{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 19206, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0, "grad_norm": 531.315111998153, "learning_rate": 3.466204506065858e-08, "loss": 5.0363, "step": 1 }, { "epoch": 0.0, "grad_norm": 292.2964725663925, "learning_rate": 6.932409012131716e-08, "loss": 4.6321, "step": 2 }, { "epoch": 0.0, "grad_norm": 264.51170744342073, "learning_rate": 1.0398613518197575e-07, "loss": 4.7518, "step": 3 }, { "epoch": 0.0, "grad_norm": 65.54403336268102, "learning_rate": 1.386481802426343e-07, "loss": 4.7551, "step": 4 }, { "epoch": 0.0, "grad_norm": 71.62633373356861, "learning_rate": 1.733102253032929e-07, "loss": 4.9667, "step": 5 }, { "epoch": 0.0, "grad_norm": 93.71744734115397, "learning_rate": 2.079722703639515e-07, "loss": 4.4864, "step": 6 }, { "epoch": 0.0, "grad_norm": 79.13891063463899, "learning_rate": 2.426343154246101e-07, "loss": 4.3112, "step": 7 }, { "epoch": 0.0, "grad_norm": 130.9824844691712, "learning_rate": 2.772963604852686e-07, "loss": 4.7113, "step": 8 }, { "epoch": 0.0, "grad_norm": 62.15746440823066, "learning_rate": 3.119584055459272e-07, "loss": 4.5315, "step": 9 }, { "epoch": 0.0, "grad_norm": 63.06252314526499, "learning_rate": 3.466204506065858e-07, "loss": 4.1675, "step": 10 }, { "epoch": 0.0, "grad_norm": 63.181877274700796, "learning_rate": 3.8128249566724436e-07, "loss": 4.7039, "step": 11 }, { "epoch": 0.0, "grad_norm": 77.56295637293104, "learning_rate": 4.15944540727903e-07, "loss": 4.7308, "step": 12 }, { "epoch": 0.0, "grad_norm": 90.59912217138084, "learning_rate": 4.5060658578856156e-07, "loss": 4.268, "step": 13 }, { "epoch": 0.0, "grad_norm": 60.27261208385014, "learning_rate": 4.852686308492202e-07, "loss": 4.0223, "step": 14 }, { "epoch": 0.0, "grad_norm": 53.092113647032356, "learning_rate": 5.199306759098788e-07, "loss": 4.0547, "step": 15 }, { "epoch": 0.0, "grad_norm": 194.42885862477203, "learning_rate": 5.545927209705372e-07, "loss": 4.1719, "step": 16 }, { "epoch": 0.0, "grad_norm": 192.72886528486674, "learning_rate": 5.89254766031196e-07, "loss": 3.9482, "step": 17 }, { "epoch": 0.0, "grad_norm": 52.923874944802805, "learning_rate": 6.239168110918544e-07, "loss": 4.0473, "step": 18 }, { "epoch": 0.0, "grad_norm": 243.21030215791566, "learning_rate": 6.58578856152513e-07, "loss": 3.7208, "step": 19 }, { "epoch": 0.0, "grad_norm": 199.76016694550418, "learning_rate": 6.932409012131716e-07, "loss": 3.5057, "step": 20 }, { "epoch": 0.0, "grad_norm": 219.9332157291184, "learning_rate": 7.279029462738301e-07, "loss": 3.8037, "step": 21 }, { "epoch": 0.0, "grad_norm": 224.8460521206362, "learning_rate": 7.625649913344887e-07, "loss": 3.6361, "step": 22 }, { "epoch": 0.0, "grad_norm": 52.96817437593409, "learning_rate": 7.972270363951473e-07, "loss": 3.451, "step": 23 }, { "epoch": 0.0, "grad_norm": 176.07245204045535, "learning_rate": 8.31889081455806e-07, "loss": 3.1902, "step": 24 }, { "epoch": 0.0, "grad_norm": 196.34398136614035, "learning_rate": 8.665511265164646e-07, "loss": 2.9131, "step": 25 }, { "epoch": 0.0, "grad_norm": 32.43285611524579, "learning_rate": 9.012131715771231e-07, "loss": 2.8942, "step": 26 }, { "epoch": 0.0, "grad_norm": 130.1069982354433, "learning_rate": 9.358752166377817e-07, "loss": 2.521, "step": 27 }, { "epoch": 0.0, "grad_norm": 30.298152068413035, "learning_rate": 9.705372616984403e-07, "loss": 2.6266, "step": 28 }, { "epoch": 0.0, "grad_norm": 23.695189025738316, "learning_rate": 1.0051993067590988e-06, "loss": 2.5869, "step": 29 }, { "epoch": 0.0, "grad_norm": 28.64249612592944, "learning_rate": 1.0398613518197575e-06, "loss": 2.4856, "step": 30 }, { "epoch": 0.0, "grad_norm": 23.675691242861983, "learning_rate": 1.074523396880416e-06, "loss": 2.4556, "step": 31 }, { "epoch": 0.0, "grad_norm": 23.53407161761756, "learning_rate": 1.1091854419410745e-06, "loss": 2.318, "step": 32 }, { "epoch": 0.01, "grad_norm": 96.18604569702083, "learning_rate": 1.1438474870017332e-06, "loss": 2.3074, "step": 33 }, { "epoch": 0.01, "grad_norm": 141.04263290664068, "learning_rate": 1.178509532062392e-06, "loss": 2.1149, "step": 34 }, { "epoch": 0.01, "grad_norm": 24.26582216230617, "learning_rate": 1.2131715771230504e-06, "loss": 2.0832, "step": 35 }, { "epoch": 0.01, "grad_norm": 28.986545105175942, "learning_rate": 1.2478336221837089e-06, "loss": 2.2623, "step": 36 }, { "epoch": 0.01, "grad_norm": 80.66247728466575, "learning_rate": 1.2824956672443676e-06, "loss": 2.1439, "step": 37 }, { "epoch": 0.01, "grad_norm": 21.463273681199404, "learning_rate": 1.317157712305026e-06, "loss": 1.9362, "step": 38 }, { "epoch": 0.01, "grad_norm": 84.87951823415129, "learning_rate": 1.3518197573656846e-06, "loss": 2.054, "step": 39 }, { "epoch": 0.01, "grad_norm": 18.529187478472057, "learning_rate": 1.3864818024263433e-06, "loss": 2.0543, "step": 40 }, { "epoch": 0.01, "grad_norm": 57.58101304987958, "learning_rate": 1.4211438474870018e-06, "loss": 1.9546, "step": 41 }, { "epoch": 0.01, "grad_norm": 59.388993477434354, "learning_rate": 1.4558058925476603e-06, "loss": 2.0146, "step": 42 }, { "epoch": 0.01, "grad_norm": 16.39079349038478, "learning_rate": 1.490467937608319e-06, "loss": 2.0469, "step": 43 }, { "epoch": 0.01, "grad_norm": 17.011224687411563, "learning_rate": 1.5251299826689774e-06, "loss": 1.8853, "step": 44 }, { "epoch": 0.01, "grad_norm": 57.65040422903312, "learning_rate": 1.5597920277296362e-06, "loss": 1.8947, "step": 45 }, { "epoch": 0.01, "grad_norm": 18.209375496951544, "learning_rate": 1.5944540727902946e-06, "loss": 1.8545, "step": 46 }, { "epoch": 0.01, "grad_norm": 44.473361066444234, "learning_rate": 1.6291161178509536e-06, "loss": 1.7729, "step": 47 }, { "epoch": 0.01, "grad_norm": 59.65041968269149, "learning_rate": 1.663778162911612e-06, "loss": 1.9303, "step": 48 }, { "epoch": 0.01, "grad_norm": 47.739525868265204, "learning_rate": 1.6984402079722705e-06, "loss": 1.8462, "step": 49 }, { "epoch": 0.01, "grad_norm": 64.66907700879658, "learning_rate": 1.7331022530329292e-06, "loss": 1.8848, "step": 50 }, { "epoch": 0.01, "grad_norm": 59.48751044620968, "learning_rate": 1.7677642980935877e-06, "loss": 1.8173, "step": 51 }, { "epoch": 0.01, "grad_norm": 14.627369497177005, "learning_rate": 1.8024263431542462e-06, "loss": 1.7882, "step": 52 }, { "epoch": 0.01, "grad_norm": 55.94019378697077, "learning_rate": 1.837088388214905e-06, "loss": 1.8156, "step": 53 }, { "epoch": 0.01, "grad_norm": 14.79310617919457, "learning_rate": 1.8717504332755634e-06, "loss": 1.7695, "step": 54 }, { "epoch": 0.01, "grad_norm": 12.858413094829237, "learning_rate": 1.906412478336222e-06, "loss": 1.6991, "step": 55 }, { "epoch": 0.01, "grad_norm": 12.585260260538933, "learning_rate": 1.9410745233968806e-06, "loss": 1.6335, "step": 56 }, { "epoch": 0.01, "grad_norm": 65.95190152730467, "learning_rate": 1.9757365684575393e-06, "loss": 1.8287, "step": 57 }, { "epoch": 0.01, "grad_norm": 28.052890780604574, "learning_rate": 2.0103986135181976e-06, "loss": 1.5994, "step": 58 }, { "epoch": 0.01, "grad_norm": 52.96283885189271, "learning_rate": 2.0450606585788563e-06, "loss": 1.7196, "step": 59 }, { "epoch": 0.01, "grad_norm": 56.70262797924363, "learning_rate": 2.079722703639515e-06, "loss": 1.7748, "step": 60 }, { "epoch": 0.01, "grad_norm": 37.195107480088836, "learning_rate": 2.1143847487001733e-06, "loss": 1.6626, "step": 61 }, { "epoch": 0.01, "grad_norm": 63.72151010848557, "learning_rate": 2.149046793760832e-06, "loss": 1.7699, "step": 62 }, { "epoch": 0.01, "grad_norm": 38.40982890234492, "learning_rate": 2.1837088388214907e-06, "loss": 1.6679, "step": 63 }, { "epoch": 0.01, "grad_norm": 43.77033326258082, "learning_rate": 2.218370883882149e-06, "loss": 1.8076, "step": 64 }, { "epoch": 0.01, "grad_norm": 39.80828537183788, "learning_rate": 2.2530329289428077e-06, "loss": 1.6574, "step": 65 }, { "epoch": 0.01, "grad_norm": 39.0141134889743, "learning_rate": 2.2876949740034664e-06, "loss": 1.7012, "step": 66 }, { "epoch": 0.01, "grad_norm": 8.550052719832058, "learning_rate": 2.322357019064125e-06, "loss": 1.495, "step": 67 }, { "epoch": 0.01, "grad_norm": 42.31678881948093, "learning_rate": 2.357019064124784e-06, "loss": 1.6645, "step": 68 }, { "epoch": 0.01, "grad_norm": 44.22241947059129, "learning_rate": 2.391681109185442e-06, "loss": 1.6002, "step": 69 }, { "epoch": 0.01, "grad_norm": 11.846153629350137, "learning_rate": 2.4263431542461008e-06, "loss": 1.7184, "step": 70 }, { "epoch": 0.01, "grad_norm": 46.693782384986555, "learning_rate": 2.4610051993067595e-06, "loss": 1.5383, "step": 71 }, { "epoch": 0.01, "grad_norm": 65.46057084460385, "learning_rate": 2.4956672443674178e-06, "loss": 1.5847, "step": 72 }, { "epoch": 0.01, "grad_norm": 48.29185609030196, "learning_rate": 2.530329289428076e-06, "loss": 1.609, "step": 73 }, { "epoch": 0.01, "grad_norm": 23.895269717082964, "learning_rate": 2.564991334488735e-06, "loss": 1.4175, "step": 74 }, { "epoch": 0.01, "grad_norm": 45.28353952262731, "learning_rate": 2.599653379549394e-06, "loss": 1.5396, "step": 75 }, { "epoch": 0.01, "grad_norm": 41.48833822367313, "learning_rate": 2.634315424610052e-06, "loss": 1.537, "step": 76 }, { "epoch": 0.01, "grad_norm": 33.32863144952529, "learning_rate": 2.668977469670711e-06, "loss": 1.3886, "step": 77 }, { "epoch": 0.01, "grad_norm": 34.59213100046043, "learning_rate": 2.703639514731369e-06, "loss": 1.6539, "step": 78 }, { "epoch": 0.01, "grad_norm": 31.78870746070221, "learning_rate": 2.7383015597920283e-06, "loss": 1.5099, "step": 79 }, { "epoch": 0.01, "grad_norm": 32.31520248932208, "learning_rate": 2.7729636048526865e-06, "loss": 1.5438, "step": 80 }, { "epoch": 0.01, "grad_norm": 42.59625680645732, "learning_rate": 2.8076256499133452e-06, "loss": 1.5118, "step": 81 }, { "epoch": 0.01, "grad_norm": 56.6037912144551, "learning_rate": 2.8422876949740035e-06, "loss": 1.6787, "step": 82 }, { "epoch": 0.01, "grad_norm": 35.51296773654188, "learning_rate": 2.8769497400346622e-06, "loss": 1.6533, "step": 83 }, { "epoch": 0.01, "grad_norm": 65.8743678692378, "learning_rate": 2.9116117850953205e-06, "loss": 1.4466, "step": 84 }, { "epoch": 0.01, "grad_norm": 7.660212235000567, "learning_rate": 2.9462738301559796e-06, "loss": 1.5002, "step": 85 }, { "epoch": 0.01, "grad_norm": 8.080726230549912, "learning_rate": 2.980935875216638e-06, "loss": 1.5862, "step": 86 }, { "epoch": 0.01, "grad_norm": 36.809914237432615, "learning_rate": 3.0155979202772966e-06, "loss": 1.4045, "step": 87 }, { "epoch": 0.01, "grad_norm": 26.38411280641693, "learning_rate": 3.050259965337955e-06, "loss": 1.4151, "step": 88 }, { "epoch": 0.01, "grad_norm": 31.09621377509721, "learning_rate": 3.084922010398614e-06, "loss": 1.3729, "step": 89 }, { "epoch": 0.01, "grad_norm": 56.80399612113042, "learning_rate": 3.1195840554592723e-06, "loss": 1.4924, "step": 90 }, { "epoch": 0.01, "grad_norm": 33.03162578163297, "learning_rate": 3.154246100519931e-06, "loss": 1.2896, "step": 91 }, { "epoch": 0.01, "grad_norm": 46.05831648463537, "learning_rate": 3.1889081455805893e-06, "loss": 1.5209, "step": 92 }, { "epoch": 0.01, "grad_norm": 35.381577279632964, "learning_rate": 3.223570190641248e-06, "loss": 1.2987, "step": 93 }, { "epoch": 0.01, "grad_norm": 50.67155855969887, "learning_rate": 3.258232235701907e-06, "loss": 1.4132, "step": 94 }, { "epoch": 0.01, "grad_norm": 27.14601135720065, "learning_rate": 3.2928942807625654e-06, "loss": 1.3058, "step": 95 }, { "epoch": 0.01, "grad_norm": 52.753150739630186, "learning_rate": 3.327556325823224e-06, "loss": 1.5529, "step": 96 }, { "epoch": 0.02, "grad_norm": 41.63375309679603, "learning_rate": 3.3622183708838824e-06, "loss": 1.3607, "step": 97 }, { "epoch": 0.02, "grad_norm": 38.060557282953845, "learning_rate": 3.396880415944541e-06, "loss": 1.5151, "step": 98 }, { "epoch": 0.02, "grad_norm": 36.341843029383874, "learning_rate": 3.4315424610051994e-06, "loss": 1.4832, "step": 99 }, { "epoch": 0.02, "grad_norm": 32.71183195182694, "learning_rate": 3.4662045060658585e-06, "loss": 1.3898, "step": 100 }, { "epoch": 0.02, "grad_norm": 35.364513544716985, "learning_rate": 3.5008665511265168e-06, "loss": 1.2867, "step": 101 }, { "epoch": 0.02, "grad_norm": 34.96477526522823, "learning_rate": 3.5355285961871755e-06, "loss": 1.3614, "step": 102 }, { "epoch": 0.02, "grad_norm": 54.31042409667524, "learning_rate": 3.5701906412478338e-06, "loss": 1.3225, "step": 103 }, { "epoch": 0.02, "grad_norm": 37.162779470946084, "learning_rate": 3.6048526863084925e-06, "loss": 1.4249, "step": 104 }, { "epoch": 0.02, "grad_norm": 29.29934761904572, "learning_rate": 3.6395147313691507e-06, "loss": 1.2458, "step": 105 }, { "epoch": 0.02, "grad_norm": 34.47241906103389, "learning_rate": 3.67417677642981e-06, "loss": 1.4181, "step": 106 }, { "epoch": 0.02, "grad_norm": 42.256401884767726, "learning_rate": 3.708838821490468e-06, "loss": 1.4879, "step": 107 }, { "epoch": 0.02, "grad_norm": 43.925025114520146, "learning_rate": 3.743500866551127e-06, "loss": 1.3717, "step": 108 }, { "epoch": 0.02, "grad_norm": 29.95344397624258, "learning_rate": 3.778162911611785e-06, "loss": 1.2605, "step": 109 }, { "epoch": 0.02, "grad_norm": 44.758083549766795, "learning_rate": 3.812824956672444e-06, "loss": 1.4066, "step": 110 }, { "epoch": 0.02, "grad_norm": 53.9073924565535, "learning_rate": 3.8474870017331025e-06, "loss": 1.3514, "step": 111 }, { "epoch": 0.02, "grad_norm": 41.10458204572215, "learning_rate": 3.882149046793761e-06, "loss": 1.2649, "step": 112 }, { "epoch": 0.02, "grad_norm": 53.05955697604764, "learning_rate": 3.91681109185442e-06, "loss": 1.4632, "step": 113 }, { "epoch": 0.02, "grad_norm": 10.46621436256183, "learning_rate": 3.951473136915079e-06, "loss": 1.4173, "step": 114 }, { "epoch": 0.02, "grad_norm": 9.335895509505711, "learning_rate": 3.986135181975737e-06, "loss": 1.2472, "step": 115 }, { "epoch": 0.02, "grad_norm": 53.528985041276094, "learning_rate": 4.020797227036395e-06, "loss": 1.3109, "step": 116 }, { "epoch": 0.02, "grad_norm": 44.74688266972805, "learning_rate": 4.055459272097054e-06, "loss": 1.5135, "step": 117 }, { "epoch": 0.02, "grad_norm": 39.26028178330096, "learning_rate": 4.090121317157713e-06, "loss": 1.3563, "step": 118 }, { "epoch": 0.02, "grad_norm": 44.316768802456025, "learning_rate": 4.124783362218371e-06, "loss": 1.3292, "step": 119 }, { "epoch": 0.02, "grad_norm": 29.574670927485485, "learning_rate": 4.15944540727903e-06, "loss": 1.1467, "step": 120 }, { "epoch": 0.02, "grad_norm": 45.81860769341073, "learning_rate": 4.194107452339689e-06, "loss": 1.3843, "step": 121 }, { "epoch": 0.02, "grad_norm": 57.43559259894182, "learning_rate": 4.228769497400347e-06, "loss": 1.4919, "step": 122 }, { "epoch": 0.02, "grad_norm": 5.415087089635171, "learning_rate": 4.263431542461005e-06, "loss": 1.1586, "step": 123 }, { "epoch": 0.02, "grad_norm": 13.678235459508098, "learning_rate": 4.298093587521664e-06, "loss": 1.3752, "step": 124 }, { "epoch": 0.02, "grad_norm": 44.934836609796854, "learning_rate": 4.332755632582323e-06, "loss": 1.294, "step": 125 }, { "epoch": 0.02, "grad_norm": 33.2067769266012, "learning_rate": 4.367417677642981e-06, "loss": 1.2223, "step": 126 }, { "epoch": 0.02, "grad_norm": 31.158718020945006, "learning_rate": 4.40207972270364e-06, "loss": 1.2844, "step": 127 }, { "epoch": 0.02, "grad_norm": 33.88473113813146, "learning_rate": 4.436741767764298e-06, "loss": 1.3532, "step": 128 }, { "epoch": 0.02, "grad_norm": 35.647074162323214, "learning_rate": 4.471403812824957e-06, "loss": 1.3752, "step": 129 }, { "epoch": 0.02, "grad_norm": 36.23125015937893, "learning_rate": 4.506065857885615e-06, "loss": 1.2189, "step": 130 }, { "epoch": 0.02, "grad_norm": 41.133569313035245, "learning_rate": 4.540727902946274e-06, "loss": 1.4709, "step": 131 }, { "epoch": 0.02, "grad_norm": 36.33225915991336, "learning_rate": 4.575389948006933e-06, "loss": 1.4186, "step": 132 }, { "epoch": 0.02, "grad_norm": 35.414972508619215, "learning_rate": 4.6100519930675915e-06, "loss": 1.3723, "step": 133 }, { "epoch": 0.02, "grad_norm": 33.35428797151585, "learning_rate": 4.64471403812825e-06, "loss": 1.3848, "step": 134 }, { "epoch": 0.02, "grad_norm": 39.34666713235231, "learning_rate": 4.679376083188908e-06, "loss": 1.3439, "step": 135 }, { "epoch": 0.02, "grad_norm": 29.078741925934583, "learning_rate": 4.714038128249568e-06, "loss": 1.3974, "step": 136 }, { "epoch": 0.02, "grad_norm": 26.90329138232109, "learning_rate": 4.7487001733102254e-06, "loss": 1.1737, "step": 137 }, { "epoch": 0.02, "grad_norm": 7.292796210039388, "learning_rate": 4.783362218370884e-06, "loss": 1.3312, "step": 138 }, { "epoch": 0.02, "grad_norm": 37.7018850322623, "learning_rate": 4.818024263431543e-06, "loss": 1.2078, "step": 139 }, { "epoch": 0.02, "grad_norm": 45.284279681800335, "learning_rate": 4.8526863084922016e-06, "loss": 1.2895, "step": 140 }, { "epoch": 0.02, "grad_norm": 51.891835734719514, "learning_rate": 4.88734835355286e-06, "loss": 1.4174, "step": 141 }, { "epoch": 0.02, "grad_norm": 22.47660619461644, "learning_rate": 4.922010398613519e-06, "loss": 1.2463, "step": 142 }, { "epoch": 0.02, "grad_norm": 29.76010427665964, "learning_rate": 4.956672443674177e-06, "loss": 1.381, "step": 143 }, { "epoch": 0.02, "grad_norm": 35.70995866748017, "learning_rate": 4.9913344887348355e-06, "loss": 1.3176, "step": 144 }, { "epoch": 0.02, "grad_norm": 39.20985658428473, "learning_rate": 5.025996533795494e-06, "loss": 1.2154, "step": 145 }, { "epoch": 0.02, "grad_norm": 39.78268072622544, "learning_rate": 5.060658578856152e-06, "loss": 1.3954, "step": 146 }, { "epoch": 0.02, "grad_norm": 37.29081364869007, "learning_rate": 5.095320623916812e-06, "loss": 1.2138, "step": 147 }, { "epoch": 0.02, "grad_norm": 42.82941003268194, "learning_rate": 5.12998266897747e-06, "loss": 1.2528, "step": 148 }, { "epoch": 0.02, "grad_norm": 29.308513069270575, "learning_rate": 5.164644714038128e-06, "loss": 1.3094, "step": 149 }, { "epoch": 0.02, "grad_norm": 37.716104024319435, "learning_rate": 5.199306759098788e-06, "loss": 1.197, "step": 150 }, { "epoch": 0.02, "grad_norm": 44.569028238421126, "learning_rate": 5.2339688041594464e-06, "loss": 1.2321, "step": 151 }, { "epoch": 0.02, "grad_norm": 30.039305431241836, "learning_rate": 5.268630849220104e-06, "loss": 1.225, "step": 152 }, { "epoch": 0.02, "grad_norm": 30.957562751900497, "learning_rate": 5.303292894280763e-06, "loss": 1.2773, "step": 153 }, { "epoch": 0.02, "grad_norm": 28.24003345089031, "learning_rate": 5.337954939341422e-06, "loss": 1.1255, "step": 154 }, { "epoch": 0.02, "grad_norm": 7.444971755777517, "learning_rate": 5.37261698440208e-06, "loss": 1.1667, "step": 155 }, { "epoch": 0.02, "grad_norm": 46.517308003223164, "learning_rate": 5.407279029462738e-06, "loss": 1.3639, "step": 156 }, { "epoch": 0.02, "grad_norm": 44.84407460430838, "learning_rate": 5.441941074523397e-06, "loss": 1.3066, "step": 157 }, { "epoch": 0.02, "grad_norm": 27.237931197168404, "learning_rate": 5.4766031195840565e-06, "loss": 1.2181, "step": 158 }, { "epoch": 0.02, "grad_norm": 42.470842864262124, "learning_rate": 5.511265164644714e-06, "loss": 1.2999, "step": 159 }, { "epoch": 0.02, "grad_norm": 51.05946680302711, "learning_rate": 5.545927209705373e-06, "loss": 1.4571, "step": 160 }, { "epoch": 0.03, "grad_norm": 22.331443568035873, "learning_rate": 5.580589254766031e-06, "loss": 1.17, "step": 161 }, { "epoch": 0.03, "grad_norm": 31.660909220397603, "learning_rate": 5.6152512998266905e-06, "loss": 1.1699, "step": 162 }, { "epoch": 0.03, "grad_norm": 39.83129881241251, "learning_rate": 5.649913344887349e-06, "loss": 1.3412, "step": 163 }, { "epoch": 0.03, "grad_norm": 41.582122805500425, "learning_rate": 5.684575389948007e-06, "loss": 1.2571, "step": 164 }, { "epoch": 0.03, "grad_norm": 34.92599689727244, "learning_rate": 5.719237435008666e-06, "loss": 1.3948, "step": 165 }, { "epoch": 0.03, "grad_norm": 35.588642926181166, "learning_rate": 5.7538994800693245e-06, "loss": 1.3808, "step": 166 }, { "epoch": 0.03, "grad_norm": 32.375995972310854, "learning_rate": 5.788561525129983e-06, "loss": 1.2056, "step": 167 }, { "epoch": 0.03, "grad_norm": 7.049289160851054, "learning_rate": 5.823223570190641e-06, "loss": 1.1249, "step": 168 }, { "epoch": 0.03, "grad_norm": 30.980834700967822, "learning_rate": 5.8578856152513006e-06, "loss": 1.2299, "step": 169 }, { "epoch": 0.03, "grad_norm": 7.81297797838982, "learning_rate": 5.892547660311959e-06, "loss": 1.2217, "step": 170 }, { "epoch": 0.03, "grad_norm": 24.62331531360481, "learning_rate": 5.927209705372617e-06, "loss": 1.1349, "step": 171 }, { "epoch": 0.03, "grad_norm": 38.67789488822055, "learning_rate": 5.961871750433276e-06, "loss": 1.1445, "step": 172 }, { "epoch": 0.03, "grad_norm": 36.04467601914481, "learning_rate": 5.996533795493935e-06, "loss": 1.234, "step": 173 }, { "epoch": 0.03, "grad_norm": 39.50498866391282, "learning_rate": 6.031195840554593e-06, "loss": 1.3125, "step": 174 }, { "epoch": 0.03, "grad_norm": 25.497032026697326, "learning_rate": 6.065857885615252e-06, "loss": 1.3066, "step": 175 }, { "epoch": 0.03, "grad_norm": 33.757574274360955, "learning_rate": 6.10051993067591e-06, "loss": 1.1588, "step": 176 }, { "epoch": 0.03, "grad_norm": 37.29514032396307, "learning_rate": 6.135181975736569e-06, "loss": 1.2298, "step": 177 }, { "epoch": 0.03, "grad_norm": 40.78940463621289, "learning_rate": 6.169844020797228e-06, "loss": 1.3352, "step": 178 }, { "epoch": 0.03, "grad_norm": 23.524942823101373, "learning_rate": 6.204506065857886e-06, "loss": 1.1822, "step": 179 }, { "epoch": 0.03, "grad_norm": 36.69774783822274, "learning_rate": 6.239168110918545e-06, "loss": 1.2495, "step": 180 }, { "epoch": 0.03, "grad_norm": 41.3084844464238, "learning_rate": 6.273830155979203e-06, "loss": 1.1862, "step": 181 }, { "epoch": 0.03, "grad_norm": 34.43259940959318, "learning_rate": 6.308492201039862e-06, "loss": 1.2432, "step": 182 }, { "epoch": 0.03, "grad_norm": 38.24744721071966, "learning_rate": 6.34315424610052e-06, "loss": 1.1918, "step": 183 }, { "epoch": 0.03, "grad_norm": 41.74737424574073, "learning_rate": 6.377816291161179e-06, "loss": 1.1328, "step": 184 }, { "epoch": 0.03, "grad_norm": 25.256152917656, "learning_rate": 6.412478336221838e-06, "loss": 1.1292, "step": 185 }, { "epoch": 0.03, "grad_norm": 10.65204682336752, "learning_rate": 6.447140381282496e-06, "loss": 1.3824, "step": 186 }, { "epoch": 0.03, "grad_norm": 35.39689166268264, "learning_rate": 6.481802426343155e-06, "loss": 1.2992, "step": 187 }, { "epoch": 0.03, "grad_norm": 43.78390060477098, "learning_rate": 6.516464471403814e-06, "loss": 1.3357, "step": 188 }, { "epoch": 0.03, "grad_norm": 29.22914734988391, "learning_rate": 6.551126516464472e-06, "loss": 1.225, "step": 189 }, { "epoch": 0.03, "grad_norm": 23.597944332714018, "learning_rate": 6.585788561525131e-06, "loss": 1.2756, "step": 190 }, { "epoch": 0.03, "grad_norm": 419.35612647258307, "learning_rate": 6.620450606585789e-06, "loss": 1.2129, "step": 191 }, { "epoch": 0.03, "grad_norm": 46.917815459962284, "learning_rate": 6.655112651646448e-06, "loss": 1.2689, "step": 192 }, { "epoch": 0.03, "grad_norm": 31.56025284047614, "learning_rate": 6.689774696707106e-06, "loss": 1.1644, "step": 193 }, { "epoch": 0.03, "grad_norm": 6.313856039421796, "learning_rate": 6.724436741767765e-06, "loss": 1.1856, "step": 194 }, { "epoch": 0.03, "grad_norm": 30.530156384407103, "learning_rate": 6.759098786828423e-06, "loss": 1.2193, "step": 195 }, { "epoch": 0.03, "grad_norm": 36.97730140906862, "learning_rate": 6.793760831889082e-06, "loss": 1.177, "step": 196 }, { "epoch": 0.03, "grad_norm": 7.392897250545163, "learning_rate": 6.828422876949741e-06, "loss": 1.231, "step": 197 }, { "epoch": 0.03, "grad_norm": 32.92835037394705, "learning_rate": 6.863084922010399e-06, "loss": 1.146, "step": 198 }, { "epoch": 0.03, "grad_norm": 19.552219517077233, "learning_rate": 6.8977469670710574e-06, "loss": 1.0737, "step": 199 }, { "epoch": 0.03, "grad_norm": 26.903193126268295, "learning_rate": 6.932409012131717e-06, "loss": 1.1864, "step": 200 }, { "epoch": 0.03, "grad_norm": 38.42820331679975, "learning_rate": 6.967071057192375e-06, "loss": 1.1963, "step": 201 }, { "epoch": 0.03, "grad_norm": 6.780253042602195, "learning_rate": 7.0017331022530336e-06, "loss": 1.2246, "step": 202 }, { "epoch": 0.03, "grad_norm": 41.88507083311122, "learning_rate": 7.036395147313691e-06, "loss": 1.253, "step": 203 }, { "epoch": 0.03, "grad_norm": 28.588234394474906, "learning_rate": 7.071057192374351e-06, "loss": 1.1283, "step": 204 }, { "epoch": 0.03, "grad_norm": 7.2968956254753365, "learning_rate": 7.10571923743501e-06, "loss": 1.2647, "step": 205 }, { "epoch": 0.03, "grad_norm": 6.7423780177881625, "learning_rate": 7.1403812824956675e-06, "loss": 1.1301, "step": 206 }, { "epoch": 0.03, "grad_norm": 40.2782497605626, "learning_rate": 7.175043327556327e-06, "loss": 1.416, "step": 207 }, { "epoch": 0.03, "grad_norm": 32.16224681685655, "learning_rate": 7.209705372616985e-06, "loss": 1.1316, "step": 208 }, { "epoch": 0.03, "grad_norm": 36.03994387862284, "learning_rate": 7.244367417677644e-06, "loss": 1.0575, "step": 209 }, { "epoch": 0.03, "grad_norm": 45.38313356254118, "learning_rate": 7.2790294627383015e-06, "loss": 1.2639, "step": 210 }, { "epoch": 0.03, "grad_norm": 29.74022686936235, "learning_rate": 7.313691507798961e-06, "loss": 1.2833, "step": 211 }, { "epoch": 0.03, "grad_norm": 36.19193833892053, "learning_rate": 7.34835355285962e-06, "loss": 1.271, "step": 212 }, { "epoch": 0.03, "grad_norm": 32.62994222002236, "learning_rate": 7.383015597920278e-06, "loss": 1.1843, "step": 213 }, { "epoch": 0.03, "grad_norm": 36.961412282025826, "learning_rate": 7.417677642980936e-06, "loss": 1.1769, "step": 214 }, { "epoch": 0.03, "grad_norm": 20.229385803664805, "learning_rate": 7.452339688041596e-06, "loss": 1.1415, "step": 215 }, { "epoch": 0.03, "grad_norm": 49.63053238151394, "learning_rate": 7.487001733102254e-06, "loss": 1.2606, "step": 216 }, { "epoch": 0.03, "grad_norm": 36.18979464060009, "learning_rate": 7.521663778162912e-06, "loss": 1.0297, "step": 217 }, { "epoch": 0.03, "grad_norm": 37.33706410383136, "learning_rate": 7.55632582322357e-06, "loss": 1.2468, "step": 218 }, { "epoch": 0.03, "grad_norm": 37.466290187188605, "learning_rate": 7.59098786828423e-06, "loss": 1.2497, "step": 219 }, { "epoch": 0.03, "grad_norm": 31.785716426582937, "learning_rate": 7.625649913344888e-06, "loss": 1.148, "step": 220 }, { "epoch": 0.03, "grad_norm": 6.186234469992598, "learning_rate": 7.660311958405546e-06, "loss": 1.1191, "step": 221 }, { "epoch": 0.03, "grad_norm": 21.136424431825276, "learning_rate": 7.694974003466205e-06, "loss": 1.1053, "step": 222 }, { "epoch": 0.03, "grad_norm": 36.5331661009021, "learning_rate": 7.729636048526865e-06, "loss": 1.1752, "step": 223 }, { "epoch": 0.03, "grad_norm": 11.825170021241153, "learning_rate": 7.764298093587522e-06, "loss": 1.1805, "step": 224 }, { "epoch": 0.04, "grad_norm": 49.98371478649682, "learning_rate": 7.79896013864818e-06, "loss": 1.2036, "step": 225 }, { "epoch": 0.04, "grad_norm": 25.76367878778229, "learning_rate": 7.83362218370884e-06, "loss": 1.1512, "step": 226 }, { "epoch": 0.04, "grad_norm": 6.343654337053652, "learning_rate": 7.868284228769498e-06, "loss": 1.2117, "step": 227 }, { "epoch": 0.04, "grad_norm": 33.74923088998774, "learning_rate": 7.902946273830157e-06, "loss": 1.1138, "step": 228 }, { "epoch": 0.04, "grad_norm": 36.786056645264466, "learning_rate": 7.937608318890815e-06, "loss": 1.1355, "step": 229 }, { "epoch": 0.04, "grad_norm": 48.2760855668748, "learning_rate": 7.972270363951475e-06, "loss": 1.2969, "step": 230 }, { "epoch": 0.04, "grad_norm": 30.061946352264094, "learning_rate": 8.006932409012133e-06, "loss": 1.2115, "step": 231 }, { "epoch": 0.04, "grad_norm": 28.30631954006009, "learning_rate": 8.04159445407279e-06, "loss": 1.0828, "step": 232 }, { "epoch": 0.04, "grad_norm": 36.30636139112351, "learning_rate": 8.076256499133448e-06, "loss": 1.2671, "step": 233 }, { "epoch": 0.04, "grad_norm": 25.827200155898385, "learning_rate": 8.110918544194108e-06, "loss": 1.2521, "step": 234 }, { "epoch": 0.04, "grad_norm": 30.46692480831601, "learning_rate": 8.145580589254767e-06, "loss": 1.1863, "step": 235 }, { "epoch": 0.04, "grad_norm": 33.95893658100235, "learning_rate": 8.180242634315425e-06, "loss": 1.1368, "step": 236 }, { "epoch": 0.04, "grad_norm": 6.481549417206551, "learning_rate": 8.214904679376083e-06, "loss": 1.2642, "step": 237 }, { "epoch": 0.04, "grad_norm": 26.55777034389227, "learning_rate": 8.249566724436743e-06, "loss": 1.1898, "step": 238 }, { "epoch": 0.04, "grad_norm": 31.537696253165546, "learning_rate": 8.2842287694974e-06, "loss": 1.1685, "step": 239 }, { "epoch": 0.04, "grad_norm": 40.98763564530311, "learning_rate": 8.31889081455806e-06, "loss": 1.1621, "step": 240 }, { "epoch": 0.04, "grad_norm": 22.768020223733394, "learning_rate": 8.353552859618718e-06, "loss": 1.0423, "step": 241 }, { "epoch": 0.04, "grad_norm": 30.991154264281352, "learning_rate": 8.388214904679377e-06, "loss": 1.1565, "step": 242 }, { "epoch": 0.04, "grad_norm": 35.33399157297096, "learning_rate": 8.422876949740035e-06, "loss": 1.1692, "step": 243 }, { "epoch": 0.04, "grad_norm": 30.435295857435555, "learning_rate": 8.457538994800693e-06, "loss": 1.0559, "step": 244 }, { "epoch": 0.04, "grad_norm": 26.817123491290523, "learning_rate": 8.492201039861353e-06, "loss": 1.1404, "step": 245 }, { "epoch": 0.04, "grad_norm": 35.386770881490634, "learning_rate": 8.52686308492201e-06, "loss": 1.2227, "step": 246 }, { "epoch": 0.04, "grad_norm": 45.657356716165744, "learning_rate": 8.56152512998267e-06, "loss": 1.2336, "step": 247 }, { "epoch": 0.04, "grad_norm": 8.792431903517928, "learning_rate": 8.596187175043328e-06, "loss": 1.1526, "step": 248 }, { "epoch": 0.04, "grad_norm": 43.45146147895134, "learning_rate": 8.630849220103988e-06, "loss": 1.3344, "step": 249 }, { "epoch": 0.04, "grad_norm": 40.00593644734271, "learning_rate": 8.665511265164645e-06, "loss": 1.3203, "step": 250 }, { "epoch": 0.04, "grad_norm": 27.483074029490002, "learning_rate": 8.700173310225303e-06, "loss": 1.1013, "step": 251 }, { "epoch": 0.04, "grad_norm": 32.86979396650612, "learning_rate": 8.734835355285963e-06, "loss": 1.2232, "step": 252 }, { "epoch": 0.04, "grad_norm": 61.6369277483482, "learning_rate": 8.769497400346622e-06, "loss": 1.3066, "step": 253 }, { "epoch": 0.04, "grad_norm": 32.427511507502366, "learning_rate": 8.80415944540728e-06, "loss": 1.0331, "step": 254 }, { "epoch": 0.04, "grad_norm": 32.41941191543784, "learning_rate": 8.838821490467938e-06, "loss": 1.193, "step": 255 }, { "epoch": 0.04, "grad_norm": 32.89093912658737, "learning_rate": 8.873483535528596e-06, "loss": 1.188, "step": 256 }, { "epoch": 0.04, "grad_norm": 6.7963108025890575, "learning_rate": 8.908145580589255e-06, "loss": 1.0219, "step": 257 }, { "epoch": 0.04, "grad_norm": 39.48217111225304, "learning_rate": 8.942807625649913e-06, "loss": 1.287, "step": 258 }, { "epoch": 0.04, "grad_norm": 17.591255719236734, "learning_rate": 8.977469670710573e-06, "loss": 0.993, "step": 259 }, { "epoch": 0.04, "grad_norm": 30.271428492673127, "learning_rate": 9.01213171577123e-06, "loss": 1.0541, "step": 260 }, { "epoch": 0.04, "grad_norm": 20.960546409549913, "learning_rate": 9.04679376083189e-06, "loss": 1.114, "step": 261 }, { "epoch": 0.04, "grad_norm": 23.37091576688421, "learning_rate": 9.081455805892548e-06, "loss": 1.0278, "step": 262 }, { "epoch": 0.04, "grad_norm": 32.44645666671875, "learning_rate": 9.116117850953206e-06, "loss": 1.0882, "step": 263 }, { "epoch": 0.04, "grad_norm": 29.394409995098258, "learning_rate": 9.150779896013866e-06, "loss": 1.098, "step": 264 }, { "epoch": 0.04, "grad_norm": 32.84000688314728, "learning_rate": 9.185441941074525e-06, "loss": 1.1333, "step": 265 }, { "epoch": 0.04, "grad_norm": 31.90842895091659, "learning_rate": 9.220103986135183e-06, "loss": 1.1737, "step": 266 }, { "epoch": 0.04, "grad_norm": 29.000285479151675, "learning_rate": 9.25476603119584e-06, "loss": 1.1174, "step": 267 }, { "epoch": 0.04, "grad_norm": 39.56155290117829, "learning_rate": 9.2894280762565e-06, "loss": 1.1217, "step": 268 }, { "epoch": 0.04, "grad_norm": 133.62511300634418, "learning_rate": 9.324090121317158e-06, "loss": 1.2347, "step": 269 }, { "epoch": 0.04, "grad_norm": 28.648703538269523, "learning_rate": 9.358752166377816e-06, "loss": 1.2441, "step": 270 }, { "epoch": 0.04, "grad_norm": 26.332707601116592, "learning_rate": 9.393414211438476e-06, "loss": 1.1396, "step": 271 }, { "epoch": 0.04, "grad_norm": 22.044852545885977, "learning_rate": 9.428076256499135e-06, "loss": 1.2006, "step": 272 }, { "epoch": 0.04, "grad_norm": 6.57984200583586, "learning_rate": 9.462738301559793e-06, "loss": 1.1745, "step": 273 }, { "epoch": 0.04, "grad_norm": 37.63920262944199, "learning_rate": 9.497400346620451e-06, "loss": 1.1589, "step": 274 }, { "epoch": 0.04, "grad_norm": 20.540283066738453, "learning_rate": 9.532062391681109e-06, "loss": 0.986, "step": 275 }, { "epoch": 0.04, "grad_norm": 39.58208785113726, "learning_rate": 9.566724436741768e-06, "loss": 1.2153, "step": 276 }, { "epoch": 0.04, "grad_norm": 7.028840029223187, "learning_rate": 9.601386481802428e-06, "loss": 1.179, "step": 277 }, { "epoch": 0.04, "grad_norm": 27.1549390118637, "learning_rate": 9.636048526863086e-06, "loss": 1.1244, "step": 278 }, { "epoch": 0.04, "grad_norm": 36.940692187597406, "learning_rate": 9.670710571923744e-06, "loss": 1.3229, "step": 279 }, { "epoch": 0.04, "grad_norm": 39.666642069642, "learning_rate": 9.705372616984403e-06, "loss": 1.1767, "step": 280 }, { "epoch": 0.04, "grad_norm": 47.50090416624944, "learning_rate": 9.740034662045061e-06, "loss": 1.2691, "step": 281 }, { "epoch": 0.04, "grad_norm": 19.807914091835052, "learning_rate": 9.77469670710572e-06, "loss": 1.0357, "step": 282 }, { "epoch": 0.04, "grad_norm": 22.087099286974343, "learning_rate": 9.809358752166378e-06, "loss": 1.208, "step": 283 }, { "epoch": 0.04, "grad_norm": 36.14497004182539, "learning_rate": 9.844020797227038e-06, "loss": 1.1656, "step": 284 }, { "epoch": 0.04, "grad_norm": 30.562239852531384, "learning_rate": 9.878682842287696e-06, "loss": 1.2537, "step": 285 }, { "epoch": 0.04, "grad_norm": 28.507637794902266, "learning_rate": 9.913344887348354e-06, "loss": 1.0488, "step": 286 }, { "epoch": 0.04, "grad_norm": 30.378105736734124, "learning_rate": 9.948006932409013e-06, "loss": 1.222, "step": 287 }, { "epoch": 0.04, "grad_norm": 33.37978941506294, "learning_rate": 9.982668977469671e-06, "loss": 1.1733, "step": 288 }, { "epoch": 0.05, "grad_norm": 32.412566171836765, "learning_rate": 1.001733102253033e-05, "loss": 1.2059, "step": 289 }, { "epoch": 0.05, "grad_norm": 28.658334281858885, "learning_rate": 1.0051993067590988e-05, "loss": 1.1613, "step": 290 }, { "epoch": 0.05, "grad_norm": 34.339859763827214, "learning_rate": 1.0086655112651646e-05, "loss": 1.2065, "step": 291 }, { "epoch": 0.05, "grad_norm": 20.60806040298398, "learning_rate": 1.0121317157712304e-05, "loss": 1.1927, "step": 292 }, { "epoch": 0.05, "grad_norm": 28.020750138656787, "learning_rate": 1.0155979202772965e-05, "loss": 1.1303, "step": 293 }, { "epoch": 0.05, "grad_norm": 32.55648462321204, "learning_rate": 1.0190641247833623e-05, "loss": 1.2168, "step": 294 }, { "epoch": 0.05, "grad_norm": 28.584010805483356, "learning_rate": 1.0225303292894281e-05, "loss": 1.1082, "step": 295 }, { "epoch": 0.05, "grad_norm": 27.66448217384437, "learning_rate": 1.025996533795494e-05, "loss": 1.0592, "step": 296 }, { "epoch": 0.05, "grad_norm": 33.284408458864455, "learning_rate": 1.0294627383015599e-05, "loss": 1.0929, "step": 297 }, { "epoch": 0.05, "grad_norm": 31.98890294225414, "learning_rate": 1.0329289428076256e-05, "loss": 1.169, "step": 298 }, { "epoch": 0.05, "grad_norm": 30.00410883601317, "learning_rate": 1.0363951473136914e-05, "loss": 1.058, "step": 299 }, { "epoch": 0.05, "grad_norm": 27.910588839190368, "learning_rate": 1.0398613518197575e-05, "loss": 1.1519, "step": 300 }, { "epoch": 0.05, "grad_norm": 36.227928446836465, "learning_rate": 1.0433275563258233e-05, "loss": 1.0383, "step": 301 }, { "epoch": 0.05, "grad_norm": 26.57893303187256, "learning_rate": 1.0467937608318893e-05, "loss": 1.2146, "step": 302 }, { "epoch": 0.05, "grad_norm": 26.806770479481848, "learning_rate": 1.050259965337955e-05, "loss": 1.1068, "step": 303 }, { "epoch": 0.05, "grad_norm": 26.19032893913023, "learning_rate": 1.0537261698440209e-05, "loss": 1.0901, "step": 304 }, { "epoch": 0.05, "grad_norm": 28.416192711982273, "learning_rate": 1.0571923743500866e-05, "loss": 1.0974, "step": 305 }, { "epoch": 0.05, "grad_norm": 23.870990538838257, "learning_rate": 1.0606585788561526e-05, "loss": 1.0428, "step": 306 }, { "epoch": 0.05, "grad_norm": 26.31028593040287, "learning_rate": 1.0641247833622184e-05, "loss": 1.0996, "step": 307 }, { "epoch": 0.05, "grad_norm": 20.12638582913621, "learning_rate": 1.0675909878682843e-05, "loss": 1.0697, "step": 308 }, { "epoch": 0.05, "grad_norm": 26.229273797125238, "learning_rate": 1.0710571923743503e-05, "loss": 1.1433, "step": 309 }, { "epoch": 0.05, "grad_norm": 26.102329681320533, "learning_rate": 1.074523396880416e-05, "loss": 1.105, "step": 310 }, { "epoch": 0.05, "grad_norm": 26.304508250537534, "learning_rate": 1.0779896013864819e-05, "loss": 1.1613, "step": 311 }, { "epoch": 0.05, "grad_norm": 45.92040450216617, "learning_rate": 1.0814558058925477e-05, "loss": 1.1904, "step": 312 }, { "epoch": 0.05, "grad_norm": 26.0916480782526, "learning_rate": 1.0849220103986136e-05, "loss": 1.0855, "step": 313 }, { "epoch": 0.05, "grad_norm": 43.67952947804621, "learning_rate": 1.0883882149046794e-05, "loss": 1.3718, "step": 314 }, { "epoch": 0.05, "grad_norm": 21.554504791956532, "learning_rate": 1.0918544194107452e-05, "loss": 0.9619, "step": 315 }, { "epoch": 0.05, "grad_norm": 22.608862729388665, "learning_rate": 1.0953206239168113e-05, "loss": 1.0781, "step": 316 }, { "epoch": 0.05, "grad_norm": 41.32110408068733, "learning_rate": 1.0987868284228771e-05, "loss": 1.1264, "step": 317 }, { "epoch": 0.05, "grad_norm": 30.537403696445477, "learning_rate": 1.1022530329289429e-05, "loss": 1.1479, "step": 318 }, { "epoch": 0.05, "grad_norm": 22.88435630109961, "learning_rate": 1.1057192374350088e-05, "loss": 1.1415, "step": 319 }, { "epoch": 0.05, "grad_norm": 34.390235825883956, "learning_rate": 1.1091854419410746e-05, "loss": 1.0666, "step": 320 }, { "epoch": 0.05, "grad_norm": 19.016075798543312, "learning_rate": 1.1126516464471404e-05, "loss": 1.0004, "step": 321 }, { "epoch": 0.05, "grad_norm": 30.738052474932843, "learning_rate": 1.1161178509532062e-05, "loss": 1.0154, "step": 322 }, { "epoch": 0.05, "grad_norm": 38.4046694746161, "learning_rate": 1.1195840554592723e-05, "loss": 1.2249, "step": 323 }, { "epoch": 0.05, "grad_norm": 5.28829747954234, "learning_rate": 1.1230502599653381e-05, "loss": 1.1102, "step": 324 }, { "epoch": 0.05, "grad_norm": 23.448439893190514, "learning_rate": 1.1265164644714039e-05, "loss": 1.0582, "step": 325 }, { "epoch": 0.05, "grad_norm": 23.83991091682964, "learning_rate": 1.1299826689774698e-05, "loss": 1.1886, "step": 326 }, { "epoch": 0.05, "grad_norm": 36.654051427770966, "learning_rate": 1.1334488734835356e-05, "loss": 1.1055, "step": 327 }, { "epoch": 0.05, "grad_norm": 26.0320110581874, "learning_rate": 1.1369150779896014e-05, "loss": 0.9618, "step": 328 }, { "epoch": 0.05, "grad_norm": 33.79684069416141, "learning_rate": 1.1403812824956672e-05, "loss": 1.2244, "step": 329 }, { "epoch": 0.05, "grad_norm": 27.228386863544692, "learning_rate": 1.1438474870017332e-05, "loss": 1.077, "step": 330 }, { "epoch": 0.05, "grad_norm": 31.836119884132362, "learning_rate": 1.1473136915077991e-05, "loss": 1.0394, "step": 331 }, { "epoch": 0.05, "grad_norm": 25.110325038671736, "learning_rate": 1.1507798960138649e-05, "loss": 1.2516, "step": 332 }, { "epoch": 0.05, "grad_norm": 34.72559738791931, "learning_rate": 1.1542461005199308e-05, "loss": 1.1126, "step": 333 }, { "epoch": 0.05, "grad_norm": 29.478969825496325, "learning_rate": 1.1577123050259966e-05, "loss": 0.9767, "step": 334 }, { "epoch": 0.05, "grad_norm": 20.48757251280962, "learning_rate": 1.1611785095320624e-05, "loss": 1.2302, "step": 335 }, { "epoch": 0.05, "grad_norm": 43.73903702776825, "learning_rate": 1.1646447140381282e-05, "loss": 1.153, "step": 336 }, { "epoch": 0.05, "grad_norm": 36.286243693056086, "learning_rate": 1.1681109185441942e-05, "loss": 1.1822, "step": 337 }, { "epoch": 0.05, "grad_norm": 23.34518100764557, "learning_rate": 1.1715771230502601e-05, "loss": 1.0814, "step": 338 }, { "epoch": 0.05, "grad_norm": 25.58957807542444, "learning_rate": 1.175043327556326e-05, "loss": 1.1042, "step": 339 }, { "epoch": 0.05, "grad_norm": 20.286081019374777, "learning_rate": 1.1785095320623919e-05, "loss": 1.0859, "step": 340 }, { "epoch": 0.05, "grad_norm": 32.51010524518503, "learning_rate": 1.1819757365684576e-05, "loss": 1.0435, "step": 341 }, { "epoch": 0.05, "grad_norm": 23.181531261302148, "learning_rate": 1.1854419410745234e-05, "loss": 1.107, "step": 342 }, { "epoch": 0.05, "grad_norm": 35.47835515156875, "learning_rate": 1.1889081455805894e-05, "loss": 1.0314, "step": 343 }, { "epoch": 0.05, "grad_norm": 24.60361534656248, "learning_rate": 1.1923743500866552e-05, "loss": 1.0581, "step": 344 }, { "epoch": 0.05, "grad_norm": 29.603447091746283, "learning_rate": 1.195840554592721e-05, "loss": 1.0639, "step": 345 }, { "epoch": 0.05, "grad_norm": 42.005933401381604, "learning_rate": 1.199306759098787e-05, "loss": 1.119, "step": 346 }, { "epoch": 0.05, "grad_norm": 31.325769802994927, "learning_rate": 1.2027729636048529e-05, "loss": 1.0979, "step": 347 }, { "epoch": 0.05, "grad_norm": 46.88281582331888, "learning_rate": 1.2062391681109186e-05, "loss": 1.1392, "step": 348 }, { "epoch": 0.05, "grad_norm": 31.996595149395624, "learning_rate": 1.2097053726169844e-05, "loss": 1.0662, "step": 349 }, { "epoch": 0.05, "grad_norm": 32.306233032599955, "learning_rate": 1.2131715771230504e-05, "loss": 1.0391, "step": 350 }, { "epoch": 0.05, "grad_norm": 33.385801572063706, "learning_rate": 1.2166377816291162e-05, "loss": 1.2242, "step": 351 }, { "epoch": 0.05, "grad_norm": 22.43179488371909, "learning_rate": 1.220103986135182e-05, "loss": 1.1773, "step": 352 }, { "epoch": 0.06, "grad_norm": 33.50602676407605, "learning_rate": 1.2235701906412477e-05, "loss": 1.0898, "step": 353 }, { "epoch": 0.06, "grad_norm": 19.631638215145294, "learning_rate": 1.2270363951473139e-05, "loss": 1.0632, "step": 354 }, { "epoch": 0.06, "grad_norm": 25.926041056193505, "learning_rate": 1.2305025996533797e-05, "loss": 0.9388, "step": 355 }, { "epoch": 0.06, "grad_norm": 26.932012060913973, "learning_rate": 1.2339688041594456e-05, "loss": 1.2701, "step": 356 }, { "epoch": 0.06, "grad_norm": 26.65525816696688, "learning_rate": 1.2374350086655114e-05, "loss": 0.9718, "step": 357 }, { "epoch": 0.06, "grad_norm": 22.096018642350085, "learning_rate": 1.2409012131715772e-05, "loss": 0.9923, "step": 358 }, { "epoch": 0.06, "grad_norm": 23.929265322745625, "learning_rate": 1.244367417677643e-05, "loss": 1.1121, "step": 359 }, { "epoch": 0.06, "grad_norm": 28.227952983086723, "learning_rate": 1.247833622183709e-05, "loss": 1.1241, "step": 360 }, { "epoch": 0.06, "grad_norm": 21.766410493745894, "learning_rate": 1.2512998266897749e-05, "loss": 1.1074, "step": 361 }, { "epoch": 0.06, "grad_norm": 26.994232906235982, "learning_rate": 1.2547660311958407e-05, "loss": 1.1329, "step": 362 }, { "epoch": 0.06, "grad_norm": 28.510957715316916, "learning_rate": 1.2582322357019066e-05, "loss": 1.0699, "step": 363 }, { "epoch": 0.06, "grad_norm": 24.935369772262348, "learning_rate": 1.2616984402079724e-05, "loss": 1.1374, "step": 364 }, { "epoch": 0.06, "grad_norm": 36.271485986247896, "learning_rate": 1.2651646447140382e-05, "loss": 1.0808, "step": 365 }, { "epoch": 0.06, "grad_norm": 28.99020265257391, "learning_rate": 1.268630849220104e-05, "loss": 1.0781, "step": 366 }, { "epoch": 0.06, "grad_norm": 37.36522216302166, "learning_rate": 1.27209705372617e-05, "loss": 1.0592, "step": 367 }, { "epoch": 0.06, "grad_norm": 31.96383324342438, "learning_rate": 1.2755632582322357e-05, "loss": 0.9815, "step": 368 }, { "epoch": 0.06, "grad_norm": 33.51334153868519, "learning_rate": 1.2790294627383017e-05, "loss": 1.1268, "step": 369 }, { "epoch": 0.06, "grad_norm": 5.693184440634166, "learning_rate": 1.2824956672443676e-05, "loss": 1.0338, "step": 370 }, { "epoch": 0.06, "grad_norm": 30.48413772334245, "learning_rate": 1.2859618717504334e-05, "loss": 1.2143, "step": 371 }, { "epoch": 0.06, "grad_norm": 26.667324067868737, "learning_rate": 1.2894280762564992e-05, "loss": 1.0773, "step": 372 }, { "epoch": 0.06, "grad_norm": 32.613217908327364, "learning_rate": 1.2928942807625652e-05, "loss": 1.0994, "step": 373 }, { "epoch": 0.06, "grad_norm": 25.86133089764893, "learning_rate": 1.296360485268631e-05, "loss": 1.0735, "step": 374 }, { "epoch": 0.06, "grad_norm": 31.189272337112506, "learning_rate": 1.2998266897746967e-05, "loss": 1.1056, "step": 375 }, { "epoch": 0.06, "grad_norm": 19.21329440551253, "learning_rate": 1.3032928942807628e-05, "loss": 1.0915, "step": 376 }, { "epoch": 0.06, "grad_norm": 29.346472574260808, "learning_rate": 1.3067590987868286e-05, "loss": 1.1536, "step": 377 }, { "epoch": 0.06, "grad_norm": 45.79309598200069, "learning_rate": 1.3102253032928944e-05, "loss": 1.1109, "step": 378 }, { "epoch": 0.06, "grad_norm": 29.733980734491936, "learning_rate": 1.3136915077989602e-05, "loss": 1.0806, "step": 379 }, { "epoch": 0.06, "grad_norm": 26.973894157979053, "learning_rate": 1.3171577123050262e-05, "loss": 1.093, "step": 380 }, { "epoch": 0.06, "grad_norm": 27.54901906636627, "learning_rate": 1.320623916811092e-05, "loss": 1.0221, "step": 381 }, { "epoch": 0.06, "grad_norm": 22.908550257544434, "learning_rate": 1.3240901213171577e-05, "loss": 1.0472, "step": 382 }, { "epoch": 0.06, "grad_norm": 27.727359550673427, "learning_rate": 1.3275563258232235e-05, "loss": 1.0048, "step": 383 }, { "epoch": 0.06, "grad_norm": 29.224659504713923, "learning_rate": 1.3310225303292896e-05, "loss": 1.1251, "step": 384 }, { "epoch": 0.06, "grad_norm": 37.47984058913266, "learning_rate": 1.3344887348353554e-05, "loss": 1.0694, "step": 385 }, { "epoch": 0.06, "grad_norm": 34.247752152418364, "learning_rate": 1.3379549393414212e-05, "loss": 1.2244, "step": 386 }, { "epoch": 0.06, "grad_norm": 26.877945774559905, "learning_rate": 1.3414211438474872e-05, "loss": 1.0694, "step": 387 }, { "epoch": 0.06, "grad_norm": 27.140336946003078, "learning_rate": 1.344887348353553e-05, "loss": 1.2015, "step": 388 }, { "epoch": 0.06, "grad_norm": 27.9027450262441, "learning_rate": 1.3483535528596187e-05, "loss": 1.086, "step": 389 }, { "epoch": 0.06, "grad_norm": 27.171382283179494, "learning_rate": 1.3518197573656845e-05, "loss": 1.1288, "step": 390 }, { "epoch": 0.06, "grad_norm": 21.796142295773222, "learning_rate": 1.3552859618717506e-05, "loss": 1.1133, "step": 391 }, { "epoch": 0.06, "grad_norm": 37.9444241544439, "learning_rate": 1.3587521663778164e-05, "loss": 1.1072, "step": 392 }, { "epoch": 0.06, "grad_norm": 28.169641143832457, "learning_rate": 1.3622183708838824e-05, "loss": 1.0657, "step": 393 }, { "epoch": 0.06, "grad_norm": 34.603979212792716, "learning_rate": 1.3656845753899482e-05, "loss": 1.2969, "step": 394 }, { "epoch": 0.06, "grad_norm": 32.377532200853615, "learning_rate": 1.369150779896014e-05, "loss": 1.0666, "step": 395 }, { "epoch": 0.06, "grad_norm": 17.266813074204492, "learning_rate": 1.3726169844020797e-05, "loss": 1.0658, "step": 396 }, { "epoch": 0.06, "grad_norm": 34.32559842190941, "learning_rate": 1.3760831889081457e-05, "loss": 1.0985, "step": 397 }, { "epoch": 0.06, "grad_norm": 34.371715400867, "learning_rate": 1.3795493934142115e-05, "loss": 1.3105, "step": 398 }, { "epoch": 0.06, "grad_norm": 41.63190000424927, "learning_rate": 1.3830155979202774e-05, "loss": 1.0433, "step": 399 }, { "epoch": 0.06, "grad_norm": 27.90569959847415, "learning_rate": 1.3864818024263434e-05, "loss": 1.0908, "step": 400 }, { "epoch": 0.06, "grad_norm": 21.6998805764918, "learning_rate": 1.3899480069324092e-05, "loss": 1.0031, "step": 401 }, { "epoch": 0.06, "grad_norm": 30.075689023941184, "learning_rate": 1.393414211438475e-05, "loss": 1.0568, "step": 402 }, { "epoch": 0.06, "grad_norm": 37.3133373278367, "learning_rate": 1.3968804159445408e-05, "loss": 1.1699, "step": 403 }, { "epoch": 0.06, "grad_norm": 27.4763675646989, "learning_rate": 1.4003466204506067e-05, "loss": 1.2245, "step": 404 }, { "epoch": 0.06, "grad_norm": 31.103725701166116, "learning_rate": 1.4038128249566725e-05, "loss": 1.0939, "step": 405 }, { "epoch": 0.06, "grad_norm": 16.757187190093457, "learning_rate": 1.4072790294627383e-05, "loss": 1.0749, "step": 406 }, { "epoch": 0.06, "grad_norm": 22.835036799977743, "learning_rate": 1.4107452339688044e-05, "loss": 1.031, "step": 407 }, { "epoch": 0.06, "grad_norm": 42.772867923933184, "learning_rate": 1.4142114384748702e-05, "loss": 1.1827, "step": 408 }, { "epoch": 0.06, "grad_norm": 37.04203269843442, "learning_rate": 1.417677642980936e-05, "loss": 1.2603, "step": 409 }, { "epoch": 0.06, "grad_norm": 27.941152088039125, "learning_rate": 1.421143847487002e-05, "loss": 1.0402, "step": 410 }, { "epoch": 0.06, "grad_norm": 30.012488225595373, "learning_rate": 1.4246100519930677e-05, "loss": 1.0436, "step": 411 }, { "epoch": 0.06, "grad_norm": 34.68960934660663, "learning_rate": 1.4280762564991335e-05, "loss": 1.0056, "step": 412 }, { "epoch": 0.06, "grad_norm": 23.378956623070927, "learning_rate": 1.4315424610051993e-05, "loss": 1.1198, "step": 413 }, { "epoch": 0.06, "grad_norm": 28.51518555812069, "learning_rate": 1.4350086655112654e-05, "loss": 1.1016, "step": 414 }, { "epoch": 0.06, "grad_norm": 20.430786335206683, "learning_rate": 1.4384748700173312e-05, "loss": 1.0082, "step": 415 }, { "epoch": 0.06, "grad_norm": 26.425432187951778, "learning_rate": 1.441941074523397e-05, "loss": 1.0608, "step": 416 }, { "epoch": 0.07, "grad_norm": 20.56158873892592, "learning_rate": 1.445407279029463e-05, "loss": 1.0575, "step": 417 }, { "epoch": 0.07, "grad_norm": 5.993745699464292, "learning_rate": 1.4488734835355287e-05, "loss": 1.1353, "step": 418 }, { "epoch": 0.07, "grad_norm": 33.698564188109685, "learning_rate": 1.4523396880415945e-05, "loss": 1.1354, "step": 419 }, { "epoch": 0.07, "grad_norm": 13.738939505418845, "learning_rate": 1.4558058925476603e-05, "loss": 1.0725, "step": 420 }, { "epoch": 0.07, "grad_norm": 30.872998443969035, "learning_rate": 1.4592720970537263e-05, "loss": 1.1156, "step": 421 }, { "epoch": 0.07, "grad_norm": 18.35365116093291, "learning_rate": 1.4627383015597922e-05, "loss": 1.0658, "step": 422 }, { "epoch": 0.07, "grad_norm": 22.853792234356817, "learning_rate": 1.466204506065858e-05, "loss": 1.0421, "step": 423 }, { "epoch": 0.07, "grad_norm": 32.3637560867717, "learning_rate": 1.469670710571924e-05, "loss": 1.13, "step": 424 }, { "epoch": 0.07, "grad_norm": 24.375143306873333, "learning_rate": 1.4731369150779897e-05, "loss": 1.1187, "step": 425 }, { "epoch": 0.07, "grad_norm": 24.315598036728183, "learning_rate": 1.4766031195840555e-05, "loss": 1.0749, "step": 426 }, { "epoch": 0.07, "grad_norm": 31.578187560218176, "learning_rate": 1.4800693240901213e-05, "loss": 1.0663, "step": 427 }, { "epoch": 0.07, "grad_norm": 29.277862572733177, "learning_rate": 1.4835355285961873e-05, "loss": 1.0366, "step": 428 }, { "epoch": 0.07, "grad_norm": 27.773854849425298, "learning_rate": 1.4870017331022532e-05, "loss": 1.0361, "step": 429 }, { "epoch": 0.07, "grad_norm": 30.07149383967753, "learning_rate": 1.4904679376083192e-05, "loss": 1.1342, "step": 430 }, { "epoch": 0.07, "grad_norm": 28.905491040550434, "learning_rate": 1.493934142114385e-05, "loss": 1.2152, "step": 431 }, { "epoch": 0.07, "grad_norm": 18.66915981630242, "learning_rate": 1.4974003466204507e-05, "loss": 1.0779, "step": 432 }, { "epoch": 0.07, "grad_norm": 31.81603142861209, "learning_rate": 1.5008665511265165e-05, "loss": 1.0872, "step": 433 }, { "epoch": 0.07, "grad_norm": 34.11991082135092, "learning_rate": 1.5043327556325825e-05, "loss": 0.9974, "step": 434 }, { "epoch": 0.07, "grad_norm": 26.143410269225893, "learning_rate": 1.5077989601386483e-05, "loss": 1.1808, "step": 435 }, { "epoch": 0.07, "grad_norm": 25.641715173600016, "learning_rate": 1.511265164644714e-05, "loss": 1.1308, "step": 436 }, { "epoch": 0.07, "grad_norm": 33.98886923001349, "learning_rate": 1.5147313691507802e-05, "loss": 1.1661, "step": 437 }, { "epoch": 0.07, "grad_norm": 18.572967687726788, "learning_rate": 1.518197573656846e-05, "loss": 1.0132, "step": 438 }, { "epoch": 0.07, "grad_norm": 26.362442319213617, "learning_rate": 1.5216637781629117e-05, "loss": 1.0149, "step": 439 }, { "epoch": 0.07, "grad_norm": 27.28730137556571, "learning_rate": 1.5251299826689775e-05, "loss": 1.0923, "step": 440 }, { "epoch": 0.07, "grad_norm": 19.457901932680777, "learning_rate": 1.5285961871750435e-05, "loss": 0.9266, "step": 441 }, { "epoch": 0.07, "grad_norm": 26.42685430960279, "learning_rate": 1.532062391681109e-05, "loss": 1.0687, "step": 442 }, { "epoch": 0.07, "grad_norm": 25.110720276179347, "learning_rate": 1.535528596187175e-05, "loss": 1.022, "step": 443 }, { "epoch": 0.07, "grad_norm": 27.072955662218703, "learning_rate": 1.538994800693241e-05, "loss": 1.0953, "step": 444 }, { "epoch": 0.07, "grad_norm": 20.11605472079033, "learning_rate": 1.542461005199307e-05, "loss": 0.9869, "step": 445 }, { "epoch": 0.07, "grad_norm": 34.431652101331636, "learning_rate": 1.545927209705373e-05, "loss": 1.3131, "step": 446 }, { "epoch": 0.07, "grad_norm": 33.46294148670095, "learning_rate": 1.5493934142114385e-05, "loss": 0.9587, "step": 447 }, { "epoch": 0.07, "grad_norm": 25.8286323566764, "learning_rate": 1.5528596187175045e-05, "loss": 1.1004, "step": 448 }, { "epoch": 0.07, "grad_norm": 19.15366308486822, "learning_rate": 1.5563258232235705e-05, "loss": 0.9616, "step": 449 }, { "epoch": 0.07, "grad_norm": 17.821871415050428, "learning_rate": 1.559792027729636e-05, "loss": 0.8986, "step": 450 }, { "epoch": 0.07, "grad_norm": 35.99595845901309, "learning_rate": 1.563258232235702e-05, "loss": 1.2051, "step": 451 }, { "epoch": 0.07, "grad_norm": 20.096743531257545, "learning_rate": 1.566724436741768e-05, "loss": 0.9922, "step": 452 }, { "epoch": 0.07, "grad_norm": 5.701941811616413, "learning_rate": 1.570190641247834e-05, "loss": 0.9759, "step": 453 }, { "epoch": 0.07, "grad_norm": 19.0741338296344, "learning_rate": 1.5736568457538996e-05, "loss": 1.0567, "step": 454 }, { "epoch": 0.07, "grad_norm": 5.2711761486261555, "learning_rate": 1.5771230502599655e-05, "loss": 1.1068, "step": 455 }, { "epoch": 0.07, "grad_norm": 22.120557194163602, "learning_rate": 1.5805892547660315e-05, "loss": 0.9328, "step": 456 }, { "epoch": 0.07, "grad_norm": 25.54513505382531, "learning_rate": 1.584055459272097e-05, "loss": 1.1424, "step": 457 }, { "epoch": 0.07, "grad_norm": 29.169606155817124, "learning_rate": 1.587521663778163e-05, "loss": 0.9743, "step": 458 }, { "epoch": 0.07, "grad_norm": 23.73216498446794, "learning_rate": 1.5909878682842286e-05, "loss": 0.9859, "step": 459 }, { "epoch": 0.07, "grad_norm": 20.04924244626764, "learning_rate": 1.594454072790295e-05, "loss": 1.0824, "step": 460 }, { "epoch": 0.07, "grad_norm": 26.32800409759019, "learning_rate": 1.5979202772963606e-05, "loss": 1.0437, "step": 461 }, { "epoch": 0.07, "grad_norm": 21.947383935631013, "learning_rate": 1.6013864818024265e-05, "loss": 1.0657, "step": 462 }, { "epoch": 0.07, "grad_norm": 42.90413702383083, "learning_rate": 1.6048526863084925e-05, "loss": 1.1631, "step": 463 }, { "epoch": 0.07, "grad_norm": 22.11730306306142, "learning_rate": 1.608318890814558e-05, "loss": 1.2399, "step": 464 }, { "epoch": 0.07, "grad_norm": 25.04111791854184, "learning_rate": 1.611785095320624e-05, "loss": 0.952, "step": 465 }, { "epoch": 0.07, "grad_norm": 15.07962514994669, "learning_rate": 1.6152512998266897e-05, "loss": 0.9327, "step": 466 }, { "epoch": 0.07, "grad_norm": 24.38630126346466, "learning_rate": 1.618717504332756e-05, "loss": 1.0544, "step": 467 }, { "epoch": 0.07, "grad_norm": 32.06929738788105, "learning_rate": 1.6221837088388216e-05, "loss": 1.1331, "step": 468 }, { "epoch": 0.07, "grad_norm": 20.664537084714688, "learning_rate": 1.6256499133448875e-05, "loss": 0.9988, "step": 469 }, { "epoch": 0.07, "grad_norm": 21.206259630000375, "learning_rate": 1.6291161178509535e-05, "loss": 1.0925, "step": 470 }, { "epoch": 0.07, "grad_norm": 36.306049239295874, "learning_rate": 1.632582322357019e-05, "loss": 1.0819, "step": 471 }, { "epoch": 0.07, "grad_norm": 27.65730606476609, "learning_rate": 1.636048526863085e-05, "loss": 1.0467, "step": 472 }, { "epoch": 0.07, "grad_norm": 31.947233780879255, "learning_rate": 1.639514731369151e-05, "loss": 1.0235, "step": 473 }, { "epoch": 0.07, "grad_norm": 26.346397050271698, "learning_rate": 1.6429809358752166e-05, "loss": 0.9507, "step": 474 }, { "epoch": 0.07, "grad_norm": 32.20872637450095, "learning_rate": 1.6464471403812826e-05, "loss": 1.1141, "step": 475 }, { "epoch": 0.07, "grad_norm": 40.187479039923865, "learning_rate": 1.6499133448873485e-05, "loss": 1.0833, "step": 476 }, { "epoch": 0.07, "grad_norm": 27.63407629745231, "learning_rate": 1.6533795493934145e-05, "loss": 1.0913, "step": 477 }, { "epoch": 0.07, "grad_norm": 26.55700920799157, "learning_rate": 1.65684575389948e-05, "loss": 1.0217, "step": 478 }, { "epoch": 0.07, "grad_norm": 21.303006357296574, "learning_rate": 1.660311958405546e-05, "loss": 0.9185, "step": 479 }, { "epoch": 0.07, "grad_norm": 34.83074933968313, "learning_rate": 1.663778162911612e-05, "loss": 1.1405, "step": 480 }, { "epoch": 0.08, "grad_norm": 21.905102922878694, "learning_rate": 1.6672443674176776e-05, "loss": 1.0552, "step": 481 }, { "epoch": 0.08, "grad_norm": 22.5374601706932, "learning_rate": 1.6707105719237436e-05, "loss": 1.0506, "step": 482 }, { "epoch": 0.08, "grad_norm": 30.868805725347034, "learning_rate": 1.6741767764298095e-05, "loss": 1.2388, "step": 483 }, { "epoch": 0.08, "grad_norm": 31.809013872207533, "learning_rate": 1.6776429809358755e-05, "loss": 1.1573, "step": 484 }, { "epoch": 0.08, "grad_norm": 22.584330183913444, "learning_rate": 1.681109185441941e-05, "loss": 1.0159, "step": 485 }, { "epoch": 0.08, "grad_norm": 18.359652479793855, "learning_rate": 1.684575389948007e-05, "loss": 0.9319, "step": 486 }, { "epoch": 0.08, "grad_norm": 24.035349084560117, "learning_rate": 1.688041594454073e-05, "loss": 1.192, "step": 487 }, { "epoch": 0.08, "grad_norm": 21.091984735728435, "learning_rate": 1.6915077989601386e-05, "loss": 0.993, "step": 488 }, { "epoch": 0.08, "grad_norm": 25.26835354784831, "learning_rate": 1.6949740034662046e-05, "loss": 1.1226, "step": 489 }, { "epoch": 0.08, "grad_norm": 26.006399020570456, "learning_rate": 1.6984402079722705e-05, "loss": 0.9572, "step": 490 }, { "epoch": 0.08, "grad_norm": 21.67672143114982, "learning_rate": 1.7019064124783365e-05, "loss": 1.0041, "step": 491 }, { "epoch": 0.08, "grad_norm": 35.59097800039987, "learning_rate": 1.705372616984402e-05, "loss": 1.0966, "step": 492 }, { "epoch": 0.08, "grad_norm": 24.007039133085687, "learning_rate": 1.708838821490468e-05, "loss": 1.0351, "step": 493 }, { "epoch": 0.08, "grad_norm": 16.6685605548639, "learning_rate": 1.712305025996534e-05, "loss": 0.9102, "step": 494 }, { "epoch": 0.08, "grad_norm": 23.69384850275509, "learning_rate": 1.7157712305025996e-05, "loss": 1.0426, "step": 495 }, { "epoch": 0.08, "grad_norm": 29.480033542751137, "learning_rate": 1.7192374350086656e-05, "loss": 1.0882, "step": 496 }, { "epoch": 0.08, "grad_norm": 40.07920094211409, "learning_rate": 1.7227036395147316e-05, "loss": 1.0148, "step": 497 }, { "epoch": 0.08, "grad_norm": 29.233060679818937, "learning_rate": 1.7261698440207975e-05, "loss": 1.0744, "step": 498 }, { "epoch": 0.08, "grad_norm": 35.182566386669606, "learning_rate": 1.729636048526863e-05, "loss": 0.9997, "step": 499 }, { "epoch": 0.08, "grad_norm": 30.269976589123253, "learning_rate": 1.733102253032929e-05, "loss": 1.1004, "step": 500 }, { "epoch": 0.08, "grad_norm": 24.245753203984723, "learning_rate": 1.736568457538995e-05, "loss": 1.0803, "step": 501 }, { "epoch": 0.08, "grad_norm": 39.04873017456366, "learning_rate": 1.7400346620450606e-05, "loss": 1.1467, "step": 502 }, { "epoch": 0.08, "grad_norm": 24.792476515138087, "learning_rate": 1.7435008665511266e-05, "loss": 1.0947, "step": 503 }, { "epoch": 0.08, "grad_norm": 30.2379304968761, "learning_rate": 1.7469670710571926e-05, "loss": 1.1267, "step": 504 }, { "epoch": 0.08, "grad_norm": 33.65685682891974, "learning_rate": 1.7504332755632585e-05, "loss": 1.1169, "step": 505 }, { "epoch": 0.08, "grad_norm": 24.56601044270328, "learning_rate": 1.7538994800693245e-05, "loss": 1.0628, "step": 506 }, { "epoch": 0.08, "grad_norm": 21.60060010011299, "learning_rate": 1.75736568457539e-05, "loss": 1.0078, "step": 507 }, { "epoch": 0.08, "grad_norm": 24.11059699605151, "learning_rate": 1.760831889081456e-05, "loss": 1.046, "step": 508 }, { "epoch": 0.08, "grad_norm": 34.78563216246366, "learning_rate": 1.7642980935875217e-05, "loss": 1.2045, "step": 509 }, { "epoch": 0.08, "grad_norm": 27.789317608451398, "learning_rate": 1.7677642980935876e-05, "loss": 0.9659, "step": 510 }, { "epoch": 0.08, "grad_norm": 27.727447236338673, "learning_rate": 1.7712305025996536e-05, "loss": 1.0327, "step": 511 }, { "epoch": 0.08, "grad_norm": 25.80408909837495, "learning_rate": 1.7746967071057192e-05, "loss": 1.0152, "step": 512 }, { "epoch": 0.08, "grad_norm": 18.355020428703057, "learning_rate": 1.7781629116117855e-05, "loss": 1.0644, "step": 513 }, { "epoch": 0.08, "grad_norm": 22.682545495976207, "learning_rate": 1.781629116117851e-05, "loss": 1.0329, "step": 514 }, { "epoch": 0.08, "grad_norm": 25.989554582298343, "learning_rate": 1.785095320623917e-05, "loss": 1.1096, "step": 515 }, { "epoch": 0.08, "grad_norm": 26.79673932655621, "learning_rate": 1.7885615251299827e-05, "loss": 1.1577, "step": 516 }, { "epoch": 0.08, "grad_norm": 25.84256013993326, "learning_rate": 1.7920277296360486e-05, "loss": 1.0689, "step": 517 }, { "epoch": 0.08, "grad_norm": 24.533577428342877, "learning_rate": 1.7954939341421146e-05, "loss": 0.8793, "step": 518 }, { "epoch": 0.08, "grad_norm": 23.370691730497956, "learning_rate": 1.7989601386481802e-05, "loss": 1.0736, "step": 519 }, { "epoch": 0.08, "grad_norm": 24.610226379684057, "learning_rate": 1.802426343154246e-05, "loss": 0.987, "step": 520 }, { "epoch": 0.08, "grad_norm": 20.44208765087892, "learning_rate": 1.805892547660312e-05, "loss": 1.0504, "step": 521 }, { "epoch": 0.08, "grad_norm": 18.144404591753734, "learning_rate": 1.809358752166378e-05, "loss": 1.0604, "step": 522 }, { "epoch": 0.08, "grad_norm": 16.877103069616282, "learning_rate": 1.812824956672444e-05, "loss": 0.9898, "step": 523 }, { "epoch": 0.08, "grad_norm": 23.737747217762195, "learning_rate": 1.8162911611785096e-05, "loss": 1.0038, "step": 524 }, { "epoch": 0.08, "grad_norm": 16.25951547791112, "learning_rate": 1.8197573656845756e-05, "loss": 1.0588, "step": 525 }, { "epoch": 0.08, "grad_norm": 27.599355801073518, "learning_rate": 1.8232235701906412e-05, "loss": 0.9713, "step": 526 }, { "epoch": 0.08, "grad_norm": 33.122993457048636, "learning_rate": 1.826689774696707e-05, "loss": 1.0201, "step": 527 }, { "epoch": 0.08, "grad_norm": 24.223397469833518, "learning_rate": 1.830155979202773e-05, "loss": 1.0946, "step": 528 }, { "epoch": 0.08, "grad_norm": 29.740458575239316, "learning_rate": 1.833622183708839e-05, "loss": 0.9838, "step": 529 }, { "epoch": 0.08, "grad_norm": 21.341621904414357, "learning_rate": 1.837088388214905e-05, "loss": 1.0443, "step": 530 }, { "epoch": 0.08, "grad_norm": 26.34550028831831, "learning_rate": 1.8405545927209706e-05, "loss": 1.0417, "step": 531 }, { "epoch": 0.08, "grad_norm": 22.180889743363394, "learning_rate": 1.8440207972270366e-05, "loss": 1.1049, "step": 532 }, { "epoch": 0.08, "grad_norm": 25.271695696375218, "learning_rate": 1.8474870017331022e-05, "loss": 1.0915, "step": 533 }, { "epoch": 0.08, "grad_norm": 23.83622464957578, "learning_rate": 1.850953206239168e-05, "loss": 1.1916, "step": 534 }, { "epoch": 0.08, "grad_norm": 23.2894505708719, "learning_rate": 1.854419410745234e-05, "loss": 1.073, "step": 535 }, { "epoch": 0.08, "grad_norm": 28.2656482092286, "learning_rate": 1.8578856152513e-05, "loss": 1.0285, "step": 536 }, { "epoch": 0.08, "grad_norm": 20.59298381914133, "learning_rate": 1.861351819757366e-05, "loss": 1.1132, "step": 537 }, { "epoch": 0.08, "grad_norm": 23.60061650304393, "learning_rate": 1.8648180242634316e-05, "loss": 1.1276, "step": 538 }, { "epoch": 0.08, "grad_norm": 35.98301939371637, "learning_rate": 1.8682842287694976e-05, "loss": 1.159, "step": 539 }, { "epoch": 0.08, "grad_norm": 4.4888981427578205, "learning_rate": 1.8717504332755632e-05, "loss": 1.0028, "step": 540 }, { "epoch": 0.08, "grad_norm": 36.99781273324386, "learning_rate": 1.8752166377816292e-05, "loss": 1.1377, "step": 541 }, { "epoch": 0.08, "grad_norm": 15.939578263032633, "learning_rate": 1.878682842287695e-05, "loss": 0.999, "step": 542 }, { "epoch": 0.08, "grad_norm": 22.85660085708705, "learning_rate": 1.882149046793761e-05, "loss": 1.0134, "step": 543 }, { "epoch": 0.08, "grad_norm": 17.241163366830648, "learning_rate": 1.885615251299827e-05, "loss": 1.0081, "step": 544 }, { "epoch": 0.09, "grad_norm": 25.474464093786764, "learning_rate": 1.8890814558058927e-05, "loss": 1.1213, "step": 545 }, { "epoch": 0.09, "grad_norm": 33.11335194940052, "learning_rate": 1.8925476603119586e-05, "loss": 1.0378, "step": 546 }, { "epoch": 0.09, "grad_norm": 31.546985247203498, "learning_rate": 1.8960138648180246e-05, "loss": 1.0594, "step": 547 }, { "epoch": 0.09, "grad_norm": 27.353861552798367, "learning_rate": 1.8994800693240902e-05, "loss": 1.0908, "step": 548 }, { "epoch": 0.09, "grad_norm": 18.9522994570028, "learning_rate": 1.902946273830156e-05, "loss": 0.9971, "step": 549 }, { "epoch": 0.09, "grad_norm": 26.431519891655817, "learning_rate": 1.9064124783362217e-05, "loss": 1.1073, "step": 550 }, { "epoch": 0.09, "grad_norm": 24.854707471528116, "learning_rate": 1.909878682842288e-05, "loss": 1.0473, "step": 551 }, { "epoch": 0.09, "grad_norm": 19.821719178716894, "learning_rate": 1.9133448873483537e-05, "loss": 0.987, "step": 552 }, { "epoch": 0.09, "grad_norm": 19.670015765342832, "learning_rate": 1.9168110918544196e-05, "loss": 1.0873, "step": 553 }, { "epoch": 0.09, "grad_norm": 20.380559817983787, "learning_rate": 1.9202772963604856e-05, "loss": 0.9886, "step": 554 }, { "epoch": 0.09, "grad_norm": 27.144432556811733, "learning_rate": 1.9237435008665512e-05, "loss": 0.9142, "step": 555 }, { "epoch": 0.09, "grad_norm": 26.895634272720514, "learning_rate": 1.927209705372617e-05, "loss": 1.1352, "step": 556 }, { "epoch": 0.09, "grad_norm": 22.938473895183744, "learning_rate": 1.9306759098786828e-05, "loss": 1.1945, "step": 557 }, { "epoch": 0.09, "grad_norm": 19.095345260358535, "learning_rate": 1.9341421143847487e-05, "loss": 1.0434, "step": 558 }, { "epoch": 0.09, "grad_norm": 41.15274760623393, "learning_rate": 1.9376083188908147e-05, "loss": 1.1775, "step": 559 }, { "epoch": 0.09, "grad_norm": 23.886517844676984, "learning_rate": 1.9410745233968806e-05, "loss": 1.0895, "step": 560 }, { "epoch": 0.09, "grad_norm": 22.219768592031212, "learning_rate": 1.9445407279029466e-05, "loss": 1.022, "step": 561 }, { "epoch": 0.09, "grad_norm": 21.961321974964306, "learning_rate": 1.9480069324090122e-05, "loss": 1.0518, "step": 562 }, { "epoch": 0.09, "grad_norm": 32.005822631718004, "learning_rate": 1.951473136915078e-05, "loss": 1.0212, "step": 563 }, { "epoch": 0.09, "grad_norm": 21.961686493739755, "learning_rate": 1.954939341421144e-05, "loss": 1.0504, "step": 564 }, { "epoch": 0.09, "grad_norm": 29.25877929944018, "learning_rate": 1.9584055459272097e-05, "loss": 1.0929, "step": 565 }, { "epoch": 0.09, "grad_norm": 24.163975500572413, "learning_rate": 1.9618717504332757e-05, "loss": 1.0706, "step": 566 }, { "epoch": 0.09, "grad_norm": 26.611705318472513, "learning_rate": 1.9653379549393416e-05, "loss": 0.9836, "step": 567 }, { "epoch": 0.09, "grad_norm": 28.184492445566853, "learning_rate": 1.9688041594454076e-05, "loss": 0.9916, "step": 568 }, { "epoch": 0.09, "grad_norm": 25.68156266800803, "learning_rate": 1.9722703639514732e-05, "loss": 1.1573, "step": 569 }, { "epoch": 0.09, "grad_norm": 28.378879126518083, "learning_rate": 1.975736568457539e-05, "loss": 0.9497, "step": 570 }, { "epoch": 0.09, "grad_norm": 19.356736233473026, "learning_rate": 1.979202772963605e-05, "loss": 0.9251, "step": 571 }, { "epoch": 0.09, "grad_norm": 25.592310467899786, "learning_rate": 1.9826689774696707e-05, "loss": 1.0297, "step": 572 }, { "epoch": 0.09, "grad_norm": 27.14714174210939, "learning_rate": 1.9861351819757367e-05, "loss": 1.0315, "step": 573 }, { "epoch": 0.09, "grad_norm": 28.635279627490927, "learning_rate": 1.9896013864818026e-05, "loss": 1.0889, "step": 574 }, { "epoch": 0.09, "grad_norm": 24.14251563807322, "learning_rate": 1.9930675909878686e-05, "loss": 1.0212, "step": 575 }, { "epoch": 0.09, "grad_norm": 25.605607340399068, "learning_rate": 1.9965337954939342e-05, "loss": 1.055, "step": 576 }, { "epoch": 0.09, "grad_norm": 22.861064176734494, "learning_rate": 2e-05, "loss": 0.9411, "step": 577 }, { "epoch": 0.09, "grad_norm": 32.708221310020775, "learning_rate": 1.9999999857802926e-05, "loss": 1.0222, "step": 578 }, { "epoch": 0.09, "grad_norm": 25.77872904734455, "learning_rate": 1.9999999431211706e-05, "loss": 1.0681, "step": 579 }, { "epoch": 0.09, "grad_norm": 20.068217319997437, "learning_rate": 1.999999872022635e-05, "loss": 1.1066, "step": 580 }, { "epoch": 0.09, "grad_norm": 31.490787994547425, "learning_rate": 1.9999997724846883e-05, "loss": 1.0383, "step": 581 }, { "epoch": 0.09, "grad_norm": 34.98645871334549, "learning_rate": 1.9999996445073327e-05, "loss": 1.1206, "step": 582 }, { "epoch": 0.09, "grad_norm": 35.632614390703424, "learning_rate": 1.9999994880905726e-05, "loss": 1.0542, "step": 583 }, { "epoch": 0.09, "grad_norm": 25.23781468122964, "learning_rate": 1.9999993032344115e-05, "loss": 1.0387, "step": 584 }, { "epoch": 0.09, "grad_norm": 27.61966249849483, "learning_rate": 1.9999990899388556e-05, "loss": 1.0255, "step": 585 }, { "epoch": 0.09, "grad_norm": 24.504192646772623, "learning_rate": 1.9999988482039104e-05, "loss": 1.0037, "step": 586 }, { "epoch": 0.09, "grad_norm": 21.648244447373916, "learning_rate": 1.999998578029583e-05, "loss": 1.2731, "step": 587 }, { "epoch": 0.09, "grad_norm": 21.92577731908766, "learning_rate": 1.999998279415881e-05, "loss": 1.0188, "step": 588 }, { "epoch": 0.09, "grad_norm": 23.714237684191083, "learning_rate": 1.999997952362813e-05, "loss": 1.0955, "step": 589 }, { "epoch": 0.09, "grad_norm": 18.875496663575905, "learning_rate": 1.999997596870388e-05, "loss": 0.9669, "step": 590 }, { "epoch": 0.09, "grad_norm": 17.41305678438287, "learning_rate": 1.9999972129386165e-05, "loss": 0.9065, "step": 591 }, { "epoch": 0.09, "grad_norm": 20.794859932565107, "learning_rate": 1.999996800567509e-05, "loss": 1.0044, "step": 592 }, { "epoch": 0.09, "grad_norm": 23.429124072996856, "learning_rate": 1.999996359757078e-05, "loss": 1.0774, "step": 593 }, { "epoch": 0.09, "grad_norm": 23.424359437948585, "learning_rate": 1.9999958905073352e-05, "loss": 1.0994, "step": 594 }, { "epoch": 0.09, "grad_norm": 18.345026243969027, "learning_rate": 1.9999953928182942e-05, "loss": 1.0255, "step": 595 }, { "epoch": 0.09, "grad_norm": 30.414133266324512, "learning_rate": 1.9999948666899695e-05, "loss": 1.0921, "step": 596 }, { "epoch": 0.09, "grad_norm": 24.384991495585798, "learning_rate": 1.9999943121223753e-05, "loss": 1.1104, "step": 597 }, { "epoch": 0.09, "grad_norm": 18.39591885961205, "learning_rate": 1.999993729115528e-05, "loss": 1.0308, "step": 598 }, { "epoch": 0.09, "grad_norm": 36.55302360842166, "learning_rate": 1.999993117669444e-05, "loss": 1.1033, "step": 599 }, { "epoch": 0.09, "grad_norm": 27.528707666170977, "learning_rate": 1.999992477784141e-05, "loss": 1.002, "step": 600 }, { "epoch": 0.09, "grad_norm": 39.0482325048875, "learning_rate": 1.999991809459637e-05, "loss": 0.9924, "step": 601 }, { "epoch": 0.09, "grad_norm": 31.36379419691972, "learning_rate": 1.9999911126959503e-05, "loss": 1.0878, "step": 602 }, { "epoch": 0.09, "grad_norm": 20.230094375873414, "learning_rate": 1.9999903874931017e-05, "loss": 0.9643, "step": 603 }, { "epoch": 0.09, "grad_norm": 17.097369602337533, "learning_rate": 1.9999896338511117e-05, "loss": 1.0073, "step": 604 }, { "epoch": 0.09, "grad_norm": 22.538517270030404, "learning_rate": 1.999988851770001e-05, "loss": 1.0647, "step": 605 }, { "epoch": 0.09, "grad_norm": 25.087550676595903, "learning_rate": 1.9999880412497927e-05, "loss": 1.1512, "step": 606 }, { "epoch": 0.09, "grad_norm": 22.644576876286553, "learning_rate": 1.9999872022905094e-05, "loss": 1.0329, "step": 607 }, { "epoch": 0.09, "grad_norm": 25.57077115583161, "learning_rate": 1.9999863348921748e-05, "loss": 1.0519, "step": 608 }, { "epoch": 0.1, "grad_norm": 22.806895674814548, "learning_rate": 1.999985439054814e-05, "loss": 1.0923, "step": 609 }, { "epoch": 0.1, "grad_norm": 17.1697596113921, "learning_rate": 1.9999845147784526e-05, "loss": 1.0938, "step": 610 }, { "epoch": 0.1, "grad_norm": 34.984784936143015, "learning_rate": 1.999983562063116e-05, "loss": 1.0604, "step": 611 }, { "epoch": 0.1, "grad_norm": 17.31641290786309, "learning_rate": 1.999982580908832e-05, "loss": 1.0975, "step": 612 }, { "epoch": 0.1, "grad_norm": 28.700540192601025, "learning_rate": 1.9999815713156285e-05, "loss": 1.1786, "step": 613 }, { "epoch": 0.1, "grad_norm": 21.784075607083317, "learning_rate": 1.9999805332835344e-05, "loss": 1.1711, "step": 614 }, { "epoch": 0.1, "grad_norm": 31.566909129873377, "learning_rate": 1.9999794668125784e-05, "loss": 0.979, "step": 615 }, { "epoch": 0.1, "grad_norm": 32.344774729394906, "learning_rate": 1.9999783719027913e-05, "loss": 1.0891, "step": 616 }, { "epoch": 0.1, "grad_norm": 27.350961898141318, "learning_rate": 1.9999772485542048e-05, "loss": 1.0251, "step": 617 }, { "epoch": 0.1, "grad_norm": 30.323243079901548, "learning_rate": 1.99997609676685e-05, "loss": 1.0448, "step": 618 }, { "epoch": 0.1, "grad_norm": 26.559370164481408, "learning_rate": 1.99997491654076e-05, "loss": 1.019, "step": 619 }, { "epoch": 0.1, "grad_norm": 25.116834511656986, "learning_rate": 1.9999737078759684e-05, "loss": 1.0316, "step": 620 }, { "epoch": 0.1, "grad_norm": 38.23014262882626, "learning_rate": 1.999972470772509e-05, "loss": 1.0279, "step": 621 }, { "epoch": 0.1, "grad_norm": 27.264683982109712, "learning_rate": 1.999971205230418e-05, "loss": 1.0609, "step": 622 }, { "epoch": 0.1, "grad_norm": 37.16865727262475, "learning_rate": 1.999969911249731e-05, "loss": 0.9348, "step": 623 }, { "epoch": 0.1, "grad_norm": 15.669900577725162, "learning_rate": 1.9999685888304844e-05, "loss": 1.0294, "step": 624 }, { "epoch": 0.1, "grad_norm": 25.08810207445628, "learning_rate": 1.9999672379727165e-05, "loss": 1.0945, "step": 625 }, { "epoch": 0.1, "grad_norm": 5.214393464839559, "learning_rate": 1.999965858676465e-05, "loss": 0.9841, "step": 626 }, { "epoch": 0.1, "grad_norm": 31.103198322971277, "learning_rate": 1.9999644509417694e-05, "loss": 1.0875, "step": 627 }, { "epoch": 0.1, "grad_norm": 20.915001681033875, "learning_rate": 1.99996301476867e-05, "loss": 0.9758, "step": 628 }, { "epoch": 0.1, "grad_norm": 28.30079704900265, "learning_rate": 1.9999615501572073e-05, "loss": 1.0322, "step": 629 }, { "epoch": 0.1, "grad_norm": 38.04965149385254, "learning_rate": 1.999960057107423e-05, "loss": 1.0074, "step": 630 }, { "epoch": 0.1, "grad_norm": 25.414347805622835, "learning_rate": 1.9999585356193597e-05, "loss": 0.9241, "step": 631 }, { "epoch": 0.1, "grad_norm": 28.90992076557785, "learning_rate": 1.9999569856930604e-05, "loss": 1.0194, "step": 632 }, { "epoch": 0.1, "grad_norm": 25.110853662612588, "learning_rate": 1.9999554073285695e-05, "loss": 1.0284, "step": 633 }, { "epoch": 0.1, "grad_norm": 22.778630374216906, "learning_rate": 1.999953800525932e-05, "loss": 1.1036, "step": 634 }, { "epoch": 0.1, "grad_norm": 24.904941753025636, "learning_rate": 1.9999521652851933e-05, "loss": 0.9953, "step": 635 }, { "epoch": 0.1, "grad_norm": 28.34261364640905, "learning_rate": 1.9999505016063998e-05, "loss": 1.012, "step": 636 }, { "epoch": 0.1, "grad_norm": 8.114153859611994, "learning_rate": 1.9999488094895992e-05, "loss": 0.9224, "step": 637 }, { "epoch": 0.1, "grad_norm": 9.201733390049078, "learning_rate": 1.9999470889348394e-05, "loss": 1.0225, "step": 638 }, { "epoch": 0.1, "grad_norm": 16.33693891034599, "learning_rate": 1.9999453399421692e-05, "loss": 0.9935, "step": 639 }, { "epoch": 0.1, "grad_norm": 18.755063072714822, "learning_rate": 1.9999435625116383e-05, "loss": 0.8937, "step": 640 }, { "epoch": 0.1, "grad_norm": 31.135516419968262, "learning_rate": 1.999941756643298e-05, "loss": 1.0474, "step": 641 }, { "epoch": 0.1, "grad_norm": 22.210619093269454, "learning_rate": 1.9999399223371988e-05, "loss": 0.9777, "step": 642 }, { "epoch": 0.1, "grad_norm": 27.874397763131537, "learning_rate": 1.999938059593393e-05, "loss": 1.0197, "step": 643 }, { "epoch": 0.1, "grad_norm": 23.09541794840785, "learning_rate": 1.999936168411934e-05, "loss": 1.0258, "step": 644 }, { "epoch": 0.1, "grad_norm": 19.123838546802762, "learning_rate": 1.999934248792875e-05, "loss": 0.9936, "step": 645 }, { "epoch": 0.1, "grad_norm": 20.733233766116115, "learning_rate": 1.9999323007362708e-05, "loss": 0.9803, "step": 646 }, { "epoch": 0.1, "grad_norm": 26.073300094137274, "learning_rate": 1.9999303242421773e-05, "loss": 0.9174, "step": 647 }, { "epoch": 0.1, "grad_norm": 18.286722855667904, "learning_rate": 1.9999283193106504e-05, "loss": 0.944, "step": 648 }, { "epoch": 0.1, "grad_norm": 32.56092169939891, "learning_rate": 1.9999262859417466e-05, "loss": 1.019, "step": 649 }, { "epoch": 0.1, "grad_norm": 17.95125505355276, "learning_rate": 1.9999242241355247e-05, "loss": 0.8869, "step": 650 }, { "epoch": 0.1, "grad_norm": 32.132816929611685, "learning_rate": 1.9999221338920424e-05, "loss": 1.1021, "step": 651 }, { "epoch": 0.1, "grad_norm": 17.434045022076624, "learning_rate": 1.9999200152113595e-05, "loss": 0.938, "step": 652 }, { "epoch": 0.1, "grad_norm": 32.50571419843277, "learning_rate": 1.9999178680935366e-05, "loss": 1.0048, "step": 653 }, { "epoch": 0.1, "grad_norm": 16.556043944205204, "learning_rate": 1.9999156925386344e-05, "loss": 0.9398, "step": 654 }, { "epoch": 0.1, "grad_norm": 18.766440333330717, "learning_rate": 1.9999134885467148e-05, "loss": 0.9585, "step": 655 }, { "epoch": 0.1, "grad_norm": 9.354974463671109, "learning_rate": 1.9999112561178402e-05, "loss": 1.074, "step": 656 }, { "epoch": 0.1, "grad_norm": 28.99983671056059, "learning_rate": 1.9999089952520746e-05, "loss": 1.1193, "step": 657 }, { "epoch": 0.1, "grad_norm": 31.083429746626358, "learning_rate": 1.9999067059494824e-05, "loss": 0.9585, "step": 658 }, { "epoch": 0.1, "grad_norm": 22.075384931081132, "learning_rate": 1.999904388210128e-05, "loss": 0.9859, "step": 659 }, { "epoch": 0.1, "grad_norm": 23.781617919943116, "learning_rate": 1.9999020420340782e-05, "loss": 0.9583, "step": 660 }, { "epoch": 0.1, "grad_norm": 31.966704287289733, "learning_rate": 1.999899667421399e-05, "loss": 1.098, "step": 661 }, { "epoch": 0.1, "grad_norm": 25.534238631631656, "learning_rate": 1.999897264372158e-05, "loss": 1.0995, "step": 662 }, { "epoch": 0.1, "grad_norm": 23.068356514703574, "learning_rate": 1.999894832886424e-05, "loss": 0.984, "step": 663 }, { "epoch": 0.1, "grad_norm": 20.9964647538874, "learning_rate": 1.9998923729642657e-05, "loss": 0.998, "step": 664 }, { "epoch": 0.1, "grad_norm": 6.393814403269237, "learning_rate": 1.999889884605753e-05, "loss": 0.9782, "step": 665 }, { "epoch": 0.1, "grad_norm": 24.982013960016147, "learning_rate": 1.9998873678109575e-05, "loss": 1.0013, "step": 666 }, { "epoch": 0.1, "grad_norm": 18.385080431888543, "learning_rate": 1.99988482257995e-05, "loss": 1.0315, "step": 667 }, { "epoch": 0.1, "grad_norm": 21.403720561477744, "learning_rate": 1.9998822489128028e-05, "loss": 0.937, "step": 668 }, { "epoch": 0.1, "grad_norm": 32.37210180407703, "learning_rate": 1.9998796468095897e-05, "loss": 1.0377, "step": 669 }, { "epoch": 0.1, "grad_norm": 16.621343531451927, "learning_rate": 1.999877016270384e-05, "loss": 0.999, "step": 670 }, { "epoch": 0.1, "grad_norm": 27.807004663220358, "learning_rate": 1.9998743572952608e-05, "loss": 0.9875, "step": 671 }, { "epoch": 0.1, "grad_norm": 27.530692159822845, "learning_rate": 1.999871669884296e-05, "loss": 1.0033, "step": 672 }, { "epoch": 0.11, "grad_norm": 17.477830139147436, "learning_rate": 1.999868954037566e-05, "loss": 0.9291, "step": 673 }, { "epoch": 0.11, "grad_norm": 26.30354990763825, "learning_rate": 1.9998662097551475e-05, "loss": 0.9682, "step": 674 }, { "epoch": 0.11, "grad_norm": 26.701506649907408, "learning_rate": 1.9998634370371192e-05, "loss": 1.0579, "step": 675 }, { "epoch": 0.11, "grad_norm": 16.09844673750356, "learning_rate": 1.9998606358835596e-05, "loss": 1.0579, "step": 676 }, { "epoch": 0.11, "grad_norm": 26.042153693159076, "learning_rate": 1.9998578062945483e-05, "loss": 1.1038, "step": 677 }, { "epoch": 0.11, "grad_norm": 17.87003977935117, "learning_rate": 1.999854948270166e-05, "loss": 0.9282, "step": 678 }, { "epoch": 0.11, "grad_norm": 1471.0013591518905, "learning_rate": 1.999852061810494e-05, "loss": 1.2118, "step": 679 }, { "epoch": 0.11, "grad_norm": 35.0744775570728, "learning_rate": 1.9998491469156137e-05, "loss": 1.0116, "step": 680 }, { "epoch": 0.11, "grad_norm": 28.13299186170543, "learning_rate": 1.999846203585609e-05, "loss": 0.9564, "step": 681 }, { "epoch": 0.11, "grad_norm": 23.11076190518904, "learning_rate": 1.9998432318205632e-05, "loss": 0.9886, "step": 682 }, { "epoch": 0.11, "grad_norm": 21.001125157958516, "learning_rate": 1.9998402316205606e-05, "loss": 1.0623, "step": 683 }, { "epoch": 0.11, "grad_norm": 28.818401774857087, "learning_rate": 1.9998372029856866e-05, "loss": 1.0083, "step": 684 }, { "epoch": 0.11, "grad_norm": 20.18902011512905, "learning_rate": 1.9998341459160277e-05, "loss": 0.9179, "step": 685 }, { "epoch": 0.11, "grad_norm": 35.677553504388904, "learning_rate": 1.9998310604116704e-05, "loss": 1.0216, "step": 686 }, { "epoch": 0.11, "grad_norm": 20.768781962357206, "learning_rate": 1.999827946472703e-05, "loss": 0.9849, "step": 687 }, { "epoch": 0.11, "grad_norm": 31.64937660972853, "learning_rate": 1.999824804099213e-05, "loss": 0.9782, "step": 688 }, { "epoch": 0.11, "grad_norm": 24.918096941199437, "learning_rate": 1.9998216332912908e-05, "loss": 1.0171, "step": 689 }, { "epoch": 0.11, "grad_norm": 20.61013836708593, "learning_rate": 1.9998184340490264e-05, "loss": 0.885, "step": 690 }, { "epoch": 0.11, "grad_norm": 22.565527580076296, "learning_rate": 1.9998152063725107e-05, "loss": 1.0084, "step": 691 }, { "epoch": 0.11, "grad_norm": 20.157218866202182, "learning_rate": 1.9998119502618353e-05, "loss": 0.9549, "step": 692 }, { "epoch": 0.11, "grad_norm": 17.942288719847813, "learning_rate": 1.9998086657170925e-05, "loss": 0.9984, "step": 693 }, { "epoch": 0.11, "grad_norm": 18.81983038671551, "learning_rate": 1.9998053527383766e-05, "loss": 1.1064, "step": 694 }, { "epoch": 0.11, "grad_norm": 17.702851369532596, "learning_rate": 1.999802011325781e-05, "loss": 0.9075, "step": 695 }, { "epoch": 0.11, "grad_norm": 31.219935790826188, "learning_rate": 1.9997986414794012e-05, "loss": 1.0162, "step": 696 }, { "epoch": 0.11, "grad_norm": 14.062142120681791, "learning_rate": 1.999795243199333e-05, "loss": 0.9574, "step": 697 }, { "epoch": 0.11, "grad_norm": 16.163135072367126, "learning_rate": 1.9997918164856728e-05, "loss": 0.9591, "step": 698 }, { "epoch": 0.11, "grad_norm": 22.240807100852727, "learning_rate": 1.9997883613385184e-05, "loss": 0.9251, "step": 699 }, { "epoch": 0.11, "grad_norm": 28.1818260352272, "learning_rate": 1.999784877757968e-05, "loss": 0.9941, "step": 700 }, { "epoch": 0.11, "grad_norm": 20.27154914499387, "learning_rate": 1.99978136574412e-05, "loss": 0.9307, "step": 701 }, { "epoch": 0.11, "grad_norm": 25.716527018880733, "learning_rate": 1.999777825297075e-05, "loss": 1.0928, "step": 702 }, { "epoch": 0.11, "grad_norm": 20.930286027268007, "learning_rate": 1.9997742564169335e-05, "loss": 0.9775, "step": 703 }, { "epoch": 0.11, "grad_norm": 32.953649056374836, "learning_rate": 1.999770659103797e-05, "loss": 1.0377, "step": 704 }, { "epoch": 0.11, "grad_norm": 31.720956645746604, "learning_rate": 1.999767033357768e-05, "loss": 1.136, "step": 705 }, { "epoch": 0.11, "grad_norm": 23.675923263181794, "learning_rate": 1.9997633791789496e-05, "loss": 0.9852, "step": 706 }, { "epoch": 0.11, "grad_norm": 19.159710022438254, "learning_rate": 1.999759696567445e-05, "loss": 0.9578, "step": 707 }, { "epoch": 0.11, "grad_norm": 20.615629888993382, "learning_rate": 1.9997559855233596e-05, "loss": 0.983, "step": 708 }, { "epoch": 0.11, "grad_norm": 33.146591727613576, "learning_rate": 1.999752246046799e-05, "loss": 1.1633, "step": 709 }, { "epoch": 0.11, "grad_norm": 18.228318464618354, "learning_rate": 1.9997484781378694e-05, "loss": 1.0657, "step": 710 }, { "epoch": 0.11, "grad_norm": 18.935915363694495, "learning_rate": 1.9997446817966776e-05, "loss": 1.0634, "step": 711 }, { "epoch": 0.11, "grad_norm": 23.081959929447248, "learning_rate": 1.9997408570233322e-05, "loss": 1.0599, "step": 712 }, { "epoch": 0.11, "grad_norm": 24.53073736736404, "learning_rate": 1.9997370038179414e-05, "loss": 1.1398, "step": 713 }, { "epoch": 0.11, "grad_norm": 21.904704015913204, "learning_rate": 1.9997331221806152e-05, "loss": 0.9441, "step": 714 }, { "epoch": 0.11, "grad_norm": 19.297954312924443, "learning_rate": 1.999729212111464e-05, "loss": 0.9148, "step": 715 }, { "epoch": 0.11, "grad_norm": 21.77771079126166, "learning_rate": 1.9997252736105985e-05, "loss": 0.9018, "step": 716 }, { "epoch": 0.11, "grad_norm": 21.4628075742039, "learning_rate": 1.9997213066781312e-05, "loss": 0.9501, "step": 717 }, { "epoch": 0.11, "grad_norm": 25.887196421443424, "learning_rate": 1.9997173113141747e-05, "loss": 1.0388, "step": 718 }, { "epoch": 0.11, "grad_norm": 20.88366988841301, "learning_rate": 1.9997132875188427e-05, "loss": 0.8158, "step": 719 }, { "epoch": 0.11, "grad_norm": 37.05347981778112, "learning_rate": 1.9997092352922495e-05, "loss": 1.176, "step": 720 }, { "epoch": 0.11, "grad_norm": 31.72231868155152, "learning_rate": 1.999705154634511e-05, "loss": 1.2076, "step": 721 }, { "epoch": 0.11, "grad_norm": 24.056440302541347, "learning_rate": 1.999701045545742e-05, "loss": 1.0438, "step": 722 }, { "epoch": 0.11, "grad_norm": 28.075445910281495, "learning_rate": 1.9996969080260604e-05, "loss": 1.1965, "step": 723 }, { "epoch": 0.11, "grad_norm": 26.17635262973931, "learning_rate": 1.999692742075584e-05, "loss": 1.0819, "step": 724 }, { "epoch": 0.11, "grad_norm": 32.90122458402218, "learning_rate": 1.99968854769443e-05, "loss": 0.9422, "step": 725 }, { "epoch": 0.11, "grad_norm": 25.10317667173797, "learning_rate": 1.999684324882719e-05, "loss": 1.031, "step": 726 }, { "epoch": 0.11, "grad_norm": 24.35435118631165, "learning_rate": 1.9996800736405702e-05, "loss": 1.0282, "step": 727 }, { "epoch": 0.11, "grad_norm": 33.61465806374499, "learning_rate": 1.9996757939681052e-05, "loss": 1.1021, "step": 728 }, { "epoch": 0.11, "grad_norm": 29.424962678920473, "learning_rate": 1.9996714858654456e-05, "loss": 1.095, "step": 729 }, { "epoch": 0.11, "grad_norm": 23.8259090862828, "learning_rate": 1.9996671493327135e-05, "loss": 0.9476, "step": 730 }, { "epoch": 0.11, "grad_norm": 17.119920837066687, "learning_rate": 1.9996627843700325e-05, "loss": 0.9104, "step": 731 }, { "epoch": 0.11, "grad_norm": 27.554752234049488, "learning_rate": 1.9996583909775267e-05, "loss": 1.0307, "step": 732 }, { "epoch": 0.11, "grad_norm": 18.391961434436325, "learning_rate": 1.999653969155321e-05, "loss": 1.0621, "step": 733 }, { "epoch": 0.11, "grad_norm": 44.55982382056265, "learning_rate": 1.999649518903541e-05, "loss": 1.0835, "step": 734 }, { "epoch": 0.11, "grad_norm": 19.406889639247517, "learning_rate": 1.9996450402223137e-05, "loss": 0.9581, "step": 735 }, { "epoch": 0.11, "grad_norm": 20.53286648400594, "learning_rate": 1.9996405331117662e-05, "loss": 1.0734, "step": 736 }, { "epoch": 0.12, "grad_norm": 20.201748348929918, "learning_rate": 1.999635997572027e-05, "loss": 1.0315, "step": 737 }, { "epoch": 0.12, "grad_norm": 31.566733431409812, "learning_rate": 1.9996314336032243e-05, "loss": 1.0203, "step": 738 }, { "epoch": 0.12, "grad_norm": 25.12963293665336, "learning_rate": 1.9996268412054887e-05, "loss": 0.9966, "step": 739 }, { "epoch": 0.12, "grad_norm": 6.609661739748793, "learning_rate": 1.9996222203789504e-05, "loss": 0.9146, "step": 740 }, { "epoch": 0.12, "grad_norm": 25.767601571223842, "learning_rate": 1.9996175711237406e-05, "loss": 1.0671, "step": 741 }, { "epoch": 0.12, "grad_norm": 38.12139600385158, "learning_rate": 1.9996128934399923e-05, "loss": 0.9169, "step": 742 }, { "epoch": 0.12, "grad_norm": 36.19068369218754, "learning_rate": 1.999608187327838e-05, "loss": 1.0309, "step": 743 }, { "epoch": 0.12, "grad_norm": 38.01410823168409, "learning_rate": 1.9996034527874117e-05, "loss": 0.951, "step": 744 }, { "epoch": 0.12, "grad_norm": 35.589617285805026, "learning_rate": 1.9995986898188477e-05, "loss": 1.0413, "step": 745 }, { "epoch": 0.12, "grad_norm": 29.23197348215027, "learning_rate": 1.9995938984222815e-05, "loss": 1.0136, "step": 746 }, { "epoch": 0.12, "grad_norm": 34.717249418117866, "learning_rate": 1.99958907859785e-05, "loss": 1.1298, "step": 747 }, { "epoch": 0.12, "grad_norm": 30.097320466640863, "learning_rate": 1.99958423034569e-05, "loss": 0.9868, "step": 748 }, { "epoch": 0.12, "grad_norm": 30.448825783240018, "learning_rate": 1.9995793536659388e-05, "loss": 1.1305, "step": 749 }, { "epoch": 0.12, "grad_norm": 25.343293259970014, "learning_rate": 1.9995744485587356e-05, "loss": 0.9858, "step": 750 }, { "epoch": 0.12, "grad_norm": 22.89076918989473, "learning_rate": 1.99956951502422e-05, "loss": 0.9851, "step": 751 }, { "epoch": 0.12, "grad_norm": 39.1549199883858, "learning_rate": 1.999564553062532e-05, "loss": 1.0863, "step": 752 }, { "epoch": 0.12, "grad_norm": 24.20988632682434, "learning_rate": 1.9995595626738128e-05, "loss": 0.9575, "step": 753 }, { "epoch": 0.12, "grad_norm": 28.49339147910342, "learning_rate": 1.9995545438582044e-05, "loss": 0.9938, "step": 754 }, { "epoch": 0.12, "grad_norm": 19.020720661897847, "learning_rate": 1.9995494966158494e-05, "loss": 0.9864, "step": 755 }, { "epoch": 0.12, "grad_norm": 27.414747939478183, "learning_rate": 1.9995444209468916e-05, "loss": 1.0082, "step": 756 }, { "epoch": 0.12, "grad_norm": 20.755829020368772, "learning_rate": 1.999539316851475e-05, "loss": 0.9256, "step": 757 }, { "epoch": 0.12, "grad_norm": 24.231491864276187, "learning_rate": 1.999534184329745e-05, "loss": 0.9442, "step": 758 }, { "epoch": 0.12, "grad_norm": 25.269750093204454, "learning_rate": 1.9995290233818475e-05, "loss": 0.9794, "step": 759 }, { "epoch": 0.12, "grad_norm": 25.29033883226424, "learning_rate": 1.9995238340079295e-05, "loss": 1.0591, "step": 760 }, { "epoch": 0.12, "grad_norm": 31.637327954914582, "learning_rate": 1.9995186162081384e-05, "loss": 1.1424, "step": 761 }, { "epoch": 0.12, "grad_norm": 28.39544905269339, "learning_rate": 1.9995133699826222e-05, "loss": 1.1801, "step": 762 }, { "epoch": 0.12, "grad_norm": 37.082381837224915, "learning_rate": 1.999508095331531e-05, "loss": 1.0525, "step": 763 }, { "epoch": 0.12, "grad_norm": 31.00075740331179, "learning_rate": 1.9995027922550137e-05, "loss": 0.9518, "step": 764 }, { "epoch": 0.12, "grad_norm": 28.91964572175146, "learning_rate": 1.999497460753222e-05, "loss": 1.0819, "step": 765 }, { "epoch": 0.12, "grad_norm": 15.666236160603669, "learning_rate": 1.9994921008263072e-05, "loss": 0.8689, "step": 766 }, { "epoch": 0.12, "grad_norm": 16.07462339621955, "learning_rate": 1.9994867124744216e-05, "loss": 0.8584, "step": 767 }, { "epoch": 0.12, "grad_norm": 29.995975931807806, "learning_rate": 1.9994812956977183e-05, "loss": 1.0508, "step": 768 }, { "epoch": 0.12, "grad_norm": 15.138358958316012, "learning_rate": 1.9994758504963522e-05, "loss": 0.9026, "step": 769 }, { "epoch": 0.12, "grad_norm": 41.30043257083839, "learning_rate": 1.9994703768704773e-05, "loss": 1.0871, "step": 770 }, { "epoch": 0.12, "grad_norm": 20.84981941206687, "learning_rate": 1.9994648748202493e-05, "loss": 0.9018, "step": 771 }, { "epoch": 0.12, "grad_norm": 19.056270868284336, "learning_rate": 1.9994593443458252e-05, "loss": 0.9941, "step": 772 }, { "epoch": 0.12, "grad_norm": 18.231511167349787, "learning_rate": 1.999453785447362e-05, "loss": 0.9491, "step": 773 }, { "epoch": 0.12, "grad_norm": 19.790377846130333, "learning_rate": 1.999448198125018e-05, "loss": 0.957, "step": 774 }, { "epoch": 0.12, "grad_norm": 32.21334604847917, "learning_rate": 1.9994425823789517e-05, "loss": 1.1117, "step": 775 }, { "epoch": 0.12, "grad_norm": 40.859264248174526, "learning_rate": 1.999436938209323e-05, "loss": 1.0816, "step": 776 }, { "epoch": 0.12, "grad_norm": 22.681576869897665, "learning_rate": 1.9994312656162928e-05, "loss": 0.9506, "step": 777 }, { "epoch": 0.12, "grad_norm": 15.761793379861027, "learning_rate": 1.9994255646000217e-05, "loss": 0.9511, "step": 778 }, { "epoch": 0.12, "grad_norm": 23.009516921325208, "learning_rate": 1.999419835160672e-05, "loss": 1.1356, "step": 779 }, { "epoch": 0.12, "grad_norm": 29.058749703471978, "learning_rate": 1.999414077298407e-05, "loss": 0.9984, "step": 780 }, { "epoch": 0.12, "grad_norm": 12.967857301390442, "learning_rate": 1.9994082910133903e-05, "loss": 0.9311, "step": 781 }, { "epoch": 0.12, "grad_norm": 25.28353516797878, "learning_rate": 1.9994024763057865e-05, "loss": 0.9967, "step": 782 }, { "epoch": 0.12, "grad_norm": 28.361055374804256, "learning_rate": 1.9993966331757607e-05, "loss": 0.9965, "step": 783 }, { "epoch": 0.12, "grad_norm": 17.399942802295453, "learning_rate": 1.9993907616234796e-05, "loss": 1.1351, "step": 784 }, { "epoch": 0.12, "grad_norm": 27.1752552356069, "learning_rate": 1.9993848616491097e-05, "loss": 1.0399, "step": 785 }, { "epoch": 0.12, "grad_norm": 19.665572190710133, "learning_rate": 1.9993789332528193e-05, "loss": 0.9757, "step": 786 }, { "epoch": 0.12, "grad_norm": 17.87341129638875, "learning_rate": 1.9993729764347763e-05, "loss": 0.9788, "step": 787 }, { "epoch": 0.12, "grad_norm": 13.404118302845683, "learning_rate": 1.9993669911951504e-05, "loss": 0.9547, "step": 788 }, { "epoch": 0.12, "grad_norm": 23.788661705048366, "learning_rate": 1.999360977534112e-05, "loss": 0.9624, "step": 789 }, { "epoch": 0.12, "grad_norm": 26.020252454309638, "learning_rate": 1.999354935451832e-05, "loss": 1.0428, "step": 790 }, { "epoch": 0.12, "grad_norm": 17.643653534481263, "learning_rate": 1.999348864948482e-05, "loss": 1.0729, "step": 791 }, { "epoch": 0.12, "grad_norm": 29.491574148911653, "learning_rate": 1.9993427660242356e-05, "loss": 1.0208, "step": 792 }, { "epoch": 0.12, "grad_norm": 21.411371135016022, "learning_rate": 1.999336638679265e-05, "loss": 1.0212, "step": 793 }, { "epoch": 0.12, "grad_norm": 14.16741036518218, "learning_rate": 1.999330482913745e-05, "loss": 0.9471, "step": 794 }, { "epoch": 0.12, "grad_norm": 26.060560856406358, "learning_rate": 1.9993242987278508e-05, "loss": 1.0415, "step": 795 }, { "epoch": 0.12, "grad_norm": 24.240362704137674, "learning_rate": 1.999318086121758e-05, "loss": 0.8658, "step": 796 }, { "epoch": 0.12, "grad_norm": 33.59682973451131, "learning_rate": 1.9993118450956434e-05, "loss": 1.0131, "step": 797 }, { "epoch": 0.12, "grad_norm": 24.413119608867582, "learning_rate": 1.9993055756496845e-05, "loss": 0.8949, "step": 798 }, { "epoch": 0.12, "grad_norm": 22.063337245324533, "learning_rate": 1.9992992777840596e-05, "loss": 1.0127, "step": 799 }, { "epoch": 0.12, "grad_norm": 32.5013716032295, "learning_rate": 1.999292951498948e-05, "loss": 0.9796, "step": 800 }, { "epoch": 0.13, "grad_norm": 21.4678596495198, "learning_rate": 1.9992865967945295e-05, "loss": 0.8518, "step": 801 }, { "epoch": 0.13, "grad_norm": 24.86441435108581, "learning_rate": 1.9992802136709842e-05, "loss": 0.9583, "step": 802 }, { "epoch": 0.13, "grad_norm": 33.19805786961399, "learning_rate": 1.999273802128495e-05, "loss": 1.0568, "step": 803 }, { "epoch": 0.13, "grad_norm": 30.22292863501949, "learning_rate": 1.9992673621672427e-05, "loss": 1.0313, "step": 804 }, { "epoch": 0.13, "grad_norm": 23.81481619456222, "learning_rate": 1.9992608937874115e-05, "loss": 1.0087, "step": 805 }, { "epoch": 0.13, "grad_norm": 20.22193855322229, "learning_rate": 1.999254396989185e-05, "loss": 1.0421, "step": 806 }, { "epoch": 0.13, "grad_norm": 27.403442756654155, "learning_rate": 1.9992478717727478e-05, "loss": 1.0045, "step": 807 }, { "epoch": 0.13, "grad_norm": 28.41463838234398, "learning_rate": 1.999241318138286e-05, "loss": 0.9212, "step": 808 }, { "epoch": 0.13, "grad_norm": 22.336751760909568, "learning_rate": 1.9992347360859858e-05, "loss": 0.9983, "step": 809 }, { "epoch": 0.13, "grad_norm": 21.56880793631396, "learning_rate": 1.9992281256160337e-05, "loss": 0.9342, "step": 810 }, { "epoch": 0.13, "grad_norm": 15.966927639020398, "learning_rate": 1.9992214867286182e-05, "loss": 0.8636, "step": 811 }, { "epoch": 0.13, "grad_norm": 15.393939609188461, "learning_rate": 1.9992148194239287e-05, "loss": 0.9692, "step": 812 }, { "epoch": 0.13, "grad_norm": 19.863124719263677, "learning_rate": 1.999208123702154e-05, "loss": 1.0059, "step": 813 }, { "epoch": 0.13, "grad_norm": 25.308754605893263, "learning_rate": 1.999201399563485e-05, "loss": 1.0584, "step": 814 }, { "epoch": 0.13, "grad_norm": 19.94069254585308, "learning_rate": 1.9991946470081124e-05, "loss": 0.9221, "step": 815 }, { "epoch": 0.13, "grad_norm": 26.48242480544358, "learning_rate": 1.9991878660362285e-05, "loss": 1.0481, "step": 816 }, { "epoch": 0.13, "grad_norm": 28.650056717809818, "learning_rate": 1.9991810566480264e-05, "loss": 0.9757, "step": 817 }, { "epoch": 0.13, "grad_norm": 20.304980313698515, "learning_rate": 1.9991742188436992e-05, "loss": 0.9911, "step": 818 }, { "epoch": 0.13, "grad_norm": 26.18954040059897, "learning_rate": 1.999167352623442e-05, "loss": 1.0247, "step": 819 }, { "epoch": 0.13, "grad_norm": 29.615591504789602, "learning_rate": 1.99916045798745e-05, "loss": 0.9229, "step": 820 }, { "epoch": 0.13, "grad_norm": 15.042097360141371, "learning_rate": 1.999153534935919e-05, "loss": 0.9028, "step": 821 }, { "epoch": 0.13, "grad_norm": 19.36765037774426, "learning_rate": 1.999146583469046e-05, "loss": 0.9344, "step": 822 }, { "epoch": 0.13, "grad_norm": 21.649725520853114, "learning_rate": 1.9991396035870282e-05, "loss": 0.9665, "step": 823 }, { "epoch": 0.13, "grad_norm": 4.752910489352573, "learning_rate": 1.999132595290065e-05, "loss": 1.0209, "step": 824 }, { "epoch": 0.13, "grad_norm": 20.927958070642642, "learning_rate": 1.9991255585783547e-05, "loss": 1.0627, "step": 825 }, { "epoch": 0.13, "grad_norm": 24.393859252958695, "learning_rate": 1.9991184934520987e-05, "loss": 0.9001, "step": 826 }, { "epoch": 0.13, "grad_norm": 17.021183249807795, "learning_rate": 1.9991113999114966e-05, "loss": 0.9242, "step": 827 }, { "epoch": 0.13, "grad_norm": 5.814606630902126, "learning_rate": 1.9991042779567512e-05, "loss": 1.0353, "step": 828 }, { "epoch": 0.13, "grad_norm": 21.908477894047138, "learning_rate": 1.9990971275880643e-05, "loss": 0.883, "step": 829 }, { "epoch": 0.13, "grad_norm": 24.140590699180752, "learning_rate": 1.9990899488056396e-05, "loss": 0.9375, "step": 830 }, { "epoch": 0.13, "grad_norm": 21.82090466464421, "learning_rate": 1.9990827416096813e-05, "loss": 0.9946, "step": 831 }, { "epoch": 0.13, "grad_norm": 20.062615461106372, "learning_rate": 1.9990755060003945e-05, "loss": 0.939, "step": 832 }, { "epoch": 0.13, "grad_norm": 24.180641766823623, "learning_rate": 1.9990682419779844e-05, "loss": 0.9904, "step": 833 }, { "epoch": 0.13, "grad_norm": 16.594210069190982, "learning_rate": 1.9990609495426582e-05, "loss": 0.8201, "step": 834 }, { "epoch": 0.13, "grad_norm": 23.217124454000132, "learning_rate": 1.9990536286946227e-05, "loss": 0.9544, "step": 835 }, { "epoch": 0.13, "grad_norm": 17.726451528781027, "learning_rate": 1.9990462794340864e-05, "loss": 0.9834, "step": 836 }, { "epoch": 0.13, "grad_norm": 17.424770998493283, "learning_rate": 1.9990389017612587e-05, "loss": 0.9333, "step": 837 }, { "epoch": 0.13, "grad_norm": 31.88116710317783, "learning_rate": 1.999031495676349e-05, "loss": 1.0023, "step": 838 }, { "epoch": 0.13, "grad_norm": 29.989587294002025, "learning_rate": 1.999024061179568e-05, "loss": 0.9795, "step": 839 }, { "epoch": 0.13, "grad_norm": 19.427036379858826, "learning_rate": 1.9990165982711266e-05, "loss": 0.9978, "step": 840 }, { "epoch": 0.13, "grad_norm": 18.00849579116745, "learning_rate": 1.999009106951238e-05, "loss": 0.932, "step": 841 }, { "epoch": 0.13, "grad_norm": 16.269670381901356, "learning_rate": 1.9990015872201147e-05, "loss": 1.0195, "step": 842 }, { "epoch": 0.13, "grad_norm": 23.72704759716728, "learning_rate": 1.998994039077971e-05, "loss": 0.9288, "step": 843 }, { "epoch": 0.13, "grad_norm": 21.77555586998277, "learning_rate": 1.9989864625250206e-05, "loss": 0.91, "step": 844 }, { "epoch": 0.13, "grad_norm": 17.9001177653043, "learning_rate": 1.99897885756148e-05, "loss": 0.9871, "step": 845 }, { "epoch": 0.13, "grad_norm": 18.365308538926303, "learning_rate": 1.998971224187565e-05, "loss": 0.9918, "step": 846 }, { "epoch": 0.13, "grad_norm": 6.15001487553259, "learning_rate": 1.998963562403493e-05, "loss": 0.9024, "step": 847 }, { "epoch": 0.13, "grad_norm": 30.6216208734604, "learning_rate": 1.9989558722094813e-05, "loss": 1.1355, "step": 848 }, { "epoch": 0.13, "grad_norm": 17.538922120278034, "learning_rate": 1.998948153605749e-05, "loss": 0.9668, "step": 849 }, { "epoch": 0.13, "grad_norm": 23.142611519432652, "learning_rate": 1.9989404065925157e-05, "loss": 0.9824, "step": 850 }, { "epoch": 0.13, "grad_norm": 18.998781814548558, "learning_rate": 1.9989326311700016e-05, "loss": 0.9729, "step": 851 }, { "epoch": 0.13, "grad_norm": 24.815911225495, "learning_rate": 1.9989248273384278e-05, "loss": 1.1024, "step": 852 }, { "epoch": 0.13, "grad_norm": 22.568448812719364, "learning_rate": 1.998916995098016e-05, "loss": 1.0269, "step": 853 }, { "epoch": 0.13, "grad_norm": 37.97529794878723, "learning_rate": 1.9989091344489896e-05, "loss": 1.1285, "step": 854 }, { "epoch": 0.13, "grad_norm": 18.093833307130453, "learning_rate": 1.9989012453915718e-05, "loss": 0.9447, "step": 855 }, { "epoch": 0.13, "grad_norm": 21.977681582043218, "learning_rate": 1.998893327925987e-05, "loss": 1.0056, "step": 856 }, { "epoch": 0.13, "grad_norm": 26.978096634156046, "learning_rate": 1.9988853820524596e-05, "loss": 0.9743, "step": 857 }, { "epoch": 0.13, "grad_norm": 11.30994013082935, "learning_rate": 1.998877407771217e-05, "loss": 1.0, "step": 858 }, { "epoch": 0.13, "grad_norm": 22.29930427913624, "learning_rate": 1.9988694050824847e-05, "loss": 1.0081, "step": 859 }, { "epoch": 0.13, "grad_norm": 22.667376292754565, "learning_rate": 1.998861373986491e-05, "loss": 0.8904, "step": 860 }, { "epoch": 0.13, "grad_norm": 25.86367086933168, "learning_rate": 1.9988533144834642e-05, "loss": 1.1446, "step": 861 }, { "epoch": 0.13, "grad_norm": 29.355136942558662, "learning_rate": 1.9988452265736335e-05, "loss": 0.9395, "step": 862 }, { "epoch": 0.13, "grad_norm": 27.30893269829037, "learning_rate": 1.998837110257229e-05, "loss": 0.961, "step": 863 }, { "epoch": 0.13, "grad_norm": 23.567106449316867, "learning_rate": 1.9988289655344814e-05, "loss": 0.9266, "step": 864 }, { "epoch": 0.14, "grad_norm": 25.23541728308947, "learning_rate": 1.998820792405622e-05, "loss": 0.8713, "step": 865 }, { "epoch": 0.14, "grad_norm": 17.765799393213108, "learning_rate": 1.998812590870884e-05, "loss": 0.9929, "step": 866 }, { "epoch": 0.14, "grad_norm": 27.135648330436652, "learning_rate": 1.9988043609304995e-05, "loss": 0.9062, "step": 867 }, { "epoch": 0.14, "grad_norm": 29.160797839762104, "learning_rate": 1.9987961025847037e-05, "loss": 1.0473, "step": 868 }, { "epoch": 0.14, "grad_norm": 28.63223994100243, "learning_rate": 1.998787815833731e-05, "loss": 0.9395, "step": 869 }, { "epoch": 0.14, "grad_norm": 37.465919881683654, "learning_rate": 1.9987795006778173e-05, "loss": 0.9055, "step": 870 }, { "epoch": 0.14, "grad_norm": 19.221696657207115, "learning_rate": 1.9987711571171987e-05, "loss": 1.0072, "step": 871 }, { "epoch": 0.14, "grad_norm": 34.105648468036904, "learning_rate": 1.9987627851521126e-05, "loss": 1.0029, "step": 872 }, { "epoch": 0.14, "grad_norm": 28.93263817186818, "learning_rate": 1.998754384782797e-05, "loss": 1.0123, "step": 873 }, { "epoch": 0.14, "grad_norm": 24.908467970607894, "learning_rate": 1.998745956009491e-05, "loss": 0.9502, "step": 874 }, { "epoch": 0.14, "grad_norm": 13.42165320838223, "learning_rate": 1.9987374988324345e-05, "loss": 0.9002, "step": 875 }, { "epoch": 0.14, "grad_norm": 20.078146314795195, "learning_rate": 1.9987290132518677e-05, "loss": 0.9438, "step": 876 }, { "epoch": 0.14, "grad_norm": 33.27848961893949, "learning_rate": 1.9987204992680322e-05, "loss": 0.9485, "step": 877 }, { "epoch": 0.14, "grad_norm": 17.26684856394545, "learning_rate": 1.9987119568811693e-05, "loss": 1.0242, "step": 878 }, { "epoch": 0.14, "grad_norm": 20.897137022108495, "learning_rate": 1.998703386091523e-05, "loss": 0.868, "step": 879 }, { "epoch": 0.14, "grad_norm": 20.083151933156167, "learning_rate": 1.9986947868993366e-05, "loss": 0.9719, "step": 880 }, { "epoch": 0.14, "grad_norm": 23.88446296027853, "learning_rate": 1.998686159304855e-05, "loss": 0.8459, "step": 881 }, { "epoch": 0.14, "grad_norm": 22.437631533617076, "learning_rate": 1.9986775033083225e-05, "loss": 0.861, "step": 882 }, { "epoch": 0.14, "grad_norm": 18.935485884994645, "learning_rate": 1.9986688189099865e-05, "loss": 0.9921, "step": 883 }, { "epoch": 0.14, "grad_norm": 21.621648396422263, "learning_rate": 1.9986601061100934e-05, "loss": 0.9493, "step": 884 }, { "epoch": 0.14, "grad_norm": 18.00615028190211, "learning_rate": 1.9986513649088915e-05, "loss": 0.9605, "step": 885 }, { "epoch": 0.14, "grad_norm": 25.535952044605807, "learning_rate": 1.9986425953066284e-05, "loss": 0.9192, "step": 886 }, { "epoch": 0.14, "grad_norm": 28.33890047192677, "learning_rate": 1.9986337973035542e-05, "loss": 1.1137, "step": 887 }, { "epoch": 0.14, "grad_norm": 28.633470350257046, "learning_rate": 1.9986249708999194e-05, "loss": 1.0084, "step": 888 }, { "epoch": 0.14, "grad_norm": 30.547819182250514, "learning_rate": 1.998616116095974e-05, "loss": 0.8784, "step": 889 }, { "epoch": 0.14, "grad_norm": 20.728594514284588, "learning_rate": 1.9986072328919707e-05, "loss": 0.929, "step": 890 }, { "epoch": 0.14, "grad_norm": 20.206547964184086, "learning_rate": 1.9985983212881618e-05, "loss": 1.0093, "step": 891 }, { "epoch": 0.14, "grad_norm": 27.868262481702807, "learning_rate": 1.9985893812848007e-05, "loss": 0.9966, "step": 892 }, { "epoch": 0.14, "grad_norm": 18.120365907755325, "learning_rate": 1.998580412882142e-05, "loss": 0.9585, "step": 893 }, { "epoch": 0.14, "grad_norm": 23.40060378094494, "learning_rate": 1.9985714160804406e-05, "loss": 0.951, "step": 894 }, { "epoch": 0.14, "grad_norm": 63.25595737001749, "learning_rate": 1.9985623908799518e-05, "loss": 1.0075, "step": 895 }, { "epoch": 0.14, "grad_norm": 21.184525813103157, "learning_rate": 1.9985533372809335e-05, "loss": 0.9598, "step": 896 }, { "epoch": 0.14, "grad_norm": 25.37430623943003, "learning_rate": 1.9985442552836417e-05, "loss": 1.0131, "step": 897 }, { "epoch": 0.14, "grad_norm": 15.614081869721268, "learning_rate": 1.9985351448883358e-05, "loss": 0.8407, "step": 898 }, { "epoch": 0.14, "grad_norm": 40.66036947009204, "learning_rate": 1.9985260060952742e-05, "loss": 1.1067, "step": 899 }, { "epoch": 0.14, "grad_norm": 23.771311306654475, "learning_rate": 1.9985168389047175e-05, "loss": 0.9295, "step": 900 }, { "epoch": 0.14, "grad_norm": 19.97576504675759, "learning_rate": 1.9985076433169258e-05, "loss": 0.9276, "step": 901 }, { "epoch": 0.14, "grad_norm": 33.185904152894416, "learning_rate": 1.9984984193321608e-05, "loss": 0.9348, "step": 902 }, { "epoch": 0.14, "grad_norm": 25.139107118380018, "learning_rate": 1.9984891669506847e-05, "loss": 1.0378, "step": 903 }, { "epoch": 0.14, "grad_norm": 29.13446842949091, "learning_rate": 1.9984798861727614e-05, "loss": 1.1061, "step": 904 }, { "epoch": 0.14, "grad_norm": 33.56381521010012, "learning_rate": 1.9984705769986538e-05, "loss": 0.9512, "step": 905 }, { "epoch": 0.14, "grad_norm": 18.831056176741296, "learning_rate": 1.998461239428627e-05, "loss": 0.9442, "step": 906 }, { "epoch": 0.14, "grad_norm": 25.284614438729086, "learning_rate": 1.9984518734629467e-05, "loss": 1.0857, "step": 907 }, { "epoch": 0.14, "grad_norm": 4.470259374024849, "learning_rate": 1.998442479101879e-05, "loss": 0.8212, "step": 908 }, { "epoch": 0.14, "grad_norm": 18.079067228749622, "learning_rate": 1.9984330563456918e-05, "loss": 1.0071, "step": 909 }, { "epoch": 0.14, "grad_norm": 18.548201359288534, "learning_rate": 1.998423605194652e-05, "loss": 0.9289, "step": 910 }, { "epoch": 0.14, "grad_norm": 22.867643656089143, "learning_rate": 1.9984141256490294e-05, "loss": 0.8817, "step": 911 }, { "epoch": 0.14, "grad_norm": 25.75903194116444, "learning_rate": 1.9984046177090926e-05, "loss": 0.9227, "step": 912 }, { "epoch": 0.14, "grad_norm": 20.776527642640172, "learning_rate": 1.998395081375113e-05, "loss": 0.9352, "step": 913 }, { "epoch": 0.14, "grad_norm": 17.641440914128953, "learning_rate": 1.9983855166473613e-05, "loss": 0.9707, "step": 914 }, { "epoch": 0.14, "grad_norm": 16.1738760286084, "learning_rate": 1.998375923526109e-05, "loss": 0.9254, "step": 915 }, { "epoch": 0.14, "grad_norm": 25.244305857370346, "learning_rate": 1.99836630201163e-05, "loss": 0.9369, "step": 916 }, { "epoch": 0.14, "grad_norm": 19.632949080259692, "learning_rate": 1.9983566521041973e-05, "loss": 1.0031, "step": 917 }, { "epoch": 0.14, "grad_norm": 26.584330628760394, "learning_rate": 1.9983469738040852e-05, "loss": 0.9407, "step": 918 }, { "epoch": 0.14, "grad_norm": 26.7209982779378, "learning_rate": 1.9983372671115696e-05, "loss": 0.8765, "step": 919 }, { "epoch": 0.14, "grad_norm": 20.68512370905215, "learning_rate": 1.998327532026926e-05, "loss": 0.9519, "step": 920 }, { "epoch": 0.14, "grad_norm": 20.818700823930502, "learning_rate": 1.9983177685504312e-05, "loss": 1.0332, "step": 921 }, { "epoch": 0.14, "grad_norm": 22.28754939570171, "learning_rate": 1.998307976682363e-05, "loss": 0.9153, "step": 922 }, { "epoch": 0.14, "grad_norm": 26.84776960948699, "learning_rate": 1.998298156423e-05, "loss": 0.9509, "step": 923 }, { "epoch": 0.14, "grad_norm": 17.2049697740919, "learning_rate": 1.9982883077726218e-05, "loss": 0.9338, "step": 924 }, { "epoch": 0.14, "grad_norm": 18.359839754106535, "learning_rate": 1.998278430731508e-05, "loss": 0.9355, "step": 925 }, { "epoch": 0.14, "grad_norm": 25.97638165661955, "learning_rate": 1.9982685252999393e-05, "loss": 0.8382, "step": 926 }, { "epoch": 0.14, "grad_norm": 17.32619890080394, "learning_rate": 1.998258591478198e-05, "loss": 0.959, "step": 927 }, { "epoch": 0.14, "grad_norm": 20.89415149625322, "learning_rate": 1.998248629266566e-05, "loss": 0.8957, "step": 928 }, { "epoch": 0.15, "grad_norm": 31.943558460292774, "learning_rate": 1.9982386386653273e-05, "loss": 0.9798, "step": 929 }, { "epoch": 0.15, "grad_norm": 24.017375857712963, "learning_rate": 1.9982286196747653e-05, "loss": 0.8992, "step": 930 }, { "epoch": 0.15, "grad_norm": 42.161691614895396, "learning_rate": 1.9982185722951655e-05, "loss": 1.0017, "step": 931 }, { "epoch": 0.15, "grad_norm": 20.98815800644647, "learning_rate": 1.9982084965268134e-05, "loss": 0.9948, "step": 932 }, { "epoch": 0.15, "grad_norm": 19.64485681263662, "learning_rate": 1.9981983923699956e-05, "loss": 0.9194, "step": 933 }, { "epoch": 0.15, "grad_norm": 16.91240020274571, "learning_rate": 1.9981882598249996e-05, "loss": 0.9707, "step": 934 }, { "epoch": 0.15, "grad_norm": 35.70244139970563, "learning_rate": 1.998178098892113e-05, "loss": 0.9954, "step": 935 }, { "epoch": 0.15, "grad_norm": 19.285258117678723, "learning_rate": 1.998167909571626e-05, "loss": 0.9304, "step": 936 }, { "epoch": 0.15, "grad_norm": 14.17099899625401, "learning_rate": 1.998157691863827e-05, "loss": 0.9262, "step": 937 }, { "epoch": 0.15, "grad_norm": 19.25734694190907, "learning_rate": 1.998147445769007e-05, "loss": 1.0516, "step": 938 }, { "epoch": 0.15, "grad_norm": 22.78016361673385, "learning_rate": 1.9981371712874576e-05, "loss": 0.8487, "step": 939 }, { "epoch": 0.15, "grad_norm": 25.60884712680659, "learning_rate": 1.9981268684194712e-05, "loss": 0.8976, "step": 940 }, { "epoch": 0.15, "grad_norm": 20.543851541489303, "learning_rate": 1.9981165371653404e-05, "loss": 1.0043, "step": 941 }, { "epoch": 0.15, "grad_norm": 23.880149045154408, "learning_rate": 1.998106177525359e-05, "loss": 0.9128, "step": 942 }, { "epoch": 0.15, "grad_norm": 22.143342445306903, "learning_rate": 1.998095789499822e-05, "loss": 1.0114, "step": 943 }, { "epoch": 0.15, "grad_norm": 19.71753376324862, "learning_rate": 1.9980853730890248e-05, "loss": 0.9462, "step": 944 }, { "epoch": 0.15, "grad_norm": 15.723056235586299, "learning_rate": 1.998074928293263e-05, "loss": 0.853, "step": 945 }, { "epoch": 0.15, "grad_norm": 15.305508550183559, "learning_rate": 1.9980644551128347e-05, "loss": 0.947, "step": 946 }, { "epoch": 0.15, "grad_norm": 26.7884693475016, "learning_rate": 1.9980539535480365e-05, "loss": 0.9528, "step": 947 }, { "epoch": 0.15, "grad_norm": 17.684553332254552, "learning_rate": 1.998043423599168e-05, "loss": 0.9229, "step": 948 }, { "epoch": 0.15, "grad_norm": 19.45037526225891, "learning_rate": 1.9980328652665285e-05, "loss": 0.93, "step": 949 }, { "epoch": 0.15, "grad_norm": 22.342525647370522, "learning_rate": 1.9980222785504178e-05, "loss": 0.8307, "step": 950 }, { "epoch": 0.15, "grad_norm": 15.949869235113345, "learning_rate": 1.9980116634511377e-05, "loss": 0.9576, "step": 951 }, { "epoch": 0.15, "grad_norm": 25.881216256166415, "learning_rate": 1.9980010199689893e-05, "loss": 0.9439, "step": 952 }, { "epoch": 0.15, "grad_norm": 21.92329329418154, "learning_rate": 1.997990348104276e-05, "loss": 1.0542, "step": 953 }, { "epoch": 0.15, "grad_norm": 15.721333365378953, "learning_rate": 1.997979647857301e-05, "loss": 0.9466, "step": 954 }, { "epoch": 0.15, "grad_norm": 19.140817454402058, "learning_rate": 1.9979689192283683e-05, "loss": 1.0371, "step": 955 }, { "epoch": 0.15, "grad_norm": 27.79666640560971, "learning_rate": 1.997958162217784e-05, "loss": 0.9012, "step": 956 }, { "epoch": 0.15, "grad_norm": 20.34264524598088, "learning_rate": 1.9979473768258528e-05, "loss": 0.9864, "step": 957 }, { "epoch": 0.15, "grad_norm": 13.877202936832614, "learning_rate": 1.9979365630528824e-05, "loss": 0.8326, "step": 958 }, { "epoch": 0.15, "grad_norm": 18.292802280839137, "learning_rate": 1.9979257208991796e-05, "loss": 1.0353, "step": 959 }, { "epoch": 0.15, "grad_norm": 21.998349341868792, "learning_rate": 1.9979148503650528e-05, "loss": 0.9579, "step": 960 }, { "epoch": 0.15, "grad_norm": 21.46610453602181, "learning_rate": 1.997903951450812e-05, "loss": 0.8989, "step": 961 }, { "epoch": 0.15, "grad_norm": 19.396557091882116, "learning_rate": 1.9978930241567662e-05, "loss": 1.072, "step": 962 }, { "epoch": 0.15, "grad_norm": 22.311744914247736, "learning_rate": 1.9978820684832268e-05, "loss": 0.8628, "step": 963 }, { "epoch": 0.15, "grad_norm": 21.23925123443074, "learning_rate": 1.997871084430505e-05, "loss": 0.9007, "step": 964 }, { "epoch": 0.15, "grad_norm": 15.155163351411675, "learning_rate": 1.997860071998913e-05, "loss": 0.8953, "step": 965 }, { "epoch": 0.15, "grad_norm": 27.784837092670617, "learning_rate": 1.997849031188765e-05, "loss": 0.9702, "step": 966 }, { "epoch": 0.15, "grad_norm": 22.56300106113026, "learning_rate": 1.9978379620003737e-05, "loss": 0.9037, "step": 967 }, { "epoch": 0.15, "grad_norm": 30.88380557334971, "learning_rate": 1.9978268644340545e-05, "loss": 0.9354, "step": 968 }, { "epoch": 0.15, "grad_norm": 24.44037004968756, "learning_rate": 1.997815738490123e-05, "loss": 0.9041, "step": 969 }, { "epoch": 0.15, "grad_norm": 15.061041783148795, "learning_rate": 1.997804584168896e-05, "loss": 0.8689, "step": 970 }, { "epoch": 0.15, "grad_norm": 23.00842304682578, "learning_rate": 1.9977934014706898e-05, "loss": 1.0025, "step": 971 }, { "epoch": 0.15, "grad_norm": 22.164024982279454, "learning_rate": 1.997782190395823e-05, "loss": 1.036, "step": 972 }, { "epoch": 0.15, "grad_norm": 23.160050762989524, "learning_rate": 1.9977709509446148e-05, "loss": 1.0333, "step": 973 }, { "epoch": 0.15, "grad_norm": 28.18236742319079, "learning_rate": 1.9977596831173843e-05, "loss": 0.9852, "step": 974 }, { "epoch": 0.15, "grad_norm": 21.215368577298406, "learning_rate": 1.9977483869144522e-05, "loss": 0.8775, "step": 975 }, { "epoch": 0.15, "grad_norm": 27.592001121889975, "learning_rate": 1.9977370623361393e-05, "loss": 1.0141, "step": 976 }, { "epoch": 0.15, "grad_norm": 22.88561976656229, "learning_rate": 1.997725709382768e-05, "loss": 0.9635, "step": 977 }, { "epoch": 0.15, "grad_norm": 24.50692665461813, "learning_rate": 1.9977143280546616e-05, "loss": 1.0061, "step": 978 }, { "epoch": 0.15, "grad_norm": 25.888740135974892, "learning_rate": 1.997702918352143e-05, "loss": 1.0239, "step": 979 }, { "epoch": 0.15, "grad_norm": 26.65698507822647, "learning_rate": 1.997691480275537e-05, "loss": 0.8671, "step": 980 }, { "epoch": 0.15, "grad_norm": 24.87193977279463, "learning_rate": 1.997680013825169e-05, "loss": 0.8638, "step": 981 }, { "epoch": 0.15, "grad_norm": 20.821167913617344, "learning_rate": 1.997668519001365e-05, "loss": 0.9667, "step": 982 }, { "epoch": 0.15, "grad_norm": 29.648484524776272, "learning_rate": 1.9976569958044523e-05, "loss": 0.985, "step": 983 }, { "epoch": 0.15, "grad_norm": 19.75908343510108, "learning_rate": 1.997645444234758e-05, "loss": 0.9432, "step": 984 }, { "epoch": 0.15, "grad_norm": 22.14695877945067, "learning_rate": 1.997633864292611e-05, "loss": 0.9516, "step": 985 }, { "epoch": 0.15, "grad_norm": 22.29064336987488, "learning_rate": 1.99762225597834e-05, "loss": 0.9803, "step": 986 }, { "epoch": 0.15, "grad_norm": 17.519886680918606, "learning_rate": 1.997610619292276e-05, "loss": 1.0261, "step": 987 }, { "epoch": 0.15, "grad_norm": 26.018074294943958, "learning_rate": 1.9975989542347495e-05, "loss": 1.0793, "step": 988 }, { "epoch": 0.15, "grad_norm": 19.608446275899777, "learning_rate": 1.9975872608060927e-05, "loss": 0.8461, "step": 989 }, { "epoch": 0.15, "grad_norm": 20.69574199675815, "learning_rate": 1.9975755390066373e-05, "loss": 1.0362, "step": 990 }, { "epoch": 0.15, "grad_norm": 18.82663372196337, "learning_rate": 1.9975637888367173e-05, "loss": 1.0071, "step": 991 }, { "epoch": 0.15, "grad_norm": 20.896183397888986, "learning_rate": 1.9975520102966667e-05, "loss": 1.0048, "step": 992 }, { "epoch": 0.16, "grad_norm": 19.794927310425322, "learning_rate": 1.9975402033868207e-05, "loss": 0.9567, "step": 993 }, { "epoch": 0.16, "grad_norm": 27.72274353370509, "learning_rate": 1.9975283681075142e-05, "loss": 0.9253, "step": 994 }, { "epoch": 0.16, "grad_norm": 25.81541402234388, "learning_rate": 1.9975165044590853e-05, "loss": 0.9425, "step": 995 }, { "epoch": 0.16, "grad_norm": 24.675916738713358, "learning_rate": 1.99750461244187e-05, "loss": 1.0206, "step": 996 }, { "epoch": 0.16, "grad_norm": 25.905428859595194, "learning_rate": 1.997492692056207e-05, "loss": 0.9438, "step": 997 }, { "epoch": 0.16, "grad_norm": 28.522451249444654, "learning_rate": 1.9974807433024357e-05, "loss": 0.9945, "step": 998 }, { "epoch": 0.16, "grad_norm": 19.718876927446388, "learning_rate": 1.9974687661808954e-05, "loss": 0.9148, "step": 999 }, { "epoch": 0.16, "grad_norm": 22.091939756214508, "learning_rate": 1.997456760691927e-05, "loss": 0.9874, "step": 1000 }, { "epoch": 0.16, "grad_norm": 23.11514146058491, "learning_rate": 1.9974447268358715e-05, "loss": 0.9015, "step": 1001 }, { "epoch": 0.16, "grad_norm": 17.804668716079288, "learning_rate": 1.9974326646130716e-05, "loss": 0.8472, "step": 1002 }, { "epoch": 0.16, "grad_norm": 16.656475030728036, "learning_rate": 1.9974205740238704e-05, "loss": 0.8391, "step": 1003 }, { "epoch": 0.16, "grad_norm": 21.558633175644697, "learning_rate": 1.9974084550686116e-05, "loss": 0.9494, "step": 1004 }, { "epoch": 0.16, "grad_norm": 27.818598239427306, "learning_rate": 1.9973963077476394e-05, "loss": 1.0851, "step": 1005 }, { "epoch": 0.16, "grad_norm": 14.655754652375885, "learning_rate": 1.9973841320613e-05, "loss": 0.9687, "step": 1006 }, { "epoch": 0.16, "grad_norm": 24.40657534730464, "learning_rate": 1.997371928009939e-05, "loss": 1.0059, "step": 1007 }, { "epoch": 0.16, "grad_norm": 18.40173965431018, "learning_rate": 1.9973596955939043e-05, "loss": 1.0217, "step": 1008 }, { "epoch": 0.16, "grad_norm": 24.04782350799938, "learning_rate": 1.997347434813543e-05, "loss": 0.9063, "step": 1009 }, { "epoch": 0.16, "grad_norm": 25.56124047007055, "learning_rate": 1.997335145669204e-05, "loss": 0.8219, "step": 1010 }, { "epoch": 0.16, "grad_norm": 19.831400271602874, "learning_rate": 1.997322828161237e-05, "loss": 1.0246, "step": 1011 }, { "epoch": 0.16, "grad_norm": 17.31427635365123, "learning_rate": 1.9973104822899923e-05, "loss": 0.9491, "step": 1012 }, { "epoch": 0.16, "grad_norm": 19.734833288421257, "learning_rate": 1.9972981080558208e-05, "loss": 1.1044, "step": 1013 }, { "epoch": 0.16, "grad_norm": 59.49551648289499, "learning_rate": 1.9972857054590744e-05, "loss": 1.0689, "step": 1014 }, { "epoch": 0.16, "grad_norm": 14.384798855558653, "learning_rate": 1.997273274500106e-05, "loss": 0.8931, "step": 1015 }, { "epoch": 0.16, "grad_norm": 25.769802069531156, "learning_rate": 1.997260815179269e-05, "loss": 1.1256, "step": 1016 }, { "epoch": 0.16, "grad_norm": 22.464246187071343, "learning_rate": 1.997248327496918e-05, "loss": 1.0696, "step": 1017 }, { "epoch": 0.16, "grad_norm": 25.956797254439103, "learning_rate": 1.9972358114534074e-05, "loss": 0.9252, "step": 1018 }, { "epoch": 0.16, "grad_norm": 25.760436841003276, "learning_rate": 1.9972232670490945e-05, "loss": 0.8813, "step": 1019 }, { "epoch": 0.16, "grad_norm": 18.583151045220244, "learning_rate": 1.997210694284335e-05, "loss": 0.8999, "step": 1020 }, { "epoch": 0.16, "grad_norm": 24.44323274196326, "learning_rate": 1.9971980931594865e-05, "loss": 0.9655, "step": 1021 }, { "epoch": 0.16, "grad_norm": 16.494483652381803, "learning_rate": 1.9971854636749075e-05, "loss": 0.9329, "step": 1022 }, { "epoch": 0.16, "grad_norm": 21.267103554759874, "learning_rate": 1.9971728058309572e-05, "loss": 0.987, "step": 1023 }, { "epoch": 0.16, "grad_norm": 22.118992572482618, "learning_rate": 1.9971601196279963e-05, "loss": 0.9983, "step": 1024 }, { "epoch": 0.16, "grad_norm": 19.842380598831724, "learning_rate": 1.997147405066384e-05, "loss": 0.9695, "step": 1025 }, { "epoch": 0.16, "grad_norm": 14.65298669391427, "learning_rate": 1.9971346621464833e-05, "loss": 0.9535, "step": 1026 }, { "epoch": 0.16, "grad_norm": 27.512256742170667, "learning_rate": 1.9971218908686563e-05, "loss": 0.9578, "step": 1027 }, { "epoch": 0.16, "grad_norm": 24.574149753059544, "learning_rate": 1.9971090912332657e-05, "loss": 0.9132, "step": 1028 }, { "epoch": 0.16, "grad_norm": 29.33446986848635, "learning_rate": 1.9970962632406757e-05, "loss": 1.0906, "step": 1029 }, { "epoch": 0.16, "grad_norm": 20.516292745642946, "learning_rate": 1.9970834068912515e-05, "loss": 0.9737, "step": 1030 }, { "epoch": 0.16, "grad_norm": 20.62443621688516, "learning_rate": 1.9970705221853584e-05, "loss": 0.9541, "step": 1031 }, { "epoch": 0.16, "grad_norm": 22.22644135185706, "learning_rate": 1.9970576091233633e-05, "loss": 0.9547, "step": 1032 }, { "epoch": 0.16, "grad_norm": 18.726208484949435, "learning_rate": 1.9970446677056325e-05, "loss": 0.8472, "step": 1033 }, { "epoch": 0.16, "grad_norm": 16.621540227646772, "learning_rate": 1.997031697932535e-05, "loss": 0.9316, "step": 1034 }, { "epoch": 0.16, "grad_norm": 21.326850876810315, "learning_rate": 1.9970186998044392e-05, "loss": 0.8996, "step": 1035 }, { "epoch": 0.16, "grad_norm": 18.65304831806845, "learning_rate": 1.9970056733217147e-05, "loss": 0.9552, "step": 1036 }, { "epoch": 0.16, "grad_norm": 18.16398263318951, "learning_rate": 1.996992618484732e-05, "loss": 0.8852, "step": 1037 }, { "epoch": 0.16, "grad_norm": 16.376545159769574, "learning_rate": 1.9969795352938625e-05, "loss": 0.9859, "step": 1038 }, { "epoch": 0.16, "grad_norm": 22.187092733003354, "learning_rate": 1.9969664237494785e-05, "loss": 1.0501, "step": 1039 }, { "epoch": 0.16, "grad_norm": 17.335407337507426, "learning_rate": 1.9969532838519524e-05, "loss": 0.865, "step": 1040 }, { "epoch": 0.16, "grad_norm": 21.160217237182188, "learning_rate": 1.996940115601658e-05, "loss": 1.1142, "step": 1041 }, { "epoch": 0.16, "grad_norm": 18.337951351272025, "learning_rate": 1.99692691899897e-05, "loss": 1.0136, "step": 1042 }, { "epoch": 0.16, "grad_norm": 27.48985530503833, "learning_rate": 1.9969136940442636e-05, "loss": 0.9256, "step": 1043 }, { "epoch": 0.16, "grad_norm": 4.359953718842061, "learning_rate": 1.996900440737915e-05, "loss": 0.9117, "step": 1044 }, { "epoch": 0.16, "grad_norm": 16.059047708332876, "learning_rate": 1.996887159080301e-05, "loss": 0.9515, "step": 1045 }, { "epoch": 0.16, "grad_norm": 14.98626266797299, "learning_rate": 1.9968738490717992e-05, "loss": 0.9373, "step": 1046 }, { "epoch": 0.16, "grad_norm": 23.413081707680554, "learning_rate": 1.9968605107127886e-05, "loss": 0.9226, "step": 1047 }, { "epoch": 0.16, "grad_norm": 26.178540115576087, "learning_rate": 1.996847144003648e-05, "loss": 0.897, "step": 1048 }, { "epoch": 0.16, "grad_norm": 19.96063426285729, "learning_rate": 1.9968337489447578e-05, "loss": 0.9423, "step": 1049 }, { "epoch": 0.16, "grad_norm": 22.224057065412012, "learning_rate": 1.996820325536499e-05, "loss": 0.8617, "step": 1050 }, { "epoch": 0.16, "grad_norm": 15.948927645674216, "learning_rate": 1.996806873779253e-05, "loss": 0.9121, "step": 1051 }, { "epoch": 0.16, "grad_norm": 26.049640858786617, "learning_rate": 1.9967933936734027e-05, "loss": 0.9288, "step": 1052 }, { "epoch": 0.16, "grad_norm": 21.044905889730735, "learning_rate": 1.996779885219332e-05, "loss": 0.8825, "step": 1053 }, { "epoch": 0.16, "grad_norm": 30.567094353191365, "learning_rate": 1.9967663484174235e-05, "loss": 0.8961, "step": 1054 }, { "epoch": 0.16, "grad_norm": 19.90970061560877, "learning_rate": 1.9967527832680637e-05, "loss": 0.9373, "step": 1055 }, { "epoch": 0.16, "grad_norm": 26.129488848548302, "learning_rate": 1.996739189771638e-05, "loss": 0.9245, "step": 1056 }, { "epoch": 0.17, "grad_norm": 25.387900679932635, "learning_rate": 1.9967255679285324e-05, "loss": 0.9912, "step": 1057 }, { "epoch": 0.17, "grad_norm": 7.449547174792245, "learning_rate": 1.9967119177391348e-05, "loss": 0.9457, "step": 1058 }, { "epoch": 0.17, "grad_norm": 20.37084016139191, "learning_rate": 1.9966982392038335e-05, "loss": 1.0184, "step": 1059 }, { "epoch": 0.17, "grad_norm": 77.16469358476006, "learning_rate": 1.996684532323017e-05, "loss": 0.9994, "step": 1060 }, { "epoch": 0.17, "grad_norm": 16.957716178638062, "learning_rate": 1.9966707970970756e-05, "loss": 0.9131, "step": 1061 }, { "epoch": 0.17, "grad_norm": 21.762517296446227, "learning_rate": 1.9966570335264e-05, "loss": 0.8824, "step": 1062 }, { "epoch": 0.17, "grad_norm": 20.06801353583546, "learning_rate": 1.9966432416113812e-05, "loss": 0.8856, "step": 1063 }, { "epoch": 0.17, "grad_norm": 19.171373317220162, "learning_rate": 1.9966294213524117e-05, "loss": 0.8615, "step": 1064 }, { "epoch": 0.17, "grad_norm": 21.024850750346925, "learning_rate": 1.9966155727498843e-05, "loss": 0.9013, "step": 1065 }, { "epoch": 0.17, "grad_norm": 22.164824136200476, "learning_rate": 1.9966016958041933e-05, "loss": 1.0358, "step": 1066 }, { "epoch": 0.17, "grad_norm": 41.681919746720965, "learning_rate": 1.996587790515733e-05, "loss": 0.8274, "step": 1067 }, { "epoch": 0.17, "grad_norm": 15.945059858878466, "learning_rate": 1.9965738568848986e-05, "loss": 0.9953, "step": 1068 }, { "epoch": 0.17, "grad_norm": 22.953576014056228, "learning_rate": 1.996559894912087e-05, "loss": 1.0221, "step": 1069 }, { "epoch": 0.17, "grad_norm": 17.972018525935646, "learning_rate": 1.996545904597695e-05, "loss": 0.9417, "step": 1070 }, { "epoch": 0.17, "grad_norm": 15.388072707490695, "learning_rate": 1.9965318859421203e-05, "loss": 0.9903, "step": 1071 }, { "epoch": 0.17, "grad_norm": 23.47781751645356, "learning_rate": 1.9965178389457617e-05, "loss": 0.8525, "step": 1072 }, { "epoch": 0.17, "grad_norm": 20.860240417558686, "learning_rate": 1.9965037636090187e-05, "loss": 0.9391, "step": 1073 }, { "epoch": 0.17, "grad_norm": 26.294374092027503, "learning_rate": 1.9964896599322917e-05, "loss": 0.9244, "step": 1074 }, { "epoch": 0.17, "grad_norm": 19.660765057313604, "learning_rate": 1.9964755279159816e-05, "loss": 0.8199, "step": 1075 }, { "epoch": 0.17, "grad_norm": 31.299620579344733, "learning_rate": 1.9964613675604904e-05, "loss": 0.9738, "step": 1076 }, { "epoch": 0.17, "grad_norm": 22.760483900897643, "learning_rate": 1.996447178866221e-05, "loss": 0.9592, "step": 1077 }, { "epoch": 0.17, "grad_norm": 14.650392983694106, "learning_rate": 1.9964329618335766e-05, "loss": 0.9399, "step": 1078 }, { "epoch": 0.17, "grad_norm": 28.959922380667738, "learning_rate": 1.9964187164629617e-05, "loss": 0.976, "step": 1079 }, { "epoch": 0.17, "grad_norm": 25.05245508419107, "learning_rate": 1.996404442754781e-05, "loss": 0.9548, "step": 1080 }, { "epoch": 0.17, "grad_norm": 26.07033700544596, "learning_rate": 1.9963901407094415e-05, "loss": 0.9631, "step": 1081 }, { "epoch": 0.17, "grad_norm": 23.19212974742126, "learning_rate": 1.996375810327349e-05, "loss": 1.0318, "step": 1082 }, { "epoch": 0.17, "grad_norm": 14.661445227363606, "learning_rate": 1.9963614516089112e-05, "loss": 0.9228, "step": 1083 }, { "epoch": 0.17, "grad_norm": 21.85009642256209, "learning_rate": 1.9963470645545365e-05, "loss": 1.0432, "step": 1084 }, { "epoch": 0.17, "grad_norm": 20.32518228904275, "learning_rate": 1.9963326491646346e-05, "loss": 0.9614, "step": 1085 }, { "epoch": 0.17, "grad_norm": 20.71690331977722, "learning_rate": 1.9963182054396144e-05, "loss": 0.8802, "step": 1086 }, { "epoch": 0.17, "grad_norm": 30.1355175023656, "learning_rate": 1.9963037333798876e-05, "loss": 1.0874, "step": 1087 }, { "epoch": 0.17, "grad_norm": 15.292365728501006, "learning_rate": 1.9962892329858655e-05, "loss": 0.876, "step": 1088 }, { "epoch": 0.17, "grad_norm": 21.214769426839712, "learning_rate": 1.99627470425796e-05, "loss": 0.8733, "step": 1089 }, { "epoch": 0.17, "grad_norm": 17.379554352322735, "learning_rate": 1.9962601471965854e-05, "loss": 0.9827, "step": 1090 }, { "epoch": 0.17, "grad_norm": 29.110130115979604, "learning_rate": 1.9962455618021545e-05, "loss": 0.8946, "step": 1091 }, { "epoch": 0.17, "grad_norm": 19.795895266959587, "learning_rate": 1.9962309480750827e-05, "loss": 1.029, "step": 1092 }, { "epoch": 0.17, "grad_norm": 21.176858967088037, "learning_rate": 1.9962163060157854e-05, "loss": 1.0959, "step": 1093 }, { "epoch": 0.17, "grad_norm": 36.20100217748054, "learning_rate": 1.9962016356246794e-05, "loss": 0.9817, "step": 1094 }, { "epoch": 0.17, "grad_norm": 17.19508124130565, "learning_rate": 1.9961869369021814e-05, "loss": 0.8772, "step": 1095 }, { "epoch": 0.17, "grad_norm": 23.21362086243986, "learning_rate": 1.9961722098487098e-05, "loss": 0.9472, "step": 1096 }, { "epoch": 0.17, "grad_norm": 19.058492890904425, "learning_rate": 1.9961574544646835e-05, "loss": 0.8941, "step": 1097 }, { "epoch": 0.17, "grad_norm": 195.05581634123945, "learning_rate": 1.9961426707505217e-05, "loss": 0.8981, "step": 1098 }, { "epoch": 0.17, "grad_norm": 39.32973305071126, "learning_rate": 1.996127858706645e-05, "loss": 1.0174, "step": 1099 }, { "epoch": 0.17, "grad_norm": 24.445888327195416, "learning_rate": 1.996113018333475e-05, "loss": 0.9646, "step": 1100 }, { "epoch": 0.17, "grad_norm": 23.171377947074845, "learning_rate": 1.996098149631433e-05, "loss": 0.9686, "step": 1101 }, { "epoch": 0.17, "grad_norm": 24.874052487351488, "learning_rate": 1.9960832526009427e-05, "loss": 1.0018, "step": 1102 }, { "epoch": 0.17, "grad_norm": 20.598162584973974, "learning_rate": 1.996068327242427e-05, "loss": 0.9975, "step": 1103 }, { "epoch": 0.17, "grad_norm": 21.004610504949117, "learning_rate": 1.9960533735563113e-05, "loss": 0.8323, "step": 1104 }, { "epoch": 0.17, "grad_norm": 21.582257565653794, "learning_rate": 1.99603839154302e-05, "loss": 0.7476, "step": 1105 }, { "epoch": 0.17, "grad_norm": 19.541556438261914, "learning_rate": 1.9960233812029798e-05, "loss": 0.8683, "step": 1106 }, { "epoch": 0.17, "grad_norm": 38.115180145076536, "learning_rate": 1.996008342536617e-05, "loss": 0.9944, "step": 1107 }, { "epoch": 0.17, "grad_norm": 5.129276843311849, "learning_rate": 1.9959932755443596e-05, "loss": 0.8668, "step": 1108 }, { "epoch": 0.17, "grad_norm": 16.878704671929537, "learning_rate": 1.995978180226636e-05, "loss": 0.9611, "step": 1109 }, { "epoch": 0.17, "grad_norm": 30.11822019357628, "learning_rate": 1.9959630565838758e-05, "loss": 1.0163, "step": 1110 }, { "epoch": 0.17, "grad_norm": 21.82556571920938, "learning_rate": 1.9959479046165088e-05, "loss": 1.0279, "step": 1111 }, { "epoch": 0.17, "grad_norm": 15.952955102622985, "learning_rate": 1.9959327243249662e-05, "loss": 0.8513, "step": 1112 }, { "epoch": 0.17, "grad_norm": 17.68419848419876, "learning_rate": 1.9959175157096794e-05, "loss": 0.9806, "step": 1113 }, { "epoch": 0.17, "grad_norm": 24.483035655291626, "learning_rate": 1.995902278771081e-05, "loss": 0.9821, "step": 1114 }, { "epoch": 0.17, "grad_norm": 15.279112038316622, "learning_rate": 1.9958870135096044e-05, "loss": 0.9228, "step": 1115 }, { "epoch": 0.17, "grad_norm": 23.95422021734671, "learning_rate": 1.9958717199256833e-05, "loss": 0.9523, "step": 1116 }, { "epoch": 0.17, "grad_norm": 13.930983981026705, "learning_rate": 1.9958563980197536e-05, "loss": 0.9921, "step": 1117 }, { "epoch": 0.17, "grad_norm": 23.897593610216855, "learning_rate": 1.9958410477922504e-05, "loss": 0.8384, "step": 1118 }, { "epoch": 0.17, "grad_norm": 25.011398549303884, "learning_rate": 1.9958256692436103e-05, "loss": 0.9378, "step": 1119 }, { "epoch": 0.17, "grad_norm": 30.645943474306943, "learning_rate": 1.995810262374271e-05, "loss": 0.9036, "step": 1120 }, { "epoch": 0.18, "grad_norm": 14.178524056826843, "learning_rate": 1.99579482718467e-05, "loss": 0.7769, "step": 1121 }, { "epoch": 0.18, "grad_norm": 23.82688153633873, "learning_rate": 1.9957793636752466e-05, "loss": 0.8656, "step": 1122 }, { "epoch": 0.18, "grad_norm": 28.657159114881118, "learning_rate": 1.9957638718464405e-05, "loss": 1.0352, "step": 1123 }, { "epoch": 0.18, "grad_norm": 25.372462772966394, "learning_rate": 1.995748351698693e-05, "loss": 1.0062, "step": 1124 }, { "epoch": 0.18, "grad_norm": 31.516940010486838, "learning_rate": 1.995732803232444e-05, "loss": 0.9846, "step": 1125 }, { "epoch": 0.18, "grad_norm": 21.428904983246213, "learning_rate": 1.995717226448137e-05, "loss": 1.0084, "step": 1126 }, { "epoch": 0.18, "grad_norm": 26.96075121589446, "learning_rate": 1.995701621346215e-05, "loss": 0.9589, "step": 1127 }, { "epoch": 0.18, "grad_norm": 21.407298769632664, "learning_rate": 1.9956859879271207e-05, "loss": 0.891, "step": 1128 }, { "epoch": 0.18, "grad_norm": 23.547919443225897, "learning_rate": 1.9956703261913e-05, "loss": 0.8514, "step": 1129 }, { "epoch": 0.18, "grad_norm": 19.761614780186456, "learning_rate": 1.995654636139197e-05, "loss": 0.9503, "step": 1130 }, { "epoch": 0.18, "grad_norm": 20.186137096558895, "learning_rate": 1.9956389177712592e-05, "loss": 1.0059, "step": 1131 }, { "epoch": 0.18, "grad_norm": 20.500806487315653, "learning_rate": 1.9956231710879325e-05, "loss": 1.0018, "step": 1132 }, { "epoch": 0.18, "grad_norm": 25.397915764093298, "learning_rate": 1.9956073960896652e-05, "loss": 0.8797, "step": 1133 }, { "epoch": 0.18, "grad_norm": 25.37684573662613, "learning_rate": 1.9955915927769066e-05, "loss": 0.9778, "step": 1134 }, { "epoch": 0.18, "grad_norm": 20.25531568966331, "learning_rate": 1.995575761150105e-05, "loss": 1.0669, "step": 1135 }, { "epoch": 0.18, "grad_norm": 13.679965697748836, "learning_rate": 1.995559901209711e-05, "loss": 0.896, "step": 1136 }, { "epoch": 0.18, "grad_norm": 23.073898200835217, "learning_rate": 1.995544012956176e-05, "loss": 0.9498, "step": 1137 }, { "epoch": 0.18, "grad_norm": 25.17758573778675, "learning_rate": 1.995528096389952e-05, "loss": 0.9395, "step": 1138 }, { "epoch": 0.18, "grad_norm": 19.8285652424402, "learning_rate": 1.995512151511491e-05, "loss": 0.9703, "step": 1139 }, { "epoch": 0.18, "grad_norm": 18.66749442919373, "learning_rate": 1.995496178321247e-05, "loss": 1.0551, "step": 1140 }, { "epoch": 0.18, "grad_norm": 23.76111052344706, "learning_rate": 1.9954801768196738e-05, "loss": 0.8594, "step": 1141 }, { "epoch": 0.18, "grad_norm": 16.19758400764559, "learning_rate": 1.9954641470072267e-05, "loss": 0.8637, "step": 1142 }, { "epoch": 0.18, "grad_norm": 17.223655637944525, "learning_rate": 1.9954480888843614e-05, "loss": 1.0662, "step": 1143 }, { "epoch": 0.18, "grad_norm": 18.395337053731495, "learning_rate": 1.995432002451535e-05, "loss": 0.9162, "step": 1144 }, { "epoch": 0.18, "grad_norm": 27.209186781300918, "learning_rate": 1.995415887709205e-05, "loss": 0.8881, "step": 1145 }, { "epoch": 0.18, "grad_norm": 24.438681956827622, "learning_rate": 1.995399744657829e-05, "loss": 0.9011, "step": 1146 }, { "epoch": 0.18, "grad_norm": 17.813041791707164, "learning_rate": 1.9953835732978668e-05, "loss": 0.8796, "step": 1147 }, { "epoch": 0.18, "grad_norm": 24.719100036988042, "learning_rate": 1.995367373629778e-05, "loss": 0.9395, "step": 1148 }, { "epoch": 0.18, "grad_norm": 18.08496977604367, "learning_rate": 1.9953511456540233e-05, "loss": 0.9645, "step": 1149 }, { "epoch": 0.18, "grad_norm": 20.237033522061687, "learning_rate": 1.9953348893710645e-05, "loss": 0.91, "step": 1150 }, { "epoch": 0.18, "grad_norm": 13.488883503013803, "learning_rate": 1.9953186047813637e-05, "loss": 0.8988, "step": 1151 }, { "epoch": 0.18, "grad_norm": 36.55713919113992, "learning_rate": 1.9953022918853837e-05, "loss": 0.9923, "step": 1152 }, { "epoch": 0.18, "grad_norm": 30.13020995109024, "learning_rate": 1.9952859506835887e-05, "loss": 1.0171, "step": 1153 }, { "epoch": 0.18, "grad_norm": 19.313885660675833, "learning_rate": 1.995269581176444e-05, "loss": 0.925, "step": 1154 }, { "epoch": 0.18, "grad_norm": 25.380561423525133, "learning_rate": 1.9952531833644145e-05, "loss": 0.9817, "step": 1155 }, { "epoch": 0.18, "grad_norm": 33.27406470233171, "learning_rate": 1.9952367572479664e-05, "loss": 1.102, "step": 1156 }, { "epoch": 0.18, "grad_norm": 27.73252246633459, "learning_rate": 1.995220302827567e-05, "loss": 0.9488, "step": 1157 }, { "epoch": 0.18, "grad_norm": 34.5026125177882, "learning_rate": 1.9952038201036848e-05, "loss": 1.0748, "step": 1158 }, { "epoch": 0.18, "grad_norm": 28.870464497190003, "learning_rate": 1.995187309076788e-05, "loss": 0.8552, "step": 1159 }, { "epoch": 0.18, "grad_norm": 14.031554636165179, "learning_rate": 1.9951707697473465e-05, "loss": 0.8569, "step": 1160 }, { "epoch": 0.18, "grad_norm": 26.63769626711829, "learning_rate": 1.99515420211583e-05, "loss": 1.0523, "step": 1161 }, { "epoch": 0.18, "grad_norm": 21.890918687515374, "learning_rate": 1.9951376061827104e-05, "loss": 0.9819, "step": 1162 }, { "epoch": 0.18, "grad_norm": 22.280379834779016, "learning_rate": 1.995120981948459e-05, "loss": 0.966, "step": 1163 }, { "epoch": 0.18, "grad_norm": 17.07628401442733, "learning_rate": 1.9951043294135498e-05, "loss": 0.8894, "step": 1164 }, { "epoch": 0.18, "grad_norm": 5.540979091455273, "learning_rate": 1.995087648578455e-05, "loss": 0.8823, "step": 1165 }, { "epoch": 0.18, "grad_norm": 19.13708005658681, "learning_rate": 1.9950709394436497e-05, "loss": 1.0229, "step": 1166 }, { "epoch": 0.18, "grad_norm": 21.64973108975881, "learning_rate": 1.9950542020096094e-05, "loss": 0.9317, "step": 1167 }, { "epoch": 0.18, "grad_norm": 21.215047049702857, "learning_rate": 1.995037436276809e-05, "loss": 0.8126, "step": 1168 }, { "epoch": 0.18, "grad_norm": 23.32932582071813, "learning_rate": 1.9950206422457263e-05, "loss": 0.9556, "step": 1169 }, { "epoch": 0.18, "grad_norm": 18.990490524916908, "learning_rate": 1.9950038199168387e-05, "loss": 1.0087, "step": 1170 }, { "epoch": 0.18, "grad_norm": 26.641587413955936, "learning_rate": 1.9949869692906245e-05, "loss": 0.9149, "step": 1171 }, { "epoch": 0.18, "grad_norm": 26.106088595271515, "learning_rate": 1.994970090367563e-05, "loss": 0.9971, "step": 1172 }, { "epoch": 0.18, "grad_norm": 23.41581929549556, "learning_rate": 1.994953183148134e-05, "loss": 0.856, "step": 1173 }, { "epoch": 0.18, "grad_norm": 47.672523506627265, "learning_rate": 1.9949362476328184e-05, "loss": 1.0399, "step": 1174 }, { "epoch": 0.18, "grad_norm": 13.695419951962345, "learning_rate": 1.9949192838220982e-05, "loss": 0.8773, "step": 1175 }, { "epoch": 0.18, "grad_norm": 23.926550004091787, "learning_rate": 1.9949022917164556e-05, "loss": 0.8676, "step": 1176 }, { "epoch": 0.18, "grad_norm": 30.105655016483126, "learning_rate": 1.9948852713163737e-05, "loss": 0.9021, "step": 1177 }, { "epoch": 0.18, "grad_norm": 18.316633414260945, "learning_rate": 1.9948682226223366e-05, "loss": 0.9636, "step": 1178 }, { "epoch": 0.18, "grad_norm": 23.736809104705028, "learning_rate": 1.9948511456348295e-05, "loss": 0.9362, "step": 1179 }, { "epoch": 0.18, "grad_norm": 23.08183740911144, "learning_rate": 1.9948340403543375e-05, "loss": 1.0331, "step": 1180 }, { "epoch": 0.18, "grad_norm": 21.57154819026003, "learning_rate": 1.9948169067813473e-05, "loss": 0.983, "step": 1181 }, { "epoch": 0.18, "grad_norm": 18.184887788579513, "learning_rate": 1.9947997449163467e-05, "loss": 1.0024, "step": 1182 }, { "epoch": 0.18, "grad_norm": 21.90526965392405, "learning_rate": 1.9947825547598226e-05, "loss": 0.9809, "step": 1183 }, { "epoch": 0.18, "grad_norm": 29.46472945046384, "learning_rate": 1.994765336312265e-05, "loss": 0.8987, "step": 1184 }, { "epoch": 0.19, "grad_norm": 22.38233590675501, "learning_rate": 1.994748089574163e-05, "loss": 0.9454, "step": 1185 }, { "epoch": 0.19, "grad_norm": 23.310221537703704, "learning_rate": 1.9947308145460073e-05, "loss": 0.9608, "step": 1186 }, { "epoch": 0.19, "grad_norm": 24.30178722481752, "learning_rate": 1.9947135112282894e-05, "loss": 0.9571, "step": 1187 }, { "epoch": 0.19, "grad_norm": 20.044473009121155, "learning_rate": 1.9946961796215006e-05, "loss": 0.8899, "step": 1188 }, { "epoch": 0.19, "grad_norm": 14.507704191568795, "learning_rate": 1.994678819726135e-05, "loss": 0.8034, "step": 1189 }, { "epoch": 0.19, "grad_norm": 17.730711127433754, "learning_rate": 1.994661431542685e-05, "loss": 0.9218, "step": 1190 }, { "epoch": 0.19, "grad_norm": 21.246142138176882, "learning_rate": 1.994644015071646e-05, "loss": 0.8613, "step": 1191 }, { "epoch": 0.19, "grad_norm": 22.45738185901497, "learning_rate": 1.994626570313513e-05, "loss": 0.9275, "step": 1192 }, { "epoch": 0.19, "grad_norm": 23.320499622610207, "learning_rate": 1.9946090972687827e-05, "loss": 1.0286, "step": 1193 }, { "epoch": 0.19, "grad_norm": 23.75997874661781, "learning_rate": 1.994591595937951e-05, "loss": 0.7884, "step": 1194 }, { "epoch": 0.19, "grad_norm": 16.398642040283757, "learning_rate": 1.994574066321516e-05, "loss": 0.9726, "step": 1195 }, { "epoch": 0.19, "grad_norm": 20.392093268692573, "learning_rate": 1.9945565084199766e-05, "loss": 0.9629, "step": 1196 }, { "epoch": 0.19, "grad_norm": 13.446296983902577, "learning_rate": 1.994538922233832e-05, "loss": 0.9321, "step": 1197 }, { "epoch": 0.19, "grad_norm": 21.243092139911507, "learning_rate": 1.994521307763582e-05, "loss": 0.8464, "step": 1198 }, { "epoch": 0.19, "grad_norm": 32.68458859395483, "learning_rate": 1.994503665009728e-05, "loss": 0.953, "step": 1199 }, { "epoch": 0.19, "grad_norm": 25.726231668744806, "learning_rate": 1.9944859939727714e-05, "loss": 0.9653, "step": 1200 }, { "epoch": 0.19, "grad_norm": 16.352486735112418, "learning_rate": 1.994468294653215e-05, "loss": 0.7899, "step": 1201 }, { "epoch": 0.19, "grad_norm": 16.72271390253157, "learning_rate": 1.994450567051562e-05, "loss": 0.9373, "step": 1202 }, { "epoch": 0.19, "grad_norm": 19.202867844926203, "learning_rate": 1.9944328111683166e-05, "loss": 0.9778, "step": 1203 }, { "epoch": 0.19, "grad_norm": 14.555628186642041, "learning_rate": 1.9944150270039837e-05, "loss": 0.9264, "step": 1204 }, { "epoch": 0.19, "grad_norm": 31.11643087054771, "learning_rate": 1.994397214559069e-05, "loss": 0.8862, "step": 1205 }, { "epoch": 0.19, "grad_norm": 16.276235170561964, "learning_rate": 1.9943793738340796e-05, "loss": 0.9889, "step": 1206 }, { "epoch": 0.19, "grad_norm": 21.994496461120395, "learning_rate": 1.9943615048295224e-05, "loss": 1.0607, "step": 1207 }, { "epoch": 0.19, "grad_norm": 31.373176717403403, "learning_rate": 1.9943436075459056e-05, "loss": 0.9986, "step": 1208 }, { "epoch": 0.19, "grad_norm": 21.367489900837658, "learning_rate": 1.994325681983738e-05, "loss": 0.8561, "step": 1209 }, { "epoch": 0.19, "grad_norm": 20.57649557531624, "learning_rate": 1.9943077281435302e-05, "loss": 0.9088, "step": 1210 }, { "epoch": 0.19, "grad_norm": 19.22798190056659, "learning_rate": 1.994289746025792e-05, "loss": 0.9386, "step": 1211 }, { "epoch": 0.19, "grad_norm": 20.051370076626796, "learning_rate": 1.994271735631035e-05, "loss": 0.9943, "step": 1212 }, { "epoch": 0.19, "grad_norm": 24.49291742688194, "learning_rate": 1.9942536969597713e-05, "loss": 0.9867, "step": 1213 }, { "epoch": 0.19, "grad_norm": 23.380714742365978, "learning_rate": 1.9942356300125146e-05, "loss": 1.01, "step": 1214 }, { "epoch": 0.19, "grad_norm": 17.769314347622814, "learning_rate": 1.994217534789778e-05, "loss": 1.009, "step": 1215 }, { "epoch": 0.19, "grad_norm": 25.32979480338763, "learning_rate": 1.994199411292076e-05, "loss": 0.97, "step": 1216 }, { "epoch": 0.19, "grad_norm": 23.86239682642862, "learning_rate": 1.9941812595199246e-05, "loss": 0.9607, "step": 1217 }, { "epoch": 0.19, "grad_norm": 24.97029570902465, "learning_rate": 1.99416307947384e-05, "loss": 0.8529, "step": 1218 }, { "epoch": 0.19, "grad_norm": 22.543302660121476, "learning_rate": 1.9941448711543386e-05, "loss": 0.9853, "step": 1219 }, { "epoch": 0.19, "grad_norm": 16.608061158043856, "learning_rate": 1.994126634561939e-05, "loss": 0.9182, "step": 1220 }, { "epoch": 0.19, "grad_norm": 22.731217702565235, "learning_rate": 1.994108369697159e-05, "loss": 0.9641, "step": 1221 }, { "epoch": 0.19, "grad_norm": 33.05863261724831, "learning_rate": 1.9940900765605187e-05, "loss": 1.0102, "step": 1222 }, { "epoch": 0.19, "grad_norm": 39.6450780495734, "learning_rate": 1.994071755152538e-05, "loss": 1.0147, "step": 1223 }, { "epoch": 0.19, "grad_norm": 7.845016571956475, "learning_rate": 1.9940534054737383e-05, "loss": 1.0116, "step": 1224 }, { "epoch": 0.19, "grad_norm": 23.703273159160023, "learning_rate": 1.9940350275246412e-05, "loss": 0.8897, "step": 1225 }, { "epoch": 0.19, "grad_norm": 20.373733486995256, "learning_rate": 1.9940166213057693e-05, "loss": 0.9425, "step": 1226 }, { "epoch": 0.19, "grad_norm": 22.455188548883385, "learning_rate": 1.9939981868176462e-05, "loss": 1.0323, "step": 1227 }, { "epoch": 0.19, "grad_norm": 28.822589758446107, "learning_rate": 1.9939797240607966e-05, "loss": 0.9633, "step": 1228 }, { "epoch": 0.19, "grad_norm": 16.82033621737185, "learning_rate": 1.9939612330357446e-05, "loss": 0.9004, "step": 1229 }, { "epoch": 0.19, "grad_norm": 16.329713536063725, "learning_rate": 1.993942713743017e-05, "loss": 0.9775, "step": 1230 }, { "epoch": 0.19, "grad_norm": 19.118670545951, "learning_rate": 1.9939241661831398e-05, "loss": 0.8798, "step": 1231 }, { "epoch": 0.19, "grad_norm": 27.92895927653587, "learning_rate": 1.9939055903566403e-05, "loss": 0.826, "step": 1232 }, { "epoch": 0.19, "grad_norm": 24.968752351958887, "learning_rate": 1.9938869862640476e-05, "loss": 0.9785, "step": 1233 }, { "epoch": 0.19, "grad_norm": 20.587930629160464, "learning_rate": 1.9938683539058906e-05, "loss": 0.9025, "step": 1234 }, { "epoch": 0.19, "grad_norm": 21.091215742957534, "learning_rate": 1.9938496932826988e-05, "loss": 0.8837, "step": 1235 }, { "epoch": 0.19, "grad_norm": 22.754134579278293, "learning_rate": 1.993831004395003e-05, "loss": 1.0415, "step": 1236 }, { "epoch": 0.19, "grad_norm": 11.773719017875198, "learning_rate": 1.9938122872433353e-05, "loss": 0.7607, "step": 1237 }, { "epoch": 0.19, "grad_norm": 7.629002864946615, "learning_rate": 1.9937935418282274e-05, "loss": 0.8634, "step": 1238 }, { "epoch": 0.19, "grad_norm": 18.14983628556474, "learning_rate": 1.9937747681502122e-05, "loss": 0.878, "step": 1239 }, { "epoch": 0.19, "grad_norm": 28.180830389525347, "learning_rate": 1.993755966209824e-05, "loss": 1.002, "step": 1240 }, { "epoch": 0.19, "grad_norm": 16.552035578828285, "learning_rate": 1.9937371360075977e-05, "loss": 0.9009, "step": 1241 }, { "epoch": 0.19, "grad_norm": 22.728857726390583, "learning_rate": 1.9937182775440683e-05, "loss": 1.0715, "step": 1242 }, { "epoch": 0.19, "grad_norm": 21.58531103284727, "learning_rate": 1.9936993908197725e-05, "loss": 0.9315, "step": 1243 }, { "epoch": 0.19, "grad_norm": 37.56450138694873, "learning_rate": 1.9936804758352473e-05, "loss": 1.0043, "step": 1244 }, { "epoch": 0.19, "grad_norm": 18.127500708486227, "learning_rate": 1.9936615325910308e-05, "loss": 0.8986, "step": 1245 }, { "epoch": 0.19, "grad_norm": 23.70305438398933, "learning_rate": 1.9936425610876614e-05, "loss": 0.9206, "step": 1246 }, { "epoch": 0.19, "grad_norm": 23.75094908712745, "learning_rate": 1.993623561325679e-05, "loss": 0.9542, "step": 1247 }, { "epoch": 0.19, "grad_norm": 24.105175353425047, "learning_rate": 1.9936045333056235e-05, "loss": 0.854, "step": 1248 }, { "epoch": 0.2, "grad_norm": 27.642600995924038, "learning_rate": 1.9935854770280362e-05, "loss": 0.8935, "step": 1249 }, { "epoch": 0.2, "grad_norm": 14.381684001037932, "learning_rate": 1.9935663924934596e-05, "loss": 0.8901, "step": 1250 }, { "epoch": 0.2, "grad_norm": 12.971888514001433, "learning_rate": 1.993547279702436e-05, "loss": 0.8007, "step": 1251 }, { "epoch": 0.2, "grad_norm": 16.821220718931485, "learning_rate": 1.9935281386555087e-05, "loss": 0.9687, "step": 1252 }, { "epoch": 0.2, "grad_norm": 20.53065080363474, "learning_rate": 1.9935089693532225e-05, "loss": 0.9721, "step": 1253 }, { "epoch": 0.2, "grad_norm": 16.75973181592774, "learning_rate": 1.9934897717961223e-05, "loss": 0.8984, "step": 1254 }, { "epoch": 0.2, "grad_norm": 19.57298354615714, "learning_rate": 1.9934705459847544e-05, "loss": 0.9365, "step": 1255 }, { "epoch": 0.2, "grad_norm": 16.953908912711572, "learning_rate": 1.9934512919196647e-05, "loss": 0.8488, "step": 1256 }, { "epoch": 0.2, "grad_norm": 28.33725745547857, "learning_rate": 1.9934320096014022e-05, "loss": 0.9324, "step": 1257 }, { "epoch": 0.2, "grad_norm": 21.281497199769635, "learning_rate": 1.9934126990305142e-05, "loss": 0.8084, "step": 1258 }, { "epoch": 0.2, "grad_norm": 7.136670117142064, "learning_rate": 1.99339336020755e-05, "loss": 0.9868, "step": 1259 }, { "epoch": 0.2, "grad_norm": 18.091622239433082, "learning_rate": 1.99337399313306e-05, "loss": 0.8913, "step": 1260 }, { "epoch": 0.2, "grad_norm": 18.635426654212395, "learning_rate": 1.9933545978075942e-05, "loss": 0.8554, "step": 1261 }, { "epoch": 0.2, "grad_norm": 21.474915783037556, "learning_rate": 1.9933351742317053e-05, "loss": 0.983, "step": 1262 }, { "epoch": 0.2, "grad_norm": 29.234804159467206, "learning_rate": 1.993315722405945e-05, "loss": 0.966, "step": 1263 }, { "epoch": 0.2, "grad_norm": 11.672568471062904, "learning_rate": 1.9932962423308666e-05, "loss": 0.8412, "step": 1264 }, { "epoch": 0.2, "grad_norm": 14.449711114972242, "learning_rate": 1.9932767340070243e-05, "loss": 0.9385, "step": 1265 }, { "epoch": 0.2, "grad_norm": 18.669071223225533, "learning_rate": 1.9932571974349724e-05, "loss": 0.8169, "step": 1266 }, { "epoch": 0.2, "grad_norm": 19.992868998436204, "learning_rate": 1.9932376326152672e-05, "loss": 0.9201, "step": 1267 }, { "epoch": 0.2, "grad_norm": 27.272789298085467, "learning_rate": 1.9932180395484644e-05, "loss": 0.9691, "step": 1268 }, { "epoch": 0.2, "grad_norm": 20.28794206515459, "learning_rate": 1.993198418235122e-05, "loss": 0.9517, "step": 1269 }, { "epoch": 0.2, "grad_norm": 17.36011207674172, "learning_rate": 1.993178768675797e-05, "loss": 1.0202, "step": 1270 }, { "epoch": 0.2, "grad_norm": 21.541887003351604, "learning_rate": 1.9931590908710495e-05, "loss": 0.9676, "step": 1271 }, { "epoch": 0.2, "grad_norm": 18.141704428640352, "learning_rate": 1.9931393848214376e-05, "loss": 0.8703, "step": 1272 }, { "epoch": 0.2, "grad_norm": 20.695246094783986, "learning_rate": 1.9931196505275233e-05, "loss": 1.0039, "step": 1273 }, { "epoch": 0.2, "grad_norm": 8.346715071676691, "learning_rate": 1.993099887989867e-05, "loss": 0.8852, "step": 1274 }, { "epoch": 0.2, "grad_norm": 25.672661702505575, "learning_rate": 1.9930800972090307e-05, "loss": 0.9104, "step": 1275 }, { "epoch": 0.2, "grad_norm": 25.618545684106746, "learning_rate": 1.9930602781855772e-05, "loss": 0.8934, "step": 1276 }, { "epoch": 0.2, "grad_norm": 37.476588063711446, "learning_rate": 1.9930404309200706e-05, "loss": 0.9254, "step": 1277 }, { "epoch": 0.2, "grad_norm": 18.0360428722522, "learning_rate": 1.9930205554130748e-05, "loss": 0.9899, "step": 1278 }, { "epoch": 0.2, "grad_norm": 33.663050132821624, "learning_rate": 1.9930006516651555e-05, "loss": 0.9813, "step": 1279 }, { "epoch": 0.2, "grad_norm": 22.84790344808927, "learning_rate": 1.9929807196768785e-05, "loss": 0.9439, "step": 1280 }, { "epoch": 0.2, "grad_norm": 17.98229335110451, "learning_rate": 1.9929607594488104e-05, "loss": 0.8303, "step": 1281 }, { "epoch": 0.2, "grad_norm": 30.035470062652568, "learning_rate": 1.9929407709815197e-05, "loss": 0.9388, "step": 1282 }, { "epoch": 0.2, "grad_norm": 23.215968611692105, "learning_rate": 1.992920754275574e-05, "loss": 0.8859, "step": 1283 }, { "epoch": 0.2, "grad_norm": 4.943706562053954, "learning_rate": 1.992900709331543e-05, "loss": 0.9961, "step": 1284 }, { "epoch": 0.2, "grad_norm": 23.8796046794665, "learning_rate": 1.9928806361499965e-05, "loss": 0.8896, "step": 1285 }, { "epoch": 0.2, "grad_norm": 14.615686040819751, "learning_rate": 1.9928605347315057e-05, "loss": 0.9176, "step": 1286 }, { "epoch": 0.2, "grad_norm": 17.386918984886165, "learning_rate": 1.992840405076642e-05, "loss": 0.8916, "step": 1287 }, { "epoch": 0.2, "grad_norm": 24.019404373913307, "learning_rate": 1.992820247185978e-05, "loss": 1.0235, "step": 1288 }, { "epoch": 0.2, "grad_norm": 23.637710906025376, "learning_rate": 1.992800061060087e-05, "loss": 0.9788, "step": 1289 }, { "epoch": 0.2, "grad_norm": 19.1480366265285, "learning_rate": 1.992779846699543e-05, "loss": 0.9326, "step": 1290 }, { "epoch": 0.2, "grad_norm": 27.699055472836726, "learning_rate": 1.992759604104921e-05, "loss": 0.9894, "step": 1291 }, { "epoch": 0.2, "grad_norm": 28.074772851204248, "learning_rate": 1.9927393332767965e-05, "loss": 1.0129, "step": 1292 }, { "epoch": 0.2, "grad_norm": 22.953500482332803, "learning_rate": 1.9927190342157462e-05, "loss": 1.0021, "step": 1293 }, { "epoch": 0.2, "grad_norm": 21.84006127194746, "learning_rate": 1.9926987069223475e-05, "loss": 0.9293, "step": 1294 }, { "epoch": 0.2, "grad_norm": 20.89379450621844, "learning_rate": 1.9926783513971777e-05, "loss": 0.9532, "step": 1295 }, { "epoch": 0.2, "grad_norm": 35.378460087236796, "learning_rate": 1.9926579676408168e-05, "loss": 1.0128, "step": 1296 }, { "epoch": 0.2, "grad_norm": 17.318854205624405, "learning_rate": 1.992637555653844e-05, "loss": 0.9178, "step": 1297 }, { "epoch": 0.2, "grad_norm": 19.897438525977122, "learning_rate": 1.9926171154368393e-05, "loss": 0.8841, "step": 1298 }, { "epoch": 0.2, "grad_norm": 16.500584955485103, "learning_rate": 1.992596646990385e-05, "loss": 1.0633, "step": 1299 }, { "epoch": 0.2, "grad_norm": 34.49422380219095, "learning_rate": 1.992576150315062e-05, "loss": 0.7963, "step": 1300 }, { "epoch": 0.2, "grad_norm": 19.81966630425843, "learning_rate": 1.9925556254114547e-05, "loss": 0.8769, "step": 1301 }, { "epoch": 0.2, "grad_norm": 29.164915812711097, "learning_rate": 1.9925350722801458e-05, "loss": 0.9607, "step": 1302 }, { "epoch": 0.2, "grad_norm": 25.58943896263914, "learning_rate": 1.9925144909217197e-05, "loss": 0.9324, "step": 1303 }, { "epoch": 0.2, "grad_norm": 25.13568356508364, "learning_rate": 1.9924938813367626e-05, "loss": 0.8545, "step": 1304 }, { "epoch": 0.2, "grad_norm": 25.45426296385659, "learning_rate": 1.99247324352586e-05, "loss": 1.0189, "step": 1305 }, { "epoch": 0.2, "grad_norm": 31.72773203400263, "learning_rate": 1.9924525774895986e-05, "loss": 0.9538, "step": 1306 }, { "epoch": 0.2, "grad_norm": 27.50644130479907, "learning_rate": 1.9924318832285667e-05, "loss": 1.0431, "step": 1307 }, { "epoch": 0.2, "grad_norm": 20.018262046216933, "learning_rate": 1.9924111607433528e-05, "loss": 0.9644, "step": 1308 }, { "epoch": 0.2, "grad_norm": 16.072629528814268, "learning_rate": 1.992390410034546e-05, "loss": 0.9031, "step": 1309 }, { "epoch": 0.2, "grad_norm": 24.638114951632378, "learning_rate": 1.992369631102736e-05, "loss": 1.0506, "step": 1310 }, { "epoch": 0.2, "grad_norm": 21.19323599795892, "learning_rate": 1.9923488239485146e-05, "loss": 0.9868, "step": 1311 }, { "epoch": 0.2, "grad_norm": 23.022933296867507, "learning_rate": 1.992327988572473e-05, "loss": 1.0442, "step": 1312 }, { "epoch": 0.21, "grad_norm": 38.60077075343159, "learning_rate": 1.9923071249752038e-05, "loss": 0.9415, "step": 1313 }, { "epoch": 0.21, "grad_norm": 16.935327725801237, "learning_rate": 1.9922862331573007e-05, "loss": 0.8662, "step": 1314 }, { "epoch": 0.21, "grad_norm": 21.549857823175042, "learning_rate": 1.9922653131193575e-05, "loss": 0.9445, "step": 1315 }, { "epoch": 0.21, "grad_norm": 16.42482200842557, "learning_rate": 1.9922443648619696e-05, "loss": 0.8671, "step": 1316 }, { "epoch": 0.21, "grad_norm": 31.910562303238194, "learning_rate": 1.992223388385732e-05, "loss": 0.8879, "step": 1317 }, { "epoch": 0.21, "grad_norm": 15.252798493002622, "learning_rate": 1.9922023836912418e-05, "loss": 0.8011, "step": 1318 }, { "epoch": 0.21, "grad_norm": 14.150754687231103, "learning_rate": 1.992181350779096e-05, "loss": 0.8793, "step": 1319 }, { "epoch": 0.21, "grad_norm": 17.31881387900814, "learning_rate": 1.9921602896498934e-05, "loss": 0.8197, "step": 1320 }, { "epoch": 0.21, "grad_norm": 19.93742356016639, "learning_rate": 1.9921392003042322e-05, "loss": 0.9119, "step": 1321 }, { "epoch": 0.21, "grad_norm": 13.878665614920763, "learning_rate": 1.9921180827427126e-05, "loss": 0.8363, "step": 1322 }, { "epoch": 0.21, "grad_norm": 4.5208541018877675, "learning_rate": 1.9920969369659354e-05, "loss": 1.0458, "step": 1323 }, { "epoch": 0.21, "grad_norm": 21.1772940981394, "learning_rate": 1.9920757629745015e-05, "loss": 0.9305, "step": 1324 }, { "epoch": 0.21, "grad_norm": 17.442677599117722, "learning_rate": 1.992054560769013e-05, "loss": 0.9774, "step": 1325 }, { "epoch": 0.21, "grad_norm": 22.854559087963043, "learning_rate": 1.9920333303500734e-05, "loss": 0.9629, "step": 1326 }, { "epoch": 0.21, "grad_norm": 35.30302763377855, "learning_rate": 1.992012071718286e-05, "loss": 0.9763, "step": 1327 }, { "epoch": 0.21, "grad_norm": 22.153192085498706, "learning_rate": 1.9919907848742556e-05, "loss": 0.9926, "step": 1328 }, { "epoch": 0.21, "grad_norm": 17.260157248420533, "learning_rate": 1.991969469818588e-05, "loss": 0.9913, "step": 1329 }, { "epoch": 0.21, "grad_norm": 19.96376624247682, "learning_rate": 1.9919481265518885e-05, "loss": 0.9022, "step": 1330 }, { "epoch": 0.21, "grad_norm": 34.18917028402202, "learning_rate": 1.9919267550747647e-05, "loss": 0.9388, "step": 1331 }, { "epoch": 0.21, "grad_norm": 14.98196083705158, "learning_rate": 1.991905355387824e-05, "loss": 0.8616, "step": 1332 }, { "epoch": 0.21, "grad_norm": 16.595385975033494, "learning_rate": 1.9918839274916757e-05, "loss": 0.8426, "step": 1333 }, { "epoch": 0.21, "grad_norm": 21.307162634521717, "learning_rate": 1.9918624713869287e-05, "loss": 0.895, "step": 1334 }, { "epoch": 0.21, "grad_norm": 16.081120207404, "learning_rate": 1.9918409870741928e-05, "loss": 0.7825, "step": 1335 }, { "epoch": 0.21, "grad_norm": 19.539929219347595, "learning_rate": 1.9918194745540798e-05, "loss": 0.8481, "step": 1336 }, { "epoch": 0.21, "grad_norm": 33.29577674826702, "learning_rate": 1.991797933827201e-05, "loss": 0.895, "step": 1337 }, { "epoch": 0.21, "grad_norm": 22.000071211179446, "learning_rate": 1.9917763648941692e-05, "loss": 1.023, "step": 1338 }, { "epoch": 0.21, "grad_norm": 16.508815965063587, "learning_rate": 1.991754767755598e-05, "loss": 0.8633, "step": 1339 }, { "epoch": 0.21, "grad_norm": 17.74053182565702, "learning_rate": 1.991733142412101e-05, "loss": 0.9433, "step": 1340 }, { "epoch": 0.21, "grad_norm": 15.644052303855188, "learning_rate": 1.9917114888642937e-05, "loss": 0.8993, "step": 1341 }, { "epoch": 0.21, "grad_norm": 23.775108462957714, "learning_rate": 1.9916898071127915e-05, "loss": 0.9787, "step": 1342 }, { "epoch": 0.21, "grad_norm": 22.003199432061866, "learning_rate": 1.9916680971582116e-05, "loss": 0.9631, "step": 1343 }, { "epoch": 0.21, "grad_norm": 14.829538243320078, "learning_rate": 1.9916463590011713e-05, "loss": 0.8365, "step": 1344 }, { "epoch": 0.21, "grad_norm": 24.878333483592566, "learning_rate": 1.9916245926422883e-05, "loss": 0.9273, "step": 1345 }, { "epoch": 0.21, "grad_norm": 16.257384510446926, "learning_rate": 1.9916027980821823e-05, "loss": 0.8607, "step": 1346 }, { "epoch": 0.21, "grad_norm": 17.07316436252357, "learning_rate": 1.9915809753214725e-05, "loss": 0.8915, "step": 1347 }, { "epoch": 0.21, "grad_norm": 13.959282926977835, "learning_rate": 1.99155912436078e-05, "loss": 0.741, "step": 1348 }, { "epoch": 0.21, "grad_norm": 20.978979063158075, "learning_rate": 1.991537245200726e-05, "loss": 0.8688, "step": 1349 }, { "epoch": 0.21, "grad_norm": 27.744051462182586, "learning_rate": 1.991515337841933e-05, "loss": 1.12, "step": 1350 }, { "epoch": 0.21, "grad_norm": 15.25120064531158, "learning_rate": 1.9914934022850233e-05, "loss": 0.8152, "step": 1351 }, { "epoch": 0.21, "grad_norm": 23.29446485142575, "learning_rate": 1.9914714385306217e-05, "loss": 0.9278, "step": 1352 }, { "epoch": 0.21, "grad_norm": 16.049437390735815, "learning_rate": 1.9914494465793522e-05, "loss": 0.8964, "step": 1353 }, { "epoch": 0.21, "grad_norm": 21.324527902453276, "learning_rate": 1.9914274264318402e-05, "loss": 0.9232, "step": 1354 }, { "epoch": 0.21, "grad_norm": 12.591530142767086, "learning_rate": 1.9914053780887122e-05, "loss": 0.8671, "step": 1355 }, { "epoch": 0.21, "grad_norm": 21.597951831475005, "learning_rate": 1.9913833015505957e-05, "loss": 0.9686, "step": 1356 }, { "epoch": 0.21, "grad_norm": 20.237243762700256, "learning_rate": 1.9913611968181178e-05, "loss": 0.925, "step": 1357 }, { "epoch": 0.21, "grad_norm": 18.727077098296117, "learning_rate": 1.9913390638919074e-05, "loss": 0.879, "step": 1358 }, { "epoch": 0.21, "grad_norm": 27.987888497407347, "learning_rate": 1.9913169027725936e-05, "loss": 1.0334, "step": 1359 }, { "epoch": 0.21, "grad_norm": 30.733868300949965, "learning_rate": 1.9912947134608073e-05, "loss": 0.9026, "step": 1360 }, { "epoch": 0.21, "grad_norm": 16.173910454258483, "learning_rate": 1.9912724959571793e-05, "loss": 0.9208, "step": 1361 }, { "epoch": 0.21, "grad_norm": 26.230580679171098, "learning_rate": 1.9912502502623413e-05, "loss": 0.9103, "step": 1362 }, { "epoch": 0.21, "grad_norm": 19.052439439329678, "learning_rate": 1.991227976376926e-05, "loss": 0.824, "step": 1363 }, { "epoch": 0.21, "grad_norm": 17.76117529959484, "learning_rate": 1.991205674301567e-05, "loss": 0.8669, "step": 1364 }, { "epoch": 0.21, "grad_norm": 17.210222016837943, "learning_rate": 1.9911833440368982e-05, "loss": 0.7947, "step": 1365 }, { "epoch": 0.21, "grad_norm": 33.605481570448795, "learning_rate": 1.9911609855835553e-05, "loss": 0.9422, "step": 1366 }, { "epoch": 0.21, "grad_norm": 18.18952214004595, "learning_rate": 1.9911385989421736e-05, "loss": 0.8542, "step": 1367 }, { "epoch": 0.21, "grad_norm": 21.163667112580974, "learning_rate": 1.9911161841133898e-05, "loss": 0.8978, "step": 1368 }, { "epoch": 0.21, "grad_norm": 30.03465533772165, "learning_rate": 1.9910937410978418e-05, "loss": 0.8659, "step": 1369 }, { "epoch": 0.21, "grad_norm": 14.56129051678673, "learning_rate": 1.9910712698961673e-05, "loss": 0.8997, "step": 1370 }, { "epoch": 0.21, "grad_norm": 21.7784257478019, "learning_rate": 1.991048770509006e-05, "loss": 0.9045, "step": 1371 }, { "epoch": 0.21, "grad_norm": 14.464371663980677, "learning_rate": 1.991026242936997e-05, "loss": 0.8788, "step": 1372 }, { "epoch": 0.21, "grad_norm": 23.28065532590489, "learning_rate": 1.9910036871807817e-05, "loss": 1.0487, "step": 1373 }, { "epoch": 0.21, "grad_norm": 26.378756413259477, "learning_rate": 1.9909811032410012e-05, "loss": 0.8824, "step": 1374 }, { "epoch": 0.21, "grad_norm": 24.43542564349211, "learning_rate": 1.9909584911182977e-05, "loss": 0.9517, "step": 1375 }, { "epoch": 0.21, "grad_norm": 26.59946889494999, "learning_rate": 1.9909358508133145e-05, "loss": 0.9472, "step": 1376 }, { "epoch": 0.22, "grad_norm": 24.575767925465115, "learning_rate": 1.990913182326695e-05, "loss": 0.9322, "step": 1377 }, { "epoch": 0.22, "grad_norm": 24.283341835010557, "learning_rate": 1.990890485659085e-05, "loss": 0.877, "step": 1378 }, { "epoch": 0.22, "grad_norm": 24.015921680538895, "learning_rate": 1.9908677608111287e-05, "loss": 0.9034, "step": 1379 }, { "epoch": 0.22, "grad_norm": 32.386463181298566, "learning_rate": 1.9908450077834732e-05, "loss": 1.018, "step": 1380 }, { "epoch": 0.22, "grad_norm": 23.62281652293308, "learning_rate": 1.9908222265767653e-05, "loss": 0.9683, "step": 1381 }, { "epoch": 0.22, "grad_norm": 12.893433320441366, "learning_rate": 1.9907994171916524e-05, "loss": 0.9804, "step": 1382 }, { "epoch": 0.22, "grad_norm": 19.677590592003387, "learning_rate": 1.9907765796287844e-05, "loss": 0.9403, "step": 1383 }, { "epoch": 0.22, "grad_norm": 22.765449230010383, "learning_rate": 1.9907537138888097e-05, "loss": 0.9905, "step": 1384 }, { "epoch": 0.22, "grad_norm": 18.90626779981043, "learning_rate": 1.990730819972379e-05, "loss": 0.8958, "step": 1385 }, { "epoch": 0.22, "grad_norm": 22.605297923163597, "learning_rate": 1.9907078978801432e-05, "loss": 0.8924, "step": 1386 }, { "epoch": 0.22, "grad_norm": 22.121787663737475, "learning_rate": 1.9906849476127545e-05, "loss": 0.8506, "step": 1387 }, { "epoch": 0.22, "grad_norm": 25.914299010726207, "learning_rate": 1.9906619691708652e-05, "loss": 0.9328, "step": 1388 }, { "epoch": 0.22, "grad_norm": 23.406456402733113, "learning_rate": 1.9906389625551294e-05, "loss": 0.9968, "step": 1389 }, { "epoch": 0.22, "grad_norm": 21.848600183369733, "learning_rate": 1.990615927766201e-05, "loss": 0.9835, "step": 1390 }, { "epoch": 0.22, "grad_norm": 26.99942841962401, "learning_rate": 1.9905928648047346e-05, "loss": 0.9886, "step": 1391 }, { "epoch": 0.22, "grad_norm": 14.034098767161495, "learning_rate": 1.990569773671387e-05, "loss": 0.9693, "step": 1392 }, { "epoch": 0.22, "grad_norm": 35.71487357663943, "learning_rate": 1.9905466543668143e-05, "loss": 1.1704, "step": 1393 }, { "epoch": 0.22, "grad_norm": 41.45938206197676, "learning_rate": 1.9905235068916745e-05, "loss": 0.9234, "step": 1394 }, { "epoch": 0.22, "grad_norm": 22.495776853003562, "learning_rate": 1.9905003312466255e-05, "loss": 0.9915, "step": 1395 }, { "epoch": 0.22, "grad_norm": 13.212968858485853, "learning_rate": 1.9904771274323262e-05, "loss": 0.8888, "step": 1396 }, { "epoch": 0.22, "grad_norm": 19.54383243693455, "learning_rate": 1.990453895449437e-05, "loss": 0.9925, "step": 1397 }, { "epoch": 0.22, "grad_norm": 13.750532358874487, "learning_rate": 1.9904306352986183e-05, "loss": 0.8366, "step": 1398 }, { "epoch": 0.22, "grad_norm": 19.479072838232593, "learning_rate": 1.990407346980532e-05, "loss": 0.959, "step": 1399 }, { "epoch": 0.22, "grad_norm": 28.19172560606223, "learning_rate": 1.99038403049584e-05, "loss": 0.8447, "step": 1400 }, { "epoch": 0.22, "grad_norm": 27.135755440848115, "learning_rate": 1.9903606858452056e-05, "loss": 0.879, "step": 1401 }, { "epoch": 0.22, "grad_norm": 23.06888160182277, "learning_rate": 1.9903373130292928e-05, "loss": 0.9948, "step": 1402 }, { "epoch": 0.22, "grad_norm": 17.61414992076343, "learning_rate": 1.9903139120487654e-05, "loss": 0.8504, "step": 1403 }, { "epoch": 0.22, "grad_norm": 16.871522748279556, "learning_rate": 1.9902904829042902e-05, "loss": 0.9806, "step": 1404 }, { "epoch": 0.22, "grad_norm": 29.369601845853666, "learning_rate": 1.9902670255965333e-05, "loss": 0.9354, "step": 1405 }, { "epoch": 0.22, "grad_norm": 19.74389234094852, "learning_rate": 1.9902435401261608e-05, "loss": 0.8639, "step": 1406 }, { "epoch": 0.22, "grad_norm": 17.603213032638298, "learning_rate": 1.990220026493842e-05, "loss": 0.9023, "step": 1407 }, { "epoch": 0.22, "grad_norm": 20.97678157365228, "learning_rate": 1.9901964847002442e-05, "loss": 1.014, "step": 1408 }, { "epoch": 0.22, "grad_norm": 16.73880095593121, "learning_rate": 1.990172914746038e-05, "loss": 0.9541, "step": 1409 }, { "epoch": 0.22, "grad_norm": 15.593196662706202, "learning_rate": 1.9901493166318934e-05, "loss": 0.8117, "step": 1410 }, { "epoch": 0.22, "grad_norm": 20.13659180417085, "learning_rate": 1.9901256903584812e-05, "loss": 0.8954, "step": 1411 }, { "epoch": 0.22, "grad_norm": 24.86775771340296, "learning_rate": 1.9901020359264738e-05, "loss": 1.0768, "step": 1412 }, { "epoch": 0.22, "grad_norm": 18.893536761352216, "learning_rate": 1.9900783533365437e-05, "loss": 0.8369, "step": 1413 }, { "epoch": 0.22, "grad_norm": 30.969588877520685, "learning_rate": 1.9900546425893644e-05, "loss": 0.851, "step": 1414 }, { "epoch": 0.22, "grad_norm": 30.364502580785054, "learning_rate": 1.99003090368561e-05, "loss": 0.9364, "step": 1415 }, { "epoch": 0.22, "grad_norm": 25.64482005553018, "learning_rate": 1.990007136625956e-05, "loss": 0.9606, "step": 1416 }, { "epoch": 0.22, "grad_norm": 23.860730601099355, "learning_rate": 1.989983341411078e-05, "loss": 0.8902, "step": 1417 }, { "epoch": 0.22, "grad_norm": 26.945570604122867, "learning_rate": 1.9899595180416532e-05, "loss": 1.036, "step": 1418 }, { "epoch": 0.22, "grad_norm": 22.585343567181834, "learning_rate": 1.9899356665183584e-05, "loss": 0.9409, "step": 1419 }, { "epoch": 0.22, "grad_norm": 19.592404568031984, "learning_rate": 1.989911786841873e-05, "loss": 0.8498, "step": 1420 }, { "epoch": 0.22, "grad_norm": 19.34782284981256, "learning_rate": 1.9898878790128752e-05, "loss": 0.862, "step": 1421 }, { "epoch": 0.22, "grad_norm": 22.330109461349725, "learning_rate": 1.9898639430320448e-05, "loss": 0.9395, "step": 1422 }, { "epoch": 0.22, "grad_norm": 17.958950574966494, "learning_rate": 1.9898399789000633e-05, "loss": 0.8653, "step": 1423 }, { "epoch": 0.22, "grad_norm": 27.022031537393666, "learning_rate": 1.9898159866176115e-05, "loss": 0.8856, "step": 1424 }, { "epoch": 0.22, "grad_norm": 21.055090281468015, "learning_rate": 1.9897919661853723e-05, "loss": 0.9617, "step": 1425 }, { "epoch": 0.22, "grad_norm": 19.808843135298243, "learning_rate": 1.9897679176040284e-05, "loss": 0.8153, "step": 1426 }, { "epoch": 0.22, "grad_norm": 26.98273301223038, "learning_rate": 1.9897438408742644e-05, "loss": 0.9232, "step": 1427 }, { "epoch": 0.22, "grad_norm": 29.467110019854843, "learning_rate": 1.9897197359967642e-05, "loss": 0.899, "step": 1428 }, { "epoch": 0.22, "grad_norm": 26.321039026185428, "learning_rate": 1.9896956029722138e-05, "loss": 1.0245, "step": 1429 }, { "epoch": 0.22, "grad_norm": 29.247209941626622, "learning_rate": 1.9896714418012993e-05, "loss": 0.8231, "step": 1430 }, { "epoch": 0.22, "grad_norm": 26.348569622001488, "learning_rate": 1.9896472524847078e-05, "loss": 0.961, "step": 1431 }, { "epoch": 0.22, "grad_norm": 37.56406826166652, "learning_rate": 1.9896230350231274e-05, "loss": 0.8571, "step": 1432 }, { "epoch": 0.22, "grad_norm": 20.53945002910316, "learning_rate": 1.9895987894172474e-05, "loss": 0.8621, "step": 1433 }, { "epoch": 0.22, "grad_norm": 24.920721265341765, "learning_rate": 1.989574515667756e-05, "loss": 1.022, "step": 1434 }, { "epoch": 0.22, "grad_norm": 16.206127859641747, "learning_rate": 1.9895502137753448e-05, "loss": 0.8861, "step": 1435 }, { "epoch": 0.22, "grad_norm": 17.513001768723406, "learning_rate": 1.989525883740704e-05, "loss": 0.8813, "step": 1436 }, { "epoch": 0.22, "grad_norm": 25.080606141750987, "learning_rate": 1.989501525564526e-05, "loss": 0.9389, "step": 1437 }, { "epoch": 0.22, "grad_norm": 26.902371726278396, "learning_rate": 1.9894771392475036e-05, "loss": 1.035, "step": 1438 }, { "epoch": 0.22, "grad_norm": 29.802421383683235, "learning_rate": 1.98945272479033e-05, "loss": 0.9444, "step": 1439 }, { "epoch": 0.22, "grad_norm": 17.20849078913029, "learning_rate": 1.9894282821936995e-05, "loss": 0.7965, "step": 1440 }, { "epoch": 0.23, "grad_norm": 23.171240186323416, "learning_rate": 1.989403811458308e-05, "loss": 0.9591, "step": 1441 }, { "epoch": 0.23, "grad_norm": 39.5019219963036, "learning_rate": 1.9893793125848506e-05, "loss": 1.0449, "step": 1442 }, { "epoch": 0.23, "grad_norm": 25.663329008585723, "learning_rate": 1.9893547855740243e-05, "loss": 0.9246, "step": 1443 }, { "epoch": 0.23, "grad_norm": 24.112498416673166, "learning_rate": 1.9893302304265267e-05, "loss": 0.8921, "step": 1444 }, { "epoch": 0.23, "grad_norm": 20.039729486857773, "learning_rate": 1.989305647143056e-05, "loss": 0.8282, "step": 1445 }, { "epoch": 0.23, "grad_norm": 20.691181532359877, "learning_rate": 1.9892810357243116e-05, "loss": 0.8882, "step": 1446 }, { "epoch": 0.23, "grad_norm": 19.225087008849833, "learning_rate": 1.989256396170993e-05, "loss": 0.8138, "step": 1447 }, { "epoch": 0.23, "grad_norm": 24.41466549670376, "learning_rate": 1.9892317284838016e-05, "loss": 1.0354, "step": 1448 }, { "epoch": 0.23, "grad_norm": 15.987971303075874, "learning_rate": 1.9892070326634383e-05, "loss": 0.9045, "step": 1449 }, { "epoch": 0.23, "grad_norm": 33.14052844179611, "learning_rate": 1.9891823087106057e-05, "loss": 0.9687, "step": 1450 }, { "epoch": 0.23, "grad_norm": 18.543345362147758, "learning_rate": 1.9891575566260068e-05, "loss": 0.8688, "step": 1451 }, { "epoch": 0.23, "grad_norm": 31.82596494286104, "learning_rate": 1.9891327764103456e-05, "loss": 0.8913, "step": 1452 }, { "epoch": 0.23, "grad_norm": 15.102029975724584, "learning_rate": 1.989107968064327e-05, "loss": 0.8815, "step": 1453 }, { "epoch": 0.23, "grad_norm": 25.800652465672165, "learning_rate": 1.9890831315886566e-05, "loss": 0.9337, "step": 1454 }, { "epoch": 0.23, "grad_norm": 23.492450865502047, "learning_rate": 1.98905826698404e-05, "loss": 0.9639, "step": 1455 }, { "epoch": 0.23, "grad_norm": 36.36599028160405, "learning_rate": 1.989033374251185e-05, "loss": 0.9794, "step": 1456 }, { "epoch": 0.23, "grad_norm": 18.531828198249837, "learning_rate": 1.9890084533907998e-05, "loss": 0.836, "step": 1457 }, { "epoch": 0.23, "grad_norm": 16.454867441883128, "learning_rate": 1.9889835044035925e-05, "loss": 0.9107, "step": 1458 }, { "epoch": 0.23, "grad_norm": 29.831723663035085, "learning_rate": 1.988958527290273e-05, "loss": 0.9699, "step": 1459 }, { "epoch": 0.23, "grad_norm": 30.106522068630024, "learning_rate": 1.9889335220515512e-05, "loss": 0.902, "step": 1460 }, { "epoch": 0.23, "grad_norm": 17.6693855491818, "learning_rate": 1.9889084886881387e-05, "loss": 0.843, "step": 1461 }, { "epoch": 0.23, "grad_norm": 18.156992833080583, "learning_rate": 1.9888834272007475e-05, "loss": 0.8507, "step": 1462 }, { "epoch": 0.23, "grad_norm": 16.58564273789724, "learning_rate": 1.9888583375900897e-05, "loss": 0.8502, "step": 1463 }, { "epoch": 0.23, "grad_norm": 20.48048829094658, "learning_rate": 1.9888332198568794e-05, "loss": 0.8605, "step": 1464 }, { "epoch": 0.23, "grad_norm": 24.514160969565783, "learning_rate": 1.988808074001831e-05, "loss": 0.8924, "step": 1465 }, { "epoch": 0.23, "grad_norm": 23.93567949722883, "learning_rate": 1.9887829000256592e-05, "loss": 0.8573, "step": 1466 }, { "epoch": 0.23, "grad_norm": 18.38912034665384, "learning_rate": 1.9887576979290803e-05, "loss": 0.8493, "step": 1467 }, { "epoch": 0.23, "grad_norm": 26.441573061468294, "learning_rate": 1.9887324677128107e-05, "loss": 0.8848, "step": 1468 }, { "epoch": 0.23, "grad_norm": 20.710959270815522, "learning_rate": 1.988707209377568e-05, "loss": 1.0571, "step": 1469 }, { "epoch": 0.23, "grad_norm": 14.66331877496578, "learning_rate": 1.9886819229240708e-05, "loss": 0.763, "step": 1470 }, { "epoch": 0.23, "grad_norm": 20.84368962246273, "learning_rate": 1.9886566083530382e-05, "loss": 0.9172, "step": 1471 }, { "epoch": 0.23, "grad_norm": 18.960246015749057, "learning_rate": 1.9886312656651897e-05, "loss": 0.9751, "step": 1472 }, { "epoch": 0.23, "grad_norm": 13.7503540177721, "learning_rate": 1.9886058948612466e-05, "loss": 0.827, "step": 1473 }, { "epoch": 0.23, "grad_norm": 18.23471947012008, "learning_rate": 1.98858049594193e-05, "loss": 0.8813, "step": 1474 }, { "epoch": 0.23, "grad_norm": 27.443001082697744, "learning_rate": 1.9885550689079624e-05, "loss": 0.9107, "step": 1475 }, { "epoch": 0.23, "grad_norm": 23.963675917420925, "learning_rate": 1.988529613760067e-05, "loss": 1.0011, "step": 1476 }, { "epoch": 0.23, "grad_norm": 15.62294501476367, "learning_rate": 1.9885041304989682e-05, "loss": 0.8256, "step": 1477 }, { "epoch": 0.23, "grad_norm": 22.379305682327313, "learning_rate": 1.9884786191253895e-05, "loss": 0.9107, "step": 1478 }, { "epoch": 0.23, "grad_norm": 17.226568236750126, "learning_rate": 1.988453079640057e-05, "loss": 0.9533, "step": 1479 }, { "epoch": 0.23, "grad_norm": 22.931625383750685, "learning_rate": 1.9884275120436976e-05, "loss": 0.8969, "step": 1480 }, { "epoch": 0.23, "grad_norm": 28.762895368604905, "learning_rate": 1.988401916337038e-05, "loss": 0.947, "step": 1481 }, { "epoch": 0.23, "grad_norm": 17.57032215038423, "learning_rate": 1.9883762925208058e-05, "loss": 0.8551, "step": 1482 }, { "epoch": 0.23, "grad_norm": 24.22208747140622, "learning_rate": 1.98835064059573e-05, "loss": 0.8877, "step": 1483 }, { "epoch": 0.23, "grad_norm": 37.12283804808617, "learning_rate": 1.9883249605625403e-05, "loss": 1.0072, "step": 1484 }, { "epoch": 0.23, "grad_norm": 14.543397514422557, "learning_rate": 1.988299252421967e-05, "loss": 0.9646, "step": 1485 }, { "epoch": 0.23, "grad_norm": 31.81226664028502, "learning_rate": 1.9882735161747407e-05, "loss": 0.8968, "step": 1486 }, { "epoch": 0.23, "grad_norm": 16.148444937919745, "learning_rate": 1.9882477518215938e-05, "loss": 0.8184, "step": 1487 }, { "epoch": 0.23, "grad_norm": 22.605271616238163, "learning_rate": 1.9882219593632595e-05, "loss": 0.8806, "step": 1488 }, { "epoch": 0.23, "grad_norm": 22.401248724863805, "learning_rate": 1.9881961388004702e-05, "loss": 0.8844, "step": 1489 }, { "epoch": 0.23, "grad_norm": 16.85256456659565, "learning_rate": 1.988170290133961e-05, "loss": 0.8483, "step": 1490 }, { "epoch": 0.23, "grad_norm": 22.243625712281947, "learning_rate": 1.9881444133644668e-05, "loss": 0.918, "step": 1491 }, { "epoch": 0.23, "grad_norm": 19.86633504313018, "learning_rate": 1.9881185084927234e-05, "loss": 0.914, "step": 1492 }, { "epoch": 0.23, "grad_norm": 23.082115977569647, "learning_rate": 1.9880925755194677e-05, "loss": 0.9888, "step": 1493 }, { "epoch": 0.23, "grad_norm": 18.567946105511787, "learning_rate": 1.9880666144454372e-05, "loss": 0.8667, "step": 1494 }, { "epoch": 0.23, "grad_norm": 18.103235278591296, "learning_rate": 1.9880406252713703e-05, "loss": 0.8576, "step": 1495 }, { "epoch": 0.23, "grad_norm": 32.05414053766156, "learning_rate": 1.988014607998006e-05, "loss": 0.9628, "step": 1496 }, { "epoch": 0.23, "grad_norm": 20.314446426447713, "learning_rate": 1.987988562626084e-05, "loss": 1.0049, "step": 1497 }, { "epoch": 0.23, "grad_norm": 22.489690572149616, "learning_rate": 1.9879624891563456e-05, "loss": 0.8725, "step": 1498 }, { "epoch": 0.23, "grad_norm": 19.839622826800472, "learning_rate": 1.9879363875895317e-05, "loss": 0.9889, "step": 1499 }, { "epoch": 0.23, "grad_norm": 25.44151166791262, "learning_rate": 1.987910257926385e-05, "loss": 0.9919, "step": 1500 }, { "epoch": 0.23, "grad_norm": 21.1580910476791, "learning_rate": 1.9878841001676484e-05, "loss": 0.846, "step": 1501 }, { "epoch": 0.23, "grad_norm": 12.442634325784066, "learning_rate": 1.9878579143140656e-05, "loss": 0.8613, "step": 1502 }, { "epoch": 0.23, "grad_norm": 25.65284426453401, "learning_rate": 1.987831700366382e-05, "loss": 0.9577, "step": 1503 }, { "epoch": 0.23, "grad_norm": 23.49152889666891, "learning_rate": 1.9878054583253426e-05, "loss": 0.8942, "step": 1504 }, { "epoch": 0.24, "grad_norm": 17.020052578138305, "learning_rate": 1.9877791881916935e-05, "loss": 0.8903, "step": 1505 }, { "epoch": 0.24, "grad_norm": 16.929546441871096, "learning_rate": 1.9877528899661826e-05, "loss": 0.8664, "step": 1506 }, { "epoch": 0.24, "grad_norm": 16.08767658742868, "learning_rate": 1.9877265636495574e-05, "loss": 0.8503, "step": 1507 }, { "epoch": 0.24, "grad_norm": 24.611948125225855, "learning_rate": 1.987700209242566e-05, "loss": 0.9032, "step": 1508 }, { "epoch": 0.24, "grad_norm": 17.24670966874891, "learning_rate": 1.9876738267459587e-05, "loss": 0.8426, "step": 1509 }, { "epoch": 0.24, "grad_norm": 17.202767041853868, "learning_rate": 1.9876474161604858e-05, "loss": 0.8817, "step": 1510 }, { "epoch": 0.24, "grad_norm": 21.355531537159216, "learning_rate": 1.9876209774868975e-05, "loss": 0.9919, "step": 1511 }, { "epoch": 0.24, "grad_norm": 18.44822621871114, "learning_rate": 1.9875945107259467e-05, "loss": 0.9514, "step": 1512 }, { "epoch": 0.24, "grad_norm": 15.600634870616602, "learning_rate": 1.987568015878386e-05, "loss": 0.8976, "step": 1513 }, { "epoch": 0.24, "grad_norm": 42.65773799033624, "learning_rate": 1.9875414929449684e-05, "loss": 0.9613, "step": 1514 }, { "epoch": 0.24, "grad_norm": 23.868428440579983, "learning_rate": 1.9875149419264483e-05, "loss": 0.8894, "step": 1515 }, { "epoch": 0.24, "grad_norm": 21.1866930186212, "learning_rate": 1.987488362823581e-05, "loss": 0.8675, "step": 1516 }, { "epoch": 0.24, "grad_norm": 19.6598789347875, "learning_rate": 1.9874617556371224e-05, "loss": 0.8718, "step": 1517 }, { "epoch": 0.24, "grad_norm": 27.132711136285618, "learning_rate": 1.987435120367829e-05, "loss": 0.9302, "step": 1518 }, { "epoch": 0.24, "grad_norm": 28.074359770835063, "learning_rate": 1.987408457016459e-05, "loss": 0.8548, "step": 1519 }, { "epoch": 0.24, "grad_norm": 18.889135313023882, "learning_rate": 1.9873817655837695e-05, "loss": 0.877, "step": 1520 }, { "epoch": 0.24, "grad_norm": 20.15093699153894, "learning_rate": 1.9873550460705203e-05, "loss": 0.7814, "step": 1521 }, { "epoch": 0.24, "grad_norm": 41.35476704008727, "learning_rate": 1.987328298477471e-05, "loss": 0.8904, "step": 1522 }, { "epoch": 0.24, "grad_norm": 16.969927016775248, "learning_rate": 1.9873015228053828e-05, "loss": 0.8934, "step": 1523 }, { "epoch": 0.24, "grad_norm": 20.84439129393353, "learning_rate": 1.9872747190550167e-05, "loss": 0.987, "step": 1524 }, { "epoch": 0.24, "grad_norm": 16.562192149270235, "learning_rate": 1.987247887227135e-05, "loss": 0.8728, "step": 1525 }, { "epoch": 0.24, "grad_norm": 20.610440172458183, "learning_rate": 1.9872210273225012e-05, "loss": 0.845, "step": 1526 }, { "epoch": 0.24, "grad_norm": 28.046246185266774, "learning_rate": 1.9871941393418788e-05, "loss": 0.8761, "step": 1527 }, { "epoch": 0.24, "grad_norm": 24.19092754233467, "learning_rate": 1.9871672232860323e-05, "loss": 0.9944, "step": 1528 }, { "epoch": 0.24, "grad_norm": 14.301384599738446, "learning_rate": 1.9871402791557276e-05, "loss": 0.9345, "step": 1529 }, { "epoch": 0.24, "grad_norm": 9.007479441077614, "learning_rate": 1.987113306951731e-05, "loss": 0.9111, "step": 1530 }, { "epoch": 0.24, "grad_norm": 18.403992620522057, "learning_rate": 1.9870863066748092e-05, "loss": 0.8233, "step": 1531 }, { "epoch": 0.24, "grad_norm": 24.482426596277037, "learning_rate": 1.9870592783257303e-05, "loss": 0.8811, "step": 1532 }, { "epoch": 0.24, "grad_norm": 18.977830127577914, "learning_rate": 1.987032221905263e-05, "loss": 0.8781, "step": 1533 }, { "epoch": 0.24, "grad_norm": 14.989255471340897, "learning_rate": 1.9870051374141765e-05, "loss": 0.7822, "step": 1534 }, { "epoch": 0.24, "grad_norm": 24.703784181026588, "learning_rate": 1.9869780248532416e-05, "loss": 0.963, "step": 1535 }, { "epoch": 0.24, "grad_norm": 21.42849824799235, "learning_rate": 1.986950884223229e-05, "loss": 0.8673, "step": 1536 }, { "epoch": 0.24, "grad_norm": 19.104362050169783, "learning_rate": 1.9869237155249105e-05, "loss": 0.9335, "step": 1537 }, { "epoch": 0.24, "grad_norm": 19.762347184048217, "learning_rate": 1.9868965187590588e-05, "loss": 0.9469, "step": 1538 }, { "epoch": 0.24, "grad_norm": 26.174756447752902, "learning_rate": 1.9868692939264474e-05, "loss": 0.9615, "step": 1539 }, { "epoch": 0.24, "grad_norm": 27.08999698213284, "learning_rate": 1.9868420410278506e-05, "loss": 0.9476, "step": 1540 }, { "epoch": 0.24, "grad_norm": 23.4324296667565, "learning_rate": 1.9868147600640433e-05, "loss": 0.8804, "step": 1541 }, { "epoch": 0.24, "grad_norm": 15.979844326593657, "learning_rate": 1.9867874510358017e-05, "loss": 0.8384, "step": 1542 }, { "epoch": 0.24, "grad_norm": 19.081608419897385, "learning_rate": 1.9867601139439023e-05, "loss": 0.8873, "step": 1543 }, { "epoch": 0.24, "grad_norm": 24.779687846062483, "learning_rate": 1.9867327487891224e-05, "loss": 0.9797, "step": 1544 }, { "epoch": 0.24, "grad_norm": 24.391452958535215, "learning_rate": 1.9867053555722406e-05, "loss": 0.8472, "step": 1545 }, { "epoch": 0.24, "grad_norm": 24.469499017498077, "learning_rate": 1.9866779342940354e-05, "loss": 0.9956, "step": 1546 }, { "epoch": 0.24, "grad_norm": 21.310513414299503, "learning_rate": 1.986650484955287e-05, "loss": 0.9349, "step": 1547 }, { "epoch": 0.24, "grad_norm": 14.279051936045528, "learning_rate": 1.986623007556776e-05, "loss": 0.8944, "step": 1548 }, { "epoch": 0.24, "grad_norm": 18.777165605094, "learning_rate": 1.9865955020992837e-05, "loss": 0.9278, "step": 1549 }, { "epoch": 0.24, "grad_norm": 24.286095031490465, "learning_rate": 1.9865679685835924e-05, "loss": 0.8482, "step": 1550 }, { "epoch": 0.24, "grad_norm": 21.06437172765661, "learning_rate": 1.9865404070104857e-05, "loss": 0.8321, "step": 1551 }, { "epoch": 0.24, "grad_norm": 24.569456568911974, "learning_rate": 1.9865128173807465e-05, "loss": 0.9967, "step": 1552 }, { "epoch": 0.24, "grad_norm": 23.787221183987047, "learning_rate": 1.98648519969516e-05, "loss": 0.978, "step": 1553 }, { "epoch": 0.24, "grad_norm": 25.99072945819066, "learning_rate": 1.9864575539545115e-05, "loss": 0.8469, "step": 1554 }, { "epoch": 0.24, "grad_norm": 16.420302041685954, "learning_rate": 1.986429880159587e-05, "loss": 0.9421, "step": 1555 }, { "epoch": 0.24, "grad_norm": 27.393284233906577, "learning_rate": 1.986402178311174e-05, "loss": 0.9912, "step": 1556 }, { "epoch": 0.24, "grad_norm": 19.083219956199336, "learning_rate": 1.98637444841006e-05, "loss": 1.0749, "step": 1557 }, { "epoch": 0.24, "grad_norm": 24.371095586728668, "learning_rate": 1.9863466904570338e-05, "loss": 0.9149, "step": 1558 }, { "epoch": 0.24, "grad_norm": 18.69632820410229, "learning_rate": 1.9863189044528847e-05, "loss": 0.8413, "step": 1559 }, { "epoch": 0.24, "grad_norm": 19.669799542191065, "learning_rate": 1.986291090398403e-05, "loss": 0.9167, "step": 1560 }, { "epoch": 0.24, "grad_norm": 21.127502376430776, "learning_rate": 1.9862632482943796e-05, "loss": 0.9223, "step": 1561 }, { "epoch": 0.24, "grad_norm": 31.630367421031174, "learning_rate": 1.9862353781416062e-05, "loss": 0.8649, "step": 1562 }, { "epoch": 0.24, "grad_norm": 19.87502752758015, "learning_rate": 1.9862074799408757e-05, "loss": 0.8531, "step": 1563 }, { "epoch": 0.24, "grad_norm": 39.3331372197143, "learning_rate": 1.9861795536929814e-05, "loss": 0.9515, "step": 1564 }, { "epoch": 0.24, "grad_norm": 21.271972655549494, "learning_rate": 1.9861515993987174e-05, "loss": 0.8518, "step": 1565 }, { "epoch": 0.24, "grad_norm": 18.494631913935663, "learning_rate": 1.986123617058879e-05, "loss": 0.801, "step": 1566 }, { "epoch": 0.24, "grad_norm": 17.919234674150502, "learning_rate": 1.9860956066742613e-05, "loss": 1.007, "step": 1567 }, { "epoch": 0.24, "grad_norm": 24.858569040426747, "learning_rate": 1.986067568245662e-05, "loss": 0.8046, "step": 1568 }, { "epoch": 0.25, "grad_norm": 23.534347097329405, "learning_rate": 1.9860395017738778e-05, "loss": 0.82, "step": 1569 }, { "epoch": 0.25, "grad_norm": 22.73581702617474, "learning_rate": 1.9860114072597064e-05, "loss": 0.8114, "step": 1570 }, { "epoch": 0.25, "grad_norm": 22.757802254792043, "learning_rate": 1.985983284703948e-05, "loss": 0.8302, "step": 1571 }, { "epoch": 0.25, "grad_norm": 23.301606886146196, "learning_rate": 1.9859551341074016e-05, "loss": 0.9559, "step": 1572 }, { "epoch": 0.25, "grad_norm": 28.235940291350914, "learning_rate": 1.985926955470868e-05, "loss": 0.894, "step": 1573 }, { "epoch": 0.25, "grad_norm": 25.379316689200724, "learning_rate": 1.9858987487951485e-05, "loss": 0.9196, "step": 1574 }, { "epoch": 0.25, "grad_norm": 19.96398671967239, "learning_rate": 1.9858705140810453e-05, "loss": 0.9733, "step": 1575 }, { "epoch": 0.25, "grad_norm": 33.444085546192184, "learning_rate": 1.985842251329361e-05, "loss": 0.9762, "step": 1576 }, { "epoch": 0.25, "grad_norm": 22.0310834093762, "learning_rate": 1.9858139605409005e-05, "loss": 0.9813, "step": 1577 }, { "epoch": 0.25, "grad_norm": 28.530349919890558, "learning_rate": 1.985785641716467e-05, "loss": 0.9046, "step": 1578 }, { "epoch": 0.25, "grad_norm": 20.473781409203127, "learning_rate": 1.985757294856867e-05, "loss": 0.891, "step": 1579 }, { "epoch": 0.25, "grad_norm": 24.782652740656243, "learning_rate": 1.985728919962906e-05, "loss": 0.969, "step": 1580 }, { "epoch": 0.25, "grad_norm": 27.928155643095884, "learning_rate": 1.985700517035391e-05, "loss": 0.9972, "step": 1581 }, { "epoch": 0.25, "grad_norm": 15.171645769485439, "learning_rate": 1.98567208607513e-05, "loss": 0.779, "step": 1582 }, { "epoch": 0.25, "grad_norm": 16.15141236688541, "learning_rate": 1.9856436270829315e-05, "loss": 0.8682, "step": 1583 }, { "epoch": 0.25, "grad_norm": 30.34893832160513, "learning_rate": 1.985615140059605e-05, "loss": 0.8446, "step": 1584 }, { "epoch": 0.25, "grad_norm": 14.90170651294565, "learning_rate": 1.9855866250059604e-05, "loss": 0.9663, "step": 1585 }, { "epoch": 0.25, "grad_norm": 20.814766740364764, "learning_rate": 1.9855580819228087e-05, "loss": 0.9283, "step": 1586 }, { "epoch": 0.25, "grad_norm": 25.826641225339994, "learning_rate": 1.985529510810962e-05, "loss": 0.9141, "step": 1587 }, { "epoch": 0.25, "grad_norm": 24.461194642870954, "learning_rate": 1.9855009116712318e-05, "loss": 0.9014, "step": 1588 }, { "epoch": 0.25, "grad_norm": 16.67600278512863, "learning_rate": 1.985472284504433e-05, "loss": 0.9292, "step": 1589 }, { "epoch": 0.25, "grad_norm": 18.464522902281754, "learning_rate": 1.9854436293113782e-05, "loss": 0.9108, "step": 1590 }, { "epoch": 0.25, "grad_norm": 27.952304762485202, "learning_rate": 1.9854149460928832e-05, "loss": 0.7976, "step": 1591 }, { "epoch": 0.25, "grad_norm": 13.89888782104047, "learning_rate": 1.9853862348497638e-05, "loss": 0.8497, "step": 1592 }, { "epoch": 0.25, "grad_norm": 20.30094980944647, "learning_rate": 1.9853574955828364e-05, "loss": 0.919, "step": 1593 }, { "epoch": 0.25, "grad_norm": 29.379491346000943, "learning_rate": 1.985328728292918e-05, "loss": 0.8353, "step": 1594 }, { "epoch": 0.25, "grad_norm": 24.51259496756525, "learning_rate": 1.985299932980827e-05, "loss": 0.9382, "step": 1595 }, { "epoch": 0.25, "grad_norm": 19.055782837566728, "learning_rate": 1.9852711096473824e-05, "loss": 1.0492, "step": 1596 }, { "epoch": 0.25, "grad_norm": 17.092632275464698, "learning_rate": 1.9852422582934037e-05, "loss": 0.8701, "step": 1597 }, { "epoch": 0.25, "grad_norm": 14.837682850336902, "learning_rate": 1.9852133789197117e-05, "loss": 0.7924, "step": 1598 }, { "epoch": 0.25, "grad_norm": 24.70587231115513, "learning_rate": 1.9851844715271272e-05, "loss": 0.9711, "step": 1599 }, { "epoch": 0.25, "grad_norm": 26.768006772182794, "learning_rate": 1.985155536116473e-05, "loss": 0.913, "step": 1600 }, { "epoch": 0.25, "grad_norm": 18.889000930105173, "learning_rate": 1.9851265726885716e-05, "loss": 0.9164, "step": 1601 }, { "epoch": 0.25, "grad_norm": 29.56957839094889, "learning_rate": 1.9850975812442465e-05, "loss": 0.9676, "step": 1602 }, { "epoch": 0.25, "grad_norm": 20.131166672112233, "learning_rate": 1.9850685617843225e-05, "loss": 0.929, "step": 1603 }, { "epoch": 0.25, "grad_norm": 20.227839084474525, "learning_rate": 1.985039514309625e-05, "loss": 0.85, "step": 1604 }, { "epoch": 0.25, "grad_norm": 15.506373025882, "learning_rate": 1.9850104388209802e-05, "loss": 0.87, "step": 1605 }, { "epoch": 0.25, "grad_norm": 17.749856350545816, "learning_rate": 1.9849813353192142e-05, "loss": 0.7983, "step": 1606 }, { "epoch": 0.25, "grad_norm": 22.040139069914062, "learning_rate": 1.9849522038051553e-05, "loss": 0.8483, "step": 1607 }, { "epoch": 0.25, "grad_norm": 21.499774206716868, "learning_rate": 1.9849230442796324e-05, "loss": 0.9599, "step": 1608 }, { "epoch": 0.25, "grad_norm": 17.094841699068112, "learning_rate": 1.9848938567434737e-05, "loss": 0.9141, "step": 1609 }, { "epoch": 0.25, "grad_norm": 16.47469423698453, "learning_rate": 1.98486464119751e-05, "loss": 0.899, "step": 1610 }, { "epoch": 0.25, "grad_norm": 19.056744094228375, "learning_rate": 1.9848353976425723e-05, "loss": 0.876, "step": 1611 }, { "epoch": 0.25, "grad_norm": 16.517360578017563, "learning_rate": 1.984806126079492e-05, "loss": 0.8606, "step": 1612 }, { "epoch": 0.25, "grad_norm": 21.184089523301797, "learning_rate": 1.9847768265091015e-05, "loss": 0.8799, "step": 1613 }, { "epoch": 0.25, "grad_norm": 21.220808366496115, "learning_rate": 1.984747498932234e-05, "loss": 0.9153, "step": 1614 }, { "epoch": 0.25, "grad_norm": 20.191952781195642, "learning_rate": 1.984718143349724e-05, "loss": 0.9018, "step": 1615 }, { "epoch": 0.25, "grad_norm": 20.696585650972292, "learning_rate": 1.9846887597624054e-05, "loss": 0.8602, "step": 1616 }, { "epoch": 0.25, "grad_norm": 19.645189871587714, "learning_rate": 1.984659348171115e-05, "loss": 0.9109, "step": 1617 }, { "epoch": 0.25, "grad_norm": 21.80176946317184, "learning_rate": 1.9846299085766888e-05, "loss": 0.7923, "step": 1618 }, { "epoch": 0.25, "grad_norm": 12.40837045415924, "learning_rate": 1.9846004409799636e-05, "loss": 0.8333, "step": 1619 }, { "epoch": 0.25, "grad_norm": 18.21026819611962, "learning_rate": 1.984570945381778e-05, "loss": 0.9582, "step": 1620 }, { "epoch": 0.25, "grad_norm": 21.057845668281168, "learning_rate": 1.9845414217829708e-05, "loss": 0.8639, "step": 1621 }, { "epoch": 0.25, "grad_norm": 16.807790616028218, "learning_rate": 1.9845118701843816e-05, "loss": 0.8879, "step": 1622 }, { "epoch": 0.25, "grad_norm": 20.871412538595905, "learning_rate": 1.9844822905868507e-05, "loss": 0.9686, "step": 1623 }, { "epoch": 0.25, "grad_norm": 21.205071110481367, "learning_rate": 1.984452682991219e-05, "loss": 0.9883, "step": 1624 }, { "epoch": 0.25, "grad_norm": 19.853528600595627, "learning_rate": 1.984423047398329e-05, "loss": 0.9944, "step": 1625 }, { "epoch": 0.25, "grad_norm": 17.91780237453702, "learning_rate": 1.9843933838090236e-05, "loss": 0.8593, "step": 1626 }, { "epoch": 0.25, "grad_norm": 21.217180569043126, "learning_rate": 1.9843636922241463e-05, "loss": 0.9105, "step": 1627 }, { "epoch": 0.25, "grad_norm": 13.632563895879928, "learning_rate": 1.984333972644541e-05, "loss": 0.818, "step": 1628 }, { "epoch": 0.25, "grad_norm": 18.235173435921375, "learning_rate": 1.984304225071054e-05, "loss": 0.8356, "step": 1629 }, { "epoch": 0.25, "grad_norm": 20.182657224892584, "learning_rate": 1.98427444950453e-05, "loss": 0.7999, "step": 1630 }, { "epoch": 0.25, "grad_norm": 15.394914966399828, "learning_rate": 1.9842446459458166e-05, "loss": 0.8587, "step": 1631 }, { "epoch": 0.25, "grad_norm": 18.2234754879083, "learning_rate": 1.9842148143957612e-05, "loss": 0.9409, "step": 1632 }, { "epoch": 0.26, "grad_norm": 20.972849386188827, "learning_rate": 1.9841849548552124e-05, "loss": 0.8701, "step": 1633 }, { "epoch": 0.26, "grad_norm": 20.238464828067983, "learning_rate": 1.9841550673250192e-05, "loss": 0.8618, "step": 1634 }, { "epoch": 0.26, "grad_norm": 21.860763723830463, "learning_rate": 1.9841251518060314e-05, "loss": 0.9003, "step": 1635 }, { "epoch": 0.26, "grad_norm": 16.63649729993102, "learning_rate": 1.9840952082991e-05, "loss": 0.8897, "step": 1636 }, { "epoch": 0.26, "grad_norm": 15.24084487468591, "learning_rate": 1.984065236805077e-05, "loss": 0.8665, "step": 1637 }, { "epoch": 0.26, "grad_norm": 18.135258355377267, "learning_rate": 1.984035237324814e-05, "loss": 0.8578, "step": 1638 }, { "epoch": 0.26, "grad_norm": 28.447627242906936, "learning_rate": 1.9840052098591645e-05, "loss": 0.9812, "step": 1639 }, { "epoch": 0.26, "grad_norm": 17.872861871021367, "learning_rate": 1.9839751544089827e-05, "loss": 0.8685, "step": 1640 }, { "epoch": 0.26, "grad_norm": 7.810502189781224, "learning_rate": 1.983945070975123e-05, "loss": 0.834, "step": 1641 }, { "epoch": 0.26, "grad_norm": 19.15290131322366, "learning_rate": 1.983914959558441e-05, "loss": 0.9041, "step": 1642 }, { "epoch": 0.26, "grad_norm": 16.15181768591629, "learning_rate": 1.983884820159793e-05, "loss": 0.878, "step": 1643 }, { "epoch": 0.26, "grad_norm": 20.195868435407395, "learning_rate": 1.9838546527800364e-05, "loss": 0.8713, "step": 1644 }, { "epoch": 0.26, "grad_norm": 34.692844200851354, "learning_rate": 1.983824457420029e-05, "loss": 0.984, "step": 1645 }, { "epoch": 0.26, "grad_norm": 18.70194492425561, "learning_rate": 1.983794234080629e-05, "loss": 0.9004, "step": 1646 }, { "epoch": 0.26, "grad_norm": 18.841944503640804, "learning_rate": 1.9837639827626973e-05, "loss": 0.8657, "step": 1647 }, { "epoch": 0.26, "grad_norm": 29.375172863613965, "learning_rate": 1.983733703467093e-05, "loss": 0.9512, "step": 1648 }, { "epoch": 0.26, "grad_norm": 25.117005789234447, "learning_rate": 1.983703396194678e-05, "loss": 0.9423, "step": 1649 }, { "epoch": 0.26, "grad_norm": 22.010961772423034, "learning_rate": 1.9836730609463134e-05, "loss": 0.8295, "step": 1650 }, { "epoch": 0.26, "grad_norm": 17.48029181875695, "learning_rate": 1.983642697722863e-05, "loss": 0.8895, "step": 1651 }, { "epoch": 0.26, "grad_norm": 18.08662586900136, "learning_rate": 1.9836123065251892e-05, "loss": 0.8811, "step": 1652 }, { "epoch": 0.26, "grad_norm": 14.468787104680578, "learning_rate": 1.9835818873541572e-05, "loss": 0.8907, "step": 1653 }, { "epoch": 0.26, "grad_norm": 42.75017693549901, "learning_rate": 1.983551440210631e-05, "loss": 1.0122, "step": 1654 }, { "epoch": 0.26, "grad_norm": 37.603883522249845, "learning_rate": 1.9835209650954775e-05, "loss": 1.016, "step": 1655 }, { "epoch": 0.26, "grad_norm": 16.902028193039648, "learning_rate": 1.9834904620095636e-05, "loss": 0.8202, "step": 1656 }, { "epoch": 0.26, "grad_norm": 20.05601838766581, "learning_rate": 1.9834599309537557e-05, "loss": 0.8901, "step": 1657 }, { "epoch": 0.26, "grad_norm": 21.000108851990827, "learning_rate": 1.9834293719289227e-05, "loss": 0.9947, "step": 1658 }, { "epoch": 0.26, "grad_norm": 52.55156834831746, "learning_rate": 1.9833987849359338e-05, "loss": 1.0453, "step": 1659 }, { "epoch": 0.26, "grad_norm": 18.137705460376477, "learning_rate": 1.9833681699756592e-05, "loss": 0.8133, "step": 1660 }, { "epoch": 0.26, "grad_norm": 24.64539404819129, "learning_rate": 1.9833375270489682e-05, "loss": 0.9252, "step": 1661 }, { "epoch": 0.26, "grad_norm": 29.720110765826092, "learning_rate": 1.983306856156734e-05, "loss": 0.8509, "step": 1662 }, { "epoch": 0.26, "grad_norm": 10.061173508564595, "learning_rate": 1.9832761572998274e-05, "loss": 0.8002, "step": 1663 }, { "epoch": 0.26, "grad_norm": 19.487527124277147, "learning_rate": 1.9832454304791224e-05, "loss": 0.9104, "step": 1664 }, { "epoch": 0.26, "grad_norm": 19.2215853506947, "learning_rate": 1.9832146756954925e-05, "loss": 0.9163, "step": 1665 }, { "epoch": 0.26, "grad_norm": 16.120777669608227, "learning_rate": 1.9831838929498123e-05, "loss": 0.855, "step": 1666 }, { "epoch": 0.26, "grad_norm": 33.39743699387739, "learning_rate": 1.983153082242957e-05, "loss": 0.8794, "step": 1667 }, { "epoch": 0.26, "grad_norm": 16.19784093810489, "learning_rate": 1.983122243575804e-05, "loss": 0.7899, "step": 1668 }, { "epoch": 0.26, "grad_norm": 19.899923464553627, "learning_rate": 1.9830913769492288e-05, "loss": 0.896, "step": 1669 }, { "epoch": 0.26, "grad_norm": 17.8511313622163, "learning_rate": 1.9830604823641103e-05, "loss": 0.985, "step": 1670 }, { "epoch": 0.26, "grad_norm": 31.44816884401661, "learning_rate": 1.9830295598213263e-05, "loss": 0.8494, "step": 1671 }, { "epoch": 0.26, "grad_norm": 21.58686807623665, "learning_rate": 1.9829986093217575e-05, "loss": 0.9452, "step": 1672 }, { "epoch": 0.26, "grad_norm": 29.097783068819563, "learning_rate": 1.9829676308662826e-05, "loss": 0.9155, "step": 1673 }, { "epoch": 0.26, "grad_norm": 15.029132243761168, "learning_rate": 1.9829366244557837e-05, "loss": 0.8357, "step": 1674 }, { "epoch": 0.26, "grad_norm": 26.770385605535186, "learning_rate": 1.982905590091142e-05, "loss": 0.8228, "step": 1675 }, { "epoch": 0.26, "grad_norm": 16.883739146354582, "learning_rate": 1.98287452777324e-05, "loss": 0.9237, "step": 1676 }, { "epoch": 0.26, "grad_norm": 18.522797857862518, "learning_rate": 1.982843437502962e-05, "loss": 0.9413, "step": 1677 }, { "epoch": 0.26, "grad_norm": 15.670813278734217, "learning_rate": 1.982812319281191e-05, "loss": 0.7638, "step": 1678 }, { "epoch": 0.26, "grad_norm": 22.824570396700686, "learning_rate": 1.982781173108813e-05, "loss": 0.8331, "step": 1679 }, { "epoch": 0.26, "grad_norm": 6.102055059619059, "learning_rate": 1.982749998986713e-05, "loss": 0.9727, "step": 1680 }, { "epoch": 0.26, "grad_norm": 18.613612673391664, "learning_rate": 1.982718796915778e-05, "loss": 0.9464, "step": 1681 }, { "epoch": 0.26, "grad_norm": 21.58938208456188, "learning_rate": 1.9826875668968956e-05, "loss": 0.8991, "step": 1682 }, { "epoch": 0.26, "grad_norm": 13.565050239136012, "learning_rate": 1.9826563089309534e-05, "loss": 0.8076, "step": 1683 }, { "epoch": 0.26, "grad_norm": 33.96244031969681, "learning_rate": 1.982625023018841e-05, "loss": 1.0186, "step": 1684 }, { "epoch": 0.26, "grad_norm": 18.330985058321044, "learning_rate": 1.982593709161447e-05, "loss": 0.8724, "step": 1685 }, { "epoch": 0.26, "grad_norm": 16.11666479201521, "learning_rate": 1.9825623673596634e-05, "loss": 0.7933, "step": 1686 }, { "epoch": 0.26, "grad_norm": 15.77564466044492, "learning_rate": 1.9825309976143807e-05, "loss": 0.897, "step": 1687 }, { "epoch": 0.26, "grad_norm": 19.416525951954807, "learning_rate": 1.982499599926491e-05, "loss": 0.798, "step": 1688 }, { "epoch": 0.26, "grad_norm": 26.160292368889735, "learning_rate": 1.9824681742968876e-05, "loss": 0.812, "step": 1689 }, { "epoch": 0.26, "grad_norm": 34.88096240990854, "learning_rate": 1.9824367207264638e-05, "loss": 0.9091, "step": 1690 }, { "epoch": 0.26, "grad_norm": 36.20317577149408, "learning_rate": 1.9824052392161148e-05, "loss": 0.9105, "step": 1691 }, { "epoch": 0.26, "grad_norm": 20.68920813017233, "learning_rate": 1.982373729766735e-05, "loss": 0.8965, "step": 1692 }, { "epoch": 0.26, "grad_norm": 21.33353376344843, "learning_rate": 1.9823421923792213e-05, "loss": 0.9168, "step": 1693 }, { "epoch": 0.26, "grad_norm": 25.1289374411941, "learning_rate": 1.98231062705447e-05, "loss": 0.9557, "step": 1694 }, { "epoch": 0.26, "grad_norm": 22.92384134919337, "learning_rate": 1.9822790337933795e-05, "loss": 0.9023, "step": 1695 }, { "epoch": 0.26, "grad_norm": 18.142957059410243, "learning_rate": 1.9822474125968478e-05, "loss": 0.7806, "step": 1696 }, { "epoch": 0.27, "grad_norm": 18.00885413543742, "learning_rate": 1.9822157634657742e-05, "loss": 0.8202, "step": 1697 }, { "epoch": 0.27, "grad_norm": 20.582682493009642, "learning_rate": 1.9821840864010588e-05, "loss": 0.8614, "step": 1698 }, { "epoch": 0.27, "grad_norm": 17.35402551523379, "learning_rate": 1.9821523814036026e-05, "loss": 0.8824, "step": 1699 }, { "epoch": 0.27, "grad_norm": 16.520273603716277, "learning_rate": 1.9821206484743073e-05, "loss": 0.8615, "step": 1700 }, { "epoch": 0.27, "grad_norm": 12.00684525162505, "learning_rate": 1.9820888876140752e-05, "loss": 0.8243, "step": 1701 }, { "epoch": 0.27, "grad_norm": 20.61042011017325, "learning_rate": 1.9820570988238096e-05, "loss": 0.8826, "step": 1702 }, { "epoch": 0.27, "grad_norm": 21.07278778761485, "learning_rate": 1.982025282104415e-05, "loss": 0.7674, "step": 1703 }, { "epoch": 0.27, "grad_norm": 14.739100708209826, "learning_rate": 1.981993437456795e-05, "loss": 0.8416, "step": 1704 }, { "epoch": 0.27, "grad_norm": 17.387826597579927, "learning_rate": 1.9819615648818568e-05, "loss": 0.8929, "step": 1705 }, { "epoch": 0.27, "grad_norm": 24.54800838449656, "learning_rate": 1.9819296643805058e-05, "loss": 0.947, "step": 1706 }, { "epoch": 0.27, "grad_norm": 27.16360430129431, "learning_rate": 1.98189773595365e-05, "loss": 0.9132, "step": 1707 }, { "epoch": 0.27, "grad_norm": 21.937256023756593, "learning_rate": 1.9818657796021963e-05, "loss": 0.8758, "step": 1708 }, { "epoch": 0.27, "grad_norm": 21.491781477037854, "learning_rate": 1.9818337953270544e-05, "loss": 0.9212, "step": 1709 }, { "epoch": 0.27, "grad_norm": 22.04077979195327, "learning_rate": 1.981801783129134e-05, "loss": 0.8816, "step": 1710 }, { "epoch": 0.27, "grad_norm": 18.636449283785787, "learning_rate": 1.981769743009345e-05, "loss": 0.8137, "step": 1711 }, { "epoch": 0.27, "grad_norm": 29.107738028061473, "learning_rate": 1.9817376749685994e-05, "loss": 0.8802, "step": 1712 }, { "epoch": 0.27, "grad_norm": 26.23181844435767, "learning_rate": 1.981705579007808e-05, "loss": 1.1021, "step": 1713 }, { "epoch": 0.27, "grad_norm": 17.399339037603546, "learning_rate": 1.9816734551278843e-05, "loss": 0.8471, "step": 1714 }, { "epoch": 0.27, "grad_norm": 16.343173807355683, "learning_rate": 1.981641303329742e-05, "loss": 0.938, "step": 1715 }, { "epoch": 0.27, "grad_norm": 19.64286212425664, "learning_rate": 1.981609123614295e-05, "loss": 0.9954, "step": 1716 }, { "epoch": 0.27, "grad_norm": 30.577373505150504, "learning_rate": 1.981576915982459e-05, "loss": 0.9168, "step": 1717 }, { "epoch": 0.27, "grad_norm": 22.085219672148874, "learning_rate": 1.9815446804351495e-05, "loss": 0.8749, "step": 1718 }, { "epoch": 0.27, "grad_norm": 20.99618885817167, "learning_rate": 1.9815124169732834e-05, "loss": 0.8716, "step": 1719 }, { "epoch": 0.27, "grad_norm": 24.651759864107493, "learning_rate": 1.9814801255977786e-05, "loss": 0.9354, "step": 1720 }, { "epoch": 0.27, "grad_norm": 23.447190975021126, "learning_rate": 1.9814478063095533e-05, "loss": 1.0228, "step": 1721 }, { "epoch": 0.27, "grad_norm": 27.748623575943913, "learning_rate": 1.9814154591095262e-05, "loss": 0.8155, "step": 1722 }, { "epoch": 0.27, "grad_norm": 15.932456202019589, "learning_rate": 1.9813830839986177e-05, "loss": 0.8719, "step": 1723 }, { "epoch": 0.27, "grad_norm": 16.80784905725702, "learning_rate": 1.9813506809777483e-05, "loss": 0.9, "step": 1724 }, { "epoch": 0.27, "grad_norm": 28.68492081003693, "learning_rate": 1.9813182500478395e-05, "loss": 0.9751, "step": 1725 }, { "epoch": 0.27, "grad_norm": 18.15240916369025, "learning_rate": 1.981285791209814e-05, "loss": 0.8592, "step": 1726 }, { "epoch": 0.27, "grad_norm": 28.619245552542825, "learning_rate": 1.9812533044645945e-05, "loss": 0.9925, "step": 1727 }, { "epoch": 0.27, "grad_norm": 32.54050154142809, "learning_rate": 1.981220789813105e-05, "loss": 0.9115, "step": 1728 }, { "epoch": 0.27, "grad_norm": 18.369549080638826, "learning_rate": 1.9811882472562703e-05, "loss": 0.9084, "step": 1729 }, { "epoch": 0.27, "grad_norm": 28.557472124537792, "learning_rate": 1.9811556767950158e-05, "loss": 0.9147, "step": 1730 }, { "epoch": 0.27, "grad_norm": 20.197237848997407, "learning_rate": 1.981123078430268e-05, "loss": 0.8611, "step": 1731 }, { "epoch": 0.27, "grad_norm": 23.070902991884388, "learning_rate": 1.9810904521629534e-05, "loss": 0.813, "step": 1732 }, { "epoch": 0.27, "grad_norm": 20.363541672221512, "learning_rate": 1.9810577979940006e-05, "loss": 0.8972, "step": 1733 }, { "epoch": 0.27, "grad_norm": 20.100392094047965, "learning_rate": 1.9810251159243378e-05, "loss": 0.9072, "step": 1734 }, { "epoch": 0.27, "grad_norm": 23.17521346846764, "learning_rate": 1.9809924059548944e-05, "loss": 0.8894, "step": 1735 }, { "epoch": 0.27, "grad_norm": 14.59224159932445, "learning_rate": 1.980959668086601e-05, "loss": 0.9422, "step": 1736 }, { "epoch": 0.27, "grad_norm": 26.431800013074078, "learning_rate": 1.9809269023203887e-05, "loss": 0.8817, "step": 1737 }, { "epoch": 0.27, "grad_norm": 21.204420327753652, "learning_rate": 1.9808941086571888e-05, "loss": 0.8591, "step": 1738 }, { "epoch": 0.27, "grad_norm": 24.24064240124899, "learning_rate": 1.9808612870979346e-05, "loss": 0.8312, "step": 1739 }, { "epoch": 0.27, "grad_norm": 16.829441547252312, "learning_rate": 1.9808284376435588e-05, "loss": 0.7915, "step": 1740 }, { "epoch": 0.27, "grad_norm": 16.05777680683614, "learning_rate": 1.9807955602949962e-05, "loss": 0.8875, "step": 1741 }, { "epoch": 0.27, "grad_norm": 28.82942523608552, "learning_rate": 1.9807626550531816e-05, "loss": 0.9477, "step": 1742 }, { "epoch": 0.27, "grad_norm": 13.526141031429328, "learning_rate": 1.9807297219190508e-05, "loss": 0.865, "step": 1743 }, { "epoch": 0.27, "grad_norm": 21.71654978111612, "learning_rate": 1.9806967608935403e-05, "loss": 0.8723, "step": 1744 }, { "epoch": 0.27, "grad_norm": 20.122209608729346, "learning_rate": 1.980663771977588e-05, "loss": 0.8894, "step": 1745 }, { "epoch": 0.27, "grad_norm": 18.36792613612658, "learning_rate": 1.9806307551721313e-05, "loss": 0.9009, "step": 1746 }, { "epoch": 0.27, "grad_norm": 15.668873128133317, "learning_rate": 1.9805977104781096e-05, "loss": 0.858, "step": 1747 }, { "epoch": 0.27, "grad_norm": 21.70445764376114, "learning_rate": 1.9805646378964626e-05, "loss": 0.8944, "step": 1748 }, { "epoch": 0.27, "grad_norm": 24.62086584393321, "learning_rate": 1.9805315374281314e-05, "loss": 0.8313, "step": 1749 }, { "epoch": 0.27, "grad_norm": 26.815314747600738, "learning_rate": 1.9804984090740563e-05, "loss": 0.9241, "step": 1750 }, { "epoch": 0.27, "grad_norm": 31.07602689554703, "learning_rate": 1.9804652528351806e-05, "loss": 0.8719, "step": 1751 }, { "epoch": 0.27, "grad_norm": 4.43906421023547, "learning_rate": 1.980432068712446e-05, "loss": 0.9143, "step": 1752 }, { "epoch": 0.27, "grad_norm": 17.522436014708436, "learning_rate": 1.9803988567067975e-05, "loss": 0.9915, "step": 1753 }, { "epoch": 0.27, "grad_norm": 22.50613093822696, "learning_rate": 1.9803656168191786e-05, "loss": 0.901, "step": 1754 }, { "epoch": 0.27, "grad_norm": 17.236064247078225, "learning_rate": 1.9803323490505352e-05, "loss": 0.8827, "step": 1755 }, { "epoch": 0.27, "grad_norm": 17.635590551628425, "learning_rate": 1.9802990534018135e-05, "loss": 0.8642, "step": 1756 }, { "epoch": 0.27, "grad_norm": 18.513616355557605, "learning_rate": 1.9802657298739598e-05, "loss": 0.9179, "step": 1757 }, { "epoch": 0.27, "grad_norm": 14.952587424952117, "learning_rate": 1.9802323784679225e-05, "loss": 0.8081, "step": 1758 }, { "epoch": 0.27, "grad_norm": 34.995843557082345, "learning_rate": 1.9801989991846495e-05, "loss": 0.9588, "step": 1759 }, { "epoch": 0.27, "grad_norm": 18.433058171333993, "learning_rate": 1.9801655920250903e-05, "loss": 0.9047, "step": 1760 }, { "epoch": 0.28, "grad_norm": 27.19085417697937, "learning_rate": 1.9801321569901955e-05, "loss": 0.8589, "step": 1761 }, { "epoch": 0.28, "grad_norm": 11.490056017876276, "learning_rate": 1.980098694080915e-05, "loss": 0.7968, "step": 1762 }, { "epoch": 0.28, "grad_norm": 10.330236244397957, "learning_rate": 1.9800652032982013e-05, "loss": 1.0385, "step": 1763 }, { "epoch": 0.28, "grad_norm": 24.55127119426707, "learning_rate": 1.9800316846430064e-05, "loss": 0.9348, "step": 1764 }, { "epoch": 0.28, "grad_norm": 21.18259629186154, "learning_rate": 1.979998138116284e-05, "loss": 0.9195, "step": 1765 }, { "epoch": 0.28, "grad_norm": 20.289223537753408, "learning_rate": 1.9799645637189875e-05, "loss": 0.9477, "step": 1766 }, { "epoch": 0.28, "grad_norm": 14.693346282413629, "learning_rate": 1.979930961452072e-05, "loss": 0.783, "step": 1767 }, { "epoch": 0.28, "grad_norm": 25.326541681708854, "learning_rate": 1.9798973313164934e-05, "loss": 0.881, "step": 1768 }, { "epoch": 0.28, "grad_norm": 31.889067523030125, "learning_rate": 1.979863673313208e-05, "loss": 0.9513, "step": 1769 }, { "epoch": 0.28, "grad_norm": 21.741223782349362, "learning_rate": 1.979829987443173e-05, "loss": 0.881, "step": 1770 }, { "epoch": 0.28, "grad_norm": 15.742427767299679, "learning_rate": 1.9797962737073456e-05, "loss": 0.8532, "step": 1771 }, { "epoch": 0.28, "grad_norm": 26.50449467949628, "learning_rate": 1.9797625321066863e-05, "loss": 1.0252, "step": 1772 }, { "epoch": 0.28, "grad_norm": 21.619867725984435, "learning_rate": 1.9797287626421534e-05, "loss": 0.8524, "step": 1773 }, { "epoch": 0.28, "grad_norm": 26.785282296602777, "learning_rate": 1.9796949653147073e-05, "loss": 0.8381, "step": 1774 }, { "epoch": 0.28, "grad_norm": 19.71690646061289, "learning_rate": 1.97966114012531e-05, "loss": 0.9848, "step": 1775 }, { "epoch": 0.28, "grad_norm": 14.441185843013914, "learning_rate": 1.9796272870749225e-05, "loss": 0.8482, "step": 1776 }, { "epoch": 0.28, "grad_norm": 19.0942923208159, "learning_rate": 1.9795934061645084e-05, "loss": 0.8812, "step": 1777 }, { "epoch": 0.28, "grad_norm": 12.787196562122961, "learning_rate": 1.9795594973950305e-05, "loss": 0.9134, "step": 1778 }, { "epoch": 0.28, "grad_norm": 25.1654330487038, "learning_rate": 1.979525560767454e-05, "loss": 0.9075, "step": 1779 }, { "epoch": 0.28, "grad_norm": 16.38385296156733, "learning_rate": 1.9794915962827427e-05, "loss": 0.7823, "step": 1780 }, { "epoch": 0.28, "grad_norm": 22.93106760878871, "learning_rate": 1.9794576039418643e-05, "loss": 0.8569, "step": 1781 }, { "epoch": 0.28, "grad_norm": 17.028411733726802, "learning_rate": 1.979423583745784e-05, "loss": 0.8344, "step": 1782 }, { "epoch": 0.28, "grad_norm": 19.801500937619682, "learning_rate": 1.9793895356954704e-05, "loss": 0.8412, "step": 1783 }, { "epoch": 0.28, "grad_norm": 21.849405129691057, "learning_rate": 1.9793554597918908e-05, "loss": 0.9314, "step": 1784 }, { "epoch": 0.28, "grad_norm": 19.8297964628564, "learning_rate": 1.9793213560360153e-05, "loss": 0.8966, "step": 1785 }, { "epoch": 0.28, "grad_norm": 19.357410765379026, "learning_rate": 1.9792872244288132e-05, "loss": 0.8239, "step": 1786 }, { "epoch": 0.28, "grad_norm": 17.137299185445123, "learning_rate": 1.9792530649712552e-05, "loss": 0.8157, "step": 1787 }, { "epoch": 0.28, "grad_norm": 15.96843058682852, "learning_rate": 1.979218877664313e-05, "loss": 0.8441, "step": 1788 }, { "epoch": 0.28, "grad_norm": 22.4538187763964, "learning_rate": 1.9791846625089588e-05, "loss": 0.7619, "step": 1789 }, { "epoch": 0.28, "grad_norm": 29.816475872016984, "learning_rate": 1.9791504195061655e-05, "loss": 0.8856, "step": 1790 }, { "epoch": 0.28, "grad_norm": 19.022937439097497, "learning_rate": 1.9791161486569072e-05, "loss": 0.9113, "step": 1791 }, { "epoch": 0.28, "grad_norm": 24.92024545155266, "learning_rate": 1.979081849962158e-05, "loss": 0.984, "step": 1792 }, { "epoch": 0.28, "grad_norm": 15.34857719844794, "learning_rate": 1.979047523422894e-05, "loss": 0.8896, "step": 1793 }, { "epoch": 0.28, "grad_norm": 23.357809398328623, "learning_rate": 1.9790131690400913e-05, "loss": 0.9505, "step": 1794 }, { "epoch": 0.28, "grad_norm": 13.204969814461755, "learning_rate": 1.9789787868147265e-05, "loss": 0.7692, "step": 1795 }, { "epoch": 0.28, "grad_norm": 19.921512275834388, "learning_rate": 1.9789443767477778e-05, "loss": 0.9257, "step": 1796 }, { "epoch": 0.28, "grad_norm": 21.092290704193367, "learning_rate": 1.978909938840224e-05, "loss": 0.8839, "step": 1797 }, { "epoch": 0.28, "grad_norm": 21.555757245780182, "learning_rate": 1.9788754730930436e-05, "loss": 1.0024, "step": 1798 }, { "epoch": 0.28, "grad_norm": 16.588832850024062, "learning_rate": 1.9788409795072175e-05, "loss": 0.8748, "step": 1799 }, { "epoch": 0.28, "grad_norm": 17.08744310269793, "learning_rate": 1.978806458083727e-05, "loss": 0.9688, "step": 1800 }, { "epoch": 0.28, "grad_norm": 13.231640853636327, "learning_rate": 1.978771908823553e-05, "loss": 0.8112, "step": 1801 }, { "epoch": 0.28, "grad_norm": 22.850053553523697, "learning_rate": 1.978737331727679e-05, "loss": 0.9087, "step": 1802 }, { "epoch": 0.28, "grad_norm": 24.854705950518575, "learning_rate": 1.9787027267970873e-05, "loss": 0.8585, "step": 1803 }, { "epoch": 0.28, "grad_norm": 26.570283879099517, "learning_rate": 1.9786680940327626e-05, "loss": 0.9648, "step": 1804 }, { "epoch": 0.28, "grad_norm": 24.773776774758392, "learning_rate": 1.97863343343569e-05, "loss": 0.762, "step": 1805 }, { "epoch": 0.28, "grad_norm": 21.047349295762505, "learning_rate": 1.978598745006855e-05, "loss": 0.8808, "step": 1806 }, { "epoch": 0.28, "grad_norm": 22.56834834060168, "learning_rate": 1.978564028747244e-05, "loss": 0.889, "step": 1807 }, { "epoch": 0.28, "grad_norm": 23.922200495363604, "learning_rate": 1.9785292846578446e-05, "loss": 0.8565, "step": 1808 }, { "epoch": 0.28, "grad_norm": 27.401746957361137, "learning_rate": 1.9784945127396446e-05, "loss": 0.9528, "step": 1809 }, { "epoch": 0.28, "grad_norm": 14.761854402797546, "learning_rate": 1.9784597129936333e-05, "loss": 0.753, "step": 1810 }, { "epoch": 0.28, "grad_norm": 14.52069704734337, "learning_rate": 1.9784248854208e-05, "loss": 0.9209, "step": 1811 }, { "epoch": 0.28, "grad_norm": 10.558703475252177, "learning_rate": 1.9783900300221354e-05, "loss": 0.7836, "step": 1812 }, { "epoch": 0.28, "grad_norm": 20.251583693824134, "learning_rate": 1.9783551467986307e-05, "loss": 0.873, "step": 1813 }, { "epoch": 0.28, "grad_norm": 26.622178118444687, "learning_rate": 1.978320235751278e-05, "loss": 0.9453, "step": 1814 }, { "epoch": 0.28, "grad_norm": 19.12187551550468, "learning_rate": 1.97828529688107e-05, "loss": 0.9581, "step": 1815 }, { "epoch": 0.28, "grad_norm": 23.286965899324226, "learning_rate": 1.9782503301890004e-05, "loss": 0.9024, "step": 1816 }, { "epoch": 0.28, "grad_norm": 22.827516362290183, "learning_rate": 1.9782153356760635e-05, "loss": 0.8556, "step": 1817 }, { "epoch": 0.28, "grad_norm": 21.34215166787253, "learning_rate": 1.978180313343255e-05, "loss": 0.8911, "step": 1818 }, { "epoch": 0.28, "grad_norm": 18.600726753311086, "learning_rate": 1.9781452631915705e-05, "loss": 0.8825, "step": 1819 }, { "epoch": 0.28, "grad_norm": 13.253950137496588, "learning_rate": 1.9781101852220074e-05, "loss": 0.8998, "step": 1820 }, { "epoch": 0.28, "grad_norm": 22.370409260145085, "learning_rate": 1.978075079435562e-05, "loss": 0.9154, "step": 1821 }, { "epoch": 0.28, "grad_norm": 19.678942426605634, "learning_rate": 1.978039945833234e-05, "loss": 0.9338, "step": 1822 }, { "epoch": 0.28, "grad_norm": 17.99321255846877, "learning_rate": 1.978004784416022e-05, "loss": 0.9495, "step": 1823 }, { "epoch": 0.28, "grad_norm": 15.751243995365353, "learning_rate": 1.977969595184926e-05, "loss": 0.895, "step": 1824 }, { "epoch": 0.29, "grad_norm": 17.825625049592, "learning_rate": 1.9779343781409466e-05, "loss": 0.8204, "step": 1825 }, { "epoch": 0.29, "grad_norm": 20.513645984582833, "learning_rate": 1.9778991332850857e-05, "loss": 0.8813, "step": 1826 }, { "epoch": 0.29, "grad_norm": 20.849047481639463, "learning_rate": 1.9778638606183453e-05, "loss": 1.0108, "step": 1827 }, { "epoch": 0.29, "grad_norm": 15.838968960727696, "learning_rate": 1.977828560141729e-05, "loss": 0.8283, "step": 1828 }, { "epoch": 0.29, "grad_norm": 24.336703018712903, "learning_rate": 1.9777932318562407e-05, "loss": 0.9114, "step": 1829 }, { "epoch": 0.29, "grad_norm": 23.559767149740388, "learning_rate": 1.9777578757628845e-05, "loss": 0.9584, "step": 1830 }, { "epoch": 0.29, "grad_norm": 18.806822963396797, "learning_rate": 1.9777224918626662e-05, "loss": 0.8323, "step": 1831 }, { "epoch": 0.29, "grad_norm": 19.578962190177567, "learning_rate": 1.9776870801565924e-05, "loss": 0.9118, "step": 1832 }, { "epoch": 0.29, "grad_norm": 23.522935496446518, "learning_rate": 1.97765164064567e-05, "loss": 0.8278, "step": 1833 }, { "epoch": 0.29, "grad_norm": 19.73905184759165, "learning_rate": 1.977616173330907e-05, "loss": 0.8432, "step": 1834 }, { "epoch": 0.29, "grad_norm": 15.574097636000438, "learning_rate": 1.9775806782133114e-05, "loss": 0.7777, "step": 1835 }, { "epoch": 0.29, "grad_norm": 32.37142113747569, "learning_rate": 1.9775451552938937e-05, "loss": 0.9256, "step": 1836 }, { "epoch": 0.29, "grad_norm": 23.42497999339028, "learning_rate": 1.9775096045736634e-05, "loss": 0.8323, "step": 1837 }, { "epoch": 0.29, "grad_norm": 21.867717619627314, "learning_rate": 1.9774740260536318e-05, "loss": 0.7863, "step": 1838 }, { "epoch": 0.29, "grad_norm": 17.43668407151477, "learning_rate": 1.977438419734811e-05, "loss": 0.8583, "step": 1839 }, { "epoch": 0.29, "grad_norm": 23.98656454871355, "learning_rate": 1.977402785618213e-05, "loss": 0.9369, "step": 1840 }, { "epoch": 0.29, "grad_norm": 18.480920694582508, "learning_rate": 1.9773671237048514e-05, "loss": 0.8054, "step": 1841 }, { "epoch": 0.29, "grad_norm": 26.789256845271577, "learning_rate": 1.977331433995741e-05, "loss": 0.9995, "step": 1842 }, { "epoch": 0.29, "grad_norm": 24.08763435503969, "learning_rate": 1.977295716491896e-05, "loss": 0.8417, "step": 1843 }, { "epoch": 0.29, "grad_norm": 42.645517418034, "learning_rate": 1.9772599711943323e-05, "loss": 0.9394, "step": 1844 }, { "epoch": 0.29, "grad_norm": 17.063569339318114, "learning_rate": 1.977224198104067e-05, "loss": 0.864, "step": 1845 }, { "epoch": 0.29, "grad_norm": 21.15715535845419, "learning_rate": 1.9771883972221174e-05, "loss": 0.8319, "step": 1846 }, { "epoch": 0.29, "grad_norm": 27.82915968741137, "learning_rate": 1.977152568549501e-05, "loss": 0.9725, "step": 1847 }, { "epoch": 0.29, "grad_norm": 33.0771888236441, "learning_rate": 1.9771167120872375e-05, "loss": 0.8645, "step": 1848 }, { "epoch": 0.29, "grad_norm": 20.287435605199015, "learning_rate": 1.977080827836346e-05, "loss": 0.9815, "step": 1849 }, { "epoch": 0.29, "grad_norm": 12.977668332700398, "learning_rate": 1.9770449157978475e-05, "loss": 0.7676, "step": 1850 }, { "epoch": 0.29, "grad_norm": 16.51914198257357, "learning_rate": 1.9770089759727634e-05, "loss": 0.8214, "step": 1851 }, { "epoch": 0.29, "grad_norm": 22.3756017737208, "learning_rate": 1.976973008362115e-05, "loss": 0.8112, "step": 1852 }, { "epoch": 0.29, "grad_norm": 22.287185008240183, "learning_rate": 1.976937012966926e-05, "loss": 0.7605, "step": 1853 }, { "epoch": 0.29, "grad_norm": 17.322853428520098, "learning_rate": 1.97690098978822e-05, "loss": 0.9402, "step": 1854 }, { "epoch": 0.29, "grad_norm": 38.27820211694408, "learning_rate": 1.9768649388270213e-05, "loss": 0.8972, "step": 1855 }, { "epoch": 0.29, "grad_norm": 17.129150240127114, "learning_rate": 1.976828860084355e-05, "loss": 0.87, "step": 1856 }, { "epoch": 0.29, "grad_norm": 18.448595222631848, "learning_rate": 1.9767927535612473e-05, "loss": 0.8432, "step": 1857 }, { "epoch": 0.29, "grad_norm": 16.16610402139183, "learning_rate": 1.9767566192587253e-05, "loss": 0.8021, "step": 1858 }, { "epoch": 0.29, "grad_norm": 25.94836865062978, "learning_rate": 1.9767204571778163e-05, "loss": 0.9226, "step": 1859 }, { "epoch": 0.29, "grad_norm": 20.758907658204592, "learning_rate": 1.976684267319549e-05, "loss": 0.8218, "step": 1860 }, { "epoch": 0.29, "grad_norm": 26.600530844694067, "learning_rate": 1.9766480496849527e-05, "loss": 0.8165, "step": 1861 }, { "epoch": 0.29, "grad_norm": 19.589273431906122, "learning_rate": 1.9766118042750566e-05, "loss": 0.7289, "step": 1862 }, { "epoch": 0.29, "grad_norm": 20.449142527324344, "learning_rate": 1.9765755310908925e-05, "loss": 0.8735, "step": 1863 }, { "epoch": 0.29, "grad_norm": 19.13288460527496, "learning_rate": 1.9765392301334912e-05, "loss": 0.8423, "step": 1864 }, { "epoch": 0.29, "grad_norm": 24.610665193675995, "learning_rate": 1.976502901403886e-05, "loss": 0.9759, "step": 1865 }, { "epoch": 0.29, "grad_norm": 17.8569561545966, "learning_rate": 1.9764665449031093e-05, "loss": 0.9654, "step": 1866 }, { "epoch": 0.29, "grad_norm": 25.353380535986272, "learning_rate": 1.976430160632195e-05, "loss": 0.8497, "step": 1867 }, { "epoch": 0.29, "grad_norm": 30.508107354647823, "learning_rate": 1.9763937485921785e-05, "loss": 0.9529, "step": 1868 }, { "epoch": 0.29, "grad_norm": 22.969245304229776, "learning_rate": 1.976357308784095e-05, "loss": 0.8159, "step": 1869 }, { "epoch": 0.29, "grad_norm": 37.45412320798896, "learning_rate": 1.9763208412089802e-05, "loss": 0.785, "step": 1870 }, { "epoch": 0.29, "grad_norm": 16.87231431358377, "learning_rate": 1.9762843458678723e-05, "loss": 0.8465, "step": 1871 }, { "epoch": 0.29, "grad_norm": 25.83016601028346, "learning_rate": 1.976247822761809e-05, "loss": 0.8687, "step": 1872 }, { "epoch": 0.29, "grad_norm": 18.515231235357167, "learning_rate": 1.976211271891828e-05, "loss": 0.9696, "step": 1873 }, { "epoch": 0.29, "grad_norm": 20.620431983707377, "learning_rate": 1.97617469325897e-05, "loss": 0.8897, "step": 1874 }, { "epoch": 0.29, "grad_norm": 28.371820075238965, "learning_rate": 1.9761380868642745e-05, "loss": 1.0011, "step": 1875 }, { "epoch": 0.29, "grad_norm": 14.977174386233976, "learning_rate": 1.9761014527087828e-05, "loss": 0.9175, "step": 1876 }, { "epoch": 0.29, "grad_norm": 19.64465399427102, "learning_rate": 1.9760647907935372e-05, "loss": 0.7933, "step": 1877 }, { "epoch": 0.29, "grad_norm": 18.79532450577986, "learning_rate": 1.9760281011195793e-05, "loss": 0.8207, "step": 1878 }, { "epoch": 0.29, "grad_norm": 24.920172704359203, "learning_rate": 1.9759913836879535e-05, "loss": 0.8858, "step": 1879 }, { "epoch": 0.29, "grad_norm": 19.638220166399208, "learning_rate": 1.9759546384997042e-05, "loss": 0.9311, "step": 1880 }, { "epoch": 0.29, "grad_norm": 14.781828814539178, "learning_rate": 1.975917865555875e-05, "loss": 0.8638, "step": 1881 }, { "epoch": 0.29, "grad_norm": 21.054452944353176, "learning_rate": 1.9758810648575132e-05, "loss": 0.8943, "step": 1882 }, { "epoch": 0.29, "grad_norm": 26.283434521357467, "learning_rate": 1.9758442364056643e-05, "loss": 0.873, "step": 1883 }, { "epoch": 0.29, "grad_norm": 15.699712329327701, "learning_rate": 1.975807380201377e-05, "loss": 0.8444, "step": 1884 }, { "epoch": 0.29, "grad_norm": 14.492083412886533, "learning_rate": 1.9757704962456978e-05, "loss": 0.9192, "step": 1885 }, { "epoch": 0.29, "grad_norm": 27.698089892390986, "learning_rate": 1.9757335845396768e-05, "loss": 0.898, "step": 1886 }, { "epoch": 0.29, "grad_norm": 35.97715392895058, "learning_rate": 1.9756966450843635e-05, "loss": 0.9176, "step": 1887 }, { "epoch": 0.29, "grad_norm": 18.065809891214382, "learning_rate": 1.9756596778808085e-05, "loss": 0.862, "step": 1888 }, { "epoch": 0.3, "grad_norm": 29.38781764166271, "learning_rate": 1.9756226829300628e-05, "loss": 0.8988, "step": 1889 }, { "epoch": 0.3, "grad_norm": 16.457850544725364, "learning_rate": 1.9755856602331788e-05, "loss": 0.8768, "step": 1890 }, { "epoch": 0.3, "grad_norm": 13.788459257811533, "learning_rate": 1.9755486097912092e-05, "loss": 0.8156, "step": 1891 }, { "epoch": 0.3, "grad_norm": 23.2677302948353, "learning_rate": 1.9755115316052076e-05, "loss": 0.7801, "step": 1892 }, { "epoch": 0.3, "grad_norm": 22.212946392362642, "learning_rate": 1.9754744256762293e-05, "loss": 0.8462, "step": 1893 }, { "epoch": 0.3, "grad_norm": 24.414810513443292, "learning_rate": 1.9754372920053284e-05, "loss": 0.9012, "step": 1894 }, { "epoch": 0.3, "grad_norm": 19.37392721590177, "learning_rate": 1.975400130593562e-05, "loss": 0.8598, "step": 1895 }, { "epoch": 0.3, "grad_norm": 25.597424221282452, "learning_rate": 1.975362941441986e-05, "loss": 0.9248, "step": 1896 }, { "epoch": 0.3, "grad_norm": 15.675500822089871, "learning_rate": 1.975325724551659e-05, "loss": 0.9456, "step": 1897 }, { "epoch": 0.3, "grad_norm": 15.55818149337485, "learning_rate": 1.9752884799236387e-05, "loss": 0.7981, "step": 1898 }, { "epoch": 0.3, "grad_norm": 25.309976139340897, "learning_rate": 1.9752512075589844e-05, "loss": 0.9044, "step": 1899 }, { "epoch": 0.3, "grad_norm": 22.143745146314007, "learning_rate": 1.9752139074587563e-05, "loss": 0.8209, "step": 1900 }, { "epoch": 0.3, "grad_norm": 20.12003667117563, "learning_rate": 1.9751765796240153e-05, "loss": 0.8165, "step": 1901 }, { "epoch": 0.3, "grad_norm": 14.426752350094178, "learning_rate": 1.9751392240558228e-05, "loss": 0.9043, "step": 1902 }, { "epoch": 0.3, "grad_norm": 20.94066089000282, "learning_rate": 1.9751018407552417e-05, "loss": 0.9336, "step": 1903 }, { "epoch": 0.3, "grad_norm": 18.89384598324315, "learning_rate": 1.9750644297233342e-05, "loss": 0.8236, "step": 1904 }, { "epoch": 0.3, "grad_norm": 19.065600030567253, "learning_rate": 1.9750269909611645e-05, "loss": 0.9207, "step": 1905 }, { "epoch": 0.3, "grad_norm": 19.548708685621445, "learning_rate": 1.9749895244697978e-05, "loss": 0.8082, "step": 1906 }, { "epoch": 0.3, "grad_norm": 18.02616298028141, "learning_rate": 1.9749520302502993e-05, "loss": 0.7811, "step": 1907 }, { "epoch": 0.3, "grad_norm": 10.92885723191646, "learning_rate": 1.9749145083037354e-05, "loss": 0.8268, "step": 1908 }, { "epoch": 0.3, "grad_norm": 21.275090474531414, "learning_rate": 1.9748769586311732e-05, "loss": 0.7387, "step": 1909 }, { "epoch": 0.3, "grad_norm": 17.556192712972138, "learning_rate": 1.9748393812336807e-05, "loss": 0.798, "step": 1910 }, { "epoch": 0.3, "grad_norm": 29.336131973739484, "learning_rate": 1.9748017761123265e-05, "loss": 0.8965, "step": 1911 }, { "epoch": 0.3, "grad_norm": 22.347891191708836, "learning_rate": 1.97476414326818e-05, "loss": 0.925, "step": 1912 }, { "epoch": 0.3, "grad_norm": 26.474903759362682, "learning_rate": 1.9747264827023113e-05, "loss": 0.8234, "step": 1913 }, { "epoch": 0.3, "grad_norm": 20.728841426987962, "learning_rate": 1.974688794415792e-05, "loss": 0.9392, "step": 1914 }, { "epoch": 0.3, "grad_norm": 18.138012152042396, "learning_rate": 1.9746510784096928e-05, "loss": 0.7939, "step": 1915 }, { "epoch": 0.3, "grad_norm": 16.234423786012925, "learning_rate": 1.974613334685088e-05, "loss": 0.8185, "step": 1916 }, { "epoch": 0.3, "grad_norm": 22.95480920560591, "learning_rate": 1.9745755632430496e-05, "loss": 0.925, "step": 1917 }, { "epoch": 0.3, "grad_norm": 21.165180219028983, "learning_rate": 1.9745377640846523e-05, "loss": 0.9045, "step": 1918 }, { "epoch": 0.3, "grad_norm": 16.59620959701154, "learning_rate": 1.974499937210971e-05, "loss": 0.8652, "step": 1919 }, { "epoch": 0.3, "grad_norm": 14.839764076434742, "learning_rate": 1.974462082623082e-05, "loss": 0.7006, "step": 1920 }, { "epoch": 0.3, "grad_norm": 34.00717524065735, "learning_rate": 1.9744242003220607e-05, "loss": 0.9767, "step": 1921 }, { "epoch": 0.3, "grad_norm": 38.31057661931976, "learning_rate": 1.9743862903089857e-05, "loss": 0.9409, "step": 1922 }, { "epoch": 0.3, "grad_norm": 25.645945820431308, "learning_rate": 1.9743483525849345e-05, "loss": 0.8992, "step": 1923 }, { "epoch": 0.3, "grad_norm": 20.038228969081157, "learning_rate": 1.974310387150986e-05, "loss": 0.8825, "step": 1924 }, { "epoch": 0.3, "grad_norm": 22.159937852966443, "learning_rate": 1.9742723940082203e-05, "loss": 0.8062, "step": 1925 }, { "epoch": 0.3, "grad_norm": 29.16880505023091, "learning_rate": 1.9742343731577173e-05, "loss": 0.8642, "step": 1926 }, { "epoch": 0.3, "grad_norm": 15.390617242801474, "learning_rate": 1.9741963246005592e-05, "loss": 0.8687, "step": 1927 }, { "epoch": 0.3, "grad_norm": 18.389926695388457, "learning_rate": 1.974158248337827e-05, "loss": 0.8399, "step": 1928 }, { "epoch": 0.3, "grad_norm": 20.723959075932196, "learning_rate": 1.9741201443706042e-05, "loss": 0.7871, "step": 1929 }, { "epoch": 0.3, "grad_norm": 17.139668398846528, "learning_rate": 1.9740820126999743e-05, "loss": 0.9488, "step": 1930 }, { "epoch": 0.3, "grad_norm": 26.27604148544649, "learning_rate": 1.9740438533270218e-05, "loss": 0.8128, "step": 1931 }, { "epoch": 0.3, "grad_norm": 17.72623610921065, "learning_rate": 1.9740056662528318e-05, "loss": 0.8657, "step": 1932 }, { "epoch": 0.3, "grad_norm": 24.8832880342997, "learning_rate": 1.9739674514784906e-05, "loss": 0.8405, "step": 1933 }, { "epoch": 0.3, "grad_norm": 29.337330406431807, "learning_rate": 1.973929209005085e-05, "loss": 0.918, "step": 1934 }, { "epoch": 0.3, "grad_norm": 30.514234999501607, "learning_rate": 1.9738909388337023e-05, "loss": 0.9271, "step": 1935 }, { "epoch": 0.3, "grad_norm": 16.198109752315794, "learning_rate": 1.973852640965431e-05, "loss": 0.7544, "step": 1936 }, { "epoch": 0.3, "grad_norm": 29.291033052491546, "learning_rate": 1.9738143154013603e-05, "loss": 0.9277, "step": 1937 }, { "epoch": 0.3, "grad_norm": 23.874831114790798, "learning_rate": 1.9737759621425797e-05, "loss": 0.8836, "step": 1938 }, { "epoch": 0.3, "grad_norm": 17.85217318481541, "learning_rate": 1.973737581190181e-05, "loss": 0.8899, "step": 1939 }, { "epoch": 0.3, "grad_norm": 17.54350282921674, "learning_rate": 1.973699172545255e-05, "loss": 0.7964, "step": 1940 }, { "epoch": 0.3, "grad_norm": 26.825471069158436, "learning_rate": 1.973660736208894e-05, "loss": 0.9965, "step": 1941 }, { "epoch": 0.3, "grad_norm": 30.974556993406168, "learning_rate": 1.973622272182191e-05, "loss": 0.9037, "step": 1942 }, { "epoch": 0.3, "grad_norm": 22.06262056315065, "learning_rate": 1.9735837804662404e-05, "loss": 0.9677, "step": 1943 }, { "epoch": 0.3, "grad_norm": 15.281524040440718, "learning_rate": 1.9735452610621366e-05, "loss": 0.7922, "step": 1944 }, { "epoch": 0.3, "grad_norm": 23.790383033869862, "learning_rate": 1.9735067139709753e-05, "loss": 0.8615, "step": 1945 }, { "epoch": 0.3, "grad_norm": 28.5259578759439, "learning_rate": 1.973468139193852e-05, "loss": 0.9008, "step": 1946 }, { "epoch": 0.3, "grad_norm": 30.496352395834368, "learning_rate": 1.9734295367318648e-05, "loss": 0.9198, "step": 1947 }, { "epoch": 0.3, "grad_norm": 15.299607119627156, "learning_rate": 1.9733909065861107e-05, "loss": 0.8284, "step": 1948 }, { "epoch": 0.3, "grad_norm": 20.589582219757954, "learning_rate": 1.9733522487576884e-05, "loss": 0.8235, "step": 1949 }, { "epoch": 0.3, "grad_norm": 25.360756131942562, "learning_rate": 1.973313563247698e-05, "loss": 0.8538, "step": 1950 }, { "epoch": 0.3, "grad_norm": 23.774548061178905, "learning_rate": 1.9732748500572395e-05, "loss": 0.8256, "step": 1951 }, { "epoch": 0.3, "grad_norm": 13.7938465135125, "learning_rate": 1.973236109187413e-05, "loss": 0.8148, "step": 1952 }, { "epoch": 0.31, "grad_norm": 20.9068540534322, "learning_rate": 1.973197340639321e-05, "loss": 0.8526, "step": 1953 }, { "epoch": 0.31, "grad_norm": 24.386475881564493, "learning_rate": 1.973158544414066e-05, "loss": 0.8164, "step": 1954 }, { "epoch": 0.31, "grad_norm": 17.91021109918913, "learning_rate": 1.9731197205127514e-05, "loss": 0.7654, "step": 1955 }, { "epoch": 0.31, "grad_norm": 15.011544735068918, "learning_rate": 1.973080868936481e-05, "loss": 0.8016, "step": 1956 }, { "epoch": 0.31, "grad_norm": 16.43234175575187, "learning_rate": 1.9730419896863603e-05, "loss": 0.7588, "step": 1957 }, { "epoch": 0.31, "grad_norm": 21.63898731879432, "learning_rate": 1.9730030827634942e-05, "loss": 0.9247, "step": 1958 }, { "epoch": 0.31, "grad_norm": 20.99724500329837, "learning_rate": 1.9729641481689897e-05, "loss": 0.8635, "step": 1959 }, { "epoch": 0.31, "grad_norm": 20.48537461833763, "learning_rate": 1.972925185903954e-05, "loss": 0.8497, "step": 1960 }, { "epoch": 0.31, "grad_norm": 22.691299742929573, "learning_rate": 1.9728861959694957e-05, "loss": 0.8473, "step": 1961 }, { "epoch": 0.31, "grad_norm": 17.4133431263369, "learning_rate": 1.9728471783667226e-05, "loss": 0.7965, "step": 1962 }, { "epoch": 0.31, "grad_norm": 15.797636299579123, "learning_rate": 1.9728081330967448e-05, "loss": 0.8817, "step": 1963 }, { "epoch": 0.31, "grad_norm": 21.551570678047835, "learning_rate": 1.972769060160673e-05, "loss": 0.9161, "step": 1964 }, { "epoch": 0.31, "grad_norm": 24.364141629164624, "learning_rate": 1.9727299595596178e-05, "loss": 0.9401, "step": 1965 }, { "epoch": 0.31, "grad_norm": 19.65152058387248, "learning_rate": 1.9726908312946918e-05, "loss": 0.7555, "step": 1966 }, { "epoch": 0.31, "grad_norm": 19.519633299946516, "learning_rate": 1.9726516753670078e-05, "loss": 0.9234, "step": 1967 }, { "epoch": 0.31, "grad_norm": 20.187621768784943, "learning_rate": 1.972612491777679e-05, "loss": 0.8401, "step": 1968 }, { "epoch": 0.31, "grad_norm": 21.113322729939835, "learning_rate": 1.9725732805278198e-05, "loss": 0.8784, "step": 1969 }, { "epoch": 0.31, "grad_norm": 24.121764926298493, "learning_rate": 1.9725340416185456e-05, "loss": 0.7902, "step": 1970 }, { "epoch": 0.31, "grad_norm": 14.984586533215857, "learning_rate": 1.9724947750509718e-05, "loss": 0.8074, "step": 1971 }, { "epoch": 0.31, "grad_norm": 19.058733451536096, "learning_rate": 1.9724554808262157e-05, "loss": 0.9017, "step": 1972 }, { "epoch": 0.31, "grad_norm": 32.59790955222826, "learning_rate": 1.9724161589453948e-05, "loss": 1.061, "step": 1973 }, { "epoch": 0.31, "grad_norm": 20.22872112332218, "learning_rate": 1.9723768094096268e-05, "loss": 0.8682, "step": 1974 }, { "epoch": 0.31, "grad_norm": 4.163242602012352, "learning_rate": 1.9723374322200315e-05, "loss": 0.8198, "step": 1975 }, { "epoch": 0.31, "grad_norm": 21.033946309405977, "learning_rate": 1.9722980273777284e-05, "loss": 0.7993, "step": 1976 }, { "epoch": 0.31, "grad_norm": 28.124299212043177, "learning_rate": 1.9722585948838378e-05, "loss": 0.7894, "step": 1977 }, { "epoch": 0.31, "grad_norm": 22.55154649185102, "learning_rate": 1.972219134739482e-05, "loss": 0.8277, "step": 1978 }, { "epoch": 0.31, "grad_norm": 21.439044298984843, "learning_rate": 1.9721796469457827e-05, "loss": 0.8154, "step": 1979 }, { "epoch": 0.31, "grad_norm": 24.263510828712764, "learning_rate": 1.9721401315038628e-05, "loss": 0.867, "step": 1980 }, { "epoch": 0.31, "grad_norm": 22.413829884031742, "learning_rate": 1.972100588414846e-05, "loss": 0.8369, "step": 1981 }, { "epoch": 0.31, "grad_norm": 21.421277764785366, "learning_rate": 1.9720610176798574e-05, "loss": 0.9176, "step": 1982 }, { "epoch": 0.31, "grad_norm": 16.591302858287282, "learning_rate": 1.972021419300022e-05, "loss": 0.8975, "step": 1983 }, { "epoch": 0.31, "grad_norm": 17.44606981628151, "learning_rate": 1.9719817932764662e-05, "loss": 0.7829, "step": 1984 }, { "epoch": 0.31, "grad_norm": 25.60036458020331, "learning_rate": 1.9719421396103167e-05, "loss": 0.7589, "step": 1985 }, { "epoch": 0.31, "grad_norm": 19.488065013806025, "learning_rate": 1.9719024583027012e-05, "loss": 0.9107, "step": 1986 }, { "epoch": 0.31, "grad_norm": 19.800749859434116, "learning_rate": 1.9718627493547486e-05, "loss": 0.744, "step": 1987 }, { "epoch": 0.31, "grad_norm": 19.87122036874998, "learning_rate": 1.9718230127675877e-05, "loss": 0.8792, "step": 1988 }, { "epoch": 0.31, "grad_norm": 11.451930796982744, "learning_rate": 1.971783248542349e-05, "loss": 0.7478, "step": 1989 }, { "epoch": 0.31, "grad_norm": 23.369709856981036, "learning_rate": 1.9717434566801633e-05, "loss": 0.9079, "step": 1990 }, { "epoch": 0.31, "grad_norm": 10.6883050569356, "learning_rate": 1.9717036371821615e-05, "loss": 0.8945, "step": 1991 }, { "epoch": 0.31, "grad_norm": 18.590627946723554, "learning_rate": 1.9716637900494775e-05, "loss": 0.739, "step": 1992 }, { "epoch": 0.31, "grad_norm": 32.17525068408096, "learning_rate": 1.9716239152832434e-05, "loss": 0.9047, "step": 1993 }, { "epoch": 0.31, "grad_norm": 18.223546792008175, "learning_rate": 1.9715840128845934e-05, "loss": 0.746, "step": 1994 }, { "epoch": 0.31, "grad_norm": 26.474822254232627, "learning_rate": 1.9715440828546626e-05, "loss": 0.9029, "step": 1995 }, { "epoch": 0.31, "grad_norm": 12.990517024875519, "learning_rate": 1.9715041251945863e-05, "loss": 0.7252, "step": 1996 }, { "epoch": 0.31, "grad_norm": 23.633635011972196, "learning_rate": 1.9714641399055007e-05, "loss": 0.8276, "step": 1997 }, { "epoch": 0.31, "grad_norm": 30.676383652935883, "learning_rate": 1.9714241269885436e-05, "loss": 0.9074, "step": 1998 }, { "epoch": 0.31, "grad_norm": 18.616173736210115, "learning_rate": 1.9713840864448528e-05, "loss": 0.8446, "step": 1999 }, { "epoch": 0.31, "grad_norm": 18.878151480522458, "learning_rate": 1.9713440182755665e-05, "loss": 0.8643, "step": 2000 }, { "epoch": 0.31, "grad_norm": 17.67374095374624, "learning_rate": 1.9713039224818247e-05, "loss": 0.8124, "step": 2001 }, { "epoch": 0.31, "grad_norm": 19.68957343006568, "learning_rate": 1.9712637990647674e-05, "loss": 0.8725, "step": 2002 }, { "epoch": 0.31, "grad_norm": 16.224064555261798, "learning_rate": 1.971223648025536e-05, "loss": 0.8174, "step": 2003 }, { "epoch": 0.31, "grad_norm": 18.320630612895478, "learning_rate": 1.9711834693652722e-05, "loss": 0.8064, "step": 2004 }, { "epoch": 0.31, "grad_norm": 24.548316005124242, "learning_rate": 1.9711432630851182e-05, "loss": 0.8951, "step": 2005 }, { "epoch": 0.31, "grad_norm": 22.668361165424816, "learning_rate": 1.9711030291862187e-05, "loss": 0.8711, "step": 2006 }, { "epoch": 0.31, "grad_norm": 16.07097755932726, "learning_rate": 1.9710627676697167e-05, "loss": 0.8192, "step": 2007 }, { "epoch": 0.31, "grad_norm": 20.01017858545481, "learning_rate": 1.9710224785367575e-05, "loss": 0.8742, "step": 2008 }, { "epoch": 0.31, "grad_norm": 17.438593319146708, "learning_rate": 1.9709821617884873e-05, "loss": 0.9124, "step": 2009 }, { "epoch": 0.31, "grad_norm": 20.11446710507202, "learning_rate": 1.9709418174260523e-05, "loss": 0.8485, "step": 2010 }, { "epoch": 0.31, "grad_norm": 20.509390048068454, "learning_rate": 1.9709014454506e-05, "loss": 0.8336, "step": 2011 }, { "epoch": 0.31, "grad_norm": 16.8540006601597, "learning_rate": 1.9708610458632783e-05, "loss": 0.9212, "step": 2012 }, { "epoch": 0.31, "grad_norm": 17.426905372360444, "learning_rate": 1.970820618665237e-05, "loss": 0.8895, "step": 2013 }, { "epoch": 0.31, "grad_norm": 21.478918339061252, "learning_rate": 1.9707801638576246e-05, "loss": 0.8906, "step": 2014 }, { "epoch": 0.31, "grad_norm": 22.198459180637528, "learning_rate": 1.9707396814415927e-05, "loss": 0.8595, "step": 2015 }, { "epoch": 0.31, "grad_norm": 23.220622442841528, "learning_rate": 1.970699171418292e-05, "loss": 0.8944, "step": 2016 }, { "epoch": 0.32, "grad_norm": 19.090429001609277, "learning_rate": 1.9706586337888744e-05, "loss": 0.9746, "step": 2017 }, { "epoch": 0.32, "grad_norm": 37.80171079558046, "learning_rate": 1.9706180685544933e-05, "loss": 0.9084, "step": 2018 }, { "epoch": 0.32, "grad_norm": 23.693300957325114, "learning_rate": 1.9705774757163022e-05, "loss": 0.7714, "step": 2019 }, { "epoch": 0.32, "grad_norm": 36.16578761922622, "learning_rate": 1.9705368552754554e-05, "loss": 0.9715, "step": 2020 }, { "epoch": 0.32, "grad_norm": 30.37952921600066, "learning_rate": 1.970496207233108e-05, "loss": 1.0402, "step": 2021 }, { "epoch": 0.32, "grad_norm": 13.97949704174473, "learning_rate": 1.9704555315904164e-05, "loss": 0.7923, "step": 2022 }, { "epoch": 0.32, "grad_norm": 23.425513981446954, "learning_rate": 1.9704148283485374e-05, "loss": 0.8088, "step": 2023 }, { "epoch": 0.32, "grad_norm": 17.737425028472487, "learning_rate": 1.9703740975086282e-05, "loss": 0.9466, "step": 2024 }, { "epoch": 0.32, "grad_norm": 15.688795316457627, "learning_rate": 1.970333339071847e-05, "loss": 0.788, "step": 2025 }, { "epoch": 0.32, "grad_norm": 18.67685110749523, "learning_rate": 1.9702925530393535e-05, "loss": 0.8963, "step": 2026 }, { "epoch": 0.32, "grad_norm": 20.37720486713858, "learning_rate": 1.970251739412307e-05, "loss": 0.8601, "step": 2027 }, { "epoch": 0.32, "grad_norm": 21.424242169193125, "learning_rate": 1.9702108981918693e-05, "loss": 0.8474, "step": 2028 }, { "epoch": 0.32, "grad_norm": 28.64494758115085, "learning_rate": 1.970170029379201e-05, "loss": 0.9745, "step": 2029 }, { "epoch": 0.32, "grad_norm": 16.170713737276607, "learning_rate": 1.9701291329754645e-05, "loss": 0.7781, "step": 2030 }, { "epoch": 0.32, "grad_norm": 26.473651958448663, "learning_rate": 1.970088208981823e-05, "loss": 0.8505, "step": 2031 }, { "epoch": 0.32, "grad_norm": 25.099027657055245, "learning_rate": 1.9700472573994403e-05, "loss": 0.8165, "step": 2032 }, { "epoch": 0.32, "grad_norm": 16.389292650100625, "learning_rate": 1.970006278229481e-05, "loss": 0.8138, "step": 2033 }, { "epoch": 0.32, "grad_norm": 28.028228690501816, "learning_rate": 1.9699652714731106e-05, "loss": 0.8045, "step": 2034 }, { "epoch": 0.32, "grad_norm": 22.785452550579784, "learning_rate": 1.9699242371314952e-05, "loss": 0.8292, "step": 2035 }, { "epoch": 0.32, "grad_norm": 19.201463608848353, "learning_rate": 1.9698831752058023e-05, "loss": 0.863, "step": 2036 }, { "epoch": 0.32, "grad_norm": 18.340812277280527, "learning_rate": 1.969842085697199e-05, "loss": 0.938, "step": 2037 }, { "epoch": 0.32, "grad_norm": 17.485254576668932, "learning_rate": 1.969800968606854e-05, "loss": 0.8792, "step": 2038 }, { "epoch": 0.32, "grad_norm": 23.50361377495418, "learning_rate": 1.9697598239359368e-05, "loss": 0.9366, "step": 2039 }, { "epoch": 0.32, "grad_norm": 28.004593364380916, "learning_rate": 1.9697186516856177e-05, "loss": 0.8729, "step": 2040 }, { "epoch": 0.32, "grad_norm": 18.27001093393149, "learning_rate": 1.9696774518570673e-05, "loss": 0.81, "step": 2041 }, { "epoch": 0.32, "grad_norm": 20.275109917781354, "learning_rate": 1.9696362244514576e-05, "loss": 0.8853, "step": 2042 }, { "epoch": 0.32, "grad_norm": 15.747320829069553, "learning_rate": 1.9695949694699604e-05, "loss": 0.8787, "step": 2043 }, { "epoch": 0.32, "grad_norm": 18.324589701361806, "learning_rate": 1.96955368691375e-05, "loss": 0.9621, "step": 2044 }, { "epoch": 0.32, "grad_norm": 12.62062350748662, "learning_rate": 1.9695123767839995e-05, "loss": 0.7259, "step": 2045 }, { "epoch": 0.32, "grad_norm": 15.868806449305948, "learning_rate": 1.9694710390818844e-05, "loss": 0.8744, "step": 2046 }, { "epoch": 0.32, "grad_norm": 20.289745679185543, "learning_rate": 1.9694296738085802e-05, "loss": 0.8401, "step": 2047 }, { "epoch": 0.32, "grad_norm": 14.319830790116374, "learning_rate": 1.9693882809652626e-05, "loss": 0.8339, "step": 2048 }, { "epoch": 0.32, "grad_norm": 19.35813448628382, "learning_rate": 1.96934686055311e-05, "loss": 0.8886, "step": 2049 }, { "epoch": 0.32, "grad_norm": 28.498450702459422, "learning_rate": 1.9693054125732995e-05, "loss": 0.9545, "step": 2050 }, { "epoch": 0.32, "grad_norm": 23.41585665636211, "learning_rate": 1.96926393702701e-05, "loss": 0.7897, "step": 2051 }, { "epoch": 0.32, "grad_norm": 20.60940300575881, "learning_rate": 1.9692224339154215e-05, "loss": 0.8198, "step": 2052 }, { "epoch": 0.32, "grad_norm": 13.23249166491358, "learning_rate": 1.9691809032397135e-05, "loss": 0.7522, "step": 2053 }, { "epoch": 0.32, "grad_norm": 15.991811089588687, "learning_rate": 1.9691393450010675e-05, "loss": 0.8621, "step": 2054 }, { "epoch": 0.32, "grad_norm": 26.00682109812143, "learning_rate": 1.969097759200666e-05, "loss": 0.9021, "step": 2055 }, { "epoch": 0.32, "grad_norm": 14.863507928510778, "learning_rate": 1.969056145839691e-05, "loss": 0.8484, "step": 2056 }, { "epoch": 0.32, "grad_norm": 28.320363565333135, "learning_rate": 1.9690145049193257e-05, "loss": 0.8124, "step": 2057 }, { "epoch": 0.32, "grad_norm": 24.337468244865462, "learning_rate": 1.968972836440755e-05, "loss": 1.0166, "step": 2058 }, { "epoch": 0.32, "grad_norm": 23.886848211054595, "learning_rate": 1.9689311404051638e-05, "loss": 0.9355, "step": 2059 }, { "epoch": 0.32, "grad_norm": 20.150662392893672, "learning_rate": 1.9688894168137372e-05, "loss": 0.8225, "step": 2060 }, { "epoch": 0.32, "grad_norm": 16.100422275118852, "learning_rate": 1.9688476656676628e-05, "loss": 0.8357, "step": 2061 }, { "epoch": 0.32, "grad_norm": 15.912121041966605, "learning_rate": 1.9688058869681277e-05, "loss": 0.7751, "step": 2062 }, { "epoch": 0.32, "grad_norm": 22.385547289120343, "learning_rate": 1.96876408071632e-05, "loss": 0.9013, "step": 2063 }, { "epoch": 0.32, "grad_norm": 20.674005406574164, "learning_rate": 1.968722246913428e-05, "loss": 0.8803, "step": 2064 }, { "epoch": 0.32, "grad_norm": 14.651857143774105, "learning_rate": 1.9686803855606422e-05, "loss": 0.8493, "step": 2065 }, { "epoch": 0.32, "grad_norm": 20.83345569066038, "learning_rate": 1.968638496659153e-05, "loss": 0.8327, "step": 2066 }, { "epoch": 0.32, "grad_norm": 14.981675972967828, "learning_rate": 1.9685965802101517e-05, "loss": 0.8561, "step": 2067 }, { "epoch": 0.32, "grad_norm": 20.980884605410015, "learning_rate": 1.96855463621483e-05, "loss": 1.0125, "step": 2068 }, { "epoch": 0.32, "grad_norm": 16.170432441982427, "learning_rate": 1.9685126646743814e-05, "loss": 0.7933, "step": 2069 }, { "epoch": 0.32, "grad_norm": 25.31564892738416, "learning_rate": 1.9684706655899988e-05, "loss": 1.0028, "step": 2070 }, { "epoch": 0.32, "grad_norm": 27.402884714954798, "learning_rate": 1.968428638962877e-05, "loss": 0.8449, "step": 2071 }, { "epoch": 0.32, "grad_norm": 20.701878523820394, "learning_rate": 1.9683865847942116e-05, "loss": 0.8012, "step": 2072 }, { "epoch": 0.32, "grad_norm": 19.698104536417045, "learning_rate": 1.968344503085198e-05, "loss": 0.9933, "step": 2073 }, { "epoch": 0.32, "grad_norm": 21.286533086458956, "learning_rate": 1.968302393837033e-05, "loss": 0.8337, "step": 2074 }, { "epoch": 0.32, "grad_norm": 17.971950636545188, "learning_rate": 1.9682602570509147e-05, "loss": 0.8593, "step": 2075 }, { "epoch": 0.32, "grad_norm": 17.314689908288504, "learning_rate": 1.968218092728041e-05, "loss": 0.9049, "step": 2076 }, { "epoch": 0.32, "grad_norm": 22.8035843946017, "learning_rate": 1.968175900869611e-05, "loss": 0.8038, "step": 2077 }, { "epoch": 0.32, "grad_norm": 23.762809406054114, "learning_rate": 1.968133681476825e-05, "loss": 0.9007, "step": 2078 }, { "epoch": 0.32, "grad_norm": 18.792393175444175, "learning_rate": 1.9680914345508834e-05, "loss": 0.8621, "step": 2079 }, { "epoch": 0.32, "grad_norm": 15.112575483961884, "learning_rate": 1.9680491600929874e-05, "loss": 0.9009, "step": 2080 }, { "epoch": 0.33, "grad_norm": 18.51078749449631, "learning_rate": 1.9680068581043396e-05, "loss": 0.8398, "step": 2081 }, { "epoch": 0.33, "grad_norm": 20.781832085251615, "learning_rate": 1.9679645285861433e-05, "loss": 0.8318, "step": 2082 }, { "epoch": 0.33, "grad_norm": 17.82291140287787, "learning_rate": 1.9679221715396018e-05, "loss": 0.8048, "step": 2083 }, { "epoch": 0.33, "grad_norm": 32.944377575626525, "learning_rate": 1.9678797869659204e-05, "loss": 0.9733, "step": 2084 }, { "epoch": 0.33, "grad_norm": 15.676625373293733, "learning_rate": 1.9678373748663037e-05, "loss": 0.8363, "step": 2085 }, { "epoch": 0.33, "grad_norm": 30.685482164690686, "learning_rate": 1.967794935241958e-05, "loss": 0.95, "step": 2086 }, { "epoch": 0.33, "grad_norm": 21.79509052642431, "learning_rate": 1.967752468094091e-05, "loss": 0.8782, "step": 2087 }, { "epoch": 0.33, "grad_norm": 23.101179041048873, "learning_rate": 1.9677099734239098e-05, "loss": 0.9321, "step": 2088 }, { "epoch": 0.33, "grad_norm": 23.373758789247308, "learning_rate": 1.967667451232623e-05, "loss": 0.8857, "step": 2089 }, { "epoch": 0.33, "grad_norm": 15.729766105341012, "learning_rate": 1.9676249015214396e-05, "loss": 0.7793, "step": 2090 }, { "epoch": 0.33, "grad_norm": 45.513667037997, "learning_rate": 1.9675823242915705e-05, "loss": 0.8488, "step": 2091 }, { "epoch": 0.33, "grad_norm": 16.170709642632474, "learning_rate": 1.967539719544226e-05, "loss": 0.8632, "step": 2092 }, { "epoch": 0.33, "grad_norm": 23.873762529006765, "learning_rate": 1.967497087280618e-05, "loss": 0.8697, "step": 2093 }, { "epoch": 0.33, "grad_norm": 15.146375161618987, "learning_rate": 1.9674544275019583e-05, "loss": 0.8427, "step": 2094 }, { "epoch": 0.33, "grad_norm": 18.39465103304834, "learning_rate": 1.9674117402094612e-05, "loss": 0.7769, "step": 2095 }, { "epoch": 0.33, "grad_norm": 30.994700505673816, "learning_rate": 1.96736902540434e-05, "loss": 0.8662, "step": 2096 }, { "epoch": 0.33, "grad_norm": 17.297744865219787, "learning_rate": 1.9673262830878094e-05, "loss": 0.8444, "step": 2097 }, { "epoch": 0.33, "grad_norm": 18.50302667844549, "learning_rate": 1.9672835132610856e-05, "loss": 0.8278, "step": 2098 }, { "epoch": 0.33, "grad_norm": 20.231258831369892, "learning_rate": 1.967240715925384e-05, "loss": 0.9185, "step": 2099 }, { "epoch": 0.33, "grad_norm": 17.586851130579518, "learning_rate": 1.9671978910819225e-05, "loss": 0.8729, "step": 2100 }, { "epoch": 0.33, "grad_norm": 17.596507171474503, "learning_rate": 1.9671550387319188e-05, "loss": 0.7587, "step": 2101 }, { "epoch": 0.33, "grad_norm": 21.54986398351388, "learning_rate": 1.9671121588765913e-05, "loss": 0.8134, "step": 2102 }, { "epoch": 0.33, "grad_norm": 21.44040746963521, "learning_rate": 1.96706925151716e-05, "loss": 0.9036, "step": 2103 }, { "epoch": 0.33, "grad_norm": 19.642517391678766, "learning_rate": 1.967026316654845e-05, "loss": 0.9083, "step": 2104 }, { "epoch": 0.33, "grad_norm": 25.546774618435716, "learning_rate": 1.966983354290867e-05, "loss": 0.8881, "step": 2105 }, { "epoch": 0.33, "grad_norm": 12.580669462025064, "learning_rate": 1.9669403644264485e-05, "loss": 0.8202, "step": 2106 }, { "epoch": 0.33, "grad_norm": 3.9494752545919876, "learning_rate": 1.966897347062811e-05, "loss": 0.8271, "step": 2107 }, { "epoch": 0.33, "grad_norm": 28.918628136324678, "learning_rate": 1.9668543022011792e-05, "loss": 0.8817, "step": 2108 }, { "epoch": 0.33, "grad_norm": 30.775240789369356, "learning_rate": 1.9668112298427764e-05, "loss": 0.9015, "step": 2109 }, { "epoch": 0.33, "grad_norm": 19.737897706509454, "learning_rate": 1.9667681299888277e-05, "loss": 0.9715, "step": 2110 }, { "epoch": 0.33, "grad_norm": 23.843163249344535, "learning_rate": 1.966725002640559e-05, "loss": 0.8514, "step": 2111 }, { "epoch": 0.33, "grad_norm": 29.238157946085167, "learning_rate": 1.9666818477991967e-05, "loss": 0.8035, "step": 2112 }, { "epoch": 0.33, "grad_norm": 25.25716850174795, "learning_rate": 1.9666386654659683e-05, "loss": 0.7557, "step": 2113 }, { "epoch": 0.33, "grad_norm": 23.080429387928188, "learning_rate": 1.9665954556421014e-05, "loss": 0.8524, "step": 2114 }, { "epoch": 0.33, "grad_norm": 23.46195452223196, "learning_rate": 1.9665522183288253e-05, "loss": 0.9517, "step": 2115 }, { "epoch": 0.33, "grad_norm": 14.22997900690526, "learning_rate": 1.9665089535273698e-05, "loss": 0.7616, "step": 2116 }, { "epoch": 0.33, "grad_norm": 14.71597199941219, "learning_rate": 1.966465661238965e-05, "loss": 0.7954, "step": 2117 }, { "epoch": 0.33, "grad_norm": 16.747367255209674, "learning_rate": 1.966422341464842e-05, "loss": 0.819, "step": 2118 }, { "epoch": 0.33, "grad_norm": 25.337880005125236, "learning_rate": 1.966378994206233e-05, "loss": 0.8782, "step": 2119 }, { "epoch": 0.33, "grad_norm": 21.6161545450506, "learning_rate": 1.9663356194643704e-05, "loss": 0.7597, "step": 2120 }, { "epoch": 0.33, "grad_norm": 18.478990774292374, "learning_rate": 1.9662922172404887e-05, "loss": 0.8034, "step": 2121 }, { "epoch": 0.33, "grad_norm": 21.93094670261835, "learning_rate": 1.966248787535821e-05, "loss": 0.8329, "step": 2122 }, { "epoch": 0.33, "grad_norm": 18.203284656659473, "learning_rate": 1.9662053303516036e-05, "loss": 0.9106, "step": 2123 }, { "epoch": 0.33, "grad_norm": 18.648103820474752, "learning_rate": 1.9661618456890714e-05, "loss": 0.7647, "step": 2124 }, { "epoch": 0.33, "grad_norm": 16.951290720050416, "learning_rate": 1.9661183335494617e-05, "loss": 0.8088, "step": 2125 }, { "epoch": 0.33, "grad_norm": 20.791425707541396, "learning_rate": 1.9660747939340116e-05, "loss": 0.8276, "step": 2126 }, { "epoch": 0.33, "grad_norm": 17.346590899726824, "learning_rate": 1.9660312268439593e-05, "loss": 0.8467, "step": 2127 }, { "epoch": 0.33, "grad_norm": 31.015011462333717, "learning_rate": 1.9659876322805448e-05, "loss": 0.852, "step": 2128 }, { "epoch": 0.33, "grad_norm": 20.85371623488054, "learning_rate": 1.965944010245006e-05, "loss": 0.7844, "step": 2129 }, { "epoch": 0.33, "grad_norm": 20.3166608686285, "learning_rate": 1.9659003607385857e-05, "loss": 0.731, "step": 2130 }, { "epoch": 0.33, "grad_norm": 18.230175135283112, "learning_rate": 1.9658566837625236e-05, "loss": 0.8378, "step": 2131 }, { "epoch": 0.33, "grad_norm": 20.268960548845886, "learning_rate": 1.9658129793180625e-05, "loss": 0.8499, "step": 2132 }, { "epoch": 0.33, "grad_norm": 21.393021354586935, "learning_rate": 1.965769247406445e-05, "loss": 0.8268, "step": 2133 }, { "epoch": 0.33, "grad_norm": 24.690563458490438, "learning_rate": 1.9657254880289155e-05, "loss": 0.8692, "step": 2134 }, { "epoch": 0.33, "grad_norm": 18.730910040437305, "learning_rate": 1.965681701186718e-05, "loss": 0.8778, "step": 2135 }, { "epoch": 0.33, "grad_norm": 34.76191543504597, "learning_rate": 1.9656378868810978e-05, "loss": 0.9182, "step": 2136 }, { "epoch": 0.33, "grad_norm": 22.923119795595042, "learning_rate": 1.965594045113301e-05, "loss": 0.8778, "step": 2137 }, { "epoch": 0.33, "grad_norm": 35.14006837969686, "learning_rate": 1.965550175884574e-05, "loss": 0.9905, "step": 2138 }, { "epoch": 0.33, "grad_norm": 20.81176056135414, "learning_rate": 1.9655062791961654e-05, "loss": 0.9221, "step": 2139 }, { "epoch": 0.33, "grad_norm": 12.778697089127505, "learning_rate": 1.9654623550493227e-05, "loss": 0.7828, "step": 2140 }, { "epoch": 0.33, "grad_norm": 13.075896231015909, "learning_rate": 1.9654184034452955e-05, "loss": 0.7573, "step": 2141 }, { "epoch": 0.33, "grad_norm": 18.250945541957204, "learning_rate": 1.9653744243853333e-05, "loss": 0.7862, "step": 2142 }, { "epoch": 0.33, "grad_norm": 17.914711142341105, "learning_rate": 1.9653304178706877e-05, "loss": 0.8773, "step": 2143 }, { "epoch": 0.33, "grad_norm": 13.355644536765228, "learning_rate": 1.96528638390261e-05, "loss": 0.8611, "step": 2144 }, { "epoch": 0.34, "grad_norm": 19.78372677028951, "learning_rate": 1.9652423224823515e-05, "loss": 0.8258, "step": 2145 }, { "epoch": 0.34, "grad_norm": 13.400282524238868, "learning_rate": 1.965198233611166e-05, "loss": 0.8081, "step": 2146 }, { "epoch": 0.34, "grad_norm": 22.44229291155315, "learning_rate": 1.9651541172903076e-05, "loss": 0.8738, "step": 2147 }, { "epoch": 0.34, "grad_norm": 17.966595726243217, "learning_rate": 1.965109973521031e-05, "loss": 0.8329, "step": 2148 }, { "epoch": 0.34, "grad_norm": 18.382324163627537, "learning_rate": 1.965065802304591e-05, "loss": 0.7689, "step": 2149 }, { "epoch": 0.34, "grad_norm": 18.824299573279532, "learning_rate": 1.965021603642244e-05, "loss": 0.8131, "step": 2150 }, { "epoch": 0.34, "grad_norm": 18.696418335757354, "learning_rate": 1.9649773775352475e-05, "loss": 0.7632, "step": 2151 }, { "epoch": 0.34, "grad_norm": 22.731280950027507, "learning_rate": 1.9649331239848587e-05, "loss": 0.8744, "step": 2152 }, { "epoch": 0.34, "grad_norm": 14.206521976165025, "learning_rate": 1.9648888429923364e-05, "loss": 0.7901, "step": 2153 }, { "epoch": 0.34, "grad_norm": 16.19688452062193, "learning_rate": 1.9648445345589395e-05, "loss": 0.8523, "step": 2154 }, { "epoch": 0.34, "grad_norm": 14.333005006512693, "learning_rate": 1.9648001986859288e-05, "loss": 0.7507, "step": 2155 }, { "epoch": 0.34, "grad_norm": 21.647414526418764, "learning_rate": 1.9647558353745645e-05, "loss": 0.8648, "step": 2156 }, { "epoch": 0.34, "grad_norm": 25.029609753829504, "learning_rate": 1.964711444626109e-05, "loss": 0.8985, "step": 2157 }, { "epoch": 0.34, "grad_norm": 35.82979041602923, "learning_rate": 1.9646670264418242e-05, "loss": 0.8636, "step": 2158 }, { "epoch": 0.34, "grad_norm": 14.343582252070068, "learning_rate": 1.9646225808229734e-05, "loss": 0.7577, "step": 2159 }, { "epoch": 0.34, "grad_norm": 18.348562087888574, "learning_rate": 1.9645781077708208e-05, "loss": 0.7974, "step": 2160 }, { "epoch": 0.34, "grad_norm": 21.810396517951595, "learning_rate": 1.9645336072866312e-05, "loss": 0.8364, "step": 2161 }, { "epoch": 0.34, "grad_norm": 17.723141038253587, "learning_rate": 1.9644890793716697e-05, "loss": 0.9236, "step": 2162 }, { "epoch": 0.34, "grad_norm": 25.359374407699494, "learning_rate": 1.9644445240272033e-05, "loss": 0.9071, "step": 2163 }, { "epoch": 0.34, "grad_norm": 36.453060436232484, "learning_rate": 1.964399941254499e-05, "loss": 0.87, "step": 2164 }, { "epoch": 0.34, "grad_norm": 19.467608515584082, "learning_rate": 1.964355331054824e-05, "loss": 0.9243, "step": 2165 }, { "epoch": 0.34, "grad_norm": 19.248844364392504, "learning_rate": 1.9643106934294476e-05, "loss": 0.8163, "step": 2166 }, { "epoch": 0.34, "grad_norm": 19.122933555051425, "learning_rate": 1.9642660283796395e-05, "loss": 0.952, "step": 2167 }, { "epoch": 0.34, "grad_norm": 21.93321172529305, "learning_rate": 1.9642213359066694e-05, "loss": 0.8702, "step": 2168 }, { "epoch": 0.34, "grad_norm": 22.994449727132253, "learning_rate": 1.9641766160118086e-05, "loss": 1.0528, "step": 2169 }, { "epoch": 0.34, "grad_norm": 29.239068988388972, "learning_rate": 1.964131868696329e-05, "loss": 0.8495, "step": 2170 }, { "epoch": 0.34, "grad_norm": 28.44738303728948, "learning_rate": 1.964087093961503e-05, "loss": 0.9238, "step": 2171 }, { "epoch": 0.34, "grad_norm": 26.948207012360225, "learning_rate": 1.9640422918086038e-05, "loss": 0.8579, "step": 2172 }, { "epoch": 0.34, "grad_norm": 20.89967049998885, "learning_rate": 1.963997462238906e-05, "loss": 0.7927, "step": 2173 }, { "epoch": 0.34, "grad_norm": 17.8201842213537, "learning_rate": 1.963952605253684e-05, "loss": 0.7746, "step": 2174 }, { "epoch": 0.34, "grad_norm": 19.545274448273872, "learning_rate": 1.963907720854214e-05, "loss": 0.9211, "step": 2175 }, { "epoch": 0.34, "grad_norm": 23.93813130975505, "learning_rate": 1.9638628090417727e-05, "loss": 0.8126, "step": 2176 }, { "epoch": 0.34, "grad_norm": 4.84867541245477, "learning_rate": 1.9638178698176364e-05, "loss": 0.8616, "step": 2177 }, { "epoch": 0.34, "grad_norm": 21.82145434641559, "learning_rate": 1.963772903183084e-05, "loss": 0.844, "step": 2178 }, { "epoch": 0.34, "grad_norm": 23.43618542481379, "learning_rate": 1.963727909139394e-05, "loss": 0.782, "step": 2179 }, { "epoch": 0.34, "grad_norm": 22.283206081398543, "learning_rate": 1.9636828876878456e-05, "loss": 0.8548, "step": 2180 }, { "epoch": 0.34, "grad_norm": 16.47930205510977, "learning_rate": 1.96363783882972e-05, "loss": 0.7673, "step": 2181 }, { "epoch": 0.34, "grad_norm": 24.63215972502464, "learning_rate": 1.963592762566298e-05, "loss": 0.8863, "step": 2182 }, { "epoch": 0.34, "grad_norm": 27.53223359410934, "learning_rate": 1.9635476588988615e-05, "loss": 0.8786, "step": 2183 }, { "epoch": 0.34, "grad_norm": 29.538060661131315, "learning_rate": 1.963502527828693e-05, "loss": 0.8706, "step": 2184 }, { "epoch": 0.34, "grad_norm": 57.84225302005477, "learning_rate": 1.9634573693570767e-05, "loss": 0.8658, "step": 2185 }, { "epoch": 0.34, "grad_norm": 23.793039716665785, "learning_rate": 1.963412183485296e-05, "loss": 0.7691, "step": 2186 }, { "epoch": 0.34, "grad_norm": 16.26578418496224, "learning_rate": 1.9633669702146366e-05, "loss": 0.8299, "step": 2187 }, { "epoch": 0.34, "grad_norm": 21.478741276842545, "learning_rate": 1.963321729546384e-05, "loss": 0.8549, "step": 2188 }, { "epoch": 0.34, "grad_norm": 37.167273619461376, "learning_rate": 1.963276461481825e-05, "loss": 0.9106, "step": 2189 }, { "epoch": 0.34, "grad_norm": 56.31049650419636, "learning_rate": 1.963231166022247e-05, "loss": 0.861, "step": 2190 }, { "epoch": 0.34, "grad_norm": 23.201976639815516, "learning_rate": 1.9631858431689382e-05, "loss": 0.8063, "step": 2191 }, { "epoch": 0.34, "grad_norm": 16.645263961030828, "learning_rate": 1.963140492923187e-05, "loss": 0.8363, "step": 2192 }, { "epoch": 0.34, "grad_norm": 20.49710724500291, "learning_rate": 1.963095115286284e-05, "loss": 0.8837, "step": 2193 }, { "epoch": 0.34, "grad_norm": 15.65235460289088, "learning_rate": 1.963049710259519e-05, "loss": 0.9756, "step": 2194 }, { "epoch": 0.34, "grad_norm": 29.373686438471058, "learning_rate": 1.963004277844184e-05, "loss": 0.8135, "step": 2195 }, { "epoch": 0.34, "grad_norm": 4.171045728122307, "learning_rate": 1.9629588180415707e-05, "loss": 0.8234, "step": 2196 }, { "epoch": 0.34, "grad_norm": 19.958338018415482, "learning_rate": 1.9629133308529717e-05, "loss": 0.9029, "step": 2197 }, { "epoch": 0.34, "grad_norm": 22.054987007780277, "learning_rate": 1.962867816279681e-05, "loss": 0.8789, "step": 2198 }, { "epoch": 0.34, "grad_norm": 22.724933992763166, "learning_rate": 1.9628222743229927e-05, "loss": 0.8741, "step": 2199 }, { "epoch": 0.34, "grad_norm": 21.22813346884525, "learning_rate": 1.9627767049842022e-05, "loss": 0.9541, "step": 2200 }, { "epoch": 0.34, "grad_norm": 19.565110368362504, "learning_rate": 1.9627311082646053e-05, "loss": 0.8735, "step": 2201 }, { "epoch": 0.34, "grad_norm": 18.17061207401581, "learning_rate": 1.962685484165499e-05, "loss": 0.7636, "step": 2202 }, { "epoch": 0.34, "grad_norm": 21.33147770307174, "learning_rate": 1.9626398326881805e-05, "loss": 0.8396, "step": 2203 }, { "epoch": 0.34, "grad_norm": 22.692971774283702, "learning_rate": 1.9625941538339484e-05, "loss": 0.8401, "step": 2204 }, { "epoch": 0.34, "grad_norm": 15.696540156952329, "learning_rate": 1.962548447604102e-05, "loss": 0.8128, "step": 2205 }, { "epoch": 0.34, "grad_norm": 14.082897626280321, "learning_rate": 1.9625027139999404e-05, "loss": 0.7894, "step": 2206 }, { "epoch": 0.34, "grad_norm": 12.58972804448063, "learning_rate": 1.962456953022765e-05, "loss": 0.7516, "step": 2207 }, { "epoch": 0.34, "grad_norm": 24.880836772182267, "learning_rate": 1.9624111646738767e-05, "loss": 0.9179, "step": 2208 }, { "epoch": 0.35, "grad_norm": 17.28436038367192, "learning_rate": 1.962365348954578e-05, "loss": 0.8, "step": 2209 }, { "epoch": 0.35, "grad_norm": 23.63477719802138, "learning_rate": 1.9623195058661716e-05, "loss": 0.8046, "step": 2210 }, { "epoch": 0.35, "grad_norm": 17.092068807439325, "learning_rate": 1.9622736354099612e-05, "loss": 0.7285, "step": 2211 }, { "epoch": 0.35, "grad_norm": 25.19775477201366, "learning_rate": 1.9622277375872518e-05, "loss": 0.8251, "step": 2212 }, { "epoch": 0.35, "grad_norm": 21.554167467807773, "learning_rate": 1.9621818123993482e-05, "loss": 0.9185, "step": 2213 }, { "epoch": 0.35, "grad_norm": 29.398751526067016, "learning_rate": 1.962135859847557e-05, "loss": 0.9394, "step": 2214 }, { "epoch": 0.35, "grad_norm": 23.06735841237785, "learning_rate": 1.9620898799331848e-05, "loss": 0.9402, "step": 2215 }, { "epoch": 0.35, "grad_norm": 30.64704539316894, "learning_rate": 1.962043872657539e-05, "loss": 0.989, "step": 2216 }, { "epoch": 0.35, "grad_norm": 21.050664734708178, "learning_rate": 1.9619978380219282e-05, "loss": 0.8904, "step": 2217 }, { "epoch": 0.35, "grad_norm": 16.489511197724184, "learning_rate": 1.9619517760276618e-05, "loss": 0.889, "step": 2218 }, { "epoch": 0.35, "grad_norm": 27.021809114490775, "learning_rate": 1.9619056866760497e-05, "loss": 0.8951, "step": 2219 }, { "epoch": 0.35, "grad_norm": 21.09823309624137, "learning_rate": 1.9618595699684026e-05, "loss": 0.8278, "step": 2220 }, { "epoch": 0.35, "grad_norm": 20.472222781964913, "learning_rate": 1.9618134259060318e-05, "loss": 0.8743, "step": 2221 }, { "epoch": 0.35, "grad_norm": 27.254923119860816, "learning_rate": 1.96176725449025e-05, "loss": 0.9064, "step": 2222 }, { "epoch": 0.35, "grad_norm": 23.05791563768528, "learning_rate": 1.9617210557223697e-05, "loss": 0.8709, "step": 2223 }, { "epoch": 0.35, "grad_norm": 16.543076767281683, "learning_rate": 1.9616748296037052e-05, "loss": 0.8798, "step": 2224 }, { "epoch": 0.35, "grad_norm": 22.666738148373785, "learning_rate": 1.9616285761355717e-05, "loss": 0.895, "step": 2225 }, { "epoch": 0.35, "grad_norm": 16.746350567830284, "learning_rate": 1.9615822953192833e-05, "loss": 0.8131, "step": 2226 }, { "epoch": 0.35, "grad_norm": 15.436412622403108, "learning_rate": 1.9615359871561574e-05, "loss": 0.7679, "step": 2227 }, { "epoch": 0.35, "grad_norm": 16.810434706482013, "learning_rate": 1.9614896516475107e-05, "loss": 0.79, "step": 2228 }, { "epoch": 0.35, "grad_norm": 35.932532497907, "learning_rate": 1.96144328879466e-05, "loss": 0.8941, "step": 2229 }, { "epoch": 0.35, "grad_norm": 27.831174199343533, "learning_rate": 1.9613968985989253e-05, "loss": 1.0721, "step": 2230 }, { "epoch": 0.35, "grad_norm": 25.270256096950266, "learning_rate": 1.961350481061625e-05, "loss": 0.9382, "step": 2231 }, { "epoch": 0.35, "grad_norm": 24.38343000500924, "learning_rate": 1.9613040361840792e-05, "loss": 0.9236, "step": 2232 }, { "epoch": 0.35, "grad_norm": 20.443885365676206, "learning_rate": 1.961257563967609e-05, "loss": 0.759, "step": 2233 }, { "epoch": 0.35, "grad_norm": 19.715840613201532, "learning_rate": 1.9612110644135365e-05, "loss": 0.8116, "step": 2234 }, { "epoch": 0.35, "grad_norm": 20.429731530895157, "learning_rate": 1.961164537523183e-05, "loss": 0.9412, "step": 2235 }, { "epoch": 0.35, "grad_norm": 17.716173067865874, "learning_rate": 1.9611179832978725e-05, "loss": 0.8309, "step": 2236 }, { "epoch": 0.35, "grad_norm": 28.72478583669382, "learning_rate": 1.961071401738929e-05, "loss": 0.9226, "step": 2237 }, { "epoch": 0.35, "grad_norm": 13.306935361831256, "learning_rate": 1.961024792847677e-05, "loss": 0.8484, "step": 2238 }, { "epoch": 0.35, "grad_norm": 17.03491843244981, "learning_rate": 1.960978156625442e-05, "loss": 0.8104, "step": 2239 }, { "epoch": 0.35, "grad_norm": 18.021771311788267, "learning_rate": 1.96093149307355e-05, "loss": 0.8226, "step": 2240 }, { "epoch": 0.35, "grad_norm": 21.19508629061962, "learning_rate": 1.960884802193329e-05, "loss": 0.8646, "step": 2241 }, { "epoch": 0.35, "grad_norm": 17.193580278916478, "learning_rate": 1.960838083986106e-05, "loss": 0.8639, "step": 2242 }, { "epoch": 0.35, "grad_norm": 18.843368560416156, "learning_rate": 1.9607913384532102e-05, "loss": 0.886, "step": 2243 }, { "epoch": 0.35, "grad_norm": 13.13145673744327, "learning_rate": 1.9607445655959707e-05, "loss": 0.8167, "step": 2244 }, { "epoch": 0.35, "grad_norm": 19.90926852762867, "learning_rate": 1.9606977654157176e-05, "loss": 0.763, "step": 2245 }, { "epoch": 0.35, "grad_norm": 12.687757201124638, "learning_rate": 1.9606509379137823e-05, "loss": 0.7763, "step": 2246 }, { "epoch": 0.35, "grad_norm": 20.68120718669408, "learning_rate": 1.960604083091496e-05, "loss": 0.877, "step": 2247 }, { "epoch": 0.35, "grad_norm": 13.947481220770838, "learning_rate": 1.9605572009501918e-05, "loss": 0.7808, "step": 2248 }, { "epoch": 0.35, "grad_norm": 24.670105893666836, "learning_rate": 1.9605102914912025e-05, "loss": 0.8162, "step": 2249 }, { "epoch": 0.35, "grad_norm": 13.617112052826652, "learning_rate": 1.960463354715863e-05, "loss": 0.8944, "step": 2250 }, { "epoch": 0.35, "grad_norm": 28.60815847009406, "learning_rate": 1.9604163906255068e-05, "loss": 0.8596, "step": 2251 }, { "epoch": 0.35, "grad_norm": 19.786423520382826, "learning_rate": 1.9603693992214707e-05, "loss": 0.8678, "step": 2252 }, { "epoch": 0.35, "grad_norm": 18.845369301044503, "learning_rate": 1.9603223805050903e-05, "loss": 0.8616, "step": 2253 }, { "epoch": 0.35, "grad_norm": 21.754836853200725, "learning_rate": 1.9602753344777036e-05, "loss": 0.8292, "step": 2254 }, { "epoch": 0.35, "grad_norm": 26.716550201073716, "learning_rate": 1.960228261140648e-05, "loss": 0.9063, "step": 2255 }, { "epoch": 0.35, "grad_norm": 18.451365058301914, "learning_rate": 1.960181160495262e-05, "loss": 0.8574, "step": 2256 }, { "epoch": 0.35, "grad_norm": 21.877343921425027, "learning_rate": 1.960134032542886e-05, "loss": 0.8011, "step": 2257 }, { "epoch": 0.35, "grad_norm": 15.849301502671569, "learning_rate": 1.9600868772848595e-05, "loss": 0.8001, "step": 2258 }, { "epoch": 0.35, "grad_norm": 19.364172779712053, "learning_rate": 1.9600396947225238e-05, "loss": 0.8061, "step": 2259 }, { "epoch": 0.35, "grad_norm": 17.25579954351772, "learning_rate": 1.959992484857221e-05, "loss": 0.8852, "step": 2260 }, { "epoch": 0.35, "grad_norm": 18.53819251681691, "learning_rate": 1.9599452476902934e-05, "loss": 0.7397, "step": 2261 }, { "epoch": 0.35, "grad_norm": 17.778921804544083, "learning_rate": 1.9598979832230846e-05, "loss": 0.8344, "step": 2262 }, { "epoch": 0.35, "grad_norm": 19.35354442268991, "learning_rate": 1.9598506914569383e-05, "loss": 0.789, "step": 2263 }, { "epoch": 0.35, "grad_norm": 14.017006439206416, "learning_rate": 1.9598033723932004e-05, "loss": 0.8317, "step": 2264 }, { "epoch": 0.35, "grad_norm": 28.88808198024425, "learning_rate": 1.959756026033216e-05, "loss": 0.8481, "step": 2265 }, { "epoch": 0.35, "grad_norm": 21.361581158227146, "learning_rate": 1.9597086523783312e-05, "loss": 0.9174, "step": 2266 }, { "epoch": 0.35, "grad_norm": 21.32743899042107, "learning_rate": 1.959661251429894e-05, "loss": 0.9045, "step": 2267 }, { "epoch": 0.35, "grad_norm": 43.97896255880699, "learning_rate": 1.959613823189252e-05, "loss": 1.0044, "step": 2268 }, { "epoch": 0.35, "grad_norm": 19.26575233791744, "learning_rate": 1.9595663676577543e-05, "loss": 0.8442, "step": 2269 }, { "epoch": 0.35, "grad_norm": 18.3437053194103, "learning_rate": 1.9595188848367507e-05, "loss": 0.7827, "step": 2270 }, { "epoch": 0.35, "grad_norm": 17.962052756168983, "learning_rate": 1.959471374727591e-05, "loss": 0.7672, "step": 2271 }, { "epoch": 0.35, "grad_norm": 29.265136007937627, "learning_rate": 1.959423837331627e-05, "loss": 0.9122, "step": 2272 }, { "epoch": 0.36, "grad_norm": 26.90766591102539, "learning_rate": 1.95937627265021e-05, "loss": 0.9845, "step": 2273 }, { "epoch": 0.36, "grad_norm": 22.351767860703873, "learning_rate": 1.9593286806846935e-05, "loss": 0.6728, "step": 2274 }, { "epoch": 0.36, "grad_norm": 28.274015619945366, "learning_rate": 1.95928106143643e-05, "loss": 0.9657, "step": 2275 }, { "epoch": 0.36, "grad_norm": 21.033006236249136, "learning_rate": 1.9592334149067745e-05, "loss": 0.7861, "step": 2276 }, { "epoch": 0.36, "grad_norm": 27.176641004022738, "learning_rate": 1.9591857410970818e-05, "loss": 0.8808, "step": 2277 }, { "epoch": 0.36, "grad_norm": 22.834020162842634, "learning_rate": 1.9591380400087077e-05, "loss": 0.9017, "step": 2278 }, { "epoch": 0.36, "grad_norm": 20.08037626915052, "learning_rate": 1.9590903116430087e-05, "loss": 0.8966, "step": 2279 }, { "epoch": 0.36, "grad_norm": 20.240594464788924, "learning_rate": 1.9590425560013427e-05, "loss": 0.8006, "step": 2280 }, { "epoch": 0.36, "grad_norm": 16.109508313127826, "learning_rate": 1.9589947730850668e-05, "loss": 0.7795, "step": 2281 }, { "epoch": 0.36, "grad_norm": 23.75527198715787, "learning_rate": 1.958946962895541e-05, "loss": 0.7473, "step": 2282 }, { "epoch": 0.36, "grad_norm": 24.926261129035797, "learning_rate": 1.9588991254341243e-05, "loss": 0.8775, "step": 2283 }, { "epoch": 0.36, "grad_norm": 17.210590480261835, "learning_rate": 1.9588512607021777e-05, "loss": 0.7519, "step": 2284 }, { "epoch": 0.36, "grad_norm": 32.38915544885124, "learning_rate": 1.958803368701062e-05, "loss": 0.9165, "step": 2285 }, { "epoch": 0.36, "grad_norm": 22.4473599510166, "learning_rate": 1.9587554494321394e-05, "loss": 0.7747, "step": 2286 }, { "epoch": 0.36, "grad_norm": 20.650569556725742, "learning_rate": 1.958707502896773e-05, "loss": 0.8813, "step": 2287 }, { "epoch": 0.36, "grad_norm": 23.36865528908284, "learning_rate": 1.9586595290963254e-05, "loss": 0.8279, "step": 2288 }, { "epoch": 0.36, "grad_norm": 23.36321122746258, "learning_rate": 1.958611528032162e-05, "loss": 0.7433, "step": 2289 }, { "epoch": 0.36, "grad_norm": 23.60395808911962, "learning_rate": 1.9585634997056472e-05, "loss": 0.9214, "step": 2290 }, { "epoch": 0.36, "grad_norm": 19.629458455531346, "learning_rate": 1.9585154441181475e-05, "loss": 0.7866, "step": 2291 }, { "epoch": 0.36, "grad_norm": 21.64594494261864, "learning_rate": 1.9584673612710293e-05, "loss": 0.8852, "step": 2292 }, { "epoch": 0.36, "grad_norm": 18.88986593059227, "learning_rate": 1.95841925116566e-05, "loss": 0.811, "step": 2293 }, { "epoch": 0.36, "grad_norm": 16.920206414764746, "learning_rate": 1.9583711138034075e-05, "loss": 0.9009, "step": 2294 }, { "epoch": 0.36, "grad_norm": 36.281520411250064, "learning_rate": 1.958322949185642e-05, "loss": 0.8237, "step": 2295 }, { "epoch": 0.36, "grad_norm": 17.502897981405205, "learning_rate": 1.9582747573137315e-05, "loss": 0.9108, "step": 2296 }, { "epoch": 0.36, "grad_norm": 14.900124518863047, "learning_rate": 1.958226538189048e-05, "loss": 0.8748, "step": 2297 }, { "epoch": 0.36, "grad_norm": 19.353509063519404, "learning_rate": 1.958178291812962e-05, "loss": 0.9391, "step": 2298 }, { "epoch": 0.36, "grad_norm": 22.069704368787633, "learning_rate": 1.9581300181868462e-05, "loss": 0.7822, "step": 2299 }, { "epoch": 0.36, "grad_norm": 19.563794908993, "learning_rate": 1.958081717312073e-05, "loss": 0.9021, "step": 2300 }, { "epoch": 0.36, "grad_norm": 20.7725747077412, "learning_rate": 1.9580333891900164e-05, "loss": 0.8624, "step": 2301 }, { "epoch": 0.36, "grad_norm": 15.52007982411239, "learning_rate": 1.9579850338220502e-05, "loss": 0.8048, "step": 2302 }, { "epoch": 0.36, "grad_norm": 31.01456514921061, "learning_rate": 1.9579366512095504e-05, "loss": 0.8553, "step": 2303 }, { "epoch": 0.36, "grad_norm": 21.39003111946264, "learning_rate": 1.9578882413538928e-05, "loss": 0.8755, "step": 2304 }, { "epoch": 0.36, "grad_norm": 21.87317373622345, "learning_rate": 1.9578398042564534e-05, "loss": 0.8392, "step": 2305 }, { "epoch": 0.36, "grad_norm": 28.639158294850937, "learning_rate": 1.9577913399186106e-05, "loss": 0.8533, "step": 2306 }, { "epoch": 0.36, "grad_norm": 16.981001838335168, "learning_rate": 1.9577428483417424e-05, "loss": 0.8814, "step": 2307 }, { "epoch": 0.36, "grad_norm": 17.767328364359244, "learning_rate": 1.957694329527228e-05, "loss": 0.9141, "step": 2308 }, { "epoch": 0.36, "grad_norm": 19.60064072758327, "learning_rate": 1.9576457834764468e-05, "loss": 0.8803, "step": 2309 }, { "epoch": 0.36, "grad_norm": 13.785497650368177, "learning_rate": 1.9575972101907798e-05, "loss": 0.8541, "step": 2310 }, { "epoch": 0.36, "grad_norm": 18.31662293800399, "learning_rate": 1.9575486096716086e-05, "loss": 0.8953, "step": 2311 }, { "epoch": 0.36, "grad_norm": 21.52787343403029, "learning_rate": 1.957499981920315e-05, "loss": 0.7918, "step": 2312 }, { "epoch": 0.36, "grad_norm": 21.865508904039164, "learning_rate": 1.957451326938282e-05, "loss": 0.8077, "step": 2313 }, { "epoch": 0.36, "grad_norm": 25.315322527875505, "learning_rate": 1.957402644726893e-05, "loss": 0.8776, "step": 2314 }, { "epoch": 0.36, "grad_norm": 11.644909527203188, "learning_rate": 1.9573539352875333e-05, "loss": 0.9013, "step": 2315 }, { "epoch": 0.36, "grad_norm": 21.987844389972054, "learning_rate": 1.9573051986215873e-05, "loss": 0.7833, "step": 2316 }, { "epoch": 0.36, "grad_norm": 21.81496192809406, "learning_rate": 1.9572564347304418e-05, "loss": 0.7088, "step": 2317 }, { "epoch": 0.36, "grad_norm": 14.478514597867123, "learning_rate": 1.957207643615483e-05, "loss": 0.8507, "step": 2318 }, { "epoch": 0.36, "grad_norm": 16.714268848564387, "learning_rate": 1.9571588252780992e-05, "loss": 0.8358, "step": 2319 }, { "epoch": 0.36, "grad_norm": 24.02872563318518, "learning_rate": 1.9571099797196778e-05, "loss": 0.7872, "step": 2320 }, { "epoch": 0.36, "grad_norm": 21.76587062007009, "learning_rate": 1.957061106941609e-05, "loss": 0.8648, "step": 2321 }, { "epoch": 0.36, "grad_norm": 18.02049166892051, "learning_rate": 1.9570122069452817e-05, "loss": 0.8226, "step": 2322 }, { "epoch": 0.36, "grad_norm": 17.835083095121952, "learning_rate": 1.9569632797320876e-05, "loss": 0.7784, "step": 2323 }, { "epoch": 0.36, "grad_norm": 20.276543367213247, "learning_rate": 1.9569143253034172e-05, "loss": 0.8056, "step": 2324 }, { "epoch": 0.36, "grad_norm": 18.884539515346667, "learning_rate": 1.9568653436606632e-05, "loss": 0.9212, "step": 2325 }, { "epoch": 0.36, "grad_norm": 17.54858209589304, "learning_rate": 1.9568163348052186e-05, "loss": 0.7636, "step": 2326 }, { "epoch": 0.36, "grad_norm": 23.923779145843948, "learning_rate": 1.9567672987384774e-05, "loss": 0.9018, "step": 2327 }, { "epoch": 0.36, "grad_norm": 22.46457753946557, "learning_rate": 1.9567182354618338e-05, "loss": 0.9344, "step": 2328 }, { "epoch": 0.36, "grad_norm": 29.285314327737893, "learning_rate": 1.956669144976683e-05, "loss": 0.8852, "step": 2329 }, { "epoch": 0.36, "grad_norm": 10.760738511625377, "learning_rate": 1.9566200272844216e-05, "loss": 0.716, "step": 2330 }, { "epoch": 0.36, "grad_norm": 19.21331647187897, "learning_rate": 1.956570882386446e-05, "loss": 0.8791, "step": 2331 }, { "epoch": 0.36, "grad_norm": 19.41563576630927, "learning_rate": 1.9565217102841543e-05, "loss": 0.8382, "step": 2332 }, { "epoch": 0.36, "grad_norm": 17.750899110356364, "learning_rate": 1.9564725109789447e-05, "loss": 0.7417, "step": 2333 }, { "epoch": 0.36, "grad_norm": 20.760458215179696, "learning_rate": 1.9564232844722166e-05, "loss": 0.7939, "step": 2334 }, { "epoch": 0.36, "grad_norm": 16.114660816668017, "learning_rate": 1.9563740307653693e-05, "loss": 0.8228, "step": 2335 }, { "epoch": 0.36, "grad_norm": 20.373569290927932, "learning_rate": 1.9563247498598045e-05, "loss": 0.8, "step": 2336 }, { "epoch": 0.37, "grad_norm": 18.201653685477662, "learning_rate": 1.956275441756923e-05, "loss": 0.8496, "step": 2337 }, { "epoch": 0.37, "grad_norm": 19.45357286711351, "learning_rate": 1.9562261064581272e-05, "loss": 0.8678, "step": 2338 }, { "epoch": 0.37, "grad_norm": 18.698160175413193, "learning_rate": 1.956176743964821e-05, "loss": 0.9822, "step": 2339 }, { "epoch": 0.37, "grad_norm": 17.29147152809738, "learning_rate": 1.9561273542784067e-05, "loss": 0.8554, "step": 2340 }, { "epoch": 0.37, "grad_norm": 19.468092203891214, "learning_rate": 1.9560779374002902e-05, "loss": 0.9422, "step": 2341 }, { "epoch": 0.37, "grad_norm": 16.90812846481586, "learning_rate": 1.9560284933318763e-05, "loss": 0.9153, "step": 2342 }, { "epoch": 0.37, "grad_norm": 14.881066794651561, "learning_rate": 1.9559790220745715e-05, "loss": 0.7362, "step": 2343 }, { "epoch": 0.37, "grad_norm": 18.834799798238674, "learning_rate": 1.9559295236297822e-05, "loss": 0.8588, "step": 2344 }, { "epoch": 0.37, "grad_norm": 19.44682119945334, "learning_rate": 1.9558799979989165e-05, "loss": 0.8622, "step": 2345 }, { "epoch": 0.37, "grad_norm": 19.7818948774698, "learning_rate": 1.9558304451833828e-05, "loss": 0.9244, "step": 2346 }, { "epoch": 0.37, "grad_norm": 32.15240395829633, "learning_rate": 1.955780865184591e-05, "loss": 0.8391, "step": 2347 }, { "epoch": 0.37, "grad_norm": 19.52928040606101, "learning_rate": 1.95573125800395e-05, "loss": 0.8631, "step": 2348 }, { "epoch": 0.37, "grad_norm": 13.014948379107024, "learning_rate": 1.955681623642871e-05, "loss": 0.7669, "step": 2349 }, { "epoch": 0.37, "grad_norm": 7.3606479985429365, "learning_rate": 1.955631962102766e-05, "loss": 0.9177, "step": 2350 }, { "epoch": 0.37, "grad_norm": 16.70675317168515, "learning_rate": 1.955582273385047e-05, "loss": 0.7743, "step": 2351 }, { "epoch": 0.37, "grad_norm": 16.959628858998226, "learning_rate": 1.955532557491127e-05, "loss": 0.7911, "step": 2352 }, { "epoch": 0.37, "grad_norm": 22.870408728266078, "learning_rate": 1.95548281442242e-05, "loss": 0.7857, "step": 2353 }, { "epoch": 0.37, "grad_norm": 16.12068729290322, "learning_rate": 1.9554330441803406e-05, "loss": 0.8328, "step": 2354 }, { "epoch": 0.37, "grad_norm": 21.39010542509534, "learning_rate": 1.9553832467663048e-05, "loss": 0.8622, "step": 2355 }, { "epoch": 0.37, "grad_norm": 18.484470397819127, "learning_rate": 1.955333422181728e-05, "loss": 0.9214, "step": 2356 }, { "epoch": 0.37, "grad_norm": 13.757632950277799, "learning_rate": 1.955283570428027e-05, "loss": 0.8439, "step": 2357 }, { "epoch": 0.37, "grad_norm": 18.790934236685473, "learning_rate": 1.955233691506621e-05, "loss": 0.9286, "step": 2358 }, { "epoch": 0.37, "grad_norm": 20.639654567005557, "learning_rate": 1.9551837854189274e-05, "loss": 0.8256, "step": 2359 }, { "epoch": 0.37, "grad_norm": 15.339632657461078, "learning_rate": 1.9551338521663656e-05, "loss": 0.8338, "step": 2360 }, { "epoch": 0.37, "grad_norm": 20.92224329452618, "learning_rate": 1.9550838917503557e-05, "loss": 0.937, "step": 2361 }, { "epoch": 0.37, "grad_norm": 24.22799428542671, "learning_rate": 1.9550339041723188e-05, "loss": 1.0115, "step": 2362 }, { "epoch": 0.37, "grad_norm": 15.104370162075357, "learning_rate": 1.954983889433676e-05, "loss": 0.7641, "step": 2363 }, { "epoch": 0.37, "grad_norm": 26.204035727222347, "learning_rate": 1.9549338475358505e-05, "loss": 0.9896, "step": 2364 }, { "epoch": 0.37, "grad_norm": 15.977562247710003, "learning_rate": 1.9548837784802646e-05, "loss": 0.7483, "step": 2365 }, { "epoch": 0.37, "grad_norm": 18.580521959104235, "learning_rate": 1.9548336822683426e-05, "loss": 0.8389, "step": 2366 }, { "epoch": 0.37, "grad_norm": 14.841327111116088, "learning_rate": 1.9547835589015095e-05, "loss": 0.7908, "step": 2367 }, { "epoch": 0.37, "grad_norm": 19.442799154015507, "learning_rate": 1.9547334083811906e-05, "loss": 0.9086, "step": 2368 }, { "epoch": 0.37, "grad_norm": 23.129577828623827, "learning_rate": 1.954683230708812e-05, "loss": 0.8035, "step": 2369 }, { "epoch": 0.37, "grad_norm": 16.687118091034925, "learning_rate": 1.9546330258858007e-05, "loss": 0.9282, "step": 2370 }, { "epoch": 0.37, "grad_norm": 18.267640544320162, "learning_rate": 1.9545827939135846e-05, "loss": 0.8739, "step": 2371 }, { "epoch": 0.37, "grad_norm": 21.876459420089844, "learning_rate": 1.9545325347935923e-05, "loss": 0.8541, "step": 2372 }, { "epoch": 0.37, "grad_norm": 15.80787492667919, "learning_rate": 1.9544822485272528e-05, "loss": 0.8464, "step": 2373 }, { "epoch": 0.37, "grad_norm": 19.987684051973616, "learning_rate": 1.954431935115997e-05, "loss": 0.876, "step": 2374 }, { "epoch": 0.37, "grad_norm": 19.834476586173817, "learning_rate": 1.9543815945612548e-05, "loss": 0.8347, "step": 2375 }, { "epoch": 0.37, "grad_norm": 22.49566676745921, "learning_rate": 1.9543312268644586e-05, "loss": 0.9176, "step": 2376 }, { "epoch": 0.37, "grad_norm": 11.844859224781588, "learning_rate": 1.9542808320270405e-05, "loss": 0.7221, "step": 2377 }, { "epoch": 0.37, "grad_norm": 21.313890675370274, "learning_rate": 1.954230410050434e-05, "loss": 0.8976, "step": 2378 }, { "epoch": 0.37, "grad_norm": 16.046858249402987, "learning_rate": 1.9541799609360727e-05, "loss": 0.8804, "step": 2379 }, { "epoch": 0.37, "grad_norm": 15.667118069864024, "learning_rate": 1.9541294846853914e-05, "loss": 0.7766, "step": 2380 }, { "epoch": 0.37, "grad_norm": 17.285024985122114, "learning_rate": 1.954078981299826e-05, "loss": 0.7479, "step": 2381 }, { "epoch": 0.37, "grad_norm": 18.918423588302655, "learning_rate": 1.954028450780812e-05, "loss": 0.8779, "step": 2382 }, { "epoch": 0.37, "grad_norm": 16.615739365173884, "learning_rate": 1.9539778931297874e-05, "loss": 0.7884, "step": 2383 }, { "epoch": 0.37, "grad_norm": 20.350674587510465, "learning_rate": 1.9539273083481894e-05, "loss": 0.8458, "step": 2384 }, { "epoch": 0.37, "grad_norm": 15.364339834697589, "learning_rate": 1.953876696437457e-05, "loss": 0.7645, "step": 2385 }, { "epoch": 0.37, "grad_norm": 22.216220994129927, "learning_rate": 1.953826057399029e-05, "loss": 0.833, "step": 2386 }, { "epoch": 0.37, "grad_norm": 18.280495522490607, "learning_rate": 1.953775391234346e-05, "loss": 0.8504, "step": 2387 }, { "epoch": 0.37, "grad_norm": 19.523350656292592, "learning_rate": 1.953724697944849e-05, "loss": 0.83, "step": 2388 }, { "epoch": 0.37, "grad_norm": 16.70498830699112, "learning_rate": 1.9536739775319793e-05, "loss": 0.7669, "step": 2389 }, { "epoch": 0.37, "grad_norm": 16.407102977740667, "learning_rate": 1.9536232299971797e-05, "loss": 0.7656, "step": 2390 }, { "epoch": 0.37, "grad_norm": 18.86788595577168, "learning_rate": 1.953572455341893e-05, "loss": 0.9123, "step": 2391 }, { "epoch": 0.37, "grad_norm": 23.757444997301512, "learning_rate": 1.9535216535675636e-05, "loss": 0.8602, "step": 2392 }, { "epoch": 0.37, "grad_norm": 19.303482902670957, "learning_rate": 1.9534708246756364e-05, "loss": 0.8665, "step": 2393 }, { "epoch": 0.37, "grad_norm": 26.802081116741597, "learning_rate": 1.9534199686675567e-05, "loss": 0.8638, "step": 2394 }, { "epoch": 0.37, "grad_norm": 19.25971291935905, "learning_rate": 1.9533690855447706e-05, "loss": 0.7855, "step": 2395 }, { "epoch": 0.37, "grad_norm": 39.078096568216615, "learning_rate": 1.9533181753087255e-05, "loss": 0.8725, "step": 2396 }, { "epoch": 0.37, "grad_norm": 22.16279272140576, "learning_rate": 1.953267237960869e-05, "loss": 0.9653, "step": 2397 }, { "epoch": 0.37, "grad_norm": 23.145180519751115, "learning_rate": 1.9532162735026503e-05, "loss": 0.8258, "step": 2398 }, { "epoch": 0.37, "grad_norm": 19.526106660730825, "learning_rate": 1.9531652819355184e-05, "loss": 0.795, "step": 2399 }, { "epoch": 0.37, "grad_norm": 21.05502578024802, "learning_rate": 1.953114263260923e-05, "loss": 0.7586, "step": 2400 }, { "epoch": 0.38, "grad_norm": 23.07608855269181, "learning_rate": 1.9530632174803158e-05, "loss": 0.8031, "step": 2401 }, { "epoch": 0.38, "grad_norm": 18.335398720951122, "learning_rate": 1.953012144595148e-05, "loss": 0.8231, "step": 2402 }, { "epoch": 0.38, "grad_norm": 20.26518463444549, "learning_rate": 1.9529610446068727e-05, "loss": 0.7897, "step": 2403 }, { "epoch": 0.38, "grad_norm": 17.46569208530518, "learning_rate": 1.9529099175169425e-05, "loss": 0.8447, "step": 2404 }, { "epoch": 0.38, "grad_norm": 12.051389937276333, "learning_rate": 1.952858763326812e-05, "loss": 0.7447, "step": 2405 }, { "epoch": 0.38, "grad_norm": 13.646038065046703, "learning_rate": 1.9528075820379353e-05, "loss": 0.732, "step": 2406 }, { "epoch": 0.38, "grad_norm": 19.79895137956602, "learning_rate": 1.9527563736517683e-05, "loss": 0.788, "step": 2407 }, { "epoch": 0.38, "grad_norm": 20.013451145876225, "learning_rate": 1.952705138169768e-05, "loss": 0.8152, "step": 2408 }, { "epoch": 0.38, "grad_norm": 16.993917869946696, "learning_rate": 1.95265387559339e-05, "loss": 0.7993, "step": 2409 }, { "epoch": 0.38, "grad_norm": 27.754758625511705, "learning_rate": 1.9526025859240936e-05, "loss": 0.8487, "step": 2410 }, { "epoch": 0.38, "grad_norm": 15.465520094814222, "learning_rate": 1.952551269163337e-05, "loss": 0.9287, "step": 2411 }, { "epoch": 0.38, "grad_norm": 19.208621756966533, "learning_rate": 1.9524999253125792e-05, "loss": 0.7842, "step": 2412 }, { "epoch": 0.38, "grad_norm": 22.480105149736534, "learning_rate": 1.952448554373281e-05, "loss": 0.9469, "step": 2413 }, { "epoch": 0.38, "grad_norm": 21.46800276512284, "learning_rate": 1.9523971563469026e-05, "loss": 0.9319, "step": 2414 }, { "epoch": 0.38, "grad_norm": 28.447086995510453, "learning_rate": 1.9523457312349067e-05, "loss": 0.832, "step": 2415 }, { "epoch": 0.38, "grad_norm": 30.189566063717486, "learning_rate": 1.9522942790387552e-05, "loss": 0.9581, "step": 2416 }, { "epoch": 0.38, "grad_norm": 14.866945686719145, "learning_rate": 1.9522427997599113e-05, "loss": 0.7988, "step": 2417 }, { "epoch": 0.38, "grad_norm": 21.18456878190828, "learning_rate": 1.9521912933998392e-05, "loss": 0.9309, "step": 2418 }, { "epoch": 0.38, "grad_norm": 24.084679660748574, "learning_rate": 1.9521397599600042e-05, "loss": 0.788, "step": 2419 }, { "epoch": 0.38, "grad_norm": 24.8868244206621, "learning_rate": 1.952088199441871e-05, "loss": 0.8429, "step": 2420 }, { "epoch": 0.38, "grad_norm": 20.5815178602056, "learning_rate": 1.9520366118469064e-05, "loss": 0.9069, "step": 2421 }, { "epoch": 0.38, "grad_norm": 19.245618282799214, "learning_rate": 1.9519849971765775e-05, "loss": 0.8289, "step": 2422 }, { "epoch": 0.38, "grad_norm": 19.110846508041238, "learning_rate": 1.951933355432352e-05, "loss": 0.9091, "step": 2423 }, { "epoch": 0.38, "grad_norm": 18.453560795753642, "learning_rate": 1.9518816866156994e-05, "loss": 0.844, "step": 2424 }, { "epoch": 0.38, "grad_norm": 23.633751101179815, "learning_rate": 1.951829990728088e-05, "loss": 0.8443, "step": 2425 }, { "epoch": 0.38, "grad_norm": 18.87846786728831, "learning_rate": 1.951778267770988e-05, "loss": 0.8341, "step": 2426 }, { "epoch": 0.38, "grad_norm": 21.67956006674215, "learning_rate": 1.9517265177458716e-05, "loss": 0.8208, "step": 2427 }, { "epoch": 0.38, "grad_norm": 27.744037822142193, "learning_rate": 1.9516747406542097e-05, "loss": 0.789, "step": 2428 }, { "epoch": 0.38, "grad_norm": 19.707457612097446, "learning_rate": 1.9516229364974748e-05, "loss": 0.8049, "step": 2429 }, { "epoch": 0.38, "grad_norm": 21.061050355247907, "learning_rate": 1.95157110527714e-05, "loss": 0.9428, "step": 2430 }, { "epoch": 0.38, "grad_norm": 17.28179708653034, "learning_rate": 1.95151924699468e-05, "loss": 0.8558, "step": 2431 }, { "epoch": 0.38, "grad_norm": 18.939199239990796, "learning_rate": 1.9514673616515686e-05, "loss": 0.8189, "step": 2432 }, { "epoch": 0.38, "grad_norm": 25.907680305183728, "learning_rate": 1.9514154492492827e-05, "loss": 0.7942, "step": 2433 }, { "epoch": 0.38, "grad_norm": 23.527868800517016, "learning_rate": 1.9513635097892975e-05, "loss": 0.9862, "step": 2434 }, { "epoch": 0.38, "grad_norm": 26.291971377888153, "learning_rate": 1.951311543273091e-05, "loss": 0.859, "step": 2435 }, { "epoch": 0.38, "grad_norm": 21.189825695882313, "learning_rate": 1.9512595497021406e-05, "loss": 0.8974, "step": 2436 }, { "epoch": 0.38, "grad_norm": 21.843778903543672, "learning_rate": 1.951207529077925e-05, "loss": 0.8477, "step": 2437 }, { "epoch": 0.38, "grad_norm": 27.592198517817323, "learning_rate": 1.9511554814019237e-05, "loss": 0.8692, "step": 2438 }, { "epoch": 0.38, "grad_norm": 29.270361096022043, "learning_rate": 1.951103406675617e-05, "loss": 0.8998, "step": 2439 }, { "epoch": 0.38, "grad_norm": 27.333147997392555, "learning_rate": 1.951051304900486e-05, "loss": 0.8525, "step": 2440 }, { "epoch": 0.38, "grad_norm": 16.34717099325237, "learning_rate": 1.9509991760780117e-05, "loss": 0.8892, "step": 2441 }, { "epoch": 0.38, "grad_norm": 23.99948084662135, "learning_rate": 1.9509470202096774e-05, "loss": 0.8871, "step": 2442 }, { "epoch": 0.38, "grad_norm": 19.181937820280993, "learning_rate": 1.9508948372969664e-05, "loss": 0.9362, "step": 2443 }, { "epoch": 0.38, "grad_norm": 21.371574580522392, "learning_rate": 1.950842627341362e-05, "loss": 0.8514, "step": 2444 }, { "epoch": 0.38, "grad_norm": 21.015313346849275, "learning_rate": 1.95079039034435e-05, "loss": 0.8166, "step": 2445 }, { "epoch": 0.38, "grad_norm": 22.168714169644574, "learning_rate": 1.9507381263074153e-05, "loss": 0.7802, "step": 2446 }, { "epoch": 0.38, "grad_norm": 17.710394869090457, "learning_rate": 1.9506858352320444e-05, "loss": 0.8928, "step": 2447 }, { "epoch": 0.38, "grad_norm": 24.27983543026391, "learning_rate": 1.9506335171197243e-05, "loss": 0.8349, "step": 2448 }, { "epoch": 0.38, "grad_norm": 16.263049157155898, "learning_rate": 1.9505811719719435e-05, "loss": 0.8699, "step": 2449 }, { "epoch": 0.38, "grad_norm": 12.720897475693256, "learning_rate": 1.95052879979019e-05, "loss": 0.8266, "step": 2450 }, { "epoch": 0.38, "grad_norm": 15.566849225297409, "learning_rate": 1.9504764005759534e-05, "loss": 0.8193, "step": 2451 }, { "epoch": 0.38, "grad_norm": 14.10746439220069, "learning_rate": 1.950423974330724e-05, "loss": 0.7747, "step": 2452 }, { "epoch": 0.38, "grad_norm": 26.526712596026236, "learning_rate": 1.9503715210559928e-05, "loss": 0.8887, "step": 2453 }, { "epoch": 0.38, "grad_norm": 20.950628753504276, "learning_rate": 1.9503190407532514e-05, "loss": 0.8675, "step": 2454 }, { "epoch": 0.38, "grad_norm": 16.83666826684625, "learning_rate": 1.9502665334239926e-05, "loss": 0.7574, "step": 2455 }, { "epoch": 0.38, "grad_norm": 21.02049736775882, "learning_rate": 1.9502139990697095e-05, "loss": 0.8922, "step": 2456 }, { "epoch": 0.38, "grad_norm": 22.486613397471082, "learning_rate": 1.9501614376918963e-05, "loss": 0.8107, "step": 2457 }, { "epoch": 0.38, "grad_norm": 25.76034383741656, "learning_rate": 1.9501088492920472e-05, "loss": 0.7128, "step": 2458 }, { "epoch": 0.38, "grad_norm": 21.59133177161005, "learning_rate": 1.9500562338716584e-05, "loss": 0.892, "step": 2459 }, { "epoch": 0.38, "grad_norm": 14.671197644036523, "learning_rate": 1.9500035914322263e-05, "loss": 0.7916, "step": 2460 }, { "epoch": 0.38, "grad_norm": 24.787109359369694, "learning_rate": 1.9499509219752473e-05, "loss": 0.8417, "step": 2461 }, { "epoch": 0.38, "grad_norm": 15.881299648109724, "learning_rate": 1.94989822550222e-05, "loss": 0.8222, "step": 2462 }, { "epoch": 0.38, "grad_norm": 31.415744304819388, "learning_rate": 1.9498455020146428e-05, "loss": 0.8926, "step": 2463 }, { "epoch": 0.38, "grad_norm": 29.347522763218166, "learning_rate": 1.9497927515140155e-05, "loss": 0.9163, "step": 2464 }, { "epoch": 0.39, "grad_norm": 14.668232378260434, "learning_rate": 1.9497399740018376e-05, "loss": 0.7106, "step": 2465 }, { "epoch": 0.39, "grad_norm": 19.630541805620698, "learning_rate": 1.9496871694796106e-05, "loss": 0.7943, "step": 2466 }, { "epoch": 0.39, "grad_norm": 23.891356055337113, "learning_rate": 1.949634337948836e-05, "loss": 0.8403, "step": 2467 }, { "epoch": 0.39, "grad_norm": 16.755547365964734, "learning_rate": 1.9495814794110165e-05, "loss": 0.6969, "step": 2468 }, { "epoch": 0.39, "grad_norm": 21.946137380188254, "learning_rate": 1.949528593867655e-05, "loss": 0.8803, "step": 2469 }, { "epoch": 0.39, "grad_norm": 25.55641275230682, "learning_rate": 1.9494756813202558e-05, "loss": 0.8794, "step": 2470 }, { "epoch": 0.39, "grad_norm": 16.42324417424368, "learning_rate": 1.9494227417703237e-05, "loss": 0.8305, "step": 2471 }, { "epoch": 0.39, "grad_norm": 38.59207416980971, "learning_rate": 1.949369775219364e-05, "loss": 0.9413, "step": 2472 }, { "epoch": 0.39, "grad_norm": 16.52672246064103, "learning_rate": 1.9493167816688837e-05, "loss": 0.8335, "step": 2473 }, { "epoch": 0.39, "grad_norm": 37.97518438284426, "learning_rate": 1.949263761120389e-05, "loss": 0.9332, "step": 2474 }, { "epoch": 0.39, "grad_norm": 26.705251099045753, "learning_rate": 1.9492107135753884e-05, "loss": 0.7934, "step": 2475 }, { "epoch": 0.39, "grad_norm": 29.23079161444644, "learning_rate": 1.9491576390353906e-05, "loss": 0.9349, "step": 2476 }, { "epoch": 0.39, "grad_norm": 25.89753302696428, "learning_rate": 1.9491045375019043e-05, "loss": 0.8761, "step": 2477 }, { "epoch": 0.39, "grad_norm": 29.637885229490248, "learning_rate": 1.949051408976441e-05, "loss": 0.7511, "step": 2478 }, { "epoch": 0.39, "grad_norm": 16.165437200175504, "learning_rate": 1.9489982534605098e-05, "loss": 0.7051, "step": 2479 }, { "epoch": 0.39, "grad_norm": 23.88011490371271, "learning_rate": 1.948945070955624e-05, "loss": 0.7667, "step": 2480 }, { "epoch": 0.39, "grad_norm": 18.006794537053363, "learning_rate": 1.9488918614632953e-05, "loss": 0.6903, "step": 2481 }, { "epoch": 0.39, "grad_norm": 22.41213134631574, "learning_rate": 1.9488386249850372e-05, "loss": 0.8073, "step": 2482 }, { "epoch": 0.39, "grad_norm": 24.74164763735822, "learning_rate": 1.9487853615223636e-05, "loss": 0.8618, "step": 2483 }, { "epoch": 0.39, "grad_norm": 28.02158039608608, "learning_rate": 1.9487320710767898e-05, "loss": 0.9661, "step": 2484 }, { "epoch": 0.39, "grad_norm": 25.41546536046868, "learning_rate": 1.9486787536498304e-05, "loss": 0.8838, "step": 2485 }, { "epoch": 0.39, "grad_norm": 16.02531547412535, "learning_rate": 1.9486254092430022e-05, "loss": 0.9066, "step": 2486 }, { "epoch": 0.39, "grad_norm": 25.248648046662286, "learning_rate": 1.9485720378578225e-05, "loss": 0.8172, "step": 2487 }, { "epoch": 0.39, "grad_norm": 21.50523668204783, "learning_rate": 1.948518639495809e-05, "loss": 0.7969, "step": 2488 }, { "epoch": 0.39, "grad_norm": 34.12783320336865, "learning_rate": 1.94846521415848e-05, "loss": 0.9122, "step": 2489 }, { "epoch": 0.39, "grad_norm": 30.81649489649323, "learning_rate": 1.9484117618473553e-05, "loss": 0.8958, "step": 2490 }, { "epoch": 0.39, "grad_norm": 29.140569533430803, "learning_rate": 1.948358282563955e-05, "loss": 0.8575, "step": 2491 }, { "epoch": 0.39, "grad_norm": 17.109909442061948, "learning_rate": 1.9483047763097994e-05, "loss": 0.8475, "step": 2492 }, { "epoch": 0.39, "grad_norm": 29.933604053862425, "learning_rate": 1.9482512430864113e-05, "loss": 0.8546, "step": 2493 }, { "epoch": 0.39, "grad_norm": 21.782840037220737, "learning_rate": 1.9481976828953126e-05, "loss": 0.848, "step": 2494 }, { "epoch": 0.39, "grad_norm": 22.515427145793463, "learning_rate": 1.9481440957380258e-05, "loss": 0.8328, "step": 2495 }, { "epoch": 0.39, "grad_norm": 21.258223283970217, "learning_rate": 1.948090481616076e-05, "loss": 0.9235, "step": 2496 }, { "epoch": 0.39, "grad_norm": 17.32457242172472, "learning_rate": 1.9480368405309876e-05, "loss": 0.8043, "step": 2497 }, { "epoch": 0.39, "grad_norm": 28.00501584092179, "learning_rate": 1.9479831724842862e-05, "loss": 0.8732, "step": 2498 }, { "epoch": 0.39, "grad_norm": 15.98044102978033, "learning_rate": 1.9479294774774974e-05, "loss": 0.8157, "step": 2499 }, { "epoch": 0.39, "grad_norm": 20.99142914663275, "learning_rate": 1.9478757555121493e-05, "loss": 0.8674, "step": 2500 }, { "epoch": 0.39, "grad_norm": 15.768989350516703, "learning_rate": 1.9478220065897687e-05, "loss": 0.7731, "step": 2501 }, { "epoch": 0.39, "grad_norm": 21.897985533324526, "learning_rate": 1.947768230711885e-05, "loss": 0.9334, "step": 2502 }, { "epoch": 0.39, "grad_norm": 27.057438971284018, "learning_rate": 1.9477144278800275e-05, "loss": 0.8583, "step": 2503 }, { "epoch": 0.39, "grad_norm": 23.44126465952835, "learning_rate": 1.9476605980957258e-05, "loss": 0.8965, "step": 2504 }, { "epoch": 0.39, "grad_norm": 13.897611667484993, "learning_rate": 1.947606741360511e-05, "loss": 0.7554, "step": 2505 }, { "epoch": 0.39, "grad_norm": 32.64812153674515, "learning_rate": 1.9475528576759147e-05, "loss": 0.8532, "step": 2506 }, { "epoch": 0.39, "grad_norm": 33.427141424698995, "learning_rate": 1.9474989470434695e-05, "loss": 0.7965, "step": 2507 }, { "epoch": 0.39, "grad_norm": 16.03586781371496, "learning_rate": 1.9474450094647087e-05, "loss": 0.7687, "step": 2508 }, { "epoch": 0.39, "grad_norm": 22.1414637793719, "learning_rate": 1.947391044941166e-05, "loss": 0.7306, "step": 2509 }, { "epoch": 0.39, "grad_norm": 16.98787337028632, "learning_rate": 1.947337053474376e-05, "loss": 0.9005, "step": 2510 }, { "epoch": 0.39, "grad_norm": 13.887055852476287, "learning_rate": 1.9472830350658748e-05, "loss": 0.7767, "step": 2511 }, { "epoch": 0.39, "grad_norm": 17.22012925811569, "learning_rate": 1.9472289897171978e-05, "loss": 0.8296, "step": 2512 }, { "epoch": 0.39, "grad_norm": 27.639276062421043, "learning_rate": 1.9471749174298827e-05, "loss": 0.8719, "step": 2513 }, { "epoch": 0.39, "grad_norm": 13.703361496140543, "learning_rate": 1.9471208182054668e-05, "loss": 0.735, "step": 2514 }, { "epoch": 0.39, "grad_norm": 14.538841045503059, "learning_rate": 1.947066692045489e-05, "loss": 0.8326, "step": 2515 }, { "epoch": 0.39, "grad_norm": 16.81425147168844, "learning_rate": 1.9470125389514884e-05, "loss": 0.813, "step": 2516 }, { "epoch": 0.39, "grad_norm": 19.290948836591507, "learning_rate": 1.9469583589250055e-05, "loss": 0.8913, "step": 2517 }, { "epoch": 0.39, "grad_norm": 20.048454680311696, "learning_rate": 1.9469041519675806e-05, "loss": 0.8645, "step": 2518 }, { "epoch": 0.39, "grad_norm": 24.305522683022993, "learning_rate": 1.9468499180807554e-05, "loss": 0.8648, "step": 2519 }, { "epoch": 0.39, "grad_norm": 19.836333264507807, "learning_rate": 1.9467956572660727e-05, "loss": 0.7702, "step": 2520 }, { "epoch": 0.39, "grad_norm": 20.05402738118547, "learning_rate": 1.9467413695250753e-05, "loss": 0.9407, "step": 2521 }, { "epoch": 0.39, "grad_norm": 18.635108099384144, "learning_rate": 1.946687054859307e-05, "loss": 0.9345, "step": 2522 }, { "epoch": 0.39, "grad_norm": 29.09553264643005, "learning_rate": 1.9466327132703128e-05, "loss": 0.8608, "step": 2523 }, { "epoch": 0.39, "grad_norm": 20.853735737980077, "learning_rate": 1.946578344759638e-05, "loss": 0.8349, "step": 2524 }, { "epoch": 0.39, "grad_norm": 21.318348150735257, "learning_rate": 1.946523949328829e-05, "loss": 0.9876, "step": 2525 }, { "epoch": 0.39, "grad_norm": 17.541723103066253, "learning_rate": 1.9464695269794322e-05, "loss": 0.8498, "step": 2526 }, { "epoch": 0.39, "grad_norm": 18.278090788039147, "learning_rate": 1.9464150777129956e-05, "loss": 0.848, "step": 2527 }, { "epoch": 0.39, "grad_norm": 17.57808443785332, "learning_rate": 1.9463606015310684e-05, "loss": 0.8322, "step": 2528 }, { "epoch": 0.4, "grad_norm": 24.491282169821478, "learning_rate": 1.9463060984351988e-05, "loss": 0.8076, "step": 2529 }, { "epoch": 0.4, "grad_norm": 19.797416678848354, "learning_rate": 1.946251568426938e-05, "loss": 0.8709, "step": 2530 }, { "epoch": 0.4, "grad_norm": 24.77472292459336, "learning_rate": 1.9461970115078356e-05, "loss": 0.8743, "step": 2531 }, { "epoch": 0.4, "grad_norm": 25.31539705440156, "learning_rate": 1.9461424276794435e-05, "loss": 0.8091, "step": 2532 }, { "epoch": 0.4, "grad_norm": 28.366568187486834, "learning_rate": 1.9460878169433147e-05, "loss": 1.0113, "step": 2533 }, { "epoch": 0.4, "grad_norm": 32.48096707042649, "learning_rate": 1.9460331793010018e-05, "loss": 0.9509, "step": 2534 }, { "epoch": 0.4, "grad_norm": 20.78370720239767, "learning_rate": 1.9459785147540585e-05, "loss": 0.9013, "step": 2535 }, { "epoch": 0.4, "grad_norm": 16.215463192254308, "learning_rate": 1.9459238233040393e-05, "loss": 0.7608, "step": 2536 }, { "epoch": 0.4, "grad_norm": 29.078353551943156, "learning_rate": 1.9458691049525003e-05, "loss": 0.7252, "step": 2537 }, { "epoch": 0.4, "grad_norm": 23.396372730432347, "learning_rate": 1.9458143597009974e-05, "loss": 0.7558, "step": 2538 }, { "epoch": 0.4, "grad_norm": 21.206016682786856, "learning_rate": 1.9457595875510874e-05, "loss": 0.8054, "step": 2539 }, { "epoch": 0.4, "grad_norm": 14.420907792507922, "learning_rate": 1.9457047885043275e-05, "loss": 0.7454, "step": 2540 }, { "epoch": 0.4, "grad_norm": 31.388612438205207, "learning_rate": 1.945649962562277e-05, "loss": 0.8577, "step": 2541 }, { "epoch": 0.4, "grad_norm": 19.179107481660534, "learning_rate": 1.9455951097264947e-05, "loss": 0.8072, "step": 2542 }, { "epoch": 0.4, "grad_norm": 27.118587854484836, "learning_rate": 1.9455402299985408e-05, "loss": 0.8226, "step": 2543 }, { "epoch": 0.4, "grad_norm": 34.098309812510784, "learning_rate": 1.9454853233799756e-05, "loss": 0.8972, "step": 2544 }, { "epoch": 0.4, "grad_norm": 16.84101018093514, "learning_rate": 1.9454303898723613e-05, "loss": 0.8312, "step": 2545 }, { "epoch": 0.4, "grad_norm": 16.21854358745712, "learning_rate": 1.9453754294772593e-05, "loss": 0.8243, "step": 2546 }, { "epoch": 0.4, "grad_norm": 27.23783987916674, "learning_rate": 1.9453204421962334e-05, "loss": 0.846, "step": 2547 }, { "epoch": 0.4, "grad_norm": 22.92782513595387, "learning_rate": 1.945265428030847e-05, "loss": 0.8237, "step": 2548 }, { "epoch": 0.4, "grad_norm": 20.45996879303278, "learning_rate": 1.945210386982665e-05, "loss": 0.9023, "step": 2549 }, { "epoch": 0.4, "grad_norm": 16.74723628226207, "learning_rate": 1.9451553190532523e-05, "loss": 0.7803, "step": 2550 }, { "epoch": 0.4, "grad_norm": 21.662748966228154, "learning_rate": 1.945100224244175e-05, "loss": 0.8573, "step": 2551 }, { "epoch": 0.4, "grad_norm": 26.167088374355984, "learning_rate": 1.9450451025570006e-05, "loss": 0.8045, "step": 2552 }, { "epoch": 0.4, "grad_norm": 17.369647147470108, "learning_rate": 1.9449899539932962e-05, "loss": 0.8749, "step": 2553 }, { "epoch": 0.4, "grad_norm": 14.154262547420428, "learning_rate": 1.9449347785546303e-05, "loss": 0.8593, "step": 2554 }, { "epoch": 0.4, "grad_norm": 22.46712800738409, "learning_rate": 1.9448795762425722e-05, "loss": 0.8296, "step": 2555 }, { "epoch": 0.4, "grad_norm": 17.195440759834735, "learning_rate": 1.9448243470586913e-05, "loss": 0.8126, "step": 2556 }, { "epoch": 0.4, "grad_norm": 16.662371809438405, "learning_rate": 1.944769091004559e-05, "loss": 0.8672, "step": 2557 }, { "epoch": 0.4, "grad_norm": 17.915135451291725, "learning_rate": 1.9447138080817466e-05, "loss": 0.8059, "step": 2558 }, { "epoch": 0.4, "grad_norm": 33.48498326513195, "learning_rate": 1.9446584982918256e-05, "loss": 0.9164, "step": 2559 }, { "epoch": 0.4, "grad_norm": 22.587054317708027, "learning_rate": 1.9446031616363702e-05, "loss": 0.9321, "step": 2560 }, { "epoch": 0.4, "grad_norm": 19.981016468882483, "learning_rate": 1.9445477981169528e-05, "loss": 0.7125, "step": 2561 }, { "epoch": 0.4, "grad_norm": 12.959233508270154, "learning_rate": 1.944492407735149e-05, "loss": 0.7989, "step": 2562 }, { "epoch": 0.4, "grad_norm": 21.3023220892182, "learning_rate": 1.9444369904925337e-05, "loss": 0.9101, "step": 2563 }, { "epoch": 0.4, "grad_norm": 28.232542782831064, "learning_rate": 1.9443815463906826e-05, "loss": 0.8185, "step": 2564 }, { "epoch": 0.4, "grad_norm": 18.32370665246752, "learning_rate": 1.9443260754311728e-05, "loss": 0.8727, "step": 2565 }, { "epoch": 0.4, "grad_norm": 23.90503022940415, "learning_rate": 1.944270577615582e-05, "loss": 0.8984, "step": 2566 }, { "epoch": 0.4, "grad_norm": 17.48925276003296, "learning_rate": 1.944215052945488e-05, "loss": 0.8324, "step": 2567 }, { "epoch": 0.4, "grad_norm": 25.314506234030222, "learning_rate": 1.9441595014224702e-05, "loss": 0.8118, "step": 2568 }, { "epoch": 0.4, "grad_norm": 15.248241293835397, "learning_rate": 1.944103923048109e-05, "loss": 0.8533, "step": 2569 }, { "epoch": 0.4, "grad_norm": 21.064490976242134, "learning_rate": 1.944048317823984e-05, "loss": 0.8003, "step": 2570 }, { "epoch": 0.4, "grad_norm": 23.62154523721757, "learning_rate": 1.9439926857516777e-05, "loss": 0.8221, "step": 2571 }, { "epoch": 0.4, "grad_norm": 21.815765695892782, "learning_rate": 1.9439370268327712e-05, "loss": 0.9321, "step": 2572 }, { "epoch": 0.4, "grad_norm": 17.541433321944204, "learning_rate": 1.9438813410688478e-05, "loss": 0.7801, "step": 2573 }, { "epoch": 0.4, "grad_norm": 16.184646929430254, "learning_rate": 1.9438256284614915e-05, "loss": 0.8541, "step": 2574 }, { "epoch": 0.4, "grad_norm": 20.487637588092507, "learning_rate": 1.943769889012286e-05, "loss": 0.9258, "step": 2575 }, { "epoch": 0.4, "grad_norm": 24.381051461654067, "learning_rate": 1.9437141227228175e-05, "loss": 0.8346, "step": 2576 }, { "epoch": 0.4, "grad_norm": 27.422308540720923, "learning_rate": 1.943658329594671e-05, "loss": 0.9246, "step": 2577 }, { "epoch": 0.4, "grad_norm": 16.992872255931847, "learning_rate": 1.9436025096294337e-05, "loss": 0.8081, "step": 2578 }, { "epoch": 0.4, "grad_norm": 14.661583718384255, "learning_rate": 1.9435466628286933e-05, "loss": 0.7142, "step": 2579 }, { "epoch": 0.4, "grad_norm": 13.723917551601362, "learning_rate": 1.9434907891940376e-05, "loss": 0.7743, "step": 2580 }, { "epoch": 0.4, "grad_norm": 23.244948795001786, "learning_rate": 1.943434888727056e-05, "loss": 0.8379, "step": 2581 }, { "epoch": 0.4, "grad_norm": 22.210386887312627, "learning_rate": 1.9433789614293376e-05, "loss": 0.8381, "step": 2582 }, { "epoch": 0.4, "grad_norm": 15.886106102683641, "learning_rate": 1.9433230073024737e-05, "loss": 0.8676, "step": 2583 }, { "epoch": 0.4, "grad_norm": 18.00643830713997, "learning_rate": 1.943267026348055e-05, "loss": 0.7641, "step": 2584 }, { "epoch": 0.4, "grad_norm": 19.577011405528797, "learning_rate": 1.943211018567674e-05, "loss": 0.745, "step": 2585 }, { "epoch": 0.4, "grad_norm": 18.638061927585586, "learning_rate": 1.9431549839629235e-05, "loss": 0.7064, "step": 2586 }, { "epoch": 0.4, "grad_norm": 19.29325120004328, "learning_rate": 1.943098922535397e-05, "loss": 0.7488, "step": 2587 }, { "epoch": 0.4, "grad_norm": 22.640773912233634, "learning_rate": 1.9430428342866888e-05, "loss": 0.7827, "step": 2588 }, { "epoch": 0.4, "grad_norm": 18.03750906162206, "learning_rate": 1.942986719218394e-05, "loss": 0.7597, "step": 2589 }, { "epoch": 0.4, "grad_norm": 18.373635234477586, "learning_rate": 1.9429305773321085e-05, "loss": 0.834, "step": 2590 }, { "epoch": 0.4, "grad_norm": 26.200292549633932, "learning_rate": 1.9428744086294293e-05, "loss": 0.8667, "step": 2591 }, { "epoch": 0.4, "grad_norm": 23.750456096009135, "learning_rate": 1.942818213111953e-05, "loss": 0.8944, "step": 2592 }, { "epoch": 0.41, "grad_norm": 21.175852747481397, "learning_rate": 1.9427619907812788e-05, "loss": 0.8616, "step": 2593 }, { "epoch": 0.41, "grad_norm": 21.656538642126716, "learning_rate": 1.9427057416390048e-05, "loss": 0.8284, "step": 2594 }, { "epoch": 0.41, "grad_norm": 24.168777550937367, "learning_rate": 1.942649465686731e-05, "loss": 0.9372, "step": 2595 }, { "epoch": 0.41, "grad_norm": 21.447199351089612, "learning_rate": 1.9425931629260578e-05, "loss": 0.8351, "step": 2596 }, { "epoch": 0.41, "grad_norm": 23.96381596581886, "learning_rate": 1.9425368333585862e-05, "loss": 0.8431, "step": 2597 }, { "epoch": 0.41, "grad_norm": 18.22198750003726, "learning_rate": 1.9424804769859188e-05, "loss": 0.8937, "step": 2598 }, { "epoch": 0.41, "grad_norm": 19.90344687896506, "learning_rate": 1.942424093809658e-05, "loss": 0.8646, "step": 2599 }, { "epoch": 0.41, "grad_norm": 18.041467011567903, "learning_rate": 1.9423676838314067e-05, "loss": 0.8619, "step": 2600 }, { "epoch": 0.41, "grad_norm": 22.060994326514724, "learning_rate": 1.9423112470527703e-05, "loss": 0.9244, "step": 2601 }, { "epoch": 0.41, "grad_norm": 27.994162290491946, "learning_rate": 1.9422547834753532e-05, "loss": 0.9067, "step": 2602 }, { "epoch": 0.41, "grad_norm": 17.774520219794255, "learning_rate": 1.942198293100761e-05, "loss": 0.9444, "step": 2603 }, { "epoch": 0.41, "grad_norm": 15.34939937930872, "learning_rate": 1.9421417759306006e-05, "loss": 0.8641, "step": 2604 }, { "epoch": 0.41, "grad_norm": 19.431907047704055, "learning_rate": 1.9420852319664797e-05, "loss": 0.8796, "step": 2605 }, { "epoch": 0.41, "grad_norm": 28.351289316323932, "learning_rate": 1.9420286612100054e-05, "loss": 0.8465, "step": 2606 }, { "epoch": 0.41, "grad_norm": 16.141287664490484, "learning_rate": 1.9419720636627874e-05, "loss": 0.7885, "step": 2607 }, { "epoch": 0.41, "grad_norm": 14.238178940393663, "learning_rate": 1.9419154393264346e-05, "loss": 0.7726, "step": 2608 }, { "epoch": 0.41, "grad_norm": 16.774143510626377, "learning_rate": 1.9418587882025578e-05, "loss": 0.8392, "step": 2609 }, { "epoch": 0.41, "grad_norm": 18.471464637756934, "learning_rate": 1.941802110292768e-05, "loss": 0.8838, "step": 2610 }, { "epoch": 0.41, "grad_norm": 17.16966640337463, "learning_rate": 1.941745405598677e-05, "loss": 0.6908, "step": 2611 }, { "epoch": 0.41, "grad_norm": 15.489898343229815, "learning_rate": 1.941688674121898e-05, "loss": 0.8888, "step": 2612 }, { "epoch": 0.41, "grad_norm": 16.863847662501758, "learning_rate": 1.941631915864044e-05, "loss": 0.7215, "step": 2613 }, { "epoch": 0.41, "grad_norm": 21.77807440109385, "learning_rate": 1.9415751308267284e-05, "loss": 0.8139, "step": 2614 }, { "epoch": 0.41, "grad_norm": 18.827306565822177, "learning_rate": 1.9415183190115678e-05, "loss": 0.8615, "step": 2615 }, { "epoch": 0.41, "grad_norm": 27.65043099442133, "learning_rate": 1.9414614804201764e-05, "loss": 0.8808, "step": 2616 }, { "epoch": 0.41, "grad_norm": 20.045503889813485, "learning_rate": 1.9414046150541712e-05, "loss": 0.8402, "step": 2617 }, { "epoch": 0.41, "grad_norm": 24.587246551792816, "learning_rate": 1.94134772291517e-05, "loss": 0.7904, "step": 2618 }, { "epoch": 0.41, "grad_norm": 21.43876005868107, "learning_rate": 1.9412908040047894e-05, "loss": 0.8568, "step": 2619 }, { "epoch": 0.41, "grad_norm": 16.123789396234184, "learning_rate": 1.94123385832465e-05, "loss": 0.8212, "step": 2620 }, { "epoch": 0.41, "grad_norm": 17.576751786741397, "learning_rate": 1.9411768858763695e-05, "loss": 0.8884, "step": 2621 }, { "epoch": 0.41, "grad_norm": 13.457738781891761, "learning_rate": 1.941119886661569e-05, "loss": 0.7221, "step": 2622 }, { "epoch": 0.41, "grad_norm": 21.87818287717947, "learning_rate": 1.9410628606818696e-05, "loss": 0.7998, "step": 2623 }, { "epoch": 0.41, "grad_norm": 26.376370772903343, "learning_rate": 1.9410058079388933e-05, "loss": 0.8713, "step": 2624 }, { "epoch": 0.41, "grad_norm": 22.514787003821464, "learning_rate": 1.940948728434262e-05, "loss": 0.7811, "step": 2625 }, { "epoch": 0.41, "grad_norm": 18.618198144833002, "learning_rate": 1.940891622169599e-05, "loss": 0.8024, "step": 2626 }, { "epoch": 0.41, "grad_norm": 19.111914775500992, "learning_rate": 1.940834489146529e-05, "loss": 0.8077, "step": 2627 }, { "epoch": 0.41, "grad_norm": 13.8638801579698, "learning_rate": 1.9407773293666764e-05, "loss": 0.813, "step": 2628 }, { "epoch": 0.41, "grad_norm": 15.381499617072896, "learning_rate": 1.940720142831667e-05, "loss": 0.8414, "step": 2629 }, { "epoch": 0.41, "grad_norm": 5.59328042022629, "learning_rate": 1.940662929543127e-05, "loss": 0.8662, "step": 2630 }, { "epoch": 0.41, "grad_norm": 20.48884477740405, "learning_rate": 1.9406056895026837e-05, "loss": 0.887, "step": 2631 }, { "epoch": 0.41, "grad_norm": 24.64496365595773, "learning_rate": 1.9405484227119646e-05, "loss": 0.7966, "step": 2632 }, { "epoch": 0.41, "grad_norm": 18.36209585684469, "learning_rate": 1.9404911291725985e-05, "loss": 0.8656, "step": 2633 }, { "epoch": 0.41, "grad_norm": 13.859547374754367, "learning_rate": 1.9404338088862152e-05, "loss": 0.8752, "step": 2634 }, { "epoch": 0.41, "grad_norm": 14.376115107822649, "learning_rate": 1.940376461854444e-05, "loss": 0.827, "step": 2635 }, { "epoch": 0.41, "grad_norm": 23.630232947310596, "learning_rate": 1.940319088078917e-05, "loss": 0.826, "step": 2636 }, { "epoch": 0.41, "grad_norm": 25.305490575839116, "learning_rate": 1.9402616875612645e-05, "loss": 0.8631, "step": 2637 }, { "epoch": 0.41, "grad_norm": 42.794256984693206, "learning_rate": 1.9402042603031202e-05, "loss": 0.8284, "step": 2638 }, { "epoch": 0.41, "grad_norm": 46.730669959461885, "learning_rate": 1.9401468063061165e-05, "loss": 0.7978, "step": 2639 }, { "epoch": 0.41, "grad_norm": 23.69849547185854, "learning_rate": 1.9400893255718874e-05, "loss": 0.852, "step": 2640 }, { "epoch": 0.41, "grad_norm": 20.864571354155228, "learning_rate": 1.9400318181020682e-05, "loss": 0.7427, "step": 2641 }, { "epoch": 0.41, "grad_norm": 22.975606694813866, "learning_rate": 1.9399742838982937e-05, "loss": 0.8502, "step": 2642 }, { "epoch": 0.41, "grad_norm": 23.738481673925747, "learning_rate": 1.9399167229622008e-05, "loss": 0.8133, "step": 2643 }, { "epoch": 0.41, "grad_norm": 18.74722582957297, "learning_rate": 1.9398591352954258e-05, "loss": 0.8737, "step": 2644 }, { "epoch": 0.41, "grad_norm": 21.801209265631115, "learning_rate": 1.939801520899607e-05, "loss": 0.8404, "step": 2645 }, { "epoch": 0.41, "grad_norm": 18.29072114103525, "learning_rate": 1.9397438797763825e-05, "loss": 0.8134, "step": 2646 }, { "epoch": 0.41, "grad_norm": 14.701624742579497, "learning_rate": 1.9396862119273918e-05, "loss": 0.805, "step": 2647 }, { "epoch": 0.41, "grad_norm": 15.658906172625787, "learning_rate": 1.939628517354275e-05, "loss": 0.8634, "step": 2648 }, { "epoch": 0.41, "grad_norm": 15.305061373737125, "learning_rate": 1.939570796058673e-05, "loss": 0.8961, "step": 2649 }, { "epoch": 0.41, "grad_norm": 21.0478111361923, "learning_rate": 1.939513048042227e-05, "loss": 0.8788, "step": 2650 }, { "epoch": 0.41, "grad_norm": 22.228470704640216, "learning_rate": 1.9394552733065797e-05, "loss": 0.8253, "step": 2651 }, { "epoch": 0.41, "grad_norm": 23.60915031174712, "learning_rate": 1.939397471853374e-05, "loss": 0.8131, "step": 2652 }, { "epoch": 0.41, "grad_norm": 17.542882152081166, "learning_rate": 1.9393396436842537e-05, "loss": 0.8406, "step": 2653 }, { "epoch": 0.41, "grad_norm": 22.779320073894784, "learning_rate": 1.9392817888008632e-05, "loss": 0.8619, "step": 2654 }, { "epoch": 0.41, "grad_norm": 22.323311619922016, "learning_rate": 1.939223907204848e-05, "loss": 0.8435, "step": 2655 }, { "epoch": 0.41, "grad_norm": 16.03902363482122, "learning_rate": 1.9391659988978546e-05, "loss": 0.8012, "step": 2656 }, { "epoch": 0.42, "grad_norm": 18.83149713891515, "learning_rate": 1.9391080638815295e-05, "loss": 0.7867, "step": 2657 }, { "epoch": 0.42, "grad_norm": 20.66758118725523, "learning_rate": 1.9390501021575203e-05, "loss": 0.8648, "step": 2658 }, { "epoch": 0.42, "grad_norm": 15.167484054660793, "learning_rate": 1.938992113727476e-05, "loss": 0.7431, "step": 2659 }, { "epoch": 0.42, "grad_norm": 21.54112188354704, "learning_rate": 1.9389340985930447e-05, "loss": 0.8308, "step": 2660 }, { "epoch": 0.42, "grad_norm": 22.94534245710849, "learning_rate": 1.938876056755877e-05, "loss": 0.8617, "step": 2661 }, { "epoch": 0.42, "grad_norm": 17.41685642053697, "learning_rate": 1.9388179882176237e-05, "loss": 0.7977, "step": 2662 }, { "epoch": 0.42, "grad_norm": 19.65406110649243, "learning_rate": 1.938759892979936e-05, "loss": 0.9213, "step": 2663 }, { "epoch": 0.42, "grad_norm": 16.703268422098926, "learning_rate": 1.9387017710444662e-05, "loss": 0.8146, "step": 2664 }, { "epoch": 0.42, "grad_norm": 15.150328875169098, "learning_rate": 1.9386436224128668e-05, "loss": 0.7223, "step": 2665 }, { "epoch": 0.42, "grad_norm": 18.439974196960556, "learning_rate": 1.938585447086792e-05, "loss": 0.8488, "step": 2666 }, { "epoch": 0.42, "grad_norm": 15.982365029709545, "learning_rate": 1.9385272450678966e-05, "loss": 0.7891, "step": 2667 }, { "epoch": 0.42, "grad_norm": 18.709679270988122, "learning_rate": 1.938469016357835e-05, "loss": 0.7952, "step": 2668 }, { "epoch": 0.42, "grad_norm": 23.691875524818492, "learning_rate": 1.938410760958263e-05, "loss": 0.8298, "step": 2669 }, { "epoch": 0.42, "grad_norm": 14.818480974198497, "learning_rate": 1.9383524788708387e-05, "loss": 0.8908, "step": 2670 }, { "epoch": 0.42, "grad_norm": 18.902835379961296, "learning_rate": 1.9382941700972188e-05, "loss": 0.7914, "step": 2671 }, { "epoch": 0.42, "grad_norm": 18.726659806356395, "learning_rate": 1.938235834639061e-05, "loss": 0.8642, "step": 2672 }, { "epoch": 0.42, "grad_norm": 28.178781863414006, "learning_rate": 1.9381774724980253e-05, "loss": 0.8986, "step": 2673 }, { "epoch": 0.42, "grad_norm": 21.916274957607293, "learning_rate": 1.9381190836757712e-05, "loss": 0.7313, "step": 2674 }, { "epoch": 0.42, "grad_norm": 15.844817039102114, "learning_rate": 1.938060668173959e-05, "loss": 0.8021, "step": 2675 }, { "epoch": 0.42, "grad_norm": 19.15385254455125, "learning_rate": 1.93800222599425e-05, "loss": 0.8223, "step": 2676 }, { "epoch": 0.42, "grad_norm": 25.65244877768913, "learning_rate": 1.9379437571383067e-05, "loss": 0.7941, "step": 2677 }, { "epoch": 0.42, "grad_norm": 19.381240955530274, "learning_rate": 1.9378852616077915e-05, "loss": 0.8056, "step": 2678 }, { "epoch": 0.42, "grad_norm": 20.230558983335847, "learning_rate": 1.9378267394043678e-05, "loss": 0.8657, "step": 2679 }, { "epoch": 0.42, "grad_norm": 23.617910331165994, "learning_rate": 1.9377681905297007e-05, "loss": 0.9116, "step": 2680 }, { "epoch": 0.42, "grad_norm": 22.23549219837821, "learning_rate": 1.937709614985455e-05, "loss": 0.9101, "step": 2681 }, { "epoch": 0.42, "grad_norm": 23.55923682796077, "learning_rate": 1.937651012773296e-05, "loss": 0.8404, "step": 2682 }, { "epoch": 0.42, "grad_norm": 24.6342227567183, "learning_rate": 1.9375923838948907e-05, "loss": 0.8992, "step": 2683 }, { "epoch": 0.42, "grad_norm": 15.826724038442235, "learning_rate": 1.9375337283519067e-05, "loss": 0.7664, "step": 2684 }, { "epoch": 0.42, "grad_norm": 17.281880268499513, "learning_rate": 1.9374750461460117e-05, "loss": 0.7582, "step": 2685 }, { "epoch": 0.42, "grad_norm": 17.387846112902928, "learning_rate": 1.9374163372788748e-05, "loss": 0.8163, "step": 2686 }, { "epoch": 0.42, "grad_norm": 21.980031414030144, "learning_rate": 1.9373576017521657e-05, "loss": 0.8424, "step": 2687 }, { "epoch": 0.42, "grad_norm": 18.226036465604434, "learning_rate": 1.9372988395675547e-05, "loss": 0.7509, "step": 2688 }, { "epoch": 0.42, "grad_norm": 17.09676331821948, "learning_rate": 1.9372400507267132e-05, "loss": 0.9732, "step": 2689 }, { "epoch": 0.42, "grad_norm": 20.227840404579943, "learning_rate": 1.9371812352313125e-05, "loss": 0.8197, "step": 2690 }, { "epoch": 0.42, "grad_norm": 16.38703806842231, "learning_rate": 1.9371223930830264e-05, "loss": 0.848, "step": 2691 }, { "epoch": 0.42, "grad_norm": 30.472192629947614, "learning_rate": 1.9370635242835272e-05, "loss": 1.0408, "step": 2692 }, { "epoch": 0.42, "grad_norm": 18.427443078613916, "learning_rate": 1.9370046288344894e-05, "loss": 0.784, "step": 2693 }, { "epoch": 0.42, "grad_norm": 22.309832968515117, "learning_rate": 1.9369457067375884e-05, "loss": 0.8382, "step": 2694 }, { "epoch": 0.42, "grad_norm": 15.275171236141343, "learning_rate": 1.9368867579944994e-05, "loss": 0.7785, "step": 2695 }, { "epoch": 0.42, "grad_norm": 15.936757177918732, "learning_rate": 1.936827782606899e-05, "loss": 0.802, "step": 2696 }, { "epoch": 0.42, "grad_norm": 21.114803473183333, "learning_rate": 1.9367687805764647e-05, "loss": 0.9395, "step": 2697 }, { "epoch": 0.42, "grad_norm": 22.621051054466136, "learning_rate": 1.936709751904874e-05, "loss": 0.8338, "step": 2698 }, { "epoch": 0.42, "grad_norm": 22.16405750203337, "learning_rate": 1.936650696593806e-05, "loss": 0.9549, "step": 2699 }, { "epoch": 0.42, "grad_norm": 17.09112454742946, "learning_rate": 1.9365916146449405e-05, "loss": 0.872, "step": 2700 }, { "epoch": 0.42, "grad_norm": 22.780151092780315, "learning_rate": 1.936532506059957e-05, "loss": 0.938, "step": 2701 }, { "epoch": 0.42, "grad_norm": 23.783845257351054, "learning_rate": 1.936473370840537e-05, "loss": 0.9177, "step": 2702 }, { "epoch": 0.42, "grad_norm": 20.529436400045263, "learning_rate": 1.936414208988362e-05, "loss": 0.8504, "step": 2703 }, { "epoch": 0.42, "grad_norm": 27.865737750064977, "learning_rate": 1.936355020505115e-05, "loss": 0.9391, "step": 2704 }, { "epoch": 0.42, "grad_norm": 19.11100780236159, "learning_rate": 1.9362958053924786e-05, "loss": 0.8241, "step": 2705 }, { "epoch": 0.42, "grad_norm": 16.930240394988935, "learning_rate": 1.9362365636521377e-05, "loss": 0.8171, "step": 2706 }, { "epoch": 0.42, "grad_norm": 18.082579719991905, "learning_rate": 1.9361772952857762e-05, "loss": 0.8348, "step": 2707 }, { "epoch": 0.42, "grad_norm": 15.997880650018738, "learning_rate": 1.93611800029508e-05, "loss": 0.7933, "step": 2708 }, { "epoch": 0.42, "grad_norm": 25.995473620715693, "learning_rate": 1.9360586786817355e-05, "loss": 0.8625, "step": 2709 }, { "epoch": 0.42, "grad_norm": 22.979720218080338, "learning_rate": 1.9359993304474302e-05, "loss": 0.8743, "step": 2710 }, { "epoch": 0.42, "grad_norm": 16.680965757860424, "learning_rate": 1.935939955593851e-05, "loss": 0.8663, "step": 2711 }, { "epoch": 0.42, "grad_norm": 30.115294095895603, "learning_rate": 1.9358805541226872e-05, "loss": 0.8895, "step": 2712 }, { "epoch": 0.42, "grad_norm": 18.827004143683194, "learning_rate": 1.9358211260356282e-05, "loss": 0.8517, "step": 2713 }, { "epoch": 0.42, "grad_norm": 19.58248676962657, "learning_rate": 1.9357616713343633e-05, "loss": 0.857, "step": 2714 }, { "epoch": 0.42, "grad_norm": 17.93136013431971, "learning_rate": 1.935702190020584e-05, "loss": 0.8615, "step": 2715 }, { "epoch": 0.42, "grad_norm": 15.677871383670924, "learning_rate": 1.9356426820959817e-05, "loss": 0.8615, "step": 2716 }, { "epoch": 0.42, "grad_norm": 18.538824027642022, "learning_rate": 1.935583147562249e-05, "loss": 0.9242, "step": 2717 }, { "epoch": 0.42, "grad_norm": 26.46955974493218, "learning_rate": 1.9355235864210792e-05, "loss": 1.0344, "step": 2718 }, { "epoch": 0.42, "grad_norm": 20.194503384287984, "learning_rate": 1.9354639986741653e-05, "loss": 0.9864, "step": 2719 }, { "epoch": 0.42, "grad_norm": 24.302758743922624, "learning_rate": 1.9354043843232028e-05, "loss": 0.9412, "step": 2720 }, { "epoch": 0.43, "grad_norm": 16.898592952387137, "learning_rate": 1.935344743369887e-05, "loss": 0.8174, "step": 2721 }, { "epoch": 0.43, "grad_norm": 19.005242735707103, "learning_rate": 1.9352850758159136e-05, "loss": 0.8357, "step": 2722 }, { "epoch": 0.43, "grad_norm": 28.29533521537663, "learning_rate": 1.9352253816629796e-05, "loss": 0.8839, "step": 2723 }, { "epoch": 0.43, "grad_norm": 13.290798532477393, "learning_rate": 1.9351656609127833e-05, "loss": 0.758, "step": 2724 }, { "epoch": 0.43, "grad_norm": 14.259434209592417, "learning_rate": 1.9351059135670222e-05, "loss": 0.827, "step": 2725 }, { "epoch": 0.43, "grad_norm": 27.97107893874645, "learning_rate": 1.9350461396273963e-05, "loss": 0.8841, "step": 2726 }, { "epoch": 0.43, "grad_norm": 23.08908614714832, "learning_rate": 1.934986339095605e-05, "loss": 0.8445, "step": 2727 }, { "epoch": 0.43, "grad_norm": 17.594918757657876, "learning_rate": 1.934926511973349e-05, "loss": 0.7765, "step": 2728 }, { "epoch": 0.43, "grad_norm": 20.439456423537294, "learning_rate": 1.9348666582623302e-05, "loss": 0.7948, "step": 2729 }, { "epoch": 0.43, "grad_norm": 19.495766741869282, "learning_rate": 1.9348067779642506e-05, "loss": 0.8457, "step": 2730 }, { "epoch": 0.43, "grad_norm": 23.42110042932668, "learning_rate": 1.9347468710808128e-05, "loss": 0.8191, "step": 2731 }, { "epoch": 0.43, "grad_norm": 20.988561329125673, "learning_rate": 1.9346869376137206e-05, "loss": 0.8307, "step": 2732 }, { "epoch": 0.43, "grad_norm": 22.340376920452577, "learning_rate": 1.9346269775646793e-05, "loss": 0.8083, "step": 2733 }, { "epoch": 0.43, "grad_norm": 22.13230733836942, "learning_rate": 1.9345669909353934e-05, "loss": 0.8327, "step": 2734 }, { "epoch": 0.43, "grad_norm": 14.948771506687264, "learning_rate": 1.9345069777275685e-05, "loss": 0.7678, "step": 2735 }, { "epoch": 0.43, "grad_norm": 17.722435713386524, "learning_rate": 1.934446937942912e-05, "loss": 0.7494, "step": 2736 }, { "epoch": 0.43, "grad_norm": 16.34788575734836, "learning_rate": 1.9343868715831313e-05, "loss": 0.7902, "step": 2737 }, { "epoch": 0.43, "grad_norm": 23.809935082445637, "learning_rate": 1.9343267786499346e-05, "loss": 0.7964, "step": 2738 }, { "epoch": 0.43, "grad_norm": 21.479036582074325, "learning_rate": 1.9342666591450307e-05, "loss": 0.8675, "step": 2739 }, { "epoch": 0.43, "grad_norm": 21.61778473520191, "learning_rate": 1.9342065130701297e-05, "loss": 0.8963, "step": 2740 }, { "epoch": 0.43, "grad_norm": 23.35420822044292, "learning_rate": 1.9341463404269422e-05, "loss": 0.8754, "step": 2741 }, { "epoch": 0.43, "grad_norm": 19.1031477576541, "learning_rate": 1.934086141217179e-05, "loss": 0.7699, "step": 2742 }, { "epoch": 0.43, "grad_norm": 14.799582350019515, "learning_rate": 1.9340259154425524e-05, "loss": 0.7482, "step": 2743 }, { "epoch": 0.43, "grad_norm": 26.268775707852193, "learning_rate": 1.933965663104775e-05, "loss": 0.878, "step": 2744 }, { "epoch": 0.43, "grad_norm": 18.606657994922337, "learning_rate": 1.9339053842055606e-05, "loss": 0.8324, "step": 2745 }, { "epoch": 0.43, "grad_norm": 23.30469492574248, "learning_rate": 1.9338450787466234e-05, "loss": 0.7594, "step": 2746 }, { "epoch": 0.43, "grad_norm": 18.01005121246038, "learning_rate": 1.9337847467296783e-05, "loss": 0.7594, "step": 2747 }, { "epoch": 0.43, "grad_norm": 20.100149317739525, "learning_rate": 1.9337243881564417e-05, "loss": 0.7973, "step": 2748 }, { "epoch": 0.43, "grad_norm": 22.149178918071595, "learning_rate": 1.9336640030286293e-05, "loss": 0.8795, "step": 2749 }, { "epoch": 0.43, "grad_norm": 23.784110354669203, "learning_rate": 1.933603591347959e-05, "loss": 0.9595, "step": 2750 }, { "epoch": 0.43, "grad_norm": 3.477578478672543, "learning_rate": 1.9335431531161486e-05, "loss": 0.7808, "step": 2751 }, { "epoch": 0.43, "grad_norm": 15.147766612888061, "learning_rate": 1.933482688334917e-05, "loss": 0.7993, "step": 2752 }, { "epoch": 0.43, "grad_norm": 15.406671432201627, "learning_rate": 1.9334221970059837e-05, "loss": 0.7119, "step": 2753 }, { "epoch": 0.43, "grad_norm": 15.062424950640239, "learning_rate": 1.9333616791310696e-05, "loss": 0.7349, "step": 2754 }, { "epoch": 0.43, "grad_norm": 17.54979717570123, "learning_rate": 1.9333011347118953e-05, "loss": 0.9561, "step": 2755 }, { "epoch": 0.43, "grad_norm": 22.768112161300735, "learning_rate": 1.9332405637501823e-05, "loss": 0.8222, "step": 2756 }, { "epoch": 0.43, "grad_norm": 20.354478960966798, "learning_rate": 1.9331799662476537e-05, "loss": 0.945, "step": 2757 }, { "epoch": 0.43, "grad_norm": 16.542335764767387, "learning_rate": 1.933119342206033e-05, "loss": 0.8322, "step": 2758 }, { "epoch": 0.43, "grad_norm": 16.29031479227513, "learning_rate": 1.933058691627044e-05, "loss": 0.8184, "step": 2759 }, { "epoch": 0.43, "grad_norm": 18.46496870572795, "learning_rate": 1.9329980145124115e-05, "loss": 0.8366, "step": 2760 }, { "epoch": 0.43, "grad_norm": 18.949423400994245, "learning_rate": 1.9329373108638614e-05, "loss": 0.7697, "step": 2761 }, { "epoch": 0.43, "grad_norm": 18.251941269839598, "learning_rate": 1.93287658068312e-05, "loss": 0.8186, "step": 2762 }, { "epoch": 0.43, "grad_norm": 19.620956735969866, "learning_rate": 1.932815823971914e-05, "loss": 0.7985, "step": 2763 }, { "epoch": 0.43, "grad_norm": 22.906531339026152, "learning_rate": 1.9327550407319717e-05, "loss": 0.9988, "step": 2764 }, { "epoch": 0.43, "grad_norm": 17.057307608259677, "learning_rate": 1.932694230965022e-05, "loss": 0.7492, "step": 2765 }, { "epoch": 0.43, "grad_norm": 22.8766000479919, "learning_rate": 1.9326333946727938e-05, "loss": 0.8733, "step": 2766 }, { "epoch": 0.43, "grad_norm": 22.366252670778806, "learning_rate": 1.932572531857017e-05, "loss": 0.8315, "step": 2767 }, { "epoch": 0.43, "grad_norm": 22.97704605480904, "learning_rate": 1.9325116425194235e-05, "loss": 0.873, "step": 2768 }, { "epoch": 0.43, "grad_norm": 17.93085267014662, "learning_rate": 1.9324507266617444e-05, "loss": 0.864, "step": 2769 }, { "epoch": 0.43, "grad_norm": 25.14933070516083, "learning_rate": 1.932389784285712e-05, "loss": 0.9263, "step": 2770 }, { "epoch": 0.43, "grad_norm": 19.647679178912497, "learning_rate": 1.9323288153930595e-05, "loss": 0.7297, "step": 2771 }, { "epoch": 0.43, "grad_norm": 14.797743218961239, "learning_rate": 1.932267819985521e-05, "loss": 0.8237, "step": 2772 }, { "epoch": 0.43, "grad_norm": 21.2180337131879, "learning_rate": 1.932206798064831e-05, "loss": 0.8558, "step": 2773 }, { "epoch": 0.43, "grad_norm": 19.404121098252574, "learning_rate": 1.932145749632725e-05, "loss": 0.8191, "step": 2774 }, { "epoch": 0.43, "grad_norm": 20.558964219507182, "learning_rate": 1.932084674690939e-05, "loss": 0.796, "step": 2775 }, { "epoch": 0.43, "grad_norm": 25.654297185751886, "learning_rate": 1.9320235732412104e-05, "loss": 0.9271, "step": 2776 }, { "epoch": 0.43, "grad_norm": 15.460692308822745, "learning_rate": 1.9319624452852765e-05, "loss": 0.7587, "step": 2777 }, { "epoch": 0.43, "grad_norm": 25.55434798502312, "learning_rate": 1.931901290824876e-05, "loss": 0.8581, "step": 2778 }, { "epoch": 0.43, "grad_norm": 17.45949459984029, "learning_rate": 1.9318401098617475e-05, "loss": 0.782, "step": 2779 }, { "epoch": 0.43, "grad_norm": 16.010583702854948, "learning_rate": 1.9317789023976314e-05, "loss": 0.7699, "step": 2780 }, { "epoch": 0.43, "grad_norm": 19.76329166300339, "learning_rate": 1.9317176684342685e-05, "loss": 0.8075, "step": 2781 }, { "epoch": 0.43, "grad_norm": 23.381196447646076, "learning_rate": 1.9316564079734005e-05, "loss": 0.813, "step": 2782 }, { "epoch": 0.43, "grad_norm": 14.620988882648282, "learning_rate": 1.931595121016769e-05, "loss": 0.8484, "step": 2783 }, { "epoch": 0.43, "grad_norm": 23.305390716618515, "learning_rate": 1.9315338075661172e-05, "loss": 0.8336, "step": 2784 }, { "epoch": 0.44, "grad_norm": 17.44467630605795, "learning_rate": 1.931472467623189e-05, "loss": 0.772, "step": 2785 }, { "epoch": 0.44, "grad_norm": 18.793004956683315, "learning_rate": 1.9314111011897285e-05, "loss": 0.9619, "step": 2786 }, { "epoch": 0.44, "grad_norm": 23.540675940737646, "learning_rate": 1.9313497082674813e-05, "loss": 0.9043, "step": 2787 }, { "epoch": 0.44, "grad_norm": 21.164699910261472, "learning_rate": 1.931288288858193e-05, "loss": 0.8337, "step": 2788 }, { "epoch": 0.44, "grad_norm": 13.454483928343135, "learning_rate": 1.9312268429636108e-05, "loss": 0.6702, "step": 2789 }, { "epoch": 0.44, "grad_norm": 20.001847195258666, "learning_rate": 1.9311653705854817e-05, "loss": 0.7726, "step": 2790 }, { "epoch": 0.44, "grad_norm": 29.434434865819178, "learning_rate": 1.9311038717255542e-05, "loss": 1.0484, "step": 2791 }, { "epoch": 0.44, "grad_norm": 19.90567106486018, "learning_rate": 1.9310423463855774e-05, "loss": 0.9265, "step": 2792 }, { "epoch": 0.44, "grad_norm": 26.145564394716416, "learning_rate": 1.930980794567301e-05, "loss": 0.8952, "step": 2793 }, { "epoch": 0.44, "grad_norm": 42.86813095108527, "learning_rate": 1.9309192162724756e-05, "loss": 0.9466, "step": 2794 }, { "epoch": 0.44, "grad_norm": 14.99347205769062, "learning_rate": 1.9308576115028515e-05, "loss": 0.7671, "step": 2795 }, { "epoch": 0.44, "grad_norm": 27.096805186308767, "learning_rate": 1.9307959802601824e-05, "loss": 0.856, "step": 2796 }, { "epoch": 0.44, "grad_norm": 17.20832649231591, "learning_rate": 1.9307343225462195e-05, "loss": 0.7815, "step": 2797 }, { "epoch": 0.44, "grad_norm": 20.836475105773456, "learning_rate": 1.930672638362717e-05, "loss": 0.8059, "step": 2798 }, { "epoch": 0.44, "grad_norm": 20.827430523571735, "learning_rate": 1.9306109277114292e-05, "loss": 0.8531, "step": 2799 }, { "epoch": 0.44, "grad_norm": 23.487774256797177, "learning_rate": 1.930549190594111e-05, "loss": 0.7303, "step": 2800 }, { "epoch": 0.44, "grad_norm": 23.2695687591317, "learning_rate": 1.930487427012518e-05, "loss": 0.9313, "step": 2801 }, { "epoch": 0.44, "grad_norm": 23.644877286452623, "learning_rate": 1.930425636968407e-05, "loss": 0.8546, "step": 2802 }, { "epoch": 0.44, "grad_norm": 29.70245533147644, "learning_rate": 1.9303638204635354e-05, "loss": 0.8352, "step": 2803 }, { "epoch": 0.44, "grad_norm": 18.042241174530986, "learning_rate": 1.930301977499661e-05, "loss": 0.8288, "step": 2804 }, { "epoch": 0.44, "grad_norm": 22.242139636016038, "learning_rate": 1.930240108078542e-05, "loss": 0.8485, "step": 2805 }, { "epoch": 0.44, "grad_norm": 14.388418784413442, "learning_rate": 1.9301782122019392e-05, "loss": 0.7612, "step": 2806 }, { "epoch": 0.44, "grad_norm": 13.451465471103637, "learning_rate": 1.9301162898716116e-05, "loss": 0.8578, "step": 2807 }, { "epoch": 0.44, "grad_norm": 19.405939431400558, "learning_rate": 1.9300543410893213e-05, "loss": 0.8356, "step": 2808 }, { "epoch": 0.44, "grad_norm": 27.165890540818857, "learning_rate": 1.9299923658568294e-05, "loss": 0.7887, "step": 2809 }, { "epoch": 0.44, "grad_norm": 18.407376016101875, "learning_rate": 1.9299303641758986e-05, "loss": 0.7502, "step": 2810 }, { "epoch": 0.44, "grad_norm": 16.465288114829953, "learning_rate": 1.9298683360482923e-05, "loss": 0.8396, "step": 2811 }, { "epoch": 0.44, "grad_norm": 34.92772760867393, "learning_rate": 1.9298062814757746e-05, "loss": 0.9241, "step": 2812 }, { "epoch": 0.44, "grad_norm": 18.657996869052813, "learning_rate": 1.92974420046011e-05, "loss": 0.7611, "step": 2813 }, { "epoch": 0.44, "grad_norm": 13.418309645230106, "learning_rate": 1.9296820930030642e-05, "loss": 0.7252, "step": 2814 }, { "epoch": 0.44, "grad_norm": 14.719098678739265, "learning_rate": 1.9296199591064036e-05, "loss": 0.7518, "step": 2815 }, { "epoch": 0.44, "grad_norm": 13.900043727099296, "learning_rate": 1.9295577987718953e-05, "loss": 0.8918, "step": 2816 }, { "epoch": 0.44, "grad_norm": 17.353443178660008, "learning_rate": 1.929495612001307e-05, "loss": 0.8403, "step": 2817 }, { "epoch": 0.44, "grad_norm": 29.80150160354507, "learning_rate": 1.9294333987964067e-05, "loss": 0.8056, "step": 2818 }, { "epoch": 0.44, "grad_norm": 15.495019609607217, "learning_rate": 1.9293711591589645e-05, "loss": 0.8668, "step": 2819 }, { "epoch": 0.44, "grad_norm": 20.93091527891047, "learning_rate": 1.9293088930907505e-05, "loss": 0.7623, "step": 2820 }, { "epoch": 0.44, "grad_norm": 15.685629673804485, "learning_rate": 1.9292466005935352e-05, "loss": 0.7573, "step": 2821 }, { "epoch": 0.44, "grad_norm": 17.229591706205877, "learning_rate": 1.9291842816690898e-05, "loss": 0.7855, "step": 2822 }, { "epoch": 0.44, "grad_norm": 14.20478784208699, "learning_rate": 1.9291219363191873e-05, "loss": 0.8322, "step": 2823 }, { "epoch": 0.44, "grad_norm": 14.574667659646149, "learning_rate": 1.9290595645456003e-05, "loss": 0.7588, "step": 2824 }, { "epoch": 0.44, "grad_norm": 13.720982019705515, "learning_rate": 1.9289971663501027e-05, "loss": 0.8363, "step": 2825 }, { "epoch": 0.44, "grad_norm": 20.736009845229447, "learning_rate": 1.9289347417344694e-05, "loss": 0.768, "step": 2826 }, { "epoch": 0.44, "grad_norm": 23.70867224969513, "learning_rate": 1.928872290700475e-05, "loss": 0.8619, "step": 2827 }, { "epoch": 0.44, "grad_norm": 35.94005813864722, "learning_rate": 1.9288098132498966e-05, "loss": 0.817, "step": 2828 }, { "epoch": 0.44, "grad_norm": 15.729340832316538, "learning_rate": 1.9287473093845102e-05, "loss": 0.7662, "step": 2829 }, { "epoch": 0.44, "grad_norm": 25.33096850540217, "learning_rate": 1.9286847791060937e-05, "loss": 0.8533, "step": 2830 }, { "epoch": 0.44, "grad_norm": 3.731406791945428, "learning_rate": 1.928622222416425e-05, "loss": 0.7649, "step": 2831 }, { "epoch": 0.44, "grad_norm": 25.22571892330583, "learning_rate": 1.928559639317284e-05, "loss": 0.8109, "step": 2832 }, { "epoch": 0.44, "grad_norm": 21.07722226960523, "learning_rate": 1.9284970298104495e-05, "loss": 0.7884, "step": 2833 }, { "epoch": 0.44, "grad_norm": 16.24392876091876, "learning_rate": 1.928434393897703e-05, "loss": 0.8451, "step": 2834 }, { "epoch": 0.44, "grad_norm": 17.055815044775574, "learning_rate": 1.9283717315808255e-05, "loss": 0.8385, "step": 2835 }, { "epoch": 0.44, "grad_norm": 23.294156363632272, "learning_rate": 1.928309042861599e-05, "loss": 0.7455, "step": 2836 }, { "epoch": 0.44, "grad_norm": 21.476296005177545, "learning_rate": 1.9282463277418062e-05, "loss": 0.911, "step": 2837 }, { "epoch": 0.44, "grad_norm": 16.64769102971277, "learning_rate": 1.928183586223231e-05, "loss": 0.8264, "step": 2838 }, { "epoch": 0.44, "grad_norm": 18.59383549390099, "learning_rate": 1.9281208183076576e-05, "loss": 0.8206, "step": 2839 }, { "epoch": 0.44, "grad_norm": 22.910436317962915, "learning_rate": 1.9280580239968708e-05, "loss": 0.8734, "step": 2840 }, { "epoch": 0.44, "grad_norm": 22.43049350806382, "learning_rate": 1.927995203292657e-05, "loss": 0.88, "step": 2841 }, { "epoch": 0.44, "grad_norm": 15.987091362221824, "learning_rate": 1.9279323561968024e-05, "loss": 0.7627, "step": 2842 }, { "epoch": 0.44, "grad_norm": 12.835625789244995, "learning_rate": 1.927869482711094e-05, "loss": 0.7323, "step": 2843 }, { "epoch": 0.44, "grad_norm": 21.991106596922414, "learning_rate": 1.927806582837321e-05, "loss": 0.8016, "step": 2844 }, { "epoch": 0.44, "grad_norm": 22.10333686844645, "learning_rate": 1.927743656577271e-05, "loss": 0.9001, "step": 2845 }, { "epoch": 0.44, "grad_norm": 28.852537134557487, "learning_rate": 1.9276807039327344e-05, "loss": 0.9389, "step": 2846 }, { "epoch": 0.44, "grad_norm": 13.031735257963781, "learning_rate": 1.9276177249055012e-05, "loss": 0.7254, "step": 2847 }, { "epoch": 0.44, "grad_norm": 11.636012338566934, "learning_rate": 1.9275547194973626e-05, "loss": 0.879, "step": 2848 }, { "epoch": 0.45, "grad_norm": 19.620915853524366, "learning_rate": 1.9274916877101104e-05, "loss": 0.843, "step": 2849 }, { "epoch": 0.45, "grad_norm": 15.696749001023305, "learning_rate": 1.9274286295455373e-05, "loss": 0.6976, "step": 2850 }, { "epoch": 0.45, "grad_norm": 27.36699915098846, "learning_rate": 1.927365545005436e-05, "loss": 0.8912, "step": 2851 }, { "epoch": 0.45, "grad_norm": 30.587851371664442, "learning_rate": 1.9273024340916015e-05, "loss": 0.9167, "step": 2852 }, { "epoch": 0.45, "grad_norm": 16.504500437316278, "learning_rate": 1.9272392968058282e-05, "loss": 0.7861, "step": 2853 }, { "epoch": 0.45, "grad_norm": 32.04225498861467, "learning_rate": 1.9271761331499114e-05, "loss": 0.9072, "step": 2854 }, { "epoch": 0.45, "grad_norm": 17.428991956182465, "learning_rate": 1.927112943125648e-05, "loss": 0.8342, "step": 2855 }, { "epoch": 0.45, "grad_norm": 30.74141216754505, "learning_rate": 1.9270497267348348e-05, "loss": 0.8617, "step": 2856 }, { "epoch": 0.45, "grad_norm": 19.899881867271972, "learning_rate": 1.9269864839792697e-05, "loss": 0.8082, "step": 2857 }, { "epoch": 0.45, "grad_norm": 23.883514953630044, "learning_rate": 1.9269232148607515e-05, "loss": 0.8591, "step": 2858 }, { "epoch": 0.45, "grad_norm": 13.003423591312604, "learning_rate": 1.926859919381079e-05, "loss": 0.7978, "step": 2859 }, { "epoch": 0.45, "grad_norm": 15.679820444582184, "learning_rate": 1.926796597542053e-05, "loss": 0.9388, "step": 2860 }, { "epoch": 0.45, "grad_norm": 22.481095958348472, "learning_rate": 1.9267332493454732e-05, "loss": 0.7668, "step": 2861 }, { "epoch": 0.45, "grad_norm": 18.287048395966394, "learning_rate": 1.9266698747931425e-05, "loss": 0.7206, "step": 2862 }, { "epoch": 0.45, "grad_norm": 19.923573275715178, "learning_rate": 1.9266064738868625e-05, "loss": 0.847, "step": 2863 }, { "epoch": 0.45, "grad_norm": 23.292903726689474, "learning_rate": 1.9265430466284362e-05, "loss": 0.871, "step": 2864 }, { "epoch": 0.45, "grad_norm": 32.18323858162131, "learning_rate": 1.9264795930196677e-05, "loss": 0.9068, "step": 2865 }, { "epoch": 0.45, "grad_norm": 17.809343413020226, "learning_rate": 1.9264161130623618e-05, "loss": 0.9016, "step": 2866 }, { "epoch": 0.45, "grad_norm": 26.926370773381045, "learning_rate": 1.9263526067583235e-05, "loss": 0.8996, "step": 2867 }, { "epoch": 0.45, "grad_norm": 15.92374671214326, "learning_rate": 1.926289074109359e-05, "loss": 0.7603, "step": 2868 }, { "epoch": 0.45, "grad_norm": 26.096887386183415, "learning_rate": 1.9262255151172752e-05, "loss": 0.8825, "step": 2869 }, { "epoch": 0.45, "grad_norm": 14.555201177905374, "learning_rate": 1.9261619297838794e-05, "loss": 0.8921, "step": 2870 }, { "epoch": 0.45, "grad_norm": 17.373893600938462, "learning_rate": 1.92609831811098e-05, "loss": 0.8113, "step": 2871 }, { "epoch": 0.45, "grad_norm": 18.639356349355918, "learning_rate": 1.9260346801003864e-05, "loss": 0.8037, "step": 2872 }, { "epoch": 0.45, "grad_norm": 16.585625110995945, "learning_rate": 1.925971015753908e-05, "loss": 0.8173, "step": 2873 }, { "epoch": 0.45, "grad_norm": 21.404036637200182, "learning_rate": 1.925907325073356e-05, "loss": 0.8761, "step": 2874 }, { "epoch": 0.45, "grad_norm": 26.06670253819081, "learning_rate": 1.925843608060541e-05, "loss": 0.938, "step": 2875 }, { "epoch": 0.45, "grad_norm": 15.330411233051962, "learning_rate": 1.925779864717275e-05, "loss": 0.7903, "step": 2876 }, { "epoch": 0.45, "grad_norm": 21.969001946964816, "learning_rate": 1.925716095045372e-05, "loss": 0.8028, "step": 2877 }, { "epoch": 0.45, "grad_norm": 19.030716748432223, "learning_rate": 1.9256522990466445e-05, "loss": 0.8031, "step": 2878 }, { "epoch": 0.45, "grad_norm": 20.625710126905126, "learning_rate": 1.925588476722907e-05, "loss": 0.8059, "step": 2879 }, { "epoch": 0.45, "grad_norm": 36.493411691443036, "learning_rate": 1.9255246280759747e-05, "loss": 0.802, "step": 2880 }, { "epoch": 0.45, "grad_norm": 12.679559905608317, "learning_rate": 1.9254607531076633e-05, "loss": 0.735, "step": 2881 }, { "epoch": 0.45, "grad_norm": 19.48163898319547, "learning_rate": 1.9253968518197896e-05, "loss": 0.8374, "step": 2882 }, { "epoch": 0.45, "grad_norm": 22.901193054065537, "learning_rate": 1.925332924214171e-05, "loss": 0.8883, "step": 2883 }, { "epoch": 0.45, "grad_norm": 19.677948101766702, "learning_rate": 1.925268970292625e-05, "loss": 0.8174, "step": 2884 }, { "epoch": 0.45, "grad_norm": 25.600996014098943, "learning_rate": 1.9252049900569707e-05, "loss": 0.8718, "step": 2885 }, { "epoch": 0.45, "grad_norm": 20.30245772040695, "learning_rate": 1.925140983509028e-05, "loss": 0.7605, "step": 2886 }, { "epoch": 0.45, "grad_norm": 19.85453456772631, "learning_rate": 1.9250769506506164e-05, "loss": 0.8221, "step": 2887 }, { "epoch": 0.45, "grad_norm": 21.577647359604168, "learning_rate": 1.925012891483558e-05, "loss": 0.7792, "step": 2888 }, { "epoch": 0.45, "grad_norm": 18.794540209930005, "learning_rate": 1.924948806009674e-05, "loss": 0.7336, "step": 2889 }, { "epoch": 0.45, "grad_norm": 16.356493493907163, "learning_rate": 1.9248846942307867e-05, "loss": 0.7764, "step": 2890 }, { "epoch": 0.45, "grad_norm": 15.396885541591361, "learning_rate": 1.92482055614872e-05, "loss": 0.8077, "step": 2891 }, { "epoch": 0.45, "grad_norm": 29.39138071912793, "learning_rate": 1.9247563917652978e-05, "loss": 0.74, "step": 2892 }, { "epoch": 0.45, "grad_norm": 23.058876910531712, "learning_rate": 1.9246922010823445e-05, "loss": 0.7889, "step": 2893 }, { "epoch": 0.45, "grad_norm": 20.35058201205367, "learning_rate": 1.924627984101686e-05, "loss": 0.8713, "step": 2894 }, { "epoch": 0.45, "grad_norm": 23.303365962775334, "learning_rate": 1.924563740825149e-05, "loss": 0.7908, "step": 2895 }, { "epoch": 0.45, "grad_norm": 19.218575881541433, "learning_rate": 1.9244994712545596e-05, "loss": 0.7346, "step": 2896 }, { "epoch": 0.45, "grad_norm": 20.45784247951282, "learning_rate": 1.924435175391746e-05, "loss": 0.8048, "step": 2897 }, { "epoch": 0.45, "grad_norm": 18.769205526806456, "learning_rate": 1.924370853238537e-05, "loss": 0.7424, "step": 2898 }, { "epoch": 0.45, "grad_norm": 16.916617140223753, "learning_rate": 1.9243065047967614e-05, "loss": 0.7788, "step": 2899 }, { "epoch": 0.45, "grad_norm": 20.888614085555783, "learning_rate": 1.92424213006825e-05, "loss": 0.8761, "step": 2900 }, { "epoch": 0.45, "grad_norm": 25.87728224239912, "learning_rate": 1.924177729054833e-05, "loss": 0.894, "step": 2901 }, { "epoch": 0.45, "grad_norm": 38.50258540399339, "learning_rate": 1.9241133017583416e-05, "loss": 1.0238, "step": 2902 }, { "epoch": 0.45, "grad_norm": 16.942786794876184, "learning_rate": 1.9240488481806086e-05, "loss": 0.772, "step": 2903 }, { "epoch": 0.45, "grad_norm": 16.247008721428116, "learning_rate": 1.923984368323467e-05, "loss": 0.7697, "step": 2904 }, { "epoch": 0.45, "grad_norm": 27.128511498190324, "learning_rate": 1.9239198621887505e-05, "loss": 0.8602, "step": 2905 }, { "epoch": 0.45, "grad_norm": 23.063971694778907, "learning_rate": 1.9238553297782937e-05, "loss": 0.7644, "step": 2906 }, { "epoch": 0.45, "grad_norm": 16.009838367951684, "learning_rate": 1.9237907710939317e-05, "loss": 0.7811, "step": 2907 }, { "epoch": 0.45, "grad_norm": 19.240008564197918, "learning_rate": 1.9237261861375004e-05, "loss": 0.9248, "step": 2908 }, { "epoch": 0.45, "grad_norm": 27.23992104588431, "learning_rate": 1.923661574910837e-05, "loss": 0.7512, "step": 2909 }, { "epoch": 0.45, "grad_norm": 21.939374063872858, "learning_rate": 1.9235969374157786e-05, "loss": 0.9421, "step": 2910 }, { "epoch": 0.45, "grad_norm": 17.181644877669257, "learning_rate": 1.9235322736541635e-05, "loss": 0.8945, "step": 2911 }, { "epoch": 0.45, "grad_norm": 22.533710042537482, "learning_rate": 1.9234675836278308e-05, "loss": 0.846, "step": 2912 }, { "epoch": 0.46, "grad_norm": 5.960547458816901, "learning_rate": 1.9234028673386205e-05, "loss": 0.7696, "step": 2913 }, { "epoch": 0.46, "grad_norm": 18.396502906181674, "learning_rate": 1.9233381247883724e-05, "loss": 0.8041, "step": 2914 }, { "epoch": 0.46, "grad_norm": 18.387321258067743, "learning_rate": 1.9232733559789286e-05, "loss": 0.796, "step": 2915 }, { "epoch": 0.46, "grad_norm": 16.071487851572833, "learning_rate": 1.9232085609121305e-05, "loss": 0.8013, "step": 2916 }, { "epoch": 0.46, "grad_norm": 14.002491058816016, "learning_rate": 1.9231437395898207e-05, "loss": 0.7197, "step": 2917 }, { "epoch": 0.46, "grad_norm": 17.68117358451436, "learning_rate": 1.9230788920138433e-05, "loss": 0.8017, "step": 2918 }, { "epoch": 0.46, "grad_norm": 21.04612653430889, "learning_rate": 1.923014018186042e-05, "loss": 0.7759, "step": 2919 }, { "epoch": 0.46, "grad_norm": 19.44459704258811, "learning_rate": 1.922949118108262e-05, "loss": 0.8843, "step": 2920 }, { "epoch": 0.46, "grad_norm": 22.341203785582618, "learning_rate": 1.9228841917823492e-05, "loss": 0.8406, "step": 2921 }, { "epoch": 0.46, "grad_norm": 20.879511315024136, "learning_rate": 1.92281923921015e-05, "loss": 0.8005, "step": 2922 }, { "epoch": 0.46, "grad_norm": 15.180477080078573, "learning_rate": 1.922754260393511e-05, "loss": 0.7988, "step": 2923 }, { "epoch": 0.46, "grad_norm": 15.55693439175142, "learning_rate": 1.9226892553342808e-05, "loss": 0.7678, "step": 2924 }, { "epoch": 0.46, "grad_norm": 23.63735886531153, "learning_rate": 1.9226242240343082e-05, "loss": 0.8642, "step": 2925 }, { "epoch": 0.46, "grad_norm": 17.689179542921874, "learning_rate": 1.9225591664954423e-05, "loss": 0.7621, "step": 2926 }, { "epoch": 0.46, "grad_norm": 23.62090388367294, "learning_rate": 1.922494082719533e-05, "loss": 0.8511, "step": 2927 }, { "epoch": 0.46, "grad_norm": 18.44901606835908, "learning_rate": 1.922428972708432e-05, "loss": 0.8152, "step": 2928 }, { "epoch": 0.46, "grad_norm": 273.49283357951964, "learning_rate": 1.9223638364639902e-05, "loss": 0.8089, "step": 2929 }, { "epoch": 0.46, "grad_norm": 26.336963973047034, "learning_rate": 1.9222986739880607e-05, "loss": 0.8745, "step": 2930 }, { "epoch": 0.46, "grad_norm": 15.416894450523952, "learning_rate": 1.9222334852824966e-05, "loss": 0.7592, "step": 2931 }, { "epoch": 0.46, "grad_norm": 30.967808432945866, "learning_rate": 1.922168270349152e-05, "loss": 0.8561, "step": 2932 }, { "epoch": 0.46, "grad_norm": 27.312385426811293, "learning_rate": 1.9221030291898802e-05, "loss": 0.8278, "step": 2933 }, { "epoch": 0.46, "grad_norm": 16.97397118798427, "learning_rate": 1.9220377618065383e-05, "loss": 0.7038, "step": 2934 }, { "epoch": 0.46, "grad_norm": 21.4901065990765, "learning_rate": 1.921972468200982e-05, "loss": 0.7889, "step": 2935 }, { "epoch": 0.46, "grad_norm": 22.625603199526246, "learning_rate": 1.9219071483750678e-05, "loss": 0.8768, "step": 2936 }, { "epoch": 0.46, "grad_norm": 16.20967454333816, "learning_rate": 1.9218418023306536e-05, "loss": 0.8274, "step": 2937 }, { "epoch": 0.46, "grad_norm": 24.141432391247864, "learning_rate": 1.9217764300695977e-05, "loss": 0.7748, "step": 2938 }, { "epoch": 0.46, "grad_norm": 21.316808523052693, "learning_rate": 1.9217110315937597e-05, "loss": 0.835, "step": 2939 }, { "epoch": 0.46, "grad_norm": 30.469695912863475, "learning_rate": 1.921645606904999e-05, "loss": 0.891, "step": 2940 }, { "epoch": 0.46, "grad_norm": 21.706821970594095, "learning_rate": 1.9215801560051764e-05, "loss": 0.8276, "step": 2941 }, { "epoch": 0.46, "grad_norm": 14.769118284834967, "learning_rate": 1.921514678896153e-05, "loss": 0.7584, "step": 2942 }, { "epoch": 0.46, "grad_norm": 11.253685453527048, "learning_rate": 1.9214491755797916e-05, "loss": 0.7512, "step": 2943 }, { "epoch": 0.46, "grad_norm": 23.465678317509454, "learning_rate": 1.9213836460579546e-05, "loss": 0.8682, "step": 2944 }, { "epoch": 0.46, "grad_norm": 15.11941303362085, "learning_rate": 1.9213180903325056e-05, "loss": 0.8247, "step": 2945 }, { "epoch": 0.46, "grad_norm": 24.692209704748254, "learning_rate": 1.921252508405309e-05, "loss": 0.8497, "step": 2946 }, { "epoch": 0.46, "grad_norm": 19.01600024794213, "learning_rate": 1.92118690027823e-05, "loss": 0.8728, "step": 2947 }, { "epoch": 0.46, "grad_norm": 21.891205517207375, "learning_rate": 1.9211212659531345e-05, "loss": 0.809, "step": 2948 }, { "epoch": 0.46, "grad_norm": 21.101412431874632, "learning_rate": 1.9210556054318886e-05, "loss": 0.7563, "step": 2949 }, { "epoch": 0.46, "grad_norm": 16.680172858223028, "learning_rate": 1.9209899187163606e-05, "loss": 0.8506, "step": 2950 }, { "epoch": 0.46, "grad_norm": 15.491085455744027, "learning_rate": 1.920924205808418e-05, "loss": 0.7501, "step": 2951 }, { "epoch": 0.46, "grad_norm": 17.866830131057615, "learning_rate": 1.92085846670993e-05, "loss": 0.7544, "step": 2952 }, { "epoch": 0.46, "grad_norm": 17.84464375166633, "learning_rate": 1.9207927014227653e-05, "loss": 0.8491, "step": 2953 }, { "epoch": 0.46, "grad_norm": 16.44809408698163, "learning_rate": 1.9207269099487953e-05, "loss": 0.8276, "step": 2954 }, { "epoch": 0.46, "grad_norm": 14.07766197027408, "learning_rate": 1.9206610922898904e-05, "loss": 0.7609, "step": 2955 }, { "epoch": 0.46, "grad_norm": 23.54914848905122, "learning_rate": 1.9205952484479225e-05, "loss": 0.8494, "step": 2956 }, { "epoch": 0.46, "grad_norm": 26.0218185979425, "learning_rate": 1.920529378424764e-05, "loss": 0.8645, "step": 2957 }, { "epoch": 0.46, "grad_norm": 21.94251281096728, "learning_rate": 1.920463482222289e-05, "loss": 0.8819, "step": 2958 }, { "epoch": 0.46, "grad_norm": 16.004192332047143, "learning_rate": 1.920397559842371e-05, "loss": 0.8254, "step": 2959 }, { "epoch": 0.46, "grad_norm": 17.4700718577476, "learning_rate": 1.9203316112868844e-05, "loss": 0.7932, "step": 2960 }, { "epoch": 0.46, "grad_norm": 23.28652544928423, "learning_rate": 1.9202656365577056e-05, "loss": 0.7781, "step": 2961 }, { "epoch": 0.46, "grad_norm": 40.843648733817595, "learning_rate": 1.9201996356567104e-05, "loss": 0.9355, "step": 2962 }, { "epoch": 0.46, "grad_norm": 15.72787100895518, "learning_rate": 1.9201336085857757e-05, "loss": 0.8533, "step": 2963 }, { "epoch": 0.46, "grad_norm": 17.437005000967584, "learning_rate": 1.9200675553467793e-05, "loss": 0.7738, "step": 2964 }, { "epoch": 0.46, "grad_norm": 17.514193885590164, "learning_rate": 1.9200014759416002e-05, "loss": 0.7954, "step": 2965 }, { "epoch": 0.46, "grad_norm": 14.800568750003228, "learning_rate": 1.919935370372117e-05, "loss": 0.8311, "step": 2966 }, { "epoch": 0.46, "grad_norm": 15.624515791542912, "learning_rate": 1.91986923864021e-05, "loss": 0.7354, "step": 2967 }, { "epoch": 0.46, "grad_norm": 18.092666715845724, "learning_rate": 1.91980308074776e-05, "loss": 0.909, "step": 2968 }, { "epoch": 0.46, "grad_norm": 20.732371659651548, "learning_rate": 1.9197368966966486e-05, "loss": 0.8271, "step": 2969 }, { "epoch": 0.46, "grad_norm": 24.69417819706931, "learning_rate": 1.919670686488758e-05, "loss": 0.7414, "step": 2970 }, { "epoch": 0.46, "grad_norm": 20.122450704110175, "learning_rate": 1.9196044501259706e-05, "loss": 0.8145, "step": 2971 }, { "epoch": 0.46, "grad_norm": 17.592437767195474, "learning_rate": 1.919538187610171e-05, "loss": 0.7751, "step": 2972 }, { "epoch": 0.46, "grad_norm": 15.147962863759687, "learning_rate": 1.9194718989432434e-05, "loss": 0.7556, "step": 2973 }, { "epoch": 0.46, "grad_norm": 13.081973025903448, "learning_rate": 1.9194055841270724e-05, "loss": 0.7244, "step": 2974 }, { "epoch": 0.46, "grad_norm": 22.151120993107924, "learning_rate": 1.9193392431635447e-05, "loss": 0.7764, "step": 2975 }, { "epoch": 0.46, "grad_norm": 18.013099177799887, "learning_rate": 1.9192728760545466e-05, "loss": 0.8906, "step": 2976 }, { "epoch": 0.47, "grad_norm": 23.82556209101294, "learning_rate": 1.9192064828019657e-05, "loss": 0.8022, "step": 2977 }, { "epoch": 0.47, "grad_norm": 29.16419849113684, "learning_rate": 1.91914006340769e-05, "loss": 0.8661, "step": 2978 }, { "epoch": 0.47, "grad_norm": 17.058282397280657, "learning_rate": 1.919073617873609e-05, "loss": 0.7733, "step": 2979 }, { "epoch": 0.47, "grad_norm": 20.393799984046787, "learning_rate": 1.9190071462016115e-05, "loss": 0.7316, "step": 2980 }, { "epoch": 0.47, "grad_norm": 25.690095128816516, "learning_rate": 1.918940648393589e-05, "loss": 0.8779, "step": 2981 }, { "epoch": 0.47, "grad_norm": 16.530499667083717, "learning_rate": 1.9188741244514312e-05, "loss": 0.8052, "step": 2982 }, { "epoch": 0.47, "grad_norm": 22.38641859368205, "learning_rate": 1.9188075743770312e-05, "loss": 0.8937, "step": 2983 }, { "epoch": 0.47, "grad_norm": 25.43112931390468, "learning_rate": 1.9187409981722814e-05, "loss": 0.801, "step": 2984 }, { "epoch": 0.47, "grad_norm": 16.866776106651184, "learning_rate": 1.918674395839075e-05, "loss": 0.7824, "step": 2985 }, { "epoch": 0.47, "grad_norm": 26.97578799817641, "learning_rate": 1.9186077673793062e-05, "loss": 0.8566, "step": 2986 }, { "epoch": 0.47, "grad_norm": 25.05341934632158, "learning_rate": 1.9185411127948695e-05, "loss": 0.9179, "step": 2987 }, { "epoch": 0.47, "grad_norm": 18.067672739124912, "learning_rate": 1.9184744320876612e-05, "loss": 0.8418, "step": 2988 }, { "epoch": 0.47, "grad_norm": 19.47557472848811, "learning_rate": 1.9184077252595777e-05, "loss": 0.8147, "step": 2989 }, { "epoch": 0.47, "grad_norm": 15.80401669538672, "learning_rate": 1.918340992312515e-05, "loss": 0.8732, "step": 2990 }, { "epoch": 0.47, "grad_norm": 13.328644617154742, "learning_rate": 1.9182742332483722e-05, "loss": 0.715, "step": 2991 }, { "epoch": 0.47, "grad_norm": 18.63828946756317, "learning_rate": 1.9182074480690472e-05, "loss": 0.7943, "step": 2992 }, { "epoch": 0.47, "grad_norm": 13.833982534164743, "learning_rate": 1.9181406367764398e-05, "loss": 0.7129, "step": 2993 }, { "epoch": 0.47, "grad_norm": 21.7322433730387, "learning_rate": 1.9180737993724494e-05, "loss": 0.8075, "step": 2994 }, { "epoch": 0.47, "grad_norm": 14.652519000273282, "learning_rate": 1.9180069358589773e-05, "loss": 0.7774, "step": 2995 }, { "epoch": 0.47, "grad_norm": 21.711949560619978, "learning_rate": 1.9179400462379248e-05, "loss": 0.8935, "step": 2996 }, { "epoch": 0.47, "grad_norm": 19.270920943284427, "learning_rate": 1.9178731305111952e-05, "loss": 0.7492, "step": 2997 }, { "epoch": 0.47, "grad_norm": 20.175697308021366, "learning_rate": 1.91780618868069e-05, "loss": 0.8474, "step": 2998 }, { "epoch": 0.47, "grad_norm": 24.968055430767055, "learning_rate": 1.9177392207483138e-05, "loss": 0.7235, "step": 2999 }, { "epoch": 0.47, "grad_norm": 15.887994508659474, "learning_rate": 1.9176722267159706e-05, "loss": 0.8417, "step": 3000 }, { "epoch": 0.47, "grad_norm": 19.83000504108188, "learning_rate": 1.9176052065855666e-05, "loss": 0.8184, "step": 3001 }, { "epoch": 0.47, "grad_norm": 26.695328402940113, "learning_rate": 1.917538160359007e-05, "loss": 0.8408, "step": 3002 }, { "epoch": 0.47, "grad_norm": 25.982270747971615, "learning_rate": 1.9174710880381995e-05, "loss": 0.8644, "step": 3003 }, { "epoch": 0.47, "grad_norm": 22.912944427613844, "learning_rate": 1.9174039896250505e-05, "loss": 0.7888, "step": 3004 }, { "epoch": 0.47, "grad_norm": 20.032150266921676, "learning_rate": 1.9173368651214686e-05, "loss": 0.7972, "step": 3005 }, { "epoch": 0.47, "grad_norm": 19.035552993951598, "learning_rate": 1.917269714529363e-05, "loss": 0.7865, "step": 3006 }, { "epoch": 0.47, "grad_norm": 22.93616797064195, "learning_rate": 1.9172025378506434e-05, "loss": 0.7995, "step": 3007 }, { "epoch": 0.47, "grad_norm": 21.41459353378958, "learning_rate": 1.9171353350872203e-05, "loss": 0.767, "step": 3008 }, { "epoch": 0.47, "grad_norm": 20.700018681023064, "learning_rate": 1.9170681062410042e-05, "loss": 0.7218, "step": 3009 }, { "epoch": 0.47, "grad_norm": 31.0410510740856, "learning_rate": 1.917000851313908e-05, "loss": 0.8386, "step": 3010 }, { "epoch": 0.47, "grad_norm": 23.574986941661635, "learning_rate": 1.9169335703078443e-05, "loss": 0.757, "step": 3011 }, { "epoch": 0.47, "grad_norm": 21.38841900934235, "learning_rate": 1.916866263224726e-05, "loss": 0.7719, "step": 3012 }, { "epoch": 0.47, "grad_norm": 22.931303081752528, "learning_rate": 1.9167989300664675e-05, "loss": 0.8328, "step": 3013 }, { "epoch": 0.47, "grad_norm": 20.136696520403664, "learning_rate": 1.9167315708349835e-05, "loss": 0.8111, "step": 3014 }, { "epoch": 0.47, "grad_norm": 28.187525535622676, "learning_rate": 1.9166641855321902e-05, "loss": 0.7314, "step": 3015 }, { "epoch": 0.47, "grad_norm": 20.280160694669522, "learning_rate": 1.9165967741600038e-05, "loss": 0.7884, "step": 3016 }, { "epoch": 0.47, "grad_norm": 14.690755728793388, "learning_rate": 1.916529336720341e-05, "loss": 0.74, "step": 3017 }, { "epoch": 0.47, "grad_norm": 17.09338758509844, "learning_rate": 1.9164618732151202e-05, "loss": 0.835, "step": 3018 }, { "epoch": 0.47, "grad_norm": 48.7374454853749, "learning_rate": 1.91639438364626e-05, "loss": 0.8324, "step": 3019 }, { "epoch": 0.47, "grad_norm": 27.207459095194455, "learning_rate": 1.9163268680156793e-05, "loss": 0.879, "step": 3020 }, { "epoch": 0.47, "grad_norm": 24.994638515091175, "learning_rate": 1.9162593263252988e-05, "loss": 0.7789, "step": 3021 }, { "epoch": 0.47, "grad_norm": 17.173834867637698, "learning_rate": 1.916191758577039e-05, "loss": 0.8177, "step": 3022 }, { "epoch": 0.47, "grad_norm": 78.70199739861276, "learning_rate": 1.9161241647728213e-05, "loss": 1.0178, "step": 3023 }, { "epoch": 0.47, "grad_norm": 23.1262693230769, "learning_rate": 1.9160565449145687e-05, "loss": 0.787, "step": 3024 }, { "epoch": 0.47, "grad_norm": 13.807728325083367, "learning_rate": 1.9159888990042033e-05, "loss": 0.708, "step": 3025 }, { "epoch": 0.47, "grad_norm": 10.274863449463217, "learning_rate": 1.91592122704365e-05, "loss": 0.6912, "step": 3026 }, { "epoch": 0.47, "grad_norm": 13.52421657493265, "learning_rate": 1.915853529034832e-05, "loss": 0.8269, "step": 3027 }, { "epoch": 0.47, "grad_norm": 19.22484300063159, "learning_rate": 1.915785804979676e-05, "loss": 0.7822, "step": 3028 }, { "epoch": 0.47, "grad_norm": 29.935498456404957, "learning_rate": 1.915718054880107e-05, "loss": 0.9521, "step": 3029 }, { "epoch": 0.47, "grad_norm": 28.4065199819226, "learning_rate": 1.9156502787380527e-05, "loss": 0.7704, "step": 3030 }, { "epoch": 0.47, "grad_norm": 21.14438303372769, "learning_rate": 1.91558247655544e-05, "loss": 0.8861, "step": 3031 }, { "epoch": 0.47, "grad_norm": 188.43611629193143, "learning_rate": 1.9155146483341972e-05, "loss": 0.9285, "step": 3032 }, { "epoch": 0.47, "grad_norm": 23.070695464394742, "learning_rate": 1.9154467940762534e-05, "loss": 0.7855, "step": 3033 }, { "epoch": 0.47, "grad_norm": 19.187993590781332, "learning_rate": 1.9153789137835384e-05, "loss": 0.7894, "step": 3034 }, { "epoch": 0.47, "grad_norm": 68.40038764914176, "learning_rate": 1.9153110074579823e-05, "loss": 0.866, "step": 3035 }, { "epoch": 0.47, "grad_norm": 14.83281251750027, "learning_rate": 1.915243075101517e-05, "loss": 0.8398, "step": 3036 }, { "epoch": 0.47, "grad_norm": 15.808878613981356, "learning_rate": 1.9151751167160734e-05, "loss": 0.7947, "step": 3037 }, { "epoch": 0.47, "grad_norm": 19.707452924922112, "learning_rate": 1.9151071323035856e-05, "loss": 0.7885, "step": 3038 }, { "epoch": 0.47, "grad_norm": 22.65563795249573, "learning_rate": 1.9150391218659855e-05, "loss": 0.8874, "step": 3039 }, { "epoch": 0.47, "grad_norm": 32.72912747263712, "learning_rate": 1.9149710854052087e-05, "loss": 0.7335, "step": 3040 }, { "epoch": 0.48, "grad_norm": 21.214797116661902, "learning_rate": 1.914903022923189e-05, "loss": 0.8206, "step": 3041 }, { "epoch": 0.48, "grad_norm": 28.518177596767742, "learning_rate": 1.914834934421863e-05, "loss": 0.8378, "step": 3042 }, { "epoch": 0.48, "grad_norm": 23.781896890045363, "learning_rate": 1.9147668199031664e-05, "loss": 0.8055, "step": 3043 }, { "epoch": 0.48, "grad_norm": 27.430209166027087, "learning_rate": 1.9146986793690363e-05, "loss": 0.9628, "step": 3044 }, { "epoch": 0.48, "grad_norm": 18.027555733919122, "learning_rate": 1.914630512821411e-05, "loss": 0.8761, "step": 3045 }, { "epoch": 0.48, "grad_norm": 22.445006547376877, "learning_rate": 1.9145623202622293e-05, "loss": 0.8882, "step": 3046 }, { "epoch": 0.48, "grad_norm": 14.351071810860862, "learning_rate": 1.91449410169343e-05, "loss": 0.7529, "step": 3047 }, { "epoch": 0.48, "grad_norm": 16.19633968461515, "learning_rate": 1.9144258571169533e-05, "loss": 0.8132, "step": 3048 }, { "epoch": 0.48, "grad_norm": 22.80901606358736, "learning_rate": 1.9143575865347405e-05, "loss": 0.9359, "step": 3049 }, { "epoch": 0.48, "grad_norm": 15.474570051261475, "learning_rate": 1.9142892899487323e-05, "loss": 0.8576, "step": 3050 }, { "epoch": 0.48, "grad_norm": 16.161613733349938, "learning_rate": 1.9142209673608717e-05, "loss": 0.879, "step": 3051 }, { "epoch": 0.48, "grad_norm": 25.155900803505816, "learning_rate": 1.914152618773102e-05, "loss": 0.8732, "step": 3052 }, { "epoch": 0.48, "grad_norm": 24.691004289524503, "learning_rate": 1.914084244187366e-05, "loss": 0.8542, "step": 3053 }, { "epoch": 0.48, "grad_norm": 23.733439208562597, "learning_rate": 1.914015843605609e-05, "loss": 0.7865, "step": 3054 }, { "epoch": 0.48, "grad_norm": 30.483586455117734, "learning_rate": 1.9139474170297764e-05, "loss": 0.8413, "step": 3055 }, { "epoch": 0.48, "grad_norm": 21.977388508824916, "learning_rate": 1.9138789644618138e-05, "loss": 0.8357, "step": 3056 }, { "epoch": 0.48, "grad_norm": 21.508870975541917, "learning_rate": 1.9138104859036678e-05, "loss": 0.8364, "step": 3057 }, { "epoch": 0.48, "grad_norm": 21.79289521148358, "learning_rate": 1.913741981357286e-05, "loss": 0.9342, "step": 3058 }, { "epoch": 0.48, "grad_norm": 22.60082752919683, "learning_rate": 1.913673450824617e-05, "loss": 0.7614, "step": 3059 }, { "epoch": 0.48, "grad_norm": 14.314914554878236, "learning_rate": 1.9136048943076098e-05, "loss": 0.7982, "step": 3060 }, { "epoch": 0.48, "grad_norm": 16.21496649621919, "learning_rate": 1.9135363118082138e-05, "loss": 0.7698, "step": 3061 }, { "epoch": 0.48, "grad_norm": 14.68564685432869, "learning_rate": 1.913467703328379e-05, "loss": 0.813, "step": 3062 }, { "epoch": 0.48, "grad_norm": 14.238265147154276, "learning_rate": 1.9133990688700578e-05, "loss": 0.7063, "step": 3063 }, { "epoch": 0.48, "grad_norm": 20.684032162134578, "learning_rate": 1.9133304084352006e-05, "loss": 0.794, "step": 3064 }, { "epoch": 0.48, "grad_norm": 22.306271986791536, "learning_rate": 1.9132617220257613e-05, "loss": 0.8492, "step": 3065 }, { "epoch": 0.48, "grad_norm": 17.816758462426602, "learning_rate": 1.913193009643693e-05, "loss": 0.8228, "step": 3066 }, { "epoch": 0.48, "grad_norm": 20.44329454946185, "learning_rate": 1.9131242712909497e-05, "loss": 0.913, "step": 3067 }, { "epoch": 0.48, "grad_norm": 14.711856183312651, "learning_rate": 1.913055506969486e-05, "loss": 0.8675, "step": 3068 }, { "epoch": 0.48, "grad_norm": 16.10911783078787, "learning_rate": 1.9129867166812583e-05, "loss": 0.7506, "step": 3069 }, { "epoch": 0.48, "grad_norm": 22.39252835287759, "learning_rate": 1.9129179004282218e-05, "loss": 0.7542, "step": 3070 }, { "epoch": 0.48, "grad_norm": 19.396309603383, "learning_rate": 1.9128490582123346e-05, "loss": 0.8345, "step": 3071 }, { "epoch": 0.48, "grad_norm": 16.322095694118957, "learning_rate": 1.912780190035554e-05, "loss": 0.7505, "step": 3072 }, { "epoch": 0.48, "grad_norm": 19.454810197446502, "learning_rate": 1.9127112958998392e-05, "loss": 0.8345, "step": 3073 }, { "epoch": 0.48, "grad_norm": 105.19515645780156, "learning_rate": 1.9126423758071486e-05, "loss": 0.9433, "step": 3074 }, { "epoch": 0.48, "grad_norm": 20.251784971534057, "learning_rate": 1.912573429759443e-05, "loss": 0.863, "step": 3075 }, { "epoch": 0.48, "grad_norm": 18.192685568633546, "learning_rate": 1.9125044577586828e-05, "loss": 0.784, "step": 3076 }, { "epoch": 0.48, "grad_norm": 15.968912127543344, "learning_rate": 1.9124354598068294e-05, "loss": 0.8748, "step": 3077 }, { "epoch": 0.48, "grad_norm": 26.428298764865872, "learning_rate": 1.912366435905846e-05, "loss": 0.9027, "step": 3078 }, { "epoch": 0.48, "grad_norm": 24.129201116824394, "learning_rate": 1.9122973860576942e-05, "loss": 0.9924, "step": 3079 }, { "epoch": 0.48, "grad_norm": 18.870318116325166, "learning_rate": 1.9122283102643383e-05, "loss": 0.7567, "step": 3080 }, { "epoch": 0.48, "grad_norm": 20.323786018100822, "learning_rate": 1.9121592085277432e-05, "loss": 0.7717, "step": 3081 }, { "epoch": 0.48, "grad_norm": 25.112334402216227, "learning_rate": 1.912090080849874e-05, "loss": 0.8616, "step": 3082 }, { "epoch": 0.48, "grad_norm": 25.934243678587745, "learning_rate": 1.9120209272326962e-05, "loss": 0.7796, "step": 3083 }, { "epoch": 0.48, "grad_norm": 24.82672576550508, "learning_rate": 1.9119517476781766e-05, "loss": 0.8885, "step": 3084 }, { "epoch": 0.48, "grad_norm": 28.10484160412152, "learning_rate": 1.911882542188283e-05, "loss": 0.8709, "step": 3085 }, { "epoch": 0.48, "grad_norm": 16.074270100007478, "learning_rate": 1.9118133107649833e-05, "loss": 0.756, "step": 3086 }, { "epoch": 0.48, "grad_norm": 31.769694056903443, "learning_rate": 1.9117440534102466e-05, "loss": 0.9146, "step": 3087 }, { "epoch": 0.48, "grad_norm": 31.294511954370197, "learning_rate": 1.911674770126042e-05, "loss": 0.8921, "step": 3088 }, { "epoch": 0.48, "grad_norm": 19.86619813370443, "learning_rate": 1.9116054609143408e-05, "loss": 0.8159, "step": 3089 }, { "epoch": 0.48, "grad_norm": 19.140514074510172, "learning_rate": 1.9115361257771133e-05, "loss": 0.7716, "step": 3090 }, { "epoch": 0.48, "grad_norm": 15.32158532854001, "learning_rate": 1.911466764716331e-05, "loss": 0.9614, "step": 3091 }, { "epoch": 0.48, "grad_norm": 16.20811950273685, "learning_rate": 1.911397377733968e-05, "loss": 0.7836, "step": 3092 }, { "epoch": 0.48, "grad_norm": 13.772607778353215, "learning_rate": 1.9113279648319964e-05, "loss": 0.7588, "step": 3093 }, { "epoch": 0.48, "grad_norm": 34.73969914570977, "learning_rate": 1.9112585260123906e-05, "loss": 0.9183, "step": 3094 }, { "epoch": 0.48, "grad_norm": 19.482908049519, "learning_rate": 1.9111890612771252e-05, "loss": 0.886, "step": 3095 }, { "epoch": 0.48, "grad_norm": 17.53036064214237, "learning_rate": 1.911119570628176e-05, "loss": 0.7775, "step": 3096 }, { "epoch": 0.48, "grad_norm": 21.3446514356776, "learning_rate": 1.9110500540675194e-05, "loss": 0.9258, "step": 3097 }, { "epoch": 0.48, "grad_norm": 26.49974695458246, "learning_rate": 1.910980511597132e-05, "loss": 0.8614, "step": 3098 }, { "epoch": 0.48, "grad_norm": 27.22658414251488, "learning_rate": 1.910910943218992e-05, "loss": 0.8753, "step": 3099 }, { "epoch": 0.48, "grad_norm": 20.928771941239514, "learning_rate": 1.9108413489350775e-05, "loss": 0.8703, "step": 3100 }, { "epoch": 0.48, "grad_norm": 22.631568721769806, "learning_rate": 1.910771728747368e-05, "loss": 0.9159, "step": 3101 }, { "epoch": 0.48, "grad_norm": 19.876356522433195, "learning_rate": 1.910702082657843e-05, "loss": 0.7536, "step": 3102 }, { "epoch": 0.48, "grad_norm": 17.274547291080765, "learning_rate": 1.910632410668484e-05, "loss": 0.6968, "step": 3103 }, { "epoch": 0.48, "grad_norm": 14.671944836418826, "learning_rate": 1.9105627127812717e-05, "loss": 0.8034, "step": 3104 }, { "epoch": 0.49, "grad_norm": 22.764355759594338, "learning_rate": 1.910492988998188e-05, "loss": 0.8499, "step": 3105 }, { "epoch": 0.49, "grad_norm": 29.226975306267484, "learning_rate": 1.910423239321217e-05, "loss": 0.8224, "step": 3106 }, { "epoch": 0.49, "grad_norm": 12.579159736720895, "learning_rate": 1.9103534637523414e-05, "loss": 0.7808, "step": 3107 }, { "epoch": 0.49, "grad_norm": 21.72727288419119, "learning_rate": 1.910283662293546e-05, "loss": 0.8869, "step": 3108 }, { "epoch": 0.49, "grad_norm": 18.856564437993306, "learning_rate": 1.9102138349468154e-05, "loss": 0.8029, "step": 3109 }, { "epoch": 0.49, "grad_norm": 18.953049028249062, "learning_rate": 1.910143981714136e-05, "loss": 0.8347, "step": 3110 }, { "epoch": 0.49, "grad_norm": 16.9489918808129, "learning_rate": 1.9100741025974943e-05, "loss": 0.7513, "step": 3111 }, { "epoch": 0.49, "grad_norm": 27.66708381117652, "learning_rate": 1.9100041975988776e-05, "loss": 0.8121, "step": 3112 }, { "epoch": 0.49, "grad_norm": 21.065126588513714, "learning_rate": 1.9099342667202733e-05, "loss": 0.754, "step": 3113 }, { "epoch": 0.49, "grad_norm": 15.67327893729973, "learning_rate": 1.9098643099636714e-05, "loss": 0.7429, "step": 3114 }, { "epoch": 0.49, "grad_norm": 19.743642416760856, "learning_rate": 1.9097943273310603e-05, "loss": 0.8474, "step": 3115 }, { "epoch": 0.49, "grad_norm": 16.69908849475767, "learning_rate": 1.909724318824431e-05, "loss": 0.8815, "step": 3116 }, { "epoch": 0.49, "grad_norm": 25.1918363766617, "learning_rate": 1.909654284445774e-05, "loss": 0.8509, "step": 3117 }, { "epoch": 0.49, "grad_norm": 106.24448763193921, "learning_rate": 1.9095842241970817e-05, "loss": 0.8482, "step": 3118 }, { "epoch": 0.49, "grad_norm": 16.94313310015852, "learning_rate": 1.9095141380803457e-05, "loss": 0.7856, "step": 3119 }, { "epoch": 0.49, "grad_norm": 16.2955491385221, "learning_rate": 1.90944402609756e-05, "loss": 0.7584, "step": 3120 }, { "epoch": 0.49, "grad_norm": 27.399234865636398, "learning_rate": 1.909373888250718e-05, "loss": 0.8746, "step": 3121 }, { "epoch": 0.49, "grad_norm": 19.065127674726245, "learning_rate": 1.9093037245418147e-05, "loss": 0.8033, "step": 3122 }, { "epoch": 0.49, "grad_norm": 17.415516942768967, "learning_rate": 1.909233534972845e-05, "loss": 0.8502, "step": 3123 }, { "epoch": 0.49, "grad_norm": 19.13339042803202, "learning_rate": 1.9091633195458062e-05, "loss": 0.82, "step": 3124 }, { "epoch": 0.49, "grad_norm": 14.54999807055041, "learning_rate": 1.9090930782626943e-05, "loss": 0.7479, "step": 3125 }, { "epoch": 0.49, "grad_norm": 21.961036975028698, "learning_rate": 1.9090228111255066e-05, "loss": 0.8619, "step": 3126 }, { "epoch": 0.49, "grad_norm": 21.472341947734442, "learning_rate": 1.908952518136242e-05, "loss": 0.8502, "step": 3127 }, { "epoch": 0.49, "grad_norm": 16.592292588739333, "learning_rate": 1.9088821992969e-05, "loss": 0.7517, "step": 3128 }, { "epoch": 0.49, "grad_norm": 14.21307371670834, "learning_rate": 1.9088118546094793e-05, "loss": 0.8045, "step": 3129 }, { "epoch": 0.49, "grad_norm": 19.981963037825658, "learning_rate": 1.9087414840759812e-05, "loss": 0.8468, "step": 3130 }, { "epoch": 0.49, "grad_norm": 22.076005426268907, "learning_rate": 1.9086710876984075e-05, "loss": 0.7823, "step": 3131 }, { "epoch": 0.49, "grad_norm": 19.660020628441846, "learning_rate": 1.9086006654787588e-05, "loss": 0.7982, "step": 3132 }, { "epoch": 0.49, "grad_norm": 16.416622407082215, "learning_rate": 1.908530217419039e-05, "loss": 0.8989, "step": 3133 }, { "epoch": 0.49, "grad_norm": 13.823122805150962, "learning_rate": 1.9084597435212517e-05, "loss": 0.8361, "step": 3134 }, { "epoch": 0.49, "grad_norm": 17.627198711174408, "learning_rate": 1.9083892437874004e-05, "loss": 0.8226, "step": 3135 }, { "epoch": 0.49, "grad_norm": 20.515210547695123, "learning_rate": 1.90831871821949e-05, "loss": 0.7261, "step": 3136 }, { "epoch": 0.49, "grad_norm": 27.63891188206161, "learning_rate": 1.9082481668195273e-05, "loss": 0.8063, "step": 3137 }, { "epoch": 0.49, "grad_norm": 21.223533788113098, "learning_rate": 1.9081775895895177e-05, "loss": 0.7253, "step": 3138 }, { "epoch": 0.49, "grad_norm": 21.209520861701765, "learning_rate": 1.9081069865314688e-05, "loss": 0.7902, "step": 3139 }, { "epoch": 0.49, "grad_norm": 25.65344581020031, "learning_rate": 1.9080363576473883e-05, "loss": 0.7909, "step": 3140 }, { "epoch": 0.49, "grad_norm": 16.616322906054116, "learning_rate": 1.907965702939285e-05, "loss": 0.7744, "step": 3141 }, { "epoch": 0.49, "grad_norm": 15.884314015271608, "learning_rate": 1.9078950224091685e-05, "loss": 0.8179, "step": 3142 }, { "epoch": 0.49, "grad_norm": 18.443519655416107, "learning_rate": 1.907824316059048e-05, "loss": 0.8592, "step": 3143 }, { "epoch": 0.49, "grad_norm": 18.7428535376474, "learning_rate": 1.9077535838909356e-05, "loss": 0.7896, "step": 3144 }, { "epoch": 0.49, "grad_norm": 16.010459291984223, "learning_rate": 1.9076828259068422e-05, "loss": 0.7939, "step": 3145 }, { "epoch": 0.49, "grad_norm": 18.17897469664532, "learning_rate": 1.9076120421087806e-05, "loss": 0.7815, "step": 3146 }, { "epoch": 0.49, "grad_norm": 20.02230568097062, "learning_rate": 1.907541232498763e-05, "loss": 0.9493, "step": 3147 }, { "epoch": 0.49, "grad_norm": 23.369242738149158, "learning_rate": 1.9074703970788032e-05, "loss": 0.9246, "step": 3148 }, { "epoch": 0.49, "grad_norm": 20.075400649309046, "learning_rate": 1.907399535850917e-05, "loss": 0.8102, "step": 3149 }, { "epoch": 0.49, "grad_norm": 21.134030874954295, "learning_rate": 1.9073286488171184e-05, "loss": 0.7831, "step": 3150 }, { "epoch": 0.49, "grad_norm": 17.20100989319069, "learning_rate": 1.9072577359794237e-05, "loss": 0.797, "step": 3151 }, { "epoch": 0.49, "grad_norm": 16.31817177208678, "learning_rate": 1.9071867973398502e-05, "loss": 0.8151, "step": 3152 }, { "epoch": 0.49, "grad_norm": 4.3416210736141725, "learning_rate": 1.9071158329004144e-05, "loss": 0.8087, "step": 3153 }, { "epoch": 0.49, "grad_norm": 18.61621300103439, "learning_rate": 1.907044842663135e-05, "loss": 0.7402, "step": 3154 }, { "epoch": 0.49, "grad_norm": 16.41108041951863, "learning_rate": 1.906973826630031e-05, "loss": 0.865, "step": 3155 }, { "epoch": 0.49, "grad_norm": 20.024465899899187, "learning_rate": 1.906902784803122e-05, "loss": 0.8306, "step": 3156 }, { "epoch": 0.49, "grad_norm": 26.99403006507065, "learning_rate": 1.9068317171844284e-05, "loss": 0.9317, "step": 3157 }, { "epoch": 0.49, "grad_norm": 9.96428759991952, "learning_rate": 1.9067606237759712e-05, "loss": 0.7189, "step": 3158 }, { "epoch": 0.49, "grad_norm": 17.963645590690497, "learning_rate": 1.906689504579772e-05, "loss": 0.8662, "step": 3159 }, { "epoch": 0.49, "grad_norm": 17.218922496925103, "learning_rate": 1.906618359597854e-05, "loss": 0.9281, "step": 3160 }, { "epoch": 0.49, "grad_norm": 17.79946724003632, "learning_rate": 1.90654718883224e-05, "loss": 0.6847, "step": 3161 }, { "epoch": 0.49, "grad_norm": 18.17157078272671, "learning_rate": 1.9064759922849544e-05, "loss": 0.8478, "step": 3162 }, { "epoch": 0.49, "grad_norm": 16.78562059796633, "learning_rate": 1.906404769958022e-05, "loss": 0.6997, "step": 3163 }, { "epoch": 0.49, "grad_norm": 17.497370269004637, "learning_rate": 1.9063335218534677e-05, "loss": 0.7554, "step": 3164 }, { "epoch": 0.49, "grad_norm": 16.649693669473898, "learning_rate": 1.906262247973319e-05, "loss": 0.7674, "step": 3165 }, { "epoch": 0.49, "grad_norm": 15.00603353579781, "learning_rate": 1.9061909483196012e-05, "loss": 0.8077, "step": 3166 }, { "epoch": 0.49, "grad_norm": 21.96696829319825, "learning_rate": 1.9061196228943436e-05, "loss": 0.9424, "step": 3167 }, { "epoch": 0.49, "grad_norm": 14.666303651373918, "learning_rate": 1.906048271699574e-05, "loss": 0.7462, "step": 3168 }, { "epoch": 0.5, "grad_norm": 21.56346788169085, "learning_rate": 1.905976894737321e-05, "loss": 0.8058, "step": 3169 }, { "epoch": 0.5, "grad_norm": 14.983457982262538, "learning_rate": 1.9059054920096155e-05, "loss": 0.8306, "step": 3170 }, { "epoch": 0.5, "grad_norm": 22.155661529398557, "learning_rate": 1.9058340635184877e-05, "loss": 0.8405, "step": 3171 }, { "epoch": 0.5, "grad_norm": 14.57549470322433, "learning_rate": 1.9057626092659692e-05, "loss": 0.7334, "step": 3172 }, { "epoch": 0.5, "grad_norm": 15.827567746896081, "learning_rate": 1.9056911292540915e-05, "loss": 0.7195, "step": 3173 }, { "epoch": 0.5, "grad_norm": 21.972028228600916, "learning_rate": 1.905619623484888e-05, "loss": 0.7214, "step": 3174 }, { "epoch": 0.5, "grad_norm": 13.866056987188042, "learning_rate": 1.9055480919603924e-05, "loss": 0.7584, "step": 3175 }, { "epoch": 0.5, "grad_norm": 26.330229548889065, "learning_rate": 1.905476534682639e-05, "loss": 0.8473, "step": 3176 }, { "epoch": 0.5, "grad_norm": 31.325558457872724, "learning_rate": 1.905404951653662e-05, "loss": 0.8722, "step": 3177 }, { "epoch": 0.5, "grad_norm": 17.211834858164256, "learning_rate": 1.9053333428754986e-05, "loss": 0.7664, "step": 3178 }, { "epoch": 0.5, "grad_norm": 15.083965391892713, "learning_rate": 1.9052617083501837e-05, "loss": 0.85, "step": 3179 }, { "epoch": 0.5, "grad_norm": 17.158817002806252, "learning_rate": 1.905190048079756e-05, "loss": 0.8052, "step": 3180 }, { "epoch": 0.5, "grad_norm": 12.372182607235231, "learning_rate": 1.9051183620662526e-05, "loss": 0.7049, "step": 3181 }, { "epoch": 0.5, "grad_norm": 22.96415398086274, "learning_rate": 1.9050466503117127e-05, "loss": 0.8397, "step": 3182 }, { "epoch": 0.5, "grad_norm": 25.250823988311357, "learning_rate": 1.9049749128181752e-05, "loss": 0.9214, "step": 3183 }, { "epoch": 0.5, "grad_norm": 20.628738158483294, "learning_rate": 1.904903149587681e-05, "loss": 0.8246, "step": 3184 }, { "epoch": 0.5, "grad_norm": 13.580946174331254, "learning_rate": 1.90483136062227e-05, "loss": 0.7727, "step": 3185 }, { "epoch": 0.5, "grad_norm": 17.16508975837801, "learning_rate": 1.9047595459239848e-05, "loss": 0.7698, "step": 3186 }, { "epoch": 0.5, "grad_norm": 17.30438426938258, "learning_rate": 1.904687705494867e-05, "loss": 0.7347, "step": 3187 }, { "epoch": 0.5, "grad_norm": 21.562595660232425, "learning_rate": 1.9046158393369608e-05, "loss": 0.9778, "step": 3188 }, { "epoch": 0.5, "grad_norm": 13.90905310314675, "learning_rate": 1.9045439474523086e-05, "loss": 0.8563, "step": 3189 }, { "epoch": 0.5, "grad_norm": 15.126788763602624, "learning_rate": 1.904472029842956e-05, "loss": 0.7797, "step": 3190 }, { "epoch": 0.5, "grad_norm": 26.624222759758187, "learning_rate": 1.904400086510948e-05, "loss": 0.8578, "step": 3191 }, { "epoch": 0.5, "grad_norm": 24.69015382433746, "learning_rate": 1.9043281174583305e-05, "loss": 0.7512, "step": 3192 }, { "epoch": 0.5, "grad_norm": 28.265565583233567, "learning_rate": 1.9042561226871506e-05, "loss": 0.8479, "step": 3193 }, { "epoch": 0.5, "grad_norm": 18.815323731396376, "learning_rate": 1.9041841021994552e-05, "loss": 0.7862, "step": 3194 }, { "epoch": 0.5, "grad_norm": 17.924109704560887, "learning_rate": 1.9041120559972927e-05, "loss": 0.7745, "step": 3195 }, { "epoch": 0.5, "grad_norm": 3.877079630596655, "learning_rate": 1.904039984082713e-05, "loss": 0.8712, "step": 3196 }, { "epoch": 0.5, "grad_norm": 18.77250016992269, "learning_rate": 1.9039678864577642e-05, "loss": 0.8153, "step": 3197 }, { "epoch": 0.5, "grad_norm": 27.76538514853176, "learning_rate": 1.903895763124498e-05, "loss": 0.8957, "step": 3198 }, { "epoch": 0.5, "grad_norm": 23.117283871505574, "learning_rate": 1.9038236140849648e-05, "loss": 0.8031, "step": 3199 }, { "epoch": 0.5, "grad_norm": 18.047662480005094, "learning_rate": 1.9037514393412167e-05, "loss": 0.8145, "step": 3200 }, { "epoch": 0.5, "grad_norm": 14.283046792237744, "learning_rate": 1.9036792388953066e-05, "loss": 0.7254, "step": 3201 }, { "epoch": 0.5, "grad_norm": 23.19781966004275, "learning_rate": 1.9036070127492875e-05, "loss": 0.76, "step": 3202 }, { "epoch": 0.5, "grad_norm": 20.17193392088283, "learning_rate": 1.903534760905213e-05, "loss": 0.8017, "step": 3203 }, { "epoch": 0.5, "grad_norm": 16.10453323656041, "learning_rate": 1.9034624833651393e-05, "loss": 0.7997, "step": 3204 }, { "epoch": 0.5, "grad_norm": 18.030486963000833, "learning_rate": 1.9033901801311207e-05, "loss": 0.8689, "step": 3205 }, { "epoch": 0.5, "grad_norm": 23.46829393479594, "learning_rate": 1.903317851205214e-05, "loss": 0.839, "step": 3206 }, { "epoch": 0.5, "grad_norm": 20.5378164201416, "learning_rate": 1.903245496589476e-05, "loss": 0.7806, "step": 3207 }, { "epoch": 0.5, "grad_norm": 22.730279853850114, "learning_rate": 1.9031731162859642e-05, "loss": 0.8162, "step": 3208 }, { "epoch": 0.5, "grad_norm": 16.84158372661201, "learning_rate": 1.9031007102967375e-05, "loss": 0.7447, "step": 3209 }, { "epoch": 0.5, "grad_norm": 26.09868887228873, "learning_rate": 1.903028278623855e-05, "loss": 0.888, "step": 3210 }, { "epoch": 0.5, "grad_norm": 14.788062582253868, "learning_rate": 1.9029558212693764e-05, "loss": 0.7907, "step": 3211 }, { "epoch": 0.5, "grad_norm": 23.156128065789513, "learning_rate": 1.9028833382353624e-05, "loss": 0.8118, "step": 3212 }, { "epoch": 0.5, "grad_norm": 13.067312722915801, "learning_rate": 1.9028108295238745e-05, "loss": 0.7429, "step": 3213 }, { "epoch": 0.5, "grad_norm": 15.730143143699177, "learning_rate": 1.902738295136975e-05, "loss": 0.7797, "step": 3214 }, { "epoch": 0.5, "grad_norm": 20.737672631233288, "learning_rate": 1.902665735076726e-05, "loss": 0.7292, "step": 3215 }, { "epoch": 0.5, "grad_norm": 25.47696348750332, "learning_rate": 1.9025931493451917e-05, "loss": 0.8176, "step": 3216 }, { "epoch": 0.5, "grad_norm": 18.45868765326416, "learning_rate": 1.9025205379444362e-05, "loss": 0.7827, "step": 3217 }, { "epoch": 0.5, "grad_norm": 25.305542756219992, "learning_rate": 1.9024479008765246e-05, "loss": 0.8556, "step": 3218 }, { "epoch": 0.5, "grad_norm": 22.36812396126054, "learning_rate": 1.9023752381435227e-05, "loss": 0.7753, "step": 3219 }, { "epoch": 0.5, "grad_norm": 15.08174138604477, "learning_rate": 1.9023025497474968e-05, "loss": 0.7337, "step": 3220 }, { "epoch": 0.5, "grad_norm": 42.66182033372968, "learning_rate": 1.9022298356905146e-05, "loss": 0.9066, "step": 3221 }, { "epoch": 0.5, "grad_norm": 17.446188973439455, "learning_rate": 1.902157095974643e-05, "loss": 0.8542, "step": 3222 }, { "epoch": 0.5, "grad_norm": 45.87349920130624, "learning_rate": 1.902084330601952e-05, "loss": 0.854, "step": 3223 }, { "epoch": 0.5, "grad_norm": 19.09499898530948, "learning_rate": 1.9020115395745098e-05, "loss": 0.7287, "step": 3224 }, { "epoch": 0.5, "grad_norm": 18.85777698342505, "learning_rate": 1.9019387228943872e-05, "loss": 0.8088, "step": 3225 }, { "epoch": 0.5, "grad_norm": 17.911027894602718, "learning_rate": 1.9018658805636553e-05, "loss": 0.8078, "step": 3226 }, { "epoch": 0.5, "grad_norm": 25.314574331483886, "learning_rate": 1.901793012584385e-05, "loss": 0.7852, "step": 3227 }, { "epoch": 0.5, "grad_norm": 13.346102203471863, "learning_rate": 1.901720118958649e-05, "loss": 0.8695, "step": 3228 }, { "epoch": 0.5, "grad_norm": 27.236645349492782, "learning_rate": 1.9016471996885202e-05, "loss": 0.839, "step": 3229 }, { "epoch": 0.5, "grad_norm": 16.432241548763045, "learning_rate": 1.9015742547760726e-05, "loss": 0.9157, "step": 3230 }, { "epoch": 0.5, "grad_norm": 24.47201569275253, "learning_rate": 1.9015012842233807e-05, "loss": 0.8392, "step": 3231 }, { "epoch": 0.5, "grad_norm": 23.995745037065667, "learning_rate": 1.9014282880325194e-05, "loss": 0.7897, "step": 3232 }, { "epoch": 0.5, "grad_norm": 23.31086148008573, "learning_rate": 1.9013552662055652e-05, "loss": 0.859, "step": 3233 }, { "epoch": 0.51, "grad_norm": 16.7200266199151, "learning_rate": 1.9012822187445944e-05, "loss": 0.7593, "step": 3234 }, { "epoch": 0.51, "grad_norm": 19.441460770125662, "learning_rate": 1.9012091456516843e-05, "loss": 0.8184, "step": 3235 }, { "epoch": 0.51, "grad_norm": 23.760640379030786, "learning_rate": 1.9011360469289138e-05, "loss": 0.9675, "step": 3236 }, { "epoch": 0.51, "grad_norm": 16.91864224073, "learning_rate": 1.901062922578361e-05, "loss": 0.7626, "step": 3237 }, { "epoch": 0.51, "grad_norm": 16.97133714710073, "learning_rate": 1.9009897726021058e-05, "loss": 0.8719, "step": 3238 }, { "epoch": 0.51, "grad_norm": 21.490950684369828, "learning_rate": 1.9009165970022282e-05, "loss": 0.7341, "step": 3239 }, { "epoch": 0.51, "grad_norm": 19.52146375768583, "learning_rate": 1.9008433957808102e-05, "loss": 0.7873, "step": 3240 }, { "epoch": 0.51, "grad_norm": 11.896604934104056, "learning_rate": 1.9007701689399323e-05, "loss": 0.7226, "step": 3241 }, { "epoch": 0.51, "grad_norm": 20.72922314743194, "learning_rate": 1.9006969164816784e-05, "loss": 0.7948, "step": 3242 }, { "epoch": 0.51, "grad_norm": 19.929676465637428, "learning_rate": 1.9006236384081306e-05, "loss": 0.7807, "step": 3243 }, { "epoch": 0.51, "grad_norm": 12.201355750938077, "learning_rate": 1.9005503347213738e-05, "loss": 0.8165, "step": 3244 }, { "epoch": 0.51, "grad_norm": 29.4712549006632, "learning_rate": 1.900477005423492e-05, "loss": 0.852, "step": 3245 }, { "epoch": 0.51, "grad_norm": 19.618645060894476, "learning_rate": 1.9004036505165708e-05, "loss": 1.0108, "step": 3246 }, { "epoch": 0.51, "grad_norm": 18.086511241966033, "learning_rate": 1.9003302700026968e-05, "loss": 0.6943, "step": 3247 }, { "epoch": 0.51, "grad_norm": 19.071567673811153, "learning_rate": 1.9002568638839566e-05, "loss": 0.8185, "step": 3248 }, { "epoch": 0.51, "grad_norm": 33.88515531423689, "learning_rate": 1.9001834321624378e-05, "loss": 0.9174, "step": 3249 }, { "epoch": 0.51, "grad_norm": 16.64538851824194, "learning_rate": 1.9001099748402287e-05, "loss": 0.7773, "step": 3250 }, { "epoch": 0.51, "grad_norm": 16.382574772326507, "learning_rate": 1.9000364919194185e-05, "loss": 0.7932, "step": 3251 }, { "epoch": 0.51, "grad_norm": 36.10558050811683, "learning_rate": 1.899962983402097e-05, "loss": 0.9106, "step": 3252 }, { "epoch": 0.51, "grad_norm": 19.984378653824976, "learning_rate": 1.8998894492903545e-05, "loss": 0.8735, "step": 3253 }, { "epoch": 0.51, "grad_norm": 24.14826342527058, "learning_rate": 1.8998158895862827e-05, "loss": 0.7821, "step": 3254 }, { "epoch": 0.51, "grad_norm": 15.4967278527968, "learning_rate": 1.8997423042919734e-05, "loss": 0.7565, "step": 3255 }, { "epoch": 0.51, "grad_norm": 24.466214145418185, "learning_rate": 1.899668693409519e-05, "loss": 0.7963, "step": 3256 }, { "epoch": 0.51, "grad_norm": 26.11605946292533, "learning_rate": 1.8995950569410136e-05, "loss": 0.9111, "step": 3257 }, { "epoch": 0.51, "grad_norm": 15.69279396543573, "learning_rate": 1.8995213948885508e-05, "loss": 0.7317, "step": 3258 }, { "epoch": 0.51, "grad_norm": 15.875436113216862, "learning_rate": 1.899447707254226e-05, "loss": 0.7685, "step": 3259 }, { "epoch": 0.51, "grad_norm": 17.43809216700467, "learning_rate": 1.8993739940401342e-05, "loss": 0.7964, "step": 3260 }, { "epoch": 0.51, "grad_norm": 47.01185875219237, "learning_rate": 1.8993002552483726e-05, "loss": 0.8537, "step": 3261 }, { "epoch": 0.51, "grad_norm": 17.80303413004833, "learning_rate": 1.8992264908810373e-05, "loss": 0.7608, "step": 3262 }, { "epoch": 0.51, "grad_norm": 12.31721858616252, "learning_rate": 1.899152700940227e-05, "loss": 0.7321, "step": 3263 }, { "epoch": 0.51, "grad_norm": 17.495838070581346, "learning_rate": 1.8990788854280397e-05, "loss": 0.7526, "step": 3264 }, { "epoch": 0.51, "grad_norm": 16.489194545562345, "learning_rate": 1.899005044346575e-05, "loss": 0.7448, "step": 3265 }, { "epoch": 0.51, "grad_norm": 12.4744470208579, "learning_rate": 1.8989311776979326e-05, "loss": 0.8306, "step": 3266 }, { "epoch": 0.51, "grad_norm": 14.526798336522805, "learning_rate": 1.8988572854842133e-05, "loss": 0.7805, "step": 3267 }, { "epoch": 0.51, "grad_norm": 18.05069596271059, "learning_rate": 1.898783367707519e-05, "loss": 0.8492, "step": 3268 }, { "epoch": 0.51, "grad_norm": 16.82385227201902, "learning_rate": 1.898709424369951e-05, "loss": 0.7442, "step": 3269 }, { "epoch": 0.51, "grad_norm": 19.30131590781488, "learning_rate": 1.898635455473613e-05, "loss": 0.8237, "step": 3270 }, { "epoch": 0.51, "grad_norm": 20.42282232759359, "learning_rate": 1.898561461020608e-05, "loss": 0.8917, "step": 3271 }, { "epoch": 0.51, "grad_norm": 14.499431660115283, "learning_rate": 1.8984874410130413e-05, "loss": 0.7379, "step": 3272 }, { "epoch": 0.51, "grad_norm": 15.93915737921503, "learning_rate": 1.898413395453017e-05, "loss": 0.8256, "step": 3273 }, { "epoch": 0.51, "grad_norm": 15.358397824315356, "learning_rate": 1.8983393243426413e-05, "loss": 0.8956, "step": 3274 }, { "epoch": 0.51, "grad_norm": 17.31304296526907, "learning_rate": 1.8982652276840205e-05, "loss": 0.726, "step": 3275 }, { "epoch": 0.51, "grad_norm": 19.093361744468556, "learning_rate": 1.8981911054792625e-05, "loss": 0.8032, "step": 3276 }, { "epoch": 0.51, "grad_norm": 36.918961722001136, "learning_rate": 1.8981169577304747e-05, "loss": 0.7639, "step": 3277 }, { "epoch": 0.51, "grad_norm": 16.241479004072122, "learning_rate": 1.898042784439766e-05, "loss": 0.8295, "step": 3278 }, { "epoch": 0.51, "grad_norm": 17.829385569119307, "learning_rate": 1.897968585609246e-05, "loss": 0.8646, "step": 3279 }, { "epoch": 0.51, "grad_norm": 25.805204674449193, "learning_rate": 1.8978943612410245e-05, "loss": 0.8207, "step": 3280 }, { "epoch": 0.51, "grad_norm": 18.210242630062517, "learning_rate": 1.8978201113372128e-05, "loss": 0.7854, "step": 3281 }, { "epoch": 0.51, "grad_norm": 16.48862852748904, "learning_rate": 1.8977458358999222e-05, "loss": 0.7114, "step": 3282 }, { "epoch": 0.51, "grad_norm": 22.770164608552005, "learning_rate": 1.8976715349312652e-05, "loss": 0.8092, "step": 3283 }, { "epoch": 0.51, "grad_norm": 19.67252407677198, "learning_rate": 1.897597208433355e-05, "loss": 0.7627, "step": 3284 }, { "epoch": 0.51, "grad_norm": 20.654441019912475, "learning_rate": 1.8975228564083052e-05, "loss": 0.8817, "step": 3285 }, { "epoch": 0.51, "grad_norm": 24.432568941648473, "learning_rate": 1.8974484788582303e-05, "loss": 0.8926, "step": 3286 }, { "epoch": 0.51, "grad_norm": 17.773174942927838, "learning_rate": 1.897374075785246e-05, "loss": 0.7816, "step": 3287 }, { "epoch": 0.51, "grad_norm": 13.417549328863489, "learning_rate": 1.8972996471914674e-05, "loss": 0.8842, "step": 3288 }, { "epoch": 0.51, "grad_norm": 15.497277064876052, "learning_rate": 1.8972251930790124e-05, "loss": 0.7477, "step": 3289 }, { "epoch": 0.51, "grad_norm": 22.686390993504194, "learning_rate": 1.8971507134499974e-05, "loss": 0.7553, "step": 3290 }, { "epoch": 0.51, "grad_norm": 33.51949868035587, "learning_rate": 1.8970762083065408e-05, "loss": 0.981, "step": 3291 }, { "epoch": 0.51, "grad_norm": 17.377263269082434, "learning_rate": 1.897001677650762e-05, "loss": 0.8271, "step": 3292 }, { "epoch": 0.51, "grad_norm": 15.536138019612448, "learning_rate": 1.8969271214847795e-05, "loss": 0.7148, "step": 3293 }, { "epoch": 0.51, "grad_norm": 18.923094759355415, "learning_rate": 1.896852539810715e-05, "loss": 0.8304, "step": 3294 }, { "epoch": 0.51, "grad_norm": 21.894773826650535, "learning_rate": 1.8967779326306884e-05, "loss": 0.9322, "step": 3295 }, { "epoch": 0.51, "grad_norm": 19.304494596696205, "learning_rate": 1.8967032999468226e-05, "loss": 0.7831, "step": 3296 }, { "epoch": 0.51, "grad_norm": 14.043791583152023, "learning_rate": 1.8966286417612393e-05, "loss": 0.8117, "step": 3297 }, { "epoch": 0.52, "grad_norm": 16.07606661292513, "learning_rate": 1.8965539580760615e-05, "loss": 0.8522, "step": 3298 }, { "epoch": 0.52, "grad_norm": 19.131179547515075, "learning_rate": 1.8964792488934143e-05, "loss": 0.8286, "step": 3299 }, { "epoch": 0.52, "grad_norm": 15.86586056611989, "learning_rate": 1.8964045142154212e-05, "loss": 0.8386, "step": 3300 }, { "epoch": 0.52, "grad_norm": 25.713210022562002, "learning_rate": 1.896329754044208e-05, "loss": 0.896, "step": 3301 }, { "epoch": 0.52, "grad_norm": 24.974547955524066, "learning_rate": 1.8962549683819013e-05, "loss": 0.8399, "step": 3302 }, { "epoch": 0.52, "grad_norm": 21.41597401849312, "learning_rate": 1.8961801572306276e-05, "loss": 0.8076, "step": 3303 }, { "epoch": 0.52, "grad_norm": 21.818217938092527, "learning_rate": 1.8961053205925143e-05, "loss": 0.7319, "step": 3304 }, { "epoch": 0.52, "grad_norm": 20.541694103242634, "learning_rate": 1.89603045846969e-05, "loss": 0.7611, "step": 3305 }, { "epoch": 0.52, "grad_norm": 19.173817859470883, "learning_rate": 1.8959555708642835e-05, "loss": 0.8252, "step": 3306 }, { "epoch": 0.52, "grad_norm": 20.150288053493515, "learning_rate": 1.895880657778425e-05, "loss": 0.7268, "step": 3307 }, { "epoch": 0.52, "grad_norm": 19.08003356690973, "learning_rate": 1.895805719214244e-05, "loss": 0.7844, "step": 3308 }, { "epoch": 0.52, "grad_norm": 19.322439721785628, "learning_rate": 1.8957307551738727e-05, "loss": 0.926, "step": 3309 }, { "epoch": 0.52, "grad_norm": 19.02751792682329, "learning_rate": 1.895655765659443e-05, "loss": 0.7907, "step": 3310 }, { "epoch": 0.52, "grad_norm": 33.30861495607743, "learning_rate": 1.8955807506730872e-05, "loss": 0.804, "step": 3311 }, { "epoch": 0.52, "grad_norm": 21.57224672039511, "learning_rate": 1.8955057102169385e-05, "loss": 0.7755, "step": 3312 }, { "epoch": 0.52, "grad_norm": 21.243166927249245, "learning_rate": 1.8954306442931315e-05, "loss": 0.8541, "step": 3313 }, { "epoch": 0.52, "grad_norm": 25.078610801744933, "learning_rate": 1.8953555529038006e-05, "loss": 0.8728, "step": 3314 }, { "epoch": 0.52, "grad_norm": 14.274344722894911, "learning_rate": 1.8952804360510816e-05, "loss": 0.8128, "step": 3315 }, { "epoch": 0.52, "grad_norm": 18.425515175977296, "learning_rate": 1.8952052937371105e-05, "loss": 0.8284, "step": 3316 }, { "epoch": 0.52, "grad_norm": 32.74106691373, "learning_rate": 1.8951301259640252e-05, "loss": 0.8772, "step": 3317 }, { "epoch": 0.52, "grad_norm": 25.996978524597257, "learning_rate": 1.895054932733962e-05, "loss": 0.8976, "step": 3318 }, { "epoch": 0.52, "grad_norm": 30.3681125249725, "learning_rate": 1.8949797140490607e-05, "loss": 0.9461, "step": 3319 }, { "epoch": 0.52, "grad_norm": 15.100203911570139, "learning_rate": 1.8949044699114592e-05, "loss": 0.7677, "step": 3320 }, { "epoch": 0.52, "grad_norm": 20.378046988326368, "learning_rate": 1.8948292003232987e-05, "loss": 0.9179, "step": 3321 }, { "epoch": 0.52, "grad_norm": 22.65943203810927, "learning_rate": 1.894753905286719e-05, "loss": 0.8169, "step": 3322 }, { "epoch": 0.52, "grad_norm": 24.6908656602609, "learning_rate": 1.8946785848038614e-05, "loss": 0.9243, "step": 3323 }, { "epoch": 0.52, "grad_norm": 19.642141462670683, "learning_rate": 1.8946032388768687e-05, "loss": 0.8762, "step": 3324 }, { "epoch": 0.52, "grad_norm": 17.503034261356934, "learning_rate": 1.8945278675078828e-05, "loss": 0.8297, "step": 3325 }, { "epoch": 0.52, "grad_norm": 18.890176946029367, "learning_rate": 1.894452470699048e-05, "loss": 0.817, "step": 3326 }, { "epoch": 0.52, "grad_norm": 17.339149383852504, "learning_rate": 1.894377048452508e-05, "loss": 0.9539, "step": 3327 }, { "epoch": 0.52, "grad_norm": 25.530667503206878, "learning_rate": 1.8943016007704078e-05, "loss": 0.8668, "step": 3328 }, { "epoch": 0.52, "grad_norm": 18.460202050300555, "learning_rate": 1.8942261276548932e-05, "loss": 0.8424, "step": 3329 }, { "epoch": 0.52, "grad_norm": 19.204140872949036, "learning_rate": 1.894150629108111e-05, "loss": 0.8521, "step": 3330 }, { "epoch": 0.52, "grad_norm": 24.107550123649787, "learning_rate": 1.8940751051322075e-05, "loss": 0.7128, "step": 3331 }, { "epoch": 0.52, "grad_norm": 18.54912264364053, "learning_rate": 1.8939995557293315e-05, "loss": 0.7438, "step": 3332 }, { "epoch": 0.52, "grad_norm": 13.289986236868678, "learning_rate": 1.8939239809016306e-05, "loss": 0.7825, "step": 3333 }, { "epoch": 0.52, "grad_norm": 15.173740732385943, "learning_rate": 1.893848380651255e-05, "loss": 0.8543, "step": 3334 }, { "epoch": 0.52, "grad_norm": 16.182607373474102, "learning_rate": 1.893772754980354e-05, "loss": 0.7827, "step": 3335 }, { "epoch": 0.52, "grad_norm": 18.840961295812292, "learning_rate": 1.8936971038910792e-05, "loss": 0.8808, "step": 3336 }, { "epoch": 0.52, "grad_norm": 18.335262489328013, "learning_rate": 1.8936214273855813e-05, "loss": 0.8171, "step": 3337 }, { "epoch": 0.52, "grad_norm": 20.641496398805156, "learning_rate": 1.8935457254660128e-05, "loss": 0.8496, "step": 3338 }, { "epoch": 0.52, "grad_norm": 24.289102779439688, "learning_rate": 1.8934699981345265e-05, "loss": 0.806, "step": 3339 }, { "epoch": 0.52, "grad_norm": 16.84246254797321, "learning_rate": 1.893394245393276e-05, "loss": 0.8426, "step": 3340 }, { "epoch": 0.52, "grad_norm": 16.598108331705184, "learning_rate": 1.893318467244416e-05, "loss": 0.8253, "step": 3341 }, { "epoch": 0.52, "grad_norm": 25.54094295632851, "learning_rate": 1.893242663690101e-05, "loss": 0.8221, "step": 3342 }, { "epoch": 0.52, "grad_norm": 19.41202666202375, "learning_rate": 1.8931668347324877e-05, "loss": 0.887, "step": 3343 }, { "epoch": 0.52, "grad_norm": 15.909508859381846, "learning_rate": 1.893090980373732e-05, "loss": 0.7678, "step": 3344 }, { "epoch": 0.52, "grad_norm": 21.05664375927195, "learning_rate": 1.8930151006159908e-05, "loss": 0.7798, "step": 3345 }, { "epoch": 0.52, "grad_norm": 17.522019220476068, "learning_rate": 1.892939195461423e-05, "loss": 0.7704, "step": 3346 }, { "epoch": 0.52, "grad_norm": 37.00418295928479, "learning_rate": 1.8928632649121867e-05, "loss": 0.8342, "step": 3347 }, { "epoch": 0.52, "grad_norm": 30.24032633658489, "learning_rate": 1.8927873089704416e-05, "loss": 0.8104, "step": 3348 }, { "epoch": 0.52, "grad_norm": 26.553372156383084, "learning_rate": 1.8927113276383475e-05, "loss": 0.8248, "step": 3349 }, { "epoch": 0.52, "grad_norm": 25.53241335052198, "learning_rate": 1.8926353209180655e-05, "loss": 0.826, "step": 3350 }, { "epoch": 0.52, "grad_norm": 13.55322187838542, "learning_rate": 1.892559288811757e-05, "loss": 0.8026, "step": 3351 }, { "epoch": 0.52, "grad_norm": 26.519803613153854, "learning_rate": 1.892483231321585e-05, "loss": 0.9084, "step": 3352 }, { "epoch": 0.52, "grad_norm": 21.11096696335605, "learning_rate": 1.8924071484497114e-05, "loss": 0.8003, "step": 3353 }, { "epoch": 0.52, "grad_norm": 24.335539141926084, "learning_rate": 1.892331040198301e-05, "loss": 0.8166, "step": 3354 }, { "epoch": 0.52, "grad_norm": 14.240287257457208, "learning_rate": 1.8922549065695172e-05, "loss": 0.7634, "step": 3355 }, { "epoch": 0.52, "grad_norm": 24.75685767574243, "learning_rate": 1.8921787475655265e-05, "loss": 0.9598, "step": 3356 }, { "epoch": 0.52, "grad_norm": 17.757156519389437, "learning_rate": 1.8921025631884938e-05, "loss": 0.8711, "step": 3357 }, { "epoch": 0.52, "grad_norm": 17.124454988209056, "learning_rate": 1.8920263534405858e-05, "loss": 0.7593, "step": 3358 }, { "epoch": 0.52, "grad_norm": 21.969145722440434, "learning_rate": 1.8919501183239705e-05, "loss": 0.8531, "step": 3359 }, { "epoch": 0.52, "grad_norm": 21.07152944592918, "learning_rate": 1.8918738578408157e-05, "loss": 0.7959, "step": 3360 }, { "epoch": 0.52, "grad_norm": 15.900326921630372, "learning_rate": 1.8917975719932898e-05, "loss": 0.7134, "step": 3361 }, { "epoch": 0.53, "grad_norm": 17.994220511945322, "learning_rate": 1.891721260783563e-05, "loss": 0.7959, "step": 3362 }, { "epoch": 0.53, "grad_norm": 15.25241374432694, "learning_rate": 1.8916449242138047e-05, "loss": 0.8172, "step": 3363 }, { "epoch": 0.53, "grad_norm": 26.73083024684305, "learning_rate": 1.8915685622861864e-05, "loss": 0.8052, "step": 3364 }, { "epoch": 0.53, "grad_norm": 18.64199132380459, "learning_rate": 1.8914921750028804e-05, "loss": 0.8448, "step": 3365 }, { "epoch": 0.53, "grad_norm": 12.664915251452088, "learning_rate": 1.891415762366058e-05, "loss": 0.7772, "step": 3366 }, { "epoch": 0.53, "grad_norm": 22.610154367840074, "learning_rate": 1.8913393243778927e-05, "loss": 0.7887, "step": 3367 }, { "epoch": 0.53, "grad_norm": 15.32555711475076, "learning_rate": 1.8912628610405585e-05, "loss": 0.8175, "step": 3368 }, { "epoch": 0.53, "grad_norm": 17.45858352462135, "learning_rate": 1.89118637235623e-05, "loss": 0.7907, "step": 3369 }, { "epoch": 0.53, "grad_norm": 21.87556641526553, "learning_rate": 1.891109858327082e-05, "loss": 0.9231, "step": 3370 }, { "epoch": 0.53, "grad_norm": 10.770753878588453, "learning_rate": 1.8910333189552913e-05, "loss": 0.6939, "step": 3371 }, { "epoch": 0.53, "grad_norm": 15.903194134884835, "learning_rate": 1.8909567542430344e-05, "loss": 0.7623, "step": 3372 }, { "epoch": 0.53, "grad_norm": 15.033742888282342, "learning_rate": 1.890880164192488e-05, "loss": 0.7945, "step": 3373 }, { "epoch": 0.53, "grad_norm": 11.502029831964597, "learning_rate": 1.8908035488058316e-05, "loss": 0.7376, "step": 3374 }, { "epoch": 0.53, "grad_norm": 18.609909045149866, "learning_rate": 1.8907269080852432e-05, "loss": 0.8078, "step": 3375 }, { "epoch": 0.53, "grad_norm": 13.357426894513708, "learning_rate": 1.8906502420329022e-05, "loss": 0.7522, "step": 3376 }, { "epoch": 0.53, "grad_norm": 21.214632161433357, "learning_rate": 1.8905735506509897e-05, "loss": 0.8444, "step": 3377 }, { "epoch": 0.53, "grad_norm": 21.704193312769636, "learning_rate": 1.890496833941686e-05, "loss": 0.8611, "step": 3378 }, { "epoch": 0.53, "grad_norm": 28.767409683180908, "learning_rate": 1.890420091907174e-05, "loss": 0.8217, "step": 3379 }, { "epoch": 0.53, "grad_norm": 13.99141770017904, "learning_rate": 1.8903433245496348e-05, "loss": 0.736, "step": 3380 }, { "epoch": 0.53, "grad_norm": 24.66708461176765, "learning_rate": 1.8902665318712527e-05, "loss": 0.8313, "step": 3381 }, { "epoch": 0.53, "grad_norm": 14.466920263218196, "learning_rate": 1.8901897138742107e-05, "loss": 0.8133, "step": 3382 }, { "epoch": 0.53, "grad_norm": 20.826265505340814, "learning_rate": 1.890112870560694e-05, "loss": 0.8638, "step": 3383 }, { "epoch": 0.53, "grad_norm": 19.936165660345143, "learning_rate": 1.8900360019328885e-05, "loss": 0.9087, "step": 3384 }, { "epoch": 0.53, "grad_norm": 17.720476132451292, "learning_rate": 1.8899591079929796e-05, "loss": 0.7638, "step": 3385 }, { "epoch": 0.53, "grad_norm": 32.02774678123373, "learning_rate": 1.8898821887431543e-05, "loss": 0.8345, "step": 3386 }, { "epoch": 0.53, "grad_norm": 16.123609201016286, "learning_rate": 1.8898052441855997e-05, "loss": 0.7579, "step": 3387 }, { "epoch": 0.53, "grad_norm": 20.596070170489416, "learning_rate": 1.8897282743225048e-05, "loss": 0.7905, "step": 3388 }, { "epoch": 0.53, "grad_norm": 23.662250099283213, "learning_rate": 1.8896512791560584e-05, "loss": 0.761, "step": 3389 }, { "epoch": 0.53, "grad_norm": 27.909954965665126, "learning_rate": 1.8895742586884502e-05, "loss": 0.8593, "step": 3390 }, { "epoch": 0.53, "grad_norm": 36.3871593925754, "learning_rate": 1.88949721292187e-05, "loss": 0.6885, "step": 3391 }, { "epoch": 0.53, "grad_norm": 18.06501881648598, "learning_rate": 1.8894201418585094e-05, "loss": 0.9004, "step": 3392 }, { "epoch": 0.53, "grad_norm": 19.926497168169508, "learning_rate": 1.8893430455005604e-05, "loss": 0.8322, "step": 3393 }, { "epoch": 0.53, "grad_norm": 17.99428544627486, "learning_rate": 1.8892659238502154e-05, "loss": 0.8215, "step": 3394 }, { "epoch": 0.53, "grad_norm": 21.148237302133445, "learning_rate": 1.889188776909668e-05, "loss": 0.7444, "step": 3395 }, { "epoch": 0.53, "grad_norm": 20.47423651160239, "learning_rate": 1.8891116046811116e-05, "loss": 0.8691, "step": 3396 }, { "epoch": 0.53, "grad_norm": 26.444399667196294, "learning_rate": 1.8890344071667415e-05, "loss": 0.8061, "step": 3397 }, { "epoch": 0.53, "grad_norm": 21.23073772972432, "learning_rate": 1.888957184368753e-05, "loss": 0.7917, "step": 3398 }, { "epoch": 0.53, "grad_norm": 14.70512289607133, "learning_rate": 1.888879936289342e-05, "loss": 0.8484, "step": 3399 }, { "epoch": 0.53, "grad_norm": 23.84407015806765, "learning_rate": 1.8888026629307056e-05, "loss": 0.815, "step": 3400 }, { "epoch": 0.53, "grad_norm": 23.48505899279683, "learning_rate": 1.8887253642950414e-05, "loss": 0.6795, "step": 3401 }, { "epoch": 0.53, "grad_norm": 19.848641319132316, "learning_rate": 1.8886480403845477e-05, "loss": 0.8493, "step": 3402 }, { "epoch": 0.53, "grad_norm": 15.291422213017542, "learning_rate": 1.888570691201424e-05, "loss": 0.7337, "step": 3403 }, { "epoch": 0.53, "grad_norm": 21.811800274651663, "learning_rate": 1.8884933167478688e-05, "loss": 0.8441, "step": 3404 }, { "epoch": 0.53, "grad_norm": 15.537586278207844, "learning_rate": 1.888415917026084e-05, "loss": 0.7338, "step": 3405 }, { "epoch": 0.53, "grad_norm": 10.878420723399753, "learning_rate": 1.8883384920382703e-05, "loss": 0.8015, "step": 3406 }, { "epoch": 0.53, "grad_norm": 13.533743463410543, "learning_rate": 1.8882610417866295e-05, "loss": 0.7818, "step": 3407 }, { "epoch": 0.53, "grad_norm": 27.510857839030827, "learning_rate": 1.888183566273364e-05, "loss": 0.8128, "step": 3408 }, { "epoch": 0.53, "grad_norm": 18.53714177437603, "learning_rate": 1.888106065500678e-05, "loss": 0.7775, "step": 3409 }, { "epoch": 0.53, "grad_norm": 17.42868004840664, "learning_rate": 1.8880285394707743e-05, "loss": 0.7933, "step": 3410 }, { "epoch": 0.53, "grad_norm": 17.942344776774117, "learning_rate": 1.8879509881858587e-05, "loss": 0.8449, "step": 3411 }, { "epoch": 0.53, "grad_norm": 13.97399866407785, "learning_rate": 1.8878734116481364e-05, "loss": 0.7983, "step": 3412 }, { "epoch": 0.53, "grad_norm": 26.976028293476137, "learning_rate": 1.887795809859814e-05, "loss": 0.8435, "step": 3413 }, { "epoch": 0.53, "grad_norm": 21.749414522413606, "learning_rate": 1.8877181828230978e-05, "loss": 0.803, "step": 3414 }, { "epoch": 0.53, "grad_norm": 26.371180048696683, "learning_rate": 1.887640530540196e-05, "loss": 0.7834, "step": 3415 }, { "epoch": 0.53, "grad_norm": 20.418450622973566, "learning_rate": 1.8875628530133164e-05, "loss": 0.8457, "step": 3416 }, { "epoch": 0.53, "grad_norm": 23.53891280145962, "learning_rate": 1.887485150244669e-05, "loss": 0.9259, "step": 3417 }, { "epoch": 0.53, "grad_norm": 15.1538160886726, "learning_rate": 1.8874074222364626e-05, "loss": 0.7541, "step": 3418 }, { "epoch": 0.53, "grad_norm": 23.298196457700744, "learning_rate": 1.8873296689909084e-05, "loss": 0.8898, "step": 3419 }, { "epoch": 0.53, "grad_norm": 20.409744392510575, "learning_rate": 1.8872518905102177e-05, "loss": 0.8425, "step": 3420 }, { "epoch": 0.53, "grad_norm": 17.777615546988404, "learning_rate": 1.887174086796602e-05, "loss": 0.8014, "step": 3421 }, { "epoch": 0.53, "grad_norm": 19.753797718660437, "learning_rate": 1.8870962578522746e-05, "loss": 0.7904, "step": 3422 }, { "epoch": 0.53, "grad_norm": 19.093606963144772, "learning_rate": 1.887018403679448e-05, "loss": 0.8219, "step": 3423 }, { "epoch": 0.53, "grad_norm": 16.201060743220935, "learning_rate": 1.8869405242803373e-05, "loss": 0.7504, "step": 3424 }, { "epoch": 0.53, "grad_norm": 31.126851586435176, "learning_rate": 1.886862619657157e-05, "loss": 0.9006, "step": 3425 }, { "epoch": 0.54, "grad_norm": 23.103589940691286, "learning_rate": 1.8867846898121223e-05, "loss": 0.9261, "step": 3426 }, { "epoch": 0.54, "grad_norm": 15.898382804457848, "learning_rate": 1.8867067347474498e-05, "loss": 0.8594, "step": 3427 }, { "epoch": 0.54, "grad_norm": 13.037837845812941, "learning_rate": 1.8866287544653566e-05, "loss": 0.7686, "step": 3428 }, { "epoch": 0.54, "grad_norm": 24.176376983851878, "learning_rate": 1.8865507489680602e-05, "loss": 0.9257, "step": 3429 }, { "epoch": 0.54, "grad_norm": 15.022398017057679, "learning_rate": 1.8864727182577792e-05, "loss": 0.8068, "step": 3430 }, { "epoch": 0.54, "grad_norm": 23.161244457387706, "learning_rate": 1.8863946623367324e-05, "loss": 0.7708, "step": 3431 }, { "epoch": 0.54, "grad_norm": 14.072629002916553, "learning_rate": 1.88631658120714e-05, "loss": 0.7199, "step": 3432 }, { "epoch": 0.54, "grad_norm": 23.8485785428595, "learning_rate": 1.886238474871223e-05, "loss": 0.8088, "step": 3433 }, { "epoch": 0.54, "grad_norm": 19.682636580650488, "learning_rate": 1.8861603433312017e-05, "loss": 0.8052, "step": 3434 }, { "epoch": 0.54, "grad_norm": 16.554400508601866, "learning_rate": 1.8860821865892988e-05, "loss": 0.7221, "step": 3435 }, { "epoch": 0.54, "grad_norm": 16.396140113928965, "learning_rate": 1.8860040046477367e-05, "loss": 0.7557, "step": 3436 }, { "epoch": 0.54, "grad_norm": 14.612653212013303, "learning_rate": 1.8859257975087395e-05, "loss": 0.7959, "step": 3437 }, { "epoch": 0.54, "grad_norm": 19.94524016854678, "learning_rate": 1.8858475651745304e-05, "loss": 0.8156, "step": 3438 }, { "epoch": 0.54, "grad_norm": 18.27030400103325, "learning_rate": 1.8857693076473348e-05, "loss": 0.7762, "step": 3439 }, { "epoch": 0.54, "grad_norm": 14.113184566678713, "learning_rate": 1.8856910249293783e-05, "loss": 0.7738, "step": 3440 }, { "epoch": 0.54, "grad_norm": 21.12987140278447, "learning_rate": 1.8856127170228873e-05, "loss": 0.8317, "step": 3441 }, { "epoch": 0.54, "grad_norm": 15.703873130546501, "learning_rate": 1.8855343839300885e-05, "loss": 0.7512, "step": 3442 }, { "epoch": 0.54, "grad_norm": 21.85220304606607, "learning_rate": 1.8854560256532098e-05, "loss": 0.7905, "step": 3443 }, { "epoch": 0.54, "grad_norm": 20.6116934174334, "learning_rate": 1.88537764219448e-05, "loss": 0.7489, "step": 3444 }, { "epoch": 0.54, "grad_norm": 16.650546684540593, "learning_rate": 1.885299233556128e-05, "loss": 0.7357, "step": 3445 }, { "epoch": 0.54, "grad_norm": 23.8245985911209, "learning_rate": 1.8852207997403835e-05, "loss": 0.7486, "step": 3446 }, { "epoch": 0.54, "grad_norm": 22.30396428647669, "learning_rate": 1.8851423407494774e-05, "loss": 0.7502, "step": 3447 }, { "epoch": 0.54, "grad_norm": 16.147550420567963, "learning_rate": 1.8850638565856406e-05, "loss": 0.7028, "step": 3448 }, { "epoch": 0.54, "grad_norm": 25.946758679084766, "learning_rate": 1.8849853472511057e-05, "loss": 0.7767, "step": 3449 }, { "epoch": 0.54, "grad_norm": 20.138954583112035, "learning_rate": 1.8849068127481055e-05, "loss": 0.7856, "step": 3450 }, { "epoch": 0.54, "grad_norm": 22.999988545539324, "learning_rate": 1.884828253078873e-05, "loss": 0.808, "step": 3451 }, { "epoch": 0.54, "grad_norm": 17.264519551217536, "learning_rate": 1.8847496682456422e-05, "loss": 0.8131, "step": 3452 }, { "epoch": 0.54, "grad_norm": 18.128320698075523, "learning_rate": 1.884671058250649e-05, "loss": 0.7629, "step": 3453 }, { "epoch": 0.54, "grad_norm": 22.566536444575174, "learning_rate": 1.8845924230961278e-05, "loss": 0.8165, "step": 3454 }, { "epoch": 0.54, "grad_norm": 35.766168226097186, "learning_rate": 1.884513762784316e-05, "loss": 0.8274, "step": 3455 }, { "epoch": 0.54, "grad_norm": 28.15575892515542, "learning_rate": 1.8844350773174502e-05, "loss": 0.7508, "step": 3456 }, { "epoch": 0.54, "grad_norm": 18.40047403957846, "learning_rate": 1.884356366697768e-05, "loss": 0.8564, "step": 3457 }, { "epoch": 0.54, "grad_norm": 17.091814960256134, "learning_rate": 1.8842776309275083e-05, "loss": 0.7324, "step": 3458 }, { "epoch": 0.54, "grad_norm": 19.41725716508165, "learning_rate": 1.8841988700089096e-05, "loss": 0.8842, "step": 3459 }, { "epoch": 0.54, "grad_norm": 20.600226610318046, "learning_rate": 1.8841200839442128e-05, "loss": 0.8848, "step": 3460 }, { "epoch": 0.54, "grad_norm": 16.44865655447884, "learning_rate": 1.8840412727356576e-05, "loss": 0.776, "step": 3461 }, { "epoch": 0.54, "grad_norm": 27.0947566516384, "learning_rate": 1.883962436385486e-05, "loss": 0.7059, "step": 3462 }, { "epoch": 0.54, "grad_norm": 34.218234190689394, "learning_rate": 1.88388357489594e-05, "loss": 0.8564, "step": 3463 }, { "epoch": 0.54, "grad_norm": 15.261275124625653, "learning_rate": 1.883804688269262e-05, "loss": 0.8148, "step": 3464 }, { "epoch": 0.54, "grad_norm": 19.006673130392, "learning_rate": 1.8837257765076956e-05, "loss": 0.8582, "step": 3465 }, { "epoch": 0.54, "grad_norm": 21.97831617132807, "learning_rate": 1.883646839613485e-05, "loss": 0.8575, "step": 3466 }, { "epoch": 0.54, "grad_norm": 19.531856205878626, "learning_rate": 1.8835678775888752e-05, "loss": 0.8249, "step": 3467 }, { "epoch": 0.54, "grad_norm": 12.242212969833199, "learning_rate": 1.883488890436112e-05, "loss": 0.844, "step": 3468 }, { "epoch": 0.54, "grad_norm": 23.218738505526197, "learning_rate": 1.8834098781574415e-05, "loss": 0.785, "step": 3469 }, { "epoch": 0.54, "grad_norm": 14.179441163258888, "learning_rate": 1.883330840755111e-05, "loss": 0.7026, "step": 3470 }, { "epoch": 0.54, "grad_norm": 18.914691652235785, "learning_rate": 1.883251778231368e-05, "loss": 0.837, "step": 3471 }, { "epoch": 0.54, "grad_norm": 35.636151214867326, "learning_rate": 1.883172690588461e-05, "loss": 0.7955, "step": 3472 }, { "epoch": 0.54, "grad_norm": 26.70029326860049, "learning_rate": 1.8830935778286393e-05, "loss": 0.782, "step": 3473 }, { "epoch": 0.54, "grad_norm": 18.354191565379452, "learning_rate": 1.8830144399541533e-05, "loss": 0.7538, "step": 3474 }, { "epoch": 0.54, "grad_norm": 25.22030474964758, "learning_rate": 1.8829352769672525e-05, "loss": 0.8834, "step": 3475 }, { "epoch": 0.54, "grad_norm": 25.52685199868065, "learning_rate": 1.8828560888701895e-05, "loss": 0.7973, "step": 3476 }, { "epoch": 0.54, "grad_norm": 26.280431588977667, "learning_rate": 1.8827768756652155e-05, "loss": 0.8578, "step": 3477 }, { "epoch": 0.54, "grad_norm": 33.79368563561777, "learning_rate": 1.8826976373545838e-05, "loss": 0.9458, "step": 3478 }, { "epoch": 0.54, "grad_norm": 13.586325938225523, "learning_rate": 1.8826183739405475e-05, "loss": 0.7706, "step": 3479 }, { "epoch": 0.54, "grad_norm": 17.495325419543924, "learning_rate": 1.8825390854253605e-05, "loss": 0.8013, "step": 3480 }, { "epoch": 0.54, "grad_norm": 17.841185596979305, "learning_rate": 1.882459771811279e-05, "loss": 0.7905, "step": 3481 }, { "epoch": 0.54, "grad_norm": 27.790002438057645, "learning_rate": 1.8823804331005573e-05, "loss": 0.9765, "step": 3482 }, { "epoch": 0.54, "grad_norm": 15.265754893403322, "learning_rate": 1.8823010692954523e-05, "loss": 0.723, "step": 3483 }, { "epoch": 0.54, "grad_norm": 19.58633743732954, "learning_rate": 1.8822216803982214e-05, "loss": 0.7811, "step": 3484 }, { "epoch": 0.54, "grad_norm": 22.22621006744884, "learning_rate": 1.8821422664111218e-05, "loss": 0.9039, "step": 3485 }, { "epoch": 0.54, "grad_norm": 15.987284765591708, "learning_rate": 1.882062827336412e-05, "loss": 0.787, "step": 3486 }, { "epoch": 0.54, "grad_norm": 10.46565408548851, "learning_rate": 1.8819833631763514e-05, "loss": 0.683, "step": 3487 }, { "epoch": 0.54, "grad_norm": 20.111958333714732, "learning_rate": 1.8819038739332e-05, "loss": 0.8128, "step": 3488 }, { "epoch": 0.54, "grad_norm": 21.848316055120737, "learning_rate": 1.8818243596092182e-05, "loss": 0.8297, "step": 3489 }, { "epoch": 0.55, "grad_norm": 23.662014347126476, "learning_rate": 1.8817448202066677e-05, "loss": 0.88, "step": 3490 }, { "epoch": 0.55, "grad_norm": 14.752111273733133, "learning_rate": 1.88166525572781e-05, "loss": 0.7116, "step": 3491 }, { "epoch": 0.55, "grad_norm": 14.515634826229782, "learning_rate": 1.8815856661749085e-05, "loss": 0.7744, "step": 3492 }, { "epoch": 0.55, "grad_norm": 15.680512253111425, "learning_rate": 1.8815060515502262e-05, "loss": 0.7396, "step": 3493 }, { "epoch": 0.55, "grad_norm": 14.066291974934833, "learning_rate": 1.8814264118560274e-05, "loss": 0.736, "step": 3494 }, { "epoch": 0.55, "grad_norm": 14.829860335486249, "learning_rate": 1.8813467470945774e-05, "loss": 0.761, "step": 3495 }, { "epoch": 0.55, "grad_norm": 38.75692336622185, "learning_rate": 1.8812670572681413e-05, "loss": 0.7411, "step": 3496 }, { "epoch": 0.55, "grad_norm": 17.85813735818634, "learning_rate": 1.8811873423789853e-05, "loss": 0.7823, "step": 3497 }, { "epoch": 0.55, "grad_norm": 18.627602908166725, "learning_rate": 1.8811076024293774e-05, "loss": 0.8384, "step": 3498 }, { "epoch": 0.55, "grad_norm": 33.77782872172686, "learning_rate": 1.881027837421584e-05, "loss": 0.819, "step": 3499 }, { "epoch": 0.55, "grad_norm": 33.12887411315028, "learning_rate": 1.880948047357875e-05, "loss": 0.7636, "step": 3500 }, { "epoch": 0.55, "grad_norm": 15.36707713713431, "learning_rate": 1.8808682322405184e-05, "loss": 0.7357, "step": 3501 }, { "epoch": 0.55, "grad_norm": 15.492143628995628, "learning_rate": 1.8807883920717844e-05, "loss": 0.865, "step": 3502 }, { "epoch": 0.55, "grad_norm": 15.756720148068348, "learning_rate": 1.8807085268539443e-05, "loss": 0.77, "step": 3503 }, { "epoch": 0.55, "grad_norm": 23.94826734523092, "learning_rate": 1.8806286365892685e-05, "loss": 0.8267, "step": 3504 }, { "epoch": 0.55, "grad_norm": 18.55035426808554, "learning_rate": 1.8805487212800297e-05, "loss": 0.7403, "step": 3505 }, { "epoch": 0.55, "grad_norm": 17.823150509122005, "learning_rate": 1.8804687809285003e-05, "loss": 0.8168, "step": 3506 }, { "epoch": 0.55, "grad_norm": 15.015405160582734, "learning_rate": 1.8803888155369538e-05, "loss": 0.7991, "step": 3507 }, { "epoch": 0.55, "grad_norm": 16.309295285606257, "learning_rate": 1.8803088251076642e-05, "loss": 0.8021, "step": 3508 }, { "epoch": 0.55, "grad_norm": 20.422907803898124, "learning_rate": 1.880228809642907e-05, "loss": 0.7982, "step": 3509 }, { "epoch": 0.55, "grad_norm": 23.902390401289402, "learning_rate": 1.880148769144957e-05, "loss": 0.808, "step": 3510 }, { "epoch": 0.55, "grad_norm": 15.874141597630716, "learning_rate": 1.880068703616091e-05, "loss": 0.7047, "step": 3511 }, { "epoch": 0.55, "grad_norm": 24.998691193772636, "learning_rate": 1.8799886130585858e-05, "loss": 0.7697, "step": 3512 }, { "epoch": 0.55, "grad_norm": 19.20763225325, "learning_rate": 1.8799084974747193e-05, "loss": 0.8053, "step": 3513 }, { "epoch": 0.55, "grad_norm": 18.560378502900267, "learning_rate": 1.8798283568667698e-05, "loss": 0.7729, "step": 3514 }, { "epoch": 0.55, "grad_norm": 27.044188167107126, "learning_rate": 1.8797481912370167e-05, "loss": 0.7679, "step": 3515 }, { "epoch": 0.55, "grad_norm": 18.8551137071639, "learning_rate": 1.8796680005877397e-05, "loss": 0.8702, "step": 3516 }, { "epoch": 0.55, "grad_norm": 18.63964457223816, "learning_rate": 1.879587784921219e-05, "loss": 0.794, "step": 3517 }, { "epoch": 0.55, "grad_norm": 14.036531375690243, "learning_rate": 1.8795075442397367e-05, "loss": 0.7565, "step": 3518 }, { "epoch": 0.55, "grad_norm": 22.444531661007105, "learning_rate": 1.879427278545574e-05, "loss": 0.7602, "step": 3519 }, { "epoch": 0.55, "grad_norm": 20.818201867740747, "learning_rate": 1.8793469878410142e-05, "loss": 0.7963, "step": 3520 }, { "epoch": 0.55, "grad_norm": 19.78712336711334, "learning_rate": 1.8792666721283406e-05, "loss": 0.8398, "step": 3521 }, { "epoch": 0.55, "grad_norm": 13.980181112045726, "learning_rate": 1.8791863314098368e-05, "loss": 0.8076, "step": 3522 }, { "epoch": 0.55, "grad_norm": 20.180916078675132, "learning_rate": 1.879105965687788e-05, "loss": 0.8217, "step": 3523 }, { "epoch": 0.55, "grad_norm": 18.73869338943173, "learning_rate": 1.87902557496448e-05, "loss": 0.7327, "step": 3524 }, { "epoch": 0.55, "grad_norm": 17.342797615777204, "learning_rate": 1.878945159242199e-05, "loss": 0.7908, "step": 3525 }, { "epoch": 0.55, "grad_norm": 20.183876213546846, "learning_rate": 1.878864718523232e-05, "loss": 0.7959, "step": 3526 }, { "epoch": 0.55, "grad_norm": 13.645566801084128, "learning_rate": 1.8787842528098657e-05, "loss": 0.7731, "step": 3527 }, { "epoch": 0.55, "grad_norm": 25.787845821507243, "learning_rate": 1.87870376210439e-05, "loss": 0.757, "step": 3528 }, { "epoch": 0.55, "grad_norm": 23.189468052211144, "learning_rate": 1.878623246409093e-05, "loss": 0.8377, "step": 3529 }, { "epoch": 0.55, "grad_norm": 18.148868128724928, "learning_rate": 1.878542705726265e-05, "loss": 0.7908, "step": 3530 }, { "epoch": 0.55, "grad_norm": 17.809226666395197, "learning_rate": 1.8784621400581966e-05, "loss": 0.8775, "step": 3531 }, { "epoch": 0.55, "grad_norm": 16.48633325163417, "learning_rate": 1.8783815494071786e-05, "loss": 0.7944, "step": 3532 }, { "epoch": 0.55, "grad_norm": 18.74523896647049, "learning_rate": 1.8783009337755032e-05, "loss": 0.8008, "step": 3533 }, { "epoch": 0.55, "grad_norm": 20.606981366940992, "learning_rate": 1.878220293165463e-05, "loss": 0.8824, "step": 3534 }, { "epoch": 0.55, "grad_norm": 20.213001332379317, "learning_rate": 1.8781396275793516e-05, "loss": 0.7521, "step": 3535 }, { "epoch": 0.55, "grad_norm": 18.088980811287946, "learning_rate": 1.8780589370194626e-05, "loss": 0.8775, "step": 3536 }, { "epoch": 0.55, "grad_norm": 28.292222384602326, "learning_rate": 1.8779782214880912e-05, "loss": 0.8227, "step": 3537 }, { "epoch": 0.55, "grad_norm": 20.035381930205684, "learning_rate": 1.8778974809875326e-05, "loss": 0.8039, "step": 3538 }, { "epoch": 0.55, "grad_norm": 16.780408127763714, "learning_rate": 1.8778167155200836e-05, "loss": 0.7841, "step": 3539 }, { "epoch": 0.55, "grad_norm": 15.79215376886102, "learning_rate": 1.8777359250880404e-05, "loss": 0.8581, "step": 3540 }, { "epoch": 0.55, "grad_norm": 26.351232227480317, "learning_rate": 1.8776551096937014e-05, "loss": 0.9017, "step": 3541 }, { "epoch": 0.55, "grad_norm": 17.091008297288706, "learning_rate": 1.8775742693393638e-05, "loss": 0.8004, "step": 3542 }, { "epoch": 0.55, "grad_norm": 18.097193069179173, "learning_rate": 1.8774934040273278e-05, "loss": 0.8364, "step": 3543 }, { "epoch": 0.55, "grad_norm": 19.74878090252714, "learning_rate": 1.8774125137598926e-05, "loss": 0.7653, "step": 3544 }, { "epoch": 0.55, "grad_norm": 20.47398130237187, "learning_rate": 1.877331598539359e-05, "loss": 0.8068, "step": 3545 }, { "epoch": 0.55, "grad_norm": 28.430191692316722, "learning_rate": 1.8772506583680273e-05, "loss": 0.9048, "step": 3546 }, { "epoch": 0.55, "grad_norm": 24.8082207946181, "learning_rate": 1.8771696932482008e-05, "loss": 0.8238, "step": 3547 }, { "epoch": 0.55, "grad_norm": 24.417594390523444, "learning_rate": 1.8770887031821813e-05, "loss": 0.903, "step": 3548 }, { "epoch": 0.55, "grad_norm": 20.11493896140036, "learning_rate": 1.8770076881722717e-05, "loss": 0.675, "step": 3549 }, { "epoch": 0.55, "grad_norm": 18.486352414912705, "learning_rate": 1.8769266482207766e-05, "loss": 0.8944, "step": 3550 }, { "epoch": 0.55, "grad_norm": 17.881971707791365, "learning_rate": 1.8768455833300007e-05, "loss": 0.8369, "step": 3551 }, { "epoch": 0.55, "grad_norm": 26.010122825104496, "learning_rate": 1.8767644935022496e-05, "loss": 0.8127, "step": 3552 }, { "epoch": 0.55, "grad_norm": 11.726223891806299, "learning_rate": 1.876683378739829e-05, "loss": 0.6649, "step": 3553 }, { "epoch": 0.56, "grad_norm": 33.54011793628526, "learning_rate": 1.876602239045046e-05, "loss": 0.9012, "step": 3554 }, { "epoch": 0.56, "grad_norm": 17.61798279990791, "learning_rate": 1.876521074420208e-05, "loss": 0.7523, "step": 3555 }, { "epoch": 0.56, "grad_norm": 13.164041866659867, "learning_rate": 1.8764398848676232e-05, "loss": 0.856, "step": 3556 }, { "epoch": 0.56, "grad_norm": 17.812691038345378, "learning_rate": 1.876358670389601e-05, "loss": 0.8713, "step": 3557 }, { "epoch": 0.56, "grad_norm": 19.691827570897786, "learning_rate": 1.8762774309884508e-05, "loss": 0.7943, "step": 3558 }, { "epoch": 0.56, "grad_norm": 21.22391859148357, "learning_rate": 1.876196166666483e-05, "loss": 0.7599, "step": 3559 }, { "epoch": 0.56, "grad_norm": 22.280253101801907, "learning_rate": 1.876114877426009e-05, "loss": 0.7013, "step": 3560 }, { "epoch": 0.56, "grad_norm": 21.241165103207855, "learning_rate": 1.8760335632693404e-05, "loss": 0.7498, "step": 3561 }, { "epoch": 0.56, "grad_norm": 21.492375995276703, "learning_rate": 1.8759522241987896e-05, "loss": 0.8936, "step": 3562 }, { "epoch": 0.56, "grad_norm": 18.047379567577668, "learning_rate": 1.87587086021667e-05, "loss": 0.7544, "step": 3563 }, { "epoch": 0.56, "grad_norm": 21.261741974576328, "learning_rate": 1.8757894713252956e-05, "loss": 0.7377, "step": 3564 }, { "epoch": 0.56, "grad_norm": 12.04700849047223, "learning_rate": 1.8757080575269808e-05, "loss": 0.7349, "step": 3565 }, { "epoch": 0.56, "grad_norm": 16.582964809993896, "learning_rate": 1.8756266188240413e-05, "loss": 0.8271, "step": 3566 }, { "epoch": 0.56, "grad_norm": 19.764698830116274, "learning_rate": 1.875545155218793e-05, "loss": 0.7742, "step": 3567 }, { "epoch": 0.56, "grad_norm": 17.89655761220839, "learning_rate": 1.8754636667135523e-05, "loss": 0.8057, "step": 3568 }, { "epoch": 0.56, "grad_norm": 13.626023173192689, "learning_rate": 1.8753821533106373e-05, "loss": 0.8656, "step": 3569 }, { "epoch": 0.56, "grad_norm": 24.137613945736003, "learning_rate": 1.8753006150123662e-05, "loss": 0.7823, "step": 3570 }, { "epoch": 0.56, "grad_norm": 20.26057195828669, "learning_rate": 1.8752190518210572e-05, "loss": 0.8447, "step": 3571 }, { "epoch": 0.56, "grad_norm": 24.026301947296844, "learning_rate": 1.8751374637390304e-05, "loss": 0.8068, "step": 3572 }, { "epoch": 0.56, "grad_norm": 20.285851063941212, "learning_rate": 1.8750558507686065e-05, "loss": 0.8071, "step": 3573 }, { "epoch": 0.56, "grad_norm": 24.57400438624219, "learning_rate": 1.8749742129121058e-05, "loss": 0.8453, "step": 3574 }, { "epoch": 0.56, "grad_norm": 16.30709197570784, "learning_rate": 1.87489255017185e-05, "loss": 0.7422, "step": 3575 }, { "epoch": 0.56, "grad_norm": 21.731462238209005, "learning_rate": 1.8748108625501623e-05, "loss": 0.9153, "step": 3576 }, { "epoch": 0.56, "grad_norm": 25.09779022359157, "learning_rate": 1.8747291500493653e-05, "loss": 0.7667, "step": 3577 }, { "epoch": 0.56, "grad_norm": 17.817954822661456, "learning_rate": 1.874647412671783e-05, "loss": 0.809, "step": 3578 }, { "epoch": 0.56, "grad_norm": 23.494904352463458, "learning_rate": 1.87456565041974e-05, "loss": 0.863, "step": 3579 }, { "epoch": 0.56, "grad_norm": 20.312299238826743, "learning_rate": 1.8744838632955617e-05, "loss": 0.8786, "step": 3580 }, { "epoch": 0.56, "grad_norm": 29.869929817642944, "learning_rate": 1.8744020513015733e-05, "loss": 0.915, "step": 3581 }, { "epoch": 0.56, "grad_norm": 19.34152127146327, "learning_rate": 1.8743202144401027e-05, "loss": 0.7927, "step": 3582 }, { "epoch": 0.56, "grad_norm": 19.800781097231496, "learning_rate": 1.874238352713476e-05, "loss": 0.8538, "step": 3583 }, { "epoch": 0.56, "grad_norm": 16.98585337606707, "learning_rate": 1.8741564661240223e-05, "loss": 0.7454, "step": 3584 }, { "epoch": 0.56, "grad_norm": 24.135990223814957, "learning_rate": 1.87407455467407e-05, "loss": 0.9235, "step": 3585 }, { "epoch": 0.56, "grad_norm": 14.53984675067897, "learning_rate": 1.8739926183659492e-05, "loss": 0.8527, "step": 3586 }, { "epoch": 0.56, "grad_norm": 18.051889720266622, "learning_rate": 1.873910657201989e-05, "loss": 0.7914, "step": 3587 }, { "epoch": 0.56, "grad_norm": 12.075225415406303, "learning_rate": 1.873828671184521e-05, "loss": 0.6688, "step": 3588 }, { "epoch": 0.56, "grad_norm": 13.689917363989505, "learning_rate": 1.873746660315877e-05, "loss": 0.6685, "step": 3589 }, { "epoch": 0.56, "grad_norm": 20.073484476381477, "learning_rate": 1.873664624598389e-05, "loss": 0.8412, "step": 3590 }, { "epoch": 0.56, "grad_norm": 18.233145370390293, "learning_rate": 1.8735825640343904e-05, "loss": 0.7767, "step": 3591 }, { "epoch": 0.56, "grad_norm": 16.73616713990554, "learning_rate": 1.8735004786262142e-05, "loss": 0.7929, "step": 3592 }, { "epoch": 0.56, "grad_norm": 19.424976564636232, "learning_rate": 1.8734183683761957e-05, "loss": 0.7167, "step": 3593 }, { "epoch": 0.56, "grad_norm": 16.365261378791256, "learning_rate": 1.8733362332866698e-05, "loss": 0.7045, "step": 3594 }, { "epoch": 0.56, "grad_norm": 11.050867687142146, "learning_rate": 1.873254073359972e-05, "loss": 0.7009, "step": 3595 }, { "epoch": 0.56, "grad_norm": 15.590000056221823, "learning_rate": 1.8731718885984396e-05, "loss": 0.7316, "step": 3596 }, { "epoch": 0.56, "grad_norm": 16.923169066313335, "learning_rate": 1.8730896790044092e-05, "loss": 0.8225, "step": 3597 }, { "epoch": 0.56, "grad_norm": 15.301342875710873, "learning_rate": 1.873007444580219e-05, "loss": 0.6974, "step": 3598 }, { "epoch": 0.56, "grad_norm": 19.55983581164699, "learning_rate": 1.872925185328208e-05, "loss": 0.8101, "step": 3599 }, { "epoch": 0.56, "grad_norm": 21.681606536335355, "learning_rate": 1.872842901250715e-05, "loss": 0.8205, "step": 3600 }, { "epoch": 0.56, "grad_norm": 18.115853660671004, "learning_rate": 1.872760592350081e-05, "loss": 0.7441, "step": 3601 }, { "epoch": 0.56, "grad_norm": 15.91763958736906, "learning_rate": 1.872678258628646e-05, "loss": 0.7799, "step": 3602 }, { "epoch": 0.56, "grad_norm": 30.306542040101633, "learning_rate": 1.872595900088752e-05, "loss": 0.8051, "step": 3603 }, { "epoch": 0.56, "grad_norm": 27.23274052060126, "learning_rate": 1.872513516732741e-05, "loss": 0.847, "step": 3604 }, { "epoch": 0.56, "grad_norm": 21.02108160886153, "learning_rate": 1.872431108562956e-05, "loss": 0.8093, "step": 3605 }, { "epoch": 0.56, "grad_norm": 26.913484732152558, "learning_rate": 1.8723486755817405e-05, "loss": 0.7594, "step": 3606 }, { "epoch": 0.56, "grad_norm": 20.98816288542478, "learning_rate": 1.872266217791439e-05, "loss": 0.8622, "step": 3607 }, { "epoch": 0.56, "grad_norm": 16.278605511314908, "learning_rate": 1.8721837351943964e-05, "loss": 0.7945, "step": 3608 }, { "epoch": 0.56, "grad_norm": 17.05202768757929, "learning_rate": 1.8721012277929588e-05, "loss": 0.7563, "step": 3609 }, { "epoch": 0.56, "grad_norm": 13.60643878756732, "learning_rate": 1.8720186955894722e-05, "loss": 0.7403, "step": 3610 }, { "epoch": 0.56, "grad_norm": 23.26520798311354, "learning_rate": 1.8719361385862843e-05, "loss": 0.6803, "step": 3611 }, { "epoch": 0.56, "grad_norm": 15.350552483449457, "learning_rate": 1.8718535567857426e-05, "loss": 0.8566, "step": 3612 }, { "epoch": 0.56, "grad_norm": 14.610469631250647, "learning_rate": 1.8717709501901956e-05, "loss": 0.8075, "step": 3613 }, { "epoch": 0.56, "grad_norm": 22.323171734891655, "learning_rate": 1.8716883188019932e-05, "loss": 0.733, "step": 3614 }, { "epoch": 0.56, "grad_norm": 25.128167239825995, "learning_rate": 1.8716056626234848e-05, "loss": 0.9516, "step": 3615 }, { "epoch": 0.56, "grad_norm": 18.73150825360642, "learning_rate": 1.871522981657021e-05, "loss": 0.8248, "step": 3616 }, { "epoch": 0.56, "grad_norm": 22.36607078445375, "learning_rate": 1.8714402759049535e-05, "loss": 0.8285, "step": 3617 }, { "epoch": 0.57, "grad_norm": 19.57147295931924, "learning_rate": 1.8713575453696344e-05, "loss": 0.9086, "step": 3618 }, { "epoch": 0.57, "grad_norm": 21.40856968082394, "learning_rate": 1.8712747900534164e-05, "loss": 0.7795, "step": 3619 }, { "epoch": 0.57, "grad_norm": 12.859809590037942, "learning_rate": 1.871192009958653e-05, "loss": 0.7441, "step": 3620 }, { "epoch": 0.57, "grad_norm": 19.53025016819523, "learning_rate": 1.8711092050876982e-05, "loss": 0.9138, "step": 3621 }, { "epoch": 0.57, "grad_norm": 34.76080812152451, "learning_rate": 1.8710263754429076e-05, "loss": 0.8394, "step": 3622 }, { "epoch": 0.57, "grad_norm": 14.259602599968307, "learning_rate": 1.870943521026636e-05, "loss": 0.7928, "step": 3623 }, { "epoch": 0.57, "grad_norm": 11.41764770275707, "learning_rate": 1.8708606418412408e-05, "loss": 0.7217, "step": 3624 }, { "epoch": 0.57, "grad_norm": 22.228803090629167, "learning_rate": 1.8707777378890778e-05, "loss": 0.7196, "step": 3625 }, { "epoch": 0.57, "grad_norm": 13.779308315302089, "learning_rate": 1.870694809172506e-05, "loss": 0.6867, "step": 3626 }, { "epoch": 0.57, "grad_norm": 17.212870471543894, "learning_rate": 1.8706118556938826e-05, "loss": 0.8893, "step": 3627 }, { "epoch": 0.57, "grad_norm": 21.288949869540392, "learning_rate": 1.870528877455567e-05, "loss": 0.7834, "step": 3628 }, { "epoch": 0.57, "grad_norm": 27.652253198347413, "learning_rate": 1.8704458744599202e-05, "loss": 0.7959, "step": 3629 }, { "epoch": 0.57, "grad_norm": 24.136165796846026, "learning_rate": 1.8703628467093017e-05, "loss": 0.7148, "step": 3630 }, { "epoch": 0.57, "grad_norm": 18.19935266429516, "learning_rate": 1.8702797942060725e-05, "loss": 0.7861, "step": 3631 }, { "epoch": 0.57, "grad_norm": 14.491594517855711, "learning_rate": 1.8701967169525958e-05, "loss": 0.7346, "step": 3632 }, { "epoch": 0.57, "grad_norm": 25.613146952174763, "learning_rate": 1.8701136149512328e-05, "loss": 0.8124, "step": 3633 }, { "epoch": 0.57, "grad_norm": 25.47404466251857, "learning_rate": 1.870030488204348e-05, "loss": 0.7995, "step": 3634 }, { "epoch": 0.57, "grad_norm": 18.611119590517344, "learning_rate": 1.869947336714305e-05, "loss": 0.7885, "step": 3635 }, { "epoch": 0.57, "grad_norm": 18.579567387238896, "learning_rate": 1.8698641604834687e-05, "loss": 0.7388, "step": 3636 }, { "epoch": 0.57, "grad_norm": 30.468340198010637, "learning_rate": 1.8697809595142042e-05, "loss": 0.8654, "step": 3637 }, { "epoch": 0.57, "grad_norm": 21.634382934161103, "learning_rate": 1.869697733808878e-05, "loss": 0.7932, "step": 3638 }, { "epoch": 0.57, "grad_norm": 15.187628459485262, "learning_rate": 1.8696144833698574e-05, "loss": 0.645, "step": 3639 }, { "epoch": 0.57, "grad_norm": 22.130600509361752, "learning_rate": 1.8695312081995096e-05, "loss": 0.8204, "step": 3640 }, { "epoch": 0.57, "grad_norm": 27.831058635649146, "learning_rate": 1.8694479083002027e-05, "loss": 0.8136, "step": 3641 }, { "epoch": 0.57, "grad_norm": 22.409522530523994, "learning_rate": 1.8693645836743062e-05, "loss": 0.8237, "step": 3642 }, { "epoch": 0.57, "grad_norm": 17.235013782217933, "learning_rate": 1.8692812343241892e-05, "loss": 0.7363, "step": 3643 }, { "epoch": 0.57, "grad_norm": 17.526979839661074, "learning_rate": 1.8691978602522226e-05, "loss": 0.8215, "step": 3644 }, { "epoch": 0.57, "grad_norm": 18.583651183965696, "learning_rate": 1.869114461460777e-05, "loss": 0.7235, "step": 3645 }, { "epoch": 0.57, "grad_norm": 20.26246913640462, "learning_rate": 1.8690310379522247e-05, "loss": 0.7706, "step": 3646 }, { "epoch": 0.57, "grad_norm": 19.183451956300864, "learning_rate": 1.8689475897289383e-05, "loss": 0.8115, "step": 3647 }, { "epoch": 0.57, "grad_norm": 19.555517784457376, "learning_rate": 1.8688641167932906e-05, "loss": 0.7744, "step": 3648 }, { "epoch": 0.57, "grad_norm": 17.314369426079622, "learning_rate": 1.8687806191476555e-05, "loss": 0.8448, "step": 3649 }, { "epoch": 0.57, "grad_norm": 57.21910759185438, "learning_rate": 1.8686970967944082e-05, "loss": 0.911, "step": 3650 }, { "epoch": 0.57, "grad_norm": 20.292246053565233, "learning_rate": 1.8686135497359233e-05, "loss": 0.7652, "step": 3651 }, { "epoch": 0.57, "grad_norm": 28.760178358034118, "learning_rate": 1.868529977974577e-05, "loss": 0.8735, "step": 3652 }, { "epoch": 0.57, "grad_norm": 18.26116852854488, "learning_rate": 1.868446381512747e-05, "loss": 0.763, "step": 3653 }, { "epoch": 0.57, "grad_norm": 21.453039699809953, "learning_rate": 1.868362760352809e-05, "loss": 0.818, "step": 3654 }, { "epoch": 0.57, "grad_norm": 26.40508058486648, "learning_rate": 1.8682791144971426e-05, "loss": 0.7489, "step": 3655 }, { "epoch": 0.57, "grad_norm": 17.359591276708297, "learning_rate": 1.8681954439481262e-05, "loss": 0.7248, "step": 3656 }, { "epoch": 0.57, "grad_norm": 15.872460952596, "learning_rate": 1.868111748708139e-05, "loss": 0.7021, "step": 3657 }, { "epoch": 0.57, "grad_norm": 28.487588763402098, "learning_rate": 1.8680280287795612e-05, "loss": 0.7719, "step": 3658 }, { "epoch": 0.57, "grad_norm": 24.57308009136288, "learning_rate": 1.8679442841647747e-05, "loss": 0.7908, "step": 3659 }, { "epoch": 0.57, "grad_norm": 21.86155825162964, "learning_rate": 1.8678605148661602e-05, "loss": 0.7919, "step": 3660 }, { "epoch": 0.57, "grad_norm": 27.30456815592562, "learning_rate": 1.8677767208861e-05, "loss": 0.8657, "step": 3661 }, { "epoch": 0.57, "grad_norm": 16.672394869080815, "learning_rate": 1.867692902226978e-05, "loss": 0.7507, "step": 3662 }, { "epoch": 0.57, "grad_norm": 17.984745160195164, "learning_rate": 1.8676090588911774e-05, "loss": 0.8663, "step": 3663 }, { "epoch": 0.57, "grad_norm": 18.687202469983326, "learning_rate": 1.8675251908810824e-05, "loss": 0.8457, "step": 3664 }, { "epoch": 0.57, "grad_norm": 20.454990390052906, "learning_rate": 1.867441298199079e-05, "loss": 0.8896, "step": 3665 }, { "epoch": 0.57, "grad_norm": 16.650155715851447, "learning_rate": 1.867357380847552e-05, "loss": 0.8179, "step": 3666 }, { "epoch": 0.57, "grad_norm": 18.549687040618135, "learning_rate": 1.8672734388288883e-05, "loss": 0.8302, "step": 3667 }, { "epoch": 0.57, "grad_norm": 36.85061458123028, "learning_rate": 1.867189472145476e-05, "loss": 0.8611, "step": 3668 }, { "epoch": 0.57, "grad_norm": 25.57156035807646, "learning_rate": 1.867105480799702e-05, "loss": 0.8497, "step": 3669 }, { "epoch": 0.57, "grad_norm": 22.301634964109933, "learning_rate": 1.867021464793955e-05, "loss": 0.9199, "step": 3670 }, { "epoch": 0.57, "grad_norm": 28.41043453638851, "learning_rate": 1.8669374241306256e-05, "loss": 0.8815, "step": 3671 }, { "epoch": 0.57, "grad_norm": 18.0655999081036, "learning_rate": 1.8668533588121026e-05, "loss": 0.6696, "step": 3672 }, { "epoch": 0.57, "grad_norm": 28.456877733810106, "learning_rate": 1.866769268840777e-05, "loss": 0.7239, "step": 3673 }, { "epoch": 0.57, "grad_norm": 21.17869941548231, "learning_rate": 1.8666851542190406e-05, "loss": 0.8522, "step": 3674 }, { "epoch": 0.57, "grad_norm": 15.974652400007232, "learning_rate": 1.8666010149492853e-05, "loss": 0.8909, "step": 3675 }, { "epoch": 0.57, "grad_norm": 21.414487590279492, "learning_rate": 1.8665168510339043e-05, "loss": 0.8369, "step": 3676 }, { "epoch": 0.57, "grad_norm": 15.600487558833734, "learning_rate": 1.866432662475291e-05, "loss": 0.7624, "step": 3677 }, { "epoch": 0.57, "grad_norm": 21.683341203338003, "learning_rate": 1.8663484492758393e-05, "loss": 0.7371, "step": 3678 }, { "epoch": 0.57, "grad_norm": 22.098718588319226, "learning_rate": 1.8662642114379444e-05, "loss": 0.8288, "step": 3679 }, { "epoch": 0.57, "grad_norm": 28.80406669918209, "learning_rate": 1.8661799489640025e-05, "loss": 0.8039, "step": 3680 }, { "epoch": 0.57, "grad_norm": 30.536797920397884, "learning_rate": 1.8660956618564095e-05, "loss": 0.8121, "step": 3681 }, { "epoch": 0.58, "grad_norm": 5.463780093074181, "learning_rate": 1.8660113501175622e-05, "loss": 0.8281, "step": 3682 }, { "epoch": 0.58, "grad_norm": 13.134779266009994, "learning_rate": 1.8659270137498588e-05, "loss": 0.8206, "step": 3683 }, { "epoch": 0.58, "grad_norm": 15.628277560276816, "learning_rate": 1.865842652755698e-05, "loss": 0.7461, "step": 3684 }, { "epoch": 0.58, "grad_norm": 11.614731937008038, "learning_rate": 1.8657582671374787e-05, "loss": 0.6644, "step": 3685 }, { "epoch": 0.58, "grad_norm": 19.265353233388268, "learning_rate": 1.8656738568976003e-05, "loss": 0.8624, "step": 3686 }, { "epoch": 0.58, "grad_norm": 40.75501088477457, "learning_rate": 1.865589422038464e-05, "loss": 0.9502, "step": 3687 }, { "epoch": 0.58, "grad_norm": 15.772324339522127, "learning_rate": 1.865504962562471e-05, "loss": 0.8153, "step": 3688 }, { "epoch": 0.58, "grad_norm": 33.304020535159474, "learning_rate": 1.865420478472023e-05, "loss": 0.8258, "step": 3689 }, { "epoch": 0.58, "grad_norm": 15.301013521861792, "learning_rate": 1.865335969769523e-05, "loss": 0.792, "step": 3690 }, { "epoch": 0.58, "grad_norm": 16.082822833479007, "learning_rate": 1.865251436457374e-05, "loss": 0.7324, "step": 3691 }, { "epoch": 0.58, "grad_norm": 24.465329748446127, "learning_rate": 1.8651668785379806e-05, "loss": 0.8932, "step": 3692 }, { "epoch": 0.58, "grad_norm": 16.513643689980302, "learning_rate": 1.865082296013747e-05, "loss": 0.7207, "step": 3693 }, { "epoch": 0.58, "grad_norm": 16.25021015911653, "learning_rate": 1.864997688887079e-05, "loss": 0.8304, "step": 3694 }, { "epoch": 0.58, "grad_norm": 16.24651118111633, "learning_rate": 1.864913057160383e-05, "loss": 0.7584, "step": 3695 }, { "epoch": 0.58, "grad_norm": 16.393855193902642, "learning_rate": 1.8648284008360654e-05, "loss": 0.7448, "step": 3696 }, { "epoch": 0.58, "grad_norm": 15.497330229931306, "learning_rate": 1.864743719916534e-05, "loss": 0.7827, "step": 3697 }, { "epoch": 0.58, "grad_norm": 16.0206140060649, "learning_rate": 1.8646590144041972e-05, "loss": 0.7568, "step": 3698 }, { "epoch": 0.58, "grad_norm": 18.920893952006146, "learning_rate": 1.8645742843014635e-05, "loss": 0.7157, "step": 3699 }, { "epoch": 0.58, "grad_norm": 20.773234533872827, "learning_rate": 1.8644895296107432e-05, "loss": 0.8918, "step": 3700 }, { "epoch": 0.58, "grad_norm": 14.331823485254711, "learning_rate": 1.8644047503344465e-05, "loss": 0.7402, "step": 3701 }, { "epoch": 0.58, "grad_norm": 17.397191009440277, "learning_rate": 1.8643199464749843e-05, "loss": 0.8009, "step": 3702 }, { "epoch": 0.58, "grad_norm": 17.073646718228165, "learning_rate": 1.8642351180347682e-05, "loss": 0.7898, "step": 3703 }, { "epoch": 0.58, "grad_norm": 22.889933136258644, "learning_rate": 1.864150265016211e-05, "loss": 0.8683, "step": 3704 }, { "epoch": 0.58, "grad_norm": 22.081720816621452, "learning_rate": 1.8640653874217257e-05, "loss": 0.7606, "step": 3705 }, { "epoch": 0.58, "grad_norm": 20.892132729651188, "learning_rate": 1.8639804852537264e-05, "loss": 0.749, "step": 3706 }, { "epoch": 0.58, "grad_norm": 22.552952171787748, "learning_rate": 1.8638955585146277e-05, "loss": 0.8249, "step": 3707 }, { "epoch": 0.58, "grad_norm": 17.715828096238656, "learning_rate": 1.8638106072068443e-05, "loss": 0.8625, "step": 3708 }, { "epoch": 0.58, "grad_norm": 24.97202987995221, "learning_rate": 1.863725631332793e-05, "loss": 0.7885, "step": 3709 }, { "epoch": 0.58, "grad_norm": 20.576101891571074, "learning_rate": 1.8636406308948894e-05, "loss": 0.716, "step": 3710 }, { "epoch": 0.58, "grad_norm": 18.46503447903622, "learning_rate": 1.863555605895552e-05, "loss": 0.8587, "step": 3711 }, { "epoch": 0.58, "grad_norm": 16.752132537564197, "learning_rate": 1.863470556337198e-05, "loss": 0.7506, "step": 3712 }, { "epoch": 0.58, "grad_norm": 14.947742279101046, "learning_rate": 1.8633854822222467e-05, "loss": 0.8411, "step": 3713 }, { "epoch": 0.58, "grad_norm": 28.89918343127105, "learning_rate": 1.8633003835531172e-05, "loss": 0.8768, "step": 3714 }, { "epoch": 0.58, "grad_norm": 21.52785553286638, "learning_rate": 1.86321526033223e-05, "loss": 0.8796, "step": 3715 }, { "epoch": 0.58, "grad_norm": 19.747557255120682, "learning_rate": 1.8631301125620056e-05, "loss": 0.887, "step": 3716 }, { "epoch": 0.58, "grad_norm": 15.558054906778917, "learning_rate": 1.8630449402448655e-05, "loss": 0.8198, "step": 3717 }, { "epoch": 0.58, "grad_norm": 12.103131864803347, "learning_rate": 1.8629597433832326e-05, "loss": 0.7212, "step": 3718 }, { "epoch": 0.58, "grad_norm": 16.514996598774914, "learning_rate": 1.8628745219795292e-05, "loss": 0.8049, "step": 3719 }, { "epoch": 0.58, "grad_norm": 15.416992104377439, "learning_rate": 1.862789276036179e-05, "loss": 0.8523, "step": 3720 }, { "epoch": 0.58, "grad_norm": 17.69594792705124, "learning_rate": 1.862704005555607e-05, "loss": 0.7291, "step": 3721 }, { "epoch": 0.58, "grad_norm": 4.649249725546298, "learning_rate": 1.8626187105402373e-05, "loss": 0.8157, "step": 3722 }, { "epoch": 0.58, "grad_norm": 19.471889943738677, "learning_rate": 1.8625333909924964e-05, "loss": 0.8382, "step": 3723 }, { "epoch": 0.58, "grad_norm": 10.930556376606772, "learning_rate": 1.8624480469148103e-05, "loss": 0.7533, "step": 3724 }, { "epoch": 0.58, "grad_norm": 18.18798221443737, "learning_rate": 1.862362678309606e-05, "loss": 0.7067, "step": 3725 }, { "epoch": 0.58, "grad_norm": 33.96096729249176, "learning_rate": 1.862277285179312e-05, "loss": 0.7739, "step": 3726 }, { "epoch": 0.58, "grad_norm": 23.539033528287007, "learning_rate": 1.8621918675263564e-05, "loss": 0.8901, "step": 3727 }, { "epoch": 0.58, "grad_norm": 24.943651439129837, "learning_rate": 1.862106425353168e-05, "loss": 0.8123, "step": 3728 }, { "epoch": 0.58, "grad_norm": 15.068993937394005, "learning_rate": 1.8620209586621778e-05, "loss": 0.6734, "step": 3729 }, { "epoch": 0.58, "grad_norm": 13.885760329232188, "learning_rate": 1.8619354674558155e-05, "loss": 0.7728, "step": 3730 }, { "epoch": 0.58, "grad_norm": 20.392613563181637, "learning_rate": 1.8618499517365127e-05, "loss": 0.8404, "step": 3731 }, { "epoch": 0.58, "grad_norm": 14.644081427299254, "learning_rate": 1.8617644115067013e-05, "loss": 0.6939, "step": 3732 }, { "epoch": 0.58, "grad_norm": 26.357701271049024, "learning_rate": 1.8616788467688146e-05, "loss": 0.7877, "step": 3733 }, { "epoch": 0.58, "grad_norm": 18.70186861458413, "learning_rate": 1.8615932575252855e-05, "loss": 0.8398, "step": 3734 }, { "epoch": 0.58, "grad_norm": 15.065060152623122, "learning_rate": 1.8615076437785474e-05, "loss": 0.8109, "step": 3735 }, { "epoch": 0.58, "grad_norm": 17.992245761414257, "learning_rate": 1.8614220055310367e-05, "loss": 0.7348, "step": 3736 }, { "epoch": 0.58, "grad_norm": 15.539017242837538, "learning_rate": 1.8613363427851875e-05, "loss": 0.77, "step": 3737 }, { "epoch": 0.58, "grad_norm": 18.89321930033038, "learning_rate": 1.8612506555434372e-05, "loss": 0.748, "step": 3738 }, { "epoch": 0.58, "grad_norm": 16.259032747631917, "learning_rate": 1.861164943808222e-05, "loss": 0.7775, "step": 3739 }, { "epoch": 0.58, "grad_norm": 22.41070148109956, "learning_rate": 1.861079207581979e-05, "loss": 0.8158, "step": 3740 }, { "epoch": 0.58, "grad_norm": 17.338031472144, "learning_rate": 1.8609934468671474e-05, "loss": 0.848, "step": 3741 }, { "epoch": 0.58, "grad_norm": 20.376155894349836, "learning_rate": 1.8609076616661658e-05, "loss": 0.8179, "step": 3742 }, { "epoch": 0.58, "grad_norm": 20.957199003573866, "learning_rate": 1.860821851981474e-05, "loss": 0.8948, "step": 3743 }, { "epoch": 0.58, "grad_norm": 20.75913222270166, "learning_rate": 1.8607360178155123e-05, "loss": 0.7616, "step": 3744 }, { "epoch": 0.58, "grad_norm": 12.465591898348475, "learning_rate": 1.8606501591707217e-05, "loss": 0.8039, "step": 3745 }, { "epoch": 0.59, "grad_norm": 18.581807346691082, "learning_rate": 1.860564276049544e-05, "loss": 0.7629, "step": 3746 }, { "epoch": 0.59, "grad_norm": 14.018448676515995, "learning_rate": 1.860478368454422e-05, "loss": 0.8146, "step": 3747 }, { "epoch": 0.59, "grad_norm": 22.529173518563663, "learning_rate": 1.860392436387798e-05, "loss": 0.875, "step": 3748 }, { "epoch": 0.59, "grad_norm": 21.22668664928939, "learning_rate": 1.8603064798521168e-05, "loss": 0.8169, "step": 3749 }, { "epoch": 0.59, "grad_norm": 17.648865502914557, "learning_rate": 1.8602204988498224e-05, "loss": 0.6834, "step": 3750 }, { "epoch": 0.59, "grad_norm": 24.235708710342703, "learning_rate": 1.8601344933833605e-05, "loss": 0.787, "step": 3751 }, { "epoch": 0.59, "grad_norm": 16.519268344095945, "learning_rate": 1.8600484634551767e-05, "loss": 0.8487, "step": 3752 }, { "epoch": 0.59, "grad_norm": 16.483473850314812, "learning_rate": 1.8599624090677174e-05, "loss": 0.8656, "step": 3753 }, { "epoch": 0.59, "grad_norm": 22.485350888528718, "learning_rate": 1.8598763302234306e-05, "loss": 0.8484, "step": 3754 }, { "epoch": 0.59, "grad_norm": 14.958964163665327, "learning_rate": 1.859790226924764e-05, "loss": 0.7221, "step": 3755 }, { "epoch": 0.59, "grad_norm": 24.43862765361714, "learning_rate": 1.8597040991741663e-05, "loss": 0.8133, "step": 3756 }, { "epoch": 0.59, "grad_norm": 14.426216796530715, "learning_rate": 1.8596179469740863e-05, "loss": 0.8019, "step": 3757 }, { "epoch": 0.59, "grad_norm": 17.811719792830015, "learning_rate": 1.8595317703269756e-05, "loss": 0.8006, "step": 3758 }, { "epoch": 0.59, "grad_norm": 29.485282200702745, "learning_rate": 1.859445569235284e-05, "loss": 0.7719, "step": 3759 }, { "epoch": 0.59, "grad_norm": 18.848862857046242, "learning_rate": 1.8593593437014627e-05, "loss": 0.9183, "step": 3760 }, { "epoch": 0.59, "grad_norm": 32.01885737200636, "learning_rate": 1.8592730937279647e-05, "loss": 0.8939, "step": 3761 }, { "epoch": 0.59, "grad_norm": 15.799788601583767, "learning_rate": 1.8591868193172423e-05, "loss": 0.7175, "step": 3762 }, { "epoch": 0.59, "grad_norm": 23.225276019629412, "learning_rate": 1.8591005204717498e-05, "loss": 0.7856, "step": 3763 }, { "epoch": 0.59, "grad_norm": 22.592885455927046, "learning_rate": 1.859014197193941e-05, "loss": 0.7464, "step": 3764 }, { "epoch": 0.59, "grad_norm": 23.39088362396005, "learning_rate": 1.8589278494862707e-05, "loss": 0.8755, "step": 3765 }, { "epoch": 0.59, "grad_norm": 21.106150836220248, "learning_rate": 1.858841477351195e-05, "loss": 0.7518, "step": 3766 }, { "epoch": 0.59, "grad_norm": 21.785924575441783, "learning_rate": 1.85875508079117e-05, "loss": 0.8277, "step": 3767 }, { "epoch": 0.59, "grad_norm": 32.84390877051048, "learning_rate": 1.858668659808653e-05, "loss": 0.8027, "step": 3768 }, { "epoch": 0.59, "grad_norm": 19.61427228954143, "learning_rate": 1.8585822144061013e-05, "loss": 0.8436, "step": 3769 }, { "epoch": 0.59, "grad_norm": 15.665475394438053, "learning_rate": 1.858495744585974e-05, "loss": 0.6795, "step": 3770 }, { "epoch": 0.59, "grad_norm": 13.333938397397688, "learning_rate": 1.8584092503507296e-05, "loss": 0.7503, "step": 3771 }, { "epoch": 0.59, "grad_norm": 15.391209292576, "learning_rate": 1.8583227317028287e-05, "loss": 0.7334, "step": 3772 }, { "epoch": 0.59, "grad_norm": 21.32468939906611, "learning_rate": 1.8582361886447315e-05, "loss": 0.84, "step": 3773 }, { "epoch": 0.59, "grad_norm": 21.89209096820718, "learning_rate": 1.8581496211788985e-05, "loss": 0.722, "step": 3774 }, { "epoch": 0.59, "grad_norm": 15.059009598705948, "learning_rate": 1.8580630293077925e-05, "loss": 0.7838, "step": 3775 }, { "epoch": 0.59, "grad_norm": 21.446796155028977, "learning_rate": 1.857976413033876e-05, "loss": 0.7523, "step": 3776 }, { "epoch": 0.59, "grad_norm": 15.836283488509583, "learning_rate": 1.8578897723596124e-05, "loss": 0.818, "step": 3777 }, { "epoch": 0.59, "grad_norm": 24.57292858200032, "learning_rate": 1.8578031072874655e-05, "loss": 0.6849, "step": 3778 }, { "epoch": 0.59, "grad_norm": 28.16611248179845, "learning_rate": 1.8577164178198995e-05, "loss": 0.8842, "step": 3779 }, { "epoch": 0.59, "grad_norm": 21.157711751917876, "learning_rate": 1.857629703959381e-05, "loss": 0.7568, "step": 3780 }, { "epoch": 0.59, "grad_norm": 19.410373774042984, "learning_rate": 1.8575429657083753e-05, "loss": 0.7815, "step": 3781 }, { "epoch": 0.59, "grad_norm": 22.342904453684604, "learning_rate": 1.8574562030693492e-05, "loss": 0.7891, "step": 3782 }, { "epoch": 0.59, "grad_norm": 20.830380088738504, "learning_rate": 1.85736941604477e-05, "loss": 0.7813, "step": 3783 }, { "epoch": 0.59, "grad_norm": 28.31186738581136, "learning_rate": 1.8572826046371063e-05, "loss": 0.8201, "step": 3784 }, { "epoch": 0.59, "grad_norm": 14.20612771751097, "learning_rate": 1.8571957688488268e-05, "loss": 0.7892, "step": 3785 }, { "epoch": 0.59, "grad_norm": 29.547951486546346, "learning_rate": 1.8571089086824014e-05, "loss": 0.9203, "step": 3786 }, { "epoch": 0.59, "grad_norm": 18.98118268868699, "learning_rate": 1.8570220241402996e-05, "loss": 0.8133, "step": 3787 }, { "epoch": 0.59, "grad_norm": 20.630320797576523, "learning_rate": 1.8569351152249932e-05, "loss": 0.7927, "step": 3788 }, { "epoch": 0.59, "grad_norm": 19.285110558572953, "learning_rate": 1.856848181938953e-05, "loss": 0.8186, "step": 3789 }, { "epoch": 0.59, "grad_norm": 14.457573311308336, "learning_rate": 1.856761224284652e-05, "loss": 0.783, "step": 3790 }, { "epoch": 0.59, "grad_norm": 20.97183920820715, "learning_rate": 1.856674242264563e-05, "loss": 0.7642, "step": 3791 }, { "epoch": 0.59, "grad_norm": 13.592564884134067, "learning_rate": 1.8565872358811594e-05, "loss": 0.8134, "step": 3792 }, { "epoch": 0.59, "grad_norm": 14.40412498552598, "learning_rate": 1.856500205136916e-05, "loss": 0.7384, "step": 3793 }, { "epoch": 0.59, "grad_norm": 15.685917406879609, "learning_rate": 1.8564131500343082e-05, "loss": 0.8745, "step": 3794 }, { "epoch": 0.59, "grad_norm": 17.667348752559906, "learning_rate": 1.8563260705758114e-05, "loss": 0.7676, "step": 3795 }, { "epoch": 0.59, "grad_norm": 26.836344030929713, "learning_rate": 1.8562389667639016e-05, "loss": 0.803, "step": 3796 }, { "epoch": 0.59, "grad_norm": 15.011330895596231, "learning_rate": 1.8561518386010567e-05, "loss": 0.7189, "step": 3797 }, { "epoch": 0.59, "grad_norm": 20.048253176286806, "learning_rate": 1.856064686089754e-05, "loss": 0.7731, "step": 3798 }, { "epoch": 0.59, "grad_norm": 12.730891031388188, "learning_rate": 1.855977509232473e-05, "loss": 0.7178, "step": 3799 }, { "epoch": 0.59, "grad_norm": 18.19745705653203, "learning_rate": 1.8558903080316922e-05, "loss": 0.9754, "step": 3800 }, { "epoch": 0.59, "grad_norm": 15.072025045541388, "learning_rate": 1.8558030824898915e-05, "loss": 0.8115, "step": 3801 }, { "epoch": 0.59, "grad_norm": 17.687996764460728, "learning_rate": 1.8557158326095515e-05, "loss": 0.7821, "step": 3802 }, { "epoch": 0.59, "grad_norm": 27.291128992773302, "learning_rate": 1.8556285583931547e-05, "loss": 0.8319, "step": 3803 }, { "epoch": 0.59, "grad_norm": 31.95447375148788, "learning_rate": 1.8555412598431814e-05, "loss": 0.6964, "step": 3804 }, { "epoch": 0.59, "grad_norm": 18.714737086269803, "learning_rate": 1.8554539369621156e-05, "loss": 0.9218, "step": 3805 }, { "epoch": 0.59, "grad_norm": 33.00982659509593, "learning_rate": 1.85536658975244e-05, "loss": 0.8843, "step": 3806 }, { "epoch": 0.59, "grad_norm": 16.160791135197275, "learning_rate": 1.8552792182166393e-05, "loss": 0.7222, "step": 3807 }, { "epoch": 0.59, "grad_norm": 18.08386377841724, "learning_rate": 1.8551918223571975e-05, "loss": 0.7301, "step": 3808 }, { "epoch": 0.59, "grad_norm": 17.394686121050924, "learning_rate": 1.855104402176601e-05, "loss": 0.7419, "step": 3809 }, { "epoch": 0.6, "grad_norm": 23.767436149750548, "learning_rate": 1.8550169576773353e-05, "loss": 0.7551, "step": 3810 }, { "epoch": 0.6, "grad_norm": 15.201831240165006, "learning_rate": 1.8549294888618872e-05, "loss": 0.7064, "step": 3811 }, { "epoch": 0.6, "grad_norm": 21.679244004258322, "learning_rate": 1.8548419957327454e-05, "loss": 0.7568, "step": 3812 }, { "epoch": 0.6, "grad_norm": 43.70672313633746, "learning_rate": 1.8547544782923963e-05, "loss": 0.809, "step": 3813 }, { "epoch": 0.6, "grad_norm": 31.798978130979503, "learning_rate": 1.8546669365433305e-05, "loss": 0.7657, "step": 3814 }, { "epoch": 0.6, "grad_norm": 17.444162700736577, "learning_rate": 1.8545793704880367e-05, "loss": 0.8721, "step": 3815 }, { "epoch": 0.6, "grad_norm": 26.541801941792407, "learning_rate": 1.854491780129006e-05, "loss": 0.6517, "step": 3816 }, { "epoch": 0.6, "grad_norm": 23.685965498701293, "learning_rate": 1.8544041654687282e-05, "loss": 0.7482, "step": 3817 }, { "epoch": 0.6, "grad_norm": 23.180410440072112, "learning_rate": 1.8543165265096962e-05, "loss": 0.793, "step": 3818 }, { "epoch": 0.6, "grad_norm": 17.74174721447432, "learning_rate": 1.854228863254402e-05, "loss": 0.7929, "step": 3819 }, { "epoch": 0.6, "grad_norm": 16.68191396786097, "learning_rate": 1.854141175705338e-05, "loss": 0.755, "step": 3820 }, { "epoch": 0.6, "grad_norm": 18.026995382528284, "learning_rate": 1.8540534638649994e-05, "loss": 0.8245, "step": 3821 }, { "epoch": 0.6, "grad_norm": 18.431292069860227, "learning_rate": 1.8539657277358798e-05, "loss": 0.7911, "step": 3822 }, { "epoch": 0.6, "grad_norm": 19.52814984093385, "learning_rate": 1.8538779673204743e-05, "loss": 0.8422, "step": 3823 }, { "epoch": 0.6, "grad_norm": 13.641874603503572, "learning_rate": 1.8537901826212786e-05, "loss": 0.7668, "step": 3824 }, { "epoch": 0.6, "grad_norm": 15.763300835906904, "learning_rate": 1.8537023736407898e-05, "loss": 0.7732, "step": 3825 }, { "epoch": 0.6, "grad_norm": 17.46968069224584, "learning_rate": 1.853614540381505e-05, "loss": 0.8077, "step": 3826 }, { "epoch": 0.6, "grad_norm": 23.594141751630684, "learning_rate": 1.853526682845922e-05, "loss": 0.8224, "step": 3827 }, { "epoch": 0.6, "grad_norm": 20.16988856553982, "learning_rate": 1.8534388010365397e-05, "loss": 0.839, "step": 3828 }, { "epoch": 0.6, "grad_norm": 64.85110000732195, "learning_rate": 1.8533508949558564e-05, "loss": 0.84, "step": 3829 }, { "epoch": 0.6, "grad_norm": 20.32852115215774, "learning_rate": 1.8532629646063737e-05, "loss": 0.8396, "step": 3830 }, { "epoch": 0.6, "grad_norm": 15.524113179146838, "learning_rate": 1.853175009990591e-05, "loss": 0.7346, "step": 3831 }, { "epoch": 0.6, "grad_norm": 23.23100621134276, "learning_rate": 1.85308703111101e-05, "loss": 0.7434, "step": 3832 }, { "epoch": 0.6, "grad_norm": 18.980683415202517, "learning_rate": 1.852999027970133e-05, "loss": 0.7483, "step": 3833 }, { "epoch": 0.6, "grad_norm": 30.459384500970334, "learning_rate": 1.8529110005704627e-05, "loss": 0.8256, "step": 3834 }, { "epoch": 0.6, "grad_norm": 16.785462742635673, "learning_rate": 1.8528229489145023e-05, "loss": 0.7698, "step": 3835 }, { "epoch": 0.6, "grad_norm": 14.966389276290089, "learning_rate": 1.8527348730047563e-05, "loss": 0.7426, "step": 3836 }, { "epoch": 0.6, "grad_norm": 25.438100229558053, "learning_rate": 1.8526467728437292e-05, "loss": 0.7875, "step": 3837 }, { "epoch": 0.6, "grad_norm": 30.353504072658048, "learning_rate": 1.852558648433927e-05, "loss": 0.762, "step": 3838 }, { "epoch": 0.6, "grad_norm": 24.14142603363089, "learning_rate": 1.852470499777855e-05, "loss": 0.919, "step": 3839 }, { "epoch": 0.6, "grad_norm": 21.39316143660563, "learning_rate": 1.8523823268780214e-05, "loss": 0.7851, "step": 3840 }, { "epoch": 0.6, "grad_norm": 14.969173797708438, "learning_rate": 1.8522941297369324e-05, "loss": 0.7629, "step": 3841 }, { "epoch": 0.6, "grad_norm": 14.20639261315958, "learning_rate": 1.8522059083570972e-05, "loss": 0.7311, "step": 3842 }, { "epoch": 0.6, "grad_norm": 29.990470734251467, "learning_rate": 1.8521176627410246e-05, "loss": 0.7724, "step": 3843 }, { "epoch": 0.6, "grad_norm": 19.021110580348523, "learning_rate": 1.852029392891224e-05, "loss": 0.8001, "step": 3844 }, { "epoch": 0.6, "grad_norm": 15.79630895066451, "learning_rate": 1.8519410988102058e-05, "loss": 0.7395, "step": 3845 }, { "epoch": 0.6, "grad_norm": 23.08821192734763, "learning_rate": 1.8518527805004812e-05, "loss": 0.7995, "step": 3846 }, { "epoch": 0.6, "grad_norm": 15.586739605506358, "learning_rate": 1.851764437964562e-05, "loss": 0.7433, "step": 3847 }, { "epoch": 0.6, "grad_norm": 16.41879667817107, "learning_rate": 1.8516760712049603e-05, "loss": 0.7605, "step": 3848 }, { "epoch": 0.6, "grad_norm": 12.800195746349994, "learning_rate": 1.8515876802241894e-05, "loss": 0.6754, "step": 3849 }, { "epoch": 0.6, "grad_norm": 20.735721712568434, "learning_rate": 1.8514992650247627e-05, "loss": 0.8175, "step": 3850 }, { "epoch": 0.6, "grad_norm": 25.940824373759135, "learning_rate": 1.8514108256091953e-05, "loss": 0.8489, "step": 3851 }, { "epoch": 0.6, "grad_norm": 14.821477867824576, "learning_rate": 1.8513223619800022e-05, "loss": 0.7689, "step": 3852 }, { "epoch": 0.6, "grad_norm": 17.62874136775155, "learning_rate": 1.851233874139699e-05, "loss": 0.8623, "step": 3853 }, { "epoch": 0.6, "grad_norm": 16.45067523042239, "learning_rate": 1.8511453620908027e-05, "loss": 0.7721, "step": 3854 }, { "epoch": 0.6, "grad_norm": 18.27249941593221, "learning_rate": 1.8510568258358298e-05, "loss": 0.7848, "step": 3855 }, { "epoch": 0.6, "grad_norm": 19.267452031925085, "learning_rate": 1.850968265377299e-05, "loss": 0.7377, "step": 3856 }, { "epoch": 0.6, "grad_norm": 16.04300267727044, "learning_rate": 1.850879680717728e-05, "loss": 0.8161, "step": 3857 }, { "epoch": 0.6, "grad_norm": 14.973969469370363, "learning_rate": 1.850791071859637e-05, "loss": 0.7655, "step": 3858 }, { "epoch": 0.6, "grad_norm": 13.764558773012604, "learning_rate": 1.8507024388055456e-05, "loss": 0.8755, "step": 3859 }, { "epoch": 0.6, "grad_norm": 20.076852204468675, "learning_rate": 1.8506137815579744e-05, "loss": 0.7986, "step": 3860 }, { "epoch": 0.6, "grad_norm": 23.58434569656616, "learning_rate": 1.8505251001194445e-05, "loss": 0.8154, "step": 3861 }, { "epoch": 0.6, "grad_norm": 11.40001003903859, "learning_rate": 1.850436394492479e-05, "loss": 0.7052, "step": 3862 }, { "epoch": 0.6, "grad_norm": 15.656556720356061, "learning_rate": 1.8503476646795996e-05, "loss": 0.6883, "step": 3863 }, { "epoch": 0.6, "grad_norm": 17.207146983455665, "learning_rate": 1.85025891068333e-05, "loss": 0.8264, "step": 3864 }, { "epoch": 0.6, "grad_norm": 13.353768880110001, "learning_rate": 1.8501701325061943e-05, "loss": 0.7402, "step": 3865 }, { "epoch": 0.6, "grad_norm": 15.248055625298326, "learning_rate": 1.8500813301507174e-05, "loss": 0.7408, "step": 3866 }, { "epoch": 0.6, "grad_norm": 22.465835975885266, "learning_rate": 1.8499925036194247e-05, "loss": 0.8808, "step": 3867 }, { "epoch": 0.6, "grad_norm": 15.006840873991008, "learning_rate": 1.849903652914843e-05, "loss": 0.785, "step": 3868 }, { "epoch": 0.6, "grad_norm": 16.900283448110795, "learning_rate": 1.849814778039498e-05, "loss": 0.8314, "step": 3869 }, { "epoch": 0.6, "grad_norm": 21.997378637458333, "learning_rate": 1.849725878995918e-05, "loss": 0.8069, "step": 3870 }, { "epoch": 0.6, "grad_norm": 18.670353523418115, "learning_rate": 1.849636955786631e-05, "loss": 0.7975, "step": 3871 }, { "epoch": 0.6, "grad_norm": 12.908022661833616, "learning_rate": 1.8495480084141662e-05, "loss": 0.8225, "step": 3872 }, { "epoch": 0.6, "grad_norm": 26.254595375864266, "learning_rate": 1.849459036881053e-05, "loss": 0.8323, "step": 3873 }, { "epoch": 0.61, "grad_norm": 29.21850568801818, "learning_rate": 1.8493700411898213e-05, "loss": 0.8218, "step": 3874 }, { "epoch": 0.61, "grad_norm": 18.41556248323299, "learning_rate": 1.8492810213430027e-05, "loss": 0.6852, "step": 3875 }, { "epoch": 0.61, "grad_norm": 22.42568565445289, "learning_rate": 1.849191977343129e-05, "loss": 0.8137, "step": 3876 }, { "epoch": 0.61, "grad_norm": 20.13855530497266, "learning_rate": 1.849102909192732e-05, "loss": 0.8085, "step": 3877 }, { "epoch": 0.61, "grad_norm": 24.071614659543368, "learning_rate": 1.849013816894345e-05, "loss": 0.8206, "step": 3878 }, { "epoch": 0.61, "grad_norm": 16.880815030354963, "learning_rate": 1.8489247004505017e-05, "loss": 0.8588, "step": 3879 }, { "epoch": 0.61, "grad_norm": 20.932620809459976, "learning_rate": 1.8488355598637362e-05, "loss": 0.8097, "step": 3880 }, { "epoch": 0.61, "grad_norm": 19.821513297094455, "learning_rate": 1.8487463951365842e-05, "loss": 0.6766, "step": 3881 }, { "epoch": 0.61, "grad_norm": 25.381460687751073, "learning_rate": 1.8486572062715814e-05, "loss": 0.7676, "step": 3882 }, { "epoch": 0.61, "grad_norm": 12.694321796201884, "learning_rate": 1.848567993271264e-05, "loss": 0.851, "step": 3883 }, { "epoch": 0.61, "grad_norm": 23.91939712986684, "learning_rate": 1.848478756138169e-05, "loss": 0.7627, "step": 3884 }, { "epoch": 0.61, "grad_norm": 27.575489992472782, "learning_rate": 1.8483894948748348e-05, "loss": 0.9076, "step": 3885 }, { "epoch": 0.61, "grad_norm": 20.200503250736677, "learning_rate": 1.8483002094837998e-05, "loss": 0.7739, "step": 3886 }, { "epoch": 0.61, "grad_norm": 24.171025317348672, "learning_rate": 1.848210899967603e-05, "loss": 0.7689, "step": 3887 }, { "epoch": 0.61, "grad_norm": 18.639976547742176, "learning_rate": 1.848121566328784e-05, "loss": 0.7419, "step": 3888 }, { "epoch": 0.61, "grad_norm": 25.321154784917912, "learning_rate": 1.848032208569884e-05, "loss": 0.8569, "step": 3889 }, { "epoch": 0.61, "grad_norm": 20.313079087732586, "learning_rate": 1.8479428266934442e-05, "loss": 0.8096, "step": 3890 }, { "epoch": 0.61, "grad_norm": 11.55765342883998, "learning_rate": 1.8478534207020066e-05, "loss": 0.7158, "step": 3891 }, { "epoch": 0.61, "grad_norm": 24.29365973096387, "learning_rate": 1.8477639905981135e-05, "loss": 0.8127, "step": 3892 }, { "epoch": 0.61, "grad_norm": 27.826071288049587, "learning_rate": 1.8476745363843082e-05, "loss": 0.7439, "step": 3893 }, { "epoch": 0.61, "grad_norm": 18.669471236448274, "learning_rate": 1.847585058063135e-05, "loss": 0.8184, "step": 3894 }, { "epoch": 0.61, "grad_norm": 27.851623443601042, "learning_rate": 1.847495555637139e-05, "loss": 0.7586, "step": 3895 }, { "epoch": 0.61, "grad_norm": 18.347098516052323, "learning_rate": 1.8474060291088646e-05, "loss": 0.7504, "step": 3896 }, { "epoch": 0.61, "grad_norm": 20.727692649101545, "learning_rate": 1.8473164784808592e-05, "loss": 0.8458, "step": 3897 }, { "epoch": 0.61, "grad_norm": 20.36514862589803, "learning_rate": 1.8472269037556683e-05, "loss": 0.7923, "step": 3898 }, { "epoch": 0.61, "grad_norm": 30.892618550586896, "learning_rate": 1.84713730493584e-05, "loss": 0.7549, "step": 3899 }, { "epoch": 0.61, "grad_norm": 16.007721708458217, "learning_rate": 1.847047682023922e-05, "loss": 0.7426, "step": 3900 }, { "epoch": 0.61, "grad_norm": 20.439207536485934, "learning_rate": 1.8469580350224636e-05, "loss": 0.7434, "step": 3901 }, { "epoch": 0.61, "grad_norm": 30.340700094476045, "learning_rate": 1.8468683639340145e-05, "loss": 0.8639, "step": 3902 }, { "epoch": 0.61, "grad_norm": 16.80755338434516, "learning_rate": 1.846778668761124e-05, "loss": 0.9562, "step": 3903 }, { "epoch": 0.61, "grad_norm": 5.444557099233771, "learning_rate": 1.846688949506344e-05, "loss": 0.773, "step": 3904 }, { "epoch": 0.61, "grad_norm": 13.873548635833972, "learning_rate": 1.8465992061722255e-05, "loss": 0.7355, "step": 3905 }, { "epoch": 0.61, "grad_norm": 24.731886725897116, "learning_rate": 1.8465094387613204e-05, "loss": 0.8077, "step": 3906 }, { "epoch": 0.61, "grad_norm": 19.668701540451185, "learning_rate": 1.8464196472761825e-05, "loss": 0.9629, "step": 3907 }, { "epoch": 0.61, "grad_norm": 13.96158451594049, "learning_rate": 1.8463298317193647e-05, "loss": 0.8042, "step": 3908 }, { "epoch": 0.61, "grad_norm": 22.66219790157908, "learning_rate": 1.846239992093422e-05, "loss": 0.7923, "step": 3909 }, { "epoch": 0.61, "grad_norm": 16.93281560075687, "learning_rate": 1.8461501284009085e-05, "loss": 0.7495, "step": 3910 }, { "epoch": 0.61, "grad_norm": 17.829182751826917, "learning_rate": 1.8460602406443802e-05, "loss": 0.7514, "step": 3911 }, { "epoch": 0.61, "grad_norm": 18.564662925341402, "learning_rate": 1.845970328826394e-05, "loss": 0.81, "step": 3912 }, { "epoch": 0.61, "grad_norm": 15.968870285685993, "learning_rate": 1.8458803929495062e-05, "loss": 0.8361, "step": 3913 }, { "epoch": 0.61, "grad_norm": 27.28492253503119, "learning_rate": 1.8457904330162752e-05, "loss": 0.8537, "step": 3914 }, { "epoch": 0.61, "grad_norm": 18.33909896703304, "learning_rate": 1.8457004490292588e-05, "loss": 0.7779, "step": 3915 }, { "epoch": 0.61, "grad_norm": 24.2744878738717, "learning_rate": 1.8456104409910165e-05, "loss": 0.8338, "step": 3916 }, { "epoch": 0.61, "grad_norm": 30.352666432201893, "learning_rate": 1.845520408904108e-05, "loss": 0.9, "step": 3917 }, { "epoch": 0.61, "grad_norm": 26.65282672610522, "learning_rate": 1.8454303527710934e-05, "loss": 0.7512, "step": 3918 }, { "epoch": 0.61, "grad_norm": 15.784527142814259, "learning_rate": 1.8453402725945343e-05, "loss": 0.7653, "step": 3919 }, { "epoch": 0.61, "grad_norm": 13.254694785066972, "learning_rate": 1.8452501683769924e-05, "loss": 0.7724, "step": 3920 }, { "epoch": 0.61, "grad_norm": 15.85627894130136, "learning_rate": 1.8451600401210303e-05, "loss": 0.7332, "step": 3921 }, { "epoch": 0.61, "grad_norm": 26.047388374930797, "learning_rate": 1.8450698878292106e-05, "loss": 1.0138, "step": 3922 }, { "epoch": 0.61, "grad_norm": 17.418785834027496, "learning_rate": 1.844979711504098e-05, "loss": 0.8606, "step": 3923 }, { "epoch": 0.61, "grad_norm": 21.816517696932664, "learning_rate": 1.8448895111482565e-05, "loss": 0.8204, "step": 3924 }, { "epoch": 0.61, "grad_norm": 17.214806980589778, "learning_rate": 1.8447992867642518e-05, "loss": 0.8153, "step": 3925 }, { "epoch": 0.61, "grad_norm": 20.205753759461043, "learning_rate": 1.8447090383546494e-05, "loss": 0.689, "step": 3926 }, { "epoch": 0.61, "grad_norm": 13.492803046007635, "learning_rate": 1.8446187659220163e-05, "loss": 0.698, "step": 3927 }, { "epoch": 0.61, "grad_norm": 13.529158688303893, "learning_rate": 1.8445284694689193e-05, "loss": 0.8182, "step": 3928 }, { "epoch": 0.61, "grad_norm": 15.343326230616558, "learning_rate": 1.8444381489979267e-05, "loss": 0.7947, "step": 3929 }, { "epoch": 0.61, "grad_norm": 16.393176317521146, "learning_rate": 1.8443478045116072e-05, "loss": 0.7321, "step": 3930 }, { "epoch": 0.61, "grad_norm": 22.799329624887452, "learning_rate": 1.84425743601253e-05, "loss": 0.7459, "step": 3931 }, { "epoch": 0.61, "grad_norm": 23.987049691732953, "learning_rate": 1.8441670435032654e-05, "loss": 0.7579, "step": 3932 }, { "epoch": 0.61, "grad_norm": 20.089735649065783, "learning_rate": 1.8440766269863838e-05, "loss": 0.7613, "step": 3933 }, { "epoch": 0.61, "grad_norm": 24.310592131543064, "learning_rate": 1.8439861864644567e-05, "loss": 0.8116, "step": 3934 }, { "epoch": 0.61, "grad_norm": 29.16116392434744, "learning_rate": 1.843895721940056e-05, "loss": 0.8648, "step": 3935 }, { "epoch": 0.61, "grad_norm": 28.05762151270861, "learning_rate": 1.843805233415755e-05, "loss": 0.8286, "step": 3936 }, { "epoch": 0.61, "grad_norm": 13.51066513673727, "learning_rate": 1.8437147208941265e-05, "loss": 0.8016, "step": 3937 }, { "epoch": 0.62, "grad_norm": 28.732777431316652, "learning_rate": 1.843624184377745e-05, "loss": 0.7766, "step": 3938 }, { "epoch": 0.62, "grad_norm": 18.46432995267984, "learning_rate": 1.8435336238691853e-05, "loss": 0.7877, "step": 3939 }, { "epoch": 0.62, "grad_norm": 26.58450417602139, "learning_rate": 1.8434430393710224e-05, "loss": 0.8457, "step": 3940 }, { "epoch": 0.62, "grad_norm": 22.97657099194777, "learning_rate": 1.8433524308858334e-05, "loss": 0.8222, "step": 3941 }, { "epoch": 0.62, "grad_norm": 22.80400193802145, "learning_rate": 1.843261798416194e-05, "loss": 0.7666, "step": 3942 }, { "epoch": 0.62, "grad_norm": 14.739879411047784, "learning_rate": 1.8431711419646826e-05, "loss": 0.7505, "step": 3943 }, { "epoch": 0.62, "grad_norm": 13.493484900760645, "learning_rate": 1.8430804615338774e-05, "loss": 0.704, "step": 3944 }, { "epoch": 0.62, "grad_norm": 18.62165709232176, "learning_rate": 1.842989757126357e-05, "loss": 0.8114, "step": 3945 }, { "epoch": 0.62, "grad_norm": 19.9087229641836, "learning_rate": 1.8428990287447006e-05, "loss": 0.7421, "step": 3946 }, { "epoch": 0.62, "grad_norm": 20.44888273325592, "learning_rate": 1.8428082763914893e-05, "loss": 0.7429, "step": 3947 }, { "epoch": 0.62, "grad_norm": 27.407702641282988, "learning_rate": 1.842717500069304e-05, "loss": 0.7829, "step": 3948 }, { "epoch": 0.62, "grad_norm": 10.696606699158387, "learning_rate": 1.8426266997807252e-05, "loss": 0.766, "step": 3949 }, { "epoch": 0.62, "grad_norm": 19.182272009347923, "learning_rate": 1.8425358755283365e-05, "loss": 0.7823, "step": 3950 }, { "epoch": 0.62, "grad_norm": 19.51357873522703, "learning_rate": 1.84244502731472e-05, "loss": 0.8084, "step": 3951 }, { "epoch": 0.62, "grad_norm": 21.2355762588216, "learning_rate": 1.84235415514246e-05, "loss": 0.7781, "step": 3952 }, { "epoch": 0.62, "grad_norm": 18.900462960225724, "learning_rate": 1.8422632590141407e-05, "loss": 0.7362, "step": 3953 }, { "epoch": 0.62, "grad_norm": 15.374814147232982, "learning_rate": 1.8421723389323468e-05, "loss": 0.6842, "step": 3954 }, { "epoch": 0.62, "grad_norm": 16.95554232832515, "learning_rate": 1.8420813948996643e-05, "loss": 0.7623, "step": 3955 }, { "epoch": 0.62, "grad_norm": 26.167381137093408, "learning_rate": 1.8419904269186797e-05, "loss": 0.8981, "step": 3956 }, { "epoch": 0.62, "grad_norm": 23.079421702367732, "learning_rate": 1.8418994349919798e-05, "loss": 0.9015, "step": 3957 }, { "epoch": 0.62, "grad_norm": 21.286503494870637, "learning_rate": 1.8418084191221522e-05, "loss": 0.9776, "step": 3958 }, { "epoch": 0.62, "grad_norm": 21.228907204678052, "learning_rate": 1.8417173793117857e-05, "loss": 0.7629, "step": 3959 }, { "epoch": 0.62, "grad_norm": 16.949775742036135, "learning_rate": 1.8416263155634695e-05, "loss": 0.7898, "step": 3960 }, { "epoch": 0.62, "grad_norm": 22.86129622296271, "learning_rate": 1.8415352278797932e-05, "loss": 0.8817, "step": 3961 }, { "epoch": 0.62, "grad_norm": 15.527071654136954, "learning_rate": 1.8414441162633473e-05, "loss": 0.7815, "step": 3962 }, { "epoch": 0.62, "grad_norm": 11.621064669147938, "learning_rate": 1.8413529807167228e-05, "loss": 0.7699, "step": 3963 }, { "epoch": 0.62, "grad_norm": 13.06265569895257, "learning_rate": 1.8412618212425116e-05, "loss": 0.7837, "step": 3964 }, { "epoch": 0.62, "grad_norm": 24.55817505382642, "learning_rate": 1.841170637843307e-05, "loss": 0.7721, "step": 3965 }, { "epoch": 0.62, "grad_norm": 19.239676548274712, "learning_rate": 1.8410794305217006e-05, "loss": 0.7911, "step": 3966 }, { "epoch": 0.62, "grad_norm": 22.03393289111414, "learning_rate": 1.8409881992802882e-05, "loss": 0.8064, "step": 3967 }, { "epoch": 0.62, "grad_norm": 23.318176207464848, "learning_rate": 1.8408969441216628e-05, "loss": 0.8338, "step": 3968 }, { "epoch": 0.62, "grad_norm": 32.12564241846807, "learning_rate": 1.8408056650484204e-05, "loss": 0.7723, "step": 3969 }, { "epoch": 0.62, "grad_norm": 16.174953355677708, "learning_rate": 1.8407143620631564e-05, "loss": 0.7423, "step": 3970 }, { "epoch": 0.62, "grad_norm": 18.970926900663347, "learning_rate": 1.840623035168468e-05, "loss": 0.7913, "step": 3971 }, { "epoch": 0.62, "grad_norm": 14.744496493201806, "learning_rate": 1.8405316843669524e-05, "loss": 0.6713, "step": 3972 }, { "epoch": 0.62, "grad_norm": 27.633855168561823, "learning_rate": 1.840440309661207e-05, "loss": 0.7697, "step": 3973 }, { "epoch": 0.62, "grad_norm": 39.12806122126247, "learning_rate": 1.8403489110538312e-05, "loss": 0.7226, "step": 3974 }, { "epoch": 0.62, "grad_norm": 29.946253078307834, "learning_rate": 1.8402574885474238e-05, "loss": 0.8742, "step": 3975 }, { "epoch": 0.62, "grad_norm": 25.150682981912503, "learning_rate": 1.840166042144585e-05, "loss": 0.7331, "step": 3976 }, { "epoch": 0.62, "grad_norm": 15.814120574315359, "learning_rate": 1.8400745718479155e-05, "loss": 0.8319, "step": 3977 }, { "epoch": 0.62, "grad_norm": 14.955471699729365, "learning_rate": 1.8399830776600164e-05, "loss": 0.7965, "step": 3978 }, { "epoch": 0.62, "grad_norm": 21.239781488610827, "learning_rate": 1.83989155958349e-05, "loss": 0.7555, "step": 3979 }, { "epoch": 0.62, "grad_norm": 23.13886507426497, "learning_rate": 1.839800017620939e-05, "loss": 0.8598, "step": 3980 }, { "epoch": 0.62, "grad_norm": 17.824127672720582, "learning_rate": 1.8397084517749668e-05, "loss": 0.8674, "step": 3981 }, { "epoch": 0.62, "grad_norm": 24.81647318274528, "learning_rate": 1.8396168620481772e-05, "loss": 0.8192, "step": 3982 }, { "epoch": 0.62, "grad_norm": 17.2916935686875, "learning_rate": 1.8395252484431758e-05, "loss": 0.8028, "step": 3983 }, { "epoch": 0.62, "grad_norm": 21.95360975443012, "learning_rate": 1.839433610962567e-05, "loss": 0.8024, "step": 3984 }, { "epoch": 0.62, "grad_norm": 13.174789954633832, "learning_rate": 1.8393419496089577e-05, "loss": 0.7694, "step": 3985 }, { "epoch": 0.62, "grad_norm": 23.955312698652683, "learning_rate": 1.839250264384954e-05, "loss": 0.8709, "step": 3986 }, { "epoch": 0.62, "grad_norm": 25.585178347674756, "learning_rate": 1.839158555293164e-05, "loss": 0.7702, "step": 3987 }, { "epoch": 0.62, "grad_norm": 21.922154810201178, "learning_rate": 1.839066822336195e-05, "loss": 0.7046, "step": 3988 }, { "epoch": 0.62, "grad_norm": 20.116644496070613, "learning_rate": 1.8389750655166572e-05, "loss": 0.7483, "step": 3989 }, { "epoch": 0.62, "grad_norm": 31.005086219098988, "learning_rate": 1.8388832848371588e-05, "loss": 0.8069, "step": 3990 }, { "epoch": 0.62, "grad_norm": 21.008058943035326, "learning_rate": 1.8387914803003106e-05, "loss": 0.7936, "step": 3991 }, { "epoch": 0.62, "grad_norm": 21.647187100838746, "learning_rate": 1.8386996519087236e-05, "loss": 0.7547, "step": 3992 }, { "epoch": 0.62, "grad_norm": 19.94987403505995, "learning_rate": 1.8386077996650086e-05, "loss": 0.7883, "step": 3993 }, { "epoch": 0.62, "grad_norm": 13.975370734778773, "learning_rate": 1.838515923571779e-05, "loss": 0.8114, "step": 3994 }, { "epoch": 0.62, "grad_norm": 20.02200142409825, "learning_rate": 1.838424023631647e-05, "loss": 0.8497, "step": 3995 }, { "epoch": 0.62, "grad_norm": 13.750925009943543, "learning_rate": 1.838332099847226e-05, "loss": 0.6907, "step": 3996 }, { "epoch": 0.62, "grad_norm": 27.149989611490202, "learning_rate": 1.8382401522211304e-05, "loss": 0.8797, "step": 3997 }, { "epoch": 0.62, "grad_norm": 22.249574071512278, "learning_rate": 1.838148180755975e-05, "loss": 0.71, "step": 3998 }, { "epoch": 0.62, "grad_norm": 25.632264021819047, "learning_rate": 1.8380561854543763e-05, "loss": 0.838, "step": 3999 }, { "epoch": 0.62, "grad_norm": 15.405313161566008, "learning_rate": 1.8379641663189493e-05, "loss": 0.8205, "step": 4000 }, { "epoch": 0.62, "grad_norm": 21.319739297686727, "learning_rate": 1.837872123352312e-05, "loss": 0.7407, "step": 4001 }, { "epoch": 0.63, "grad_norm": 14.747272550442679, "learning_rate": 1.837780056557082e-05, "loss": 0.7633, "step": 4002 }, { "epoch": 0.63, "grad_norm": 21.668417350866296, "learning_rate": 1.8376879659358767e-05, "loss": 0.7911, "step": 4003 }, { "epoch": 0.63, "grad_norm": 16.531101403262607, "learning_rate": 1.8375958514913158e-05, "loss": 0.7659, "step": 4004 }, { "epoch": 0.63, "grad_norm": 15.707772077855429, "learning_rate": 1.8375037132260187e-05, "loss": 0.7484, "step": 4005 }, { "epoch": 0.63, "grad_norm": 22.826776165706086, "learning_rate": 1.837411551142606e-05, "loss": 0.8314, "step": 4006 }, { "epoch": 0.63, "grad_norm": 21.591839304191762, "learning_rate": 1.837319365243699e-05, "loss": 0.7947, "step": 4007 }, { "epoch": 0.63, "grad_norm": 21.507425071948727, "learning_rate": 1.8372271555319185e-05, "loss": 0.7932, "step": 4008 }, { "epoch": 0.63, "grad_norm": 27.13151214310836, "learning_rate": 1.837134922009888e-05, "loss": 0.802, "step": 4009 }, { "epoch": 0.63, "grad_norm": 19.195559047098417, "learning_rate": 1.8370426646802298e-05, "loss": 0.7496, "step": 4010 }, { "epoch": 0.63, "grad_norm": 26.029585745047147, "learning_rate": 1.8369503835455676e-05, "loss": 0.9137, "step": 4011 }, { "epoch": 0.63, "grad_norm": 18.172161446902468, "learning_rate": 1.8368580786085265e-05, "loss": 0.7937, "step": 4012 }, { "epoch": 0.63, "grad_norm": 17.169397638376193, "learning_rate": 1.836765749871731e-05, "loss": 0.8181, "step": 4013 }, { "epoch": 0.63, "grad_norm": 21.25379491607554, "learning_rate": 1.8366733973378074e-05, "loss": 0.8097, "step": 4014 }, { "epoch": 0.63, "grad_norm": 19.523381642616236, "learning_rate": 1.8365810210093814e-05, "loss": 0.7363, "step": 4015 }, { "epoch": 0.63, "grad_norm": 24.965855560785187, "learning_rate": 1.836488620889081e-05, "loss": 0.8354, "step": 4016 }, { "epoch": 0.63, "grad_norm": 19.242756028629827, "learning_rate": 1.836396196979533e-05, "loss": 0.8586, "step": 4017 }, { "epoch": 0.63, "grad_norm": 27.550900611148332, "learning_rate": 1.8363037492833667e-05, "loss": 0.9081, "step": 4018 }, { "epoch": 0.63, "grad_norm": 27.709086276462795, "learning_rate": 1.836211277803211e-05, "loss": 0.7596, "step": 4019 }, { "epoch": 0.63, "grad_norm": 12.05138982784889, "learning_rate": 1.836118782541696e-05, "loss": 0.7134, "step": 4020 }, { "epoch": 0.63, "grad_norm": 44.184207096855985, "learning_rate": 1.8360262635014513e-05, "loss": 0.8022, "step": 4021 }, { "epoch": 0.63, "grad_norm": 17.15848588444529, "learning_rate": 1.8359337206851094e-05, "loss": 0.7862, "step": 4022 }, { "epoch": 0.63, "grad_norm": 32.793212961511045, "learning_rate": 1.8358411540953008e-05, "loss": 0.8714, "step": 4023 }, { "epoch": 0.63, "grad_norm": 26.86531251475134, "learning_rate": 1.835748563734659e-05, "loss": 0.841, "step": 4024 }, { "epoch": 0.63, "grad_norm": 24.677800799251358, "learning_rate": 1.8356559496058172e-05, "loss": 0.7518, "step": 4025 }, { "epoch": 0.63, "grad_norm": 19.5701404761941, "learning_rate": 1.835563311711409e-05, "loss": 0.8105, "step": 4026 }, { "epoch": 0.63, "grad_norm": 13.806335132849288, "learning_rate": 1.835470650054069e-05, "loss": 0.7603, "step": 4027 }, { "epoch": 0.63, "grad_norm": 22.27433458205608, "learning_rate": 1.8353779646364323e-05, "loss": 0.822, "step": 4028 }, { "epoch": 0.63, "grad_norm": 21.796653723875387, "learning_rate": 1.835285255461135e-05, "loss": 0.9318, "step": 4029 }, { "epoch": 0.63, "grad_norm": 23.65776073187368, "learning_rate": 1.8351925225308136e-05, "loss": 0.8439, "step": 4030 }, { "epoch": 0.63, "grad_norm": 16.733131728483166, "learning_rate": 1.8350997658481054e-05, "loss": 0.7087, "step": 4031 }, { "epoch": 0.63, "grad_norm": 20.27642042085711, "learning_rate": 1.835006985415649e-05, "loss": 0.7264, "step": 4032 }, { "epoch": 0.63, "grad_norm": 31.554432419541385, "learning_rate": 1.8349141812360815e-05, "loss": 0.8506, "step": 4033 }, { "epoch": 0.63, "grad_norm": 20.956954534626647, "learning_rate": 1.8348213533120437e-05, "loss": 0.7942, "step": 4034 }, { "epoch": 0.63, "grad_norm": 31.273529697752117, "learning_rate": 1.8347285016461745e-05, "loss": 0.8884, "step": 4035 }, { "epoch": 0.63, "grad_norm": 23.249950737670268, "learning_rate": 1.8346356262411157e-05, "loss": 0.7631, "step": 4036 }, { "epoch": 0.63, "grad_norm": 24.02906163759453, "learning_rate": 1.8345427270995072e-05, "loss": 0.7787, "step": 4037 }, { "epoch": 0.63, "grad_norm": 21.05649738565577, "learning_rate": 1.8344498042239922e-05, "loss": 0.7353, "step": 4038 }, { "epoch": 0.63, "grad_norm": 14.38100596628698, "learning_rate": 1.8343568576172127e-05, "loss": 0.7436, "step": 4039 }, { "epoch": 0.63, "grad_norm": 21.546504893730834, "learning_rate": 1.834263887281812e-05, "loss": 0.82, "step": 4040 }, { "epoch": 0.63, "grad_norm": 16.93560101889014, "learning_rate": 1.834170893220435e-05, "loss": 0.6939, "step": 4041 }, { "epoch": 0.63, "grad_norm": 17.58899066922581, "learning_rate": 1.8340778754357252e-05, "loss": 0.7755, "step": 4042 }, { "epoch": 0.63, "grad_norm": 25.40402393981906, "learning_rate": 1.833984833930329e-05, "loss": 0.8215, "step": 4043 }, { "epoch": 0.63, "grad_norm": 18.259975279949824, "learning_rate": 1.8338917687068916e-05, "loss": 0.7213, "step": 4044 }, { "epoch": 0.63, "grad_norm": 19.97638652598689, "learning_rate": 1.8337986797680604e-05, "loss": 0.8617, "step": 4045 }, { "epoch": 0.63, "grad_norm": 16.826331171398678, "learning_rate": 1.8337055671164826e-05, "loss": 0.7763, "step": 4046 }, { "epoch": 0.63, "grad_norm": 22.21938759892234, "learning_rate": 1.833612430754806e-05, "loss": 0.7074, "step": 4047 }, { "epoch": 0.63, "grad_norm": 17.975197932986642, "learning_rate": 1.8335192706856792e-05, "loss": 0.7524, "step": 4048 }, { "epoch": 0.63, "grad_norm": 13.644701512975637, "learning_rate": 1.8334260869117524e-05, "loss": 0.8038, "step": 4049 }, { "epoch": 0.63, "grad_norm": 36.40626696966864, "learning_rate": 1.833332879435675e-05, "loss": 0.7759, "step": 4050 }, { "epoch": 0.63, "grad_norm": 29.939665329817146, "learning_rate": 1.833239648260098e-05, "loss": 0.7455, "step": 4051 }, { "epoch": 0.63, "grad_norm": 19.847727211801143, "learning_rate": 1.833146393387673e-05, "loss": 0.7344, "step": 4052 }, { "epoch": 0.63, "grad_norm": 17.72879469205248, "learning_rate": 1.8330531148210516e-05, "loss": 0.8269, "step": 4053 }, { "epoch": 0.63, "grad_norm": 19.034760398827235, "learning_rate": 1.8329598125628873e-05, "loss": 0.79, "step": 4054 }, { "epoch": 0.63, "grad_norm": 12.130774426903695, "learning_rate": 1.8328664866158332e-05, "loss": 0.7332, "step": 4055 }, { "epoch": 0.63, "grad_norm": 16.734666720286576, "learning_rate": 1.8327731369825432e-05, "loss": 0.788, "step": 4056 }, { "epoch": 0.63, "grad_norm": 18.344273597094165, "learning_rate": 1.8326797636656723e-05, "loss": 0.8264, "step": 4057 }, { "epoch": 0.63, "grad_norm": 20.630659119575697, "learning_rate": 1.8325863666678764e-05, "loss": 0.7591, "step": 4058 }, { "epoch": 0.63, "grad_norm": 36.55776862643885, "learning_rate": 1.832492945991811e-05, "loss": 0.7203, "step": 4059 }, { "epoch": 0.63, "grad_norm": 13.70146592654616, "learning_rate": 1.8323995016401332e-05, "loss": 0.7348, "step": 4060 }, { "epoch": 0.63, "grad_norm": 16.597404789158826, "learning_rate": 1.8323060336155008e-05, "loss": 0.7367, "step": 4061 }, { "epoch": 0.63, "grad_norm": 18.41308691752615, "learning_rate": 1.8322125419205713e-05, "loss": 0.7802, "step": 4062 }, { "epoch": 0.63, "grad_norm": 23.890983435397562, "learning_rate": 1.832119026558004e-05, "loss": 0.8569, "step": 4063 }, { "epoch": 0.63, "grad_norm": 12.789768444191667, "learning_rate": 1.8320254875304585e-05, "loss": 0.7193, "step": 4064 }, { "epoch": 0.63, "grad_norm": 17.398659155709403, "learning_rate": 1.831931924840595e-05, "loss": 0.6454, "step": 4065 }, { "epoch": 0.64, "grad_norm": 16.721580559436095, "learning_rate": 1.831838338491074e-05, "loss": 0.7313, "step": 4066 }, { "epoch": 0.64, "grad_norm": 19.189410233777284, "learning_rate": 1.8317447284845573e-05, "loss": 0.7123, "step": 4067 }, { "epoch": 0.64, "grad_norm": 25.44081514018268, "learning_rate": 1.831651094823707e-05, "loss": 0.7311, "step": 4068 }, { "epoch": 0.64, "grad_norm": 23.73241113633123, "learning_rate": 1.831557437511186e-05, "loss": 0.8133, "step": 4069 }, { "epoch": 0.64, "grad_norm": 16.433668434747723, "learning_rate": 1.8314637565496584e-05, "loss": 0.6534, "step": 4070 }, { "epoch": 0.64, "grad_norm": 30.62790181026401, "learning_rate": 1.8313700519417876e-05, "loss": 0.8467, "step": 4071 }, { "epoch": 0.64, "grad_norm": 19.14620854273327, "learning_rate": 1.8312763236902387e-05, "loss": 0.8438, "step": 4072 }, { "epoch": 0.64, "grad_norm": 17.66952688602762, "learning_rate": 1.831182571797678e-05, "loss": 0.7178, "step": 4073 }, { "epoch": 0.64, "grad_norm": 16.692903620687968, "learning_rate": 1.8310887962667706e-05, "loss": 0.7031, "step": 4074 }, { "epoch": 0.64, "grad_norm": 22.533743314309838, "learning_rate": 1.830994997100184e-05, "loss": 0.884, "step": 4075 }, { "epoch": 0.64, "grad_norm": 23.27041583951733, "learning_rate": 1.8309011743005863e-05, "loss": 0.7584, "step": 4076 }, { "epoch": 0.64, "grad_norm": 86.65536448497078, "learning_rate": 1.8308073278706448e-05, "loss": 0.7623, "step": 4077 }, { "epoch": 0.64, "grad_norm": 17.39951506978944, "learning_rate": 1.8307134578130293e-05, "loss": 0.7753, "step": 4078 }, { "epoch": 0.64, "grad_norm": 16.932615527768895, "learning_rate": 1.8306195641304088e-05, "loss": 0.8052, "step": 4079 }, { "epoch": 0.64, "grad_norm": 15.343993609897169, "learning_rate": 1.8305256468254537e-05, "loss": 0.8088, "step": 4080 }, { "epoch": 0.64, "grad_norm": 19.93329382866461, "learning_rate": 1.830431705900835e-05, "loss": 0.7737, "step": 4081 }, { "epoch": 0.64, "grad_norm": 18.29276556553434, "learning_rate": 1.8303377413592248e-05, "loss": 0.6871, "step": 4082 }, { "epoch": 0.64, "grad_norm": 17.800445241228395, "learning_rate": 1.8302437532032945e-05, "loss": 0.7648, "step": 4083 }, { "epoch": 0.64, "grad_norm": 21.153387648377237, "learning_rate": 1.830149741435718e-05, "loss": 0.8411, "step": 4084 }, { "epoch": 0.64, "grad_norm": 19.861353873533375, "learning_rate": 1.830055706059168e-05, "loss": 0.8067, "step": 4085 }, { "epoch": 0.64, "grad_norm": 18.48304642508233, "learning_rate": 1.8299616470763196e-05, "loss": 0.7107, "step": 4086 }, { "epoch": 0.64, "grad_norm": 20.179646688753063, "learning_rate": 1.829867564489847e-05, "loss": 0.9648, "step": 4087 }, { "epoch": 0.64, "grad_norm": 28.52022475833779, "learning_rate": 1.8297734583024273e-05, "loss": 0.808, "step": 4088 }, { "epoch": 0.64, "grad_norm": 16.980345340979714, "learning_rate": 1.8296793285167348e-05, "loss": 0.6603, "step": 4089 }, { "epoch": 0.64, "grad_norm": 24.50393409247012, "learning_rate": 1.8295851751354482e-05, "loss": 0.7467, "step": 4090 }, { "epoch": 0.64, "grad_norm": 22.699693493333474, "learning_rate": 1.8294909981612443e-05, "loss": 0.7744, "step": 4091 }, { "epoch": 0.64, "grad_norm": 24.123405910362294, "learning_rate": 1.8293967975968016e-05, "loss": 0.7457, "step": 4092 }, { "epoch": 0.64, "grad_norm": 15.135385663796493, "learning_rate": 1.8293025734447992e-05, "loss": 0.7948, "step": 4093 }, { "epoch": 0.64, "grad_norm": 20.12608952194189, "learning_rate": 1.8292083257079166e-05, "loss": 0.7302, "step": 4094 }, { "epoch": 0.64, "grad_norm": 19.64810996948637, "learning_rate": 1.8291140543888347e-05, "loss": 0.8145, "step": 4095 }, { "epoch": 0.64, "grad_norm": 23.120497737278335, "learning_rate": 1.8290197594902337e-05, "loss": 0.7315, "step": 4096 }, { "epoch": 0.64, "grad_norm": 14.403996646144535, "learning_rate": 1.8289254410147958e-05, "loss": 0.8329, "step": 4097 }, { "epoch": 0.64, "grad_norm": 14.843392320640191, "learning_rate": 1.8288310989652035e-05, "loss": 0.7139, "step": 4098 }, { "epoch": 0.64, "grad_norm": 17.479391158151, "learning_rate": 1.828736733344139e-05, "loss": 0.7458, "step": 4099 }, { "epoch": 0.64, "grad_norm": 11.982364150750014, "learning_rate": 1.8286423441542874e-05, "loss": 0.7826, "step": 4100 }, { "epoch": 0.64, "grad_norm": 14.865509935831454, "learning_rate": 1.828547931398332e-05, "loss": 0.7012, "step": 4101 }, { "epoch": 0.64, "grad_norm": 20.914275430842952, "learning_rate": 1.8284534950789578e-05, "loss": 0.7364, "step": 4102 }, { "epoch": 0.64, "grad_norm": 16.00571356771551, "learning_rate": 1.828359035198851e-05, "loss": 0.7736, "step": 4103 }, { "epoch": 0.64, "grad_norm": 29.542125641956837, "learning_rate": 1.8282645517606977e-05, "loss": 0.9103, "step": 4104 }, { "epoch": 0.64, "grad_norm": 15.097036599178933, "learning_rate": 1.8281700447671852e-05, "loss": 0.8486, "step": 4105 }, { "epoch": 0.64, "grad_norm": 16.684310183173224, "learning_rate": 1.828075514221001e-05, "loss": 0.6891, "step": 4106 }, { "epoch": 0.64, "grad_norm": 14.84088828107749, "learning_rate": 1.8279809601248342e-05, "loss": 0.831, "step": 4107 }, { "epoch": 0.64, "grad_norm": 23.74038264025481, "learning_rate": 1.8278863824813726e-05, "loss": 0.8437, "step": 4108 }, { "epoch": 0.64, "grad_norm": 16.528611614866424, "learning_rate": 1.8277917812933068e-05, "loss": 0.7473, "step": 4109 }, { "epoch": 0.64, "grad_norm": 28.163180080034113, "learning_rate": 1.827697156563327e-05, "loss": 0.8711, "step": 4110 }, { "epoch": 0.64, "grad_norm": 12.354580146850937, "learning_rate": 1.827602508294124e-05, "loss": 0.7358, "step": 4111 }, { "epoch": 0.64, "grad_norm": 15.790304353132047, "learning_rate": 1.82750783648839e-05, "loss": 0.7273, "step": 4112 }, { "epoch": 0.64, "grad_norm": 33.003127082226484, "learning_rate": 1.8274131411488172e-05, "loss": 0.8199, "step": 4113 }, { "epoch": 0.64, "grad_norm": 17.171991094486625, "learning_rate": 1.8273184222780984e-05, "loss": 0.794, "step": 4114 }, { "epoch": 0.64, "grad_norm": 14.696733095273618, "learning_rate": 1.827223679878928e-05, "loss": 0.7188, "step": 4115 }, { "epoch": 0.64, "grad_norm": 23.425840492548925, "learning_rate": 1.8271289139539997e-05, "loss": 0.8428, "step": 4116 }, { "epoch": 0.64, "grad_norm": 26.09084472929248, "learning_rate": 1.8270341245060095e-05, "loss": 0.822, "step": 4117 }, { "epoch": 0.64, "grad_norm": 21.6793608611648, "learning_rate": 1.8269393115376522e-05, "loss": 0.8358, "step": 4118 }, { "epoch": 0.64, "grad_norm": 36.155094100226705, "learning_rate": 1.8268444750516246e-05, "loss": 0.7757, "step": 4119 }, { "epoch": 0.64, "grad_norm": 18.56333594133736, "learning_rate": 1.8267496150506242e-05, "loss": 0.706, "step": 4120 }, { "epoch": 0.64, "grad_norm": 25.576181748058023, "learning_rate": 1.826654731537348e-05, "loss": 0.8794, "step": 4121 }, { "epoch": 0.64, "grad_norm": 27.442736306317713, "learning_rate": 1.826559824514495e-05, "loss": 0.8858, "step": 4122 }, { "epoch": 0.64, "grad_norm": 28.96593825882559, "learning_rate": 1.8264648939847642e-05, "loss": 0.9125, "step": 4123 }, { "epoch": 0.64, "grad_norm": 19.547110085785135, "learning_rate": 1.826369939950855e-05, "loss": 0.7612, "step": 4124 }, { "epoch": 0.64, "grad_norm": 18.927598558929994, "learning_rate": 1.8262749624154684e-05, "loss": 0.8409, "step": 4125 }, { "epoch": 0.64, "grad_norm": 33.57686183261153, "learning_rate": 1.8261799613813053e-05, "loss": 0.7878, "step": 4126 }, { "epoch": 0.64, "grad_norm": 14.26701776238318, "learning_rate": 1.826084936851067e-05, "loss": 0.6831, "step": 4127 }, { "epoch": 0.64, "grad_norm": 23.320788959766865, "learning_rate": 1.8259898888274564e-05, "loss": 0.7782, "step": 4128 }, { "epoch": 0.64, "grad_norm": 19.862091160834396, "learning_rate": 1.8258948173131766e-05, "loss": 0.7341, "step": 4129 }, { "epoch": 0.65, "grad_norm": 14.021932432436662, "learning_rate": 1.8257997223109317e-05, "loss": 0.7314, "step": 4130 }, { "epoch": 0.65, "grad_norm": 22.693422991939816, "learning_rate": 1.825704603823425e-05, "loss": 0.7429, "step": 4131 }, { "epoch": 0.65, "grad_norm": 19.957814172696747, "learning_rate": 1.8256094618533633e-05, "loss": 0.8141, "step": 4132 }, { "epoch": 0.65, "grad_norm": 13.163449491609454, "learning_rate": 1.825514296403451e-05, "loss": 0.7616, "step": 4133 }, { "epoch": 0.65, "grad_norm": 16.848775350195158, "learning_rate": 1.825419107476395e-05, "loss": 0.8326, "step": 4134 }, { "epoch": 0.65, "grad_norm": 14.569001354101406, "learning_rate": 1.8253238950749025e-05, "loss": 0.8025, "step": 4135 }, { "epoch": 0.65, "grad_norm": 19.710723629341924, "learning_rate": 1.8252286592016812e-05, "loss": 0.8281, "step": 4136 }, { "epoch": 0.65, "grad_norm": 19.947511865867735, "learning_rate": 1.8251333998594395e-05, "loss": 0.7433, "step": 4137 }, { "epoch": 0.65, "grad_norm": 13.028457690058588, "learning_rate": 1.825038117050887e-05, "loss": 0.7615, "step": 4138 }, { "epoch": 0.65, "grad_norm": 26.60533198398627, "learning_rate": 1.824942810778733e-05, "loss": 0.7839, "step": 4139 }, { "epoch": 0.65, "grad_norm": 17.996097847880318, "learning_rate": 1.8248474810456874e-05, "loss": 0.8439, "step": 4140 }, { "epoch": 0.65, "grad_norm": 19.433477798959455, "learning_rate": 1.8247521278544625e-05, "loss": 0.7769, "step": 4141 }, { "epoch": 0.65, "grad_norm": 15.530227256158849, "learning_rate": 1.8246567512077694e-05, "loss": 0.7159, "step": 4142 }, { "epoch": 0.65, "grad_norm": 14.970803228492773, "learning_rate": 1.824561351108321e-05, "loss": 0.7677, "step": 4143 }, { "epoch": 0.65, "grad_norm": 23.85053833675316, "learning_rate": 1.8244659275588298e-05, "loss": 0.8731, "step": 4144 }, { "epoch": 0.65, "grad_norm": 24.81336237324034, "learning_rate": 1.82437048056201e-05, "loss": 0.7852, "step": 4145 }, { "epoch": 0.65, "grad_norm": 17.01315384405582, "learning_rate": 1.8242750101205757e-05, "loss": 0.7627, "step": 4146 }, { "epoch": 0.65, "grad_norm": 19.354121519100417, "learning_rate": 1.8241795162372426e-05, "loss": 0.8619, "step": 4147 }, { "epoch": 0.65, "grad_norm": 16.814955547995996, "learning_rate": 1.8240839989147264e-05, "loss": 0.6823, "step": 4148 }, { "epoch": 0.65, "grad_norm": 12.764312652754464, "learning_rate": 1.823988458155743e-05, "loss": 0.7452, "step": 4149 }, { "epoch": 0.65, "grad_norm": 15.218329197629386, "learning_rate": 1.8238928939630097e-05, "loss": 0.7756, "step": 4150 }, { "epoch": 0.65, "grad_norm": 24.214087248146637, "learning_rate": 1.823797306339245e-05, "loss": 0.7437, "step": 4151 }, { "epoch": 0.65, "grad_norm": 27.496878017594884, "learning_rate": 1.8237016952871664e-05, "loss": 0.7958, "step": 4152 }, { "epoch": 0.65, "grad_norm": 22.105387664446674, "learning_rate": 1.8236060608094938e-05, "loss": 0.7972, "step": 4153 }, { "epoch": 0.65, "grad_norm": 20.260814013339807, "learning_rate": 1.8235104029089465e-05, "loss": 0.8071, "step": 4154 }, { "epoch": 0.65, "grad_norm": 23.197953402440604, "learning_rate": 1.823414721588245e-05, "loss": 0.7347, "step": 4155 }, { "epoch": 0.65, "grad_norm": 24.597249801764658, "learning_rate": 1.8233190168501104e-05, "loss": 0.7891, "step": 4156 }, { "epoch": 0.65, "grad_norm": 20.397370830305267, "learning_rate": 1.8232232886972643e-05, "loss": 0.7062, "step": 4157 }, { "epoch": 0.65, "grad_norm": 15.84046343709538, "learning_rate": 1.8231275371324303e-05, "loss": 0.696, "step": 4158 }, { "epoch": 0.65, "grad_norm": 20.091233622436835, "learning_rate": 1.82303176215833e-05, "loss": 0.7187, "step": 4159 }, { "epoch": 0.65, "grad_norm": 13.506800333979673, "learning_rate": 1.8229359637776883e-05, "loss": 0.746, "step": 4160 }, { "epoch": 0.65, "grad_norm": 20.212309330351744, "learning_rate": 1.8228401419932287e-05, "loss": 0.8117, "step": 4161 }, { "epoch": 0.65, "grad_norm": 17.157183395940265, "learning_rate": 1.822744296807677e-05, "loss": 0.7365, "step": 4162 }, { "epoch": 0.65, "grad_norm": 31.04302911307338, "learning_rate": 1.8226484282237587e-05, "loss": 0.768, "step": 4163 }, { "epoch": 0.65, "grad_norm": 19.208916504686204, "learning_rate": 1.822552536244201e-05, "loss": 0.6897, "step": 4164 }, { "epoch": 0.65, "grad_norm": 28.746442931343466, "learning_rate": 1.8224566208717296e-05, "loss": 0.7417, "step": 4165 }, { "epoch": 0.65, "grad_norm": 20.255061158054005, "learning_rate": 1.8223606821090733e-05, "loss": 0.7198, "step": 4166 }, { "epoch": 0.65, "grad_norm": 13.535106175537841, "learning_rate": 1.82226471995896e-05, "loss": 0.7751, "step": 4167 }, { "epoch": 0.65, "grad_norm": 15.6580760161256, "learning_rate": 1.8221687344241193e-05, "loss": 0.7885, "step": 4168 }, { "epoch": 0.65, "grad_norm": 17.306401910524993, "learning_rate": 1.822072725507281e-05, "loss": 0.7807, "step": 4169 }, { "epoch": 0.65, "grad_norm": 15.240129507863395, "learning_rate": 1.8219766932111747e-05, "loss": 0.7656, "step": 4170 }, { "epoch": 0.65, "grad_norm": 22.08386420108185, "learning_rate": 1.8218806375385326e-05, "loss": 0.8176, "step": 4171 }, { "epoch": 0.65, "grad_norm": 18.282687719982253, "learning_rate": 1.8217845584920858e-05, "loss": 0.7178, "step": 4172 }, { "epoch": 0.65, "grad_norm": 16.343011743673248, "learning_rate": 1.821688456074567e-05, "loss": 0.757, "step": 4173 }, { "epoch": 0.65, "grad_norm": 17.277970841041398, "learning_rate": 1.8215923302887085e-05, "loss": 0.8011, "step": 4174 }, { "epoch": 0.65, "grad_norm": 70.69089067862052, "learning_rate": 1.8214961811372455e-05, "loss": 0.8367, "step": 4175 }, { "epoch": 0.65, "grad_norm": 26.88569046519337, "learning_rate": 1.8214000086229115e-05, "loss": 0.7145, "step": 4176 }, { "epoch": 0.65, "grad_norm": 10.862646877911239, "learning_rate": 1.8213038127484415e-05, "loss": 0.724, "step": 4177 }, { "epoch": 0.65, "grad_norm": 23.664314450660363, "learning_rate": 1.821207593516572e-05, "loss": 0.8284, "step": 4178 }, { "epoch": 0.65, "grad_norm": 21.66913914893567, "learning_rate": 1.8211113509300384e-05, "loss": 0.9162, "step": 4179 }, { "epoch": 0.65, "grad_norm": 13.477847513495092, "learning_rate": 1.8210150849915787e-05, "loss": 0.7381, "step": 4180 }, { "epoch": 0.65, "grad_norm": 34.222411037368175, "learning_rate": 1.82091879570393e-05, "loss": 0.7785, "step": 4181 }, { "epoch": 0.65, "grad_norm": 24.01902531121858, "learning_rate": 1.8208224830698312e-05, "loss": 0.8747, "step": 4182 }, { "epoch": 0.65, "grad_norm": 20.111924042495076, "learning_rate": 1.820726147092021e-05, "loss": 0.849, "step": 4183 }, { "epoch": 0.65, "grad_norm": 12.247972078479735, "learning_rate": 1.8206297877732393e-05, "loss": 0.6994, "step": 4184 }, { "epoch": 0.65, "grad_norm": 16.675256512218837, "learning_rate": 1.8205334051162264e-05, "loss": 0.7962, "step": 4185 }, { "epoch": 0.65, "grad_norm": 18.884641425402634, "learning_rate": 1.8204369991237237e-05, "loss": 0.8158, "step": 4186 }, { "epoch": 0.65, "grad_norm": 21.349592604424227, "learning_rate": 1.8203405697984726e-05, "loss": 0.7835, "step": 4187 }, { "epoch": 0.65, "grad_norm": 15.53768262207262, "learning_rate": 1.8202441171432156e-05, "loss": 0.8492, "step": 4188 }, { "epoch": 0.65, "grad_norm": 19.984575034237253, "learning_rate": 1.8201476411606956e-05, "loss": 0.7321, "step": 4189 }, { "epoch": 0.65, "grad_norm": 18.785201187460796, "learning_rate": 1.8200511418536563e-05, "loss": 0.6348, "step": 4190 }, { "epoch": 0.65, "grad_norm": 20.782846553680805, "learning_rate": 1.819954619224843e-05, "loss": 0.9021, "step": 4191 }, { "epoch": 0.65, "grad_norm": 25.22696632069129, "learning_rate": 1.8198580732769992e-05, "loss": 0.8692, "step": 4192 }, { "epoch": 0.65, "grad_norm": 23.90325030911037, "learning_rate": 1.819761504012872e-05, "loss": 0.8521, "step": 4193 }, { "epoch": 0.66, "grad_norm": 15.209610022450809, "learning_rate": 1.8196649114352065e-05, "loss": 0.8209, "step": 4194 }, { "epoch": 0.66, "grad_norm": 17.128908647560987, "learning_rate": 1.819568295546751e-05, "loss": 0.7317, "step": 4195 }, { "epoch": 0.66, "grad_norm": 21.094954425899743, "learning_rate": 1.819471656350252e-05, "loss": 0.7932, "step": 4196 }, { "epoch": 0.66, "grad_norm": 22.65058012686775, "learning_rate": 1.819374993848459e-05, "loss": 0.8494, "step": 4197 }, { "epoch": 0.66, "grad_norm": 24.688260305744706, "learning_rate": 1.8192783080441204e-05, "loss": 0.8413, "step": 4198 }, { "epoch": 0.66, "grad_norm": 17.12246524496961, "learning_rate": 1.819181598939986e-05, "loss": 0.7947, "step": 4199 }, { "epoch": 0.66, "grad_norm": 17.72511579142144, "learning_rate": 1.819084866538806e-05, "loss": 0.8739, "step": 4200 }, { "epoch": 0.66, "grad_norm": 23.70355082401617, "learning_rate": 1.8189881108433317e-05, "loss": 0.8834, "step": 4201 }, { "epoch": 0.66, "grad_norm": 21.189093626553777, "learning_rate": 1.8188913318563145e-05, "loss": 0.8128, "step": 4202 }, { "epoch": 0.66, "grad_norm": 22.59124555399027, "learning_rate": 1.8187945295805066e-05, "loss": 0.6995, "step": 4203 }, { "epoch": 0.66, "grad_norm": 17.390925392819383, "learning_rate": 1.8186977040186615e-05, "loss": 0.8919, "step": 4204 }, { "epoch": 0.66, "grad_norm": 14.713384502269177, "learning_rate": 1.8186008551735327e-05, "loss": 0.7108, "step": 4205 }, { "epoch": 0.66, "grad_norm": 17.107845542862567, "learning_rate": 1.8185039830478745e-05, "loss": 0.8084, "step": 4206 }, { "epoch": 0.66, "grad_norm": 18.85657301601761, "learning_rate": 1.8184070876444417e-05, "loss": 0.7317, "step": 4207 }, { "epoch": 0.66, "grad_norm": 24.990322371369285, "learning_rate": 1.8183101689659902e-05, "loss": 0.7218, "step": 4208 }, { "epoch": 0.66, "grad_norm": 16.208851136761886, "learning_rate": 1.818213227015276e-05, "loss": 0.7399, "step": 4209 }, { "epoch": 0.66, "grad_norm": 16.142991538272394, "learning_rate": 1.8181162617950563e-05, "loss": 0.76, "step": 4210 }, { "epoch": 0.66, "grad_norm": 22.846959313289375, "learning_rate": 1.8180192733080887e-05, "loss": 0.7868, "step": 4211 }, { "epoch": 0.66, "grad_norm": 34.90326480343949, "learning_rate": 1.817922261557132e-05, "loss": 0.7735, "step": 4212 }, { "epoch": 0.66, "grad_norm": 21.493681543379306, "learning_rate": 1.817825226544944e-05, "loss": 0.8084, "step": 4213 }, { "epoch": 0.66, "grad_norm": 23.45435698397495, "learning_rate": 1.817728168274285e-05, "loss": 0.8268, "step": 4214 }, { "epoch": 0.66, "grad_norm": 22.28494940819731, "learning_rate": 1.8176310867479155e-05, "loss": 0.7959, "step": 4215 }, { "epoch": 0.66, "grad_norm": 12.652858458670133, "learning_rate": 1.8175339819685965e-05, "loss": 0.6771, "step": 4216 }, { "epoch": 0.66, "grad_norm": 14.71053667378346, "learning_rate": 1.8174368539390888e-05, "loss": 0.7457, "step": 4217 }, { "epoch": 0.66, "grad_norm": 21.2796520327813, "learning_rate": 1.8173397026621554e-05, "loss": 0.7368, "step": 4218 }, { "epoch": 0.66, "grad_norm": 12.584496892201871, "learning_rate": 1.817242528140559e-05, "loss": 0.7424, "step": 4219 }, { "epoch": 0.66, "grad_norm": 16.700335026067325, "learning_rate": 1.8171453303770628e-05, "loss": 0.8336, "step": 4220 }, { "epoch": 0.66, "grad_norm": 23.60266369823785, "learning_rate": 1.8170481093744317e-05, "loss": 0.7677, "step": 4221 }, { "epoch": 0.66, "grad_norm": 20.796280385952652, "learning_rate": 1.81695086513543e-05, "loss": 0.825, "step": 4222 }, { "epoch": 0.66, "grad_norm": 20.339328609160805, "learning_rate": 1.8168535976628242e-05, "loss": 0.8076, "step": 4223 }, { "epoch": 0.66, "grad_norm": 20.664036607996508, "learning_rate": 1.8167563069593796e-05, "loss": 0.8796, "step": 4224 }, { "epoch": 0.66, "grad_norm": 18.685501119311006, "learning_rate": 1.8166589930278637e-05, "loss": 0.8069, "step": 4225 }, { "epoch": 0.66, "grad_norm": 18.0908378763667, "learning_rate": 1.816561655871043e-05, "loss": 0.731, "step": 4226 }, { "epoch": 0.66, "grad_norm": 33.870150718719664, "learning_rate": 1.816464295491687e-05, "loss": 0.6768, "step": 4227 }, { "epoch": 0.66, "grad_norm": 14.729729567852427, "learning_rate": 1.816366911892564e-05, "loss": 0.7056, "step": 4228 }, { "epoch": 0.66, "grad_norm": 25.702781940478168, "learning_rate": 1.8162695050764437e-05, "loss": 0.8006, "step": 4229 }, { "epoch": 0.66, "grad_norm": 46.41708567786615, "learning_rate": 1.8161720750460963e-05, "loss": 0.771, "step": 4230 }, { "epoch": 0.66, "grad_norm": 19.36104449712154, "learning_rate": 1.8160746218042923e-05, "loss": 0.7014, "step": 4231 }, { "epoch": 0.66, "grad_norm": 13.573114581859631, "learning_rate": 1.8159771453538034e-05, "loss": 0.6727, "step": 4232 }, { "epoch": 0.66, "grad_norm": 15.327036339677111, "learning_rate": 1.8158796456974023e-05, "loss": 0.7674, "step": 4233 }, { "epoch": 0.66, "grad_norm": 16.592549386565253, "learning_rate": 1.8157821228378607e-05, "loss": 0.8118, "step": 4234 }, { "epoch": 0.66, "grad_norm": 18.06741184879512, "learning_rate": 1.8156845767779532e-05, "loss": 0.8578, "step": 4235 }, { "epoch": 0.66, "grad_norm": 19.853984133347314, "learning_rate": 1.815587007520453e-05, "loss": 0.7859, "step": 4236 }, { "epoch": 0.66, "grad_norm": 23.56419704343956, "learning_rate": 1.815489415068136e-05, "loss": 0.7317, "step": 4237 }, { "epoch": 0.66, "grad_norm": 15.464204674312242, "learning_rate": 1.8153917994237765e-05, "loss": 0.7163, "step": 4238 }, { "epoch": 0.66, "grad_norm": 25.546874955670187, "learning_rate": 1.8152941605901518e-05, "loss": 0.8373, "step": 4239 }, { "epoch": 0.66, "grad_norm": 20.286881393131456, "learning_rate": 1.8151964985700375e-05, "loss": 0.7966, "step": 4240 }, { "epoch": 0.66, "grad_norm": 16.69725180837408, "learning_rate": 1.815098813366212e-05, "loss": 0.6484, "step": 4241 }, { "epoch": 0.66, "grad_norm": 20.485323425701363, "learning_rate": 1.8150011049814534e-05, "loss": 0.7915, "step": 4242 }, { "epoch": 0.66, "grad_norm": 16.325577534439056, "learning_rate": 1.8149033734185396e-05, "loss": 0.7345, "step": 4243 }, { "epoch": 0.66, "grad_norm": 20.35399091586787, "learning_rate": 1.814805618680251e-05, "loss": 0.8457, "step": 4244 }, { "epoch": 0.66, "grad_norm": 26.326481407887844, "learning_rate": 1.814707840769367e-05, "loss": 0.796, "step": 4245 }, { "epoch": 0.66, "grad_norm": 19.757716482222545, "learning_rate": 1.8146100396886685e-05, "loss": 0.7758, "step": 4246 }, { "epoch": 0.66, "grad_norm": 18.672847433984025, "learning_rate": 1.814512215440937e-05, "loss": 0.7704, "step": 4247 }, { "epoch": 0.66, "grad_norm": 19.76956626890496, "learning_rate": 1.8144143680289547e-05, "loss": 0.777, "step": 4248 }, { "epoch": 0.66, "grad_norm": 16.811800221589184, "learning_rate": 1.8143164974555042e-05, "loss": 0.724, "step": 4249 }, { "epoch": 0.66, "grad_norm": 19.37250055988275, "learning_rate": 1.8142186037233687e-05, "loss": 0.8216, "step": 4250 }, { "epoch": 0.66, "grad_norm": 17.467740504008308, "learning_rate": 1.8141206868353327e-05, "loss": 0.7109, "step": 4251 }, { "epoch": 0.66, "grad_norm": 14.09127753365462, "learning_rate": 1.8140227467941803e-05, "loss": 0.6957, "step": 4252 }, { "epoch": 0.66, "grad_norm": 17.512332261362477, "learning_rate": 1.8139247836026977e-05, "loss": 0.8107, "step": 4253 }, { "epoch": 0.66, "grad_norm": 15.053714789069293, "learning_rate": 1.8138267972636697e-05, "loss": 0.8666, "step": 4254 }, { "epoch": 0.66, "grad_norm": 27.58978856186665, "learning_rate": 1.8137287877798842e-05, "loss": 0.8388, "step": 4255 }, { "epoch": 0.66, "grad_norm": 21.86409343623787, "learning_rate": 1.8136307551541276e-05, "loss": 0.8195, "step": 4256 }, { "epoch": 0.66, "grad_norm": 17.1160706365079, "learning_rate": 1.8135326993891886e-05, "loss": 0.8557, "step": 4257 }, { "epoch": 0.67, "grad_norm": 21.49182803457618, "learning_rate": 1.8134346204878553e-05, "loss": 0.8223, "step": 4258 }, { "epoch": 0.67, "grad_norm": 15.701361360959423, "learning_rate": 1.8133365184529176e-05, "loss": 0.7515, "step": 4259 }, { "epoch": 0.67, "grad_norm": 31.154165733893155, "learning_rate": 1.8132383932871645e-05, "loss": 0.6681, "step": 4260 }, { "epoch": 0.67, "grad_norm": 32.96153690202711, "learning_rate": 1.8131402449933875e-05, "loss": 0.7608, "step": 4261 }, { "epoch": 0.67, "grad_norm": 17.93937119749214, "learning_rate": 1.813042073574378e-05, "loss": 0.7109, "step": 4262 }, { "epoch": 0.67, "grad_norm": 17.42305734839047, "learning_rate": 1.8129438790329272e-05, "loss": 0.7843, "step": 4263 }, { "epoch": 0.67, "grad_norm": 15.325273479441474, "learning_rate": 1.8128456613718278e-05, "loss": 0.7861, "step": 4264 }, { "epoch": 0.67, "grad_norm": 18.82262801814742, "learning_rate": 1.812747420593874e-05, "loss": 0.7791, "step": 4265 }, { "epoch": 0.67, "grad_norm": 20.475188254677505, "learning_rate": 1.8126491567018582e-05, "loss": 0.7828, "step": 4266 }, { "epoch": 0.67, "grad_norm": 19.533285579654216, "learning_rate": 1.8125508696985763e-05, "loss": 0.8698, "step": 4267 }, { "epoch": 0.67, "grad_norm": 21.120126518310595, "learning_rate": 1.8124525595868228e-05, "loss": 0.7757, "step": 4268 }, { "epoch": 0.67, "grad_norm": 20.919112486785913, "learning_rate": 1.8123542263693935e-05, "loss": 0.7934, "step": 4269 }, { "epoch": 0.67, "grad_norm": 11.94620908456108, "learning_rate": 1.8122558700490857e-05, "loss": 0.7079, "step": 4270 }, { "epoch": 0.67, "grad_norm": 19.697023743339116, "learning_rate": 1.812157490628696e-05, "loss": 0.7158, "step": 4271 }, { "epoch": 0.67, "grad_norm": 21.91522053125438, "learning_rate": 1.812059088111022e-05, "loss": 0.7142, "step": 4272 }, { "epoch": 0.67, "grad_norm": 18.362811639425157, "learning_rate": 1.8119606624988632e-05, "loss": 0.7401, "step": 4273 }, { "epoch": 0.67, "grad_norm": 15.109241423859642, "learning_rate": 1.8118622137950174e-05, "loss": 0.7763, "step": 4274 }, { "epoch": 0.67, "grad_norm": 27.86274346394573, "learning_rate": 1.8117637420022858e-05, "loss": 0.8347, "step": 4275 }, { "epoch": 0.67, "grad_norm": 15.750907907969598, "learning_rate": 1.8116652471234682e-05, "loss": 0.7227, "step": 4276 }, { "epoch": 0.67, "grad_norm": 13.535649071841243, "learning_rate": 1.8115667291613653e-05, "loss": 0.7447, "step": 4277 }, { "epoch": 0.67, "grad_norm": 18.85389388916283, "learning_rate": 1.8114681881187795e-05, "loss": 0.7143, "step": 4278 }, { "epoch": 0.67, "grad_norm": 17.848976007142422, "learning_rate": 1.8113696239985134e-05, "loss": 0.8332, "step": 4279 }, { "epoch": 0.67, "grad_norm": 12.079137332828445, "learning_rate": 1.81127103680337e-05, "loss": 0.7348, "step": 4280 }, { "epoch": 0.67, "grad_norm": 14.437528922415328, "learning_rate": 1.8111724265361525e-05, "loss": 0.6703, "step": 4281 }, { "epoch": 0.67, "grad_norm": 17.77181728523707, "learning_rate": 1.8110737931996658e-05, "loss": 0.7071, "step": 4282 }, { "epoch": 0.67, "grad_norm": 19.94980380782242, "learning_rate": 1.8109751367967148e-05, "loss": 0.7469, "step": 4283 }, { "epoch": 0.67, "grad_norm": 19.21179685497273, "learning_rate": 1.8108764573301054e-05, "loss": 0.7592, "step": 4284 }, { "epoch": 0.67, "grad_norm": 30.20020004284759, "learning_rate": 1.810777754802644e-05, "loss": 0.7037, "step": 4285 }, { "epoch": 0.67, "grad_norm": 22.464562474179527, "learning_rate": 1.8106790292171378e-05, "loss": 0.7459, "step": 4286 }, { "epoch": 0.67, "grad_norm": 17.493058091409495, "learning_rate": 1.8105802805763938e-05, "loss": 0.7197, "step": 4287 }, { "epoch": 0.67, "grad_norm": 25.784912404683094, "learning_rate": 1.810481508883221e-05, "loss": 0.7689, "step": 4288 }, { "epoch": 0.67, "grad_norm": 16.012711676814728, "learning_rate": 1.810382714140428e-05, "loss": 0.7586, "step": 4289 }, { "epoch": 0.67, "grad_norm": 19.877625776276066, "learning_rate": 1.8102838963508247e-05, "loss": 0.7816, "step": 4290 }, { "epoch": 0.67, "grad_norm": 26.141985244497537, "learning_rate": 1.8101850555172214e-05, "loss": 0.9162, "step": 4291 }, { "epoch": 0.67, "grad_norm": 25.660256589872496, "learning_rate": 1.8100861916424293e-05, "loss": 0.8101, "step": 4292 }, { "epoch": 0.67, "grad_norm": 21.79914055128638, "learning_rate": 1.8099873047292596e-05, "loss": 0.8228, "step": 4293 }, { "epoch": 0.67, "grad_norm": 24.090390228412115, "learning_rate": 1.8098883947805247e-05, "loss": 0.8054, "step": 4294 }, { "epoch": 0.67, "grad_norm": 29.20721872618258, "learning_rate": 1.8097894617990378e-05, "loss": 0.7194, "step": 4295 }, { "epoch": 0.67, "grad_norm": 16.476324899080826, "learning_rate": 1.809690505787612e-05, "loss": 0.8397, "step": 4296 }, { "epoch": 0.67, "grad_norm": 17.93750872429027, "learning_rate": 1.8095915267490625e-05, "loss": 0.7262, "step": 4297 }, { "epoch": 0.67, "grad_norm": 18.124958189143264, "learning_rate": 1.809492524686203e-05, "loss": 0.7357, "step": 4298 }, { "epoch": 0.67, "grad_norm": 28.13115606275349, "learning_rate": 1.80939349960185e-05, "loss": 0.8305, "step": 4299 }, { "epoch": 0.67, "grad_norm": 23.035539915949922, "learning_rate": 1.8092944514988193e-05, "loss": 0.9096, "step": 4300 }, { "epoch": 0.67, "grad_norm": 16.872842539960725, "learning_rate": 1.8091953803799275e-05, "loss": 0.6916, "step": 4301 }, { "epoch": 0.67, "grad_norm": 20.77223099773228, "learning_rate": 1.8090962862479927e-05, "loss": 0.8311, "step": 4302 }, { "epoch": 0.67, "grad_norm": 16.082046418904415, "learning_rate": 1.8089971691058327e-05, "loss": 0.8602, "step": 4303 }, { "epoch": 0.67, "grad_norm": 29.242166431510118, "learning_rate": 1.8088980289562666e-05, "loss": 0.8083, "step": 4304 }, { "epoch": 0.67, "grad_norm": 20.3028884639645, "learning_rate": 1.8087988658021137e-05, "loss": 0.7272, "step": 4305 }, { "epoch": 0.67, "grad_norm": 16.733338198346054, "learning_rate": 1.8086996796461943e-05, "loss": 0.7799, "step": 4306 }, { "epoch": 0.67, "grad_norm": 18.619174167735295, "learning_rate": 1.808600470491329e-05, "loss": 0.8411, "step": 4307 }, { "epoch": 0.67, "grad_norm": 20.88761445214009, "learning_rate": 1.8085012383403392e-05, "loss": 0.7286, "step": 4308 }, { "epoch": 0.67, "grad_norm": 21.059224807944098, "learning_rate": 1.8084019831960474e-05, "loss": 0.7332, "step": 4309 }, { "epoch": 0.67, "grad_norm": 13.722668229425524, "learning_rate": 1.808302705061276e-05, "loss": 0.6538, "step": 4310 }, { "epoch": 0.67, "grad_norm": 16.36886703664029, "learning_rate": 1.8082034039388486e-05, "loss": 0.7814, "step": 4311 }, { "epoch": 0.67, "grad_norm": 33.47505662217455, "learning_rate": 1.808104079831589e-05, "loss": 0.8421, "step": 4312 }, { "epoch": 0.67, "grad_norm": 16.327617185398385, "learning_rate": 1.808004732742322e-05, "loss": 0.7285, "step": 4313 }, { "epoch": 0.67, "grad_norm": 17.1307651272322, "learning_rate": 1.8079053626738732e-05, "loss": 0.7086, "step": 4314 }, { "epoch": 0.67, "grad_norm": 43.19829755649961, "learning_rate": 1.8078059696290685e-05, "loss": 0.8283, "step": 4315 }, { "epoch": 0.67, "grad_norm": 17.407922130239847, "learning_rate": 1.8077065536107347e-05, "loss": 0.8116, "step": 4316 }, { "epoch": 0.67, "grad_norm": 15.62592827042257, "learning_rate": 1.8076071146216988e-05, "loss": 0.8064, "step": 4317 }, { "epoch": 0.67, "grad_norm": 28.946234029702975, "learning_rate": 1.8075076526647892e-05, "loss": 0.7978, "step": 4318 }, { "epoch": 0.67, "grad_norm": 16.55398774569627, "learning_rate": 1.807408167742834e-05, "loss": 0.8183, "step": 4319 }, { "epoch": 0.67, "grad_norm": 32.310680777675756, "learning_rate": 1.8073086598586634e-05, "loss": 0.8057, "step": 4320 }, { "epoch": 0.67, "grad_norm": 22.56848686112926, "learning_rate": 1.8072091290151063e-05, "loss": 0.8163, "step": 4321 }, { "epoch": 0.68, "grad_norm": 28.680198467444463, "learning_rate": 1.807109575214994e-05, "loss": 0.7474, "step": 4322 }, { "epoch": 0.68, "grad_norm": 14.888657405865755, "learning_rate": 1.8070099984611575e-05, "loss": 0.6191, "step": 4323 }, { "epoch": 0.68, "grad_norm": 17.62726442014608, "learning_rate": 1.8069103987564286e-05, "loss": 0.741, "step": 4324 }, { "epoch": 0.68, "grad_norm": 24.28482620434681, "learning_rate": 1.8068107761036402e-05, "loss": 0.8525, "step": 4325 }, { "epoch": 0.68, "grad_norm": 21.900999161629514, "learning_rate": 1.806711130505625e-05, "loss": 0.8357, "step": 4326 }, { "epoch": 0.68, "grad_norm": 25.085255612111194, "learning_rate": 1.806611461965217e-05, "loss": 0.7387, "step": 4327 }, { "epoch": 0.68, "grad_norm": 20.54601720172465, "learning_rate": 1.8065117704852518e-05, "loss": 0.8159, "step": 4328 }, { "epoch": 0.68, "grad_norm": 22.934420265192585, "learning_rate": 1.8064120560685626e-05, "loss": 0.813, "step": 4329 }, { "epoch": 0.68, "grad_norm": 18.78721142610371, "learning_rate": 1.806312318717987e-05, "loss": 0.8022, "step": 4330 }, { "epoch": 0.68, "grad_norm": 19.46167464757203, "learning_rate": 1.8062125584363605e-05, "loss": 0.8014, "step": 4331 }, { "epoch": 0.68, "grad_norm": 19.9684230010178, "learning_rate": 1.8061127752265203e-05, "loss": 0.7071, "step": 4332 }, { "epoch": 0.68, "grad_norm": 17.1168461778214, "learning_rate": 1.8060129690913045e-05, "loss": 0.7522, "step": 4333 }, { "epoch": 0.68, "grad_norm": 20.285889726700216, "learning_rate": 1.8059131400335512e-05, "loss": 0.8363, "step": 4334 }, { "epoch": 0.68, "grad_norm": 14.347939714102884, "learning_rate": 1.8058132880560998e-05, "loss": 0.7959, "step": 4335 }, { "epoch": 0.68, "grad_norm": 25.419360362360546, "learning_rate": 1.80571341316179e-05, "loss": 0.7799, "step": 4336 }, { "epoch": 0.68, "grad_norm": 17.75484865077361, "learning_rate": 1.8056135153534616e-05, "loss": 0.7569, "step": 4337 }, { "epoch": 0.68, "grad_norm": 18.53483295500399, "learning_rate": 1.8055135946339563e-05, "loss": 0.8076, "step": 4338 }, { "epoch": 0.68, "grad_norm": 14.414970112993563, "learning_rate": 1.8054136510061156e-05, "loss": 0.7859, "step": 4339 }, { "epoch": 0.68, "grad_norm": 15.435482671334562, "learning_rate": 1.8053136844727822e-05, "loss": 0.7089, "step": 4340 }, { "epoch": 0.68, "grad_norm": 18.503406182091712, "learning_rate": 1.8052136950367983e-05, "loss": 0.7825, "step": 4341 }, { "epoch": 0.68, "grad_norm": 21.00745055602217, "learning_rate": 1.8051136827010077e-05, "loss": 0.8883, "step": 4342 }, { "epoch": 0.68, "grad_norm": 16.391674330898983, "learning_rate": 1.8050136474682553e-05, "loss": 0.7553, "step": 4343 }, { "epoch": 0.68, "grad_norm": 18.554786039391917, "learning_rate": 1.8049135893413858e-05, "loss": 0.812, "step": 4344 }, { "epoch": 0.68, "grad_norm": 12.314777588105425, "learning_rate": 1.8048135083232446e-05, "loss": 0.6781, "step": 4345 }, { "epoch": 0.68, "grad_norm": 17.949065210061367, "learning_rate": 1.804713404416678e-05, "loss": 0.8333, "step": 4346 }, { "epoch": 0.68, "grad_norm": 4.591601797200667, "learning_rate": 1.8046132776245327e-05, "loss": 0.6883, "step": 4347 }, { "epoch": 0.68, "grad_norm": 14.932887640126669, "learning_rate": 1.804513127949657e-05, "loss": 0.8157, "step": 4348 }, { "epoch": 0.68, "grad_norm": 15.288021167190276, "learning_rate": 1.8044129553948978e-05, "loss": 0.8966, "step": 4349 }, { "epoch": 0.68, "grad_norm": 16.91643776899268, "learning_rate": 1.8043127599631053e-05, "loss": 0.7158, "step": 4350 }, { "epoch": 0.68, "grad_norm": 20.067793257401863, "learning_rate": 1.8042125416571284e-05, "loss": 0.7801, "step": 4351 }, { "epoch": 0.68, "grad_norm": 19.39998942130946, "learning_rate": 1.8041123004798173e-05, "loss": 0.7697, "step": 4352 }, { "epoch": 0.68, "grad_norm": 20.100686927107667, "learning_rate": 1.8040120364340223e-05, "loss": 0.8217, "step": 4353 }, { "epoch": 0.68, "grad_norm": 28.562866033299674, "learning_rate": 1.8039117495225956e-05, "loss": 0.9237, "step": 4354 }, { "epoch": 0.68, "grad_norm": 13.732006388989692, "learning_rate": 1.803811439748389e-05, "loss": 0.7574, "step": 4355 }, { "epoch": 0.68, "grad_norm": 21.841406363579924, "learning_rate": 1.8037111071142552e-05, "loss": 0.8425, "step": 4356 }, { "epoch": 0.68, "grad_norm": 15.001851837765855, "learning_rate": 1.8036107516230478e-05, "loss": 0.7416, "step": 4357 }, { "epoch": 0.68, "grad_norm": 27.63443504792, "learning_rate": 1.8035103732776206e-05, "loss": 0.8568, "step": 4358 }, { "epoch": 0.68, "grad_norm": 40.96006777616897, "learning_rate": 1.803409972080829e-05, "loss": 0.7273, "step": 4359 }, { "epoch": 0.68, "grad_norm": 16.101674452924147, "learning_rate": 1.803309548035527e-05, "loss": 0.7821, "step": 4360 }, { "epoch": 0.68, "grad_norm": 20.56604787182817, "learning_rate": 1.8032091011445714e-05, "loss": 0.7877, "step": 4361 }, { "epoch": 0.68, "grad_norm": 26.465669799806857, "learning_rate": 1.803108631410819e-05, "loss": 0.7313, "step": 4362 }, { "epoch": 0.68, "grad_norm": 3.1204724421504864, "learning_rate": 1.803008138837127e-05, "loss": 0.676, "step": 4363 }, { "epoch": 0.68, "grad_norm": 40.72773430428415, "learning_rate": 1.8029076234263534e-05, "loss": 0.7996, "step": 4364 }, { "epoch": 0.68, "grad_norm": 16.66358343026831, "learning_rate": 1.8028070851813566e-05, "loss": 0.7804, "step": 4365 }, { "epoch": 0.68, "grad_norm": 18.6108139596374, "learning_rate": 1.802706524104996e-05, "loss": 0.8305, "step": 4366 }, { "epoch": 0.68, "grad_norm": 18.040803518821154, "learning_rate": 1.8026059402001313e-05, "loss": 0.6919, "step": 4367 }, { "epoch": 0.68, "grad_norm": 13.762513375167515, "learning_rate": 1.8025053334696232e-05, "loss": 0.7765, "step": 4368 }, { "epoch": 0.68, "grad_norm": 16.129159322551544, "learning_rate": 1.8024047039163325e-05, "loss": 0.7246, "step": 4369 }, { "epoch": 0.68, "grad_norm": 23.06298176332034, "learning_rate": 1.802304051543122e-05, "loss": 0.7454, "step": 4370 }, { "epoch": 0.68, "grad_norm": 24.823940025223486, "learning_rate": 1.8022033763528534e-05, "loss": 0.8304, "step": 4371 }, { "epoch": 0.68, "grad_norm": 17.53250667161487, "learning_rate": 1.80210267834839e-05, "loss": 0.707, "step": 4372 }, { "epoch": 0.68, "grad_norm": 15.50934033879026, "learning_rate": 1.802001957532596e-05, "loss": 0.7716, "step": 4373 }, { "epoch": 0.68, "grad_norm": 16.948848468000165, "learning_rate": 1.801901213908335e-05, "loss": 0.7268, "step": 4374 }, { "epoch": 0.68, "grad_norm": 19.420163334666416, "learning_rate": 1.801800447478473e-05, "loss": 0.896, "step": 4375 }, { "epoch": 0.68, "grad_norm": 26.897851962525642, "learning_rate": 1.8016996582458746e-05, "loss": 0.7131, "step": 4376 }, { "epoch": 0.68, "grad_norm": 13.634898020198941, "learning_rate": 1.8015988462134073e-05, "loss": 0.7771, "step": 4377 }, { "epoch": 0.68, "grad_norm": 16.821859129615284, "learning_rate": 1.801498011383938e-05, "loss": 0.7458, "step": 4378 }, { "epoch": 0.68, "grad_norm": 16.127979127194294, "learning_rate": 1.8013971537603336e-05, "loss": 0.7777, "step": 4379 }, { "epoch": 0.68, "grad_norm": 20.924722870477922, "learning_rate": 1.8012962733454636e-05, "loss": 0.7691, "step": 4380 }, { "epoch": 0.68, "grad_norm": 14.171522156088209, "learning_rate": 1.8011953701421957e-05, "loss": 0.7542, "step": 4381 }, { "epoch": 0.68, "grad_norm": 22.07723572297356, "learning_rate": 1.8010944441534004e-05, "loss": 0.7337, "step": 4382 }, { "epoch": 0.68, "grad_norm": 17.69890807634116, "learning_rate": 1.800993495381948e-05, "loss": 0.7926, "step": 4383 }, { "epoch": 0.68, "grad_norm": 21.605652896121875, "learning_rate": 1.800892523830709e-05, "loss": 0.8293, "step": 4384 }, { "epoch": 0.68, "grad_norm": 26.276327549891764, "learning_rate": 1.8007915295025552e-05, "loss": 0.7653, "step": 4385 }, { "epoch": 0.69, "grad_norm": 16.608032089735506, "learning_rate": 1.8006905124003583e-05, "loss": 0.8014, "step": 4386 }, { "epoch": 0.69, "grad_norm": 17.580945152395373, "learning_rate": 1.8005894725269918e-05, "loss": 0.8588, "step": 4387 }, { "epoch": 0.69, "grad_norm": 22.8424159468008, "learning_rate": 1.8004884098853296e-05, "loss": 0.7619, "step": 4388 }, { "epoch": 0.69, "grad_norm": 12.82480906982031, "learning_rate": 1.8003873244782447e-05, "loss": 0.7729, "step": 4389 }, { "epoch": 0.69, "grad_norm": 15.69975581915073, "learning_rate": 1.8002862163086127e-05, "loss": 0.7613, "step": 4390 }, { "epoch": 0.69, "grad_norm": 17.51018684676257, "learning_rate": 1.8001850853793084e-05, "loss": 0.7006, "step": 4391 }, { "epoch": 0.69, "grad_norm": 15.82898322524861, "learning_rate": 1.8000839316932094e-05, "loss": 0.7644, "step": 4392 }, { "epoch": 0.69, "grad_norm": 20.407322404147806, "learning_rate": 1.7999827552531905e-05, "loss": 0.7579, "step": 4393 }, { "epoch": 0.69, "grad_norm": 15.723444178959944, "learning_rate": 1.7998815560621305e-05, "loss": 0.8213, "step": 4394 }, { "epoch": 0.69, "grad_norm": 21.889634416972665, "learning_rate": 1.7997803341229067e-05, "loss": 0.8172, "step": 4395 }, { "epoch": 0.69, "grad_norm": 11.723829215174922, "learning_rate": 1.7996790894383982e-05, "loss": 0.6714, "step": 4396 }, { "epoch": 0.69, "grad_norm": 27.59433678373567, "learning_rate": 1.799577822011484e-05, "loss": 0.7723, "step": 4397 }, { "epoch": 0.69, "grad_norm": 14.131878716031913, "learning_rate": 1.7994765318450446e-05, "loss": 0.7368, "step": 4398 }, { "epoch": 0.69, "grad_norm": 22.187226373704544, "learning_rate": 1.7993752189419602e-05, "loss": 0.7389, "step": 4399 }, { "epoch": 0.69, "grad_norm": 16.6787192713573, "learning_rate": 1.7992738833051122e-05, "loss": 0.7607, "step": 4400 }, { "epoch": 0.69, "grad_norm": 25.246798452805926, "learning_rate": 1.7991725249373822e-05, "loss": 0.8209, "step": 4401 }, { "epoch": 0.69, "grad_norm": 19.80020465075268, "learning_rate": 1.7990711438416536e-05, "loss": 0.7652, "step": 4402 }, { "epoch": 0.69, "grad_norm": 18.88989684970984, "learning_rate": 1.798969740020809e-05, "loss": 0.7749, "step": 4403 }, { "epoch": 0.69, "grad_norm": 26.274208877065714, "learning_rate": 1.798868313477732e-05, "loss": 0.6947, "step": 4404 }, { "epoch": 0.69, "grad_norm": 14.882254386309226, "learning_rate": 1.7987668642153075e-05, "loss": 0.7086, "step": 4405 }, { "epoch": 0.69, "grad_norm": 16.090227360052566, "learning_rate": 1.798665392236421e-05, "loss": 0.6781, "step": 4406 }, { "epoch": 0.69, "grad_norm": 16.378504637878486, "learning_rate": 1.7985638975439578e-05, "loss": 0.7496, "step": 4407 }, { "epoch": 0.69, "grad_norm": 26.369406659588574, "learning_rate": 1.7984623801408046e-05, "loss": 0.7258, "step": 4408 }, { "epoch": 0.69, "grad_norm": 15.660206540360726, "learning_rate": 1.798360840029848e-05, "loss": 0.7535, "step": 4409 }, { "epoch": 0.69, "grad_norm": 15.986224646399426, "learning_rate": 1.7982592772139766e-05, "loss": 0.7267, "step": 4410 }, { "epoch": 0.69, "grad_norm": 15.791240883508046, "learning_rate": 1.7981576916960784e-05, "loss": 0.8889, "step": 4411 }, { "epoch": 0.69, "grad_norm": 13.433677329768178, "learning_rate": 1.798056083479042e-05, "loss": 0.732, "step": 4412 }, { "epoch": 0.69, "grad_norm": 15.085880260665142, "learning_rate": 1.7979544525657578e-05, "loss": 0.7393, "step": 4413 }, { "epoch": 0.69, "grad_norm": 17.73487400201541, "learning_rate": 1.797852798959115e-05, "loss": 0.7295, "step": 4414 }, { "epoch": 0.69, "grad_norm": 20.464276245659956, "learning_rate": 1.7977511226620062e-05, "loss": 0.8915, "step": 4415 }, { "epoch": 0.69, "grad_norm": 21.289352732438633, "learning_rate": 1.7976494236773217e-05, "loss": 0.8119, "step": 4416 }, { "epoch": 0.69, "grad_norm": 18.103478341841534, "learning_rate": 1.7975477020079544e-05, "loss": 0.7636, "step": 4417 }, { "epoch": 0.69, "grad_norm": 18.893535765235537, "learning_rate": 1.7974459576567968e-05, "loss": 0.7475, "step": 4418 }, { "epoch": 0.69, "grad_norm": 27.040377117318236, "learning_rate": 1.797344190626743e-05, "loss": 0.8772, "step": 4419 }, { "epoch": 0.69, "grad_norm": 17.792159452154106, "learning_rate": 1.7972424009206865e-05, "loss": 0.7556, "step": 4420 }, { "epoch": 0.69, "grad_norm": 22.696463105651482, "learning_rate": 1.7971405885415228e-05, "loss": 0.7266, "step": 4421 }, { "epoch": 0.69, "grad_norm": 16.240044561619914, "learning_rate": 1.7970387534921467e-05, "loss": 0.7683, "step": 4422 }, { "epoch": 0.69, "grad_norm": 23.60420954626879, "learning_rate": 1.796936895775455e-05, "loss": 0.7696, "step": 4423 }, { "epoch": 0.69, "grad_norm": 22.518673024693154, "learning_rate": 1.7968350153943443e-05, "loss": 0.8025, "step": 4424 }, { "epoch": 0.69, "grad_norm": 10.974492961960626, "learning_rate": 1.7967331123517114e-05, "loss": 0.6361, "step": 4425 }, { "epoch": 0.69, "grad_norm": 15.23256368530721, "learning_rate": 1.7966311866504552e-05, "loss": 0.7467, "step": 4426 }, { "epoch": 0.69, "grad_norm": 14.19721673234662, "learning_rate": 1.796529238293474e-05, "loss": 0.8107, "step": 4427 }, { "epoch": 0.69, "grad_norm": 27.03278740716948, "learning_rate": 1.7964272672836675e-05, "loss": 0.8674, "step": 4428 }, { "epoch": 0.69, "grad_norm": 21.983059411990997, "learning_rate": 1.7963252736239347e-05, "loss": 0.7714, "step": 4429 }, { "epoch": 0.69, "grad_norm": 18.646595055233774, "learning_rate": 1.7962232573171772e-05, "loss": 0.8475, "step": 4430 }, { "epoch": 0.69, "grad_norm": 16.200130679363628, "learning_rate": 1.7961212183662967e-05, "loss": 0.7553, "step": 4431 }, { "epoch": 0.69, "grad_norm": 19.025344381774524, "learning_rate": 1.796019156774194e-05, "loss": 0.8178, "step": 4432 }, { "epoch": 0.69, "grad_norm": 247.01706348116528, "learning_rate": 1.795917072543772e-05, "loss": 0.7495, "step": 4433 }, { "epoch": 0.69, "grad_norm": 11.127014079637586, "learning_rate": 1.7958149656779343e-05, "loss": 0.7655, "step": 4434 }, { "epoch": 0.69, "grad_norm": 15.928762044211297, "learning_rate": 1.7957128361795844e-05, "loss": 0.7268, "step": 4435 }, { "epoch": 0.69, "grad_norm": 21.02890202621121, "learning_rate": 1.7956106840516268e-05, "loss": 0.8222, "step": 4436 }, { "epoch": 0.69, "grad_norm": 11.901316062245872, "learning_rate": 1.7955085092969668e-05, "loss": 0.7505, "step": 4437 }, { "epoch": 0.69, "grad_norm": 17.681103566629638, "learning_rate": 1.7954063119185105e-05, "loss": 0.7627, "step": 4438 }, { "epoch": 0.69, "grad_norm": 20.414639107509593, "learning_rate": 1.7953040919191635e-05, "loss": 0.8111, "step": 4439 }, { "epoch": 0.69, "grad_norm": 15.012859023733897, "learning_rate": 1.7952018493018335e-05, "loss": 0.7608, "step": 4440 }, { "epoch": 0.69, "grad_norm": 18.005937312750195, "learning_rate": 1.7950995840694282e-05, "loss": 0.8028, "step": 4441 }, { "epoch": 0.69, "grad_norm": 14.745034940992863, "learning_rate": 1.794997296224856e-05, "loss": 0.6908, "step": 4442 }, { "epoch": 0.69, "grad_norm": 15.67019884439328, "learning_rate": 1.7948949857710256e-05, "loss": 0.8683, "step": 4443 }, { "epoch": 0.69, "grad_norm": 28.939592474780678, "learning_rate": 1.7947926527108467e-05, "loss": 0.8034, "step": 4444 }, { "epoch": 0.69, "grad_norm": 25.855332846946425, "learning_rate": 1.7946902970472297e-05, "loss": 0.7717, "step": 4445 }, { "epoch": 0.69, "grad_norm": 13.151155115022215, "learning_rate": 1.7945879187830857e-05, "loss": 0.7681, "step": 4446 }, { "epoch": 0.69, "grad_norm": 17.3831178068288, "learning_rate": 1.794485517921326e-05, "loss": 0.7679, "step": 4447 }, { "epoch": 0.69, "grad_norm": 16.126139486420012, "learning_rate": 1.794383094464863e-05, "loss": 0.7383, "step": 4448 }, { "epoch": 0.69, "grad_norm": 18.3920308420861, "learning_rate": 1.7942806484166096e-05, "loss": 0.7641, "step": 4449 }, { "epoch": 0.7, "grad_norm": 20.317655831876813, "learning_rate": 1.794178179779479e-05, "loss": 0.8207, "step": 4450 }, { "epoch": 0.7, "grad_norm": 18.12111928353769, "learning_rate": 1.7940756885563855e-05, "loss": 0.6349, "step": 4451 }, { "epoch": 0.7, "grad_norm": 21.89606140772132, "learning_rate": 1.793973174750244e-05, "loss": 0.8081, "step": 4452 }, { "epoch": 0.7, "grad_norm": 14.8642243536, "learning_rate": 1.79387063836397e-05, "loss": 0.6497, "step": 4453 }, { "epoch": 0.7, "grad_norm": 17.988074198969503, "learning_rate": 1.7937680794004793e-05, "loss": 0.8457, "step": 4454 }, { "epoch": 0.7, "grad_norm": 22.446891520665506, "learning_rate": 1.793665497862689e-05, "loss": 0.7774, "step": 4455 }, { "epoch": 0.7, "grad_norm": 14.243367416713927, "learning_rate": 1.793562893753516e-05, "loss": 0.8724, "step": 4456 }, { "epoch": 0.7, "grad_norm": 14.285394610783442, "learning_rate": 1.7934602670758786e-05, "loss": 0.7699, "step": 4457 }, { "epoch": 0.7, "grad_norm": 21.909584013113722, "learning_rate": 1.7933576178326952e-05, "loss": 0.7427, "step": 4458 }, { "epoch": 0.7, "grad_norm": 24.405535239329332, "learning_rate": 1.7932549460268856e-05, "loss": 0.8373, "step": 4459 }, { "epoch": 0.7, "grad_norm": 15.428939132891044, "learning_rate": 1.7931522516613688e-05, "loss": 0.6933, "step": 4460 }, { "epoch": 0.7, "grad_norm": 21.68097958812621, "learning_rate": 1.7930495347390665e-05, "loss": 0.8012, "step": 4461 }, { "epoch": 0.7, "grad_norm": 20.95167454486104, "learning_rate": 1.792946795262899e-05, "loss": 0.756, "step": 4462 }, { "epoch": 0.7, "grad_norm": 25.440243898720556, "learning_rate": 1.7928440332357885e-05, "loss": 0.8529, "step": 4463 }, { "epoch": 0.7, "grad_norm": 15.31729846039767, "learning_rate": 1.7927412486606578e-05, "loss": 0.7239, "step": 4464 }, { "epoch": 0.7, "grad_norm": 15.547319519934067, "learning_rate": 1.7926384415404298e-05, "loss": 0.7182, "step": 4465 }, { "epoch": 0.7, "grad_norm": 19.749350629917217, "learning_rate": 1.792535611878028e-05, "loss": 0.8117, "step": 4466 }, { "epoch": 0.7, "grad_norm": 19.079585312386996, "learning_rate": 1.792432759676377e-05, "loss": 0.6649, "step": 4467 }, { "epoch": 0.7, "grad_norm": 26.710050192703306, "learning_rate": 1.792329884938402e-05, "loss": 0.785, "step": 4468 }, { "epoch": 0.7, "grad_norm": 12.633466685949253, "learning_rate": 1.7922269876670282e-05, "loss": 0.8311, "step": 4469 }, { "epoch": 0.7, "grad_norm": 18.615603397575217, "learning_rate": 1.7921240678651823e-05, "loss": 0.7499, "step": 4470 }, { "epoch": 0.7, "grad_norm": 14.947341529955526, "learning_rate": 1.7920211255357918e-05, "loss": 0.7379, "step": 4471 }, { "epoch": 0.7, "grad_norm": 14.11531134437023, "learning_rate": 1.7919181606817836e-05, "loss": 0.821, "step": 4472 }, { "epoch": 0.7, "grad_norm": 27.957349265596267, "learning_rate": 1.791815173306086e-05, "loss": 0.8164, "step": 4473 }, { "epoch": 0.7, "grad_norm": 18.651135891057883, "learning_rate": 1.7917121634116284e-05, "loss": 0.8202, "step": 4474 }, { "epoch": 0.7, "grad_norm": 18.631555885950647, "learning_rate": 1.7916091310013396e-05, "loss": 0.8001, "step": 4475 }, { "epoch": 0.7, "grad_norm": 20.496797906155138, "learning_rate": 1.7915060760781505e-05, "loss": 0.7961, "step": 4476 }, { "epoch": 0.7, "grad_norm": 17.339943354218075, "learning_rate": 1.7914029986449916e-05, "loss": 0.7516, "step": 4477 }, { "epoch": 0.7, "grad_norm": 18.98117911121771, "learning_rate": 1.791299898704794e-05, "loss": 0.7873, "step": 4478 }, { "epoch": 0.7, "grad_norm": 20.449011366411074, "learning_rate": 1.7911967762604905e-05, "loss": 0.7652, "step": 4479 }, { "epoch": 0.7, "grad_norm": 19.8835194482637, "learning_rate": 1.7910936313150137e-05, "loss": 0.8287, "step": 4480 }, { "epoch": 0.7, "grad_norm": 16.320333878908084, "learning_rate": 1.7909904638712963e-05, "loss": 0.7859, "step": 4481 }, { "epoch": 0.7, "grad_norm": 16.95907926110049, "learning_rate": 1.790887273932273e-05, "loss": 0.8135, "step": 4482 }, { "epoch": 0.7, "grad_norm": 18.83241715398812, "learning_rate": 1.7907840615008787e-05, "loss": 0.7302, "step": 4483 }, { "epoch": 0.7, "grad_norm": 14.654744967972178, "learning_rate": 1.790680826580048e-05, "loss": 0.7286, "step": 4484 }, { "epoch": 0.7, "grad_norm": 17.764342193790647, "learning_rate": 1.7905775691727168e-05, "loss": 0.8639, "step": 4485 }, { "epoch": 0.7, "grad_norm": 18.18982506913134, "learning_rate": 1.7904742892818225e-05, "loss": 0.8222, "step": 4486 }, { "epoch": 0.7, "grad_norm": 20.311592257830757, "learning_rate": 1.7903709869103018e-05, "loss": 0.7469, "step": 4487 }, { "epoch": 0.7, "grad_norm": 17.281143091571263, "learning_rate": 1.7902676620610925e-05, "loss": 0.7178, "step": 4488 }, { "epoch": 0.7, "grad_norm": 20.40254109882613, "learning_rate": 1.7901643147371328e-05, "loss": 0.8199, "step": 4489 }, { "epoch": 0.7, "grad_norm": 17.87214955697475, "learning_rate": 1.7900609449413626e-05, "loss": 0.8005, "step": 4490 }, { "epoch": 0.7, "grad_norm": 17.232942132534024, "learning_rate": 1.7899575526767214e-05, "loss": 0.7777, "step": 4491 }, { "epoch": 0.7, "grad_norm": 16.33288877982655, "learning_rate": 1.7898541379461494e-05, "loss": 0.7689, "step": 4492 }, { "epoch": 0.7, "grad_norm": 21.66610174799004, "learning_rate": 1.7897507007525873e-05, "loss": 0.7062, "step": 4493 }, { "epoch": 0.7, "grad_norm": 15.251050478095443, "learning_rate": 1.789647241098978e-05, "loss": 0.8165, "step": 4494 }, { "epoch": 0.7, "grad_norm": 23.139680554668924, "learning_rate": 1.7895437589882627e-05, "loss": 0.814, "step": 4495 }, { "epoch": 0.7, "grad_norm": 19.387941302325746, "learning_rate": 1.7894402544233846e-05, "loss": 0.8096, "step": 4496 }, { "epoch": 0.7, "grad_norm": 21.50296622152407, "learning_rate": 1.789336727407288e-05, "loss": 0.7492, "step": 4497 }, { "epoch": 0.7, "grad_norm": 17.119100092542315, "learning_rate": 1.7892331779429164e-05, "loss": 0.6993, "step": 4498 }, { "epoch": 0.7, "grad_norm": 14.749492042491251, "learning_rate": 1.789129606033215e-05, "loss": 0.7714, "step": 4499 }, { "epoch": 0.7, "grad_norm": 28.973994530305404, "learning_rate": 1.789026011681129e-05, "loss": 0.8197, "step": 4500 }, { "epoch": 0.7, "grad_norm": 28.70944916662741, "learning_rate": 1.7889223948896046e-05, "loss": 0.7195, "step": 4501 }, { "epoch": 0.7, "grad_norm": 22.64568218703805, "learning_rate": 1.788818755661589e-05, "loss": 0.7309, "step": 4502 }, { "epoch": 0.7, "grad_norm": 15.431507958326039, "learning_rate": 1.7887150940000296e-05, "loss": 0.722, "step": 4503 }, { "epoch": 0.7, "grad_norm": 12.803628546058574, "learning_rate": 1.788611409907874e-05, "loss": 0.7555, "step": 4504 }, { "epoch": 0.7, "grad_norm": 15.174416813826769, "learning_rate": 1.7885077033880714e-05, "loss": 0.6677, "step": 4505 }, { "epoch": 0.7, "grad_norm": 18.519255028776655, "learning_rate": 1.788403974443571e-05, "loss": 0.7111, "step": 4506 }, { "epoch": 0.7, "grad_norm": 18.945335141197912, "learning_rate": 1.788300223077323e-05, "loss": 0.7666, "step": 4507 }, { "epoch": 0.7, "grad_norm": 30.26109322944965, "learning_rate": 1.7881964492922774e-05, "loss": 0.7937, "step": 4508 }, { "epoch": 0.7, "grad_norm": 20.18380947710191, "learning_rate": 1.7880926530913863e-05, "loss": 0.7532, "step": 4509 }, { "epoch": 0.7, "grad_norm": 13.311576581211785, "learning_rate": 1.7879888344776008e-05, "loss": 0.7534, "step": 4510 }, { "epoch": 0.7, "grad_norm": 12.425904345629691, "learning_rate": 1.787884993453874e-05, "loss": 0.7623, "step": 4511 }, { "epoch": 0.7, "grad_norm": 22.30198697712689, "learning_rate": 1.7877811300231587e-05, "loss": 0.7699, "step": 4512 }, { "epoch": 0.7, "grad_norm": 20.377571393475165, "learning_rate": 1.7876772441884093e-05, "loss": 0.7981, "step": 4513 }, { "epoch": 0.71, "grad_norm": 16.58466978839017, "learning_rate": 1.7875733359525793e-05, "loss": 0.8134, "step": 4514 }, { "epoch": 0.71, "grad_norm": 26.9714839462326, "learning_rate": 1.7874694053186246e-05, "loss": 0.8238, "step": 4515 }, { "epoch": 0.71, "grad_norm": 17.728531718417514, "learning_rate": 1.7873654522895006e-05, "loss": 0.8276, "step": 4516 }, { "epoch": 0.71, "grad_norm": 17.510589664470007, "learning_rate": 1.7872614768681638e-05, "loss": 0.8108, "step": 4517 }, { "epoch": 0.71, "grad_norm": 20.24816512107083, "learning_rate": 1.7871574790575713e-05, "loss": 0.8559, "step": 4518 }, { "epoch": 0.71, "grad_norm": 24.814463942464236, "learning_rate": 1.7870534588606804e-05, "loss": 0.8228, "step": 4519 }, { "epoch": 0.71, "grad_norm": 21.126995740101478, "learning_rate": 1.7869494162804492e-05, "loss": 0.7478, "step": 4520 }, { "epoch": 0.71, "grad_norm": 21.225116719811254, "learning_rate": 1.7868453513198375e-05, "loss": 0.7609, "step": 4521 }, { "epoch": 0.71, "grad_norm": 22.06792835085101, "learning_rate": 1.786741263981804e-05, "loss": 0.724, "step": 4522 }, { "epoch": 0.71, "grad_norm": 25.883283735753, "learning_rate": 1.7866371542693094e-05, "loss": 0.7266, "step": 4523 }, { "epoch": 0.71, "grad_norm": 21.20164873034085, "learning_rate": 1.786533022185314e-05, "loss": 0.7889, "step": 4524 }, { "epoch": 0.71, "grad_norm": 30.722225829817987, "learning_rate": 1.78642886773278e-05, "loss": 0.8972, "step": 4525 }, { "epoch": 0.71, "grad_norm": 15.33453424711936, "learning_rate": 1.7863246909146688e-05, "loss": 0.7599, "step": 4526 }, { "epoch": 0.71, "grad_norm": 14.690056751794366, "learning_rate": 1.786220491733943e-05, "loss": 0.7537, "step": 4527 }, { "epoch": 0.71, "grad_norm": 17.930520901810457, "learning_rate": 1.7861162701935672e-05, "loss": 0.7096, "step": 4528 }, { "epoch": 0.71, "grad_norm": 19.71866787237424, "learning_rate": 1.7860120262965038e-05, "loss": 0.7578, "step": 4529 }, { "epoch": 0.71, "grad_norm": 27.19743173764247, "learning_rate": 1.7859077600457184e-05, "loss": 0.7474, "step": 4530 }, { "epoch": 0.71, "grad_norm": 14.763969372970255, "learning_rate": 1.7858034714441764e-05, "loss": 0.7404, "step": 4531 }, { "epoch": 0.71, "grad_norm": 21.289749176430142, "learning_rate": 1.7856991604948428e-05, "loss": 0.689, "step": 4532 }, { "epoch": 0.71, "grad_norm": 21.31819391225716, "learning_rate": 1.7855948272006848e-05, "loss": 0.7654, "step": 4533 }, { "epoch": 0.71, "grad_norm": 18.21917595537159, "learning_rate": 1.7854904715646697e-05, "loss": 0.7669, "step": 4534 }, { "epoch": 0.71, "grad_norm": 18.18341888851737, "learning_rate": 1.785386093589765e-05, "loss": 0.7457, "step": 4535 }, { "epoch": 0.71, "grad_norm": 22.213289459037025, "learning_rate": 1.7852816932789392e-05, "loss": 0.85, "step": 4536 }, { "epoch": 0.71, "grad_norm": 14.041481135578346, "learning_rate": 1.7851772706351614e-05, "loss": 0.7068, "step": 4537 }, { "epoch": 0.71, "grad_norm": 20.176504845582247, "learning_rate": 1.785072825661401e-05, "loss": 0.7969, "step": 4538 }, { "epoch": 0.71, "grad_norm": 15.287263358936663, "learning_rate": 1.7849683583606293e-05, "loss": 0.7074, "step": 4539 }, { "epoch": 0.71, "grad_norm": 12.276838393556602, "learning_rate": 1.7848638687358164e-05, "loss": 0.6571, "step": 4540 }, { "epoch": 0.71, "grad_norm": 18.05278429987945, "learning_rate": 1.7847593567899344e-05, "loss": 0.7659, "step": 4541 }, { "epoch": 0.71, "grad_norm": 15.298943066125636, "learning_rate": 1.7846548225259553e-05, "loss": 0.7017, "step": 4542 }, { "epoch": 0.71, "grad_norm": 13.871929997530419, "learning_rate": 1.7845502659468522e-05, "loss": 0.7217, "step": 4543 }, { "epoch": 0.71, "grad_norm": 11.840309628514953, "learning_rate": 1.784445687055598e-05, "loss": 0.7104, "step": 4544 }, { "epoch": 0.71, "grad_norm": 16.54040551532769, "learning_rate": 1.7843410858551674e-05, "loss": 0.6895, "step": 4545 }, { "epoch": 0.71, "grad_norm": 23.885929705359707, "learning_rate": 1.7842364623485356e-05, "loss": 0.8651, "step": 4546 }, { "epoch": 0.71, "grad_norm": 19.742020265499644, "learning_rate": 1.784131816538677e-05, "loss": 0.8096, "step": 4547 }, { "epoch": 0.71, "grad_norm": 18.32213060918708, "learning_rate": 1.7840271484285687e-05, "loss": 0.7455, "step": 4548 }, { "epoch": 0.71, "grad_norm": 16.294218046096653, "learning_rate": 1.783922458021187e-05, "loss": 0.7091, "step": 4549 }, { "epoch": 0.71, "grad_norm": 12.886128865663519, "learning_rate": 1.783817745319509e-05, "loss": 0.693, "step": 4550 }, { "epoch": 0.71, "grad_norm": 16.304773134884957, "learning_rate": 1.783713010326513e-05, "loss": 0.7035, "step": 4551 }, { "epoch": 0.71, "grad_norm": 26.510328769850396, "learning_rate": 1.7836082530451772e-05, "loss": 0.8074, "step": 4552 }, { "epoch": 0.71, "grad_norm": 12.01065139370198, "learning_rate": 1.783503473478481e-05, "loss": 0.6972, "step": 4553 }, { "epoch": 0.71, "grad_norm": 19.235472748525677, "learning_rate": 1.7833986716294046e-05, "loss": 0.7201, "step": 4554 }, { "epoch": 0.71, "grad_norm": 14.144588477717663, "learning_rate": 1.783293847500928e-05, "loss": 0.7781, "step": 4555 }, { "epoch": 0.71, "grad_norm": 14.747317467243905, "learning_rate": 1.783189001096033e-05, "loss": 0.6988, "step": 4556 }, { "epoch": 0.71, "grad_norm": 13.169701209631764, "learning_rate": 1.7830841324177006e-05, "loss": 0.7843, "step": 4557 }, { "epoch": 0.71, "grad_norm": 17.938530348425555, "learning_rate": 1.7829792414689135e-05, "loss": 0.7218, "step": 4558 }, { "epoch": 0.71, "grad_norm": 13.487101280375038, "learning_rate": 1.782874328252655e-05, "loss": 0.7515, "step": 4559 }, { "epoch": 0.71, "grad_norm": 27.69630725295481, "learning_rate": 1.7827693927719086e-05, "loss": 0.8335, "step": 4560 }, { "epoch": 0.71, "grad_norm": 13.951567465727559, "learning_rate": 1.7826644350296588e-05, "loss": 0.7452, "step": 4561 }, { "epoch": 0.71, "grad_norm": 23.347824798761902, "learning_rate": 1.7825594550288898e-05, "loss": 0.7723, "step": 4562 }, { "epoch": 0.71, "grad_norm": 22.25889149156369, "learning_rate": 1.782454452772588e-05, "loss": 0.719, "step": 4563 }, { "epoch": 0.71, "grad_norm": 17.542686348007297, "learning_rate": 1.782349428263739e-05, "loss": 0.7425, "step": 4564 }, { "epoch": 0.71, "grad_norm": 17.523572289632572, "learning_rate": 1.7822443815053305e-05, "loss": 0.848, "step": 4565 }, { "epoch": 0.71, "grad_norm": 19.67062176082445, "learning_rate": 1.782139312500349e-05, "loss": 0.7283, "step": 4566 }, { "epoch": 0.71, "grad_norm": 21.117571666016065, "learning_rate": 1.7820342212517826e-05, "loss": 0.714, "step": 4567 }, { "epoch": 0.71, "grad_norm": 18.19939445039066, "learning_rate": 1.7819291077626212e-05, "loss": 0.8413, "step": 4568 }, { "epoch": 0.71, "grad_norm": 15.50637241374402, "learning_rate": 1.781823972035853e-05, "loss": 0.8397, "step": 4569 }, { "epoch": 0.71, "grad_norm": 13.80469806867194, "learning_rate": 1.7817188140744682e-05, "loss": 0.7116, "step": 4570 }, { "epoch": 0.71, "grad_norm": 23.981495182970207, "learning_rate": 1.7816136338814577e-05, "loss": 0.7207, "step": 4571 }, { "epoch": 0.71, "grad_norm": 15.554049720698364, "learning_rate": 1.781508431459813e-05, "loss": 0.7284, "step": 4572 }, { "epoch": 0.71, "grad_norm": 22.198146139689523, "learning_rate": 1.7814032068125254e-05, "loss": 0.8335, "step": 4573 }, { "epoch": 0.71, "grad_norm": 13.120963710368788, "learning_rate": 1.781297959942588e-05, "loss": 0.7382, "step": 4574 }, { "epoch": 0.71, "grad_norm": 18.32251608311107, "learning_rate": 1.7811926908529934e-05, "loss": 0.7331, "step": 4575 }, { "epoch": 0.71, "grad_norm": 22.51043225540429, "learning_rate": 1.781087399546736e-05, "loss": 0.7078, "step": 4576 }, { "epoch": 0.71, "grad_norm": 16.47989384276147, "learning_rate": 1.7809820860268096e-05, "loss": 0.7913, "step": 4577 }, { "epoch": 0.72, "grad_norm": 15.634376014463157, "learning_rate": 1.7808767502962094e-05, "loss": 0.72, "step": 4578 }, { "epoch": 0.72, "grad_norm": 22.849754926480593, "learning_rate": 1.7807713923579313e-05, "loss": 0.6631, "step": 4579 }, { "epoch": 0.72, "grad_norm": 20.006859729488482, "learning_rate": 1.7806660122149714e-05, "loss": 0.7598, "step": 4580 }, { "epoch": 0.72, "grad_norm": 14.412321651336214, "learning_rate": 1.7805606098703273e-05, "loss": 0.699, "step": 4581 }, { "epoch": 0.72, "grad_norm": 24.106188192033642, "learning_rate": 1.7804551853269957e-05, "loss": 0.7818, "step": 4582 }, { "epoch": 0.72, "grad_norm": 24.032803466648353, "learning_rate": 1.7803497385879752e-05, "loss": 0.8391, "step": 4583 }, { "epoch": 0.72, "grad_norm": 18.95640167772439, "learning_rate": 1.7802442696562646e-05, "loss": 0.6838, "step": 4584 }, { "epoch": 0.72, "grad_norm": 17.11034547457215, "learning_rate": 1.7801387785348637e-05, "loss": 0.7524, "step": 4585 }, { "epoch": 0.72, "grad_norm": 14.478175168142961, "learning_rate": 1.780033265226772e-05, "loss": 0.6551, "step": 4586 }, { "epoch": 0.72, "grad_norm": 13.932257244091351, "learning_rate": 1.7799277297349908e-05, "loss": 0.6978, "step": 4587 }, { "epoch": 0.72, "grad_norm": 18.86384159850776, "learning_rate": 1.7798221720625205e-05, "loss": 0.6796, "step": 4588 }, { "epoch": 0.72, "grad_norm": 14.21155762223803, "learning_rate": 1.779716592212365e-05, "loss": 0.7782, "step": 4589 }, { "epoch": 0.72, "grad_norm": 18.086023444183674, "learning_rate": 1.7796109901875246e-05, "loss": 0.7912, "step": 4590 }, { "epoch": 0.72, "grad_norm": 16.905189667765715, "learning_rate": 1.7795053659910044e-05, "loss": 0.822, "step": 4591 }, { "epoch": 0.72, "grad_norm": 17.280428154633523, "learning_rate": 1.7793997196258074e-05, "loss": 0.8209, "step": 4592 }, { "epoch": 0.72, "grad_norm": 14.618783952315601, "learning_rate": 1.7792940510949383e-05, "loss": 0.7516, "step": 4593 }, { "epoch": 0.72, "grad_norm": 26.049443477568158, "learning_rate": 1.7791883604014023e-05, "loss": 0.7825, "step": 4594 }, { "epoch": 0.72, "grad_norm": 24.4910808696523, "learning_rate": 1.7790826475482047e-05, "loss": 0.812, "step": 4595 }, { "epoch": 0.72, "grad_norm": 20.17971809472022, "learning_rate": 1.7789769125383528e-05, "loss": 0.7576, "step": 4596 }, { "epoch": 0.72, "grad_norm": 14.355798234337353, "learning_rate": 1.7788711553748533e-05, "loss": 0.7372, "step": 4597 }, { "epoch": 0.72, "grad_norm": 17.416064380501787, "learning_rate": 1.7787653760607134e-05, "loss": 0.7403, "step": 4598 }, { "epoch": 0.72, "grad_norm": 27.892054013099575, "learning_rate": 1.778659574598942e-05, "loss": 0.7828, "step": 4599 }, { "epoch": 0.72, "grad_norm": 18.44548862474979, "learning_rate": 1.7785537509925478e-05, "loss": 0.721, "step": 4600 }, { "epoch": 0.72, "grad_norm": 11.884567290745816, "learning_rate": 1.77844790524454e-05, "loss": 0.7116, "step": 4601 }, { "epoch": 0.72, "grad_norm": 30.758239248976416, "learning_rate": 1.77834203735793e-05, "loss": 0.7893, "step": 4602 }, { "epoch": 0.72, "grad_norm": 28.94109036209768, "learning_rate": 1.778236147335727e-05, "loss": 0.6847, "step": 4603 }, { "epoch": 0.72, "grad_norm": 12.537867110602596, "learning_rate": 1.7781302351809437e-05, "loss": 0.691, "step": 4604 }, { "epoch": 0.72, "grad_norm": 17.825930645070017, "learning_rate": 1.7780243008965915e-05, "loss": 0.652, "step": 4605 }, { "epoch": 0.72, "grad_norm": 15.45764941718774, "learning_rate": 1.7779183444856833e-05, "loss": 0.7485, "step": 4606 }, { "epoch": 0.72, "grad_norm": 26.59329265824312, "learning_rate": 1.7778123659512326e-05, "loss": 0.7516, "step": 4607 }, { "epoch": 0.72, "grad_norm": 17.52337750132507, "learning_rate": 1.777706365296253e-05, "loss": 0.6939, "step": 4608 }, { "epoch": 0.72, "grad_norm": 18.87398120162674, "learning_rate": 1.7776003425237592e-05, "loss": 0.6667, "step": 4609 }, { "epoch": 0.72, "grad_norm": 15.049088492545417, "learning_rate": 1.7774942976367668e-05, "loss": 0.7756, "step": 4610 }, { "epoch": 0.72, "grad_norm": 18.44443025623951, "learning_rate": 1.7773882306382913e-05, "loss": 0.6874, "step": 4611 }, { "epoch": 0.72, "grad_norm": 24.064901833986912, "learning_rate": 1.7772821415313493e-05, "loss": 0.7053, "step": 4612 }, { "epoch": 0.72, "grad_norm": 15.36597587180189, "learning_rate": 1.777176030318958e-05, "loss": 0.7054, "step": 4613 }, { "epoch": 0.72, "grad_norm": 17.365403532037313, "learning_rate": 1.777069897004135e-05, "loss": 0.7461, "step": 4614 }, { "epoch": 0.72, "grad_norm": 24.22883278371235, "learning_rate": 1.7769637415898982e-05, "loss": 0.6973, "step": 4615 }, { "epoch": 0.72, "grad_norm": 9.706301176988791, "learning_rate": 1.776857564079268e-05, "loss": 0.7002, "step": 4616 }, { "epoch": 0.72, "grad_norm": 29.914397830832083, "learning_rate": 1.7767513644752624e-05, "loss": 0.8495, "step": 4617 }, { "epoch": 0.72, "grad_norm": 16.378578762289482, "learning_rate": 1.7766451427809026e-05, "loss": 0.6789, "step": 4618 }, { "epoch": 0.72, "grad_norm": 20.225273640111205, "learning_rate": 1.7765388989992093e-05, "loss": 0.7316, "step": 4619 }, { "epoch": 0.72, "grad_norm": 23.14733087145608, "learning_rate": 1.776432633133204e-05, "loss": 0.7645, "step": 4620 }, { "epoch": 0.72, "grad_norm": 16.36780120500211, "learning_rate": 1.776326345185909e-05, "loss": 0.7196, "step": 4621 }, { "epoch": 0.72, "grad_norm": 18.967696973710073, "learning_rate": 1.7762200351603465e-05, "loss": 0.7228, "step": 4622 }, { "epoch": 0.72, "grad_norm": 17.57145546333111, "learning_rate": 1.7761137030595406e-05, "loss": 0.7221, "step": 4623 }, { "epoch": 0.72, "grad_norm": 16.224495556378056, "learning_rate": 1.776007348886515e-05, "loss": 0.7173, "step": 4624 }, { "epoch": 0.72, "grad_norm": 13.778919065555696, "learning_rate": 1.775900972644294e-05, "loss": 0.7658, "step": 4625 }, { "epoch": 0.72, "grad_norm": 14.968809065971032, "learning_rate": 1.7757945743359033e-05, "loss": 0.7949, "step": 4626 }, { "epoch": 0.72, "grad_norm": 20.726221838120367, "learning_rate": 1.775688153964369e-05, "loss": 0.8367, "step": 4627 }, { "epoch": 0.72, "grad_norm": 16.838458730080607, "learning_rate": 1.775581711532717e-05, "loss": 0.7264, "step": 4628 }, { "epoch": 0.72, "grad_norm": 17.315085156088408, "learning_rate": 1.7754752470439755e-05, "loss": 0.752, "step": 4629 }, { "epoch": 0.72, "grad_norm": 12.852962364319751, "learning_rate": 1.7753687605011707e-05, "loss": 0.763, "step": 4630 }, { "epoch": 0.72, "grad_norm": 15.063412276120008, "learning_rate": 1.7752622519073327e-05, "loss": 0.7749, "step": 4631 }, { "epoch": 0.72, "grad_norm": 17.92250672784313, "learning_rate": 1.7751557212654893e-05, "loss": 0.7456, "step": 4632 }, { "epoch": 0.72, "grad_norm": 17.54124495642337, "learning_rate": 1.7750491685786708e-05, "loss": 0.8077, "step": 4633 }, { "epoch": 0.72, "grad_norm": 15.289543972708875, "learning_rate": 1.7749425938499073e-05, "loss": 0.7602, "step": 4634 }, { "epoch": 0.72, "grad_norm": 30.640675451080558, "learning_rate": 1.77483599708223e-05, "loss": 0.7679, "step": 4635 }, { "epoch": 0.72, "grad_norm": 18.525962541647345, "learning_rate": 1.7747293782786697e-05, "loss": 0.8328, "step": 4636 }, { "epoch": 0.72, "grad_norm": 14.983633664461529, "learning_rate": 1.7746227374422597e-05, "loss": 0.7545, "step": 4637 }, { "epoch": 0.72, "grad_norm": 29.179535389699073, "learning_rate": 1.7745160745760317e-05, "loss": 0.7919, "step": 4638 }, { "epoch": 0.72, "grad_norm": 21.79949564065988, "learning_rate": 1.7744093896830196e-05, "loss": 0.7646, "step": 4639 }, { "epoch": 0.72, "grad_norm": 26.00539878640754, "learning_rate": 1.7743026827662578e-05, "loss": 0.7665, "step": 4640 }, { "epoch": 0.72, "grad_norm": 20.839287454604097, "learning_rate": 1.7741959538287807e-05, "loss": 0.7677, "step": 4641 }, { "epoch": 0.73, "grad_norm": 24.868581926904834, "learning_rate": 1.7740892028736233e-05, "loss": 0.7512, "step": 4642 }, { "epoch": 0.73, "grad_norm": 21.953774704601337, "learning_rate": 1.7739824299038217e-05, "loss": 0.8183, "step": 4643 }, { "epoch": 0.73, "grad_norm": 13.993655319477101, "learning_rate": 1.773875634922413e-05, "loss": 0.7837, "step": 4644 }, { "epoch": 0.73, "grad_norm": 18.672320268022368, "learning_rate": 1.7737688179324334e-05, "loss": 0.789, "step": 4645 }, { "epoch": 0.73, "grad_norm": 17.89504836425524, "learning_rate": 1.773661978936922e-05, "loss": 0.7179, "step": 4646 }, { "epoch": 0.73, "grad_norm": 18.98533202926206, "learning_rate": 1.7735551179389157e-05, "loss": 0.7097, "step": 4647 }, { "epoch": 0.73, "grad_norm": 18.47458291286288, "learning_rate": 1.7734482349414547e-05, "loss": 0.7666, "step": 4648 }, { "epoch": 0.73, "grad_norm": 18.43875464033775, "learning_rate": 1.7733413299475787e-05, "loss": 0.8031, "step": 4649 }, { "epoch": 0.73, "grad_norm": 17.87916699853392, "learning_rate": 1.7732344029603273e-05, "loss": 0.7293, "step": 4650 }, { "epoch": 0.73, "grad_norm": 19.626097062298985, "learning_rate": 1.773127453982742e-05, "loss": 0.9176, "step": 4651 }, { "epoch": 0.73, "grad_norm": 19.283764035368232, "learning_rate": 1.7730204830178638e-05, "loss": 0.7812, "step": 4652 }, { "epoch": 0.73, "grad_norm": 21.78481579672387, "learning_rate": 1.7729134900687354e-05, "loss": 0.7556, "step": 4653 }, { "epoch": 0.73, "grad_norm": 25.059231830123245, "learning_rate": 1.7728064751383997e-05, "loss": 0.784, "step": 4654 }, { "epoch": 0.73, "grad_norm": 22.44174358356745, "learning_rate": 1.7726994382299e-05, "loss": 0.758, "step": 4655 }, { "epoch": 0.73, "grad_norm": 24.80834578322753, "learning_rate": 1.7725923793462798e-05, "loss": 0.8446, "step": 4656 }, { "epoch": 0.73, "grad_norm": 14.215338581736098, "learning_rate": 1.7724852984905844e-05, "loss": 0.6924, "step": 4657 }, { "epoch": 0.73, "grad_norm": 13.856078966059064, "learning_rate": 1.7723781956658593e-05, "loss": 0.7276, "step": 4658 }, { "epoch": 0.73, "grad_norm": 21.507102864187885, "learning_rate": 1.77227107087515e-05, "loss": 0.7842, "step": 4659 }, { "epoch": 0.73, "grad_norm": 19.76527188781026, "learning_rate": 1.7721639241215034e-05, "loss": 0.7244, "step": 4660 }, { "epoch": 0.73, "grad_norm": 17.368469322252448, "learning_rate": 1.772056755407966e-05, "loss": 0.7384, "step": 4661 }, { "epoch": 0.73, "grad_norm": 22.35696192981688, "learning_rate": 1.771949564737587e-05, "loss": 0.769, "step": 4662 }, { "epoch": 0.73, "grad_norm": 13.98192081147761, "learning_rate": 1.771842352113413e-05, "loss": 0.7962, "step": 4663 }, { "epoch": 0.73, "grad_norm": 19.801141597518985, "learning_rate": 1.7717351175384948e-05, "loss": 0.7734, "step": 4664 }, { "epoch": 0.73, "grad_norm": 28.0424575461503, "learning_rate": 1.7716278610158806e-05, "loss": 0.8066, "step": 4665 }, { "epoch": 0.73, "grad_norm": 13.600696584361431, "learning_rate": 1.7715205825486218e-05, "loss": 0.7815, "step": 4666 }, { "epoch": 0.73, "grad_norm": 21.76127266345328, "learning_rate": 1.771413282139769e-05, "loss": 0.7115, "step": 4667 }, { "epoch": 0.73, "grad_norm": 24.362943337772816, "learning_rate": 1.7713059597923735e-05, "loss": 0.7848, "step": 4668 }, { "epoch": 0.73, "grad_norm": 21.439576638543528, "learning_rate": 1.771198615509488e-05, "loss": 0.756, "step": 4669 }, { "epoch": 0.73, "grad_norm": 14.119958743836833, "learning_rate": 1.771091249294165e-05, "loss": 0.6665, "step": 4670 }, { "epoch": 0.73, "grad_norm": 19.20309282801442, "learning_rate": 1.7709838611494577e-05, "loss": 0.7591, "step": 4671 }, { "epoch": 0.73, "grad_norm": 18.794715676459475, "learning_rate": 1.7708764510784205e-05, "loss": 0.725, "step": 4672 }, { "epoch": 0.73, "grad_norm": 18.402110985595833, "learning_rate": 1.770769019084108e-05, "loss": 0.6992, "step": 4673 }, { "epoch": 0.73, "grad_norm": 33.52548618675277, "learning_rate": 1.7706615651695754e-05, "loss": 0.8105, "step": 4674 }, { "epoch": 0.73, "grad_norm": 12.938989561018383, "learning_rate": 1.7705540893378787e-05, "loss": 0.7831, "step": 4675 }, { "epoch": 0.73, "grad_norm": 15.43314064498119, "learning_rate": 1.7704465915920745e-05, "loss": 0.815, "step": 4676 }, { "epoch": 0.73, "grad_norm": 20.685658191031653, "learning_rate": 1.77033907193522e-05, "loss": 0.7443, "step": 4677 }, { "epoch": 0.73, "grad_norm": 12.049858843156683, "learning_rate": 1.770231530370373e-05, "loss": 0.7456, "step": 4678 }, { "epoch": 0.73, "grad_norm": 15.708947774163132, "learning_rate": 1.7701239669005916e-05, "loss": 0.8426, "step": 4679 }, { "epoch": 0.73, "grad_norm": 19.484571068946565, "learning_rate": 1.7700163815289353e-05, "loss": 0.724, "step": 4680 }, { "epoch": 0.73, "grad_norm": 17.628196660609422, "learning_rate": 1.7699087742584636e-05, "loss": 0.7379, "step": 4681 }, { "epoch": 0.73, "grad_norm": 26.81608904812798, "learning_rate": 1.769801145092237e-05, "loss": 0.7381, "step": 4682 }, { "epoch": 0.73, "grad_norm": 19.802430812466564, "learning_rate": 1.7696934940333155e-05, "loss": 0.8898, "step": 4683 }, { "epoch": 0.73, "grad_norm": 27.12962424616862, "learning_rate": 1.7695858210847615e-05, "loss": 0.8014, "step": 4684 }, { "epoch": 0.73, "grad_norm": 18.1015214166352, "learning_rate": 1.769478126249637e-05, "loss": 0.7168, "step": 4685 }, { "epoch": 0.73, "grad_norm": 20.520628887616834, "learning_rate": 1.769370409531005e-05, "loss": 0.7749, "step": 4686 }, { "epoch": 0.73, "grad_norm": 20.73359294572111, "learning_rate": 1.7692626709319283e-05, "loss": 0.7404, "step": 4687 }, { "epoch": 0.73, "grad_norm": 26.08046606022863, "learning_rate": 1.769154910455471e-05, "loss": 0.7131, "step": 4688 }, { "epoch": 0.73, "grad_norm": 19.21366844840311, "learning_rate": 1.7690471281046982e-05, "loss": 0.7684, "step": 4689 }, { "epoch": 0.73, "grad_norm": 15.744214996514318, "learning_rate": 1.768939323882675e-05, "loss": 0.7194, "step": 4690 }, { "epoch": 0.73, "grad_norm": 24.950315766400433, "learning_rate": 1.7688314977924674e-05, "loss": 0.7763, "step": 4691 }, { "epoch": 0.73, "grad_norm": 21.612398469106186, "learning_rate": 1.768723649837142e-05, "loss": 0.744, "step": 4692 }, { "epoch": 0.73, "grad_norm": 12.773618786659869, "learning_rate": 1.7686157800197652e-05, "loss": 0.7515, "step": 4693 }, { "epoch": 0.73, "grad_norm": 21.05816044164395, "learning_rate": 1.7685078883434054e-05, "loss": 0.744, "step": 4694 }, { "epoch": 0.73, "grad_norm": 19.406283108288324, "learning_rate": 1.7683999748111306e-05, "loss": 0.831, "step": 4695 }, { "epoch": 0.73, "grad_norm": 18.617075782901715, "learning_rate": 1.7682920394260102e-05, "loss": 0.7737, "step": 4696 }, { "epoch": 0.73, "grad_norm": 24.916444636448617, "learning_rate": 1.768184082191114e-05, "loss": 0.7374, "step": 4697 }, { "epoch": 0.73, "grad_norm": 15.094554564256278, "learning_rate": 1.7680761031095113e-05, "loss": 0.6707, "step": 4698 }, { "epoch": 0.73, "grad_norm": 16.64583492298938, "learning_rate": 1.7679681021842738e-05, "loss": 0.7114, "step": 4699 }, { "epoch": 0.73, "grad_norm": 15.201698725408875, "learning_rate": 1.767860079418473e-05, "loss": 0.7649, "step": 4700 }, { "epoch": 0.73, "grad_norm": 27.219623840705353, "learning_rate": 1.7677520348151805e-05, "loss": 0.798, "step": 4701 }, { "epoch": 0.73, "grad_norm": 18.068731173647688, "learning_rate": 1.767643968377469e-05, "loss": 0.7095, "step": 4702 }, { "epoch": 0.73, "grad_norm": 17.347959692297604, "learning_rate": 1.7675358801084122e-05, "loss": 0.6746, "step": 4703 }, { "epoch": 0.73, "grad_norm": 20.983950511220705, "learning_rate": 1.767427770011084e-05, "loss": 0.7586, "step": 4704 }, { "epoch": 0.73, "grad_norm": 22.607776450884366, "learning_rate": 1.767319638088559e-05, "loss": 0.8404, "step": 4705 }, { "epoch": 0.74, "grad_norm": 18.93052476896135, "learning_rate": 1.7672114843439127e-05, "loss": 0.7683, "step": 4706 }, { "epoch": 0.74, "grad_norm": 25.206016053666044, "learning_rate": 1.76710330878022e-05, "loss": 0.8398, "step": 4707 }, { "epoch": 0.74, "grad_norm": 16.813816645625376, "learning_rate": 1.7669951114005584e-05, "loss": 0.7192, "step": 4708 }, { "epoch": 0.74, "grad_norm": 16.644285375681033, "learning_rate": 1.766886892208004e-05, "loss": 0.7189, "step": 4709 }, { "epoch": 0.74, "grad_norm": 18.0204235644049, "learning_rate": 1.7667786512056352e-05, "loss": 0.8104, "step": 4710 }, { "epoch": 0.74, "grad_norm": 16.506676078951678, "learning_rate": 1.7666703883965305e-05, "loss": 0.7505, "step": 4711 }, { "epoch": 0.74, "grad_norm": 19.526104483491093, "learning_rate": 1.766562103783768e-05, "loss": 0.8487, "step": 4712 }, { "epoch": 0.74, "grad_norm": 16.166371296326044, "learning_rate": 1.766453797370428e-05, "loss": 0.8902, "step": 4713 }, { "epoch": 0.74, "grad_norm": 16.098966153656235, "learning_rate": 1.76634546915959e-05, "loss": 0.8063, "step": 4714 }, { "epoch": 0.74, "grad_norm": 16.266027946684616, "learning_rate": 1.7662371191543355e-05, "loss": 0.7278, "step": 4715 }, { "epoch": 0.74, "grad_norm": 28.156781562832112, "learning_rate": 1.7661287473577457e-05, "loss": 0.7446, "step": 4716 }, { "epoch": 0.74, "grad_norm": 22.59327000389072, "learning_rate": 1.766020353772902e-05, "loss": 0.8598, "step": 4717 }, { "epoch": 0.74, "grad_norm": 22.87572390552177, "learning_rate": 1.765911938402888e-05, "loss": 0.7998, "step": 4718 }, { "epoch": 0.74, "grad_norm": 23.873782269145302, "learning_rate": 1.7658035012507863e-05, "loss": 0.757, "step": 4719 }, { "epoch": 0.74, "grad_norm": 20.234898673833182, "learning_rate": 1.7656950423196807e-05, "loss": 0.7828, "step": 4720 }, { "epoch": 0.74, "grad_norm": 18.95689105247914, "learning_rate": 1.7655865616126564e-05, "loss": 0.7172, "step": 4721 }, { "epoch": 0.74, "grad_norm": 19.574414710725613, "learning_rate": 1.7654780591327983e-05, "loss": 0.7864, "step": 4722 }, { "epoch": 0.74, "grad_norm": 14.966824203599673, "learning_rate": 1.7653695348831915e-05, "loss": 0.7277, "step": 4723 }, { "epoch": 0.74, "grad_norm": 21.66688520859272, "learning_rate": 1.7652609888669234e-05, "loss": 0.6598, "step": 4724 }, { "epoch": 0.74, "grad_norm": 19.200763538571433, "learning_rate": 1.76515242108708e-05, "loss": 0.8736, "step": 4725 }, { "epoch": 0.74, "grad_norm": 19.356492148095814, "learning_rate": 1.7650438315467494e-05, "loss": 0.7293, "step": 4726 }, { "epoch": 0.74, "grad_norm": 15.214323658311407, "learning_rate": 1.7649352202490198e-05, "loss": 0.7564, "step": 4727 }, { "epoch": 0.74, "grad_norm": 20.56004466670747, "learning_rate": 1.7648265871969803e-05, "loss": 0.768, "step": 4728 }, { "epoch": 0.74, "grad_norm": 22.702671486249205, "learning_rate": 1.76471793239372e-05, "loss": 0.7815, "step": 4729 }, { "epoch": 0.74, "grad_norm": 24.945495801921826, "learning_rate": 1.7646092558423288e-05, "loss": 0.8258, "step": 4730 }, { "epoch": 0.74, "grad_norm": 19.69115327136434, "learning_rate": 1.7645005575458977e-05, "loss": 0.7168, "step": 4731 }, { "epoch": 0.74, "grad_norm": 3.0180134369763225, "learning_rate": 1.764391837507518e-05, "loss": 0.6225, "step": 4732 }, { "epoch": 0.74, "grad_norm": 14.709364330858422, "learning_rate": 1.7642830957302815e-05, "loss": 0.7384, "step": 4733 }, { "epoch": 0.74, "grad_norm": 13.751418677749186, "learning_rate": 1.7641743322172812e-05, "loss": 0.7155, "step": 4734 }, { "epoch": 0.74, "grad_norm": 29.490568879027517, "learning_rate": 1.7640655469716096e-05, "loss": 0.8194, "step": 4735 }, { "epoch": 0.74, "grad_norm": 17.936220906422324, "learning_rate": 1.7639567399963607e-05, "loss": 0.7671, "step": 4736 }, { "epoch": 0.74, "grad_norm": 20.913759602721246, "learning_rate": 1.7638479112946294e-05, "loss": 0.7645, "step": 4737 }, { "epoch": 0.74, "grad_norm": 24.82486069972475, "learning_rate": 1.7637390608695102e-05, "loss": 0.8512, "step": 4738 }, { "epoch": 0.74, "grad_norm": 19.809928657989634, "learning_rate": 1.7636301887240987e-05, "loss": 0.7897, "step": 4739 }, { "epoch": 0.74, "grad_norm": 16.471959093668424, "learning_rate": 1.7635212948614915e-05, "loss": 0.6799, "step": 4740 }, { "epoch": 0.74, "grad_norm": 18.685886680214303, "learning_rate": 1.7634123792847854e-05, "loss": 0.7134, "step": 4741 }, { "epoch": 0.74, "grad_norm": 19.1684793871101, "learning_rate": 1.7633034419970775e-05, "loss": 0.796, "step": 4742 }, { "epoch": 0.74, "grad_norm": 19.809886761536184, "learning_rate": 1.7631944830014663e-05, "loss": 0.7088, "step": 4743 }, { "epoch": 0.74, "grad_norm": 14.96983445565587, "learning_rate": 1.763085502301051e-05, "loss": 0.6873, "step": 4744 }, { "epoch": 0.74, "grad_norm": 23.76149004516207, "learning_rate": 1.76297649989893e-05, "loss": 0.7967, "step": 4745 }, { "epoch": 0.74, "grad_norm": 19.463363868330706, "learning_rate": 1.7628674757982037e-05, "loss": 0.8489, "step": 4746 }, { "epoch": 0.74, "grad_norm": 15.672467960747674, "learning_rate": 1.7627584300019727e-05, "loss": 0.6577, "step": 4747 }, { "epoch": 0.74, "grad_norm": 25.878883422841493, "learning_rate": 1.7626493625133377e-05, "loss": 0.7587, "step": 4748 }, { "epoch": 0.74, "grad_norm": 20.673150511219983, "learning_rate": 1.7625402733354015e-05, "loss": 0.8432, "step": 4749 }, { "epoch": 0.74, "grad_norm": 18.784277057915933, "learning_rate": 1.7624311624712657e-05, "loss": 0.6969, "step": 4750 }, { "epoch": 0.74, "grad_norm": 18.754912266190452, "learning_rate": 1.762322029924034e-05, "loss": 0.8593, "step": 4751 }, { "epoch": 0.74, "grad_norm": 20.59320934205328, "learning_rate": 1.7622128756968095e-05, "loss": 0.9065, "step": 4752 }, { "epoch": 0.74, "grad_norm": 16.267190841368144, "learning_rate": 1.762103699792697e-05, "loss": 0.6885, "step": 4753 }, { "epoch": 0.74, "grad_norm": 20.573975020037427, "learning_rate": 1.7619945022148008e-05, "loss": 0.7677, "step": 4754 }, { "epoch": 0.74, "grad_norm": 15.137598308973239, "learning_rate": 1.7618852829662264e-05, "loss": 0.7774, "step": 4755 }, { "epoch": 0.74, "grad_norm": 18.130612327390836, "learning_rate": 1.7617760420500806e-05, "loss": 0.7532, "step": 4756 }, { "epoch": 0.74, "grad_norm": 19.903173049405172, "learning_rate": 1.7616667794694697e-05, "loss": 0.7776, "step": 4757 }, { "epoch": 0.74, "grad_norm": 21.739616063788986, "learning_rate": 1.7615574952275012e-05, "loss": 0.8543, "step": 4758 }, { "epoch": 0.74, "grad_norm": 13.746722267982145, "learning_rate": 1.761448189327283e-05, "loss": 0.7372, "step": 4759 }, { "epoch": 0.74, "grad_norm": 17.923825014880418, "learning_rate": 1.761338861771924e-05, "loss": 0.7715, "step": 4760 }, { "epoch": 0.74, "grad_norm": 24.34899404771331, "learning_rate": 1.761229512564533e-05, "loss": 0.8064, "step": 4761 }, { "epoch": 0.74, "grad_norm": 15.528797054775271, "learning_rate": 1.7611201417082196e-05, "loss": 0.7003, "step": 4762 }, { "epoch": 0.74, "grad_norm": 22.192845139928203, "learning_rate": 1.761010749206095e-05, "loss": 0.8327, "step": 4763 }, { "epoch": 0.74, "grad_norm": 18.279638844594935, "learning_rate": 1.7609013350612696e-05, "loss": 0.6791, "step": 4764 }, { "epoch": 0.74, "grad_norm": 20.375197317112875, "learning_rate": 1.7607918992768554e-05, "loss": 0.8303, "step": 4765 }, { "epoch": 0.74, "grad_norm": 30.204422041251554, "learning_rate": 1.7606824418559648e-05, "loss": 0.8284, "step": 4766 }, { "epoch": 0.74, "grad_norm": 15.738304781049953, "learning_rate": 1.7605729628017108e-05, "loss": 0.7289, "step": 4767 }, { "epoch": 0.74, "grad_norm": 23.424944146577246, "learning_rate": 1.760463462117206e-05, "loss": 0.7679, "step": 4768 }, { "epoch": 0.74, "grad_norm": 18.6465918798675, "learning_rate": 1.7603539398055658e-05, "loss": 0.7444, "step": 4769 }, { "epoch": 0.75, "grad_norm": 25.41213395370621, "learning_rate": 1.760244395869904e-05, "loss": 0.7664, "step": 4770 }, { "epoch": 0.75, "grad_norm": 18.23403088172201, "learning_rate": 1.7601348303133364e-05, "loss": 0.8096, "step": 4771 }, { "epoch": 0.75, "grad_norm": 19.55227978095691, "learning_rate": 1.760025243138979e-05, "loss": 0.767, "step": 4772 }, { "epoch": 0.75, "grad_norm": 25.0073222339007, "learning_rate": 1.7599156343499482e-05, "loss": 0.8258, "step": 4773 }, { "epoch": 0.75, "grad_norm": 17.34440598750834, "learning_rate": 1.7598060039493613e-05, "loss": 0.6841, "step": 4774 }, { "epoch": 0.75, "grad_norm": 16.792987881152875, "learning_rate": 1.759696351940336e-05, "loss": 0.8004, "step": 4775 }, { "epoch": 0.75, "grad_norm": 13.556558176506321, "learning_rate": 1.759586678325991e-05, "loss": 0.7607, "step": 4776 }, { "epoch": 0.75, "grad_norm": 15.958002349925415, "learning_rate": 1.7594769831094452e-05, "loss": 0.7425, "step": 4777 }, { "epoch": 0.75, "grad_norm": 13.902998572006434, "learning_rate": 1.759367266293818e-05, "loss": 0.7837, "step": 4778 }, { "epoch": 0.75, "grad_norm": 18.127361967484298, "learning_rate": 1.7592575278822304e-05, "loss": 0.7381, "step": 4779 }, { "epoch": 0.75, "grad_norm": 19.620887369463702, "learning_rate": 1.7591477678778027e-05, "loss": 0.8289, "step": 4780 }, { "epoch": 0.75, "grad_norm": 17.46171538137161, "learning_rate": 1.7590379862836565e-05, "loss": 0.7354, "step": 4781 }, { "epoch": 0.75, "grad_norm": 16.731126361314573, "learning_rate": 1.758928183102914e-05, "loss": 0.7713, "step": 4782 }, { "epoch": 0.75, "grad_norm": 16.345787646147702, "learning_rate": 1.7588183583386982e-05, "loss": 0.7797, "step": 4783 }, { "epoch": 0.75, "grad_norm": 18.611504364860586, "learning_rate": 1.7587085119941318e-05, "loss": 0.7315, "step": 4784 }, { "epoch": 0.75, "grad_norm": 24.601739066305928, "learning_rate": 1.758598644072339e-05, "loss": 0.678, "step": 4785 }, { "epoch": 0.75, "grad_norm": 19.040475408622893, "learning_rate": 1.7584887545764452e-05, "loss": 0.6183, "step": 4786 }, { "epoch": 0.75, "grad_norm": 18.977348198296745, "learning_rate": 1.7583788435095746e-05, "loss": 0.7622, "step": 4787 }, { "epoch": 0.75, "grad_norm": 13.658938106779566, "learning_rate": 1.758268910874853e-05, "loss": 0.7202, "step": 4788 }, { "epoch": 0.75, "grad_norm": 19.002904135147123, "learning_rate": 1.7581589566754076e-05, "loss": 0.6931, "step": 4789 }, { "epoch": 0.75, "grad_norm": 16.676365673898395, "learning_rate": 1.7580489809143648e-05, "loss": 0.6883, "step": 4790 }, { "epoch": 0.75, "grad_norm": 14.98718536532281, "learning_rate": 1.7579389835948525e-05, "loss": 0.7581, "step": 4791 }, { "epoch": 0.75, "grad_norm": 18.523750962171704, "learning_rate": 1.757828964719999e-05, "loss": 0.8242, "step": 4792 }, { "epoch": 0.75, "grad_norm": 18.675163396059236, "learning_rate": 1.7577189242929325e-05, "loss": 0.6967, "step": 4793 }, { "epoch": 0.75, "grad_norm": 12.879273057364701, "learning_rate": 1.7576088623167838e-05, "loss": 0.7332, "step": 4794 }, { "epoch": 0.75, "grad_norm": 19.176967468374464, "learning_rate": 1.7574987787946817e-05, "loss": 0.87, "step": 4795 }, { "epoch": 0.75, "grad_norm": 20.13362650544997, "learning_rate": 1.7573886737297575e-05, "loss": 0.7196, "step": 4796 }, { "epoch": 0.75, "grad_norm": 21.67366215048563, "learning_rate": 1.757278547125143e-05, "loss": 0.8146, "step": 4797 }, { "epoch": 0.75, "grad_norm": 22.81321570893647, "learning_rate": 1.7571683989839693e-05, "loss": 0.7774, "step": 4798 }, { "epoch": 0.75, "grad_norm": 17.753856486805756, "learning_rate": 1.757058229309369e-05, "loss": 0.7166, "step": 4799 }, { "epoch": 0.75, "grad_norm": 23.24128792472426, "learning_rate": 1.7569480381044758e-05, "loss": 0.7229, "step": 4800 }, { "epoch": 0.75, "grad_norm": 15.485896989409332, "learning_rate": 1.756837825372423e-05, "loss": 0.7756, "step": 4801 }, { "epoch": 0.75, "grad_norm": 16.792856612255914, "learning_rate": 1.7567275911163454e-05, "loss": 0.7584, "step": 4802 }, { "epoch": 0.75, "grad_norm": 20.3441967373095, "learning_rate": 1.7566173353393778e-05, "loss": 0.7485, "step": 4803 }, { "epoch": 0.75, "grad_norm": 20.569963981785115, "learning_rate": 1.7565070580446556e-05, "loss": 0.6857, "step": 4804 }, { "epoch": 0.75, "grad_norm": 16.88349512548097, "learning_rate": 1.7563967592353152e-05, "loss": 0.852, "step": 4805 }, { "epoch": 0.75, "grad_norm": 12.22333099633842, "learning_rate": 1.7562864389144936e-05, "loss": 0.7922, "step": 4806 }, { "epoch": 0.75, "grad_norm": 22.25885159312121, "learning_rate": 1.756176097085328e-05, "loss": 0.7922, "step": 4807 }, { "epoch": 0.75, "grad_norm": 16.06409000577669, "learning_rate": 1.7560657337509567e-05, "loss": 0.826, "step": 4808 }, { "epoch": 0.75, "grad_norm": 21.984912386875923, "learning_rate": 1.755955348914518e-05, "loss": 0.6804, "step": 4809 }, { "epoch": 0.75, "grad_norm": 29.425783566771443, "learning_rate": 1.7558449425791515e-05, "loss": 0.7631, "step": 4810 }, { "epoch": 0.75, "grad_norm": 16.734042760078, "learning_rate": 1.7557345147479968e-05, "loss": 0.7474, "step": 4811 }, { "epoch": 0.75, "grad_norm": 33.94893108040185, "learning_rate": 1.755624065424195e-05, "loss": 0.7507, "step": 4812 }, { "epoch": 0.75, "grad_norm": 17.545682669575285, "learning_rate": 1.7555135946108866e-05, "loss": 0.6758, "step": 4813 }, { "epoch": 0.75, "grad_norm": 26.479266283301598, "learning_rate": 1.7554031023112134e-05, "loss": 0.8218, "step": 4814 }, { "epoch": 0.75, "grad_norm": 13.218782299196178, "learning_rate": 1.755292588528318e-05, "loss": 0.7695, "step": 4815 }, { "epoch": 0.75, "grad_norm": 24.94502200905209, "learning_rate": 1.755182053265343e-05, "loss": 0.7807, "step": 4816 }, { "epoch": 0.75, "grad_norm": 17.34075820996803, "learning_rate": 1.7550714965254325e-05, "loss": 0.7112, "step": 4817 }, { "epoch": 0.75, "grad_norm": 15.989275127157576, "learning_rate": 1.7549609183117305e-05, "loss": 0.7459, "step": 4818 }, { "epoch": 0.75, "grad_norm": 15.15197623845428, "learning_rate": 1.7548503186273812e-05, "loss": 0.7195, "step": 4819 }, { "epoch": 0.75, "grad_norm": 27.064862393196414, "learning_rate": 1.7547396974755307e-05, "loss": 0.7602, "step": 4820 }, { "epoch": 0.75, "grad_norm": 19.261950742693383, "learning_rate": 1.7546290548593242e-05, "loss": 0.7356, "step": 4821 }, { "epoch": 0.75, "grad_norm": 31.706864058009234, "learning_rate": 1.7545183907819094e-05, "loss": 0.7768, "step": 4822 }, { "epoch": 0.75, "grad_norm": 21.51565134792058, "learning_rate": 1.7544077052464327e-05, "loss": 0.7456, "step": 4823 }, { "epoch": 0.75, "grad_norm": 12.91997191231309, "learning_rate": 1.7542969982560424e-05, "loss": 0.7127, "step": 4824 }, { "epoch": 0.75, "grad_norm": 13.831934449967568, "learning_rate": 1.754186269813886e-05, "loss": 0.6598, "step": 4825 }, { "epoch": 0.75, "grad_norm": 20.17699316988014, "learning_rate": 1.7540755199231145e-05, "loss": 0.7775, "step": 4826 }, { "epoch": 0.75, "grad_norm": 19.772209228806588, "learning_rate": 1.7539647485868753e-05, "loss": 0.6638, "step": 4827 }, { "epoch": 0.75, "grad_norm": 14.126662285225219, "learning_rate": 1.75385395580832e-05, "loss": 0.8396, "step": 4828 }, { "epoch": 0.75, "grad_norm": 23.78153301652085, "learning_rate": 1.7537431415905995e-05, "loss": 0.7789, "step": 4829 }, { "epoch": 0.75, "grad_norm": 13.539708284450187, "learning_rate": 1.7536323059368644e-05, "loss": 0.8224, "step": 4830 }, { "epoch": 0.75, "grad_norm": 15.345940725830468, "learning_rate": 1.7535214488502677e-05, "loss": 0.6338, "step": 4831 }, { "epoch": 0.75, "grad_norm": 22.021739634453333, "learning_rate": 1.753410570333962e-05, "loss": 0.7438, "step": 4832 }, { "epoch": 0.75, "grad_norm": 20.167628832709756, "learning_rate": 1.7532996703911002e-05, "loss": 0.7499, "step": 4833 }, { "epoch": 0.76, "grad_norm": 17.612009739025503, "learning_rate": 1.7531887490248364e-05, "loss": 0.7419, "step": 4834 }, { "epoch": 0.76, "grad_norm": 17.86547834964394, "learning_rate": 1.7530778062383253e-05, "loss": 0.7748, "step": 4835 }, { "epoch": 0.76, "grad_norm": 17.43928118037929, "learning_rate": 1.752966842034722e-05, "loss": 0.7467, "step": 4836 }, { "epoch": 0.76, "grad_norm": 15.462477024758247, "learning_rate": 1.752855856417182e-05, "loss": 0.7031, "step": 4837 }, { "epoch": 0.76, "grad_norm": 31.15741089797868, "learning_rate": 1.752744849388862e-05, "loss": 0.8256, "step": 4838 }, { "epoch": 0.76, "grad_norm": 11.898801636149011, "learning_rate": 1.7526338209529184e-05, "loss": 0.7411, "step": 4839 }, { "epoch": 0.76, "grad_norm": 18.635275707134667, "learning_rate": 1.7525227711125098e-05, "loss": 0.7647, "step": 4840 }, { "epoch": 0.76, "grad_norm": 18.956492013688138, "learning_rate": 1.7524116998707933e-05, "loss": 0.6666, "step": 4841 }, { "epoch": 0.76, "grad_norm": 15.724817775681023, "learning_rate": 1.7523006072309286e-05, "loss": 0.7866, "step": 4842 }, { "epoch": 0.76, "grad_norm": 30.70384077252155, "learning_rate": 1.7521894931960742e-05, "loss": 0.8051, "step": 4843 }, { "epoch": 0.76, "grad_norm": 15.734744386778676, "learning_rate": 1.7520783577693912e-05, "loss": 0.7635, "step": 4844 }, { "epoch": 0.76, "grad_norm": 13.164976594243466, "learning_rate": 1.7519672009540394e-05, "loss": 0.6252, "step": 4845 }, { "epoch": 0.76, "grad_norm": 25.772837018721546, "learning_rate": 1.7518560227531806e-05, "loss": 0.7612, "step": 4846 }, { "epoch": 0.76, "grad_norm": 21.238284836671102, "learning_rate": 1.7517448231699758e-05, "loss": 0.6688, "step": 4847 }, { "epoch": 0.76, "grad_norm": 29.834792932520127, "learning_rate": 1.7516336022075883e-05, "loss": 0.7549, "step": 4848 }, { "epoch": 0.76, "grad_norm": 20.412196228438255, "learning_rate": 1.751522359869181e-05, "loss": 0.7305, "step": 4849 }, { "epoch": 0.76, "grad_norm": 12.79203448662454, "learning_rate": 1.751411096157917e-05, "loss": 0.7108, "step": 4850 }, { "epoch": 0.76, "grad_norm": 13.71893000538938, "learning_rate": 1.7512998110769613e-05, "loss": 0.782, "step": 4851 }, { "epoch": 0.76, "grad_norm": 15.70798477915841, "learning_rate": 1.7511885046294782e-05, "loss": 0.7218, "step": 4852 }, { "epoch": 0.76, "grad_norm": 18.088597666822793, "learning_rate": 1.751077176818634e-05, "loss": 0.7287, "step": 4853 }, { "epoch": 0.76, "grad_norm": 17.38114358112564, "learning_rate": 1.7509658276475936e-05, "loss": 0.7659, "step": 4854 }, { "epoch": 0.76, "grad_norm": 22.80805134834718, "learning_rate": 1.750854457119525e-05, "loss": 0.8347, "step": 4855 }, { "epoch": 0.76, "grad_norm": 19.59051237553145, "learning_rate": 1.7507430652375943e-05, "loss": 0.7803, "step": 4856 }, { "epoch": 0.76, "grad_norm": 21.52014780433687, "learning_rate": 1.7506316520049704e-05, "loss": 0.6656, "step": 4857 }, { "epoch": 0.76, "grad_norm": 19.592379901597482, "learning_rate": 1.750520217424821e-05, "loss": 0.7752, "step": 4858 }, { "epoch": 0.76, "grad_norm": 19.259305076713584, "learning_rate": 1.750408761500316e-05, "loss": 0.7609, "step": 4859 }, { "epoch": 0.76, "grad_norm": 18.62497678217625, "learning_rate": 1.7502972842346248e-05, "loss": 0.738, "step": 4860 }, { "epoch": 0.76, "grad_norm": 16.26723370305961, "learning_rate": 1.7501857856309176e-05, "loss": 0.7739, "step": 4861 }, { "epoch": 0.76, "grad_norm": 33.64519298872203, "learning_rate": 1.7500742656923658e-05, "loss": 0.7796, "step": 4862 }, { "epoch": 0.76, "grad_norm": 20.616099507779385, "learning_rate": 1.7499627244221403e-05, "loss": 0.7832, "step": 4863 }, { "epoch": 0.76, "grad_norm": 15.383847361173087, "learning_rate": 1.7498511618234134e-05, "loss": 0.7442, "step": 4864 }, { "epoch": 0.76, "grad_norm": 13.867646029776123, "learning_rate": 1.7497395778993585e-05, "loss": 0.6778, "step": 4865 }, { "epoch": 0.76, "grad_norm": 24.667337178461022, "learning_rate": 1.749627972653149e-05, "loss": 0.7646, "step": 4866 }, { "epoch": 0.76, "grad_norm": 14.585988483431796, "learning_rate": 1.749516346087958e-05, "loss": 0.6406, "step": 4867 }, { "epoch": 0.76, "grad_norm": 16.86307876541342, "learning_rate": 1.7494046982069604e-05, "loss": 0.7236, "step": 4868 }, { "epoch": 0.76, "grad_norm": 15.67129283852544, "learning_rate": 1.749293029013332e-05, "loss": 0.7443, "step": 4869 }, { "epoch": 0.76, "grad_norm": 16.84446944078611, "learning_rate": 1.749181338510248e-05, "loss": 0.7368, "step": 4870 }, { "epoch": 0.76, "grad_norm": 20.127623775734374, "learning_rate": 1.7490696267008845e-05, "loss": 0.6603, "step": 4871 }, { "epoch": 0.76, "grad_norm": 19.399211668374484, "learning_rate": 1.74895789358842e-05, "loss": 0.6618, "step": 4872 }, { "epoch": 0.76, "grad_norm": 25.01011514832804, "learning_rate": 1.7488461391760304e-05, "loss": 0.8076, "step": 4873 }, { "epoch": 0.76, "grad_norm": 19.397100702130807, "learning_rate": 1.748734363466895e-05, "loss": 0.6843, "step": 4874 }, { "epoch": 0.76, "grad_norm": 20.45345986152622, "learning_rate": 1.748622566464192e-05, "loss": 0.7117, "step": 4875 }, { "epoch": 0.76, "grad_norm": 19.40187664018827, "learning_rate": 1.7485107481711014e-05, "loss": 0.6721, "step": 4876 }, { "epoch": 0.76, "grad_norm": 33.539235193912916, "learning_rate": 1.7483989085908027e-05, "loss": 0.8621, "step": 4877 }, { "epoch": 0.76, "grad_norm": 20.342273169264676, "learning_rate": 1.748287047726477e-05, "loss": 0.7188, "step": 4878 }, { "epoch": 0.76, "grad_norm": 25.199967304182035, "learning_rate": 1.7481751655813056e-05, "loss": 0.7636, "step": 4879 }, { "epoch": 0.76, "grad_norm": 15.462494589111566, "learning_rate": 1.74806326215847e-05, "loss": 0.8107, "step": 4880 }, { "epoch": 0.76, "grad_norm": 16.340582240645666, "learning_rate": 1.747951337461153e-05, "loss": 0.5871, "step": 4881 }, { "epoch": 0.76, "grad_norm": 29.550130129219415, "learning_rate": 1.747839391492537e-05, "loss": 0.7369, "step": 4882 }, { "epoch": 0.76, "grad_norm": 19.411405450608832, "learning_rate": 1.7477274242558064e-05, "loss": 0.7911, "step": 4883 }, { "epoch": 0.76, "grad_norm": 17.689339783440087, "learning_rate": 1.7476154357541455e-05, "loss": 0.8549, "step": 4884 }, { "epoch": 0.76, "grad_norm": 17.88998864887052, "learning_rate": 1.747503425990739e-05, "loss": 0.8061, "step": 4885 }, { "epoch": 0.76, "grad_norm": 14.532968591767025, "learning_rate": 1.747391394968772e-05, "loss": 0.7671, "step": 4886 }, { "epoch": 0.76, "grad_norm": 14.844411140575536, "learning_rate": 1.7472793426914313e-05, "loss": 0.7971, "step": 4887 }, { "epoch": 0.76, "grad_norm": 16.986351136859465, "learning_rate": 1.7471672691619028e-05, "loss": 0.704, "step": 4888 }, { "epoch": 0.76, "grad_norm": 21.54876558019364, "learning_rate": 1.7470551743833747e-05, "loss": 0.7295, "step": 4889 }, { "epoch": 0.76, "grad_norm": 20.1104018557556, "learning_rate": 1.746943058359034e-05, "loss": 0.8514, "step": 4890 }, { "epoch": 0.76, "grad_norm": 18.648603297160975, "learning_rate": 1.74683092109207e-05, "loss": 0.7439, "step": 4891 }, { "epoch": 0.76, "grad_norm": 21.10021493246341, "learning_rate": 1.7467187625856713e-05, "loss": 0.7427, "step": 4892 }, { "epoch": 0.76, "grad_norm": 20.892852968974157, "learning_rate": 1.7466065828430277e-05, "loss": 0.7124, "step": 4893 }, { "epoch": 0.76, "grad_norm": 18.42721033852216, "learning_rate": 1.7464943818673302e-05, "loss": 0.7044, "step": 4894 }, { "epoch": 0.76, "grad_norm": 22.740172822091314, "learning_rate": 1.7463821596617686e-05, "loss": 0.8082, "step": 4895 }, { "epoch": 0.76, "grad_norm": 21.424625993600092, "learning_rate": 1.7462699162295353e-05, "loss": 0.7107, "step": 4896 }, { "epoch": 0.76, "grad_norm": 14.28094236211586, "learning_rate": 1.746157651573822e-05, "loss": 0.7964, "step": 4897 }, { "epoch": 0.77, "grad_norm": 20.424930058330958, "learning_rate": 1.7460453656978217e-05, "loss": 0.7084, "step": 4898 }, { "epoch": 0.77, "grad_norm": 12.99073109747475, "learning_rate": 1.7459330586047273e-05, "loss": 0.6947, "step": 4899 }, { "epoch": 0.77, "grad_norm": 18.261263635849495, "learning_rate": 1.7458207302977333e-05, "loss": 0.7776, "step": 4900 }, { "epoch": 0.77, "grad_norm": 21.060236984020634, "learning_rate": 1.7457083807800342e-05, "loss": 0.8081, "step": 4901 }, { "epoch": 0.77, "grad_norm": 18.611709005554005, "learning_rate": 1.745596010054825e-05, "loss": 0.8036, "step": 4902 }, { "epoch": 0.77, "grad_norm": 18.718040050439416, "learning_rate": 1.745483618125301e-05, "loss": 0.7284, "step": 4903 }, { "epoch": 0.77, "grad_norm": 32.20314612982518, "learning_rate": 1.745371204994659e-05, "loss": 0.8614, "step": 4904 }, { "epoch": 0.77, "grad_norm": 22.952299269384426, "learning_rate": 1.7452587706660967e-05, "loss": 0.7266, "step": 4905 }, { "epoch": 0.77, "grad_norm": 15.38758085808322, "learning_rate": 1.7451463151428104e-05, "loss": 0.8472, "step": 4906 }, { "epoch": 0.77, "grad_norm": 26.177508221646757, "learning_rate": 1.7450338384279985e-05, "loss": 0.7725, "step": 4907 }, { "epoch": 0.77, "grad_norm": 13.94373766611307, "learning_rate": 1.7449213405248607e-05, "loss": 0.6757, "step": 4908 }, { "epoch": 0.77, "grad_norm": 11.264865436291009, "learning_rate": 1.7448088214365953e-05, "loss": 0.7085, "step": 4909 }, { "epoch": 0.77, "grad_norm": 13.981295953661448, "learning_rate": 1.744696281166403e-05, "loss": 0.7812, "step": 4910 }, { "epoch": 0.77, "grad_norm": 26.35133221025492, "learning_rate": 1.7445837197174836e-05, "loss": 0.665, "step": 4911 }, { "epoch": 0.77, "grad_norm": 15.685280684676613, "learning_rate": 1.744471137093039e-05, "loss": 0.7484, "step": 4912 }, { "epoch": 0.77, "grad_norm": 22.87073515685023, "learning_rate": 1.7443585332962715e-05, "loss": 0.6615, "step": 4913 }, { "epoch": 0.77, "grad_norm": 18.15167632578306, "learning_rate": 1.744245908330382e-05, "loss": 0.7008, "step": 4914 }, { "epoch": 0.77, "grad_norm": 16.078062662109442, "learning_rate": 1.7441332621985742e-05, "loss": 0.7807, "step": 4915 }, { "epoch": 0.77, "grad_norm": 17.922468637256276, "learning_rate": 1.744020594904052e-05, "loss": 0.7061, "step": 4916 }, { "epoch": 0.77, "grad_norm": 25.984085288414054, "learning_rate": 1.7439079064500193e-05, "loss": 0.7583, "step": 4917 }, { "epoch": 0.77, "grad_norm": 375.8586389997877, "learning_rate": 1.7437951968396808e-05, "loss": 0.7949, "step": 4918 }, { "epoch": 0.77, "grad_norm": 14.152516417959635, "learning_rate": 1.743682466076242e-05, "loss": 0.7221, "step": 4919 }, { "epoch": 0.77, "grad_norm": 20.097579075819894, "learning_rate": 1.7435697141629087e-05, "loss": 0.7473, "step": 4920 }, { "epoch": 0.77, "grad_norm": 15.952375938958086, "learning_rate": 1.7434569411028883e-05, "loss": 0.8087, "step": 4921 }, { "epoch": 0.77, "grad_norm": 18.23459097673864, "learning_rate": 1.743344146899387e-05, "loss": 0.7221, "step": 4922 }, { "epoch": 0.77, "grad_norm": 21.46926016388806, "learning_rate": 1.7432313315556132e-05, "loss": 0.7627, "step": 4923 }, { "epoch": 0.77, "grad_norm": 12.914724127061033, "learning_rate": 1.743118495074775e-05, "loss": 0.6511, "step": 4924 }, { "epoch": 0.77, "grad_norm": 14.53497233552534, "learning_rate": 1.7430056374600813e-05, "loss": 0.6766, "step": 4925 }, { "epoch": 0.77, "grad_norm": 29.61839743366324, "learning_rate": 1.7428927587147422e-05, "loss": 0.7791, "step": 4926 }, { "epoch": 0.77, "grad_norm": 22.54962927979832, "learning_rate": 1.7427798588419674e-05, "loss": 0.7249, "step": 4927 }, { "epoch": 0.77, "grad_norm": 17.21038336777368, "learning_rate": 1.7426669378449685e-05, "loss": 0.7438, "step": 4928 }, { "epoch": 0.77, "grad_norm": 19.168165871737898, "learning_rate": 1.7425539957269557e-05, "loss": 0.6793, "step": 4929 }, { "epoch": 0.77, "grad_norm": 20.76502212412781, "learning_rate": 1.742441032491142e-05, "loss": 0.8016, "step": 4930 }, { "epoch": 0.77, "grad_norm": 30.406634951680513, "learning_rate": 1.7423280481407393e-05, "loss": 0.8101, "step": 4931 }, { "epoch": 0.77, "grad_norm": 19.897400229519402, "learning_rate": 1.7422150426789613e-05, "loss": 0.7459, "step": 4932 }, { "epoch": 0.77, "grad_norm": 19.69640883254658, "learning_rate": 1.7421020161090216e-05, "loss": 0.6963, "step": 4933 }, { "epoch": 0.77, "grad_norm": 27.31423682794454, "learning_rate": 1.741988968434135e-05, "loss": 0.7501, "step": 4934 }, { "epoch": 0.77, "grad_norm": 18.355686037896728, "learning_rate": 1.741875899657516e-05, "loss": 0.6719, "step": 4935 }, { "epoch": 0.77, "grad_norm": 15.52960706373916, "learning_rate": 1.7417628097823802e-05, "loss": 0.7426, "step": 4936 }, { "epoch": 0.77, "grad_norm": 24.38273804178698, "learning_rate": 1.741649698811944e-05, "loss": 0.715, "step": 4937 }, { "epoch": 0.77, "grad_norm": 16.799724867602897, "learning_rate": 1.7415365667494248e-05, "loss": 0.7854, "step": 4938 }, { "epoch": 0.77, "grad_norm": 26.92200338982838, "learning_rate": 1.741423413598039e-05, "loss": 0.84, "step": 4939 }, { "epoch": 0.77, "grad_norm": 27.316727055810073, "learning_rate": 1.741310239361005e-05, "loss": 0.8222, "step": 4940 }, { "epoch": 0.77, "grad_norm": 17.01069974595615, "learning_rate": 1.7411970440415418e-05, "loss": 0.7209, "step": 4941 }, { "epoch": 0.77, "grad_norm": 16.595927858845673, "learning_rate": 1.7410838276428677e-05, "loss": 0.7543, "step": 4942 }, { "epoch": 0.77, "grad_norm": 16.564628310031143, "learning_rate": 1.7409705901682033e-05, "loss": 0.804, "step": 4943 }, { "epoch": 0.77, "grad_norm": 15.485886607309236, "learning_rate": 1.740857331620769e-05, "loss": 0.6732, "step": 4944 }, { "epoch": 0.77, "grad_norm": 16.846433901650936, "learning_rate": 1.740744052003785e-05, "loss": 0.7495, "step": 4945 }, { "epoch": 0.77, "grad_norm": 18.38669637237412, "learning_rate": 1.7406307513204742e-05, "loss": 0.7389, "step": 4946 }, { "epoch": 0.77, "grad_norm": 11.912706189107379, "learning_rate": 1.7405174295740577e-05, "loss": 0.6628, "step": 4947 }, { "epoch": 0.77, "grad_norm": 14.065252702231586, "learning_rate": 1.7404040867677587e-05, "loss": 0.6699, "step": 4948 }, { "epoch": 0.77, "grad_norm": 18.96734736895513, "learning_rate": 1.740290722904801e-05, "loss": 0.7858, "step": 4949 }, { "epoch": 0.77, "grad_norm": 17.967239962029407, "learning_rate": 1.7401773379884078e-05, "loss": 0.7964, "step": 4950 }, { "epoch": 0.77, "grad_norm": 16.241067718393687, "learning_rate": 1.7400639320218042e-05, "loss": 0.7308, "step": 4951 }, { "epoch": 0.77, "grad_norm": 11.299745099140768, "learning_rate": 1.739950505008215e-05, "loss": 0.6156, "step": 4952 }, { "epoch": 0.77, "grad_norm": 27.602985599130843, "learning_rate": 1.7398370569508667e-05, "loss": 0.7483, "step": 4953 }, { "epoch": 0.77, "grad_norm": 13.337192921021662, "learning_rate": 1.739723587852985e-05, "loss": 0.7972, "step": 4954 }, { "epoch": 0.77, "grad_norm": 18.357646844570873, "learning_rate": 1.7396100977177975e-05, "loss": 0.7109, "step": 4955 }, { "epoch": 0.77, "grad_norm": 17.871261652590274, "learning_rate": 1.7394965865485312e-05, "loss": 0.7369, "step": 4956 }, { "epoch": 0.77, "grad_norm": 20.456156405465862, "learning_rate": 1.7393830543484147e-05, "loss": 0.7759, "step": 4957 }, { "epoch": 0.77, "grad_norm": 25.044864504838507, "learning_rate": 1.7392695011206768e-05, "loss": 0.7819, "step": 4958 }, { "epoch": 0.77, "grad_norm": 11.588033656940288, "learning_rate": 1.7391559268685464e-05, "loss": 0.7667, "step": 4959 }, { "epoch": 0.77, "grad_norm": 21.153757569746126, "learning_rate": 1.739042331595254e-05, "loss": 0.8124, "step": 4960 }, { "epoch": 0.77, "grad_norm": 16.839408350346112, "learning_rate": 1.73892871530403e-05, "loss": 0.692, "step": 4961 }, { "epoch": 0.78, "grad_norm": 26.906280886403078, "learning_rate": 1.7388150779981057e-05, "loss": 0.7894, "step": 4962 }, { "epoch": 0.78, "grad_norm": 18.98817814852142, "learning_rate": 1.738701419680713e-05, "loss": 0.7946, "step": 4963 }, { "epoch": 0.78, "grad_norm": 14.921376115637958, "learning_rate": 1.7385877403550836e-05, "loss": 0.7439, "step": 4964 }, { "epoch": 0.78, "grad_norm": 18.48194463783497, "learning_rate": 1.738474040024451e-05, "loss": 0.7844, "step": 4965 }, { "epoch": 0.78, "grad_norm": 25.02973950099229, "learning_rate": 1.738360318692049e-05, "loss": 0.7565, "step": 4966 }, { "epoch": 0.78, "grad_norm": 10.903436595742544, "learning_rate": 1.738246576361111e-05, "loss": 0.6621, "step": 4967 }, { "epoch": 0.78, "grad_norm": 14.457844870285887, "learning_rate": 1.7381328130348727e-05, "loss": 0.6961, "step": 4968 }, { "epoch": 0.78, "grad_norm": 25.22048976456105, "learning_rate": 1.7380190287165686e-05, "loss": 0.7874, "step": 4969 }, { "epoch": 0.78, "grad_norm": 22.782271585661586, "learning_rate": 1.7379052234094353e-05, "loss": 0.7587, "step": 4970 }, { "epoch": 0.78, "grad_norm": 21.653046819361595, "learning_rate": 1.737791397116709e-05, "loss": 0.7627, "step": 4971 }, { "epoch": 0.78, "grad_norm": 26.451389024828025, "learning_rate": 1.737677549841627e-05, "loss": 0.8129, "step": 4972 }, { "epoch": 0.78, "grad_norm": 17.51047892184048, "learning_rate": 1.7375636815874273e-05, "loss": 0.7612, "step": 4973 }, { "epoch": 0.78, "grad_norm": 15.190522751359087, "learning_rate": 1.7374497923573473e-05, "loss": 0.6705, "step": 4974 }, { "epoch": 0.78, "grad_norm": 25.356071201725552, "learning_rate": 1.7373358821546272e-05, "loss": 0.7768, "step": 4975 }, { "epoch": 0.78, "grad_norm": 22.612871549515546, "learning_rate": 1.7372219509825056e-05, "loss": 0.6925, "step": 4976 }, { "epoch": 0.78, "grad_norm": 18.09954965700338, "learning_rate": 1.737107998844223e-05, "loss": 0.7263, "step": 4977 }, { "epoch": 0.78, "grad_norm": 16.89917224533569, "learning_rate": 1.7369940257430203e-05, "loss": 0.6812, "step": 4978 }, { "epoch": 0.78, "grad_norm": 34.32157037706258, "learning_rate": 1.7368800316821387e-05, "loss": 0.7067, "step": 4979 }, { "epoch": 0.78, "grad_norm": 20.21420758448821, "learning_rate": 1.7367660166648197e-05, "loss": 0.7596, "step": 4980 }, { "epoch": 0.78, "grad_norm": 17.927504539990547, "learning_rate": 1.7366519806943067e-05, "loss": 0.7706, "step": 4981 }, { "epoch": 0.78, "grad_norm": 23.745854640258234, "learning_rate": 1.736537923773842e-05, "loss": 0.7822, "step": 4982 }, { "epoch": 0.78, "grad_norm": 15.82011657499333, "learning_rate": 1.736423845906669e-05, "loss": 0.7312, "step": 4983 }, { "epoch": 0.78, "grad_norm": 40.469379937486714, "learning_rate": 1.7363097470960336e-05, "loss": 0.7486, "step": 4984 }, { "epoch": 0.78, "grad_norm": 21.289888085272274, "learning_rate": 1.736195627345179e-05, "loss": 0.7448, "step": 4985 }, { "epoch": 0.78, "grad_norm": 20.01380473487693, "learning_rate": 1.736081486657352e-05, "loss": 0.7457, "step": 4986 }, { "epoch": 0.78, "grad_norm": 16.8077806962746, "learning_rate": 1.7359673250357977e-05, "loss": 0.687, "step": 4987 }, { "epoch": 0.78, "grad_norm": 25.211705960943725, "learning_rate": 1.7358531424837633e-05, "loss": 0.7321, "step": 4988 }, { "epoch": 0.78, "grad_norm": 19.987147723469857, "learning_rate": 1.7357389390044965e-05, "loss": 0.7628, "step": 4989 }, { "epoch": 0.78, "grad_norm": 26.934819811715208, "learning_rate": 1.735624714601244e-05, "loss": 0.7881, "step": 4990 }, { "epoch": 0.78, "grad_norm": 17.968630449477843, "learning_rate": 1.735510469277255e-05, "loss": 0.7534, "step": 4991 }, { "epoch": 0.78, "grad_norm": 29.594671351864825, "learning_rate": 1.7353962030357786e-05, "loss": 0.8809, "step": 4992 }, { "epoch": 0.78, "grad_norm": 29.502093776252696, "learning_rate": 1.7352819158800646e-05, "loss": 0.9089, "step": 4993 }, { "epoch": 0.78, "grad_norm": 19.798676835568042, "learning_rate": 1.735167607813363e-05, "loss": 0.7955, "step": 4994 }, { "epoch": 0.78, "grad_norm": 18.52557333945106, "learning_rate": 1.7350532788389248e-05, "loss": 0.7633, "step": 4995 }, { "epoch": 0.78, "grad_norm": 17.960497202377574, "learning_rate": 1.734938928960001e-05, "loss": 0.7301, "step": 4996 }, { "epoch": 0.78, "grad_norm": 14.114733823159321, "learning_rate": 1.7348245581798438e-05, "loss": 0.6224, "step": 4997 }, { "epoch": 0.78, "grad_norm": 15.798089532073249, "learning_rate": 1.7347101665017064e-05, "loss": 0.6808, "step": 4998 }, { "epoch": 0.78, "grad_norm": 15.943326265187848, "learning_rate": 1.7345957539288417e-05, "loss": 0.6658, "step": 4999 }, { "epoch": 0.78, "grad_norm": 35.25788677916932, "learning_rate": 1.7344813204645033e-05, "loss": 0.8039, "step": 5000 }, { "epoch": 0.78, "grad_norm": 21.701814065027627, "learning_rate": 1.734366866111946e-05, "loss": 0.7383, "step": 5001 }, { "epoch": 0.78, "grad_norm": 24.47408580973461, "learning_rate": 1.7342523908744246e-05, "loss": 0.6607, "step": 5002 }, { "epoch": 0.78, "grad_norm": 46.325402325204784, "learning_rate": 1.7341378947551946e-05, "loss": 0.7624, "step": 5003 }, { "epoch": 0.78, "grad_norm": 21.05457513775683, "learning_rate": 1.7340233777575125e-05, "loss": 0.7722, "step": 5004 }, { "epoch": 0.78, "grad_norm": 21.92265491821371, "learning_rate": 1.733908839884635e-05, "loss": 0.7098, "step": 5005 }, { "epoch": 0.78, "grad_norm": 19.640089091452882, "learning_rate": 1.733794281139819e-05, "loss": 0.6812, "step": 5006 }, { "epoch": 0.78, "grad_norm": 18.545110240138342, "learning_rate": 1.733679701526323e-05, "loss": 0.7906, "step": 5007 }, { "epoch": 0.78, "grad_norm": 24.377396656467127, "learning_rate": 1.7335651010474057e-05, "loss": 0.7975, "step": 5008 }, { "epoch": 0.78, "grad_norm": 21.033133173062968, "learning_rate": 1.7334504797063257e-05, "loss": 0.8084, "step": 5009 }, { "epoch": 0.78, "grad_norm": 19.281571613547705, "learning_rate": 1.733335837506344e-05, "loss": 0.7777, "step": 5010 }, { "epoch": 0.78, "grad_norm": 20.245447107638785, "learning_rate": 1.733221174450719e-05, "loss": 0.782, "step": 5011 }, { "epoch": 0.78, "grad_norm": 23.19563933656489, "learning_rate": 1.7331064905427133e-05, "loss": 0.7788, "step": 5012 }, { "epoch": 0.78, "grad_norm": 15.869574781846504, "learning_rate": 1.7329917857855875e-05, "loss": 0.696, "step": 5013 }, { "epoch": 0.78, "grad_norm": 19.764099954869234, "learning_rate": 1.7328770601826047e-05, "loss": 0.673, "step": 5014 }, { "epoch": 0.78, "grad_norm": 14.066046134154801, "learning_rate": 1.7327623137370265e-05, "loss": 0.8146, "step": 5015 }, { "epoch": 0.78, "grad_norm": 25.078479193328253, "learning_rate": 1.732647546452117e-05, "loss": 0.7619, "step": 5016 }, { "epoch": 0.78, "grad_norm": 14.420310542099985, "learning_rate": 1.7325327583311396e-05, "loss": 0.609, "step": 5017 }, { "epoch": 0.78, "grad_norm": 19.662384773600046, "learning_rate": 1.7324179493773593e-05, "loss": 0.8095, "step": 5018 }, { "epoch": 0.78, "grad_norm": 14.057423188069045, "learning_rate": 1.732303119594041e-05, "loss": 0.7724, "step": 5019 }, { "epoch": 0.78, "grad_norm": 13.762989985979699, "learning_rate": 1.73218826898445e-05, "loss": 0.704, "step": 5020 }, { "epoch": 0.78, "grad_norm": 15.115908829384493, "learning_rate": 1.7320733975518533e-05, "loss": 0.6469, "step": 5021 }, { "epoch": 0.78, "grad_norm": 21.836261484816877, "learning_rate": 1.7319585052995177e-05, "loss": 0.7122, "step": 5022 }, { "epoch": 0.78, "grad_norm": 29.730244977642954, "learning_rate": 1.7318435922307097e-05, "loss": 0.6995, "step": 5023 }, { "epoch": 0.78, "grad_norm": 19.261123366428865, "learning_rate": 1.7317286583486983e-05, "loss": 0.8975, "step": 5024 }, { "epoch": 0.78, "grad_norm": 13.467894593355588, "learning_rate": 1.7316137036567523e-05, "loss": 0.6568, "step": 5025 }, { "epoch": 0.79, "grad_norm": 17.87838022659654, "learning_rate": 1.7314987281581403e-05, "loss": 0.7126, "step": 5026 }, { "epoch": 0.79, "grad_norm": 26.72086776237663, "learning_rate": 1.731383731856132e-05, "loss": 0.8774, "step": 5027 }, { "epoch": 0.79, "grad_norm": 16.906667107057004, "learning_rate": 1.731268714753999e-05, "loss": 0.7681, "step": 5028 }, { "epoch": 0.79, "grad_norm": 21.15878858084662, "learning_rate": 1.7311536768550107e-05, "loss": 0.7387, "step": 5029 }, { "epoch": 0.79, "grad_norm": 24.56685865958884, "learning_rate": 1.73103861816244e-05, "loss": 0.9333, "step": 5030 }, { "epoch": 0.79, "grad_norm": 16.591399114734493, "learning_rate": 1.7309235386795585e-05, "loss": 0.7584, "step": 5031 }, { "epoch": 0.79, "grad_norm": 17.610174713067547, "learning_rate": 1.7308084384096395e-05, "loss": 0.7689, "step": 5032 }, { "epoch": 0.79, "grad_norm": 17.458683640351946, "learning_rate": 1.7306933173559554e-05, "loss": 0.8298, "step": 5033 }, { "epoch": 0.79, "grad_norm": 15.872977671443623, "learning_rate": 1.7305781755217812e-05, "loss": 0.6979, "step": 5034 }, { "epoch": 0.79, "grad_norm": 33.247439795895005, "learning_rate": 1.730463012910391e-05, "loss": 0.7918, "step": 5035 }, { "epoch": 0.79, "grad_norm": 28.85241462860156, "learning_rate": 1.73034782952506e-05, "loss": 0.8116, "step": 5036 }, { "epoch": 0.79, "grad_norm": 13.957824468826395, "learning_rate": 1.7302326253690643e-05, "loss": 0.7945, "step": 5037 }, { "epoch": 0.79, "grad_norm": 18.111495393334224, "learning_rate": 1.7301174004456794e-05, "loss": 0.6709, "step": 5038 }, { "epoch": 0.79, "grad_norm": 16.74365873854793, "learning_rate": 1.730002154758183e-05, "loss": 0.7245, "step": 5039 }, { "epoch": 0.79, "grad_norm": 16.507523721912314, "learning_rate": 1.729886888309852e-05, "loss": 0.7429, "step": 5040 }, { "epoch": 0.79, "grad_norm": 21.073937723373685, "learning_rate": 1.7297716011039654e-05, "loss": 0.8302, "step": 5041 }, { "epoch": 0.79, "grad_norm": 14.857892900153706, "learning_rate": 1.729656293143801e-05, "loss": 0.7171, "step": 5042 }, { "epoch": 0.79, "grad_norm": 15.253516744410266, "learning_rate": 1.7295409644326387e-05, "loss": 0.6826, "step": 5043 }, { "epoch": 0.79, "grad_norm": 4.94960649562677, "learning_rate": 1.7294256149737577e-05, "loss": 0.7902, "step": 5044 }, { "epoch": 0.79, "grad_norm": 12.635953946089323, "learning_rate": 1.7293102447704395e-05, "loss": 0.6542, "step": 5045 }, { "epoch": 0.79, "grad_norm": 19.098848532733925, "learning_rate": 1.7291948538259644e-05, "loss": 0.687, "step": 5046 }, { "epoch": 0.79, "grad_norm": 19.597190720623495, "learning_rate": 1.7290794421436138e-05, "loss": 0.7154, "step": 5047 }, { "epoch": 0.79, "grad_norm": 17.319660975909798, "learning_rate": 1.7289640097266712e-05, "loss": 0.739, "step": 5048 }, { "epoch": 0.79, "grad_norm": 17.850288444146614, "learning_rate": 1.728848556578418e-05, "loss": 0.8241, "step": 5049 }, { "epoch": 0.79, "grad_norm": 18.412644354745783, "learning_rate": 1.7287330827021382e-05, "loss": 0.7195, "step": 5050 }, { "epoch": 0.79, "grad_norm": 15.979657389637712, "learning_rate": 1.7286175881011158e-05, "loss": 0.7461, "step": 5051 }, { "epoch": 0.79, "grad_norm": 20.853701563777296, "learning_rate": 1.7285020727786354e-05, "loss": 0.7505, "step": 5052 }, { "epoch": 0.79, "grad_norm": 25.178964584961466, "learning_rate": 1.7283865367379826e-05, "loss": 0.7019, "step": 5053 }, { "epoch": 0.79, "grad_norm": 18.59595817869395, "learning_rate": 1.7282709799824428e-05, "loss": 0.7261, "step": 5054 }, { "epoch": 0.79, "grad_norm": 25.532895160090938, "learning_rate": 1.7281554025153018e-05, "loss": 0.9205, "step": 5055 }, { "epoch": 0.79, "grad_norm": 18.24995895276614, "learning_rate": 1.7280398043398478e-05, "loss": 0.8256, "step": 5056 }, { "epoch": 0.79, "grad_norm": 33.58273686706776, "learning_rate": 1.727924185459367e-05, "loss": 0.8316, "step": 5057 }, { "epoch": 0.79, "grad_norm": 26.184429837137692, "learning_rate": 1.7278085458771485e-05, "loss": 0.7576, "step": 5058 }, { "epoch": 0.79, "grad_norm": 26.901705563019643, "learning_rate": 1.7276928855964805e-05, "loss": 0.685, "step": 5059 }, { "epoch": 0.79, "grad_norm": 23.299143703119697, "learning_rate": 1.727577204620653e-05, "loss": 0.716, "step": 5060 }, { "epoch": 0.79, "grad_norm": 42.928552792021215, "learning_rate": 1.727461502952955e-05, "loss": 0.8264, "step": 5061 }, { "epoch": 0.79, "grad_norm": 17.698791213334815, "learning_rate": 1.7273457805966773e-05, "loss": 0.8002, "step": 5062 }, { "epoch": 0.79, "grad_norm": 33.83940608694808, "learning_rate": 1.7272300375551116e-05, "loss": 0.6599, "step": 5063 }, { "epoch": 0.79, "grad_norm": 19.194257719646686, "learning_rate": 1.7271142738315485e-05, "loss": 0.6953, "step": 5064 }, { "epoch": 0.79, "grad_norm": 30.302458308170536, "learning_rate": 1.7269984894292814e-05, "loss": 0.8831, "step": 5065 }, { "epoch": 0.79, "grad_norm": 16.28154251088473, "learning_rate": 1.726882684351602e-05, "loss": 0.7863, "step": 5066 }, { "epoch": 0.79, "grad_norm": 17.52729123853482, "learning_rate": 1.7267668586018044e-05, "loss": 0.729, "step": 5067 }, { "epoch": 0.79, "grad_norm": 22.744382800151502, "learning_rate": 1.7266510121831824e-05, "loss": 0.7486, "step": 5068 }, { "epoch": 0.79, "grad_norm": 15.712542562381573, "learning_rate": 1.726535145099031e-05, "loss": 0.6335, "step": 5069 }, { "epoch": 0.79, "grad_norm": 20.561469044247666, "learning_rate": 1.726419257352645e-05, "loss": 0.6944, "step": 5070 }, { "epoch": 0.79, "grad_norm": 24.059782782473214, "learning_rate": 1.7263033489473202e-05, "loss": 0.7326, "step": 5071 }, { "epoch": 0.79, "grad_norm": 28.439633716454974, "learning_rate": 1.7261874198863533e-05, "loss": 0.8421, "step": 5072 }, { "epoch": 0.79, "grad_norm": 16.85165307626801, "learning_rate": 1.7260714701730405e-05, "loss": 0.6602, "step": 5073 }, { "epoch": 0.79, "grad_norm": 24.32739867765844, "learning_rate": 1.72595549981068e-05, "loss": 0.7712, "step": 5074 }, { "epoch": 0.79, "grad_norm": 18.7142273975174, "learning_rate": 1.72583950880257e-05, "loss": 0.763, "step": 5075 }, { "epoch": 0.79, "grad_norm": 17.549305123802387, "learning_rate": 1.7257234971520086e-05, "loss": 0.7109, "step": 5076 }, { "epoch": 0.79, "grad_norm": 21.870855238142497, "learning_rate": 1.7256074648622958e-05, "loss": 0.7423, "step": 5077 }, { "epoch": 0.79, "grad_norm": 18.662565530964063, "learning_rate": 1.725491411936731e-05, "loss": 0.7172, "step": 5078 }, { "epoch": 0.79, "grad_norm": 15.344096756724776, "learning_rate": 1.7253753383786148e-05, "loss": 0.771, "step": 5079 }, { "epoch": 0.79, "grad_norm": 22.20295165341256, "learning_rate": 1.7252592441912487e-05, "loss": 0.7137, "step": 5080 }, { "epoch": 0.79, "grad_norm": 22.099638587167057, "learning_rate": 1.7251431293779334e-05, "loss": 0.7776, "step": 5081 }, { "epoch": 0.79, "grad_norm": 35.31761512252748, "learning_rate": 1.725026993941972e-05, "loss": 0.7658, "step": 5082 }, { "epoch": 0.79, "grad_norm": 18.967257706907937, "learning_rate": 1.724910837886667e-05, "loss": 0.7261, "step": 5083 }, { "epoch": 0.79, "grad_norm": 29.378218776803575, "learning_rate": 1.7247946612153215e-05, "loss": 0.8825, "step": 5084 }, { "epoch": 0.79, "grad_norm": 18.34233699259649, "learning_rate": 1.7246784639312403e-05, "loss": 0.805, "step": 5085 }, { "epoch": 0.79, "grad_norm": 19.742365402292652, "learning_rate": 1.7245622460377274e-05, "loss": 0.7195, "step": 5086 }, { "epoch": 0.79, "grad_norm": 18.991661550158632, "learning_rate": 1.724446007538088e-05, "loss": 0.658, "step": 5087 }, { "epoch": 0.79, "grad_norm": 19.917805067842313, "learning_rate": 1.724329748435628e-05, "loss": 0.7155, "step": 5088 }, { "epoch": 0.79, "grad_norm": 18.351676272709078, "learning_rate": 1.7242134687336535e-05, "loss": 0.701, "step": 5089 }, { "epoch": 0.8, "grad_norm": 19.300366654504995, "learning_rate": 1.7240971684354717e-05, "loss": 0.6904, "step": 5090 }, { "epoch": 0.8, "grad_norm": 24.543106932966747, "learning_rate": 1.72398084754439e-05, "loss": 0.8452, "step": 5091 }, { "epoch": 0.8, "grad_norm": 19.219439792659678, "learning_rate": 1.7238645060637165e-05, "loss": 0.7456, "step": 5092 }, { "epoch": 0.8, "grad_norm": 31.200695625694124, "learning_rate": 1.7237481439967602e-05, "loss": 0.8066, "step": 5093 }, { "epoch": 0.8, "grad_norm": 31.05674417462991, "learning_rate": 1.7236317613468294e-05, "loss": 0.7463, "step": 5094 }, { "epoch": 0.8, "grad_norm": 14.318735471840476, "learning_rate": 1.7235153581172353e-05, "loss": 0.7005, "step": 5095 }, { "epoch": 0.8, "grad_norm": 16.241663489024187, "learning_rate": 1.7233989343112872e-05, "loss": 0.7386, "step": 5096 }, { "epoch": 0.8, "grad_norm": 18.114435285438898, "learning_rate": 1.723282489932297e-05, "loss": 0.6815, "step": 5097 }, { "epoch": 0.8, "grad_norm": 19.384308548542357, "learning_rate": 1.7231660249835756e-05, "loss": 0.6392, "step": 5098 }, { "epoch": 0.8, "grad_norm": 17.817426268006322, "learning_rate": 1.7230495394684353e-05, "loss": 0.7207, "step": 5099 }, { "epoch": 0.8, "grad_norm": 17.575994480720233, "learning_rate": 1.7229330333901895e-05, "loss": 0.7451, "step": 5100 }, { "epoch": 0.8, "grad_norm": 13.362047778565351, "learning_rate": 1.722816506752151e-05, "loss": 0.6762, "step": 5101 }, { "epoch": 0.8, "grad_norm": 27.31812986382629, "learning_rate": 1.722699959557634e-05, "loss": 0.6989, "step": 5102 }, { "epoch": 0.8, "grad_norm": 29.037783050881803, "learning_rate": 1.7225833918099527e-05, "loss": 0.8387, "step": 5103 }, { "epoch": 0.8, "grad_norm": 20.149734461983485, "learning_rate": 1.7224668035124224e-05, "loss": 0.7087, "step": 5104 }, { "epoch": 0.8, "grad_norm": 20.376804377223998, "learning_rate": 1.722350194668359e-05, "loss": 0.8158, "step": 5105 }, { "epoch": 0.8, "grad_norm": 23.28636350223484, "learning_rate": 1.7222335652810788e-05, "loss": 0.7634, "step": 5106 }, { "epoch": 0.8, "grad_norm": 12.50021746565481, "learning_rate": 1.7221169153538978e-05, "loss": 0.7332, "step": 5107 }, { "epoch": 0.8, "grad_norm": 17.11290429002052, "learning_rate": 1.7220002448901346e-05, "loss": 0.6832, "step": 5108 }, { "epoch": 0.8, "grad_norm": 23.245410589283114, "learning_rate": 1.721883553893107e-05, "loss": 0.7127, "step": 5109 }, { "epoch": 0.8, "grad_norm": 18.672782294523607, "learning_rate": 1.7217668423661332e-05, "loss": 0.6993, "step": 5110 }, { "epoch": 0.8, "grad_norm": 18.904317338145937, "learning_rate": 1.7216501103125326e-05, "loss": 0.6789, "step": 5111 }, { "epoch": 0.8, "grad_norm": 26.703068857597138, "learning_rate": 1.721533357735625e-05, "loss": 0.7675, "step": 5112 }, { "epoch": 0.8, "grad_norm": 11.342364338243668, "learning_rate": 1.721416584638731e-05, "loss": 0.6744, "step": 5113 }, { "epoch": 0.8, "grad_norm": 16.11973672871839, "learning_rate": 1.721299791025171e-05, "loss": 0.7176, "step": 5114 }, { "epoch": 0.8, "grad_norm": 25.50128034049107, "learning_rate": 1.7211829768982672e-05, "loss": 0.8299, "step": 5115 }, { "epoch": 0.8, "grad_norm": 19.21175173677941, "learning_rate": 1.7210661422613412e-05, "loss": 0.8219, "step": 5116 }, { "epoch": 0.8, "grad_norm": 22.4106500739761, "learning_rate": 1.720949287117716e-05, "loss": 0.806, "step": 5117 }, { "epoch": 0.8, "grad_norm": 38.73741170044382, "learning_rate": 1.720832411470715e-05, "loss": 0.8398, "step": 5118 }, { "epoch": 0.8, "grad_norm": 19.31007118908587, "learning_rate": 1.720715515323662e-05, "loss": 0.7324, "step": 5119 }, { "epoch": 0.8, "grad_norm": 16.226876249847756, "learning_rate": 1.7205985986798808e-05, "loss": 0.8309, "step": 5120 }, { "epoch": 0.8, "grad_norm": 15.664302245705011, "learning_rate": 1.7204816615426972e-05, "loss": 0.7168, "step": 5121 }, { "epoch": 0.8, "grad_norm": 17.232590517474655, "learning_rate": 1.7203647039154367e-05, "loss": 0.6735, "step": 5122 }, { "epoch": 0.8, "grad_norm": 11.68743278871682, "learning_rate": 1.7202477258014256e-05, "loss": 0.7258, "step": 5123 }, { "epoch": 0.8, "grad_norm": 25.935429112939584, "learning_rate": 1.72013072720399e-05, "loss": 0.7146, "step": 5124 }, { "epoch": 0.8, "grad_norm": 13.39228771974679, "learning_rate": 1.7200137081264584e-05, "loss": 0.7438, "step": 5125 }, { "epoch": 0.8, "grad_norm": 23.290735597346607, "learning_rate": 1.719896668572158e-05, "loss": 0.805, "step": 5126 }, { "epoch": 0.8, "grad_norm": 21.979866063634727, "learning_rate": 1.7197796085444173e-05, "loss": 0.7777, "step": 5127 }, { "epoch": 0.8, "grad_norm": 21.221921175017695, "learning_rate": 1.7196625280465656e-05, "loss": 0.6502, "step": 5128 }, { "epoch": 0.8, "grad_norm": 38.09581345984922, "learning_rate": 1.7195454270819326e-05, "loss": 0.7592, "step": 5129 }, { "epoch": 0.8, "grad_norm": 14.941661498438886, "learning_rate": 1.7194283056538486e-05, "loss": 0.8059, "step": 5130 }, { "epoch": 0.8, "grad_norm": 18.06014691975666, "learning_rate": 1.7193111637656445e-05, "loss": 0.697, "step": 5131 }, { "epoch": 0.8, "grad_norm": 43.032591067507944, "learning_rate": 1.7191940014206518e-05, "loss": 0.7742, "step": 5132 }, { "epoch": 0.8, "grad_norm": 14.028730004207219, "learning_rate": 1.719076818622202e-05, "loss": 0.7147, "step": 5133 }, { "epoch": 0.8, "grad_norm": 17.75989389284088, "learning_rate": 1.7189596153736285e-05, "loss": 0.6818, "step": 5134 }, { "epoch": 0.8, "grad_norm": 22.58132415862778, "learning_rate": 1.7188423916782637e-05, "loss": 0.7869, "step": 5135 }, { "epoch": 0.8, "grad_norm": 16.37672344494824, "learning_rate": 1.7187251475394423e-05, "loss": 0.6452, "step": 5136 }, { "epoch": 0.8, "grad_norm": 32.285660965811566, "learning_rate": 1.7186078829604978e-05, "loss": 0.7822, "step": 5137 }, { "epoch": 0.8, "grad_norm": 17.285971889756457, "learning_rate": 1.7184905979447655e-05, "loss": 0.685, "step": 5138 }, { "epoch": 0.8, "grad_norm": 18.100395532166885, "learning_rate": 1.7183732924955808e-05, "loss": 0.7418, "step": 5139 }, { "epoch": 0.8, "grad_norm": 16.721195950690603, "learning_rate": 1.7182559666162802e-05, "loss": 0.737, "step": 5140 }, { "epoch": 0.8, "grad_norm": 15.197672171376754, "learning_rate": 1.7181386203102e-05, "loss": 0.6763, "step": 5141 }, { "epoch": 0.8, "grad_norm": 21.89372631094041, "learning_rate": 1.7180212535806776e-05, "loss": 0.7672, "step": 5142 }, { "epoch": 0.8, "grad_norm": 14.410474215830988, "learning_rate": 1.717903866431051e-05, "loss": 0.7046, "step": 5143 }, { "epoch": 0.8, "grad_norm": 24.094264874727255, "learning_rate": 1.717786458864658e-05, "loss": 0.7272, "step": 5144 }, { "epoch": 0.8, "grad_norm": 13.63603840510089, "learning_rate": 1.717669030884838e-05, "loss": 0.7827, "step": 5145 }, { "epoch": 0.8, "grad_norm": 29.835625627140793, "learning_rate": 1.7175515824949304e-05, "loss": 0.721, "step": 5146 }, { "epoch": 0.8, "grad_norm": 20.555538623113, "learning_rate": 1.7174341136982763e-05, "loss": 0.707, "step": 5147 }, { "epoch": 0.8, "grad_norm": 23.694979288158706, "learning_rate": 1.717316624498215e-05, "loss": 0.7533, "step": 5148 }, { "epoch": 0.8, "grad_norm": 14.30406713104073, "learning_rate": 1.7171991148980888e-05, "loss": 0.7635, "step": 5149 }, { "epoch": 0.8, "grad_norm": 15.469966176659288, "learning_rate": 1.717081584901239e-05, "loss": 0.6965, "step": 5150 }, { "epoch": 0.8, "grad_norm": 17.156253639248597, "learning_rate": 1.7169640345110088e-05, "loss": 0.7239, "step": 5151 }, { "epoch": 0.8, "grad_norm": 13.503101974870242, "learning_rate": 1.7168464637307408e-05, "loss": 0.7429, "step": 5152 }, { "epoch": 0.8, "grad_norm": 13.495866157500753, "learning_rate": 1.7167288725637786e-05, "loss": 0.7084, "step": 5153 }, { "epoch": 0.81, "grad_norm": 20.819476779676247, "learning_rate": 1.7166112610134665e-05, "loss": 0.7504, "step": 5154 }, { "epoch": 0.81, "grad_norm": 12.253693590638479, "learning_rate": 1.7164936290831494e-05, "loss": 0.7233, "step": 5155 }, { "epoch": 0.81, "grad_norm": 14.691760306642331, "learning_rate": 1.7163759767761727e-05, "loss": 0.6635, "step": 5156 }, { "epoch": 0.81, "grad_norm": 21.058269663818134, "learning_rate": 1.716258304095882e-05, "loss": 0.7584, "step": 5157 }, { "epoch": 0.81, "grad_norm": 21.380375913891744, "learning_rate": 1.716140611045625e-05, "loss": 0.7411, "step": 5158 }, { "epoch": 0.81, "grad_norm": 21.897040454233423, "learning_rate": 1.716022897628747e-05, "loss": 0.6835, "step": 5159 }, { "epoch": 0.81, "grad_norm": 16.92237858185876, "learning_rate": 1.7159051638485967e-05, "loss": 0.7479, "step": 5160 }, { "epoch": 0.81, "grad_norm": 25.31088891612536, "learning_rate": 1.715787409708523e-05, "loss": 0.8402, "step": 5161 }, { "epoch": 0.81, "grad_norm": 20.15829211025891, "learning_rate": 1.7156696352118735e-05, "loss": 0.762, "step": 5162 }, { "epoch": 0.81, "grad_norm": 14.895683898265137, "learning_rate": 1.7155518403619983e-05, "loss": 0.6637, "step": 5163 }, { "epoch": 0.81, "grad_norm": 19.200923417126887, "learning_rate": 1.7154340251622476e-05, "loss": 0.7266, "step": 5164 }, { "epoch": 0.81, "grad_norm": 15.680402059006028, "learning_rate": 1.7153161896159717e-05, "loss": 0.8016, "step": 5165 }, { "epoch": 0.81, "grad_norm": 22.499018163583965, "learning_rate": 1.7151983337265217e-05, "loss": 0.6742, "step": 5166 }, { "epoch": 0.81, "grad_norm": 12.139926349637031, "learning_rate": 1.7150804574972496e-05, "loss": 0.6492, "step": 5167 }, { "epoch": 0.81, "grad_norm": 17.32334262778569, "learning_rate": 1.7149625609315076e-05, "loss": 0.7421, "step": 5168 }, { "epoch": 0.81, "grad_norm": 15.758706509086169, "learning_rate": 1.7148446440326485e-05, "loss": 0.7625, "step": 5169 }, { "epoch": 0.81, "grad_norm": 19.378582217554925, "learning_rate": 1.714726706804026e-05, "loss": 0.7941, "step": 5170 }, { "epoch": 0.81, "grad_norm": 23.14861364873735, "learning_rate": 1.714608749248994e-05, "loss": 0.7745, "step": 5171 }, { "epoch": 0.81, "grad_norm": 12.184340133993782, "learning_rate": 1.714490771370907e-05, "loss": 0.6667, "step": 5172 }, { "epoch": 0.81, "grad_norm": 27.235216965958795, "learning_rate": 1.7143727731731204e-05, "loss": 0.8347, "step": 5173 }, { "epoch": 0.81, "grad_norm": 19.78384076910924, "learning_rate": 1.7142547546589903e-05, "loss": 0.7187, "step": 5174 }, { "epoch": 0.81, "grad_norm": 22.327430603182503, "learning_rate": 1.714136715831873e-05, "loss": 0.7614, "step": 5175 }, { "epoch": 0.81, "grad_norm": 21.341016486852183, "learning_rate": 1.7140186566951248e-05, "loss": 0.6902, "step": 5176 }, { "epoch": 0.81, "grad_norm": 18.309475983495677, "learning_rate": 1.7139005772521038e-05, "loss": 0.7205, "step": 5177 }, { "epoch": 0.81, "grad_norm": 18.225050607200238, "learning_rate": 1.7137824775061682e-05, "loss": 0.735, "step": 5178 }, { "epoch": 0.81, "grad_norm": 19.36372698497285, "learning_rate": 1.7136643574606764e-05, "loss": 0.7876, "step": 5179 }, { "epoch": 0.81, "grad_norm": 13.597623428331413, "learning_rate": 1.7135462171189877e-05, "loss": 0.6934, "step": 5180 }, { "epoch": 0.81, "grad_norm": 19.694557004463306, "learning_rate": 1.713428056484462e-05, "loss": 0.6687, "step": 5181 }, { "epoch": 0.81, "grad_norm": 23.611076331549317, "learning_rate": 1.7133098755604595e-05, "loss": 0.7197, "step": 5182 }, { "epoch": 0.81, "grad_norm": 16.26785181360096, "learning_rate": 1.713191674350342e-05, "loss": 0.7426, "step": 5183 }, { "epoch": 0.81, "grad_norm": 18.46083528019795, "learning_rate": 1.71307345285747e-05, "loss": 0.691, "step": 5184 }, { "epoch": 0.81, "grad_norm": 18.027515995777403, "learning_rate": 1.7129552110852064e-05, "loss": 0.7468, "step": 5185 }, { "epoch": 0.81, "grad_norm": 14.067577070744028, "learning_rate": 1.7128369490369134e-05, "loss": 0.6679, "step": 5186 }, { "epoch": 0.81, "grad_norm": 15.996270422140327, "learning_rate": 1.7127186667159547e-05, "loss": 0.7136, "step": 5187 }, { "epoch": 0.81, "grad_norm": 28.687060999991512, "learning_rate": 1.712600364125694e-05, "loss": 0.7427, "step": 5188 }, { "epoch": 0.81, "grad_norm": 20.87675124948807, "learning_rate": 1.712482041269496e-05, "loss": 0.7439, "step": 5189 }, { "epoch": 0.81, "grad_norm": 23.023500437888337, "learning_rate": 1.7123636981507256e-05, "loss": 0.7542, "step": 5190 }, { "epoch": 0.81, "grad_norm": 14.226060276080307, "learning_rate": 1.712245334772748e-05, "loss": 0.7316, "step": 5191 }, { "epoch": 0.81, "grad_norm": 21.60619267513184, "learning_rate": 1.7121269511389298e-05, "loss": 0.784, "step": 5192 }, { "epoch": 0.81, "grad_norm": 23.594555092016382, "learning_rate": 1.7120085472526377e-05, "loss": 0.6827, "step": 5193 }, { "epoch": 0.81, "grad_norm": 19.276446784427318, "learning_rate": 1.711890123117239e-05, "loss": 0.7894, "step": 5194 }, { "epoch": 0.81, "grad_norm": 18.86603492865205, "learning_rate": 1.711771678736102e-05, "loss": 0.7585, "step": 5195 }, { "epoch": 0.81, "grad_norm": 23.78643156013516, "learning_rate": 1.7116532141125947e-05, "loss": 0.7483, "step": 5196 }, { "epoch": 0.81, "grad_norm": 18.21745251915064, "learning_rate": 1.711534729250086e-05, "loss": 0.7554, "step": 5197 }, { "epoch": 0.81, "grad_norm": 20.214312822676618, "learning_rate": 1.711416224151946e-05, "loss": 0.6873, "step": 5198 }, { "epoch": 0.81, "grad_norm": 25.00368890249207, "learning_rate": 1.7112976988215445e-05, "loss": 0.6451, "step": 5199 }, { "epoch": 0.81, "grad_norm": 29.31716589810737, "learning_rate": 1.711179153262253e-05, "loss": 0.7897, "step": 5200 }, { "epoch": 0.81, "grad_norm": 24.645334688695545, "learning_rate": 1.711060587477442e-05, "loss": 0.8364, "step": 5201 }, { "epoch": 0.81, "grad_norm": 13.248071867907846, "learning_rate": 1.710942001470484e-05, "loss": 0.6988, "step": 5202 }, { "epoch": 0.81, "grad_norm": 25.189003249359537, "learning_rate": 1.7108233952447516e-05, "loss": 0.7757, "step": 5203 }, { "epoch": 0.81, "grad_norm": 23.825042063029482, "learning_rate": 1.7107047688036175e-05, "loss": 0.6276, "step": 5204 }, { "epoch": 0.81, "grad_norm": 19.303921080303425, "learning_rate": 1.7105861221504555e-05, "loss": 0.7765, "step": 5205 }, { "epoch": 0.81, "grad_norm": 20.345665834899343, "learning_rate": 1.71046745528864e-05, "loss": 0.7833, "step": 5206 }, { "epoch": 0.81, "grad_norm": 22.304084153708335, "learning_rate": 1.7103487682215454e-05, "loss": 0.8053, "step": 5207 }, { "epoch": 0.81, "grad_norm": 16.49489460334743, "learning_rate": 1.710230060952548e-05, "loss": 0.7643, "step": 5208 }, { "epoch": 0.81, "grad_norm": 26.111518871150807, "learning_rate": 1.7101113334850224e-05, "loss": 0.7566, "step": 5209 }, { "epoch": 0.81, "grad_norm": 26.556357640680986, "learning_rate": 1.709992585822346e-05, "loss": 0.8432, "step": 5210 }, { "epoch": 0.81, "grad_norm": 18.020274380374257, "learning_rate": 1.709873817967896e-05, "loss": 0.7846, "step": 5211 }, { "epoch": 0.81, "grad_norm": 16.61036679706035, "learning_rate": 1.70975502992505e-05, "loss": 0.7227, "step": 5212 }, { "epoch": 0.81, "grad_norm": 17.708249171187916, "learning_rate": 1.7096362216971857e-05, "loss": 0.7751, "step": 5213 }, { "epoch": 0.81, "grad_norm": 19.708748120169744, "learning_rate": 1.7095173932876827e-05, "loss": 0.7299, "step": 5214 }, { "epoch": 0.81, "grad_norm": 18.909691180597793, "learning_rate": 1.7093985446999203e-05, "loss": 0.6862, "step": 5215 }, { "epoch": 0.81, "grad_norm": 16.467779997791684, "learning_rate": 1.7092796759372778e-05, "loss": 0.7572, "step": 5216 }, { "epoch": 0.81, "grad_norm": 19.466025192414246, "learning_rate": 1.7091607870031362e-05, "loss": 0.8308, "step": 5217 }, { "epoch": 0.82, "grad_norm": 23.435405656064976, "learning_rate": 1.7090418779008772e-05, "loss": 0.6645, "step": 5218 }, { "epoch": 0.82, "grad_norm": 17.10645770933609, "learning_rate": 1.7089229486338815e-05, "loss": 0.6955, "step": 5219 }, { "epoch": 0.82, "grad_norm": 15.041225599807925, "learning_rate": 1.708803999205532e-05, "loss": 0.7755, "step": 5220 }, { "epoch": 0.82, "grad_norm": 24.964851359696137, "learning_rate": 1.7086850296192118e-05, "loss": 0.7442, "step": 5221 }, { "epoch": 0.82, "grad_norm": 19.45873126706079, "learning_rate": 1.7085660398783032e-05, "loss": 0.7241, "step": 5222 }, { "epoch": 0.82, "grad_norm": 25.149234360924705, "learning_rate": 1.7084470299861915e-05, "loss": 0.6826, "step": 5223 }, { "epoch": 0.82, "grad_norm": 26.213680701389535, "learning_rate": 1.7083279999462603e-05, "loss": 0.8048, "step": 5224 }, { "epoch": 0.82, "grad_norm": 18.257926227654465, "learning_rate": 1.7082089497618955e-05, "loss": 0.7918, "step": 5225 }, { "epoch": 0.82, "grad_norm": 22.216825118832926, "learning_rate": 1.708089879436482e-05, "loss": 0.7294, "step": 5226 }, { "epoch": 0.82, "grad_norm": 18.91550840486508, "learning_rate": 1.707970788973407e-05, "loss": 0.6509, "step": 5227 }, { "epoch": 0.82, "grad_norm": 23.22206902174581, "learning_rate": 1.7078516783760568e-05, "loss": 0.7002, "step": 5228 }, { "epoch": 0.82, "grad_norm": 21.471909689941164, "learning_rate": 1.707732547647819e-05, "loss": 0.7251, "step": 5229 }, { "epoch": 0.82, "grad_norm": 24.20088935978388, "learning_rate": 1.7076133967920813e-05, "loss": 0.7751, "step": 5230 }, { "epoch": 0.82, "grad_norm": 30.134268155096944, "learning_rate": 1.7074942258122326e-05, "loss": 0.7393, "step": 5231 }, { "epoch": 0.82, "grad_norm": 30.24635652893284, "learning_rate": 1.707375034711662e-05, "loss": 0.7391, "step": 5232 }, { "epoch": 0.82, "grad_norm": 17.511066867966253, "learning_rate": 1.7072558234937597e-05, "loss": 0.7057, "step": 5233 }, { "epoch": 0.82, "grad_norm": 20.00314918847954, "learning_rate": 1.707136592161915e-05, "loss": 0.7347, "step": 5234 }, { "epoch": 0.82, "grad_norm": 27.868116686264358, "learning_rate": 1.7070173407195193e-05, "loss": 0.8165, "step": 5235 }, { "epoch": 0.82, "grad_norm": 18.786545019757952, "learning_rate": 1.7068980691699638e-05, "loss": 0.7533, "step": 5236 }, { "epoch": 0.82, "grad_norm": 18.024889323655994, "learning_rate": 1.7067787775166414e-05, "loss": 0.7822, "step": 5237 }, { "epoch": 0.82, "grad_norm": 26.497529064221553, "learning_rate": 1.7066594657629435e-05, "loss": 0.6936, "step": 5238 }, { "epoch": 0.82, "grad_norm": 14.644817973230241, "learning_rate": 1.706540133912264e-05, "loss": 0.7498, "step": 5239 }, { "epoch": 0.82, "grad_norm": 18.682022740892513, "learning_rate": 1.7064207819679964e-05, "loss": 0.8008, "step": 5240 }, { "epoch": 0.82, "grad_norm": 15.142779933096879, "learning_rate": 1.7063014099335353e-05, "loss": 0.7062, "step": 5241 }, { "epoch": 0.82, "grad_norm": 29.532680526276256, "learning_rate": 1.7061820178122746e-05, "loss": 0.8107, "step": 5242 }, { "epoch": 0.82, "grad_norm": 18.24548046361106, "learning_rate": 1.7060626056076107e-05, "loss": 0.7246, "step": 5243 }, { "epoch": 0.82, "grad_norm": 33.7868676935813, "learning_rate": 1.705943173322939e-05, "loss": 0.8513, "step": 5244 }, { "epoch": 0.82, "grad_norm": 17.52759771671025, "learning_rate": 1.705823720961657e-05, "loss": 0.7322, "step": 5245 }, { "epoch": 0.82, "grad_norm": 16.44824281848892, "learning_rate": 1.705704248527161e-05, "loss": 0.7418, "step": 5246 }, { "epoch": 0.82, "grad_norm": 19.02752852306113, "learning_rate": 1.7055847560228495e-05, "loss": 0.7536, "step": 5247 }, { "epoch": 0.82, "grad_norm": 22.9162490112364, "learning_rate": 1.7054652434521197e-05, "loss": 0.6988, "step": 5248 }, { "epoch": 0.82, "grad_norm": 17.729013259289736, "learning_rate": 1.7053457108183713e-05, "loss": 0.7439, "step": 5249 }, { "epoch": 0.82, "grad_norm": 15.012753659661463, "learning_rate": 1.7052261581250034e-05, "loss": 0.7762, "step": 5250 }, { "epoch": 0.82, "grad_norm": 20.4150008472449, "learning_rate": 1.705106585375416e-05, "loss": 0.7424, "step": 5251 }, { "epoch": 0.82, "grad_norm": 28.2915625090407, "learning_rate": 1.70498699257301e-05, "loss": 0.7448, "step": 5252 }, { "epoch": 0.82, "grad_norm": 19.18176275901338, "learning_rate": 1.7048673797211863e-05, "loss": 0.7461, "step": 5253 }, { "epoch": 0.82, "grad_norm": 22.212802045895916, "learning_rate": 1.7047477468233467e-05, "loss": 0.8229, "step": 5254 }, { "epoch": 0.82, "grad_norm": 15.20314254348072, "learning_rate": 1.7046280938828932e-05, "loss": 0.7776, "step": 5255 }, { "epoch": 0.82, "grad_norm": 15.645115960251603, "learning_rate": 1.704508420903229e-05, "loss": 0.6577, "step": 5256 }, { "epoch": 0.82, "grad_norm": 25.789337101271453, "learning_rate": 1.7043887278877577e-05, "loss": 0.6788, "step": 5257 }, { "epoch": 0.82, "grad_norm": 20.30974265603834, "learning_rate": 1.7042690148398825e-05, "loss": 0.8397, "step": 5258 }, { "epoch": 0.82, "grad_norm": 18.83332377320303, "learning_rate": 1.704149281763009e-05, "loss": 0.7066, "step": 5259 }, { "epoch": 0.82, "grad_norm": 25.308804916958696, "learning_rate": 1.7040295286605415e-05, "loss": 0.7674, "step": 5260 }, { "epoch": 0.82, "grad_norm": 26.81848377757063, "learning_rate": 1.7039097555358862e-05, "loss": 0.8434, "step": 5261 }, { "epoch": 0.82, "grad_norm": 22.03986832890927, "learning_rate": 1.7037899623924495e-05, "loss": 0.7236, "step": 5262 }, { "epoch": 0.82, "grad_norm": 22.090543859637418, "learning_rate": 1.7036701492336378e-05, "loss": 0.7969, "step": 5263 }, { "epoch": 0.82, "grad_norm": 19.970197314733156, "learning_rate": 1.7035503160628585e-05, "loss": 0.7631, "step": 5264 }, { "epoch": 0.82, "grad_norm": 15.222801017904905, "learning_rate": 1.7034304628835198e-05, "loss": 0.6801, "step": 5265 }, { "epoch": 0.82, "grad_norm": 18.482936078396236, "learning_rate": 1.70331058969903e-05, "loss": 0.7057, "step": 5266 }, { "epoch": 0.82, "grad_norm": 17.027679157388516, "learning_rate": 1.703190696512799e-05, "loss": 0.6521, "step": 5267 }, { "epoch": 0.82, "grad_norm": 22.23955269185322, "learning_rate": 1.703070783328236e-05, "loss": 0.8073, "step": 5268 }, { "epoch": 0.82, "grad_norm": 14.96725247722746, "learning_rate": 1.702950850148751e-05, "loss": 0.6897, "step": 5269 }, { "epoch": 0.82, "grad_norm": 21.440301662020264, "learning_rate": 1.7028308969777548e-05, "loss": 0.7676, "step": 5270 }, { "epoch": 0.82, "grad_norm": 18.081220980339747, "learning_rate": 1.7027109238186593e-05, "loss": 0.753, "step": 5271 }, { "epoch": 0.82, "grad_norm": 16.28191990490078, "learning_rate": 1.7025909306748757e-05, "loss": 0.7668, "step": 5272 }, { "epoch": 0.82, "grad_norm": 8.205375073967923, "learning_rate": 1.7024709175498174e-05, "loss": 0.6975, "step": 5273 }, { "epoch": 0.82, "grad_norm": 21.929077265993786, "learning_rate": 1.7023508844468974e-05, "loss": 0.6462, "step": 5274 }, { "epoch": 0.82, "grad_norm": 18.905851414711375, "learning_rate": 1.702230831369529e-05, "loss": 0.8903, "step": 5275 }, { "epoch": 0.82, "grad_norm": 16.44763775331927, "learning_rate": 1.7021107583211266e-05, "loss": 0.7497, "step": 5276 }, { "epoch": 0.82, "grad_norm": 21.32928537279355, "learning_rate": 1.7019906653051045e-05, "loss": 0.7195, "step": 5277 }, { "epoch": 0.82, "grad_norm": 23.683651035194888, "learning_rate": 1.701870552324879e-05, "loss": 0.7452, "step": 5278 }, { "epoch": 0.82, "grad_norm": 17.664000125118573, "learning_rate": 1.7017504193838654e-05, "loss": 0.7115, "step": 5279 }, { "epoch": 0.82, "grad_norm": 17.395220431268605, "learning_rate": 1.7016302664854802e-05, "loss": 0.728, "step": 5280 }, { "epoch": 0.82, "grad_norm": 16.27474301998892, "learning_rate": 1.701510093633141e-05, "loss": 0.6826, "step": 5281 }, { "epoch": 0.83, "grad_norm": 20.15930575118074, "learning_rate": 1.701389900830265e-05, "loss": 0.8046, "step": 5282 }, { "epoch": 0.83, "grad_norm": 19.83775068066402, "learning_rate": 1.70126968808027e-05, "loss": 0.7946, "step": 5283 }, { "epoch": 0.83, "grad_norm": 15.934965622871642, "learning_rate": 1.7011494553865762e-05, "loss": 0.7219, "step": 5284 }, { "epoch": 0.83, "grad_norm": 18.094198983144597, "learning_rate": 1.7010292027526016e-05, "loss": 0.6905, "step": 5285 }, { "epoch": 0.83, "grad_norm": 16.386529025228718, "learning_rate": 1.7009089301817666e-05, "loss": 0.7556, "step": 5286 }, { "epoch": 0.83, "grad_norm": 16.50589419609408, "learning_rate": 1.7007886376774917e-05, "loss": 0.7015, "step": 5287 }, { "epoch": 0.83, "grad_norm": 17.505807325138306, "learning_rate": 1.7006683252431977e-05, "loss": 0.6908, "step": 5288 }, { "epoch": 0.83, "grad_norm": 18.468875830523434, "learning_rate": 1.7005479928823066e-05, "loss": 0.7265, "step": 5289 }, { "epoch": 0.83, "grad_norm": 17.890223480283264, "learning_rate": 1.7004276405982404e-05, "loss": 0.7221, "step": 5290 }, { "epoch": 0.83, "grad_norm": 19.138049348636994, "learning_rate": 1.7003072683944214e-05, "loss": 0.7783, "step": 5291 }, { "epoch": 0.83, "grad_norm": 17.008168949672765, "learning_rate": 1.700186876274274e-05, "loss": 0.803, "step": 5292 }, { "epoch": 0.83, "grad_norm": 21.046225805401793, "learning_rate": 1.700066464241221e-05, "loss": 0.8066, "step": 5293 }, { "epoch": 0.83, "grad_norm": 11.41782958252252, "learning_rate": 1.6999460322986876e-05, "loss": 0.7059, "step": 5294 }, { "epoch": 0.83, "grad_norm": 15.223645780850129, "learning_rate": 1.6998255804500983e-05, "loss": 0.7338, "step": 5295 }, { "epoch": 0.83, "grad_norm": 18.688801233732843, "learning_rate": 1.699705108698879e-05, "loss": 0.7164, "step": 5296 }, { "epoch": 0.83, "grad_norm": 14.211439051488393, "learning_rate": 1.6995846170484554e-05, "loss": 0.7625, "step": 5297 }, { "epoch": 0.83, "grad_norm": 16.793240609090326, "learning_rate": 1.699464105502255e-05, "loss": 0.7275, "step": 5298 }, { "epoch": 0.83, "grad_norm": 31.8135925809831, "learning_rate": 1.6993435740637042e-05, "loss": 0.8877, "step": 5299 }, { "epoch": 0.83, "grad_norm": 14.932537832555084, "learning_rate": 1.6992230227362317e-05, "loss": 0.6649, "step": 5300 }, { "epoch": 0.83, "grad_norm": 14.850596382446989, "learning_rate": 1.6991024515232652e-05, "loss": 0.7352, "step": 5301 }, { "epoch": 0.83, "grad_norm": 19.527752671497687, "learning_rate": 1.698981860428234e-05, "loss": 0.7362, "step": 5302 }, { "epoch": 0.83, "grad_norm": 12.234730511952906, "learning_rate": 1.6988612494545673e-05, "loss": 0.6077, "step": 5303 }, { "epoch": 0.83, "grad_norm": 15.151526856633746, "learning_rate": 1.698740618605696e-05, "loss": 0.7386, "step": 5304 }, { "epoch": 0.83, "grad_norm": 17.161854040613125, "learning_rate": 1.69861996788505e-05, "loss": 0.6559, "step": 5305 }, { "epoch": 0.83, "grad_norm": 21.160182836669758, "learning_rate": 1.6984992972960606e-05, "loss": 0.756, "step": 5306 }, { "epoch": 0.83, "grad_norm": 13.711337702091864, "learning_rate": 1.69837860684216e-05, "loss": 0.6666, "step": 5307 }, { "epoch": 0.83, "grad_norm": 15.52688328378112, "learning_rate": 1.6982578965267805e-05, "loss": 0.7041, "step": 5308 }, { "epoch": 0.83, "grad_norm": 16.04005928769861, "learning_rate": 1.6981371663533543e-05, "loss": 0.7199, "step": 5309 }, { "epoch": 0.83, "grad_norm": 15.322963190580786, "learning_rate": 1.698016416325316e-05, "loss": 0.7607, "step": 5310 }, { "epoch": 0.83, "grad_norm": 23.216787203310748, "learning_rate": 1.6978956464460993e-05, "loss": 0.7545, "step": 5311 }, { "epoch": 0.83, "grad_norm": 16.328090675901993, "learning_rate": 1.6977748567191384e-05, "loss": 0.729, "step": 5312 }, { "epoch": 0.83, "grad_norm": 19.564781120714, "learning_rate": 1.697654047147869e-05, "loss": 0.7159, "step": 5313 }, { "epoch": 0.83, "grad_norm": 22.83253972749377, "learning_rate": 1.6975332177357258e-05, "loss": 0.8002, "step": 5314 }, { "epoch": 0.83, "grad_norm": 26.683272427810657, "learning_rate": 1.697412368486147e-05, "loss": 0.7561, "step": 5315 }, { "epoch": 0.83, "grad_norm": 26.439257556016987, "learning_rate": 1.697291499402568e-05, "loss": 0.7028, "step": 5316 }, { "epoch": 0.83, "grad_norm": 21.944124655796955, "learning_rate": 1.6971706104884263e-05, "loss": 0.7289, "step": 5317 }, { "epoch": 0.83, "grad_norm": 18.98941214379324, "learning_rate": 1.6970497017471608e-05, "loss": 0.694, "step": 5318 }, { "epoch": 0.83, "grad_norm": 25.85206005266668, "learning_rate": 1.6969287731822092e-05, "loss": 0.7589, "step": 5319 }, { "epoch": 0.83, "grad_norm": 25.109176435557835, "learning_rate": 1.6968078247970116e-05, "loss": 0.8027, "step": 5320 }, { "epoch": 0.83, "grad_norm": 17.40351651197082, "learning_rate": 1.6966868565950062e-05, "loss": 0.696, "step": 5321 }, { "epoch": 0.83, "grad_norm": 33.89123415982081, "learning_rate": 1.6965658685796347e-05, "loss": 0.8422, "step": 5322 }, { "epoch": 0.83, "grad_norm": 18.86977291992447, "learning_rate": 1.6964448607543375e-05, "loss": 0.7781, "step": 5323 }, { "epoch": 0.83, "grad_norm": 21.757045640563643, "learning_rate": 1.696323833122555e-05, "loss": 0.7318, "step": 5324 }, { "epoch": 0.83, "grad_norm": 19.32630022111386, "learning_rate": 1.696202785687731e-05, "loss": 0.7198, "step": 5325 }, { "epoch": 0.83, "grad_norm": 15.799283365704994, "learning_rate": 1.6960817184533068e-05, "loss": 0.7747, "step": 5326 }, { "epoch": 0.83, "grad_norm": 20.77355076994614, "learning_rate": 1.6959606314227254e-05, "loss": 0.7008, "step": 5327 }, { "epoch": 0.83, "grad_norm": 4.142022414391043, "learning_rate": 1.695839524599431e-05, "loss": 0.8104, "step": 5328 }, { "epoch": 0.83, "grad_norm": 23.127038314271974, "learning_rate": 1.6957183979868673e-05, "loss": 0.7259, "step": 5329 }, { "epoch": 0.83, "grad_norm": 13.896891192166652, "learning_rate": 1.6955972515884798e-05, "loss": 0.6573, "step": 5330 }, { "epoch": 0.83, "grad_norm": 27.293947471736764, "learning_rate": 1.695476085407713e-05, "loss": 0.804, "step": 5331 }, { "epoch": 0.83, "grad_norm": 23.47729253411391, "learning_rate": 1.695354899448013e-05, "loss": 0.8239, "step": 5332 }, { "epoch": 0.83, "grad_norm": 29.055762164147385, "learning_rate": 1.695233693712827e-05, "loss": 0.7221, "step": 5333 }, { "epoch": 0.83, "grad_norm": 24.08650277415116, "learning_rate": 1.695112468205601e-05, "loss": 0.7172, "step": 5334 }, { "epoch": 0.83, "grad_norm": 24.250135975018978, "learning_rate": 1.6949912229297834e-05, "loss": 0.6906, "step": 5335 }, { "epoch": 0.83, "grad_norm": 21.04451229679336, "learning_rate": 1.6948699578888218e-05, "loss": 0.6787, "step": 5336 }, { "epoch": 0.83, "grad_norm": 11.778587713069367, "learning_rate": 1.6947486730861652e-05, "loss": 0.5909, "step": 5337 }, { "epoch": 0.83, "grad_norm": 16.90332788855892, "learning_rate": 1.694627368525263e-05, "loss": 0.6412, "step": 5338 }, { "epoch": 0.83, "grad_norm": 19.719465036471096, "learning_rate": 1.6945060442095642e-05, "loss": 0.7676, "step": 5339 }, { "epoch": 0.83, "grad_norm": 16.381134694038188, "learning_rate": 1.69438470014252e-05, "loss": 0.7599, "step": 5340 }, { "epoch": 0.83, "grad_norm": 22.198136504812283, "learning_rate": 1.6942633363275812e-05, "loss": 0.7734, "step": 5341 }, { "epoch": 0.83, "grad_norm": 19.840605630082592, "learning_rate": 1.694141952768199e-05, "loss": 0.6985, "step": 5342 }, { "epoch": 0.83, "grad_norm": 14.279210167727102, "learning_rate": 1.694020549467826e-05, "loss": 0.7811, "step": 5343 }, { "epoch": 0.83, "grad_norm": 19.04089971795267, "learning_rate": 1.6938991264299145e-05, "loss": 0.8073, "step": 5344 }, { "epoch": 0.83, "grad_norm": 15.756741233425233, "learning_rate": 1.693777683657918e-05, "loss": 0.7204, "step": 5345 }, { "epoch": 0.84, "grad_norm": 21.743478395737206, "learning_rate": 1.6936562211552895e-05, "loss": 0.7173, "step": 5346 }, { "epoch": 0.84, "grad_norm": 18.310044583768175, "learning_rate": 1.693534738925484e-05, "loss": 0.7546, "step": 5347 }, { "epoch": 0.84, "grad_norm": 17.8291652193765, "learning_rate": 1.6934132369719563e-05, "loss": 0.8646, "step": 5348 }, { "epoch": 0.84, "grad_norm": 13.072921132344042, "learning_rate": 1.693291715298162e-05, "loss": 0.6123, "step": 5349 }, { "epoch": 0.84, "grad_norm": 17.40706843792547, "learning_rate": 1.6931701739075568e-05, "loss": 0.7233, "step": 5350 }, { "epoch": 0.84, "grad_norm": 13.430716859603752, "learning_rate": 1.6930486128035974e-05, "loss": 0.7303, "step": 5351 }, { "epoch": 0.84, "grad_norm": 17.321403549064605, "learning_rate": 1.6929270319897407e-05, "loss": 0.796, "step": 5352 }, { "epoch": 0.84, "grad_norm": 35.42729013989812, "learning_rate": 1.6928054314694443e-05, "loss": 0.8944, "step": 5353 }, { "epoch": 0.84, "grad_norm": 11.109657402043945, "learning_rate": 1.6926838112461668e-05, "loss": 0.6913, "step": 5354 }, { "epoch": 0.84, "grad_norm": 24.642529293682223, "learning_rate": 1.692562171323367e-05, "loss": 0.7497, "step": 5355 }, { "epoch": 0.84, "grad_norm": 15.971605644649218, "learning_rate": 1.6924405117045046e-05, "loss": 0.7153, "step": 5356 }, { "epoch": 0.84, "grad_norm": 23.764910216620862, "learning_rate": 1.692318832393039e-05, "loss": 0.7474, "step": 5357 }, { "epoch": 0.84, "grad_norm": 17.43158527696729, "learning_rate": 1.6921971333924304e-05, "loss": 0.7534, "step": 5358 }, { "epoch": 0.84, "grad_norm": 18.385573703015968, "learning_rate": 1.6920754147061406e-05, "loss": 0.6754, "step": 5359 }, { "epoch": 0.84, "grad_norm": 19.13515113838597, "learning_rate": 1.691953676337631e-05, "loss": 0.7512, "step": 5360 }, { "epoch": 0.84, "grad_norm": 29.899136083319725, "learning_rate": 1.691831918290363e-05, "loss": 0.8366, "step": 5361 }, { "epoch": 0.84, "grad_norm": 18.289847191325563, "learning_rate": 1.6917101405678004e-05, "loss": 0.7825, "step": 5362 }, { "epoch": 0.84, "grad_norm": 18.41337792309947, "learning_rate": 1.691588343173406e-05, "loss": 0.7244, "step": 5363 }, { "epoch": 0.84, "grad_norm": 24.716262148435145, "learning_rate": 1.6914665261106434e-05, "loss": 0.8078, "step": 5364 }, { "epoch": 0.84, "grad_norm": 18.652928943341834, "learning_rate": 1.691344689382978e-05, "loss": 0.7657, "step": 5365 }, { "epoch": 0.84, "grad_norm": 20.2795215187537, "learning_rate": 1.6912228329938734e-05, "loss": 0.7703, "step": 5366 }, { "epoch": 0.84, "grad_norm": 13.675518267624597, "learning_rate": 1.691100956946796e-05, "loss": 0.6667, "step": 5367 }, { "epoch": 0.84, "grad_norm": 12.976858003617421, "learning_rate": 1.6909790612452114e-05, "loss": 0.6166, "step": 5368 }, { "epoch": 0.84, "grad_norm": 18.270833606640224, "learning_rate": 1.690857145892587e-05, "loss": 0.6824, "step": 5369 }, { "epoch": 0.84, "grad_norm": 23.864918181674533, "learning_rate": 1.6907352108923892e-05, "loss": 0.7064, "step": 5370 }, { "epoch": 0.84, "grad_norm": 29.027192003091926, "learning_rate": 1.690613256248086e-05, "loss": 0.7339, "step": 5371 }, { "epoch": 0.84, "grad_norm": 19.325723336958, "learning_rate": 1.690491281963146e-05, "loss": 0.7569, "step": 5372 }, { "epoch": 0.84, "grad_norm": 17.778871219588932, "learning_rate": 1.690369288041038e-05, "loss": 0.6858, "step": 5373 }, { "epoch": 0.84, "grad_norm": 21.28704337037985, "learning_rate": 1.690247274485231e-05, "loss": 0.7345, "step": 5374 }, { "epoch": 0.84, "grad_norm": 17.14197846774523, "learning_rate": 1.690125241299195e-05, "loss": 0.8074, "step": 5375 }, { "epoch": 0.84, "grad_norm": 11.866728701584494, "learning_rate": 1.6900031884864013e-05, "loss": 0.7064, "step": 5376 }, { "epoch": 0.84, "grad_norm": 12.400844693638945, "learning_rate": 1.6898811160503204e-05, "loss": 0.7522, "step": 5377 }, { "epoch": 0.84, "grad_norm": 28.037670620170804, "learning_rate": 1.6897590239944242e-05, "loss": 0.8515, "step": 5378 }, { "epoch": 0.84, "grad_norm": 34.703549224891866, "learning_rate": 1.6896369123221852e-05, "loss": 0.7814, "step": 5379 }, { "epoch": 0.84, "grad_norm": 17.171722151099456, "learning_rate": 1.6895147810370755e-05, "loss": 0.6885, "step": 5380 }, { "epoch": 0.84, "grad_norm": 21.14283453523767, "learning_rate": 1.6893926301425685e-05, "loss": 0.7007, "step": 5381 }, { "epoch": 0.84, "grad_norm": 18.788615669917164, "learning_rate": 1.6892704596421386e-05, "loss": 0.7271, "step": 5382 }, { "epoch": 0.84, "grad_norm": 17.446338854851334, "learning_rate": 1.6891482695392603e-05, "loss": 0.7586, "step": 5383 }, { "epoch": 0.84, "grad_norm": 15.167327045008195, "learning_rate": 1.6890260598374083e-05, "loss": 0.6385, "step": 5384 }, { "epoch": 0.84, "grad_norm": 9.075636376512431, "learning_rate": 1.688903830540058e-05, "loss": 0.5901, "step": 5385 }, { "epoch": 0.84, "grad_norm": 14.96259297581637, "learning_rate": 1.6887815816506858e-05, "loss": 0.7269, "step": 5386 }, { "epoch": 0.84, "grad_norm": 14.875292372935663, "learning_rate": 1.6886593131727687e-05, "loss": 0.6637, "step": 5387 }, { "epoch": 0.84, "grad_norm": 17.162009929500407, "learning_rate": 1.688537025109783e-05, "loss": 0.6942, "step": 5388 }, { "epoch": 0.84, "grad_norm": 19.860024641923033, "learning_rate": 1.6884147174652077e-05, "loss": 0.8074, "step": 5389 }, { "epoch": 0.84, "grad_norm": 17.332970647263213, "learning_rate": 1.68829239024252e-05, "loss": 0.6906, "step": 5390 }, { "epoch": 0.84, "grad_norm": 14.801020061311059, "learning_rate": 1.6881700434451996e-05, "loss": 0.6766, "step": 5391 }, { "epoch": 0.84, "grad_norm": 30.397282887052942, "learning_rate": 1.6880476770767256e-05, "loss": 0.7628, "step": 5392 }, { "epoch": 0.84, "grad_norm": 23.59440561482592, "learning_rate": 1.6879252911405782e-05, "loss": 0.7597, "step": 5393 }, { "epoch": 0.84, "grad_norm": 15.572461786664956, "learning_rate": 1.6878028856402382e-05, "loss": 0.5419, "step": 5394 }, { "epoch": 0.84, "grad_norm": 17.872457975420073, "learning_rate": 1.6876804605791864e-05, "loss": 0.7453, "step": 5395 }, { "epoch": 0.84, "grad_norm": 19.52819200151108, "learning_rate": 1.6875580159609044e-05, "loss": 0.65, "step": 5396 }, { "epoch": 0.84, "grad_norm": 12.497709574665823, "learning_rate": 1.6874355517888747e-05, "loss": 0.7285, "step": 5397 }, { "epoch": 0.84, "grad_norm": 18.48034715582227, "learning_rate": 1.6873130680665798e-05, "loss": 0.7966, "step": 5398 }, { "epoch": 0.84, "grad_norm": 16.77543020116055, "learning_rate": 1.6871905647975038e-05, "loss": 0.6778, "step": 5399 }, { "epoch": 0.84, "grad_norm": 18.77171730148631, "learning_rate": 1.6870680419851297e-05, "loss": 0.7279, "step": 5400 }, { "epoch": 0.84, "grad_norm": 23.7211224584472, "learning_rate": 1.686945499632943e-05, "loss": 0.8393, "step": 5401 }, { "epoch": 0.84, "grad_norm": 24.29289405996517, "learning_rate": 1.6868229377444276e-05, "loss": 0.6865, "step": 5402 }, { "epoch": 0.84, "grad_norm": 15.51783646692002, "learning_rate": 1.6867003563230697e-05, "loss": 0.7408, "step": 5403 }, { "epoch": 0.84, "grad_norm": 17.786608894086974, "learning_rate": 1.6865777553723558e-05, "loss": 0.676, "step": 5404 }, { "epoch": 0.84, "grad_norm": 18.13278356432301, "learning_rate": 1.6864551348957717e-05, "loss": 0.723, "step": 5405 }, { "epoch": 0.84, "grad_norm": 13.032227648738216, "learning_rate": 1.686332494896805e-05, "loss": 0.7232, "step": 5406 }, { "epoch": 0.84, "grad_norm": 13.030171626454523, "learning_rate": 1.686209835378944e-05, "loss": 0.8027, "step": 5407 }, { "epoch": 0.84, "grad_norm": 14.975415536603201, "learning_rate": 1.6860871563456766e-05, "loss": 0.7442, "step": 5408 }, { "epoch": 0.84, "grad_norm": 19.163654531435537, "learning_rate": 1.685964457800492e-05, "loss": 0.7449, "step": 5409 }, { "epoch": 0.85, "grad_norm": 13.836815275514981, "learning_rate": 1.6858417397468792e-05, "loss": 0.6989, "step": 5410 }, { "epoch": 0.85, "grad_norm": 16.020592201147846, "learning_rate": 1.6857190021883287e-05, "loss": 0.7022, "step": 5411 }, { "epoch": 0.85, "grad_norm": 21.64287639089137, "learning_rate": 1.685596245128331e-05, "loss": 0.6749, "step": 5412 }, { "epoch": 0.85, "grad_norm": 23.36718262545153, "learning_rate": 1.685473468570377e-05, "loss": 0.746, "step": 5413 }, { "epoch": 0.85, "grad_norm": 147.37386729526654, "learning_rate": 1.6853506725179584e-05, "loss": 0.6308, "step": 5414 }, { "epoch": 0.85, "grad_norm": 19.43013961996076, "learning_rate": 1.6852278569745678e-05, "loss": 0.8726, "step": 5415 }, { "epoch": 0.85, "grad_norm": 17.542861201709016, "learning_rate": 1.685105021943698e-05, "loss": 0.6851, "step": 5416 }, { "epoch": 0.85, "grad_norm": 19.034092360070094, "learning_rate": 1.6849821674288418e-05, "loss": 0.6692, "step": 5417 }, { "epoch": 0.85, "grad_norm": 21.483809558016567, "learning_rate": 1.6848592934334934e-05, "loss": 0.6107, "step": 5418 }, { "epoch": 0.85, "grad_norm": 23.327963325773197, "learning_rate": 1.6847363999611475e-05, "loss": 0.8643, "step": 5419 }, { "epoch": 0.85, "grad_norm": 23.43470713696328, "learning_rate": 1.6846134870152987e-05, "loss": 0.6726, "step": 5420 }, { "epoch": 0.85, "grad_norm": 25.91558776474682, "learning_rate": 1.684490554599443e-05, "loss": 0.7237, "step": 5421 }, { "epoch": 0.85, "grad_norm": 20.068149151594326, "learning_rate": 1.6843676027170764e-05, "loss": 0.7284, "step": 5422 }, { "epoch": 0.85, "grad_norm": 22.815319879861377, "learning_rate": 1.6842446313716957e-05, "loss": 0.7744, "step": 5423 }, { "epoch": 0.85, "grad_norm": 21.963717502821574, "learning_rate": 1.6841216405667976e-05, "loss": 0.7257, "step": 5424 }, { "epoch": 0.85, "grad_norm": 24.63696057715217, "learning_rate": 1.6839986303058803e-05, "loss": 0.7225, "step": 5425 }, { "epoch": 0.85, "grad_norm": 15.247224644113391, "learning_rate": 1.6838756005924425e-05, "loss": 0.7258, "step": 5426 }, { "epoch": 0.85, "grad_norm": 14.773143098158165, "learning_rate": 1.6837525514299823e-05, "loss": 0.7017, "step": 5427 }, { "epoch": 0.85, "grad_norm": 14.448104473049431, "learning_rate": 1.6836294828219997e-05, "loss": 0.6968, "step": 5428 }, { "epoch": 0.85, "grad_norm": 23.374370569026592, "learning_rate": 1.6835063947719943e-05, "loss": 0.7632, "step": 5429 }, { "epoch": 0.85, "grad_norm": 21.36995378960079, "learning_rate": 1.683383287283467e-05, "loss": 0.7148, "step": 5430 }, { "epoch": 0.85, "grad_norm": 18.512365207142757, "learning_rate": 1.683260160359919e-05, "loss": 0.7483, "step": 5431 }, { "epoch": 0.85, "grad_norm": 47.606828254324384, "learning_rate": 1.6831370140048513e-05, "loss": 0.7077, "step": 5432 }, { "epoch": 0.85, "grad_norm": 14.440612262962913, "learning_rate": 1.6830138482217667e-05, "loss": 0.6902, "step": 5433 }, { "epoch": 0.85, "grad_norm": 27.612076012484934, "learning_rate": 1.6828906630141678e-05, "loss": 0.7381, "step": 5434 }, { "epoch": 0.85, "grad_norm": 15.960302204889638, "learning_rate": 1.682767458385558e-05, "loss": 0.7493, "step": 5435 }, { "epoch": 0.85, "grad_norm": 23.36174247061062, "learning_rate": 1.682644234339441e-05, "loss": 0.777, "step": 5436 }, { "epoch": 0.85, "grad_norm": 16.07842362981971, "learning_rate": 1.6825209908793217e-05, "loss": 0.6903, "step": 5437 }, { "epoch": 0.85, "grad_norm": 20.78099443880849, "learning_rate": 1.682397728008704e-05, "loss": 0.7493, "step": 5438 }, { "epoch": 0.85, "grad_norm": 24.07761205111984, "learning_rate": 1.6822744457310948e-05, "loss": 0.8222, "step": 5439 }, { "epoch": 0.85, "grad_norm": 16.066881692623568, "learning_rate": 1.682151144049999e-05, "loss": 0.7605, "step": 5440 }, { "epoch": 0.85, "grad_norm": 20.300542475118444, "learning_rate": 1.682027822968924e-05, "loss": 0.7117, "step": 5441 }, { "epoch": 0.85, "grad_norm": 32.28297065403295, "learning_rate": 1.6819044824913762e-05, "loss": 0.7171, "step": 5442 }, { "epoch": 0.85, "grad_norm": 26.16915410203333, "learning_rate": 1.6817811226208643e-05, "loss": 0.7454, "step": 5443 }, { "epoch": 0.85, "grad_norm": 20.004041628462364, "learning_rate": 1.681657743360896e-05, "loss": 0.6326, "step": 5444 }, { "epoch": 0.85, "grad_norm": 25.7243775616582, "learning_rate": 1.6815343447149803e-05, "loss": 0.6992, "step": 5445 }, { "epoch": 0.85, "grad_norm": 19.069653091375866, "learning_rate": 1.681410926686626e-05, "loss": 0.7899, "step": 5446 }, { "epoch": 0.85, "grad_norm": 16.86751397237882, "learning_rate": 1.6812874892793443e-05, "loss": 0.6546, "step": 5447 }, { "epoch": 0.85, "grad_norm": 14.301640517827305, "learning_rate": 1.6811640324966446e-05, "loss": 0.7471, "step": 5448 }, { "epoch": 0.85, "grad_norm": 26.96025782314002, "learning_rate": 1.681040556342038e-05, "loss": 0.7041, "step": 5449 }, { "epoch": 0.85, "grad_norm": 17.10642878801702, "learning_rate": 1.680917060819037e-05, "loss": 0.7166, "step": 5450 }, { "epoch": 0.85, "grad_norm": 20.825284947407024, "learning_rate": 1.6807935459311528e-05, "loss": 0.7401, "step": 5451 }, { "epoch": 0.85, "grad_norm": 24.84335733979094, "learning_rate": 1.6806700116818982e-05, "loss": 0.7844, "step": 5452 }, { "epoch": 0.85, "grad_norm": 28.750581994592896, "learning_rate": 1.680546458074787e-05, "loss": 0.6945, "step": 5453 }, { "epoch": 0.85, "grad_norm": 21.368784116409262, "learning_rate": 1.6804228851133326e-05, "loss": 0.7021, "step": 5454 }, { "epoch": 0.85, "grad_norm": 12.474745071948833, "learning_rate": 1.6802992928010496e-05, "loss": 0.7248, "step": 5455 }, { "epoch": 0.85, "grad_norm": 21.29534970388788, "learning_rate": 1.680175681141452e-05, "loss": 0.7332, "step": 5456 }, { "epoch": 0.85, "grad_norm": 20.158171505233792, "learning_rate": 1.6800520501380564e-05, "loss": 0.7238, "step": 5457 }, { "epoch": 0.85, "grad_norm": 18.72776420089135, "learning_rate": 1.679928399794378e-05, "loss": 0.7926, "step": 5458 }, { "epoch": 0.85, "grad_norm": 18.651488879121537, "learning_rate": 1.6798047301139338e-05, "loss": 0.7782, "step": 5459 }, { "epoch": 0.85, "grad_norm": 21.378991828550618, "learning_rate": 1.679681041100241e-05, "loss": 0.692, "step": 5460 }, { "epoch": 0.85, "grad_norm": 14.95584336348539, "learning_rate": 1.6795573327568168e-05, "loss": 0.7196, "step": 5461 }, { "epoch": 0.85, "grad_norm": 23.752709687151274, "learning_rate": 1.6794336050871797e-05, "loss": 0.6768, "step": 5462 }, { "epoch": 0.85, "grad_norm": 17.039901966332902, "learning_rate": 1.6793098580948482e-05, "loss": 0.7626, "step": 5463 }, { "epoch": 0.85, "grad_norm": 20.38153448188531, "learning_rate": 1.6791860917833417e-05, "loss": 0.7007, "step": 5464 }, { "epoch": 0.85, "grad_norm": 22.905846184974084, "learning_rate": 1.67906230615618e-05, "loss": 0.7494, "step": 5465 }, { "epoch": 0.85, "grad_norm": 14.117005847887409, "learning_rate": 1.6789385012168836e-05, "loss": 0.7091, "step": 5466 }, { "epoch": 0.85, "grad_norm": 16.86673968499415, "learning_rate": 1.6788146769689734e-05, "loss": 0.7533, "step": 5467 }, { "epoch": 0.85, "grad_norm": 32.723153092673385, "learning_rate": 1.678690833415971e-05, "loss": 0.7846, "step": 5468 }, { "epoch": 0.85, "grad_norm": 31.071843660843896, "learning_rate": 1.678566970561398e-05, "loss": 0.8239, "step": 5469 }, { "epoch": 0.85, "grad_norm": 18.769982936477426, "learning_rate": 1.678443088408778e-05, "loss": 0.7651, "step": 5470 }, { "epoch": 0.85, "grad_norm": 15.14182436378412, "learning_rate": 1.6783191869616327e-05, "loss": 0.678, "step": 5471 }, { "epoch": 0.85, "grad_norm": 17.497020667499044, "learning_rate": 1.678195266223487e-05, "loss": 0.6751, "step": 5472 }, { "epoch": 0.85, "grad_norm": 28.61302556831773, "learning_rate": 1.6780713261978646e-05, "loss": 0.742, "step": 5473 }, { "epoch": 0.86, "grad_norm": 18.251215526847925, "learning_rate": 1.67794736688829e-05, "loss": 0.7207, "step": 5474 }, { "epoch": 0.86, "grad_norm": 20.657820768118793, "learning_rate": 1.6778233882982894e-05, "loss": 0.7102, "step": 5475 }, { "epoch": 0.86, "grad_norm": 28.817946007768445, "learning_rate": 1.6776993904313875e-05, "loss": 0.6995, "step": 5476 }, { "epoch": 0.86, "grad_norm": 17.36170447777714, "learning_rate": 1.677575373291112e-05, "loss": 0.8064, "step": 5477 }, { "epoch": 0.86, "grad_norm": 16.105219809241028, "learning_rate": 1.6774513368809887e-05, "loss": 0.8047, "step": 5478 }, { "epoch": 0.86, "grad_norm": 17.222777176517035, "learning_rate": 1.677327281204546e-05, "loss": 0.6031, "step": 5479 }, { "epoch": 0.86, "grad_norm": 17.26237220675424, "learning_rate": 1.6772032062653115e-05, "loss": 0.723, "step": 5480 }, { "epoch": 0.86, "grad_norm": 26.956443900727262, "learning_rate": 1.677079112066814e-05, "loss": 0.6891, "step": 5481 }, { "epoch": 0.86, "grad_norm": 27.18991411715872, "learning_rate": 1.6769549986125827e-05, "loss": 0.7149, "step": 5482 }, { "epoch": 0.86, "grad_norm": 15.601994523287907, "learning_rate": 1.6768308659061474e-05, "loss": 0.7284, "step": 5483 }, { "epoch": 0.86, "grad_norm": 23.101024903307696, "learning_rate": 1.6767067139510383e-05, "loss": 0.8209, "step": 5484 }, { "epoch": 0.86, "grad_norm": 14.222146911217841, "learning_rate": 1.6765825427507855e-05, "loss": 0.7532, "step": 5485 }, { "epoch": 0.86, "grad_norm": 23.919897616914454, "learning_rate": 1.6764583523089214e-05, "loss": 0.7917, "step": 5486 }, { "epoch": 0.86, "grad_norm": 24.025062797275268, "learning_rate": 1.6763341426289773e-05, "loss": 0.67, "step": 5487 }, { "epoch": 0.86, "grad_norm": 11.989780899025297, "learning_rate": 1.676209913714486e-05, "loss": 0.7046, "step": 5488 }, { "epoch": 0.86, "grad_norm": 18.214736099628603, "learning_rate": 1.6760856655689804e-05, "loss": 0.6675, "step": 5489 }, { "epoch": 0.86, "grad_norm": 22.572673796557577, "learning_rate": 1.675961398195994e-05, "loss": 0.7588, "step": 5490 }, { "epoch": 0.86, "grad_norm": 18.24116876277585, "learning_rate": 1.6758371115990607e-05, "loss": 0.7645, "step": 5491 }, { "epoch": 0.86, "grad_norm": 13.527954935404674, "learning_rate": 1.6757128057817154e-05, "loss": 0.6699, "step": 5492 }, { "epoch": 0.86, "grad_norm": 18.71155999089906, "learning_rate": 1.675588480747493e-05, "loss": 0.6759, "step": 5493 }, { "epoch": 0.86, "grad_norm": 18.509173275191113, "learning_rate": 1.6754641364999297e-05, "loss": 0.7605, "step": 5494 }, { "epoch": 0.86, "grad_norm": 22.03881733773878, "learning_rate": 1.675339773042561e-05, "loss": 0.755, "step": 5495 }, { "epoch": 0.86, "grad_norm": 20.29853177962315, "learning_rate": 1.6752153903789247e-05, "loss": 0.7541, "step": 5496 }, { "epoch": 0.86, "grad_norm": 18.967974491386407, "learning_rate": 1.6750909885125575e-05, "loss": 0.7504, "step": 5497 }, { "epoch": 0.86, "grad_norm": 25.734535862408897, "learning_rate": 1.6749665674469975e-05, "loss": 0.7387, "step": 5498 }, { "epoch": 0.86, "grad_norm": 12.494666704626571, "learning_rate": 1.6748421271857835e-05, "loss": 0.7133, "step": 5499 }, { "epoch": 0.86, "grad_norm": 13.040793883338267, "learning_rate": 1.6747176677324535e-05, "loss": 0.6771, "step": 5500 }, { "epoch": 0.86, "grad_norm": 18.44159683520143, "learning_rate": 1.674593189090548e-05, "loss": 0.6774, "step": 5501 }, { "epoch": 0.86, "grad_norm": 28.33537642841377, "learning_rate": 1.674468691263607e-05, "loss": 0.737, "step": 5502 }, { "epoch": 0.86, "grad_norm": 22.05528572404925, "learning_rate": 1.674344174255171e-05, "loss": 0.7308, "step": 5503 }, { "epoch": 0.86, "grad_norm": 14.338982081574864, "learning_rate": 1.674219638068781e-05, "loss": 0.6879, "step": 5504 }, { "epoch": 0.86, "grad_norm": 20.994028790795618, "learning_rate": 1.6740950827079786e-05, "loss": 0.7206, "step": 5505 }, { "epoch": 0.86, "grad_norm": 28.307530655447184, "learning_rate": 1.673970508176307e-05, "loss": 0.6481, "step": 5506 }, { "epoch": 0.86, "grad_norm": 22.46851334128871, "learning_rate": 1.673845914477308e-05, "loss": 0.7108, "step": 5507 }, { "epoch": 0.86, "grad_norm": 18.1834380711694, "learning_rate": 1.6737213016145256e-05, "loss": 0.6724, "step": 5508 }, { "epoch": 0.86, "grad_norm": 20.63379417909364, "learning_rate": 1.6735966695915036e-05, "loss": 0.6387, "step": 5509 }, { "epoch": 0.86, "grad_norm": 16.719482008337145, "learning_rate": 1.673472018411786e-05, "loss": 0.7228, "step": 5510 }, { "epoch": 0.86, "grad_norm": 16.721704262125755, "learning_rate": 1.6733473480789183e-05, "loss": 0.7622, "step": 5511 }, { "epoch": 0.86, "grad_norm": 18.955363955890217, "learning_rate": 1.673222658596446e-05, "loss": 0.7596, "step": 5512 }, { "epoch": 0.86, "grad_norm": 33.99476577021787, "learning_rate": 1.673097949967915e-05, "loss": 0.8085, "step": 5513 }, { "epoch": 0.86, "grad_norm": 18.60209281426596, "learning_rate": 1.6729732221968722e-05, "loss": 0.7025, "step": 5514 }, { "epoch": 0.86, "grad_norm": 28.20081843509955, "learning_rate": 1.6728484752868644e-05, "loss": 0.6539, "step": 5515 }, { "epoch": 0.86, "grad_norm": 16.372695521366598, "learning_rate": 1.6727237092414397e-05, "loss": 0.7045, "step": 5516 }, { "epoch": 0.86, "grad_norm": 16.56112279789204, "learning_rate": 1.6725989240641466e-05, "loss": 0.7136, "step": 5517 }, { "epoch": 0.86, "grad_norm": 20.575310276352525, "learning_rate": 1.672474119758533e-05, "loss": 0.7474, "step": 5518 }, { "epoch": 0.86, "grad_norm": 17.956381313236854, "learning_rate": 1.6723492963281492e-05, "loss": 0.6584, "step": 5519 }, { "epoch": 0.86, "grad_norm": 20.492598974159456, "learning_rate": 1.6722244537765444e-05, "loss": 0.7113, "step": 5520 }, { "epoch": 0.86, "grad_norm": 24.99055338242651, "learning_rate": 1.6720995921072698e-05, "loss": 0.6908, "step": 5521 }, { "epoch": 0.86, "grad_norm": 16.617679999352113, "learning_rate": 1.6719747113238754e-05, "loss": 0.6961, "step": 5522 }, { "epoch": 0.86, "grad_norm": 21.938635807575874, "learning_rate": 1.6718498114299138e-05, "loss": 0.7129, "step": 5523 }, { "epoch": 0.86, "grad_norm": 16.346676757710142, "learning_rate": 1.6717248924289363e-05, "loss": 0.7396, "step": 5524 }, { "epoch": 0.86, "grad_norm": 16.889399115692704, "learning_rate": 1.671599954324496e-05, "loss": 0.8214, "step": 5525 }, { "epoch": 0.86, "grad_norm": 15.567382454027097, "learning_rate": 1.6714749971201457e-05, "loss": 0.7353, "step": 5526 }, { "epoch": 0.86, "grad_norm": 32.615304135524894, "learning_rate": 1.6713500208194395e-05, "loss": 0.7089, "step": 5527 }, { "epoch": 0.86, "grad_norm": 14.31624027848153, "learning_rate": 1.6712250254259313e-05, "loss": 0.6713, "step": 5528 }, { "epoch": 0.86, "grad_norm": 22.378537556372276, "learning_rate": 1.6711000109431757e-05, "loss": 0.7627, "step": 5529 }, { "epoch": 0.86, "grad_norm": 13.144232575649506, "learning_rate": 1.6709749773747292e-05, "loss": 0.6812, "step": 5530 }, { "epoch": 0.86, "grad_norm": 38.75782443736315, "learning_rate": 1.670849924724146e-05, "loss": 0.7546, "step": 5531 }, { "epoch": 0.86, "grad_norm": 20.827908124889206, "learning_rate": 1.670724852994984e-05, "loss": 0.7323, "step": 5532 }, { "epoch": 0.86, "grad_norm": 19.193446520468456, "learning_rate": 1.6705997621907993e-05, "loss": 0.7449, "step": 5533 }, { "epoch": 0.86, "grad_norm": 21.531535550896763, "learning_rate": 1.6704746523151498e-05, "loss": 0.7342, "step": 5534 }, { "epoch": 0.86, "grad_norm": 19.871039653514718, "learning_rate": 1.670349523371593e-05, "loss": 0.8125, "step": 5535 }, { "epoch": 0.86, "grad_norm": 18.494552064670486, "learning_rate": 1.670224375363688e-05, "loss": 0.7303, "step": 5536 }, { "epoch": 0.86, "grad_norm": 16.702676063156925, "learning_rate": 1.670099208294994e-05, "loss": 0.7046, "step": 5537 }, { "epoch": 0.87, "grad_norm": 19.632325735255055, "learning_rate": 1.6699740221690706e-05, "loss": 0.8022, "step": 5538 }, { "epoch": 0.87, "grad_norm": 14.587098073427606, "learning_rate": 1.669848816989478e-05, "loss": 0.7318, "step": 5539 }, { "epoch": 0.87, "grad_norm": 14.253821634348277, "learning_rate": 1.6697235927597763e-05, "loss": 0.7121, "step": 5540 }, { "epoch": 0.87, "grad_norm": 23.327291501224117, "learning_rate": 1.6695983494835283e-05, "loss": 0.7967, "step": 5541 }, { "epoch": 0.87, "grad_norm": 21.954009283318214, "learning_rate": 1.669473087164294e-05, "loss": 0.7708, "step": 5542 }, { "epoch": 0.87, "grad_norm": 15.513382041747036, "learning_rate": 1.6693478058056375e-05, "loss": 0.7594, "step": 5543 }, { "epoch": 0.87, "grad_norm": 13.237630234348744, "learning_rate": 1.6692225054111207e-05, "loss": 0.6948, "step": 5544 }, { "epoch": 0.87, "grad_norm": 15.505037142989893, "learning_rate": 1.669097185984307e-05, "loss": 0.696, "step": 5545 }, { "epoch": 0.87, "grad_norm": 19.21335438607501, "learning_rate": 1.668971847528761e-05, "loss": 0.6997, "step": 5546 }, { "epoch": 0.87, "grad_norm": 14.882450275568763, "learning_rate": 1.668846490048047e-05, "loss": 0.677, "step": 5547 }, { "epoch": 0.87, "grad_norm": 20.483848613025124, "learning_rate": 1.6687211135457304e-05, "loss": 0.8614, "step": 5548 }, { "epoch": 0.87, "grad_norm": 18.915116410356028, "learning_rate": 1.668595718025376e-05, "loss": 0.7381, "step": 5549 }, { "epoch": 0.87, "grad_norm": 20.84288955139238, "learning_rate": 1.6684703034905507e-05, "loss": 0.7433, "step": 5550 }, { "epoch": 0.87, "grad_norm": 18.270744241049968, "learning_rate": 1.668344869944821e-05, "loss": 0.6813, "step": 5551 }, { "epoch": 0.87, "grad_norm": 18.405522495247816, "learning_rate": 1.6682194173917543e-05, "loss": 0.7991, "step": 5552 }, { "epoch": 0.87, "grad_norm": 21.065625722874085, "learning_rate": 1.6680939458349184e-05, "loss": 0.6658, "step": 5553 }, { "epoch": 0.87, "grad_norm": 16.149458870886434, "learning_rate": 1.667968455277881e-05, "loss": 0.8056, "step": 5554 }, { "epoch": 0.87, "grad_norm": 20.546887985921607, "learning_rate": 1.667842945724212e-05, "loss": 0.675, "step": 5555 }, { "epoch": 0.87, "grad_norm": 13.620954666543469, "learning_rate": 1.6677174171774798e-05, "loss": 0.6429, "step": 5556 }, { "epoch": 0.87, "grad_norm": 18.3793106010298, "learning_rate": 1.6675918696412552e-05, "loss": 0.6852, "step": 5557 }, { "epoch": 0.87, "grad_norm": 14.533755819574274, "learning_rate": 1.6674663031191084e-05, "loss": 0.7247, "step": 5558 }, { "epoch": 0.87, "grad_norm": 19.292760146952514, "learning_rate": 1.66734071761461e-05, "loss": 0.7152, "step": 5559 }, { "epoch": 0.87, "grad_norm": 23.396429482018345, "learning_rate": 1.6672151131313324e-05, "loss": 0.7848, "step": 5560 }, { "epoch": 0.87, "grad_norm": 19.14182856454198, "learning_rate": 1.6670894896728472e-05, "loss": 0.7138, "step": 5561 }, { "epoch": 0.87, "grad_norm": 22.90061012471951, "learning_rate": 1.666963847242727e-05, "loss": 0.7846, "step": 5562 }, { "epoch": 0.87, "grad_norm": 18.584828852010293, "learning_rate": 1.666838185844545e-05, "loss": 0.7686, "step": 5563 }, { "epoch": 0.87, "grad_norm": 19.569462244899825, "learning_rate": 1.6667125054818756e-05, "loss": 0.6719, "step": 5564 }, { "epoch": 0.87, "grad_norm": 26.307846550925944, "learning_rate": 1.6665868061582922e-05, "loss": 0.7473, "step": 5565 }, { "epoch": 0.87, "grad_norm": 31.89063735243966, "learning_rate": 1.66646108787737e-05, "loss": 0.6967, "step": 5566 }, { "epoch": 0.87, "grad_norm": 19.311779746417574, "learning_rate": 1.6663353506426845e-05, "loss": 0.7495, "step": 5567 }, { "epoch": 0.87, "grad_norm": 23.69390050301414, "learning_rate": 1.6662095944578112e-05, "loss": 0.6886, "step": 5568 }, { "epoch": 0.87, "grad_norm": 19.175396334712502, "learning_rate": 1.6660838193263268e-05, "loss": 0.744, "step": 5569 }, { "epoch": 0.87, "grad_norm": 13.753154560751634, "learning_rate": 1.6659580252518083e-05, "loss": 0.7076, "step": 5570 }, { "epoch": 0.87, "grad_norm": 16.184822257793588, "learning_rate": 1.665832212237833e-05, "loss": 0.7598, "step": 5571 }, { "epoch": 0.87, "grad_norm": 14.138235499359693, "learning_rate": 1.6657063802879795e-05, "loss": 0.6107, "step": 5572 }, { "epoch": 0.87, "grad_norm": 22.665448903465165, "learning_rate": 1.6655805294058256e-05, "loss": 0.8026, "step": 5573 }, { "epoch": 0.87, "grad_norm": 26.5927127239132, "learning_rate": 1.665454659594951e-05, "loss": 0.7968, "step": 5574 }, { "epoch": 0.87, "grad_norm": 19.468444241270436, "learning_rate": 1.665328770858935e-05, "loss": 0.7041, "step": 5575 }, { "epoch": 0.87, "grad_norm": 25.703204099191392, "learning_rate": 1.665202863201358e-05, "loss": 0.7069, "step": 5576 }, { "epoch": 0.87, "grad_norm": 18.347136781098744, "learning_rate": 1.6650769366258007e-05, "loss": 0.6781, "step": 5577 }, { "epoch": 0.87, "grad_norm": 19.703343744901467, "learning_rate": 1.6649509911358444e-05, "loss": 0.6943, "step": 5578 }, { "epoch": 0.87, "grad_norm": 18.322211212314883, "learning_rate": 1.6648250267350708e-05, "loss": 0.8158, "step": 5579 }, { "epoch": 0.87, "grad_norm": 16.781358037455213, "learning_rate": 1.6646990434270625e-05, "loss": 0.686, "step": 5580 }, { "epoch": 0.87, "grad_norm": 25.481525669403204, "learning_rate": 1.664573041215402e-05, "loss": 0.7547, "step": 5581 }, { "epoch": 0.87, "grad_norm": 18.18280275060542, "learning_rate": 1.6644470201036732e-05, "loss": 0.7663, "step": 5582 }, { "epoch": 0.87, "grad_norm": 19.40391433849428, "learning_rate": 1.6643209800954597e-05, "loss": 0.7589, "step": 5583 }, { "epoch": 0.87, "grad_norm": 22.155087411586884, "learning_rate": 1.664194921194346e-05, "loss": 0.6845, "step": 5584 }, { "epoch": 0.87, "grad_norm": 19.26437062954492, "learning_rate": 1.664068843403918e-05, "loss": 0.8006, "step": 5585 }, { "epoch": 0.87, "grad_norm": 16.616893719083563, "learning_rate": 1.6639427467277597e-05, "loss": 0.7166, "step": 5586 }, { "epoch": 0.87, "grad_norm": 28.234724825838313, "learning_rate": 1.663816631169459e-05, "loss": 0.7036, "step": 5587 }, { "epoch": 0.87, "grad_norm": 20.511589595465793, "learning_rate": 1.663690496732601e-05, "loss": 0.7027, "step": 5588 }, { "epoch": 0.87, "grad_norm": 22.159613335204206, "learning_rate": 1.663564343420773e-05, "loss": 0.7976, "step": 5589 }, { "epoch": 0.87, "grad_norm": 16.108872972837347, "learning_rate": 1.6634381712375643e-05, "loss": 0.7092, "step": 5590 }, { "epoch": 0.87, "grad_norm": 20.131143095134703, "learning_rate": 1.6633119801865617e-05, "loss": 0.7527, "step": 5591 }, { "epoch": 0.87, "grad_norm": 20.755097063373377, "learning_rate": 1.6631857702713544e-05, "loss": 0.7212, "step": 5592 }, { "epoch": 0.87, "grad_norm": 21.457837596779466, "learning_rate": 1.6630595414955314e-05, "loss": 0.6246, "step": 5593 }, { "epoch": 0.87, "grad_norm": 13.411398728124018, "learning_rate": 1.6629332938626836e-05, "loss": 0.6982, "step": 5594 }, { "epoch": 0.87, "grad_norm": 15.84759840415231, "learning_rate": 1.6628070273764002e-05, "loss": 0.7234, "step": 5595 }, { "epoch": 0.87, "grad_norm": 25.660024459391256, "learning_rate": 1.6626807420402727e-05, "loss": 0.7209, "step": 5596 }, { "epoch": 0.87, "grad_norm": 18.33829364136699, "learning_rate": 1.662554437857893e-05, "loss": 0.8041, "step": 5597 }, { "epoch": 0.87, "grad_norm": 17.02438790774433, "learning_rate": 1.662428114832852e-05, "loss": 0.7279, "step": 5598 }, { "epoch": 0.87, "grad_norm": 22.76716925463418, "learning_rate": 1.6623017729687432e-05, "loss": 0.7577, "step": 5599 }, { "epoch": 0.87, "grad_norm": 14.27211329754222, "learning_rate": 1.6621754122691596e-05, "loss": 0.6983, "step": 5600 }, { "epoch": 0.87, "grad_norm": 17.032776306599608, "learning_rate": 1.6620490327376947e-05, "loss": 0.721, "step": 5601 }, { "epoch": 0.88, "grad_norm": 21.787681220196095, "learning_rate": 1.6619226343779424e-05, "loss": 0.6814, "step": 5602 }, { "epoch": 0.88, "grad_norm": 23.348860665079567, "learning_rate": 1.6617962171934975e-05, "loss": 0.7318, "step": 5603 }, { "epoch": 0.88, "grad_norm": 21.521396659286662, "learning_rate": 1.6616697811879553e-05, "loss": 0.7363, "step": 5604 }, { "epoch": 0.88, "grad_norm": 20.860291335163293, "learning_rate": 1.661543326364912e-05, "loss": 0.7472, "step": 5605 }, { "epoch": 0.88, "grad_norm": 15.75152451747405, "learning_rate": 1.661416852727963e-05, "loss": 0.7367, "step": 5606 }, { "epoch": 0.88, "grad_norm": 17.039202729775827, "learning_rate": 1.661290360280706e-05, "loss": 0.674, "step": 5607 }, { "epoch": 0.88, "grad_norm": 16.370202808262032, "learning_rate": 1.6611638490267375e-05, "loss": 0.6773, "step": 5608 }, { "epoch": 0.88, "grad_norm": 20.7575965664809, "learning_rate": 1.6610373189696565e-05, "loss": 0.7473, "step": 5609 }, { "epoch": 0.88, "grad_norm": 19.7891086912162, "learning_rate": 1.6609107701130603e-05, "loss": 0.6954, "step": 5610 }, { "epoch": 0.88, "grad_norm": 22.139049392867005, "learning_rate": 1.660784202460549e-05, "loss": 0.7718, "step": 5611 }, { "epoch": 0.88, "grad_norm": 23.288890370907275, "learning_rate": 1.6606576160157212e-05, "loss": 0.7744, "step": 5612 }, { "epoch": 0.88, "grad_norm": 24.609885958256594, "learning_rate": 1.6605310107821774e-05, "loss": 0.7104, "step": 5613 }, { "epoch": 0.88, "grad_norm": 16.668146188800378, "learning_rate": 1.660404386763518e-05, "loss": 0.7293, "step": 5614 }, { "epoch": 0.88, "grad_norm": 17.494055250018512, "learning_rate": 1.660277743963344e-05, "loss": 0.6627, "step": 5615 }, { "epoch": 0.88, "grad_norm": 15.059095645956646, "learning_rate": 1.6601510823852574e-05, "loss": 0.7107, "step": 5616 }, { "epoch": 0.88, "grad_norm": 15.99842022057446, "learning_rate": 1.66002440203286e-05, "loss": 0.7479, "step": 5617 }, { "epoch": 0.88, "grad_norm": 15.643079812755305, "learning_rate": 1.659897702909755e-05, "loss": 0.8548, "step": 5618 }, { "epoch": 0.88, "grad_norm": 25.324072378295437, "learning_rate": 1.659770985019545e-05, "loss": 0.7261, "step": 5619 }, { "epoch": 0.88, "grad_norm": 19.60092128892326, "learning_rate": 1.6596442483658345e-05, "loss": 0.792, "step": 5620 }, { "epoch": 0.88, "grad_norm": 21.09456805373649, "learning_rate": 1.6595174929522273e-05, "loss": 0.7742, "step": 5621 }, { "epoch": 0.88, "grad_norm": 19.991880673151805, "learning_rate": 1.6593907187823284e-05, "loss": 0.7054, "step": 5622 }, { "epoch": 0.88, "grad_norm": 22.437698529741215, "learning_rate": 1.6592639258597432e-05, "loss": 0.7062, "step": 5623 }, { "epoch": 0.88, "grad_norm": 15.745114927499325, "learning_rate": 1.6591371141880776e-05, "loss": 0.7459, "step": 5624 }, { "epoch": 0.88, "grad_norm": 13.622390470927279, "learning_rate": 1.6590102837709382e-05, "loss": 0.7225, "step": 5625 }, { "epoch": 0.88, "grad_norm": 21.870437478685915, "learning_rate": 1.6588834346119315e-05, "loss": 0.7657, "step": 5626 }, { "epoch": 0.88, "grad_norm": 21.268626183483757, "learning_rate": 1.6587565667146657e-05, "loss": 0.887, "step": 5627 }, { "epoch": 0.88, "grad_norm": 19.831310196369493, "learning_rate": 1.6586296800827483e-05, "loss": 0.7503, "step": 5628 }, { "epoch": 0.88, "grad_norm": 13.596742024491475, "learning_rate": 1.6585027747197882e-05, "loss": 0.76, "step": 5629 }, { "epoch": 0.88, "grad_norm": 14.465565701814533, "learning_rate": 1.6583758506293945e-05, "loss": 0.6691, "step": 5630 }, { "epoch": 0.88, "grad_norm": 18.34487366382551, "learning_rate": 1.6582489078151765e-05, "loss": 0.7564, "step": 5631 }, { "epoch": 0.88, "grad_norm": 13.745352710718397, "learning_rate": 1.6581219462807444e-05, "loss": 0.6095, "step": 5632 }, { "epoch": 0.88, "grad_norm": 20.98266594176781, "learning_rate": 1.6579949660297098e-05, "loss": 0.8123, "step": 5633 }, { "epoch": 0.88, "grad_norm": 16.608125391575303, "learning_rate": 1.657867967065683e-05, "loss": 0.6883, "step": 5634 }, { "epoch": 0.88, "grad_norm": 18.35142454758878, "learning_rate": 1.657740949392276e-05, "loss": 0.7156, "step": 5635 }, { "epoch": 0.88, "grad_norm": 16.73731976555211, "learning_rate": 1.657613913013101e-05, "loss": 0.7523, "step": 5636 }, { "epoch": 0.88, "grad_norm": 16.179669520627378, "learning_rate": 1.657486857931771e-05, "loss": 0.72, "step": 5637 }, { "epoch": 0.88, "grad_norm": 18.431933500835687, "learning_rate": 1.6573597841518995e-05, "loss": 0.6616, "step": 5638 }, { "epoch": 0.88, "grad_norm": 16.440175977244216, "learning_rate": 1.6572326916771008e-05, "loss": 0.7945, "step": 5639 }, { "epoch": 0.88, "grad_norm": 11.693828836488256, "learning_rate": 1.657105580510988e-05, "loss": 0.7288, "step": 5640 }, { "epoch": 0.88, "grad_norm": 14.244733091183958, "learning_rate": 1.6569784506571772e-05, "loss": 0.7198, "step": 5641 }, { "epoch": 0.88, "grad_norm": 16.140556573491953, "learning_rate": 1.6568513021192837e-05, "loss": 0.7317, "step": 5642 }, { "epoch": 0.88, "grad_norm": 17.960201168688616, "learning_rate": 1.6567241349009235e-05, "loss": 0.6911, "step": 5643 }, { "epoch": 0.88, "grad_norm": 17.151401328821926, "learning_rate": 1.6565969490057128e-05, "loss": 0.6568, "step": 5644 }, { "epoch": 0.88, "grad_norm": 27.846305203315232, "learning_rate": 1.656469744437269e-05, "loss": 0.7271, "step": 5645 }, { "epoch": 0.88, "grad_norm": 18.376494627860403, "learning_rate": 1.65634252119921e-05, "loss": 0.6899, "step": 5646 }, { "epoch": 0.88, "grad_norm": 13.58101857633776, "learning_rate": 1.6562152792951534e-05, "loss": 0.8136, "step": 5647 }, { "epoch": 0.88, "grad_norm": 11.968154964645576, "learning_rate": 1.656088018728718e-05, "loss": 0.6698, "step": 5648 }, { "epoch": 0.88, "grad_norm": 13.635239206441062, "learning_rate": 1.6559607395035234e-05, "loss": 0.6785, "step": 5649 }, { "epoch": 0.88, "grad_norm": 17.85055862458894, "learning_rate": 1.655833441623189e-05, "loss": 0.7595, "step": 5650 }, { "epoch": 0.88, "grad_norm": 11.452070321654034, "learning_rate": 1.6557061250913352e-05, "loss": 0.7528, "step": 5651 }, { "epoch": 0.88, "grad_norm": 12.448101144352076, "learning_rate": 1.6555787899115827e-05, "loss": 0.7606, "step": 5652 }, { "epoch": 0.88, "grad_norm": 13.298406981692136, "learning_rate": 1.6554514360875528e-05, "loss": 0.7146, "step": 5653 }, { "epoch": 0.88, "grad_norm": 22.357977843134766, "learning_rate": 1.655324063622868e-05, "loss": 0.6706, "step": 5654 }, { "epoch": 0.88, "grad_norm": 20.054341309148466, "learning_rate": 1.6551966725211497e-05, "loss": 0.7434, "step": 5655 }, { "epoch": 0.88, "grad_norm": 21.89844506270747, "learning_rate": 1.6550692627860213e-05, "loss": 0.6801, "step": 5656 }, { "epoch": 0.88, "grad_norm": 12.865825831171021, "learning_rate": 1.6549418344211066e-05, "loss": 0.679, "step": 5657 }, { "epoch": 0.88, "grad_norm": 12.409735980795922, "learning_rate": 1.6548143874300292e-05, "loss": 0.6125, "step": 5658 }, { "epoch": 0.88, "grad_norm": 24.94414499133065, "learning_rate": 1.654686921816413e-05, "loss": 0.6911, "step": 5659 }, { "epoch": 0.88, "grad_norm": 18.445974171891383, "learning_rate": 1.6545594375838846e-05, "loss": 0.6618, "step": 5660 }, { "epoch": 0.88, "grad_norm": 23.57349780480745, "learning_rate": 1.6544319347360685e-05, "loss": 0.8199, "step": 5661 }, { "epoch": 0.88, "grad_norm": 27.230779203626952, "learning_rate": 1.6543044132765907e-05, "loss": 0.6828, "step": 5662 }, { "epoch": 0.88, "grad_norm": 18.855952965347512, "learning_rate": 1.6541768732090784e-05, "loss": 0.7227, "step": 5663 }, { "epoch": 0.88, "grad_norm": 23.783346947768226, "learning_rate": 1.6540493145371582e-05, "loss": 0.6645, "step": 5664 }, { "epoch": 0.88, "grad_norm": 17.76466457619321, "learning_rate": 1.6539217372644585e-05, "loss": 0.6301, "step": 5665 }, { "epoch": 0.89, "grad_norm": 16.164926600786643, "learning_rate": 1.653794141394607e-05, "loss": 0.7843, "step": 5666 }, { "epoch": 0.89, "grad_norm": 17.844293454140125, "learning_rate": 1.6536665269312324e-05, "loss": 0.6663, "step": 5667 }, { "epoch": 0.89, "grad_norm": 19.45526192123226, "learning_rate": 1.6535388938779644e-05, "loss": 0.7432, "step": 5668 }, { "epoch": 0.89, "grad_norm": 18.74445819496282, "learning_rate": 1.6534112422384324e-05, "loss": 0.7317, "step": 5669 }, { "epoch": 0.89, "grad_norm": 14.787744545486513, "learning_rate": 1.653283572016267e-05, "loss": 0.6547, "step": 5670 }, { "epoch": 0.89, "grad_norm": 22.759511184284943, "learning_rate": 1.653155883215099e-05, "loss": 0.7599, "step": 5671 }, { "epoch": 0.89, "grad_norm": 21.623840955587834, "learning_rate": 1.6530281758385597e-05, "loss": 0.7579, "step": 5672 }, { "epoch": 0.89, "grad_norm": 14.249405641458004, "learning_rate": 1.652900449890281e-05, "loss": 0.7172, "step": 5673 }, { "epoch": 0.89, "grad_norm": 18.338067834282093, "learning_rate": 1.6527727053738957e-05, "loss": 0.6848, "step": 5674 }, { "epoch": 0.89, "grad_norm": 16.435249072962097, "learning_rate": 1.6526449422930363e-05, "loss": 0.6931, "step": 5675 }, { "epoch": 0.89, "grad_norm": 21.02453084021368, "learning_rate": 1.6525171606513368e-05, "loss": 0.7353, "step": 5676 }, { "epoch": 0.89, "grad_norm": 21.410508837247384, "learning_rate": 1.6523893604524304e-05, "loss": 0.6824, "step": 5677 }, { "epoch": 0.89, "grad_norm": 16.91434361640482, "learning_rate": 1.652261541699953e-05, "loss": 0.6474, "step": 5678 }, { "epoch": 0.89, "grad_norm": 19.18647938135595, "learning_rate": 1.6521337043975384e-05, "loss": 0.6542, "step": 5679 }, { "epoch": 0.89, "grad_norm": 20.790304567604757, "learning_rate": 1.652005848548823e-05, "loss": 0.7275, "step": 5680 }, { "epoch": 0.89, "grad_norm": 20.48159908862092, "learning_rate": 1.6518779741574422e-05, "loss": 0.7223, "step": 5681 }, { "epoch": 0.89, "grad_norm": 18.67648162039196, "learning_rate": 1.6517500812270335e-05, "loss": 0.7433, "step": 5682 }, { "epoch": 0.89, "grad_norm": 18.224208182662455, "learning_rate": 1.6516221697612338e-05, "loss": 0.6883, "step": 5683 }, { "epoch": 0.89, "grad_norm": 28.35615649915233, "learning_rate": 1.6514942397636807e-05, "loss": 0.8017, "step": 5684 }, { "epoch": 0.89, "grad_norm": 15.18832427651014, "learning_rate": 1.6513662912380123e-05, "loss": 0.793, "step": 5685 }, { "epoch": 0.89, "grad_norm": 15.798340100303275, "learning_rate": 1.651238324187868e-05, "loss": 0.6869, "step": 5686 }, { "epoch": 0.89, "grad_norm": 23.429544259524203, "learning_rate": 1.6511103386168867e-05, "loss": 0.76, "step": 5687 }, { "epoch": 0.89, "grad_norm": 20.18241841323595, "learning_rate": 1.650982334528708e-05, "loss": 0.6465, "step": 5688 }, { "epoch": 0.89, "grad_norm": 19.43709596600368, "learning_rate": 1.6508543119269727e-05, "loss": 0.7059, "step": 5689 }, { "epoch": 0.89, "grad_norm": 20.585352701860142, "learning_rate": 1.6507262708153215e-05, "loss": 0.7288, "step": 5690 }, { "epoch": 0.89, "grad_norm": 17.692208607133892, "learning_rate": 1.6505982111973956e-05, "loss": 0.7269, "step": 5691 }, { "epoch": 0.89, "grad_norm": 13.412397001311842, "learning_rate": 1.6504701330768372e-05, "loss": 0.6766, "step": 5692 }, { "epoch": 0.89, "grad_norm": 15.154341760434438, "learning_rate": 1.650342036457289e-05, "loss": 0.7598, "step": 5693 }, { "epoch": 0.89, "grad_norm": 24.8203046995012, "learning_rate": 1.6502139213423936e-05, "loss": 0.6902, "step": 5694 }, { "epoch": 0.89, "grad_norm": 27.61772130210199, "learning_rate": 1.650085787735795e-05, "loss": 0.7521, "step": 5695 }, { "epoch": 0.89, "grad_norm": 25.034694769455974, "learning_rate": 1.649957635641136e-05, "loss": 0.7226, "step": 5696 }, { "epoch": 0.89, "grad_norm": 16.32790611567556, "learning_rate": 1.6498294650620628e-05, "loss": 0.6905, "step": 5697 }, { "epoch": 0.89, "grad_norm": 20.756171411085806, "learning_rate": 1.6497012760022197e-05, "loss": 0.6105, "step": 5698 }, { "epoch": 0.89, "grad_norm": 22.9530410349578, "learning_rate": 1.649573068465252e-05, "loss": 0.8234, "step": 5699 }, { "epoch": 0.89, "grad_norm": 21.5295224375633, "learning_rate": 1.6494448424548065e-05, "loss": 0.7711, "step": 5700 }, { "epoch": 0.89, "grad_norm": 20.882921491167963, "learning_rate": 1.6493165979745294e-05, "loss": 0.7079, "step": 5701 }, { "epoch": 0.89, "grad_norm": 14.860371266449944, "learning_rate": 1.649188335028068e-05, "loss": 0.7249, "step": 5702 }, { "epoch": 0.89, "grad_norm": 26.544225030604608, "learning_rate": 1.6490600536190705e-05, "loss": 0.8041, "step": 5703 }, { "epoch": 0.89, "grad_norm": 18.9381631775993, "learning_rate": 1.6489317537511846e-05, "loss": 0.6826, "step": 5704 }, { "epoch": 0.89, "grad_norm": 17.407314707999674, "learning_rate": 1.6488034354280592e-05, "loss": 0.7103, "step": 5705 }, { "epoch": 0.89, "grad_norm": 27.182579225513994, "learning_rate": 1.648675098653344e-05, "loss": 0.762, "step": 5706 }, { "epoch": 0.89, "grad_norm": 14.006120731722394, "learning_rate": 1.648546743430688e-05, "loss": 0.7462, "step": 5707 }, { "epoch": 0.89, "grad_norm": 20.798014200121603, "learning_rate": 1.648418369763742e-05, "loss": 0.7125, "step": 5708 }, { "epoch": 0.89, "grad_norm": 14.193339055305193, "learning_rate": 1.648289977656157e-05, "loss": 0.6597, "step": 5709 }, { "epoch": 0.89, "grad_norm": 15.733403721864674, "learning_rate": 1.6481615671115845e-05, "loss": 0.7473, "step": 5710 }, { "epoch": 0.89, "grad_norm": 11.467838994361564, "learning_rate": 1.6480331381336757e-05, "loss": 0.6704, "step": 5711 }, { "epoch": 0.89, "grad_norm": 21.564381874933485, "learning_rate": 1.647904690726084e-05, "loss": 0.661, "step": 5712 }, { "epoch": 0.89, "grad_norm": 26.289721502515405, "learning_rate": 1.6477762248924616e-05, "loss": 0.6799, "step": 5713 }, { "epoch": 0.89, "grad_norm": 18.554585598715992, "learning_rate": 1.6476477406364623e-05, "loss": 0.7488, "step": 5714 }, { "epoch": 0.89, "grad_norm": 17.731416276488957, "learning_rate": 1.6475192379617405e-05, "loss": 0.7656, "step": 5715 }, { "epoch": 0.89, "grad_norm": 22.411782964186976, "learning_rate": 1.64739071687195e-05, "loss": 0.7907, "step": 5716 }, { "epoch": 0.89, "grad_norm": 14.706146514036433, "learning_rate": 1.6472621773707463e-05, "loss": 0.7054, "step": 5717 }, { "epoch": 0.89, "grad_norm": 17.554123431838466, "learning_rate": 1.647133619461785e-05, "loss": 0.717, "step": 5718 }, { "epoch": 0.89, "grad_norm": 26.95087189918306, "learning_rate": 1.647005043148722e-05, "loss": 0.762, "step": 5719 }, { "epoch": 0.89, "grad_norm": 27.028863548489635, "learning_rate": 1.646876448435214e-05, "loss": 0.7717, "step": 5720 }, { "epoch": 0.89, "grad_norm": 18.460620633641483, "learning_rate": 1.6467478353249186e-05, "loss": 0.7567, "step": 5721 }, { "epoch": 0.89, "grad_norm": 13.78109727286138, "learning_rate": 1.6466192038214928e-05, "loss": 0.684, "step": 5722 }, { "epoch": 0.89, "grad_norm": 19.78884535322842, "learning_rate": 1.646490553928595e-05, "loss": 0.6817, "step": 5723 }, { "epoch": 0.89, "grad_norm": 16.732138404346333, "learning_rate": 1.6463618856498844e-05, "loss": 0.7988, "step": 5724 }, { "epoch": 0.89, "grad_norm": 24.295008041801893, "learning_rate": 1.64623319898902e-05, "loss": 0.7215, "step": 5725 }, { "epoch": 0.89, "grad_norm": 16.610719198450784, "learning_rate": 1.6461044939496606e-05, "loss": 0.6587, "step": 5726 }, { "epoch": 0.89, "grad_norm": 17.817251600648486, "learning_rate": 1.645975770535468e-05, "loss": 0.6557, "step": 5727 }, { "epoch": 0.89, "grad_norm": 17.32195410909108, "learning_rate": 1.6458470287501024e-05, "loss": 0.7095, "step": 5728 }, { "epoch": 0.89, "grad_norm": 16.27110069371097, "learning_rate": 1.6457182685972248e-05, "loss": 0.6062, "step": 5729 }, { "epoch": 0.9, "grad_norm": 25.095901566257552, "learning_rate": 1.6455894900804974e-05, "loss": 0.6872, "step": 5730 }, { "epoch": 0.9, "grad_norm": 24.916391614298977, "learning_rate": 1.6454606932035827e-05, "loss": 0.7936, "step": 5731 }, { "epoch": 0.9, "grad_norm": 15.674545463712919, "learning_rate": 1.6453318779701434e-05, "loss": 0.6512, "step": 5732 }, { "epoch": 0.9, "grad_norm": 22.96112751890239, "learning_rate": 1.645203044383843e-05, "loss": 0.7633, "step": 5733 }, { "epoch": 0.9, "grad_norm": 35.18483719736039, "learning_rate": 1.645074192448346e-05, "loss": 0.7853, "step": 5734 }, { "epoch": 0.9, "grad_norm": 17.53176577382229, "learning_rate": 1.6449453221673154e-05, "loss": 0.6666, "step": 5735 }, { "epoch": 0.9, "grad_norm": 14.473599699417615, "learning_rate": 1.6448164335444178e-05, "loss": 0.6325, "step": 5736 }, { "epoch": 0.9, "grad_norm": 19.893537718634708, "learning_rate": 1.6446875265833178e-05, "loss": 0.6931, "step": 5737 }, { "epoch": 0.9, "grad_norm": 22.35430756545539, "learning_rate": 1.6445586012876814e-05, "loss": 0.7357, "step": 5738 }, { "epoch": 0.9, "grad_norm": 18.71951126857608, "learning_rate": 1.644429657661176e-05, "loss": 0.7224, "step": 5739 }, { "epoch": 0.9, "grad_norm": 19.413789780042524, "learning_rate": 1.6443006957074675e-05, "loss": 0.8399, "step": 5740 }, { "epoch": 0.9, "grad_norm": 22.443157903238202, "learning_rate": 1.6441717154302246e-05, "loss": 0.7918, "step": 5741 }, { "epoch": 0.9, "grad_norm": 15.37276849941754, "learning_rate": 1.644042716833115e-05, "loss": 0.6584, "step": 5742 }, { "epoch": 0.9, "grad_norm": 23.25333712134639, "learning_rate": 1.6439136999198068e-05, "loss": 0.7796, "step": 5743 }, { "epoch": 0.9, "grad_norm": 23.405334323599018, "learning_rate": 1.64378466469397e-05, "loss": 0.8338, "step": 5744 }, { "epoch": 0.9, "grad_norm": 16.330854093823742, "learning_rate": 1.643655611159274e-05, "loss": 0.7011, "step": 5745 }, { "epoch": 0.9, "grad_norm": 18.47278854667994, "learning_rate": 1.6435265393193886e-05, "loss": 0.6948, "step": 5746 }, { "epoch": 0.9, "grad_norm": 15.217504019519719, "learning_rate": 1.6433974491779853e-05, "loss": 0.7489, "step": 5747 }, { "epoch": 0.9, "grad_norm": 21.288441049447194, "learning_rate": 1.6432683407387346e-05, "loss": 0.7664, "step": 5748 }, { "epoch": 0.9, "grad_norm": 14.15909214309748, "learning_rate": 1.6431392140053088e-05, "loss": 0.7481, "step": 5749 }, { "epoch": 0.9, "grad_norm": 28.147548078110866, "learning_rate": 1.6430100689813797e-05, "loss": 0.8429, "step": 5750 }, { "epoch": 0.9, "grad_norm": 15.291706603134074, "learning_rate": 1.642880905670621e-05, "loss": 0.7116, "step": 5751 }, { "epoch": 0.9, "grad_norm": 15.039495611518436, "learning_rate": 1.6427517240767046e-05, "loss": 0.6976, "step": 5752 }, { "epoch": 0.9, "grad_norm": 16.935437611792857, "learning_rate": 1.6426225242033055e-05, "loss": 0.7023, "step": 5753 }, { "epoch": 0.9, "grad_norm": 13.821594706584087, "learning_rate": 1.6424933060540978e-05, "loss": 0.7242, "step": 5754 }, { "epoch": 0.9, "grad_norm": 20.602660503572892, "learning_rate": 1.6423640696327564e-05, "loss": 0.7304, "step": 5755 }, { "epoch": 0.9, "grad_norm": 4.888132968633195, "learning_rate": 1.6422348149429566e-05, "loss": 0.694, "step": 5756 }, { "epoch": 0.9, "grad_norm": 15.49301044136683, "learning_rate": 1.6421055419883744e-05, "loss": 0.6861, "step": 5757 }, { "epoch": 0.9, "grad_norm": 13.656531414497566, "learning_rate": 1.641976250772686e-05, "loss": 0.7489, "step": 5758 }, { "epoch": 0.9, "grad_norm": 26.013145738750463, "learning_rate": 1.6418469412995685e-05, "loss": 0.7409, "step": 5759 }, { "epoch": 0.9, "grad_norm": 13.819613326906847, "learning_rate": 1.6417176135726998e-05, "loss": 0.7258, "step": 5760 }, { "epoch": 0.9, "grad_norm": 24.179174931135993, "learning_rate": 1.641588267595757e-05, "loss": 0.6894, "step": 5761 }, { "epoch": 0.9, "grad_norm": 24.704633209495057, "learning_rate": 1.6414589033724197e-05, "loss": 0.673, "step": 5762 }, { "epoch": 0.9, "grad_norm": 22.870078511105422, "learning_rate": 1.6413295209063663e-05, "loss": 0.7022, "step": 5763 }, { "epoch": 0.9, "grad_norm": 22.8587057441996, "learning_rate": 1.6412001202012768e-05, "loss": 0.744, "step": 5764 }, { "epoch": 0.9, "grad_norm": 11.374960559996676, "learning_rate": 1.6410707012608303e-05, "loss": 0.6695, "step": 5765 }, { "epoch": 0.9, "grad_norm": 19.30811657895733, "learning_rate": 1.640941264088709e-05, "loss": 0.7162, "step": 5766 }, { "epoch": 0.9, "grad_norm": 14.24656049153652, "learning_rate": 1.6408118086885925e-05, "loss": 0.7177, "step": 5767 }, { "epoch": 0.9, "grad_norm": 20.595495521942496, "learning_rate": 1.6406823350641628e-05, "loss": 0.7463, "step": 5768 }, { "epoch": 0.9, "grad_norm": 16.459095175127864, "learning_rate": 1.6405528432191027e-05, "loss": 0.7578, "step": 5769 }, { "epoch": 0.9, "grad_norm": 17.044689857474776, "learning_rate": 1.6404233331570944e-05, "loss": 0.6874, "step": 5770 }, { "epoch": 0.9, "grad_norm": 4.205197236482149, "learning_rate": 1.6402938048818208e-05, "loss": 0.6507, "step": 5771 }, { "epoch": 0.9, "grad_norm": 20.028392994915404, "learning_rate": 1.640164258396966e-05, "loss": 0.7461, "step": 5772 }, { "epoch": 0.9, "grad_norm": 27.12965333100808, "learning_rate": 1.6400346937062147e-05, "loss": 0.7316, "step": 5773 }, { "epoch": 0.9, "grad_norm": 18.985086341372234, "learning_rate": 1.6399051108132507e-05, "loss": 0.6895, "step": 5774 }, { "epoch": 0.9, "grad_norm": 22.008121465288855, "learning_rate": 1.6397755097217598e-05, "loss": 0.7771, "step": 5775 }, { "epoch": 0.9, "grad_norm": 21.732444575985188, "learning_rate": 1.6396458904354275e-05, "loss": 0.7354, "step": 5776 }, { "epoch": 0.9, "grad_norm": 25.80816470182595, "learning_rate": 1.6395162529579406e-05, "loss": 0.7311, "step": 5777 }, { "epoch": 0.9, "grad_norm": 13.955277632507821, "learning_rate": 1.6393865972929853e-05, "loss": 0.7192, "step": 5778 }, { "epoch": 0.9, "grad_norm": 31.729889325866502, "learning_rate": 1.6392569234442492e-05, "loss": 0.7409, "step": 5779 }, { "epoch": 0.9, "grad_norm": 10.554252143990125, "learning_rate": 1.63912723141542e-05, "loss": 0.6934, "step": 5780 }, { "epoch": 0.9, "grad_norm": 18.4984187532941, "learning_rate": 1.6389975212101867e-05, "loss": 0.7349, "step": 5781 }, { "epoch": 0.9, "grad_norm": 24.212569607342665, "learning_rate": 1.6388677928322376e-05, "loss": 0.6987, "step": 5782 }, { "epoch": 0.9, "grad_norm": 23.265658918500762, "learning_rate": 1.638738046285262e-05, "loss": 0.6853, "step": 5783 }, { "epoch": 0.9, "grad_norm": 20.651439859095806, "learning_rate": 1.63860828157295e-05, "loss": 0.7496, "step": 5784 }, { "epoch": 0.9, "grad_norm": 14.696158599730714, "learning_rate": 1.6384784986989918e-05, "loss": 0.7441, "step": 5785 }, { "epoch": 0.9, "grad_norm": 22.1359620780321, "learning_rate": 1.638348697667079e-05, "loss": 0.7198, "step": 5786 }, { "epoch": 0.9, "grad_norm": 18.99321591202826, "learning_rate": 1.6382188784809027e-05, "loss": 0.7325, "step": 5787 }, { "epoch": 0.9, "grad_norm": 16.02321955580521, "learning_rate": 1.638089041144155e-05, "loss": 0.7557, "step": 5788 }, { "epoch": 0.9, "grad_norm": 18.966457452291284, "learning_rate": 1.6379591856605274e-05, "loss": 0.7063, "step": 5789 }, { "epoch": 0.9, "grad_norm": 17.163418201252806, "learning_rate": 1.6378293120337144e-05, "loss": 0.7001, "step": 5790 }, { "epoch": 0.9, "grad_norm": 20.953390495273855, "learning_rate": 1.637699420267409e-05, "loss": 0.7787, "step": 5791 }, { "epoch": 0.9, "grad_norm": 19.369755507357898, "learning_rate": 1.6375695103653047e-05, "loss": 0.7026, "step": 5792 }, { "epoch": 0.9, "grad_norm": 9.90137034189558, "learning_rate": 1.637439582331097e-05, "loss": 0.5927, "step": 5793 }, { "epoch": 0.91, "grad_norm": 21.616159394648477, "learning_rate": 1.63730963616848e-05, "loss": 0.6738, "step": 5794 }, { "epoch": 0.91, "grad_norm": 12.90839275544625, "learning_rate": 1.6371796718811496e-05, "loss": 0.6792, "step": 5795 }, { "epoch": 0.91, "grad_norm": 25.755384194776852, "learning_rate": 1.637049689472803e-05, "loss": 0.8311, "step": 5796 }, { "epoch": 0.91, "grad_norm": 26.98814318289358, "learning_rate": 1.636919688947135e-05, "loss": 0.8479, "step": 5797 }, { "epoch": 0.91, "grad_norm": 28.09579774842622, "learning_rate": 1.636789670307844e-05, "loss": 0.7434, "step": 5798 }, { "epoch": 0.91, "grad_norm": 23.489971238392123, "learning_rate": 1.636659633558627e-05, "loss": 0.7615, "step": 5799 }, { "epoch": 0.91, "grad_norm": 13.174140411815452, "learning_rate": 1.6365295787031827e-05, "loss": 0.6884, "step": 5800 }, { "epoch": 0.91, "grad_norm": 32.195328238283686, "learning_rate": 1.6363995057452098e-05, "loss": 0.7759, "step": 5801 }, { "epoch": 0.91, "grad_norm": 22.505019659154755, "learning_rate": 1.6362694146884067e-05, "loss": 0.7392, "step": 5802 }, { "epoch": 0.91, "grad_norm": 23.42531078884606, "learning_rate": 1.636139305536474e-05, "loss": 0.6837, "step": 5803 }, { "epoch": 0.91, "grad_norm": 15.879879409041795, "learning_rate": 1.6360091782931115e-05, "loss": 0.7237, "step": 5804 }, { "epoch": 0.91, "grad_norm": 13.782277755517892, "learning_rate": 1.63587903296202e-05, "loss": 0.6456, "step": 5805 }, { "epoch": 0.91, "grad_norm": 12.95610790913143, "learning_rate": 1.635748869546901e-05, "loss": 0.6609, "step": 5806 }, { "epoch": 0.91, "grad_norm": 21.08536098097721, "learning_rate": 1.6356186880514556e-05, "loss": 0.6788, "step": 5807 }, { "epoch": 0.91, "grad_norm": 16.278555583952873, "learning_rate": 1.635488488479387e-05, "loss": 0.7367, "step": 5808 }, { "epoch": 0.91, "grad_norm": 28.759383400674686, "learning_rate": 1.635358270834397e-05, "loss": 0.6626, "step": 5809 }, { "epoch": 0.91, "grad_norm": 20.078319271053218, "learning_rate": 1.6352280351201898e-05, "loss": 0.7372, "step": 5810 }, { "epoch": 0.91, "grad_norm": 20.144959247533215, "learning_rate": 1.635097781340469e-05, "loss": 0.6672, "step": 5811 }, { "epoch": 0.91, "grad_norm": 16.14920839463731, "learning_rate": 1.6349675094989387e-05, "loss": 0.7133, "step": 5812 }, { "epoch": 0.91, "grad_norm": 12.656466561101208, "learning_rate": 1.6348372195993037e-05, "loss": 0.7127, "step": 5813 }, { "epoch": 0.91, "grad_norm": 18.666831581568328, "learning_rate": 1.63470691164527e-05, "loss": 0.712, "step": 5814 }, { "epoch": 0.91, "grad_norm": 11.696959494882545, "learning_rate": 1.6345765856405424e-05, "loss": 0.6774, "step": 5815 }, { "epoch": 0.91, "grad_norm": 18.490884041950114, "learning_rate": 1.6344462415888283e-05, "loss": 0.7901, "step": 5816 }, { "epoch": 0.91, "grad_norm": 19.732721353459777, "learning_rate": 1.6343158794938342e-05, "loss": 0.6554, "step": 5817 }, { "epoch": 0.91, "grad_norm": 25.560658768655045, "learning_rate": 1.6341854993592674e-05, "loss": 0.6604, "step": 5818 }, { "epoch": 0.91, "grad_norm": 24.328422611942173, "learning_rate": 1.634055101188836e-05, "loss": 0.7234, "step": 5819 }, { "epoch": 0.91, "grad_norm": 19.392988013327987, "learning_rate": 1.6339246849862488e-05, "loss": 0.8028, "step": 5820 }, { "epoch": 0.91, "grad_norm": 20.42406229215249, "learning_rate": 1.633794250755214e-05, "loss": 0.7765, "step": 5821 }, { "epoch": 0.91, "grad_norm": 18.514115378247727, "learning_rate": 1.6336637984994418e-05, "loss": 0.7536, "step": 5822 }, { "epoch": 0.91, "grad_norm": 14.812951527727318, "learning_rate": 1.6335333282226414e-05, "loss": 0.6994, "step": 5823 }, { "epoch": 0.91, "grad_norm": 13.910757446580629, "learning_rate": 1.6334028399285243e-05, "loss": 0.6646, "step": 5824 }, { "epoch": 0.91, "grad_norm": 23.882888151803336, "learning_rate": 1.6332723336208007e-05, "loss": 0.7778, "step": 5825 }, { "epoch": 0.91, "grad_norm": 14.67503314209466, "learning_rate": 1.633141809303182e-05, "loss": 0.7605, "step": 5826 }, { "epoch": 0.91, "grad_norm": 16.940135701907863, "learning_rate": 1.633011266979381e-05, "loss": 0.6965, "step": 5827 }, { "epoch": 0.91, "grad_norm": 14.114629175374471, "learning_rate": 1.6328807066531102e-05, "loss": 0.7662, "step": 5828 }, { "epoch": 0.91, "grad_norm": 20.789468579230917, "learning_rate": 1.6327501283280816e-05, "loss": 0.7287, "step": 5829 }, { "epoch": 0.91, "grad_norm": 24.233231541498746, "learning_rate": 1.6326195320080103e-05, "loss": 0.6999, "step": 5830 }, { "epoch": 0.91, "grad_norm": 29.19205879100767, "learning_rate": 1.632488917696609e-05, "loss": 0.6944, "step": 5831 }, { "epoch": 0.91, "grad_norm": 17.753578174610723, "learning_rate": 1.632358285397593e-05, "loss": 0.7462, "step": 5832 }, { "epoch": 0.91, "grad_norm": 20.644718473915397, "learning_rate": 1.6322276351146774e-05, "loss": 0.7603, "step": 5833 }, { "epoch": 0.91, "grad_norm": 21.903770592158505, "learning_rate": 1.6320969668515773e-05, "loss": 0.8326, "step": 5834 }, { "epoch": 0.91, "grad_norm": 22.879420381005456, "learning_rate": 1.63196628061201e-05, "loss": 0.7076, "step": 5835 }, { "epoch": 0.91, "grad_norm": 19.43826235528824, "learning_rate": 1.6318355763996908e-05, "loss": 0.7249, "step": 5836 }, { "epoch": 0.91, "grad_norm": 16.35383209283138, "learning_rate": 1.6317048542183375e-05, "loss": 0.7659, "step": 5837 }, { "epoch": 0.91, "grad_norm": 17.356027427162466, "learning_rate": 1.6315741140716675e-05, "loss": 0.7482, "step": 5838 }, { "epoch": 0.91, "grad_norm": 16.405609994848234, "learning_rate": 1.6314433559633992e-05, "loss": 0.7291, "step": 5839 }, { "epoch": 0.91, "grad_norm": 21.420958412422095, "learning_rate": 1.6313125798972514e-05, "loss": 0.7589, "step": 5840 }, { "epoch": 0.91, "grad_norm": 15.491793530351103, "learning_rate": 1.6311817858769434e-05, "loss": 0.6358, "step": 5841 }, { "epoch": 0.91, "grad_norm": 23.172076856540357, "learning_rate": 1.6310509739061946e-05, "loss": 0.7548, "step": 5842 }, { "epoch": 0.91, "grad_norm": 17.173773223292894, "learning_rate": 1.6309201439887248e-05, "loss": 0.7527, "step": 5843 }, { "epoch": 0.91, "grad_norm": 16.80129068219147, "learning_rate": 1.6307892961282555e-05, "loss": 0.6585, "step": 5844 }, { "epoch": 0.91, "grad_norm": 23.375898180116764, "learning_rate": 1.6306584303285077e-05, "loss": 0.6452, "step": 5845 }, { "epoch": 0.91, "grad_norm": 19.877060956532212, "learning_rate": 1.6305275465932027e-05, "loss": 0.833, "step": 5846 }, { "epoch": 0.91, "grad_norm": 12.77071229112161, "learning_rate": 1.6303966449260636e-05, "loss": 0.6422, "step": 5847 }, { "epoch": 0.91, "grad_norm": 25.207241713057428, "learning_rate": 1.630265725330812e-05, "loss": 0.6187, "step": 5848 }, { "epoch": 0.91, "grad_norm": 21.220453171151078, "learning_rate": 1.6301347878111726e-05, "loss": 0.8335, "step": 5849 }, { "epoch": 0.91, "grad_norm": 15.177799544773135, "learning_rate": 1.630003832370868e-05, "loss": 0.6804, "step": 5850 }, { "epoch": 0.91, "grad_norm": 16.285428087688317, "learning_rate": 1.629872859013623e-05, "loss": 0.7265, "step": 5851 }, { "epoch": 0.91, "grad_norm": 19.01305889244609, "learning_rate": 1.6297418677431625e-05, "loss": 0.6638, "step": 5852 }, { "epoch": 0.91, "grad_norm": 34.01237027495073, "learning_rate": 1.629610858563212e-05, "loss": 0.6628, "step": 5853 }, { "epoch": 0.91, "grad_norm": 18.18372891680082, "learning_rate": 1.6294798314774963e-05, "loss": 0.7069, "step": 5854 }, { "epoch": 0.91, "grad_norm": 20.38925319017957, "learning_rate": 1.6293487864897425e-05, "loss": 0.748, "step": 5855 }, { "epoch": 0.91, "grad_norm": 20.62885086421975, "learning_rate": 1.6292177236036776e-05, "loss": 0.6876, "step": 5856 }, { "epoch": 0.91, "grad_norm": 19.8710086442742, "learning_rate": 1.629086642823029e-05, "loss": 0.6719, "step": 5857 }, { "epoch": 0.92, "grad_norm": 31.722702970495018, "learning_rate": 1.628955544151524e-05, "loss": 0.7324, "step": 5858 }, { "epoch": 0.92, "grad_norm": 29.46370384451972, "learning_rate": 1.6288244275928912e-05, "loss": 0.7323, "step": 5859 }, { "epoch": 0.92, "grad_norm": 19.859204978125025, "learning_rate": 1.6286932931508596e-05, "loss": 0.7308, "step": 5860 }, { "epoch": 0.92, "grad_norm": 17.1169048200672, "learning_rate": 1.6285621408291584e-05, "loss": 0.7692, "step": 5861 }, { "epoch": 0.92, "grad_norm": 17.490830373486926, "learning_rate": 1.6284309706315178e-05, "loss": 0.7694, "step": 5862 }, { "epoch": 0.92, "grad_norm": 19.378687726053442, "learning_rate": 1.6282997825616676e-05, "loss": 0.7079, "step": 5863 }, { "epoch": 0.92, "grad_norm": 15.007806749159085, "learning_rate": 1.62816857662334e-05, "loss": 0.7497, "step": 5864 }, { "epoch": 0.92, "grad_norm": 14.422473914666208, "learning_rate": 1.6280373528202648e-05, "loss": 0.7115, "step": 5865 }, { "epoch": 0.92, "grad_norm": 18.863384409275017, "learning_rate": 1.627906111156175e-05, "loss": 0.7473, "step": 5866 }, { "epoch": 0.92, "grad_norm": 14.712622773388253, "learning_rate": 1.6277748516348025e-05, "loss": 0.638, "step": 5867 }, { "epoch": 0.92, "grad_norm": 13.5264682178003, "learning_rate": 1.6276435742598807e-05, "loss": 0.6744, "step": 5868 }, { "epoch": 0.92, "grad_norm": 16.79368025227889, "learning_rate": 1.6275122790351426e-05, "loss": 0.6629, "step": 5869 }, { "epoch": 0.92, "grad_norm": 16.75003657096813, "learning_rate": 1.6273809659643226e-05, "loss": 0.6624, "step": 5870 }, { "epoch": 0.92, "grad_norm": 17.537043094562865, "learning_rate": 1.6272496350511547e-05, "loss": 0.7322, "step": 5871 }, { "epoch": 0.92, "grad_norm": 19.29564125696257, "learning_rate": 1.6271182862993743e-05, "loss": 0.6634, "step": 5872 }, { "epoch": 0.92, "grad_norm": 21.113132191854387, "learning_rate": 1.6269869197127162e-05, "loss": 0.8624, "step": 5873 }, { "epoch": 0.92, "grad_norm": 18.490609787042935, "learning_rate": 1.6268555352949175e-05, "loss": 0.6903, "step": 5874 }, { "epoch": 0.92, "grad_norm": 15.2939526951468, "learning_rate": 1.626724133049714e-05, "loss": 0.6174, "step": 5875 }, { "epoch": 0.92, "grad_norm": 11.409671632112559, "learning_rate": 1.6265927129808426e-05, "loss": 0.627, "step": 5876 }, { "epoch": 0.92, "grad_norm": 27.136673234973063, "learning_rate": 1.6264612750920406e-05, "loss": 0.7712, "step": 5877 }, { "epoch": 0.92, "grad_norm": 14.666245259950959, "learning_rate": 1.6263298193870468e-05, "loss": 0.6477, "step": 5878 }, { "epoch": 0.92, "grad_norm": 17.55641363525317, "learning_rate": 1.6261983458695996e-05, "loss": 0.7585, "step": 5879 }, { "epoch": 0.92, "grad_norm": 25.48309068437332, "learning_rate": 1.6260668545434375e-05, "loss": 0.7301, "step": 5880 }, { "epoch": 0.92, "grad_norm": 27.784376618553992, "learning_rate": 1.6259353454123003e-05, "loss": 0.7419, "step": 5881 }, { "epoch": 0.92, "grad_norm": 16.801995266137457, "learning_rate": 1.625803818479928e-05, "loss": 0.6598, "step": 5882 }, { "epoch": 0.92, "grad_norm": 15.573069531722815, "learning_rate": 1.6256722737500613e-05, "loss": 0.7259, "step": 5883 }, { "epoch": 0.92, "grad_norm": 16.47241078321202, "learning_rate": 1.625540711226441e-05, "loss": 0.7121, "step": 5884 }, { "epoch": 0.92, "grad_norm": 13.727059569719156, "learning_rate": 1.625409130912809e-05, "loss": 0.7054, "step": 5885 }, { "epoch": 0.92, "grad_norm": 22.850694802267544, "learning_rate": 1.6252775328129067e-05, "loss": 0.82, "step": 5886 }, { "epoch": 0.92, "grad_norm": 16.648720053714523, "learning_rate": 1.6251459169304776e-05, "loss": 0.6951, "step": 5887 }, { "epoch": 0.92, "grad_norm": 32.58796360555267, "learning_rate": 1.6250142832692643e-05, "loss": 0.7245, "step": 5888 }, { "epoch": 0.92, "grad_norm": 28.586115250369748, "learning_rate": 1.6248826318330103e-05, "loss": 0.8198, "step": 5889 }, { "epoch": 0.92, "grad_norm": 19.75473040052826, "learning_rate": 1.6247509626254597e-05, "loss": 0.767, "step": 5890 }, { "epoch": 0.92, "grad_norm": 18.847430644488675, "learning_rate": 1.6246192756503572e-05, "loss": 0.6209, "step": 5891 }, { "epoch": 0.92, "grad_norm": 24.909309793357508, "learning_rate": 1.624487570911448e-05, "loss": 0.7406, "step": 5892 }, { "epoch": 0.92, "grad_norm": 14.864794776139332, "learning_rate": 1.6243558484124778e-05, "loss": 0.7169, "step": 5893 }, { "epoch": 0.92, "grad_norm": 13.195730899481068, "learning_rate": 1.6242241081571923e-05, "loss": 0.6453, "step": 5894 }, { "epoch": 0.92, "grad_norm": 19.801300941874892, "learning_rate": 1.6240923501493383e-05, "loss": 0.7624, "step": 5895 }, { "epoch": 0.92, "grad_norm": 17.3320670730396, "learning_rate": 1.6239605743926632e-05, "loss": 0.8266, "step": 5896 }, { "epoch": 0.92, "grad_norm": 16.097127002272018, "learning_rate": 1.6238287808909137e-05, "loss": 0.7413, "step": 5897 }, { "epoch": 0.92, "grad_norm": 19.67196590260174, "learning_rate": 1.6236969696478393e-05, "loss": 0.6988, "step": 5898 }, { "epoch": 0.92, "grad_norm": 21.623937590361077, "learning_rate": 1.6235651406671875e-05, "loss": 0.6864, "step": 5899 }, { "epoch": 0.92, "grad_norm": 17.586423718896402, "learning_rate": 1.623433293952708e-05, "loss": 0.6383, "step": 5900 }, { "epoch": 0.92, "grad_norm": 27.757080110335508, "learning_rate": 1.6233014295081505e-05, "loss": 0.6652, "step": 5901 }, { "epoch": 0.92, "grad_norm": 19.259816567253296, "learning_rate": 1.623169547337265e-05, "loss": 0.6892, "step": 5902 }, { "epoch": 0.92, "grad_norm": 25.2420301400005, "learning_rate": 1.6230376474438018e-05, "loss": 0.6876, "step": 5903 }, { "epoch": 0.92, "grad_norm": 13.592569898485612, "learning_rate": 1.6229057298315123e-05, "loss": 0.7102, "step": 5904 }, { "epoch": 0.92, "grad_norm": 12.735847455884493, "learning_rate": 1.6227737945041485e-05, "loss": 0.5997, "step": 5905 }, { "epoch": 0.92, "grad_norm": 20.085673896160745, "learning_rate": 1.6226418414654625e-05, "loss": 0.7951, "step": 5906 }, { "epoch": 0.92, "grad_norm": 14.890752029777746, "learning_rate": 1.6225098707192063e-05, "loss": 0.7078, "step": 5907 }, { "epoch": 0.92, "grad_norm": 27.410334507826146, "learning_rate": 1.622377882269134e-05, "loss": 0.7008, "step": 5908 }, { "epoch": 0.92, "grad_norm": 22.162051671703466, "learning_rate": 1.6222458761189984e-05, "loss": 0.7213, "step": 5909 }, { "epoch": 0.92, "grad_norm": 20.494290652244455, "learning_rate": 1.622113852272554e-05, "loss": 0.6921, "step": 5910 }, { "epoch": 0.92, "grad_norm": 24.428519991445437, "learning_rate": 1.621981810733556e-05, "loss": 0.7837, "step": 5911 }, { "epoch": 0.92, "grad_norm": 20.87060602279722, "learning_rate": 1.6218497515057593e-05, "loss": 0.7829, "step": 5912 }, { "epoch": 0.92, "grad_norm": 21.086965935111138, "learning_rate": 1.621717674592919e-05, "loss": 0.7273, "step": 5913 }, { "epoch": 0.92, "grad_norm": 19.247245796677582, "learning_rate": 1.621585579998792e-05, "loss": 0.7406, "step": 5914 }, { "epoch": 0.92, "grad_norm": 21.437697422186208, "learning_rate": 1.6214534677271344e-05, "loss": 0.6834, "step": 5915 }, { "epoch": 0.92, "grad_norm": 14.005778743206983, "learning_rate": 1.6213213377817035e-05, "loss": 0.6712, "step": 5916 }, { "epoch": 0.92, "grad_norm": 13.28936195148279, "learning_rate": 1.6211891901662576e-05, "loss": 0.6866, "step": 5917 }, { "epoch": 0.92, "grad_norm": 17.603544973130766, "learning_rate": 1.621057024884555e-05, "loss": 0.7248, "step": 5918 }, { "epoch": 0.92, "grad_norm": 15.067000101702876, "learning_rate": 1.620924841940353e-05, "loss": 0.6844, "step": 5919 }, { "epoch": 0.92, "grad_norm": 21.65909563224819, "learning_rate": 1.6207926413374124e-05, "loss": 0.7633, "step": 5920 }, { "epoch": 0.92, "grad_norm": 21.061215646648854, "learning_rate": 1.620660423079492e-05, "loss": 0.7529, "step": 5921 }, { "epoch": 0.93, "grad_norm": 29.69768886484089, "learning_rate": 1.620528187170352e-05, "loss": 0.6891, "step": 5922 }, { "epoch": 0.93, "grad_norm": 19.060081408626154, "learning_rate": 1.6203959336137538e-05, "loss": 0.6552, "step": 5923 }, { "epoch": 0.93, "grad_norm": 13.476902658896494, "learning_rate": 1.620263662413458e-05, "loss": 0.6676, "step": 5924 }, { "epoch": 0.93, "grad_norm": 14.676627481719384, "learning_rate": 1.6201313735732265e-05, "loss": 0.6213, "step": 5925 }, { "epoch": 0.93, "grad_norm": 14.727930664639576, "learning_rate": 1.619999067096821e-05, "loss": 0.6276, "step": 5926 }, { "epoch": 0.93, "grad_norm": 19.159879646287866, "learning_rate": 1.6198667429880054e-05, "loss": 0.745, "step": 5927 }, { "epoch": 0.93, "grad_norm": 16.112949457625067, "learning_rate": 1.619734401250542e-05, "loss": 0.7489, "step": 5928 }, { "epoch": 0.93, "grad_norm": 15.354114584357836, "learning_rate": 1.6196020418881947e-05, "loss": 0.691, "step": 5929 }, { "epoch": 0.93, "grad_norm": 17.80471084889467, "learning_rate": 1.6194696649047278e-05, "loss": 0.7834, "step": 5930 }, { "epoch": 0.93, "grad_norm": 21.076001775085388, "learning_rate": 1.619337270303906e-05, "loss": 0.6526, "step": 5931 }, { "epoch": 0.93, "grad_norm": 20.21434072043704, "learning_rate": 1.6192048580894946e-05, "loss": 0.7788, "step": 5932 }, { "epoch": 0.93, "grad_norm": 16.87706993176504, "learning_rate": 1.6190724282652594e-05, "loss": 0.7708, "step": 5933 }, { "epoch": 0.93, "grad_norm": 17.404084460435197, "learning_rate": 1.6189399808349663e-05, "loss": 0.6706, "step": 5934 }, { "epoch": 0.93, "grad_norm": 21.152339116005287, "learning_rate": 1.6188075158023825e-05, "loss": 0.7311, "step": 5935 }, { "epoch": 0.93, "grad_norm": 17.15266799359661, "learning_rate": 1.6186750331712744e-05, "loss": 0.7385, "step": 5936 }, { "epoch": 0.93, "grad_norm": 21.42989187876736, "learning_rate": 1.6185425329454106e-05, "loss": 0.7341, "step": 5937 }, { "epoch": 0.93, "grad_norm": 17.289202157347265, "learning_rate": 1.618410015128559e-05, "loss": 0.7335, "step": 5938 }, { "epoch": 0.93, "grad_norm": 14.004547217529506, "learning_rate": 1.6182774797244882e-05, "loss": 0.6932, "step": 5939 }, { "epoch": 0.93, "grad_norm": 38.503908652520025, "learning_rate": 1.6181449267369678e-05, "loss": 0.8746, "step": 5940 }, { "epoch": 0.93, "grad_norm": 19.426658899734946, "learning_rate": 1.6180123561697672e-05, "loss": 0.6522, "step": 5941 }, { "epoch": 0.93, "grad_norm": 17.043240801405812, "learning_rate": 1.6178797680266566e-05, "loss": 0.6749, "step": 5942 }, { "epoch": 0.93, "grad_norm": 21.961162540880018, "learning_rate": 1.6177471623114068e-05, "loss": 0.7554, "step": 5943 }, { "epoch": 0.93, "grad_norm": 19.96801580950909, "learning_rate": 1.6176145390277893e-05, "loss": 0.7957, "step": 5944 }, { "epoch": 0.93, "grad_norm": 22.16415807833147, "learning_rate": 1.6174818981795756e-05, "loss": 0.6562, "step": 5945 }, { "epoch": 0.93, "grad_norm": 16.53753529643634, "learning_rate": 1.6173492397705376e-05, "loss": 0.7054, "step": 5946 }, { "epoch": 0.93, "grad_norm": 17.335069969031636, "learning_rate": 1.6172165638044486e-05, "loss": 0.6904, "step": 5947 }, { "epoch": 0.93, "grad_norm": 19.89637617765377, "learning_rate": 1.6170838702850814e-05, "loss": 0.7049, "step": 5948 }, { "epoch": 0.93, "grad_norm": 27.271486570438373, "learning_rate": 1.61695115921621e-05, "loss": 0.7481, "step": 5949 }, { "epoch": 0.93, "grad_norm": 20.494015256206872, "learning_rate": 1.6168184306016092e-05, "loss": 0.7544, "step": 5950 }, { "epoch": 0.93, "grad_norm": 12.90461561884123, "learning_rate": 1.6166856844450522e-05, "loss": 0.6851, "step": 5951 }, { "epoch": 0.93, "grad_norm": 13.308063031837321, "learning_rate": 1.6165529207503156e-05, "loss": 0.7029, "step": 5952 }, { "epoch": 0.93, "grad_norm": 16.858244233132446, "learning_rate": 1.616420139521174e-05, "loss": 0.7325, "step": 5953 }, { "epoch": 0.93, "grad_norm": 16.418704637893903, "learning_rate": 1.6162873407614047e-05, "loss": 0.7511, "step": 5954 }, { "epoch": 0.93, "grad_norm": 21.27601236079115, "learning_rate": 1.616154524474784e-05, "loss": 0.7231, "step": 5955 }, { "epoch": 0.93, "grad_norm": 20.736412367587192, "learning_rate": 1.616021690665089e-05, "loss": 0.7177, "step": 5956 }, { "epoch": 0.93, "grad_norm": 15.583090055979834, "learning_rate": 1.615888839336097e-05, "loss": 0.6823, "step": 5957 }, { "epoch": 0.93, "grad_norm": 30.736905367607875, "learning_rate": 1.6157559704915873e-05, "loss": 0.7765, "step": 5958 }, { "epoch": 0.93, "grad_norm": 22.72996953313437, "learning_rate": 1.6156230841353376e-05, "loss": 0.6867, "step": 5959 }, { "epoch": 0.93, "grad_norm": 21.616841350675035, "learning_rate": 1.6154901802711277e-05, "loss": 0.6481, "step": 5960 }, { "epoch": 0.93, "grad_norm": 14.959188555290131, "learning_rate": 1.615357258902737e-05, "loss": 0.6785, "step": 5961 }, { "epoch": 0.93, "grad_norm": 22.279277267440737, "learning_rate": 1.6152243200339462e-05, "loss": 0.7434, "step": 5962 }, { "epoch": 0.93, "grad_norm": 19.872918304102438, "learning_rate": 1.6150913636685356e-05, "loss": 0.8089, "step": 5963 }, { "epoch": 0.93, "grad_norm": 22.24059477336781, "learning_rate": 1.614958389810286e-05, "loss": 0.6465, "step": 5964 }, { "epoch": 0.93, "grad_norm": 15.286999985837308, "learning_rate": 1.6148253984629798e-05, "loss": 0.7133, "step": 5965 }, { "epoch": 0.93, "grad_norm": 33.91779360560661, "learning_rate": 1.614692389630399e-05, "loss": 0.7208, "step": 5966 }, { "epoch": 0.93, "grad_norm": 18.756621761095253, "learning_rate": 1.6145593633163258e-05, "loss": 0.7476, "step": 5967 }, { "epoch": 0.93, "grad_norm": 13.215467946248461, "learning_rate": 1.614426319524544e-05, "loss": 0.7045, "step": 5968 }, { "epoch": 0.93, "grad_norm": 21.00972725245103, "learning_rate": 1.6142932582588375e-05, "loss": 0.7449, "step": 5969 }, { "epoch": 0.93, "grad_norm": 17.352999536228005, "learning_rate": 1.61416017952299e-05, "loss": 0.7178, "step": 5970 }, { "epoch": 0.93, "grad_norm": 16.142891198991848, "learning_rate": 1.614027083320786e-05, "loss": 0.6726, "step": 5971 }, { "epoch": 0.93, "grad_norm": 20.701998655521543, "learning_rate": 1.613893969656011e-05, "loss": 0.805, "step": 5972 }, { "epoch": 0.93, "grad_norm": 12.263784887695612, "learning_rate": 1.6137608385324505e-05, "loss": 0.6631, "step": 5973 }, { "epoch": 0.93, "grad_norm": 17.414910505893396, "learning_rate": 1.613627689953891e-05, "loss": 0.6877, "step": 5974 }, { "epoch": 0.93, "grad_norm": 11.738871387239564, "learning_rate": 1.613494523924119e-05, "loss": 0.6301, "step": 5975 }, { "epoch": 0.93, "grad_norm": 21.638921087368857, "learning_rate": 1.6133613404469216e-05, "loss": 0.7647, "step": 5976 }, { "epoch": 0.93, "grad_norm": 15.135337843386694, "learning_rate": 1.6132281395260863e-05, "loss": 0.6678, "step": 5977 }, { "epoch": 0.93, "grad_norm": 27.808605120678386, "learning_rate": 1.6130949211654017e-05, "loss": 0.6802, "step": 5978 }, { "epoch": 0.93, "grad_norm": 25.32361928761944, "learning_rate": 1.612961685368656e-05, "loss": 0.6662, "step": 5979 }, { "epoch": 0.93, "grad_norm": 21.059270969583373, "learning_rate": 1.6128284321396385e-05, "loss": 0.6913, "step": 5980 }, { "epoch": 0.93, "grad_norm": 29.265600688677793, "learning_rate": 1.6126951614821388e-05, "loss": 0.8176, "step": 5981 }, { "epoch": 0.93, "grad_norm": 19.253895606148106, "learning_rate": 1.6125618733999476e-05, "loss": 0.7087, "step": 5982 }, { "epoch": 0.93, "grad_norm": 26.790769623312986, "learning_rate": 1.6124285678968546e-05, "loss": 0.6605, "step": 5983 }, { "epoch": 0.93, "grad_norm": 18.50153300660963, "learning_rate": 1.6122952449766516e-05, "loss": 0.7991, "step": 5984 }, { "epoch": 0.93, "grad_norm": 17.855735052856446, "learning_rate": 1.6121619046431297e-05, "loss": 0.7114, "step": 5985 }, { "epoch": 0.94, "grad_norm": 27.134727186948577, "learning_rate": 1.612028546900082e-05, "loss": 0.8299, "step": 5986 }, { "epoch": 0.94, "grad_norm": 22.983470155317562, "learning_rate": 1.6118951717513e-05, "loss": 0.6333, "step": 5987 }, { "epoch": 0.94, "grad_norm": 13.225499314885921, "learning_rate": 1.611761779200577e-05, "loss": 0.6733, "step": 5988 }, { "epoch": 0.94, "grad_norm": 17.23512549465239, "learning_rate": 1.6116283692517075e-05, "loss": 0.7622, "step": 5989 }, { "epoch": 0.94, "grad_norm": 16.93436108514862, "learning_rate": 1.6114949419084846e-05, "loss": 0.6438, "step": 5990 }, { "epoch": 0.94, "grad_norm": 26.437615919648117, "learning_rate": 1.6113614971747032e-05, "loss": 0.7415, "step": 5991 }, { "epoch": 0.94, "grad_norm": 32.42633262967853, "learning_rate": 1.6112280350541584e-05, "loss": 0.7304, "step": 5992 }, { "epoch": 0.94, "grad_norm": 18.67108212467494, "learning_rate": 1.6110945555506462e-05, "loss": 0.69, "step": 5993 }, { "epoch": 0.94, "grad_norm": 19.339503193389636, "learning_rate": 1.6109610586679624e-05, "loss": 0.7079, "step": 5994 }, { "epoch": 0.94, "grad_norm": 29.556020902744823, "learning_rate": 1.610827544409903e-05, "loss": 0.7038, "step": 5995 }, { "epoch": 0.94, "grad_norm": 16.64112554880865, "learning_rate": 1.6106940127802658e-05, "loss": 0.7173, "step": 5996 }, { "epoch": 0.94, "grad_norm": 9.624462066208018, "learning_rate": 1.610560463782848e-05, "loss": 0.7088, "step": 5997 }, { "epoch": 0.94, "grad_norm": 31.290247628273566, "learning_rate": 1.610426897421448e-05, "loss": 0.802, "step": 5998 }, { "epoch": 0.94, "grad_norm": 13.606065118792602, "learning_rate": 1.610293313699864e-05, "loss": 0.6306, "step": 5999 }, { "epoch": 0.94, "grad_norm": 29.70868308336053, "learning_rate": 1.610159712621895e-05, "loss": 0.6919, "step": 6000 }, { "epoch": 0.94, "grad_norm": 16.370156368943082, "learning_rate": 1.610026094191341e-05, "loss": 0.7404, "step": 6001 }, { "epoch": 0.94, "grad_norm": 15.185485222007488, "learning_rate": 1.6098924584120017e-05, "loss": 0.7808, "step": 6002 }, { "epoch": 0.94, "grad_norm": 14.173276347237808, "learning_rate": 1.6097588052876774e-05, "loss": 0.7459, "step": 6003 }, { "epoch": 0.94, "grad_norm": 20.47823266964914, "learning_rate": 1.60962513482217e-05, "loss": 0.7217, "step": 6004 }, { "epoch": 0.94, "grad_norm": 16.99578884853381, "learning_rate": 1.6094914470192796e-05, "loss": 0.656, "step": 6005 }, { "epoch": 0.94, "grad_norm": 18.3157444486773, "learning_rate": 1.6093577418828094e-05, "loss": 0.6596, "step": 6006 }, { "epoch": 0.94, "grad_norm": 26.44876873752978, "learning_rate": 1.6092240194165618e-05, "loss": 0.7488, "step": 6007 }, { "epoch": 0.94, "grad_norm": 19.546222867233915, "learning_rate": 1.6090902796243386e-05, "loss": 0.698, "step": 6008 }, { "epoch": 0.94, "grad_norm": 22.527975887467075, "learning_rate": 1.608956522509945e-05, "loss": 0.6801, "step": 6009 }, { "epoch": 0.94, "grad_norm": 35.15430544752382, "learning_rate": 1.6088227480771835e-05, "loss": 0.7016, "step": 6010 }, { "epoch": 0.94, "grad_norm": 14.079673579247963, "learning_rate": 1.6086889563298594e-05, "loss": 0.6352, "step": 6011 }, { "epoch": 0.94, "grad_norm": 18.300598659846223, "learning_rate": 1.6085551472717775e-05, "loss": 0.6901, "step": 6012 }, { "epoch": 0.94, "grad_norm": 17.018386023503165, "learning_rate": 1.6084213209067433e-05, "loss": 0.7082, "step": 6013 }, { "epoch": 0.94, "grad_norm": 14.904186541309263, "learning_rate": 1.6082874772385628e-05, "loss": 0.6874, "step": 6014 }, { "epoch": 0.94, "grad_norm": 14.714295879602227, "learning_rate": 1.6081536162710418e-05, "loss": 0.6371, "step": 6015 }, { "epoch": 0.94, "grad_norm": 17.127558041170627, "learning_rate": 1.6080197380079883e-05, "loss": 0.7249, "step": 6016 }, { "epoch": 0.94, "grad_norm": 20.548216729359247, "learning_rate": 1.6078858424532088e-05, "loss": 0.6625, "step": 6017 }, { "epoch": 0.94, "grad_norm": 17.252166740678206, "learning_rate": 1.6077519296105112e-05, "loss": 0.7389, "step": 6018 }, { "epoch": 0.94, "grad_norm": 19.756865178148285, "learning_rate": 1.6076179994837045e-05, "loss": 0.7765, "step": 6019 }, { "epoch": 0.94, "grad_norm": 21.465302958969573, "learning_rate": 1.6074840520765976e-05, "loss": 0.7671, "step": 6020 }, { "epoch": 0.94, "grad_norm": 16.76529910418772, "learning_rate": 1.607350087393e-05, "loss": 0.7215, "step": 6021 }, { "epoch": 0.94, "grad_norm": 24.15083210407022, "learning_rate": 1.60721610543672e-05, "loss": 0.8208, "step": 6022 }, { "epoch": 0.94, "grad_norm": 21.675696418363845, "learning_rate": 1.6070821062115704e-05, "loss": 0.7002, "step": 6023 }, { "epoch": 0.94, "grad_norm": 17.67845007038022, "learning_rate": 1.60694808972136e-05, "loss": 0.7252, "step": 6024 }, { "epoch": 0.94, "grad_norm": 11.678972298276326, "learning_rate": 1.606814055969901e-05, "loss": 0.5876, "step": 6025 }, { "epoch": 0.94, "grad_norm": 16.47609256700064, "learning_rate": 1.606680004961006e-05, "loss": 0.6981, "step": 6026 }, { "epoch": 0.94, "grad_norm": 20.30427774542063, "learning_rate": 1.6065459366984856e-05, "loss": 0.6276, "step": 6027 }, { "epoch": 0.94, "grad_norm": 19.338619783642514, "learning_rate": 1.6064118511861543e-05, "loss": 0.7362, "step": 6028 }, { "epoch": 0.94, "grad_norm": 17.555603238376158, "learning_rate": 1.6062777484278245e-05, "loss": 0.6852, "step": 6029 }, { "epoch": 0.94, "grad_norm": 21.551533864113665, "learning_rate": 1.60614362842731e-05, "loss": 0.7836, "step": 6030 }, { "epoch": 0.94, "grad_norm": 17.280541871102116, "learning_rate": 1.6060094911884255e-05, "loss": 0.6696, "step": 6031 }, { "epoch": 0.94, "grad_norm": 17.955626101863103, "learning_rate": 1.6058753367149854e-05, "loss": 0.7, "step": 6032 }, { "epoch": 0.94, "grad_norm": 20.42183603685802, "learning_rate": 1.6057411650108053e-05, "loss": 0.7145, "step": 6033 }, { "epoch": 0.94, "grad_norm": 19.09152443537757, "learning_rate": 1.6056069760797006e-05, "loss": 0.7615, "step": 6034 }, { "epoch": 0.94, "grad_norm": 18.943154608542795, "learning_rate": 1.605472769925488e-05, "loss": 0.701, "step": 6035 }, { "epoch": 0.94, "grad_norm": 21.687209125321623, "learning_rate": 1.6053385465519838e-05, "loss": 0.6862, "step": 6036 }, { "epoch": 0.94, "grad_norm": 18.024486516594113, "learning_rate": 1.6052043059630058e-05, "loss": 0.699, "step": 6037 }, { "epoch": 0.94, "grad_norm": 16.422543449254004, "learning_rate": 1.6050700481623712e-05, "loss": 0.6818, "step": 6038 }, { "epoch": 0.94, "grad_norm": 24.572686157436372, "learning_rate": 1.604935773153898e-05, "loss": 0.7413, "step": 6039 }, { "epoch": 0.94, "grad_norm": 18.045914884516726, "learning_rate": 1.6048014809414057e-05, "loss": 0.7308, "step": 6040 }, { "epoch": 0.94, "grad_norm": 15.39324619124694, "learning_rate": 1.604667171528713e-05, "loss": 0.8003, "step": 6041 }, { "epoch": 0.94, "grad_norm": 12.848160861965429, "learning_rate": 1.60453284491964e-05, "loss": 0.6883, "step": 6042 }, { "epoch": 0.94, "grad_norm": 24.661372682422005, "learning_rate": 1.6043985011180062e-05, "loss": 0.7748, "step": 6043 }, { "epoch": 0.94, "grad_norm": 13.149216708158821, "learning_rate": 1.6042641401276326e-05, "loss": 0.666, "step": 6044 }, { "epoch": 0.94, "grad_norm": 17.574026378062864, "learning_rate": 1.6041297619523405e-05, "loss": 0.685, "step": 6045 }, { "epoch": 0.94, "grad_norm": 16.945768089026576, "learning_rate": 1.603995366595951e-05, "loss": 0.7174, "step": 6046 }, { "epoch": 0.94, "grad_norm": 21.517366707254787, "learning_rate": 1.603860954062287e-05, "loss": 0.685, "step": 6047 }, { "epoch": 0.94, "grad_norm": 15.701941173207373, "learning_rate": 1.6037265243551704e-05, "loss": 0.6415, "step": 6048 }, { "epoch": 0.94, "grad_norm": 24.461719324391474, "learning_rate": 1.6035920774784247e-05, "loss": 0.7319, "step": 6049 }, { "epoch": 0.95, "grad_norm": 14.784512617949902, "learning_rate": 1.6034576134358736e-05, "loss": 0.7536, "step": 6050 }, { "epoch": 0.95, "grad_norm": 21.90841418621444, "learning_rate": 1.6033231322313406e-05, "loss": 0.7027, "step": 6051 }, { "epoch": 0.95, "grad_norm": 14.118060865550813, "learning_rate": 1.603188633868651e-05, "loss": 0.7073, "step": 6052 }, { "epoch": 0.95, "grad_norm": 17.89676469955705, "learning_rate": 1.6030541183516292e-05, "loss": 0.783, "step": 6053 }, { "epoch": 0.95, "grad_norm": 19.16690837320704, "learning_rate": 1.6029195856841015e-05, "loss": 0.6625, "step": 6054 }, { "epoch": 0.95, "grad_norm": 12.620054953732446, "learning_rate": 1.602785035869893e-05, "loss": 0.6994, "step": 6055 }, { "epoch": 0.95, "grad_norm": 22.97839752050187, "learning_rate": 1.602650468912831e-05, "loss": 0.7622, "step": 6056 }, { "epoch": 0.95, "grad_norm": 16.639519712070097, "learning_rate": 1.602515884816742e-05, "loss": 0.6513, "step": 6057 }, { "epoch": 0.95, "grad_norm": 18.51940529456008, "learning_rate": 1.602381283585454e-05, "loss": 0.8177, "step": 6058 }, { "epoch": 0.95, "grad_norm": 18.49837469842179, "learning_rate": 1.6022466652227944e-05, "loss": 0.7179, "step": 6059 }, { "epoch": 0.95, "grad_norm": 18.829366897843116, "learning_rate": 1.602112029732592e-05, "loss": 0.7255, "step": 6060 }, { "epoch": 0.95, "grad_norm": 18.660995152306217, "learning_rate": 1.6019773771186756e-05, "loss": 0.792, "step": 6061 }, { "epoch": 0.95, "grad_norm": 17.198536408158073, "learning_rate": 1.601842707384875e-05, "loss": 0.7614, "step": 6062 }, { "epoch": 0.95, "grad_norm": 21.285158852870904, "learning_rate": 1.60170802053502e-05, "loss": 0.7865, "step": 6063 }, { "epoch": 0.95, "grad_norm": 15.801970257246602, "learning_rate": 1.6015733165729406e-05, "loss": 0.6287, "step": 6064 }, { "epoch": 0.95, "grad_norm": 18.967032262083727, "learning_rate": 1.601438595502468e-05, "loss": 0.7677, "step": 6065 }, { "epoch": 0.95, "grad_norm": 19.766403354804233, "learning_rate": 1.6013038573274336e-05, "loss": 0.719, "step": 6066 }, { "epoch": 0.95, "grad_norm": 18.585718313963557, "learning_rate": 1.6011691020516696e-05, "loss": 0.7175, "step": 6067 }, { "epoch": 0.95, "grad_norm": 20.55071192695079, "learning_rate": 1.6010343296790074e-05, "loss": 0.7438, "step": 6068 }, { "epoch": 0.95, "grad_norm": 16.923081299536978, "learning_rate": 1.6008995402132807e-05, "loss": 0.7245, "step": 6069 }, { "epoch": 0.95, "grad_norm": 19.24145848716582, "learning_rate": 1.6007647336583227e-05, "loss": 0.6264, "step": 6070 }, { "epoch": 0.95, "grad_norm": 16.482393296613985, "learning_rate": 1.600629910017967e-05, "loss": 0.6931, "step": 6071 }, { "epoch": 0.95, "grad_norm": 18.306325195697898, "learning_rate": 1.600495069296048e-05, "loss": 0.6905, "step": 6072 }, { "epoch": 0.95, "grad_norm": 15.890905039259907, "learning_rate": 1.6003602114964007e-05, "loss": 0.6545, "step": 6073 }, { "epoch": 0.95, "grad_norm": 17.400955227430988, "learning_rate": 1.60022533662286e-05, "loss": 0.7456, "step": 6074 }, { "epoch": 0.95, "grad_norm": 18.88797996672957, "learning_rate": 1.600090444679262e-05, "loss": 0.7343, "step": 6075 }, { "epoch": 0.95, "grad_norm": 16.141422330507574, "learning_rate": 1.5999555356694426e-05, "loss": 0.6491, "step": 6076 }, { "epoch": 0.95, "grad_norm": 13.750025277764754, "learning_rate": 1.599820609597239e-05, "loss": 0.6739, "step": 6077 }, { "epoch": 0.95, "grad_norm": 20.70318905442603, "learning_rate": 1.5996856664664876e-05, "loss": 0.6953, "step": 6078 }, { "epoch": 0.95, "grad_norm": 16.838339184236734, "learning_rate": 1.599550706281027e-05, "loss": 0.8145, "step": 6079 }, { "epoch": 0.95, "grad_norm": 43.52356748710453, "learning_rate": 1.5994157290446953e-05, "loss": 0.6234, "step": 6080 }, { "epoch": 0.95, "grad_norm": 13.221769869464637, "learning_rate": 1.5992807347613305e-05, "loss": 0.7163, "step": 6081 }, { "epoch": 0.95, "grad_norm": 22.25877150012426, "learning_rate": 1.5991457234347725e-05, "loss": 0.7037, "step": 6082 }, { "epoch": 0.95, "grad_norm": 17.356749499155963, "learning_rate": 1.5990106950688605e-05, "loss": 0.6282, "step": 6083 }, { "epoch": 0.95, "grad_norm": 18.581752400514258, "learning_rate": 1.5988756496674345e-05, "loss": 0.6403, "step": 6084 }, { "epoch": 0.95, "grad_norm": 15.290055549234074, "learning_rate": 1.5987405872343358e-05, "loss": 0.7299, "step": 6085 }, { "epoch": 0.95, "grad_norm": 35.39351909164021, "learning_rate": 1.5986055077734046e-05, "loss": 0.6297, "step": 6086 }, { "epoch": 0.95, "grad_norm": 15.03817085595391, "learning_rate": 1.5984704112884834e-05, "loss": 0.7216, "step": 6087 }, { "epoch": 0.95, "grad_norm": 157.62999318720173, "learning_rate": 1.5983352977834132e-05, "loss": 0.7822, "step": 6088 }, { "epoch": 0.95, "grad_norm": 12.102644789452551, "learning_rate": 1.5982001672620376e-05, "loss": 0.6764, "step": 6089 }, { "epoch": 0.95, "grad_norm": 16.830539053822466, "learning_rate": 1.5980650197281994e-05, "loss": 0.784, "step": 6090 }, { "epoch": 0.95, "grad_norm": 16.20774198726569, "learning_rate": 1.597929855185741e-05, "loss": 0.7037, "step": 6091 }, { "epoch": 0.95, "grad_norm": 17.601887583958202, "learning_rate": 1.597794673638508e-05, "loss": 0.6495, "step": 6092 }, { "epoch": 0.95, "grad_norm": 13.91086018416681, "learning_rate": 1.597659475090344e-05, "loss": 0.6718, "step": 6093 }, { "epoch": 0.95, "grad_norm": 16.735718613066542, "learning_rate": 1.5975242595450946e-05, "loss": 0.6484, "step": 6094 }, { "epoch": 0.95, "grad_norm": 21.485949560554143, "learning_rate": 1.5973890270066042e-05, "loss": 0.694, "step": 6095 }, { "epoch": 0.95, "grad_norm": 13.755962155331988, "learning_rate": 1.5972537774787194e-05, "loss": 0.6725, "step": 6096 }, { "epoch": 0.95, "grad_norm": 25.94634781643838, "learning_rate": 1.597118510965287e-05, "loss": 0.8582, "step": 6097 }, { "epoch": 0.95, "grad_norm": 15.60803522865118, "learning_rate": 1.5969832274701528e-05, "loss": 0.803, "step": 6098 }, { "epoch": 0.95, "grad_norm": 23.74992923318208, "learning_rate": 1.5968479269971654e-05, "loss": 0.8157, "step": 6099 }, { "epoch": 0.95, "grad_norm": 23.218071783898644, "learning_rate": 1.596712609550172e-05, "loss": 0.736, "step": 6100 }, { "epoch": 0.95, "grad_norm": 12.684420037970854, "learning_rate": 1.5965772751330207e-05, "loss": 0.6614, "step": 6101 }, { "epoch": 0.95, "grad_norm": 15.545800636779923, "learning_rate": 1.596441923749561e-05, "loss": 0.6814, "step": 6102 }, { "epoch": 0.95, "grad_norm": 14.094761465892764, "learning_rate": 1.5963065554036418e-05, "loss": 0.6432, "step": 6103 }, { "epoch": 0.95, "grad_norm": 26.097896790077538, "learning_rate": 1.596171170099113e-05, "loss": 0.775, "step": 6104 }, { "epoch": 0.95, "grad_norm": 16.60526115145929, "learning_rate": 1.596035767839825e-05, "loss": 0.7764, "step": 6105 }, { "epoch": 0.95, "grad_norm": 16.457857204946226, "learning_rate": 1.595900348629628e-05, "loss": 0.7346, "step": 6106 }, { "epoch": 0.95, "grad_norm": 15.03574651350869, "learning_rate": 1.5957649124723743e-05, "loss": 0.6761, "step": 6107 }, { "epoch": 0.95, "grad_norm": 18.79131792890106, "learning_rate": 1.5956294593719147e-05, "loss": 0.6428, "step": 6108 }, { "epoch": 0.95, "grad_norm": 26.21022323169579, "learning_rate": 1.5954939893321016e-05, "loss": 0.7119, "step": 6109 }, { "epoch": 0.95, "grad_norm": 20.838029359524974, "learning_rate": 1.595358502356788e-05, "loss": 0.7915, "step": 6110 }, { "epoch": 0.95, "grad_norm": 19.35842348443619, "learning_rate": 1.5952229984498265e-05, "loss": 0.7209, "step": 6111 }, { "epoch": 0.95, "grad_norm": 16.590018111377717, "learning_rate": 1.5950874776150715e-05, "loss": 0.7623, "step": 6112 }, { "epoch": 0.95, "grad_norm": 18.048860509236775, "learning_rate": 1.5949519398563766e-05, "loss": 0.767, "step": 6113 }, { "epoch": 0.96, "grad_norm": 30.0661945402471, "learning_rate": 1.594816385177597e-05, "loss": 0.6822, "step": 6114 }, { "epoch": 0.96, "grad_norm": 13.486607023168693, "learning_rate": 1.5946808135825866e-05, "loss": 0.6685, "step": 6115 }, { "epoch": 0.96, "grad_norm": 17.553040729827732, "learning_rate": 1.5945452250752022e-05, "loss": 0.7848, "step": 6116 }, { "epoch": 0.96, "grad_norm": 24.321822710506687, "learning_rate": 1.5944096196592994e-05, "loss": 0.7468, "step": 6117 }, { "epoch": 0.96, "grad_norm": 18.95007752930136, "learning_rate": 1.5942739973387345e-05, "loss": 0.6842, "step": 6118 }, { "epoch": 0.96, "grad_norm": 16.568679781289745, "learning_rate": 1.594138358117365e-05, "loss": 0.782, "step": 6119 }, { "epoch": 0.96, "grad_norm": 19.57447788772685, "learning_rate": 1.594002701999048e-05, "loss": 0.745, "step": 6120 }, { "epoch": 0.96, "grad_norm": 19.92677475008517, "learning_rate": 1.5938670289876418e-05, "loss": 0.7654, "step": 6121 }, { "epoch": 0.96, "grad_norm": 20.45918067710494, "learning_rate": 1.5937313390870045e-05, "loss": 0.666, "step": 6122 }, { "epoch": 0.96, "grad_norm": 24.467953735978377, "learning_rate": 1.5935956323009954e-05, "loss": 0.7535, "step": 6123 }, { "epoch": 0.96, "grad_norm": 13.270819109929526, "learning_rate": 1.5934599086334733e-05, "loss": 0.618, "step": 6124 }, { "epoch": 0.96, "grad_norm": 20.83316831777706, "learning_rate": 1.593324168088299e-05, "loss": 0.7432, "step": 6125 }, { "epoch": 0.96, "grad_norm": 11.483902816731982, "learning_rate": 1.5931884106693326e-05, "loss": 0.6655, "step": 6126 }, { "epoch": 0.96, "grad_norm": 21.763631881407356, "learning_rate": 1.5930526363804344e-05, "loss": 0.7791, "step": 6127 }, { "epoch": 0.96, "grad_norm": 21.520092686101282, "learning_rate": 1.5929168452254664e-05, "loss": 0.7591, "step": 6128 }, { "epoch": 0.96, "grad_norm": 20.464502903706038, "learning_rate": 1.5927810372082903e-05, "loss": 0.7043, "step": 6129 }, { "epoch": 0.96, "grad_norm": 13.67035408170661, "learning_rate": 1.592645212332768e-05, "loss": 0.674, "step": 6130 }, { "epoch": 0.96, "grad_norm": 23.229569894486914, "learning_rate": 1.5925093706027625e-05, "loss": 0.6965, "step": 6131 }, { "epoch": 0.96, "grad_norm": 24.16192466042968, "learning_rate": 1.5923735120221372e-05, "loss": 0.7232, "step": 6132 }, { "epoch": 0.96, "grad_norm": 17.857956823309355, "learning_rate": 1.5922376365947557e-05, "loss": 0.6704, "step": 6133 }, { "epoch": 0.96, "grad_norm": 15.807482009906417, "learning_rate": 1.5921017443244825e-05, "loss": 0.7533, "step": 6134 }, { "epoch": 0.96, "grad_norm": 30.69490538977884, "learning_rate": 1.5919658352151818e-05, "loss": 0.6534, "step": 6135 }, { "epoch": 0.96, "grad_norm": 15.612192071799736, "learning_rate": 1.5918299092707194e-05, "loss": 0.707, "step": 6136 }, { "epoch": 0.96, "grad_norm": 18.23309693192679, "learning_rate": 1.5916939664949602e-05, "loss": 0.6635, "step": 6137 }, { "epoch": 0.96, "grad_norm": 28.64551615845766, "learning_rate": 1.591558006891771e-05, "loss": 0.7467, "step": 6138 }, { "epoch": 0.96, "grad_norm": 22.691285514904354, "learning_rate": 1.5914220304650177e-05, "loss": 0.693, "step": 6139 }, { "epoch": 0.96, "grad_norm": 13.408445064087257, "learning_rate": 1.5912860372185683e-05, "loss": 0.7431, "step": 6140 }, { "epoch": 0.96, "grad_norm": 17.917755096572193, "learning_rate": 1.5911500271562895e-05, "loss": 0.726, "step": 6141 }, { "epoch": 0.96, "grad_norm": 16.26411704320793, "learning_rate": 1.5910140002820503e-05, "loss": 0.5939, "step": 6142 }, { "epoch": 0.96, "grad_norm": 21.180750087450953, "learning_rate": 1.5908779565997182e-05, "loss": 0.6788, "step": 6143 }, { "epoch": 0.96, "grad_norm": 28.832142271379315, "learning_rate": 1.5907418961131628e-05, "loss": 0.8077, "step": 6144 }, { "epoch": 0.96, "grad_norm": 23.37061146476504, "learning_rate": 1.5906058188262534e-05, "loss": 0.6322, "step": 6145 }, { "epoch": 0.96, "grad_norm": 18.14986540833794, "learning_rate": 1.5904697247428602e-05, "loss": 0.6985, "step": 6146 }, { "epoch": 0.96, "grad_norm": 16.1414003195211, "learning_rate": 1.5903336138668532e-05, "loss": 0.6955, "step": 6147 }, { "epoch": 0.96, "grad_norm": 18.395505968840943, "learning_rate": 1.5901974862021036e-05, "loss": 0.753, "step": 6148 }, { "epoch": 0.96, "grad_norm": 14.91349934992944, "learning_rate": 1.5900613417524827e-05, "loss": 0.744, "step": 6149 }, { "epoch": 0.96, "grad_norm": 14.357660189111426, "learning_rate": 1.5899251805218625e-05, "loss": 0.7316, "step": 6150 }, { "epoch": 0.96, "grad_norm": 27.979583490308016, "learning_rate": 1.5897890025141153e-05, "loss": 0.8225, "step": 6151 }, { "epoch": 0.96, "grad_norm": 17.383404537791456, "learning_rate": 1.589652807733114e-05, "loss": 0.6453, "step": 6152 }, { "epoch": 0.96, "grad_norm": 14.601501822409327, "learning_rate": 1.5895165961827315e-05, "loss": 0.6967, "step": 6153 }, { "epoch": 0.96, "grad_norm": 26.27885179032878, "learning_rate": 1.5893803678668424e-05, "loss": 0.6976, "step": 6154 }, { "epoch": 0.96, "grad_norm": 29.475166394151895, "learning_rate": 1.5892441227893197e-05, "loss": 0.7154, "step": 6155 }, { "epoch": 0.96, "grad_norm": 19.082518067099613, "learning_rate": 1.589107860954039e-05, "loss": 0.768, "step": 6156 }, { "epoch": 0.96, "grad_norm": 17.038815927273703, "learning_rate": 1.588971582364876e-05, "loss": 0.6498, "step": 6157 }, { "epoch": 0.96, "grad_norm": 38.421322803865195, "learning_rate": 1.588835287025705e-05, "loss": 0.7573, "step": 6158 }, { "epoch": 0.96, "grad_norm": 21.50476274161912, "learning_rate": 1.588698974940403e-05, "loss": 0.6594, "step": 6159 }, { "epoch": 0.96, "grad_norm": 19.396100140412013, "learning_rate": 1.5885626461128467e-05, "loss": 0.7322, "step": 6160 }, { "epoch": 0.96, "grad_norm": 20.76339265350174, "learning_rate": 1.588426300546913e-05, "loss": 0.7459, "step": 6161 }, { "epoch": 0.96, "grad_norm": 28.619583614843886, "learning_rate": 1.5882899382464795e-05, "loss": 0.7357, "step": 6162 }, { "epoch": 0.96, "grad_norm": 18.08433901353271, "learning_rate": 1.5881535592154245e-05, "loss": 0.7771, "step": 6163 }, { "epoch": 0.96, "grad_norm": 17.57934639374916, "learning_rate": 1.5880171634576262e-05, "loss": 0.7524, "step": 6164 }, { "epoch": 0.96, "grad_norm": 18.70357114893932, "learning_rate": 1.587880750976964e-05, "loss": 0.6779, "step": 6165 }, { "epoch": 0.96, "grad_norm": 17.966815701738245, "learning_rate": 1.5877443217773166e-05, "loss": 0.7679, "step": 6166 }, { "epoch": 0.96, "grad_norm": 20.02784629787974, "learning_rate": 1.5876078758625647e-05, "loss": 0.7229, "step": 6167 }, { "epoch": 0.96, "grad_norm": 14.905083016833325, "learning_rate": 1.5874714132365887e-05, "loss": 0.7336, "step": 6168 }, { "epoch": 0.96, "grad_norm": 14.604266264866526, "learning_rate": 1.5873349339032693e-05, "loss": 0.6868, "step": 6169 }, { "epoch": 0.96, "grad_norm": 13.742797501879371, "learning_rate": 1.587198437866488e-05, "loss": 0.6729, "step": 6170 }, { "epoch": 0.96, "grad_norm": 19.6087345679823, "learning_rate": 1.5870619251301267e-05, "loss": 0.7424, "step": 6171 }, { "epoch": 0.96, "grad_norm": 14.605849624038399, "learning_rate": 1.5869253956980672e-05, "loss": 0.6427, "step": 6172 }, { "epoch": 0.96, "grad_norm": 26.16999020180721, "learning_rate": 1.5867888495741934e-05, "loss": 0.7078, "step": 6173 }, { "epoch": 0.96, "grad_norm": 16.37528765949897, "learning_rate": 1.5866522867623876e-05, "loss": 0.7846, "step": 6174 }, { "epoch": 0.96, "grad_norm": 16.227007441281597, "learning_rate": 1.586515707266534e-05, "loss": 0.6812, "step": 6175 }, { "epoch": 0.96, "grad_norm": 17.90701736600108, "learning_rate": 1.5863791110905172e-05, "loss": 0.7431, "step": 6176 }, { "epoch": 0.96, "grad_norm": 12.573309327177764, "learning_rate": 1.586242498238221e-05, "loss": 0.6916, "step": 6177 }, { "epoch": 0.97, "grad_norm": 23.951134870883674, "learning_rate": 1.5861058687135314e-05, "loss": 0.8912, "step": 6178 }, { "epoch": 0.97, "grad_norm": 14.686810916603525, "learning_rate": 1.5859692225203335e-05, "loss": 0.6475, "step": 6179 }, { "epoch": 0.97, "grad_norm": 18.335740569056334, "learning_rate": 1.585832559662514e-05, "loss": 0.66, "step": 6180 }, { "epoch": 0.97, "grad_norm": 15.311238841262115, "learning_rate": 1.585695880143959e-05, "loss": 0.7251, "step": 6181 }, { "epoch": 0.97, "grad_norm": 19.062592071560996, "learning_rate": 1.5855591839685556e-05, "loss": 0.7371, "step": 6182 }, { "epoch": 0.97, "grad_norm": 20.24589320544766, "learning_rate": 1.585422471140192e-05, "loss": 0.7716, "step": 6183 }, { "epoch": 0.97, "grad_norm": 13.928098998528, "learning_rate": 1.5852857416627552e-05, "loss": 0.6516, "step": 6184 }, { "epoch": 0.97, "grad_norm": 28.28717905668244, "learning_rate": 1.585148995540135e-05, "loss": 0.7295, "step": 6185 }, { "epoch": 0.97, "grad_norm": 17.73335874813438, "learning_rate": 1.585012232776219e-05, "loss": 0.7134, "step": 6186 }, { "epoch": 0.97, "grad_norm": 19.872737960476485, "learning_rate": 1.584875453374898e-05, "loss": 0.7236, "step": 6187 }, { "epoch": 0.97, "grad_norm": 31.06385799630569, "learning_rate": 1.5847386573400605e-05, "loss": 0.6823, "step": 6188 }, { "epoch": 0.97, "grad_norm": 25.418897220507894, "learning_rate": 1.584601844675598e-05, "loss": 0.8347, "step": 6189 }, { "epoch": 0.97, "grad_norm": 17.48595862593136, "learning_rate": 1.584465015385401e-05, "loss": 0.6605, "step": 6190 }, { "epoch": 0.97, "grad_norm": 15.049688708968837, "learning_rate": 1.584328169473361e-05, "loss": 0.702, "step": 6191 }, { "epoch": 0.97, "grad_norm": 17.301021634979367, "learning_rate": 1.58419130694337e-05, "loss": 0.7398, "step": 6192 }, { "epoch": 0.97, "grad_norm": 18.44046627465891, "learning_rate": 1.5840544277993193e-05, "loss": 0.7179, "step": 6193 }, { "epoch": 0.97, "grad_norm": 13.966605734474095, "learning_rate": 1.5839175320451027e-05, "loss": 0.7269, "step": 6194 }, { "epoch": 0.97, "grad_norm": 25.76288541822951, "learning_rate": 1.583780619684613e-05, "loss": 0.6416, "step": 6195 }, { "epoch": 0.97, "grad_norm": 21.82144262567529, "learning_rate": 1.5836436907217438e-05, "loss": 0.7189, "step": 6196 }, { "epoch": 0.97, "grad_norm": 14.499975888359641, "learning_rate": 1.58350674516039e-05, "loss": 0.6673, "step": 6197 }, { "epoch": 0.97, "grad_norm": 12.512830787258151, "learning_rate": 1.583369783004445e-05, "loss": 0.6559, "step": 6198 }, { "epoch": 0.97, "grad_norm": 31.764767172207975, "learning_rate": 1.5832328042578047e-05, "loss": 0.6729, "step": 6199 }, { "epoch": 0.97, "grad_norm": 20.18189812003746, "learning_rate": 1.5830958089243652e-05, "loss": 0.7005, "step": 6200 }, { "epoch": 0.97, "grad_norm": 16.926758816958422, "learning_rate": 1.5829587970080217e-05, "loss": 0.6892, "step": 6201 }, { "epoch": 0.97, "grad_norm": 14.682554361445563, "learning_rate": 1.582821768512671e-05, "loss": 0.6898, "step": 6202 }, { "epoch": 0.97, "grad_norm": 13.80032111620127, "learning_rate": 1.5826847234422102e-05, "loss": 0.5931, "step": 6203 }, { "epoch": 0.97, "grad_norm": 31.578395719655997, "learning_rate": 1.5825476618005363e-05, "loss": 0.7904, "step": 6204 }, { "epoch": 0.97, "grad_norm": 14.478891744221988, "learning_rate": 1.582410583591548e-05, "loss": 0.6769, "step": 6205 }, { "epoch": 0.97, "grad_norm": 17.466471831445197, "learning_rate": 1.5822734888191437e-05, "loss": 0.6986, "step": 6206 }, { "epoch": 0.97, "grad_norm": 19.877754720031678, "learning_rate": 1.5821363774872218e-05, "loss": 0.7492, "step": 6207 }, { "epoch": 0.97, "grad_norm": 20.61935295009391, "learning_rate": 1.5819992495996818e-05, "loss": 0.6531, "step": 6208 }, { "epoch": 0.97, "grad_norm": 26.650594045296724, "learning_rate": 1.5818621051604234e-05, "loss": 0.6713, "step": 6209 }, { "epoch": 0.97, "grad_norm": 13.13050746091519, "learning_rate": 1.5817249441733475e-05, "loss": 0.6609, "step": 6210 }, { "epoch": 0.97, "grad_norm": 20.896457561928887, "learning_rate": 1.5815877666423543e-05, "loss": 0.733, "step": 6211 }, { "epoch": 0.97, "grad_norm": 23.620864088589446, "learning_rate": 1.5814505725713454e-05, "loss": 0.6689, "step": 6212 }, { "epoch": 0.97, "grad_norm": 15.294414013500338, "learning_rate": 1.581313361964222e-05, "loss": 0.6283, "step": 6213 }, { "epoch": 0.97, "grad_norm": 16.857289407117033, "learning_rate": 1.5811761348248872e-05, "loss": 0.6373, "step": 6214 }, { "epoch": 0.97, "grad_norm": 22.443122265630272, "learning_rate": 1.5810388911572424e-05, "loss": 0.7406, "step": 6215 }, { "epoch": 0.97, "grad_norm": 31.48483365011168, "learning_rate": 1.5809016309651916e-05, "loss": 0.7768, "step": 6216 }, { "epoch": 0.97, "grad_norm": 14.811447672544205, "learning_rate": 1.5807643542526387e-05, "loss": 0.749, "step": 6217 }, { "epoch": 0.97, "grad_norm": 24.078537042571483, "learning_rate": 1.580627061023487e-05, "loss": 0.6296, "step": 6218 }, { "epoch": 0.97, "grad_norm": 17.313146821825555, "learning_rate": 1.5804897512816416e-05, "loss": 0.733, "step": 6219 }, { "epoch": 0.97, "grad_norm": 22.227180493819073, "learning_rate": 1.580352425031007e-05, "loss": 0.6935, "step": 6220 }, { "epoch": 0.97, "grad_norm": 20.230630329294836, "learning_rate": 1.5802150822754888e-05, "loss": 0.8322, "step": 6221 }, { "epoch": 0.97, "grad_norm": 14.605318888456289, "learning_rate": 1.580077723018993e-05, "loss": 0.7045, "step": 6222 }, { "epoch": 0.97, "grad_norm": 14.056639086029023, "learning_rate": 1.5799403472654266e-05, "loss": 0.6404, "step": 6223 }, { "epoch": 0.97, "grad_norm": 18.59353403238485, "learning_rate": 1.5798029550186957e-05, "loss": 0.692, "step": 6224 }, { "epoch": 0.97, "grad_norm": 15.638346383175973, "learning_rate": 1.5796655462827077e-05, "loss": 0.6574, "step": 6225 }, { "epoch": 0.97, "grad_norm": 20.147086033456212, "learning_rate": 1.5795281210613708e-05, "loss": 0.6948, "step": 6226 }, { "epoch": 0.97, "grad_norm": 15.524126123015039, "learning_rate": 1.5793906793585935e-05, "loss": 0.6924, "step": 6227 }, { "epoch": 0.97, "grad_norm": 13.899402099788624, "learning_rate": 1.5792532211782837e-05, "loss": 0.6904, "step": 6228 }, { "epoch": 0.97, "grad_norm": 17.832425310976404, "learning_rate": 1.5791157465243517e-05, "loss": 0.7059, "step": 6229 }, { "epoch": 0.97, "grad_norm": 16.189572613082234, "learning_rate": 1.5789782554007063e-05, "loss": 0.7952, "step": 6230 }, { "epoch": 0.97, "grad_norm": 25.967314395288053, "learning_rate": 1.578840747811258e-05, "loss": 0.7456, "step": 6231 }, { "epoch": 0.97, "grad_norm": 18.490645421287475, "learning_rate": 1.5787032237599173e-05, "loss": 0.6799, "step": 6232 }, { "epoch": 0.97, "grad_norm": 17.400697747167488, "learning_rate": 1.5785656832505956e-05, "loss": 0.8336, "step": 6233 }, { "epoch": 0.97, "grad_norm": 25.58609244321543, "learning_rate": 1.5784281262872046e-05, "loss": 0.6934, "step": 6234 }, { "epoch": 0.97, "grad_norm": 12.762061707343198, "learning_rate": 1.5782905528736558e-05, "loss": 0.7065, "step": 6235 }, { "epoch": 0.97, "grad_norm": 20.433959511728663, "learning_rate": 1.578152963013862e-05, "loss": 0.6244, "step": 6236 }, { "epoch": 0.97, "grad_norm": 14.306377118909536, "learning_rate": 1.578015356711736e-05, "loss": 0.6028, "step": 6237 }, { "epoch": 0.97, "grad_norm": 21.675394127297274, "learning_rate": 1.5778777339711914e-05, "loss": 0.8232, "step": 6238 }, { "epoch": 0.97, "grad_norm": 17.718009130944154, "learning_rate": 1.577740094796142e-05, "loss": 0.6942, "step": 6239 }, { "epoch": 0.97, "grad_norm": 24.662029480421435, "learning_rate": 1.5776024391905026e-05, "loss": 0.6917, "step": 6240 }, { "epoch": 0.97, "grad_norm": 32.679605801722104, "learning_rate": 1.5774647671581878e-05, "loss": 0.7454, "step": 6241 }, { "epoch": 0.98, "grad_norm": 22.049254008571342, "learning_rate": 1.5773270787031124e-05, "loss": 0.7414, "step": 6242 }, { "epoch": 0.98, "grad_norm": 24.852855324957886, "learning_rate": 1.577189373829193e-05, "loss": 0.6756, "step": 6243 }, { "epoch": 0.98, "grad_norm": 13.62704146910007, "learning_rate": 1.5770516525403453e-05, "loss": 0.7319, "step": 6244 }, { "epoch": 0.98, "grad_norm": 12.562153743543485, "learning_rate": 1.576913914840486e-05, "loss": 0.6966, "step": 6245 }, { "epoch": 0.98, "grad_norm": 18.28206774969318, "learning_rate": 1.5767761607335327e-05, "loss": 0.7001, "step": 6246 }, { "epoch": 0.98, "grad_norm": 22.321958583315887, "learning_rate": 1.5766383902234026e-05, "loss": 0.6744, "step": 6247 }, { "epoch": 0.98, "grad_norm": 23.249578056813633, "learning_rate": 1.5765006033140142e-05, "loss": 0.6593, "step": 6248 }, { "epoch": 0.98, "grad_norm": 16.780097770914747, "learning_rate": 1.5763628000092858e-05, "loss": 0.7383, "step": 6249 }, { "epoch": 0.98, "grad_norm": 26.861128854522114, "learning_rate": 1.5762249803131365e-05, "loss": 0.7002, "step": 6250 }, { "epoch": 0.98, "grad_norm": 21.856258944628756, "learning_rate": 1.5760871442294856e-05, "loss": 0.8652, "step": 6251 }, { "epoch": 0.98, "grad_norm": 11.94426362208592, "learning_rate": 1.5759492917622537e-05, "loss": 0.6385, "step": 6252 }, { "epoch": 0.98, "grad_norm": 20.115648218541228, "learning_rate": 1.5758114229153606e-05, "loss": 0.7313, "step": 6253 }, { "epoch": 0.98, "grad_norm": 13.829432571142346, "learning_rate": 1.575673537692728e-05, "loss": 0.6776, "step": 6254 }, { "epoch": 0.98, "grad_norm": 26.05287736126642, "learning_rate": 1.575535636098276e-05, "loss": 0.6779, "step": 6255 }, { "epoch": 0.98, "grad_norm": 21.580785969844754, "learning_rate": 1.5753977181359277e-05, "loss": 0.7603, "step": 6256 }, { "epoch": 0.98, "grad_norm": 21.562898043745616, "learning_rate": 1.5752597838096046e-05, "loss": 0.6555, "step": 6257 }, { "epoch": 0.98, "grad_norm": 19.524721865101647, "learning_rate": 1.57512183312323e-05, "loss": 0.693, "step": 6258 }, { "epoch": 0.98, "grad_norm": 20.6832497298869, "learning_rate": 1.574983866080727e-05, "loss": 0.7587, "step": 6259 }, { "epoch": 0.98, "grad_norm": 19.216483982755733, "learning_rate": 1.5748458826860185e-05, "loss": 0.683, "step": 6260 }, { "epoch": 0.98, "grad_norm": 15.548388677261059, "learning_rate": 1.5747078829430302e-05, "loss": 0.7515, "step": 6261 }, { "epoch": 0.98, "grad_norm": 21.187351731257, "learning_rate": 1.5745698668556856e-05, "loss": 0.7154, "step": 6262 }, { "epoch": 0.98, "grad_norm": 15.402076516673862, "learning_rate": 1.5744318344279103e-05, "loss": 0.6895, "step": 6263 }, { "epoch": 0.98, "grad_norm": 25.176644906322444, "learning_rate": 1.5742937856636294e-05, "loss": 0.6538, "step": 6264 }, { "epoch": 0.98, "grad_norm": 14.097045165379267, "learning_rate": 1.5741557205667688e-05, "loss": 0.697, "step": 6265 }, { "epoch": 0.98, "grad_norm": 21.51688942833326, "learning_rate": 1.574017639141256e-05, "loss": 0.7147, "step": 6266 }, { "epoch": 0.98, "grad_norm": 15.356930145712544, "learning_rate": 1.5738795413910174e-05, "loss": 0.674, "step": 6267 }, { "epoch": 0.98, "grad_norm": 31.798673484233632, "learning_rate": 1.57374142731998e-05, "loss": 0.6925, "step": 6268 }, { "epoch": 0.98, "grad_norm": 13.950738697464558, "learning_rate": 1.573603296932072e-05, "loss": 0.7904, "step": 6269 }, { "epoch": 0.98, "grad_norm": 20.508249749744827, "learning_rate": 1.5734651502312218e-05, "loss": 0.7407, "step": 6270 }, { "epoch": 0.98, "grad_norm": 16.141409384066787, "learning_rate": 1.5733269872213583e-05, "loss": 0.7864, "step": 6271 }, { "epoch": 0.98, "grad_norm": 18.444272982668206, "learning_rate": 1.5731888079064107e-05, "loss": 0.6863, "step": 6272 }, { "epoch": 0.98, "grad_norm": 14.12951265467563, "learning_rate": 1.5730506122903086e-05, "loss": 0.7873, "step": 6273 }, { "epoch": 0.98, "grad_norm": 14.93230822650624, "learning_rate": 1.5729124003769826e-05, "loss": 0.7465, "step": 6274 }, { "epoch": 0.98, "grad_norm": 17.134391908430306, "learning_rate": 1.572774172170363e-05, "loss": 0.7148, "step": 6275 }, { "epoch": 0.98, "grad_norm": 15.246212959067847, "learning_rate": 1.5726359276743808e-05, "loss": 0.6541, "step": 6276 }, { "epoch": 0.98, "grad_norm": 13.883252963354481, "learning_rate": 1.5724976668929678e-05, "loss": 0.6974, "step": 6277 }, { "epoch": 0.98, "grad_norm": 35.41424468561854, "learning_rate": 1.5723593898300562e-05, "loss": 0.7017, "step": 6278 }, { "epoch": 0.98, "grad_norm": 23.743387526486405, "learning_rate": 1.572221096489578e-05, "loss": 0.65, "step": 6279 }, { "epoch": 0.98, "grad_norm": 11.916708543579952, "learning_rate": 1.572082786875467e-05, "loss": 0.6345, "step": 6280 }, { "epoch": 0.98, "grad_norm": 26.015664969519534, "learning_rate": 1.5719444609916564e-05, "loss": 0.7147, "step": 6281 }, { "epoch": 0.98, "grad_norm": 30.076012645710914, "learning_rate": 1.5718061188420793e-05, "loss": 0.7767, "step": 6282 }, { "epoch": 0.98, "grad_norm": 14.62416221234385, "learning_rate": 1.571667760430671e-05, "loss": 0.7267, "step": 6283 }, { "epoch": 0.98, "grad_norm": 14.473050718298492, "learning_rate": 1.5715293857613662e-05, "loss": 0.7076, "step": 6284 }, { "epoch": 0.98, "grad_norm": 21.892231357457838, "learning_rate": 1.5713909948380995e-05, "loss": 0.7583, "step": 6285 }, { "epoch": 0.98, "grad_norm": 18.017917791540224, "learning_rate": 1.5712525876648076e-05, "loss": 0.7583, "step": 6286 }, { "epoch": 0.98, "grad_norm": 20.676341254821256, "learning_rate": 1.5711141642454258e-05, "loss": 0.7764, "step": 6287 }, { "epoch": 0.98, "grad_norm": 20.796402382012346, "learning_rate": 1.5709757245838918e-05, "loss": 0.7369, "step": 6288 }, { "epoch": 0.98, "grad_norm": 17.376078308044274, "learning_rate": 1.570837268684142e-05, "loss": 0.7264, "step": 6289 }, { "epoch": 0.98, "grad_norm": 19.632087278955975, "learning_rate": 1.5706987965501142e-05, "loss": 0.7711, "step": 6290 }, { "epoch": 0.98, "grad_norm": 15.923638613941863, "learning_rate": 1.5705603081857463e-05, "loss": 0.6989, "step": 6291 }, { "epoch": 0.98, "grad_norm": 18.242577810420823, "learning_rate": 1.5704218035949773e-05, "loss": 0.7212, "step": 6292 }, { "epoch": 0.98, "grad_norm": 15.279137905210119, "learning_rate": 1.570283282781746e-05, "loss": 0.7621, "step": 6293 }, { "epoch": 0.98, "grad_norm": 18.450926928379797, "learning_rate": 1.5701447457499914e-05, "loss": 0.791, "step": 6294 }, { "epoch": 0.98, "grad_norm": 17.837420604916066, "learning_rate": 1.570006192503654e-05, "loss": 0.77, "step": 6295 }, { "epoch": 0.98, "grad_norm": 18.909350604393193, "learning_rate": 1.5698676230466735e-05, "loss": 0.7334, "step": 6296 }, { "epoch": 0.98, "grad_norm": 20.886295373527055, "learning_rate": 1.5697290373829913e-05, "loss": 0.6799, "step": 6297 }, { "epoch": 0.98, "grad_norm": 22.233097679083436, "learning_rate": 1.5695904355165486e-05, "loss": 0.7532, "step": 6298 }, { "epoch": 0.98, "grad_norm": 16.108783529765777, "learning_rate": 1.5694518174512873e-05, "loss": 0.6611, "step": 6299 }, { "epoch": 0.98, "grad_norm": 17.657668291364995, "learning_rate": 1.5693131831911494e-05, "loss": 0.7844, "step": 6300 }, { "epoch": 0.98, "grad_norm": 25.34681562116457, "learning_rate": 1.5691745327400776e-05, "loss": 0.7776, "step": 6301 }, { "epoch": 0.98, "grad_norm": 30.17676812592073, "learning_rate": 1.569035866102015e-05, "loss": 0.7234, "step": 6302 }, { "epoch": 0.98, "grad_norm": 17.39102641685156, "learning_rate": 1.568897183280905e-05, "loss": 0.7063, "step": 6303 }, { "epoch": 0.98, "grad_norm": 16.93110483175266, "learning_rate": 1.5687584842806925e-05, "loss": 0.6177, "step": 6304 }, { "epoch": 0.98, "grad_norm": 15.526691742169396, "learning_rate": 1.568619769105321e-05, "loss": 0.6045, "step": 6305 }, { "epoch": 0.99, "grad_norm": 22.57082940305604, "learning_rate": 1.568481037758736e-05, "loss": 0.7763, "step": 6306 }, { "epoch": 0.99, "grad_norm": 19.24234523855676, "learning_rate": 1.5683422902448827e-05, "loss": 0.7059, "step": 6307 }, { "epoch": 0.99, "grad_norm": 20.985835218432385, "learning_rate": 1.5682035265677074e-05, "loss": 0.6476, "step": 6308 }, { "epoch": 0.99, "grad_norm": 24.889131048595345, "learning_rate": 1.568064746731156e-05, "loss": 0.7627, "step": 6309 }, { "epoch": 0.99, "grad_norm": 27.590260461076955, "learning_rate": 1.5679259507391755e-05, "loss": 0.7267, "step": 6310 }, { "epoch": 0.99, "grad_norm": 16.37743047520126, "learning_rate": 1.5677871385957134e-05, "loss": 0.731, "step": 6311 }, { "epoch": 0.99, "grad_norm": 18.17826006522586, "learning_rate": 1.567648310304717e-05, "loss": 0.7181, "step": 6312 }, { "epoch": 0.99, "grad_norm": 15.444323788353781, "learning_rate": 1.567509465870135e-05, "loss": 0.7175, "step": 6313 }, { "epoch": 0.99, "grad_norm": 18.491751149568916, "learning_rate": 1.567370605295915e-05, "loss": 0.7675, "step": 6314 }, { "epoch": 0.99, "grad_norm": 15.453460728283682, "learning_rate": 1.5672317285860076e-05, "loss": 0.6342, "step": 6315 }, { "epoch": 0.99, "grad_norm": 22.80354555471989, "learning_rate": 1.5670928357443617e-05, "loss": 0.7829, "step": 6316 }, { "epoch": 0.99, "grad_norm": 21.4134458670892, "learning_rate": 1.566953926774927e-05, "loss": 0.7413, "step": 6317 }, { "epoch": 0.99, "grad_norm": 14.142592560890225, "learning_rate": 1.5668150016816545e-05, "loss": 0.7177, "step": 6318 }, { "epoch": 0.99, "grad_norm": 16.270432504738718, "learning_rate": 1.5666760604684947e-05, "loss": 0.7444, "step": 6319 }, { "epoch": 0.99, "grad_norm": 18.252765065559505, "learning_rate": 1.5665371031393994e-05, "loss": 0.8004, "step": 6320 }, { "epoch": 0.99, "grad_norm": 22.988404591763434, "learning_rate": 1.56639812969832e-05, "loss": 0.7443, "step": 6321 }, { "epoch": 0.99, "grad_norm": 16.969777669751174, "learning_rate": 1.5662591401492096e-05, "loss": 0.7221, "step": 6322 }, { "epoch": 0.99, "grad_norm": 17.280438987023008, "learning_rate": 1.5661201344960203e-05, "loss": 0.7201, "step": 6323 }, { "epoch": 0.99, "grad_norm": 37.57959628327843, "learning_rate": 1.5659811127427053e-05, "loss": 0.767, "step": 6324 }, { "epoch": 0.99, "grad_norm": 19.437502706932296, "learning_rate": 1.5658420748932187e-05, "loss": 0.6743, "step": 6325 }, { "epoch": 0.99, "grad_norm": 22.148492767628333, "learning_rate": 1.5657030209515146e-05, "loss": 0.7354, "step": 6326 }, { "epoch": 0.99, "grad_norm": 20.79360772813145, "learning_rate": 1.5655639509215476e-05, "loss": 0.6183, "step": 6327 }, { "epoch": 0.99, "grad_norm": 15.334730897115843, "learning_rate": 1.565424864807273e-05, "loss": 0.6637, "step": 6328 }, { "epoch": 0.99, "grad_norm": 22.47819394584062, "learning_rate": 1.565285762612645e-05, "loss": 0.6701, "step": 6329 }, { "epoch": 0.99, "grad_norm": 16.261798390201033, "learning_rate": 1.5651466443416213e-05, "loss": 0.6882, "step": 6330 }, { "epoch": 0.99, "grad_norm": 15.828376330727263, "learning_rate": 1.5650075099981573e-05, "loss": 0.7075, "step": 6331 }, { "epoch": 0.99, "grad_norm": 20.908258572557916, "learning_rate": 1.564868359586211e-05, "loss": 0.7339, "step": 6332 }, { "epoch": 0.99, "grad_norm": 24.5335566250218, "learning_rate": 1.564729193109738e-05, "loss": 0.6724, "step": 6333 }, { "epoch": 0.99, "grad_norm": 19.72041759943538, "learning_rate": 1.5645900105726976e-05, "loss": 0.6783, "step": 6334 }, { "epoch": 0.99, "grad_norm": 25.438276170094106, "learning_rate": 1.5644508119790477e-05, "loss": 0.7547, "step": 6335 }, { "epoch": 0.99, "grad_norm": 21.496643036512626, "learning_rate": 1.5643115973327464e-05, "loss": 0.7009, "step": 6336 }, { "epoch": 0.99, "grad_norm": 26.56720763395691, "learning_rate": 1.5641723666377536e-05, "loss": 0.7125, "step": 6337 }, { "epoch": 0.99, "grad_norm": 19.754529238248484, "learning_rate": 1.564033119898029e-05, "loss": 0.7256, "step": 6338 }, { "epoch": 0.99, "grad_norm": 16.333374418697414, "learning_rate": 1.5638938571175324e-05, "loss": 0.6783, "step": 6339 }, { "epoch": 0.99, "grad_norm": 16.92363618146433, "learning_rate": 1.563754578300224e-05, "loss": 0.6673, "step": 6340 }, { "epoch": 0.99, "grad_norm": 21.209951178308156, "learning_rate": 1.5636152834500654e-05, "loss": 0.6956, "step": 6341 }, { "epoch": 0.99, "grad_norm": 19.734447474584446, "learning_rate": 1.5634759725710178e-05, "loss": 0.6897, "step": 6342 }, { "epoch": 0.99, "grad_norm": 18.960526927169756, "learning_rate": 1.5633366456670433e-05, "loss": 0.7769, "step": 6343 }, { "epoch": 0.99, "grad_norm": 15.17234637542746, "learning_rate": 1.563197302742104e-05, "loss": 0.6268, "step": 6344 }, { "epoch": 0.99, "grad_norm": 13.327554166181422, "learning_rate": 1.5630579438001626e-05, "loss": 0.671, "step": 6345 }, { "epoch": 0.99, "grad_norm": 21.89072928584197, "learning_rate": 1.562918568845183e-05, "loss": 0.6727, "step": 6346 }, { "epoch": 0.99, "grad_norm": 24.526235828507357, "learning_rate": 1.562779177881129e-05, "loss": 0.6639, "step": 6347 }, { "epoch": 0.99, "grad_norm": 13.72861549057825, "learning_rate": 1.5626397709119638e-05, "loss": 0.6892, "step": 6348 }, { "epoch": 0.99, "grad_norm": 22.15658985786569, "learning_rate": 1.562500347941653e-05, "loss": 0.6799, "step": 6349 }, { "epoch": 0.99, "grad_norm": 21.507176640634484, "learning_rate": 1.5623609089741608e-05, "loss": 0.7295, "step": 6350 }, { "epoch": 0.99, "grad_norm": 23.77365799424598, "learning_rate": 1.5622214540134536e-05, "loss": 0.7676, "step": 6351 }, { "epoch": 0.99, "grad_norm": 13.633065273269677, "learning_rate": 1.5620819830634975e-05, "loss": 0.72, "step": 6352 }, { "epoch": 0.99, "grad_norm": 20.624250783461548, "learning_rate": 1.5619424961282585e-05, "loss": 0.6405, "step": 6353 }, { "epoch": 0.99, "grad_norm": 15.837879893131927, "learning_rate": 1.5618029932117035e-05, "loss": 0.672, "step": 6354 }, { "epoch": 0.99, "grad_norm": 16.2073531832548, "learning_rate": 1.5616634743177996e-05, "loss": 0.6685, "step": 6355 }, { "epoch": 0.99, "grad_norm": 23.607158163602943, "learning_rate": 1.5615239394505154e-05, "loss": 0.7397, "step": 6356 }, { "epoch": 0.99, "grad_norm": 21.430247388166993, "learning_rate": 1.5613843886138192e-05, "loss": 0.6731, "step": 6357 }, { "epoch": 0.99, "grad_norm": 15.615028495457203, "learning_rate": 1.561244821811679e-05, "loss": 0.7111, "step": 6358 }, { "epoch": 0.99, "grad_norm": 20.608515768526523, "learning_rate": 1.5611052390480646e-05, "loss": 0.7206, "step": 6359 }, { "epoch": 0.99, "grad_norm": 12.936369934115836, "learning_rate": 1.5609656403269452e-05, "loss": 0.6903, "step": 6360 }, { "epoch": 0.99, "grad_norm": 15.944315726390709, "learning_rate": 1.560826025652291e-05, "loss": 0.7447, "step": 6361 }, { "epoch": 0.99, "grad_norm": 20.259277886198085, "learning_rate": 1.560686395028073e-05, "loss": 0.6132, "step": 6362 }, { "epoch": 0.99, "grad_norm": 13.02463172180988, "learning_rate": 1.560546748458262e-05, "loss": 0.6949, "step": 6363 }, { "epoch": 0.99, "grad_norm": 30.59667730938663, "learning_rate": 1.5604070859468292e-05, "loss": 0.6782, "step": 6364 }, { "epoch": 0.99, "grad_norm": 11.042902305654117, "learning_rate": 1.5602674074977467e-05, "loss": 0.7063, "step": 6365 }, { "epoch": 0.99, "grad_norm": 24.800659256294253, "learning_rate": 1.560127713114987e-05, "loss": 0.7687, "step": 6366 }, { "epoch": 0.99, "grad_norm": 15.89127829661326, "learning_rate": 1.559988002802523e-05, "loss": 0.6818, "step": 6367 }, { "epoch": 0.99, "grad_norm": 14.499096024221645, "learning_rate": 1.5598482765643273e-05, "loss": 0.6369, "step": 6368 }, { "epoch": 0.99, "grad_norm": 17.145687953391345, "learning_rate": 1.5597085344043742e-05, "loss": 0.6976, "step": 6369 }, { "epoch": 1.0, "grad_norm": 27.72505787980115, "learning_rate": 1.5595687763266378e-05, "loss": 0.7195, "step": 6370 }, { "epoch": 1.0, "grad_norm": 17.090530167227787, "learning_rate": 1.5594290023350933e-05, "loss": 0.6804, "step": 6371 }, { "epoch": 1.0, "grad_norm": 23.81879930735099, "learning_rate": 1.5592892124337145e-05, "loss": 0.7284, "step": 6372 }, { "epoch": 1.0, "grad_norm": 19.79588388477555, "learning_rate": 1.559149406626478e-05, "loss": 0.7053, "step": 6373 }, { "epoch": 1.0, "grad_norm": 13.071680652907293, "learning_rate": 1.5590095849173597e-05, "loss": 0.6733, "step": 6374 }, { "epoch": 1.0, "grad_norm": 12.068402627502268, "learning_rate": 1.5588697473103355e-05, "loss": 0.6621, "step": 6375 }, { "epoch": 1.0, "grad_norm": 23.994208110113572, "learning_rate": 1.5587298938093828e-05, "loss": 0.6938, "step": 6376 }, { "epoch": 1.0, "grad_norm": 27.89978263360765, "learning_rate": 1.5585900244184785e-05, "loss": 0.7703, "step": 6377 }, { "epoch": 1.0, "grad_norm": 22.50729676979544, "learning_rate": 1.5584501391416008e-05, "loss": 0.6832, "step": 6378 }, { "epoch": 1.0, "grad_norm": 15.379438339225747, "learning_rate": 1.558310237982728e-05, "loss": 0.6573, "step": 6379 }, { "epoch": 1.0, "grad_norm": 20.45984504218609, "learning_rate": 1.558170320945838e-05, "loss": 0.6878, "step": 6380 }, { "epoch": 1.0, "grad_norm": 20.599490474879698, "learning_rate": 1.5580303880349113e-05, "loss": 0.6284, "step": 6381 }, { "epoch": 1.0, "grad_norm": 20.223205712944264, "learning_rate": 1.5578904392539268e-05, "loss": 0.6712, "step": 6382 }, { "epoch": 1.0, "grad_norm": 35.117554979449, "learning_rate": 1.557750474606864e-05, "loss": 0.7686, "step": 6383 }, { "epoch": 1.0, "grad_norm": 18.909667976630182, "learning_rate": 1.5576104940977045e-05, "loss": 0.6952, "step": 6384 }, { "epoch": 1.0, "grad_norm": 21.715535244308136, "learning_rate": 1.5574704977304286e-05, "loss": 0.7224, "step": 6385 }, { "epoch": 1.0, "grad_norm": 17.4192327668256, "learning_rate": 1.557330485509018e-05, "loss": 0.6875, "step": 6386 }, { "epoch": 1.0, "grad_norm": 19.048427537679707, "learning_rate": 1.5571904574374543e-05, "loss": 0.7214, "step": 6387 }, { "epoch": 1.0, "grad_norm": 17.260240005029573, "learning_rate": 1.55705041351972e-05, "loss": 0.7188, "step": 6388 }, { "epoch": 1.0, "grad_norm": 20.199730289557138, "learning_rate": 1.5569103537597972e-05, "loss": 0.7141, "step": 6389 }, { "epoch": 1.0, "grad_norm": 25.794480954963547, "learning_rate": 1.55677027816167e-05, "loss": 0.736, "step": 6390 }, { "epoch": 1.0, "grad_norm": 18.265690492589275, "learning_rate": 1.5566301867293223e-05, "loss": 0.6804, "step": 6391 }, { "epoch": 1.0, "grad_norm": 18.674947584000783, "learning_rate": 1.5564900794667372e-05, "loss": 0.6766, "step": 6392 }, { "epoch": 1.0, "grad_norm": 14.604538746373462, "learning_rate": 1.5563499563779002e-05, "loss": 0.7749, "step": 6393 }, { "epoch": 1.0, "grad_norm": 12.775019820371973, "learning_rate": 1.5562098174667957e-05, "loss": 0.6711, "step": 6394 }, { "epoch": 1.0, "grad_norm": 27.382511781573122, "learning_rate": 1.556069662737409e-05, "loss": 0.8826, "step": 6395 }, { "epoch": 1.0, "grad_norm": 17.75612464784674, "learning_rate": 1.5559294921937272e-05, "loss": 0.7504, "step": 6396 }, { "epoch": 1.0, "grad_norm": 18.36544359471965, "learning_rate": 1.555789305839735e-05, "loss": 0.5571, "step": 6397 }, { "epoch": 1.0, "grad_norm": 15.937426314464096, "learning_rate": 1.5556491036794204e-05, "loss": 0.6322, "step": 6398 }, { "epoch": 1.0, "grad_norm": 12.087657573041733, "learning_rate": 1.5555088857167703e-05, "loss": 0.6066, "step": 6399 }, { "epoch": 1.0, "grad_norm": 19.06324289781649, "learning_rate": 1.5553686519557726e-05, "loss": 0.7523, "step": 6400 }, { "epoch": 1.0, "grad_norm": 12.220169261789898, "learning_rate": 1.5552284024004154e-05, "loss": 0.5992, "step": 6401 }, { "epoch": 1.0, "grad_norm": 8.19145050819094, "learning_rate": 1.555088137054687e-05, "loss": 0.6211, "step": 6402 }, { "epoch": 1.0, "grad_norm": 16.2437545282828, "learning_rate": 1.554947855922577e-05, "loss": 0.7152, "step": 6403 }, { "epoch": 1.0, "grad_norm": 27.16781968871853, "learning_rate": 1.5548075590080745e-05, "loss": 0.7184, "step": 6404 }, { "epoch": 1.0, "grad_norm": 28.54908774558578, "learning_rate": 1.5546672463151695e-05, "loss": 0.6511, "step": 6405 }, { "epoch": 1.0, "grad_norm": 62.210625402081355, "learning_rate": 1.5545269178478523e-05, "loss": 0.6291, "step": 6406 }, { "epoch": 1.0, "grad_norm": 16.581450136064625, "learning_rate": 1.554386573610114e-05, "loss": 0.6526, "step": 6407 }, { "epoch": 1.0, "grad_norm": 15.038554100261715, "learning_rate": 1.554246213605946e-05, "loss": 0.6116, "step": 6408 }, { "epoch": 1.0, "grad_norm": 19.306176302260706, "learning_rate": 1.55410583783934e-05, "loss": 0.7377, "step": 6409 }, { "epoch": 1.0, "grad_norm": 29.4897848532468, "learning_rate": 1.5539654463142878e-05, "loss": 0.6722, "step": 6410 }, { "epoch": 1.0, "grad_norm": 24.97828358986796, "learning_rate": 1.5538250390347825e-05, "loss": 0.7174, "step": 6411 }, { "epoch": 1.0, "grad_norm": 20.532038375590826, "learning_rate": 1.5536846160048172e-05, "loss": 0.6915, "step": 6412 }, { "epoch": 1.0, "grad_norm": 30.187267639221538, "learning_rate": 1.553544177228385e-05, "loss": 0.6716, "step": 6413 }, { "epoch": 1.0, "grad_norm": 17.411902530744527, "learning_rate": 1.5534037227094807e-05, "loss": 0.6789, "step": 6414 }, { "epoch": 1.0, "grad_norm": 21.86502560255622, "learning_rate": 1.553263252452098e-05, "loss": 0.7738, "step": 6415 }, { "epoch": 1.0, "grad_norm": 17.6030213927747, "learning_rate": 1.553122766460232e-05, "loss": 0.6946, "step": 6416 }, { "epoch": 1.0, "grad_norm": 22.28326880563784, "learning_rate": 1.552982264737878e-05, "loss": 0.7725, "step": 6417 }, { "epoch": 1.0, "grad_norm": 20.524584318138132, "learning_rate": 1.5528417472890324e-05, "loss": 0.6993, "step": 6418 }, { "epoch": 1.0, "grad_norm": 19.205418527364696, "learning_rate": 1.5527012141176904e-05, "loss": 0.6616, "step": 6419 }, { "epoch": 1.0, "grad_norm": 12.703448832232999, "learning_rate": 1.5525606652278493e-05, "loss": 0.6215, "step": 6420 }, { "epoch": 1.0, "grad_norm": 20.331133644745545, "learning_rate": 1.5524201006235063e-05, "loss": 0.7806, "step": 6421 }, { "epoch": 1.0, "grad_norm": 11.713294211999813, "learning_rate": 1.5522795203086584e-05, "loss": 0.7002, "step": 6422 }, { "epoch": 1.0, "grad_norm": 25.876963129760735, "learning_rate": 1.5521389242873044e-05, "loss": 0.6968, "step": 6423 }, { "epoch": 1.0, "grad_norm": 17.581924863610826, "learning_rate": 1.551998312563442e-05, "loss": 0.6997, "step": 6424 }, { "epoch": 1.0, "grad_norm": 15.814924504690554, "learning_rate": 1.551857685141071e-05, "loss": 0.6691, "step": 6425 }, { "epoch": 1.0, "grad_norm": 19.574550622803667, "learning_rate": 1.5517170420241897e-05, "loss": 0.7012, "step": 6426 }, { "epoch": 1.0, "grad_norm": 17.427129710839353, "learning_rate": 1.551576383216799e-05, "loss": 0.6208, "step": 6427 }, { "epoch": 1.0, "grad_norm": 16.742300797727502, "learning_rate": 1.5514357087228985e-05, "loss": 0.7013, "step": 6428 }, { "epoch": 1.0, "grad_norm": 12.469801133051082, "learning_rate": 1.551295018546489e-05, "loss": 0.556, "step": 6429 }, { "epoch": 1.0, "grad_norm": 20.812763396449327, "learning_rate": 1.5511543126915713e-05, "loss": 0.6248, "step": 6430 }, { "epoch": 1.0, "grad_norm": 25.462182426274687, "learning_rate": 1.5510135911621474e-05, "loss": 0.8001, "step": 6431 }, { "epoch": 1.0, "grad_norm": 17.442260031825697, "learning_rate": 1.5508728539622196e-05, "loss": 0.6867, "step": 6432 }, { "epoch": 1.0, "grad_norm": 12.667367790946475, "learning_rate": 1.5507321010957903e-05, "loss": 0.6411, "step": 6433 }, { "epoch": 1.0, "grad_norm": 26.82405009618746, "learning_rate": 1.550591332566862e-05, "loss": 0.8276, "step": 6434 }, { "epoch": 1.01, "grad_norm": 28.317963098612843, "learning_rate": 1.550450548379438e-05, "loss": 0.678, "step": 6435 }, { "epoch": 1.01, "grad_norm": 20.657453142697225, "learning_rate": 1.5503097485375228e-05, "loss": 0.635, "step": 6436 }, { "epoch": 1.01, "grad_norm": 17.13644110396231, "learning_rate": 1.5501689330451203e-05, "loss": 0.6308, "step": 6437 }, { "epoch": 1.01, "grad_norm": 17.74563325137112, "learning_rate": 1.5500281019062347e-05, "loss": 0.6649, "step": 6438 }, { "epoch": 1.01, "grad_norm": 21.28041175908318, "learning_rate": 1.5498872551248722e-05, "loss": 0.7186, "step": 6439 }, { "epoch": 1.01, "grad_norm": 11.692231973433241, "learning_rate": 1.5497463927050375e-05, "loss": 0.6133, "step": 6440 }, { "epoch": 1.01, "grad_norm": 24.996811089384508, "learning_rate": 1.5496055146507368e-05, "loss": 0.6893, "step": 6441 }, { "epoch": 1.01, "grad_norm": 18.039620496579285, "learning_rate": 1.5494646209659775e-05, "loss": 0.6556, "step": 6442 }, { "epoch": 1.01, "grad_norm": 18.81479343267713, "learning_rate": 1.549323711654765e-05, "loss": 0.6677, "step": 6443 }, { "epoch": 1.01, "grad_norm": 16.438537720370615, "learning_rate": 1.549182786721108e-05, "loss": 0.5729, "step": 6444 }, { "epoch": 1.01, "grad_norm": 13.437771074477322, "learning_rate": 1.5490418461690137e-05, "loss": 0.6372, "step": 6445 }, { "epoch": 1.01, "grad_norm": 23.703700780947685, "learning_rate": 1.5489008900024903e-05, "loss": 0.6979, "step": 6446 }, { "epoch": 1.01, "grad_norm": 12.84046898071911, "learning_rate": 1.5487599182255467e-05, "loss": 0.7347, "step": 6447 }, { "epoch": 1.01, "grad_norm": 25.446927423152637, "learning_rate": 1.5486189308421922e-05, "loss": 0.7067, "step": 6448 }, { "epoch": 1.01, "grad_norm": 15.861818366852788, "learning_rate": 1.5484779278564363e-05, "loss": 0.739, "step": 6449 }, { "epoch": 1.01, "grad_norm": 19.445507919874448, "learning_rate": 1.5483369092722888e-05, "loss": 0.6521, "step": 6450 }, { "epoch": 1.01, "grad_norm": 19.52276352149899, "learning_rate": 1.5481958750937605e-05, "loss": 0.7105, "step": 6451 }, { "epoch": 1.01, "grad_norm": 19.278302783551332, "learning_rate": 1.548054825324862e-05, "loss": 0.7706, "step": 6452 }, { "epoch": 1.01, "grad_norm": 12.19005566113112, "learning_rate": 1.547913759969605e-05, "loss": 0.6834, "step": 6453 }, { "epoch": 1.01, "grad_norm": 17.101977136359437, "learning_rate": 1.5477726790320012e-05, "loss": 0.6386, "step": 6454 }, { "epoch": 1.01, "grad_norm": 20.9440510386669, "learning_rate": 1.547631582516063e-05, "loss": 0.6406, "step": 6455 }, { "epoch": 1.01, "grad_norm": 14.728639640331524, "learning_rate": 1.5474904704258027e-05, "loss": 0.5915, "step": 6456 }, { "epoch": 1.01, "grad_norm": 24.502157493094778, "learning_rate": 1.5473493427652337e-05, "loss": 0.708, "step": 6457 }, { "epoch": 1.01, "grad_norm": 16.493871406810868, "learning_rate": 1.54720819953837e-05, "loss": 0.69, "step": 6458 }, { "epoch": 1.01, "grad_norm": 17.440185088053983, "learning_rate": 1.547067040749225e-05, "loss": 0.702, "step": 6459 }, { "epoch": 1.01, "grad_norm": 21.26024440494361, "learning_rate": 1.546925866401813e-05, "loss": 0.7452, "step": 6460 }, { "epoch": 1.01, "grad_norm": 30.164224661248184, "learning_rate": 1.5467846765001496e-05, "loss": 0.7826, "step": 6461 }, { "epoch": 1.01, "grad_norm": 20.484245985135512, "learning_rate": 1.54664347104825e-05, "loss": 0.6766, "step": 6462 }, { "epoch": 1.01, "grad_norm": 21.16069597947086, "learning_rate": 1.5465022500501294e-05, "loss": 0.6465, "step": 6463 }, { "epoch": 1.01, "grad_norm": 20.559388702718227, "learning_rate": 1.5463610135098048e-05, "loss": 0.7666, "step": 6464 }, { "epoch": 1.01, "grad_norm": 16.008984004839487, "learning_rate": 1.5462197614312926e-05, "loss": 0.7048, "step": 6465 }, { "epoch": 1.01, "grad_norm": 17.238208929654018, "learning_rate": 1.54607849381861e-05, "loss": 0.6362, "step": 6466 }, { "epoch": 1.01, "grad_norm": 19.94097853826782, "learning_rate": 1.5459372106757742e-05, "loss": 0.7061, "step": 6467 }, { "epoch": 1.01, "grad_norm": 17.899308479426832, "learning_rate": 1.5457959120068036e-05, "loss": 0.6635, "step": 6468 }, { "epoch": 1.01, "grad_norm": 14.943075280311621, "learning_rate": 1.5456545978157168e-05, "loss": 0.6428, "step": 6469 }, { "epoch": 1.01, "grad_norm": 19.13668959818432, "learning_rate": 1.545513268106532e-05, "loss": 0.6835, "step": 6470 }, { "epoch": 1.01, "grad_norm": 57.91829494776299, "learning_rate": 1.5453719228832692e-05, "loss": 0.6759, "step": 6471 }, { "epoch": 1.01, "grad_norm": 17.221148865546716, "learning_rate": 1.5452305621499483e-05, "loss": 0.7349, "step": 6472 }, { "epoch": 1.01, "grad_norm": 17.936886695518524, "learning_rate": 1.5450891859105884e-05, "loss": 0.7431, "step": 6473 }, { "epoch": 1.01, "grad_norm": 15.768308261848315, "learning_rate": 1.5449477941692114e-05, "loss": 0.6369, "step": 6474 }, { "epoch": 1.01, "grad_norm": 23.73339342343991, "learning_rate": 1.544806386929838e-05, "loss": 0.7103, "step": 6475 }, { "epoch": 1.01, "grad_norm": 19.01587585681051, "learning_rate": 1.5446649641964895e-05, "loss": 0.634, "step": 6476 }, { "epoch": 1.01, "grad_norm": 13.830770775546783, "learning_rate": 1.544523525973188e-05, "loss": 0.7456, "step": 6477 }, { "epoch": 1.01, "grad_norm": 15.09747496837828, "learning_rate": 1.544382072263956e-05, "loss": 0.6198, "step": 6478 }, { "epoch": 1.01, "grad_norm": 16.985267152490426, "learning_rate": 1.5442406030728166e-05, "loss": 0.6579, "step": 6479 }, { "epoch": 1.01, "grad_norm": 13.73792426638232, "learning_rate": 1.5440991184037924e-05, "loss": 0.6999, "step": 6480 }, { "epoch": 1.01, "grad_norm": 21.700077622907656, "learning_rate": 1.5439576182609077e-05, "loss": 0.7502, "step": 6481 }, { "epoch": 1.01, "grad_norm": 22.47298910821798, "learning_rate": 1.5438161026481866e-05, "loss": 0.6358, "step": 6482 }, { "epoch": 1.01, "grad_norm": 26.84028716228371, "learning_rate": 1.5436745715696535e-05, "loss": 0.744, "step": 6483 }, { "epoch": 1.01, "grad_norm": 31.05960894616586, "learning_rate": 1.543533025029334e-05, "loss": 0.7034, "step": 6484 }, { "epoch": 1.01, "grad_norm": 23.039469146826423, "learning_rate": 1.5433914630312527e-05, "loss": 0.6496, "step": 6485 }, { "epoch": 1.01, "grad_norm": 14.360632751056993, "learning_rate": 1.5432498855794364e-05, "loss": 0.6024, "step": 6486 }, { "epoch": 1.01, "grad_norm": 17.719325236245606, "learning_rate": 1.543108292677911e-05, "loss": 0.6599, "step": 6487 }, { "epoch": 1.01, "grad_norm": 17.95426578067958, "learning_rate": 1.5429666843307035e-05, "loss": 0.6399, "step": 6488 }, { "epoch": 1.01, "grad_norm": 13.729858326264344, "learning_rate": 1.5428250605418414e-05, "loss": 0.6643, "step": 6489 }, { "epoch": 1.01, "grad_norm": 21.274789936001337, "learning_rate": 1.542683421315352e-05, "loss": 0.6547, "step": 6490 }, { "epoch": 1.01, "grad_norm": 18.05461939242172, "learning_rate": 1.5425417666552635e-05, "loss": 0.6833, "step": 6491 }, { "epoch": 1.01, "grad_norm": 19.758131550316673, "learning_rate": 1.5424000965656042e-05, "loss": 0.707, "step": 6492 }, { "epoch": 1.01, "grad_norm": 12.878210758802147, "learning_rate": 1.542258411050404e-05, "loss": 0.6852, "step": 6493 }, { "epoch": 1.01, "grad_norm": 17.76383856789012, "learning_rate": 1.5421167101136917e-05, "loss": 0.6823, "step": 6494 }, { "epoch": 1.01, "grad_norm": 21.582764347105275, "learning_rate": 1.5419749937594967e-05, "loss": 0.7516, "step": 6495 }, { "epoch": 1.01, "grad_norm": 16.840335613120832, "learning_rate": 1.5418332619918507e-05, "loss": 0.6507, "step": 6496 }, { "epoch": 1.01, "grad_norm": 17.363139546935002, "learning_rate": 1.541691514814783e-05, "loss": 0.6675, "step": 6497 }, { "epoch": 1.01, "grad_norm": 23.466826583583153, "learning_rate": 1.541549752232326e-05, "loss": 0.7122, "step": 6498 }, { "epoch": 1.02, "grad_norm": 17.442843736145985, "learning_rate": 1.541407974248511e-05, "loss": 0.6983, "step": 6499 }, { "epoch": 1.02, "grad_norm": 16.599299483498584, "learning_rate": 1.5412661808673694e-05, "loss": 0.6219, "step": 6500 }, { "epoch": 1.02, "grad_norm": 12.347073533096834, "learning_rate": 1.5411243720929342e-05, "loss": 0.6531, "step": 6501 }, { "epoch": 1.02, "grad_norm": 21.872325090621768, "learning_rate": 1.5409825479292388e-05, "loss": 0.6819, "step": 6502 }, { "epoch": 1.02, "grad_norm": 12.181120917833676, "learning_rate": 1.5408407083803162e-05, "loss": 0.6597, "step": 6503 }, { "epoch": 1.02, "grad_norm": 16.683661268486603, "learning_rate": 1.5406988534502002e-05, "loss": 0.6862, "step": 6504 }, { "epoch": 1.02, "grad_norm": 28.451001613869806, "learning_rate": 1.5405569831429247e-05, "loss": 0.6581, "step": 6505 }, { "epoch": 1.02, "grad_norm": 23.754498773432474, "learning_rate": 1.5404150974625254e-05, "loss": 0.7414, "step": 6506 }, { "epoch": 1.02, "grad_norm": 20.520233758083858, "learning_rate": 1.5402731964130365e-05, "loss": 0.6471, "step": 6507 }, { "epoch": 1.02, "grad_norm": 16.36371344976559, "learning_rate": 1.5401312799984943e-05, "loss": 0.5999, "step": 6508 }, { "epoch": 1.02, "grad_norm": 21.94754732621128, "learning_rate": 1.539989348222934e-05, "loss": 0.6978, "step": 6509 }, { "epoch": 1.02, "grad_norm": 16.382679134675595, "learning_rate": 1.5398474010903927e-05, "loss": 0.7457, "step": 6510 }, { "epoch": 1.02, "grad_norm": 16.69835778231014, "learning_rate": 1.5397054386049072e-05, "loss": 0.6914, "step": 6511 }, { "epoch": 1.02, "grad_norm": 18.07778185244271, "learning_rate": 1.5395634607705145e-05, "loss": 0.6611, "step": 6512 }, { "epoch": 1.02, "grad_norm": 20.33259452530617, "learning_rate": 1.539421467591253e-05, "loss": 0.7345, "step": 6513 }, { "epoch": 1.02, "grad_norm": 16.867619554428977, "learning_rate": 1.5392794590711605e-05, "loss": 0.6659, "step": 6514 }, { "epoch": 1.02, "grad_norm": 14.37208631571248, "learning_rate": 1.5391374352142752e-05, "loss": 0.6822, "step": 6515 }, { "epoch": 1.02, "grad_norm": 19.438450565047656, "learning_rate": 1.538995396024637e-05, "loss": 0.714, "step": 6516 }, { "epoch": 1.02, "grad_norm": 27.06842814846263, "learning_rate": 1.5388533415062848e-05, "loss": 0.5554, "step": 6517 }, { "epoch": 1.02, "grad_norm": 27.561295686305304, "learning_rate": 1.5387112716632594e-05, "loss": 0.6792, "step": 6518 }, { "epoch": 1.02, "grad_norm": 16.8032830643274, "learning_rate": 1.5385691864995998e-05, "loss": 0.6521, "step": 6519 }, { "epoch": 1.02, "grad_norm": 15.67270866080271, "learning_rate": 1.5384270860193477e-05, "loss": 0.6598, "step": 6520 }, { "epoch": 1.02, "grad_norm": 10.561964563900505, "learning_rate": 1.5382849702265447e-05, "loss": 0.6752, "step": 6521 }, { "epoch": 1.02, "grad_norm": 28.522186184979937, "learning_rate": 1.538142839125232e-05, "loss": 0.7348, "step": 6522 }, { "epoch": 1.02, "grad_norm": 24.088772860199423, "learning_rate": 1.538000692719451e-05, "loss": 0.7279, "step": 6523 }, { "epoch": 1.02, "grad_norm": 19.539333245538113, "learning_rate": 1.5378585310132458e-05, "loss": 0.7308, "step": 6524 }, { "epoch": 1.02, "grad_norm": 14.89188983239716, "learning_rate": 1.5377163540106582e-05, "loss": 0.6546, "step": 6525 }, { "epoch": 1.02, "grad_norm": 14.668147509891359, "learning_rate": 1.5375741617157324e-05, "loss": 0.6505, "step": 6526 }, { "epoch": 1.02, "grad_norm": 15.063535418279532, "learning_rate": 1.5374319541325114e-05, "loss": 0.7976, "step": 6527 }, { "epoch": 1.02, "grad_norm": 16.380752379373973, "learning_rate": 1.5372897312650406e-05, "loss": 0.7622, "step": 6528 }, { "epoch": 1.02, "grad_norm": 26.530187130588722, "learning_rate": 1.5371474931173638e-05, "loss": 0.7456, "step": 6529 }, { "epoch": 1.02, "grad_norm": 19.945712580293577, "learning_rate": 1.5370052396935268e-05, "loss": 0.7097, "step": 6530 }, { "epoch": 1.02, "grad_norm": 16.622305601523422, "learning_rate": 1.5368629709975745e-05, "loss": 0.7191, "step": 6531 }, { "epoch": 1.02, "grad_norm": 11.725128365690392, "learning_rate": 1.5367206870335536e-05, "loss": 0.6044, "step": 6532 }, { "epoch": 1.02, "grad_norm": 13.28518469812207, "learning_rate": 1.5365783878055103e-05, "loss": 0.6909, "step": 6533 }, { "epoch": 1.02, "grad_norm": 22.472772158677415, "learning_rate": 1.5364360733174916e-05, "loss": 0.6582, "step": 6534 }, { "epoch": 1.02, "grad_norm": 17.542231148209925, "learning_rate": 1.536293743573545e-05, "loss": 0.6751, "step": 6535 }, { "epoch": 1.02, "grad_norm": 16.47603248188668, "learning_rate": 1.5361513985777175e-05, "loss": 0.6323, "step": 6536 }, { "epoch": 1.02, "grad_norm": 16.942876716357123, "learning_rate": 1.536009038334058e-05, "loss": 0.6564, "step": 6537 }, { "epoch": 1.02, "grad_norm": 16.88624460991718, "learning_rate": 1.5358666628466154e-05, "loss": 0.6908, "step": 6538 }, { "epoch": 1.02, "grad_norm": 14.097432343778436, "learning_rate": 1.535724272119438e-05, "loss": 0.6468, "step": 6539 }, { "epoch": 1.02, "grad_norm": 16.533646284586915, "learning_rate": 1.535581866156576e-05, "loss": 0.6792, "step": 6540 }, { "epoch": 1.02, "grad_norm": 16.373033787711773, "learning_rate": 1.535439444962079e-05, "loss": 0.6745, "step": 6541 }, { "epoch": 1.02, "grad_norm": 16.30349281989993, "learning_rate": 1.535297008539997e-05, "loss": 0.6978, "step": 6542 }, { "epoch": 1.02, "grad_norm": 18.16978710001151, "learning_rate": 1.5351545568943818e-05, "loss": 0.7, "step": 6543 }, { "epoch": 1.02, "grad_norm": 20.654371946662845, "learning_rate": 1.5350120900292833e-05, "loss": 0.6959, "step": 6544 }, { "epoch": 1.02, "grad_norm": 15.19838647026938, "learning_rate": 1.5348696079487547e-05, "loss": 0.7448, "step": 6545 }, { "epoch": 1.02, "grad_norm": 19.045007251105677, "learning_rate": 1.534727110656847e-05, "loss": 0.6897, "step": 6546 }, { "epoch": 1.02, "grad_norm": 26.992515178849985, "learning_rate": 1.534584598157613e-05, "loss": 0.7007, "step": 6547 }, { "epoch": 1.02, "grad_norm": 19.780094752167326, "learning_rate": 1.5344420704551058e-05, "loss": 0.647, "step": 6548 }, { "epoch": 1.02, "grad_norm": 18.98574558852048, "learning_rate": 1.534299527553379e-05, "loss": 0.6448, "step": 6549 }, { "epoch": 1.02, "grad_norm": 22.06125811020938, "learning_rate": 1.534156969456486e-05, "loss": 0.683, "step": 6550 }, { "epoch": 1.02, "grad_norm": 18.182842330895035, "learning_rate": 1.5340143961684813e-05, "loss": 0.6714, "step": 6551 }, { "epoch": 1.02, "grad_norm": 21.8774928974933, "learning_rate": 1.5338718076934195e-05, "loss": 0.7379, "step": 6552 }, { "epoch": 1.02, "grad_norm": 13.287842793250592, "learning_rate": 1.5337292040353555e-05, "loss": 0.6541, "step": 6553 }, { "epoch": 1.02, "grad_norm": 16.490607593925503, "learning_rate": 1.5335865851983456e-05, "loss": 0.6495, "step": 6554 }, { "epoch": 1.02, "grad_norm": 16.359068395388427, "learning_rate": 1.5334439511864453e-05, "loss": 0.7372, "step": 6555 }, { "epoch": 1.02, "grad_norm": 15.46820770526866, "learning_rate": 1.533301302003711e-05, "loss": 0.6705, "step": 6556 }, { "epoch": 1.02, "grad_norm": 18.33734881215871, "learning_rate": 1.5331586376541997e-05, "loss": 0.7181, "step": 6557 }, { "epoch": 1.02, "grad_norm": 26.14072399439881, "learning_rate": 1.5330159581419687e-05, "loss": 0.8224, "step": 6558 }, { "epoch": 1.02, "grad_norm": 22.334981350149064, "learning_rate": 1.532873263471075e-05, "loss": 0.8401, "step": 6559 }, { "epoch": 1.02, "grad_norm": 18.26068576114692, "learning_rate": 1.5327305536455786e-05, "loss": 0.7529, "step": 6560 }, { "epoch": 1.02, "grad_norm": 21.45708392616138, "learning_rate": 1.5325878286695362e-05, "loss": 0.6644, "step": 6561 }, { "epoch": 1.02, "grad_norm": 19.169113866224432, "learning_rate": 1.5324450885470078e-05, "loss": 0.6645, "step": 6562 }, { "epoch": 1.03, "grad_norm": 26.22562730353031, "learning_rate": 1.5323023332820517e-05, "loss": 0.7402, "step": 6563 }, { "epoch": 1.03, "grad_norm": 14.552562716194432, "learning_rate": 1.5321595628787297e-05, "loss": 0.623, "step": 6564 }, { "epoch": 1.03, "grad_norm": 21.47646629420186, "learning_rate": 1.5320167773411004e-05, "loss": 0.6306, "step": 6565 }, { "epoch": 1.03, "grad_norm": 19.24444339180681, "learning_rate": 1.5318739766732255e-05, "loss": 0.6996, "step": 6566 }, { "epoch": 1.03, "grad_norm": 13.445902434090621, "learning_rate": 1.5317311608791656e-05, "loss": 0.5829, "step": 6567 }, { "epoch": 1.03, "grad_norm": 16.940875641665073, "learning_rate": 1.5315883299629825e-05, "loss": 0.6225, "step": 6568 }, { "epoch": 1.03, "grad_norm": 16.88560715845788, "learning_rate": 1.531445483928738e-05, "loss": 0.656, "step": 6569 }, { "epoch": 1.03, "grad_norm": 17.068383973428357, "learning_rate": 1.531302622780496e-05, "loss": 0.7427, "step": 6570 }, { "epoch": 1.03, "grad_norm": 19.72585627283115, "learning_rate": 1.5311597465223173e-05, "loss": 0.6016, "step": 6571 }, { "epoch": 1.03, "grad_norm": 18.24498861073896, "learning_rate": 1.531016855158266e-05, "loss": 0.6832, "step": 6572 }, { "epoch": 1.03, "grad_norm": 16.651745441165993, "learning_rate": 1.5308739486924064e-05, "loss": 0.7824, "step": 6573 }, { "epoch": 1.03, "grad_norm": 20.544739473709814, "learning_rate": 1.530731027128802e-05, "loss": 0.7702, "step": 6574 }, { "epoch": 1.03, "grad_norm": 17.67457569626726, "learning_rate": 1.5305880904715177e-05, "loss": 0.6796, "step": 6575 }, { "epoch": 1.03, "grad_norm": 17.914617544826108, "learning_rate": 1.530445138724619e-05, "loss": 0.6844, "step": 6576 }, { "epoch": 1.03, "grad_norm": 12.174804068189223, "learning_rate": 1.53030217189217e-05, "loss": 0.6634, "step": 6577 }, { "epoch": 1.03, "grad_norm": 24.172706655658924, "learning_rate": 1.5301591899782376e-05, "loss": 0.6934, "step": 6578 }, { "epoch": 1.03, "grad_norm": 12.762627190097165, "learning_rate": 1.5300161929868886e-05, "loss": 0.7323, "step": 6579 }, { "epoch": 1.03, "grad_norm": 15.243914724334267, "learning_rate": 1.5298731809221886e-05, "loss": 0.6541, "step": 6580 }, { "epoch": 1.03, "grad_norm": 21.242675904007758, "learning_rate": 1.5297301537882055e-05, "loss": 0.7242, "step": 6581 }, { "epoch": 1.03, "grad_norm": 17.2606390878493, "learning_rate": 1.5295871115890066e-05, "loss": 0.7013, "step": 6582 }, { "epoch": 1.03, "grad_norm": 22.768022064124573, "learning_rate": 1.52944405432866e-05, "loss": 0.7681, "step": 6583 }, { "epoch": 1.03, "grad_norm": 20.229685884432772, "learning_rate": 1.5293009820112345e-05, "loss": 0.7712, "step": 6584 }, { "epoch": 1.03, "grad_norm": 23.52838174973978, "learning_rate": 1.5291578946407985e-05, "loss": 0.7438, "step": 6585 }, { "epoch": 1.03, "grad_norm": 25.288551781476045, "learning_rate": 1.5290147922214212e-05, "loss": 0.736, "step": 6586 }, { "epoch": 1.03, "grad_norm": 18.289904916199976, "learning_rate": 1.5288716747571735e-05, "loss": 0.7038, "step": 6587 }, { "epoch": 1.03, "grad_norm": 17.69966445189918, "learning_rate": 1.5287285422521238e-05, "loss": 0.6714, "step": 6588 }, { "epoch": 1.03, "grad_norm": 24.50131665980954, "learning_rate": 1.5285853947103446e-05, "loss": 0.682, "step": 6589 }, { "epoch": 1.03, "grad_norm": 18.383778157224622, "learning_rate": 1.5284422321359054e-05, "loss": 0.6443, "step": 6590 }, { "epoch": 1.03, "grad_norm": 16.396775495388287, "learning_rate": 1.5282990545328782e-05, "loss": 0.6823, "step": 6591 }, { "epoch": 1.03, "grad_norm": 21.96607641078348, "learning_rate": 1.5281558619053353e-05, "loss": 0.7675, "step": 6592 }, { "epoch": 1.03, "grad_norm": 21.713466718998358, "learning_rate": 1.5280126542573484e-05, "loss": 0.7566, "step": 6593 }, { "epoch": 1.03, "grad_norm": 22.48877243942122, "learning_rate": 1.5278694315929906e-05, "loss": 0.816, "step": 6594 }, { "epoch": 1.03, "grad_norm": 18.04490309713365, "learning_rate": 1.527726193916335e-05, "loss": 0.6772, "step": 6595 }, { "epoch": 1.03, "grad_norm": 22.346228167348286, "learning_rate": 1.5275829412314547e-05, "loss": 0.6375, "step": 6596 }, { "epoch": 1.03, "grad_norm": 19.052906827864728, "learning_rate": 1.5274396735424244e-05, "loss": 0.664, "step": 6597 }, { "epoch": 1.03, "grad_norm": 19.91804025692924, "learning_rate": 1.5272963908533184e-05, "loss": 0.6322, "step": 6598 }, { "epoch": 1.03, "grad_norm": 17.815045410789473, "learning_rate": 1.5271530931682116e-05, "loss": 0.7188, "step": 6599 }, { "epoch": 1.03, "grad_norm": 17.794946515808697, "learning_rate": 1.5270097804911794e-05, "loss": 0.7057, "step": 6600 }, { "epoch": 1.03, "grad_norm": 18.807764284614507, "learning_rate": 1.526866452826297e-05, "loss": 0.7643, "step": 6601 }, { "epoch": 1.03, "grad_norm": 19.98111661753733, "learning_rate": 1.526723110177641e-05, "loss": 0.6432, "step": 6602 }, { "epoch": 1.03, "grad_norm": 15.439288304104851, "learning_rate": 1.5265797525492878e-05, "loss": 0.7841, "step": 6603 }, { "epoch": 1.03, "grad_norm": 19.356104020241695, "learning_rate": 1.526436379945315e-05, "loss": 0.6421, "step": 6604 }, { "epoch": 1.03, "grad_norm": 17.35263202201593, "learning_rate": 1.5262929923697986e-05, "loss": 0.739, "step": 6605 }, { "epoch": 1.03, "grad_norm": 15.834669111293513, "learning_rate": 1.526149589826818e-05, "loss": 0.6509, "step": 6606 }, { "epoch": 1.03, "grad_norm": 16.840777058710355, "learning_rate": 1.5260061723204506e-05, "loss": 0.7117, "step": 6607 }, { "epoch": 1.03, "grad_norm": 17.481618551021654, "learning_rate": 1.5258627398547754e-05, "loss": 0.6568, "step": 6608 }, { "epoch": 1.03, "grad_norm": 17.80696078053773, "learning_rate": 1.5257192924338715e-05, "loss": 0.6788, "step": 6609 }, { "epoch": 1.03, "grad_norm": 17.284141943746125, "learning_rate": 1.5255758300618184e-05, "loss": 0.7012, "step": 6610 }, { "epoch": 1.03, "grad_norm": 23.337081794919776, "learning_rate": 1.5254323527426964e-05, "loss": 0.6966, "step": 6611 }, { "epoch": 1.03, "grad_norm": 13.714745961228799, "learning_rate": 1.5252888604805853e-05, "loss": 0.586, "step": 6612 }, { "epoch": 1.03, "grad_norm": 25.742324016596896, "learning_rate": 1.5251453532795663e-05, "loss": 0.7441, "step": 6613 }, { "epoch": 1.03, "grad_norm": 26.83939254152738, "learning_rate": 1.5250018311437212e-05, "loss": 0.7659, "step": 6614 }, { "epoch": 1.03, "grad_norm": 14.680834414274619, "learning_rate": 1.5248582940771306e-05, "loss": 0.7014, "step": 6615 }, { "epoch": 1.03, "grad_norm": 16.952776197281985, "learning_rate": 1.5247147420838774e-05, "loss": 0.7046, "step": 6616 }, { "epoch": 1.03, "grad_norm": 23.594671322979067, "learning_rate": 1.5245711751680438e-05, "loss": 0.7383, "step": 6617 }, { "epoch": 1.03, "grad_norm": 23.595836471358915, "learning_rate": 1.524427593333713e-05, "loss": 0.6726, "step": 6618 }, { "epoch": 1.03, "grad_norm": 14.10038194654784, "learning_rate": 1.524283996584968e-05, "loss": 0.6717, "step": 6619 }, { "epoch": 1.03, "grad_norm": 16.71986501598634, "learning_rate": 1.524140384925893e-05, "loss": 0.681, "step": 6620 }, { "epoch": 1.03, "grad_norm": 22.48799162203246, "learning_rate": 1.5239967583605719e-05, "loss": 0.6372, "step": 6621 }, { "epoch": 1.03, "grad_norm": 20.083716389523538, "learning_rate": 1.5238531168930894e-05, "loss": 0.6297, "step": 6622 }, { "epoch": 1.03, "grad_norm": 16.48344413153892, "learning_rate": 1.5237094605275311e-05, "loss": 0.6848, "step": 6623 }, { "epoch": 1.03, "grad_norm": 19.151736963888386, "learning_rate": 1.5235657892679818e-05, "loss": 0.7363, "step": 6624 }, { "epoch": 1.03, "grad_norm": 21.90782770211155, "learning_rate": 1.523422103118528e-05, "loss": 0.6543, "step": 6625 }, { "epoch": 1.03, "grad_norm": 28.46399421910907, "learning_rate": 1.5232784020832556e-05, "loss": 0.6591, "step": 6626 }, { "epoch": 1.04, "grad_norm": 23.43643280905722, "learning_rate": 1.5231346861662518e-05, "loss": 0.7411, "step": 6627 }, { "epoch": 1.04, "grad_norm": 20.530727861752872, "learning_rate": 1.5229909553716032e-05, "loss": 0.6726, "step": 6628 }, { "epoch": 1.04, "grad_norm": 19.942418488788668, "learning_rate": 1.5228472097033979e-05, "loss": 0.7344, "step": 6629 }, { "epoch": 1.04, "grad_norm": 15.667152133851134, "learning_rate": 1.5227034491657235e-05, "loss": 0.6592, "step": 6630 }, { "epoch": 1.04, "grad_norm": 22.635822310648, "learning_rate": 1.5225596737626695e-05, "loss": 0.7934, "step": 6631 }, { "epoch": 1.04, "grad_norm": 19.207611311074004, "learning_rate": 1.5224158834983234e-05, "loss": 0.6699, "step": 6632 }, { "epoch": 1.04, "grad_norm": 30.850181217816736, "learning_rate": 1.5222720783767755e-05, "loss": 0.6371, "step": 6633 }, { "epoch": 1.04, "grad_norm": 22.081180126253468, "learning_rate": 1.522128258402115e-05, "loss": 0.6717, "step": 6634 }, { "epoch": 1.04, "grad_norm": 13.79146770108563, "learning_rate": 1.5219844235784326e-05, "loss": 0.6334, "step": 6635 }, { "epoch": 1.04, "grad_norm": 15.525451205576074, "learning_rate": 1.5218405739098183e-05, "loss": 0.6134, "step": 6636 }, { "epoch": 1.04, "grad_norm": 19.671374621780412, "learning_rate": 1.5216967094003633e-05, "loss": 0.6702, "step": 6637 }, { "epoch": 1.04, "grad_norm": 17.72162200427294, "learning_rate": 1.5215528300541593e-05, "loss": 0.6383, "step": 6638 }, { "epoch": 1.04, "grad_norm": 20.024548302186307, "learning_rate": 1.5214089358752979e-05, "loss": 0.7096, "step": 6639 }, { "epoch": 1.04, "grad_norm": 18.861294975777007, "learning_rate": 1.5212650268678711e-05, "loss": 0.7022, "step": 6640 }, { "epoch": 1.04, "grad_norm": 13.541978602912499, "learning_rate": 1.5211211030359719e-05, "loss": 0.6873, "step": 6641 }, { "epoch": 1.04, "grad_norm": 18.905069210285983, "learning_rate": 1.5209771643836937e-05, "loss": 0.7557, "step": 6642 }, { "epoch": 1.04, "grad_norm": 13.949176372184894, "learning_rate": 1.5208332109151295e-05, "loss": 0.7601, "step": 6643 }, { "epoch": 1.04, "grad_norm": 14.168979017672584, "learning_rate": 1.5206892426343736e-05, "loss": 0.6954, "step": 6644 }, { "epoch": 1.04, "grad_norm": 13.81713949689299, "learning_rate": 1.5205452595455198e-05, "loss": 0.6907, "step": 6645 }, { "epoch": 1.04, "grad_norm": 12.719257058760887, "learning_rate": 1.5204012616526637e-05, "loss": 0.6739, "step": 6646 }, { "epoch": 1.04, "grad_norm": 19.61770966904786, "learning_rate": 1.5202572489598998e-05, "loss": 0.713, "step": 6647 }, { "epoch": 1.04, "grad_norm": 16.00530202475908, "learning_rate": 1.5201132214713249e-05, "loss": 0.7099, "step": 6648 }, { "epoch": 1.04, "grad_norm": 15.75996954853697, "learning_rate": 1.5199691791910333e-05, "loss": 0.7586, "step": 6649 }, { "epoch": 1.04, "grad_norm": 21.906346931052607, "learning_rate": 1.5198251221231232e-05, "loss": 0.7403, "step": 6650 }, { "epoch": 1.04, "grad_norm": 13.89578159295, "learning_rate": 1.51968105027169e-05, "loss": 0.6511, "step": 6651 }, { "epoch": 1.04, "grad_norm": 25.635977239549483, "learning_rate": 1.5195369636408322e-05, "loss": 0.7877, "step": 6652 }, { "epoch": 1.04, "grad_norm": 16.404539385126125, "learning_rate": 1.519392862234647e-05, "loss": 0.5677, "step": 6653 }, { "epoch": 1.04, "grad_norm": 20.03915373203264, "learning_rate": 1.5192487460572327e-05, "loss": 0.7096, "step": 6654 }, { "epoch": 1.04, "grad_norm": 18.096133952151117, "learning_rate": 1.5191046151126876e-05, "loss": 0.7408, "step": 6655 }, { "epoch": 1.04, "grad_norm": 16.70834941359846, "learning_rate": 1.5189604694051111e-05, "loss": 0.6242, "step": 6656 }, { "epoch": 1.04, "grad_norm": 17.453598259372303, "learning_rate": 1.5188163089386022e-05, "loss": 0.6273, "step": 6657 }, { "epoch": 1.04, "grad_norm": 17.856765901077996, "learning_rate": 1.5186721337172614e-05, "loss": 0.6754, "step": 6658 }, { "epoch": 1.04, "grad_norm": 19.852426534897962, "learning_rate": 1.5185279437451881e-05, "loss": 0.7661, "step": 6659 }, { "epoch": 1.04, "grad_norm": 10.698674770647406, "learning_rate": 1.5183837390264833e-05, "loss": 0.6751, "step": 6660 }, { "epoch": 1.04, "grad_norm": 21.809745454791692, "learning_rate": 1.5182395195652487e-05, "loss": 0.6831, "step": 6661 }, { "epoch": 1.04, "grad_norm": 14.689005475298254, "learning_rate": 1.518095285365585e-05, "loss": 0.6706, "step": 6662 }, { "epoch": 1.04, "grad_norm": 14.43701011788531, "learning_rate": 1.5179510364315945e-05, "loss": 0.6119, "step": 6663 }, { "epoch": 1.04, "grad_norm": 14.094374598430834, "learning_rate": 1.5178067727673797e-05, "loss": 0.6518, "step": 6664 }, { "epoch": 1.04, "grad_norm": 19.6080674989379, "learning_rate": 1.517662494377043e-05, "loss": 0.7072, "step": 6665 }, { "epoch": 1.04, "grad_norm": 21.187186529997053, "learning_rate": 1.5175182012646877e-05, "loss": 0.6354, "step": 6666 }, { "epoch": 1.04, "grad_norm": 12.323068565102645, "learning_rate": 1.5173738934344178e-05, "loss": 0.59, "step": 6667 }, { "epoch": 1.04, "grad_norm": 17.293548632210413, "learning_rate": 1.5172295708903368e-05, "loss": 0.7842, "step": 6668 }, { "epoch": 1.04, "grad_norm": 17.640274083217776, "learning_rate": 1.5170852336365494e-05, "loss": 0.6901, "step": 6669 }, { "epoch": 1.04, "grad_norm": 28.419465718381375, "learning_rate": 1.5169408816771605e-05, "loss": 0.6439, "step": 6670 }, { "epoch": 1.04, "grad_norm": 24.974292597093342, "learning_rate": 1.5167965150162753e-05, "loss": 0.6539, "step": 6671 }, { "epoch": 1.04, "grad_norm": 17.167377202645568, "learning_rate": 1.5166521336579996e-05, "loss": 0.6445, "step": 6672 }, { "epoch": 1.04, "grad_norm": 13.12948753071187, "learning_rate": 1.5165077376064394e-05, "loss": 0.6076, "step": 6673 }, { "epoch": 1.04, "grad_norm": 25.008362679118434, "learning_rate": 1.516363326865701e-05, "loss": 0.7471, "step": 6674 }, { "epoch": 1.04, "grad_norm": 20.526367413997082, "learning_rate": 1.5162189014398923e-05, "loss": 0.7342, "step": 6675 }, { "epoch": 1.04, "grad_norm": 28.163684441387435, "learning_rate": 1.5160744613331196e-05, "loss": 0.6954, "step": 6676 }, { "epoch": 1.04, "grad_norm": 27.353512265996105, "learning_rate": 1.5159300065494912e-05, "loss": 0.8157, "step": 6677 }, { "epoch": 1.04, "grad_norm": 17.217737586967633, "learning_rate": 1.515785537093115e-05, "loss": 0.7461, "step": 6678 }, { "epoch": 1.04, "grad_norm": 14.31736374654659, "learning_rate": 1.5156410529681e-05, "loss": 0.73, "step": 6679 }, { "epoch": 1.04, "grad_norm": 28.388256885394128, "learning_rate": 1.5154965541785554e-05, "loss": 0.6695, "step": 6680 }, { "epoch": 1.04, "grad_norm": 19.050016300627473, "learning_rate": 1.51535204072859e-05, "loss": 0.6573, "step": 6681 }, { "epoch": 1.04, "grad_norm": 18.55720826704021, "learning_rate": 1.515207512622314e-05, "loss": 0.6634, "step": 6682 }, { "epoch": 1.04, "grad_norm": 17.29011207442339, "learning_rate": 1.515062969863838e-05, "loss": 0.712, "step": 6683 }, { "epoch": 1.04, "grad_norm": 17.737950136059283, "learning_rate": 1.514918412457272e-05, "loss": 0.7008, "step": 6684 }, { "epoch": 1.04, "grad_norm": 15.189511755562448, "learning_rate": 1.5147738404067279e-05, "loss": 0.6075, "step": 6685 }, { "epoch": 1.04, "grad_norm": 20.262714429181987, "learning_rate": 1.5146292537163168e-05, "loss": 0.6746, "step": 6686 }, { "epoch": 1.04, "grad_norm": 17.636403245099917, "learning_rate": 1.5144846523901508e-05, "loss": 0.7292, "step": 6687 }, { "epoch": 1.04, "grad_norm": 17.09836613456225, "learning_rate": 1.5143400364323424e-05, "loss": 0.6327, "step": 6688 }, { "epoch": 1.04, "grad_norm": 16.782418149998275, "learning_rate": 1.5141954058470041e-05, "loss": 0.6115, "step": 6689 }, { "epoch": 1.04, "grad_norm": 24.250980679970176, "learning_rate": 1.5140507606382496e-05, "loss": 0.7613, "step": 6690 }, { "epoch": 1.05, "grad_norm": 17.284831227840556, "learning_rate": 1.5139061008101915e-05, "loss": 0.6453, "step": 6691 }, { "epoch": 1.05, "grad_norm": 30.36856014956248, "learning_rate": 1.5137614263669451e-05, "loss": 0.7937, "step": 6692 }, { "epoch": 1.05, "grad_norm": 27.24527272262378, "learning_rate": 1.5136167373126239e-05, "loss": 0.7274, "step": 6693 }, { "epoch": 1.05, "grad_norm": 27.303322647975598, "learning_rate": 1.5134720336513433e-05, "loss": 0.6853, "step": 6694 }, { "epoch": 1.05, "grad_norm": 13.510953574668997, "learning_rate": 1.5133273153872186e-05, "loss": 0.6834, "step": 6695 }, { "epoch": 1.05, "grad_norm": 17.93447534633121, "learning_rate": 1.5131825825243651e-05, "loss": 0.6649, "step": 6696 }, { "epoch": 1.05, "grad_norm": 12.860722140994472, "learning_rate": 1.5130378350668996e-05, "loss": 0.724, "step": 6697 }, { "epoch": 1.05, "grad_norm": 17.81838713951837, "learning_rate": 1.5128930730189377e-05, "loss": 0.6582, "step": 6698 }, { "epoch": 1.05, "grad_norm": 19.702926904739805, "learning_rate": 1.5127482963845968e-05, "loss": 0.7186, "step": 6699 }, { "epoch": 1.05, "grad_norm": 29.46530149899022, "learning_rate": 1.5126035051679946e-05, "loss": 0.7247, "step": 6700 }, { "epoch": 1.05, "grad_norm": 17.876680611502888, "learning_rate": 1.5124586993732483e-05, "loss": 0.7157, "step": 6701 }, { "epoch": 1.05, "grad_norm": 15.079512483867564, "learning_rate": 1.5123138790044767e-05, "loss": 0.6618, "step": 6702 }, { "epoch": 1.05, "grad_norm": 16.463695534052768, "learning_rate": 1.5121690440657977e-05, "loss": 0.6404, "step": 6703 }, { "epoch": 1.05, "grad_norm": 18.689674616117216, "learning_rate": 1.512024194561331e-05, "loss": 0.7385, "step": 6704 }, { "epoch": 1.05, "grad_norm": 13.592732213822007, "learning_rate": 1.5118793304951955e-05, "loss": 0.6684, "step": 6705 }, { "epoch": 1.05, "grad_norm": 15.832450765138345, "learning_rate": 1.5117344518715115e-05, "loss": 0.6442, "step": 6706 }, { "epoch": 1.05, "grad_norm": 22.80305041043113, "learning_rate": 1.511589558694399e-05, "loss": 0.6963, "step": 6707 }, { "epoch": 1.05, "grad_norm": 19.56725129076971, "learning_rate": 1.5114446509679783e-05, "loss": 0.7202, "step": 6708 }, { "epoch": 1.05, "grad_norm": 18.534769864974145, "learning_rate": 1.5112997286963715e-05, "loss": 0.6659, "step": 6709 }, { "epoch": 1.05, "grad_norm": 17.26281817306663, "learning_rate": 1.5111547918836994e-05, "loss": 0.6833, "step": 6710 }, { "epoch": 1.05, "grad_norm": 17.506014359211708, "learning_rate": 1.5110098405340838e-05, "loss": 0.6883, "step": 6711 }, { "epoch": 1.05, "grad_norm": 12.400366255735799, "learning_rate": 1.5108648746516475e-05, "loss": 0.6027, "step": 6712 }, { "epoch": 1.05, "grad_norm": 12.308729778132559, "learning_rate": 1.5107198942405128e-05, "loss": 0.6443, "step": 6713 }, { "epoch": 1.05, "grad_norm": 22.632348640831825, "learning_rate": 1.5105748993048032e-05, "loss": 0.7034, "step": 6714 }, { "epoch": 1.05, "grad_norm": 21.452454173052796, "learning_rate": 1.5104298898486423e-05, "loss": 0.7327, "step": 6715 }, { "epoch": 1.05, "grad_norm": 24.4268050596143, "learning_rate": 1.5102848658761535e-05, "loss": 0.739, "step": 6716 }, { "epoch": 1.05, "grad_norm": 14.614295445019987, "learning_rate": 1.5101398273914623e-05, "loss": 0.6471, "step": 6717 }, { "epoch": 1.05, "grad_norm": 20.486895905852382, "learning_rate": 1.509994774398692e-05, "loss": 0.7768, "step": 6718 }, { "epoch": 1.05, "grad_norm": 22.19455573091933, "learning_rate": 1.5098497069019693e-05, "loss": 0.794, "step": 6719 }, { "epoch": 1.05, "grad_norm": 27.27445002489998, "learning_rate": 1.5097046249054187e-05, "loss": 0.7013, "step": 6720 }, { "epoch": 1.05, "grad_norm": 15.220557597984667, "learning_rate": 1.5095595284131674e-05, "loss": 0.7077, "step": 6721 }, { "epoch": 1.05, "grad_norm": 16.517174452293958, "learning_rate": 1.5094144174293406e-05, "loss": 0.6298, "step": 6722 }, { "epoch": 1.05, "grad_norm": 16.60737034958921, "learning_rate": 1.509269291958066e-05, "loss": 0.7088, "step": 6723 }, { "epoch": 1.05, "grad_norm": 15.646346818630947, "learning_rate": 1.5091241520034705e-05, "loss": 0.7184, "step": 6724 }, { "epoch": 1.05, "grad_norm": 21.632190975079098, "learning_rate": 1.508978997569682e-05, "loss": 0.6687, "step": 6725 }, { "epoch": 1.05, "grad_norm": 21.012759653616886, "learning_rate": 1.5088338286608287e-05, "loss": 0.7022, "step": 6726 }, { "epoch": 1.05, "grad_norm": 21.635143369369796, "learning_rate": 1.508688645281039e-05, "loss": 0.7386, "step": 6727 }, { "epoch": 1.05, "grad_norm": 13.433604095379572, "learning_rate": 1.5085434474344416e-05, "loss": 0.6502, "step": 6728 }, { "epoch": 1.05, "grad_norm": 27.862097299616963, "learning_rate": 1.508398235125166e-05, "loss": 0.6882, "step": 6729 }, { "epoch": 1.05, "grad_norm": 20.576157288620845, "learning_rate": 1.5082530083573421e-05, "loss": 0.7651, "step": 6730 }, { "epoch": 1.05, "grad_norm": 19.890640653962166, "learning_rate": 1.5081077671350998e-05, "loss": 0.7228, "step": 6731 }, { "epoch": 1.05, "grad_norm": 15.263088205884493, "learning_rate": 1.5079625114625701e-05, "loss": 0.6851, "step": 6732 }, { "epoch": 1.05, "grad_norm": 16.431106882462366, "learning_rate": 1.5078172413438836e-05, "loss": 0.6179, "step": 6733 }, { "epoch": 1.05, "grad_norm": 21.37782525191314, "learning_rate": 1.507671956783172e-05, "loss": 0.7089, "step": 6734 }, { "epoch": 1.05, "grad_norm": 17.936303321450204, "learning_rate": 1.5075266577845668e-05, "loss": 0.6976, "step": 6735 }, { "epoch": 1.05, "grad_norm": 18.63058520146432, "learning_rate": 1.5073813443522005e-05, "loss": 0.6687, "step": 6736 }, { "epoch": 1.05, "grad_norm": 14.870071077366125, "learning_rate": 1.5072360164902052e-05, "loss": 0.5894, "step": 6737 }, { "epoch": 1.05, "grad_norm": 26.251515197166363, "learning_rate": 1.5070906742027145e-05, "loss": 0.7214, "step": 6738 }, { "epoch": 1.05, "grad_norm": 32.10956064806358, "learning_rate": 1.5069453174938618e-05, "loss": 0.6366, "step": 6739 }, { "epoch": 1.05, "grad_norm": 16.493373581536158, "learning_rate": 1.5067999463677807e-05, "loss": 0.7083, "step": 6740 }, { "epoch": 1.05, "grad_norm": 16.7424982863581, "learning_rate": 1.5066545608286056e-05, "loss": 0.7417, "step": 6741 }, { "epoch": 1.05, "grad_norm": 23.184725924503475, "learning_rate": 1.5065091608804712e-05, "loss": 0.7666, "step": 6742 }, { "epoch": 1.05, "grad_norm": 18.12462423938081, "learning_rate": 1.5063637465275127e-05, "loss": 0.7064, "step": 6743 }, { "epoch": 1.05, "grad_norm": 21.929054207961638, "learning_rate": 1.5062183177738652e-05, "loss": 0.6309, "step": 6744 }, { "epoch": 1.05, "grad_norm": 18.467203612465383, "learning_rate": 1.5060728746236648e-05, "loss": 0.6572, "step": 6745 }, { "epoch": 1.05, "grad_norm": 23.7776663156548, "learning_rate": 1.5059274170810483e-05, "loss": 0.7138, "step": 6746 }, { "epoch": 1.05, "grad_norm": 16.73596376469182, "learning_rate": 1.505781945150152e-05, "loss": 0.6955, "step": 6747 }, { "epoch": 1.05, "grad_norm": 21.814254819269365, "learning_rate": 1.5056364588351127e-05, "loss": 0.6977, "step": 6748 }, { "epoch": 1.05, "grad_norm": 12.190434135687397, "learning_rate": 1.5054909581400683e-05, "loss": 0.6426, "step": 6749 }, { "epoch": 1.05, "grad_norm": 24.528145942382583, "learning_rate": 1.505345443069157e-05, "loss": 0.6864, "step": 6750 }, { "epoch": 1.05, "grad_norm": 18.28437851602784, "learning_rate": 1.505199913626517e-05, "loss": 0.6202, "step": 6751 }, { "epoch": 1.05, "grad_norm": 12.074780002212918, "learning_rate": 1.5050543698162865e-05, "loss": 0.6105, "step": 6752 }, { "epoch": 1.05, "grad_norm": 17.03098004109674, "learning_rate": 1.5049088116426056e-05, "loss": 0.64, "step": 6753 }, { "epoch": 1.05, "grad_norm": 14.2625640121314, "learning_rate": 1.5047632391096132e-05, "loss": 0.7421, "step": 6754 }, { "epoch": 1.06, "grad_norm": 11.196997694040093, "learning_rate": 1.5046176522214497e-05, "loss": 0.6098, "step": 6755 }, { "epoch": 1.06, "grad_norm": 21.544319961933265, "learning_rate": 1.5044720509822553e-05, "loss": 0.6651, "step": 6756 }, { "epoch": 1.06, "grad_norm": 15.107893537816315, "learning_rate": 1.5043264353961711e-05, "loss": 0.6649, "step": 6757 }, { "epoch": 1.06, "grad_norm": 16.688034368292268, "learning_rate": 1.5041808054673379e-05, "loss": 0.6494, "step": 6758 }, { "epoch": 1.06, "grad_norm": 27.510079136349486, "learning_rate": 1.5040351611998975e-05, "loss": 0.698, "step": 6759 }, { "epoch": 1.06, "grad_norm": 26.915014561283886, "learning_rate": 1.5038895025979919e-05, "loss": 0.7072, "step": 6760 }, { "epoch": 1.06, "grad_norm": 15.561983282825631, "learning_rate": 1.503743829665764e-05, "loss": 0.6981, "step": 6761 }, { "epoch": 1.06, "grad_norm": 19.339591041413275, "learning_rate": 1.503598142407356e-05, "loss": 0.5992, "step": 6762 }, { "epoch": 1.06, "grad_norm": 16.477260059333883, "learning_rate": 1.5034524408269115e-05, "loss": 0.6849, "step": 6763 }, { "epoch": 1.06, "grad_norm": 22.58687186312458, "learning_rate": 1.503306724928574e-05, "loss": 0.7011, "step": 6764 }, { "epoch": 1.06, "grad_norm": 19.31629383606017, "learning_rate": 1.5031609947164876e-05, "loss": 0.6922, "step": 6765 }, { "epoch": 1.06, "grad_norm": 18.86701573990697, "learning_rate": 1.503015250194797e-05, "loss": 0.6886, "step": 6766 }, { "epoch": 1.06, "grad_norm": 16.194127112590767, "learning_rate": 1.5028694913676469e-05, "loss": 0.6928, "step": 6767 }, { "epoch": 1.06, "grad_norm": 26.367849941550556, "learning_rate": 1.5027237182391825e-05, "loss": 0.7701, "step": 6768 }, { "epoch": 1.06, "grad_norm": 17.472990147525692, "learning_rate": 1.5025779308135499e-05, "loss": 0.792, "step": 6769 }, { "epoch": 1.06, "grad_norm": 14.092443508716261, "learning_rate": 1.5024321290948948e-05, "loss": 0.7054, "step": 6770 }, { "epoch": 1.06, "grad_norm": 20.347386482692883, "learning_rate": 1.502286313087364e-05, "loss": 0.6532, "step": 6771 }, { "epoch": 1.06, "grad_norm": 14.677010330138664, "learning_rate": 1.5021404827951039e-05, "loss": 0.6664, "step": 6772 }, { "epoch": 1.06, "grad_norm": 16.927849904431202, "learning_rate": 1.5019946382222626e-05, "loss": 0.6334, "step": 6773 }, { "epoch": 1.06, "grad_norm": 16.654999269114985, "learning_rate": 1.5018487793729875e-05, "loss": 0.6635, "step": 6774 }, { "epoch": 1.06, "grad_norm": 20.958678695046423, "learning_rate": 1.5017029062514263e-05, "loss": 0.6941, "step": 6775 }, { "epoch": 1.06, "grad_norm": 18.987532676438875, "learning_rate": 1.5015570188617283e-05, "loss": 0.7844, "step": 6776 }, { "epoch": 1.06, "grad_norm": 22.552739288482833, "learning_rate": 1.5014111172080416e-05, "loss": 0.604, "step": 6777 }, { "epoch": 1.06, "grad_norm": 25.16805239729693, "learning_rate": 1.5012652012945169e-05, "loss": 0.6479, "step": 6778 }, { "epoch": 1.06, "grad_norm": 32.77274945324569, "learning_rate": 1.5011192711253021e-05, "loss": 0.7088, "step": 6779 }, { "epoch": 1.06, "grad_norm": 12.730486535911156, "learning_rate": 1.5009733267045492e-05, "loss": 0.631, "step": 6780 }, { "epoch": 1.06, "grad_norm": 15.1626431667129, "learning_rate": 1.5008273680364074e-05, "loss": 0.7112, "step": 6781 }, { "epoch": 1.06, "grad_norm": 18.652303983635978, "learning_rate": 1.5006813951250286e-05, "loss": 0.7087, "step": 6782 }, { "epoch": 1.06, "grad_norm": 27.013477595308252, "learning_rate": 1.5005354079745636e-05, "loss": 0.8159, "step": 6783 }, { "epoch": 1.06, "grad_norm": 18.46465218051016, "learning_rate": 1.5003894065891644e-05, "loss": 0.7353, "step": 6784 }, { "epoch": 1.06, "grad_norm": 24.82916107560699, "learning_rate": 1.5002433909729832e-05, "loss": 0.6299, "step": 6785 }, { "epoch": 1.06, "grad_norm": 19.352828822555992, "learning_rate": 1.5000973611301729e-05, "loss": 0.8654, "step": 6786 }, { "epoch": 1.06, "grad_norm": 25.386211684563552, "learning_rate": 1.4999513170648856e-05, "loss": 0.6896, "step": 6787 }, { "epoch": 1.06, "grad_norm": 31.012217609550927, "learning_rate": 1.4998052587812759e-05, "loss": 0.6279, "step": 6788 }, { "epoch": 1.06, "grad_norm": 44.16336133377703, "learning_rate": 1.4996591862834965e-05, "loss": 0.756, "step": 6789 }, { "epoch": 1.06, "grad_norm": 15.678696744111086, "learning_rate": 1.4995130995757029e-05, "loss": 0.7094, "step": 6790 }, { "epoch": 1.06, "grad_norm": 12.709091299218962, "learning_rate": 1.499366998662048e-05, "loss": 0.6427, "step": 6791 }, { "epoch": 1.06, "grad_norm": 16.878654305948793, "learning_rate": 1.4992208835466884e-05, "loss": 0.725, "step": 6792 }, { "epoch": 1.06, "grad_norm": 32.343372325258606, "learning_rate": 1.4990747542337787e-05, "loss": 0.7414, "step": 6793 }, { "epoch": 1.06, "grad_norm": 17.317289096279897, "learning_rate": 1.498928610727475e-05, "loss": 0.6178, "step": 6794 }, { "epoch": 1.06, "grad_norm": 13.479834453502043, "learning_rate": 1.4987824530319335e-05, "loss": 0.7074, "step": 6795 }, { "epoch": 1.06, "grad_norm": 22.55327039355677, "learning_rate": 1.4986362811513107e-05, "loss": 0.6459, "step": 6796 }, { "epoch": 1.06, "grad_norm": 41.87650463475585, "learning_rate": 1.4984900950897638e-05, "loss": 0.7785, "step": 6797 }, { "epoch": 1.06, "grad_norm": 19.145241705308184, "learning_rate": 1.4983438948514501e-05, "loss": 0.7369, "step": 6798 }, { "epoch": 1.06, "grad_norm": 44.90480650013255, "learning_rate": 1.4981976804405278e-05, "loss": 0.76, "step": 6799 }, { "epoch": 1.06, "grad_norm": 16.24738458163355, "learning_rate": 1.4980514518611546e-05, "loss": 0.6368, "step": 6800 }, { "epoch": 1.06, "grad_norm": 19.602200210767613, "learning_rate": 1.49790520911749e-05, "loss": 0.7114, "step": 6801 }, { "epoch": 1.06, "grad_norm": 28.55989867119618, "learning_rate": 1.497758952213692e-05, "loss": 0.6931, "step": 6802 }, { "epoch": 1.06, "grad_norm": 17.88186985496182, "learning_rate": 1.4976126811539207e-05, "loss": 0.697, "step": 6803 }, { "epoch": 1.06, "grad_norm": 23.79902455808823, "learning_rate": 1.4974663959423358e-05, "loss": 0.6749, "step": 6804 }, { "epoch": 1.06, "grad_norm": 15.462700658944087, "learning_rate": 1.497320096583098e-05, "loss": 0.7276, "step": 6805 }, { "epoch": 1.06, "grad_norm": 20.465791524138826, "learning_rate": 1.497173783080367e-05, "loss": 0.7689, "step": 6806 }, { "epoch": 1.06, "grad_norm": 15.511313547679306, "learning_rate": 1.497027455438305e-05, "loss": 0.5923, "step": 6807 }, { "epoch": 1.06, "grad_norm": 14.123060524762046, "learning_rate": 1.4968811136610724e-05, "loss": 0.6836, "step": 6808 }, { "epoch": 1.06, "grad_norm": 15.695988198598952, "learning_rate": 1.4967347577528319e-05, "loss": 0.7611, "step": 6809 }, { "epoch": 1.06, "grad_norm": 13.851156695683098, "learning_rate": 1.4965883877177454e-05, "loss": 0.6637, "step": 6810 }, { "epoch": 1.06, "grad_norm": 27.71474204552171, "learning_rate": 1.4964420035599756e-05, "loss": 0.7271, "step": 6811 }, { "epoch": 1.06, "grad_norm": 18.816600197094832, "learning_rate": 1.4962956052836858e-05, "loss": 0.7371, "step": 6812 }, { "epoch": 1.06, "grad_norm": 14.377530578813502, "learning_rate": 1.4961491928930394e-05, "loss": 0.6727, "step": 6813 }, { "epoch": 1.06, "grad_norm": 16.31630213409514, "learning_rate": 1.4960027663921996e-05, "loss": 0.6912, "step": 6814 }, { "epoch": 1.06, "grad_norm": 17.73835271599576, "learning_rate": 1.495856325785332e-05, "loss": 0.7391, "step": 6815 }, { "epoch": 1.06, "grad_norm": 20.810605633427603, "learning_rate": 1.4957098710766e-05, "loss": 0.6262, "step": 6816 }, { "epoch": 1.06, "grad_norm": 14.876832552132404, "learning_rate": 1.4955634022701695e-05, "loss": 0.6713, "step": 6817 }, { "epoch": 1.06, "grad_norm": 24.96261159479596, "learning_rate": 1.4954169193702058e-05, "loss": 0.6871, "step": 6818 }, { "epoch": 1.07, "grad_norm": 16.196235310255712, "learning_rate": 1.4952704223808745e-05, "loss": 0.6898, "step": 6819 }, { "epoch": 1.07, "grad_norm": 29.31788666994457, "learning_rate": 1.4951239113063425e-05, "loss": 0.7327, "step": 6820 }, { "epoch": 1.07, "grad_norm": 20.302621169001725, "learning_rate": 1.4949773861507754e-05, "loss": 0.6682, "step": 6821 }, { "epoch": 1.07, "grad_norm": 22.85392665305955, "learning_rate": 1.4948308469183419e-05, "loss": 0.7063, "step": 6822 }, { "epoch": 1.07, "grad_norm": 22.082330922027204, "learning_rate": 1.494684293613208e-05, "loss": 0.666, "step": 6823 }, { "epoch": 1.07, "grad_norm": 15.04374310965196, "learning_rate": 1.4945377262395425e-05, "loss": 0.6601, "step": 6824 }, { "epoch": 1.07, "grad_norm": 16.389755900554096, "learning_rate": 1.4943911448015129e-05, "loss": 0.6056, "step": 6825 }, { "epoch": 1.07, "grad_norm": 15.100140596913102, "learning_rate": 1.4942445493032887e-05, "loss": 0.6693, "step": 6826 }, { "epoch": 1.07, "grad_norm": 13.478692802455363, "learning_rate": 1.4940979397490386e-05, "loss": 0.5297, "step": 6827 }, { "epoch": 1.07, "grad_norm": 21.680751222512853, "learning_rate": 1.493951316142932e-05, "loss": 0.5804, "step": 6828 }, { "epoch": 1.07, "grad_norm": 11.251738047120982, "learning_rate": 1.493804678489139e-05, "loss": 0.6372, "step": 6829 }, { "epoch": 1.07, "grad_norm": 28.469253560709763, "learning_rate": 1.4936580267918299e-05, "loss": 0.8205, "step": 6830 }, { "epoch": 1.07, "grad_norm": 16.202536620085596, "learning_rate": 1.4935113610551751e-05, "loss": 0.6718, "step": 6831 }, { "epoch": 1.07, "grad_norm": 14.015048666122139, "learning_rate": 1.4933646812833461e-05, "loss": 0.6796, "step": 6832 }, { "epoch": 1.07, "grad_norm": 15.696228174848482, "learning_rate": 1.493217987480514e-05, "loss": 0.7411, "step": 6833 }, { "epoch": 1.07, "grad_norm": 14.884923190534414, "learning_rate": 1.4930712796508511e-05, "loss": 0.6621, "step": 6834 }, { "epoch": 1.07, "grad_norm": 24.478785677170265, "learning_rate": 1.4929245577985292e-05, "loss": 0.7677, "step": 6835 }, { "epoch": 1.07, "grad_norm": 22.19751390149639, "learning_rate": 1.4927778219277214e-05, "loss": 0.6948, "step": 6836 }, { "epoch": 1.07, "grad_norm": 19.090736125151793, "learning_rate": 1.4926310720426005e-05, "loss": 0.6565, "step": 6837 }, { "epoch": 1.07, "grad_norm": 22.064623228708854, "learning_rate": 1.4924843081473401e-05, "loss": 0.7807, "step": 6838 }, { "epoch": 1.07, "grad_norm": 24.472623109604786, "learning_rate": 1.492337530246114e-05, "loss": 0.6843, "step": 6839 }, { "epoch": 1.07, "grad_norm": 14.504632696364414, "learning_rate": 1.4921907383430964e-05, "loss": 0.7186, "step": 6840 }, { "epoch": 1.07, "grad_norm": 19.362482820556192, "learning_rate": 1.4920439324424627e-05, "loss": 0.6502, "step": 6841 }, { "epoch": 1.07, "grad_norm": 13.316981648675004, "learning_rate": 1.491897112548387e-05, "loss": 0.6199, "step": 6842 }, { "epoch": 1.07, "grad_norm": 19.835712021789067, "learning_rate": 1.4917502786650451e-05, "loss": 0.7515, "step": 6843 }, { "epoch": 1.07, "grad_norm": 18.10075321578142, "learning_rate": 1.4916034307966132e-05, "loss": 0.7038, "step": 6844 }, { "epoch": 1.07, "grad_norm": 17.994877366130115, "learning_rate": 1.4914565689472671e-05, "loss": 0.7293, "step": 6845 }, { "epoch": 1.07, "grad_norm": 16.10547906115209, "learning_rate": 1.4913096931211839e-05, "loss": 0.6755, "step": 6846 }, { "epoch": 1.07, "grad_norm": 20.547108326881645, "learning_rate": 1.4911628033225402e-05, "loss": 0.7262, "step": 6847 }, { "epoch": 1.07, "grad_norm": 20.29024413453594, "learning_rate": 1.4910158995555138e-05, "loss": 0.6144, "step": 6848 }, { "epoch": 1.07, "grad_norm": 25.26053868031611, "learning_rate": 1.4908689818242826e-05, "loss": 0.7477, "step": 6849 }, { "epoch": 1.07, "grad_norm": 15.024619263793335, "learning_rate": 1.4907220501330245e-05, "loss": 0.6162, "step": 6850 }, { "epoch": 1.07, "grad_norm": 15.205784198534877, "learning_rate": 1.4905751044859185e-05, "loss": 0.7409, "step": 6851 }, { "epoch": 1.07, "grad_norm": 12.162922534434607, "learning_rate": 1.4904281448871431e-05, "loss": 0.6259, "step": 6852 }, { "epoch": 1.07, "grad_norm": 15.590950670662059, "learning_rate": 1.4902811713408784e-05, "loss": 0.6334, "step": 6853 }, { "epoch": 1.07, "grad_norm": 15.42584840697611, "learning_rate": 1.4901341838513044e-05, "loss": 0.7229, "step": 6854 }, { "epoch": 1.07, "grad_norm": 20.918537038141352, "learning_rate": 1.4899871824226004e-05, "loss": 0.683, "step": 6855 }, { "epoch": 1.07, "grad_norm": 18.438816199764084, "learning_rate": 1.4898401670589478e-05, "loss": 0.6605, "step": 6856 }, { "epoch": 1.07, "grad_norm": 17.56973616255662, "learning_rate": 1.4896931377645273e-05, "loss": 0.6847, "step": 6857 }, { "epoch": 1.07, "grad_norm": 19.68732531901476, "learning_rate": 1.4895460945435205e-05, "loss": 0.6687, "step": 6858 }, { "epoch": 1.07, "grad_norm": 14.047093463976806, "learning_rate": 1.489399037400109e-05, "loss": 0.6741, "step": 6859 }, { "epoch": 1.07, "grad_norm": 24.990936668935863, "learning_rate": 1.4892519663384755e-05, "loss": 0.6475, "step": 6860 }, { "epoch": 1.07, "grad_norm": 17.909315085755992, "learning_rate": 1.489104881362802e-05, "loss": 0.6919, "step": 6861 }, { "epoch": 1.07, "grad_norm": 15.86212752560986, "learning_rate": 1.4889577824772719e-05, "loss": 0.6404, "step": 6862 }, { "epoch": 1.07, "grad_norm": 17.620987791412155, "learning_rate": 1.4888106696860684e-05, "loss": 0.6714, "step": 6863 }, { "epoch": 1.07, "grad_norm": 18.86707473784369, "learning_rate": 1.4886635429933756e-05, "loss": 0.6344, "step": 6864 }, { "epoch": 1.07, "grad_norm": 21.752094637922625, "learning_rate": 1.4885164024033771e-05, "loss": 0.7448, "step": 6865 }, { "epoch": 1.07, "grad_norm": 14.351449011788484, "learning_rate": 1.4883692479202582e-05, "loss": 0.6105, "step": 6866 }, { "epoch": 1.07, "grad_norm": 28.396978297255078, "learning_rate": 1.4882220795482037e-05, "loss": 0.7575, "step": 6867 }, { "epoch": 1.07, "grad_norm": 17.887766752180045, "learning_rate": 1.4880748972913987e-05, "loss": 0.7021, "step": 6868 }, { "epoch": 1.07, "grad_norm": 19.62241245670707, "learning_rate": 1.4879277011540289e-05, "loss": 0.698, "step": 6869 }, { "epoch": 1.07, "grad_norm": 27.94651450317409, "learning_rate": 1.4877804911402809e-05, "loss": 0.7187, "step": 6870 }, { "epoch": 1.07, "grad_norm": 20.232966576535382, "learning_rate": 1.4876332672543412e-05, "loss": 0.7409, "step": 6871 }, { "epoch": 1.07, "grad_norm": 22.66010890127837, "learning_rate": 1.4874860295003965e-05, "loss": 0.6974, "step": 6872 }, { "epoch": 1.07, "grad_norm": 14.804826238581002, "learning_rate": 1.4873387778826344e-05, "loss": 0.61, "step": 6873 }, { "epoch": 1.07, "grad_norm": 9.506525522659032, "learning_rate": 1.4871915124052426e-05, "loss": 0.5513, "step": 6874 }, { "epoch": 1.07, "grad_norm": 24.77126192268069, "learning_rate": 1.4870442330724088e-05, "loss": 0.6856, "step": 6875 }, { "epoch": 1.07, "grad_norm": 13.463418757623485, "learning_rate": 1.4868969398883224e-05, "loss": 0.6515, "step": 6876 }, { "epoch": 1.07, "grad_norm": 26.449551904890306, "learning_rate": 1.4867496328571716e-05, "loss": 0.6854, "step": 6877 }, { "epoch": 1.07, "grad_norm": 20.694815587233965, "learning_rate": 1.4866023119831461e-05, "loss": 0.7199, "step": 6878 }, { "epoch": 1.07, "grad_norm": 18.633412048002835, "learning_rate": 1.4864549772704353e-05, "loss": 0.7222, "step": 6879 }, { "epoch": 1.07, "grad_norm": 18.048754914532285, "learning_rate": 1.4863076287232297e-05, "loss": 0.6516, "step": 6880 }, { "epoch": 1.07, "grad_norm": 20.338989017140193, "learning_rate": 1.4861602663457195e-05, "loss": 0.702, "step": 6881 }, { "epoch": 1.07, "grad_norm": 23.753267535878702, "learning_rate": 1.4860128901420958e-05, "loss": 0.7323, "step": 6882 }, { "epoch": 1.08, "grad_norm": 14.328161696134313, "learning_rate": 1.4858655001165498e-05, "loss": 0.557, "step": 6883 }, { "epoch": 1.08, "grad_norm": 19.837909827298652, "learning_rate": 1.4857180962732731e-05, "loss": 0.7135, "step": 6884 }, { "epoch": 1.08, "grad_norm": 16.3281535883019, "learning_rate": 1.4855706786164579e-05, "loss": 0.7526, "step": 6885 }, { "epoch": 1.08, "grad_norm": 23.341651623031126, "learning_rate": 1.4854232471502967e-05, "loss": 0.6791, "step": 6886 }, { "epoch": 1.08, "grad_norm": 21.346725633794186, "learning_rate": 1.4852758018789826e-05, "loss": 0.6881, "step": 6887 }, { "epoch": 1.08, "grad_norm": 17.20036394845212, "learning_rate": 1.4851283428067083e-05, "loss": 0.7463, "step": 6888 }, { "epoch": 1.08, "grad_norm": 17.835894446335644, "learning_rate": 1.4849808699376675e-05, "loss": 0.6627, "step": 6889 }, { "epoch": 1.08, "grad_norm": 18.724042714175972, "learning_rate": 1.4848333832760549e-05, "loss": 0.6696, "step": 6890 }, { "epoch": 1.08, "grad_norm": 18.012475708687663, "learning_rate": 1.4846858828260642e-05, "loss": 0.6936, "step": 6891 }, { "epoch": 1.08, "grad_norm": 18.062443667954245, "learning_rate": 1.4845383685918906e-05, "loss": 0.7597, "step": 6892 }, { "epoch": 1.08, "grad_norm": 16.44012516103373, "learning_rate": 1.4843908405777294e-05, "loss": 0.7649, "step": 6893 }, { "epoch": 1.08, "grad_norm": 18.611921070592302, "learning_rate": 1.4842432987877758e-05, "loss": 0.716, "step": 6894 }, { "epoch": 1.08, "grad_norm": 13.432880388216796, "learning_rate": 1.4840957432262265e-05, "loss": 0.6406, "step": 6895 }, { "epoch": 1.08, "grad_norm": 16.203478788616906, "learning_rate": 1.483948173897277e-05, "loss": 0.7051, "step": 6896 }, { "epoch": 1.08, "grad_norm": 20.47160488268398, "learning_rate": 1.4838005908051245e-05, "loss": 0.6831, "step": 6897 }, { "epoch": 1.08, "grad_norm": 16.68274434016696, "learning_rate": 1.4836529939539665e-05, "loss": 0.6279, "step": 6898 }, { "epoch": 1.08, "grad_norm": 25.558568519814205, "learning_rate": 1.4835053833480003e-05, "loss": 0.6641, "step": 6899 }, { "epoch": 1.08, "grad_norm": 19.304881011551856, "learning_rate": 1.4833577589914234e-05, "loss": 0.6803, "step": 6900 }, { "epoch": 1.08, "grad_norm": 18.183987779783546, "learning_rate": 1.4832101208884352e-05, "loss": 0.6123, "step": 6901 }, { "epoch": 1.08, "grad_norm": 19.32523948960862, "learning_rate": 1.4830624690432331e-05, "loss": 0.6468, "step": 6902 }, { "epoch": 1.08, "grad_norm": 27.145288723905725, "learning_rate": 1.4829148034600174e-05, "loss": 0.6132, "step": 6903 }, { "epoch": 1.08, "grad_norm": 22.06759863853126, "learning_rate": 1.4827671241429873e-05, "loss": 0.6542, "step": 6904 }, { "epoch": 1.08, "grad_norm": 19.76076857723192, "learning_rate": 1.4826194310963424e-05, "loss": 0.7168, "step": 6905 }, { "epoch": 1.08, "grad_norm": 18.705553563992417, "learning_rate": 1.4824717243242835e-05, "loss": 0.6569, "step": 6906 }, { "epoch": 1.08, "grad_norm": 19.847961762781033, "learning_rate": 1.4823240038310108e-05, "loss": 0.6935, "step": 6907 }, { "epoch": 1.08, "grad_norm": 12.936835498837047, "learning_rate": 1.4821762696207254e-05, "loss": 0.5734, "step": 6908 }, { "epoch": 1.08, "grad_norm": 21.75644804650147, "learning_rate": 1.4820285216976288e-05, "loss": 0.7417, "step": 6909 }, { "epoch": 1.08, "grad_norm": 16.68685259125747, "learning_rate": 1.4818807600659237e-05, "loss": 0.6866, "step": 6910 }, { "epoch": 1.08, "grad_norm": 24.248879402485336, "learning_rate": 1.4817329847298112e-05, "loss": 0.6538, "step": 6911 }, { "epoch": 1.08, "grad_norm": 17.528503137326087, "learning_rate": 1.4815851956934948e-05, "loss": 0.6704, "step": 6912 }, { "epoch": 1.08, "grad_norm": 4.891502665721167, "learning_rate": 1.4814373929611768e-05, "loss": 0.7168, "step": 6913 }, { "epoch": 1.08, "grad_norm": 16.781617483807377, "learning_rate": 1.4812895765370609e-05, "loss": 0.7302, "step": 6914 }, { "epoch": 1.08, "grad_norm": 16.535920767769795, "learning_rate": 1.4811417464253512e-05, "loss": 0.6499, "step": 6915 }, { "epoch": 1.08, "grad_norm": 23.921355271128462, "learning_rate": 1.4809939026302518e-05, "loss": 0.6869, "step": 6916 }, { "epoch": 1.08, "grad_norm": 15.897900614003222, "learning_rate": 1.4808460451559669e-05, "loss": 0.6351, "step": 6917 }, { "epoch": 1.08, "grad_norm": 17.154011517809057, "learning_rate": 1.4806981740067018e-05, "loss": 0.588, "step": 6918 }, { "epoch": 1.08, "grad_norm": 18.81200540722553, "learning_rate": 1.480550289186662e-05, "loss": 0.644, "step": 6919 }, { "epoch": 1.08, "grad_norm": 20.645432092217323, "learning_rate": 1.480402390700053e-05, "loss": 0.7193, "step": 6920 }, { "epoch": 1.08, "grad_norm": 18.905216777564288, "learning_rate": 1.4802544785510811e-05, "loss": 0.7288, "step": 6921 }, { "epoch": 1.08, "grad_norm": 12.696237168603231, "learning_rate": 1.4801065527439526e-05, "loss": 0.5994, "step": 6922 }, { "epoch": 1.08, "grad_norm": 14.590824371024818, "learning_rate": 1.4799586132828747e-05, "loss": 0.6722, "step": 6923 }, { "epoch": 1.08, "grad_norm": 22.933085539106568, "learning_rate": 1.4798106601720545e-05, "loss": 0.6759, "step": 6924 }, { "epoch": 1.08, "grad_norm": 18.985911898940657, "learning_rate": 1.4796626934157e-05, "loss": 0.6826, "step": 6925 }, { "epoch": 1.08, "grad_norm": 21.291782400686994, "learning_rate": 1.4795147130180192e-05, "loss": 0.6581, "step": 6926 }, { "epoch": 1.08, "grad_norm": 23.520752775450195, "learning_rate": 1.4793667189832201e-05, "loss": 0.8795, "step": 6927 }, { "epoch": 1.08, "grad_norm": 20.867146188830034, "learning_rate": 1.4792187113155121e-05, "loss": 0.6364, "step": 6928 }, { "epoch": 1.08, "grad_norm": 19.228691108022982, "learning_rate": 1.4790706900191043e-05, "loss": 0.6169, "step": 6929 }, { "epoch": 1.08, "grad_norm": 16.889255036721124, "learning_rate": 1.4789226550982062e-05, "loss": 0.7175, "step": 6930 }, { "epoch": 1.08, "grad_norm": 17.9396156249684, "learning_rate": 1.4787746065570282e-05, "loss": 0.6739, "step": 6931 }, { "epoch": 1.08, "grad_norm": 23.950394583679003, "learning_rate": 1.4786265443997805e-05, "loss": 0.6497, "step": 6932 }, { "epoch": 1.08, "grad_norm": 19.86424577334854, "learning_rate": 1.4784784686306737e-05, "loss": 0.7775, "step": 6933 }, { "epoch": 1.08, "grad_norm": 25.279852644021776, "learning_rate": 1.478330379253919e-05, "loss": 0.7147, "step": 6934 }, { "epoch": 1.08, "grad_norm": 17.13845804282142, "learning_rate": 1.4781822762737284e-05, "loss": 0.6882, "step": 6935 }, { "epoch": 1.08, "grad_norm": 12.626746143671477, "learning_rate": 1.4780341596943133e-05, "loss": 0.6497, "step": 6936 }, { "epoch": 1.08, "grad_norm": 17.964296727852883, "learning_rate": 1.4778860295198869e-05, "loss": 0.644, "step": 6937 }, { "epoch": 1.08, "grad_norm": 21.62900263410205, "learning_rate": 1.4777378857546606e-05, "loss": 0.6984, "step": 6938 }, { "epoch": 1.08, "grad_norm": 21.706145554232766, "learning_rate": 1.4775897284028491e-05, "loss": 0.6706, "step": 6939 }, { "epoch": 1.08, "grad_norm": 23.973669696114257, "learning_rate": 1.4774415574686645e-05, "loss": 0.7294, "step": 6940 }, { "epoch": 1.08, "grad_norm": 14.739109649954834, "learning_rate": 1.4772933729563215e-05, "loss": 0.6569, "step": 6941 }, { "epoch": 1.08, "grad_norm": 13.057039554068108, "learning_rate": 1.4771451748700344e-05, "loss": 0.6426, "step": 6942 }, { "epoch": 1.08, "grad_norm": 15.33816397406472, "learning_rate": 1.4769969632140173e-05, "loss": 0.6911, "step": 6943 }, { "epoch": 1.08, "grad_norm": 24.74663158515733, "learning_rate": 1.4768487379924857e-05, "loss": 0.6781, "step": 6944 }, { "epoch": 1.08, "grad_norm": 17.80884483756154, "learning_rate": 1.476700499209655e-05, "loss": 0.7354, "step": 6945 }, { "epoch": 1.08, "grad_norm": 17.843543977182733, "learning_rate": 1.4765522468697406e-05, "loss": 0.7084, "step": 6946 }, { "epoch": 1.09, "grad_norm": 15.51182711738, "learning_rate": 1.4764039809769598e-05, "loss": 0.7018, "step": 6947 }, { "epoch": 1.09, "grad_norm": 16.914673627688813, "learning_rate": 1.4762557015355278e-05, "loss": 0.6176, "step": 6948 }, { "epoch": 1.09, "grad_norm": 31.854363101035872, "learning_rate": 1.4761074085496625e-05, "loss": 0.6629, "step": 6949 }, { "epoch": 1.09, "grad_norm": 10.88423216077385, "learning_rate": 1.4759591020235812e-05, "loss": 0.699, "step": 6950 }, { "epoch": 1.09, "grad_norm": 26.588928320100546, "learning_rate": 1.4758107819615014e-05, "loss": 0.6286, "step": 6951 }, { "epoch": 1.09, "grad_norm": 19.338798535155828, "learning_rate": 1.4756624483676412e-05, "loss": 0.6679, "step": 6952 }, { "epoch": 1.09, "grad_norm": 20.15223280847857, "learning_rate": 1.475514101246219e-05, "loss": 0.6892, "step": 6953 }, { "epoch": 1.09, "grad_norm": 16.254974697743698, "learning_rate": 1.4753657406014545e-05, "loss": 0.6922, "step": 6954 }, { "epoch": 1.09, "grad_norm": 13.536834734925447, "learning_rate": 1.475217366437566e-05, "loss": 0.6449, "step": 6955 }, { "epoch": 1.09, "grad_norm": 14.775998131666926, "learning_rate": 1.4750689787587735e-05, "loss": 0.6154, "step": 6956 }, { "epoch": 1.09, "grad_norm": 15.990829974346735, "learning_rate": 1.4749205775692974e-05, "loss": 0.7237, "step": 6957 }, { "epoch": 1.09, "grad_norm": 16.495954436179797, "learning_rate": 1.4747721628733578e-05, "loss": 0.6836, "step": 6958 }, { "epoch": 1.09, "grad_norm": 19.21748560181008, "learning_rate": 1.4746237346751757e-05, "loss": 0.6481, "step": 6959 }, { "epoch": 1.09, "grad_norm": 13.270806569916324, "learning_rate": 1.4744752929789723e-05, "loss": 0.5648, "step": 6960 }, { "epoch": 1.09, "grad_norm": 13.705429537995972, "learning_rate": 1.474326837788969e-05, "loss": 0.6576, "step": 6961 }, { "epoch": 1.09, "grad_norm": 20.46108868467707, "learning_rate": 1.4741783691093879e-05, "loss": 0.6488, "step": 6962 }, { "epoch": 1.09, "grad_norm": 14.672031985733904, "learning_rate": 1.4740298869444512e-05, "loss": 0.6764, "step": 6963 }, { "epoch": 1.09, "grad_norm": 11.330488535854496, "learning_rate": 1.4738813912983823e-05, "loss": 0.6726, "step": 6964 }, { "epoch": 1.09, "grad_norm": 15.997999848576214, "learning_rate": 1.4737328821754034e-05, "loss": 0.7138, "step": 6965 }, { "epoch": 1.09, "grad_norm": 17.873263037092144, "learning_rate": 1.4735843595797388e-05, "loss": 0.656, "step": 6966 }, { "epoch": 1.09, "grad_norm": 19.484086140844646, "learning_rate": 1.4734358235156117e-05, "loss": 0.6322, "step": 6967 }, { "epoch": 1.09, "grad_norm": 20.733840856449074, "learning_rate": 1.4732872739872469e-05, "loss": 0.6889, "step": 6968 }, { "epoch": 1.09, "grad_norm": 13.444808700051054, "learning_rate": 1.4731387109988689e-05, "loss": 0.7014, "step": 6969 }, { "epoch": 1.09, "grad_norm": 18.971813462003162, "learning_rate": 1.4729901345547027e-05, "loss": 0.6809, "step": 6970 }, { "epoch": 1.09, "grad_norm": 14.401088114612397, "learning_rate": 1.4728415446589736e-05, "loss": 0.6512, "step": 6971 }, { "epoch": 1.09, "grad_norm": 16.771706887609707, "learning_rate": 1.4726929413159077e-05, "loss": 0.6208, "step": 6972 }, { "epoch": 1.09, "grad_norm": 25.89304416188458, "learning_rate": 1.472544324529731e-05, "loss": 0.7429, "step": 6973 }, { "epoch": 1.09, "grad_norm": 22.627261270755298, "learning_rate": 1.4723956943046705e-05, "loss": 0.7454, "step": 6974 }, { "epoch": 1.09, "grad_norm": 11.86831949760473, "learning_rate": 1.4722470506449525e-05, "loss": 0.6715, "step": 6975 }, { "epoch": 1.09, "grad_norm": 20.250413323455156, "learning_rate": 1.4720983935548048e-05, "loss": 0.7234, "step": 6976 }, { "epoch": 1.09, "grad_norm": 17.19040646253769, "learning_rate": 1.4719497230384548e-05, "loss": 0.6732, "step": 6977 }, { "epoch": 1.09, "grad_norm": 15.156288867094405, "learning_rate": 1.4718010391001306e-05, "loss": 0.6864, "step": 6978 }, { "epoch": 1.09, "grad_norm": 19.736330097090796, "learning_rate": 1.4716523417440613e-05, "loss": 0.7096, "step": 6979 }, { "epoch": 1.09, "grad_norm": 14.31613064249538, "learning_rate": 1.4715036309744749e-05, "loss": 0.6621, "step": 6980 }, { "epoch": 1.09, "grad_norm": 20.727156947986735, "learning_rate": 1.4713549067956011e-05, "loss": 0.7025, "step": 6981 }, { "epoch": 1.09, "grad_norm": 20.937049214648837, "learning_rate": 1.4712061692116697e-05, "loss": 0.6454, "step": 6982 }, { "epoch": 1.09, "grad_norm": 19.215261549031933, "learning_rate": 1.4710574182269103e-05, "loss": 0.6641, "step": 6983 }, { "epoch": 1.09, "grad_norm": 20.26781923507785, "learning_rate": 1.4709086538455537e-05, "loss": 0.5997, "step": 6984 }, { "epoch": 1.09, "grad_norm": 28.266171831774347, "learning_rate": 1.4707598760718305e-05, "loss": 0.6695, "step": 6985 }, { "epoch": 1.09, "grad_norm": 25.94063095590854, "learning_rate": 1.4706110849099715e-05, "loss": 0.8163, "step": 6986 }, { "epoch": 1.09, "grad_norm": 19.96683123747085, "learning_rate": 1.4704622803642089e-05, "loss": 0.7017, "step": 6987 }, { "epoch": 1.09, "grad_norm": 25.847383674981895, "learning_rate": 1.470313462438774e-05, "loss": 0.6739, "step": 6988 }, { "epoch": 1.09, "grad_norm": 18.503805068844486, "learning_rate": 1.4701646311378994e-05, "loss": 0.7649, "step": 6989 }, { "epoch": 1.09, "grad_norm": 18.18007624496995, "learning_rate": 1.4700157864658176e-05, "loss": 0.6699, "step": 6990 }, { "epoch": 1.09, "grad_norm": 15.102198666801241, "learning_rate": 1.4698669284267619e-05, "loss": 0.613, "step": 6991 }, { "epoch": 1.09, "grad_norm": 19.507259196056836, "learning_rate": 1.4697180570249655e-05, "loss": 0.6672, "step": 6992 }, { "epoch": 1.09, "grad_norm": 13.20135285938748, "learning_rate": 1.4695691722646626e-05, "loss": 0.6565, "step": 6993 }, { "epoch": 1.09, "grad_norm": 17.399815078758454, "learning_rate": 1.4694202741500871e-05, "loss": 0.7691, "step": 6994 }, { "epoch": 1.09, "grad_norm": 17.254121472892848, "learning_rate": 1.4692713626854735e-05, "loss": 0.6645, "step": 6995 }, { "epoch": 1.09, "grad_norm": 10.544987441451005, "learning_rate": 1.4691224378750569e-05, "loss": 0.5736, "step": 6996 }, { "epoch": 1.09, "grad_norm": 14.133386995122116, "learning_rate": 1.4689734997230724e-05, "loss": 0.6855, "step": 6997 }, { "epoch": 1.09, "grad_norm": 17.569562101305245, "learning_rate": 1.4688245482337565e-05, "loss": 0.7389, "step": 6998 }, { "epoch": 1.09, "grad_norm": 19.67870523712877, "learning_rate": 1.468675583411344e-05, "loss": 0.7144, "step": 6999 }, { "epoch": 1.09, "grad_norm": 14.019001725156665, "learning_rate": 1.4685266052600724e-05, "loss": 0.6206, "step": 7000 }, { "epoch": 1.09, "grad_norm": 17.342001245159917, "learning_rate": 1.468377613784178e-05, "loss": 0.6379, "step": 7001 }, { "epoch": 1.09, "grad_norm": 20.443360173874197, "learning_rate": 1.4682286089878985e-05, "loss": 0.7224, "step": 7002 }, { "epoch": 1.09, "grad_norm": 23.846929963695523, "learning_rate": 1.4680795908754713e-05, "loss": 0.703, "step": 7003 }, { "epoch": 1.09, "grad_norm": 24.87026691275611, "learning_rate": 1.4679305594511344e-05, "loss": 0.7048, "step": 7004 }, { "epoch": 1.09, "grad_norm": 18.959445866404707, "learning_rate": 1.4677815147191256e-05, "loss": 0.7042, "step": 7005 }, { "epoch": 1.09, "grad_norm": 14.11895754842285, "learning_rate": 1.4676324566836848e-05, "loss": 0.7079, "step": 7006 }, { "epoch": 1.09, "grad_norm": 17.83222527330517, "learning_rate": 1.4674833853490502e-05, "loss": 0.7459, "step": 7007 }, { "epoch": 1.09, "grad_norm": 11.030453288455655, "learning_rate": 1.4673343007194618e-05, "loss": 0.6139, "step": 7008 }, { "epoch": 1.09, "grad_norm": 19.907334622091476, "learning_rate": 1.4671852027991589e-05, "loss": 0.6868, "step": 7009 }, { "epoch": 1.09, "grad_norm": 22.42533027097129, "learning_rate": 1.4670360915923824e-05, "loss": 0.6339, "step": 7010 }, { "epoch": 1.1, "grad_norm": 16.657633555780176, "learning_rate": 1.4668869671033726e-05, "loss": 0.6438, "step": 7011 }, { "epoch": 1.1, "grad_norm": 20.88928070322481, "learning_rate": 1.4667378293363706e-05, "loss": 0.7048, "step": 7012 }, { "epoch": 1.1, "grad_norm": 16.70095618885727, "learning_rate": 1.4665886782956178e-05, "loss": 0.61, "step": 7013 }, { "epoch": 1.1, "grad_norm": 19.453035742007238, "learning_rate": 1.4664395139853558e-05, "loss": 0.7035, "step": 7014 }, { "epoch": 1.1, "grad_norm": 26.709407210272033, "learning_rate": 1.466290336409827e-05, "loss": 0.6475, "step": 7015 }, { "epoch": 1.1, "grad_norm": 24.27223713843969, "learning_rate": 1.4661411455732735e-05, "loss": 0.7413, "step": 7016 }, { "epoch": 1.1, "grad_norm": 19.31140263843974, "learning_rate": 1.4659919414799389e-05, "loss": 0.6015, "step": 7017 }, { "epoch": 1.1, "grad_norm": 22.056088089260825, "learning_rate": 1.465842724134066e-05, "loss": 0.6879, "step": 7018 }, { "epoch": 1.1, "grad_norm": 20.38249931811291, "learning_rate": 1.4656934935398983e-05, "loss": 0.6906, "step": 7019 }, { "epoch": 1.1, "grad_norm": 17.207129877611006, "learning_rate": 1.4655442497016802e-05, "loss": 0.6343, "step": 7020 }, { "epoch": 1.1, "grad_norm": 20.646688022751317, "learning_rate": 1.4653949926236562e-05, "loss": 0.6299, "step": 7021 }, { "epoch": 1.1, "grad_norm": 17.525901557415224, "learning_rate": 1.4652457223100704e-05, "loss": 0.7107, "step": 7022 }, { "epoch": 1.1, "grad_norm": 21.2278939265862, "learning_rate": 1.4650964387651692e-05, "loss": 0.715, "step": 7023 }, { "epoch": 1.1, "grad_norm": 22.366730467816783, "learning_rate": 1.4649471419931964e-05, "loss": 0.6328, "step": 7024 }, { "epoch": 1.1, "grad_norm": 20.22701759975661, "learning_rate": 1.4647978319983998e-05, "loss": 0.6437, "step": 7025 }, { "epoch": 1.1, "grad_norm": 25.290032370833387, "learning_rate": 1.4646485087850239e-05, "loss": 0.7371, "step": 7026 }, { "epoch": 1.1, "grad_norm": 16.123669322687615, "learning_rate": 1.4644991723573167e-05, "loss": 0.7021, "step": 7027 }, { "epoch": 1.1, "grad_norm": 17.399931850758918, "learning_rate": 1.4643498227195247e-05, "loss": 0.7094, "step": 7028 }, { "epoch": 1.1, "grad_norm": 15.0638921235817, "learning_rate": 1.4642004598758955e-05, "loss": 0.6051, "step": 7029 }, { "epoch": 1.1, "grad_norm": 19.992809223377687, "learning_rate": 1.4640510838306765e-05, "loss": 0.6004, "step": 7030 }, { "epoch": 1.1, "grad_norm": 33.715305468849856, "learning_rate": 1.4639016945881164e-05, "loss": 0.7904, "step": 7031 }, { "epoch": 1.1, "grad_norm": 28.687614468349267, "learning_rate": 1.4637522921524632e-05, "loss": 0.6982, "step": 7032 }, { "epoch": 1.1, "grad_norm": 12.598117797439969, "learning_rate": 1.4636028765279662e-05, "loss": 0.5972, "step": 7033 }, { "epoch": 1.1, "grad_norm": 15.615879165286403, "learning_rate": 1.4634534477188745e-05, "loss": 0.6154, "step": 7034 }, { "epoch": 1.1, "grad_norm": 13.878544153149125, "learning_rate": 1.4633040057294379e-05, "loss": 0.7251, "step": 7035 }, { "epoch": 1.1, "grad_norm": 16.490215179963524, "learning_rate": 1.4631545505639066e-05, "loss": 0.638, "step": 7036 }, { "epoch": 1.1, "grad_norm": 21.998263878485687, "learning_rate": 1.4630050822265306e-05, "loss": 0.6917, "step": 7037 }, { "epoch": 1.1, "grad_norm": 18.251567147849286, "learning_rate": 1.4628556007215612e-05, "loss": 0.6591, "step": 7038 }, { "epoch": 1.1, "grad_norm": 22.090590785622393, "learning_rate": 1.462706106053249e-05, "loss": 0.6737, "step": 7039 }, { "epoch": 1.1, "grad_norm": 18.11842552024782, "learning_rate": 1.4625565982258461e-05, "loss": 0.7529, "step": 7040 }, { "epoch": 1.1, "grad_norm": 22.220165905724727, "learning_rate": 1.4624070772436037e-05, "loss": 0.6819, "step": 7041 }, { "epoch": 1.1, "grad_norm": 45.387024255695025, "learning_rate": 1.4622575431107748e-05, "loss": 0.7692, "step": 7042 }, { "epoch": 1.1, "grad_norm": 17.686625203750875, "learning_rate": 1.4621079958316118e-05, "loss": 0.6618, "step": 7043 }, { "epoch": 1.1, "grad_norm": 20.755306786830555, "learning_rate": 1.4619584354103675e-05, "loss": 0.6575, "step": 7044 }, { "epoch": 1.1, "grad_norm": 24.8441490561147, "learning_rate": 1.4618088618512957e-05, "loss": 0.7926, "step": 7045 }, { "epoch": 1.1, "grad_norm": 18.99545148704661, "learning_rate": 1.4616592751586503e-05, "loss": 0.6647, "step": 7046 }, { "epoch": 1.1, "grad_norm": 18.447742875621604, "learning_rate": 1.4615096753366849e-05, "loss": 0.715, "step": 7047 }, { "epoch": 1.1, "grad_norm": 14.451496408826745, "learning_rate": 1.4613600623896542e-05, "loss": 0.6422, "step": 7048 }, { "epoch": 1.1, "grad_norm": 16.301495811328188, "learning_rate": 1.4612104363218132e-05, "loss": 0.6394, "step": 7049 }, { "epoch": 1.1, "grad_norm": 20.492657231754652, "learning_rate": 1.4610607971374175e-05, "loss": 0.733, "step": 7050 }, { "epoch": 1.1, "grad_norm": 15.673587667932145, "learning_rate": 1.460911144840722e-05, "loss": 0.7125, "step": 7051 }, { "epoch": 1.1, "grad_norm": 18.09370370967656, "learning_rate": 1.4607614794359836e-05, "loss": 0.6866, "step": 7052 }, { "epoch": 1.1, "grad_norm": 19.41064252760542, "learning_rate": 1.4606118009274582e-05, "loss": 0.6593, "step": 7053 }, { "epoch": 1.1, "grad_norm": 24.503598446450553, "learning_rate": 1.4604621093194024e-05, "loss": 0.684, "step": 7054 }, { "epoch": 1.1, "grad_norm": 16.500413183706783, "learning_rate": 1.4603124046160737e-05, "loss": 0.7107, "step": 7055 }, { "epoch": 1.1, "grad_norm": 17.284920481997407, "learning_rate": 1.4601626868217296e-05, "loss": 0.6495, "step": 7056 }, { "epoch": 1.1, "grad_norm": 22.54281996045134, "learning_rate": 1.4600129559406278e-05, "loss": 0.7752, "step": 7057 }, { "epoch": 1.1, "grad_norm": 20.973435400765425, "learning_rate": 1.4598632119770266e-05, "loss": 0.7222, "step": 7058 }, { "epoch": 1.1, "grad_norm": 26.655250829102528, "learning_rate": 1.4597134549351845e-05, "loss": 0.6997, "step": 7059 }, { "epoch": 1.1, "grad_norm": 24.968223785852004, "learning_rate": 1.459563684819361e-05, "loss": 0.7215, "step": 7060 }, { "epoch": 1.1, "grad_norm": 16.77558051936856, "learning_rate": 1.4594139016338149e-05, "loss": 0.6302, "step": 7061 }, { "epoch": 1.1, "grad_norm": 20.65315032446035, "learning_rate": 1.4592641053828063e-05, "loss": 0.5942, "step": 7062 }, { "epoch": 1.1, "grad_norm": 19.807305845723146, "learning_rate": 1.4591142960705951e-05, "loss": 0.7034, "step": 7063 }, { "epoch": 1.1, "grad_norm": 18.018211518192775, "learning_rate": 1.4589644737014418e-05, "loss": 0.6423, "step": 7064 }, { "epoch": 1.1, "grad_norm": 21.051010367016783, "learning_rate": 1.4588146382796075e-05, "loss": 0.6319, "step": 7065 }, { "epoch": 1.1, "grad_norm": 20.951180637386592, "learning_rate": 1.4586647898093534e-05, "loss": 0.6474, "step": 7066 }, { "epoch": 1.1, "grad_norm": 19.7070851335649, "learning_rate": 1.4585149282949409e-05, "loss": 0.6891, "step": 7067 }, { "epoch": 1.1, "grad_norm": 16.384739329077902, "learning_rate": 1.458365053740632e-05, "loss": 0.6133, "step": 7068 }, { "epoch": 1.1, "grad_norm": 17.041241613600196, "learning_rate": 1.4582151661506894e-05, "loss": 0.5978, "step": 7069 }, { "epoch": 1.1, "grad_norm": 18.890526369782368, "learning_rate": 1.458065265529375e-05, "loss": 0.6614, "step": 7070 }, { "epoch": 1.1, "grad_norm": 17.048859252621583, "learning_rate": 1.4579153518809527e-05, "loss": 0.7275, "step": 7071 }, { "epoch": 1.1, "grad_norm": 13.468198100574117, "learning_rate": 1.4577654252096856e-05, "loss": 0.6838, "step": 7072 }, { "epoch": 1.1, "grad_norm": 20.25671740439634, "learning_rate": 1.4576154855198377e-05, "loss": 0.6564, "step": 7073 }, { "epoch": 1.1, "grad_norm": 17.427955231875117, "learning_rate": 1.457465532815673e-05, "loss": 0.7075, "step": 7074 }, { "epoch": 1.11, "grad_norm": 23.81805510441936, "learning_rate": 1.4573155671014563e-05, "loss": 0.729, "step": 7075 }, { "epoch": 1.11, "grad_norm": 20.555953099584958, "learning_rate": 1.457165588381452e-05, "loss": 0.7012, "step": 7076 }, { "epoch": 1.11, "grad_norm": 15.117698708104072, "learning_rate": 1.4570155966599264e-05, "loss": 0.6668, "step": 7077 }, { "epoch": 1.11, "grad_norm": 15.652709015139353, "learning_rate": 1.456865591941144e-05, "loss": 0.6553, "step": 7078 }, { "epoch": 1.11, "grad_norm": 16.639431213934756, "learning_rate": 1.4567155742293717e-05, "loss": 0.6589, "step": 7079 }, { "epoch": 1.11, "grad_norm": 28.755967781325303, "learning_rate": 1.4565655435288756e-05, "loss": 0.7819, "step": 7080 }, { "epoch": 1.11, "grad_norm": 12.085008097855225, "learning_rate": 1.4564154998439225e-05, "loss": 0.656, "step": 7081 }, { "epoch": 1.11, "grad_norm": 15.128384263745025, "learning_rate": 1.4562654431787797e-05, "loss": 0.6186, "step": 7082 }, { "epoch": 1.11, "grad_norm": 18.361402765257665, "learning_rate": 1.4561153735377147e-05, "loss": 0.6589, "step": 7083 }, { "epoch": 1.11, "grad_norm": 18.039083891368954, "learning_rate": 1.455965290924995e-05, "loss": 0.7627, "step": 7084 }, { "epoch": 1.11, "grad_norm": 19.872707496763248, "learning_rate": 1.4558151953448891e-05, "loss": 0.7493, "step": 7085 }, { "epoch": 1.11, "grad_norm": 20.48439826934913, "learning_rate": 1.4556650868016661e-05, "loss": 0.6464, "step": 7086 }, { "epoch": 1.11, "grad_norm": 18.80123666528361, "learning_rate": 1.4555149652995942e-05, "loss": 0.6105, "step": 7087 }, { "epoch": 1.11, "grad_norm": 20.48556201952367, "learning_rate": 1.4553648308429433e-05, "loss": 0.677, "step": 7088 }, { "epoch": 1.11, "grad_norm": 19.585760968648692, "learning_rate": 1.455214683435983e-05, "loss": 0.6422, "step": 7089 }, { "epoch": 1.11, "grad_norm": 15.374523158354936, "learning_rate": 1.4550645230829837e-05, "loss": 0.6592, "step": 7090 }, { "epoch": 1.11, "grad_norm": 23.031335417745673, "learning_rate": 1.4549143497882152e-05, "loss": 0.6826, "step": 7091 }, { "epoch": 1.11, "grad_norm": 19.136101080664933, "learning_rate": 1.4547641635559488e-05, "loss": 0.7095, "step": 7092 }, { "epoch": 1.11, "grad_norm": 29.673752854120632, "learning_rate": 1.4546139643904553e-05, "loss": 0.7743, "step": 7093 }, { "epoch": 1.11, "grad_norm": 28.004237428043275, "learning_rate": 1.4544637522960072e-05, "loss": 0.6392, "step": 7094 }, { "epoch": 1.11, "grad_norm": 17.403888674916924, "learning_rate": 1.4543135272768753e-05, "loss": 0.7561, "step": 7095 }, { "epoch": 1.11, "grad_norm": 22.2972724280328, "learning_rate": 1.454163289337333e-05, "loss": 0.7921, "step": 7096 }, { "epoch": 1.11, "grad_norm": 14.257958892697793, "learning_rate": 1.454013038481652e-05, "loss": 0.6995, "step": 7097 }, { "epoch": 1.11, "grad_norm": 14.417911662751203, "learning_rate": 1.4538627747141058e-05, "loss": 0.665, "step": 7098 }, { "epoch": 1.11, "grad_norm": 25.90885183627213, "learning_rate": 1.4537124980389678e-05, "loss": 0.7621, "step": 7099 }, { "epoch": 1.11, "grad_norm": 21.131756108475784, "learning_rate": 1.4535622084605119e-05, "loss": 0.6533, "step": 7100 }, { "epoch": 1.11, "grad_norm": 28.440700924110484, "learning_rate": 1.453411905983012e-05, "loss": 0.6031, "step": 7101 }, { "epoch": 1.11, "grad_norm": 15.446422980777664, "learning_rate": 1.453261590610743e-05, "loss": 0.6583, "step": 7102 }, { "epoch": 1.11, "grad_norm": 17.393097887192482, "learning_rate": 1.453111262347979e-05, "loss": 0.6708, "step": 7103 }, { "epoch": 1.11, "grad_norm": 25.235441142846927, "learning_rate": 1.4529609211989962e-05, "loss": 0.6485, "step": 7104 }, { "epoch": 1.11, "grad_norm": 15.911867624274608, "learning_rate": 1.4528105671680697e-05, "loss": 0.6828, "step": 7105 }, { "epoch": 1.11, "grad_norm": 25.534884267697496, "learning_rate": 1.4526602002594755e-05, "loss": 0.645, "step": 7106 }, { "epoch": 1.11, "grad_norm": 18.25805458823257, "learning_rate": 1.45250982047749e-05, "loss": 0.6715, "step": 7107 }, { "epoch": 1.11, "grad_norm": 19.48968898836983, "learning_rate": 1.4523594278263899e-05, "loss": 0.7343, "step": 7108 }, { "epoch": 1.11, "grad_norm": 17.322576181327083, "learning_rate": 1.4522090223104523e-05, "loss": 0.6218, "step": 7109 }, { "epoch": 1.11, "grad_norm": 11.337220602420146, "learning_rate": 1.4520586039339545e-05, "loss": 0.6086, "step": 7110 }, { "epoch": 1.11, "grad_norm": 18.37134331126699, "learning_rate": 1.4519081727011747e-05, "loss": 0.7534, "step": 7111 }, { "epoch": 1.11, "grad_norm": 14.238374376450317, "learning_rate": 1.4517577286163906e-05, "loss": 0.651, "step": 7112 }, { "epoch": 1.11, "grad_norm": 32.46718639999983, "learning_rate": 1.4516072716838815e-05, "loss": 0.6485, "step": 7113 }, { "epoch": 1.11, "grad_norm": 20.217345109327145, "learning_rate": 1.4514568019079252e-05, "loss": 0.8135, "step": 7114 }, { "epoch": 1.11, "grad_norm": 17.922151751322875, "learning_rate": 1.4513063192928016e-05, "loss": 0.6534, "step": 7115 }, { "epoch": 1.11, "grad_norm": 17.125813466611653, "learning_rate": 1.4511558238427905e-05, "loss": 0.6134, "step": 7116 }, { "epoch": 1.11, "grad_norm": 21.48041140043487, "learning_rate": 1.4510053155621714e-05, "loss": 0.6565, "step": 7117 }, { "epoch": 1.11, "grad_norm": 23.86290110105597, "learning_rate": 1.4508547944552254e-05, "loss": 0.6473, "step": 7118 }, { "epoch": 1.11, "grad_norm": 24.96522334530517, "learning_rate": 1.4507042605262324e-05, "loss": 0.6947, "step": 7119 }, { "epoch": 1.11, "grad_norm": 18.667638920649857, "learning_rate": 1.450553713779474e-05, "loss": 0.7727, "step": 7120 }, { "epoch": 1.11, "grad_norm": 17.191781196154576, "learning_rate": 1.4504031542192319e-05, "loss": 0.6742, "step": 7121 }, { "epoch": 1.11, "grad_norm": 14.86807658352979, "learning_rate": 1.450252581849787e-05, "loss": 0.6663, "step": 7122 }, { "epoch": 1.11, "grad_norm": 22.07577298831755, "learning_rate": 1.4501019966754225e-05, "loss": 0.6357, "step": 7123 }, { "epoch": 1.11, "grad_norm": 26.80208455819756, "learning_rate": 1.4499513987004201e-05, "loss": 0.7979, "step": 7124 }, { "epoch": 1.11, "grad_norm": 30.814780257309796, "learning_rate": 1.4498007879290635e-05, "loss": 0.7274, "step": 7125 }, { "epoch": 1.11, "grad_norm": 13.112242110632847, "learning_rate": 1.4496501643656356e-05, "loss": 0.6785, "step": 7126 }, { "epoch": 1.11, "grad_norm": 18.779309510510934, "learning_rate": 1.4494995280144198e-05, "loss": 0.5376, "step": 7127 }, { "epoch": 1.11, "grad_norm": 19.730295629502738, "learning_rate": 1.4493488788797006e-05, "loss": 0.6951, "step": 7128 }, { "epoch": 1.11, "grad_norm": 13.67425831866003, "learning_rate": 1.449198216965762e-05, "loss": 0.679, "step": 7129 }, { "epoch": 1.11, "grad_norm": 43.810620102977005, "learning_rate": 1.449047542276889e-05, "loss": 0.7325, "step": 7130 }, { "epoch": 1.11, "grad_norm": 15.433666699274976, "learning_rate": 1.4488968548173662e-05, "loss": 0.6103, "step": 7131 }, { "epoch": 1.11, "grad_norm": 16.24231067535535, "learning_rate": 1.4487461545914801e-05, "loss": 0.6429, "step": 7132 }, { "epoch": 1.11, "grad_norm": 11.28116405538859, "learning_rate": 1.4485954416035152e-05, "loss": 0.6813, "step": 7133 }, { "epoch": 1.11, "grad_norm": 18.072778834871823, "learning_rate": 1.4484447158577588e-05, "loss": 0.6368, "step": 7134 }, { "epoch": 1.11, "grad_norm": 25.47481399146278, "learning_rate": 1.448293977358497e-05, "loss": 0.6792, "step": 7135 }, { "epoch": 1.11, "grad_norm": 14.532519097950544, "learning_rate": 1.4481432261100168e-05, "loss": 0.6942, "step": 7136 }, { "epoch": 1.11, "grad_norm": 18.272208164300157, "learning_rate": 1.4479924621166051e-05, "loss": 0.6786, "step": 7137 }, { "epoch": 1.11, "grad_norm": 19.632029391016218, "learning_rate": 1.4478416853825502e-05, "loss": 0.6437, "step": 7138 }, { "epoch": 1.12, "grad_norm": 18.045931425652277, "learning_rate": 1.4476908959121394e-05, "loss": 0.7058, "step": 7139 }, { "epoch": 1.12, "grad_norm": 16.9929101364993, "learning_rate": 1.4475400937096617e-05, "loss": 0.628, "step": 7140 }, { "epoch": 1.12, "grad_norm": 19.128778170156263, "learning_rate": 1.4473892787794053e-05, "loss": 0.7265, "step": 7141 }, { "epoch": 1.12, "grad_norm": 21.684913205118107, "learning_rate": 1.4472384511256597e-05, "loss": 0.6601, "step": 7142 }, { "epoch": 1.12, "grad_norm": 21.101327858799248, "learning_rate": 1.4470876107527141e-05, "loss": 0.7012, "step": 7143 }, { "epoch": 1.12, "grad_norm": 13.549378449755283, "learning_rate": 1.4469367576648582e-05, "loss": 0.683, "step": 7144 }, { "epoch": 1.12, "grad_norm": 16.643840399672822, "learning_rate": 1.4467858918663826e-05, "loss": 0.652, "step": 7145 }, { "epoch": 1.12, "grad_norm": 19.099977081906715, "learning_rate": 1.4466350133615776e-05, "loss": 0.6943, "step": 7146 }, { "epoch": 1.12, "grad_norm": 18.526825632591546, "learning_rate": 1.4464841221547339e-05, "loss": 0.6407, "step": 7147 }, { "epoch": 1.12, "grad_norm": 14.440793436888754, "learning_rate": 1.4463332182501431e-05, "loss": 0.6954, "step": 7148 }, { "epoch": 1.12, "grad_norm": 19.649252492030804, "learning_rate": 1.4461823016520967e-05, "loss": 0.7201, "step": 7149 }, { "epoch": 1.12, "grad_norm": 15.52141929028158, "learning_rate": 1.4460313723648866e-05, "loss": 0.6271, "step": 7150 }, { "epoch": 1.12, "grad_norm": 7.022794139236988, "learning_rate": 1.4458804303928053e-05, "loss": 0.6861, "step": 7151 }, { "epoch": 1.12, "grad_norm": 16.31227642761135, "learning_rate": 1.4457294757401452e-05, "loss": 0.6074, "step": 7152 }, { "epoch": 1.12, "grad_norm": 30.454482830572918, "learning_rate": 1.4455785084111997e-05, "loss": 0.6284, "step": 7153 }, { "epoch": 1.12, "grad_norm": 17.638165825402233, "learning_rate": 1.445427528410262e-05, "loss": 0.688, "step": 7154 }, { "epoch": 1.12, "grad_norm": 24.754279296529734, "learning_rate": 1.4452765357416261e-05, "loss": 0.6578, "step": 7155 }, { "epoch": 1.12, "grad_norm": 30.252084088847063, "learning_rate": 1.4451255304095857e-05, "loss": 0.6368, "step": 7156 }, { "epoch": 1.12, "grad_norm": 14.34131684838313, "learning_rate": 1.444974512418436e-05, "loss": 0.6447, "step": 7157 }, { "epoch": 1.12, "grad_norm": 16.375665156126406, "learning_rate": 1.444823481772471e-05, "loss": 0.6622, "step": 7158 }, { "epoch": 1.12, "grad_norm": 25.45230349496516, "learning_rate": 1.4446724384759869e-05, "loss": 0.6687, "step": 7159 }, { "epoch": 1.12, "grad_norm": 21.27005710284545, "learning_rate": 1.4445213825332784e-05, "loss": 0.6446, "step": 7160 }, { "epoch": 1.12, "grad_norm": 15.183637495789984, "learning_rate": 1.444370313948642e-05, "loss": 0.5617, "step": 7161 }, { "epoch": 1.12, "grad_norm": 20.80311784112368, "learning_rate": 1.4442192327263737e-05, "loss": 0.5712, "step": 7162 }, { "epoch": 1.12, "grad_norm": 20.468068875926573, "learning_rate": 1.4440681388707704e-05, "loss": 0.6664, "step": 7163 }, { "epoch": 1.12, "grad_norm": 18.19284891834503, "learning_rate": 1.4439170323861289e-05, "loss": 0.7225, "step": 7164 }, { "epoch": 1.12, "grad_norm": 27.27110475766207, "learning_rate": 1.443765913276747e-05, "loss": 0.7459, "step": 7165 }, { "epoch": 1.12, "grad_norm": 23.593908107935167, "learning_rate": 1.4436147815469217e-05, "loss": 0.6545, "step": 7166 }, { "epoch": 1.12, "grad_norm": 20.296511303974057, "learning_rate": 1.4434636372009516e-05, "loss": 0.7165, "step": 7167 }, { "epoch": 1.12, "grad_norm": 15.056383111186385, "learning_rate": 1.4433124802431353e-05, "loss": 0.6576, "step": 7168 }, { "epoch": 1.12, "grad_norm": 18.743213363914144, "learning_rate": 1.4431613106777712e-05, "loss": 0.5647, "step": 7169 }, { "epoch": 1.12, "grad_norm": 19.870715013580075, "learning_rate": 1.4430101285091587e-05, "loss": 0.6001, "step": 7170 }, { "epoch": 1.12, "grad_norm": 15.314093600279405, "learning_rate": 1.4428589337415975e-05, "loss": 0.6606, "step": 7171 }, { "epoch": 1.12, "grad_norm": 17.44392831472725, "learning_rate": 1.442707726379387e-05, "loss": 0.6873, "step": 7172 }, { "epoch": 1.12, "grad_norm": 15.4437503236935, "learning_rate": 1.4425565064268276e-05, "loss": 0.6569, "step": 7173 }, { "epoch": 1.12, "grad_norm": 22.267867267873793, "learning_rate": 1.4424052738882203e-05, "loss": 0.6765, "step": 7174 }, { "epoch": 1.12, "grad_norm": 15.534561724241765, "learning_rate": 1.442254028767866e-05, "loss": 0.6394, "step": 7175 }, { "epoch": 1.12, "grad_norm": 24.281917991495554, "learning_rate": 1.4421027710700655e-05, "loss": 0.7031, "step": 7176 }, { "epoch": 1.12, "grad_norm": 13.474470189530354, "learning_rate": 1.441951500799121e-05, "loss": 0.7319, "step": 7177 }, { "epoch": 1.12, "grad_norm": 21.35123254428233, "learning_rate": 1.4418002179593345e-05, "loss": 0.64, "step": 7178 }, { "epoch": 1.12, "grad_norm": 20.71189170849506, "learning_rate": 1.441648922555008e-05, "loss": 0.6774, "step": 7179 }, { "epoch": 1.12, "grad_norm": 12.746596146425777, "learning_rate": 1.4414976145904445e-05, "loss": 0.5948, "step": 7180 }, { "epoch": 1.12, "grad_norm": 19.093170712107323, "learning_rate": 1.441346294069947e-05, "loss": 0.6454, "step": 7181 }, { "epoch": 1.12, "grad_norm": 19.990572226482872, "learning_rate": 1.4411949609978196e-05, "loss": 0.6853, "step": 7182 }, { "epoch": 1.12, "grad_norm": 21.34061084299404, "learning_rate": 1.441043615378365e-05, "loss": 0.6357, "step": 7183 }, { "epoch": 1.12, "grad_norm": 14.196777537404106, "learning_rate": 1.4408922572158883e-05, "loss": 0.5732, "step": 7184 }, { "epoch": 1.12, "grad_norm": 15.584729596707398, "learning_rate": 1.4407408865146935e-05, "loss": 0.6176, "step": 7185 }, { "epoch": 1.12, "grad_norm": 13.18308444134311, "learning_rate": 1.4405895032790858e-05, "loss": 0.6675, "step": 7186 }, { "epoch": 1.12, "grad_norm": 15.587270234188065, "learning_rate": 1.4404381075133706e-05, "loss": 0.6121, "step": 7187 }, { "epoch": 1.12, "grad_norm": 20.64533149878886, "learning_rate": 1.4402866992218529e-05, "loss": 0.6496, "step": 7188 }, { "epoch": 1.12, "grad_norm": 30.97312875118572, "learning_rate": 1.4401352784088394e-05, "loss": 0.7026, "step": 7189 }, { "epoch": 1.12, "grad_norm": 22.50748127962067, "learning_rate": 1.4399838450786358e-05, "loss": 0.6849, "step": 7190 }, { "epoch": 1.12, "grad_norm": 16.264482633260354, "learning_rate": 1.439832399235549e-05, "loss": 0.7039, "step": 7191 }, { "epoch": 1.12, "grad_norm": 15.890241899626659, "learning_rate": 1.4396809408838862e-05, "loss": 0.588, "step": 7192 }, { "epoch": 1.12, "grad_norm": 24.062036142584752, "learning_rate": 1.4395294700279545e-05, "loss": 0.5902, "step": 7193 }, { "epoch": 1.12, "grad_norm": 23.765563902378236, "learning_rate": 1.439377986672062e-05, "loss": 0.7383, "step": 7194 }, { "epoch": 1.12, "grad_norm": 13.300892123349808, "learning_rate": 1.4392264908205165e-05, "loss": 0.7094, "step": 7195 }, { "epoch": 1.12, "grad_norm": 15.3762345138238, "learning_rate": 1.4390749824776264e-05, "loss": 0.6302, "step": 7196 }, { "epoch": 1.12, "grad_norm": 17.207671099661752, "learning_rate": 1.438923461647701e-05, "loss": 0.7919, "step": 7197 }, { "epoch": 1.12, "grad_norm": 15.203450093611192, "learning_rate": 1.4387719283350488e-05, "loss": 0.663, "step": 7198 }, { "epoch": 1.12, "grad_norm": 18.132287068378385, "learning_rate": 1.43862038254398e-05, "loss": 0.5977, "step": 7199 }, { "epoch": 1.12, "grad_norm": 15.118131337038932, "learning_rate": 1.4384688242788034e-05, "loss": 0.727, "step": 7200 }, { "epoch": 1.12, "grad_norm": 17.18136887261186, "learning_rate": 1.4383172535438306e-05, "loss": 0.6412, "step": 7201 }, { "epoch": 1.12, "grad_norm": 21.190909346982185, "learning_rate": 1.4381656703433713e-05, "loss": 0.7353, "step": 7202 }, { "epoch": 1.13, "grad_norm": 21.15093742302803, "learning_rate": 1.4380140746817366e-05, "loss": 0.6698, "step": 7203 }, { "epoch": 1.13, "grad_norm": 18.195003159654558, "learning_rate": 1.4378624665632379e-05, "loss": 0.6264, "step": 7204 }, { "epoch": 1.13, "grad_norm": 19.480302149440988, "learning_rate": 1.4377108459921867e-05, "loss": 0.6304, "step": 7205 }, { "epoch": 1.13, "grad_norm": 16.97901000046173, "learning_rate": 1.4375592129728952e-05, "loss": 0.6901, "step": 7206 }, { "epoch": 1.13, "grad_norm": 12.494333391847798, "learning_rate": 1.4374075675096757e-05, "loss": 0.5668, "step": 7207 }, { "epoch": 1.13, "grad_norm": 20.394876210865057, "learning_rate": 1.4372559096068406e-05, "loss": 0.6654, "step": 7208 }, { "epoch": 1.13, "grad_norm": 18.987401928186944, "learning_rate": 1.4371042392687036e-05, "loss": 0.7057, "step": 7209 }, { "epoch": 1.13, "grad_norm": 24.61080116741353, "learning_rate": 1.4369525564995774e-05, "loss": 0.6659, "step": 7210 }, { "epoch": 1.13, "grad_norm": 18.792474335419126, "learning_rate": 1.4368008613037763e-05, "loss": 0.6541, "step": 7211 }, { "epoch": 1.13, "grad_norm": 16.998459234696274, "learning_rate": 1.436649153685614e-05, "loss": 0.6462, "step": 7212 }, { "epoch": 1.13, "grad_norm": 27.13593411714945, "learning_rate": 1.4364974336494055e-05, "loss": 0.6775, "step": 7213 }, { "epoch": 1.13, "grad_norm": 15.118755586011249, "learning_rate": 1.436345701199465e-05, "loss": 0.5968, "step": 7214 }, { "epoch": 1.13, "grad_norm": 21.300974833288958, "learning_rate": 1.4361939563401082e-05, "loss": 0.6587, "step": 7215 }, { "epoch": 1.13, "grad_norm": 22.01647242383209, "learning_rate": 1.4360421990756506e-05, "loss": 0.6577, "step": 7216 }, { "epoch": 1.13, "grad_norm": 11.977488267357753, "learning_rate": 1.4358904294104076e-05, "loss": 0.6365, "step": 7217 }, { "epoch": 1.13, "grad_norm": 16.596867110520396, "learning_rate": 1.4357386473486961e-05, "loss": 0.7022, "step": 7218 }, { "epoch": 1.13, "grad_norm": 26.604504721962428, "learning_rate": 1.435586852894832e-05, "loss": 0.7153, "step": 7219 }, { "epoch": 1.13, "grad_norm": 23.031523037796934, "learning_rate": 1.435435046053133e-05, "loss": 0.7289, "step": 7220 }, { "epoch": 1.13, "grad_norm": 17.795522085783244, "learning_rate": 1.4352832268279158e-05, "loss": 0.7591, "step": 7221 }, { "epoch": 1.13, "grad_norm": 19.248858315808214, "learning_rate": 1.4351313952234985e-05, "loss": 0.684, "step": 7222 }, { "epoch": 1.13, "grad_norm": 14.92443869022873, "learning_rate": 1.4349795512441984e-05, "loss": 0.6102, "step": 7223 }, { "epoch": 1.13, "grad_norm": 15.636637438034517, "learning_rate": 1.434827694894335e-05, "loss": 0.6419, "step": 7224 }, { "epoch": 1.13, "grad_norm": 27.59543016584774, "learning_rate": 1.4346758261782256e-05, "loss": 0.7201, "step": 7225 }, { "epoch": 1.13, "grad_norm": 15.881622650791597, "learning_rate": 1.4345239451001905e-05, "loss": 0.7057, "step": 7226 }, { "epoch": 1.13, "grad_norm": 16.200292531914716, "learning_rate": 1.4343720516645482e-05, "loss": 0.7032, "step": 7227 }, { "epoch": 1.13, "grad_norm": 20.024773365993283, "learning_rate": 1.4342201458756193e-05, "loss": 0.6261, "step": 7228 }, { "epoch": 1.13, "grad_norm": 13.231735378911848, "learning_rate": 1.434068227737723e-05, "loss": 0.6029, "step": 7229 }, { "epoch": 1.13, "grad_norm": 11.76573305409677, "learning_rate": 1.4339162972551806e-05, "loss": 0.5652, "step": 7230 }, { "epoch": 1.13, "grad_norm": 17.957174920035747, "learning_rate": 1.4337643544323124e-05, "loss": 0.6785, "step": 7231 }, { "epoch": 1.13, "grad_norm": 18.60017781909677, "learning_rate": 1.4336123992734396e-05, "loss": 0.6322, "step": 7232 }, { "epoch": 1.13, "grad_norm": 28.529308365836304, "learning_rate": 1.433460431782884e-05, "loss": 0.7221, "step": 7233 }, { "epoch": 1.13, "grad_norm": 19.9133651038106, "learning_rate": 1.4333084519649671e-05, "loss": 0.675, "step": 7234 }, { "epoch": 1.13, "grad_norm": 13.231114544916153, "learning_rate": 1.4331564598240112e-05, "loss": 0.7332, "step": 7235 }, { "epoch": 1.13, "grad_norm": 18.315115823194645, "learning_rate": 1.4330044553643391e-05, "loss": 0.7017, "step": 7236 }, { "epoch": 1.13, "grad_norm": 32.258094438063104, "learning_rate": 1.4328524385902738e-05, "loss": 0.6993, "step": 7237 }, { "epoch": 1.13, "grad_norm": 21.244949787649077, "learning_rate": 1.432700409506138e-05, "loss": 0.632, "step": 7238 }, { "epoch": 1.13, "grad_norm": 19.06964998762685, "learning_rate": 1.4325483681162562e-05, "loss": 0.6341, "step": 7239 }, { "epoch": 1.13, "grad_norm": 17.866102818706402, "learning_rate": 1.4323963144249514e-05, "loss": 0.6796, "step": 7240 }, { "epoch": 1.13, "grad_norm": 15.021833442279746, "learning_rate": 1.4322442484365486e-05, "loss": 0.6807, "step": 7241 }, { "epoch": 1.13, "grad_norm": 17.452541188727096, "learning_rate": 1.432092170155372e-05, "loss": 0.6833, "step": 7242 }, { "epoch": 1.13, "grad_norm": 18.42122978030149, "learning_rate": 1.4319400795857472e-05, "loss": 0.6333, "step": 7243 }, { "epoch": 1.13, "grad_norm": 21.716899455771948, "learning_rate": 1.431787976731999e-05, "loss": 0.6892, "step": 7244 }, { "epoch": 1.13, "grad_norm": 19.069079270938037, "learning_rate": 1.4316358615984532e-05, "loss": 0.6718, "step": 7245 }, { "epoch": 1.13, "grad_norm": 13.104257318416137, "learning_rate": 1.4314837341894361e-05, "loss": 0.7103, "step": 7246 }, { "epoch": 1.13, "grad_norm": 16.881999526249366, "learning_rate": 1.431331594509274e-05, "loss": 0.6932, "step": 7247 }, { "epoch": 1.13, "grad_norm": 16.72241974066135, "learning_rate": 1.4311794425622937e-05, "loss": 0.6597, "step": 7248 }, { "epoch": 1.13, "grad_norm": 13.534884557157698, "learning_rate": 1.4310272783528223e-05, "loss": 0.6009, "step": 7249 }, { "epoch": 1.13, "grad_norm": 25.739369021894525, "learning_rate": 1.4308751018851872e-05, "loss": 0.7261, "step": 7250 }, { "epoch": 1.13, "grad_norm": 13.952255815194379, "learning_rate": 1.4307229131637163e-05, "loss": 0.5514, "step": 7251 }, { "epoch": 1.13, "grad_norm": 26.721540954729356, "learning_rate": 1.4305707121927377e-05, "loss": 0.7217, "step": 7252 }, { "epoch": 1.13, "grad_norm": 14.57838382215206, "learning_rate": 1.43041849897658e-05, "loss": 0.6106, "step": 7253 }, { "epoch": 1.13, "grad_norm": 25.813005878857766, "learning_rate": 1.4302662735195717e-05, "loss": 0.7072, "step": 7254 }, { "epoch": 1.13, "grad_norm": 19.81141440722146, "learning_rate": 1.4301140358260426e-05, "loss": 0.5864, "step": 7255 }, { "epoch": 1.13, "grad_norm": 23.280836446020455, "learning_rate": 1.4299617859003218e-05, "loss": 0.6065, "step": 7256 }, { "epoch": 1.13, "grad_norm": 29.162200611807993, "learning_rate": 1.4298095237467394e-05, "loss": 0.7162, "step": 7257 }, { "epoch": 1.13, "grad_norm": 23.580302761666395, "learning_rate": 1.4296572493696255e-05, "loss": 0.7014, "step": 7258 }, { "epoch": 1.13, "grad_norm": 18.698055707703222, "learning_rate": 1.4295049627733107e-05, "loss": 0.7062, "step": 7259 }, { "epoch": 1.13, "grad_norm": 22.704559497119337, "learning_rate": 1.4293526639621262e-05, "loss": 0.7014, "step": 7260 }, { "epoch": 1.13, "grad_norm": 14.237680704039285, "learning_rate": 1.4292003529404028e-05, "loss": 0.6506, "step": 7261 }, { "epoch": 1.13, "grad_norm": 20.000050231751302, "learning_rate": 1.4290480297124726e-05, "loss": 0.7142, "step": 7262 }, { "epoch": 1.13, "grad_norm": 15.553125203381537, "learning_rate": 1.4288956942826674e-05, "loss": 0.6221, "step": 7263 }, { "epoch": 1.13, "grad_norm": 15.705421841070356, "learning_rate": 1.4287433466553198e-05, "loss": 0.6185, "step": 7264 }, { "epoch": 1.13, "grad_norm": 16.954092152866856, "learning_rate": 1.4285909868347621e-05, "loss": 0.6653, "step": 7265 }, { "epoch": 1.13, "grad_norm": 25.50243836086091, "learning_rate": 1.4284386148253273e-05, "loss": 0.6396, "step": 7266 }, { "epoch": 1.14, "grad_norm": 19.404806212376876, "learning_rate": 1.4282862306313488e-05, "loss": 0.6636, "step": 7267 }, { "epoch": 1.14, "grad_norm": 26.71300850298545, "learning_rate": 1.4281338342571609e-05, "loss": 0.7065, "step": 7268 }, { "epoch": 1.14, "grad_norm": 25.39621370050001, "learning_rate": 1.4279814257070967e-05, "loss": 0.6672, "step": 7269 }, { "epoch": 1.14, "grad_norm": 26.62580982619853, "learning_rate": 1.4278290049854917e-05, "loss": 0.7053, "step": 7270 }, { "epoch": 1.14, "grad_norm": 18.372850870615956, "learning_rate": 1.4276765720966797e-05, "loss": 0.7387, "step": 7271 }, { "epoch": 1.14, "grad_norm": 26.858369018973544, "learning_rate": 1.4275241270449962e-05, "loss": 0.6253, "step": 7272 }, { "epoch": 1.14, "grad_norm": 16.310326414065248, "learning_rate": 1.4273716698347766e-05, "loss": 0.6862, "step": 7273 }, { "epoch": 1.14, "grad_norm": 14.1063625762404, "learning_rate": 1.4272192004703569e-05, "loss": 0.5218, "step": 7274 }, { "epoch": 1.14, "grad_norm": 13.870113432159268, "learning_rate": 1.4270667189560727e-05, "loss": 0.5224, "step": 7275 }, { "epoch": 1.14, "grad_norm": 19.723494026688144, "learning_rate": 1.426914225296261e-05, "loss": 0.6658, "step": 7276 }, { "epoch": 1.14, "grad_norm": 17.30637815410105, "learning_rate": 1.4267617194952588e-05, "loss": 0.6535, "step": 7277 }, { "epoch": 1.14, "grad_norm": 24.63036808573288, "learning_rate": 1.4266092015574027e-05, "loss": 0.707, "step": 7278 }, { "epoch": 1.14, "grad_norm": 17.63144918739665, "learning_rate": 1.4264566714870301e-05, "loss": 0.6415, "step": 7279 }, { "epoch": 1.14, "grad_norm": 27.568212890235827, "learning_rate": 1.4263041292884795e-05, "loss": 0.643, "step": 7280 }, { "epoch": 1.14, "grad_norm": 27.427697458550004, "learning_rate": 1.426151574966089e-05, "loss": 0.7036, "step": 7281 }, { "epoch": 1.14, "grad_norm": 23.635511133263815, "learning_rate": 1.425999008524197e-05, "loss": 0.6852, "step": 7282 }, { "epoch": 1.14, "grad_norm": 14.44841963596834, "learning_rate": 1.4258464299671425e-05, "loss": 0.6534, "step": 7283 }, { "epoch": 1.14, "grad_norm": 23.55068954080406, "learning_rate": 1.4256938392992641e-05, "loss": 0.6783, "step": 7284 }, { "epoch": 1.14, "grad_norm": 22.273292658109778, "learning_rate": 1.4255412365249027e-05, "loss": 0.6484, "step": 7285 }, { "epoch": 1.14, "grad_norm": 21.905888167398263, "learning_rate": 1.4253886216483968e-05, "loss": 0.7714, "step": 7286 }, { "epoch": 1.14, "grad_norm": 41.88415731093706, "learning_rate": 1.4252359946740877e-05, "loss": 0.7661, "step": 7287 }, { "epoch": 1.14, "grad_norm": 17.98892526540278, "learning_rate": 1.4250833556063155e-05, "loss": 0.6841, "step": 7288 }, { "epoch": 1.14, "grad_norm": 17.74233157436609, "learning_rate": 1.4249307044494214e-05, "loss": 0.7232, "step": 7289 }, { "epoch": 1.14, "grad_norm": 28.876401321252764, "learning_rate": 1.4247780412077466e-05, "loss": 0.8119, "step": 7290 }, { "epoch": 1.14, "grad_norm": 19.898143076431495, "learning_rate": 1.4246253658856329e-05, "loss": 0.669, "step": 7291 }, { "epoch": 1.14, "grad_norm": 19.584322122119293, "learning_rate": 1.4244726784874221e-05, "loss": 0.609, "step": 7292 }, { "epoch": 1.14, "grad_norm": 17.693950548453664, "learning_rate": 1.4243199790174566e-05, "loss": 0.742, "step": 7293 }, { "epoch": 1.14, "grad_norm": 17.46314194949273, "learning_rate": 1.4241672674800791e-05, "loss": 0.6365, "step": 7294 }, { "epoch": 1.14, "grad_norm": 19.011452578360043, "learning_rate": 1.4240145438796329e-05, "loss": 0.8068, "step": 7295 }, { "epoch": 1.14, "grad_norm": 15.842204373113004, "learning_rate": 1.4238618082204609e-05, "loss": 0.7043, "step": 7296 }, { "epoch": 1.14, "grad_norm": 20.18058377223884, "learning_rate": 1.4237090605069072e-05, "loss": 0.6929, "step": 7297 }, { "epoch": 1.14, "grad_norm": 13.439286518547199, "learning_rate": 1.4235563007433153e-05, "loss": 0.6226, "step": 7298 }, { "epoch": 1.14, "grad_norm": 19.97953463894396, "learning_rate": 1.4234035289340304e-05, "loss": 0.5827, "step": 7299 }, { "epoch": 1.14, "grad_norm": 25.155805540601076, "learning_rate": 1.4232507450833966e-05, "loss": 0.6884, "step": 7300 }, { "epoch": 1.14, "grad_norm": 15.895742233843588, "learning_rate": 1.4230979491957592e-05, "loss": 0.7232, "step": 7301 }, { "epoch": 1.14, "grad_norm": 22.15858242606363, "learning_rate": 1.4229451412754638e-05, "loss": 0.6397, "step": 7302 }, { "epoch": 1.14, "grad_norm": 24.3522504691592, "learning_rate": 1.4227923213268557e-05, "loss": 0.7252, "step": 7303 }, { "epoch": 1.14, "grad_norm": 40.257306912484495, "learning_rate": 1.4226394893542817e-05, "loss": 0.7522, "step": 7304 }, { "epoch": 1.14, "grad_norm": 19.102134458095758, "learning_rate": 1.4224866453620875e-05, "loss": 0.6615, "step": 7305 }, { "epoch": 1.14, "grad_norm": 18.660909969686745, "learning_rate": 1.4223337893546206e-05, "loss": 0.7609, "step": 7306 }, { "epoch": 1.14, "grad_norm": 21.064511993637204, "learning_rate": 1.4221809213362275e-05, "loss": 0.7111, "step": 7307 }, { "epoch": 1.14, "grad_norm": 17.592858773860158, "learning_rate": 1.422028041311256e-05, "loss": 0.6496, "step": 7308 }, { "epoch": 1.14, "grad_norm": 17.298560536797847, "learning_rate": 1.421875149284054e-05, "loss": 0.7658, "step": 7309 }, { "epoch": 1.14, "grad_norm": 17.27792632629873, "learning_rate": 1.4217222452589694e-05, "loss": 0.6584, "step": 7310 }, { "epoch": 1.14, "grad_norm": 29.62450832952857, "learning_rate": 1.4215693292403508e-05, "loss": 0.6047, "step": 7311 }, { "epoch": 1.14, "grad_norm": 18.945433246974968, "learning_rate": 1.4214164012325475e-05, "loss": 0.6657, "step": 7312 }, { "epoch": 1.14, "grad_norm": 17.511089778686312, "learning_rate": 1.421263461239908e-05, "loss": 0.6407, "step": 7313 }, { "epoch": 1.14, "grad_norm": 15.651154781658498, "learning_rate": 1.421110509266782e-05, "loss": 0.5592, "step": 7314 }, { "epoch": 1.14, "grad_norm": 14.131092057490608, "learning_rate": 1.4209575453175195e-05, "loss": 0.5954, "step": 7315 }, { "epoch": 1.14, "grad_norm": 16.539222297112726, "learning_rate": 1.4208045693964707e-05, "loss": 0.5758, "step": 7316 }, { "epoch": 1.14, "grad_norm": 21.98470227391721, "learning_rate": 1.4206515815079862e-05, "loss": 0.7989, "step": 7317 }, { "epoch": 1.14, "grad_norm": 21.938613006826344, "learning_rate": 1.4204985816564167e-05, "loss": 0.7022, "step": 7318 }, { "epoch": 1.14, "grad_norm": 16.651870943641747, "learning_rate": 1.4203455698461135e-05, "loss": 0.6421, "step": 7319 }, { "epoch": 1.14, "grad_norm": 27.187722911140124, "learning_rate": 1.4201925460814282e-05, "loss": 0.7116, "step": 7320 }, { "epoch": 1.14, "grad_norm": 24.278937345022634, "learning_rate": 1.4200395103667126e-05, "loss": 0.6897, "step": 7321 }, { "epoch": 1.14, "grad_norm": 19.002033966297134, "learning_rate": 1.4198864627063194e-05, "loss": 0.7879, "step": 7322 }, { "epoch": 1.14, "grad_norm": 11.94732913452051, "learning_rate": 1.4197334031046004e-05, "loss": 0.5438, "step": 7323 }, { "epoch": 1.14, "grad_norm": 17.17885160212382, "learning_rate": 1.4195803315659092e-05, "loss": 0.6195, "step": 7324 }, { "epoch": 1.14, "grad_norm": 24.56809037171448, "learning_rate": 1.4194272480945987e-05, "loss": 0.6291, "step": 7325 }, { "epoch": 1.14, "grad_norm": 18.22303539589026, "learning_rate": 1.4192741526950226e-05, "loss": 0.6927, "step": 7326 }, { "epoch": 1.14, "grad_norm": 18.75937794270382, "learning_rate": 1.4191210453715351e-05, "loss": 0.6513, "step": 7327 }, { "epoch": 1.14, "grad_norm": 20.299243211425, "learning_rate": 1.4189679261284899e-05, "loss": 0.699, "step": 7328 }, { "epoch": 1.14, "grad_norm": 16.46224169867028, "learning_rate": 1.4188147949702425e-05, "loss": 0.6191, "step": 7329 }, { "epoch": 1.14, "grad_norm": 19.030742795884645, "learning_rate": 1.4186616519011467e-05, "loss": 0.7157, "step": 7330 }, { "epoch": 1.15, "grad_norm": 17.007342884044387, "learning_rate": 1.4185084969255589e-05, "loss": 0.6542, "step": 7331 }, { "epoch": 1.15, "grad_norm": 17.038201794405612, "learning_rate": 1.4183553300478339e-05, "loss": 0.643, "step": 7332 }, { "epoch": 1.15, "grad_norm": 14.23321810154991, "learning_rate": 1.4182021512723283e-05, "loss": 0.6499, "step": 7333 }, { "epoch": 1.15, "grad_norm": 18.29804193438696, "learning_rate": 1.418048960603398e-05, "loss": 0.7239, "step": 7334 }, { "epoch": 1.15, "grad_norm": 18.06691928320673, "learning_rate": 1.4178957580454e-05, "loss": 0.7259, "step": 7335 }, { "epoch": 1.15, "grad_norm": 19.185520240060747, "learning_rate": 1.417742543602691e-05, "loss": 0.6354, "step": 7336 }, { "epoch": 1.15, "grad_norm": 18.189090704259083, "learning_rate": 1.4175893172796285e-05, "loss": 0.5948, "step": 7337 }, { "epoch": 1.15, "grad_norm": 12.50046685807299, "learning_rate": 1.4174360790805699e-05, "loss": 0.7353, "step": 7338 }, { "epoch": 1.15, "grad_norm": 18.518134735725692, "learning_rate": 1.417282829009874e-05, "loss": 0.5918, "step": 7339 }, { "epoch": 1.15, "grad_norm": 17.761593482433817, "learning_rate": 1.4171295670718976e-05, "loss": 0.6593, "step": 7340 }, { "epoch": 1.15, "grad_norm": 14.163263791284107, "learning_rate": 1.4169762932710012e-05, "loss": 0.6502, "step": 7341 }, { "epoch": 1.15, "grad_norm": 14.779312368198045, "learning_rate": 1.4168230076115424e-05, "loss": 0.5405, "step": 7342 }, { "epoch": 1.15, "grad_norm": 16.719819136330365, "learning_rate": 1.4166697100978812e-05, "loss": 0.6644, "step": 7343 }, { "epoch": 1.15, "grad_norm": 18.919184132740725, "learning_rate": 1.4165164007343772e-05, "loss": 0.634, "step": 7344 }, { "epoch": 1.15, "grad_norm": 16.712967475570874, "learning_rate": 1.4163630795253904e-05, "loss": 0.688, "step": 7345 }, { "epoch": 1.15, "grad_norm": 16.3363748803825, "learning_rate": 1.4162097464752815e-05, "loss": 0.6943, "step": 7346 }, { "epoch": 1.15, "grad_norm": 13.560537049527449, "learning_rate": 1.4160564015884103e-05, "loss": 0.5626, "step": 7347 }, { "epoch": 1.15, "grad_norm": 15.223194824677954, "learning_rate": 1.415903044869139e-05, "loss": 0.6359, "step": 7348 }, { "epoch": 1.15, "grad_norm": 11.69957061072878, "learning_rate": 1.4157496763218278e-05, "loss": 0.6065, "step": 7349 }, { "epoch": 1.15, "grad_norm": 20.588844511724094, "learning_rate": 1.4155962959508396e-05, "loss": 0.7005, "step": 7350 }, { "epoch": 1.15, "grad_norm": 20.55260639971917, "learning_rate": 1.4154429037605358e-05, "loss": 0.7367, "step": 7351 }, { "epoch": 1.15, "grad_norm": 15.867184974711531, "learning_rate": 1.4152894997552786e-05, "loss": 0.636, "step": 7352 }, { "epoch": 1.15, "grad_norm": 21.56582770342848, "learning_rate": 1.415136083939431e-05, "loss": 0.7422, "step": 7353 }, { "epoch": 1.15, "grad_norm": 18.27299115539783, "learning_rate": 1.414982656317356e-05, "loss": 0.7147, "step": 7354 }, { "epoch": 1.15, "grad_norm": 16.87205477638366, "learning_rate": 1.414829216893417e-05, "loss": 0.6581, "step": 7355 }, { "epoch": 1.15, "grad_norm": 25.777476602169482, "learning_rate": 1.4146757656719781e-05, "loss": 0.622, "step": 7356 }, { "epoch": 1.15, "grad_norm": 20.911269385805255, "learning_rate": 1.4145223026574027e-05, "loss": 0.6752, "step": 7357 }, { "epoch": 1.15, "grad_norm": 26.591560678767447, "learning_rate": 1.4143688278540554e-05, "loss": 0.6516, "step": 7358 }, { "epoch": 1.15, "grad_norm": 27.284983232587212, "learning_rate": 1.4142153412663012e-05, "loss": 0.7673, "step": 7359 }, { "epoch": 1.15, "grad_norm": 30.947000097337614, "learning_rate": 1.414061842898505e-05, "loss": 0.6534, "step": 7360 }, { "epoch": 1.15, "grad_norm": 16.162060374390233, "learning_rate": 1.413908332755032e-05, "loss": 0.7408, "step": 7361 }, { "epoch": 1.15, "grad_norm": 18.071875773411218, "learning_rate": 1.4137548108402483e-05, "loss": 0.607, "step": 7362 }, { "epoch": 1.15, "grad_norm": 16.91990479229881, "learning_rate": 1.41360127715852e-05, "loss": 0.6592, "step": 7363 }, { "epoch": 1.15, "grad_norm": 13.913176749864778, "learning_rate": 1.4134477317142133e-05, "loss": 0.7209, "step": 7364 }, { "epoch": 1.15, "grad_norm": 22.581044234735447, "learning_rate": 1.4132941745116946e-05, "loss": 0.6888, "step": 7365 }, { "epoch": 1.15, "grad_norm": 22.878128883946122, "learning_rate": 1.4131406055553316e-05, "loss": 0.7104, "step": 7366 }, { "epoch": 1.15, "grad_norm": 15.10784189075606, "learning_rate": 1.4129870248494913e-05, "loss": 0.6707, "step": 7367 }, { "epoch": 1.15, "grad_norm": 24.9504778036491, "learning_rate": 1.4128334323985418e-05, "loss": 0.7455, "step": 7368 }, { "epoch": 1.15, "grad_norm": 12.608233045903027, "learning_rate": 1.412679828206851e-05, "loss": 0.6429, "step": 7369 }, { "epoch": 1.15, "grad_norm": 34.300653763023575, "learning_rate": 1.4125262122787872e-05, "loss": 0.621, "step": 7370 }, { "epoch": 1.15, "grad_norm": 19.980337563510073, "learning_rate": 1.4123725846187193e-05, "loss": 0.6928, "step": 7371 }, { "epoch": 1.15, "grad_norm": 13.370839955994054, "learning_rate": 1.412218945231016e-05, "loss": 0.6911, "step": 7372 }, { "epoch": 1.15, "grad_norm": 17.750375623657416, "learning_rate": 1.4120652941200477e-05, "loss": 0.6599, "step": 7373 }, { "epoch": 1.15, "grad_norm": 17.144720114282386, "learning_rate": 1.4119116312901828e-05, "loss": 0.6486, "step": 7374 }, { "epoch": 1.15, "grad_norm": 15.228050529146017, "learning_rate": 1.4117579567457927e-05, "loss": 0.5787, "step": 7375 }, { "epoch": 1.15, "grad_norm": 16.31204141916774, "learning_rate": 1.4116042704912465e-05, "loss": 0.7116, "step": 7376 }, { "epoch": 1.15, "grad_norm": 19.973238864326664, "learning_rate": 1.411450572530916e-05, "loss": 0.686, "step": 7377 }, { "epoch": 1.15, "grad_norm": 13.533872217040669, "learning_rate": 1.411296862869172e-05, "loss": 0.6718, "step": 7378 }, { "epoch": 1.15, "grad_norm": 24.361131588851002, "learning_rate": 1.4111431415103858e-05, "loss": 0.716, "step": 7379 }, { "epoch": 1.15, "grad_norm": 21.509238595237658, "learning_rate": 1.4109894084589291e-05, "loss": 0.6428, "step": 7380 }, { "epoch": 1.15, "grad_norm": 15.586743325361361, "learning_rate": 1.410835663719174e-05, "loss": 0.6645, "step": 7381 }, { "epoch": 1.15, "grad_norm": 21.186221632749692, "learning_rate": 1.410681907295493e-05, "loss": 0.6604, "step": 7382 }, { "epoch": 1.15, "grad_norm": 15.939096133396044, "learning_rate": 1.410528139192259e-05, "loss": 0.7495, "step": 7383 }, { "epoch": 1.15, "grad_norm": 22.933165789733653, "learning_rate": 1.4103743594138443e-05, "loss": 0.6805, "step": 7384 }, { "epoch": 1.15, "grad_norm": 20.45622464524285, "learning_rate": 1.4102205679646236e-05, "loss": 0.7013, "step": 7385 }, { "epoch": 1.15, "grad_norm": 15.055066577056435, "learning_rate": 1.4100667648489692e-05, "loss": 0.6227, "step": 7386 }, { "epoch": 1.15, "grad_norm": 17.882454329212287, "learning_rate": 1.4099129500712562e-05, "loss": 0.5995, "step": 7387 }, { "epoch": 1.15, "grad_norm": 24.79465860119624, "learning_rate": 1.4097591236358588e-05, "loss": 0.7247, "step": 7388 }, { "epoch": 1.15, "grad_norm": 15.272643075951262, "learning_rate": 1.4096052855471519e-05, "loss": 0.6833, "step": 7389 }, { "epoch": 1.15, "grad_norm": 17.842453650218186, "learning_rate": 1.4094514358095096e-05, "loss": 0.6822, "step": 7390 }, { "epoch": 1.15, "grad_norm": 17.675001128036836, "learning_rate": 1.4092975744273082e-05, "loss": 0.5861, "step": 7391 }, { "epoch": 1.15, "grad_norm": 20.759574079585438, "learning_rate": 1.4091437014049234e-05, "loss": 0.6755, "step": 7392 }, { "epoch": 1.15, "grad_norm": 14.376489050078144, "learning_rate": 1.408989816746731e-05, "loss": 0.7123, "step": 7393 }, { "epoch": 1.15, "grad_norm": 24.247837955897612, "learning_rate": 1.4088359204571076e-05, "loss": 0.7154, "step": 7394 }, { "epoch": 1.16, "grad_norm": 15.001454248068562, "learning_rate": 1.4086820125404299e-05, "loss": 0.6979, "step": 7395 }, { "epoch": 1.16, "grad_norm": 23.785621788868074, "learning_rate": 1.4085280930010745e-05, "loss": 0.6524, "step": 7396 }, { "epoch": 1.16, "grad_norm": 15.620393958684955, "learning_rate": 1.4083741618434192e-05, "loss": 0.6626, "step": 7397 }, { "epoch": 1.16, "grad_norm": 24.826052385045198, "learning_rate": 1.4082202190718417e-05, "loss": 0.6586, "step": 7398 }, { "epoch": 1.16, "grad_norm": 22.273155315649465, "learning_rate": 1.40806626469072e-05, "loss": 0.6882, "step": 7399 }, { "epoch": 1.16, "grad_norm": 19.282204565720075, "learning_rate": 1.4079122987044324e-05, "loss": 0.6899, "step": 7400 }, { "epoch": 1.16, "grad_norm": 15.549876274179365, "learning_rate": 1.4077583211173575e-05, "loss": 0.717, "step": 7401 }, { "epoch": 1.16, "grad_norm": 18.33697810834557, "learning_rate": 1.4076043319338748e-05, "loss": 0.5725, "step": 7402 }, { "epoch": 1.16, "grad_norm": 16.315063533646104, "learning_rate": 1.407450331158363e-05, "loss": 0.7549, "step": 7403 }, { "epoch": 1.16, "grad_norm": 14.744429739733686, "learning_rate": 1.4072963187952023e-05, "loss": 0.5895, "step": 7404 }, { "epoch": 1.16, "grad_norm": 14.463991743229743, "learning_rate": 1.4071422948487725e-05, "loss": 0.6726, "step": 7405 }, { "epoch": 1.16, "grad_norm": 15.788281130249198, "learning_rate": 1.406988259323454e-05, "loss": 0.6566, "step": 7406 }, { "epoch": 1.16, "grad_norm": 23.20010528951628, "learning_rate": 1.4068342122236275e-05, "loss": 0.7284, "step": 7407 }, { "epoch": 1.16, "grad_norm": 28.13856353601223, "learning_rate": 1.406680153553674e-05, "loss": 0.7362, "step": 7408 }, { "epoch": 1.16, "grad_norm": 13.419744186157724, "learning_rate": 1.4065260833179748e-05, "loss": 0.6689, "step": 7409 }, { "epoch": 1.16, "grad_norm": 22.797216267413933, "learning_rate": 1.4063720015209117e-05, "loss": 0.6516, "step": 7410 }, { "epoch": 1.16, "grad_norm": 20.152084730649868, "learning_rate": 1.4062179081668665e-05, "loss": 0.6049, "step": 7411 }, { "epoch": 1.16, "grad_norm": 18.964435438638063, "learning_rate": 1.4060638032602218e-05, "loss": 0.6738, "step": 7412 }, { "epoch": 1.16, "grad_norm": 19.80899551607031, "learning_rate": 1.4059096868053601e-05, "loss": 0.6962, "step": 7413 }, { "epoch": 1.16, "grad_norm": 18.636531243613337, "learning_rate": 1.405755558806664e-05, "loss": 0.7195, "step": 7414 }, { "epoch": 1.16, "grad_norm": 17.075681713180828, "learning_rate": 1.4056014192685175e-05, "loss": 0.6948, "step": 7415 }, { "epoch": 1.16, "grad_norm": 16.800611271161323, "learning_rate": 1.4054472681953035e-05, "loss": 0.6461, "step": 7416 }, { "epoch": 1.16, "grad_norm": 18.98961262095591, "learning_rate": 1.405293105591407e-05, "loss": 0.6318, "step": 7417 }, { "epoch": 1.16, "grad_norm": 20.766827578629822, "learning_rate": 1.4051389314612112e-05, "loss": 0.6247, "step": 7418 }, { "epoch": 1.16, "grad_norm": 16.22760036980565, "learning_rate": 1.4049847458091014e-05, "loss": 0.6329, "step": 7419 }, { "epoch": 1.16, "grad_norm": 15.506367963971051, "learning_rate": 1.404830548639462e-05, "loss": 0.6172, "step": 7420 }, { "epoch": 1.16, "grad_norm": 19.971184970853464, "learning_rate": 1.404676339956679e-05, "loss": 0.686, "step": 7421 }, { "epoch": 1.16, "grad_norm": 18.00354369947476, "learning_rate": 1.4045221197651375e-05, "loss": 0.7335, "step": 7422 }, { "epoch": 1.16, "grad_norm": 13.898873788162456, "learning_rate": 1.4043678880692235e-05, "loss": 0.5801, "step": 7423 }, { "epoch": 1.16, "grad_norm": 23.175083200681893, "learning_rate": 1.4042136448733235e-05, "loss": 0.6682, "step": 7424 }, { "epoch": 1.16, "grad_norm": 14.586331558829869, "learning_rate": 1.4040593901818239e-05, "loss": 0.6266, "step": 7425 }, { "epoch": 1.16, "grad_norm": 17.76058851999795, "learning_rate": 1.403905123999111e-05, "loss": 0.6591, "step": 7426 }, { "epoch": 1.16, "grad_norm": 15.896263202572664, "learning_rate": 1.4037508463295735e-05, "loss": 0.5713, "step": 7427 }, { "epoch": 1.16, "grad_norm": 19.76547623133555, "learning_rate": 1.4035965571775976e-05, "loss": 0.6287, "step": 7428 }, { "epoch": 1.16, "grad_norm": 15.94958560604843, "learning_rate": 1.403442256547572e-05, "loss": 0.6983, "step": 7429 }, { "epoch": 1.16, "grad_norm": 15.673931826429664, "learning_rate": 1.4032879444438843e-05, "loss": 0.6669, "step": 7430 }, { "epoch": 1.16, "grad_norm": 23.755173992948183, "learning_rate": 1.4031336208709236e-05, "loss": 0.6923, "step": 7431 }, { "epoch": 1.16, "grad_norm": 16.853443592981684, "learning_rate": 1.4029792858330783e-05, "loss": 0.7087, "step": 7432 }, { "epoch": 1.16, "grad_norm": 18.04611829145561, "learning_rate": 1.4028249393347382e-05, "loss": 0.605, "step": 7433 }, { "epoch": 1.16, "grad_norm": 14.575987522186704, "learning_rate": 1.4026705813802923e-05, "loss": 0.6012, "step": 7434 }, { "epoch": 1.16, "grad_norm": 18.225078286878, "learning_rate": 1.4025162119741304e-05, "loss": 0.5948, "step": 7435 }, { "epoch": 1.16, "grad_norm": 20.87707332705314, "learning_rate": 1.4023618311206432e-05, "loss": 0.7036, "step": 7436 }, { "epoch": 1.16, "grad_norm": 25.970216250655813, "learning_rate": 1.4022074388242208e-05, "loss": 0.7093, "step": 7437 }, { "epoch": 1.16, "grad_norm": 20.026213796162736, "learning_rate": 1.4020530350892542e-05, "loss": 0.7042, "step": 7438 }, { "epoch": 1.16, "grad_norm": 19.980603976272196, "learning_rate": 1.4018986199201345e-05, "loss": 0.761, "step": 7439 }, { "epoch": 1.16, "grad_norm": 24.443586857918696, "learning_rate": 1.4017441933212532e-05, "loss": 0.6455, "step": 7440 }, { "epoch": 1.16, "grad_norm": 18.55953781980156, "learning_rate": 1.4015897552970017e-05, "loss": 0.6394, "step": 7441 }, { "epoch": 1.16, "grad_norm": 23.26280079205925, "learning_rate": 1.401435305851773e-05, "loss": 0.6492, "step": 7442 }, { "epoch": 1.16, "grad_norm": 15.57450492333809, "learning_rate": 1.4012808449899584e-05, "loss": 0.6636, "step": 7443 }, { "epoch": 1.16, "grad_norm": 25.879526449869036, "learning_rate": 1.4011263727159521e-05, "loss": 0.6698, "step": 7444 }, { "epoch": 1.16, "grad_norm": 14.09169694808246, "learning_rate": 1.4009718890341457e-05, "loss": 0.6293, "step": 7445 }, { "epoch": 1.16, "grad_norm": 28.297327064344987, "learning_rate": 1.4008173939489338e-05, "loss": 0.6776, "step": 7446 }, { "epoch": 1.16, "grad_norm": 18.47554753763949, "learning_rate": 1.4006628874647094e-05, "loss": 0.6542, "step": 7447 }, { "epoch": 1.16, "grad_norm": 15.084212500370837, "learning_rate": 1.400508369585867e-05, "loss": 0.6238, "step": 7448 }, { "epoch": 1.16, "grad_norm": 21.652779760758925, "learning_rate": 1.400353840316801e-05, "loss": 0.7362, "step": 7449 }, { "epoch": 1.16, "grad_norm": 19.954309514971452, "learning_rate": 1.4001992996619056e-05, "loss": 0.6846, "step": 7450 }, { "epoch": 1.16, "grad_norm": 19.143083624679395, "learning_rate": 1.4000447476255765e-05, "loss": 0.7219, "step": 7451 }, { "epoch": 1.16, "grad_norm": 17.73512113971914, "learning_rate": 1.3998901842122088e-05, "loss": 0.6412, "step": 7452 }, { "epoch": 1.16, "grad_norm": 14.6357970465466, "learning_rate": 1.3997356094261977e-05, "loss": 0.6264, "step": 7453 }, { "epoch": 1.16, "grad_norm": 23.022154026352275, "learning_rate": 1.3995810232719405e-05, "loss": 0.6172, "step": 7454 }, { "epoch": 1.16, "grad_norm": 20.373682419762957, "learning_rate": 1.3994264257538324e-05, "loss": 0.696, "step": 7455 }, { "epoch": 1.16, "grad_norm": 22.068361951330044, "learning_rate": 1.3992718168762702e-05, "loss": 0.5801, "step": 7456 }, { "epoch": 1.16, "grad_norm": 18.87333762075109, "learning_rate": 1.3991171966436513e-05, "loss": 0.7128, "step": 7457 }, { "epoch": 1.16, "grad_norm": 11.790901233762101, "learning_rate": 1.3989625650603729e-05, "loss": 0.5648, "step": 7458 }, { "epoch": 1.17, "grad_norm": 20.56375106814535, "learning_rate": 1.3988079221308323e-05, "loss": 0.619, "step": 7459 }, { "epoch": 1.17, "grad_norm": 18.73698580098692, "learning_rate": 1.3986532678594277e-05, "loss": 0.6288, "step": 7460 }, { "epoch": 1.17, "grad_norm": 21.035665876046952, "learning_rate": 1.3984986022505578e-05, "loss": 0.7057, "step": 7461 }, { "epoch": 1.17, "grad_norm": 17.590239793799036, "learning_rate": 1.3983439253086201e-05, "loss": 0.6575, "step": 7462 }, { "epoch": 1.17, "grad_norm": 15.885276763696012, "learning_rate": 1.3981892370380146e-05, "loss": 0.6823, "step": 7463 }, { "epoch": 1.17, "grad_norm": 22.377524375366338, "learning_rate": 1.3980345374431401e-05, "loss": 0.7452, "step": 7464 }, { "epoch": 1.17, "grad_norm": 18.54329056116013, "learning_rate": 1.3978798265283962e-05, "loss": 0.6429, "step": 7465 }, { "epoch": 1.17, "grad_norm": 25.903684229042685, "learning_rate": 1.3977251042981828e-05, "loss": 0.7628, "step": 7466 }, { "epoch": 1.17, "grad_norm": 12.999764541180905, "learning_rate": 1.3975703707569e-05, "loss": 0.6401, "step": 7467 }, { "epoch": 1.17, "grad_norm": 21.223170027261705, "learning_rate": 1.3974156259089486e-05, "loss": 0.7002, "step": 7468 }, { "epoch": 1.17, "grad_norm": 13.750197318135148, "learning_rate": 1.3972608697587292e-05, "loss": 0.6167, "step": 7469 }, { "epoch": 1.17, "grad_norm": 18.489287971411944, "learning_rate": 1.3971061023106428e-05, "loss": 0.6141, "step": 7470 }, { "epoch": 1.17, "grad_norm": 19.059324109406212, "learning_rate": 1.3969513235690915e-05, "loss": 0.6861, "step": 7471 }, { "epoch": 1.17, "grad_norm": 23.256802116335816, "learning_rate": 1.3967965335384766e-05, "loss": 0.6624, "step": 7472 }, { "epoch": 1.17, "grad_norm": 15.706032848147292, "learning_rate": 1.3966417322232006e-05, "loss": 0.6457, "step": 7473 }, { "epoch": 1.17, "grad_norm": 17.277762907093944, "learning_rate": 1.3964869196276654e-05, "loss": 0.659, "step": 7474 }, { "epoch": 1.17, "grad_norm": 20.595718087145244, "learning_rate": 1.3963320957562743e-05, "loss": 0.7015, "step": 7475 }, { "epoch": 1.17, "grad_norm": 20.606199690884935, "learning_rate": 1.3961772606134303e-05, "loss": 0.6491, "step": 7476 }, { "epoch": 1.17, "grad_norm": 17.200989268409902, "learning_rate": 1.3960224142035368e-05, "loss": 0.6213, "step": 7477 }, { "epoch": 1.17, "grad_norm": 21.51283058533687, "learning_rate": 1.3958675565309974e-05, "loss": 0.6469, "step": 7478 }, { "epoch": 1.17, "grad_norm": 16.902464307051382, "learning_rate": 1.3957126876002164e-05, "loss": 0.7062, "step": 7479 }, { "epoch": 1.17, "grad_norm": 20.246684168649182, "learning_rate": 1.3955578074155978e-05, "loss": 0.6201, "step": 7480 }, { "epoch": 1.17, "grad_norm": 14.684449043338914, "learning_rate": 1.3954029159815468e-05, "loss": 0.5924, "step": 7481 }, { "epoch": 1.17, "grad_norm": 12.218361384951006, "learning_rate": 1.3952480133024682e-05, "loss": 0.6665, "step": 7482 }, { "epoch": 1.17, "grad_norm": 23.73341601683904, "learning_rate": 1.3950930993827671e-05, "loss": 0.6646, "step": 7483 }, { "epoch": 1.17, "grad_norm": 12.65514275334491, "learning_rate": 1.3949381742268496e-05, "loss": 0.6029, "step": 7484 }, { "epoch": 1.17, "grad_norm": 15.227009257813034, "learning_rate": 1.394783237839121e-05, "loss": 0.5666, "step": 7485 }, { "epoch": 1.17, "grad_norm": 23.874543880928986, "learning_rate": 1.3946282902239886e-05, "loss": 0.6344, "step": 7486 }, { "epoch": 1.17, "grad_norm": 18.48261663396154, "learning_rate": 1.3944733313858583e-05, "loss": 0.7229, "step": 7487 }, { "epoch": 1.17, "grad_norm": 21.218143652509674, "learning_rate": 1.3943183613291374e-05, "loss": 0.73, "step": 7488 }, { "epoch": 1.17, "grad_norm": 18.645593324817906, "learning_rate": 1.3941633800582325e-05, "loss": 0.7254, "step": 7489 }, { "epoch": 1.17, "grad_norm": 19.409587059496285, "learning_rate": 1.394008387577552e-05, "loss": 0.7066, "step": 7490 }, { "epoch": 1.17, "grad_norm": 16.0758242745071, "learning_rate": 1.3938533838915033e-05, "loss": 0.6805, "step": 7491 }, { "epoch": 1.17, "grad_norm": 20.481029815308045, "learning_rate": 1.3936983690044949e-05, "loss": 0.6164, "step": 7492 }, { "epoch": 1.17, "grad_norm": 15.771087089235197, "learning_rate": 1.3935433429209352e-05, "loss": 0.6318, "step": 7493 }, { "epoch": 1.17, "grad_norm": 20.97081639388146, "learning_rate": 1.3933883056452331e-05, "loss": 0.7006, "step": 7494 }, { "epoch": 1.17, "grad_norm": 27.4972055204725, "learning_rate": 1.3932332571817975e-05, "loss": 0.6624, "step": 7495 }, { "epoch": 1.17, "grad_norm": 24.643876100745285, "learning_rate": 1.3930781975350382e-05, "loss": 0.6949, "step": 7496 }, { "epoch": 1.17, "grad_norm": 20.89933613779166, "learning_rate": 1.3929231267093646e-05, "loss": 0.7052, "step": 7497 }, { "epoch": 1.17, "grad_norm": 12.711042071980877, "learning_rate": 1.3927680447091876e-05, "loss": 0.811, "step": 7498 }, { "epoch": 1.17, "grad_norm": 10.793785704220694, "learning_rate": 1.3926129515389168e-05, "loss": 0.5723, "step": 7499 }, { "epoch": 1.17, "grad_norm": 24.612089721566733, "learning_rate": 1.3924578472029637e-05, "loss": 0.6871, "step": 7500 }, { "epoch": 1.17, "grad_norm": 14.881192241790526, "learning_rate": 1.3923027317057388e-05, "loss": 0.7151, "step": 7501 }, { "epoch": 1.17, "grad_norm": 32.59567292101353, "learning_rate": 1.3921476050516538e-05, "loss": 0.7365, "step": 7502 }, { "epoch": 1.17, "grad_norm": 14.47455473428712, "learning_rate": 1.3919924672451201e-05, "loss": 0.7011, "step": 7503 }, { "epoch": 1.17, "grad_norm": 16.764728500842555, "learning_rate": 1.3918373182905501e-05, "loss": 0.6463, "step": 7504 }, { "epoch": 1.17, "grad_norm": 20.85860007557551, "learning_rate": 1.391682158192356e-05, "loss": 0.7198, "step": 7505 }, { "epoch": 1.17, "grad_norm": 18.907522179228543, "learning_rate": 1.3915269869549504e-05, "loss": 0.6414, "step": 7506 }, { "epoch": 1.17, "grad_norm": 15.510044998498573, "learning_rate": 1.3913718045827462e-05, "loss": 0.6663, "step": 7507 }, { "epoch": 1.17, "grad_norm": 27.59371658033721, "learning_rate": 1.391216611080157e-05, "loss": 0.6849, "step": 7508 }, { "epoch": 1.17, "grad_norm": 13.287951147669038, "learning_rate": 1.3910614064515964e-05, "loss": 0.6739, "step": 7509 }, { "epoch": 1.17, "grad_norm": 23.478886626050734, "learning_rate": 1.3909061907014781e-05, "loss": 0.6944, "step": 7510 }, { "epoch": 1.17, "grad_norm": 14.072511998812043, "learning_rate": 1.390750963834216e-05, "loss": 0.5905, "step": 7511 }, { "epoch": 1.17, "grad_norm": 42.05425884238126, "learning_rate": 1.3905957258542253e-05, "loss": 0.7121, "step": 7512 }, { "epoch": 1.17, "grad_norm": 16.01832958738399, "learning_rate": 1.390440476765921e-05, "loss": 0.6673, "step": 7513 }, { "epoch": 1.17, "grad_norm": 16.41394322718299, "learning_rate": 1.3902852165737172e-05, "loss": 0.683, "step": 7514 }, { "epoch": 1.17, "grad_norm": 15.456126261052685, "learning_rate": 1.390129945282031e-05, "loss": 0.678, "step": 7515 }, { "epoch": 1.17, "grad_norm": 19.400980989672785, "learning_rate": 1.3899746628952766e-05, "loss": 0.7271, "step": 7516 }, { "epoch": 1.17, "grad_norm": 21.431509641391173, "learning_rate": 1.3898193694178714e-05, "loss": 0.7039, "step": 7517 }, { "epoch": 1.17, "grad_norm": 24.610124583257402, "learning_rate": 1.3896640648542312e-05, "loss": 0.6537, "step": 7518 }, { "epoch": 1.17, "grad_norm": 17.39909699504327, "learning_rate": 1.389508749208773e-05, "loss": 0.7293, "step": 7519 }, { "epoch": 1.17, "grad_norm": 16.21891755594496, "learning_rate": 1.389353422485914e-05, "loss": 0.6648, "step": 7520 }, { "epoch": 1.17, "grad_norm": 23.923362457703828, "learning_rate": 1.3891980846900712e-05, "loss": 0.7032, "step": 7521 }, { "epoch": 1.17, "grad_norm": 19.66710044209375, "learning_rate": 1.3890427358256626e-05, "loss": 0.6923, "step": 7522 }, { "epoch": 1.18, "grad_norm": 13.958698425630674, "learning_rate": 1.388887375897106e-05, "loss": 0.6708, "step": 7523 }, { "epoch": 1.18, "grad_norm": 20.38599063342712, "learning_rate": 1.3887320049088202e-05, "loss": 0.6418, "step": 7524 }, { "epoch": 1.18, "grad_norm": 13.83205121493252, "learning_rate": 1.3885766228652235e-05, "loss": 0.598, "step": 7525 }, { "epoch": 1.18, "grad_norm": 19.79574681968676, "learning_rate": 1.3884212297707349e-05, "loss": 0.6237, "step": 7526 }, { "epoch": 1.18, "grad_norm": 15.677812889038439, "learning_rate": 1.3882658256297738e-05, "loss": 0.7044, "step": 7527 }, { "epoch": 1.18, "grad_norm": 14.830279366624833, "learning_rate": 1.3881104104467598e-05, "loss": 0.6114, "step": 7528 }, { "epoch": 1.18, "grad_norm": 15.913860664342462, "learning_rate": 1.3879549842261123e-05, "loss": 0.6359, "step": 7529 }, { "epoch": 1.18, "grad_norm": 13.892387249923937, "learning_rate": 1.3877995469722527e-05, "loss": 0.6188, "step": 7530 }, { "epoch": 1.18, "grad_norm": 16.58478861911699, "learning_rate": 1.3876440986896004e-05, "loss": 0.6768, "step": 7531 }, { "epoch": 1.18, "grad_norm": 24.869863317629406, "learning_rate": 1.387488639382577e-05, "loss": 0.7038, "step": 7532 }, { "epoch": 1.18, "grad_norm": 30.157638454592195, "learning_rate": 1.387333169055603e-05, "loss": 0.6805, "step": 7533 }, { "epoch": 1.18, "grad_norm": 21.382775782984638, "learning_rate": 1.3871776877131007e-05, "loss": 0.7045, "step": 7534 }, { "epoch": 1.18, "grad_norm": 17.424394742648765, "learning_rate": 1.3870221953594912e-05, "loss": 0.6368, "step": 7535 }, { "epoch": 1.18, "grad_norm": 21.88825740236287, "learning_rate": 1.3868666919991966e-05, "loss": 0.6372, "step": 7536 }, { "epoch": 1.18, "grad_norm": 16.92365602822914, "learning_rate": 1.38671117763664e-05, "loss": 0.6643, "step": 7537 }, { "epoch": 1.18, "grad_norm": 16.729266756545332, "learning_rate": 1.3865556522762438e-05, "loss": 0.679, "step": 7538 }, { "epoch": 1.18, "grad_norm": 13.845997522974661, "learning_rate": 1.3864001159224306e-05, "loss": 0.6784, "step": 7539 }, { "epoch": 1.18, "grad_norm": 27.252948094053966, "learning_rate": 1.3862445685796244e-05, "loss": 0.6914, "step": 7540 }, { "epoch": 1.18, "grad_norm": 26.367243305288532, "learning_rate": 1.3860890102522482e-05, "loss": 0.7281, "step": 7541 }, { "epoch": 1.18, "grad_norm": 15.747856897861515, "learning_rate": 1.385933440944727e-05, "loss": 0.6155, "step": 7542 }, { "epoch": 1.18, "grad_norm": 21.461379410839402, "learning_rate": 1.385777860661484e-05, "loss": 0.6093, "step": 7543 }, { "epoch": 1.18, "grad_norm": 20.55451364925784, "learning_rate": 1.3856222694069446e-05, "loss": 0.6032, "step": 7544 }, { "epoch": 1.18, "grad_norm": 26.457505973072525, "learning_rate": 1.3854666671855334e-05, "loss": 0.624, "step": 7545 }, { "epoch": 1.18, "grad_norm": 15.36800395976475, "learning_rate": 1.3853110540016759e-05, "loss": 0.649, "step": 7546 }, { "epoch": 1.18, "grad_norm": 20.234854057709228, "learning_rate": 1.3851554298597971e-05, "loss": 0.7901, "step": 7547 }, { "epoch": 1.18, "grad_norm": 23.349697126593004, "learning_rate": 1.384999794764323e-05, "loss": 0.6993, "step": 7548 }, { "epoch": 1.18, "grad_norm": 36.34435634976441, "learning_rate": 1.3848441487196804e-05, "loss": 0.7236, "step": 7549 }, { "epoch": 1.18, "grad_norm": 14.937564966427612, "learning_rate": 1.3846884917302952e-05, "loss": 0.6597, "step": 7550 }, { "epoch": 1.18, "grad_norm": 19.575933954124604, "learning_rate": 1.3845328238005942e-05, "loss": 0.7012, "step": 7551 }, { "epoch": 1.18, "grad_norm": 28.011320616404007, "learning_rate": 1.3843771449350047e-05, "loss": 0.7467, "step": 7552 }, { "epoch": 1.18, "grad_norm": 14.13286322652592, "learning_rate": 1.3842214551379542e-05, "loss": 0.6014, "step": 7553 }, { "epoch": 1.18, "grad_norm": 25.82566865028165, "learning_rate": 1.3840657544138702e-05, "loss": 0.7456, "step": 7554 }, { "epoch": 1.18, "grad_norm": 13.244024722615029, "learning_rate": 1.3839100427671808e-05, "loss": 0.6363, "step": 7555 }, { "epoch": 1.18, "grad_norm": 13.811069034658267, "learning_rate": 1.3837543202023143e-05, "loss": 0.6291, "step": 7556 }, { "epoch": 1.18, "grad_norm": 15.160090923291442, "learning_rate": 1.3835985867236995e-05, "loss": 0.652, "step": 7557 }, { "epoch": 1.18, "grad_norm": 15.11189345926525, "learning_rate": 1.383442842335765e-05, "loss": 0.6006, "step": 7558 }, { "epoch": 1.18, "grad_norm": 26.496299290692853, "learning_rate": 1.383287087042941e-05, "loss": 0.7171, "step": 7559 }, { "epoch": 1.18, "grad_norm": 13.572744585304314, "learning_rate": 1.383131320849656e-05, "loss": 0.5934, "step": 7560 }, { "epoch": 1.18, "grad_norm": 16.945321114284212, "learning_rate": 1.3829755437603405e-05, "loss": 0.6502, "step": 7561 }, { "epoch": 1.18, "grad_norm": 21.953599670817265, "learning_rate": 1.3828197557794245e-05, "loss": 0.6227, "step": 7562 }, { "epoch": 1.18, "grad_norm": 16.30852525159876, "learning_rate": 1.3826639569113384e-05, "loss": 0.5657, "step": 7563 }, { "epoch": 1.18, "grad_norm": 27.66963524803588, "learning_rate": 1.3825081471605137e-05, "loss": 0.6587, "step": 7564 }, { "epoch": 1.18, "grad_norm": 14.762473263502336, "learning_rate": 1.3823523265313808e-05, "loss": 0.6205, "step": 7565 }, { "epoch": 1.18, "grad_norm": 15.477765238289495, "learning_rate": 1.3821964950283713e-05, "loss": 0.5983, "step": 7566 }, { "epoch": 1.18, "grad_norm": 14.924384916632853, "learning_rate": 1.3820406526559169e-05, "loss": 0.6099, "step": 7567 }, { "epoch": 1.18, "grad_norm": 16.781352330984465, "learning_rate": 1.38188479941845e-05, "loss": 0.6457, "step": 7568 }, { "epoch": 1.18, "grad_norm": 27.584528475708602, "learning_rate": 1.381728935320403e-05, "loss": 0.6183, "step": 7569 }, { "epoch": 1.18, "grad_norm": 25.564806360203946, "learning_rate": 1.3815730603662083e-05, "loss": 0.6289, "step": 7570 }, { "epoch": 1.18, "grad_norm": 20.089715600091765, "learning_rate": 1.3814171745602991e-05, "loss": 0.6958, "step": 7571 }, { "epoch": 1.18, "grad_norm": 23.6109904551384, "learning_rate": 1.3812612779071084e-05, "loss": 0.7949, "step": 7572 }, { "epoch": 1.18, "grad_norm": 21.26732133080964, "learning_rate": 1.3811053704110697e-05, "loss": 0.6563, "step": 7573 }, { "epoch": 1.18, "grad_norm": 19.909600846430628, "learning_rate": 1.3809494520766179e-05, "loss": 0.5822, "step": 7574 }, { "epoch": 1.18, "grad_norm": 17.463115556294614, "learning_rate": 1.3807935229081859e-05, "loss": 0.6293, "step": 7575 }, { "epoch": 1.18, "grad_norm": 16.67937550398115, "learning_rate": 1.3806375829102092e-05, "loss": 0.6278, "step": 7576 }, { "epoch": 1.18, "grad_norm": 17.35971127037589, "learning_rate": 1.3804816320871221e-05, "loss": 0.6333, "step": 7577 }, { "epoch": 1.18, "grad_norm": 16.773158144834433, "learning_rate": 1.38032567044336e-05, "loss": 0.6524, "step": 7578 }, { "epoch": 1.18, "grad_norm": 31.53549174216908, "learning_rate": 1.3801696979833586e-05, "loss": 0.7427, "step": 7579 }, { "epoch": 1.18, "grad_norm": 14.63057206910241, "learning_rate": 1.3800137147115531e-05, "loss": 0.7508, "step": 7580 }, { "epoch": 1.18, "grad_norm": 19.8101798712467, "learning_rate": 1.37985772063238e-05, "loss": 0.7527, "step": 7581 }, { "epoch": 1.18, "grad_norm": 23.001134149949873, "learning_rate": 1.3797017157502754e-05, "loss": 0.6924, "step": 7582 }, { "epoch": 1.18, "grad_norm": 31.23867863927879, "learning_rate": 1.3795457000696762e-05, "loss": 0.6913, "step": 7583 }, { "epoch": 1.18, "grad_norm": 23.772363631571118, "learning_rate": 1.3793896735950195e-05, "loss": 0.6638, "step": 7584 }, { "epoch": 1.18, "grad_norm": 13.612161568413985, "learning_rate": 1.3792336363307419e-05, "loss": 0.6242, "step": 7585 }, { "epoch": 1.18, "grad_norm": 19.39163862709829, "learning_rate": 1.3790775882812818e-05, "loss": 0.7153, "step": 7586 }, { "epoch": 1.19, "grad_norm": 28.381249904275546, "learning_rate": 1.378921529451077e-05, "loss": 0.7509, "step": 7587 }, { "epoch": 1.19, "grad_norm": 16.591754432341034, "learning_rate": 1.3787654598445656e-05, "loss": 0.777, "step": 7588 }, { "epoch": 1.19, "grad_norm": 20.40724545142196, "learning_rate": 1.3786093794661858e-05, "loss": 0.6377, "step": 7589 }, { "epoch": 1.19, "grad_norm": 16.488630621416277, "learning_rate": 1.3784532883203769e-05, "loss": 0.7281, "step": 7590 }, { "epoch": 1.19, "grad_norm": 26.80473266926859, "learning_rate": 1.378297186411578e-05, "loss": 0.6796, "step": 7591 }, { "epoch": 1.19, "grad_norm": 22.218804443536794, "learning_rate": 1.378141073744228e-05, "loss": 0.7338, "step": 7592 }, { "epoch": 1.19, "grad_norm": 20.18899216318968, "learning_rate": 1.3779849503227675e-05, "loss": 0.5817, "step": 7593 }, { "epoch": 1.19, "grad_norm": 20.51721928104307, "learning_rate": 1.377828816151636e-05, "loss": 0.7058, "step": 7594 }, { "epoch": 1.19, "grad_norm": 19.66716454915918, "learning_rate": 1.377672671235274e-05, "loss": 0.7502, "step": 7595 }, { "epoch": 1.19, "grad_norm": 17.583078490297684, "learning_rate": 1.377516515578122e-05, "loss": 0.6846, "step": 7596 }, { "epoch": 1.19, "grad_norm": 39.52117862531305, "learning_rate": 1.3773603491846213e-05, "loss": 0.6456, "step": 7597 }, { "epoch": 1.19, "grad_norm": 25.83714727863196, "learning_rate": 1.3772041720592131e-05, "loss": 0.719, "step": 7598 }, { "epoch": 1.19, "grad_norm": 16.26837292379662, "learning_rate": 1.3770479842063387e-05, "loss": 0.5968, "step": 7599 }, { "epoch": 1.19, "grad_norm": 14.494083477936199, "learning_rate": 1.37689178563044e-05, "loss": 0.6826, "step": 7600 }, { "epoch": 1.19, "grad_norm": 16.190142649637103, "learning_rate": 1.37673557633596e-05, "loss": 0.7216, "step": 7601 }, { "epoch": 1.19, "grad_norm": 21.799027146975, "learning_rate": 1.3765793563273402e-05, "loss": 0.6388, "step": 7602 }, { "epoch": 1.19, "grad_norm": 18.595024924285916, "learning_rate": 1.376423125609024e-05, "loss": 0.6552, "step": 7603 }, { "epoch": 1.19, "grad_norm": 24.91833612434623, "learning_rate": 1.376266884185454e-05, "loss": 0.7241, "step": 7604 }, { "epoch": 1.19, "grad_norm": 33.587980181204244, "learning_rate": 1.376110632061074e-05, "loss": 0.6916, "step": 7605 }, { "epoch": 1.19, "grad_norm": 25.88662066425963, "learning_rate": 1.3759543692403278e-05, "loss": 0.6966, "step": 7606 }, { "epoch": 1.19, "grad_norm": 18.996353307665114, "learning_rate": 1.3757980957276594e-05, "loss": 0.7129, "step": 7607 }, { "epoch": 1.19, "grad_norm": 18.07633098406556, "learning_rate": 1.3756418115275128e-05, "loss": 0.6685, "step": 7608 }, { "epoch": 1.19, "grad_norm": 20.98864413724588, "learning_rate": 1.3754855166443326e-05, "loss": 0.6743, "step": 7609 }, { "epoch": 1.19, "grad_norm": 22.32496062053142, "learning_rate": 1.375329211082564e-05, "loss": 0.6902, "step": 7610 }, { "epoch": 1.19, "grad_norm": 18.280417487759227, "learning_rate": 1.3751728948466526e-05, "loss": 0.6503, "step": 7611 }, { "epoch": 1.19, "grad_norm": 15.466755148350478, "learning_rate": 1.3750165679410431e-05, "loss": 0.6718, "step": 7612 }, { "epoch": 1.19, "grad_norm": 13.280107986493373, "learning_rate": 1.3748602303701821e-05, "loss": 0.6518, "step": 7613 }, { "epoch": 1.19, "grad_norm": 11.770372852212237, "learning_rate": 1.3747038821385155e-05, "loss": 0.6022, "step": 7614 }, { "epoch": 1.19, "grad_norm": 15.899196220495579, "learning_rate": 1.3745475232504895e-05, "loss": 0.6409, "step": 7615 }, { "epoch": 1.19, "grad_norm": 12.657098477968363, "learning_rate": 1.374391153710551e-05, "loss": 0.6713, "step": 7616 }, { "epoch": 1.19, "grad_norm": 16.680950969218298, "learning_rate": 1.374234773523147e-05, "loss": 0.7287, "step": 7617 }, { "epoch": 1.19, "grad_norm": 16.114378093032236, "learning_rate": 1.3740783826927252e-05, "loss": 0.7616, "step": 7618 }, { "epoch": 1.19, "grad_norm": 22.38611065463507, "learning_rate": 1.373921981223733e-05, "loss": 0.7034, "step": 7619 }, { "epoch": 1.19, "grad_norm": 18.81119139213464, "learning_rate": 1.3737655691206184e-05, "loss": 0.6361, "step": 7620 }, { "epoch": 1.19, "grad_norm": 29.82213783823889, "learning_rate": 1.3736091463878296e-05, "loss": 0.6597, "step": 7621 }, { "epoch": 1.19, "grad_norm": 31.98501895615954, "learning_rate": 1.3734527130298153e-05, "loss": 0.7065, "step": 7622 }, { "epoch": 1.19, "grad_norm": 19.824982803353503, "learning_rate": 1.373296269051024e-05, "loss": 0.6366, "step": 7623 }, { "epoch": 1.19, "grad_norm": 13.33518289711661, "learning_rate": 1.3731398144559056e-05, "loss": 0.6591, "step": 7624 }, { "epoch": 1.19, "grad_norm": 19.55279041450296, "learning_rate": 1.3729833492489091e-05, "loss": 0.6656, "step": 7625 }, { "epoch": 1.19, "grad_norm": 20.840946987019695, "learning_rate": 1.3728268734344842e-05, "loss": 0.6558, "step": 7626 }, { "epoch": 1.19, "grad_norm": 18.403385587937322, "learning_rate": 1.372670387017081e-05, "loss": 0.7045, "step": 7627 }, { "epoch": 1.19, "grad_norm": 13.664214010145676, "learning_rate": 1.3725138900011503e-05, "loss": 0.6638, "step": 7628 }, { "epoch": 1.19, "grad_norm": 17.38587578731434, "learning_rate": 1.372357382391142e-05, "loss": 0.6434, "step": 7629 }, { "epoch": 1.19, "grad_norm": 14.366065858470517, "learning_rate": 1.3722008641915081e-05, "loss": 0.7008, "step": 7630 }, { "epoch": 1.19, "grad_norm": 20.743109412817446, "learning_rate": 1.372044335406699e-05, "loss": 0.6805, "step": 7631 }, { "epoch": 1.19, "grad_norm": 21.621152475237675, "learning_rate": 1.3718877960411669e-05, "loss": 0.6004, "step": 7632 }, { "epoch": 1.19, "grad_norm": 14.985923084486366, "learning_rate": 1.3717312460993632e-05, "loss": 0.6458, "step": 7633 }, { "epoch": 1.19, "grad_norm": 14.831499571936895, "learning_rate": 1.3715746855857407e-05, "loss": 0.5965, "step": 7634 }, { "epoch": 1.19, "grad_norm": 19.378227400774456, "learning_rate": 1.3714181145047512e-05, "loss": 0.682, "step": 7635 }, { "epoch": 1.19, "grad_norm": 21.8385788468174, "learning_rate": 1.3712615328608476e-05, "loss": 0.6703, "step": 7636 }, { "epoch": 1.19, "grad_norm": 20.541157472343233, "learning_rate": 1.3711049406584836e-05, "loss": 0.6786, "step": 7637 }, { "epoch": 1.19, "grad_norm": 28.222698862644222, "learning_rate": 1.3709483379021118e-05, "loss": 0.6697, "step": 7638 }, { "epoch": 1.19, "grad_norm": 19.572778544125555, "learning_rate": 1.3707917245961864e-05, "loss": 0.7053, "step": 7639 }, { "epoch": 1.19, "grad_norm": 19.73125718557368, "learning_rate": 1.3706351007451616e-05, "loss": 0.6334, "step": 7640 }, { "epoch": 1.19, "grad_norm": 14.35657112144822, "learning_rate": 1.3704784663534909e-05, "loss": 0.6475, "step": 7641 }, { "epoch": 1.19, "grad_norm": 18.2966139470998, "learning_rate": 1.3703218214256295e-05, "loss": 0.7055, "step": 7642 }, { "epoch": 1.19, "grad_norm": 19.124191457199956, "learning_rate": 1.3701651659660322e-05, "loss": 0.6314, "step": 7643 }, { "epoch": 1.19, "grad_norm": 16.737879939036816, "learning_rate": 1.3700084999791539e-05, "loss": 0.6366, "step": 7644 }, { "epoch": 1.19, "grad_norm": 15.617573194716968, "learning_rate": 1.3698518234694507e-05, "loss": 0.7464, "step": 7645 }, { "epoch": 1.19, "grad_norm": 19.01912822833064, "learning_rate": 1.3696951364413776e-05, "loss": 0.654, "step": 7646 }, { "epoch": 1.19, "grad_norm": 23.066088949309595, "learning_rate": 1.3695384388993915e-05, "loss": 0.7175, "step": 7647 }, { "epoch": 1.19, "grad_norm": 14.099934492948222, "learning_rate": 1.369381730847948e-05, "loss": 0.6197, "step": 7648 }, { "epoch": 1.19, "grad_norm": 16.964683048858276, "learning_rate": 1.3692250122915042e-05, "loss": 0.6001, "step": 7649 }, { "epoch": 1.19, "grad_norm": 18.877150163643428, "learning_rate": 1.3690682832345172e-05, "loss": 0.6811, "step": 7650 }, { "epoch": 1.2, "grad_norm": 16.667626102270013, "learning_rate": 1.3689115436814442e-05, "loss": 0.6262, "step": 7651 }, { "epoch": 1.2, "grad_norm": 17.954731895040663, "learning_rate": 1.3687547936367424e-05, "loss": 0.6491, "step": 7652 }, { "epoch": 1.2, "grad_norm": 27.787930592887133, "learning_rate": 1.36859803310487e-05, "loss": 0.7097, "step": 7653 }, { "epoch": 1.2, "grad_norm": 17.17398208354797, "learning_rate": 1.3684412620902854e-05, "loss": 0.7381, "step": 7654 }, { "epoch": 1.2, "grad_norm": 18.47236551599288, "learning_rate": 1.3682844805974466e-05, "loss": 0.7036, "step": 7655 }, { "epoch": 1.2, "grad_norm": 23.80344222823844, "learning_rate": 1.3681276886308129e-05, "loss": 0.6652, "step": 7656 }, { "epoch": 1.2, "grad_norm": 13.467410604034352, "learning_rate": 1.3679708861948429e-05, "loss": 0.6617, "step": 7657 }, { "epoch": 1.2, "grad_norm": 22.298707388184887, "learning_rate": 1.3678140732939962e-05, "loss": 0.6192, "step": 7658 }, { "epoch": 1.2, "grad_norm": 15.506839285419602, "learning_rate": 1.3676572499327322e-05, "loss": 0.677, "step": 7659 }, { "epoch": 1.2, "grad_norm": 15.5937809117265, "learning_rate": 1.3675004161155116e-05, "loss": 0.688, "step": 7660 }, { "epoch": 1.2, "grad_norm": 18.131816096625517, "learning_rate": 1.3673435718467936e-05, "loss": 0.6501, "step": 7661 }, { "epoch": 1.2, "grad_norm": 16.64486544268422, "learning_rate": 1.3671867171310398e-05, "loss": 0.6203, "step": 7662 }, { "epoch": 1.2, "grad_norm": 19.945936554089887, "learning_rate": 1.36702985197271e-05, "loss": 0.6201, "step": 7663 }, { "epoch": 1.2, "grad_norm": 19.82345189607954, "learning_rate": 1.3668729763762664e-05, "loss": 0.6751, "step": 7664 }, { "epoch": 1.2, "grad_norm": 23.04941698441705, "learning_rate": 1.3667160903461699e-05, "loss": 0.7807, "step": 7665 }, { "epoch": 1.2, "grad_norm": 16.25502799540227, "learning_rate": 1.366559193886882e-05, "loss": 0.6261, "step": 7666 }, { "epoch": 1.2, "grad_norm": 25.027603123063447, "learning_rate": 1.3664022870028655e-05, "loss": 0.7956, "step": 7667 }, { "epoch": 1.2, "grad_norm": 19.668726061547922, "learning_rate": 1.366245369698582e-05, "loss": 0.6413, "step": 7668 }, { "epoch": 1.2, "grad_norm": 30.615400659331648, "learning_rate": 1.3660884419784944e-05, "loss": 0.6724, "step": 7669 }, { "epoch": 1.2, "grad_norm": 18.61355365501805, "learning_rate": 1.3659315038470657e-05, "loss": 0.6764, "step": 7670 }, { "epoch": 1.2, "grad_norm": 25.317622178655764, "learning_rate": 1.3657745553087591e-05, "loss": 0.6982, "step": 7671 }, { "epoch": 1.2, "grad_norm": 16.344091877725926, "learning_rate": 1.3656175963680385e-05, "loss": 0.6288, "step": 7672 }, { "epoch": 1.2, "grad_norm": 26.068582226163738, "learning_rate": 1.3654606270293666e-05, "loss": 0.6353, "step": 7673 }, { "epoch": 1.2, "grad_norm": 21.689203383696714, "learning_rate": 1.3653036472972088e-05, "loss": 0.67, "step": 7674 }, { "epoch": 1.2, "grad_norm": 17.55984097125743, "learning_rate": 1.365146657176029e-05, "loss": 0.6526, "step": 7675 }, { "epoch": 1.2, "grad_norm": 17.987938094205383, "learning_rate": 1.3649896566702916e-05, "loss": 0.6221, "step": 7676 }, { "epoch": 1.2, "grad_norm": 17.697086014618723, "learning_rate": 1.364832645784462e-05, "loss": 0.685, "step": 7677 }, { "epoch": 1.2, "grad_norm": 18.614157721106924, "learning_rate": 1.3646756245230056e-05, "loss": 0.7203, "step": 7678 }, { "epoch": 1.2, "grad_norm": 24.74984298431076, "learning_rate": 1.3645185928903874e-05, "loss": 0.5951, "step": 7679 }, { "epoch": 1.2, "grad_norm": 17.58654666237059, "learning_rate": 1.3643615508910734e-05, "loss": 0.5818, "step": 7680 }, { "epoch": 1.2, "grad_norm": 16.597103354334525, "learning_rate": 1.3642044985295307e-05, "loss": 0.6439, "step": 7681 }, { "epoch": 1.2, "grad_norm": 21.33460279471346, "learning_rate": 1.3640474358102247e-05, "loss": 0.7883, "step": 7682 }, { "epoch": 1.2, "grad_norm": 12.261608012690708, "learning_rate": 1.3638903627376227e-05, "loss": 0.66, "step": 7683 }, { "epoch": 1.2, "grad_norm": 20.19728792220861, "learning_rate": 1.3637332793161918e-05, "loss": 0.7208, "step": 7684 }, { "epoch": 1.2, "grad_norm": 23.352946889048283, "learning_rate": 1.363576185550399e-05, "loss": 0.6549, "step": 7685 }, { "epoch": 1.2, "grad_norm": 14.845646273305002, "learning_rate": 1.3634190814447124e-05, "loss": 0.6223, "step": 7686 }, { "epoch": 1.2, "grad_norm": 27.27922149754853, "learning_rate": 1.3632619670035996e-05, "loss": 0.7209, "step": 7687 }, { "epoch": 1.2, "grad_norm": 24.391918448658416, "learning_rate": 1.3631048422315288e-05, "loss": 0.6488, "step": 7688 }, { "epoch": 1.2, "grad_norm": 27.26403424560335, "learning_rate": 1.3629477071329689e-05, "loss": 0.5684, "step": 7689 }, { "epoch": 1.2, "grad_norm": 14.135181514972738, "learning_rate": 1.3627905617123884e-05, "loss": 0.5704, "step": 7690 }, { "epoch": 1.2, "grad_norm": 11.742921262836303, "learning_rate": 1.3626334059742567e-05, "loss": 0.5532, "step": 7691 }, { "epoch": 1.2, "grad_norm": 26.149796006398997, "learning_rate": 1.362476239923043e-05, "loss": 0.6653, "step": 7692 }, { "epoch": 1.2, "grad_norm": 22.8417530230639, "learning_rate": 1.3623190635632171e-05, "loss": 0.6207, "step": 7693 }, { "epoch": 1.2, "grad_norm": 21.328355889939534, "learning_rate": 1.3621618768992488e-05, "loss": 0.5995, "step": 7694 }, { "epoch": 1.2, "grad_norm": 12.363896644185626, "learning_rate": 1.3620046799356088e-05, "loss": 0.5512, "step": 7695 }, { "epoch": 1.2, "grad_norm": 17.73881009138261, "learning_rate": 1.3618474726767675e-05, "loss": 0.6583, "step": 7696 }, { "epoch": 1.2, "grad_norm": 24.47576820942793, "learning_rate": 1.3616902551271955e-05, "loss": 0.7518, "step": 7697 }, { "epoch": 1.2, "grad_norm": 14.461286948632727, "learning_rate": 1.3615330272913643e-05, "loss": 0.6328, "step": 7698 }, { "epoch": 1.2, "grad_norm": 30.251965504230686, "learning_rate": 1.3613757891737453e-05, "loss": 0.7129, "step": 7699 }, { "epoch": 1.2, "grad_norm": 13.70995665462603, "learning_rate": 1.3612185407788104e-05, "loss": 0.5815, "step": 7700 }, { "epoch": 1.2, "grad_norm": 18.64415549573565, "learning_rate": 1.3610612821110315e-05, "loss": 0.6377, "step": 7701 }, { "epoch": 1.2, "grad_norm": 27.733259576745162, "learning_rate": 1.360904013174881e-05, "loss": 0.6813, "step": 7702 }, { "epoch": 1.2, "grad_norm": 18.172275078371847, "learning_rate": 1.360746733974831e-05, "loss": 0.7182, "step": 7703 }, { "epoch": 1.2, "grad_norm": 21.055222941510106, "learning_rate": 1.3605894445153556e-05, "loss": 0.7182, "step": 7704 }, { "epoch": 1.2, "grad_norm": 15.476187080902122, "learning_rate": 1.3604321448009268e-05, "loss": 0.5663, "step": 7705 }, { "epoch": 1.2, "grad_norm": 17.040039513294065, "learning_rate": 1.3602748348360191e-05, "loss": 0.6797, "step": 7706 }, { "epoch": 1.2, "grad_norm": 16.51030181028526, "learning_rate": 1.3601175146251056e-05, "loss": 0.5196, "step": 7707 }, { "epoch": 1.2, "grad_norm": 18.308671933221458, "learning_rate": 1.3599601841726608e-05, "loss": 0.6526, "step": 7708 }, { "epoch": 1.2, "grad_norm": 18.208077474119694, "learning_rate": 1.3598028434831584e-05, "loss": 0.6512, "step": 7709 }, { "epoch": 1.2, "grad_norm": 23.929046116626616, "learning_rate": 1.3596454925610742e-05, "loss": 0.6529, "step": 7710 }, { "epoch": 1.2, "grad_norm": 25.957286642294974, "learning_rate": 1.3594881314108823e-05, "loss": 0.7335, "step": 7711 }, { "epoch": 1.2, "grad_norm": 22.292747626136237, "learning_rate": 1.3593307600370585e-05, "loss": 0.6245, "step": 7712 }, { "epoch": 1.2, "grad_norm": 26.95357431894152, "learning_rate": 1.359173378444078e-05, "loss": 0.6813, "step": 7713 }, { "epoch": 1.2, "grad_norm": 17.831349505024185, "learning_rate": 1.3590159866364166e-05, "loss": 0.6319, "step": 7714 }, { "epoch": 1.21, "grad_norm": 20.648316742756453, "learning_rate": 1.3588585846185502e-05, "loss": 0.7271, "step": 7715 }, { "epoch": 1.21, "grad_norm": 14.373605419492204, "learning_rate": 1.3587011723949564e-05, "loss": 0.5684, "step": 7716 }, { "epoch": 1.21, "grad_norm": 28.810962346352422, "learning_rate": 1.3585437499701104e-05, "loss": 0.7036, "step": 7717 }, { "epoch": 1.21, "grad_norm": 24.28804142187009, "learning_rate": 1.3583863173484902e-05, "loss": 0.6169, "step": 7718 }, { "epoch": 1.21, "grad_norm": 21.76797237766549, "learning_rate": 1.3582288745345728e-05, "loss": 0.6861, "step": 7719 }, { "epoch": 1.21, "grad_norm": 14.63998005565017, "learning_rate": 1.3580714215328357e-05, "loss": 0.5862, "step": 7720 }, { "epoch": 1.21, "grad_norm": 24.53303063818445, "learning_rate": 1.357913958347757e-05, "loss": 0.6792, "step": 7721 }, { "epoch": 1.21, "grad_norm": 13.826674323848138, "learning_rate": 1.3577564849838146e-05, "loss": 0.6953, "step": 7722 }, { "epoch": 1.21, "grad_norm": 13.313937702809529, "learning_rate": 1.3575990014454869e-05, "loss": 0.5925, "step": 7723 }, { "epoch": 1.21, "grad_norm": 23.865880868610812, "learning_rate": 1.3574415077372527e-05, "loss": 0.6402, "step": 7724 }, { "epoch": 1.21, "grad_norm": 23.010248769582482, "learning_rate": 1.3572840038635913e-05, "loss": 0.6418, "step": 7725 }, { "epoch": 1.21, "grad_norm": 16.122639224503757, "learning_rate": 1.357126489828982e-05, "loss": 0.6886, "step": 7726 }, { "epoch": 1.21, "grad_norm": 22.33183266466506, "learning_rate": 1.356968965637904e-05, "loss": 0.6792, "step": 7727 }, { "epoch": 1.21, "grad_norm": 17.824989547549205, "learning_rate": 1.3568114312948375e-05, "loss": 0.5982, "step": 7728 }, { "epoch": 1.21, "grad_norm": 18.848025552165538, "learning_rate": 1.3566538868042627e-05, "loss": 0.6607, "step": 7729 }, { "epoch": 1.21, "grad_norm": 24.38716803079882, "learning_rate": 1.3564963321706597e-05, "loss": 0.7229, "step": 7730 }, { "epoch": 1.21, "grad_norm": 16.13679198167889, "learning_rate": 1.35633876739851e-05, "loss": 0.5827, "step": 7731 }, { "epoch": 1.21, "grad_norm": 13.391859682018913, "learning_rate": 1.3561811924922939e-05, "loss": 0.6427, "step": 7732 }, { "epoch": 1.21, "grad_norm": 28.225278217525755, "learning_rate": 1.3560236074564935e-05, "loss": 0.6938, "step": 7733 }, { "epoch": 1.21, "grad_norm": 17.87479051137091, "learning_rate": 1.3558660122955894e-05, "loss": 0.5692, "step": 7734 }, { "epoch": 1.21, "grad_norm": 16.634308373254914, "learning_rate": 1.3557084070140644e-05, "loss": 0.6371, "step": 7735 }, { "epoch": 1.21, "grad_norm": 17.132658715521163, "learning_rate": 1.3555507916164002e-05, "loss": 0.6319, "step": 7736 }, { "epoch": 1.21, "grad_norm": 59.79078517262928, "learning_rate": 1.3553931661070796e-05, "loss": 0.7993, "step": 7737 }, { "epoch": 1.21, "grad_norm": 14.675090233356418, "learning_rate": 1.3552355304905851e-05, "loss": 0.6695, "step": 7738 }, { "epoch": 1.21, "grad_norm": 18.822836955615433, "learning_rate": 1.3550778847714001e-05, "loss": 0.6723, "step": 7739 }, { "epoch": 1.21, "grad_norm": 18.56423243472237, "learning_rate": 1.3549202289540077e-05, "loss": 0.6465, "step": 7740 }, { "epoch": 1.21, "grad_norm": 12.799417812501762, "learning_rate": 1.3547625630428919e-05, "loss": 0.6235, "step": 7741 }, { "epoch": 1.21, "grad_norm": 23.001394396734433, "learning_rate": 1.3546048870425356e-05, "loss": 0.7434, "step": 7742 }, { "epoch": 1.21, "grad_norm": 16.618670841687017, "learning_rate": 1.3544472009574243e-05, "loss": 0.6331, "step": 7743 }, { "epoch": 1.21, "grad_norm": 22.760969114618792, "learning_rate": 1.354289504792042e-05, "loss": 0.6606, "step": 7744 }, { "epoch": 1.21, "grad_norm": 12.166953750820301, "learning_rate": 1.3541317985508733e-05, "loss": 0.7009, "step": 7745 }, { "epoch": 1.21, "grad_norm": 19.23684946609206, "learning_rate": 1.3539740822384034e-05, "loss": 0.6593, "step": 7746 }, { "epoch": 1.21, "grad_norm": 17.166807897073213, "learning_rate": 1.3538163558591175e-05, "loss": 0.6959, "step": 7747 }, { "epoch": 1.21, "grad_norm": 15.08195940191766, "learning_rate": 1.3536586194175019e-05, "loss": 0.7125, "step": 7748 }, { "epoch": 1.21, "grad_norm": 17.443448226338507, "learning_rate": 1.3535008729180414e-05, "loss": 0.7189, "step": 7749 }, { "epoch": 1.21, "grad_norm": 15.9016326289895, "learning_rate": 1.3533431163652235e-05, "loss": 0.7998, "step": 7750 }, { "epoch": 1.21, "grad_norm": 20.443605478528706, "learning_rate": 1.3531853497635336e-05, "loss": 0.6639, "step": 7751 }, { "epoch": 1.21, "grad_norm": 14.645828533533575, "learning_rate": 1.3530275731174591e-05, "loss": 0.694, "step": 7752 }, { "epoch": 1.21, "grad_norm": 27.10189880268002, "learning_rate": 1.352869786431487e-05, "loss": 0.6703, "step": 7753 }, { "epoch": 1.21, "grad_norm": 18.97391631528998, "learning_rate": 1.3527119897101046e-05, "loss": 0.6051, "step": 7754 }, { "epoch": 1.21, "grad_norm": 16.39083757071754, "learning_rate": 1.3525541829577997e-05, "loss": 0.6977, "step": 7755 }, { "epoch": 1.21, "grad_norm": 11.660695229985553, "learning_rate": 1.3523963661790597e-05, "loss": 0.566, "step": 7756 }, { "epoch": 1.21, "grad_norm": 18.514916987503454, "learning_rate": 1.3522385393783735e-05, "loss": 0.6181, "step": 7757 }, { "epoch": 1.21, "grad_norm": 18.44983075362788, "learning_rate": 1.3520807025602291e-05, "loss": 0.6142, "step": 7758 }, { "epoch": 1.21, "grad_norm": 20.112730421360908, "learning_rate": 1.3519228557291156e-05, "loss": 0.7191, "step": 7759 }, { "epoch": 1.21, "grad_norm": 23.87400981712163, "learning_rate": 1.3517649988895221e-05, "loss": 0.641, "step": 7760 }, { "epoch": 1.21, "grad_norm": 19.363589170268963, "learning_rate": 1.3516071320459374e-05, "loss": 0.6954, "step": 7761 }, { "epoch": 1.21, "grad_norm": 16.762579793334773, "learning_rate": 1.3514492552028518e-05, "loss": 0.7501, "step": 7762 }, { "epoch": 1.21, "grad_norm": 19.25704190722339, "learning_rate": 1.351291368364755e-05, "loss": 0.6863, "step": 7763 }, { "epoch": 1.21, "grad_norm": 17.79564088841133, "learning_rate": 1.3511334715361371e-05, "loss": 0.6296, "step": 7764 }, { "epoch": 1.21, "grad_norm": 17.96447351733195, "learning_rate": 1.3509755647214888e-05, "loss": 0.6559, "step": 7765 }, { "epoch": 1.21, "grad_norm": 22.17274124564975, "learning_rate": 1.3508176479253003e-05, "loss": 0.6416, "step": 7766 }, { "epoch": 1.21, "grad_norm": 12.323950186945373, "learning_rate": 1.3506597211520638e-05, "loss": 0.6761, "step": 7767 }, { "epoch": 1.21, "grad_norm": 21.553193918228306, "learning_rate": 1.3505017844062693e-05, "loss": 0.6436, "step": 7768 }, { "epoch": 1.21, "grad_norm": 17.341149612035274, "learning_rate": 1.3503438376924096e-05, "loss": 0.6985, "step": 7769 }, { "epoch": 1.21, "grad_norm": 18.187834744797613, "learning_rate": 1.350185881014976e-05, "loss": 0.7258, "step": 7770 }, { "epoch": 1.21, "grad_norm": 21.118010345781553, "learning_rate": 1.3500279143784606e-05, "loss": 0.5968, "step": 7771 }, { "epoch": 1.21, "grad_norm": 19.766183982076797, "learning_rate": 1.3498699377873563e-05, "loss": 0.6071, "step": 7772 }, { "epoch": 1.21, "grad_norm": 15.648475346017172, "learning_rate": 1.3497119512461555e-05, "loss": 0.5594, "step": 7773 }, { "epoch": 1.21, "grad_norm": 23.256902129731618, "learning_rate": 1.3495539547593512e-05, "loss": 0.6319, "step": 7774 }, { "epoch": 1.21, "grad_norm": 21.95187187555169, "learning_rate": 1.3493959483314373e-05, "loss": 0.73, "step": 7775 }, { "epoch": 1.21, "grad_norm": 26.797598910771534, "learning_rate": 1.349237931966907e-05, "loss": 0.7152, "step": 7776 }, { "epoch": 1.21, "grad_norm": 22.137483614216, "learning_rate": 1.3490799056702542e-05, "loss": 0.6849, "step": 7777 }, { "epoch": 1.21, "grad_norm": 30.073841916592357, "learning_rate": 1.3489218694459727e-05, "loss": 0.6337, "step": 7778 }, { "epoch": 1.22, "grad_norm": 36.22687657442029, "learning_rate": 1.3487638232985575e-05, "loss": 0.7377, "step": 7779 }, { "epoch": 1.22, "grad_norm": 17.285290256993726, "learning_rate": 1.3486057672325035e-05, "loss": 0.7236, "step": 7780 }, { "epoch": 1.22, "grad_norm": 39.5672733664932, "learning_rate": 1.348447701252305e-05, "loss": 0.8154, "step": 7781 }, { "epoch": 1.22, "grad_norm": 23.84282254868855, "learning_rate": 1.348289625362458e-05, "loss": 0.7073, "step": 7782 }, { "epoch": 1.22, "grad_norm": 24.170662522190558, "learning_rate": 1.3481315395674577e-05, "loss": 0.6956, "step": 7783 }, { "epoch": 1.22, "grad_norm": 12.927685363625908, "learning_rate": 1.3479734438718002e-05, "loss": 0.581, "step": 7784 }, { "epoch": 1.22, "grad_norm": 11.836569437461584, "learning_rate": 1.3478153382799812e-05, "loss": 0.6611, "step": 7785 }, { "epoch": 1.22, "grad_norm": 24.520857340134633, "learning_rate": 1.3476572227964976e-05, "loss": 0.6539, "step": 7786 }, { "epoch": 1.22, "grad_norm": 12.932599178936254, "learning_rate": 1.347499097425846e-05, "loss": 0.6455, "step": 7787 }, { "epoch": 1.22, "grad_norm": 14.924515073712305, "learning_rate": 1.3473409621725233e-05, "loss": 0.6104, "step": 7788 }, { "epoch": 1.22, "grad_norm": 22.843174330924167, "learning_rate": 1.3471828170410268e-05, "loss": 0.6356, "step": 7789 }, { "epoch": 1.22, "grad_norm": 17.22999083816796, "learning_rate": 1.3470246620358543e-05, "loss": 0.645, "step": 7790 }, { "epoch": 1.22, "grad_norm": 18.155008353193594, "learning_rate": 1.3468664971615031e-05, "loss": 0.6822, "step": 7791 }, { "epoch": 1.22, "grad_norm": 23.88314308827424, "learning_rate": 1.3467083224224719e-05, "loss": 0.692, "step": 7792 }, { "epoch": 1.22, "grad_norm": 15.887869052967515, "learning_rate": 1.3465501378232586e-05, "loss": 0.7156, "step": 7793 }, { "epoch": 1.22, "grad_norm": 19.55371340810835, "learning_rate": 1.3463919433683624e-05, "loss": 0.7328, "step": 7794 }, { "epoch": 1.22, "grad_norm": 14.815173644195019, "learning_rate": 1.3462337390622814e-05, "loss": 0.6379, "step": 7795 }, { "epoch": 1.22, "grad_norm": 13.7295557383263, "learning_rate": 1.3460755249095161e-05, "loss": 0.6373, "step": 7796 }, { "epoch": 1.22, "grad_norm": 35.688595746460194, "learning_rate": 1.345917300914565e-05, "loss": 0.7018, "step": 7797 }, { "epoch": 1.22, "grad_norm": 18.029232767640135, "learning_rate": 1.3457590670819283e-05, "loss": 0.6712, "step": 7798 }, { "epoch": 1.22, "grad_norm": 16.215186818242362, "learning_rate": 1.3456008234161057e-05, "loss": 0.6219, "step": 7799 }, { "epoch": 1.22, "grad_norm": 28.221219862978312, "learning_rate": 1.3454425699215982e-05, "loss": 0.6038, "step": 7800 }, { "epoch": 1.22, "grad_norm": 20.461458556760373, "learning_rate": 1.3452843066029058e-05, "loss": 0.5516, "step": 7801 }, { "epoch": 1.22, "grad_norm": 12.928907133449648, "learning_rate": 1.3451260334645299e-05, "loss": 0.66, "step": 7802 }, { "epoch": 1.22, "grad_norm": 16.87503209949902, "learning_rate": 1.3449677505109716e-05, "loss": 0.5877, "step": 7803 }, { "epoch": 1.22, "grad_norm": 27.80302327163075, "learning_rate": 1.3448094577467323e-05, "loss": 0.6431, "step": 7804 }, { "epoch": 1.22, "grad_norm": 39.054457456823506, "learning_rate": 1.3446511551763133e-05, "loss": 0.6847, "step": 7805 }, { "epoch": 1.22, "grad_norm": 15.944321259074611, "learning_rate": 1.3444928428042175e-05, "loss": 0.724, "step": 7806 }, { "epoch": 1.22, "grad_norm": 17.451371262771637, "learning_rate": 1.3443345206349466e-05, "loss": 0.6378, "step": 7807 }, { "epoch": 1.22, "grad_norm": 18.279431429503372, "learning_rate": 1.3441761886730033e-05, "loss": 0.6591, "step": 7808 }, { "epoch": 1.22, "grad_norm": 30.09793328981426, "learning_rate": 1.3440178469228906e-05, "loss": 0.7033, "step": 7809 }, { "epoch": 1.22, "grad_norm": 25.74143649237283, "learning_rate": 1.3438594953891113e-05, "loss": 0.606, "step": 7810 }, { "epoch": 1.22, "grad_norm": 23.527783235374834, "learning_rate": 1.3437011340761697e-05, "loss": 0.7218, "step": 7811 }, { "epoch": 1.22, "grad_norm": 15.67283128367126, "learning_rate": 1.3435427629885684e-05, "loss": 0.6986, "step": 7812 }, { "epoch": 1.22, "grad_norm": 19.904130784686412, "learning_rate": 1.343384382130812e-05, "loss": 0.7566, "step": 7813 }, { "epoch": 1.22, "grad_norm": 21.393523155060652, "learning_rate": 1.3432259915074048e-05, "loss": 0.6636, "step": 7814 }, { "epoch": 1.22, "grad_norm": 13.681988834630545, "learning_rate": 1.3430675911228512e-05, "loss": 0.5694, "step": 7815 }, { "epoch": 1.22, "grad_norm": 17.390931557749273, "learning_rate": 1.342909180981656e-05, "loss": 0.6919, "step": 7816 }, { "epoch": 1.22, "grad_norm": 17.62703428212765, "learning_rate": 1.3427507610883242e-05, "loss": 0.5966, "step": 7817 }, { "epoch": 1.22, "grad_norm": 14.082857439457646, "learning_rate": 1.342592331447361e-05, "loss": 0.5937, "step": 7818 }, { "epoch": 1.22, "grad_norm": 17.16480701483057, "learning_rate": 1.342433892063273e-05, "loss": 0.6725, "step": 7819 }, { "epoch": 1.22, "grad_norm": 15.29070869203043, "learning_rate": 1.3422754429405647e-05, "loss": 0.6131, "step": 7820 }, { "epoch": 1.22, "grad_norm": 16.209781259413884, "learning_rate": 1.3421169840837433e-05, "loss": 0.6267, "step": 7821 }, { "epoch": 1.22, "grad_norm": 13.400291058307083, "learning_rate": 1.341958515497315e-05, "loss": 0.6059, "step": 7822 }, { "epoch": 1.22, "grad_norm": 20.10039043456538, "learning_rate": 1.3418000371857866e-05, "loss": 0.7005, "step": 7823 }, { "epoch": 1.22, "grad_norm": 40.55081024845028, "learning_rate": 1.341641549153665e-05, "loss": 0.7274, "step": 7824 }, { "epoch": 1.22, "grad_norm": 18.49873915770387, "learning_rate": 1.3414830514054578e-05, "loss": 0.6495, "step": 7825 }, { "epoch": 1.22, "grad_norm": 37.50741681027481, "learning_rate": 1.3413245439456721e-05, "loss": 0.5725, "step": 7826 }, { "epoch": 1.22, "grad_norm": 17.86678812325382, "learning_rate": 1.3411660267788163e-05, "loss": 0.6922, "step": 7827 }, { "epoch": 1.22, "grad_norm": 12.908302403716277, "learning_rate": 1.341007499909398e-05, "loss": 0.5855, "step": 7828 }, { "epoch": 1.22, "grad_norm": 16.360341223899788, "learning_rate": 1.3408489633419261e-05, "loss": 0.7458, "step": 7829 }, { "epoch": 1.22, "grad_norm": 19.839485972258483, "learning_rate": 1.340690417080909e-05, "loss": 0.6703, "step": 7830 }, { "epoch": 1.22, "grad_norm": 11.7909383248384, "learning_rate": 1.3405318611308558e-05, "loss": 0.5847, "step": 7831 }, { "epoch": 1.22, "grad_norm": 14.451240739427591, "learning_rate": 1.3403732954962755e-05, "loss": 0.6302, "step": 7832 }, { "epoch": 1.22, "grad_norm": 12.27548092829371, "learning_rate": 1.3402147201816779e-05, "loss": 0.6838, "step": 7833 }, { "epoch": 1.22, "grad_norm": 12.857950330962739, "learning_rate": 1.3400561351915728e-05, "loss": 0.6174, "step": 7834 }, { "epoch": 1.22, "grad_norm": 20.737197755700027, "learning_rate": 1.33989754053047e-05, "loss": 0.6696, "step": 7835 }, { "epoch": 1.22, "grad_norm": 16.101315306506343, "learning_rate": 1.3397389362028802e-05, "loss": 0.573, "step": 7836 }, { "epoch": 1.22, "grad_norm": 20.274495057910308, "learning_rate": 1.3395803222133133e-05, "loss": 0.741, "step": 7837 }, { "epoch": 1.22, "grad_norm": 14.494150524395614, "learning_rate": 1.3394216985662813e-05, "loss": 0.7125, "step": 7838 }, { "epoch": 1.22, "grad_norm": 20.088645319283465, "learning_rate": 1.3392630652662943e-05, "loss": 0.6642, "step": 7839 }, { "epoch": 1.22, "grad_norm": 19.428706659746524, "learning_rate": 1.3391044223178641e-05, "loss": 0.6716, "step": 7840 }, { "epoch": 1.22, "grad_norm": 16.56534645434312, "learning_rate": 1.338945769725503e-05, "loss": 0.5711, "step": 7841 }, { "epoch": 1.22, "grad_norm": 16.620159165927667, "learning_rate": 1.3387871074937222e-05, "loss": 0.7104, "step": 7842 }, { "epoch": 1.23, "grad_norm": 18.010273246035972, "learning_rate": 1.3386284356270342e-05, "loss": 0.6918, "step": 7843 }, { "epoch": 1.23, "grad_norm": 24.633946700862996, "learning_rate": 1.3384697541299518e-05, "loss": 0.6386, "step": 7844 }, { "epoch": 1.23, "grad_norm": 18.301930476555217, "learning_rate": 1.3383110630069874e-05, "loss": 0.6633, "step": 7845 }, { "epoch": 1.23, "grad_norm": 18.57836072690391, "learning_rate": 1.3381523622626544e-05, "loss": 0.7097, "step": 7846 }, { "epoch": 1.23, "grad_norm": 19.97680065682913, "learning_rate": 1.3379936519014658e-05, "loss": 0.6742, "step": 7847 }, { "epoch": 1.23, "grad_norm": 21.171551114944574, "learning_rate": 1.3378349319279358e-05, "loss": 0.6585, "step": 7848 }, { "epoch": 1.23, "grad_norm": 25.342515185104073, "learning_rate": 1.3376762023465775e-05, "loss": 0.7317, "step": 7849 }, { "epoch": 1.23, "grad_norm": 17.011294051955705, "learning_rate": 1.337517463161906e-05, "loss": 0.6369, "step": 7850 }, { "epoch": 1.23, "grad_norm": 13.294206516236073, "learning_rate": 1.337358714378435e-05, "loss": 0.6676, "step": 7851 }, { "epoch": 1.23, "grad_norm": 16.85916209683056, "learning_rate": 1.3371999560006796e-05, "loss": 0.6662, "step": 7852 }, { "epoch": 1.23, "grad_norm": 15.731471469049188, "learning_rate": 1.3370411880331549e-05, "loss": 0.6359, "step": 7853 }, { "epoch": 1.23, "grad_norm": 11.577569937946773, "learning_rate": 1.3368824104803755e-05, "loss": 0.5964, "step": 7854 }, { "epoch": 1.23, "grad_norm": 16.292603830930343, "learning_rate": 1.3367236233468579e-05, "loss": 0.6249, "step": 7855 }, { "epoch": 1.23, "grad_norm": 12.053976986132536, "learning_rate": 1.3365648266371171e-05, "loss": 0.6138, "step": 7856 }, { "epoch": 1.23, "grad_norm": 20.281809365988988, "learning_rate": 1.3364060203556697e-05, "loss": 0.6376, "step": 7857 }, { "epoch": 1.23, "grad_norm": 12.587373227775677, "learning_rate": 1.3362472045070318e-05, "loss": 0.607, "step": 7858 }, { "epoch": 1.23, "grad_norm": 13.037851861391571, "learning_rate": 1.3360883790957202e-05, "loss": 0.6241, "step": 7859 }, { "epoch": 1.23, "grad_norm": 24.69572943528869, "learning_rate": 1.3359295441262516e-05, "loss": 0.677, "step": 7860 }, { "epoch": 1.23, "grad_norm": 25.246291058144273, "learning_rate": 1.3357706996031434e-05, "loss": 0.7144, "step": 7861 }, { "epoch": 1.23, "grad_norm": 26.800505409439165, "learning_rate": 1.3356118455309127e-05, "loss": 0.6671, "step": 7862 }, { "epoch": 1.23, "grad_norm": 21.097743590918743, "learning_rate": 1.3354529819140777e-05, "loss": 0.6569, "step": 7863 }, { "epoch": 1.23, "grad_norm": 21.808594838449327, "learning_rate": 1.3352941087571558e-05, "loss": 0.6306, "step": 7864 }, { "epoch": 1.23, "grad_norm": 19.444113129241654, "learning_rate": 1.3351352260646662e-05, "loss": 0.6454, "step": 7865 }, { "epoch": 1.23, "grad_norm": 14.65124130678045, "learning_rate": 1.3349763338411261e-05, "loss": 0.6305, "step": 7866 }, { "epoch": 1.23, "grad_norm": 17.59297812853889, "learning_rate": 1.3348174320910554e-05, "loss": 0.5736, "step": 7867 }, { "epoch": 1.23, "grad_norm": 3.205577460427556, "learning_rate": 1.3346585208189727e-05, "loss": 0.5189, "step": 7868 }, { "epoch": 1.23, "grad_norm": 16.536255799702765, "learning_rate": 1.3344996000293976e-05, "loss": 0.6064, "step": 7869 }, { "epoch": 1.23, "grad_norm": 31.37655068788581, "learning_rate": 1.3343406697268493e-05, "loss": 0.7551, "step": 7870 }, { "epoch": 1.23, "grad_norm": 22.403108360591908, "learning_rate": 1.334181729915848e-05, "loss": 0.6887, "step": 7871 }, { "epoch": 1.23, "grad_norm": 17.3271059961191, "learning_rate": 1.3340227806009138e-05, "loss": 0.6999, "step": 7872 }, { "epoch": 1.23, "grad_norm": 10.470230556526593, "learning_rate": 1.3338638217865673e-05, "loss": 0.6047, "step": 7873 }, { "epoch": 1.23, "grad_norm": 21.523624682434654, "learning_rate": 1.333704853477329e-05, "loss": 0.6704, "step": 7874 }, { "epoch": 1.23, "grad_norm": 18.07322435052991, "learning_rate": 1.3335458756777195e-05, "loss": 0.6204, "step": 7875 }, { "epoch": 1.23, "grad_norm": 19.890217170444934, "learning_rate": 1.3333868883922606e-05, "loss": 0.6307, "step": 7876 }, { "epoch": 1.23, "grad_norm": 23.18885976989232, "learning_rate": 1.3332278916254739e-05, "loss": 0.6319, "step": 7877 }, { "epoch": 1.23, "grad_norm": 22.45499080585571, "learning_rate": 1.3330688853818808e-05, "loss": 0.6339, "step": 7878 }, { "epoch": 1.23, "grad_norm": 18.178983751781235, "learning_rate": 1.332909869666003e-05, "loss": 0.5528, "step": 7879 }, { "epoch": 1.23, "grad_norm": 14.828388373565346, "learning_rate": 1.3327508444823637e-05, "loss": 0.6013, "step": 7880 }, { "epoch": 1.23, "grad_norm": 35.98520209943347, "learning_rate": 1.3325918098354848e-05, "loss": 0.6541, "step": 7881 }, { "epoch": 1.23, "grad_norm": 12.243383701530203, "learning_rate": 1.33243276572989e-05, "loss": 0.6377, "step": 7882 }, { "epoch": 1.23, "grad_norm": 19.21215025543746, "learning_rate": 1.3322737121701011e-05, "loss": 0.6345, "step": 7883 }, { "epoch": 1.23, "grad_norm": 16.014958809743952, "learning_rate": 1.3321146491606423e-05, "loss": 0.6191, "step": 7884 }, { "epoch": 1.23, "grad_norm": 20.491245505540082, "learning_rate": 1.3319555767060373e-05, "loss": 0.6163, "step": 7885 }, { "epoch": 1.23, "grad_norm": 14.035706632181144, "learning_rate": 1.33179649481081e-05, "loss": 0.6931, "step": 7886 }, { "epoch": 1.23, "grad_norm": 16.461097830565055, "learning_rate": 1.3316374034794844e-05, "loss": 0.7712, "step": 7887 }, { "epoch": 1.23, "grad_norm": 18.399169550948923, "learning_rate": 1.331478302716585e-05, "loss": 0.6629, "step": 7888 }, { "epoch": 1.23, "grad_norm": 20.88543885916358, "learning_rate": 1.3313191925266362e-05, "loss": 0.721, "step": 7889 }, { "epoch": 1.23, "grad_norm": 17.65867786762948, "learning_rate": 1.3311600729141641e-05, "loss": 0.6391, "step": 7890 }, { "epoch": 1.23, "grad_norm": 12.215032964594407, "learning_rate": 1.3310009438836925e-05, "loss": 0.6456, "step": 7891 }, { "epoch": 1.23, "grad_norm": 13.163176239584654, "learning_rate": 1.3308418054397485e-05, "loss": 0.5854, "step": 7892 }, { "epoch": 1.23, "grad_norm": 22.397403010577648, "learning_rate": 1.3306826575868565e-05, "loss": 0.6801, "step": 7893 }, { "epoch": 1.23, "grad_norm": 17.50753423426254, "learning_rate": 1.3305235003295434e-05, "loss": 0.6, "step": 7894 }, { "epoch": 1.23, "grad_norm": 11.560138784772928, "learning_rate": 1.330364333672335e-05, "loss": 0.5809, "step": 7895 }, { "epoch": 1.23, "grad_norm": 15.55651368342618, "learning_rate": 1.3302051576197586e-05, "loss": 0.6925, "step": 7896 }, { "epoch": 1.23, "grad_norm": 14.831075461221058, "learning_rate": 1.3300459721763405e-05, "loss": 0.6789, "step": 7897 }, { "epoch": 1.23, "grad_norm": 20.710369055180365, "learning_rate": 1.329886777346608e-05, "loss": 0.6549, "step": 7898 }, { "epoch": 1.23, "grad_norm": 21.264327182730028, "learning_rate": 1.3297275731350885e-05, "loss": 0.6392, "step": 7899 }, { "epoch": 1.23, "grad_norm": 14.441937706898234, "learning_rate": 1.3295683595463099e-05, "loss": 0.6144, "step": 7900 }, { "epoch": 1.23, "grad_norm": 19.786126687444444, "learning_rate": 1.3294091365847998e-05, "loss": 0.6627, "step": 7901 }, { "epoch": 1.23, "grad_norm": 24.377213004717987, "learning_rate": 1.3292499042550869e-05, "loss": 0.7348, "step": 7902 }, { "epoch": 1.23, "grad_norm": 23.047238160521076, "learning_rate": 1.3290906625616988e-05, "loss": 0.6707, "step": 7903 }, { "epoch": 1.23, "grad_norm": 24.753436743419407, "learning_rate": 1.3289314115091649e-05, "loss": 0.6353, "step": 7904 }, { "epoch": 1.23, "grad_norm": 18.532643287509494, "learning_rate": 1.3287721511020145e-05, "loss": 0.6173, "step": 7905 }, { "epoch": 1.23, "grad_norm": 18.753650588851634, "learning_rate": 1.3286128813447759e-05, "loss": 0.6767, "step": 7906 }, { "epoch": 1.24, "grad_norm": 27.803578055323058, "learning_rate": 1.3284536022419796e-05, "loss": 0.7903, "step": 7907 }, { "epoch": 1.24, "grad_norm": 14.273160275093604, "learning_rate": 1.3282943137981546e-05, "loss": 0.6562, "step": 7908 }, { "epoch": 1.24, "grad_norm": 18.305561522661897, "learning_rate": 1.3281350160178317e-05, "loss": 0.6933, "step": 7909 }, { "epoch": 1.24, "grad_norm": 15.470618459293842, "learning_rate": 1.3279757089055404e-05, "loss": 0.6052, "step": 7910 }, { "epoch": 1.24, "grad_norm": 20.91177093019644, "learning_rate": 1.3278163924658123e-05, "loss": 0.622, "step": 7911 }, { "epoch": 1.24, "grad_norm": 18.094124083530875, "learning_rate": 1.3276570667031777e-05, "loss": 0.6626, "step": 7912 }, { "epoch": 1.24, "grad_norm": 13.997377877039073, "learning_rate": 1.3274977316221678e-05, "loss": 0.6484, "step": 7913 }, { "epoch": 1.24, "grad_norm": 34.02435372977003, "learning_rate": 1.3273383872273137e-05, "loss": 0.7073, "step": 7914 }, { "epoch": 1.24, "grad_norm": 12.765552197205572, "learning_rate": 1.3271790335231476e-05, "loss": 0.6007, "step": 7915 }, { "epoch": 1.24, "grad_norm": 10.841112129305039, "learning_rate": 1.327019670514201e-05, "loss": 0.6273, "step": 7916 }, { "epoch": 1.24, "grad_norm": 16.69000467959322, "learning_rate": 1.3268602982050064e-05, "loss": 0.611, "step": 7917 }, { "epoch": 1.24, "grad_norm": 19.28029239987682, "learning_rate": 1.3267009166000961e-05, "loss": 0.7759, "step": 7918 }, { "epoch": 1.24, "grad_norm": 20.22753663293937, "learning_rate": 1.3265415257040031e-05, "loss": 0.7013, "step": 7919 }, { "epoch": 1.24, "grad_norm": 13.891110487103646, "learning_rate": 1.3263821255212598e-05, "loss": 0.5992, "step": 7920 }, { "epoch": 1.24, "grad_norm": 12.740556329509829, "learning_rate": 1.3262227160564001e-05, "loss": 0.588, "step": 7921 }, { "epoch": 1.24, "grad_norm": 15.409217790425616, "learning_rate": 1.3260632973139569e-05, "loss": 0.5627, "step": 7922 }, { "epoch": 1.24, "grad_norm": 32.057686791152825, "learning_rate": 1.3259038692984643e-05, "loss": 0.7109, "step": 7923 }, { "epoch": 1.24, "grad_norm": 23.50835141204126, "learning_rate": 1.3257444320144566e-05, "loss": 0.6567, "step": 7924 }, { "epoch": 1.24, "grad_norm": 18.554456563616693, "learning_rate": 1.3255849854664674e-05, "loss": 0.6055, "step": 7925 }, { "epoch": 1.24, "grad_norm": 20.78730080790689, "learning_rate": 1.325425529659032e-05, "loss": 0.6669, "step": 7926 }, { "epoch": 1.24, "grad_norm": 26.201441033424327, "learning_rate": 1.3252660645966846e-05, "loss": 0.6913, "step": 7927 }, { "epoch": 1.24, "grad_norm": 15.291437704780622, "learning_rate": 1.3251065902839608e-05, "loss": 0.5889, "step": 7928 }, { "epoch": 1.24, "grad_norm": 14.25152501407792, "learning_rate": 1.3249471067253957e-05, "loss": 0.6122, "step": 7929 }, { "epoch": 1.24, "grad_norm": 18.182677067842906, "learning_rate": 1.3247876139255249e-05, "loss": 0.666, "step": 7930 }, { "epoch": 1.24, "grad_norm": 28.480784499600276, "learning_rate": 1.3246281118888843e-05, "loss": 0.6199, "step": 7931 }, { "epoch": 1.24, "grad_norm": 18.85401371600529, "learning_rate": 1.3244686006200105e-05, "loss": 0.7349, "step": 7932 }, { "epoch": 1.24, "grad_norm": 13.55603224050496, "learning_rate": 1.324309080123439e-05, "loss": 0.6342, "step": 7933 }, { "epoch": 1.24, "grad_norm": 14.490307000093008, "learning_rate": 1.3241495504037074e-05, "loss": 0.6602, "step": 7934 }, { "epoch": 1.24, "grad_norm": 19.061803130403938, "learning_rate": 1.323990011465352e-05, "loss": 0.6368, "step": 7935 }, { "epoch": 1.24, "grad_norm": 21.654416692014685, "learning_rate": 1.3238304633129102e-05, "loss": 0.6791, "step": 7936 }, { "epoch": 1.24, "grad_norm": 15.013629017532137, "learning_rate": 1.3236709059509194e-05, "loss": 0.6531, "step": 7937 }, { "epoch": 1.24, "grad_norm": 17.749109220651647, "learning_rate": 1.3235113393839174e-05, "loss": 0.6401, "step": 7938 }, { "epoch": 1.24, "grad_norm": 13.597698698166603, "learning_rate": 1.3233517636164423e-05, "loss": 0.5637, "step": 7939 }, { "epoch": 1.24, "grad_norm": 19.845483500689486, "learning_rate": 1.3231921786530323e-05, "loss": 0.6072, "step": 7940 }, { "epoch": 1.24, "grad_norm": 23.05340353228807, "learning_rate": 1.3230325844982254e-05, "loss": 0.6468, "step": 7941 }, { "epoch": 1.24, "grad_norm": 20.41042156977892, "learning_rate": 1.3228729811565608e-05, "loss": 0.6402, "step": 7942 }, { "epoch": 1.24, "grad_norm": 19.345713386359616, "learning_rate": 1.3227133686325777e-05, "loss": 0.6406, "step": 7943 }, { "epoch": 1.24, "grad_norm": 17.9661174778001, "learning_rate": 1.3225537469308152e-05, "loss": 0.8235, "step": 7944 }, { "epoch": 1.24, "grad_norm": 21.523520140653662, "learning_rate": 1.3223941160558127e-05, "loss": 0.5726, "step": 7945 }, { "epoch": 1.24, "grad_norm": 18.432950747673917, "learning_rate": 1.3222344760121105e-05, "loss": 0.6464, "step": 7946 }, { "epoch": 1.24, "grad_norm": 22.092475980503007, "learning_rate": 1.3220748268042478e-05, "loss": 0.7446, "step": 7947 }, { "epoch": 1.24, "grad_norm": 21.004634284806414, "learning_rate": 1.3219151684367656e-05, "loss": 0.6777, "step": 7948 }, { "epoch": 1.24, "grad_norm": 14.929152988664814, "learning_rate": 1.3217555009142044e-05, "loss": 0.602, "step": 7949 }, { "epoch": 1.24, "grad_norm": 10.8817185716788, "learning_rate": 1.3215958242411048e-05, "loss": 0.7217, "step": 7950 }, { "epoch": 1.24, "grad_norm": 15.319623139418203, "learning_rate": 1.3214361384220084e-05, "loss": 0.6298, "step": 7951 }, { "epoch": 1.24, "grad_norm": 21.567336582154155, "learning_rate": 1.321276443461456e-05, "loss": 0.7221, "step": 7952 }, { "epoch": 1.24, "grad_norm": 29.23238888410719, "learning_rate": 1.3211167393639898e-05, "loss": 0.6523, "step": 7953 }, { "epoch": 1.24, "grad_norm": 22.35537994463111, "learning_rate": 1.320957026134151e-05, "loss": 0.6673, "step": 7954 }, { "epoch": 1.24, "grad_norm": 19.716767447961214, "learning_rate": 1.3207973037764822e-05, "loss": 0.6886, "step": 7955 }, { "epoch": 1.24, "grad_norm": 13.993860901341591, "learning_rate": 1.3206375722955261e-05, "loss": 0.6252, "step": 7956 }, { "epoch": 1.24, "grad_norm": 34.21098975756675, "learning_rate": 1.3204778316958247e-05, "loss": 0.6027, "step": 7957 }, { "epoch": 1.24, "grad_norm": 11.75832085191696, "learning_rate": 1.3203180819819213e-05, "loss": 0.6426, "step": 7958 }, { "epoch": 1.24, "grad_norm": 14.815084283267451, "learning_rate": 1.320158323158359e-05, "loss": 0.6695, "step": 7959 }, { "epoch": 1.24, "grad_norm": 11.100789063992094, "learning_rate": 1.319998555229681e-05, "loss": 0.603, "step": 7960 }, { "epoch": 1.24, "grad_norm": 25.813543146220944, "learning_rate": 1.3198387782004315e-05, "loss": 0.7146, "step": 7961 }, { "epoch": 1.24, "grad_norm": 13.048277472308529, "learning_rate": 1.3196789920751546e-05, "loss": 0.6135, "step": 7962 }, { "epoch": 1.24, "grad_norm": 25.871516541486958, "learning_rate": 1.3195191968583938e-05, "loss": 0.689, "step": 7963 }, { "epoch": 1.24, "grad_norm": 20.541956828209123, "learning_rate": 1.3193593925546941e-05, "loss": 0.5964, "step": 7964 }, { "epoch": 1.24, "grad_norm": 24.38098700551929, "learning_rate": 1.3191995791686002e-05, "loss": 0.6473, "step": 7965 }, { "epoch": 1.24, "grad_norm": 18.753004758398816, "learning_rate": 1.3190397567046569e-05, "loss": 0.6561, "step": 7966 }, { "epoch": 1.24, "grad_norm": 19.73712064953473, "learning_rate": 1.3188799251674092e-05, "loss": 0.6794, "step": 7967 }, { "epoch": 1.24, "grad_norm": 16.066982157870687, "learning_rate": 1.3187200845614035e-05, "loss": 0.6959, "step": 7968 }, { "epoch": 1.24, "grad_norm": 19.87475163601495, "learning_rate": 1.3185602348911846e-05, "loss": 0.5211, "step": 7969 }, { "epoch": 1.24, "grad_norm": 15.02567790223042, "learning_rate": 1.318400376161299e-05, "loss": 0.5565, "step": 7970 }, { "epoch": 1.25, "grad_norm": 20.72316915540688, "learning_rate": 1.3182405083762935e-05, "loss": 0.6656, "step": 7971 }, { "epoch": 1.25, "grad_norm": 22.445719403217705, "learning_rate": 1.3180806315407137e-05, "loss": 0.6301, "step": 7972 }, { "epoch": 1.25, "grad_norm": 13.544344864032784, "learning_rate": 1.3179207456591067e-05, "loss": 0.6285, "step": 7973 }, { "epoch": 1.25, "grad_norm": 19.56743709737352, "learning_rate": 1.31776085073602e-05, "loss": 0.6398, "step": 7974 }, { "epoch": 1.25, "grad_norm": 20.49886792134699, "learning_rate": 1.3176009467760005e-05, "loss": 0.6281, "step": 7975 }, { "epoch": 1.25, "grad_norm": 18.215587319370943, "learning_rate": 1.3174410337835956e-05, "loss": 0.6139, "step": 7976 }, { "epoch": 1.25, "grad_norm": 16.98417351690066, "learning_rate": 1.3172811117633536e-05, "loss": 0.622, "step": 7977 }, { "epoch": 1.25, "grad_norm": 18.407426807415938, "learning_rate": 1.3171211807198225e-05, "loss": 0.6029, "step": 7978 }, { "epoch": 1.25, "grad_norm": 14.876736337735665, "learning_rate": 1.3169612406575501e-05, "loss": 0.5869, "step": 7979 }, { "epoch": 1.25, "grad_norm": 18.51903827206621, "learning_rate": 1.316801291581086e-05, "loss": 0.6657, "step": 7980 }, { "epoch": 1.25, "grad_norm": 15.993202123709066, "learning_rate": 1.316641333494978e-05, "loss": 0.622, "step": 7981 }, { "epoch": 1.25, "grad_norm": 16.914745675136928, "learning_rate": 1.316481366403776e-05, "loss": 0.6857, "step": 7982 }, { "epoch": 1.25, "grad_norm": 14.067066034141959, "learning_rate": 1.316321390312029e-05, "loss": 0.5794, "step": 7983 }, { "epoch": 1.25, "grad_norm": 21.279088552644822, "learning_rate": 1.3161614052242869e-05, "loss": 0.644, "step": 7984 }, { "epoch": 1.25, "grad_norm": 22.278151384964875, "learning_rate": 1.3160014111450989e-05, "loss": 0.6082, "step": 7985 }, { "epoch": 1.25, "grad_norm": 24.751250024738408, "learning_rate": 1.3158414080790158e-05, "loss": 0.6466, "step": 7986 }, { "epoch": 1.25, "grad_norm": 25.5915203357572, "learning_rate": 1.315681396030588e-05, "loss": 0.6963, "step": 7987 }, { "epoch": 1.25, "grad_norm": 14.11976135716292, "learning_rate": 1.3155213750043658e-05, "loss": 0.6684, "step": 7988 }, { "epoch": 1.25, "grad_norm": 20.56733587035152, "learning_rate": 1.3153613450049006e-05, "loss": 0.6152, "step": 7989 }, { "epoch": 1.25, "grad_norm": 16.830153845485786, "learning_rate": 1.3152013060367428e-05, "loss": 0.8028, "step": 7990 }, { "epoch": 1.25, "grad_norm": 20.26611231509897, "learning_rate": 1.3150412581044446e-05, "loss": 0.672, "step": 7991 }, { "epoch": 1.25, "grad_norm": 14.139062711391306, "learning_rate": 1.314881201212557e-05, "loss": 0.6382, "step": 7992 }, { "epoch": 1.25, "grad_norm": 12.39103672851601, "learning_rate": 1.3147211353656327e-05, "loss": 0.5999, "step": 7993 }, { "epoch": 1.25, "grad_norm": 21.83432799329184, "learning_rate": 1.314561060568223e-05, "loss": 0.6059, "step": 7994 }, { "epoch": 1.25, "grad_norm": 16.23246007104005, "learning_rate": 1.3144009768248808e-05, "loss": 0.655, "step": 7995 }, { "epoch": 1.25, "grad_norm": 18.06712140910257, "learning_rate": 1.3142408841401586e-05, "loss": 0.6001, "step": 7996 }, { "epoch": 1.25, "grad_norm": 15.775752023871869, "learning_rate": 1.3140807825186098e-05, "loss": 0.5959, "step": 7997 }, { "epoch": 1.25, "grad_norm": 25.43728491038459, "learning_rate": 1.313920671964787e-05, "loss": 0.6157, "step": 7998 }, { "epoch": 1.25, "grad_norm": 18.1862003240563, "learning_rate": 1.3137605524832442e-05, "loss": 0.7027, "step": 7999 }, { "epoch": 1.25, "grad_norm": 17.532901553101294, "learning_rate": 1.3136004240785348e-05, "loss": 0.6413, "step": 8000 }, { "epoch": 1.25, "grad_norm": 14.918692224404275, "learning_rate": 1.3134402867552125e-05, "loss": 0.6516, "step": 8001 }, { "epoch": 1.25, "grad_norm": 26.022916189083826, "learning_rate": 1.3132801405178322e-05, "loss": 0.7436, "step": 8002 }, { "epoch": 1.25, "grad_norm": 28.172128322898967, "learning_rate": 1.3131199853709477e-05, "loss": 0.6289, "step": 8003 }, { "epoch": 1.25, "grad_norm": 33.08401007639743, "learning_rate": 1.3129598213191139e-05, "loss": 0.6346, "step": 8004 }, { "epoch": 1.25, "grad_norm": 22.320825942732977, "learning_rate": 1.3127996483668862e-05, "loss": 0.6322, "step": 8005 }, { "epoch": 1.25, "grad_norm": 13.723695398040176, "learning_rate": 1.3126394665188191e-05, "loss": 0.5866, "step": 8006 }, { "epoch": 1.25, "grad_norm": 19.3880531225369, "learning_rate": 1.3124792757794687e-05, "loss": 0.6105, "step": 8007 }, { "epoch": 1.25, "grad_norm": 18.773889711396897, "learning_rate": 1.3123190761533903e-05, "loss": 0.6621, "step": 8008 }, { "epoch": 1.25, "grad_norm": 19.007287362743273, "learning_rate": 1.3121588676451405e-05, "loss": 0.6335, "step": 8009 }, { "epoch": 1.25, "grad_norm": 22.835750770914707, "learning_rate": 1.3119986502592747e-05, "loss": 0.8118, "step": 8010 }, { "epoch": 1.25, "grad_norm": 16.21119332996405, "learning_rate": 1.3118384240003498e-05, "loss": 0.6193, "step": 8011 }, { "epoch": 1.25, "grad_norm": 16.051947208073248, "learning_rate": 1.311678188872923e-05, "loss": 0.6799, "step": 8012 }, { "epoch": 1.25, "grad_norm": 22.9895501395088, "learning_rate": 1.3115179448815503e-05, "loss": 0.7574, "step": 8013 }, { "epoch": 1.25, "grad_norm": 15.752326133175949, "learning_rate": 1.3113576920307896e-05, "loss": 0.6685, "step": 8014 }, { "epoch": 1.25, "grad_norm": 29.550321243598614, "learning_rate": 1.3111974303251984e-05, "loss": 0.6871, "step": 8015 }, { "epoch": 1.25, "grad_norm": 17.699380170060376, "learning_rate": 1.3110371597693344e-05, "loss": 0.5708, "step": 8016 }, { "epoch": 1.25, "grad_norm": 21.613952519430622, "learning_rate": 1.3108768803677554e-05, "loss": 0.6425, "step": 8017 }, { "epoch": 1.25, "grad_norm": 17.555677728944183, "learning_rate": 1.3107165921250197e-05, "loss": 0.7518, "step": 8018 }, { "epoch": 1.25, "grad_norm": 22.73028546974145, "learning_rate": 1.3105562950456857e-05, "loss": 0.6634, "step": 8019 }, { "epoch": 1.25, "grad_norm": 16.786130834689537, "learning_rate": 1.310395989134313e-05, "loss": 0.702, "step": 8020 }, { "epoch": 1.25, "grad_norm": 20.777253248459047, "learning_rate": 1.3102356743954594e-05, "loss": 0.6694, "step": 8021 }, { "epoch": 1.25, "grad_norm": 23.333587093903443, "learning_rate": 1.3100753508336852e-05, "loss": 0.6967, "step": 8022 }, { "epoch": 1.25, "grad_norm": 22.969389769343948, "learning_rate": 1.3099150184535487e-05, "loss": 0.6445, "step": 8023 }, { "epoch": 1.25, "grad_norm": 23.006206591118794, "learning_rate": 1.309754677259611e-05, "loss": 0.6807, "step": 8024 }, { "epoch": 1.25, "grad_norm": 20.049078108190155, "learning_rate": 1.3095943272564312e-05, "loss": 0.618, "step": 8025 }, { "epoch": 1.25, "grad_norm": 18.950898794490673, "learning_rate": 1.3094339684485698e-05, "loss": 0.5541, "step": 8026 }, { "epoch": 1.25, "grad_norm": 13.673019773239924, "learning_rate": 1.3092736008405875e-05, "loss": 0.6066, "step": 8027 }, { "epoch": 1.25, "grad_norm": 15.261889639083932, "learning_rate": 1.3091132244370449e-05, "loss": 0.5696, "step": 8028 }, { "epoch": 1.25, "grad_norm": 16.930248221286238, "learning_rate": 1.3089528392425029e-05, "loss": 0.581, "step": 8029 }, { "epoch": 1.25, "grad_norm": 19.690711249989143, "learning_rate": 1.3087924452615228e-05, "loss": 0.7141, "step": 8030 }, { "epoch": 1.25, "grad_norm": 34.32094288704457, "learning_rate": 1.3086320424986662e-05, "loss": 0.6354, "step": 8031 }, { "epoch": 1.25, "grad_norm": 20.845228194962765, "learning_rate": 1.308471630958495e-05, "loss": 0.6104, "step": 8032 }, { "epoch": 1.25, "grad_norm": 20.53215942226686, "learning_rate": 1.308311210645571e-05, "loss": 0.7366, "step": 8033 }, { "epoch": 1.25, "grad_norm": 16.13644579044986, "learning_rate": 1.3081507815644567e-05, "loss": 0.7059, "step": 8034 }, { "epoch": 1.26, "grad_norm": 16.96785430537651, "learning_rate": 1.3079903437197143e-05, "loss": 0.6275, "step": 8035 }, { "epoch": 1.26, "grad_norm": 20.643482824029313, "learning_rate": 1.3078298971159067e-05, "loss": 0.6371, "step": 8036 }, { "epoch": 1.26, "grad_norm": 16.96079941038394, "learning_rate": 1.307669441757597e-05, "loss": 0.7108, "step": 8037 }, { "epoch": 1.26, "grad_norm": 18.606243149464927, "learning_rate": 1.3075089776493481e-05, "loss": 0.6766, "step": 8038 }, { "epoch": 1.26, "grad_norm": 22.976607181592534, "learning_rate": 1.3073485047957242e-05, "loss": 0.753, "step": 8039 }, { "epoch": 1.26, "grad_norm": 16.59863223795007, "learning_rate": 1.3071880232012881e-05, "loss": 0.7284, "step": 8040 }, { "epoch": 1.26, "grad_norm": 26.25509165928565, "learning_rate": 1.3070275328706047e-05, "loss": 0.6254, "step": 8041 }, { "epoch": 1.26, "grad_norm": 18.182875356608985, "learning_rate": 1.306867033808238e-05, "loss": 0.591, "step": 8042 }, { "epoch": 1.26, "grad_norm": 27.29877226011105, "learning_rate": 1.306706526018752e-05, "loss": 0.6691, "step": 8043 }, { "epoch": 1.26, "grad_norm": 19.16888109006233, "learning_rate": 1.306546009506712e-05, "loss": 0.6923, "step": 8044 }, { "epoch": 1.26, "grad_norm": 16.08747598980473, "learning_rate": 1.3063854842766828e-05, "loss": 0.6637, "step": 8045 }, { "epoch": 1.26, "grad_norm": 28.459688069690944, "learning_rate": 1.3062249503332294e-05, "loss": 0.7626, "step": 8046 }, { "epoch": 1.26, "grad_norm": 18.987700725973603, "learning_rate": 1.3060644076809178e-05, "loss": 0.689, "step": 8047 }, { "epoch": 1.26, "grad_norm": 17.207565630561497, "learning_rate": 1.3059038563243134e-05, "loss": 0.6706, "step": 8048 }, { "epoch": 1.26, "grad_norm": 18.32752337163448, "learning_rate": 1.3057432962679823e-05, "loss": 0.634, "step": 8049 }, { "epoch": 1.26, "grad_norm": 17.357077477434743, "learning_rate": 1.3055827275164908e-05, "loss": 0.6368, "step": 8050 }, { "epoch": 1.26, "grad_norm": 14.405643486569828, "learning_rate": 1.305422150074405e-05, "loss": 0.5992, "step": 8051 }, { "epoch": 1.26, "grad_norm": 18.089647997557105, "learning_rate": 1.3052615639462923e-05, "loss": 0.6943, "step": 8052 }, { "epoch": 1.26, "grad_norm": 23.605909577930653, "learning_rate": 1.3051009691367191e-05, "loss": 0.6964, "step": 8053 }, { "epoch": 1.26, "grad_norm": 24.936611174038248, "learning_rate": 1.304940365650253e-05, "loss": 0.6061, "step": 8054 }, { "epoch": 1.26, "grad_norm": 17.87232370907216, "learning_rate": 1.3047797534914609e-05, "loss": 0.6753, "step": 8055 }, { "epoch": 1.26, "grad_norm": 15.715961010743476, "learning_rate": 1.3046191326649114e-05, "loss": 0.7218, "step": 8056 }, { "epoch": 1.26, "grad_norm": 16.881813148601875, "learning_rate": 1.3044585031751716e-05, "loss": 0.6388, "step": 8057 }, { "epoch": 1.26, "grad_norm": 19.289576357327363, "learning_rate": 1.3042978650268102e-05, "loss": 0.6415, "step": 8058 }, { "epoch": 1.26, "grad_norm": 30.86277932209255, "learning_rate": 1.3041372182243955e-05, "loss": 0.6875, "step": 8059 }, { "epoch": 1.26, "grad_norm": 17.82818851325902, "learning_rate": 1.3039765627724961e-05, "loss": 0.6484, "step": 8060 }, { "epoch": 1.26, "grad_norm": 12.779006779809293, "learning_rate": 1.3038158986756812e-05, "loss": 0.6049, "step": 8061 }, { "epoch": 1.26, "grad_norm": 38.29816063334219, "learning_rate": 1.30365522593852e-05, "loss": 0.687, "step": 8062 }, { "epoch": 1.26, "grad_norm": 23.63552919362852, "learning_rate": 1.3034945445655813e-05, "loss": 0.5901, "step": 8063 }, { "epoch": 1.26, "grad_norm": 15.039137659256628, "learning_rate": 1.303333854561436e-05, "loss": 0.6291, "step": 8064 }, { "epoch": 1.26, "grad_norm": 19.518064989659592, "learning_rate": 1.3031731559306528e-05, "loss": 0.6836, "step": 8065 }, { "epoch": 1.26, "grad_norm": 19.162309839383838, "learning_rate": 1.3030124486778029e-05, "loss": 0.6957, "step": 8066 }, { "epoch": 1.26, "grad_norm": 13.673784059481592, "learning_rate": 1.3028517328074557e-05, "loss": 0.5744, "step": 8067 }, { "epoch": 1.26, "grad_norm": 19.798210789957974, "learning_rate": 1.3026910083241825e-05, "loss": 0.6016, "step": 8068 }, { "epoch": 1.26, "grad_norm": 20.60857404840023, "learning_rate": 1.3025302752325544e-05, "loss": 0.68, "step": 8069 }, { "epoch": 1.26, "grad_norm": 23.599724971756796, "learning_rate": 1.302369533537142e-05, "loss": 0.6848, "step": 8070 }, { "epoch": 1.26, "grad_norm": 29.297590207334487, "learning_rate": 1.302208783242517e-05, "loss": 0.7169, "step": 8071 }, { "epoch": 1.26, "grad_norm": 19.50109058549094, "learning_rate": 1.302048024353251e-05, "loss": 0.6008, "step": 8072 }, { "epoch": 1.26, "grad_norm": 13.230349797811378, "learning_rate": 1.3018872568739158e-05, "loss": 0.5577, "step": 8073 }, { "epoch": 1.26, "grad_norm": 13.740862755684619, "learning_rate": 1.3017264808090834e-05, "loss": 0.6639, "step": 8074 }, { "epoch": 1.26, "grad_norm": 16.590514992000543, "learning_rate": 1.3015656961633267e-05, "loss": 0.6889, "step": 8075 }, { "epoch": 1.26, "grad_norm": 12.703502957913768, "learning_rate": 1.301404902941218e-05, "loss": 0.6794, "step": 8076 }, { "epoch": 1.26, "grad_norm": 13.48843429438833, "learning_rate": 1.3012441011473302e-05, "loss": 0.6521, "step": 8077 }, { "epoch": 1.26, "grad_norm": 25.176779369281245, "learning_rate": 1.3010832907862363e-05, "loss": 0.6505, "step": 8078 }, { "epoch": 1.26, "grad_norm": 14.06897907515434, "learning_rate": 1.30092247186251e-05, "loss": 0.607, "step": 8079 }, { "epoch": 1.26, "grad_norm": 29.02435386207123, "learning_rate": 1.300761644380724e-05, "loss": 0.6733, "step": 8080 }, { "epoch": 1.26, "grad_norm": 17.404721088481462, "learning_rate": 1.3006008083454535e-05, "loss": 0.6356, "step": 8081 }, { "epoch": 1.26, "grad_norm": 17.623189423948606, "learning_rate": 1.3004399637612712e-05, "loss": 0.5975, "step": 8082 }, { "epoch": 1.26, "grad_norm": 14.212860929356678, "learning_rate": 1.3002791106327528e-05, "loss": 0.6029, "step": 8083 }, { "epoch": 1.26, "grad_norm": 14.77139714966941, "learning_rate": 1.3001182489644715e-05, "loss": 0.5814, "step": 8084 }, { "epoch": 1.26, "grad_norm": 10.927539945752825, "learning_rate": 1.2999573787610031e-05, "loss": 0.6126, "step": 8085 }, { "epoch": 1.26, "grad_norm": 15.795299804758624, "learning_rate": 1.2997965000269223e-05, "loss": 0.5802, "step": 8086 }, { "epoch": 1.26, "grad_norm": 15.780677363893393, "learning_rate": 1.2996356127668044e-05, "loss": 0.6255, "step": 8087 }, { "epoch": 1.26, "grad_norm": 22.077401849251853, "learning_rate": 1.2994747169852249e-05, "loss": 0.6261, "step": 8088 }, { "epoch": 1.26, "grad_norm": 19.559667540661923, "learning_rate": 1.2993138126867597e-05, "loss": 0.6299, "step": 8089 }, { "epoch": 1.26, "grad_norm": 28.457817103440483, "learning_rate": 1.2991528998759844e-05, "loss": 0.6094, "step": 8090 }, { "epoch": 1.26, "grad_norm": 27.339257345874962, "learning_rate": 1.2989919785574764e-05, "loss": 0.7234, "step": 8091 }, { "epoch": 1.26, "grad_norm": 19.6435415185879, "learning_rate": 1.2988310487358107e-05, "loss": 0.6359, "step": 8092 }, { "epoch": 1.26, "grad_norm": 23.987047190713486, "learning_rate": 1.298670110415565e-05, "loss": 0.6762, "step": 8093 }, { "epoch": 1.26, "grad_norm": 18.850866606065082, "learning_rate": 1.2985091636013163e-05, "loss": 0.6225, "step": 8094 }, { "epoch": 1.26, "grad_norm": 20.571227599151083, "learning_rate": 1.2983482082976412e-05, "loss": 0.6764, "step": 8095 }, { "epoch": 1.26, "grad_norm": 38.67571502732276, "learning_rate": 1.2981872445091179e-05, "loss": 0.7227, "step": 8096 }, { "epoch": 1.26, "grad_norm": 21.467369648786548, "learning_rate": 1.2980262722403238e-05, "loss": 0.6298, "step": 8097 }, { "epoch": 1.26, "grad_norm": 16.703392703334963, "learning_rate": 1.297865291495837e-05, "loss": 0.6477, "step": 8098 }, { "epoch": 1.27, "grad_norm": 18.284576594262212, "learning_rate": 1.2977043022802351e-05, "loss": 0.6747, "step": 8099 }, { "epoch": 1.27, "grad_norm": 19.193384234372612, "learning_rate": 1.2975433045980975e-05, "loss": 0.6105, "step": 8100 }, { "epoch": 1.27, "grad_norm": 15.67586365123682, "learning_rate": 1.2973822984540018e-05, "loss": 0.6995, "step": 8101 }, { "epoch": 1.27, "grad_norm": 17.10651846798882, "learning_rate": 1.2972212838525278e-05, "loss": 0.6199, "step": 8102 }, { "epoch": 1.27, "grad_norm": 15.45786557803001, "learning_rate": 1.2970602607982546e-05, "loss": 0.7185, "step": 8103 }, { "epoch": 1.27, "grad_norm": 18.34438554224968, "learning_rate": 1.2968992292957609e-05, "loss": 0.5187, "step": 8104 }, { "epoch": 1.27, "grad_norm": 21.936576690697496, "learning_rate": 1.296738189349627e-05, "loss": 0.6408, "step": 8105 }, { "epoch": 1.27, "grad_norm": 19.64420570169021, "learning_rate": 1.2965771409644328e-05, "loss": 0.6539, "step": 8106 }, { "epoch": 1.27, "grad_norm": 23.30442986456231, "learning_rate": 1.2964160841447577e-05, "loss": 0.618, "step": 8107 }, { "epoch": 1.27, "grad_norm": 26.527710628823286, "learning_rate": 1.2962550188951832e-05, "loss": 0.6227, "step": 8108 }, { "epoch": 1.27, "grad_norm": 20.65186116964494, "learning_rate": 1.2960939452202888e-05, "loss": 0.6377, "step": 8109 }, { "epoch": 1.27, "grad_norm": 17.48742288029853, "learning_rate": 1.2959328631246562e-05, "loss": 0.6404, "step": 8110 }, { "epoch": 1.27, "grad_norm": 17.599566165277086, "learning_rate": 1.2957717726128656e-05, "loss": 0.6358, "step": 8111 }, { "epoch": 1.27, "grad_norm": 16.049314629175925, "learning_rate": 1.295610673689499e-05, "loss": 0.721, "step": 8112 }, { "epoch": 1.27, "grad_norm": 14.902807623466806, "learning_rate": 1.295449566359138e-05, "loss": 0.6448, "step": 8113 }, { "epoch": 1.27, "grad_norm": 16.78937737953778, "learning_rate": 1.2952884506263641e-05, "loss": 0.5574, "step": 8114 }, { "epoch": 1.27, "grad_norm": 17.79619761208542, "learning_rate": 1.2951273264957591e-05, "loss": 0.6552, "step": 8115 }, { "epoch": 1.27, "grad_norm": 16.187879162107276, "learning_rate": 1.2949661939719059e-05, "loss": 0.7062, "step": 8116 }, { "epoch": 1.27, "grad_norm": 22.66637270677793, "learning_rate": 1.2948050530593863e-05, "loss": 0.556, "step": 8117 }, { "epoch": 1.27, "grad_norm": 16.678856294276525, "learning_rate": 1.2946439037627838e-05, "loss": 0.6095, "step": 8118 }, { "epoch": 1.27, "grad_norm": 18.89260742317735, "learning_rate": 1.2944827460866809e-05, "loss": 0.6259, "step": 8119 }, { "epoch": 1.27, "grad_norm": 25.78214248384012, "learning_rate": 1.294321580035661e-05, "loss": 0.6526, "step": 8120 }, { "epoch": 1.27, "grad_norm": 16.834424581502493, "learning_rate": 1.2941604056143078e-05, "loss": 0.6312, "step": 8121 }, { "epoch": 1.27, "grad_norm": 24.361240146110788, "learning_rate": 1.2939992228272043e-05, "loss": 0.69, "step": 8122 }, { "epoch": 1.27, "grad_norm": 14.877071231404777, "learning_rate": 1.2938380316789352e-05, "loss": 0.5819, "step": 8123 }, { "epoch": 1.27, "grad_norm": 14.473555748512167, "learning_rate": 1.293676832174084e-05, "loss": 0.6196, "step": 8124 }, { "epoch": 1.27, "grad_norm": 16.73949488865913, "learning_rate": 1.293515624317236e-05, "loss": 0.6337, "step": 8125 }, { "epoch": 1.27, "grad_norm": 15.448364607954518, "learning_rate": 1.293354408112975e-05, "loss": 0.5691, "step": 8126 }, { "epoch": 1.27, "grad_norm": 16.694756018415156, "learning_rate": 1.2931931835658864e-05, "loss": 0.6521, "step": 8127 }, { "epoch": 1.27, "grad_norm": 38.066722819233796, "learning_rate": 1.293031950680555e-05, "loss": 0.737, "step": 8128 }, { "epoch": 1.27, "grad_norm": 21.798866402502114, "learning_rate": 1.2928707094615665e-05, "loss": 0.6981, "step": 8129 }, { "epoch": 1.27, "grad_norm": 18.41769476532043, "learning_rate": 1.2927094599135064e-05, "loss": 0.5406, "step": 8130 }, { "epoch": 1.27, "grad_norm": 19.64114958708201, "learning_rate": 1.2925482020409604e-05, "loss": 0.6444, "step": 8131 }, { "epoch": 1.27, "grad_norm": 15.039166211848157, "learning_rate": 1.2923869358485147e-05, "loss": 0.6988, "step": 8132 }, { "epoch": 1.27, "grad_norm": 17.34886971034679, "learning_rate": 1.2922256613407553e-05, "loss": 0.5802, "step": 8133 }, { "epoch": 1.27, "grad_norm": 15.549011014723588, "learning_rate": 1.292064378522269e-05, "loss": 0.6199, "step": 8134 }, { "epoch": 1.27, "grad_norm": 20.731477949708044, "learning_rate": 1.2919030873976433e-05, "loss": 0.6691, "step": 8135 }, { "epoch": 1.27, "grad_norm": 16.514022325330988, "learning_rate": 1.291741787971464e-05, "loss": 0.6005, "step": 8136 }, { "epoch": 1.27, "grad_norm": 27.042457301323097, "learning_rate": 1.2915804802483192e-05, "loss": 0.8196, "step": 8137 }, { "epoch": 1.27, "grad_norm": 30.26239914656705, "learning_rate": 1.2914191642327958e-05, "loss": 0.6548, "step": 8138 }, { "epoch": 1.27, "grad_norm": 16.61272050246207, "learning_rate": 1.2912578399294822e-05, "loss": 0.5848, "step": 8139 }, { "epoch": 1.27, "grad_norm": 23.194163024471106, "learning_rate": 1.2910965073429657e-05, "loss": 0.6923, "step": 8140 }, { "epoch": 1.27, "grad_norm": 28.114345680387576, "learning_rate": 1.2909351664778352e-05, "loss": 0.689, "step": 8141 }, { "epoch": 1.27, "grad_norm": 18.409550353111786, "learning_rate": 1.2907738173386783e-05, "loss": 0.6354, "step": 8142 }, { "epoch": 1.27, "grad_norm": 17.480181081951525, "learning_rate": 1.2906124599300842e-05, "loss": 0.6481, "step": 8143 }, { "epoch": 1.27, "grad_norm": 22.429377995836642, "learning_rate": 1.2904510942566421e-05, "loss": 0.6804, "step": 8144 }, { "epoch": 1.27, "grad_norm": 15.722834172160264, "learning_rate": 1.2902897203229406e-05, "loss": 0.5646, "step": 8145 }, { "epoch": 1.27, "grad_norm": 22.759729081395175, "learning_rate": 1.2901283381335691e-05, "loss": 0.6659, "step": 8146 }, { "epoch": 1.27, "grad_norm": 21.444978165930042, "learning_rate": 1.2899669476931176e-05, "loss": 0.5811, "step": 8147 }, { "epoch": 1.27, "grad_norm": 21.518306735423433, "learning_rate": 1.2898055490061757e-05, "loss": 0.6521, "step": 8148 }, { "epoch": 1.27, "grad_norm": 18.733359029327055, "learning_rate": 1.2896441420773335e-05, "loss": 0.6341, "step": 8149 }, { "epoch": 1.27, "grad_norm": 11.477306501553883, "learning_rate": 1.2894827269111811e-05, "loss": 0.6213, "step": 8150 }, { "epoch": 1.27, "grad_norm": 17.630328221854942, "learning_rate": 1.2893213035123096e-05, "loss": 0.6439, "step": 8151 }, { "epoch": 1.27, "grad_norm": 25.360006763758854, "learning_rate": 1.2891598718853096e-05, "loss": 0.7555, "step": 8152 }, { "epoch": 1.27, "grad_norm": 13.679010002465652, "learning_rate": 1.2889984320347715e-05, "loss": 0.6289, "step": 8153 }, { "epoch": 1.27, "grad_norm": 19.306412535242888, "learning_rate": 1.2888369839652875e-05, "loss": 0.6526, "step": 8154 }, { "epoch": 1.27, "grad_norm": 17.080063890437046, "learning_rate": 1.2886755276814484e-05, "loss": 0.6543, "step": 8155 }, { "epoch": 1.27, "grad_norm": 18.375367553028894, "learning_rate": 1.2885140631878463e-05, "loss": 0.6382, "step": 8156 }, { "epoch": 1.27, "grad_norm": 16.28122932117928, "learning_rate": 1.2883525904890728e-05, "loss": 0.5818, "step": 8157 }, { "epoch": 1.27, "grad_norm": 22.21215980378753, "learning_rate": 1.2881911095897206e-05, "loss": 0.5374, "step": 8158 }, { "epoch": 1.27, "grad_norm": 18.70605472892297, "learning_rate": 1.2880296204943816e-05, "loss": 0.65, "step": 8159 }, { "epoch": 1.27, "grad_norm": 22.843607854466118, "learning_rate": 1.287868123207649e-05, "loss": 0.6771, "step": 8160 }, { "epoch": 1.27, "grad_norm": 18.619776545342837, "learning_rate": 1.2877066177341148e-05, "loss": 0.6139, "step": 8161 }, { "epoch": 1.27, "grad_norm": 15.732893863656638, "learning_rate": 1.287545104078373e-05, "loss": 0.5971, "step": 8162 }, { "epoch": 1.28, "grad_norm": 22.347713784442327, "learning_rate": 1.2873835822450168e-05, "loss": 0.7457, "step": 8163 }, { "epoch": 1.28, "grad_norm": 19.417072269309283, "learning_rate": 1.2872220522386396e-05, "loss": 0.6432, "step": 8164 }, { "epoch": 1.28, "grad_norm": 12.638863247439584, "learning_rate": 1.2870605140638352e-05, "loss": 0.596, "step": 8165 }, { "epoch": 1.28, "grad_norm": 18.743124205002776, "learning_rate": 1.2868989677251976e-05, "loss": 0.6887, "step": 8166 }, { "epoch": 1.28, "grad_norm": 12.636283051586897, "learning_rate": 1.2867374132273214e-05, "loss": 0.6665, "step": 8167 }, { "epoch": 1.28, "grad_norm": 20.321634925503275, "learning_rate": 1.2865758505748006e-05, "loss": 0.5936, "step": 8168 }, { "epoch": 1.28, "grad_norm": 23.366738158522285, "learning_rate": 1.2864142797722305e-05, "loss": 0.6217, "step": 8169 }, { "epoch": 1.28, "grad_norm": 16.724378762419924, "learning_rate": 1.2862527008242056e-05, "loss": 0.6125, "step": 8170 }, { "epoch": 1.28, "grad_norm": 16.261670719380938, "learning_rate": 1.2860911137353219e-05, "loss": 0.6728, "step": 8171 }, { "epoch": 1.28, "grad_norm": 17.35691244855049, "learning_rate": 1.2859295185101736e-05, "loss": 0.6272, "step": 8172 }, { "epoch": 1.28, "grad_norm": 16.572375423177817, "learning_rate": 1.2857679151533576e-05, "loss": 0.7258, "step": 8173 }, { "epoch": 1.28, "grad_norm": 20.28698023090884, "learning_rate": 1.2856063036694688e-05, "loss": 0.6612, "step": 8174 }, { "epoch": 1.28, "grad_norm": 13.575975106319408, "learning_rate": 1.2854446840631041e-05, "loss": 0.6047, "step": 8175 }, { "epoch": 1.28, "grad_norm": 14.183492852947104, "learning_rate": 1.2852830563388597e-05, "loss": 0.5923, "step": 8176 }, { "epoch": 1.28, "grad_norm": 19.053828227886665, "learning_rate": 1.2851214205013317e-05, "loss": 0.6942, "step": 8177 }, { "epoch": 1.28, "grad_norm": 14.7969470888414, "learning_rate": 1.2849597765551171e-05, "loss": 0.6402, "step": 8178 }, { "epoch": 1.28, "grad_norm": 19.286436343268512, "learning_rate": 1.2847981245048137e-05, "loss": 0.6593, "step": 8179 }, { "epoch": 1.28, "grad_norm": 19.417870226809434, "learning_rate": 1.284636464355018e-05, "loss": 0.6639, "step": 8180 }, { "epoch": 1.28, "grad_norm": 19.517238998406224, "learning_rate": 1.2844747961103279e-05, "loss": 0.6272, "step": 8181 }, { "epoch": 1.28, "grad_norm": 27.43945814826806, "learning_rate": 1.2843131197753408e-05, "loss": 0.6415, "step": 8182 }, { "epoch": 1.28, "grad_norm": 26.910037510156272, "learning_rate": 1.2841514353546549e-05, "loss": 0.6448, "step": 8183 }, { "epoch": 1.28, "grad_norm": 12.863559745081396, "learning_rate": 1.2839897428528687e-05, "loss": 0.594, "step": 8184 }, { "epoch": 1.28, "grad_norm": 19.722278407000786, "learning_rate": 1.2838280422745801e-05, "loss": 0.5537, "step": 8185 }, { "epoch": 1.28, "grad_norm": 23.28998282903744, "learning_rate": 1.283666333624388e-05, "loss": 0.7031, "step": 8186 }, { "epoch": 1.28, "grad_norm": 24.678736125870845, "learning_rate": 1.2835046169068912e-05, "loss": 0.6506, "step": 8187 }, { "epoch": 1.28, "grad_norm": 18.86576640145763, "learning_rate": 1.283342892126689e-05, "loss": 0.6597, "step": 8188 }, { "epoch": 1.28, "grad_norm": 20.68417514707934, "learning_rate": 1.283181159288381e-05, "loss": 0.6305, "step": 8189 }, { "epoch": 1.28, "grad_norm": 14.088556117146169, "learning_rate": 1.2830194183965663e-05, "loss": 0.6077, "step": 8190 }, { "epoch": 1.28, "grad_norm": 12.664025697295775, "learning_rate": 1.2828576694558447e-05, "loss": 0.5667, "step": 8191 }, { "epoch": 1.28, "grad_norm": 24.491983022855344, "learning_rate": 1.2826959124708167e-05, "loss": 0.767, "step": 8192 }, { "epoch": 1.28, "grad_norm": 23.348674336485477, "learning_rate": 1.2825341474460824e-05, "loss": 0.6835, "step": 8193 }, { "epoch": 1.28, "grad_norm": 27.10513878849774, "learning_rate": 1.2823723743862422e-05, "loss": 0.7139, "step": 8194 }, { "epoch": 1.28, "grad_norm": 17.087777241478815, "learning_rate": 1.2822105932958964e-05, "loss": 0.5762, "step": 8195 }, { "epoch": 1.28, "grad_norm": 27.34988598560056, "learning_rate": 1.282048804179647e-05, "loss": 0.5978, "step": 8196 }, { "epoch": 1.28, "grad_norm": 19.475743284566807, "learning_rate": 1.2818870070420942e-05, "loss": 0.6952, "step": 8197 }, { "epoch": 1.28, "grad_norm": 24.90319452575244, "learning_rate": 1.2817252018878401e-05, "loss": 0.6428, "step": 8198 }, { "epoch": 1.28, "grad_norm": 19.821598848943367, "learning_rate": 1.2815633887214857e-05, "loss": 0.6362, "step": 8199 }, { "epoch": 1.28, "grad_norm": 14.267088470829696, "learning_rate": 1.2814015675476336e-05, "loss": 0.6174, "step": 8200 }, { "epoch": 1.28, "grad_norm": 32.68403077269346, "learning_rate": 1.2812397383708854e-05, "loss": 0.6853, "step": 8201 }, { "epoch": 1.28, "grad_norm": 23.39607547817, "learning_rate": 1.2810779011958437e-05, "loss": 0.6587, "step": 8202 }, { "epoch": 1.28, "grad_norm": 14.088205178262257, "learning_rate": 1.2809160560271108e-05, "loss": 0.6183, "step": 8203 }, { "epoch": 1.28, "grad_norm": 20.81991632310064, "learning_rate": 1.2807542028692894e-05, "loss": 0.7255, "step": 8204 }, { "epoch": 1.28, "grad_norm": 17.652369874199216, "learning_rate": 1.2805923417269828e-05, "loss": 0.6284, "step": 8205 }, { "epoch": 1.28, "grad_norm": 18.30337913744467, "learning_rate": 1.2804304726047943e-05, "loss": 0.61, "step": 8206 }, { "epoch": 1.28, "grad_norm": 18.251616359030656, "learning_rate": 1.2802685955073271e-05, "loss": 0.6537, "step": 8207 }, { "epoch": 1.28, "grad_norm": 18.35665278828155, "learning_rate": 1.280106710439185e-05, "loss": 0.645, "step": 8208 }, { "epoch": 1.28, "grad_norm": 14.34360182130856, "learning_rate": 1.279944817404972e-05, "loss": 0.5877, "step": 8209 }, { "epoch": 1.28, "grad_norm": 12.940694352410677, "learning_rate": 1.279782916409292e-05, "loss": 0.6063, "step": 8210 }, { "epoch": 1.28, "grad_norm": 14.1460525655355, "learning_rate": 1.2796210074567499e-05, "loss": 0.6644, "step": 8211 }, { "epoch": 1.28, "grad_norm": 14.391992489499236, "learning_rate": 1.2794590905519494e-05, "loss": 0.5939, "step": 8212 }, { "epoch": 1.28, "grad_norm": 27.558598731004682, "learning_rate": 1.2792971656994963e-05, "loss": 0.6897, "step": 8213 }, { "epoch": 1.28, "grad_norm": 21.736035336015192, "learning_rate": 1.2791352329039948e-05, "loss": 0.5729, "step": 8214 }, { "epoch": 1.28, "grad_norm": 15.35966969126824, "learning_rate": 1.278973292170051e-05, "loss": 0.6441, "step": 8215 }, { "epoch": 1.28, "grad_norm": 26.81296011334332, "learning_rate": 1.2788113435022696e-05, "loss": 0.7154, "step": 8216 }, { "epoch": 1.28, "grad_norm": 12.376780871287782, "learning_rate": 1.2786493869052568e-05, "loss": 0.5998, "step": 8217 }, { "epoch": 1.28, "grad_norm": 29.08429770774486, "learning_rate": 1.2784874223836184e-05, "loss": 0.6733, "step": 8218 }, { "epoch": 1.28, "grad_norm": 10.842001650702121, "learning_rate": 1.2783254499419607e-05, "loss": 0.5726, "step": 8219 }, { "epoch": 1.28, "grad_norm": 18.339599638625465, "learning_rate": 1.2781634695848899e-05, "loss": 0.6483, "step": 8220 }, { "epoch": 1.28, "grad_norm": 24.412345451094303, "learning_rate": 1.278001481317013e-05, "loss": 0.649, "step": 8221 }, { "epoch": 1.28, "grad_norm": 15.097347138905876, "learning_rate": 1.2778394851429361e-05, "loss": 0.5526, "step": 8222 }, { "epoch": 1.28, "grad_norm": 20.73414152023992, "learning_rate": 1.2776774810672674e-05, "loss": 0.632, "step": 8223 }, { "epoch": 1.28, "grad_norm": 15.325303280716902, "learning_rate": 1.277515469094613e-05, "loss": 0.5003, "step": 8224 }, { "epoch": 1.28, "grad_norm": 10.971576566258243, "learning_rate": 1.2773534492295813e-05, "loss": 0.5308, "step": 8225 }, { "epoch": 1.28, "grad_norm": 15.055650524859043, "learning_rate": 1.2771914214767798e-05, "loss": 0.6434, "step": 8226 }, { "epoch": 1.29, "grad_norm": 14.336058272079114, "learning_rate": 1.2770293858408166e-05, "loss": 0.6633, "step": 8227 }, { "epoch": 1.29, "grad_norm": 14.182235426854886, "learning_rate": 1.2768673423262992e-05, "loss": 0.7339, "step": 8228 }, { "epoch": 1.29, "grad_norm": 22.68572989959388, "learning_rate": 1.2767052909378367e-05, "loss": 0.7338, "step": 8229 }, { "epoch": 1.29, "grad_norm": 19.320254660896573, "learning_rate": 1.2765432316800379e-05, "loss": 0.626, "step": 8230 }, { "epoch": 1.29, "grad_norm": 19.825806270231375, "learning_rate": 1.276381164557511e-05, "loss": 0.6617, "step": 8231 }, { "epoch": 1.29, "grad_norm": 23.305797496759737, "learning_rate": 1.2762190895748657e-05, "loss": 0.6881, "step": 8232 }, { "epoch": 1.29, "grad_norm": 12.713697541779094, "learning_rate": 1.2760570067367111e-05, "loss": 0.6689, "step": 8233 }, { "epoch": 1.29, "grad_norm": 18.242289927193006, "learning_rate": 1.2758949160476567e-05, "loss": 0.6251, "step": 8234 }, { "epoch": 1.29, "grad_norm": 19.520502193029696, "learning_rate": 1.2757328175123122e-05, "loss": 0.5961, "step": 8235 }, { "epoch": 1.29, "grad_norm": 23.25639740864642, "learning_rate": 1.2755707111352877e-05, "loss": 0.7622, "step": 8236 }, { "epoch": 1.29, "grad_norm": 16.405445122730182, "learning_rate": 1.2754085969211932e-05, "loss": 0.6287, "step": 8237 }, { "epoch": 1.29, "grad_norm": 21.009959486758213, "learning_rate": 1.2752464748746395e-05, "loss": 0.6654, "step": 8238 }, { "epoch": 1.29, "grad_norm": 29.488844121426467, "learning_rate": 1.275084345000237e-05, "loss": 0.6218, "step": 8239 }, { "epoch": 1.29, "grad_norm": 15.191574689186826, "learning_rate": 1.2749222073025968e-05, "loss": 0.5641, "step": 8240 }, { "epoch": 1.29, "grad_norm": 22.50691927839262, "learning_rate": 1.2747600617863296e-05, "loss": 0.6369, "step": 8241 }, { "epoch": 1.29, "grad_norm": 14.675275036425337, "learning_rate": 1.2745979084560474e-05, "loss": 0.6113, "step": 8242 }, { "epoch": 1.29, "grad_norm": 19.42540606595315, "learning_rate": 1.274435747316361e-05, "loss": 0.7135, "step": 8243 }, { "epoch": 1.29, "grad_norm": 18.502848249262144, "learning_rate": 1.2742735783718825e-05, "loss": 0.6044, "step": 8244 }, { "epoch": 1.29, "grad_norm": 12.234961526703646, "learning_rate": 1.2741114016272239e-05, "loss": 0.581, "step": 8245 }, { "epoch": 1.29, "grad_norm": 20.207085752975964, "learning_rate": 1.2739492170869973e-05, "loss": 0.6448, "step": 8246 }, { "epoch": 1.29, "grad_norm": 12.858919183687561, "learning_rate": 1.2737870247558152e-05, "loss": 0.5923, "step": 8247 }, { "epoch": 1.29, "grad_norm": 15.561342703826355, "learning_rate": 1.2736248246382906e-05, "loss": 0.6666, "step": 8248 }, { "epoch": 1.29, "grad_norm": 12.343705813715111, "learning_rate": 1.2734626167390356e-05, "loss": 0.5232, "step": 8249 }, { "epoch": 1.29, "grad_norm": 17.694151070144503, "learning_rate": 1.2733004010626639e-05, "loss": 0.7189, "step": 8250 }, { "epoch": 1.29, "grad_norm": 14.76889183044537, "learning_rate": 1.273138177613789e-05, "loss": 0.673, "step": 8251 }, { "epoch": 1.29, "grad_norm": 19.191555895339366, "learning_rate": 1.2729759463970238e-05, "loss": 0.6596, "step": 8252 }, { "epoch": 1.29, "grad_norm": 19.413788901767116, "learning_rate": 1.2728137074169824e-05, "loss": 0.622, "step": 8253 }, { "epoch": 1.29, "grad_norm": 21.702806584416553, "learning_rate": 1.2726514606782785e-05, "loss": 0.7032, "step": 8254 }, { "epoch": 1.29, "grad_norm": 19.454562870184347, "learning_rate": 1.2724892061855271e-05, "loss": 0.6099, "step": 8255 }, { "epoch": 1.29, "grad_norm": 16.534468667838016, "learning_rate": 1.2723269439433416e-05, "loss": 0.5489, "step": 8256 }, { "epoch": 1.29, "grad_norm": 15.893511429360114, "learning_rate": 1.2721646739563378e-05, "loss": 0.6107, "step": 8257 }, { "epoch": 1.29, "grad_norm": 25.034105700567117, "learning_rate": 1.2720023962291292e-05, "loss": 0.697, "step": 8258 }, { "epoch": 1.29, "grad_norm": 13.347027819371986, "learning_rate": 1.2718401107663318e-05, "loss": 0.6945, "step": 8259 }, { "epoch": 1.29, "grad_norm": 20.475692892839113, "learning_rate": 1.2716778175725606e-05, "loss": 0.6346, "step": 8260 }, { "epoch": 1.29, "grad_norm": 17.219003174438452, "learning_rate": 1.2715155166524313e-05, "loss": 0.6401, "step": 8261 }, { "epoch": 1.29, "grad_norm": 21.103182003521006, "learning_rate": 1.2713532080105595e-05, "loss": 0.6422, "step": 8262 }, { "epoch": 1.29, "grad_norm": 22.474250747499287, "learning_rate": 1.2711908916515613e-05, "loss": 0.617, "step": 8263 }, { "epoch": 1.29, "grad_norm": 18.380625495399574, "learning_rate": 1.2710285675800527e-05, "loss": 0.5819, "step": 8264 }, { "epoch": 1.29, "grad_norm": 17.53238287748573, "learning_rate": 1.2708662358006504e-05, "loss": 0.5705, "step": 8265 }, { "epoch": 1.29, "grad_norm": 23.17753445404153, "learning_rate": 1.2707038963179704e-05, "loss": 0.717, "step": 8266 }, { "epoch": 1.29, "grad_norm": 17.869701037519455, "learning_rate": 1.2705415491366303e-05, "loss": 0.6855, "step": 8267 }, { "epoch": 1.29, "grad_norm": 17.1292536875374, "learning_rate": 1.2703791942612462e-05, "loss": 0.5989, "step": 8268 }, { "epoch": 1.29, "grad_norm": 15.419536762127363, "learning_rate": 1.2702168316964365e-05, "loss": 0.5862, "step": 8269 }, { "epoch": 1.29, "grad_norm": 20.78543399158514, "learning_rate": 1.2700544614468181e-05, "loss": 0.6635, "step": 8270 }, { "epoch": 1.29, "grad_norm": 18.072822686719643, "learning_rate": 1.2698920835170088e-05, "loss": 0.6921, "step": 8271 }, { "epoch": 1.29, "grad_norm": 21.085295042332895, "learning_rate": 1.2697296979116265e-05, "loss": 0.6539, "step": 8272 }, { "epoch": 1.29, "grad_norm": 34.33598617663096, "learning_rate": 1.2695673046352893e-05, "loss": 0.619, "step": 8273 }, { "epoch": 1.29, "grad_norm": 12.727146128274091, "learning_rate": 1.2694049036926159e-05, "loss": 0.5424, "step": 8274 }, { "epoch": 1.29, "grad_norm": 16.767142420036183, "learning_rate": 1.2692424950882242e-05, "loss": 0.6286, "step": 8275 }, { "epoch": 1.29, "grad_norm": 20.435146818243783, "learning_rate": 1.2690800788267338e-05, "loss": 0.5762, "step": 8276 }, { "epoch": 1.29, "grad_norm": 16.64200408620659, "learning_rate": 1.2689176549127633e-05, "loss": 0.7087, "step": 8277 }, { "epoch": 1.29, "grad_norm": 16.426749474636896, "learning_rate": 1.2687552233509322e-05, "loss": 0.6936, "step": 8278 }, { "epoch": 1.29, "grad_norm": 16.95753997236476, "learning_rate": 1.2685927841458593e-05, "loss": 0.6355, "step": 8279 }, { "epoch": 1.29, "grad_norm": 19.704809227276208, "learning_rate": 1.2684303373021651e-05, "loss": 0.6469, "step": 8280 }, { "epoch": 1.29, "grad_norm": 16.786845307106773, "learning_rate": 1.268267882824469e-05, "loss": 0.6742, "step": 8281 }, { "epoch": 1.29, "grad_norm": 21.523714175237412, "learning_rate": 1.2681054207173915e-05, "loss": 0.599, "step": 8282 }, { "epoch": 1.29, "grad_norm": 19.42877414153336, "learning_rate": 1.2679429509855521e-05, "loss": 0.659, "step": 8283 }, { "epoch": 1.29, "grad_norm": 12.899460323960357, "learning_rate": 1.2677804736335726e-05, "loss": 0.5583, "step": 8284 }, { "epoch": 1.29, "grad_norm": 20.53406389686106, "learning_rate": 1.2676179886660725e-05, "loss": 0.6189, "step": 8285 }, { "epoch": 1.29, "grad_norm": 18.087217950500186, "learning_rate": 1.2674554960876737e-05, "loss": 0.6769, "step": 8286 }, { "epoch": 1.29, "grad_norm": 24.029417775453116, "learning_rate": 1.2672929959029969e-05, "loss": 0.6117, "step": 8287 }, { "epoch": 1.29, "grad_norm": 24.828266838393592, "learning_rate": 1.2671304881166638e-05, "loss": 0.6575, "step": 8288 }, { "epoch": 1.29, "grad_norm": 16.012288101359115, "learning_rate": 1.2669679727332957e-05, "loss": 0.5936, "step": 8289 }, { "epoch": 1.29, "grad_norm": 20.185367081076276, "learning_rate": 1.2668054497575147e-05, "loss": 0.6373, "step": 8290 }, { "epoch": 1.3, "grad_norm": 17.471258515197395, "learning_rate": 1.2666429191939428e-05, "loss": 0.6036, "step": 8291 }, { "epoch": 1.3, "grad_norm": 32.78390001926139, "learning_rate": 1.2664803810472025e-05, "loss": 0.672, "step": 8292 }, { "epoch": 1.3, "grad_norm": 20.973422872550596, "learning_rate": 1.2663178353219154e-05, "loss": 0.6369, "step": 8293 }, { "epoch": 1.3, "grad_norm": 11.30912832403255, "learning_rate": 1.2661552820227053e-05, "loss": 0.6168, "step": 8294 }, { "epoch": 1.3, "grad_norm": 20.21009902999088, "learning_rate": 1.2659927211541946e-05, "loss": 0.6063, "step": 8295 }, { "epoch": 1.3, "grad_norm": 18.766984977642572, "learning_rate": 1.2658301527210066e-05, "loss": 0.6186, "step": 8296 }, { "epoch": 1.3, "grad_norm": 14.254903317822391, "learning_rate": 1.2656675767277647e-05, "loss": 0.5818, "step": 8297 }, { "epoch": 1.3, "grad_norm": 40.45768699840152, "learning_rate": 1.2655049931790919e-05, "loss": 0.6823, "step": 8298 }, { "epoch": 1.3, "grad_norm": 16.93386568161306, "learning_rate": 1.2653424020796128e-05, "loss": 0.5915, "step": 8299 }, { "epoch": 1.3, "grad_norm": 27.277270190650405, "learning_rate": 1.2651798034339506e-05, "loss": 0.6445, "step": 8300 }, { "epoch": 1.3, "grad_norm": 17.89493076950185, "learning_rate": 1.2650171972467304e-05, "loss": 0.6528, "step": 8301 }, { "epoch": 1.3, "grad_norm": 24.165653551312477, "learning_rate": 1.2648545835225758e-05, "loss": 0.6742, "step": 8302 }, { "epoch": 1.3, "grad_norm": 15.142076057643568, "learning_rate": 1.2646919622661119e-05, "loss": 0.6438, "step": 8303 }, { "epoch": 1.3, "grad_norm": 10.194847990832569, "learning_rate": 1.2645293334819635e-05, "loss": 0.5588, "step": 8304 }, { "epoch": 1.3, "grad_norm": 19.47180693523747, "learning_rate": 1.2643666971747555e-05, "loss": 0.6949, "step": 8305 }, { "epoch": 1.3, "grad_norm": 19.225884641457455, "learning_rate": 1.2642040533491133e-05, "loss": 0.6261, "step": 8306 }, { "epoch": 1.3, "grad_norm": 20.36221919133966, "learning_rate": 1.2640414020096626e-05, "loss": 0.5965, "step": 8307 }, { "epoch": 1.3, "grad_norm": 22.204110906216208, "learning_rate": 1.2638787431610285e-05, "loss": 0.6384, "step": 8308 }, { "epoch": 1.3, "grad_norm": 16.597296453442517, "learning_rate": 1.2637160768078376e-05, "loss": 0.6544, "step": 8309 }, { "epoch": 1.3, "grad_norm": 23.160980914446505, "learning_rate": 1.2635534029547155e-05, "loss": 0.6486, "step": 8310 }, { "epoch": 1.3, "grad_norm": 15.00884000496649, "learning_rate": 1.263390721606289e-05, "loss": 0.6161, "step": 8311 }, { "epoch": 1.3, "grad_norm": 23.374112095773256, "learning_rate": 1.2632280327671842e-05, "loss": 0.6349, "step": 8312 }, { "epoch": 1.3, "grad_norm": 19.90759129436889, "learning_rate": 1.2630653364420285e-05, "loss": 0.7162, "step": 8313 }, { "epoch": 1.3, "grad_norm": 28.444149062689792, "learning_rate": 1.2629026326354481e-05, "loss": 0.6268, "step": 8314 }, { "epoch": 1.3, "grad_norm": 16.3915209329876, "learning_rate": 1.262739921352071e-05, "loss": 0.6736, "step": 8315 }, { "epoch": 1.3, "grad_norm": 21.4954904953057, "learning_rate": 1.2625772025965241e-05, "loss": 0.6483, "step": 8316 }, { "epoch": 1.3, "grad_norm": 17.94987849039004, "learning_rate": 1.262414476373435e-05, "loss": 0.644, "step": 8317 }, { "epoch": 1.3, "grad_norm": 19.480263768799166, "learning_rate": 1.2622517426874322e-05, "loss": 0.6107, "step": 8318 }, { "epoch": 1.3, "grad_norm": 13.973122043753486, "learning_rate": 1.2620890015431427e-05, "loss": 0.5114, "step": 8319 }, { "epoch": 1.3, "grad_norm": 22.78530759285365, "learning_rate": 1.2619262529451954e-05, "loss": 0.6009, "step": 8320 }, { "epoch": 1.3, "grad_norm": 16.290259431183483, "learning_rate": 1.2617634968982189e-05, "loss": 0.6141, "step": 8321 }, { "epoch": 1.3, "grad_norm": 10.916252844856091, "learning_rate": 1.2616007334068417e-05, "loss": 0.6312, "step": 8322 }, { "epoch": 1.3, "grad_norm": 21.10366944305978, "learning_rate": 1.2614379624756924e-05, "loss": 0.6477, "step": 8323 }, { "epoch": 1.3, "grad_norm": 21.974929096587086, "learning_rate": 1.2612751841094006e-05, "loss": 0.6621, "step": 8324 }, { "epoch": 1.3, "grad_norm": 44.209934620285786, "learning_rate": 1.2611123983125951e-05, "loss": 0.6666, "step": 8325 }, { "epoch": 1.3, "grad_norm": 16.474935030700554, "learning_rate": 1.2609496050899061e-05, "loss": 0.6136, "step": 8326 }, { "epoch": 1.3, "grad_norm": 20.948634831066823, "learning_rate": 1.2607868044459629e-05, "loss": 0.5512, "step": 8327 }, { "epoch": 1.3, "grad_norm": 18.758153153662324, "learning_rate": 1.2606239963853957e-05, "loss": 0.645, "step": 8328 }, { "epoch": 1.3, "grad_norm": 21.338863175005827, "learning_rate": 1.260461180912834e-05, "loss": 0.666, "step": 8329 }, { "epoch": 1.3, "grad_norm": 21.871535718989445, "learning_rate": 1.260298358032909e-05, "loss": 0.6818, "step": 8330 }, { "epoch": 1.3, "grad_norm": 16.4711589734319, "learning_rate": 1.260135527750251e-05, "loss": 0.6062, "step": 8331 }, { "epoch": 1.3, "grad_norm": 18.597299414938952, "learning_rate": 1.2599726900694909e-05, "loss": 0.6556, "step": 8332 }, { "epoch": 1.3, "grad_norm": 22.01644488202975, "learning_rate": 1.2598098449952594e-05, "loss": 0.6689, "step": 8333 }, { "epoch": 1.3, "grad_norm": 22.478950747518574, "learning_rate": 1.2596469925321877e-05, "loss": 0.6645, "step": 8334 }, { "epoch": 1.3, "grad_norm": 17.898143156471157, "learning_rate": 1.2594841326849075e-05, "loss": 0.671, "step": 8335 }, { "epoch": 1.3, "grad_norm": 14.316397881603008, "learning_rate": 1.2593212654580507e-05, "loss": 0.7539, "step": 8336 }, { "epoch": 1.3, "grad_norm": 12.919082616649789, "learning_rate": 1.2591583908562483e-05, "loss": 0.5255, "step": 8337 }, { "epoch": 1.3, "grad_norm": 43.385797618939, "learning_rate": 1.2589955088841332e-05, "loss": 0.5732, "step": 8338 }, { "epoch": 1.3, "grad_norm": 24.524400495887633, "learning_rate": 1.2588326195463373e-05, "loss": 0.6069, "step": 8339 }, { "epoch": 1.3, "grad_norm": 17.70334773584294, "learning_rate": 1.2586697228474931e-05, "loss": 0.6646, "step": 8340 }, { "epoch": 1.3, "grad_norm": 19.06490269003772, "learning_rate": 1.2585068187922333e-05, "loss": 0.6353, "step": 8341 }, { "epoch": 1.3, "grad_norm": 21.61264025552755, "learning_rate": 1.2583439073851907e-05, "loss": 0.6275, "step": 8342 }, { "epoch": 1.3, "grad_norm": 15.231405004863914, "learning_rate": 1.2581809886309989e-05, "loss": 0.5737, "step": 8343 }, { "epoch": 1.3, "grad_norm": 17.215214434693078, "learning_rate": 1.25801806253429e-05, "loss": 0.5884, "step": 8344 }, { "epoch": 1.3, "grad_norm": 19.457123827986855, "learning_rate": 1.2578551290996993e-05, "loss": 0.6552, "step": 8345 }, { "epoch": 1.3, "grad_norm": 18.390290753966006, "learning_rate": 1.2576921883318589e-05, "loss": 0.6073, "step": 8346 }, { "epoch": 1.3, "grad_norm": 13.580823521516413, "learning_rate": 1.2575292402354036e-05, "loss": 0.6572, "step": 8347 }, { "epoch": 1.3, "grad_norm": 19.342939219638488, "learning_rate": 1.2573662848149674e-05, "loss": 0.6532, "step": 8348 }, { "epoch": 1.3, "grad_norm": 11.704701681590626, "learning_rate": 1.2572033220751844e-05, "loss": 0.6629, "step": 8349 }, { "epoch": 1.3, "grad_norm": 18.744592104506047, "learning_rate": 1.2570403520206898e-05, "loss": 0.6081, "step": 8350 }, { "epoch": 1.3, "grad_norm": 21.026847720474173, "learning_rate": 1.2568773746561179e-05, "loss": 0.5171, "step": 8351 }, { "epoch": 1.3, "grad_norm": 15.078173749479243, "learning_rate": 1.2567143899861031e-05, "loss": 0.6911, "step": 8352 }, { "epoch": 1.3, "grad_norm": 15.63746392616011, "learning_rate": 1.2565513980152822e-05, "loss": 0.6007, "step": 8353 }, { "epoch": 1.3, "grad_norm": 17.722548822687365, "learning_rate": 1.2563883987482887e-05, "loss": 0.5726, "step": 8354 }, { "epoch": 1.31, "grad_norm": 18.49407482065009, "learning_rate": 1.2562253921897598e-05, "loss": 0.6477, "step": 8355 }, { "epoch": 1.31, "grad_norm": 14.053773822740483, "learning_rate": 1.2560623783443302e-05, "loss": 0.6122, "step": 8356 }, { "epoch": 1.31, "grad_norm": 22.992982872981976, "learning_rate": 1.2558993572166365e-05, "loss": 0.6202, "step": 8357 }, { "epoch": 1.31, "grad_norm": 21.554575835820508, "learning_rate": 1.255736328811315e-05, "loss": 0.6036, "step": 8358 }, { "epoch": 1.31, "grad_norm": 28.007069121355418, "learning_rate": 1.2555732931330015e-05, "loss": 0.6098, "step": 8359 }, { "epoch": 1.31, "grad_norm": 16.900518265884386, "learning_rate": 1.2554102501863333e-05, "loss": 0.5865, "step": 8360 }, { "epoch": 1.31, "grad_norm": 17.19558522659384, "learning_rate": 1.2552471999759469e-05, "loss": 0.5392, "step": 8361 }, { "epoch": 1.31, "grad_norm": 16.110822537651774, "learning_rate": 1.2550841425064795e-05, "loss": 0.5385, "step": 8362 }, { "epoch": 1.31, "grad_norm": 13.938318039318656, "learning_rate": 1.254921077782568e-05, "loss": 0.5992, "step": 8363 }, { "epoch": 1.31, "grad_norm": 22.815408864791664, "learning_rate": 1.2547580058088507e-05, "loss": 0.7195, "step": 8364 }, { "epoch": 1.31, "grad_norm": 21.682037749183138, "learning_rate": 1.2545949265899645e-05, "loss": 0.6883, "step": 8365 }, { "epoch": 1.31, "grad_norm": 21.833900433071957, "learning_rate": 1.2544318401305476e-05, "loss": 0.7107, "step": 8366 }, { "epoch": 1.31, "grad_norm": 17.506565240216386, "learning_rate": 1.2542687464352378e-05, "loss": 0.6388, "step": 8367 }, { "epoch": 1.31, "grad_norm": 19.664891985924683, "learning_rate": 1.2541056455086737e-05, "loss": 0.6404, "step": 8368 }, { "epoch": 1.31, "grad_norm": 22.537311302131393, "learning_rate": 1.2539425373554937e-05, "loss": 0.6769, "step": 8369 }, { "epoch": 1.31, "grad_norm": 18.35579949645393, "learning_rate": 1.2537794219803368e-05, "loss": 0.6158, "step": 8370 }, { "epoch": 1.31, "grad_norm": 20.55089074986457, "learning_rate": 1.2536162993878409e-05, "loss": 0.5631, "step": 8371 }, { "epoch": 1.31, "grad_norm": 17.08641090051207, "learning_rate": 1.2534531695826466e-05, "loss": 0.66, "step": 8372 }, { "epoch": 1.31, "grad_norm": 18.393151781789907, "learning_rate": 1.2532900325693917e-05, "loss": 0.6303, "step": 8373 }, { "epoch": 1.31, "grad_norm": 24.21301103091029, "learning_rate": 1.2531268883527168e-05, "loss": 0.7126, "step": 8374 }, { "epoch": 1.31, "grad_norm": 18.535537246309605, "learning_rate": 1.2529637369372614e-05, "loss": 0.6424, "step": 8375 }, { "epoch": 1.31, "grad_norm": 17.814339562844467, "learning_rate": 1.2528005783276652e-05, "loss": 0.5745, "step": 8376 }, { "epoch": 1.31, "grad_norm": 13.077784393687985, "learning_rate": 1.2526374125285683e-05, "loss": 0.5842, "step": 8377 }, { "epoch": 1.31, "grad_norm": 17.764549310036568, "learning_rate": 1.2524742395446112e-05, "loss": 0.6442, "step": 8378 }, { "epoch": 1.31, "grad_norm": 24.32706219404166, "learning_rate": 1.2523110593804344e-05, "loss": 0.6665, "step": 8379 }, { "epoch": 1.31, "grad_norm": 17.587623230692884, "learning_rate": 1.252147872040679e-05, "loss": 0.6504, "step": 8380 }, { "epoch": 1.31, "grad_norm": 23.64738030869867, "learning_rate": 1.251984677529985e-05, "loss": 0.6724, "step": 8381 }, { "epoch": 1.31, "grad_norm": 16.66866432322756, "learning_rate": 1.2518214758529947e-05, "loss": 0.6555, "step": 8382 }, { "epoch": 1.31, "grad_norm": 19.135922908489277, "learning_rate": 1.251658267014349e-05, "loss": 0.5536, "step": 8383 }, { "epoch": 1.31, "grad_norm": 22.683570287004873, "learning_rate": 1.2514950510186892e-05, "loss": 0.5341, "step": 8384 }, { "epoch": 1.31, "grad_norm": 19.497442436918337, "learning_rate": 1.2513318278706573e-05, "loss": 0.6684, "step": 8385 }, { "epoch": 1.31, "grad_norm": 17.794121792194833, "learning_rate": 1.2511685975748948e-05, "loss": 0.6374, "step": 8386 }, { "epoch": 1.31, "grad_norm": 25.733633631743633, "learning_rate": 1.251005360136045e-05, "loss": 0.7394, "step": 8387 }, { "epoch": 1.31, "grad_norm": 46.83819900012865, "learning_rate": 1.2508421155587492e-05, "loss": 0.635, "step": 8388 }, { "epoch": 1.31, "grad_norm": 17.8407095466611, "learning_rate": 1.2506788638476506e-05, "loss": 0.5833, "step": 8389 }, { "epoch": 1.31, "grad_norm": 21.778555733521987, "learning_rate": 1.2505156050073914e-05, "loss": 0.5958, "step": 8390 }, { "epoch": 1.31, "grad_norm": 27.671684704432664, "learning_rate": 1.2503523390426153e-05, "loss": 0.7228, "step": 8391 }, { "epoch": 1.31, "grad_norm": 18.50419597507629, "learning_rate": 1.2501890659579649e-05, "loss": 0.6231, "step": 8392 }, { "epoch": 1.31, "grad_norm": 28.585759525417487, "learning_rate": 1.250025785758084e-05, "loss": 0.6347, "step": 8393 }, { "epoch": 1.31, "grad_norm": 17.109868363012563, "learning_rate": 1.249862498447616e-05, "loss": 0.5274, "step": 8394 }, { "epoch": 1.31, "grad_norm": 20.593890241020958, "learning_rate": 1.2496992040312045e-05, "loss": 0.655, "step": 8395 }, { "epoch": 1.31, "grad_norm": 16.28844610735789, "learning_rate": 1.2495359025134939e-05, "loss": 0.5972, "step": 8396 }, { "epoch": 1.31, "grad_norm": 15.615189672600541, "learning_rate": 1.2493725938991283e-05, "loss": 0.5638, "step": 8397 }, { "epoch": 1.31, "grad_norm": 19.034943350656782, "learning_rate": 1.2492092781927517e-05, "loss": 0.6228, "step": 8398 }, { "epoch": 1.31, "grad_norm": 16.604180662260138, "learning_rate": 1.2490459553990095e-05, "loss": 0.6158, "step": 8399 }, { "epoch": 1.31, "grad_norm": 25.844669714295254, "learning_rate": 1.2488826255225455e-05, "loss": 0.6275, "step": 8400 }, { "epoch": 1.31, "grad_norm": 15.912448924384213, "learning_rate": 1.2487192885680053e-05, "loss": 0.6135, "step": 8401 }, { "epoch": 1.31, "grad_norm": 20.16203081307279, "learning_rate": 1.2485559445400343e-05, "loss": 0.6634, "step": 8402 }, { "epoch": 1.31, "grad_norm": 14.671311987717258, "learning_rate": 1.2483925934432774e-05, "loss": 0.6482, "step": 8403 }, { "epoch": 1.31, "grad_norm": 25.97376114848387, "learning_rate": 1.2482292352823806e-05, "loss": 0.6776, "step": 8404 }, { "epoch": 1.31, "grad_norm": 25.836135582904237, "learning_rate": 1.2480658700619891e-05, "loss": 0.6545, "step": 8405 }, { "epoch": 1.31, "grad_norm": 17.747727832968042, "learning_rate": 1.24790249778675e-05, "loss": 0.6735, "step": 8406 }, { "epoch": 1.31, "grad_norm": 23.52613928250092, "learning_rate": 1.2477391184613086e-05, "loss": 0.6163, "step": 8407 }, { "epoch": 1.31, "grad_norm": 15.685962541820313, "learning_rate": 1.2475757320903117e-05, "loss": 0.7279, "step": 8408 }, { "epoch": 1.31, "grad_norm": 22.845276350411908, "learning_rate": 1.2474123386784059e-05, "loss": 0.6353, "step": 8409 }, { "epoch": 1.31, "grad_norm": 16.96359114795371, "learning_rate": 1.2472489382302377e-05, "loss": 0.648, "step": 8410 }, { "epoch": 1.31, "grad_norm": 19.820878426393136, "learning_rate": 1.2470855307504544e-05, "loss": 0.6637, "step": 8411 }, { "epoch": 1.31, "grad_norm": 14.835597795712292, "learning_rate": 1.2469221162437033e-05, "loss": 0.6047, "step": 8412 }, { "epoch": 1.31, "grad_norm": 14.10453886893205, "learning_rate": 1.2467586947146313e-05, "loss": 0.5672, "step": 8413 }, { "epoch": 1.31, "grad_norm": 20.020047285174645, "learning_rate": 1.2465952661678866e-05, "loss": 0.6774, "step": 8414 }, { "epoch": 1.31, "grad_norm": 23.203443262377164, "learning_rate": 1.2464318306081167e-05, "loss": 0.6614, "step": 8415 }, { "epoch": 1.31, "grad_norm": 25.966718314210883, "learning_rate": 1.24626838803997e-05, "loss": 0.626, "step": 8416 }, { "epoch": 1.31, "grad_norm": 20.429948645357744, "learning_rate": 1.2461049384680938e-05, "loss": 0.6385, "step": 8417 }, { "epoch": 1.31, "grad_norm": 17.410800452892463, "learning_rate": 1.2459414818971376e-05, "loss": 0.7324, "step": 8418 }, { "epoch": 1.32, "grad_norm": 18.28931554132851, "learning_rate": 1.2457780183317496e-05, "loss": 0.6756, "step": 8419 }, { "epoch": 1.32, "grad_norm": 17.80160122486069, "learning_rate": 1.2456145477765782e-05, "loss": 0.6449, "step": 8420 }, { "epoch": 1.32, "grad_norm": 26.596933002921876, "learning_rate": 1.245451070236273e-05, "loss": 0.7206, "step": 8421 }, { "epoch": 1.32, "grad_norm": 14.168266110261689, "learning_rate": 1.2452875857154827e-05, "loss": 0.6589, "step": 8422 }, { "epoch": 1.32, "grad_norm": 19.552074673880554, "learning_rate": 1.2451240942188569e-05, "loss": 0.7514, "step": 8423 }, { "epoch": 1.32, "grad_norm": 15.089566779995492, "learning_rate": 1.2449605957510456e-05, "loss": 0.5992, "step": 8424 }, { "epoch": 1.32, "grad_norm": 18.558456475792575, "learning_rate": 1.244797090316698e-05, "loss": 0.7114, "step": 8425 }, { "epoch": 1.32, "grad_norm": 21.750146685851046, "learning_rate": 1.2446335779204647e-05, "loss": 0.7159, "step": 8426 }, { "epoch": 1.32, "grad_norm": 15.94879042559237, "learning_rate": 1.2444700585669952e-05, "loss": 0.5895, "step": 8427 }, { "epoch": 1.32, "grad_norm": 19.606569309814386, "learning_rate": 1.2443065322609402e-05, "loss": 0.6628, "step": 8428 }, { "epoch": 1.32, "grad_norm": 16.28941968393588, "learning_rate": 1.2441429990069507e-05, "loss": 0.6623, "step": 8429 }, { "epoch": 1.32, "grad_norm": 19.82479284197508, "learning_rate": 1.2439794588096767e-05, "loss": 0.6832, "step": 8430 }, { "epoch": 1.32, "grad_norm": 12.99441871821392, "learning_rate": 1.24381591167377e-05, "loss": 0.6229, "step": 8431 }, { "epoch": 1.32, "grad_norm": 17.94365852748525, "learning_rate": 1.243652357603881e-05, "loss": 0.681, "step": 8432 }, { "epoch": 1.32, "grad_norm": 31.147005642907533, "learning_rate": 1.243488796604662e-05, "loss": 0.6554, "step": 8433 }, { "epoch": 1.32, "grad_norm": 28.43136384596959, "learning_rate": 1.2433252286807635e-05, "loss": 0.6552, "step": 8434 }, { "epoch": 1.32, "grad_norm": 21.172208024675033, "learning_rate": 1.2431616538368383e-05, "loss": 0.6674, "step": 8435 }, { "epoch": 1.32, "grad_norm": 13.567540897048898, "learning_rate": 1.2429980720775376e-05, "loss": 0.6981, "step": 8436 }, { "epoch": 1.32, "grad_norm": 16.623454990605445, "learning_rate": 1.2428344834075142e-05, "loss": 0.6922, "step": 8437 }, { "epoch": 1.32, "grad_norm": 21.0994253181439, "learning_rate": 1.2426708878314197e-05, "loss": 0.6782, "step": 8438 }, { "epoch": 1.32, "grad_norm": 19.673935467175266, "learning_rate": 1.2425072853539076e-05, "loss": 0.6562, "step": 8439 }, { "epoch": 1.32, "grad_norm": 21.8969765086303, "learning_rate": 1.2423436759796296e-05, "loss": 0.6404, "step": 8440 }, { "epoch": 1.32, "grad_norm": 19.58288717159023, "learning_rate": 1.2421800597132399e-05, "loss": 0.6153, "step": 8441 }, { "epoch": 1.32, "grad_norm": 19.26700704340193, "learning_rate": 1.2420164365593903e-05, "loss": 0.5948, "step": 8442 }, { "epoch": 1.32, "grad_norm": 20.903703649971675, "learning_rate": 1.2418528065227354e-05, "loss": 0.5366, "step": 8443 }, { "epoch": 1.32, "grad_norm": 19.758806024914684, "learning_rate": 1.2416891696079276e-05, "loss": 0.6088, "step": 8444 }, { "epoch": 1.32, "grad_norm": 19.369139554869104, "learning_rate": 1.2415255258196215e-05, "loss": 0.6122, "step": 8445 }, { "epoch": 1.32, "grad_norm": 24.64057546391706, "learning_rate": 1.2413618751624708e-05, "loss": 0.6863, "step": 8446 }, { "epoch": 1.32, "grad_norm": 22.249161263125092, "learning_rate": 1.2411982176411294e-05, "loss": 0.6737, "step": 8447 }, { "epoch": 1.32, "grad_norm": 19.616167868424817, "learning_rate": 1.2410345532602518e-05, "loss": 0.679, "step": 8448 }, { "epoch": 1.32, "grad_norm": 17.542266887724022, "learning_rate": 1.2408708820244926e-05, "loss": 0.6725, "step": 8449 }, { "epoch": 1.32, "grad_norm": 21.2780886429179, "learning_rate": 1.2407072039385064e-05, "loss": 0.6734, "step": 8450 }, { "epoch": 1.32, "grad_norm": 23.760429935599344, "learning_rate": 1.2405435190069481e-05, "loss": 0.6293, "step": 8451 }, { "epoch": 1.32, "grad_norm": 22.352998041416104, "learning_rate": 1.2403798272344729e-05, "loss": 0.7218, "step": 8452 }, { "epoch": 1.32, "grad_norm": 19.34735796232701, "learning_rate": 1.240216128625736e-05, "loss": 0.6762, "step": 8453 }, { "epoch": 1.32, "grad_norm": 22.848365222889896, "learning_rate": 1.2400524231853929e-05, "loss": 0.6442, "step": 8454 }, { "epoch": 1.32, "grad_norm": 17.527076288966455, "learning_rate": 1.2398887109180992e-05, "loss": 0.651, "step": 8455 }, { "epoch": 1.32, "grad_norm": 19.76830119459625, "learning_rate": 1.2397249918285113e-05, "loss": 0.5973, "step": 8456 }, { "epoch": 1.32, "grad_norm": 17.009791511833946, "learning_rate": 1.2395612659212844e-05, "loss": 0.6102, "step": 8457 }, { "epoch": 1.32, "grad_norm": 31.919847215104042, "learning_rate": 1.2393975332010757e-05, "loss": 0.6027, "step": 8458 }, { "epoch": 1.32, "grad_norm": 21.250803343229645, "learning_rate": 1.2392337936725408e-05, "loss": 0.6422, "step": 8459 }, { "epoch": 1.32, "grad_norm": 9.980697480658229, "learning_rate": 1.239070047340337e-05, "loss": 0.5356, "step": 8460 }, { "epoch": 1.32, "grad_norm": 36.12678984436896, "learning_rate": 1.2389062942091209e-05, "loss": 0.6401, "step": 8461 }, { "epoch": 1.32, "grad_norm": 17.517682775584007, "learning_rate": 1.2387425342835492e-05, "loss": 0.6006, "step": 8462 }, { "epoch": 1.32, "grad_norm": 17.33834159567959, "learning_rate": 1.2385787675682799e-05, "loss": 0.61, "step": 8463 }, { "epoch": 1.32, "grad_norm": 22.59121940021831, "learning_rate": 1.2384149940679697e-05, "loss": 0.7406, "step": 8464 }, { "epoch": 1.32, "grad_norm": 12.764711292442753, "learning_rate": 1.2382512137872769e-05, "loss": 0.6407, "step": 8465 }, { "epoch": 1.32, "grad_norm": 19.010763261304934, "learning_rate": 1.2380874267308586e-05, "loss": 0.656, "step": 8466 }, { "epoch": 1.32, "grad_norm": 15.710717990713249, "learning_rate": 1.237923632903373e-05, "loss": 0.6055, "step": 8467 }, { "epoch": 1.32, "grad_norm": 34.495774283715484, "learning_rate": 1.2377598323094788e-05, "loss": 0.6495, "step": 8468 }, { "epoch": 1.32, "grad_norm": 16.69431043267568, "learning_rate": 1.2375960249538341e-05, "loss": 0.6576, "step": 8469 }, { "epoch": 1.32, "grad_norm": 17.511810830278588, "learning_rate": 1.2374322108410974e-05, "loss": 0.6334, "step": 8470 }, { "epoch": 1.32, "grad_norm": 22.853325801537327, "learning_rate": 1.2372683899759274e-05, "loss": 0.6293, "step": 8471 }, { "epoch": 1.32, "grad_norm": 16.866740655037848, "learning_rate": 1.2371045623629834e-05, "loss": 0.5808, "step": 8472 }, { "epoch": 1.32, "grad_norm": 22.115381657602597, "learning_rate": 1.2369407280069241e-05, "loss": 0.6865, "step": 8473 }, { "epoch": 1.32, "grad_norm": 20.801582747227275, "learning_rate": 1.2367768869124091e-05, "loss": 0.6638, "step": 8474 }, { "epoch": 1.32, "grad_norm": 18.664340221147935, "learning_rate": 1.2366130390840982e-05, "loss": 0.6152, "step": 8475 }, { "epoch": 1.32, "grad_norm": 21.415242830972065, "learning_rate": 1.2364491845266506e-05, "loss": 0.5968, "step": 8476 }, { "epoch": 1.32, "grad_norm": 19.067682735023382, "learning_rate": 1.2362853232447267e-05, "loss": 0.5581, "step": 8477 }, { "epoch": 1.32, "grad_norm": 20.29645373881489, "learning_rate": 1.2361214552429863e-05, "loss": 0.6734, "step": 8478 }, { "epoch": 1.32, "grad_norm": 20.764378467558146, "learning_rate": 1.23595758052609e-05, "loss": 0.6377, "step": 8479 }, { "epoch": 1.32, "grad_norm": 26.29545478929641, "learning_rate": 1.235793699098698e-05, "loss": 0.5976, "step": 8480 }, { "epoch": 1.32, "grad_norm": 20.952746884278923, "learning_rate": 1.2356298109654712e-05, "loss": 0.6185, "step": 8481 }, { "epoch": 1.32, "grad_norm": 19.902903533041194, "learning_rate": 1.2354659161310704e-05, "loss": 0.6704, "step": 8482 }, { "epoch": 1.33, "grad_norm": 19.132129194364175, "learning_rate": 1.2353020146001568e-05, "loss": 0.648, "step": 8483 }, { "epoch": 1.33, "grad_norm": 18.529667552345707, "learning_rate": 1.2351381063773913e-05, "loss": 0.6833, "step": 8484 }, { "epoch": 1.33, "grad_norm": 18.40850758105997, "learning_rate": 1.2349741914674361e-05, "loss": 0.6454, "step": 8485 }, { "epoch": 1.33, "grad_norm": 19.17867539293414, "learning_rate": 1.2348102698749518e-05, "loss": 0.6472, "step": 8486 }, { "epoch": 1.33, "grad_norm": 21.911434192083497, "learning_rate": 1.234646341604601e-05, "loss": 0.6109, "step": 8487 }, { "epoch": 1.33, "grad_norm": 23.287101348233122, "learning_rate": 1.2344824066610454e-05, "loss": 0.6755, "step": 8488 }, { "epoch": 1.33, "grad_norm": 18.26716566563346, "learning_rate": 1.2343184650489476e-05, "loss": 0.6091, "step": 8489 }, { "epoch": 1.33, "grad_norm": 13.40588813351155, "learning_rate": 1.2341545167729693e-05, "loss": 0.5919, "step": 8490 }, { "epoch": 1.33, "grad_norm": 20.77361262600216, "learning_rate": 1.2339905618377739e-05, "loss": 0.6741, "step": 8491 }, { "epoch": 1.33, "grad_norm": 21.582979000219552, "learning_rate": 1.2338266002480237e-05, "loss": 0.5845, "step": 8492 }, { "epoch": 1.33, "grad_norm": 18.768318299826905, "learning_rate": 1.2336626320083816e-05, "loss": 0.6034, "step": 8493 }, { "epoch": 1.33, "grad_norm": 15.772503873865713, "learning_rate": 1.233498657123511e-05, "loss": 0.6968, "step": 8494 }, { "epoch": 1.33, "grad_norm": 25.65378679074082, "learning_rate": 1.2333346755980753e-05, "loss": 0.7412, "step": 8495 }, { "epoch": 1.33, "grad_norm": 15.7526211061879, "learning_rate": 1.233170687436738e-05, "loss": 0.6636, "step": 8496 }, { "epoch": 1.33, "grad_norm": 24.085844378766186, "learning_rate": 1.2330066926441626e-05, "loss": 0.6107, "step": 8497 }, { "epoch": 1.33, "grad_norm": 20.880085868092525, "learning_rate": 1.232842691225013e-05, "loss": 0.6404, "step": 8498 }, { "epoch": 1.33, "grad_norm": 13.821906419986114, "learning_rate": 1.2326786831839536e-05, "loss": 0.5547, "step": 8499 }, { "epoch": 1.33, "grad_norm": 11.788811926836328, "learning_rate": 1.2325146685256489e-05, "loss": 0.6134, "step": 8500 }, { "epoch": 1.33, "grad_norm": 16.759688831202755, "learning_rate": 1.2323506472547626e-05, "loss": 0.6144, "step": 8501 }, { "epoch": 1.33, "grad_norm": 24.276056399474264, "learning_rate": 1.2321866193759602e-05, "loss": 0.6455, "step": 8502 }, { "epoch": 1.33, "grad_norm": 13.869654232969632, "learning_rate": 1.2320225848939059e-05, "loss": 0.6269, "step": 8503 }, { "epoch": 1.33, "grad_norm": 22.588172582308815, "learning_rate": 1.2318585438132654e-05, "loss": 0.6206, "step": 8504 }, { "epoch": 1.33, "grad_norm": 23.0132422032763, "learning_rate": 1.2316944961387028e-05, "loss": 0.5531, "step": 8505 }, { "epoch": 1.33, "grad_norm": 17.95920531327305, "learning_rate": 1.2315304418748848e-05, "loss": 0.6361, "step": 8506 }, { "epoch": 1.33, "grad_norm": 21.701476311867843, "learning_rate": 1.2313663810264762e-05, "loss": 0.6137, "step": 8507 }, { "epoch": 1.33, "grad_norm": 18.3641054747069, "learning_rate": 1.2312023135981434e-05, "loss": 0.6001, "step": 8508 }, { "epoch": 1.33, "grad_norm": 16.87934186055937, "learning_rate": 1.231038239594552e-05, "loss": 0.616, "step": 8509 }, { "epoch": 1.33, "grad_norm": 31.62906589022556, "learning_rate": 1.230874159020368e-05, "loss": 0.6292, "step": 8510 }, { "epoch": 1.33, "grad_norm": 17.017662707481655, "learning_rate": 1.2307100718802579e-05, "loss": 0.5669, "step": 8511 }, { "epoch": 1.33, "grad_norm": 15.850245509893645, "learning_rate": 1.2305459781788885e-05, "loss": 0.6347, "step": 8512 }, { "epoch": 1.33, "grad_norm": 24.579043334833766, "learning_rate": 1.2303818779209264e-05, "loss": 0.6614, "step": 8513 }, { "epoch": 1.33, "grad_norm": 18.089340029140562, "learning_rate": 1.2302177711110384e-05, "loss": 0.6952, "step": 8514 }, { "epoch": 1.33, "grad_norm": 21.053311086757766, "learning_rate": 1.2300536577538917e-05, "loss": 0.6103, "step": 8515 }, { "epoch": 1.33, "grad_norm": 19.962015982455814, "learning_rate": 1.2298895378541536e-05, "loss": 0.5827, "step": 8516 }, { "epoch": 1.33, "grad_norm": 15.934177167292416, "learning_rate": 1.2297254114164914e-05, "loss": 0.6862, "step": 8517 }, { "epoch": 1.33, "grad_norm": 22.84563689781987, "learning_rate": 1.2295612784455728e-05, "loss": 0.6806, "step": 8518 }, { "epoch": 1.33, "grad_norm": 17.973864316809564, "learning_rate": 1.2293971389460659e-05, "loss": 0.6555, "step": 8519 }, { "epoch": 1.33, "grad_norm": 19.418891846811245, "learning_rate": 1.2292329929226385e-05, "loss": 0.6581, "step": 8520 }, { "epoch": 1.33, "grad_norm": 18.18932821156568, "learning_rate": 1.229068840379959e-05, "loss": 0.606, "step": 8521 }, { "epoch": 1.33, "grad_norm": 13.743693873701552, "learning_rate": 1.2289046813226954e-05, "loss": 0.6379, "step": 8522 }, { "epoch": 1.33, "grad_norm": 18.77122972883461, "learning_rate": 1.2287405157555168e-05, "loss": 0.6134, "step": 8523 }, { "epoch": 1.33, "grad_norm": 23.03307996425988, "learning_rate": 1.2285763436830917e-05, "loss": 0.675, "step": 8524 }, { "epoch": 1.33, "grad_norm": 13.659364245108755, "learning_rate": 1.2284121651100891e-05, "loss": 0.5764, "step": 8525 }, { "epoch": 1.33, "grad_norm": 26.19314833728055, "learning_rate": 1.228247980041178e-05, "loss": 0.6735, "step": 8526 }, { "epoch": 1.33, "grad_norm": 12.713701830037026, "learning_rate": 1.2280837884810282e-05, "loss": 0.591, "step": 8527 }, { "epoch": 1.33, "grad_norm": 19.15943430902438, "learning_rate": 1.2279195904343084e-05, "loss": 0.6002, "step": 8528 }, { "epoch": 1.33, "grad_norm": 16.048214151727812, "learning_rate": 1.2277553859056894e-05, "loss": 0.6848, "step": 8529 }, { "epoch": 1.33, "grad_norm": 13.690646851684782, "learning_rate": 1.22759117489984e-05, "loss": 0.5433, "step": 8530 }, { "epoch": 1.33, "grad_norm": 24.987558771553786, "learning_rate": 1.227426957421431e-05, "loss": 0.5566, "step": 8531 }, { "epoch": 1.33, "grad_norm": 18.66383854167598, "learning_rate": 1.227262733475132e-05, "loss": 0.6899, "step": 8532 }, { "epoch": 1.33, "grad_norm": 11.022869830178324, "learning_rate": 1.2270985030656139e-05, "loss": 0.5529, "step": 8533 }, { "epoch": 1.33, "grad_norm": 19.35945595900608, "learning_rate": 1.2269342661975474e-05, "loss": 0.6456, "step": 8534 }, { "epoch": 1.33, "grad_norm": 16.18422387360619, "learning_rate": 1.2267700228756033e-05, "loss": 0.6298, "step": 8535 }, { "epoch": 1.33, "grad_norm": 14.660415692858141, "learning_rate": 1.2266057731044521e-05, "loss": 0.6397, "step": 8536 }, { "epoch": 1.33, "grad_norm": 15.309528642527022, "learning_rate": 1.2264415168887651e-05, "loss": 0.5893, "step": 8537 }, { "epoch": 1.33, "grad_norm": 21.20277945083012, "learning_rate": 1.226277254233214e-05, "loss": 0.6605, "step": 8538 }, { "epoch": 1.33, "grad_norm": 13.896931020350566, "learning_rate": 1.2261129851424703e-05, "loss": 0.5987, "step": 8539 }, { "epoch": 1.33, "grad_norm": 27.659557426552695, "learning_rate": 1.2259487096212055e-05, "loss": 0.66, "step": 8540 }, { "epoch": 1.33, "grad_norm": 17.572156330212678, "learning_rate": 1.2257844276740916e-05, "loss": 0.6837, "step": 8541 }, { "epoch": 1.33, "grad_norm": 27.48955546162416, "learning_rate": 1.2256201393058006e-05, "loss": 0.6782, "step": 8542 }, { "epoch": 1.33, "grad_norm": 12.208081511215612, "learning_rate": 1.2254558445210048e-05, "loss": 0.5595, "step": 8543 }, { "epoch": 1.33, "grad_norm": 15.310728373998613, "learning_rate": 1.2252915433243768e-05, "loss": 0.5796, "step": 8544 }, { "epoch": 1.33, "grad_norm": 18.754076669892413, "learning_rate": 1.225127235720589e-05, "loss": 0.6264, "step": 8545 }, { "epoch": 1.33, "grad_norm": 14.674181956880906, "learning_rate": 1.2249629217143143e-05, "loss": 0.5549, "step": 8546 }, { "epoch": 1.34, "grad_norm": 12.92216467619158, "learning_rate": 1.2247986013102258e-05, "loss": 0.6989, "step": 8547 }, { "epoch": 1.34, "grad_norm": 18.064078355667892, "learning_rate": 1.2246342745129964e-05, "loss": 0.6911, "step": 8548 }, { "epoch": 1.34, "grad_norm": 19.170920627284424, "learning_rate": 1.2244699413272998e-05, "loss": 0.6604, "step": 8549 }, { "epoch": 1.34, "grad_norm": 15.94594602994399, "learning_rate": 1.2243056017578095e-05, "loss": 0.6326, "step": 8550 }, { "epoch": 1.34, "grad_norm": 13.776293552586147, "learning_rate": 1.2241412558091988e-05, "loss": 0.6862, "step": 8551 }, { "epoch": 1.34, "grad_norm": 17.479472675607774, "learning_rate": 1.2239769034861423e-05, "loss": 0.6175, "step": 8552 }, { "epoch": 1.34, "grad_norm": 13.63204416552825, "learning_rate": 1.2238125447933134e-05, "loss": 0.6443, "step": 8553 }, { "epoch": 1.34, "grad_norm": 20.82843613445841, "learning_rate": 1.2236481797353865e-05, "loss": 0.6338, "step": 8554 }, { "epoch": 1.34, "grad_norm": 17.37390498443152, "learning_rate": 1.2234838083170362e-05, "loss": 0.5719, "step": 8555 }, { "epoch": 1.34, "grad_norm": 15.549117574979292, "learning_rate": 1.2233194305429375e-05, "loss": 0.6037, "step": 8556 }, { "epoch": 1.34, "grad_norm": 19.57356127272504, "learning_rate": 1.2231550464177646e-05, "loss": 0.5215, "step": 8557 }, { "epoch": 1.34, "grad_norm": 21.570649246300786, "learning_rate": 1.2229906559461925e-05, "loss": 0.6638, "step": 8558 }, { "epoch": 1.34, "grad_norm": 18.945698632700093, "learning_rate": 1.222826259132897e-05, "loss": 0.6179, "step": 8559 }, { "epoch": 1.34, "grad_norm": 26.915847750115947, "learning_rate": 1.2226618559825529e-05, "loss": 0.6985, "step": 8560 }, { "epoch": 1.34, "grad_norm": 15.076808096148143, "learning_rate": 1.222497446499836e-05, "loss": 0.5829, "step": 8561 }, { "epoch": 1.34, "grad_norm": 25.728779161227575, "learning_rate": 1.2223330306894215e-05, "loss": 0.6485, "step": 8562 }, { "epoch": 1.34, "grad_norm": 19.566633584150182, "learning_rate": 1.222168608555986e-05, "loss": 0.6531, "step": 8563 }, { "epoch": 1.34, "grad_norm": 15.665807427988359, "learning_rate": 1.222004180104205e-05, "loss": 0.6951, "step": 8564 }, { "epoch": 1.34, "grad_norm": 18.09666819832233, "learning_rate": 1.2218397453387551e-05, "loss": 0.8099, "step": 8565 }, { "epoch": 1.34, "grad_norm": 22.251314463926256, "learning_rate": 1.2216753042643128e-05, "loss": 0.656, "step": 8566 }, { "epoch": 1.34, "grad_norm": 17.174028588709266, "learning_rate": 1.2215108568855545e-05, "loss": 0.5178, "step": 8567 }, { "epoch": 1.34, "grad_norm": 15.88891556827831, "learning_rate": 1.2213464032071567e-05, "loss": 0.5962, "step": 8568 }, { "epoch": 1.34, "grad_norm": 19.799382599683735, "learning_rate": 1.221181943233797e-05, "loss": 0.5768, "step": 8569 }, { "epoch": 1.34, "grad_norm": 18.91518573033513, "learning_rate": 1.221017476970152e-05, "loss": 0.6383, "step": 8570 }, { "epoch": 1.34, "grad_norm": 17.456165007982904, "learning_rate": 1.2208530044208995e-05, "loss": 0.6282, "step": 8571 }, { "epoch": 1.34, "grad_norm": 16.63572919042834, "learning_rate": 1.2206885255907163e-05, "loss": 0.6224, "step": 8572 }, { "epoch": 1.34, "grad_norm": 22.210550375297608, "learning_rate": 1.2205240404842811e-05, "loss": 0.5794, "step": 8573 }, { "epoch": 1.34, "grad_norm": 25.877061355759512, "learning_rate": 1.2203595491062707e-05, "loss": 0.6282, "step": 8574 }, { "epoch": 1.34, "grad_norm": 16.423044212995645, "learning_rate": 1.2201950514613638e-05, "loss": 0.6323, "step": 8575 }, { "epoch": 1.34, "grad_norm": 21.954159829670836, "learning_rate": 1.2200305475542385e-05, "loss": 0.6524, "step": 8576 }, { "epoch": 1.34, "grad_norm": 16.71081158954826, "learning_rate": 1.2198660373895731e-05, "loss": 0.5714, "step": 8577 }, { "epoch": 1.34, "grad_norm": 17.267509780902145, "learning_rate": 1.2197015209720462e-05, "loss": 0.5689, "step": 8578 }, { "epoch": 1.34, "grad_norm": 26.785402962476226, "learning_rate": 1.2195369983063368e-05, "loss": 0.6925, "step": 8579 }, { "epoch": 1.34, "grad_norm": 21.482483312382136, "learning_rate": 1.2193724693971235e-05, "loss": 0.6086, "step": 8580 }, { "epoch": 1.34, "grad_norm": 20.127896025743652, "learning_rate": 1.2192079342490851e-05, "loss": 0.6026, "step": 8581 }, { "epoch": 1.34, "grad_norm": 14.353481856921645, "learning_rate": 1.2190433928669015e-05, "loss": 0.5335, "step": 8582 }, { "epoch": 1.34, "grad_norm": 19.15072360140765, "learning_rate": 1.218878845255252e-05, "loss": 0.6624, "step": 8583 }, { "epoch": 1.34, "grad_norm": 11.802699797114263, "learning_rate": 1.2187142914188161e-05, "loss": 0.6546, "step": 8584 }, { "epoch": 1.34, "grad_norm": 21.262017076436383, "learning_rate": 1.218549731362274e-05, "loss": 0.6721, "step": 8585 }, { "epoch": 1.34, "grad_norm": 12.179441932199305, "learning_rate": 1.218385165090305e-05, "loss": 0.5582, "step": 8586 }, { "epoch": 1.34, "grad_norm": 17.147349544241326, "learning_rate": 1.2182205926075899e-05, "loss": 0.6305, "step": 8587 }, { "epoch": 1.34, "grad_norm": 21.79689105727175, "learning_rate": 1.2180560139188088e-05, "loss": 0.6199, "step": 8588 }, { "epoch": 1.34, "grad_norm": 12.124370497161896, "learning_rate": 1.217891429028642e-05, "loss": 0.5722, "step": 8589 }, { "epoch": 1.34, "grad_norm": 18.527256327436305, "learning_rate": 1.2177268379417708e-05, "loss": 0.6247, "step": 8590 }, { "epoch": 1.34, "grad_norm": 19.956928681494368, "learning_rate": 1.2175622406628754e-05, "loss": 0.549, "step": 8591 }, { "epoch": 1.34, "grad_norm": 29.97446677005685, "learning_rate": 1.2173976371966372e-05, "loss": 0.6974, "step": 8592 }, { "epoch": 1.34, "grad_norm": 15.493586167850015, "learning_rate": 1.2172330275477374e-05, "loss": 0.6115, "step": 8593 }, { "epoch": 1.34, "grad_norm": 21.941489858062244, "learning_rate": 1.2170684117208573e-05, "loss": 0.6264, "step": 8594 }, { "epoch": 1.34, "grad_norm": 19.27002217046163, "learning_rate": 1.2169037897206787e-05, "loss": 0.6026, "step": 8595 }, { "epoch": 1.34, "grad_norm": 21.086005392739107, "learning_rate": 1.2167391615518831e-05, "loss": 0.6347, "step": 8596 }, { "epoch": 1.34, "grad_norm": 19.84576977340459, "learning_rate": 1.2165745272191524e-05, "loss": 0.6643, "step": 8597 }, { "epoch": 1.34, "grad_norm": 21.689365290258255, "learning_rate": 1.2164098867271694e-05, "loss": 0.7244, "step": 8598 }, { "epoch": 1.34, "grad_norm": 13.649201631581684, "learning_rate": 1.2162452400806151e-05, "loss": 0.5708, "step": 8599 }, { "epoch": 1.34, "grad_norm": 29.300732506662012, "learning_rate": 1.216080587284173e-05, "loss": 0.6756, "step": 8600 }, { "epoch": 1.34, "grad_norm": 18.766556312517203, "learning_rate": 1.2159159283425254e-05, "loss": 0.6162, "step": 8601 }, { "epoch": 1.34, "grad_norm": 26.39327628356211, "learning_rate": 1.2157512632603553e-05, "loss": 0.6694, "step": 8602 }, { "epoch": 1.34, "grad_norm": 24.74339040488667, "learning_rate": 1.215586592042345e-05, "loss": 0.6752, "step": 8603 }, { "epoch": 1.34, "grad_norm": 18.849812490823286, "learning_rate": 1.2154219146931786e-05, "loss": 0.6391, "step": 8604 }, { "epoch": 1.34, "grad_norm": 15.199938153408523, "learning_rate": 1.2152572312175388e-05, "loss": 0.579, "step": 8605 }, { "epoch": 1.34, "grad_norm": 19.43259540466348, "learning_rate": 1.2150925416201091e-05, "loss": 0.6109, "step": 8606 }, { "epoch": 1.34, "grad_norm": 18.49210279563776, "learning_rate": 1.2149278459055737e-05, "loss": 0.6209, "step": 8607 }, { "epoch": 1.34, "grad_norm": 20.70374795012466, "learning_rate": 1.2147631440786156e-05, "loss": 0.6015, "step": 8608 }, { "epoch": 1.34, "grad_norm": 17.84303984924125, "learning_rate": 1.2145984361439197e-05, "loss": 0.6145, "step": 8609 }, { "epoch": 1.34, "grad_norm": 18.83683998891923, "learning_rate": 1.2144337221061697e-05, "loss": 0.6278, "step": 8610 }, { "epoch": 1.35, "grad_norm": 18.27728890070193, "learning_rate": 1.2142690019700503e-05, "loss": 0.6035, "step": 8611 }, { "epoch": 1.35, "grad_norm": 15.829886528611222, "learning_rate": 1.2141042757402454e-05, "loss": 0.59, "step": 8612 }, { "epoch": 1.35, "grad_norm": 18.899656288015105, "learning_rate": 1.2139395434214406e-05, "loss": 0.584, "step": 8613 }, { "epoch": 1.35, "grad_norm": 17.5394635526347, "learning_rate": 1.2137748050183197e-05, "loss": 0.6026, "step": 8614 }, { "epoch": 1.35, "grad_norm": 13.832506498603806, "learning_rate": 1.2136100605355691e-05, "loss": 0.618, "step": 8615 }, { "epoch": 1.35, "grad_norm": 31.684148808281535, "learning_rate": 1.213445309977873e-05, "loss": 0.6659, "step": 8616 }, { "epoch": 1.35, "grad_norm": 20.11390394060753, "learning_rate": 1.2132805533499172e-05, "loss": 0.6353, "step": 8617 }, { "epoch": 1.35, "grad_norm": 16.570365225381163, "learning_rate": 1.213115790656387e-05, "loss": 0.6731, "step": 8618 }, { "epoch": 1.35, "grad_norm": 18.545611620505994, "learning_rate": 1.2129510219019686e-05, "loss": 0.6404, "step": 8619 }, { "epoch": 1.35, "grad_norm": 16.941143432735313, "learning_rate": 1.2127862470913478e-05, "loss": 0.5982, "step": 8620 }, { "epoch": 1.35, "grad_norm": 14.405501643083866, "learning_rate": 1.2126214662292106e-05, "loss": 0.6401, "step": 8621 }, { "epoch": 1.35, "grad_norm": 21.49025598247731, "learning_rate": 1.2124566793202432e-05, "loss": 0.6717, "step": 8622 }, { "epoch": 1.35, "grad_norm": 20.20007527781142, "learning_rate": 1.2122918863691322e-05, "loss": 0.6624, "step": 8623 }, { "epoch": 1.35, "grad_norm": 18.363786412918277, "learning_rate": 1.2121270873805638e-05, "loss": 0.5814, "step": 8624 }, { "epoch": 1.35, "grad_norm": 17.313352477778924, "learning_rate": 1.2119622823592254e-05, "loss": 0.6372, "step": 8625 }, { "epoch": 1.35, "grad_norm": 18.40802895622932, "learning_rate": 1.2117974713098038e-05, "loss": 0.6714, "step": 8626 }, { "epoch": 1.35, "grad_norm": 17.594557231218044, "learning_rate": 1.2116326542369859e-05, "loss": 0.5989, "step": 8627 }, { "epoch": 1.35, "grad_norm": 15.18178038960349, "learning_rate": 1.211467831145459e-05, "loss": 0.5986, "step": 8628 }, { "epoch": 1.35, "grad_norm": 13.942473229376064, "learning_rate": 1.2113030020399107e-05, "loss": 0.6953, "step": 8629 }, { "epoch": 1.35, "grad_norm": 16.832735425818143, "learning_rate": 1.2111381669250288e-05, "loss": 0.6188, "step": 8630 }, { "epoch": 1.35, "grad_norm": 26.321500638717836, "learning_rate": 1.2109733258055007e-05, "loss": 0.6906, "step": 8631 }, { "epoch": 1.35, "grad_norm": 21.573466512152592, "learning_rate": 1.210808478686015e-05, "loss": 0.6716, "step": 8632 }, { "epoch": 1.35, "grad_norm": 19.15335020999387, "learning_rate": 1.210643625571259e-05, "loss": 0.6121, "step": 8633 }, { "epoch": 1.35, "grad_norm": 19.133454025061795, "learning_rate": 1.2104787664659221e-05, "loss": 0.6633, "step": 8634 }, { "epoch": 1.35, "grad_norm": 15.280080881719982, "learning_rate": 1.2103139013746919e-05, "loss": 0.6343, "step": 8635 }, { "epoch": 1.35, "grad_norm": 27.531134442791593, "learning_rate": 1.2101490303022571e-05, "loss": 0.702, "step": 8636 }, { "epoch": 1.35, "grad_norm": 16.797471289585328, "learning_rate": 1.2099841532533073e-05, "loss": 0.6256, "step": 8637 }, { "epoch": 1.35, "grad_norm": 18.276726431224354, "learning_rate": 1.2098192702325309e-05, "loss": 0.6478, "step": 8638 }, { "epoch": 1.35, "grad_norm": 17.89552403801863, "learning_rate": 1.209654381244617e-05, "loss": 0.6902, "step": 8639 }, { "epoch": 1.35, "grad_norm": 11.958346612008572, "learning_rate": 1.2094894862942552e-05, "loss": 0.6106, "step": 8640 }, { "epoch": 1.35, "grad_norm": 12.678519720501198, "learning_rate": 1.2093245853861349e-05, "loss": 0.657, "step": 8641 }, { "epoch": 1.35, "grad_norm": 15.707797702739459, "learning_rate": 1.2091596785249461e-05, "loss": 0.6076, "step": 8642 }, { "epoch": 1.35, "grad_norm": 21.43371740228696, "learning_rate": 1.208994765715378e-05, "loss": 0.647, "step": 8643 }, { "epoch": 1.35, "grad_norm": 27.12832496025796, "learning_rate": 1.2088298469621214e-05, "loss": 0.6718, "step": 8644 }, { "epoch": 1.35, "grad_norm": 24.98985447894854, "learning_rate": 1.208664922269866e-05, "loss": 0.5796, "step": 8645 }, { "epoch": 1.35, "grad_norm": 20.767594889857637, "learning_rate": 1.2084999916433022e-05, "loss": 0.5757, "step": 8646 }, { "epoch": 1.35, "grad_norm": 18.86950795375119, "learning_rate": 1.2083350550871206e-05, "loss": 0.6389, "step": 8647 }, { "epoch": 1.35, "grad_norm": 16.182343188120832, "learning_rate": 1.2081701126060121e-05, "loss": 0.6121, "step": 8648 }, { "epoch": 1.35, "grad_norm": 25.74199996439996, "learning_rate": 1.2080051642046674e-05, "loss": 0.6565, "step": 8649 }, { "epoch": 1.35, "grad_norm": 17.432182822891118, "learning_rate": 1.2078402098877771e-05, "loss": 0.6103, "step": 8650 }, { "epoch": 1.35, "grad_norm": 17.90295544119717, "learning_rate": 1.2076752496600333e-05, "loss": 0.6326, "step": 8651 }, { "epoch": 1.35, "grad_norm": 18.870214430344216, "learning_rate": 1.2075102835261264e-05, "loss": 0.6406, "step": 8652 }, { "epoch": 1.35, "grad_norm": 23.562156110943068, "learning_rate": 1.2073453114907485e-05, "loss": 0.665, "step": 8653 }, { "epoch": 1.35, "grad_norm": 15.820579893890073, "learning_rate": 1.2071803335585917e-05, "loss": 0.6558, "step": 8654 }, { "epoch": 1.35, "grad_norm": 42.492516257039505, "learning_rate": 1.207015349734347e-05, "loss": 0.6626, "step": 8655 }, { "epoch": 1.35, "grad_norm": 16.752472217332123, "learning_rate": 1.206850360022707e-05, "loss": 0.7339, "step": 8656 }, { "epoch": 1.35, "grad_norm": 16.302521378499183, "learning_rate": 1.2066853644283639e-05, "loss": 0.6109, "step": 8657 }, { "epoch": 1.35, "grad_norm": 15.986431140602697, "learning_rate": 1.2065203629560093e-05, "loss": 0.5829, "step": 8658 }, { "epoch": 1.35, "grad_norm": 18.375647208678277, "learning_rate": 1.2063553556103372e-05, "loss": 0.5178, "step": 8659 }, { "epoch": 1.35, "grad_norm": 19.58411080173229, "learning_rate": 1.206190342396039e-05, "loss": 0.67, "step": 8660 }, { "epoch": 1.35, "grad_norm": 15.888372249392383, "learning_rate": 1.2060253233178086e-05, "loss": 0.5886, "step": 8661 }, { "epoch": 1.35, "grad_norm": 13.246519981283104, "learning_rate": 1.2058602983803378e-05, "loss": 0.6475, "step": 8662 }, { "epoch": 1.35, "grad_norm": 12.864418853513957, "learning_rate": 1.2056952675883208e-05, "loss": 0.6008, "step": 8663 }, { "epoch": 1.35, "grad_norm": 20.578250072884106, "learning_rate": 1.2055302309464509e-05, "loss": 0.5911, "step": 8664 }, { "epoch": 1.35, "grad_norm": 20.716578899167196, "learning_rate": 1.2053651884594215e-05, "loss": 0.6514, "step": 8665 }, { "epoch": 1.35, "grad_norm": 23.838624148899495, "learning_rate": 1.2052001401319262e-05, "loss": 0.6989, "step": 8666 }, { "epoch": 1.35, "grad_norm": 18.869598103963412, "learning_rate": 1.205035085968659e-05, "loss": 0.7058, "step": 8667 }, { "epoch": 1.35, "grad_norm": 25.229050145928856, "learning_rate": 1.2048700259743136e-05, "loss": 0.6901, "step": 8668 }, { "epoch": 1.35, "grad_norm": 4.2123019603452105, "learning_rate": 1.2047049601535847e-05, "loss": 0.602, "step": 8669 }, { "epoch": 1.35, "grad_norm": 21.594150492574343, "learning_rate": 1.2045398885111665e-05, "loss": 0.6425, "step": 8670 }, { "epoch": 1.35, "grad_norm": 18.441651733548678, "learning_rate": 1.2043748110517536e-05, "loss": 0.6666, "step": 8671 }, { "epoch": 1.35, "grad_norm": 15.520323122735798, "learning_rate": 1.2042097277800406e-05, "loss": 0.5862, "step": 8672 }, { "epoch": 1.35, "grad_norm": 15.330987991507707, "learning_rate": 1.2040446387007222e-05, "loss": 0.6194, "step": 8673 }, { "epoch": 1.35, "grad_norm": 14.146208037695654, "learning_rate": 1.203879543818494e-05, "loss": 0.6213, "step": 8674 }, { "epoch": 1.36, "grad_norm": 19.66988023487055, "learning_rate": 1.2037144431380506e-05, "loss": 0.5863, "step": 8675 }, { "epoch": 1.36, "grad_norm": 17.40602767817765, "learning_rate": 1.2035493366640879e-05, "loss": 0.5943, "step": 8676 }, { "epoch": 1.36, "grad_norm": 23.25615465959148, "learning_rate": 1.2033842244013006e-05, "loss": 0.5958, "step": 8677 }, { "epoch": 1.36, "grad_norm": 17.37008311490887, "learning_rate": 1.2032191063543855e-05, "loss": 0.6931, "step": 8678 }, { "epoch": 1.36, "grad_norm": 13.431357786931859, "learning_rate": 1.2030539825280373e-05, "loss": 0.5809, "step": 8679 }, { "epoch": 1.36, "grad_norm": 22.634978234535726, "learning_rate": 1.202888852926953e-05, "loss": 0.706, "step": 8680 }, { "epoch": 1.36, "grad_norm": 19.180166034206074, "learning_rate": 1.2027237175558283e-05, "loss": 0.6727, "step": 8681 }, { "epoch": 1.36, "grad_norm": 18.274704392230202, "learning_rate": 1.2025585764193597e-05, "loss": 0.6655, "step": 8682 }, { "epoch": 1.36, "grad_norm": 15.758746647604836, "learning_rate": 1.2023934295222437e-05, "loss": 0.589, "step": 8683 }, { "epoch": 1.36, "grad_norm": 22.147222840172123, "learning_rate": 1.202228276869177e-05, "loss": 0.7426, "step": 8684 }, { "epoch": 1.36, "grad_norm": 18.242841270216683, "learning_rate": 1.202063118464856e-05, "loss": 0.5979, "step": 8685 }, { "epoch": 1.36, "grad_norm": 22.265055188701282, "learning_rate": 1.2018979543139788e-05, "loss": 0.6719, "step": 8686 }, { "epoch": 1.36, "grad_norm": 20.970906259754297, "learning_rate": 1.2017327844212414e-05, "loss": 0.573, "step": 8687 }, { "epoch": 1.36, "grad_norm": 13.97843936655783, "learning_rate": 1.2015676087913418e-05, "loss": 0.6731, "step": 8688 }, { "epoch": 1.36, "grad_norm": 26.719685570471505, "learning_rate": 1.2014024274289773e-05, "loss": 0.7976, "step": 8689 }, { "epoch": 1.36, "grad_norm": 17.657509474436512, "learning_rate": 1.2012372403388457e-05, "loss": 0.6603, "step": 8690 }, { "epoch": 1.36, "grad_norm": 14.92491280555814, "learning_rate": 1.2010720475256446e-05, "loss": 0.6932, "step": 8691 }, { "epoch": 1.36, "grad_norm": 15.058209762427303, "learning_rate": 1.200906848994072e-05, "loss": 0.6104, "step": 8692 }, { "epoch": 1.36, "grad_norm": 15.012169296408299, "learning_rate": 1.2007416447488263e-05, "loss": 0.6158, "step": 8693 }, { "epoch": 1.36, "grad_norm": 16.49370475224746, "learning_rate": 1.2005764347946053e-05, "loss": 0.6646, "step": 8694 }, { "epoch": 1.36, "grad_norm": 19.485652090188943, "learning_rate": 1.200411219136108e-05, "loss": 0.5479, "step": 8695 }, { "epoch": 1.36, "grad_norm": 14.097223393070108, "learning_rate": 1.2002459977780331e-05, "loss": 0.6201, "step": 8696 }, { "epoch": 1.36, "grad_norm": 17.79586073218258, "learning_rate": 1.200080770725079e-05, "loss": 0.6067, "step": 8697 }, { "epoch": 1.36, "grad_norm": 11.52462496183499, "learning_rate": 1.1999155379819449e-05, "loss": 0.5849, "step": 8698 }, { "epoch": 1.36, "grad_norm": 21.27744664362888, "learning_rate": 1.1997502995533299e-05, "loss": 0.7032, "step": 8699 }, { "epoch": 1.36, "grad_norm": 23.678167306299112, "learning_rate": 1.1995850554439332e-05, "loss": 0.6332, "step": 8700 }, { "epoch": 1.36, "grad_norm": 27.32049155237977, "learning_rate": 1.199419805658454e-05, "loss": 0.7102, "step": 8701 }, { "epoch": 1.36, "grad_norm": 15.668013411662498, "learning_rate": 1.1992545502015923e-05, "loss": 0.6285, "step": 8702 }, { "epoch": 1.36, "grad_norm": 16.35178649391694, "learning_rate": 1.1990892890780482e-05, "loss": 0.6692, "step": 8703 }, { "epoch": 1.36, "grad_norm": 16.67873565852468, "learning_rate": 1.1989240222925206e-05, "loss": 0.6064, "step": 8704 }, { "epoch": 1.36, "grad_norm": 11.204817147330086, "learning_rate": 1.1987587498497107e-05, "loss": 0.6411, "step": 8705 }, { "epoch": 1.36, "grad_norm": 14.827820825019513, "learning_rate": 1.1985934717543178e-05, "loss": 0.6193, "step": 8706 }, { "epoch": 1.36, "grad_norm": 18.466607331382953, "learning_rate": 1.1984281880110427e-05, "loss": 0.664, "step": 8707 }, { "epoch": 1.36, "grad_norm": 13.920055425695603, "learning_rate": 1.198262898624586e-05, "loss": 0.6509, "step": 8708 }, { "epoch": 1.36, "grad_norm": 18.04460874009785, "learning_rate": 1.1980976035996488e-05, "loss": 0.5718, "step": 8709 }, { "epoch": 1.36, "grad_norm": 14.65760851930483, "learning_rate": 1.1979323029409316e-05, "loss": 0.6145, "step": 8710 }, { "epoch": 1.36, "grad_norm": 18.045625629775152, "learning_rate": 1.1977669966531353e-05, "loss": 0.616, "step": 8711 }, { "epoch": 1.36, "grad_norm": 20.667504673998017, "learning_rate": 1.197601684740961e-05, "loss": 0.6646, "step": 8712 }, { "epoch": 1.36, "grad_norm": 16.207071307030063, "learning_rate": 1.197436367209111e-05, "loss": 0.6228, "step": 8713 }, { "epoch": 1.36, "grad_norm": 21.933617073651327, "learning_rate": 1.1972710440622858e-05, "loss": 0.6474, "step": 8714 }, { "epoch": 1.36, "grad_norm": 18.28544016770247, "learning_rate": 1.1971057153051878e-05, "loss": 0.707, "step": 8715 }, { "epoch": 1.36, "grad_norm": 18.115497778856586, "learning_rate": 1.1969403809425183e-05, "loss": 0.702, "step": 8716 }, { "epoch": 1.36, "grad_norm": 24.303170521308342, "learning_rate": 1.1967750409789796e-05, "loss": 0.6456, "step": 8717 }, { "epoch": 1.36, "grad_norm": 21.41790776206668, "learning_rate": 1.196609695419274e-05, "loss": 0.6288, "step": 8718 }, { "epoch": 1.36, "grad_norm": 19.29997157684022, "learning_rate": 1.1964443442681036e-05, "loss": 0.667, "step": 8719 }, { "epoch": 1.36, "grad_norm": 18.997276983135546, "learning_rate": 1.196278987530171e-05, "loss": 0.5949, "step": 8720 }, { "epoch": 1.36, "grad_norm": 15.344192490004842, "learning_rate": 1.1961136252101786e-05, "loss": 0.5479, "step": 8721 }, { "epoch": 1.36, "grad_norm": 21.402597818887436, "learning_rate": 1.1959482573128298e-05, "loss": 0.659, "step": 8722 }, { "epoch": 1.36, "grad_norm": 22.250074908346086, "learning_rate": 1.1957828838428269e-05, "loss": 0.6589, "step": 8723 }, { "epoch": 1.36, "grad_norm": 14.78168894829748, "learning_rate": 1.1956175048048734e-05, "loss": 0.5787, "step": 8724 }, { "epoch": 1.36, "grad_norm": 26.64106913053749, "learning_rate": 1.1954521202036726e-05, "loss": 0.6823, "step": 8725 }, { "epoch": 1.36, "grad_norm": 25.159660220161182, "learning_rate": 1.1952867300439276e-05, "loss": 0.6809, "step": 8726 }, { "epoch": 1.36, "grad_norm": 15.948444492476462, "learning_rate": 1.1951213343303425e-05, "loss": 0.7293, "step": 8727 }, { "epoch": 1.36, "grad_norm": 18.259032591867555, "learning_rate": 1.1949559330676209e-05, "loss": 0.695, "step": 8728 }, { "epoch": 1.36, "grad_norm": 12.339335834764851, "learning_rate": 1.194790526260466e-05, "loss": 0.6094, "step": 8729 }, { "epoch": 1.36, "grad_norm": 20.742763378389107, "learning_rate": 1.1946251139135831e-05, "loss": 0.62, "step": 8730 }, { "epoch": 1.36, "grad_norm": 18.939018681191225, "learning_rate": 1.1944596960316755e-05, "loss": 0.5929, "step": 8731 }, { "epoch": 1.36, "grad_norm": 15.431779897274023, "learning_rate": 1.1942942726194477e-05, "loss": 0.6364, "step": 8732 }, { "epoch": 1.36, "grad_norm": 22.683718808267724, "learning_rate": 1.194128843681605e-05, "loss": 0.6156, "step": 8733 }, { "epoch": 1.36, "grad_norm": 24.38668062835554, "learning_rate": 1.1939634092228511e-05, "loss": 0.6481, "step": 8734 }, { "epoch": 1.36, "grad_norm": 20.017589097508985, "learning_rate": 1.1937979692478915e-05, "loss": 0.6234, "step": 8735 }, { "epoch": 1.36, "grad_norm": 20.01555311471506, "learning_rate": 1.1936325237614312e-05, "loss": 0.6355, "step": 8736 }, { "epoch": 1.36, "grad_norm": 14.86445135205235, "learning_rate": 1.193467072768175e-05, "loss": 0.5934, "step": 8737 }, { "epoch": 1.36, "grad_norm": 22.803704844700196, "learning_rate": 1.1933016162728281e-05, "loss": 0.5933, "step": 8738 }, { "epoch": 1.37, "grad_norm": 15.139628692497936, "learning_rate": 1.1931361542800968e-05, "loss": 0.5912, "step": 8739 }, { "epoch": 1.37, "grad_norm": 12.402821710042435, "learning_rate": 1.192970686794686e-05, "loss": 0.5911, "step": 8740 }, { "epoch": 1.37, "grad_norm": 17.761679372666222, "learning_rate": 1.192805213821302e-05, "loss": 0.6233, "step": 8741 }, { "epoch": 1.37, "grad_norm": 16.400814986295323, "learning_rate": 1.1926397353646501e-05, "loss": 0.602, "step": 8742 }, { "epoch": 1.37, "grad_norm": 22.277909358812714, "learning_rate": 1.1924742514294371e-05, "loss": 0.7104, "step": 8743 }, { "epoch": 1.37, "grad_norm": 12.739870507452032, "learning_rate": 1.1923087620203688e-05, "loss": 0.5877, "step": 8744 }, { "epoch": 1.37, "grad_norm": 15.39542170293932, "learning_rate": 1.1921432671421523e-05, "loss": 0.5963, "step": 8745 }, { "epoch": 1.37, "grad_norm": 23.54429346999875, "learning_rate": 1.1919777667994932e-05, "loss": 0.5965, "step": 8746 }, { "epoch": 1.37, "grad_norm": 27.29421663577674, "learning_rate": 1.191812260997099e-05, "loss": 0.6987, "step": 8747 }, { "epoch": 1.37, "grad_norm": 14.545700450931495, "learning_rate": 1.1916467497396759e-05, "loss": 0.6507, "step": 8748 }, { "epoch": 1.37, "grad_norm": 18.808383496813654, "learning_rate": 1.1914812330319318e-05, "loss": 0.6615, "step": 8749 }, { "epoch": 1.37, "grad_norm": 22.436963995758067, "learning_rate": 1.1913157108785731e-05, "loss": 0.7119, "step": 8750 }, { "epoch": 1.37, "grad_norm": 13.583041417119194, "learning_rate": 1.1911501832843077e-05, "loss": 0.6371, "step": 8751 }, { "epoch": 1.37, "grad_norm": 12.972580252186157, "learning_rate": 1.1909846502538429e-05, "loss": 0.6289, "step": 8752 }, { "epoch": 1.37, "grad_norm": 20.192975607421552, "learning_rate": 1.1908191117918864e-05, "loss": 0.6453, "step": 8753 }, { "epoch": 1.37, "grad_norm": 36.280189820668895, "learning_rate": 1.190653567903146e-05, "loss": 0.6793, "step": 8754 }, { "epoch": 1.37, "grad_norm": 12.86481619178849, "learning_rate": 1.1904880185923295e-05, "loss": 0.5888, "step": 8755 }, { "epoch": 1.37, "grad_norm": 19.206713653849533, "learning_rate": 1.190322463864145e-05, "loss": 0.5912, "step": 8756 }, { "epoch": 1.37, "grad_norm": 18.95514795879096, "learning_rate": 1.1901569037233012e-05, "loss": 0.6591, "step": 8757 }, { "epoch": 1.37, "grad_norm": 16.45718708322206, "learning_rate": 1.1899913381745062e-05, "loss": 0.6098, "step": 8758 }, { "epoch": 1.37, "grad_norm": 22.174229752905077, "learning_rate": 1.189825767222469e-05, "loss": 0.6077, "step": 8759 }, { "epoch": 1.37, "grad_norm": 14.924854490395116, "learning_rate": 1.1896601908718979e-05, "loss": 0.555, "step": 8760 }, { "epoch": 1.37, "grad_norm": 21.766709660096474, "learning_rate": 1.1894946091275014e-05, "loss": 0.6823, "step": 8761 }, { "epoch": 1.37, "grad_norm": 15.711140559234362, "learning_rate": 1.1893290219939899e-05, "loss": 0.6426, "step": 8762 }, { "epoch": 1.37, "grad_norm": 21.665283668070913, "learning_rate": 1.1891634294760713e-05, "loss": 0.6003, "step": 8763 }, { "epoch": 1.37, "grad_norm": 36.673144886994365, "learning_rate": 1.1889978315784557e-05, "loss": 0.6776, "step": 8764 }, { "epoch": 1.37, "grad_norm": 15.903327367053109, "learning_rate": 1.1888322283058517e-05, "loss": 0.5077, "step": 8765 }, { "epoch": 1.37, "grad_norm": 20.94289715722972, "learning_rate": 1.1886666196629701e-05, "loss": 0.6484, "step": 8766 }, { "epoch": 1.37, "grad_norm": 25.498087574032546, "learning_rate": 1.1885010056545204e-05, "loss": 0.673, "step": 8767 }, { "epoch": 1.37, "grad_norm": 12.40149669127506, "learning_rate": 1.1883353862852121e-05, "loss": 0.6498, "step": 8768 }, { "epoch": 1.37, "grad_norm": 11.29558471312626, "learning_rate": 1.1881697615597554e-05, "loss": 0.5583, "step": 8769 }, { "epoch": 1.37, "grad_norm": 17.698319888606086, "learning_rate": 1.188004131482861e-05, "loss": 0.6484, "step": 8770 }, { "epoch": 1.37, "grad_norm": 21.05487926657686, "learning_rate": 1.187838496059239e-05, "loss": 0.7046, "step": 8771 }, { "epoch": 1.37, "grad_norm": 17.347937588728534, "learning_rate": 1.1876728552936e-05, "loss": 0.6062, "step": 8772 }, { "epoch": 1.37, "grad_norm": 11.477226872362207, "learning_rate": 1.1875072091906547e-05, "loss": 0.5326, "step": 8773 }, { "epoch": 1.37, "grad_norm": 21.41050072044705, "learning_rate": 1.1873415577551146e-05, "loss": 0.6239, "step": 8774 }, { "epoch": 1.37, "grad_norm": 29.094114957911238, "learning_rate": 1.1871759009916897e-05, "loss": 0.7097, "step": 8775 }, { "epoch": 1.37, "grad_norm": 26.523528699479314, "learning_rate": 1.1870102389050917e-05, "loss": 0.6356, "step": 8776 }, { "epoch": 1.37, "grad_norm": 22.427232202669828, "learning_rate": 1.186844571500032e-05, "loss": 0.5878, "step": 8777 }, { "epoch": 1.37, "grad_norm": 19.328231193694823, "learning_rate": 1.1866788987812219e-05, "loss": 0.5914, "step": 8778 }, { "epoch": 1.37, "grad_norm": 17.47172187406481, "learning_rate": 1.1865132207533731e-05, "loss": 0.6255, "step": 8779 }, { "epoch": 1.37, "grad_norm": 12.157636511901286, "learning_rate": 1.1863475374211974e-05, "loss": 0.5833, "step": 8780 }, { "epoch": 1.37, "grad_norm": 14.718110224006761, "learning_rate": 1.1861818487894072e-05, "loss": 0.6306, "step": 8781 }, { "epoch": 1.37, "grad_norm": 16.783794854776737, "learning_rate": 1.1860161548627137e-05, "loss": 0.592, "step": 8782 }, { "epoch": 1.37, "grad_norm": 21.493314127303073, "learning_rate": 1.1858504556458294e-05, "loss": 0.6051, "step": 8783 }, { "epoch": 1.37, "grad_norm": 18.525089320586517, "learning_rate": 1.1856847511434673e-05, "loss": 0.575, "step": 8784 }, { "epoch": 1.37, "grad_norm": 18.28510709241336, "learning_rate": 1.1855190413603392e-05, "loss": 0.605, "step": 8785 }, { "epoch": 1.37, "grad_norm": 23.72661662166227, "learning_rate": 1.1853533263011583e-05, "loss": 0.6353, "step": 8786 }, { "epoch": 1.37, "grad_norm": 17.017655812338674, "learning_rate": 1.185187605970637e-05, "loss": 0.6292, "step": 8787 }, { "epoch": 1.37, "grad_norm": 21.77108743548212, "learning_rate": 1.1850218803734886e-05, "loss": 0.6616, "step": 8788 }, { "epoch": 1.37, "grad_norm": 19.064875782206286, "learning_rate": 1.1848561495144263e-05, "loss": 0.6098, "step": 8789 }, { "epoch": 1.37, "grad_norm": 15.372480403586309, "learning_rate": 1.184690413398163e-05, "loss": 0.6271, "step": 8790 }, { "epoch": 1.37, "grad_norm": 23.617858075109126, "learning_rate": 1.1845246720294129e-05, "loss": 0.5457, "step": 8791 }, { "epoch": 1.37, "grad_norm": 26.57863554500613, "learning_rate": 1.1843589254128884e-05, "loss": 0.6431, "step": 8792 }, { "epoch": 1.37, "grad_norm": 17.010458838038158, "learning_rate": 1.1841931735533043e-05, "loss": 0.6162, "step": 8793 }, { "epoch": 1.37, "grad_norm": 23.40619553481973, "learning_rate": 1.184027416455374e-05, "loss": 0.6197, "step": 8794 }, { "epoch": 1.37, "grad_norm": 16.926978034401095, "learning_rate": 1.1838616541238115e-05, "loss": 0.5773, "step": 8795 }, { "epoch": 1.37, "grad_norm": 24.16491994313919, "learning_rate": 1.1836958865633315e-05, "loss": 0.6239, "step": 8796 }, { "epoch": 1.37, "grad_norm": 16.247293061595126, "learning_rate": 1.1835301137786476e-05, "loss": 0.7009, "step": 8797 }, { "epoch": 1.37, "grad_norm": 17.60317748373623, "learning_rate": 1.1833643357744747e-05, "loss": 0.5932, "step": 8798 }, { "epoch": 1.37, "grad_norm": 15.256813926854264, "learning_rate": 1.1831985525555274e-05, "loss": 0.6461, "step": 8799 }, { "epoch": 1.37, "grad_norm": 17.181852749825417, "learning_rate": 1.1830327641265202e-05, "loss": 0.6034, "step": 8800 }, { "epoch": 1.37, "grad_norm": 10.984551365590939, "learning_rate": 1.1828669704921685e-05, "loss": 0.5431, "step": 8801 }, { "epoch": 1.37, "grad_norm": 27.292590424626432, "learning_rate": 1.182701171657187e-05, "loss": 0.6692, "step": 8802 }, { "epoch": 1.38, "grad_norm": 21.036929265693516, "learning_rate": 1.1825353676262914e-05, "loss": 0.615, "step": 8803 }, { "epoch": 1.38, "grad_norm": 20.25751258357393, "learning_rate": 1.1823695584041963e-05, "loss": 0.613, "step": 8804 }, { "epoch": 1.38, "grad_norm": 27.153551680277655, "learning_rate": 1.1822037439956178e-05, "loss": 0.6862, "step": 8805 }, { "epoch": 1.38, "grad_norm": 15.63115168470582, "learning_rate": 1.1820379244052715e-05, "loss": 0.653, "step": 8806 }, { "epoch": 1.38, "grad_norm": 18.23232320913089, "learning_rate": 1.1818720996378729e-05, "loss": 0.5638, "step": 8807 }, { "epoch": 1.38, "grad_norm": 15.930645730054778, "learning_rate": 1.1817062696981384e-05, "loss": 0.6471, "step": 8808 }, { "epoch": 1.38, "grad_norm": 21.33081934192666, "learning_rate": 1.1815404345907837e-05, "loss": 0.6343, "step": 8809 }, { "epoch": 1.38, "grad_norm": 19.683788137352334, "learning_rate": 1.1813745943205254e-05, "loss": 0.5866, "step": 8810 }, { "epoch": 1.38, "grad_norm": 14.684608961215451, "learning_rate": 1.1812087488920798e-05, "loss": 0.5788, "step": 8811 }, { "epoch": 1.38, "grad_norm": 16.180297513575756, "learning_rate": 1.1810428983101632e-05, "loss": 0.6191, "step": 8812 }, { "epoch": 1.38, "grad_norm": 14.243397929638935, "learning_rate": 1.1808770425794927e-05, "loss": 0.5692, "step": 8813 }, { "epoch": 1.38, "grad_norm": 13.72967744177661, "learning_rate": 1.1807111817047846e-05, "loss": 0.6357, "step": 8814 }, { "epoch": 1.38, "grad_norm": 20.66109854889212, "learning_rate": 1.1805453156907562e-05, "loss": 0.6164, "step": 8815 }, { "epoch": 1.38, "grad_norm": 21.436350847256517, "learning_rate": 1.1803794445421251e-05, "loss": 0.6708, "step": 8816 }, { "epoch": 1.38, "grad_norm": 17.339018356708436, "learning_rate": 1.1802135682636076e-05, "loss": 0.6642, "step": 8817 }, { "epoch": 1.38, "grad_norm": 19.560486225980682, "learning_rate": 1.1800476868599222e-05, "loss": 0.653, "step": 8818 }, { "epoch": 1.38, "grad_norm": 16.680373201413264, "learning_rate": 1.1798818003357853e-05, "loss": 0.558, "step": 8819 }, { "epoch": 1.38, "grad_norm": 22.154495874987695, "learning_rate": 1.1797159086959156e-05, "loss": 0.5709, "step": 8820 }, { "epoch": 1.38, "grad_norm": 20.7590174255286, "learning_rate": 1.1795500119450305e-05, "loss": 0.6297, "step": 8821 }, { "epoch": 1.38, "grad_norm": 15.793729425057617, "learning_rate": 1.179384110087848e-05, "loss": 0.602, "step": 8822 }, { "epoch": 1.38, "grad_norm": 15.675922302858655, "learning_rate": 1.1792182031290867e-05, "loss": 0.6355, "step": 8823 }, { "epoch": 1.38, "grad_norm": 25.56581067935375, "learning_rate": 1.1790522910734638e-05, "loss": 0.592, "step": 8824 }, { "epoch": 1.38, "grad_norm": 21.423739930853916, "learning_rate": 1.1788863739256992e-05, "loss": 0.6738, "step": 8825 }, { "epoch": 1.38, "grad_norm": 14.300026753623985, "learning_rate": 1.1787204516905104e-05, "loss": 0.5973, "step": 8826 }, { "epoch": 1.38, "grad_norm": 16.190096682563716, "learning_rate": 1.1785545243726166e-05, "loss": 0.5915, "step": 8827 }, { "epoch": 1.38, "grad_norm": 19.673453136647232, "learning_rate": 1.1783885919767368e-05, "loss": 0.6839, "step": 8828 }, { "epoch": 1.38, "grad_norm": 18.876111351146506, "learning_rate": 1.1782226545075896e-05, "loss": 0.6366, "step": 8829 }, { "epoch": 1.38, "grad_norm": 18.039812517661556, "learning_rate": 1.1780567119698944e-05, "loss": 0.6643, "step": 8830 }, { "epoch": 1.38, "grad_norm": 26.43417260015495, "learning_rate": 1.1778907643683704e-05, "loss": 0.5339, "step": 8831 }, { "epoch": 1.38, "grad_norm": 19.071657149789413, "learning_rate": 1.1777248117077371e-05, "loss": 0.59, "step": 8832 }, { "epoch": 1.38, "grad_norm": 16.47647137885765, "learning_rate": 1.1775588539927142e-05, "loss": 0.4877, "step": 8833 }, { "epoch": 1.38, "grad_norm": 21.44229360361136, "learning_rate": 1.1773928912280213e-05, "loss": 0.6178, "step": 8834 }, { "epoch": 1.38, "grad_norm": 27.87886765509306, "learning_rate": 1.1772269234183786e-05, "loss": 0.6472, "step": 8835 }, { "epoch": 1.38, "grad_norm": 30.0246233759278, "learning_rate": 1.1770609505685056e-05, "loss": 0.6759, "step": 8836 }, { "epoch": 1.38, "grad_norm": 22.69169194991702, "learning_rate": 1.1768949726831228e-05, "loss": 0.6553, "step": 8837 }, { "epoch": 1.38, "grad_norm": 17.54049871464825, "learning_rate": 1.1767289897669505e-05, "loss": 0.696, "step": 8838 }, { "epoch": 1.38, "grad_norm": 24.74826716608812, "learning_rate": 1.1765630018247089e-05, "loss": 0.6519, "step": 8839 }, { "epoch": 1.38, "grad_norm": 13.132074428380475, "learning_rate": 1.1763970088611192e-05, "loss": 0.5551, "step": 8840 }, { "epoch": 1.38, "grad_norm": 16.130480533561737, "learning_rate": 1.1762310108809017e-05, "loss": 0.6302, "step": 8841 }, { "epoch": 1.38, "grad_norm": 18.78296848692802, "learning_rate": 1.176065007888777e-05, "loss": 0.6442, "step": 8842 }, { "epoch": 1.38, "grad_norm": 21.249799158044773, "learning_rate": 1.1758989998894667e-05, "loss": 0.67, "step": 8843 }, { "epoch": 1.38, "grad_norm": 18.989843346260734, "learning_rate": 1.1757329868876917e-05, "loss": 0.5932, "step": 8844 }, { "epoch": 1.38, "grad_norm": 15.585443668142796, "learning_rate": 1.1755669688881732e-05, "loss": 0.6167, "step": 8845 }, { "epoch": 1.38, "grad_norm": 15.87361622114679, "learning_rate": 1.175400945895633e-05, "loss": 0.5859, "step": 8846 }, { "epoch": 1.38, "grad_norm": 19.710688289680192, "learning_rate": 1.1752349179147926e-05, "loss": 0.6055, "step": 8847 }, { "epoch": 1.38, "grad_norm": 14.391839388802753, "learning_rate": 1.1750688849503735e-05, "loss": 0.6189, "step": 8848 }, { "epoch": 1.38, "grad_norm": 12.569850741263114, "learning_rate": 1.1749028470070975e-05, "loss": 0.6183, "step": 8849 }, { "epoch": 1.38, "grad_norm": 29.473755080491895, "learning_rate": 1.1747368040896875e-05, "loss": 0.657, "step": 8850 }, { "epoch": 1.38, "grad_norm": 18.53333045561157, "learning_rate": 1.1745707562028643e-05, "loss": 0.5385, "step": 8851 }, { "epoch": 1.38, "grad_norm": 22.16195506698316, "learning_rate": 1.1744047033513514e-05, "loss": 0.6817, "step": 8852 }, { "epoch": 1.38, "grad_norm": 14.38801996331766, "learning_rate": 1.1742386455398704e-05, "loss": 0.5677, "step": 8853 }, { "epoch": 1.38, "grad_norm": 12.618734703035798, "learning_rate": 1.1740725827731446e-05, "loss": 0.6374, "step": 8854 }, { "epoch": 1.38, "grad_norm": 16.860257317008685, "learning_rate": 1.1739065150558961e-05, "loss": 0.606, "step": 8855 }, { "epoch": 1.38, "grad_norm": 19.30032536088489, "learning_rate": 1.1737404423928482e-05, "loss": 0.6202, "step": 8856 }, { "epoch": 1.38, "grad_norm": 13.39155890836942, "learning_rate": 1.1735743647887237e-05, "loss": 0.6448, "step": 8857 }, { "epoch": 1.38, "grad_norm": 14.356398879559249, "learning_rate": 1.1734082822482457e-05, "loss": 0.5771, "step": 8858 }, { "epoch": 1.38, "grad_norm": 17.78085065068071, "learning_rate": 1.1732421947761377e-05, "loss": 0.5897, "step": 8859 }, { "epoch": 1.38, "grad_norm": 37.65702955657859, "learning_rate": 1.173076102377123e-05, "loss": 0.714, "step": 8860 }, { "epoch": 1.38, "grad_norm": 15.294290677908684, "learning_rate": 1.1729100050559252e-05, "loss": 0.6297, "step": 8861 }, { "epoch": 1.38, "grad_norm": 14.82242889957776, "learning_rate": 1.1727439028172682e-05, "loss": 0.5688, "step": 8862 }, { "epoch": 1.38, "grad_norm": 18.457085996274138, "learning_rate": 1.1725777956658752e-05, "loss": 0.6314, "step": 8863 }, { "epoch": 1.38, "grad_norm": 24.38195712133182, "learning_rate": 1.172411683606471e-05, "loss": 0.7306, "step": 8864 }, { "epoch": 1.38, "grad_norm": 18.683424076152892, "learning_rate": 1.1722455666437793e-05, "loss": 0.643, "step": 8865 }, { "epoch": 1.38, "grad_norm": 17.81138611707172, "learning_rate": 1.1720794447825245e-05, "loss": 0.6803, "step": 8866 }, { "epoch": 1.39, "grad_norm": 13.127828186591692, "learning_rate": 1.171913318027431e-05, "loss": 0.6178, "step": 8867 }, { "epoch": 1.39, "grad_norm": 34.470922097308126, "learning_rate": 1.1717471863832231e-05, "loss": 0.683, "step": 8868 }, { "epoch": 1.39, "grad_norm": 15.81591433011889, "learning_rate": 1.1715810498546259e-05, "loss": 0.5453, "step": 8869 }, { "epoch": 1.39, "grad_norm": 16.681676905898172, "learning_rate": 1.1714149084463638e-05, "loss": 0.6379, "step": 8870 }, { "epoch": 1.39, "grad_norm": 13.621112405574397, "learning_rate": 1.1712487621631621e-05, "loss": 0.6087, "step": 8871 }, { "epoch": 1.39, "grad_norm": 17.987682156107702, "learning_rate": 1.1710826110097457e-05, "loss": 0.5933, "step": 8872 }, { "epoch": 1.39, "grad_norm": 10.562717508864173, "learning_rate": 1.17091645499084e-05, "loss": 0.5919, "step": 8873 }, { "epoch": 1.39, "grad_norm": 18.549489103415727, "learning_rate": 1.1707502941111704e-05, "loss": 0.5727, "step": 8874 }, { "epoch": 1.39, "grad_norm": 12.998997583450262, "learning_rate": 1.1705841283754622e-05, "loss": 0.5683, "step": 8875 }, { "epoch": 1.39, "grad_norm": 17.374022289081182, "learning_rate": 1.170417957788441e-05, "loss": 0.6948, "step": 8876 }, { "epoch": 1.39, "grad_norm": 14.439125885229366, "learning_rate": 1.1702517823548332e-05, "loss": 0.6025, "step": 8877 }, { "epoch": 1.39, "grad_norm": 21.614318817235965, "learning_rate": 1.1700856020793639e-05, "loss": 0.5508, "step": 8878 }, { "epoch": 1.39, "grad_norm": 15.54863954143879, "learning_rate": 1.1699194169667598e-05, "loss": 0.5586, "step": 8879 }, { "epoch": 1.39, "grad_norm": 14.1742295062101, "learning_rate": 1.1697532270217466e-05, "loss": 0.6709, "step": 8880 }, { "epoch": 1.39, "grad_norm": 19.145459674676427, "learning_rate": 1.1695870322490512e-05, "loss": 0.5692, "step": 8881 }, { "epoch": 1.39, "grad_norm": 15.531136037419943, "learning_rate": 1.1694208326533997e-05, "loss": 0.6319, "step": 8882 }, { "epoch": 1.39, "grad_norm": 16.172624563730935, "learning_rate": 1.169254628239519e-05, "loss": 0.5992, "step": 8883 }, { "epoch": 1.39, "grad_norm": 18.61402977301959, "learning_rate": 1.1690884190121356e-05, "loss": 0.6144, "step": 8884 }, { "epoch": 1.39, "grad_norm": 20.10082069893346, "learning_rate": 1.1689222049759765e-05, "loss": 0.6038, "step": 8885 }, { "epoch": 1.39, "grad_norm": 16.112785282709446, "learning_rate": 1.1687559861357685e-05, "loss": 0.6546, "step": 8886 }, { "epoch": 1.39, "grad_norm": 12.74565906233532, "learning_rate": 1.1685897624962392e-05, "loss": 0.5352, "step": 8887 }, { "epoch": 1.39, "grad_norm": 18.639195117563393, "learning_rate": 1.1684235340621155e-05, "loss": 0.6038, "step": 8888 }, { "epoch": 1.39, "grad_norm": 17.48490717796426, "learning_rate": 1.1682573008381252e-05, "loss": 0.6278, "step": 8889 }, { "epoch": 1.39, "grad_norm": 19.658684348570834, "learning_rate": 1.1680910628289956e-05, "loss": 0.6125, "step": 8890 }, { "epoch": 1.39, "grad_norm": 19.835226346700075, "learning_rate": 1.1679248200394546e-05, "loss": 0.7187, "step": 8891 }, { "epoch": 1.39, "grad_norm": 17.99453620761519, "learning_rate": 1.1677585724742298e-05, "loss": 0.6022, "step": 8892 }, { "epoch": 1.39, "grad_norm": 22.909605502560105, "learning_rate": 1.1675923201380493e-05, "loss": 0.622, "step": 8893 }, { "epoch": 1.39, "grad_norm": 14.935382409610092, "learning_rate": 1.1674260630356415e-05, "loss": 0.5615, "step": 8894 }, { "epoch": 1.39, "grad_norm": 21.553511302307438, "learning_rate": 1.167259801171734e-05, "loss": 0.627, "step": 8895 }, { "epoch": 1.39, "grad_norm": 20.566553538810293, "learning_rate": 1.1670935345510561e-05, "loss": 0.5479, "step": 8896 }, { "epoch": 1.39, "grad_norm": 17.748620073649047, "learning_rate": 1.1669272631783354e-05, "loss": 0.5502, "step": 8897 }, { "epoch": 1.39, "grad_norm": 18.17891408769377, "learning_rate": 1.1667609870583012e-05, "loss": 0.6716, "step": 8898 }, { "epoch": 1.39, "grad_norm": 21.30397120615585, "learning_rate": 1.1665947061956821e-05, "loss": 0.6013, "step": 8899 }, { "epoch": 1.39, "grad_norm": 44.10363793773203, "learning_rate": 1.166428420595207e-05, "loss": 0.6198, "step": 8900 }, { "epoch": 1.39, "grad_norm": 19.83430853108769, "learning_rate": 1.166262130261605e-05, "loss": 0.5914, "step": 8901 }, { "epoch": 1.39, "grad_norm": 14.042748495398728, "learning_rate": 1.1660958351996051e-05, "loss": 0.6349, "step": 8902 }, { "epoch": 1.39, "grad_norm": 16.83199374171938, "learning_rate": 1.1659295354139368e-05, "loss": 0.5785, "step": 8903 }, { "epoch": 1.39, "grad_norm": 15.144907496287269, "learning_rate": 1.1657632309093301e-05, "loss": 0.6123, "step": 8904 }, { "epoch": 1.39, "grad_norm": 18.796133662545763, "learning_rate": 1.1655969216905134e-05, "loss": 0.6385, "step": 8905 }, { "epoch": 1.39, "grad_norm": 15.98031227577902, "learning_rate": 1.1654306077622176e-05, "loss": 0.5588, "step": 8906 }, { "epoch": 1.39, "grad_norm": 14.62861061204906, "learning_rate": 1.1652642891291717e-05, "loss": 0.5689, "step": 8907 }, { "epoch": 1.39, "grad_norm": 19.20944149014515, "learning_rate": 1.1650979657961063e-05, "loss": 0.6959, "step": 8908 }, { "epoch": 1.39, "grad_norm": 19.67912927951192, "learning_rate": 1.1649316377677513e-05, "loss": 0.6151, "step": 8909 }, { "epoch": 1.39, "grad_norm": 15.476384363588675, "learning_rate": 1.164765305048837e-05, "loss": 0.5469, "step": 8910 }, { "epoch": 1.39, "grad_norm": 23.744040991437682, "learning_rate": 1.1645989676440938e-05, "loss": 0.6541, "step": 8911 }, { "epoch": 1.39, "grad_norm": 28.569482728186248, "learning_rate": 1.1644326255582523e-05, "loss": 0.6799, "step": 8912 }, { "epoch": 1.39, "grad_norm": 22.658019062560932, "learning_rate": 1.164266278796043e-05, "loss": 0.6158, "step": 8913 }, { "epoch": 1.39, "grad_norm": 12.828817107270327, "learning_rate": 1.1640999273621969e-05, "loss": 0.5203, "step": 8914 }, { "epoch": 1.39, "grad_norm": 16.081379738490785, "learning_rate": 1.1639335712614451e-05, "loss": 0.5795, "step": 8915 }, { "epoch": 1.39, "grad_norm": 19.46435336575244, "learning_rate": 1.1637672104985182e-05, "loss": 0.6023, "step": 8916 }, { "epoch": 1.39, "grad_norm": 24.036083264315902, "learning_rate": 1.1636008450781475e-05, "loss": 0.597, "step": 8917 }, { "epoch": 1.39, "grad_norm": 18.408367795334165, "learning_rate": 1.1634344750050648e-05, "loss": 0.6426, "step": 8918 }, { "epoch": 1.39, "grad_norm": 19.13800700053045, "learning_rate": 1.163268100284001e-05, "loss": 0.5833, "step": 8919 }, { "epoch": 1.39, "grad_norm": 14.642906726672628, "learning_rate": 1.1631017209196878e-05, "loss": 0.5988, "step": 8920 }, { "epoch": 1.39, "grad_norm": 35.048421595169614, "learning_rate": 1.1629353369168574e-05, "loss": 0.6942, "step": 8921 }, { "epoch": 1.39, "grad_norm": 16.33704772617256, "learning_rate": 1.162768948280241e-05, "loss": 0.5742, "step": 8922 }, { "epoch": 1.39, "grad_norm": 19.085674109185405, "learning_rate": 1.1626025550145714e-05, "loss": 0.5937, "step": 8923 }, { "epoch": 1.39, "grad_norm": 20.09362559480258, "learning_rate": 1.16243615712458e-05, "loss": 0.7127, "step": 8924 }, { "epoch": 1.39, "grad_norm": 21.298571636574305, "learning_rate": 1.1622697546149992e-05, "loss": 0.5892, "step": 8925 }, { "epoch": 1.39, "grad_norm": 17.749609536590906, "learning_rate": 1.1621033474905617e-05, "loss": 0.6755, "step": 8926 }, { "epoch": 1.39, "grad_norm": 13.381152615711747, "learning_rate": 1.1619369357559998e-05, "loss": 0.533, "step": 8927 }, { "epoch": 1.39, "grad_norm": 17.306365367498675, "learning_rate": 1.161770519416046e-05, "loss": 0.5943, "step": 8928 }, { "epoch": 1.39, "grad_norm": 18.434557268926508, "learning_rate": 1.1616040984754333e-05, "loss": 0.6459, "step": 8929 }, { "epoch": 1.39, "grad_norm": 16.75157432399176, "learning_rate": 1.1614376729388946e-05, "loss": 0.5812, "step": 8930 }, { "epoch": 1.4, "grad_norm": 12.785748338378092, "learning_rate": 1.161271242811163e-05, "loss": 0.5867, "step": 8931 }, { "epoch": 1.4, "grad_norm": 16.899215954112893, "learning_rate": 1.1611048080969715e-05, "loss": 0.5809, "step": 8932 }, { "epoch": 1.4, "grad_norm": 15.941007743290774, "learning_rate": 1.1609383688010536e-05, "loss": 0.5935, "step": 8933 }, { "epoch": 1.4, "grad_norm": 24.552889559950803, "learning_rate": 1.1607719249281426e-05, "loss": 0.6503, "step": 8934 }, { "epoch": 1.4, "grad_norm": 26.732238530327162, "learning_rate": 1.1606054764829723e-05, "loss": 0.668, "step": 8935 }, { "epoch": 1.4, "grad_norm": 18.736266633633413, "learning_rate": 1.1604390234702758e-05, "loss": 0.5471, "step": 8936 }, { "epoch": 1.4, "grad_norm": 17.658529238310255, "learning_rate": 1.1602725658947873e-05, "loss": 0.5252, "step": 8937 }, { "epoch": 1.4, "grad_norm": 37.67316527591635, "learning_rate": 1.1601061037612413e-05, "loss": 0.5912, "step": 8938 }, { "epoch": 1.4, "grad_norm": 20.72212938866781, "learning_rate": 1.1599396370743707e-05, "loss": 0.6876, "step": 8939 }, { "epoch": 1.4, "grad_norm": 12.025153776868871, "learning_rate": 1.159773165838911e-05, "loss": 0.5517, "step": 8940 }, { "epoch": 1.4, "grad_norm": 26.624871202166297, "learning_rate": 1.1596066900595954e-05, "loss": 0.6611, "step": 8941 }, { "epoch": 1.4, "grad_norm": 17.767632991852402, "learning_rate": 1.1594402097411595e-05, "loss": 0.6476, "step": 8942 }, { "epoch": 1.4, "grad_norm": 24.77659046484016, "learning_rate": 1.159273724888337e-05, "loss": 0.6358, "step": 8943 }, { "epoch": 1.4, "grad_norm": 23.103510608716086, "learning_rate": 1.1591072355058629e-05, "loss": 0.6641, "step": 8944 }, { "epoch": 1.4, "grad_norm": 16.655814361064117, "learning_rate": 1.1589407415984721e-05, "loss": 0.6319, "step": 8945 }, { "epoch": 1.4, "grad_norm": 15.430856029348964, "learning_rate": 1.1587742431708997e-05, "loss": 0.5775, "step": 8946 }, { "epoch": 1.4, "grad_norm": 20.641534294354134, "learning_rate": 1.1586077402278804e-05, "loss": 0.5636, "step": 8947 }, { "epoch": 1.4, "grad_norm": 14.833329482247713, "learning_rate": 1.1584412327741501e-05, "loss": 0.6544, "step": 8948 }, { "epoch": 1.4, "grad_norm": 31.28449221794735, "learning_rate": 1.1582747208144435e-05, "loss": 0.612, "step": 8949 }, { "epoch": 1.4, "grad_norm": 15.982731508595581, "learning_rate": 1.1581082043534968e-05, "loss": 0.6152, "step": 8950 }, { "epoch": 1.4, "grad_norm": 26.69315728320189, "learning_rate": 1.157941683396045e-05, "loss": 0.6447, "step": 8951 }, { "epoch": 1.4, "grad_norm": 17.134849385511526, "learning_rate": 1.1577751579468242e-05, "loss": 0.6076, "step": 8952 }, { "epoch": 1.4, "grad_norm": 25.03793340984905, "learning_rate": 1.1576086280105702e-05, "loss": 0.6063, "step": 8953 }, { "epoch": 1.4, "grad_norm": 17.249844101619445, "learning_rate": 1.1574420935920192e-05, "loss": 0.6639, "step": 8954 }, { "epoch": 1.4, "grad_norm": 14.4426587226952, "learning_rate": 1.157275554695907e-05, "loss": 0.6062, "step": 8955 }, { "epoch": 1.4, "grad_norm": 16.042862773607467, "learning_rate": 1.1571090113269697e-05, "loss": 0.6121, "step": 8956 }, { "epoch": 1.4, "grad_norm": 16.273812716384683, "learning_rate": 1.1569424634899441e-05, "loss": 0.6205, "step": 8957 }, { "epoch": 1.4, "grad_norm": 18.980340108535618, "learning_rate": 1.156775911189567e-05, "loss": 0.5797, "step": 8958 }, { "epoch": 1.4, "grad_norm": 14.760529708721084, "learning_rate": 1.1566093544305747e-05, "loss": 0.6132, "step": 8959 }, { "epoch": 1.4, "grad_norm": 20.046573704293593, "learning_rate": 1.1564427932177037e-05, "loss": 0.6122, "step": 8960 }, { "epoch": 1.4, "grad_norm": 17.435986591621763, "learning_rate": 1.1562762275556916e-05, "loss": 0.615, "step": 8961 }, { "epoch": 1.4, "grad_norm": 20.714940123573943, "learning_rate": 1.1561096574492745e-05, "loss": 0.5803, "step": 8962 }, { "epoch": 1.4, "grad_norm": 12.705839174480742, "learning_rate": 1.1559430829031905e-05, "loss": 0.5994, "step": 8963 }, { "epoch": 1.4, "grad_norm": 14.009704724260164, "learning_rate": 1.155776503922176e-05, "loss": 0.5652, "step": 8964 }, { "epoch": 1.4, "grad_norm": 17.796618555212422, "learning_rate": 1.1556099205109694e-05, "loss": 0.5592, "step": 8965 }, { "epoch": 1.4, "grad_norm": 17.947415085453642, "learning_rate": 1.155443332674307e-05, "loss": 0.5901, "step": 8966 }, { "epoch": 1.4, "grad_norm": 12.57799767968796, "learning_rate": 1.1552767404169281e-05, "loss": 0.5742, "step": 8967 }, { "epoch": 1.4, "grad_norm": 15.846545519231139, "learning_rate": 1.1551101437435686e-05, "loss": 0.5587, "step": 8968 }, { "epoch": 1.4, "grad_norm": 19.203446933385028, "learning_rate": 1.1549435426589678e-05, "loss": 0.6159, "step": 8969 }, { "epoch": 1.4, "grad_norm": 17.861264901860107, "learning_rate": 1.1547769371678632e-05, "loss": 0.6122, "step": 8970 }, { "epoch": 1.4, "grad_norm": 27.995382434851084, "learning_rate": 1.1546103272749931e-05, "loss": 0.5866, "step": 8971 }, { "epoch": 1.4, "grad_norm": 21.95391270082982, "learning_rate": 1.1544437129850956e-05, "loss": 0.515, "step": 8972 }, { "epoch": 1.4, "grad_norm": 15.54622634895558, "learning_rate": 1.1542770943029096e-05, "loss": 0.5953, "step": 8973 }, { "epoch": 1.4, "grad_norm": 17.626418754952542, "learning_rate": 1.1541104712331728e-05, "loss": 0.6332, "step": 8974 }, { "epoch": 1.4, "grad_norm": 20.772675761717576, "learning_rate": 1.1539438437806245e-05, "loss": 0.7193, "step": 8975 }, { "epoch": 1.4, "grad_norm": 14.878241759765308, "learning_rate": 1.1537772119500035e-05, "loss": 0.6371, "step": 8976 }, { "epoch": 1.4, "grad_norm": 19.42295789116523, "learning_rate": 1.1536105757460485e-05, "loss": 0.5887, "step": 8977 }, { "epoch": 1.4, "grad_norm": 16.367424480206616, "learning_rate": 1.1534439351734986e-05, "loss": 0.6164, "step": 8978 }, { "epoch": 1.4, "grad_norm": 14.34381620775164, "learning_rate": 1.1532772902370929e-05, "loss": 0.5721, "step": 8979 }, { "epoch": 1.4, "grad_norm": 16.17088806577735, "learning_rate": 1.1531106409415706e-05, "loss": 0.6079, "step": 8980 }, { "epoch": 1.4, "grad_norm": 25.690804804924554, "learning_rate": 1.1529439872916712e-05, "loss": 0.6134, "step": 8981 }, { "epoch": 1.4, "grad_norm": 16.90920049372513, "learning_rate": 1.1527773292921347e-05, "loss": 0.5996, "step": 8982 }, { "epoch": 1.4, "grad_norm": 21.57414107107544, "learning_rate": 1.1526106669476999e-05, "loss": 0.6256, "step": 8983 }, { "epoch": 1.4, "grad_norm": 13.237908356120657, "learning_rate": 1.1524440002631073e-05, "loss": 0.6122, "step": 8984 }, { "epoch": 1.4, "grad_norm": 16.84641441723242, "learning_rate": 1.1522773292430963e-05, "loss": 0.6135, "step": 8985 }, { "epoch": 1.4, "grad_norm": 22.795206548805538, "learning_rate": 1.1521106538924071e-05, "loss": 0.6405, "step": 8986 }, { "epoch": 1.4, "grad_norm": 15.33282839749868, "learning_rate": 1.15194397421578e-05, "loss": 0.5875, "step": 8987 }, { "epoch": 1.4, "grad_norm": 21.667473735969466, "learning_rate": 1.1517772902179554e-05, "loss": 0.6352, "step": 8988 }, { "epoch": 1.4, "grad_norm": 15.271874362929152, "learning_rate": 1.1516106019036733e-05, "loss": 0.5384, "step": 8989 }, { "epoch": 1.4, "grad_norm": 15.605000335146926, "learning_rate": 1.1514439092776741e-05, "loss": 0.6122, "step": 8990 }, { "epoch": 1.4, "grad_norm": 16.496008624354705, "learning_rate": 1.1512772123446988e-05, "loss": 0.5473, "step": 8991 }, { "epoch": 1.4, "grad_norm": 21.405462524102404, "learning_rate": 1.1511105111094884e-05, "loss": 0.6454, "step": 8992 }, { "epoch": 1.4, "grad_norm": 12.47929706894094, "learning_rate": 1.150943805576783e-05, "loss": 0.5362, "step": 8993 }, { "epoch": 1.4, "grad_norm": 13.107262922064121, "learning_rate": 1.1507770957513245e-05, "loss": 0.6412, "step": 8994 }, { "epoch": 1.41, "grad_norm": 19.85739110563627, "learning_rate": 1.1506103816378533e-05, "loss": 0.6541, "step": 8995 }, { "epoch": 1.41, "grad_norm": 19.063767656794987, "learning_rate": 1.1504436632411112e-05, "loss": 0.5636, "step": 8996 }, { "epoch": 1.41, "grad_norm": 15.551905877143591, "learning_rate": 1.150276940565839e-05, "loss": 0.5732, "step": 8997 }, { "epoch": 1.41, "grad_norm": 19.483792454592585, "learning_rate": 1.1501102136167788e-05, "loss": 0.6391, "step": 8998 }, { "epoch": 1.41, "grad_norm": 21.324170060319688, "learning_rate": 1.1499434823986719e-05, "loss": 0.6567, "step": 8999 }, { "epoch": 1.41, "grad_norm": 15.33599461930103, "learning_rate": 1.1497767469162598e-05, "loss": 0.5386, "step": 9000 }, { "epoch": 1.41, "grad_norm": 18.664917872369813, "learning_rate": 1.1496100071742849e-05, "loss": 0.5914, "step": 9001 }, { "epoch": 1.41, "grad_norm": 23.243480797696296, "learning_rate": 1.1494432631774888e-05, "loss": 0.5619, "step": 9002 }, { "epoch": 1.41, "grad_norm": 12.05223211151562, "learning_rate": 1.1492765149306137e-05, "loss": 0.5855, "step": 9003 }, { "epoch": 1.41, "grad_norm": 17.276544287648605, "learning_rate": 1.149109762438402e-05, "loss": 0.5184, "step": 9004 }, { "epoch": 1.41, "grad_norm": 23.335126113067542, "learning_rate": 1.1489430057055959e-05, "loss": 0.5605, "step": 9005 }, { "epoch": 1.41, "grad_norm": 35.99915808737943, "learning_rate": 1.1487762447369375e-05, "loss": 0.6487, "step": 9006 }, { "epoch": 1.41, "grad_norm": 25.421618981411598, "learning_rate": 1.1486094795371703e-05, "loss": 0.6516, "step": 9007 }, { "epoch": 1.41, "grad_norm": 17.423986668758584, "learning_rate": 1.1484427101110359e-05, "loss": 0.6055, "step": 9008 }, { "epoch": 1.41, "grad_norm": 16.111500551723328, "learning_rate": 1.148275936463278e-05, "loss": 0.6073, "step": 9009 }, { "epoch": 1.41, "grad_norm": 17.382029610480792, "learning_rate": 1.148109158598639e-05, "loss": 0.6652, "step": 9010 }, { "epoch": 1.41, "grad_norm": 22.41212379704429, "learning_rate": 1.1479423765218625e-05, "loss": 0.6017, "step": 9011 }, { "epoch": 1.41, "grad_norm": 15.243962914789135, "learning_rate": 1.1477755902376908e-05, "loss": 0.5556, "step": 9012 }, { "epoch": 1.41, "grad_norm": 17.75227774273833, "learning_rate": 1.147608799750868e-05, "loss": 0.6113, "step": 9013 }, { "epoch": 1.41, "grad_norm": 20.742528265328865, "learning_rate": 1.1474420050661374e-05, "loss": 0.6155, "step": 9014 }, { "epoch": 1.41, "grad_norm": 16.653549596306117, "learning_rate": 1.1472752061882426e-05, "loss": 0.5651, "step": 9015 }, { "epoch": 1.41, "grad_norm": 16.244776740031117, "learning_rate": 1.1471084031219268e-05, "loss": 0.5875, "step": 9016 }, { "epoch": 1.41, "grad_norm": 18.00375170692068, "learning_rate": 1.146941595871934e-05, "loss": 0.5834, "step": 9017 }, { "epoch": 1.41, "grad_norm": 19.79248906652147, "learning_rate": 1.1467747844430082e-05, "loss": 0.5934, "step": 9018 }, { "epoch": 1.41, "grad_norm": 20.18208711647645, "learning_rate": 1.1466079688398937e-05, "loss": 0.5653, "step": 9019 }, { "epoch": 1.41, "grad_norm": 15.989205817157451, "learning_rate": 1.146441149067334e-05, "loss": 0.6278, "step": 9020 }, { "epoch": 1.41, "grad_norm": 21.380178522762158, "learning_rate": 1.1462743251300739e-05, "loss": 0.7111, "step": 9021 }, { "epoch": 1.41, "grad_norm": 13.215860721788788, "learning_rate": 1.1461074970328574e-05, "loss": 0.5509, "step": 9022 }, { "epoch": 1.41, "grad_norm": 21.994802569519802, "learning_rate": 1.1459406647804295e-05, "loss": 0.6439, "step": 9023 }, { "epoch": 1.41, "grad_norm": 17.916998712009644, "learning_rate": 1.1457738283775339e-05, "loss": 0.5621, "step": 9024 }, { "epoch": 1.41, "grad_norm": 16.976651007592288, "learning_rate": 1.1456069878289161e-05, "loss": 0.6414, "step": 9025 }, { "epoch": 1.41, "grad_norm": 17.180718418326556, "learning_rate": 1.1454401431393212e-05, "loss": 0.5068, "step": 9026 }, { "epoch": 1.41, "grad_norm": 28.58602278530663, "learning_rate": 1.145273294313493e-05, "loss": 0.5983, "step": 9027 }, { "epoch": 1.41, "grad_norm": 26.57496777821415, "learning_rate": 1.1451064413561776e-05, "loss": 0.6932, "step": 9028 }, { "epoch": 1.41, "grad_norm": 17.214888109323557, "learning_rate": 1.1449395842721201e-05, "loss": 0.5989, "step": 9029 }, { "epoch": 1.41, "grad_norm": 21.505302495435384, "learning_rate": 1.1447727230660654e-05, "loss": 0.6125, "step": 9030 }, { "epoch": 1.41, "grad_norm": 22.371601697784598, "learning_rate": 1.1446058577427592e-05, "loss": 0.5455, "step": 9031 }, { "epoch": 1.41, "grad_norm": 31.773109916459667, "learning_rate": 1.1444389883069471e-05, "loss": 0.6116, "step": 9032 }, { "epoch": 1.41, "grad_norm": 18.058310238120455, "learning_rate": 1.1442721147633744e-05, "loss": 0.6278, "step": 9033 }, { "epoch": 1.41, "grad_norm": 19.038650264041706, "learning_rate": 1.1441052371167876e-05, "loss": 0.6426, "step": 9034 }, { "epoch": 1.41, "grad_norm": 17.568615878545533, "learning_rate": 1.1439383553719315e-05, "loss": 0.5955, "step": 9035 }, { "epoch": 1.41, "grad_norm": 14.927676643839913, "learning_rate": 1.1437714695335534e-05, "loss": 0.5605, "step": 9036 }, { "epoch": 1.41, "grad_norm": 24.555699644064862, "learning_rate": 1.1436045796063983e-05, "loss": 0.6343, "step": 9037 }, { "epoch": 1.41, "grad_norm": 19.433361387157586, "learning_rate": 1.1434376855952136e-05, "loss": 0.5932, "step": 9038 }, { "epoch": 1.41, "grad_norm": 18.801446360782016, "learning_rate": 1.1432707875047445e-05, "loss": 0.6008, "step": 9039 }, { "epoch": 1.41, "grad_norm": 12.820679051484227, "learning_rate": 1.143103885339738e-05, "loss": 0.6079, "step": 9040 }, { "epoch": 1.41, "grad_norm": 17.998013283692515, "learning_rate": 1.1429369791049409e-05, "loss": 0.6169, "step": 9041 }, { "epoch": 1.41, "grad_norm": 17.860633692269566, "learning_rate": 1.1427700688050998e-05, "loss": 0.5423, "step": 9042 }, { "epoch": 1.41, "grad_norm": 14.42527479279007, "learning_rate": 1.1426031544449614e-05, "loss": 0.5335, "step": 9043 }, { "epoch": 1.41, "grad_norm": 17.489213408677326, "learning_rate": 1.1424362360292725e-05, "loss": 0.5786, "step": 9044 }, { "epoch": 1.41, "grad_norm": 17.797013522957908, "learning_rate": 1.1422693135627804e-05, "loss": 0.5931, "step": 9045 }, { "epoch": 1.41, "grad_norm": 22.216643216115255, "learning_rate": 1.1421023870502324e-05, "loss": 0.5823, "step": 9046 }, { "epoch": 1.41, "grad_norm": 14.452861149415751, "learning_rate": 1.1419354564963756e-05, "loss": 0.6471, "step": 9047 }, { "epoch": 1.41, "grad_norm": 16.712113179015002, "learning_rate": 1.1417685219059576e-05, "loss": 0.6075, "step": 9048 }, { "epoch": 1.41, "grad_norm": 17.09751338432782, "learning_rate": 1.1416015832837258e-05, "loss": 0.601, "step": 9049 }, { "epoch": 1.41, "grad_norm": 21.08564511640922, "learning_rate": 1.1414346406344272e-05, "loss": 0.6791, "step": 9050 }, { "epoch": 1.41, "grad_norm": 20.037675992504557, "learning_rate": 1.1412676939628108e-05, "loss": 0.6227, "step": 9051 }, { "epoch": 1.41, "grad_norm": 19.526436426946656, "learning_rate": 1.1411007432736236e-05, "loss": 0.6669, "step": 9052 }, { "epoch": 1.41, "grad_norm": 14.49169064310133, "learning_rate": 1.1409337885716141e-05, "loss": 0.6462, "step": 9053 }, { "epoch": 1.41, "grad_norm": 18.349309727778657, "learning_rate": 1.1407668298615296e-05, "loss": 0.5469, "step": 9054 }, { "epoch": 1.41, "grad_norm": 20.755873802426294, "learning_rate": 1.1405998671481191e-05, "loss": 0.6872, "step": 9055 }, { "epoch": 1.41, "grad_norm": 14.277805061036986, "learning_rate": 1.1404329004361306e-05, "loss": 0.6122, "step": 9056 }, { "epoch": 1.41, "grad_norm": 17.948319764812553, "learning_rate": 1.1402659297303124e-05, "loss": 0.6444, "step": 9057 }, { "epoch": 1.41, "grad_norm": 18.131996583476063, "learning_rate": 1.1400989550354133e-05, "loss": 0.5869, "step": 9058 }, { "epoch": 1.42, "grad_norm": 16.669039271476105, "learning_rate": 1.1399319763561821e-05, "loss": 0.6279, "step": 9059 }, { "epoch": 1.42, "grad_norm": 24.808556724404184, "learning_rate": 1.1397649936973672e-05, "loss": 0.6098, "step": 9060 }, { "epoch": 1.42, "grad_norm": 28.779127564593985, "learning_rate": 1.1395980070637175e-05, "loss": 0.6201, "step": 9061 }, { "epoch": 1.42, "grad_norm": 13.949160058980471, "learning_rate": 1.139431016459982e-05, "loss": 0.6068, "step": 9062 }, { "epoch": 1.42, "grad_norm": 21.81452301011464, "learning_rate": 1.1392640218909103e-05, "loss": 0.5838, "step": 9063 }, { "epoch": 1.42, "grad_norm": 22.88530604372359, "learning_rate": 1.1390970233612513e-05, "loss": 0.5541, "step": 9064 }, { "epoch": 1.42, "grad_norm": 16.12488839846441, "learning_rate": 1.138930020875754e-05, "loss": 0.6342, "step": 9065 }, { "epoch": 1.42, "grad_norm": 17.314105152263146, "learning_rate": 1.1387630144391685e-05, "loss": 0.5846, "step": 9066 }, { "epoch": 1.42, "grad_norm": 20.729774355647102, "learning_rate": 1.138596004056244e-05, "loss": 0.6847, "step": 9067 }, { "epoch": 1.42, "grad_norm": 16.456214600661525, "learning_rate": 1.1384289897317302e-05, "loss": 0.5532, "step": 9068 }, { "epoch": 1.42, "grad_norm": 15.074690685609944, "learning_rate": 1.1382619714703767e-05, "loss": 0.5645, "step": 9069 }, { "epoch": 1.42, "grad_norm": 15.098515367923971, "learning_rate": 1.1380949492769339e-05, "loss": 0.5844, "step": 9070 }, { "epoch": 1.42, "grad_norm": 13.28447161701956, "learning_rate": 1.1379279231561514e-05, "loss": 0.6042, "step": 9071 }, { "epoch": 1.42, "grad_norm": 18.510734081140757, "learning_rate": 1.1377608931127792e-05, "loss": 0.6153, "step": 9072 }, { "epoch": 1.42, "grad_norm": 28.839027655715256, "learning_rate": 1.137593859151568e-05, "loss": 0.6751, "step": 9073 }, { "epoch": 1.42, "grad_norm": 16.811467209898392, "learning_rate": 1.137426821277268e-05, "loss": 0.68, "step": 9074 }, { "epoch": 1.42, "grad_norm": 22.72735095959442, "learning_rate": 1.1372597794946298e-05, "loss": 0.6713, "step": 9075 }, { "epoch": 1.42, "grad_norm": 20.814761549418616, "learning_rate": 1.1370927338084035e-05, "loss": 0.5791, "step": 9076 }, { "epoch": 1.42, "grad_norm": 19.49466500218632, "learning_rate": 1.1369256842233399e-05, "loss": 0.6257, "step": 9077 }, { "epoch": 1.42, "grad_norm": 13.443216470792667, "learning_rate": 1.1367586307441902e-05, "loss": 0.6048, "step": 9078 }, { "epoch": 1.42, "grad_norm": 21.067267870846962, "learning_rate": 1.136591573375705e-05, "loss": 0.6292, "step": 9079 }, { "epoch": 1.42, "grad_norm": 17.285803065725048, "learning_rate": 1.1364245121226354e-05, "loss": 0.6121, "step": 9080 }, { "epoch": 1.42, "grad_norm": 14.894933058358427, "learning_rate": 1.1362574469897322e-05, "loss": 0.6188, "step": 9081 }, { "epoch": 1.42, "grad_norm": 29.614628034449026, "learning_rate": 1.136090377981747e-05, "loss": 0.6406, "step": 9082 }, { "epoch": 1.42, "grad_norm": 14.403628824033497, "learning_rate": 1.1359233051034316e-05, "loss": 0.6175, "step": 9083 }, { "epoch": 1.42, "grad_norm": 26.84342563830639, "learning_rate": 1.1357562283595363e-05, "loss": 0.6301, "step": 9084 }, { "epoch": 1.42, "grad_norm": 23.60368557270249, "learning_rate": 1.1355891477548136e-05, "loss": 0.6775, "step": 9085 }, { "epoch": 1.42, "grad_norm": 23.069978213576917, "learning_rate": 1.1354220632940147e-05, "loss": 0.6347, "step": 9086 }, { "epoch": 1.42, "grad_norm": 21.572896261741263, "learning_rate": 1.1352549749818916e-05, "loss": 0.6391, "step": 9087 }, { "epoch": 1.42, "grad_norm": 13.69603869095432, "learning_rate": 1.135087882823196e-05, "loss": 0.5302, "step": 9088 }, { "epoch": 1.42, "grad_norm": 25.01557777371735, "learning_rate": 1.1349207868226801e-05, "loss": 0.5949, "step": 9089 }, { "epoch": 1.42, "grad_norm": 15.342854894218238, "learning_rate": 1.1347536869850959e-05, "loss": 0.7235, "step": 9090 }, { "epoch": 1.42, "grad_norm": 21.25456287398254, "learning_rate": 1.134586583315196e-05, "loss": 0.6007, "step": 9091 }, { "epoch": 1.42, "grad_norm": 19.083450642548637, "learning_rate": 1.134419475817732e-05, "loss": 0.638, "step": 9092 }, { "epoch": 1.42, "grad_norm": 16.231451875127494, "learning_rate": 1.1342523644974567e-05, "loss": 0.632, "step": 9093 }, { "epoch": 1.42, "grad_norm": 19.93877235301933, "learning_rate": 1.134085249359123e-05, "loss": 0.6454, "step": 9094 }, { "epoch": 1.42, "grad_norm": 22.033638809775347, "learning_rate": 1.1339181304074833e-05, "loss": 0.647, "step": 9095 }, { "epoch": 1.42, "grad_norm": 29.423520161219532, "learning_rate": 1.13375100764729e-05, "loss": 0.6064, "step": 9096 }, { "epoch": 1.42, "grad_norm": 28.54665340897153, "learning_rate": 1.1335838810832966e-05, "loss": 0.7019, "step": 9097 }, { "epoch": 1.42, "grad_norm": 22.72381920385396, "learning_rate": 1.1334167507202553e-05, "loss": 0.5706, "step": 9098 }, { "epoch": 1.42, "grad_norm": 38.26470266899626, "learning_rate": 1.1332496165629201e-05, "loss": 0.6931, "step": 9099 }, { "epoch": 1.42, "grad_norm": 17.353207287815692, "learning_rate": 1.1330824786160437e-05, "loss": 0.6489, "step": 9100 }, { "epoch": 1.42, "grad_norm": 20.343490737370498, "learning_rate": 1.1329153368843793e-05, "loss": 0.6839, "step": 9101 }, { "epoch": 1.42, "grad_norm": 21.03310480837816, "learning_rate": 1.1327481913726806e-05, "loss": 0.6535, "step": 9102 }, { "epoch": 1.42, "grad_norm": 12.648559477832315, "learning_rate": 1.1325810420857009e-05, "loss": 0.617, "step": 9103 }, { "epoch": 1.42, "grad_norm": 13.355797718121442, "learning_rate": 1.1324138890281937e-05, "loss": 0.6064, "step": 9104 }, { "epoch": 1.42, "grad_norm": 22.46330109342206, "learning_rate": 1.1322467322049135e-05, "loss": 0.5981, "step": 9105 }, { "epoch": 1.42, "grad_norm": 14.75941963592894, "learning_rate": 1.1320795716206132e-05, "loss": 0.7807, "step": 9106 }, { "epoch": 1.42, "grad_norm": 21.167837482896456, "learning_rate": 1.1319124072800471e-05, "loss": 0.5619, "step": 9107 }, { "epoch": 1.42, "grad_norm": 34.71152468366489, "learning_rate": 1.1317452391879695e-05, "loss": 0.5719, "step": 9108 }, { "epoch": 1.42, "grad_norm": 14.4066642659776, "learning_rate": 1.1315780673491344e-05, "loss": 0.6372, "step": 9109 }, { "epoch": 1.42, "grad_norm": 16.55319533344844, "learning_rate": 1.131410891768296e-05, "loss": 0.6287, "step": 9110 }, { "epoch": 1.42, "grad_norm": 26.391365222701964, "learning_rate": 1.1312437124502086e-05, "loss": 0.6055, "step": 9111 }, { "epoch": 1.42, "grad_norm": 19.06958664821942, "learning_rate": 1.131076529399627e-05, "loss": 0.5371, "step": 9112 }, { "epoch": 1.42, "grad_norm": 17.792279032666656, "learning_rate": 1.1309093426213051e-05, "loss": 0.6333, "step": 9113 }, { "epoch": 1.42, "grad_norm": 23.371635136812454, "learning_rate": 1.1307421521199988e-05, "loss": 0.6905, "step": 9114 }, { "epoch": 1.42, "grad_norm": 19.001999132647843, "learning_rate": 1.1305749579004618e-05, "loss": 0.5812, "step": 9115 }, { "epoch": 1.42, "grad_norm": 15.828004745260056, "learning_rate": 1.1304077599674493e-05, "loss": 0.6056, "step": 9116 }, { "epoch": 1.42, "grad_norm": 19.639540626435203, "learning_rate": 1.1302405583257163e-05, "loss": 0.6281, "step": 9117 }, { "epoch": 1.42, "grad_norm": 19.16487097871779, "learning_rate": 1.1300733529800183e-05, "loss": 0.5788, "step": 9118 }, { "epoch": 1.42, "grad_norm": 23.968955207577906, "learning_rate": 1.1299061439351102e-05, "loss": 0.6387, "step": 9119 }, { "epoch": 1.42, "grad_norm": 14.807406211577261, "learning_rate": 1.1297389311957472e-05, "loss": 0.5559, "step": 9120 }, { "epoch": 1.42, "grad_norm": 18.176759074569787, "learning_rate": 1.1295717147666848e-05, "loss": 0.6406, "step": 9121 }, { "epoch": 1.42, "grad_norm": 14.275687869233185, "learning_rate": 1.129404494652679e-05, "loss": 0.6397, "step": 9122 }, { "epoch": 1.43, "grad_norm": 20.290152658163905, "learning_rate": 1.1292372708584847e-05, "loss": 0.7365, "step": 9123 }, { "epoch": 1.43, "grad_norm": 14.600511221700568, "learning_rate": 1.1290700433888584e-05, "loss": 0.6836, "step": 9124 }, { "epoch": 1.43, "grad_norm": 19.36856989679862, "learning_rate": 1.1289028122485553e-05, "loss": 0.6109, "step": 9125 }, { "epoch": 1.43, "grad_norm": 18.96650136820152, "learning_rate": 1.1287355774423316e-05, "loss": 0.6098, "step": 9126 }, { "epoch": 1.43, "grad_norm": 21.440399361721234, "learning_rate": 1.1285683389749434e-05, "loss": 0.6564, "step": 9127 }, { "epoch": 1.43, "grad_norm": 18.37273353754732, "learning_rate": 1.128401096851147e-05, "loss": 0.6347, "step": 9128 }, { "epoch": 1.43, "grad_norm": 28.22226318455248, "learning_rate": 1.1282338510756985e-05, "loss": 0.6322, "step": 9129 }, { "epoch": 1.43, "grad_norm": 14.428904535895661, "learning_rate": 1.1280666016533542e-05, "loss": 0.5889, "step": 9130 }, { "epoch": 1.43, "grad_norm": 16.379043486184198, "learning_rate": 1.1278993485888704e-05, "loss": 0.6049, "step": 9131 }, { "epoch": 1.43, "grad_norm": 15.240538401824047, "learning_rate": 1.1277320918870043e-05, "loss": 0.617, "step": 9132 }, { "epoch": 1.43, "grad_norm": 20.327043346213088, "learning_rate": 1.1275648315525123e-05, "loss": 0.6593, "step": 9133 }, { "epoch": 1.43, "grad_norm": 18.678445781150614, "learning_rate": 1.127397567590151e-05, "loss": 0.6154, "step": 9134 }, { "epoch": 1.43, "grad_norm": 20.86344744757711, "learning_rate": 1.1272303000046774e-05, "loss": 0.6913, "step": 9135 }, { "epoch": 1.43, "grad_norm": 16.504715011914662, "learning_rate": 1.1270630288008484e-05, "loss": 0.6695, "step": 9136 }, { "epoch": 1.43, "grad_norm": 15.092482862634203, "learning_rate": 1.1268957539834213e-05, "loss": 0.6672, "step": 9137 }, { "epoch": 1.43, "grad_norm": 18.48369333117928, "learning_rate": 1.1267284755571528e-05, "loss": 0.5981, "step": 9138 }, { "epoch": 1.43, "grad_norm": 15.308538836365852, "learning_rate": 1.1265611935268013e-05, "loss": 0.5965, "step": 9139 }, { "epoch": 1.43, "grad_norm": 14.073746797439478, "learning_rate": 1.1263939078971229e-05, "loss": 0.6243, "step": 9140 }, { "epoch": 1.43, "grad_norm": 21.810150379953935, "learning_rate": 1.1262266186728763e-05, "loss": 0.6773, "step": 9141 }, { "epoch": 1.43, "grad_norm": 13.019139879984305, "learning_rate": 1.1260593258588179e-05, "loss": 0.5797, "step": 9142 }, { "epoch": 1.43, "grad_norm": 14.68618465276088, "learning_rate": 1.1258920294597063e-05, "loss": 0.5153, "step": 9143 }, { "epoch": 1.43, "grad_norm": 21.780031276852068, "learning_rate": 1.1257247294802988e-05, "loss": 0.5815, "step": 9144 }, { "epoch": 1.43, "grad_norm": 15.872738059414592, "learning_rate": 1.125557425925354e-05, "loss": 0.5487, "step": 9145 }, { "epoch": 1.43, "grad_norm": 22.979703139241888, "learning_rate": 1.125390118799629e-05, "loss": 0.6997, "step": 9146 }, { "epoch": 1.43, "grad_norm": 20.524404827507347, "learning_rate": 1.1252228081078826e-05, "loss": 0.5704, "step": 9147 }, { "epoch": 1.43, "grad_norm": 16.949117590326956, "learning_rate": 1.1250554938548726e-05, "loss": 0.6567, "step": 9148 }, { "epoch": 1.43, "grad_norm": 22.889068687313657, "learning_rate": 1.1248881760453578e-05, "loss": 0.6515, "step": 9149 }, { "epoch": 1.43, "grad_norm": 16.429801151464815, "learning_rate": 1.124720854684096e-05, "loss": 0.6001, "step": 9150 }, { "epoch": 1.43, "grad_norm": 13.819143058741489, "learning_rate": 1.1245535297758463e-05, "loss": 0.5381, "step": 9151 }, { "epoch": 1.43, "grad_norm": 15.710729237204703, "learning_rate": 1.124386201325367e-05, "loss": 0.6662, "step": 9152 }, { "epoch": 1.43, "grad_norm": 11.605071221177864, "learning_rate": 1.124218869337417e-05, "loss": 0.5742, "step": 9153 }, { "epoch": 1.43, "grad_norm": 18.094322992716332, "learning_rate": 1.1240515338167548e-05, "loss": 0.5979, "step": 9154 }, { "epoch": 1.43, "grad_norm": 15.685898919014859, "learning_rate": 1.1238841947681399e-05, "loss": 0.5981, "step": 9155 }, { "epoch": 1.43, "grad_norm": 17.8066106956894, "learning_rate": 1.1237168521963307e-05, "loss": 0.5302, "step": 9156 }, { "epoch": 1.43, "grad_norm": 18.51211495266274, "learning_rate": 1.1235495061060864e-05, "loss": 0.5423, "step": 9157 }, { "epoch": 1.43, "grad_norm": 14.017995690202573, "learning_rate": 1.123382156502167e-05, "loss": 0.5371, "step": 9158 }, { "epoch": 1.43, "grad_norm": 12.274334587612588, "learning_rate": 1.1232148033893307e-05, "loss": 0.4745, "step": 9159 }, { "epoch": 1.43, "grad_norm": 21.28087899645096, "learning_rate": 1.1230474467723376e-05, "loss": 0.594, "step": 9160 }, { "epoch": 1.43, "grad_norm": 22.606996366315286, "learning_rate": 1.1228800866559472e-05, "loss": 0.5989, "step": 9161 }, { "epoch": 1.43, "grad_norm": 18.246749941264525, "learning_rate": 1.1227127230449191e-05, "loss": 0.5773, "step": 9162 }, { "epoch": 1.43, "grad_norm": 11.688941519878224, "learning_rate": 1.1225453559440128e-05, "loss": 0.6094, "step": 9163 }, { "epoch": 1.43, "grad_norm": 16.601185173683334, "learning_rate": 1.1223779853579884e-05, "loss": 0.6012, "step": 9164 }, { "epoch": 1.43, "grad_norm": 13.220288273403998, "learning_rate": 1.1222106112916053e-05, "loss": 0.6064, "step": 9165 }, { "epoch": 1.43, "grad_norm": 22.02275981584391, "learning_rate": 1.1220432337496244e-05, "loss": 0.6672, "step": 9166 }, { "epoch": 1.43, "grad_norm": 28.058959748511075, "learning_rate": 1.121875852736805e-05, "loss": 0.5952, "step": 9167 }, { "epoch": 1.43, "grad_norm": 17.952123799971428, "learning_rate": 1.1217084682579077e-05, "loss": 0.6151, "step": 9168 }, { "epoch": 1.43, "grad_norm": 19.647652856467833, "learning_rate": 1.1215410803176927e-05, "loss": 0.5902, "step": 9169 }, { "epoch": 1.43, "grad_norm": 22.473547970343066, "learning_rate": 1.1213736889209207e-05, "loss": 0.6666, "step": 9170 }, { "epoch": 1.43, "grad_norm": 23.354751545277264, "learning_rate": 1.1212062940723519e-05, "loss": 0.5764, "step": 9171 }, { "epoch": 1.43, "grad_norm": 26.54377848980917, "learning_rate": 1.1210388957767471e-05, "loss": 0.5804, "step": 9172 }, { "epoch": 1.43, "grad_norm": 17.864543439766816, "learning_rate": 1.1208714940388668e-05, "loss": 0.6083, "step": 9173 }, { "epoch": 1.43, "grad_norm": 17.573375387447467, "learning_rate": 1.1207040888634719e-05, "loss": 0.5952, "step": 9174 }, { "epoch": 1.43, "grad_norm": 20.071871097091563, "learning_rate": 1.1205366802553231e-05, "loss": 0.5872, "step": 9175 }, { "epoch": 1.43, "grad_norm": 18.281087903176285, "learning_rate": 1.1203692682191819e-05, "loss": 0.6939, "step": 9176 }, { "epoch": 1.43, "grad_norm": 15.202771754635325, "learning_rate": 1.1202018527598091e-05, "loss": 0.6457, "step": 9177 }, { "epoch": 1.43, "grad_norm": 20.67415285283405, "learning_rate": 1.120034433881966e-05, "loss": 0.59, "step": 9178 }, { "epoch": 1.43, "grad_norm": 19.601946449548517, "learning_rate": 1.1198670115904136e-05, "loss": 0.6297, "step": 9179 }, { "epoch": 1.43, "grad_norm": 30.470285551551125, "learning_rate": 1.1196995858899138e-05, "loss": 0.6937, "step": 9180 }, { "epoch": 1.43, "grad_norm": 14.413605388640091, "learning_rate": 1.1195321567852275e-05, "loss": 0.5675, "step": 9181 }, { "epoch": 1.43, "grad_norm": 18.069825486226854, "learning_rate": 1.1193647242811165e-05, "loss": 0.5391, "step": 9182 }, { "epoch": 1.43, "grad_norm": 11.292613018269247, "learning_rate": 1.1191972883823433e-05, "loss": 0.5359, "step": 9183 }, { "epoch": 1.43, "grad_norm": 14.862600239254476, "learning_rate": 1.119029849093668e-05, "loss": 0.647, "step": 9184 }, { "epoch": 1.43, "grad_norm": 25.713910930367874, "learning_rate": 1.118862406419854e-05, "loss": 0.5467, "step": 9185 }, { "epoch": 1.43, "grad_norm": 23.153794562571782, "learning_rate": 1.1186949603656624e-05, "loss": 0.7108, "step": 9186 }, { "epoch": 1.44, "grad_norm": 19.312513565935227, "learning_rate": 1.1185275109358558e-05, "loss": 0.5972, "step": 9187 }, { "epoch": 1.44, "grad_norm": 19.569038125764617, "learning_rate": 1.118360058135196e-05, "loss": 0.6932, "step": 9188 }, { "epoch": 1.44, "grad_norm": 29.725426942939055, "learning_rate": 1.1181926019684454e-05, "loss": 0.6175, "step": 9189 }, { "epoch": 1.44, "grad_norm": 15.614473932249652, "learning_rate": 1.1180251424403666e-05, "loss": 0.5703, "step": 9190 }, { "epoch": 1.44, "grad_norm": 17.54772774245434, "learning_rate": 1.1178576795557214e-05, "loss": 0.5395, "step": 9191 }, { "epoch": 1.44, "grad_norm": 19.772009570027226, "learning_rate": 1.117690213319273e-05, "loss": 0.7171, "step": 9192 }, { "epoch": 1.44, "grad_norm": 37.63290593937698, "learning_rate": 1.1175227437357835e-05, "loss": 0.6536, "step": 9193 }, { "epoch": 1.44, "grad_norm": 17.0844609992272, "learning_rate": 1.1173552708100162e-05, "loss": 0.5932, "step": 9194 }, { "epoch": 1.44, "grad_norm": 22.895576886320796, "learning_rate": 1.1171877945467333e-05, "loss": 0.6676, "step": 9195 }, { "epoch": 1.44, "grad_norm": 22.679667293790565, "learning_rate": 1.1170203149506984e-05, "loss": 0.5905, "step": 9196 }, { "epoch": 1.44, "grad_norm": 21.82699749983028, "learning_rate": 1.1168528320266743e-05, "loss": 0.5795, "step": 9197 }, { "epoch": 1.44, "grad_norm": 26.148849002528173, "learning_rate": 1.1166853457794238e-05, "loss": 0.5694, "step": 9198 }, { "epoch": 1.44, "grad_norm": 33.59946377445537, "learning_rate": 1.1165178562137103e-05, "loss": 0.6562, "step": 9199 }, { "epoch": 1.44, "grad_norm": 22.98983943847259, "learning_rate": 1.1163503633342971e-05, "loss": 0.7734, "step": 9200 }, { "epoch": 1.44, "grad_norm": 19.280237031384548, "learning_rate": 1.1161828671459475e-05, "loss": 0.615, "step": 9201 }, { "epoch": 1.44, "grad_norm": 26.66468511156852, "learning_rate": 1.1160153676534256e-05, "loss": 0.5803, "step": 9202 }, { "epoch": 1.44, "grad_norm": 21.658845253320656, "learning_rate": 1.115847864861494e-05, "loss": 0.6408, "step": 9203 }, { "epoch": 1.44, "grad_norm": 26.169203517702094, "learning_rate": 1.1156803587749172e-05, "loss": 0.5945, "step": 9204 }, { "epoch": 1.44, "grad_norm": 21.953103423871607, "learning_rate": 1.1155128493984588e-05, "loss": 0.6681, "step": 9205 }, { "epoch": 1.44, "grad_norm": 21.969191934066426, "learning_rate": 1.1153453367368824e-05, "loss": 0.6249, "step": 9206 }, { "epoch": 1.44, "grad_norm": 22.91817980087681, "learning_rate": 1.115177820794952e-05, "loss": 0.5894, "step": 9207 }, { "epoch": 1.44, "grad_norm": 21.48646017948929, "learning_rate": 1.1150103015774318e-05, "loss": 0.6155, "step": 9208 }, { "epoch": 1.44, "grad_norm": 23.897741915870306, "learning_rate": 1.1148427790890857e-05, "loss": 0.6374, "step": 9209 }, { "epoch": 1.44, "grad_norm": 21.232398398685255, "learning_rate": 1.1146752533346786e-05, "loss": 0.6724, "step": 9210 }, { "epoch": 1.44, "grad_norm": 19.359026797213556, "learning_rate": 1.1145077243189739e-05, "loss": 0.6039, "step": 9211 }, { "epoch": 1.44, "grad_norm": 13.296470177581037, "learning_rate": 1.114340192046737e-05, "loss": 0.575, "step": 9212 }, { "epoch": 1.44, "grad_norm": 32.605814467960386, "learning_rate": 1.1141726565227313e-05, "loss": 0.7205, "step": 9213 }, { "epoch": 1.44, "grad_norm": 32.400132068023424, "learning_rate": 1.1140051177517224e-05, "loss": 0.6781, "step": 9214 }, { "epoch": 1.44, "grad_norm": 11.485535025884852, "learning_rate": 1.1138375757384747e-05, "loss": 0.5879, "step": 9215 }, { "epoch": 1.44, "grad_norm": 18.18111534724461, "learning_rate": 1.113670030487753e-05, "loss": 0.6068, "step": 9216 }, { "epoch": 1.44, "grad_norm": 20.93906984991635, "learning_rate": 1.1135024820043218e-05, "loss": 0.5994, "step": 9217 }, { "epoch": 1.44, "grad_norm": 17.576505744951646, "learning_rate": 1.1133349302929468e-05, "loss": 0.6009, "step": 9218 }, { "epoch": 1.44, "grad_norm": 23.1509549950664, "learning_rate": 1.1131673753583922e-05, "loss": 0.5916, "step": 9219 }, { "epoch": 1.44, "grad_norm": 14.72266107176036, "learning_rate": 1.112999817205424e-05, "loss": 0.6482, "step": 9220 }, { "epoch": 1.44, "grad_norm": 11.915140540956552, "learning_rate": 1.1128322558388071e-05, "loss": 0.6096, "step": 9221 }, { "epoch": 1.44, "grad_norm": 18.446700158244614, "learning_rate": 1.1126646912633068e-05, "loss": 0.5952, "step": 9222 }, { "epoch": 1.44, "grad_norm": 18.73237618249963, "learning_rate": 1.1124971234836886e-05, "loss": 0.589, "step": 9223 }, { "epoch": 1.44, "grad_norm": 19.254016859650473, "learning_rate": 1.1123295525047178e-05, "loss": 0.6781, "step": 9224 }, { "epoch": 1.44, "grad_norm": 13.664382352297014, "learning_rate": 1.1121619783311606e-05, "loss": 0.5769, "step": 9225 }, { "epoch": 1.44, "grad_norm": 25.017923541978643, "learning_rate": 1.111994400967782e-05, "loss": 0.6411, "step": 9226 }, { "epoch": 1.44, "grad_norm": 16.551425547660248, "learning_rate": 1.1118268204193485e-05, "loss": 0.6588, "step": 9227 }, { "epoch": 1.44, "grad_norm": 13.759691531412997, "learning_rate": 1.1116592366906254e-05, "loss": 0.5924, "step": 9228 }, { "epoch": 1.44, "grad_norm": 14.396146381919781, "learning_rate": 1.111491649786379e-05, "loss": 0.5966, "step": 9229 }, { "epoch": 1.44, "grad_norm": 18.314983633208765, "learning_rate": 1.1113240597113752e-05, "loss": 0.6199, "step": 9230 }, { "epoch": 1.44, "grad_norm": 32.73033330998373, "learning_rate": 1.1111564664703805e-05, "loss": 0.7141, "step": 9231 }, { "epoch": 1.44, "grad_norm": 11.42948871421432, "learning_rate": 1.1109888700681608e-05, "loss": 0.5544, "step": 9232 }, { "epoch": 1.44, "grad_norm": 21.724222380238306, "learning_rate": 1.1108212705094826e-05, "loss": 0.6953, "step": 9233 }, { "epoch": 1.44, "grad_norm": 26.683155058442054, "learning_rate": 1.1106536677991122e-05, "loss": 0.6156, "step": 9234 }, { "epoch": 1.44, "grad_norm": 18.15696689943428, "learning_rate": 1.1104860619418162e-05, "loss": 0.5368, "step": 9235 }, { "epoch": 1.44, "grad_norm": 18.71613632600167, "learning_rate": 1.1103184529423612e-05, "loss": 0.6409, "step": 9236 }, { "epoch": 1.44, "grad_norm": 19.456197802359966, "learning_rate": 1.1101508408055141e-05, "loss": 0.6639, "step": 9237 }, { "epoch": 1.44, "grad_norm": 24.49740585848573, "learning_rate": 1.1099832255360412e-05, "loss": 0.5518, "step": 9238 }, { "epoch": 1.44, "grad_norm": 19.69709484610711, "learning_rate": 1.1098156071387101e-05, "loss": 0.6742, "step": 9239 }, { "epoch": 1.44, "grad_norm": 11.395890251304616, "learning_rate": 1.1096479856182872e-05, "loss": 0.5318, "step": 9240 }, { "epoch": 1.44, "grad_norm": 12.377089384249047, "learning_rate": 1.1094803609795398e-05, "loss": 0.5615, "step": 9241 }, { "epoch": 1.44, "grad_norm": 20.295350144498446, "learning_rate": 1.109312733227235e-05, "loss": 0.6399, "step": 9242 }, { "epoch": 1.44, "grad_norm": 19.574306593655788, "learning_rate": 1.10914510236614e-05, "loss": 0.6261, "step": 9243 }, { "epoch": 1.44, "grad_norm": 26.288961762324547, "learning_rate": 1.1089774684010219e-05, "loss": 0.6474, "step": 9244 }, { "epoch": 1.44, "grad_norm": 21.94701582591727, "learning_rate": 1.1088098313366484e-05, "loss": 0.5891, "step": 9245 }, { "epoch": 1.44, "grad_norm": 19.822434631643524, "learning_rate": 1.108642191177787e-05, "loss": 0.573, "step": 9246 }, { "epoch": 1.44, "grad_norm": 15.142014519928907, "learning_rate": 1.1084745479292054e-05, "loss": 0.5446, "step": 9247 }, { "epoch": 1.44, "grad_norm": 18.005720786212926, "learning_rate": 1.108306901595671e-05, "loss": 0.6057, "step": 9248 }, { "epoch": 1.44, "grad_norm": 21.07260389531164, "learning_rate": 1.1081392521819518e-05, "loss": 0.6217, "step": 9249 }, { "epoch": 1.44, "grad_norm": 17.374317807588838, "learning_rate": 1.1079715996928156e-05, "loss": 0.672, "step": 9250 }, { "epoch": 1.45, "grad_norm": 14.466625453375757, "learning_rate": 1.1078039441330298e-05, "loss": 0.6079, "step": 9251 }, { "epoch": 1.45, "grad_norm": 19.851873231007477, "learning_rate": 1.1076362855073635e-05, "loss": 0.598, "step": 9252 }, { "epoch": 1.45, "grad_norm": 28.064885948167905, "learning_rate": 1.1074686238205837e-05, "loss": 0.6714, "step": 9253 }, { "epoch": 1.45, "grad_norm": 14.867303519064384, "learning_rate": 1.1073009590774595e-05, "loss": 0.5728, "step": 9254 }, { "epoch": 1.45, "grad_norm": 15.335319741740042, "learning_rate": 1.1071332912827584e-05, "loss": 0.6313, "step": 9255 }, { "epoch": 1.45, "grad_norm": 13.180464268866153, "learning_rate": 1.1069656204412496e-05, "loss": 0.5479, "step": 9256 }, { "epoch": 1.45, "grad_norm": 15.80640386217679, "learning_rate": 1.1067979465577005e-05, "loss": 0.617, "step": 9257 }, { "epoch": 1.45, "grad_norm": 36.29597278537783, "learning_rate": 1.106630269636881e-05, "loss": 0.5618, "step": 9258 }, { "epoch": 1.45, "grad_norm": 22.1431439002648, "learning_rate": 1.1064625896835585e-05, "loss": 0.5572, "step": 9259 }, { "epoch": 1.45, "grad_norm": 16.42095537813848, "learning_rate": 1.1062949067025024e-05, "loss": 0.6221, "step": 9260 }, { "epoch": 1.45, "grad_norm": 20.78267029483533, "learning_rate": 1.1061272206984815e-05, "loss": 0.6412, "step": 9261 }, { "epoch": 1.45, "grad_norm": 18.15203782380583, "learning_rate": 1.1059595316762644e-05, "loss": 0.5633, "step": 9262 }, { "epoch": 1.45, "grad_norm": 18.537579408795157, "learning_rate": 1.1057918396406197e-05, "loss": 0.5938, "step": 9263 }, { "epoch": 1.45, "grad_norm": 16.498481395676407, "learning_rate": 1.1056241445963177e-05, "loss": 0.6043, "step": 9264 }, { "epoch": 1.45, "grad_norm": 17.583496270867375, "learning_rate": 1.1054564465481263e-05, "loss": 0.6441, "step": 9265 }, { "epoch": 1.45, "grad_norm": 16.38927262129966, "learning_rate": 1.1052887455008157e-05, "loss": 0.5766, "step": 9266 }, { "epoch": 1.45, "grad_norm": 15.729221358477853, "learning_rate": 1.1051210414591544e-05, "loss": 0.5274, "step": 9267 }, { "epoch": 1.45, "grad_norm": 18.39367556733507, "learning_rate": 1.1049533344279122e-05, "loss": 0.599, "step": 9268 }, { "epoch": 1.45, "grad_norm": 25.45838002144944, "learning_rate": 1.1047856244118591e-05, "loss": 0.6777, "step": 9269 }, { "epoch": 1.45, "grad_norm": 16.46941887856235, "learning_rate": 1.1046179114157636e-05, "loss": 0.6293, "step": 9270 }, { "epoch": 1.45, "grad_norm": 20.313819413982287, "learning_rate": 1.1044501954443962e-05, "loss": 0.5981, "step": 9271 }, { "epoch": 1.45, "grad_norm": 9.764063283786262, "learning_rate": 1.104282476502526e-05, "loss": 0.4817, "step": 9272 }, { "epoch": 1.45, "grad_norm": 17.95439232245066, "learning_rate": 1.1041147545949233e-05, "loss": 0.5305, "step": 9273 }, { "epoch": 1.45, "grad_norm": 17.06889737895641, "learning_rate": 1.103947029726358e-05, "loss": 0.6047, "step": 9274 }, { "epoch": 1.45, "grad_norm": 16.000177674173145, "learning_rate": 1.1037793019015999e-05, "loss": 0.5771, "step": 9275 }, { "epoch": 1.45, "grad_norm": 16.596997755010054, "learning_rate": 1.1036115711254191e-05, "loss": 0.5705, "step": 9276 }, { "epoch": 1.45, "grad_norm": 17.040014524156327, "learning_rate": 1.103443837402586e-05, "loss": 0.5791, "step": 9277 }, { "epoch": 1.45, "grad_norm": 28.342455143566227, "learning_rate": 1.1032761007378708e-05, "loss": 0.6149, "step": 9278 }, { "epoch": 1.45, "grad_norm": 10.76329390046126, "learning_rate": 1.1031083611360433e-05, "loss": 0.5815, "step": 9279 }, { "epoch": 1.45, "grad_norm": 21.4693686369783, "learning_rate": 1.1029406186018745e-05, "loss": 0.6036, "step": 9280 }, { "epoch": 1.45, "grad_norm": 22.6934949150342, "learning_rate": 1.102772873140135e-05, "loss": 0.6247, "step": 9281 }, { "epoch": 1.45, "grad_norm": 23.703464840066538, "learning_rate": 1.1026051247555947e-05, "loss": 0.5954, "step": 9282 }, { "epoch": 1.45, "grad_norm": 18.936623447971115, "learning_rate": 1.1024373734530247e-05, "loss": 0.6855, "step": 9283 }, { "epoch": 1.45, "grad_norm": 20.189429309421666, "learning_rate": 1.102269619237196e-05, "loss": 0.6611, "step": 9284 }, { "epoch": 1.45, "grad_norm": 14.165220812394546, "learning_rate": 1.102101862112879e-05, "loss": 0.6039, "step": 9285 }, { "epoch": 1.45, "grad_norm": 22.82262879304276, "learning_rate": 1.1019341020848449e-05, "loss": 0.6157, "step": 9286 }, { "epoch": 1.45, "grad_norm": 20.60149013056682, "learning_rate": 1.1017663391578644e-05, "loss": 0.6286, "step": 9287 }, { "epoch": 1.45, "grad_norm": 13.476861249404273, "learning_rate": 1.1015985733367087e-05, "loss": 0.6079, "step": 9288 }, { "epoch": 1.45, "grad_norm": 31.370785933627776, "learning_rate": 1.1014308046261491e-05, "loss": 0.6477, "step": 9289 }, { "epoch": 1.45, "grad_norm": 12.87850610610669, "learning_rate": 1.1012630330309567e-05, "loss": 0.5596, "step": 9290 }, { "epoch": 1.45, "grad_norm": 20.24132793096575, "learning_rate": 1.1010952585559029e-05, "loss": 0.6938, "step": 9291 }, { "epoch": 1.45, "grad_norm": 14.959268132691284, "learning_rate": 1.1009274812057592e-05, "loss": 0.5619, "step": 9292 }, { "epoch": 1.45, "grad_norm": 21.89586029783041, "learning_rate": 1.1007597009852968e-05, "loss": 0.6731, "step": 9293 }, { "epoch": 1.45, "grad_norm": 21.91133291751002, "learning_rate": 1.1005919178992876e-05, "loss": 0.524, "step": 9294 }, { "epoch": 1.45, "grad_norm": 20.36805678174434, "learning_rate": 1.1004241319525028e-05, "loss": 0.6244, "step": 9295 }, { "epoch": 1.45, "grad_norm": 18.889540809995655, "learning_rate": 1.1002563431497151e-05, "loss": 0.6113, "step": 9296 }, { "epoch": 1.45, "grad_norm": 15.685101366562206, "learning_rate": 1.1000885514956949e-05, "loss": 0.618, "step": 9297 }, { "epoch": 1.45, "grad_norm": 17.93623806920666, "learning_rate": 1.0999207569952155e-05, "loss": 0.7182, "step": 9298 }, { "epoch": 1.45, "grad_norm": 20.81101575060987, "learning_rate": 1.0997529596530477e-05, "loss": 0.6416, "step": 9299 }, { "epoch": 1.45, "grad_norm": 19.326877634694508, "learning_rate": 1.0995851594739642e-05, "loss": 0.6346, "step": 9300 }, { "epoch": 1.45, "grad_norm": 6.503003877467766, "learning_rate": 1.099417356462737e-05, "loss": 0.603, "step": 9301 }, { "epoch": 1.45, "grad_norm": 18.843762097267835, "learning_rate": 1.0992495506241387e-05, "loss": 0.645, "step": 9302 }, { "epoch": 1.45, "grad_norm": 18.308978150167626, "learning_rate": 1.0990817419629409e-05, "loss": 0.6139, "step": 9303 }, { "epoch": 1.45, "grad_norm": 18.311669976135995, "learning_rate": 1.0989139304839165e-05, "loss": 0.6507, "step": 9304 }, { "epoch": 1.45, "grad_norm": 24.02842702456249, "learning_rate": 1.0987461161918375e-05, "loss": 0.6218, "step": 9305 }, { "epoch": 1.45, "grad_norm": 20.82349988290152, "learning_rate": 1.0985782990914771e-05, "loss": 0.6529, "step": 9306 }, { "epoch": 1.45, "grad_norm": 20.726658065638848, "learning_rate": 1.098410479187607e-05, "loss": 0.6127, "step": 9307 }, { "epoch": 1.45, "grad_norm": 18.259433790246426, "learning_rate": 1.0982426564850009e-05, "loss": 0.5724, "step": 9308 }, { "epoch": 1.45, "grad_norm": 19.976708911408448, "learning_rate": 1.098074830988431e-05, "loss": 0.5547, "step": 9309 }, { "epoch": 1.45, "grad_norm": 16.170615903933054, "learning_rate": 1.0979070027026703e-05, "loss": 0.5575, "step": 9310 }, { "epoch": 1.45, "grad_norm": 18.485875205866122, "learning_rate": 1.0977391716324916e-05, "loss": 0.5526, "step": 9311 }, { "epoch": 1.45, "grad_norm": 16.027373340606335, "learning_rate": 1.097571337782668e-05, "loss": 0.5961, "step": 9312 }, { "epoch": 1.45, "grad_norm": 36.743595569193275, "learning_rate": 1.0974035011579731e-05, "loss": 0.6316, "step": 9313 }, { "epoch": 1.45, "grad_norm": 13.71870219454767, "learning_rate": 1.097235661763179e-05, "loss": 0.6095, "step": 9314 }, { "epoch": 1.46, "grad_norm": 20.657334626731686, "learning_rate": 1.09706781960306e-05, "loss": 0.584, "step": 9315 }, { "epoch": 1.46, "grad_norm": 18.001788818695566, "learning_rate": 1.0968999746823888e-05, "loss": 0.5891, "step": 9316 }, { "epoch": 1.46, "grad_norm": 18.441643764078986, "learning_rate": 1.0967321270059391e-05, "loss": 0.6439, "step": 9317 }, { "epoch": 1.46, "grad_norm": 16.053219437131116, "learning_rate": 1.0965642765784841e-05, "loss": 0.6044, "step": 9318 }, { "epoch": 1.46, "grad_norm": 22.90568478291575, "learning_rate": 1.0963964234047977e-05, "loss": 0.6336, "step": 9319 }, { "epoch": 1.46, "grad_norm": 79.01080314115897, "learning_rate": 1.0962285674896534e-05, "loss": 0.7097, "step": 9320 }, { "epoch": 1.46, "grad_norm": 21.373823724694624, "learning_rate": 1.096060708837825e-05, "loss": 0.6142, "step": 9321 }, { "epoch": 1.46, "grad_norm": 17.400886317917156, "learning_rate": 1.0958928474540857e-05, "loss": 0.6652, "step": 9322 }, { "epoch": 1.46, "grad_norm": 21.61111164975765, "learning_rate": 1.0957249833432108e-05, "loss": 0.6009, "step": 9323 }, { "epoch": 1.46, "grad_norm": 29.695094435206546, "learning_rate": 1.0955571165099728e-05, "loss": 0.6241, "step": 9324 }, { "epoch": 1.46, "grad_norm": 13.494431626007302, "learning_rate": 1.0953892469591464e-05, "loss": 0.5911, "step": 9325 }, { "epoch": 1.46, "grad_norm": 27.481600748807796, "learning_rate": 1.0952213746955056e-05, "loss": 0.6171, "step": 9326 }, { "epoch": 1.46, "grad_norm": 25.74845349832007, "learning_rate": 1.0950534997238246e-05, "loss": 0.5717, "step": 9327 }, { "epoch": 1.46, "grad_norm": 15.79694738805393, "learning_rate": 1.0948856220488779e-05, "loss": 0.652, "step": 9328 }, { "epoch": 1.46, "grad_norm": 20.7749297189614, "learning_rate": 1.0947177416754396e-05, "loss": 0.6805, "step": 9329 }, { "epoch": 1.46, "grad_norm": 15.015635277052905, "learning_rate": 1.0945498586082838e-05, "loss": 0.5753, "step": 9330 }, { "epoch": 1.46, "grad_norm": 20.101593173226394, "learning_rate": 1.0943819728521854e-05, "loss": 0.5702, "step": 9331 }, { "epoch": 1.46, "grad_norm": 18.889657821696332, "learning_rate": 1.0942140844119194e-05, "loss": 0.6042, "step": 9332 }, { "epoch": 1.46, "grad_norm": 22.450217602308165, "learning_rate": 1.0940461932922593e-05, "loss": 0.5951, "step": 9333 }, { "epoch": 1.46, "grad_norm": 16.44554068663033, "learning_rate": 1.0938782994979808e-05, "loss": 0.6493, "step": 9334 }, { "epoch": 1.46, "grad_norm": 18.847142159759, "learning_rate": 1.0937104030338584e-05, "loss": 0.5713, "step": 9335 }, { "epoch": 1.46, "grad_norm": 16.85219721732854, "learning_rate": 1.0935425039046668e-05, "loss": 0.5854, "step": 9336 }, { "epoch": 1.46, "grad_norm": 16.14891029592234, "learning_rate": 1.0933746021151813e-05, "loss": 0.5597, "step": 9337 }, { "epoch": 1.46, "grad_norm": 21.617218879894757, "learning_rate": 1.0932066976701767e-05, "loss": 0.5958, "step": 9338 }, { "epoch": 1.46, "grad_norm": 17.19051853896591, "learning_rate": 1.0930387905744278e-05, "loss": 0.6437, "step": 9339 }, { "epoch": 1.46, "grad_norm": 27.79981532287448, "learning_rate": 1.0928708808327107e-05, "loss": 0.7169, "step": 9340 }, { "epoch": 1.46, "grad_norm": 11.750440615270596, "learning_rate": 1.0927029684497995e-05, "loss": 0.5916, "step": 9341 }, { "epoch": 1.46, "grad_norm": 17.773517717452904, "learning_rate": 1.0925350534304706e-05, "loss": 0.596, "step": 9342 }, { "epoch": 1.46, "grad_norm": 16.23432686737092, "learning_rate": 1.0923671357794986e-05, "loss": 0.6124, "step": 9343 }, { "epoch": 1.46, "grad_norm": 17.46536881756696, "learning_rate": 1.0921992155016594e-05, "loss": 0.6399, "step": 9344 }, { "epoch": 1.46, "grad_norm": 13.835094566774238, "learning_rate": 1.0920312926017285e-05, "loss": 0.5395, "step": 9345 }, { "epoch": 1.46, "grad_norm": 18.85248532785278, "learning_rate": 1.0918633670844813e-05, "loss": 0.6304, "step": 9346 }, { "epoch": 1.46, "grad_norm": 19.692073893187192, "learning_rate": 1.0916954389546939e-05, "loss": 0.5905, "step": 9347 }, { "epoch": 1.46, "grad_norm": 13.087433340637027, "learning_rate": 1.0915275082171419e-05, "loss": 0.5698, "step": 9348 }, { "epoch": 1.46, "grad_norm": 12.552102892874942, "learning_rate": 1.0913595748766009e-05, "loss": 0.5512, "step": 9349 }, { "epoch": 1.46, "grad_norm": 24.961806478880785, "learning_rate": 1.0911916389378473e-05, "loss": 0.5684, "step": 9350 }, { "epoch": 1.46, "grad_norm": 25.915952094357614, "learning_rate": 1.0910237004056563e-05, "loss": 0.6457, "step": 9351 }, { "epoch": 1.46, "grad_norm": 17.217591032081387, "learning_rate": 1.0908557592848048e-05, "loss": 0.6163, "step": 9352 }, { "epoch": 1.46, "grad_norm": 23.803704479040587, "learning_rate": 1.0906878155800686e-05, "loss": 0.7503, "step": 9353 }, { "epoch": 1.46, "grad_norm": 17.070667471784837, "learning_rate": 1.0905198692962241e-05, "loss": 0.6108, "step": 9354 }, { "epoch": 1.46, "grad_norm": 14.253869194386121, "learning_rate": 1.0903519204380475e-05, "loss": 0.605, "step": 9355 }, { "epoch": 1.46, "grad_norm": 18.43850178502428, "learning_rate": 1.090183969010315e-05, "loss": 0.5053, "step": 9356 }, { "epoch": 1.46, "grad_norm": 21.94831951734522, "learning_rate": 1.0900160150178034e-05, "loss": 0.6282, "step": 9357 }, { "epoch": 1.46, "grad_norm": 20.79610114732136, "learning_rate": 1.0898480584652887e-05, "loss": 0.6296, "step": 9358 }, { "epoch": 1.46, "grad_norm": 18.24718245874271, "learning_rate": 1.0896800993575482e-05, "loss": 0.6255, "step": 9359 }, { "epoch": 1.46, "grad_norm": 15.759223427063468, "learning_rate": 1.0895121376993576e-05, "loss": 0.5951, "step": 9360 }, { "epoch": 1.46, "grad_norm": 27.871080497733203, "learning_rate": 1.0893441734954944e-05, "loss": 0.5584, "step": 9361 }, { "epoch": 1.46, "grad_norm": 20.92422331578606, "learning_rate": 1.0891762067507353e-05, "loss": 0.6447, "step": 9362 }, { "epoch": 1.46, "grad_norm": 21.245075625056977, "learning_rate": 1.089008237469857e-05, "loss": 0.5667, "step": 9363 }, { "epoch": 1.46, "grad_norm": 18.180661107784466, "learning_rate": 1.0888402656576364e-05, "loss": 0.5551, "step": 9364 }, { "epoch": 1.46, "grad_norm": 15.099482853924005, "learning_rate": 1.0886722913188508e-05, "loss": 0.6097, "step": 9365 }, { "epoch": 1.46, "grad_norm": 18.237800511740094, "learning_rate": 1.0885043144582765e-05, "loss": 0.578, "step": 9366 }, { "epoch": 1.46, "grad_norm": 24.35403595069551, "learning_rate": 1.0883363350806922e-05, "loss": 0.5912, "step": 9367 }, { "epoch": 1.46, "grad_norm": 13.135482664141634, "learning_rate": 1.0881683531908733e-05, "loss": 0.6131, "step": 9368 }, { "epoch": 1.46, "grad_norm": 21.470869481661918, "learning_rate": 1.0880003687935988e-05, "loss": 0.6347, "step": 9369 }, { "epoch": 1.46, "grad_norm": 16.97363563802261, "learning_rate": 1.0878323818936448e-05, "loss": 0.5712, "step": 9370 }, { "epoch": 1.46, "grad_norm": 20.5706278316438, "learning_rate": 1.0876643924957892e-05, "loss": 0.6928, "step": 9371 }, { "epoch": 1.46, "grad_norm": 15.891631202386465, "learning_rate": 1.0874964006048098e-05, "loss": 0.6189, "step": 9372 }, { "epoch": 1.46, "grad_norm": 17.001800649486196, "learning_rate": 1.087328406225484e-05, "loss": 0.5915, "step": 9373 }, { "epoch": 1.46, "grad_norm": 24.739046890599845, "learning_rate": 1.0871604093625894e-05, "loss": 0.5962, "step": 9374 }, { "epoch": 1.46, "grad_norm": 13.372756715849441, "learning_rate": 1.0869924100209032e-05, "loss": 0.5949, "step": 9375 }, { "epoch": 1.46, "grad_norm": 15.676144707632373, "learning_rate": 1.0868244082052047e-05, "loss": 0.6773, "step": 9376 }, { "epoch": 1.46, "grad_norm": 17.00720326423703, "learning_rate": 1.0866564039202701e-05, "loss": 0.7093, "step": 9377 }, { "epoch": 1.46, "grad_norm": 24.900872300899252, "learning_rate": 1.0864883971708782e-05, "loss": 0.5787, "step": 9378 }, { "epoch": 1.47, "grad_norm": 16.036677754152258, "learning_rate": 1.0863203879618069e-05, "loss": 0.5634, "step": 9379 }, { "epoch": 1.47, "grad_norm": 18.653135864055255, "learning_rate": 1.0861523762978343e-05, "loss": 0.6148, "step": 9380 }, { "epoch": 1.47, "grad_norm": 17.19453208812757, "learning_rate": 1.0859843621837386e-05, "loss": 0.6133, "step": 9381 }, { "epoch": 1.47, "grad_norm": 15.224947890261403, "learning_rate": 1.085816345624298e-05, "loss": 0.5717, "step": 9382 }, { "epoch": 1.47, "grad_norm": 15.338802059912311, "learning_rate": 1.0856483266242903e-05, "loss": 0.4999, "step": 9383 }, { "epoch": 1.47, "grad_norm": 15.246108394990104, "learning_rate": 1.085480305188495e-05, "loss": 0.5871, "step": 9384 }, { "epoch": 1.47, "grad_norm": 15.480610727703029, "learning_rate": 1.0853122813216891e-05, "loss": 0.5697, "step": 9385 }, { "epoch": 1.47, "grad_norm": 17.552867714016873, "learning_rate": 1.0851442550286524e-05, "loss": 0.5627, "step": 9386 }, { "epoch": 1.47, "grad_norm": 18.694715413826017, "learning_rate": 1.0849762263141625e-05, "loss": 0.6263, "step": 9387 }, { "epoch": 1.47, "grad_norm": 17.15524028019837, "learning_rate": 1.0848081951829985e-05, "loss": 0.5954, "step": 9388 }, { "epoch": 1.47, "grad_norm": 19.591500600347356, "learning_rate": 1.0846401616399392e-05, "loss": 0.6272, "step": 9389 }, { "epoch": 1.47, "grad_norm": 24.016059650664904, "learning_rate": 1.0844721256897633e-05, "loss": 0.6562, "step": 9390 }, { "epoch": 1.47, "grad_norm": 19.775444815102258, "learning_rate": 1.0843040873372493e-05, "loss": 0.6982, "step": 9391 }, { "epoch": 1.47, "grad_norm": 16.60872089662641, "learning_rate": 1.0841360465871765e-05, "loss": 0.5756, "step": 9392 }, { "epoch": 1.47, "grad_norm": 20.712522712256504, "learning_rate": 1.0839680034443234e-05, "loss": 0.5979, "step": 9393 }, { "epoch": 1.47, "grad_norm": 21.86911710516593, "learning_rate": 1.0837999579134699e-05, "loss": 0.6265, "step": 9394 }, { "epoch": 1.47, "grad_norm": 18.528542272546858, "learning_rate": 1.0836319099993944e-05, "loss": 0.6186, "step": 9395 }, { "epoch": 1.47, "grad_norm": 20.391496309029126, "learning_rate": 1.0834638597068763e-05, "loss": 0.5755, "step": 9396 }, { "epoch": 1.47, "grad_norm": 11.280628048508916, "learning_rate": 1.0832958070406949e-05, "loss": 0.548, "step": 9397 }, { "epoch": 1.47, "grad_norm": 35.738299791033334, "learning_rate": 1.0831277520056296e-05, "loss": 0.6854, "step": 9398 }, { "epoch": 1.47, "grad_norm": 16.076698314995113, "learning_rate": 1.0829596946064595e-05, "loss": 0.6017, "step": 9399 }, { "epoch": 1.47, "grad_norm": 15.164490710427323, "learning_rate": 1.0827916348479641e-05, "loss": 0.627, "step": 9400 }, { "epoch": 1.47, "grad_norm": 14.47180817728054, "learning_rate": 1.0826235727349235e-05, "loss": 0.6082, "step": 9401 }, { "epoch": 1.47, "grad_norm": 17.532146106264022, "learning_rate": 1.0824555082721161e-05, "loss": 0.622, "step": 9402 }, { "epoch": 1.47, "grad_norm": 21.87756875146567, "learning_rate": 1.0822874414643229e-05, "loss": 0.6658, "step": 9403 }, { "epoch": 1.47, "grad_norm": 21.87691566185493, "learning_rate": 1.0821193723163228e-05, "loss": 0.6399, "step": 9404 }, { "epoch": 1.47, "grad_norm": 317.9635921230517, "learning_rate": 1.0819513008328957e-05, "loss": 0.592, "step": 9405 }, { "epoch": 1.47, "grad_norm": 18.982118411228477, "learning_rate": 1.0817832270188217e-05, "loss": 0.5933, "step": 9406 }, { "epoch": 1.47, "grad_norm": 45.05185597761114, "learning_rate": 1.0816151508788804e-05, "loss": 0.6157, "step": 9407 }, { "epoch": 1.47, "grad_norm": 19.376543944123615, "learning_rate": 1.081447072417852e-05, "loss": 0.5731, "step": 9408 }, { "epoch": 1.47, "grad_norm": 22.80013261806668, "learning_rate": 1.0812789916405167e-05, "loss": 0.6376, "step": 9409 }, { "epoch": 1.47, "grad_norm": 24.585975999077736, "learning_rate": 1.081110908551654e-05, "loss": 0.5928, "step": 9410 }, { "epoch": 1.47, "grad_norm": 13.844602445573592, "learning_rate": 1.0809428231560451e-05, "loss": 0.6208, "step": 9411 }, { "epoch": 1.47, "grad_norm": 15.132977277481936, "learning_rate": 1.0807747354584692e-05, "loss": 0.5376, "step": 9412 }, { "epoch": 1.47, "grad_norm": 14.942128926670657, "learning_rate": 1.0806066454637074e-05, "loss": 0.6012, "step": 9413 }, { "epoch": 1.47, "grad_norm": 22.963368986364024, "learning_rate": 1.0804385531765394e-05, "loss": 0.6952, "step": 9414 }, { "epoch": 1.47, "grad_norm": 17.20483350455384, "learning_rate": 1.0802704586017463e-05, "loss": 0.5912, "step": 9415 }, { "epoch": 1.47, "grad_norm": 16.641716544747602, "learning_rate": 1.0801023617441082e-05, "loss": 0.5807, "step": 9416 }, { "epoch": 1.47, "grad_norm": 16.490538248553587, "learning_rate": 1.0799342626084057e-05, "loss": 0.632, "step": 9417 }, { "epoch": 1.47, "grad_norm": 16.217425944648927, "learning_rate": 1.0797661611994196e-05, "loss": 0.5509, "step": 9418 }, { "epoch": 1.47, "grad_norm": 18.932601520218793, "learning_rate": 1.0795980575219305e-05, "loss": 0.6256, "step": 9419 }, { "epoch": 1.47, "grad_norm": 22.35474050889783, "learning_rate": 1.0794299515807191e-05, "loss": 0.6549, "step": 9420 }, { "epoch": 1.47, "grad_norm": 17.936730895849756, "learning_rate": 1.0792618433805666e-05, "loss": 0.6483, "step": 9421 }, { "epoch": 1.47, "grad_norm": 21.44849353224946, "learning_rate": 1.0790937329262537e-05, "loss": 0.6675, "step": 9422 }, { "epoch": 1.47, "grad_norm": 17.453121421799935, "learning_rate": 1.0789256202225611e-05, "loss": 0.6725, "step": 9423 }, { "epoch": 1.47, "grad_norm": 19.045690951106884, "learning_rate": 1.07875750527427e-05, "loss": 0.6397, "step": 9424 }, { "epoch": 1.47, "grad_norm": 25.129528171380745, "learning_rate": 1.0785893880861616e-05, "loss": 0.6001, "step": 9425 }, { "epoch": 1.47, "grad_norm": 18.170946033509956, "learning_rate": 1.0784212686630171e-05, "loss": 0.5264, "step": 9426 }, { "epoch": 1.47, "grad_norm": 15.591933308486732, "learning_rate": 1.0782531470096171e-05, "loss": 0.6915, "step": 9427 }, { "epoch": 1.47, "grad_norm": 21.98558754210368, "learning_rate": 1.078085023130744e-05, "loss": 0.6217, "step": 9428 }, { "epoch": 1.47, "grad_norm": 15.663204068755721, "learning_rate": 1.0779168970311782e-05, "loss": 0.5677, "step": 9429 }, { "epoch": 1.47, "grad_norm": 19.071986070294194, "learning_rate": 1.0777487687157018e-05, "loss": 0.5633, "step": 9430 }, { "epoch": 1.47, "grad_norm": 13.05711221520912, "learning_rate": 1.0775806381890951e-05, "loss": 0.6033, "step": 9431 }, { "epoch": 1.47, "grad_norm": 13.609692545754523, "learning_rate": 1.0774125054561411e-05, "loss": 0.6702, "step": 9432 }, { "epoch": 1.47, "grad_norm": 17.828614111617625, "learning_rate": 1.0772443705216206e-05, "loss": 0.6436, "step": 9433 }, { "epoch": 1.47, "grad_norm": 21.221197034597598, "learning_rate": 1.0770762333903154e-05, "loss": 0.5892, "step": 9434 }, { "epoch": 1.47, "grad_norm": 14.655164095824736, "learning_rate": 1.0769080940670072e-05, "loss": 0.6246, "step": 9435 }, { "epoch": 1.47, "grad_norm": 19.871111394935483, "learning_rate": 1.0767399525564777e-05, "loss": 0.5908, "step": 9436 }, { "epoch": 1.47, "grad_norm": 19.680634806307495, "learning_rate": 1.0765718088635086e-05, "loss": 0.6012, "step": 9437 }, { "epoch": 1.47, "grad_norm": 13.00025038638523, "learning_rate": 1.0764036629928827e-05, "loss": 0.57, "step": 9438 }, { "epoch": 1.47, "grad_norm": 11.24499639660738, "learning_rate": 1.0762355149493808e-05, "loss": 0.5749, "step": 9439 }, { "epoch": 1.47, "grad_norm": 25.77828202936811, "learning_rate": 1.0760673647377855e-05, "loss": 0.6088, "step": 9440 }, { "epoch": 1.47, "grad_norm": 23.354121597287953, "learning_rate": 1.0758992123628792e-05, "loss": 0.7406, "step": 9441 }, { "epoch": 1.47, "grad_norm": 41.03289745928189, "learning_rate": 1.0757310578294433e-05, "loss": 0.7082, "step": 9442 }, { "epoch": 1.48, "grad_norm": 19.45371777845439, "learning_rate": 1.0755629011422605e-05, "loss": 0.5759, "step": 9443 }, { "epoch": 1.48, "grad_norm": 19.571716547566247, "learning_rate": 1.0753947423061131e-05, "loss": 0.6467, "step": 9444 }, { "epoch": 1.48, "grad_norm": 23.299547865296514, "learning_rate": 1.0752265813257833e-05, "loss": 0.6574, "step": 9445 }, { "epoch": 1.48, "grad_norm": 16.86579688039223, "learning_rate": 1.0750584182060535e-05, "loss": 0.6207, "step": 9446 }, { "epoch": 1.48, "grad_norm": 19.80949575604331, "learning_rate": 1.0748902529517065e-05, "loss": 0.5757, "step": 9447 }, { "epoch": 1.48, "grad_norm": 21.01128324591708, "learning_rate": 1.074722085567524e-05, "loss": 0.5821, "step": 9448 }, { "epoch": 1.48, "grad_norm": 17.561519825384572, "learning_rate": 1.0745539160582894e-05, "loss": 0.5665, "step": 9449 }, { "epoch": 1.48, "grad_norm": 16.751996929248858, "learning_rate": 1.074385744428785e-05, "loss": 0.5297, "step": 9450 }, { "epoch": 1.48, "grad_norm": 16.69825900169725, "learning_rate": 1.0742175706837936e-05, "loss": 0.6072, "step": 9451 }, { "epoch": 1.48, "grad_norm": 19.325386852186696, "learning_rate": 1.0740493948280976e-05, "loss": 0.6216, "step": 9452 }, { "epoch": 1.48, "grad_norm": 30.758691158442254, "learning_rate": 1.0738812168664804e-05, "loss": 0.7611, "step": 9453 }, { "epoch": 1.48, "grad_norm": 16.43667945880691, "learning_rate": 1.0737130368037245e-05, "loss": 0.5615, "step": 9454 }, { "epoch": 1.48, "grad_norm": 13.917657728250465, "learning_rate": 1.0735448546446132e-05, "loss": 0.5749, "step": 9455 }, { "epoch": 1.48, "grad_norm": 19.572068139474975, "learning_rate": 1.073376670393929e-05, "loss": 0.6511, "step": 9456 }, { "epoch": 1.48, "grad_norm": 21.302822306266066, "learning_rate": 1.0732084840564556e-05, "loss": 0.6853, "step": 9457 }, { "epoch": 1.48, "grad_norm": 18.795132239927618, "learning_rate": 1.0730402956369753e-05, "loss": 0.5679, "step": 9458 }, { "epoch": 1.48, "grad_norm": 14.018679810606356, "learning_rate": 1.0728721051402718e-05, "loss": 0.568, "step": 9459 }, { "epoch": 1.48, "grad_norm": 17.406772044107694, "learning_rate": 1.0727039125711283e-05, "loss": 0.5909, "step": 9460 }, { "epoch": 1.48, "grad_norm": 25.010215222578704, "learning_rate": 1.0725357179343282e-05, "loss": 0.5392, "step": 9461 }, { "epoch": 1.48, "grad_norm": 15.532561240117714, "learning_rate": 1.0723675212346546e-05, "loss": 0.6206, "step": 9462 }, { "epoch": 1.48, "grad_norm": 13.360520181037266, "learning_rate": 1.0721993224768908e-05, "loss": 0.5557, "step": 9463 }, { "epoch": 1.48, "grad_norm": 22.81751436097232, "learning_rate": 1.0720311216658205e-05, "loss": 0.6083, "step": 9464 }, { "epoch": 1.48, "grad_norm": 17.101448116207504, "learning_rate": 1.0718629188062275e-05, "loss": 0.5691, "step": 9465 }, { "epoch": 1.48, "grad_norm": 18.293912554498625, "learning_rate": 1.0716947139028953e-05, "loss": 0.563, "step": 9466 }, { "epoch": 1.48, "grad_norm": 22.40208876114644, "learning_rate": 1.071526506960607e-05, "loss": 0.5423, "step": 9467 }, { "epoch": 1.48, "grad_norm": 14.219565201540604, "learning_rate": 1.071358297984147e-05, "loss": 0.6123, "step": 9468 }, { "epoch": 1.48, "grad_norm": 22.08994336835152, "learning_rate": 1.0711900869782983e-05, "loss": 0.6473, "step": 9469 }, { "epoch": 1.48, "grad_norm": 11.957967786928343, "learning_rate": 1.0710218739478457e-05, "loss": 0.5852, "step": 9470 }, { "epoch": 1.48, "grad_norm": 20.091591723560867, "learning_rate": 1.0708536588975721e-05, "loss": 0.5488, "step": 9471 }, { "epoch": 1.48, "grad_norm": 17.706371141067113, "learning_rate": 1.0706854418322625e-05, "loss": 0.57, "step": 9472 }, { "epoch": 1.48, "grad_norm": 10.67907114768848, "learning_rate": 1.0705172227566996e-05, "loss": 0.5924, "step": 9473 }, { "epoch": 1.48, "grad_norm": 25.80960656458376, "learning_rate": 1.0703490016756688e-05, "loss": 0.6129, "step": 9474 }, { "epoch": 1.48, "grad_norm": 27.17041024948368, "learning_rate": 1.070180778593953e-05, "loss": 0.7127, "step": 9475 }, { "epoch": 1.48, "grad_norm": 21.205340816419813, "learning_rate": 1.0700125535163371e-05, "loss": 0.65, "step": 9476 }, { "epoch": 1.48, "grad_norm": 15.982293288930224, "learning_rate": 1.0698443264476051e-05, "loss": 0.5533, "step": 9477 }, { "epoch": 1.48, "grad_norm": 18.031886832876715, "learning_rate": 1.0696760973925413e-05, "loss": 0.5697, "step": 9478 }, { "epoch": 1.48, "grad_norm": 18.236372267085525, "learning_rate": 1.0695078663559302e-05, "loss": 0.6043, "step": 9479 }, { "epoch": 1.48, "grad_norm": 18.486247462818596, "learning_rate": 1.0693396333425559e-05, "loss": 0.6826, "step": 9480 }, { "epoch": 1.48, "grad_norm": 20.5916181887827, "learning_rate": 1.0691713983572028e-05, "loss": 0.5957, "step": 9481 }, { "epoch": 1.48, "grad_norm": 19.68124914541815, "learning_rate": 1.069003161404656e-05, "loss": 0.6007, "step": 9482 }, { "epoch": 1.48, "grad_norm": 17.45992795001345, "learning_rate": 1.0688349224896991e-05, "loss": 0.6118, "step": 9483 }, { "epoch": 1.48, "grad_norm": 18.61345398813733, "learning_rate": 1.0686666816171179e-05, "loss": 0.6545, "step": 9484 }, { "epoch": 1.48, "grad_norm": 16.089503070496587, "learning_rate": 1.0684984387916961e-05, "loss": 0.5167, "step": 9485 }, { "epoch": 1.48, "grad_norm": 24.119442864267423, "learning_rate": 1.068330194018219e-05, "loss": 0.6023, "step": 9486 }, { "epoch": 1.48, "grad_norm": 18.792477187993107, "learning_rate": 1.068161947301471e-05, "loss": 0.6376, "step": 9487 }, { "epoch": 1.48, "grad_norm": 30.869152872577324, "learning_rate": 1.067993698646237e-05, "loss": 0.6115, "step": 9488 }, { "epoch": 1.48, "grad_norm": 15.931252073446462, "learning_rate": 1.0678254480573023e-05, "loss": 0.6299, "step": 9489 }, { "epoch": 1.48, "grad_norm": 23.67480474037045, "learning_rate": 1.067657195539451e-05, "loss": 0.5673, "step": 9490 }, { "epoch": 1.48, "grad_norm": 18.791128503380907, "learning_rate": 1.0674889410974689e-05, "loss": 0.6741, "step": 9491 }, { "epoch": 1.48, "grad_norm": 15.574516935055764, "learning_rate": 1.0673206847361407e-05, "loss": 0.5793, "step": 9492 }, { "epoch": 1.48, "grad_norm": 14.44956903891582, "learning_rate": 1.067152426460252e-05, "loss": 0.6147, "step": 9493 }, { "epoch": 1.48, "grad_norm": 25.631647409052487, "learning_rate": 1.0669841662745874e-05, "loss": 0.5661, "step": 9494 }, { "epoch": 1.48, "grad_norm": 18.798459480931673, "learning_rate": 1.0668159041839321e-05, "loss": 0.5407, "step": 9495 }, { "epoch": 1.48, "grad_norm": 21.97414835428551, "learning_rate": 1.066647640193072e-05, "loss": 0.6451, "step": 9496 }, { "epoch": 1.48, "grad_norm": 28.361269138596125, "learning_rate": 1.0664793743067916e-05, "loss": 0.6634, "step": 9497 }, { "epoch": 1.48, "grad_norm": 12.077724281735431, "learning_rate": 1.0663111065298766e-05, "loss": 0.5143, "step": 9498 }, { "epoch": 1.48, "grad_norm": 15.987191370713397, "learning_rate": 1.066142836867113e-05, "loss": 0.6361, "step": 9499 }, { "epoch": 1.48, "grad_norm": 14.609474364876062, "learning_rate": 1.0659745653232851e-05, "loss": 0.546, "step": 9500 }, { "epoch": 1.48, "grad_norm": 17.46218057470345, "learning_rate": 1.06580629190318e-05, "loss": 0.5314, "step": 9501 }, { "epoch": 1.48, "grad_norm": 15.539981436084707, "learning_rate": 1.0656380166115818e-05, "loss": 0.5221, "step": 9502 }, { "epoch": 1.48, "grad_norm": 19.82093256580783, "learning_rate": 1.065469739453277e-05, "loss": 0.5585, "step": 9503 }, { "epoch": 1.48, "grad_norm": 24.485869001173825, "learning_rate": 1.0653014604330511e-05, "loss": 0.6606, "step": 9504 }, { "epoch": 1.48, "grad_norm": 21.91504538604698, "learning_rate": 1.06513317955569e-05, "loss": 0.6476, "step": 9505 }, { "epoch": 1.48, "grad_norm": 23.354664022589944, "learning_rate": 1.0649648968259793e-05, "loss": 0.6103, "step": 9506 }, { "epoch": 1.49, "grad_norm": 19.770386579560626, "learning_rate": 1.0647966122487048e-05, "loss": 0.5838, "step": 9507 }, { "epoch": 1.49, "grad_norm": 20.284554726570335, "learning_rate": 1.0646283258286524e-05, "loss": 0.6077, "step": 9508 }, { "epoch": 1.49, "grad_norm": 25.092795261804845, "learning_rate": 1.0644600375706087e-05, "loss": 0.5511, "step": 9509 }, { "epoch": 1.49, "grad_norm": 17.014322802909845, "learning_rate": 1.0642917474793591e-05, "loss": 0.6354, "step": 9510 }, { "epoch": 1.49, "grad_norm": 12.469612704215882, "learning_rate": 1.0641234555596898e-05, "loss": 0.5781, "step": 9511 }, { "epoch": 1.49, "grad_norm": 17.41696005460803, "learning_rate": 1.0639551618163869e-05, "loss": 0.5833, "step": 9512 }, { "epoch": 1.49, "grad_norm": 14.243668227611723, "learning_rate": 1.0637868662542364e-05, "loss": 0.5534, "step": 9513 }, { "epoch": 1.49, "grad_norm": 26.532094638430074, "learning_rate": 1.0636185688780253e-05, "loss": 0.5653, "step": 9514 }, { "epoch": 1.49, "grad_norm": 13.319982775189816, "learning_rate": 1.0634502696925387e-05, "loss": 0.5705, "step": 9515 }, { "epoch": 1.49, "grad_norm": 22.95894349296172, "learning_rate": 1.063281968702564e-05, "loss": 0.6782, "step": 9516 }, { "epoch": 1.49, "grad_norm": 15.063393052919158, "learning_rate": 1.0631136659128867e-05, "loss": 0.5694, "step": 9517 }, { "epoch": 1.49, "grad_norm": 18.209102662146336, "learning_rate": 1.062945361328294e-05, "loss": 0.6827, "step": 9518 }, { "epoch": 1.49, "grad_norm": 18.475584376052943, "learning_rate": 1.0627770549535716e-05, "loss": 0.5483, "step": 9519 }, { "epoch": 1.49, "grad_norm": 23.914881369756916, "learning_rate": 1.0626087467935069e-05, "loss": 0.6358, "step": 9520 }, { "epoch": 1.49, "grad_norm": 13.735009675664795, "learning_rate": 1.062440436852886e-05, "loss": 0.5627, "step": 9521 }, { "epoch": 1.49, "grad_norm": 15.783430771235727, "learning_rate": 1.0622721251364955e-05, "loss": 0.6351, "step": 9522 }, { "epoch": 1.49, "grad_norm": 13.958779100398177, "learning_rate": 1.0621038116491224e-05, "loss": 0.5411, "step": 9523 }, { "epoch": 1.49, "grad_norm": 17.768034254116973, "learning_rate": 1.061935496395553e-05, "loss": 0.5827, "step": 9524 }, { "epoch": 1.49, "grad_norm": 14.388536022047735, "learning_rate": 1.0617671793805743e-05, "loss": 0.6838, "step": 9525 }, { "epoch": 1.49, "grad_norm": 17.26562426036298, "learning_rate": 1.0615988606089733e-05, "loss": 0.5505, "step": 9526 }, { "epoch": 1.49, "grad_norm": 22.565596474686416, "learning_rate": 1.0614305400855367e-05, "loss": 0.7014, "step": 9527 }, { "epoch": 1.49, "grad_norm": 20.974896961204593, "learning_rate": 1.0612622178150515e-05, "loss": 0.6463, "step": 9528 }, { "epoch": 1.49, "grad_norm": 18.238998235587502, "learning_rate": 1.0610938938023047e-05, "loss": 0.6253, "step": 9529 }, { "epoch": 1.49, "grad_norm": 14.923957649410756, "learning_rate": 1.060925568052083e-05, "loss": 0.6337, "step": 9530 }, { "epoch": 1.49, "grad_norm": 21.41670397875525, "learning_rate": 1.0607572405691741e-05, "loss": 0.6089, "step": 9531 }, { "epoch": 1.49, "grad_norm": 11.801162031710726, "learning_rate": 1.0605889113583647e-05, "loss": 0.5919, "step": 9532 }, { "epoch": 1.49, "grad_norm": 19.469134183065833, "learning_rate": 1.0604205804244424e-05, "loss": 0.5946, "step": 9533 }, { "epoch": 1.49, "grad_norm": 15.567008141113567, "learning_rate": 1.0602522477721938e-05, "loss": 0.6264, "step": 9534 }, { "epoch": 1.49, "grad_norm": 22.87978994529843, "learning_rate": 1.0600839134064065e-05, "loss": 0.7207, "step": 9535 }, { "epoch": 1.49, "grad_norm": 19.125638520654952, "learning_rate": 1.0599155773318679e-05, "loss": 0.5884, "step": 9536 }, { "epoch": 1.49, "grad_norm": 26.762880063197887, "learning_rate": 1.0597472395533654e-05, "loss": 0.6003, "step": 9537 }, { "epoch": 1.49, "grad_norm": 17.77436759191914, "learning_rate": 1.0595789000756864e-05, "loss": 0.6117, "step": 9538 }, { "epoch": 1.49, "grad_norm": 14.16403719104961, "learning_rate": 1.0594105589036182e-05, "loss": 0.4892, "step": 9539 }, { "epoch": 1.49, "grad_norm": 17.848443429109874, "learning_rate": 1.0592422160419484e-05, "loss": 0.6423, "step": 9540 }, { "epoch": 1.49, "grad_norm": 24.226938673467696, "learning_rate": 1.0590738714954652e-05, "loss": 0.6831, "step": 9541 }, { "epoch": 1.49, "grad_norm": 18.35935245390786, "learning_rate": 1.0589055252689549e-05, "loss": 0.5699, "step": 9542 }, { "epoch": 1.49, "grad_norm": 16.08264945118136, "learning_rate": 1.0587371773672064e-05, "loss": 0.6028, "step": 9543 }, { "epoch": 1.49, "grad_norm": 29.98770900765187, "learning_rate": 1.0585688277950065e-05, "loss": 0.7055, "step": 9544 }, { "epoch": 1.49, "grad_norm": 15.587539367293392, "learning_rate": 1.058400476557144e-05, "loss": 0.6007, "step": 9545 }, { "epoch": 1.49, "grad_norm": 15.978174737292955, "learning_rate": 1.0582321236584057e-05, "loss": 0.6042, "step": 9546 }, { "epoch": 1.49, "grad_norm": 27.281801388598783, "learning_rate": 1.0580637691035799e-05, "loss": 0.6295, "step": 9547 }, { "epoch": 1.49, "grad_norm": 20.556255756318215, "learning_rate": 1.0578954128974546e-05, "loss": 0.6692, "step": 9548 }, { "epoch": 1.49, "grad_norm": 16.6087337362928, "learning_rate": 1.0577270550448175e-05, "loss": 0.615, "step": 9549 }, { "epoch": 1.49, "grad_norm": 21.79501263445483, "learning_rate": 1.0575586955504568e-05, "loss": 0.6839, "step": 9550 }, { "epoch": 1.49, "grad_norm": 18.24112870045532, "learning_rate": 1.0573903344191603e-05, "loss": 0.6651, "step": 9551 }, { "epoch": 1.49, "grad_norm": 18.81415157755609, "learning_rate": 1.0572219716557163e-05, "loss": 0.6202, "step": 9552 }, { "epoch": 1.49, "grad_norm": 17.39179908597005, "learning_rate": 1.0570536072649132e-05, "loss": 0.5771, "step": 9553 }, { "epoch": 1.49, "grad_norm": 15.791777579774434, "learning_rate": 1.0568852412515388e-05, "loss": 0.5681, "step": 9554 }, { "epoch": 1.49, "grad_norm": 17.152439581947128, "learning_rate": 1.0567168736203811e-05, "loss": 0.6505, "step": 9555 }, { "epoch": 1.49, "grad_norm": 13.366736823599535, "learning_rate": 1.0565485043762289e-05, "loss": 0.6095, "step": 9556 }, { "epoch": 1.49, "grad_norm": 17.203889834130816, "learning_rate": 1.0563801335238701e-05, "loss": 0.6203, "step": 9557 }, { "epoch": 1.49, "grad_norm": 15.4492931060352, "learning_rate": 1.0562117610680938e-05, "loss": 0.584, "step": 9558 }, { "epoch": 1.49, "grad_norm": 24.79157263366647, "learning_rate": 1.0560433870136872e-05, "loss": 0.7251, "step": 9559 }, { "epoch": 1.49, "grad_norm": 13.480030784467429, "learning_rate": 1.0558750113654404e-05, "loss": 0.5692, "step": 9560 }, { "epoch": 1.49, "grad_norm": 22.889954727696995, "learning_rate": 1.0557066341281402e-05, "loss": 0.5767, "step": 9561 }, { "epoch": 1.49, "grad_norm": 21.438089618415116, "learning_rate": 1.055538255306576e-05, "loss": 0.6005, "step": 9562 }, { "epoch": 1.49, "grad_norm": 17.30729982622911, "learning_rate": 1.0553698749055365e-05, "loss": 0.5957, "step": 9563 }, { "epoch": 1.49, "grad_norm": 24.809103160319772, "learning_rate": 1.05520149292981e-05, "loss": 0.5624, "step": 9564 }, { "epoch": 1.49, "grad_norm": 15.908112386932967, "learning_rate": 1.0550331093841855e-05, "loss": 0.5825, "step": 9565 }, { "epoch": 1.49, "grad_norm": 15.515388137377007, "learning_rate": 1.0548647242734516e-05, "loss": 0.5655, "step": 9566 }, { "epoch": 1.49, "grad_norm": 17.679372729467417, "learning_rate": 1.054696337602397e-05, "loss": 0.6545, "step": 9567 }, { "epoch": 1.49, "grad_norm": 16.1263260896761, "learning_rate": 1.0545279493758103e-05, "loss": 0.56, "step": 9568 }, { "epoch": 1.49, "grad_norm": 24.053409338032864, "learning_rate": 1.0543595595984806e-05, "loss": 0.6447, "step": 9569 }, { "epoch": 1.49, "grad_norm": 22.056430020817814, "learning_rate": 1.0541911682751971e-05, "loss": 0.6836, "step": 9570 }, { "epoch": 1.5, "grad_norm": 13.916570906419091, "learning_rate": 1.0540227754107483e-05, "loss": 0.6045, "step": 9571 }, { "epoch": 1.5, "grad_norm": 16.304461112504757, "learning_rate": 1.0538543810099236e-05, "loss": 0.6043, "step": 9572 }, { "epoch": 1.5, "grad_norm": 15.328828931096027, "learning_rate": 1.0536859850775117e-05, "loss": 0.66, "step": 9573 }, { "epoch": 1.5, "grad_norm": 14.321388573598336, "learning_rate": 1.0535175876183017e-05, "loss": 0.6903, "step": 9574 }, { "epoch": 1.5, "grad_norm": 18.06154264029838, "learning_rate": 1.0533491886370828e-05, "loss": 0.594, "step": 9575 }, { "epoch": 1.5, "grad_norm": 28.40824175399614, "learning_rate": 1.0531807881386442e-05, "loss": 0.6531, "step": 9576 }, { "epoch": 1.5, "grad_norm": 18.491114377761125, "learning_rate": 1.0530123861277752e-05, "loss": 0.5708, "step": 9577 }, { "epoch": 1.5, "grad_norm": 20.745317276142934, "learning_rate": 1.0528439826092649e-05, "loss": 0.6742, "step": 9578 }, { "epoch": 1.5, "grad_norm": 26.819592832648894, "learning_rate": 1.0526755775879024e-05, "loss": 0.6408, "step": 9579 }, { "epoch": 1.5, "grad_norm": 16.210495938641355, "learning_rate": 1.0525071710684777e-05, "loss": 0.6574, "step": 9580 }, { "epoch": 1.5, "grad_norm": 14.755272966114962, "learning_rate": 1.0523387630557797e-05, "loss": 0.649, "step": 9581 }, { "epoch": 1.5, "grad_norm": 18.359098214031285, "learning_rate": 1.0521703535545977e-05, "loss": 0.5388, "step": 9582 }, { "epoch": 1.5, "grad_norm": 14.95129733191838, "learning_rate": 1.0520019425697217e-05, "loss": 0.5968, "step": 9583 }, { "epoch": 1.5, "grad_norm": 25.611221677474138, "learning_rate": 1.0518335301059404e-05, "loss": 0.6536, "step": 9584 }, { "epoch": 1.5, "grad_norm": 32.14734087634544, "learning_rate": 1.0516651161680443e-05, "loss": 0.6968, "step": 9585 }, { "epoch": 1.5, "grad_norm": 12.636278067544522, "learning_rate": 1.0514967007608221e-05, "loss": 0.5266, "step": 9586 }, { "epoch": 1.5, "grad_norm": 14.785277131063943, "learning_rate": 1.0513282838890642e-05, "loss": 0.5664, "step": 9587 }, { "epoch": 1.5, "grad_norm": 16.669899890082892, "learning_rate": 1.0511598655575594e-05, "loss": 0.537, "step": 9588 }, { "epoch": 1.5, "grad_norm": 24.3401458567208, "learning_rate": 1.0509914457710986e-05, "loss": 0.6169, "step": 9589 }, { "epoch": 1.5, "grad_norm": 12.943686247889197, "learning_rate": 1.0508230245344707e-05, "loss": 0.6526, "step": 9590 }, { "epoch": 1.5, "grad_norm": 12.400058174393402, "learning_rate": 1.0506546018524656e-05, "loss": 0.5398, "step": 9591 }, { "epoch": 1.5, "grad_norm": 19.38221197724052, "learning_rate": 1.0504861777298732e-05, "loss": 0.6262, "step": 9592 }, { "epoch": 1.5, "grad_norm": 24.8586760242244, "learning_rate": 1.0503177521714836e-05, "loss": 0.614, "step": 9593 }, { "epoch": 1.5, "grad_norm": 22.04308591355548, "learning_rate": 1.0501493251820864e-05, "loss": 0.6216, "step": 9594 }, { "epoch": 1.5, "grad_norm": 19.803606390929883, "learning_rate": 1.0499808967664717e-05, "loss": 0.5423, "step": 9595 }, { "epoch": 1.5, "grad_norm": 26.97331554537966, "learning_rate": 1.0498124669294296e-05, "loss": 0.6118, "step": 9596 }, { "epoch": 1.5, "grad_norm": 20.264583169691303, "learning_rate": 1.0496440356757502e-05, "loss": 0.557, "step": 9597 }, { "epoch": 1.5, "grad_norm": 15.91924624038395, "learning_rate": 1.0494756030102236e-05, "loss": 0.5841, "step": 9598 }, { "epoch": 1.5, "grad_norm": 24.78175202927894, "learning_rate": 1.0493071689376396e-05, "loss": 0.6376, "step": 9599 }, { "epoch": 1.5, "grad_norm": 19.143372913772797, "learning_rate": 1.0491387334627884e-05, "loss": 0.5542, "step": 9600 }, { "epoch": 1.5, "grad_norm": 18.07029006864258, "learning_rate": 1.0489702965904605e-05, "loss": 0.6338, "step": 9601 }, { "epoch": 1.5, "grad_norm": 22.01711013149036, "learning_rate": 1.0488018583254463e-05, "loss": 0.6842, "step": 9602 }, { "epoch": 1.5, "grad_norm": 21.069095836699088, "learning_rate": 1.0486334186725353e-05, "loss": 0.6613, "step": 9603 }, { "epoch": 1.5, "grad_norm": 16.043266099041606, "learning_rate": 1.0484649776365189e-05, "loss": 0.6321, "step": 9604 }, { "epoch": 1.5, "grad_norm": 23.33969613639088, "learning_rate": 1.0482965352221863e-05, "loss": 0.552, "step": 9605 }, { "epoch": 1.5, "grad_norm": 15.475998266671306, "learning_rate": 1.0481280914343289e-05, "loss": 0.5734, "step": 9606 }, { "epoch": 1.5, "grad_norm": 23.847528959992133, "learning_rate": 1.0479596462777363e-05, "loss": 0.5331, "step": 9607 }, { "epoch": 1.5, "grad_norm": 13.475077926024703, "learning_rate": 1.0477911997571998e-05, "loss": 0.6345, "step": 9608 }, { "epoch": 1.5, "grad_norm": 10.699984906394286, "learning_rate": 1.0476227518775095e-05, "loss": 0.5554, "step": 9609 }, { "epoch": 1.5, "grad_norm": 15.42390074512959, "learning_rate": 1.0474543026434559e-05, "loss": 0.6587, "step": 9610 }, { "epoch": 1.5, "grad_norm": 14.81933469467535, "learning_rate": 1.0472858520598295e-05, "loss": 0.638, "step": 9611 }, { "epoch": 1.5, "grad_norm": 16.350814940764952, "learning_rate": 1.0471174001314215e-05, "loss": 0.5785, "step": 9612 }, { "epoch": 1.5, "grad_norm": 28.812973572982727, "learning_rate": 1.0469489468630218e-05, "loss": 0.6487, "step": 9613 }, { "epoch": 1.5, "grad_norm": 15.2756985602967, "learning_rate": 1.0467804922594218e-05, "loss": 0.5419, "step": 9614 }, { "epoch": 1.5, "grad_norm": 18.91706930602366, "learning_rate": 1.0466120363254119e-05, "loss": 0.5769, "step": 9615 }, { "epoch": 1.5, "grad_norm": 20.756687172567716, "learning_rate": 1.0464435790657833e-05, "loss": 0.5769, "step": 9616 }, { "epoch": 1.5, "grad_norm": 21.618957540812264, "learning_rate": 1.0462751204853262e-05, "loss": 0.5736, "step": 9617 }, { "epoch": 1.5, "grad_norm": 21.11631268909565, "learning_rate": 1.0461066605888314e-05, "loss": 0.6871, "step": 9618 }, { "epoch": 1.5, "grad_norm": 25.81731376749823, "learning_rate": 1.0459381993810906e-05, "loss": 0.5945, "step": 9619 }, { "epoch": 1.5, "grad_norm": 18.639913323143578, "learning_rate": 1.045769736866894e-05, "loss": 0.6161, "step": 9620 }, { "epoch": 1.5, "grad_norm": 18.973731759155125, "learning_rate": 1.0456012730510332e-05, "loss": 0.6132, "step": 9621 }, { "epoch": 1.5, "grad_norm": 14.093217327408952, "learning_rate": 1.0454328079382985e-05, "loss": 0.5851, "step": 9622 }, { "epoch": 1.5, "grad_norm": 16.378059627342083, "learning_rate": 1.0452643415334816e-05, "loss": 0.5905, "step": 9623 }, { "epoch": 1.5, "grad_norm": 24.635351020585734, "learning_rate": 1.045095873841373e-05, "loss": 0.6017, "step": 9624 }, { "epoch": 1.5, "grad_norm": 13.787613591048734, "learning_rate": 1.0449274048667644e-05, "loss": 0.565, "step": 9625 }, { "epoch": 1.5, "grad_norm": 23.71521270771573, "learning_rate": 1.0447589346144467e-05, "loss": 0.629, "step": 9626 }, { "epoch": 1.5, "grad_norm": 15.837488716681479, "learning_rate": 1.044590463089211e-05, "loss": 0.6665, "step": 9627 }, { "epoch": 1.5, "grad_norm": 24.80534046064591, "learning_rate": 1.0444219902958483e-05, "loss": 0.6202, "step": 9628 }, { "epoch": 1.5, "grad_norm": 14.543291504337233, "learning_rate": 1.0442535162391508e-05, "loss": 0.5971, "step": 9629 }, { "epoch": 1.5, "grad_norm": 18.99881099247838, "learning_rate": 1.0440850409239085e-05, "loss": 0.5446, "step": 9630 }, { "epoch": 1.5, "grad_norm": 15.765826762735369, "learning_rate": 1.043916564354914e-05, "loss": 0.5486, "step": 9631 }, { "epoch": 1.5, "grad_norm": 20.81523294733946, "learning_rate": 1.0437480865369577e-05, "loss": 0.6378, "step": 9632 }, { "epoch": 1.5, "grad_norm": 13.563178170735709, "learning_rate": 1.0435796074748316e-05, "loss": 0.5161, "step": 9633 }, { "epoch": 1.5, "grad_norm": 18.48273473960778, "learning_rate": 1.0434111271733271e-05, "loss": 0.5999, "step": 9634 }, { "epoch": 1.5, "grad_norm": 19.222243011752166, "learning_rate": 1.0432426456372356e-05, "loss": 0.5401, "step": 9635 }, { "epoch": 1.51, "grad_norm": 17.073605794994442, "learning_rate": 1.0430741628713482e-05, "loss": 0.6115, "step": 9636 }, { "epoch": 1.51, "grad_norm": 17.715624759873922, "learning_rate": 1.042905678880457e-05, "loss": 0.5834, "step": 9637 }, { "epoch": 1.51, "grad_norm": 30.531316117768554, "learning_rate": 1.0427371936693531e-05, "loss": 0.6356, "step": 9638 }, { "epoch": 1.51, "grad_norm": 21.905429731539908, "learning_rate": 1.0425687072428288e-05, "loss": 0.5975, "step": 9639 }, { "epoch": 1.51, "grad_norm": 28.15837052136694, "learning_rate": 1.042400219605675e-05, "loss": 0.6498, "step": 9640 }, { "epoch": 1.51, "grad_norm": 16.623438171112767, "learning_rate": 1.0422317307626842e-05, "loss": 0.5523, "step": 9641 }, { "epoch": 1.51, "grad_norm": 23.70465619693213, "learning_rate": 1.0420632407186475e-05, "loss": 0.6668, "step": 9642 }, { "epoch": 1.51, "grad_norm": 19.58035066257976, "learning_rate": 1.0418947494783567e-05, "loss": 0.6056, "step": 9643 }, { "epoch": 1.51, "grad_norm": 13.5340884592538, "learning_rate": 1.041726257046604e-05, "loss": 0.5483, "step": 9644 }, { "epoch": 1.51, "grad_norm": 12.576004154279504, "learning_rate": 1.0415577634281806e-05, "loss": 0.544, "step": 9645 }, { "epoch": 1.51, "grad_norm": 18.879241438387634, "learning_rate": 1.0413892686278791e-05, "loss": 0.5846, "step": 9646 }, { "epoch": 1.51, "grad_norm": 19.072237066510255, "learning_rate": 1.0412207726504906e-05, "loss": 0.5579, "step": 9647 }, { "epoch": 1.51, "grad_norm": 12.506065142355302, "learning_rate": 1.041052275500808e-05, "loss": 0.596, "step": 9648 }, { "epoch": 1.51, "grad_norm": 23.049002962311587, "learning_rate": 1.0408837771836222e-05, "loss": 0.5536, "step": 9649 }, { "epoch": 1.51, "grad_norm": 14.397936068666075, "learning_rate": 1.0407152777037258e-05, "loss": 0.6283, "step": 9650 }, { "epoch": 1.51, "grad_norm": 15.235977550239177, "learning_rate": 1.0405467770659107e-05, "loss": 0.6029, "step": 9651 }, { "epoch": 1.51, "grad_norm": 23.61655253112933, "learning_rate": 1.0403782752749692e-05, "loss": 0.6601, "step": 9652 }, { "epoch": 1.51, "grad_norm": 17.83600702410067, "learning_rate": 1.040209772335693e-05, "loss": 0.5642, "step": 9653 }, { "epoch": 1.51, "grad_norm": 17.185093523963605, "learning_rate": 1.0400412682528745e-05, "loss": 0.6029, "step": 9654 }, { "epoch": 1.51, "grad_norm": 20.58923856554717, "learning_rate": 1.0398727630313055e-05, "loss": 0.5541, "step": 9655 }, { "epoch": 1.51, "grad_norm": 18.243442389314847, "learning_rate": 1.0397042566757789e-05, "loss": 0.6346, "step": 9656 }, { "epoch": 1.51, "grad_norm": 19.87343456668084, "learning_rate": 1.039535749191086e-05, "loss": 0.6218, "step": 9657 }, { "epoch": 1.51, "grad_norm": 22.669560513443166, "learning_rate": 1.0393672405820197e-05, "loss": 0.5286, "step": 9658 }, { "epoch": 1.51, "grad_norm": 24.856694818496706, "learning_rate": 1.0391987308533722e-05, "loss": 0.6374, "step": 9659 }, { "epoch": 1.51, "grad_norm": 12.917608547927562, "learning_rate": 1.0390302200099355e-05, "loss": 0.5294, "step": 9660 }, { "epoch": 1.51, "grad_norm": 22.980607044193555, "learning_rate": 1.0388617080565024e-05, "loss": 0.6468, "step": 9661 }, { "epoch": 1.51, "grad_norm": 11.413652209010364, "learning_rate": 1.0386931949978649e-05, "loss": 0.5366, "step": 9662 }, { "epoch": 1.51, "grad_norm": 23.617944231443634, "learning_rate": 1.038524680838816e-05, "loss": 0.5906, "step": 9663 }, { "epoch": 1.51, "grad_norm": 18.45518717600868, "learning_rate": 1.038356165584147e-05, "loss": 0.6459, "step": 9664 }, { "epoch": 1.51, "grad_norm": 14.670259877357442, "learning_rate": 1.0381876492386516e-05, "loss": 0.5649, "step": 9665 }, { "epoch": 1.51, "grad_norm": 15.59739134506479, "learning_rate": 1.0380191318071215e-05, "loss": 0.5716, "step": 9666 }, { "epoch": 1.51, "grad_norm": 17.128297436985324, "learning_rate": 1.0378506132943497e-05, "loss": 0.61, "step": 9667 }, { "epoch": 1.51, "grad_norm": 20.822045197951343, "learning_rate": 1.0376820937051286e-05, "loss": 0.6517, "step": 9668 }, { "epoch": 1.51, "grad_norm": 18.933589958913025, "learning_rate": 1.0375135730442507e-05, "loss": 0.5457, "step": 9669 }, { "epoch": 1.51, "grad_norm": 25.78790087367503, "learning_rate": 1.0373450513165089e-05, "loss": 0.5006, "step": 9670 }, { "epoch": 1.51, "grad_norm": 24.879292001607123, "learning_rate": 1.0371765285266957e-05, "loss": 0.7449, "step": 9671 }, { "epoch": 1.51, "grad_norm": 14.096462902150305, "learning_rate": 1.0370080046796034e-05, "loss": 0.5806, "step": 9672 }, { "epoch": 1.51, "grad_norm": 19.20883964014725, "learning_rate": 1.0368394797800256e-05, "loss": 0.5285, "step": 9673 }, { "epoch": 1.51, "grad_norm": 22.80245911197219, "learning_rate": 1.0366709538327542e-05, "loss": 0.5422, "step": 9674 }, { "epoch": 1.51, "grad_norm": 22.57351551851732, "learning_rate": 1.0365024268425826e-05, "loss": 0.6182, "step": 9675 }, { "epoch": 1.51, "grad_norm": 20.05579679874009, "learning_rate": 1.036333898814303e-05, "loss": 0.578, "step": 9676 }, { "epoch": 1.51, "grad_norm": 17.21522120266898, "learning_rate": 1.0361653697527088e-05, "loss": 0.5919, "step": 9677 }, { "epoch": 1.51, "grad_norm": 21.639236719654644, "learning_rate": 1.0359968396625925e-05, "loss": 0.5879, "step": 9678 }, { "epoch": 1.51, "grad_norm": 16.53797098980881, "learning_rate": 1.0358283085487473e-05, "loss": 0.5471, "step": 9679 }, { "epoch": 1.51, "grad_norm": 16.453302982016123, "learning_rate": 1.0356597764159659e-05, "loss": 0.6067, "step": 9680 }, { "epoch": 1.51, "grad_norm": 24.8029974288082, "learning_rate": 1.0354912432690412e-05, "loss": 0.6423, "step": 9681 }, { "epoch": 1.51, "grad_norm": 14.81667495944494, "learning_rate": 1.0353227091127662e-05, "loss": 0.5919, "step": 9682 }, { "epoch": 1.51, "grad_norm": 19.322391265133675, "learning_rate": 1.0351541739519341e-05, "loss": 0.6508, "step": 9683 }, { "epoch": 1.51, "grad_norm": 23.739329140593767, "learning_rate": 1.034985637791338e-05, "loss": 0.6226, "step": 9684 }, { "epoch": 1.51, "grad_norm": 19.4631897523884, "learning_rate": 1.034817100635771e-05, "loss": 0.6291, "step": 9685 }, { "epoch": 1.51, "grad_norm": 24.728615029887486, "learning_rate": 1.0346485624900258e-05, "loss": 0.5992, "step": 9686 }, { "epoch": 1.51, "grad_norm": 25.357985935608827, "learning_rate": 1.0344800233588959e-05, "loss": 0.6013, "step": 9687 }, { "epoch": 1.51, "grad_norm": 26.728368036087755, "learning_rate": 1.0343114832471742e-05, "loss": 0.5807, "step": 9688 }, { "epoch": 1.51, "grad_norm": 14.641954210006624, "learning_rate": 1.0341429421596538e-05, "loss": 0.5927, "step": 9689 }, { "epoch": 1.51, "grad_norm": 16.661158702157888, "learning_rate": 1.0339744001011285e-05, "loss": 0.6064, "step": 9690 }, { "epoch": 1.51, "grad_norm": 15.535338796147864, "learning_rate": 1.0338058570763907e-05, "loss": 0.593, "step": 9691 }, { "epoch": 1.51, "grad_norm": 16.891379072240596, "learning_rate": 1.0336373130902346e-05, "loss": 0.5489, "step": 9692 }, { "epoch": 1.51, "grad_norm": 20.70750703680616, "learning_rate": 1.0334687681474524e-05, "loss": 0.5354, "step": 9693 }, { "epoch": 1.51, "grad_norm": 23.374109674521367, "learning_rate": 1.0333002222528384e-05, "loss": 0.6331, "step": 9694 }, { "epoch": 1.51, "grad_norm": 13.964810037638912, "learning_rate": 1.0331316754111855e-05, "loss": 0.5579, "step": 9695 }, { "epoch": 1.51, "grad_norm": 19.67437888035102, "learning_rate": 1.032963127627287e-05, "loss": 0.5782, "step": 9696 }, { "epoch": 1.51, "grad_norm": 14.720702438182393, "learning_rate": 1.0327945789059366e-05, "loss": 0.5521, "step": 9697 }, { "epoch": 1.51, "grad_norm": 16.055180070100995, "learning_rate": 1.0326260292519274e-05, "loss": 0.6125, "step": 9698 }, { "epoch": 1.51, "grad_norm": 22.082120845605495, "learning_rate": 1.032457478670053e-05, "loss": 0.5681, "step": 9699 }, { "epoch": 1.52, "grad_norm": 27.765011397984036, "learning_rate": 1.0322889271651073e-05, "loss": 0.5868, "step": 9700 }, { "epoch": 1.52, "grad_norm": 16.549520035750025, "learning_rate": 1.0321203747418829e-05, "loss": 0.6422, "step": 9701 }, { "epoch": 1.52, "grad_norm": 24.63145328937301, "learning_rate": 1.031951821405174e-05, "loss": 0.5735, "step": 9702 }, { "epoch": 1.52, "grad_norm": 18.452257447679084, "learning_rate": 1.031783267159774e-05, "loss": 0.6196, "step": 9703 }, { "epoch": 1.52, "grad_norm": 16.704933470054325, "learning_rate": 1.0316147120104767e-05, "loss": 0.5698, "step": 9704 }, { "epoch": 1.52, "grad_norm": 22.007496829167547, "learning_rate": 1.0314461559620752e-05, "loss": 0.5526, "step": 9705 }, { "epoch": 1.52, "grad_norm": 13.626932779203724, "learning_rate": 1.0312775990193635e-05, "loss": 0.6084, "step": 9706 }, { "epoch": 1.52, "grad_norm": 18.193909153640224, "learning_rate": 1.0311090411871354e-05, "loss": 0.6134, "step": 9707 }, { "epoch": 1.52, "grad_norm": 22.992071896105898, "learning_rate": 1.030940482470184e-05, "loss": 0.653, "step": 9708 }, { "epoch": 1.52, "grad_norm": 23.32786916092986, "learning_rate": 1.0307719228733034e-05, "loss": 0.5311, "step": 9709 }, { "epoch": 1.52, "grad_norm": 20.04289358448721, "learning_rate": 1.0306033624012875e-05, "loss": 0.5914, "step": 9710 }, { "epoch": 1.52, "grad_norm": 24.094784621958546, "learning_rate": 1.0304348010589298e-05, "loss": 0.6218, "step": 9711 }, { "epoch": 1.52, "grad_norm": 26.112077356290023, "learning_rate": 1.0302662388510243e-05, "loss": 0.7058, "step": 9712 }, { "epoch": 1.52, "grad_norm": 17.5109792447267, "learning_rate": 1.0300976757823643e-05, "loss": 0.6278, "step": 9713 }, { "epoch": 1.52, "grad_norm": 21.711790830596787, "learning_rate": 1.0299291118577445e-05, "loss": 0.5624, "step": 9714 }, { "epoch": 1.52, "grad_norm": 22.12122326879831, "learning_rate": 1.029760547081958e-05, "loss": 0.6401, "step": 9715 }, { "epoch": 1.52, "grad_norm": 19.863390983195917, "learning_rate": 1.0295919814597988e-05, "loss": 0.5783, "step": 9716 }, { "epoch": 1.52, "grad_norm": 24.91616310685399, "learning_rate": 1.0294234149960614e-05, "loss": 0.6954, "step": 9717 }, { "epoch": 1.52, "grad_norm": 20.848052646670258, "learning_rate": 1.0292548476955389e-05, "loss": 0.6291, "step": 9718 }, { "epoch": 1.52, "grad_norm": 19.508515081407044, "learning_rate": 1.0290862795630261e-05, "loss": 0.6066, "step": 9719 }, { "epoch": 1.52, "grad_norm": 15.698917700203406, "learning_rate": 1.028917710603316e-05, "loss": 0.5862, "step": 9720 }, { "epoch": 1.52, "grad_norm": 19.166175449769707, "learning_rate": 1.0287491408212031e-05, "loss": 0.5832, "step": 9721 }, { "epoch": 1.52, "grad_norm": 15.907137928444762, "learning_rate": 1.0285805702214817e-05, "loss": 0.509, "step": 9722 }, { "epoch": 1.52, "grad_norm": 25.613246921192253, "learning_rate": 1.0284119988089458e-05, "loss": 0.621, "step": 9723 }, { "epoch": 1.52, "grad_norm": 24.33246406149417, "learning_rate": 1.028243426588389e-05, "loss": 0.6292, "step": 9724 }, { "epoch": 1.52, "grad_norm": 22.49944751511538, "learning_rate": 1.0280748535646058e-05, "loss": 0.6251, "step": 9725 }, { "epoch": 1.52, "grad_norm": 17.96674256182304, "learning_rate": 1.0279062797423899e-05, "loss": 0.5056, "step": 9726 }, { "epoch": 1.52, "grad_norm": 19.99551118157452, "learning_rate": 1.0277377051265361e-05, "loss": 0.5474, "step": 9727 }, { "epoch": 1.52, "grad_norm": 21.519572105791326, "learning_rate": 1.027569129721838e-05, "loss": 0.5619, "step": 9728 }, { "epoch": 1.52, "grad_norm": 19.134570249747842, "learning_rate": 1.02740055353309e-05, "loss": 0.5142, "step": 9729 }, { "epoch": 1.52, "grad_norm": 18.55793940025469, "learning_rate": 1.0272319765650862e-05, "loss": 0.551, "step": 9730 }, { "epoch": 1.52, "grad_norm": 16.757582540646, "learning_rate": 1.0270633988226209e-05, "loss": 0.5913, "step": 9731 }, { "epoch": 1.52, "grad_norm": 15.880201362449023, "learning_rate": 1.0268948203104888e-05, "loss": 0.5933, "step": 9732 }, { "epoch": 1.52, "grad_norm": 20.41890471454992, "learning_rate": 1.026726241033483e-05, "loss": 0.6509, "step": 9733 }, { "epoch": 1.52, "grad_norm": 16.561051001233153, "learning_rate": 1.0265576609963994e-05, "loss": 0.5906, "step": 9734 }, { "epoch": 1.52, "grad_norm": 25.22125263748386, "learning_rate": 1.0263890802040308e-05, "loss": 0.6282, "step": 9735 }, { "epoch": 1.52, "grad_norm": 19.599353912923885, "learning_rate": 1.0262204986611726e-05, "loss": 0.6209, "step": 9736 }, { "epoch": 1.52, "grad_norm": 17.644582756360847, "learning_rate": 1.0260519163726183e-05, "loss": 0.5601, "step": 9737 }, { "epoch": 1.52, "grad_norm": 17.62967762103688, "learning_rate": 1.025883333343163e-05, "loss": 0.6239, "step": 9738 }, { "epoch": 1.52, "grad_norm": 18.741691860675274, "learning_rate": 1.025714749577601e-05, "loss": 0.556, "step": 9739 }, { "epoch": 1.52, "grad_norm": 20.94721018892791, "learning_rate": 1.0255461650807264e-05, "loss": 0.5684, "step": 9740 }, { "epoch": 1.52, "grad_norm": 22.220464958127668, "learning_rate": 1.025377579857334e-05, "loss": 0.6047, "step": 9741 }, { "epoch": 1.52, "grad_norm": 16.498191713185314, "learning_rate": 1.0252089939122179e-05, "loss": 0.5286, "step": 9742 }, { "epoch": 1.52, "grad_norm": 17.820507427288362, "learning_rate": 1.0250404072501724e-05, "loss": 0.6005, "step": 9743 }, { "epoch": 1.52, "grad_norm": 18.218123499631304, "learning_rate": 1.024871819875993e-05, "loss": 0.5803, "step": 9744 }, { "epoch": 1.52, "grad_norm": 19.56208453483713, "learning_rate": 1.0247032317944731e-05, "loss": 0.6825, "step": 9745 }, { "epoch": 1.52, "grad_norm": 18.358616007679327, "learning_rate": 1.0245346430104082e-05, "loss": 0.5834, "step": 9746 }, { "epoch": 1.52, "grad_norm": 16.489950339584198, "learning_rate": 1.024366053528592e-05, "loss": 0.6595, "step": 9747 }, { "epoch": 1.52, "grad_norm": 25.524704981387046, "learning_rate": 1.0241974633538198e-05, "loss": 0.6182, "step": 9748 }, { "epoch": 1.52, "grad_norm": 16.349640465368353, "learning_rate": 1.0240288724908858e-05, "loss": 0.6046, "step": 9749 }, { "epoch": 1.52, "grad_norm": 22.977115747664868, "learning_rate": 1.0238602809445846e-05, "loss": 0.6919, "step": 9750 }, { "epoch": 1.52, "grad_norm": 18.00082695529993, "learning_rate": 1.023691688719711e-05, "loss": 0.6042, "step": 9751 }, { "epoch": 1.52, "grad_norm": 21.24340528159094, "learning_rate": 1.0235230958210596e-05, "loss": 0.546, "step": 9752 }, { "epoch": 1.52, "grad_norm": 25.84671425165797, "learning_rate": 1.0233545022534252e-05, "loss": 0.6432, "step": 9753 }, { "epoch": 1.52, "grad_norm": 19.40913042170192, "learning_rate": 1.0231859080216025e-05, "loss": 0.6516, "step": 9754 }, { "epoch": 1.52, "grad_norm": 26.12026396343735, "learning_rate": 1.023017313130386e-05, "loss": 0.5668, "step": 9755 }, { "epoch": 1.52, "grad_norm": 18.141763595544948, "learning_rate": 1.0228487175845707e-05, "loss": 0.6057, "step": 9756 }, { "epoch": 1.52, "grad_norm": 20.81871215724684, "learning_rate": 1.0226801213889512e-05, "loss": 0.6012, "step": 9757 }, { "epoch": 1.52, "grad_norm": 30.97131783389649, "learning_rate": 1.0225115245483219e-05, "loss": 0.6591, "step": 9758 }, { "epoch": 1.52, "grad_norm": 21.083537548890842, "learning_rate": 1.0223429270674788e-05, "loss": 0.6875, "step": 9759 }, { "epoch": 1.52, "grad_norm": 16.42947557206542, "learning_rate": 1.0221743289512153e-05, "loss": 0.6043, "step": 9760 }, { "epoch": 1.52, "grad_norm": 23.06860331990219, "learning_rate": 1.0220057302043273e-05, "loss": 0.6349, "step": 9761 }, { "epoch": 1.52, "grad_norm": 12.762559951179993, "learning_rate": 1.0218371308316089e-05, "loss": 0.5893, "step": 9762 }, { "epoch": 1.52, "grad_norm": 20.131425721705238, "learning_rate": 1.0216685308378556e-05, "loss": 0.6209, "step": 9763 }, { "epoch": 1.53, "grad_norm": 20.56900243104382, "learning_rate": 1.0214999302278614e-05, "loss": 0.5722, "step": 9764 }, { "epoch": 1.53, "grad_norm": 17.030102627278943, "learning_rate": 1.0213313290064222e-05, "loss": 0.5976, "step": 9765 }, { "epoch": 1.53, "grad_norm": 20.47728830861015, "learning_rate": 1.0211627271783323e-05, "loss": 0.5158, "step": 9766 }, { "epoch": 1.53, "grad_norm": 17.42929441760261, "learning_rate": 1.0209941247483868e-05, "loss": 0.574, "step": 9767 }, { "epoch": 1.53, "grad_norm": 23.15787797487095, "learning_rate": 1.0208255217213809e-05, "loss": 0.6116, "step": 9768 }, { "epoch": 1.53, "grad_norm": 25.321843004212354, "learning_rate": 1.0206569181021092e-05, "loss": 0.6432, "step": 9769 }, { "epoch": 1.53, "grad_norm": 37.74511198148553, "learning_rate": 1.0204883138953666e-05, "loss": 0.6437, "step": 9770 }, { "epoch": 1.53, "grad_norm": 19.193600862023636, "learning_rate": 1.0203197091059485e-05, "loss": 0.6095, "step": 9771 }, { "epoch": 1.53, "grad_norm": 22.42449413513622, "learning_rate": 1.02015110373865e-05, "loss": 0.5764, "step": 9772 }, { "epoch": 1.53, "grad_norm": 11.61353653422856, "learning_rate": 1.0199824977982658e-05, "loss": 0.5876, "step": 9773 }, { "epoch": 1.53, "grad_norm": 15.01537566618072, "learning_rate": 1.0198138912895907e-05, "loss": 0.5989, "step": 9774 }, { "epoch": 1.53, "grad_norm": 20.777430230276448, "learning_rate": 1.0196452842174202e-05, "loss": 0.5752, "step": 9775 }, { "epoch": 1.53, "grad_norm": 21.44965588164675, "learning_rate": 1.0194766765865498e-05, "loss": 0.6001, "step": 9776 }, { "epoch": 1.53, "grad_norm": 26.35495229814092, "learning_rate": 1.0193080684017737e-05, "loss": 0.6678, "step": 9777 }, { "epoch": 1.53, "grad_norm": 15.17195117393095, "learning_rate": 1.0191394596678879e-05, "loss": 0.6193, "step": 9778 }, { "epoch": 1.53, "grad_norm": 21.180960499109634, "learning_rate": 1.0189708503896865e-05, "loss": 0.5787, "step": 9779 }, { "epoch": 1.53, "grad_norm": 22.98836179604109, "learning_rate": 1.0188022405719652e-05, "loss": 0.6599, "step": 9780 }, { "epoch": 1.53, "grad_norm": 17.355594134001255, "learning_rate": 1.0186336302195197e-05, "loss": 0.655, "step": 9781 }, { "epoch": 1.53, "grad_norm": 23.255886298603194, "learning_rate": 1.0184650193371444e-05, "loss": 0.6365, "step": 9782 }, { "epoch": 1.53, "grad_norm": 19.904635072682954, "learning_rate": 1.0182964079296347e-05, "loss": 0.5763, "step": 9783 }, { "epoch": 1.53, "grad_norm": 17.628876284568133, "learning_rate": 1.0181277960017856e-05, "loss": 0.5941, "step": 9784 }, { "epoch": 1.53, "grad_norm": 20.917966169113157, "learning_rate": 1.017959183558393e-05, "loss": 0.5622, "step": 9785 }, { "epoch": 1.53, "grad_norm": 23.77326155944365, "learning_rate": 1.0177905706042517e-05, "loss": 0.623, "step": 9786 }, { "epoch": 1.53, "grad_norm": 14.67839546872158, "learning_rate": 1.0176219571441565e-05, "loss": 0.552, "step": 9787 }, { "epoch": 1.53, "grad_norm": 17.126178967570326, "learning_rate": 1.0174533431829039e-05, "loss": 0.6912, "step": 9788 }, { "epoch": 1.53, "grad_norm": 25.925455739672177, "learning_rate": 1.0172847287252878e-05, "loss": 0.5756, "step": 9789 }, { "epoch": 1.53, "grad_norm": 20.118236408110047, "learning_rate": 1.0171161137761042e-05, "loss": 0.5954, "step": 9790 }, { "epoch": 1.53, "grad_norm": 15.25787566846281, "learning_rate": 1.0169474983401488e-05, "loss": 0.5304, "step": 9791 }, { "epoch": 1.53, "grad_norm": 19.24041767631331, "learning_rate": 1.016778882422216e-05, "loss": 0.5783, "step": 9792 }, { "epoch": 1.53, "grad_norm": 16.971278162102124, "learning_rate": 1.0166102660271018e-05, "loss": 0.5405, "step": 9793 }, { "epoch": 1.53, "grad_norm": 17.206598312785587, "learning_rate": 1.0164416491596014e-05, "loss": 0.5874, "step": 9794 }, { "epoch": 1.53, "grad_norm": 17.00980527651768, "learning_rate": 1.01627303182451e-05, "loss": 0.5576, "step": 9795 }, { "epoch": 1.53, "grad_norm": 19.555684191899413, "learning_rate": 1.016104414026623e-05, "loss": 0.5853, "step": 9796 }, { "epoch": 1.53, "grad_norm": 26.607579592369152, "learning_rate": 1.015935795770736e-05, "loss": 0.6985, "step": 9797 }, { "epoch": 1.53, "grad_norm": 20.729322284245352, "learning_rate": 1.0157671770616444e-05, "loss": 0.659, "step": 9798 }, { "epoch": 1.53, "grad_norm": 11.630172165972308, "learning_rate": 1.0155985579041434e-05, "loss": 0.5162, "step": 9799 }, { "epoch": 1.53, "grad_norm": 18.524771154057753, "learning_rate": 1.0154299383030287e-05, "loss": 0.5533, "step": 9800 }, { "epoch": 1.53, "grad_norm": 21.371466626955154, "learning_rate": 1.0152613182630953e-05, "loss": 0.684, "step": 9801 }, { "epoch": 1.53, "grad_norm": 19.852210292883242, "learning_rate": 1.0150926977891388e-05, "loss": 0.6126, "step": 9802 }, { "epoch": 1.53, "grad_norm": 28.77283713568664, "learning_rate": 1.0149240768859554e-05, "loss": 0.6412, "step": 9803 }, { "epoch": 1.53, "grad_norm": 14.271240985111056, "learning_rate": 1.0147554555583394e-05, "loss": 0.6132, "step": 9804 }, { "epoch": 1.53, "grad_norm": 17.990411848866618, "learning_rate": 1.0145868338110873e-05, "loss": 0.585, "step": 9805 }, { "epoch": 1.53, "grad_norm": 21.628319372530903, "learning_rate": 1.0144182116489938e-05, "loss": 0.6258, "step": 9806 }, { "epoch": 1.53, "grad_norm": 30.878107570909272, "learning_rate": 1.0142495890768551e-05, "loss": 0.6224, "step": 9807 }, { "epoch": 1.53, "grad_norm": 23.719539571393554, "learning_rate": 1.0140809660994663e-05, "loss": 0.6045, "step": 9808 }, { "epoch": 1.53, "grad_norm": 22.16270071141561, "learning_rate": 1.0139123427216231e-05, "loss": 0.6214, "step": 9809 }, { "epoch": 1.53, "grad_norm": 15.264404714201888, "learning_rate": 1.013743718948121e-05, "loss": 0.5899, "step": 9810 }, { "epoch": 1.53, "grad_norm": 18.782291226396453, "learning_rate": 1.0135750947837558e-05, "loss": 0.5736, "step": 9811 }, { "epoch": 1.53, "grad_norm": 14.00041017878756, "learning_rate": 1.0134064702333225e-05, "loss": 0.5152, "step": 9812 }, { "epoch": 1.53, "grad_norm": 17.09147224906994, "learning_rate": 1.0132378453016171e-05, "loss": 0.6397, "step": 9813 }, { "epoch": 1.53, "grad_norm": 17.93878225236053, "learning_rate": 1.013069219993435e-05, "loss": 0.5874, "step": 9814 }, { "epoch": 1.53, "grad_norm": 18.597211456268017, "learning_rate": 1.0129005943135721e-05, "loss": 0.5429, "step": 9815 }, { "epoch": 1.53, "grad_norm": 13.85037321276858, "learning_rate": 1.012731968266824e-05, "loss": 0.5659, "step": 9816 }, { "epoch": 1.53, "grad_norm": 23.714440965821364, "learning_rate": 1.012563341857986e-05, "loss": 0.5713, "step": 9817 }, { "epoch": 1.53, "grad_norm": 22.695734829018026, "learning_rate": 1.0123947150918539e-05, "loss": 0.5793, "step": 9818 }, { "epoch": 1.53, "grad_norm": 18.717277534218006, "learning_rate": 1.0122260879732231e-05, "loss": 0.6535, "step": 9819 }, { "epoch": 1.53, "grad_norm": 20.35976604736056, "learning_rate": 1.01205746050689e-05, "loss": 0.7365, "step": 9820 }, { "epoch": 1.53, "grad_norm": 26.98959759102438, "learning_rate": 1.0118888326976494e-05, "loss": 0.6402, "step": 9821 }, { "epoch": 1.53, "grad_norm": 20.390692643532052, "learning_rate": 1.0117202045502978e-05, "loss": 0.6214, "step": 9822 }, { "epoch": 1.53, "grad_norm": 25.895291304471748, "learning_rate": 1.01155157606963e-05, "loss": 0.5579, "step": 9823 }, { "epoch": 1.53, "grad_norm": 20.940143999718817, "learning_rate": 1.0113829472604422e-05, "loss": 0.6262, "step": 9824 }, { "epoch": 1.53, "grad_norm": 18.693191947661525, "learning_rate": 1.0112143181275302e-05, "loss": 0.5833, "step": 9825 }, { "epoch": 1.53, "grad_norm": 12.92902628311404, "learning_rate": 1.0110456886756894e-05, "loss": 0.5885, "step": 9826 }, { "epoch": 1.53, "grad_norm": 18.488499609300288, "learning_rate": 1.010877058909716e-05, "loss": 0.6217, "step": 9827 }, { "epoch": 1.54, "grad_norm": 14.201822237923107, "learning_rate": 1.0107084288344052e-05, "loss": 0.6189, "step": 9828 }, { "epoch": 1.54, "grad_norm": 20.196419569325847, "learning_rate": 1.0105397984545524e-05, "loss": 0.6501, "step": 9829 }, { "epoch": 1.54, "grad_norm": 22.0816349336105, "learning_rate": 1.0103711677749548e-05, "loss": 0.5687, "step": 9830 }, { "epoch": 1.54, "grad_norm": 23.91595385392772, "learning_rate": 1.0102025368004066e-05, "loss": 0.5631, "step": 9831 }, { "epoch": 1.54, "grad_norm": 19.674529830649426, "learning_rate": 1.0100339055357048e-05, "loss": 0.6132, "step": 9832 }, { "epoch": 1.54, "grad_norm": 16.405248610498568, "learning_rate": 1.0098652739856441e-05, "loss": 0.6614, "step": 9833 }, { "epoch": 1.54, "grad_norm": 18.41468111471084, "learning_rate": 1.0096966421550209e-05, "loss": 0.5909, "step": 9834 }, { "epoch": 1.54, "grad_norm": 19.370706133186534, "learning_rate": 1.0095280100486309e-05, "loss": 0.6184, "step": 9835 }, { "epoch": 1.54, "grad_norm": 16.965911400366622, "learning_rate": 1.00935937767127e-05, "loss": 0.5726, "step": 9836 }, { "epoch": 1.54, "grad_norm": 14.131970856564038, "learning_rate": 1.0091907450277338e-05, "loss": 0.531, "step": 9837 }, { "epoch": 1.54, "grad_norm": 17.09407529094273, "learning_rate": 1.0090221121228178e-05, "loss": 0.5342, "step": 9838 }, { "epoch": 1.54, "grad_norm": 28.24465590330699, "learning_rate": 1.0088534789613188e-05, "loss": 0.6801, "step": 9839 }, { "epoch": 1.54, "grad_norm": 20.876139976588874, "learning_rate": 1.0086848455480318e-05, "loss": 0.6493, "step": 9840 }, { "epoch": 1.54, "grad_norm": 24.241602855250118, "learning_rate": 1.0085162118877527e-05, "loss": 0.5267, "step": 9841 }, { "epoch": 1.54, "grad_norm": 25.365857996756986, "learning_rate": 1.0083475779852778e-05, "loss": 0.6316, "step": 9842 }, { "epoch": 1.54, "grad_norm": 17.95590279190697, "learning_rate": 1.0081789438454026e-05, "loss": 0.5928, "step": 9843 }, { "epoch": 1.54, "grad_norm": 18.109198327474008, "learning_rate": 1.0080103094729229e-05, "loss": 0.5614, "step": 9844 }, { "epoch": 1.54, "grad_norm": 22.289320802512265, "learning_rate": 1.007841674872635e-05, "loss": 0.6248, "step": 9845 }, { "epoch": 1.54, "grad_norm": 16.843820585021568, "learning_rate": 1.007673040049334e-05, "loss": 0.5457, "step": 9846 }, { "epoch": 1.54, "grad_norm": 28.343440797542915, "learning_rate": 1.0075044050078166e-05, "loss": 0.6412, "step": 9847 }, { "epoch": 1.54, "grad_norm": 16.585986569009034, "learning_rate": 1.0073357697528779e-05, "loss": 0.6133, "step": 9848 }, { "epoch": 1.54, "grad_norm": 22.969192878190256, "learning_rate": 1.0071671342893148e-05, "loss": 0.5725, "step": 9849 }, { "epoch": 1.54, "grad_norm": 18.826596777377638, "learning_rate": 1.0069984986219219e-05, "loss": 0.6182, "step": 9850 }, { "epoch": 1.54, "grad_norm": 17.93768335125752, "learning_rate": 1.0068298627554962e-05, "loss": 0.5551, "step": 9851 }, { "epoch": 1.54, "grad_norm": 27.673201439254342, "learning_rate": 1.0066612266948333e-05, "loss": 0.5886, "step": 9852 }, { "epoch": 1.54, "grad_norm": 16.402230321188068, "learning_rate": 1.0064925904447288e-05, "loss": 0.6064, "step": 9853 }, { "epoch": 1.54, "grad_norm": 22.52270246081074, "learning_rate": 1.006323954009979e-05, "loss": 0.6101, "step": 9854 }, { "epoch": 1.54, "grad_norm": 20.883609994830113, "learning_rate": 1.0061553173953794e-05, "loss": 0.6599, "step": 9855 }, { "epoch": 1.54, "grad_norm": 15.423958749690176, "learning_rate": 1.0059866806057263e-05, "loss": 0.5467, "step": 9856 }, { "epoch": 1.54, "grad_norm": 28.857281790198073, "learning_rate": 1.0058180436458156e-05, "loss": 0.5315, "step": 9857 }, { "epoch": 1.54, "grad_norm": 15.796284954170929, "learning_rate": 1.0056494065204428e-05, "loss": 0.5457, "step": 9858 }, { "epoch": 1.54, "grad_norm": 17.452180206763426, "learning_rate": 1.0054807692344045e-05, "loss": 0.6638, "step": 9859 }, { "epoch": 1.54, "grad_norm": 14.82281976866086, "learning_rate": 1.0053121317924963e-05, "loss": 0.516, "step": 9860 }, { "epoch": 1.54, "grad_norm": 22.84377747911087, "learning_rate": 1.0051434941995142e-05, "loss": 0.6408, "step": 9861 }, { "epoch": 1.54, "grad_norm": 20.496517203922064, "learning_rate": 1.004974856460254e-05, "loss": 0.5712, "step": 9862 }, { "epoch": 1.54, "grad_norm": 16.528885550589, "learning_rate": 1.0048062185795117e-05, "loss": 0.6111, "step": 9863 }, { "epoch": 1.54, "grad_norm": 15.583634769062323, "learning_rate": 1.0046375805620838e-05, "loss": 0.5953, "step": 9864 }, { "epoch": 1.54, "grad_norm": 28.530671231331553, "learning_rate": 1.004468942412765e-05, "loss": 0.6013, "step": 9865 }, { "epoch": 1.54, "grad_norm": 28.883483566276666, "learning_rate": 1.0043003041363531e-05, "loss": 0.5862, "step": 9866 }, { "epoch": 1.54, "grad_norm": 11.668462437624736, "learning_rate": 1.0041316657376426e-05, "loss": 0.5047, "step": 9867 }, { "epoch": 1.54, "grad_norm": 19.723909886349947, "learning_rate": 1.0039630272214299e-05, "loss": 0.5517, "step": 9868 }, { "epoch": 1.54, "grad_norm": 15.050017125355206, "learning_rate": 1.003794388592511e-05, "loss": 0.5672, "step": 9869 }, { "epoch": 1.54, "grad_norm": 13.931392360492978, "learning_rate": 1.0036257498556821e-05, "loss": 0.6131, "step": 9870 }, { "epoch": 1.54, "grad_norm": 17.521188137504225, "learning_rate": 1.0034571110157388e-05, "loss": 0.5368, "step": 9871 }, { "epoch": 1.54, "grad_norm": 16.832271467979282, "learning_rate": 1.0032884720774773e-05, "loss": 0.6052, "step": 9872 }, { "epoch": 1.54, "grad_norm": 21.914611576973698, "learning_rate": 1.0031198330456936e-05, "loss": 0.5725, "step": 9873 }, { "epoch": 1.54, "grad_norm": 19.94124677633545, "learning_rate": 1.002951193925184e-05, "loss": 0.6069, "step": 9874 }, { "epoch": 1.54, "grad_norm": 15.148054018719586, "learning_rate": 1.0027825547207435e-05, "loss": 0.5259, "step": 9875 }, { "epoch": 1.54, "grad_norm": 27.810999591946036, "learning_rate": 1.0026139154371694e-05, "loss": 0.5989, "step": 9876 }, { "epoch": 1.54, "grad_norm": 14.218846892980409, "learning_rate": 1.0024452760792566e-05, "loss": 0.6176, "step": 9877 }, { "epoch": 1.54, "grad_norm": 22.009688907103346, "learning_rate": 1.0022766366518018e-05, "loss": 0.6078, "step": 9878 }, { "epoch": 1.54, "grad_norm": 26.936615283271358, "learning_rate": 1.0021079971596009e-05, "loss": 0.6418, "step": 9879 }, { "epoch": 1.54, "grad_norm": 14.099479744977542, "learning_rate": 1.0019393576074497e-05, "loss": 0.4746, "step": 9880 }, { "epoch": 1.54, "grad_norm": 14.946486662262915, "learning_rate": 1.0017707180001443e-05, "loss": 0.5979, "step": 9881 }, { "epoch": 1.54, "grad_norm": 20.441687089393028, "learning_rate": 1.0016020783424805e-05, "loss": 0.6242, "step": 9882 }, { "epoch": 1.54, "grad_norm": 19.200995576274416, "learning_rate": 1.001433438639255e-05, "loss": 0.5817, "step": 9883 }, { "epoch": 1.54, "grad_norm": 25.13079098251688, "learning_rate": 1.0012647988952628e-05, "loss": 0.7288, "step": 9884 }, { "epoch": 1.54, "grad_norm": 12.194260132108658, "learning_rate": 1.0010961591153008e-05, "loss": 0.6085, "step": 9885 }, { "epoch": 1.54, "grad_norm": 15.953252793714666, "learning_rate": 1.0009275193041645e-05, "loss": 0.6196, "step": 9886 }, { "epoch": 1.54, "grad_norm": 29.297800509520645, "learning_rate": 1.0007588794666503e-05, "loss": 0.6428, "step": 9887 }, { "epoch": 1.54, "grad_norm": 15.571010031342773, "learning_rate": 1.000590239607554e-05, "loss": 0.5867, "step": 9888 }, { "epoch": 1.54, "grad_norm": 15.933519257083526, "learning_rate": 1.0004215997316715e-05, "loss": 0.5932, "step": 9889 }, { "epoch": 1.54, "grad_norm": 16.45958911313961, "learning_rate": 1.0002529598437988e-05, "loss": 0.5464, "step": 9890 }, { "epoch": 1.54, "grad_norm": 16.00742481785263, "learning_rate": 1.0000843199487325e-05, "loss": 0.6235, "step": 9891 }, { "epoch": 1.55, "grad_norm": 16.62902593804824, "learning_rate": 9.99915680051268e-06, "loss": 0.6296, "step": 9892 }, { "epoch": 1.55, "grad_norm": 24.121323145343172, "learning_rate": 9.997470401562015e-06, "loss": 0.5748, "step": 9893 }, { "epoch": 1.55, "grad_norm": 19.20997601089092, "learning_rate": 9.995784002683288e-06, "loss": 0.6191, "step": 9894 }, { "epoch": 1.55, "grad_norm": 13.328343513168837, "learning_rate": 9.994097603924462e-06, "loss": 0.5634, "step": 9895 }, { "epoch": 1.55, "grad_norm": 16.472865471744875, "learning_rate": 9.992411205333498e-06, "loss": 0.6051, "step": 9896 }, { "epoch": 1.55, "grad_norm": 21.655178533817047, "learning_rate": 9.990724806958358e-06, "loss": 0.5335, "step": 9897 }, { "epoch": 1.55, "grad_norm": 21.636031693794646, "learning_rate": 9.989038408846996e-06, "loss": 0.6225, "step": 9898 }, { "epoch": 1.55, "grad_norm": 21.846735055191772, "learning_rate": 9.987352011047374e-06, "loss": 0.6031, "step": 9899 }, { "epoch": 1.55, "grad_norm": 22.382934042715444, "learning_rate": 9.985665613607454e-06, "loss": 0.5799, "step": 9900 }, { "epoch": 1.55, "grad_norm": 16.21675362455437, "learning_rate": 9.983979216575195e-06, "loss": 0.534, "step": 9901 }, { "epoch": 1.55, "grad_norm": 20.166205386545933, "learning_rate": 9.982292819998562e-06, "loss": 0.6495, "step": 9902 }, { "epoch": 1.55, "grad_norm": 17.94326492206648, "learning_rate": 9.980606423925506e-06, "loss": 0.7226, "step": 9903 }, { "epoch": 1.55, "grad_norm": 19.587813356402894, "learning_rate": 9.978920028403995e-06, "loss": 0.6983, "step": 9904 }, { "epoch": 1.55, "grad_norm": 23.414201246632665, "learning_rate": 9.977233633481984e-06, "loss": 0.5428, "step": 9905 }, { "epoch": 1.55, "grad_norm": 18.64245689516531, "learning_rate": 9.975547239207435e-06, "loss": 0.5448, "step": 9906 }, { "epoch": 1.55, "grad_norm": 24.025556916818953, "learning_rate": 9.973860845628311e-06, "loss": 0.6219, "step": 9907 }, { "epoch": 1.55, "grad_norm": 17.963216960531298, "learning_rate": 9.972174452792568e-06, "loss": 0.537, "step": 9908 }, { "epoch": 1.55, "grad_norm": 18.694681263352486, "learning_rate": 9.970488060748164e-06, "loss": 0.585, "step": 9909 }, { "epoch": 1.55, "grad_norm": 22.050040062475464, "learning_rate": 9.968801669543066e-06, "loss": 0.6326, "step": 9910 }, { "epoch": 1.55, "grad_norm": 21.306401981461846, "learning_rate": 9.967115279225228e-06, "loss": 0.6385, "step": 9911 }, { "epoch": 1.55, "grad_norm": 20.722230479390333, "learning_rate": 9.965428889842617e-06, "loss": 0.5485, "step": 9912 }, { "epoch": 1.55, "grad_norm": 14.268524689447355, "learning_rate": 9.963742501443184e-06, "loss": 0.5791, "step": 9913 }, { "epoch": 1.55, "grad_norm": 15.611129566668549, "learning_rate": 9.962056114074893e-06, "loss": 0.5641, "step": 9914 }, { "epoch": 1.55, "grad_norm": 18.79455886041507, "learning_rate": 9.960369727785703e-06, "loss": 0.687, "step": 9915 }, { "epoch": 1.55, "grad_norm": 15.246771988185504, "learning_rate": 9.958683342623579e-06, "loss": 0.5719, "step": 9916 }, { "epoch": 1.55, "grad_norm": 18.210505642019125, "learning_rate": 9.956996958636474e-06, "loss": 0.5923, "step": 9917 }, { "epoch": 1.55, "grad_norm": 24.489068636129506, "learning_rate": 9.955310575872351e-06, "loss": 0.6331, "step": 9918 }, { "epoch": 1.55, "grad_norm": 21.104498795216216, "learning_rate": 9.953624194379165e-06, "loss": 0.5477, "step": 9919 }, { "epoch": 1.55, "grad_norm": 11.122084049287043, "learning_rate": 9.951937814204884e-06, "loss": 0.5328, "step": 9920 }, { "epoch": 1.55, "grad_norm": 20.779803803681574, "learning_rate": 9.950251435397466e-06, "loss": 0.6101, "step": 9921 }, { "epoch": 1.55, "grad_norm": 19.895115824106036, "learning_rate": 9.948565058004863e-06, "loss": 0.6086, "step": 9922 }, { "epoch": 1.55, "grad_norm": 20.166533984156313, "learning_rate": 9.94687868207504e-06, "loss": 0.5932, "step": 9923 }, { "epoch": 1.55, "grad_norm": 15.573150429252294, "learning_rate": 9.945192307655959e-06, "loss": 0.5164, "step": 9924 }, { "epoch": 1.55, "grad_norm": 16.22028948068708, "learning_rate": 9.943505934795572e-06, "loss": 0.6668, "step": 9925 }, { "epoch": 1.55, "grad_norm": 24.384813425233407, "learning_rate": 9.941819563541849e-06, "loss": 0.644, "step": 9926 }, { "epoch": 1.55, "grad_norm": 18.500511899354784, "learning_rate": 9.940133193942742e-06, "loss": 0.576, "step": 9927 }, { "epoch": 1.55, "grad_norm": 26.124508283290584, "learning_rate": 9.938446826046209e-06, "loss": 0.614, "step": 9928 }, { "epoch": 1.55, "grad_norm": 22.123440994681633, "learning_rate": 9.936760459900215e-06, "loss": 0.5415, "step": 9929 }, { "epoch": 1.55, "grad_norm": 21.66765646729247, "learning_rate": 9.935074095552714e-06, "loss": 0.6245, "step": 9930 }, { "epoch": 1.55, "grad_norm": 16.512151878962634, "learning_rate": 9.933387733051672e-06, "loss": 0.6438, "step": 9931 }, { "epoch": 1.55, "grad_norm": 16.247368819620945, "learning_rate": 9.93170137244504e-06, "loss": 0.6089, "step": 9932 }, { "epoch": 1.55, "grad_norm": 15.046404442505159, "learning_rate": 9.930015013780783e-06, "loss": 0.5949, "step": 9933 }, { "epoch": 1.55, "grad_norm": 15.964185564414901, "learning_rate": 9.928328657106855e-06, "loss": 0.4891, "step": 9934 }, { "epoch": 1.55, "grad_norm": 19.93954144238853, "learning_rate": 9.92664230247122e-06, "loss": 0.644, "step": 9935 }, { "epoch": 1.55, "grad_norm": 19.822924170105942, "learning_rate": 9.924955949921839e-06, "loss": 0.5922, "step": 9936 }, { "epoch": 1.55, "grad_norm": 26.63956348843764, "learning_rate": 9.923269599506664e-06, "loss": 0.6541, "step": 9937 }, { "epoch": 1.55, "grad_norm": 24.025018378244766, "learning_rate": 9.921583251273654e-06, "loss": 0.5801, "step": 9938 }, { "epoch": 1.55, "grad_norm": 21.50971012217428, "learning_rate": 9.919896905270772e-06, "loss": 0.6773, "step": 9939 }, { "epoch": 1.55, "grad_norm": 12.90671493267816, "learning_rate": 9.918210561545974e-06, "loss": 0.559, "step": 9940 }, { "epoch": 1.55, "grad_norm": 20.98238292336455, "learning_rate": 9.916524220147224e-06, "loss": 0.6073, "step": 9941 }, { "epoch": 1.55, "grad_norm": 14.50327156007637, "learning_rate": 9.914837881122474e-06, "loss": 0.4855, "step": 9942 }, { "epoch": 1.55, "grad_norm": 16.27992628559294, "learning_rate": 9.913151544519685e-06, "loss": 0.5874, "step": 9943 }, { "epoch": 1.55, "grad_norm": 22.04891448232537, "learning_rate": 9.911465210386813e-06, "loss": 0.6289, "step": 9944 }, { "epoch": 1.55, "grad_norm": 16.440772380831724, "learning_rate": 9.909778878771822e-06, "loss": 0.5556, "step": 9945 }, { "epoch": 1.55, "grad_norm": 24.39156764333135, "learning_rate": 9.908092549722667e-06, "loss": 0.6782, "step": 9946 }, { "epoch": 1.55, "grad_norm": 22.937471535272632, "learning_rate": 9.906406223287304e-06, "loss": 0.6178, "step": 9947 }, { "epoch": 1.55, "grad_norm": 16.38511420886467, "learning_rate": 9.904719899513693e-06, "loss": 0.5479, "step": 9948 }, { "epoch": 1.55, "grad_norm": 15.071545944442061, "learning_rate": 9.903033578449793e-06, "loss": 0.5566, "step": 9949 }, { "epoch": 1.55, "grad_norm": 19.42644736235071, "learning_rate": 9.90134726014356e-06, "loss": 0.5795, "step": 9950 }, { "epoch": 1.55, "grad_norm": 26.848196280376573, "learning_rate": 9.899660944642957e-06, "loss": 0.5194, "step": 9951 }, { "epoch": 1.55, "grad_norm": 18.89129961708667, "learning_rate": 9.897974631995937e-06, "loss": 0.6012, "step": 9952 }, { "epoch": 1.55, "grad_norm": 15.843725417473586, "learning_rate": 9.896288322250455e-06, "loss": 0.5459, "step": 9953 }, { "epoch": 1.55, "grad_norm": 27.010483674494523, "learning_rate": 9.894602015454476e-06, "loss": 0.603, "step": 9954 }, { "epoch": 1.55, "grad_norm": 13.783049845427678, "learning_rate": 9.892915711655953e-06, "loss": 0.5753, "step": 9955 }, { "epoch": 1.56, "grad_norm": 27.184418986024316, "learning_rate": 9.891229410902846e-06, "loss": 0.6342, "step": 9956 }, { "epoch": 1.56, "grad_norm": 33.941184736712835, "learning_rate": 9.88954311324311e-06, "loss": 0.5585, "step": 9957 }, { "epoch": 1.56, "grad_norm": 18.565286535572284, "learning_rate": 9.887856818724702e-06, "loss": 0.6155, "step": 9958 }, { "epoch": 1.56, "grad_norm": 15.423762633025335, "learning_rate": 9.88617052739558e-06, "loss": 0.5807, "step": 9959 }, { "epoch": 1.56, "grad_norm": 24.059084364987115, "learning_rate": 9.8844842393037e-06, "loss": 0.6225, "step": 9960 }, { "epoch": 1.56, "grad_norm": 15.25280894387938, "learning_rate": 9.882797954497028e-06, "loss": 0.5436, "step": 9961 }, { "epoch": 1.56, "grad_norm": 18.743540728517914, "learning_rate": 9.881111673023509e-06, "loss": 0.5998, "step": 9962 }, { "epoch": 1.56, "grad_norm": 16.814166057385815, "learning_rate": 9.879425394931103e-06, "loss": 0.577, "step": 9963 }, { "epoch": 1.56, "grad_norm": 16.453038243484357, "learning_rate": 9.877739120267769e-06, "loss": 0.5843, "step": 9964 }, { "epoch": 1.56, "grad_norm": 13.67659237539244, "learning_rate": 9.876052849081467e-06, "loss": 0.6041, "step": 9965 }, { "epoch": 1.56, "grad_norm": 12.768110269139063, "learning_rate": 9.874366581420144e-06, "loss": 0.624, "step": 9966 }, { "epoch": 1.56, "grad_norm": 19.144665052704863, "learning_rate": 9.872680317331764e-06, "loss": 0.5898, "step": 9967 }, { "epoch": 1.56, "grad_norm": 16.512891213380023, "learning_rate": 9.87099405686428e-06, "loss": 0.5721, "step": 9968 }, { "epoch": 1.56, "grad_norm": 16.02635306769787, "learning_rate": 9.869307800065651e-06, "loss": 0.5754, "step": 9969 }, { "epoch": 1.56, "grad_norm": 19.452335579109473, "learning_rate": 9.867621546983834e-06, "loss": 0.5932, "step": 9970 }, { "epoch": 1.56, "grad_norm": 15.942141635817212, "learning_rate": 9.86593529766678e-06, "loss": 0.623, "step": 9971 }, { "epoch": 1.56, "grad_norm": 16.76215330603838, "learning_rate": 9.864249052162447e-06, "loss": 0.504, "step": 9972 }, { "epoch": 1.56, "grad_norm": 16.11229616887807, "learning_rate": 9.862562810518792e-06, "loss": 0.5537, "step": 9973 }, { "epoch": 1.56, "grad_norm": 17.006137157785712, "learning_rate": 9.860876572783772e-06, "loss": 0.6332, "step": 9974 }, { "epoch": 1.56, "grad_norm": 14.069077922642027, "learning_rate": 9.859190339005342e-06, "loss": 0.5788, "step": 9975 }, { "epoch": 1.56, "grad_norm": 21.022021846320726, "learning_rate": 9.857504109231452e-06, "loss": 0.6415, "step": 9976 }, { "epoch": 1.56, "grad_norm": 16.438184969558296, "learning_rate": 9.855817883510063e-06, "loss": 0.5972, "step": 9977 }, { "epoch": 1.56, "grad_norm": 19.07915214285458, "learning_rate": 9.854131661889129e-06, "loss": 0.627, "step": 9978 }, { "epoch": 1.56, "grad_norm": 17.43861554578508, "learning_rate": 9.852445444416606e-06, "loss": 0.6388, "step": 9979 }, { "epoch": 1.56, "grad_norm": 17.964867473436495, "learning_rate": 9.850759231140451e-06, "loss": 0.5946, "step": 9980 }, { "epoch": 1.56, "grad_norm": 20.13246155737529, "learning_rate": 9.849073022108613e-06, "loss": 0.5865, "step": 9981 }, { "epoch": 1.56, "grad_norm": 19.534122396204978, "learning_rate": 9.84738681736905e-06, "loss": 0.6329, "step": 9982 }, { "epoch": 1.56, "grad_norm": 20.400952487656827, "learning_rate": 9.845700616969718e-06, "loss": 0.7185, "step": 9983 }, { "epoch": 1.56, "grad_norm": 14.862095055385375, "learning_rate": 9.844014420958567e-06, "loss": 0.611, "step": 9984 }, { "epoch": 1.56, "grad_norm": 24.366300341082948, "learning_rate": 9.84232822938356e-06, "loss": 0.5397, "step": 9985 }, { "epoch": 1.56, "grad_norm": 24.757542661964088, "learning_rate": 9.840642042292643e-06, "loss": 0.6127, "step": 9986 }, { "epoch": 1.56, "grad_norm": 19.247783334322346, "learning_rate": 9.838955859733773e-06, "loss": 0.6427, "step": 9987 }, { "epoch": 1.56, "grad_norm": 20.24752014923365, "learning_rate": 9.837269681754901e-06, "loss": 0.6351, "step": 9988 }, { "epoch": 1.56, "grad_norm": 19.357579347341186, "learning_rate": 9.83558350840399e-06, "loss": 0.6147, "step": 9989 }, { "epoch": 1.56, "grad_norm": 27.714189794566042, "learning_rate": 9.833897339728987e-06, "loss": 0.6498, "step": 9990 }, { "epoch": 1.56, "grad_norm": 21.38621738142277, "learning_rate": 9.832211175777841e-06, "loss": 0.626, "step": 9991 }, { "epoch": 1.56, "grad_norm": 18.86324138088125, "learning_rate": 9.830525016598515e-06, "loss": 0.6015, "step": 9992 }, { "epoch": 1.56, "grad_norm": 18.786048721210143, "learning_rate": 9.82883886223896e-06, "loss": 0.6237, "step": 9993 }, { "epoch": 1.56, "grad_norm": 16.315495899431646, "learning_rate": 9.827152712747122e-06, "loss": 0.5222, "step": 9994 }, { "epoch": 1.56, "grad_norm": 13.484828884977205, "learning_rate": 9.825466568170966e-06, "loss": 0.5161, "step": 9995 }, { "epoch": 1.56, "grad_norm": 23.56019167002632, "learning_rate": 9.823780428558437e-06, "loss": 0.5658, "step": 9996 }, { "epoch": 1.56, "grad_norm": 25.197611497185168, "learning_rate": 9.822094293957486e-06, "loss": 0.6232, "step": 9997 }, { "epoch": 1.56, "grad_norm": 19.839589939946194, "learning_rate": 9.820408164416071e-06, "loss": 0.5993, "step": 9998 }, { "epoch": 1.56, "grad_norm": 18.453414585254144, "learning_rate": 9.818722039982145e-06, "loss": 0.5559, "step": 9999 }, { "epoch": 1.56, "grad_norm": 17.429415153109442, "learning_rate": 9.81703592070366e-06, "loss": 0.5934, "step": 10000 }, { "epoch": 1.56, "grad_norm": 12.833704595167802, "learning_rate": 9.81534980662856e-06, "loss": 0.627, "step": 10001 }, { "epoch": 1.56, "grad_norm": 22.602764708174913, "learning_rate": 9.813663697804808e-06, "loss": 0.6003, "step": 10002 }, { "epoch": 1.56, "grad_norm": 25.716150741705444, "learning_rate": 9.811977594280348e-06, "loss": 0.5747, "step": 10003 }, { "epoch": 1.56, "grad_norm": 23.605941266512165, "learning_rate": 9.810291496103137e-06, "loss": 0.6323, "step": 10004 }, { "epoch": 1.56, "grad_norm": 32.268516663109075, "learning_rate": 9.808605403321128e-06, "loss": 0.7072, "step": 10005 }, { "epoch": 1.56, "grad_norm": 18.580157368921157, "learning_rate": 9.806919315982266e-06, "loss": 0.709, "step": 10006 }, { "epoch": 1.56, "grad_norm": 20.00704672918317, "learning_rate": 9.805233234134504e-06, "loss": 0.5557, "step": 10007 }, { "epoch": 1.56, "grad_norm": 17.86993196473675, "learning_rate": 9.803547157825796e-06, "loss": 0.6805, "step": 10008 }, { "epoch": 1.56, "grad_norm": 23.02845911167214, "learning_rate": 9.801861087104093e-06, "loss": 0.5853, "step": 10009 }, { "epoch": 1.56, "grad_norm": 20.83648681667443, "learning_rate": 9.800175022017346e-06, "loss": 0.5768, "step": 10010 }, { "epoch": 1.56, "grad_norm": 19.236295876082348, "learning_rate": 9.798488962613503e-06, "loss": 0.5331, "step": 10011 }, { "epoch": 1.56, "grad_norm": 10.249045818244447, "learning_rate": 9.796802908940516e-06, "loss": 0.5372, "step": 10012 }, { "epoch": 1.56, "grad_norm": 20.572018369707923, "learning_rate": 9.795116861046334e-06, "loss": 0.6242, "step": 10013 }, { "epoch": 1.56, "grad_norm": 19.72488461600909, "learning_rate": 9.793430818978913e-06, "loss": 0.6959, "step": 10014 }, { "epoch": 1.56, "grad_norm": 24.468377731677613, "learning_rate": 9.791744782786196e-06, "loss": 0.5991, "step": 10015 }, { "epoch": 1.56, "grad_norm": 14.222245572684825, "learning_rate": 9.790058752516134e-06, "loss": 0.6418, "step": 10016 }, { "epoch": 1.56, "grad_norm": 16.746107808403732, "learning_rate": 9.788372728216679e-06, "loss": 0.5471, "step": 10017 }, { "epoch": 1.56, "grad_norm": 16.25826421638985, "learning_rate": 9.786686709935781e-06, "loss": 0.5759, "step": 10018 }, { "epoch": 1.56, "grad_norm": 17.3216778129315, "learning_rate": 9.785000697721391e-06, "loss": 0.5616, "step": 10019 }, { "epoch": 1.57, "grad_norm": 16.998181417608414, "learning_rate": 9.783314691621451e-06, "loss": 0.6413, "step": 10020 }, { "epoch": 1.57, "grad_norm": 14.35888629741709, "learning_rate": 9.781628691683916e-06, "loss": 0.6424, "step": 10021 }, { "epoch": 1.57, "grad_norm": 17.459458089193458, "learning_rate": 9.77994269795673e-06, "loss": 0.5769, "step": 10022 }, { "epoch": 1.57, "grad_norm": 15.25476294638253, "learning_rate": 9.778256710487849e-06, "loss": 0.5612, "step": 10023 }, { "epoch": 1.57, "grad_norm": 23.980834826409183, "learning_rate": 9.776570729325217e-06, "loss": 0.6082, "step": 10024 }, { "epoch": 1.57, "grad_norm": 14.90638193236052, "learning_rate": 9.774884754516783e-06, "loss": 0.6031, "step": 10025 }, { "epoch": 1.57, "grad_norm": 15.400623857259639, "learning_rate": 9.773198786110492e-06, "loss": 0.6128, "step": 10026 }, { "epoch": 1.57, "grad_norm": 19.086633911575074, "learning_rate": 9.771512824154297e-06, "loss": 0.5134, "step": 10027 }, { "epoch": 1.57, "grad_norm": 18.330108139346816, "learning_rate": 9.76982686869614e-06, "loss": 0.572, "step": 10028 }, { "epoch": 1.57, "grad_norm": 18.356847875643304, "learning_rate": 9.768140919783979e-06, "loss": 0.5971, "step": 10029 }, { "epoch": 1.57, "grad_norm": 19.6674742646926, "learning_rate": 9.766454977465751e-06, "loss": 0.6008, "step": 10030 }, { "epoch": 1.57, "grad_norm": 25.692674453288213, "learning_rate": 9.764769041789408e-06, "loss": 0.5852, "step": 10031 }, { "epoch": 1.57, "grad_norm": 23.52018672952823, "learning_rate": 9.763083112802891e-06, "loss": 0.558, "step": 10032 }, { "epoch": 1.57, "grad_norm": 14.73908533955712, "learning_rate": 9.761397190554156e-06, "loss": 0.6074, "step": 10033 }, { "epoch": 1.57, "grad_norm": 16.529178335437962, "learning_rate": 9.759711275091149e-06, "loss": 0.5673, "step": 10034 }, { "epoch": 1.57, "grad_norm": 14.023308214107018, "learning_rate": 9.758025366461805e-06, "loss": 0.5685, "step": 10035 }, { "epoch": 1.57, "grad_norm": 15.723803368653815, "learning_rate": 9.756339464714081e-06, "loss": 0.5399, "step": 10036 }, { "epoch": 1.57, "grad_norm": 14.34524316952747, "learning_rate": 9.754653569895922e-06, "loss": 0.5646, "step": 10037 }, { "epoch": 1.57, "grad_norm": 27.518265496216117, "learning_rate": 9.752967682055269e-06, "loss": 0.582, "step": 10038 }, { "epoch": 1.57, "grad_norm": 16.7688568391926, "learning_rate": 9.751281801240075e-06, "loss": 0.756, "step": 10039 }, { "epoch": 1.57, "grad_norm": 16.830683029405115, "learning_rate": 9.749595927498277e-06, "loss": 0.664, "step": 10040 }, { "epoch": 1.57, "grad_norm": 15.67313052897807, "learning_rate": 9.747910060877824e-06, "loss": 0.5482, "step": 10041 }, { "epoch": 1.57, "grad_norm": 16.71254616101865, "learning_rate": 9.746224201426663e-06, "loss": 0.566, "step": 10042 }, { "epoch": 1.57, "grad_norm": 14.922229060621065, "learning_rate": 9.744538349192736e-06, "loss": 0.5583, "step": 10043 }, { "epoch": 1.57, "grad_norm": 16.172194401263614, "learning_rate": 9.742852504223995e-06, "loss": 0.5571, "step": 10044 }, { "epoch": 1.57, "grad_norm": 21.68311685880578, "learning_rate": 9.741166666568371e-06, "loss": 0.5765, "step": 10045 }, { "epoch": 1.57, "grad_norm": 18.700981563864698, "learning_rate": 9.739480836273819e-06, "loss": 0.5952, "step": 10046 }, { "epoch": 1.57, "grad_norm": 26.562755268216712, "learning_rate": 9.737795013388277e-06, "loss": 0.6485, "step": 10047 }, { "epoch": 1.57, "grad_norm": 26.416795483710313, "learning_rate": 9.736109197959693e-06, "loss": 0.6672, "step": 10048 }, { "epoch": 1.57, "grad_norm": 20.315048663829586, "learning_rate": 9.734423390036011e-06, "loss": 0.5546, "step": 10049 }, { "epoch": 1.57, "grad_norm": 19.861351698157037, "learning_rate": 9.732737589665171e-06, "loss": 0.6097, "step": 10050 }, { "epoch": 1.57, "grad_norm": 18.340926464405918, "learning_rate": 9.731051796895116e-06, "loss": 0.55, "step": 10051 }, { "epoch": 1.57, "grad_norm": 24.10408875257175, "learning_rate": 9.729366011773793e-06, "loss": 0.6593, "step": 10052 }, { "epoch": 1.57, "grad_norm": 15.62370364724231, "learning_rate": 9.727680234349138e-06, "loss": 0.5941, "step": 10053 }, { "epoch": 1.57, "grad_norm": 15.556889430353527, "learning_rate": 9.725994464669103e-06, "loss": 0.592, "step": 10054 }, { "epoch": 1.57, "grad_norm": 24.830167263869914, "learning_rate": 9.724308702781625e-06, "loss": 0.617, "step": 10055 }, { "epoch": 1.57, "grad_norm": 22.948813933028337, "learning_rate": 9.722622948734644e-06, "loss": 0.5507, "step": 10056 }, { "epoch": 1.57, "grad_norm": 17.237573565261453, "learning_rate": 9.720937202576101e-06, "loss": 0.579, "step": 10057 }, { "epoch": 1.57, "grad_norm": 14.538579105661437, "learning_rate": 9.719251464353944e-06, "loss": 0.5689, "step": 10058 }, { "epoch": 1.57, "grad_norm": 19.259593188770477, "learning_rate": 9.717565734116114e-06, "loss": 0.5896, "step": 10059 }, { "epoch": 1.57, "grad_norm": 13.278848890750854, "learning_rate": 9.715880011910545e-06, "loss": 0.5417, "step": 10060 }, { "epoch": 1.57, "grad_norm": 13.032823222949075, "learning_rate": 9.714194297785184e-06, "loss": 0.5356, "step": 10061 }, { "epoch": 1.57, "grad_norm": 18.800956271110064, "learning_rate": 9.71250859178797e-06, "loss": 0.6126, "step": 10062 }, { "epoch": 1.57, "grad_norm": 27.2278365407668, "learning_rate": 9.710822893966845e-06, "loss": 0.6029, "step": 10063 }, { "epoch": 1.57, "grad_norm": 15.88481036378908, "learning_rate": 9.709137204369746e-06, "loss": 0.6065, "step": 10064 }, { "epoch": 1.57, "grad_norm": 19.89216972459948, "learning_rate": 9.707451523044614e-06, "loss": 0.6157, "step": 10065 }, { "epoch": 1.57, "grad_norm": 19.80632905571142, "learning_rate": 9.705765850039388e-06, "loss": 0.5651, "step": 10066 }, { "epoch": 1.57, "grad_norm": 17.021337088967, "learning_rate": 9.704080185402012e-06, "loss": 0.5574, "step": 10067 }, { "epoch": 1.57, "grad_norm": 20.36421736968079, "learning_rate": 9.702394529180424e-06, "loss": 0.5428, "step": 10068 }, { "epoch": 1.57, "grad_norm": 14.457185828416597, "learning_rate": 9.70070888142256e-06, "loss": 0.5958, "step": 10069 }, { "epoch": 1.57, "grad_norm": 18.866692643557005, "learning_rate": 9.699023242176358e-06, "loss": 0.5505, "step": 10070 }, { "epoch": 1.57, "grad_norm": 12.840051949893022, "learning_rate": 9.69733761148976e-06, "loss": 0.5661, "step": 10071 }, { "epoch": 1.57, "grad_norm": 23.585973973986793, "learning_rate": 9.695651989410702e-06, "loss": 0.6339, "step": 10072 }, { "epoch": 1.57, "grad_norm": 25.877395682531404, "learning_rate": 9.693966375987128e-06, "loss": 0.5814, "step": 10073 }, { "epoch": 1.57, "grad_norm": 12.72589435134867, "learning_rate": 9.692280771266969e-06, "loss": 0.503, "step": 10074 }, { "epoch": 1.57, "grad_norm": 16.221366208720603, "learning_rate": 9.690595175298164e-06, "loss": 0.4694, "step": 10075 }, { "epoch": 1.57, "grad_norm": 18.782637790812743, "learning_rate": 9.68890958812865e-06, "loss": 0.577, "step": 10076 }, { "epoch": 1.57, "grad_norm": 29.370909077281556, "learning_rate": 9.687224009806366e-06, "loss": 0.7095, "step": 10077 }, { "epoch": 1.57, "grad_norm": 14.02968395707663, "learning_rate": 9.685538440379253e-06, "loss": 0.5078, "step": 10078 }, { "epoch": 1.57, "grad_norm": 15.367498899872174, "learning_rate": 9.683852879895236e-06, "loss": 0.5463, "step": 10079 }, { "epoch": 1.57, "grad_norm": 17.613962039747314, "learning_rate": 9.682167328402261e-06, "loss": 0.4927, "step": 10080 }, { "epoch": 1.57, "grad_norm": 16.779577303303533, "learning_rate": 9.680481785948263e-06, "loss": 0.551, "step": 10081 }, { "epoch": 1.57, "grad_norm": 16.584457012445984, "learning_rate": 9.678796252581171e-06, "loss": 0.6895, "step": 10082 }, { "epoch": 1.57, "grad_norm": 17.44529614103979, "learning_rate": 9.677110728348932e-06, "loss": 0.6271, "step": 10083 }, { "epoch": 1.58, "grad_norm": 26.204992829401718, "learning_rate": 9.675425213299471e-06, "loss": 0.6812, "step": 10084 }, { "epoch": 1.58, "grad_norm": 25.054014717301907, "learning_rate": 9.673739707480727e-06, "loss": 0.6447, "step": 10085 }, { "epoch": 1.58, "grad_norm": 14.26093790209592, "learning_rate": 9.672054210940638e-06, "loss": 0.564, "step": 10086 }, { "epoch": 1.58, "grad_norm": 23.031045110184525, "learning_rate": 9.670368723727131e-06, "loss": 0.5727, "step": 10087 }, { "epoch": 1.58, "grad_norm": 21.579104807331625, "learning_rate": 9.66868324588815e-06, "loss": 0.5525, "step": 10088 }, { "epoch": 1.58, "grad_norm": 14.635028514488914, "learning_rate": 9.66699777747162e-06, "loss": 0.5841, "step": 10089 }, { "epoch": 1.58, "grad_norm": 25.389476404473285, "learning_rate": 9.665312318525478e-06, "loss": 0.5501, "step": 10090 }, { "epoch": 1.58, "grad_norm": 16.356013336116614, "learning_rate": 9.663626869097657e-06, "loss": 0.6834, "step": 10091 }, { "epoch": 1.58, "grad_norm": 22.067622225700482, "learning_rate": 9.661941429236094e-06, "loss": 0.6668, "step": 10092 }, { "epoch": 1.58, "grad_norm": 16.438632758348085, "learning_rate": 9.66025599898872e-06, "loss": 0.5709, "step": 10093 }, { "epoch": 1.58, "grad_norm": 24.151288601924566, "learning_rate": 9.658570578403465e-06, "loss": 0.5981, "step": 10094 }, { "epoch": 1.58, "grad_norm": 16.026308965783524, "learning_rate": 9.656885167528261e-06, "loss": 0.5739, "step": 10095 }, { "epoch": 1.58, "grad_norm": 23.103073550600058, "learning_rate": 9.655199766411044e-06, "loss": 0.6453, "step": 10096 }, { "epoch": 1.58, "grad_norm": 16.13750904108484, "learning_rate": 9.653514375099742e-06, "loss": 0.5524, "step": 10097 }, { "epoch": 1.58, "grad_norm": 16.252009208884083, "learning_rate": 9.651828993642293e-06, "loss": 0.6176, "step": 10098 }, { "epoch": 1.58, "grad_norm": 15.69996785979489, "learning_rate": 9.650143622086621e-06, "loss": 0.5696, "step": 10099 }, { "epoch": 1.58, "grad_norm": 18.49341584530932, "learning_rate": 9.64845826048066e-06, "loss": 0.6005, "step": 10100 }, { "epoch": 1.58, "grad_norm": 14.68649876027647, "learning_rate": 9.646772908872338e-06, "loss": 0.4704, "step": 10101 }, { "epoch": 1.58, "grad_norm": 26.16287216606658, "learning_rate": 9.64508756730959e-06, "loss": 0.6085, "step": 10102 }, { "epoch": 1.58, "grad_norm": 15.04830133901125, "learning_rate": 9.643402235840346e-06, "loss": 0.4905, "step": 10103 }, { "epoch": 1.58, "grad_norm": 19.06883674265519, "learning_rate": 9.641716914512532e-06, "loss": 0.55, "step": 10104 }, { "epoch": 1.58, "grad_norm": 17.27897008856571, "learning_rate": 9.640031603374078e-06, "loss": 0.5096, "step": 10105 }, { "epoch": 1.58, "grad_norm": 15.226691759753798, "learning_rate": 9.638346302472916e-06, "loss": 0.4599, "step": 10106 }, { "epoch": 1.58, "grad_norm": 15.126980546467149, "learning_rate": 9.636661011856971e-06, "loss": 0.543, "step": 10107 }, { "epoch": 1.58, "grad_norm": 18.663240713244512, "learning_rate": 9.63497573157418e-06, "loss": 0.5798, "step": 10108 }, { "epoch": 1.58, "grad_norm": 14.931403170806357, "learning_rate": 9.633290461672463e-06, "loss": 0.6677, "step": 10109 }, { "epoch": 1.58, "grad_norm": 17.958064945119073, "learning_rate": 9.631605202199748e-06, "loss": 0.619, "step": 10110 }, { "epoch": 1.58, "grad_norm": 19.973811260379875, "learning_rate": 9.629919953203966e-06, "loss": 0.5956, "step": 10111 }, { "epoch": 1.58, "grad_norm": 23.270058285730773, "learning_rate": 9.628234714733048e-06, "loss": 0.5969, "step": 10112 }, { "epoch": 1.58, "grad_norm": 16.911981900015906, "learning_rate": 9.626549486834916e-06, "loss": 0.6023, "step": 10113 }, { "epoch": 1.58, "grad_norm": 19.619658509364537, "learning_rate": 9.624864269557495e-06, "loss": 0.6119, "step": 10114 }, { "epoch": 1.58, "grad_norm": 26.55474463117469, "learning_rate": 9.623179062948716e-06, "loss": 0.5962, "step": 10115 }, { "epoch": 1.58, "grad_norm": 16.891516908957403, "learning_rate": 9.621493867056505e-06, "loss": 0.556, "step": 10116 }, { "epoch": 1.58, "grad_norm": 16.622133667595797, "learning_rate": 9.619808681928788e-06, "loss": 0.4833, "step": 10117 }, { "epoch": 1.58, "grad_norm": 30.217225328225783, "learning_rate": 9.618123507613487e-06, "loss": 0.5976, "step": 10118 }, { "epoch": 1.58, "grad_norm": 19.319260472563556, "learning_rate": 9.616438344158533e-06, "loss": 0.6221, "step": 10119 }, { "epoch": 1.58, "grad_norm": 15.064963212064075, "learning_rate": 9.614753191611846e-06, "loss": 0.4999, "step": 10120 }, { "epoch": 1.58, "grad_norm": 17.670909195385725, "learning_rate": 9.613068050021353e-06, "loss": 0.605, "step": 10121 }, { "epoch": 1.58, "grad_norm": 28.09931893999679, "learning_rate": 9.61138291943498e-06, "loss": 0.6798, "step": 10122 }, { "epoch": 1.58, "grad_norm": 21.721724371359706, "learning_rate": 9.609697799900647e-06, "loss": 0.6061, "step": 10123 }, { "epoch": 1.58, "grad_norm": 14.680934739331123, "learning_rate": 9.608012691466281e-06, "loss": 0.5393, "step": 10124 }, { "epoch": 1.58, "grad_norm": 22.99617350340761, "learning_rate": 9.606327594179806e-06, "loss": 0.5694, "step": 10125 }, { "epoch": 1.58, "grad_norm": 15.884327146865957, "learning_rate": 9.604642508089142e-06, "loss": 0.6111, "step": 10126 }, { "epoch": 1.58, "grad_norm": 26.003422813282558, "learning_rate": 9.602957433242218e-06, "loss": 0.6758, "step": 10127 }, { "epoch": 1.58, "grad_norm": 22.745249444803786, "learning_rate": 9.601272369686948e-06, "loss": 0.6849, "step": 10128 }, { "epoch": 1.58, "grad_norm": 13.502951511291759, "learning_rate": 9.599587317471259e-06, "loss": 0.5306, "step": 10129 }, { "epoch": 1.58, "grad_norm": 18.108790051654342, "learning_rate": 9.597902276643074e-06, "loss": 0.4684, "step": 10130 }, { "epoch": 1.58, "grad_norm": 17.67386112216378, "learning_rate": 9.59621724725031e-06, "loss": 0.5954, "step": 10131 }, { "epoch": 1.58, "grad_norm": 27.41976670422977, "learning_rate": 9.594532229340898e-06, "loss": 0.5848, "step": 10132 }, { "epoch": 1.58, "grad_norm": 13.702352275831599, "learning_rate": 9.592847222962744e-06, "loss": 0.543, "step": 10133 }, { "epoch": 1.58, "grad_norm": 18.863435959077865, "learning_rate": 9.591162228163781e-06, "loss": 0.6405, "step": 10134 }, { "epoch": 1.58, "grad_norm": 15.884657693043843, "learning_rate": 9.589477244991924e-06, "loss": 0.526, "step": 10135 }, { "epoch": 1.58, "grad_norm": 21.56997672941892, "learning_rate": 9.587792273495095e-06, "loss": 0.6221, "step": 10136 }, { "epoch": 1.58, "grad_norm": 19.7318703307114, "learning_rate": 9.586107313721214e-06, "loss": 0.5262, "step": 10137 }, { "epoch": 1.58, "grad_norm": 24.19203957877898, "learning_rate": 9.584422365718197e-06, "loss": 0.61, "step": 10138 }, { "epoch": 1.58, "grad_norm": 19.729414635116736, "learning_rate": 9.582737429533964e-06, "loss": 0.5727, "step": 10139 }, { "epoch": 1.58, "grad_norm": 25.388957499401354, "learning_rate": 9.581052505216434e-06, "loss": 0.5099, "step": 10140 }, { "epoch": 1.58, "grad_norm": 18.947607268739354, "learning_rate": 9.579367592813526e-06, "loss": 0.5567, "step": 10141 }, { "epoch": 1.58, "grad_norm": 13.280214226668772, "learning_rate": 9.577682692373161e-06, "loss": 0.5128, "step": 10142 }, { "epoch": 1.58, "grad_norm": 20.638860567901208, "learning_rate": 9.575997803943251e-06, "loss": 0.6247, "step": 10143 }, { "epoch": 1.58, "grad_norm": 18.800169965538153, "learning_rate": 9.574312927571717e-06, "loss": 0.6513, "step": 10144 }, { "epoch": 1.58, "grad_norm": 23.66225216641796, "learning_rate": 9.572628063306469e-06, "loss": 0.5887, "step": 10145 }, { "epoch": 1.58, "grad_norm": 26.067352312053234, "learning_rate": 9.570943211195434e-06, "loss": 0.6061, "step": 10146 }, { "epoch": 1.58, "grad_norm": 18.45671744872066, "learning_rate": 9.569258371286523e-06, "loss": 0.5249, "step": 10147 }, { "epoch": 1.59, "grad_norm": 16.880283281002388, "learning_rate": 9.567573543627649e-06, "loss": 0.5792, "step": 10148 }, { "epoch": 1.59, "grad_norm": 30.50596970329945, "learning_rate": 9.565888728266732e-06, "loss": 0.6021, "step": 10149 }, { "epoch": 1.59, "grad_norm": 21.981365633669338, "learning_rate": 9.564203925251685e-06, "loss": 0.6029, "step": 10150 }, { "epoch": 1.59, "grad_norm": 15.153166468078364, "learning_rate": 9.562519134630423e-06, "loss": 0.5361, "step": 10151 }, { "epoch": 1.59, "grad_norm": 19.73050006055113, "learning_rate": 9.560834356450864e-06, "loss": 0.614, "step": 10152 }, { "epoch": 1.59, "grad_norm": 18.54892938270337, "learning_rate": 9.559149590760917e-06, "loss": 0.6027, "step": 10153 }, { "epoch": 1.59, "grad_norm": 21.098824253816073, "learning_rate": 9.557464837608496e-06, "loss": 0.6092, "step": 10154 }, { "epoch": 1.59, "grad_norm": 18.676721788558627, "learning_rate": 9.555780097041517e-06, "loss": 0.6006, "step": 10155 }, { "epoch": 1.59, "grad_norm": 20.382317378740684, "learning_rate": 9.554095369107892e-06, "loss": 0.592, "step": 10156 }, { "epoch": 1.59, "grad_norm": 21.335653481955433, "learning_rate": 9.55241065385554e-06, "loss": 0.6207, "step": 10157 }, { "epoch": 1.59, "grad_norm": 19.10824624515318, "learning_rate": 9.55072595133236e-06, "loss": 0.5659, "step": 10158 }, { "epoch": 1.59, "grad_norm": 18.221329543216243, "learning_rate": 9.549041261586273e-06, "loss": 0.5967, "step": 10159 }, { "epoch": 1.59, "grad_norm": 19.922817758430813, "learning_rate": 9.547356584665185e-06, "loss": 0.6424, "step": 10160 }, { "epoch": 1.59, "grad_norm": 22.874842735698397, "learning_rate": 9.545671920617018e-06, "loss": 0.5457, "step": 10161 }, { "epoch": 1.59, "grad_norm": 15.082334397319096, "learning_rate": 9.543987269489673e-06, "loss": 0.5322, "step": 10162 }, { "epoch": 1.59, "grad_norm": 17.206464654208467, "learning_rate": 9.542302631331063e-06, "loss": 0.5889, "step": 10163 }, { "epoch": 1.59, "grad_norm": 17.305205657727395, "learning_rate": 9.540618006189096e-06, "loss": 0.5444, "step": 10164 }, { "epoch": 1.59, "grad_norm": 18.114067945740302, "learning_rate": 9.538933394111687e-06, "loss": 0.6235, "step": 10165 }, { "epoch": 1.59, "grad_norm": 21.498871242063327, "learning_rate": 9.537248795146745e-06, "loss": 0.5464, "step": 10166 }, { "epoch": 1.59, "grad_norm": 15.411073713948925, "learning_rate": 9.535564209342172e-06, "loss": 0.5993, "step": 10167 }, { "epoch": 1.59, "grad_norm": 20.395969403728373, "learning_rate": 9.533879636745883e-06, "loss": 0.704, "step": 10168 }, { "epoch": 1.59, "grad_norm": 18.301484906847854, "learning_rate": 9.532195077405784e-06, "loss": 0.5667, "step": 10169 }, { "epoch": 1.59, "grad_norm": 12.219251545566367, "learning_rate": 9.53051053136978e-06, "loss": 0.5725, "step": 10170 }, { "epoch": 1.59, "grad_norm": 23.52548695153024, "learning_rate": 9.528825998685788e-06, "loss": 0.6348, "step": 10171 }, { "epoch": 1.59, "grad_norm": 28.5215812639141, "learning_rate": 9.527141479401708e-06, "loss": 0.6318, "step": 10172 }, { "epoch": 1.59, "grad_norm": 20.428166581208274, "learning_rate": 9.525456973565443e-06, "loss": 0.5895, "step": 10173 }, { "epoch": 1.59, "grad_norm": 22.104338422728773, "learning_rate": 9.523772481224908e-06, "loss": 0.5357, "step": 10174 }, { "epoch": 1.59, "grad_norm": 23.521313544488233, "learning_rate": 9.522088002428003e-06, "loss": 0.5764, "step": 10175 }, { "epoch": 1.59, "grad_norm": 19.492518070911466, "learning_rate": 9.520403537222642e-06, "loss": 0.6186, "step": 10176 }, { "epoch": 1.59, "grad_norm": 16.469337373280638, "learning_rate": 9.518719085656716e-06, "loss": 0.5808, "step": 10177 }, { "epoch": 1.59, "grad_norm": 20.01032854613012, "learning_rate": 9.51703464777814e-06, "loss": 0.5827, "step": 10178 }, { "epoch": 1.59, "grad_norm": 17.454706514809335, "learning_rate": 9.515350223634815e-06, "loss": 0.5495, "step": 10179 }, { "epoch": 1.59, "grad_norm": 11.2733111774296, "learning_rate": 9.513665813274647e-06, "loss": 0.58, "step": 10180 }, { "epoch": 1.59, "grad_norm": 22.128237953431082, "learning_rate": 9.511981416745542e-06, "loss": 0.5409, "step": 10181 }, { "epoch": 1.59, "grad_norm": 14.02137841432915, "learning_rate": 9.510297034095398e-06, "loss": 0.5669, "step": 10182 }, { "epoch": 1.59, "grad_norm": 21.796168025320355, "learning_rate": 9.508612665372117e-06, "loss": 0.595, "step": 10183 }, { "epoch": 1.59, "grad_norm": 16.964037031263683, "learning_rate": 9.506928310623608e-06, "loss": 0.5848, "step": 10184 }, { "epoch": 1.59, "grad_norm": 24.160202415810765, "learning_rate": 9.505243969897766e-06, "loss": 0.5386, "step": 10185 }, { "epoch": 1.59, "grad_norm": 21.004701400328855, "learning_rate": 9.5035596432425e-06, "loss": 0.5449, "step": 10186 }, { "epoch": 1.59, "grad_norm": 19.47876873289876, "learning_rate": 9.501875330705706e-06, "loss": 0.5268, "step": 10187 }, { "epoch": 1.59, "grad_norm": 17.744135956666508, "learning_rate": 9.500191032335286e-06, "loss": 0.5836, "step": 10188 }, { "epoch": 1.59, "grad_norm": 24.923679271546753, "learning_rate": 9.498506748179137e-06, "loss": 0.5676, "step": 10189 }, { "epoch": 1.59, "grad_norm": 14.577169493325691, "learning_rate": 9.496822478285167e-06, "loss": 0.5193, "step": 10190 }, { "epoch": 1.59, "grad_norm": 21.95495695228826, "learning_rate": 9.495138222701273e-06, "loss": 0.6165, "step": 10191 }, { "epoch": 1.59, "grad_norm": 23.45231711942303, "learning_rate": 9.493453981475348e-06, "loss": 0.5866, "step": 10192 }, { "epoch": 1.59, "grad_norm": 12.599289539531185, "learning_rate": 9.491769754655298e-06, "loss": 0.5462, "step": 10193 }, { "epoch": 1.59, "grad_norm": 18.851061260612095, "learning_rate": 9.490085542289016e-06, "loss": 0.6032, "step": 10194 }, { "epoch": 1.59, "grad_norm": 13.497221020491342, "learning_rate": 9.488401344424404e-06, "loss": 0.5381, "step": 10195 }, { "epoch": 1.59, "grad_norm": 15.088175688090733, "learning_rate": 9.486717161109363e-06, "loss": 0.5695, "step": 10196 }, { "epoch": 1.59, "grad_norm": 15.392560985232791, "learning_rate": 9.485032992391782e-06, "loss": 0.531, "step": 10197 }, { "epoch": 1.59, "grad_norm": 20.9133322840861, "learning_rate": 9.48334883831956e-06, "loss": 0.5628, "step": 10198 }, { "epoch": 1.59, "grad_norm": 12.150192333934168, "learning_rate": 9.481664698940598e-06, "loss": 0.5566, "step": 10199 }, { "epoch": 1.59, "grad_norm": 21.434743320288796, "learning_rate": 9.479980574302785e-06, "loss": 0.5734, "step": 10200 }, { "epoch": 1.59, "grad_norm": 14.037543672653364, "learning_rate": 9.478296464454028e-06, "loss": 0.5512, "step": 10201 }, { "epoch": 1.59, "grad_norm": 26.748778909427127, "learning_rate": 9.476612369442207e-06, "loss": 0.5786, "step": 10202 }, { "epoch": 1.59, "grad_norm": 13.593833432818476, "learning_rate": 9.474928289315224e-06, "loss": 0.5504, "step": 10203 }, { "epoch": 1.59, "grad_norm": 21.797855319944865, "learning_rate": 9.473244224120974e-06, "loss": 0.6067, "step": 10204 }, { "epoch": 1.59, "grad_norm": 27.362343431585103, "learning_rate": 9.471560173907353e-06, "loss": 0.6348, "step": 10205 }, { "epoch": 1.59, "grad_norm": 23.205063470470282, "learning_rate": 9.469876138722252e-06, "loss": 0.5843, "step": 10206 }, { "epoch": 1.59, "grad_norm": 15.564455273848642, "learning_rate": 9.46819211861356e-06, "loss": 0.5634, "step": 10207 }, { "epoch": 1.59, "grad_norm": 24.979014592798812, "learning_rate": 9.466508113629174e-06, "loss": 0.5812, "step": 10208 }, { "epoch": 1.59, "grad_norm": 17.5927799162375, "learning_rate": 9.464824123816986e-06, "loss": 0.5749, "step": 10209 }, { "epoch": 1.59, "grad_norm": 23.331526192155177, "learning_rate": 9.463140149224888e-06, "loss": 0.6353, "step": 10210 }, { "epoch": 1.59, "grad_norm": 20.18905818317108, "learning_rate": 9.461456189900767e-06, "loss": 0.6544, "step": 10211 }, { "epoch": 1.6, "grad_norm": 17.73927568287242, "learning_rate": 9.459772245892518e-06, "loss": 0.5988, "step": 10212 }, { "epoch": 1.6, "grad_norm": 16.912787478362144, "learning_rate": 9.458088317248032e-06, "loss": 0.5653, "step": 10213 }, { "epoch": 1.6, "grad_norm": 22.484053835122573, "learning_rate": 9.456404404015194e-06, "loss": 0.6067, "step": 10214 }, { "epoch": 1.6, "grad_norm": 18.348222035518095, "learning_rate": 9.454720506241902e-06, "loss": 0.4943, "step": 10215 }, { "epoch": 1.6, "grad_norm": 21.978898508385594, "learning_rate": 9.453036623976036e-06, "loss": 0.59, "step": 10216 }, { "epoch": 1.6, "grad_norm": 20.48585443515506, "learning_rate": 9.451352757265488e-06, "loss": 0.5905, "step": 10217 }, { "epoch": 1.6, "grad_norm": 15.623685701407165, "learning_rate": 9.449668906158149e-06, "loss": 0.5601, "step": 10218 }, { "epoch": 1.6, "grad_norm": 14.59599014381848, "learning_rate": 9.4479850707019e-06, "loss": 0.5843, "step": 10219 }, { "epoch": 1.6, "grad_norm": 10.582674865188112, "learning_rate": 9.446301250944641e-06, "loss": 0.5157, "step": 10220 }, { "epoch": 1.6, "grad_norm": 26.282548043426566, "learning_rate": 9.444617446934244e-06, "loss": 0.5368, "step": 10221 }, { "epoch": 1.6, "grad_norm": 23.083932482277575, "learning_rate": 9.442933658718603e-06, "loss": 0.5931, "step": 10222 }, { "epoch": 1.6, "grad_norm": 12.950197334194463, "learning_rate": 9.4412498863456e-06, "loss": 0.5582, "step": 10223 }, { "epoch": 1.6, "grad_norm": 31.458807319213445, "learning_rate": 9.439566129863126e-06, "loss": 0.6908, "step": 10224 }, { "epoch": 1.6, "grad_norm": 19.418340543595967, "learning_rate": 9.437882389319067e-06, "loss": 0.5632, "step": 10225 }, { "epoch": 1.6, "grad_norm": 14.601004524550662, "learning_rate": 9.436198664761302e-06, "loss": 0.5978, "step": 10226 }, { "epoch": 1.6, "grad_norm": 14.946124442638643, "learning_rate": 9.434514956237714e-06, "loss": 0.5696, "step": 10227 }, { "epoch": 1.6, "grad_norm": 18.578052067031987, "learning_rate": 9.43283126379619e-06, "loss": 0.5703, "step": 10228 }, { "epoch": 1.6, "grad_norm": 21.660376256129847, "learning_rate": 9.431147587484614e-06, "loss": 0.5427, "step": 10229 }, { "epoch": 1.6, "grad_norm": 16.171986702980117, "learning_rate": 9.429463927350872e-06, "loss": 0.5693, "step": 10230 }, { "epoch": 1.6, "grad_norm": 14.131356021402402, "learning_rate": 9.427780283442838e-06, "loss": 0.6187, "step": 10231 }, { "epoch": 1.6, "grad_norm": 18.673093993805384, "learning_rate": 9.4260966558084e-06, "loss": 0.6402, "step": 10232 }, { "epoch": 1.6, "grad_norm": 20.810651001117048, "learning_rate": 9.424413044495435e-06, "loss": 0.6816, "step": 10233 }, { "epoch": 1.6, "grad_norm": 17.537029502815066, "learning_rate": 9.422729449551828e-06, "loss": 0.5753, "step": 10234 }, { "epoch": 1.6, "grad_norm": 16.174861105675564, "learning_rate": 9.42104587102546e-06, "loss": 0.5041, "step": 10235 }, { "epoch": 1.6, "grad_norm": 24.18466711211708, "learning_rate": 9.419362308964203e-06, "loss": 0.6166, "step": 10236 }, { "epoch": 1.6, "grad_norm": 21.901788555175294, "learning_rate": 9.417678763415948e-06, "loss": 0.5414, "step": 10237 }, { "epoch": 1.6, "grad_norm": 17.198155662471024, "learning_rate": 9.415995234428563e-06, "loss": 0.6062, "step": 10238 }, { "epoch": 1.6, "grad_norm": 22.553335830466292, "learning_rate": 9.414311722049935e-06, "loss": 0.5701, "step": 10239 }, { "epoch": 1.6, "grad_norm": 19.9277178868868, "learning_rate": 9.41262822632794e-06, "loss": 0.6295, "step": 10240 }, { "epoch": 1.6, "grad_norm": 26.466209562185433, "learning_rate": 9.410944747310454e-06, "loss": 0.5657, "step": 10241 }, { "epoch": 1.6, "grad_norm": 14.962371337647367, "learning_rate": 9.409261285045352e-06, "loss": 0.5876, "step": 10242 }, { "epoch": 1.6, "grad_norm": 19.82596976378527, "learning_rate": 9.407577839580516e-06, "loss": 0.5939, "step": 10243 }, { "epoch": 1.6, "grad_norm": 22.59301528553886, "learning_rate": 9.405894410963817e-06, "loss": 0.622, "step": 10244 }, { "epoch": 1.6, "grad_norm": 15.98443494060237, "learning_rate": 9.404210999243141e-06, "loss": 0.6171, "step": 10245 }, { "epoch": 1.6, "grad_norm": 16.873800868238796, "learning_rate": 9.402527604466347e-06, "loss": 0.5381, "step": 10246 }, { "epoch": 1.6, "grad_norm": 13.757366974704107, "learning_rate": 9.400844226681324e-06, "loss": 0.5755, "step": 10247 }, { "epoch": 1.6, "grad_norm": 23.078581079742847, "learning_rate": 9.399160865935936e-06, "loss": 0.5406, "step": 10248 }, { "epoch": 1.6, "grad_norm": 20.536315443335315, "learning_rate": 9.397477522278064e-06, "loss": 0.5606, "step": 10249 }, { "epoch": 1.6, "grad_norm": 16.518486491893302, "learning_rate": 9.395794195755581e-06, "loss": 0.6363, "step": 10250 }, { "epoch": 1.6, "grad_norm": 17.10358513276388, "learning_rate": 9.394110886416355e-06, "loss": 0.6003, "step": 10251 }, { "epoch": 1.6, "grad_norm": 15.416257672963312, "learning_rate": 9.39242759430826e-06, "loss": 0.526, "step": 10252 }, { "epoch": 1.6, "grad_norm": 24.113152665994722, "learning_rate": 9.390744319479171e-06, "loss": 0.6683, "step": 10253 }, { "epoch": 1.6, "grad_norm": 15.690606755692588, "learning_rate": 9.389061061976958e-06, "loss": 0.5294, "step": 10254 }, { "epoch": 1.6, "grad_norm": 35.46687030432764, "learning_rate": 9.387377821849489e-06, "loss": 0.7158, "step": 10255 }, { "epoch": 1.6, "grad_norm": 28.0977567736958, "learning_rate": 9.385694599144636e-06, "loss": 0.6548, "step": 10256 }, { "epoch": 1.6, "grad_norm": 17.389204608186773, "learning_rate": 9.38401139391027e-06, "loss": 0.5865, "step": 10257 }, { "epoch": 1.6, "grad_norm": 16.95918528178506, "learning_rate": 9.382328206194259e-06, "loss": 0.6258, "step": 10258 }, { "epoch": 1.6, "grad_norm": 15.568563356356565, "learning_rate": 9.380645036044473e-06, "loss": 0.502, "step": 10259 }, { "epoch": 1.6, "grad_norm": 17.88837911135321, "learning_rate": 9.37896188350878e-06, "loss": 0.6104, "step": 10260 }, { "epoch": 1.6, "grad_norm": 22.81118628233423, "learning_rate": 9.377278748635046e-06, "loss": 0.6765, "step": 10261 }, { "epoch": 1.6, "grad_norm": 19.87957169964455, "learning_rate": 9.375595631471143e-06, "loss": 0.578, "step": 10262 }, { "epoch": 1.6, "grad_norm": 15.256536082207402, "learning_rate": 9.373912532064931e-06, "loss": 0.5445, "step": 10263 }, { "epoch": 1.6, "grad_norm": 21.100147125699564, "learning_rate": 9.372229450464287e-06, "loss": 0.579, "step": 10264 }, { "epoch": 1.6, "grad_norm": 12.915311870614163, "learning_rate": 9.370546386717065e-06, "loss": 0.5259, "step": 10265 }, { "epoch": 1.6, "grad_norm": 31.69585569337966, "learning_rate": 9.368863340871137e-06, "loss": 0.5983, "step": 10266 }, { "epoch": 1.6, "grad_norm": 16.1828908723917, "learning_rate": 9.367180312974364e-06, "loss": 0.5635, "step": 10267 }, { "epoch": 1.6, "grad_norm": 12.697943695885563, "learning_rate": 9.365497303074615e-06, "loss": 0.5772, "step": 10268 }, { "epoch": 1.6, "grad_norm": 24.679322525325297, "learning_rate": 9.363814311219754e-06, "loss": 0.6061, "step": 10269 }, { "epoch": 1.6, "grad_norm": 13.954191442661077, "learning_rate": 9.36213133745764e-06, "loss": 0.5482, "step": 10270 }, { "epoch": 1.6, "grad_norm": 14.922208037579464, "learning_rate": 9.360448381836134e-06, "loss": 0.5827, "step": 10271 }, { "epoch": 1.6, "grad_norm": 23.75693757472985, "learning_rate": 9.358765444403107e-06, "loss": 0.7006, "step": 10272 }, { "epoch": 1.6, "grad_norm": 18.04679844827641, "learning_rate": 9.35708252520641e-06, "loss": 0.548, "step": 10273 }, { "epoch": 1.6, "grad_norm": 22.92478961980648, "learning_rate": 9.355399624293917e-06, "loss": 0.5839, "step": 10274 }, { "epoch": 1.6, "grad_norm": 16.703282960371414, "learning_rate": 9.353716741713477e-06, "loss": 0.6449, "step": 10275 }, { "epoch": 1.61, "grad_norm": 17.442910593222667, "learning_rate": 9.352033877512957e-06, "loss": 0.5102, "step": 10276 }, { "epoch": 1.61, "grad_norm": 15.887638379641604, "learning_rate": 9.35035103174021e-06, "loss": 0.6407, "step": 10277 }, { "epoch": 1.61, "grad_norm": 18.359682947376314, "learning_rate": 9.348668204443103e-06, "loss": 0.5838, "step": 10278 }, { "epoch": 1.61, "grad_norm": 13.313177205712227, "learning_rate": 9.346985395669494e-06, "loss": 0.5284, "step": 10279 }, { "epoch": 1.61, "grad_norm": 21.927042933145184, "learning_rate": 9.345302605467232e-06, "loss": 0.6332, "step": 10280 }, { "epoch": 1.61, "grad_norm": 21.39632241741886, "learning_rate": 9.343619833884186e-06, "loss": 0.5846, "step": 10281 }, { "epoch": 1.61, "grad_norm": 20.727835286498337, "learning_rate": 9.341937080968204e-06, "loss": 0.5179, "step": 10282 }, { "epoch": 1.61, "grad_norm": 21.1071667395017, "learning_rate": 9.340254346767149e-06, "loss": 0.6129, "step": 10283 }, { "epoch": 1.61, "grad_norm": 20.245811454546278, "learning_rate": 9.338571631328877e-06, "loss": 0.5398, "step": 10284 }, { "epoch": 1.61, "grad_norm": 17.270080420479953, "learning_rate": 9.336888934701238e-06, "loss": 0.6099, "step": 10285 }, { "epoch": 1.61, "grad_norm": 19.16792531977402, "learning_rate": 9.335206256932088e-06, "loss": 0.6419, "step": 10286 }, { "epoch": 1.61, "grad_norm": 14.657754697808723, "learning_rate": 9.333523598069286e-06, "loss": 0.5355, "step": 10287 }, { "epoch": 1.61, "grad_norm": 19.473529822096957, "learning_rate": 9.331840958160679e-06, "loss": 0.6247, "step": 10288 }, { "epoch": 1.61, "grad_norm": 11.271632504055837, "learning_rate": 9.330158337254131e-06, "loss": 0.4567, "step": 10289 }, { "epoch": 1.61, "grad_norm": 28.09334661726697, "learning_rate": 9.328475735397483e-06, "loss": 0.6163, "step": 10290 }, { "epoch": 1.61, "grad_norm": 18.25708516173471, "learning_rate": 9.326793152638594e-06, "loss": 0.5547, "step": 10291 }, { "epoch": 1.61, "grad_norm": 15.876319783460378, "learning_rate": 9.325110589025311e-06, "loss": 0.6222, "step": 10292 }, { "epoch": 1.61, "grad_norm": 18.7870769791939, "learning_rate": 9.323428044605491e-06, "loss": 0.5774, "step": 10293 }, { "epoch": 1.61, "grad_norm": 18.43082015578266, "learning_rate": 9.321745519426984e-06, "loss": 0.6114, "step": 10294 }, { "epoch": 1.61, "grad_norm": 18.380068503266173, "learning_rate": 9.320063013537634e-06, "loss": 0.5857, "step": 10295 }, { "epoch": 1.61, "grad_norm": 20.748713204256827, "learning_rate": 9.318380526985293e-06, "loss": 0.6269, "step": 10296 }, { "epoch": 1.61, "grad_norm": 22.65661006641371, "learning_rate": 9.316698059817814e-06, "loss": 0.5935, "step": 10297 }, { "epoch": 1.61, "grad_norm": 19.79811898427557, "learning_rate": 9.315015612083039e-06, "loss": 0.6588, "step": 10298 }, { "epoch": 1.61, "grad_norm": 21.167883736073115, "learning_rate": 9.313333183828824e-06, "loss": 0.6254, "step": 10299 }, { "epoch": 1.61, "grad_norm": 20.774981369258086, "learning_rate": 9.31165077510301e-06, "loss": 0.5082, "step": 10300 }, { "epoch": 1.61, "grad_norm": 16.732890730213768, "learning_rate": 9.309968385953445e-06, "loss": 0.6986, "step": 10301 }, { "epoch": 1.61, "grad_norm": 21.58489678773787, "learning_rate": 9.308286016427974e-06, "loss": 0.5383, "step": 10302 }, { "epoch": 1.61, "grad_norm": 18.716575974740497, "learning_rate": 9.306603666574446e-06, "loss": 0.5544, "step": 10303 }, { "epoch": 1.61, "grad_norm": 15.021266107854519, "learning_rate": 9.304921336440704e-06, "loss": 0.5209, "step": 10304 }, { "epoch": 1.61, "grad_norm": 19.063998376866504, "learning_rate": 9.303239026074589e-06, "loss": 0.5744, "step": 10305 }, { "epoch": 1.61, "grad_norm": 28.43306292145674, "learning_rate": 9.301556735523952e-06, "loss": 0.6, "step": 10306 }, { "epoch": 1.61, "grad_norm": 21.218778917099094, "learning_rate": 9.29987446483663e-06, "loss": 0.5853, "step": 10307 }, { "epoch": 1.61, "grad_norm": 12.283182325207553, "learning_rate": 9.298192214060476e-06, "loss": 0.5186, "step": 10308 }, { "epoch": 1.61, "grad_norm": 26.22087330342056, "learning_rate": 9.296509983243319e-06, "loss": 0.6404, "step": 10309 }, { "epoch": 1.61, "grad_norm": 20.1578283098604, "learning_rate": 9.294827772433006e-06, "loss": 0.5412, "step": 10310 }, { "epoch": 1.61, "grad_norm": 18.916514023106846, "learning_rate": 9.293145581677378e-06, "loss": 0.5919, "step": 10311 }, { "epoch": 1.61, "grad_norm": 14.71944951867443, "learning_rate": 9.291463411024279e-06, "loss": 0.648, "step": 10312 }, { "epoch": 1.61, "grad_norm": 21.223280638580093, "learning_rate": 9.289781260521548e-06, "loss": 0.5693, "step": 10313 }, { "epoch": 1.61, "grad_norm": 13.29924722303305, "learning_rate": 9.288099130217018e-06, "loss": 0.5729, "step": 10314 }, { "epoch": 1.61, "grad_norm": 20.990288877855676, "learning_rate": 9.286417020158533e-06, "loss": 0.6412, "step": 10315 }, { "epoch": 1.61, "grad_norm": 12.722236580115561, "learning_rate": 9.284734930393931e-06, "loss": 0.5158, "step": 10316 }, { "epoch": 1.61, "grad_norm": 27.177101535810518, "learning_rate": 9.283052860971049e-06, "loss": 0.6303, "step": 10317 }, { "epoch": 1.61, "grad_norm": 17.53566191274117, "learning_rate": 9.281370811937726e-06, "loss": 0.5773, "step": 10318 }, { "epoch": 1.61, "grad_norm": 24.816391767896608, "learning_rate": 9.279688783341796e-06, "loss": 0.5675, "step": 10319 }, { "epoch": 1.61, "grad_norm": 20.432124444753892, "learning_rate": 9.278006775231097e-06, "loss": 0.6174, "step": 10320 }, { "epoch": 1.61, "grad_norm": 13.286689421389843, "learning_rate": 9.276324787653458e-06, "loss": 0.5627, "step": 10321 }, { "epoch": 1.61, "grad_norm": 26.682728795398713, "learning_rate": 9.274642820656722e-06, "loss": 0.6473, "step": 10322 }, { "epoch": 1.61, "grad_norm": 23.291114155622914, "learning_rate": 9.272960874288722e-06, "loss": 0.6316, "step": 10323 }, { "epoch": 1.61, "grad_norm": 17.73245335851438, "learning_rate": 9.271278948597285e-06, "loss": 0.5468, "step": 10324 }, { "epoch": 1.61, "grad_norm": 19.81599449829306, "learning_rate": 9.269597043630252e-06, "loss": 0.6296, "step": 10325 }, { "epoch": 1.61, "grad_norm": 20.962175262337958, "learning_rate": 9.267915159435447e-06, "loss": 0.5747, "step": 10326 }, { "epoch": 1.61, "grad_norm": 20.38652485369521, "learning_rate": 9.266233296060712e-06, "loss": 0.6067, "step": 10327 }, { "epoch": 1.61, "grad_norm": 13.194172448009406, "learning_rate": 9.264551453553872e-06, "loss": 0.6119, "step": 10328 }, { "epoch": 1.61, "grad_norm": 15.653573340854845, "learning_rate": 9.262869631962758e-06, "loss": 0.5876, "step": 10329 }, { "epoch": 1.61, "grad_norm": 19.48438943970978, "learning_rate": 9.261187831335197e-06, "loss": 0.4992, "step": 10330 }, { "epoch": 1.61, "grad_norm": 14.082981925122127, "learning_rate": 9.259506051719025e-06, "loss": 0.6096, "step": 10331 }, { "epoch": 1.61, "grad_norm": 17.502835289291347, "learning_rate": 9.257824293162066e-06, "loss": 0.5047, "step": 10332 }, { "epoch": 1.61, "grad_norm": 6.420905425209333, "learning_rate": 9.256142555712156e-06, "loss": 0.6344, "step": 10333 }, { "epoch": 1.61, "grad_norm": 20.50588945073572, "learning_rate": 9.25446083941711e-06, "loss": 0.6006, "step": 10334 }, { "epoch": 1.61, "grad_norm": 17.39603997529431, "learning_rate": 9.252779144324763e-06, "loss": 0.5787, "step": 10335 }, { "epoch": 1.61, "grad_norm": 22.832102139952735, "learning_rate": 9.251097470482938e-06, "loss": 0.5998, "step": 10336 }, { "epoch": 1.61, "grad_norm": 26.953039387000295, "learning_rate": 9.249415817939465e-06, "loss": 0.5463, "step": 10337 }, { "epoch": 1.61, "grad_norm": 27.634023958526196, "learning_rate": 9.24773418674217e-06, "loss": 0.6105, "step": 10338 }, { "epoch": 1.61, "grad_norm": 18.144275607637432, "learning_rate": 9.246052576938874e-06, "loss": 0.6086, "step": 10339 }, { "epoch": 1.62, "grad_norm": 16.076926761811993, "learning_rate": 9.244370988577396e-06, "loss": 0.593, "step": 10340 }, { "epoch": 1.62, "grad_norm": 16.39934634324917, "learning_rate": 9.24268942170557e-06, "loss": 0.5132, "step": 10341 }, { "epoch": 1.62, "grad_norm": 22.67235096524518, "learning_rate": 9.24100787637121e-06, "loss": 0.5689, "step": 10342 }, { "epoch": 1.62, "grad_norm": 12.510444718569524, "learning_rate": 9.239326352622146e-06, "loss": 0.5262, "step": 10343 }, { "epoch": 1.62, "grad_norm": 26.69376720592254, "learning_rate": 9.237644850506196e-06, "loss": 0.6201, "step": 10344 }, { "epoch": 1.62, "grad_norm": 18.471294593275754, "learning_rate": 9.235963370071176e-06, "loss": 0.6485, "step": 10345 }, { "epoch": 1.62, "grad_norm": 29.448793617415234, "learning_rate": 9.234281911364914e-06, "loss": 0.6395, "step": 10346 }, { "epoch": 1.62, "grad_norm": 19.550272676024864, "learning_rate": 9.232600474435226e-06, "loss": 0.5727, "step": 10347 }, { "epoch": 1.62, "grad_norm": 29.68396072773288, "learning_rate": 9.230919059329934e-06, "loss": 0.6261, "step": 10348 }, { "epoch": 1.62, "grad_norm": 14.850594901922985, "learning_rate": 9.229237666096848e-06, "loss": 0.5576, "step": 10349 }, { "epoch": 1.62, "grad_norm": 16.38601101971303, "learning_rate": 9.227556294783797e-06, "loss": 0.5621, "step": 10350 }, { "epoch": 1.62, "grad_norm": 23.363081351609598, "learning_rate": 9.22587494543859e-06, "loss": 0.6408, "step": 10351 }, { "epoch": 1.62, "grad_norm": 28.623481177396595, "learning_rate": 9.224193618109052e-06, "loss": 0.5417, "step": 10352 }, { "epoch": 1.62, "grad_norm": 23.912896584319487, "learning_rate": 9.222512312842988e-06, "loss": 0.5638, "step": 10353 }, { "epoch": 1.62, "grad_norm": 22.642217769632843, "learning_rate": 9.220831029688222e-06, "loss": 0.6226, "step": 10354 }, { "epoch": 1.62, "grad_norm": 21.777555452694852, "learning_rate": 9.219149768692563e-06, "loss": 0.5332, "step": 10355 }, { "epoch": 1.62, "grad_norm": 19.636761167687975, "learning_rate": 9.217468529903829e-06, "loss": 0.596, "step": 10356 }, { "epoch": 1.62, "grad_norm": 20.28010750823603, "learning_rate": 9.215787313369836e-06, "loss": 0.576, "step": 10357 }, { "epoch": 1.62, "grad_norm": 22.846388950768564, "learning_rate": 9.214106119138388e-06, "loss": 0.5896, "step": 10358 }, { "epoch": 1.62, "grad_norm": 16.684558197454713, "learning_rate": 9.212424947257302e-06, "loss": 0.5528, "step": 10359 }, { "epoch": 1.62, "grad_norm": 16.25708366017856, "learning_rate": 9.210743797774392e-06, "loss": 0.559, "step": 10360 }, { "epoch": 1.62, "grad_norm": 26.2853679756167, "learning_rate": 9.209062670737465e-06, "loss": 0.5536, "step": 10361 }, { "epoch": 1.62, "grad_norm": 14.790140533854814, "learning_rate": 9.207381566194336e-06, "loss": 0.4897, "step": 10362 }, { "epoch": 1.62, "grad_norm": 30.496283108795332, "learning_rate": 9.20570048419281e-06, "loss": 0.5982, "step": 10363 }, { "epoch": 1.62, "grad_norm": 17.592540602423043, "learning_rate": 9.2040194247807e-06, "loss": 0.6106, "step": 10364 }, { "epoch": 1.62, "grad_norm": 21.310805134488394, "learning_rate": 9.202338388005807e-06, "loss": 0.5794, "step": 10365 }, { "epoch": 1.62, "grad_norm": 21.701146314865305, "learning_rate": 9.200657373915946e-06, "loss": 0.5905, "step": 10366 }, { "epoch": 1.62, "grad_norm": 14.291560195112899, "learning_rate": 9.198976382558924e-06, "loss": 0.5828, "step": 10367 }, { "epoch": 1.62, "grad_norm": 12.652736740745688, "learning_rate": 9.19729541398254e-06, "loss": 0.4989, "step": 10368 }, { "epoch": 1.62, "grad_norm": 15.452074973242688, "learning_rate": 9.19561446823461e-06, "loss": 0.5452, "step": 10369 }, { "epoch": 1.62, "grad_norm": 18.520115284540324, "learning_rate": 9.19393354536293e-06, "loss": 0.6126, "step": 10370 }, { "epoch": 1.62, "grad_norm": 13.569610225599869, "learning_rate": 9.19225264541531e-06, "loss": 0.4944, "step": 10371 }, { "epoch": 1.62, "grad_norm": 22.269979223842032, "learning_rate": 9.190571768439554e-06, "loss": 0.6588, "step": 10372 }, { "epoch": 1.62, "grad_norm": 13.070464657766006, "learning_rate": 9.188890914483462e-06, "loss": 0.5684, "step": 10373 }, { "epoch": 1.62, "grad_norm": 22.193578442888327, "learning_rate": 9.187210083594835e-06, "loss": 0.4986, "step": 10374 }, { "epoch": 1.62, "grad_norm": 12.756416496127612, "learning_rate": 9.185529275821481e-06, "loss": 0.4903, "step": 10375 }, { "epoch": 1.62, "grad_norm": 36.45561857070713, "learning_rate": 9.183848491211196e-06, "loss": 0.5849, "step": 10376 }, { "epoch": 1.62, "grad_norm": 20.188243915370844, "learning_rate": 9.182167729811788e-06, "loss": 0.5845, "step": 10377 }, { "epoch": 1.62, "grad_norm": 12.224510569736154, "learning_rate": 9.180486991671046e-06, "loss": 0.4929, "step": 10378 }, { "epoch": 1.62, "grad_norm": 16.828061450813465, "learning_rate": 9.178806276836777e-06, "loss": 0.5227, "step": 10379 }, { "epoch": 1.62, "grad_norm": 12.600475383227142, "learning_rate": 9.177125585356773e-06, "loss": 0.4719, "step": 10380 }, { "epoch": 1.62, "grad_norm": 22.9438112967308, "learning_rate": 9.175444917278839e-06, "loss": 0.5348, "step": 10381 }, { "epoch": 1.62, "grad_norm": 24.027224423666176, "learning_rate": 9.173764272650771e-06, "loss": 0.5896, "step": 10382 }, { "epoch": 1.62, "grad_norm": 24.211070452728517, "learning_rate": 9.172083651520362e-06, "loss": 0.683, "step": 10383 }, { "epoch": 1.62, "grad_norm": 18.378095623142716, "learning_rate": 9.170403053935408e-06, "loss": 0.5744, "step": 10384 }, { "epoch": 1.62, "grad_norm": 19.88163814186779, "learning_rate": 9.168722479943707e-06, "loss": 0.5265, "step": 10385 }, { "epoch": 1.62, "grad_norm": 17.597047180491167, "learning_rate": 9.167041929593051e-06, "loss": 0.5158, "step": 10386 }, { "epoch": 1.62, "grad_norm": 16.541592896320616, "learning_rate": 9.16536140293124e-06, "loss": 0.5669, "step": 10387 }, { "epoch": 1.62, "grad_norm": 20.123741828790738, "learning_rate": 9.16368090000606e-06, "loss": 0.6116, "step": 10388 }, { "epoch": 1.62, "grad_norm": 11.70749623638826, "learning_rate": 9.162000420865303e-06, "loss": 0.5384, "step": 10389 }, { "epoch": 1.62, "grad_norm": 21.061476114470807, "learning_rate": 9.160319965556765e-06, "loss": 0.5703, "step": 10390 }, { "epoch": 1.62, "grad_norm": 16.41892315738222, "learning_rate": 9.158639534128239e-06, "loss": 0.5311, "step": 10391 }, { "epoch": 1.62, "grad_norm": 16.964652011410568, "learning_rate": 9.156959126627512e-06, "loss": 0.5812, "step": 10392 }, { "epoch": 1.62, "grad_norm": 21.382889950618612, "learning_rate": 9.155278743102372e-06, "loss": 0.6956, "step": 10393 }, { "epoch": 1.62, "grad_norm": 15.852998802875378, "learning_rate": 9.15359838360061e-06, "loss": 0.5987, "step": 10394 }, { "epoch": 1.62, "grad_norm": 24.414755959389733, "learning_rate": 9.151918048170015e-06, "loss": 0.5732, "step": 10395 }, { "epoch": 1.62, "grad_norm": 17.29753428997181, "learning_rate": 9.150237736858376e-06, "loss": 0.5638, "step": 10396 }, { "epoch": 1.62, "grad_norm": 23.442121723980062, "learning_rate": 9.148557449713481e-06, "loss": 0.5895, "step": 10397 }, { "epoch": 1.62, "grad_norm": 19.856061955456987, "learning_rate": 9.146877186783112e-06, "loss": 0.5639, "step": 10398 }, { "epoch": 1.62, "grad_norm": 22.664234970078237, "learning_rate": 9.145196948115053e-06, "loss": 0.5464, "step": 10399 }, { "epoch": 1.62, "grad_norm": 22.588811622575363, "learning_rate": 9.143516733757097e-06, "loss": 0.5261, "step": 10400 }, { "epoch": 1.62, "grad_norm": 15.106962277375844, "learning_rate": 9.141836543757025e-06, "loss": 0.6426, "step": 10401 }, { "epoch": 1.62, "grad_norm": 25.18564480262938, "learning_rate": 9.140156378162619e-06, "loss": 0.6504, "step": 10402 }, { "epoch": 1.62, "grad_norm": 20.423011503722876, "learning_rate": 9.138476237021659e-06, "loss": 0.5862, "step": 10403 }, { "epoch": 1.63, "grad_norm": 25.163164573407695, "learning_rate": 9.136796120381933e-06, "loss": 0.6266, "step": 10404 }, { "epoch": 1.63, "grad_norm": 22.171129386030675, "learning_rate": 9.13511602829122e-06, "loss": 0.6082, "step": 10405 }, { "epoch": 1.63, "grad_norm": 15.648151915025393, "learning_rate": 9.133435960797304e-06, "loss": 0.5139, "step": 10406 }, { "epoch": 1.63, "grad_norm": 34.61505707495903, "learning_rate": 9.131755917947958e-06, "loss": 0.6058, "step": 10407 }, { "epoch": 1.63, "grad_norm": 23.20616615479577, "learning_rate": 9.13007589979097e-06, "loss": 0.6402, "step": 10408 }, { "epoch": 1.63, "grad_norm": 13.013305828952747, "learning_rate": 9.12839590637411e-06, "loss": 0.5154, "step": 10409 }, { "epoch": 1.63, "grad_norm": 15.371907195460599, "learning_rate": 9.126715937745163e-06, "loss": 0.585, "step": 10410 }, { "epoch": 1.63, "grad_norm": 14.81631373047951, "learning_rate": 9.125035993951907e-06, "loss": 0.6006, "step": 10411 }, { "epoch": 1.63, "grad_norm": 18.413609229850618, "learning_rate": 9.12335607504211e-06, "loss": 0.581, "step": 10412 }, { "epoch": 1.63, "grad_norm": 18.397150209266, "learning_rate": 9.121676181063556e-06, "loss": 0.5582, "step": 10413 }, { "epoch": 1.63, "grad_norm": 20.1919420724341, "learning_rate": 9.119996312064014e-06, "loss": 0.5781, "step": 10414 }, { "epoch": 1.63, "grad_norm": 18.907842755811654, "learning_rate": 9.118316468091267e-06, "loss": 0.525, "step": 10415 }, { "epoch": 1.63, "grad_norm": 18.881090241103244, "learning_rate": 9.116636649193085e-06, "loss": 0.5209, "step": 10416 }, { "epoch": 1.63, "grad_norm": 15.368469373418282, "learning_rate": 9.114956855417236e-06, "loss": 0.5508, "step": 10417 }, { "epoch": 1.63, "grad_norm": 24.752123591082036, "learning_rate": 9.113277086811495e-06, "loss": 0.503, "step": 10418 }, { "epoch": 1.63, "grad_norm": 16.305718111075542, "learning_rate": 9.11159734342364e-06, "loss": 0.5498, "step": 10419 }, { "epoch": 1.63, "grad_norm": 15.812254986730656, "learning_rate": 9.109917625301432e-06, "loss": 0.5825, "step": 10420 }, { "epoch": 1.63, "grad_norm": 14.583515730608436, "learning_rate": 9.108237932492652e-06, "loss": 0.5404, "step": 10421 }, { "epoch": 1.63, "grad_norm": 15.505171441865173, "learning_rate": 9.106558265045058e-06, "loss": 0.5211, "step": 10422 }, { "epoch": 1.63, "grad_norm": 22.580835701071518, "learning_rate": 9.104878623006427e-06, "loss": 0.583, "step": 10423 }, { "epoch": 1.63, "grad_norm": 21.24192872894691, "learning_rate": 9.103199006424521e-06, "loss": 0.5479, "step": 10424 }, { "epoch": 1.63, "grad_norm": 21.851000703934886, "learning_rate": 9.101519415347113e-06, "loss": 0.5859, "step": 10425 }, { "epoch": 1.63, "grad_norm": 17.38091153155122, "learning_rate": 9.099839849821971e-06, "loss": 0.5956, "step": 10426 }, { "epoch": 1.63, "grad_norm": 62.709451744069504, "learning_rate": 9.098160309896852e-06, "loss": 0.5458, "step": 10427 }, { "epoch": 1.63, "grad_norm": 17.208093056625394, "learning_rate": 9.096480795619527e-06, "loss": 0.5215, "step": 10428 }, { "epoch": 1.63, "grad_norm": 18.630082533037218, "learning_rate": 9.09480130703776e-06, "loss": 0.5565, "step": 10429 }, { "epoch": 1.63, "grad_norm": 27.96890927061625, "learning_rate": 9.093121844199312e-06, "loss": 0.6058, "step": 10430 }, { "epoch": 1.63, "grad_norm": 22.805145402828263, "learning_rate": 9.091442407151955e-06, "loss": 0.5889, "step": 10431 }, { "epoch": 1.63, "grad_norm": 23.358026581697406, "learning_rate": 9.08976299594344e-06, "loss": 0.5732, "step": 10432 }, { "epoch": 1.63, "grad_norm": 17.60574966927669, "learning_rate": 9.088083610621532e-06, "loss": 0.5694, "step": 10433 }, { "epoch": 1.63, "grad_norm": 27.29064929366573, "learning_rate": 9.086404251233993e-06, "loss": 0.5882, "step": 10434 }, { "epoch": 1.63, "grad_norm": 18.62426147660861, "learning_rate": 9.084724917828585e-06, "loss": 0.5358, "step": 10435 }, { "epoch": 1.63, "grad_norm": 14.684225797940414, "learning_rate": 9.083045610453065e-06, "loss": 0.5923, "step": 10436 }, { "epoch": 1.63, "grad_norm": 14.8363555208507, "learning_rate": 9.081366329155188e-06, "loss": 0.6223, "step": 10437 }, { "epoch": 1.63, "grad_norm": 12.02217791405987, "learning_rate": 9.079687073982719e-06, "loss": 0.5745, "step": 10438 }, { "epoch": 1.63, "grad_norm": 18.561908722139812, "learning_rate": 9.078007844983406e-06, "loss": 0.6294, "step": 10439 }, { "epoch": 1.63, "grad_norm": 29.790703838079608, "learning_rate": 9.076328642205015e-06, "loss": 0.6192, "step": 10440 }, { "epoch": 1.63, "grad_norm": 13.040004356222106, "learning_rate": 9.074649465695298e-06, "loss": 0.5477, "step": 10441 }, { "epoch": 1.63, "grad_norm": 20.168233502881762, "learning_rate": 9.072970315502008e-06, "loss": 0.6195, "step": 10442 }, { "epoch": 1.63, "grad_norm": 19.03000217490477, "learning_rate": 9.071291191672896e-06, "loss": 0.621, "step": 10443 }, { "epoch": 1.63, "grad_norm": 25.93311040169013, "learning_rate": 9.069612094255722e-06, "loss": 0.588, "step": 10444 }, { "epoch": 1.63, "grad_norm": 20.210828988381525, "learning_rate": 9.067933023298234e-06, "loss": 0.5322, "step": 10445 }, { "epoch": 1.63, "grad_norm": 13.698145656581431, "learning_rate": 9.066253978848192e-06, "loss": 0.5393, "step": 10446 }, { "epoch": 1.63, "grad_norm": 20.219936736948377, "learning_rate": 9.064574960953334e-06, "loss": 0.5373, "step": 10447 }, { "epoch": 1.63, "grad_norm": 30.702779209820246, "learning_rate": 9.06289596966142e-06, "loss": 0.6247, "step": 10448 }, { "epoch": 1.63, "grad_norm": 27.47793611840959, "learning_rate": 9.061217005020194e-06, "loss": 0.5536, "step": 10449 }, { "epoch": 1.63, "grad_norm": 22.542843646619687, "learning_rate": 9.05953806707741e-06, "loss": 0.4856, "step": 10450 }, { "epoch": 1.63, "grad_norm": 29.10338815121388, "learning_rate": 9.057859155880811e-06, "loss": 0.5466, "step": 10451 }, { "epoch": 1.63, "grad_norm": 14.75676527896899, "learning_rate": 9.05618027147815e-06, "loss": 0.5425, "step": 10452 }, { "epoch": 1.63, "grad_norm": 21.432956507410985, "learning_rate": 9.054501413917163e-06, "loss": 0.5824, "step": 10453 }, { "epoch": 1.63, "grad_norm": 23.85335939962612, "learning_rate": 9.052822583245608e-06, "loss": 0.5536, "step": 10454 }, { "epoch": 1.63, "grad_norm": 13.364184912418985, "learning_rate": 9.051143779511226e-06, "loss": 0.5782, "step": 10455 }, { "epoch": 1.63, "grad_norm": 26.731295591658434, "learning_rate": 9.049465002761756e-06, "loss": 0.6328, "step": 10456 }, { "epoch": 1.63, "grad_norm": 20.87272332962382, "learning_rate": 9.047786253044945e-06, "loss": 0.5642, "step": 10457 }, { "epoch": 1.63, "grad_norm": 24.542904040311868, "learning_rate": 9.046107530408537e-06, "loss": 0.5402, "step": 10458 }, { "epoch": 1.63, "grad_norm": 20.292259341223126, "learning_rate": 9.044428834900274e-06, "loss": 0.5507, "step": 10459 }, { "epoch": 1.63, "grad_norm": 21.090231742617792, "learning_rate": 9.042750166567898e-06, "loss": 0.6195, "step": 10460 }, { "epoch": 1.63, "grad_norm": 17.310976359907976, "learning_rate": 9.041071525459145e-06, "loss": 0.564, "step": 10461 }, { "epoch": 1.63, "grad_norm": 15.651425541524818, "learning_rate": 9.039392911621754e-06, "loss": 0.5527, "step": 10462 }, { "epoch": 1.63, "grad_norm": 14.544464442729812, "learning_rate": 9.03771432510347e-06, "loss": 0.5885, "step": 10463 }, { "epoch": 1.63, "grad_norm": 14.909516456475053, "learning_rate": 9.036035765952023e-06, "loss": 0.5726, "step": 10464 }, { "epoch": 1.63, "grad_norm": 18.64231401738984, "learning_rate": 9.034357234215164e-06, "loss": 0.5706, "step": 10465 }, { "epoch": 1.63, "grad_norm": 17.89626764825909, "learning_rate": 9.032678729940614e-06, "loss": 0.5832, "step": 10466 }, { "epoch": 1.63, "grad_norm": 24.214936565917274, "learning_rate": 9.031000253176117e-06, "loss": 0.5661, "step": 10467 }, { "epoch": 1.64, "grad_norm": 24.995543725587893, "learning_rate": 9.029321803969402e-06, "loss": 0.6112, "step": 10468 }, { "epoch": 1.64, "grad_norm": 21.20016321517818, "learning_rate": 9.02764338236821e-06, "loss": 0.5867, "step": 10469 }, { "epoch": 1.64, "grad_norm": 16.50360846075877, "learning_rate": 9.025964988420274e-06, "loss": 0.6435, "step": 10470 }, { "epoch": 1.64, "grad_norm": 17.068692340826775, "learning_rate": 9.024286622173323e-06, "loss": 0.6163, "step": 10471 }, { "epoch": 1.64, "grad_norm": 16.56694383205472, "learning_rate": 9.022608283675086e-06, "loss": 0.6321, "step": 10472 }, { "epoch": 1.64, "grad_norm": 13.86294535479617, "learning_rate": 9.0209299729733e-06, "loss": 0.592, "step": 10473 }, { "epoch": 1.64, "grad_norm": 18.245385621951677, "learning_rate": 9.019251690115692e-06, "loss": 0.6287, "step": 10474 }, { "epoch": 1.64, "grad_norm": 25.085652060178237, "learning_rate": 9.017573435149994e-06, "loss": 0.606, "step": 10475 }, { "epoch": 1.64, "grad_norm": 25.030911748127032, "learning_rate": 9.015895208123933e-06, "loss": 0.6451, "step": 10476 }, { "epoch": 1.64, "grad_norm": 21.749620147928773, "learning_rate": 9.014217009085232e-06, "loss": 0.622, "step": 10477 }, { "epoch": 1.64, "grad_norm": 17.241953794194444, "learning_rate": 9.012538838081626e-06, "loss": 0.6088, "step": 10478 }, { "epoch": 1.64, "grad_norm": 16.791576828274543, "learning_rate": 9.010860695160839e-06, "loss": 0.5898, "step": 10479 }, { "epoch": 1.64, "grad_norm": 18.109710640467018, "learning_rate": 9.009182580370596e-06, "loss": 0.5762, "step": 10480 }, { "epoch": 1.64, "grad_norm": 14.69851413024346, "learning_rate": 9.007504493758616e-06, "loss": 0.502, "step": 10481 }, { "epoch": 1.64, "grad_norm": 15.708258823029182, "learning_rate": 9.005826435372631e-06, "loss": 0.5825, "step": 10482 }, { "epoch": 1.64, "grad_norm": 32.349294069837086, "learning_rate": 9.004148405260358e-06, "loss": 0.5218, "step": 10483 }, { "epoch": 1.64, "grad_norm": 16.486705073880735, "learning_rate": 9.002470403469525e-06, "loss": 0.5413, "step": 10484 }, { "epoch": 1.64, "grad_norm": 19.943930851873372, "learning_rate": 9.000792430047852e-06, "loss": 0.6559, "step": 10485 }, { "epoch": 1.64, "grad_norm": 17.393029896664174, "learning_rate": 8.999114485043053e-06, "loss": 0.6228, "step": 10486 }, { "epoch": 1.64, "grad_norm": 14.855324695579824, "learning_rate": 8.997436568502852e-06, "loss": 0.5897, "step": 10487 }, { "epoch": 1.64, "grad_norm": 38.82854908916895, "learning_rate": 8.995758680474972e-06, "loss": 0.6062, "step": 10488 }, { "epoch": 1.64, "grad_norm": 30.91201636922907, "learning_rate": 8.994080821007124e-06, "loss": 0.5471, "step": 10489 }, { "epoch": 1.64, "grad_norm": 16.952062258832903, "learning_rate": 8.992402990147035e-06, "loss": 0.596, "step": 10490 }, { "epoch": 1.64, "grad_norm": 14.700295361483954, "learning_rate": 8.99072518794241e-06, "loss": 0.5504, "step": 10491 }, { "epoch": 1.64, "grad_norm": 26.66921100374453, "learning_rate": 8.989047414440973e-06, "loss": 0.6041, "step": 10492 }, { "epoch": 1.64, "grad_norm": 23.783490233980313, "learning_rate": 8.987369669690433e-06, "loss": 0.6787, "step": 10493 }, { "epoch": 1.64, "grad_norm": 22.90884712817259, "learning_rate": 8.98569195373851e-06, "loss": 0.5389, "step": 10494 }, { "epoch": 1.64, "grad_norm": 15.559373820232615, "learning_rate": 8.984014266632916e-06, "loss": 0.5523, "step": 10495 }, { "epoch": 1.64, "grad_norm": 22.224304031512577, "learning_rate": 8.98233660842136e-06, "loss": 0.5691, "step": 10496 }, { "epoch": 1.64, "grad_norm": 14.789810081162711, "learning_rate": 8.980658979151556e-06, "loss": 0.6064, "step": 10497 }, { "epoch": 1.64, "grad_norm": 16.598343913717517, "learning_rate": 8.978981378871213e-06, "loss": 0.5818, "step": 10498 }, { "epoch": 1.64, "grad_norm": 14.935223357032948, "learning_rate": 8.977303807628046e-06, "loss": 0.5364, "step": 10499 }, { "epoch": 1.64, "grad_norm": 15.71346149770786, "learning_rate": 8.975626265469755e-06, "loss": 0.5531, "step": 10500 }, { "epoch": 1.64, "grad_norm": 15.912235980585484, "learning_rate": 8.973948752444057e-06, "loss": 0.5786, "step": 10501 }, { "epoch": 1.64, "grad_norm": 15.822584951635, "learning_rate": 8.972271268598653e-06, "loss": 0.5186, "step": 10502 }, { "epoch": 1.64, "grad_norm": 12.646987580410274, "learning_rate": 8.970593813981257e-06, "loss": 0.527, "step": 10503 }, { "epoch": 1.64, "grad_norm": 21.37215044870707, "learning_rate": 8.96891638863957e-06, "loss": 0.5504, "step": 10504 }, { "epoch": 1.64, "grad_norm": 18.32014397053585, "learning_rate": 8.967238992621297e-06, "loss": 0.5552, "step": 10505 }, { "epoch": 1.64, "grad_norm": 32.506098366750244, "learning_rate": 8.965561625974142e-06, "loss": 0.6567, "step": 10506 }, { "epoch": 1.64, "grad_norm": 22.63123209883333, "learning_rate": 8.96388428874581e-06, "loss": 0.5346, "step": 10507 }, { "epoch": 1.64, "grad_norm": 29.01387683201534, "learning_rate": 8.962206980984001e-06, "loss": 0.5601, "step": 10508 }, { "epoch": 1.64, "grad_norm": 20.579383760185625, "learning_rate": 8.960529702736425e-06, "loss": 0.6136, "step": 10509 }, { "epoch": 1.64, "grad_norm": 33.94865908966314, "learning_rate": 8.95885245405077e-06, "loss": 0.6253, "step": 10510 }, { "epoch": 1.64, "grad_norm": 14.326960453414634, "learning_rate": 8.957175234974744e-06, "loss": 0.5111, "step": 10511 }, { "epoch": 1.64, "grad_norm": 14.295587042864588, "learning_rate": 8.955498045556041e-06, "loss": 0.5929, "step": 10512 }, { "epoch": 1.64, "grad_norm": 15.619679746855416, "learning_rate": 8.953820885842366e-06, "loss": 0.5436, "step": 10513 }, { "epoch": 1.64, "grad_norm": 18.316659410239936, "learning_rate": 8.952143755881416e-06, "loss": 0.5652, "step": 10514 }, { "epoch": 1.64, "grad_norm": 19.005804529680333, "learning_rate": 8.95046665572088e-06, "loss": 0.5638, "step": 10515 }, { "epoch": 1.64, "grad_norm": 23.193856865295665, "learning_rate": 8.948789585408457e-06, "loss": 0.6839, "step": 10516 }, { "epoch": 1.64, "grad_norm": 29.485420764435943, "learning_rate": 8.947112544991846e-06, "loss": 0.691, "step": 10517 }, { "epoch": 1.64, "grad_norm": 18.825134452574616, "learning_rate": 8.945435534518737e-06, "loss": 0.535, "step": 10518 }, { "epoch": 1.64, "grad_norm": 17.698285867340033, "learning_rate": 8.943758554036828e-06, "loss": 0.6034, "step": 10519 }, { "epoch": 1.64, "grad_norm": 18.116419948227406, "learning_rate": 8.942081603593805e-06, "loss": 0.5716, "step": 10520 }, { "epoch": 1.64, "grad_norm": 11.646230375318012, "learning_rate": 8.94040468323736e-06, "loss": 0.4644, "step": 10521 }, { "epoch": 1.64, "grad_norm": 15.147505503801593, "learning_rate": 8.93872779301519e-06, "loss": 0.5879, "step": 10522 }, { "epoch": 1.64, "grad_norm": 23.506815008398426, "learning_rate": 8.93705093297498e-06, "loss": 0.6004, "step": 10523 }, { "epoch": 1.64, "grad_norm": 28.961036983470393, "learning_rate": 8.93537410316442e-06, "loss": 0.5787, "step": 10524 }, { "epoch": 1.64, "grad_norm": 25.411648502851467, "learning_rate": 8.933697303631195e-06, "loss": 0.5981, "step": 10525 }, { "epoch": 1.64, "grad_norm": 16.85277461758098, "learning_rate": 8.932020534422997e-06, "loss": 0.5784, "step": 10526 }, { "epoch": 1.64, "grad_norm": 20.567264533179394, "learning_rate": 8.930343795587508e-06, "loss": 0.6279, "step": 10527 }, { "epoch": 1.64, "grad_norm": 19.783303706361327, "learning_rate": 8.928667087172417e-06, "loss": 0.5405, "step": 10528 }, { "epoch": 1.64, "grad_norm": 26.242620891570777, "learning_rate": 8.92699040922541e-06, "loss": 0.598, "step": 10529 }, { "epoch": 1.64, "grad_norm": 24.446708085871386, "learning_rate": 8.925313761794166e-06, "loss": 0.649, "step": 10530 }, { "epoch": 1.64, "grad_norm": 26.43444869004903, "learning_rate": 8.92363714492637e-06, "loss": 0.5425, "step": 10531 }, { "epoch": 1.65, "grad_norm": 19.426807286335617, "learning_rate": 8.921960558669704e-06, "loss": 0.5696, "step": 10532 }, { "epoch": 1.65, "grad_norm": 22.3181047966698, "learning_rate": 8.920284003071846e-06, "loss": 0.5564, "step": 10533 }, { "epoch": 1.65, "grad_norm": 19.588846155782246, "learning_rate": 8.918607478180487e-06, "loss": 0.6613, "step": 10534 }, { "epoch": 1.65, "grad_norm": 21.827995510850453, "learning_rate": 8.916930984043291e-06, "loss": 0.6416, "step": 10535 }, { "epoch": 1.65, "grad_norm": 15.902758208505269, "learning_rate": 8.915254520707948e-06, "loss": 0.6196, "step": 10536 }, { "epoch": 1.65, "grad_norm": 22.381999434105282, "learning_rate": 8.913578088222129e-06, "loss": 0.627, "step": 10537 }, { "epoch": 1.65, "grad_norm": 15.943552720985142, "learning_rate": 8.911901686633516e-06, "loss": 0.5887, "step": 10538 }, { "epoch": 1.65, "grad_norm": 20.470312676219738, "learning_rate": 8.910225315989786e-06, "loss": 0.532, "step": 10539 }, { "epoch": 1.65, "grad_norm": 17.918104668553518, "learning_rate": 8.908548976338604e-06, "loss": 0.5603, "step": 10540 }, { "epoch": 1.65, "grad_norm": 26.809946006343832, "learning_rate": 8.906872667727652e-06, "loss": 0.5891, "step": 10541 }, { "epoch": 1.65, "grad_norm": 17.78282057566971, "learning_rate": 8.905196390204604e-06, "loss": 0.5432, "step": 10542 }, { "epoch": 1.65, "grad_norm": 22.873296661550523, "learning_rate": 8.903520143817128e-06, "loss": 0.4863, "step": 10543 }, { "epoch": 1.65, "grad_norm": 20.609685488784915, "learning_rate": 8.901843928612902e-06, "loss": 0.5901, "step": 10544 }, { "epoch": 1.65, "grad_norm": 19.091840284059156, "learning_rate": 8.90016774463959e-06, "loss": 0.5348, "step": 10545 }, { "epoch": 1.65, "grad_norm": 29.813559242190617, "learning_rate": 8.89849159194486e-06, "loss": 0.5508, "step": 10546 }, { "epoch": 1.65, "grad_norm": 13.20410714700571, "learning_rate": 8.89681547057639e-06, "loss": 0.5794, "step": 10547 }, { "epoch": 1.65, "grad_norm": 21.54428253732981, "learning_rate": 8.895139380581843e-06, "loss": 0.5544, "step": 10548 }, { "epoch": 1.65, "grad_norm": 15.134956746082892, "learning_rate": 8.893463322008885e-06, "loss": 0.5286, "step": 10549 }, { "epoch": 1.65, "grad_norm": 18.96443540760124, "learning_rate": 8.891787294905178e-06, "loss": 0.5442, "step": 10550 }, { "epoch": 1.65, "grad_norm": 14.288512659939148, "learning_rate": 8.890111299318396e-06, "loss": 0.5372, "step": 10551 }, { "epoch": 1.65, "grad_norm": 20.022725706215542, "learning_rate": 8.888435335296197e-06, "loss": 0.6456, "step": 10552 }, { "epoch": 1.65, "grad_norm": 25.12157082642232, "learning_rate": 8.886759402886253e-06, "loss": 0.621, "step": 10553 }, { "epoch": 1.65, "grad_norm": 23.40396885521449, "learning_rate": 8.885083502136214e-06, "loss": 0.6463, "step": 10554 }, { "epoch": 1.65, "grad_norm": 26.201191068195104, "learning_rate": 8.883407633093751e-06, "loss": 0.6021, "step": 10555 }, { "epoch": 1.65, "grad_norm": 19.472412542503527, "learning_rate": 8.881731795806517e-06, "loss": 0.5157, "step": 10556 }, { "epoch": 1.65, "grad_norm": 29.813747006053468, "learning_rate": 8.88005599032218e-06, "loss": 0.6652, "step": 10557 }, { "epoch": 1.65, "grad_norm": 24.104706436439592, "learning_rate": 8.878380216688399e-06, "loss": 0.5574, "step": 10558 }, { "epoch": 1.65, "grad_norm": 14.11552890909985, "learning_rate": 8.876704474952826e-06, "loss": 0.527, "step": 10559 }, { "epoch": 1.65, "grad_norm": 21.01454380969632, "learning_rate": 8.875028765163117e-06, "loss": 0.5797, "step": 10560 }, { "epoch": 1.65, "grad_norm": 46.84552553615185, "learning_rate": 8.873353087366936e-06, "loss": 0.6101, "step": 10561 }, { "epoch": 1.65, "grad_norm": 16.20300028154275, "learning_rate": 8.87167744161193e-06, "loss": 0.5251, "step": 10562 }, { "epoch": 1.65, "grad_norm": 18.605342130132044, "learning_rate": 8.870001827945764e-06, "loss": 0.6308, "step": 10563 }, { "epoch": 1.65, "grad_norm": 22.411500302351627, "learning_rate": 8.86832624641608e-06, "loss": 0.5474, "step": 10564 }, { "epoch": 1.65, "grad_norm": 21.25021404270756, "learning_rate": 8.866650697070535e-06, "loss": 0.5891, "step": 10565 }, { "epoch": 1.65, "grad_norm": 23.708686623737233, "learning_rate": 8.864975179956784e-06, "loss": 0.6242, "step": 10566 }, { "epoch": 1.65, "grad_norm": 23.342905163318473, "learning_rate": 8.863299695122473e-06, "loss": 0.6278, "step": 10567 }, { "epoch": 1.65, "grad_norm": 12.250185552111372, "learning_rate": 8.861624242615258e-06, "loss": 0.5149, "step": 10568 }, { "epoch": 1.65, "grad_norm": 15.698050931771068, "learning_rate": 8.859948822482777e-06, "loss": 0.6111, "step": 10569 }, { "epoch": 1.65, "grad_norm": 15.970391682744916, "learning_rate": 8.85827343477269e-06, "loss": 0.6045, "step": 10570 }, { "epoch": 1.65, "grad_norm": 18.980238250287478, "learning_rate": 8.856598079532633e-06, "loss": 0.5814, "step": 10571 }, { "epoch": 1.65, "grad_norm": 22.126611172746667, "learning_rate": 8.854922756810263e-06, "loss": 0.615, "step": 10572 }, { "epoch": 1.65, "grad_norm": 15.35933075934371, "learning_rate": 8.85324746665322e-06, "loss": 0.524, "step": 10573 }, { "epoch": 1.65, "grad_norm": 17.341025619636287, "learning_rate": 8.851572209109147e-06, "loss": 0.5864, "step": 10574 }, { "epoch": 1.65, "grad_norm": 17.709495743204776, "learning_rate": 8.849896984225685e-06, "loss": 0.5425, "step": 10575 }, { "epoch": 1.65, "grad_norm": 15.90207491511352, "learning_rate": 8.848221792050484e-06, "loss": 0.5611, "step": 10576 }, { "epoch": 1.65, "grad_norm": 26.862340315170645, "learning_rate": 8.846546632631178e-06, "loss": 0.5316, "step": 10577 }, { "epoch": 1.65, "grad_norm": 15.112449246456217, "learning_rate": 8.844871506015417e-06, "loss": 0.5679, "step": 10578 }, { "epoch": 1.65, "grad_norm": 17.957798917973193, "learning_rate": 8.84319641225083e-06, "loss": 0.642, "step": 10579 }, { "epoch": 1.65, "grad_norm": 18.40861249716418, "learning_rate": 8.841521351385061e-06, "loss": 0.6312, "step": 10580 }, { "epoch": 1.65, "grad_norm": 26.820813264069383, "learning_rate": 8.839846323465745e-06, "loss": 0.5613, "step": 10581 }, { "epoch": 1.65, "grad_norm": 19.41690770797413, "learning_rate": 8.838171328540524e-06, "loss": 0.4693, "step": 10582 }, { "epoch": 1.65, "grad_norm": 14.989719391963709, "learning_rate": 8.836496366657032e-06, "loss": 0.562, "step": 10583 }, { "epoch": 1.65, "grad_norm": 16.3326707033684, "learning_rate": 8.834821437862899e-06, "loss": 0.5911, "step": 10584 }, { "epoch": 1.65, "grad_norm": 18.576962384099044, "learning_rate": 8.833146542205765e-06, "loss": 0.6162, "step": 10585 }, { "epoch": 1.65, "grad_norm": 13.4642771952231, "learning_rate": 8.831471679733262e-06, "loss": 0.5255, "step": 10586 }, { "epoch": 1.65, "grad_norm": 20.243070515432077, "learning_rate": 8.829796850493016e-06, "loss": 0.5509, "step": 10587 }, { "epoch": 1.65, "grad_norm": 12.86933225291331, "learning_rate": 8.828122054532669e-06, "loss": 0.5437, "step": 10588 }, { "epoch": 1.65, "grad_norm": 20.259582024807646, "learning_rate": 8.826447291899842e-06, "loss": 0.5232, "step": 10589 }, { "epoch": 1.65, "grad_norm": 19.975049566519548, "learning_rate": 8.824772562642166e-06, "loss": 0.5528, "step": 10590 }, { "epoch": 1.65, "grad_norm": 19.093213555733982, "learning_rate": 8.823097866807272e-06, "loss": 0.6207, "step": 10591 }, { "epoch": 1.65, "grad_norm": 14.031383719263756, "learning_rate": 8.82142320444279e-06, "loss": 0.5709, "step": 10592 }, { "epoch": 1.65, "grad_norm": 11.665876311398485, "learning_rate": 8.81974857559634e-06, "loss": 0.5485, "step": 10593 }, { "epoch": 1.65, "grad_norm": 19.212568141739474, "learning_rate": 8.818073980315547e-06, "loss": 0.5291, "step": 10594 }, { "epoch": 1.65, "grad_norm": 20.588865287298297, "learning_rate": 8.816399418648041e-06, "loss": 0.5733, "step": 10595 }, { "epoch": 1.66, "grad_norm": 22.849516488449012, "learning_rate": 8.814724890641444e-06, "loss": 0.5425, "step": 10596 }, { "epoch": 1.66, "grad_norm": 21.87554829097664, "learning_rate": 8.81305039634338e-06, "loss": 0.5708, "step": 10597 }, { "epoch": 1.66, "grad_norm": 15.465789852701272, "learning_rate": 8.811375935801463e-06, "loss": 0.5993, "step": 10598 }, { "epoch": 1.66, "grad_norm": 28.559402022352625, "learning_rate": 8.809701509063322e-06, "loss": 0.5763, "step": 10599 }, { "epoch": 1.66, "grad_norm": 22.03008716436355, "learning_rate": 8.808027116176572e-06, "loss": 0.606, "step": 10600 }, { "epoch": 1.66, "grad_norm": 14.245845244938598, "learning_rate": 8.806352757188835e-06, "loss": 0.5874, "step": 10601 }, { "epoch": 1.66, "grad_norm": 19.14003062356893, "learning_rate": 8.80467843214773e-06, "loss": 0.6026, "step": 10602 }, { "epoch": 1.66, "grad_norm": 14.204199934927258, "learning_rate": 8.803004141100869e-06, "loss": 0.5457, "step": 10603 }, { "epoch": 1.66, "grad_norm": 20.529404622585364, "learning_rate": 8.801329884095866e-06, "loss": 0.5848, "step": 10604 }, { "epoch": 1.66, "grad_norm": 11.298156957054113, "learning_rate": 8.799655661180344e-06, "loss": 0.5413, "step": 10605 }, { "epoch": 1.66, "grad_norm": 13.737168289497772, "learning_rate": 8.79798147240191e-06, "loss": 0.5144, "step": 10606 }, { "epoch": 1.66, "grad_norm": 23.805284703584743, "learning_rate": 8.796307317808183e-06, "loss": 0.58, "step": 10607 }, { "epoch": 1.66, "grad_norm": 25.893084368389925, "learning_rate": 8.79463319744677e-06, "loss": 0.6639, "step": 10608 }, { "epoch": 1.66, "grad_norm": 12.783285408542032, "learning_rate": 8.792959111365285e-06, "loss": 0.5912, "step": 10609 }, { "epoch": 1.66, "grad_norm": 17.15976555473086, "learning_rate": 8.791285059611335e-06, "loss": 0.5486, "step": 10610 }, { "epoch": 1.66, "grad_norm": 58.08449727410499, "learning_rate": 8.789611042232534e-06, "loss": 0.5791, "step": 10611 }, { "epoch": 1.66, "grad_norm": 14.858577084603429, "learning_rate": 8.787937059276486e-06, "loss": 0.6202, "step": 10612 }, { "epoch": 1.66, "grad_norm": 11.93035156485214, "learning_rate": 8.786263110790796e-06, "loss": 0.5729, "step": 10613 }, { "epoch": 1.66, "grad_norm": 15.78064060734215, "learning_rate": 8.784589196823075e-06, "loss": 0.5756, "step": 10614 }, { "epoch": 1.66, "grad_norm": 24.37020895162406, "learning_rate": 8.782915317420924e-06, "loss": 0.6149, "step": 10615 }, { "epoch": 1.66, "grad_norm": 12.33864730379393, "learning_rate": 8.781241472631952e-06, "loss": 0.5862, "step": 10616 }, { "epoch": 1.66, "grad_norm": 13.152805862941696, "learning_rate": 8.779567662503761e-06, "loss": 0.5163, "step": 10617 }, { "epoch": 1.66, "grad_norm": 24.74101156120906, "learning_rate": 8.77789388708395e-06, "loss": 0.6642, "step": 10618 }, { "epoch": 1.66, "grad_norm": 19.764350783510245, "learning_rate": 8.77622014642012e-06, "loss": 0.5889, "step": 10619 }, { "epoch": 1.66, "grad_norm": 18.104367664027645, "learning_rate": 8.774546440559874e-06, "loss": 0.6471, "step": 10620 }, { "epoch": 1.66, "grad_norm": 15.159442081747295, "learning_rate": 8.77287276955081e-06, "loss": 0.5624, "step": 10621 }, { "epoch": 1.66, "grad_norm": 17.761670993650963, "learning_rate": 8.771199133440533e-06, "loss": 0.6047, "step": 10622 }, { "epoch": 1.66, "grad_norm": 12.538438136756746, "learning_rate": 8.769525532276627e-06, "loss": 0.5557, "step": 10623 }, { "epoch": 1.66, "grad_norm": 22.501896265708194, "learning_rate": 8.767851966106696e-06, "loss": 0.5563, "step": 10624 }, { "epoch": 1.66, "grad_norm": 13.367783037672163, "learning_rate": 8.766178434978334e-06, "loss": 0.5335, "step": 10625 }, { "epoch": 1.66, "grad_norm": 15.82841159677827, "learning_rate": 8.764504938939137e-06, "loss": 0.5462, "step": 10626 }, { "epoch": 1.66, "grad_norm": 20.8786251580896, "learning_rate": 8.762831478036698e-06, "loss": 0.5568, "step": 10627 }, { "epoch": 1.66, "grad_norm": 16.848564690399293, "learning_rate": 8.761158052318605e-06, "loss": 0.5251, "step": 10628 }, { "epoch": 1.66, "grad_norm": 19.7890619201901, "learning_rate": 8.759484661832454e-06, "loss": 0.6059, "step": 10629 }, { "epoch": 1.66, "grad_norm": 16.41973266728501, "learning_rate": 8.757811306625833e-06, "loss": 0.5684, "step": 10630 }, { "epoch": 1.66, "grad_norm": 26.163483730326483, "learning_rate": 8.75613798674633e-06, "loss": 0.6227, "step": 10631 }, { "epoch": 1.66, "grad_norm": 16.754502767668455, "learning_rate": 8.754464702241539e-06, "loss": 0.5203, "step": 10632 }, { "epoch": 1.66, "grad_norm": 33.083697424926214, "learning_rate": 8.752791453159041e-06, "loss": 0.6494, "step": 10633 }, { "epoch": 1.66, "grad_norm": 18.333628676123528, "learning_rate": 8.751118239546424e-06, "loss": 0.5658, "step": 10634 }, { "epoch": 1.66, "grad_norm": 18.01461747942771, "learning_rate": 8.749445061451274e-06, "loss": 0.5694, "step": 10635 }, { "epoch": 1.66, "grad_norm": 20.762830054612646, "learning_rate": 8.747771918921176e-06, "loss": 0.5754, "step": 10636 }, { "epoch": 1.66, "grad_norm": 18.04331935765732, "learning_rate": 8.746098812003714e-06, "loss": 0.5217, "step": 10637 }, { "epoch": 1.66, "grad_norm": 21.803258697348074, "learning_rate": 8.744425740746464e-06, "loss": 0.6335, "step": 10638 }, { "epoch": 1.66, "grad_norm": 35.64794688542407, "learning_rate": 8.742752705197013e-06, "loss": 0.6455, "step": 10639 }, { "epoch": 1.66, "grad_norm": 16.46138732668498, "learning_rate": 8.741079705402939e-06, "loss": 0.5211, "step": 10640 }, { "epoch": 1.66, "grad_norm": 18.39230899749596, "learning_rate": 8.739406741411826e-06, "loss": 0.6125, "step": 10641 }, { "epoch": 1.66, "grad_norm": 24.056763465866894, "learning_rate": 8.737733813271244e-06, "loss": 0.6236, "step": 10642 }, { "epoch": 1.66, "grad_norm": 23.11978770809794, "learning_rate": 8.736060921028773e-06, "loss": 0.6278, "step": 10643 }, { "epoch": 1.66, "grad_norm": 18.55293153906435, "learning_rate": 8.73438806473199e-06, "loss": 0.6099, "step": 10644 }, { "epoch": 1.66, "grad_norm": 31.703294073925896, "learning_rate": 8.732715244428472e-06, "loss": 0.5905, "step": 10645 }, { "epoch": 1.66, "grad_norm": 20.044599065678185, "learning_rate": 8.731042460165793e-06, "loss": 0.6493, "step": 10646 }, { "epoch": 1.66, "grad_norm": 17.53069151279472, "learning_rate": 8.72936971199152e-06, "loss": 0.5679, "step": 10647 }, { "epoch": 1.66, "grad_norm": 19.404807442570014, "learning_rate": 8.72769699995323e-06, "loss": 0.5657, "step": 10648 }, { "epoch": 1.66, "grad_norm": 20.64926481592163, "learning_rate": 8.726024324098494e-06, "loss": 0.5606, "step": 10649 }, { "epoch": 1.66, "grad_norm": 19.096280654487593, "learning_rate": 8.724351684474878e-06, "loss": 0.6145, "step": 10650 }, { "epoch": 1.66, "grad_norm": 13.983145720513793, "learning_rate": 8.72267908112996e-06, "loss": 0.5798, "step": 10651 }, { "epoch": 1.66, "grad_norm": 14.717085467814213, "learning_rate": 8.721006514111299e-06, "loss": 0.6065, "step": 10652 }, { "epoch": 1.66, "grad_norm": 24.342680007155987, "learning_rate": 8.719333983466462e-06, "loss": 0.573, "step": 10653 }, { "epoch": 1.66, "grad_norm": 17.836395298734068, "learning_rate": 8.71766148924302e-06, "loss": 0.5968, "step": 10654 }, { "epoch": 1.66, "grad_norm": 19.802264576610416, "learning_rate": 8.715989031488534e-06, "loss": 0.5596, "step": 10655 }, { "epoch": 1.66, "grad_norm": 31.443259978035083, "learning_rate": 8.71431661025057e-06, "loss": 0.6468, "step": 10656 }, { "epoch": 1.66, "grad_norm": 23.129382336842877, "learning_rate": 8.712644225576687e-06, "loss": 0.5922, "step": 10657 }, { "epoch": 1.66, "grad_norm": 28.295943588734268, "learning_rate": 8.710971877514452e-06, "loss": 0.6159, "step": 10658 }, { "epoch": 1.66, "grad_norm": 15.424142085832925, "learning_rate": 8.709299566111418e-06, "loss": 0.5413, "step": 10659 }, { "epoch": 1.67, "grad_norm": 22.136337155818854, "learning_rate": 8.707627291415153e-06, "loss": 0.5498, "step": 10660 }, { "epoch": 1.67, "grad_norm": 17.46602635314065, "learning_rate": 8.705955053473215e-06, "loss": 0.5709, "step": 10661 }, { "epoch": 1.67, "grad_norm": 13.894539876768423, "learning_rate": 8.704282852333154e-06, "loss": 0.5348, "step": 10662 }, { "epoch": 1.67, "grad_norm": 20.28782670389082, "learning_rate": 8.702610688042531e-06, "loss": 0.5645, "step": 10663 }, { "epoch": 1.67, "grad_norm": 24.659024680775442, "learning_rate": 8.700938560648901e-06, "loss": 0.5586, "step": 10664 }, { "epoch": 1.67, "grad_norm": 20.07548852227552, "learning_rate": 8.699266470199817e-06, "loss": 0.5867, "step": 10665 }, { "epoch": 1.67, "grad_norm": 24.687843035475538, "learning_rate": 8.697594416742842e-06, "loss": 0.567, "step": 10666 }, { "epoch": 1.67, "grad_norm": 31.690082866130616, "learning_rate": 8.695922400325512e-06, "loss": 0.6163, "step": 10667 }, { "epoch": 1.67, "grad_norm": 19.467159651740882, "learning_rate": 8.694250420995387e-06, "loss": 0.5532, "step": 10668 }, { "epoch": 1.67, "grad_norm": 22.305395753775834, "learning_rate": 8.692578478800015e-06, "loss": 0.545, "step": 10669 }, { "epoch": 1.67, "grad_norm": 12.87434396299192, "learning_rate": 8.690906573786949e-06, "loss": 0.5222, "step": 10670 }, { "epoch": 1.67, "grad_norm": 18.899565599573474, "learning_rate": 8.689234706003735e-06, "loss": 0.5471, "step": 10671 }, { "epoch": 1.67, "grad_norm": 19.70416979057833, "learning_rate": 8.687562875497915e-06, "loss": 0.5633, "step": 10672 }, { "epoch": 1.67, "grad_norm": 15.019830241904353, "learning_rate": 8.685891082317043e-06, "loss": 0.5505, "step": 10673 }, { "epoch": 1.67, "grad_norm": 22.802964553488113, "learning_rate": 8.684219326508657e-06, "loss": 0.5658, "step": 10674 }, { "epoch": 1.67, "grad_norm": 21.074838988021675, "learning_rate": 8.682547608120305e-06, "loss": 0.6734, "step": 10675 }, { "epoch": 1.67, "grad_norm": 14.346922444668317, "learning_rate": 8.68087592719953e-06, "loss": 0.5758, "step": 10676 }, { "epoch": 1.67, "grad_norm": 17.69825046983035, "learning_rate": 8.679204283793872e-06, "loss": 0.5865, "step": 10677 }, { "epoch": 1.67, "grad_norm": 21.119922815723925, "learning_rate": 8.677532677950868e-06, "loss": 0.5694, "step": 10678 }, { "epoch": 1.67, "grad_norm": 12.771917218837862, "learning_rate": 8.675861109718064e-06, "loss": 0.5888, "step": 10679 }, { "epoch": 1.67, "grad_norm": 14.020549771981512, "learning_rate": 8.674189579142993e-06, "loss": 0.5295, "step": 10680 }, { "epoch": 1.67, "grad_norm": 15.487481324367025, "learning_rate": 8.672518086273199e-06, "loss": 0.5509, "step": 10681 }, { "epoch": 1.67, "grad_norm": 20.103027997523373, "learning_rate": 8.67084663115621e-06, "loss": 0.6162, "step": 10682 }, { "epoch": 1.67, "grad_norm": 22.89042324849556, "learning_rate": 8.669175213839566e-06, "loss": 0.6217, "step": 10683 }, { "epoch": 1.67, "grad_norm": 19.26186337375132, "learning_rate": 8.6675038343708e-06, "loss": 0.5555, "step": 10684 }, { "epoch": 1.67, "grad_norm": 16.550121457856772, "learning_rate": 8.665832492797447e-06, "loss": 0.624, "step": 10685 }, { "epoch": 1.67, "grad_norm": 39.97594128846003, "learning_rate": 8.664161189167039e-06, "loss": 0.5122, "step": 10686 }, { "epoch": 1.67, "grad_norm": 19.74867915085563, "learning_rate": 8.662489923527104e-06, "loss": 0.6164, "step": 10687 }, { "epoch": 1.67, "grad_norm": 30.633618695158294, "learning_rate": 8.66081869592517e-06, "loss": 0.5545, "step": 10688 }, { "epoch": 1.67, "grad_norm": 19.27831230397225, "learning_rate": 8.659147506408771e-06, "loss": 0.5698, "step": 10689 }, { "epoch": 1.67, "grad_norm": 18.733921291307386, "learning_rate": 8.657476355025436e-06, "loss": 0.5785, "step": 10690 }, { "epoch": 1.67, "grad_norm": 15.975913549921286, "learning_rate": 8.655805241822683e-06, "loss": 0.5286, "step": 10691 }, { "epoch": 1.67, "grad_norm": 26.078959152841342, "learning_rate": 8.654134166848045e-06, "loss": 0.6512, "step": 10692 }, { "epoch": 1.67, "grad_norm": 22.678011360871974, "learning_rate": 8.652463130149044e-06, "loss": 0.5497, "step": 10693 }, { "epoch": 1.67, "grad_norm": 18.936375516138952, "learning_rate": 8.6507921317732e-06, "loss": 0.5309, "step": 10694 }, { "epoch": 1.67, "grad_norm": 18.004562876209715, "learning_rate": 8.649121171768045e-06, "loss": 0.5622, "step": 10695 }, { "epoch": 1.67, "grad_norm": 14.525471573860942, "learning_rate": 8.64745025018109e-06, "loss": 0.5093, "step": 10696 }, { "epoch": 1.67, "grad_norm": 19.033996785871246, "learning_rate": 8.645779367059856e-06, "loss": 0.5196, "step": 10697 }, { "epoch": 1.67, "grad_norm": 19.416615364736174, "learning_rate": 8.644108522451868e-06, "loss": 0.5451, "step": 10698 }, { "epoch": 1.67, "grad_norm": 24.653989905853603, "learning_rate": 8.642437716404639e-06, "loss": 0.6033, "step": 10699 }, { "epoch": 1.67, "grad_norm": 29.13510300191042, "learning_rate": 8.64076694896569e-06, "loss": 0.6148, "step": 10700 }, { "epoch": 1.67, "grad_norm": 14.439080940319844, "learning_rate": 8.63909622018253e-06, "loss": 0.5496, "step": 10701 }, { "epoch": 1.67, "grad_norm": 15.81245478191997, "learning_rate": 8.63742553010268e-06, "loss": 0.4813, "step": 10702 }, { "epoch": 1.67, "grad_norm": 18.147938339259284, "learning_rate": 8.635754878773647e-06, "loss": 0.5894, "step": 10703 }, { "epoch": 1.67, "grad_norm": 13.729983585734047, "learning_rate": 8.63408426624295e-06, "loss": 0.5983, "step": 10704 }, { "epoch": 1.67, "grad_norm": 18.491812111122975, "learning_rate": 8.632413692558101e-06, "loss": 0.6297, "step": 10705 }, { "epoch": 1.67, "grad_norm": 10.5818421440928, "learning_rate": 8.630743157766603e-06, "loss": 0.5013, "step": 10706 }, { "epoch": 1.67, "grad_norm": 12.654129999293646, "learning_rate": 8.629072661915969e-06, "loss": 0.5424, "step": 10707 }, { "epoch": 1.67, "grad_norm": 32.61416526818084, "learning_rate": 8.627402205053705e-06, "loss": 0.6774, "step": 10708 }, { "epoch": 1.67, "grad_norm": 23.55944077973013, "learning_rate": 8.62573178722732e-06, "loss": 0.6439, "step": 10709 }, { "epoch": 1.67, "grad_norm": 37.969366467341764, "learning_rate": 8.624061408484323e-06, "loss": 0.6523, "step": 10710 }, { "epoch": 1.67, "grad_norm": 23.16996973632973, "learning_rate": 8.62239106887221e-06, "loss": 0.5698, "step": 10711 }, { "epoch": 1.67, "grad_norm": 14.657162908334145, "learning_rate": 8.620720768438491e-06, "loss": 0.5486, "step": 10712 }, { "epoch": 1.67, "grad_norm": 18.22964584719875, "learning_rate": 8.619050507230664e-06, "loss": 0.512, "step": 10713 }, { "epoch": 1.67, "grad_norm": 20.45094193864924, "learning_rate": 8.617380285296235e-06, "loss": 0.5656, "step": 10714 }, { "epoch": 1.67, "grad_norm": 12.745405615218226, "learning_rate": 8.615710102682704e-06, "loss": 0.541, "step": 10715 }, { "epoch": 1.67, "grad_norm": 16.072208825359976, "learning_rate": 8.614039959437563e-06, "loss": 0.4817, "step": 10716 }, { "epoch": 1.67, "grad_norm": 20.896459487998627, "learning_rate": 8.612369855608318e-06, "loss": 0.5105, "step": 10717 }, { "epoch": 1.67, "grad_norm": 15.380521483569984, "learning_rate": 8.610699791242463e-06, "loss": 0.5645, "step": 10718 }, { "epoch": 1.67, "grad_norm": 23.418514226902982, "learning_rate": 8.609029766387489e-06, "loss": 0.5872, "step": 10719 }, { "epoch": 1.67, "grad_norm": 20.166731689153888, "learning_rate": 8.6073597810909e-06, "loss": 0.5525, "step": 10720 }, { "epoch": 1.67, "grad_norm": 16.744072309163542, "learning_rate": 8.60568983540018e-06, "loss": 0.5425, "step": 10721 }, { "epoch": 1.67, "grad_norm": 16.926109861894297, "learning_rate": 8.604019929362826e-06, "loss": 0.5313, "step": 10722 }, { "epoch": 1.67, "grad_norm": 18.165968500856906, "learning_rate": 8.602350063026332e-06, "loss": 0.6692, "step": 10723 }, { "epoch": 1.68, "grad_norm": 23.68867824160515, "learning_rate": 8.600680236438182e-06, "loss": 0.6164, "step": 10724 }, { "epoch": 1.68, "grad_norm": 14.540329024768237, "learning_rate": 8.59901044964587e-06, "loss": 0.5649, "step": 10725 }, { "epoch": 1.68, "grad_norm": 18.795344318937275, "learning_rate": 8.597340702696879e-06, "loss": 0.5707, "step": 10726 }, { "epoch": 1.68, "grad_norm": 17.778287348288444, "learning_rate": 8.595670995638697e-06, "loss": 0.6534, "step": 10727 }, { "epoch": 1.68, "grad_norm": 16.846555598264512, "learning_rate": 8.59400132851881e-06, "loss": 0.6313, "step": 10728 }, { "epoch": 1.68, "grad_norm": 60.657481253762995, "learning_rate": 8.592331701384704e-06, "loss": 0.5221, "step": 10729 }, { "epoch": 1.68, "grad_norm": 14.445630685882481, "learning_rate": 8.590662114283864e-06, "loss": 0.5594, "step": 10730 }, { "epoch": 1.68, "grad_norm": 19.772839558640303, "learning_rate": 8.588992567263767e-06, "loss": 0.5873, "step": 10731 }, { "epoch": 1.68, "grad_norm": 18.228508343026352, "learning_rate": 8.587323060371893e-06, "loss": 0.5972, "step": 10732 }, { "epoch": 1.68, "grad_norm": 16.554875978096312, "learning_rate": 8.585653593655728e-06, "loss": 0.4995, "step": 10733 }, { "epoch": 1.68, "grad_norm": 32.764185484088166, "learning_rate": 8.583984167162744e-06, "loss": 0.5808, "step": 10734 }, { "epoch": 1.68, "grad_norm": 18.758167474796675, "learning_rate": 8.582314780940426e-06, "loss": 0.6025, "step": 10735 }, { "epoch": 1.68, "grad_norm": 24.464260946486014, "learning_rate": 8.580645435036246e-06, "loss": 0.6673, "step": 10736 }, { "epoch": 1.68, "grad_norm": 19.463984912020386, "learning_rate": 8.578976129497678e-06, "loss": 0.6484, "step": 10737 }, { "epoch": 1.68, "grad_norm": 15.224395771497912, "learning_rate": 8.577306864372196e-06, "loss": 0.5168, "step": 10738 }, { "epoch": 1.68, "grad_norm": 22.178047040568764, "learning_rate": 8.575637639707279e-06, "loss": 0.5852, "step": 10739 }, { "epoch": 1.68, "grad_norm": 14.675893029240875, "learning_rate": 8.573968455550393e-06, "loss": 0.5868, "step": 10740 }, { "epoch": 1.68, "grad_norm": 17.05917112309347, "learning_rate": 8.572299311949005e-06, "loss": 0.5661, "step": 10741 }, { "epoch": 1.68, "grad_norm": 15.85306827364698, "learning_rate": 8.570630208950593e-06, "loss": 0.5354, "step": 10742 }, { "epoch": 1.68, "grad_norm": 24.342044462683226, "learning_rate": 8.568961146602623e-06, "loss": 0.7152, "step": 10743 }, { "epoch": 1.68, "grad_norm": 17.53403104858795, "learning_rate": 8.567292124952562e-06, "loss": 0.5077, "step": 10744 }, { "epoch": 1.68, "grad_norm": 16.330861984728294, "learning_rate": 8.565623144047869e-06, "loss": 0.6172, "step": 10745 }, { "epoch": 1.68, "grad_norm": 11.933010554071792, "learning_rate": 8.563954203936018e-06, "loss": 0.5138, "step": 10746 }, { "epoch": 1.68, "grad_norm": 14.617867703211141, "learning_rate": 8.562285304664468e-06, "loss": 0.5158, "step": 10747 }, { "epoch": 1.68, "grad_norm": 19.220326867581598, "learning_rate": 8.560616446280685e-06, "loss": 0.5887, "step": 10748 }, { "epoch": 1.68, "grad_norm": 21.163282920783544, "learning_rate": 8.558947628832131e-06, "loss": 0.547, "step": 10749 }, { "epoch": 1.68, "grad_norm": 18.178377573961438, "learning_rate": 8.55727885236626e-06, "loss": 0.501, "step": 10750 }, { "epoch": 1.68, "grad_norm": 17.539447230513943, "learning_rate": 8.555610116930532e-06, "loss": 0.5629, "step": 10751 }, { "epoch": 1.68, "grad_norm": 17.153772731149665, "learning_rate": 8.55394142257241e-06, "loss": 0.6401, "step": 10752 }, { "epoch": 1.68, "grad_norm": 22.523729855680674, "learning_rate": 8.552272769339346e-06, "loss": 0.5685, "step": 10753 }, { "epoch": 1.68, "grad_norm": 15.144733790670937, "learning_rate": 8.550604157278804e-06, "loss": 0.5591, "step": 10754 }, { "epoch": 1.68, "grad_norm": 18.987875599972817, "learning_rate": 8.548935586438226e-06, "loss": 0.5283, "step": 10755 }, { "epoch": 1.68, "grad_norm": 30.957807026477884, "learning_rate": 8.547267056865072e-06, "loss": 0.6091, "step": 10756 }, { "epoch": 1.68, "grad_norm": 16.298346499211956, "learning_rate": 8.545598568606791e-06, "loss": 0.5476, "step": 10757 }, { "epoch": 1.68, "grad_norm": 20.67125036199832, "learning_rate": 8.54393012171084e-06, "loss": 0.6027, "step": 10758 }, { "epoch": 1.68, "grad_norm": 18.273155631003064, "learning_rate": 8.542261716224664e-06, "loss": 0.602, "step": 10759 }, { "epoch": 1.68, "grad_norm": 29.03129477851625, "learning_rate": 8.54059335219571e-06, "loss": 0.7013, "step": 10760 }, { "epoch": 1.68, "grad_norm": 19.30868392574454, "learning_rate": 8.538925029671428e-06, "loss": 0.5596, "step": 10761 }, { "epoch": 1.68, "grad_norm": 12.693635683094321, "learning_rate": 8.537256748699264e-06, "loss": 0.5799, "step": 10762 }, { "epoch": 1.68, "grad_norm": 27.46103990950827, "learning_rate": 8.53558850932666e-06, "loss": 0.5265, "step": 10763 }, { "epoch": 1.68, "grad_norm": 19.19505132964049, "learning_rate": 8.533920311601068e-06, "loss": 0.592, "step": 10764 }, { "epoch": 1.68, "grad_norm": 16.687741969440236, "learning_rate": 8.53225215556992e-06, "loss": 0.5944, "step": 10765 }, { "epoch": 1.68, "grad_norm": 12.572429562279567, "learning_rate": 8.530584041280661e-06, "loss": 0.4828, "step": 10766 }, { "epoch": 1.68, "grad_norm": 24.930239366180977, "learning_rate": 8.528915968780735e-06, "loss": 0.5969, "step": 10767 }, { "epoch": 1.68, "grad_norm": 21.761951885231092, "learning_rate": 8.527247938117577e-06, "loss": 0.5746, "step": 10768 }, { "epoch": 1.68, "grad_norm": 19.295019880691335, "learning_rate": 8.525579949338631e-06, "loss": 0.557, "step": 10769 }, { "epoch": 1.68, "grad_norm": 22.540317906369257, "learning_rate": 8.523912002491321e-06, "loss": 0.5871, "step": 10770 }, { "epoch": 1.68, "grad_norm": 19.08280525004356, "learning_rate": 8.522244097623095e-06, "loss": 0.5634, "step": 10771 }, { "epoch": 1.68, "grad_norm": 15.854267211107338, "learning_rate": 8.520576234781379e-06, "loss": 0.6086, "step": 10772 }, { "epoch": 1.68, "grad_norm": 26.61927199983796, "learning_rate": 8.518908414013612e-06, "loss": 0.5546, "step": 10773 }, { "epoch": 1.68, "grad_norm": 20.498914969037678, "learning_rate": 8.517240635367225e-06, "loss": 0.5376, "step": 10774 }, { "epoch": 1.68, "grad_norm": 16.791312800465825, "learning_rate": 8.515572898889645e-06, "loss": 0.6344, "step": 10775 }, { "epoch": 1.68, "grad_norm": 18.368579324180516, "learning_rate": 8.513905204628302e-06, "loss": 0.596, "step": 10776 }, { "epoch": 1.68, "grad_norm": 16.730523377658212, "learning_rate": 8.512237552630625e-06, "loss": 0.5299, "step": 10777 }, { "epoch": 1.68, "grad_norm": 19.58309404923637, "learning_rate": 8.510569942944043e-06, "loss": 0.5225, "step": 10778 }, { "epoch": 1.68, "grad_norm": 16.148507973932883, "learning_rate": 8.508902375615982e-06, "loss": 0.5601, "step": 10779 }, { "epoch": 1.68, "grad_norm": 23.22874099415552, "learning_rate": 8.507234850693864e-06, "loss": 0.549, "step": 10780 }, { "epoch": 1.68, "grad_norm": 15.78281593201867, "learning_rate": 8.505567368225116e-06, "loss": 0.5283, "step": 10781 }, { "epoch": 1.68, "grad_norm": 17.257811337587967, "learning_rate": 8.503899928257151e-06, "loss": 0.5914, "step": 10782 }, { "epoch": 1.68, "grad_norm": 17.45303843612514, "learning_rate": 8.502232530837404e-06, "loss": 0.6183, "step": 10783 }, { "epoch": 1.68, "grad_norm": 13.4942155240457, "learning_rate": 8.500565176013286e-06, "loss": 0.5439, "step": 10784 }, { "epoch": 1.68, "grad_norm": 26.19587388704196, "learning_rate": 8.498897863832215e-06, "loss": 0.6469, "step": 10785 }, { "epoch": 1.68, "grad_norm": 28.86250391636306, "learning_rate": 8.497230594341612e-06, "loss": 0.55, "step": 10786 }, { "epoch": 1.68, "grad_norm": 16.092384997136644, "learning_rate": 8.495563367588893e-06, "loss": 0.5076, "step": 10787 }, { "epoch": 1.69, "grad_norm": 19.610718187934484, "learning_rate": 8.493896183621474e-06, "loss": 0.5576, "step": 10788 }, { "epoch": 1.69, "grad_norm": 15.281071674538994, "learning_rate": 8.49222904248676e-06, "loss": 0.5307, "step": 10789 }, { "epoch": 1.69, "grad_norm": 14.53231879887502, "learning_rate": 8.490561944232173e-06, "loss": 0.5903, "step": 10790 }, { "epoch": 1.69, "grad_norm": 14.234581263813762, "learning_rate": 8.48889488890512e-06, "loss": 0.5798, "step": 10791 }, { "epoch": 1.69, "grad_norm": 21.00969939803462, "learning_rate": 8.487227876553012e-06, "loss": 0.5739, "step": 10792 }, { "epoch": 1.69, "grad_norm": 26.15902377305113, "learning_rate": 8.485560907223264e-06, "loss": 0.54, "step": 10793 }, { "epoch": 1.69, "grad_norm": 16.751103074845837, "learning_rate": 8.483893980963274e-06, "loss": 0.5303, "step": 10794 }, { "epoch": 1.69, "grad_norm": 15.350415813533033, "learning_rate": 8.48222709782045e-06, "loss": 0.5704, "step": 10795 }, { "epoch": 1.69, "grad_norm": 19.66552754341187, "learning_rate": 8.480560257842201e-06, "loss": 0.5487, "step": 10796 }, { "epoch": 1.69, "grad_norm": 13.400866645839347, "learning_rate": 8.478893461075929e-06, "loss": 0.5899, "step": 10797 }, { "epoch": 1.69, "grad_norm": 14.682271762916503, "learning_rate": 8.477226707569042e-06, "loss": 0.4899, "step": 10798 }, { "epoch": 1.69, "grad_norm": 24.790982739180222, "learning_rate": 8.475559997368932e-06, "loss": 0.5653, "step": 10799 }, { "epoch": 1.69, "grad_norm": 22.081901957836713, "learning_rate": 8.473893330523004e-06, "loss": 0.5363, "step": 10800 }, { "epoch": 1.69, "grad_norm": 35.674554635624865, "learning_rate": 8.472226707078655e-06, "loss": 0.5174, "step": 10801 }, { "epoch": 1.69, "grad_norm": 23.486239177828867, "learning_rate": 8.470560127083288e-06, "loss": 0.7111, "step": 10802 }, { "epoch": 1.69, "grad_norm": 19.824546100939184, "learning_rate": 8.468893590584299e-06, "loss": 0.5417, "step": 10803 }, { "epoch": 1.69, "grad_norm": 18.76853911033336, "learning_rate": 8.467227097629075e-06, "loss": 0.5925, "step": 10804 }, { "epoch": 1.69, "grad_norm": 18.867199731510848, "learning_rate": 8.465560648265017e-06, "loss": 0.5799, "step": 10805 }, { "epoch": 1.69, "grad_norm": 21.116085691283033, "learning_rate": 8.463894242539518e-06, "loss": 0.5612, "step": 10806 }, { "epoch": 1.69, "grad_norm": 18.025610195884738, "learning_rate": 8.462227880499967e-06, "loss": 0.5085, "step": 10807 }, { "epoch": 1.69, "grad_norm": 31.240371653239404, "learning_rate": 8.460561562193758e-06, "loss": 0.6057, "step": 10808 }, { "epoch": 1.69, "grad_norm": 14.539389937135326, "learning_rate": 8.458895287668277e-06, "loss": 0.5588, "step": 10809 }, { "epoch": 1.69, "grad_norm": 20.450996276492145, "learning_rate": 8.457229056970908e-06, "loss": 0.4768, "step": 10810 }, { "epoch": 1.69, "grad_norm": 19.282977499772514, "learning_rate": 8.455562870149046e-06, "loss": 0.5475, "step": 10811 }, { "epoch": 1.69, "grad_norm": 25.531087087510674, "learning_rate": 8.453896727250072e-06, "loss": 0.5638, "step": 10812 }, { "epoch": 1.69, "grad_norm": 13.864832133815286, "learning_rate": 8.452230628321373e-06, "loss": 0.5117, "step": 10813 }, { "epoch": 1.69, "grad_norm": 12.856022948887485, "learning_rate": 8.450564573410324e-06, "loss": 0.5839, "step": 10814 }, { "epoch": 1.69, "grad_norm": 13.973730064148045, "learning_rate": 8.448898562564316e-06, "loss": 0.5701, "step": 10815 }, { "epoch": 1.69, "grad_norm": 18.4664471573499, "learning_rate": 8.447232595830724e-06, "loss": 0.5822, "step": 10816 }, { "epoch": 1.69, "grad_norm": 23.41364848235764, "learning_rate": 8.445566673256928e-06, "loss": 0.4963, "step": 10817 }, { "epoch": 1.69, "grad_norm": 18.801153053286477, "learning_rate": 8.443900794890311e-06, "loss": 0.6126, "step": 10818 }, { "epoch": 1.69, "grad_norm": 18.906502736191467, "learning_rate": 8.442234960778242e-06, "loss": 0.5466, "step": 10819 }, { "epoch": 1.69, "grad_norm": 24.460889732779616, "learning_rate": 8.440569170968098e-06, "loss": 0.5557, "step": 10820 }, { "epoch": 1.69, "grad_norm": 17.172252852193743, "learning_rate": 8.438903425507257e-06, "loss": 0.5847, "step": 10821 }, { "epoch": 1.69, "grad_norm": 26.27801680209744, "learning_rate": 8.437237724443086e-06, "loss": 0.5818, "step": 10822 }, { "epoch": 1.69, "grad_norm": 14.469337238790432, "learning_rate": 8.435572067822964e-06, "loss": 0.5871, "step": 10823 }, { "epoch": 1.69, "grad_norm": 14.37746908192004, "learning_rate": 8.433906455694256e-06, "loss": 0.5186, "step": 10824 }, { "epoch": 1.69, "grad_norm": 17.13722574333145, "learning_rate": 8.432240888104331e-06, "loss": 0.5721, "step": 10825 }, { "epoch": 1.69, "grad_norm": 12.924817153930416, "learning_rate": 8.430575365100557e-06, "loss": 0.526, "step": 10826 }, { "epoch": 1.69, "grad_norm": 24.800927392310008, "learning_rate": 8.428909886730303e-06, "loss": 0.578, "step": 10827 }, { "epoch": 1.69, "grad_norm": 22.606140092472344, "learning_rate": 8.427244453040937e-06, "loss": 0.5831, "step": 10828 }, { "epoch": 1.69, "grad_norm": 22.35319666936993, "learning_rate": 8.425579064079811e-06, "loss": 0.5031, "step": 10829 }, { "epoch": 1.69, "grad_norm": 25.111957761121204, "learning_rate": 8.423913719894301e-06, "loss": 0.6381, "step": 10830 }, { "epoch": 1.69, "grad_norm": 20.767845607352733, "learning_rate": 8.42224842053176e-06, "loss": 0.6046, "step": 10831 }, { "epoch": 1.69, "grad_norm": 17.303750973367805, "learning_rate": 8.420583166039551e-06, "loss": 0.552, "step": 10832 }, { "epoch": 1.69, "grad_norm": 14.924034830635621, "learning_rate": 8.418917956465037e-06, "loss": 0.5613, "step": 10833 }, { "epoch": 1.69, "grad_norm": 21.41116487785557, "learning_rate": 8.417252791855566e-06, "loss": 0.5754, "step": 10834 }, { "epoch": 1.69, "grad_norm": 21.25801904501597, "learning_rate": 8.4155876722585e-06, "loss": 0.659, "step": 10835 }, { "epoch": 1.69, "grad_norm": 19.872164505998658, "learning_rate": 8.413922597721197e-06, "loss": 0.5593, "step": 10836 }, { "epoch": 1.69, "grad_norm": 14.992201474627034, "learning_rate": 8.41225756829101e-06, "loss": 0.5238, "step": 10837 }, { "epoch": 1.69, "grad_norm": 28.645610031066756, "learning_rate": 8.410592584015284e-06, "loss": 0.6622, "step": 10838 }, { "epoch": 1.69, "grad_norm": 21.259955414190802, "learning_rate": 8.408927644941373e-06, "loss": 0.553, "step": 10839 }, { "epoch": 1.69, "grad_norm": 15.493482280724884, "learning_rate": 8.407262751116633e-06, "loss": 0.5705, "step": 10840 }, { "epoch": 1.69, "grad_norm": 16.236174416706437, "learning_rate": 8.405597902588407e-06, "loss": 0.5587, "step": 10841 }, { "epoch": 1.69, "grad_norm": 18.273024350079066, "learning_rate": 8.403933099404047e-06, "loss": 0.6023, "step": 10842 }, { "epoch": 1.69, "grad_norm": 19.518787243893357, "learning_rate": 8.402268341610895e-06, "loss": 0.5304, "step": 10843 }, { "epoch": 1.69, "grad_norm": 21.61697461072004, "learning_rate": 8.400603629256294e-06, "loss": 0.5343, "step": 10844 }, { "epoch": 1.69, "grad_norm": 24.30347646101296, "learning_rate": 8.39893896238759e-06, "loss": 0.59, "step": 10845 }, { "epoch": 1.69, "grad_norm": 31.802763261309277, "learning_rate": 8.397274341052128e-06, "loss": 0.6912, "step": 10846 }, { "epoch": 1.69, "grad_norm": 28.9689021548912, "learning_rate": 8.395609765297249e-06, "loss": 0.6296, "step": 10847 }, { "epoch": 1.69, "grad_norm": 13.51220117238763, "learning_rate": 8.393945235170283e-06, "loss": 0.5139, "step": 10848 }, { "epoch": 1.69, "grad_norm": 23.253475208923728, "learning_rate": 8.392280750718577e-06, "loss": 0.6178, "step": 10849 }, { "epoch": 1.69, "grad_norm": 40.39131619804619, "learning_rate": 8.390616311989468e-06, "loss": 0.5895, "step": 10850 }, { "epoch": 1.69, "grad_norm": 17.26240991027494, "learning_rate": 8.388951919030287e-06, "loss": 0.5462, "step": 10851 }, { "epoch": 1.7, "grad_norm": 18.505393947093093, "learning_rate": 8.387287571888373e-06, "loss": 0.5715, "step": 10852 }, { "epoch": 1.7, "grad_norm": 29.347960694063094, "learning_rate": 8.385623270611058e-06, "loss": 0.621, "step": 10853 }, { "epoch": 1.7, "grad_norm": 16.533378821857262, "learning_rate": 8.383959015245669e-06, "loss": 0.5487, "step": 10854 }, { "epoch": 1.7, "grad_norm": 17.14716468032409, "learning_rate": 8.382294805839543e-06, "loss": 0.6003, "step": 10855 }, { "epoch": 1.7, "grad_norm": 30.36464852636328, "learning_rate": 8.380630642440006e-06, "loss": 0.5592, "step": 10856 }, { "epoch": 1.7, "grad_norm": 16.669617198803504, "learning_rate": 8.378966525094388e-06, "loss": 0.5366, "step": 10857 }, { "epoch": 1.7, "grad_norm": 18.349752383971616, "learning_rate": 8.377302453850012e-06, "loss": 0.5522, "step": 10858 }, { "epoch": 1.7, "grad_norm": 13.422825150170677, "learning_rate": 8.375638428754204e-06, "loss": 0.4631, "step": 10859 }, { "epoch": 1.7, "grad_norm": 21.44501896224781, "learning_rate": 8.373974449854288e-06, "loss": 0.594, "step": 10860 }, { "epoch": 1.7, "grad_norm": 14.357106091811323, "learning_rate": 8.37231051719759e-06, "loss": 0.541, "step": 10861 }, { "epoch": 1.7, "grad_norm": 21.665164062469536, "learning_rate": 8.37064663083143e-06, "loss": 0.4646, "step": 10862 }, { "epoch": 1.7, "grad_norm": 16.425120253536413, "learning_rate": 8.368982790803124e-06, "loss": 0.5297, "step": 10863 }, { "epoch": 1.7, "grad_norm": 19.342211771554663, "learning_rate": 8.367318997159992e-06, "loss": 0.6073, "step": 10864 }, { "epoch": 1.7, "grad_norm": 17.74727391746629, "learning_rate": 8.365655249949355e-06, "loss": 0.5718, "step": 10865 }, { "epoch": 1.7, "grad_norm": 14.425060577013328, "learning_rate": 8.363991549218525e-06, "loss": 0.4823, "step": 10866 }, { "epoch": 1.7, "grad_norm": 22.13968438827088, "learning_rate": 8.362327895014821e-06, "loss": 0.6008, "step": 10867 }, { "epoch": 1.7, "grad_norm": 17.737013640701857, "learning_rate": 8.360664287385552e-06, "loss": 0.6384, "step": 10868 }, { "epoch": 1.7, "grad_norm": 19.90273808471432, "learning_rate": 8.359000726378033e-06, "loss": 0.6097, "step": 10869 }, { "epoch": 1.7, "grad_norm": 16.962587155541268, "learning_rate": 8.35733721203957e-06, "loss": 0.5433, "step": 10870 }, { "epoch": 1.7, "grad_norm": 23.109271828844765, "learning_rate": 8.355673744417479e-06, "loss": 0.5872, "step": 10871 }, { "epoch": 1.7, "grad_norm": 15.762518960996564, "learning_rate": 8.354010323559065e-06, "loss": 0.567, "step": 10872 }, { "epoch": 1.7, "grad_norm": 19.263026035887965, "learning_rate": 8.352346949511632e-06, "loss": 0.5727, "step": 10873 }, { "epoch": 1.7, "grad_norm": 20.912423635715026, "learning_rate": 8.35068362232249e-06, "loss": 0.5527, "step": 10874 }, { "epoch": 1.7, "grad_norm": 11.301853715817181, "learning_rate": 8.34902034203894e-06, "loss": 0.5117, "step": 10875 }, { "epoch": 1.7, "grad_norm": 17.68786742870043, "learning_rate": 8.347357108708284e-06, "loss": 0.6024, "step": 10876 }, { "epoch": 1.7, "grad_norm": 21.077632427469204, "learning_rate": 8.345693922377829e-06, "loss": 0.5627, "step": 10877 }, { "epoch": 1.7, "grad_norm": 26.141580930151132, "learning_rate": 8.344030783094869e-06, "loss": 0.5991, "step": 10878 }, { "epoch": 1.7, "grad_norm": 15.240592496703146, "learning_rate": 8.342367690906702e-06, "loss": 0.521, "step": 10879 }, { "epoch": 1.7, "grad_norm": 17.5548296749909, "learning_rate": 8.340704645860632e-06, "loss": 0.6031, "step": 10880 }, { "epoch": 1.7, "grad_norm": 13.405925888531877, "learning_rate": 8.339041648003952e-06, "loss": 0.5575, "step": 10881 }, { "epoch": 1.7, "grad_norm": 20.430878572772418, "learning_rate": 8.337378697383956e-06, "loss": 0.6364, "step": 10882 }, { "epoch": 1.7, "grad_norm": 27.631832317885262, "learning_rate": 8.335715794047933e-06, "loss": 0.6396, "step": 10883 }, { "epoch": 1.7, "grad_norm": 13.86446759722926, "learning_rate": 8.33405293804318e-06, "loss": 0.5707, "step": 10884 }, { "epoch": 1.7, "grad_norm": 18.897243542465798, "learning_rate": 8.332390129416988e-06, "loss": 0.5846, "step": 10885 }, { "epoch": 1.7, "grad_norm": 13.279694136198131, "learning_rate": 8.330727368216647e-06, "loss": 0.6303, "step": 10886 }, { "epoch": 1.7, "grad_norm": 11.776156146395545, "learning_rate": 8.329064654489442e-06, "loss": 0.488, "step": 10887 }, { "epoch": 1.7, "grad_norm": 13.848680261969525, "learning_rate": 8.327401988282661e-06, "loss": 0.5125, "step": 10888 }, { "epoch": 1.7, "grad_norm": 15.010487715179941, "learning_rate": 8.325739369643588e-06, "loss": 0.5131, "step": 10889 }, { "epoch": 1.7, "grad_norm": 14.783162116317529, "learning_rate": 8.324076798619508e-06, "loss": 0.5867, "step": 10890 }, { "epoch": 1.7, "grad_norm": 20.27836650840371, "learning_rate": 8.322414275257707e-06, "loss": 0.6239, "step": 10891 }, { "epoch": 1.7, "grad_norm": 15.562074142205898, "learning_rate": 8.320751799605458e-06, "loss": 0.5891, "step": 10892 }, { "epoch": 1.7, "grad_norm": 19.156873202026848, "learning_rate": 8.319089371710048e-06, "loss": 0.5657, "step": 10893 }, { "epoch": 1.7, "grad_norm": 14.307776239000717, "learning_rate": 8.317426991618751e-06, "loss": 0.5497, "step": 10894 }, { "epoch": 1.7, "grad_norm": 21.456159758357202, "learning_rate": 8.315764659378845e-06, "loss": 0.6096, "step": 10895 }, { "epoch": 1.7, "grad_norm": 20.34234110323499, "learning_rate": 8.314102375037611e-06, "loss": 0.5512, "step": 10896 }, { "epoch": 1.7, "grad_norm": 20.335310418143504, "learning_rate": 8.312440138642319e-06, "loss": 0.5436, "step": 10897 }, { "epoch": 1.7, "grad_norm": 14.981360918954534, "learning_rate": 8.310777950240238e-06, "loss": 0.5684, "step": 10898 }, { "epoch": 1.7, "grad_norm": 18.586375878749468, "learning_rate": 8.309115809878646e-06, "loss": 0.5508, "step": 10899 }, { "epoch": 1.7, "grad_norm": 18.032988858217905, "learning_rate": 8.307453717604811e-06, "loss": 0.5258, "step": 10900 }, { "epoch": 1.7, "grad_norm": 15.502474815603955, "learning_rate": 8.305791673466006e-06, "loss": 0.5734, "step": 10901 }, { "epoch": 1.7, "grad_norm": 19.328851826862802, "learning_rate": 8.304129677509491e-06, "loss": 0.5672, "step": 10902 }, { "epoch": 1.7, "grad_norm": 21.237094704980088, "learning_rate": 8.302467729782535e-06, "loss": 0.5341, "step": 10903 }, { "epoch": 1.7, "grad_norm": 23.74807560803915, "learning_rate": 8.300805830332404e-06, "loss": 0.6041, "step": 10904 }, { "epoch": 1.7, "grad_norm": 21.99702414860258, "learning_rate": 8.299143979206363e-06, "loss": 0.5708, "step": 10905 }, { "epoch": 1.7, "grad_norm": 27.44408096202091, "learning_rate": 8.297482176451675e-06, "loss": 0.5676, "step": 10906 }, { "epoch": 1.7, "grad_norm": 23.123638509287442, "learning_rate": 8.295820422115594e-06, "loss": 0.5608, "step": 10907 }, { "epoch": 1.7, "grad_norm": 18.28172198974164, "learning_rate": 8.294158716245381e-06, "loss": 0.5305, "step": 10908 }, { "epoch": 1.7, "grad_norm": 21.354148299201572, "learning_rate": 8.2924970588883e-06, "loss": 0.5862, "step": 10909 }, { "epoch": 1.7, "grad_norm": 22.55309224210255, "learning_rate": 8.290835450091601e-06, "loss": 0.5633, "step": 10910 }, { "epoch": 1.7, "grad_norm": 21.20277382848681, "learning_rate": 8.289173889902546e-06, "loss": 0.5293, "step": 10911 }, { "epoch": 1.7, "grad_norm": 17.133885749202733, "learning_rate": 8.287512378368384e-06, "loss": 0.5776, "step": 10912 }, { "epoch": 1.7, "grad_norm": 16.523097200248724, "learning_rate": 8.285850915536367e-06, "loss": 0.5001, "step": 10913 }, { "epoch": 1.7, "grad_norm": 16.10297550986689, "learning_rate": 8.284189501453745e-06, "loss": 0.5975, "step": 10914 }, { "epoch": 1.7, "grad_norm": 15.999201031243887, "learning_rate": 8.28252813616777e-06, "loss": 0.5366, "step": 10915 }, { "epoch": 1.71, "grad_norm": 21.51660198564347, "learning_rate": 8.280866819725695e-06, "loss": 0.5432, "step": 10916 }, { "epoch": 1.71, "grad_norm": 19.794595704593096, "learning_rate": 8.279205552174758e-06, "loss": 0.6163, "step": 10917 }, { "epoch": 1.71, "grad_norm": 19.992468481262, "learning_rate": 8.27754433356221e-06, "loss": 0.6185, "step": 10918 }, { "epoch": 1.71, "grad_norm": 17.496271285399576, "learning_rate": 8.275883163935292e-06, "loss": 0.5806, "step": 10919 }, { "epoch": 1.71, "grad_norm": 22.46996722452469, "learning_rate": 8.274222043341247e-06, "loss": 0.6158, "step": 10920 }, { "epoch": 1.71, "grad_norm": 13.716957907184327, "learning_rate": 8.272560971827323e-06, "loss": 0.5498, "step": 10921 }, { "epoch": 1.71, "grad_norm": 20.150285231556925, "learning_rate": 8.270899949440751e-06, "loss": 0.4829, "step": 10922 }, { "epoch": 1.71, "grad_norm": 17.496475492785464, "learning_rate": 8.269238976228771e-06, "loss": 0.4964, "step": 10923 }, { "epoch": 1.71, "grad_norm": 4.502838690621552, "learning_rate": 8.267578052238624e-06, "loss": 0.5263, "step": 10924 }, { "epoch": 1.71, "grad_norm": 21.25713961797008, "learning_rate": 8.265917177517545e-06, "loss": 0.5561, "step": 10925 }, { "epoch": 1.71, "grad_norm": 20.398053843916482, "learning_rate": 8.264256352112768e-06, "loss": 0.5813, "step": 10926 }, { "epoch": 1.71, "grad_norm": 37.41464395443639, "learning_rate": 8.262595576071521e-06, "loss": 0.4747, "step": 10927 }, { "epoch": 1.71, "grad_norm": 17.847317025928152, "learning_rate": 8.260934849441042e-06, "loss": 0.4976, "step": 10928 }, { "epoch": 1.71, "grad_norm": 19.635553731026775, "learning_rate": 8.259274172268556e-06, "loss": 0.5807, "step": 10929 }, { "epoch": 1.71, "grad_norm": 23.640106078867394, "learning_rate": 8.2576135446013e-06, "loss": 0.6105, "step": 10930 }, { "epoch": 1.71, "grad_norm": 15.383362906505662, "learning_rate": 8.255952966486491e-06, "loss": 0.6145, "step": 10931 }, { "epoch": 1.71, "grad_norm": 11.106375030346186, "learning_rate": 8.25429243797136e-06, "loss": 0.524, "step": 10932 }, { "epoch": 1.71, "grad_norm": 30.64138129670902, "learning_rate": 8.252631959103129e-06, "loss": 0.6047, "step": 10933 }, { "epoch": 1.71, "grad_norm": 12.419194950835276, "learning_rate": 8.250971529929026e-06, "loss": 0.5218, "step": 10934 }, { "epoch": 1.71, "grad_norm": 20.669199615435286, "learning_rate": 8.249311150496271e-06, "loss": 0.5736, "step": 10935 }, { "epoch": 1.71, "grad_norm": 26.472133173678266, "learning_rate": 8.247650820852078e-06, "loss": 0.6087, "step": 10936 }, { "epoch": 1.71, "grad_norm": 29.620755318065687, "learning_rate": 8.245990541043672e-06, "loss": 0.5594, "step": 10937 }, { "epoch": 1.71, "grad_norm": 25.733572474892174, "learning_rate": 8.24433031111827e-06, "loss": 0.5994, "step": 10938 }, { "epoch": 1.71, "grad_norm": 17.432677710961624, "learning_rate": 8.242670131123085e-06, "loss": 0.5492, "step": 10939 }, { "epoch": 1.71, "grad_norm": 17.384402292824348, "learning_rate": 8.241010001105338e-06, "loss": 0.6059, "step": 10940 }, { "epoch": 1.71, "grad_norm": 20.774879041869337, "learning_rate": 8.239349921112235e-06, "loss": 0.5671, "step": 10941 }, { "epoch": 1.71, "grad_norm": 19.625399671269154, "learning_rate": 8.237689891190988e-06, "loss": 0.5578, "step": 10942 }, { "epoch": 1.71, "grad_norm": 19.42894168015925, "learning_rate": 8.236029911388811e-06, "loss": 0.5748, "step": 10943 }, { "epoch": 1.71, "grad_norm": 21.770875552165393, "learning_rate": 8.234369981752913e-06, "loss": 0.5848, "step": 10944 }, { "epoch": 1.71, "grad_norm": 17.350162655744025, "learning_rate": 8.232710102330502e-06, "loss": 0.5975, "step": 10945 }, { "epoch": 1.71, "grad_norm": 17.78063106926223, "learning_rate": 8.231050273168776e-06, "loss": 0.562, "step": 10946 }, { "epoch": 1.71, "grad_norm": 5.502297131688767, "learning_rate": 8.229390494314949e-06, "loss": 0.5946, "step": 10947 }, { "epoch": 1.71, "grad_norm": 15.264131519330666, "learning_rate": 8.227730765816216e-06, "loss": 0.5049, "step": 10948 }, { "epoch": 1.71, "grad_norm": 11.951478238830493, "learning_rate": 8.226071087719789e-06, "loss": 0.5444, "step": 10949 }, { "epoch": 1.71, "grad_norm": 19.942149590621575, "learning_rate": 8.224411460072863e-06, "loss": 0.5804, "step": 10950 }, { "epoch": 1.71, "grad_norm": 13.968534621022663, "learning_rate": 8.222751882922632e-06, "loss": 0.6098, "step": 10951 }, { "epoch": 1.71, "grad_norm": 16.37887084960751, "learning_rate": 8.221092356316298e-06, "loss": 0.5557, "step": 10952 }, { "epoch": 1.71, "grad_norm": 17.608654212247025, "learning_rate": 8.21943288030106e-06, "loss": 0.6502, "step": 10953 }, { "epoch": 1.71, "grad_norm": 16.563353896621614, "learning_rate": 8.217773454924105e-06, "loss": 0.5859, "step": 10954 }, { "epoch": 1.71, "grad_norm": 17.051143185334322, "learning_rate": 8.216114080232635e-06, "loss": 0.6179, "step": 10955 }, { "epoch": 1.71, "grad_norm": 15.33739395143801, "learning_rate": 8.214454756273836e-06, "loss": 0.5227, "step": 10956 }, { "epoch": 1.71, "grad_norm": 24.930093385983994, "learning_rate": 8.2127954830949e-06, "loss": 0.571, "step": 10957 }, { "epoch": 1.71, "grad_norm": 27.78895490859864, "learning_rate": 8.21113626074301e-06, "loss": 0.6368, "step": 10958 }, { "epoch": 1.71, "grad_norm": 18.97172718471198, "learning_rate": 8.20947708926536e-06, "loss": 0.5462, "step": 10959 }, { "epoch": 1.71, "grad_norm": 14.152619516337708, "learning_rate": 8.20781796870914e-06, "loss": 0.4816, "step": 10960 }, { "epoch": 1.71, "grad_norm": 14.294996336195624, "learning_rate": 8.20615889912152e-06, "loss": 0.5204, "step": 10961 }, { "epoch": 1.71, "grad_norm": 24.888929213512366, "learning_rate": 8.204499880549699e-06, "loss": 0.5903, "step": 10962 }, { "epoch": 1.71, "grad_norm": 24.16416867859061, "learning_rate": 8.202840913040847e-06, "loss": 0.593, "step": 10963 }, { "epoch": 1.71, "grad_norm": 34.83630193385815, "learning_rate": 8.201181996642147e-06, "loss": 0.6155, "step": 10964 }, { "epoch": 1.71, "grad_norm": 15.768592730950118, "learning_rate": 8.199523131400783e-06, "loss": 0.5226, "step": 10965 }, { "epoch": 1.71, "grad_norm": 15.132701464724713, "learning_rate": 8.197864317363926e-06, "loss": 0.5371, "step": 10966 }, { "epoch": 1.71, "grad_norm": 21.337489980956185, "learning_rate": 8.19620555457875e-06, "loss": 0.5384, "step": 10967 }, { "epoch": 1.71, "grad_norm": 14.497313984422355, "learning_rate": 8.194546843092438e-06, "loss": 0.5255, "step": 10968 }, { "epoch": 1.71, "grad_norm": 14.670657211968459, "learning_rate": 8.192888182952155e-06, "loss": 0.4893, "step": 10969 }, { "epoch": 1.71, "grad_norm": 19.008764870268358, "learning_rate": 8.191229574205078e-06, "loss": 0.5432, "step": 10970 }, { "epoch": 1.71, "grad_norm": 21.819143573333196, "learning_rate": 8.18957101689837e-06, "loss": 0.6081, "step": 10971 }, { "epoch": 1.71, "grad_norm": 26.228774009600517, "learning_rate": 8.187912511079205e-06, "loss": 0.5523, "step": 10972 }, { "epoch": 1.71, "grad_norm": 18.20772256216007, "learning_rate": 8.186254056794747e-06, "loss": 0.5818, "step": 10973 }, { "epoch": 1.71, "grad_norm": 21.06983739156714, "learning_rate": 8.184595654092161e-06, "loss": 0.5562, "step": 10974 }, { "epoch": 1.71, "grad_norm": 25.424231585890986, "learning_rate": 8.182937303018619e-06, "loss": 0.6361, "step": 10975 }, { "epoch": 1.71, "grad_norm": 19.16481744528762, "learning_rate": 8.181279003621274e-06, "loss": 0.5153, "step": 10976 }, { "epoch": 1.71, "grad_norm": 30.802604154974652, "learning_rate": 8.179620755947287e-06, "loss": 0.5733, "step": 10977 }, { "epoch": 1.71, "grad_norm": 16.742537468560272, "learning_rate": 8.177962560043824e-06, "loss": 0.6117, "step": 10978 }, { "epoch": 1.71, "grad_norm": 28.38632288276588, "learning_rate": 8.17630441595804e-06, "loss": 0.517, "step": 10979 }, { "epoch": 1.72, "grad_norm": 15.923263957167489, "learning_rate": 8.17464632373709e-06, "loss": 0.5239, "step": 10980 }, { "epoch": 1.72, "grad_norm": 19.490897551455543, "learning_rate": 8.172988283428132e-06, "loss": 0.5829, "step": 10981 }, { "epoch": 1.72, "grad_norm": 23.940319390755732, "learning_rate": 8.171330295078318e-06, "loss": 0.5117, "step": 10982 }, { "epoch": 1.72, "grad_norm": 19.43473985982462, "learning_rate": 8.1696723587348e-06, "loss": 0.6099, "step": 10983 }, { "epoch": 1.72, "grad_norm": 17.632587529878155, "learning_rate": 8.168014474444731e-06, "loss": 0.5982, "step": 10984 }, { "epoch": 1.72, "grad_norm": 20.675668228564096, "learning_rate": 8.166356642255258e-06, "loss": 0.584, "step": 10985 }, { "epoch": 1.72, "grad_norm": 20.01815195830891, "learning_rate": 8.164698862213527e-06, "loss": 0.6065, "step": 10986 }, { "epoch": 1.72, "grad_norm": 13.87360210561644, "learning_rate": 8.16304113436669e-06, "loss": 0.5353, "step": 10987 }, { "epoch": 1.72, "grad_norm": 18.041664347683344, "learning_rate": 8.161383458761887e-06, "loss": 0.5885, "step": 10988 }, { "epoch": 1.72, "grad_norm": 16.997703264741872, "learning_rate": 8.159725835446266e-06, "loss": 0.5921, "step": 10989 }, { "epoch": 1.72, "grad_norm": 27.729567331896416, "learning_rate": 8.15806826446696e-06, "loss": 0.5812, "step": 10990 }, { "epoch": 1.72, "grad_norm": 19.69375402328852, "learning_rate": 8.156410745871119e-06, "loss": 0.542, "step": 10991 }, { "epoch": 1.72, "grad_norm": 15.188270581676644, "learning_rate": 8.154753279705875e-06, "loss": 0.5461, "step": 10992 }, { "epoch": 1.72, "grad_norm": 24.136801629850993, "learning_rate": 8.15309586601837e-06, "loss": 0.5889, "step": 10993 }, { "epoch": 1.72, "grad_norm": 20.047274211675738, "learning_rate": 8.151438504855742e-06, "loss": 0.5705, "step": 10994 }, { "epoch": 1.72, "grad_norm": 17.155994891960784, "learning_rate": 8.149781196265117e-06, "loss": 0.5442, "step": 10995 }, { "epoch": 1.72, "grad_norm": 18.199399782235506, "learning_rate": 8.148123940293632e-06, "loss": 0.5261, "step": 10996 }, { "epoch": 1.72, "grad_norm": 21.092233468394266, "learning_rate": 8.146466736988422e-06, "loss": 0.5723, "step": 10997 }, { "epoch": 1.72, "grad_norm": 18.21096789645126, "learning_rate": 8.144809586396608e-06, "loss": 0.5656, "step": 10998 }, { "epoch": 1.72, "grad_norm": 14.233921976553873, "learning_rate": 8.143152488565332e-06, "loss": 0.5529, "step": 10999 }, { "epoch": 1.72, "grad_norm": 16.137995196396947, "learning_rate": 8.141495443541708e-06, "loss": 0.5669, "step": 11000 }, { "epoch": 1.72, "grad_norm": 20.661002369888248, "learning_rate": 8.139838451372868e-06, "loss": 0.534, "step": 11001 }, { "epoch": 1.72, "grad_norm": 25.811475719029517, "learning_rate": 8.138181512105931e-06, "loss": 0.5449, "step": 11002 }, { "epoch": 1.72, "grad_norm": 13.560522016809891, "learning_rate": 8.136524625788026e-06, "loss": 0.4659, "step": 11003 }, { "epoch": 1.72, "grad_norm": 18.104156449874377, "learning_rate": 8.134867792466272e-06, "loss": 0.5013, "step": 11004 }, { "epoch": 1.72, "grad_norm": 21.90485664669444, "learning_rate": 8.133211012187783e-06, "loss": 0.56, "step": 11005 }, { "epoch": 1.72, "grad_norm": 27.519362035032106, "learning_rate": 8.131554284999683e-06, "loss": 0.5678, "step": 11006 }, { "epoch": 1.72, "grad_norm": 21.27402890749563, "learning_rate": 8.129897610949086e-06, "loss": 0.502, "step": 11007 }, { "epoch": 1.72, "grad_norm": 25.254044915782583, "learning_rate": 8.128240990083104e-06, "loss": 0.6059, "step": 11008 }, { "epoch": 1.72, "grad_norm": 18.026716506973344, "learning_rate": 8.12658442244886e-06, "loss": 0.6353, "step": 11009 }, { "epoch": 1.72, "grad_norm": 23.869026824952083, "learning_rate": 8.124927908093455e-06, "loss": 0.4833, "step": 11010 }, { "epoch": 1.72, "grad_norm": 22.541212007840347, "learning_rate": 8.123271447064001e-06, "loss": 0.5512, "step": 11011 }, { "epoch": 1.72, "grad_norm": 24.436588693539342, "learning_rate": 8.121615039407613e-06, "loss": 0.5784, "step": 11012 }, { "epoch": 1.72, "grad_norm": 16.515478452900282, "learning_rate": 8.119958685171392e-06, "loss": 0.5607, "step": 11013 }, { "epoch": 1.72, "grad_norm": 19.27485765158441, "learning_rate": 8.11830238440245e-06, "loss": 0.5034, "step": 11014 }, { "epoch": 1.72, "grad_norm": 13.417919135522526, "learning_rate": 8.116646137147884e-06, "loss": 0.6027, "step": 11015 }, { "epoch": 1.72, "grad_norm": 14.62591573352725, "learning_rate": 8.114989943454801e-06, "loss": 0.5586, "step": 11016 }, { "epoch": 1.72, "grad_norm": 17.09641837558083, "learning_rate": 8.113333803370297e-06, "loss": 0.5503, "step": 11017 }, { "epoch": 1.72, "grad_norm": 27.03796482806511, "learning_rate": 8.111677716941481e-06, "loss": 0.5911, "step": 11018 }, { "epoch": 1.72, "grad_norm": 17.330280045255513, "learning_rate": 8.110021684215448e-06, "loss": 0.545, "step": 11019 }, { "epoch": 1.72, "grad_norm": 24.73453217201637, "learning_rate": 8.10836570523929e-06, "loss": 0.5973, "step": 11020 }, { "epoch": 1.72, "grad_norm": 14.325531198520604, "learning_rate": 8.106709780060103e-06, "loss": 0.5642, "step": 11021 }, { "epoch": 1.72, "grad_norm": 22.658649594400202, "learning_rate": 8.105053908724985e-06, "loss": 0.5774, "step": 11022 }, { "epoch": 1.72, "grad_norm": 17.606361811712446, "learning_rate": 8.103398091281023e-06, "loss": 0.5776, "step": 11023 }, { "epoch": 1.72, "grad_norm": 26.57523617190926, "learning_rate": 8.101742327775312e-06, "loss": 0.5409, "step": 11024 }, { "epoch": 1.72, "grad_norm": 16.01286506504692, "learning_rate": 8.10008661825494e-06, "loss": 0.5582, "step": 11025 }, { "epoch": 1.72, "grad_norm": 17.546346017841536, "learning_rate": 8.09843096276699e-06, "loss": 0.5599, "step": 11026 }, { "epoch": 1.72, "grad_norm": 21.43382008611894, "learning_rate": 8.096775361358552e-06, "loss": 0.5823, "step": 11027 }, { "epoch": 1.72, "grad_norm": 21.18612216552755, "learning_rate": 8.095119814076711e-06, "loss": 0.5712, "step": 11028 }, { "epoch": 1.72, "grad_norm": 17.592964890005003, "learning_rate": 8.093464320968547e-06, "loss": 0.5272, "step": 11029 }, { "epoch": 1.72, "grad_norm": 25.499410651680005, "learning_rate": 8.09180888208114e-06, "loss": 0.5841, "step": 11030 }, { "epoch": 1.72, "grad_norm": 19.41703335771521, "learning_rate": 8.090153497461573e-06, "loss": 0.6221, "step": 11031 }, { "epoch": 1.72, "grad_norm": 16.54118439577805, "learning_rate": 8.088498167156926e-06, "loss": 0.4778, "step": 11032 }, { "epoch": 1.72, "grad_norm": 26.262342527585577, "learning_rate": 8.086842891214274e-06, "loss": 0.5956, "step": 11033 }, { "epoch": 1.72, "grad_norm": 21.358215950784594, "learning_rate": 8.085187669680687e-06, "loss": 0.5643, "step": 11034 }, { "epoch": 1.72, "grad_norm": 13.09042014762866, "learning_rate": 8.083532502603243e-06, "loss": 0.5982, "step": 11035 }, { "epoch": 1.72, "grad_norm": 15.472888105967124, "learning_rate": 8.081877390029013e-06, "loss": 0.5203, "step": 11036 }, { "epoch": 1.72, "grad_norm": 20.36259903488524, "learning_rate": 8.08022233200507e-06, "loss": 0.615, "step": 11037 }, { "epoch": 1.72, "grad_norm": 19.797824194055696, "learning_rate": 8.078567328578482e-06, "loss": 0.5085, "step": 11038 }, { "epoch": 1.72, "grad_norm": 27.77186851093607, "learning_rate": 8.076912379796314e-06, "loss": 0.5671, "step": 11039 }, { "epoch": 1.72, "grad_norm": 18.64235028247073, "learning_rate": 8.07525748570563e-06, "loss": 0.5634, "step": 11040 }, { "epoch": 1.72, "grad_norm": 30.995082032506957, "learning_rate": 8.0736026463535e-06, "loss": 0.6884, "step": 11041 }, { "epoch": 1.72, "grad_norm": 18.509702462811813, "learning_rate": 8.071947861786982e-06, "loss": 0.5493, "step": 11042 }, { "epoch": 1.72, "grad_norm": 20.382712219564755, "learning_rate": 8.070293132053143e-06, "loss": 0.5787, "step": 11043 }, { "epoch": 1.73, "grad_norm": 13.06460155802515, "learning_rate": 8.068638457199037e-06, "loss": 0.5473, "step": 11044 }, { "epoch": 1.73, "grad_norm": 14.342234244697979, "learning_rate": 8.066983837271722e-06, "loss": 0.5552, "step": 11045 }, { "epoch": 1.73, "grad_norm": 16.61376142491823, "learning_rate": 8.065329272318255e-06, "loss": 0.6151, "step": 11046 }, { "epoch": 1.73, "grad_norm": 18.626125978110455, "learning_rate": 8.063674762385691e-06, "loss": 0.553, "step": 11047 }, { "epoch": 1.73, "grad_norm": 18.643831769416725, "learning_rate": 8.06202030752109e-06, "loss": 0.5936, "step": 11048 }, { "epoch": 1.73, "grad_norm": 26.662655345895203, "learning_rate": 8.06036590777149e-06, "loss": 0.5426, "step": 11049 }, { "epoch": 1.73, "grad_norm": 13.247502883389618, "learning_rate": 8.058711563183955e-06, "loss": 0.4977, "step": 11050 }, { "epoch": 1.73, "grad_norm": 18.289681275470457, "learning_rate": 8.057057273805525e-06, "loss": 0.5044, "step": 11051 }, { "epoch": 1.73, "grad_norm": 28.097842980216058, "learning_rate": 8.055403039683247e-06, "loss": 0.661, "step": 11052 }, { "epoch": 1.73, "grad_norm": 31.535869156552145, "learning_rate": 8.053748860864174e-06, "loss": 0.6149, "step": 11053 }, { "epoch": 1.73, "grad_norm": 17.49580416106502, "learning_rate": 8.052094737395343e-06, "loss": 0.5196, "step": 11054 }, { "epoch": 1.73, "grad_norm": 20.72396495256864, "learning_rate": 8.050440669323796e-06, "loss": 0.5294, "step": 11055 }, { "epoch": 1.73, "grad_norm": 17.234785005909544, "learning_rate": 8.048786656696579e-06, "loss": 0.6251, "step": 11056 }, { "epoch": 1.73, "grad_norm": 20.220223077620975, "learning_rate": 8.047132699560725e-06, "loss": 0.5548, "step": 11057 }, { "epoch": 1.73, "grad_norm": 17.359246668096613, "learning_rate": 8.04547879796328e-06, "loss": 0.57, "step": 11058 }, { "epoch": 1.73, "grad_norm": 18.67673597067759, "learning_rate": 8.04382495195127e-06, "loss": 0.5596, "step": 11059 }, { "epoch": 1.73, "grad_norm": 12.909389040797807, "learning_rate": 8.042171161571734e-06, "loss": 0.5106, "step": 11060 }, { "epoch": 1.73, "grad_norm": 20.535049992591745, "learning_rate": 8.040517426871703e-06, "loss": 0.5295, "step": 11061 }, { "epoch": 1.73, "grad_norm": 21.898674181429076, "learning_rate": 8.038863747898214e-06, "loss": 0.5764, "step": 11062 }, { "epoch": 1.73, "grad_norm": 24.233164874520643, "learning_rate": 8.037210124698294e-06, "loss": 0.5359, "step": 11063 }, { "epoch": 1.73, "grad_norm": 17.149139086680933, "learning_rate": 8.03555655731897e-06, "loss": 0.4864, "step": 11064 }, { "epoch": 1.73, "grad_norm": 21.420859164603485, "learning_rate": 8.033903045807262e-06, "loss": 0.5552, "step": 11065 }, { "epoch": 1.73, "grad_norm": 26.332369953580862, "learning_rate": 8.032249590210206e-06, "loss": 0.5435, "step": 11066 }, { "epoch": 1.73, "grad_norm": 16.057074256631132, "learning_rate": 8.030596190574818e-06, "loss": 0.592, "step": 11067 }, { "epoch": 1.73, "grad_norm": 23.374759156789157, "learning_rate": 8.028942846948126e-06, "loss": 0.5417, "step": 11068 }, { "epoch": 1.73, "grad_norm": 23.16419267022803, "learning_rate": 8.027289559377145e-06, "loss": 0.5208, "step": 11069 }, { "epoch": 1.73, "grad_norm": 14.173438081687703, "learning_rate": 8.025636327908895e-06, "loss": 0.5357, "step": 11070 }, { "epoch": 1.73, "grad_norm": 21.554575367103908, "learning_rate": 8.02398315259039e-06, "loss": 0.5781, "step": 11071 }, { "epoch": 1.73, "grad_norm": 12.602296407876821, "learning_rate": 8.02233003346865e-06, "loss": 0.4713, "step": 11072 }, { "epoch": 1.73, "grad_norm": 23.593174744849943, "learning_rate": 8.02067697059069e-06, "loss": 0.6106, "step": 11073 }, { "epoch": 1.73, "grad_norm": 27.016106065666666, "learning_rate": 8.019023964003513e-06, "loss": 0.5307, "step": 11074 }, { "epoch": 1.73, "grad_norm": 16.047224747190242, "learning_rate": 8.01737101375414e-06, "loss": 0.5816, "step": 11075 }, { "epoch": 1.73, "grad_norm": 14.723276295792006, "learning_rate": 8.015718119889576e-06, "loss": 0.5206, "step": 11076 }, { "epoch": 1.73, "grad_norm": 14.158261089440472, "learning_rate": 8.014065282456829e-06, "loss": 0.5438, "step": 11077 }, { "epoch": 1.73, "grad_norm": 23.76313922040878, "learning_rate": 8.0124125015029e-06, "loss": 0.5314, "step": 11078 }, { "epoch": 1.73, "grad_norm": 17.643851173575644, "learning_rate": 8.010759777074796e-06, "loss": 0.622, "step": 11079 }, { "epoch": 1.73, "grad_norm": 15.529692492113387, "learning_rate": 8.009107109219522e-06, "loss": 0.5564, "step": 11080 }, { "epoch": 1.73, "grad_norm": 16.519515489222094, "learning_rate": 8.007454497984078e-06, "loss": 0.5922, "step": 11081 }, { "epoch": 1.73, "grad_norm": 15.377649951148507, "learning_rate": 8.005801943415462e-06, "loss": 0.5335, "step": 11082 }, { "epoch": 1.73, "grad_norm": 25.989080296956804, "learning_rate": 8.004149445560675e-06, "loss": 0.5507, "step": 11083 }, { "epoch": 1.73, "grad_norm": 26.98497781704211, "learning_rate": 8.002497004466703e-06, "loss": 0.6189, "step": 11084 }, { "epoch": 1.73, "grad_norm": 16.69837751890889, "learning_rate": 8.000844620180553e-06, "loss": 0.6157, "step": 11085 }, { "epoch": 1.73, "grad_norm": 12.52657075512683, "learning_rate": 7.999192292749209e-06, "loss": 0.5011, "step": 11086 }, { "epoch": 1.73, "grad_norm": 20.089404243929433, "learning_rate": 7.997540022219672e-06, "loss": 0.6038, "step": 11087 }, { "epoch": 1.73, "grad_norm": 17.70165493375579, "learning_rate": 7.995887808638921e-06, "loss": 0.6156, "step": 11088 }, { "epoch": 1.73, "grad_norm": 21.615571556863518, "learning_rate": 7.99423565205395e-06, "loss": 0.5861, "step": 11089 }, { "epoch": 1.73, "grad_norm": 15.954431413536655, "learning_rate": 7.99258355251174e-06, "loss": 0.5519, "step": 11090 }, { "epoch": 1.73, "grad_norm": 17.510564341840514, "learning_rate": 7.990931510059281e-06, "loss": 0.5643, "step": 11091 }, { "epoch": 1.73, "grad_norm": 26.357858273602496, "learning_rate": 7.989279524743559e-06, "loss": 0.532, "step": 11092 }, { "epoch": 1.73, "grad_norm": 22.877190438779387, "learning_rate": 7.987627596611546e-06, "loss": 0.625, "step": 11093 }, { "epoch": 1.73, "grad_norm": 28.456516538218356, "learning_rate": 7.985975725710229e-06, "loss": 0.6112, "step": 11094 }, { "epoch": 1.73, "grad_norm": 14.390847832925022, "learning_rate": 7.984323912086584e-06, "loss": 0.5136, "step": 11095 }, { "epoch": 1.73, "grad_norm": 28.019716719828942, "learning_rate": 7.982672155787586e-06, "loss": 0.58, "step": 11096 }, { "epoch": 1.73, "grad_norm": 12.003477062195799, "learning_rate": 7.981020456860215e-06, "loss": 0.583, "step": 11097 }, { "epoch": 1.73, "grad_norm": 25.403469419461544, "learning_rate": 7.979368815351441e-06, "loss": 0.5502, "step": 11098 }, { "epoch": 1.73, "grad_norm": 17.524190774060003, "learning_rate": 7.977717231308233e-06, "loss": 0.4403, "step": 11099 }, { "epoch": 1.73, "grad_norm": 27.22380789500585, "learning_rate": 7.976065704777566e-06, "loss": 0.5688, "step": 11100 }, { "epoch": 1.73, "grad_norm": 24.66113094994034, "learning_rate": 7.974414235806403e-06, "loss": 0.5801, "step": 11101 }, { "epoch": 1.73, "grad_norm": 19.72633698551709, "learning_rate": 7.972762824441722e-06, "loss": 0.5958, "step": 11102 }, { "epoch": 1.73, "grad_norm": 23.229137339523575, "learning_rate": 7.971111470730474e-06, "loss": 0.5563, "step": 11103 }, { "epoch": 1.73, "grad_norm": 21.096566558281705, "learning_rate": 7.96946017471963e-06, "loss": 0.565, "step": 11104 }, { "epoch": 1.73, "grad_norm": 21.96685706485057, "learning_rate": 7.967808936456149e-06, "loss": 0.5936, "step": 11105 }, { "epoch": 1.73, "grad_norm": 23.269102688477492, "learning_rate": 7.966157755986994e-06, "loss": 0.6045, "step": 11106 }, { "epoch": 1.73, "grad_norm": 15.209484271912398, "learning_rate": 7.964506633359128e-06, "loss": 0.5603, "step": 11107 }, { "epoch": 1.74, "grad_norm": 20.844999399586634, "learning_rate": 7.962855568619499e-06, "loss": 0.5773, "step": 11108 }, { "epoch": 1.74, "grad_norm": 20.72385245923531, "learning_rate": 7.961204561815063e-06, "loss": 0.4946, "step": 11109 }, { "epoch": 1.74, "grad_norm": 14.107827367319523, "learning_rate": 7.95955361299278e-06, "loss": 0.5071, "step": 11110 }, { "epoch": 1.74, "grad_norm": 18.403746216851392, "learning_rate": 7.957902722199596e-06, "loss": 0.6051, "step": 11111 }, { "epoch": 1.74, "grad_norm": 35.29953233308446, "learning_rate": 7.956251889482467e-06, "loss": 0.5955, "step": 11112 }, { "epoch": 1.74, "grad_norm": 14.51073173821916, "learning_rate": 7.954601114888338e-06, "loss": 0.5469, "step": 11113 }, { "epoch": 1.74, "grad_norm": 21.361051458775115, "learning_rate": 7.952950398464156e-06, "loss": 0.5672, "step": 11114 }, { "epoch": 1.74, "grad_norm": 21.68827777891515, "learning_rate": 7.951299740256865e-06, "loss": 0.585, "step": 11115 }, { "epoch": 1.74, "grad_norm": 18.981508707375458, "learning_rate": 7.949649140313413e-06, "loss": 0.5476, "step": 11116 }, { "epoch": 1.74, "grad_norm": 15.699210480390972, "learning_rate": 7.947998598680743e-06, "loss": 0.5574, "step": 11117 }, { "epoch": 1.74, "grad_norm": 22.554949270353937, "learning_rate": 7.946348115405788e-06, "loss": 0.5768, "step": 11118 }, { "epoch": 1.74, "grad_norm": 20.234936766024894, "learning_rate": 7.944697690535493e-06, "loss": 0.5614, "step": 11119 }, { "epoch": 1.74, "grad_norm": 16.38727312207389, "learning_rate": 7.943047324116793e-06, "loss": 0.5479, "step": 11120 }, { "epoch": 1.74, "grad_norm": 16.145180182217953, "learning_rate": 7.941397016196622e-06, "loss": 0.5471, "step": 11121 }, { "epoch": 1.74, "grad_norm": 39.59424449322826, "learning_rate": 7.939746766821921e-06, "loss": 0.6819, "step": 11122 }, { "epoch": 1.74, "grad_norm": 22.470554431222713, "learning_rate": 7.938096576039613e-06, "loss": 0.5799, "step": 11123 }, { "epoch": 1.74, "grad_norm": 18.87952576742975, "learning_rate": 7.93644644389663e-06, "loss": 0.5676, "step": 11124 }, { "epoch": 1.74, "grad_norm": 13.474331828724617, "learning_rate": 7.934796370439907e-06, "loss": 0.5295, "step": 11125 }, { "epoch": 1.74, "grad_norm": 19.151053640199148, "learning_rate": 7.933146355716368e-06, "loss": 0.5266, "step": 11126 }, { "epoch": 1.74, "grad_norm": 21.41669492814009, "learning_rate": 7.931496399772935e-06, "loss": 0.5283, "step": 11127 }, { "epoch": 1.74, "grad_norm": 17.02942631083082, "learning_rate": 7.929846502656533e-06, "loss": 0.5478, "step": 11128 }, { "epoch": 1.74, "grad_norm": 17.93103972345277, "learning_rate": 7.928196664414088e-06, "loss": 0.5782, "step": 11129 }, { "epoch": 1.74, "grad_norm": 17.59302446290554, "learning_rate": 7.926546885092515e-06, "loss": 0.5762, "step": 11130 }, { "epoch": 1.74, "grad_norm": 12.436897528615656, "learning_rate": 7.92489716473874e-06, "loss": 0.5658, "step": 11131 }, { "epoch": 1.74, "grad_norm": 13.341677152386612, "learning_rate": 7.923247503399674e-06, "loss": 0.5375, "step": 11132 }, { "epoch": 1.74, "grad_norm": 15.199581514459307, "learning_rate": 7.921597901122234e-06, "loss": 0.6023, "step": 11133 }, { "epoch": 1.74, "grad_norm": 21.14620217712568, "learning_rate": 7.919948357953332e-06, "loss": 0.5263, "step": 11134 }, { "epoch": 1.74, "grad_norm": 19.87994058848624, "learning_rate": 7.918298873939882e-06, "loss": 0.6351, "step": 11135 }, { "epoch": 1.74, "grad_norm": 15.894872621895544, "learning_rate": 7.916649449128799e-06, "loss": 0.5342, "step": 11136 }, { "epoch": 1.74, "grad_norm": 20.574830035472903, "learning_rate": 7.915000083566981e-06, "loss": 0.5463, "step": 11137 }, { "epoch": 1.74, "grad_norm": 13.873451409610581, "learning_rate": 7.913350777301342e-06, "loss": 0.5196, "step": 11138 }, { "epoch": 1.74, "grad_norm": 20.12385858459081, "learning_rate": 7.911701530378789e-06, "loss": 0.5146, "step": 11139 }, { "epoch": 1.74, "grad_norm": 25.481924751387364, "learning_rate": 7.910052342846219e-06, "loss": 0.5504, "step": 11140 }, { "epoch": 1.74, "grad_norm": 13.437353532569539, "learning_rate": 7.908403214750544e-06, "loss": 0.5345, "step": 11141 }, { "epoch": 1.74, "grad_norm": 17.383152699516856, "learning_rate": 7.906754146138655e-06, "loss": 0.5909, "step": 11142 }, { "epoch": 1.74, "grad_norm": 36.781543884667066, "learning_rate": 7.905105137057451e-06, "loss": 0.5767, "step": 11143 }, { "epoch": 1.74, "grad_norm": 22.89474109448701, "learning_rate": 7.903456187553833e-06, "loss": 0.6353, "step": 11144 }, { "epoch": 1.74, "grad_norm": 16.728992731936774, "learning_rate": 7.901807297674693e-06, "loss": 0.5532, "step": 11145 }, { "epoch": 1.74, "grad_norm": 15.184728292008584, "learning_rate": 7.900158467466932e-06, "loss": 0.4741, "step": 11146 }, { "epoch": 1.74, "grad_norm": 21.181395474145667, "learning_rate": 7.89850969697743e-06, "loss": 0.5903, "step": 11147 }, { "epoch": 1.74, "grad_norm": 16.120405041472555, "learning_rate": 7.896860986253086e-06, "loss": 0.5352, "step": 11148 }, { "epoch": 1.74, "grad_norm": 19.06258247740679, "learning_rate": 7.89521233534078e-06, "loss": 0.5398, "step": 11149 }, { "epoch": 1.74, "grad_norm": 19.13804934630115, "learning_rate": 7.89356374428741e-06, "loss": 0.5708, "step": 11150 }, { "epoch": 1.74, "grad_norm": 21.187892427720595, "learning_rate": 7.891915213139855e-06, "loss": 0.5356, "step": 11151 }, { "epoch": 1.74, "grad_norm": 18.343560072420896, "learning_rate": 7.890266741944995e-06, "loss": 0.5867, "step": 11152 }, { "epoch": 1.74, "grad_norm": 18.77213299013415, "learning_rate": 7.888618330749715e-06, "loss": 0.5711, "step": 11153 }, { "epoch": 1.74, "grad_norm": 15.805220979143153, "learning_rate": 7.886969979600894e-06, "loss": 0.5862, "step": 11154 }, { "epoch": 1.74, "grad_norm": 14.429068554675853, "learning_rate": 7.885321688545412e-06, "loss": 0.5516, "step": 11155 }, { "epoch": 1.74, "grad_norm": 16.692821746198884, "learning_rate": 7.883673457630144e-06, "loss": 0.503, "step": 11156 }, { "epoch": 1.74, "grad_norm": 21.367965746136644, "learning_rate": 7.882025286901965e-06, "loss": 0.5323, "step": 11157 }, { "epoch": 1.74, "grad_norm": 23.50679528370471, "learning_rate": 7.880377176407749e-06, "loss": 0.5922, "step": 11158 }, { "epoch": 1.74, "grad_norm": 17.70563036251502, "learning_rate": 7.878729126194362e-06, "loss": 0.5291, "step": 11159 }, { "epoch": 1.74, "grad_norm": 15.653253196150233, "learning_rate": 7.877081136308681e-06, "loss": 0.5786, "step": 11160 }, { "epoch": 1.74, "grad_norm": 19.693096101244436, "learning_rate": 7.875433206797573e-06, "loss": 0.6015, "step": 11161 }, { "epoch": 1.74, "grad_norm": 19.594477678776837, "learning_rate": 7.873785337707897e-06, "loss": 0.5464, "step": 11162 }, { "epoch": 1.74, "grad_norm": 18.992648678118947, "learning_rate": 7.872137529086525e-06, "loss": 0.594, "step": 11163 }, { "epoch": 1.74, "grad_norm": 14.856251143566597, "learning_rate": 7.870489780980315e-06, "loss": 0.5132, "step": 11164 }, { "epoch": 1.74, "grad_norm": 15.567361704222945, "learning_rate": 7.86884209343613e-06, "loss": 0.5292, "step": 11165 }, { "epoch": 1.74, "grad_norm": 23.6082497971756, "learning_rate": 7.867194466500831e-06, "loss": 0.5751, "step": 11166 }, { "epoch": 1.74, "grad_norm": 23.46617471518683, "learning_rate": 7.865546900221274e-06, "loss": 0.5088, "step": 11167 }, { "epoch": 1.74, "grad_norm": 18.359587600775672, "learning_rate": 7.863899394644312e-06, "loss": 0.6115, "step": 11168 }, { "epoch": 1.74, "grad_norm": 16.852719261937743, "learning_rate": 7.862251949816803e-06, "loss": 0.5957, "step": 11169 }, { "epoch": 1.74, "grad_norm": 14.061317520228739, "learning_rate": 7.860604565785596e-06, "loss": 0.5912, "step": 11170 }, { "epoch": 1.74, "grad_norm": 16.93777975066035, "learning_rate": 7.85895724259755e-06, "loss": 0.5291, "step": 11171 }, { "epoch": 1.75, "grad_norm": 25.539383992995287, "learning_rate": 7.857309980299502e-06, "loss": 0.5292, "step": 11172 }, { "epoch": 1.75, "grad_norm": 23.411902252602395, "learning_rate": 7.855662778938306e-06, "loss": 0.5422, "step": 11173 }, { "epoch": 1.75, "grad_norm": 14.621579744342313, "learning_rate": 7.854015638560804e-06, "loss": 0.5601, "step": 11174 }, { "epoch": 1.75, "grad_norm": 23.57756516314625, "learning_rate": 7.852368559213846e-06, "loss": 0.6073, "step": 11175 }, { "epoch": 1.75, "grad_norm": 14.900995452002077, "learning_rate": 7.850721540944268e-06, "loss": 0.5832, "step": 11176 }, { "epoch": 1.75, "grad_norm": 19.737593953849398, "learning_rate": 7.849074583798912e-06, "loss": 0.547, "step": 11177 }, { "epoch": 1.75, "grad_norm": 16.23704238232806, "learning_rate": 7.847427687824615e-06, "loss": 0.4946, "step": 11178 }, { "epoch": 1.75, "grad_norm": 25.272781334072516, "learning_rate": 7.845780853068217e-06, "loss": 0.5513, "step": 11179 }, { "epoch": 1.75, "grad_norm": 13.3597406125372, "learning_rate": 7.844134079576553e-06, "loss": 0.5056, "step": 11180 }, { "epoch": 1.75, "grad_norm": 15.034055145233486, "learning_rate": 7.84248736739645e-06, "loss": 0.4983, "step": 11181 }, { "epoch": 1.75, "grad_norm": 14.196566362067875, "learning_rate": 7.840840716574748e-06, "loss": 0.5368, "step": 11182 }, { "epoch": 1.75, "grad_norm": 10.627838873271484, "learning_rate": 7.839194127158271e-06, "loss": 0.5258, "step": 11183 }, { "epoch": 1.75, "grad_norm": 15.224867405885616, "learning_rate": 7.837547599193849e-06, "loss": 0.5772, "step": 11184 }, { "epoch": 1.75, "grad_norm": 13.95064474362585, "learning_rate": 7.835901132728311e-06, "loss": 0.4517, "step": 11185 }, { "epoch": 1.75, "grad_norm": 16.45076166153509, "learning_rate": 7.834254727808477e-06, "loss": 0.5257, "step": 11186 }, { "epoch": 1.75, "grad_norm": 15.820391013810106, "learning_rate": 7.83260838448117e-06, "loss": 0.5656, "step": 11187 }, { "epoch": 1.75, "grad_norm": 23.667336873699867, "learning_rate": 7.830962102793214e-06, "loss": 0.6235, "step": 11188 }, { "epoch": 1.75, "grad_norm": 11.754841785972543, "learning_rate": 7.829315882791426e-06, "loss": 0.5182, "step": 11189 }, { "epoch": 1.75, "grad_norm": 22.898584723934768, "learning_rate": 7.827669724522632e-06, "loss": 0.5112, "step": 11190 }, { "epoch": 1.75, "grad_norm": 20.03620929445909, "learning_rate": 7.826023628033631e-06, "loss": 0.5311, "step": 11191 }, { "epoch": 1.75, "grad_norm": 24.418773772718986, "learning_rate": 7.824377593371248e-06, "loss": 0.5441, "step": 11192 }, { "epoch": 1.75, "grad_norm": 20.220247998822664, "learning_rate": 7.822731620582294e-06, "loss": 0.562, "step": 11193 }, { "epoch": 1.75, "grad_norm": 17.274615775975082, "learning_rate": 7.821085709713581e-06, "loss": 0.6609, "step": 11194 }, { "epoch": 1.75, "grad_norm": 20.167098561320827, "learning_rate": 7.819439860811915e-06, "loss": 0.5042, "step": 11195 }, { "epoch": 1.75, "grad_norm": 21.093025764465423, "learning_rate": 7.817794073924104e-06, "loss": 0.5515, "step": 11196 }, { "epoch": 1.75, "grad_norm": 18.877998192084533, "learning_rate": 7.816148349096951e-06, "loss": 0.5659, "step": 11197 }, { "epoch": 1.75, "grad_norm": 17.18021799363955, "learning_rate": 7.814502686377263e-06, "loss": 0.5763, "step": 11198 }, { "epoch": 1.75, "grad_norm": 18.60497919127739, "learning_rate": 7.812857085811837e-06, "loss": 0.5274, "step": 11199 }, { "epoch": 1.75, "grad_norm": 24.518095578832366, "learning_rate": 7.811211547447483e-06, "loss": 0.5378, "step": 11200 }, { "epoch": 1.75, "grad_norm": 12.476006045798519, "learning_rate": 7.809566071330987e-06, "loss": 0.5343, "step": 11201 }, { "epoch": 1.75, "grad_norm": 15.502766660779288, "learning_rate": 7.807920657509152e-06, "loss": 0.5502, "step": 11202 }, { "epoch": 1.75, "grad_norm": 15.12532621615114, "learning_rate": 7.80627530602877e-06, "loss": 0.6286, "step": 11203 }, { "epoch": 1.75, "grad_norm": 12.686116191255312, "learning_rate": 7.804630016936636e-06, "loss": 0.5292, "step": 11204 }, { "epoch": 1.75, "grad_norm": 18.902276441245746, "learning_rate": 7.802984790279542e-06, "loss": 0.5782, "step": 11205 }, { "epoch": 1.75, "grad_norm": 17.75679950367359, "learning_rate": 7.801339626104272e-06, "loss": 0.5057, "step": 11206 }, { "epoch": 1.75, "grad_norm": 25.95664332252799, "learning_rate": 7.799694524457616e-06, "loss": 0.5221, "step": 11207 }, { "epoch": 1.75, "grad_norm": 17.29322166637064, "learning_rate": 7.798049485386365e-06, "loss": 0.6383, "step": 11208 }, { "epoch": 1.75, "grad_norm": 18.69645899817208, "learning_rate": 7.796404508937295e-06, "loss": 0.5263, "step": 11209 }, { "epoch": 1.75, "grad_norm": 17.98804970425537, "learning_rate": 7.794759595157194e-06, "loss": 0.5728, "step": 11210 }, { "epoch": 1.75, "grad_norm": 14.8381526613064, "learning_rate": 7.79311474409284e-06, "loss": 0.5681, "step": 11211 }, { "epoch": 1.75, "grad_norm": 19.553867530167153, "learning_rate": 7.791469955791009e-06, "loss": 0.5167, "step": 11212 }, { "epoch": 1.75, "grad_norm": 28.675711620388135, "learning_rate": 7.789825230298481e-06, "loss": 0.5042, "step": 11213 }, { "epoch": 1.75, "grad_norm": 24.621927666949734, "learning_rate": 7.788180567662031e-06, "loss": 0.5363, "step": 11214 }, { "epoch": 1.75, "grad_norm": 13.111166937280643, "learning_rate": 7.786535967928437e-06, "loss": 0.5568, "step": 11215 }, { "epoch": 1.75, "grad_norm": 19.22113113849329, "learning_rate": 7.784891431144459e-06, "loss": 0.5156, "step": 11216 }, { "epoch": 1.75, "grad_norm": 16.3711243235276, "learning_rate": 7.783246957356876e-06, "loss": 0.4867, "step": 11217 }, { "epoch": 1.75, "grad_norm": 19.481162776941453, "learning_rate": 7.781602546612447e-06, "loss": 0.5881, "step": 11218 }, { "epoch": 1.75, "grad_norm": 18.39490523396599, "learning_rate": 7.779958198957951e-06, "loss": 0.5699, "step": 11219 }, { "epoch": 1.75, "grad_norm": 18.988333990215676, "learning_rate": 7.778313914440143e-06, "loss": 0.5531, "step": 11220 }, { "epoch": 1.75, "grad_norm": 26.344981249615987, "learning_rate": 7.776669693105786e-06, "loss": 0.6281, "step": 11221 }, { "epoch": 1.75, "grad_norm": 20.98876758312302, "learning_rate": 7.775025535001643e-06, "loss": 0.6004, "step": 11222 }, { "epoch": 1.75, "grad_norm": 15.00271330621962, "learning_rate": 7.773381440174473e-06, "loss": 0.5866, "step": 11223 }, { "epoch": 1.75, "grad_norm": 17.538014427764445, "learning_rate": 7.771737408671034e-06, "loss": 0.5326, "step": 11224 }, { "epoch": 1.75, "grad_norm": 13.40138392845931, "learning_rate": 7.770093440538076e-06, "loss": 0.6238, "step": 11225 }, { "epoch": 1.75, "grad_norm": 19.467097297984274, "learning_rate": 7.768449535822357e-06, "loss": 0.5971, "step": 11226 }, { "epoch": 1.75, "grad_norm": 16.94261482295833, "learning_rate": 7.766805694570629e-06, "loss": 0.5379, "step": 11227 }, { "epoch": 1.75, "grad_norm": 16.03707658929044, "learning_rate": 7.76516191682964e-06, "loss": 0.5704, "step": 11228 }, { "epoch": 1.75, "grad_norm": 22.489801493630956, "learning_rate": 7.76351820264614e-06, "loss": 0.5618, "step": 11229 }, { "epoch": 1.75, "grad_norm": 39.26357738622217, "learning_rate": 7.761874552066873e-06, "loss": 0.6505, "step": 11230 }, { "epoch": 1.75, "grad_norm": 30.675719590576257, "learning_rate": 7.760230965138582e-06, "loss": 0.6188, "step": 11231 }, { "epoch": 1.75, "grad_norm": 18.640713466872427, "learning_rate": 7.758587441908014e-06, "loss": 0.5371, "step": 11232 }, { "epoch": 1.75, "grad_norm": 29.7058679482082, "learning_rate": 7.756943982421907e-06, "loss": 0.6301, "step": 11233 }, { "epoch": 1.75, "grad_norm": 17.67222206237001, "learning_rate": 7.755300586727007e-06, "loss": 0.5135, "step": 11234 }, { "epoch": 1.75, "grad_norm": 19.478443852101808, "learning_rate": 7.753657254870038e-06, "loss": 0.576, "step": 11235 }, { "epoch": 1.76, "grad_norm": 32.647392756461734, "learning_rate": 7.752013986897747e-06, "loss": 0.6269, "step": 11236 }, { "epoch": 1.76, "grad_norm": 15.111928499682369, "learning_rate": 7.750370782856858e-06, "loss": 0.4867, "step": 11237 }, { "epoch": 1.76, "grad_norm": 17.0842524601561, "learning_rate": 7.748727642794111e-06, "loss": 0.5763, "step": 11238 }, { "epoch": 1.76, "grad_norm": 23.249422957709474, "learning_rate": 7.747084566756237e-06, "loss": 0.6144, "step": 11239 }, { "epoch": 1.76, "grad_norm": 21.60997163917072, "learning_rate": 7.745441554789956e-06, "loss": 0.4776, "step": 11240 }, { "epoch": 1.76, "grad_norm": 9.54723510593718, "learning_rate": 7.743798606941997e-06, "loss": 0.5605, "step": 11241 }, { "epoch": 1.76, "grad_norm": 17.03156441258331, "learning_rate": 7.742155723259086e-06, "loss": 0.6029, "step": 11242 }, { "epoch": 1.76, "grad_norm": 15.844558310768933, "learning_rate": 7.740512903787945e-06, "loss": 0.5783, "step": 11243 }, { "epoch": 1.76, "grad_norm": 20.290430426064916, "learning_rate": 7.738870148575299e-06, "loss": 0.5824, "step": 11244 }, { "epoch": 1.76, "grad_norm": 14.021835985888213, "learning_rate": 7.737227457667861e-06, "loss": 0.6041, "step": 11245 }, { "epoch": 1.76, "grad_norm": 21.910213307840092, "learning_rate": 7.735584831112352e-06, "loss": 0.5481, "step": 11246 }, { "epoch": 1.76, "grad_norm": 23.54598192117341, "learning_rate": 7.733942268955482e-06, "loss": 0.5601, "step": 11247 }, { "epoch": 1.76, "grad_norm": 15.875992057808219, "learning_rate": 7.732299771243972e-06, "loss": 0.6435, "step": 11248 }, { "epoch": 1.76, "grad_norm": 20.111678844400902, "learning_rate": 7.73065733802453e-06, "loss": 0.5505, "step": 11249 }, { "epoch": 1.76, "grad_norm": 20.713366736911624, "learning_rate": 7.729014969343863e-06, "loss": 0.5367, "step": 11250 }, { "epoch": 1.76, "grad_norm": 14.51284570335611, "learning_rate": 7.727372665248682e-06, "loss": 0.5855, "step": 11251 }, { "epoch": 1.76, "grad_norm": 25.156948780045354, "learning_rate": 7.725730425785694e-06, "loss": 0.6051, "step": 11252 }, { "epoch": 1.76, "grad_norm": 19.73512527900115, "learning_rate": 7.724088251001602e-06, "loss": 0.5376, "step": 11253 }, { "epoch": 1.76, "grad_norm": 15.11760176748968, "learning_rate": 7.722446140943111e-06, "loss": 0.5813, "step": 11254 }, { "epoch": 1.76, "grad_norm": 16.03475619663871, "learning_rate": 7.720804095656918e-06, "loss": 0.4859, "step": 11255 }, { "epoch": 1.76, "grad_norm": 14.216014199055966, "learning_rate": 7.71916211518972e-06, "loss": 0.5153, "step": 11256 }, { "epoch": 1.76, "grad_norm": 18.174982535027215, "learning_rate": 7.717520199588222e-06, "loss": 0.5669, "step": 11257 }, { "epoch": 1.76, "grad_norm": 18.71978624911367, "learning_rate": 7.71587834889911e-06, "loss": 0.5883, "step": 11258 }, { "epoch": 1.76, "grad_norm": 17.758882855347803, "learning_rate": 7.714236563169088e-06, "loss": 0.4991, "step": 11259 }, { "epoch": 1.76, "grad_norm": 19.58478125235607, "learning_rate": 7.712594842444836e-06, "loss": 0.4861, "step": 11260 }, { "epoch": 1.76, "grad_norm": 14.183877571773113, "learning_rate": 7.710953186773048e-06, "loss": 0.6171, "step": 11261 }, { "epoch": 1.76, "grad_norm": 13.82640345035016, "learning_rate": 7.709311596200412e-06, "loss": 0.5654, "step": 11262 }, { "epoch": 1.76, "grad_norm": 30.559027839765587, "learning_rate": 7.707670070773616e-06, "loss": 0.5939, "step": 11263 }, { "epoch": 1.76, "grad_norm": 28.142003183966946, "learning_rate": 7.706028610539345e-06, "loss": 0.6379, "step": 11264 }, { "epoch": 1.76, "grad_norm": 18.554621440569086, "learning_rate": 7.704387215544276e-06, "loss": 0.5122, "step": 11265 }, { "epoch": 1.76, "grad_norm": 21.55907975554662, "learning_rate": 7.70274588583509e-06, "loss": 0.5367, "step": 11266 }, { "epoch": 1.76, "grad_norm": 17.607315172325947, "learning_rate": 7.701104621458467e-06, "loss": 0.5735, "step": 11267 }, { "epoch": 1.76, "grad_norm": 19.2734713567615, "learning_rate": 7.69946342246109e-06, "loss": 0.5533, "step": 11268 }, { "epoch": 1.76, "grad_norm": 19.717117067772456, "learning_rate": 7.697822288889617e-06, "loss": 0.5291, "step": 11269 }, { "epoch": 1.76, "grad_norm": 19.384736506442632, "learning_rate": 7.69618122079074e-06, "loss": 0.5018, "step": 11270 }, { "epoch": 1.76, "grad_norm": 23.973631029852697, "learning_rate": 7.694540218211117e-06, "loss": 0.563, "step": 11271 }, { "epoch": 1.76, "grad_norm": 18.628703270044507, "learning_rate": 7.692899281197421e-06, "loss": 0.4823, "step": 11272 }, { "epoch": 1.76, "grad_norm": 17.57420443406871, "learning_rate": 7.691258409796324e-06, "loss": 0.4832, "step": 11273 }, { "epoch": 1.76, "grad_norm": 13.447646746363091, "learning_rate": 7.689617604054487e-06, "loss": 0.5394, "step": 11274 }, { "epoch": 1.76, "grad_norm": 19.838827575073203, "learning_rate": 7.68797686401857e-06, "loss": 0.5503, "step": 11275 }, { "epoch": 1.76, "grad_norm": 19.091192857792542, "learning_rate": 7.68633618973524e-06, "loss": 0.4895, "step": 11276 }, { "epoch": 1.76, "grad_norm": 20.817442342235797, "learning_rate": 7.684695581251153e-06, "loss": 0.5912, "step": 11277 }, { "epoch": 1.76, "grad_norm": 23.593332278109564, "learning_rate": 7.683055038612977e-06, "loss": 0.6045, "step": 11278 }, { "epoch": 1.76, "grad_norm": 21.482524264990158, "learning_rate": 7.681414561867353e-06, "loss": 0.4914, "step": 11279 }, { "epoch": 1.76, "grad_norm": 23.650770275846927, "learning_rate": 7.679774151060945e-06, "loss": 0.5341, "step": 11280 }, { "epoch": 1.76, "grad_norm": 25.13054877720958, "learning_rate": 7.6781338062404e-06, "loss": 0.5412, "step": 11281 }, { "epoch": 1.76, "grad_norm": 21.12693658041485, "learning_rate": 7.676493527452374e-06, "loss": 0.5932, "step": 11282 }, { "epoch": 1.76, "grad_norm": 19.663614451730098, "learning_rate": 7.674853314743516e-06, "loss": 0.4899, "step": 11283 }, { "epoch": 1.76, "grad_norm": 24.244453653723927, "learning_rate": 7.673213168160465e-06, "loss": 0.5234, "step": 11284 }, { "epoch": 1.76, "grad_norm": 11.741632808532874, "learning_rate": 7.671573087749871e-06, "loss": 0.4883, "step": 11285 }, { "epoch": 1.76, "grad_norm": 16.626228054121917, "learning_rate": 7.669933073558377e-06, "loss": 0.6313, "step": 11286 }, { "epoch": 1.76, "grad_norm": 16.419755841583367, "learning_rate": 7.668293125632621e-06, "loss": 0.5558, "step": 11287 }, { "epoch": 1.76, "grad_norm": 24.203634339225513, "learning_rate": 7.66665324401925e-06, "loss": 0.6161, "step": 11288 }, { "epoch": 1.76, "grad_norm": 24.611350398632172, "learning_rate": 7.665013428764892e-06, "loss": 0.5002, "step": 11289 }, { "epoch": 1.76, "grad_norm": 24.788121986761205, "learning_rate": 7.663373679916187e-06, "loss": 0.5877, "step": 11290 }, { "epoch": 1.76, "grad_norm": 45.5453902615997, "learning_rate": 7.661733997519765e-06, "loss": 0.676, "step": 11291 }, { "epoch": 1.76, "grad_norm": 21.680113970711655, "learning_rate": 7.660094381622265e-06, "loss": 0.5259, "step": 11292 }, { "epoch": 1.76, "grad_norm": 18.286117485253868, "learning_rate": 7.658454832270312e-06, "loss": 0.5003, "step": 11293 }, { "epoch": 1.76, "grad_norm": 18.25049209012945, "learning_rate": 7.656815349510528e-06, "loss": 0.5075, "step": 11294 }, { "epoch": 1.76, "grad_norm": 24.28092716007329, "learning_rate": 7.65517593338955e-06, "loss": 0.5332, "step": 11295 }, { "epoch": 1.76, "grad_norm": 26.04700770455865, "learning_rate": 7.653536583953993e-06, "loss": 0.593, "step": 11296 }, { "epoch": 1.76, "grad_norm": 18.678745228459746, "learning_rate": 7.651897301250484e-06, "loss": 0.5501, "step": 11297 }, { "epoch": 1.76, "grad_norm": 13.003487137617729, "learning_rate": 7.650258085325646e-06, "loss": 0.5254, "step": 11298 }, { "epoch": 1.76, "grad_norm": 18.865165146944452, "learning_rate": 7.64861893622609e-06, "loss": 0.5383, "step": 11299 }, { "epoch": 1.77, "grad_norm": 18.664046491118665, "learning_rate": 7.646979853998434e-06, "loss": 0.5551, "step": 11300 }, { "epoch": 1.77, "grad_norm": 20.161354677915533, "learning_rate": 7.645340838689297e-06, "loss": 0.5478, "step": 11301 }, { "epoch": 1.77, "grad_norm": 17.640657624247467, "learning_rate": 7.643701890345288e-06, "loss": 0.5305, "step": 11302 }, { "epoch": 1.77, "grad_norm": 18.743547150944273, "learning_rate": 7.642063009013025e-06, "loss": 0.5061, "step": 11303 }, { "epoch": 1.77, "grad_norm": 21.06998142425038, "learning_rate": 7.640424194739102e-06, "loss": 0.5848, "step": 11304 }, { "epoch": 1.77, "grad_norm": 22.29649272435952, "learning_rate": 7.63878544757014e-06, "loss": 0.6403, "step": 11305 }, { "epoch": 1.77, "grad_norm": 13.715199545490968, "learning_rate": 7.637146767552735e-06, "loss": 0.5738, "step": 11306 }, { "epoch": 1.77, "grad_norm": 25.079013950718778, "learning_rate": 7.635508154733494e-06, "loss": 0.5993, "step": 11307 }, { "epoch": 1.77, "grad_norm": 18.884186294227668, "learning_rate": 7.633869609159023e-06, "loss": 0.5479, "step": 11308 }, { "epoch": 1.77, "grad_norm": 16.411200973428215, "learning_rate": 7.632231130875912e-06, "loss": 0.4966, "step": 11309 }, { "epoch": 1.77, "grad_norm": 21.025665686084285, "learning_rate": 7.63059271993076e-06, "loss": 0.493, "step": 11310 }, { "epoch": 1.77, "grad_norm": 19.17998847181433, "learning_rate": 7.628954376370169e-06, "loss": 0.5698, "step": 11311 }, { "epoch": 1.77, "grad_norm": 20.30178264902838, "learning_rate": 7.627316100240726e-06, "loss": 0.62, "step": 11312 }, { "epoch": 1.77, "grad_norm": 19.115320294558252, "learning_rate": 7.625677891589029e-06, "loss": 0.546, "step": 11313 }, { "epoch": 1.77, "grad_norm": 10.920930267645089, "learning_rate": 7.6240397504616615e-06, "loss": 0.5092, "step": 11314 }, { "epoch": 1.77, "grad_norm": 23.25450947570427, "learning_rate": 7.622401676905214e-06, "loss": 0.5265, "step": 11315 }, { "epoch": 1.77, "grad_norm": 22.207353591928037, "learning_rate": 7.6207636709662694e-06, "loss": 0.5301, "step": 11316 }, { "epoch": 1.77, "grad_norm": 21.44018048163546, "learning_rate": 7.619125732691419e-06, "loss": 0.5206, "step": 11317 }, { "epoch": 1.77, "grad_norm": 29.436667042284412, "learning_rate": 7.617487862127238e-06, "loss": 0.6249, "step": 11318 }, { "epoch": 1.77, "grad_norm": 20.89768052200801, "learning_rate": 7.615850059320306e-06, "loss": 0.5658, "step": 11319 }, { "epoch": 1.77, "grad_norm": 15.396333062932055, "learning_rate": 7.614212324317205e-06, "loss": 0.5533, "step": 11320 }, { "epoch": 1.77, "grad_norm": 15.466935711726588, "learning_rate": 7.612574657164509e-06, "loss": 0.5844, "step": 11321 }, { "epoch": 1.77, "grad_norm": 15.202536005017654, "learning_rate": 7.610937057908798e-06, "loss": 0.5077, "step": 11322 }, { "epoch": 1.77, "grad_norm": 17.427338583373395, "learning_rate": 7.609299526596634e-06, "loss": 0.5474, "step": 11323 }, { "epoch": 1.77, "grad_norm": 20.521451742194035, "learning_rate": 7.607662063274595e-06, "loss": 0.5075, "step": 11324 }, { "epoch": 1.77, "grad_norm": 17.76787564709022, "learning_rate": 7.606024667989246e-06, "loss": 0.5888, "step": 11325 }, { "epoch": 1.77, "grad_norm": 18.35670487486769, "learning_rate": 7.604387340787156e-06, "loss": 0.6043, "step": 11326 }, { "epoch": 1.77, "grad_norm": 16.377373961453696, "learning_rate": 7.6027500817148915e-06, "loss": 0.5456, "step": 11327 }, { "epoch": 1.77, "grad_norm": 24.70214588944195, "learning_rate": 7.6011128908190105e-06, "loss": 0.5512, "step": 11328 }, { "epoch": 1.77, "grad_norm": 21.17046677526658, "learning_rate": 7.599475768146072e-06, "loss": 0.5254, "step": 11329 }, { "epoch": 1.77, "grad_norm": 28.881291500838543, "learning_rate": 7.5978387137426425e-06, "loss": 0.5213, "step": 11330 }, { "epoch": 1.77, "grad_norm": 27.06366082855771, "learning_rate": 7.59620172765527e-06, "loss": 0.6224, "step": 11331 }, { "epoch": 1.77, "grad_norm": 17.986773955641482, "learning_rate": 7.5945648099305206e-06, "loss": 0.536, "step": 11332 }, { "epoch": 1.77, "grad_norm": 13.104959198871104, "learning_rate": 7.592927960614939e-06, "loss": 0.5259, "step": 11333 }, { "epoch": 1.77, "grad_norm": 27.427609701510946, "learning_rate": 7.591291179755077e-06, "loss": 0.5567, "step": 11334 }, { "epoch": 1.77, "grad_norm": 30.908048655047114, "learning_rate": 7.5896544673974825e-06, "loss": 0.6181, "step": 11335 }, { "epoch": 1.77, "grad_norm": 13.155516628645152, "learning_rate": 7.588017823588707e-06, "loss": 0.5286, "step": 11336 }, { "epoch": 1.77, "grad_norm": 13.862126498204118, "learning_rate": 7.586381248375296e-06, "loss": 0.542, "step": 11337 }, { "epoch": 1.77, "grad_norm": 22.68626587114657, "learning_rate": 7.584744741803787e-06, "loss": 0.6334, "step": 11338 }, { "epoch": 1.77, "grad_norm": 13.741204965630933, "learning_rate": 7.583108303920726e-06, "loss": 0.5653, "step": 11339 }, { "epoch": 1.77, "grad_norm": 16.622610320171088, "learning_rate": 7.581471934772649e-06, "loss": 0.5145, "step": 11340 }, { "epoch": 1.77, "grad_norm": 26.460607001777873, "learning_rate": 7.579835634406097e-06, "loss": 0.5746, "step": 11341 }, { "epoch": 1.77, "grad_norm": 17.09364872715884, "learning_rate": 7.578199402867607e-06, "loss": 0.5498, "step": 11342 }, { "epoch": 1.77, "grad_norm": 22.428643651102213, "learning_rate": 7.576563240203707e-06, "loss": 0.5437, "step": 11343 }, { "epoch": 1.77, "grad_norm": 27.391532778453815, "learning_rate": 7.574927146460928e-06, "loss": 0.7232, "step": 11344 }, { "epoch": 1.77, "grad_norm": 12.036940668630008, "learning_rate": 7.573291121685805e-06, "loss": 0.5555, "step": 11345 }, { "epoch": 1.77, "grad_norm": 14.541138106240838, "learning_rate": 7.571655165924861e-06, "loss": 0.5121, "step": 11346 }, { "epoch": 1.77, "grad_norm": 18.702217574801235, "learning_rate": 7.570019279224629e-06, "loss": 0.5851, "step": 11347 }, { "epoch": 1.77, "grad_norm": 21.994134952919435, "learning_rate": 7.568383461631621e-06, "loss": 0.5711, "step": 11348 }, { "epoch": 1.77, "grad_norm": 18.010437819121165, "learning_rate": 7.566747713192368e-06, "loss": 0.5566, "step": 11349 }, { "epoch": 1.77, "grad_norm": 14.631605681039963, "learning_rate": 7.5651120339533836e-06, "loss": 0.5361, "step": 11350 }, { "epoch": 1.77, "grad_norm": 26.537630574268555, "learning_rate": 7.56347642396119e-06, "loss": 0.4726, "step": 11351 }, { "epoch": 1.77, "grad_norm": 16.490625056277086, "learning_rate": 7.5618408832623044e-06, "loss": 0.5939, "step": 11352 }, { "epoch": 1.77, "grad_norm": 19.01667213887202, "learning_rate": 7.5602054119032364e-06, "loss": 0.581, "step": 11353 }, { "epoch": 1.77, "grad_norm": 18.18123170895015, "learning_rate": 7.558570009930497e-06, "loss": 0.5122, "step": 11354 }, { "epoch": 1.77, "grad_norm": 13.46694993671408, "learning_rate": 7.556934677390599e-06, "loss": 0.5276, "step": 11355 }, { "epoch": 1.77, "grad_norm": 15.678534258165081, "learning_rate": 7.555299414330048e-06, "loss": 0.5347, "step": 11356 }, { "epoch": 1.77, "grad_norm": 22.84877471547157, "learning_rate": 7.553664220795357e-06, "loss": 0.624, "step": 11357 }, { "epoch": 1.77, "grad_norm": 19.20675257685648, "learning_rate": 7.5520290968330226e-06, "loss": 0.5292, "step": 11358 }, { "epoch": 1.77, "grad_norm": 16.244811129173595, "learning_rate": 7.550394042489547e-06, "loss": 0.4978, "step": 11359 }, { "epoch": 1.77, "grad_norm": 22.691635399920745, "learning_rate": 7.54875905781143e-06, "loss": 0.5368, "step": 11360 }, { "epoch": 1.77, "grad_norm": 21.108495215344774, "learning_rate": 7.547124142845174e-06, "loss": 0.5894, "step": 11361 }, { "epoch": 1.77, "grad_norm": 19.787358983628412, "learning_rate": 7.545489297637275e-06, "loss": 0.5217, "step": 11362 }, { "epoch": 1.77, "grad_norm": 16.75864957839539, "learning_rate": 7.54385452223422e-06, "loss": 0.5147, "step": 11363 }, { "epoch": 1.78, "grad_norm": 17.827183936404236, "learning_rate": 7.542219816682508e-06, "loss": 0.5432, "step": 11364 }, { "epoch": 1.78, "grad_norm": 13.936623261335924, "learning_rate": 7.5405851810286236e-06, "loss": 0.52, "step": 11365 }, { "epoch": 1.78, "grad_norm": 31.788461221524262, "learning_rate": 7.538950615319065e-06, "loss": 0.5846, "step": 11366 }, { "epoch": 1.78, "grad_norm": 18.16915610236504, "learning_rate": 7.537316119600305e-06, "loss": 0.4864, "step": 11367 }, { "epoch": 1.78, "grad_norm": 26.443898616504786, "learning_rate": 7.535681693918836e-06, "loss": 0.6274, "step": 11368 }, { "epoch": 1.78, "grad_norm": 20.05773937853591, "learning_rate": 7.534047338321135e-06, "loss": 0.6017, "step": 11369 }, { "epoch": 1.78, "grad_norm": 16.0723196411894, "learning_rate": 7.5324130528536885e-06, "loss": 0.544, "step": 11370 }, { "epoch": 1.78, "grad_norm": 16.615076700172942, "learning_rate": 7.530778837562973e-06, "loss": 0.5261, "step": 11371 }, { "epoch": 1.78, "grad_norm": 17.103055651792662, "learning_rate": 7.5291446924954604e-06, "loss": 0.5812, "step": 11372 }, { "epoch": 1.78, "grad_norm": 24.764978800505, "learning_rate": 7.527510617697627e-06, "loss": 0.5989, "step": 11373 }, { "epoch": 1.78, "grad_norm": 14.735712433650383, "learning_rate": 7.525876613215944e-06, "loss": 0.5288, "step": 11374 }, { "epoch": 1.78, "grad_norm": 19.60952320707782, "learning_rate": 7.524242679096884e-06, "loss": 0.5349, "step": 11375 }, { "epoch": 1.78, "grad_norm": 15.33266028845027, "learning_rate": 7.522608815386916e-06, "loss": 0.5461, "step": 11376 }, { "epoch": 1.78, "grad_norm": 39.29299032375867, "learning_rate": 7.520975022132503e-06, "loss": 0.6018, "step": 11377 }, { "epoch": 1.78, "grad_norm": 20.944527482201728, "learning_rate": 7.5193412993801096e-06, "loss": 0.5803, "step": 11378 }, { "epoch": 1.78, "grad_norm": 16.319588604965567, "learning_rate": 7.517707647176197e-06, "loss": 0.5136, "step": 11379 }, { "epoch": 1.78, "grad_norm": 17.154907010275167, "learning_rate": 7.5160740655672285e-06, "loss": 0.5337, "step": 11380 }, { "epoch": 1.78, "grad_norm": 35.0472940895238, "learning_rate": 7.514440554599662e-06, "loss": 0.6092, "step": 11381 }, { "epoch": 1.78, "grad_norm": 12.427804427824013, "learning_rate": 7.5128071143199485e-06, "loss": 0.4463, "step": 11382 }, { "epoch": 1.78, "grad_norm": 28.468810734647416, "learning_rate": 7.511173744774548e-06, "loss": 0.6599, "step": 11383 }, { "epoch": 1.78, "grad_norm": 22.806755517760198, "learning_rate": 7.509540446009909e-06, "loss": 0.5261, "step": 11384 }, { "epoch": 1.78, "grad_norm": 24.650154336906553, "learning_rate": 7.5079072180724834e-06, "loss": 0.5738, "step": 11385 }, { "epoch": 1.78, "grad_norm": 20.552376672694674, "learning_rate": 7.506274061008721e-06, "loss": 0.5766, "step": 11386 }, { "epoch": 1.78, "grad_norm": 18.52665775612717, "learning_rate": 7.504640974865065e-06, "loss": 0.5652, "step": 11387 }, { "epoch": 1.78, "grad_norm": 18.364114213791307, "learning_rate": 7.503007959687956e-06, "loss": 0.5082, "step": 11388 }, { "epoch": 1.78, "grad_norm": 16.630305653625925, "learning_rate": 7.501375015523843e-06, "loss": 0.52, "step": 11389 }, { "epoch": 1.78, "grad_norm": 24.990781722460444, "learning_rate": 7.4997421424191595e-06, "loss": 0.5826, "step": 11390 }, { "epoch": 1.78, "grad_norm": 16.385785030547925, "learning_rate": 7.498109340420354e-06, "loss": 0.5185, "step": 11391 }, { "epoch": 1.78, "grad_norm": 21.994585668737507, "learning_rate": 7.496476609573851e-06, "loss": 0.5332, "step": 11392 }, { "epoch": 1.78, "grad_norm": 19.288584503548858, "learning_rate": 7.494843949926088e-06, "loss": 0.5416, "step": 11393 }, { "epoch": 1.78, "grad_norm": 14.799751232809594, "learning_rate": 7.493211361523496e-06, "loss": 0.496, "step": 11394 }, { "epoch": 1.78, "grad_norm": 24.41847322767009, "learning_rate": 7.49157884441251e-06, "loss": 0.5305, "step": 11395 }, { "epoch": 1.78, "grad_norm": 18.666975920498224, "learning_rate": 7.489946398639555e-06, "loss": 0.557, "step": 11396 }, { "epoch": 1.78, "grad_norm": 18.181201309500835, "learning_rate": 7.488314024251055e-06, "loss": 0.5348, "step": 11397 }, { "epoch": 1.78, "grad_norm": 17.335047304902012, "learning_rate": 7.486681721293432e-06, "loss": 0.5491, "step": 11398 }, { "epoch": 1.78, "grad_norm": 24.997225798224527, "learning_rate": 7.485049489813112e-06, "loss": 0.5846, "step": 11399 }, { "epoch": 1.78, "grad_norm": 21.045463066134577, "learning_rate": 7.483417329856513e-06, "loss": 0.6344, "step": 11400 }, { "epoch": 1.78, "grad_norm": 19.379279360863514, "learning_rate": 7.481785241470055e-06, "loss": 0.5108, "step": 11401 }, { "epoch": 1.78, "grad_norm": 30.37656194493515, "learning_rate": 7.48015322470015e-06, "loss": 0.6035, "step": 11402 }, { "epoch": 1.78, "grad_norm": 28.551575829058095, "learning_rate": 7.478521279593213e-06, "loss": 0.5514, "step": 11403 }, { "epoch": 1.78, "grad_norm": 30.489734825170196, "learning_rate": 7.476889406195656e-06, "loss": 0.6415, "step": 11404 }, { "epoch": 1.78, "grad_norm": 19.29606572975482, "learning_rate": 7.4752576045538894e-06, "loss": 0.5571, "step": 11405 }, { "epoch": 1.78, "grad_norm": 24.47288837429332, "learning_rate": 7.473625874714322e-06, "loss": 0.5259, "step": 11406 }, { "epoch": 1.78, "grad_norm": 26.90683662127291, "learning_rate": 7.471994216723352e-06, "loss": 0.6256, "step": 11407 }, { "epoch": 1.78, "grad_norm": 28.548577318448874, "learning_rate": 7.47036263062739e-06, "loss": 0.5733, "step": 11408 }, { "epoch": 1.78, "grad_norm": 24.489789801600327, "learning_rate": 7.4687311164728315e-06, "loss": 0.6144, "step": 11409 }, { "epoch": 1.78, "grad_norm": 27.41217402065633, "learning_rate": 7.467099674306083e-06, "loss": 0.6344, "step": 11410 }, { "epoch": 1.78, "grad_norm": 16.61036505489193, "learning_rate": 7.46546830417354e-06, "loss": 0.6296, "step": 11411 }, { "epoch": 1.78, "grad_norm": 18.603542227338057, "learning_rate": 7.463837006121593e-06, "loss": 0.5792, "step": 11412 }, { "epoch": 1.78, "grad_norm": 10.910512073640668, "learning_rate": 7.462205780196637e-06, "loss": 0.4822, "step": 11413 }, { "epoch": 1.78, "grad_norm": 16.151463022633337, "learning_rate": 7.460574626445065e-06, "loss": 0.4946, "step": 11414 }, { "epoch": 1.78, "grad_norm": 12.15470215668673, "learning_rate": 7.458943544913266e-06, "loss": 0.4978, "step": 11415 }, { "epoch": 1.78, "grad_norm": 18.915036437568993, "learning_rate": 7.457312535647627e-06, "loss": 0.5425, "step": 11416 }, { "epoch": 1.78, "grad_norm": 29.983166331350837, "learning_rate": 7.455681598694529e-06, "loss": 0.5607, "step": 11417 }, { "epoch": 1.78, "grad_norm": 13.344378133859822, "learning_rate": 7.454050734100358e-06, "loss": 0.5259, "step": 11418 }, { "epoch": 1.78, "grad_norm": 13.006090705488193, "learning_rate": 7.452419941911495e-06, "loss": 0.4909, "step": 11419 }, { "epoch": 1.78, "grad_norm": 36.77931888197404, "learning_rate": 7.450789222174322e-06, "loss": 0.5748, "step": 11420 }, { "epoch": 1.78, "grad_norm": 17.665778530168314, "learning_rate": 7.449158574935209e-06, "loss": 0.5435, "step": 11421 }, { "epoch": 1.78, "grad_norm": 15.403401709261521, "learning_rate": 7.4475280002405355e-06, "loss": 0.4883, "step": 11422 }, { "epoch": 1.78, "grad_norm": 19.565923136502217, "learning_rate": 7.445897498136671e-06, "loss": 0.5907, "step": 11423 }, { "epoch": 1.78, "grad_norm": 24.554561143137953, "learning_rate": 7.444267068669988e-06, "loss": 0.5442, "step": 11424 }, { "epoch": 1.78, "grad_norm": 15.825035906550575, "learning_rate": 7.442636711886857e-06, "loss": 0.5484, "step": 11425 }, { "epoch": 1.78, "grad_norm": 17.329956392178588, "learning_rate": 7.441006427833638e-06, "loss": 0.4986, "step": 11426 }, { "epoch": 1.78, "grad_norm": 25.342584678444084, "learning_rate": 7.439376216556702e-06, "loss": 0.6236, "step": 11427 }, { "epoch": 1.79, "grad_norm": 16.12708209146642, "learning_rate": 7.437746078102406e-06, "loss": 0.5069, "step": 11428 }, { "epoch": 1.79, "grad_norm": 13.136200042452291, "learning_rate": 7.4361160125171135e-06, "loss": 0.522, "step": 11429 }, { "epoch": 1.79, "grad_norm": 17.614532100569896, "learning_rate": 7.4344860198471845e-06, "loss": 0.5555, "step": 11430 }, { "epoch": 1.79, "grad_norm": 16.02168763759666, "learning_rate": 7.432856100138971e-06, "loss": 0.5387, "step": 11431 }, { "epoch": 1.79, "grad_norm": 24.54406809162352, "learning_rate": 7.431226253438826e-06, "loss": 0.5536, "step": 11432 }, { "epoch": 1.79, "grad_norm": 5.297411764849363, "learning_rate": 7.4295964797931045e-06, "loss": 0.5904, "step": 11433 }, { "epoch": 1.79, "grad_norm": 17.62735902863111, "learning_rate": 7.427966779248155e-06, "loss": 0.5318, "step": 11434 }, { "epoch": 1.79, "grad_norm": 12.214972048562355, "learning_rate": 7.4263371518503314e-06, "loss": 0.544, "step": 11435 }, { "epoch": 1.79, "grad_norm": 19.4808209052573, "learning_rate": 7.424707597645967e-06, "loss": 0.5807, "step": 11436 }, { "epoch": 1.79, "grad_norm": 23.361949452178724, "learning_rate": 7.423078116681415e-06, "loss": 0.5074, "step": 11437 }, { "epoch": 1.79, "grad_norm": 20.813417483202834, "learning_rate": 7.42144870900301e-06, "loss": 0.5444, "step": 11438 }, { "epoch": 1.79, "grad_norm": 19.101800580091467, "learning_rate": 7.419819374657099e-06, "loss": 0.4982, "step": 11439 }, { "epoch": 1.79, "grad_norm": 17.75164397108254, "learning_rate": 7.418190113690018e-06, "loss": 0.5825, "step": 11440 }, { "epoch": 1.79, "grad_norm": 27.168074552109232, "learning_rate": 7.416560926148096e-06, "loss": 0.5696, "step": 11441 }, { "epoch": 1.79, "grad_norm": 24.425701566438384, "learning_rate": 7.414931812077669e-06, "loss": 0.5407, "step": 11442 }, { "epoch": 1.79, "grad_norm": 16.379802725722428, "learning_rate": 7.413302771525071e-06, "loss": 0.5773, "step": 11443 }, { "epoch": 1.79, "grad_norm": 17.548462701218444, "learning_rate": 7.4116738045366275e-06, "loss": 0.5163, "step": 11444 }, { "epoch": 1.79, "grad_norm": 15.31737987577868, "learning_rate": 7.410044911158671e-06, "loss": 0.4955, "step": 11445 }, { "epoch": 1.79, "grad_norm": 14.409990786943924, "learning_rate": 7.408416091437519e-06, "loss": 0.5088, "step": 11446 }, { "epoch": 1.79, "grad_norm": 21.85818226923015, "learning_rate": 7.406787345419496e-06, "loss": 0.4689, "step": 11447 }, { "epoch": 1.79, "grad_norm": 40.5033319472836, "learning_rate": 7.405158673150925e-06, "loss": 0.5989, "step": 11448 }, { "epoch": 1.79, "grad_norm": 23.966572658288015, "learning_rate": 7.403530074678124e-06, "loss": 0.6299, "step": 11449 }, { "epoch": 1.79, "grad_norm": 26.144250024422643, "learning_rate": 7.401901550047412e-06, "loss": 0.5756, "step": 11450 }, { "epoch": 1.79, "grad_norm": 41.9853284824794, "learning_rate": 7.400273099305095e-06, "loss": 0.5159, "step": 11451 }, { "epoch": 1.79, "grad_norm": 17.188745112402138, "learning_rate": 7.398644722497492e-06, "loss": 0.58, "step": 11452 }, { "epoch": 1.79, "grad_norm": 23.155762553715245, "learning_rate": 7.39701641967091e-06, "loss": 0.6029, "step": 11453 }, { "epoch": 1.79, "grad_norm": 14.716441421418159, "learning_rate": 7.39538819087166e-06, "loss": 0.4462, "step": 11454 }, { "epoch": 1.79, "grad_norm": 21.049621076851558, "learning_rate": 7.393760036146049e-06, "loss": 0.5585, "step": 11455 }, { "epoch": 1.79, "grad_norm": 11.969821100140978, "learning_rate": 7.392131955540375e-06, "loss": 0.5433, "step": 11456 }, { "epoch": 1.79, "grad_norm": 29.09790260854883, "learning_rate": 7.39050394910094e-06, "loss": 0.5265, "step": 11457 }, { "epoch": 1.79, "grad_norm": 29.0838616271026, "learning_rate": 7.388876016874049e-06, "loss": 0.5435, "step": 11458 }, { "epoch": 1.79, "grad_norm": 23.600742562606495, "learning_rate": 7.3872481589059955e-06, "loss": 0.5656, "step": 11459 }, { "epoch": 1.79, "grad_norm": 19.187812296674355, "learning_rate": 7.3856203752430815e-06, "loss": 0.5298, "step": 11460 }, { "epoch": 1.79, "grad_norm": 16.464390609315895, "learning_rate": 7.383992665931587e-06, "loss": 0.5759, "step": 11461 }, { "epoch": 1.79, "grad_norm": 26.536347147839994, "learning_rate": 7.382365031017815e-06, "loss": 0.6713, "step": 11462 }, { "epoch": 1.79, "grad_norm": 26.35868706113572, "learning_rate": 7.380737470548047e-06, "loss": 0.5641, "step": 11463 }, { "epoch": 1.79, "grad_norm": 17.0942186094347, "learning_rate": 7.379109984568578e-06, "loss": 0.4985, "step": 11464 }, { "epoch": 1.79, "grad_norm": 18.604261128864643, "learning_rate": 7.377482573125685e-06, "loss": 0.5172, "step": 11465 }, { "epoch": 1.79, "grad_norm": 16.106487769309915, "learning_rate": 7.375855236265653e-06, "loss": 0.4756, "step": 11466 }, { "epoch": 1.79, "grad_norm": 20.625194477840566, "learning_rate": 7.374227974034762e-06, "loss": 0.5454, "step": 11467 }, { "epoch": 1.79, "grad_norm": 20.210040203417375, "learning_rate": 7.372600786479292e-06, "loss": 0.6038, "step": 11468 }, { "epoch": 1.79, "grad_norm": 18.766965638741333, "learning_rate": 7.370973673645523e-06, "loss": 0.5922, "step": 11469 }, { "epoch": 1.79, "grad_norm": 19.34738017269755, "learning_rate": 7.369346635579719e-06, "loss": 0.6014, "step": 11470 }, { "epoch": 1.79, "grad_norm": 21.399821268375863, "learning_rate": 7.36771967232816e-06, "loss": 0.6503, "step": 11471 }, { "epoch": 1.79, "grad_norm": 19.107412149332678, "learning_rate": 7.366092783937112e-06, "loss": 0.5697, "step": 11472 }, { "epoch": 1.79, "grad_norm": 10.495972166645576, "learning_rate": 7.364465970452846e-06, "loss": 0.5355, "step": 11473 }, { "epoch": 1.79, "grad_norm": 18.396476812833953, "learning_rate": 7.362839231921629e-06, "loss": 0.5564, "step": 11474 }, { "epoch": 1.79, "grad_norm": 12.382363428583044, "learning_rate": 7.36121256838972e-06, "loss": 0.5647, "step": 11475 }, { "epoch": 1.79, "grad_norm": 25.483142155800156, "learning_rate": 7.359585979903378e-06, "loss": 0.536, "step": 11476 }, { "epoch": 1.79, "grad_norm": 21.117138162908805, "learning_rate": 7.357959466508869e-06, "loss": 0.5983, "step": 11477 }, { "epoch": 1.79, "grad_norm": 24.221263222270228, "learning_rate": 7.356333028252445e-06, "loss": 0.6431, "step": 11478 }, { "epoch": 1.79, "grad_norm": 19.757440566532612, "learning_rate": 7.35470666518037e-06, "loss": 0.538, "step": 11479 }, { "epoch": 1.79, "grad_norm": 17.592670854771782, "learning_rate": 7.353080377338884e-06, "loss": 0.535, "step": 11480 }, { "epoch": 1.79, "grad_norm": 24.04995887655184, "learning_rate": 7.351454164774246e-06, "loss": 0.535, "step": 11481 }, { "epoch": 1.79, "grad_norm": 14.531234554569691, "learning_rate": 7.349828027532697e-06, "loss": 0.5391, "step": 11482 }, { "epoch": 1.79, "grad_norm": 12.007842090547893, "learning_rate": 7.348201965660493e-06, "loss": 0.4662, "step": 11483 }, { "epoch": 1.79, "grad_norm": 17.89742319054473, "learning_rate": 7.346575979203876e-06, "loss": 0.6283, "step": 11484 }, { "epoch": 1.79, "grad_norm": 19.997308287685787, "learning_rate": 7.344950068209085e-06, "loss": 0.5943, "step": 11485 }, { "epoch": 1.79, "grad_norm": 16.95218685340715, "learning_rate": 7.343324232722358e-06, "loss": 0.5205, "step": 11486 }, { "epoch": 1.79, "grad_norm": 16.901178688783077, "learning_rate": 7.341698472789937e-06, "loss": 0.5508, "step": 11487 }, { "epoch": 1.79, "grad_norm": 32.234490718439694, "learning_rate": 7.340072788458054e-06, "loss": 0.6428, "step": 11488 }, { "epoch": 1.79, "grad_norm": 16.13596092864975, "learning_rate": 7.33844717977295e-06, "loss": 0.5074, "step": 11489 }, { "epoch": 1.79, "grad_norm": 18.318878771675195, "learning_rate": 7.336821646780848e-06, "loss": 0.612, "step": 11490 }, { "epoch": 1.79, "grad_norm": 18.813702198757596, "learning_rate": 7.33519618952798e-06, "loss": 0.5537, "step": 11491 }, { "epoch": 1.8, "grad_norm": 33.8220451371144, "learning_rate": 7.333570808060575e-06, "loss": 0.6062, "step": 11492 }, { "epoch": 1.8, "grad_norm": 18.838004598079962, "learning_rate": 7.331945502424856e-06, "loss": 0.5966, "step": 11493 }, { "epoch": 1.8, "grad_norm": 23.606448647933572, "learning_rate": 7.330320272667048e-06, "loss": 0.5019, "step": 11494 }, { "epoch": 1.8, "grad_norm": 24.672707554502434, "learning_rate": 7.328695118833366e-06, "loss": 0.5954, "step": 11495 }, { "epoch": 1.8, "grad_norm": 19.290324147070788, "learning_rate": 7.327070040970034e-06, "loss": 0.5974, "step": 11496 }, { "epoch": 1.8, "grad_norm": 17.957349660297997, "learning_rate": 7.325445039123264e-06, "loss": 0.5039, "step": 11497 }, { "epoch": 1.8, "grad_norm": 13.023864402707884, "learning_rate": 7.323820113339275e-06, "loss": 0.5635, "step": 11498 }, { "epoch": 1.8, "grad_norm": 16.764988161436822, "learning_rate": 7.32219526366428e-06, "loss": 0.4926, "step": 11499 }, { "epoch": 1.8, "grad_norm": 19.2724340484563, "learning_rate": 7.320570490144481e-06, "loss": 0.56, "step": 11500 }, { "epoch": 1.8, "grad_norm": 14.689650348748872, "learning_rate": 7.318945792826089e-06, "loss": 0.4847, "step": 11501 }, { "epoch": 1.8, "grad_norm": 16.03860160808342, "learning_rate": 7.317321171755312e-06, "loss": 0.5869, "step": 11502 }, { "epoch": 1.8, "grad_norm": 21.71263439675476, "learning_rate": 7.315696626978349e-06, "loss": 0.582, "step": 11503 }, { "epoch": 1.8, "grad_norm": 13.618109297616618, "learning_rate": 7.31407215854141e-06, "loss": 0.5609, "step": 11504 }, { "epoch": 1.8, "grad_norm": 19.716229185459664, "learning_rate": 7.312447766490682e-06, "loss": 0.5531, "step": 11505 }, { "epoch": 1.8, "grad_norm": 28.59327396194591, "learning_rate": 7.310823450872368e-06, "loss": 0.6018, "step": 11506 }, { "epoch": 1.8, "grad_norm": 23.28109653729545, "learning_rate": 7.309199211732662e-06, "loss": 0.5789, "step": 11507 }, { "epoch": 1.8, "grad_norm": 25.067616773364556, "learning_rate": 7.307575049117758e-06, "loss": 0.529, "step": 11508 }, { "epoch": 1.8, "grad_norm": 12.129865554982795, "learning_rate": 7.305950963073845e-06, "loss": 0.5656, "step": 11509 }, { "epoch": 1.8, "grad_norm": 18.387748238030866, "learning_rate": 7.30432695364711e-06, "loss": 0.5259, "step": 11510 }, { "epoch": 1.8, "grad_norm": 24.88808603023079, "learning_rate": 7.3027030208837365e-06, "loss": 0.5565, "step": 11511 }, { "epoch": 1.8, "grad_norm": 19.058768320662143, "learning_rate": 7.301079164829914e-06, "loss": 0.5268, "step": 11512 }, { "epoch": 1.8, "grad_norm": 17.521769197177548, "learning_rate": 7.299455385531824e-06, "loss": 0.5398, "step": 11513 }, { "epoch": 1.8, "grad_norm": 15.795225292172237, "learning_rate": 7.297831683035638e-06, "loss": 0.5016, "step": 11514 }, { "epoch": 1.8, "grad_norm": 15.60226270664066, "learning_rate": 7.296208057387539e-06, "loss": 0.5119, "step": 11515 }, { "epoch": 1.8, "grad_norm": 15.838540825132426, "learning_rate": 7.294584508633702e-06, "loss": 0.4934, "step": 11516 }, { "epoch": 1.8, "grad_norm": 24.4410294206056, "learning_rate": 7.292961036820299e-06, "loss": 0.5948, "step": 11517 }, { "epoch": 1.8, "grad_norm": 18.67166955667262, "learning_rate": 7.291337641993503e-06, "loss": 0.5506, "step": 11518 }, { "epoch": 1.8, "grad_norm": 25.66219568880412, "learning_rate": 7.289714324199477e-06, "loss": 0.5738, "step": 11519 }, { "epoch": 1.8, "grad_norm": 25.18033534041247, "learning_rate": 7.28809108348439e-06, "loss": 0.5238, "step": 11520 }, { "epoch": 1.8, "grad_norm": 30.523125982537806, "learning_rate": 7.2864679198944065e-06, "loss": 0.5832, "step": 11521 }, { "epoch": 1.8, "grad_norm": 20.100377477474808, "learning_rate": 7.284844833475687e-06, "loss": 0.6028, "step": 11522 }, { "epoch": 1.8, "grad_norm": 19.453355127528393, "learning_rate": 7.2832218242743976e-06, "loss": 0.5583, "step": 11523 }, { "epoch": 1.8, "grad_norm": 16.263042211913042, "learning_rate": 7.281598892336685e-06, "loss": 0.4963, "step": 11524 }, { "epoch": 1.8, "grad_norm": 17.999787035483596, "learning_rate": 7.279976037708712e-06, "loss": 0.5325, "step": 11525 }, { "epoch": 1.8, "grad_norm": 11.882125789505263, "learning_rate": 7.278353260436626e-06, "loss": 0.4867, "step": 11526 }, { "epoch": 1.8, "grad_norm": 16.530496046922735, "learning_rate": 7.276730560566583e-06, "loss": 0.4782, "step": 11527 }, { "epoch": 1.8, "grad_norm": 19.621875188426237, "learning_rate": 7.275107938144732e-06, "loss": 0.5357, "step": 11528 }, { "epoch": 1.8, "grad_norm": 39.83509778942412, "learning_rate": 7.273485393217217e-06, "loss": 0.6171, "step": 11529 }, { "epoch": 1.8, "grad_norm": 20.770321480678714, "learning_rate": 7.2718629258301786e-06, "loss": 0.5575, "step": 11530 }, { "epoch": 1.8, "grad_norm": 15.559136687930513, "learning_rate": 7.270240536029765e-06, "loss": 0.5941, "step": 11531 }, { "epoch": 1.8, "grad_norm": 22.22612460263177, "learning_rate": 7.268618223862112e-06, "loss": 0.5206, "step": 11532 }, { "epoch": 1.8, "grad_norm": 30.797401464366427, "learning_rate": 7.266995989373362e-06, "loss": 0.6178, "step": 11533 }, { "epoch": 1.8, "grad_norm": 15.26367688901969, "learning_rate": 7.265373832609647e-06, "loss": 0.4966, "step": 11534 }, { "epoch": 1.8, "grad_norm": 15.968609561041333, "learning_rate": 7.263751753617096e-06, "loss": 0.4368, "step": 11535 }, { "epoch": 1.8, "grad_norm": 12.96177088798451, "learning_rate": 7.262129752441849e-06, "loss": 0.5324, "step": 11536 }, { "epoch": 1.8, "grad_norm": 34.13775319082707, "learning_rate": 7.26050782913003e-06, "loss": 0.5062, "step": 11537 }, { "epoch": 1.8, "grad_norm": 15.345901915224099, "learning_rate": 7.258885983727767e-06, "loss": 0.5407, "step": 11538 }, { "epoch": 1.8, "grad_norm": 18.20253414292048, "learning_rate": 7.257264216281179e-06, "loss": 0.5549, "step": 11539 }, { "epoch": 1.8, "grad_norm": 16.602094976076273, "learning_rate": 7.255642526836395e-06, "loss": 0.5279, "step": 11540 }, { "epoch": 1.8, "grad_norm": 19.932284910787157, "learning_rate": 7.254020915439528e-06, "loss": 0.5675, "step": 11541 }, { "epoch": 1.8, "grad_norm": 23.63453724990942, "learning_rate": 7.252399382136703e-06, "loss": 0.5577, "step": 11542 }, { "epoch": 1.8, "grad_norm": 29.52740181771988, "learning_rate": 7.250777926974035e-06, "loss": 0.5173, "step": 11543 }, { "epoch": 1.8, "grad_norm": 19.16748262686469, "learning_rate": 7.2491565499976335e-06, "loss": 0.4873, "step": 11544 }, { "epoch": 1.8, "grad_norm": 22.219209069393244, "learning_rate": 7.247535251253606e-06, "loss": 0.5414, "step": 11545 }, { "epoch": 1.8, "grad_norm": 22.595532408326665, "learning_rate": 7.245914030788069e-06, "loss": 0.6052, "step": 11546 }, { "epoch": 1.8, "grad_norm": 19.39046671400855, "learning_rate": 7.2442928886471246e-06, "loss": 0.4945, "step": 11547 }, { "epoch": 1.8, "grad_norm": 14.75083033600665, "learning_rate": 7.242671824876884e-06, "loss": 0.493, "step": 11548 }, { "epoch": 1.8, "grad_norm": 17.200690403828464, "learning_rate": 7.241050839523437e-06, "loss": 0.532, "step": 11549 }, { "epoch": 1.8, "grad_norm": 24.070537502632565, "learning_rate": 7.239429932632892e-06, "loss": 0.5129, "step": 11550 }, { "epoch": 1.8, "grad_norm": 17.56690887688945, "learning_rate": 7.237809104251343e-06, "loss": 0.5329, "step": 11551 }, { "epoch": 1.8, "grad_norm": 25.513129948030343, "learning_rate": 7.23618835442489e-06, "loss": 0.5995, "step": 11552 }, { "epoch": 1.8, "grad_norm": 17.327299232516882, "learning_rate": 7.234567683199624e-06, "loss": 0.5152, "step": 11553 }, { "epoch": 1.8, "grad_norm": 24.434607061388434, "learning_rate": 7.232947090621633e-06, "loss": 0.6099, "step": 11554 }, { "epoch": 1.8, "grad_norm": 19.68106319469668, "learning_rate": 7.23132657673701e-06, "loss": 0.5755, "step": 11555 }, { "epoch": 1.81, "grad_norm": 17.69053986252105, "learning_rate": 7.2297061415918394e-06, "loss": 0.5416, "step": 11556 }, { "epoch": 1.81, "grad_norm": 25.534710592366903, "learning_rate": 7.228085785232206e-06, "loss": 0.538, "step": 11557 }, { "epoch": 1.81, "grad_norm": 14.670140467561515, "learning_rate": 7.226465507704189e-06, "loss": 0.5287, "step": 11558 }, { "epoch": 1.81, "grad_norm": 24.034391567416467, "learning_rate": 7.224845309053872e-06, "loss": 0.5244, "step": 11559 }, { "epoch": 1.81, "grad_norm": 23.254307325144577, "learning_rate": 7.2232251893273295e-06, "loss": 0.554, "step": 11560 }, { "epoch": 1.81, "grad_norm": 18.428917244941026, "learning_rate": 7.2216051485706385e-06, "loss": 0.6522, "step": 11561 }, { "epoch": 1.81, "grad_norm": 15.35913087877865, "learning_rate": 7.219985186829877e-06, "loss": 0.5901, "step": 11562 }, { "epoch": 1.81, "grad_norm": 13.886915921771122, "learning_rate": 7.2183653041511045e-06, "loss": 0.4914, "step": 11563 }, { "epoch": 1.81, "grad_norm": 12.316887925659467, "learning_rate": 7.216745500580396e-06, "loss": 0.4751, "step": 11564 }, { "epoch": 1.81, "grad_norm": 15.381351965388394, "learning_rate": 7.215125776163818e-06, "loss": 0.53, "step": 11565 }, { "epoch": 1.81, "grad_norm": 22.229950311112308, "learning_rate": 7.213506130947434e-06, "loss": 0.5857, "step": 11566 }, { "epoch": 1.81, "grad_norm": 14.602523431748542, "learning_rate": 7.21188656497731e-06, "loss": 0.4648, "step": 11567 }, { "epoch": 1.81, "grad_norm": 18.704346830809087, "learning_rate": 7.210267078299495e-06, "loss": 0.6302, "step": 11568 }, { "epoch": 1.81, "grad_norm": 17.25239431136035, "learning_rate": 7.208647670960055e-06, "loss": 0.5672, "step": 11569 }, { "epoch": 1.81, "grad_norm": 24.816507427838555, "learning_rate": 7.2070283430050405e-06, "loss": 0.5629, "step": 11570 }, { "epoch": 1.81, "grad_norm": 15.329995378554532, "learning_rate": 7.205409094480506e-06, "loss": 0.6058, "step": 11571 }, { "epoch": 1.81, "grad_norm": 18.91797965965295, "learning_rate": 7.203789925432507e-06, "loss": 0.5032, "step": 11572 }, { "epoch": 1.81, "grad_norm": 19.999618177196492, "learning_rate": 7.202170835907083e-06, "loss": 0.512, "step": 11573 }, { "epoch": 1.81, "grad_norm": 19.35564411903794, "learning_rate": 7.200551825950281e-06, "loss": 0.5569, "step": 11574 }, { "epoch": 1.81, "grad_norm": 25.16508216059275, "learning_rate": 7.198932895608153e-06, "loss": 0.5442, "step": 11575 }, { "epoch": 1.81, "grad_norm": 19.32847007128881, "learning_rate": 7.197314044926729e-06, "loss": 0.5186, "step": 11576 }, { "epoch": 1.81, "grad_norm": 20.986982417213863, "learning_rate": 7.1956952739520605e-06, "loss": 0.6003, "step": 11577 }, { "epoch": 1.81, "grad_norm": 16.43661062931299, "learning_rate": 7.194076582730174e-06, "loss": 0.5839, "step": 11578 }, { "epoch": 1.81, "grad_norm": 19.077134116025597, "learning_rate": 7.192457971307107e-06, "loss": 0.4945, "step": 11579 }, { "epoch": 1.81, "grad_norm": 21.3660693365981, "learning_rate": 7.190839439728896e-06, "loss": 0.6116, "step": 11580 }, { "epoch": 1.81, "grad_norm": 15.944513625187975, "learning_rate": 7.1892209880415665e-06, "loss": 0.5307, "step": 11581 }, { "epoch": 1.81, "grad_norm": 16.509966789598643, "learning_rate": 7.18760261629115e-06, "loss": 0.5284, "step": 11582 }, { "epoch": 1.81, "grad_norm": 14.926267509656231, "learning_rate": 7.185984324523667e-06, "loss": 0.6299, "step": 11583 }, { "epoch": 1.81, "grad_norm": 18.47240983617567, "learning_rate": 7.184366112785144e-06, "loss": 0.5129, "step": 11584 }, { "epoch": 1.81, "grad_norm": 13.714475371439356, "learning_rate": 7.1827479811216e-06, "loss": 0.4596, "step": 11585 }, { "epoch": 1.81, "grad_norm": 18.759201687900756, "learning_rate": 7.181129929579058e-06, "loss": 0.6174, "step": 11586 }, { "epoch": 1.81, "grad_norm": 14.957772109820837, "learning_rate": 7.179511958203535e-06, "loss": 0.5165, "step": 11587 }, { "epoch": 1.81, "grad_norm": 16.264233700982434, "learning_rate": 7.177894067041038e-06, "loss": 0.495, "step": 11588 }, { "epoch": 1.81, "grad_norm": 14.501972470768314, "learning_rate": 7.176276256137582e-06, "loss": 0.5209, "step": 11589 }, { "epoch": 1.81, "grad_norm": 19.99326854017606, "learning_rate": 7.174658525539179e-06, "loss": 0.5556, "step": 11590 }, { "epoch": 1.81, "grad_norm": 18.133672481410485, "learning_rate": 7.173040875291832e-06, "loss": 0.5433, "step": 11591 }, { "epoch": 1.81, "grad_norm": 25.741492951767032, "learning_rate": 7.171423305441556e-06, "loss": 0.6295, "step": 11592 }, { "epoch": 1.81, "grad_norm": 15.711441463504482, "learning_rate": 7.169805816034341e-06, "loss": 0.5306, "step": 11593 }, { "epoch": 1.81, "grad_norm": 17.01586118994831, "learning_rate": 7.168188407116194e-06, "loss": 0.4987, "step": 11594 }, { "epoch": 1.81, "grad_norm": 18.13723441201615, "learning_rate": 7.166571078733109e-06, "loss": 0.5472, "step": 11595 }, { "epoch": 1.81, "grad_norm": 22.416324911487486, "learning_rate": 7.164953830931089e-06, "loss": 0.5724, "step": 11596 }, { "epoch": 1.81, "grad_norm": 23.745983958755303, "learning_rate": 7.1633366637561245e-06, "loss": 0.453, "step": 11597 }, { "epoch": 1.81, "grad_norm": 23.556714797830292, "learning_rate": 7.161719577254203e-06, "loss": 0.5831, "step": 11598 }, { "epoch": 1.81, "grad_norm": 24.194467292067234, "learning_rate": 7.160102571471317e-06, "loss": 0.5713, "step": 11599 }, { "epoch": 1.81, "grad_norm": 17.796453285077092, "learning_rate": 7.158485646453452e-06, "loss": 0.5181, "step": 11600 }, { "epoch": 1.81, "grad_norm": 31.683945862802087, "learning_rate": 7.156868802246593e-06, "loss": 0.6248, "step": 11601 }, { "epoch": 1.81, "grad_norm": 13.174943732778496, "learning_rate": 7.155252038896725e-06, "loss": 0.5787, "step": 11602 }, { "epoch": 1.81, "grad_norm": 17.143311059709728, "learning_rate": 7.153635356449824e-06, "loss": 0.5478, "step": 11603 }, { "epoch": 1.81, "grad_norm": 24.6078015893191, "learning_rate": 7.152018754951864e-06, "loss": 0.57, "step": 11604 }, { "epoch": 1.81, "grad_norm": 19.892040079862152, "learning_rate": 7.150402234448829e-06, "loss": 0.5277, "step": 11605 }, { "epoch": 1.81, "grad_norm": 16.082032184911203, "learning_rate": 7.1487857949866875e-06, "loss": 0.5345, "step": 11606 }, { "epoch": 1.81, "grad_norm": 17.600869583344398, "learning_rate": 7.14716943661141e-06, "loss": 0.484, "step": 11607 }, { "epoch": 1.81, "grad_norm": 17.463483634175, "learning_rate": 7.145553159368962e-06, "loss": 0.5432, "step": 11608 }, { "epoch": 1.81, "grad_norm": 21.550240639009058, "learning_rate": 7.143936963305314e-06, "loss": 0.4511, "step": 11609 }, { "epoch": 1.81, "grad_norm": 18.326305962561612, "learning_rate": 7.142320848466427e-06, "loss": 0.6216, "step": 11610 }, { "epoch": 1.81, "grad_norm": 20.925693796618475, "learning_rate": 7.140704814898268e-06, "loss": 0.5114, "step": 11611 }, { "epoch": 1.81, "grad_norm": 21.197542415629997, "learning_rate": 7.1390888626467855e-06, "loss": 0.5002, "step": 11612 }, { "epoch": 1.81, "grad_norm": 16.991477213955214, "learning_rate": 7.137472991757945e-06, "loss": 0.5065, "step": 11613 }, { "epoch": 1.81, "grad_norm": 20.739434624590164, "learning_rate": 7.1358572022776965e-06, "loss": 0.4908, "step": 11614 }, { "epoch": 1.81, "grad_norm": 15.775300688662563, "learning_rate": 7.134241494251994e-06, "loss": 0.5454, "step": 11615 }, { "epoch": 1.81, "grad_norm": 13.93031653162277, "learning_rate": 7.1326258677267916e-06, "loss": 0.5446, "step": 11616 }, { "epoch": 1.81, "grad_norm": 13.415612695785184, "learning_rate": 7.131010322748029e-06, "loss": 0.4873, "step": 11617 }, { "epoch": 1.81, "grad_norm": 16.02539989929117, "learning_rate": 7.129394859361652e-06, "loss": 0.559, "step": 11618 }, { "epoch": 1.81, "grad_norm": 14.872117397347287, "learning_rate": 7.127779477613608e-06, "loss": 0.5077, "step": 11619 }, { "epoch": 1.82, "grad_norm": 24.65724182914066, "learning_rate": 7.126164177549832e-06, "loss": 0.6218, "step": 11620 }, { "epoch": 1.82, "grad_norm": 13.759147034514728, "learning_rate": 7.124548959216272e-06, "loss": 0.5329, "step": 11621 }, { "epoch": 1.82, "grad_norm": 19.038905292185966, "learning_rate": 7.122933822658855e-06, "loss": 0.4956, "step": 11622 }, { "epoch": 1.82, "grad_norm": 29.62218741201781, "learning_rate": 7.121318767923514e-06, "loss": 0.4754, "step": 11623 }, { "epoch": 1.82, "grad_norm": 19.502984850110522, "learning_rate": 7.1197037950561855e-06, "loss": 0.4947, "step": 11624 }, { "epoch": 1.82, "grad_norm": 16.272880233786456, "learning_rate": 7.118088904102798e-06, "loss": 0.5762, "step": 11625 }, { "epoch": 1.82, "grad_norm": 20.44003194179857, "learning_rate": 7.116474095109276e-06, "loss": 0.6131, "step": 11626 }, { "epoch": 1.82, "grad_norm": 28.673043687165784, "learning_rate": 7.11485936812154e-06, "loss": 0.5435, "step": 11627 }, { "epoch": 1.82, "grad_norm": 21.987354130936072, "learning_rate": 7.113244723185519e-06, "loss": 0.5232, "step": 11628 }, { "epoch": 1.82, "grad_norm": 16.152391405969357, "learning_rate": 7.111630160347127e-06, "loss": 0.5511, "step": 11629 }, { "epoch": 1.82, "grad_norm": 18.23919644881341, "learning_rate": 7.110015679652286e-06, "loss": 0.5772, "step": 11630 }, { "epoch": 1.82, "grad_norm": 18.039844517465994, "learning_rate": 7.1084012811469105e-06, "loss": 0.5794, "step": 11631 }, { "epoch": 1.82, "grad_norm": 13.622477521039478, "learning_rate": 7.106786964876909e-06, "loss": 0.5101, "step": 11632 }, { "epoch": 1.82, "grad_norm": 14.455026629330053, "learning_rate": 7.10517273088819e-06, "loss": 0.5568, "step": 11633 }, { "epoch": 1.82, "grad_norm": 17.477765151496527, "learning_rate": 7.103558579226668e-06, "loss": 0.6092, "step": 11634 }, { "epoch": 1.82, "grad_norm": 16.827347758734685, "learning_rate": 7.101944509938244e-06, "loss": 0.5387, "step": 11635 }, { "epoch": 1.82, "grad_norm": 18.319090684113004, "learning_rate": 7.100330523068829e-06, "loss": 0.5566, "step": 11636 }, { "epoch": 1.82, "grad_norm": 12.285856624828533, "learning_rate": 7.098716618664312e-06, "loss": 0.5493, "step": 11637 }, { "epoch": 1.82, "grad_norm": 20.543538877210672, "learning_rate": 7.097102796770598e-06, "loss": 0.6521, "step": 11638 }, { "epoch": 1.82, "grad_norm": 16.06960988490768, "learning_rate": 7.0954890574335814e-06, "loss": 0.5151, "step": 11639 }, { "epoch": 1.82, "grad_norm": 15.641734816259934, "learning_rate": 7.0938754006991584e-06, "loss": 0.456, "step": 11640 }, { "epoch": 1.82, "grad_norm": 12.988555649461693, "learning_rate": 7.092261826613221e-06, "loss": 0.4699, "step": 11641 }, { "epoch": 1.82, "grad_norm": 16.667874990610347, "learning_rate": 7.0906483352216525e-06, "loss": 0.5523, "step": 11642 }, { "epoch": 1.82, "grad_norm": 26.79399074291014, "learning_rate": 7.0890349265703455e-06, "loss": 0.5444, "step": 11643 }, { "epoch": 1.82, "grad_norm": 16.387598566527046, "learning_rate": 7.0874216007051826e-06, "loss": 0.5266, "step": 11644 }, { "epoch": 1.82, "grad_norm": 19.485098884246934, "learning_rate": 7.0858083576720416e-06, "loss": 0.5479, "step": 11645 }, { "epoch": 1.82, "grad_norm": 20.359197228988386, "learning_rate": 7.084195197516812e-06, "loss": 0.5273, "step": 11646 }, { "epoch": 1.82, "grad_norm": 15.397537041574818, "learning_rate": 7.082582120285363e-06, "loss": 0.5431, "step": 11647 }, { "epoch": 1.82, "grad_norm": 17.962690917416055, "learning_rate": 7.080969126023569e-06, "loss": 0.5433, "step": 11648 }, { "epoch": 1.82, "grad_norm": 11.65443252515218, "learning_rate": 7.079356214777309e-06, "loss": 0.5319, "step": 11649 }, { "epoch": 1.82, "grad_norm": 18.089367226996785, "learning_rate": 7.077743386592448e-06, "loss": 0.5205, "step": 11650 }, { "epoch": 1.82, "grad_norm": 18.824622964645325, "learning_rate": 7.07613064151486e-06, "loss": 0.5486, "step": 11651 }, { "epoch": 1.82, "grad_norm": 17.933723391905808, "learning_rate": 7.0745179795904005e-06, "loss": 0.5626, "step": 11652 }, { "epoch": 1.82, "grad_norm": 22.729480015211607, "learning_rate": 7.07290540086494e-06, "loss": 0.5183, "step": 11653 }, { "epoch": 1.82, "grad_norm": 17.888364682400976, "learning_rate": 7.071292905384336e-06, "loss": 0.5771, "step": 11654 }, { "epoch": 1.82, "grad_norm": 14.516893766652776, "learning_rate": 7.069680493194455e-06, "loss": 0.5666, "step": 11655 }, { "epoch": 1.82, "grad_norm": 12.383855888560012, "learning_rate": 7.068068164341139e-06, "loss": 0.4768, "step": 11656 }, { "epoch": 1.82, "grad_norm": 26.092905400935205, "learning_rate": 7.066455918870254e-06, "loss": 0.6187, "step": 11657 }, { "epoch": 1.82, "grad_norm": 17.99880162031453, "learning_rate": 7.064843756827641e-06, "loss": 0.5263, "step": 11658 }, { "epoch": 1.82, "grad_norm": 20.83135382013805, "learning_rate": 7.06323167825916e-06, "loss": 0.4394, "step": 11659 }, { "epoch": 1.82, "grad_norm": 19.115081215869775, "learning_rate": 7.0616196832106535e-06, "loss": 0.5675, "step": 11660 }, { "epoch": 1.82, "grad_norm": 17.19547223039035, "learning_rate": 7.060007771727961e-06, "loss": 0.5267, "step": 11661 }, { "epoch": 1.82, "grad_norm": 19.36535662075284, "learning_rate": 7.058395943856926e-06, "loss": 0.4983, "step": 11662 }, { "epoch": 1.82, "grad_norm": 19.881207718754755, "learning_rate": 7.056784199643392e-06, "loss": 0.615, "step": 11663 }, { "epoch": 1.82, "grad_norm": 37.98192371899534, "learning_rate": 7.05517253913319e-06, "loss": 0.7017, "step": 11664 }, { "epoch": 1.82, "grad_norm": 18.144408628215704, "learning_rate": 7.053560962372166e-06, "loss": 0.5324, "step": 11665 }, { "epoch": 1.82, "grad_norm": 41.209823439305936, "learning_rate": 7.051949469406139e-06, "loss": 0.6122, "step": 11666 }, { "epoch": 1.82, "grad_norm": 14.731807680259813, "learning_rate": 7.0503380602809435e-06, "loss": 0.532, "step": 11667 }, { "epoch": 1.82, "grad_norm": 17.82789525494998, "learning_rate": 7.04872673504241e-06, "loss": 0.4863, "step": 11668 }, { "epoch": 1.82, "grad_norm": 11.934443048076906, "learning_rate": 7.047115493736363e-06, "loss": 0.5891, "step": 11669 }, { "epoch": 1.82, "grad_norm": 25.771589173584143, "learning_rate": 7.045504336408625e-06, "loss": 0.5339, "step": 11670 }, { "epoch": 1.82, "grad_norm": 21.56569298862143, "learning_rate": 7.043893263105012e-06, "loss": 0.5708, "step": 11671 }, { "epoch": 1.82, "grad_norm": 32.98548931595973, "learning_rate": 7.042282273871346e-06, "loss": 0.5527, "step": 11672 }, { "epoch": 1.82, "grad_norm": 23.69957079357906, "learning_rate": 7.04067136875344e-06, "loss": 0.5612, "step": 11673 }, { "epoch": 1.82, "grad_norm": 20.38158897442927, "learning_rate": 7.039060547797112e-06, "loss": 0.6501, "step": 11674 }, { "epoch": 1.82, "grad_norm": 17.600871070000377, "learning_rate": 7.037449811048172e-06, "loss": 0.5599, "step": 11675 }, { "epoch": 1.82, "grad_norm": 19.809189468344815, "learning_rate": 7.035839158552424e-06, "loss": 0.5837, "step": 11676 }, { "epoch": 1.82, "grad_norm": 16.348039741530055, "learning_rate": 7.034228590355674e-06, "loss": 0.5329, "step": 11677 }, { "epoch": 1.82, "grad_norm": 19.98229254698548, "learning_rate": 7.03261810650373e-06, "loss": 0.6243, "step": 11678 }, { "epoch": 1.82, "grad_norm": 24.72234095593728, "learning_rate": 7.03100770704239e-06, "loss": 0.636, "step": 11679 }, { "epoch": 1.82, "grad_norm": 16.60186140150378, "learning_rate": 7.029397392017461e-06, "loss": 0.4827, "step": 11680 }, { "epoch": 1.82, "grad_norm": 20.502826870839673, "learning_rate": 7.027787161474724e-06, "loss": 0.5632, "step": 11681 }, { "epoch": 1.82, "grad_norm": 18.442930845763893, "learning_rate": 7.026177015459984e-06, "loss": 0.4844, "step": 11682 }, { "epoch": 1.82, "grad_norm": 18.428943477520082, "learning_rate": 7.024566954019028e-06, "loss": 0.5357, "step": 11683 }, { "epoch": 1.83, "grad_norm": 13.090731012707813, "learning_rate": 7.0229569771976505e-06, "loss": 0.4444, "step": 11684 }, { "epoch": 1.83, "grad_norm": 21.401333984377093, "learning_rate": 7.021347085041636e-06, "loss": 0.4988, "step": 11685 }, { "epoch": 1.83, "grad_norm": 13.31659277496896, "learning_rate": 7.019737277596763e-06, "loss": 0.5421, "step": 11686 }, { "epoch": 1.83, "grad_norm": 15.635348465567645, "learning_rate": 7.018127554908823e-06, "loss": 0.5676, "step": 11687 }, { "epoch": 1.83, "grad_norm": 17.71528537624553, "learning_rate": 7.016517917023588e-06, "loss": 0.5824, "step": 11688 }, { "epoch": 1.83, "grad_norm": 16.41710880322763, "learning_rate": 7.014908363986839e-06, "loss": 0.5984, "step": 11689 }, { "epoch": 1.83, "grad_norm": 17.762436271924965, "learning_rate": 7.013298895844352e-06, "loss": 0.5494, "step": 11690 }, { "epoch": 1.83, "grad_norm": 21.50616798945086, "learning_rate": 7.011689512641896e-06, "loss": 0.5149, "step": 11691 }, { "epoch": 1.83, "grad_norm": 15.878454605473662, "learning_rate": 7.01008021442524e-06, "loss": 0.5081, "step": 11692 }, { "epoch": 1.83, "grad_norm": 26.63039866301779, "learning_rate": 7.008471001240156e-06, "loss": 0.6059, "step": 11693 }, { "epoch": 1.83, "grad_norm": 14.296200632127004, "learning_rate": 7.006861873132406e-06, "loss": 0.5353, "step": 11694 }, { "epoch": 1.83, "grad_norm": 13.308023601777744, "learning_rate": 7.005252830147756e-06, "loss": 0.4995, "step": 11695 }, { "epoch": 1.83, "grad_norm": 13.613516585780115, "learning_rate": 7.00364387233196e-06, "loss": 0.5259, "step": 11696 }, { "epoch": 1.83, "grad_norm": 15.367796623519187, "learning_rate": 7.00203499973078e-06, "loss": 0.5112, "step": 11697 }, { "epoch": 1.83, "grad_norm": 22.999478345894985, "learning_rate": 7.00042621238997e-06, "loss": 0.5607, "step": 11698 }, { "epoch": 1.83, "grad_norm": 20.2043551788129, "learning_rate": 6.998817510355285e-06, "loss": 0.4995, "step": 11699 }, { "epoch": 1.83, "grad_norm": 24.29378257709534, "learning_rate": 6.997208893672478e-06, "loss": 0.5981, "step": 11700 }, { "epoch": 1.83, "grad_norm": 20.001025599766976, "learning_rate": 6.9956003623872895e-06, "loss": 0.5671, "step": 11701 }, { "epoch": 1.83, "grad_norm": 20.665063599420087, "learning_rate": 6.993991916545468e-06, "loss": 0.541, "step": 11702 }, { "epoch": 1.83, "grad_norm": 16.794120899011663, "learning_rate": 6.99238355619276e-06, "loss": 0.5195, "step": 11703 }, { "epoch": 1.83, "grad_norm": 19.06642043260306, "learning_rate": 6.990775281374907e-06, "loss": 0.5742, "step": 11704 }, { "epoch": 1.83, "grad_norm": 21.570380307139004, "learning_rate": 6.989167092137639e-06, "loss": 0.5394, "step": 11705 }, { "epoch": 1.83, "grad_norm": 18.745364045974263, "learning_rate": 6.987558988526702e-06, "loss": 0.5228, "step": 11706 }, { "epoch": 1.83, "grad_norm": 18.606251088876917, "learning_rate": 6.985950970587823e-06, "loss": 0.5828, "step": 11707 }, { "epoch": 1.83, "grad_norm": 19.743709976491427, "learning_rate": 6.984343038366733e-06, "loss": 0.5984, "step": 11708 }, { "epoch": 1.83, "grad_norm": 19.52210655499473, "learning_rate": 6.982735191909168e-06, "loss": 0.5115, "step": 11709 }, { "epoch": 1.83, "grad_norm": 20.525566713777664, "learning_rate": 6.981127431260846e-06, "loss": 0.5047, "step": 11710 }, { "epoch": 1.83, "grad_norm": 15.782272385537256, "learning_rate": 6.9795197564674935e-06, "loss": 0.5157, "step": 11711 }, { "epoch": 1.83, "grad_norm": 28.74157636220329, "learning_rate": 6.977912167574833e-06, "loss": 0.4437, "step": 11712 }, { "epoch": 1.83, "grad_norm": 24.292225681735477, "learning_rate": 6.9763046646285836e-06, "loss": 0.5652, "step": 11713 }, { "epoch": 1.83, "grad_norm": 18.522042738018346, "learning_rate": 6.974697247674462e-06, "loss": 0.5312, "step": 11714 }, { "epoch": 1.83, "grad_norm": 27.193514223376404, "learning_rate": 6.973089916758177e-06, "loss": 0.5129, "step": 11715 }, { "epoch": 1.83, "grad_norm": 20.90361913170271, "learning_rate": 6.971482671925446e-06, "loss": 0.5479, "step": 11716 }, { "epoch": 1.83, "grad_norm": 28.873867614946988, "learning_rate": 6.969875513221974e-06, "loss": 0.5831, "step": 11717 }, { "epoch": 1.83, "grad_norm": 15.782363282017418, "learning_rate": 6.968268440693472e-06, "loss": 0.5187, "step": 11718 }, { "epoch": 1.83, "grad_norm": 20.702251415175613, "learning_rate": 6.966661454385644e-06, "loss": 0.5572, "step": 11719 }, { "epoch": 1.83, "grad_norm": 29.902716813603863, "learning_rate": 6.965054554344188e-06, "loss": 0.5399, "step": 11720 }, { "epoch": 1.83, "grad_norm": 16.625732002420186, "learning_rate": 6.963447740614804e-06, "loss": 0.5546, "step": 11721 }, { "epoch": 1.83, "grad_norm": 19.208926691726003, "learning_rate": 6.96184101324319e-06, "loss": 0.5155, "step": 11722 }, { "epoch": 1.83, "grad_norm": 20.97838514141988, "learning_rate": 6.960234372275039e-06, "loss": 0.4979, "step": 11723 }, { "epoch": 1.83, "grad_norm": 17.449116927089577, "learning_rate": 6.958627817756052e-06, "loss": 0.5685, "step": 11724 }, { "epoch": 1.83, "grad_norm": 18.50965678122747, "learning_rate": 6.957021349731902e-06, "loss": 0.4852, "step": 11725 }, { "epoch": 1.83, "grad_norm": 27.984516194560854, "learning_rate": 6.955414968248289e-06, "loss": 0.5442, "step": 11726 }, { "epoch": 1.83, "grad_norm": 16.848203408271477, "learning_rate": 6.953808673350889e-06, "loss": 0.5545, "step": 11727 }, { "epoch": 1.83, "grad_norm": 21.553414325506136, "learning_rate": 6.952202465085391e-06, "loss": 0.5492, "step": 11728 }, { "epoch": 1.83, "grad_norm": 23.345977558031564, "learning_rate": 6.950596343497475e-06, "loss": 0.5099, "step": 11729 }, { "epoch": 1.83, "grad_norm": 24.98191970118878, "learning_rate": 6.9489903086328106e-06, "loss": 0.5756, "step": 11730 }, { "epoch": 1.83, "grad_norm": 35.24253605048893, "learning_rate": 6.94738436053708e-06, "loss": 0.4795, "step": 11731 }, { "epoch": 1.83, "grad_norm": 21.177290150022948, "learning_rate": 6.945778499255951e-06, "loss": 0.5683, "step": 11732 }, { "epoch": 1.83, "grad_norm": 21.957408071805656, "learning_rate": 6.944172724835093e-06, "loss": 0.6336, "step": 11733 }, { "epoch": 1.83, "grad_norm": 22.72423147978298, "learning_rate": 6.94256703732018e-06, "loss": 0.5884, "step": 11734 }, { "epoch": 1.83, "grad_norm": 26.55956095816552, "learning_rate": 6.94096143675687e-06, "loss": 0.5428, "step": 11735 }, { "epoch": 1.83, "grad_norm": 22.664277537580542, "learning_rate": 6.939355923190823e-06, "loss": 0.5702, "step": 11736 }, { "epoch": 1.83, "grad_norm": 14.13854389946916, "learning_rate": 6.9377504966677075e-06, "loss": 0.4787, "step": 11737 }, { "epoch": 1.83, "grad_norm": 11.473880337545149, "learning_rate": 6.936145157233176e-06, "loss": 0.5476, "step": 11738 }, { "epoch": 1.83, "grad_norm": 19.92305827347941, "learning_rate": 6.934539904932887e-06, "loss": 0.5446, "step": 11739 }, { "epoch": 1.83, "grad_norm": 17.016222793545023, "learning_rate": 6.9329347398124825e-06, "loss": 0.4996, "step": 11740 }, { "epoch": 1.83, "grad_norm": 24.33511644389723, "learning_rate": 6.9313296619176255e-06, "loss": 0.5833, "step": 11741 }, { "epoch": 1.83, "grad_norm": 15.728913362676044, "learning_rate": 6.929724671293954e-06, "loss": 0.5381, "step": 11742 }, { "epoch": 1.83, "grad_norm": 14.216073908335357, "learning_rate": 6.928119767987118e-06, "loss": 0.5825, "step": 11743 }, { "epoch": 1.83, "grad_norm": 15.530455562260306, "learning_rate": 6.926514952042762e-06, "loss": 0.5411, "step": 11744 }, { "epoch": 1.83, "grad_norm": 15.774554030780376, "learning_rate": 6.9249102235065205e-06, "loss": 0.5609, "step": 11745 }, { "epoch": 1.83, "grad_norm": 21.532517514202322, "learning_rate": 6.923305582424031e-06, "loss": 0.4781, "step": 11746 }, { "epoch": 1.83, "grad_norm": 28.746084930422715, "learning_rate": 6.9217010288409335e-06, "loss": 0.5183, "step": 11747 }, { "epoch": 1.84, "grad_norm": 22.037601828894335, "learning_rate": 6.920096562802856e-06, "loss": 0.5638, "step": 11748 }, { "epoch": 1.84, "grad_norm": 25.63200652236906, "learning_rate": 6.918492184355434e-06, "loss": 0.5939, "step": 11749 }, { "epoch": 1.84, "grad_norm": 20.951470353109308, "learning_rate": 6.9168878935442905e-06, "loss": 0.5804, "step": 11750 }, { "epoch": 1.84, "grad_norm": 17.41521443287422, "learning_rate": 6.915283690415051e-06, "loss": 0.6116, "step": 11751 }, { "epoch": 1.84, "grad_norm": 19.170861591270473, "learning_rate": 6.913679575013338e-06, "loss": 0.5257, "step": 11752 }, { "epoch": 1.84, "grad_norm": 18.677820440642755, "learning_rate": 6.9120755473847755e-06, "loss": 0.5979, "step": 11753 }, { "epoch": 1.84, "grad_norm": 15.912057962892913, "learning_rate": 6.9104716075749755e-06, "loss": 0.5724, "step": 11754 }, { "epoch": 1.84, "grad_norm": 19.12830214182361, "learning_rate": 6.908867755629555e-06, "loss": 0.4904, "step": 11755 }, { "epoch": 1.84, "grad_norm": 17.174629708546878, "learning_rate": 6.907263991594129e-06, "loss": 0.4541, "step": 11756 }, { "epoch": 1.84, "grad_norm": 22.04055811483667, "learning_rate": 6.905660315514304e-06, "loss": 0.554, "step": 11757 }, { "epoch": 1.84, "grad_norm": 12.510908032083996, "learning_rate": 6.904056727435694e-06, "loss": 0.4692, "step": 11758 }, { "epoch": 1.84, "grad_norm": 22.405438107846976, "learning_rate": 6.9024532274038935e-06, "loss": 0.5845, "step": 11759 }, { "epoch": 1.84, "grad_norm": 49.87042182361547, "learning_rate": 6.900849815464515e-06, "loss": 0.5984, "step": 11760 }, { "epoch": 1.84, "grad_norm": 17.197483714257224, "learning_rate": 6.8992464916631515e-06, "loss": 0.5514, "step": 11761 }, { "epoch": 1.84, "grad_norm": 19.98222614662644, "learning_rate": 6.897643256045407e-06, "loss": 0.5275, "step": 11762 }, { "epoch": 1.84, "grad_norm": 14.394025212454915, "learning_rate": 6.896040108656874e-06, "loss": 0.5384, "step": 11763 }, { "epoch": 1.84, "grad_norm": 22.957535848675242, "learning_rate": 6.894437049543144e-06, "loss": 0.5513, "step": 11764 }, { "epoch": 1.84, "grad_norm": 17.594980462785127, "learning_rate": 6.892834078749805e-06, "loss": 0.5522, "step": 11765 }, { "epoch": 1.84, "grad_norm": 28.172844572303415, "learning_rate": 6.89123119632245e-06, "loss": 0.5277, "step": 11766 }, { "epoch": 1.84, "grad_norm": 28.18619036645395, "learning_rate": 6.8896284023066585e-06, "loss": 0.5676, "step": 11767 }, { "epoch": 1.84, "grad_norm": 18.165212246182225, "learning_rate": 6.888025696748021e-06, "loss": 0.5836, "step": 11768 }, { "epoch": 1.84, "grad_norm": 19.903844497147052, "learning_rate": 6.886423079692106e-06, "loss": 0.4911, "step": 11769 }, { "epoch": 1.84, "grad_norm": 15.889666266293897, "learning_rate": 6.884820551184501e-06, "loss": 0.5259, "step": 11770 }, { "epoch": 1.84, "grad_norm": 59.62691749432491, "learning_rate": 6.883218111270774e-06, "loss": 0.6656, "step": 11771 }, { "epoch": 1.84, "grad_norm": 19.16000784953565, "learning_rate": 6.8816157599965025e-06, "loss": 0.5916, "step": 11772 }, { "epoch": 1.84, "grad_norm": 22.41139580947437, "learning_rate": 6.880013497407257e-06, "loss": 0.5374, "step": 11773 }, { "epoch": 1.84, "grad_norm": 22.153799914221107, "learning_rate": 6.878411323548599e-06, "loss": 0.565, "step": 11774 }, { "epoch": 1.84, "grad_norm": 14.224518948631207, "learning_rate": 6.8768092384660975e-06, "loss": 0.5482, "step": 11775 }, { "epoch": 1.84, "grad_norm": 20.499790664511192, "learning_rate": 6.875207242205316e-06, "loss": 0.544, "step": 11776 }, { "epoch": 1.84, "grad_norm": 19.849667799125257, "learning_rate": 6.8736053348118106e-06, "loss": 0.552, "step": 11777 }, { "epoch": 1.84, "grad_norm": 18.69364399524402, "learning_rate": 6.8720035163311425e-06, "loss": 0.5775, "step": 11778 }, { "epoch": 1.84, "grad_norm": 18.042298678013218, "learning_rate": 6.870401786808865e-06, "loss": 0.5117, "step": 11779 }, { "epoch": 1.84, "grad_norm": 23.697660090276667, "learning_rate": 6.868800146290526e-06, "loss": 0.5191, "step": 11780 }, { "epoch": 1.84, "grad_norm": 22.543585380670198, "learning_rate": 6.867198594821683e-06, "loss": 0.6217, "step": 11781 }, { "epoch": 1.84, "grad_norm": 22.137636251504162, "learning_rate": 6.8655971324478764e-06, "loss": 0.5715, "step": 11782 }, { "epoch": 1.84, "grad_norm": 17.380098124966214, "learning_rate": 6.863995759214659e-06, "loss": 0.5126, "step": 11783 }, { "epoch": 1.84, "grad_norm": 17.56032807183403, "learning_rate": 6.862394475167562e-06, "loss": 0.6195, "step": 11784 }, { "epoch": 1.84, "grad_norm": 18.78594554179766, "learning_rate": 6.860793280352132e-06, "loss": 0.4863, "step": 11785 }, { "epoch": 1.84, "grad_norm": 15.041057080859524, "learning_rate": 6.859192174813905e-06, "loss": 0.5814, "step": 11786 }, { "epoch": 1.84, "grad_norm": 12.23394728968936, "learning_rate": 6.857591158598415e-06, "loss": 0.4819, "step": 11787 }, { "epoch": 1.84, "grad_norm": 15.65858811681614, "learning_rate": 6.855990231751197e-06, "loss": 0.5488, "step": 11788 }, { "epoch": 1.84, "grad_norm": 30.701395605851804, "learning_rate": 6.854389394317776e-06, "loss": 0.6204, "step": 11789 }, { "epoch": 1.84, "grad_norm": 15.008493917346742, "learning_rate": 6.852788646343678e-06, "loss": 0.5578, "step": 11790 }, { "epoch": 1.84, "grad_norm": 21.51846641001329, "learning_rate": 6.851187987874432e-06, "loss": 0.5132, "step": 11791 }, { "epoch": 1.84, "grad_norm": 16.708946223078858, "learning_rate": 6.849587418955554e-06, "loss": 0.5397, "step": 11792 }, { "epoch": 1.84, "grad_norm": 11.564967495303193, "learning_rate": 6.847986939632574e-06, "loss": 0.4913, "step": 11793 }, { "epoch": 1.84, "grad_norm": 18.902148200622115, "learning_rate": 6.846386549950998e-06, "loss": 0.5779, "step": 11794 }, { "epoch": 1.84, "grad_norm": 16.169481890982244, "learning_rate": 6.844786249956343e-06, "loss": 0.5989, "step": 11795 }, { "epoch": 1.84, "grad_norm": 15.29195636306816, "learning_rate": 6.8431860396941205e-06, "loss": 0.5063, "step": 11796 }, { "epoch": 1.84, "grad_norm": 16.995587490005533, "learning_rate": 6.841585919209842e-06, "loss": 0.4843, "step": 11797 }, { "epoch": 1.84, "grad_norm": 15.299465661928968, "learning_rate": 6.839985888549015e-06, "loss": 0.4884, "step": 11798 }, { "epoch": 1.84, "grad_norm": 18.042442160481034, "learning_rate": 6.838385947757136e-06, "loss": 0.4733, "step": 11799 }, { "epoch": 1.84, "grad_norm": 11.234073573858344, "learning_rate": 6.836786096879713e-06, "loss": 0.4793, "step": 11800 }, { "epoch": 1.84, "grad_norm": 21.637346360869344, "learning_rate": 6.8351863359622424e-06, "loss": 0.5195, "step": 11801 }, { "epoch": 1.84, "grad_norm": 19.11047107254283, "learning_rate": 6.833586665050225e-06, "loss": 0.5508, "step": 11802 }, { "epoch": 1.84, "grad_norm": 16.622951219543683, "learning_rate": 6.831987084189144e-06, "loss": 0.5396, "step": 11803 }, { "epoch": 1.84, "grad_norm": 21.988993444953024, "learning_rate": 6.830387593424501e-06, "loss": 0.5571, "step": 11804 }, { "epoch": 1.84, "grad_norm": 14.64473752405375, "learning_rate": 6.828788192801779e-06, "loss": 0.5243, "step": 11805 }, { "epoch": 1.84, "grad_norm": 22.002541021424317, "learning_rate": 6.827188882366467e-06, "loss": 0.5972, "step": 11806 }, { "epoch": 1.84, "grad_norm": 13.862318553427183, "learning_rate": 6.825589662164049e-06, "loss": 0.4388, "step": 11807 }, { "epoch": 1.84, "grad_norm": 17.240215211181084, "learning_rate": 6.823990532240001e-06, "loss": 0.5396, "step": 11808 }, { "epoch": 1.84, "grad_norm": 21.415745600830192, "learning_rate": 6.822391492639804e-06, "loss": 0.5604, "step": 11809 }, { "epoch": 1.84, "grad_norm": 19.99398208284026, "learning_rate": 6.820792543408935e-06, "loss": 0.5525, "step": 11810 }, { "epoch": 1.84, "grad_norm": 12.17063712961563, "learning_rate": 6.819193684592864e-06, "loss": 0.5205, "step": 11811 }, { "epoch": 1.85, "grad_norm": 14.598539516072515, "learning_rate": 6.817594916237071e-06, "loss": 0.5296, "step": 11812 }, { "epoch": 1.85, "grad_norm": 19.106494857291004, "learning_rate": 6.815996238387011e-06, "loss": 0.5202, "step": 11813 }, { "epoch": 1.85, "grad_norm": 25.761730921744594, "learning_rate": 6.814397651088156e-06, "loss": 0.5598, "step": 11814 }, { "epoch": 1.85, "grad_norm": 13.505484719089626, "learning_rate": 6.812799154385967e-06, "loss": 0.5047, "step": 11815 }, { "epoch": 1.85, "grad_norm": 26.217824920729925, "learning_rate": 6.811200748325908e-06, "loss": 0.5667, "step": 11816 }, { "epoch": 1.85, "grad_norm": 16.357447255129127, "learning_rate": 6.8096024329534376e-06, "loss": 0.5417, "step": 11817 }, { "epoch": 1.85, "grad_norm": 19.722383620791216, "learning_rate": 6.808004208314001e-06, "loss": 0.5542, "step": 11818 }, { "epoch": 1.85, "grad_norm": 14.253706437495703, "learning_rate": 6.806406074453061e-06, "loss": 0.5359, "step": 11819 }, { "epoch": 1.85, "grad_norm": 25.598648277932625, "learning_rate": 6.804808031416064e-06, "loss": 0.5253, "step": 11820 }, { "epoch": 1.85, "grad_norm": 28.686280740433627, "learning_rate": 6.803210079248454e-06, "loss": 0.5844, "step": 11821 }, { "epoch": 1.85, "grad_norm": 17.462911925171134, "learning_rate": 6.8016122179956855e-06, "loss": 0.5614, "step": 11822 }, { "epoch": 1.85, "grad_norm": 24.004407491641857, "learning_rate": 6.800014447703191e-06, "loss": 0.6014, "step": 11823 }, { "epoch": 1.85, "grad_norm": 24.385313763184758, "learning_rate": 6.798416768416412e-06, "loss": 0.5567, "step": 11824 }, { "epoch": 1.85, "grad_norm": 24.957210859857593, "learning_rate": 6.7968191801807894e-06, "loss": 0.5931, "step": 11825 }, { "epoch": 1.85, "grad_norm": 20.17053963594402, "learning_rate": 6.795221683041756e-06, "loss": 0.5224, "step": 11826 }, { "epoch": 1.85, "grad_norm": 25.049286904697535, "learning_rate": 6.793624277044745e-06, "loss": 0.6028, "step": 11827 }, { "epoch": 1.85, "grad_norm": 21.505051819156407, "learning_rate": 6.792026962235178e-06, "loss": 0.551, "step": 11828 }, { "epoch": 1.85, "grad_norm": 17.627967005178707, "learning_rate": 6.790429738658493e-06, "loss": 0.5519, "step": 11829 }, { "epoch": 1.85, "grad_norm": 34.89721143971475, "learning_rate": 6.788832606360104e-06, "loss": 0.5496, "step": 11830 }, { "epoch": 1.85, "grad_norm": 21.622222009458074, "learning_rate": 6.787235565385439e-06, "loss": 0.5669, "step": 11831 }, { "epoch": 1.85, "grad_norm": 14.831788367733381, "learning_rate": 6.785638615779919e-06, "loss": 0.4961, "step": 11832 }, { "epoch": 1.85, "grad_norm": 15.221369891869672, "learning_rate": 6.784041757588954e-06, "loss": 0.4772, "step": 11833 }, { "epoch": 1.85, "grad_norm": 19.665110942426757, "learning_rate": 6.782444990857957e-06, "loss": 0.5192, "step": 11834 }, { "epoch": 1.85, "grad_norm": 20.58718310719083, "learning_rate": 6.780848315632344e-06, "loss": 0.5618, "step": 11835 }, { "epoch": 1.85, "grad_norm": 15.653080811745662, "learning_rate": 6.779251731957522e-06, "loss": 0.4769, "step": 11836 }, { "epoch": 1.85, "grad_norm": 19.05049685073707, "learning_rate": 6.7776552398788995e-06, "loss": 0.5144, "step": 11837 }, { "epoch": 1.85, "grad_norm": 25.362459280753125, "learning_rate": 6.776058839441875e-06, "loss": 0.6139, "step": 11838 }, { "epoch": 1.85, "grad_norm": 20.530316306937102, "learning_rate": 6.77446253069185e-06, "loss": 0.5723, "step": 11839 }, { "epoch": 1.85, "grad_norm": 18.7788743057357, "learning_rate": 6.772866313674223e-06, "loss": 0.5287, "step": 11840 }, { "epoch": 1.85, "grad_norm": 23.03395121596905, "learning_rate": 6.771270188434392e-06, "loss": 0.5809, "step": 11841 }, { "epoch": 1.85, "grad_norm": 21.256488767245816, "learning_rate": 6.7696741550177505e-06, "loss": 0.5326, "step": 11842 }, { "epoch": 1.85, "grad_norm": 14.767931118769736, "learning_rate": 6.768078213469682e-06, "loss": 0.526, "step": 11843 }, { "epoch": 1.85, "grad_norm": 17.13373259923442, "learning_rate": 6.76648236383558e-06, "loss": 0.5511, "step": 11844 }, { "epoch": 1.85, "grad_norm": 23.087789328284906, "learning_rate": 6.764886606160828e-06, "loss": 0.6134, "step": 11845 }, { "epoch": 1.85, "grad_norm": 18.54309932356595, "learning_rate": 6.763290940490807e-06, "loss": 0.5314, "step": 11846 }, { "epoch": 1.85, "grad_norm": 26.978805143862637, "learning_rate": 6.761695366870902e-06, "loss": 0.5518, "step": 11847 }, { "epoch": 1.85, "grad_norm": 23.006521624635567, "learning_rate": 6.760099885346485e-06, "loss": 0.5675, "step": 11848 }, { "epoch": 1.85, "grad_norm": 19.727629834750847, "learning_rate": 6.758504495962929e-06, "loss": 0.5407, "step": 11849 }, { "epoch": 1.85, "grad_norm": 17.421520238950748, "learning_rate": 6.756909198765611e-06, "loss": 0.5079, "step": 11850 }, { "epoch": 1.85, "grad_norm": 23.702382093039137, "learning_rate": 6.755313993799901e-06, "loss": 0.5862, "step": 11851 }, { "epoch": 1.85, "grad_norm": 18.637445580775932, "learning_rate": 6.75371888111116e-06, "loss": 0.598, "step": 11852 }, { "epoch": 1.85, "grad_norm": 18.19318262773005, "learning_rate": 6.7521238607447536e-06, "loss": 0.5322, "step": 11853 }, { "epoch": 1.85, "grad_norm": 18.023378612685022, "learning_rate": 6.750528932746047e-06, "loss": 0.5321, "step": 11854 }, { "epoch": 1.85, "grad_norm": 21.177275724017136, "learning_rate": 6.7489340971603935e-06, "loss": 0.5131, "step": 11855 }, { "epoch": 1.85, "grad_norm": 24.2220833791688, "learning_rate": 6.747339354033158e-06, "loss": 0.591, "step": 11856 }, { "epoch": 1.85, "grad_norm": 15.46797529746724, "learning_rate": 6.745744703409685e-06, "loss": 0.4741, "step": 11857 }, { "epoch": 1.85, "grad_norm": 14.043386639357967, "learning_rate": 6.74415014533533e-06, "loss": 0.5669, "step": 11858 }, { "epoch": 1.85, "grad_norm": 13.599418667835762, "learning_rate": 6.742555679855438e-06, "loss": 0.5139, "step": 11859 }, { "epoch": 1.85, "grad_norm": 18.676745969282397, "learning_rate": 6.740961307015359e-06, "loss": 0.5714, "step": 11860 }, { "epoch": 1.85, "grad_norm": 18.501435230617314, "learning_rate": 6.739367026860435e-06, "loss": 0.5779, "step": 11861 }, { "epoch": 1.85, "grad_norm": 22.944800473462895, "learning_rate": 6.737772839436003e-06, "loss": 0.5178, "step": 11862 }, { "epoch": 1.85, "grad_norm": 17.066471334164344, "learning_rate": 6.736178744787403e-06, "loss": 0.5734, "step": 11863 }, { "epoch": 1.85, "grad_norm": 17.15696972152039, "learning_rate": 6.734584742959972e-06, "loss": 0.5637, "step": 11864 }, { "epoch": 1.85, "grad_norm": 19.14204942502275, "learning_rate": 6.732990833999038e-06, "loss": 0.5619, "step": 11865 }, { "epoch": 1.85, "grad_norm": 27.13584361484866, "learning_rate": 6.731397017949937e-06, "loss": 0.5247, "step": 11866 }, { "epoch": 1.85, "grad_norm": 13.631445407909792, "learning_rate": 6.7298032948579925e-06, "loss": 0.4517, "step": 11867 }, { "epoch": 1.85, "grad_norm": 16.83198742297547, "learning_rate": 6.728209664768525e-06, "loss": 0.5518, "step": 11868 }, { "epoch": 1.85, "grad_norm": 18.781420560544717, "learning_rate": 6.726616127726863e-06, "loss": 0.4721, "step": 11869 }, { "epoch": 1.85, "grad_norm": 17.300662697260762, "learning_rate": 6.725022683778326e-06, "loss": 0.524, "step": 11870 }, { "epoch": 1.85, "grad_norm": 15.412647118527575, "learning_rate": 6.723429332968227e-06, "loss": 0.4846, "step": 11871 }, { "epoch": 1.85, "grad_norm": 25.994617002918577, "learning_rate": 6.721836075341879e-06, "loss": 0.5776, "step": 11872 }, { "epoch": 1.85, "grad_norm": 16.225958231151278, "learning_rate": 6.720242910944596e-06, "loss": 0.5125, "step": 11873 }, { "epoch": 1.85, "grad_norm": 14.809634243317463, "learning_rate": 6.718649839821686e-06, "loss": 0.5774, "step": 11874 }, { "epoch": 1.85, "grad_norm": 21.659502838570962, "learning_rate": 6.717056862018456e-06, "loss": 0.5375, "step": 11875 }, { "epoch": 1.86, "grad_norm": 18.93651919531852, "learning_rate": 6.71546397758021e-06, "loss": 0.4983, "step": 11876 }, { "epoch": 1.86, "grad_norm": 32.62854315940374, "learning_rate": 6.713871186552246e-06, "loss": 0.5915, "step": 11877 }, { "epoch": 1.86, "grad_norm": 17.088300326161473, "learning_rate": 6.7122784889798596e-06, "loss": 0.4891, "step": 11878 }, { "epoch": 1.86, "grad_norm": 15.14300445188734, "learning_rate": 6.710685884908352e-06, "loss": 0.5869, "step": 11879 }, { "epoch": 1.86, "grad_norm": 15.329452902371624, "learning_rate": 6.709093374383012e-06, "loss": 0.5971, "step": 11880 }, { "epoch": 1.86, "grad_norm": 19.83084185153524, "learning_rate": 6.7075009574491355e-06, "loss": 0.4683, "step": 11881 }, { "epoch": 1.86, "grad_norm": 16.93847415733815, "learning_rate": 6.7059086341520045e-06, "loss": 0.5264, "step": 11882 }, { "epoch": 1.86, "grad_norm": 16.919098557608034, "learning_rate": 6.704316404536904e-06, "loss": 0.523, "step": 11883 }, { "epoch": 1.86, "grad_norm": 16.745501041267495, "learning_rate": 6.702724268649114e-06, "loss": 0.5162, "step": 11884 }, { "epoch": 1.86, "grad_norm": 20.12752802442818, "learning_rate": 6.70113222653392e-06, "loss": 0.5366, "step": 11885 }, { "epoch": 1.86, "grad_norm": 22.86254287679728, "learning_rate": 6.699540278236598e-06, "loss": 0.536, "step": 11886 }, { "epoch": 1.86, "grad_norm": 31.498146309124703, "learning_rate": 6.697948423802416e-06, "loss": 0.5708, "step": 11887 }, { "epoch": 1.86, "grad_norm": 22.63014622476013, "learning_rate": 6.696356663276651e-06, "loss": 0.5801, "step": 11888 }, { "epoch": 1.86, "grad_norm": 23.92612413246314, "learning_rate": 6.694764996704569e-06, "loss": 0.6027, "step": 11889 }, { "epoch": 1.86, "grad_norm": 18.431043361548078, "learning_rate": 6.6931734241314365e-06, "loss": 0.5927, "step": 11890 }, { "epoch": 1.86, "grad_norm": 16.900931632374874, "learning_rate": 6.69158194560252e-06, "loss": 0.5302, "step": 11891 }, { "epoch": 1.86, "grad_norm": 15.44972644217673, "learning_rate": 6.689990561163075e-06, "loss": 0.483, "step": 11892 }, { "epoch": 1.86, "grad_norm": 25.53810183053534, "learning_rate": 6.688399270858362e-06, "loss": 0.5208, "step": 11893 }, { "epoch": 1.86, "grad_norm": 14.32393672489673, "learning_rate": 6.686808074733638e-06, "loss": 0.6336, "step": 11894 }, { "epoch": 1.86, "grad_norm": 14.550682988187054, "learning_rate": 6.6852169728341565e-06, "loss": 0.5505, "step": 11895 }, { "epoch": 1.86, "grad_norm": 28.163360533517643, "learning_rate": 6.683625965205162e-06, "loss": 0.5482, "step": 11896 }, { "epoch": 1.86, "grad_norm": 18.439995082846295, "learning_rate": 6.6820350518919044e-06, "loss": 0.5604, "step": 11897 }, { "epoch": 1.86, "grad_norm": 13.926460986782294, "learning_rate": 6.680444232939629e-06, "loss": 0.5129, "step": 11898 }, { "epoch": 1.86, "grad_norm": 13.509597936661928, "learning_rate": 6.6788535083935764e-06, "loss": 0.4443, "step": 11899 }, { "epoch": 1.86, "grad_norm": 17.854731308975214, "learning_rate": 6.677262878298993e-06, "loss": 0.5309, "step": 11900 }, { "epoch": 1.86, "grad_norm": 18.2357011253982, "learning_rate": 6.675672342701107e-06, "loss": 0.5588, "step": 11901 }, { "epoch": 1.86, "grad_norm": 14.498992077658713, "learning_rate": 6.674081901645153e-06, "loss": 0.5059, "step": 11902 }, { "epoch": 1.86, "grad_norm": 23.24220808973923, "learning_rate": 6.672491555176363e-06, "loss": 0.548, "step": 11903 }, { "epoch": 1.86, "grad_norm": 23.62993069630609, "learning_rate": 6.6709013033399695e-06, "loss": 0.5014, "step": 11904 }, { "epoch": 1.86, "grad_norm": 16.23498499360388, "learning_rate": 6.6693111461811975e-06, "loss": 0.5233, "step": 11905 }, { "epoch": 1.86, "grad_norm": 13.433866590705799, "learning_rate": 6.6677210837452645e-06, "loss": 0.5078, "step": 11906 }, { "epoch": 1.86, "grad_norm": 15.8680186939078, "learning_rate": 6.6661311160773945e-06, "loss": 0.531, "step": 11907 }, { "epoch": 1.86, "grad_norm": 27.097072217239205, "learning_rate": 6.664541243222806e-06, "loss": 0.5464, "step": 11908 }, { "epoch": 1.86, "grad_norm": 19.21038081373631, "learning_rate": 6.662951465226713e-06, "loss": 0.5657, "step": 11909 }, { "epoch": 1.86, "grad_norm": 20.1176575815036, "learning_rate": 6.66136178213433e-06, "loss": 0.5954, "step": 11910 }, { "epoch": 1.86, "grad_norm": 13.379564099253406, "learning_rate": 6.659772193990865e-06, "loss": 0.4821, "step": 11911 }, { "epoch": 1.86, "grad_norm": 20.86837813145166, "learning_rate": 6.658182700841521e-06, "loss": 0.5862, "step": 11912 }, { "epoch": 1.86, "grad_norm": 24.83615664193141, "learning_rate": 6.656593302731509e-06, "loss": 0.5975, "step": 11913 }, { "epoch": 1.86, "grad_norm": 27.50810701342046, "learning_rate": 6.655003999706027e-06, "loss": 0.5414, "step": 11914 }, { "epoch": 1.86, "grad_norm": 26.595309570946146, "learning_rate": 6.653414791810277e-06, "loss": 0.5938, "step": 11915 }, { "epoch": 1.86, "grad_norm": 21.01384182990767, "learning_rate": 6.6518256790894484e-06, "loss": 0.5025, "step": 11916 }, { "epoch": 1.86, "grad_norm": 14.942821893884318, "learning_rate": 6.6502366615887415e-06, "loss": 0.4598, "step": 11917 }, { "epoch": 1.86, "grad_norm": 16.74590904804912, "learning_rate": 6.648647739353342e-06, "loss": 0.6213, "step": 11918 }, { "epoch": 1.86, "grad_norm": 20.531460046782406, "learning_rate": 6.6470589124284414e-06, "loss": 0.5528, "step": 11919 }, { "epoch": 1.86, "grad_norm": 19.49027521427508, "learning_rate": 6.645470180859227e-06, "loss": 0.5561, "step": 11920 }, { "epoch": 1.86, "grad_norm": 21.957432561850307, "learning_rate": 6.643881544690876e-06, "loss": 0.5572, "step": 11921 }, { "epoch": 1.86, "grad_norm": 23.841515348867155, "learning_rate": 6.6422930039685695e-06, "loss": 0.5858, "step": 11922 }, { "epoch": 1.86, "grad_norm": 14.962875077399254, "learning_rate": 6.640704558737487e-06, "loss": 0.5437, "step": 11923 }, { "epoch": 1.86, "grad_norm": 21.801763986755386, "learning_rate": 6.639116209042798e-06, "loss": 0.5837, "step": 11924 }, { "epoch": 1.86, "grad_norm": 14.553471164792626, "learning_rate": 6.637527954929685e-06, "loss": 0.5125, "step": 11925 }, { "epoch": 1.86, "grad_norm": 18.343271173697165, "learning_rate": 6.635939796443306e-06, "loss": 0.5093, "step": 11926 }, { "epoch": 1.86, "grad_norm": 16.16702925649879, "learning_rate": 6.634351733628832e-06, "loss": 0.4375, "step": 11927 }, { "epoch": 1.86, "grad_norm": 15.231680656249214, "learning_rate": 6.632763766531423e-06, "loss": 0.5767, "step": 11928 }, { "epoch": 1.86, "grad_norm": 14.807245196131904, "learning_rate": 6.631175895196245e-06, "loss": 0.4982, "step": 11929 }, { "epoch": 1.86, "grad_norm": 23.58400584091568, "learning_rate": 6.629588119668458e-06, "loss": 0.5558, "step": 11930 }, { "epoch": 1.86, "grad_norm": 25.211921388987264, "learning_rate": 6.628000439993207e-06, "loss": 0.581, "step": 11931 }, { "epoch": 1.86, "grad_norm": 15.20751505527256, "learning_rate": 6.626412856215653e-06, "loss": 0.5769, "step": 11932 }, { "epoch": 1.86, "grad_norm": 23.05257301884939, "learning_rate": 6.624825368380944e-06, "loss": 0.5327, "step": 11933 }, { "epoch": 1.86, "grad_norm": 16.78064218766533, "learning_rate": 6.623237976534225e-06, "loss": 0.5307, "step": 11934 }, { "epoch": 1.86, "grad_norm": 20.038876802576763, "learning_rate": 6.6216506807206475e-06, "loss": 0.5385, "step": 11935 }, { "epoch": 1.86, "grad_norm": 27.964040149215766, "learning_rate": 6.620063480985346e-06, "loss": 0.591, "step": 11936 }, { "epoch": 1.86, "grad_norm": 18.065915284779983, "learning_rate": 6.61847637737346e-06, "loss": 0.5, "step": 11937 }, { "epoch": 1.86, "grad_norm": 25.694112426091873, "learning_rate": 6.6168893699301286e-06, "loss": 0.6282, "step": 11938 }, { "epoch": 1.86, "grad_norm": 22.02899333407536, "learning_rate": 6.6153024587004855e-06, "loss": 0.5439, "step": 11939 }, { "epoch": 1.87, "grad_norm": 21.54619790065212, "learning_rate": 6.613715643729661e-06, "loss": 0.52, "step": 11940 }, { "epoch": 1.87, "grad_norm": 14.125489193222258, "learning_rate": 6.612128925062781e-06, "loss": 0.5128, "step": 11941 }, { "epoch": 1.87, "grad_norm": 15.600904049204516, "learning_rate": 6.610542302744973e-06, "loss": 0.4713, "step": 11942 }, { "epoch": 1.87, "grad_norm": 16.244118339483844, "learning_rate": 6.608955776821357e-06, "loss": 0.5841, "step": 11943 }, { "epoch": 1.87, "grad_norm": 18.010179813372737, "learning_rate": 6.60736934733706e-06, "loss": 0.5324, "step": 11944 }, { "epoch": 1.87, "grad_norm": 20.17303308343498, "learning_rate": 6.605783014337192e-06, "loss": 0.58, "step": 11945 }, { "epoch": 1.87, "grad_norm": 24.99588703267362, "learning_rate": 6.6041967778668695e-06, "loss": 0.5321, "step": 11946 }, { "epoch": 1.87, "grad_norm": 12.759150767172045, "learning_rate": 6.602610637971201e-06, "loss": 0.5313, "step": 11947 }, { "epoch": 1.87, "grad_norm": 15.140154150185321, "learning_rate": 6.601024594695302e-06, "loss": 0.5859, "step": 11948 }, { "epoch": 1.87, "grad_norm": 26.09471790037621, "learning_rate": 6.599438648084277e-06, "loss": 0.5384, "step": 11949 }, { "epoch": 1.87, "grad_norm": 22.696077316843915, "learning_rate": 6.597852798183222e-06, "loss": 0.5021, "step": 11950 }, { "epoch": 1.87, "grad_norm": 22.99404933000497, "learning_rate": 6.596267045037247e-06, "loss": 0.4996, "step": 11951 }, { "epoch": 1.87, "grad_norm": 17.844594934604636, "learning_rate": 6.594681388691445e-06, "loss": 0.5581, "step": 11952 }, { "epoch": 1.87, "grad_norm": 18.823244354999975, "learning_rate": 6.593095829190911e-06, "loss": 0.5456, "step": 11953 }, { "epoch": 1.87, "grad_norm": 18.63540224995561, "learning_rate": 6.591510366580742e-06, "loss": 0.5803, "step": 11954 }, { "epoch": 1.87, "grad_norm": 25.34283976913497, "learning_rate": 6.589925000906023e-06, "loss": 0.6194, "step": 11955 }, { "epoch": 1.87, "grad_norm": 19.68611024235323, "learning_rate": 6.588339732211839e-06, "loss": 0.5292, "step": 11956 }, { "epoch": 1.87, "grad_norm": 20.242855904238755, "learning_rate": 6.586754560543281e-06, "loss": 0.5524, "step": 11957 }, { "epoch": 1.87, "grad_norm": 16.19504177016277, "learning_rate": 6.585169485945425e-06, "loss": 0.5544, "step": 11958 }, { "epoch": 1.87, "grad_norm": 17.036393206621653, "learning_rate": 6.583584508463354e-06, "loss": 0.5281, "step": 11959 }, { "epoch": 1.87, "grad_norm": 16.209893802917698, "learning_rate": 6.581999628142137e-06, "loss": 0.5102, "step": 11960 }, { "epoch": 1.87, "grad_norm": 21.44764380716063, "learning_rate": 6.580414845026853e-06, "loss": 0.6065, "step": 11961 }, { "epoch": 1.87, "grad_norm": 14.453680625751677, "learning_rate": 6.578830159162568e-06, "loss": 0.5405, "step": 11962 }, { "epoch": 1.87, "grad_norm": 12.52716847491306, "learning_rate": 6.577245570594355e-06, "loss": 0.5187, "step": 11963 }, { "epoch": 1.87, "grad_norm": 14.793590153852755, "learning_rate": 6.575661079367277e-06, "loss": 0.5581, "step": 11964 }, { "epoch": 1.87, "grad_norm": 16.835090525624217, "learning_rate": 6.574076685526392e-06, "loss": 0.5464, "step": 11965 }, { "epoch": 1.87, "grad_norm": 20.724116884577217, "learning_rate": 6.572492389116761e-06, "loss": 0.5823, "step": 11966 }, { "epoch": 1.87, "grad_norm": 15.399030876242543, "learning_rate": 6.570908190183442e-06, "loss": 0.575, "step": 11967 }, { "epoch": 1.87, "grad_norm": 16.239860788201426, "learning_rate": 6.569324088771488e-06, "loss": 0.5382, "step": 11968 }, { "epoch": 1.87, "grad_norm": 16.204370771723788, "learning_rate": 6.5677400849259535e-06, "loss": 0.5312, "step": 11969 }, { "epoch": 1.87, "grad_norm": 21.074923913451805, "learning_rate": 6.566156178691882e-06, "loss": 0.5588, "step": 11970 }, { "epoch": 1.87, "grad_norm": 13.565930506182092, "learning_rate": 6.564572370114318e-06, "loss": 0.5463, "step": 11971 }, { "epoch": 1.87, "grad_norm": 13.476208829247197, "learning_rate": 6.562988659238305e-06, "loss": 0.4763, "step": 11972 }, { "epoch": 1.87, "grad_norm": 19.253953913921627, "learning_rate": 6.561405046108886e-06, "loss": 0.5453, "step": 11973 }, { "epoch": 1.87, "grad_norm": 31.662751848120546, "learning_rate": 6.5598215307710975e-06, "loss": 0.5514, "step": 11974 }, { "epoch": 1.87, "grad_norm": 16.379043718630413, "learning_rate": 6.55823811326997e-06, "loss": 0.4958, "step": 11975 }, { "epoch": 1.87, "grad_norm": 13.306475433529185, "learning_rate": 6.5566547936505375e-06, "loss": 0.4861, "step": 11976 }, { "epoch": 1.87, "grad_norm": 15.781586888938156, "learning_rate": 6.5550715719578295e-06, "loss": 0.5202, "step": 11977 }, { "epoch": 1.87, "grad_norm": 13.454892737963295, "learning_rate": 6.5534884482368665e-06, "loss": 0.5321, "step": 11978 }, { "epoch": 1.87, "grad_norm": 30.386922081657993, "learning_rate": 6.551905422532683e-06, "loss": 0.5595, "step": 11979 }, { "epoch": 1.87, "grad_norm": 13.129213434522658, "learning_rate": 6.550322494890289e-06, "loss": 0.5749, "step": 11980 }, { "epoch": 1.87, "grad_norm": 31.121168642850527, "learning_rate": 6.548739665354702e-06, "loss": 0.6267, "step": 11981 }, { "epoch": 1.87, "grad_norm": 26.492847846690314, "learning_rate": 6.547156933970943e-06, "loss": 0.4809, "step": 11982 }, { "epoch": 1.87, "grad_norm": 21.42820274279954, "learning_rate": 6.545574300784021e-06, "loss": 0.5006, "step": 11983 }, { "epoch": 1.87, "grad_norm": 13.74027669340231, "learning_rate": 6.543991765838946e-06, "loss": 0.4851, "step": 11984 }, { "epoch": 1.87, "grad_norm": 21.6468925805872, "learning_rate": 6.542409329180721e-06, "loss": 0.5231, "step": 11985 }, { "epoch": 1.87, "grad_norm": 16.411628075991704, "learning_rate": 6.540826990854353e-06, "loss": 0.4959, "step": 11986 }, { "epoch": 1.87, "grad_norm": 19.310887480824633, "learning_rate": 6.53924475090484e-06, "loss": 0.641, "step": 11987 }, { "epoch": 1.87, "grad_norm": 14.091250560901004, "learning_rate": 6.537662609377184e-06, "loss": 0.4852, "step": 11988 }, { "epoch": 1.87, "grad_norm": 14.301149613096703, "learning_rate": 6.53608056631638e-06, "loss": 0.536, "step": 11989 }, { "epoch": 1.87, "grad_norm": 18.0324518614721, "learning_rate": 6.534498621767418e-06, "loss": 0.568, "step": 11990 }, { "epoch": 1.87, "grad_norm": 16.685525064723162, "learning_rate": 6.5329167757752835e-06, "loss": 0.5087, "step": 11991 }, { "epoch": 1.87, "grad_norm": 21.139792904897877, "learning_rate": 6.531335028384971e-06, "loss": 0.5631, "step": 11992 }, { "epoch": 1.87, "grad_norm": 30.70380287198627, "learning_rate": 6.529753379641462e-06, "loss": 0.5151, "step": 11993 }, { "epoch": 1.87, "grad_norm": 15.420217631754886, "learning_rate": 6.528171829589734e-06, "loss": 0.5444, "step": 11994 }, { "epoch": 1.87, "grad_norm": 17.095784423662707, "learning_rate": 6.52659037827477e-06, "loss": 0.5532, "step": 11995 }, { "epoch": 1.87, "grad_norm": 16.702125868921176, "learning_rate": 6.525009025741543e-06, "loss": 0.5825, "step": 11996 }, { "epoch": 1.87, "grad_norm": 22.49969019246063, "learning_rate": 6.523427772035024e-06, "loss": 0.569, "step": 11997 }, { "epoch": 1.87, "grad_norm": 18.672128056363587, "learning_rate": 6.521846617200191e-06, "loss": 0.5452, "step": 11998 }, { "epoch": 1.87, "grad_norm": 20.031964799807888, "learning_rate": 6.520265561282004e-06, "loss": 0.4891, "step": 11999 }, { "epoch": 1.87, "grad_norm": 18.824851233042082, "learning_rate": 6.5186846043254246e-06, "loss": 0.5455, "step": 12000 }, { "epoch": 1.87, "grad_norm": 20.722747281923528, "learning_rate": 6.5171037463754224e-06, "loss": 0.4926, "step": 12001 }, { "epoch": 1.87, "grad_norm": 20.635381943777627, "learning_rate": 6.515522987476952e-06, "loss": 0.5578, "step": 12002 }, { "epoch": 1.87, "grad_norm": 20.64048423798505, "learning_rate": 6.513942327674971e-06, "loss": 0.5128, "step": 12003 }, { "epoch": 1.88, "grad_norm": 16.054445207350017, "learning_rate": 6.512361767014427e-06, "loss": 0.4807, "step": 12004 }, { "epoch": 1.88, "grad_norm": 26.537284136804757, "learning_rate": 6.510781305540277e-06, "loss": 0.5757, "step": 12005 }, { "epoch": 1.88, "grad_norm": 24.2552813390368, "learning_rate": 6.509200943297463e-06, "loss": 0.5545, "step": 12006 }, { "epoch": 1.88, "grad_norm": 18.618808972273953, "learning_rate": 6.507620680330933e-06, "loss": 0.558, "step": 12007 }, { "epoch": 1.88, "grad_norm": 17.196480106087815, "learning_rate": 6.50604051668563e-06, "loss": 0.4874, "step": 12008 }, { "epoch": 1.88, "grad_norm": 18.234040437030032, "learning_rate": 6.504460452406489e-06, "loss": 0.5577, "step": 12009 }, { "epoch": 1.88, "grad_norm": 13.218088589459896, "learning_rate": 6.502880487538448e-06, "loss": 0.5357, "step": 12010 }, { "epoch": 1.88, "grad_norm": 16.23254004343529, "learning_rate": 6.50130062212644e-06, "loss": 0.4843, "step": 12011 }, { "epoch": 1.88, "grad_norm": 18.948171562680933, "learning_rate": 6.499720856215395e-06, "loss": 0.5707, "step": 12012 }, { "epoch": 1.88, "grad_norm": 19.484562485962197, "learning_rate": 6.498141189850243e-06, "loss": 0.5083, "step": 12013 }, { "epoch": 1.88, "grad_norm": 22.168233041406186, "learning_rate": 6.496561623075907e-06, "loss": 0.6047, "step": 12014 }, { "epoch": 1.88, "grad_norm": 15.80088044407739, "learning_rate": 6.494982155937308e-06, "loss": 0.5362, "step": 12015 }, { "epoch": 1.88, "grad_norm": 24.10934125733502, "learning_rate": 6.493402788479365e-06, "loss": 0.614, "step": 12016 }, { "epoch": 1.88, "grad_norm": 15.099278274792237, "learning_rate": 6.491823520746996e-06, "loss": 0.4768, "step": 12017 }, { "epoch": 1.88, "grad_norm": 18.835058640332303, "learning_rate": 6.4902443527851175e-06, "loss": 0.4981, "step": 12018 }, { "epoch": 1.88, "grad_norm": 17.48080520220848, "learning_rate": 6.488665284638631e-06, "loss": 0.5364, "step": 12019 }, { "epoch": 1.88, "grad_norm": 16.531853567924134, "learning_rate": 6.487086316352453e-06, "loss": 0.4992, "step": 12020 }, { "epoch": 1.88, "grad_norm": 23.790879642131458, "learning_rate": 6.4855074479714845e-06, "loss": 0.588, "step": 12021 }, { "epoch": 1.88, "grad_norm": 23.51269961058938, "learning_rate": 6.483928679540627e-06, "loss": 0.5755, "step": 12022 }, { "epoch": 1.88, "grad_norm": 27.226335474214363, "learning_rate": 6.482350011104785e-06, "loss": 0.6003, "step": 12023 }, { "epoch": 1.88, "grad_norm": 18.332537974278345, "learning_rate": 6.480771442708847e-06, "loss": 0.4889, "step": 12024 }, { "epoch": 1.88, "grad_norm": 20.805310986666903, "learning_rate": 6.47919297439771e-06, "loss": 0.581, "step": 12025 }, { "epoch": 1.88, "grad_norm": 20.600358841777485, "learning_rate": 6.477614606216267e-06, "loss": 0.5528, "step": 12026 }, { "epoch": 1.88, "grad_norm": 18.761596828209886, "learning_rate": 6.476036338209404e-06, "loss": 0.613, "step": 12027 }, { "epoch": 1.88, "grad_norm": 16.9691599011309, "learning_rate": 6.474458170422009e-06, "loss": 0.5067, "step": 12028 }, { "epoch": 1.88, "grad_norm": 22.95342617798133, "learning_rate": 6.4728801028989556e-06, "loss": 0.584, "step": 12029 }, { "epoch": 1.88, "grad_norm": 18.458857956363488, "learning_rate": 6.471302135685131e-06, "loss": 0.5122, "step": 12030 }, { "epoch": 1.88, "grad_norm": 20.434727997712663, "learning_rate": 6.4697242688254095e-06, "loss": 0.5729, "step": 12031 }, { "epoch": 1.88, "grad_norm": 14.332635926544123, "learning_rate": 6.468146502364665e-06, "loss": 0.5738, "step": 12032 }, { "epoch": 1.88, "grad_norm": 21.16065222525277, "learning_rate": 6.46656883634777e-06, "loss": 0.5258, "step": 12033 }, { "epoch": 1.88, "grad_norm": 16.332691181118758, "learning_rate": 6.464991270819587e-06, "loss": 0.5065, "step": 12034 }, { "epoch": 1.88, "grad_norm": 21.425591228550076, "learning_rate": 6.463413805824985e-06, "loss": 0.5185, "step": 12035 }, { "epoch": 1.88, "grad_norm": 25.27596465451524, "learning_rate": 6.4618364414088264e-06, "loss": 0.5113, "step": 12036 }, { "epoch": 1.88, "grad_norm": 25.13137946699103, "learning_rate": 6.460259177615966e-06, "loss": 0.5017, "step": 12037 }, { "epoch": 1.88, "grad_norm": 19.078665961157515, "learning_rate": 6.458682014491271e-06, "loss": 0.5111, "step": 12038 }, { "epoch": 1.88, "grad_norm": 22.039267404476618, "learning_rate": 6.4571049520795836e-06, "loss": 0.6013, "step": 12039 }, { "epoch": 1.88, "grad_norm": 16.297309446350216, "learning_rate": 6.45552799042576e-06, "loss": 0.4705, "step": 12040 }, { "epoch": 1.88, "grad_norm": 15.868311334626522, "learning_rate": 6.453951129574644e-06, "loss": 0.5293, "step": 12041 }, { "epoch": 1.88, "grad_norm": 16.002769508056005, "learning_rate": 6.452374369571088e-06, "loss": 0.4852, "step": 12042 }, { "epoch": 1.88, "grad_norm": 14.531897925985582, "learning_rate": 6.4507977104599275e-06, "loss": 0.5412, "step": 12043 }, { "epoch": 1.88, "grad_norm": 15.400456852549771, "learning_rate": 6.449221152286001e-06, "loss": 0.552, "step": 12044 }, { "epoch": 1.88, "grad_norm": 15.919692352216606, "learning_rate": 6.447644695094151e-06, "loss": 0.5007, "step": 12045 }, { "epoch": 1.88, "grad_norm": 15.454047197725393, "learning_rate": 6.446068338929208e-06, "loss": 0.4966, "step": 12046 }, { "epoch": 1.88, "grad_norm": 21.529276509635686, "learning_rate": 6.444492083836004e-06, "loss": 0.5484, "step": 12047 }, { "epoch": 1.88, "grad_norm": 13.080353742944991, "learning_rate": 6.442915929859361e-06, "loss": 0.5258, "step": 12048 }, { "epoch": 1.88, "grad_norm": 17.57137340668942, "learning_rate": 6.44133987704411e-06, "loss": 0.4786, "step": 12049 }, { "epoch": 1.88, "grad_norm": 20.809829112026197, "learning_rate": 6.43976392543507e-06, "loss": 0.5064, "step": 12050 }, { "epoch": 1.88, "grad_norm": 21.35816842140582, "learning_rate": 6.438188075077062e-06, "loss": 0.5983, "step": 12051 }, { "epoch": 1.88, "grad_norm": 13.744822945028577, "learning_rate": 6.436612326014904e-06, "loss": 0.4876, "step": 12052 }, { "epoch": 1.88, "grad_norm": 13.071159078858736, "learning_rate": 6.435036678293405e-06, "loss": 0.4982, "step": 12053 }, { "epoch": 1.88, "grad_norm": 18.206086003062207, "learning_rate": 6.433461131957376e-06, "loss": 0.5777, "step": 12054 }, { "epoch": 1.88, "grad_norm": 13.565210411623358, "learning_rate": 6.431885687051628e-06, "loss": 0.5207, "step": 12055 }, { "epoch": 1.88, "grad_norm": 11.904620525914225, "learning_rate": 6.43031034362096e-06, "loss": 0.521, "step": 12056 }, { "epoch": 1.88, "grad_norm": 23.268963539661176, "learning_rate": 6.428735101710184e-06, "loss": 0.5785, "step": 12057 }, { "epoch": 1.88, "grad_norm": 19.266283307890372, "learning_rate": 6.427159961364089e-06, "loss": 0.5278, "step": 12058 }, { "epoch": 1.88, "grad_norm": 15.946035343048436, "learning_rate": 6.425584922627476e-06, "loss": 0.6558, "step": 12059 }, { "epoch": 1.88, "grad_norm": 20.4069344170661, "learning_rate": 6.424009985545134e-06, "loss": 0.5802, "step": 12060 }, { "epoch": 1.88, "grad_norm": 26.070471611371907, "learning_rate": 6.422435150161859e-06, "loss": 0.5259, "step": 12061 }, { "epoch": 1.88, "grad_norm": 14.002173108227096, "learning_rate": 6.4208604165224366e-06, "loss": 0.5054, "step": 12062 }, { "epoch": 1.88, "grad_norm": 29.44864237138187, "learning_rate": 6.419285784671645e-06, "loss": 0.5329, "step": 12063 }, { "epoch": 1.88, "grad_norm": 14.694834885448381, "learning_rate": 6.417711254654275e-06, "loss": 0.5357, "step": 12064 }, { "epoch": 1.88, "grad_norm": 21.190361391668823, "learning_rate": 6.4161368265151e-06, "loss": 0.5836, "step": 12065 }, { "epoch": 1.88, "grad_norm": 23.61209067726471, "learning_rate": 6.414562500298896e-06, "loss": 0.5106, "step": 12066 }, { "epoch": 1.88, "grad_norm": 23.921850059005422, "learning_rate": 6.412988276050441e-06, "loss": 0.5209, "step": 12067 }, { "epoch": 1.89, "grad_norm": 17.85677272130333, "learning_rate": 6.411414153814499e-06, "loss": 0.5596, "step": 12068 }, { "epoch": 1.89, "grad_norm": 26.65474902585938, "learning_rate": 6.409840133635837e-06, "loss": 0.5578, "step": 12069 }, { "epoch": 1.89, "grad_norm": 23.700509780796263, "learning_rate": 6.408266215559225e-06, "loss": 0.5157, "step": 12070 }, { "epoch": 1.89, "grad_norm": 20.823636016248656, "learning_rate": 6.406692399629418e-06, "loss": 0.5578, "step": 12071 }, { "epoch": 1.89, "grad_norm": 23.114453731968332, "learning_rate": 6.405118685891181e-06, "loss": 0.5969, "step": 12072 }, { "epoch": 1.89, "grad_norm": 25.228801245186332, "learning_rate": 6.403545074389261e-06, "loss": 0.6072, "step": 12073 }, { "epoch": 1.89, "grad_norm": 11.891383374977051, "learning_rate": 6.401971565168416e-06, "loss": 0.5766, "step": 12074 }, { "epoch": 1.89, "grad_norm": 19.5082393992152, "learning_rate": 6.400398158273395e-06, "loss": 0.613, "step": 12075 }, { "epoch": 1.89, "grad_norm": 12.321157001973747, "learning_rate": 6.398824853748946e-06, "loss": 0.491, "step": 12076 }, { "epoch": 1.89, "grad_norm": 23.11424374184085, "learning_rate": 6.397251651639813e-06, "loss": 0.5516, "step": 12077 }, { "epoch": 1.89, "grad_norm": 12.155877556280513, "learning_rate": 6.395678551990735e-06, "loss": 0.5044, "step": 12078 }, { "epoch": 1.89, "grad_norm": 20.834961961666178, "learning_rate": 6.394105554846447e-06, "loss": 0.5645, "step": 12079 }, { "epoch": 1.89, "grad_norm": 27.123841943828932, "learning_rate": 6.3925326602516904e-06, "loss": 0.5897, "step": 12080 }, { "epoch": 1.89, "grad_norm": 28.80286386540562, "learning_rate": 6.390959868251192e-06, "loss": 0.5403, "step": 12081 }, { "epoch": 1.89, "grad_norm": 25.309276944286207, "learning_rate": 6.389387178889688e-06, "loss": 0.5799, "step": 12082 }, { "epoch": 1.89, "grad_norm": 15.868157630875439, "learning_rate": 6.3878145922118985e-06, "loss": 0.5274, "step": 12083 }, { "epoch": 1.89, "grad_norm": 24.53124128967179, "learning_rate": 6.3862421082625495e-06, "loss": 0.4512, "step": 12084 }, { "epoch": 1.89, "grad_norm": 20.752414053072727, "learning_rate": 6.3846697270863576e-06, "loss": 0.557, "step": 12085 }, { "epoch": 1.89, "grad_norm": 23.740643366025363, "learning_rate": 6.383097448728046e-06, "loss": 0.5528, "step": 12086 }, { "epoch": 1.89, "grad_norm": 15.205294359221694, "learning_rate": 6.3815252732323304e-06, "loss": 0.5682, "step": 12087 }, { "epoch": 1.89, "grad_norm": 15.470416003879741, "learning_rate": 6.379953200643915e-06, "loss": 0.5613, "step": 12088 }, { "epoch": 1.89, "grad_norm": 12.122439780259775, "learning_rate": 6.378381231007514e-06, "loss": 0.472, "step": 12089 }, { "epoch": 1.89, "grad_norm": 24.79463706046399, "learning_rate": 6.376809364367833e-06, "loss": 0.5723, "step": 12090 }, { "epoch": 1.89, "grad_norm": 20.655461484572395, "learning_rate": 6.375237600769575e-06, "loss": 0.5336, "step": 12091 }, { "epoch": 1.89, "grad_norm": 17.200802081020502, "learning_rate": 6.373665940257436e-06, "loss": 0.4941, "step": 12092 }, { "epoch": 1.89, "grad_norm": 27.97962465335117, "learning_rate": 6.37209438287612e-06, "loss": 0.5426, "step": 12093 }, { "epoch": 1.89, "grad_norm": 16.244034309401638, "learning_rate": 6.370522928670313e-06, "loss": 0.5486, "step": 12094 }, { "epoch": 1.89, "grad_norm": 23.552523041911183, "learning_rate": 6.368951577684713e-06, "loss": 0.5198, "step": 12095 }, { "epoch": 1.89, "grad_norm": 17.469898056492717, "learning_rate": 6.367380329964009e-06, "loss": 0.534, "step": 12096 }, { "epoch": 1.89, "grad_norm": 19.157528948134143, "learning_rate": 6.36580918555288e-06, "loss": 0.5386, "step": 12097 }, { "epoch": 1.89, "grad_norm": 17.377336372007253, "learning_rate": 6.364238144496011e-06, "loss": 0.5521, "step": 12098 }, { "epoch": 1.89, "grad_norm": 24.245526389650948, "learning_rate": 6.362667206838085e-06, "loss": 0.5953, "step": 12099 }, { "epoch": 1.89, "grad_norm": 16.804114905249342, "learning_rate": 6.361096372623773e-06, "loss": 0.5421, "step": 12100 }, { "epoch": 1.89, "grad_norm": 18.342537276912662, "learning_rate": 6.359525641897755e-06, "loss": 0.5526, "step": 12101 }, { "epoch": 1.89, "grad_norm": 13.77301974320269, "learning_rate": 6.357955014704698e-06, "loss": 0.4785, "step": 12102 }, { "epoch": 1.89, "grad_norm": 18.091841981690127, "learning_rate": 6.356384491089267e-06, "loss": 0.4871, "step": 12103 }, { "epoch": 1.89, "grad_norm": 24.33809527510175, "learning_rate": 6.354814071096129e-06, "loss": 0.593, "step": 12104 }, { "epoch": 1.89, "grad_norm": 17.102190394788074, "learning_rate": 6.353243754769948e-06, "loss": 0.5027, "step": 12105 }, { "epoch": 1.89, "grad_norm": 18.060306430312203, "learning_rate": 6.351673542155385e-06, "loss": 0.4962, "step": 12106 }, { "epoch": 1.89, "grad_norm": 20.6670855382289, "learning_rate": 6.350103433297087e-06, "loss": 0.5085, "step": 12107 }, { "epoch": 1.89, "grad_norm": 21.680164550481418, "learning_rate": 6.348533428239714e-06, "loss": 0.634, "step": 12108 }, { "epoch": 1.89, "grad_norm": 13.489945181727462, "learning_rate": 6.346963527027914e-06, "loss": 0.5525, "step": 12109 }, { "epoch": 1.89, "grad_norm": 16.19784392701362, "learning_rate": 6.345393729706333e-06, "loss": 0.546, "step": 12110 }, { "epoch": 1.89, "grad_norm": 19.954025546209017, "learning_rate": 6.343824036319622e-06, "loss": 0.5526, "step": 12111 }, { "epoch": 1.89, "grad_norm": 23.34881135030741, "learning_rate": 6.3422544469124116e-06, "loss": 0.5157, "step": 12112 }, { "epoch": 1.89, "grad_norm": 20.118293256647107, "learning_rate": 6.340684961529344e-06, "loss": 0.5003, "step": 12113 }, { "epoch": 1.89, "grad_norm": 29.154600797479603, "learning_rate": 6.3391155802150585e-06, "loss": 0.5921, "step": 12114 }, { "epoch": 1.89, "grad_norm": 21.12816342673484, "learning_rate": 6.337546303014182e-06, "loss": 0.58, "step": 12115 }, { "epoch": 1.89, "grad_norm": 20.339794443128238, "learning_rate": 6.3359771299713515e-06, "loss": 0.5267, "step": 12116 }, { "epoch": 1.89, "grad_norm": 18.634952107227992, "learning_rate": 6.334408061131182e-06, "loss": 0.6128, "step": 12117 }, { "epoch": 1.89, "grad_norm": 17.08689824993263, "learning_rate": 6.332839096538306e-06, "loss": 0.5128, "step": 12118 }, { "epoch": 1.89, "grad_norm": 32.108857005412496, "learning_rate": 6.331270236237337e-06, "loss": 0.6248, "step": 12119 }, { "epoch": 1.89, "grad_norm": 21.527985321504907, "learning_rate": 6.3297014802729e-06, "loss": 0.5667, "step": 12120 }, { "epoch": 1.89, "grad_norm": 21.067530380097878, "learning_rate": 6.328132828689607e-06, "loss": 0.5479, "step": 12121 }, { "epoch": 1.89, "grad_norm": 15.4930213472013, "learning_rate": 6.326564281532066e-06, "loss": 0.5874, "step": 12122 }, { "epoch": 1.89, "grad_norm": 24.555763306628627, "learning_rate": 6.324995838844887e-06, "loss": 0.5761, "step": 12123 }, { "epoch": 1.89, "grad_norm": 13.541109242948549, "learning_rate": 6.3234275006726785e-06, "loss": 0.5583, "step": 12124 }, { "epoch": 1.89, "grad_norm": 22.056090374825953, "learning_rate": 6.321859267060039e-06, "loss": 0.5633, "step": 12125 }, { "epoch": 1.89, "grad_norm": 22.46125347346427, "learning_rate": 6.320291138051574e-06, "loss": 0.5139, "step": 12126 }, { "epoch": 1.89, "grad_norm": 18.564255330388068, "learning_rate": 6.318723113691874e-06, "loss": 0.5711, "step": 12127 }, { "epoch": 1.89, "grad_norm": 16.729528991683672, "learning_rate": 6.317155194025536e-06, "loss": 0.514, "step": 12128 }, { "epoch": 1.89, "grad_norm": 15.953630778900546, "learning_rate": 6.315587379097148e-06, "loss": 0.4967, "step": 12129 }, { "epoch": 1.89, "grad_norm": 19.196151424732893, "learning_rate": 6.314019668951301e-06, "loss": 0.4802, "step": 12130 }, { "epoch": 1.89, "grad_norm": 28.745074756786046, "learning_rate": 6.312452063632581e-06, "loss": 0.6421, "step": 12131 }, { "epoch": 1.9, "grad_norm": 18.89685114620413, "learning_rate": 6.310884563185562e-06, "loss": 0.5391, "step": 12132 }, { "epoch": 1.9, "grad_norm": 19.51968135207296, "learning_rate": 6.309317167654832e-06, "loss": 0.5695, "step": 12133 }, { "epoch": 1.9, "grad_norm": 16.21754217152341, "learning_rate": 6.307749877084961e-06, "loss": 0.5368, "step": 12134 }, { "epoch": 1.9, "grad_norm": 15.802725015764079, "learning_rate": 6.306182691520522e-06, "loss": 0.5337, "step": 12135 }, { "epoch": 1.9, "grad_norm": 21.475203937732335, "learning_rate": 6.304615611006091e-06, "loss": 0.5122, "step": 12136 }, { "epoch": 1.9, "grad_norm": 18.803607309789403, "learning_rate": 6.303048635586226e-06, "loss": 0.5234, "step": 12137 }, { "epoch": 1.9, "grad_norm": 18.6900878393442, "learning_rate": 6.301481765305495e-06, "loss": 0.4845, "step": 12138 }, { "epoch": 1.9, "grad_norm": 14.536274392563245, "learning_rate": 6.2999150002084605e-06, "loss": 0.4481, "step": 12139 }, { "epoch": 1.9, "grad_norm": 13.34631101077166, "learning_rate": 6.298348340339681e-06, "loss": 0.4614, "step": 12140 }, { "epoch": 1.9, "grad_norm": 18.733572981706885, "learning_rate": 6.296781785743708e-06, "loss": 0.4746, "step": 12141 }, { "epoch": 1.9, "grad_norm": 20.696732479182458, "learning_rate": 6.295215336465094e-06, "loss": 0.5664, "step": 12142 }, { "epoch": 1.9, "grad_norm": 23.773605769800138, "learning_rate": 6.293648992548389e-06, "loss": 0.5913, "step": 12143 }, { "epoch": 1.9, "grad_norm": 27.63732583878765, "learning_rate": 6.292082754038135e-06, "loss": 0.5831, "step": 12144 }, { "epoch": 1.9, "grad_norm": 16.396089979180317, "learning_rate": 6.2905166209788845e-06, "loss": 0.552, "step": 12145 }, { "epoch": 1.9, "grad_norm": 17.33877369740794, "learning_rate": 6.288950593415168e-06, "loss": 0.4786, "step": 12146 }, { "epoch": 1.9, "grad_norm": 22.863567746112626, "learning_rate": 6.2873846713915276e-06, "loss": 0.5624, "step": 12147 }, { "epoch": 1.9, "grad_norm": 35.01740979961583, "learning_rate": 6.285818854952492e-06, "loss": 0.5152, "step": 12148 }, { "epoch": 1.9, "grad_norm": 26.63655857307103, "learning_rate": 6.284253144142598e-06, "loss": 0.5377, "step": 12149 }, { "epoch": 1.9, "grad_norm": 10.547387084538498, "learning_rate": 6.282687539006371e-06, "loss": 0.4411, "step": 12150 }, { "epoch": 1.9, "grad_norm": 22.093835304451684, "learning_rate": 6.2811220395883346e-06, "loss": 0.5887, "step": 12151 }, { "epoch": 1.9, "grad_norm": 12.83116213730814, "learning_rate": 6.279556645933011e-06, "loss": 0.4992, "step": 12152 }, { "epoch": 1.9, "grad_norm": 19.305860641229682, "learning_rate": 6.277991358084923e-06, "loss": 0.4904, "step": 12153 }, { "epoch": 1.9, "grad_norm": 23.656308158944118, "learning_rate": 6.276426176088579e-06, "loss": 0.5123, "step": 12154 }, { "epoch": 1.9, "grad_norm": 37.919970446881415, "learning_rate": 6.274861099988502e-06, "loss": 0.4972, "step": 12155 }, { "epoch": 1.9, "grad_norm": 22.10413727202575, "learning_rate": 6.273296129829194e-06, "loss": 0.4898, "step": 12156 }, { "epoch": 1.9, "grad_norm": 14.42871961564321, "learning_rate": 6.271731265655161e-06, "loss": 0.4595, "step": 12157 }, { "epoch": 1.9, "grad_norm": 19.745079032566384, "learning_rate": 6.270166507510913e-06, "loss": 0.5628, "step": 12158 }, { "epoch": 1.9, "grad_norm": 19.419873320898557, "learning_rate": 6.268601855440944e-06, "loss": 0.5304, "step": 12159 }, { "epoch": 1.9, "grad_norm": 18.7436566319673, "learning_rate": 6.267037309489763e-06, "loss": 0.5514, "step": 12160 }, { "epoch": 1.9, "grad_norm": 30.18045657385437, "learning_rate": 6.265472869701851e-06, "loss": 0.5706, "step": 12161 }, { "epoch": 1.9, "grad_norm": 20.326844633949047, "learning_rate": 6.2639085361217075e-06, "loss": 0.5038, "step": 12162 }, { "epoch": 1.9, "grad_norm": 13.806943304338365, "learning_rate": 6.262344308793818e-06, "loss": 0.5406, "step": 12163 }, { "epoch": 1.9, "grad_norm": 20.500973005289495, "learning_rate": 6.2607801877626715e-06, "loss": 0.5545, "step": 12164 }, { "epoch": 1.9, "grad_norm": 20.16940634386228, "learning_rate": 6.259216173072751e-06, "loss": 0.5182, "step": 12165 }, { "epoch": 1.9, "grad_norm": 18.769873816215153, "learning_rate": 6.257652264768532e-06, "loss": 0.5135, "step": 12166 }, { "epoch": 1.9, "grad_norm": 22.583136667244826, "learning_rate": 6.256088462894492e-06, "loss": 0.4973, "step": 12167 }, { "epoch": 1.9, "grad_norm": 24.71962303544123, "learning_rate": 6.254524767495108e-06, "loss": 0.5374, "step": 12168 }, { "epoch": 1.9, "grad_norm": 15.828915716531528, "learning_rate": 6.252961178614847e-06, "loss": 0.5273, "step": 12169 }, { "epoch": 1.9, "grad_norm": 17.488331191192067, "learning_rate": 6.251397696298181e-06, "loss": 0.5523, "step": 12170 }, { "epoch": 1.9, "grad_norm": 25.26685122142957, "learning_rate": 6.24983432058957e-06, "loss": 0.4952, "step": 12171 }, { "epoch": 1.9, "grad_norm": 22.96508958934432, "learning_rate": 6.2482710515334775e-06, "loss": 0.5503, "step": 12172 }, { "epoch": 1.9, "grad_norm": 25.732194557995523, "learning_rate": 6.246707889174359e-06, "loss": 0.6013, "step": 12173 }, { "epoch": 1.9, "grad_norm": 20.552280537380508, "learning_rate": 6.245144833556675e-06, "loss": 0.5129, "step": 12174 }, { "epoch": 1.9, "grad_norm": 23.539379121541863, "learning_rate": 6.243581884724879e-06, "loss": 0.5449, "step": 12175 }, { "epoch": 1.9, "grad_norm": 13.429760648118778, "learning_rate": 6.242019042723411e-06, "loss": 0.4941, "step": 12176 }, { "epoch": 1.9, "grad_norm": 19.374425665741384, "learning_rate": 6.240456307596725e-06, "loss": 0.5319, "step": 12177 }, { "epoch": 1.9, "grad_norm": 24.315272535988825, "learning_rate": 6.238893679389262e-06, "loss": 0.5312, "step": 12178 }, { "epoch": 1.9, "grad_norm": 19.871648856812556, "learning_rate": 6.237331158145461e-06, "loss": 0.5329, "step": 12179 }, { "epoch": 1.9, "grad_norm": 14.468476990643177, "learning_rate": 6.2357687439097645e-06, "loss": 0.5216, "step": 12180 }, { "epoch": 1.9, "grad_norm": 14.178170901760858, "learning_rate": 6.234206436726602e-06, "loss": 0.5029, "step": 12181 }, { "epoch": 1.9, "grad_norm": 13.410333693919632, "learning_rate": 6.232644236640403e-06, "loss": 0.468, "step": 12182 }, { "epoch": 1.9, "grad_norm": 24.020185020285673, "learning_rate": 6.231082143695599e-06, "loss": 0.4947, "step": 12183 }, { "epoch": 1.9, "grad_norm": 18.449760025430884, "learning_rate": 6.229520157936617e-06, "loss": 0.538, "step": 12184 }, { "epoch": 1.9, "grad_norm": 24.12944773331522, "learning_rate": 6.2279582794078755e-06, "loss": 0.4938, "step": 12185 }, { "epoch": 1.9, "grad_norm": 18.59313569376448, "learning_rate": 6.226396508153789e-06, "loss": 0.5377, "step": 12186 }, { "epoch": 1.9, "grad_norm": 18.33269462872682, "learning_rate": 6.2248348442187825e-06, "loss": 0.5134, "step": 12187 }, { "epoch": 1.9, "grad_norm": 16.566752857629098, "learning_rate": 6.223273287647262e-06, "loss": 0.5049, "step": 12188 }, { "epoch": 1.9, "grad_norm": 18.16600884295014, "learning_rate": 6.221711838483644e-06, "loss": 0.569, "step": 12189 }, { "epoch": 1.9, "grad_norm": 12.89655233224021, "learning_rate": 6.220150496772328e-06, "loss": 0.4793, "step": 12190 }, { "epoch": 1.9, "grad_norm": 19.262054701911623, "learning_rate": 6.2185892625577215e-06, "loss": 0.5134, "step": 12191 }, { "epoch": 1.9, "grad_norm": 18.098562206397553, "learning_rate": 6.217028135884222e-06, "loss": 0.5987, "step": 12192 }, { "epoch": 1.9, "grad_norm": 17.694832545628266, "learning_rate": 6.215467116796232e-06, "loss": 0.5321, "step": 12193 }, { "epoch": 1.9, "grad_norm": 26.71760695616933, "learning_rate": 6.213906205338145e-06, "loss": 0.6175, "step": 12194 }, { "epoch": 1.9, "grad_norm": 16.29930276882442, "learning_rate": 6.212345401554347e-06, "loss": 0.493, "step": 12195 }, { "epoch": 1.91, "grad_norm": 14.941679654082067, "learning_rate": 6.2107847054892325e-06, "loss": 0.5197, "step": 12196 }, { "epoch": 1.91, "grad_norm": 29.9872709792075, "learning_rate": 6.209224117187184e-06, "loss": 0.5919, "step": 12197 }, { "epoch": 1.91, "grad_norm": 27.558368416452314, "learning_rate": 6.207663636692581e-06, "loss": 0.5752, "step": 12198 }, { "epoch": 1.91, "grad_norm": 13.595299440629072, "learning_rate": 6.2061032640498095e-06, "loss": 0.4722, "step": 12199 }, { "epoch": 1.91, "grad_norm": 22.196052407349097, "learning_rate": 6.204542999303242e-06, "loss": 0.5753, "step": 12200 }, { "epoch": 1.91, "grad_norm": 13.92598435466475, "learning_rate": 6.202982842497247e-06, "loss": 0.5, "step": 12201 }, { "epoch": 1.91, "grad_norm": 22.196931852423365, "learning_rate": 6.201422793676203e-06, "loss": 0.5354, "step": 12202 }, { "epoch": 1.91, "grad_norm": 25.322023037926495, "learning_rate": 6.1998628528844685e-06, "loss": 0.5301, "step": 12203 }, { "epoch": 1.91, "grad_norm": 14.072040389281618, "learning_rate": 6.198303020166418e-06, "loss": 0.4978, "step": 12204 }, { "epoch": 1.91, "grad_norm": 19.72833382236004, "learning_rate": 6.196743295566401e-06, "loss": 0.5288, "step": 12205 }, { "epoch": 1.91, "grad_norm": 17.988414108663815, "learning_rate": 6.195183679128781e-06, "loss": 0.5215, "step": 12206 }, { "epoch": 1.91, "grad_norm": 18.723997816160026, "learning_rate": 6.19362417089791e-06, "loss": 0.461, "step": 12207 }, { "epoch": 1.91, "grad_norm": 17.68155221802161, "learning_rate": 6.1920647709181425e-06, "loss": 0.5168, "step": 12208 }, { "epoch": 1.91, "grad_norm": 15.470433069431035, "learning_rate": 6.190505479233826e-06, "loss": 0.5784, "step": 12209 }, { "epoch": 1.91, "grad_norm": 31.764688419846152, "learning_rate": 6.188946295889304e-06, "loss": 0.5499, "step": 12210 }, { "epoch": 1.91, "grad_norm": 52.303380862688996, "learning_rate": 6.18738722092892e-06, "loss": 0.5165, "step": 12211 }, { "epoch": 1.91, "grad_norm": 20.527049362147796, "learning_rate": 6.185828254397012e-06, "loss": 0.5718, "step": 12212 }, { "epoch": 1.91, "grad_norm": 24.46383839094235, "learning_rate": 6.184269396337916e-06, "loss": 0.544, "step": 12213 }, { "epoch": 1.91, "grad_norm": 15.638061376217884, "learning_rate": 6.182710646795972e-06, "loss": 0.5732, "step": 12214 }, { "epoch": 1.91, "grad_norm": 28.396531618539328, "learning_rate": 6.181152005815501e-06, "loss": 0.6066, "step": 12215 }, { "epoch": 1.91, "grad_norm": 25.555496253753066, "learning_rate": 6.179593473440833e-06, "loss": 0.5093, "step": 12216 }, { "epoch": 1.91, "grad_norm": 19.767739276362914, "learning_rate": 6.178035049716289e-06, "loss": 0.5647, "step": 12217 }, { "epoch": 1.91, "grad_norm": 22.583684578647365, "learning_rate": 6.1764767346861955e-06, "loss": 0.5463, "step": 12218 }, { "epoch": 1.91, "grad_norm": 26.37654932045606, "learning_rate": 6.174918528394868e-06, "loss": 0.5759, "step": 12219 }, { "epoch": 1.91, "grad_norm": 11.013167751793873, "learning_rate": 6.173360430886617e-06, "loss": 0.4951, "step": 12220 }, { "epoch": 1.91, "grad_norm": 25.632850781812976, "learning_rate": 6.171802442205759e-06, "loss": 0.5838, "step": 12221 }, { "epoch": 1.91, "grad_norm": 29.75469137214884, "learning_rate": 6.170244562396599e-06, "loss": 0.6622, "step": 12222 }, { "epoch": 1.91, "grad_norm": 20.208161660740046, "learning_rate": 6.1686867915034425e-06, "loss": 0.5114, "step": 12223 }, { "epoch": 1.91, "grad_norm": 16.09749480076051, "learning_rate": 6.167129129570595e-06, "loss": 0.5446, "step": 12224 }, { "epoch": 1.91, "grad_norm": 21.50417245428113, "learning_rate": 6.165571576642351e-06, "loss": 0.5308, "step": 12225 }, { "epoch": 1.91, "grad_norm": 22.708743762687742, "learning_rate": 6.164014132763008e-06, "loss": 0.5796, "step": 12226 }, { "epoch": 1.91, "grad_norm": 15.435235427625951, "learning_rate": 6.162456797976859e-06, "loss": 0.4247, "step": 12227 }, { "epoch": 1.91, "grad_norm": 15.944371516691, "learning_rate": 6.1608995723281925e-06, "loss": 0.584, "step": 12228 }, { "epoch": 1.91, "grad_norm": 14.31472861152017, "learning_rate": 6.159342455861304e-06, "loss": 0.4976, "step": 12229 }, { "epoch": 1.91, "grad_norm": 20.092443165619823, "learning_rate": 6.157785448620461e-06, "loss": 0.5718, "step": 12230 }, { "epoch": 1.91, "grad_norm": 15.706088023786904, "learning_rate": 6.156228550649954e-06, "loss": 0.4975, "step": 12231 }, { "epoch": 1.91, "grad_norm": 25.66864942322433, "learning_rate": 6.154671761994058e-06, "loss": 0.6087, "step": 12232 }, { "epoch": 1.91, "grad_norm": 15.474472487283657, "learning_rate": 6.153115082697052e-06, "loss": 0.4825, "step": 12233 }, { "epoch": 1.91, "grad_norm": 19.818215031021587, "learning_rate": 6.1515585128032e-06, "loss": 0.5691, "step": 12234 }, { "epoch": 1.91, "grad_norm": 15.501704291506618, "learning_rate": 6.150002052356772e-06, "loss": 0.5322, "step": 12235 }, { "epoch": 1.91, "grad_norm": 12.194695851182095, "learning_rate": 6.148445701402032e-06, "loss": 0.5778, "step": 12236 }, { "epoch": 1.91, "grad_norm": 16.76790458525802, "learning_rate": 6.1468894599832455e-06, "loss": 0.5472, "step": 12237 }, { "epoch": 1.91, "grad_norm": 19.855332722142574, "learning_rate": 6.145333328144669e-06, "loss": 0.4362, "step": 12238 }, { "epoch": 1.91, "grad_norm": 18.30343730953803, "learning_rate": 6.143777305930556e-06, "loss": 0.5867, "step": 12239 }, { "epoch": 1.91, "grad_norm": 14.508747785039262, "learning_rate": 6.142221393385161e-06, "loss": 0.4782, "step": 12240 }, { "epoch": 1.91, "grad_norm": 13.681509216897357, "learning_rate": 6.140665590552735e-06, "loss": 0.5278, "step": 12241 }, { "epoch": 1.91, "grad_norm": 21.12231851381048, "learning_rate": 6.139109897477517e-06, "loss": 0.499, "step": 12242 }, { "epoch": 1.91, "grad_norm": 16.80724646913212, "learning_rate": 6.137554314203762e-06, "loss": 0.5027, "step": 12243 }, { "epoch": 1.91, "grad_norm": 16.74158231929338, "learning_rate": 6.135998840775699e-06, "loss": 0.4842, "step": 12244 }, { "epoch": 1.91, "grad_norm": 19.675780274552274, "learning_rate": 6.134443477237566e-06, "loss": 0.5263, "step": 12245 }, { "epoch": 1.91, "grad_norm": 14.52745672419481, "learning_rate": 6.132888223633602e-06, "loss": 0.5024, "step": 12246 }, { "epoch": 1.91, "grad_norm": 14.004724199356977, "learning_rate": 6.131333080008033e-06, "loss": 0.4343, "step": 12247 }, { "epoch": 1.91, "grad_norm": 20.733886475834016, "learning_rate": 6.129778046405095e-06, "loss": 0.5556, "step": 12248 }, { "epoch": 1.91, "grad_norm": 13.8758789978313, "learning_rate": 6.128223122868997e-06, "loss": 0.5325, "step": 12249 }, { "epoch": 1.91, "grad_norm": 14.666615292904387, "learning_rate": 6.126668309443973e-06, "loss": 0.4913, "step": 12250 }, { "epoch": 1.91, "grad_norm": 17.96252052545536, "learning_rate": 6.125113606174233e-06, "loss": 0.4449, "step": 12251 }, { "epoch": 1.91, "grad_norm": 17.053416924322395, "learning_rate": 6.123559013103997e-06, "loss": 0.5862, "step": 12252 }, { "epoch": 1.91, "grad_norm": 19.105968538219564, "learning_rate": 6.122004530277476e-06, "loss": 0.5546, "step": 12253 }, { "epoch": 1.91, "grad_norm": 23.602386607329148, "learning_rate": 6.1204501577388775e-06, "loss": 0.5671, "step": 12254 }, { "epoch": 1.91, "grad_norm": 16.400506416576217, "learning_rate": 6.118895895532405e-06, "loss": 0.4884, "step": 12255 }, { "epoch": 1.91, "grad_norm": 18.01123151614482, "learning_rate": 6.117341743702265e-06, "loss": 0.5161, "step": 12256 }, { "epoch": 1.91, "grad_norm": 20.794278583194735, "learning_rate": 6.115787702292651e-06, "loss": 0.5328, "step": 12257 }, { "epoch": 1.91, "grad_norm": 19.752328734634908, "learning_rate": 6.1142337713477685e-06, "loss": 0.4854, "step": 12258 }, { "epoch": 1.91, "grad_norm": 22.945047820567815, "learning_rate": 6.112679950911802e-06, "loss": 0.5748, "step": 12259 }, { "epoch": 1.92, "grad_norm": 10.582051918908004, "learning_rate": 6.111126241028942e-06, "loss": 0.4871, "step": 12260 }, { "epoch": 1.92, "grad_norm": 15.339243105559248, "learning_rate": 6.1095726417433765e-06, "loss": 0.5092, "step": 12261 }, { "epoch": 1.92, "grad_norm": 19.14320682434728, "learning_rate": 6.10801915309929e-06, "loss": 0.546, "step": 12262 }, { "epoch": 1.92, "grad_norm": 20.53945640434366, "learning_rate": 6.106465775140865e-06, "loss": 0.5227, "step": 12263 }, { "epoch": 1.92, "grad_norm": 14.576915372890655, "learning_rate": 6.104912507912272e-06, "loss": 0.4823, "step": 12264 }, { "epoch": 1.92, "grad_norm": 23.49053580678368, "learning_rate": 6.1033593514576895e-06, "loss": 0.5771, "step": 12265 }, { "epoch": 1.92, "grad_norm": 17.141406715618732, "learning_rate": 6.101806305821288e-06, "loss": 0.4874, "step": 12266 }, { "epoch": 1.92, "grad_norm": 23.96907863864128, "learning_rate": 6.100253371047233e-06, "loss": 0.5957, "step": 12267 }, { "epoch": 1.92, "grad_norm": 22.537269674362996, "learning_rate": 6.098700547179695e-06, "loss": 0.559, "step": 12268 }, { "epoch": 1.92, "grad_norm": 15.360425778300087, "learning_rate": 6.0971478342628285e-06, "loss": 0.5299, "step": 12269 }, { "epoch": 1.92, "grad_norm": 18.44405969940147, "learning_rate": 6.095595232340794e-06, "loss": 0.488, "step": 12270 }, { "epoch": 1.92, "grad_norm": 23.080720858805734, "learning_rate": 6.094042741457749e-06, "loss": 0.5644, "step": 12271 }, { "epoch": 1.92, "grad_norm": 23.63498438983016, "learning_rate": 6.0924903616578394e-06, "loss": 0.5554, "step": 12272 }, { "epoch": 1.92, "grad_norm": 19.989096452521217, "learning_rate": 6.090938092985226e-06, "loss": 0.546, "step": 12273 }, { "epoch": 1.92, "grad_norm": 14.255078179482236, "learning_rate": 6.08938593548404e-06, "loss": 0.4866, "step": 12274 }, { "epoch": 1.92, "grad_norm": 14.732335913330358, "learning_rate": 6.087833889198431e-06, "loss": 0.4742, "step": 12275 }, { "epoch": 1.92, "grad_norm": 22.263861475589888, "learning_rate": 6.086281954172537e-06, "loss": 0.5251, "step": 12276 }, { "epoch": 1.92, "grad_norm": 23.329257043351287, "learning_rate": 6.084730130450497e-06, "loss": 0.5346, "step": 12277 }, { "epoch": 1.92, "grad_norm": 22.871124746430304, "learning_rate": 6.083178418076443e-06, "loss": 0.5068, "step": 12278 }, { "epoch": 1.92, "grad_norm": 17.967703829454823, "learning_rate": 6.081626817094502e-06, "loss": 0.4768, "step": 12279 }, { "epoch": 1.92, "grad_norm": 22.833495254505166, "learning_rate": 6.0800753275488e-06, "loss": 0.5139, "step": 12280 }, { "epoch": 1.92, "grad_norm": 17.877622778332483, "learning_rate": 6.078523949483465e-06, "loss": 0.473, "step": 12281 }, { "epoch": 1.92, "grad_norm": 16.82979669122015, "learning_rate": 6.076972682942617e-06, "loss": 0.5988, "step": 12282 }, { "epoch": 1.92, "grad_norm": 35.828291760774825, "learning_rate": 6.075421527970366e-06, "loss": 0.5938, "step": 12283 }, { "epoch": 1.92, "grad_norm": 17.579769136467974, "learning_rate": 6.073870484610833e-06, "loss": 0.4486, "step": 12284 }, { "epoch": 1.92, "grad_norm": 18.436900194562686, "learning_rate": 6.072319552908128e-06, "loss": 0.5465, "step": 12285 }, { "epoch": 1.92, "grad_norm": 22.484018060529767, "learning_rate": 6.070768732906353e-06, "loss": 0.5934, "step": 12286 }, { "epoch": 1.92, "grad_norm": 22.157476642748698, "learning_rate": 6.069218024649622e-06, "loss": 0.526, "step": 12287 }, { "epoch": 1.92, "grad_norm": 33.04431796631102, "learning_rate": 6.0676674281820294e-06, "loss": 0.5257, "step": 12288 }, { "epoch": 1.92, "grad_norm": 33.817975569553894, "learning_rate": 6.066116943547673e-06, "loss": 0.5153, "step": 12289 }, { "epoch": 1.92, "grad_norm": 26.919243660167318, "learning_rate": 6.064566570790651e-06, "loss": 0.5411, "step": 12290 }, { "epoch": 1.92, "grad_norm": 22.007482900307902, "learning_rate": 6.063016309955051e-06, "loss": 0.5515, "step": 12291 }, { "epoch": 1.92, "grad_norm": 16.910521128157438, "learning_rate": 6.06146616108497e-06, "loss": 0.5549, "step": 12292 }, { "epoch": 1.92, "grad_norm": 24.922693304899678, "learning_rate": 6.059916124224483e-06, "loss": 0.5857, "step": 12293 }, { "epoch": 1.92, "grad_norm": 18.935425710856062, "learning_rate": 6.058366199417676e-06, "loss": 0.5129, "step": 12294 }, { "epoch": 1.92, "grad_norm": 15.906686051767723, "learning_rate": 6.056816386708628e-06, "loss": 0.5667, "step": 12295 }, { "epoch": 1.92, "grad_norm": 24.12557874118802, "learning_rate": 6.055266686141418e-06, "loss": 0.5148, "step": 12296 }, { "epoch": 1.92, "grad_norm": 16.554082193330338, "learning_rate": 6.053717097760117e-06, "loss": 0.6317, "step": 12297 }, { "epoch": 1.92, "grad_norm": 11.323846431926478, "learning_rate": 6.052167621608791e-06, "loss": 0.4545, "step": 12298 }, { "epoch": 1.92, "grad_norm": 19.044315923526742, "learning_rate": 6.050618257731507e-06, "loss": 0.588, "step": 12299 }, { "epoch": 1.92, "grad_norm": 21.87047303360174, "learning_rate": 6.049069006172331e-06, "loss": 0.6194, "step": 12300 }, { "epoch": 1.92, "grad_norm": 18.025605170167378, "learning_rate": 6.04751986697532e-06, "loss": 0.5724, "step": 12301 }, { "epoch": 1.92, "grad_norm": 20.72592136212466, "learning_rate": 6.045970840184534e-06, "loss": 0.561, "step": 12302 }, { "epoch": 1.92, "grad_norm": 15.776708591536986, "learning_rate": 6.044421925844024e-06, "loss": 0.577, "step": 12303 }, { "epoch": 1.92, "grad_norm": 17.23633041943759, "learning_rate": 6.0428731239978415e-06, "loss": 0.5055, "step": 12304 }, { "epoch": 1.92, "grad_norm": 15.852625788170199, "learning_rate": 6.041324434690028e-06, "loss": 0.5349, "step": 12305 }, { "epoch": 1.92, "grad_norm": 24.325067232427184, "learning_rate": 6.0397758579646335e-06, "loss": 0.5686, "step": 12306 }, { "epoch": 1.92, "grad_norm": 19.65158011481245, "learning_rate": 6.038227393865701e-06, "loss": 0.5007, "step": 12307 }, { "epoch": 1.92, "grad_norm": 16.732066002306635, "learning_rate": 6.036679042437259e-06, "loss": 0.5151, "step": 12308 }, { "epoch": 1.92, "grad_norm": 17.113880423875756, "learning_rate": 6.035130803723349e-06, "loss": 0.5191, "step": 12309 }, { "epoch": 1.92, "grad_norm": 18.109060157198147, "learning_rate": 6.033582677767997e-06, "loss": 0.4982, "step": 12310 }, { "epoch": 1.92, "grad_norm": 17.981712496775902, "learning_rate": 6.032034664615236e-06, "loss": 0.5105, "step": 12311 }, { "epoch": 1.92, "grad_norm": 17.46247899208498, "learning_rate": 6.030486764309089e-06, "loss": 0.4542, "step": 12312 }, { "epoch": 1.92, "grad_norm": 15.68847238236059, "learning_rate": 6.028938976893575e-06, "loss": 0.5514, "step": 12313 }, { "epoch": 1.92, "grad_norm": 19.300428745900966, "learning_rate": 6.027391302412711e-06, "loss": 0.4951, "step": 12314 }, { "epoch": 1.92, "grad_norm": 27.54649278545439, "learning_rate": 6.025843740910518e-06, "loss": 0.5783, "step": 12315 }, { "epoch": 1.92, "grad_norm": 14.536354303130112, "learning_rate": 6.024296292431e-06, "loss": 0.4886, "step": 12316 }, { "epoch": 1.92, "grad_norm": 20.83737265854583, "learning_rate": 6.022748957018176e-06, "loss": 0.5059, "step": 12317 }, { "epoch": 1.92, "grad_norm": 16.784168414085872, "learning_rate": 6.02120173471604e-06, "loss": 0.5209, "step": 12318 }, { "epoch": 1.92, "grad_norm": 17.342161685798253, "learning_rate": 6.0196546255686016e-06, "loss": 0.5904, "step": 12319 }, { "epoch": 1.92, "grad_norm": 16.253202693117608, "learning_rate": 6.018107629619854e-06, "loss": 0.5323, "step": 12320 }, { "epoch": 1.92, "grad_norm": 19.855083079835506, "learning_rate": 6.0165607469137984e-06, "loss": 0.5526, "step": 12321 }, { "epoch": 1.92, "grad_norm": 24.258805807173978, "learning_rate": 6.015013977494429e-06, "loss": 0.4956, "step": 12322 }, { "epoch": 1.92, "grad_norm": 17.880375996946327, "learning_rate": 6.013467321405726e-06, "loss": 0.5218, "step": 12323 }, { "epoch": 1.93, "grad_norm": 16.002634858758192, "learning_rate": 6.011920778691679e-06, "loss": 0.5165, "step": 12324 }, { "epoch": 1.93, "grad_norm": 23.227114741114562, "learning_rate": 6.010374349396274e-06, "loss": 0.558, "step": 12325 }, { "epoch": 1.93, "grad_norm": 25.40746386292164, "learning_rate": 6.008828033563487e-06, "loss": 0.5004, "step": 12326 }, { "epoch": 1.93, "grad_norm": 21.305761380857877, "learning_rate": 6.0072818312373e-06, "loss": 0.4808, "step": 12327 }, { "epoch": 1.93, "grad_norm": 13.371554597161667, "learning_rate": 6.005735742461681e-06, "loss": 0.4352, "step": 12328 }, { "epoch": 1.93, "grad_norm": 19.46274298321452, "learning_rate": 6.004189767280599e-06, "loss": 0.4739, "step": 12329 }, { "epoch": 1.93, "grad_norm": 20.270289648941908, "learning_rate": 6.002643905738022e-06, "loss": 0.5381, "step": 12330 }, { "epoch": 1.93, "grad_norm": 12.718616033827995, "learning_rate": 6.0010981578779175e-06, "loss": 0.4453, "step": 12331 }, { "epoch": 1.93, "grad_norm": 17.35109225747623, "learning_rate": 5.9995525237442395e-06, "loss": 0.4429, "step": 12332 }, { "epoch": 1.93, "grad_norm": 21.473640577903073, "learning_rate": 5.9980070033809455e-06, "loss": 0.5189, "step": 12333 }, { "epoch": 1.93, "grad_norm": 25.72697366906619, "learning_rate": 5.996461596831995e-06, "loss": 0.5608, "step": 12334 }, { "epoch": 1.93, "grad_norm": 14.514171329426617, "learning_rate": 5.9949163041413315e-06, "loss": 0.4483, "step": 12335 }, { "epoch": 1.93, "grad_norm": 17.781371233548622, "learning_rate": 5.9933711253529115e-06, "loss": 0.5658, "step": 12336 }, { "epoch": 1.93, "grad_norm": 20.80831062263204, "learning_rate": 5.991826060510666e-06, "loss": 0.5643, "step": 12337 }, { "epoch": 1.93, "grad_norm": 25.85296637077086, "learning_rate": 5.990281109658545e-06, "loss": 0.5359, "step": 12338 }, { "epoch": 1.93, "grad_norm": 28.357611902460572, "learning_rate": 5.988736272840483e-06, "loss": 0.5674, "step": 12339 }, { "epoch": 1.93, "grad_norm": 20.35844605610881, "learning_rate": 5.987191550100415e-06, "loss": 0.5319, "step": 12340 }, { "epoch": 1.93, "grad_norm": 22.82902687218169, "learning_rate": 5.985646941482275e-06, "loss": 0.5442, "step": 12341 }, { "epoch": 1.93, "grad_norm": 18.13460488522913, "learning_rate": 5.984102447029986e-06, "loss": 0.5105, "step": 12342 }, { "epoch": 1.93, "grad_norm": 21.501221263792665, "learning_rate": 5.9825580667874715e-06, "loss": 0.4686, "step": 12343 }, { "epoch": 1.93, "grad_norm": 19.02378032069939, "learning_rate": 5.981013800798657e-06, "loss": 0.4675, "step": 12344 }, { "epoch": 1.93, "grad_norm": 14.986058449378904, "learning_rate": 5.9794696491074576e-06, "loss": 0.4835, "step": 12345 }, { "epoch": 1.93, "grad_norm": 13.695985333615948, "learning_rate": 5.977925611757794e-06, "loss": 0.4871, "step": 12346 }, { "epoch": 1.93, "grad_norm": 25.67034463970409, "learning_rate": 5.976381688793571e-06, "loss": 0.5343, "step": 12347 }, { "epoch": 1.93, "grad_norm": 19.512696596220714, "learning_rate": 5.974837880258697e-06, "loss": 0.487, "step": 12348 }, { "epoch": 1.93, "grad_norm": 21.34622884831244, "learning_rate": 5.973294186197079e-06, "loss": 0.6084, "step": 12349 }, { "epoch": 1.93, "grad_norm": 18.68531712609804, "learning_rate": 5.971750606652621e-06, "loss": 0.5363, "step": 12350 }, { "epoch": 1.93, "grad_norm": 20.314529720279726, "learning_rate": 5.97020714166922e-06, "loss": 0.5161, "step": 12351 }, { "epoch": 1.93, "grad_norm": 11.415219791041318, "learning_rate": 5.9686637912907675e-06, "loss": 0.5051, "step": 12352 }, { "epoch": 1.93, "grad_norm": 16.814206456863005, "learning_rate": 5.96712055556116e-06, "loss": 0.545, "step": 12353 }, { "epoch": 1.93, "grad_norm": 17.706709550619333, "learning_rate": 5.965577434524283e-06, "loss": 0.4624, "step": 12354 }, { "epoch": 1.93, "grad_norm": 14.581329845397073, "learning_rate": 5.964034428224026e-06, "loss": 0.5049, "step": 12355 }, { "epoch": 1.93, "grad_norm": 13.117178325638507, "learning_rate": 5.9624915367042705e-06, "loss": 0.492, "step": 12356 }, { "epoch": 1.93, "grad_norm": 13.125009604859244, "learning_rate": 5.960948760008891e-06, "loss": 0.5081, "step": 12357 }, { "epoch": 1.93, "grad_norm": 23.20827378614795, "learning_rate": 5.959406098181766e-06, "loss": 0.5871, "step": 12358 }, { "epoch": 1.93, "grad_norm": 17.45444301351702, "learning_rate": 5.9578635512667686e-06, "loss": 0.5759, "step": 12359 }, { "epoch": 1.93, "grad_norm": 27.47755231964424, "learning_rate": 5.9563211193077644e-06, "loss": 0.5985, "step": 12360 }, { "epoch": 1.93, "grad_norm": 18.32948695242102, "learning_rate": 5.9547788023486296e-06, "loss": 0.5493, "step": 12361 }, { "epoch": 1.93, "grad_norm": 28.60447621319841, "learning_rate": 5.953236600433212e-06, "loss": 0.6283, "step": 12362 }, { "epoch": 1.93, "grad_norm": 22.70753012069293, "learning_rate": 5.951694513605381e-06, "loss": 0.541, "step": 12363 }, { "epoch": 1.93, "grad_norm": 16.780237856413557, "learning_rate": 5.950152541908988e-06, "loss": 0.4784, "step": 12364 }, { "epoch": 1.93, "grad_norm": 18.073217379480376, "learning_rate": 5.94861068538789e-06, "loss": 0.4856, "step": 12365 }, { "epoch": 1.93, "grad_norm": 24.70224930033288, "learning_rate": 5.947068944085934e-06, "loss": 0.5123, "step": 12366 }, { "epoch": 1.93, "grad_norm": 15.355420320444683, "learning_rate": 5.9455273180469666e-06, "loss": 0.5719, "step": 12367 }, { "epoch": 1.93, "grad_norm": 18.02773195553403, "learning_rate": 5.943985807314827e-06, "loss": 0.4973, "step": 12368 }, { "epoch": 1.93, "grad_norm": 15.900555744632838, "learning_rate": 5.942444411933361e-06, "loss": 0.5651, "step": 12369 }, { "epoch": 1.93, "grad_norm": 25.365440457697158, "learning_rate": 5.940903131946401e-06, "loss": 0.5803, "step": 12370 }, { "epoch": 1.93, "grad_norm": 19.917058685633137, "learning_rate": 5.939361967397785e-06, "loss": 0.6036, "step": 12371 }, { "epoch": 1.93, "grad_norm": 12.248547168624674, "learning_rate": 5.937820918331336e-06, "loss": 0.458, "step": 12372 }, { "epoch": 1.93, "grad_norm": 18.06914817004524, "learning_rate": 5.9362799847908845e-06, "loss": 0.4907, "step": 12373 }, { "epoch": 1.93, "grad_norm": 23.20746151830743, "learning_rate": 5.934739166820252e-06, "loss": 0.6095, "step": 12374 }, { "epoch": 1.93, "grad_norm": 17.4524110427762, "learning_rate": 5.9331984644632615e-06, "loss": 0.5806, "step": 12375 }, { "epoch": 1.93, "grad_norm": 26.480536295771884, "learning_rate": 5.931657877763728e-06, "loss": 0.4927, "step": 12376 }, { "epoch": 1.93, "grad_norm": 22.82148895182661, "learning_rate": 5.930117406765462e-06, "loss": 0.6186, "step": 12377 }, { "epoch": 1.93, "grad_norm": 25.16527062831763, "learning_rate": 5.928577051512277e-06, "loss": 0.5995, "step": 12378 }, { "epoch": 1.93, "grad_norm": 15.246434517017754, "learning_rate": 5.927036812047978e-06, "loss": 0.5412, "step": 12379 }, { "epoch": 1.93, "grad_norm": 22.70932936674966, "learning_rate": 5.925496688416374e-06, "loss": 0.5898, "step": 12380 }, { "epoch": 1.93, "grad_norm": 16.30599333081173, "learning_rate": 5.923956680661257e-06, "loss": 0.5145, "step": 12381 }, { "epoch": 1.93, "grad_norm": 26.839283217938156, "learning_rate": 5.922416788826429e-06, "loss": 0.5827, "step": 12382 }, { "epoch": 1.93, "grad_norm": 21.621006501287976, "learning_rate": 5.920877012955679e-06, "loss": 0.5624, "step": 12383 }, { "epoch": 1.93, "grad_norm": 19.77983446013133, "learning_rate": 5.919337353092802e-06, "loss": 0.5661, "step": 12384 }, { "epoch": 1.93, "grad_norm": 13.844974066521289, "learning_rate": 5.917797809281587e-06, "loss": 0.4736, "step": 12385 }, { "epoch": 1.93, "grad_norm": 14.97412940604854, "learning_rate": 5.916258381565811e-06, "loss": 0.4732, "step": 12386 }, { "epoch": 1.93, "grad_norm": 21.183570081700054, "learning_rate": 5.914719069989257e-06, "loss": 0.4821, "step": 12387 }, { "epoch": 1.94, "grad_norm": 17.848445057527147, "learning_rate": 5.913179874595705e-06, "loss": 0.542, "step": 12388 }, { "epoch": 1.94, "grad_norm": 15.3524597269773, "learning_rate": 5.911640795428923e-06, "loss": 0.5594, "step": 12389 }, { "epoch": 1.94, "grad_norm": 25.311570329430268, "learning_rate": 5.910101832532691e-06, "loss": 0.5641, "step": 12390 }, { "epoch": 1.94, "grad_norm": 21.574542657705813, "learning_rate": 5.908562985950768e-06, "loss": 0.5433, "step": 12391 }, { "epoch": 1.94, "grad_norm": 19.314135868731526, "learning_rate": 5.907024255726919e-06, "loss": 0.5528, "step": 12392 }, { "epoch": 1.94, "grad_norm": 21.87727639547193, "learning_rate": 5.905485641904904e-06, "loss": 0.531, "step": 12393 }, { "epoch": 1.94, "grad_norm": 18.0783898903737, "learning_rate": 5.903947144528486e-06, "loss": 0.5324, "step": 12394 }, { "epoch": 1.94, "grad_norm": 24.25238748815649, "learning_rate": 5.902408763641416e-06, "loss": 0.5721, "step": 12395 }, { "epoch": 1.94, "grad_norm": 19.923012906656197, "learning_rate": 5.9008704992874384e-06, "loss": 0.5137, "step": 12396 }, { "epoch": 1.94, "grad_norm": 16.70871145537197, "learning_rate": 5.89933235151031e-06, "loss": 0.4857, "step": 12397 }, { "epoch": 1.94, "grad_norm": 22.25025273029509, "learning_rate": 5.897794320353768e-06, "loss": 0.5418, "step": 12398 }, { "epoch": 1.94, "grad_norm": 15.853369733894144, "learning_rate": 5.896256405861558e-06, "loss": 0.5445, "step": 12399 }, { "epoch": 1.94, "grad_norm": 16.20968496516432, "learning_rate": 5.8947186080774165e-06, "loss": 0.4262, "step": 12400 }, { "epoch": 1.94, "grad_norm": 22.339365901195418, "learning_rate": 5.8931809270450746e-06, "loss": 0.4989, "step": 12401 }, { "epoch": 1.94, "grad_norm": 22.328759201097878, "learning_rate": 5.8916433628082626e-06, "loss": 0.5199, "step": 12402 }, { "epoch": 1.94, "grad_norm": 21.190822115815195, "learning_rate": 5.890105915410712e-06, "loss": 0.5436, "step": 12403 }, { "epoch": 1.94, "grad_norm": 15.322672321278281, "learning_rate": 5.8885685848961424e-06, "loss": 0.4906, "step": 12404 }, { "epoch": 1.94, "grad_norm": 14.316512197750402, "learning_rate": 5.8870313713082845e-06, "loss": 0.4905, "step": 12405 }, { "epoch": 1.94, "grad_norm": 24.476398468076734, "learning_rate": 5.885494274690842e-06, "loss": 0.5817, "step": 12406 }, { "epoch": 1.94, "grad_norm": 18.981877596757233, "learning_rate": 5.883957295087537e-06, "loss": 0.5369, "step": 12407 }, { "epoch": 1.94, "grad_norm": 19.790324660696253, "learning_rate": 5.882420432542077e-06, "loss": 0.6368, "step": 12408 }, { "epoch": 1.94, "grad_norm": 20.776741641622746, "learning_rate": 5.8808836870981725e-06, "loss": 0.5382, "step": 12409 }, { "epoch": 1.94, "grad_norm": 28.8304841045478, "learning_rate": 5.87934705879953e-06, "loss": 0.4809, "step": 12410 }, { "epoch": 1.94, "grad_norm": 18.49730136397342, "learning_rate": 5.8778105476898415e-06, "loss": 0.569, "step": 12411 }, { "epoch": 1.94, "grad_norm": 24.831637009879625, "learning_rate": 5.87627415381281e-06, "loss": 0.5326, "step": 12412 }, { "epoch": 1.94, "grad_norm": 21.52488268034948, "learning_rate": 5.874737877212131e-06, "loss": 0.5532, "step": 12413 }, { "epoch": 1.94, "grad_norm": 17.600109363752566, "learning_rate": 5.873201717931492e-06, "loss": 0.4851, "step": 12414 }, { "epoch": 1.94, "grad_norm": 17.071249126516147, "learning_rate": 5.871665676014584e-06, "loss": 0.4928, "step": 12415 }, { "epoch": 1.94, "grad_norm": 12.959499104560754, "learning_rate": 5.870129751505089e-06, "loss": 0.5253, "step": 12416 }, { "epoch": 1.94, "grad_norm": 19.189228694467246, "learning_rate": 5.868593944446688e-06, "loss": 0.5779, "step": 12417 }, { "epoch": 1.94, "grad_norm": 16.355434399466912, "learning_rate": 5.867058254883056e-06, "loss": 0.5513, "step": 12418 }, { "epoch": 1.94, "grad_norm": 18.660743889390798, "learning_rate": 5.8655226828578715e-06, "loss": 0.4516, "step": 12419 }, { "epoch": 1.94, "grad_norm": 16.286007793706315, "learning_rate": 5.863987228414805e-06, "loss": 0.5183, "step": 12420 }, { "epoch": 1.94, "grad_norm": 18.253188452800565, "learning_rate": 5.8624518915975185e-06, "loss": 0.5251, "step": 12421 }, { "epoch": 1.94, "grad_norm": 23.41976905119453, "learning_rate": 5.860916672449682e-06, "loss": 0.5562, "step": 12422 }, { "epoch": 1.94, "grad_norm": 16.102702458140445, "learning_rate": 5.859381571014951e-06, "loss": 0.5111, "step": 12423 }, { "epoch": 1.94, "grad_norm": 26.352578266560638, "learning_rate": 5.857846587336989e-06, "loss": 0.5896, "step": 12424 }, { "epoch": 1.94, "grad_norm": 13.961129258199225, "learning_rate": 5.856311721459449e-06, "loss": 0.5243, "step": 12425 }, { "epoch": 1.94, "grad_norm": 23.310855484682733, "learning_rate": 5.854776973425978e-06, "loss": 0.6174, "step": 12426 }, { "epoch": 1.94, "grad_norm": 13.128513415428728, "learning_rate": 5.853242343280222e-06, "loss": 0.4883, "step": 12427 }, { "epoch": 1.94, "grad_norm": 20.5058426664811, "learning_rate": 5.85170783106583e-06, "loss": 0.5232, "step": 12428 }, { "epoch": 1.94, "grad_norm": 15.839485978626312, "learning_rate": 5.8501734368264425e-06, "loss": 0.5185, "step": 12429 }, { "epoch": 1.94, "grad_norm": 16.23613590313458, "learning_rate": 5.848639160605694e-06, "loss": 0.5017, "step": 12430 }, { "epoch": 1.94, "grad_norm": 17.551606047043208, "learning_rate": 5.847105002447218e-06, "loss": 0.5182, "step": 12431 }, { "epoch": 1.94, "grad_norm": 23.34640627229531, "learning_rate": 5.845570962394647e-06, "loss": 0.569, "step": 12432 }, { "epoch": 1.94, "grad_norm": 17.691517026642444, "learning_rate": 5.8440370404916035e-06, "loss": 0.4934, "step": 12433 }, { "epoch": 1.94, "grad_norm": 18.552155417085057, "learning_rate": 5.842503236781722e-06, "loss": 0.4917, "step": 12434 }, { "epoch": 1.94, "grad_norm": 20.389523060572717, "learning_rate": 5.840969551308614e-06, "loss": 0.4808, "step": 12435 }, { "epoch": 1.94, "grad_norm": 15.692684038938735, "learning_rate": 5.839435984115899e-06, "loss": 0.5329, "step": 12436 }, { "epoch": 1.94, "grad_norm": 24.00601334414222, "learning_rate": 5.8379025352471905e-06, "loss": 0.5689, "step": 12437 }, { "epoch": 1.94, "grad_norm": 19.812143364337903, "learning_rate": 5.836369204746097e-06, "loss": 0.5513, "step": 12438 }, { "epoch": 1.94, "grad_norm": 14.104422500629422, "learning_rate": 5.834835992656232e-06, "loss": 0.5193, "step": 12439 }, { "epoch": 1.94, "grad_norm": 17.40942907408857, "learning_rate": 5.833302899021191e-06, "loss": 0.5167, "step": 12440 }, { "epoch": 1.94, "grad_norm": 18.757174954111548, "learning_rate": 5.831769923884579e-06, "loss": 0.5556, "step": 12441 }, { "epoch": 1.94, "grad_norm": 22.52624071654354, "learning_rate": 5.830237067289993e-06, "loss": 0.51, "step": 12442 }, { "epoch": 1.94, "grad_norm": 16.920302141397592, "learning_rate": 5.828704329281024e-06, "loss": 0.5263, "step": 12443 }, { "epoch": 1.94, "grad_norm": 18.55781347183722, "learning_rate": 5.827171709901267e-06, "loss": 0.5711, "step": 12444 }, { "epoch": 1.94, "grad_norm": 28.532334896719814, "learning_rate": 5.825639209194302e-06, "loss": 0.5597, "step": 12445 }, { "epoch": 1.94, "grad_norm": 20.352330935389503, "learning_rate": 5.824106827203719e-06, "loss": 0.5696, "step": 12446 }, { "epoch": 1.94, "grad_norm": 23.64854656392188, "learning_rate": 5.822574563973091e-06, "loss": 0.5965, "step": 12447 }, { "epoch": 1.94, "grad_norm": 13.898015455986016, "learning_rate": 5.8210424195460005e-06, "loss": 0.5788, "step": 12448 }, { "epoch": 1.94, "grad_norm": 21.882164503947042, "learning_rate": 5.8195103939660214e-06, "loss": 0.4394, "step": 12449 }, { "epoch": 1.94, "grad_norm": 21.478105566625683, "learning_rate": 5.817978487276722e-06, "loss": 0.4859, "step": 12450 }, { "epoch": 1.94, "grad_norm": 15.224604271771039, "learning_rate": 5.816446699521663e-06, "loss": 0.5257, "step": 12451 }, { "epoch": 1.95, "grad_norm": 24.33418179052358, "learning_rate": 5.814915030744414e-06, "loss": 0.4796, "step": 12452 }, { "epoch": 1.95, "grad_norm": 21.10577788189234, "learning_rate": 5.813383480988533e-06, "loss": 0.4948, "step": 12453 }, { "epoch": 1.95, "grad_norm": 23.879545363926923, "learning_rate": 5.811852050297579e-06, "loss": 0.5374, "step": 12454 }, { "epoch": 1.95, "grad_norm": 19.97696263006178, "learning_rate": 5.810320738715104e-06, "loss": 0.506, "step": 12455 }, { "epoch": 1.95, "grad_norm": 21.110892698703324, "learning_rate": 5.808789546284652e-06, "loss": 0.476, "step": 12456 }, { "epoch": 1.95, "grad_norm": 24.43157343451516, "learning_rate": 5.807258473049774e-06, "loss": 0.5584, "step": 12457 }, { "epoch": 1.95, "grad_norm": 26.064357158531386, "learning_rate": 5.805727519054017e-06, "loss": 0.5798, "step": 12458 }, { "epoch": 1.95, "grad_norm": 15.64100087892689, "learning_rate": 5.8041966843409135e-06, "loss": 0.4604, "step": 12459 }, { "epoch": 1.95, "grad_norm": 16.739041076827338, "learning_rate": 5.802665968954e-06, "loss": 0.5202, "step": 12460 }, { "epoch": 1.95, "grad_norm": 24.643162402588274, "learning_rate": 5.801135372936809e-06, "loss": 0.5942, "step": 12461 }, { "epoch": 1.95, "grad_norm": 21.75121216164051, "learning_rate": 5.7996048963328775e-06, "loss": 0.4859, "step": 12462 }, { "epoch": 1.95, "grad_norm": 17.670004865447886, "learning_rate": 5.798074539185721e-06, "loss": 0.5276, "step": 12463 }, { "epoch": 1.95, "grad_norm": 20.13919389795665, "learning_rate": 5.79654430153887e-06, "loss": 0.493, "step": 12464 }, { "epoch": 1.95, "grad_norm": 16.92708306068029, "learning_rate": 5.7950141834358365e-06, "loss": 0.5175, "step": 12465 }, { "epoch": 1.95, "grad_norm": 18.451526262504714, "learning_rate": 5.793484184920139e-06, "loss": 0.5432, "step": 12466 }, { "epoch": 1.95, "grad_norm": 18.563684716292574, "learning_rate": 5.7919543060352965e-06, "loss": 0.5377, "step": 12467 }, { "epoch": 1.95, "grad_norm": 24.197960593496667, "learning_rate": 5.790424546824806e-06, "loss": 0.476, "step": 12468 }, { "epoch": 1.95, "grad_norm": 14.942537784489552, "learning_rate": 5.788894907332184e-06, "loss": 0.4994, "step": 12469 }, { "epoch": 1.95, "grad_norm": 16.053730136514798, "learning_rate": 5.787365387600923e-06, "loss": 0.4765, "step": 12470 }, { "epoch": 1.95, "grad_norm": 27.214363119087164, "learning_rate": 5.78583598767453e-06, "loss": 0.5595, "step": 12471 }, { "epoch": 1.95, "grad_norm": 15.915624174855418, "learning_rate": 5.784306707596492e-06, "loss": 0.5166, "step": 12472 }, { "epoch": 1.95, "grad_norm": 16.060402106052805, "learning_rate": 5.782777547410305e-06, "loss": 0.5342, "step": 12473 }, { "epoch": 1.95, "grad_norm": 18.914268547678464, "learning_rate": 5.781248507159463e-06, "loss": 0.5307, "step": 12474 }, { "epoch": 1.95, "grad_norm": 31.899785463421146, "learning_rate": 5.7797195868874445e-06, "loss": 0.5482, "step": 12475 }, { "epoch": 1.95, "grad_norm": 15.001721067835165, "learning_rate": 5.778190786637729e-06, "loss": 0.5411, "step": 12476 }, { "epoch": 1.95, "grad_norm": 12.444153370730817, "learning_rate": 5.776662106453797e-06, "loss": 0.4678, "step": 12477 }, { "epoch": 1.95, "grad_norm": 20.353632849090427, "learning_rate": 5.775133546379128e-06, "loss": 0.5077, "step": 12478 }, { "epoch": 1.95, "grad_norm": 10.834760320730712, "learning_rate": 5.773605106457185e-06, "loss": 0.4954, "step": 12479 }, { "epoch": 1.95, "grad_norm": 16.308350709486163, "learning_rate": 5.7720767867314464e-06, "loss": 0.5926, "step": 12480 }, { "epoch": 1.95, "grad_norm": 18.609722802013273, "learning_rate": 5.7705485872453645e-06, "loss": 0.5273, "step": 12481 }, { "epoch": 1.95, "grad_norm": 23.27119436783684, "learning_rate": 5.769020508042408e-06, "loss": 0.5826, "step": 12482 }, { "epoch": 1.95, "grad_norm": 18.239027591810153, "learning_rate": 5.7674925491660365e-06, "loss": 0.5343, "step": 12483 }, { "epoch": 1.95, "grad_norm": 22.908663200060055, "learning_rate": 5.765964710659702e-06, "loss": 0.5001, "step": 12484 }, { "epoch": 1.95, "grad_norm": 19.89939788976453, "learning_rate": 5.764436992566849e-06, "loss": 0.5365, "step": 12485 }, { "epoch": 1.95, "grad_norm": 17.136950492813643, "learning_rate": 5.762909394930931e-06, "loss": 0.4934, "step": 12486 }, { "epoch": 1.95, "grad_norm": 21.235445422265318, "learning_rate": 5.761381917795394e-06, "loss": 0.6292, "step": 12487 }, { "epoch": 1.95, "grad_norm": 23.285681019017932, "learning_rate": 5.759854561203677e-06, "loss": 0.5552, "step": 12488 }, { "epoch": 1.95, "grad_norm": 16.197960103509278, "learning_rate": 5.7583273251992115e-06, "loss": 0.4705, "step": 12489 }, { "epoch": 1.95, "grad_norm": 16.631962528254324, "learning_rate": 5.756800209825435e-06, "loss": 0.5682, "step": 12490 }, { "epoch": 1.95, "grad_norm": 17.57362087103549, "learning_rate": 5.75527321512578e-06, "loss": 0.5323, "step": 12491 }, { "epoch": 1.95, "grad_norm": 23.030932422969112, "learning_rate": 5.753746341143674e-06, "loss": 0.5403, "step": 12492 }, { "epoch": 1.95, "grad_norm": 16.337260179503378, "learning_rate": 5.752219587922538e-06, "loss": 0.5667, "step": 12493 }, { "epoch": 1.95, "grad_norm": 15.85073213202555, "learning_rate": 5.75069295550579e-06, "loss": 0.5049, "step": 12494 }, { "epoch": 1.95, "grad_norm": 15.73827537788363, "learning_rate": 5.749166443936847e-06, "loss": 0.5169, "step": 12495 }, { "epoch": 1.95, "grad_norm": 28.790778922850944, "learning_rate": 5.747640053259127e-06, "loss": 0.5765, "step": 12496 }, { "epoch": 1.95, "grad_norm": 14.127927161419743, "learning_rate": 5.746113783516034e-06, "loss": 0.5171, "step": 12497 }, { "epoch": 1.95, "grad_norm": 16.800661820548402, "learning_rate": 5.74458763475098e-06, "loss": 0.4633, "step": 12498 }, { "epoch": 1.95, "grad_norm": 23.60879236975523, "learning_rate": 5.743061607007359e-06, "loss": 0.4808, "step": 12499 }, { "epoch": 1.95, "grad_norm": 20.959288420981103, "learning_rate": 5.741535700328581e-06, "loss": 0.5148, "step": 12500 }, { "epoch": 1.95, "grad_norm": 15.711860192404695, "learning_rate": 5.740009914758032e-06, "loss": 0.5283, "step": 12501 }, { "epoch": 1.95, "grad_norm": 26.531583622159197, "learning_rate": 5.738484250339109e-06, "loss": 0.5762, "step": 12502 }, { "epoch": 1.95, "grad_norm": 19.132486755725896, "learning_rate": 5.7369587071152055e-06, "loss": 0.4862, "step": 12503 }, { "epoch": 1.95, "grad_norm": 23.722684751858132, "learning_rate": 5.735433285129699e-06, "loss": 0.5213, "step": 12504 }, { "epoch": 1.95, "grad_norm": 28.15577624268529, "learning_rate": 5.733907984425979e-06, "loss": 0.5755, "step": 12505 }, { "epoch": 1.95, "grad_norm": 5.741721753286174, "learning_rate": 5.732382805047416e-06, "loss": 0.5924, "step": 12506 }, { "epoch": 1.95, "grad_norm": 33.274203655077116, "learning_rate": 5.730857747037389e-06, "loss": 0.5293, "step": 12507 }, { "epoch": 1.95, "grad_norm": 20.296892509717193, "learning_rate": 5.729332810439274e-06, "loss": 0.4954, "step": 12508 }, { "epoch": 1.95, "grad_norm": 18.053697510222843, "learning_rate": 5.727807995296437e-06, "loss": 0.5509, "step": 12509 }, { "epoch": 1.95, "grad_norm": 17.581892688055213, "learning_rate": 5.7262833016522366e-06, "loss": 0.4933, "step": 12510 }, { "epoch": 1.95, "grad_norm": 18.662274235466818, "learning_rate": 5.72475872955004e-06, "loss": 0.4975, "step": 12511 }, { "epoch": 1.95, "grad_norm": 18.684752196093253, "learning_rate": 5.723234279033207e-06, "loss": 0.5613, "step": 12512 }, { "epoch": 1.95, "grad_norm": 19.316687034363607, "learning_rate": 5.721709950145089e-06, "loss": 0.4734, "step": 12513 }, { "epoch": 1.95, "grad_norm": 19.93026752592397, "learning_rate": 5.720185742929034e-06, "loss": 0.4768, "step": 12514 }, { "epoch": 1.95, "grad_norm": 17.1710369393287, "learning_rate": 5.718661657428393e-06, "loss": 0.568, "step": 12515 }, { "epoch": 1.96, "grad_norm": 23.876264666595794, "learning_rate": 5.717137693686509e-06, "loss": 0.557, "step": 12516 }, { "epoch": 1.96, "grad_norm": 39.55901711471774, "learning_rate": 5.7156138517467285e-06, "loss": 0.5792, "step": 12517 }, { "epoch": 1.96, "grad_norm": 22.886133900409952, "learning_rate": 5.714090131652385e-06, "loss": 0.4748, "step": 12518 }, { "epoch": 1.96, "grad_norm": 21.304057770134662, "learning_rate": 5.712566533446805e-06, "loss": 0.5342, "step": 12519 }, { "epoch": 1.96, "grad_norm": 29.1376990048434, "learning_rate": 5.711043057173326e-06, "loss": 0.5208, "step": 12520 }, { "epoch": 1.96, "grad_norm": 13.883859511732858, "learning_rate": 5.709519702875277e-06, "loss": 0.5307, "step": 12521 }, { "epoch": 1.96, "grad_norm": 19.384569057745946, "learning_rate": 5.707996470595977e-06, "loss": 0.5088, "step": 12522 }, { "epoch": 1.96, "grad_norm": 23.603573246586848, "learning_rate": 5.706473360378743e-06, "loss": 0.5319, "step": 12523 }, { "epoch": 1.96, "grad_norm": 18.797908347876046, "learning_rate": 5.704950372266895e-06, "loss": 0.5121, "step": 12524 }, { "epoch": 1.96, "grad_norm": 14.752522660479476, "learning_rate": 5.70342750630375e-06, "loss": 0.5141, "step": 12525 }, { "epoch": 1.96, "grad_norm": 18.492231915303353, "learning_rate": 5.70190476253261e-06, "loss": 0.5276, "step": 12526 }, { "epoch": 1.96, "grad_norm": 20.262984210376192, "learning_rate": 5.700382140996787e-06, "loss": 0.5458, "step": 12527 }, { "epoch": 1.96, "grad_norm": 14.646237699000984, "learning_rate": 5.698859641739578e-06, "loss": 0.5303, "step": 12528 }, { "epoch": 1.96, "grad_norm": 18.33721121600938, "learning_rate": 5.697337264804283e-06, "loss": 0.5405, "step": 12529 }, { "epoch": 1.96, "grad_norm": 20.62713172184386, "learning_rate": 5.695815010234204e-06, "loss": 0.5091, "step": 12530 }, { "epoch": 1.96, "grad_norm": 15.779170723887841, "learning_rate": 5.694292878072625e-06, "loss": 0.5264, "step": 12531 }, { "epoch": 1.96, "grad_norm": 18.651083869925394, "learning_rate": 5.6927708683628415e-06, "loss": 0.4928, "step": 12532 }, { "epoch": 1.96, "grad_norm": 22.997205040947005, "learning_rate": 5.6912489811481295e-06, "loss": 0.5407, "step": 12533 }, { "epoch": 1.96, "grad_norm": 16.428851128556964, "learning_rate": 5.689727216471782e-06, "loss": 0.5003, "step": 12534 }, { "epoch": 1.96, "grad_norm": 27.057714057259094, "learning_rate": 5.688205574377066e-06, "loss": 0.5358, "step": 12535 }, { "epoch": 1.96, "grad_norm": 13.40300789085051, "learning_rate": 5.686684054907261e-06, "loss": 0.5165, "step": 12536 }, { "epoch": 1.96, "grad_norm": 20.309578242343726, "learning_rate": 5.685162658105643e-06, "loss": 0.5323, "step": 12537 }, { "epoch": 1.96, "grad_norm": 13.518517897377475, "learning_rate": 5.683641384015475e-06, "loss": 0.5376, "step": 12538 }, { "epoch": 1.96, "grad_norm": 13.366438832991074, "learning_rate": 5.682120232680015e-06, "loss": 0.4875, "step": 12539 }, { "epoch": 1.96, "grad_norm": 16.21755931963509, "learning_rate": 5.6805992041425315e-06, "loss": 0.4978, "step": 12540 }, { "epoch": 1.96, "grad_norm": 13.553028789026607, "learning_rate": 5.679078298446279e-06, "loss": 0.4798, "step": 12541 }, { "epoch": 1.96, "grad_norm": 18.120778420877425, "learning_rate": 5.677557515634517e-06, "loss": 0.5202, "step": 12542 }, { "epoch": 1.96, "grad_norm": 13.682213563651375, "learning_rate": 5.67603685575049e-06, "loss": 0.498, "step": 12543 }, { "epoch": 1.96, "grad_norm": 22.76760529484355, "learning_rate": 5.674516318837442e-06, "loss": 0.5423, "step": 12544 }, { "epoch": 1.96, "grad_norm": 26.01898118499872, "learning_rate": 5.6729959049386185e-06, "loss": 0.4723, "step": 12545 }, { "epoch": 1.96, "grad_norm": 17.906333548665486, "learning_rate": 5.6714756140972645e-06, "loss": 0.5309, "step": 12546 }, { "epoch": 1.96, "grad_norm": 23.24678918499856, "learning_rate": 5.669955446356612e-06, "loss": 0.5461, "step": 12547 }, { "epoch": 1.96, "grad_norm": 21.26010737006464, "learning_rate": 5.668435401759891e-06, "loss": 0.5405, "step": 12548 }, { "epoch": 1.96, "grad_norm": 14.728930170174033, "learning_rate": 5.666915480350332e-06, "loss": 0.4694, "step": 12549 }, { "epoch": 1.96, "grad_norm": 18.655714290122393, "learning_rate": 5.665395682171166e-06, "loss": 0.56, "step": 12550 }, { "epoch": 1.96, "grad_norm": 28.772033980332132, "learning_rate": 5.6638760072656075e-06, "loss": 0.6019, "step": 12551 }, { "epoch": 1.96, "grad_norm": 21.37753722750831, "learning_rate": 5.662356455676882e-06, "loss": 0.5345, "step": 12552 }, { "epoch": 1.96, "grad_norm": 16.169553315475262, "learning_rate": 5.660837027448198e-06, "loss": 0.5226, "step": 12553 }, { "epoch": 1.96, "grad_norm": 20.76049640319283, "learning_rate": 5.65931772262277e-06, "loss": 0.5174, "step": 12554 }, { "epoch": 1.96, "grad_norm": 13.494866472412605, "learning_rate": 5.657798541243812e-06, "loss": 0.5609, "step": 12555 }, { "epoch": 1.96, "grad_norm": 17.063626158757213, "learning_rate": 5.656279483354519e-06, "loss": 0.5339, "step": 12556 }, { "epoch": 1.96, "grad_norm": 20.754580984163205, "learning_rate": 5.6547605489981e-06, "loss": 0.5504, "step": 12557 }, { "epoch": 1.96, "grad_norm": 27.982667719669724, "learning_rate": 5.653241738217745e-06, "loss": 0.5279, "step": 12558 }, { "epoch": 1.96, "grad_norm": 20.728815621649602, "learning_rate": 5.651723051056657e-06, "loss": 0.4881, "step": 12559 }, { "epoch": 1.96, "grad_norm": 23.595384337175233, "learning_rate": 5.650204487558016e-06, "loss": 0.5373, "step": 12560 }, { "epoch": 1.96, "grad_norm": 30.876869151838356, "learning_rate": 5.648686047765017e-06, "loss": 0.5955, "step": 12561 }, { "epoch": 1.96, "grad_norm": 13.985678155480695, "learning_rate": 5.647167731720844e-06, "loss": 0.5058, "step": 12562 }, { "epoch": 1.96, "grad_norm": 25.238870857978693, "learning_rate": 5.645649539468675e-06, "loss": 0.5549, "step": 12563 }, { "epoch": 1.96, "grad_norm": 19.525420028330117, "learning_rate": 5.644131471051681e-06, "loss": 0.4633, "step": 12564 }, { "epoch": 1.96, "grad_norm": 19.56415096756817, "learning_rate": 5.642613526513041e-06, "loss": 0.4476, "step": 12565 }, { "epoch": 1.96, "grad_norm": 13.182992717464638, "learning_rate": 5.641095705895923e-06, "loss": 0.4575, "step": 12566 }, { "epoch": 1.96, "grad_norm": 20.488173297783145, "learning_rate": 5.639578009243496e-06, "loss": 0.5252, "step": 12567 }, { "epoch": 1.96, "grad_norm": 24.166150841195538, "learning_rate": 5.638060436598922e-06, "loss": 0.5236, "step": 12568 }, { "epoch": 1.96, "grad_norm": 20.61527252302159, "learning_rate": 5.636542988005351e-06, "loss": 0.5729, "step": 12569 }, { "epoch": 1.96, "grad_norm": 22.638191243712967, "learning_rate": 5.635025663505946e-06, "loss": 0.5291, "step": 12570 }, { "epoch": 1.96, "grad_norm": 22.562467559199376, "learning_rate": 5.633508463143862e-06, "loss": 0.6122, "step": 12571 }, { "epoch": 1.96, "grad_norm": 18.02950975459572, "learning_rate": 5.631991386962243e-06, "loss": 0.5858, "step": 12572 }, { "epoch": 1.96, "grad_norm": 22.289733018226038, "learning_rate": 5.6304744350042295e-06, "loss": 0.6044, "step": 12573 }, { "epoch": 1.96, "grad_norm": 17.089825940804054, "learning_rate": 5.628957607312967e-06, "loss": 0.501, "step": 12574 }, { "epoch": 1.96, "grad_norm": 25.87329719303149, "learning_rate": 5.627440903931598e-06, "loss": 0.5379, "step": 12575 }, { "epoch": 1.96, "grad_norm": 17.764890367891415, "learning_rate": 5.62592432490325e-06, "loss": 0.5123, "step": 12576 }, { "epoch": 1.96, "grad_norm": 21.741319689084367, "learning_rate": 5.624407870271052e-06, "loss": 0.496, "step": 12577 }, { "epoch": 1.96, "grad_norm": 25.34575921924254, "learning_rate": 5.622891540078135e-06, "loss": 0.5614, "step": 12578 }, { "epoch": 1.96, "grad_norm": 21.20204664593074, "learning_rate": 5.621375334367622e-06, "loss": 0.5523, "step": 12579 }, { "epoch": 1.97, "grad_norm": 23.332139501443475, "learning_rate": 5.619859253182638e-06, "loss": 0.4856, "step": 12580 }, { "epoch": 1.97, "grad_norm": 20.622692533943, "learning_rate": 5.618343296566293e-06, "loss": 0.4649, "step": 12581 }, { "epoch": 1.97, "grad_norm": 14.584562452381421, "learning_rate": 5.616827464561698e-06, "loss": 0.5671, "step": 12582 }, { "epoch": 1.97, "grad_norm": 16.894356240485116, "learning_rate": 5.615311757211965e-06, "loss": 0.468, "step": 12583 }, { "epoch": 1.97, "grad_norm": 14.573504952153666, "learning_rate": 5.613796174560207e-06, "loss": 0.4361, "step": 12584 }, { "epoch": 1.97, "grad_norm": 17.565973201874296, "learning_rate": 5.612280716649514e-06, "loss": 0.5162, "step": 12585 }, { "epoch": 1.97, "grad_norm": 20.376518821859634, "learning_rate": 5.6107653835229954e-06, "loss": 0.554, "step": 12586 }, { "epoch": 1.97, "grad_norm": 16.229109762060997, "learning_rate": 5.609250175223737e-06, "loss": 0.4645, "step": 12587 }, { "epoch": 1.97, "grad_norm": 17.404187656848006, "learning_rate": 5.607735091794839e-06, "loss": 0.5476, "step": 12588 }, { "epoch": 1.97, "grad_norm": 20.088113569729536, "learning_rate": 5.606220133279383e-06, "loss": 0.512, "step": 12589 }, { "epoch": 1.97, "grad_norm": 13.489813118036757, "learning_rate": 5.604705299720455e-06, "loss": 0.4846, "step": 12590 }, { "epoch": 1.97, "grad_norm": 19.402525817803287, "learning_rate": 5.603190591161141e-06, "loss": 0.5612, "step": 12591 }, { "epoch": 1.97, "grad_norm": 17.287011542456632, "learning_rate": 5.601676007644511e-06, "loss": 0.5346, "step": 12592 }, { "epoch": 1.97, "grad_norm": 25.26303079305595, "learning_rate": 5.600161549213647e-06, "loss": 0.6273, "step": 12593 }, { "epoch": 1.97, "grad_norm": 20.737577954910897, "learning_rate": 5.598647215911609e-06, "loss": 0.4365, "step": 12594 }, { "epoch": 1.97, "grad_norm": 18.259214475553833, "learning_rate": 5.59713300778147e-06, "loss": 0.536, "step": 12595 }, { "epoch": 1.97, "grad_norm": 19.789865134882238, "learning_rate": 5.595618924866298e-06, "loss": 0.5557, "step": 12596 }, { "epoch": 1.97, "grad_norm": 30.925207838876528, "learning_rate": 5.594104967209146e-06, "loss": 0.5139, "step": 12597 }, { "epoch": 1.97, "grad_norm": 20.160048251129773, "learning_rate": 5.592591134853067e-06, "loss": 0.4601, "step": 12598 }, { "epoch": 1.97, "grad_norm": 33.102190886856214, "learning_rate": 5.591077427841118e-06, "loss": 0.5281, "step": 12599 }, { "epoch": 1.97, "grad_norm": 32.24142014145156, "learning_rate": 5.5895638462163536e-06, "loss": 0.5306, "step": 12600 }, { "epoch": 1.97, "grad_norm": 15.513560001143286, "learning_rate": 5.588050390021812e-06, "loss": 0.5155, "step": 12601 }, { "epoch": 1.97, "grad_norm": 13.357463083037707, "learning_rate": 5.586537059300532e-06, "loss": 0.5161, "step": 12602 }, { "epoch": 1.97, "grad_norm": 17.789975246355397, "learning_rate": 5.585023854095557e-06, "loss": 0.5413, "step": 12603 }, { "epoch": 1.97, "grad_norm": 14.613335914478983, "learning_rate": 5.5835107744499205e-06, "loss": 0.4683, "step": 12604 }, { "epoch": 1.97, "grad_norm": 24.80942499891135, "learning_rate": 5.581997820406659e-06, "loss": 0.5057, "step": 12605 }, { "epoch": 1.97, "grad_norm": 18.141944585903207, "learning_rate": 5.580484992008795e-06, "loss": 0.5648, "step": 12606 }, { "epoch": 1.97, "grad_norm": 25.615456157019434, "learning_rate": 5.5789722892993466e-06, "loss": 0.6131, "step": 12607 }, { "epoch": 1.97, "grad_norm": 36.313936375821235, "learning_rate": 5.577459712321341e-06, "loss": 0.5668, "step": 12608 }, { "epoch": 1.97, "grad_norm": 20.30870921782114, "learning_rate": 5.575947261117798e-06, "loss": 0.5608, "step": 12609 }, { "epoch": 1.97, "grad_norm": 22.2800572030574, "learning_rate": 5.574434935731723e-06, "loss": 0.4784, "step": 12610 }, { "epoch": 1.97, "grad_norm": 26.826145475115595, "learning_rate": 5.572922736206135e-06, "loss": 0.502, "step": 12611 }, { "epoch": 1.97, "grad_norm": 13.839644765217468, "learning_rate": 5.571410662584029e-06, "loss": 0.4732, "step": 12612 }, { "epoch": 1.97, "grad_norm": 12.342510382058945, "learning_rate": 5.5698987149084174e-06, "loss": 0.5158, "step": 12613 }, { "epoch": 1.97, "grad_norm": 17.544607838055274, "learning_rate": 5.568386893222291e-06, "loss": 0.5141, "step": 12614 }, { "epoch": 1.97, "grad_norm": 17.28824598855327, "learning_rate": 5.5668751975686485e-06, "loss": 0.6036, "step": 12615 }, { "epoch": 1.97, "grad_norm": 35.074450238780216, "learning_rate": 5.565363627990485e-06, "loss": 0.5349, "step": 12616 }, { "epoch": 1.97, "grad_norm": 23.567612451319665, "learning_rate": 5.563852184530784e-06, "loss": 0.4689, "step": 12617 }, { "epoch": 1.97, "grad_norm": 17.957472684793785, "learning_rate": 5.562340867232535e-06, "loss": 0.6178, "step": 12618 }, { "epoch": 1.97, "grad_norm": 19.258346950241314, "learning_rate": 5.560829676138712e-06, "loss": 0.5912, "step": 12619 }, { "epoch": 1.97, "grad_norm": 19.857396440666978, "learning_rate": 5.559318611292299e-06, "loss": 0.5387, "step": 12620 }, { "epoch": 1.97, "grad_norm": 22.970008307004164, "learning_rate": 5.557807672736264e-06, "loss": 0.5527, "step": 12621 }, { "epoch": 1.97, "grad_norm": 25.550777935103092, "learning_rate": 5.556296860513584e-06, "loss": 0.4918, "step": 12622 }, { "epoch": 1.97, "grad_norm": 19.66604013467039, "learning_rate": 5.554786174667217e-06, "loss": 0.5096, "step": 12623 }, { "epoch": 1.97, "grad_norm": 21.533930410269456, "learning_rate": 5.553275615240132e-06, "loss": 0.5125, "step": 12624 }, { "epoch": 1.97, "grad_norm": 14.539640807673022, "learning_rate": 5.551765182275292e-06, "loss": 0.4658, "step": 12625 }, { "epoch": 1.97, "grad_norm": 15.966272654662532, "learning_rate": 5.550254875815646e-06, "loss": 0.519, "step": 12626 }, { "epoch": 1.97, "grad_norm": 19.805065626939996, "learning_rate": 5.548744695904145e-06, "loss": 0.4964, "step": 12627 }, { "epoch": 1.97, "grad_norm": 30.900625258009978, "learning_rate": 5.5472346425837405e-06, "loss": 0.56, "step": 12628 }, { "epoch": 1.97, "grad_norm": 16.063630668201945, "learning_rate": 5.545724715897381e-06, "loss": 0.5138, "step": 12629 }, { "epoch": 1.97, "grad_norm": 21.596001596001994, "learning_rate": 5.544214915888006e-06, "loss": 0.5456, "step": 12630 }, { "epoch": 1.97, "grad_norm": 15.340409046731004, "learning_rate": 5.542705242598552e-06, "loss": 0.5157, "step": 12631 }, { "epoch": 1.97, "grad_norm": 18.661060137275072, "learning_rate": 5.54119569607195e-06, "loss": 0.5138, "step": 12632 }, { "epoch": 1.97, "grad_norm": 22.36441971592129, "learning_rate": 5.539686276351135e-06, "loss": 0.495, "step": 12633 }, { "epoch": 1.97, "grad_norm": 17.35537810670071, "learning_rate": 5.538176983479036e-06, "loss": 0.5449, "step": 12634 }, { "epoch": 1.97, "grad_norm": 15.332534711553333, "learning_rate": 5.536667817498573e-06, "loss": 0.4595, "step": 12635 }, { "epoch": 1.97, "grad_norm": 25.484735979836614, "learning_rate": 5.535158778452664e-06, "loss": 0.5956, "step": 12636 }, { "epoch": 1.97, "grad_norm": 19.2746785231696, "learning_rate": 5.533649866384226e-06, "loss": 0.5036, "step": 12637 }, { "epoch": 1.97, "grad_norm": 21.945022102042472, "learning_rate": 5.532141081336177e-06, "loss": 0.5224, "step": 12638 }, { "epoch": 1.97, "grad_norm": 13.935658346049768, "learning_rate": 5.530632423351421e-06, "loss": 0.4443, "step": 12639 }, { "epoch": 1.97, "grad_norm": 18.51050762449517, "learning_rate": 5.5291238924728654e-06, "loss": 0.4302, "step": 12640 }, { "epoch": 1.97, "grad_norm": 14.027996218024988, "learning_rate": 5.5276154887434075e-06, "loss": 0.5304, "step": 12641 }, { "epoch": 1.97, "grad_norm": 17.159128770952982, "learning_rate": 5.526107212205949e-06, "loss": 0.5162, "step": 12642 }, { "epoch": 1.97, "grad_norm": 19.50456143483398, "learning_rate": 5.524599062903388e-06, "loss": 0.512, "step": 12643 }, { "epoch": 1.98, "grad_norm": 16.944525703500673, "learning_rate": 5.523091040878608e-06, "loss": 0.5381, "step": 12644 }, { "epoch": 1.98, "grad_norm": 14.98428168009199, "learning_rate": 5.521583146174503e-06, "loss": 0.4688, "step": 12645 }, { "epoch": 1.98, "grad_norm": 15.4442419198208, "learning_rate": 5.5200753788339515e-06, "loss": 0.5068, "step": 12646 }, { "epoch": 1.98, "grad_norm": 15.575833924074296, "learning_rate": 5.518567738899838e-06, "loss": 0.5214, "step": 12647 }, { "epoch": 1.98, "grad_norm": 17.677035146254948, "learning_rate": 5.517060226415032e-06, "loss": 0.4799, "step": 12648 }, { "epoch": 1.98, "grad_norm": 16.97043014080274, "learning_rate": 5.515552841422412e-06, "loss": 0.509, "step": 12649 }, { "epoch": 1.98, "grad_norm": 16.025432938561142, "learning_rate": 5.514045583964848e-06, "loss": 0.4991, "step": 12650 }, { "epoch": 1.98, "grad_norm": 21.943064070286184, "learning_rate": 5.512538454085206e-06, "loss": 0.5119, "step": 12651 }, { "epoch": 1.98, "grad_norm": 12.590342849984218, "learning_rate": 5.51103145182634e-06, "loss": 0.493, "step": 12652 }, { "epoch": 1.98, "grad_norm": 16.1280554974113, "learning_rate": 5.509524577231114e-06, "loss": 0.5113, "step": 12653 }, { "epoch": 1.98, "grad_norm": 17.01443975128287, "learning_rate": 5.50801783034238e-06, "loss": 0.4957, "step": 12654 }, { "epoch": 1.98, "grad_norm": 18.94761608709194, "learning_rate": 5.506511211202997e-06, "loss": 0.5017, "step": 12655 }, { "epoch": 1.98, "grad_norm": 18.506820794649318, "learning_rate": 5.505004719855806e-06, "loss": 0.4814, "step": 12656 }, { "epoch": 1.98, "grad_norm": 15.043393147162146, "learning_rate": 5.503498356343648e-06, "loss": 0.4844, "step": 12657 }, { "epoch": 1.98, "grad_norm": 15.508233658479071, "learning_rate": 5.501992120709367e-06, "loss": 0.4701, "step": 12658 }, { "epoch": 1.98, "grad_norm": 14.224943003368498, "learning_rate": 5.500486012995801e-06, "loss": 0.4635, "step": 12659 }, { "epoch": 1.98, "grad_norm": 16.72161179764742, "learning_rate": 5.498980033245782e-06, "loss": 0.5972, "step": 12660 }, { "epoch": 1.98, "grad_norm": 18.86460772424473, "learning_rate": 5.4974741815021336e-06, "loss": 0.5446, "step": 12661 }, { "epoch": 1.98, "grad_norm": 21.606701164842864, "learning_rate": 5.4959684578076855e-06, "loss": 0.4804, "step": 12662 }, { "epoch": 1.98, "grad_norm": 23.86005248636155, "learning_rate": 5.494462862205263e-06, "loss": 0.5453, "step": 12663 }, { "epoch": 1.98, "grad_norm": 18.910636331555022, "learning_rate": 5.492957394737677e-06, "loss": 0.5336, "step": 12664 }, { "epoch": 1.98, "grad_norm": 16.01784423432177, "learning_rate": 5.49145205544775e-06, "loss": 0.4901, "step": 12665 }, { "epoch": 1.98, "grad_norm": 26.87061895546319, "learning_rate": 5.4899468443782864e-06, "loss": 0.4757, "step": 12666 }, { "epoch": 1.98, "grad_norm": 31.910062698953737, "learning_rate": 5.488441761572096e-06, "loss": 0.5706, "step": 12667 }, { "epoch": 1.98, "grad_norm": 25.82587874949454, "learning_rate": 5.486936807071986e-06, "loss": 0.532, "step": 12668 }, { "epoch": 1.98, "grad_norm": 18.129131046965366, "learning_rate": 5.485431980920753e-06, "loss": 0.5269, "step": 12669 }, { "epoch": 1.98, "grad_norm": 23.77725639534025, "learning_rate": 5.4839272831611905e-06, "loss": 0.5058, "step": 12670 }, { "epoch": 1.98, "grad_norm": 19.692392764572553, "learning_rate": 5.482422713836094e-06, "loss": 0.4975, "step": 12671 }, { "epoch": 1.98, "grad_norm": 22.37550091415856, "learning_rate": 5.480918272988256e-06, "loss": 0.5867, "step": 12672 }, { "epoch": 1.98, "grad_norm": 12.749561082540527, "learning_rate": 5.4794139606604556e-06, "loss": 0.4644, "step": 12673 }, { "epoch": 1.98, "grad_norm": 19.30856260188741, "learning_rate": 5.477909776895481e-06, "loss": 0.5362, "step": 12674 }, { "epoch": 1.98, "grad_norm": 15.235047056584289, "learning_rate": 5.476405721736104e-06, "loss": 0.5213, "step": 12675 }, { "epoch": 1.98, "grad_norm": 19.307820831949023, "learning_rate": 5.474901795225105e-06, "loss": 0.4949, "step": 12676 }, { "epoch": 1.98, "grad_norm": 26.21044316117008, "learning_rate": 5.473397997405249e-06, "loss": 0.591, "step": 12677 }, { "epoch": 1.98, "grad_norm": 24.42783966066164, "learning_rate": 5.471894328319305e-06, "loss": 0.5754, "step": 12678 }, { "epoch": 1.98, "grad_norm": 20.827380236440565, "learning_rate": 5.470390788010042e-06, "loss": 0.4074, "step": 12679 }, { "epoch": 1.98, "grad_norm": 16.652490441152498, "learning_rate": 5.4688873765202114e-06, "loss": 0.476, "step": 12680 }, { "epoch": 1.98, "grad_norm": 16.85093569175643, "learning_rate": 5.467384093892576e-06, "loss": 0.5046, "step": 12681 }, { "epoch": 1.98, "grad_norm": 21.883627884861234, "learning_rate": 5.465880940169881e-06, "loss": 0.4855, "step": 12682 }, { "epoch": 1.98, "grad_norm": 22.787404535248456, "learning_rate": 5.464377915394882e-06, "loss": 0.4596, "step": 12683 }, { "epoch": 1.98, "grad_norm": 23.843085502224415, "learning_rate": 5.4628750196103245e-06, "loss": 0.5642, "step": 12684 }, { "epoch": 1.98, "grad_norm": 15.258187935913085, "learning_rate": 5.461372252858948e-06, "loss": 0.4234, "step": 12685 }, { "epoch": 1.98, "grad_norm": 20.812965002678975, "learning_rate": 5.459869615183484e-06, "loss": 0.5264, "step": 12686 }, { "epoch": 1.98, "grad_norm": 20.075905620799684, "learning_rate": 5.458367106626674e-06, "loss": 0.4839, "step": 12687 }, { "epoch": 1.98, "grad_norm": 31.14433998849154, "learning_rate": 5.456864727231246e-06, "loss": 0.5968, "step": 12688 }, { "epoch": 1.98, "grad_norm": 15.173216462374063, "learning_rate": 5.455362477039935e-06, "loss": 0.4966, "step": 12689 }, { "epoch": 1.98, "grad_norm": 23.078196366625278, "learning_rate": 5.453860356095448e-06, "loss": 0.515, "step": 12690 }, { "epoch": 1.98, "grad_norm": 16.39923988531013, "learning_rate": 5.452358364440515e-06, "loss": 0.4839, "step": 12691 }, { "epoch": 1.98, "grad_norm": 34.13339292234002, "learning_rate": 5.450856502117849e-06, "loss": 0.5461, "step": 12692 }, { "epoch": 1.98, "grad_norm": 46.60784001688967, "learning_rate": 5.449354769170168e-06, "loss": 0.4725, "step": 12693 }, { "epoch": 1.98, "grad_norm": 16.84592823884347, "learning_rate": 5.447853165640173e-06, "loss": 0.4901, "step": 12694 }, { "epoch": 1.98, "grad_norm": 16.6065309328846, "learning_rate": 5.4463516915705684e-06, "loss": 0.5196, "step": 12695 }, { "epoch": 1.98, "grad_norm": 16.75612199254925, "learning_rate": 5.444850347004058e-06, "loss": 0.4887, "step": 12696 }, { "epoch": 1.98, "grad_norm": 19.72679947134772, "learning_rate": 5.443349131983343e-06, "loss": 0.5495, "step": 12697 }, { "epoch": 1.98, "grad_norm": 16.374805875166246, "learning_rate": 5.441848046551108e-06, "loss": 0.4772, "step": 12698 }, { "epoch": 1.98, "grad_norm": 23.827177509168333, "learning_rate": 5.440347090750053e-06, "loss": 0.4797, "step": 12699 }, { "epoch": 1.98, "grad_norm": 18.745513655058687, "learning_rate": 5.438846264622857e-06, "loss": 0.5091, "step": 12700 }, { "epoch": 1.98, "grad_norm": 14.107582728295293, "learning_rate": 5.437345568212207e-06, "loss": 0.4981, "step": 12701 }, { "epoch": 1.98, "grad_norm": 27.19873679217685, "learning_rate": 5.435845001560775e-06, "loss": 0.6202, "step": 12702 }, { "epoch": 1.98, "grad_norm": 22.438463846506803, "learning_rate": 5.434344564711244e-06, "loss": 0.534, "step": 12703 }, { "epoch": 1.98, "grad_norm": 34.288939828016595, "learning_rate": 5.432844257706285e-06, "loss": 0.5175, "step": 12704 }, { "epoch": 1.98, "grad_norm": 22.999118775092427, "learning_rate": 5.431344080588561e-06, "loss": 0.479, "step": 12705 }, { "epoch": 1.98, "grad_norm": 14.19710137900284, "learning_rate": 5.4298440334007415e-06, "loss": 0.4579, "step": 12706 }, { "epoch": 1.98, "grad_norm": 13.66095380333795, "learning_rate": 5.42834411618548e-06, "loss": 0.5239, "step": 12707 }, { "epoch": 1.99, "grad_norm": 19.466056785167183, "learning_rate": 5.426844328985439e-06, "loss": 0.4562, "step": 12708 }, { "epoch": 1.99, "grad_norm": 14.137920294845857, "learning_rate": 5.4253446718432724e-06, "loss": 0.5183, "step": 12709 }, { "epoch": 1.99, "grad_norm": 22.31210413416006, "learning_rate": 5.423845144801627e-06, "loss": 0.449, "step": 12710 }, { "epoch": 1.99, "grad_norm": 18.312527322911524, "learning_rate": 5.422345747903146e-06, "loss": 0.581, "step": 12711 }, { "epoch": 1.99, "grad_norm": 11.636423452958091, "learning_rate": 5.420846481190473e-06, "loss": 0.4649, "step": 12712 }, { "epoch": 1.99, "grad_norm": 20.911916545742397, "learning_rate": 5.41934734470625e-06, "loss": 0.443, "step": 12713 }, { "epoch": 1.99, "grad_norm": 22.308307962485706, "learning_rate": 5.417848338493114e-06, "loss": 0.4783, "step": 12714 }, { "epoch": 1.99, "grad_norm": 12.178909942491316, "learning_rate": 5.416349462593684e-06, "loss": 0.4977, "step": 12715 }, { "epoch": 1.99, "grad_norm": 19.61768315827325, "learning_rate": 5.414850717050593e-06, "loss": 0.567, "step": 12716 }, { "epoch": 1.99, "grad_norm": 31.15318871120512, "learning_rate": 5.413352101906466e-06, "loss": 0.5068, "step": 12717 }, { "epoch": 1.99, "grad_norm": 15.16018160429881, "learning_rate": 5.411853617203926e-06, "loss": 0.5339, "step": 12718 }, { "epoch": 1.99, "grad_norm": 19.35363165088891, "learning_rate": 5.410355262985585e-06, "loss": 0.6084, "step": 12719 }, { "epoch": 1.99, "grad_norm": 24.946300817947172, "learning_rate": 5.408857039294052e-06, "loss": 0.606, "step": 12720 }, { "epoch": 1.99, "grad_norm": 17.25809714569129, "learning_rate": 5.407358946171939e-06, "loss": 0.5121, "step": 12721 }, { "epoch": 1.99, "grad_norm": 20.165364431807888, "learning_rate": 5.405860983661854e-06, "loss": 0.5182, "step": 12722 }, { "epoch": 1.99, "grad_norm": 15.306420280230563, "learning_rate": 5.404363151806397e-06, "loss": 0.5631, "step": 12723 }, { "epoch": 1.99, "grad_norm": 21.31934463978182, "learning_rate": 5.402865450648158e-06, "loss": 0.5226, "step": 12724 }, { "epoch": 1.99, "grad_norm": 19.09855279634109, "learning_rate": 5.401367880229737e-06, "loss": 0.4902, "step": 12725 }, { "epoch": 1.99, "grad_norm": 20.092424936343008, "learning_rate": 5.399870440593726e-06, "loss": 0.4783, "step": 12726 }, { "epoch": 1.99, "grad_norm": 24.567248699527127, "learning_rate": 5.3983731317827075e-06, "loss": 0.4904, "step": 12727 }, { "epoch": 1.99, "grad_norm": 24.319415452431414, "learning_rate": 5.396875953839267e-06, "loss": 0.5109, "step": 12728 }, { "epoch": 1.99, "grad_norm": 22.834170988666447, "learning_rate": 5.3953789068059785e-06, "loss": 0.47, "step": 12729 }, { "epoch": 1.99, "grad_norm": 15.93428220054176, "learning_rate": 5.3938819907254204e-06, "loss": 0.499, "step": 12730 }, { "epoch": 1.99, "grad_norm": 15.772634888247694, "learning_rate": 5.392385205640167e-06, "loss": 0.4686, "step": 12731 }, { "epoch": 1.99, "grad_norm": 28.175442800283278, "learning_rate": 5.390888551592779e-06, "loss": 0.5549, "step": 12732 }, { "epoch": 1.99, "grad_norm": 22.662389325299316, "learning_rate": 5.3893920286258285e-06, "loss": 0.4863, "step": 12733 }, { "epoch": 1.99, "grad_norm": 22.161478477555537, "learning_rate": 5.387895636781868e-06, "loss": 0.531, "step": 12734 }, { "epoch": 1.99, "grad_norm": 15.2288172786045, "learning_rate": 5.386399376103462e-06, "loss": 0.5023, "step": 12735 }, { "epoch": 1.99, "grad_norm": 29.600610627620956, "learning_rate": 5.3849032466331555e-06, "loss": 0.5666, "step": 12736 }, { "epoch": 1.99, "grad_norm": 22.337931923220637, "learning_rate": 5.3834072484134995e-06, "loss": 0.4963, "step": 12737 }, { "epoch": 1.99, "grad_norm": 30.882099074854956, "learning_rate": 5.381911381487044e-06, "loss": 0.5255, "step": 12738 }, { "epoch": 1.99, "grad_norm": 17.9306973341253, "learning_rate": 5.380415645896329e-06, "loss": 0.5578, "step": 12739 }, { "epoch": 1.99, "grad_norm": 23.92875505300171, "learning_rate": 5.378920041683886e-06, "loss": 0.5569, "step": 12740 }, { "epoch": 1.99, "grad_norm": 25.04182570708978, "learning_rate": 5.377424568892253e-06, "loss": 0.5425, "step": 12741 }, { "epoch": 1.99, "grad_norm": 16.041537775774856, "learning_rate": 5.375929227563963e-06, "loss": 0.4817, "step": 12742 }, { "epoch": 1.99, "grad_norm": 15.55573722194232, "learning_rate": 5.374434017741543e-06, "loss": 0.5313, "step": 12743 }, { "epoch": 1.99, "grad_norm": 23.24591995787106, "learning_rate": 5.372938939467514e-06, "loss": 0.5362, "step": 12744 }, { "epoch": 1.99, "grad_norm": 23.268212947544505, "learning_rate": 5.371443992784391e-06, "loss": 0.5324, "step": 12745 }, { "epoch": 1.99, "grad_norm": 14.161217592649967, "learning_rate": 5.3699491777346935e-06, "loss": 0.4901, "step": 12746 }, { "epoch": 1.99, "grad_norm": 22.94016474127129, "learning_rate": 5.3684544943609375e-06, "loss": 0.5367, "step": 12747 }, { "epoch": 1.99, "grad_norm": 19.399479291889282, "learning_rate": 5.366959942705625e-06, "loss": 0.567, "step": 12748 }, { "epoch": 1.99, "grad_norm": 15.848442611949254, "learning_rate": 5.3654655228112574e-06, "loss": 0.5463, "step": 12749 }, { "epoch": 1.99, "grad_norm": 21.540081386106916, "learning_rate": 5.36397123472034e-06, "loss": 0.5711, "step": 12750 }, { "epoch": 1.99, "grad_norm": 17.756696683615296, "learning_rate": 5.362477078475372e-06, "loss": 0.4785, "step": 12751 }, { "epoch": 1.99, "grad_norm": 27.4221936579608, "learning_rate": 5.36098305411884e-06, "loss": 0.5048, "step": 12752 }, { "epoch": 1.99, "grad_norm": 16.355368281571305, "learning_rate": 5.35948916169324e-06, "loss": 0.4539, "step": 12753 }, { "epoch": 1.99, "grad_norm": 19.192543244568437, "learning_rate": 5.357995401241049e-06, "loss": 0.5127, "step": 12754 }, { "epoch": 1.99, "grad_norm": 17.156664353024194, "learning_rate": 5.3565017728047545e-06, "loss": 0.5508, "step": 12755 }, { "epoch": 1.99, "grad_norm": 22.408044845592613, "learning_rate": 5.355008276426836e-06, "loss": 0.5727, "step": 12756 }, { "epoch": 1.99, "grad_norm": 27.378817416501576, "learning_rate": 5.353514912149761e-06, "loss": 0.5143, "step": 12757 }, { "epoch": 1.99, "grad_norm": 15.563650196440346, "learning_rate": 5.3520216800160085e-06, "loss": 0.489, "step": 12758 }, { "epoch": 1.99, "grad_norm": 22.96476733199368, "learning_rate": 5.350528580068035e-06, "loss": 0.545, "step": 12759 }, { "epoch": 1.99, "grad_norm": 23.688638459101494, "learning_rate": 5.349035612348314e-06, "loss": 0.4843, "step": 12760 }, { "epoch": 1.99, "grad_norm": 16.728049830840234, "learning_rate": 5.347542776899295e-06, "loss": 0.4693, "step": 12761 }, { "epoch": 1.99, "grad_norm": 23.359305906672024, "learning_rate": 5.346050073763438e-06, "loss": 0.5438, "step": 12762 }, { "epoch": 1.99, "grad_norm": 19.39442096690375, "learning_rate": 5.344557502983198e-06, "loss": 0.6188, "step": 12763 }, { "epoch": 1.99, "grad_norm": 15.823107072761507, "learning_rate": 5.343065064601021e-06, "loss": 0.5734, "step": 12764 }, { "epoch": 1.99, "grad_norm": 24.274501514472956, "learning_rate": 5.3415727586593435e-06, "loss": 0.5177, "step": 12765 }, { "epoch": 1.99, "grad_norm": 18.339050194956446, "learning_rate": 5.340080585200612e-06, "loss": 0.5186, "step": 12766 }, { "epoch": 1.99, "grad_norm": 15.009594334932329, "learning_rate": 5.338588544267267e-06, "loss": 0.5078, "step": 12767 }, { "epoch": 1.99, "grad_norm": 21.263938690498964, "learning_rate": 5.3370966359017325e-06, "loss": 0.5454, "step": 12768 }, { "epoch": 1.99, "grad_norm": 16.471205769851714, "learning_rate": 5.335604860146446e-06, "loss": 0.4558, "step": 12769 }, { "epoch": 1.99, "grad_norm": 12.348412244402393, "learning_rate": 5.334113217043826e-06, "loss": 0.4622, "step": 12770 }, { "epoch": 1.99, "grad_norm": 28.880391896280504, "learning_rate": 5.332621706636295e-06, "loss": 0.5254, "step": 12771 }, { "epoch": 2.0, "grad_norm": 18.312226957834174, "learning_rate": 5.331130328966276e-06, "loss": 0.5503, "step": 12772 }, { "epoch": 2.0, "grad_norm": 19.21481608547601, "learning_rate": 5.329639084076181e-06, "loss": 0.5063, "step": 12773 }, { "epoch": 2.0, "grad_norm": 23.63318441555478, "learning_rate": 5.3281479720084125e-06, "loss": 0.4939, "step": 12774 }, { "epoch": 2.0, "grad_norm": 17.814329478358903, "learning_rate": 5.326656992805384e-06, "loss": 0.4858, "step": 12775 }, { "epoch": 2.0, "grad_norm": 17.21986106136835, "learning_rate": 5.325166146509497e-06, "loss": 0.4915, "step": 12776 }, { "epoch": 2.0, "grad_norm": 17.45065739390407, "learning_rate": 5.323675433163158e-06, "loss": 0.5109, "step": 12777 }, { "epoch": 2.0, "grad_norm": 18.690695171323338, "learning_rate": 5.322184852808745e-06, "loss": 0.5427, "step": 12778 }, { "epoch": 2.0, "grad_norm": 21.935566199145306, "learning_rate": 5.32069440548866e-06, "loss": 0.4739, "step": 12779 }, { "epoch": 2.0, "grad_norm": 15.900462675804087, "learning_rate": 5.319204091245287e-06, "loss": 0.5493, "step": 12780 }, { "epoch": 2.0, "grad_norm": 16.957816886321336, "learning_rate": 5.317713910121016e-06, "loss": 0.5106, "step": 12781 }, { "epoch": 2.0, "grad_norm": 12.700480878612106, "learning_rate": 5.316223862158223e-06, "loss": 0.4849, "step": 12782 }, { "epoch": 2.0, "grad_norm": 22.165340633469405, "learning_rate": 5.31473394739928e-06, "loss": 0.486, "step": 12783 }, { "epoch": 2.0, "grad_norm": 19.21583252138848, "learning_rate": 5.313244165886562e-06, "loss": 0.5137, "step": 12784 }, { "epoch": 2.0, "grad_norm": 15.628529801300667, "learning_rate": 5.311754517662442e-06, "loss": 0.5165, "step": 12785 }, { "epoch": 2.0, "grad_norm": 28.03665938858337, "learning_rate": 5.310265002769277e-06, "loss": 0.5386, "step": 12786 }, { "epoch": 2.0, "grad_norm": 26.411248315271912, "learning_rate": 5.308775621249435e-06, "loss": 0.5839, "step": 12787 }, { "epoch": 2.0, "grad_norm": 19.33712262346072, "learning_rate": 5.3072863731452674e-06, "loss": 0.5538, "step": 12788 }, { "epoch": 2.0, "grad_norm": 19.014096949390407, "learning_rate": 5.305797258499134e-06, "loss": 0.5904, "step": 12789 }, { "epoch": 2.0, "grad_norm": 24.258354472127433, "learning_rate": 5.304308277353376e-06, "loss": 0.5926, "step": 12790 }, { "epoch": 2.0, "grad_norm": 28.179351920976487, "learning_rate": 5.3028194297503445e-06, "loss": 0.5284, "step": 12791 }, { "epoch": 2.0, "grad_norm": 23.945392149252918, "learning_rate": 5.301330715732385e-06, "loss": 0.5319, "step": 12792 }, { "epoch": 2.0, "grad_norm": 18.357873436416508, "learning_rate": 5.299842135341825e-06, "loss": 0.5489, "step": 12793 }, { "epoch": 2.0, "grad_norm": 20.088004487734466, "learning_rate": 5.29835368862101e-06, "loss": 0.5624, "step": 12794 }, { "epoch": 2.0, "grad_norm": 16.059239912012366, "learning_rate": 5.296865375612264e-06, "loss": 0.5117, "step": 12795 }, { "epoch": 2.0, "grad_norm": 19.88174735815471, "learning_rate": 5.295377196357914e-06, "loss": 0.5244, "step": 12796 }, { "epoch": 2.0, "grad_norm": 24.131125571783016, "learning_rate": 5.293889150900287e-06, "loss": 0.5465, "step": 12797 }, { "epoch": 2.0, "grad_norm": 19.5814515033147, "learning_rate": 5.2924012392817014e-06, "loss": 0.5624, "step": 12798 }, { "epoch": 2.0, "grad_norm": 25.22358408168549, "learning_rate": 5.290913461544466e-06, "loss": 0.5068, "step": 12799 }, { "epoch": 2.0, "grad_norm": 17.86146255158927, "learning_rate": 5.289425817730897e-06, "loss": 0.4743, "step": 12800 }, { "epoch": 2.0, "grad_norm": 22.716995784140252, "learning_rate": 5.287938307883302e-06, "loss": 0.5308, "step": 12801 }, { "epoch": 2.0, "grad_norm": 20.709844466138364, "learning_rate": 5.286450932043994e-06, "loss": 0.4986, "step": 12802 }, { "epoch": 2.0, "grad_norm": 15.515052494807907, "learning_rate": 5.284963690255254e-06, "loss": 0.4749, "step": 12803 }, { "epoch": 2.0, "grad_norm": 14.939860179421112, "learning_rate": 5.28347658255939e-06, "loss": 0.5057, "step": 12804 }, { "epoch": 2.0, "grad_norm": 17.851529592571353, "learning_rate": 5.281989608998693e-06, "loss": 0.5132, "step": 12805 }, { "epoch": 2.0, "grad_norm": 14.747981189919482, "learning_rate": 5.280502769615456e-06, "loss": 0.4984, "step": 12806 }, { "epoch": 2.0, "grad_norm": 20.71458497372247, "learning_rate": 5.279016064451959e-06, "loss": 0.5556, "step": 12807 }, { "epoch": 2.0, "grad_norm": 15.483231372346973, "learning_rate": 5.277529493550478e-06, "loss": 0.5781, "step": 12808 }, { "epoch": 2.0, "grad_norm": 14.501442855437578, "learning_rate": 5.276043056953297e-06, "loss": 0.4606, "step": 12809 }, { "epoch": 2.0, "grad_norm": 34.23938197375652, "learning_rate": 5.274556754702691e-06, "loss": 0.521, "step": 12810 }, { "epoch": 2.0, "grad_norm": 22.855925610686196, "learning_rate": 5.273070586840925e-06, "loss": 0.5214, "step": 12811 }, { "epoch": 2.0, "grad_norm": 25.92509189624372, "learning_rate": 5.271584553410267e-06, "loss": 0.5487, "step": 12812 }, { "epoch": 2.0, "grad_norm": 15.282506018392025, "learning_rate": 5.270098654452977e-06, "loss": 0.5312, "step": 12813 }, { "epoch": 2.0, "grad_norm": 15.534534498495953, "learning_rate": 5.268612890011318e-06, "loss": 0.4754, "step": 12814 }, { "epoch": 2.0, "grad_norm": 13.077242536635687, "learning_rate": 5.267127260127536e-06, "loss": 0.5234, "step": 12815 }, { "epoch": 2.0, "grad_norm": 19.95447469177707, "learning_rate": 5.26564176484389e-06, "loss": 0.5423, "step": 12816 }, { "epoch": 2.0, "grad_norm": 20.944773742849737, "learning_rate": 5.264156404202618e-06, "loss": 0.5707, "step": 12817 }, { "epoch": 2.0, "grad_norm": 21.50944322472432, "learning_rate": 5.262671178245968e-06, "loss": 0.5324, "step": 12818 }, { "epoch": 2.0, "grad_norm": 25.443451093896535, "learning_rate": 5.261186087016183e-06, "loss": 0.533, "step": 12819 }, { "epoch": 2.0, "grad_norm": 13.053129724739613, "learning_rate": 5.259701130555489e-06, "loss": 0.4852, "step": 12820 }, { "epoch": 2.0, "grad_norm": 20.67333586642384, "learning_rate": 5.2582163089061255e-06, "loss": 0.5785, "step": 12821 }, { "epoch": 2.0, "grad_norm": 14.678764346717594, "learning_rate": 5.256731622110314e-06, "loss": 0.4252, "step": 12822 }, { "epoch": 2.0, "grad_norm": 20.637510476205016, "learning_rate": 5.2552470702102835e-06, "loss": 0.5379, "step": 12823 }, { "epoch": 2.0, "grad_norm": 19.331672049249445, "learning_rate": 5.253762653248245e-06, "loss": 0.526, "step": 12824 }, { "epoch": 2.0, "grad_norm": 16.219938923691952, "learning_rate": 5.252278371266422e-06, "loss": 0.449, "step": 12825 }, { "epoch": 2.0, "grad_norm": 16.178244468847648, "learning_rate": 5.250794224307028e-06, "loss": 0.4775, "step": 12826 }, { "epoch": 2.0, "grad_norm": 15.999490883948951, "learning_rate": 5.249310212412269e-06, "loss": 0.503, "step": 12827 }, { "epoch": 2.0, "grad_norm": 20.402767786844866, "learning_rate": 5.247826335624344e-06, "loss": 0.5055, "step": 12828 }, { "epoch": 2.0, "grad_norm": 19.37863443841836, "learning_rate": 5.246342593985458e-06, "loss": 0.5109, "step": 12829 }, { "epoch": 2.0, "grad_norm": 15.144753117740438, "learning_rate": 5.244858987537808e-06, "loss": 0.482, "step": 12830 }, { "epoch": 2.0, "grad_norm": 17.419721878328797, "learning_rate": 5.243375516323591e-06, "loss": 0.5129, "step": 12831 }, { "epoch": 2.0, "grad_norm": 18.602522359449324, "learning_rate": 5.241892180384991e-06, "loss": 0.511, "step": 12832 }, { "epoch": 2.0, "grad_norm": 22.599632476802746, "learning_rate": 5.240408979764191e-06, "loss": 0.4747, "step": 12833 }, { "epoch": 2.0, "grad_norm": 17.39301927661398, "learning_rate": 5.2389259145033744e-06, "loss": 0.5152, "step": 12834 }, { "epoch": 2.0, "grad_norm": 21.495350274620353, "learning_rate": 5.237442984644724e-06, "loss": 0.4846, "step": 12835 }, { "epoch": 2.0, "grad_norm": 15.860988160947693, "learning_rate": 5.235960190230409e-06, "loss": 0.4342, "step": 12836 }, { "epoch": 2.01, "grad_norm": 19.270513452675903, "learning_rate": 5.234477531302595e-06, "loss": 0.5372, "step": 12837 }, { "epoch": 2.01, "grad_norm": 16.419396131694374, "learning_rate": 5.232995007903453e-06, "loss": 0.534, "step": 12838 }, { "epoch": 2.01, "grad_norm": 33.4992244745453, "learning_rate": 5.231512620075143e-06, "loss": 0.566, "step": 12839 }, { "epoch": 2.01, "grad_norm": 14.161195104967208, "learning_rate": 5.23003036785983e-06, "loss": 0.4307, "step": 12840 }, { "epoch": 2.01, "grad_norm": 19.680810002096436, "learning_rate": 5.228548251299663e-06, "loss": 0.429, "step": 12841 }, { "epoch": 2.01, "grad_norm": 22.076746387999645, "learning_rate": 5.227066270436788e-06, "loss": 0.5123, "step": 12842 }, { "epoch": 2.01, "grad_norm": 24.717862502103365, "learning_rate": 5.225584425313357e-06, "loss": 0.5131, "step": 12843 }, { "epoch": 2.01, "grad_norm": 18.42245341178339, "learning_rate": 5.224102715971515e-06, "loss": 0.4998, "step": 12844 }, { "epoch": 2.01, "grad_norm": 11.873891543429119, "learning_rate": 5.222621142453394e-06, "loss": 0.506, "step": 12845 }, { "epoch": 2.01, "grad_norm": 24.739271958784194, "learning_rate": 5.221139704801137e-06, "loss": 0.4974, "step": 12846 }, { "epoch": 2.01, "grad_norm": 18.322773141463294, "learning_rate": 5.219658403056867e-06, "loss": 0.4874, "step": 12847 }, { "epoch": 2.01, "grad_norm": 18.303040486121585, "learning_rate": 5.218177237262721e-06, "loss": 0.5105, "step": 12848 }, { "epoch": 2.01, "grad_norm": 18.723988803755613, "learning_rate": 5.216696207460812e-06, "loss": 0.5677, "step": 12849 }, { "epoch": 2.01, "grad_norm": 14.079114685983859, "learning_rate": 5.215215313693265e-06, "loss": 0.4846, "step": 12850 }, { "epoch": 2.01, "grad_norm": 19.86693381951216, "learning_rate": 5.213734556002199e-06, "loss": 0.5356, "step": 12851 }, { "epoch": 2.01, "grad_norm": 23.769928964439313, "learning_rate": 5.212253934429723e-06, "loss": 0.5335, "step": 12852 }, { "epoch": 2.01, "grad_norm": 24.150151082419267, "learning_rate": 5.210773449017939e-06, "loss": 0.5059, "step": 12853 }, { "epoch": 2.01, "grad_norm": 19.212346614400335, "learning_rate": 5.209293099808959e-06, "loss": 0.4648, "step": 12854 }, { "epoch": 2.01, "grad_norm": 24.03153420504713, "learning_rate": 5.207812886844879e-06, "loss": 0.5475, "step": 12855 }, { "epoch": 2.01, "grad_norm": 23.486372312635062, "learning_rate": 5.2063328101678004e-06, "loss": 0.5562, "step": 12856 }, { "epoch": 2.01, "grad_norm": 15.595165511905176, "learning_rate": 5.204852869819814e-06, "loss": 0.4704, "step": 12857 }, { "epoch": 2.01, "grad_norm": 20.74521933199456, "learning_rate": 5.203373065843003e-06, "loss": 0.5267, "step": 12858 }, { "epoch": 2.01, "grad_norm": 22.16987454864099, "learning_rate": 5.201893398279454e-06, "loss": 0.5634, "step": 12859 }, { "epoch": 2.01, "grad_norm": 55.38955412584117, "learning_rate": 5.2004138671712555e-06, "loss": 0.5711, "step": 12860 }, { "epoch": 2.01, "grad_norm": 20.553992401398364, "learning_rate": 5.198934472560479e-06, "loss": 0.5495, "step": 12861 }, { "epoch": 2.01, "grad_norm": 18.043493991234104, "learning_rate": 5.197455214489193e-06, "loss": 0.4802, "step": 12862 }, { "epoch": 2.01, "grad_norm": 20.467267454489008, "learning_rate": 5.195976092999472e-06, "loss": 0.5754, "step": 12863 }, { "epoch": 2.01, "grad_norm": 15.545558530574674, "learning_rate": 5.194497108133381e-06, "loss": 0.5344, "step": 12864 }, { "epoch": 2.01, "grad_norm": 12.526643115518555, "learning_rate": 5.193018259932987e-06, "loss": 0.4985, "step": 12865 }, { "epoch": 2.01, "grad_norm": 20.22003788214398, "learning_rate": 5.191539548440336e-06, "loss": 0.5001, "step": 12866 }, { "epoch": 2.01, "grad_norm": 22.295615203409778, "learning_rate": 5.190060973697486e-06, "loss": 0.5713, "step": 12867 }, { "epoch": 2.01, "grad_norm": 21.880064723342546, "learning_rate": 5.188582535746489e-06, "loss": 0.4764, "step": 12868 }, { "epoch": 2.01, "grad_norm": 24.816532976525576, "learning_rate": 5.187104234629394e-06, "loss": 0.4935, "step": 12869 }, { "epoch": 2.01, "grad_norm": 14.477431978855728, "learning_rate": 5.185626070388239e-06, "loss": 0.4337, "step": 12870 }, { "epoch": 2.01, "grad_norm": 21.118029455207974, "learning_rate": 5.184148043065058e-06, "loss": 0.5271, "step": 12871 }, { "epoch": 2.01, "grad_norm": 17.57273564775897, "learning_rate": 5.182670152701889e-06, "loss": 0.496, "step": 12872 }, { "epoch": 2.01, "grad_norm": 21.403528377403912, "learning_rate": 5.181192399340768e-06, "loss": 0.561, "step": 12873 }, { "epoch": 2.01, "grad_norm": 23.00317202921992, "learning_rate": 5.179714783023711e-06, "loss": 0.5042, "step": 12874 }, { "epoch": 2.01, "grad_norm": 18.11761677830964, "learning_rate": 5.17823730379275e-06, "loss": 0.4443, "step": 12875 }, { "epoch": 2.01, "grad_norm": 21.650502661525685, "learning_rate": 5.1767599616898965e-06, "loss": 0.5648, "step": 12876 }, { "epoch": 2.01, "grad_norm": 15.093180487730304, "learning_rate": 5.175282756757172e-06, "loss": 0.5177, "step": 12877 }, { "epoch": 2.01, "grad_norm": 21.59031068543, "learning_rate": 5.1738056890365775e-06, "loss": 0.4623, "step": 12878 }, { "epoch": 2.01, "grad_norm": 30.362432434131314, "learning_rate": 5.1723287585701285e-06, "loss": 0.5351, "step": 12879 }, { "epoch": 2.01, "grad_norm": 16.879641186058198, "learning_rate": 5.1708519653998275e-06, "loss": 0.5023, "step": 12880 }, { "epoch": 2.01, "grad_norm": 16.007617434015483, "learning_rate": 5.169375309567669e-06, "loss": 0.5119, "step": 12881 }, { "epoch": 2.01, "grad_norm": 31.5426782331286, "learning_rate": 5.167898791115654e-06, "loss": 0.5462, "step": 12882 }, { "epoch": 2.01, "grad_norm": 16.6085095704948, "learning_rate": 5.166422410085767e-06, "loss": 0.498, "step": 12883 }, { "epoch": 2.01, "grad_norm": 21.62602963228964, "learning_rate": 5.16494616652e-06, "loss": 0.5543, "step": 12884 }, { "epoch": 2.01, "grad_norm": 27.532708059870203, "learning_rate": 5.163470060460338e-06, "loss": 0.5182, "step": 12885 }, { "epoch": 2.01, "grad_norm": 13.561550753199148, "learning_rate": 5.161994091948759e-06, "loss": 0.5083, "step": 12886 }, { "epoch": 2.01, "grad_norm": 21.228462194603114, "learning_rate": 5.160518261027234e-06, "loss": 0.5143, "step": 12887 }, { "epoch": 2.01, "grad_norm": 15.861113874822482, "learning_rate": 5.159042567737739e-06, "loss": 0.5461, "step": 12888 }, { "epoch": 2.01, "grad_norm": 43.83926517526602, "learning_rate": 5.157567012122241e-06, "loss": 0.5495, "step": 12889 }, { "epoch": 2.01, "grad_norm": 14.808685678283762, "learning_rate": 5.1560915942227115e-06, "loss": 0.4999, "step": 12890 }, { "epoch": 2.01, "grad_norm": 27.285094730905854, "learning_rate": 5.154616314081098e-06, "loss": 0.4821, "step": 12891 }, { "epoch": 2.01, "grad_norm": 20.975509671121834, "learning_rate": 5.15314117173936e-06, "loss": 0.4987, "step": 12892 }, { "epoch": 2.01, "grad_norm": 21.795217110536342, "learning_rate": 5.151666167239452e-06, "loss": 0.576, "step": 12893 }, { "epoch": 2.01, "grad_norm": 24.030537334907187, "learning_rate": 5.150191300623326e-06, "loss": 0.5055, "step": 12894 }, { "epoch": 2.01, "grad_norm": 18.954326292021857, "learning_rate": 5.1487165719329225e-06, "loss": 0.4767, "step": 12895 }, { "epoch": 2.01, "grad_norm": 14.781848346946983, "learning_rate": 5.147241981210178e-06, "loss": 0.4882, "step": 12896 }, { "epoch": 2.01, "grad_norm": 26.721712180230348, "learning_rate": 5.145767528497032e-06, "loss": 0.495, "step": 12897 }, { "epoch": 2.01, "grad_norm": 16.71099786092121, "learning_rate": 5.1442932138354225e-06, "loss": 0.5591, "step": 12898 }, { "epoch": 2.01, "grad_norm": 16.50726376843265, "learning_rate": 5.142819037267271e-06, "loss": 0.4738, "step": 12899 }, { "epoch": 2.01, "grad_norm": 13.43351219853536, "learning_rate": 5.1413449988345055e-06, "loss": 0.446, "step": 12900 }, { "epoch": 2.02, "grad_norm": 14.86939762003122, "learning_rate": 5.139871098579045e-06, "loss": 0.4269, "step": 12901 }, { "epoch": 2.02, "grad_norm": 12.580438248351527, "learning_rate": 5.138397336542809e-06, "loss": 0.4845, "step": 12902 }, { "epoch": 2.02, "grad_norm": 30.333315181639414, "learning_rate": 5.136923712767706e-06, "loss": 0.4742, "step": 12903 }, { "epoch": 2.02, "grad_norm": 17.771202178233892, "learning_rate": 5.1354502272956486e-06, "loss": 0.5283, "step": 12904 }, { "epoch": 2.02, "grad_norm": 26.203682428366672, "learning_rate": 5.133976880168543e-06, "loss": 0.5423, "step": 12905 }, { "epoch": 2.02, "grad_norm": 17.895838072514895, "learning_rate": 5.132503671428286e-06, "loss": 0.4642, "step": 12906 }, { "epoch": 2.02, "grad_norm": 17.36616706681718, "learning_rate": 5.13103060111678e-06, "loss": 0.4566, "step": 12907 }, { "epoch": 2.02, "grad_norm": 15.740179368285217, "learning_rate": 5.129557669275913e-06, "loss": 0.5351, "step": 12908 }, { "epoch": 2.02, "grad_norm": 20.187535147887303, "learning_rate": 5.128084875947579e-06, "loss": 0.4946, "step": 12909 }, { "epoch": 2.02, "grad_norm": 19.08145395379845, "learning_rate": 5.126612221173659e-06, "loss": 0.4965, "step": 12910 }, { "epoch": 2.02, "grad_norm": 20.949772486213806, "learning_rate": 5.125139704996038e-06, "loss": 0.5379, "step": 12911 }, { "epoch": 2.02, "grad_norm": 18.47096047268034, "learning_rate": 5.123667327456591e-06, "loss": 0.4741, "step": 12912 }, { "epoch": 2.02, "grad_norm": 20.483268776302502, "learning_rate": 5.12219508859719e-06, "loss": 0.5142, "step": 12913 }, { "epoch": 2.02, "grad_norm": 18.783053883549393, "learning_rate": 5.1207229884597135e-06, "loss": 0.5842, "step": 12914 }, { "epoch": 2.02, "grad_norm": 16.618585379566344, "learning_rate": 5.119251027086019e-06, "loss": 0.4189, "step": 12915 }, { "epoch": 2.02, "grad_norm": 21.84419775846366, "learning_rate": 5.117779204517967e-06, "loss": 0.4689, "step": 12916 }, { "epoch": 2.02, "grad_norm": 21.524304974100705, "learning_rate": 5.116307520797419e-06, "loss": 0.5453, "step": 12917 }, { "epoch": 2.02, "grad_norm": 14.32835194981006, "learning_rate": 5.114835975966228e-06, "loss": 0.4066, "step": 12918 }, { "epoch": 2.02, "grad_norm": 24.56690370841834, "learning_rate": 5.113364570066248e-06, "loss": 0.5124, "step": 12919 }, { "epoch": 2.02, "grad_norm": 20.504089780884723, "learning_rate": 5.11189330313932e-06, "loss": 0.4682, "step": 12920 }, { "epoch": 2.02, "grad_norm": 21.077138459862923, "learning_rate": 5.110422175227284e-06, "loss": 0.5939, "step": 12921 }, { "epoch": 2.02, "grad_norm": 16.649965935713773, "learning_rate": 5.1089511863719824e-06, "loss": 0.4629, "step": 12922 }, { "epoch": 2.02, "grad_norm": 32.15808636562768, "learning_rate": 5.10748033661525e-06, "loss": 0.5891, "step": 12923 }, { "epoch": 2.02, "grad_norm": 19.038751400052202, "learning_rate": 5.106009625998913e-06, "loss": 0.4786, "step": 12924 }, { "epoch": 2.02, "grad_norm": 28.30107336689926, "learning_rate": 5.104539054564799e-06, "loss": 0.4997, "step": 12925 }, { "epoch": 2.02, "grad_norm": 15.10869506323825, "learning_rate": 5.103068622354729e-06, "loss": 0.5007, "step": 12926 }, { "epoch": 2.02, "grad_norm": 28.111912110840457, "learning_rate": 5.101598329410522e-06, "loss": 0.4682, "step": 12927 }, { "epoch": 2.02, "grad_norm": 20.111038363739276, "learning_rate": 5.100128175773998e-06, "loss": 0.4905, "step": 12928 }, { "epoch": 2.02, "grad_norm": 20.552109248155773, "learning_rate": 5.098658161486962e-06, "loss": 0.5043, "step": 12929 }, { "epoch": 2.02, "grad_norm": 21.00488729032167, "learning_rate": 5.097188286591217e-06, "loss": 0.4849, "step": 12930 }, { "epoch": 2.02, "grad_norm": 22.301912926835485, "learning_rate": 5.095718551128569e-06, "loss": 0.579, "step": 12931 }, { "epoch": 2.02, "grad_norm": 17.263912913502743, "learning_rate": 5.09424895514082e-06, "loss": 0.5646, "step": 12932 }, { "epoch": 2.02, "grad_norm": 26.040774268245876, "learning_rate": 5.092779498669758e-06, "loss": 0.4881, "step": 12933 }, { "epoch": 2.02, "grad_norm": 25.22300925885845, "learning_rate": 5.091310181757178e-06, "loss": 0.4967, "step": 12934 }, { "epoch": 2.02, "grad_norm": 15.216214411333613, "learning_rate": 5.089841004444864e-06, "loss": 0.5145, "step": 12935 }, { "epoch": 2.02, "grad_norm": 25.828643816694733, "learning_rate": 5.088371966774601e-06, "loss": 0.5384, "step": 12936 }, { "epoch": 2.02, "grad_norm": 16.97114113562809, "learning_rate": 5.086903068788163e-06, "loss": 0.5048, "step": 12937 }, { "epoch": 2.02, "grad_norm": 12.293208981455534, "learning_rate": 5.085434310527329e-06, "loss": 0.4853, "step": 12938 }, { "epoch": 2.02, "grad_norm": 14.845856726872723, "learning_rate": 5.083965692033871e-06, "loss": 0.4793, "step": 12939 }, { "epoch": 2.02, "grad_norm": 15.982750894963123, "learning_rate": 5.0824972133495535e-06, "loss": 0.5018, "step": 12940 }, { "epoch": 2.02, "grad_norm": 15.962981736401444, "learning_rate": 5.081028874516134e-06, "loss": 0.4511, "step": 12941 }, { "epoch": 2.02, "grad_norm": 25.991586146780836, "learning_rate": 5.0795606755753745e-06, "loss": 0.4761, "step": 12942 }, { "epoch": 2.02, "grad_norm": 18.59586374946476, "learning_rate": 5.0780926165690325e-06, "loss": 0.4913, "step": 12943 }, { "epoch": 2.02, "grad_norm": 26.261921909221112, "learning_rate": 5.076624697538861e-06, "loss": 0.5173, "step": 12944 }, { "epoch": 2.02, "grad_norm": 23.850943563273177, "learning_rate": 5.075156918526604e-06, "loss": 0.5424, "step": 12945 }, { "epoch": 2.02, "grad_norm": 16.629691340431602, "learning_rate": 5.073689279573999e-06, "loss": 0.4819, "step": 12946 }, { "epoch": 2.02, "grad_norm": 20.535175868433708, "learning_rate": 5.072221780722788e-06, "loss": 0.4638, "step": 12947 }, { "epoch": 2.02, "grad_norm": 14.420423019238626, "learning_rate": 5.070754422014711e-06, "loss": 0.4447, "step": 12948 }, { "epoch": 2.02, "grad_norm": 24.360977288989826, "learning_rate": 5.069287203491495e-06, "loss": 0.5058, "step": 12949 }, { "epoch": 2.02, "grad_norm": 16.359228791145526, "learning_rate": 5.0678201251948625e-06, "loss": 0.5314, "step": 12950 }, { "epoch": 2.02, "grad_norm": 18.21461601000422, "learning_rate": 5.066353187166542e-06, "loss": 0.5161, "step": 12951 }, { "epoch": 2.02, "grad_norm": 22.560840450553414, "learning_rate": 5.064886389448249e-06, "loss": 0.4922, "step": 12952 }, { "epoch": 2.02, "grad_norm": 20.521824713462816, "learning_rate": 5.063419732081704e-06, "loss": 0.5454, "step": 12953 }, { "epoch": 2.02, "grad_norm": 23.041056475886474, "learning_rate": 5.061953215108614e-06, "loss": 0.511, "step": 12954 }, { "epoch": 2.02, "grad_norm": 12.945096319738319, "learning_rate": 5.060486838570682e-06, "loss": 0.44, "step": 12955 }, { "epoch": 2.02, "grad_norm": 18.523696602369135, "learning_rate": 5.059020602509616e-06, "loss": 0.5487, "step": 12956 }, { "epoch": 2.02, "grad_norm": 26.189207472023572, "learning_rate": 5.057554506967116e-06, "loss": 0.5123, "step": 12957 }, { "epoch": 2.02, "grad_norm": 17.71969545792104, "learning_rate": 5.056088551984876e-06, "loss": 0.5108, "step": 12958 }, { "epoch": 2.02, "grad_norm": 21.007571547783737, "learning_rate": 5.0546227376045796e-06, "loss": 0.4888, "step": 12959 }, { "epoch": 2.02, "grad_norm": 18.61477990856951, "learning_rate": 5.053157063867922e-06, "loss": 0.5057, "step": 12960 }, { "epoch": 2.02, "grad_norm": 20.44929291894792, "learning_rate": 5.051691530816586e-06, "loss": 0.5329, "step": 12961 }, { "epoch": 2.02, "grad_norm": 16.669504545917782, "learning_rate": 5.050226138492245e-06, "loss": 0.4971, "step": 12962 }, { "epoch": 2.02, "grad_norm": 20.916932108291594, "learning_rate": 5.048760886936581e-06, "loss": 0.6495, "step": 12963 }, { "epoch": 2.02, "grad_norm": 18.93328422811248, "learning_rate": 5.0472957761912565e-06, "loss": 0.4565, "step": 12964 }, { "epoch": 2.03, "grad_norm": 16.554134242636305, "learning_rate": 5.0458308062979466e-06, "loss": 0.5064, "step": 12965 }, { "epoch": 2.03, "grad_norm": 31.569692142568133, "learning_rate": 5.0443659772983064e-06, "loss": 0.4578, "step": 12966 }, { "epoch": 2.03, "grad_norm": 16.86317067754007, "learning_rate": 5.042901289234001e-06, "loss": 0.4656, "step": 12967 }, { "epoch": 2.03, "grad_norm": 13.796203357704613, "learning_rate": 5.041436742146685e-06, "loss": 0.4735, "step": 12968 }, { "epoch": 2.03, "grad_norm": 23.234503629593902, "learning_rate": 5.039972336078004e-06, "loss": 0.5157, "step": 12969 }, { "epoch": 2.03, "grad_norm": 17.97618756199585, "learning_rate": 5.038508071069612e-06, "loss": 0.4808, "step": 12970 }, { "epoch": 2.03, "grad_norm": 23.701606671683642, "learning_rate": 5.037043947163145e-06, "loss": 0.5264, "step": 12971 }, { "epoch": 2.03, "grad_norm": 18.91798585272234, "learning_rate": 5.035579964400244e-06, "loss": 0.4856, "step": 12972 }, { "epoch": 2.03, "grad_norm": 19.52615643650113, "learning_rate": 5.034116122822547e-06, "loss": 0.4913, "step": 12973 }, { "epoch": 2.03, "grad_norm": 23.718765009371467, "learning_rate": 5.032652422471685e-06, "loss": 0.5045, "step": 12974 }, { "epoch": 2.03, "grad_norm": 24.632496651204306, "learning_rate": 5.031188863389278e-06, "loss": 0.5444, "step": 12975 }, { "epoch": 2.03, "grad_norm": 18.311159396900184, "learning_rate": 5.029725445616953e-06, "loss": 0.4897, "step": 12976 }, { "epoch": 2.03, "grad_norm": 13.556354788440725, "learning_rate": 5.028262169196329e-06, "loss": 0.4971, "step": 12977 }, { "epoch": 2.03, "grad_norm": 19.03218709035426, "learning_rate": 5.0267990341690275e-06, "loss": 0.5537, "step": 12978 }, { "epoch": 2.03, "grad_norm": 15.586155611199526, "learning_rate": 5.025336040576645e-06, "loss": 0.475, "step": 12979 }, { "epoch": 2.03, "grad_norm": 23.218267322505994, "learning_rate": 5.023873188460795e-06, "loss": 0.5585, "step": 12980 }, { "epoch": 2.03, "grad_norm": 18.621293800068443, "learning_rate": 5.02241047786308e-06, "loss": 0.5174, "step": 12981 }, { "epoch": 2.03, "grad_norm": 14.338429233924517, "learning_rate": 5.020947908825104e-06, "loss": 0.463, "step": 12982 }, { "epoch": 2.03, "grad_norm": 17.489757113012747, "learning_rate": 5.019485481388457e-06, "loss": 0.4849, "step": 12983 }, { "epoch": 2.03, "grad_norm": 20.381331134448654, "learning_rate": 5.018023195594726e-06, "loss": 0.4659, "step": 12984 }, { "epoch": 2.03, "grad_norm": 16.696198371658305, "learning_rate": 5.0165610514855005e-06, "loss": 0.545, "step": 12985 }, { "epoch": 2.03, "grad_norm": 27.280164352917183, "learning_rate": 5.0150990491023655e-06, "loss": 0.5153, "step": 12986 }, { "epoch": 2.03, "grad_norm": 15.186738514157582, "learning_rate": 5.013637188486895e-06, "loss": 0.4551, "step": 12987 }, { "epoch": 2.03, "grad_norm": 23.209177101321878, "learning_rate": 5.0121754696806705e-06, "loss": 0.5723, "step": 12988 }, { "epoch": 2.03, "grad_norm": 20.233382817101287, "learning_rate": 5.010713892725253e-06, "loss": 0.5307, "step": 12989 }, { "epoch": 2.03, "grad_norm": 24.51935531142266, "learning_rate": 5.009252457662215e-06, "loss": 0.528, "step": 12990 }, { "epoch": 2.03, "grad_norm": 24.229875616045803, "learning_rate": 5.00779116453312e-06, "loss": 0.5104, "step": 12991 }, { "epoch": 2.03, "grad_norm": 21.693085909591737, "learning_rate": 5.00633001337952e-06, "loss": 0.517, "step": 12992 }, { "epoch": 2.03, "grad_norm": 25.112386556125134, "learning_rate": 5.004869004242978e-06, "loss": 0.5073, "step": 12993 }, { "epoch": 2.03, "grad_norm": 19.9949028866268, "learning_rate": 5.0034081371650355e-06, "loss": 0.4546, "step": 12994 }, { "epoch": 2.03, "grad_norm": 20.344539489310822, "learning_rate": 5.001947412187246e-06, "loss": 0.5482, "step": 12995 }, { "epoch": 2.03, "grad_norm": 35.425448719903216, "learning_rate": 5.000486829351145e-06, "loss": 0.4497, "step": 12996 }, { "epoch": 2.03, "grad_norm": 23.94215794828907, "learning_rate": 4.999026388698274e-06, "loss": 0.4724, "step": 12997 }, { "epoch": 2.03, "grad_norm": 18.277678036389432, "learning_rate": 4.9975660902701695e-06, "loss": 0.5749, "step": 12998 }, { "epoch": 2.03, "grad_norm": 24.646998735564413, "learning_rate": 4.996105934108361e-06, "loss": 0.5397, "step": 12999 }, { "epoch": 2.03, "grad_norm": 18.888224960434503, "learning_rate": 4.9946459202543675e-06, "loss": 0.4837, "step": 13000 }, { "epoch": 2.03, "grad_norm": 18.404559647187043, "learning_rate": 4.993186048749717e-06, "loss": 0.4851, "step": 13001 }, { "epoch": 2.03, "grad_norm": 18.407126168639, "learning_rate": 4.991726319635925e-06, "loss": 0.5178, "step": 13002 }, { "epoch": 2.03, "grad_norm": 39.285891033021166, "learning_rate": 4.990266732954515e-06, "loss": 0.5457, "step": 13003 }, { "epoch": 2.03, "grad_norm": 19.235065248718087, "learning_rate": 4.98880728874698e-06, "loss": 0.5051, "step": 13004 }, { "epoch": 2.03, "grad_norm": 17.03958188909016, "learning_rate": 4.987347987054835e-06, "loss": 0.4714, "step": 13005 }, { "epoch": 2.03, "grad_norm": 18.90473123471736, "learning_rate": 4.9858888279195815e-06, "loss": 0.4831, "step": 13006 }, { "epoch": 2.03, "grad_norm": 19.53985441781149, "learning_rate": 4.984429811382721e-06, "loss": 0.468, "step": 13007 }, { "epoch": 2.03, "grad_norm": 18.609928158114585, "learning_rate": 4.982970937485741e-06, "loss": 0.5079, "step": 13008 }, { "epoch": 2.03, "grad_norm": 17.073784736051554, "learning_rate": 4.981512206270129e-06, "loss": 0.4932, "step": 13009 }, { "epoch": 2.03, "grad_norm": 22.075368560333942, "learning_rate": 4.980053617777375e-06, "loss": 0.4835, "step": 13010 }, { "epoch": 2.03, "grad_norm": 24.64095544623726, "learning_rate": 4.978595172048963e-06, "loss": 0.5551, "step": 13011 }, { "epoch": 2.03, "grad_norm": 17.01881777896616, "learning_rate": 4.977136869126366e-06, "loss": 0.4885, "step": 13012 }, { "epoch": 2.03, "grad_norm": 15.801097233917018, "learning_rate": 4.975678709051056e-06, "loss": 0.4988, "step": 13013 }, { "epoch": 2.03, "grad_norm": 22.741799172034856, "learning_rate": 4.974220691864503e-06, "loss": 0.4692, "step": 13014 }, { "epoch": 2.03, "grad_norm": 12.919911083698452, "learning_rate": 4.972762817608174e-06, "loss": 0.4664, "step": 13015 }, { "epoch": 2.03, "grad_norm": 29.579764216190913, "learning_rate": 4.971305086323534e-06, "loss": 0.5139, "step": 13016 }, { "epoch": 2.03, "grad_norm": 20.2319391323564, "learning_rate": 4.969847498052035e-06, "loss": 0.474, "step": 13017 }, { "epoch": 2.03, "grad_norm": 18.199246220057237, "learning_rate": 4.968390052835127e-06, "loss": 0.5309, "step": 13018 }, { "epoch": 2.03, "grad_norm": 15.906873815366412, "learning_rate": 4.966932750714262e-06, "loss": 0.4528, "step": 13019 }, { "epoch": 2.03, "grad_norm": 20.783104280692573, "learning_rate": 4.96547559173089e-06, "loss": 0.5059, "step": 13020 }, { "epoch": 2.03, "grad_norm": 19.0585573699973, "learning_rate": 4.964018575926442e-06, "loss": 0.4826, "step": 13021 }, { "epoch": 2.03, "grad_norm": 18.384747476709627, "learning_rate": 4.962561703342364e-06, "loss": 0.4838, "step": 13022 }, { "epoch": 2.03, "grad_norm": 23.105715717920873, "learning_rate": 4.961104974020081e-06, "loss": 0.5308, "step": 13023 }, { "epoch": 2.03, "grad_norm": 12.207606773542302, "learning_rate": 4.959648388001028e-06, "loss": 0.4892, "step": 13024 }, { "epoch": 2.03, "grad_norm": 19.095427183483118, "learning_rate": 4.958191945326624e-06, "loss": 0.4947, "step": 13025 }, { "epoch": 2.03, "grad_norm": 25.755325375488038, "learning_rate": 4.9567356460382895e-06, "loss": 0.4677, "step": 13026 }, { "epoch": 2.03, "grad_norm": 15.235071031858206, "learning_rate": 4.955279490177448e-06, "loss": 0.515, "step": 13027 }, { "epoch": 2.03, "grad_norm": 16.819117463320943, "learning_rate": 4.953823477785507e-06, "loss": 0.4928, "step": 13028 }, { "epoch": 2.04, "grad_norm": 22.049638209353752, "learning_rate": 4.952367608903871e-06, "loss": 0.4752, "step": 13029 }, { "epoch": 2.04, "grad_norm": 20.620339938601706, "learning_rate": 4.9509118835739465e-06, "loss": 0.5261, "step": 13030 }, { "epoch": 2.04, "grad_norm": 20.11174514781223, "learning_rate": 4.949456301837134e-06, "loss": 0.4821, "step": 13031 }, { "epoch": 2.04, "grad_norm": 14.63069351136901, "learning_rate": 4.948000863734834e-06, "loss": 0.4823, "step": 13032 }, { "epoch": 2.04, "grad_norm": 16.461578355081606, "learning_rate": 4.946545569308436e-06, "loss": 0.4603, "step": 13033 }, { "epoch": 2.04, "grad_norm": 24.697868419755885, "learning_rate": 4.94509041859932e-06, "loss": 0.4841, "step": 13034 }, { "epoch": 2.04, "grad_norm": 23.784272872751654, "learning_rate": 4.943635411648875e-06, "loss": 0.5846, "step": 13035 }, { "epoch": 2.04, "grad_norm": 23.470043085795815, "learning_rate": 4.942180548498485e-06, "loss": 0.4584, "step": 13036 }, { "epoch": 2.04, "grad_norm": 18.63525237817207, "learning_rate": 4.940725829189523e-06, "loss": 0.5578, "step": 13037 }, { "epoch": 2.04, "grad_norm": 21.616966067201265, "learning_rate": 4.939271253763354e-06, "loss": 0.4953, "step": 13038 }, { "epoch": 2.04, "grad_norm": 13.966022760732278, "learning_rate": 4.93781682226135e-06, "loss": 0.4928, "step": 13039 }, { "epoch": 2.04, "grad_norm": 18.258842047480677, "learning_rate": 4.936362534724874e-06, "loss": 0.4683, "step": 13040 }, { "epoch": 2.04, "grad_norm": 17.542794078296232, "learning_rate": 4.93490839119529e-06, "loss": 0.5421, "step": 13041 }, { "epoch": 2.04, "grad_norm": 16.147552542113733, "learning_rate": 4.933454391713948e-06, "loss": 0.5064, "step": 13042 }, { "epoch": 2.04, "grad_norm": 20.626187050500658, "learning_rate": 4.932000536322196e-06, "loss": 0.4845, "step": 13043 }, { "epoch": 2.04, "grad_norm": 21.654846846411715, "learning_rate": 4.930546825061383e-06, "loss": 0.4447, "step": 13044 }, { "epoch": 2.04, "grad_norm": 19.889471016107183, "learning_rate": 4.929093257972857e-06, "loss": 0.425, "step": 13045 }, { "epoch": 2.04, "grad_norm": 19.998922219144394, "learning_rate": 4.927639835097948e-06, "loss": 0.617, "step": 13046 }, { "epoch": 2.04, "grad_norm": 43.946394668478995, "learning_rate": 4.926186556478e-06, "loss": 0.5313, "step": 13047 }, { "epoch": 2.04, "grad_norm": 17.354368619718787, "learning_rate": 4.924733422154333e-06, "loss": 0.4354, "step": 13048 }, { "epoch": 2.04, "grad_norm": 24.574407637122906, "learning_rate": 4.9232804321682835e-06, "loss": 0.6144, "step": 13049 }, { "epoch": 2.04, "grad_norm": 17.508368702453623, "learning_rate": 4.921827586561164e-06, "loss": 0.5333, "step": 13050 }, { "epoch": 2.04, "grad_norm": 17.70410935391592, "learning_rate": 4.9203748853742986e-06, "loss": 0.4878, "step": 13051 }, { "epoch": 2.04, "grad_norm": 15.846075923695057, "learning_rate": 4.918922328649003e-06, "loss": 0.5222, "step": 13052 }, { "epoch": 2.04, "grad_norm": 25.70642182532485, "learning_rate": 4.917469916426584e-06, "loss": 0.4936, "step": 13053 }, { "epoch": 2.04, "grad_norm": 29.422650817755528, "learning_rate": 4.916017648748342e-06, "loss": 0.4421, "step": 13054 }, { "epoch": 2.04, "grad_norm": 11.939127934634389, "learning_rate": 4.914565525655587e-06, "loss": 0.46, "step": 13055 }, { "epoch": 2.04, "grad_norm": 16.252299034270948, "learning_rate": 4.913113547189615e-06, "loss": 0.4824, "step": 13056 }, { "epoch": 2.04, "grad_norm": 13.09355082792873, "learning_rate": 4.911661713391716e-06, "loss": 0.4894, "step": 13057 }, { "epoch": 2.04, "grad_norm": 16.789493086642114, "learning_rate": 4.9102100243031835e-06, "loss": 0.4764, "step": 13058 }, { "epoch": 2.04, "grad_norm": 14.479533381129174, "learning_rate": 4.9087584799652976e-06, "loss": 0.4745, "step": 13059 }, { "epoch": 2.04, "grad_norm": 22.950148162202517, "learning_rate": 4.907307080419341e-06, "loss": 0.4781, "step": 13060 }, { "epoch": 2.04, "grad_norm": 21.722010297797063, "learning_rate": 4.905855825706597e-06, "loss": 0.4307, "step": 13061 }, { "epoch": 2.04, "grad_norm": 22.026886453767627, "learning_rate": 4.904404715868333e-06, "loss": 0.4888, "step": 13062 }, { "epoch": 2.04, "grad_norm": 19.4081582668416, "learning_rate": 4.902953750945814e-06, "loss": 0.5287, "step": 13063 }, { "epoch": 2.04, "grad_norm": 22.495950213857196, "learning_rate": 4.90150293098031e-06, "loss": 0.482, "step": 13064 }, { "epoch": 2.04, "grad_norm": 16.550060380763576, "learning_rate": 4.900052256013078e-06, "loss": 0.4732, "step": 13065 }, { "epoch": 2.04, "grad_norm": 24.514216969149448, "learning_rate": 4.898601726085385e-06, "loss": 0.5483, "step": 13066 }, { "epoch": 2.04, "grad_norm": 25.881271781050895, "learning_rate": 4.897151341238468e-06, "loss": 0.5618, "step": 13067 }, { "epoch": 2.04, "grad_norm": 19.2862616224728, "learning_rate": 4.89570110151358e-06, "loss": 0.5062, "step": 13068 }, { "epoch": 2.04, "grad_norm": 17.636930824770552, "learning_rate": 4.894251006951968e-06, "loss": 0.4818, "step": 13069 }, { "epoch": 2.04, "grad_norm": 39.210850895103675, "learning_rate": 4.892801057594874e-06, "loss": 0.5468, "step": 13070 }, { "epoch": 2.04, "grad_norm": 17.21575051696606, "learning_rate": 4.891351253483529e-06, "loss": 0.5224, "step": 13071 }, { "epoch": 2.04, "grad_norm": 12.857765260387817, "learning_rate": 4.8899015946591645e-06, "loss": 0.4473, "step": 13072 }, { "epoch": 2.04, "grad_norm": 21.41204916929831, "learning_rate": 4.888452081163007e-06, "loss": 0.6123, "step": 13073 }, { "epoch": 2.04, "grad_norm": 14.0952821628801, "learning_rate": 4.8870027130362885e-06, "loss": 0.4573, "step": 13074 }, { "epoch": 2.04, "grad_norm": 17.262400707975143, "learning_rate": 4.8855534903202165e-06, "loss": 0.523, "step": 13075 }, { "epoch": 2.04, "grad_norm": 27.61781785777043, "learning_rate": 4.884104413056015e-06, "loss": 0.5429, "step": 13076 }, { "epoch": 2.04, "grad_norm": 29.266838248685563, "learning_rate": 4.882655481284888e-06, "loss": 0.5581, "step": 13077 }, { "epoch": 2.04, "grad_norm": 19.540519602827015, "learning_rate": 4.881206695048046e-06, "loss": 0.5006, "step": 13078 }, { "epoch": 2.04, "grad_norm": 22.898472919000717, "learning_rate": 4.879758054386694e-06, "loss": 0.4967, "step": 13079 }, { "epoch": 2.04, "grad_norm": 18.996039434218446, "learning_rate": 4.8783095593420234e-06, "loss": 0.6114, "step": 13080 }, { "epoch": 2.04, "grad_norm": 24.872649070934497, "learning_rate": 4.876861209955238e-06, "loss": 0.4357, "step": 13081 }, { "epoch": 2.04, "grad_norm": 18.342807587975297, "learning_rate": 4.875413006267518e-06, "loss": 0.4837, "step": 13082 }, { "epoch": 2.04, "grad_norm": 16.1471705149605, "learning_rate": 4.873964948320059e-06, "loss": 0.5326, "step": 13083 }, { "epoch": 2.04, "grad_norm": 16.54139311762064, "learning_rate": 4.8725170361540345e-06, "loss": 0.5116, "step": 13084 }, { "epoch": 2.04, "grad_norm": 22.049095065687027, "learning_rate": 4.871069269810626e-06, "loss": 0.4685, "step": 13085 }, { "epoch": 2.04, "grad_norm": 24.9775664480053, "learning_rate": 4.869621649331009e-06, "loss": 0.5759, "step": 13086 }, { "epoch": 2.04, "grad_norm": 19.462539014305552, "learning_rate": 4.868174174756353e-06, "loss": 0.5591, "step": 13087 }, { "epoch": 2.04, "grad_norm": 24.31731374249837, "learning_rate": 4.866726846127817e-06, "loss": 0.4638, "step": 13088 }, { "epoch": 2.04, "grad_norm": 24.251425048652923, "learning_rate": 4.865279663486567e-06, "loss": 0.5928, "step": 13089 }, { "epoch": 2.04, "grad_norm": 18.183029788047318, "learning_rate": 4.863832626873759e-06, "loss": 0.5242, "step": 13090 }, { "epoch": 2.04, "grad_norm": 24.660302493324785, "learning_rate": 4.862385736330555e-06, "loss": 0.5345, "step": 13091 }, { "epoch": 2.04, "grad_norm": 16.32070945925332, "learning_rate": 4.860938991898088e-06, "loss": 0.5238, "step": 13092 }, { "epoch": 2.05, "grad_norm": 13.224183372765292, "learning_rate": 4.859492393617509e-06, "loss": 0.4141, "step": 13093 }, { "epoch": 2.05, "grad_norm": 19.471059750200975, "learning_rate": 4.8580459415299585e-06, "loss": 0.6052, "step": 13094 }, { "epoch": 2.05, "grad_norm": 19.971992064574472, "learning_rate": 4.856599635676578e-06, "loss": 0.5123, "step": 13095 }, { "epoch": 2.05, "grad_norm": 11.373945844321241, "learning_rate": 4.8551534760984954e-06, "loss": 0.4766, "step": 13096 }, { "epoch": 2.05, "grad_norm": 19.025852800447048, "learning_rate": 4.853707462836834e-06, "loss": 0.4895, "step": 13097 }, { "epoch": 2.05, "grad_norm": 14.710918251264056, "learning_rate": 4.852261595932721e-06, "loss": 0.5298, "step": 13098 }, { "epoch": 2.05, "grad_norm": 23.674850255173887, "learning_rate": 4.850815875427283e-06, "loss": 0.4622, "step": 13099 }, { "epoch": 2.05, "grad_norm": 23.87107094045906, "learning_rate": 4.849370301361623e-06, "loss": 0.5502, "step": 13100 }, { "epoch": 2.05, "grad_norm": 13.496457546868037, "learning_rate": 4.8479248737768644e-06, "loss": 0.4528, "step": 13101 }, { "epoch": 2.05, "grad_norm": 18.58280930364715, "learning_rate": 4.846479592714104e-06, "loss": 0.4763, "step": 13102 }, { "epoch": 2.05, "grad_norm": 20.63788096808724, "learning_rate": 4.845034458214449e-06, "loss": 0.4793, "step": 13103 }, { "epoch": 2.05, "grad_norm": 15.55182021770802, "learning_rate": 4.8435894703190026e-06, "loss": 0.4668, "step": 13104 }, { "epoch": 2.05, "grad_norm": 26.505994733864373, "learning_rate": 4.842144629068854e-06, "loss": 0.476, "step": 13105 }, { "epoch": 2.05, "grad_norm": 26.48734237116329, "learning_rate": 4.8406999345050934e-06, "loss": 0.5079, "step": 13106 }, { "epoch": 2.05, "grad_norm": 40.49507249221655, "learning_rate": 4.839255386668806e-06, "loss": 0.4895, "step": 13107 }, { "epoch": 2.05, "grad_norm": 23.317551708540375, "learning_rate": 4.8378109856010825e-06, "loss": 0.5622, "step": 13108 }, { "epoch": 2.05, "grad_norm": 18.86885238236193, "learning_rate": 4.8363667313429896e-06, "loss": 0.5886, "step": 13109 }, { "epoch": 2.05, "grad_norm": 16.538093366415712, "learning_rate": 4.83492262393561e-06, "loss": 0.4805, "step": 13110 }, { "epoch": 2.05, "grad_norm": 15.875291456058891, "learning_rate": 4.833478663420007e-06, "loss": 0.5045, "step": 13111 }, { "epoch": 2.05, "grad_norm": 26.29727931910288, "learning_rate": 4.832034849837252e-06, "loss": 0.5028, "step": 13112 }, { "epoch": 2.05, "grad_norm": 12.391570617988279, "learning_rate": 4.830591183228398e-06, "loss": 0.5129, "step": 13113 }, { "epoch": 2.05, "grad_norm": 14.781565945849648, "learning_rate": 4.829147663634507e-06, "loss": 0.475, "step": 13114 }, { "epoch": 2.05, "grad_norm": 21.271750278507955, "learning_rate": 4.827704291096635e-06, "loss": 0.5685, "step": 13115 }, { "epoch": 2.05, "grad_norm": 26.570186180615327, "learning_rate": 4.826261065655828e-06, "loss": 0.5113, "step": 13116 }, { "epoch": 2.05, "grad_norm": 24.171336103891505, "learning_rate": 4.824817987353126e-06, "loss": 0.4951, "step": 13117 }, { "epoch": 2.05, "grad_norm": 25.022904260856897, "learning_rate": 4.823375056229573e-06, "loss": 0.4439, "step": 13118 }, { "epoch": 2.05, "grad_norm": 18.6026222659298, "learning_rate": 4.821932272326204e-06, "loss": 0.445, "step": 13119 }, { "epoch": 2.05, "grad_norm": 21.25628924003059, "learning_rate": 4.820489635684057e-06, "loss": 0.5066, "step": 13120 }, { "epoch": 2.05, "grad_norm": 19.04497601537879, "learning_rate": 4.819047146344154e-06, "loss": 0.4717, "step": 13121 }, { "epoch": 2.05, "grad_norm": 26.909828417992323, "learning_rate": 4.817604804347517e-06, "loss": 0.4742, "step": 13122 }, { "epoch": 2.05, "grad_norm": 17.844475594447097, "learning_rate": 4.816162609735167e-06, "loss": 0.5907, "step": 13123 }, { "epoch": 2.05, "grad_norm": 28.258651678751484, "learning_rate": 4.814720562548123e-06, "loss": 0.5051, "step": 13124 }, { "epoch": 2.05, "grad_norm": 29.189905794985034, "learning_rate": 4.8132786628273945e-06, "loss": 0.4528, "step": 13125 }, { "epoch": 2.05, "grad_norm": 15.55679207420983, "learning_rate": 4.811836910613982e-06, "loss": 0.4475, "step": 13126 }, { "epoch": 2.05, "grad_norm": 19.25663783067621, "learning_rate": 4.810395305948892e-06, "loss": 0.4709, "step": 13127 }, { "epoch": 2.05, "grad_norm": 26.461609445830266, "learning_rate": 4.8089538488731244e-06, "loss": 0.4838, "step": 13128 }, { "epoch": 2.05, "grad_norm": 16.427072531874575, "learning_rate": 4.807512539427677e-06, "loss": 0.514, "step": 13129 }, { "epoch": 2.05, "grad_norm": 18.82424251842245, "learning_rate": 4.806071377653534e-06, "loss": 0.4567, "step": 13130 }, { "epoch": 2.05, "grad_norm": 20.610791011663075, "learning_rate": 4.80463036359168e-06, "loss": 0.5322, "step": 13131 }, { "epoch": 2.05, "grad_norm": 23.69776457447809, "learning_rate": 4.803189497283099e-06, "loss": 0.4828, "step": 13132 }, { "epoch": 2.05, "grad_norm": 15.26149548849849, "learning_rate": 4.801748778768772e-06, "loss": 0.5067, "step": 13133 }, { "epoch": 2.05, "grad_norm": 19.147832899069844, "learning_rate": 4.800308208089666e-06, "loss": 0.4711, "step": 13134 }, { "epoch": 2.05, "grad_norm": 16.39570157316439, "learning_rate": 4.798867785286756e-06, "loss": 0.4451, "step": 13135 }, { "epoch": 2.05, "grad_norm": 23.326819546857475, "learning_rate": 4.797427510401001e-06, "loss": 0.4599, "step": 13136 }, { "epoch": 2.05, "grad_norm": 23.71725273558126, "learning_rate": 4.795987383473366e-06, "loss": 0.53, "step": 13137 }, { "epoch": 2.05, "grad_norm": 18.39146658252815, "learning_rate": 4.794547404544802e-06, "loss": 0.5757, "step": 13138 }, { "epoch": 2.05, "grad_norm": 18.27019751220681, "learning_rate": 4.793107573656266e-06, "loss": 0.4434, "step": 13139 }, { "epoch": 2.05, "grad_norm": 18.784911968176218, "learning_rate": 4.791667890848708e-06, "loss": 0.5118, "step": 13140 }, { "epoch": 2.05, "grad_norm": 16.650569769826888, "learning_rate": 4.790228356163065e-06, "loss": 0.496, "step": 13141 }, { "epoch": 2.05, "grad_norm": 26.366631113469268, "learning_rate": 4.788788969640283e-06, "loss": 0.5115, "step": 13142 }, { "epoch": 2.05, "grad_norm": 18.826351075873383, "learning_rate": 4.787349731321291e-06, "loss": 0.4305, "step": 13143 }, { "epoch": 2.05, "grad_norm": 19.161040994357972, "learning_rate": 4.785910641247022e-06, "loss": 0.5053, "step": 13144 }, { "epoch": 2.05, "grad_norm": 17.033952075636236, "learning_rate": 4.78447169945841e-06, "loss": 0.511, "step": 13145 }, { "epoch": 2.05, "grad_norm": 18.50736386909542, "learning_rate": 4.78303290599637e-06, "loss": 0.5146, "step": 13146 }, { "epoch": 2.05, "grad_norm": 29.35894583936064, "learning_rate": 4.78159426090182e-06, "loss": 0.5213, "step": 13147 }, { "epoch": 2.05, "grad_norm": 37.52666531820444, "learning_rate": 4.780155764215676e-06, "loss": 0.4146, "step": 13148 }, { "epoch": 2.05, "grad_norm": 24.906601628725465, "learning_rate": 4.7787174159788514e-06, "loss": 0.4949, "step": 13149 }, { "epoch": 2.05, "grad_norm": 19.291174776819464, "learning_rate": 4.7772792162322505e-06, "loss": 0.4918, "step": 13150 }, { "epoch": 2.05, "grad_norm": 20.13246260432585, "learning_rate": 4.775841165016769e-06, "loss": 0.4286, "step": 13151 }, { "epoch": 2.05, "grad_norm": 18.380334995204837, "learning_rate": 4.774403262373308e-06, "loss": 0.4636, "step": 13152 }, { "epoch": 2.05, "grad_norm": 24.10448252085768, "learning_rate": 4.772965508342763e-06, "loss": 0.4568, "step": 13153 }, { "epoch": 2.05, "grad_norm": 25.580935339550056, "learning_rate": 4.771527902966028e-06, "loss": 0.4981, "step": 13154 }, { "epoch": 2.05, "grad_norm": 23.631121422976026, "learning_rate": 4.7700904462839735e-06, "loss": 0.4566, "step": 13155 }, { "epoch": 2.05, "grad_norm": 21.537421263248078, "learning_rate": 4.7686531383374865e-06, "loss": 0.5223, "step": 13156 }, { "epoch": 2.06, "grad_norm": 24.658410671153835, "learning_rate": 4.767215979167445e-06, "loss": 0.4918, "step": 13157 }, { "epoch": 2.06, "grad_norm": 15.134114788897476, "learning_rate": 4.765778968814724e-06, "loss": 0.4591, "step": 13158 }, { "epoch": 2.06, "grad_norm": 25.363092060755346, "learning_rate": 4.764342107320186e-06, "loss": 0.5202, "step": 13159 }, { "epoch": 2.06, "grad_norm": 22.8174915006642, "learning_rate": 4.762905394724691e-06, "loss": 0.4964, "step": 13160 }, { "epoch": 2.06, "grad_norm": 29.040706876349738, "learning_rate": 4.761468831069106e-06, "loss": 0.5863, "step": 13161 }, { "epoch": 2.06, "grad_norm": 11.987952909326758, "learning_rate": 4.7600324163942845e-06, "loss": 0.4474, "step": 13162 }, { "epoch": 2.06, "grad_norm": 24.6635138154888, "learning_rate": 4.758596150741073e-06, "loss": 0.4867, "step": 13163 }, { "epoch": 2.06, "grad_norm": 13.164974662691726, "learning_rate": 4.757160034150324e-06, "loss": 0.4996, "step": 13164 }, { "epoch": 2.06, "grad_norm": 28.721278538236707, "learning_rate": 4.7557240666628735e-06, "loss": 0.5115, "step": 13165 }, { "epoch": 2.06, "grad_norm": 22.04895459389824, "learning_rate": 4.754288248319563e-06, "loss": 0.4863, "step": 13166 }, { "epoch": 2.06, "grad_norm": 23.734362158354347, "learning_rate": 4.752852579161229e-06, "loss": 0.5033, "step": 13167 }, { "epoch": 2.06, "grad_norm": 14.837193166256506, "learning_rate": 4.751417059228695e-06, "loss": 0.5426, "step": 13168 }, { "epoch": 2.06, "grad_norm": 15.734374005602717, "learning_rate": 4.749981688562792e-06, "loss": 0.4455, "step": 13169 }, { "epoch": 2.06, "grad_norm": 19.256447441515185, "learning_rate": 4.748546467204336e-06, "loss": 0.4139, "step": 13170 }, { "epoch": 2.06, "grad_norm": 15.689209340064366, "learning_rate": 4.747111395194149e-06, "loss": 0.472, "step": 13171 }, { "epoch": 2.06, "grad_norm": 17.07914701686187, "learning_rate": 4.7456764725730385e-06, "loss": 0.474, "step": 13172 }, { "epoch": 2.06, "grad_norm": 13.888771738598066, "learning_rate": 4.744241699381816e-06, "loss": 0.4863, "step": 13173 }, { "epoch": 2.06, "grad_norm": 19.261870969048253, "learning_rate": 4.742807075661288e-06, "loss": 0.5512, "step": 13174 }, { "epoch": 2.06, "grad_norm": 17.238974395861504, "learning_rate": 4.741372601452251e-06, "loss": 0.52, "step": 13175 }, { "epoch": 2.06, "grad_norm": 19.092522108737782, "learning_rate": 4.739938276795498e-06, "loss": 0.5036, "step": 13176 }, { "epoch": 2.06, "grad_norm": 14.593687492845298, "learning_rate": 4.738504101731824e-06, "loss": 0.4155, "step": 13177 }, { "epoch": 2.06, "grad_norm": 21.249497735875668, "learning_rate": 4.737070076302014e-06, "loss": 0.4787, "step": 13178 }, { "epoch": 2.06, "grad_norm": 17.961678617833368, "learning_rate": 4.735636200546859e-06, "loss": 0.4691, "step": 13179 }, { "epoch": 2.06, "grad_norm": 13.321922905751014, "learning_rate": 4.734202474507126e-06, "loss": 0.5002, "step": 13180 }, { "epoch": 2.06, "grad_norm": 28.150140581110264, "learning_rate": 4.7327688982235916e-06, "loss": 0.4799, "step": 13181 }, { "epoch": 2.06, "grad_norm": 25.732883946361678, "learning_rate": 4.731335471737031e-06, "loss": 0.525, "step": 13182 }, { "epoch": 2.06, "grad_norm": 17.31312165323388, "learning_rate": 4.72990219508821e-06, "loss": 0.5178, "step": 13183 }, { "epoch": 2.06, "grad_norm": 27.750640312960908, "learning_rate": 4.728469068317887e-06, "loss": 0.5771, "step": 13184 }, { "epoch": 2.06, "grad_norm": 20.267775892763424, "learning_rate": 4.727036091466816e-06, "loss": 0.5371, "step": 13185 }, { "epoch": 2.06, "grad_norm": 19.92896450978873, "learning_rate": 4.7256032645757555e-06, "loss": 0.4909, "step": 13186 }, { "epoch": 2.06, "grad_norm": 19.03684667372953, "learning_rate": 4.724170587685456e-06, "loss": 0.496, "step": 13187 }, { "epoch": 2.06, "grad_norm": 25.026947929011325, "learning_rate": 4.722738060836654e-06, "loss": 0.6308, "step": 13188 }, { "epoch": 2.06, "grad_norm": 17.34865867297164, "learning_rate": 4.7213056840701e-06, "loss": 0.4742, "step": 13189 }, { "epoch": 2.06, "grad_norm": 25.45888529978834, "learning_rate": 4.719873457426519e-06, "loss": 0.5104, "step": 13190 }, { "epoch": 2.06, "grad_norm": 15.4916437604768, "learning_rate": 4.718441380946649e-06, "loss": 0.448, "step": 13191 }, { "epoch": 2.06, "grad_norm": 16.640372947789828, "learning_rate": 4.717009454671221e-06, "loss": 0.4748, "step": 13192 }, { "epoch": 2.06, "grad_norm": 17.926840234874444, "learning_rate": 4.715577678640948e-06, "loss": 0.5192, "step": 13193 }, { "epoch": 2.06, "grad_norm": 19.48118511662771, "learning_rate": 4.714146052896559e-06, "loss": 0.4897, "step": 13194 }, { "epoch": 2.06, "grad_norm": 16.640480900530083, "learning_rate": 4.7127145774787604e-06, "loss": 0.5584, "step": 13195 }, { "epoch": 2.06, "grad_norm": 31.634562008018673, "learning_rate": 4.71128325242827e-06, "loss": 0.5316, "step": 13196 }, { "epoch": 2.06, "grad_norm": 15.578033159314309, "learning_rate": 4.709852077785787e-06, "loss": 0.5237, "step": 13197 }, { "epoch": 2.06, "grad_norm": 25.785091844179796, "learning_rate": 4.708421053592019e-06, "loss": 0.4978, "step": 13198 }, { "epoch": 2.06, "grad_norm": 22.001608689512103, "learning_rate": 4.706990179887658e-06, "loss": 0.4839, "step": 13199 }, { "epoch": 2.06, "grad_norm": 16.638333979575602, "learning_rate": 4.705559456713403e-06, "loss": 0.456, "step": 13200 }, { "epoch": 2.06, "grad_norm": 23.374595600342712, "learning_rate": 4.704128884109936e-06, "loss": 0.5992, "step": 13201 }, { "epoch": 2.06, "grad_norm": 16.259441866014114, "learning_rate": 4.702698462117946e-06, "loss": 0.5045, "step": 13202 }, { "epoch": 2.06, "grad_norm": 15.117531366897676, "learning_rate": 4.701268190778117e-06, "loss": 0.4629, "step": 13203 }, { "epoch": 2.06, "grad_norm": 16.42074207069252, "learning_rate": 4.6998380701311196e-06, "loss": 0.4112, "step": 13204 }, { "epoch": 2.06, "grad_norm": 14.662762698111324, "learning_rate": 4.6984081002176256e-06, "loss": 0.484, "step": 13205 }, { "epoch": 2.06, "grad_norm": 19.72600408840137, "learning_rate": 4.696978281078302e-06, "loss": 0.4784, "step": 13206 }, { "epoch": 2.06, "grad_norm": 14.682652369952763, "learning_rate": 4.695548612753814e-06, "loss": 0.5063, "step": 13207 }, { "epoch": 2.06, "grad_norm": 20.305321777816935, "learning_rate": 4.694119095284825e-06, "loss": 0.5034, "step": 13208 }, { "epoch": 2.06, "grad_norm": 21.726412949223977, "learning_rate": 4.6926897287119845e-06, "loss": 0.5166, "step": 13209 }, { "epoch": 2.06, "grad_norm": 16.03169614977865, "learning_rate": 4.6912605130759396e-06, "loss": 0.4976, "step": 13210 }, { "epoch": 2.06, "grad_norm": 22.463874376767897, "learning_rate": 4.68983144841734e-06, "loss": 0.4748, "step": 13211 }, { "epoch": 2.06, "grad_norm": 18.00008583131168, "learning_rate": 4.688402534776832e-06, "loss": 0.5078, "step": 13212 }, { "epoch": 2.06, "grad_norm": 14.768752517691908, "learning_rate": 4.686973772195048e-06, "loss": 0.4953, "step": 13213 }, { "epoch": 2.06, "grad_norm": 16.489240478831842, "learning_rate": 4.685545160712619e-06, "loss": 0.4815, "step": 13214 }, { "epoch": 2.06, "grad_norm": 22.387293940516482, "learning_rate": 4.684116700370176e-06, "loss": 0.4804, "step": 13215 }, { "epoch": 2.06, "grad_norm": 17.5873130809419, "learning_rate": 4.682688391208345e-06, "loss": 0.4599, "step": 13216 }, { "epoch": 2.06, "grad_norm": 16.849759722572045, "learning_rate": 4.681260233267749e-06, "loss": 0.4729, "step": 13217 }, { "epoch": 2.06, "grad_norm": 32.49318304002005, "learning_rate": 4.679832226589001e-06, "loss": 0.5261, "step": 13218 }, { "epoch": 2.06, "grad_norm": 12.763241618638316, "learning_rate": 4.6784043712127084e-06, "loss": 0.4391, "step": 13219 }, { "epoch": 2.06, "grad_norm": 21.010813152628668, "learning_rate": 4.676976667179482e-06, "loss": 0.4879, "step": 13220 }, { "epoch": 2.07, "grad_norm": 14.51285047854393, "learning_rate": 4.675549114529929e-06, "loss": 0.4461, "step": 13221 }, { "epoch": 2.07, "grad_norm": 19.272838068293154, "learning_rate": 4.6741217133046415e-06, "loss": 0.4978, "step": 13222 }, { "epoch": 2.07, "grad_norm": 19.640360671110585, "learning_rate": 4.67269446354422e-06, "loss": 0.4897, "step": 13223 }, { "epoch": 2.07, "grad_norm": 24.475874627240326, "learning_rate": 4.671267365289247e-06, "loss": 0.5642, "step": 13224 }, { "epoch": 2.07, "grad_norm": 15.514364817556473, "learning_rate": 4.669840418580318e-06, "loss": 0.5352, "step": 13225 }, { "epoch": 2.07, "grad_norm": 15.470593647954168, "learning_rate": 4.668413623458006e-06, "loss": 0.5017, "step": 13226 }, { "epoch": 2.07, "grad_norm": 18.7617868985732, "learning_rate": 4.666986979962891e-06, "loss": 0.4471, "step": 13227 }, { "epoch": 2.07, "grad_norm": 23.45572303852927, "learning_rate": 4.6655604881355495e-06, "loss": 0.5002, "step": 13228 }, { "epoch": 2.07, "grad_norm": 20.608994056732502, "learning_rate": 4.664134148016545e-06, "loss": 0.5192, "step": 13229 }, { "epoch": 2.07, "grad_norm": 42.26989431695043, "learning_rate": 4.662707959646446e-06, "loss": 0.53, "step": 13230 }, { "epoch": 2.07, "grad_norm": 17.433607868562383, "learning_rate": 4.661281923065808e-06, "loss": 0.5339, "step": 13231 }, { "epoch": 2.07, "grad_norm": 24.983701504211606, "learning_rate": 4.659856038315188e-06, "loss": 0.5411, "step": 13232 }, { "epoch": 2.07, "grad_norm": 17.651993867294077, "learning_rate": 4.658430305435143e-06, "loss": 0.5539, "step": 13233 }, { "epoch": 2.07, "grad_norm": 21.453543533221193, "learning_rate": 4.657004724466216e-06, "loss": 0.4818, "step": 13234 }, { "epoch": 2.07, "grad_norm": 20.137962185257834, "learning_rate": 4.655579295448944e-06, "loss": 0.4841, "step": 13235 }, { "epoch": 2.07, "grad_norm": 16.48465977418263, "learning_rate": 4.654154018423871e-06, "loss": 0.4693, "step": 13236 }, { "epoch": 2.07, "grad_norm": 16.99336340332264, "learning_rate": 4.652728893431534e-06, "loss": 0.5159, "step": 13237 }, { "epoch": 2.07, "grad_norm": 16.420228762644086, "learning_rate": 4.651303920512459e-06, "loss": 0.5163, "step": 13238 }, { "epoch": 2.07, "grad_norm": 27.681051048955716, "learning_rate": 4.649879099707168e-06, "loss": 0.4894, "step": 13239 }, { "epoch": 2.07, "grad_norm": 15.828555037257193, "learning_rate": 4.648454431056186e-06, "loss": 0.5398, "step": 13240 }, { "epoch": 2.07, "grad_norm": 19.027205134078546, "learning_rate": 4.64702991460003e-06, "loss": 0.4548, "step": 13241 }, { "epoch": 2.07, "grad_norm": 26.730373530683064, "learning_rate": 4.645605550379214e-06, "loss": 0.5444, "step": 13242 }, { "epoch": 2.07, "grad_norm": 24.670789631336856, "learning_rate": 4.644181338434245e-06, "loss": 0.4692, "step": 13243 }, { "epoch": 2.07, "grad_norm": 16.388460423661712, "learning_rate": 4.642757278805622e-06, "loss": 0.4848, "step": 13244 }, { "epoch": 2.07, "grad_norm": 17.01839281634388, "learning_rate": 4.641333371533848e-06, "loss": 0.4741, "step": 13245 }, { "epoch": 2.07, "grad_norm": 18.859873675000927, "learning_rate": 4.6399096166594215e-06, "loss": 0.565, "step": 13246 }, { "epoch": 2.07, "grad_norm": 20.440530171019105, "learning_rate": 4.638486014222831e-06, "loss": 0.558, "step": 13247 }, { "epoch": 2.07, "grad_norm": 22.89752531235786, "learning_rate": 4.6370625642645565e-06, "loss": 0.5609, "step": 13248 }, { "epoch": 2.07, "grad_norm": 22.214453554088877, "learning_rate": 4.635639266825086e-06, "loss": 0.4755, "step": 13249 }, { "epoch": 2.07, "grad_norm": 17.558230368107555, "learning_rate": 4.634216121944901e-06, "loss": 0.4882, "step": 13250 }, { "epoch": 2.07, "grad_norm": 18.937022105566925, "learning_rate": 4.632793129664466e-06, "loss": 0.5074, "step": 13251 }, { "epoch": 2.07, "grad_norm": 12.170384326781141, "learning_rate": 4.63137029002426e-06, "loss": 0.4778, "step": 13252 }, { "epoch": 2.07, "grad_norm": 16.679424250309843, "learning_rate": 4.629947603064737e-06, "loss": 0.4871, "step": 13253 }, { "epoch": 2.07, "grad_norm": 18.605160811761372, "learning_rate": 4.6285250688263625e-06, "loss": 0.478, "step": 13254 }, { "epoch": 2.07, "grad_norm": 19.4407472684658, "learning_rate": 4.627102687349599e-06, "loss": 0.4601, "step": 13255 }, { "epoch": 2.07, "grad_norm": 22.16531912004903, "learning_rate": 4.625680458674886e-06, "loss": 0.4667, "step": 13256 }, { "epoch": 2.07, "grad_norm": 16.87519743062133, "learning_rate": 4.624258382842681e-06, "loss": 0.5047, "step": 13257 }, { "epoch": 2.07, "grad_norm": 19.26658181477529, "learning_rate": 4.622836459893419e-06, "loss": 0.5049, "step": 13258 }, { "epoch": 2.07, "grad_norm": 21.944766816321177, "learning_rate": 4.621414689867547e-06, "loss": 0.5219, "step": 13259 }, { "epoch": 2.07, "grad_norm": 18.991635752752213, "learning_rate": 4.619993072805491e-06, "loss": 0.4795, "step": 13260 }, { "epoch": 2.07, "grad_norm": 21.84476442191785, "learning_rate": 4.618571608747685e-06, "loss": 0.5565, "step": 13261 }, { "epoch": 2.07, "grad_norm": 21.35563997458557, "learning_rate": 4.617150297734557e-06, "loss": 0.5183, "step": 13262 }, { "epoch": 2.07, "grad_norm": 24.05433216185614, "learning_rate": 4.615729139806527e-06, "loss": 0.5606, "step": 13263 }, { "epoch": 2.07, "grad_norm": 17.802504232805475, "learning_rate": 4.614308135004006e-06, "loss": 0.4718, "step": 13264 }, { "epoch": 2.07, "grad_norm": 23.50101327740678, "learning_rate": 4.612887283367411e-06, "loss": 0.5271, "step": 13265 }, { "epoch": 2.07, "grad_norm": 17.01612142358905, "learning_rate": 4.611466584937151e-06, "loss": 0.515, "step": 13266 }, { "epoch": 2.07, "grad_norm": 22.621576061979578, "learning_rate": 4.610046039753636e-06, "loss": 0.5371, "step": 13267 }, { "epoch": 2.07, "grad_norm": 17.89313635461756, "learning_rate": 4.608625647857251e-06, "loss": 0.4967, "step": 13268 }, { "epoch": 2.07, "grad_norm": 17.527823419300745, "learning_rate": 4.6072054092884e-06, "loss": 0.5123, "step": 13269 }, { "epoch": 2.07, "grad_norm": 24.466714567903377, "learning_rate": 4.60578532408747e-06, "loss": 0.4647, "step": 13270 }, { "epoch": 2.07, "grad_norm": 28.490170040516933, "learning_rate": 4.604365392294856e-06, "loss": 0.5206, "step": 13271 }, { "epoch": 2.07, "grad_norm": 24.953122054179286, "learning_rate": 4.602945613950933e-06, "loss": 0.5469, "step": 13272 }, { "epoch": 2.07, "grad_norm": 23.444795847522744, "learning_rate": 4.6015259890960765e-06, "loss": 0.4765, "step": 13273 }, { "epoch": 2.07, "grad_norm": 33.2344108369721, "learning_rate": 4.600106517770662e-06, "loss": 0.5262, "step": 13274 }, { "epoch": 2.07, "grad_norm": 25.078606215283795, "learning_rate": 4.598687200015063e-06, "loss": 0.5199, "step": 13275 }, { "epoch": 2.07, "grad_norm": 12.847109196383375, "learning_rate": 4.597268035869636e-06, "loss": 0.4408, "step": 13276 }, { "epoch": 2.07, "grad_norm": 17.273272357843112, "learning_rate": 4.59584902537475e-06, "loss": 0.48, "step": 13277 }, { "epoch": 2.07, "grad_norm": 20.891020378788014, "learning_rate": 4.594430168570753e-06, "loss": 0.5481, "step": 13278 }, { "epoch": 2.07, "grad_norm": 35.46058444922919, "learning_rate": 4.593011465498e-06, "loss": 0.5408, "step": 13279 }, { "epoch": 2.07, "grad_norm": 18.44592376198636, "learning_rate": 4.591592916196841e-06, "loss": 0.5321, "step": 13280 }, { "epoch": 2.07, "grad_norm": 19.537570353016452, "learning_rate": 4.590174520707612e-06, "loss": 0.4878, "step": 13281 }, { "epoch": 2.07, "grad_norm": 17.821963766807723, "learning_rate": 4.58875627907066e-06, "loss": 0.4343, "step": 13282 }, { "epoch": 2.07, "grad_norm": 18.44724914384452, "learning_rate": 4.587338191326308e-06, "loss": 0.449, "step": 13283 }, { "epoch": 2.07, "grad_norm": 30.103092306209025, "learning_rate": 4.585920257514897e-06, "loss": 0.4938, "step": 13284 }, { "epoch": 2.08, "grad_norm": 20.819799432641574, "learning_rate": 4.584502477676742e-06, "loss": 0.5244, "step": 13285 }, { "epoch": 2.08, "grad_norm": 20.27474386383267, "learning_rate": 4.583084851852169e-06, "loss": 0.5628, "step": 13286 }, { "epoch": 2.08, "grad_norm": 22.788345147770126, "learning_rate": 4.581667380081497e-06, "loss": 0.4706, "step": 13287 }, { "epoch": 2.08, "grad_norm": 20.6980046119383, "learning_rate": 4.580250062405036e-06, "loss": 0.4697, "step": 13288 }, { "epoch": 2.08, "grad_norm": 17.228160775360433, "learning_rate": 4.578832898863088e-06, "loss": 0.4847, "step": 13289 }, { "epoch": 2.08, "grad_norm": 32.10342693043912, "learning_rate": 4.577415889495962e-06, "loss": 0.5837, "step": 13290 }, { "epoch": 2.08, "grad_norm": 20.98573331491425, "learning_rate": 4.575999034343957e-06, "loss": 0.5588, "step": 13291 }, { "epoch": 2.08, "grad_norm": 24.260059327578197, "learning_rate": 4.5745823334473685e-06, "loss": 0.4674, "step": 13292 }, { "epoch": 2.08, "grad_norm": 20.381798869962854, "learning_rate": 4.573165786846485e-06, "loss": 0.5284, "step": 13293 }, { "epoch": 2.08, "grad_norm": 24.528074104199415, "learning_rate": 4.571749394581588e-06, "loss": 0.5152, "step": 13294 }, { "epoch": 2.08, "grad_norm": 19.18882387638936, "learning_rate": 4.570333156692963e-06, "loss": 0.4948, "step": 13295 }, { "epoch": 2.08, "grad_norm": 19.183718532373142, "learning_rate": 4.568917073220891e-06, "loss": 0.4781, "step": 13296 }, { "epoch": 2.08, "grad_norm": 26.013057232625478, "learning_rate": 4.567501144205639e-06, "loss": 0.5371, "step": 13297 }, { "epoch": 2.08, "grad_norm": 16.105198645333353, "learning_rate": 4.566085369687475e-06, "loss": 0.4781, "step": 13298 }, { "epoch": 2.08, "grad_norm": 20.81561349514247, "learning_rate": 4.564669749706663e-06, "loss": 0.4589, "step": 13299 }, { "epoch": 2.08, "grad_norm": 21.75364095533292, "learning_rate": 4.563254284303468e-06, "loss": 0.5483, "step": 13300 }, { "epoch": 2.08, "grad_norm": 21.414261980529798, "learning_rate": 4.56183897351814e-06, "loss": 0.4292, "step": 13301 }, { "epoch": 2.08, "grad_norm": 28.169139517149215, "learning_rate": 4.560423817390927e-06, "loss": 0.5308, "step": 13302 }, { "epoch": 2.08, "grad_norm": 16.864232061451382, "learning_rate": 4.559008815962078e-06, "loss": 0.4324, "step": 13303 }, { "epoch": 2.08, "grad_norm": 21.460231120868325, "learning_rate": 4.557593969271836e-06, "loss": 0.528, "step": 13304 }, { "epoch": 2.08, "grad_norm": 20.877083541629784, "learning_rate": 4.556179277360442e-06, "loss": 0.4705, "step": 13305 }, { "epoch": 2.08, "grad_norm": 28.45493918474053, "learning_rate": 4.554764740268124e-06, "loss": 0.525, "step": 13306 }, { "epoch": 2.08, "grad_norm": 30.070636441276545, "learning_rate": 4.553350358035108e-06, "loss": 0.5167, "step": 13307 }, { "epoch": 2.08, "grad_norm": 14.855782165752, "learning_rate": 4.551936130701622e-06, "loss": 0.4943, "step": 13308 }, { "epoch": 2.08, "grad_norm": 18.96931533348104, "learning_rate": 4.550522058307888e-06, "loss": 0.4773, "step": 13309 }, { "epoch": 2.08, "grad_norm": 16.585443390966994, "learning_rate": 4.549108140894117e-06, "loss": 0.473, "step": 13310 }, { "epoch": 2.08, "grad_norm": 16.401202400835626, "learning_rate": 4.547694378500523e-06, "loss": 0.4329, "step": 13311 }, { "epoch": 2.08, "grad_norm": 19.704926195938093, "learning_rate": 4.54628077116731e-06, "loss": 0.5286, "step": 13312 }, { "epoch": 2.08, "grad_norm": 18.107338179007186, "learning_rate": 4.544867318934684e-06, "loss": 0.4678, "step": 13313 }, { "epoch": 2.08, "grad_norm": 16.743244162718618, "learning_rate": 4.543454021842836e-06, "loss": 0.4638, "step": 13314 }, { "epoch": 2.08, "grad_norm": 13.776271012553853, "learning_rate": 4.5420408799319646e-06, "loss": 0.45, "step": 13315 }, { "epoch": 2.08, "grad_norm": 23.885712340231894, "learning_rate": 4.540627893242261e-06, "loss": 0.4533, "step": 13316 }, { "epoch": 2.08, "grad_norm": 20.790081447185585, "learning_rate": 4.539215061813903e-06, "loss": 0.5177, "step": 13317 }, { "epoch": 2.08, "grad_norm": 35.4578125192979, "learning_rate": 4.537802385687078e-06, "loss": 0.4472, "step": 13318 }, { "epoch": 2.08, "grad_norm": 19.718204867464728, "learning_rate": 4.5363898649019545e-06, "loss": 0.4364, "step": 13319 }, { "epoch": 2.08, "grad_norm": 14.74725590503136, "learning_rate": 4.534977499498706e-06, "loss": 0.4313, "step": 13320 }, { "epoch": 2.08, "grad_norm": 23.19691321457073, "learning_rate": 4.533565289517506e-06, "loss": 0.5193, "step": 13321 }, { "epoch": 2.08, "grad_norm": 21.306957474357283, "learning_rate": 4.5321532349985095e-06, "loss": 0.4197, "step": 13322 }, { "epoch": 2.08, "grad_norm": 12.334155559786513, "learning_rate": 4.530741335981874e-06, "loss": 0.4571, "step": 13323 }, { "epoch": 2.08, "grad_norm": 13.120207059017911, "learning_rate": 4.529329592507755e-06, "loss": 0.4476, "step": 13324 }, { "epoch": 2.08, "grad_norm": 18.673845227092745, "learning_rate": 4.527918004616305e-06, "loss": 0.46, "step": 13325 }, { "epoch": 2.08, "grad_norm": 22.132544220666155, "learning_rate": 4.5265065723476675e-06, "loss": 0.5036, "step": 13326 }, { "epoch": 2.08, "grad_norm": 23.62823446652317, "learning_rate": 4.525095295741976e-06, "loss": 0.5255, "step": 13327 }, { "epoch": 2.08, "grad_norm": 28.649807073615964, "learning_rate": 4.523684174839372e-06, "loss": 0.5218, "step": 13328 }, { "epoch": 2.08, "grad_norm": 16.79047333020339, "learning_rate": 4.5222732096799875e-06, "loss": 0.431, "step": 13329 }, { "epoch": 2.08, "grad_norm": 19.665507084310978, "learning_rate": 4.520862400303951e-06, "loss": 0.5162, "step": 13330 }, { "epoch": 2.08, "grad_norm": 21.49819586497654, "learning_rate": 4.519451746751383e-06, "loss": 0.4882, "step": 13331 }, { "epoch": 2.08, "grad_norm": 4.160024828916969, "learning_rate": 4.518041249062398e-06, "loss": 0.5908, "step": 13332 }, { "epoch": 2.08, "grad_norm": 14.103632722671467, "learning_rate": 4.516630907277113e-06, "loss": 0.4805, "step": 13333 }, { "epoch": 2.08, "grad_norm": 23.395934853428344, "learning_rate": 4.5152207214356406e-06, "loss": 0.5737, "step": 13334 }, { "epoch": 2.08, "grad_norm": 18.24113983478634, "learning_rate": 4.513810691578079e-06, "loss": 0.5527, "step": 13335 }, { "epoch": 2.08, "grad_norm": 20.32820993767835, "learning_rate": 4.512400817744535e-06, "loss": 0.4805, "step": 13336 }, { "epoch": 2.08, "grad_norm": 20.00100803889626, "learning_rate": 4.510991099975098e-06, "loss": 0.4277, "step": 13337 }, { "epoch": 2.08, "grad_norm": 31.046224498989424, "learning_rate": 4.509581538309867e-06, "loss": 0.5508, "step": 13338 }, { "epoch": 2.08, "grad_norm": 17.51467379080833, "learning_rate": 4.508172132788923e-06, "loss": 0.5532, "step": 13339 }, { "epoch": 2.08, "grad_norm": 14.355440968077888, "learning_rate": 4.506762883452349e-06, "loss": 0.4422, "step": 13340 }, { "epoch": 2.08, "grad_norm": 17.804332656805983, "learning_rate": 4.50535379034023e-06, "loss": 0.4851, "step": 13341 }, { "epoch": 2.08, "grad_norm": 34.906183021177206, "learning_rate": 4.503944853492631e-06, "loss": 0.542, "step": 13342 }, { "epoch": 2.08, "grad_norm": 16.73444871325246, "learning_rate": 4.502536072949628e-06, "loss": 0.4982, "step": 13343 }, { "epoch": 2.08, "grad_norm": 28.978346040808887, "learning_rate": 4.5011274487512806e-06, "loss": 0.4867, "step": 13344 }, { "epoch": 2.08, "grad_norm": 20.602943925251328, "learning_rate": 4.4997189809376554e-06, "loss": 0.5291, "step": 13345 }, { "epoch": 2.08, "grad_norm": 14.135052178232254, "learning_rate": 4.498310669548801e-06, "loss": 0.4534, "step": 13346 }, { "epoch": 2.08, "grad_norm": 19.818277409403674, "learning_rate": 4.496902514624775e-06, "loss": 0.5213, "step": 13347 }, { "epoch": 2.08, "grad_norm": 19.446230919170148, "learning_rate": 4.495494516205621e-06, "loss": 0.5111, "step": 13348 }, { "epoch": 2.09, "grad_norm": 15.29595463617056, "learning_rate": 4.494086674331382e-06, "loss": 0.4961, "step": 13349 }, { "epoch": 2.09, "grad_norm": 13.931181608762111, "learning_rate": 4.492678989042099e-06, "loss": 0.4893, "step": 13350 }, { "epoch": 2.09, "grad_norm": 20.38331068840339, "learning_rate": 4.491271460377806e-06, "loss": 0.5184, "step": 13351 }, { "epoch": 2.09, "grad_norm": 21.977967451624647, "learning_rate": 4.489864088378526e-06, "loss": 0.4909, "step": 13352 }, { "epoch": 2.09, "grad_norm": 27.461327212784816, "learning_rate": 4.488456873084288e-06, "loss": 0.4533, "step": 13353 }, { "epoch": 2.09, "grad_norm": 30.113063848583863, "learning_rate": 4.487049814535112e-06, "loss": 0.4814, "step": 13354 }, { "epoch": 2.09, "grad_norm": 23.05653861450223, "learning_rate": 4.4856429127710224e-06, "loss": 0.5623, "step": 13355 }, { "epoch": 2.09, "grad_norm": 15.800967790895884, "learning_rate": 4.484236167832015e-06, "loss": 0.4503, "step": 13356 }, { "epoch": 2.09, "grad_norm": 15.436910201979742, "learning_rate": 4.482829579758103e-06, "loss": 0.4479, "step": 13357 }, { "epoch": 2.09, "grad_norm": 14.097455583234527, "learning_rate": 4.481423148589292e-06, "loss": 0.5083, "step": 13358 }, { "epoch": 2.09, "grad_norm": 19.07132455874364, "learning_rate": 4.4800168743655814e-06, "loss": 0.5065, "step": 13359 }, { "epoch": 2.09, "grad_norm": 18.15619025901775, "learning_rate": 4.478610757126962e-06, "loss": 0.4788, "step": 13360 }, { "epoch": 2.09, "grad_norm": 23.43294365272516, "learning_rate": 4.477204796913419e-06, "loss": 0.4948, "step": 13361 }, { "epoch": 2.09, "grad_norm": 22.072391561296246, "learning_rate": 4.475798993764941e-06, "loss": 0.4632, "step": 13362 }, { "epoch": 2.09, "grad_norm": 15.355250719725868, "learning_rate": 4.474393347721511e-06, "loss": 0.472, "step": 13363 }, { "epoch": 2.09, "grad_norm": 17.806147210482905, "learning_rate": 4.472987858823099e-06, "loss": 0.4732, "step": 13364 }, { "epoch": 2.09, "grad_norm": 30.2367927836394, "learning_rate": 4.471582527109683e-06, "loss": 0.5207, "step": 13365 }, { "epoch": 2.09, "grad_norm": 26.766143714746494, "learning_rate": 4.470177352621221e-06, "loss": 0.5428, "step": 13366 }, { "epoch": 2.09, "grad_norm": 13.279677872452869, "learning_rate": 4.468772335397681e-06, "loss": 0.5143, "step": 13367 }, { "epoch": 2.09, "grad_norm": 30.028609946729038, "learning_rate": 4.467367475479023e-06, "loss": 0.5886, "step": 13368 }, { "epoch": 2.09, "grad_norm": 22.703485283549373, "learning_rate": 4.465962772905195e-06, "loss": 0.47, "step": 13369 }, { "epoch": 2.09, "grad_norm": 26.02288981834509, "learning_rate": 4.464558227716152e-06, "loss": 0.4235, "step": 13370 }, { "epoch": 2.09, "grad_norm": 17.08937653998288, "learning_rate": 4.463153839951829e-06, "loss": 0.4421, "step": 13371 }, { "epoch": 2.09, "grad_norm": 25.071619293970567, "learning_rate": 4.461749609652179e-06, "loss": 0.5131, "step": 13372 }, { "epoch": 2.09, "grad_norm": 20.046640020481444, "learning_rate": 4.460345536857124e-06, "loss": 0.481, "step": 13373 }, { "epoch": 2.09, "grad_norm": 11.804403857693078, "learning_rate": 4.458941621606602e-06, "loss": 0.4692, "step": 13374 }, { "epoch": 2.09, "grad_norm": 28.887366923500668, "learning_rate": 4.457537863940543e-06, "loss": 0.4929, "step": 13375 }, { "epoch": 2.09, "grad_norm": 16.737847099771972, "learning_rate": 4.4561342638988645e-06, "loss": 0.4731, "step": 13376 }, { "epoch": 2.09, "grad_norm": 17.01579895841372, "learning_rate": 4.4547308215214815e-06, "loss": 0.4864, "step": 13377 }, { "epoch": 2.09, "grad_norm": 21.57507625521603, "learning_rate": 4.453327536848309e-06, "loss": 0.5053, "step": 13378 }, { "epoch": 2.09, "grad_norm": 14.295256455841429, "learning_rate": 4.451924409919257e-06, "loss": 0.4498, "step": 13379 }, { "epoch": 2.09, "grad_norm": 13.032001856742218, "learning_rate": 4.450521440774233e-06, "loss": 0.4372, "step": 13380 }, { "epoch": 2.09, "grad_norm": 15.853086634235515, "learning_rate": 4.449118629453133e-06, "loss": 0.4574, "step": 13381 }, { "epoch": 2.09, "grad_norm": 16.162096127968514, "learning_rate": 4.447715975995848e-06, "loss": 0.4785, "step": 13382 }, { "epoch": 2.09, "grad_norm": 21.77159990249389, "learning_rate": 4.446313480442272e-06, "loss": 0.516, "step": 13383 }, { "epoch": 2.09, "grad_norm": 17.727921858709138, "learning_rate": 4.444911142832297e-06, "loss": 0.5095, "step": 13384 }, { "epoch": 2.09, "grad_norm": 19.91745334847935, "learning_rate": 4.4435089632058e-06, "loss": 0.545, "step": 13385 }, { "epoch": 2.09, "grad_norm": 20.937821783610683, "learning_rate": 4.442106941602652e-06, "loss": 0.3844, "step": 13386 }, { "epoch": 2.09, "grad_norm": 15.151497540172905, "learning_rate": 4.440705078062732e-06, "loss": 0.4793, "step": 13387 }, { "epoch": 2.09, "grad_norm": 14.833597902207083, "learning_rate": 4.4393033726259116e-06, "loss": 0.4458, "step": 13388 }, { "epoch": 2.09, "grad_norm": 22.930245868701412, "learning_rate": 4.437901825332046e-06, "loss": 0.5224, "step": 13389 }, { "epoch": 2.09, "grad_norm": 35.93038643051206, "learning_rate": 4.436500436221003e-06, "loss": 0.5245, "step": 13390 }, { "epoch": 2.09, "grad_norm": 14.300825086649915, "learning_rate": 4.4350992053326295e-06, "loss": 0.4769, "step": 13391 }, { "epoch": 2.09, "grad_norm": 15.15378430110901, "learning_rate": 4.433698132706779e-06, "loss": 0.4482, "step": 13392 }, { "epoch": 2.09, "grad_norm": 23.1344869407657, "learning_rate": 4.4322972183833e-06, "loss": 0.4981, "step": 13393 }, { "epoch": 2.09, "grad_norm": 14.851429273959964, "learning_rate": 4.430896462402033e-06, "loss": 0.4633, "step": 13394 }, { "epoch": 2.09, "grad_norm": 22.324897390681926, "learning_rate": 4.429495864802808e-06, "loss": 0.5283, "step": 13395 }, { "epoch": 2.09, "grad_norm": 21.50423002562317, "learning_rate": 4.428095425625462e-06, "loss": 0.5526, "step": 13396 }, { "epoch": 2.09, "grad_norm": 15.829657741191447, "learning_rate": 4.426695144909826e-06, "loss": 0.4498, "step": 13397 }, { "epoch": 2.09, "grad_norm": 15.97235357949818, "learning_rate": 4.425295022695716e-06, "loss": 0.4363, "step": 13398 }, { "epoch": 2.09, "grad_norm": 17.15532984810896, "learning_rate": 4.423895059022959e-06, "loss": 0.5423, "step": 13399 }, { "epoch": 2.09, "grad_norm": 29.374390386789514, "learning_rate": 4.42249525393136e-06, "loss": 0.4868, "step": 13400 }, { "epoch": 2.09, "grad_norm": 19.370441946581966, "learning_rate": 4.421095607460738e-06, "loss": 0.5241, "step": 13401 }, { "epoch": 2.09, "grad_norm": 24.046294385656193, "learning_rate": 4.419696119650889e-06, "loss": 0.5022, "step": 13402 }, { "epoch": 2.09, "grad_norm": 14.471081369764708, "learning_rate": 4.418296790541618e-06, "loss": 0.453, "step": 13403 }, { "epoch": 2.09, "grad_norm": 17.778671540287874, "learning_rate": 4.4168976201727255e-06, "loss": 0.5484, "step": 13404 }, { "epoch": 2.09, "grad_norm": 22.642828665612175, "learning_rate": 4.415498608583993e-06, "loss": 0.498, "step": 13405 }, { "epoch": 2.09, "grad_norm": 19.196734116869575, "learning_rate": 4.414099755815219e-06, "loss": 0.5427, "step": 13406 }, { "epoch": 2.09, "grad_norm": 24.414930914730913, "learning_rate": 4.412701061906176e-06, "loss": 0.4866, "step": 13407 }, { "epoch": 2.09, "grad_norm": 19.613471442240435, "learning_rate": 4.411302526896646e-06, "loss": 0.4846, "step": 13408 }, { "epoch": 2.09, "grad_norm": 15.281869920587228, "learning_rate": 4.409904150826408e-06, "loss": 0.487, "step": 13409 }, { "epoch": 2.09, "grad_norm": 16.914756785320684, "learning_rate": 4.408505933735223e-06, "loss": 0.4572, "step": 13410 }, { "epoch": 2.09, "grad_norm": 31.268555643467717, "learning_rate": 4.407107875662857e-06, "loss": 0.5634, "step": 13411 }, { "epoch": 2.09, "grad_norm": 40.664296413192254, "learning_rate": 4.405709976649069e-06, "loss": 0.4784, "step": 13412 }, { "epoch": 2.1, "grad_norm": 26.10843387513603, "learning_rate": 4.404312236733623e-06, "loss": 0.5861, "step": 13413 }, { "epoch": 2.1, "grad_norm": 20.622896061499468, "learning_rate": 4.402914655956262e-06, "loss": 0.4474, "step": 13414 }, { "epoch": 2.1, "grad_norm": 19.505891463197088, "learning_rate": 4.401517234356731e-06, "loss": 0.469, "step": 13415 }, { "epoch": 2.1, "grad_norm": 17.116856751582834, "learning_rate": 4.400119971974774e-06, "loss": 0.4995, "step": 13416 }, { "epoch": 2.1, "grad_norm": 18.487069225015794, "learning_rate": 4.398722868850131e-06, "loss": 0.5086, "step": 13417 }, { "epoch": 2.1, "grad_norm": 27.17987141460619, "learning_rate": 4.3973259250225355e-06, "loss": 0.5204, "step": 13418 }, { "epoch": 2.1, "grad_norm": 22.958732782735332, "learning_rate": 4.395929140531713e-06, "loss": 0.5012, "step": 13419 }, { "epoch": 2.1, "grad_norm": 18.360661793517504, "learning_rate": 4.3945325154173834e-06, "loss": 0.4872, "step": 13420 }, { "epoch": 2.1, "grad_norm": 14.569555511957606, "learning_rate": 4.39313604971927e-06, "loss": 0.487, "step": 13421 }, { "epoch": 2.1, "grad_norm": 17.395513677277506, "learning_rate": 4.391739743477092e-06, "loss": 0.4899, "step": 13422 }, { "epoch": 2.1, "grad_norm": 15.362455836163539, "learning_rate": 4.390343596730551e-06, "loss": 0.4725, "step": 13423 }, { "epoch": 2.1, "grad_norm": 16.110999586455385, "learning_rate": 4.38894760951936e-06, "loss": 0.4512, "step": 13424 }, { "epoch": 2.1, "grad_norm": 34.52331145484327, "learning_rate": 4.387551781883213e-06, "loss": 0.4879, "step": 13425 }, { "epoch": 2.1, "grad_norm": 14.1126875864702, "learning_rate": 4.386156113861814e-06, "loss": 0.514, "step": 13426 }, { "epoch": 2.1, "grad_norm": 21.91621939521298, "learning_rate": 4.384760605494847e-06, "loss": 0.4568, "step": 13427 }, { "epoch": 2.1, "grad_norm": 21.71690157702032, "learning_rate": 4.383365256822003e-06, "loss": 0.5882, "step": 13428 }, { "epoch": 2.1, "grad_norm": 28.110465908033323, "learning_rate": 4.3819700678829705e-06, "loss": 0.5001, "step": 13429 }, { "epoch": 2.1, "grad_norm": 24.28365850715801, "learning_rate": 4.380575038717419e-06, "loss": 0.4645, "step": 13430 }, { "epoch": 2.1, "grad_norm": 17.778307378775217, "learning_rate": 4.37918016936503e-06, "loss": 0.4423, "step": 13431 }, { "epoch": 2.1, "grad_norm": 18.601300170022775, "learning_rate": 4.377785459865463e-06, "loss": 0.4522, "step": 13432 }, { "epoch": 2.1, "grad_norm": 22.816120012352762, "learning_rate": 4.376390910258391e-06, "loss": 0.5215, "step": 13433 }, { "epoch": 2.1, "grad_norm": 15.929450510871082, "learning_rate": 4.374996520583474e-06, "loss": 0.4921, "step": 13434 }, { "epoch": 2.1, "grad_norm": 21.439954293978737, "learning_rate": 4.373602290880367e-06, "loss": 0.5448, "step": 13435 }, { "epoch": 2.1, "grad_norm": 19.480410820748844, "learning_rate": 4.372208221188715e-06, "loss": 0.5549, "step": 13436 }, { "epoch": 2.1, "grad_norm": 18.031498653574978, "learning_rate": 4.370814311548168e-06, "loss": 0.5153, "step": 13437 }, { "epoch": 2.1, "grad_norm": 15.58080708478236, "learning_rate": 4.3694205619983744e-06, "loss": 0.4444, "step": 13438 }, { "epoch": 2.1, "grad_norm": 17.081969274360034, "learning_rate": 4.368026972578965e-06, "loss": 0.4861, "step": 13439 }, { "epoch": 2.1, "grad_norm": 19.746778438933905, "learning_rate": 4.366633543329571e-06, "loss": 0.4883, "step": 13440 }, { "epoch": 2.1, "grad_norm": 20.18715196160379, "learning_rate": 4.365240274289824e-06, "loss": 0.5169, "step": 13441 }, { "epoch": 2.1, "grad_norm": 21.94904135873195, "learning_rate": 4.363847165499347e-06, "loss": 0.5225, "step": 13442 }, { "epoch": 2.1, "grad_norm": 21.299634151607687, "learning_rate": 4.362454216997761e-06, "loss": 0.5753, "step": 13443 }, { "epoch": 2.1, "grad_norm": 15.082429075885264, "learning_rate": 4.361061428824682e-06, "loss": 0.4669, "step": 13444 }, { "epoch": 2.1, "grad_norm": 14.104073344794825, "learning_rate": 4.359668801019713e-06, "loss": 0.4348, "step": 13445 }, { "epoch": 2.1, "grad_norm": 23.718910643004673, "learning_rate": 4.358276333622463e-06, "loss": 0.4432, "step": 13446 }, { "epoch": 2.1, "grad_norm": 22.711772740623914, "learning_rate": 4.356884026672537e-06, "loss": 0.5723, "step": 13447 }, { "epoch": 2.1, "grad_norm": 25.126831003982765, "learning_rate": 4.35549188020953e-06, "loss": 0.5013, "step": 13448 }, { "epoch": 2.1, "grad_norm": 20.404262327988558, "learning_rate": 4.354099894273027e-06, "loss": 0.4589, "step": 13449 }, { "epoch": 2.1, "grad_norm": 20.73178193177861, "learning_rate": 4.352708068902621e-06, "loss": 0.5353, "step": 13450 }, { "epoch": 2.1, "grad_norm": 11.986171861251247, "learning_rate": 4.351316404137898e-06, "loss": 0.4418, "step": 13451 }, { "epoch": 2.1, "grad_norm": 20.675801292465934, "learning_rate": 4.349924900018427e-06, "loss": 0.4957, "step": 13452 }, { "epoch": 2.1, "grad_norm": 26.15825558052395, "learning_rate": 4.3485335565837905e-06, "loss": 0.467, "step": 13453 }, { "epoch": 2.1, "grad_norm": 19.360053995739154, "learning_rate": 4.347142373873551e-06, "loss": 0.5268, "step": 13454 }, { "epoch": 2.1, "grad_norm": 18.354302844395008, "learning_rate": 4.345751351927275e-06, "loss": 0.4907, "step": 13455 }, { "epoch": 2.1, "grad_norm": 12.607745848555282, "learning_rate": 4.344360490784526e-06, "loss": 0.3871, "step": 13456 }, { "epoch": 2.1, "grad_norm": 15.682853313083948, "learning_rate": 4.342969790484853e-06, "loss": 0.507, "step": 13457 }, { "epoch": 2.1, "grad_norm": 12.511992220193015, "learning_rate": 4.341579251067815e-06, "loss": 0.5151, "step": 13458 }, { "epoch": 2.1, "grad_norm": 20.429363591706245, "learning_rate": 4.3401888725729465e-06, "loss": 0.4728, "step": 13459 }, { "epoch": 2.1, "grad_norm": 22.251778645359995, "learning_rate": 4.338798655039802e-06, "loss": 0.5422, "step": 13460 }, { "epoch": 2.1, "grad_norm": 28.283733727263364, "learning_rate": 4.3374085985079075e-06, "loss": 0.4668, "step": 13461 }, { "epoch": 2.1, "grad_norm": 27.149063186570853, "learning_rate": 4.336018703016799e-06, "loss": 0.6065, "step": 13462 }, { "epoch": 2.1, "grad_norm": 15.774726462637535, "learning_rate": 4.33462896860601e-06, "loss": 0.5066, "step": 13463 }, { "epoch": 2.1, "grad_norm": 25.6124587660728, "learning_rate": 4.3332393953150574e-06, "loss": 0.5152, "step": 13464 }, { "epoch": 2.1, "grad_norm": 20.04154843680813, "learning_rate": 4.331849983183459e-06, "loss": 0.4485, "step": 13465 }, { "epoch": 2.1, "grad_norm": 20.307506383220737, "learning_rate": 4.330460732250732e-06, "loss": 0.5035, "step": 13466 }, { "epoch": 2.1, "grad_norm": 20.697163901889468, "learning_rate": 4.329071642556384e-06, "loss": 0.504, "step": 13467 }, { "epoch": 2.1, "grad_norm": 16.534570776876762, "learning_rate": 4.327682714139925e-06, "loss": 0.5129, "step": 13468 }, { "epoch": 2.1, "grad_norm": 23.343498125849504, "learning_rate": 4.326293947040851e-06, "loss": 0.5047, "step": 13469 }, { "epoch": 2.1, "grad_norm": 17.346328722351895, "learning_rate": 4.324905341298655e-06, "loss": 0.4915, "step": 13470 }, { "epoch": 2.1, "grad_norm": 13.040168434746302, "learning_rate": 4.323516896952832e-06, "loss": 0.4433, "step": 13471 }, { "epoch": 2.1, "grad_norm": 22.014367330946357, "learning_rate": 4.3221286140428695e-06, "loss": 0.5257, "step": 13472 }, { "epoch": 2.1, "grad_norm": 19.90869229918086, "learning_rate": 4.32074049260825e-06, "loss": 0.4855, "step": 13473 }, { "epoch": 2.1, "grad_norm": 19.19758465507929, "learning_rate": 4.319352532688444e-06, "loss": 0.4622, "step": 13474 }, { "epoch": 2.1, "grad_norm": 18.783367663592294, "learning_rate": 4.317964734322928e-06, "loss": 0.4441, "step": 13475 }, { "epoch": 2.1, "grad_norm": 22.85024333048001, "learning_rate": 4.316577097551176e-06, "loss": 0.4744, "step": 13476 }, { "epoch": 2.11, "grad_norm": 15.207394375839547, "learning_rate": 4.315189622412642e-06, "loss": 0.4636, "step": 13477 }, { "epoch": 2.11, "grad_norm": 23.860458434674364, "learning_rate": 4.313802308946794e-06, "loss": 0.5367, "step": 13478 }, { "epoch": 2.11, "grad_norm": 12.696411620464664, "learning_rate": 4.312415157193078e-06, "loss": 0.4785, "step": 13479 }, { "epoch": 2.11, "grad_norm": 20.291071754104014, "learning_rate": 4.3110281671909484e-06, "loss": 0.5058, "step": 13480 }, { "epoch": 2.11, "grad_norm": 19.7198849206344, "learning_rate": 4.309641338979853e-06, "loss": 0.5133, "step": 13481 }, { "epoch": 2.11, "grad_norm": 14.280293193668603, "learning_rate": 4.308254672599225e-06, "loss": 0.47, "step": 13482 }, { "epoch": 2.11, "grad_norm": 17.55920090012835, "learning_rate": 4.306868168088508e-06, "loss": 0.4729, "step": 13483 }, { "epoch": 2.11, "grad_norm": 16.34642124592735, "learning_rate": 4.305481825487128e-06, "loss": 0.4877, "step": 13484 }, { "epoch": 2.11, "grad_norm": 19.00180157595025, "learning_rate": 4.304095644834516e-06, "loss": 0.5126, "step": 13485 }, { "epoch": 2.11, "grad_norm": 16.27862461031544, "learning_rate": 4.302709626170089e-06, "loss": 0.5341, "step": 13486 }, { "epoch": 2.11, "grad_norm": 19.11646340000066, "learning_rate": 4.30132376953327e-06, "loss": 0.3857, "step": 13487 }, { "epoch": 2.11, "grad_norm": 21.20590603693105, "learning_rate": 4.299938074963465e-06, "loss": 0.4701, "step": 13488 }, { "epoch": 2.11, "grad_norm": 18.47539332846092, "learning_rate": 4.298552542500093e-06, "loss": 0.502, "step": 13489 }, { "epoch": 2.11, "grad_norm": 42.398325496980426, "learning_rate": 4.297167172182546e-06, "loss": 0.5287, "step": 13490 }, { "epoch": 2.11, "grad_norm": 38.68794039397025, "learning_rate": 4.295781964050229e-06, "loss": 0.5194, "step": 13491 }, { "epoch": 2.11, "grad_norm": 16.276072685491886, "learning_rate": 4.2943969181425395e-06, "loss": 0.4434, "step": 13492 }, { "epoch": 2.11, "grad_norm": 21.53435643712485, "learning_rate": 4.2930120344988605e-06, "loss": 0.4705, "step": 13493 }, { "epoch": 2.11, "grad_norm": 24.75365813631957, "learning_rate": 4.291627313158584e-06, "loss": 0.5483, "step": 13494 }, { "epoch": 2.11, "grad_norm": 19.243731719798483, "learning_rate": 4.290242754161084e-06, "loss": 0.493, "step": 13495 }, { "epoch": 2.11, "grad_norm": 16.60385053777235, "learning_rate": 4.28885835754574e-06, "loss": 0.4554, "step": 13496 }, { "epoch": 2.11, "grad_norm": 32.31787061053695, "learning_rate": 4.287474123351928e-06, "loss": 0.541, "step": 13497 }, { "epoch": 2.11, "grad_norm": 20.96526104970375, "learning_rate": 4.286090051619008e-06, "loss": 0.4559, "step": 13498 }, { "epoch": 2.11, "grad_norm": 21.100422787102886, "learning_rate": 4.284706142386342e-06, "loss": 0.5356, "step": 13499 }, { "epoch": 2.11, "grad_norm": 19.722016532497694, "learning_rate": 4.2833223956932916e-06, "loss": 0.5214, "step": 13500 }, { "epoch": 2.11, "grad_norm": 17.78271697635198, "learning_rate": 4.2819388115792095e-06, "loss": 0.4669, "step": 13501 }, { "epoch": 2.11, "grad_norm": 12.97942187587495, "learning_rate": 4.280555390083443e-06, "loss": 0.4304, "step": 13502 }, { "epoch": 2.11, "grad_norm": 18.659132246796773, "learning_rate": 4.279172131245332e-06, "loss": 0.4608, "step": 13503 }, { "epoch": 2.11, "grad_norm": 21.692096957162224, "learning_rate": 4.27778903510422e-06, "loss": 0.5481, "step": 13504 }, { "epoch": 2.11, "grad_norm": 24.752425590297722, "learning_rate": 4.2764061016994386e-06, "loss": 0.538, "step": 13505 }, { "epoch": 2.11, "grad_norm": 20.18439080037953, "learning_rate": 4.275023331070324e-06, "loss": 0.5294, "step": 13506 }, { "epoch": 2.11, "grad_norm": 20.458798299958378, "learning_rate": 4.2736407232561975e-06, "loss": 0.4268, "step": 13507 }, { "epoch": 2.11, "grad_norm": 23.10961786649343, "learning_rate": 4.272258278296374e-06, "loss": 0.4963, "step": 13508 }, { "epoch": 2.11, "grad_norm": 17.70235840348874, "learning_rate": 4.270875996230176e-06, "loss": 0.5808, "step": 13509 }, { "epoch": 2.11, "grad_norm": 28.62649659082115, "learning_rate": 4.269493877096915e-06, "loss": 0.5652, "step": 13510 }, { "epoch": 2.11, "grad_norm": 20.99787160008829, "learning_rate": 4.268111920935893e-06, "loss": 0.4426, "step": 13511 }, { "epoch": 2.11, "grad_norm": 20.321086393620956, "learning_rate": 4.266730127786419e-06, "loss": 0.4878, "step": 13512 }, { "epoch": 2.11, "grad_norm": 27.343704060107193, "learning_rate": 4.265348497687784e-06, "loss": 0.495, "step": 13513 }, { "epoch": 2.11, "grad_norm": 24.921945488091644, "learning_rate": 4.263967030679284e-06, "loss": 0.4941, "step": 13514 }, { "epoch": 2.11, "grad_norm": 20.520433813811366, "learning_rate": 4.262585726800204e-06, "loss": 0.5092, "step": 13515 }, { "epoch": 2.11, "grad_norm": 17.808686122499925, "learning_rate": 4.26120458608983e-06, "loss": 0.5165, "step": 13516 }, { "epoch": 2.11, "grad_norm": 18.30822553360003, "learning_rate": 4.259823608587443e-06, "loss": 0.505, "step": 13517 }, { "epoch": 2.11, "grad_norm": 16.909309206936634, "learning_rate": 4.258442794332311e-06, "loss": 0.4927, "step": 13518 }, { "epoch": 2.11, "grad_norm": 28.05286728858601, "learning_rate": 4.2570621433637116e-06, "loss": 0.5266, "step": 13519 }, { "epoch": 2.11, "grad_norm": 28.709137145186844, "learning_rate": 4.255681655720901e-06, "loss": 0.6329, "step": 13520 }, { "epoch": 2.11, "grad_norm": 20.58938696985488, "learning_rate": 4.254301331443145e-06, "loss": 0.5211, "step": 13521 }, { "epoch": 2.11, "grad_norm": 16.575775673928216, "learning_rate": 4.252921170569701e-06, "loss": 0.4887, "step": 13522 }, { "epoch": 2.11, "grad_norm": 19.976739653636162, "learning_rate": 4.251541173139816e-06, "loss": 0.4615, "step": 13523 }, { "epoch": 2.11, "grad_norm": 16.604192278645606, "learning_rate": 4.250161339192734e-06, "loss": 0.4779, "step": 13524 }, { "epoch": 2.11, "grad_norm": 21.061330305781027, "learning_rate": 4.2487816687677006e-06, "loss": 0.4923, "step": 13525 }, { "epoch": 2.11, "grad_norm": 22.94314081684299, "learning_rate": 4.247402161903956e-06, "loss": 0.4928, "step": 13526 }, { "epoch": 2.11, "grad_norm": 17.352750244607382, "learning_rate": 4.246022818640728e-06, "loss": 0.5382, "step": 13527 }, { "epoch": 2.11, "grad_norm": 18.118142313798433, "learning_rate": 4.244643639017242e-06, "loss": 0.513, "step": 13528 }, { "epoch": 2.11, "grad_norm": 19.816358548555687, "learning_rate": 4.243264623072725e-06, "loss": 0.5031, "step": 13529 }, { "epoch": 2.11, "grad_norm": 28.05101485501066, "learning_rate": 4.241885770846392e-06, "loss": 0.5212, "step": 13530 }, { "epoch": 2.11, "grad_norm": 20.47354865135328, "learning_rate": 4.240507082377465e-06, "loss": 0.5088, "step": 13531 }, { "epoch": 2.11, "grad_norm": 26.268185417557632, "learning_rate": 4.239128557705146e-06, "loss": 0.4698, "step": 13532 }, { "epoch": 2.11, "grad_norm": 28.84944940364741, "learning_rate": 4.237750196868639e-06, "loss": 0.5317, "step": 13533 }, { "epoch": 2.11, "grad_norm": 20.476987320569467, "learning_rate": 4.236371999907144e-06, "loss": 0.4917, "step": 13534 }, { "epoch": 2.11, "grad_norm": 14.891998469844339, "learning_rate": 4.234993966859862e-06, "loss": 0.4485, "step": 13535 }, { "epoch": 2.11, "grad_norm": 18.146306924811462, "learning_rate": 4.233616097765979e-06, "loss": 0.5415, "step": 13536 }, { "epoch": 2.11, "grad_norm": 20.932540197922695, "learning_rate": 4.232238392664677e-06, "loss": 0.4883, "step": 13537 }, { "epoch": 2.11, "grad_norm": 13.163851399526058, "learning_rate": 4.2308608515951414e-06, "loss": 0.4369, "step": 13538 }, { "epoch": 2.11, "grad_norm": 15.091191036385965, "learning_rate": 4.229483474596552e-06, "loss": 0.4475, "step": 13539 }, { "epoch": 2.11, "grad_norm": 23.806749958748377, "learning_rate": 4.228106261708073e-06, "loss": 0.4853, "step": 13540 }, { "epoch": 2.12, "grad_norm": 23.165884479575702, "learning_rate": 4.22672921296888e-06, "loss": 0.4928, "step": 13541 }, { "epoch": 2.12, "grad_norm": 21.29297327151099, "learning_rate": 4.225352328418126e-06, "loss": 0.5042, "step": 13542 }, { "epoch": 2.12, "grad_norm": 18.2225418156734, "learning_rate": 4.223975608094975e-06, "loss": 0.5866, "step": 13543 }, { "epoch": 2.12, "grad_norm": 15.721819138129119, "learning_rate": 4.222599052038581e-06, "loss": 0.4423, "step": 13544 }, { "epoch": 2.12, "grad_norm": 24.805819150086123, "learning_rate": 4.221222660288087e-06, "loss": 0.4885, "step": 13545 }, { "epoch": 2.12, "grad_norm": 20.034232436800302, "learning_rate": 4.219846432882644e-06, "loss": 0.4366, "step": 13546 }, { "epoch": 2.12, "grad_norm": 29.151975361289153, "learning_rate": 4.218470369861384e-06, "loss": 0.6061, "step": 13547 }, { "epoch": 2.12, "grad_norm": 26.154425497011548, "learning_rate": 4.217094471263447e-06, "loss": 0.5386, "step": 13548 }, { "epoch": 2.12, "grad_norm": 16.32975497437622, "learning_rate": 4.2157187371279585e-06, "loss": 0.4813, "step": 13549 }, { "epoch": 2.12, "grad_norm": 26.540746655056783, "learning_rate": 4.214343167494044e-06, "loss": 0.4965, "step": 13550 }, { "epoch": 2.12, "grad_norm": 15.448068747658212, "learning_rate": 4.2129677624008294e-06, "loss": 0.4188, "step": 13551 }, { "epoch": 2.12, "grad_norm": 14.746551157949762, "learning_rate": 4.211592521887426e-06, "loss": 0.4561, "step": 13552 }, { "epoch": 2.12, "grad_norm": 13.01736470911061, "learning_rate": 4.2102174459929424e-06, "loss": 0.4634, "step": 13553 }, { "epoch": 2.12, "grad_norm": 19.461717377866623, "learning_rate": 4.2088425347564866e-06, "loss": 0.4935, "step": 13554 }, { "epoch": 2.12, "grad_norm": 15.913210044486908, "learning_rate": 4.207467788217162e-06, "loss": 0.4567, "step": 13555 }, { "epoch": 2.12, "grad_norm": 13.193984821439942, "learning_rate": 4.206093206414069e-06, "loss": 0.4453, "step": 13556 }, { "epoch": 2.12, "grad_norm": 35.940176690803796, "learning_rate": 4.204718789386295e-06, "loss": 0.5471, "step": 13557 }, { "epoch": 2.12, "grad_norm": 19.406448302584195, "learning_rate": 4.203344537172925e-06, "loss": 0.4823, "step": 13558 }, { "epoch": 2.12, "grad_norm": 17.516960817506494, "learning_rate": 4.201970449813045e-06, "loss": 0.427, "step": 13559 }, { "epoch": 2.12, "grad_norm": 20.517757946971244, "learning_rate": 4.200596527345738e-06, "loss": 0.4655, "step": 13560 }, { "epoch": 2.12, "grad_norm": 19.124877200012897, "learning_rate": 4.199222769810074e-06, "loss": 0.5051, "step": 13561 }, { "epoch": 2.12, "grad_norm": 19.653873230921988, "learning_rate": 4.197849177245117e-06, "loss": 0.5575, "step": 13562 }, { "epoch": 2.12, "grad_norm": 15.744192549943213, "learning_rate": 4.196475749689934e-06, "loss": 0.5311, "step": 13563 }, { "epoch": 2.12, "grad_norm": 20.507579584161434, "learning_rate": 4.19510248718359e-06, "loss": 0.5005, "step": 13564 }, { "epoch": 2.12, "grad_norm": 25.340815557395395, "learning_rate": 4.1937293897651324e-06, "loss": 0.4535, "step": 13565 }, { "epoch": 2.12, "grad_norm": 15.670890821891394, "learning_rate": 4.192356457473617e-06, "loss": 0.4888, "step": 13566 }, { "epoch": 2.12, "grad_norm": 23.3920332498085, "learning_rate": 4.190983690348084e-06, "loss": 0.5754, "step": 13567 }, { "epoch": 2.12, "grad_norm": 19.55890911024046, "learning_rate": 4.189611088427576e-06, "loss": 0.5032, "step": 13568 }, { "epoch": 2.12, "grad_norm": 15.953563424492723, "learning_rate": 4.188238651751134e-06, "loss": 0.4359, "step": 13569 }, { "epoch": 2.12, "grad_norm": 16.11835830233997, "learning_rate": 4.186866380357782e-06, "loss": 0.4383, "step": 13570 }, { "epoch": 2.12, "grad_norm": 13.627039716220475, "learning_rate": 4.1854942742865514e-06, "loss": 0.4585, "step": 13571 }, { "epoch": 2.12, "grad_norm": 16.443986626544245, "learning_rate": 4.184122333576459e-06, "loss": 0.4323, "step": 13572 }, { "epoch": 2.12, "grad_norm": 25.55811324404679, "learning_rate": 4.1827505582665295e-06, "loss": 0.5244, "step": 13573 }, { "epoch": 2.12, "grad_norm": 23.951393495349535, "learning_rate": 4.181378948395767e-06, "loss": 0.4781, "step": 13574 }, { "epoch": 2.12, "grad_norm": 16.91530885927774, "learning_rate": 4.180007504003183e-06, "loss": 0.4867, "step": 13575 }, { "epoch": 2.12, "grad_norm": 23.51367503981383, "learning_rate": 4.178636225127786e-06, "loss": 0.5121, "step": 13576 }, { "epoch": 2.12, "grad_norm": 21.956888057565543, "learning_rate": 4.177265111808568e-06, "loss": 0.5423, "step": 13577 }, { "epoch": 2.12, "grad_norm": 32.657642296492234, "learning_rate": 4.175894164084521e-06, "loss": 0.4856, "step": 13578 }, { "epoch": 2.12, "grad_norm": 24.39869156722586, "learning_rate": 4.174523381994637e-06, "loss": 0.5382, "step": 13579 }, { "epoch": 2.12, "grad_norm": 16.979040378898556, "learning_rate": 4.1731527655779e-06, "loss": 0.4163, "step": 13580 }, { "epoch": 2.12, "grad_norm": 24.721661672297405, "learning_rate": 4.171782314873294e-06, "loss": 0.5212, "step": 13581 }, { "epoch": 2.12, "grad_norm": 14.207655940404102, "learning_rate": 4.170412029919788e-06, "loss": 0.4455, "step": 13582 }, { "epoch": 2.12, "grad_norm": 15.62061479067443, "learning_rate": 4.169041910756352e-06, "loss": 0.4677, "step": 13583 }, { "epoch": 2.12, "grad_norm": 28.295346041299343, "learning_rate": 4.167671957421952e-06, "loss": 0.4471, "step": 13584 }, { "epoch": 2.12, "grad_norm": 19.790220923991725, "learning_rate": 4.166302169955553e-06, "loss": 0.4968, "step": 13585 }, { "epoch": 2.12, "grad_norm": 17.582258738732836, "learning_rate": 4.164932548396108e-06, "loss": 0.5303, "step": 13586 }, { "epoch": 2.12, "grad_norm": 21.44405705715216, "learning_rate": 4.163563092782564e-06, "loss": 0.4366, "step": 13587 }, { "epoch": 2.12, "grad_norm": 14.801282382393591, "learning_rate": 4.162193803153872e-06, "loss": 0.545, "step": 13588 }, { "epoch": 2.12, "grad_norm": 14.452182708128388, "learning_rate": 4.160824679548977e-06, "loss": 0.4392, "step": 13589 }, { "epoch": 2.12, "grad_norm": 30.298217108255777, "learning_rate": 4.1594557220068125e-06, "loss": 0.4864, "step": 13590 }, { "epoch": 2.12, "grad_norm": 26.0626089839349, "learning_rate": 4.158086930566307e-06, "loss": 0.61, "step": 13591 }, { "epoch": 2.12, "grad_norm": 29.86377432037528, "learning_rate": 4.15671830526639e-06, "loss": 0.4633, "step": 13592 }, { "epoch": 2.12, "grad_norm": 18.084666647865678, "learning_rate": 4.155349846145988e-06, "loss": 0.4707, "step": 13593 }, { "epoch": 2.12, "grad_norm": 18.109577938535764, "learning_rate": 4.153981553244021e-06, "loss": 0.4985, "step": 13594 }, { "epoch": 2.12, "grad_norm": 23.339063911976808, "learning_rate": 4.152613426599398e-06, "loss": 0.6181, "step": 13595 }, { "epoch": 2.12, "grad_norm": 18.537277732569315, "learning_rate": 4.151245466251025e-06, "loss": 0.4634, "step": 13596 }, { "epoch": 2.12, "grad_norm": 20.67606763859391, "learning_rate": 4.14987767223781e-06, "loss": 0.4626, "step": 13597 }, { "epoch": 2.12, "grad_norm": 17.820083548125744, "learning_rate": 4.148510044598655e-06, "loss": 0.4513, "step": 13598 }, { "epoch": 2.12, "grad_norm": 14.20304431619588, "learning_rate": 4.147142583372448e-06, "loss": 0.4403, "step": 13599 }, { "epoch": 2.12, "grad_norm": 19.956380531482687, "learning_rate": 4.145775288598085e-06, "loss": 0.4714, "step": 13600 }, { "epoch": 2.12, "grad_norm": 18.326358972281653, "learning_rate": 4.1444081603144445e-06, "loss": 0.5179, "step": 13601 }, { "epoch": 2.12, "grad_norm": 20.695428501969044, "learning_rate": 4.1430411985604145e-06, "loss": 0.4596, "step": 13602 }, { "epoch": 2.12, "grad_norm": 24.93097910976537, "learning_rate": 4.141674403374864e-06, "loss": 0.5294, "step": 13603 }, { "epoch": 2.12, "grad_norm": 22.41000683498459, "learning_rate": 4.1403077747966646e-06, "loss": 0.5187, "step": 13604 }, { "epoch": 2.13, "grad_norm": 18.835924438098633, "learning_rate": 4.13894131286469e-06, "loss": 0.4493, "step": 13605 }, { "epoch": 2.13, "grad_norm": 19.468391398315873, "learning_rate": 4.13757501761779e-06, "loss": 0.5106, "step": 13606 }, { "epoch": 2.13, "grad_norm": 19.169558327599734, "learning_rate": 4.136208889094832e-06, "loss": 0.5085, "step": 13607 }, { "epoch": 2.13, "grad_norm": 30.760694053416294, "learning_rate": 4.1348429273346595e-06, "loss": 0.4817, "step": 13608 }, { "epoch": 2.13, "grad_norm": 18.130547217234085, "learning_rate": 4.133477132376122e-06, "loss": 0.4943, "step": 13609 }, { "epoch": 2.13, "grad_norm": 20.134660244498594, "learning_rate": 4.132111504258067e-06, "loss": 0.4764, "step": 13610 }, { "epoch": 2.13, "grad_norm": 18.11397962311846, "learning_rate": 4.130746043019329e-06, "loss": 0.47, "step": 13611 }, { "epoch": 2.13, "grad_norm": 11.031046037436356, "learning_rate": 4.129380748698737e-06, "loss": 0.51, "step": 13612 }, { "epoch": 2.13, "grad_norm": 21.06777510111775, "learning_rate": 4.128015621335121e-06, "loss": 0.5151, "step": 13613 }, { "epoch": 2.13, "grad_norm": 15.195392775139672, "learning_rate": 4.12665066096731e-06, "loss": 0.5055, "step": 13614 }, { "epoch": 2.13, "grad_norm": 17.38425398003947, "learning_rate": 4.1252858676341175e-06, "loss": 0.4768, "step": 13615 }, { "epoch": 2.13, "grad_norm": 23.492607801792264, "learning_rate": 4.123921241374355e-06, "loss": 0.516, "step": 13616 }, { "epoch": 2.13, "grad_norm": 23.164307575955736, "learning_rate": 4.1225567822268365e-06, "loss": 0.4869, "step": 13617 }, { "epoch": 2.13, "grad_norm": 13.589675209840214, "learning_rate": 4.121192490230363e-06, "loss": 0.4767, "step": 13618 }, { "epoch": 2.13, "grad_norm": 16.242954502096712, "learning_rate": 4.11982836542374e-06, "loss": 0.4503, "step": 13619 }, { "epoch": 2.13, "grad_norm": 18.02628616834752, "learning_rate": 4.118464407845759e-06, "loss": 0.5123, "step": 13620 }, { "epoch": 2.13, "grad_norm": 13.525161193664482, "learning_rate": 4.117100617535207e-06, "loss": 0.4797, "step": 13621 }, { "epoch": 2.13, "grad_norm": 17.23901406084867, "learning_rate": 4.11573699453087e-06, "loss": 0.4386, "step": 13622 }, { "epoch": 2.13, "grad_norm": 17.815428997454436, "learning_rate": 4.114373538871535e-06, "loss": 0.4999, "step": 13623 }, { "epoch": 2.13, "grad_norm": 11.88352985085275, "learning_rate": 4.1130102505959715e-06, "loss": 0.4624, "step": 13624 }, { "epoch": 2.13, "grad_norm": 18.60170207397827, "learning_rate": 4.111647129742954e-06, "loss": 0.5037, "step": 13625 }, { "epoch": 2.13, "grad_norm": 15.241897778712433, "learning_rate": 4.110284176351245e-06, "loss": 0.4716, "step": 13626 }, { "epoch": 2.13, "grad_norm": 19.909195686506425, "learning_rate": 4.108921390459612e-06, "loss": 0.536, "step": 13627 }, { "epoch": 2.13, "grad_norm": 23.83249178786733, "learning_rate": 4.107558772106805e-06, "loss": 0.5445, "step": 13628 }, { "epoch": 2.13, "grad_norm": 19.100331026489787, "learning_rate": 4.106196321331581e-06, "loss": 0.4798, "step": 13629 }, { "epoch": 2.13, "grad_norm": 14.76686234714087, "learning_rate": 4.104834038172687e-06, "loss": 0.5654, "step": 13630 }, { "epoch": 2.13, "grad_norm": 16.727406996709465, "learning_rate": 4.103471922668862e-06, "loss": 0.4911, "step": 13631 }, { "epoch": 2.13, "grad_norm": 22.848115422367645, "learning_rate": 4.10210997485885e-06, "loss": 0.5234, "step": 13632 }, { "epoch": 2.13, "grad_norm": 22.365964335774482, "learning_rate": 4.100748194781376e-06, "loss": 0.4239, "step": 13633 }, { "epoch": 2.13, "grad_norm": 20.449688821689175, "learning_rate": 4.099386582475175e-06, "loss": 0.5003, "step": 13634 }, { "epoch": 2.13, "grad_norm": 31.09837584950147, "learning_rate": 4.0980251379789655e-06, "loss": 0.4586, "step": 13635 }, { "epoch": 2.13, "grad_norm": 22.19372386342677, "learning_rate": 4.096663861331472e-06, "loss": 0.4624, "step": 13636 }, { "epoch": 2.13, "grad_norm": 21.027454313568263, "learning_rate": 4.095302752571402e-06, "loss": 0.4547, "step": 13637 }, { "epoch": 2.13, "grad_norm": 30.405046929771117, "learning_rate": 4.093941811737466e-06, "loss": 0.5329, "step": 13638 }, { "epoch": 2.13, "grad_norm": 20.270770578509428, "learning_rate": 4.092581038868375e-06, "loss": 0.4578, "step": 13639 }, { "epoch": 2.13, "grad_norm": 24.343442198750296, "learning_rate": 4.0912204340028224e-06, "loss": 0.521, "step": 13640 }, { "epoch": 2.13, "grad_norm": 15.023524792437934, "learning_rate": 4.089859997179502e-06, "loss": 0.4562, "step": 13641 }, { "epoch": 2.13, "grad_norm": 21.9503108633932, "learning_rate": 4.088499728437104e-06, "loss": 0.5366, "step": 13642 }, { "epoch": 2.13, "grad_norm": 14.036514319506963, "learning_rate": 4.087139627814317e-06, "loss": 0.4861, "step": 13643 }, { "epoch": 2.13, "grad_norm": 21.313620862264777, "learning_rate": 4.085779695349823e-06, "loss": 0.5295, "step": 13644 }, { "epoch": 2.13, "grad_norm": 26.80087441051656, "learning_rate": 4.084419931082295e-06, "loss": 0.5164, "step": 13645 }, { "epoch": 2.13, "grad_norm": 20.693608031999965, "learning_rate": 4.083060335050401e-06, "loss": 0.5041, "step": 13646 }, { "epoch": 2.13, "grad_norm": 13.884248269917485, "learning_rate": 4.0817009072928085e-06, "loss": 0.3982, "step": 13647 }, { "epoch": 2.13, "grad_norm": 14.967372945924364, "learning_rate": 4.080341647848184e-06, "loss": 0.4749, "step": 13648 }, { "epoch": 2.13, "grad_norm": 17.246629287595802, "learning_rate": 4.07898255675518e-06, "loss": 0.4649, "step": 13649 }, { "epoch": 2.13, "grad_norm": 22.401588402596303, "learning_rate": 4.077623634052445e-06, "loss": 0.5824, "step": 13650 }, { "epoch": 2.13, "grad_norm": 21.589883329087062, "learning_rate": 4.076264879778629e-06, "loss": 0.4992, "step": 13651 }, { "epoch": 2.13, "grad_norm": 22.885499242978852, "learning_rate": 4.074906293972378e-06, "loss": 0.4625, "step": 13652 }, { "epoch": 2.13, "grad_norm": 19.665082111681873, "learning_rate": 4.073547876672323e-06, "loss": 0.5504, "step": 13653 }, { "epoch": 2.13, "grad_norm": 17.166712005149897, "learning_rate": 4.072189627917102e-06, "loss": 0.4586, "step": 13654 }, { "epoch": 2.13, "grad_norm": 19.487445647461794, "learning_rate": 4.070831547745338e-06, "loss": 0.4891, "step": 13655 }, { "epoch": 2.13, "grad_norm": 21.933707751744688, "learning_rate": 4.0694736361956555e-06, "loss": 0.4906, "step": 13656 }, { "epoch": 2.13, "grad_norm": 22.028510284517882, "learning_rate": 4.068115893306678e-06, "loss": 0.4744, "step": 13657 }, { "epoch": 2.13, "grad_norm": 18.826276854669484, "learning_rate": 4.06675831911701e-06, "loss": 0.4593, "step": 13658 }, { "epoch": 2.13, "grad_norm": 28.371458327776494, "learning_rate": 4.0654009136652685e-06, "loss": 0.4562, "step": 13659 }, { "epoch": 2.13, "grad_norm": 18.532878240921082, "learning_rate": 4.06404367699005e-06, "loss": 0.4931, "step": 13660 }, { "epoch": 2.13, "grad_norm": 21.0797959864339, "learning_rate": 4.0626866091299595e-06, "loss": 0.4673, "step": 13661 }, { "epoch": 2.13, "grad_norm": 15.074621787791056, "learning_rate": 4.061329710123586e-06, "loss": 0.4647, "step": 13662 }, { "epoch": 2.13, "grad_norm": 15.404031571572393, "learning_rate": 4.059972980009522e-06, "loss": 0.4751, "step": 13663 }, { "epoch": 2.13, "grad_norm": 18.388421726196704, "learning_rate": 4.058616418826355e-06, "loss": 0.4938, "step": 13664 }, { "epoch": 2.13, "grad_norm": 29.828047124322392, "learning_rate": 4.05726002661266e-06, "loss": 0.5372, "step": 13665 }, { "epoch": 2.13, "grad_norm": 15.08296897638668, "learning_rate": 4.055903803407011e-06, "loss": 0.4706, "step": 13666 }, { "epoch": 2.13, "grad_norm": 22.473197783283254, "learning_rate": 4.05454774924798e-06, "loss": 0.5085, "step": 13667 }, { "epoch": 2.13, "grad_norm": 28.580788632144923, "learning_rate": 4.0531918641741344e-06, "loss": 0.4934, "step": 13668 }, { "epoch": 2.14, "grad_norm": 21.64413508858331, "learning_rate": 4.051836148224035e-06, "loss": 0.5595, "step": 13669 }, { "epoch": 2.14, "grad_norm": 23.084197259882018, "learning_rate": 4.050480601436237e-06, "loss": 0.4821, "step": 13670 }, { "epoch": 2.14, "grad_norm": 19.58997925229432, "learning_rate": 4.049125223849287e-06, "loss": 0.4646, "step": 13671 }, { "epoch": 2.14, "grad_norm": 31.123418109979106, "learning_rate": 4.047770015501734e-06, "loss": 0.5362, "step": 13672 }, { "epoch": 2.14, "grad_norm": 18.521583464744715, "learning_rate": 4.046414976432124e-06, "loss": 0.4303, "step": 13673 }, { "epoch": 2.14, "grad_norm": 19.726309800748407, "learning_rate": 4.045060106678989e-06, "loss": 0.5317, "step": 13674 }, { "epoch": 2.14, "grad_norm": 19.419603646916862, "learning_rate": 4.043705406280857e-06, "loss": 0.494, "step": 13675 }, { "epoch": 2.14, "grad_norm": 18.07930658199789, "learning_rate": 4.04235087527626e-06, "loss": 0.4651, "step": 13676 }, { "epoch": 2.14, "grad_norm": 17.151086766839835, "learning_rate": 4.040996513703721e-06, "loss": 0.5083, "step": 13677 }, { "epoch": 2.14, "grad_norm": 19.109216745552544, "learning_rate": 4.039642321601753e-06, "loss": 0.4849, "step": 13678 }, { "epoch": 2.14, "grad_norm": 20.10280049311328, "learning_rate": 4.0382882990088735e-06, "loss": 0.4456, "step": 13679 }, { "epoch": 2.14, "grad_norm": 29.768161518423963, "learning_rate": 4.036934445963584e-06, "loss": 0.4621, "step": 13680 }, { "epoch": 2.14, "grad_norm": 20.009388870482887, "learning_rate": 4.035580762504391e-06, "loss": 0.5105, "step": 13681 }, { "epoch": 2.14, "grad_norm": 20.039032976600957, "learning_rate": 4.034227248669794e-06, "loss": 0.5935, "step": 13682 }, { "epoch": 2.14, "grad_norm": 20.264620074412747, "learning_rate": 4.032873904498286e-06, "loss": 0.5224, "step": 13683 }, { "epoch": 2.14, "grad_norm": 22.122891115448716, "learning_rate": 4.031520730028348e-06, "loss": 0.4953, "step": 13684 }, { "epoch": 2.14, "grad_norm": 20.5715564868856, "learning_rate": 4.030167725298472e-06, "loss": 0.5739, "step": 13685 }, { "epoch": 2.14, "grad_norm": 21.738966846193932, "learning_rate": 4.028814890347134e-06, "loss": 0.5055, "step": 13686 }, { "epoch": 2.14, "grad_norm": 13.60996665555389, "learning_rate": 4.027462225212806e-06, "loss": 0.4751, "step": 13687 }, { "epoch": 2.14, "grad_norm": 16.13213868708344, "learning_rate": 4.026109729933962e-06, "loss": 0.4357, "step": 13688 }, { "epoch": 2.14, "grad_norm": 23.05932294470333, "learning_rate": 4.024757404549058e-06, "loss": 0.5339, "step": 13689 }, { "epoch": 2.14, "grad_norm": 16.319212736306508, "learning_rate": 4.023405249096561e-06, "loss": 0.5883, "step": 13690 }, { "epoch": 2.14, "grad_norm": 15.859497610802075, "learning_rate": 4.022053263614921e-06, "loss": 0.4219, "step": 13691 }, { "epoch": 2.14, "grad_norm": 30.85810576931222, "learning_rate": 4.0207014481425875e-06, "loss": 0.5327, "step": 13692 }, { "epoch": 2.14, "grad_norm": 19.55922623288415, "learning_rate": 4.019349802718011e-06, "loss": 0.4878, "step": 13693 }, { "epoch": 2.14, "grad_norm": 13.247477699593244, "learning_rate": 4.017998327379626e-06, "loss": 0.4509, "step": 13694 }, { "epoch": 2.14, "grad_norm": 16.51693598173506, "learning_rate": 4.01664702216587e-06, "loss": 0.4849, "step": 13695 }, { "epoch": 2.14, "grad_norm": 14.843596417485822, "learning_rate": 4.015295887115169e-06, "loss": 0.5144, "step": 13696 }, { "epoch": 2.14, "grad_norm": 16.52286735209337, "learning_rate": 4.013944922265953e-06, "loss": 0.4234, "step": 13697 }, { "epoch": 2.14, "grad_norm": 24.705827313065996, "learning_rate": 4.012594127656646e-06, "loss": 0.4958, "step": 13698 }, { "epoch": 2.14, "grad_norm": 13.127258333558165, "learning_rate": 4.011243503325658e-06, "loss": 0.5063, "step": 13699 }, { "epoch": 2.14, "grad_norm": 24.312314186023546, "learning_rate": 4.009893049311399e-06, "loss": 0.4913, "step": 13700 }, { "epoch": 2.14, "grad_norm": 14.626372023357817, "learning_rate": 4.008542765652277e-06, "loss": 0.4502, "step": 13701 }, { "epoch": 2.14, "grad_norm": 21.61066268606607, "learning_rate": 4.007192652386695e-06, "loss": 0.5332, "step": 13702 }, { "epoch": 2.14, "grad_norm": 19.217029566634995, "learning_rate": 4.005842709553053e-06, "loss": 0.5055, "step": 13703 }, { "epoch": 2.14, "grad_norm": 31.327060903518884, "learning_rate": 4.0044929371897325e-06, "loss": 0.4788, "step": 13704 }, { "epoch": 2.14, "grad_norm": 18.38063009206433, "learning_rate": 4.003143335335124e-06, "loss": 0.4758, "step": 13705 }, { "epoch": 2.14, "grad_norm": 24.571766981610075, "learning_rate": 4.0017939040276136e-06, "loss": 0.4352, "step": 13706 }, { "epoch": 2.14, "grad_norm": 17.570630728357337, "learning_rate": 4.000444643305577e-06, "loss": 0.5377, "step": 13707 }, { "epoch": 2.14, "grad_norm": 20.537448221804265, "learning_rate": 3.999095553207386e-06, "loss": 0.4685, "step": 13708 }, { "epoch": 2.14, "grad_norm": 24.748541355499185, "learning_rate": 3.997746633771403e-06, "loss": 0.4966, "step": 13709 }, { "epoch": 2.14, "grad_norm": 20.882406352810595, "learning_rate": 3.996397885035995e-06, "loss": 0.4785, "step": 13710 }, { "epoch": 2.14, "grad_norm": 18.273152824164345, "learning_rate": 3.9950493070395235e-06, "loss": 0.4481, "step": 13711 }, { "epoch": 2.14, "grad_norm": 20.574758165891964, "learning_rate": 3.993700899820332e-06, "loss": 0.4331, "step": 13712 }, { "epoch": 2.14, "grad_norm": 21.10326420723819, "learning_rate": 3.9923526634167775e-06, "loss": 0.5087, "step": 13713 }, { "epoch": 2.14, "grad_norm": 20.482406304830853, "learning_rate": 3.991004597867195e-06, "loss": 0.5555, "step": 13714 }, { "epoch": 2.14, "grad_norm": 19.86213545143624, "learning_rate": 3.989656703209931e-06, "loss": 0.5258, "step": 13715 }, { "epoch": 2.14, "grad_norm": 20.263404000223726, "learning_rate": 3.9883089794833094e-06, "loss": 0.4341, "step": 13716 }, { "epoch": 2.14, "grad_norm": 14.18826280236325, "learning_rate": 3.986961426725665e-06, "loss": 0.5836, "step": 13717 }, { "epoch": 2.14, "grad_norm": 28.938851417639622, "learning_rate": 3.985614044975323e-06, "loss": 0.5013, "step": 13718 }, { "epoch": 2.14, "grad_norm": 15.860513615254497, "learning_rate": 3.9842668342705956e-06, "loss": 0.4221, "step": 13719 }, { "epoch": 2.14, "grad_norm": 19.91892328610376, "learning_rate": 3.982919794649805e-06, "loss": 0.5074, "step": 13720 }, { "epoch": 2.14, "grad_norm": 14.067444423764387, "learning_rate": 3.98157292615125e-06, "loss": 0.4327, "step": 13721 }, { "epoch": 2.14, "grad_norm": 16.54187269306718, "learning_rate": 3.980226228813242e-06, "loss": 0.4802, "step": 13722 }, { "epoch": 2.14, "grad_norm": 22.339117434908303, "learning_rate": 3.978879702674081e-06, "loss": 0.4776, "step": 13723 }, { "epoch": 2.14, "grad_norm": 17.754255318646006, "learning_rate": 3.97753334777206e-06, "loss": 0.4561, "step": 13724 }, { "epoch": 2.14, "grad_norm": 20.517606919861695, "learning_rate": 3.976187164145463e-06, "loss": 0.4778, "step": 13725 }, { "epoch": 2.14, "grad_norm": 17.43586719082391, "learning_rate": 3.97484115183258e-06, "loss": 0.4658, "step": 13726 }, { "epoch": 2.14, "grad_norm": 24.798569817845294, "learning_rate": 3.97349531087169e-06, "loss": 0.5521, "step": 13727 }, { "epoch": 2.14, "grad_norm": 25.420551036386378, "learning_rate": 3.972149641301075e-06, "loss": 0.4836, "step": 13728 }, { "epoch": 2.14, "grad_norm": 17.13754603500082, "learning_rate": 3.97080414315899e-06, "loss": 0.4835, "step": 13729 }, { "epoch": 2.14, "grad_norm": 14.881668349463343, "learning_rate": 3.969458816483709e-06, "loss": 0.4099, "step": 13730 }, { "epoch": 2.14, "grad_norm": 14.04242205769323, "learning_rate": 3.968113661313492e-06, "loss": 0.4811, "step": 13731 }, { "epoch": 2.14, "grad_norm": 18.945751120612176, "learning_rate": 3.966768677686597e-06, "loss": 0.5218, "step": 13732 }, { "epoch": 2.15, "grad_norm": 22.603341586719335, "learning_rate": 3.9654238656412715e-06, "loss": 0.434, "step": 13733 }, { "epoch": 2.15, "grad_norm": 22.39277913818106, "learning_rate": 3.9640792252157564e-06, "loss": 0.5095, "step": 13734 }, { "epoch": 2.15, "grad_norm": 27.45804126437764, "learning_rate": 3.962734756448299e-06, "loss": 0.4849, "step": 13735 }, { "epoch": 2.15, "grad_norm": 27.199613024017832, "learning_rate": 3.961390459377136e-06, "loss": 0.4819, "step": 13736 }, { "epoch": 2.15, "grad_norm": 15.847953142935125, "learning_rate": 3.960046334040496e-06, "loss": 0.4422, "step": 13737 }, { "epoch": 2.15, "grad_norm": 29.35578340017858, "learning_rate": 3.958702380476602e-06, "loss": 0.5015, "step": 13738 }, { "epoch": 2.15, "grad_norm": 18.535883657896086, "learning_rate": 3.957358598723677e-06, "loss": 0.4936, "step": 13739 }, { "epoch": 2.15, "grad_norm": 21.06538448052129, "learning_rate": 3.956014988819944e-06, "loss": 0.437, "step": 13740 }, { "epoch": 2.15, "grad_norm": 33.44518593485961, "learning_rate": 3.954671550803604e-06, "loss": 0.584, "step": 13741 }, { "epoch": 2.15, "grad_norm": 28.37576228800434, "learning_rate": 3.953328284712873e-06, "loss": 0.4868, "step": 13742 }, { "epoch": 2.15, "grad_norm": 24.72052935358145, "learning_rate": 3.951985190585944e-06, "loss": 0.4664, "step": 13743 }, { "epoch": 2.15, "grad_norm": 18.911912218378284, "learning_rate": 3.950642268461019e-06, "loss": 0.4844, "step": 13744 }, { "epoch": 2.15, "grad_norm": 27.67270996938472, "learning_rate": 3.9492995183762925e-06, "loss": 0.5117, "step": 13745 }, { "epoch": 2.15, "grad_norm": 34.96199081283332, "learning_rate": 3.947956940369944e-06, "loss": 0.519, "step": 13746 }, { "epoch": 2.15, "grad_norm": 25.21025821466923, "learning_rate": 3.946614534480164e-06, "loss": 0.5094, "step": 13747 }, { "epoch": 2.15, "grad_norm": 19.969785803733526, "learning_rate": 3.945272300745122e-06, "loss": 0.5673, "step": 13748 }, { "epoch": 2.15, "grad_norm": 18.7217589835283, "learning_rate": 3.943930239202996e-06, "loss": 0.4993, "step": 13749 }, { "epoch": 2.15, "grad_norm": 17.70408681628699, "learning_rate": 3.94258834989195e-06, "loss": 0.4487, "step": 13750 }, { "epoch": 2.15, "grad_norm": 17.773733450972227, "learning_rate": 3.941246632850146e-06, "loss": 0.4801, "step": 13751 }, { "epoch": 2.15, "grad_norm": 18.074960599819416, "learning_rate": 3.939905088115749e-06, "loss": 0.5416, "step": 13752 }, { "epoch": 2.15, "grad_norm": 13.103058469709426, "learning_rate": 3.938563715726904e-06, "loss": 0.4651, "step": 13753 }, { "epoch": 2.15, "grad_norm": 17.58381256575493, "learning_rate": 3.937222515721759e-06, "loss": 0.4427, "step": 13754 }, { "epoch": 2.15, "grad_norm": 12.944561510052417, "learning_rate": 3.935881488138459e-06, "loss": 0.4169, "step": 13755 }, { "epoch": 2.15, "grad_norm": 18.743133884370458, "learning_rate": 3.934540633015141e-06, "loss": 0.4329, "step": 13756 }, { "epoch": 2.15, "grad_norm": 18.08461541190231, "learning_rate": 3.933199950389944e-06, "loss": 0.4711, "step": 13757 }, { "epoch": 2.15, "grad_norm": 37.73339844539813, "learning_rate": 3.931859440300991e-06, "loss": 0.5231, "step": 13758 }, { "epoch": 2.15, "grad_norm": 16.505439799912597, "learning_rate": 3.930519102786402e-06, "loss": 0.4466, "step": 13759 }, { "epoch": 2.15, "grad_norm": 25.546909714888105, "learning_rate": 3.9291789378843e-06, "loss": 0.4311, "step": 13760 }, { "epoch": 2.15, "grad_norm": 24.2298276042473, "learning_rate": 3.927838945632799e-06, "loss": 0.5691, "step": 13761 }, { "epoch": 2.15, "grad_norm": 16.52777270267802, "learning_rate": 3.926499126070008e-06, "loss": 0.461, "step": 13762 }, { "epoch": 2.15, "grad_norm": 24.432323475238753, "learning_rate": 3.925159479234025e-06, "loss": 0.4728, "step": 13763 }, { "epoch": 2.15, "grad_norm": 15.988836755522414, "learning_rate": 3.923820005162954e-06, "loss": 0.4175, "step": 13764 }, { "epoch": 2.15, "grad_norm": 21.657416414800597, "learning_rate": 3.92248070389489e-06, "loss": 0.5605, "step": 13765 }, { "epoch": 2.15, "grad_norm": 17.00641292382174, "learning_rate": 3.9211415754679164e-06, "loss": 0.4619, "step": 13766 }, { "epoch": 2.15, "grad_norm": 23.691330852215604, "learning_rate": 3.919802619920122e-06, "loss": 0.492, "step": 13767 }, { "epoch": 2.15, "grad_norm": 26.29551294015667, "learning_rate": 3.9184638372895835e-06, "loss": 0.4238, "step": 13768 }, { "epoch": 2.15, "grad_norm": 38.15365430428977, "learning_rate": 3.9171252276143745e-06, "loss": 0.498, "step": 13769 }, { "epoch": 2.15, "grad_norm": 20.59322324956117, "learning_rate": 3.9157867909325684e-06, "loss": 0.4847, "step": 13770 }, { "epoch": 2.15, "grad_norm": 12.18020199795622, "learning_rate": 3.914448527282225e-06, "loss": 0.4729, "step": 13771 }, { "epoch": 2.15, "grad_norm": 16.175924663004682, "learning_rate": 3.9131104367014085e-06, "loss": 0.4657, "step": 13772 }, { "epoch": 2.15, "grad_norm": 15.820109088673199, "learning_rate": 3.9117725192281664e-06, "loss": 0.5235, "step": 13773 }, { "epoch": 2.15, "grad_norm": 30.926313180072196, "learning_rate": 3.910434774900555e-06, "loss": 0.452, "step": 13774 }, { "epoch": 2.15, "grad_norm": 18.734153041526948, "learning_rate": 3.909097203756615e-06, "loss": 0.4625, "step": 13775 }, { "epoch": 2.15, "grad_norm": 22.465166973454593, "learning_rate": 3.907759805834387e-06, "loss": 0.4865, "step": 13776 }, { "epoch": 2.15, "grad_norm": 17.328716967956478, "learning_rate": 3.906422581171908e-06, "loss": 0.4502, "step": 13777 }, { "epoch": 2.15, "grad_norm": 33.67026838819909, "learning_rate": 3.905085529807208e-06, "loss": 0.4552, "step": 13778 }, { "epoch": 2.15, "grad_norm": 18.702986331894113, "learning_rate": 3.903748651778306e-06, "loss": 0.5045, "step": 13779 }, { "epoch": 2.15, "grad_norm": 13.366622138092893, "learning_rate": 3.902411947123226e-06, "loss": 0.4295, "step": 13780 }, { "epoch": 2.15, "grad_norm": 18.022108940959637, "learning_rate": 3.901075415879987e-06, "loss": 0.4728, "step": 13781 }, { "epoch": 2.15, "grad_norm": 19.92704491944772, "learning_rate": 3.8997390580865915e-06, "loss": 0.5135, "step": 13782 }, { "epoch": 2.15, "grad_norm": 22.0508792601334, "learning_rate": 3.898402873781052e-06, "loss": 0.5274, "step": 13783 }, { "epoch": 2.15, "grad_norm": 15.433416757073212, "learning_rate": 3.897066863001363e-06, "loss": 0.4083, "step": 13784 }, { "epoch": 2.15, "grad_norm": 13.850326068728911, "learning_rate": 3.895731025785521e-06, "loss": 0.401, "step": 13785 }, { "epoch": 2.15, "grad_norm": 16.382674369902773, "learning_rate": 3.894395362171523e-06, "loss": 0.4531, "step": 13786 }, { "epoch": 2.15, "grad_norm": 23.78860367330921, "learning_rate": 3.893059872197347e-06, "loss": 0.5513, "step": 13787 }, { "epoch": 2.15, "grad_norm": 24.47919211095475, "learning_rate": 3.8917245559009735e-06, "loss": 0.4, "step": 13788 }, { "epoch": 2.15, "grad_norm": 24.180328592974686, "learning_rate": 3.89038941332038e-06, "loss": 0.4821, "step": 13789 }, { "epoch": 2.15, "grad_norm": 23.990397153291415, "learning_rate": 3.889054444493537e-06, "loss": 0.4787, "step": 13790 }, { "epoch": 2.15, "grad_norm": 17.76833099618584, "learning_rate": 3.887719649458419e-06, "loss": 0.4812, "step": 13791 }, { "epoch": 2.15, "grad_norm": 18.83301022059549, "learning_rate": 3.8863850282529715e-06, "loss": 0.5184, "step": 13792 }, { "epoch": 2.15, "grad_norm": 20.655549034598337, "learning_rate": 3.8850505809151574e-06, "loss": 0.5091, "step": 13793 }, { "epoch": 2.15, "grad_norm": 37.917152107933724, "learning_rate": 3.883716307482928e-06, "loss": 0.4166, "step": 13794 }, { "epoch": 2.15, "grad_norm": 18.995171467479597, "learning_rate": 3.882382207994231e-06, "loss": 0.4807, "step": 13795 }, { "epoch": 2.15, "grad_norm": 18.210267603273245, "learning_rate": 3.881048282487005e-06, "loss": 0.4251, "step": 13796 }, { "epoch": 2.16, "grad_norm": 16.194573550702966, "learning_rate": 3.879714530999185e-06, "loss": 0.5272, "step": 13797 }, { "epoch": 2.16, "grad_norm": 31.432147372145565, "learning_rate": 3.878380953568702e-06, "loss": 0.4632, "step": 13798 }, { "epoch": 2.16, "grad_norm": 16.33612165889988, "learning_rate": 3.877047550233488e-06, "loss": 0.4312, "step": 13799 }, { "epoch": 2.16, "grad_norm": 18.39905848497902, "learning_rate": 3.875714321031455e-06, "loss": 0.489, "step": 13800 }, { "epoch": 2.16, "grad_norm": 22.022634447605213, "learning_rate": 3.874381266000529e-06, "loss": 0.4236, "step": 13801 }, { "epoch": 2.16, "grad_norm": 22.33871689375885, "learning_rate": 3.873048385178613e-06, "loss": 0.4222, "step": 13802 }, { "epoch": 2.16, "grad_norm": 18.614947080571543, "learning_rate": 3.8717156786036194e-06, "loss": 0.4792, "step": 13803 }, { "epoch": 2.16, "grad_norm": 16.12220059720625, "learning_rate": 3.870383146313444e-06, "loss": 0.478, "step": 13804 }, { "epoch": 2.16, "grad_norm": 22.6458914932082, "learning_rate": 3.869050788345985e-06, "loss": 0.5189, "step": 13805 }, { "epoch": 2.16, "grad_norm": 16.606458162450263, "learning_rate": 3.86771860473914e-06, "loss": 0.5411, "step": 13806 }, { "epoch": 2.16, "grad_norm": 17.049167188300654, "learning_rate": 3.866386595530787e-06, "loss": 0.5451, "step": 13807 }, { "epoch": 2.16, "grad_norm": 17.226807028144304, "learning_rate": 3.865054760758813e-06, "loss": 0.462, "step": 13808 }, { "epoch": 2.16, "grad_norm": 23.703487373353195, "learning_rate": 3.863723100461091e-06, "loss": 0.5312, "step": 13809 }, { "epoch": 2.16, "grad_norm": 15.720219999366716, "learning_rate": 3.8623916146754935e-06, "loss": 0.4453, "step": 13810 }, { "epoch": 2.16, "grad_norm": 23.43698700216267, "learning_rate": 3.861060303439892e-06, "loss": 0.5466, "step": 13811 }, { "epoch": 2.16, "grad_norm": 28.113386443644814, "learning_rate": 3.859729166792144e-06, "loss": 0.5718, "step": 13812 }, { "epoch": 2.16, "grad_norm": 17.54450793873083, "learning_rate": 3.858398204770104e-06, "loss": 0.4871, "step": 13813 }, { "epoch": 2.16, "grad_norm": 18.94733574632788, "learning_rate": 3.8570674174116256e-06, "loss": 0.4339, "step": 13814 }, { "epoch": 2.16, "grad_norm": 14.756846983481807, "learning_rate": 3.855736804754556e-06, "loss": 0.4843, "step": 13815 }, { "epoch": 2.16, "grad_norm": 22.177096987878823, "learning_rate": 3.854406366836745e-06, "loss": 0.4191, "step": 13816 }, { "epoch": 2.16, "grad_norm": 27.469610137296737, "learning_rate": 3.853076103696015e-06, "loss": 0.4479, "step": 13817 }, { "epoch": 2.16, "grad_norm": 19.972465716415062, "learning_rate": 3.851746015370204e-06, "loss": 0.4721, "step": 13818 }, { "epoch": 2.16, "grad_norm": 21.57357047932348, "learning_rate": 3.8504161018971395e-06, "loss": 0.532, "step": 13819 }, { "epoch": 2.16, "grad_norm": 15.066666471030683, "learning_rate": 3.849086363314648e-06, "loss": 0.4783, "step": 13820 }, { "epoch": 2.16, "grad_norm": 21.251182363377925, "learning_rate": 3.847756799660543e-06, "loss": 0.5119, "step": 13821 }, { "epoch": 2.16, "grad_norm": 17.0781719236785, "learning_rate": 3.84642741097263e-06, "loss": 0.505, "step": 13822 }, { "epoch": 2.16, "grad_norm": 28.662032923425876, "learning_rate": 3.845098197288724e-06, "loss": 0.4965, "step": 13823 }, { "epoch": 2.16, "grad_norm": 15.116391376852972, "learning_rate": 3.843769158646626e-06, "loss": 0.4629, "step": 13824 }, { "epoch": 2.16, "grad_norm": 19.739721609921304, "learning_rate": 3.842440295084133e-06, "loss": 0.4611, "step": 13825 }, { "epoch": 2.16, "grad_norm": 22.773359891389546, "learning_rate": 3.841111606639032e-06, "loss": 0.5289, "step": 13826 }, { "epoch": 2.16, "grad_norm": 21.240824536072004, "learning_rate": 3.839783093349114e-06, "loss": 0.5214, "step": 13827 }, { "epoch": 2.16, "grad_norm": 19.626175148974685, "learning_rate": 3.838454755252166e-06, "loss": 0.5176, "step": 13828 }, { "epoch": 2.16, "grad_norm": 24.017673629666405, "learning_rate": 3.837126592385956e-06, "loss": 0.5024, "step": 13829 }, { "epoch": 2.16, "grad_norm": 16.998351240671198, "learning_rate": 3.8357986047882635e-06, "loss": 0.4954, "step": 13830 }, { "epoch": 2.16, "grad_norm": 14.535264263659025, "learning_rate": 3.8344707924968494e-06, "loss": 0.4635, "step": 13831 }, { "epoch": 2.16, "grad_norm": 29.739895736581154, "learning_rate": 3.833143155549479e-06, "loss": 0.5204, "step": 13832 }, { "epoch": 2.16, "grad_norm": 26.93640113567417, "learning_rate": 3.831815693983915e-06, "loss": 0.5732, "step": 13833 }, { "epoch": 2.16, "grad_norm": 29.060360410338088, "learning_rate": 3.830488407837899e-06, "loss": 0.4942, "step": 13834 }, { "epoch": 2.16, "grad_norm": 34.74055317559641, "learning_rate": 3.829161297149187e-06, "loss": 0.4653, "step": 13835 }, { "epoch": 2.16, "grad_norm": 28.806554128051562, "learning_rate": 3.8278343619555155e-06, "loss": 0.5492, "step": 13836 }, { "epoch": 2.16, "grad_norm": 26.49196965975394, "learning_rate": 3.826507602294628e-06, "loss": 0.5491, "step": 13837 }, { "epoch": 2.16, "grad_norm": 22.616057039250588, "learning_rate": 3.825181018204248e-06, "loss": 0.5572, "step": 13838 }, { "epoch": 2.16, "grad_norm": 16.009056243128285, "learning_rate": 3.823854609722108e-06, "loss": 0.5575, "step": 13839 }, { "epoch": 2.16, "grad_norm": 22.919704019155592, "learning_rate": 3.8225283768859345e-06, "loss": 0.5026, "step": 13840 }, { "epoch": 2.16, "grad_norm": 16.00686820611522, "learning_rate": 3.8212023197334395e-06, "loss": 0.5281, "step": 13841 }, { "epoch": 2.16, "grad_norm": 23.563345590388938, "learning_rate": 3.819876438302332e-06, "loss": 0.5126, "step": 13842 }, { "epoch": 2.16, "grad_norm": 17.395522280632267, "learning_rate": 3.8185507326303236e-06, "loss": 0.4583, "step": 13843 }, { "epoch": 2.16, "grad_norm": 23.343331406311542, "learning_rate": 3.817225202755117e-06, "loss": 0.4793, "step": 13844 }, { "epoch": 2.16, "grad_norm": 21.359305721413385, "learning_rate": 3.815899848714412e-06, "loss": 0.4604, "step": 13845 }, { "epoch": 2.16, "grad_norm": 35.65342793509411, "learning_rate": 3.8145746705458976e-06, "loss": 0.5287, "step": 13846 }, { "epoch": 2.16, "grad_norm": 24.1185799540555, "learning_rate": 3.8132496682872576e-06, "loss": 0.4408, "step": 13847 }, { "epoch": 2.16, "grad_norm": 20.168136318174888, "learning_rate": 3.811924841976178e-06, "loss": 0.4905, "step": 13848 }, { "epoch": 2.16, "grad_norm": 30.395899087003755, "learning_rate": 3.81060019165034e-06, "loss": 0.4961, "step": 13849 }, { "epoch": 2.16, "grad_norm": 26.04443045554107, "learning_rate": 3.8092757173474114e-06, "loss": 0.4972, "step": 13850 }, { "epoch": 2.16, "grad_norm": 19.447656623516373, "learning_rate": 3.807951419105056e-06, "loss": 0.4863, "step": 13851 }, { "epoch": 2.16, "grad_norm": 23.089494493786223, "learning_rate": 3.8066272969609407e-06, "loss": 0.499, "step": 13852 }, { "epoch": 2.16, "grad_norm": 13.885366479673738, "learning_rate": 3.8053033509527214e-06, "loss": 0.4918, "step": 13853 }, { "epoch": 2.16, "grad_norm": 21.0940775160618, "learning_rate": 3.8039795811180547e-06, "loss": 0.5497, "step": 13854 }, { "epoch": 2.16, "grad_norm": 30.623567587576517, "learning_rate": 3.8026559874945844e-06, "loss": 0.4488, "step": 13855 }, { "epoch": 2.16, "grad_norm": 23.204946987146617, "learning_rate": 3.801332570119949e-06, "loss": 0.523, "step": 13856 }, { "epoch": 2.16, "grad_norm": 33.90051573857576, "learning_rate": 3.8000093290317886e-06, "loss": 0.5347, "step": 13857 }, { "epoch": 2.16, "grad_norm": 27.360146308496923, "learning_rate": 3.7986862642677402e-06, "loss": 0.5333, "step": 13858 }, { "epoch": 2.16, "grad_norm": 18.113558550125575, "learning_rate": 3.7973633758654225e-06, "loss": 0.5234, "step": 13859 }, { "epoch": 2.16, "grad_norm": 18.666177408655454, "learning_rate": 3.796040663862467e-06, "loss": 0.5465, "step": 13860 }, { "epoch": 2.17, "grad_norm": 24.423976211983565, "learning_rate": 3.7947181282964806e-06, "loss": 0.5468, "step": 13861 }, { "epoch": 2.17, "grad_norm": 16.03031949267016, "learning_rate": 3.793395769205085e-06, "loss": 0.4362, "step": 13862 }, { "epoch": 2.17, "grad_norm": 27.25323634877798, "learning_rate": 3.7920735866258796e-06, "loss": 0.4984, "step": 13863 }, { "epoch": 2.17, "grad_norm": 20.221239961071152, "learning_rate": 3.790751580596469e-06, "loss": 0.5372, "step": 13864 }, { "epoch": 2.17, "grad_norm": 27.802734656379144, "learning_rate": 3.7894297511544553e-06, "loss": 0.5342, "step": 13865 }, { "epoch": 2.17, "grad_norm": 17.28318000315127, "learning_rate": 3.7881080983374263e-06, "loss": 0.4139, "step": 13866 }, { "epoch": 2.17, "grad_norm": 23.103619880902194, "learning_rate": 3.786786622182966e-06, "loss": 0.5184, "step": 13867 }, { "epoch": 2.17, "grad_norm": 14.232947886224364, "learning_rate": 3.7854653227286586e-06, "loss": 0.4749, "step": 13868 }, { "epoch": 2.17, "grad_norm": 23.69926999905366, "learning_rate": 3.784144200012083e-06, "loss": 0.5395, "step": 13869 }, { "epoch": 2.17, "grad_norm": 26.86781979166525, "learning_rate": 3.782823254070813e-06, "loss": 0.4364, "step": 13870 }, { "epoch": 2.17, "grad_norm": 22.672117333855482, "learning_rate": 3.781502484942413e-06, "loss": 0.5121, "step": 13871 }, { "epoch": 2.17, "grad_norm": 30.376665610301327, "learning_rate": 3.7801818926644416e-06, "loss": 0.5527, "step": 13872 }, { "epoch": 2.17, "grad_norm": 18.870592599040634, "learning_rate": 3.778861477274458e-06, "loss": 0.4935, "step": 13873 }, { "epoch": 2.17, "grad_norm": 16.214417900681852, "learning_rate": 3.7775412388100186e-06, "loss": 0.4581, "step": 13874 }, { "epoch": 2.17, "grad_norm": 13.50413907142526, "learning_rate": 3.7762211773086656e-06, "loss": 0.4296, "step": 13875 }, { "epoch": 2.17, "grad_norm": 17.086538747282624, "learning_rate": 3.7749012928079387e-06, "loss": 0.4754, "step": 13876 }, { "epoch": 2.17, "grad_norm": 18.153668502430836, "learning_rate": 3.773581585345377e-06, "loss": 0.518, "step": 13877 }, { "epoch": 2.17, "grad_norm": 14.565319608962314, "learning_rate": 3.772262054958513e-06, "loss": 0.4488, "step": 13878 }, { "epoch": 2.17, "grad_norm": 19.002450101918896, "learning_rate": 3.77094270168488e-06, "loss": 0.5108, "step": 13879 }, { "epoch": 2.17, "grad_norm": 21.186942853466455, "learning_rate": 3.769623525561986e-06, "loss": 0.4658, "step": 13880 }, { "epoch": 2.17, "grad_norm": 23.06550099585411, "learning_rate": 3.768304526627354e-06, "loss": 0.4874, "step": 13881 }, { "epoch": 2.17, "grad_norm": 22.429163810757675, "learning_rate": 3.7669857049184953e-06, "loss": 0.5271, "step": 13882 }, { "epoch": 2.17, "grad_norm": 17.727649414142682, "learning_rate": 3.765667060472922e-06, "loss": 0.4413, "step": 13883 }, { "epoch": 2.17, "grad_norm": 27.112849995071905, "learning_rate": 3.764348593328129e-06, "loss": 0.4824, "step": 13884 }, { "epoch": 2.17, "grad_norm": 21.889569403433157, "learning_rate": 3.763030303521611e-06, "loss": 0.5256, "step": 13885 }, { "epoch": 2.17, "grad_norm": 14.050841578719153, "learning_rate": 3.7617121910908627e-06, "loss": 0.4345, "step": 13886 }, { "epoch": 2.17, "grad_norm": 22.25122012117449, "learning_rate": 3.760394256073374e-06, "loss": 0.4876, "step": 13887 }, { "epoch": 2.17, "grad_norm": 19.327057029496785, "learning_rate": 3.7590764985066187e-06, "loss": 0.4769, "step": 13888 }, { "epoch": 2.17, "grad_norm": 12.126113154129019, "learning_rate": 3.7577589184280817e-06, "loss": 0.4417, "step": 13889 }, { "epoch": 2.17, "grad_norm": 23.267813302315506, "learning_rate": 3.7564415158752244e-06, "loss": 0.5399, "step": 13890 }, { "epoch": 2.17, "grad_norm": 15.934242441075234, "learning_rate": 3.755124290885522e-06, "loss": 0.4572, "step": 13891 }, { "epoch": 2.17, "grad_norm": 25.72777193854207, "learning_rate": 3.753807243496429e-06, "loss": 0.5363, "step": 13892 }, { "epoch": 2.17, "grad_norm": 18.68732316912678, "learning_rate": 3.752490373745403e-06, "loss": 0.4636, "step": 13893 }, { "epoch": 2.17, "grad_norm": 18.195430534114255, "learning_rate": 3.751173681669901e-06, "loss": 0.4551, "step": 13894 }, { "epoch": 2.17, "grad_norm": 25.770825307151473, "learning_rate": 3.749857167307359e-06, "loss": 0.492, "step": 13895 }, { "epoch": 2.17, "grad_norm": 19.805678432736446, "learning_rate": 3.7485408306952263e-06, "loss": 0.4557, "step": 13896 }, { "epoch": 2.17, "grad_norm": 22.36118923347034, "learning_rate": 3.7472246718709325e-06, "loss": 0.4547, "step": 13897 }, { "epoch": 2.17, "grad_norm": 16.697549062766893, "learning_rate": 3.7459086908719124e-06, "loss": 0.4868, "step": 13898 }, { "epoch": 2.17, "grad_norm": 18.198767491674225, "learning_rate": 3.744592887735592e-06, "loss": 0.4764, "step": 13899 }, { "epoch": 2.17, "grad_norm": 26.00196618316421, "learning_rate": 3.7432772624993917e-06, "loss": 0.4639, "step": 13900 }, { "epoch": 2.17, "grad_norm": 25.156482588836, "learning_rate": 3.741961815200723e-06, "loss": 0.4867, "step": 13901 }, { "epoch": 2.17, "grad_norm": 20.34395287012653, "learning_rate": 3.740646545876998e-06, "loss": 0.493, "step": 13902 }, { "epoch": 2.17, "grad_norm": 17.052540452182416, "learning_rate": 3.7393314545656243e-06, "loss": 0.5237, "step": 13903 }, { "epoch": 2.17, "grad_norm": 27.072504637392218, "learning_rate": 3.7380165413040092e-06, "loss": 0.4997, "step": 13904 }, { "epoch": 2.17, "grad_norm": 22.62259429581696, "learning_rate": 3.7367018061295335e-06, "loss": 0.5068, "step": 13905 }, { "epoch": 2.17, "grad_norm": 16.021429367556948, "learning_rate": 3.735387249079594e-06, "loss": 0.415, "step": 13906 }, { "epoch": 2.17, "grad_norm": 20.917348662659062, "learning_rate": 3.7340728701915764e-06, "loss": 0.4499, "step": 13907 }, { "epoch": 2.17, "grad_norm": 15.856144422902132, "learning_rate": 3.732758669502864e-06, "loss": 0.4747, "step": 13908 }, { "epoch": 2.17, "grad_norm": 12.802182053117328, "learning_rate": 3.731444647050829e-06, "loss": 0.478, "step": 13909 }, { "epoch": 2.17, "grad_norm": 18.648342386425874, "learning_rate": 3.730130802872839e-06, "loss": 0.5104, "step": 13910 }, { "epoch": 2.17, "grad_norm": 22.40224088064852, "learning_rate": 3.72881713700626e-06, "loss": 0.4798, "step": 13911 }, { "epoch": 2.17, "grad_norm": 16.22119826652589, "learning_rate": 3.7275036494884563e-06, "loss": 0.4925, "step": 13912 }, { "epoch": 2.17, "grad_norm": 18.040495374066186, "learning_rate": 3.7261903403567767e-06, "loss": 0.4351, "step": 13913 }, { "epoch": 2.17, "grad_norm": 18.476561855614552, "learning_rate": 3.7248772096485773e-06, "loss": 0.4969, "step": 13914 }, { "epoch": 2.17, "grad_norm": 29.656008212929844, "learning_rate": 3.7235642574011955e-06, "loss": 0.5233, "step": 13915 }, { "epoch": 2.17, "grad_norm": 18.42924940373642, "learning_rate": 3.7222514836519784e-06, "loss": 0.418, "step": 13916 }, { "epoch": 2.17, "grad_norm": 19.142868668758883, "learning_rate": 3.7209388884382526e-06, "loss": 0.4841, "step": 13917 }, { "epoch": 2.17, "grad_norm": 18.182305342595264, "learning_rate": 3.719626471797352e-06, "loss": 0.5755, "step": 13918 }, { "epoch": 2.17, "grad_norm": 16.20518447272807, "learning_rate": 3.7183142337666045e-06, "loss": 0.4834, "step": 13919 }, { "epoch": 2.17, "grad_norm": 17.57457942554384, "learning_rate": 3.7170021743833216e-06, "loss": 0.3893, "step": 13920 }, { "epoch": 2.17, "grad_norm": 21.992605063775144, "learning_rate": 3.715690293684825e-06, "loss": 0.4778, "step": 13921 }, { "epoch": 2.17, "grad_norm": 15.958230542950803, "learning_rate": 3.7143785917084163e-06, "loss": 0.3823, "step": 13922 }, { "epoch": 2.17, "grad_norm": 14.730368526606386, "learning_rate": 3.7130670684914083e-06, "loss": 0.5007, "step": 13923 }, { "epoch": 2.17, "grad_norm": 28.378319004228697, "learning_rate": 3.71175572407109e-06, "loss": 0.5193, "step": 13924 }, { "epoch": 2.18, "grad_norm": 17.744855643999525, "learning_rate": 3.7104445584847647e-06, "loss": 0.442, "step": 13925 }, { "epoch": 2.18, "grad_norm": 18.921790652791913, "learning_rate": 3.709133571769713e-06, "loss": 0.4352, "step": 13926 }, { "epoch": 2.18, "grad_norm": 31.174525690791548, "learning_rate": 3.7078227639632234e-06, "loss": 0.5035, "step": 13927 }, { "epoch": 2.18, "grad_norm": 14.620553649009429, "learning_rate": 3.7065121351025758e-06, "loss": 0.4392, "step": 13928 }, { "epoch": 2.18, "grad_norm": 29.90004100392075, "learning_rate": 3.705201685225043e-06, "loss": 0.6798, "step": 13929 }, { "epoch": 2.18, "grad_norm": 17.527296288113387, "learning_rate": 3.7038914143678874e-06, "loss": 0.4699, "step": 13930 }, { "epoch": 2.18, "grad_norm": 22.30784560191277, "learning_rate": 3.702581322568376e-06, "loss": 0.5408, "step": 13931 }, { "epoch": 2.18, "grad_norm": 15.861843915730232, "learning_rate": 3.701271409863769e-06, "loss": 0.4325, "step": 13932 }, { "epoch": 2.18, "grad_norm": 18.81794535048324, "learning_rate": 3.6999616762913226e-06, "loss": 0.4409, "step": 13933 }, { "epoch": 2.18, "grad_norm": 22.7578377024586, "learning_rate": 3.6986521218882798e-06, "loss": 0.5059, "step": 13934 }, { "epoch": 2.18, "grad_norm": 13.806213747355129, "learning_rate": 3.697342746691881e-06, "loss": 0.4334, "step": 13935 }, { "epoch": 2.18, "grad_norm": 26.173465803394198, "learning_rate": 3.6960335507393672e-06, "loss": 0.5271, "step": 13936 }, { "epoch": 2.18, "grad_norm": 16.64270022441161, "learning_rate": 3.694724534067976e-06, "loss": 0.4328, "step": 13937 }, { "epoch": 2.18, "grad_norm": 18.916564799516234, "learning_rate": 3.693415696714929e-06, "loss": 0.45, "step": 13938 }, { "epoch": 2.18, "grad_norm": 14.592918441737208, "learning_rate": 3.6921070387174484e-06, "loss": 0.5372, "step": 13939 }, { "epoch": 2.18, "grad_norm": 17.651884181892083, "learning_rate": 3.690798560112754e-06, "loss": 0.4198, "step": 13940 }, { "epoch": 2.18, "grad_norm": 16.509405655447036, "learning_rate": 3.6894902609380566e-06, "loss": 0.5016, "step": 13941 }, { "epoch": 2.18, "grad_norm": 18.27971133497859, "learning_rate": 3.6881821412305683e-06, "loss": 0.4705, "step": 13942 }, { "epoch": 2.18, "grad_norm": 16.39685677519553, "learning_rate": 3.686874201027487e-06, "loss": 0.4337, "step": 13943 }, { "epoch": 2.18, "grad_norm": 24.002955028784868, "learning_rate": 3.6855664403660073e-06, "loss": 0.4793, "step": 13944 }, { "epoch": 2.18, "grad_norm": 32.031929770676626, "learning_rate": 3.684258859283325e-06, "loss": 0.4854, "step": 13945 }, { "epoch": 2.18, "grad_norm": 15.166517287957952, "learning_rate": 3.6829514578166293e-06, "loss": 0.433, "step": 13946 }, { "epoch": 2.18, "grad_norm": 26.03985727769971, "learning_rate": 3.6816442360030946e-06, "loss": 0.5459, "step": 13947 }, { "epoch": 2.18, "grad_norm": 23.580314624301515, "learning_rate": 3.680337193879906e-06, "loss": 0.4546, "step": 13948 }, { "epoch": 2.18, "grad_norm": 14.81311772462806, "learning_rate": 3.679030331484227e-06, "loss": 0.4476, "step": 13949 }, { "epoch": 2.18, "grad_norm": 15.472065570750203, "learning_rate": 3.677723648853231e-06, "loss": 0.4231, "step": 13950 }, { "epoch": 2.18, "grad_norm": 20.031267716612696, "learning_rate": 3.676417146024073e-06, "loss": 0.4422, "step": 13951 }, { "epoch": 2.18, "grad_norm": 24.97955650829259, "learning_rate": 3.6751108230339116e-06, "loss": 0.4984, "step": 13952 }, { "epoch": 2.18, "grad_norm": 19.24901122721355, "learning_rate": 3.673804679919901e-06, "loss": 0.5214, "step": 13953 }, { "epoch": 2.18, "grad_norm": 15.329999098499357, "learning_rate": 3.672498716719186e-06, "loss": 0.4278, "step": 13954 }, { "epoch": 2.18, "grad_norm": 29.249313411642508, "learning_rate": 3.6711929334689024e-06, "loss": 0.5053, "step": 13955 }, { "epoch": 2.18, "grad_norm": 19.4968305344219, "learning_rate": 3.6698873302061888e-06, "loss": 0.505, "step": 13956 }, { "epoch": 2.18, "grad_norm": 16.45472467240973, "learning_rate": 3.668581906968177e-06, "loss": 0.4573, "step": 13957 }, { "epoch": 2.18, "grad_norm": 19.04060933564639, "learning_rate": 3.6672766637919967e-06, "loss": 0.4659, "step": 13958 }, { "epoch": 2.18, "grad_norm": 24.05959855520941, "learning_rate": 3.6659716007147616e-06, "loss": 0.5075, "step": 13959 }, { "epoch": 2.18, "grad_norm": 18.095191734239712, "learning_rate": 3.664666717773586e-06, "loss": 0.5157, "step": 13960 }, { "epoch": 2.18, "grad_norm": 16.977994121589116, "learning_rate": 3.6633620150055835e-06, "loss": 0.4636, "step": 13961 }, { "epoch": 2.18, "grad_norm": 13.195037928153063, "learning_rate": 3.6620574924478615e-06, "loss": 0.4557, "step": 13962 }, { "epoch": 2.18, "grad_norm": 20.064549108242435, "learning_rate": 3.660753150137517e-06, "loss": 0.4572, "step": 13963 }, { "epoch": 2.18, "grad_norm": 19.54747176460059, "learning_rate": 3.65944898811164e-06, "loss": 0.5224, "step": 13964 }, { "epoch": 2.18, "grad_norm": 25.491890843892687, "learning_rate": 3.6581450064073265e-06, "loss": 0.4815, "step": 13965 }, { "epoch": 2.18, "grad_norm": 18.26828795487254, "learning_rate": 3.6568412050616574e-06, "loss": 0.5211, "step": 13966 }, { "epoch": 2.18, "grad_norm": 17.160762294224657, "learning_rate": 3.655537584111718e-06, "loss": 0.4263, "step": 13967 }, { "epoch": 2.18, "grad_norm": 20.093935823248746, "learning_rate": 3.654234143594578e-06, "loss": 0.458, "step": 13968 }, { "epoch": 2.18, "grad_norm": 19.317338236743893, "learning_rate": 3.652930883547304e-06, "loss": 0.4909, "step": 13969 }, { "epoch": 2.18, "grad_norm": 18.61116934100178, "learning_rate": 3.651627804006963e-06, "loss": 0.4267, "step": 13970 }, { "epoch": 2.18, "grad_norm": 18.55893812941986, "learning_rate": 3.650324905010616e-06, "loss": 0.4397, "step": 13971 }, { "epoch": 2.18, "grad_norm": 27.044436155293575, "learning_rate": 3.6490221865953146e-06, "loss": 0.5589, "step": 13972 }, { "epoch": 2.18, "grad_norm": 19.380076979906473, "learning_rate": 3.6477196487981036e-06, "loss": 0.5088, "step": 13973 }, { "epoch": 2.18, "grad_norm": 23.389732167335104, "learning_rate": 3.6464172916560305e-06, "loss": 0.4705, "step": 13974 }, { "epoch": 2.18, "grad_norm": 17.589091577782902, "learning_rate": 3.645115115206136e-06, "loss": 0.4937, "step": 13975 }, { "epoch": 2.18, "grad_norm": 29.01265602554712, "learning_rate": 3.643813119485445e-06, "loss": 0.4521, "step": 13976 }, { "epoch": 2.18, "grad_norm": 13.64416319719017, "learning_rate": 3.642511304530996e-06, "loss": 0.4766, "step": 13977 }, { "epoch": 2.18, "grad_norm": 18.352320732961957, "learning_rate": 3.641209670379803e-06, "loss": 0.4317, "step": 13978 }, { "epoch": 2.18, "grad_norm": 19.73647408334706, "learning_rate": 3.6399082170688893e-06, "loss": 0.5278, "step": 13979 }, { "epoch": 2.18, "grad_norm": 17.12766188186991, "learning_rate": 3.6386069446352632e-06, "loss": 0.5229, "step": 13980 }, { "epoch": 2.18, "grad_norm": 24.355657629092867, "learning_rate": 3.6373058531159332e-06, "loss": 0.4636, "step": 13981 }, { "epoch": 2.18, "grad_norm": 20.65096667990874, "learning_rate": 3.636004942547907e-06, "loss": 0.4917, "step": 13982 }, { "epoch": 2.18, "grad_norm": 24.37103479607225, "learning_rate": 3.634704212968174e-06, "loss": 0.4963, "step": 13983 }, { "epoch": 2.18, "grad_norm": 15.788994113711732, "learning_rate": 3.6334036644137328e-06, "loss": 0.4964, "step": 13984 }, { "epoch": 2.18, "grad_norm": 16.224723738366823, "learning_rate": 3.632103296921563e-06, "loss": 0.4695, "step": 13985 }, { "epoch": 2.18, "grad_norm": 19.8523101380616, "learning_rate": 3.630803110528651e-06, "loss": 0.5534, "step": 13986 }, { "epoch": 2.18, "grad_norm": 19.84165634201634, "learning_rate": 3.6295031052719766e-06, "loss": 0.5135, "step": 13987 }, { "epoch": 2.18, "grad_norm": 23.956733448610258, "learning_rate": 3.628203281188506e-06, "loss": 0.4013, "step": 13988 }, { "epoch": 2.19, "grad_norm": 35.406843996915335, "learning_rate": 3.6269036383152032e-06, "loss": 0.5603, "step": 13989 }, { "epoch": 2.19, "grad_norm": 23.49508580228283, "learning_rate": 3.6256041766890337e-06, "loss": 0.4794, "step": 13990 }, { "epoch": 2.19, "grad_norm": 25.701856949435804, "learning_rate": 3.624304896346953e-06, "loss": 0.43, "step": 13991 }, { "epoch": 2.19, "grad_norm": 26.33248877711425, "learning_rate": 3.623005797325917e-06, "loss": 0.4812, "step": 13992 }, { "epoch": 2.19, "grad_norm": 16.093145042456435, "learning_rate": 3.6217068796628597e-06, "loss": 0.4626, "step": 13993 }, { "epoch": 2.19, "grad_norm": 22.07658711791868, "learning_rate": 3.6204081433947268e-06, "loss": 0.4784, "step": 13994 }, { "epoch": 2.19, "grad_norm": 18.614420989407897, "learning_rate": 3.619109588558455e-06, "loss": 0.5323, "step": 13995 }, { "epoch": 2.19, "grad_norm": 23.058644835738473, "learning_rate": 3.6178112151909763e-06, "loss": 0.4095, "step": 13996 }, { "epoch": 2.19, "grad_norm": 19.812510704389624, "learning_rate": 3.6165130233292133e-06, "loss": 0.5238, "step": 13997 }, { "epoch": 2.19, "grad_norm": 23.198471185753913, "learning_rate": 3.6152150130100825e-06, "loss": 0.5558, "step": 13998 }, { "epoch": 2.19, "grad_norm": 25.21144552460755, "learning_rate": 3.6139171842705024e-06, "loss": 0.5057, "step": 13999 }, { "epoch": 2.19, "grad_norm": 22.192608312420063, "learning_rate": 3.612619537147385e-06, "loss": 0.5618, "step": 14000 }, { "epoch": 2.19, "grad_norm": 15.8332970868708, "learning_rate": 3.611322071677628e-06, "loss": 0.477, "step": 14001 }, { "epoch": 2.19, "grad_norm": 14.833045054788366, "learning_rate": 3.6100247878981365e-06, "loss": 0.389, "step": 14002 }, { "epoch": 2.19, "grad_norm": 15.222395402544583, "learning_rate": 3.6087276858458e-06, "loss": 0.4335, "step": 14003 }, { "epoch": 2.19, "grad_norm": 24.119977674108185, "learning_rate": 3.607430765557508e-06, "loss": 0.5015, "step": 14004 }, { "epoch": 2.19, "grad_norm": 14.618093805031508, "learning_rate": 3.606134027070151e-06, "loss": 0.4526, "step": 14005 }, { "epoch": 2.19, "grad_norm": 15.710519401435027, "learning_rate": 3.604837470420596e-06, "loss": 0.4316, "step": 14006 }, { "epoch": 2.19, "grad_norm": 16.891181656083717, "learning_rate": 3.603541095645727e-06, "loss": 0.4568, "step": 14007 }, { "epoch": 2.19, "grad_norm": 27.573844077065104, "learning_rate": 3.6022449027824035e-06, "loss": 0.4954, "step": 14008 }, { "epoch": 2.19, "grad_norm": 20.6284477473315, "learning_rate": 3.600948891867496e-06, "loss": 0.5733, "step": 14009 }, { "epoch": 2.19, "grad_norm": 18.12823187197001, "learning_rate": 3.599653062937856e-06, "loss": 0.5448, "step": 14010 }, { "epoch": 2.19, "grad_norm": 22.564854645638107, "learning_rate": 3.598357416030338e-06, "loss": 0.4427, "step": 14011 }, { "epoch": 2.19, "grad_norm": 21.107277889542466, "learning_rate": 3.597061951181794e-06, "loss": 0.498, "step": 14012 }, { "epoch": 2.19, "grad_norm": 24.69182937783355, "learning_rate": 3.5957666684290626e-06, "loss": 0.508, "step": 14013 }, { "epoch": 2.19, "grad_norm": 18.85321189709224, "learning_rate": 3.594471567808977e-06, "loss": 0.4794, "step": 14014 }, { "epoch": 2.19, "grad_norm": 16.42785927445323, "learning_rate": 3.593176649358373e-06, "loss": 0.4412, "step": 14015 }, { "epoch": 2.19, "grad_norm": 34.66148731861961, "learning_rate": 3.591881913114077e-06, "loss": 0.4727, "step": 14016 }, { "epoch": 2.19, "grad_norm": 17.502460511120397, "learning_rate": 3.590587359112917e-06, "loss": 0.4875, "step": 14017 }, { "epoch": 2.19, "grad_norm": 14.031174962728825, "learning_rate": 3.589292987391697e-06, "loss": 0.4412, "step": 14018 }, { "epoch": 2.19, "grad_norm": 24.394031926978407, "learning_rate": 3.5879987979872342e-06, "loss": 0.5227, "step": 14019 }, { "epoch": 2.19, "grad_norm": 17.69219731579314, "learning_rate": 3.586704790936335e-06, "loss": 0.4277, "step": 14020 }, { "epoch": 2.19, "grad_norm": 17.419275426006198, "learning_rate": 3.5854109662758043e-06, "loss": 0.4646, "step": 14021 }, { "epoch": 2.19, "grad_norm": 20.173975894212102, "learning_rate": 3.584117324042431e-06, "loss": 0.4418, "step": 14022 }, { "epoch": 2.19, "grad_norm": 24.775346166971545, "learning_rate": 3.5828238642730063e-06, "loss": 0.5118, "step": 14023 }, { "epoch": 2.19, "grad_norm": 22.174961773109384, "learning_rate": 3.581530587004316e-06, "loss": 0.4566, "step": 14024 }, { "epoch": 2.19, "grad_norm": 19.649343930034874, "learning_rate": 3.580237492273144e-06, "loss": 0.5392, "step": 14025 }, { "epoch": 2.19, "grad_norm": 18.97642455476868, "learning_rate": 3.578944580116264e-06, "loss": 0.4857, "step": 14026 }, { "epoch": 2.19, "grad_norm": 29.003735713778482, "learning_rate": 3.5776518505704383e-06, "loss": 0.5662, "step": 14027 }, { "epoch": 2.19, "grad_norm": 21.726996247050618, "learning_rate": 3.5763593036724387e-06, "loss": 0.4362, "step": 14028 }, { "epoch": 2.19, "grad_norm": 24.769312331371186, "learning_rate": 3.575066939459022e-06, "loss": 0.5074, "step": 14029 }, { "epoch": 2.19, "grad_norm": 18.008584782554852, "learning_rate": 3.5737747579669468e-06, "loss": 0.4826, "step": 14030 }, { "epoch": 2.19, "grad_norm": 31.0675369195081, "learning_rate": 3.572482759232958e-06, "loss": 0.5604, "step": 14031 }, { "epoch": 2.19, "grad_norm": 18.034944622705922, "learning_rate": 3.5711909432937964e-06, "loss": 0.4204, "step": 14032 }, { "epoch": 2.19, "grad_norm": 14.918155319393698, "learning_rate": 3.5698993101862034e-06, "loss": 0.4572, "step": 14033 }, { "epoch": 2.19, "grad_norm": 15.828490565598276, "learning_rate": 3.5686078599469166e-06, "loss": 0.4274, "step": 14034 }, { "epoch": 2.19, "grad_norm": 20.154704688221837, "learning_rate": 3.5673165926126553e-06, "loss": 0.4479, "step": 14035 }, { "epoch": 2.19, "grad_norm": 17.335644179274187, "learning_rate": 3.566025508220151e-06, "loss": 0.4564, "step": 14036 }, { "epoch": 2.19, "grad_norm": 20.158496724080653, "learning_rate": 3.5647346068061152e-06, "loss": 0.4476, "step": 14037 }, { "epoch": 2.19, "grad_norm": 18.988967250210617, "learning_rate": 3.5634438884072653e-06, "loss": 0.4973, "step": 14038 }, { "epoch": 2.19, "grad_norm": 17.731892454006914, "learning_rate": 3.5621533530603024e-06, "loss": 0.5157, "step": 14039 }, { "epoch": 2.19, "grad_norm": 24.45232979270624, "learning_rate": 3.560863000801932e-06, "loss": 0.4275, "step": 14040 }, { "epoch": 2.19, "grad_norm": 21.361726270071014, "learning_rate": 3.559572831668855e-06, "loss": 0.4794, "step": 14041 }, { "epoch": 2.19, "grad_norm": 20.57647011807346, "learning_rate": 3.558282845697758e-06, "loss": 0.4836, "step": 14042 }, { "epoch": 2.19, "grad_norm": 16.13212645690718, "learning_rate": 3.5569930429253263e-06, "loss": 0.4892, "step": 14043 }, { "epoch": 2.19, "grad_norm": 18.072692573341673, "learning_rate": 3.5557034233882426e-06, "loss": 0.4775, "step": 14044 }, { "epoch": 2.19, "grad_norm": 20.109924302234905, "learning_rate": 3.5544139871231842e-06, "loss": 0.4951, "step": 14045 }, { "epoch": 2.19, "grad_norm": 26.79231211125581, "learning_rate": 3.5531247341668253e-06, "loss": 0.4937, "step": 14046 }, { "epoch": 2.19, "grad_norm": 19.738250184690546, "learning_rate": 3.551835664555827e-06, "loss": 0.5187, "step": 14047 }, { "epoch": 2.19, "grad_norm": 19.87518642146174, "learning_rate": 3.5505467783268465e-06, "loss": 0.3845, "step": 14048 }, { "epoch": 2.19, "grad_norm": 17.32296510824594, "learning_rate": 3.549258075516544e-06, "loss": 0.4773, "step": 14049 }, { "epoch": 2.19, "grad_norm": 24.356241592345082, "learning_rate": 3.54796955616157e-06, "loss": 0.559, "step": 14050 }, { "epoch": 2.19, "grad_norm": 15.505514144658026, "learning_rate": 3.546681220298569e-06, "loss": 0.4726, "step": 14051 }, { "epoch": 2.19, "grad_norm": 27.18552797005648, "learning_rate": 3.545393067964176e-06, "loss": 0.4635, "step": 14052 }, { "epoch": 2.2, "grad_norm": 18.11100749671277, "learning_rate": 3.5441050991950264e-06, "loss": 0.4708, "step": 14053 }, { "epoch": 2.2, "grad_norm": 17.290062820689727, "learning_rate": 3.5428173140277523e-06, "loss": 0.4224, "step": 14054 }, { "epoch": 2.2, "grad_norm": 20.972685798429524, "learning_rate": 3.54152971249898e-06, "loss": 0.4253, "step": 14055 }, { "epoch": 2.2, "grad_norm": 19.22771292387033, "learning_rate": 3.5402422946453242e-06, "loss": 0.4478, "step": 14056 }, { "epoch": 2.2, "grad_norm": 19.75963982291962, "learning_rate": 3.5389550605033963e-06, "loss": 0.4764, "step": 14057 }, { "epoch": 2.2, "grad_norm": 19.94142548189347, "learning_rate": 3.537668010109805e-06, "loss": 0.4443, "step": 14058 }, { "epoch": 2.2, "grad_norm": 21.8522734309868, "learning_rate": 3.5363811435011598e-06, "loss": 0.4693, "step": 14059 }, { "epoch": 2.2, "grad_norm": 19.312032718622437, "learning_rate": 3.53509446071405e-06, "loss": 0.4541, "step": 14060 }, { "epoch": 2.2, "grad_norm": 40.733736502176065, "learning_rate": 3.5338079617850762e-06, "loss": 0.5451, "step": 14061 }, { "epoch": 2.2, "grad_norm": 15.812510162470874, "learning_rate": 3.5325216467508172e-06, "loss": 0.4658, "step": 14062 }, { "epoch": 2.2, "grad_norm": 26.776922491698524, "learning_rate": 3.5312355156478616e-06, "loss": 0.5021, "step": 14063 }, { "epoch": 2.2, "grad_norm": 22.546906582921906, "learning_rate": 3.5299495685127816e-06, "loss": 0.5025, "step": 14064 }, { "epoch": 2.2, "grad_norm": 20.470269510054994, "learning_rate": 3.528663805382151e-06, "loss": 0.4629, "step": 14065 }, { "epoch": 2.2, "grad_norm": 17.87054624667286, "learning_rate": 3.5273782262925392e-06, "loss": 0.4486, "step": 14066 }, { "epoch": 2.2, "grad_norm": 18.641677316730252, "learning_rate": 3.5260928312805042e-06, "loss": 0.4712, "step": 14067 }, { "epoch": 2.2, "grad_norm": 15.922997956361876, "learning_rate": 3.5248076203825987e-06, "loss": 0.4194, "step": 14068 }, { "epoch": 2.2, "grad_norm": 21.050922270157553, "learning_rate": 3.5235225936353767e-06, "loss": 0.4295, "step": 14069 }, { "epoch": 2.2, "grad_norm": 21.062036976125352, "learning_rate": 3.522237751075387e-06, "loss": 0.5112, "step": 14070 }, { "epoch": 2.2, "grad_norm": 18.444524533294278, "learning_rate": 3.5209530927391632e-06, "loss": 0.4451, "step": 14071 }, { "epoch": 2.2, "grad_norm": 25.213072462088874, "learning_rate": 3.5196686186632457e-06, "loss": 0.4932, "step": 14072 }, { "epoch": 2.2, "grad_norm": 30.96334830925032, "learning_rate": 3.518384328884159e-06, "loss": 0.4499, "step": 14073 }, { "epoch": 2.2, "grad_norm": 18.86835308699928, "learning_rate": 3.517100223438431e-06, "loss": 0.5184, "step": 14074 }, { "epoch": 2.2, "grad_norm": 19.84810799443279, "learning_rate": 3.5158163023625825e-06, "loss": 0.5209, "step": 14075 }, { "epoch": 2.2, "grad_norm": 11.432368447763544, "learning_rate": 3.514532565693126e-06, "loss": 0.4604, "step": 14076 }, { "epoch": 2.2, "grad_norm": 17.796697835065537, "learning_rate": 3.513249013466565e-06, "loss": 0.5287, "step": 14077 }, { "epoch": 2.2, "grad_norm": 23.644830117321195, "learning_rate": 3.5119656457194086e-06, "loss": 0.5124, "step": 14078 }, { "epoch": 2.2, "grad_norm": 19.887887654731266, "learning_rate": 3.510682462488153e-06, "loss": 0.4844, "step": 14079 }, { "epoch": 2.2, "grad_norm": 27.712158869350212, "learning_rate": 3.5093994638093e-06, "loss": 0.4934, "step": 14080 }, { "epoch": 2.2, "grad_norm": 18.37520570542549, "learning_rate": 3.5081166497193208e-06, "loss": 0.4673, "step": 14081 }, { "epoch": 2.2, "grad_norm": 20.733917116285223, "learning_rate": 3.506834020254708e-06, "loss": 0.4224, "step": 14082 }, { "epoch": 2.2, "grad_norm": 28.46560013468761, "learning_rate": 3.5055515754519363e-06, "loss": 0.5899, "step": 14083 }, { "epoch": 2.2, "grad_norm": 16.0879471205812, "learning_rate": 3.504269315347483e-06, "loss": 0.4107, "step": 14084 }, { "epoch": 2.2, "grad_norm": 23.056191021463018, "learning_rate": 3.50298723997781e-06, "loss": 0.5308, "step": 14085 }, { "epoch": 2.2, "grad_norm": 22.03328130569702, "learning_rate": 3.501705349379375e-06, "loss": 0.5034, "step": 14086 }, { "epoch": 2.2, "grad_norm": 14.221858021841893, "learning_rate": 3.5004236435886395e-06, "loss": 0.4313, "step": 14087 }, { "epoch": 2.2, "grad_norm": 28.614419932964108, "learning_rate": 3.4991421226420563e-06, "loss": 0.5076, "step": 14088 }, { "epoch": 2.2, "grad_norm": 18.563292301814137, "learning_rate": 3.497860786576065e-06, "loss": 0.4621, "step": 14089 }, { "epoch": 2.2, "grad_norm": 18.712796202145565, "learning_rate": 3.496579635427113e-06, "loss": 0.5354, "step": 14090 }, { "epoch": 2.2, "grad_norm": 19.276930815999908, "learning_rate": 3.4952986692316283e-06, "loss": 0.4711, "step": 14091 }, { "epoch": 2.2, "grad_norm": 22.864254241709762, "learning_rate": 3.4940178880260446e-06, "loss": 0.4399, "step": 14092 }, { "epoch": 2.2, "grad_norm": 17.459176633574064, "learning_rate": 3.4927372918467893e-06, "loss": 0.4139, "step": 14093 }, { "epoch": 2.2, "grad_norm": 17.626736464080295, "learning_rate": 3.491456880730275e-06, "loss": 0.4625, "step": 14094 }, { "epoch": 2.2, "grad_norm": 15.652212032909171, "learning_rate": 3.4901766547129234e-06, "loss": 0.3932, "step": 14095 }, { "epoch": 2.2, "grad_norm": 21.399589499208492, "learning_rate": 3.4888966138311363e-06, "loss": 0.4526, "step": 14096 }, { "epoch": 2.2, "grad_norm": 23.10224476859147, "learning_rate": 3.487616758121324e-06, "loss": 0.501, "step": 14097 }, { "epoch": 2.2, "grad_norm": 26.427406652385116, "learning_rate": 3.4863370876198777e-06, "loss": 0.4978, "step": 14098 }, { "epoch": 2.2, "grad_norm": 15.963691570911946, "learning_rate": 3.4850576023631942e-06, "loss": 0.4076, "step": 14099 }, { "epoch": 2.2, "grad_norm": 14.642865323550817, "learning_rate": 3.483778302387665e-06, "loss": 0.4195, "step": 14100 }, { "epoch": 2.2, "grad_norm": 22.715794370707034, "learning_rate": 3.4824991877296687e-06, "loss": 0.5094, "step": 14101 }, { "epoch": 2.2, "grad_norm": 19.806229377481852, "learning_rate": 3.4812202584255805e-06, "loss": 0.5296, "step": 14102 }, { "epoch": 2.2, "grad_norm": 15.787694634893695, "learning_rate": 3.4799415145117743e-06, "loss": 0.5387, "step": 14103 }, { "epoch": 2.2, "grad_norm": 16.17787377227078, "learning_rate": 3.4786629560246166e-06, "loss": 0.4788, "step": 14104 }, { "epoch": 2.2, "grad_norm": 19.324468011522743, "learning_rate": 3.477384583000477e-06, "loss": 0.4665, "step": 14105 }, { "epoch": 2.2, "grad_norm": 17.87162415806508, "learning_rate": 3.4761063954756967e-06, "loss": 0.4935, "step": 14106 }, { "epoch": 2.2, "grad_norm": 16.881506885263878, "learning_rate": 3.4748283934866354e-06, "loss": 0.4713, "step": 14107 }, { "epoch": 2.2, "grad_norm": 18.943740553383154, "learning_rate": 3.4735505770696365e-06, "loss": 0.5316, "step": 14108 }, { "epoch": 2.2, "grad_norm": 20.812310162099454, "learning_rate": 3.472272946261045e-06, "loss": 0.4961, "step": 14109 }, { "epoch": 2.2, "grad_norm": 18.55296125191944, "learning_rate": 3.4709955010971928e-06, "loss": 0.4191, "step": 14110 }, { "epoch": 2.2, "grad_norm": 19.224162283638634, "learning_rate": 3.4697182416144047e-06, "loss": 0.4456, "step": 14111 }, { "epoch": 2.2, "grad_norm": 21.882020101803512, "learning_rate": 3.4684411678490104e-06, "loss": 0.4243, "step": 14112 }, { "epoch": 2.2, "grad_norm": 24.784082809121216, "learning_rate": 3.4671642798373316e-06, "loss": 0.5193, "step": 14113 }, { "epoch": 2.2, "grad_norm": 19.32385857737832, "learning_rate": 3.4658875776156763e-06, "loss": 0.4954, "step": 14114 }, { "epoch": 2.2, "grad_norm": 19.0715255234308, "learning_rate": 3.4646110612203586e-06, "loss": 0.3921, "step": 14115 }, { "epoch": 2.2, "grad_norm": 18.061098432941833, "learning_rate": 3.4633347306876763e-06, "loss": 0.4915, "step": 14116 }, { "epoch": 2.21, "grad_norm": 15.773273701924971, "learning_rate": 3.4620585860539303e-06, "loss": 0.4743, "step": 14117 }, { "epoch": 2.21, "grad_norm": 34.74314170763986, "learning_rate": 3.460782627355417e-06, "loss": 0.4791, "step": 14118 }, { "epoch": 2.21, "grad_norm": 14.900922653014524, "learning_rate": 3.4595068546284207e-06, "loss": 0.4289, "step": 14119 }, { "epoch": 2.21, "grad_norm": 16.947930371650763, "learning_rate": 3.458231267909219e-06, "loss": 0.4502, "step": 14120 }, { "epoch": 2.21, "grad_norm": 18.370486386702545, "learning_rate": 3.4569558672340943e-06, "loss": 0.4923, "step": 14121 }, { "epoch": 2.21, "grad_norm": 22.91746958684292, "learning_rate": 3.455680652639319e-06, "loss": 0.4596, "step": 14122 }, { "epoch": 2.21, "grad_norm": 25.278477306884255, "learning_rate": 3.4544056241611556e-06, "loss": 0.5325, "step": 14123 }, { "epoch": 2.21, "grad_norm": 24.186439899382407, "learning_rate": 3.4531307818358705e-06, "loss": 0.468, "step": 14124 }, { "epoch": 2.21, "grad_norm": 18.243319270050254, "learning_rate": 3.451856125699713e-06, "loss": 0.4905, "step": 14125 }, { "epoch": 2.21, "grad_norm": 16.746459396353284, "learning_rate": 3.4505816557889393e-06, "loss": 0.4814, "step": 14126 }, { "epoch": 2.21, "grad_norm": 22.467693026117686, "learning_rate": 3.4493073721397895e-06, "loss": 0.5108, "step": 14127 }, { "epoch": 2.21, "grad_norm": 14.660902901563624, "learning_rate": 3.4480332747885047e-06, "loss": 0.4562, "step": 14128 }, { "epoch": 2.21, "grad_norm": 24.14380614360332, "learning_rate": 3.446759363771326e-06, "loss": 0.4643, "step": 14129 }, { "epoch": 2.21, "grad_norm": 14.309130709073447, "learning_rate": 3.445485639124476e-06, "loss": 0.4471, "step": 14130 }, { "epoch": 2.21, "grad_norm": 23.896155444501538, "learning_rate": 3.444212100884177e-06, "loss": 0.4594, "step": 14131 }, { "epoch": 2.21, "grad_norm": 17.981652467205958, "learning_rate": 3.4429387490866506e-06, "loss": 0.5047, "step": 14132 }, { "epoch": 2.21, "grad_norm": 19.197004302292797, "learning_rate": 3.4416655837681113e-06, "loss": 0.4436, "step": 14133 }, { "epoch": 2.21, "grad_norm": 21.28478034953453, "learning_rate": 3.4403926049647684e-06, "loss": 0.4566, "step": 14134 }, { "epoch": 2.21, "grad_norm": 32.408551509154584, "learning_rate": 3.4391198127128232e-06, "loss": 0.5468, "step": 14135 }, { "epoch": 2.21, "grad_norm": 18.636295686396245, "learning_rate": 3.4378472070484704e-06, "loss": 0.462, "step": 14136 }, { "epoch": 2.21, "grad_norm": 20.610298432492023, "learning_rate": 3.4365747880079027e-06, "loss": 0.4794, "step": 14137 }, { "epoch": 2.21, "grad_norm": 32.36538118814052, "learning_rate": 3.435302555627312e-06, "loss": 0.511, "step": 14138 }, { "epoch": 2.21, "grad_norm": 28.38483892126981, "learning_rate": 3.4340305099428774e-06, "loss": 0.4952, "step": 14139 }, { "epoch": 2.21, "grad_norm": 22.034427193231025, "learning_rate": 3.4327586509907695e-06, "loss": 0.4728, "step": 14140 }, { "epoch": 2.21, "grad_norm": 20.309065176248062, "learning_rate": 3.431486978807165e-06, "loss": 0.4917, "step": 14141 }, { "epoch": 2.21, "grad_norm": 22.13643800481047, "learning_rate": 3.4302154934282272e-06, "loss": 0.4239, "step": 14142 }, { "epoch": 2.21, "grad_norm": 21.289588030006954, "learning_rate": 3.4289441948901214e-06, "loss": 0.5347, "step": 14143 }, { "epoch": 2.21, "grad_norm": 27.630478786016152, "learning_rate": 3.4276730832289997e-06, "loss": 0.541, "step": 14144 }, { "epoch": 2.21, "grad_norm": 21.937499302404888, "learning_rate": 3.4264021584810058e-06, "loss": 0.4982, "step": 14145 }, { "epoch": 2.21, "grad_norm": 16.395984152637848, "learning_rate": 3.4251314206822894e-06, "loss": 0.4164, "step": 14146 }, { "epoch": 2.21, "grad_norm": 26.127649796864702, "learning_rate": 3.4238608698689933e-06, "loss": 0.4489, "step": 14147 }, { "epoch": 2.21, "grad_norm": 22.792726029359738, "learning_rate": 3.4225905060772436e-06, "loss": 0.5104, "step": 14148 }, { "epoch": 2.21, "grad_norm": 17.616790053119878, "learning_rate": 3.421320329343175e-06, "loss": 0.4441, "step": 14149 }, { "epoch": 2.21, "grad_norm": 32.529334857498206, "learning_rate": 3.420050339702905e-06, "loss": 0.4474, "step": 14150 }, { "epoch": 2.21, "grad_norm": 31.246698348868602, "learning_rate": 3.4187805371925575e-06, "loss": 0.5011, "step": 14151 }, { "epoch": 2.21, "grad_norm": 15.64517496239613, "learning_rate": 3.4175109218482383e-06, "loss": 0.4612, "step": 14152 }, { "epoch": 2.21, "grad_norm": 18.80270826449453, "learning_rate": 3.4162414937060584e-06, "loss": 0.4647, "step": 14153 }, { "epoch": 2.21, "grad_norm": 23.96895271150894, "learning_rate": 3.414972252802121e-06, "loss": 0.4539, "step": 14154 }, { "epoch": 2.21, "grad_norm": 23.328766394329747, "learning_rate": 3.413703199172519e-06, "loss": 0.4923, "step": 14155 }, { "epoch": 2.21, "grad_norm": 20.68504558291149, "learning_rate": 3.4124343328533473e-06, "loss": 0.5028, "step": 14156 }, { "epoch": 2.21, "grad_norm": 25.70867271556117, "learning_rate": 3.411165653880686e-06, "loss": 0.4463, "step": 14157 }, { "epoch": 2.21, "grad_norm": 23.55297796394052, "learning_rate": 3.40989716229062e-06, "loss": 0.5025, "step": 14158 }, { "epoch": 2.21, "grad_norm": 17.30916769874984, "learning_rate": 3.408628858119226e-06, "loss": 0.4585, "step": 14159 }, { "epoch": 2.21, "grad_norm": 17.766876085226116, "learning_rate": 3.4073607414025723e-06, "loss": 0.4396, "step": 14160 }, { "epoch": 2.21, "grad_norm": 16.99950692394386, "learning_rate": 3.406092812176719e-06, "loss": 0.4851, "step": 14161 }, { "epoch": 2.21, "grad_norm": 17.813836869884632, "learning_rate": 3.404825070477729e-06, "loss": 0.4791, "step": 14162 }, { "epoch": 2.21, "grad_norm": 20.32462966381847, "learning_rate": 3.403557516341658e-06, "loss": 0.504, "step": 14163 }, { "epoch": 2.21, "grad_norm": 20.55743554542306, "learning_rate": 3.4022901498045536e-06, "loss": 0.4485, "step": 14164 }, { "epoch": 2.21, "grad_norm": 27.66898178751445, "learning_rate": 3.4010229709024545e-06, "loss": 0.462, "step": 14165 }, { "epoch": 2.21, "grad_norm": 27.434941073638207, "learning_rate": 3.399755979671401e-06, "loss": 0.4548, "step": 14166 }, { "epoch": 2.21, "grad_norm": 19.783287359241562, "learning_rate": 3.398489176147427e-06, "loss": 0.4205, "step": 14167 }, { "epoch": 2.21, "grad_norm": 23.71350405721002, "learning_rate": 3.3972225603665653e-06, "loss": 0.4487, "step": 14168 }, { "epoch": 2.21, "grad_norm": 25.63923409250133, "learning_rate": 3.395956132364825e-06, "loss": 0.5299, "step": 14169 }, { "epoch": 2.21, "grad_norm": 23.060362616664143, "learning_rate": 3.394689892178229e-06, "loss": 0.4602, "step": 14170 }, { "epoch": 2.21, "grad_norm": 14.928723529214714, "learning_rate": 3.3934238398427886e-06, "loss": 0.4494, "step": 14171 }, { "epoch": 2.21, "grad_norm": 19.48356566778734, "learning_rate": 3.3921579753945132e-06, "loss": 0.4911, "step": 14172 }, { "epoch": 2.21, "grad_norm": 23.46514293542787, "learning_rate": 3.3908922988693995e-06, "loss": 0.4977, "step": 14173 }, { "epoch": 2.21, "grad_norm": 20.995072645330605, "learning_rate": 3.389626810303439e-06, "loss": 0.4918, "step": 14174 }, { "epoch": 2.21, "grad_norm": 12.043434922224902, "learning_rate": 3.388361509732625e-06, "loss": 0.3946, "step": 14175 }, { "epoch": 2.21, "grad_norm": 25.518325539529023, "learning_rate": 3.387096397192945e-06, "loss": 0.5225, "step": 14176 }, { "epoch": 2.21, "grad_norm": 20.24785047840458, "learning_rate": 3.3858314727203724e-06, "loss": 0.4514, "step": 14177 }, { "epoch": 2.21, "grad_norm": 19.376635347076668, "learning_rate": 3.384566736350886e-06, "loss": 0.4733, "step": 14178 }, { "epoch": 2.21, "grad_norm": 18.212229146527353, "learning_rate": 3.3833021881204487e-06, "loss": 0.4301, "step": 14179 }, { "epoch": 2.21, "grad_norm": 23.13177084131551, "learning_rate": 3.3820378280650267e-06, "loss": 0.5764, "step": 14180 }, { "epoch": 2.22, "grad_norm": 14.98748482006631, "learning_rate": 3.3807736562205805e-06, "loss": 0.4465, "step": 14181 }, { "epoch": 2.22, "grad_norm": 18.408600330064246, "learning_rate": 3.3795096726230558e-06, "loss": 0.4226, "step": 14182 }, { "epoch": 2.22, "grad_norm": 19.44167334212344, "learning_rate": 3.3782458773084072e-06, "loss": 0.4467, "step": 14183 }, { "epoch": 2.22, "grad_norm": 21.3822110439587, "learning_rate": 3.3769822703125686e-06, "loss": 0.4714, "step": 14184 }, { "epoch": 2.22, "grad_norm": 21.633649028751396, "learning_rate": 3.3757188516714833e-06, "loss": 0.5298, "step": 14185 }, { "epoch": 2.22, "grad_norm": 17.040769237496903, "learning_rate": 3.3744556214210756e-06, "loss": 0.5229, "step": 14186 }, { "epoch": 2.22, "grad_norm": 16.53573065318118, "learning_rate": 3.3731925795972743e-06, "loss": 0.4111, "step": 14187 }, { "epoch": 2.22, "grad_norm": 17.835639341237453, "learning_rate": 3.371929726236002e-06, "loss": 0.4892, "step": 14188 }, { "epoch": 2.22, "grad_norm": 18.292185956232, "learning_rate": 3.370667061373172e-06, "loss": 0.4838, "step": 14189 }, { "epoch": 2.22, "grad_norm": 20.642229667564486, "learning_rate": 3.369404585044689e-06, "loss": 0.4707, "step": 14190 }, { "epoch": 2.22, "grad_norm": 16.12206559686894, "learning_rate": 3.3681422972864598e-06, "loss": 0.4677, "step": 14191 }, { "epoch": 2.22, "grad_norm": 25.649980219197595, "learning_rate": 3.366880198134386e-06, "loss": 0.5158, "step": 14192 }, { "epoch": 2.22, "grad_norm": 25.220100786178264, "learning_rate": 3.3656182876243636e-06, "loss": 0.4718, "step": 14193 }, { "epoch": 2.22, "grad_norm": 29.121310873264775, "learning_rate": 3.3643565657922704e-06, "loss": 0.5004, "step": 14194 }, { "epoch": 2.22, "grad_norm": 22.303880807214757, "learning_rate": 3.3630950326739954e-06, "loss": 0.4644, "step": 14195 }, { "epoch": 2.22, "grad_norm": 23.889365711292687, "learning_rate": 3.3618336883054146e-06, "loss": 0.4265, "step": 14196 }, { "epoch": 2.22, "grad_norm": 18.29784727567313, "learning_rate": 3.360572532722404e-06, "loss": 0.4522, "step": 14197 }, { "epoch": 2.22, "grad_norm": 15.26974162583579, "learning_rate": 3.359311565960827e-06, "loss": 0.3747, "step": 14198 }, { "epoch": 2.22, "grad_norm": 28.503402181682258, "learning_rate": 3.3580507880565406e-06, "loss": 0.5239, "step": 14199 }, { "epoch": 2.22, "grad_norm": 31.36989860932342, "learning_rate": 3.3567901990454043e-06, "loss": 0.5603, "step": 14200 }, { "epoch": 2.22, "grad_norm": 21.294901828499274, "learning_rate": 3.355529798963272e-06, "loss": 0.498, "step": 14201 }, { "epoch": 2.22, "grad_norm": 21.89706129689245, "learning_rate": 3.354269587845982e-06, "loss": 0.4806, "step": 14202 }, { "epoch": 2.22, "grad_norm": 16.72062211116757, "learning_rate": 3.3530095657293803e-06, "loss": 0.4153, "step": 14203 }, { "epoch": 2.22, "grad_norm": 25.057245536089876, "learning_rate": 3.3517497326492952e-06, "loss": 0.5043, "step": 14204 }, { "epoch": 2.22, "grad_norm": 18.124092487109326, "learning_rate": 3.350490088641558e-06, "loss": 0.4259, "step": 14205 }, { "epoch": 2.22, "grad_norm": 17.50410646427527, "learning_rate": 3.3492306337419967e-06, "loss": 0.4986, "step": 14206 }, { "epoch": 2.22, "grad_norm": 19.564168096832738, "learning_rate": 3.3479713679864223e-06, "loss": 0.4927, "step": 14207 }, { "epoch": 2.22, "grad_norm": 24.4743156375904, "learning_rate": 3.346712291410654e-06, "loss": 0.5083, "step": 14208 }, { "epoch": 2.22, "grad_norm": 20.03059779145361, "learning_rate": 3.345453404050493e-06, "loss": 0.4428, "step": 14209 }, { "epoch": 2.22, "grad_norm": 18.726774609971446, "learning_rate": 3.344194705941748e-06, "loss": 0.4248, "step": 14210 }, { "epoch": 2.22, "grad_norm": 19.253499794708, "learning_rate": 3.342936197120208e-06, "loss": 0.4438, "step": 14211 }, { "epoch": 2.22, "grad_norm": 18.849995816354262, "learning_rate": 3.341677877621672e-06, "loss": 0.4895, "step": 14212 }, { "epoch": 2.22, "grad_norm": 14.839533799346864, "learning_rate": 3.340419747481919e-06, "loss": 0.4433, "step": 14213 }, { "epoch": 2.22, "grad_norm": 24.960924142884693, "learning_rate": 3.3391618067367347e-06, "loss": 0.4278, "step": 14214 }, { "epoch": 2.22, "grad_norm": 25.56766996589953, "learning_rate": 3.33790405542189e-06, "loss": 0.5227, "step": 14215 }, { "epoch": 2.22, "grad_norm": 20.760033100704494, "learning_rate": 3.3366464935731566e-06, "loss": 0.4874, "step": 14216 }, { "epoch": 2.22, "grad_norm": 14.74126868121376, "learning_rate": 3.3353891212263023e-06, "loss": 0.4802, "step": 14217 }, { "epoch": 2.22, "grad_norm": 26.367003042635503, "learning_rate": 3.3341319384170824e-06, "loss": 0.5098, "step": 14218 }, { "epoch": 2.22, "grad_norm": 19.144786309187655, "learning_rate": 3.332874945181248e-06, "loss": 0.4565, "step": 14219 }, { "epoch": 2.22, "grad_norm": 16.01811562315315, "learning_rate": 3.33161814155455e-06, "loss": 0.452, "step": 14220 }, { "epoch": 2.22, "grad_norm": 24.598941370793774, "learning_rate": 3.330361527572731e-06, "loss": 0.4083, "step": 14221 }, { "epoch": 2.22, "grad_norm": 17.59739929756781, "learning_rate": 3.3291051032715317e-06, "loss": 0.4356, "step": 14222 }, { "epoch": 2.22, "grad_norm": 21.428833061356688, "learning_rate": 3.3278488686866807e-06, "loss": 0.4809, "step": 14223 }, { "epoch": 2.22, "grad_norm": 19.427374598967635, "learning_rate": 3.3265928238539015e-06, "loss": 0.4905, "step": 14224 }, { "epoch": 2.22, "grad_norm": 16.67122250775418, "learning_rate": 3.3253369688089186e-06, "loss": 0.4164, "step": 14225 }, { "epoch": 2.22, "grad_norm": 23.149478178291545, "learning_rate": 3.324081303587451e-06, "loss": 0.4805, "step": 14226 }, { "epoch": 2.22, "grad_norm": 14.9630818272006, "learning_rate": 3.3228258282252056e-06, "loss": 0.4986, "step": 14227 }, { "epoch": 2.22, "grad_norm": 20.311415303319023, "learning_rate": 3.321570542757885e-06, "loss": 0.4531, "step": 14228 }, { "epoch": 2.22, "grad_norm": 20.470234888394227, "learning_rate": 3.320315447221191e-06, "loss": 0.479, "step": 14229 }, { "epoch": 2.22, "grad_norm": 26.985279540786987, "learning_rate": 3.3190605416508182e-06, "loss": 0.4606, "step": 14230 }, { "epoch": 2.22, "grad_norm": 23.78908937762767, "learning_rate": 3.3178058260824586e-06, "loss": 0.4094, "step": 14231 }, { "epoch": 2.22, "grad_norm": 21.78107048204239, "learning_rate": 3.3165513005517925e-06, "loss": 0.5466, "step": 14232 }, { "epoch": 2.22, "grad_norm": 16.38778097645092, "learning_rate": 3.3152969650944943e-06, "loss": 0.4879, "step": 14233 }, { "epoch": 2.22, "grad_norm": 21.175010316681853, "learning_rate": 3.3140428197462406e-06, "loss": 0.4802, "step": 14234 }, { "epoch": 2.22, "grad_norm": 17.772871014297134, "learning_rate": 3.312788864542701e-06, "loss": 0.4472, "step": 14235 }, { "epoch": 2.22, "grad_norm": 21.060031068760054, "learning_rate": 3.3115350995195293e-06, "loss": 0.4908, "step": 14236 }, { "epoch": 2.22, "grad_norm": 29.84109969878593, "learning_rate": 3.310281524712392e-06, "loss": 0.4939, "step": 14237 }, { "epoch": 2.22, "grad_norm": 19.69208608810965, "learning_rate": 3.3090281401569302e-06, "loss": 0.4536, "step": 14238 }, { "epoch": 2.22, "grad_norm": 27.794362378113373, "learning_rate": 3.307774945888799e-06, "loss": 0.507, "step": 14239 }, { "epoch": 2.22, "grad_norm": 17.428111228643306, "learning_rate": 3.306521941943628e-06, "loss": 0.4862, "step": 14240 }, { "epoch": 2.22, "grad_norm": 14.649227753240204, "learning_rate": 3.3052691283570593e-06, "loss": 0.5169, "step": 14241 }, { "epoch": 2.22, "grad_norm": 25.769626700522476, "learning_rate": 3.304016505164722e-06, "loss": 0.4817, "step": 14242 }, { "epoch": 2.22, "grad_norm": 20.108087637774734, "learning_rate": 3.3027640724022357e-06, "loss": 0.4683, "step": 14243 }, { "epoch": 2.22, "grad_norm": 17.233353210137388, "learning_rate": 3.301511830105225e-06, "loss": 0.4916, "step": 14244 }, { "epoch": 2.23, "grad_norm": 25.25821851126149, "learning_rate": 3.300259778309296e-06, "loss": 0.5044, "step": 14245 }, { "epoch": 2.23, "grad_norm": 16.10562831236159, "learning_rate": 3.2990079170500587e-06, "loss": 0.4736, "step": 14246 }, { "epoch": 2.23, "grad_norm": 27.241259908897085, "learning_rate": 3.297756246363121e-06, "loss": 0.4797, "step": 14247 }, { "epoch": 2.23, "grad_norm": 13.764458495389457, "learning_rate": 3.2965047662840744e-06, "loss": 0.4578, "step": 14248 }, { "epoch": 2.23, "grad_norm": 21.533729231931737, "learning_rate": 3.2952534768485066e-06, "loss": 0.4996, "step": 14249 }, { "epoch": 2.23, "grad_norm": 33.62612132149112, "learning_rate": 3.294002378092008e-06, "loss": 0.6179, "step": 14250 }, { "epoch": 2.23, "grad_norm": 24.766559414729766, "learning_rate": 3.292751470050164e-06, "loss": 0.3969, "step": 14251 }, { "epoch": 2.23, "grad_norm": 12.868531518816555, "learning_rate": 3.291500752758543e-06, "loss": 0.5056, "step": 14252 }, { "epoch": 2.23, "grad_norm": 19.674855912340163, "learning_rate": 3.2902502262527124e-06, "loss": 0.5254, "step": 14253 }, { "epoch": 2.23, "grad_norm": 19.660867243506882, "learning_rate": 3.288999890568242e-06, "loss": 0.5089, "step": 14254 }, { "epoch": 2.23, "grad_norm": 16.399274103197467, "learning_rate": 3.287749745740687e-06, "loss": 0.5188, "step": 14255 }, { "epoch": 2.23, "grad_norm": 30.479011577077774, "learning_rate": 3.2864997918056075e-06, "loss": 0.5113, "step": 14256 }, { "epoch": 2.23, "grad_norm": 25.150002384944067, "learning_rate": 3.2852500287985455e-06, "loss": 0.5246, "step": 14257 }, { "epoch": 2.23, "grad_norm": 17.388451270724662, "learning_rate": 3.284000456755042e-06, "loss": 0.4819, "step": 14258 }, { "epoch": 2.23, "grad_norm": 18.012601831596356, "learning_rate": 3.282751075710637e-06, "loss": 0.4771, "step": 14259 }, { "epoch": 2.23, "grad_norm": 18.74239715452549, "learning_rate": 3.2815018857008653e-06, "loss": 0.41, "step": 14260 }, { "epoch": 2.23, "grad_norm": 23.97333561576591, "learning_rate": 3.2802528867612494e-06, "loss": 0.5099, "step": 14261 }, { "epoch": 2.23, "grad_norm": 24.139812240992367, "learning_rate": 3.279004078927307e-06, "loss": 0.4836, "step": 14262 }, { "epoch": 2.23, "grad_norm": 19.145610888241855, "learning_rate": 3.2777554622345563e-06, "loss": 0.4683, "step": 14263 }, { "epoch": 2.23, "grad_norm": 25.393912403621083, "learning_rate": 3.2765070367185126e-06, "loss": 0.5101, "step": 14264 }, { "epoch": 2.23, "grad_norm": 16.908920665919428, "learning_rate": 3.2752588024146714e-06, "loss": 0.4527, "step": 14265 }, { "epoch": 2.23, "grad_norm": 25.12345903330322, "learning_rate": 3.274010759358539e-06, "loss": 0.478, "step": 14266 }, { "epoch": 2.23, "grad_norm": 15.874270747172954, "learning_rate": 3.272762907585604e-06, "loss": 0.456, "step": 14267 }, { "epoch": 2.23, "grad_norm": 41.42134939908545, "learning_rate": 3.271515247131356e-06, "loss": 0.4637, "step": 14268 }, { "epoch": 2.23, "grad_norm": 19.79649799246871, "learning_rate": 3.270267778031281e-06, "loss": 0.4652, "step": 14269 }, { "epoch": 2.23, "grad_norm": 21.23659705366554, "learning_rate": 3.2690205003208508e-06, "loss": 0.4655, "step": 14270 }, { "epoch": 2.23, "grad_norm": 21.757774548523162, "learning_rate": 3.2677734140355434e-06, "loss": 0.4932, "step": 14271 }, { "epoch": 2.23, "grad_norm": 24.056692628958345, "learning_rate": 3.2665265192108186e-06, "loss": 0.4383, "step": 14272 }, { "epoch": 2.23, "grad_norm": 17.407862603978565, "learning_rate": 3.265279815882144e-06, "loss": 0.5083, "step": 14273 }, { "epoch": 2.23, "grad_norm": 24.034531624729695, "learning_rate": 3.264033304084968e-06, "loss": 0.5098, "step": 14274 }, { "epoch": 2.23, "grad_norm": 23.019349531035484, "learning_rate": 3.2627869838547454e-06, "loss": 0.4443, "step": 14275 }, { "epoch": 2.23, "grad_norm": 18.186384294640565, "learning_rate": 3.2615408552269223e-06, "loss": 0.4873, "step": 14276 }, { "epoch": 2.23, "grad_norm": 16.194729800490645, "learning_rate": 3.260294918236935e-06, "loss": 0.4052, "step": 14277 }, { "epoch": 2.23, "grad_norm": 20.94586027824207, "learning_rate": 3.2590491729202146e-06, "loss": 0.427, "step": 14278 }, { "epoch": 2.23, "grad_norm": 18.449611347066973, "learning_rate": 3.257803619312193e-06, "loss": 0.4804, "step": 14279 }, { "epoch": 2.23, "grad_norm": 22.939002629026504, "learning_rate": 3.256558257448292e-06, "loss": 0.4719, "step": 14280 }, { "epoch": 2.23, "grad_norm": 19.642706943152906, "learning_rate": 3.255313087363936e-06, "loss": 0.4459, "step": 14281 }, { "epoch": 2.23, "grad_norm": 18.899193243596365, "learning_rate": 3.2540681090945235e-06, "loss": 0.5118, "step": 14282 }, { "epoch": 2.23, "grad_norm": 19.39349957168287, "learning_rate": 3.252823322675468e-06, "loss": 0.4634, "step": 14283 }, { "epoch": 2.23, "grad_norm": 21.177561836556563, "learning_rate": 3.2515787281421697e-06, "loss": 0.587, "step": 14284 }, { "epoch": 2.23, "grad_norm": 15.530327032668758, "learning_rate": 3.250334325530029e-06, "loss": 0.4582, "step": 14285 }, { "epoch": 2.23, "grad_norm": 21.160174986162176, "learning_rate": 3.24909011487443e-06, "loss": 0.4731, "step": 14286 }, { "epoch": 2.23, "grad_norm": 23.226236104355927, "learning_rate": 3.247846096210757e-06, "loss": 0.5326, "step": 14287 }, { "epoch": 2.23, "grad_norm": 20.96990783169642, "learning_rate": 3.2466022695743905e-06, "loss": 0.4769, "step": 14288 }, { "epoch": 2.23, "grad_norm": 14.472697040948256, "learning_rate": 3.2453586350007084e-06, "loss": 0.4267, "step": 14289 }, { "epoch": 2.23, "grad_norm": 18.966648976634477, "learning_rate": 3.2441151925250724e-06, "loss": 0.4255, "step": 14290 }, { "epoch": 2.23, "grad_norm": 13.185181563092526, "learning_rate": 3.2428719421828515e-06, "loss": 0.4687, "step": 14291 }, { "epoch": 2.23, "grad_norm": 24.446849352864483, "learning_rate": 3.2416288840093955e-06, "loss": 0.5577, "step": 14292 }, { "epoch": 2.23, "grad_norm": 22.577778325834785, "learning_rate": 3.2403860180400615e-06, "loss": 0.4748, "step": 14293 }, { "epoch": 2.23, "grad_norm": 27.08563303329335, "learning_rate": 3.2391433443101984e-06, "loss": 0.4926, "step": 14294 }, { "epoch": 2.23, "grad_norm": 16.28430442522078, "learning_rate": 3.237900862855139e-06, "loss": 0.3785, "step": 14295 }, { "epoch": 2.23, "grad_norm": 13.686519460034482, "learning_rate": 3.236658573710227e-06, "loss": 0.4874, "step": 14296 }, { "epoch": 2.23, "grad_norm": 24.982863507525025, "learning_rate": 3.2354164769107866e-06, "loss": 0.5085, "step": 14297 }, { "epoch": 2.23, "grad_norm": 23.35585327265751, "learning_rate": 3.234174572492147e-06, "loss": 0.4952, "step": 14298 }, { "epoch": 2.23, "grad_norm": 19.697228652851315, "learning_rate": 3.2329328604896217e-06, "loss": 0.472, "step": 14299 }, { "epoch": 2.23, "grad_norm": 18.83689424639408, "learning_rate": 3.2316913409385263e-06, "loss": 0.4262, "step": 14300 }, { "epoch": 2.23, "grad_norm": 20.1018932877213, "learning_rate": 3.2304500138741736e-06, "loss": 0.5658, "step": 14301 }, { "epoch": 2.23, "grad_norm": 24.114858241742233, "learning_rate": 3.229208879331862e-06, "loss": 0.4258, "step": 14302 }, { "epoch": 2.23, "grad_norm": 26.834629937623834, "learning_rate": 3.2279679373468866e-06, "loss": 0.4812, "step": 14303 }, { "epoch": 2.23, "grad_norm": 24.88807828150885, "learning_rate": 3.2267271879545413e-06, "loss": 0.5089, "step": 14304 }, { "epoch": 2.23, "grad_norm": 17.1155344353335, "learning_rate": 3.2254866311901125e-06, "loss": 0.4876, "step": 14305 }, { "epoch": 2.23, "grad_norm": 21.317829679443097, "learning_rate": 3.224246267088884e-06, "loss": 0.5531, "step": 14306 }, { "epoch": 2.23, "grad_norm": 33.593595767886946, "learning_rate": 3.2230060956861285e-06, "loss": 0.5258, "step": 14307 }, { "epoch": 2.23, "grad_norm": 15.675784659747722, "learning_rate": 3.221766117017111e-06, "loss": 0.4679, "step": 14308 }, { "epoch": 2.24, "grad_norm": 17.130109254529824, "learning_rate": 3.220526331117101e-06, "loss": 0.4636, "step": 14309 }, { "epoch": 2.24, "grad_norm": 19.189185045228566, "learning_rate": 3.219286738021359e-06, "loss": 0.4753, "step": 14310 }, { "epoch": 2.24, "grad_norm": 23.383957187050587, "learning_rate": 3.2180473377651355e-06, "loss": 0.5106, "step": 14311 }, { "epoch": 2.24, "grad_norm": 20.01716723519588, "learning_rate": 3.216808130383675e-06, "loss": 0.4941, "step": 14312 }, { "epoch": 2.24, "grad_norm": 26.421769824909966, "learning_rate": 3.2155691159122236e-06, "loss": 0.5627, "step": 14313 }, { "epoch": 2.24, "grad_norm": 21.824993583398598, "learning_rate": 3.214330294386021e-06, "loss": 0.5243, "step": 14314 }, { "epoch": 2.24, "grad_norm": 21.072516955797546, "learning_rate": 3.2130916658402956e-06, "loss": 0.5274, "step": 14315 }, { "epoch": 2.24, "grad_norm": 17.153106139095012, "learning_rate": 3.211853230310269e-06, "loss": 0.5096, "step": 14316 }, { "epoch": 2.24, "grad_norm": 19.136361590087546, "learning_rate": 3.2106149878311655e-06, "loss": 0.3973, "step": 14317 }, { "epoch": 2.24, "grad_norm": 24.237866439159284, "learning_rate": 3.209376938438201e-06, "loss": 0.4129, "step": 14318 }, { "epoch": 2.24, "grad_norm": 18.565251571926947, "learning_rate": 3.208139082166587e-06, "loss": 0.4767, "step": 14319 }, { "epoch": 2.24, "grad_norm": 22.440590835511284, "learning_rate": 3.2069014190515237e-06, "loss": 0.5808, "step": 14320 }, { "epoch": 2.24, "grad_norm": 16.453265147866635, "learning_rate": 3.205663949128207e-06, "loss": 0.4778, "step": 14321 }, { "epoch": 2.24, "grad_norm": 331.2600062478427, "learning_rate": 3.204426672431834e-06, "loss": 0.5467, "step": 14322 }, { "epoch": 2.24, "grad_norm": 27.51695419872413, "learning_rate": 3.2031895889975926e-06, "loss": 0.4934, "step": 14323 }, { "epoch": 2.24, "grad_norm": 18.623731021886798, "learning_rate": 3.201952698860662e-06, "loss": 0.5092, "step": 14324 }, { "epoch": 2.24, "grad_norm": 19.77207562124895, "learning_rate": 3.200716002056222e-06, "loss": 0.496, "step": 14325 }, { "epoch": 2.24, "grad_norm": 15.125458357699145, "learning_rate": 3.199479498619439e-06, "loss": 0.5078, "step": 14326 }, { "epoch": 2.24, "grad_norm": 24.328589947618013, "learning_rate": 3.1982431885854837e-06, "loss": 0.457, "step": 14327 }, { "epoch": 2.24, "grad_norm": 14.784039748250384, "learning_rate": 3.19700707198951e-06, "loss": 0.4783, "step": 14328 }, { "epoch": 2.24, "grad_norm": 20.502971612606256, "learning_rate": 3.1957711488666756e-06, "loss": 0.4858, "step": 14329 }, { "epoch": 2.24, "grad_norm": 25.170697397193248, "learning_rate": 3.1945354192521316e-06, "loss": 0.47, "step": 14330 }, { "epoch": 2.24, "grad_norm": 20.939671265055, "learning_rate": 3.193299883181017e-06, "loss": 0.4573, "step": 14331 }, { "epoch": 2.24, "grad_norm": 18.847025219725495, "learning_rate": 3.192064540688475e-06, "loss": 0.5015, "step": 14332 }, { "epoch": 2.24, "grad_norm": 14.79447557840037, "learning_rate": 3.190829391809631e-06, "loss": 0.4768, "step": 14333 }, { "epoch": 2.24, "grad_norm": 11.246256180200069, "learning_rate": 3.189594436579617e-06, "loss": 0.3996, "step": 14334 }, { "epoch": 2.24, "grad_norm": 20.952643798576595, "learning_rate": 3.1883596750335554e-06, "loss": 0.4828, "step": 14335 }, { "epoch": 2.24, "grad_norm": 23.894220802905878, "learning_rate": 3.1871251072065613e-06, "loss": 0.4751, "step": 14336 }, { "epoch": 2.24, "grad_norm": 22.197414232890242, "learning_rate": 3.1858907331337397e-06, "loss": 0.4944, "step": 14337 }, { "epoch": 2.24, "grad_norm": 23.07762755501611, "learning_rate": 3.1846565528501993e-06, "loss": 0.4641, "step": 14338 }, { "epoch": 2.24, "grad_norm": 24.98848517833969, "learning_rate": 3.1834225663910436e-06, "loss": 0.547, "step": 14339 }, { "epoch": 2.24, "grad_norm": 18.141554250910612, "learning_rate": 3.1821887737913613e-06, "loss": 0.3794, "step": 14340 }, { "epoch": 2.24, "grad_norm": 16.376573819594306, "learning_rate": 3.18095517508624e-06, "loss": 0.3943, "step": 14341 }, { "epoch": 2.24, "grad_norm": 34.161243254792005, "learning_rate": 3.1797217703107643e-06, "loss": 0.5386, "step": 14342 }, { "epoch": 2.24, "grad_norm": 31.969370179206262, "learning_rate": 3.1784885595000114e-06, "loss": 0.5642, "step": 14343 }, { "epoch": 2.24, "grad_norm": 14.332285108628783, "learning_rate": 3.177255542689056e-06, "loss": 0.5116, "step": 14344 }, { "epoch": 2.24, "grad_norm": 15.646392088672556, "learning_rate": 3.1760227199129623e-06, "loss": 0.4162, "step": 14345 }, { "epoch": 2.24, "grad_norm": 22.409295328701088, "learning_rate": 3.1747900912067877e-06, "loss": 0.451, "step": 14346 }, { "epoch": 2.24, "grad_norm": 31.4810930893793, "learning_rate": 3.17355765660559e-06, "loss": 0.5056, "step": 14347 }, { "epoch": 2.24, "grad_norm": 23.86838834835201, "learning_rate": 3.172325416144423e-06, "loss": 0.4957, "step": 14348 }, { "epoch": 2.24, "grad_norm": 20.439822092020375, "learning_rate": 3.171093369858322e-06, "loss": 0.4592, "step": 14349 }, { "epoch": 2.24, "grad_norm": 13.048567435999786, "learning_rate": 3.1698615177823357e-06, "loss": 0.4631, "step": 14350 }, { "epoch": 2.24, "grad_norm": 18.6653666146451, "learning_rate": 3.1686298599514887e-06, "loss": 0.5, "step": 14351 }, { "epoch": 2.24, "grad_norm": 19.060286325369788, "learning_rate": 3.167398396400816e-06, "loss": 0.4454, "step": 14352 }, { "epoch": 2.24, "grad_norm": 20.511651940746397, "learning_rate": 3.1661671271653317e-06, "loss": 0.4407, "step": 14353 }, { "epoch": 2.24, "grad_norm": 20.465336946040818, "learning_rate": 3.1649360522800578e-06, "loss": 0.509, "step": 14354 }, { "epoch": 2.24, "grad_norm": 22.67990111678465, "learning_rate": 3.163705171780006e-06, "loss": 0.454, "step": 14355 }, { "epoch": 2.24, "grad_norm": 14.015911591618107, "learning_rate": 3.1624744857001776e-06, "loss": 0.4932, "step": 14356 }, { "epoch": 2.24, "grad_norm": 28.546697440375567, "learning_rate": 3.161243994075578e-06, "loss": 0.4614, "step": 14357 }, { "epoch": 2.24, "grad_norm": 18.98538722177228, "learning_rate": 3.1600136969411954e-06, "loss": 0.4479, "step": 14358 }, { "epoch": 2.24, "grad_norm": 21.639863646021713, "learning_rate": 3.1587835943320256e-06, "loss": 0.4667, "step": 14359 }, { "epoch": 2.24, "grad_norm": 12.654913511470536, "learning_rate": 3.1575536862830445e-06, "loss": 0.4452, "step": 14360 }, { "epoch": 2.24, "grad_norm": 21.19234611852943, "learning_rate": 3.156323972829237e-06, "loss": 0.4799, "step": 14361 }, { "epoch": 2.24, "grad_norm": 23.202184871932534, "learning_rate": 3.1550944540055693e-06, "loss": 0.4994, "step": 14362 }, { "epoch": 2.24, "grad_norm": 11.69935143448063, "learning_rate": 3.1538651298470126e-06, "loss": 0.485, "step": 14363 }, { "epoch": 2.24, "grad_norm": 17.52985505896903, "learning_rate": 3.1526360003885283e-06, "loss": 0.4512, "step": 14364 }, { "epoch": 2.24, "grad_norm": 24.83612723456546, "learning_rate": 3.151407065665071e-06, "loss": 0.5222, "step": 14365 }, { "epoch": 2.24, "grad_norm": 18.234942443733026, "learning_rate": 3.150178325711587e-06, "loss": 0.4532, "step": 14366 }, { "epoch": 2.24, "grad_norm": 53.68903504176631, "learning_rate": 3.1489497805630243e-06, "loss": 0.4524, "step": 14367 }, { "epoch": 2.24, "grad_norm": 23.256895821842843, "learning_rate": 3.1477214302543225e-06, "loss": 0.51, "step": 14368 }, { "epoch": 2.24, "grad_norm": 17.933806884037544, "learning_rate": 3.1464932748204215e-06, "loss": 0.4733, "step": 14369 }, { "epoch": 2.24, "grad_norm": 12.700028696694204, "learning_rate": 3.1452653142962353e-06, "loss": 0.4283, "step": 14370 }, { "epoch": 2.24, "grad_norm": 34.95162712369848, "learning_rate": 3.144037548716694e-06, "loss": 0.5144, "step": 14371 }, { "epoch": 2.24, "grad_norm": 22.451144056398366, "learning_rate": 3.142809978116714e-06, "loss": 0.5001, "step": 14372 }, { "epoch": 2.25, "grad_norm": 16.573215571137066, "learning_rate": 3.1415826025312103e-06, "loss": 0.412, "step": 14373 }, { "epoch": 2.25, "grad_norm": 29.050526628485077, "learning_rate": 3.140355421995086e-06, "loss": 0.5389, "step": 14374 }, { "epoch": 2.25, "grad_norm": 14.248755929833392, "learning_rate": 3.139128436543236e-06, "loss": 0.468, "step": 14375 }, { "epoch": 2.25, "grad_norm": 20.8957388994746, "learning_rate": 3.1379016462105617e-06, "loss": 0.3903, "step": 14376 }, { "epoch": 2.25, "grad_norm": 16.93119597084899, "learning_rate": 3.136675051031952e-06, "loss": 0.3946, "step": 14377 }, { "epoch": 2.25, "grad_norm": 15.80356864297048, "learning_rate": 3.1354486510422865e-06, "loss": 0.4358, "step": 14378 }, { "epoch": 2.25, "grad_norm": 26.34947878145194, "learning_rate": 3.134222446276448e-06, "loss": 0.4947, "step": 14379 }, { "epoch": 2.25, "grad_norm": 15.728335436484356, "learning_rate": 3.1329964367693042e-06, "loss": 0.4604, "step": 14380 }, { "epoch": 2.25, "grad_norm": 16.813707834769023, "learning_rate": 3.1317706225557254e-06, "loss": 0.4187, "step": 14381 }, { "epoch": 2.25, "grad_norm": 18.45159610225471, "learning_rate": 3.130545003670574e-06, "loss": 0.4904, "step": 14382 }, { "epoch": 2.25, "grad_norm": 21.566818894589602, "learning_rate": 3.129319580148702e-06, "loss": 0.5048, "step": 14383 }, { "epoch": 2.25, "grad_norm": 23.911631635401953, "learning_rate": 3.128094352024965e-06, "loss": 0.4119, "step": 14384 }, { "epoch": 2.25, "grad_norm": 26.220533922953038, "learning_rate": 3.1268693193342014e-06, "loss": 0.4568, "step": 14385 }, { "epoch": 2.25, "grad_norm": 14.011667214802133, "learning_rate": 3.1256444821112573e-06, "loss": 0.4442, "step": 14386 }, { "epoch": 2.25, "grad_norm": 28.49043746878992, "learning_rate": 3.1244198403909587e-06, "loss": 0.4872, "step": 14387 }, { "epoch": 2.25, "grad_norm": 14.50332712568118, "learning_rate": 3.1231953942081385e-06, "loss": 0.4169, "step": 14388 }, { "epoch": 2.25, "grad_norm": 14.501733297614937, "learning_rate": 3.1219711435976207e-06, "loss": 0.4344, "step": 14389 }, { "epoch": 2.25, "grad_norm": 32.946003909564176, "learning_rate": 3.1207470885942213e-06, "loss": 0.4389, "step": 14390 }, { "epoch": 2.25, "grad_norm": 20.428630474746697, "learning_rate": 3.119523229232746e-06, "loss": 0.4551, "step": 14391 }, { "epoch": 2.25, "grad_norm": 28.246480268878578, "learning_rate": 3.1182995655480065e-06, "loss": 0.4917, "step": 14392 }, { "epoch": 2.25, "grad_norm": 20.894016106189262, "learning_rate": 3.1170760975748006e-06, "loss": 0.4509, "step": 14393 }, { "epoch": 2.25, "grad_norm": 15.825113068778577, "learning_rate": 3.115852825347927e-06, "loss": 0.4263, "step": 14394 }, { "epoch": 2.25, "grad_norm": 21.229917379061057, "learning_rate": 3.114629748902173e-06, "loss": 0.449, "step": 14395 }, { "epoch": 2.25, "grad_norm": 17.263691951888067, "learning_rate": 3.1134068682723173e-06, "loss": 0.4603, "step": 14396 }, { "epoch": 2.25, "grad_norm": 22.43525973906959, "learning_rate": 3.112184183493142e-06, "loss": 0.473, "step": 14397 }, { "epoch": 2.25, "grad_norm": 15.820768709080474, "learning_rate": 3.1109616945994225e-06, "loss": 0.4678, "step": 14398 }, { "epoch": 2.25, "grad_norm": 33.057930132058054, "learning_rate": 3.109739401625922e-06, "loss": 0.5474, "step": 14399 }, { "epoch": 2.25, "grad_norm": 23.04651367615477, "learning_rate": 3.108517304607399e-06, "loss": 0.5315, "step": 14400 }, { "epoch": 2.25, "grad_norm": 19.247604440900464, "learning_rate": 3.107295403578613e-06, "loss": 0.4507, "step": 14401 }, { "epoch": 2.25, "grad_norm": 24.72954990365314, "learning_rate": 3.1060736985743166e-06, "loss": 0.6107, "step": 14402 }, { "epoch": 2.25, "grad_norm": 24.574510548656622, "learning_rate": 3.1048521896292483e-06, "loss": 0.4751, "step": 14403 }, { "epoch": 2.25, "grad_norm": 22.150621748501255, "learning_rate": 3.103630876778153e-06, "loss": 0.4813, "step": 14404 }, { "epoch": 2.25, "grad_norm": 20.581079570714497, "learning_rate": 3.102409760055758e-06, "loss": 0.4464, "step": 14405 }, { "epoch": 2.25, "grad_norm": 26.680110779472226, "learning_rate": 3.101188839496795e-06, "loss": 0.5548, "step": 14406 }, { "epoch": 2.25, "grad_norm": 17.091697645504162, "learning_rate": 3.099968115135988e-06, "loss": 0.5001, "step": 14407 }, { "epoch": 2.25, "grad_norm": 27.30240287882424, "learning_rate": 3.098747587008053e-06, "loss": 0.5052, "step": 14408 }, { "epoch": 2.25, "grad_norm": 24.883326689529127, "learning_rate": 3.0975272551476953e-06, "loss": 0.4574, "step": 14409 }, { "epoch": 2.25, "grad_norm": 21.06161724272564, "learning_rate": 3.0963071195896243e-06, "loss": 0.5273, "step": 14410 }, { "epoch": 2.25, "grad_norm": 29.230492738792673, "learning_rate": 3.0950871803685435e-06, "loss": 0.4941, "step": 14411 }, { "epoch": 2.25, "grad_norm": 38.369405423690246, "learning_rate": 3.0938674375191415e-06, "loss": 0.5153, "step": 14412 }, { "epoch": 2.25, "grad_norm": 18.70718803467509, "learning_rate": 3.0926478910761127e-06, "loss": 0.4729, "step": 14413 }, { "epoch": 2.25, "grad_norm": 27.151596696762624, "learning_rate": 3.091428541074134e-06, "loss": 0.4964, "step": 14414 }, { "epoch": 2.25, "grad_norm": 30.89761117922585, "learning_rate": 3.090209387547889e-06, "loss": 0.5068, "step": 14415 }, { "epoch": 2.25, "grad_norm": 22.268542439245305, "learning_rate": 3.088990430532044e-06, "loss": 0.5266, "step": 14416 }, { "epoch": 2.25, "grad_norm": 24.64029710015681, "learning_rate": 3.0877716700612683e-06, "loss": 0.4689, "step": 14417 }, { "epoch": 2.25, "grad_norm": 17.933843844827553, "learning_rate": 3.086553106170226e-06, "loss": 0.4081, "step": 14418 }, { "epoch": 2.25, "grad_norm": 14.797660422602728, "learning_rate": 3.085334738893565e-06, "loss": 0.4636, "step": 14419 }, { "epoch": 2.25, "grad_norm": 19.296978701571533, "learning_rate": 3.0841165682659436e-06, "loss": 0.431, "step": 14420 }, { "epoch": 2.25, "grad_norm": 20.908002233275102, "learning_rate": 3.082898594321998e-06, "loss": 0.4549, "step": 14421 }, { "epoch": 2.25, "grad_norm": 14.254789524200397, "learning_rate": 3.0816808170963697e-06, "loss": 0.4357, "step": 14422 }, { "epoch": 2.25, "grad_norm": 15.83150593497665, "learning_rate": 3.080463236623694e-06, "loss": 0.4414, "step": 14423 }, { "epoch": 2.25, "grad_norm": 18.63845862288901, "learning_rate": 3.079245852938597e-06, "loss": 0.471, "step": 14424 }, { "epoch": 2.25, "grad_norm": 20.340703832860676, "learning_rate": 3.0780286660756964e-06, "loss": 0.5241, "step": 14425 }, { "epoch": 2.25, "grad_norm": 18.455858806933296, "learning_rate": 3.0768116760696108e-06, "loss": 0.4209, "step": 14426 }, { "epoch": 2.25, "grad_norm": 16.35202524830817, "learning_rate": 3.0755948829549552e-06, "loss": 0.505, "step": 14427 }, { "epoch": 2.25, "grad_norm": 22.689368047929896, "learning_rate": 3.0743782867663297e-06, "loss": 0.5471, "step": 14428 }, { "epoch": 2.25, "grad_norm": 15.994484678204449, "learning_rate": 3.0731618875383318e-06, "loss": 0.5364, "step": 14429 }, { "epoch": 2.25, "grad_norm": 26.42212318905037, "learning_rate": 3.071945685305557e-06, "loss": 0.5211, "step": 14430 }, { "epoch": 2.25, "grad_norm": 4.191327406256336, "learning_rate": 3.0707296801025954e-06, "loss": 0.5204, "step": 14431 }, { "epoch": 2.25, "grad_norm": 21.991302221039223, "learning_rate": 3.0695138719640307e-06, "loss": 0.4586, "step": 14432 }, { "epoch": 2.25, "grad_norm": 20.888557334610837, "learning_rate": 3.068298260924436e-06, "loss": 0.501, "step": 14433 }, { "epoch": 2.25, "grad_norm": 21.2318372367857, "learning_rate": 3.0670828470183823e-06, "loss": 0.4796, "step": 14434 }, { "epoch": 2.25, "grad_norm": 16.624803972164138, "learning_rate": 3.065867630280436e-06, "loss": 0.4517, "step": 14435 }, { "epoch": 2.25, "grad_norm": 20.187882251486435, "learning_rate": 3.0646526107451615e-06, "loss": 0.4536, "step": 14436 }, { "epoch": 2.26, "grad_norm": 18.443662702139886, "learning_rate": 3.0634377884471055e-06, "loss": 0.4794, "step": 14437 }, { "epoch": 2.26, "grad_norm": 26.660167494551413, "learning_rate": 3.0622231634208254e-06, "loss": 0.4736, "step": 14438 }, { "epoch": 2.26, "grad_norm": 24.6690069782025, "learning_rate": 3.0610087357008567e-06, "loss": 0.4688, "step": 14439 }, { "epoch": 2.26, "grad_norm": 16.528152093975113, "learning_rate": 3.0597945053217435e-06, "loss": 0.4876, "step": 14440 }, { "epoch": 2.26, "grad_norm": 21.846604173722543, "learning_rate": 3.0585804723180115e-06, "loss": 0.473, "step": 14441 }, { "epoch": 2.26, "grad_norm": 19.46423862074733, "learning_rate": 3.0573666367241893e-06, "loss": 0.47, "step": 14442 }, { "epoch": 2.26, "grad_norm": 21.817058423026626, "learning_rate": 3.056152998574803e-06, "loss": 0.4354, "step": 14443 }, { "epoch": 2.26, "grad_norm": 28.24880661384868, "learning_rate": 3.0549395579043593e-06, "loss": 0.4707, "step": 14444 }, { "epoch": 2.26, "grad_norm": 21.011777732696835, "learning_rate": 3.053726314747376e-06, "loss": 0.4621, "step": 14445 }, { "epoch": 2.26, "grad_norm": 37.511315925657414, "learning_rate": 3.0525132691383486e-06, "loss": 0.5196, "step": 14446 }, { "epoch": 2.26, "grad_norm": 19.35770411127352, "learning_rate": 3.0513004211117805e-06, "loss": 0.5356, "step": 14447 }, { "epoch": 2.26, "grad_norm": 26.269570943242257, "learning_rate": 3.050087770702167e-06, "loss": 0.5597, "step": 14448 }, { "epoch": 2.26, "grad_norm": 11.864862249370894, "learning_rate": 3.048875317943991e-06, "loss": 0.4689, "step": 14449 }, { "epoch": 2.26, "grad_norm": 15.075530169860757, "learning_rate": 3.047663062871731e-06, "loss": 0.4594, "step": 14450 }, { "epoch": 2.26, "grad_norm": 23.95961564916996, "learning_rate": 3.046451005519867e-06, "loss": 0.4948, "step": 14451 }, { "epoch": 2.26, "grad_norm": 20.648139900230777, "learning_rate": 3.0452391459228726e-06, "loss": 0.4036, "step": 14452 }, { "epoch": 2.26, "grad_norm": 13.6742962693194, "learning_rate": 3.0440274841152072e-06, "loss": 0.3943, "step": 14453 }, { "epoch": 2.26, "grad_norm": 20.8202832262932, "learning_rate": 3.0428160201313295e-06, "loss": 0.4934, "step": 14454 }, { "epoch": 2.26, "grad_norm": 25.224949656697138, "learning_rate": 3.0416047540056924e-06, "loss": 0.5326, "step": 14455 }, { "epoch": 2.26, "grad_norm": 20.351467316124854, "learning_rate": 3.040393685772747e-06, "loss": 0.451, "step": 14456 }, { "epoch": 2.26, "grad_norm": 21.01128670964009, "learning_rate": 3.0391828154669367e-06, "loss": 0.4905, "step": 14457 }, { "epoch": 2.26, "grad_norm": 18.223739676859527, "learning_rate": 3.037972143122695e-06, "loss": 0.5233, "step": 14458 }, { "epoch": 2.26, "grad_norm": 12.127069508035094, "learning_rate": 3.0367616687744504e-06, "loss": 0.5177, "step": 14459 }, { "epoch": 2.26, "grad_norm": 16.784847572725155, "learning_rate": 3.035551392456629e-06, "loss": 0.5001, "step": 14460 }, { "epoch": 2.26, "grad_norm": 17.118037957009825, "learning_rate": 3.0343413142036573e-06, "loss": 0.4609, "step": 14461 }, { "epoch": 2.26, "grad_norm": 24.86200295673555, "learning_rate": 3.0331314340499416e-06, "loss": 0.3973, "step": 14462 }, { "epoch": 2.26, "grad_norm": 27.53457395620035, "learning_rate": 3.0319217520298904e-06, "loss": 0.5044, "step": 14463 }, { "epoch": 2.26, "grad_norm": 15.766893642032095, "learning_rate": 3.0307122681779088e-06, "loss": 0.4725, "step": 14464 }, { "epoch": 2.26, "grad_norm": 18.08258026387261, "learning_rate": 3.0295029825283952e-06, "loss": 0.4086, "step": 14465 }, { "epoch": 2.26, "grad_norm": 18.277191850390157, "learning_rate": 3.028293895115737e-06, "loss": 0.5337, "step": 14466 }, { "epoch": 2.26, "grad_norm": 14.337090513963576, "learning_rate": 3.027085005974325e-06, "loss": 0.4672, "step": 14467 }, { "epoch": 2.26, "grad_norm": 18.845246605093315, "learning_rate": 3.025876315138533e-06, "loss": 0.47, "step": 14468 }, { "epoch": 2.26, "grad_norm": 30.07893215334974, "learning_rate": 3.0246678226427396e-06, "loss": 0.514, "step": 14469 }, { "epoch": 2.26, "grad_norm": 18.79710428473836, "learning_rate": 3.0234595285213154e-06, "loss": 0.4601, "step": 14470 }, { "epoch": 2.26, "grad_norm": 25.590624931709048, "learning_rate": 3.022251432808618e-06, "loss": 0.4572, "step": 14471 }, { "epoch": 2.26, "grad_norm": 16.60547368954316, "learning_rate": 3.021043535539011e-06, "loss": 0.492, "step": 14472 }, { "epoch": 2.26, "grad_norm": 23.489112097051457, "learning_rate": 3.0198358367468407e-06, "loss": 0.5081, "step": 14473 }, { "epoch": 2.26, "grad_norm": 17.924892681158354, "learning_rate": 3.018628336466458e-06, "loss": 0.4881, "step": 14474 }, { "epoch": 2.26, "grad_norm": 27.067810916223017, "learning_rate": 3.0174210347322e-06, "loss": 0.4992, "step": 14475 }, { "epoch": 2.26, "grad_norm": 21.110252014848722, "learning_rate": 3.016213931578401e-06, "loss": 0.3848, "step": 14476 }, { "epoch": 2.26, "grad_norm": 20.053493181843194, "learning_rate": 3.0150070270393973e-06, "loss": 0.4943, "step": 14477 }, { "epoch": 2.26, "grad_norm": 16.591367301746114, "learning_rate": 3.013800321149506e-06, "loss": 0.4513, "step": 14478 }, { "epoch": 2.26, "grad_norm": 27.25668355364714, "learning_rate": 3.012593813943044e-06, "loss": 0.4823, "step": 14479 }, { "epoch": 2.26, "grad_norm": 24.236099586389955, "learning_rate": 3.0113875054543273e-06, "loss": 0.4419, "step": 14480 }, { "epoch": 2.26, "grad_norm": 14.793343441803447, "learning_rate": 3.0101813957176617e-06, "loss": 0.4485, "step": 14481 }, { "epoch": 2.26, "grad_norm": 22.148656435629523, "learning_rate": 3.008975484767351e-06, "loss": 0.4498, "step": 14482 }, { "epoch": 2.26, "grad_norm": 26.316310479237725, "learning_rate": 3.0077697726376873e-06, "loss": 0.4948, "step": 14483 }, { "epoch": 2.26, "grad_norm": 23.967292667981102, "learning_rate": 3.0065642593629597e-06, "loss": 0.4571, "step": 14484 }, { "epoch": 2.26, "grad_norm": 19.033081494290624, "learning_rate": 3.005358944977452e-06, "loss": 0.422, "step": 14485 }, { "epoch": 2.26, "grad_norm": 19.42182222943798, "learning_rate": 3.004153829515447e-06, "loss": 0.3917, "step": 14486 }, { "epoch": 2.26, "grad_norm": 21.248015574051667, "learning_rate": 3.0029489130112155e-06, "loss": 0.4413, "step": 14487 }, { "epoch": 2.26, "grad_norm": 24.56216708532793, "learning_rate": 3.00174419549902e-06, "loss": 0.5005, "step": 14488 }, { "epoch": 2.26, "grad_norm": 16.607647846384427, "learning_rate": 3.0005396770131266e-06, "loss": 0.4137, "step": 14489 }, { "epoch": 2.26, "grad_norm": 19.657152513001353, "learning_rate": 2.9993353575877926e-06, "loss": 0.5015, "step": 14490 }, { "epoch": 2.26, "grad_norm": 19.530034866680325, "learning_rate": 2.9981312372572624e-06, "loss": 0.4064, "step": 14491 }, { "epoch": 2.26, "grad_norm": 27.66560088660425, "learning_rate": 2.9969273160557877e-06, "loss": 0.4708, "step": 14492 }, { "epoch": 2.26, "grad_norm": 22.908500331863422, "learning_rate": 2.9957235940176e-06, "loss": 0.4383, "step": 14493 }, { "epoch": 2.26, "grad_norm": 16.767054634609384, "learning_rate": 2.9945200711769362e-06, "loss": 0.4238, "step": 14494 }, { "epoch": 2.26, "grad_norm": 12.695565296883993, "learning_rate": 2.993316747568026e-06, "loss": 0.4209, "step": 14495 }, { "epoch": 2.26, "grad_norm": 17.785178843667577, "learning_rate": 2.992113623225086e-06, "loss": 0.5235, "step": 14496 }, { "epoch": 2.26, "grad_norm": 18.993745771290246, "learning_rate": 2.990910698182338e-06, "loss": 0.4338, "step": 14497 }, { "epoch": 2.26, "grad_norm": 22.035853734141558, "learning_rate": 2.9897079724739866e-06, "loss": 0.486, "step": 14498 }, { "epoch": 2.26, "grad_norm": 18.84878956762315, "learning_rate": 2.9885054461342423e-06, "loss": 0.4751, "step": 14499 }, { "epoch": 2.26, "grad_norm": 23.292491253505947, "learning_rate": 2.987303119197299e-06, "loss": 0.571, "step": 14500 }, { "epoch": 2.27, "grad_norm": 26.748964980414147, "learning_rate": 2.986100991697355e-06, "loss": 0.503, "step": 14501 }, { "epoch": 2.27, "grad_norm": 17.626106715288486, "learning_rate": 2.9848990636685937e-06, "loss": 0.4401, "step": 14502 }, { "epoch": 2.27, "grad_norm": 19.825547914341616, "learning_rate": 2.9836973351452014e-06, "loss": 0.4166, "step": 14503 }, { "epoch": 2.27, "grad_norm": 20.36021487039967, "learning_rate": 2.9824958061613497e-06, "loss": 0.5354, "step": 14504 }, { "epoch": 2.27, "grad_norm": 14.280289947932223, "learning_rate": 2.9812944767512132e-06, "loss": 0.4738, "step": 14505 }, { "epoch": 2.27, "grad_norm": 23.28615169192425, "learning_rate": 2.980093346948958e-06, "loss": 0.4233, "step": 14506 }, { "epoch": 2.27, "grad_norm": 14.652010355293557, "learning_rate": 2.978892416788738e-06, "loss": 0.4452, "step": 14507 }, { "epoch": 2.27, "grad_norm": 25.86114506152174, "learning_rate": 2.9776916863047146e-06, "loss": 0.513, "step": 14508 }, { "epoch": 2.27, "grad_norm": 18.08327401077484, "learning_rate": 2.976491155531027e-06, "loss": 0.3844, "step": 14509 }, { "epoch": 2.27, "grad_norm": 20.971513364263334, "learning_rate": 2.975290824501824e-06, "loss": 0.4355, "step": 14510 }, { "epoch": 2.27, "grad_norm": 24.137544789759183, "learning_rate": 2.9740906932512427e-06, "loss": 0.4916, "step": 14511 }, { "epoch": 2.27, "grad_norm": 21.506983227994724, "learning_rate": 2.9728907618134127e-06, "loss": 0.4789, "step": 14512 }, { "epoch": 2.27, "grad_norm": 26.674275112045073, "learning_rate": 2.9716910302224556e-06, "loss": 0.4751, "step": 14513 }, { "epoch": 2.27, "grad_norm": 24.903675721692522, "learning_rate": 2.970491498512493e-06, "loss": 0.4841, "step": 14514 }, { "epoch": 2.27, "grad_norm": 27.091565882834058, "learning_rate": 2.9692921667176443e-06, "loss": 0.495, "step": 14515 }, { "epoch": 2.27, "grad_norm": 23.164713719495037, "learning_rate": 2.968093034872014e-06, "loss": 0.4176, "step": 14516 }, { "epoch": 2.27, "grad_norm": 19.855802217633038, "learning_rate": 2.9668941030097e-06, "loss": 0.5089, "step": 14517 }, { "epoch": 2.27, "grad_norm": 18.513754404920117, "learning_rate": 2.965695371164804e-06, "loss": 0.4824, "step": 14518 }, { "epoch": 2.27, "grad_norm": 17.153989606483556, "learning_rate": 2.964496839371417e-06, "loss": 0.475, "step": 14519 }, { "epoch": 2.27, "grad_norm": 23.54719415535711, "learning_rate": 2.963298507663627e-06, "loss": 0.4146, "step": 14520 }, { "epoch": 2.27, "grad_norm": 15.587431353251151, "learning_rate": 2.9621003760755105e-06, "loss": 0.4653, "step": 14521 }, { "epoch": 2.27, "grad_norm": 17.765670173238664, "learning_rate": 2.9609024446411394e-06, "loss": 0.4635, "step": 14522 }, { "epoch": 2.27, "grad_norm": 33.189888607785285, "learning_rate": 2.9597047133945855e-06, "loss": 0.4649, "step": 14523 }, { "epoch": 2.27, "grad_norm": 17.530166295839656, "learning_rate": 2.9585071823699143e-06, "loss": 0.5239, "step": 14524 }, { "epoch": 2.27, "grad_norm": 19.76675658021288, "learning_rate": 2.9573098516011756e-06, "loss": 0.437, "step": 14525 }, { "epoch": 2.27, "grad_norm": 21.74801631731782, "learning_rate": 2.9561127211224283e-06, "loss": 0.4819, "step": 14526 }, { "epoch": 2.27, "grad_norm": 20.54625519662564, "learning_rate": 2.9549157909677116e-06, "loss": 0.4643, "step": 14527 }, { "epoch": 2.27, "grad_norm": 24.086998798894857, "learning_rate": 2.953719061171072e-06, "loss": 0.5299, "step": 14528 }, { "epoch": 2.27, "grad_norm": 21.47009228448441, "learning_rate": 2.9525225317665375e-06, "loss": 0.4952, "step": 14529 }, { "epoch": 2.27, "grad_norm": 20.299977458215515, "learning_rate": 2.9513262027881396e-06, "loss": 0.4498, "step": 14530 }, { "epoch": 2.27, "grad_norm": 24.708110349709333, "learning_rate": 2.9501300742699036e-06, "loss": 0.4598, "step": 14531 }, { "epoch": 2.27, "grad_norm": 18.90220833643359, "learning_rate": 2.948934146245841e-06, "loss": 0.4472, "step": 14532 }, { "epoch": 2.27, "grad_norm": 29.114908057001653, "learning_rate": 2.94773841874997e-06, "loss": 0.5214, "step": 14533 }, { "epoch": 2.27, "grad_norm": 25.30533818311973, "learning_rate": 2.94654289181629e-06, "loss": 0.5607, "step": 14534 }, { "epoch": 2.27, "grad_norm": 25.211873931090388, "learning_rate": 2.9453475654788032e-06, "loss": 0.4365, "step": 14535 }, { "epoch": 2.27, "grad_norm": 25.13345323023959, "learning_rate": 2.9441524397715094e-06, "loss": 0.4351, "step": 14536 }, { "epoch": 2.27, "grad_norm": 25.27511627659148, "learning_rate": 2.942957514728392e-06, "loss": 0.5023, "step": 14537 }, { "epoch": 2.27, "grad_norm": 11.422833012431038, "learning_rate": 2.9417627903834313e-06, "loss": 0.46, "step": 14538 }, { "epoch": 2.27, "grad_norm": 22.46845957842912, "learning_rate": 2.940568266770607e-06, "loss": 0.5018, "step": 14539 }, { "epoch": 2.27, "grad_norm": 30.015772882203493, "learning_rate": 2.9393739439238955e-06, "loss": 0.463, "step": 14540 }, { "epoch": 2.27, "grad_norm": 24.04022529065963, "learning_rate": 2.9381798218772594e-06, "loss": 0.4812, "step": 14541 }, { "epoch": 2.27, "grad_norm": 12.619839533049767, "learning_rate": 2.936985900664654e-06, "loss": 0.4966, "step": 14542 }, { "epoch": 2.27, "grad_norm": 19.654944393936418, "learning_rate": 2.9357921803200386e-06, "loss": 0.4499, "step": 14543 }, { "epoch": 2.27, "grad_norm": 24.553267308308644, "learning_rate": 2.93459866087736e-06, "loss": 0.5098, "step": 14544 }, { "epoch": 2.27, "grad_norm": 16.593171497541185, "learning_rate": 2.9334053423705666e-06, "loss": 0.4264, "step": 14545 }, { "epoch": 2.27, "grad_norm": 22.457023025407675, "learning_rate": 2.9322122248335905e-06, "loss": 0.4872, "step": 14546 }, { "epoch": 2.27, "grad_norm": 13.542750911649627, "learning_rate": 2.9310193083003614e-06, "loss": 0.4392, "step": 14547 }, { "epoch": 2.27, "grad_norm": 21.489082211932182, "learning_rate": 2.9298265928048086e-06, "loss": 0.4723, "step": 14548 }, { "epoch": 2.27, "grad_norm": 22.081831631044146, "learning_rate": 2.9286340783808543e-06, "loss": 0.5261, "step": 14549 }, { "epoch": 2.27, "grad_norm": 28.07680276868442, "learning_rate": 2.9274417650624097e-06, "loss": 0.4656, "step": 14550 }, { "epoch": 2.27, "grad_norm": 18.449567615829828, "learning_rate": 2.926249652883382e-06, "loss": 0.5031, "step": 14551 }, { "epoch": 2.27, "grad_norm": 16.441363449929163, "learning_rate": 2.925057741877675e-06, "loss": 0.431, "step": 14552 }, { "epoch": 2.27, "grad_norm": 27.545289395895157, "learning_rate": 2.92386603207919e-06, "loss": 0.4956, "step": 14553 }, { "epoch": 2.27, "grad_norm": 19.43183017108756, "learning_rate": 2.9226745235218134e-06, "loss": 0.446, "step": 14554 }, { "epoch": 2.27, "grad_norm": 19.270468638981548, "learning_rate": 2.921483216239436e-06, "loss": 0.4876, "step": 14555 }, { "epoch": 2.27, "grad_norm": 17.45729308615934, "learning_rate": 2.920292110265932e-06, "loss": 0.4861, "step": 14556 }, { "epoch": 2.27, "grad_norm": 27.16373185797766, "learning_rate": 2.9191012056351787e-06, "loss": 0.4482, "step": 14557 }, { "epoch": 2.27, "grad_norm": 18.588607991492953, "learning_rate": 2.917910502381048e-06, "loss": 0.4362, "step": 14558 }, { "epoch": 2.27, "grad_norm": 33.51284107837178, "learning_rate": 2.9167200005373976e-06, "loss": 0.5505, "step": 14559 }, { "epoch": 2.27, "grad_norm": 23.74736918384356, "learning_rate": 2.915529700138089e-06, "loss": 0.442, "step": 14560 }, { "epoch": 2.27, "grad_norm": 23.059512406185483, "learning_rate": 2.9143396012169677e-06, "loss": 0.4956, "step": 14561 }, { "epoch": 2.27, "grad_norm": 17.80184154521149, "learning_rate": 2.913149703807887e-06, "loss": 0.4861, "step": 14562 }, { "epoch": 2.27, "grad_norm": 23.87458520970228, "learning_rate": 2.9119600079446807e-06, "loss": 0.5143, "step": 14563 }, { "epoch": 2.27, "grad_norm": 20.69032047095443, "learning_rate": 2.910770513661184e-06, "loss": 0.4832, "step": 14564 }, { "epoch": 2.28, "grad_norm": 23.498750248229264, "learning_rate": 2.909581220991231e-06, "loss": 0.5434, "step": 14565 }, { "epoch": 2.28, "grad_norm": 18.0219330277558, "learning_rate": 2.9083921299686403e-06, "loss": 0.442, "step": 14566 }, { "epoch": 2.28, "grad_norm": 21.758304284939708, "learning_rate": 2.9072032406272243e-06, "loss": 0.4921, "step": 14567 }, { "epoch": 2.28, "grad_norm": 20.624430396421985, "learning_rate": 2.9060145530008013e-06, "loss": 0.4476, "step": 14568 }, { "epoch": 2.28, "grad_norm": 21.387851107157815, "learning_rate": 2.9048260671231733e-06, "loss": 0.4768, "step": 14569 }, { "epoch": 2.28, "grad_norm": 24.096355680163803, "learning_rate": 2.9036377830281435e-06, "loss": 0.4922, "step": 14570 }, { "epoch": 2.28, "grad_norm": 16.7804446582132, "learning_rate": 2.902449700749506e-06, "loss": 0.4775, "step": 14571 }, { "epoch": 2.28, "grad_norm": 20.879931748187758, "learning_rate": 2.901261820321042e-06, "loss": 0.5038, "step": 14572 }, { "epoch": 2.28, "grad_norm": 24.898715280707513, "learning_rate": 2.90007414177654e-06, "loss": 0.4704, "step": 14573 }, { "epoch": 2.28, "grad_norm": 35.882618256385136, "learning_rate": 2.8988866651497793e-06, "loss": 0.558, "step": 14574 }, { "epoch": 2.28, "grad_norm": 25.85289553343663, "learning_rate": 2.8976993904745277e-06, "loss": 0.4738, "step": 14575 }, { "epoch": 2.28, "grad_norm": 13.857581692419895, "learning_rate": 2.8965123177845476e-06, "loss": 0.479, "step": 14576 }, { "epoch": 2.28, "grad_norm": 20.481075737700465, "learning_rate": 2.8953254471136017e-06, "loss": 0.475, "step": 14577 }, { "epoch": 2.28, "grad_norm": 22.950848007157003, "learning_rate": 2.8941387784954476e-06, "loss": 0.5141, "step": 14578 }, { "epoch": 2.28, "grad_norm": 19.821943100227863, "learning_rate": 2.8929523119638257e-06, "loss": 0.4449, "step": 14579 }, { "epoch": 2.28, "grad_norm": 19.016087738776136, "learning_rate": 2.8917660475524867e-06, "loss": 0.5287, "step": 14580 }, { "epoch": 2.28, "grad_norm": 16.85752968946482, "learning_rate": 2.89057998529516e-06, "loss": 0.4505, "step": 14581 }, { "epoch": 2.28, "grad_norm": 18.64946738988201, "learning_rate": 2.8893941252255784e-06, "loss": 0.4415, "step": 14582 }, { "epoch": 2.28, "grad_norm": 15.013762917310192, "learning_rate": 2.8882084673774722e-06, "loss": 0.4651, "step": 14583 }, { "epoch": 2.28, "grad_norm": 33.37288981278104, "learning_rate": 2.8870230117845543e-06, "loss": 0.4923, "step": 14584 }, { "epoch": 2.28, "grad_norm": 26.872206953092036, "learning_rate": 2.885837758480543e-06, "loss": 0.4836, "step": 14585 }, { "epoch": 2.28, "grad_norm": 18.644142551234598, "learning_rate": 2.8846527074991427e-06, "loss": 0.4874, "step": 14586 }, { "epoch": 2.28, "grad_norm": 14.325005004417585, "learning_rate": 2.8834678588740585e-06, "loss": 0.4944, "step": 14587 }, { "epoch": 2.28, "grad_norm": 23.662272654185333, "learning_rate": 2.8822832126389834e-06, "loss": 0.4829, "step": 14588 }, { "epoch": 2.28, "grad_norm": 21.74556383159142, "learning_rate": 2.8810987688276092e-06, "loss": 0.4504, "step": 14589 }, { "epoch": 2.28, "grad_norm": 18.91071675869982, "learning_rate": 2.8799145274736253e-06, "loss": 0.4825, "step": 14590 }, { "epoch": 2.28, "grad_norm": 17.21911702937514, "learning_rate": 2.878730488610706e-06, "loss": 0.4193, "step": 14591 }, { "epoch": 2.28, "grad_norm": 19.372383373215637, "learning_rate": 2.8775466522725238e-06, "loss": 0.4663, "step": 14592 }, { "epoch": 2.28, "grad_norm": 13.75069859854102, "learning_rate": 2.876363018492747e-06, "loss": 0.4074, "step": 14593 }, { "epoch": 2.28, "grad_norm": 27.8066160873516, "learning_rate": 2.8751795873050405e-06, "loss": 0.5877, "step": 14594 }, { "epoch": 2.28, "grad_norm": 13.661622441894131, "learning_rate": 2.873996358743061e-06, "loss": 0.3917, "step": 14595 }, { "epoch": 2.28, "grad_norm": 15.017143054186315, "learning_rate": 2.872813332840456e-06, "loss": 0.4441, "step": 14596 }, { "epoch": 2.28, "grad_norm": 21.565455806318877, "learning_rate": 2.8716305096308674e-06, "loss": 0.4235, "step": 14597 }, { "epoch": 2.28, "grad_norm": 15.927970018637607, "learning_rate": 2.870447889147938e-06, "loss": 0.5055, "step": 14598 }, { "epoch": 2.28, "grad_norm": 20.692718635637814, "learning_rate": 2.869265471425302e-06, "loss": 0.4674, "step": 14599 }, { "epoch": 2.28, "grad_norm": 24.29028742650249, "learning_rate": 2.8680832564965853e-06, "loss": 0.4597, "step": 14600 }, { "epoch": 2.28, "grad_norm": 21.760241108927712, "learning_rate": 2.866901244395406e-06, "loss": 0.4893, "step": 14601 }, { "epoch": 2.28, "grad_norm": 23.41395928274524, "learning_rate": 2.8657194351553817e-06, "loss": 0.5354, "step": 14602 }, { "epoch": 2.28, "grad_norm": 18.196611814185918, "learning_rate": 2.8645378288101267e-06, "loss": 0.4533, "step": 14603 }, { "epoch": 2.28, "grad_norm": 16.016700890852174, "learning_rate": 2.863356425393242e-06, "loss": 0.4446, "step": 14604 }, { "epoch": 2.28, "grad_norm": 3.901529817688725, "learning_rate": 2.862175224938323e-06, "loss": 0.4981, "step": 14605 }, { "epoch": 2.28, "grad_norm": 23.93669328376025, "learning_rate": 2.860994227478964e-06, "loss": 0.4859, "step": 14606 }, { "epoch": 2.28, "grad_norm": 11.275653836699915, "learning_rate": 2.8598134330487536e-06, "loss": 0.4262, "step": 14607 }, { "epoch": 2.28, "grad_norm": 29.23866465692385, "learning_rate": 2.8586328416812746e-06, "loss": 0.479, "step": 14608 }, { "epoch": 2.28, "grad_norm": 18.303862175792137, "learning_rate": 2.857452453410101e-06, "loss": 0.5062, "step": 14609 }, { "epoch": 2.28, "grad_norm": 20.070503608147156, "learning_rate": 2.856272268268798e-06, "loss": 0.4274, "step": 14610 }, { "epoch": 2.28, "grad_norm": 24.416829497740473, "learning_rate": 2.8550922862909327e-06, "loss": 0.4651, "step": 14611 }, { "epoch": 2.28, "grad_norm": 53.24683257429877, "learning_rate": 2.853912507510066e-06, "loss": 0.4817, "step": 14612 }, { "epoch": 2.28, "grad_norm": 23.6163698181681, "learning_rate": 2.852732931959744e-06, "loss": 0.5131, "step": 14613 }, { "epoch": 2.28, "grad_norm": 21.85994082182788, "learning_rate": 2.85155355967352e-06, "loss": 0.4233, "step": 14614 }, { "epoch": 2.28, "grad_norm": 25.748390077405716, "learning_rate": 2.850374390684928e-06, "loss": 0.4581, "step": 14615 }, { "epoch": 2.28, "grad_norm": 16.71247043596305, "learning_rate": 2.849195425027509e-06, "loss": 0.577, "step": 14616 }, { "epoch": 2.28, "grad_norm": 18.36669564409993, "learning_rate": 2.8480166627347847e-06, "loss": 0.4809, "step": 14617 }, { "epoch": 2.28, "grad_norm": 21.013068291095262, "learning_rate": 2.8468381038402836e-06, "loss": 0.4681, "step": 14618 }, { "epoch": 2.28, "grad_norm": 21.458456391071184, "learning_rate": 2.845659748377526e-06, "loss": 0.5167, "step": 14619 }, { "epoch": 2.28, "grad_norm": 20.23870538157785, "learning_rate": 2.8444815963800164e-06, "loss": 0.5133, "step": 14620 }, { "epoch": 2.28, "grad_norm": 21.536674247422006, "learning_rate": 2.843303647881267e-06, "loss": 0.4429, "step": 14621 }, { "epoch": 2.28, "grad_norm": 18.70212843796634, "learning_rate": 2.842125902914773e-06, "loss": 0.5036, "step": 14622 }, { "epoch": 2.28, "grad_norm": 37.736611541272914, "learning_rate": 2.840948361514031e-06, "loss": 0.5101, "step": 14623 }, { "epoch": 2.28, "grad_norm": 18.606048121652947, "learning_rate": 2.839771023712532e-06, "loss": 0.4418, "step": 14624 }, { "epoch": 2.28, "grad_norm": 23.13825398493852, "learning_rate": 2.838593889543757e-06, "loss": 0.4611, "step": 14625 }, { "epoch": 2.28, "grad_norm": 18.694140225746896, "learning_rate": 2.8374169590411793e-06, "loss": 0.4813, "step": 14626 }, { "epoch": 2.28, "grad_norm": 24.938039316712928, "learning_rate": 2.836240232238274e-06, "loss": 0.49, "step": 14627 }, { "epoch": 2.28, "grad_norm": 18.672149234018818, "learning_rate": 2.8350637091685085e-06, "loss": 0.4539, "step": 14628 }, { "epoch": 2.29, "grad_norm": 22.839086796349058, "learning_rate": 2.8338873898653394e-06, "loss": 0.4864, "step": 14629 }, { "epoch": 2.29, "grad_norm": 22.876517253987576, "learning_rate": 2.8327112743622176e-06, "loss": 0.4772, "step": 14630 }, { "epoch": 2.29, "grad_norm": 29.03922370566224, "learning_rate": 2.8315353626925943e-06, "loss": 0.4416, "step": 14631 }, { "epoch": 2.29, "grad_norm": 25.506289814340988, "learning_rate": 2.8303596548899127e-06, "loss": 0.5167, "step": 14632 }, { "epoch": 2.29, "grad_norm": 23.45906272877004, "learning_rate": 2.829184150987612e-06, "loss": 0.4793, "step": 14633 }, { "epoch": 2.29, "grad_norm": 22.315959965353123, "learning_rate": 2.8280088510191174e-06, "loss": 0.5183, "step": 14634 }, { "epoch": 2.29, "grad_norm": 25.80685360739133, "learning_rate": 2.8268337550178537e-06, "loss": 0.4263, "step": 14635 }, { "epoch": 2.29, "grad_norm": 23.11465568640033, "learning_rate": 2.8256588630172412e-06, "loss": 0.4544, "step": 14636 }, { "epoch": 2.29, "grad_norm": 23.103817308213625, "learning_rate": 2.8244841750506967e-06, "loss": 0.4515, "step": 14637 }, { "epoch": 2.29, "grad_norm": 21.62563470366555, "learning_rate": 2.823309691151622e-06, "loss": 0.4442, "step": 14638 }, { "epoch": 2.29, "grad_norm": 26.325458061004305, "learning_rate": 2.8221354113534248e-06, "loss": 0.4463, "step": 14639 }, { "epoch": 2.29, "grad_norm": 25.48125682982296, "learning_rate": 2.8209613356894948e-06, "loss": 0.4377, "step": 14640 }, { "epoch": 2.29, "grad_norm": 20.810976844044117, "learning_rate": 2.8197874641932276e-06, "loss": 0.4796, "step": 14641 }, { "epoch": 2.29, "grad_norm": 18.447786302624664, "learning_rate": 2.818613796898001e-06, "loss": 0.4873, "step": 14642 }, { "epoch": 2.29, "grad_norm": 20.518754891044363, "learning_rate": 2.8174403338371968e-06, "loss": 0.4801, "step": 14643 }, { "epoch": 2.29, "grad_norm": 23.140094227148214, "learning_rate": 2.8162670750441924e-06, "loss": 0.487, "step": 14644 }, { "epoch": 2.29, "grad_norm": 16.97101558398097, "learning_rate": 2.815094020552346e-06, "loss": 0.5129, "step": 14645 }, { "epoch": 2.29, "grad_norm": 19.28130223562377, "learning_rate": 2.8139211703950255e-06, "loss": 0.6144, "step": 14646 }, { "epoch": 2.29, "grad_norm": 15.737472836752787, "learning_rate": 2.81274852460558e-06, "loss": 0.4402, "step": 14647 }, { "epoch": 2.29, "grad_norm": 19.490587807186053, "learning_rate": 2.8115760832173657e-06, "loss": 0.4549, "step": 14648 }, { "epoch": 2.29, "grad_norm": 15.184511076016577, "learning_rate": 2.8104038462637195e-06, "loss": 0.4527, "step": 14649 }, { "epoch": 2.29, "grad_norm": 16.66197647132942, "learning_rate": 2.809231813777984e-06, "loss": 0.4474, "step": 14650 }, { "epoch": 2.29, "grad_norm": 20.516430213623416, "learning_rate": 2.8080599857934876e-06, "loss": 0.5242, "step": 14651 }, { "epoch": 2.29, "grad_norm": 22.360133258856674, "learning_rate": 2.806888362343557e-06, "loss": 0.5067, "step": 14652 }, { "epoch": 2.29, "grad_norm": 14.378408133727232, "learning_rate": 2.805716943461517e-06, "loss": 0.4669, "step": 14653 }, { "epoch": 2.29, "grad_norm": 12.125692484808297, "learning_rate": 2.8045457291806787e-06, "loss": 0.3607, "step": 14654 }, { "epoch": 2.29, "grad_norm": 18.92809975105807, "learning_rate": 2.8033747195343474e-06, "loss": 0.5073, "step": 14655 }, { "epoch": 2.29, "grad_norm": 16.888604003515766, "learning_rate": 2.802203914555829e-06, "loss": 0.4903, "step": 14656 }, { "epoch": 2.29, "grad_norm": 24.927222034641275, "learning_rate": 2.8010333142784217e-06, "loss": 0.5385, "step": 14657 }, { "epoch": 2.29, "grad_norm": 18.71245044373287, "learning_rate": 2.7998629187354178e-06, "loss": 0.4453, "step": 14658 }, { "epoch": 2.29, "grad_norm": 14.554409654087648, "learning_rate": 2.798692727960101e-06, "loss": 0.4856, "step": 14659 }, { "epoch": 2.29, "grad_norm": 16.986017007424994, "learning_rate": 2.7975227419857466e-06, "loss": 0.4349, "step": 14660 }, { "epoch": 2.29, "grad_norm": 24.87303181061066, "learning_rate": 2.796352960845632e-06, "loss": 0.4533, "step": 14661 }, { "epoch": 2.29, "grad_norm": 29.373021142333087, "learning_rate": 2.795183384573029e-06, "loss": 0.4723, "step": 14662 }, { "epoch": 2.29, "grad_norm": 15.222709169900291, "learning_rate": 2.7940140132011963e-06, "loss": 0.4216, "step": 14663 }, { "epoch": 2.29, "grad_norm": 23.032443900070554, "learning_rate": 2.792844846763385e-06, "loss": 0.4792, "step": 14664 }, { "epoch": 2.29, "grad_norm": 22.9586632307366, "learning_rate": 2.7916758852928514e-06, "loss": 0.4608, "step": 14665 }, { "epoch": 2.29, "grad_norm": 24.399868598568535, "learning_rate": 2.7905071288228415e-06, "loss": 0.4484, "step": 14666 }, { "epoch": 2.29, "grad_norm": 28.753317653444675, "learning_rate": 2.789338577386589e-06, "loss": 0.5067, "step": 14667 }, { "epoch": 2.29, "grad_norm": 18.383428350103912, "learning_rate": 2.788170231017332e-06, "loss": 0.4084, "step": 14668 }, { "epoch": 2.29, "grad_norm": 31.82335384846937, "learning_rate": 2.787002089748291e-06, "loss": 0.4722, "step": 14669 }, { "epoch": 2.29, "grad_norm": 24.38579140469859, "learning_rate": 2.785834153612692e-06, "loss": 0.4914, "step": 14670 }, { "epoch": 2.29, "grad_norm": 16.433195744947856, "learning_rate": 2.7846664226437524e-06, "loss": 0.4063, "step": 14671 }, { "epoch": 2.29, "grad_norm": 18.670153166576075, "learning_rate": 2.783498896874676e-06, "loss": 0.4822, "step": 14672 }, { "epoch": 2.29, "grad_norm": 26.748641396273573, "learning_rate": 2.7823315763386715e-06, "loss": 0.4548, "step": 14673 }, { "epoch": 2.29, "grad_norm": 17.134115833809354, "learning_rate": 2.7811644610689327e-06, "loss": 0.4557, "step": 14674 }, { "epoch": 2.29, "grad_norm": 22.29928887699729, "learning_rate": 2.7799975510986566e-06, "loss": 0.4867, "step": 14675 }, { "epoch": 2.29, "grad_norm": 18.249070043826354, "learning_rate": 2.7788308464610225e-06, "loss": 0.4484, "step": 14676 }, { "epoch": 2.29, "grad_norm": 21.684559613134798, "learning_rate": 2.777664347189216e-06, "loss": 0.4743, "step": 14677 }, { "epoch": 2.29, "grad_norm": 18.679171289820214, "learning_rate": 2.776498053316413e-06, "loss": 0.4336, "step": 14678 }, { "epoch": 2.29, "grad_norm": 15.897845134899944, "learning_rate": 2.7753319648757802e-06, "loss": 0.4376, "step": 14679 }, { "epoch": 2.29, "grad_norm": 23.94907233449032, "learning_rate": 2.7741660819004766e-06, "loss": 0.493, "step": 14680 }, { "epoch": 2.29, "grad_norm": 16.118537295277786, "learning_rate": 2.7730004044236636e-06, "loss": 0.3929, "step": 14681 }, { "epoch": 2.29, "grad_norm": 19.018668022160856, "learning_rate": 2.7718349324784897e-06, "loss": 0.4568, "step": 14682 }, { "epoch": 2.29, "grad_norm": 21.37691721558114, "learning_rate": 2.770669666098107e-06, "loss": 0.545, "step": 14683 }, { "epoch": 2.29, "grad_norm": 28.157284896494332, "learning_rate": 2.769504605315648e-06, "loss": 0.4321, "step": 14684 }, { "epoch": 2.29, "grad_norm": 19.420540112757475, "learning_rate": 2.7683397501642463e-06, "loss": 0.4862, "step": 14685 }, { "epoch": 2.29, "grad_norm": 16.18713358101687, "learning_rate": 2.767175100677032e-06, "loss": 0.4293, "step": 14686 }, { "epoch": 2.29, "grad_norm": 21.996247025641516, "learning_rate": 2.7660106568871293e-06, "loss": 0.4966, "step": 14687 }, { "epoch": 2.29, "grad_norm": 21.484005550675867, "learning_rate": 2.7648464188276514e-06, "loss": 0.4995, "step": 14688 }, { "epoch": 2.29, "grad_norm": 18.23665222348996, "learning_rate": 2.7636823865317065e-06, "loss": 0.4938, "step": 14689 }, { "epoch": 2.29, "grad_norm": 19.48146588649827, "learning_rate": 2.7625185600324013e-06, "loss": 0.4615, "step": 14690 }, { "epoch": 2.29, "grad_norm": 11.65652048721396, "learning_rate": 2.761354939362838e-06, "loss": 0.4101, "step": 14691 }, { "epoch": 2.29, "grad_norm": 17.754384520145393, "learning_rate": 2.7601915245561017e-06, "loss": 0.4562, "step": 14692 }, { "epoch": 2.3, "grad_norm": 18.28091433127257, "learning_rate": 2.7590283156452867e-06, "loss": 0.4447, "step": 14693 }, { "epoch": 2.3, "grad_norm": 18.169771750070954, "learning_rate": 2.757865312663468e-06, "loss": 0.5174, "step": 14694 }, { "epoch": 2.3, "grad_norm": 12.597543016264888, "learning_rate": 2.7567025156437223e-06, "loss": 0.4409, "step": 14695 }, { "epoch": 2.3, "grad_norm": 16.911334087744464, "learning_rate": 2.755539924619124e-06, "loss": 0.4581, "step": 14696 }, { "epoch": 2.3, "grad_norm": 17.502788704677496, "learning_rate": 2.754377539622731e-06, "loss": 0.4437, "step": 14697 }, { "epoch": 2.3, "grad_norm": 24.431116527730488, "learning_rate": 2.7532153606876e-06, "loss": 0.4441, "step": 14698 }, { "epoch": 2.3, "grad_norm": 20.527162249277215, "learning_rate": 2.7520533878467846e-06, "loss": 0.4626, "step": 14699 }, { "epoch": 2.3, "grad_norm": 23.97678055389082, "learning_rate": 2.7508916211333346e-06, "loss": 0.4527, "step": 14700 }, { "epoch": 2.3, "grad_norm": 24.68243090762094, "learning_rate": 2.7497300605802823e-06, "loss": 0.4519, "step": 14701 }, { "epoch": 2.3, "grad_norm": 20.909088308455285, "learning_rate": 2.74856870622067e-06, "loss": 0.5036, "step": 14702 }, { "epoch": 2.3, "grad_norm": 23.995748140782542, "learning_rate": 2.7474075580875182e-06, "loss": 0.5154, "step": 14703 }, { "epoch": 2.3, "grad_norm": 18.834694657208953, "learning_rate": 2.7462466162138557e-06, "loss": 0.4256, "step": 14704 }, { "epoch": 2.3, "grad_norm": 27.08003131299765, "learning_rate": 2.745085880632693e-06, "loss": 0.5153, "step": 14705 }, { "epoch": 2.3, "grad_norm": 20.49231549338284, "learning_rate": 2.7439253513770434e-06, "loss": 0.484, "step": 14706 }, { "epoch": 2.3, "grad_norm": 15.393991719809545, "learning_rate": 2.7427650284799157e-06, "loss": 0.4556, "step": 14707 }, { "epoch": 2.3, "grad_norm": 23.16881591696427, "learning_rate": 2.7416049119743025e-06, "loss": 0.4203, "step": 14708 }, { "epoch": 2.3, "grad_norm": 22.339026026426843, "learning_rate": 2.7404450018932027e-06, "loss": 0.5047, "step": 14709 }, { "epoch": 2.3, "grad_norm": 26.102454892123095, "learning_rate": 2.739285298269596e-06, "loss": 0.5579, "step": 14710 }, { "epoch": 2.3, "grad_norm": 19.587943772420655, "learning_rate": 2.7381258011364697e-06, "loss": 0.4283, "step": 14711 }, { "epoch": 2.3, "grad_norm": 15.980272688650185, "learning_rate": 2.7369665105267995e-06, "loss": 0.4591, "step": 14712 }, { "epoch": 2.3, "grad_norm": 27.826995916663137, "learning_rate": 2.7358074264735534e-06, "loss": 0.4661, "step": 14713 }, { "epoch": 2.3, "grad_norm": 28.78130568686368, "learning_rate": 2.7346485490096908e-06, "loss": 0.4651, "step": 14714 }, { "epoch": 2.3, "grad_norm": 22.15933784413965, "learning_rate": 2.733489878168175e-06, "loss": 0.4658, "step": 14715 }, { "epoch": 2.3, "grad_norm": 18.930780814506697, "learning_rate": 2.732331413981958e-06, "loss": 0.4281, "step": 14716 }, { "epoch": 2.3, "grad_norm": 15.150278522548328, "learning_rate": 2.731173156483985e-06, "loss": 0.3993, "step": 14717 }, { "epoch": 2.3, "grad_norm": 25.11578189245044, "learning_rate": 2.730015105707191e-06, "loss": 0.5256, "step": 14718 }, { "epoch": 2.3, "grad_norm": 18.060144420985747, "learning_rate": 2.7288572616845154e-06, "loss": 0.4226, "step": 14719 }, { "epoch": 2.3, "grad_norm": 22.270587037624495, "learning_rate": 2.7276996244488864e-06, "loss": 0.4851, "step": 14720 }, { "epoch": 2.3, "grad_norm": 23.69489096465625, "learning_rate": 2.7265421940332283e-06, "loss": 0.4996, "step": 14721 }, { "epoch": 2.3, "grad_norm": 17.65475279781367, "learning_rate": 2.725384970470455e-06, "loss": 0.4392, "step": 14722 }, { "epoch": 2.3, "grad_norm": 23.89639241926165, "learning_rate": 2.724227953793475e-06, "loss": 0.4943, "step": 14723 }, { "epoch": 2.3, "grad_norm": 19.830173977256038, "learning_rate": 2.7230711440351952e-06, "loss": 0.4506, "step": 14724 }, { "epoch": 2.3, "grad_norm": 18.002047370994212, "learning_rate": 2.7219145412285187e-06, "loss": 0.5654, "step": 14725 }, { "epoch": 2.3, "grad_norm": 30.83761012448493, "learning_rate": 2.7207581454063314e-06, "loss": 0.4628, "step": 14726 }, { "epoch": 2.3, "grad_norm": 19.420598344895975, "learning_rate": 2.7196019566015286e-06, "loss": 0.4241, "step": 14727 }, { "epoch": 2.3, "grad_norm": 19.726341596820642, "learning_rate": 2.7184459748469836e-06, "loss": 0.442, "step": 14728 }, { "epoch": 2.3, "grad_norm": 28.157998006417163, "learning_rate": 2.7172902001755787e-06, "loss": 0.4655, "step": 14729 }, { "epoch": 2.3, "grad_norm": 18.637164559726063, "learning_rate": 2.7161346326201763e-06, "loss": 0.4612, "step": 14730 }, { "epoch": 2.3, "grad_norm": 19.990402636917032, "learning_rate": 2.7149792722136446e-06, "loss": 0.424, "step": 14731 }, { "epoch": 2.3, "grad_norm": 17.8277824629726, "learning_rate": 2.713824118988844e-06, "loss": 0.4213, "step": 14732 }, { "epoch": 2.3, "grad_norm": 22.966103460986012, "learning_rate": 2.7126691729786202e-06, "loss": 0.3914, "step": 14733 }, { "epoch": 2.3, "grad_norm": 18.871155453871058, "learning_rate": 2.711514434215825e-06, "loss": 0.4956, "step": 14734 }, { "epoch": 2.3, "grad_norm": 18.273026851950597, "learning_rate": 2.710359902733293e-06, "loss": 0.4185, "step": 14735 }, { "epoch": 2.3, "grad_norm": 22.39572743871331, "learning_rate": 2.709205578563859e-06, "loss": 0.4317, "step": 14736 }, { "epoch": 2.3, "grad_norm": 23.75973473988637, "learning_rate": 2.7080514617403584e-06, "loss": 0.485, "step": 14737 }, { "epoch": 2.3, "grad_norm": 22.986755810526127, "learning_rate": 2.706897552295609e-06, "loss": 0.4726, "step": 14738 }, { "epoch": 2.3, "grad_norm": 18.84436533763036, "learning_rate": 2.705743850262422e-06, "loss": 0.4531, "step": 14739 }, { "epoch": 2.3, "grad_norm": 21.006405353056806, "learning_rate": 2.704590355673614e-06, "loss": 0.4994, "step": 14740 }, { "epoch": 2.3, "grad_norm": 19.168732031082126, "learning_rate": 2.7034370685619894e-06, "loss": 0.4639, "step": 14741 }, { "epoch": 2.3, "grad_norm": 25.70326264063381, "learning_rate": 2.7022839889603514e-06, "loss": 0.5261, "step": 14742 }, { "epoch": 2.3, "grad_norm": 26.200516599920157, "learning_rate": 2.7011311169014807e-06, "loss": 0.5424, "step": 14743 }, { "epoch": 2.3, "grad_norm": 20.137120982314368, "learning_rate": 2.6999784524181727e-06, "loss": 0.4332, "step": 14744 }, { "epoch": 2.3, "grad_norm": 18.503260156884863, "learning_rate": 2.698825995543207e-06, "loss": 0.4426, "step": 14745 }, { "epoch": 2.3, "grad_norm": 17.890467453975056, "learning_rate": 2.6976737463093617e-06, "loss": 0.5504, "step": 14746 }, { "epoch": 2.3, "grad_norm": 3.410389858341277, "learning_rate": 2.696521704749403e-06, "loss": 0.4595, "step": 14747 }, { "epoch": 2.3, "grad_norm": 28.532383847138505, "learning_rate": 2.6953698708960916e-06, "loss": 0.4939, "step": 14748 }, { "epoch": 2.3, "grad_norm": 24.730443736033425, "learning_rate": 2.6942182447821884e-06, "loss": 0.5174, "step": 14749 }, { "epoch": 2.3, "grad_norm": 19.030762659920697, "learning_rate": 2.693066826440447e-06, "loss": 0.3762, "step": 14750 }, { "epoch": 2.3, "grad_norm": 20.309325667082327, "learning_rate": 2.691915615903611e-06, "loss": 0.4684, "step": 14751 }, { "epoch": 2.3, "grad_norm": 15.336966322184672, "learning_rate": 2.6907646132044174e-06, "loss": 0.4399, "step": 14752 }, { "epoch": 2.3, "grad_norm": 22.96643673056194, "learning_rate": 2.6896138183756015e-06, "loss": 0.466, "step": 14753 }, { "epoch": 2.3, "grad_norm": 28.01286912651057, "learning_rate": 2.6884632314498947e-06, "loss": 0.5197, "step": 14754 }, { "epoch": 2.3, "grad_norm": 28.32817280360805, "learning_rate": 2.6873128524600144e-06, "loss": 0.4914, "step": 14755 }, { "epoch": 2.3, "grad_norm": 24.038694696894883, "learning_rate": 2.686162681438682e-06, "loss": 0.4478, "step": 14756 }, { "epoch": 2.31, "grad_norm": 20.72828883096704, "learning_rate": 2.6850127184186015e-06, "loss": 0.4786, "step": 14757 }, { "epoch": 2.31, "grad_norm": 15.499255774401032, "learning_rate": 2.6838629634324796e-06, "loss": 0.4649, "step": 14758 }, { "epoch": 2.31, "grad_norm": 20.49377016157838, "learning_rate": 2.6827134165130175e-06, "loss": 0.4751, "step": 14759 }, { "epoch": 2.31, "grad_norm": 33.80068617339733, "learning_rate": 2.681564077692903e-06, "loss": 0.465, "step": 14760 }, { "epoch": 2.31, "grad_norm": 19.10800867350413, "learning_rate": 2.6804149470048293e-06, "loss": 0.4257, "step": 14761 }, { "epoch": 2.31, "grad_norm": 18.92176748177484, "learning_rate": 2.6792660244814683e-06, "loss": 0.5182, "step": 14762 }, { "epoch": 2.31, "grad_norm": 21.61863688678932, "learning_rate": 2.6781173101555014e-06, "loss": 0.5147, "step": 14763 }, { "epoch": 2.31, "grad_norm": 22.51446621022123, "learning_rate": 2.6769688040595933e-06, "loss": 0.418, "step": 14764 }, { "epoch": 2.31, "grad_norm": 35.37467467424671, "learning_rate": 2.6758205062264076e-06, "loss": 0.6174, "step": 14765 }, { "epoch": 2.31, "grad_norm": 17.19856184104096, "learning_rate": 2.6746724166886063e-06, "loss": 0.4986, "step": 14766 }, { "epoch": 2.31, "grad_norm": 21.73393693066997, "learning_rate": 2.6735245354788354e-06, "loss": 0.4074, "step": 14767 }, { "epoch": 2.31, "grad_norm": 29.977719126804683, "learning_rate": 2.6723768626297387e-06, "loss": 0.5491, "step": 14768 }, { "epoch": 2.31, "grad_norm": 15.036516328754196, "learning_rate": 2.671229398173957e-06, "loss": 0.4181, "step": 14769 }, { "epoch": 2.31, "grad_norm": 18.795193939475638, "learning_rate": 2.670082142144124e-06, "loss": 0.4695, "step": 14770 }, { "epoch": 2.31, "grad_norm": 21.281777913103817, "learning_rate": 2.66893509457287e-06, "loss": 0.4283, "step": 14771 }, { "epoch": 2.31, "grad_norm": 18.715814940672992, "learning_rate": 2.6677882554928126e-06, "loss": 0.4336, "step": 14772 }, { "epoch": 2.31, "grad_norm": 18.29970615856538, "learning_rate": 2.666641624936566e-06, "loss": 0.4705, "step": 14773 }, { "epoch": 2.31, "grad_norm": 25.7762620869203, "learning_rate": 2.665495202936741e-06, "loss": 0.3934, "step": 14774 }, { "epoch": 2.31, "grad_norm": 14.45398289686087, "learning_rate": 2.6643489895259454e-06, "loss": 0.4339, "step": 14775 }, { "epoch": 2.31, "grad_norm": 31.787859629787935, "learning_rate": 2.6632029847367734e-06, "loss": 0.4828, "step": 14776 }, { "epoch": 2.31, "grad_norm": 17.304642750443232, "learning_rate": 2.662057188601813e-06, "loss": 0.4754, "step": 14777 }, { "epoch": 2.31, "grad_norm": 21.273346968341556, "learning_rate": 2.660911601153654e-06, "loss": 0.4228, "step": 14778 }, { "epoch": 2.31, "grad_norm": 16.552807188520944, "learning_rate": 2.659766222424879e-06, "loss": 0.4679, "step": 14779 }, { "epoch": 2.31, "grad_norm": 17.11000004850306, "learning_rate": 2.658621052448055e-06, "loss": 0.4796, "step": 14780 }, { "epoch": 2.31, "grad_norm": 18.583140371604497, "learning_rate": 2.6574760912557575e-06, "loss": 0.4617, "step": 14781 }, { "epoch": 2.31, "grad_norm": 22.13267904780665, "learning_rate": 2.656331338880541e-06, "loss": 0.4948, "step": 14782 }, { "epoch": 2.31, "grad_norm": 15.483142060774096, "learning_rate": 2.6551867953549657e-06, "loss": 0.4608, "step": 14783 }, { "epoch": 2.31, "grad_norm": 19.424854003297458, "learning_rate": 2.6540424607115845e-06, "loss": 0.4336, "step": 14784 }, { "epoch": 2.31, "grad_norm": 17.966072058776174, "learning_rate": 2.6528983349829353e-06, "loss": 0.4938, "step": 14785 }, { "epoch": 2.31, "grad_norm": 25.3838812854603, "learning_rate": 2.651754418201563e-06, "loss": 0.5059, "step": 14786 }, { "epoch": 2.31, "grad_norm": 20.55757092470731, "learning_rate": 2.6506107103999924e-06, "loss": 0.5013, "step": 14787 }, { "epoch": 2.31, "grad_norm": 21.67083547969026, "learning_rate": 2.6494672116107577e-06, "loss": 0.4672, "step": 14788 }, { "epoch": 2.31, "grad_norm": 12.768631326653288, "learning_rate": 2.6483239218663724e-06, "loss": 0.4431, "step": 14789 }, { "epoch": 2.31, "grad_norm": 23.775635248215437, "learning_rate": 2.647180841199354e-06, "loss": 0.4165, "step": 14790 }, { "epoch": 2.31, "grad_norm": 20.37594359719969, "learning_rate": 2.6460379696422146e-06, "loss": 0.5215, "step": 14791 }, { "epoch": 2.31, "grad_norm": 17.149163491907018, "learning_rate": 2.6448953072274532e-06, "loss": 0.4261, "step": 14792 }, { "epoch": 2.31, "grad_norm": 27.66035867979796, "learning_rate": 2.6437528539875633e-06, "loss": 0.4685, "step": 14793 }, { "epoch": 2.31, "grad_norm": 18.53604255743931, "learning_rate": 2.6426106099550396e-06, "loss": 0.4199, "step": 14794 }, { "epoch": 2.31, "grad_norm": 20.808610632910877, "learning_rate": 2.6414685751623682e-06, "loss": 0.5242, "step": 14795 }, { "epoch": 2.31, "grad_norm": 26.38101476407635, "learning_rate": 2.6403267496420238e-06, "loss": 0.4226, "step": 14796 }, { "epoch": 2.31, "grad_norm": 20.953883919708073, "learning_rate": 2.639185133426484e-06, "loss": 0.4586, "step": 14797 }, { "epoch": 2.31, "grad_norm": 24.96119778576508, "learning_rate": 2.6380437265482107e-06, "loss": 0.4967, "step": 14798 }, { "epoch": 2.31, "grad_norm": 19.063426896917715, "learning_rate": 2.6369025290396664e-06, "loss": 0.471, "step": 14799 }, { "epoch": 2.31, "grad_norm": 24.121129323338522, "learning_rate": 2.6357615409333092e-06, "loss": 0.5235, "step": 14800 }, { "epoch": 2.31, "grad_norm": 28.479236777572076, "learning_rate": 2.634620762261587e-06, "loss": 0.5212, "step": 14801 }, { "epoch": 2.31, "grad_norm": 26.14978696533939, "learning_rate": 2.6334801930569387e-06, "loss": 0.4359, "step": 14802 }, { "epoch": 2.31, "grad_norm": 15.385133531371865, "learning_rate": 2.6323398333518047e-06, "loss": 0.3778, "step": 14803 }, { "epoch": 2.31, "grad_norm": 18.217725480424736, "learning_rate": 2.631199683178615e-06, "loss": 0.4403, "step": 14804 }, { "epoch": 2.31, "grad_norm": 15.753549080849846, "learning_rate": 2.6300597425698015e-06, "loss": 0.4559, "step": 14805 }, { "epoch": 2.31, "grad_norm": 23.199802247465566, "learning_rate": 2.6289200115577727e-06, "loss": 0.469, "step": 14806 }, { "epoch": 2.31, "grad_norm": 16.925972175709823, "learning_rate": 2.627780490174946e-06, "loss": 0.4779, "step": 14807 }, { "epoch": 2.31, "grad_norm": 23.888678641712463, "learning_rate": 2.6266411784537294e-06, "loss": 0.4734, "step": 14808 }, { "epoch": 2.31, "grad_norm": 29.56573948742745, "learning_rate": 2.625502076426528e-06, "loss": 0.4755, "step": 14809 }, { "epoch": 2.31, "grad_norm": 24.176629647454032, "learning_rate": 2.6243631841257332e-06, "loss": 0.486, "step": 14810 }, { "epoch": 2.31, "grad_norm": 17.585281193322523, "learning_rate": 2.6232245015837323e-06, "loss": 0.4237, "step": 14811 }, { "epoch": 2.31, "grad_norm": 23.804607391848606, "learning_rate": 2.6220860288329108e-06, "loss": 0.4012, "step": 14812 }, { "epoch": 2.31, "grad_norm": 17.361003595040447, "learning_rate": 2.6209477659056503e-06, "loss": 0.4458, "step": 14813 }, { "epoch": 2.31, "grad_norm": 29.78269713009057, "learning_rate": 2.6198097128343147e-06, "loss": 0.4823, "step": 14814 }, { "epoch": 2.31, "grad_norm": 27.436299516911774, "learning_rate": 2.618671869651278e-06, "loss": 0.4859, "step": 14815 }, { "epoch": 2.31, "grad_norm": 26.03135524398759, "learning_rate": 2.617534236388891e-06, "loss": 0.426, "step": 14816 }, { "epoch": 2.31, "grad_norm": 24.54870931885095, "learning_rate": 2.616396813079516e-06, "loss": 0.5183, "step": 14817 }, { "epoch": 2.31, "grad_norm": 19.097912615598116, "learning_rate": 2.615259599755492e-06, "loss": 0.4959, "step": 14818 }, { "epoch": 2.31, "grad_norm": 18.04726304797163, "learning_rate": 2.614122596449166e-06, "loss": 0.4563, "step": 14819 }, { "epoch": 2.31, "grad_norm": 22.833553091845644, "learning_rate": 2.6129858031928746e-06, "loss": 0.4487, "step": 14820 }, { "epoch": 2.32, "grad_norm": 18.616677906779472, "learning_rate": 2.611849220018944e-06, "loss": 0.4976, "step": 14821 }, { "epoch": 2.32, "grad_norm": 22.643266251972197, "learning_rate": 2.610712846959702e-06, "loss": 0.4414, "step": 14822 }, { "epoch": 2.32, "grad_norm": 15.073181399367094, "learning_rate": 2.609576684047461e-06, "loss": 0.4214, "step": 14823 }, { "epoch": 2.32, "grad_norm": 27.644867918566085, "learning_rate": 2.6084407313145354e-06, "loss": 0.5168, "step": 14824 }, { "epoch": 2.32, "grad_norm": 24.276763378764976, "learning_rate": 2.6073049887932357e-06, "loss": 0.4718, "step": 14825 }, { "epoch": 2.32, "grad_norm": 17.10417396931099, "learning_rate": 2.606169456515857e-06, "loss": 0.4461, "step": 14826 }, { "epoch": 2.32, "grad_norm": 18.48619740757821, "learning_rate": 2.6050341345146902e-06, "loss": 0.4862, "step": 14827 }, { "epoch": 2.32, "grad_norm": 25.439154093150872, "learning_rate": 2.6038990228220275e-06, "loss": 0.478, "step": 14828 }, { "epoch": 2.32, "grad_norm": 20.906699752389017, "learning_rate": 2.6027641214701495e-06, "loss": 0.4522, "step": 14829 }, { "epoch": 2.32, "grad_norm": 20.301037713919737, "learning_rate": 2.601629430491338e-06, "loss": 0.4749, "step": 14830 }, { "epoch": 2.32, "grad_norm": 15.233602452374694, "learning_rate": 2.6004949499178522e-06, "loss": 0.4284, "step": 14831 }, { "epoch": 2.32, "grad_norm": 22.47842621460263, "learning_rate": 2.5993606797819616e-06, "loss": 0.4687, "step": 14832 }, { "epoch": 2.32, "grad_norm": 15.425854689750977, "learning_rate": 2.5982266201159236e-06, "loss": 0.4504, "step": 14833 }, { "epoch": 2.32, "grad_norm": 21.461040963172433, "learning_rate": 2.5970927709519942e-06, "loss": 0.5189, "step": 14834 }, { "epoch": 2.32, "grad_norm": 22.111971924386424, "learning_rate": 2.5959591323224152e-06, "loss": 0.4433, "step": 14835 }, { "epoch": 2.32, "grad_norm": 17.55038206833877, "learning_rate": 2.5948257042594248e-06, "loss": 0.4911, "step": 14836 }, { "epoch": 2.32, "grad_norm": 20.941442181129986, "learning_rate": 2.5936924867952596e-06, "loss": 0.491, "step": 14837 }, { "epoch": 2.32, "grad_norm": 28.30753601001517, "learning_rate": 2.5925594799621503e-06, "loss": 0.4588, "step": 14838 }, { "epoch": 2.32, "grad_norm": 15.024396194283737, "learning_rate": 2.5914266837923153e-06, "loss": 0.4093, "step": 14839 }, { "epoch": 2.32, "grad_norm": 21.974316825679026, "learning_rate": 2.5902940983179694e-06, "loss": 0.5072, "step": 14840 }, { "epoch": 2.32, "grad_norm": 22.16822372357109, "learning_rate": 2.589161723571324e-06, "loss": 0.4847, "step": 14841 }, { "epoch": 2.32, "grad_norm": 17.215014280924883, "learning_rate": 2.5880295595845885e-06, "loss": 0.5208, "step": 14842 }, { "epoch": 2.32, "grad_norm": 22.876328135326183, "learning_rate": 2.586897606389952e-06, "loss": 0.4503, "step": 14843 }, { "epoch": 2.32, "grad_norm": 34.072124884356555, "learning_rate": 2.585765864019615e-06, "loss": 0.408, "step": 14844 }, { "epoch": 2.32, "grad_norm": 18.428078083498058, "learning_rate": 2.5846343325057556e-06, "loss": 0.4284, "step": 14845 }, { "epoch": 2.32, "grad_norm": 15.794546595276005, "learning_rate": 2.583503011880558e-06, "loss": 0.4672, "step": 14846 }, { "epoch": 2.32, "grad_norm": 30.61878933990655, "learning_rate": 2.5823719021761993e-06, "loss": 0.5086, "step": 14847 }, { "epoch": 2.32, "grad_norm": 26.845127475043842, "learning_rate": 2.5812410034248426e-06, "loss": 0.4499, "step": 14848 }, { "epoch": 2.32, "grad_norm": 24.689449787403717, "learning_rate": 2.5801103156586536e-06, "loss": 0.4536, "step": 14849 }, { "epoch": 2.32, "grad_norm": 21.768953118210405, "learning_rate": 2.5789798389097855e-06, "loss": 0.4558, "step": 14850 }, { "epoch": 2.32, "grad_norm": 14.603273176706368, "learning_rate": 2.5778495732103914e-06, "loss": 0.4389, "step": 14851 }, { "epoch": 2.32, "grad_norm": 15.59337215787625, "learning_rate": 2.5767195185926098e-06, "loss": 0.4187, "step": 14852 }, { "epoch": 2.32, "grad_norm": 18.49086633224654, "learning_rate": 2.5755896750885834e-06, "loss": 0.4934, "step": 14853 }, { "epoch": 2.32, "grad_norm": 21.443562248250814, "learning_rate": 2.5744600427304467e-06, "loss": 0.4737, "step": 14854 }, { "epoch": 2.32, "grad_norm": 27.265632918228267, "learning_rate": 2.573330621550323e-06, "loss": 0.4643, "step": 14855 }, { "epoch": 2.32, "grad_norm": 18.102236767700052, "learning_rate": 2.5722014115803284e-06, "loss": 0.5043, "step": 14856 }, { "epoch": 2.32, "grad_norm": 42.18891286850637, "learning_rate": 2.57107241285258e-06, "loss": 0.4472, "step": 14857 }, { "epoch": 2.32, "grad_norm": 24.464202794195778, "learning_rate": 2.569943625399187e-06, "loss": 0.4397, "step": 14858 }, { "epoch": 2.32, "grad_norm": 14.301314694360217, "learning_rate": 2.5688150492522533e-06, "loss": 0.4834, "step": 14859 }, { "epoch": 2.32, "grad_norm": 38.41842983805012, "learning_rate": 2.567686684443873e-06, "loss": 0.5286, "step": 14860 }, { "epoch": 2.32, "grad_norm": 21.73256821324706, "learning_rate": 2.5665585310061327e-06, "loss": 0.4968, "step": 14861 }, { "epoch": 2.32, "grad_norm": 14.696909525666674, "learning_rate": 2.5654305889711182e-06, "loss": 0.3983, "step": 14862 }, { "epoch": 2.32, "grad_norm": 18.193662637424165, "learning_rate": 2.564302858370913e-06, "loss": 0.5031, "step": 14863 }, { "epoch": 2.32, "grad_norm": 23.51864185688541, "learning_rate": 2.5631753392375846e-06, "loss": 0.438, "step": 14864 }, { "epoch": 2.32, "grad_norm": 15.578503771494018, "learning_rate": 2.562048031603196e-06, "loss": 0.4669, "step": 14865 }, { "epoch": 2.32, "grad_norm": 24.248392692124835, "learning_rate": 2.5609209354998097e-06, "loss": 0.5014, "step": 14866 }, { "epoch": 2.32, "grad_norm": 15.274470300706877, "learning_rate": 2.5597940509594845e-06, "loss": 0.4124, "step": 14867 }, { "epoch": 2.32, "grad_norm": 17.404641410115246, "learning_rate": 2.55866737801426e-06, "loss": 0.4448, "step": 14868 }, { "epoch": 2.32, "grad_norm": 20.988768344908845, "learning_rate": 2.557540916696185e-06, "loss": 0.4244, "step": 14869 }, { "epoch": 2.32, "grad_norm": 19.524144897910464, "learning_rate": 2.5564146670372904e-06, "loss": 0.5063, "step": 14870 }, { "epoch": 2.32, "grad_norm": 20.849123852546313, "learning_rate": 2.5552886290696075e-06, "loss": 0.491, "step": 14871 }, { "epoch": 2.32, "grad_norm": 20.02197733938225, "learning_rate": 2.554162802825165e-06, "loss": 0.5005, "step": 14872 }, { "epoch": 2.32, "grad_norm": 28.52587704033205, "learning_rate": 2.553037188335973e-06, "loss": 0.5299, "step": 14873 }, { "epoch": 2.32, "grad_norm": 30.571850955114936, "learning_rate": 2.5519117856340503e-06, "loss": 0.444, "step": 14874 }, { "epoch": 2.32, "grad_norm": 31.508207762774102, "learning_rate": 2.5507865947513966e-06, "loss": 0.5001, "step": 14875 }, { "epoch": 2.32, "grad_norm": 26.745713434179905, "learning_rate": 2.549661615720017e-06, "loss": 0.4894, "step": 14876 }, { "epoch": 2.32, "grad_norm": 20.509532882464033, "learning_rate": 2.5485368485719e-06, "loss": 0.4895, "step": 14877 }, { "epoch": 2.32, "grad_norm": 27.204400296694757, "learning_rate": 2.5474122933390355e-06, "loss": 0.4808, "step": 14878 }, { "epoch": 2.32, "grad_norm": 24.238772374540634, "learning_rate": 2.54628795005341e-06, "loss": 0.4826, "step": 14879 }, { "epoch": 2.32, "grad_norm": 19.43901433442772, "learning_rate": 2.545163818746994e-06, "loss": 0.4471, "step": 14880 }, { "epoch": 2.32, "grad_norm": 21.967828119451124, "learning_rate": 2.5440398994517546e-06, "loss": 0.4474, "step": 14881 }, { "epoch": 2.32, "grad_norm": 23.352975465021924, "learning_rate": 2.54291619219966e-06, "loss": 0.4161, "step": 14882 }, { "epoch": 2.32, "grad_norm": 20.75014815507415, "learning_rate": 2.541792697022666e-06, "loss": 0.508, "step": 14883 }, { "epoch": 2.32, "grad_norm": 32.89221903569449, "learning_rate": 2.540669413952728e-06, "loss": 0.5474, "step": 14884 }, { "epoch": 2.33, "grad_norm": 21.21871007766676, "learning_rate": 2.5395463430217883e-06, "loss": 0.4734, "step": 14885 }, { "epoch": 2.33, "grad_norm": 15.473334047437193, "learning_rate": 2.538423484261783e-06, "loss": 0.4387, "step": 14886 }, { "epoch": 2.33, "grad_norm": 23.590986072886295, "learning_rate": 2.537300837704648e-06, "loss": 0.4992, "step": 14887 }, { "epoch": 2.33, "grad_norm": 15.750634755861585, "learning_rate": 2.536178403382317e-06, "loss": 0.4633, "step": 14888 }, { "epoch": 2.33, "grad_norm": 16.826223110092187, "learning_rate": 2.535056181326704e-06, "loss": 0.4211, "step": 14889 }, { "epoch": 2.33, "grad_norm": 25.123645260806665, "learning_rate": 2.533934171569724e-06, "loss": 0.4863, "step": 14890 }, { "epoch": 2.33, "grad_norm": 19.663848545394245, "learning_rate": 2.5328123741432887e-06, "loss": 0.4774, "step": 14891 }, { "epoch": 2.33, "grad_norm": 17.10033263763426, "learning_rate": 2.5316907890793007e-06, "loss": 0.4488, "step": 14892 }, { "epoch": 2.33, "grad_norm": 24.353011283085245, "learning_rate": 2.530569416409665e-06, "loss": 0.43, "step": 14893 }, { "epoch": 2.33, "grad_norm": 19.914013441645707, "learning_rate": 2.529448256166259e-06, "loss": 0.5523, "step": 14894 }, { "epoch": 2.33, "grad_norm": 22.388623450954917, "learning_rate": 2.5283273083809746e-06, "loss": 0.4799, "step": 14895 }, { "epoch": 2.33, "grad_norm": 28.360423670314166, "learning_rate": 2.52720657308569e-06, "loss": 0.5104, "step": 14896 }, { "epoch": 2.33, "grad_norm": 21.648751782449107, "learning_rate": 2.526086050312283e-06, "loss": 0.4867, "step": 14897 }, { "epoch": 2.33, "grad_norm": 23.202380727461488, "learning_rate": 2.5249657400926165e-06, "loss": 0.4579, "step": 14898 }, { "epoch": 2.33, "grad_norm": 27.208175456570928, "learning_rate": 2.5238456424585477e-06, "loss": 0.5395, "step": 14899 }, { "epoch": 2.33, "grad_norm": 18.22642357450763, "learning_rate": 2.522725757441936e-06, "loss": 0.523, "step": 14900 }, { "epoch": 2.33, "grad_norm": 23.96946688690302, "learning_rate": 2.5216060850746326e-06, "loss": 0.4238, "step": 14901 }, { "epoch": 2.33, "grad_norm": 24.70682862873449, "learning_rate": 2.520486625388475e-06, "loss": 0.5179, "step": 14902 }, { "epoch": 2.33, "grad_norm": 21.26506501552932, "learning_rate": 2.519367378415305e-06, "loss": 0.4175, "step": 14903 }, { "epoch": 2.33, "grad_norm": 28.265522268687448, "learning_rate": 2.5182483441869477e-06, "loss": 0.4958, "step": 14904 }, { "epoch": 2.33, "grad_norm": 16.64060572681296, "learning_rate": 2.5171295227352332e-06, "loss": 0.4378, "step": 14905 }, { "epoch": 2.33, "grad_norm": 16.50999705361037, "learning_rate": 2.516010914091975e-06, "loss": 0.5289, "step": 14906 }, { "epoch": 2.33, "grad_norm": 27.060710645198675, "learning_rate": 2.514892518288988e-06, "loss": 0.5828, "step": 14907 }, { "epoch": 2.33, "grad_norm": 18.108628868188458, "learning_rate": 2.5137743353580833e-06, "loss": 0.4716, "step": 14908 }, { "epoch": 2.33, "grad_norm": 35.21383427481106, "learning_rate": 2.512656365331053e-06, "loss": 0.5287, "step": 14909 }, { "epoch": 2.33, "grad_norm": 22.300545590803143, "learning_rate": 2.5115386082396996e-06, "loss": 0.4895, "step": 14910 }, { "epoch": 2.33, "grad_norm": 18.891203170996736, "learning_rate": 2.5104210641158045e-06, "loss": 0.4978, "step": 14911 }, { "epoch": 2.33, "grad_norm": 18.390122345186295, "learning_rate": 2.509303732991153e-06, "loss": 0.4928, "step": 14912 }, { "epoch": 2.33, "grad_norm": 20.891904415135357, "learning_rate": 2.508186614897524e-06, "loss": 0.5075, "step": 14913 }, { "epoch": 2.33, "grad_norm": 24.8286407349919, "learning_rate": 2.507069709866685e-06, "loss": 0.5637, "step": 14914 }, { "epoch": 2.33, "grad_norm": 4.558267714111223, "learning_rate": 2.5059530179303983e-06, "loss": 0.509, "step": 14915 }, { "epoch": 2.33, "grad_norm": 21.68190799579549, "learning_rate": 2.5048365391204233e-06, "loss": 0.4622, "step": 14916 }, { "epoch": 2.33, "grad_norm": 14.198108666433214, "learning_rate": 2.5037202734685117e-06, "loss": 0.4392, "step": 14917 }, { "epoch": 2.33, "grad_norm": 44.612409332816014, "learning_rate": 2.502604221006417e-06, "loss": 0.5159, "step": 14918 }, { "epoch": 2.33, "grad_norm": 27.79677335389258, "learning_rate": 2.5014883817658662e-06, "loss": 0.4584, "step": 14919 }, { "epoch": 2.33, "grad_norm": 21.223816181167074, "learning_rate": 2.5003727557785997e-06, "loss": 0.4899, "step": 14920 }, { "epoch": 2.33, "grad_norm": 27.338412216169772, "learning_rate": 2.4992573430763447e-06, "loss": 0.4704, "step": 14921 }, { "epoch": 2.33, "grad_norm": 19.266150324296394, "learning_rate": 2.498142143690826e-06, "loss": 0.4251, "step": 14922 }, { "epoch": 2.33, "grad_norm": 26.36628844981907, "learning_rate": 2.4970271576537563e-06, "loss": 0.4463, "step": 14923 }, { "epoch": 2.33, "grad_norm": 19.369843771302754, "learning_rate": 2.4959123849968427e-06, "loss": 0.4128, "step": 14924 }, { "epoch": 2.33, "grad_norm": 15.515989629680922, "learning_rate": 2.4947978257517903e-06, "loss": 0.4476, "step": 14925 }, { "epoch": 2.33, "grad_norm": 15.347493744395427, "learning_rate": 2.4936834799503005e-06, "loss": 0.4164, "step": 14926 }, { "epoch": 2.33, "grad_norm": 17.77561540779018, "learning_rate": 2.4925693476240587e-06, "loss": 0.4677, "step": 14927 }, { "epoch": 2.33, "grad_norm": 20.057707899789197, "learning_rate": 2.4914554288047553e-06, "loss": 0.5281, "step": 14928 }, { "epoch": 2.33, "grad_norm": 19.505151410565166, "learning_rate": 2.4903417235240647e-06, "loss": 0.4587, "step": 14929 }, { "epoch": 2.33, "grad_norm": 22.570211867494013, "learning_rate": 2.489228231813665e-06, "loss": 0.4689, "step": 14930 }, { "epoch": 2.33, "grad_norm": 26.963006089739416, "learning_rate": 2.4881149537052183e-06, "loss": 0.4297, "step": 14931 }, { "epoch": 2.33, "grad_norm": 21.576649079476116, "learning_rate": 2.4870018892303883e-06, "loss": 0.4819, "step": 14932 }, { "epoch": 2.33, "grad_norm": 12.821338733951984, "learning_rate": 2.485889038420831e-06, "loss": 0.4319, "step": 14933 }, { "epoch": 2.33, "grad_norm": 21.333859014522588, "learning_rate": 2.4847764013081923e-06, "loss": 0.4518, "step": 14934 }, { "epoch": 2.33, "grad_norm": 17.750719861759315, "learning_rate": 2.483663977924119e-06, "loss": 0.5296, "step": 14935 }, { "epoch": 2.33, "grad_norm": 19.035616928027732, "learning_rate": 2.482551768300242e-06, "loss": 0.4475, "step": 14936 }, { "epoch": 2.33, "grad_norm": 18.39575879709027, "learning_rate": 2.4814397724681995e-06, "loss": 0.4389, "step": 14937 }, { "epoch": 2.33, "grad_norm": 26.18367423355775, "learning_rate": 2.4803279904596067e-06, "loss": 0.5237, "step": 14938 }, { "epoch": 2.33, "grad_norm": 21.494608101793883, "learning_rate": 2.479216422306091e-06, "loss": 0.3787, "step": 14939 }, { "epoch": 2.33, "grad_norm": 18.0936613904269, "learning_rate": 2.4781050680392572e-06, "loss": 0.4652, "step": 14940 }, { "epoch": 2.33, "grad_norm": 27.019365501805396, "learning_rate": 2.4769939276907163e-06, "loss": 0.4932, "step": 14941 }, { "epoch": 2.33, "grad_norm": 24.607130837220044, "learning_rate": 2.4758830012920688e-06, "loss": 0.4356, "step": 14942 }, { "epoch": 2.33, "grad_norm": 32.46969584848756, "learning_rate": 2.4747722888749082e-06, "loss": 0.5453, "step": 14943 }, { "epoch": 2.33, "grad_norm": 24.45369661778227, "learning_rate": 2.4736617904708184e-06, "loss": 0.4634, "step": 14944 }, { "epoch": 2.33, "grad_norm": 21.218689123468057, "learning_rate": 2.4725515061113835e-06, "loss": 0.5306, "step": 14945 }, { "epoch": 2.33, "grad_norm": 27.013655443471304, "learning_rate": 2.471441435828181e-06, "loss": 0.4706, "step": 14946 }, { "epoch": 2.33, "grad_norm": 20.41876768538191, "learning_rate": 2.470331579652784e-06, "loss": 0.5321, "step": 14947 }, { "epoch": 2.33, "grad_norm": 24.032984111707147, "learning_rate": 2.4692219376167504e-06, "loss": 0.417, "step": 14948 }, { "epoch": 2.34, "grad_norm": 15.417769252703222, "learning_rate": 2.4681125097516376e-06, "loss": 0.4101, "step": 14949 }, { "epoch": 2.34, "grad_norm": 16.586297617976975, "learning_rate": 2.467003296088999e-06, "loss": 0.3846, "step": 14950 }, { "epoch": 2.34, "grad_norm": 15.620353305817154, "learning_rate": 2.465894296660383e-06, "loss": 0.4324, "step": 14951 }, { "epoch": 2.34, "grad_norm": 18.054366319771617, "learning_rate": 2.464785511497325e-06, "loss": 0.5084, "step": 14952 }, { "epoch": 2.34, "grad_norm": 28.02748522360581, "learning_rate": 2.4636769406313576e-06, "loss": 0.4487, "step": 14953 }, { "epoch": 2.34, "grad_norm": 15.005331870006021, "learning_rate": 2.4625685840940083e-06, "loss": 0.419, "step": 14954 }, { "epoch": 2.34, "grad_norm": 28.732713900766228, "learning_rate": 2.4614604419167997e-06, "loss": 0.4639, "step": 14955 }, { "epoch": 2.34, "grad_norm": 19.951818094790486, "learning_rate": 2.460352514131249e-06, "loss": 0.4381, "step": 14956 }, { "epoch": 2.34, "grad_norm": 19.367471802114718, "learning_rate": 2.459244800768862e-06, "loss": 0.4936, "step": 14957 }, { "epoch": 2.34, "grad_norm": 20.924971382003047, "learning_rate": 2.458137301861139e-06, "loss": 0.4543, "step": 14958 }, { "epoch": 2.34, "grad_norm": 25.393652212000195, "learning_rate": 2.457030017439579e-06, "loss": 0.4744, "step": 14959 }, { "epoch": 2.34, "grad_norm": 31.35668543375843, "learning_rate": 2.455922947535675e-06, "loss": 0.4553, "step": 14960 }, { "epoch": 2.34, "grad_norm": 21.097606762702817, "learning_rate": 2.4548160921809073e-06, "loss": 0.5306, "step": 14961 }, { "epoch": 2.34, "grad_norm": 19.973161160730598, "learning_rate": 2.4537094514067596e-06, "loss": 0.4066, "step": 14962 }, { "epoch": 2.34, "grad_norm": 17.01353757137663, "learning_rate": 2.4526030252446963e-06, "loss": 0.4172, "step": 14963 }, { "epoch": 2.34, "grad_norm": 22.944544113328586, "learning_rate": 2.4514968137261918e-06, "loss": 0.436, "step": 14964 }, { "epoch": 2.34, "grad_norm": 29.676451297500044, "learning_rate": 2.4503908168826985e-06, "loss": 0.4781, "step": 14965 }, { "epoch": 2.34, "grad_norm": 22.34743399293661, "learning_rate": 2.449285034745674e-06, "loss": 0.4258, "step": 14966 }, { "epoch": 2.34, "grad_norm": 24.51332246192332, "learning_rate": 2.4481794673465697e-06, "loss": 0.4292, "step": 14967 }, { "epoch": 2.34, "grad_norm": 22.55031654431513, "learning_rate": 2.4470741147168242e-06, "loss": 0.423, "step": 14968 }, { "epoch": 2.34, "grad_norm": 24.123962968071744, "learning_rate": 2.445968976887868e-06, "loss": 0.4538, "step": 14969 }, { "epoch": 2.34, "grad_norm": 25.20522426428361, "learning_rate": 2.4448640538911363e-06, "loss": 0.4326, "step": 14970 }, { "epoch": 2.34, "grad_norm": 21.067320327973807, "learning_rate": 2.443759345758051e-06, "loss": 0.4336, "step": 14971 }, { "epoch": 2.34, "grad_norm": 19.172972162745936, "learning_rate": 2.442654852520032e-06, "loss": 0.4923, "step": 14972 }, { "epoch": 2.34, "grad_norm": 23.07625028037539, "learning_rate": 2.441550574208489e-06, "loss": 0.5377, "step": 14973 }, { "epoch": 2.34, "grad_norm": 22.12302248948293, "learning_rate": 2.4404465108548226e-06, "loss": 0.534, "step": 14974 }, { "epoch": 2.34, "grad_norm": 36.095100328031485, "learning_rate": 2.4393426624904347e-06, "loss": 0.5104, "step": 14975 }, { "epoch": 2.34, "grad_norm": 19.292043007967393, "learning_rate": 2.438239029146723e-06, "loss": 0.4153, "step": 14976 }, { "epoch": 2.34, "grad_norm": 31.048400311565103, "learning_rate": 2.437135610855068e-06, "loss": 0.4563, "step": 14977 }, { "epoch": 2.34, "grad_norm": 16.961414223713774, "learning_rate": 2.436032407646851e-06, "loss": 0.4316, "step": 14978 }, { "epoch": 2.34, "grad_norm": 17.57240716296963, "learning_rate": 2.434929419553447e-06, "loss": 0.4457, "step": 14979 }, { "epoch": 2.34, "grad_norm": 23.265405732012503, "learning_rate": 2.4338266466062242e-06, "loss": 0.4394, "step": 14980 }, { "epoch": 2.34, "grad_norm": 27.10760443124565, "learning_rate": 2.432724088836549e-06, "loss": 0.4543, "step": 14981 }, { "epoch": 2.34, "grad_norm": 34.44885512425127, "learning_rate": 2.431621746275773e-06, "loss": 0.5066, "step": 14982 }, { "epoch": 2.34, "grad_norm": 27.4942005230256, "learning_rate": 2.430519618955245e-06, "loss": 0.4659, "step": 14983 }, { "epoch": 2.34, "grad_norm": 23.874659131397916, "learning_rate": 2.429417706906311e-06, "loss": 0.4105, "step": 14984 }, { "epoch": 2.34, "grad_norm": 16.379084559060704, "learning_rate": 2.428316010160312e-06, "loss": 0.5344, "step": 14985 }, { "epoch": 2.34, "grad_norm": 14.240202050230563, "learning_rate": 2.427214528748576e-06, "loss": 0.4811, "step": 14986 }, { "epoch": 2.34, "grad_norm": 31.282736381387473, "learning_rate": 2.4261132627024254e-06, "loss": 0.497, "step": 14987 }, { "epoch": 2.34, "grad_norm": 26.072178603805863, "learning_rate": 2.425012212053184e-06, "loss": 0.4797, "step": 14988 }, { "epoch": 2.34, "grad_norm": 17.55136766776394, "learning_rate": 2.4239113768321666e-06, "loss": 0.4536, "step": 14989 }, { "epoch": 2.34, "grad_norm": 25.996251690024376, "learning_rate": 2.422810757070675e-06, "loss": 0.4558, "step": 14990 }, { "epoch": 2.34, "grad_norm": 15.55493883545863, "learning_rate": 2.4217103528000153e-06, "loss": 0.4331, "step": 14991 }, { "epoch": 2.34, "grad_norm": 21.073009212313934, "learning_rate": 2.4206101640514778e-06, "loss": 0.4405, "step": 14992 }, { "epoch": 2.34, "grad_norm": 23.764856347388857, "learning_rate": 2.4195101908563556e-06, "loss": 0.5174, "step": 14993 }, { "epoch": 2.34, "grad_norm": 15.598587684706972, "learning_rate": 2.4184104332459267e-06, "loss": 0.4127, "step": 14994 }, { "epoch": 2.34, "grad_norm": 16.26352661808254, "learning_rate": 2.4173108912514696e-06, "loss": 0.4606, "step": 14995 }, { "epoch": 2.34, "grad_norm": 27.17008360660415, "learning_rate": 2.4162115649042584e-06, "loss": 0.516, "step": 14996 }, { "epoch": 2.34, "grad_norm": 23.425129906671753, "learning_rate": 2.415112454235551e-06, "loss": 0.4293, "step": 14997 }, { "epoch": 2.34, "grad_norm": 24.44199735875639, "learning_rate": 2.41401355927661e-06, "loss": 0.4917, "step": 14998 }, { "epoch": 2.34, "grad_norm": 19.851185743651754, "learning_rate": 2.4129148800586846e-06, "loss": 0.4333, "step": 14999 }, { "epoch": 2.34, "grad_norm": 22.86725538508925, "learning_rate": 2.411816416613021e-06, "loss": 0.4597, "step": 15000 }, { "epoch": 2.34, "grad_norm": 40.98778962989567, "learning_rate": 2.4107181689708614e-06, "loss": 0.5131, "step": 15001 }, { "epoch": 2.34, "grad_norm": 11.802075302152716, "learning_rate": 2.4096201371634387e-06, "loss": 0.4535, "step": 15002 }, { "epoch": 2.34, "grad_norm": 17.372028739991396, "learning_rate": 2.4085223212219767e-06, "loss": 0.4252, "step": 15003 }, { "epoch": 2.34, "grad_norm": 20.576019779724053, "learning_rate": 2.407424721177698e-06, "loss": 0.4688, "step": 15004 }, { "epoch": 2.34, "grad_norm": 12.118607587257904, "learning_rate": 2.406327337061818e-06, "loss": 0.4039, "step": 15005 }, { "epoch": 2.34, "grad_norm": 16.77158144537063, "learning_rate": 2.4052301689055536e-06, "loss": 0.5015, "step": 15006 }, { "epoch": 2.34, "grad_norm": 24.119996615644077, "learning_rate": 2.4041332167400944e-06, "loss": 0.5551, "step": 15007 }, { "epoch": 2.34, "grad_norm": 25.989182301652665, "learning_rate": 2.4030364805966423e-06, "loss": 0.4872, "step": 15008 }, { "epoch": 2.34, "grad_norm": 20.823894124675242, "learning_rate": 2.401939960506389e-06, "loss": 0.4754, "step": 15009 }, { "epoch": 2.34, "grad_norm": 14.71575180119783, "learning_rate": 2.400843656500521e-06, "loss": 0.4743, "step": 15010 }, { "epoch": 2.34, "grad_norm": 16.538955506135007, "learning_rate": 2.3997475686102147e-06, "loss": 0.4732, "step": 15011 }, { "epoch": 2.34, "grad_norm": 23.860486072159112, "learning_rate": 2.3986516968666385e-06, "loss": 0.4324, "step": 15012 }, { "epoch": 2.35, "grad_norm": 18.621210674465352, "learning_rate": 2.397556041300961e-06, "loss": 0.4393, "step": 15013 }, { "epoch": 2.35, "grad_norm": 16.63412330997982, "learning_rate": 2.396460601944346e-06, "loss": 0.414, "step": 15014 }, { "epoch": 2.35, "grad_norm": 25.2569740656539, "learning_rate": 2.3953653788279395e-06, "loss": 0.5455, "step": 15015 }, { "epoch": 2.35, "grad_norm": 27.54138978255306, "learning_rate": 2.3942703719828965e-06, "loss": 0.531, "step": 15016 }, { "epoch": 2.35, "grad_norm": 22.553068131286206, "learning_rate": 2.393175581440352e-06, "loss": 0.4352, "step": 15017 }, { "epoch": 2.35, "grad_norm": 24.916161770207733, "learning_rate": 2.3920810072314473e-06, "loss": 0.5124, "step": 15018 }, { "epoch": 2.35, "grad_norm": 17.102864184970702, "learning_rate": 2.3909866493873048e-06, "loss": 0.4557, "step": 15019 }, { "epoch": 2.35, "grad_norm": 12.91727923458385, "learning_rate": 2.389892507939051e-06, "loss": 0.4645, "step": 15020 }, { "epoch": 2.35, "grad_norm": 16.530191500294986, "learning_rate": 2.388798582917806e-06, "loss": 0.4213, "step": 15021 }, { "epoch": 2.35, "grad_norm": 17.047844349602368, "learning_rate": 2.387704874354674e-06, "loss": 0.4513, "step": 15022 }, { "epoch": 2.35, "grad_norm": 23.088735549655407, "learning_rate": 2.3866113822807645e-06, "loss": 0.508, "step": 15023 }, { "epoch": 2.35, "grad_norm": 22.924472581369244, "learning_rate": 2.3855181067271703e-06, "loss": 0.4427, "step": 15024 }, { "epoch": 2.35, "grad_norm": 16.682589026198844, "learning_rate": 2.3844250477249877e-06, "loss": 0.4234, "step": 15025 }, { "epoch": 2.35, "grad_norm": 30.908038435221503, "learning_rate": 2.3833322053053045e-06, "loss": 0.4948, "step": 15026 }, { "epoch": 2.35, "grad_norm": 18.868917020265116, "learning_rate": 2.382239579499198e-06, "loss": 0.4493, "step": 15027 }, { "epoch": 2.35, "grad_norm": 25.83841338404869, "learning_rate": 2.3811471703377376e-06, "loss": 0.4761, "step": 15028 }, { "epoch": 2.35, "grad_norm": 22.192612713754983, "learning_rate": 2.3800549778519956e-06, "loss": 0.4263, "step": 15029 }, { "epoch": 2.35, "grad_norm": 18.85421361513422, "learning_rate": 2.3789630020730327e-06, "loss": 0.5324, "step": 15030 }, { "epoch": 2.35, "grad_norm": 18.814362989740594, "learning_rate": 2.3778712430319095e-06, "loss": 0.419, "step": 15031 }, { "epoch": 2.35, "grad_norm": 24.775991052577602, "learning_rate": 2.376779700759664e-06, "loss": 0.4638, "step": 15032 }, { "epoch": 2.35, "grad_norm": 26.492217591205897, "learning_rate": 2.3756883752873426e-06, "loss": 0.4974, "step": 15033 }, { "epoch": 2.35, "grad_norm": 17.598084423178136, "learning_rate": 2.3745972666459847e-06, "loss": 0.4467, "step": 15034 }, { "epoch": 2.35, "grad_norm": 19.346625671265098, "learning_rate": 2.373506374866623e-06, "loss": 0.4888, "step": 15035 }, { "epoch": 2.35, "grad_norm": 17.6641484038716, "learning_rate": 2.3724156999802795e-06, "loss": 0.4539, "step": 15036 }, { "epoch": 2.35, "grad_norm": 16.10505575538282, "learning_rate": 2.3713252420179668e-06, "loss": 0.4668, "step": 15037 }, { "epoch": 2.35, "grad_norm": 14.419118771799873, "learning_rate": 2.3702350010107023e-06, "loss": 0.4351, "step": 15038 }, { "epoch": 2.35, "grad_norm": 17.85497914012074, "learning_rate": 2.3691449769894946e-06, "loss": 0.4623, "step": 15039 }, { "epoch": 2.35, "grad_norm": 19.339209468551555, "learning_rate": 2.368055169985339e-06, "loss": 0.41, "step": 15040 }, { "epoch": 2.35, "grad_norm": 18.87039009224215, "learning_rate": 2.366965580029227e-06, "loss": 0.4823, "step": 15041 }, { "epoch": 2.35, "grad_norm": 22.988464601056158, "learning_rate": 2.365876207152149e-06, "loss": 0.4696, "step": 15042 }, { "epoch": 2.35, "grad_norm": 14.337716538432867, "learning_rate": 2.3647870513850857e-06, "loss": 0.4161, "step": 15043 }, { "epoch": 2.35, "grad_norm": 20.433713748895297, "learning_rate": 2.3636981127590155e-06, "loss": 0.4285, "step": 15044 }, { "epoch": 2.35, "grad_norm": 15.199931924794557, "learning_rate": 2.362609391304903e-06, "loss": 0.419, "step": 15045 }, { "epoch": 2.35, "grad_norm": 29.65361755865329, "learning_rate": 2.361520887053709e-06, "loss": 0.4675, "step": 15046 }, { "epoch": 2.35, "grad_norm": 33.776858864836804, "learning_rate": 2.360432600036393e-06, "loss": 0.4308, "step": 15047 }, { "epoch": 2.35, "grad_norm": 32.109737641080365, "learning_rate": 2.359344530283908e-06, "loss": 0.6344, "step": 15048 }, { "epoch": 2.35, "grad_norm": 19.87092471627326, "learning_rate": 2.358256677827191e-06, "loss": 0.3783, "step": 15049 }, { "epoch": 2.35, "grad_norm": 25.31178849185736, "learning_rate": 2.357169042697187e-06, "loss": 0.5111, "step": 15050 }, { "epoch": 2.35, "grad_norm": 21.185630606155875, "learning_rate": 2.356081624924822e-06, "loss": 0.423, "step": 15051 }, { "epoch": 2.35, "grad_norm": 15.59275958254778, "learning_rate": 2.354994424541027e-06, "loss": 0.4502, "step": 15052 }, { "epoch": 2.35, "grad_norm": 26.230387707866985, "learning_rate": 2.353907441576715e-06, "loss": 0.4714, "step": 15053 }, { "epoch": 2.35, "grad_norm": 23.36885248844274, "learning_rate": 2.3528206760628026e-06, "loss": 0.5002, "step": 15054 }, { "epoch": 2.35, "grad_norm": 14.731884886518342, "learning_rate": 2.3517341280301997e-06, "loss": 0.5039, "step": 15055 }, { "epoch": 2.35, "grad_norm": 20.666482198200214, "learning_rate": 2.350647797509804e-06, "loss": 0.5093, "step": 15056 }, { "epoch": 2.35, "grad_norm": 22.761040849150536, "learning_rate": 2.3495616845325074e-06, "loss": 0.4546, "step": 15057 }, { "epoch": 2.35, "grad_norm": 15.422941218373836, "learning_rate": 2.348475789129202e-06, "loss": 0.4277, "step": 15058 }, { "epoch": 2.35, "grad_norm": 20.727040200133143, "learning_rate": 2.3473901113307685e-06, "loss": 0.4747, "step": 15059 }, { "epoch": 2.35, "grad_norm": 18.24599733125058, "learning_rate": 2.3463046511680864e-06, "loss": 0.4769, "step": 15060 }, { "epoch": 2.35, "grad_norm": 20.143046850149158, "learning_rate": 2.3452194086720225e-06, "loss": 0.462, "step": 15061 }, { "epoch": 2.35, "grad_norm": 17.73811209328546, "learning_rate": 2.3441343838734375e-06, "loss": 0.4421, "step": 15062 }, { "epoch": 2.35, "grad_norm": 19.13671426798326, "learning_rate": 2.3430495768031923e-06, "loss": 0.4325, "step": 15063 }, { "epoch": 2.35, "grad_norm": 19.814185855241742, "learning_rate": 2.3419649874921413e-06, "loss": 0.4137, "step": 15064 }, { "epoch": 2.35, "grad_norm": 15.830912419901717, "learning_rate": 2.3408806159711262e-06, "loss": 0.3923, "step": 15065 }, { "epoch": 2.35, "grad_norm": 27.002734095504596, "learning_rate": 2.3397964622709825e-06, "loss": 0.5272, "step": 15066 }, { "epoch": 2.35, "grad_norm": 15.237510153146669, "learning_rate": 2.3387125264225475e-06, "loss": 0.4649, "step": 15067 }, { "epoch": 2.35, "grad_norm": 21.016762106384046, "learning_rate": 2.3376288084566445e-06, "loss": 0.4739, "step": 15068 }, { "epoch": 2.35, "grad_norm": 24.655839518356228, "learning_rate": 2.3365453084041e-06, "loss": 0.5133, "step": 15069 }, { "epoch": 2.35, "grad_norm": 23.933735255113675, "learning_rate": 2.3354620262957238e-06, "loss": 0.468, "step": 15070 }, { "epoch": 2.35, "grad_norm": 25.10690323699414, "learning_rate": 2.3343789621623213e-06, "loss": 0.4302, "step": 15071 }, { "epoch": 2.35, "grad_norm": 27.086627692293746, "learning_rate": 2.3332961160346966e-06, "loss": 0.4648, "step": 15072 }, { "epoch": 2.35, "grad_norm": 25.18997946904562, "learning_rate": 2.3322134879436487e-06, "loss": 0.4564, "step": 15073 }, { "epoch": 2.35, "grad_norm": 23.97426075171063, "learning_rate": 2.3311310779199603e-06, "loss": 0.4732, "step": 15074 }, { "epoch": 2.35, "grad_norm": 18.09005984624344, "learning_rate": 2.3300488859944217e-06, "loss": 0.4992, "step": 15075 }, { "epoch": 2.35, "grad_norm": 32.03438821631756, "learning_rate": 2.328966912197802e-06, "loss": 0.5262, "step": 15076 }, { "epoch": 2.36, "grad_norm": 21.092950243447078, "learning_rate": 2.3278851565608782e-06, "loss": 0.4489, "step": 15077 }, { "epoch": 2.36, "grad_norm": 18.48221056060619, "learning_rate": 2.3268036191144117e-06, "loss": 0.4537, "step": 15078 }, { "epoch": 2.36, "grad_norm": 22.29270156384741, "learning_rate": 2.3257222998891603e-06, "loss": 0.4318, "step": 15079 }, { "epoch": 2.36, "grad_norm": 28.135608742821066, "learning_rate": 2.32464119891588e-06, "loss": 0.5001, "step": 15080 }, { "epoch": 2.36, "grad_norm": 14.235043348267983, "learning_rate": 2.323560316225314e-06, "loss": 0.4727, "step": 15081 }, { "epoch": 2.36, "grad_norm": 16.84446794752678, "learning_rate": 2.3224796518481995e-06, "loss": 0.4851, "step": 15082 }, { "epoch": 2.36, "grad_norm": 20.97587809733731, "learning_rate": 2.3213992058152733e-06, "loss": 0.5008, "step": 15083 }, { "epoch": 2.36, "grad_norm": 26.42728099584931, "learning_rate": 2.320318978157263e-06, "loss": 0.514, "step": 15084 }, { "epoch": 2.36, "grad_norm": 34.64607227686141, "learning_rate": 2.319238968904888e-06, "loss": 0.4717, "step": 15085 }, { "epoch": 2.36, "grad_norm": 22.462435583824682, "learning_rate": 2.318159178088865e-06, "loss": 0.4446, "step": 15086 }, { "epoch": 2.36, "grad_norm": 27.888941684649076, "learning_rate": 2.317079605739898e-06, "loss": 0.4867, "step": 15087 }, { "epoch": 2.36, "grad_norm": 20.10364849224933, "learning_rate": 2.3160002518886927e-06, "loss": 0.498, "step": 15088 }, { "epoch": 2.36, "grad_norm": 23.953757768548204, "learning_rate": 2.314921116565948e-06, "loss": 0.4684, "step": 15089 }, { "epoch": 2.36, "grad_norm": 23.420475021522208, "learning_rate": 2.313842199802352e-06, "loss": 0.4878, "step": 15090 }, { "epoch": 2.36, "grad_norm": 24.036588174673145, "learning_rate": 2.312763501628584e-06, "loss": 0.5103, "step": 15091 }, { "epoch": 2.36, "grad_norm": 27.239336427453367, "learning_rate": 2.3116850220753253e-06, "loss": 0.4904, "step": 15092 }, { "epoch": 2.36, "grad_norm": 24.92729935643449, "learning_rate": 2.3106067611732473e-06, "loss": 0.4889, "step": 15093 }, { "epoch": 2.36, "grad_norm": 22.614072179281564, "learning_rate": 2.3095287189530203e-06, "loss": 0.5348, "step": 15094 }, { "epoch": 2.36, "grad_norm": 26.012088577086438, "learning_rate": 2.3084508954452923e-06, "loss": 0.4741, "step": 15095 }, { "epoch": 2.36, "grad_norm": 23.222409866091148, "learning_rate": 2.307373290680721e-06, "loss": 0.424, "step": 15096 }, { "epoch": 2.36, "grad_norm": 36.56650173711479, "learning_rate": 2.3062959046899535e-06, "loss": 0.5467, "step": 15097 }, { "epoch": 2.36, "grad_norm": 16.061730655844066, "learning_rate": 2.3052187375036327e-06, "loss": 0.4801, "step": 15098 }, { "epoch": 2.36, "grad_norm": 27.289106672585966, "learning_rate": 2.3041417891523897e-06, "loss": 0.4666, "step": 15099 }, { "epoch": 2.36, "grad_norm": 22.583160254711412, "learning_rate": 2.303065059666849e-06, "loss": 0.4162, "step": 15100 }, { "epoch": 2.36, "grad_norm": 20.483103717779745, "learning_rate": 2.301988549077636e-06, "loss": 0.4439, "step": 15101 }, { "epoch": 2.36, "grad_norm": 22.947887461905324, "learning_rate": 2.3009122574153673e-06, "loss": 0.4959, "step": 15102 }, { "epoch": 2.36, "grad_norm": 24.980107709409623, "learning_rate": 2.2998361847106487e-06, "loss": 0.45, "step": 15103 }, { "epoch": 2.36, "grad_norm": 16.862130669069074, "learning_rate": 2.2987603309940866e-06, "loss": 0.4614, "step": 15104 }, { "epoch": 2.36, "grad_norm": 17.08831933421522, "learning_rate": 2.297684696296273e-06, "loss": 0.4327, "step": 15105 }, { "epoch": 2.36, "grad_norm": 15.927605998576137, "learning_rate": 2.2966092806478e-06, "loss": 0.4253, "step": 15106 }, { "epoch": 2.36, "grad_norm": 19.731646935165134, "learning_rate": 2.2955340840792563e-06, "loss": 0.4872, "step": 15107 }, { "epoch": 2.36, "grad_norm": 23.502861993208516, "learning_rate": 2.294459106621214e-06, "loss": 0.505, "step": 15108 }, { "epoch": 2.36, "grad_norm": 27.69800669017182, "learning_rate": 2.2933843483042495e-06, "loss": 0.5039, "step": 15109 }, { "epoch": 2.36, "grad_norm": 19.162776566417605, "learning_rate": 2.2923098091589224e-06, "loss": 0.4361, "step": 15110 }, { "epoch": 2.36, "grad_norm": 23.103642994710736, "learning_rate": 2.2912354892157995e-06, "loss": 0.4665, "step": 15111 }, { "epoch": 2.36, "grad_norm": 24.710251316525817, "learning_rate": 2.2901613885054252e-06, "loss": 0.4314, "step": 15112 }, { "epoch": 2.36, "grad_norm": 18.402614824715194, "learning_rate": 2.289087507058353e-06, "loss": 0.4764, "step": 15113 }, { "epoch": 2.36, "grad_norm": 21.036047549123868, "learning_rate": 2.2880138449051227e-06, "loss": 0.4791, "step": 15114 }, { "epoch": 2.36, "grad_norm": 15.06097541706953, "learning_rate": 2.2869404020762676e-06, "loss": 0.4918, "step": 15115 }, { "epoch": 2.36, "grad_norm": 20.834690349172593, "learning_rate": 2.285867178602312e-06, "loss": 0.4796, "step": 15116 }, { "epoch": 2.36, "grad_norm": 22.41086027666266, "learning_rate": 2.2847941745137826e-06, "loss": 0.4671, "step": 15117 }, { "epoch": 2.36, "grad_norm": 22.05177698987236, "learning_rate": 2.2837213898411925e-06, "loss": 0.3635, "step": 15118 }, { "epoch": 2.36, "grad_norm": 19.04124619855663, "learning_rate": 2.282648824615059e-06, "loss": 0.5058, "step": 15119 }, { "epoch": 2.36, "grad_norm": 27.816393747904115, "learning_rate": 2.281576478865872e-06, "loss": 0.4795, "step": 15120 }, { "epoch": 2.36, "grad_norm": 21.70489589203088, "learning_rate": 2.2805043526241342e-06, "loss": 0.485, "step": 15121 }, { "epoch": 2.36, "grad_norm": 28.742116212043584, "learning_rate": 2.2794324459203377e-06, "loss": 0.4964, "step": 15122 }, { "epoch": 2.36, "grad_norm": 17.5325057086598, "learning_rate": 2.278360758784969e-06, "loss": 0.4472, "step": 15123 }, { "epoch": 2.36, "grad_norm": 23.80214907767873, "learning_rate": 2.277289291248502e-06, "loss": 0.4132, "step": 15124 }, { "epoch": 2.36, "grad_norm": 22.10269729138953, "learning_rate": 2.2762180433414083e-06, "loss": 0.4791, "step": 15125 }, { "epoch": 2.36, "grad_norm": 20.031731506011766, "learning_rate": 2.2751470150941558e-06, "loss": 0.5251, "step": 15126 }, { "epoch": 2.36, "grad_norm": 26.074122114823965, "learning_rate": 2.2740762065372046e-06, "loss": 0.3994, "step": 15127 }, { "epoch": 2.36, "grad_norm": 18.40802495369327, "learning_rate": 2.2730056177010075e-06, "loss": 0.4494, "step": 15128 }, { "epoch": 2.36, "grad_norm": 24.072359591101424, "learning_rate": 2.2719352486160074e-06, "loss": 0.468, "step": 15129 }, { "epoch": 2.36, "grad_norm": 9.795083219011353, "learning_rate": 2.270865099312648e-06, "loss": 0.3959, "step": 15130 }, { "epoch": 2.36, "grad_norm": 34.55590289826952, "learning_rate": 2.2697951698213628e-06, "loss": 0.4944, "step": 15131 }, { "epoch": 2.36, "grad_norm": 23.966692484312077, "learning_rate": 2.268725460172585e-06, "loss": 0.4744, "step": 15132 }, { "epoch": 2.36, "grad_norm": 17.845703329157487, "learning_rate": 2.2676559703967317e-06, "loss": 0.5224, "step": 15133 }, { "epoch": 2.36, "grad_norm": 25.79507267395847, "learning_rate": 2.266586700524217e-06, "loss": 0.4531, "step": 15134 }, { "epoch": 2.36, "grad_norm": 18.88192124289535, "learning_rate": 2.2655176505854527e-06, "loss": 0.5386, "step": 15135 }, { "epoch": 2.36, "grad_norm": 26.416832025854777, "learning_rate": 2.264448820610844e-06, "loss": 0.4835, "step": 15136 }, { "epoch": 2.36, "grad_norm": 27.42077197222985, "learning_rate": 2.2633802106307835e-06, "loss": 0.4408, "step": 15137 }, { "epoch": 2.36, "grad_norm": 23.94313458717503, "learning_rate": 2.2623118206756668e-06, "loss": 0.4415, "step": 15138 }, { "epoch": 2.36, "grad_norm": 16.750516646427478, "learning_rate": 2.2612436507758726e-06, "loss": 0.4662, "step": 15139 }, { "epoch": 2.36, "grad_norm": 19.288046312969854, "learning_rate": 2.260175700961785e-06, "loss": 0.4662, "step": 15140 }, { "epoch": 2.37, "grad_norm": 17.896086577545045, "learning_rate": 2.2591079712637698e-06, "loss": 0.4487, "step": 15141 }, { "epoch": 2.37, "grad_norm": 20.10662859500066, "learning_rate": 2.2580404617121954e-06, "loss": 0.4259, "step": 15142 }, { "epoch": 2.37, "grad_norm": 25.261615140085144, "learning_rate": 2.2569731723374243e-06, "loss": 0.4717, "step": 15143 }, { "epoch": 2.37, "grad_norm": 26.266033891062857, "learning_rate": 2.255906103169806e-06, "loss": 0.4583, "step": 15144 }, { "epoch": 2.37, "grad_norm": 22.01899804807438, "learning_rate": 2.2548392542396856e-06, "loss": 0.5174, "step": 15145 }, { "epoch": 2.37, "grad_norm": 25.078344651921057, "learning_rate": 2.2537726255774063e-06, "loss": 0.5355, "step": 15146 }, { "epoch": 2.37, "grad_norm": 19.49513815217672, "learning_rate": 2.252706217213302e-06, "loss": 0.4529, "step": 15147 }, { "epoch": 2.37, "grad_norm": 19.20640744803605, "learning_rate": 2.251640029177704e-06, "loss": 0.4513, "step": 15148 }, { "epoch": 2.37, "grad_norm": 22.20882558426196, "learning_rate": 2.250574061500931e-06, "loss": 0.4744, "step": 15149 }, { "epoch": 2.37, "grad_norm": 18.609538546018904, "learning_rate": 2.2495083142132944e-06, "loss": 0.4265, "step": 15150 }, { "epoch": 2.37, "grad_norm": 26.405454795195126, "learning_rate": 2.2484427873451086e-06, "loss": 0.475, "step": 15151 }, { "epoch": 2.37, "grad_norm": 19.365715220181023, "learning_rate": 2.247377480926678e-06, "loss": 0.4119, "step": 15152 }, { "epoch": 2.37, "grad_norm": 11.081622743119825, "learning_rate": 2.246312394988296e-06, "loss": 0.4392, "step": 15153 }, { "epoch": 2.37, "grad_norm": 20.932720703552434, "learning_rate": 2.2452475295602518e-06, "loss": 0.4564, "step": 15154 }, { "epoch": 2.37, "grad_norm": 18.11219220962849, "learning_rate": 2.2441828846728307e-06, "loss": 0.4462, "step": 15155 }, { "epoch": 2.37, "grad_norm": 21.680290516208032, "learning_rate": 2.243118460356312e-06, "loss": 0.4261, "step": 15156 }, { "epoch": 2.37, "grad_norm": 15.008526355659011, "learning_rate": 2.2420542566409686e-06, "loss": 0.4033, "step": 15157 }, { "epoch": 2.37, "grad_norm": 17.946891196248757, "learning_rate": 2.2409902735570643e-06, "loss": 0.4615, "step": 15158 }, { "epoch": 2.37, "grad_norm": 15.7169714617309, "learning_rate": 2.2399265111348558e-06, "loss": 0.5126, "step": 15159 }, { "epoch": 2.37, "grad_norm": 24.489491198999364, "learning_rate": 2.238862969404596e-06, "loss": 0.4774, "step": 15160 }, { "epoch": 2.37, "grad_norm": 21.558503354134128, "learning_rate": 2.2377996483965368e-06, "loss": 0.4408, "step": 15161 }, { "epoch": 2.37, "grad_norm": 24.781019896440743, "learning_rate": 2.236736548140913e-06, "loss": 0.5461, "step": 15162 }, { "epoch": 2.37, "grad_norm": 19.67023745713467, "learning_rate": 2.2356736686679624e-06, "loss": 0.504, "step": 15163 }, { "epoch": 2.37, "grad_norm": 17.553010393576393, "learning_rate": 2.2346110100079076e-06, "loss": 0.4241, "step": 15164 }, { "epoch": 2.37, "grad_norm": 14.042129224106823, "learning_rate": 2.2335485721909766e-06, "loss": 0.4849, "step": 15165 }, { "epoch": 2.37, "grad_norm": 19.516321256112047, "learning_rate": 2.2324863552473776e-06, "loss": 0.4786, "step": 15166 }, { "epoch": 2.37, "grad_norm": 30.363196988407843, "learning_rate": 2.2314243592073226e-06, "loss": 0.4663, "step": 15167 }, { "epoch": 2.37, "grad_norm": 14.034170561528624, "learning_rate": 2.230362584101018e-06, "loss": 0.4426, "step": 15168 }, { "epoch": 2.37, "grad_norm": 17.68042147095207, "learning_rate": 2.229301029958656e-06, "loss": 0.4172, "step": 15169 }, { "epoch": 2.37, "grad_norm": 17.598652276949753, "learning_rate": 2.2282396968104235e-06, "loss": 0.4323, "step": 15170 }, { "epoch": 2.37, "grad_norm": 16.253130448943175, "learning_rate": 2.227178584686509e-06, "loss": 0.4624, "step": 15171 }, { "epoch": 2.37, "grad_norm": 28.68904261295678, "learning_rate": 2.226117693617088e-06, "loss": 0.4532, "step": 15172 }, { "epoch": 2.37, "grad_norm": 18.15407636569071, "learning_rate": 2.2250570236323344e-06, "loss": 0.4531, "step": 15173 }, { "epoch": 2.37, "grad_norm": 17.526510357836578, "learning_rate": 2.2239965747624117e-06, "loss": 0.4421, "step": 15174 }, { "epoch": 2.37, "grad_norm": 22.863900434829066, "learning_rate": 2.222936347037474e-06, "loss": 0.5133, "step": 15175 }, { "epoch": 2.37, "grad_norm": 37.661697840448895, "learning_rate": 2.2218763404876775e-06, "loss": 0.5123, "step": 15176 }, { "epoch": 2.37, "grad_norm": 35.003177265298156, "learning_rate": 2.2208165551431706e-06, "loss": 0.4869, "step": 15177 }, { "epoch": 2.37, "grad_norm": 19.896395629073762, "learning_rate": 2.21975699103409e-06, "loss": 0.4571, "step": 15178 }, { "epoch": 2.37, "grad_norm": 20.09295527842895, "learning_rate": 2.2186976481905663e-06, "loss": 0.4676, "step": 15179 }, { "epoch": 2.37, "grad_norm": 33.75960256301567, "learning_rate": 2.2176385266427302e-06, "loss": 0.5048, "step": 15180 }, { "epoch": 2.37, "grad_norm": 23.864206347047276, "learning_rate": 2.216579626420702e-06, "loss": 0.4616, "step": 15181 }, { "epoch": 2.37, "grad_norm": 17.9554156526153, "learning_rate": 2.2155209475546013e-06, "loss": 0.4541, "step": 15182 }, { "epoch": 2.37, "grad_norm": 27.950230723146994, "learning_rate": 2.2144624900745247e-06, "loss": 0.4778, "step": 15183 }, { "epoch": 2.37, "grad_norm": 21.409156613119322, "learning_rate": 2.2134042540105814e-06, "loss": 0.4772, "step": 15184 }, { "epoch": 2.37, "grad_norm": 23.49391084901629, "learning_rate": 2.2123462393928663e-06, "loss": 0.4264, "step": 15185 }, { "epoch": 2.37, "grad_norm": 17.893617807793394, "learning_rate": 2.2112884462514707e-06, "loss": 0.4835, "step": 15186 }, { "epoch": 2.37, "grad_norm": 21.10285137495659, "learning_rate": 2.210230874616475e-06, "loss": 0.4756, "step": 15187 }, { "epoch": 2.37, "grad_norm": 20.442786691519522, "learning_rate": 2.2091735245179548e-06, "loss": 0.4194, "step": 15188 }, { "epoch": 2.37, "grad_norm": 25.597548588922628, "learning_rate": 2.208116395985981e-06, "loss": 0.4776, "step": 15189 }, { "epoch": 2.37, "grad_norm": 23.568131821060923, "learning_rate": 2.2070594890506216e-06, "loss": 0.4597, "step": 15190 }, { "epoch": 2.37, "grad_norm": 31.14080444529234, "learning_rate": 2.206002803741929e-06, "loss": 0.4679, "step": 15191 }, { "epoch": 2.37, "grad_norm": 23.002903793213502, "learning_rate": 2.2049463400899606e-06, "loss": 0.4415, "step": 15192 }, { "epoch": 2.37, "grad_norm": 21.99219630971295, "learning_rate": 2.2038900981247545e-06, "loss": 0.455, "step": 15193 }, { "epoch": 2.37, "grad_norm": 15.539661793398427, "learning_rate": 2.2028340778763544e-06, "loss": 0.4065, "step": 15194 }, { "epoch": 2.37, "grad_norm": 20.403671257978583, "learning_rate": 2.201778279374794e-06, "loss": 0.5221, "step": 15195 }, { "epoch": 2.37, "grad_norm": 20.47130216310369, "learning_rate": 2.2007227026500956e-06, "loss": 0.4356, "step": 15196 }, { "epoch": 2.37, "grad_norm": 23.045397497759183, "learning_rate": 2.199667347732284e-06, "loss": 0.5095, "step": 15197 }, { "epoch": 2.37, "grad_norm": 40.87522694578659, "learning_rate": 2.1986122146513654e-06, "loss": 0.4589, "step": 15198 }, { "epoch": 2.37, "grad_norm": 21.537142542464714, "learning_rate": 2.1975573034373563e-06, "loss": 0.4946, "step": 15199 }, { "epoch": 2.37, "grad_norm": 14.27352048351801, "learning_rate": 2.1965026141202495e-06, "loss": 0.4413, "step": 15200 }, { "epoch": 2.37, "grad_norm": 21.423561029841085, "learning_rate": 2.195448146730044e-06, "loss": 0.4912, "step": 15201 }, { "epoch": 2.37, "grad_norm": 26.706838841712543, "learning_rate": 2.1943939012967295e-06, "loss": 0.5352, "step": 15202 }, { "epoch": 2.37, "grad_norm": 18.325223293023893, "learning_rate": 2.193339877850288e-06, "loss": 0.4296, "step": 15203 }, { "epoch": 2.37, "grad_norm": 30.31120406408753, "learning_rate": 2.19228607642069e-06, "loss": 0.4846, "step": 15204 }, { "epoch": 2.38, "grad_norm": 22.647088281393838, "learning_rate": 2.1912324970379084e-06, "loss": 0.4687, "step": 15205 }, { "epoch": 2.38, "grad_norm": 21.04969849370558, "learning_rate": 2.190179139731906e-06, "loss": 0.4253, "step": 15206 }, { "epoch": 2.38, "grad_norm": 27.21354589410533, "learning_rate": 2.1891260045326467e-06, "loss": 0.4336, "step": 15207 }, { "epoch": 2.38, "grad_norm": 16.34410482362443, "learning_rate": 2.1880730914700687e-06, "loss": 0.3635, "step": 15208 }, { "epoch": 2.38, "grad_norm": 13.928483493779746, "learning_rate": 2.1870204005741226e-06, "loss": 0.4417, "step": 15209 }, { "epoch": 2.38, "grad_norm": 26.76386835443305, "learning_rate": 2.185967931874745e-06, "loss": 0.5163, "step": 15210 }, { "epoch": 2.38, "grad_norm": 23.337665815050382, "learning_rate": 2.1849156854018715e-06, "loss": 0.4789, "step": 15211 }, { "epoch": 2.38, "grad_norm": 21.98255458471436, "learning_rate": 2.1838636611854246e-06, "loss": 0.4906, "step": 15212 }, { "epoch": 2.38, "grad_norm": 17.883022576260995, "learning_rate": 2.1828118592553195e-06, "loss": 0.4344, "step": 15213 }, { "epoch": 2.38, "grad_norm": 19.319707726045827, "learning_rate": 2.181760279641473e-06, "loss": 0.4343, "step": 15214 }, { "epoch": 2.38, "grad_norm": 24.73896608619982, "learning_rate": 2.180708922373792e-06, "loss": 0.5158, "step": 15215 }, { "epoch": 2.38, "grad_norm": 21.008709923474523, "learning_rate": 2.1796577874821734e-06, "loss": 0.4283, "step": 15216 }, { "epoch": 2.38, "grad_norm": 26.025005388396746, "learning_rate": 2.178606874996515e-06, "loss": 0.5685, "step": 15217 }, { "epoch": 2.38, "grad_norm": 34.11862513931091, "learning_rate": 2.1775561849466987e-06, "loss": 0.5322, "step": 15218 }, { "epoch": 2.38, "grad_norm": 13.529657402317886, "learning_rate": 2.176505717362609e-06, "loss": 0.3853, "step": 15219 }, { "epoch": 2.38, "grad_norm": 23.28974909187821, "learning_rate": 2.1754554722741227e-06, "loss": 0.469, "step": 15220 }, { "epoch": 2.38, "grad_norm": 24.327234389088307, "learning_rate": 2.1744054497111033e-06, "loss": 0.5213, "step": 15221 }, { "epoch": 2.38, "grad_norm": 21.708528128637457, "learning_rate": 2.173355649703417e-06, "loss": 0.4262, "step": 15222 }, { "epoch": 2.38, "grad_norm": 25.981680667860044, "learning_rate": 2.1723060722809155e-06, "loss": 0.4433, "step": 15223 }, { "epoch": 2.38, "grad_norm": 18.021885487173225, "learning_rate": 2.1712567174734523e-06, "loss": 0.4914, "step": 15224 }, { "epoch": 2.38, "grad_norm": 16.080261608435933, "learning_rate": 2.170207585310865e-06, "loss": 0.4246, "step": 15225 }, { "epoch": 2.38, "grad_norm": 20.35669418406093, "learning_rate": 2.1691586758229976e-06, "loss": 0.4319, "step": 15226 }, { "epoch": 2.38, "grad_norm": 24.717864791727223, "learning_rate": 2.168109989039674e-06, "loss": 0.5165, "step": 15227 }, { "epoch": 2.38, "grad_norm": 19.313844705728073, "learning_rate": 2.1670615249907234e-06, "loss": 0.4082, "step": 15228 }, { "epoch": 2.38, "grad_norm": 22.27455009804577, "learning_rate": 2.1660132837059576e-06, "loss": 0.427, "step": 15229 }, { "epoch": 2.38, "grad_norm": 28.38154265305997, "learning_rate": 2.164965265215191e-06, "loss": 0.5739, "step": 15230 }, { "epoch": 2.38, "grad_norm": 21.880419604734247, "learning_rate": 2.1639174695482322e-06, "loss": 0.4886, "step": 15231 }, { "epoch": 2.38, "grad_norm": 20.93977163218643, "learning_rate": 2.162869896734876e-06, "loss": 0.5195, "step": 15232 }, { "epoch": 2.38, "grad_norm": 18.304286216475976, "learning_rate": 2.161822546804914e-06, "loss": 0.4787, "step": 15233 }, { "epoch": 2.38, "grad_norm": 18.907581546883872, "learning_rate": 2.1607754197881324e-06, "loss": 0.4781, "step": 15234 }, { "epoch": 2.38, "grad_norm": 18.136662726528343, "learning_rate": 2.1597285157143122e-06, "loss": 0.529, "step": 15235 }, { "epoch": 2.38, "grad_norm": 16.70373016542226, "learning_rate": 2.15868183461323e-06, "loss": 0.4257, "step": 15236 }, { "epoch": 2.38, "grad_norm": 15.803857775908057, "learning_rate": 2.1576353765146486e-06, "loss": 0.4258, "step": 15237 }, { "epoch": 2.38, "grad_norm": 21.926739208091387, "learning_rate": 2.1565891414483266e-06, "loss": 0.5076, "step": 15238 }, { "epoch": 2.38, "grad_norm": 17.914958365823445, "learning_rate": 2.155543129444021e-06, "loss": 0.4534, "step": 15239 }, { "epoch": 2.38, "grad_norm": 14.620965244687518, "learning_rate": 2.154497340531484e-06, "loss": 0.4115, "step": 15240 }, { "epoch": 2.38, "grad_norm": 20.489008931491618, "learning_rate": 2.1534517747404516e-06, "loss": 0.4078, "step": 15241 }, { "epoch": 2.38, "grad_norm": 21.020740197787905, "learning_rate": 2.152406432100659e-06, "loss": 0.4945, "step": 15242 }, { "epoch": 2.38, "grad_norm": 34.305891782958305, "learning_rate": 2.1513613126418364e-06, "loss": 0.5067, "step": 15243 }, { "epoch": 2.38, "grad_norm": 32.25065685682638, "learning_rate": 2.1503164163937062e-06, "loss": 0.5569, "step": 15244 }, { "epoch": 2.38, "grad_norm": 19.246492199405413, "learning_rate": 2.1492717433859887e-06, "loss": 0.4588, "step": 15245 }, { "epoch": 2.38, "grad_norm": 15.875655993289474, "learning_rate": 2.1482272936483896e-06, "loss": 0.4507, "step": 15246 }, { "epoch": 2.38, "grad_norm": 25.398889432586618, "learning_rate": 2.147183067210611e-06, "loss": 0.4635, "step": 15247 }, { "epoch": 2.38, "grad_norm": 16.829173364897283, "learning_rate": 2.1461390641023516e-06, "loss": 0.3928, "step": 15248 }, { "epoch": 2.38, "grad_norm": 18.466488108280227, "learning_rate": 2.145095284353307e-06, "loss": 0.4859, "step": 15249 }, { "epoch": 2.38, "grad_norm": 22.865140941282668, "learning_rate": 2.1440517279931528e-06, "loss": 0.4691, "step": 15250 }, { "epoch": 2.38, "grad_norm": 26.47114574214595, "learning_rate": 2.1430083950515755e-06, "loss": 0.5508, "step": 15251 }, { "epoch": 2.38, "grad_norm": 20.067723503341373, "learning_rate": 2.1419652855582406e-06, "loss": 0.4572, "step": 15252 }, { "epoch": 2.38, "grad_norm": 17.574947489603044, "learning_rate": 2.1409223995428187e-06, "loss": 0.4094, "step": 15253 }, { "epoch": 2.38, "grad_norm": 18.520740203244234, "learning_rate": 2.1398797370349644e-06, "loss": 0.4278, "step": 15254 }, { "epoch": 2.38, "grad_norm": 17.120394540791928, "learning_rate": 2.1388372980643315e-06, "loss": 0.4548, "step": 15255 }, { "epoch": 2.38, "grad_norm": 15.811261880598597, "learning_rate": 2.13779508266057e-06, "loss": 0.4311, "step": 15256 }, { "epoch": 2.38, "grad_norm": 15.838750567266487, "learning_rate": 2.136753090853314e-06, "loss": 0.4487, "step": 15257 }, { "epoch": 2.38, "grad_norm": 27.622548763507183, "learning_rate": 2.1357113226722036e-06, "loss": 0.4885, "step": 15258 }, { "epoch": 2.38, "grad_norm": 19.994836262835964, "learning_rate": 2.1346697781468593e-06, "loss": 0.4761, "step": 15259 }, { "epoch": 2.38, "grad_norm": 15.120743627985972, "learning_rate": 2.1336284573069067e-06, "loss": 0.4331, "step": 15260 }, { "epoch": 2.38, "grad_norm": 17.730562044693688, "learning_rate": 2.1325873601819613e-06, "loss": 0.4798, "step": 15261 }, { "epoch": 2.38, "grad_norm": 20.816124974843554, "learning_rate": 2.1315464868016287e-06, "loss": 0.4176, "step": 15262 }, { "epoch": 2.38, "grad_norm": 15.066580148553282, "learning_rate": 2.130505837195508e-06, "loss": 0.4189, "step": 15263 }, { "epoch": 2.38, "grad_norm": 24.81745196480578, "learning_rate": 2.129465411393198e-06, "loss": 0.4104, "step": 15264 }, { "epoch": 2.38, "grad_norm": 13.185868853329204, "learning_rate": 2.1284252094242908e-06, "loss": 0.4024, "step": 15265 }, { "epoch": 2.38, "grad_norm": 28.804102140810347, "learning_rate": 2.1273852313183663e-06, "loss": 0.5208, "step": 15266 }, { "epoch": 2.38, "grad_norm": 23.183263133904518, "learning_rate": 2.126345477104996e-06, "loss": 0.4094, "step": 15267 }, { "epoch": 2.38, "grad_norm": 15.447683793891619, "learning_rate": 2.125305946813756e-06, "loss": 0.464, "step": 15268 }, { "epoch": 2.39, "grad_norm": 17.236747514415185, "learning_rate": 2.1242666404742074e-06, "loss": 0.4207, "step": 15269 }, { "epoch": 2.39, "grad_norm": 35.1851084101385, "learning_rate": 2.1232275581159123e-06, "loss": 0.4921, "step": 15270 }, { "epoch": 2.39, "grad_norm": 27.142526780668916, "learning_rate": 2.122188699768416e-06, "loss": 0.5261, "step": 15271 }, { "epoch": 2.39, "grad_norm": 26.914903823591946, "learning_rate": 2.1211500654612625e-06, "loss": 0.4496, "step": 15272 }, { "epoch": 2.39, "grad_norm": 22.448980001916134, "learning_rate": 2.120111655223993e-06, "loss": 0.449, "step": 15273 }, { "epoch": 2.39, "grad_norm": 25.74322992625142, "learning_rate": 2.1190734690861403e-06, "loss": 0.4286, "step": 15274 }, { "epoch": 2.39, "grad_norm": 15.213273706092936, "learning_rate": 2.1180355070772287e-06, "loss": 0.3943, "step": 15275 }, { "epoch": 2.39, "grad_norm": 20.0684325212691, "learning_rate": 2.116997769226773e-06, "loss": 0.4632, "step": 15276 }, { "epoch": 2.39, "grad_norm": 17.165732116302507, "learning_rate": 2.11596025556429e-06, "loss": 0.4407, "step": 15277 }, { "epoch": 2.39, "grad_norm": 20.299523644748298, "learning_rate": 2.114922966119287e-06, "loss": 0.4801, "step": 15278 }, { "epoch": 2.39, "grad_norm": 22.816497956679257, "learning_rate": 2.113885900921261e-06, "loss": 0.4531, "step": 15279 }, { "epoch": 2.39, "grad_norm": 18.627782371408447, "learning_rate": 2.1128490599997078e-06, "loss": 0.4729, "step": 15280 }, { "epoch": 2.39, "grad_norm": 23.13071447853857, "learning_rate": 2.1118124433841114e-06, "loss": 0.4311, "step": 15281 }, { "epoch": 2.39, "grad_norm": 14.15286762311381, "learning_rate": 2.1107760511039553e-06, "loss": 0.401, "step": 15282 }, { "epoch": 2.39, "grad_norm": 14.096559783668733, "learning_rate": 2.109739883188715e-06, "loss": 0.4717, "step": 15283 }, { "epoch": 2.39, "grad_norm": 15.369880701111178, "learning_rate": 2.1087039396678544e-06, "loss": 0.4749, "step": 15284 }, { "epoch": 2.39, "grad_norm": 19.97325918166824, "learning_rate": 2.107668220570841e-06, "loss": 0.4723, "step": 15285 }, { "epoch": 2.39, "grad_norm": 19.392319367006277, "learning_rate": 2.1066327259271223e-06, "loss": 0.4417, "step": 15286 }, { "epoch": 2.39, "grad_norm": 19.358544998449325, "learning_rate": 2.1055974557661553e-06, "loss": 0.4639, "step": 15287 }, { "epoch": 2.39, "grad_norm": 15.538118530338991, "learning_rate": 2.1045624101173754e-06, "loss": 0.48, "step": 15288 }, { "epoch": 2.39, "grad_norm": 18.54379026748019, "learning_rate": 2.1035275890102214e-06, "loss": 0.4564, "step": 15289 }, { "epoch": 2.39, "grad_norm": 22.67548195167444, "learning_rate": 2.1024929924741265e-06, "loss": 0.4887, "step": 15290 }, { "epoch": 2.39, "grad_norm": 24.105772885345147, "learning_rate": 2.1014586205385113e-06, "loss": 0.451, "step": 15291 }, { "epoch": 2.39, "grad_norm": 15.880563801004168, "learning_rate": 2.1004244732327896e-06, "loss": 0.4955, "step": 15292 }, { "epoch": 2.39, "grad_norm": 18.33069642415794, "learning_rate": 2.0993905505863755e-06, "loss": 0.4027, "step": 15293 }, { "epoch": 2.39, "grad_norm": 20.187647428690234, "learning_rate": 2.098356852628671e-06, "loss": 0.4164, "step": 15294 }, { "epoch": 2.39, "grad_norm": 15.326887277138487, "learning_rate": 2.0973233793890812e-06, "loss": 0.4642, "step": 15295 }, { "epoch": 2.39, "grad_norm": 29.452539743735503, "learning_rate": 2.0962901308969864e-06, "loss": 0.4929, "step": 15296 }, { "epoch": 2.39, "grad_norm": 19.54224098553883, "learning_rate": 2.0952571071817763e-06, "loss": 0.491, "step": 15297 }, { "epoch": 2.39, "grad_norm": 17.126878092015584, "learning_rate": 2.094224308272831e-06, "loss": 0.4714, "step": 15298 }, { "epoch": 2.39, "grad_norm": 12.508026746074151, "learning_rate": 2.0931917341995233e-06, "loss": 0.4208, "step": 15299 }, { "epoch": 2.39, "grad_norm": 17.945344411732755, "learning_rate": 2.092159384991217e-06, "loss": 0.4564, "step": 15300 }, { "epoch": 2.39, "grad_norm": 29.7161328941377, "learning_rate": 2.09112726067727e-06, "loss": 0.4838, "step": 15301 }, { "epoch": 2.39, "grad_norm": 14.722281553197112, "learning_rate": 2.0900953612870367e-06, "loss": 0.4828, "step": 15302 }, { "epoch": 2.39, "grad_norm": 19.280879863388925, "learning_rate": 2.089063686849867e-06, "loss": 0.4541, "step": 15303 }, { "epoch": 2.39, "grad_norm": 24.265620390676048, "learning_rate": 2.0880322373950957e-06, "loss": 0.4254, "step": 15304 }, { "epoch": 2.39, "grad_norm": 11.847108449072214, "learning_rate": 2.0870010129520624e-06, "loss": 0.3639, "step": 15305 }, { "epoch": 2.39, "grad_norm": 19.175332719037446, "learning_rate": 2.085970013550088e-06, "loss": 0.4502, "step": 15306 }, { "epoch": 2.39, "grad_norm": 16.620323661778215, "learning_rate": 2.0849392392184963e-06, "loss": 0.4247, "step": 15307 }, { "epoch": 2.39, "grad_norm": 16.89964126482681, "learning_rate": 2.0839086899866055e-06, "loss": 0.3961, "step": 15308 }, { "epoch": 2.39, "grad_norm": 26.751031064724515, "learning_rate": 2.0828783658837194e-06, "loss": 0.503, "step": 15309 }, { "epoch": 2.39, "grad_norm": 19.3922401173626, "learning_rate": 2.0818482669391428e-06, "loss": 0.4951, "step": 15310 }, { "epoch": 2.39, "grad_norm": 29.168647014905293, "learning_rate": 2.080818393182167e-06, "loss": 0.4679, "step": 15311 }, { "epoch": 2.39, "grad_norm": 17.284134225318383, "learning_rate": 2.079788744642085e-06, "loss": 0.3895, "step": 15312 }, { "epoch": 2.39, "grad_norm": 17.468573998432976, "learning_rate": 2.078759321348177e-06, "loss": 0.4335, "step": 15313 }, { "epoch": 2.39, "grad_norm": 19.35319610402778, "learning_rate": 2.0777301233297187e-06, "loss": 0.4567, "step": 15314 }, { "epoch": 2.39, "grad_norm": 23.803433324202466, "learning_rate": 2.076701150615985e-06, "loss": 0.4792, "step": 15315 }, { "epoch": 2.39, "grad_norm": 25.24004927339067, "learning_rate": 2.075672403236235e-06, "loss": 0.4346, "step": 15316 }, { "epoch": 2.39, "grad_norm": 22.991296868892224, "learning_rate": 2.0746438812197244e-06, "loss": 0.456, "step": 15317 }, { "epoch": 2.39, "grad_norm": 27.957388735536725, "learning_rate": 2.0736155845957053e-06, "loss": 0.4773, "step": 15318 }, { "epoch": 2.39, "grad_norm": 15.605519259565634, "learning_rate": 2.0725875133934216e-06, "loss": 0.4481, "step": 15319 }, { "epoch": 2.39, "grad_norm": 19.73151073978322, "learning_rate": 2.0715596676421167e-06, "loss": 0.4388, "step": 15320 }, { "epoch": 2.39, "grad_norm": 17.991339745138355, "learning_rate": 2.070532047371012e-06, "loss": 0.4013, "step": 15321 }, { "epoch": 2.39, "grad_norm": 23.16828528988592, "learning_rate": 2.0695046526093375e-06, "loss": 0.4142, "step": 15322 }, { "epoch": 2.39, "grad_norm": 20.058899684159, "learning_rate": 2.0684774833863117e-06, "loss": 0.4261, "step": 15323 }, { "epoch": 2.39, "grad_norm": 21.119604420841213, "learning_rate": 2.067450539731148e-06, "loss": 0.465, "step": 15324 }, { "epoch": 2.39, "grad_norm": 24.24926345421895, "learning_rate": 2.0664238216730513e-06, "loss": 0.4469, "step": 15325 }, { "epoch": 2.39, "grad_norm": 16.12030764577204, "learning_rate": 2.0653973292412177e-06, "loss": 0.4801, "step": 15326 }, { "epoch": 2.39, "grad_norm": 14.922300739106715, "learning_rate": 2.0643710624648427e-06, "loss": 0.3977, "step": 15327 }, { "epoch": 2.39, "grad_norm": 16.896278905716102, "learning_rate": 2.063345021373114e-06, "loss": 0.4221, "step": 15328 }, { "epoch": 2.39, "grad_norm": 16.350592302390524, "learning_rate": 2.0623192059952114e-06, "loss": 0.4511, "step": 15329 }, { "epoch": 2.39, "grad_norm": 15.65658045513267, "learning_rate": 2.061293616360304e-06, "loss": 0.4579, "step": 15330 }, { "epoch": 2.39, "grad_norm": 32.541728066465026, "learning_rate": 2.0602682524975616e-06, "loss": 0.4114, "step": 15331 }, { "epoch": 2.39, "grad_norm": 24.310236105285252, "learning_rate": 2.0592431144361458e-06, "loss": 0.4857, "step": 15332 }, { "epoch": 2.4, "grad_norm": 25.37305633906512, "learning_rate": 2.0582182022052134e-06, "loss": 0.5092, "step": 15333 }, { "epoch": 2.4, "grad_norm": 18.066039909978777, "learning_rate": 2.0571935158339084e-06, "loss": 0.4638, "step": 15334 }, { "epoch": 2.4, "grad_norm": 24.331466619873176, "learning_rate": 2.0561690553513725e-06, "loss": 0.4662, "step": 15335 }, { "epoch": 2.4, "grad_norm": 28.017903718223245, "learning_rate": 2.0551448207867407e-06, "loss": 0.5036, "step": 15336 }, { "epoch": 2.4, "grad_norm": 21.294786548508146, "learning_rate": 2.0541208121691458e-06, "loss": 0.4489, "step": 15337 }, { "epoch": 2.4, "grad_norm": 18.567544250155628, "learning_rate": 2.0530970295277032e-06, "loss": 0.4863, "step": 15338 }, { "epoch": 2.4, "grad_norm": 17.931100286867284, "learning_rate": 2.0520734728915358e-06, "loss": 0.4708, "step": 15339 }, { "epoch": 2.4, "grad_norm": 37.74441704810459, "learning_rate": 2.0510501422897466e-06, "loss": 0.4936, "step": 15340 }, { "epoch": 2.4, "grad_norm": 23.6797677173722, "learning_rate": 2.050027037751444e-06, "loss": 0.4403, "step": 15341 }, { "epoch": 2.4, "grad_norm": 20.973351363146076, "learning_rate": 2.0490041593057185e-06, "loss": 0.4396, "step": 15342 }, { "epoch": 2.4, "grad_norm": 22.134208738840865, "learning_rate": 2.0479815069816643e-06, "loss": 0.4349, "step": 15343 }, { "epoch": 2.4, "grad_norm": 25.92914842273367, "learning_rate": 2.0469590808083674e-06, "loss": 0.4598, "step": 15344 }, { "epoch": 2.4, "grad_norm": 23.291027910721137, "learning_rate": 2.0459368808148983e-06, "loss": 0.4279, "step": 15345 }, { "epoch": 2.4, "grad_norm": 19.97231052934327, "learning_rate": 2.0449149070303344e-06, "loss": 0.5042, "step": 15346 }, { "epoch": 2.4, "grad_norm": 18.486329738739798, "learning_rate": 2.043893159483734e-06, "loss": 0.4563, "step": 15347 }, { "epoch": 2.4, "grad_norm": 32.7448715576632, "learning_rate": 2.042871638204158e-06, "loss": 0.4979, "step": 15348 }, { "epoch": 2.4, "grad_norm": 20.77696737173364, "learning_rate": 2.0418503432206604e-06, "loss": 0.5243, "step": 15349 }, { "epoch": 2.4, "grad_norm": 23.805140429332745, "learning_rate": 2.040829274562284e-06, "loss": 0.4392, "step": 15350 }, { "epoch": 2.4, "grad_norm": 23.66522374520072, "learning_rate": 2.0398084322580634e-06, "loss": 0.4811, "step": 15351 }, { "epoch": 2.4, "grad_norm": 24.87259473505785, "learning_rate": 2.0387878163370354e-06, "loss": 0.4471, "step": 15352 }, { "epoch": 2.4, "grad_norm": 17.038117560517303, "learning_rate": 2.0377674268282275e-06, "loss": 0.4631, "step": 15353 }, { "epoch": 2.4, "grad_norm": 20.373652843686266, "learning_rate": 2.0367472637606554e-06, "loss": 0.4535, "step": 15354 }, { "epoch": 2.4, "grad_norm": 18.776490327793674, "learning_rate": 2.0357273271633304e-06, "loss": 0.4974, "step": 15355 }, { "epoch": 2.4, "grad_norm": 18.12868131708863, "learning_rate": 2.0347076170652624e-06, "loss": 0.4719, "step": 15356 }, { "epoch": 2.4, "grad_norm": 26.62670633619575, "learning_rate": 2.0336881334954484e-06, "loss": 0.4814, "step": 15357 }, { "epoch": 2.4, "grad_norm": 13.178459496376743, "learning_rate": 2.0326688764828873e-06, "loss": 0.4184, "step": 15358 }, { "epoch": 2.4, "grad_norm": 16.562927807375075, "learning_rate": 2.0316498460565627e-06, "loss": 0.4516, "step": 15359 }, { "epoch": 2.4, "grad_norm": 15.897289915633605, "learning_rate": 2.030631042245452e-06, "loss": 0.4528, "step": 15360 }, { "epoch": 2.4, "grad_norm": 23.380814730794636, "learning_rate": 2.0296124650785333e-06, "loss": 0.4434, "step": 15361 }, { "epoch": 2.4, "grad_norm": 37.66922588717112, "learning_rate": 2.0285941145847754e-06, "loss": 0.5892, "step": 15362 }, { "epoch": 2.4, "grad_norm": 19.163152481233393, "learning_rate": 2.0275759907931356e-06, "loss": 0.4204, "step": 15363 }, { "epoch": 2.4, "grad_norm": 16.885929421890374, "learning_rate": 2.0265580937325735e-06, "loss": 0.4409, "step": 15364 }, { "epoch": 2.4, "grad_norm": 18.469361192656066, "learning_rate": 2.025540423432032e-06, "loss": 0.4698, "step": 15365 }, { "epoch": 2.4, "grad_norm": 21.994192649943056, "learning_rate": 2.024522979920458e-06, "loss": 0.4465, "step": 15366 }, { "epoch": 2.4, "grad_norm": 26.51679327306867, "learning_rate": 2.023505763226783e-06, "loss": 0.4988, "step": 15367 }, { "epoch": 2.4, "grad_norm": 17.77590399343352, "learning_rate": 2.0224887733799383e-06, "loss": 0.4089, "step": 15368 }, { "epoch": 2.4, "grad_norm": 32.02313432459231, "learning_rate": 2.021472010408848e-06, "loss": 0.4216, "step": 15369 }, { "epoch": 2.4, "grad_norm": 30.690017619258448, "learning_rate": 2.020455474342424e-06, "loss": 0.5699, "step": 15370 }, { "epoch": 2.4, "grad_norm": 20.010036847206578, "learning_rate": 2.019439165209581e-06, "loss": 0.4813, "step": 15371 }, { "epoch": 2.4, "grad_norm": 17.770772061073693, "learning_rate": 2.018423083039218e-06, "loss": 0.3951, "step": 15372 }, { "epoch": 2.4, "grad_norm": 25.315667106740317, "learning_rate": 2.0174072278602352e-06, "loss": 0.4572, "step": 15373 }, { "epoch": 2.4, "grad_norm": 16.238974084083967, "learning_rate": 2.0163915997015195e-06, "loss": 0.4257, "step": 15374 }, { "epoch": 2.4, "grad_norm": 22.102981106203305, "learning_rate": 2.0153761985919575e-06, "loss": 0.4603, "step": 15375 }, { "epoch": 2.4, "grad_norm": 28.179075897026795, "learning_rate": 2.014361024560424e-06, "loss": 0.52, "step": 15376 }, { "epoch": 2.4, "grad_norm": 17.324719065050832, "learning_rate": 2.0133460776357906e-06, "loss": 0.4679, "step": 15377 }, { "epoch": 2.4, "grad_norm": 14.442114778843433, "learning_rate": 2.012331357846926e-06, "loss": 0.4396, "step": 15378 }, { "epoch": 2.4, "grad_norm": 27.113192681389247, "learning_rate": 2.0113168652226843e-06, "loss": 0.5031, "step": 15379 }, { "epoch": 2.4, "grad_norm": 21.800191970076195, "learning_rate": 2.0103025997919155e-06, "loss": 0.4541, "step": 15380 }, { "epoch": 2.4, "grad_norm": 19.773256372641818, "learning_rate": 2.009288561583468e-06, "loss": 0.4666, "step": 15381 }, { "epoch": 2.4, "grad_norm": 25.079896243995304, "learning_rate": 2.008274750626178e-06, "loss": 0.4205, "step": 15382 }, { "epoch": 2.4, "grad_norm": 24.364590584423745, "learning_rate": 2.007261166948885e-06, "loss": 0.4824, "step": 15383 }, { "epoch": 2.4, "grad_norm": 26.51123863787077, "learning_rate": 2.006247810580403e-06, "loss": 0.4745, "step": 15384 }, { "epoch": 2.4, "grad_norm": 17.29417493985364, "learning_rate": 2.0052346815495573e-06, "loss": 0.5009, "step": 15385 }, { "epoch": 2.4, "grad_norm": 17.300373253885418, "learning_rate": 2.0042217798851606e-06, "loss": 0.4424, "step": 15386 }, { "epoch": 2.4, "grad_norm": 14.944193459635352, "learning_rate": 2.003209105616021e-06, "loss": 0.4249, "step": 15387 }, { "epoch": 2.4, "grad_norm": 19.081763144789335, "learning_rate": 2.0021966587709372e-06, "loss": 0.4179, "step": 15388 }, { "epoch": 2.4, "grad_norm": 26.19094276075015, "learning_rate": 2.0011844393786983e-06, "loss": 0.4952, "step": 15389 }, { "epoch": 2.4, "grad_norm": 15.347019968802632, "learning_rate": 2.0001724474680963e-06, "loss": 0.4252, "step": 15390 }, { "epoch": 2.4, "grad_norm": 20.528058081918658, "learning_rate": 1.999160683067911e-06, "loss": 0.4597, "step": 15391 }, { "epoch": 2.4, "grad_norm": 23.026360461180957, "learning_rate": 1.9981491462069146e-06, "loss": 0.4793, "step": 15392 }, { "epoch": 2.4, "grad_norm": 20.33475373845935, "learning_rate": 1.9971378369138773e-06, "loss": 0.4627, "step": 15393 }, { "epoch": 2.4, "grad_norm": 24.78181475636807, "learning_rate": 1.9961267552175558e-06, "loss": 0.4858, "step": 15394 }, { "epoch": 2.4, "grad_norm": 21.11313778125388, "learning_rate": 1.995115901146707e-06, "loss": 0.4304, "step": 15395 }, { "epoch": 2.4, "grad_norm": 26.311099526541714, "learning_rate": 1.9941052747300826e-06, "loss": 0.4489, "step": 15396 }, { "epoch": 2.41, "grad_norm": 19.42199272970177, "learning_rate": 1.9930948759964175e-06, "loss": 0.4383, "step": 15397 }, { "epoch": 2.41, "grad_norm": 28.49084642087205, "learning_rate": 1.992084704974453e-06, "loss": 0.4936, "step": 15398 }, { "epoch": 2.41, "grad_norm": 20.074301459503367, "learning_rate": 1.991074761692913e-06, "loss": 0.4381, "step": 15399 }, { "epoch": 2.41, "grad_norm": 22.3196861184539, "learning_rate": 1.9900650461805237e-06, "loss": 0.492, "step": 15400 }, { "epoch": 2.41, "grad_norm": 27.923558758099947, "learning_rate": 1.9890555584659965e-06, "loss": 0.4317, "step": 15401 }, { "epoch": 2.41, "grad_norm": 29.10296726190204, "learning_rate": 1.9880462985780423e-06, "loss": 0.4126, "step": 15402 }, { "epoch": 2.41, "grad_norm": 15.012345671436195, "learning_rate": 1.9870372665453673e-06, "loss": 0.5412, "step": 15403 }, { "epoch": 2.41, "grad_norm": 15.319882534691205, "learning_rate": 1.986028462396666e-06, "loss": 0.426, "step": 15404 }, { "epoch": 2.41, "grad_norm": 29.70867586217058, "learning_rate": 1.985019886160624e-06, "loss": 0.4587, "step": 15405 }, { "epoch": 2.41, "grad_norm": 18.001231797939358, "learning_rate": 1.9840115378659275e-06, "loss": 0.4341, "step": 15406 }, { "epoch": 2.41, "grad_norm": 20.576686755284673, "learning_rate": 1.983003417541254e-06, "loss": 0.4102, "step": 15407 }, { "epoch": 2.41, "grad_norm": 17.271254128035764, "learning_rate": 1.9819955252152755e-06, "loss": 0.4084, "step": 15408 }, { "epoch": 2.41, "grad_norm": 16.68404562964947, "learning_rate": 1.9809878609166546e-06, "loss": 0.4174, "step": 15409 }, { "epoch": 2.41, "grad_norm": 21.624490099336295, "learning_rate": 1.979980424674045e-06, "loss": 0.367, "step": 15410 }, { "epoch": 2.41, "grad_norm": 20.412502316791894, "learning_rate": 1.9789732165161e-06, "loss": 0.4526, "step": 15411 }, { "epoch": 2.41, "grad_norm": 22.2091591234103, "learning_rate": 1.977966236471468e-06, "loss": 0.4503, "step": 15412 }, { "epoch": 2.41, "grad_norm": 15.325366069873303, "learning_rate": 1.9769594845687833e-06, "loss": 0.3721, "step": 15413 }, { "epoch": 2.41, "grad_norm": 23.835925868983953, "learning_rate": 1.9759529608366744e-06, "loss": 0.4913, "step": 15414 }, { "epoch": 2.41, "grad_norm": 25.879553597181058, "learning_rate": 1.9749466653037707e-06, "loss": 0.4376, "step": 15415 }, { "epoch": 2.41, "grad_norm": 14.57163658773314, "learning_rate": 1.9739405979986904e-06, "loss": 0.4387, "step": 15416 }, { "epoch": 2.41, "grad_norm": 18.763706329939858, "learning_rate": 1.9729347589500426e-06, "loss": 0.468, "step": 15417 }, { "epoch": 2.41, "grad_norm": 13.819709399160494, "learning_rate": 1.9719291481864366e-06, "loss": 0.4323, "step": 15418 }, { "epoch": 2.41, "grad_norm": 19.61366202921762, "learning_rate": 1.970923765736468e-06, "loss": 0.4705, "step": 15419 }, { "epoch": 2.41, "grad_norm": 22.109965891069336, "learning_rate": 1.9699186116287295e-06, "loss": 0.4518, "step": 15420 }, { "epoch": 2.41, "grad_norm": 27.525286626730985, "learning_rate": 1.9689136858918112e-06, "loss": 0.4076, "step": 15421 }, { "epoch": 2.41, "grad_norm": 21.754325400294373, "learning_rate": 1.967908988554289e-06, "loss": 0.4334, "step": 15422 }, { "epoch": 2.41, "grad_norm": 35.943351270917994, "learning_rate": 1.9669045196447345e-06, "loss": 0.5155, "step": 15423 }, { "epoch": 2.41, "grad_norm": 21.22633731443197, "learning_rate": 1.9659002791917157e-06, "loss": 0.4026, "step": 15424 }, { "epoch": 2.41, "grad_norm": 24.75152008489357, "learning_rate": 1.964896267223797e-06, "loss": 0.5159, "step": 15425 }, { "epoch": 2.41, "grad_norm": 17.93363529949157, "learning_rate": 1.963892483769524e-06, "loss": 0.525, "step": 15426 }, { "epoch": 2.41, "grad_norm": 26.46723593622914, "learning_rate": 1.9628889288574514e-06, "loss": 0.5061, "step": 15427 }, { "epoch": 2.41, "grad_norm": 23.668548761977448, "learning_rate": 1.9618856025161127e-06, "loss": 0.4853, "step": 15428 }, { "epoch": 2.41, "grad_norm": 18.384002698692992, "learning_rate": 1.9608825047740486e-06, "loss": 0.4093, "step": 15429 }, { "epoch": 2.41, "grad_norm": 15.924566578618645, "learning_rate": 1.9598796356597806e-06, "loss": 0.4646, "step": 15430 }, { "epoch": 2.41, "grad_norm": 26.824214996964244, "learning_rate": 1.958876995201833e-06, "loss": 0.4867, "step": 15431 }, { "epoch": 2.41, "grad_norm": 20.065254880886613, "learning_rate": 1.9578745834287204e-06, "loss": 0.4278, "step": 15432 }, { "epoch": 2.41, "grad_norm": 15.789933509528122, "learning_rate": 1.9568724003689486e-06, "loss": 0.4337, "step": 15433 }, { "epoch": 2.41, "grad_norm": 30.280174263820577, "learning_rate": 1.9558704460510235e-06, "loss": 0.4887, "step": 15434 }, { "epoch": 2.41, "grad_norm": 30.785945669720157, "learning_rate": 1.9548687205034346e-06, "loss": 0.4184, "step": 15435 }, { "epoch": 2.41, "grad_norm": 16.040069254900477, "learning_rate": 1.953867223754673e-06, "loss": 0.4437, "step": 15436 }, { "epoch": 2.41, "grad_norm": 22.663784536914285, "learning_rate": 1.952865955833223e-06, "loss": 0.4312, "step": 15437 }, { "epoch": 2.41, "grad_norm": 15.958025664367458, "learning_rate": 1.9518649167675585e-06, "loss": 0.4752, "step": 15438 }, { "epoch": 2.41, "grad_norm": 18.61297948338482, "learning_rate": 1.9508641065861445e-06, "loss": 0.4297, "step": 15439 }, { "epoch": 2.41, "grad_norm": 20.158109998640352, "learning_rate": 1.949863525317447e-06, "loss": 0.4255, "step": 15440 }, { "epoch": 2.41, "grad_norm": 19.182126192799373, "learning_rate": 1.9488631729899243e-06, "loss": 0.4966, "step": 15441 }, { "epoch": 2.41, "grad_norm": 24.806468850398485, "learning_rate": 1.9478630496320227e-06, "loss": 0.4453, "step": 15442 }, { "epoch": 2.41, "grad_norm": 17.316897872565225, "learning_rate": 1.946863155272183e-06, "loss": 0.4356, "step": 15443 }, { "epoch": 2.41, "grad_norm": 18.5711907342403, "learning_rate": 1.9458634899388453e-06, "loss": 0.4, "step": 15444 }, { "epoch": 2.41, "grad_norm": 18.81115641857474, "learning_rate": 1.944864053660437e-06, "loss": 0.4453, "step": 15445 }, { "epoch": 2.41, "grad_norm": 18.257289195039252, "learning_rate": 1.9438648464653865e-06, "loss": 0.4563, "step": 15446 }, { "epoch": 2.41, "grad_norm": 27.242501838211776, "learning_rate": 1.9428658683821066e-06, "loss": 0.4796, "step": 15447 }, { "epoch": 2.41, "grad_norm": 21.693119120209765, "learning_rate": 1.9418671194390047e-06, "loss": 0.4608, "step": 15448 }, { "epoch": 2.41, "grad_norm": 15.888288650158334, "learning_rate": 1.9408685996644893e-06, "loss": 0.4281, "step": 15449 }, { "epoch": 2.41, "grad_norm": 23.671266004074816, "learning_rate": 1.9398703090869585e-06, "loss": 0.4544, "step": 15450 }, { "epoch": 2.41, "grad_norm": 26.165442782295703, "learning_rate": 1.938872247734799e-06, "loss": 0.4851, "step": 15451 }, { "epoch": 2.41, "grad_norm": 22.86494189204278, "learning_rate": 1.9378744156363993e-06, "loss": 0.4449, "step": 15452 }, { "epoch": 2.41, "grad_norm": 26.405277811584256, "learning_rate": 1.936876812820133e-06, "loss": 0.4911, "step": 15453 }, { "epoch": 2.41, "grad_norm": 23.213567757429583, "learning_rate": 1.9358794393143755e-06, "loss": 0.4307, "step": 15454 }, { "epoch": 2.41, "grad_norm": 17.93357932709829, "learning_rate": 1.934882295147487e-06, "loss": 0.4249, "step": 15455 }, { "epoch": 2.41, "grad_norm": 17.539829820614916, "learning_rate": 1.9338853803478274e-06, "loss": 0.4973, "step": 15456 }, { "epoch": 2.41, "grad_norm": 27.161147379474386, "learning_rate": 1.932888694943752e-06, "loss": 0.445, "step": 15457 }, { "epoch": 2.41, "grad_norm": 27.23288635140148, "learning_rate": 1.931892238963601e-06, "loss": 0.3949, "step": 15458 }, { "epoch": 2.41, "grad_norm": 17.009897063129984, "learning_rate": 1.9308960124357167e-06, "loss": 0.3791, "step": 15459 }, { "epoch": 2.41, "grad_norm": 35.84390328140876, "learning_rate": 1.9299000153884274e-06, "loss": 0.4538, "step": 15460 }, { "epoch": 2.42, "grad_norm": 21.902951677899033, "learning_rate": 1.928904247850061e-06, "loss": 0.4124, "step": 15461 }, { "epoch": 2.42, "grad_norm": 19.13498505469158, "learning_rate": 1.927908709848938e-06, "loss": 0.4538, "step": 15462 }, { "epoch": 2.42, "grad_norm": 19.210150469481583, "learning_rate": 1.9269134014133706e-06, "loss": 0.398, "step": 15463 }, { "epoch": 2.42, "grad_norm": 20.164378533589606, "learning_rate": 1.9259183225716594e-06, "loss": 0.5532, "step": 15464 }, { "epoch": 2.42, "grad_norm": 20.420596460686678, "learning_rate": 1.924923473352108e-06, "loss": 0.4582, "step": 15465 }, { "epoch": 2.42, "grad_norm": 25.620834669154583, "learning_rate": 1.9239288537830136e-06, "loss": 0.4289, "step": 15466 }, { "epoch": 2.42, "grad_norm": 32.40137726881236, "learning_rate": 1.922934463892657e-06, "loss": 0.5195, "step": 15467 }, { "epoch": 2.42, "grad_norm": 19.4995559923045, "learning_rate": 1.9219403037093164e-06, "loss": 0.5003, "step": 15468 }, { "epoch": 2.42, "grad_norm": 24.151146947461086, "learning_rate": 1.9209463732612687e-06, "loss": 0.4386, "step": 15469 }, { "epoch": 2.42, "grad_norm": 23.795229519280987, "learning_rate": 1.919952672576779e-06, "loss": 0.4837, "step": 15470 }, { "epoch": 2.42, "grad_norm": 19.702119020347897, "learning_rate": 1.9189592016841154e-06, "loss": 0.4508, "step": 15471 }, { "epoch": 2.42, "grad_norm": 23.393520058845333, "learning_rate": 1.9179659606115187e-06, "loss": 0.4745, "step": 15472 }, { "epoch": 2.42, "grad_norm": 27.81683536397295, "learning_rate": 1.9169729493872423e-06, "loss": 0.4712, "step": 15473 }, { "epoch": 2.42, "grad_norm": 25.68612940872227, "learning_rate": 1.9159801680395264e-06, "loss": 0.4178, "step": 15474 }, { "epoch": 2.42, "grad_norm": 36.52499344882476, "learning_rate": 1.9149876165966088e-06, "loss": 0.5706, "step": 15475 }, { "epoch": 2.42, "grad_norm": 14.428117968256183, "learning_rate": 1.9139952950867135e-06, "loss": 0.3901, "step": 15476 }, { "epoch": 2.42, "grad_norm": 28.240323277700302, "learning_rate": 1.9130032035380595e-06, "loss": 0.5087, "step": 15477 }, { "epoch": 2.42, "grad_norm": 18.345983534783485, "learning_rate": 1.9120113419788635e-06, "loss": 0.4124, "step": 15478 }, { "epoch": 2.42, "grad_norm": 28.73539116635506, "learning_rate": 1.911019710437335e-06, "loss": 0.4251, "step": 15479 }, { "epoch": 2.42, "grad_norm": 16.389330319163786, "learning_rate": 1.9100283089416725e-06, "loss": 0.4845, "step": 15480 }, { "epoch": 2.42, "grad_norm": 17.833135539329074, "learning_rate": 1.9090371375200755e-06, "loss": 0.4288, "step": 15481 }, { "epoch": 2.42, "grad_norm": 28.73584550535108, "learning_rate": 1.9080461962007257e-06, "loss": 0.4462, "step": 15482 }, { "epoch": 2.42, "grad_norm": 18.444229159599242, "learning_rate": 1.9070554850118095e-06, "loss": 0.3995, "step": 15483 }, { "epoch": 2.42, "grad_norm": 27.146113728480138, "learning_rate": 1.906065003981503e-06, "loss": 0.4411, "step": 15484 }, { "epoch": 2.42, "grad_norm": 20.86290123241126, "learning_rate": 1.9050747531379698e-06, "loss": 0.4998, "step": 15485 }, { "epoch": 2.42, "grad_norm": 27.4611787131556, "learning_rate": 1.9040847325093791e-06, "loss": 0.4423, "step": 15486 }, { "epoch": 2.42, "grad_norm": 20.461417487866154, "learning_rate": 1.9030949421238787e-06, "loss": 0.4245, "step": 15487 }, { "epoch": 2.42, "grad_norm": 11.842269138694503, "learning_rate": 1.9021053820096246e-06, "loss": 0.5377, "step": 15488 }, { "epoch": 2.42, "grad_norm": 20.282042748691268, "learning_rate": 1.9011160521947548e-06, "loss": 0.5069, "step": 15489 }, { "epoch": 2.42, "grad_norm": 20.012690382599313, "learning_rate": 1.900126952707405e-06, "loss": 0.4664, "step": 15490 }, { "epoch": 2.42, "grad_norm": 20.660097409166543, "learning_rate": 1.89913808357571e-06, "loss": 0.4061, "step": 15491 }, { "epoch": 2.42, "grad_norm": 17.227096898792517, "learning_rate": 1.8981494448277893e-06, "loss": 0.4035, "step": 15492 }, { "epoch": 2.42, "grad_norm": 20.70250947145214, "learning_rate": 1.897161036491756e-06, "loss": 0.4435, "step": 15493 }, { "epoch": 2.42, "grad_norm": 15.803582263133466, "learning_rate": 1.8961728585957229e-06, "loss": 0.4321, "step": 15494 }, { "epoch": 2.42, "grad_norm": 24.03617323495913, "learning_rate": 1.8951849111677922e-06, "loss": 0.5375, "step": 15495 }, { "epoch": 2.42, "grad_norm": 22.04168502338729, "learning_rate": 1.894197194236065e-06, "loss": 0.4519, "step": 15496 }, { "epoch": 2.42, "grad_norm": 16.570695008463197, "learning_rate": 1.8932097078286283e-06, "loss": 0.4613, "step": 15497 }, { "epoch": 2.42, "grad_norm": 23.397042540170354, "learning_rate": 1.8922224519735611e-06, "loss": 0.4647, "step": 15498 }, { "epoch": 2.42, "grad_norm": 20.23721492971823, "learning_rate": 1.8912354266989453e-06, "loss": 0.4537, "step": 15499 }, { "epoch": 2.42, "grad_norm": 17.640733786113206, "learning_rate": 1.890248632032854e-06, "loss": 0.4088, "step": 15500 }, { "epoch": 2.42, "grad_norm": 16.344779684642898, "learning_rate": 1.8892620680033457e-06, "loss": 0.4612, "step": 15501 }, { "epoch": 2.42, "grad_norm": 22.393334753011086, "learning_rate": 1.8882757346384785e-06, "loss": 0.4664, "step": 15502 }, { "epoch": 2.42, "grad_norm": 25.639841912752278, "learning_rate": 1.8872896319663038e-06, "loss": 0.4648, "step": 15503 }, { "epoch": 2.42, "grad_norm": 11.715601949015337, "learning_rate": 1.886303760014868e-06, "loss": 0.3739, "step": 15504 }, { "epoch": 2.42, "grad_norm": 20.487007607727623, "learning_rate": 1.885318118812205e-06, "loss": 0.4284, "step": 15505 }, { "epoch": 2.42, "grad_norm": 18.665020401948865, "learning_rate": 1.8843327083863495e-06, "loss": 0.3773, "step": 15506 }, { "epoch": 2.42, "grad_norm": 15.011841090803404, "learning_rate": 1.8833475287653225e-06, "loss": 0.4468, "step": 15507 }, { "epoch": 2.42, "grad_norm": 25.123861286468784, "learning_rate": 1.8823625799771428e-06, "loss": 0.4441, "step": 15508 }, { "epoch": 2.42, "grad_norm": 17.689392883258854, "learning_rate": 1.8813778620498268e-06, "loss": 0.4761, "step": 15509 }, { "epoch": 2.42, "grad_norm": 24.485711425186466, "learning_rate": 1.8803933750113712e-06, "loss": 0.3874, "step": 15510 }, { "epoch": 2.42, "grad_norm": 26.08969638941055, "learning_rate": 1.8794091188897812e-06, "loss": 0.4997, "step": 15511 }, { "epoch": 2.42, "grad_norm": 22.463916629092914, "learning_rate": 1.8784250937130433e-06, "loss": 0.4722, "step": 15512 }, { "epoch": 2.42, "grad_norm": 29.346236041874445, "learning_rate": 1.8774412995091462e-06, "loss": 0.5104, "step": 15513 }, { "epoch": 2.42, "grad_norm": 21.387853635175233, "learning_rate": 1.8764577363060654e-06, "loss": 0.525, "step": 15514 }, { "epoch": 2.42, "grad_norm": 26.565157835886573, "learning_rate": 1.8754744041317763e-06, "loss": 0.4754, "step": 15515 }, { "epoch": 2.42, "grad_norm": 24.772762045167756, "learning_rate": 1.8744913030142409e-06, "loss": 0.4753, "step": 15516 }, { "epoch": 2.42, "grad_norm": 21.084553604251518, "learning_rate": 1.8735084329814213e-06, "loss": 0.435, "step": 15517 }, { "epoch": 2.42, "grad_norm": 17.82897151171001, "learning_rate": 1.872525794061265e-06, "loss": 0.4226, "step": 15518 }, { "epoch": 2.42, "grad_norm": 14.005539103803844, "learning_rate": 1.8715433862817224e-06, "loss": 0.3689, "step": 15519 }, { "epoch": 2.42, "grad_norm": 21.70696593736945, "learning_rate": 1.8705612096707314e-06, "loss": 0.3947, "step": 15520 }, { "epoch": 2.42, "grad_norm": 14.32795276819128, "learning_rate": 1.8695792642562226e-06, "loss": 0.4634, "step": 15521 }, { "epoch": 2.42, "grad_norm": 16.592953810116423, "learning_rate": 1.8685975500661257e-06, "loss": 0.4318, "step": 15522 }, { "epoch": 2.42, "grad_norm": 32.822369493560196, "learning_rate": 1.867616067128355e-06, "loss": 0.4851, "step": 15523 }, { "epoch": 2.42, "grad_norm": 14.689784104714516, "learning_rate": 1.8666348154708269e-06, "loss": 0.429, "step": 15524 }, { "epoch": 2.43, "grad_norm": 21.81618102391986, "learning_rate": 1.8656537951214482e-06, "loss": 0.4085, "step": 15525 }, { "epoch": 2.43, "grad_norm": 23.91793502825176, "learning_rate": 1.8646730061081174e-06, "loss": 0.4567, "step": 15526 }, { "epoch": 2.43, "grad_norm": 22.960649132256282, "learning_rate": 1.8636924484587248e-06, "loss": 0.4243, "step": 15527 }, { "epoch": 2.43, "grad_norm": 24.55950211655765, "learning_rate": 1.86271212220116e-06, "loss": 0.4286, "step": 15528 }, { "epoch": 2.43, "grad_norm": 20.89256074221222, "learning_rate": 1.8617320273633044e-06, "loss": 0.4338, "step": 15529 }, { "epoch": 2.43, "grad_norm": 12.84907087098428, "learning_rate": 1.8607521639730285e-06, "loss": 0.3855, "step": 15530 }, { "epoch": 2.43, "grad_norm": 18.650400882214374, "learning_rate": 1.8597725320581982e-06, "loss": 0.4583, "step": 15531 }, { "epoch": 2.43, "grad_norm": 26.300842484516632, "learning_rate": 1.8587931316466746e-06, "loss": 0.4421, "step": 15532 }, { "epoch": 2.43, "grad_norm": 15.463793913339718, "learning_rate": 1.8578139627663127e-06, "loss": 0.4414, "step": 15533 }, { "epoch": 2.43, "grad_norm": 27.251963384737504, "learning_rate": 1.8568350254449596e-06, "loss": 0.5296, "step": 15534 }, { "epoch": 2.43, "grad_norm": 27.35784798210833, "learning_rate": 1.8558563197104562e-06, "loss": 0.4436, "step": 15535 }, { "epoch": 2.43, "grad_norm": 20.207881245193466, "learning_rate": 1.8548778455906325e-06, "loss": 0.4711, "step": 15536 }, { "epoch": 2.43, "grad_norm": 46.79857274066157, "learning_rate": 1.8538996031133172e-06, "loss": 0.4162, "step": 15537 }, { "epoch": 2.43, "grad_norm": 19.27897282744381, "learning_rate": 1.8529215923063348e-06, "loss": 0.4149, "step": 15538 }, { "epoch": 2.43, "grad_norm": 23.941187353591445, "learning_rate": 1.8519438131974932e-06, "loss": 0.52, "step": 15539 }, { "epoch": 2.43, "grad_norm": 21.009220502405334, "learning_rate": 1.8509662658146067e-06, "loss": 0.5146, "step": 15540 }, { "epoch": 2.43, "grad_norm": 15.457626560750542, "learning_rate": 1.8499889501854685e-06, "loss": 0.4807, "step": 15541 }, { "epoch": 2.43, "grad_norm": 19.223221916088125, "learning_rate": 1.8490118663378809e-06, "loss": 0.409, "step": 15542 }, { "epoch": 2.43, "grad_norm": 25.22119670766908, "learning_rate": 1.848035014299625e-06, "loss": 0.515, "step": 15543 }, { "epoch": 2.43, "grad_norm": 23.71409698187734, "learning_rate": 1.8470583940984832e-06, "loss": 0.434, "step": 15544 }, { "epoch": 2.43, "grad_norm": 16.1865642707045, "learning_rate": 1.8460820057622353e-06, "loss": 0.4485, "step": 15545 }, { "epoch": 2.43, "grad_norm": 28.92453661088341, "learning_rate": 1.8451058493186424e-06, "loss": 0.4738, "step": 15546 }, { "epoch": 2.43, "grad_norm": 18.645635838679198, "learning_rate": 1.8441299247954713e-06, "loss": 0.4146, "step": 15547 }, { "epoch": 2.43, "grad_norm": 13.40138518325208, "learning_rate": 1.843154232220471e-06, "loss": 0.4028, "step": 15548 }, { "epoch": 2.43, "grad_norm": 19.564279626909308, "learning_rate": 1.8421787716213934e-06, "loss": 0.4484, "step": 15549 }, { "epoch": 2.43, "grad_norm": 13.513939551734618, "learning_rate": 1.8412035430259811e-06, "loss": 0.4353, "step": 15550 }, { "epoch": 2.43, "grad_norm": 19.47360084561293, "learning_rate": 1.8402285464619673e-06, "loss": 0.4062, "step": 15551 }, { "epoch": 2.43, "grad_norm": 18.869193567127297, "learning_rate": 1.839253781957079e-06, "loss": 0.4228, "step": 15552 }, { "epoch": 2.43, "grad_norm": 20.74760828036703, "learning_rate": 1.8382792495390378e-06, "loss": 0.4246, "step": 15553 }, { "epoch": 2.43, "grad_norm": 26.291339030330214, "learning_rate": 1.8373049492355633e-06, "loss": 0.4226, "step": 15554 }, { "epoch": 2.43, "grad_norm": 17.175077504187207, "learning_rate": 1.8363308810743619e-06, "loss": 0.4958, "step": 15555 }, { "epoch": 2.43, "grad_norm": 21.893646382972268, "learning_rate": 1.8353570450831303e-06, "loss": 0.5396, "step": 15556 }, { "epoch": 2.43, "grad_norm": 20.594284514493253, "learning_rate": 1.83438344128957e-06, "loss": 0.4408, "step": 15557 }, { "epoch": 2.43, "grad_norm": 16.786191548393347, "learning_rate": 1.8334100697213664e-06, "loss": 0.4289, "step": 15558 }, { "epoch": 2.43, "grad_norm": 25.093224752731874, "learning_rate": 1.8324369304062062e-06, "loss": 0.4942, "step": 15559 }, { "epoch": 2.43, "grad_norm": 32.5159815503378, "learning_rate": 1.8314640233717617e-06, "loss": 0.4696, "step": 15560 }, { "epoch": 2.43, "grad_norm": 16.333487850142767, "learning_rate": 1.8304913486456998e-06, "loss": 0.4198, "step": 15561 }, { "epoch": 2.43, "grad_norm": 26.79393163425483, "learning_rate": 1.8295189062556851e-06, "loss": 0.5029, "step": 15562 }, { "epoch": 2.43, "grad_norm": 30.47662430370472, "learning_rate": 1.8285466962293752e-06, "loss": 0.456, "step": 15563 }, { "epoch": 2.43, "grad_norm": 18.83600360205312, "learning_rate": 1.827574718594417e-06, "loss": 0.4348, "step": 15564 }, { "epoch": 2.43, "grad_norm": 17.463346120194814, "learning_rate": 1.8266029733784506e-06, "loss": 0.4414, "step": 15565 }, { "epoch": 2.43, "grad_norm": 21.173209467846437, "learning_rate": 1.8256314606091153e-06, "loss": 0.4557, "step": 15566 }, { "epoch": 2.43, "grad_norm": 24.56113126268079, "learning_rate": 1.8246601803140407e-06, "loss": 0.4599, "step": 15567 }, { "epoch": 2.43, "grad_norm": 25.826642576888922, "learning_rate": 1.823689132520846e-06, "loss": 0.4432, "step": 15568 }, { "epoch": 2.43, "grad_norm": 19.693058125489113, "learning_rate": 1.8227183172571516e-06, "loss": 0.4092, "step": 15569 }, { "epoch": 2.43, "grad_norm": 15.781914976953203, "learning_rate": 1.8217477345505619e-06, "loss": 0.4502, "step": 15570 }, { "epoch": 2.43, "grad_norm": 27.233240033879568, "learning_rate": 1.8207773844286835e-06, "loss": 0.4368, "step": 15571 }, { "epoch": 2.43, "grad_norm": 19.52694403599269, "learning_rate": 1.8198072669191136e-06, "loss": 0.4958, "step": 15572 }, { "epoch": 2.43, "grad_norm": 22.219751880017537, "learning_rate": 1.8188373820494376e-06, "loss": 0.4521, "step": 15573 }, { "epoch": 2.43, "grad_norm": 25.045430620088553, "learning_rate": 1.8178677298472425e-06, "loss": 0.4735, "step": 15574 }, { "epoch": 2.43, "grad_norm": 22.332529526843867, "learning_rate": 1.8168983103401006e-06, "loss": 0.4764, "step": 15575 }, { "epoch": 2.43, "grad_norm": 17.68263924013169, "learning_rate": 1.8159291235555864e-06, "loss": 0.4603, "step": 15576 }, { "epoch": 2.43, "grad_norm": 26.4168065768868, "learning_rate": 1.8149601695212571e-06, "loss": 0.4721, "step": 15577 }, { "epoch": 2.43, "grad_norm": 14.256574396174575, "learning_rate": 1.8139914482646737e-06, "loss": 0.4618, "step": 15578 }, { "epoch": 2.43, "grad_norm": 35.16246848825398, "learning_rate": 1.8130229598133865e-06, "loss": 0.4945, "step": 15579 }, { "epoch": 2.43, "grad_norm": 24.050274288304607, "learning_rate": 1.8120547041949366e-06, "loss": 0.4849, "step": 15580 }, { "epoch": 2.43, "grad_norm": 24.73123641352344, "learning_rate": 1.81108668143686e-06, "loss": 0.4841, "step": 15581 }, { "epoch": 2.43, "grad_norm": 17.97760586319767, "learning_rate": 1.8101188915666867e-06, "loss": 0.4433, "step": 15582 }, { "epoch": 2.43, "grad_norm": 26.52608140765442, "learning_rate": 1.8091513346119415e-06, "loss": 0.4898, "step": 15583 }, { "epoch": 2.43, "grad_norm": 25.909912315841037, "learning_rate": 1.8081840106001435e-06, "loss": 0.4103, "step": 15584 }, { "epoch": 2.43, "grad_norm": 21.91392893215008, "learning_rate": 1.8072169195587996e-06, "loss": 0.4377, "step": 15585 }, { "epoch": 2.43, "grad_norm": 20.67713105815181, "learning_rate": 1.8062500615154121e-06, "loss": 0.5208, "step": 15586 }, { "epoch": 2.43, "grad_norm": 19.345071707390748, "learning_rate": 1.8052834364974791e-06, "loss": 0.4971, "step": 15587 }, { "epoch": 2.43, "grad_norm": 17.223992218249172, "learning_rate": 1.8043170445324943e-06, "loss": 0.401, "step": 15588 }, { "epoch": 2.44, "grad_norm": 20.01658956509949, "learning_rate": 1.8033508856479387e-06, "loss": 0.4579, "step": 15589 }, { "epoch": 2.44, "grad_norm": 16.816350958951464, "learning_rate": 1.8023849598712861e-06, "loss": 0.434, "step": 15590 }, { "epoch": 2.44, "grad_norm": 19.49204753221549, "learning_rate": 1.801419267230009e-06, "loss": 0.4251, "step": 15591 }, { "epoch": 2.44, "grad_norm": 20.476043720798017, "learning_rate": 1.8004538077515755e-06, "loss": 0.4256, "step": 15592 }, { "epoch": 2.44, "grad_norm": 14.009408024875126, "learning_rate": 1.7994885814634366e-06, "loss": 0.451, "step": 15593 }, { "epoch": 2.44, "grad_norm": 15.059859371806303, "learning_rate": 1.7985235883930475e-06, "loss": 0.4199, "step": 15594 }, { "epoch": 2.44, "grad_norm": 21.99359851057603, "learning_rate": 1.797558828567847e-06, "loss": 0.3981, "step": 15595 }, { "epoch": 2.44, "grad_norm": 22.7191569922826, "learning_rate": 1.7965943020152753e-06, "loss": 0.4991, "step": 15596 }, { "epoch": 2.44, "grad_norm": 15.35126964833335, "learning_rate": 1.7956300087627655e-06, "loss": 0.4206, "step": 15597 }, { "epoch": 2.44, "grad_norm": 22.147526625752768, "learning_rate": 1.794665948837736e-06, "loss": 0.421, "step": 15598 }, { "epoch": 2.44, "grad_norm": 15.842469635323038, "learning_rate": 1.7937021222676098e-06, "loss": 0.431, "step": 15599 }, { "epoch": 2.44, "grad_norm": 22.437492891273326, "learning_rate": 1.7927385290797917e-06, "loss": 0.4165, "step": 15600 }, { "epoch": 2.44, "grad_norm": 25.767849399350506, "learning_rate": 1.7917751693016915e-06, "loss": 0.4514, "step": 15601 }, { "epoch": 2.44, "grad_norm": 19.44545047461242, "learning_rate": 1.7908120429607024e-06, "loss": 0.421, "step": 15602 }, { "epoch": 2.44, "grad_norm": 15.466842943217578, "learning_rate": 1.7898491500842152e-06, "loss": 0.4603, "step": 15603 }, { "epoch": 2.44, "grad_norm": 14.735812061493073, "learning_rate": 1.7888864906996183e-06, "loss": 0.4109, "step": 15604 }, { "epoch": 2.44, "grad_norm": 21.371800362353035, "learning_rate": 1.7879240648342854e-06, "loss": 0.4994, "step": 15605 }, { "epoch": 2.44, "grad_norm": 20.85608643136716, "learning_rate": 1.7869618725155868e-06, "loss": 0.426, "step": 15606 }, { "epoch": 2.44, "grad_norm": 21.468068342478652, "learning_rate": 1.7859999137708872e-06, "loss": 0.4487, "step": 15607 }, { "epoch": 2.44, "grad_norm": 22.705747158243064, "learning_rate": 1.7850381886275458e-06, "loss": 0.497, "step": 15608 }, { "epoch": 2.44, "grad_norm": 18.666298408597527, "learning_rate": 1.7840766971129153e-06, "loss": 0.493, "step": 15609 }, { "epoch": 2.44, "grad_norm": 19.22067524485722, "learning_rate": 1.783115439254337e-06, "loss": 0.4137, "step": 15610 }, { "epoch": 2.44, "grad_norm": 23.730136175140682, "learning_rate": 1.7821544150791459e-06, "loss": 0.4131, "step": 15611 }, { "epoch": 2.44, "grad_norm": 22.197540813171795, "learning_rate": 1.7811936246146765e-06, "loss": 0.5077, "step": 15612 }, { "epoch": 2.44, "grad_norm": 27.45459468948094, "learning_rate": 1.7802330678882552e-06, "loss": 0.4683, "step": 15613 }, { "epoch": 2.44, "grad_norm": 19.172755624557027, "learning_rate": 1.7792727449271962e-06, "loss": 0.4453, "step": 15614 }, { "epoch": 2.44, "grad_norm": 28.151691070862654, "learning_rate": 1.7783126557588092e-06, "loss": 0.5051, "step": 15615 }, { "epoch": 2.44, "grad_norm": 17.603273924833005, "learning_rate": 1.7773528004104012e-06, "loss": 0.3779, "step": 15616 }, { "epoch": 2.44, "grad_norm": 23.866715365670732, "learning_rate": 1.7763931789092715e-06, "loss": 0.4988, "step": 15617 }, { "epoch": 2.44, "grad_norm": 23.680668224387993, "learning_rate": 1.7754337912827092e-06, "loss": 0.4164, "step": 15618 }, { "epoch": 2.44, "grad_norm": 24.94071613991043, "learning_rate": 1.7744746375579968e-06, "loss": 0.4503, "step": 15619 }, { "epoch": 2.44, "grad_norm": 19.952616256435636, "learning_rate": 1.7735157177624129e-06, "loss": 0.4294, "step": 15620 }, { "epoch": 2.44, "grad_norm": 23.878098029710657, "learning_rate": 1.77255703192323e-06, "loss": 0.4186, "step": 15621 }, { "epoch": 2.44, "grad_norm": 18.756194665738846, "learning_rate": 1.7715985800677148e-06, "loss": 0.4426, "step": 15622 }, { "epoch": 2.44, "grad_norm": 21.196275182305122, "learning_rate": 1.7706403622231228e-06, "loss": 0.4844, "step": 15623 }, { "epoch": 2.44, "grad_norm": 13.582590126210418, "learning_rate": 1.7696823784167017e-06, "loss": 0.4085, "step": 15624 }, { "epoch": 2.44, "grad_norm": 16.921045101593744, "learning_rate": 1.7687246286756999e-06, "loss": 0.4785, "step": 15625 }, { "epoch": 2.44, "grad_norm": 27.30614475842026, "learning_rate": 1.7677671130273567e-06, "loss": 0.4916, "step": 15626 }, { "epoch": 2.44, "grad_norm": 23.09746587820967, "learning_rate": 1.766809831498898e-06, "loss": 0.4793, "step": 15627 }, { "epoch": 2.44, "grad_norm": 17.991629341336196, "learning_rate": 1.7658527841175543e-06, "loss": 0.4679, "step": 15628 }, { "epoch": 2.44, "grad_norm": 19.524029075937946, "learning_rate": 1.7648959709105384e-06, "loss": 0.4695, "step": 15629 }, { "epoch": 2.44, "grad_norm": 23.787327944967217, "learning_rate": 1.7639393919050662e-06, "loss": 0.4756, "step": 15630 }, { "epoch": 2.44, "grad_norm": 19.943533902893822, "learning_rate": 1.762983047128337e-06, "loss": 0.4014, "step": 15631 }, { "epoch": 2.44, "grad_norm": 18.80827617634642, "learning_rate": 1.7620269366075514e-06, "loss": 0.4168, "step": 15632 }, { "epoch": 2.44, "grad_norm": 19.99525715500652, "learning_rate": 1.7610710603699033e-06, "loss": 0.4677, "step": 15633 }, { "epoch": 2.44, "grad_norm": 19.448211092185396, "learning_rate": 1.7601154184425717e-06, "loss": 0.4523, "step": 15634 }, { "epoch": 2.44, "grad_norm": 21.916203340889655, "learning_rate": 1.7591600108527407e-06, "loss": 0.4635, "step": 15635 }, { "epoch": 2.44, "grad_norm": 16.464627713633504, "learning_rate": 1.758204837627575e-06, "loss": 0.3869, "step": 15636 }, { "epoch": 2.44, "grad_norm": 19.811552922359514, "learning_rate": 1.757249898794242e-06, "loss": 0.4253, "step": 15637 }, { "epoch": 2.44, "grad_norm": 21.851804911960375, "learning_rate": 1.7562951943799033e-06, "loss": 0.4138, "step": 15638 }, { "epoch": 2.44, "grad_norm": 17.47046993264175, "learning_rate": 1.755340724411707e-06, "loss": 0.4075, "step": 15639 }, { "epoch": 2.44, "grad_norm": 23.026540968396244, "learning_rate": 1.754386488916795e-06, "loss": 0.5165, "step": 15640 }, { "epoch": 2.44, "grad_norm": 13.018244045936102, "learning_rate": 1.7534324879223065e-06, "loss": 0.3948, "step": 15641 }, { "epoch": 2.44, "grad_norm": 18.24360875182509, "learning_rate": 1.7524787214553774e-06, "loss": 0.5273, "step": 15642 }, { "epoch": 2.44, "grad_norm": 24.330801428254386, "learning_rate": 1.7515251895431295e-06, "loss": 0.486, "step": 15643 }, { "epoch": 2.44, "grad_norm": 30.040085542716724, "learning_rate": 1.750571892212677e-06, "loss": 0.5172, "step": 15644 }, { "epoch": 2.44, "grad_norm": 18.782636953037628, "learning_rate": 1.7496188294911332e-06, "loss": 0.5043, "step": 15645 }, { "epoch": 2.44, "grad_norm": 17.405097401911334, "learning_rate": 1.7486660014056046e-06, "loss": 0.4902, "step": 15646 }, { "epoch": 2.44, "grad_norm": 22.60867511151067, "learning_rate": 1.7477134079831893e-06, "loss": 0.4531, "step": 15647 }, { "epoch": 2.44, "grad_norm": 17.777383447618277, "learning_rate": 1.746761049250979e-06, "loss": 0.4798, "step": 15648 }, { "epoch": 2.44, "grad_norm": 20.04281763359204, "learning_rate": 1.7458089252360522e-06, "loss": 0.468, "step": 15649 }, { "epoch": 2.44, "grad_norm": 22.84701788420225, "learning_rate": 1.7448570359654925e-06, "loss": 0.4395, "step": 15650 }, { "epoch": 2.44, "grad_norm": 15.415634407676242, "learning_rate": 1.743905381466371e-06, "loss": 0.4606, "step": 15651 }, { "epoch": 2.44, "grad_norm": 22.031542974918363, "learning_rate": 1.7429539617657487e-06, "loss": 0.4578, "step": 15652 }, { "epoch": 2.45, "grad_norm": 14.683632227752598, "learning_rate": 1.7420027768906877e-06, "loss": 0.5084, "step": 15653 }, { "epoch": 2.45, "grad_norm": 20.666012341353326, "learning_rate": 1.7410518268682353e-06, "loss": 0.4138, "step": 15654 }, { "epoch": 2.45, "grad_norm": 24.73429182265147, "learning_rate": 1.7401011117254397e-06, "loss": 0.4406, "step": 15655 }, { "epoch": 2.45, "grad_norm": 23.55104984896507, "learning_rate": 1.7391506314893337e-06, "loss": 0.4558, "step": 15656 }, { "epoch": 2.45, "grad_norm": 17.390192437768842, "learning_rate": 1.7382003861869512e-06, "loss": 0.4631, "step": 15657 }, { "epoch": 2.45, "grad_norm": 18.926578607680682, "learning_rate": 1.7372503758453196e-06, "loss": 0.4201, "step": 15658 }, { "epoch": 2.45, "grad_norm": 37.3928337257797, "learning_rate": 1.7363006004914507e-06, "loss": 0.4767, "step": 15659 }, { "epoch": 2.45, "grad_norm": 16.056462213480682, "learning_rate": 1.7353510601523616e-06, "loss": 0.4543, "step": 15660 }, { "epoch": 2.45, "grad_norm": 21.124255974211724, "learning_rate": 1.7344017548550506e-06, "loss": 0.4279, "step": 15661 }, { "epoch": 2.45, "grad_norm": 20.64766689728801, "learning_rate": 1.733452684626522e-06, "loss": 0.4757, "step": 15662 }, { "epoch": 2.45, "grad_norm": 21.501590290138942, "learning_rate": 1.7325038494937597e-06, "loss": 0.418, "step": 15663 }, { "epoch": 2.45, "grad_norm": 40.71444599038201, "learning_rate": 1.7315552494837552e-06, "loss": 0.5379, "step": 15664 }, { "epoch": 2.45, "grad_norm": 16.89527322407755, "learning_rate": 1.7306068846234791e-06, "loss": 0.405, "step": 15665 }, { "epoch": 2.45, "grad_norm": 33.276344008462154, "learning_rate": 1.7296587549399057e-06, "loss": 0.4574, "step": 15666 }, { "epoch": 2.45, "grad_norm": 17.912322647377387, "learning_rate": 1.7287108604600023e-06, "loss": 0.4568, "step": 15667 }, { "epoch": 2.45, "grad_norm": 20.790070236593376, "learning_rate": 1.7277632012107236e-06, "loss": 0.4321, "step": 15668 }, { "epoch": 2.45, "grad_norm": 21.828375157511363, "learning_rate": 1.7268157772190174e-06, "loss": 0.4351, "step": 15669 }, { "epoch": 2.45, "grad_norm": 35.59040206179848, "learning_rate": 1.725868588511831e-06, "loss": 0.4765, "step": 15670 }, { "epoch": 2.45, "grad_norm": 18.292652946057988, "learning_rate": 1.7249216351161013e-06, "loss": 0.4741, "step": 15671 }, { "epoch": 2.45, "grad_norm": 32.32808279308516, "learning_rate": 1.7239749170587628e-06, "loss": 0.5315, "step": 15672 }, { "epoch": 2.45, "grad_norm": 26.633858747457225, "learning_rate": 1.723028434366736e-06, "loss": 0.4776, "step": 15673 }, { "epoch": 2.45, "grad_norm": 12.156949570465212, "learning_rate": 1.7220821870669358e-06, "loss": 0.3494, "step": 15674 }, { "epoch": 2.45, "grad_norm": 30.33912482109138, "learning_rate": 1.7211361751862754e-06, "loss": 0.556, "step": 15675 }, { "epoch": 2.45, "grad_norm": 21.094025046356762, "learning_rate": 1.720190398751662e-06, "loss": 0.4877, "step": 15676 }, { "epoch": 2.45, "grad_norm": 33.400337224103104, "learning_rate": 1.7192448577899912e-06, "loss": 0.4672, "step": 15677 }, { "epoch": 2.45, "grad_norm": 19.280156895180415, "learning_rate": 1.7182995523281488e-06, "loss": 0.3835, "step": 15678 }, { "epoch": 2.45, "grad_norm": 15.90173532132101, "learning_rate": 1.7173544823930232e-06, "loss": 0.4142, "step": 15679 }, { "epoch": 2.45, "grad_norm": 22.076843355454923, "learning_rate": 1.7164096480114934e-06, "loss": 0.3858, "step": 15680 }, { "epoch": 2.45, "grad_norm": 20.93684716986153, "learning_rate": 1.715465049210424e-06, "loss": 0.4491, "step": 15681 }, { "epoch": 2.45, "grad_norm": 17.915089700065323, "learning_rate": 1.7145206860166863e-06, "loss": 0.4833, "step": 15682 }, { "epoch": 2.45, "grad_norm": 20.768664691547805, "learning_rate": 1.7135765584571296e-06, "loss": 0.4888, "step": 15683 }, { "epoch": 2.45, "grad_norm": 23.222977411601583, "learning_rate": 1.712632666558609e-06, "loss": 0.4562, "step": 15684 }, { "epoch": 2.45, "grad_norm": 29.22592705869698, "learning_rate": 1.7116890103479688e-06, "loss": 0.4759, "step": 15685 }, { "epoch": 2.45, "grad_norm": 20.455293674475193, "learning_rate": 1.710745589852043e-06, "loss": 0.4375, "step": 15686 }, { "epoch": 2.45, "grad_norm": 26.645101678992134, "learning_rate": 1.7098024050976657e-06, "loss": 0.43, "step": 15687 }, { "epoch": 2.45, "grad_norm": 23.064482267084994, "learning_rate": 1.7088594561116567e-06, "loss": 0.4928, "step": 15688 }, { "epoch": 2.45, "grad_norm": 42.91025739196887, "learning_rate": 1.7079167429208366e-06, "loss": 0.428, "step": 15689 }, { "epoch": 2.45, "grad_norm": 13.843329713869773, "learning_rate": 1.7069742655520106e-06, "loss": 0.4348, "step": 15690 }, { "epoch": 2.45, "grad_norm": 30.96254361674733, "learning_rate": 1.706032024031985e-06, "loss": 0.4761, "step": 15691 }, { "epoch": 2.45, "grad_norm": 18.25259133208445, "learning_rate": 1.7050900183875608e-06, "loss": 0.4165, "step": 15692 }, { "epoch": 2.45, "grad_norm": 24.25572723234725, "learning_rate": 1.704148248645523e-06, "loss": 0.4813, "step": 15693 }, { "epoch": 2.45, "grad_norm": 17.79142915029655, "learning_rate": 1.7032067148326536e-06, "loss": 0.4392, "step": 15694 }, { "epoch": 2.45, "grad_norm": 21.939948738332422, "learning_rate": 1.7022654169757312e-06, "loss": 0.4084, "step": 15695 }, { "epoch": 2.45, "grad_norm": 24.65556954882072, "learning_rate": 1.701324355101528e-06, "loss": 0.4667, "step": 15696 }, { "epoch": 2.45, "grad_norm": 19.71053279245022, "learning_rate": 1.7003835292368064e-06, "loss": 0.5133, "step": 15697 }, { "epoch": 2.45, "grad_norm": 22.391741919760776, "learning_rate": 1.6994429394083235e-06, "loss": 0.4073, "step": 15698 }, { "epoch": 2.45, "grad_norm": 16.278930824709033, "learning_rate": 1.698502585642824e-06, "loss": 0.4607, "step": 15699 }, { "epoch": 2.45, "grad_norm": 21.060065826381518, "learning_rate": 1.6975624679670556e-06, "loss": 0.4082, "step": 15700 }, { "epoch": 2.45, "grad_norm": 34.31509593921933, "learning_rate": 1.6966225864077557e-06, "loss": 0.5108, "step": 15701 }, { "epoch": 2.45, "grad_norm": 20.508322050031307, "learning_rate": 1.695682940991652e-06, "loss": 0.4909, "step": 15702 }, { "epoch": 2.45, "grad_norm": 22.985910287871675, "learning_rate": 1.6947435317454653e-06, "loss": 0.4292, "step": 15703 }, { "epoch": 2.45, "grad_norm": 19.47548698427316, "learning_rate": 1.6938043586959141e-06, "loss": 0.4284, "step": 15704 }, { "epoch": 2.45, "grad_norm": 12.050977083961271, "learning_rate": 1.6928654218697105e-06, "loss": 0.3717, "step": 15705 }, { "epoch": 2.45, "grad_norm": 20.90466366723827, "learning_rate": 1.6919267212935531e-06, "loss": 0.5338, "step": 15706 }, { "epoch": 2.45, "grad_norm": 21.04390598473832, "learning_rate": 1.6909882569941405e-06, "loss": 0.5185, "step": 15707 }, { "epoch": 2.45, "grad_norm": 19.311964475897952, "learning_rate": 1.6900500289981603e-06, "loss": 0.4271, "step": 15708 }, { "epoch": 2.45, "grad_norm": 19.922750902555027, "learning_rate": 1.6891120373322955e-06, "loss": 0.4614, "step": 15709 }, { "epoch": 2.45, "grad_norm": 22.186880507350157, "learning_rate": 1.6881742820232249e-06, "loss": 0.4484, "step": 15710 }, { "epoch": 2.45, "grad_norm": 20.79142881944137, "learning_rate": 1.687236763097615e-06, "loss": 0.5004, "step": 15711 }, { "epoch": 2.45, "grad_norm": 28.39079508785594, "learning_rate": 1.6862994805821275e-06, "loss": 0.4618, "step": 15712 }, { "epoch": 2.45, "grad_norm": 34.46493770176251, "learning_rate": 1.685362434503418e-06, "loss": 0.4565, "step": 15713 }, { "epoch": 2.45, "grad_norm": 18.168428165464427, "learning_rate": 1.6844256248881408e-06, "loss": 0.4587, "step": 15714 }, { "epoch": 2.45, "grad_norm": 20.442181943364776, "learning_rate": 1.6834890517629309e-06, "loss": 0.4486, "step": 15715 }, { "epoch": 2.45, "grad_norm": 25.33710046546341, "learning_rate": 1.6825527151544307e-06, "loss": 0.4138, "step": 15716 }, { "epoch": 2.46, "grad_norm": 20.24390139624741, "learning_rate": 1.6816166150892622e-06, "loss": 0.481, "step": 15717 }, { "epoch": 2.46, "grad_norm": 23.863243445189916, "learning_rate": 1.6806807515940537e-06, "loss": 0.4497, "step": 15718 }, { "epoch": 2.46, "grad_norm": 23.33179953520906, "learning_rate": 1.6797451246954166e-06, "loss": 0.4594, "step": 15719 }, { "epoch": 2.46, "grad_norm": 15.528177594521686, "learning_rate": 1.6788097344199594e-06, "loss": 0.4584, "step": 15720 }, { "epoch": 2.46, "grad_norm": 14.745148863171943, "learning_rate": 1.6778745807942899e-06, "loss": 0.3949, "step": 15721 }, { "epoch": 2.46, "grad_norm": 21.658972075059108, "learning_rate": 1.6769396638449954e-06, "loss": 0.438, "step": 15722 }, { "epoch": 2.46, "grad_norm": 29.867973959384695, "learning_rate": 1.6760049835986703e-06, "loss": 0.4756, "step": 15723 }, { "epoch": 2.46, "grad_norm": 31.91997725680295, "learning_rate": 1.6750705400818913e-06, "loss": 0.4711, "step": 15724 }, { "epoch": 2.46, "grad_norm": 25.30707162677131, "learning_rate": 1.6741363333212368e-06, "loss": 0.4578, "step": 15725 }, { "epoch": 2.46, "grad_norm": 16.47445156765777, "learning_rate": 1.673202363343277e-06, "loss": 0.3928, "step": 15726 }, { "epoch": 2.46, "grad_norm": 23.360269565740566, "learning_rate": 1.6722686301745716e-06, "loss": 0.4526, "step": 15727 }, { "epoch": 2.46, "grad_norm": 18.272234146009826, "learning_rate": 1.6713351338416707e-06, "loss": 0.5207, "step": 15728 }, { "epoch": 2.46, "grad_norm": 18.550500597329904, "learning_rate": 1.670401874371128e-06, "loss": 0.475, "step": 15729 }, { "epoch": 2.46, "grad_norm": 20.184570663904687, "learning_rate": 1.6694688517894852e-06, "loss": 0.4715, "step": 15730 }, { "epoch": 2.46, "grad_norm": 22.688760559540253, "learning_rate": 1.6685360661232742e-06, "loss": 0.5298, "step": 15731 }, { "epoch": 2.46, "grad_norm": 21.263396481997386, "learning_rate": 1.6676035173990225e-06, "loss": 0.488, "step": 15732 }, { "epoch": 2.46, "grad_norm": 13.907477818623446, "learning_rate": 1.6666712056432522e-06, "loss": 0.4376, "step": 15733 }, { "epoch": 2.46, "grad_norm": 23.06244295360002, "learning_rate": 1.6657391308824777e-06, "loss": 0.3825, "step": 15734 }, { "epoch": 2.46, "grad_norm": 14.560467777766524, "learning_rate": 1.6648072931432091e-06, "loss": 0.4598, "step": 15735 }, { "epoch": 2.46, "grad_norm": 15.348141967163514, "learning_rate": 1.663875692451945e-06, "loss": 0.4161, "step": 15736 }, { "epoch": 2.46, "grad_norm": 28.690267492226873, "learning_rate": 1.6629443288351777e-06, "loss": 0.4758, "step": 15737 }, { "epoch": 2.46, "grad_norm": 22.162360533755802, "learning_rate": 1.6620132023193968e-06, "loss": 0.4485, "step": 15738 }, { "epoch": 2.46, "grad_norm": 17.300505327792237, "learning_rate": 1.661082312931085e-06, "loss": 0.4433, "step": 15739 }, { "epoch": 2.46, "grad_norm": 16.127373860708108, "learning_rate": 1.6601516606967127e-06, "loss": 0.4189, "step": 15740 }, { "epoch": 2.46, "grad_norm": 23.10631513778579, "learning_rate": 1.6592212456427492e-06, "loss": 0.4226, "step": 15741 }, { "epoch": 2.46, "grad_norm": 17.828958180957322, "learning_rate": 1.6582910677956532e-06, "loss": 0.4769, "step": 15742 }, { "epoch": 2.46, "grad_norm": 18.875245256520007, "learning_rate": 1.6573611271818812e-06, "loss": 0.4423, "step": 15743 }, { "epoch": 2.46, "grad_norm": 15.888399507501056, "learning_rate": 1.6564314238278755e-06, "loss": 0.4677, "step": 15744 }, { "epoch": 2.46, "grad_norm": 15.235006550144908, "learning_rate": 1.6555019577600806e-06, "loss": 0.4204, "step": 15745 }, { "epoch": 2.46, "grad_norm": 26.225437003227533, "learning_rate": 1.6545727290049307e-06, "loss": 0.5065, "step": 15746 }, { "epoch": 2.46, "grad_norm": 27.23459102798812, "learning_rate": 1.653643737588847e-06, "loss": 0.4503, "step": 15747 }, { "epoch": 2.46, "grad_norm": 15.500837972027806, "learning_rate": 1.652714983538256e-06, "loss": 0.402, "step": 15748 }, { "epoch": 2.46, "grad_norm": 20.26094469070454, "learning_rate": 1.6517864668795658e-06, "loss": 0.3832, "step": 15749 }, { "epoch": 2.46, "grad_norm": 21.15137077129673, "learning_rate": 1.650858187639185e-06, "loss": 0.457, "step": 15750 }, { "epoch": 2.46, "grad_norm": 15.12292444306921, "learning_rate": 1.6499301458435146e-06, "loss": 0.4477, "step": 15751 }, { "epoch": 2.46, "grad_norm": 34.511774226922874, "learning_rate": 1.6490023415189472e-06, "loss": 0.5406, "step": 15752 }, { "epoch": 2.46, "grad_norm": 18.108051259073665, "learning_rate": 1.6480747746918657e-06, "loss": 0.3869, "step": 15753 }, { "epoch": 2.46, "grad_norm": 14.535380076997066, "learning_rate": 1.6471474453886504e-06, "loss": 0.4209, "step": 15754 }, { "epoch": 2.46, "grad_norm": 21.087204325168425, "learning_rate": 1.6462203536356769e-06, "loss": 0.406, "step": 15755 }, { "epoch": 2.46, "grad_norm": 23.541237341642887, "learning_rate": 1.645293499459314e-06, "loss": 0.4353, "step": 15756 }, { "epoch": 2.46, "grad_norm": 21.850660523697552, "learning_rate": 1.6443668828859126e-06, "loss": 0.5183, "step": 15757 }, { "epoch": 2.46, "grad_norm": 20.681784007261435, "learning_rate": 1.6434405039418277e-06, "loss": 0.4485, "step": 15758 }, { "epoch": 2.46, "grad_norm": 24.705386693449253, "learning_rate": 1.642514362653408e-06, "loss": 0.4631, "step": 15759 }, { "epoch": 2.46, "grad_norm": 16.775212829585673, "learning_rate": 1.6415884590469922e-06, "loss": 0.437, "step": 15760 }, { "epoch": 2.46, "grad_norm": 22.09529900212891, "learning_rate": 1.6406627931489116e-06, "loss": 0.4687, "step": 15761 }, { "epoch": 2.46, "grad_norm": 22.013072223904953, "learning_rate": 1.6397373649854874e-06, "loss": 0.4063, "step": 15762 }, { "epoch": 2.46, "grad_norm": 22.860996041540066, "learning_rate": 1.6388121745830433e-06, "loss": 0.4664, "step": 15763 }, { "epoch": 2.46, "grad_norm": 26.62977977728416, "learning_rate": 1.637887221967892e-06, "loss": 0.4672, "step": 15764 }, { "epoch": 2.46, "grad_norm": 30.710837432205327, "learning_rate": 1.6369625071663365e-06, "loss": 0.4251, "step": 15765 }, { "epoch": 2.46, "grad_norm": 15.91126554833314, "learning_rate": 1.636038030204672e-06, "loss": 0.3985, "step": 15766 }, { "epoch": 2.46, "grad_norm": 25.44609519397724, "learning_rate": 1.6351137911091942e-06, "loss": 0.4975, "step": 15767 }, { "epoch": 2.46, "grad_norm": 26.950533963484457, "learning_rate": 1.6341897899061887e-06, "loss": 0.4428, "step": 15768 }, { "epoch": 2.46, "grad_norm": 17.44905099303504, "learning_rate": 1.6332660266219292e-06, "loss": 0.4662, "step": 15769 }, { "epoch": 2.46, "grad_norm": 23.980274385446613, "learning_rate": 1.632342501282692e-06, "loss": 0.5126, "step": 15770 }, { "epoch": 2.46, "grad_norm": 23.486799757696925, "learning_rate": 1.6314192139147366e-06, "loss": 0.4376, "step": 15771 }, { "epoch": 2.46, "grad_norm": 22.123059281939863, "learning_rate": 1.630496164544323e-06, "loss": 0.4476, "step": 15772 }, { "epoch": 2.46, "grad_norm": 21.677437329063675, "learning_rate": 1.6295733531977054e-06, "loss": 0.4235, "step": 15773 }, { "epoch": 2.46, "grad_norm": 23.671952821584764, "learning_rate": 1.628650779901122e-06, "loss": 0.4668, "step": 15774 }, { "epoch": 2.46, "grad_norm": 21.2288892900679, "learning_rate": 1.6277284446808162e-06, "loss": 0.3926, "step": 15775 }, { "epoch": 2.46, "grad_norm": 28.175068523002945, "learning_rate": 1.6268063475630136e-06, "loss": 0.4747, "step": 15776 }, { "epoch": 2.46, "grad_norm": 21.578656915825086, "learning_rate": 1.6258844885739434e-06, "loss": 0.4324, "step": 15777 }, { "epoch": 2.46, "grad_norm": 22.973499852119023, "learning_rate": 1.6249628677398155e-06, "loss": 0.492, "step": 15778 }, { "epoch": 2.46, "grad_norm": 20.485454415223497, "learning_rate": 1.6240414850868457e-06, "loss": 0.4259, "step": 15779 }, { "epoch": 2.46, "grad_norm": 17.160058083135137, "learning_rate": 1.6231203406412377e-06, "loss": 0.3976, "step": 15780 }, { "epoch": 2.47, "grad_norm": 32.365971471735506, "learning_rate": 1.622199434429188e-06, "loss": 0.4728, "step": 15781 }, { "epoch": 2.47, "grad_norm": 13.908113643517492, "learning_rate": 1.6212787664768826e-06, "loss": 0.3815, "step": 15782 }, { "epoch": 2.47, "grad_norm": 19.140819016076644, "learning_rate": 1.620358336810507e-06, "loss": 0.4187, "step": 15783 }, { "epoch": 2.47, "grad_norm": 22.876185456811204, "learning_rate": 1.6194381454562392e-06, "loss": 0.4372, "step": 15784 }, { "epoch": 2.47, "grad_norm": 19.691695485320647, "learning_rate": 1.6185181924402504e-06, "loss": 0.4669, "step": 15785 }, { "epoch": 2.47, "grad_norm": 16.645488250443414, "learning_rate": 1.617598477788701e-06, "loss": 0.3984, "step": 15786 }, { "epoch": 2.47, "grad_norm": 23.609772265441492, "learning_rate": 1.6166790015277446e-06, "loss": 0.4795, "step": 15787 }, { "epoch": 2.47, "grad_norm": 26.187348424522906, "learning_rate": 1.6157597636835332e-06, "loss": 0.5007, "step": 15788 }, { "epoch": 2.47, "grad_norm": 16.776798204782633, "learning_rate": 1.6148407642822128e-06, "loss": 0.4749, "step": 15789 }, { "epoch": 2.47, "grad_norm": 34.4999099260343, "learning_rate": 1.6139220033499148e-06, "loss": 0.4462, "step": 15790 }, { "epoch": 2.47, "grad_norm": 25.093255654545267, "learning_rate": 1.6130034809127682e-06, "loss": 0.4726, "step": 15791 }, { "epoch": 2.47, "grad_norm": 20.5607394258983, "learning_rate": 1.6120851969968954e-06, "loss": 0.4774, "step": 15792 }, { "epoch": 2.47, "grad_norm": 19.460412827336775, "learning_rate": 1.6111671516284155e-06, "loss": 0.5472, "step": 15793 }, { "epoch": 2.47, "grad_norm": 18.65970534525912, "learning_rate": 1.6102493448334312e-06, "loss": 0.4788, "step": 15794 }, { "epoch": 2.47, "grad_norm": 17.722224742957692, "learning_rate": 1.6093317766380511e-06, "loss": 0.4785, "step": 15795 }, { "epoch": 2.47, "grad_norm": 27.16006378801096, "learning_rate": 1.608414447068365e-06, "loss": 0.4683, "step": 15796 }, { "epoch": 2.47, "grad_norm": 17.990680896939896, "learning_rate": 1.6074973561504614e-06, "loss": 0.4504, "step": 15797 }, { "epoch": 2.47, "grad_norm": 15.260492871672495, "learning_rate": 1.6065805039104275e-06, "loss": 0.4479, "step": 15798 }, { "epoch": 2.47, "grad_norm": 19.023663228475808, "learning_rate": 1.6056638903743306e-06, "loss": 0.4239, "step": 15799 }, { "epoch": 2.47, "grad_norm": 16.1769134753493, "learning_rate": 1.6047475155682445e-06, "loss": 0.4174, "step": 15800 }, { "epoch": 2.47, "grad_norm": 22.46304274868134, "learning_rate": 1.603831379518227e-06, "loss": 0.352, "step": 15801 }, { "epoch": 2.47, "grad_norm": 14.308352918001425, "learning_rate": 1.6029154822503346e-06, "loss": 0.4251, "step": 15802 }, { "epoch": 2.47, "grad_norm": 18.25301756145544, "learning_rate": 1.601999823790611e-06, "loss": 0.4419, "step": 15803 }, { "epoch": 2.47, "grad_norm": 22.106829267536867, "learning_rate": 1.601084404165103e-06, "loss": 0.4498, "step": 15804 }, { "epoch": 2.47, "grad_norm": 18.17004120742532, "learning_rate": 1.600169223399839e-06, "loss": 0.4301, "step": 15805 }, { "epoch": 2.47, "grad_norm": 20.01652959883348, "learning_rate": 1.59925428152085e-06, "loss": 0.474, "step": 15806 }, { "epoch": 2.47, "grad_norm": 16.756729835713678, "learning_rate": 1.5983395785541533e-06, "loss": 0.3706, "step": 15807 }, { "epoch": 2.47, "grad_norm": 27.04470500967914, "learning_rate": 1.5974251145257635e-06, "loss": 0.4483, "step": 15808 }, { "epoch": 2.47, "grad_norm": 13.001271127260578, "learning_rate": 1.596510889461691e-06, "loss": 0.4275, "step": 15809 }, { "epoch": 2.47, "grad_norm": 24.1429701360449, "learning_rate": 1.5955969033879304e-06, "loss": 0.4208, "step": 15810 }, { "epoch": 2.47, "grad_norm": 19.11069303382256, "learning_rate": 1.59468315633048e-06, "loss": 0.435, "step": 15811 }, { "epoch": 2.47, "grad_norm": 19.037680370690516, "learning_rate": 1.5937696483153208e-06, "loss": 0.4321, "step": 15812 }, { "epoch": 2.47, "grad_norm": 22.222345161909207, "learning_rate": 1.592856379368436e-06, "loss": 0.45, "step": 15813 }, { "epoch": 2.47, "grad_norm": 22.399253440500903, "learning_rate": 1.5919433495157998e-06, "loss": 0.4899, "step": 15814 }, { "epoch": 2.47, "grad_norm": 20.246569874262015, "learning_rate": 1.5910305587833763e-06, "loss": 0.4882, "step": 15815 }, { "epoch": 2.47, "grad_norm": 22.49344197240153, "learning_rate": 1.5901180071971224e-06, "loss": 0.4686, "step": 15816 }, { "epoch": 2.47, "grad_norm": 19.122848832315956, "learning_rate": 1.5892056947829914e-06, "loss": 0.3949, "step": 15817 }, { "epoch": 2.47, "grad_norm": 16.435141032438963, "learning_rate": 1.588293621566932e-06, "loss": 0.4591, "step": 15818 }, { "epoch": 2.47, "grad_norm": 16.318812911775456, "learning_rate": 1.5873817875748854e-06, "loss": 0.4251, "step": 15819 }, { "epoch": 2.47, "grad_norm": 16.49486935927179, "learning_rate": 1.586470192832774e-06, "loss": 0.4878, "step": 15820 }, { "epoch": 2.47, "grad_norm": 26.086181328734426, "learning_rate": 1.5855588373665298e-06, "loss": 0.4475, "step": 15821 }, { "epoch": 2.47, "grad_norm": 21.735985332209054, "learning_rate": 1.5846477212020695e-06, "loss": 0.5028, "step": 15822 }, { "epoch": 2.47, "grad_norm": 15.885797790478811, "learning_rate": 1.5837368443653067e-06, "loss": 0.4181, "step": 15823 }, { "epoch": 2.47, "grad_norm": 30.79569302694979, "learning_rate": 1.5828262068821453e-06, "loss": 0.4493, "step": 15824 }, { "epoch": 2.47, "grad_norm": 15.925579086135874, "learning_rate": 1.5819158087784802e-06, "loss": 0.4112, "step": 15825 }, { "epoch": 2.47, "grad_norm": 20.919531799959184, "learning_rate": 1.5810056500802063e-06, "loss": 0.4401, "step": 15826 }, { "epoch": 2.47, "grad_norm": 25.579587426495788, "learning_rate": 1.580095730813208e-06, "loss": 0.4446, "step": 15827 }, { "epoch": 2.47, "grad_norm": 20.015958006751642, "learning_rate": 1.5791860510033584e-06, "loss": 0.4614, "step": 15828 }, { "epoch": 2.47, "grad_norm": 20.109761882756516, "learning_rate": 1.5782766106765357e-06, "loss": 0.4462, "step": 15829 }, { "epoch": 2.47, "grad_norm": 21.728774618941227, "learning_rate": 1.5773674098585955e-06, "loss": 0.4592, "step": 15830 }, { "epoch": 2.47, "grad_norm": 15.566478365320659, "learning_rate": 1.5764584485754031e-06, "loss": 0.4157, "step": 15831 }, { "epoch": 2.47, "grad_norm": 19.097230370856966, "learning_rate": 1.5755497268528008e-06, "loss": 0.4192, "step": 15832 }, { "epoch": 2.47, "grad_norm": 16.77031149116484, "learning_rate": 1.5746412447166371e-06, "loss": 0.4167, "step": 15833 }, { "epoch": 2.47, "grad_norm": 20.808085527581685, "learning_rate": 1.57373300219275e-06, "loss": 0.4761, "step": 15834 }, { "epoch": 2.47, "grad_norm": 27.709133620494452, "learning_rate": 1.5728249993069645e-06, "loss": 0.4376, "step": 15835 }, { "epoch": 2.47, "grad_norm": 24.312625844232592, "learning_rate": 1.5719172360851088e-06, "loss": 0.4547, "step": 15836 }, { "epoch": 2.47, "grad_norm": 28.498933016103503, "learning_rate": 1.5710097125529943e-06, "loss": 0.4462, "step": 15837 }, { "epoch": 2.47, "grad_norm": 27.262975946077834, "learning_rate": 1.570102428736432e-06, "loss": 0.4534, "step": 15838 }, { "epoch": 2.47, "grad_norm": 20.098926065143413, "learning_rate": 1.5691953846612284e-06, "loss": 0.4624, "step": 15839 }, { "epoch": 2.47, "grad_norm": 16.155902649449267, "learning_rate": 1.5682885803531767e-06, "loss": 0.3913, "step": 15840 }, { "epoch": 2.47, "grad_norm": 21.10024052475499, "learning_rate": 1.5673820158380615e-06, "loss": 0.4716, "step": 15841 }, { "epoch": 2.47, "grad_norm": 27.76756608335823, "learning_rate": 1.56647569114167e-06, "loss": 0.4665, "step": 15842 }, { "epoch": 2.47, "grad_norm": 23.606101682044013, "learning_rate": 1.5655696062897763e-06, "loss": 0.422, "step": 15843 }, { "epoch": 2.47, "grad_norm": 19.78048032088841, "learning_rate": 1.5646637613081529e-06, "loss": 0.4435, "step": 15844 }, { "epoch": 2.48, "grad_norm": 19.583449949431344, "learning_rate": 1.5637581562225535e-06, "loss": 0.4397, "step": 15845 }, { "epoch": 2.48, "grad_norm": 19.0700346068123, "learning_rate": 1.5628527910587376e-06, "loss": 0.4828, "step": 15846 }, { "epoch": 2.48, "grad_norm": 16.261357479128073, "learning_rate": 1.5619476658424516e-06, "loss": 0.4376, "step": 15847 }, { "epoch": 2.48, "grad_norm": 20.56570500358424, "learning_rate": 1.5610427805994411e-06, "loss": 0.4662, "step": 15848 }, { "epoch": 2.48, "grad_norm": 25.245674551249213, "learning_rate": 1.5601381353554369e-06, "loss": 0.3799, "step": 15849 }, { "epoch": 2.48, "grad_norm": 17.070036385265002, "learning_rate": 1.5592337301361648e-06, "loss": 0.4301, "step": 15850 }, { "epoch": 2.48, "grad_norm": 20.830985084672424, "learning_rate": 1.5583295649673468e-06, "loss": 0.4381, "step": 15851 }, { "epoch": 2.48, "grad_norm": 18.979017601429273, "learning_rate": 1.5574256398747011e-06, "loss": 0.3895, "step": 15852 }, { "epoch": 2.48, "grad_norm": 17.893138392445845, "learning_rate": 1.5565219548839317e-06, "loss": 0.3965, "step": 15853 }, { "epoch": 2.48, "grad_norm": 28.620914134829555, "learning_rate": 1.5556185100207356e-06, "loss": 0.512, "step": 15854 }, { "epoch": 2.48, "grad_norm": 17.01075433303274, "learning_rate": 1.5547153053108088e-06, "loss": 0.3863, "step": 15855 }, { "epoch": 2.48, "grad_norm": 27.012544056937227, "learning_rate": 1.5538123407798422e-06, "loss": 0.4644, "step": 15856 }, { "epoch": 2.48, "grad_norm": 24.287770911495855, "learning_rate": 1.5529096164535084e-06, "loss": 0.4781, "step": 15857 }, { "epoch": 2.48, "grad_norm": 19.251015557401683, "learning_rate": 1.552007132357486e-06, "loss": 0.3952, "step": 15858 }, { "epoch": 2.48, "grad_norm": 21.990501520770504, "learning_rate": 1.5511048885174361e-06, "loss": 0.4279, "step": 15859 }, { "epoch": 2.48, "grad_norm": 20.696254944934083, "learning_rate": 1.5502028849590212e-06, "loss": 0.5025, "step": 15860 }, { "epoch": 2.48, "grad_norm": 24.408892287630163, "learning_rate": 1.5493011217078958e-06, "loss": 0.4544, "step": 15861 }, { "epoch": 2.48, "grad_norm": 21.323693465516758, "learning_rate": 1.5483995987897006e-06, "loss": 0.4783, "step": 15862 }, { "epoch": 2.48, "grad_norm": 25.099920832368543, "learning_rate": 1.5474983162300794e-06, "loss": 0.3937, "step": 15863 }, { "epoch": 2.48, "grad_norm": 27.983222073991183, "learning_rate": 1.5465972740546586e-06, "loss": 0.4702, "step": 15864 }, { "epoch": 2.48, "grad_norm": 24.119569879496662, "learning_rate": 1.5456964722890688e-06, "loss": 0.4174, "step": 15865 }, { "epoch": 2.48, "grad_norm": 28.876207339861036, "learning_rate": 1.5447959109589228e-06, "loss": 0.5151, "step": 15866 }, { "epoch": 2.48, "grad_norm": 26.473184638557015, "learning_rate": 1.5438955900898355e-06, "loss": 0.4941, "step": 15867 }, { "epoch": 2.48, "grad_norm": 28.065103913462607, "learning_rate": 1.5429955097074133e-06, "loss": 0.4472, "step": 15868 }, { "epoch": 2.48, "grad_norm": 23.769763028245873, "learning_rate": 1.5420956698372513e-06, "loss": 0.4528, "step": 15869 }, { "epoch": 2.48, "grad_norm": 27.888035565871476, "learning_rate": 1.5411960705049389e-06, "loss": 0.4603, "step": 15870 }, { "epoch": 2.48, "grad_norm": 33.09202352888714, "learning_rate": 1.5402967117360622e-06, "loss": 0.4594, "step": 15871 }, { "epoch": 2.48, "grad_norm": 31.206891998675562, "learning_rate": 1.5393975935561977e-06, "loss": 0.4702, "step": 15872 }, { "epoch": 2.48, "grad_norm": 22.277404367015603, "learning_rate": 1.538498715990918e-06, "loss": 0.4106, "step": 15873 }, { "epoch": 2.48, "grad_norm": 17.20349427934717, "learning_rate": 1.5376000790657864e-06, "loss": 0.4515, "step": 15874 }, { "epoch": 2.48, "grad_norm": 19.52893499769126, "learning_rate": 1.5367016828063553e-06, "loss": 0.4192, "step": 15875 }, { "epoch": 2.48, "grad_norm": 22.38506387302692, "learning_rate": 1.5358035272381765e-06, "loss": 0.4357, "step": 15876 }, { "epoch": 2.48, "grad_norm": 33.91025539726895, "learning_rate": 1.5349056123867977e-06, "loss": 0.4813, "step": 15877 }, { "epoch": 2.48, "grad_norm": 25.84013416937306, "learning_rate": 1.5340079382777506e-06, "loss": 0.5086, "step": 15878 }, { "epoch": 2.48, "grad_norm": 15.89609320740819, "learning_rate": 1.5331105049365624e-06, "loss": 0.4099, "step": 15879 }, { "epoch": 2.48, "grad_norm": 13.336404787758743, "learning_rate": 1.5322133123887595e-06, "loss": 0.4687, "step": 15880 }, { "epoch": 2.48, "grad_norm": 19.779022821362457, "learning_rate": 1.5313163606598591e-06, "loss": 0.4265, "step": 15881 }, { "epoch": 2.48, "grad_norm": 22.83796914897114, "learning_rate": 1.5304196497753642e-06, "loss": 0.4438, "step": 15882 }, { "epoch": 2.48, "grad_norm": 19.019366033482452, "learning_rate": 1.5295231797607824e-06, "loss": 0.4395, "step": 15883 }, { "epoch": 2.48, "grad_norm": 25.441464013342028, "learning_rate": 1.528626950641604e-06, "loss": 0.4454, "step": 15884 }, { "epoch": 2.48, "grad_norm": 22.39590464994623, "learning_rate": 1.527730962443319e-06, "loss": 0.5054, "step": 15885 }, { "epoch": 2.48, "grad_norm": 16.341402663888733, "learning_rate": 1.5268352151914124e-06, "loss": 0.4597, "step": 15886 }, { "epoch": 2.48, "grad_norm": 19.32514305865262, "learning_rate": 1.5259397089113525e-06, "loss": 0.4985, "step": 15887 }, { "epoch": 2.48, "grad_norm": 19.520970019557804, "learning_rate": 1.5250444436286138e-06, "loss": 0.423, "step": 15888 }, { "epoch": 2.48, "grad_norm": 22.786852633997505, "learning_rate": 1.52414941936865e-06, "loss": 0.4889, "step": 15889 }, { "epoch": 2.48, "grad_norm": 21.687739286765403, "learning_rate": 1.5232546361569212e-06, "loss": 0.448, "step": 15890 }, { "epoch": 2.48, "grad_norm": 24.65784604741263, "learning_rate": 1.5223600940188688e-06, "loss": 0.5193, "step": 15891 }, { "epoch": 2.48, "grad_norm": 25.98049628967927, "learning_rate": 1.521465792979937e-06, "loss": 0.4497, "step": 15892 }, { "epoch": 2.48, "grad_norm": 21.327504345313567, "learning_rate": 1.52057173306556e-06, "loss": 0.4631, "step": 15893 }, { "epoch": 2.48, "grad_norm": 16.638703404546824, "learning_rate": 1.5196779143011619e-06, "loss": 0.4216, "step": 15894 }, { "epoch": 2.48, "grad_norm": 27.099849734228446, "learning_rate": 1.5187843367121623e-06, "loss": 0.4455, "step": 15895 }, { "epoch": 2.48, "grad_norm": 26.04739663689291, "learning_rate": 1.5178910003239732e-06, "loss": 0.4662, "step": 15896 }, { "epoch": 2.48, "grad_norm": 19.37733372815419, "learning_rate": 1.5169979051620033e-06, "loss": 0.4447, "step": 15897 }, { "epoch": 2.48, "grad_norm": 14.22351636174334, "learning_rate": 1.5161050512516528e-06, "loss": 0.4236, "step": 15898 }, { "epoch": 2.48, "grad_norm": 19.24288336920094, "learning_rate": 1.515212438618312e-06, "loss": 0.4299, "step": 15899 }, { "epoch": 2.48, "grad_norm": 20.47311512797288, "learning_rate": 1.5143200672873626e-06, "loss": 0.4924, "step": 15900 }, { "epoch": 2.48, "grad_norm": 24.937756796998947, "learning_rate": 1.5134279372841876e-06, "loss": 0.4817, "step": 15901 }, { "epoch": 2.48, "grad_norm": 24.87034107733147, "learning_rate": 1.512536048634159e-06, "loss": 0.4097, "step": 15902 }, { "epoch": 2.48, "grad_norm": 23.113704871569478, "learning_rate": 1.5116444013626407e-06, "loss": 0.4651, "step": 15903 }, { "epoch": 2.48, "grad_norm": 20.308026687591433, "learning_rate": 1.5107529954949873e-06, "loss": 0.4509, "step": 15904 }, { "epoch": 2.48, "grad_norm": 23.147841699189456, "learning_rate": 1.5098618310565527e-06, "loss": 0.434, "step": 15905 }, { "epoch": 2.48, "grad_norm": 22.195651860447384, "learning_rate": 1.5089709080726811e-06, "loss": 0.4381, "step": 15906 }, { "epoch": 2.48, "grad_norm": 23.298331607871916, "learning_rate": 1.5080802265687145e-06, "loss": 0.4811, "step": 15907 }, { "epoch": 2.48, "grad_norm": 16.20899329786817, "learning_rate": 1.5071897865699748e-06, "loss": 0.4278, "step": 15908 }, { "epoch": 2.49, "grad_norm": 11.29002141344829, "learning_rate": 1.5062995881017882e-06, "loss": 0.4126, "step": 15909 }, { "epoch": 2.49, "grad_norm": 31.470961504396215, "learning_rate": 1.5054096311894727e-06, "loss": 0.4309, "step": 15910 }, { "epoch": 2.49, "grad_norm": 21.255734776730474, "learning_rate": 1.5045199158583411e-06, "loss": 0.4699, "step": 15911 }, { "epoch": 2.49, "grad_norm": 24.366989873052283, "learning_rate": 1.5036304421336933e-06, "loss": 0.5092, "step": 15912 }, { "epoch": 2.49, "grad_norm": 25.176274394080387, "learning_rate": 1.5027412100408234e-06, "loss": 0.4551, "step": 15913 }, { "epoch": 2.49, "grad_norm": 25.33427567491345, "learning_rate": 1.5018522196050223e-06, "loss": 0.4993, "step": 15914 }, { "epoch": 2.49, "grad_norm": 22.33923602624124, "learning_rate": 1.5009634708515753e-06, "loss": 0.3885, "step": 15915 }, { "epoch": 2.49, "grad_norm": 23.750658122782152, "learning_rate": 1.5000749638057521e-06, "loss": 0.4336, "step": 15916 }, { "epoch": 2.49, "grad_norm": 19.47256696674842, "learning_rate": 1.4991866984928283e-06, "loss": 0.3814, "step": 15917 }, { "epoch": 2.49, "grad_norm": 16.77747302466293, "learning_rate": 1.498298674938059e-06, "loss": 0.3898, "step": 15918 }, { "epoch": 2.49, "grad_norm": 24.55067618627532, "learning_rate": 1.4974108931667042e-06, "loss": 0.5037, "step": 15919 }, { "epoch": 2.49, "grad_norm": 26.458333694590667, "learning_rate": 1.496523353204008e-06, "loss": 0.4397, "step": 15920 }, { "epoch": 2.49, "grad_norm": 20.780742321410045, "learning_rate": 1.4956360550752125e-06, "loss": 0.3747, "step": 15921 }, { "epoch": 2.49, "grad_norm": 20.674386250382053, "learning_rate": 1.4947489988055552e-06, "loss": 0.4839, "step": 15922 }, { "epoch": 2.49, "grad_norm": 20.032490595906165, "learning_rate": 1.4938621844202595e-06, "loss": 0.4719, "step": 15923 }, { "epoch": 2.49, "grad_norm": 21.980267139579293, "learning_rate": 1.4929756119445481e-06, "loss": 0.4887, "step": 15924 }, { "epoch": 2.49, "grad_norm": 3.9562957245774237, "learning_rate": 1.4920892814036324e-06, "loss": 0.4581, "step": 15925 }, { "epoch": 2.49, "grad_norm": 20.830082211894204, "learning_rate": 1.4912031928227211e-06, "loss": 0.434, "step": 15926 }, { "epoch": 2.49, "grad_norm": 27.20785756003081, "learning_rate": 1.4903173462270148e-06, "loss": 0.5017, "step": 15927 }, { "epoch": 2.49, "grad_norm": 19.31120153494008, "learning_rate": 1.489431741641706e-06, "loss": 0.4552, "step": 15928 }, { "epoch": 2.49, "grad_norm": 16.68809942622622, "learning_rate": 1.4885463790919774e-06, "loss": 0.386, "step": 15929 }, { "epoch": 2.49, "grad_norm": 22.93904741690915, "learning_rate": 1.4876612586030104e-06, "loss": 0.4694, "step": 15930 }, { "epoch": 2.49, "grad_norm": 18.773813326665838, "learning_rate": 1.4867763801999768e-06, "loss": 0.513, "step": 15931 }, { "epoch": 2.49, "grad_norm": 25.49439518002952, "learning_rate": 1.485891743908049e-06, "loss": 0.4459, "step": 15932 }, { "epoch": 2.49, "grad_norm": 21.820630716911978, "learning_rate": 1.4850073497523743e-06, "loss": 0.4739, "step": 15933 }, { "epoch": 2.49, "grad_norm": 27.739568880549893, "learning_rate": 1.4841231977581094e-06, "loss": 0.4289, "step": 15934 }, { "epoch": 2.49, "grad_norm": 22.17817060679014, "learning_rate": 1.4832392879503986e-06, "loss": 0.4394, "step": 15935 }, { "epoch": 2.49, "grad_norm": 35.1045746040507, "learning_rate": 1.4823556203543833e-06, "loss": 0.4507, "step": 15936 }, { "epoch": 2.49, "grad_norm": 20.855989124231517, "learning_rate": 1.4814721949951916e-06, "loss": 0.4217, "step": 15937 }, { "epoch": 2.49, "grad_norm": 19.25297416587063, "learning_rate": 1.4805890118979448e-06, "loss": 0.3914, "step": 15938 }, { "epoch": 2.49, "grad_norm": 25.95979217932501, "learning_rate": 1.479706071087762e-06, "loss": 0.469, "step": 15939 }, { "epoch": 2.49, "grad_norm": 20.43190194668352, "learning_rate": 1.4788233725897583e-06, "loss": 0.4084, "step": 15940 }, { "epoch": 2.49, "grad_norm": 18.498548627155984, "learning_rate": 1.4779409164290292e-06, "loss": 0.4377, "step": 15941 }, { "epoch": 2.49, "grad_norm": 28.901810638691483, "learning_rate": 1.4770587026306782e-06, "loss": 0.4569, "step": 15942 }, { "epoch": 2.49, "grad_norm": 21.361813167127167, "learning_rate": 1.4761767312197906e-06, "loss": 0.3807, "step": 15943 }, { "epoch": 2.49, "grad_norm": 25.39141460423376, "learning_rate": 1.4752950022214519e-06, "loss": 0.4308, "step": 15944 }, { "epoch": 2.49, "grad_norm": 18.600865196398455, "learning_rate": 1.4744135156607332e-06, "loss": 0.4703, "step": 15945 }, { "epoch": 2.49, "grad_norm": 16.636290009763123, "learning_rate": 1.4735322715627075e-06, "loss": 0.3893, "step": 15946 }, { "epoch": 2.49, "grad_norm": 24.618136628260608, "learning_rate": 1.4726512699524399e-06, "loss": 0.4129, "step": 15947 }, { "epoch": 2.49, "grad_norm": 35.03591338670897, "learning_rate": 1.4717705108549773e-06, "loss": 0.4514, "step": 15948 }, { "epoch": 2.49, "grad_norm": 29.705063591574238, "learning_rate": 1.4708899942953757e-06, "loss": 0.5014, "step": 15949 }, { "epoch": 2.49, "grad_norm": 31.817246844554766, "learning_rate": 1.4700097202986719e-06, "loss": 0.4789, "step": 15950 }, { "epoch": 2.49, "grad_norm": 17.620312435153192, "learning_rate": 1.4691296888899031e-06, "loss": 0.4129, "step": 15951 }, { "epoch": 2.49, "grad_norm": 24.37884239331442, "learning_rate": 1.468249900094093e-06, "loss": 0.5041, "step": 15952 }, { "epoch": 2.49, "grad_norm": 19.99973168714669, "learning_rate": 1.4673703539362683e-06, "loss": 0.439, "step": 15953 }, { "epoch": 2.49, "grad_norm": 23.148361787855663, "learning_rate": 1.4664910504414355e-06, "loss": 0.4214, "step": 15954 }, { "epoch": 2.49, "grad_norm": 19.214204773318375, "learning_rate": 1.4656119896346055e-06, "loss": 0.427, "step": 15955 }, { "epoch": 2.49, "grad_norm": 15.789550828908649, "learning_rate": 1.4647331715407809e-06, "loss": 0.4045, "step": 15956 }, { "epoch": 2.49, "grad_norm": 20.866609413254544, "learning_rate": 1.4638545961849525e-06, "loss": 0.4367, "step": 15957 }, { "epoch": 2.49, "grad_norm": 34.8019548818339, "learning_rate": 1.462976263592103e-06, "loss": 0.4679, "step": 15958 }, { "epoch": 2.49, "grad_norm": 20.96090897683653, "learning_rate": 1.4620981737872142e-06, "loss": 0.48, "step": 15959 }, { "epoch": 2.49, "grad_norm": 23.801750441507693, "learning_rate": 1.4612203267952584e-06, "loss": 0.4926, "step": 15960 }, { "epoch": 2.49, "grad_norm": 19.57347303746456, "learning_rate": 1.4603427226412048e-06, "loss": 0.4394, "step": 15961 }, { "epoch": 2.49, "grad_norm": 32.975855135385835, "learning_rate": 1.4594653613500086e-06, "loss": 0.4786, "step": 15962 }, { "epoch": 2.49, "grad_norm": 30.657353516450904, "learning_rate": 1.458588242946618e-06, "loss": 0.4305, "step": 15963 }, { "epoch": 2.49, "grad_norm": 17.54057945369242, "learning_rate": 1.457711367455983e-06, "loss": 0.4019, "step": 15964 }, { "epoch": 2.49, "grad_norm": 20.056506158792367, "learning_rate": 1.4568347349030398e-06, "loss": 0.4478, "step": 15965 }, { "epoch": 2.49, "grad_norm": 20.068239892757056, "learning_rate": 1.4559583453127213e-06, "loss": 0.4297, "step": 15966 }, { "epoch": 2.49, "grad_norm": 23.80717821489896, "learning_rate": 1.455082198709945e-06, "loss": 0.4576, "step": 15967 }, { "epoch": 2.49, "grad_norm": 14.432028433405515, "learning_rate": 1.4542062951196335e-06, "loss": 0.4196, "step": 15968 }, { "epoch": 2.49, "grad_norm": 20.264111352038498, "learning_rate": 1.4533306345666954e-06, "loss": 0.4649, "step": 15969 }, { "epoch": 2.49, "grad_norm": 21.010694326691546, "learning_rate": 1.4524552170760374e-06, "loss": 0.456, "step": 15970 }, { "epoch": 2.49, "grad_norm": 27.893013837435078, "learning_rate": 1.4515800426725524e-06, "loss": 0.4185, "step": 15971 }, { "epoch": 2.49, "grad_norm": 19.58392009760701, "learning_rate": 1.4507051113811277e-06, "loss": 0.45, "step": 15972 }, { "epoch": 2.5, "grad_norm": 24.573590410816404, "learning_rate": 1.449830423226649e-06, "loss": 0.4549, "step": 15973 }, { "epoch": 2.5, "grad_norm": 27.79215989695097, "learning_rate": 1.4489559782339935e-06, "loss": 0.4456, "step": 15974 }, { "epoch": 2.5, "grad_norm": 19.806146162440406, "learning_rate": 1.4480817764280254e-06, "loss": 0.4965, "step": 15975 }, { "epoch": 2.5, "grad_norm": 30.914455157782747, "learning_rate": 1.4472078178336102e-06, "loss": 0.4982, "step": 15976 }, { "epoch": 2.5, "grad_norm": 16.522800470251514, "learning_rate": 1.4463341024756006e-06, "loss": 0.4306, "step": 15977 }, { "epoch": 2.5, "grad_norm": 24.581338942711422, "learning_rate": 1.4454606303788466e-06, "loss": 0.4073, "step": 15978 }, { "epoch": 2.5, "grad_norm": 18.85094390578264, "learning_rate": 1.4445874015681872e-06, "loss": 0.4205, "step": 15979 }, { "epoch": 2.5, "grad_norm": 20.8059160010933, "learning_rate": 1.4437144160684558e-06, "loss": 0.4377, "step": 15980 }, { "epoch": 2.5, "grad_norm": 16.376876506956027, "learning_rate": 1.4428416739044836e-06, "loss": 0.4604, "step": 15981 }, { "epoch": 2.5, "grad_norm": 20.56499205433905, "learning_rate": 1.441969175101089e-06, "loss": 0.4461, "step": 15982 }, { "epoch": 2.5, "grad_norm": 17.94847217455842, "learning_rate": 1.4410969196830827e-06, "loss": 0.4406, "step": 15983 }, { "epoch": 2.5, "grad_norm": 38.01468302307978, "learning_rate": 1.4402249076752716e-06, "loss": 0.4775, "step": 15984 }, { "epoch": 2.5, "grad_norm": 29.2364358975806, "learning_rate": 1.4393531391024585e-06, "loss": 0.4504, "step": 15985 }, { "epoch": 2.5, "grad_norm": 13.376873624890948, "learning_rate": 1.4384816139894365e-06, "loss": 0.4475, "step": 15986 }, { "epoch": 2.5, "grad_norm": 21.843045444340923, "learning_rate": 1.437610332360988e-06, "loss": 0.3982, "step": 15987 }, { "epoch": 2.5, "grad_norm": 26.12486526406602, "learning_rate": 1.4367392942418912e-06, "loss": 0.46, "step": 15988 }, { "epoch": 2.5, "grad_norm": 23.313938280588555, "learning_rate": 1.4358684996569194e-06, "loss": 0.4364, "step": 15989 }, { "epoch": 2.5, "grad_norm": 27.570817531998458, "learning_rate": 1.4349979486308396e-06, "loss": 0.4876, "step": 15990 }, { "epoch": 2.5, "grad_norm": 18.717144820880137, "learning_rate": 1.4341276411884076e-06, "loss": 0.4115, "step": 15991 }, { "epoch": 2.5, "grad_norm": 21.42005007933213, "learning_rate": 1.4332575773543732e-06, "loss": 0.442, "step": 15992 }, { "epoch": 2.5, "grad_norm": 18.655672717787574, "learning_rate": 1.4323877571534817e-06, "loss": 0.4908, "step": 15993 }, { "epoch": 2.5, "grad_norm": 17.220542468255783, "learning_rate": 1.4315181806104694e-06, "loss": 0.43, "step": 15994 }, { "epoch": 2.5, "grad_norm": 16.95881503331552, "learning_rate": 1.4306488477500714e-06, "loss": 0.394, "step": 15995 }, { "epoch": 2.5, "grad_norm": 19.022329283281767, "learning_rate": 1.4297797585970063e-06, "loss": 0.4465, "step": 15996 }, { "epoch": 2.5, "grad_norm": 18.07103298678883, "learning_rate": 1.4289109131759894e-06, "loss": 0.3926, "step": 15997 }, { "epoch": 2.5, "grad_norm": 20.532406228239893, "learning_rate": 1.4280423115117316e-06, "loss": 0.411, "step": 15998 }, { "epoch": 2.5, "grad_norm": 23.49068499277464, "learning_rate": 1.4271739536289386e-06, "loss": 0.4644, "step": 15999 }, { "epoch": 2.5, "grad_norm": 24.123598302629787, "learning_rate": 1.4263058395523044e-06, "loss": 0.4253, "step": 16000 }, { "epoch": 2.5, "grad_norm": 19.33500406266326, "learning_rate": 1.4254379693065134e-06, "loss": 0.4545, "step": 16001 }, { "epoch": 2.5, "grad_norm": 25.72876902326394, "learning_rate": 1.4245703429162504e-06, "loss": 0.3899, "step": 16002 }, { "epoch": 2.5, "grad_norm": 27.910769522184466, "learning_rate": 1.4237029604061936e-06, "loss": 0.4761, "step": 16003 }, { "epoch": 2.5, "grad_norm": 23.265467622710162, "learning_rate": 1.422835821801004e-06, "loss": 0.4419, "step": 16004 }, { "epoch": 2.5, "grad_norm": 14.359258228902299, "learning_rate": 1.42196892712535e-06, "loss": 0.426, "step": 16005 }, { "epoch": 2.5, "grad_norm": 16.92926850568379, "learning_rate": 1.4211022764038785e-06, "loss": 0.4148, "step": 16006 }, { "epoch": 2.5, "grad_norm": 32.90000095425404, "learning_rate": 1.4202358696612417e-06, "loss": 0.3752, "step": 16007 }, { "epoch": 2.5, "grad_norm": 24.18432842020453, "learning_rate": 1.4193697069220758e-06, "loss": 0.425, "step": 16008 }, { "epoch": 2.5, "grad_norm": 22.54681177586437, "learning_rate": 1.4185037882110165e-06, "loss": 0.4678, "step": 16009 }, { "epoch": 2.5, "grad_norm": 31.959970348180875, "learning_rate": 1.4176381135526906e-06, "loss": 0.4374, "step": 16010 }, { "epoch": 2.5, "grad_norm": 24.57881377540959, "learning_rate": 1.4167726829717155e-06, "loss": 0.4897, "step": 16011 }, { "epoch": 2.5, "grad_norm": 20.11600947514142, "learning_rate": 1.4159074964927055e-06, "loss": 0.4423, "step": 16012 }, { "epoch": 2.5, "grad_norm": 19.540866472246, "learning_rate": 1.4150425541402623e-06, "loss": 0.4107, "step": 16013 }, { "epoch": 2.5, "grad_norm": 15.356355147104633, "learning_rate": 1.4141778559389875e-06, "loss": 0.4353, "step": 16014 }, { "epoch": 2.5, "grad_norm": 27.35814947218924, "learning_rate": 1.413313401913473e-06, "loss": 0.5225, "step": 16015 }, { "epoch": 2.5, "grad_norm": 32.787733076047395, "learning_rate": 1.4124491920883031e-06, "loss": 0.4293, "step": 16016 }, { "epoch": 2.5, "grad_norm": 21.90812415991012, "learning_rate": 1.4115852264880525e-06, "loss": 0.4075, "step": 16017 }, { "epoch": 2.5, "grad_norm": 18.233990539348376, "learning_rate": 1.4107215051372935e-06, "loss": 0.4186, "step": 16018 }, { "epoch": 2.5, "grad_norm": 19.485973633662397, "learning_rate": 1.409858028060591e-06, "loss": 0.4672, "step": 16019 }, { "epoch": 2.5, "grad_norm": 19.039306624189297, "learning_rate": 1.4089947952825057e-06, "loss": 0.4286, "step": 16020 }, { "epoch": 2.5, "grad_norm": 22.8893474463208, "learning_rate": 1.4081318068275773e-06, "loss": 0.4361, "step": 16021 }, { "epoch": 2.5, "grad_norm": 24.24454182777104, "learning_rate": 1.4072690627203544e-06, "loss": 0.4408, "step": 16022 }, { "epoch": 2.5, "grad_norm": 17.432240352625374, "learning_rate": 1.406406562985373e-06, "loss": 0.4632, "step": 16023 }, { "epoch": 2.5, "grad_norm": 22.903756152927027, "learning_rate": 1.4055443076471641e-06, "loss": 0.465, "step": 16024 }, { "epoch": 2.5, "grad_norm": 23.215776742143245, "learning_rate": 1.4046822967302477e-06, "loss": 0.4997, "step": 16025 }, { "epoch": 2.5, "grad_norm": 14.620105898580562, "learning_rate": 1.4038205302591368e-06, "loss": 0.4179, "step": 16026 }, { "epoch": 2.5, "grad_norm": 21.436592743890895, "learning_rate": 1.4029590082583411e-06, "loss": 0.4302, "step": 16027 }, { "epoch": 2.5, "grad_norm": 29.291359342036372, "learning_rate": 1.4020977307523642e-06, "loss": 0.4827, "step": 16028 }, { "epoch": 2.5, "grad_norm": 30.086728616306587, "learning_rate": 1.4012366977656954e-06, "loss": 0.442, "step": 16029 }, { "epoch": 2.5, "grad_norm": 26.25214179815325, "learning_rate": 1.4003759093228275e-06, "loss": 0.3874, "step": 16030 }, { "epoch": 2.5, "grad_norm": 23.431748455690133, "learning_rate": 1.3995153654482363e-06, "loss": 0.4182, "step": 16031 }, { "epoch": 2.5, "grad_norm": 19.557807866195223, "learning_rate": 1.3986550661663978e-06, "loss": 0.393, "step": 16032 }, { "epoch": 2.5, "grad_norm": 27.632800985802522, "learning_rate": 1.3977950115017768e-06, "loss": 0.4477, "step": 16033 }, { "epoch": 2.5, "grad_norm": 28.932862035770018, "learning_rate": 1.3969352014788328e-06, "loss": 0.457, "step": 16034 }, { "epoch": 2.5, "grad_norm": 18.31484045201061, "learning_rate": 1.3960756361220219e-06, "loss": 0.5123, "step": 16035 }, { "epoch": 2.5, "grad_norm": 30.49663488495039, "learning_rate": 1.395216315455784e-06, "loss": 0.4648, "step": 16036 }, { "epoch": 2.5, "grad_norm": 21.023862308532454, "learning_rate": 1.3943572395045624e-06, "loss": 0.4263, "step": 16037 }, { "epoch": 2.51, "grad_norm": 23.209689269179492, "learning_rate": 1.393498408292785e-06, "loss": 0.4343, "step": 16038 }, { "epoch": 2.51, "grad_norm": 32.665241953493314, "learning_rate": 1.3926398218448778e-06, "loss": 0.4288, "step": 16039 }, { "epoch": 2.51, "grad_norm": 14.577035718088354, "learning_rate": 1.3917814801852615e-06, "loss": 0.4358, "step": 16040 }, { "epoch": 2.51, "grad_norm": 19.79365517657097, "learning_rate": 1.390923383338345e-06, "loss": 0.4464, "step": 16041 }, { "epoch": 2.51, "grad_norm": 31.17829810535406, "learning_rate": 1.3900655313285283e-06, "loss": 0.416, "step": 16042 }, { "epoch": 2.51, "grad_norm": 24.73465749021419, "learning_rate": 1.3892079241802104e-06, "loss": 0.3867, "step": 16043 }, { "epoch": 2.51, "grad_norm": 25.43623042951149, "learning_rate": 1.3883505619177829e-06, "loss": 0.4216, "step": 16044 }, { "epoch": 2.51, "grad_norm": 15.731861722553425, "learning_rate": 1.3874934445656319e-06, "loss": 0.4715, "step": 16045 }, { "epoch": 2.51, "grad_norm": 20.884758223208838, "learning_rate": 1.3866365721481256e-06, "loss": 0.3903, "step": 16046 }, { "epoch": 2.51, "grad_norm": 28.662180169975024, "learning_rate": 1.3857799446896358e-06, "loss": 0.4619, "step": 16047 }, { "epoch": 2.51, "grad_norm": 27.883706859547214, "learning_rate": 1.3849235622145253e-06, "loss": 0.4339, "step": 16048 }, { "epoch": 2.51, "grad_norm": 22.12828886108539, "learning_rate": 1.3840674247471508e-06, "loss": 0.4407, "step": 16049 }, { "epoch": 2.51, "grad_norm": 16.025920230127273, "learning_rate": 1.383211532311859e-06, "loss": 0.4243, "step": 16050 }, { "epoch": 2.51, "grad_norm": 18.96181464946443, "learning_rate": 1.3823558849329877e-06, "loss": 0.4347, "step": 16051 }, { "epoch": 2.51, "grad_norm": 15.193194252977401, "learning_rate": 1.3815004826348744e-06, "loss": 0.3938, "step": 16052 }, { "epoch": 2.51, "grad_norm": 23.72867822493261, "learning_rate": 1.3806453254418483e-06, "loss": 0.4712, "step": 16053 }, { "epoch": 2.51, "grad_norm": 15.049578108965338, "learning_rate": 1.3797904133782269e-06, "loss": 0.4461, "step": 16054 }, { "epoch": 2.51, "grad_norm": 23.052904414915574, "learning_rate": 1.3789357464683206e-06, "loss": 0.4673, "step": 16055 }, { "epoch": 2.51, "grad_norm": 27.994096920293632, "learning_rate": 1.3780813247364388e-06, "loss": 0.4564, "step": 16056 }, { "epoch": 2.51, "grad_norm": 16.532372325701015, "learning_rate": 1.3772271482068799e-06, "loss": 0.4643, "step": 16057 }, { "epoch": 2.51, "grad_norm": 43.587454237994585, "learning_rate": 1.3763732169039401e-06, "loss": 0.6139, "step": 16058 }, { "epoch": 2.51, "grad_norm": 16.64526821914078, "learning_rate": 1.375519530851901e-06, "loss": 0.4037, "step": 16059 }, { "epoch": 2.51, "grad_norm": 22.01109019400029, "learning_rate": 1.3746660900750386e-06, "loss": 0.4552, "step": 16060 }, { "epoch": 2.51, "grad_norm": 20.013676850076404, "learning_rate": 1.373812894597627e-06, "loss": 0.4199, "step": 16061 }, { "epoch": 2.51, "grad_norm": 17.5749172445807, "learning_rate": 1.3729599444439323e-06, "loss": 0.4345, "step": 16062 }, { "epoch": 2.51, "grad_norm": 19.31902425804104, "learning_rate": 1.3721072396382085e-06, "loss": 0.4092, "step": 16063 }, { "epoch": 2.51, "grad_norm": 29.792735227792747, "learning_rate": 1.3712547802047094e-06, "loss": 0.5029, "step": 16064 }, { "epoch": 2.51, "grad_norm": 18.350584615520887, "learning_rate": 1.3704025661676757e-06, "loss": 0.4988, "step": 16065 }, { "epoch": 2.51, "grad_norm": 24.355428596836692, "learning_rate": 1.3695505975513456e-06, "loss": 0.4713, "step": 16066 }, { "epoch": 2.51, "grad_norm": 18.577649434470043, "learning_rate": 1.3686988743799466e-06, "loss": 0.4024, "step": 16067 }, { "epoch": 2.51, "grad_norm": 12.396153702389274, "learning_rate": 1.3678473966777018e-06, "loss": 0.4521, "step": 16068 }, { "epoch": 2.51, "grad_norm": 21.607760226309008, "learning_rate": 1.36699616446883e-06, "loss": 0.4152, "step": 16069 }, { "epoch": 2.51, "grad_norm": 29.978763376248963, "learning_rate": 1.3661451777775369e-06, "loss": 0.4588, "step": 16070 }, { "epoch": 2.51, "grad_norm": 23.434229630867517, "learning_rate": 1.3652944366280218e-06, "loss": 0.4732, "step": 16071 }, { "epoch": 2.51, "grad_norm": 23.066913181196533, "learning_rate": 1.364443941044482e-06, "loss": 0.4508, "step": 16072 }, { "epoch": 2.51, "grad_norm": 20.84857807119181, "learning_rate": 1.363593691051105e-06, "loss": 0.4867, "step": 16073 }, { "epoch": 2.51, "grad_norm": 16.880398707938138, "learning_rate": 1.3627436866720734e-06, "loss": 0.4079, "step": 16074 }, { "epoch": 2.51, "grad_norm": 18.449364661795336, "learning_rate": 1.3618939279315591e-06, "loss": 0.4162, "step": 16075 }, { "epoch": 2.51, "grad_norm": 28.60231956932494, "learning_rate": 1.3610444148537261e-06, "loss": 0.4687, "step": 16076 }, { "epoch": 2.51, "grad_norm": 12.10890978011818, "learning_rate": 1.3601951474627362e-06, "loss": 0.4715, "step": 16077 }, { "epoch": 2.51, "grad_norm": 19.36689264581611, "learning_rate": 1.3593461257827433e-06, "loss": 0.4134, "step": 16078 }, { "epoch": 2.51, "grad_norm": 18.24007244469165, "learning_rate": 1.3584973498378928e-06, "loss": 0.4382, "step": 16079 }, { "epoch": 2.51, "grad_norm": 16.53252686380978, "learning_rate": 1.3576488196523207e-06, "loss": 0.3798, "step": 16080 }, { "epoch": 2.51, "grad_norm": 24.835521443719877, "learning_rate": 1.3568005352501602e-06, "loss": 0.4876, "step": 16081 }, { "epoch": 2.51, "grad_norm": 19.419010135196174, "learning_rate": 1.355952496655536e-06, "loss": 0.4499, "step": 16082 }, { "epoch": 2.51, "grad_norm": 22.79683973428614, "learning_rate": 1.3551047038925692e-06, "loss": 0.475, "step": 16083 }, { "epoch": 2.51, "grad_norm": 16.234364634410497, "learning_rate": 1.3542571569853669e-06, "loss": 0.3969, "step": 16084 }, { "epoch": 2.51, "grad_norm": 29.244811420088652, "learning_rate": 1.353409855958031e-06, "loss": 0.4857, "step": 16085 }, { "epoch": 2.51, "grad_norm": 23.346327759358495, "learning_rate": 1.3525628008346613e-06, "loss": 0.4477, "step": 16086 }, { "epoch": 2.51, "grad_norm": 16.592895253216835, "learning_rate": 1.3517159916393485e-06, "loss": 0.3904, "step": 16087 }, { "epoch": 2.51, "grad_norm": 16.77317982561453, "learning_rate": 1.350869428396172e-06, "loss": 0.4769, "step": 16088 }, { "epoch": 2.51, "grad_norm": 17.477146218228754, "learning_rate": 1.3500231111292118e-06, "loss": 0.4318, "step": 16089 }, { "epoch": 2.51, "grad_norm": 21.350442247701558, "learning_rate": 1.3491770398625315e-06, "loss": 0.4135, "step": 16090 }, { "epoch": 2.51, "grad_norm": 23.693940265991415, "learning_rate": 1.3483312146201988e-06, "loss": 0.4339, "step": 16091 }, { "epoch": 2.51, "grad_norm": 15.730008705654967, "learning_rate": 1.3474856354262622e-06, "loss": 0.366, "step": 16092 }, { "epoch": 2.51, "grad_norm": 33.185076490374136, "learning_rate": 1.3466403023047724e-06, "loss": 0.4819, "step": 16093 }, { "epoch": 2.51, "grad_norm": 18.95929009220646, "learning_rate": 1.3457952152797727e-06, "loss": 0.4776, "step": 16094 }, { "epoch": 2.51, "grad_norm": 24.565996670989747, "learning_rate": 1.3449503743752945e-06, "loss": 0.4369, "step": 16095 }, { "epoch": 2.51, "grad_norm": 27.02435263467979, "learning_rate": 1.3441057796153634e-06, "loss": 0.4302, "step": 16096 }, { "epoch": 2.51, "grad_norm": 22.766718204264784, "learning_rate": 1.3432614310239989e-06, "loss": 0.45, "step": 16097 }, { "epoch": 2.51, "grad_norm": 22.88300408973077, "learning_rate": 1.3424173286252185e-06, "loss": 0.4317, "step": 16098 }, { "epoch": 2.51, "grad_norm": 17.461529160615942, "learning_rate": 1.3415734724430219e-06, "loss": 0.4074, "step": 16099 }, { "epoch": 2.51, "grad_norm": 14.915799160713203, "learning_rate": 1.340729862501413e-06, "loss": 0.412, "step": 16100 }, { "epoch": 2.51, "grad_norm": 37.88191126842757, "learning_rate": 1.3398864988243786e-06, "loss": 0.453, "step": 16101 }, { "epoch": 2.52, "grad_norm": 16.244575570803324, "learning_rate": 1.3390433814359072e-06, "loss": 0.4578, "step": 16102 }, { "epoch": 2.52, "grad_norm": 15.751859441621695, "learning_rate": 1.338200510359977e-06, "loss": 0.4327, "step": 16103 }, { "epoch": 2.52, "grad_norm": 16.355720114478938, "learning_rate": 1.3373578856205583e-06, "loss": 0.4698, "step": 16104 }, { "epoch": 2.52, "grad_norm": 23.373897884773598, "learning_rate": 1.3365155072416103e-06, "loss": 0.42, "step": 16105 }, { "epoch": 2.52, "grad_norm": 17.9143363783638, "learning_rate": 1.335673375247094e-06, "loss": 0.4645, "step": 16106 }, { "epoch": 2.52, "grad_norm": 26.02268170599414, "learning_rate": 1.3348314896609581e-06, "loss": 0.4158, "step": 16107 }, { "epoch": 2.52, "grad_norm": 16.200770929030377, "learning_rate": 1.3339898505071503e-06, "loss": 0.454, "step": 16108 }, { "epoch": 2.52, "grad_norm": 22.129009123358525, "learning_rate": 1.3331484578095976e-06, "loss": 0.3922, "step": 16109 }, { "epoch": 2.52, "grad_norm": 24.081902440972662, "learning_rate": 1.3323073115922324e-06, "loss": 0.3776, "step": 16110 }, { "epoch": 2.52, "grad_norm": 18.908004810139264, "learning_rate": 1.3314664118789766e-06, "loss": 0.4503, "step": 16111 }, { "epoch": 2.52, "grad_norm": 29.913964095791915, "learning_rate": 1.3306257586937476e-06, "loss": 0.4645, "step": 16112 }, { "epoch": 2.52, "grad_norm": 18.913729501256434, "learning_rate": 1.3297853520604497e-06, "loss": 0.4864, "step": 16113 }, { "epoch": 2.52, "grad_norm": 15.73875133979761, "learning_rate": 1.3289451920029838e-06, "loss": 0.4436, "step": 16114 }, { "epoch": 2.52, "grad_norm": 22.27661197764153, "learning_rate": 1.328105278545243e-06, "loss": 0.3886, "step": 16115 }, { "epoch": 2.52, "grad_norm": 22.03122522778843, "learning_rate": 1.3272656117111183e-06, "loss": 0.42, "step": 16116 }, { "epoch": 2.52, "grad_norm": 23.480596286577722, "learning_rate": 1.3264261915244836e-06, "loss": 0.4274, "step": 16117 }, { "epoch": 2.52, "grad_norm": 24.36375371864161, "learning_rate": 1.3255870180092157e-06, "loss": 0.4866, "step": 16118 }, { "epoch": 2.52, "grad_norm": 18.549188702571065, "learning_rate": 1.3247480911891775e-06, "loss": 0.4201, "step": 16119 }, { "epoch": 2.52, "grad_norm": 19.961146512215674, "learning_rate": 1.3239094110882279e-06, "loss": 0.399, "step": 16120 }, { "epoch": 2.52, "grad_norm": 32.790279165894724, "learning_rate": 1.323070977730222e-06, "loss": 0.4622, "step": 16121 }, { "epoch": 2.52, "grad_norm": 23.84024557993538, "learning_rate": 1.3222327911389988e-06, "loss": 0.4507, "step": 16122 }, { "epoch": 2.52, "grad_norm": 22.435744356362395, "learning_rate": 1.3213948513384011e-06, "loss": 0.3852, "step": 16123 }, { "epoch": 2.52, "grad_norm": 15.939901484215818, "learning_rate": 1.320557158352256e-06, "loss": 0.4418, "step": 16124 }, { "epoch": 2.52, "grad_norm": 27.66030546562115, "learning_rate": 1.3197197122043892e-06, "loss": 0.3918, "step": 16125 }, { "epoch": 2.52, "grad_norm": 34.52606753646948, "learning_rate": 1.3188825129186134e-06, "loss": 0.5061, "step": 16126 }, { "epoch": 2.52, "grad_norm": 19.435270130179433, "learning_rate": 1.3180455605187404e-06, "loss": 0.5326, "step": 16127 }, { "epoch": 2.52, "grad_norm": 21.34916310763949, "learning_rate": 1.3172088550285756e-06, "loss": 0.4382, "step": 16128 }, { "epoch": 2.52, "grad_norm": 17.489164564853382, "learning_rate": 1.3163723964719122e-06, "loss": 0.4377, "step": 16129 }, { "epoch": 2.52, "grad_norm": 16.49358801901327, "learning_rate": 1.3155361848725357e-06, "loss": 0.4037, "step": 16130 }, { "epoch": 2.52, "grad_norm": 17.431644727348466, "learning_rate": 1.3147002202542291e-06, "loss": 0.4086, "step": 16131 }, { "epoch": 2.52, "grad_norm": 22.87720014982173, "learning_rate": 1.3138645026407682e-06, "loss": 0.5487, "step": 16132 }, { "epoch": 2.52, "grad_norm": 28.105182428339685, "learning_rate": 1.3130290320559235e-06, "loss": 0.4736, "step": 16133 }, { "epoch": 2.52, "grad_norm": 16.121196676710916, "learning_rate": 1.3121938085234466e-06, "loss": 0.3873, "step": 16134 }, { "epoch": 2.52, "grad_norm": 25.141077182181313, "learning_rate": 1.3113588320670967e-06, "loss": 0.521, "step": 16135 }, { "epoch": 2.52, "grad_norm": 25.3080304067366, "learning_rate": 1.3105241027106187e-06, "loss": 0.4905, "step": 16136 }, { "epoch": 2.52, "grad_norm": 14.552060328852011, "learning_rate": 1.3096896204777532e-06, "loss": 0.4588, "step": 16137 }, { "epoch": 2.52, "grad_norm": 19.95056523741904, "learning_rate": 1.3088553853922325e-06, "loss": 0.4285, "step": 16138 }, { "epoch": 2.52, "grad_norm": 16.930672067267473, "learning_rate": 1.3080213974777766e-06, "loss": 0.4071, "step": 16139 }, { "epoch": 2.52, "grad_norm": 21.45939903035727, "learning_rate": 1.3071876567581087e-06, "loss": 0.4511, "step": 16140 }, { "epoch": 2.52, "grad_norm": 28.289480477010375, "learning_rate": 1.306354163256941e-06, "loss": 0.4329, "step": 16141 }, { "epoch": 2.52, "grad_norm": 22.187050965291913, "learning_rate": 1.3055209169979754e-06, "loss": 0.4594, "step": 16142 }, { "epoch": 2.52, "grad_norm": 20.203624089294813, "learning_rate": 1.3046879180049054e-06, "loss": 0.4578, "step": 16143 }, { "epoch": 2.52, "grad_norm": 39.72745619587268, "learning_rate": 1.3038551663014255e-06, "loss": 0.4512, "step": 16144 }, { "epoch": 2.52, "grad_norm": 22.021933170998004, "learning_rate": 1.3030226619112175e-06, "loss": 0.4401, "step": 16145 }, { "epoch": 2.52, "grad_norm": 24.099133587682196, "learning_rate": 1.3021904048579592e-06, "loss": 0.4342, "step": 16146 }, { "epoch": 2.52, "grad_norm": 19.128758235745906, "learning_rate": 1.3013583951653185e-06, "loss": 0.4436, "step": 16147 }, { "epoch": 2.52, "grad_norm": 22.877803676924344, "learning_rate": 1.3005266328569533e-06, "loss": 0.4927, "step": 16148 }, { "epoch": 2.52, "grad_norm": 21.34779687540237, "learning_rate": 1.2996951179565221e-06, "loss": 0.4455, "step": 16149 }, { "epoch": 2.52, "grad_norm": 21.475657447125872, "learning_rate": 1.2988638504876738e-06, "loss": 0.4777, "step": 16150 }, { "epoch": 2.52, "grad_norm": 26.98429490863332, "learning_rate": 1.2980328304740464e-06, "loss": 0.4661, "step": 16151 }, { "epoch": 2.52, "grad_norm": 15.879735638306125, "learning_rate": 1.2972020579392763e-06, "loss": 0.4237, "step": 16152 }, { "epoch": 2.52, "grad_norm": 23.275216031121918, "learning_rate": 1.296371532906987e-06, "loss": 0.4969, "step": 16153 }, { "epoch": 2.52, "grad_norm": 27.978751454050144, "learning_rate": 1.2955412554008018e-06, "loss": 0.5051, "step": 16154 }, { "epoch": 2.52, "grad_norm": 16.76163886252795, "learning_rate": 1.2947112254443283e-06, "loss": 0.4076, "step": 16155 }, { "epoch": 2.52, "grad_norm": 22.30297299157287, "learning_rate": 1.2938814430611756e-06, "loss": 0.4316, "step": 16156 }, { "epoch": 2.52, "grad_norm": 26.578792343866596, "learning_rate": 1.2930519082749448e-06, "loss": 0.4022, "step": 16157 }, { "epoch": 2.52, "grad_norm": 15.995794345629422, "learning_rate": 1.2922226211092237e-06, "loss": 0.4509, "step": 16158 }, { "epoch": 2.52, "grad_norm": 19.25341360948246, "learning_rate": 1.2913935815875945e-06, "loss": 0.4199, "step": 16159 }, { "epoch": 2.52, "grad_norm": 19.742479203684088, "learning_rate": 1.2905647897336382e-06, "loss": 0.4515, "step": 16160 }, { "epoch": 2.52, "grad_norm": 18.099668360077146, "learning_rate": 1.289736245570925e-06, "loss": 0.4401, "step": 16161 }, { "epoch": 2.52, "grad_norm": 13.951648367037468, "learning_rate": 1.2889079491230184e-06, "loss": 0.4362, "step": 16162 }, { "epoch": 2.52, "grad_norm": 30.03526686559841, "learning_rate": 1.2880799004134748e-06, "loss": 0.4427, "step": 16163 }, { "epoch": 2.52, "grad_norm": 17.127564521880164, "learning_rate": 1.2872520994658399e-06, "loss": 0.3623, "step": 16164 }, { "epoch": 2.52, "grad_norm": 21.367385864731812, "learning_rate": 1.2864245463036585e-06, "loss": 0.4744, "step": 16165 }, { "epoch": 2.53, "grad_norm": 19.521943041635705, "learning_rate": 1.285597240950468e-06, "loss": 0.4562, "step": 16166 }, { "epoch": 2.53, "grad_norm": 25.701737878931475, "learning_rate": 1.2847701834297944e-06, "loss": 0.4992, "step": 16167 }, { "epoch": 2.53, "grad_norm": 23.597040625228257, "learning_rate": 1.2839433737651564e-06, "loss": 0.4109, "step": 16168 }, { "epoch": 2.53, "grad_norm": 26.400038246444073, "learning_rate": 1.2831168119800696e-06, "loss": 0.4778, "step": 16169 }, { "epoch": 2.53, "grad_norm": 26.384701177462002, "learning_rate": 1.2822904980980422e-06, "loss": 0.4188, "step": 16170 }, { "epoch": 2.53, "grad_norm": 28.6465904502806, "learning_rate": 1.2814644321425752e-06, "loss": 0.4454, "step": 16171 }, { "epoch": 2.53, "grad_norm": 20.54095085520488, "learning_rate": 1.2806386141371596e-06, "loss": 0.4357, "step": 16172 }, { "epoch": 2.53, "grad_norm": 15.882080351152075, "learning_rate": 1.279813044105278e-06, "loss": 0.4967, "step": 16173 }, { "epoch": 2.53, "grad_norm": 13.645148332403927, "learning_rate": 1.2789877220704127e-06, "loss": 0.4162, "step": 16174 }, { "epoch": 2.53, "grad_norm": 19.405512592262003, "learning_rate": 1.2781626480560384e-06, "loss": 0.4495, "step": 16175 }, { "epoch": 2.53, "grad_norm": 21.214584773953103, "learning_rate": 1.2773378220856126e-06, "loss": 0.4208, "step": 16176 }, { "epoch": 2.53, "grad_norm": 20.417531345061835, "learning_rate": 1.2765132441825989e-06, "loss": 0.3994, "step": 16177 }, { "epoch": 2.53, "grad_norm": 16.558808744906163, "learning_rate": 1.2756889143704443e-06, "loss": 0.4769, "step": 16178 }, { "epoch": 2.53, "grad_norm": 29.28959923896237, "learning_rate": 1.2748648326725943e-06, "loss": 0.4233, "step": 16179 }, { "epoch": 2.53, "grad_norm": 21.396533118104976, "learning_rate": 1.2740409991124826e-06, "loss": 0.3972, "step": 16180 }, { "epoch": 2.53, "grad_norm": 19.996897906889256, "learning_rate": 1.273217413713541e-06, "loss": 0.5071, "step": 16181 }, { "epoch": 2.53, "grad_norm": 29.145206892492695, "learning_rate": 1.2723940764991927e-06, "loss": 0.5035, "step": 16182 }, { "epoch": 2.53, "grad_norm": 20.378472820613535, "learning_rate": 1.2715709874928518e-06, "loss": 0.4704, "step": 16183 }, { "epoch": 2.53, "grad_norm": 17.606376671089954, "learning_rate": 1.2707481467179228e-06, "loss": 0.4685, "step": 16184 }, { "epoch": 2.53, "grad_norm": 18.684463760661522, "learning_rate": 1.2699255541978117e-06, "loss": 0.4176, "step": 16185 }, { "epoch": 2.53, "grad_norm": 32.92323274487656, "learning_rate": 1.2691032099559097e-06, "loss": 0.5049, "step": 16186 }, { "epoch": 2.53, "grad_norm": 19.893253018705774, "learning_rate": 1.268281114015607e-06, "loss": 0.4758, "step": 16187 }, { "epoch": 2.53, "grad_norm": 29.18216352260454, "learning_rate": 1.2674592664002828e-06, "loss": 0.4583, "step": 16188 }, { "epoch": 2.53, "grad_norm": 22.359304695898565, "learning_rate": 1.2666376671333048e-06, "loss": 0.4736, "step": 16189 }, { "epoch": 2.53, "grad_norm": 23.29364880022195, "learning_rate": 1.2658163162380443e-06, "loss": 0.4767, "step": 16190 }, { "epoch": 2.53, "grad_norm": 18.124525445447688, "learning_rate": 1.2649952137378596e-06, "loss": 0.5043, "step": 16191 }, { "epoch": 2.53, "grad_norm": 26.6021441932411, "learning_rate": 1.2641743596561007e-06, "loss": 0.4006, "step": 16192 }, { "epoch": 2.53, "grad_norm": 18.132631134151765, "learning_rate": 1.2633537540161123e-06, "loss": 0.4597, "step": 16193 }, { "epoch": 2.53, "grad_norm": 20.867707415623116, "learning_rate": 1.2625333968412312e-06, "loss": 0.4097, "step": 16194 }, { "epoch": 2.53, "grad_norm": 19.063351136889246, "learning_rate": 1.261713288154789e-06, "loss": 0.4489, "step": 16195 }, { "epoch": 2.53, "grad_norm": 19.23634535290724, "learning_rate": 1.2608934279801133e-06, "loss": 0.4439, "step": 16196 }, { "epoch": 2.53, "grad_norm": 20.721883468887953, "learning_rate": 1.2600738163405124e-06, "loss": 0.3929, "step": 16197 }, { "epoch": 2.53, "grad_norm": 22.324421003465098, "learning_rate": 1.2592544532592988e-06, "loss": 0.4098, "step": 16198 }, { "epoch": 2.53, "grad_norm": 19.71749577334569, "learning_rate": 1.2584353387597758e-06, "loss": 0.4543, "step": 16199 }, { "epoch": 2.53, "grad_norm": 30.93114246878487, "learning_rate": 1.2576164728652406e-06, "loss": 0.4372, "step": 16200 }, { "epoch": 2.53, "grad_norm": 24.22293906007539, "learning_rate": 1.256797855598978e-06, "loss": 0.525, "step": 16201 }, { "epoch": 2.53, "grad_norm": 23.434392992707878, "learning_rate": 1.255979486984269e-06, "loss": 0.418, "step": 16202 }, { "epoch": 2.53, "grad_norm": 13.143526084750306, "learning_rate": 1.2551613670443862e-06, "loss": 0.4043, "step": 16203 }, { "epoch": 2.53, "grad_norm": 16.07366426422336, "learning_rate": 1.2543434958026024e-06, "loss": 0.4313, "step": 16204 }, { "epoch": 2.53, "grad_norm": 19.119309566747837, "learning_rate": 1.253525873282171e-06, "loss": 0.3971, "step": 16205 }, { "epoch": 2.53, "grad_norm": 19.379038246841734, "learning_rate": 1.2527084995063488e-06, "loss": 0.391, "step": 16206 }, { "epoch": 2.53, "grad_norm": 18.502078468770403, "learning_rate": 1.2518913744983786e-06, "loss": 0.4458, "step": 16207 }, { "epoch": 2.53, "grad_norm": 23.802937243285648, "learning_rate": 1.2510744982814993e-06, "loss": 0.4392, "step": 16208 }, { "epoch": 2.53, "grad_norm": 18.102730585342382, "learning_rate": 1.250257870878946e-06, "loss": 0.4205, "step": 16209 }, { "epoch": 2.53, "grad_norm": 21.19327764976099, "learning_rate": 1.2494414923139375e-06, "loss": 0.3977, "step": 16210 }, { "epoch": 2.53, "grad_norm": 22.61116185472731, "learning_rate": 1.2486253626096978e-06, "loss": 0.3901, "step": 16211 }, { "epoch": 2.53, "grad_norm": 27.417180792549605, "learning_rate": 1.2478094817894293e-06, "loss": 0.513, "step": 16212 }, { "epoch": 2.53, "grad_norm": 24.06952805681748, "learning_rate": 1.2469938498763423e-06, "loss": 0.4251, "step": 16213 }, { "epoch": 2.53, "grad_norm": 20.828894784814825, "learning_rate": 1.2461784668936283e-06, "loss": 0.4059, "step": 16214 }, { "epoch": 2.53, "grad_norm": 15.310972599807217, "learning_rate": 1.2453633328644765e-06, "loss": 0.4425, "step": 16215 }, { "epoch": 2.53, "grad_norm": 22.515909237493734, "learning_rate": 1.244548447812074e-06, "loss": 0.4189, "step": 16216 }, { "epoch": 2.53, "grad_norm": 16.304867524568387, "learning_rate": 1.243733811759591e-06, "loss": 0.421, "step": 16217 }, { "epoch": 2.53, "grad_norm": 29.198554136947916, "learning_rate": 1.2429194247301934e-06, "loss": 0.4726, "step": 16218 }, { "epoch": 2.53, "grad_norm": 20.18794052924772, "learning_rate": 1.2421052867470452e-06, "loss": 0.4199, "step": 16219 }, { "epoch": 2.53, "grad_norm": 15.717766903570451, "learning_rate": 1.2412913978332997e-06, "loss": 0.4405, "step": 16220 }, { "epoch": 2.53, "grad_norm": 16.601425967833137, "learning_rate": 1.2404777580121075e-06, "loss": 0.4061, "step": 16221 }, { "epoch": 2.53, "grad_norm": 31.139884177170483, "learning_rate": 1.2396643673065988e-06, "loss": 0.5397, "step": 16222 }, { "epoch": 2.53, "grad_norm": 23.946683831331104, "learning_rate": 1.238851225739911e-06, "loss": 0.476, "step": 16223 }, { "epoch": 2.53, "grad_norm": 32.56812494059177, "learning_rate": 1.2380383333351687e-06, "loss": 0.4861, "step": 16224 }, { "epoch": 2.53, "grad_norm": 28.42693705199055, "learning_rate": 1.2372256901154934e-06, "loss": 0.467, "step": 16225 }, { "epoch": 2.53, "grad_norm": 27.79674657853102, "learning_rate": 1.2364132961039933e-06, "loss": 0.4302, "step": 16226 }, { "epoch": 2.53, "grad_norm": 33.61215200084452, "learning_rate": 1.2356011513237708e-06, "loss": 0.5053, "step": 16227 }, { "epoch": 2.53, "grad_norm": 26.108951646213594, "learning_rate": 1.2347892557979236e-06, "loss": 0.4456, "step": 16228 }, { "epoch": 2.53, "grad_norm": 22.40774533005232, "learning_rate": 1.233977609549546e-06, "loss": 0.4224, "step": 16229 }, { "epoch": 2.54, "grad_norm": 27.591776182143025, "learning_rate": 1.2331662126017141e-06, "loss": 0.4777, "step": 16230 }, { "epoch": 2.54, "grad_norm": 23.88816294511219, "learning_rate": 1.2323550649775085e-06, "loss": 0.4592, "step": 16231 }, { "epoch": 2.54, "grad_norm": 19.915892017501292, "learning_rate": 1.2315441666999939e-06, "loss": 0.4355, "step": 16232 }, { "epoch": 2.54, "grad_norm": 19.90625801864774, "learning_rate": 1.2307335177922342e-06, "loss": 0.4418, "step": 16233 }, { "epoch": 2.54, "grad_norm": 19.950949541329113, "learning_rate": 1.2299231182772852e-06, "loss": 0.3922, "step": 16234 }, { "epoch": 2.54, "grad_norm": 17.943142660519563, "learning_rate": 1.229112968178191e-06, "loss": 0.4171, "step": 16235 }, { "epoch": 2.54, "grad_norm": 26.25359470413586, "learning_rate": 1.2283030675179951e-06, "loss": 0.4693, "step": 16236 }, { "epoch": 2.54, "grad_norm": 21.474521105256585, "learning_rate": 1.227493416319726e-06, "loss": 0.3891, "step": 16237 }, { "epoch": 2.54, "grad_norm": 15.272140795207816, "learning_rate": 1.226684014606414e-06, "loss": 0.413, "step": 16238 }, { "epoch": 2.54, "grad_norm": 28.175248185835784, "learning_rate": 1.2258748624010752e-06, "loss": 0.415, "step": 16239 }, { "epoch": 2.54, "grad_norm": 24.11331563107011, "learning_rate": 1.2250659597267244e-06, "loss": 0.4373, "step": 16240 }, { "epoch": 2.54, "grad_norm": 17.56547091595823, "learning_rate": 1.2242573066063623e-06, "loss": 0.3961, "step": 16241 }, { "epoch": 2.54, "grad_norm": 21.297519435494202, "learning_rate": 1.2234489030629916e-06, "loss": 0.4004, "step": 16242 }, { "epoch": 2.54, "grad_norm": 19.961800212842434, "learning_rate": 1.2226407491195969e-06, "loss": 0.4261, "step": 16243 }, { "epoch": 2.54, "grad_norm": 25.35944365916944, "learning_rate": 1.2218328447991657e-06, "loss": 0.4462, "step": 16244 }, { "epoch": 2.54, "grad_norm": 17.408130405496266, "learning_rate": 1.2210251901246739e-06, "loss": 0.4898, "step": 16245 }, { "epoch": 2.54, "grad_norm": 20.053793525757946, "learning_rate": 1.2202177851190912e-06, "loss": 0.4534, "step": 16246 }, { "epoch": 2.54, "grad_norm": 28.486410025399437, "learning_rate": 1.2194106298053765e-06, "loss": 0.4737, "step": 16247 }, { "epoch": 2.54, "grad_norm": 16.557755326887538, "learning_rate": 1.2186037242064863e-06, "loss": 0.4554, "step": 16248 }, { "epoch": 2.54, "grad_norm": 21.909335613609805, "learning_rate": 1.21779706834537e-06, "loss": 0.4745, "step": 16249 }, { "epoch": 2.54, "grad_norm": 17.707214554806292, "learning_rate": 1.2169906622449701e-06, "loss": 0.4348, "step": 16250 }, { "epoch": 2.54, "grad_norm": 20.95286784700535, "learning_rate": 1.2161845059282174e-06, "loss": 0.494, "step": 16251 }, { "epoch": 2.54, "grad_norm": 18.850565022757994, "learning_rate": 1.2153785994180366e-06, "loss": 0.4144, "step": 16252 }, { "epoch": 2.54, "grad_norm": 17.879570797585448, "learning_rate": 1.2145729427373499e-06, "loss": 0.4222, "step": 16253 }, { "epoch": 2.54, "grad_norm": 17.964010260370724, "learning_rate": 1.2137675359090705e-06, "loss": 0.4409, "step": 16254 }, { "epoch": 2.54, "grad_norm": 25.876305411875467, "learning_rate": 1.212962378956104e-06, "loss": 0.4295, "step": 16255 }, { "epoch": 2.54, "grad_norm": 25.384379962252396, "learning_rate": 1.212157471901344e-06, "loss": 0.4495, "step": 16256 }, { "epoch": 2.54, "grad_norm": 22.731816360848157, "learning_rate": 1.2113528147676855e-06, "loss": 0.3938, "step": 16257 }, { "epoch": 2.54, "grad_norm": 39.87602008363829, "learning_rate": 1.2105484075780117e-06, "loss": 0.5317, "step": 16258 }, { "epoch": 2.54, "grad_norm": 25.149658652239655, "learning_rate": 1.2097442503552003e-06, "loss": 0.4414, "step": 16259 }, { "epoch": 2.54, "grad_norm": 17.093697265390233, "learning_rate": 1.2089403431221213e-06, "loss": 0.4942, "step": 16260 }, { "epoch": 2.54, "grad_norm": 23.624666810209124, "learning_rate": 1.2081366859016353e-06, "loss": 0.5039, "step": 16261 }, { "epoch": 2.54, "grad_norm": 22.74907721754449, "learning_rate": 1.207333278716597e-06, "loss": 0.47, "step": 16262 }, { "epoch": 2.54, "grad_norm": 4.915220300859985, "learning_rate": 1.2065301215898595e-06, "loss": 0.4499, "step": 16263 }, { "epoch": 2.54, "grad_norm": 19.250262511773297, "learning_rate": 1.20572721454426e-06, "loss": 0.4474, "step": 16264 }, { "epoch": 2.54, "grad_norm": 18.62519518378732, "learning_rate": 1.2049245576026346e-06, "loss": 0.444, "step": 16265 }, { "epoch": 2.54, "grad_norm": 17.28177501991546, "learning_rate": 1.2041221507878087e-06, "loss": 0.4567, "step": 16266 }, { "epoch": 2.54, "grad_norm": 18.670112972037764, "learning_rate": 1.203319994122606e-06, "loss": 0.4854, "step": 16267 }, { "epoch": 2.54, "grad_norm": 15.713256090085531, "learning_rate": 1.2025180876298338e-06, "loss": 0.4029, "step": 16268 }, { "epoch": 2.54, "grad_norm": 34.608469280851885, "learning_rate": 1.2017164313323004e-06, "loss": 0.4417, "step": 16269 }, { "epoch": 2.54, "grad_norm": 15.796480723855371, "learning_rate": 1.200915025252808e-06, "loss": 0.371, "step": 16270 }, { "epoch": 2.54, "grad_norm": 20.9708927860443, "learning_rate": 1.2001138694141423e-06, "loss": 0.4312, "step": 16271 }, { "epoch": 2.54, "grad_norm": 19.769344572411853, "learning_rate": 1.199312963839092e-06, "loss": 0.4483, "step": 16272 }, { "epoch": 2.54, "grad_norm": 25.40493711783794, "learning_rate": 1.1985123085504312e-06, "loss": 0.4976, "step": 16273 }, { "epoch": 2.54, "grad_norm": 21.644188912463495, "learning_rate": 1.1977119035709329e-06, "loss": 0.4464, "step": 16274 }, { "epoch": 2.54, "grad_norm": 22.74301314517832, "learning_rate": 1.1969117489233595e-06, "loss": 0.4299, "step": 16275 }, { "epoch": 2.54, "grad_norm": 20.396701071155285, "learning_rate": 1.1961118446304666e-06, "loss": 0.4506, "step": 16276 }, { "epoch": 2.54, "grad_norm": 18.053488097304097, "learning_rate": 1.195312190715e-06, "loss": 0.4286, "step": 16277 }, { "epoch": 2.54, "grad_norm": 21.954936002378727, "learning_rate": 1.1945127871997042e-06, "loss": 0.4265, "step": 16278 }, { "epoch": 2.54, "grad_norm": 15.260374305349496, "learning_rate": 1.1937136341073163e-06, "loss": 0.3861, "step": 16279 }, { "epoch": 2.54, "grad_norm": 18.288779962533507, "learning_rate": 1.1929147314605617e-06, "loss": 0.4704, "step": 16280 }, { "epoch": 2.54, "grad_norm": 18.91514803069053, "learning_rate": 1.1921160792821572e-06, "loss": 0.4546, "step": 16281 }, { "epoch": 2.54, "grad_norm": 15.176547670587869, "learning_rate": 1.1913176775948187e-06, "loss": 0.4447, "step": 16282 }, { "epoch": 2.54, "grad_norm": 22.289080741813354, "learning_rate": 1.1905195264212532e-06, "loss": 0.4224, "step": 16283 }, { "epoch": 2.54, "grad_norm": 24.3583317619512, "learning_rate": 1.1897216257841605e-06, "loss": 0.455, "step": 16284 }, { "epoch": 2.54, "grad_norm": 24.582762408631325, "learning_rate": 1.1889239757062309e-06, "loss": 0.435, "step": 16285 }, { "epoch": 2.54, "grad_norm": 32.851864820851475, "learning_rate": 1.188126576210148e-06, "loss": 0.4154, "step": 16286 }, { "epoch": 2.54, "grad_norm": 20.140068313593815, "learning_rate": 1.1873294273185898e-06, "loss": 0.4222, "step": 16287 }, { "epoch": 2.54, "grad_norm": 17.573977639807598, "learning_rate": 1.1865325290542295e-06, "loss": 0.4762, "step": 16288 }, { "epoch": 2.54, "grad_norm": 17.67043129313243, "learning_rate": 1.185735881439728e-06, "loss": 0.5002, "step": 16289 }, { "epoch": 2.54, "grad_norm": 22.47103234439085, "learning_rate": 1.1849394844977402e-06, "loss": 0.4952, "step": 16290 }, { "epoch": 2.54, "grad_norm": 18.97538011588581, "learning_rate": 1.184143338250917e-06, "loss": 0.4181, "step": 16291 }, { "epoch": 2.54, "grad_norm": 15.072978286731447, "learning_rate": 1.1833474427219015e-06, "loss": 0.4126, "step": 16292 }, { "epoch": 2.54, "grad_norm": 21.14310812340901, "learning_rate": 1.1825517979333256e-06, "loss": 0.4642, "step": 16293 }, { "epoch": 2.55, "grad_norm": 12.274444614032154, "learning_rate": 1.18175640390782e-06, "loss": 0.4687, "step": 16294 }, { "epoch": 2.55, "grad_norm": 29.038151591014906, "learning_rate": 1.180961260668002e-06, "loss": 0.4576, "step": 16295 }, { "epoch": 2.55, "grad_norm": 25.431581194050647, "learning_rate": 1.1801663682364873e-06, "loss": 0.4689, "step": 16296 }, { "epoch": 2.55, "grad_norm": 18.64219839869337, "learning_rate": 1.179371726635883e-06, "loss": 0.4392, "step": 16297 }, { "epoch": 2.55, "grad_norm": 19.215664515886605, "learning_rate": 1.1785773358887854e-06, "loss": 0.4256, "step": 16298 }, { "epoch": 2.55, "grad_norm": 21.038373586691502, "learning_rate": 1.1777831960177898e-06, "loss": 0.4521, "step": 16299 }, { "epoch": 2.55, "grad_norm": 24.7798821311084, "learning_rate": 1.1769893070454774e-06, "loss": 0.4299, "step": 16300 }, { "epoch": 2.55, "grad_norm": 15.645600610730098, "learning_rate": 1.1761956689944288e-06, "loss": 0.4111, "step": 16301 }, { "epoch": 2.55, "grad_norm": 22.118569230200322, "learning_rate": 1.1754022818872123e-06, "loss": 0.3853, "step": 16302 }, { "epoch": 2.55, "grad_norm": 21.478384515151088, "learning_rate": 1.1746091457463927e-06, "loss": 0.5403, "step": 16303 }, { "epoch": 2.55, "grad_norm": 26.74090856179101, "learning_rate": 1.173816260594529e-06, "loss": 0.4452, "step": 16304 }, { "epoch": 2.55, "grad_norm": 17.75884874746568, "learning_rate": 1.1730236264541661e-06, "loss": 0.3696, "step": 16305 }, { "epoch": 2.55, "grad_norm": 24.003414411604094, "learning_rate": 1.1722312433478467e-06, "loss": 0.4272, "step": 16306 }, { "epoch": 2.55, "grad_norm": 23.056585662918472, "learning_rate": 1.1714391112981071e-06, "loss": 0.4952, "step": 16307 }, { "epoch": 2.55, "grad_norm": 17.832641771191643, "learning_rate": 1.170647230327473e-06, "loss": 0.4021, "step": 16308 }, { "epoch": 2.55, "grad_norm": 23.17824702966873, "learning_rate": 1.1698556004584728e-06, "loss": 0.4867, "step": 16309 }, { "epoch": 2.55, "grad_norm": 25.214565479448012, "learning_rate": 1.1690642217136084e-06, "loss": 0.4566, "step": 16310 }, { "epoch": 2.55, "grad_norm": 16.151760605568352, "learning_rate": 1.1682730941153918e-06, "loss": 0.3906, "step": 16311 }, { "epoch": 2.55, "grad_norm": 23.242825475319567, "learning_rate": 1.167482217686322e-06, "loss": 0.4192, "step": 16312 }, { "epoch": 2.55, "grad_norm": 28.212979516681244, "learning_rate": 1.1666915924488931e-06, "loss": 0.4565, "step": 16313 }, { "epoch": 2.55, "grad_norm": 18.35953668076768, "learning_rate": 1.165901218425588e-06, "loss": 0.4163, "step": 16314 }, { "epoch": 2.55, "grad_norm": 22.42526052880745, "learning_rate": 1.1651110956388822e-06, "loss": 0.4955, "step": 16315 }, { "epoch": 2.55, "grad_norm": 24.786270793508038, "learning_rate": 1.164321224111249e-06, "loss": 0.4752, "step": 16316 }, { "epoch": 2.55, "grad_norm": 14.603192277741512, "learning_rate": 1.1635316038651524e-06, "loss": 0.4186, "step": 16317 }, { "epoch": 2.55, "grad_norm": 16.976893413424033, "learning_rate": 1.1627422349230465e-06, "loss": 0.3975, "step": 16318 }, { "epoch": 2.55, "grad_norm": 4.2559601350352265, "learning_rate": 1.161953117307385e-06, "loss": 0.512, "step": 16319 }, { "epoch": 2.55, "grad_norm": 17.43850911033743, "learning_rate": 1.161164251040603e-06, "loss": 0.4227, "step": 16320 }, { "epoch": 2.55, "grad_norm": 31.428002993401954, "learning_rate": 1.1603756361451402e-06, "loss": 0.4544, "step": 16321 }, { "epoch": 2.55, "grad_norm": 20.824160259324255, "learning_rate": 1.1595872726434243e-06, "loss": 0.4135, "step": 16322 }, { "epoch": 2.55, "grad_norm": 24.397785517018377, "learning_rate": 1.158799160557874e-06, "loss": 0.4647, "step": 16323 }, { "epoch": 2.55, "grad_norm": 26.32288204931034, "learning_rate": 1.158011299910905e-06, "loss": 0.4082, "step": 16324 }, { "epoch": 2.55, "grad_norm": 23.529266931174835, "learning_rate": 1.15722369072492e-06, "loss": 0.4936, "step": 16325 }, { "epoch": 2.55, "grad_norm": 14.12718292792, "learning_rate": 1.1564363330223227e-06, "loss": 0.4463, "step": 16326 }, { "epoch": 2.55, "grad_norm": 23.26092624867748, "learning_rate": 1.1556492268255004e-06, "loss": 0.4761, "step": 16327 }, { "epoch": 2.55, "grad_norm": 19.714218955119456, "learning_rate": 1.1548623721568409e-06, "loss": 0.4228, "step": 16328 }, { "epoch": 2.55, "grad_norm": 27.73140780412827, "learning_rate": 1.1540757690387227e-06, "loss": 0.4478, "step": 16329 }, { "epoch": 2.55, "grad_norm": 19.243511439384427, "learning_rate": 1.153289417493515e-06, "loss": 0.4369, "step": 16330 }, { "epoch": 2.55, "grad_norm": 31.92530178660795, "learning_rate": 1.1525033175435796e-06, "loss": 0.4552, "step": 16331 }, { "epoch": 2.55, "grad_norm": 27.64650107558058, "learning_rate": 1.1517174692112742e-06, "loss": 0.4551, "step": 16332 }, { "epoch": 2.55, "grad_norm": 20.307559452401318, "learning_rate": 1.1509318725189477e-06, "loss": 0.4117, "step": 16333 }, { "epoch": 2.55, "grad_norm": 23.166770202347617, "learning_rate": 1.1501465274889457e-06, "loss": 0.4535, "step": 16334 }, { "epoch": 2.55, "grad_norm": 12.961670130533356, "learning_rate": 1.1493614341435954e-06, "loss": 0.4072, "step": 16335 }, { "epoch": 2.55, "grad_norm": 24.85060751398322, "learning_rate": 1.1485765925052294e-06, "loss": 0.4467, "step": 16336 }, { "epoch": 2.55, "grad_norm": 18.073122976941274, "learning_rate": 1.1477920025961664e-06, "loss": 0.4546, "step": 16337 }, { "epoch": 2.55, "grad_norm": 23.775591125620586, "learning_rate": 1.1470076644387229e-06, "loss": 0.4211, "step": 16338 }, { "epoch": 2.55, "grad_norm": 14.727395348309006, "learning_rate": 1.1462235780552023e-06, "loss": 0.3557, "step": 16339 }, { "epoch": 2.55, "grad_norm": 17.4260286182358, "learning_rate": 1.1454397434679022e-06, "loss": 0.3497, "step": 16340 }, { "epoch": 2.55, "grad_norm": 20.70163296055469, "learning_rate": 1.1446561606991158e-06, "loss": 0.4274, "step": 16341 }, { "epoch": 2.55, "grad_norm": 17.421664148602872, "learning_rate": 1.1438728297711288e-06, "loss": 0.3865, "step": 16342 }, { "epoch": 2.55, "grad_norm": 22.885467101816943, "learning_rate": 1.14308975070622e-06, "loss": 0.4025, "step": 16343 }, { "epoch": 2.55, "grad_norm": 21.706163560990117, "learning_rate": 1.1423069235266538e-06, "loss": 0.396, "step": 16344 }, { "epoch": 2.55, "grad_norm": 23.88431262782147, "learning_rate": 1.1415243482546977e-06, "loss": 0.4234, "step": 16345 }, { "epoch": 2.55, "grad_norm": 25.39930066883431, "learning_rate": 1.1407420249126068e-06, "loss": 0.4089, "step": 16346 }, { "epoch": 2.55, "grad_norm": 18.919922317626167, "learning_rate": 1.1399599535226324e-06, "loss": 0.5246, "step": 16347 }, { "epoch": 2.55, "grad_norm": 19.213375302989913, "learning_rate": 1.139178134107014e-06, "loss": 0.4129, "step": 16348 }, { "epoch": 2.55, "grad_norm": 24.411083641510903, "learning_rate": 1.1383965666879847e-06, "loss": 0.4604, "step": 16349 }, { "epoch": 2.55, "grad_norm": 20.891414150234976, "learning_rate": 1.1376152512877725e-06, "loss": 0.4733, "step": 16350 }, { "epoch": 2.55, "grad_norm": 28.763862318604023, "learning_rate": 1.1368341879286004e-06, "loss": 0.5224, "step": 16351 }, { "epoch": 2.55, "grad_norm": 42.95521647029882, "learning_rate": 1.1360533766326765e-06, "loss": 0.4916, "step": 16352 }, { "epoch": 2.55, "grad_norm": 18.685922021809194, "learning_rate": 1.1352728174222128e-06, "loss": 0.5037, "step": 16353 }, { "epoch": 2.55, "grad_norm": 22.051910685021223, "learning_rate": 1.1344925103194005e-06, "loss": 0.4671, "step": 16354 }, { "epoch": 2.55, "grad_norm": 17.850159651312126, "learning_rate": 1.1337124553464384e-06, "loss": 0.4175, "step": 16355 }, { "epoch": 2.55, "grad_norm": 28.218984533203432, "learning_rate": 1.1329326525255046e-06, "loss": 0.4321, "step": 16356 }, { "epoch": 2.55, "grad_norm": 27.66665776750027, "learning_rate": 1.1321531018787801e-06, "loss": 0.4387, "step": 16357 }, { "epoch": 2.56, "grad_norm": 17.757202231934077, "learning_rate": 1.131373803428435e-06, "loss": 0.4483, "step": 16358 }, { "epoch": 2.56, "grad_norm": 25.520247251566996, "learning_rate": 1.1305947571966291e-06, "loss": 0.4666, "step": 16359 }, { "epoch": 2.56, "grad_norm": 19.31917600609384, "learning_rate": 1.1298159632055228e-06, "loss": 0.403, "step": 16360 }, { "epoch": 2.56, "grad_norm": 32.47964946381058, "learning_rate": 1.1290374214772582e-06, "loss": 0.4483, "step": 16361 }, { "epoch": 2.56, "grad_norm": 21.91319271816273, "learning_rate": 1.1282591320339809e-06, "loss": 0.4727, "step": 16362 }, { "epoch": 2.56, "grad_norm": 21.226594818197338, "learning_rate": 1.1274810948978255e-06, "loss": 0.4471, "step": 16363 }, { "epoch": 2.56, "grad_norm": 24.34232487993892, "learning_rate": 1.1267033100909174e-06, "loss": 0.3947, "step": 16364 }, { "epoch": 2.56, "grad_norm": 25.37973193621713, "learning_rate": 1.125925777635375e-06, "loss": 0.4139, "step": 16365 }, { "epoch": 2.56, "grad_norm": 27.15655636806649, "learning_rate": 1.1251484975533123e-06, "loss": 0.4633, "step": 16366 }, { "epoch": 2.56, "grad_norm": 17.152954780469305, "learning_rate": 1.1243714698668363e-06, "loss": 0.4225, "step": 16367 }, { "epoch": 2.56, "grad_norm": 31.86852340582149, "learning_rate": 1.1235946945980435e-06, "loss": 0.4758, "step": 16368 }, { "epoch": 2.56, "grad_norm": 28.23079905957375, "learning_rate": 1.1228181717690234e-06, "loss": 0.4735, "step": 16369 }, { "epoch": 2.56, "grad_norm": 26.843039559854052, "learning_rate": 1.1220419014018613e-06, "loss": 0.4771, "step": 16370 }, { "epoch": 2.56, "grad_norm": 29.149476975734725, "learning_rate": 1.121265883518634e-06, "loss": 0.4881, "step": 16371 }, { "epoch": 2.56, "grad_norm": 18.689965827791, "learning_rate": 1.1204901181414141e-06, "loss": 0.3978, "step": 16372 }, { "epoch": 2.56, "grad_norm": 20.492457200085124, "learning_rate": 1.1197146052922592e-06, "loss": 0.4384, "step": 16373 }, { "epoch": 2.56, "grad_norm": 20.207400283378668, "learning_rate": 1.118939344993225e-06, "loss": 0.4618, "step": 16374 }, { "epoch": 2.56, "grad_norm": 26.166570634928924, "learning_rate": 1.1181643372663608e-06, "loss": 0.448, "step": 16375 }, { "epoch": 2.56, "grad_norm": 17.849882378177295, "learning_rate": 1.1173895821337088e-06, "loss": 0.4165, "step": 16376 }, { "epoch": 2.56, "grad_norm": 17.187464382828306, "learning_rate": 1.1166150796172981e-06, "loss": 0.3819, "step": 16377 }, { "epoch": 2.56, "grad_norm": 28.87890983405053, "learning_rate": 1.115840829739161e-06, "loss": 0.3964, "step": 16378 }, { "epoch": 2.56, "grad_norm": 26.05034665574483, "learning_rate": 1.115066832521311e-06, "loss": 0.4707, "step": 16379 }, { "epoch": 2.56, "grad_norm": 17.667140587899357, "learning_rate": 1.114293087985766e-06, "loss": 0.4569, "step": 16380 }, { "epoch": 2.56, "grad_norm": 24.16680677033514, "learning_rate": 1.1135195961545242e-06, "loss": 0.5102, "step": 16381 }, { "epoch": 2.56, "grad_norm": 16.046267554607923, "learning_rate": 1.1127463570495867e-06, "loss": 0.4581, "step": 16382 }, { "epoch": 2.56, "grad_norm": 18.73502799543351, "learning_rate": 1.111973370692947e-06, "loss": 0.376, "step": 16383 }, { "epoch": 2.56, "grad_norm": 29.466802734973157, "learning_rate": 1.111200637106582e-06, "loss": 0.45, "step": 16384 }, { "epoch": 2.56, "grad_norm": 21.16809120994578, "learning_rate": 1.1104281563124741e-06, "loss": 0.4051, "step": 16385 }, { "epoch": 2.56, "grad_norm": 30.422738142851834, "learning_rate": 1.109655928332587e-06, "loss": 0.531, "step": 16386 }, { "epoch": 2.56, "grad_norm": 19.69068650540431, "learning_rate": 1.1088839531888862e-06, "loss": 0.4703, "step": 16387 }, { "epoch": 2.56, "grad_norm": 14.187114970826515, "learning_rate": 1.108112230903322e-06, "loss": 0.4329, "step": 16388 }, { "epoch": 2.56, "grad_norm": 25.41895960636258, "learning_rate": 1.1073407614978471e-06, "loss": 0.3945, "step": 16389 }, { "epoch": 2.56, "grad_norm": 20.135008072362435, "learning_rate": 1.106569544994397e-06, "loss": 0.3688, "step": 16390 }, { "epoch": 2.56, "grad_norm": 20.880747776876863, "learning_rate": 1.1057985814149063e-06, "loss": 0.3895, "step": 16391 }, { "epoch": 2.56, "grad_norm": 20.17307082745536, "learning_rate": 1.1050278707813033e-06, "loss": 0.4243, "step": 16392 }, { "epoch": 2.56, "grad_norm": 24.12439138133963, "learning_rate": 1.1042574131155048e-06, "loss": 0.4753, "step": 16393 }, { "epoch": 2.56, "grad_norm": 17.372917637979565, "learning_rate": 1.1034872084394187e-06, "loss": 0.3861, "step": 16394 }, { "epoch": 2.56, "grad_norm": 31.82668770310337, "learning_rate": 1.1027172567749523e-06, "loss": 0.47, "step": 16395 }, { "epoch": 2.56, "grad_norm": 24.784327551633144, "learning_rate": 1.101947558144002e-06, "loss": 0.4832, "step": 16396 }, { "epoch": 2.56, "grad_norm": 23.05370878380459, "learning_rate": 1.1011781125684618e-06, "loss": 0.43, "step": 16397 }, { "epoch": 2.56, "grad_norm": 23.402403041692537, "learning_rate": 1.1004089200702072e-06, "loss": 0.4124, "step": 16398 }, { "epoch": 2.56, "grad_norm": 22.81428395768574, "learning_rate": 1.0996399806711167e-06, "loss": 0.4103, "step": 16399 }, { "epoch": 2.56, "grad_norm": 20.168087967249285, "learning_rate": 1.098871294393058e-06, "loss": 0.4625, "step": 16400 }, { "epoch": 2.56, "grad_norm": 16.887338680312844, "learning_rate": 1.0981028612578949e-06, "loss": 0.478, "step": 16401 }, { "epoch": 2.56, "grad_norm": 19.29427435026181, "learning_rate": 1.0973346812874796e-06, "loss": 0.4044, "step": 16402 }, { "epoch": 2.56, "grad_norm": 28.09919110479566, "learning_rate": 1.0965667545036552e-06, "loss": 0.5247, "step": 16403 }, { "epoch": 2.56, "grad_norm": 16.68692232232192, "learning_rate": 1.0957990809282649e-06, "loss": 0.4263, "step": 16404 }, { "epoch": 2.56, "grad_norm": 18.735905248330916, "learning_rate": 1.0950316605831413e-06, "loss": 0.5449, "step": 16405 }, { "epoch": 2.56, "grad_norm": 21.55569018693021, "learning_rate": 1.0942644934901059e-06, "loss": 0.4351, "step": 16406 }, { "epoch": 2.56, "grad_norm": 24.362020268177055, "learning_rate": 1.0934975796709801e-06, "loss": 0.4947, "step": 16407 }, { "epoch": 2.56, "grad_norm": 24.814027515230823, "learning_rate": 1.0927309191475722e-06, "loss": 0.4882, "step": 16408 }, { "epoch": 2.56, "grad_norm": 21.472303019451868, "learning_rate": 1.0919645119416855e-06, "loss": 0.4447, "step": 16409 }, { "epoch": 2.56, "grad_norm": 32.39421401356696, "learning_rate": 1.0911983580751195e-06, "loss": 0.5532, "step": 16410 }, { "epoch": 2.56, "grad_norm": 21.9577344643695, "learning_rate": 1.090432457569659e-06, "loss": 0.456, "step": 16411 }, { "epoch": 2.56, "grad_norm": 22.53984790669875, "learning_rate": 1.0896668104470886e-06, "loss": 0.419, "step": 16412 }, { "epoch": 2.56, "grad_norm": 22.15302915325032, "learning_rate": 1.08890141672918e-06, "loss": 0.4957, "step": 16413 }, { "epoch": 2.56, "grad_norm": 16.084723896009816, "learning_rate": 1.0881362764377046e-06, "loss": 0.4884, "step": 16414 }, { "epoch": 2.56, "grad_norm": 16.732870058300346, "learning_rate": 1.087371389594417e-06, "loss": 0.3784, "step": 16415 }, { "epoch": 2.56, "grad_norm": 18.86533736246151, "learning_rate": 1.0866067562210748e-06, "loss": 0.4726, "step": 16416 }, { "epoch": 2.56, "grad_norm": 22.70157432773767, "learning_rate": 1.0858423763394243e-06, "loss": 0.4598, "step": 16417 }, { "epoch": 2.56, "grad_norm": 19.265653735508337, "learning_rate": 1.085078249971201e-06, "loss": 0.4704, "step": 16418 }, { "epoch": 2.56, "grad_norm": 20.638974170667638, "learning_rate": 1.084314377138136e-06, "loss": 0.459, "step": 16419 }, { "epoch": 2.56, "grad_norm": 27.043482959470058, "learning_rate": 1.0835507578619542e-06, "loss": 0.4217, "step": 16420 }, { "epoch": 2.56, "grad_norm": 30.753017869132258, "learning_rate": 1.0827873921643727e-06, "loss": 0.4773, "step": 16421 }, { "epoch": 2.57, "grad_norm": 20.176999774484784, "learning_rate": 1.0820242800671032e-06, "loss": 0.433, "step": 16422 }, { "epoch": 2.57, "grad_norm": 17.42590269907103, "learning_rate": 1.0812614215918472e-06, "loss": 0.4405, "step": 16423 }, { "epoch": 2.57, "grad_norm": 18.14324710765302, "learning_rate": 1.080498816760296e-06, "loss": 0.4595, "step": 16424 }, { "epoch": 2.57, "grad_norm": 22.12809486888121, "learning_rate": 1.0797364655941411e-06, "loss": 0.4784, "step": 16425 }, { "epoch": 2.57, "grad_norm": 19.407784876578418, "learning_rate": 1.0789743681150656e-06, "loss": 0.476, "step": 16426 }, { "epoch": 2.57, "grad_norm": 16.517716857248157, "learning_rate": 1.0782125243447395e-06, "loss": 0.4441, "step": 16427 }, { "epoch": 2.57, "grad_norm": 18.70820166669645, "learning_rate": 1.077450934304829e-06, "loss": 0.4642, "step": 16428 }, { "epoch": 2.57, "grad_norm": 14.020234554545574, "learning_rate": 1.0766895980169933e-06, "loss": 0.3908, "step": 16429 }, { "epoch": 2.57, "grad_norm": 29.052248154618663, "learning_rate": 1.0759285155028887e-06, "loss": 0.4746, "step": 16430 }, { "epoch": 2.57, "grad_norm": 22.39138326472355, "learning_rate": 1.0751676867841553e-06, "loss": 0.4037, "step": 16431 }, { "epoch": 2.57, "grad_norm": 26.086537281988726, "learning_rate": 1.0744071118824306e-06, "loss": 0.459, "step": 16432 }, { "epoch": 2.57, "grad_norm": 14.999004045840632, "learning_rate": 1.0736467908193471e-06, "loss": 0.4116, "step": 16433 }, { "epoch": 2.57, "grad_norm": 20.219300575071028, "learning_rate": 1.0728867236165264e-06, "loss": 0.5159, "step": 16434 }, { "epoch": 2.57, "grad_norm": 25.054370111489764, "learning_rate": 1.0721269102955866e-06, "loss": 0.4311, "step": 16435 }, { "epoch": 2.57, "grad_norm": 29.298196747019396, "learning_rate": 1.0713673508781353e-06, "loss": 0.4409, "step": 16436 }, { "epoch": 2.57, "grad_norm": 30.486264198143818, "learning_rate": 1.0706080453857714e-06, "loss": 0.4789, "step": 16437 }, { "epoch": 2.57, "grad_norm": 24.735902269193858, "learning_rate": 1.0698489938400914e-06, "loss": 0.5981, "step": 16438 }, { "epoch": 2.57, "grad_norm": 21.34813898558111, "learning_rate": 1.0690901962626843e-06, "loss": 0.4232, "step": 16439 }, { "epoch": 2.57, "grad_norm": 29.209659925968698, "learning_rate": 1.0683316526751253e-06, "loss": 0.4309, "step": 16440 }, { "epoch": 2.57, "grad_norm": 24.62095876998977, "learning_rate": 1.0675733630989904e-06, "loss": 0.4082, "step": 16441 }, { "epoch": 2.57, "grad_norm": 21.230446956732127, "learning_rate": 1.0668153275558424e-06, "loss": 0.4662, "step": 16442 }, { "epoch": 2.57, "grad_norm": 18.593381128739416, "learning_rate": 1.066057546067243e-06, "loss": 0.383, "step": 16443 }, { "epoch": 2.57, "grad_norm": 15.658941749451742, "learning_rate": 1.0653000186547379e-06, "loss": 0.3876, "step": 16444 }, { "epoch": 2.57, "grad_norm": 16.735885907644683, "learning_rate": 1.0645427453398748e-06, "loss": 0.4485, "step": 16445 }, { "epoch": 2.57, "grad_norm": 20.46667836858727, "learning_rate": 1.0637857261441898e-06, "loss": 0.4401, "step": 16446 }, { "epoch": 2.57, "grad_norm": 13.609907039136365, "learning_rate": 1.0630289610892097e-06, "loss": 0.4144, "step": 16447 }, { "epoch": 2.57, "grad_norm": 28.078244097058516, "learning_rate": 1.06227245019646e-06, "loss": 0.5035, "step": 16448 }, { "epoch": 2.57, "grad_norm": 23.47437042310223, "learning_rate": 1.061516193487452e-06, "loss": 0.4938, "step": 16449 }, { "epoch": 2.57, "grad_norm": 23.657574786604457, "learning_rate": 1.060760190983694e-06, "loss": 0.4026, "step": 16450 }, { "epoch": 2.57, "grad_norm": 18.167211752626248, "learning_rate": 1.060004442706688e-06, "loss": 0.4985, "step": 16451 }, { "epoch": 2.57, "grad_norm": 19.32445025377652, "learning_rate": 1.0592489486779267e-06, "loss": 0.5047, "step": 16452 }, { "epoch": 2.57, "grad_norm": 19.524915949851813, "learning_rate": 1.0584937089188941e-06, "loss": 0.4068, "step": 16453 }, { "epoch": 2.57, "grad_norm": 12.75082880311266, "learning_rate": 1.0577387234510684e-06, "loss": 0.4077, "step": 16454 }, { "epoch": 2.57, "grad_norm": 21.761977327557, "learning_rate": 1.0569839922959247e-06, "loss": 0.4417, "step": 16455 }, { "epoch": 2.57, "grad_norm": 26.623101283704802, "learning_rate": 1.0562295154749248e-06, "loss": 0.4138, "step": 16456 }, { "epoch": 2.57, "grad_norm": 16.67506943946318, "learning_rate": 1.0554752930095236e-06, "loss": 0.4246, "step": 16457 }, { "epoch": 2.57, "grad_norm": 16.054041908445058, "learning_rate": 1.054721324921173e-06, "loss": 0.4126, "step": 16458 }, { "epoch": 2.57, "grad_norm": 21.155209200175637, "learning_rate": 1.0539676112313147e-06, "loss": 0.4115, "step": 16459 }, { "epoch": 2.57, "grad_norm": 14.782498422136245, "learning_rate": 1.053214151961386e-06, "loss": 0.4144, "step": 16460 }, { "epoch": 2.57, "grad_norm": 19.62627294384664, "learning_rate": 1.052460947132814e-06, "loss": 0.4104, "step": 16461 }, { "epoch": 2.57, "grad_norm": 23.60887920028195, "learning_rate": 1.0517079967670152e-06, "loss": 0.433, "step": 16462 }, { "epoch": 2.57, "grad_norm": 20.079174148297014, "learning_rate": 1.0509553008854067e-06, "loss": 0.4878, "step": 16463 }, { "epoch": 2.57, "grad_norm": 16.125783898461613, "learning_rate": 1.0502028595093972e-06, "loss": 0.4447, "step": 16464 }, { "epoch": 2.57, "grad_norm": 18.811373283223823, "learning_rate": 1.0494506726603804e-06, "loss": 0.4465, "step": 16465 }, { "epoch": 2.57, "grad_norm": 23.78558384809646, "learning_rate": 1.0486987403597526e-06, "loss": 0.4971, "step": 16466 }, { "epoch": 2.57, "grad_norm": 18.334436267807632, "learning_rate": 1.0479470626288946e-06, "loss": 0.4017, "step": 16467 }, { "epoch": 2.57, "grad_norm": 23.196108079437995, "learning_rate": 1.0471956394891868e-06, "loss": 0.4238, "step": 16468 }, { "epoch": 2.57, "grad_norm": 20.874548661612288, "learning_rate": 1.0464444709619959e-06, "loss": 0.432, "step": 16469 }, { "epoch": 2.57, "grad_norm": 22.229850659234515, "learning_rate": 1.0456935570686866e-06, "loss": 0.4903, "step": 16470 }, { "epoch": 2.57, "grad_norm": 23.426667251866668, "learning_rate": 1.0449428978306164e-06, "loss": 0.4599, "step": 16471 }, { "epoch": 2.57, "grad_norm": 19.81085391456848, "learning_rate": 1.0441924932691293e-06, "loss": 0.4118, "step": 16472 }, { "epoch": 2.57, "grad_norm": 21.729226681851035, "learning_rate": 1.0434423434055719e-06, "loss": 0.4039, "step": 16473 }, { "epoch": 2.57, "grad_norm": 18.3170232443151, "learning_rate": 1.042692448261272e-06, "loss": 0.4412, "step": 16474 }, { "epoch": 2.57, "grad_norm": 25.12976591168797, "learning_rate": 1.041942807857559e-06, "loss": 0.4329, "step": 16475 }, { "epoch": 2.57, "grad_norm": 23.713110554898023, "learning_rate": 1.0411934222157538e-06, "loss": 0.4795, "step": 16476 }, { "epoch": 2.57, "grad_norm": 25.977543597516572, "learning_rate": 1.0404442913571678e-06, "loss": 0.4456, "step": 16477 }, { "epoch": 2.57, "grad_norm": 22.726915575093056, "learning_rate": 1.0396954153031024e-06, "loss": 0.4091, "step": 16478 }, { "epoch": 2.57, "grad_norm": 14.9383974673595, "learning_rate": 1.0389467940748576e-06, "loss": 0.4055, "step": 16479 }, { "epoch": 2.57, "grad_norm": 14.640588296215252, "learning_rate": 1.0381984276937263e-06, "loss": 0.4168, "step": 16480 }, { "epoch": 2.57, "grad_norm": 25.45203231458933, "learning_rate": 1.037450316180989e-06, "loss": 0.4831, "step": 16481 }, { "epoch": 2.57, "grad_norm": 16.595715689208223, "learning_rate": 1.03670245955792e-06, "loss": 0.4195, "step": 16482 }, { "epoch": 2.57, "grad_norm": 19.879232477025074, "learning_rate": 1.03595485784579e-06, "loss": 0.4238, "step": 16483 }, { "epoch": 2.57, "grad_norm": 17.84826947146234, "learning_rate": 1.0352075110658588e-06, "loss": 0.4502, "step": 16484 }, { "epoch": 2.57, "grad_norm": 21.93688077949329, "learning_rate": 1.034460419239387e-06, "loss": 0.4514, "step": 16485 }, { "epoch": 2.58, "grad_norm": 25.199733733303642, "learning_rate": 1.033713582387611e-06, "loss": 0.4522, "step": 16486 }, { "epoch": 2.58, "grad_norm": 15.935862619774106, "learning_rate": 1.032967000531777e-06, "loss": 0.372, "step": 16487 }, { "epoch": 2.58, "grad_norm": 14.415324910507577, "learning_rate": 1.0322206736931152e-06, "loss": 0.404, "step": 16488 }, { "epoch": 2.58, "grad_norm": 14.555851951187535, "learning_rate": 1.0314746018928535e-06, "loss": 0.3731, "step": 16489 }, { "epoch": 2.58, "grad_norm": 19.2027583405399, "learning_rate": 1.0307287851522074e-06, "loss": 0.4799, "step": 16490 }, { "epoch": 2.58, "grad_norm": 32.87259338276433, "learning_rate": 1.0299832234923857e-06, "loss": 0.485, "step": 16491 }, { "epoch": 2.58, "grad_norm": 23.95576761204422, "learning_rate": 1.0292379169345945e-06, "loss": 0.495, "step": 16492 }, { "epoch": 2.58, "grad_norm": 24.967863538615823, "learning_rate": 1.0284928655000303e-06, "loss": 0.4787, "step": 16493 }, { "epoch": 2.58, "grad_norm": 21.18830731282949, "learning_rate": 1.0277480692098796e-06, "loss": 0.4514, "step": 16494 }, { "epoch": 2.58, "grad_norm": 28.02049827103553, "learning_rate": 1.0270035280853275e-06, "loss": 0.4805, "step": 16495 }, { "epoch": 2.58, "grad_norm": 17.312320815030127, "learning_rate": 1.0262592421475436e-06, "loss": 0.4213, "step": 16496 }, { "epoch": 2.58, "grad_norm": 16.471464950329693, "learning_rate": 1.025515211417697e-06, "loss": 0.3936, "step": 16497 }, { "epoch": 2.58, "grad_norm": 22.155199736894907, "learning_rate": 1.0247714359169502e-06, "loss": 0.4295, "step": 16498 }, { "epoch": 2.58, "grad_norm": 25.742556656432157, "learning_rate": 1.0240279156664512e-06, "loss": 0.4644, "step": 16499 }, { "epoch": 2.58, "grad_norm": 29.23617973343494, "learning_rate": 1.0232846506873495e-06, "loss": 0.3922, "step": 16500 }, { "epoch": 2.58, "grad_norm": 16.179230597439556, "learning_rate": 1.0225416410007794e-06, "loss": 0.4816, "step": 16501 }, { "epoch": 2.58, "grad_norm": 16.564784694933927, "learning_rate": 1.021798886627875e-06, "loss": 0.4516, "step": 16502 }, { "epoch": 2.58, "grad_norm": 15.216237414953463, "learning_rate": 1.0210563875897561e-06, "loss": 0.386, "step": 16503 }, { "epoch": 2.58, "grad_norm": 15.690356470878475, "learning_rate": 1.0203141439075415e-06, "loss": 0.4463, "step": 16504 }, { "epoch": 2.58, "grad_norm": 26.02801413877146, "learning_rate": 1.0195721556023409e-06, "loss": 0.4337, "step": 16505 }, { "epoch": 2.58, "grad_norm": 30.748108313299866, "learning_rate": 1.0188304226952562e-06, "loss": 0.4857, "step": 16506 }, { "epoch": 2.58, "grad_norm": 23.634515379408477, "learning_rate": 1.0180889452073772e-06, "loss": 0.4818, "step": 16507 }, { "epoch": 2.58, "grad_norm": 14.562205802827227, "learning_rate": 1.017347723159795e-06, "loss": 0.3984, "step": 16508 }, { "epoch": 2.58, "grad_norm": 22.995305402438024, "learning_rate": 1.0166067565735881e-06, "loss": 0.3964, "step": 16509 }, { "epoch": 2.58, "grad_norm": 20.219990642106335, "learning_rate": 1.015866045469832e-06, "loss": 0.4352, "step": 16510 }, { "epoch": 2.58, "grad_norm": 23.647476794914745, "learning_rate": 1.0151255898695911e-06, "loss": 0.4233, "step": 16511 }, { "epoch": 2.58, "grad_norm": 30.48425150796893, "learning_rate": 1.0143853897939193e-06, "loss": 0.4566, "step": 16512 }, { "epoch": 2.58, "grad_norm": 23.278968117401885, "learning_rate": 1.013645445263871e-06, "loss": 0.5062, "step": 16513 }, { "epoch": 2.58, "grad_norm": 23.820874956646325, "learning_rate": 1.012905756300492e-06, "loss": 0.4455, "step": 16514 }, { "epoch": 2.58, "grad_norm": 23.608701171827537, "learning_rate": 1.0121663229248145e-06, "loss": 0.4525, "step": 16515 }, { "epoch": 2.58, "grad_norm": 24.581410107235467, "learning_rate": 1.0114271451578684e-06, "loss": 0.3984, "step": 16516 }, { "epoch": 2.58, "grad_norm": 31.58511180775014, "learning_rate": 1.0106882230206749e-06, "loss": 0.5215, "step": 16517 }, { "epoch": 2.58, "grad_norm": 33.65416585690738, "learning_rate": 1.0099495565342532e-06, "loss": 0.4343, "step": 16518 }, { "epoch": 2.58, "grad_norm": 23.51411665074277, "learning_rate": 1.0092111457196041e-06, "loss": 0.4125, "step": 16519 }, { "epoch": 2.58, "grad_norm": 17.70205503767291, "learning_rate": 1.0084729905977332e-06, "loss": 0.4557, "step": 16520 }, { "epoch": 2.58, "grad_norm": 17.814874729169205, "learning_rate": 1.0077350911896278e-06, "loss": 0.4395, "step": 16521 }, { "epoch": 2.58, "grad_norm": 25.97192915817965, "learning_rate": 1.006997447516276e-06, "loss": 0.4404, "step": 16522 }, { "epoch": 2.58, "grad_norm": 20.230180421208296, "learning_rate": 1.0062600595986582e-06, "loss": 0.4285, "step": 16523 }, { "epoch": 2.58, "grad_norm": 23.677279918512994, "learning_rate": 1.0055229274577417e-06, "loss": 0.4765, "step": 16524 }, { "epoch": 2.58, "grad_norm": 21.578095830390765, "learning_rate": 1.0047860511144937e-06, "loss": 0.4114, "step": 16525 }, { "epoch": 2.58, "grad_norm": 17.861821633638986, "learning_rate": 1.004049430589865e-06, "loss": 0.507, "step": 16526 }, { "epoch": 2.58, "grad_norm": 18.770514644063674, "learning_rate": 1.0033130659048119e-06, "loss": 0.3952, "step": 16527 }, { "epoch": 2.58, "grad_norm": 13.791023707120715, "learning_rate": 1.002576957080269e-06, "loss": 0.4247, "step": 16528 }, { "epoch": 2.58, "grad_norm": 21.604076591004727, "learning_rate": 1.0018411041371756e-06, "loss": 0.4734, "step": 16529 }, { "epoch": 2.58, "grad_norm": 28.33773838349109, "learning_rate": 1.001105507096457e-06, "loss": 0.4376, "step": 16530 }, { "epoch": 2.58, "grad_norm": 20.510467834047187, "learning_rate": 1.0003701659790344e-06, "loss": 0.4145, "step": 16531 }, { "epoch": 2.58, "grad_norm": 22.457611751966294, "learning_rate": 9.996350808058175e-07, "loss": 0.3987, "step": 16532 }, { "epoch": 2.58, "grad_norm": 28.852800145939398, "learning_rate": 9.989002515977154e-07, "loss": 0.4239, "step": 16533 }, { "epoch": 2.58, "grad_norm": 16.30892513875, "learning_rate": 9.981656783756255e-07, "loss": 0.4467, "step": 16534 }, { "epoch": 2.58, "grad_norm": 16.044757461402007, "learning_rate": 9.974313611604358e-07, "loss": 0.4423, "step": 16535 }, { "epoch": 2.58, "grad_norm": 18.533637408461708, "learning_rate": 9.96697299973034e-07, "loss": 0.4163, "step": 16536 }, { "epoch": 2.58, "grad_norm": 20.661578496027914, "learning_rate": 9.959634948342923e-07, "loss": 0.4391, "step": 16537 }, { "epoch": 2.58, "grad_norm": 18.70728507022078, "learning_rate": 9.952299457650805e-07, "loss": 0.4423, "step": 16538 }, { "epoch": 2.58, "grad_norm": 18.358198417946898, "learning_rate": 9.944966527862644e-07, "loss": 0.432, "step": 16539 }, { "epoch": 2.58, "grad_norm": 18.589319925929807, "learning_rate": 9.93763615918696e-07, "loss": 0.4515, "step": 16540 }, { "epoch": 2.58, "grad_norm": 21.80563146201027, "learning_rate": 9.930308351832185e-07, "loss": 0.4338, "step": 16541 }, { "epoch": 2.58, "grad_norm": 20.925555980125676, "learning_rate": 9.922983106006766e-07, "loss": 0.4576, "step": 16542 }, { "epoch": 2.58, "grad_norm": 17.65480177265285, "learning_rate": 9.915660421919027e-07, "loss": 0.4294, "step": 16543 }, { "epoch": 2.58, "grad_norm": 20.348788224983505, "learning_rate": 9.908340299777208e-07, "loss": 0.4625, "step": 16544 }, { "epoch": 2.58, "grad_norm": 21.468527006326607, "learning_rate": 9.901022739789468e-07, "loss": 0.4469, "step": 16545 }, { "epoch": 2.58, "grad_norm": 24.437181231061157, "learning_rate": 9.893707742163926e-07, "loss": 0.4391, "step": 16546 }, { "epoch": 2.58, "grad_norm": 19.26344859691854, "learning_rate": 9.886395307108643e-07, "loss": 0.3868, "step": 16547 }, { "epoch": 2.58, "grad_norm": 21.340548640366055, "learning_rate": 9.87908543483157e-07, "loss": 0.425, "step": 16548 }, { "epoch": 2.58, "grad_norm": 23.25846971871695, "learning_rate": 9.871778125540587e-07, "loss": 0.4199, "step": 16549 }, { "epoch": 2.59, "grad_norm": 15.34363216250013, "learning_rate": 9.864473379443495e-07, "loss": 0.4126, "step": 16550 }, { "epoch": 2.59, "grad_norm": 22.553022229353086, "learning_rate": 9.85717119674806e-07, "loss": 0.4906, "step": 16551 }, { "epoch": 2.59, "grad_norm": 28.276992042104844, "learning_rate": 9.84987157766195e-07, "loss": 0.4179, "step": 16552 }, { "epoch": 2.59, "grad_norm": 23.807934996806832, "learning_rate": 9.842574522392744e-07, "loss": 0.4719, "step": 16553 }, { "epoch": 2.59, "grad_norm": 15.923307514886028, "learning_rate": 9.835280031147999e-07, "loss": 0.3939, "step": 16554 }, { "epoch": 2.59, "grad_norm": 17.91793874800877, "learning_rate": 9.827988104135122e-07, "loss": 0.4043, "step": 16555 }, { "epoch": 2.59, "grad_norm": 18.212522157107077, "learning_rate": 9.82069874156154e-07, "loss": 0.4205, "step": 16556 }, { "epoch": 2.59, "grad_norm": 29.75233433735346, "learning_rate": 9.81341194363451e-07, "loss": 0.4767, "step": 16557 }, { "epoch": 2.59, "grad_norm": 24.046657248955565, "learning_rate": 9.806127710561274e-07, "loss": 0.459, "step": 16558 }, { "epoch": 2.59, "grad_norm": 22.186223392967094, "learning_rate": 9.798846042549037e-07, "loss": 0.419, "step": 16559 }, { "epoch": 2.59, "grad_norm": 19.39623698390906, "learning_rate": 9.79156693980483e-07, "loss": 0.4234, "step": 16560 }, { "epoch": 2.59, "grad_norm": 24.044912794215403, "learning_rate": 9.784290402535713e-07, "loss": 0.4487, "step": 16561 }, { "epoch": 2.59, "grad_norm": 19.668414027335913, "learning_rate": 9.777016430948572e-07, "loss": 0.4626, "step": 16562 }, { "epoch": 2.59, "grad_norm": 29.21101457331106, "learning_rate": 9.76974502525031e-07, "loss": 0.4468, "step": 16563 }, { "epoch": 2.59, "grad_norm": 16.63610310433114, "learning_rate": 9.76247618564774e-07, "loss": 0.4276, "step": 16564 }, { "epoch": 2.59, "grad_norm": 16.28588605642223, "learning_rate": 9.755209912347552e-07, "loss": 0.4021, "step": 16565 }, { "epoch": 2.59, "grad_norm": 21.22000810491783, "learning_rate": 9.747946205556391e-07, "loss": 0.4279, "step": 16566 }, { "epoch": 2.59, "grad_norm": 38.557370057164384, "learning_rate": 9.740685065480837e-07, "loss": 0.4808, "step": 16567 }, { "epoch": 2.59, "grad_norm": 18.66066391805108, "learning_rate": 9.733426492327425e-07, "loss": 0.4066, "step": 16568 }, { "epoch": 2.59, "grad_norm": 17.276100867352753, "learning_rate": 9.726170486302554e-07, "loss": 0.4234, "step": 16569 }, { "epoch": 2.59, "grad_norm": 22.07153241997382, "learning_rate": 9.718917047612575e-07, "loss": 0.4925, "step": 16570 }, { "epoch": 2.59, "grad_norm": 21.499364029133467, "learning_rate": 9.711666176463775e-07, "loss": 0.4306, "step": 16571 }, { "epoch": 2.59, "grad_norm": 30.003989369660047, "learning_rate": 9.704417873062366e-07, "loss": 0.4829, "step": 16572 }, { "epoch": 2.59, "grad_norm": 17.47555246960147, "learning_rate": 9.697172137614518e-07, "loss": 0.4287, "step": 16573 }, { "epoch": 2.59, "grad_norm": 24.756652362469854, "learning_rate": 9.689928970326268e-07, "loss": 0.4669, "step": 16574 }, { "epoch": 2.59, "grad_norm": 14.175715192197819, "learning_rate": 9.682688371403593e-07, "loss": 0.4344, "step": 16575 }, { "epoch": 2.59, "grad_norm": 19.885279892447244, "learning_rate": 9.675450341052427e-07, "loss": 0.4255, "step": 16576 }, { "epoch": 2.59, "grad_norm": 23.888077227686253, "learning_rate": 9.668214879478622e-07, "loss": 0.5191, "step": 16577 }, { "epoch": 2.59, "grad_norm": 29.036166649908623, "learning_rate": 9.66098198688795e-07, "loss": 0.3896, "step": 16578 }, { "epoch": 2.59, "grad_norm": 25.648444942899722, "learning_rate": 9.65375166348609e-07, "loss": 0.4077, "step": 16579 }, { "epoch": 2.59, "grad_norm": 32.028053514972335, "learning_rate": 9.646523909478677e-07, "loss": 0.4307, "step": 16580 }, { "epoch": 2.59, "grad_norm": 21.693729862502174, "learning_rate": 9.63929872507129e-07, "loss": 0.4196, "step": 16581 }, { "epoch": 2.59, "grad_norm": 24.124346025032906, "learning_rate": 9.632076110469368e-07, "loss": 0.4165, "step": 16582 }, { "epoch": 2.59, "grad_norm": 18.81676246507038, "learning_rate": 9.624856065878351e-07, "loss": 0.4586, "step": 16583 }, { "epoch": 2.59, "grad_norm": 28.342226783993983, "learning_rate": 9.617638591503542e-07, "loss": 0.4957, "step": 16584 }, { "epoch": 2.59, "grad_norm": 18.93790422533783, "learning_rate": 9.61042368755023e-07, "loss": 0.4046, "step": 16585 }, { "epoch": 2.59, "grad_norm": 18.6169104061244, "learning_rate": 9.603211354223597e-07, "loss": 0.385, "step": 16586 }, { "epoch": 2.59, "grad_norm": 21.741579402639957, "learning_rate": 9.596001591728743e-07, "loss": 0.4956, "step": 16587 }, { "epoch": 2.59, "grad_norm": 30.306177006793096, "learning_rate": 9.58879440027074e-07, "loss": 0.4563, "step": 16588 }, { "epoch": 2.59, "grad_norm": 20.765973896916027, "learning_rate": 9.5815897800545e-07, "loss": 0.3972, "step": 16589 }, { "epoch": 2.59, "grad_norm": 23.11424219462224, "learning_rate": 9.574387731284984e-07, "loss": 0.4215, "step": 16590 }, { "epoch": 2.59, "grad_norm": 19.610770729820203, "learning_rate": 9.56718825416697e-07, "loss": 0.3933, "step": 16591 }, { "epoch": 2.59, "grad_norm": 33.816859833278606, "learning_rate": 9.559991348905218e-07, "loss": 0.468, "step": 16592 }, { "epoch": 2.59, "grad_norm": 20.852515207095763, "learning_rate": 9.55279701570442e-07, "loss": 0.4243, "step": 16593 }, { "epoch": 2.59, "grad_norm": 29.141677976509175, "learning_rate": 9.545605254769164e-07, "loss": 0.526, "step": 16594 }, { "epoch": 2.59, "grad_norm": 18.216068097417466, "learning_rate": 9.538416066303956e-07, "loss": 0.4218, "step": 16595 }, { "epoch": 2.59, "grad_norm": 18.26951078096083, "learning_rate": 9.531229450513291e-07, "loss": 0.4137, "step": 16596 }, { "epoch": 2.59, "grad_norm": 19.919263688774954, "learning_rate": 9.524045407601534e-07, "loss": 0.4682, "step": 16597 }, { "epoch": 2.59, "grad_norm": 27.113522541581116, "learning_rate": 9.516863937773013e-07, "loss": 0.4943, "step": 16598 }, { "epoch": 2.59, "grad_norm": 25.228212863150578, "learning_rate": 9.509685041231953e-07, "loss": 0.4431, "step": 16599 }, { "epoch": 2.59, "grad_norm": 15.440371437438177, "learning_rate": 9.502508718182491e-07, "loss": 0.4354, "step": 16600 }, { "epoch": 2.59, "grad_norm": 27.165188017538007, "learning_rate": 9.49533496882874e-07, "loss": 0.5733, "step": 16601 }, { "epoch": 2.59, "grad_norm": 26.408273239894697, "learning_rate": 9.488163793374749e-07, "loss": 0.4491, "step": 16602 }, { "epoch": 2.59, "grad_norm": 18.26673032118206, "learning_rate": 9.480995192024423e-07, "loss": 0.4764, "step": 16603 }, { "epoch": 2.59, "grad_norm": 17.712476225159275, "learning_rate": 9.473829164981629e-07, "loss": 0.4284, "step": 16604 }, { "epoch": 2.59, "grad_norm": 15.149277999189202, "learning_rate": 9.466665712450174e-07, "loss": 0.3953, "step": 16605 }, { "epoch": 2.59, "grad_norm": 17.17356222773973, "learning_rate": 9.459504834633804e-07, "loss": 0.3883, "step": 16606 }, { "epoch": 2.59, "grad_norm": 43.14834145489166, "learning_rate": 9.452346531736134e-07, "loss": 0.4145, "step": 16607 }, { "epoch": 2.59, "grad_norm": 23.048702268416047, "learning_rate": 9.44519080396078e-07, "loss": 0.4732, "step": 16608 }, { "epoch": 2.59, "grad_norm": 20.85824593299342, "learning_rate": 9.438037651511201e-07, "loss": 0.4093, "step": 16609 }, { "epoch": 2.59, "grad_norm": 25.90883123655177, "learning_rate": 9.430887074590855e-07, "loss": 0.4416, "step": 16610 }, { "epoch": 2.59, "grad_norm": 19.54133176457306, "learning_rate": 9.423739073403126e-07, "loss": 0.4391, "step": 16611 }, { "epoch": 2.59, "grad_norm": 19.732579868080226, "learning_rate": 9.416593648151251e-07, "loss": 0.4338, "step": 16612 }, { "epoch": 2.59, "grad_norm": 15.647606183870503, "learning_rate": 9.409450799038478e-07, "loss": 0.3797, "step": 16613 }, { "epoch": 2.6, "grad_norm": 30.206813752770014, "learning_rate": 9.40231052626791e-07, "loss": 0.4625, "step": 16614 }, { "epoch": 2.6, "grad_norm": 32.31348099962588, "learning_rate": 9.395172830042653e-07, "loss": 0.5482, "step": 16615 }, { "epoch": 2.6, "grad_norm": 21.6672429337497, "learning_rate": 9.388037710565667e-07, "loss": 0.5197, "step": 16616 }, { "epoch": 2.6, "grad_norm": 22.326256056476627, "learning_rate": 9.380905168039878e-07, "loss": 0.4341, "step": 16617 }, { "epoch": 2.6, "grad_norm": 23.368618083106085, "learning_rate": 9.373775202668156e-07, "loss": 0.4431, "step": 16618 }, { "epoch": 2.6, "grad_norm": 30.806549561635077, "learning_rate": 9.36664781465324e-07, "loss": 0.6125, "step": 16619 }, { "epoch": 2.6, "grad_norm": 21.9371326523099, "learning_rate": 9.359523004197835e-07, "loss": 0.414, "step": 16620 }, { "epoch": 2.6, "grad_norm": 15.144667803148963, "learning_rate": 9.352400771504566e-07, "loss": 0.3924, "step": 16621 }, { "epoch": 2.6, "grad_norm": 20.327806282623662, "learning_rate": 9.345281116775995e-07, "loss": 0.4246, "step": 16622 }, { "epoch": 2.6, "grad_norm": 18.115795528223863, "learning_rate": 9.338164040214614e-07, "loss": 0.4323, "step": 16623 }, { "epoch": 2.6, "grad_norm": 22.947981052218893, "learning_rate": 9.331049542022818e-07, "loss": 0.4235, "step": 16624 }, { "epoch": 2.6, "grad_norm": 18.063613919127054, "learning_rate": 9.323937622402912e-07, "loss": 0.4602, "step": 16625 }, { "epoch": 2.6, "grad_norm": 28.645485867676015, "learning_rate": 9.316828281557178e-07, "loss": 0.4192, "step": 16626 }, { "epoch": 2.6, "grad_norm": 18.99592090255853, "learning_rate": 9.309721519687809e-07, "loss": 0.4456, "step": 16627 }, { "epoch": 2.6, "grad_norm": 18.749788272377494, "learning_rate": 9.30261733699691e-07, "loss": 0.4229, "step": 16628 }, { "epoch": 2.6, "grad_norm": 14.854005343297715, "learning_rate": 9.295515733686511e-07, "loss": 0.456, "step": 16629 }, { "epoch": 2.6, "grad_norm": 20.565194515023208, "learning_rate": 9.28841670995857e-07, "loss": 0.451, "step": 16630 }, { "epoch": 2.6, "grad_norm": 21.278853611314297, "learning_rate": 9.281320266015015e-07, "loss": 0.4267, "step": 16631 }, { "epoch": 2.6, "grad_norm": 26.674816431988763, "learning_rate": 9.274226402057651e-07, "loss": 0.4463, "step": 16632 }, { "epoch": 2.6, "grad_norm": 18.554797704469767, "learning_rate": 9.267135118288184e-07, "loss": 0.3773, "step": 16633 }, { "epoch": 2.6, "grad_norm": 16.98594642991937, "learning_rate": 9.26004641490833e-07, "loss": 0.4135, "step": 16634 }, { "epoch": 2.6, "grad_norm": 16.150973173064106, "learning_rate": 9.252960292119661e-07, "loss": 0.4367, "step": 16635 }, { "epoch": 2.6, "grad_norm": 25.336592091477296, "learning_rate": 9.24587675012375e-07, "loss": 0.4626, "step": 16636 }, { "epoch": 2.6, "grad_norm": 29.170722339989087, "learning_rate": 9.238795789122002e-07, "loss": 0.4563, "step": 16637 }, { "epoch": 2.6, "grad_norm": 20.588816872444657, "learning_rate": 9.231717409315788e-07, "loss": 0.4561, "step": 16638 }, { "epoch": 2.6, "grad_norm": 46.91705659845594, "learning_rate": 9.224641610906437e-07, "loss": 0.4929, "step": 16639 }, { "epoch": 2.6, "grad_norm": 19.008658256639237, "learning_rate": 9.2175683940952e-07, "loss": 0.4021, "step": 16640 }, { "epoch": 2.6, "grad_norm": 25.727204710112144, "learning_rate": 9.210497759083193e-07, "loss": 0.4594, "step": 16641 }, { "epoch": 2.6, "grad_norm": 35.764165740201804, "learning_rate": 9.203429706071531e-07, "loss": 0.5151, "step": 16642 }, { "epoch": 2.6, "grad_norm": 14.089536312398872, "learning_rate": 9.196364235261202e-07, "loss": 0.4105, "step": 16643 }, { "epoch": 2.6, "grad_norm": 17.135423963964545, "learning_rate": 9.189301346853163e-07, "loss": 0.4699, "step": 16644 }, { "epoch": 2.6, "grad_norm": 20.064086434663675, "learning_rate": 9.182241041048267e-07, "loss": 0.5064, "step": 16645 }, { "epoch": 2.6, "grad_norm": 19.879351289909327, "learning_rate": 9.175183318047298e-07, "loss": 0.3794, "step": 16646 }, { "epoch": 2.6, "grad_norm": 21.104982016410478, "learning_rate": 9.168128178051005e-07, "loss": 0.4872, "step": 16647 }, { "epoch": 2.6, "grad_norm": 20.443144078723233, "learning_rate": 9.161075621259997e-07, "loss": 0.434, "step": 16648 }, { "epoch": 2.6, "grad_norm": 25.691266606296317, "learning_rate": 9.154025647874875e-07, "loss": 0.4629, "step": 16649 }, { "epoch": 2.6, "grad_norm": 21.55391299331185, "learning_rate": 9.146978258096107e-07, "loss": 0.4259, "step": 16650 }, { "epoch": 2.6, "grad_norm": 27.076821679404723, "learning_rate": 9.139933452124117e-07, "loss": 0.4911, "step": 16651 }, { "epoch": 2.6, "grad_norm": 21.259026280708042, "learning_rate": 9.132891230159302e-07, "loss": 0.4465, "step": 16652 }, { "epoch": 2.6, "grad_norm": 18.937368933457027, "learning_rate": 9.125851592401891e-07, "loss": 0.4729, "step": 16653 }, { "epoch": 2.6, "grad_norm": 14.395639017717112, "learning_rate": 9.11881453905209e-07, "loss": 0.4109, "step": 16654 }, { "epoch": 2.6, "grad_norm": 17.980382389952663, "learning_rate": 9.111780070310038e-07, "loss": 0.3972, "step": 16655 }, { "epoch": 2.6, "grad_norm": 22.66773722962771, "learning_rate": 9.104748186375811e-07, "loss": 0.4461, "step": 16656 }, { "epoch": 2.6, "grad_norm": 20.951502970467896, "learning_rate": 9.097718887449381e-07, "loss": 0.5012, "step": 16657 }, { "epoch": 2.6, "grad_norm": 23.418857307534378, "learning_rate": 9.090692173730619e-07, "loss": 0.455, "step": 16658 }, { "epoch": 2.6, "grad_norm": 24.002348904105794, "learning_rate": 9.083668045419391e-07, "loss": 0.4388, "step": 16659 }, { "epoch": 2.6, "grad_norm": 24.089457186991297, "learning_rate": 9.076646502715469e-07, "loss": 0.4302, "step": 16660 }, { "epoch": 2.6, "grad_norm": 21.07289727410144, "learning_rate": 9.069627545818549e-07, "loss": 0.4516, "step": 16661 }, { "epoch": 2.6, "grad_norm": 26.221881132746827, "learning_rate": 9.062611174928226e-07, "loss": 0.4544, "step": 16662 }, { "epoch": 2.6, "grad_norm": 22.094976478807144, "learning_rate": 9.05559739024402e-07, "loss": 0.3801, "step": 16663 }, { "epoch": 2.6, "grad_norm": 25.400617592166768, "learning_rate": 9.048586191965437e-07, "loss": 0.4404, "step": 16664 }, { "epoch": 2.6, "grad_norm": 20.691117029543463, "learning_rate": 9.04157758029186e-07, "loss": 0.4639, "step": 16665 }, { "epoch": 2.6, "grad_norm": 27.536180260436744, "learning_rate": 9.0345715554226e-07, "loss": 0.4018, "step": 16666 }, { "epoch": 2.6, "grad_norm": 22.50435428209808, "learning_rate": 9.027568117556928e-07, "loss": 0.4154, "step": 16667 }, { "epoch": 2.6, "grad_norm": 24.243344416720603, "learning_rate": 9.020567266893976e-07, "loss": 0.485, "step": 16668 }, { "epoch": 2.6, "grad_norm": 18.560779535954246, "learning_rate": 9.013569003632894e-07, "loss": 0.4223, "step": 16669 }, { "epoch": 2.6, "grad_norm": 20.521239161875705, "learning_rate": 9.006573327972667e-07, "loss": 0.4801, "step": 16670 }, { "epoch": 2.6, "grad_norm": 18.434193194990357, "learning_rate": 8.99958024011226e-07, "loss": 0.4201, "step": 16671 }, { "epoch": 2.6, "grad_norm": 23.114078630467418, "learning_rate": 8.992589740250579e-07, "loss": 0.4696, "step": 16672 }, { "epoch": 2.6, "grad_norm": 23.288510534533547, "learning_rate": 8.985601828586399e-07, "loss": 0.4432, "step": 16673 }, { "epoch": 2.6, "grad_norm": 21.627993637626343, "learning_rate": 8.978616505318461e-07, "loss": 0.4122, "step": 16674 }, { "epoch": 2.6, "grad_norm": 15.652127299895884, "learning_rate": 8.971633770645416e-07, "loss": 0.3973, "step": 16675 }, { "epoch": 2.6, "grad_norm": 32.2956595079074, "learning_rate": 8.964653624765874e-07, "loss": 0.5038, "step": 16676 }, { "epoch": 2.6, "grad_norm": 29.932339219471015, "learning_rate": 8.957676067878307e-07, "loss": 0.5604, "step": 16677 }, { "epoch": 2.61, "grad_norm": 20.154867314072508, "learning_rate": 8.950701100181203e-07, "loss": 0.4561, "step": 16678 }, { "epoch": 2.61, "grad_norm": 21.638338017033707, "learning_rate": 8.943728721872868e-07, "loss": 0.4534, "step": 16679 }, { "epoch": 2.61, "grad_norm": 20.70356290678026, "learning_rate": 8.936758933151623e-07, "loss": 0.4547, "step": 16680 }, { "epoch": 2.61, "grad_norm": 15.97695713721677, "learning_rate": 8.92979173421571e-07, "loss": 0.3834, "step": 16681 }, { "epoch": 2.61, "grad_norm": 23.435849945476544, "learning_rate": 8.922827125263234e-07, "loss": 0.3932, "step": 16682 }, { "epoch": 2.61, "grad_norm": 27.428948725196182, "learning_rate": 8.91586510649226e-07, "loss": 0.473, "step": 16683 }, { "epoch": 2.61, "grad_norm": 23.52376543878768, "learning_rate": 8.90890567810081e-07, "loss": 0.5012, "step": 16684 }, { "epoch": 2.61, "grad_norm": 27.61751614527499, "learning_rate": 8.901948840286789e-07, "loss": 0.3716, "step": 16685 }, { "epoch": 2.61, "grad_norm": 19.86593807010374, "learning_rate": 8.894994593248063e-07, "loss": 0.3881, "step": 16686 }, { "epoch": 2.61, "grad_norm": 14.410364220851308, "learning_rate": 8.888042937182406e-07, "loss": 0.4258, "step": 16687 }, { "epoch": 2.61, "grad_norm": 19.115935050366904, "learning_rate": 8.881093872287483e-07, "loss": 0.4096, "step": 16688 }, { "epoch": 2.61, "grad_norm": 20.652318385530542, "learning_rate": 8.874147398760957e-07, "loss": 0.4372, "step": 16689 }, { "epoch": 2.61, "grad_norm": 30.22928360546812, "learning_rate": 8.867203516800382e-07, "loss": 0.4408, "step": 16690 }, { "epoch": 2.61, "grad_norm": 22.39774081527456, "learning_rate": 8.860262226603234e-07, "loss": 0.3935, "step": 16691 }, { "epoch": 2.61, "grad_norm": 22.919712059188633, "learning_rate": 8.853323528366886e-07, "loss": 0.4538, "step": 16692 }, { "epoch": 2.61, "grad_norm": 35.905719633642555, "learning_rate": 8.846387422288704e-07, "loss": 0.4302, "step": 16693 }, { "epoch": 2.61, "grad_norm": 21.033654368842818, "learning_rate": 8.839453908565965e-07, "loss": 0.4794, "step": 16694 }, { "epoch": 2.61, "grad_norm": 38.43862449892962, "learning_rate": 8.832522987395798e-07, "loss": 0.4356, "step": 16695 }, { "epoch": 2.61, "grad_norm": 23.464374741397606, "learning_rate": 8.825594658975367e-07, "loss": 0.4589, "step": 16696 }, { "epoch": 2.61, "grad_norm": 17.422209357572804, "learning_rate": 8.818668923501683e-07, "loss": 0.3923, "step": 16697 }, { "epoch": 2.61, "grad_norm": 22.09562721937399, "learning_rate": 8.811745781171699e-07, "loss": 0.4206, "step": 16698 }, { "epoch": 2.61, "grad_norm": 19.442060279077392, "learning_rate": 8.804825232182345e-07, "loss": 0.4029, "step": 16699 }, { "epoch": 2.61, "grad_norm": 21.58317273293, "learning_rate": 8.7979072767304e-07, "loss": 0.474, "step": 16700 }, { "epoch": 2.61, "grad_norm": 34.61315114537455, "learning_rate": 8.790991915012636e-07, "loss": 0.5113, "step": 16701 }, { "epoch": 2.61, "grad_norm": 17.743257077126724, "learning_rate": 8.784079147225688e-07, "loss": 0.5002, "step": 16702 }, { "epoch": 2.61, "grad_norm": 13.50267092235498, "learning_rate": 8.777168973566186e-07, "loss": 0.353, "step": 16703 }, { "epoch": 2.61, "grad_norm": 20.95125831094712, "learning_rate": 8.770261394230617e-07, "loss": 0.4314, "step": 16704 }, { "epoch": 2.61, "grad_norm": 18.74905084830219, "learning_rate": 8.763356409415447e-07, "loss": 0.4074, "step": 16705 }, { "epoch": 2.61, "grad_norm": 15.579036870571148, "learning_rate": 8.756454019317063e-07, "loss": 0.4278, "step": 16706 }, { "epoch": 2.61, "grad_norm": 31.87071763962313, "learning_rate": 8.749554224131751e-07, "loss": 0.553, "step": 16707 }, { "epoch": 2.61, "grad_norm": 17.60790887329473, "learning_rate": 8.742657024055723e-07, "loss": 0.4269, "step": 16708 }, { "epoch": 2.61, "grad_norm": 26.285916786115394, "learning_rate": 8.735762419285143e-07, "loss": 0.4468, "step": 16709 }, { "epoch": 2.61, "grad_norm": 20.82882622284602, "learning_rate": 8.728870410016099e-07, "loss": 0.4227, "step": 16710 }, { "epoch": 2.61, "grad_norm": 44.476794032022404, "learning_rate": 8.721980996444601e-07, "loss": 0.4647, "step": 16711 }, { "epoch": 2.61, "grad_norm": 21.05022180011697, "learning_rate": 8.715094178766559e-07, "loss": 0.4309, "step": 16712 }, { "epoch": 2.61, "grad_norm": 24.75239388612179, "learning_rate": 8.708209957177826e-07, "loss": 0.4265, "step": 16713 }, { "epoch": 2.61, "grad_norm": 34.79433308889338, "learning_rate": 8.701328331874204e-07, "loss": 0.5209, "step": 16714 }, { "epoch": 2.61, "grad_norm": 17.99271583940879, "learning_rate": 8.694449303051411e-07, "loss": 0.4841, "step": 16715 }, { "epoch": 2.61, "grad_norm": 19.185274258323, "learning_rate": 8.68757287090507e-07, "loss": 0.4563, "step": 16716 }, { "epoch": 2.61, "grad_norm": 21.957059233477462, "learning_rate": 8.680699035630713e-07, "loss": 0.4054, "step": 16717 }, { "epoch": 2.61, "grad_norm": 21.60780446819741, "learning_rate": 8.673827797423862e-07, "loss": 0.4715, "step": 16718 }, { "epoch": 2.61, "grad_norm": 27.01343319652089, "learning_rate": 8.666959156479938e-07, "loss": 0.4525, "step": 16719 }, { "epoch": 2.61, "grad_norm": 19.50782784891804, "learning_rate": 8.660093112994261e-07, "loss": 0.4836, "step": 16720 }, { "epoch": 2.61, "grad_norm": 20.246002297155318, "learning_rate": 8.653229667162111e-07, "loss": 0.4106, "step": 16721 }, { "epoch": 2.61, "grad_norm": 14.868173611141453, "learning_rate": 8.646368819178652e-07, "loss": 0.3736, "step": 16722 }, { "epoch": 2.61, "grad_norm": 24.618808935065324, "learning_rate": 8.639510569239029e-07, "loss": 0.4321, "step": 16723 }, { "epoch": 2.61, "grad_norm": 28.00095475433834, "learning_rate": 8.632654917538297e-07, "loss": 0.4507, "step": 16724 }, { "epoch": 2.61, "grad_norm": 19.469426191949704, "learning_rate": 8.625801864271411e-07, "loss": 0.4288, "step": 16725 }, { "epoch": 2.61, "grad_norm": 26.974057184806597, "learning_rate": 8.618951409633258e-07, "loss": 0.4268, "step": 16726 }, { "epoch": 2.61, "grad_norm": 22.822644507968878, "learning_rate": 8.612103553818663e-07, "loss": 0.5234, "step": 16727 }, { "epoch": 2.61, "grad_norm": 17.76676574909621, "learning_rate": 8.605258297022401e-07, "loss": 0.3717, "step": 16728 }, { "epoch": 2.61, "grad_norm": 15.899604249361133, "learning_rate": 8.598415639439107e-07, "loss": 0.4623, "step": 16729 }, { "epoch": 2.61, "grad_norm": 11.332693949128902, "learning_rate": 8.591575581263422e-07, "loss": 0.4472, "step": 16730 }, { "epoch": 2.61, "grad_norm": 24.60243323073229, "learning_rate": 8.58473812268984e-07, "loss": 0.4391, "step": 16731 }, { "epoch": 2.61, "grad_norm": 17.63328755101761, "learning_rate": 8.577903263912846e-07, "loss": 0.3988, "step": 16732 }, { "epoch": 2.61, "grad_norm": 32.29013818102348, "learning_rate": 8.571071005126785e-07, "loss": 0.4678, "step": 16733 }, { "epoch": 2.61, "grad_norm": 20.242449249701234, "learning_rate": 8.564241346525992e-07, "loss": 0.4137, "step": 16734 }, { "epoch": 2.61, "grad_norm": 20.87398343052293, "learning_rate": 8.557414288304689e-07, "loss": 0.4265, "step": 16735 }, { "epoch": 2.61, "grad_norm": 25.34155911196266, "learning_rate": 8.550589830657019e-07, "loss": 0.3786, "step": 16736 }, { "epoch": 2.61, "grad_norm": 18.371625624494065, "learning_rate": 8.543767973777095e-07, "loss": 0.5357, "step": 16737 }, { "epoch": 2.61, "grad_norm": 22.74892713301613, "learning_rate": 8.536948717858895e-07, "loss": 0.4312, "step": 16738 }, { "epoch": 2.61, "grad_norm": 18.31856884009419, "learning_rate": 8.530132063096364e-07, "loss": 0.4346, "step": 16739 }, { "epoch": 2.61, "grad_norm": 25.60380375364114, "learning_rate": 8.52331800968339e-07, "loss": 0.426, "step": 16740 }, { "epoch": 2.61, "grad_norm": 24.597455123096058, "learning_rate": 8.516506557813742e-07, "loss": 0.4364, "step": 16741 }, { "epoch": 2.62, "grad_norm": 28.896660653956385, "learning_rate": 8.509697707681108e-07, "loss": 0.4143, "step": 16742 }, { "epoch": 2.62, "grad_norm": 19.727868880270037, "learning_rate": 8.502891459479145e-07, "loss": 0.4241, "step": 16743 }, { "epoch": 2.62, "grad_norm": 16.96661380279549, "learning_rate": 8.496087813401454e-07, "loss": 0.433, "step": 16744 }, { "epoch": 2.62, "grad_norm": 18.645807605775385, "learning_rate": 8.489286769641492e-07, "loss": 0.417, "step": 16745 }, { "epoch": 2.62, "grad_norm": 37.30492073505449, "learning_rate": 8.482488328392668e-07, "loss": 0.4635, "step": 16746 }, { "epoch": 2.62, "grad_norm": 24.241161234969553, "learning_rate": 8.475692489848342e-07, "loss": 0.4698, "step": 16747 }, { "epoch": 2.62, "grad_norm": 26.97141750841817, "learning_rate": 8.468899254201768e-07, "loss": 0.4484, "step": 16748 }, { "epoch": 2.62, "grad_norm": 19.310213648074953, "learning_rate": 8.462108621646182e-07, "loss": 0.3882, "step": 16749 }, { "epoch": 2.62, "grad_norm": 17.321206551973784, "learning_rate": 8.455320592374683e-07, "loss": 0.5108, "step": 16750 }, { "epoch": 2.62, "grad_norm": 23.945814612197484, "learning_rate": 8.448535166580286e-07, "loss": 0.4398, "step": 16751 }, { "epoch": 2.62, "grad_norm": 16.481744167599437, "learning_rate": 8.441752344456001e-07, "loss": 0.4991, "step": 16752 }, { "epoch": 2.62, "grad_norm": 21.533260033633923, "learning_rate": 8.43497212619474e-07, "loss": 0.4354, "step": 16753 }, { "epoch": 2.62, "grad_norm": 24.105916417968604, "learning_rate": 8.428194511989285e-07, "loss": 0.4483, "step": 16754 }, { "epoch": 2.62, "grad_norm": 23.228632958799906, "learning_rate": 8.421419502032425e-07, "loss": 0.406, "step": 16755 }, { "epoch": 2.62, "grad_norm": 21.048982944080215, "learning_rate": 8.414647096516804e-07, "loss": 0.4343, "step": 16756 }, { "epoch": 2.62, "grad_norm": 25.892150579569723, "learning_rate": 8.40787729563507e-07, "loss": 0.3986, "step": 16757 }, { "epoch": 2.62, "grad_norm": 15.535536964975222, "learning_rate": 8.401110099579702e-07, "loss": 0.4378, "step": 16758 }, { "epoch": 2.62, "grad_norm": 20.418707928177778, "learning_rate": 8.394345508543178e-07, "loss": 0.3899, "step": 16759 }, { "epoch": 2.62, "grad_norm": 24.74179145609242, "learning_rate": 8.387583522717901e-07, "loss": 0.4278, "step": 16760 }, { "epoch": 2.62, "grad_norm": 36.1969580306127, "learning_rate": 8.38082414229614e-07, "loss": 0.4397, "step": 16761 }, { "epoch": 2.62, "grad_norm": 20.140718870484307, "learning_rate": 8.374067367470151e-07, "loss": 0.432, "step": 16762 }, { "epoch": 2.62, "grad_norm": 19.99467742622639, "learning_rate": 8.367313198432081e-07, "loss": 0.4388, "step": 16763 }, { "epoch": 2.62, "grad_norm": 18.80601838524848, "learning_rate": 8.36056163537402e-07, "loss": 0.4025, "step": 16764 }, { "epoch": 2.62, "grad_norm": 23.8081271366024, "learning_rate": 8.353812678487993e-07, "loss": 0.4205, "step": 16765 }, { "epoch": 2.62, "grad_norm": 20.926565856637104, "learning_rate": 8.347066327965925e-07, "loss": 0.4961, "step": 16766 }, { "epoch": 2.62, "grad_norm": 35.660807793839375, "learning_rate": 8.340322583999649e-07, "loss": 0.3895, "step": 16767 }, { "epoch": 2.62, "grad_norm": 23.030089446166176, "learning_rate": 8.333581446780981e-07, "loss": 0.4632, "step": 16768 }, { "epoch": 2.62, "grad_norm": 27.060058253986636, "learning_rate": 8.326842916501654e-07, "loss": 0.5217, "step": 16769 }, { "epoch": 2.62, "grad_norm": 14.039772982655416, "learning_rate": 8.320106993353294e-07, "loss": 0.3913, "step": 16770 }, { "epoch": 2.62, "grad_norm": 25.421402582750858, "learning_rate": 8.313373677527437e-07, "loss": 0.4352, "step": 16771 }, { "epoch": 2.62, "grad_norm": 30.13481000243911, "learning_rate": 8.306642969215595e-07, "loss": 0.4793, "step": 16772 }, { "epoch": 2.62, "grad_norm": 23.837973011066698, "learning_rate": 8.299914868609193e-07, "loss": 0.423, "step": 16773 }, { "epoch": 2.62, "grad_norm": 31.082366852376566, "learning_rate": 8.293189375899579e-07, "loss": 0.4193, "step": 16774 }, { "epoch": 2.62, "grad_norm": 24.50445860318181, "learning_rate": 8.286466491278023e-07, "loss": 0.4479, "step": 16775 }, { "epoch": 2.62, "grad_norm": 27.43710492182028, "learning_rate": 8.27974621493568e-07, "loss": 0.4794, "step": 16776 }, { "epoch": 2.62, "grad_norm": 21.1983498746172, "learning_rate": 8.27302854706371e-07, "loss": 0.3988, "step": 16777 }, { "epoch": 2.62, "grad_norm": 19.101546037506935, "learning_rate": 8.266313487853162e-07, "loss": 0.4968, "step": 16778 }, { "epoch": 2.62, "grad_norm": 16.62009349238312, "learning_rate": 8.259601037494991e-07, "loss": 0.4279, "step": 16779 }, { "epoch": 2.62, "grad_norm": 16.489964474939928, "learning_rate": 8.252891196180091e-07, "loss": 0.3905, "step": 16780 }, { "epoch": 2.62, "grad_norm": 18.66843271886234, "learning_rate": 8.246183964099286e-07, "loss": 0.4721, "step": 16781 }, { "epoch": 2.62, "grad_norm": 27.24275700854788, "learning_rate": 8.239479341443357e-07, "loss": 0.4734, "step": 16782 }, { "epoch": 2.62, "grad_norm": 25.523912614546024, "learning_rate": 8.232777328402941e-07, "loss": 0.4371, "step": 16783 }, { "epoch": 2.62, "grad_norm": 18.74135799249883, "learning_rate": 8.226077925168674e-07, "loss": 0.4146, "step": 16784 }, { "epoch": 2.62, "grad_norm": 37.497727513598754, "learning_rate": 8.219381131931048e-07, "loss": 0.3987, "step": 16785 }, { "epoch": 2.62, "grad_norm": 26.18147060306356, "learning_rate": 8.212686948880521e-07, "loss": 0.4137, "step": 16786 }, { "epoch": 2.62, "grad_norm": 20.95040533477703, "learning_rate": 8.20599537620751e-07, "loss": 0.4516, "step": 16787 }, { "epoch": 2.62, "grad_norm": 33.21193701012433, "learning_rate": 8.199306414102282e-07, "loss": 0.5055, "step": 16788 }, { "epoch": 2.62, "grad_norm": 25.343243771446673, "learning_rate": 8.192620062755085e-07, "loss": 0.4725, "step": 16789 }, { "epoch": 2.62, "grad_norm": 20.176099939035836, "learning_rate": 8.185936322356047e-07, "loss": 0.4315, "step": 16790 }, { "epoch": 2.62, "grad_norm": 33.648030290338376, "learning_rate": 8.179255193095292e-07, "loss": 0.4847, "step": 16791 }, { "epoch": 2.62, "grad_norm": 26.493735355662018, "learning_rate": 8.172576675162791e-07, "loss": 0.4058, "step": 16792 }, { "epoch": 2.62, "grad_norm": 19.61509767974897, "learning_rate": 8.165900768748491e-07, "loss": 0.422, "step": 16793 }, { "epoch": 2.62, "grad_norm": 21.645227434823752, "learning_rate": 8.159227474042276e-07, "loss": 0.4167, "step": 16794 }, { "epoch": 2.62, "grad_norm": 21.955256101404252, "learning_rate": 8.15255679123389e-07, "loss": 0.4241, "step": 16795 }, { "epoch": 2.62, "grad_norm": 18.46999967207054, "learning_rate": 8.145888720513051e-07, "loss": 0.4071, "step": 16796 }, { "epoch": 2.62, "grad_norm": 22.830333613319574, "learning_rate": 8.139223262069407e-07, "loss": 0.4656, "step": 16797 }, { "epoch": 2.62, "grad_norm": 17.8787181911867, "learning_rate": 8.132560416092516e-07, "loss": 0.4108, "step": 16798 }, { "epoch": 2.62, "grad_norm": 19.966142990180433, "learning_rate": 8.125900182771874e-07, "loss": 0.514, "step": 16799 }, { "epoch": 2.62, "grad_norm": 18.276508767152627, "learning_rate": 8.119242562296892e-07, "loss": 0.3818, "step": 16800 }, { "epoch": 2.62, "grad_norm": 18.75664585271627, "learning_rate": 8.112587554856888e-07, "loss": 0.3909, "step": 16801 }, { "epoch": 2.62, "grad_norm": 16.617668005466477, "learning_rate": 8.105935160641143e-07, "loss": 0.4209, "step": 16802 }, { "epoch": 2.62, "grad_norm": 23.143456110065678, "learning_rate": 8.099285379838862e-07, "loss": 0.4811, "step": 16803 }, { "epoch": 2.62, "grad_norm": 17.81945785979983, "learning_rate": 8.092638212639136e-07, "loss": 0.4048, "step": 16804 }, { "epoch": 2.62, "grad_norm": 29.83116819485186, "learning_rate": 8.085993659231006e-07, "loss": 0.4267, "step": 16805 }, { "epoch": 2.63, "grad_norm": 19.959123716419516, "learning_rate": 8.079351719803441e-07, "loss": 0.4186, "step": 16806 }, { "epoch": 2.63, "grad_norm": 22.705781315240756, "learning_rate": 8.072712394545368e-07, "loss": 0.4208, "step": 16807 }, { "epoch": 2.63, "grad_norm": 30.147405975955014, "learning_rate": 8.066075683645557e-07, "loss": 0.5079, "step": 16808 }, { "epoch": 2.63, "grad_norm": 21.853072253400462, "learning_rate": 8.059441587292782e-07, "loss": 0.3982, "step": 16809 }, { "epoch": 2.63, "grad_norm": 20.264731188925893, "learning_rate": 8.052810105675702e-07, "loss": 0.4114, "step": 16810 }, { "epoch": 2.63, "grad_norm": 20.17411913086202, "learning_rate": 8.046181238982908e-07, "loss": 0.4288, "step": 16811 }, { "epoch": 2.63, "grad_norm": 63.012065382080834, "learning_rate": 8.039554987402942e-07, "loss": 0.45, "step": 16812 }, { "epoch": 2.63, "grad_norm": 15.451686381382414, "learning_rate": 8.03293135112423e-07, "loss": 0.4377, "step": 16813 }, { "epoch": 2.63, "grad_norm": 25.465290990800092, "learning_rate": 8.026310330335163e-07, "loss": 0.4364, "step": 16814 }, { "epoch": 2.63, "grad_norm": 27.945110257392457, "learning_rate": 8.019691925224004e-07, "loss": 0.4653, "step": 16815 }, { "epoch": 2.63, "grad_norm": 36.84208957556049, "learning_rate": 8.013076135979025e-07, "loss": 0.453, "step": 16816 }, { "epoch": 2.63, "grad_norm": 17.14770967550887, "learning_rate": 8.006462962788331e-07, "loss": 0.3888, "step": 16817 }, { "epoch": 2.63, "grad_norm": 23.031860844561525, "learning_rate": 7.999852405840025e-07, "loss": 0.4422, "step": 16818 }, { "epoch": 2.63, "grad_norm": 32.92649912212816, "learning_rate": 7.993244465322092e-07, "loss": 0.5207, "step": 16819 }, { "epoch": 2.63, "grad_norm": 14.903131962600781, "learning_rate": 7.986639141422469e-07, "loss": 0.3809, "step": 16820 }, { "epoch": 2.63, "grad_norm": 27.86737348966611, "learning_rate": 7.980036434328997e-07, "loss": 0.4695, "step": 16821 }, { "epoch": 2.63, "grad_norm": 30.168104516958334, "learning_rate": 7.973436344229458e-07, "loss": 0.4653, "step": 16822 }, { "epoch": 2.63, "grad_norm": 18.473699372465013, "learning_rate": 7.966838871311566e-07, "loss": 0.3886, "step": 16823 }, { "epoch": 2.63, "grad_norm": 17.798365960033053, "learning_rate": 7.960244015762919e-07, "loss": 0.4322, "step": 16824 }, { "epoch": 2.63, "grad_norm": 38.22287298032886, "learning_rate": 7.953651777771121e-07, "loss": 0.4198, "step": 16825 }, { "epoch": 2.63, "grad_norm": 29.210312917499845, "learning_rate": 7.947062157523589e-07, "loss": 0.4228, "step": 16826 }, { "epoch": 2.63, "grad_norm": 20.062279305625704, "learning_rate": 7.940475155207772e-07, "loss": 0.4341, "step": 16827 }, { "epoch": 2.63, "grad_norm": 15.993112556899609, "learning_rate": 7.933890771010999e-07, "loss": 0.4629, "step": 16828 }, { "epoch": 2.63, "grad_norm": 23.010568034084475, "learning_rate": 7.927309005120521e-07, "loss": 0.4348, "step": 16829 }, { "epoch": 2.63, "grad_norm": 18.577993284998563, "learning_rate": 7.920729857723486e-07, "loss": 0.4326, "step": 16830 }, { "epoch": 2.63, "grad_norm": 17.987880536647573, "learning_rate": 7.914153329007035e-07, "loss": 0.4189, "step": 16831 }, { "epoch": 2.63, "grad_norm": 19.49306578711967, "learning_rate": 7.907579419158196e-07, "loss": 0.411, "step": 16832 }, { "epoch": 2.63, "grad_norm": 17.166915742350334, "learning_rate": 7.901008128363963e-07, "loss": 0.4271, "step": 16833 }, { "epoch": 2.63, "grad_norm": 16.84249898585503, "learning_rate": 7.894439456811143e-07, "loss": 0.4914, "step": 16834 }, { "epoch": 2.63, "grad_norm": 17.18164285128788, "learning_rate": 7.887873404686586e-07, "loss": 0.4232, "step": 16835 }, { "epoch": 2.63, "grad_norm": 30.780090636964363, "learning_rate": 7.88130997217702e-07, "loss": 0.4636, "step": 16836 }, { "epoch": 2.63, "grad_norm": 23.54647295481414, "learning_rate": 7.874749159469131e-07, "loss": 0.4376, "step": 16837 }, { "epoch": 2.63, "grad_norm": 19.70252201298119, "learning_rate": 7.868190966749489e-07, "loss": 0.4249, "step": 16838 }, { "epoch": 2.63, "grad_norm": 24.30281482411287, "learning_rate": 7.861635394204581e-07, "loss": 0.4623, "step": 16839 }, { "epoch": 2.63, "grad_norm": 21.627583553703232, "learning_rate": 7.855082442020867e-07, "loss": 0.3949, "step": 16840 }, { "epoch": 2.63, "grad_norm": 18.371049234059328, "learning_rate": 7.84853211038471e-07, "loss": 0.4823, "step": 16841 }, { "epoch": 2.63, "grad_norm": 21.317755842555908, "learning_rate": 7.841984399482383e-07, "loss": 0.4953, "step": 16842 }, { "epoch": 2.63, "grad_norm": 19.099669392705323, "learning_rate": 7.835439309500126e-07, "loss": 0.4315, "step": 16843 }, { "epoch": 2.63, "grad_norm": 17.893913050184157, "learning_rate": 7.828896840624045e-07, "loss": 0.427, "step": 16844 }, { "epoch": 2.63, "grad_norm": 21.702461904718188, "learning_rate": 7.822356993040236e-07, "loss": 0.4448, "step": 16845 }, { "epoch": 2.63, "grad_norm": 21.718610975794803, "learning_rate": 7.815819766934651e-07, "loss": 0.5642, "step": 16846 }, { "epoch": 2.63, "grad_norm": 17.398796786558435, "learning_rate": 7.80928516249323e-07, "loss": 0.3895, "step": 16847 }, { "epoch": 2.63, "grad_norm": 21.030846737854098, "learning_rate": 7.802753179901823e-07, "loss": 0.426, "step": 16848 }, { "epoch": 2.63, "grad_norm": 19.90180827535106, "learning_rate": 7.796223819346171e-07, "loss": 0.3751, "step": 16849 }, { "epoch": 2.63, "grad_norm": 34.96958743022472, "learning_rate": 7.789697081011982e-07, "loss": 0.3893, "step": 16850 }, { "epoch": 2.63, "grad_norm": 16.01562875515033, "learning_rate": 7.783172965084851e-07, "loss": 0.4309, "step": 16851 }, { "epoch": 2.63, "grad_norm": 14.948257573877996, "learning_rate": 7.77665147175034e-07, "loss": 0.4026, "step": 16852 }, { "epoch": 2.63, "grad_norm": 22.418381484613125, "learning_rate": 7.770132601193936e-07, "loss": 0.46, "step": 16853 }, { "epoch": 2.63, "grad_norm": 22.844320560823682, "learning_rate": 7.763616353601e-07, "loss": 0.4746, "step": 16854 }, { "epoch": 2.63, "grad_norm": 22.733132272031753, "learning_rate": 7.75710272915684e-07, "loss": 0.4695, "step": 16855 }, { "epoch": 2.63, "grad_norm": 28.000478809596892, "learning_rate": 7.750591728046719e-07, "loss": 0.4036, "step": 16856 }, { "epoch": 2.63, "grad_norm": 24.820126679601113, "learning_rate": 7.744083350455811e-07, "loss": 0.4649, "step": 16857 }, { "epoch": 2.63, "grad_norm": 20.32344043127804, "learning_rate": 7.737577596569223e-07, "loss": 0.4875, "step": 16858 }, { "epoch": 2.63, "grad_norm": 22.400681451454044, "learning_rate": 7.731074466571942e-07, "loss": 0.398, "step": 16859 }, { "epoch": 2.63, "grad_norm": 17.71669893634057, "learning_rate": 7.724573960648907e-07, "loss": 0.3693, "step": 16860 }, { "epoch": 2.63, "grad_norm": 18.352677457157014, "learning_rate": 7.718076078985026e-07, "loss": 0.3971, "step": 16861 }, { "epoch": 2.63, "grad_norm": 19.02533477471114, "learning_rate": 7.711580821765085e-07, "loss": 0.432, "step": 16862 }, { "epoch": 2.63, "grad_norm": 17.217546166501688, "learning_rate": 7.705088189173804e-07, "loss": 0.4335, "step": 16863 }, { "epoch": 2.63, "grad_norm": 23.160675082907265, "learning_rate": 7.6985981813958e-07, "loss": 0.4206, "step": 16864 }, { "epoch": 2.63, "grad_norm": 23.202533562825728, "learning_rate": 7.692110798615682e-07, "loss": 0.4083, "step": 16865 }, { "epoch": 2.63, "grad_norm": 19.503359812365275, "learning_rate": 7.685626041017935e-07, "loss": 0.4502, "step": 16866 }, { "epoch": 2.63, "grad_norm": 17.27997414275035, "learning_rate": 7.67914390878699e-07, "loss": 0.4318, "step": 16867 }, { "epoch": 2.63, "grad_norm": 25.835566482267733, "learning_rate": 7.672664402107166e-07, "loss": 0.4624, "step": 16868 }, { "epoch": 2.63, "grad_norm": 17.61740463190695, "learning_rate": 7.66618752116276e-07, "loss": 0.3958, "step": 16869 }, { "epoch": 2.64, "grad_norm": 22.85098183309853, "learning_rate": 7.659713266137992e-07, "loss": 0.4409, "step": 16870 }, { "epoch": 2.64, "grad_norm": 18.75568043175452, "learning_rate": 7.653241637216924e-07, "loss": 0.3987, "step": 16871 }, { "epoch": 2.64, "grad_norm": 38.46801953044423, "learning_rate": 7.646772634583677e-07, "loss": 0.4885, "step": 16872 }, { "epoch": 2.64, "grad_norm": 25.697771136275506, "learning_rate": 7.64030625842217e-07, "loss": 0.4978, "step": 16873 }, { "epoch": 2.64, "grad_norm": 17.03607930650342, "learning_rate": 7.633842508916323e-07, "loss": 0.4302, "step": 16874 }, { "epoch": 2.64, "grad_norm": 29.131389335085494, "learning_rate": 7.627381386249976e-07, "loss": 0.4964, "step": 16875 }, { "epoch": 2.64, "grad_norm": 24.221972021076663, "learning_rate": 7.620922890606852e-07, "loss": 0.4262, "step": 16876 }, { "epoch": 2.64, "grad_norm": 19.861658705586958, "learning_rate": 7.614467022170658e-07, "loss": 0.4186, "step": 16877 }, { "epoch": 2.64, "grad_norm": 26.86455808652885, "learning_rate": 7.608013781124956e-07, "loss": 0.5342, "step": 16878 }, { "epoch": 2.64, "grad_norm": 23.439050092300334, "learning_rate": 7.601563167653314e-07, "loss": 0.4012, "step": 16879 }, { "epoch": 2.64, "grad_norm": 22.456579716420062, "learning_rate": 7.59511518193915e-07, "loss": 0.4512, "step": 16880 }, { "epoch": 2.64, "grad_norm": 29.554704605630015, "learning_rate": 7.588669824165851e-07, "loss": 0.4681, "step": 16881 }, { "epoch": 2.64, "grad_norm": 34.10344989705391, "learning_rate": 7.582227094516747e-07, "loss": 0.493, "step": 16882 }, { "epoch": 2.64, "grad_norm": 19.93347241115998, "learning_rate": 7.575786993175028e-07, "loss": 0.4627, "step": 16883 }, { "epoch": 2.64, "grad_norm": 29.393700149294325, "learning_rate": 7.569349520323854e-07, "loss": 0.4639, "step": 16884 }, { "epoch": 2.64, "grad_norm": 18.614876370363987, "learning_rate": 7.562914676146304e-07, "loss": 0.4296, "step": 16885 }, { "epoch": 2.64, "grad_norm": 16.77048501804122, "learning_rate": 7.556482460825399e-07, "loss": 0.3855, "step": 16886 }, { "epoch": 2.64, "grad_norm": 25.079864852779185, "learning_rate": 7.550052874544056e-07, "loss": 0.4153, "step": 16887 }, { "epoch": 2.64, "grad_norm": 27.814226345533076, "learning_rate": 7.543625917485142e-07, "loss": 0.3975, "step": 16888 }, { "epoch": 2.64, "grad_norm": 29.274484261906014, "learning_rate": 7.537201589831389e-07, "loss": 0.4465, "step": 16889 }, { "epoch": 2.64, "grad_norm": 18.42591955955083, "learning_rate": 7.530779891765549e-07, "loss": 0.3628, "step": 16890 }, { "epoch": 2.64, "grad_norm": 20.478408078269176, "learning_rate": 7.524360823470244e-07, "loss": 0.4413, "step": 16891 }, { "epoch": 2.64, "grad_norm": 23.506952992044383, "learning_rate": 7.517944385128018e-07, "loss": 0.4642, "step": 16892 }, { "epoch": 2.64, "grad_norm": 18.75456311728819, "learning_rate": 7.511530576921344e-07, "loss": 0.4452, "step": 16893 }, { "epoch": 2.64, "grad_norm": 22.437404387968996, "learning_rate": 7.505119399032623e-07, "loss": 0.4101, "step": 16894 }, { "epoch": 2.64, "grad_norm": 24.54236982975728, "learning_rate": 7.498710851644231e-07, "loss": 0.3821, "step": 16895 }, { "epoch": 2.64, "grad_norm": 29.62335673484736, "learning_rate": 7.492304934938365e-07, "loss": 0.4359, "step": 16896 }, { "epoch": 2.64, "grad_norm": 28.060790699130283, "learning_rate": 7.48590164909725e-07, "loss": 0.4738, "step": 16897 }, { "epoch": 2.64, "grad_norm": 19.914088653631087, "learning_rate": 7.479500994302957e-07, "loss": 0.3831, "step": 16898 }, { "epoch": 2.64, "grad_norm": 16.568575890819545, "learning_rate": 7.473102970737534e-07, "loss": 0.4251, "step": 16899 }, { "epoch": 2.64, "grad_norm": 25.149919860230582, "learning_rate": 7.466707578582954e-07, "loss": 0.5003, "step": 16900 }, { "epoch": 2.64, "grad_norm": 25.38762558108232, "learning_rate": 7.460314818021053e-07, "loss": 0.3956, "step": 16901 }, { "epoch": 2.64, "grad_norm": 21.40323957311481, "learning_rate": 7.453924689233693e-07, "loss": 0.4099, "step": 16902 }, { "epoch": 2.64, "grad_norm": 17.58806714822402, "learning_rate": 7.447537192402554e-07, "loss": 0.4006, "step": 16903 }, { "epoch": 2.64, "grad_norm": 33.796221153153, "learning_rate": 7.441152327709334e-07, "loss": 0.407, "step": 16904 }, { "epoch": 2.64, "grad_norm": 23.945559204473557, "learning_rate": 7.434770095335575e-07, "loss": 0.441, "step": 16905 }, { "epoch": 2.64, "grad_norm": 28.343743056156608, "learning_rate": 7.428390495462812e-07, "loss": 0.4714, "step": 16906 }, { "epoch": 2.64, "grad_norm": 33.61964477292253, "learning_rate": 7.422013528272487e-07, "loss": 0.4491, "step": 16907 }, { "epoch": 2.64, "grad_norm": 25.092504680060543, "learning_rate": 7.415639193945945e-07, "loss": 0.4407, "step": 16908 }, { "epoch": 2.64, "grad_norm": 19.522835893585672, "learning_rate": 7.409267492664441e-07, "loss": 0.5003, "step": 16909 }, { "epoch": 2.64, "grad_norm": 26.416019802692528, "learning_rate": 7.402898424609206e-07, "loss": 0.4278, "step": 16910 }, { "epoch": 2.64, "grad_norm": 21.506691327371218, "learning_rate": 7.396531989961364e-07, "loss": 0.4247, "step": 16911 }, { "epoch": 2.64, "grad_norm": 34.72702402638485, "learning_rate": 7.390168188902014e-07, "loss": 0.4221, "step": 16912 }, { "epoch": 2.64, "grad_norm": 27.332931770317877, "learning_rate": 7.383807021612089e-07, "loss": 0.4089, "step": 16913 }, { "epoch": 2.64, "grad_norm": 20.784822813677792, "learning_rate": 7.377448488272509e-07, "loss": 0.4775, "step": 16914 }, { "epoch": 2.64, "grad_norm": 23.164310796863965, "learning_rate": 7.3710925890641e-07, "loss": 0.4712, "step": 16915 }, { "epoch": 2.64, "grad_norm": 19.609184651001883, "learning_rate": 7.364739324167658e-07, "loss": 0.3752, "step": 16916 }, { "epoch": 2.64, "grad_norm": 24.25547887539159, "learning_rate": 7.35838869376384e-07, "loss": 0.4694, "step": 16917 }, { "epoch": 2.64, "grad_norm": 46.19113616541803, "learning_rate": 7.352040698033236e-07, "loss": 0.4421, "step": 16918 }, { "epoch": 2.64, "grad_norm": 37.32707569440677, "learning_rate": 7.345695337156389e-07, "loss": 0.5207, "step": 16919 }, { "epoch": 2.64, "grad_norm": 27.358150016859156, "learning_rate": 7.339352611313777e-07, "loss": 0.4851, "step": 16920 }, { "epoch": 2.64, "grad_norm": 21.542554103288225, "learning_rate": 7.33301252068579e-07, "loss": 0.5202, "step": 16921 }, { "epoch": 2.64, "grad_norm": 15.700291455294751, "learning_rate": 7.326675065452693e-07, "loss": 0.3829, "step": 16922 }, { "epoch": 2.64, "grad_norm": 39.02120483625192, "learning_rate": 7.320340245794755e-07, "loss": 0.4012, "step": 16923 }, { "epoch": 2.64, "grad_norm": 24.32930446624821, "learning_rate": 7.314008061892108e-07, "loss": 0.4506, "step": 16924 }, { "epoch": 2.64, "grad_norm": 17.33022061773795, "learning_rate": 7.307678513924877e-07, "loss": 0.3949, "step": 16925 }, { "epoch": 2.64, "grad_norm": 20.4365060809573, "learning_rate": 7.301351602073048e-07, "loss": 0.4611, "step": 16926 }, { "epoch": 2.64, "grad_norm": 17.600283221111766, "learning_rate": 7.295027326516535e-07, "loss": 0.4199, "step": 16927 }, { "epoch": 2.64, "grad_norm": 20.78572774344092, "learning_rate": 7.288705687435204e-07, "loss": 0.435, "step": 16928 }, { "epoch": 2.64, "grad_norm": 19.812328181439334, "learning_rate": 7.28238668500888e-07, "loss": 0.5015, "step": 16929 }, { "epoch": 2.64, "grad_norm": 22.96636934009789, "learning_rate": 7.276070319417216e-07, "loss": 0.4346, "step": 16930 }, { "epoch": 2.64, "grad_norm": 21.218819628644813, "learning_rate": 7.269756590839883e-07, "loss": 0.4315, "step": 16931 }, { "epoch": 2.64, "grad_norm": 23.283954821665816, "learning_rate": 7.263445499456412e-07, "loss": 0.4576, "step": 16932 }, { "epoch": 2.64, "grad_norm": 24.04858804417328, "learning_rate": 7.257137045446327e-07, "loss": 0.4329, "step": 16933 }, { "epoch": 2.65, "grad_norm": 15.926081830243813, "learning_rate": 7.250831228988986e-07, "loss": 0.4557, "step": 16934 }, { "epoch": 2.65, "grad_norm": 17.174457614433024, "learning_rate": 7.244528050263744e-07, "loss": 0.454, "step": 16935 }, { "epoch": 2.65, "grad_norm": 20.867504037493568, "learning_rate": 7.238227509449891e-07, "loss": 0.3953, "step": 16936 }, { "epoch": 2.65, "grad_norm": 34.6054061172561, "learning_rate": 7.231929606726562e-07, "loss": 0.4575, "step": 16937 }, { "epoch": 2.65, "grad_norm": 13.417321541570763, "learning_rate": 7.225634342272903e-07, "loss": 0.3648, "step": 16938 }, { "epoch": 2.65, "grad_norm": 19.1481978231408, "learning_rate": 7.219341716267925e-07, "loss": 0.4202, "step": 16939 }, { "epoch": 2.65, "grad_norm": 24.39213821181831, "learning_rate": 7.213051728890586e-07, "loss": 0.4107, "step": 16940 }, { "epoch": 2.65, "grad_norm": 29.51286279517155, "learning_rate": 7.206764380319786e-07, "loss": 0.4013, "step": 16941 }, { "epoch": 2.65, "grad_norm": 16.480239657708285, "learning_rate": 7.20047967073434e-07, "loss": 0.4018, "step": 16942 }, { "epoch": 2.65, "grad_norm": 17.85465106748411, "learning_rate": 7.194197600312936e-07, "loss": 0.4261, "step": 16943 }, { "epoch": 2.65, "grad_norm": 21.240310128614013, "learning_rate": 7.187918169234265e-07, "loss": 0.4741, "step": 16944 }, { "epoch": 2.65, "grad_norm": 18.877992688804472, "learning_rate": 7.181641377676918e-07, "loss": 0.4236, "step": 16945 }, { "epoch": 2.65, "grad_norm": 19.54637360025032, "learning_rate": 7.175367225819418e-07, "loss": 0.4088, "step": 16946 }, { "epoch": 2.65, "grad_norm": 28.202189692577342, "learning_rate": 7.169095713840135e-07, "loss": 0.4307, "step": 16947 }, { "epoch": 2.65, "grad_norm": 20.242653932280582, "learning_rate": 7.16282684191747e-07, "loss": 0.4118, "step": 16948 }, { "epoch": 2.65, "grad_norm": 25.825725198157475, "learning_rate": 7.156560610229701e-07, "loss": 0.4823, "step": 16949 }, { "epoch": 2.65, "grad_norm": 30.023298817616297, "learning_rate": 7.150297018955055e-07, "loss": 0.4348, "step": 16950 }, { "epoch": 2.65, "grad_norm": 24.26751415712457, "learning_rate": 7.144036068271654e-07, "loss": 0.4398, "step": 16951 }, { "epoch": 2.65, "grad_norm": 23.18227637387653, "learning_rate": 7.137777758357523e-07, "loss": 0.4926, "step": 16952 }, { "epoch": 2.65, "grad_norm": 17.60419619903871, "learning_rate": 7.131522089390663e-07, "loss": 0.4225, "step": 16953 }, { "epoch": 2.65, "grad_norm": 18.218780213293343, "learning_rate": 7.125269061549012e-07, "loss": 0.4487, "step": 16954 }, { "epoch": 2.65, "grad_norm": 24.474617092675352, "learning_rate": 7.119018675010369e-07, "loss": 0.4443, "step": 16955 }, { "epoch": 2.65, "grad_norm": 15.064154574624544, "learning_rate": 7.112770929952506e-07, "loss": 0.4581, "step": 16956 }, { "epoch": 2.65, "grad_norm": 24.531868571178812, "learning_rate": 7.10652582655309e-07, "loss": 0.4319, "step": 16957 }, { "epoch": 2.65, "grad_norm": 24.806323178645442, "learning_rate": 7.100283364989757e-07, "loss": 0.4148, "step": 16958 }, { "epoch": 2.65, "grad_norm": 28.25120667395479, "learning_rate": 7.094043545439999e-07, "loss": 0.4339, "step": 16959 }, { "epoch": 2.65, "grad_norm": 25.00788101906701, "learning_rate": 7.087806368081296e-07, "loss": 0.481, "step": 16960 }, { "epoch": 2.65, "grad_norm": 17.9109394047861, "learning_rate": 7.081571833091039e-07, "loss": 0.4802, "step": 16961 }, { "epoch": 2.65, "grad_norm": 18.740040837849502, "learning_rate": 7.075339940646508e-07, "loss": 0.4054, "step": 16962 }, { "epoch": 2.65, "grad_norm": 20.74466591663702, "learning_rate": 7.069110690924974e-07, "loss": 0.3905, "step": 16963 }, { "epoch": 2.65, "grad_norm": 22.29085204531449, "learning_rate": 7.06288408410355e-07, "loss": 0.385, "step": 16964 }, { "epoch": 2.65, "grad_norm": 18.844893795619, "learning_rate": 7.056660120359349e-07, "loss": 0.4143, "step": 16965 }, { "epoch": 2.65, "grad_norm": 25.0207018040152, "learning_rate": 7.050438799869352e-07, "loss": 0.4699, "step": 16966 }, { "epoch": 2.65, "grad_norm": 15.597550155580713, "learning_rate": 7.044220122810508e-07, "loss": 0.3672, "step": 16967 }, { "epoch": 2.65, "grad_norm": 23.35701236577994, "learning_rate": 7.038004089359663e-07, "loss": 0.4449, "step": 16968 }, { "epoch": 2.65, "grad_norm": 22.91420574904314, "learning_rate": 7.031790699693586e-07, "loss": 0.4751, "step": 16969 }, { "epoch": 2.65, "grad_norm": 14.909948800782285, "learning_rate": 7.025579953989026e-07, "loss": 0.4267, "step": 16970 }, { "epoch": 2.65, "grad_norm": 19.633419328048156, "learning_rate": 7.019371852422574e-07, "loss": 0.4522, "step": 16971 }, { "epoch": 2.65, "grad_norm": 40.40897275317225, "learning_rate": 7.01316639517079e-07, "loss": 0.4922, "step": 16972 }, { "epoch": 2.65, "grad_norm": 19.68186955093115, "learning_rate": 7.006963582410153e-07, "loss": 0.3753, "step": 16973 }, { "epoch": 2.65, "grad_norm": 32.352136896557326, "learning_rate": 7.000763414317069e-07, "loss": 0.4366, "step": 16974 }, { "epoch": 2.65, "grad_norm": 12.033830856347995, "learning_rate": 6.994565891067884e-07, "loss": 0.3926, "step": 16975 }, { "epoch": 2.65, "grad_norm": 19.2580439478215, "learning_rate": 6.988371012838846e-07, "loss": 0.4956, "step": 16976 }, { "epoch": 2.65, "grad_norm": 26.155731691673708, "learning_rate": 6.982178779806104e-07, "loss": 0.4942, "step": 16977 }, { "epoch": 2.65, "grad_norm": 17.71712777113979, "learning_rate": 6.975989192145782e-07, "loss": 0.4181, "step": 16978 }, { "epoch": 2.65, "grad_norm": 27.454886283890847, "learning_rate": 6.96980225003393e-07, "loss": 0.4593, "step": 16979 }, { "epoch": 2.65, "grad_norm": 23.569902573897206, "learning_rate": 6.963617953646484e-07, "loss": 0.4401, "step": 16980 }, { "epoch": 2.65, "grad_norm": 34.14101564546121, "learning_rate": 6.957436303159304e-07, "loss": 0.4662, "step": 16981 }, { "epoch": 2.65, "grad_norm": 23.973733968815406, "learning_rate": 6.951257298748204e-07, "loss": 0.4327, "step": 16982 }, { "epoch": 2.65, "grad_norm": 28.526577083748773, "learning_rate": 6.94508094058891e-07, "loss": 0.4512, "step": 16983 }, { "epoch": 2.65, "grad_norm": 17.720721383234114, "learning_rate": 6.938907228857095e-07, "loss": 0.4055, "step": 16984 }, { "epoch": 2.65, "grad_norm": 17.930534144588915, "learning_rate": 6.932736163728327e-07, "loss": 0.4427, "step": 16985 }, { "epoch": 2.65, "grad_norm": 15.495941110873215, "learning_rate": 6.926567745378076e-07, "loss": 0.3663, "step": 16986 }, { "epoch": 2.65, "grad_norm": 32.750840610351986, "learning_rate": 6.920401973981794e-07, "loss": 0.4481, "step": 16987 }, { "epoch": 2.65, "grad_norm": 21.216017385605664, "learning_rate": 6.914238849714849e-07, "loss": 0.4434, "step": 16988 }, { "epoch": 2.65, "grad_norm": 20.142577046394596, "learning_rate": 6.90807837275248e-07, "loss": 0.4575, "step": 16989 }, { "epoch": 2.65, "grad_norm": 21.24530196288164, "learning_rate": 6.901920543269925e-07, "loss": 0.4538, "step": 16990 }, { "epoch": 2.65, "grad_norm": 16.196424638366462, "learning_rate": 6.895765361442264e-07, "loss": 0.4212, "step": 16991 }, { "epoch": 2.65, "grad_norm": 18.46682812361794, "learning_rate": 6.889612827444592e-07, "loss": 0.4293, "step": 16992 }, { "epoch": 2.65, "grad_norm": 20.018479384910545, "learning_rate": 6.883462941451846e-07, "loss": 0.4197, "step": 16993 }, { "epoch": 2.65, "grad_norm": 18.42966825185873, "learning_rate": 6.877315703638943e-07, "loss": 0.4504, "step": 16994 }, { "epoch": 2.65, "grad_norm": 23.598967359315626, "learning_rate": 6.871171114180719e-07, "loss": 0.4458, "step": 16995 }, { "epoch": 2.65, "grad_norm": 13.857768877455179, "learning_rate": 6.865029173251914e-07, "loss": 0.4249, "step": 16996 }, { "epoch": 2.65, "grad_norm": 22.479247635223278, "learning_rate": 6.858889881027187e-07, "loss": 0.4384, "step": 16997 }, { "epoch": 2.66, "grad_norm": 28.270730075820964, "learning_rate": 6.852753237681131e-07, "loss": 0.4507, "step": 16998 }, { "epoch": 2.66, "grad_norm": 31.32222442805024, "learning_rate": 6.846619243388297e-07, "loss": 0.4584, "step": 16999 }, { "epoch": 2.66, "grad_norm": 24.638024602031646, "learning_rate": 6.840487898323122e-07, "loss": 0.4003, "step": 17000 }, { "epoch": 2.66, "grad_norm": 17.438656737815304, "learning_rate": 6.83435920265999e-07, "loss": 0.4725, "step": 17001 }, { "epoch": 2.66, "grad_norm": 15.78180766489473, "learning_rate": 6.82823315657315e-07, "loss": 0.4358, "step": 17002 }, { "epoch": 2.66, "grad_norm": 27.469788973800213, "learning_rate": 6.822109760236861e-07, "loss": 0.4364, "step": 17003 }, { "epoch": 2.66, "grad_norm": 19.724507139790685, "learning_rate": 6.815989013825285e-07, "loss": 0.3935, "step": 17004 }, { "epoch": 2.66, "grad_norm": 19.300029688664686, "learning_rate": 6.809870917512462e-07, "loss": 0.4098, "step": 17005 }, { "epoch": 2.66, "grad_norm": 21.41557284557138, "learning_rate": 6.803755471472384e-07, "loss": 0.4639, "step": 17006 }, { "epoch": 2.66, "grad_norm": 19.21578261332035, "learning_rate": 6.79764267587898e-07, "loss": 0.4053, "step": 17007 }, { "epoch": 2.66, "grad_norm": 19.20306508758734, "learning_rate": 6.791532530906098e-07, "loss": 0.4261, "step": 17008 }, { "epoch": 2.66, "grad_norm": 18.602186768497322, "learning_rate": 6.785425036727522e-07, "loss": 0.401, "step": 17009 }, { "epoch": 2.66, "grad_norm": 19.02387156686479, "learning_rate": 6.779320193516925e-07, "loss": 0.408, "step": 17010 }, { "epoch": 2.66, "grad_norm": 19.59933059258428, "learning_rate": 6.773218001447923e-07, "loss": 0.4902, "step": 17011 }, { "epoch": 2.66, "grad_norm": 21.781701508115393, "learning_rate": 6.767118460694056e-07, "loss": 0.4027, "step": 17012 }, { "epoch": 2.66, "grad_norm": 15.55440458234907, "learning_rate": 6.761021571428816e-07, "loss": 0.4106, "step": 17013 }, { "epoch": 2.66, "grad_norm": 39.882061410794, "learning_rate": 6.75492733382559e-07, "loss": 0.4822, "step": 17014 }, { "epoch": 2.66, "grad_norm": 31.0322179436863, "learning_rate": 6.748835748057658e-07, "loss": 0.4131, "step": 17015 }, { "epoch": 2.66, "grad_norm": 18.6244217059585, "learning_rate": 6.742746814298285e-07, "loss": 0.4519, "step": 17016 }, { "epoch": 2.66, "grad_norm": 33.92483251954795, "learning_rate": 6.736660532720662e-07, "loss": 0.4562, "step": 17017 }, { "epoch": 2.66, "grad_norm": 29.13638939002006, "learning_rate": 6.730576903497832e-07, "loss": 0.4267, "step": 17018 }, { "epoch": 2.66, "grad_norm": 20.769779602741252, "learning_rate": 6.724495926802854e-07, "loss": 0.4623, "step": 17019 }, { "epoch": 2.66, "grad_norm": 18.370774772676448, "learning_rate": 6.718417602808636e-07, "loss": 0.4245, "step": 17020 }, { "epoch": 2.66, "grad_norm": 18.943558489430227, "learning_rate": 6.71234193168806e-07, "loss": 0.3762, "step": 17021 }, { "epoch": 2.66, "grad_norm": 20.346078157515993, "learning_rate": 6.7062689136139e-07, "loss": 0.4377, "step": 17022 }, { "epoch": 2.66, "grad_norm": 20.105330236805308, "learning_rate": 6.700198548758874e-07, "loss": 0.3829, "step": 17023 }, { "epoch": 2.66, "grad_norm": 24.84923823693445, "learning_rate": 6.694130837295632e-07, "loss": 0.4291, "step": 17024 }, { "epoch": 2.66, "grad_norm": 27.275788335089313, "learning_rate": 6.688065779396713e-07, "loss": 0.485, "step": 17025 }, { "epoch": 2.66, "grad_norm": 21.590290079460623, "learning_rate": 6.682003375234636e-07, "loss": 0.4434, "step": 17026 }, { "epoch": 2.66, "grad_norm": 19.593048153860142, "learning_rate": 6.675943624981785e-07, "loss": 0.4093, "step": 17027 }, { "epoch": 2.66, "grad_norm": 18.285089568057483, "learning_rate": 6.669886528810498e-07, "loss": 0.3994, "step": 17028 }, { "epoch": 2.66, "grad_norm": 16.830597707345294, "learning_rate": 6.663832086893052e-07, "loss": 0.4229, "step": 17029 }, { "epoch": 2.66, "grad_norm": 32.45391485311872, "learning_rate": 6.657780299401628e-07, "loss": 0.4441, "step": 17030 }, { "epoch": 2.66, "grad_norm": 17.40082117049648, "learning_rate": 6.651731166508313e-07, "loss": 0.4093, "step": 17031 }, { "epoch": 2.66, "grad_norm": 28.322825016510862, "learning_rate": 6.645684688385156e-07, "loss": 0.3998, "step": 17032 }, { "epoch": 2.66, "grad_norm": 19.069347035224137, "learning_rate": 6.639640865204111e-07, "loss": 0.3896, "step": 17033 }, { "epoch": 2.66, "grad_norm": 25.01354521374338, "learning_rate": 6.633599697137105e-07, "loss": 0.4234, "step": 17034 }, { "epoch": 2.66, "grad_norm": 18.49225597143289, "learning_rate": 6.627561184355869e-07, "loss": 0.4791, "step": 17035 }, { "epoch": 2.66, "grad_norm": 22.835923343331707, "learning_rate": 6.621525327032174e-07, "loss": 0.3775, "step": 17036 }, { "epoch": 2.66, "grad_norm": 21.424137567780104, "learning_rate": 6.615492125337663e-07, "loss": 0.4314, "step": 17037 }, { "epoch": 2.66, "grad_norm": 17.96515394052094, "learning_rate": 6.609461579443954e-07, "loss": 0.4201, "step": 17038 }, { "epoch": 2.66, "grad_norm": 25.84355037628256, "learning_rate": 6.603433689522531e-07, "loss": 0.4409, "step": 17039 }, { "epoch": 2.66, "grad_norm": 29.039379297744702, "learning_rate": 6.597408455744792e-07, "loss": 0.4797, "step": 17040 }, { "epoch": 2.66, "grad_norm": 21.028265306948867, "learning_rate": 6.59138587828212e-07, "loss": 0.5161, "step": 17041 }, { "epoch": 2.66, "grad_norm": 29.6197833989168, "learning_rate": 6.585365957305812e-07, "loss": 0.489, "step": 17042 }, { "epoch": 2.66, "grad_norm": 21.549998363375323, "learning_rate": 6.57934869298703e-07, "loss": 0.4531, "step": 17043 }, { "epoch": 2.66, "grad_norm": 22.46886577563109, "learning_rate": 6.573334085496941e-07, "loss": 0.4155, "step": 17044 }, { "epoch": 2.66, "grad_norm": 19.211848286248678, "learning_rate": 6.567322135006559e-07, "loss": 0.4408, "step": 17045 }, { "epoch": 2.66, "grad_norm": 18.880414287276913, "learning_rate": 6.561312841686895e-07, "loss": 0.4522, "step": 17046 }, { "epoch": 2.66, "grad_norm": 13.521092950427791, "learning_rate": 6.555306205708812e-07, "loss": 0.4371, "step": 17047 }, { "epoch": 2.66, "grad_norm": 22.563371343641823, "learning_rate": 6.54930222724317e-07, "loss": 0.4079, "step": 17048 }, { "epoch": 2.66, "grad_norm": 22.989714800296422, "learning_rate": 6.543300906460714e-07, "loss": 0.4755, "step": 17049 }, { "epoch": 2.66, "grad_norm": 27.088749978304143, "learning_rate": 6.537302243532095e-07, "loss": 0.4328, "step": 17050 }, { "epoch": 2.66, "grad_norm": 23.439463036027046, "learning_rate": 6.531306238627932e-07, "loss": 0.4756, "step": 17051 }, { "epoch": 2.66, "grad_norm": 27.9192020244739, "learning_rate": 6.525312891918745e-07, "loss": 0.4623, "step": 17052 }, { "epoch": 2.66, "grad_norm": 25.48685072819762, "learning_rate": 6.519322203574962e-07, "loss": 0.4263, "step": 17053 }, { "epoch": 2.66, "grad_norm": 22.352761920106488, "learning_rate": 6.513334173766994e-07, "loss": 0.4176, "step": 17054 }, { "epoch": 2.66, "grad_norm": 15.283399535065746, "learning_rate": 6.507348802665115e-07, "loss": 0.4183, "step": 17055 }, { "epoch": 2.66, "grad_norm": 20.49803511140803, "learning_rate": 6.501366090439521e-07, "loss": 0.3975, "step": 17056 }, { "epoch": 2.66, "grad_norm": 19.426900729752138, "learning_rate": 6.495386037260387e-07, "loss": 0.3751, "step": 17057 }, { "epoch": 2.66, "grad_norm": 19.43626412452176, "learning_rate": 6.489408643297778e-07, "loss": 0.4494, "step": 17058 }, { "epoch": 2.66, "grad_norm": 22.855540566078112, "learning_rate": 6.483433908721715e-07, "loss": 0.4005, "step": 17059 }, { "epoch": 2.66, "grad_norm": 17.73181552104461, "learning_rate": 6.477461833702047e-07, "loss": 0.4296, "step": 17060 }, { "epoch": 2.66, "grad_norm": 20.51348699700499, "learning_rate": 6.471492418408665e-07, "loss": 0.3751, "step": 17061 }, { "epoch": 2.67, "grad_norm": 21.553816710355985, "learning_rate": 6.465525663011318e-07, "loss": 0.4578, "step": 17062 }, { "epoch": 2.67, "grad_norm": 21.542725772818944, "learning_rate": 6.459561567679728e-07, "loss": 0.4782, "step": 17063 }, { "epoch": 2.67, "grad_norm": 29.15203652187422, "learning_rate": 6.453600132583482e-07, "loss": 0.5064, "step": 17064 }, { "epoch": 2.67, "grad_norm": 23.260137969017435, "learning_rate": 6.44764135789211e-07, "loss": 0.4755, "step": 17065 }, { "epoch": 2.67, "grad_norm": 18.9456889478899, "learning_rate": 6.441685243775097e-07, "loss": 0.4129, "step": 17066 }, { "epoch": 2.67, "grad_norm": 27.729086738170775, "learning_rate": 6.435731790401833e-07, "loss": 0.5379, "step": 17067 }, { "epoch": 2.67, "grad_norm": 17.976227029806033, "learning_rate": 6.429780997941626e-07, "loss": 0.447, "step": 17068 }, { "epoch": 2.67, "grad_norm": 22.96528416714328, "learning_rate": 6.423832866563696e-07, "loss": 0.4098, "step": 17069 }, { "epoch": 2.67, "grad_norm": 30.1824875218266, "learning_rate": 6.417887396437217e-07, "loss": 0.4455, "step": 17070 }, { "epoch": 2.67, "grad_norm": 25.392324768098185, "learning_rate": 6.411944587731279e-07, "loss": 0.3924, "step": 17071 }, { "epoch": 2.67, "grad_norm": 19.80517515827337, "learning_rate": 6.406004440614911e-07, "loss": 0.409, "step": 17072 }, { "epoch": 2.67, "grad_norm": 21.29175992762999, "learning_rate": 6.400066955257023e-07, "loss": 0.4157, "step": 17073 }, { "epoch": 2.67, "grad_norm": 15.612353141023707, "learning_rate": 6.394132131826447e-07, "loss": 0.4413, "step": 17074 }, { "epoch": 2.67, "grad_norm": 24.05933694114306, "learning_rate": 6.388199970492015e-07, "loss": 0.4136, "step": 17075 }, { "epoch": 2.67, "grad_norm": 30.137079820429193, "learning_rate": 6.382270471422413e-07, "loss": 0.4603, "step": 17076 }, { "epoch": 2.67, "grad_norm": 26.955829232600486, "learning_rate": 6.376343634786264e-07, "loss": 0.4748, "step": 17077 }, { "epoch": 2.67, "grad_norm": 34.27518691882161, "learning_rate": 6.370419460752153e-07, "loss": 0.4157, "step": 17078 }, { "epoch": 2.67, "grad_norm": 27.31960892582676, "learning_rate": 6.364497949488524e-07, "loss": 0.4197, "step": 17079 }, { "epoch": 2.67, "grad_norm": 25.238343928804962, "learning_rate": 6.35857910116382e-07, "loss": 0.4026, "step": 17080 }, { "epoch": 2.67, "grad_norm": 21.92394860997047, "learning_rate": 6.352662915946317e-07, "loss": 0.3977, "step": 17081 }, { "epoch": 2.67, "grad_norm": 35.61218212490607, "learning_rate": 6.346749394004314e-07, "loss": 0.445, "step": 17082 }, { "epoch": 2.67, "grad_norm": 19.427349684647595, "learning_rate": 6.340838535505978e-07, "loss": 0.3971, "step": 17083 }, { "epoch": 2.67, "grad_norm": 17.6933223110442, "learning_rate": 6.334930340619405e-07, "loss": 0.4132, "step": 17084 }, { "epoch": 2.67, "grad_norm": 19.313131267683865, "learning_rate": 6.329024809512607e-07, "loss": 0.3959, "step": 17085 }, { "epoch": 2.67, "grad_norm": 31.41987764705529, "learning_rate": 6.323121942353549e-07, "loss": 0.469, "step": 17086 }, { "epoch": 2.67, "grad_norm": 20.329605457661362, "learning_rate": 6.317221739310109e-07, "loss": 0.4416, "step": 17087 }, { "epoch": 2.67, "grad_norm": 30.76519342568286, "learning_rate": 6.311324200550084e-07, "loss": 0.3996, "step": 17088 }, { "epoch": 2.67, "grad_norm": 19.547293628811232, "learning_rate": 6.305429326241208e-07, "loss": 0.4028, "step": 17089 }, { "epoch": 2.67, "grad_norm": 21.297110116154766, "learning_rate": 6.29953711655108e-07, "loss": 0.3779, "step": 17090 }, { "epoch": 2.67, "grad_norm": 19.02544492277749, "learning_rate": 6.29364757164731e-07, "loss": 0.4378, "step": 17091 }, { "epoch": 2.67, "grad_norm": 31.638128218862505, "learning_rate": 6.287760691697409e-07, "loss": 0.3771, "step": 17092 }, { "epoch": 2.67, "grad_norm": 28.34857753829879, "learning_rate": 6.281876476868764e-07, "loss": 0.4398, "step": 17093 }, { "epoch": 2.67, "grad_norm": 21.48143995246555, "learning_rate": 6.27599492732871e-07, "loss": 0.3982, "step": 17094 }, { "epoch": 2.67, "grad_norm": 34.10834487900786, "learning_rate": 6.270116043244545e-07, "loss": 0.4615, "step": 17095 }, { "epoch": 2.67, "grad_norm": 20.37226666944277, "learning_rate": 6.264239824783447e-07, "loss": 0.4009, "step": 17096 }, { "epoch": 2.67, "grad_norm": 18.788523795163485, "learning_rate": 6.258366272112537e-07, "loss": 0.4299, "step": 17097 }, { "epoch": 2.67, "grad_norm": 34.367802364904556, "learning_rate": 6.252495385398871e-07, "loss": 0.4405, "step": 17098 }, { "epoch": 2.67, "grad_norm": 19.54238221033468, "learning_rate": 6.246627164809371e-07, "loss": 0.4143, "step": 17099 }, { "epoch": 2.67, "grad_norm": 14.206419770341999, "learning_rate": 6.240761610510948e-07, "loss": 0.4049, "step": 17100 }, { "epoch": 2.67, "grad_norm": 25.004906530504943, "learning_rate": 6.234898722670435e-07, "loss": 0.4295, "step": 17101 }, { "epoch": 2.67, "grad_norm": 17.665223184318045, "learning_rate": 6.229038501454532e-07, "loss": 0.4223, "step": 17102 }, { "epoch": 2.67, "grad_norm": 27.555947482896414, "learning_rate": 6.223180947029939e-07, "loss": 0.4118, "step": 17103 }, { "epoch": 2.67, "grad_norm": 36.97469000283389, "learning_rate": 6.217326059563211e-07, "loss": 0.5406, "step": 17104 }, { "epoch": 2.67, "grad_norm": 21.676037536469888, "learning_rate": 6.211473839220883e-07, "loss": 0.4801, "step": 17105 }, { "epoch": 2.67, "grad_norm": 16.2837532172914, "learning_rate": 6.205624286169354e-07, "loss": 0.4184, "step": 17106 }, { "epoch": 2.67, "grad_norm": 22.598559802165607, "learning_rate": 6.199777400575013e-07, "loss": 0.4292, "step": 17107 }, { "epoch": 2.67, "grad_norm": 20.6410971124788, "learning_rate": 6.193933182604128e-07, "loss": 0.4302, "step": 17108 }, { "epoch": 2.67, "grad_norm": 22.660236184239434, "learning_rate": 6.188091632422921e-07, "loss": 0.4313, "step": 17109 }, { "epoch": 2.67, "grad_norm": 17.236551590030146, "learning_rate": 6.182252750197493e-07, "loss": 0.4422, "step": 17110 }, { "epoch": 2.67, "grad_norm": 20.96506350085633, "learning_rate": 6.17641653609391e-07, "loss": 0.3998, "step": 17111 }, { "epoch": 2.67, "grad_norm": 22.552410013389125, "learning_rate": 6.170582990278173e-07, "loss": 0.4117, "step": 17112 }, { "epoch": 2.67, "grad_norm": 18.934543313498008, "learning_rate": 6.16475211291615e-07, "loss": 0.4282, "step": 17113 }, { "epoch": 2.67, "grad_norm": 15.85941118320791, "learning_rate": 6.158923904173709e-07, "loss": 0.4317, "step": 17114 }, { "epoch": 2.67, "grad_norm": 20.408329756931018, "learning_rate": 6.153098364216548e-07, "loss": 0.4434, "step": 17115 }, { "epoch": 2.67, "grad_norm": 30.182374382344552, "learning_rate": 6.147275493210381e-07, "loss": 0.4152, "step": 17116 }, { "epoch": 2.67, "grad_norm": 23.10646288741191, "learning_rate": 6.141455291320808e-07, "loss": 0.4617, "step": 17117 }, { "epoch": 2.67, "grad_norm": 15.288819508480705, "learning_rate": 6.135637758713342e-07, "loss": 0.3848, "step": 17118 }, { "epoch": 2.67, "grad_norm": 27.006153999233906, "learning_rate": 6.129822895553417e-07, "loss": 0.4295, "step": 17119 }, { "epoch": 2.67, "grad_norm": 15.434662383401378, "learning_rate": 6.124010702006411e-07, "loss": 0.3879, "step": 17120 }, { "epoch": 2.67, "grad_norm": 16.224438716109823, "learning_rate": 6.118201178237626e-07, "loss": 0.4242, "step": 17121 }, { "epoch": 2.67, "grad_norm": 14.356076925460892, "learning_rate": 6.112394324412308e-07, "loss": 0.4398, "step": 17122 }, { "epoch": 2.67, "grad_norm": 28.671184323590666, "learning_rate": 6.106590140695545e-07, "loss": 0.4515, "step": 17123 }, { "epoch": 2.67, "grad_norm": 30.10050385512265, "learning_rate": 6.100788627252441e-07, "loss": 0.443, "step": 17124 }, { "epoch": 2.67, "grad_norm": 20.642741107107778, "learning_rate": 6.094989784247973e-07, "loss": 0.4382, "step": 17125 }, { "epoch": 2.68, "grad_norm": 24.19994401930214, "learning_rate": 6.089193611847066e-07, "loss": 0.405, "step": 17126 }, { "epoch": 2.68, "grad_norm": 24.401677884194676, "learning_rate": 6.083400110214577e-07, "loss": 0.4165, "step": 17127 }, { "epoch": 2.68, "grad_norm": 22.75780475667396, "learning_rate": 6.077609279515217e-07, "loss": 0.4232, "step": 17128 }, { "epoch": 2.68, "grad_norm": 25.733366327489033, "learning_rate": 6.071821119913713e-07, "loss": 0.4643, "step": 17129 }, { "epoch": 2.68, "grad_norm": 18.175979693518443, "learning_rate": 6.066035631574685e-07, "loss": 0.4021, "step": 17130 }, { "epoch": 2.68, "grad_norm": 26.234837500134557, "learning_rate": 6.060252814662637e-07, "loss": 0.404, "step": 17131 }, { "epoch": 2.68, "grad_norm": 24.97044382449291, "learning_rate": 6.05447266934206e-07, "loss": 0.4089, "step": 17132 }, { "epoch": 2.68, "grad_norm": 23.005256356697373, "learning_rate": 6.04869519577731e-07, "loss": 0.4979, "step": 17133 }, { "epoch": 2.68, "grad_norm": 32.22426743259202, "learning_rate": 6.042920394132712e-07, "loss": 0.4754, "step": 17134 }, { "epoch": 2.68, "grad_norm": 31.007453232056136, "learning_rate": 6.037148264572512e-07, "loss": 0.4417, "step": 17135 }, { "epoch": 2.68, "grad_norm": 20.632163141931105, "learning_rate": 6.031378807260823e-07, "loss": 0.4203, "step": 17136 }, { "epoch": 2.68, "grad_norm": 15.153728765022663, "learning_rate": 6.025612022361771e-07, "loss": 0.3838, "step": 17137 }, { "epoch": 2.68, "grad_norm": 22.110892726310215, "learning_rate": 6.019847910039334e-07, "loss": 0.4138, "step": 17138 }, { "epoch": 2.68, "grad_norm": 24.509416667543377, "learning_rate": 6.014086470457448e-07, "loss": 0.3619, "step": 17139 }, { "epoch": 2.68, "grad_norm": 22.1188525284285, "learning_rate": 6.008327703779948e-07, "loss": 0.5014, "step": 17140 }, { "epoch": 2.68, "grad_norm": 27.94531645440214, "learning_rate": 6.002571610170627e-07, "loss": 0.5094, "step": 17141 }, { "epoch": 2.68, "grad_norm": 28.45396947262181, "learning_rate": 5.996818189793207e-07, "loss": 0.4211, "step": 17142 }, { "epoch": 2.68, "grad_norm": 24.533463448691474, "learning_rate": 5.991067442811272e-07, "loss": 0.4518, "step": 17143 }, { "epoch": 2.68, "grad_norm": 19.161471037321725, "learning_rate": 5.985319369388376e-07, "loss": 0.4032, "step": 17144 }, { "epoch": 2.68, "grad_norm": 12.837735811080933, "learning_rate": 5.979573969688001e-07, "loss": 0.3626, "step": 17145 }, { "epoch": 2.68, "grad_norm": 21.65626426244929, "learning_rate": 5.973831243873551e-07, "loss": 0.4867, "step": 17146 }, { "epoch": 2.68, "grad_norm": 15.804789619662726, "learning_rate": 5.96809119210835e-07, "loss": 0.3739, "step": 17147 }, { "epoch": 2.68, "grad_norm": 25.26088088300529, "learning_rate": 5.96235381455561e-07, "loss": 0.4652, "step": 17148 }, { "epoch": 2.68, "grad_norm": 32.06862106091011, "learning_rate": 5.956619111378514e-07, "loss": 0.468, "step": 17149 }, { "epoch": 2.68, "grad_norm": 18.802936646778313, "learning_rate": 5.950887082740153e-07, "loss": 0.4661, "step": 17150 }, { "epoch": 2.68, "grad_norm": 31.468231054836213, "learning_rate": 5.945157728803563e-07, "loss": 0.4134, "step": 17151 }, { "epoch": 2.68, "grad_norm": 31.13163359167715, "learning_rate": 5.93943104973167e-07, "loss": 0.5141, "step": 17152 }, { "epoch": 2.68, "grad_norm": 17.594273805020006, "learning_rate": 5.93370704568732e-07, "loss": 0.4046, "step": 17153 }, { "epoch": 2.68, "grad_norm": 22.04453531877571, "learning_rate": 5.927985716833317e-07, "loss": 0.4324, "step": 17154 }, { "epoch": 2.68, "grad_norm": 29.099627139363037, "learning_rate": 5.922267063332376e-07, "loss": 0.454, "step": 17155 }, { "epoch": 2.68, "grad_norm": 18.01097826282531, "learning_rate": 5.916551085347134e-07, "loss": 0.4073, "step": 17156 }, { "epoch": 2.68, "grad_norm": 19.356552635662474, "learning_rate": 5.910837783040113e-07, "loss": 0.4202, "step": 17157 }, { "epoch": 2.68, "grad_norm": 26.287884465615402, "learning_rate": 5.905127156573842e-07, "loss": 0.381, "step": 17158 }, { "epoch": 2.68, "grad_norm": 22.636182347838588, "learning_rate": 5.899419206110702e-07, "loss": 0.4357, "step": 17159 }, { "epoch": 2.68, "grad_norm": 19.067440509782113, "learning_rate": 5.89371393181305e-07, "loss": 0.4253, "step": 17160 }, { "epoch": 2.68, "grad_norm": 22.297473449668768, "learning_rate": 5.888011333843113e-07, "loss": 0.4758, "step": 17161 }, { "epoch": 2.68, "grad_norm": 24.258932992607946, "learning_rate": 5.882311412363073e-07, "loss": 0.4643, "step": 17162 }, { "epoch": 2.68, "grad_norm": 18.8381388852069, "learning_rate": 5.876614167535044e-07, "loss": 0.504, "step": 17163 }, { "epoch": 2.68, "grad_norm": 16.209615633281366, "learning_rate": 5.870919599521052e-07, "loss": 0.4723, "step": 17164 }, { "epoch": 2.68, "grad_norm": 21.734926843390635, "learning_rate": 5.865227708483034e-07, "loss": 0.4168, "step": 17165 }, { "epoch": 2.68, "grad_norm": 20.323562420114825, "learning_rate": 5.859538494582895e-07, "loss": 0.4534, "step": 17166 }, { "epoch": 2.68, "grad_norm": 26.53938933118751, "learning_rate": 5.853851957982381e-07, "loss": 0.4255, "step": 17167 }, { "epoch": 2.68, "grad_norm": 19.755254837116833, "learning_rate": 5.848168098843265e-07, "loss": 0.3997, "step": 17168 }, { "epoch": 2.68, "grad_norm": 19.590963999854182, "learning_rate": 5.842486917327162e-07, "loss": 0.4001, "step": 17169 }, { "epoch": 2.68, "grad_norm": 27.10782022277645, "learning_rate": 5.836808413595641e-07, "loss": 0.4644, "step": 17170 }, { "epoch": 2.68, "grad_norm": 29.15008667856092, "learning_rate": 5.831132587810228e-07, "loss": 0.4219, "step": 17171 }, { "epoch": 2.68, "grad_norm": 27.068763426568403, "learning_rate": 5.825459440132308e-07, "loss": 0.4718, "step": 17172 }, { "epoch": 2.68, "grad_norm": 20.258163998574254, "learning_rate": 5.819788970723217e-07, "loss": 0.4259, "step": 17173 }, { "epoch": 2.68, "grad_norm": 25.124331967762906, "learning_rate": 5.814121179744248e-07, "loss": 0.4675, "step": 17174 }, { "epoch": 2.68, "grad_norm": 17.22273095093739, "learning_rate": 5.80845606735656e-07, "loss": 0.3965, "step": 17175 }, { "epoch": 2.68, "grad_norm": 20.882530648626766, "learning_rate": 5.802793633721304e-07, "loss": 0.5005, "step": 17176 }, { "epoch": 2.68, "grad_norm": 22.156578297933518, "learning_rate": 5.797133878999484e-07, "loss": 0.5105, "step": 17177 }, { "epoch": 2.68, "grad_norm": 16.17856966759336, "learning_rate": 5.791476803352058e-07, "loss": 0.3771, "step": 17178 }, { "epoch": 2.68, "grad_norm": 23.08959148490535, "learning_rate": 5.785822406939934e-07, "loss": 0.4779, "step": 17179 }, { "epoch": 2.68, "grad_norm": 33.8620337184854, "learning_rate": 5.780170689923902e-07, "loss": 0.477, "step": 17180 }, { "epoch": 2.68, "grad_norm": 21.712296244686986, "learning_rate": 5.774521652464715e-07, "loss": 0.4023, "step": 17181 }, { "epoch": 2.68, "grad_norm": 16.410673496091725, "learning_rate": 5.768875294722987e-07, "loss": 0.4226, "step": 17182 }, { "epoch": 2.68, "grad_norm": 30.004939134636484, "learning_rate": 5.763231616859333e-07, "loss": 0.4388, "step": 17183 }, { "epoch": 2.68, "grad_norm": 20.166109351432183, "learning_rate": 5.757590619034236e-07, "loss": 0.4416, "step": 17184 }, { "epoch": 2.68, "grad_norm": 19.43622743482468, "learning_rate": 5.751952301408148e-07, "loss": 0.4666, "step": 17185 }, { "epoch": 2.68, "grad_norm": 12.429942193469923, "learning_rate": 5.746316664141394e-07, "loss": 0.3997, "step": 17186 }, { "epoch": 2.68, "grad_norm": 16.667216781736396, "learning_rate": 5.740683707394256e-07, "loss": 0.3777, "step": 17187 }, { "epoch": 2.68, "grad_norm": 19.43765278814437, "learning_rate": 5.735053431326931e-07, "loss": 0.4392, "step": 17188 }, { "epoch": 2.68, "grad_norm": 28.492888375320824, "learning_rate": 5.729425836099556e-07, "loss": 0.4162, "step": 17189 }, { "epoch": 2.69, "grad_norm": 22.973017063949513, "learning_rate": 5.723800921872147e-07, "loss": 0.3937, "step": 17190 }, { "epoch": 2.69, "grad_norm": 21.098068979821022, "learning_rate": 5.718178688804699e-07, "loss": 0.4462, "step": 17191 }, { "epoch": 2.69, "grad_norm": 21.438251553087387, "learning_rate": 5.712559137057094e-07, "loss": 0.4125, "step": 17192 }, { "epoch": 2.69, "grad_norm": 21.594294559109997, "learning_rate": 5.706942266789162e-07, "loss": 0.4147, "step": 17193 }, { "epoch": 2.69, "grad_norm": 26.007051574218718, "learning_rate": 5.701328078160606e-07, "loss": 0.4217, "step": 17194 }, { "epoch": 2.69, "grad_norm": 28.82472530363959, "learning_rate": 5.695716571331134e-07, "loss": 0.4465, "step": 17195 }, { "epoch": 2.69, "grad_norm": 21.328374989759716, "learning_rate": 5.690107746460316e-07, "loss": 0.4311, "step": 17196 }, { "epoch": 2.69, "grad_norm": 28.769635109024442, "learning_rate": 5.684501603707671e-07, "loss": 0.4654, "step": 17197 }, { "epoch": 2.69, "grad_norm": 33.43446302437116, "learning_rate": 5.678898143232614e-07, "loss": 0.4013, "step": 17198 }, { "epoch": 2.69, "grad_norm": 18.164500736159326, "learning_rate": 5.673297365194508e-07, "loss": 0.4557, "step": 17199 }, { "epoch": 2.69, "grad_norm": 23.908793173259525, "learning_rate": 5.667699269752658e-07, "loss": 0.5235, "step": 17200 }, { "epoch": 2.69, "grad_norm": 17.83536273800507, "learning_rate": 5.66210385706627e-07, "loss": 0.3905, "step": 17201 }, { "epoch": 2.69, "grad_norm": 24.226073741441912, "learning_rate": 5.65651112729445e-07, "loss": 0.4059, "step": 17202 }, { "epoch": 2.69, "grad_norm": 23.18215302873196, "learning_rate": 5.650921080596261e-07, "loss": 0.4531, "step": 17203 }, { "epoch": 2.69, "grad_norm": 26.992592838627985, "learning_rate": 5.645333717130685e-07, "loss": 0.3939, "step": 17204 }, { "epoch": 2.69, "grad_norm": 24.218793059280017, "learning_rate": 5.63974903705663e-07, "loss": 0.4742, "step": 17205 }, { "epoch": 2.69, "grad_norm": 16.178349945725383, "learning_rate": 5.634167040532922e-07, "loss": 0.4189, "step": 17206 }, { "epoch": 2.69, "grad_norm": 38.621292027323015, "learning_rate": 5.628587727718282e-07, "loss": 0.4737, "step": 17207 }, { "epoch": 2.69, "grad_norm": 24.099672147540957, "learning_rate": 5.623011098771391e-07, "loss": 0.4318, "step": 17208 }, { "epoch": 2.69, "grad_norm": 22.830996172230464, "learning_rate": 5.617437153850868e-07, "loss": 0.4269, "step": 17209 }, { "epoch": 2.69, "grad_norm": 25.3327741915449, "learning_rate": 5.611865893115243e-07, "loss": 0.4992, "step": 17210 }, { "epoch": 2.69, "grad_norm": 17.42511808120721, "learning_rate": 5.60629731672292e-07, "loss": 0.4356, "step": 17211 }, { "epoch": 2.69, "grad_norm": 30.328067335322576, "learning_rate": 5.600731424832273e-07, "loss": 0.4427, "step": 17212 }, { "epoch": 2.69, "grad_norm": 13.915881625417073, "learning_rate": 5.595168217601599e-07, "loss": 0.4896, "step": 17213 }, { "epoch": 2.69, "grad_norm": 19.41446837647619, "learning_rate": 5.589607695189136e-07, "loss": 0.401, "step": 17214 }, { "epoch": 2.69, "grad_norm": 21.635645349931963, "learning_rate": 5.584049857752993e-07, "loss": 0.4354, "step": 17215 }, { "epoch": 2.69, "grad_norm": 68.99571866595082, "learning_rate": 5.578494705451232e-07, "loss": 0.5034, "step": 17216 }, { "epoch": 2.69, "grad_norm": 18.323278459515997, "learning_rate": 5.572942238441847e-07, "loss": 0.3747, "step": 17217 }, { "epoch": 2.69, "grad_norm": 22.21319699409457, "learning_rate": 5.567392456882758e-07, "loss": 0.4258, "step": 17218 }, { "epoch": 2.69, "grad_norm": 26.219894182453054, "learning_rate": 5.561845360931784e-07, "loss": 0.4587, "step": 17219 }, { "epoch": 2.69, "grad_norm": 24.46678305962246, "learning_rate": 5.556300950746684e-07, "loss": 0.409, "step": 17220 }, { "epoch": 2.69, "grad_norm": 26.733542125577035, "learning_rate": 5.550759226485126e-07, "loss": 0.508, "step": 17221 }, { "epoch": 2.69, "grad_norm": 12.443719254239705, "learning_rate": 5.545220188304723e-07, "loss": 0.3717, "step": 17222 }, { "epoch": 2.69, "grad_norm": 19.11651088669281, "learning_rate": 5.539683836363019e-07, "loss": 0.4335, "step": 17223 }, { "epoch": 2.69, "grad_norm": 21.367746253091465, "learning_rate": 5.534150170817431e-07, "loss": 0.3919, "step": 17224 }, { "epoch": 2.69, "grad_norm": 23.107033985613825, "learning_rate": 5.52861919182538e-07, "loss": 0.5067, "step": 17225 }, { "epoch": 2.69, "grad_norm": 25.81298360590927, "learning_rate": 5.523090899544104e-07, "loss": 0.3868, "step": 17226 }, { "epoch": 2.69, "grad_norm": 22.31718210774328, "learning_rate": 5.517565294130877e-07, "loss": 0.5, "step": 17227 }, { "epoch": 2.69, "grad_norm": 36.27488346861624, "learning_rate": 5.51204237574281e-07, "loss": 0.448, "step": 17228 }, { "epoch": 2.69, "grad_norm": 21.9085475282073, "learning_rate": 5.506522144536975e-07, "loss": 0.4176, "step": 17229 }, { "epoch": 2.69, "grad_norm": 27.731048607476463, "learning_rate": 5.501004600670401e-07, "loss": 0.4536, "step": 17230 }, { "epoch": 2.69, "grad_norm": 16.953519242620448, "learning_rate": 5.495489744299965e-07, "loss": 0.4309, "step": 17231 }, { "epoch": 2.69, "grad_norm": 21.110863016909242, "learning_rate": 5.489977575582506e-07, "loss": 0.3961, "step": 17232 }, { "epoch": 2.69, "grad_norm": 18.042106379043073, "learning_rate": 5.4844680946748e-07, "loss": 0.3987, "step": 17233 }, { "epoch": 2.69, "grad_norm": 19.635019648119307, "learning_rate": 5.478961301733531e-07, "loss": 0.4655, "step": 17234 }, { "epoch": 2.69, "grad_norm": 23.573123024646776, "learning_rate": 5.473457196915332e-07, "loss": 0.4032, "step": 17235 }, { "epoch": 2.69, "grad_norm": 21.71383649078869, "learning_rate": 5.467955780376688e-07, "loss": 0.424, "step": 17236 }, { "epoch": 2.69, "grad_norm": 32.21400051926023, "learning_rate": 5.462457052274084e-07, "loss": 0.4757, "step": 17237 }, { "epoch": 2.69, "grad_norm": 19.520777585629546, "learning_rate": 5.456961012763906e-07, "loss": 0.4792, "step": 17238 }, { "epoch": 2.69, "grad_norm": 19.843145934995647, "learning_rate": 5.451467662002452e-07, "loss": 0.4313, "step": 17239 }, { "epoch": 2.69, "grad_norm": 30.24670527999096, "learning_rate": 5.445977000145952e-07, "loss": 0.4449, "step": 17240 }, { "epoch": 2.69, "grad_norm": 20.887309504627083, "learning_rate": 5.440489027350548e-07, "loss": 0.4383, "step": 17241 }, { "epoch": 2.69, "grad_norm": 26.896255175288672, "learning_rate": 5.435003743772305e-07, "loss": 0.4563, "step": 17242 }, { "epoch": 2.69, "grad_norm": 15.937445373067547, "learning_rate": 5.429521149567263e-07, "loss": 0.4015, "step": 17243 }, { "epoch": 2.69, "grad_norm": 17.625888676085836, "learning_rate": 5.424041244891298e-07, "loss": 0.397, "step": 17244 }, { "epoch": 2.69, "grad_norm": 25.574446228644778, "learning_rate": 5.418564029900286e-07, "loss": 0.5172, "step": 17245 }, { "epoch": 2.69, "grad_norm": 16.00128933528401, "learning_rate": 5.413089504749979e-07, "loss": 0.4048, "step": 17246 }, { "epoch": 2.69, "grad_norm": 21.8865913729184, "learning_rate": 5.407617669596066e-07, "loss": 0.4687, "step": 17247 }, { "epoch": 2.69, "grad_norm": 26.791511789683064, "learning_rate": 5.402148524594198e-07, "loss": 0.4571, "step": 17248 }, { "epoch": 2.69, "grad_norm": 19.402168720117135, "learning_rate": 5.396682069899861e-07, "loss": 0.3675, "step": 17249 }, { "epoch": 2.69, "grad_norm": 18.038101778678637, "learning_rate": 5.391218305668566e-07, "loss": 0.4146, "step": 17250 }, { "epoch": 2.69, "grad_norm": 23.849226415182248, "learning_rate": 5.385757232055655e-07, "loss": 0.4134, "step": 17251 }, { "epoch": 2.69, "grad_norm": 20.840673754751307, "learning_rate": 5.380298849216481e-07, "loss": 0.4323, "step": 17252 }, { "epoch": 2.69, "grad_norm": 18.10852311346565, "learning_rate": 5.374843157306253e-07, "loss": 0.392, "step": 17253 }, { "epoch": 2.7, "grad_norm": 18.352512086178507, "learning_rate": 5.369390156480126e-07, "loss": 0.4611, "step": 17254 }, { "epoch": 2.7, "grad_norm": 13.725599810451797, "learning_rate": 5.363939846893184e-07, "loss": 0.4011, "step": 17255 }, { "epoch": 2.7, "grad_norm": 21.855961721961, "learning_rate": 5.35849222870044e-07, "loss": 0.465, "step": 17256 }, { "epoch": 2.7, "grad_norm": 28.07508804166514, "learning_rate": 5.353047302056802e-07, "loss": 0.4646, "step": 17257 }, { "epoch": 2.7, "grad_norm": 15.49336396481214, "learning_rate": 5.347605067117134e-07, "loss": 0.3989, "step": 17258 }, { "epoch": 2.7, "grad_norm": 20.49373117952393, "learning_rate": 5.342165524036224e-07, "loss": 0.4778, "step": 17259 }, { "epoch": 2.7, "grad_norm": 28.8572480771274, "learning_rate": 5.33672867296875e-07, "loss": 0.4625, "step": 17260 }, { "epoch": 2.7, "grad_norm": 22.662540146654848, "learning_rate": 5.331294514069318e-07, "loss": 0.4573, "step": 17261 }, { "epoch": 2.7, "grad_norm": 27.067351887518708, "learning_rate": 5.325863047492496e-07, "loss": 0.4378, "step": 17262 }, { "epoch": 2.7, "grad_norm": 23.64252606732429, "learning_rate": 5.320434273392738e-07, "loss": 0.4271, "step": 17263 }, { "epoch": 2.7, "grad_norm": 24.61442315262329, "learning_rate": 5.315008191924464e-07, "loss": 0.4566, "step": 17264 }, { "epoch": 2.7, "grad_norm": 21.27401069011807, "learning_rate": 5.309584803241973e-07, "loss": 0.4084, "step": 17265 }, { "epoch": 2.7, "grad_norm": 27.305309391170642, "learning_rate": 5.304164107499477e-07, "loss": 0.4578, "step": 17266 }, { "epoch": 2.7, "grad_norm": 22.952895964459437, "learning_rate": 5.29874610485116e-07, "loss": 0.379, "step": 17267 }, { "epoch": 2.7, "grad_norm": 22.262228147157717, "learning_rate": 5.293330795451112e-07, "loss": 0.4231, "step": 17268 }, { "epoch": 2.7, "grad_norm": 25.623663939559407, "learning_rate": 5.287918179453344e-07, "loss": 0.5229, "step": 17269 }, { "epoch": 2.7, "grad_norm": 22.013308274250857, "learning_rate": 5.282508257011764e-07, "loss": 0.4515, "step": 17270 }, { "epoch": 2.7, "grad_norm": 16.646344360743818, "learning_rate": 5.27710102828024e-07, "loss": 0.3797, "step": 17271 }, { "epoch": 2.7, "grad_norm": 26.91870644510296, "learning_rate": 5.271696493412548e-07, "loss": 0.4468, "step": 17272 }, { "epoch": 2.7, "grad_norm": 18.81347788521876, "learning_rate": 5.266294652562409e-07, "loss": 0.4902, "step": 17273 }, { "epoch": 2.7, "grad_norm": 18.685796437551218, "learning_rate": 5.260895505883423e-07, "loss": 0.4316, "step": 17274 }, { "epoch": 2.7, "grad_norm": 18.414744642470925, "learning_rate": 5.255499053529145e-07, "loss": 0.4286, "step": 17275 }, { "epoch": 2.7, "grad_norm": 34.59554654223095, "learning_rate": 5.250105295653052e-07, "loss": 0.3948, "step": 17276 }, { "epoch": 2.7, "grad_norm": 31.880595986161005, "learning_rate": 5.244714232408544e-07, "loss": 0.463, "step": 17277 }, { "epoch": 2.7, "grad_norm": 19.55392232655942, "learning_rate": 5.23932586394893e-07, "loss": 0.4426, "step": 17278 }, { "epoch": 2.7, "grad_norm": 13.202779860189478, "learning_rate": 5.233940190427456e-07, "loss": 0.4516, "step": 17279 }, { "epoch": 2.7, "grad_norm": 20.452537788284626, "learning_rate": 5.228557211997276e-07, "loss": 0.3993, "step": 17280 }, { "epoch": 2.7, "grad_norm": 28.07642177092423, "learning_rate": 5.223176928811502e-07, "loss": 0.4256, "step": 17281 }, { "epoch": 2.7, "grad_norm": 13.108226470057977, "learning_rate": 5.217799341023122e-07, "loss": 0.3946, "step": 17282 }, { "epoch": 2.7, "grad_norm": 19.841880061435862, "learning_rate": 5.212424448785092e-07, "loss": 0.4574, "step": 17283 }, { "epoch": 2.7, "grad_norm": 27.884641124032992, "learning_rate": 5.207052252250266e-07, "loss": 0.4703, "step": 17284 }, { "epoch": 2.7, "grad_norm": 18.883343293380833, "learning_rate": 5.201682751571402e-07, "loss": 0.4807, "step": 17285 }, { "epoch": 2.7, "grad_norm": 18.402956927065283, "learning_rate": 5.196315946901254e-07, "loss": 0.4299, "step": 17286 }, { "epoch": 2.7, "grad_norm": 19.7772194234793, "learning_rate": 5.1909518383924e-07, "loss": 0.4514, "step": 17287 }, { "epoch": 2.7, "grad_norm": 4.206078323145508, "learning_rate": 5.185590426197406e-07, "loss": 0.4958, "step": 17288 }, { "epoch": 2.7, "grad_norm": 24.493932784104928, "learning_rate": 5.180231710468775e-07, "loss": 0.4336, "step": 17289 }, { "epoch": 2.7, "grad_norm": 20.3163464168877, "learning_rate": 5.174875691358894e-07, "loss": 0.4442, "step": 17290 }, { "epoch": 2.7, "grad_norm": 19.3136152806508, "learning_rate": 5.169522369020052e-07, "loss": 0.4246, "step": 17291 }, { "epoch": 2.7, "grad_norm": 23.089451561457448, "learning_rate": 5.164171743604529e-07, "loss": 0.5384, "step": 17292 }, { "epoch": 2.7, "grad_norm": 21.51344816114946, "learning_rate": 5.158823815264491e-07, "loss": 0.445, "step": 17293 }, { "epoch": 2.7, "grad_norm": 30.97963196097292, "learning_rate": 5.153478584152028e-07, "loss": 0.4123, "step": 17294 }, { "epoch": 2.7, "grad_norm": 22.122142329044262, "learning_rate": 5.148136050419139e-07, "loss": 0.4176, "step": 17295 }, { "epoch": 2.7, "grad_norm": 17.935988362513783, "learning_rate": 5.14279621421776e-07, "loss": 0.3834, "step": 17296 }, { "epoch": 2.7, "grad_norm": 28.94540360780038, "learning_rate": 5.13745907569978e-07, "loss": 0.5192, "step": 17297 }, { "epoch": 2.7, "grad_norm": 22.993483268059627, "learning_rate": 5.132124635016977e-07, "loss": 0.4088, "step": 17298 }, { "epoch": 2.7, "grad_norm": 33.0552618562427, "learning_rate": 5.126792892321064e-07, "loss": 0.4924, "step": 17299 }, { "epoch": 2.7, "grad_norm": 24.015626259569817, "learning_rate": 5.121463847763641e-07, "loss": 0.442, "step": 17300 }, { "epoch": 2.7, "grad_norm": 19.004528866612098, "learning_rate": 5.116137501496288e-07, "loss": 0.4004, "step": 17301 }, { "epoch": 2.7, "grad_norm": 30.554486900496265, "learning_rate": 5.110813853670482e-07, "loss": 0.451, "step": 17302 }, { "epoch": 2.7, "grad_norm": 15.652322624921526, "learning_rate": 5.105492904437636e-07, "loss": 0.4929, "step": 17303 }, { "epoch": 2.7, "grad_norm": 20.202941979260043, "learning_rate": 5.100174653949031e-07, "loss": 0.4504, "step": 17304 }, { "epoch": 2.7, "grad_norm": 22.288963161211036, "learning_rate": 5.094859102355953e-07, "loss": 0.3748, "step": 17305 }, { "epoch": 2.7, "grad_norm": 20.11259556259227, "learning_rate": 5.089546249809584e-07, "loss": 0.4413, "step": 17306 }, { "epoch": 2.7, "grad_norm": 39.29029262048367, "learning_rate": 5.08423609646097e-07, "loss": 0.4246, "step": 17307 }, { "epoch": 2.7, "grad_norm": 16.690272692919102, "learning_rate": 5.078928642461178e-07, "loss": 0.4498, "step": 17308 }, { "epoch": 2.7, "grad_norm": 21.085286966064466, "learning_rate": 5.073623887961121e-07, "loss": 0.4678, "step": 17309 }, { "epoch": 2.7, "grad_norm": 20.842060203629657, "learning_rate": 5.068321833111667e-07, "loss": 0.4252, "step": 17310 }, { "epoch": 2.7, "grad_norm": 21.359827984218143, "learning_rate": 5.063022478063617e-07, "loss": 0.3983, "step": 17311 }, { "epoch": 2.7, "grad_norm": 24.35831420844723, "learning_rate": 5.057725822967663e-07, "loss": 0.4112, "step": 17312 }, { "epoch": 2.7, "grad_norm": 18.621102548928857, "learning_rate": 5.05243186797445e-07, "loss": 0.4945, "step": 17313 }, { "epoch": 2.7, "grad_norm": 19.109947128184082, "learning_rate": 5.047140613234524e-07, "loss": 0.3994, "step": 17314 }, { "epoch": 2.7, "grad_norm": 19.81475669737337, "learning_rate": 5.041852058898389e-07, "loss": 0.4711, "step": 17315 }, { "epoch": 2.7, "grad_norm": 22.809144760978814, "learning_rate": 5.036566205116422e-07, "loss": 0.4196, "step": 17316 }, { "epoch": 2.7, "grad_norm": 28.800104107147042, "learning_rate": 5.031283052038949e-07, "loss": 0.4705, "step": 17317 }, { "epoch": 2.71, "grad_norm": 21.571628230613605, "learning_rate": 5.026002599816248e-07, "loss": 0.4439, "step": 17318 }, { "epoch": 2.71, "grad_norm": 13.640442659638463, "learning_rate": 5.02072484859848e-07, "loss": 0.4091, "step": 17319 }, { "epoch": 2.71, "grad_norm": 23.543029136953702, "learning_rate": 5.015449798535721e-07, "loss": 0.4187, "step": 17320 }, { "epoch": 2.71, "grad_norm": 26.946093871726127, "learning_rate": 5.010177449778009e-07, "loss": 0.5137, "step": 17321 }, { "epoch": 2.71, "grad_norm": 28.47027117143808, "learning_rate": 5.004907802475278e-07, "loss": 0.4929, "step": 17322 }, { "epoch": 2.71, "grad_norm": 20.634373193810383, "learning_rate": 4.999640856777421e-07, "loss": 0.4675, "step": 17323 }, { "epoch": 2.71, "grad_norm": 22.77781522887984, "learning_rate": 4.994376612834185e-07, "loss": 0.4162, "step": 17324 }, { "epoch": 2.71, "grad_norm": 22.850471273597794, "learning_rate": 4.989115070795303e-07, "loss": 0.4495, "step": 17325 }, { "epoch": 2.71, "grad_norm": 23.957795243371162, "learning_rate": 4.983856230810402e-07, "loss": 0.5059, "step": 17326 }, { "epoch": 2.71, "grad_norm": 21.66063598569354, "learning_rate": 4.978600093029062e-07, "loss": 0.4327, "step": 17327 }, { "epoch": 2.71, "grad_norm": 26.61425020913945, "learning_rate": 4.973346657600752e-07, "loss": 0.4447, "step": 17328 }, { "epoch": 2.71, "grad_norm": 22.46283718401621, "learning_rate": 4.968095924674854e-07, "loss": 0.4474, "step": 17329 }, { "epoch": 2.71, "grad_norm": 25.523382559309375, "learning_rate": 4.962847894400724e-07, "loss": 0.4229, "step": 17330 }, { "epoch": 2.71, "grad_norm": 39.79228124626586, "learning_rate": 4.95760256692761e-07, "loss": 0.5159, "step": 17331 }, { "epoch": 2.71, "grad_norm": 25.669889569488955, "learning_rate": 4.952359942404672e-07, "loss": 0.4275, "step": 17332 }, { "epoch": 2.71, "grad_norm": 22.177408284016053, "learning_rate": 4.947120020981034e-07, "loss": 0.3933, "step": 17333 }, { "epoch": 2.71, "grad_norm": 15.664859572605161, "learning_rate": 4.941882802805675e-07, "loss": 0.4255, "step": 17334 }, { "epoch": 2.71, "grad_norm": 24.90021194727825, "learning_rate": 4.936648288027568e-07, "loss": 0.4106, "step": 17335 }, { "epoch": 2.71, "grad_norm": 18.637261371792675, "learning_rate": 4.931416476795592e-07, "loss": 0.4176, "step": 17336 }, { "epoch": 2.71, "grad_norm": 20.438985157442556, "learning_rate": 4.926187369258495e-07, "loss": 0.4423, "step": 17337 }, { "epoch": 2.71, "grad_norm": 18.940822931403154, "learning_rate": 4.920960965565036e-07, "loss": 0.4107, "step": 17338 }, { "epoch": 2.71, "grad_norm": 21.160953824999524, "learning_rate": 4.915737265863807e-07, "loss": 0.4597, "step": 17339 }, { "epoch": 2.71, "grad_norm": 14.406864437603977, "learning_rate": 4.9105162703034e-07, "loss": 0.3596, "step": 17340 }, { "epoch": 2.71, "grad_norm": 23.127785650216396, "learning_rate": 4.905297979032264e-07, "loss": 0.4405, "step": 17341 }, { "epoch": 2.71, "grad_norm": 12.541078316342178, "learning_rate": 4.900082392198835e-07, "loss": 0.4054, "step": 17342 }, { "epoch": 2.71, "grad_norm": 15.943470575163786, "learning_rate": 4.894869509951449e-07, "loss": 0.3533, "step": 17343 }, { "epoch": 2.71, "grad_norm": 22.57018770795412, "learning_rate": 4.889659332438334e-07, "loss": 0.4153, "step": 17344 }, { "epoch": 2.71, "grad_norm": 24.409024522917367, "learning_rate": 4.884451859807648e-07, "loss": 0.4465, "step": 17345 }, { "epoch": 2.71, "grad_norm": 16.92758619534769, "learning_rate": 4.879247092207518e-07, "loss": 0.4631, "step": 17346 }, { "epoch": 2.71, "grad_norm": 24.209903256734332, "learning_rate": 4.874045029785957e-07, "loss": 0.41, "step": 17347 }, { "epoch": 2.71, "grad_norm": 23.104014462954876, "learning_rate": 4.868845672690937e-07, "loss": 0.4574, "step": 17348 }, { "epoch": 2.71, "grad_norm": 25.669795151158045, "learning_rate": 4.863649021070261e-07, "loss": 0.4538, "step": 17349 }, { "epoch": 2.71, "grad_norm": 17.90745147532213, "learning_rate": 4.858455075071766e-07, "loss": 0.4072, "step": 17350 }, { "epoch": 2.71, "grad_norm": 18.577945008540926, "learning_rate": 4.853263834843136e-07, "loss": 0.3928, "step": 17351 }, { "epoch": 2.71, "grad_norm": 22.459120343549202, "learning_rate": 4.848075300532051e-07, "loss": 0.4506, "step": 17352 }, { "epoch": 2.71, "grad_norm": 16.599486700094594, "learning_rate": 4.842889472286039e-07, "loss": 0.4168, "step": 17353 }, { "epoch": 2.71, "grad_norm": 20.776638771370983, "learning_rate": 4.837706350252569e-07, "loss": 0.4259, "step": 17354 }, { "epoch": 2.71, "grad_norm": 15.100395334157533, "learning_rate": 4.832525934579058e-07, "loss": 0.399, "step": 17355 }, { "epoch": 2.71, "grad_norm": 33.67534898065224, "learning_rate": 4.827348225412864e-07, "loss": 0.4403, "step": 17356 }, { "epoch": 2.71, "grad_norm": 30.387288894532016, "learning_rate": 4.822173222901194e-07, "loss": 0.4431, "step": 17357 }, { "epoch": 2.71, "grad_norm": 15.940554654277092, "learning_rate": 4.81700092719124e-07, "loss": 0.4197, "step": 17358 }, { "epoch": 2.71, "grad_norm": 26.50262252296206, "learning_rate": 4.811831338430095e-07, "loss": 0.3843, "step": 17359 }, { "epoch": 2.71, "grad_norm": 23.92135082594988, "learning_rate": 4.806664456764787e-07, "loss": 0.4721, "step": 17360 }, { "epoch": 2.71, "grad_norm": 28.6315027965235, "learning_rate": 4.801500282342264e-07, "loss": 0.4465, "step": 17361 }, { "epoch": 2.71, "grad_norm": 27.82184890108278, "learning_rate": 4.796338815309387e-07, "loss": 0.4856, "step": 17362 }, { "epoch": 2.71, "grad_norm": 23.897452871933826, "learning_rate": 4.791180055812928e-07, "loss": 0.4594, "step": 17363 }, { "epoch": 2.71, "grad_norm": 24.95841888144139, "learning_rate": 4.786024003999612e-07, "loss": 0.4399, "step": 17364 }, { "epoch": 2.71, "grad_norm": 19.214807093508366, "learning_rate": 4.780870660016091e-07, "loss": 0.4338, "step": 17365 }, { "epoch": 2.71, "grad_norm": 24.18263758656307, "learning_rate": 4.77572002400889e-07, "loss": 0.4697, "step": 17366 }, { "epoch": 2.71, "grad_norm": 23.136460204437313, "learning_rate": 4.770572096124515e-07, "loss": 0.421, "step": 17367 }, { "epoch": 2.71, "grad_norm": 18.01549275061047, "learning_rate": 4.7654268765093604e-07, "loss": 0.4369, "step": 17368 }, { "epoch": 2.71, "grad_norm": 21.06620930056725, "learning_rate": 4.760284365309753e-07, "loss": 0.4101, "step": 17369 }, { "epoch": 2.71, "grad_norm": 29.334126203470845, "learning_rate": 4.755144562671943e-07, "loss": 0.4209, "step": 17370 }, { "epoch": 2.71, "grad_norm": 31.54646969006299, "learning_rate": 4.7500074687421017e-07, "loss": 0.4521, "step": 17371 }, { "epoch": 2.71, "grad_norm": 39.968513085421456, "learning_rate": 4.744873083666346e-07, "loss": 0.5315, "step": 17372 }, { "epoch": 2.71, "grad_norm": 24.247894269115044, "learning_rate": 4.739741407590659e-07, "loss": 0.4935, "step": 17373 }, { "epoch": 2.71, "grad_norm": 25.56898532853315, "learning_rate": 4.7346124406610125e-07, "loss": 0.4109, "step": 17374 }, { "epoch": 2.71, "grad_norm": 27.765093978651173, "learning_rate": 4.729486183023246e-07, "loss": 0.4593, "step": 17375 }, { "epoch": 2.71, "grad_norm": 18.762721227724175, "learning_rate": 4.724362634823165e-07, "loss": 0.4055, "step": 17376 }, { "epoch": 2.71, "grad_norm": 28.816990680308916, "learning_rate": 4.7192417962064865e-07, "loss": 0.439, "step": 17377 }, { "epoch": 2.71, "grad_norm": 18.274613292861563, "learning_rate": 4.714123667318837e-07, "loss": 0.4647, "step": 17378 }, { "epoch": 2.71, "grad_norm": 3.5723381890236676, "learning_rate": 4.7090082483057577e-07, "loss": 0.4093, "step": 17379 }, { "epoch": 2.71, "grad_norm": 22.15896494174935, "learning_rate": 4.7038955393127306e-07, "loss": 0.4018, "step": 17380 }, { "epoch": 2.71, "grad_norm": 13.399819966163685, "learning_rate": 4.6987855404851955e-07, "loss": 0.4273, "step": 17381 }, { "epoch": 2.72, "grad_norm": 19.085045005489427, "learning_rate": 4.6936782519684365e-07, "loss": 0.4396, "step": 17382 }, { "epoch": 2.72, "grad_norm": 20.70013423262256, "learning_rate": 4.6885736739077146e-07, "loss": 0.4026, "step": 17383 }, { "epoch": 2.72, "grad_norm": 30.19754895395075, "learning_rate": 4.683471806448192e-07, "loss": 0.4879, "step": 17384 }, { "epoch": 2.72, "grad_norm": 27.371877122015096, "learning_rate": 4.6783726497349747e-07, "loss": 0.5188, "step": 17385 }, { "epoch": 2.72, "grad_norm": 29.821238949350676, "learning_rate": 4.673276203913091e-07, "loss": 0.4776, "step": 17386 }, { "epoch": 2.72, "grad_norm": 23.906197476796812, "learning_rate": 4.668182469127469e-07, "loss": 0.4316, "step": 17387 }, { "epoch": 2.72, "grad_norm": 20.890558692781248, "learning_rate": 4.6630914455229493e-07, "loss": 0.4068, "step": 17388 }, { "epoch": 2.72, "grad_norm": 24.79089253736976, "learning_rate": 4.6580031332443487e-07, "loss": 0.4102, "step": 17389 }, { "epoch": 2.72, "grad_norm": 27.523150481244254, "learning_rate": 4.652917532436374e-07, "loss": 0.4516, "step": 17390 }, { "epoch": 2.72, "grad_norm": 15.306461668556436, "learning_rate": 4.6478346432436426e-07, "loss": 0.3934, "step": 17391 }, { "epoch": 2.72, "grad_norm": 21.29821462189279, "learning_rate": 4.642754465810717e-07, "loss": 0.4605, "step": 17392 }, { "epoch": 2.72, "grad_norm": 31.493702984748477, "learning_rate": 4.6376770002820593e-07, "loss": 0.4108, "step": 17393 }, { "epoch": 2.72, "grad_norm": 26.095407241458357, "learning_rate": 4.632602246802109e-07, "loss": 0.4612, "step": 17394 }, { "epoch": 2.72, "grad_norm": 16.646601453076475, "learning_rate": 4.6275302055151293e-07, "loss": 0.3693, "step": 17395 }, { "epoch": 2.72, "grad_norm": 18.10464786368163, "learning_rate": 4.6224608765654153e-07, "loss": 0.454, "step": 17396 }, { "epoch": 2.72, "grad_norm": 29.524280938816876, "learning_rate": 4.61739426009713e-07, "loss": 0.464, "step": 17397 }, { "epoch": 2.72, "grad_norm": 21.6426630919563, "learning_rate": 4.612330356254335e-07, "loss": 0.458, "step": 17398 }, { "epoch": 2.72, "grad_norm": 26.19173882235598, "learning_rate": 4.607269165181083e-07, "loss": 0.4576, "step": 17399 }, { "epoch": 2.72, "grad_norm": 22.28478750981967, "learning_rate": 4.6022106870212804e-07, "loss": 0.4246, "step": 17400 }, { "epoch": 2.72, "grad_norm": 28.18063883475103, "learning_rate": 4.597154921918812e-07, "loss": 0.4302, "step": 17401 }, { "epoch": 2.72, "grad_norm": 21.225910386650213, "learning_rate": 4.592101870017429e-07, "loss": 0.4514, "step": 17402 }, { "epoch": 2.72, "grad_norm": 17.98886841096316, "learning_rate": 4.587051531460873e-07, "loss": 0.4558, "step": 17403 }, { "epoch": 2.72, "grad_norm": 36.02204642984345, "learning_rate": 4.5820039063927514e-07, "loss": 0.4453, "step": 17404 }, { "epoch": 2.72, "grad_norm": 44.551800977650416, "learning_rate": 4.5769589949566153e-07, "loss": 0.4731, "step": 17405 }, { "epoch": 2.72, "grad_norm": 21.904202026760487, "learning_rate": 4.57191679729595e-07, "loss": 0.4218, "step": 17406 }, { "epoch": 2.72, "grad_norm": 29.949185903881208, "learning_rate": 4.566877313554152e-07, "loss": 0.4611, "step": 17407 }, { "epoch": 2.72, "grad_norm": 23.986283792343404, "learning_rate": 4.561840543874529e-07, "loss": 0.4625, "step": 17408 }, { "epoch": 2.72, "grad_norm": 18.171183132603076, "learning_rate": 4.5568064884003337e-07, "loss": 0.4917, "step": 17409 }, { "epoch": 2.72, "grad_norm": 21.441739249149823, "learning_rate": 4.551775147274717e-07, "loss": 0.4176, "step": 17410 }, { "epoch": 2.72, "grad_norm": 28.75684373970585, "learning_rate": 4.5467465206408103e-07, "loss": 0.4153, "step": 17411 }, { "epoch": 2.72, "grad_norm": 15.389540213874831, "learning_rate": 4.541720608641575e-07, "loss": 0.3785, "step": 17412 }, { "epoch": 2.72, "grad_norm": 16.016112998412776, "learning_rate": 4.5366974114199546e-07, "loss": 0.4106, "step": 17413 }, { "epoch": 2.72, "grad_norm": 18.931950525588608, "learning_rate": 4.5316769291188223e-07, "loss": 0.4221, "step": 17414 }, { "epoch": 2.72, "grad_norm": 29.240826921865555, "learning_rate": 4.526659161880964e-07, "loss": 0.434, "step": 17415 }, { "epoch": 2.72, "grad_norm": 31.13718517903119, "learning_rate": 4.521644109849066e-07, "loss": 0.4298, "step": 17416 }, { "epoch": 2.72, "grad_norm": 20.48098421548587, "learning_rate": 4.5166317731657363e-07, "loss": 0.4454, "step": 17417 }, { "epoch": 2.72, "grad_norm": 20.434729211114256, "learning_rate": 4.5116221519735493e-07, "loss": 0.4622, "step": 17418 }, { "epoch": 2.72, "grad_norm": 17.51366634119941, "learning_rate": 4.5066152464149914e-07, "loss": 0.444, "step": 17419 }, { "epoch": 2.72, "grad_norm": 16.712331832276472, "learning_rate": 4.5016110566324044e-07, "loss": 0.3636, "step": 17420 }, { "epoch": 2.72, "grad_norm": 19.2683599113328, "learning_rate": 4.4966095827681524e-07, "loss": 0.4288, "step": 17421 }, { "epoch": 2.72, "grad_norm": 14.892715665588616, "learning_rate": 4.491610824964454e-07, "loss": 0.4047, "step": 17422 }, { "epoch": 2.72, "grad_norm": 24.20558238929761, "learning_rate": 4.486614783363463e-07, "loss": 0.408, "step": 17423 }, { "epoch": 2.72, "grad_norm": 29.223760404762384, "learning_rate": 4.481621458107288e-07, "loss": 0.5298, "step": 17424 }, { "epoch": 2.72, "grad_norm": 29.68640664998534, "learning_rate": 4.476630849337904e-07, "loss": 0.4707, "step": 17425 }, { "epoch": 2.72, "grad_norm": 26.581314671462636, "learning_rate": 4.471642957197275e-07, "loss": 0.4379, "step": 17426 }, { "epoch": 2.72, "grad_norm": 20.671837457419493, "learning_rate": 4.4666577818272327e-07, "loss": 0.5048, "step": 17427 }, { "epoch": 2.72, "grad_norm": 28.38455317499184, "learning_rate": 4.461675323369563e-07, "loss": 0.4704, "step": 17428 }, { "epoch": 2.72, "grad_norm": 24.656114082300245, "learning_rate": 4.456695581965942e-07, "loss": 0.3824, "step": 17429 }, { "epoch": 2.72, "grad_norm": 18.952204471225592, "learning_rate": 4.4517185577580226e-07, "loss": 0.3957, "step": 17430 }, { "epoch": 2.72, "grad_norm": 24.677067259212578, "learning_rate": 4.446744250887336e-07, "loss": 0.4821, "step": 17431 }, { "epoch": 2.72, "grad_norm": 24.978744868491923, "learning_rate": 4.441772661495347e-07, "loss": 0.4703, "step": 17432 }, { "epoch": 2.72, "grad_norm": 19.496449254990207, "learning_rate": 4.4368037897234317e-07, "loss": 0.4496, "step": 17433 }, { "epoch": 2.72, "grad_norm": 17.877871280257406, "learning_rate": 4.4318376357129103e-07, "loss": 0.409, "step": 17434 }, { "epoch": 2.72, "grad_norm": 29.01844481682169, "learning_rate": 4.4268741996050244e-07, "loss": 0.5087, "step": 17435 }, { "epoch": 2.72, "grad_norm": 20.261647695163436, "learning_rate": 4.421913481540929e-07, "loss": 0.4911, "step": 17436 }, { "epoch": 2.72, "grad_norm": 18.478901469878732, "learning_rate": 4.4169554816617224e-07, "loss": 0.3723, "step": 17437 }, { "epoch": 2.72, "grad_norm": 21.70580459692987, "learning_rate": 4.4120002001083575e-07, "loss": 0.4246, "step": 17438 }, { "epoch": 2.72, "grad_norm": 22.589175481408834, "learning_rate": 4.4070476370218e-07, "loss": 0.4291, "step": 17439 }, { "epoch": 2.72, "grad_norm": 17.679987959219805, "learning_rate": 4.402097792542892e-07, "loss": 0.3547, "step": 17440 }, { "epoch": 2.72, "grad_norm": 15.396166383333188, "learning_rate": 4.397150666812411e-07, "loss": 0.4053, "step": 17441 }, { "epoch": 2.72, "grad_norm": 19.025061248631832, "learning_rate": 4.392206259971021e-07, "loss": 0.4068, "step": 17442 }, { "epoch": 2.72, "grad_norm": 19.413034781841418, "learning_rate": 4.3872645721593556e-07, "loss": 0.4552, "step": 17443 }, { "epoch": 2.72, "grad_norm": 30.036685707204484, "learning_rate": 4.382325603517956e-07, "loss": 0.4983, "step": 17444 }, { "epoch": 2.72, "grad_norm": 18.63665511456837, "learning_rate": 4.3773893541873005e-07, "loss": 0.4048, "step": 17445 }, { "epoch": 2.73, "grad_norm": 25.225865457520595, "learning_rate": 4.372455824307731e-07, "loss": 0.4295, "step": 17446 }, { "epoch": 2.73, "grad_norm": 24.630164364921853, "learning_rate": 4.36752501401958e-07, "loss": 0.4304, "step": 17447 }, { "epoch": 2.73, "grad_norm": 25.856865290300433, "learning_rate": 4.3625969234630694e-07, "loss": 0.4567, "step": 17448 }, { "epoch": 2.73, "grad_norm": 20.071757241203127, "learning_rate": 4.3576715527783755e-07, "loss": 0.3787, "step": 17449 }, { "epoch": 2.73, "grad_norm": 15.501434511668156, "learning_rate": 4.3527489021055414e-07, "loss": 0.3805, "step": 17450 }, { "epoch": 2.73, "grad_norm": 23.617238735253185, "learning_rate": 4.347828971584578e-07, "loss": 0.4372, "step": 17451 }, { "epoch": 2.73, "grad_norm": 18.29021742708971, "learning_rate": 4.342911761355395e-07, "loss": 0.3763, "step": 17452 }, { "epoch": 2.73, "grad_norm": 19.58187449331832, "learning_rate": 4.3379972715578587e-07, "loss": 0.4352, "step": 17453 }, { "epoch": 2.73, "grad_norm": 27.83295509223901, "learning_rate": 4.333085502331713e-07, "loss": 0.4433, "step": 17454 }, { "epoch": 2.73, "grad_norm": 20.9373031023361, "learning_rate": 4.328176453816657e-07, "loss": 0.4278, "step": 17455 }, { "epoch": 2.73, "grad_norm": 22.0500319547027, "learning_rate": 4.32327012615229e-07, "loss": 0.4205, "step": 17456 }, { "epoch": 2.73, "grad_norm": 18.327597823533026, "learning_rate": 4.3183665194781564e-07, "loss": 0.4277, "step": 17457 }, { "epoch": 2.73, "grad_norm": 32.74199777947312, "learning_rate": 4.3134656339337e-07, "loss": 0.4478, "step": 17458 }, { "epoch": 2.73, "grad_norm": 30.692695535246823, "learning_rate": 4.308567469658298e-07, "loss": 0.4943, "step": 17459 }, { "epoch": 2.73, "grad_norm": 23.9940120098935, "learning_rate": 4.3036720267912833e-07, "loss": 0.4214, "step": 17460 }, { "epoch": 2.73, "grad_norm": 18.34305745650129, "learning_rate": 4.2987793054718343e-07, "loss": 0.4019, "step": 17461 }, { "epoch": 2.73, "grad_norm": 18.68511982658362, "learning_rate": 4.2938893058391385e-07, "loss": 0.3996, "step": 17462 }, { "epoch": 2.73, "grad_norm": 24.150543381469717, "learning_rate": 4.289002028032219e-07, "loss": 0.4458, "step": 17463 }, { "epoch": 2.73, "grad_norm": 28.56440195766985, "learning_rate": 4.284117472190108e-07, "loss": 0.4547, "step": 17464 }, { "epoch": 2.73, "grad_norm": 20.686787437429523, "learning_rate": 4.2792356384517063e-07, "loss": 0.3967, "step": 17465 }, { "epoch": 2.73, "grad_norm": 22.997359866756085, "learning_rate": 4.2743565269558475e-07, "loss": 0.4226, "step": 17466 }, { "epoch": 2.73, "grad_norm": 18.35440114600762, "learning_rate": 4.2694801378412867e-07, "loss": 0.412, "step": 17467 }, { "epoch": 2.73, "grad_norm": 22.008676945964883, "learning_rate": 4.2646064712467016e-07, "loss": 0.485, "step": 17468 }, { "epoch": 2.73, "grad_norm": 16.297305301680044, "learning_rate": 4.2597355273107267e-07, "loss": 0.4124, "step": 17469 }, { "epoch": 2.73, "grad_norm": 23.817850589987756, "learning_rate": 4.2548673061718614e-07, "loss": 0.4001, "step": 17470 }, { "epoch": 2.73, "grad_norm": 15.62385161757375, "learning_rate": 4.25000180796854e-07, "loss": 0.4426, "step": 17471 }, { "epoch": 2.73, "grad_norm": 16.607778978955473, "learning_rate": 4.245139032839163e-07, "loss": 0.4312, "step": 17472 }, { "epoch": 2.73, "grad_norm": 26.70832361383802, "learning_rate": 4.240278980922019e-07, "loss": 0.4486, "step": 17473 }, { "epoch": 2.73, "grad_norm": 18.118358524866036, "learning_rate": 4.2354216523553314e-07, "loss": 0.4001, "step": 17474 }, { "epoch": 2.73, "grad_norm": 29.531044915915153, "learning_rate": 4.230567047277234e-07, "loss": 0.3881, "step": 17475 }, { "epoch": 2.73, "grad_norm": 22.110147238950233, "learning_rate": 4.2257151658257724e-07, "loss": 0.4405, "step": 17476 }, { "epoch": 2.73, "grad_norm": 27.63439163061701, "learning_rate": 4.2208660081389463e-07, "loss": 0.4667, "step": 17477 }, { "epoch": 2.73, "grad_norm": 25.894108281843828, "learning_rate": 4.2160195743546797e-07, "loss": 0.3913, "step": 17478 }, { "epoch": 2.73, "grad_norm": 24.97732492786066, "learning_rate": 4.2111758646107617e-07, "loss": 0.4572, "step": 17479 }, { "epoch": 2.73, "grad_norm": 14.659188849784474, "learning_rate": 4.2063348790449823e-07, "loss": 0.375, "step": 17480 }, { "epoch": 2.73, "grad_norm": 18.860107567251355, "learning_rate": 4.201496617794998e-07, "loss": 0.4538, "step": 17481 }, { "epoch": 2.73, "grad_norm": 18.708013347464217, "learning_rate": 4.19666108099841e-07, "loss": 0.4527, "step": 17482 }, { "epoch": 2.73, "grad_norm": 24.922707518062698, "learning_rate": 4.1918282687927304e-07, "loss": 0.4257, "step": 17483 }, { "epoch": 2.73, "grad_norm": 13.934500199316048, "learning_rate": 4.186998181315405e-07, "loss": 0.4286, "step": 17484 }, { "epoch": 2.73, "grad_norm": 22.812547378621478, "learning_rate": 4.1821708187038236e-07, "loss": 0.4758, "step": 17485 }, { "epoch": 2.73, "grad_norm": 25.638804827331093, "learning_rate": 4.177346181095232e-07, "loss": 0.42, "step": 17486 }, { "epoch": 2.73, "grad_norm": 20.398371528714584, "learning_rate": 4.172524268626876e-07, "loss": 0.4493, "step": 17487 }, { "epoch": 2.73, "grad_norm": 28.63009658880959, "learning_rate": 4.167705081435858e-07, "loss": 0.4378, "step": 17488 }, { "epoch": 2.73, "grad_norm": 27.343390325633028, "learning_rate": 4.162888619659244e-07, "loss": 0.4357, "step": 17489 }, { "epoch": 2.73, "grad_norm": 27.446397579123143, "learning_rate": 4.158074883434038e-07, "loss": 0.5091, "step": 17490 }, { "epoch": 2.73, "grad_norm": 18.738836955756252, "learning_rate": 4.1532638728971065e-07, "loss": 0.3896, "step": 17491 }, { "epoch": 2.73, "grad_norm": 12.552671287215457, "learning_rate": 4.148455588185274e-07, "loss": 0.3835, "step": 17492 }, { "epoch": 2.73, "grad_norm": 17.479494448313787, "learning_rate": 4.143650029435287e-07, "loss": 0.5104, "step": 17493 }, { "epoch": 2.73, "grad_norm": 20.20213456466183, "learning_rate": 4.138847196783835e-07, "loss": 0.3914, "step": 17494 }, { "epoch": 2.73, "grad_norm": 36.61430820435879, "learning_rate": 4.134047090367488e-07, "loss": 0.4465, "step": 17495 }, { "epoch": 2.73, "grad_norm": 32.111421649333316, "learning_rate": 4.129249710322758e-07, "loss": 0.4553, "step": 17496 }, { "epoch": 2.73, "grad_norm": 29.861711339807677, "learning_rate": 4.124455056786081e-07, "loss": 0.4718, "step": 17497 }, { "epoch": 2.73, "grad_norm": 21.267011851593708, "learning_rate": 4.119663129893814e-07, "loss": 0.4152, "step": 17498 }, { "epoch": 2.73, "grad_norm": 23.956345237191314, "learning_rate": 4.114873929782259e-07, "loss": 0.4485, "step": 17499 }, { "epoch": 2.73, "grad_norm": 15.606313338359346, "learning_rate": 4.1100874565875745e-07, "loss": 0.4006, "step": 17500 }, { "epoch": 2.73, "grad_norm": 15.98487525168327, "learning_rate": 4.105303710445918e-07, "loss": 0.4085, "step": 17501 }, { "epoch": 2.73, "grad_norm": 19.179010733572685, "learning_rate": 4.100522691493325e-07, "loss": 0.3811, "step": 17502 }, { "epoch": 2.73, "grad_norm": 34.54975317185062, "learning_rate": 4.0957443998657754e-07, "loss": 0.4092, "step": 17503 }, { "epoch": 2.73, "grad_norm": 18.810297646989035, "learning_rate": 4.0909688356991495e-07, "loss": 0.4638, "step": 17504 }, { "epoch": 2.73, "grad_norm": 17.72673533655055, "learning_rate": 4.086195999129261e-07, "loss": 0.3704, "step": 17505 }, { "epoch": 2.73, "grad_norm": 21.70174804765831, "learning_rate": 4.081425890291846e-07, "loss": 0.4318, "step": 17506 }, { "epoch": 2.73, "grad_norm": 16.040318933655207, "learning_rate": 4.076658509322573e-07, "loss": 0.3723, "step": 17507 }, { "epoch": 2.73, "grad_norm": 27.408368450702692, "learning_rate": 4.0718938563570233e-07, "loss": 0.4915, "step": 17508 }, { "epoch": 2.73, "grad_norm": 31.30847192503798, "learning_rate": 4.067131931530699e-07, "loss": 0.4231, "step": 17509 }, { "epoch": 2.74, "grad_norm": 39.510686933138025, "learning_rate": 4.0623727349790034e-07, "loss": 0.4788, "step": 17510 }, { "epoch": 2.74, "grad_norm": 19.3988233895468, "learning_rate": 4.0576162668373164e-07, "loss": 0.4353, "step": 17511 }, { "epoch": 2.74, "grad_norm": 18.015188683340405, "learning_rate": 4.0528625272409083e-07, "loss": 0.5014, "step": 17512 }, { "epoch": 2.74, "grad_norm": 27.904477455087267, "learning_rate": 4.0481115163249483e-07, "loss": 0.4241, "step": 17513 }, { "epoch": 2.74, "grad_norm": 20.779880772563885, "learning_rate": 4.0433632342245735e-07, "loss": 0.4098, "step": 17514 }, { "epoch": 2.74, "grad_norm": 16.437373304966645, "learning_rate": 4.038617681074808e-07, "loss": 0.3925, "step": 17515 }, { "epoch": 2.74, "grad_norm": 29.490459909934003, "learning_rate": 4.0338748570106333e-07, "loss": 0.4817, "step": 17516 }, { "epoch": 2.74, "grad_norm": 26.63319283121807, "learning_rate": 4.0291347621669084e-07, "loss": 0.4922, "step": 17517 }, { "epoch": 2.74, "grad_norm": 15.398392111616047, "learning_rate": 4.0243973966784477e-07, "loss": 0.3934, "step": 17518 }, { "epoch": 2.74, "grad_norm": 21.90392494787278, "learning_rate": 4.019662760679988e-07, "loss": 0.412, "step": 17519 }, { "epoch": 2.74, "grad_norm": 23.04840720045855, "learning_rate": 4.014930854306176e-07, "loss": 0.47, "step": 17520 }, { "epoch": 2.74, "grad_norm": 23.888840216364056, "learning_rate": 4.010201677691572e-07, "loss": 0.4637, "step": 17521 }, { "epoch": 2.74, "grad_norm": 40.307585703550416, "learning_rate": 4.005475230970679e-07, "loss": 0.4092, "step": 17522 }, { "epoch": 2.74, "grad_norm": 28.070339697621073, "learning_rate": 4.000751514277912e-07, "loss": 0.4502, "step": 17523 }, { "epoch": 2.74, "grad_norm": 25.372528754479347, "learning_rate": 3.9960305277476294e-07, "loss": 0.4244, "step": 17524 }, { "epoch": 2.74, "grad_norm": 22.54098039265196, "learning_rate": 3.9913122715140804e-07, "loss": 0.5074, "step": 17525 }, { "epoch": 2.74, "grad_norm": 42.24266727359193, "learning_rate": 3.986596745711424e-07, "loss": 0.53, "step": 17526 }, { "epoch": 2.74, "grad_norm": 23.467002523587144, "learning_rate": 3.9818839504737974e-07, "loss": 0.4493, "step": 17527 }, { "epoch": 2.74, "grad_norm": 23.461490860852557, "learning_rate": 3.977173885935237e-07, "loss": 0.3873, "step": 17528 }, { "epoch": 2.74, "grad_norm": 19.942109732590932, "learning_rate": 3.972466552229681e-07, "loss": 0.4736, "step": 17529 }, { "epoch": 2.74, "grad_norm": 16.932327548759254, "learning_rate": 3.967761949490978e-07, "loss": 0.4134, "step": 17530 }, { "epoch": 2.74, "grad_norm": 17.789274291083245, "learning_rate": 3.963060077852965e-07, "loss": 0.422, "step": 17531 }, { "epoch": 2.74, "grad_norm": 32.12679830575649, "learning_rate": 3.958360937449335e-07, "loss": 0.4283, "step": 17532 }, { "epoch": 2.74, "grad_norm": 16.195025399161796, "learning_rate": 3.953664528413737e-07, "loss": 0.4202, "step": 17533 }, { "epoch": 2.74, "grad_norm": 28.35008528858892, "learning_rate": 3.948970850879752e-07, "loss": 0.5014, "step": 17534 }, { "epoch": 2.74, "grad_norm": 22.336119845520958, "learning_rate": 3.94427990498083e-07, "loss": 0.3942, "step": 17535 }, { "epoch": 2.74, "grad_norm": 26.739048433966083, "learning_rate": 3.9395916908503863e-07, "loss": 0.4088, "step": 17536 }, { "epoch": 2.74, "grad_norm": 37.68300077363878, "learning_rate": 3.9349062086217915e-07, "loss": 0.4558, "step": 17537 }, { "epoch": 2.74, "grad_norm": 32.05418614695762, "learning_rate": 3.930223458428239e-07, "loss": 0.4647, "step": 17538 }, { "epoch": 2.74, "grad_norm": 18.312230954455963, "learning_rate": 3.9255434404029567e-07, "loss": 0.4275, "step": 17539 }, { "epoch": 2.74, "grad_norm": 22.36721689766602, "learning_rate": 3.9208661546789927e-07, "loss": 0.4145, "step": 17540 }, { "epoch": 2.74, "grad_norm": 18.389367887074666, "learning_rate": 3.9161916013894186e-07, "loss": 0.4328, "step": 17541 }, { "epoch": 2.74, "grad_norm": 32.79270567615219, "learning_rate": 3.9115197806671277e-07, "loss": 0.4913, "step": 17542 }, { "epoch": 2.74, "grad_norm": 21.452098591277245, "learning_rate": 3.9068506926450146e-07, "loss": 0.4522, "step": 17543 }, { "epoch": 2.74, "grad_norm": 22.44429547840225, "learning_rate": 3.9021843374558385e-07, "loss": 0.4402, "step": 17544 }, { "epoch": 2.74, "grad_norm": 23.951166938881396, "learning_rate": 3.897520715232339e-07, "loss": 0.4318, "step": 17545 }, { "epoch": 2.74, "grad_norm": 20.28365225407174, "learning_rate": 3.8928598261071313e-07, "loss": 0.3915, "step": 17546 }, { "epoch": 2.74, "grad_norm": 20.111141844508477, "learning_rate": 3.8882016702127544e-07, "loss": 0.4442, "step": 17547 }, { "epoch": 2.74, "grad_norm": 21.799639862152425, "learning_rate": 3.883546247681713e-07, "loss": 0.4386, "step": 17548 }, { "epoch": 2.74, "grad_norm": 20.933889535364973, "learning_rate": 3.8788935586463906e-07, "loss": 0.3988, "step": 17549 }, { "epoch": 2.74, "grad_norm": 18.307817436950778, "learning_rate": 3.8742436032391027e-07, "loss": 0.4015, "step": 17550 }, { "epoch": 2.74, "grad_norm": 18.48341164372044, "learning_rate": 3.8695963815920887e-07, "loss": 0.4712, "step": 17551 }, { "epoch": 2.74, "grad_norm": 21.99712163074532, "learning_rate": 3.864951893837521e-07, "loss": 0.4125, "step": 17552 }, { "epoch": 2.74, "grad_norm": 19.410124987344634, "learning_rate": 3.8603101401074927e-07, "loss": 0.4214, "step": 17553 }, { "epoch": 2.74, "grad_norm": 21.453021551439846, "learning_rate": 3.855671120533999e-07, "loss": 0.4249, "step": 17554 }, { "epoch": 2.74, "grad_norm": 17.90595372189574, "learning_rate": 3.851034835248979e-07, "loss": 0.4297, "step": 17555 }, { "epoch": 2.74, "grad_norm": 20.302956417423683, "learning_rate": 3.8464012843842714e-07, "loss": 0.4296, "step": 17556 }, { "epoch": 2.74, "grad_norm": 23.57739614835241, "learning_rate": 3.8417704680716704e-07, "loss": 0.4344, "step": 17557 }, { "epoch": 2.74, "grad_norm": 22.784546681606066, "learning_rate": 3.8371423864428826e-07, "loss": 0.4267, "step": 17558 }, { "epoch": 2.74, "grad_norm": 24.219278412253438, "learning_rate": 3.8325170396294907e-07, "loss": 0.4416, "step": 17559 }, { "epoch": 2.74, "grad_norm": 22.641724036574267, "learning_rate": 3.827894427763057e-07, "loss": 0.4565, "step": 17560 }, { "epoch": 2.74, "grad_norm": 28.717063972618014, "learning_rate": 3.8232745509750423e-07, "loss": 0.4422, "step": 17561 }, { "epoch": 2.74, "grad_norm": 28.222039680477796, "learning_rate": 3.818657409396853e-07, "loss": 0.4581, "step": 17562 }, { "epoch": 2.74, "grad_norm": 15.941673798151543, "learning_rate": 3.8140430031597844e-07, "loss": 0.4145, "step": 17563 }, { "epoch": 2.74, "grad_norm": 21.126597698868775, "learning_rate": 3.809431332395053e-07, "loss": 0.4306, "step": 17564 }, { "epoch": 2.74, "grad_norm": 16.492216500980064, "learning_rate": 3.804822397233832e-07, "loss": 0.3688, "step": 17565 }, { "epoch": 2.74, "grad_norm": 23.945840111002425, "learning_rate": 3.800216197807194e-07, "loss": 0.4492, "step": 17566 }, { "epoch": 2.74, "grad_norm": 17.556779334890756, "learning_rate": 3.7956127342461127e-07, "loss": 0.4067, "step": 17567 }, { "epoch": 2.74, "grad_norm": 19.31972029118095, "learning_rate": 3.791012006681549e-07, "loss": 0.4254, "step": 17568 }, { "epoch": 2.74, "grad_norm": 27.829320157970898, "learning_rate": 3.786414015244311e-07, "loss": 0.4306, "step": 17569 }, { "epoch": 2.74, "grad_norm": 27.47234722090023, "learning_rate": 3.781818760065181e-07, "loss": 0.4251, "step": 17570 }, { "epoch": 2.74, "grad_norm": 18.16782867231646, "learning_rate": 3.777226241274834e-07, "loss": 0.4364, "step": 17571 }, { "epoch": 2.74, "grad_norm": 22.465201876310815, "learning_rate": 3.772636459003887e-07, "loss": 0.434, "step": 17572 }, { "epoch": 2.74, "grad_norm": 16.965029729306384, "learning_rate": 3.768049413382868e-07, "loss": 0.3755, "step": 17573 }, { "epoch": 2.75, "grad_norm": 18.79009514268903, "learning_rate": 3.76346510454223e-07, "loss": 0.4047, "step": 17574 }, { "epoch": 2.75, "grad_norm": 35.86451489298629, "learning_rate": 3.758883532612356e-07, "loss": 0.4105, "step": 17575 }, { "epoch": 2.75, "grad_norm": 30.348565907862902, "learning_rate": 3.754304697723521e-07, "loss": 0.4163, "step": 17576 }, { "epoch": 2.75, "grad_norm": 27.512065754730056, "learning_rate": 3.749728600005953e-07, "loss": 0.3804, "step": 17577 }, { "epoch": 2.75, "grad_norm": 24.34524124876427, "learning_rate": 3.745155239589815e-07, "loss": 0.3923, "step": 17578 }, { "epoch": 2.75, "grad_norm": 22.767678273590956, "learning_rate": 3.7405846166051585e-07, "loss": 0.4289, "step": 17579 }, { "epoch": 2.75, "grad_norm": 23.07189445613623, "learning_rate": 3.736016731181946e-07, "loss": 0.4136, "step": 17580 }, { "epoch": 2.75, "grad_norm": 21.27953030308091, "learning_rate": 3.7314515834501075e-07, "loss": 0.4527, "step": 17581 }, { "epoch": 2.75, "grad_norm": 21.27630390969021, "learning_rate": 3.726889173539483e-07, "loss": 0.4432, "step": 17582 }, { "epoch": 2.75, "grad_norm": 29.0326694145637, "learning_rate": 3.7223295015798133e-07, "loss": 0.5179, "step": 17583 }, { "epoch": 2.75, "grad_norm": 16.664408589420702, "learning_rate": 3.7177725677007615e-07, "loss": 0.3659, "step": 17584 }, { "epoch": 2.75, "grad_norm": 25.147560411856237, "learning_rate": 3.713218372031935e-07, "loss": 0.5466, "step": 17585 }, { "epoch": 2.75, "grad_norm": 16.922662537900592, "learning_rate": 3.708666914702852e-07, "loss": 0.4031, "step": 17586 }, { "epoch": 2.75, "grad_norm": 23.310389091427545, "learning_rate": 3.704118195842965e-07, "loss": 0.4602, "step": 17587 }, { "epoch": 2.75, "grad_norm": 18.95408639254217, "learning_rate": 3.699572215581615e-07, "loss": 0.4711, "step": 17588 }, { "epoch": 2.75, "grad_norm": 29.79057878706563, "learning_rate": 3.695028974048098e-07, "loss": 0.5159, "step": 17589 }, { "epoch": 2.75, "grad_norm": 21.45883003666137, "learning_rate": 3.6904884713716114e-07, "loss": 0.3995, "step": 17590 }, { "epoch": 2.75, "grad_norm": 29.49774046908421, "learning_rate": 3.6859507076813073e-07, "loss": 0.4374, "step": 17591 }, { "epoch": 2.75, "grad_norm": 38.878012121898465, "learning_rate": 3.6814156831062264e-07, "loss": 0.4094, "step": 17592 }, { "epoch": 2.75, "grad_norm": 21.39699222270177, "learning_rate": 3.6768833977753214e-07, "loss": 0.4804, "step": 17593 }, { "epoch": 2.75, "grad_norm": 14.367121750254363, "learning_rate": 3.672353851817512e-07, "loss": 0.4046, "step": 17594 }, { "epoch": 2.75, "grad_norm": 23.538314317472818, "learning_rate": 3.667827045361616e-07, "loss": 0.512, "step": 17595 }, { "epoch": 2.75, "grad_norm": 28.10235191596009, "learning_rate": 3.663302978536354e-07, "loss": 0.3986, "step": 17596 }, { "epoch": 2.75, "grad_norm": 27.12716722345551, "learning_rate": 3.658781651470422e-07, "loss": 0.4979, "step": 17597 }, { "epoch": 2.75, "grad_norm": 20.902512735201658, "learning_rate": 3.654263064292363e-07, "loss": 0.4153, "step": 17598 }, { "epoch": 2.75, "grad_norm": 15.508206423505467, "learning_rate": 3.649747217130695e-07, "loss": 0.4354, "step": 17599 }, { "epoch": 2.75, "grad_norm": 21.13389008572231, "learning_rate": 3.645234110113871e-07, "loss": 0.3795, "step": 17600 }, { "epoch": 2.75, "grad_norm": 22.074589900520056, "learning_rate": 3.6407237433702116e-07, "loss": 0.4382, "step": 17601 }, { "epoch": 2.75, "grad_norm": 31.822803042074202, "learning_rate": 3.636216117028013e-07, "loss": 0.4385, "step": 17602 }, { "epoch": 2.75, "grad_norm": 16.846443769012538, "learning_rate": 3.63171123121544e-07, "loss": 0.3833, "step": 17603 }, { "epoch": 2.75, "grad_norm": 24.389533195032417, "learning_rate": 3.627209086060635e-07, "loss": 0.3963, "step": 17604 }, { "epoch": 2.75, "grad_norm": 20.93648974130658, "learning_rate": 3.6227096816916274e-07, "loss": 0.4219, "step": 17605 }, { "epoch": 2.75, "grad_norm": 22.453932798052776, "learning_rate": 3.6182130182363716e-07, "loss": 0.449, "step": 17606 }, { "epoch": 2.75, "grad_norm": 16.16350660663446, "learning_rate": 3.6137190958227655e-07, "loss": 0.3822, "step": 17607 }, { "epoch": 2.75, "grad_norm": 24.11960044964342, "learning_rate": 3.609227914578606e-07, "loss": 0.4434, "step": 17608 }, { "epoch": 2.75, "grad_norm": 35.27148795923943, "learning_rate": 3.6047394746316023e-07, "loss": 0.4816, "step": 17609 }, { "epoch": 2.75, "grad_norm": 28.587634090214188, "learning_rate": 3.6002537761094193e-07, "loss": 0.4546, "step": 17610 }, { "epoch": 2.75, "grad_norm": 29.720156458838176, "learning_rate": 3.595770819139632e-07, "loss": 0.4599, "step": 17611 }, { "epoch": 2.75, "grad_norm": 24.607773131052568, "learning_rate": 3.5912906038497287e-07, "loss": 0.5316, "step": 17612 }, { "epoch": 2.75, "grad_norm": 20.001750958060477, "learning_rate": 3.5868131303671393e-07, "loss": 0.4304, "step": 17613 }, { "epoch": 2.75, "grad_norm": 21.677063960201608, "learning_rate": 3.582338398819163e-07, "loss": 0.3819, "step": 17614 }, { "epoch": 2.75, "grad_norm": 21.54366094423086, "learning_rate": 3.577866409333075e-07, "loss": 0.3947, "step": 17615 }, { "epoch": 2.75, "grad_norm": 22.872411805571907, "learning_rate": 3.5733971620360853e-07, "loss": 0.4644, "step": 17616 }, { "epoch": 2.75, "grad_norm": 27.67797548079895, "learning_rate": 3.5689306570552584e-07, "loss": 0.4206, "step": 17617 }, { "epoch": 2.75, "grad_norm": 15.201829078776536, "learning_rate": 3.5644668945176264e-07, "loss": 0.4066, "step": 17618 }, { "epoch": 2.75, "grad_norm": 25.251090222725587, "learning_rate": 3.5600058745501543e-07, "loss": 0.4283, "step": 17619 }, { "epoch": 2.75, "grad_norm": 26.384651325590177, "learning_rate": 3.555547597279696e-07, "loss": 0.4019, "step": 17620 }, { "epoch": 2.75, "grad_norm": 24.472285705221083, "learning_rate": 3.551092062833039e-07, "loss": 0.4387, "step": 17621 }, { "epoch": 2.75, "grad_norm": 22.641588680221716, "learning_rate": 3.546639271336916e-07, "loss": 0.4783, "step": 17622 }, { "epoch": 2.75, "grad_norm": 19.627109028541238, "learning_rate": 3.5421892229179354e-07, "loss": 0.4309, "step": 17623 }, { "epoch": 2.75, "grad_norm": 23.303779076656976, "learning_rate": 3.537741917702664e-07, "loss": 0.3577, "step": 17624 }, { "epoch": 2.75, "grad_norm": 16.95909705822597, "learning_rate": 3.5332973558175996e-07, "loss": 0.4548, "step": 17625 }, { "epoch": 2.75, "grad_norm": 36.23098984089105, "learning_rate": 3.52885553738912e-07, "loss": 0.4801, "step": 17626 }, { "epoch": 2.75, "grad_norm": 17.545105013485518, "learning_rate": 3.524416462543556e-07, "loss": 0.3947, "step": 17627 }, { "epoch": 2.75, "grad_norm": 17.51014834664943, "learning_rate": 3.519980131407152e-07, "loss": 0.4006, "step": 17628 }, { "epoch": 2.75, "grad_norm": 41.87052157353007, "learning_rate": 3.515546544106074e-07, "loss": 0.3838, "step": 17629 }, { "epoch": 2.75, "grad_norm": 16.070821691892426, "learning_rate": 3.5111157007663986e-07, "loss": 0.3933, "step": 17630 }, { "epoch": 2.75, "grad_norm": 23.179131830676027, "learning_rate": 3.506687601514158e-07, "loss": 0.3897, "step": 17631 }, { "epoch": 2.75, "grad_norm": 22.031478653390828, "learning_rate": 3.502262246475285e-07, "loss": 0.4245, "step": 17632 }, { "epoch": 2.75, "grad_norm": 20.278895357513118, "learning_rate": 3.497839635775613e-07, "loss": 0.3874, "step": 17633 }, { "epoch": 2.75, "grad_norm": 24.700946334179584, "learning_rate": 3.493419769540929e-07, "loss": 0.4042, "step": 17634 }, { "epoch": 2.75, "grad_norm": 26.533459846565975, "learning_rate": 3.489002647896933e-07, "loss": 0.3964, "step": 17635 }, { "epoch": 2.75, "grad_norm": 20.071327751024754, "learning_rate": 3.4845882709692356e-07, "loss": 0.4315, "step": 17636 }, { "epoch": 2.75, "grad_norm": 18.20121481290997, "learning_rate": 3.4801766388834037e-07, "loss": 0.424, "step": 17637 }, { "epoch": 2.76, "grad_norm": 21.653808124428775, "learning_rate": 3.4757677517648916e-07, "loss": 0.438, "step": 17638 }, { "epoch": 2.76, "grad_norm": 17.261791980119334, "learning_rate": 3.471361609739055e-07, "loss": 0.4069, "step": 17639 }, { "epoch": 2.76, "grad_norm": 13.676234698956899, "learning_rate": 3.4669582129312373e-07, "loss": 0.4399, "step": 17640 }, { "epoch": 2.76, "grad_norm": 18.76643994187275, "learning_rate": 3.462557561466662e-07, "loss": 0.3931, "step": 17641 }, { "epoch": 2.76, "grad_norm": 39.28658178867971, "learning_rate": 3.458159655470483e-07, "loss": 0.4604, "step": 17642 }, { "epoch": 2.76, "grad_norm": 17.26384231158706, "learning_rate": 3.4537644950677576e-07, "loss": 0.4436, "step": 17643 }, { "epoch": 2.76, "grad_norm": 20.873303179177018, "learning_rate": 3.449372080383484e-07, "loss": 0.4575, "step": 17644 }, { "epoch": 2.76, "grad_norm": 21.13398775509945, "learning_rate": 3.4449824115426076e-07, "loss": 0.4997, "step": 17645 }, { "epoch": 2.76, "grad_norm": 18.411864926831946, "learning_rate": 3.440595488669951e-07, "loss": 0.3692, "step": 17646 }, { "epoch": 2.76, "grad_norm": 22.910887637094188, "learning_rate": 3.4362113118902475e-07, "loss": 0.4282, "step": 17647 }, { "epoch": 2.76, "grad_norm": 18.869032780226853, "learning_rate": 3.4318298813282193e-07, "loss": 0.4278, "step": 17648 }, { "epoch": 2.76, "grad_norm": 28.135475465928945, "learning_rate": 3.4274511971084556e-07, "loss": 0.4167, "step": 17649 }, { "epoch": 2.76, "grad_norm": 20.30215450484469, "learning_rate": 3.423075259355491e-07, "loss": 0.3607, "step": 17650 }, { "epoch": 2.76, "grad_norm": 40.711977948129025, "learning_rate": 3.41870206819378e-07, "loss": 0.5404, "step": 17651 }, { "epoch": 2.76, "grad_norm": 18.00839845113648, "learning_rate": 3.4143316237476686e-07, "loss": 0.3767, "step": 17652 }, { "epoch": 2.76, "grad_norm": 19.35476609541522, "learning_rate": 3.409963926141457e-07, "loss": 0.4266, "step": 17653 }, { "epoch": 2.76, "grad_norm": 19.065449786604272, "learning_rate": 3.4055989754993913e-07, "loss": 0.385, "step": 17654 }, { "epoch": 2.76, "grad_norm": 20.312340432053816, "learning_rate": 3.4012367719455595e-07, "loss": 0.4318, "step": 17655 }, { "epoch": 2.76, "grad_norm": 26.78982832485094, "learning_rate": 3.3968773156040745e-07, "loss": 0.4785, "step": 17656 }, { "epoch": 2.76, "grad_norm": 24.784937148331927, "learning_rate": 3.392520606598859e-07, "loss": 0.403, "step": 17657 }, { "epoch": 2.76, "grad_norm": 27.538531470623646, "learning_rate": 3.3881666450538695e-07, "loss": 0.4635, "step": 17658 }, { "epoch": 2.76, "grad_norm": 23.1561216262008, "learning_rate": 3.383815431092885e-07, "loss": 0.3897, "step": 17659 }, { "epoch": 2.76, "grad_norm": 14.400811593688012, "learning_rate": 3.3794669648396727e-07, "loss": 0.387, "step": 17660 }, { "epoch": 2.76, "grad_norm": 21.950143112675182, "learning_rate": 3.3751212464179007e-07, "loss": 0.5017, "step": 17661 }, { "epoch": 2.76, "grad_norm": 16.941965160337542, "learning_rate": 3.370778275951159e-07, "loss": 0.4648, "step": 17662 }, { "epoch": 2.76, "grad_norm": 27.33928110279793, "learning_rate": 3.3664380535629595e-07, "loss": 0.4601, "step": 17663 }, { "epoch": 2.76, "grad_norm": 20.795253560765687, "learning_rate": 3.362100579376726e-07, "loss": 0.4437, "step": 17664 }, { "epoch": 2.76, "grad_norm": 37.16993754871001, "learning_rate": 3.357765853515815e-07, "loss": 0.4154, "step": 17665 }, { "epoch": 2.76, "grad_norm": 21.075425902714972, "learning_rate": 3.353433876103529e-07, "loss": 0.4378, "step": 17666 }, { "epoch": 2.76, "grad_norm": 16.344221736489434, "learning_rate": 3.349104647263046e-07, "loss": 0.4249, "step": 17667 }, { "epoch": 2.76, "grad_norm": 22.077663558881213, "learning_rate": 3.3447781671174684e-07, "loss": 0.3781, "step": 17668 }, { "epoch": 2.76, "grad_norm": 22.78743554873619, "learning_rate": 3.340454435789864e-07, "loss": 0.4619, "step": 17669 }, { "epoch": 2.76, "grad_norm": 17.302493432926585, "learning_rate": 3.336133453403201e-07, "loss": 0.4299, "step": 17670 }, { "epoch": 2.76, "grad_norm": 21.881522257396004, "learning_rate": 3.331815220080359e-07, "loss": 0.4149, "step": 17671 }, { "epoch": 2.76, "grad_norm": 20.621173031454553, "learning_rate": 3.3274997359441174e-07, "loss": 0.4128, "step": 17672 }, { "epoch": 2.76, "grad_norm": 21.368764571547246, "learning_rate": 3.323187001117245e-07, "loss": 0.4402, "step": 17673 }, { "epoch": 2.76, "grad_norm": 27.149741829475385, "learning_rate": 3.318877015722377e-07, "loss": 0.5089, "step": 17674 }, { "epoch": 2.76, "grad_norm": 18.583278227643845, "learning_rate": 3.3145697798821044e-07, "loss": 0.4237, "step": 17675 }, { "epoch": 2.76, "grad_norm": 42.922950104870736, "learning_rate": 3.3102652937189064e-07, "loss": 0.5172, "step": 17676 }, { "epoch": 2.76, "grad_norm": 26.244669910682696, "learning_rate": 3.3059635573551854e-07, "loss": 0.4231, "step": 17677 }, { "epoch": 2.76, "grad_norm": 26.483313130159356, "learning_rate": 3.301664570913299e-07, "loss": 0.4778, "step": 17678 }, { "epoch": 2.76, "grad_norm": 28.538593450242786, "learning_rate": 3.297368334515516e-07, "loss": 0.4695, "step": 17679 }, { "epoch": 2.76, "grad_norm": 38.310937172454736, "learning_rate": 3.2930748482840056e-07, "loss": 0.4796, "step": 17680 }, { "epoch": 2.76, "grad_norm": 15.126211821269354, "learning_rate": 3.2887841123408817e-07, "loss": 0.3916, "step": 17681 }, { "epoch": 2.76, "grad_norm": 25.02724307084497, "learning_rate": 3.2844961268081457e-07, "loss": 0.4375, "step": 17682 }, { "epoch": 2.76, "grad_norm": 14.57181444122784, "learning_rate": 3.2802108918077777e-07, "loss": 0.3595, "step": 17683 }, { "epoch": 2.76, "grad_norm": 32.14063490115, "learning_rate": 3.275928407461615e-07, "loss": 0.4169, "step": 17684 }, { "epoch": 2.76, "grad_norm": 23.790765076268357, "learning_rate": 3.271648673891481e-07, "loss": 0.4441, "step": 17685 }, { "epoch": 2.76, "grad_norm": 24.983642147497704, "learning_rate": 3.267371691219068e-07, "loss": 0.4119, "step": 17686 }, { "epoch": 2.76, "grad_norm": 30.6282771940744, "learning_rate": 3.2630974595660226e-07, "loss": 0.4309, "step": 17687 }, { "epoch": 2.76, "grad_norm": 19.755118661930588, "learning_rate": 3.2588259790538923e-07, "loss": 0.4263, "step": 17688 }, { "epoch": 2.76, "grad_norm": 28.05257720361429, "learning_rate": 3.2545572498041576e-07, "loss": 0.4567, "step": 17689 }, { "epoch": 2.76, "grad_norm": 24.982562950701627, "learning_rate": 3.2502912719382327e-07, "loss": 0.4216, "step": 17690 }, { "epoch": 2.76, "grad_norm": 17.59691952127479, "learning_rate": 3.2460280455774096e-07, "loss": 0.3996, "step": 17691 }, { "epoch": 2.76, "grad_norm": 20.05470851463165, "learning_rate": 3.2417675708429574e-07, "loss": 0.4324, "step": 17692 }, { "epoch": 2.76, "grad_norm": 22.1273676903876, "learning_rate": 3.237509847856035e-07, "loss": 0.4311, "step": 17693 }, { "epoch": 2.76, "grad_norm": 23.049775757003967, "learning_rate": 3.2332548767377237e-07, "loss": 0.3871, "step": 17694 }, { "epoch": 2.76, "grad_norm": 37.63828354627713, "learning_rate": 3.229002657609037e-07, "loss": 0.403, "step": 17695 }, { "epoch": 2.76, "grad_norm": 19.827930429193863, "learning_rate": 3.224753190590912e-07, "loss": 0.3944, "step": 17696 }, { "epoch": 2.76, "grad_norm": 21.935924212600288, "learning_rate": 3.220506475804186e-07, "loss": 0.3809, "step": 17697 }, { "epoch": 2.76, "grad_norm": 45.98717967743818, "learning_rate": 3.21626251336965e-07, "loss": 0.4497, "step": 17698 }, { "epoch": 2.76, "grad_norm": 21.917288850400528, "learning_rate": 3.212021303407975e-07, "loss": 0.4398, "step": 17699 }, { "epoch": 2.76, "grad_norm": 27.201496306671956, "learning_rate": 3.20778284603982e-07, "loss": 0.5111, "step": 17700 }, { "epoch": 2.76, "grad_norm": 19.294124470951655, "learning_rate": 3.203547141385688e-07, "loss": 0.4211, "step": 17701 }, { "epoch": 2.77, "grad_norm": 25.9780568712668, "learning_rate": 3.1993141895660385e-07, "loss": 0.4014, "step": 17702 }, { "epoch": 2.77, "grad_norm": 14.963547719935852, "learning_rate": 3.1950839907012755e-07, "loss": 0.3566, "step": 17703 }, { "epoch": 2.77, "grad_norm": 21.675189528343374, "learning_rate": 3.1908565449117026e-07, "loss": 0.3755, "step": 17704 }, { "epoch": 2.77, "grad_norm": 27.725663897973195, "learning_rate": 3.1866318523175344e-07, "loss": 0.4299, "step": 17705 }, { "epoch": 2.77, "grad_norm": 33.222484769384224, "learning_rate": 3.18240991303892e-07, "loss": 0.5161, "step": 17706 }, { "epoch": 2.77, "grad_norm": 31.508710530122755, "learning_rate": 3.178190727195918e-07, "loss": 0.3962, "step": 17707 }, { "epoch": 2.77, "grad_norm": 23.482210403450914, "learning_rate": 3.173974294908555e-07, "loss": 0.4355, "step": 17708 }, { "epoch": 2.77, "grad_norm": 21.52061158129383, "learning_rate": 3.169760616296702e-07, "loss": 0.4382, "step": 17709 }, { "epoch": 2.77, "grad_norm": 24.220009148382967, "learning_rate": 3.1655496914802296e-07, "loss": 0.4598, "step": 17710 }, { "epoch": 2.77, "grad_norm": 27.42700087521264, "learning_rate": 3.1613415205788643e-07, "loss": 0.5331, "step": 17711 }, { "epoch": 2.77, "grad_norm": 37.06494454046371, "learning_rate": 3.1571361037122996e-07, "loss": 0.4245, "step": 17712 }, { "epoch": 2.77, "grad_norm": 19.8939808698795, "learning_rate": 3.152933441000139e-07, "loss": 0.4783, "step": 17713 }, { "epoch": 2.77, "grad_norm": 24.52823995905179, "learning_rate": 3.148733532561887e-07, "loss": 0.4167, "step": 17714 }, { "epoch": 2.77, "grad_norm": 29.28182344170294, "learning_rate": 3.1445363785170155e-07, "loss": 0.4548, "step": 17715 }, { "epoch": 2.77, "grad_norm": 40.379343304478844, "learning_rate": 3.140341978984851e-07, "loss": 0.5007, "step": 17716 }, { "epoch": 2.77, "grad_norm": 25.06055781813353, "learning_rate": 3.13615033408472e-07, "loss": 0.4108, "step": 17717 }, { "epoch": 2.77, "grad_norm": 17.125701212101035, "learning_rate": 3.1319614439357827e-07, "loss": 0.3849, "step": 17718 }, { "epoch": 2.77, "grad_norm": 22.820831586764452, "learning_rate": 3.1277753086572105e-07, "loss": 0.4135, "step": 17719 }, { "epoch": 2.77, "grad_norm": 22.55893056345587, "learning_rate": 3.123591928368042e-07, "loss": 0.4514, "step": 17720 }, { "epoch": 2.77, "grad_norm": 24.306792548790323, "learning_rate": 3.1194113031872584e-07, "loss": 0.4446, "step": 17721 }, { "epoch": 2.77, "grad_norm": 26.925897571174666, "learning_rate": 3.115233433233722e-07, "loss": 0.4256, "step": 17722 }, { "epoch": 2.77, "grad_norm": 14.986215810505279, "learning_rate": 3.111058318626281e-07, "loss": 0.3723, "step": 17723 }, { "epoch": 2.77, "grad_norm": 25.33960368579462, "learning_rate": 3.106885959483652e-07, "loss": 0.3996, "step": 17724 }, { "epoch": 2.77, "grad_norm": 22.430890990748615, "learning_rate": 3.102716355924518e-07, "loss": 0.4419, "step": 17725 }, { "epoch": 2.77, "grad_norm": 34.25280472790416, "learning_rate": 3.09854950806745e-07, "loss": 0.4201, "step": 17726 }, { "epoch": 2.77, "grad_norm": 23.198934483621148, "learning_rate": 3.094385416030943e-07, "loss": 0.4546, "step": 17727 }, { "epoch": 2.77, "grad_norm": 16.194276718604623, "learning_rate": 3.0902240799334237e-07, "loss": 0.4304, "step": 17728 }, { "epoch": 2.77, "grad_norm": 27.617083893275396, "learning_rate": 3.0860654998932536e-07, "loss": 0.4229, "step": 17729 }, { "epoch": 2.77, "grad_norm": 26.292763872018554, "learning_rate": 3.0819096760286826e-07, "loss": 0.4271, "step": 17730 }, { "epoch": 2.77, "grad_norm": 16.192417339603097, "learning_rate": 3.0777566084578934e-07, "loss": 0.3836, "step": 17731 }, { "epoch": 2.77, "grad_norm": 27.763047490997057, "learning_rate": 3.0736062972990145e-07, "loss": 0.4651, "step": 17732 }, { "epoch": 2.77, "grad_norm": 26.461177569174026, "learning_rate": 3.069458742670073e-07, "loss": 0.4566, "step": 17733 }, { "epoch": 2.77, "grad_norm": 26.637765681173875, "learning_rate": 3.065313944689019e-07, "loss": 0.4697, "step": 17734 }, { "epoch": 2.77, "grad_norm": 16.065984463969304, "learning_rate": 3.0611719034737363e-07, "loss": 0.3595, "step": 17735 }, { "epoch": 2.77, "grad_norm": 20.578260994811572, "learning_rate": 3.05703261914202e-07, "loss": 0.4632, "step": 17736 }, { "epoch": 2.77, "grad_norm": 22.242904327948846, "learning_rate": 3.0528960918115744e-07, "loss": 0.4301, "step": 17737 }, { "epoch": 2.77, "grad_norm": 29.37043488924438, "learning_rate": 3.048762321600063e-07, "loss": 0.4664, "step": 17738 }, { "epoch": 2.77, "grad_norm": 20.084968414431373, "learning_rate": 3.0446313086250346e-07, "loss": 0.4456, "step": 17739 }, { "epoch": 2.77, "grad_norm": 31.489069310879728, "learning_rate": 3.0405030530039734e-07, "loss": 0.4943, "step": 17740 }, { "epoch": 2.77, "grad_norm": 26.18330621369408, "learning_rate": 3.0363775548542753e-07, "loss": 0.5134, "step": 17741 }, { "epoch": 2.77, "grad_norm": 21.600441304473748, "learning_rate": 3.0322548142932893e-07, "loss": 0.4052, "step": 17742 }, { "epoch": 2.77, "grad_norm": 21.945506302469074, "learning_rate": 3.0281348314382453e-07, "loss": 0.4587, "step": 17743 }, { "epoch": 2.77, "grad_norm": 21.23798674940019, "learning_rate": 3.0240176064063266e-07, "loss": 0.4016, "step": 17744 }, { "epoch": 2.77, "grad_norm": 23.22841361426061, "learning_rate": 3.0199031393146174e-07, "loss": 0.5186, "step": 17745 }, { "epoch": 2.77, "grad_norm": 21.262510650425778, "learning_rate": 3.015791430280135e-07, "loss": 0.4451, "step": 17746 }, { "epoch": 2.77, "grad_norm": 34.379464918082775, "learning_rate": 3.0116824794197977e-07, "loss": 0.4826, "step": 17747 }, { "epoch": 2.77, "grad_norm": 21.097081391445617, "learning_rate": 3.0075762868504777e-07, "loss": 0.438, "step": 17748 }, { "epoch": 2.77, "grad_norm": 22.355576222488185, "learning_rate": 3.003472852688949e-07, "loss": 0.4465, "step": 17749 }, { "epoch": 2.77, "grad_norm": 17.42190627407466, "learning_rate": 2.999372177051918e-07, "loss": 0.3745, "step": 17750 }, { "epoch": 2.77, "grad_norm": 16.05724652493624, "learning_rate": 2.9952742600559917e-07, "loss": 0.4138, "step": 17751 }, { "epoch": 2.77, "grad_norm": 24.001924959113907, "learning_rate": 2.9911791018177204e-07, "loss": 0.4154, "step": 17752 }, { "epoch": 2.77, "grad_norm": 16.837550494465944, "learning_rate": 2.9870867024535675e-07, "loss": 0.4127, "step": 17753 }, { "epoch": 2.77, "grad_norm": 20.721680934436208, "learning_rate": 2.9829970620799176e-07, "loss": 0.448, "step": 17754 }, { "epoch": 2.77, "grad_norm": 21.171433216848833, "learning_rate": 2.978910180813088e-07, "loss": 0.3914, "step": 17755 }, { "epoch": 2.77, "grad_norm": 26.16145196297292, "learning_rate": 2.9748260587692867e-07, "loss": 0.4608, "step": 17756 }, { "epoch": 2.77, "grad_norm": 32.4257215006004, "learning_rate": 2.9707446960646644e-07, "loss": 0.4329, "step": 17757 }, { "epoch": 2.77, "grad_norm": 25.96223306307443, "learning_rate": 2.966666092815318e-07, "loss": 0.4311, "step": 17758 }, { "epoch": 2.77, "grad_norm": 19.596092187560284, "learning_rate": 2.962590249137232e-07, "loss": 0.4524, "step": 17759 }, { "epoch": 2.77, "grad_norm": 15.09075659731838, "learning_rate": 2.958517165146302e-07, "loss": 0.3782, "step": 17760 }, { "epoch": 2.77, "grad_norm": 32.97059382439807, "learning_rate": 2.954446840958369e-07, "loss": 0.4049, "step": 17761 }, { "epoch": 2.77, "grad_norm": 20.238962076232934, "learning_rate": 2.950379276689197e-07, "loss": 0.4492, "step": 17762 }, { "epoch": 2.77, "grad_norm": 15.979882544482278, "learning_rate": 2.946314472454481e-07, "loss": 0.4367, "step": 17763 }, { "epoch": 2.77, "grad_norm": 23.16373111494247, "learning_rate": 2.9422524283698074e-07, "loss": 0.479, "step": 17764 }, { "epoch": 2.77, "grad_norm": 14.527442167625203, "learning_rate": 2.9381931445506826e-07, "loss": 0.434, "step": 17765 }, { "epoch": 2.78, "grad_norm": 22.600064304142172, "learning_rate": 2.934136621112571e-07, "loss": 0.4528, "step": 17766 }, { "epoch": 2.78, "grad_norm": 20.506270650759074, "learning_rate": 2.9300828581708464e-07, "loss": 0.4098, "step": 17767 }, { "epoch": 2.78, "grad_norm": 16.06691320673563, "learning_rate": 2.926031855840761e-07, "loss": 0.4246, "step": 17768 }, { "epoch": 2.78, "grad_norm": 23.005656413037922, "learning_rate": 2.9219836142375557e-07, "loss": 0.3901, "step": 17769 }, { "epoch": 2.78, "grad_norm": 25.414776448735328, "learning_rate": 2.917938133476339e-07, "loss": 0.498, "step": 17770 }, { "epoch": 2.78, "grad_norm": 19.923262375038078, "learning_rate": 2.913895413672174e-07, "loss": 0.4637, "step": 17771 }, { "epoch": 2.78, "grad_norm": 16.83951212911085, "learning_rate": 2.909855454940025e-07, "loss": 0.431, "step": 17772 }, { "epoch": 2.78, "grad_norm": 32.079504482302205, "learning_rate": 2.905818257394799e-07, "loss": 0.4811, "step": 17773 }, { "epoch": 2.78, "grad_norm": 21.802675499478198, "learning_rate": 2.901783821151305e-07, "loss": 0.3864, "step": 17774 }, { "epoch": 2.78, "grad_norm": 18.217788936758698, "learning_rate": 2.897752146324262e-07, "loss": 0.4397, "step": 17775 }, { "epoch": 2.78, "grad_norm": 29.758415823459107, "learning_rate": 2.893723233028367e-07, "loss": 0.429, "step": 17776 }, { "epoch": 2.78, "grad_norm": 21.32663103816911, "learning_rate": 2.889697081378162e-07, "loss": 0.4532, "step": 17777 }, { "epoch": 2.78, "grad_norm": 21.447919822576537, "learning_rate": 2.885673691488167e-07, "loss": 0.4094, "step": 17778 }, { "epoch": 2.78, "grad_norm": 22.50888376324859, "learning_rate": 2.8816530634728e-07, "loss": 0.3904, "step": 17779 }, { "epoch": 2.78, "grad_norm": 28.601438018269526, "learning_rate": 2.877635197446427e-07, "loss": 0.4367, "step": 17780 }, { "epoch": 2.78, "grad_norm": 11.398677674365361, "learning_rate": 2.873620093523266e-07, "loss": 0.4287, "step": 17781 }, { "epoch": 2.78, "grad_norm": 19.87314811662009, "learning_rate": 2.869607751817538e-07, "loss": 0.482, "step": 17782 }, { "epoch": 2.78, "grad_norm": 24.94801116452939, "learning_rate": 2.8655981724433625e-07, "loss": 0.418, "step": 17783 }, { "epoch": 2.78, "grad_norm": 15.651201468766368, "learning_rate": 2.861591355514748e-07, "loss": 0.4093, "step": 17784 }, { "epoch": 2.78, "grad_norm": 20.469020688108923, "learning_rate": 2.857587301145637e-07, "loss": 0.4403, "step": 17785 }, { "epoch": 2.78, "grad_norm": 25.939070353022245, "learning_rate": 2.853586009449927e-07, "loss": 0.4085, "step": 17786 }, { "epoch": 2.78, "grad_norm": 16.162492029552457, "learning_rate": 2.8495874805413937e-07, "loss": 0.3995, "step": 17787 }, { "epoch": 2.78, "grad_norm": 15.75535694908689, "learning_rate": 2.845591714533769e-07, "loss": 0.4161, "step": 17788 }, { "epoch": 2.78, "grad_norm": 24.344400072171297, "learning_rate": 2.841598711540694e-07, "loss": 0.4065, "step": 17789 }, { "epoch": 2.78, "grad_norm": 31.825420460429424, "learning_rate": 2.83760847167569e-07, "loss": 0.4998, "step": 17790 }, { "epoch": 2.78, "grad_norm": 24.79331486690112, "learning_rate": 2.8336209950522774e-07, "loss": 0.4992, "step": 17791 }, { "epoch": 2.78, "grad_norm": 29.61541954485294, "learning_rate": 2.829636281783843e-07, "loss": 0.4473, "step": 17792 }, { "epoch": 2.78, "grad_norm": 27.110733862248356, "learning_rate": 2.825654331983707e-07, "loss": 0.4556, "step": 17793 }, { "epoch": 2.78, "grad_norm": 17.257772404834387, "learning_rate": 2.8216751457651124e-07, "loss": 0.4037, "step": 17794 }, { "epoch": 2.78, "grad_norm": 30.776153927814182, "learning_rate": 2.8176987232412354e-07, "loss": 0.4971, "step": 17795 }, { "epoch": 2.78, "grad_norm": 19.981997399943353, "learning_rate": 2.8137250645251636e-07, "loss": 0.4503, "step": 17796 }, { "epoch": 2.78, "grad_norm": 21.78206100896924, "learning_rate": 2.8097541697298835e-07, "loss": 0.4174, "step": 17797 }, { "epoch": 2.78, "grad_norm": 18.219222166950722, "learning_rate": 2.80578603896835e-07, "loss": 0.3872, "step": 17798 }, { "epoch": 2.78, "grad_norm": 13.712528351656207, "learning_rate": 2.8018206723533947e-07, "loss": 0.3565, "step": 17799 }, { "epoch": 2.78, "grad_norm": 16.620171348536875, "learning_rate": 2.7978580699978054e-07, "loss": 0.3975, "step": 17800 }, { "epoch": 2.78, "grad_norm": 28.43347515820931, "learning_rate": 2.7938982320142693e-07, "loss": 0.4835, "step": 17801 }, { "epoch": 2.78, "grad_norm": 33.41241235295184, "learning_rate": 2.789941158515408e-07, "loss": 0.4707, "step": 17802 }, { "epoch": 2.78, "grad_norm": 34.165840575894414, "learning_rate": 2.785986849613753e-07, "loss": 0.4095, "step": 17803 }, { "epoch": 2.78, "grad_norm": 22.33324719318697, "learning_rate": 2.78203530542176e-07, "loss": 0.4473, "step": 17804 }, { "epoch": 2.78, "grad_norm": 24.86358169649824, "learning_rate": 2.7780865260518265e-07, "loss": 0.4104, "step": 17805 }, { "epoch": 2.78, "grad_norm": 17.247536409893662, "learning_rate": 2.774140511616219e-07, "loss": 0.3936, "step": 17806 }, { "epoch": 2.78, "grad_norm": 24.528570963119602, "learning_rate": 2.770197262227181e-07, "loss": 0.4658, "step": 17807 }, { "epoch": 2.78, "grad_norm": 18.644543764248844, "learning_rate": 2.7662567779968677e-07, "loss": 0.4331, "step": 17808 }, { "epoch": 2.78, "grad_norm": 21.046297085607907, "learning_rate": 2.762319059037333e-07, "loss": 0.446, "step": 17809 }, { "epoch": 2.78, "grad_norm": 16.52821308456191, "learning_rate": 2.758384105460554e-07, "loss": 0.3938, "step": 17810 }, { "epoch": 2.78, "grad_norm": 18.39181994345637, "learning_rate": 2.7544519173784423e-07, "loss": 0.4082, "step": 17811 }, { "epoch": 2.78, "grad_norm": 24.79687175558414, "learning_rate": 2.7505224949028297e-07, "loss": 0.4646, "step": 17812 }, { "epoch": 2.78, "grad_norm": 30.372453472941118, "learning_rate": 2.7465958381454714e-07, "loss": 0.4564, "step": 17813 }, { "epoch": 2.78, "grad_norm": 20.662787697018608, "learning_rate": 2.7426719472180454e-07, "loss": 0.4665, "step": 17814 }, { "epoch": 2.78, "grad_norm": 20.80120540974446, "learning_rate": 2.738750822232128e-07, "loss": 0.4285, "step": 17815 }, { "epoch": 2.78, "grad_norm": 19.817918329884925, "learning_rate": 2.7348324632992416e-07, "loss": 0.3917, "step": 17816 }, { "epoch": 2.78, "grad_norm": 15.154300239838921, "learning_rate": 2.7309168705308196e-07, "loss": 0.3714, "step": 17817 }, { "epoch": 2.78, "grad_norm": 24.161499256870428, "learning_rate": 2.7270040440382283e-07, "loss": 0.4113, "step": 17818 }, { "epoch": 2.78, "grad_norm": 16.196009073951082, "learning_rate": 2.723093983932734e-07, "loss": 0.3644, "step": 17819 }, { "epoch": 2.78, "grad_norm": 35.6433101626748, "learning_rate": 2.719186690325537e-07, "loss": 0.4416, "step": 17820 }, { "epoch": 2.78, "grad_norm": 19.85555892314469, "learning_rate": 2.7152821633277705e-07, "loss": 0.418, "step": 17821 }, { "epoch": 2.78, "grad_norm": 18.78496606706317, "learning_rate": 2.7113804030504564e-07, "loss": 0.4408, "step": 17822 }, { "epoch": 2.78, "grad_norm": 20.062839569058323, "learning_rate": 2.7074814096045954e-07, "loss": 0.3768, "step": 17823 }, { "epoch": 2.78, "grad_norm": 14.511767594472179, "learning_rate": 2.7035851831010206e-07, "loss": 0.3782, "step": 17824 }, { "epoch": 2.78, "grad_norm": 17.075831909355863, "learning_rate": 2.699691723650577e-07, "loss": 0.4239, "step": 17825 }, { "epoch": 2.78, "grad_norm": 22.544019319984162, "learning_rate": 2.6958010313639873e-07, "loss": 0.4414, "step": 17826 }, { "epoch": 2.78, "grad_norm": 25.612221742317452, "learning_rate": 2.6919131063518956e-07, "loss": 0.4288, "step": 17827 }, { "epoch": 2.78, "grad_norm": 31.06784850113173, "learning_rate": 2.6880279487248697e-07, "loss": 0.4189, "step": 17828 }, { "epoch": 2.78, "grad_norm": 21.37054766468191, "learning_rate": 2.684145558593398e-07, "loss": 0.4404, "step": 17829 }, { "epoch": 2.79, "grad_norm": 21.728114558601035, "learning_rate": 2.680265936067905e-07, "loss": 0.42, "step": 17830 }, { "epoch": 2.79, "grad_norm": 26.826270009886528, "learning_rate": 2.6763890812587126e-07, "loss": 0.4683, "step": 17831 }, { "epoch": 2.79, "grad_norm": 26.725095256460804, "learning_rate": 2.6725149942760875e-07, "loss": 0.4641, "step": 17832 }, { "epoch": 2.79, "grad_norm": 16.005758651779562, "learning_rate": 2.668643675230198e-07, "loss": 0.4046, "step": 17833 }, { "epoch": 2.79, "grad_norm": 20.67878561715757, "learning_rate": 2.664775124231156e-07, "loss": 0.4316, "step": 17834 }, { "epoch": 2.79, "grad_norm": 18.55525916574125, "learning_rate": 2.660909341388951e-07, "loss": 0.3749, "step": 17835 }, { "epoch": 2.79, "grad_norm": 27.402857852172595, "learning_rate": 2.65704632681355e-07, "loss": 0.4265, "step": 17836 }, { "epoch": 2.79, "grad_norm": 18.315326874458894, "learning_rate": 2.653186080614811e-07, "loss": 0.3921, "step": 17837 }, { "epoch": 2.79, "grad_norm": 20.873260227104232, "learning_rate": 2.649328602902501e-07, "loss": 0.4263, "step": 17838 }, { "epoch": 2.79, "grad_norm": 21.511431394077448, "learning_rate": 2.6454738937863545e-07, "loss": 0.3848, "step": 17839 }, { "epoch": 2.79, "grad_norm": 23.758257753225937, "learning_rate": 2.6416219533759724e-07, "loss": 0.3854, "step": 17840 }, { "epoch": 2.79, "grad_norm": 20.036648640242426, "learning_rate": 2.637772781780901e-07, "loss": 0.4291, "step": 17841 }, { "epoch": 2.79, "grad_norm": 16.628179702317524, "learning_rate": 2.6339263791106294e-07, "loss": 0.4086, "step": 17842 }, { "epoch": 2.79, "grad_norm": 22.163010623028203, "learning_rate": 2.630082745474538e-07, "loss": 0.4376, "step": 17843 }, { "epoch": 2.79, "grad_norm": 20.723038656749363, "learning_rate": 2.6262418809819166e-07, "loss": 0.3908, "step": 17844 }, { "epoch": 2.79, "grad_norm": 20.039714017393614, "learning_rate": 2.6224037857420225e-07, "loss": 0.4845, "step": 17845 }, { "epoch": 2.79, "grad_norm": 24.176386528678915, "learning_rate": 2.6185684598640013e-07, "loss": 0.3959, "step": 17846 }, { "epoch": 2.79, "grad_norm": 23.52618035637394, "learning_rate": 2.614735903456933e-07, "loss": 0.4044, "step": 17847 }, { "epoch": 2.79, "grad_norm": 21.43696592738896, "learning_rate": 2.610906116629797e-07, "loss": 0.4376, "step": 17848 }, { "epoch": 2.79, "grad_norm": 22.564452098105797, "learning_rate": 2.607079099491516e-07, "loss": 0.4248, "step": 17849 }, { "epoch": 2.79, "grad_norm": 18.37316254670309, "learning_rate": 2.6032548521509384e-07, "loss": 0.4104, "step": 17850 }, { "epoch": 2.79, "grad_norm": 15.671465737161004, "learning_rate": 2.59943337471682e-07, "loss": 0.3953, "step": 17851 }, { "epoch": 2.79, "grad_norm": 27.93382537280599, "learning_rate": 2.595614667297841e-07, "loss": 0.4228, "step": 17852 }, { "epoch": 2.79, "grad_norm": 27.247811925166463, "learning_rate": 2.591798730002593e-07, "loss": 0.4455, "step": 17853 }, { "epoch": 2.79, "grad_norm": 25.55483761390252, "learning_rate": 2.587985562939599e-07, "loss": 0.4557, "step": 17854 }, { "epoch": 2.79, "grad_norm": 33.14607875360679, "learning_rate": 2.5841751662173287e-07, "loss": 0.5067, "step": 17855 }, { "epoch": 2.79, "grad_norm": 18.296160658338522, "learning_rate": 2.5803675399441173e-07, "loss": 0.5111, "step": 17856 }, { "epoch": 2.79, "grad_norm": 19.43695436414049, "learning_rate": 2.576562684228279e-07, "loss": 0.406, "step": 17857 }, { "epoch": 2.79, "grad_norm": 27.328284299679527, "learning_rate": 2.5727605991779925e-07, "loss": 0.4766, "step": 17858 }, { "epoch": 2.79, "grad_norm": 27.19147341620862, "learning_rate": 2.5689612849014166e-07, "loss": 0.4341, "step": 17859 }, { "epoch": 2.79, "grad_norm": 25.58358957947211, "learning_rate": 2.565164741506576e-07, "loss": 0.4133, "step": 17860 }, { "epoch": 2.79, "grad_norm": 17.142896297440632, "learning_rate": 2.56137096910144e-07, "loss": 0.4456, "step": 17861 }, { "epoch": 2.79, "grad_norm": 26.02469362052001, "learning_rate": 2.557579967793933e-07, "loss": 0.467, "step": 17862 }, { "epoch": 2.79, "grad_norm": 22.910104164526125, "learning_rate": 2.5537917376918466e-07, "loss": 0.4748, "step": 17863 }, { "epoch": 2.79, "grad_norm": 30.37279936389306, "learning_rate": 2.5500062789029165e-07, "loss": 0.3871, "step": 17864 }, { "epoch": 2.79, "grad_norm": 16.313061759308482, "learning_rate": 2.546223591534802e-07, "loss": 0.3915, "step": 17865 }, { "epoch": 2.79, "grad_norm": 19.73817612321626, "learning_rate": 2.542443675695072e-07, "loss": 0.4208, "step": 17866 }, { "epoch": 2.79, "grad_norm": 16.748565011865182, "learning_rate": 2.538666531491241e-07, "loss": 0.4057, "step": 17867 }, { "epoch": 2.79, "grad_norm": 17.54794355741235, "learning_rate": 2.5348921590307216e-07, "loss": 0.4045, "step": 17868 }, { "epoch": 2.79, "grad_norm": 37.00632794176606, "learning_rate": 2.5311205584208523e-07, "loss": 0.4751, "step": 17869 }, { "epoch": 2.79, "grad_norm": 37.13500278433084, "learning_rate": 2.52735172976889e-07, "loss": 0.4981, "step": 17870 }, { "epoch": 2.79, "grad_norm": 23.010762255813322, "learning_rate": 2.5235856731820276e-07, "loss": 0.4664, "step": 17871 }, { "epoch": 2.79, "grad_norm": 18.37970622446912, "learning_rate": 2.519822388767379e-07, "loss": 0.4873, "step": 17872 }, { "epoch": 2.79, "grad_norm": 48.70518842419169, "learning_rate": 2.5160618766319477e-07, "loss": 0.4545, "step": 17873 }, { "epoch": 2.79, "grad_norm": 32.41233336258351, "learning_rate": 2.512304136882682e-07, "loss": 0.4451, "step": 17874 }, { "epoch": 2.79, "grad_norm": 19.278552234495788, "learning_rate": 2.508549169626462e-07, "loss": 0.4012, "step": 17875 }, { "epoch": 2.79, "grad_norm": 17.404091322349313, "learning_rate": 2.504796974970081e-07, "loss": 0.3592, "step": 17876 }, { "epoch": 2.79, "grad_norm": 25.55736350355255, "learning_rate": 2.5010475530202414e-07, "loss": 0.3977, "step": 17877 }, { "epoch": 2.79, "grad_norm": 15.432892991775566, "learning_rate": 2.4973009038835593e-07, "loss": 0.3709, "step": 17878 }, { "epoch": 2.79, "grad_norm": 23.600486098199184, "learning_rate": 2.4935570276666157e-07, "loss": 0.4114, "step": 17879 }, { "epoch": 2.79, "grad_norm": 43.655283837355256, "learning_rate": 2.4898159244758693e-07, "loss": 0.4455, "step": 17880 }, { "epoch": 2.79, "grad_norm": 31.267924611682588, "learning_rate": 2.486077594417724e-07, "loss": 0.3964, "step": 17881 }, { "epoch": 2.79, "grad_norm": 20.25477956399121, "learning_rate": 2.482342037598473e-07, "loss": 0.5159, "step": 17882 }, { "epoch": 2.79, "grad_norm": 21.82578694458567, "learning_rate": 2.4786092541243645e-07, "loss": 0.4631, "step": 17883 }, { "epoch": 2.79, "grad_norm": 34.87979369989602, "learning_rate": 2.474879244101569e-07, "loss": 0.4869, "step": 17884 }, { "epoch": 2.79, "grad_norm": 17.024314767607805, "learning_rate": 2.471152007636157e-07, "loss": 0.4116, "step": 17885 }, { "epoch": 2.79, "grad_norm": 17.135663809825058, "learning_rate": 2.4674275448341333e-07, "loss": 0.4106, "step": 17886 }, { "epoch": 2.79, "grad_norm": 17.26663923434738, "learning_rate": 2.463705855801413e-07, "loss": 0.3772, "step": 17887 }, { "epoch": 2.79, "grad_norm": 29.899714734935777, "learning_rate": 2.4599869406438327e-07, "loss": 0.4664, "step": 17888 }, { "epoch": 2.79, "grad_norm": 26.476963347470463, "learning_rate": 2.456270799467175e-07, "loss": 0.4235, "step": 17889 }, { "epoch": 2.79, "grad_norm": 23.803210952907094, "learning_rate": 2.452557432377101e-07, "loss": 0.4169, "step": 17890 }, { "epoch": 2.79, "grad_norm": 25.70008974908491, "learning_rate": 2.4488468394792463e-07, "loss": 0.4275, "step": 17891 }, { "epoch": 2.79, "grad_norm": 27.93936782231887, "learning_rate": 2.4451390208791173e-07, "loss": 0.4253, "step": 17892 }, { "epoch": 2.79, "grad_norm": 22.847078125282508, "learning_rate": 2.4414339766821614e-07, "loss": 0.4504, "step": 17893 }, { "epoch": 2.8, "grad_norm": 28.535500378005608, "learning_rate": 2.437731706993751e-07, "loss": 0.4524, "step": 17894 }, { "epoch": 2.8, "grad_norm": 19.802235770219895, "learning_rate": 2.434032211919179e-07, "loss": 0.4161, "step": 17895 }, { "epoch": 2.8, "grad_norm": 23.109124842586482, "learning_rate": 2.4303354915636735e-07, "loss": 0.3741, "step": 17896 }, { "epoch": 2.8, "grad_norm": 22.01378010572684, "learning_rate": 2.426641546032338e-07, "loss": 0.4588, "step": 17897 }, { "epoch": 2.8, "grad_norm": 16.192167988932617, "learning_rate": 2.422950375430233e-07, "loss": 0.3965, "step": 17898 }, { "epoch": 2.8, "grad_norm": 35.99056494918064, "learning_rate": 2.419261979862342e-07, "loss": 0.4083, "step": 17899 }, { "epoch": 2.8, "grad_norm": 17.44663932228759, "learning_rate": 2.415576359433558e-07, "loss": 0.3789, "step": 17900 }, { "epoch": 2.8, "grad_norm": 21.106621868369555, "learning_rate": 2.4118935142486975e-07, "loss": 0.4439, "step": 17901 }, { "epoch": 2.8, "grad_norm": 22.056615825907002, "learning_rate": 2.408213444412499e-07, "loss": 0.4741, "step": 17902 }, { "epoch": 2.8, "grad_norm": 15.947655725701033, "learning_rate": 2.404536150029624e-07, "loss": 0.4008, "step": 17903 }, { "epoch": 2.8, "grad_norm": 27.189831583301878, "learning_rate": 2.4008616312046315e-07, "loss": 0.4449, "step": 17904 }, { "epoch": 2.8, "grad_norm": 18.155705264877064, "learning_rate": 2.397189888042062e-07, "loss": 0.4005, "step": 17905 }, { "epoch": 2.8, "grad_norm": 24.218961599102407, "learning_rate": 2.3935209206463195e-07, "loss": 0.4187, "step": 17906 }, { "epoch": 2.8, "grad_norm": 16.703775366458647, "learning_rate": 2.389854729121721e-07, "loss": 0.4173, "step": 17907 }, { "epoch": 2.8, "grad_norm": 17.050152176026867, "learning_rate": 2.3861913135725613e-07, "loss": 0.378, "step": 17908 }, { "epoch": 2.8, "grad_norm": 19.44977727219458, "learning_rate": 2.382530674103023e-07, "loss": 0.4788, "step": 17909 }, { "epoch": 2.8, "grad_norm": 28.367481589992543, "learning_rate": 2.3788728108172121e-07, "loss": 0.4312, "step": 17910 }, { "epoch": 2.8, "grad_norm": 20.97728748311856, "learning_rate": 2.3752177238191455e-07, "loss": 0.4681, "step": 17911 }, { "epoch": 2.8, "grad_norm": 21.009056153800827, "learning_rate": 2.371565413212773e-07, "loss": 0.435, "step": 17912 }, { "epoch": 2.8, "grad_norm": 22.329011798063423, "learning_rate": 2.3679158791019673e-07, "loss": 0.3975, "step": 17913 }, { "epoch": 2.8, "grad_norm": 26.850000225587483, "learning_rate": 2.3642691215905344e-07, "loss": 0.4575, "step": 17914 }, { "epoch": 2.8, "grad_norm": 25.12512308536575, "learning_rate": 2.360625140782169e-07, "loss": 0.4266, "step": 17915 }, { "epoch": 2.8, "grad_norm": 27.387586907688945, "learning_rate": 2.3569839367805103e-07, "loss": 0.4544, "step": 17916 }, { "epoch": 2.8, "grad_norm": 31.57983804281556, "learning_rate": 2.3533455096890978e-07, "loss": 0.4846, "step": 17917 }, { "epoch": 2.8, "grad_norm": 24.851156143276796, "learning_rate": 2.3497098596114265e-07, "loss": 0.4673, "step": 17918 }, { "epoch": 2.8, "grad_norm": 16.525718790305167, "learning_rate": 2.3460769866508693e-07, "loss": 0.4218, "step": 17919 }, { "epoch": 2.8, "grad_norm": 18.95308114193766, "learning_rate": 2.3424468909107656e-07, "loss": 0.4226, "step": 17920 }, { "epoch": 2.8, "grad_norm": 22.58590657467277, "learning_rate": 2.338819572494355e-07, "loss": 0.4016, "step": 17921 }, { "epoch": 2.8, "grad_norm": 24.077446403818065, "learning_rate": 2.3351950315047777e-07, "loss": 0.4855, "step": 17922 }, { "epoch": 2.8, "grad_norm": 34.780151764564344, "learning_rate": 2.3315732680451175e-07, "loss": 0.4779, "step": 17923 }, { "epoch": 2.8, "grad_norm": 23.17808452343965, "learning_rate": 2.3279542822183698e-07, "loss": 0.4154, "step": 17924 }, { "epoch": 2.8, "grad_norm": 27.08709453612449, "learning_rate": 2.3243380741274745e-07, "loss": 0.4062, "step": 17925 }, { "epoch": 2.8, "grad_norm": 16.56946141756476, "learning_rate": 2.3207246438752719e-07, "loss": 0.3592, "step": 17926 }, { "epoch": 2.8, "grad_norm": 30.321893269192223, "learning_rate": 2.3171139915645235e-07, "loss": 0.4537, "step": 17927 }, { "epoch": 2.8, "grad_norm": 21.366456171662513, "learning_rate": 2.3135061172979034e-07, "loss": 0.468, "step": 17928 }, { "epoch": 2.8, "grad_norm": 23.441128012555836, "learning_rate": 2.3099010211780183e-07, "loss": 0.4628, "step": 17929 }, { "epoch": 2.8, "grad_norm": 26.74679484617763, "learning_rate": 2.306298703307408e-07, "loss": 0.4472, "step": 17930 }, { "epoch": 2.8, "grad_norm": 27.431240642995363, "learning_rate": 2.3026991637885242e-07, "loss": 0.445, "step": 17931 }, { "epoch": 2.8, "grad_norm": 27.38865373158325, "learning_rate": 2.2991024027237075e-07, "loss": 0.4536, "step": 17932 }, { "epoch": 2.8, "grad_norm": 17.054190229574676, "learning_rate": 2.2955084202152644e-07, "loss": 0.3965, "step": 17933 }, { "epoch": 2.8, "grad_norm": 19.367665986616785, "learning_rate": 2.2919172163654134e-07, "loss": 0.405, "step": 17934 }, { "epoch": 2.8, "grad_norm": 15.21667680189889, "learning_rate": 2.2883287912762842e-07, "loss": 0.3957, "step": 17935 }, { "epoch": 2.8, "grad_norm": 24.55229995509514, "learning_rate": 2.2847431450499169e-07, "loss": 0.4233, "step": 17936 }, { "epoch": 2.8, "grad_norm": 31.149448112394793, "learning_rate": 2.281160277788297e-07, "loss": 0.4639, "step": 17937 }, { "epoch": 2.8, "grad_norm": 24.998319068664873, "learning_rate": 2.277580189593298e-07, "loss": 0.4088, "step": 17938 }, { "epoch": 2.8, "grad_norm": 17.619317761211487, "learning_rate": 2.2740028805667725e-07, "loss": 0.4614, "step": 17939 }, { "epoch": 2.8, "grad_norm": 27.025920572975707, "learning_rate": 2.2704283508104386e-07, "loss": 0.4051, "step": 17940 }, { "epoch": 2.8, "grad_norm": 26.720652305075667, "learning_rate": 2.2668566004259486e-07, "loss": 0.4812, "step": 17941 }, { "epoch": 2.8, "grad_norm": 22.52308261524274, "learning_rate": 2.2632876295148765e-07, "loss": 0.392, "step": 17942 }, { "epoch": 2.8, "grad_norm": 27.196821214205656, "learning_rate": 2.2597214381787301e-07, "loss": 0.4463, "step": 17943 }, { "epoch": 2.8, "grad_norm": 22.770853842847917, "learning_rate": 2.2561580265189397e-07, "loss": 0.4212, "step": 17944 }, { "epoch": 2.8, "grad_norm": 20.035732236463883, "learning_rate": 2.2525973946368352e-07, "loss": 0.4327, "step": 17945 }, { "epoch": 2.8, "grad_norm": 27.25822230189805, "learning_rate": 2.2490395426336687e-07, "loss": 0.4215, "step": 17946 }, { "epoch": 2.8, "grad_norm": 19.332611294960905, "learning_rate": 2.2454844706106482e-07, "loss": 0.4411, "step": 17947 }, { "epoch": 2.8, "grad_norm": 25.437310616739577, "learning_rate": 2.2419321786688485e-07, "loss": 0.4239, "step": 17948 }, { "epoch": 2.8, "grad_norm": 19.242596139103917, "learning_rate": 2.2383826669093222e-07, "loss": 0.4182, "step": 17949 }, { "epoch": 2.8, "grad_norm": 31.37208079629567, "learning_rate": 2.2348359354330106e-07, "loss": 0.4515, "step": 17950 }, { "epoch": 2.8, "grad_norm": 19.475944026074135, "learning_rate": 2.2312919843407555e-07, "loss": 0.4151, "step": 17951 }, { "epoch": 2.8, "grad_norm": 23.45641486178327, "learning_rate": 2.227750813733376e-07, "loss": 0.4315, "step": 17952 }, { "epoch": 2.8, "grad_norm": 18.957619987855367, "learning_rate": 2.2242124237115693e-07, "loss": 0.4343, "step": 17953 }, { "epoch": 2.8, "grad_norm": 24.50887749060113, "learning_rate": 2.220676814375955e-07, "loss": 0.4303, "step": 17954 }, { "epoch": 2.8, "grad_norm": 18.942143100024747, "learning_rate": 2.217143985827097e-07, "loss": 0.4136, "step": 17955 }, { "epoch": 2.8, "grad_norm": 33.555725354274415, "learning_rate": 2.2136139381654708e-07, "loss": 0.4696, "step": 17956 }, { "epoch": 2.8, "grad_norm": 20.660537463771654, "learning_rate": 2.2100866714914514e-07, "loss": 0.3986, "step": 17957 }, { "epoch": 2.81, "grad_norm": 25.221334921705278, "learning_rate": 2.2065621859053587e-07, "loss": 0.3785, "step": 17958 }, { "epoch": 2.81, "grad_norm": 19.54857315566461, "learning_rate": 2.2030404815074346e-07, "loss": 0.4516, "step": 17959 }, { "epoch": 2.81, "grad_norm": 19.719385980183056, "learning_rate": 2.1995215583978435e-07, "loss": 0.4367, "step": 17960 }, { "epoch": 2.81, "grad_norm": 18.032189383986218, "learning_rate": 2.1960054166766386e-07, "loss": 0.4147, "step": 17961 }, { "epoch": 2.81, "grad_norm": 19.170908254158707, "learning_rate": 2.1924920564438178e-07, "loss": 0.3889, "step": 17962 }, { "epoch": 2.81, "grad_norm": 13.646906441494428, "learning_rate": 2.1889814777993013e-07, "loss": 0.394, "step": 17963 }, { "epoch": 2.81, "grad_norm": 20.925506812705247, "learning_rate": 2.1854736808429533e-07, "loss": 0.4245, "step": 17964 }, { "epoch": 2.81, "grad_norm": 21.66466775180062, "learning_rate": 2.1819686656745053e-07, "loss": 0.3924, "step": 17965 }, { "epoch": 2.81, "grad_norm": 19.6367657252809, "learning_rate": 2.1784664323936445e-07, "loss": 0.4043, "step": 17966 }, { "epoch": 2.81, "grad_norm": 25.429948439189545, "learning_rate": 2.1749669810999686e-07, "loss": 0.4044, "step": 17967 }, { "epoch": 2.81, "grad_norm": 34.47296548072508, "learning_rate": 2.1714703118930203e-07, "loss": 0.4403, "step": 17968 }, { "epoch": 2.81, "grad_norm": 25.426145397725204, "learning_rate": 2.16797642487222e-07, "loss": 0.4375, "step": 17969 }, { "epoch": 2.81, "grad_norm": 21.97156176744616, "learning_rate": 2.1644853201369444e-07, "loss": 0.4265, "step": 17970 }, { "epoch": 2.81, "grad_norm": 20.83134378700535, "learning_rate": 2.1609969977864687e-07, "loss": 0.4478, "step": 17971 }, { "epoch": 2.81, "grad_norm": 21.90028893533822, "learning_rate": 2.157511457920014e-07, "loss": 0.3927, "step": 17972 }, { "epoch": 2.81, "grad_norm": 15.405559098728263, "learning_rate": 2.1540287006366788e-07, "loss": 0.3829, "step": 17973 }, { "epoch": 2.81, "grad_norm": 17.163114747719167, "learning_rate": 2.1505487260355396e-07, "loss": 0.346, "step": 17974 }, { "epoch": 2.81, "grad_norm": 30.13482991280381, "learning_rate": 2.147071534215561e-07, "loss": 0.4346, "step": 17975 }, { "epoch": 2.81, "grad_norm": 20.43033444490758, "learning_rate": 2.1435971252756093e-07, "loss": 0.4628, "step": 17976 }, { "epoch": 2.81, "grad_norm": 18.91660346603199, "learning_rate": 2.140125499314527e-07, "loss": 0.4146, "step": 17977 }, { "epoch": 2.81, "grad_norm": 19.263103447713288, "learning_rate": 2.1366566564310244e-07, "loss": 0.4877, "step": 17978 }, { "epoch": 2.81, "grad_norm": 22.674337070683304, "learning_rate": 2.1331905967237553e-07, "loss": 0.4124, "step": 17979 }, { "epoch": 2.81, "grad_norm": 25.106870838529687, "learning_rate": 2.1297273202912972e-07, "loss": 0.4446, "step": 17980 }, { "epoch": 2.81, "grad_norm": 31.21133850374774, "learning_rate": 2.1262668272321486e-07, "loss": 0.4912, "step": 17981 }, { "epoch": 2.81, "grad_norm": 22.936691970261247, "learning_rate": 2.1228091176447086e-07, "loss": 0.4554, "step": 17982 }, { "epoch": 2.81, "grad_norm": 29.217553275381192, "learning_rate": 2.1193541916273097e-07, "loss": 0.474, "step": 17983 }, { "epoch": 2.81, "grad_norm": 30.885164960164687, "learning_rate": 2.11590204927824e-07, "loss": 0.4252, "step": 17984 }, { "epoch": 2.81, "grad_norm": 18.9163463589218, "learning_rate": 2.1124526906956545e-07, "loss": 0.4332, "step": 17985 }, { "epoch": 2.81, "grad_norm": 19.514917725054374, "learning_rate": 2.109006115977641e-07, "loss": 0.4102, "step": 17986 }, { "epoch": 2.81, "grad_norm": 30.945164940878776, "learning_rate": 2.1055623252222211e-07, "loss": 0.4305, "step": 17987 }, { "epoch": 2.81, "grad_norm": 14.402236483872583, "learning_rate": 2.1021213185273504e-07, "loss": 0.3995, "step": 17988 }, { "epoch": 2.81, "grad_norm": 19.70476495208655, "learning_rate": 2.0986830959908943e-07, "loss": 0.4086, "step": 17989 }, { "epoch": 2.81, "grad_norm": 16.44522549748909, "learning_rate": 2.0952476577106085e-07, "loss": 0.3842, "step": 17990 }, { "epoch": 2.81, "grad_norm": 21.175598975506993, "learning_rate": 2.0918150037842032e-07, "loss": 0.3839, "step": 17991 }, { "epoch": 2.81, "grad_norm": 18.379939230450287, "learning_rate": 2.088385134309312e-07, "loss": 0.4379, "step": 17992 }, { "epoch": 2.81, "grad_norm": 25.57784531265054, "learning_rate": 2.0849580493834675e-07, "loss": 0.4463, "step": 17993 }, { "epoch": 2.81, "grad_norm": 26.167006690036793, "learning_rate": 2.0815337491041476e-07, "loss": 0.4224, "step": 17994 }, { "epoch": 2.81, "grad_norm": 27.843951997334646, "learning_rate": 2.0781122335687075e-07, "loss": 0.4467, "step": 17995 }, { "epoch": 2.81, "grad_norm": 17.981891019743163, "learning_rate": 2.0746935028744808e-07, "loss": 0.3986, "step": 17996 }, { "epoch": 2.81, "grad_norm": 26.026015872279658, "learning_rate": 2.0712775571186895e-07, "loss": 0.4376, "step": 17997 }, { "epoch": 2.81, "grad_norm": 17.838699151246583, "learning_rate": 2.067864396398478e-07, "loss": 0.3984, "step": 17998 }, { "epoch": 2.81, "grad_norm": 30.99544139480847, "learning_rate": 2.0644540208109133e-07, "loss": 0.4548, "step": 17999 }, { "epoch": 2.81, "grad_norm": 27.062223881244527, "learning_rate": 2.0610464304529843e-07, "loss": 0.4533, "step": 18000 }, { "epoch": 2.81, "grad_norm": 17.443422817345073, "learning_rate": 2.0576416254216024e-07, "loss": 0.4411, "step": 18001 }, { "epoch": 2.81, "grad_norm": 24.47054582219941, "learning_rate": 2.0542396058136015e-07, "loss": 0.462, "step": 18002 }, { "epoch": 2.81, "grad_norm": 23.04319876947105, "learning_rate": 2.0508403717257263e-07, "loss": 0.4435, "step": 18003 }, { "epoch": 2.81, "grad_norm": 25.397809130271202, "learning_rate": 2.047443923254655e-07, "loss": 0.3962, "step": 18004 }, { "epoch": 2.81, "grad_norm": 13.505013511246501, "learning_rate": 2.0440502604969659e-07, "loss": 0.3772, "step": 18005 }, { "epoch": 2.81, "grad_norm": 13.704697973212328, "learning_rate": 2.040659383549204e-07, "loss": 0.4494, "step": 18006 }, { "epoch": 2.81, "grad_norm": 30.39206411195344, "learning_rate": 2.0372712925077697e-07, "loss": 0.4095, "step": 18007 }, { "epoch": 2.81, "grad_norm": 26.479218452469503, "learning_rate": 2.0338859874690308e-07, "loss": 0.4731, "step": 18008 }, { "epoch": 2.81, "grad_norm": 36.94376860534872, "learning_rate": 2.030503468529288e-07, "loss": 0.4564, "step": 18009 }, { "epoch": 2.81, "grad_norm": 16.20613844833497, "learning_rate": 2.0271237357847085e-07, "loss": 0.3896, "step": 18010 }, { "epoch": 2.81, "grad_norm": 25.56121906787919, "learning_rate": 2.0237467893314044e-07, "loss": 0.423, "step": 18011 }, { "epoch": 2.81, "grad_norm": 18.851176906520124, "learning_rate": 2.0203726292654324e-07, "loss": 0.4429, "step": 18012 }, { "epoch": 2.81, "grad_norm": 17.81612367579466, "learning_rate": 2.0170012556827379e-07, "loss": 0.4194, "step": 18013 }, { "epoch": 2.81, "grad_norm": 21.072584658421118, "learning_rate": 2.0136326686792217e-07, "loss": 0.4459, "step": 18014 }, { "epoch": 2.81, "grad_norm": 18.352500836323383, "learning_rate": 2.0102668683506743e-07, "loss": 0.4194, "step": 18015 }, { "epoch": 2.81, "grad_norm": 29.785433168218354, "learning_rate": 2.0069038547928078e-07, "loss": 0.4289, "step": 18016 }, { "epoch": 2.81, "grad_norm": 27.201255488696873, "learning_rate": 2.0035436281012678e-07, "loss": 0.4176, "step": 18017 }, { "epoch": 2.81, "grad_norm": 27.335751808431972, "learning_rate": 2.0001861883716335e-07, "loss": 0.512, "step": 18018 }, { "epoch": 2.81, "grad_norm": 17.189362042739145, "learning_rate": 1.996831535699373e-07, "loss": 0.4082, "step": 18019 }, { "epoch": 2.81, "grad_norm": 24.687079651791045, "learning_rate": 1.9934796701798876e-07, "loss": 0.4497, "step": 18020 }, { "epoch": 2.81, "grad_norm": 28.828919896484148, "learning_rate": 1.9901305919085123e-07, "loss": 0.4248, "step": 18021 }, { "epoch": 2.82, "grad_norm": 27.216542038342144, "learning_rate": 1.9867843009804822e-07, "loss": 0.4025, "step": 18022 }, { "epoch": 2.82, "grad_norm": 26.15781855807853, "learning_rate": 1.9834407974909764e-07, "loss": 0.3836, "step": 18023 }, { "epoch": 2.82, "grad_norm": 25.04948801610312, "learning_rate": 1.9801000815350745e-07, "loss": 0.4651, "step": 18024 }, { "epoch": 2.82, "grad_norm": 16.298808560093953, "learning_rate": 1.9767621532077896e-07, "loss": 0.4397, "step": 18025 }, { "epoch": 2.82, "grad_norm": 25.37913525058086, "learning_rate": 1.9734270126040344e-07, "loss": 0.432, "step": 18026 }, { "epoch": 2.82, "grad_norm": 18.940686963333558, "learning_rate": 1.9700946598186887e-07, "loss": 0.4241, "step": 18027 }, { "epoch": 2.82, "grad_norm": 25.604745911987667, "learning_rate": 1.966765094946499e-07, "loss": 0.4077, "step": 18028 }, { "epoch": 2.82, "grad_norm": 28.332987679916233, "learning_rate": 1.963438318082156e-07, "loss": 0.4453, "step": 18029 }, { "epoch": 2.82, "grad_norm": 16.15408249229552, "learning_rate": 1.960114329320284e-07, "loss": 0.4226, "step": 18030 }, { "epoch": 2.82, "grad_norm": 21.414135281127027, "learning_rate": 1.9567931287554075e-07, "loss": 0.444, "step": 18031 }, { "epoch": 2.82, "grad_norm": 21.12404062488127, "learning_rate": 1.9534747164819733e-07, "loss": 0.4192, "step": 18032 }, { "epoch": 2.82, "grad_norm": 26.36948730159487, "learning_rate": 1.9501590925943726e-07, "loss": 0.4181, "step": 18033 }, { "epoch": 2.82, "grad_norm": 29.864448556483914, "learning_rate": 1.9468462571868852e-07, "loss": 0.4646, "step": 18034 }, { "epoch": 2.82, "grad_norm": 20.638843784390666, "learning_rate": 1.9435362103537357e-07, "loss": 0.3883, "step": 18035 }, { "epoch": 2.82, "grad_norm": 21.180132010990018, "learning_rate": 1.9402289521890495e-07, "loss": 0.4655, "step": 18036 }, { "epoch": 2.82, "grad_norm": 19.873556200103817, "learning_rate": 1.936924482786884e-07, "loss": 0.4052, "step": 18037 }, { "epoch": 2.82, "grad_norm": 24.944717047681387, "learning_rate": 1.9336228022412306e-07, "loss": 0.4075, "step": 18038 }, { "epoch": 2.82, "grad_norm": 15.459358404554317, "learning_rate": 1.9303239106459703e-07, "loss": 0.4124, "step": 18039 }, { "epoch": 2.82, "grad_norm": 21.725490016865255, "learning_rate": 1.927027808094939e-07, "loss": 0.402, "step": 18040 }, { "epoch": 2.82, "grad_norm": 17.46549335330472, "learning_rate": 1.9237344946818614e-07, "loss": 0.3899, "step": 18041 }, { "epoch": 2.82, "grad_norm": 21.593249390398487, "learning_rate": 1.9204439705003964e-07, "loss": 0.4879, "step": 18042 }, { "epoch": 2.82, "grad_norm": 21.53584317932088, "learning_rate": 1.9171562356441353e-07, "loss": 0.4065, "step": 18043 }, { "epoch": 2.82, "grad_norm": 18.074040117900097, "learning_rate": 1.9138712902065703e-07, "loss": 0.4057, "step": 18044 }, { "epoch": 2.82, "grad_norm": 23.156110949105237, "learning_rate": 1.9105891342811268e-07, "loss": 0.4093, "step": 18045 }, { "epoch": 2.82, "grad_norm": 17.38387574962378, "learning_rate": 1.9073097679611518e-07, "loss": 0.4512, "step": 18046 }, { "epoch": 2.82, "grad_norm": 23.113693351611204, "learning_rate": 1.9040331913398934e-07, "loss": 0.4384, "step": 18047 }, { "epoch": 2.82, "grad_norm": 19.82495614169634, "learning_rate": 1.900759404510577e-07, "loss": 0.4256, "step": 18048 }, { "epoch": 2.82, "grad_norm": 20.005652331541793, "learning_rate": 1.8974884075662503e-07, "loss": 0.4273, "step": 18049 }, { "epoch": 2.82, "grad_norm": 13.767564499450945, "learning_rate": 1.894220200599961e-07, "loss": 0.4957, "step": 18050 }, { "epoch": 2.82, "grad_norm": 24.785842358146194, "learning_rate": 1.8909547837046683e-07, "loss": 0.3231, "step": 18051 }, { "epoch": 2.82, "grad_norm": 19.597745308570083, "learning_rate": 1.887692156973231e-07, "loss": 0.457, "step": 18052 }, { "epoch": 2.82, "grad_norm": 32.00777855411114, "learning_rate": 1.8844323204984417e-07, "loss": 0.4225, "step": 18053 }, { "epoch": 2.82, "grad_norm": 26.357248174116904, "learning_rate": 1.8811752743729817e-07, "loss": 0.4255, "step": 18054 }, { "epoch": 2.82, "grad_norm": 24.549965623221635, "learning_rate": 1.8779210186895102e-07, "loss": 0.4693, "step": 18055 }, { "epoch": 2.82, "grad_norm": 22.033894527946728, "learning_rate": 1.8746695535405645e-07, "loss": 0.4267, "step": 18056 }, { "epoch": 2.82, "grad_norm": 26.834830938454907, "learning_rate": 1.871420879018615e-07, "loss": 0.4254, "step": 18057 }, { "epoch": 2.82, "grad_norm": 20.04780078453937, "learning_rate": 1.8681749952160544e-07, "loss": 0.4107, "step": 18058 }, { "epoch": 2.82, "grad_norm": 20.944762309132198, "learning_rate": 1.8649319022251865e-07, "loss": 0.4351, "step": 18059 }, { "epoch": 2.82, "grad_norm": 27.39598569404578, "learning_rate": 1.8616916001382602e-07, "loss": 0.4731, "step": 18060 }, { "epoch": 2.82, "grad_norm": 18.46669486757505, "learning_rate": 1.8584540890474012e-07, "loss": 0.329, "step": 18061 }, { "epoch": 2.82, "grad_norm": 25.708401359306325, "learning_rate": 1.855219369044703e-07, "loss": 0.4504, "step": 18062 }, { "epoch": 2.82, "grad_norm": 28.80060555401762, "learning_rate": 1.8519874402221472e-07, "loss": 0.4368, "step": 18063 }, { "epoch": 2.82, "grad_norm": 18.55700747070408, "learning_rate": 1.8487583026716604e-07, "loss": 0.4529, "step": 18064 }, { "epoch": 2.82, "grad_norm": 21.17969897434173, "learning_rate": 1.8455319564850694e-07, "loss": 0.3771, "step": 18065 }, { "epoch": 2.82, "grad_norm": 15.298909719272224, "learning_rate": 1.8423084017541227e-07, "loss": 0.4362, "step": 18066 }, { "epoch": 2.82, "grad_norm": 25.20105157117016, "learning_rate": 1.8390876385705136e-07, "loss": 0.4054, "step": 18067 }, { "epoch": 2.82, "grad_norm": 25.69374549427703, "learning_rate": 1.8358696670258357e-07, "loss": 0.4537, "step": 18068 }, { "epoch": 2.82, "grad_norm": 18.454268990957402, "learning_rate": 1.8326544872115936e-07, "loss": 0.4051, "step": 18069 }, { "epoch": 2.82, "grad_norm": 28.01332041744621, "learning_rate": 1.829442099219225e-07, "loss": 0.466, "step": 18070 }, { "epoch": 2.82, "grad_norm": 28.520798793850297, "learning_rate": 1.826232503140102e-07, "loss": 0.4633, "step": 18071 }, { "epoch": 2.82, "grad_norm": 19.453714394736885, "learning_rate": 1.823025699065495e-07, "loss": 0.4518, "step": 18072 }, { "epoch": 2.82, "grad_norm": 22.71209743885157, "learning_rate": 1.8198216870866093e-07, "loss": 0.4082, "step": 18073 }, { "epoch": 2.82, "grad_norm": 24.074791542603535, "learning_rate": 1.8166204672945496e-07, "loss": 0.5501, "step": 18074 }, { "epoch": 2.82, "grad_norm": 24.001064209887243, "learning_rate": 1.813422039780377e-07, "loss": 0.3711, "step": 18075 }, { "epoch": 2.82, "grad_norm": 22.77712935005126, "learning_rate": 1.8102264046350404e-07, "loss": 0.4231, "step": 18076 }, { "epoch": 2.82, "grad_norm": 28.90221294152296, "learning_rate": 1.8070335619494227e-07, "loss": 0.4831, "step": 18077 }, { "epoch": 2.82, "grad_norm": 24.483302189049436, "learning_rate": 1.8038435118143405e-07, "loss": 0.4072, "step": 18078 }, { "epoch": 2.82, "grad_norm": 18.329837099582967, "learning_rate": 1.8006562543204986e-07, "loss": 0.4118, "step": 18079 }, { "epoch": 2.82, "grad_norm": 25.900038995032247, "learning_rate": 1.797471789558547e-07, "loss": 0.4752, "step": 18080 }, { "epoch": 2.82, "grad_norm": 21.903649596175914, "learning_rate": 1.7942901176190574e-07, "loss": 0.4263, "step": 18081 }, { "epoch": 2.82, "grad_norm": 27.36413278391528, "learning_rate": 1.7911112385925022e-07, "loss": 0.4231, "step": 18082 }, { "epoch": 2.82, "grad_norm": 15.371269739348032, "learning_rate": 1.7879351525692978e-07, "loss": 0.4034, "step": 18083 }, { "epoch": 2.82, "grad_norm": 34.24500229798534, "learning_rate": 1.7847618596397498e-07, "loss": 0.4887, "step": 18084 }, { "epoch": 2.82, "grad_norm": 18.28627134059287, "learning_rate": 1.7815913598941303e-07, "loss": 0.4218, "step": 18085 }, { "epoch": 2.83, "grad_norm": 21.181616275644974, "learning_rate": 1.7784236534226006e-07, "loss": 0.4187, "step": 18086 }, { "epoch": 2.83, "grad_norm": 20.870334229454777, "learning_rate": 1.7752587403152444e-07, "loss": 0.3943, "step": 18087 }, { "epoch": 2.83, "grad_norm": 15.01649455262894, "learning_rate": 1.7720966206620672e-07, "loss": 0.41, "step": 18088 }, { "epoch": 2.83, "grad_norm": 20.797464520004223, "learning_rate": 1.768937294552997e-07, "loss": 0.4171, "step": 18089 }, { "epoch": 2.83, "grad_norm": 18.43060913708382, "learning_rate": 1.7657807620778956e-07, "loss": 0.4399, "step": 18090 }, { "epoch": 2.83, "grad_norm": 19.210114085611043, "learning_rate": 1.762627023326513e-07, "loss": 0.4687, "step": 18091 }, { "epoch": 2.83, "grad_norm": 28.032104896981732, "learning_rate": 1.7594760783885557e-07, "loss": 0.4251, "step": 18092 }, { "epoch": 2.83, "grad_norm": 28.886526744480392, "learning_rate": 1.7563279273536294e-07, "loss": 0.4783, "step": 18093 }, { "epoch": 2.83, "grad_norm": 18.713493722860782, "learning_rate": 1.7531825703112736e-07, "loss": 0.3652, "step": 18094 }, { "epoch": 2.83, "grad_norm": 25.35630160621016, "learning_rate": 1.7500400073509172e-07, "loss": 0.3439, "step": 18095 }, { "epoch": 2.83, "grad_norm": 46.848033351011836, "learning_rate": 1.7469002385619548e-07, "loss": 0.4755, "step": 18096 }, { "epoch": 2.83, "grad_norm": 27.098689946154227, "learning_rate": 1.7437632640336822e-07, "loss": 0.475, "step": 18097 }, { "epoch": 2.83, "grad_norm": 15.826411482122195, "learning_rate": 1.7406290838552943e-07, "loss": 0.4656, "step": 18098 }, { "epoch": 2.83, "grad_norm": 21.272662359847125, "learning_rate": 1.737497698115942e-07, "loss": 0.4917, "step": 18099 }, { "epoch": 2.83, "grad_norm": 23.69031228781328, "learning_rate": 1.7343691069046654e-07, "loss": 0.4148, "step": 18100 }, { "epoch": 2.83, "grad_norm": 22.14307886994061, "learning_rate": 1.7312433103104486e-07, "loss": 0.4642, "step": 18101 }, { "epoch": 2.83, "grad_norm": 15.091150067789329, "learning_rate": 1.7281203084221876e-07, "loss": 0.4693, "step": 18102 }, { "epoch": 2.83, "grad_norm": 24.46290276015051, "learning_rate": 1.725000101328711e-07, "loss": 0.3895, "step": 18103 }, { "epoch": 2.83, "grad_norm": 18.00665190591452, "learning_rate": 1.7218826891187258e-07, "loss": 0.3895, "step": 18104 }, { "epoch": 2.83, "grad_norm": 21.228924223964846, "learning_rate": 1.7187680718809053e-07, "loss": 0.4256, "step": 18105 }, { "epoch": 2.83, "grad_norm": 26.48815626534871, "learning_rate": 1.7156562497038342e-07, "loss": 0.4698, "step": 18106 }, { "epoch": 2.83, "grad_norm": 17.5080570814964, "learning_rate": 1.7125472226760087e-07, "loss": 0.3958, "step": 18107 }, { "epoch": 2.83, "grad_norm": 21.23897477723079, "learning_rate": 1.7094409908858357e-07, "loss": 0.4444, "step": 18108 }, { "epoch": 2.83, "grad_norm": 25.72896190740729, "learning_rate": 1.7063375544216666e-07, "loss": 0.4255, "step": 18109 }, { "epoch": 2.83, "grad_norm": 23.006760234765697, "learning_rate": 1.703236913371753e-07, "loss": 0.43, "step": 18110 }, { "epoch": 2.83, "grad_norm": 25.394987076413656, "learning_rate": 1.7001390678242803e-07, "loss": 0.4356, "step": 18111 }, { "epoch": 2.83, "grad_norm": 19.898448192981473, "learning_rate": 1.6970440178673552e-07, "loss": 0.4079, "step": 18112 }, { "epoch": 2.83, "grad_norm": 15.822246959766153, "learning_rate": 1.6939517635889857e-07, "loss": 0.3617, "step": 18113 }, { "epoch": 2.83, "grad_norm": 15.644290886983896, "learning_rate": 1.6908623050771234e-07, "loss": 0.4395, "step": 18114 }, { "epoch": 2.83, "grad_norm": 25.721337680432228, "learning_rate": 1.6877756424196312e-07, "loss": 0.4423, "step": 18115 }, { "epoch": 2.83, "grad_norm": 29.035468263154407, "learning_rate": 1.6846917757042836e-07, "loss": 0.4062, "step": 18116 }, { "epoch": 2.83, "grad_norm": 25.028969375069227, "learning_rate": 1.6816107050187992e-07, "loss": 0.4754, "step": 18117 }, { "epoch": 2.83, "grad_norm": 22.97966827731043, "learning_rate": 1.6785324304507745e-07, "loss": 0.4466, "step": 18118 }, { "epoch": 2.83, "grad_norm": 17.103186958032786, "learning_rate": 1.675456952087795e-07, "loss": 0.3774, "step": 18119 }, { "epoch": 2.83, "grad_norm": 28.04438139793631, "learning_rate": 1.67238427001728e-07, "loss": 0.4336, "step": 18120 }, { "epoch": 2.83, "grad_norm": 17.868874365569596, "learning_rate": 1.669314384326648e-07, "loss": 0.4362, "step": 18121 }, { "epoch": 2.83, "grad_norm": 20.185261840727215, "learning_rate": 1.666247295103185e-07, "loss": 0.4189, "step": 18122 }, { "epoch": 2.83, "grad_norm": 22.618388579897058, "learning_rate": 1.6631830024341323e-07, "loss": 0.4924, "step": 18123 }, { "epoch": 2.83, "grad_norm": 36.765768636932776, "learning_rate": 1.66012150640662e-07, "loss": 0.4616, "step": 18124 }, { "epoch": 2.83, "grad_norm": 14.82754277894135, "learning_rate": 1.6570628071077343e-07, "loss": 0.3961, "step": 18125 }, { "epoch": 2.83, "grad_norm": 20.195369655447088, "learning_rate": 1.65400690462445e-07, "loss": 0.4697, "step": 18126 }, { "epoch": 2.83, "grad_norm": 20.645068167612955, "learning_rate": 1.6509537990436753e-07, "loss": 0.4393, "step": 18127 }, { "epoch": 2.83, "grad_norm": 19.288813188068776, "learning_rate": 1.6479034904522518e-07, "loss": 0.4724, "step": 18128 }, { "epoch": 2.83, "grad_norm": 15.54076714896541, "learning_rate": 1.6448559789369102e-07, "loss": 0.4038, "step": 18129 }, { "epoch": 2.83, "grad_norm": 18.671157344290176, "learning_rate": 1.6418112645843253e-07, "loss": 0.3625, "step": 18130 }, { "epoch": 2.83, "grad_norm": 15.851430332127304, "learning_rate": 1.638769347481095e-07, "loss": 0.3647, "step": 18131 }, { "epoch": 2.83, "grad_norm": 20.464256768237657, "learning_rate": 1.6357302277137388e-07, "loss": 0.4647, "step": 18132 }, { "epoch": 2.83, "grad_norm": 25.426634853449738, "learning_rate": 1.632693905368654e-07, "loss": 0.4356, "step": 18133 }, { "epoch": 2.83, "grad_norm": 28.152442060722105, "learning_rate": 1.6296603805322163e-07, "loss": 0.4448, "step": 18134 }, { "epoch": 2.83, "grad_norm": 18.292625044420976, "learning_rate": 1.62662965329069e-07, "loss": 0.4173, "step": 18135 }, { "epoch": 2.83, "grad_norm": 16.925419858819957, "learning_rate": 1.6236017237302838e-07, "loss": 0.4472, "step": 18136 }, { "epoch": 2.83, "grad_norm": 16.481163865588723, "learning_rate": 1.6205765919370843e-07, "loss": 0.4739, "step": 18137 }, { "epoch": 2.83, "grad_norm": 16.99978433938535, "learning_rate": 1.6175542579971336e-07, "loss": 0.4241, "step": 18138 }, { "epoch": 2.83, "grad_norm": 17.73395309030394, "learning_rate": 1.6145347219963857e-07, "loss": 0.362, "step": 18139 }, { "epoch": 2.83, "grad_norm": 22.04112731607826, "learning_rate": 1.6115179840207274e-07, "loss": 0.4428, "step": 18140 }, { "epoch": 2.83, "grad_norm": 20.14290410309478, "learning_rate": 1.608504044155934e-07, "loss": 0.4003, "step": 18141 }, { "epoch": 2.83, "grad_norm": 16.596535828683223, "learning_rate": 1.6054929024877375e-07, "loss": 0.3913, "step": 18142 }, { "epoch": 2.83, "grad_norm": 27.040931804578047, "learning_rate": 1.6024845591017468e-07, "loss": 0.4133, "step": 18143 }, { "epoch": 2.83, "grad_norm": 29.86997905191291, "learning_rate": 1.599479014083549e-07, "loss": 0.3703, "step": 18144 }, { "epoch": 2.83, "grad_norm": 23.98163823595663, "learning_rate": 1.5964762675185984e-07, "loss": 0.4183, "step": 18145 }, { "epoch": 2.83, "grad_norm": 34.122474402284546, "learning_rate": 1.5934763194923043e-07, "loss": 0.4737, "step": 18146 }, { "epoch": 2.83, "grad_norm": 16.175138456593668, "learning_rate": 1.5904791700899758e-07, "loss": 0.4222, "step": 18147 }, { "epoch": 2.83, "grad_norm": 18.617639431288524, "learning_rate": 1.587484819396856e-07, "loss": 0.3616, "step": 18148 }, { "epoch": 2.83, "grad_norm": 21.52473055491459, "learning_rate": 1.5844932674980885e-07, "loss": 0.4336, "step": 18149 }, { "epoch": 2.84, "grad_norm": 30.533524061043977, "learning_rate": 1.58150451447876e-07, "loss": 0.3729, "step": 18150 }, { "epoch": 2.84, "grad_norm": 22.754961925460908, "learning_rate": 1.5785185604238805e-07, "loss": 0.4209, "step": 18151 }, { "epoch": 2.84, "grad_norm": 21.098454730627093, "learning_rate": 1.5755354054183491e-07, "loss": 0.3867, "step": 18152 }, { "epoch": 2.84, "grad_norm": 29.90687654450791, "learning_rate": 1.57255504954702e-07, "loss": 0.4125, "step": 18153 }, { "epoch": 2.84, "grad_norm": 18.95612312828468, "learning_rate": 1.5695774928946473e-07, "loss": 0.461, "step": 18154 }, { "epoch": 2.84, "grad_norm": 22.89656441354108, "learning_rate": 1.5666027355458967e-07, "loss": 0.417, "step": 18155 }, { "epoch": 2.84, "grad_norm": 25.8180123070717, "learning_rate": 1.563630777585401e-07, "loss": 0.3954, "step": 18156 }, { "epoch": 2.84, "grad_norm": 31.728573400762432, "learning_rate": 1.5606616190976587e-07, "loss": 0.3966, "step": 18157 }, { "epoch": 2.84, "grad_norm": 23.59168916790271, "learning_rate": 1.5576952601671135e-07, "loss": 0.4001, "step": 18158 }, { "epoch": 2.84, "grad_norm": 20.335210898568214, "learning_rate": 1.5547317008781205e-07, "loss": 0.4492, "step": 18159 }, { "epoch": 2.84, "grad_norm": 32.75339082643325, "learning_rate": 1.5517709413149674e-07, "loss": 0.4202, "step": 18160 }, { "epoch": 2.84, "grad_norm": 34.61698646202934, "learning_rate": 1.5488129815618757e-07, "loss": 0.4497, "step": 18161 }, { "epoch": 2.84, "grad_norm": 16.712190585082542, "learning_rate": 1.5458578217029342e-07, "loss": 0.4043, "step": 18162 }, { "epoch": 2.84, "grad_norm": 26.23191589203769, "learning_rate": 1.5429054618222083e-07, "loss": 0.4396, "step": 18163 }, { "epoch": 2.84, "grad_norm": 26.617654282955858, "learning_rate": 1.5399559020036537e-07, "loss": 0.4442, "step": 18164 }, { "epoch": 2.84, "grad_norm": 24.92596551157176, "learning_rate": 1.5370091423311584e-07, "loss": 0.4267, "step": 18165 }, { "epoch": 2.84, "grad_norm": 20.889899187026224, "learning_rate": 1.5340651828885223e-07, "loss": 0.4176, "step": 18166 }, { "epoch": 2.84, "grad_norm": 21.74141980706441, "learning_rate": 1.531124023759467e-07, "loss": 0.4263, "step": 18167 }, { "epoch": 2.84, "grad_norm": 18.603786962738102, "learning_rate": 1.5281856650276482e-07, "loss": 0.4044, "step": 18168 }, { "epoch": 2.84, "grad_norm": 26.053305121127813, "learning_rate": 1.5252501067766324e-07, "loss": 0.4292, "step": 18169 }, { "epoch": 2.84, "grad_norm": 13.951804333518846, "learning_rate": 1.5223173490898856e-07, "loss": 0.4112, "step": 18170 }, { "epoch": 2.84, "grad_norm": 35.875249590985746, "learning_rate": 1.5193873920508306e-07, "loss": 0.5211, "step": 18171 }, { "epoch": 2.84, "grad_norm": 15.136251961678564, "learning_rate": 1.5164602357427781e-07, "loss": 0.4034, "step": 18172 }, { "epoch": 2.84, "grad_norm": 19.237586526090553, "learning_rate": 1.5135358802489952e-07, "loss": 0.4031, "step": 18173 }, { "epoch": 2.84, "grad_norm": 20.087554054982746, "learning_rate": 1.5106143256526374e-07, "loss": 0.424, "step": 18174 }, { "epoch": 2.84, "grad_norm": 31.098079034513844, "learning_rate": 1.5076955720367937e-07, "loss": 0.4257, "step": 18175 }, { "epoch": 2.84, "grad_norm": 26.387130861713654, "learning_rate": 1.5047796194844754e-07, "loss": 0.4193, "step": 18176 }, { "epoch": 2.84, "grad_norm": 17.34765911209643, "learning_rate": 1.5018664680785943e-07, "loss": 0.4429, "step": 18177 }, { "epoch": 2.84, "grad_norm": 27.522948187908025, "learning_rate": 1.4989561179020283e-07, "loss": 0.5117, "step": 18178 }, { "epoch": 2.84, "grad_norm": 23.37303788000141, "learning_rate": 1.4960485690375115e-07, "loss": 0.4123, "step": 18179 }, { "epoch": 2.84, "grad_norm": 46.58080469463978, "learning_rate": 1.4931438215677663e-07, "loss": 0.4591, "step": 18180 }, { "epoch": 2.84, "grad_norm": 24.04519566527334, "learning_rate": 1.490241875575371e-07, "loss": 0.398, "step": 18181 }, { "epoch": 2.84, "grad_norm": 20.34164301483965, "learning_rate": 1.4873427311428822e-07, "loss": 0.3969, "step": 18182 }, { "epoch": 2.84, "grad_norm": 15.675626770634809, "learning_rate": 1.484446388352734e-07, "loss": 0.4051, "step": 18183 }, { "epoch": 2.84, "grad_norm": 17.31448853000177, "learning_rate": 1.481552847287293e-07, "loss": 0.4227, "step": 18184 }, { "epoch": 2.84, "grad_norm": 21.04627522730495, "learning_rate": 1.478662108028872e-07, "loss": 0.4129, "step": 18185 }, { "epoch": 2.84, "grad_norm": 33.439381932838494, "learning_rate": 1.4757741706596605e-07, "loss": 0.4465, "step": 18186 }, { "epoch": 2.84, "grad_norm": 28.171230519307425, "learning_rate": 1.472889035261793e-07, "loss": 0.435, "step": 18187 }, { "epoch": 2.84, "grad_norm": 18.740528923769265, "learning_rate": 1.4700067019173258e-07, "loss": 0.3903, "step": 18188 }, { "epoch": 2.84, "grad_norm": 16.00891777270265, "learning_rate": 1.4671271707082268e-07, "loss": 0.452, "step": 18189 }, { "epoch": 2.84, "grad_norm": 43.32758584806517, "learning_rate": 1.4642504417163973e-07, "loss": 0.4603, "step": 18190 }, { "epoch": 2.84, "grad_norm": 25.222767669989338, "learning_rate": 1.4613765150236382e-07, "loss": 0.4583, "step": 18191 }, { "epoch": 2.84, "grad_norm": 44.02331776166877, "learning_rate": 1.4585053907116953e-07, "loss": 0.4935, "step": 18192 }, { "epoch": 2.84, "grad_norm": 22.404369818038525, "learning_rate": 1.4556370688622036e-07, "loss": 0.3643, "step": 18193 }, { "epoch": 2.84, "grad_norm": 34.20547161922587, "learning_rate": 1.4527715495567529e-07, "loss": 0.4229, "step": 18194 }, { "epoch": 2.84, "grad_norm": 18.17256618589119, "learning_rate": 1.4499088328768341e-07, "loss": 0.391, "step": 18195 }, { "epoch": 2.84, "grad_norm": 23.934183261958513, "learning_rate": 1.4470489189038595e-07, "loss": 0.5173, "step": 18196 }, { "epoch": 2.84, "grad_norm": 21.611196711468622, "learning_rate": 1.444191807719153e-07, "loss": 0.3712, "step": 18197 }, { "epoch": 2.84, "grad_norm": 19.25601794255633, "learning_rate": 1.4413374994039718e-07, "loss": 0.4396, "step": 18198 }, { "epoch": 2.84, "grad_norm": 25.81377415364099, "learning_rate": 1.438485994039518e-07, "loss": 0.4755, "step": 18199 }, { "epoch": 2.84, "grad_norm": 21.55632520525946, "learning_rate": 1.4356372917068595e-07, "loss": 0.4468, "step": 18200 }, { "epoch": 2.84, "grad_norm": 19.516011784785015, "learning_rate": 1.4327913924870097e-07, "loss": 0.4856, "step": 18201 }, { "epoch": 2.84, "grad_norm": 20.55362997955764, "learning_rate": 1.4299482964609035e-07, "loss": 0.4636, "step": 18202 }, { "epoch": 2.84, "grad_norm": 25.437518246337195, "learning_rate": 1.427108003709421e-07, "loss": 0.4259, "step": 18203 }, { "epoch": 2.84, "grad_norm": 28.093469669801973, "learning_rate": 1.4242705143133195e-07, "loss": 0.4142, "step": 18204 }, { "epoch": 2.84, "grad_norm": 30.588926237977518, "learning_rate": 1.4214358283533124e-07, "loss": 0.3983, "step": 18205 }, { "epoch": 2.84, "grad_norm": 18.730443331926935, "learning_rate": 1.41860394590998e-07, "loss": 0.4127, "step": 18206 }, { "epoch": 2.84, "grad_norm": 32.90276969082344, "learning_rate": 1.4157748670639015e-07, "loss": 0.4189, "step": 18207 }, { "epoch": 2.84, "grad_norm": 24.323852056015433, "learning_rate": 1.412948591895502e-07, "loss": 0.4323, "step": 18208 }, { "epoch": 2.84, "grad_norm": 22.991900443859063, "learning_rate": 1.4101251204851841e-07, "loss": 0.4128, "step": 18209 }, { "epoch": 2.84, "grad_norm": 18.864021626420843, "learning_rate": 1.4073044529132273e-07, "loss": 0.4514, "step": 18210 }, { "epoch": 2.84, "grad_norm": 24.75328245576418, "learning_rate": 1.4044865892598681e-07, "loss": 0.4128, "step": 18211 }, { "epoch": 2.84, "grad_norm": 27.72153820456351, "learning_rate": 1.4016715296052198e-07, "loss": 0.4743, "step": 18212 }, { "epoch": 2.84, "grad_norm": 18.796791619506735, "learning_rate": 1.3988592740293629e-07, "loss": 0.4505, "step": 18213 }, { "epoch": 2.85, "grad_norm": 22.097491542734378, "learning_rate": 1.3960498226122554e-07, "loss": 0.3998, "step": 18214 }, { "epoch": 2.85, "grad_norm": 31.23495737136862, "learning_rate": 1.3932431754338227e-07, "loss": 0.4552, "step": 18215 }, { "epoch": 2.85, "grad_norm": 21.836687720006733, "learning_rate": 1.3904393325738785e-07, "loss": 0.5238, "step": 18216 }, { "epoch": 2.85, "grad_norm": 28.404939685266307, "learning_rate": 1.387638294112137e-07, "loss": 0.4581, "step": 18217 }, { "epoch": 2.85, "grad_norm": 22.696002639152205, "learning_rate": 1.3848400601282784e-07, "loss": 0.396, "step": 18218 }, { "epoch": 2.85, "grad_norm": 15.952713981624163, "learning_rate": 1.3820446307018843e-07, "loss": 0.3807, "step": 18219 }, { "epoch": 2.85, "grad_norm": 27.57750819329659, "learning_rate": 1.379252005912457e-07, "loss": 0.4514, "step": 18220 }, { "epoch": 2.85, "grad_norm": 24.08277278108093, "learning_rate": 1.3764621858394e-07, "loss": 0.444, "step": 18221 }, { "epoch": 2.85, "grad_norm": 28.012148032345465, "learning_rate": 1.373675170562072e-07, "loss": 0.4614, "step": 18222 }, { "epoch": 2.85, "grad_norm": 29.257655526568993, "learning_rate": 1.370890960159721e-07, "loss": 0.405, "step": 18223 }, { "epoch": 2.85, "grad_norm": 14.971149642466557, "learning_rate": 1.3681095547115497e-07, "loss": 0.4054, "step": 18224 }, { "epoch": 2.85, "grad_norm": 19.95551134139376, "learning_rate": 1.3653309542966286e-07, "loss": 0.4338, "step": 18225 }, { "epoch": 2.85, "grad_norm": 22.394114875216385, "learning_rate": 1.3625551589940056e-07, "loss": 0.4221, "step": 18226 }, { "epoch": 2.85, "grad_norm": 17.77124156926482, "learning_rate": 1.3597821688826063e-07, "loss": 0.4406, "step": 18227 }, { "epoch": 2.85, "grad_norm": 26.27295173470462, "learning_rate": 1.357011984041301e-07, "loss": 0.3948, "step": 18228 }, { "epoch": 2.85, "grad_norm": 19.550298932935227, "learning_rate": 1.3542446045488712e-07, "loss": 0.4221, "step": 18229 }, { "epoch": 2.85, "grad_norm": 24.261717487237515, "learning_rate": 1.3514800304840203e-07, "loss": 0.4513, "step": 18230 }, { "epoch": 2.85, "grad_norm": 21.378416580288324, "learning_rate": 1.3487182619253636e-07, "loss": 0.4522, "step": 18231 }, { "epoch": 2.85, "grad_norm": 18.04620413916343, "learning_rate": 1.3459592989514604e-07, "loss": 0.3865, "step": 18232 }, { "epoch": 2.85, "grad_norm": 24.92382540828731, "learning_rate": 1.3432031416407476e-07, "loss": 0.4419, "step": 18233 }, { "epoch": 2.85, "grad_norm": 25.065576182623328, "learning_rate": 1.3404497900716406e-07, "loss": 0.4844, "step": 18234 }, { "epoch": 2.85, "grad_norm": 31.668818941299143, "learning_rate": 1.3376992443224213e-07, "loss": 0.4946, "step": 18235 }, { "epoch": 2.85, "grad_norm": 16.565309548950104, "learning_rate": 1.3349515044713157e-07, "loss": 0.4246, "step": 18236 }, { "epoch": 2.85, "grad_norm": 22.682159837032373, "learning_rate": 1.3322065705964838e-07, "loss": 0.4108, "step": 18237 }, { "epoch": 2.85, "grad_norm": 21.122144044725097, "learning_rate": 1.3294644427759628e-07, "loss": 0.4337, "step": 18238 }, { "epoch": 2.85, "grad_norm": 23.58910743561029, "learning_rate": 1.3267251210877686e-07, "loss": 0.4263, "step": 18239 }, { "epoch": 2.85, "grad_norm": 16.794766502225077, "learning_rate": 1.3239886056097718e-07, "loss": 0.3911, "step": 18240 }, { "epoch": 2.85, "grad_norm": 20.03918227444374, "learning_rate": 1.3212548964198324e-07, "loss": 0.3791, "step": 18241 }, { "epoch": 2.85, "grad_norm": 26.254945583740252, "learning_rate": 1.3185239935956662e-07, "loss": 0.4239, "step": 18242 }, { "epoch": 2.85, "grad_norm": 17.113176742126637, "learning_rate": 1.3157958972149553e-07, "loss": 0.3723, "step": 18243 }, { "epoch": 2.85, "grad_norm": 25.40501717418585, "learning_rate": 1.3130706073552824e-07, "loss": 0.4449, "step": 18244 }, { "epoch": 2.85, "grad_norm": 30.65019077825835, "learning_rate": 1.310348124094152e-07, "loss": 0.4439, "step": 18245 }, { "epoch": 2.85, "grad_norm": 24.28012598840872, "learning_rate": 1.3076284475089906e-07, "loss": 0.4163, "step": 18246 }, { "epoch": 2.85, "grad_norm": 24.461155361982932, "learning_rate": 1.304911577677137e-07, "loss": 0.4483, "step": 18247 }, { "epoch": 2.85, "grad_norm": 24.86781911680022, "learning_rate": 1.3021975146758514e-07, "loss": 0.4491, "step": 18248 }, { "epoch": 2.85, "grad_norm": 21.215532946643265, "learning_rate": 1.2994862585823608e-07, "loss": 0.4408, "step": 18249 }, { "epoch": 2.85, "grad_norm": 27.474143368909964, "learning_rate": 1.296777809473726e-07, "loss": 0.4995, "step": 18250 }, { "epoch": 2.85, "grad_norm": 16.86276445619728, "learning_rate": 1.294072167426985e-07, "loss": 0.4199, "step": 18251 }, { "epoch": 2.85, "grad_norm": 32.63909716063981, "learning_rate": 1.291369332519099e-07, "loss": 0.4275, "step": 18252 }, { "epoch": 2.85, "grad_norm": 32.4958815936282, "learning_rate": 1.2886693048269284e-07, "loss": 0.4332, "step": 18253 }, { "epoch": 2.85, "grad_norm": 29.899361495708675, "learning_rate": 1.285972084427256e-07, "loss": 0.4823, "step": 18254 }, { "epoch": 2.85, "grad_norm": 20.185397513270228, "learning_rate": 1.2832776713967876e-07, "loss": 0.3765, "step": 18255 }, { "epoch": 2.85, "grad_norm": 23.96969018911251, "learning_rate": 1.2805860658121505e-07, "loss": 0.4326, "step": 18256 }, { "epoch": 2.85, "grad_norm": 20.157667397778972, "learning_rate": 1.2778972677499057e-07, "loss": 0.4084, "step": 18257 }, { "epoch": 2.85, "grad_norm": 26.398294822711463, "learning_rate": 1.2752112772865144e-07, "loss": 0.4793, "step": 18258 }, { "epoch": 2.85, "grad_norm": 17.819168584249628, "learning_rate": 1.27252809449836e-07, "loss": 0.4424, "step": 18259 }, { "epoch": 2.85, "grad_norm": 34.710535464812274, "learning_rate": 1.2698477194617475e-07, "loss": 0.4617, "step": 18260 }, { "epoch": 2.85, "grad_norm": 18.050015686515742, "learning_rate": 1.2671701522529055e-07, "loss": 0.3938, "step": 18261 }, { "epoch": 2.85, "grad_norm": 23.491837843033025, "learning_rate": 1.2644953929479954e-07, "loss": 0.4553, "step": 18262 }, { "epoch": 2.85, "grad_norm": 32.02693140364885, "learning_rate": 1.2618234416230778e-07, "loss": 0.4523, "step": 18263 }, { "epoch": 2.85, "grad_norm": 17.53436103032286, "learning_rate": 1.2591542983541371e-07, "loss": 0.4631, "step": 18264 }, { "epoch": 2.85, "grad_norm": 15.125403133844381, "learning_rate": 1.25648796321709e-07, "loss": 0.4357, "step": 18265 }, { "epoch": 2.85, "grad_norm": 30.400829232860094, "learning_rate": 1.253824436287765e-07, "loss": 0.4937, "step": 18266 }, { "epoch": 2.85, "grad_norm": 27.894364971225666, "learning_rate": 1.2511637176419012e-07, "loss": 0.4206, "step": 18267 }, { "epoch": 2.85, "grad_norm": 35.16153862190187, "learning_rate": 1.2485058073551825e-07, "loss": 0.4851, "step": 18268 }, { "epoch": 2.85, "grad_norm": 19.18343405947422, "learning_rate": 1.2458507055031822e-07, "loss": 0.4282, "step": 18269 }, { "epoch": 2.85, "grad_norm": 15.941899986257978, "learning_rate": 1.2431984121614282e-07, "loss": 0.4293, "step": 18270 }, { "epoch": 2.85, "grad_norm": 19.170185557965027, "learning_rate": 1.2405489274053273e-07, "loss": 0.4209, "step": 18271 }, { "epoch": 2.85, "grad_norm": 18.26559713335416, "learning_rate": 1.2379022513102524e-07, "loss": 0.4748, "step": 18272 }, { "epoch": 2.85, "grad_norm": 13.58268633357884, "learning_rate": 1.2352583839514655e-07, "loss": 0.3817, "step": 18273 }, { "epoch": 2.85, "grad_norm": 31.294829727891404, "learning_rate": 1.2326173254041506e-07, "loss": 0.4544, "step": 18274 }, { "epoch": 2.85, "grad_norm": 19.57193461527387, "learning_rate": 1.229979075743415e-07, "loss": 0.4164, "step": 18275 }, { "epoch": 2.85, "grad_norm": 27.49040521870134, "learning_rate": 1.227343635044298e-07, "loss": 0.4649, "step": 18276 }, { "epoch": 2.85, "grad_norm": 30.295074280515394, "learning_rate": 1.224711003381751e-07, "loss": 0.4698, "step": 18277 }, { "epoch": 2.86, "grad_norm": 22.15709087191234, "learning_rate": 1.2220811808306477e-07, "loss": 0.4272, "step": 18278 }, { "epoch": 2.86, "grad_norm": 19.525911835352204, "learning_rate": 1.219454167465761e-07, "loss": 0.4171, "step": 18279 }, { "epoch": 2.86, "grad_norm": 22.250762091936767, "learning_rate": 1.2168299633618209e-07, "loss": 0.4116, "step": 18280 }, { "epoch": 2.86, "grad_norm": 20.535413923673268, "learning_rate": 1.2142085685934446e-07, "loss": 0.446, "step": 18281 }, { "epoch": 2.86, "grad_norm": 16.227910429830516, "learning_rate": 1.2115899832351952e-07, "loss": 0.4088, "step": 18282 }, { "epoch": 2.86, "grad_norm": 20.945430437984616, "learning_rate": 1.2089742073615352e-07, "loss": 0.4462, "step": 18283 }, { "epoch": 2.86, "grad_norm": 19.46003956677633, "learning_rate": 1.2063612410468496e-07, "loss": 0.4724, "step": 18284 }, { "epoch": 2.86, "grad_norm": 13.531547658743047, "learning_rate": 1.2037510843654565e-07, "loss": 0.3948, "step": 18285 }, { "epoch": 2.86, "grad_norm": 21.28698440908482, "learning_rate": 1.201143737391597e-07, "loss": 0.4691, "step": 18286 }, { "epoch": 2.86, "grad_norm": 17.219199387627988, "learning_rate": 1.1985392001994112e-07, "loss": 0.424, "step": 18287 }, { "epoch": 2.86, "grad_norm": 17.407352152367334, "learning_rate": 1.1959374728629847e-07, "loss": 0.4237, "step": 18288 }, { "epoch": 2.86, "grad_norm": 28.14379353294335, "learning_rate": 1.1933385554562803e-07, "loss": 0.4131, "step": 18289 }, { "epoch": 2.86, "grad_norm": 19.762368673543325, "learning_rate": 1.1907424480532282e-07, "loss": 0.4638, "step": 18290 }, { "epoch": 2.86, "grad_norm": 22.293099082992246, "learning_rate": 1.1881491507276688e-07, "loss": 0.4623, "step": 18291 }, { "epoch": 2.86, "grad_norm": 27.96405503489841, "learning_rate": 1.1855586635533434e-07, "loss": 0.4386, "step": 18292 }, { "epoch": 2.86, "grad_norm": 20.452285267790614, "learning_rate": 1.1829709866039263e-07, "loss": 0.4069, "step": 18293 }, { "epoch": 2.86, "grad_norm": 16.581660493226238, "learning_rate": 1.180386119952992e-07, "loss": 0.3631, "step": 18294 }, { "epoch": 2.86, "grad_norm": 21.5751306664675, "learning_rate": 1.1778040636740928e-07, "loss": 0.3806, "step": 18295 }, { "epoch": 2.86, "grad_norm": 21.562729975952855, "learning_rate": 1.1752248178406144e-07, "loss": 0.4119, "step": 18296 }, { "epoch": 2.86, "grad_norm": 22.991797096186204, "learning_rate": 1.1726483825259427e-07, "loss": 0.4657, "step": 18297 }, { "epoch": 2.86, "grad_norm": 23.926652440305812, "learning_rate": 1.1700747578033412e-07, "loss": 0.4166, "step": 18298 }, { "epoch": 2.86, "grad_norm": 19.856143375593142, "learning_rate": 1.1675039437459956e-07, "loss": 0.4117, "step": 18299 }, { "epoch": 2.86, "grad_norm": 25.064544546774627, "learning_rate": 1.1649359404270144e-07, "loss": 0.5282, "step": 18300 }, { "epoch": 2.86, "grad_norm": 20.1012338751548, "learning_rate": 1.1623707479194501e-07, "loss": 0.4263, "step": 18301 }, { "epoch": 2.86, "grad_norm": 19.484279506911268, "learning_rate": 1.1598083662962335e-07, "loss": 0.4213, "step": 18302 }, { "epoch": 2.86, "grad_norm": 11.43540317695954, "learning_rate": 1.1572487956302613e-07, "loss": 0.4002, "step": 18303 }, { "epoch": 2.86, "grad_norm": 18.724048546019766, "learning_rate": 1.1546920359943092e-07, "loss": 0.4233, "step": 18304 }, { "epoch": 2.86, "grad_norm": 27.13283798311038, "learning_rate": 1.1521380874610854e-07, "loss": 0.5084, "step": 18305 }, { "epoch": 2.86, "grad_norm": 17.356874741693716, "learning_rate": 1.1495869501032209e-07, "loss": 0.4074, "step": 18306 }, { "epoch": 2.86, "grad_norm": 30.088977923583656, "learning_rate": 1.1470386239932907e-07, "loss": 0.4498, "step": 18307 }, { "epoch": 2.86, "grad_norm": 17.527666461746104, "learning_rate": 1.1444931092037592e-07, "loss": 0.4372, "step": 18308 }, { "epoch": 2.86, "grad_norm": 23.48198879070703, "learning_rate": 1.141950405807013e-07, "loss": 0.4184, "step": 18309 }, { "epoch": 2.86, "grad_norm": 22.287432520406412, "learning_rate": 1.1394105138753498e-07, "loss": 0.394, "step": 18310 }, { "epoch": 2.86, "grad_norm": 23.894798388320066, "learning_rate": 1.1368734334810339e-07, "loss": 0.4483, "step": 18311 }, { "epoch": 2.86, "grad_norm": 19.3576798947693, "learning_rate": 1.1343391646962076e-07, "loss": 0.4501, "step": 18312 }, { "epoch": 2.86, "grad_norm": 27.284140015930348, "learning_rate": 1.1318077075929357e-07, "loss": 0.437, "step": 18313 }, { "epoch": 2.86, "grad_norm": 19.4743317261089, "learning_rate": 1.129279062243216e-07, "loss": 0.4076, "step": 18314 }, { "epoch": 2.86, "grad_norm": 26.494946311682842, "learning_rate": 1.1267532287189576e-07, "loss": 0.4364, "step": 18315 }, { "epoch": 2.86, "grad_norm": 21.778848902942965, "learning_rate": 1.1242302070920141e-07, "loss": 0.3701, "step": 18316 }, { "epoch": 2.86, "grad_norm": 32.33168403832644, "learning_rate": 1.1217099974341173e-07, "loss": 0.5231, "step": 18317 }, { "epoch": 2.86, "grad_norm": 20.193657443112794, "learning_rate": 1.1191925998169317e-07, "loss": 0.4437, "step": 18318 }, { "epoch": 2.86, "grad_norm": 16.553740629145125, "learning_rate": 1.1166780143120781e-07, "loss": 0.376, "step": 18319 }, { "epoch": 2.86, "grad_norm": 17.83741249538844, "learning_rate": 1.1141662409910547e-07, "loss": 0.3954, "step": 18320 }, { "epoch": 2.86, "grad_norm": 19.31835728310645, "learning_rate": 1.1116572799252934e-07, "loss": 0.4073, "step": 18321 }, { "epoch": 2.86, "grad_norm": 27.507013174424475, "learning_rate": 1.109151131186148e-07, "loss": 0.4684, "step": 18322 }, { "epoch": 2.86, "grad_norm": 18.420351265117326, "learning_rate": 1.106647794844895e-07, "loss": 0.446, "step": 18323 }, { "epoch": 2.86, "grad_norm": 30.284290627564843, "learning_rate": 1.104147270972733e-07, "loss": 0.4706, "step": 18324 }, { "epoch": 2.86, "grad_norm": 21.72598521107956, "learning_rate": 1.1016495596407717e-07, "loss": 0.3406, "step": 18325 }, { "epoch": 2.86, "grad_norm": 30.89734352383938, "learning_rate": 1.0991546609200432e-07, "loss": 0.4189, "step": 18326 }, { "epoch": 2.86, "grad_norm": 32.40921275179439, "learning_rate": 1.0966625748815019e-07, "loss": 0.4185, "step": 18327 }, { "epoch": 2.86, "grad_norm": 25.249523932033956, "learning_rate": 1.0941733015960132e-07, "loss": 0.3936, "step": 18328 }, { "epoch": 2.86, "grad_norm": 27.365700176310437, "learning_rate": 1.0916868411343872e-07, "loss": 0.41, "step": 18329 }, { "epoch": 2.86, "grad_norm": 21.20717487894361, "learning_rate": 1.089203193567312e-07, "loss": 0.3419, "step": 18330 }, { "epoch": 2.86, "grad_norm": 18.1012528787738, "learning_rate": 1.0867223589654418e-07, "loss": 0.4043, "step": 18331 }, { "epoch": 2.86, "grad_norm": 20.8934833337288, "learning_rate": 1.0842443373993427e-07, "loss": 0.4408, "step": 18332 }, { "epoch": 2.86, "grad_norm": 15.506846019694269, "learning_rate": 1.0817691289394583e-07, "loss": 0.4368, "step": 18333 }, { "epoch": 2.86, "grad_norm": 25.07982944626364, "learning_rate": 1.0792967336561877e-07, "loss": 0.4592, "step": 18334 }, { "epoch": 2.86, "grad_norm": 15.647404359982936, "learning_rate": 1.0768271516198525e-07, "loss": 0.3941, "step": 18335 }, { "epoch": 2.86, "grad_norm": 21.869458228862086, "learning_rate": 1.0743603829006854e-07, "loss": 0.4166, "step": 18336 }, { "epoch": 2.86, "grad_norm": 22.32477108863656, "learning_rate": 1.0718964275688526e-07, "loss": 0.3814, "step": 18337 }, { "epoch": 2.86, "grad_norm": 23.72555269787382, "learning_rate": 1.069435285694409e-07, "loss": 0.4282, "step": 18338 }, { "epoch": 2.86, "grad_norm": 27.716896747793303, "learning_rate": 1.0669769573473432e-07, "loss": 0.4296, "step": 18339 }, { "epoch": 2.86, "grad_norm": 20.428220041937145, "learning_rate": 1.064521442597577e-07, "loss": 0.4553, "step": 18340 }, { "epoch": 2.86, "grad_norm": 16.108314938081648, "learning_rate": 1.0620687415149655e-07, "loss": 0.4194, "step": 18341 }, { "epoch": 2.87, "grad_norm": 34.529464409693254, "learning_rate": 1.0596188541692309e-07, "loss": 0.471, "step": 18342 }, { "epoch": 2.87, "grad_norm": 20.951167182996517, "learning_rate": 1.0571717806300508e-07, "loss": 0.4234, "step": 18343 }, { "epoch": 2.87, "grad_norm": 20.66355459530119, "learning_rate": 1.0547275209670249e-07, "loss": 0.3926, "step": 18344 }, { "epoch": 2.87, "grad_norm": 30.87482004653751, "learning_rate": 1.0522860752496755e-07, "loss": 0.4628, "step": 18345 }, { "epoch": 2.87, "grad_norm": 28.40349049640971, "learning_rate": 1.0498474435474249e-07, "loss": 0.413, "step": 18346 }, { "epoch": 2.87, "grad_norm": 18.8221611035089, "learning_rate": 1.0474116259296285e-07, "loss": 0.3641, "step": 18347 }, { "epoch": 2.87, "grad_norm": 17.732197671568045, "learning_rate": 1.0449786224655644e-07, "loss": 0.4381, "step": 18348 }, { "epoch": 2.87, "grad_norm": 20.640845109168804, "learning_rate": 1.0425484332244107e-07, "loss": 0.4326, "step": 18349 }, { "epoch": 2.87, "grad_norm": 14.5540106365669, "learning_rate": 1.0401210582753007e-07, "loss": 0.3778, "step": 18350 }, { "epoch": 2.87, "grad_norm": 18.943376669544964, "learning_rate": 1.0376964976872461e-07, "loss": 0.4012, "step": 18351 }, { "epoch": 2.87, "grad_norm": 20.3970843567236, "learning_rate": 1.035274751529225e-07, "loss": 0.4325, "step": 18352 }, { "epoch": 2.87, "grad_norm": 16.53049528656263, "learning_rate": 1.0328558198700933e-07, "loss": 0.4194, "step": 18353 }, { "epoch": 2.87, "grad_norm": 28.183961027611318, "learning_rate": 1.0304397027786516e-07, "loss": 0.3995, "step": 18354 }, { "epoch": 2.87, "grad_norm": 32.494961688259316, "learning_rate": 1.0280264003236006e-07, "loss": 0.4576, "step": 18355 }, { "epoch": 2.87, "grad_norm": 24.538597885877973, "learning_rate": 1.0256159125735742e-07, "loss": 0.4198, "step": 18356 }, { "epoch": 2.87, "grad_norm": 16.949595634207135, "learning_rate": 1.023208239597151e-07, "loss": 0.4104, "step": 18357 }, { "epoch": 2.87, "grad_norm": 16.961369053442983, "learning_rate": 1.0208033814627872e-07, "loss": 0.4271, "step": 18358 }, { "epoch": 2.87, "grad_norm": 24.07025856054119, "learning_rate": 1.0184013382388613e-07, "loss": 0.4391, "step": 18359 }, { "epoch": 2.87, "grad_norm": 28.14149565993997, "learning_rate": 1.0160021099936967e-07, "loss": 0.4248, "step": 18360 }, { "epoch": 2.87, "grad_norm": 27.49864984687522, "learning_rate": 1.0136056967955388e-07, "loss": 0.4734, "step": 18361 }, { "epoch": 2.87, "grad_norm": 20.68384317890662, "learning_rate": 1.0112120987125329e-07, "loss": 0.4223, "step": 18362 }, { "epoch": 2.87, "grad_norm": 24.888227326931965, "learning_rate": 1.0088213158127358e-07, "loss": 0.4164, "step": 18363 }, { "epoch": 2.87, "grad_norm": 21.193933531366707, "learning_rate": 1.0064333481641597e-07, "loss": 0.4623, "step": 18364 }, { "epoch": 2.87, "grad_norm": 24.71110605223992, "learning_rate": 1.0040481958346948e-07, "loss": 0.4058, "step": 18365 }, { "epoch": 2.87, "grad_norm": 25.67617787563051, "learning_rate": 1.0016658588922091e-07, "loss": 0.4615, "step": 18366 }, { "epoch": 2.87, "grad_norm": 26.667018233654044, "learning_rate": 9.99286337404426e-08, "loss": 0.4211, "step": 18367 }, { "epoch": 2.87, "grad_norm": 15.43700196258174, "learning_rate": 9.969096314390248e-08, "loss": 0.4526, "step": 18368 }, { "epoch": 2.87, "grad_norm": 25.116893618274897, "learning_rate": 9.945357410635959e-08, "loss": 0.478, "step": 18369 }, { "epoch": 2.87, "grad_norm": 17.175931624933508, "learning_rate": 9.92164666345663e-08, "loss": 0.4395, "step": 18370 }, { "epoch": 2.87, "grad_norm": 22.993850233103725, "learning_rate": 9.8979640735265e-08, "loss": 0.4018, "step": 18371 }, { "epoch": 2.87, "grad_norm": 18.71850141414339, "learning_rate": 9.87430964151892e-08, "loss": 0.4075, "step": 18372 }, { "epoch": 2.87, "grad_norm": 19.911529531824726, "learning_rate": 9.850683368106794e-08, "loss": 0.3908, "step": 18373 }, { "epoch": 2.87, "grad_norm": 17.664318398786268, "learning_rate": 9.827085253962144e-08, "loss": 0.3525, "step": 18374 }, { "epoch": 2.87, "grad_norm": 20.89809913931761, "learning_rate": 9.803515299755872e-08, "loss": 0.3946, "step": 18375 }, { "epoch": 2.87, "grad_norm": 25.24531524817793, "learning_rate": 9.779973506158446e-08, "loss": 0.4048, "step": 18376 }, { "epoch": 2.87, "grad_norm": 20.046205461876834, "learning_rate": 9.75645987383922e-08, "loss": 0.4428, "step": 18377 }, { "epoch": 2.87, "grad_norm": 23.8457090734149, "learning_rate": 9.732974403466988e-08, "loss": 0.4265, "step": 18378 }, { "epoch": 2.87, "grad_norm": 16.375499224514808, "learning_rate": 9.709517095709775e-08, "loss": 0.3892, "step": 18379 }, { "epoch": 2.87, "grad_norm": 30.318810043126096, "learning_rate": 9.68608795123449e-08, "loss": 0.4579, "step": 18380 }, { "epoch": 2.87, "grad_norm": 23.540822286499488, "learning_rate": 9.662686970707602e-08, "loss": 0.4104, "step": 18381 }, { "epoch": 2.87, "grad_norm": 21.78546797700525, "learning_rate": 9.639314154794576e-08, "loss": 0.45, "step": 18382 }, { "epoch": 2.87, "grad_norm": 29.403820297245723, "learning_rate": 9.615969504160106e-08, "loss": 0.4565, "step": 18383 }, { "epoch": 2.87, "grad_norm": 24.36859784779545, "learning_rate": 9.5926530194681e-08, "loss": 0.4095, "step": 18384 }, { "epoch": 2.87, "grad_norm": 27.500862742025546, "learning_rate": 9.569364701381589e-08, "loss": 0.425, "step": 18385 }, { "epoch": 2.87, "grad_norm": 18.634175997254733, "learning_rate": 9.546104550563152e-08, "loss": 0.3868, "step": 18386 }, { "epoch": 2.87, "grad_norm": 28.924624283014616, "learning_rate": 9.522872567673924e-08, "loss": 0.4375, "step": 18387 }, { "epoch": 2.87, "grad_norm": 25.80431594257756, "learning_rate": 9.499668753374936e-08, "loss": 0.4815, "step": 18388 }, { "epoch": 2.87, "grad_norm": 21.989342069424193, "learning_rate": 9.476493108325768e-08, "loss": 0.48, "step": 18389 }, { "epoch": 2.87, "grad_norm": 17.54848776409242, "learning_rate": 9.453345633185784e-08, "loss": 0.4332, "step": 18390 }, { "epoch": 2.87, "grad_norm": 28.07947207060383, "learning_rate": 9.430226328613234e-08, "loss": 0.4644, "step": 18391 }, { "epoch": 2.87, "grad_norm": 21.896055242361882, "learning_rate": 9.407135195265593e-08, "loss": 0.4584, "step": 18392 }, { "epoch": 2.87, "grad_norm": 21.43880150489254, "learning_rate": 9.384072233799447e-08, "loss": 0.44, "step": 18393 }, { "epoch": 2.87, "grad_norm": 38.68869021191799, "learning_rate": 9.361037444870824e-08, "loss": 0.3873, "step": 18394 }, { "epoch": 2.87, "grad_norm": 21.748485184951647, "learning_rate": 9.33803082913487e-08, "loss": 0.4395, "step": 18395 }, { "epoch": 2.87, "grad_norm": 22.035705717796795, "learning_rate": 9.315052387245726e-08, "loss": 0.4298, "step": 18396 }, { "epoch": 2.87, "grad_norm": 16.570742163152755, "learning_rate": 9.292102119856983e-08, "loss": 0.4361, "step": 18397 }, { "epoch": 2.87, "grad_norm": 21.56563215726859, "learning_rate": 9.269180027621228e-08, "loss": 0.4201, "step": 18398 }, { "epoch": 2.87, "grad_norm": 18.876684476042943, "learning_rate": 9.246286111190495e-08, "loss": 0.4364, "step": 18399 }, { "epoch": 2.87, "grad_norm": 22.931858661805176, "learning_rate": 9.223420371215818e-08, "loss": 0.3939, "step": 18400 }, { "epoch": 2.87, "grad_norm": 24.19178338097972, "learning_rate": 9.200582808347569e-08, "loss": 0.3766, "step": 18401 }, { "epoch": 2.87, "grad_norm": 26.112399496229294, "learning_rate": 9.177773423235004e-08, "loss": 0.4233, "step": 18402 }, { "epoch": 2.87, "grad_norm": 24.344523069302543, "learning_rate": 9.154992216526937e-08, "loss": 0.3983, "step": 18403 }, { "epoch": 2.87, "grad_norm": 35.93790694628309, "learning_rate": 9.132239188871405e-08, "loss": 0.431, "step": 18404 }, { "epoch": 2.87, "grad_norm": 20.423853945781914, "learning_rate": 9.109514340915226e-08, "loss": 0.4403, "step": 18405 }, { "epoch": 2.88, "grad_norm": 32.61197658236122, "learning_rate": 9.086817673304882e-08, "loss": 0.4037, "step": 18406 }, { "epoch": 2.88, "grad_norm": 22.570937209101274, "learning_rate": 9.064149186685744e-08, "loss": 0.4733, "step": 18407 }, { "epoch": 2.88, "grad_norm": 19.297242912460888, "learning_rate": 9.041508881702521e-08, "loss": 0.3966, "step": 18408 }, { "epoch": 2.88, "grad_norm": 20.985048350014203, "learning_rate": 9.018896758999029e-08, "loss": 0.4334, "step": 18409 }, { "epoch": 2.88, "grad_norm": 22.705161815201944, "learning_rate": 8.996312819218533e-08, "loss": 0.4409, "step": 18410 }, { "epoch": 2.88, "grad_norm": 17.741281472827755, "learning_rate": 8.973757063003075e-08, "loss": 0.371, "step": 18411 }, { "epoch": 2.88, "grad_norm": 14.623725437191284, "learning_rate": 8.95122949099425e-08, "loss": 0.4228, "step": 18412 }, { "epoch": 2.88, "grad_norm": 21.906957812172738, "learning_rate": 8.928730103832773e-08, "loss": 0.4596, "step": 18413 }, { "epoch": 2.88, "grad_norm": 17.49742682987387, "learning_rate": 8.90625890215846e-08, "loss": 0.381, "step": 18414 }, { "epoch": 2.88, "grad_norm": 39.01680808513573, "learning_rate": 8.883815886610358e-08, "loss": 0.496, "step": 18415 }, { "epoch": 2.88, "grad_norm": 22.809953867555414, "learning_rate": 8.861401057826624e-08, "loss": 0.4104, "step": 18416 }, { "epoch": 2.88, "grad_norm": 27.70306184334395, "learning_rate": 8.839014416444969e-08, "loss": 0.4811, "step": 18417 }, { "epoch": 2.88, "grad_norm": 19.58237640362796, "learning_rate": 8.816655963101883e-08, "loss": 0.4681, "step": 18418 }, { "epoch": 2.88, "grad_norm": 28.904941402404614, "learning_rate": 8.794325698433193e-08, "loss": 0.4524, "step": 18419 }, { "epoch": 2.88, "grad_norm": 18.799944844158663, "learning_rate": 8.772023623074166e-08, "loss": 0.4036, "step": 18420 }, { "epoch": 2.88, "grad_norm": 26.88421682233436, "learning_rate": 8.749749737658964e-08, "loss": 0.4637, "step": 18421 }, { "epoch": 2.88, "grad_norm": 34.75290987983933, "learning_rate": 8.727504042820966e-08, "loss": 0.4365, "step": 18422 }, { "epoch": 2.88, "grad_norm": 25.53116679907588, "learning_rate": 8.705286539192781e-08, "loss": 0.4023, "step": 18423 }, { "epoch": 2.88, "grad_norm": 20.29214367331964, "learning_rate": 8.683097227406456e-08, "loss": 0.4595, "step": 18424 }, { "epoch": 2.88, "grad_norm": 24.424526102433298, "learning_rate": 8.660936108093043e-08, "loss": 0.4449, "step": 18425 }, { "epoch": 2.88, "grad_norm": 21.800541174967424, "learning_rate": 8.638803181882483e-08, "loss": 0.4619, "step": 18426 }, { "epoch": 2.88, "grad_norm": 18.868383997538675, "learning_rate": 8.616698449404492e-08, "loss": 0.4536, "step": 18427 }, { "epoch": 2.88, "grad_norm": 27.182119321572173, "learning_rate": 8.594621911287682e-08, "loss": 0.4578, "step": 18428 }, { "epoch": 2.88, "grad_norm": 23.349822477599126, "learning_rate": 8.572573568159881e-08, "loss": 0.4155, "step": 18429 }, { "epoch": 2.88, "grad_norm": 16.235199743034283, "learning_rate": 8.550553420648145e-08, "loss": 0.426, "step": 18430 }, { "epoch": 2.88, "grad_norm": 29.155371849240236, "learning_rate": 8.528561469378637e-08, "loss": 0.4482, "step": 18431 }, { "epoch": 2.88, "grad_norm": 20.000468394635895, "learning_rate": 8.50659771497686e-08, "loss": 0.4205, "step": 18432 }, { "epoch": 2.88, "grad_norm": 27.272111640289037, "learning_rate": 8.484662158067425e-08, "loss": 0.5392, "step": 18433 }, { "epoch": 2.88, "grad_norm": 18.60949134149706, "learning_rate": 8.462754799274275e-08, "loss": 0.3962, "step": 18434 }, { "epoch": 2.88, "grad_norm": 24.73631820588152, "learning_rate": 8.440875639220247e-08, "loss": 0.3951, "step": 18435 }, { "epoch": 2.88, "grad_norm": 32.32679850814688, "learning_rate": 8.41902467852762e-08, "loss": 0.4657, "step": 18436 }, { "epoch": 2.88, "grad_norm": 20.361122131769324, "learning_rate": 8.397201917817899e-08, "loss": 0.4396, "step": 18437 }, { "epoch": 2.88, "grad_norm": 16.761292744870147, "learning_rate": 8.375407357711806e-08, "loss": 0.4351, "step": 18438 }, { "epoch": 2.88, "grad_norm": 24.202548007847533, "learning_rate": 8.35364099882896e-08, "loss": 0.4313, "step": 18439 }, { "epoch": 2.88, "grad_norm": 20.581560612045358, "learning_rate": 8.331902841788531e-08, "loss": 0.3866, "step": 18440 }, { "epoch": 2.88, "grad_norm": 25.526838452592482, "learning_rate": 8.310192887208579e-08, "loss": 0.4333, "step": 18441 }, { "epoch": 2.88, "grad_norm": 20.345680104239268, "learning_rate": 8.288511135706611e-08, "loss": 0.4141, "step": 18442 }, { "epoch": 2.88, "grad_norm": 23.12467925970497, "learning_rate": 8.266857587899357e-08, "loss": 0.4855, "step": 18443 }, { "epoch": 2.88, "grad_norm": 23.38043819880168, "learning_rate": 8.245232244402435e-08, "loss": 0.4931, "step": 18444 }, { "epoch": 2.88, "grad_norm": 26.510071820912223, "learning_rate": 8.223635105831018e-08, "loss": 0.4557, "step": 18445 }, { "epoch": 2.88, "grad_norm": 30.774422568064423, "learning_rate": 8.202066172799172e-08, "loss": 0.4404, "step": 18446 }, { "epoch": 2.88, "grad_norm": 20.156260721810025, "learning_rate": 8.180525445920407e-08, "loss": 0.4852, "step": 18447 }, { "epoch": 2.88, "grad_norm": 21.6878259796467, "learning_rate": 8.159012925807341e-08, "loss": 0.3993, "step": 18448 }, { "epoch": 2.88, "grad_norm": 39.4023906835687, "learning_rate": 8.1375286130716e-08, "loss": 0.4522, "step": 18449 }, { "epoch": 2.88, "grad_norm": 24.398988439646708, "learning_rate": 8.116072508324579e-08, "loss": 0.3749, "step": 18450 }, { "epoch": 2.88, "grad_norm": 35.94074062165958, "learning_rate": 8.094644612176017e-08, "loss": 0.4712, "step": 18451 }, { "epoch": 2.88, "grad_norm": 24.645613179361575, "learning_rate": 8.073244925235535e-08, "loss": 0.4866, "step": 18452 }, { "epoch": 2.88, "grad_norm": 24.49746228232012, "learning_rate": 8.051873448111758e-08, "loss": 0.4786, "step": 18453 }, { "epoch": 2.88, "grad_norm": 19.18045609404505, "learning_rate": 8.030530181412422e-08, "loss": 0.4345, "step": 18454 }, { "epoch": 2.88, "grad_norm": 16.963552300654136, "learning_rate": 8.009215125744486e-08, "loss": 0.4205, "step": 18455 }, { "epoch": 2.88, "grad_norm": 23.741096115634896, "learning_rate": 7.987928281714241e-08, "loss": 0.4457, "step": 18456 }, { "epoch": 2.88, "grad_norm": 17.735347494693396, "learning_rate": 7.966669649926872e-08, "loss": 0.4812, "step": 18457 }, { "epoch": 2.88, "grad_norm": 20.551751869945793, "learning_rate": 7.945439230987228e-08, "loss": 0.3879, "step": 18458 }, { "epoch": 2.88, "grad_norm": 15.915078012297402, "learning_rate": 7.924237025498938e-08, "loss": 0.3645, "step": 18459 }, { "epoch": 2.88, "grad_norm": 26.16597690965238, "learning_rate": 7.903063034064962e-08, "loss": 0.4434, "step": 18460 }, { "epoch": 2.88, "grad_norm": 22.936425450228953, "learning_rate": 7.881917257287485e-08, "loss": 0.4915, "step": 18461 }, { "epoch": 2.88, "grad_norm": 19.1757377045868, "learning_rate": 7.860799695767918e-08, "loss": 0.4435, "step": 18462 }, { "epoch": 2.88, "grad_norm": 21.778394975136973, "learning_rate": 7.839710350106888e-08, "loss": 0.4262, "step": 18463 }, { "epoch": 2.88, "grad_norm": 15.386251898154999, "learning_rate": 7.81864922090414e-08, "loss": 0.3686, "step": 18464 }, { "epoch": 2.88, "grad_norm": 16.32013498595952, "learning_rate": 7.797616308758526e-08, "loss": 0.4027, "step": 18465 }, { "epoch": 2.88, "grad_norm": 16.44788166374201, "learning_rate": 7.776611614268236e-08, "loss": 0.4164, "step": 18466 }, { "epoch": 2.88, "grad_norm": 18.394630355713797, "learning_rate": 7.755635138030682e-08, "loss": 0.4065, "step": 18467 }, { "epoch": 2.88, "grad_norm": 37.130097902971805, "learning_rate": 7.734686880642495e-08, "loss": 0.548, "step": 18468 }, { "epoch": 2.88, "grad_norm": 26.742833770222312, "learning_rate": 7.71376684269931e-08, "loss": 0.3763, "step": 18469 }, { "epoch": 2.89, "grad_norm": 30.812196427333543, "learning_rate": 7.692875024796099e-08, "loss": 0.4334, "step": 18470 }, { "epoch": 2.89, "grad_norm": 29.26335271271613, "learning_rate": 7.672011427527159e-08, "loss": 0.4692, "step": 18471 }, { "epoch": 2.89, "grad_norm": 16.540203146146805, "learning_rate": 7.651176051485576e-08, "loss": 0.4594, "step": 18472 }, { "epoch": 2.89, "grad_norm": 33.6817246831706, "learning_rate": 7.630368897264096e-08, "loss": 0.4307, "step": 18473 }, { "epoch": 2.89, "grad_norm": 21.05279806573189, "learning_rate": 7.609589965454356e-08, "loss": 0.3759, "step": 18474 }, { "epoch": 2.89, "grad_norm": 17.47577836004555, "learning_rate": 7.58883925664744e-08, "loss": 0.4349, "step": 18475 }, { "epoch": 2.89, "grad_norm": 17.057102090107257, "learning_rate": 7.56811677143332e-08, "loss": 0.4606, "step": 18476 }, { "epoch": 2.89, "grad_norm": 21.720885014053273, "learning_rate": 7.547422510401414e-08, "loss": 0.4472, "step": 18477 }, { "epoch": 2.89, "grad_norm": 22.19976653725924, "learning_rate": 7.526756474140252e-08, "loss": 0.4428, "step": 18478 }, { "epoch": 2.89, "grad_norm": 33.325196305159665, "learning_rate": 7.506118663237583e-08, "loss": 0.4944, "step": 18479 }, { "epoch": 2.89, "grad_norm": 23.749362944541655, "learning_rate": 7.485509078280384e-08, "loss": 0.4245, "step": 18480 }, { "epoch": 2.89, "grad_norm": 16.145792908361546, "learning_rate": 7.46492771985452e-08, "loss": 0.3999, "step": 18481 }, { "epoch": 2.89, "grad_norm": 18.13890644840671, "learning_rate": 7.444374588545522e-08, "loss": 0.4817, "step": 18482 }, { "epoch": 2.89, "grad_norm": 22.025304870086725, "learning_rate": 7.423849684938033e-08, "loss": 0.4045, "step": 18483 }, { "epoch": 2.89, "grad_norm": 30.917424138972326, "learning_rate": 7.403353009615588e-08, "loss": 0.4678, "step": 18484 }, { "epoch": 2.89, "grad_norm": 15.667072842114589, "learning_rate": 7.382884563161052e-08, "loss": 0.4348, "step": 18485 }, { "epoch": 2.89, "grad_norm": 19.697520489006024, "learning_rate": 7.362444346156517e-08, "loss": 0.4321, "step": 18486 }, { "epoch": 2.89, "grad_norm": 22.577317795299162, "learning_rate": 7.342032359183515e-08, "loss": 0.4574, "step": 18487 }, { "epoch": 2.89, "grad_norm": 25.125105402141795, "learning_rate": 7.321648602822473e-08, "loss": 0.4092, "step": 18488 }, { "epoch": 2.89, "grad_norm": 26.371183509091896, "learning_rate": 7.301293077653038e-08, "loss": 0.3988, "step": 18489 }, { "epoch": 2.89, "grad_norm": 19.46116850182559, "learning_rate": 7.280965784253968e-08, "loss": 0.4658, "step": 18490 }, { "epoch": 2.89, "grad_norm": 22.9599841667247, "learning_rate": 7.26066672320358e-08, "loss": 0.4495, "step": 18491 }, { "epoch": 2.89, "grad_norm": 19.241037532658318, "learning_rate": 7.240395895079189e-08, "loss": 0.4071, "step": 18492 }, { "epoch": 2.89, "grad_norm": 17.04750890978004, "learning_rate": 7.22015330045711e-08, "loss": 0.38, "step": 18493 }, { "epoch": 2.89, "grad_norm": 15.986960141430107, "learning_rate": 7.199938939913109e-08, "loss": 0.4363, "step": 18494 }, { "epoch": 2.89, "grad_norm": 14.181144282223094, "learning_rate": 7.179752814022056e-08, "loss": 0.399, "step": 18495 }, { "epoch": 2.89, "grad_norm": 15.774148673717654, "learning_rate": 7.159594923358159e-08, "loss": 0.3851, "step": 18496 }, { "epoch": 2.89, "grad_norm": 23.196637559690757, "learning_rate": 7.139465268494405e-08, "loss": 0.4522, "step": 18497 }, { "epoch": 2.89, "grad_norm": 22.357463917609813, "learning_rate": 7.119363850003558e-08, "loss": 0.4343, "step": 18498 }, { "epoch": 2.89, "grad_norm": 18.695675280088306, "learning_rate": 7.09929066845727e-08, "loss": 0.3993, "step": 18499 }, { "epoch": 2.89, "grad_norm": 17.779029934285226, "learning_rate": 7.079245724426198e-08, "loss": 0.4216, "step": 18500 }, { "epoch": 2.89, "grad_norm": 23.167377265631004, "learning_rate": 7.059229018480551e-08, "loss": 0.4308, "step": 18501 }, { "epoch": 2.89, "grad_norm": 30.58525938254278, "learning_rate": 7.039240551189653e-08, "loss": 0.4907, "step": 18502 }, { "epoch": 2.89, "grad_norm": 31.406738420740492, "learning_rate": 7.019280323121714e-08, "loss": 0.484, "step": 18503 }, { "epoch": 2.89, "grad_norm": 24.438240513838203, "learning_rate": 6.999348334844724e-08, "loss": 0.4921, "step": 18504 }, { "epoch": 2.89, "grad_norm": 24.756662620740393, "learning_rate": 6.979444586925455e-08, "loss": 0.4635, "step": 18505 }, { "epoch": 2.89, "grad_norm": 18.424191867938323, "learning_rate": 6.959569079929674e-08, "loss": 0.45, "step": 18506 }, { "epoch": 2.89, "grad_norm": 18.418451151997708, "learning_rate": 6.939721814422928e-08, "loss": 0.3599, "step": 18507 }, { "epoch": 2.89, "grad_norm": 26.611592926667754, "learning_rate": 6.919902790969657e-08, "loss": 0.5605, "step": 18508 }, { "epoch": 2.89, "grad_norm": 27.175031345401116, "learning_rate": 6.900112010133297e-08, "loss": 0.4164, "step": 18509 }, { "epoch": 2.89, "grad_norm": 23.911385013371248, "learning_rate": 6.880349472476844e-08, "loss": 0.409, "step": 18510 }, { "epoch": 2.89, "grad_norm": 25.36275611826147, "learning_rate": 6.86061517856229e-08, "loss": 0.4153, "step": 18511 }, { "epoch": 2.89, "grad_norm": 24.641146904686117, "learning_rate": 6.840909128950857e-08, "loss": 0.4182, "step": 18512 }, { "epoch": 2.89, "grad_norm": 23.634173210946592, "learning_rate": 6.821231324203093e-08, "loss": 0.4154, "step": 18513 }, { "epoch": 2.89, "grad_norm": 18.822013163736386, "learning_rate": 6.801581764878329e-08, "loss": 0.4476, "step": 18514 }, { "epoch": 2.89, "grad_norm": 15.616144023306047, "learning_rate": 6.781960451535674e-08, "loss": 0.3498, "step": 18515 }, { "epoch": 2.89, "grad_norm": 19.768925716139016, "learning_rate": 6.762367384733015e-08, "loss": 0.43, "step": 18516 }, { "epoch": 2.89, "grad_norm": 18.002053399482897, "learning_rate": 6.742802565027684e-08, "loss": 0.4196, "step": 18517 }, { "epoch": 2.89, "grad_norm": 27.32091563766875, "learning_rate": 6.723265992975903e-08, "loss": 0.4213, "step": 18518 }, { "epoch": 2.89, "grad_norm": 25.195549789612453, "learning_rate": 6.703757669133448e-08, "loss": 0.4309, "step": 18519 }, { "epoch": 2.89, "grad_norm": 16.58093273626427, "learning_rate": 6.6842775940551e-08, "loss": 0.4127, "step": 18520 }, { "epoch": 2.89, "grad_norm": 25.544681947236025, "learning_rate": 6.664825768294747e-08, "loss": 0.4339, "step": 18521 }, { "epoch": 2.89, "grad_norm": 21.074768979841554, "learning_rate": 6.645402192405726e-08, "loss": 0.4174, "step": 18522 }, { "epoch": 2.89, "grad_norm": 30.15982142829394, "learning_rate": 6.626006866940371e-08, "loss": 0.4876, "step": 18523 }, { "epoch": 2.89, "grad_norm": 22.21421075210531, "learning_rate": 6.606639792450131e-08, "loss": 0.427, "step": 18524 }, { "epoch": 2.89, "grad_norm": 19.43151432509714, "learning_rate": 6.587300969486122e-08, "loss": 0.4401, "step": 18525 }, { "epoch": 2.89, "grad_norm": 23.738934036978918, "learning_rate": 6.567990398598012e-08, "loss": 0.4326, "step": 18526 }, { "epoch": 2.89, "grad_norm": 28.927823425160025, "learning_rate": 6.548708080335143e-08, "loss": 0.415, "step": 18527 }, { "epoch": 2.89, "grad_norm": 23.358991507504868, "learning_rate": 6.529454015245962e-08, "loss": 0.4195, "step": 18528 }, { "epoch": 2.89, "grad_norm": 26.89922264475566, "learning_rate": 6.510228203877811e-08, "loss": 0.4366, "step": 18529 }, { "epoch": 2.89, "grad_norm": 29.8925049106383, "learning_rate": 6.491030646777696e-08, "loss": 0.4165, "step": 18530 }, { "epoch": 2.89, "grad_norm": 18.89408575877051, "learning_rate": 6.471861344491514e-08, "loss": 0.4147, "step": 18531 }, { "epoch": 2.89, "grad_norm": 14.746844744221828, "learning_rate": 6.452720297564275e-08, "loss": 0.4141, "step": 18532 }, { "epoch": 2.89, "grad_norm": 17.393911793127998, "learning_rate": 6.43360750654054e-08, "loss": 0.4115, "step": 18533 }, { "epoch": 2.9, "grad_norm": 24.086794263540646, "learning_rate": 6.414522971963766e-08, "loss": 0.4074, "step": 18534 }, { "epoch": 2.9, "grad_norm": 33.061442889204315, "learning_rate": 6.395466694376739e-08, "loss": 0.417, "step": 18535 }, { "epoch": 2.9, "grad_norm": 24.032101190098853, "learning_rate": 6.376438674321361e-08, "loss": 0.4324, "step": 18536 }, { "epoch": 2.9, "grad_norm": 18.81695936901427, "learning_rate": 6.357438912338864e-08, "loss": 0.4226, "step": 18537 }, { "epoch": 2.9, "grad_norm": 28.27477320593617, "learning_rate": 6.338467408969484e-08, "loss": 0.4412, "step": 18538 }, { "epoch": 2.9, "grad_norm": 23.402800398745548, "learning_rate": 6.319524164752899e-08, "loss": 0.4642, "step": 18539 }, { "epoch": 2.9, "grad_norm": 27.7534454951741, "learning_rate": 6.300609180227679e-08, "loss": 0.4372, "step": 18540 }, { "epoch": 2.9, "grad_norm": 22.486044208739152, "learning_rate": 6.281722455931839e-08, "loss": 0.3834, "step": 18541 }, { "epoch": 2.9, "grad_norm": 18.74008368530438, "learning_rate": 6.262863992402612e-08, "loss": 0.4021, "step": 18542 }, { "epoch": 2.9, "grad_norm": 27.894347682381458, "learning_rate": 6.24403379017613e-08, "loss": 0.4531, "step": 18543 }, { "epoch": 2.9, "grad_norm": 30.498933595634526, "learning_rate": 6.225231849787961e-08, "loss": 0.4429, "step": 18544 }, { "epoch": 2.9, "grad_norm": 24.38677491513527, "learning_rate": 6.2064581717729e-08, "loss": 0.4493, "step": 18545 }, { "epoch": 2.9, "grad_norm": 32.29721347417395, "learning_rate": 6.187712756664854e-08, "loss": 0.4592, "step": 18546 }, { "epoch": 2.9, "grad_norm": 36.42578139317899, "learning_rate": 6.168995604996841e-08, "loss": 0.4573, "step": 18547 }, { "epoch": 2.9, "grad_norm": 17.88275092978021, "learning_rate": 6.150306717301213e-08, "loss": 0.4129, "step": 18548 }, { "epoch": 2.9, "grad_norm": 26.442839142550525, "learning_rate": 6.131646094109545e-08, "loss": 0.4592, "step": 18549 }, { "epoch": 2.9, "grad_norm": 23.199971906040194, "learning_rate": 6.1130137359523e-08, "loss": 0.4071, "step": 18550 }, { "epoch": 2.9, "grad_norm": 41.83557046894883, "learning_rate": 6.094409643359722e-08, "loss": 0.5329, "step": 18551 }, { "epoch": 2.9, "grad_norm": 18.64986434035781, "learning_rate": 6.07583381686061e-08, "loss": 0.4239, "step": 18552 }, { "epoch": 2.9, "grad_norm": 28.978336676913905, "learning_rate": 6.05728625698343e-08, "loss": 0.4062, "step": 18553 }, { "epoch": 2.9, "grad_norm": 20.85896528963201, "learning_rate": 6.038766964255538e-08, "loss": 0.3951, "step": 18554 }, { "epoch": 2.9, "grad_norm": 17.77472125437093, "learning_rate": 6.020275939203734e-08, "loss": 0.4777, "step": 18555 }, { "epoch": 2.9, "grad_norm": 31.66742725057573, "learning_rate": 6.00181318235371e-08, "loss": 0.4444, "step": 18556 }, { "epoch": 2.9, "grad_norm": 29.886319099080577, "learning_rate": 5.983378694230713e-08, "loss": 0.4193, "step": 18557 }, { "epoch": 2.9, "grad_norm": 17.01866139274335, "learning_rate": 5.964972475358876e-08, "loss": 0.4237, "step": 18558 }, { "epoch": 2.9, "grad_norm": 17.787489964367737, "learning_rate": 5.946594526261895e-08, "loss": 0.4168, "step": 18559 }, { "epoch": 2.9, "grad_norm": 23.870874983506553, "learning_rate": 5.928244847462128e-08, "loss": 0.4386, "step": 18560 }, { "epoch": 2.9, "grad_norm": 24.5904435391408, "learning_rate": 5.909923439481491e-08, "loss": 0.4326, "step": 18561 }, { "epoch": 2.9, "grad_norm": 22.30948369156938, "learning_rate": 5.8916303028412334e-08, "loss": 0.4061, "step": 18562 }, { "epoch": 2.9, "grad_norm": 12.587707451907432, "learning_rate": 5.873365438061385e-08, "loss": 0.3877, "step": 18563 }, { "epoch": 2.9, "grad_norm": 20.80746495569008, "learning_rate": 5.8551288456616395e-08, "loss": 0.3906, "step": 18564 }, { "epoch": 2.9, "grad_norm": 25.238816500492025, "learning_rate": 5.83692052616025e-08, "loss": 0.4238, "step": 18565 }, { "epoch": 2.9, "grad_norm": 17.389540932413485, "learning_rate": 5.8187404800753575e-08, "loss": 0.4388, "step": 18566 }, { "epoch": 2.9, "grad_norm": 16.91401453385809, "learning_rate": 5.800588707923993e-08, "loss": 0.3984, "step": 18567 }, { "epoch": 2.9, "grad_norm": 28.492520582732723, "learning_rate": 5.782465210222299e-08, "loss": 0.4057, "step": 18568 }, { "epoch": 2.9, "grad_norm": 24.379133578247433, "learning_rate": 5.7643699874855296e-08, "loss": 0.4185, "step": 18569 }, { "epoch": 2.9, "grad_norm": 22.013488443175785, "learning_rate": 5.746303040228607e-08, "loss": 0.4521, "step": 18570 }, { "epoch": 2.9, "grad_norm": 22.820375565830574, "learning_rate": 5.7282643689652306e-08, "loss": 0.4176, "step": 18571 }, { "epoch": 2.9, "grad_norm": 28.053879731802162, "learning_rate": 5.710253974208324e-08, "loss": 0.4198, "step": 18572 }, { "epoch": 2.9, "grad_norm": 18.947676484006685, "learning_rate": 5.692271856470144e-08, "loss": 0.4331, "step": 18573 }, { "epoch": 2.9, "grad_norm": 19.998587581561058, "learning_rate": 5.674318016262059e-08, "loss": 0.4298, "step": 18574 }, { "epoch": 2.9, "grad_norm": 22.38055579174918, "learning_rate": 5.656392454094661e-08, "loss": 0.4834, "step": 18575 }, { "epoch": 2.9, "grad_norm": 17.908862691443375, "learning_rate": 5.638495170477876e-08, "loss": 0.4222, "step": 18576 }, { "epoch": 2.9, "grad_norm": 25.870998559218247, "learning_rate": 5.6206261659206284e-08, "loss": 0.4519, "step": 18577 }, { "epoch": 2.9, "grad_norm": 19.67904293092041, "learning_rate": 5.602785440931069e-08, "loss": 0.4067, "step": 18578 }, { "epoch": 2.9, "grad_norm": 29.013571404097277, "learning_rate": 5.584972996016569e-08, "loss": 0.4515, "step": 18579 }, { "epoch": 2.9, "grad_norm": 34.651736640621614, "learning_rate": 5.567188831683723e-08, "loss": 0.5154, "step": 18580 }, { "epoch": 2.9, "grad_norm": 22.13680798425253, "learning_rate": 5.549432948438238e-08, "loss": 0.4097, "step": 18581 }, { "epoch": 2.9, "grad_norm": 16.473877677655718, "learning_rate": 5.531705346785265e-08, "loss": 0.3921, "step": 18582 }, { "epoch": 2.9, "grad_norm": 22.161040284591447, "learning_rate": 5.514006027228735e-08, "loss": 0.4131, "step": 18583 }, { "epoch": 2.9, "grad_norm": 14.229827713016393, "learning_rate": 5.4963349902722454e-08, "loss": 0.3707, "step": 18584 }, { "epoch": 2.9, "grad_norm": 20.378601134818286, "learning_rate": 5.478692236418059e-08, "loss": 0.4258, "step": 18585 }, { "epoch": 2.9, "grad_norm": 18.933541126465354, "learning_rate": 5.4610777661681105e-08, "loss": 0.4573, "step": 18586 }, { "epoch": 2.9, "grad_norm": 21.56154401014442, "learning_rate": 5.443491580023441e-08, "loss": 0.4066, "step": 18587 }, { "epoch": 2.9, "grad_norm": 17.57596237400272, "learning_rate": 5.4259336784839855e-08, "loss": 0.4055, "step": 18588 }, { "epoch": 2.9, "grad_norm": 26.93826397190838, "learning_rate": 5.408404062049233e-08, "loss": 0.3974, "step": 18589 }, { "epoch": 2.9, "grad_norm": 17.51858298798443, "learning_rate": 5.390902731217562e-08, "loss": 0.3898, "step": 18590 }, { "epoch": 2.9, "grad_norm": 17.660792571553728, "learning_rate": 5.373429686486797e-08, "loss": 0.4059, "step": 18591 }, { "epoch": 2.9, "grad_norm": 25.495204372916653, "learning_rate": 5.355984928353986e-08, "loss": 0.468, "step": 18592 }, { "epoch": 2.9, "grad_norm": 25.989328184130365, "learning_rate": 5.338568457314952e-08, "loss": 0.4397, "step": 18593 }, { "epoch": 2.9, "grad_norm": 20.465690008419305, "learning_rate": 5.321180273865301e-08, "loss": 0.4257, "step": 18594 }, { "epoch": 2.9, "grad_norm": 20.193174553750463, "learning_rate": 5.303820378499303e-08, "loss": 0.4375, "step": 18595 }, { "epoch": 2.9, "grad_norm": 26.67642790736344, "learning_rate": 5.286488771710785e-08, "loss": 0.4524, "step": 18596 }, { "epoch": 2.9, "grad_norm": 13.756035368829533, "learning_rate": 5.2691854539926865e-08, "loss": 0.4082, "step": 18597 }, { "epoch": 2.91, "grad_norm": 21.197258258517074, "learning_rate": 5.251910425837059e-08, "loss": 0.4361, "step": 18598 }, { "epoch": 2.91, "grad_norm": 14.33317766935585, "learning_rate": 5.234663687735064e-08, "loss": 0.4043, "step": 18599 }, { "epoch": 2.91, "grad_norm": 26.62280763299338, "learning_rate": 5.217445240177421e-08, "loss": 0.5052, "step": 18600 }, { "epoch": 2.91, "grad_norm": 17.138903199289413, "learning_rate": 5.200255083653738e-08, "loss": 0.4137, "step": 18601 }, { "epoch": 2.91, "grad_norm": 22.868304257937098, "learning_rate": 5.183093218652846e-08, "loss": 0.3824, "step": 18602 }, { "epoch": 2.91, "grad_norm": 24.447223790989273, "learning_rate": 5.1659596456627995e-08, "loss": 0.4383, "step": 18603 }, { "epoch": 2.91, "grad_norm": 21.43291422741993, "learning_rate": 5.1488543651708746e-08, "loss": 0.4885, "step": 18604 }, { "epoch": 2.91, "grad_norm": 21.090242826865616, "learning_rate": 5.1317773776635715e-08, "loss": 0.4, "step": 18605 }, { "epoch": 2.91, "grad_norm": 22.938900327943976, "learning_rate": 5.114728683626613e-08, "loss": 0.4606, "step": 18606 }, { "epoch": 2.91, "grad_norm": 23.022326562030177, "learning_rate": 5.0977082835447224e-08, "loss": 0.446, "step": 18607 }, { "epoch": 2.91, "grad_norm": 15.911164662251947, "learning_rate": 5.0807161779019565e-08, "loss": 0.3709, "step": 18608 }, { "epoch": 2.91, "grad_norm": 20.155628051329376, "learning_rate": 5.0637523671817065e-08, "loss": 0.4125, "step": 18609 }, { "epoch": 2.91, "grad_norm": 16.328969247337927, "learning_rate": 5.046816851866254e-08, "loss": 0.3728, "step": 18610 }, { "epoch": 2.91, "grad_norm": 28.27759517110512, "learning_rate": 5.029909632437324e-08, "loss": 0.4255, "step": 18611 }, { "epoch": 2.91, "grad_norm": 23.982083425487772, "learning_rate": 5.013030709375644e-08, "loss": 0.4397, "step": 18612 }, { "epoch": 2.91, "grad_norm": 20.643691713699532, "learning_rate": 4.996180083161384e-08, "loss": 0.469, "step": 18613 }, { "epoch": 2.91, "grad_norm": 20.25380596551742, "learning_rate": 4.979357754273717e-08, "loss": 0.3876, "step": 18614 }, { "epoch": 2.91, "grad_norm": 24.61539945340483, "learning_rate": 4.9625637231910385e-08, "loss": 0.4598, "step": 18615 }, { "epoch": 2.91, "grad_norm": 28.110118856654594, "learning_rate": 4.9457979903909656e-08, "loss": 0.4055, "step": 18616 }, { "epoch": 2.91, "grad_norm": 24.847831118561604, "learning_rate": 4.92906055635034e-08, "loss": 0.4431, "step": 18617 }, { "epoch": 2.91, "grad_norm": 17.283652785266383, "learning_rate": 4.912351421545114e-08, "loss": 0.46, "step": 18618 }, { "epoch": 2.91, "grad_norm": 24.557359963271853, "learning_rate": 4.8956705864504625e-08, "loss": 0.4171, "step": 18619 }, { "epoch": 2.91, "grad_norm": 19.77325504479544, "learning_rate": 4.879018051540785e-08, "loss": 0.4968, "step": 18620 }, { "epoch": 2.91, "grad_norm": 26.937921242507, "learning_rate": 4.8623938172898125e-08, "loss": 0.4832, "step": 18621 }, { "epoch": 2.91, "grad_norm": 26.63163128508053, "learning_rate": 4.8457978841702777e-08, "loss": 0.4852, "step": 18622 }, { "epoch": 2.91, "grad_norm": 25.216370902881867, "learning_rate": 4.8292302526539156e-08, "loss": 0.4614, "step": 18623 }, { "epoch": 2.91, "grad_norm": 16.058903222800982, "learning_rate": 4.8126909232121265e-08, "loss": 0.4089, "step": 18624 }, { "epoch": 2.91, "grad_norm": 28.55656914493002, "learning_rate": 4.7961798963153116e-08, "loss": 0.3658, "step": 18625 }, { "epoch": 2.91, "grad_norm": 19.335173828765612, "learning_rate": 4.7796971724329844e-08, "loss": 0.4176, "step": 18626 }, { "epoch": 2.91, "grad_norm": 29.92343023157761, "learning_rate": 4.763242752033881e-08, "loss": 0.4596, "step": 18627 }, { "epoch": 2.91, "grad_norm": 26.754862113961618, "learning_rate": 4.746816635585849e-08, "loss": 0.4642, "step": 18628 }, { "epoch": 2.91, "grad_norm": 21.106521012563128, "learning_rate": 4.730418823556182e-08, "loss": 0.4528, "step": 18629 }, { "epoch": 2.91, "grad_norm": 22.291964721395384, "learning_rate": 4.7140493164112844e-08, "loss": 0.3749, "step": 18630 }, { "epoch": 2.91, "grad_norm": 23.228612524459315, "learning_rate": 4.6977081146165614e-08, "loss": 0.3591, "step": 18631 }, { "epoch": 2.91, "grad_norm": 18.39592857343386, "learning_rate": 4.6813952186366415e-08, "loss": 0.4128, "step": 18632 }, { "epoch": 2.91, "grad_norm": 26.597083295235763, "learning_rate": 4.6651106289357095e-08, "loss": 0.4548, "step": 18633 }, { "epoch": 2.91, "grad_norm": 17.806309333640822, "learning_rate": 4.648854345976839e-08, "loss": 0.3841, "step": 18634 }, { "epoch": 2.91, "grad_norm": 15.220720784791276, "learning_rate": 4.632626370222215e-08, "loss": 0.4147, "step": 18635 }, { "epoch": 2.91, "grad_norm": 25.83988797709514, "learning_rate": 4.6164267021334693e-08, "loss": 0.4305, "step": 18636 }, { "epoch": 2.91, "grad_norm": 22.425065938667647, "learning_rate": 4.6002553421711226e-08, "loss": 0.3922, "step": 18637 }, { "epoch": 2.91, "grad_norm": 30.214374762004176, "learning_rate": 4.5841122907953614e-08, "loss": 0.4502, "step": 18638 }, { "epoch": 2.91, "grad_norm": 22.165399191736235, "learning_rate": 4.567997548465153e-08, "loss": 0.4719, "step": 18639 }, { "epoch": 2.91, "grad_norm": 19.626846342036412, "learning_rate": 4.551911115638685e-08, "loss": 0.3702, "step": 18640 }, { "epoch": 2.91, "grad_norm": 18.0712803203373, "learning_rate": 4.5358529927735925e-08, "loss": 0.4592, "step": 18641 }, { "epoch": 2.91, "grad_norm": 29.73346874401523, "learning_rate": 4.5198231803265103e-08, "loss": 0.3759, "step": 18642 }, { "epoch": 2.91, "grad_norm": 26.92594201247352, "learning_rate": 4.503821678753406e-08, "loss": 0.4157, "step": 18643 }, { "epoch": 2.91, "grad_norm": 18.94960621579408, "learning_rate": 4.487848488509139e-08, "loss": 0.3873, "step": 18644 }, { "epoch": 2.91, "grad_norm": 27.034417033027776, "learning_rate": 4.4719036100481225e-08, "loss": 0.4707, "step": 18645 }, { "epoch": 2.91, "grad_norm": 20.6283469836735, "learning_rate": 4.455987043823884e-08, "loss": 0.4497, "step": 18646 }, { "epoch": 2.91, "grad_norm": 17.516282058220437, "learning_rate": 4.440098790288949e-08, "loss": 0.4272, "step": 18647 }, { "epoch": 2.91, "grad_norm": 42.846011660680006, "learning_rate": 4.4242388498951797e-08, "loss": 0.4764, "step": 18648 }, { "epoch": 2.91, "grad_norm": 18.041086199032765, "learning_rate": 4.408407223093658e-08, "loss": 0.4858, "step": 18649 }, { "epoch": 2.91, "grad_norm": 27.69422095747818, "learning_rate": 4.3926039103346915e-08, "loss": 0.4583, "step": 18650 }, { "epoch": 2.91, "grad_norm": 26.84931033681021, "learning_rate": 4.376828912067699e-08, "loss": 0.4087, "step": 18651 }, { "epoch": 2.91, "grad_norm": 26.845552992305535, "learning_rate": 4.361082228741209e-08, "loss": 0.4178, "step": 18652 }, { "epoch": 2.91, "grad_norm": 25.577283857073823, "learning_rate": 4.3453638608030865e-08, "loss": 0.481, "step": 18653 }, { "epoch": 2.91, "grad_norm": 25.728272857018755, "learning_rate": 4.329673808700308e-08, "loss": 0.5032, "step": 18654 }, { "epoch": 2.91, "grad_norm": 34.84744185745317, "learning_rate": 4.3140120728792925e-08, "loss": 0.4678, "step": 18655 }, { "epoch": 2.91, "grad_norm": 25.495810458498816, "learning_rate": 4.29837865378524e-08, "loss": 0.4702, "step": 18656 }, { "epoch": 2.91, "grad_norm": 18.704265429051265, "learning_rate": 4.2827735518629065e-08, "loss": 0.3311, "step": 18657 }, { "epoch": 2.91, "grad_norm": 14.398259129198157, "learning_rate": 4.2671967675559365e-08, "loss": 0.3712, "step": 18658 }, { "epoch": 2.91, "grad_norm": 26.14479882031989, "learning_rate": 4.2516483013074205e-08, "loss": 0.4725, "step": 18659 }, { "epoch": 2.91, "grad_norm": 17.658998022762546, "learning_rate": 4.23612815355956e-08, "loss": 0.4169, "step": 18660 }, { "epoch": 2.91, "grad_norm": 39.56902023118311, "learning_rate": 4.2206363247536684e-08, "loss": 0.3783, "step": 18661 }, { "epoch": 2.92, "grad_norm": 15.908342868291848, "learning_rate": 4.205172815330394e-08, "loss": 0.3516, "step": 18662 }, { "epoch": 2.92, "grad_norm": 18.428531260360174, "learning_rate": 4.189737625729384e-08, "loss": 0.411, "step": 18663 }, { "epoch": 2.92, "grad_norm": 43.42183158051474, "learning_rate": 4.174330756389844e-08, "loss": 0.5289, "step": 18664 }, { "epoch": 2.92, "grad_norm": 20.131858629523386, "learning_rate": 4.1589522077497556e-08, "loss": 0.3714, "step": 18665 }, { "epoch": 2.92, "grad_norm": 32.079054858466264, "learning_rate": 4.143601980246437e-08, "loss": 0.4846, "step": 18666 }, { "epoch": 2.92, "grad_norm": 25.2285427377757, "learning_rate": 4.128280074316649e-08, "loss": 0.4167, "step": 18667 }, { "epoch": 2.92, "grad_norm": 16.831271742485498, "learning_rate": 4.112986490395931e-08, "loss": 0.3816, "step": 18668 }, { "epoch": 2.92, "grad_norm": 26.932611297880097, "learning_rate": 4.0977212289192717e-08, "loss": 0.4682, "step": 18669 }, { "epoch": 2.92, "grad_norm": 18.87775375318171, "learning_rate": 4.082484290320876e-08, "loss": 0.4161, "step": 18670 }, { "epoch": 2.92, "grad_norm": 32.37367054334205, "learning_rate": 4.067275675034066e-08, "loss": 0.5205, "step": 18671 }, { "epoch": 2.92, "grad_norm": 22.177344558226117, "learning_rate": 4.0520953834912724e-08, "loss": 0.435, "step": 18672 }, { "epoch": 2.92, "grad_norm": 20.2196649354697, "learning_rate": 4.0369434161242616e-08, "loss": 0.4382, "step": 18673 }, { "epoch": 2.92, "grad_norm": 20.14482200298816, "learning_rate": 4.021819773364022e-08, "loss": 0.4702, "step": 18674 }, { "epoch": 2.92, "grad_norm": 23.06524709813954, "learning_rate": 4.0067244556405424e-08, "loss": 0.4338, "step": 18675 }, { "epoch": 2.92, "grad_norm": 16.27565151927008, "learning_rate": 3.9916574633832586e-08, "loss": 0.4441, "step": 18676 }, { "epoch": 2.92, "grad_norm": 26.532480593809655, "learning_rate": 3.976618797020493e-08, "loss": 0.4556, "step": 18677 }, { "epoch": 2.92, "grad_norm": 18.46870527326587, "learning_rate": 3.961608456980126e-08, "loss": 0.4861, "step": 18678 }, { "epoch": 2.92, "grad_norm": 19.92178228169725, "learning_rate": 3.946626443688817e-08, "loss": 0.4586, "step": 18679 }, { "epoch": 2.92, "grad_norm": 32.09177854234582, "learning_rate": 3.9316727575728906e-08, "loss": 0.4041, "step": 18680 }, { "epoch": 2.92, "grad_norm": 31.647061054597792, "learning_rate": 3.9167473990575635e-08, "loss": 0.5062, "step": 18681 }, { "epoch": 2.92, "grad_norm": 20.565054413074023, "learning_rate": 3.901850368567161e-08, "loss": 0.4193, "step": 18682 }, { "epoch": 2.92, "grad_norm": 20.209451963656534, "learning_rate": 3.886981666525347e-08, "loss": 0.381, "step": 18683 }, { "epoch": 2.92, "grad_norm": 23.65321754964633, "learning_rate": 3.8721412933552246e-08, "loss": 0.4466, "step": 18684 }, { "epoch": 2.92, "grad_norm": 22.789543695374963, "learning_rate": 3.85732924947857e-08, "loss": 0.4341, "step": 18685 }, { "epoch": 2.92, "grad_norm": 20.63322895789619, "learning_rate": 3.8425455353168215e-08, "loss": 0.4401, "step": 18686 }, { "epoch": 2.92, "grad_norm": 24.262060936557013, "learning_rate": 3.827790151290312e-08, "loss": 0.4497, "step": 18687 }, { "epoch": 2.92, "grad_norm": 18.735849732075785, "learning_rate": 3.813063097818703e-08, "loss": 0.4099, "step": 18688 }, { "epoch": 2.92, "grad_norm": 21.165894683579662, "learning_rate": 3.798364375320773e-08, "loss": 0.4682, "step": 18689 }, { "epoch": 2.92, "grad_norm": 21.28106580562114, "learning_rate": 3.783693984214743e-08, "loss": 0.386, "step": 18690 }, { "epoch": 2.92, "grad_norm": 20.474656576320363, "learning_rate": 3.7690519249174996e-08, "loss": 0.4668, "step": 18691 }, { "epoch": 2.92, "grad_norm": 28.814855488587384, "learning_rate": 3.754438197845822e-08, "loss": 0.4856, "step": 18692 }, { "epoch": 2.92, "grad_norm": 30.833775246627976, "learning_rate": 3.739852803415045e-08, "loss": 0.456, "step": 18693 }, { "epoch": 2.92, "grad_norm": 18.52357529838724, "learning_rate": 3.725295742040058e-08, "loss": 0.4246, "step": 18694 }, { "epoch": 2.92, "grad_norm": 22.3044660233226, "learning_rate": 3.710767014134864e-08, "loss": 0.4162, "step": 18695 }, { "epoch": 2.92, "grad_norm": 18.671897384199188, "learning_rate": 3.696266620112576e-08, "loss": 0.3841, "step": 18696 }, { "epoch": 2.92, "grad_norm": 21.70616058229042, "learning_rate": 3.681794560385754e-08, "loss": 0.4908, "step": 18697 }, { "epoch": 2.92, "grad_norm": 18.202102105211527, "learning_rate": 3.667350835365846e-08, "loss": 0.4185, "step": 18698 }, { "epoch": 2.92, "grad_norm": 23.94815043423386, "learning_rate": 3.6529354454635236e-08, "loss": 0.4304, "step": 18699 }, { "epoch": 2.92, "grad_norm": 31.318605188511896, "learning_rate": 3.638548391089014e-08, "loss": 0.3747, "step": 18700 }, { "epoch": 2.92, "grad_norm": 16.999345002458742, "learning_rate": 3.6241896726513236e-08, "loss": 0.4546, "step": 18701 }, { "epoch": 2.92, "grad_norm": 28.485159415881544, "learning_rate": 3.6098592905587925e-08, "loss": 0.4317, "step": 18702 }, { "epoch": 2.92, "grad_norm": 20.923521386622834, "learning_rate": 3.595557245218983e-08, "loss": 0.3923, "step": 18703 }, { "epoch": 2.92, "grad_norm": 25.324056303276155, "learning_rate": 3.581283537038571e-08, "loss": 0.4537, "step": 18704 }, { "epoch": 2.92, "grad_norm": 20.395555892423726, "learning_rate": 3.567038166423675e-08, "loss": 0.3949, "step": 18705 }, { "epoch": 2.92, "grad_norm": 29.604980440563356, "learning_rate": 3.552821133779305e-08, "loss": 0.5119, "step": 18706 }, { "epoch": 2.92, "grad_norm": 20.632383607149897, "learning_rate": 3.5386324395096924e-08, "loss": 0.5184, "step": 18707 }, { "epoch": 2.92, "grad_norm": 22.740222758894927, "learning_rate": 3.5244720840186266e-08, "loss": 0.4184, "step": 18708 }, { "epoch": 2.92, "grad_norm": 17.288677795653232, "learning_rate": 3.510340067708562e-08, "loss": 0.3519, "step": 18709 }, { "epoch": 2.92, "grad_norm": 24.570040017513822, "learning_rate": 3.4962363909815114e-08, "loss": 0.4102, "step": 18710 }, { "epoch": 2.92, "grad_norm": 25.039475040340207, "learning_rate": 3.482161054238486e-08, "loss": 0.4078, "step": 18711 }, { "epoch": 2.92, "grad_norm": 35.83570232526256, "learning_rate": 3.4681140578799453e-08, "loss": 0.4508, "step": 18712 }, { "epoch": 2.92, "grad_norm": 14.077287785597578, "learning_rate": 3.4540954023052356e-08, "loss": 0.373, "step": 18713 }, { "epoch": 2.92, "grad_norm": 28.38416494438566, "learning_rate": 3.440105087913148e-08, "loss": 0.5014, "step": 18714 }, { "epoch": 2.92, "grad_norm": 26.44152072786472, "learning_rate": 3.426143115101477e-08, "loss": 0.4134, "step": 18715 }, { "epoch": 2.92, "grad_norm": 23.044596320506358, "learning_rate": 3.412209484267237e-08, "loss": 0.3973, "step": 18716 }, { "epoch": 2.92, "grad_norm": 20.570489149932435, "learning_rate": 3.39830419580689e-08, "loss": 0.4287, "step": 18717 }, { "epoch": 2.92, "grad_norm": 17.726730926742437, "learning_rate": 3.384427250115674e-08, "loss": 0.3965, "step": 18718 }, { "epoch": 2.92, "grad_norm": 20.45951085234899, "learning_rate": 3.370578647588496e-08, "loss": 0.4517, "step": 18719 }, { "epoch": 2.92, "grad_norm": 27.034624249653447, "learning_rate": 3.3567583886189304e-08, "loss": 0.4584, "step": 18720 }, { "epoch": 2.92, "grad_norm": 33.022644246421905, "learning_rate": 3.3429664736001064e-08, "loss": 0.3956, "step": 18721 }, { "epoch": 2.92, "grad_norm": 20.400692286618987, "learning_rate": 3.3292029029243777e-08, "loss": 0.4209, "step": 18722 }, { "epoch": 2.92, "grad_norm": 24.62312803559912, "learning_rate": 3.315467676982986e-08, "loss": 0.419, "step": 18723 }, { "epoch": 2.92, "grad_norm": 19.84010695785895, "learning_rate": 3.301760796166731e-08, "loss": 0.4454, "step": 18724 }, { "epoch": 2.92, "grad_norm": 20.53350933336121, "learning_rate": 3.2880822608653e-08, "loss": 0.3966, "step": 18725 }, { "epoch": 2.93, "grad_norm": 21.667395581232263, "learning_rate": 3.274432071467826e-08, "loss": 0.3987, "step": 18726 }, { "epoch": 2.93, "grad_norm": 18.842186840818783, "learning_rate": 3.260810228362332e-08, "loss": 0.4979, "step": 18727 }, { "epoch": 2.93, "grad_norm": 22.34221306429955, "learning_rate": 3.247216731936398e-08, "loss": 0.4167, "step": 18728 }, { "epoch": 2.93, "grad_norm": 19.175157848862305, "learning_rate": 3.233651582576491e-08, "loss": 0.4058, "step": 18729 }, { "epoch": 2.93, "grad_norm": 21.698653936717903, "learning_rate": 3.220114780668415e-08, "loss": 0.4214, "step": 18730 }, { "epoch": 2.93, "grad_norm": 25.789606687807048, "learning_rate": 3.206606326597306e-08, "loss": 0.4594, "step": 18731 }, { "epoch": 2.93, "grad_norm": 20.24767071004808, "learning_rate": 3.193126220747078e-08, "loss": 0.4763, "step": 18732 }, { "epoch": 2.93, "grad_norm": 16.19364381521254, "learning_rate": 3.1796744635013144e-08, "loss": 0.4126, "step": 18733 }, { "epoch": 2.93, "grad_norm": 27.44508202963948, "learning_rate": 3.166251055242375e-08, "loss": 0.4675, "step": 18734 }, { "epoch": 2.93, "grad_norm": 23.68752461812091, "learning_rate": 3.1528559963522886e-08, "loss": 0.4252, "step": 18735 }, { "epoch": 2.93, "grad_norm": 20.882091399878316, "learning_rate": 3.139489287211639e-08, "loss": 0.3822, "step": 18736 }, { "epoch": 2.93, "grad_norm": 18.122581247374395, "learning_rate": 3.1261509282009e-08, "loss": 0.474, "step": 18737 }, { "epoch": 2.93, "grad_norm": 19.869086343954912, "learning_rate": 3.112840919699212e-08, "loss": 0.4094, "step": 18738 }, { "epoch": 2.93, "grad_norm": 20.61216789833138, "learning_rate": 3.099559262085272e-08, "loss": 0.3968, "step": 18739 }, { "epoch": 2.93, "grad_norm": 29.14144736758251, "learning_rate": 3.086305955736557e-08, "loss": 0.4646, "step": 18740 }, { "epoch": 2.93, "grad_norm": 17.577402657819984, "learning_rate": 3.073081001030209e-08, "loss": 0.4229, "step": 18741 }, { "epoch": 2.93, "grad_norm": 31.779288952076058, "learning_rate": 3.05988439834215e-08, "loss": 0.445, "step": 18742 }, { "epoch": 2.93, "grad_norm": 24.00281304771441, "learning_rate": 3.046716148047968e-08, "loss": 0.3836, "step": 18743 }, { "epoch": 2.93, "grad_norm": 28.742786190747925, "learning_rate": 3.033576250521919e-08, "loss": 0.4648, "step": 18744 }, { "epoch": 2.93, "grad_norm": 22.357514310985454, "learning_rate": 3.0204647061375936e-08, "loss": 0.3959, "step": 18745 }, { "epoch": 2.93, "grad_norm": 19.758723977431917, "learning_rate": 3.0073815152681374e-08, "loss": 0.3909, "step": 18746 }, { "epoch": 2.93, "grad_norm": 29.566795283058966, "learning_rate": 2.994326678285586e-08, "loss": 0.4181, "step": 18747 }, { "epoch": 2.93, "grad_norm": 18.89971685474977, "learning_rate": 2.981300195561088e-08, "loss": 0.4496, "step": 18748 }, { "epoch": 2.93, "grad_norm": 17.142760209003928, "learning_rate": 2.968302067465234e-08, "loss": 0.3766, "step": 18749 }, { "epoch": 2.93, "grad_norm": 17.50725785704199, "learning_rate": 2.955332294367508e-08, "loss": 0.3809, "step": 18750 }, { "epoch": 2.93, "grad_norm": 19.53937405425937, "learning_rate": 2.942390876636947e-08, "loss": 0.4405, "step": 18751 }, { "epoch": 2.93, "grad_norm": 30.427025853152482, "learning_rate": 2.9294778146415902e-08, "loss": 0.4646, "step": 18752 }, { "epoch": 2.93, "grad_norm": 24.40538684873113, "learning_rate": 2.916593108748589e-08, "loss": 0.4403, "step": 18753 }, { "epoch": 2.93, "grad_norm": 36.357294153424114, "learning_rate": 2.903736759324316e-08, "loss": 0.4464, "step": 18754 }, { "epoch": 2.93, "grad_norm": 37.04455475316805, "learning_rate": 2.8909087667345905e-08, "loss": 0.5209, "step": 18755 }, { "epoch": 2.93, "grad_norm": 19.993564961493274, "learning_rate": 2.878109131344009e-08, "loss": 0.4659, "step": 18756 }, { "epoch": 2.93, "grad_norm": 19.53477503590143, "learning_rate": 2.8653378535168363e-08, "loss": 0.3906, "step": 18757 }, { "epoch": 2.93, "grad_norm": 25.053004733778717, "learning_rate": 2.852594933616004e-08, "loss": 0.3998, "step": 18758 }, { "epoch": 2.93, "grad_norm": 25.938913936739894, "learning_rate": 2.8398803720042223e-08, "loss": 0.4132, "step": 18759 }, { "epoch": 2.93, "grad_norm": 24.73032173157626, "learning_rate": 2.8271941690427574e-08, "loss": 0.4392, "step": 18760 }, { "epoch": 2.93, "grad_norm": 15.357475297163742, "learning_rate": 2.8145363250926537e-08, "loss": 0.466, "step": 18761 }, { "epoch": 2.93, "grad_norm": 26.91957492822058, "learning_rate": 2.801906840513735e-08, "loss": 0.455, "step": 18762 }, { "epoch": 2.93, "grad_norm": 16.342303117196654, "learning_rate": 2.7893057156653802e-08, "loss": 0.3851, "step": 18763 }, { "epoch": 2.93, "grad_norm": 17.9469692244357, "learning_rate": 2.776732950905636e-08, "loss": 0.4209, "step": 18764 }, { "epoch": 2.93, "grad_norm": 17.63479290076147, "learning_rate": 2.764188546592439e-08, "loss": 0.4102, "step": 18765 }, { "epoch": 2.93, "grad_norm": 19.009678048630057, "learning_rate": 2.7516725030821702e-08, "loss": 0.4475, "step": 18766 }, { "epoch": 2.93, "grad_norm": 30.39524087934214, "learning_rate": 2.7391848207311e-08, "loss": 0.4804, "step": 18767 }, { "epoch": 2.93, "grad_norm": 16.279164359612306, "learning_rate": 2.7267254998941673e-08, "loss": 0.4328, "step": 18768 }, { "epoch": 2.93, "grad_norm": 19.07723564946217, "learning_rate": 2.714294540925866e-08, "loss": 0.4298, "step": 18769 }, { "epoch": 2.93, "grad_norm": 21.173648797225805, "learning_rate": 2.701891944179469e-08, "loss": 0.4188, "step": 18770 }, { "epoch": 2.93, "grad_norm": 33.013802930793744, "learning_rate": 2.6895177100079163e-08, "loss": 0.4474, "step": 18771 }, { "epoch": 2.93, "grad_norm": 21.179250839302412, "learning_rate": 2.6771718387631485e-08, "loss": 0.4239, "step": 18772 }, { "epoch": 2.93, "grad_norm": 36.06192174477639, "learning_rate": 2.6648543307962183e-08, "loss": 0.413, "step": 18773 }, { "epoch": 2.93, "grad_norm": 22.85214303021008, "learning_rate": 2.65256518645729e-08, "loss": 0.4541, "step": 18774 }, { "epoch": 2.93, "grad_norm": 31.70182055726893, "learning_rate": 2.6403044060959726e-08, "loss": 0.5998, "step": 18775 }, { "epoch": 2.93, "grad_norm": 24.550304011863258, "learning_rate": 2.6280719900609874e-08, "loss": 0.4539, "step": 18776 }, { "epoch": 2.93, "grad_norm": 21.817987495704166, "learning_rate": 2.6158679387002782e-08, "loss": 0.4139, "step": 18777 }, { "epoch": 2.93, "grad_norm": 33.64175094761346, "learning_rate": 2.60369225236079e-08, "loss": 0.4318, "step": 18778 }, { "epoch": 2.93, "grad_norm": 26.811499794842092, "learning_rate": 2.591544931388801e-08, "loss": 0.4256, "step": 18779 }, { "epoch": 2.93, "grad_norm": 26.57258492936507, "learning_rate": 2.5794259761298124e-08, "loss": 0.4902, "step": 18780 }, { "epoch": 2.93, "grad_norm": 32.15790848305961, "learning_rate": 2.5673353869284378e-08, "loss": 0.4376, "step": 18781 }, { "epoch": 2.93, "grad_norm": 18.634583213528288, "learning_rate": 2.5552731641286243e-08, "loss": 0.3896, "step": 18782 }, { "epoch": 2.93, "grad_norm": 27.702546031343232, "learning_rate": 2.5432393080733197e-08, "loss": 0.4426, "step": 18783 }, { "epoch": 2.93, "grad_norm": 20.141080917750035, "learning_rate": 2.5312338191048057e-08, "loss": 0.3945, "step": 18784 }, { "epoch": 2.93, "grad_norm": 20.313118559148904, "learning_rate": 2.5192566975644762e-08, "loss": 0.4106, "step": 18785 }, { "epoch": 2.93, "grad_norm": 26.031776390844172, "learning_rate": 2.507307943793058e-08, "loss": 0.443, "step": 18786 }, { "epoch": 2.93, "grad_norm": 24.179225148724576, "learning_rate": 2.495387558130169e-08, "loss": 0.4772, "step": 18787 }, { "epoch": 2.93, "grad_norm": 17.72467751902362, "learning_rate": 2.4834955409149818e-08, "loss": 0.3954, "step": 18788 }, { "epoch": 2.93, "grad_norm": 20.96829298177997, "learning_rate": 2.471631892485671e-08, "loss": 0.405, "step": 18789 }, { "epoch": 2.94, "grad_norm": 19.188812197011092, "learning_rate": 2.4597966131796324e-08, "loss": 0.4266, "step": 18790 }, { "epoch": 2.94, "grad_norm": 22.624193024996938, "learning_rate": 2.4479897033333756e-08, "loss": 0.3841, "step": 18791 }, { "epoch": 2.94, "grad_norm": 19.53607711040371, "learning_rate": 2.4362111632827423e-08, "loss": 0.374, "step": 18792 }, { "epoch": 2.94, "grad_norm": 20.580840083730862, "learning_rate": 2.4244609933627982e-08, "loss": 0.4027, "step": 18793 }, { "epoch": 2.94, "grad_norm": 30.237717700965412, "learning_rate": 2.4127391939076094e-08, "loss": 0.4185, "step": 18794 }, { "epoch": 2.94, "grad_norm": 22.415351902388192, "learning_rate": 2.401045765250465e-08, "loss": 0.3951, "step": 18795 }, { "epoch": 2.94, "grad_norm": 21.193501235139802, "learning_rate": 2.3893807077239872e-08, "loss": 0.4041, "step": 18796 }, { "epoch": 2.94, "grad_norm": 21.20760443617767, "learning_rate": 2.3777440216600224e-08, "loss": 0.4118, "step": 18797 }, { "epoch": 2.94, "grad_norm": 26.4211039557542, "learning_rate": 2.3661357073894166e-08, "loss": 0.5376, "step": 18798 }, { "epoch": 2.94, "grad_norm": 29.610038295955878, "learning_rate": 2.3545557652422392e-08, "loss": 0.4614, "step": 18799 }, { "epoch": 2.94, "grad_norm": 18.333718131824373, "learning_rate": 2.343004195547893e-08, "loss": 0.4835, "step": 18800 }, { "epoch": 2.94, "grad_norm": 22.04216062168671, "learning_rate": 2.3314809986348942e-08, "loss": 0.4097, "step": 18801 }, { "epoch": 2.94, "grad_norm": 27.712977368429215, "learning_rate": 2.3199861748309793e-08, "loss": 0.3999, "step": 18802 }, { "epoch": 2.94, "grad_norm": 22.79517534651258, "learning_rate": 2.3085197244631097e-08, "loss": 0.4006, "step": 18803 }, { "epoch": 2.94, "grad_norm": 30.395588764128835, "learning_rate": 2.2970816478572466e-08, "loss": 0.4285, "step": 18804 }, { "epoch": 2.94, "grad_norm": 22.41010286146284, "learning_rate": 2.2856719453386856e-08, "loss": 0.3604, "step": 18805 }, { "epoch": 2.94, "grad_norm": 19.367444586422824, "learning_rate": 2.2742906172320555e-08, "loss": 0.4169, "step": 18806 }, { "epoch": 2.94, "grad_norm": 16.25920666257959, "learning_rate": 2.262937663860876e-08, "loss": 0.4082, "step": 18807 }, { "epoch": 2.94, "grad_norm": 35.683292052732334, "learning_rate": 2.25161308554811e-08, "loss": 0.4507, "step": 18808 }, { "epoch": 2.94, "grad_norm": 33.708381624898436, "learning_rate": 2.240316882615834e-08, "loss": 0.5187, "step": 18809 }, { "epoch": 2.94, "grad_norm": 24.625063948136773, "learning_rate": 2.2290490553852352e-08, "loss": 0.449, "step": 18810 }, { "epoch": 2.94, "grad_norm": 24.675553624388154, "learning_rate": 2.217809604176835e-08, "loss": 0.3801, "step": 18811 }, { "epoch": 2.94, "grad_norm": 26.2654836391024, "learning_rate": 2.2065985293102664e-08, "loss": 0.423, "step": 18812 }, { "epoch": 2.94, "grad_norm": 34.24203353227539, "learning_rate": 2.195415831104275e-08, "loss": 0.4576, "step": 18813 }, { "epoch": 2.94, "grad_norm": 37.33191503746383, "learning_rate": 2.1842615098769394e-08, "loss": 0.4552, "step": 18814 }, { "epoch": 2.94, "grad_norm": 21.282757662276865, "learning_rate": 2.1731355659456722e-08, "loss": 0.4664, "step": 18815 }, { "epoch": 2.94, "grad_norm": 21.31259725987568, "learning_rate": 2.162037999626554e-08, "loss": 0.3783, "step": 18816 }, { "epoch": 2.94, "grad_norm": 38.773255630057406, "learning_rate": 2.1509688112354432e-08, "loss": 0.4386, "step": 18817 }, { "epoch": 2.94, "grad_norm": 20.559423254153288, "learning_rate": 2.139928001086977e-08, "loss": 0.4554, "step": 18818 }, { "epoch": 2.94, "grad_norm": 29.073919519978006, "learning_rate": 2.128915569495238e-08, "loss": 0.4391, "step": 18819 }, { "epoch": 2.94, "grad_norm": 18.948614969190782, "learning_rate": 2.1179315167734194e-08, "loss": 0.3962, "step": 18820 }, { "epoch": 2.94, "grad_norm": 20.99502684539474, "learning_rate": 2.1069758432339382e-08, "loss": 0.3903, "step": 18821 }, { "epoch": 2.94, "grad_norm": 15.779475235051981, "learning_rate": 2.096048549188212e-08, "loss": 0.3892, "step": 18822 }, { "epoch": 2.94, "grad_norm": 17.94539963331978, "learning_rate": 2.0851496349472144e-08, "loss": 0.4245, "step": 18823 }, { "epoch": 2.94, "grad_norm": 20.91316604901738, "learning_rate": 2.0742791008206974e-08, "loss": 0.4323, "step": 18824 }, { "epoch": 2.94, "grad_norm": 28.108933797854498, "learning_rate": 2.0634369471179692e-08, "loss": 0.5228, "step": 18825 }, { "epoch": 2.94, "grad_norm": 22.348109574930174, "learning_rate": 2.052623174147339e-08, "loss": 0.4347, "step": 18826 }, { "epoch": 2.94, "grad_norm": 22.69923514183554, "learning_rate": 2.0418377822162272e-08, "loss": 0.3567, "step": 18827 }, { "epoch": 2.94, "grad_norm": 16.375016200625435, "learning_rate": 2.0310807716316107e-08, "loss": 0.407, "step": 18828 }, { "epoch": 2.94, "grad_norm": 22.110251296777143, "learning_rate": 2.0203521426991333e-08, "loss": 0.442, "step": 18829 }, { "epoch": 2.94, "grad_norm": 23.680601394345754, "learning_rate": 2.0096518957241072e-08, "loss": 0.3985, "step": 18830 }, { "epoch": 2.94, "grad_norm": 19.201817616219056, "learning_rate": 1.9989800310107333e-08, "loss": 0.449, "step": 18831 }, { "epoch": 2.94, "grad_norm": 18.49303878118887, "learning_rate": 1.9883365488625462e-08, "loss": 0.418, "step": 18832 }, { "epoch": 2.94, "grad_norm": 34.593306186244746, "learning_rate": 1.977721449582304e-08, "loss": 0.4808, "step": 18833 }, { "epoch": 2.94, "grad_norm": 24.258314419282225, "learning_rate": 1.9671347334717648e-08, "loss": 0.4271, "step": 18834 }, { "epoch": 2.94, "grad_norm": 19.64951328495019, "learning_rate": 1.956576400832133e-08, "loss": 0.4314, "step": 18835 }, { "epoch": 2.94, "grad_norm": 25.061544618923524, "learning_rate": 1.9460464519636124e-08, "loss": 0.4569, "step": 18836 }, { "epoch": 2.94, "grad_norm": 26.005278786779574, "learning_rate": 1.9355448871657413e-08, "loss": 0.4879, "step": 18837 }, { "epoch": 2.94, "grad_norm": 23.59853772602622, "learning_rate": 1.9250717067370583e-08, "loss": 0.4838, "step": 18838 }, { "epoch": 2.94, "grad_norm": 24.536518005080442, "learning_rate": 1.9146269109755477e-08, "loss": 0.5077, "step": 18839 }, { "epoch": 2.94, "grad_norm": 24.854227551529316, "learning_rate": 1.904210500178083e-08, "loss": 0.4185, "step": 18840 }, { "epoch": 2.94, "grad_norm": 24.634960533997567, "learning_rate": 1.8938224746410938e-08, "loss": 0.384, "step": 18841 }, { "epoch": 2.94, "grad_norm": 25.961780777768983, "learning_rate": 1.8834628346598993e-08, "loss": 0.4577, "step": 18842 }, { "epoch": 2.94, "grad_norm": 18.768583418684553, "learning_rate": 1.8731315805290418e-08, "loss": 0.4575, "step": 18843 }, { "epoch": 2.94, "grad_norm": 33.07991543925951, "learning_rate": 1.862828712542508e-08, "loss": 0.4274, "step": 18844 }, { "epoch": 2.94, "grad_norm": 16.91244361565658, "learning_rate": 1.8525542309932865e-08, "loss": 0.4454, "step": 18845 }, { "epoch": 2.94, "grad_norm": 21.393036626288726, "learning_rate": 1.8423081361734762e-08, "loss": 0.4224, "step": 18846 }, { "epoch": 2.94, "grad_norm": 16.805750169499632, "learning_rate": 1.832090428374511e-08, "loss": 0.4225, "step": 18847 }, { "epoch": 2.94, "grad_norm": 17.846242157756087, "learning_rate": 1.8219011078869365e-08, "loss": 0.4427, "step": 18848 }, { "epoch": 2.94, "grad_norm": 22.013429965478007, "learning_rate": 1.8117401750006314e-08, "loss": 0.4592, "step": 18849 }, { "epoch": 2.94, "grad_norm": 31.881296388191885, "learning_rate": 1.801607630004476e-08, "loss": 0.4904, "step": 18850 }, { "epoch": 2.94, "grad_norm": 25.435922524561278, "learning_rate": 1.7915034731867952e-08, "loss": 0.4484, "step": 18851 }, { "epoch": 2.94, "grad_norm": 17.4465189340236, "learning_rate": 1.781427704834693e-08, "loss": 0.418, "step": 18852 }, { "epoch": 2.94, "grad_norm": 19.433702504222726, "learning_rate": 1.7713803252348283e-08, "loss": 0.4352, "step": 18853 }, { "epoch": 2.95, "grad_norm": 27.7159977608798, "learning_rate": 1.761361334673084e-08, "loss": 0.4113, "step": 18854 }, { "epoch": 2.95, "grad_norm": 14.719987480874398, "learning_rate": 1.7513707334341212e-08, "loss": 0.4255, "step": 18855 }, { "epoch": 2.95, "grad_norm": 24.99010997292553, "learning_rate": 1.741408521802379e-08, "loss": 0.3937, "step": 18856 }, { "epoch": 2.95, "grad_norm": 38.3508837187754, "learning_rate": 1.7314747000608532e-08, "loss": 0.4187, "step": 18857 }, { "epoch": 2.95, "grad_norm": 25.896340583041276, "learning_rate": 1.721569268492318e-08, "loss": 0.4719, "step": 18858 }, { "epoch": 2.95, "grad_norm": 24.16564306250834, "learning_rate": 1.7116922273783255e-08, "loss": 0.4503, "step": 18859 }, { "epoch": 2.95, "grad_norm": 27.82957492523101, "learning_rate": 1.7018435769998732e-08, "loss": 0.4227, "step": 18860 }, { "epoch": 2.95, "grad_norm": 18.948515972378544, "learning_rate": 1.6920233176369594e-08, "loss": 0.4106, "step": 18861 }, { "epoch": 2.95, "grad_norm": 23.739760804248764, "learning_rate": 1.6822314495689164e-08, "loss": 0.4203, "step": 18862 }, { "epoch": 2.95, "grad_norm": 16.041260531989856, "learning_rate": 1.6724679730742986e-08, "loss": 0.3665, "step": 18863 }, { "epoch": 2.95, "grad_norm": 24.729139696897622, "learning_rate": 1.6627328884305516e-08, "loss": 0.4103, "step": 18864 }, { "epoch": 2.95, "grad_norm": 27.878113369639426, "learning_rate": 1.6530261959147863e-08, "loss": 0.4326, "step": 18865 }, { "epoch": 2.95, "grad_norm": 25.231729775072946, "learning_rate": 1.6433478958028938e-08, "loss": 0.4551, "step": 18866 }, { "epoch": 2.95, "grad_norm": 20.282995790276836, "learning_rate": 1.6336979883700976e-08, "loss": 0.4121, "step": 18867 }, { "epoch": 2.95, "grad_norm": 32.43296778811208, "learning_rate": 1.6240764738909566e-08, "loss": 0.478, "step": 18868 }, { "epoch": 2.95, "grad_norm": 22.033986177707625, "learning_rate": 1.6144833526390292e-08, "loss": 0.4184, "step": 18869 }, { "epoch": 2.95, "grad_norm": 21.658520664140898, "learning_rate": 1.6049186248872084e-08, "loss": 0.4372, "step": 18870 }, { "epoch": 2.95, "grad_norm": 22.776506783762112, "learning_rate": 1.595382290907388e-08, "loss": 0.4233, "step": 18871 }, { "epoch": 2.95, "grad_norm": 16.954463981453696, "learning_rate": 1.585874350970906e-08, "loss": 0.3926, "step": 18872 }, { "epoch": 2.95, "grad_norm": 21.216906088205832, "learning_rate": 1.5763948053481026e-08, "loss": 0.3413, "step": 18873 }, { "epoch": 2.95, "grad_norm": 19.106083866466513, "learning_rate": 1.566943654308539e-08, "loss": 0.3729, "step": 18874 }, { "epoch": 2.95, "grad_norm": 21.707289195317248, "learning_rate": 1.557520898121001e-08, "loss": 0.3961, "step": 18875 }, { "epoch": 2.95, "grad_norm": 21.36446340655278, "learning_rate": 1.5481265370536068e-08, "loss": 0.4516, "step": 18876 }, { "epoch": 2.95, "grad_norm": 31.912753232155914, "learning_rate": 1.5387605713732545e-08, "loss": 0.4214, "step": 18877 }, { "epoch": 2.95, "grad_norm": 17.271925813908354, "learning_rate": 1.5294230013466194e-08, "loss": 0.3983, "step": 18878 }, { "epoch": 2.95, "grad_norm": 26.304181007673105, "learning_rate": 1.520113827239045e-08, "loss": 0.4787, "step": 18879 }, { "epoch": 2.95, "grad_norm": 17.73109102470334, "learning_rate": 1.510833049315319e-08, "loss": 0.468, "step": 18880 }, { "epoch": 2.95, "grad_norm": 28.12424159722876, "learning_rate": 1.501580667839453e-08, "loss": 0.4406, "step": 18881 }, { "epoch": 2.95, "grad_norm": 35.95885184847225, "learning_rate": 1.4923566830744586e-08, "loss": 0.3809, "step": 18882 }, { "epoch": 2.95, "grad_norm": 32.0822114380872, "learning_rate": 1.4831610952827925e-08, "loss": 0.4653, "step": 18883 }, { "epoch": 2.95, "grad_norm": 14.792258107059261, "learning_rate": 1.4739939047259122e-08, "loss": 0.4393, "step": 18884 }, { "epoch": 2.95, "grad_norm": 19.699509350934477, "learning_rate": 1.4648551116644982e-08, "loss": 0.408, "step": 18885 }, { "epoch": 2.95, "grad_norm": 17.44364783821956, "learning_rate": 1.4557447163584538e-08, "loss": 0.4689, "step": 18886 }, { "epoch": 2.95, "grad_norm": 26.215502723706983, "learning_rate": 1.4466627190669047e-08, "loss": 0.4321, "step": 18887 }, { "epoch": 2.95, "grad_norm": 21.272045555064263, "learning_rate": 1.4376091200482002e-08, "loss": 0.4471, "step": 18888 }, { "epoch": 2.95, "grad_norm": 34.28184718943699, "learning_rate": 1.4285839195596896e-08, "loss": 0.4598, "step": 18889 }, { "epoch": 2.95, "grad_norm": 27.576272715320606, "learning_rate": 1.4195871178580567e-08, "loss": 0.4785, "step": 18890 }, { "epoch": 2.95, "grad_norm": 19.07893117830633, "learning_rate": 1.410618715199319e-08, "loss": 0.4771, "step": 18891 }, { "epoch": 2.95, "grad_norm": 31.61028932803755, "learning_rate": 1.4016787118383835e-08, "loss": 0.4269, "step": 18892 }, { "epoch": 2.95, "grad_norm": 14.811103497941325, "learning_rate": 1.3927671080294913e-08, "loss": 0.4185, "step": 18893 }, { "epoch": 2.95, "grad_norm": 28.70594859420759, "learning_rate": 1.3838839040262175e-08, "loss": 0.4399, "step": 18894 }, { "epoch": 2.95, "grad_norm": 15.81090350158245, "learning_rate": 1.3750291000811377e-08, "loss": 0.4383, "step": 18895 }, { "epoch": 2.95, "grad_norm": 17.86401355509483, "learning_rate": 1.3662026964459396e-08, "loss": 0.3881, "step": 18896 }, { "epoch": 2.95, "grad_norm": 17.640748512639835, "learning_rate": 1.3574046933717556e-08, "loss": 0.3694, "step": 18897 }, { "epoch": 2.95, "grad_norm": 21.11328455583545, "learning_rate": 1.3486350911089407e-08, "loss": 0.3511, "step": 18898 }, { "epoch": 2.95, "grad_norm": 18.487984189649875, "learning_rate": 1.3398938899066294e-08, "loss": 0.401, "step": 18899 }, { "epoch": 2.95, "grad_norm": 17.72068262240934, "learning_rate": 1.3311810900135113e-08, "loss": 0.4295, "step": 18900 }, { "epoch": 2.95, "grad_norm": 18.793717919091723, "learning_rate": 1.3224966916774995e-08, "loss": 0.4255, "step": 18901 }, { "epoch": 2.95, "grad_norm": 25.787911773959113, "learning_rate": 1.3138406951453963e-08, "loss": 0.4231, "step": 18902 }, { "epoch": 2.95, "grad_norm": 23.04411393259211, "learning_rate": 1.3052131006634494e-08, "loss": 0.4177, "step": 18903 }, { "epoch": 2.95, "grad_norm": 20.70476230121021, "learning_rate": 1.296613908477018e-08, "loss": 0.4449, "step": 18904 }, { "epoch": 2.95, "grad_norm": 18.640186251891148, "learning_rate": 1.2880431188306841e-08, "loss": 0.429, "step": 18905 }, { "epoch": 2.95, "grad_norm": 47.296971730264694, "learning_rate": 1.2795007319681418e-08, "loss": 0.4389, "step": 18906 }, { "epoch": 2.95, "grad_norm": 20.84662707402527, "learning_rate": 1.270986748132419e-08, "loss": 0.4326, "step": 18907 }, { "epoch": 2.95, "grad_norm": 32.03667124438352, "learning_rate": 1.262501167565655e-08, "loss": 0.4631, "step": 18908 }, { "epoch": 2.95, "grad_norm": 24.41483813343104, "learning_rate": 1.2540439905089907e-08, "loss": 0.4659, "step": 18909 }, { "epoch": 2.95, "grad_norm": 26.103207194699998, "learning_rate": 1.2456152172031222e-08, "loss": 0.4136, "step": 18910 }, { "epoch": 2.95, "grad_norm": 22.227457055994396, "learning_rate": 1.2372148478876356e-08, "loss": 0.3986, "step": 18911 }, { "epoch": 2.95, "grad_norm": 27.45741520807344, "learning_rate": 1.228842882801562e-08, "loss": 0.419, "step": 18912 }, { "epoch": 2.95, "grad_norm": 30.814716971225845, "learning_rate": 1.2204993221829331e-08, "loss": 0.412, "step": 18913 }, { "epoch": 2.95, "grad_norm": 18.874521113630074, "learning_rate": 1.2121841662690037e-08, "loss": 0.4038, "step": 18914 }, { "epoch": 2.95, "grad_norm": 21.780510409590452, "learning_rate": 1.2038974152963623e-08, "loss": 0.4103, "step": 18915 }, { "epoch": 2.95, "grad_norm": 37.31944332104858, "learning_rate": 1.1956390695004871e-08, "loss": 0.4795, "step": 18916 }, { "epoch": 2.95, "grad_norm": 15.891607859925616, "learning_rate": 1.1874091291164125e-08, "loss": 0.4385, "step": 18917 }, { "epoch": 2.96, "grad_norm": 26.307133652937484, "learning_rate": 1.1792075943781733e-08, "loss": 0.431, "step": 18918 }, { "epoch": 2.96, "grad_norm": 22.552838197163407, "learning_rate": 1.1710344655189166e-08, "loss": 0.422, "step": 18919 }, { "epoch": 2.96, "grad_norm": 17.897930163882112, "learning_rate": 1.1628897427711227e-08, "loss": 0.4156, "step": 18920 }, { "epoch": 2.96, "grad_norm": 20.86070074197505, "learning_rate": 1.1547734263664957e-08, "loss": 0.3833, "step": 18921 }, { "epoch": 2.96, "grad_norm": 19.380461471426464, "learning_rate": 1.1466855165357393e-08, "loss": 0.4023, "step": 18922 }, { "epoch": 2.96, "grad_norm": 17.83093619687225, "learning_rate": 1.1386260135090033e-08, "loss": 0.4547, "step": 18923 }, { "epoch": 2.96, "grad_norm": 16.225807193943098, "learning_rate": 1.1305949175154374e-08, "loss": 0.4514, "step": 18924 }, { "epoch": 2.96, "grad_norm": 25.397365373777898, "learning_rate": 1.1225922287834146e-08, "loss": 0.4352, "step": 18925 }, { "epoch": 2.96, "grad_norm": 28.502076288768954, "learning_rate": 1.1146179475404195e-08, "loss": 0.5079, "step": 18926 }, { "epoch": 2.96, "grad_norm": 22.354008033831885, "learning_rate": 1.1066720740134928e-08, "loss": 0.4392, "step": 18927 }, { "epoch": 2.96, "grad_norm": 26.87022337714897, "learning_rate": 1.098754608428454e-08, "loss": 0.4118, "step": 18928 }, { "epoch": 2.96, "grad_norm": 23.40075480046642, "learning_rate": 1.090865551010456e-08, "loss": 0.4529, "step": 18929 }, { "epoch": 2.96, "grad_norm": 17.644945994855508, "learning_rate": 1.0830049019839862e-08, "loss": 0.3868, "step": 18930 }, { "epoch": 2.96, "grad_norm": 28.6227328337586, "learning_rate": 1.0751726615724212e-08, "loss": 0.4741, "step": 18931 }, { "epoch": 2.96, "grad_norm": 17.12361973400038, "learning_rate": 1.0673688299985829e-08, "loss": 0.4264, "step": 18932 }, { "epoch": 2.96, "grad_norm": 31.25673761405072, "learning_rate": 1.059593407484516e-08, "loss": 0.4484, "step": 18933 }, { "epoch": 2.96, "grad_norm": 17.018648184647915, "learning_rate": 1.0518463942511547e-08, "loss": 0.4339, "step": 18934 }, { "epoch": 2.96, "grad_norm": 20.877510915580245, "learning_rate": 1.0441277905188785e-08, "loss": 0.4812, "step": 18935 }, { "epoch": 2.96, "grad_norm": 21.089566961833093, "learning_rate": 1.0364375965074003e-08, "loss": 0.4252, "step": 18936 }, { "epoch": 2.96, "grad_norm": 21.504862675545432, "learning_rate": 1.0287758124351011e-08, "loss": 0.3679, "step": 18937 }, { "epoch": 2.96, "grad_norm": 32.761000417391585, "learning_rate": 1.0211424385201396e-08, "loss": 0.4554, "step": 18938 }, { "epoch": 2.96, "grad_norm": 24.312658174583717, "learning_rate": 1.0135374749794536e-08, "loss": 0.4521, "step": 18939 }, { "epoch": 2.96, "grad_norm": 15.280678559951753, "learning_rate": 1.0059609220293142e-08, "loss": 0.4338, "step": 18940 }, { "epoch": 2.96, "grad_norm": 17.874495016480942, "learning_rate": 9.984127798853271e-09, "loss": 0.3965, "step": 18941 }, { "epoch": 2.96, "grad_norm": 24.805881613276036, "learning_rate": 9.908930487620983e-09, "loss": 0.3996, "step": 18942 }, { "epoch": 2.96, "grad_norm": 31.750902270559063, "learning_rate": 9.834017288734565e-09, "loss": 0.4024, "step": 18943 }, { "epoch": 2.96, "grad_norm": 21.997023547056152, "learning_rate": 9.759388204323428e-09, "loss": 0.4377, "step": 18944 }, { "epoch": 2.96, "grad_norm": 18.637025379256198, "learning_rate": 9.685043236512537e-09, "loss": 0.3674, "step": 18945 }, { "epoch": 2.96, "grad_norm": 25.995770388912867, "learning_rate": 9.610982387414647e-09, "loss": 0.4357, "step": 18946 }, { "epoch": 2.96, "grad_norm": 32.8932047350986, "learning_rate": 9.53720565913585e-09, "loss": 0.4433, "step": 18947 }, { "epoch": 2.96, "grad_norm": 24.27065160741977, "learning_rate": 9.46371305377447e-09, "loss": 0.4437, "step": 18948 }, { "epoch": 2.96, "grad_norm": 21.935168933779824, "learning_rate": 9.390504573422166e-09, "loss": 0.4225, "step": 18949 }, { "epoch": 2.96, "grad_norm": 16.20277267742511, "learning_rate": 9.317580220158385e-09, "loss": 0.3808, "step": 18950 }, { "epoch": 2.96, "grad_norm": 24.955198566677257, "learning_rate": 9.244939996058134e-09, "loss": 0.4451, "step": 18951 }, { "epoch": 2.96, "grad_norm": 19.38059984314208, "learning_rate": 9.172583903187537e-09, "loss": 0.4397, "step": 18952 }, { "epoch": 2.96, "grad_norm": 14.94150959851793, "learning_rate": 9.10051194360495e-09, "loss": 0.3739, "step": 18953 }, { "epoch": 2.96, "grad_norm": 23.468157869988666, "learning_rate": 9.028724119358734e-09, "loss": 0.4121, "step": 18954 }, { "epoch": 2.96, "grad_norm": 27.10669632905505, "learning_rate": 8.957220432490587e-09, "loss": 0.4185, "step": 18955 }, { "epoch": 2.96, "grad_norm": 24.47403983111849, "learning_rate": 8.886000885034441e-09, "loss": 0.4557, "step": 18956 }, { "epoch": 2.96, "grad_norm": 16.38599176442435, "learning_rate": 8.815065479016449e-09, "loss": 0.4632, "step": 18957 }, { "epoch": 2.96, "grad_norm": 22.903400396116684, "learning_rate": 8.74441421645278e-09, "loss": 0.4089, "step": 18958 }, { "epoch": 2.96, "grad_norm": 24.430226858935495, "learning_rate": 8.674047099354044e-09, "loss": 0.4088, "step": 18959 }, { "epoch": 2.96, "grad_norm": 21.816443375817986, "learning_rate": 8.603964129719755e-09, "loss": 0.4056, "step": 18960 }, { "epoch": 2.96, "grad_norm": 29.52230697530616, "learning_rate": 8.534165309544985e-09, "loss": 0.422, "step": 18961 }, { "epoch": 2.96, "grad_norm": 18.756022535125073, "learning_rate": 8.464650640812588e-09, "loss": 0.4325, "step": 18962 }, { "epoch": 2.96, "grad_norm": 22.615046005459465, "learning_rate": 8.395420125500986e-09, "loss": 0.4399, "step": 18963 }, { "epoch": 2.96, "grad_norm": 37.68698940714089, "learning_rate": 8.326473765579713e-09, "loss": 0.4363, "step": 18964 }, { "epoch": 2.96, "grad_norm": 19.227217201578128, "learning_rate": 8.2578115630072e-09, "loss": 0.4266, "step": 18965 }, { "epoch": 2.96, "grad_norm": 29.493925372324, "learning_rate": 8.189433519738555e-09, "loss": 0.4452, "step": 18966 }, { "epoch": 2.96, "grad_norm": 19.750361215933907, "learning_rate": 8.121339637716663e-09, "loss": 0.4659, "step": 18967 }, { "epoch": 2.96, "grad_norm": 21.051687452921847, "learning_rate": 8.053529918878867e-09, "loss": 0.4494, "step": 18968 }, { "epoch": 2.96, "grad_norm": 27.385109292214995, "learning_rate": 7.986004365153621e-09, "loss": 0.5378, "step": 18969 }, { "epoch": 2.96, "grad_norm": 22.2300064321709, "learning_rate": 7.918762978461613e-09, "loss": 0.4553, "step": 18970 }, { "epoch": 2.96, "grad_norm": 20.503338150930492, "learning_rate": 7.851805760714647e-09, "loss": 0.4762, "step": 18971 }, { "epoch": 2.96, "grad_norm": 20.896910256170234, "learning_rate": 7.785132713816756e-09, "loss": 0.389, "step": 18972 }, { "epoch": 2.96, "grad_norm": 20.76167807373196, "learning_rate": 7.7187438396642e-09, "loss": 0.4352, "step": 18973 }, { "epoch": 2.96, "grad_norm": 16.797145053999174, "learning_rate": 7.652639140146579e-09, "loss": 0.4545, "step": 18974 }, { "epoch": 2.96, "grad_norm": 14.602992906713823, "learning_rate": 7.586818617141279e-09, "loss": 0.462, "step": 18975 }, { "epoch": 2.96, "grad_norm": 21.221055704374898, "learning_rate": 7.521282272521246e-09, "loss": 0.4316, "step": 18976 }, { "epoch": 2.96, "grad_norm": 26.49754136753965, "learning_rate": 7.456030108151657e-09, "loss": 0.4585, "step": 18977 }, { "epoch": 2.96, "grad_norm": 31.44579497447003, "learning_rate": 7.391062125886583e-09, "loss": 0.4044, "step": 18978 }, { "epoch": 2.96, "grad_norm": 27.599035151131698, "learning_rate": 7.326378327574546e-09, "loss": 0.3935, "step": 18979 }, { "epoch": 2.96, "grad_norm": 18.12478616851425, "learning_rate": 7.261978715054074e-09, "loss": 0.4299, "step": 18980 }, { "epoch": 2.96, "grad_norm": 23.206302658995273, "learning_rate": 7.197863290157037e-09, "loss": 0.4947, "step": 18981 }, { "epoch": 2.97, "grad_norm": 31.141215870998966, "learning_rate": 7.134032054707529e-09, "loss": 0.4481, "step": 18982 }, { "epoch": 2.97, "grad_norm": 31.301658485732723, "learning_rate": 7.070485010520767e-09, "loss": 0.4303, "step": 18983 }, { "epoch": 2.97, "grad_norm": 19.564166917408766, "learning_rate": 7.007222159404192e-09, "loss": 0.4189, "step": 18984 }, { "epoch": 2.97, "grad_norm": 16.430214103897324, "learning_rate": 6.944243503155257e-09, "loss": 0.4579, "step": 18985 }, { "epoch": 2.97, "grad_norm": 22.445940848722614, "learning_rate": 6.881549043568081e-09, "loss": 0.4568, "step": 18986 }, { "epoch": 2.97, "grad_norm": 22.023330939915716, "learning_rate": 6.819138782422352e-09, "loss": 0.4271, "step": 18987 }, { "epoch": 2.97, "grad_norm": 20.884309167028974, "learning_rate": 6.757012721494427e-09, "loss": 0.4144, "step": 18988 }, { "epoch": 2.97, "grad_norm": 24.605609217544227, "learning_rate": 6.6951708625517805e-09, "loss": 0.4263, "step": 18989 }, { "epoch": 2.97, "grad_norm": 45.597811336967645, "learning_rate": 6.633613207351897e-09, "loss": 0.4614, "step": 18990 }, { "epoch": 2.97, "grad_norm": 18.01267228656309, "learning_rate": 6.5723397576467066e-09, "loss": 0.4038, "step": 18991 }, { "epoch": 2.97, "grad_norm": 27.522545004457985, "learning_rate": 6.51135051517815e-09, "loss": 0.4144, "step": 18992 }, { "epoch": 2.97, "grad_norm": 22.75316349110637, "learning_rate": 6.450645481681506e-09, "loss": 0.3934, "step": 18993 }, { "epoch": 2.97, "grad_norm": 15.503631236625589, "learning_rate": 6.39022465888095e-09, "loss": 0.3806, "step": 18994 }, { "epoch": 2.97, "grad_norm": 20.616927052392633, "learning_rate": 6.330088048497329e-09, "loss": 0.4745, "step": 18995 }, { "epoch": 2.97, "grad_norm": 24.368009699544025, "learning_rate": 6.270235652239276e-09, "loss": 0.4141, "step": 18996 }, { "epoch": 2.97, "grad_norm": 17.230089111227763, "learning_rate": 6.2106674718098725e-09, "loss": 0.3998, "step": 18997 }, { "epoch": 2.97, "grad_norm": 20.98243307727471, "learning_rate": 6.1513835089033194e-09, "loss": 0.4365, "step": 18998 }, { "epoch": 2.97, "grad_norm": 18.82280379797896, "learning_rate": 6.0923837652038245e-09, "loss": 0.3938, "step": 18999 }, { "epoch": 2.97, "grad_norm": 25.494579999640557, "learning_rate": 6.033668242392266e-09, "loss": 0.4302, "step": 19000 }, { "epoch": 2.97, "grad_norm": 22.20595285419225, "learning_rate": 5.975236942136197e-09, "loss": 0.3808, "step": 19001 }, { "epoch": 2.97, "grad_norm": 30.59397621041931, "learning_rate": 5.917089866097625e-09, "loss": 0.4022, "step": 19002 }, { "epoch": 2.97, "grad_norm": 31.912210014301774, "learning_rate": 5.85922701593078e-09, "loss": 0.4656, "step": 19003 }, { "epoch": 2.97, "grad_norm": 27.910059455826534, "learning_rate": 5.801648393282122e-09, "loss": 0.4577, "step": 19004 }, { "epoch": 2.97, "grad_norm": 18.43865478584657, "learning_rate": 5.744353999787011e-09, "loss": 0.4491, "step": 19005 }, { "epoch": 2.97, "grad_norm": 20.1279621980003, "learning_rate": 5.6873438370763645e-09, "loss": 0.4439, "step": 19006 }, { "epoch": 2.97, "grad_norm": 15.487279225054342, "learning_rate": 5.630617906771108e-09, "loss": 0.4437, "step": 19007 }, { "epoch": 2.97, "grad_norm": 36.774827812953504, "learning_rate": 5.574176210484394e-09, "loss": 0.4529, "step": 19008 }, { "epoch": 2.97, "grad_norm": 35.57147601809623, "learning_rate": 5.518018749821607e-09, "loss": 0.4119, "step": 19009 }, { "epoch": 2.97, "grad_norm": 19.362953589589303, "learning_rate": 5.462145526380358e-09, "loss": 0.3986, "step": 19010 }, { "epoch": 2.97, "grad_norm": 16.30089443632787, "learning_rate": 5.406556541748265e-09, "loss": 0.3757, "step": 19011 }, { "epoch": 2.97, "grad_norm": 26.287451505684306, "learning_rate": 5.351251797507395e-09, "loss": 0.5059, "step": 19012 }, { "epoch": 2.97, "grad_norm": 27.050379262202735, "learning_rate": 5.296231295229826e-09, "loss": 0.4001, "step": 19013 }, { "epoch": 2.97, "grad_norm": 24.09386004708536, "learning_rate": 5.2414950364809706e-09, "loss": 0.4343, "step": 19014 }, { "epoch": 2.97, "grad_norm": 19.7126359657613, "learning_rate": 5.187043022817362e-09, "loss": 0.398, "step": 19015 }, { "epoch": 2.97, "grad_norm": 22.009488539272215, "learning_rate": 5.1328752557877615e-09, "loss": 0.4432, "step": 19016 }, { "epoch": 2.97, "grad_norm": 18.44981720515287, "learning_rate": 5.078991736932049e-09, "loss": 0.4513, "step": 19017 }, { "epoch": 2.97, "grad_norm": 20.877136236803327, "learning_rate": 5.025392467783441e-09, "loss": 0.427, "step": 19018 }, { "epoch": 2.97, "grad_norm": 27.957027941121694, "learning_rate": 4.972077449865165e-09, "loss": 0.43, "step": 19019 }, { "epoch": 2.97, "grad_norm": 17.118143936799132, "learning_rate": 4.919046684693785e-09, "loss": 0.4203, "step": 19020 }, { "epoch": 2.97, "grad_norm": 21.99521631893356, "learning_rate": 4.8663001737780934e-09, "loss": 0.4202, "step": 19021 }, { "epoch": 2.97, "grad_norm": 17.424257384974354, "learning_rate": 4.813837918618003e-09, "loss": 0.4745, "step": 19022 }, { "epoch": 2.97, "grad_norm": 20.060107964969642, "learning_rate": 4.761659920705652e-09, "loss": 0.4413, "step": 19023 }, { "epoch": 2.97, "grad_norm": 19.837777820509434, "learning_rate": 4.7097661815243e-09, "loss": 0.4484, "step": 19024 }, { "epoch": 2.97, "grad_norm": 13.61512186914786, "learning_rate": 4.6581567025494324e-09, "loss": 0.3757, "step": 19025 }, { "epoch": 2.97, "grad_norm": 13.75563206851141, "learning_rate": 4.6068314852498745e-09, "loss": 0.386, "step": 19026 }, { "epoch": 2.97, "grad_norm": 21.740098357364158, "learning_rate": 4.555790531085569e-09, "loss": 0.4401, "step": 19027 }, { "epoch": 2.97, "grad_norm": 33.30524951882078, "learning_rate": 4.505033841506468e-09, "loss": 0.4379, "step": 19028 }, { "epoch": 2.97, "grad_norm": 28.644074719748602, "learning_rate": 4.454561417958081e-09, "loss": 0.4253, "step": 19029 }, { "epoch": 2.97, "grad_norm": 26.08593718125171, "learning_rate": 4.404373261873707e-09, "loss": 0.4307, "step": 19030 }, { "epoch": 2.97, "grad_norm": 29.210141061411946, "learning_rate": 4.354469374682202e-09, "loss": 0.4129, "step": 19031 }, { "epoch": 2.97, "grad_norm": 27.10043342670859, "learning_rate": 4.304849757802432e-09, "loss": 0.4332, "step": 19032 }, { "epoch": 2.97, "grad_norm": 50.35246038649922, "learning_rate": 4.25551441264549e-09, "loss": 0.4806, "step": 19033 }, { "epoch": 2.97, "grad_norm": 23.600085213828606, "learning_rate": 4.206463340614697e-09, "loss": 0.4317, "step": 19034 }, { "epoch": 2.97, "grad_norm": 23.15291259566251, "learning_rate": 4.157696543103384e-09, "loss": 0.3772, "step": 19035 }, { "epoch": 2.97, "grad_norm": 27.431742348623246, "learning_rate": 4.109214021500441e-09, "loss": 0.5211, "step": 19036 }, { "epoch": 2.97, "grad_norm": 28.368095301477446, "learning_rate": 4.061015777184762e-09, "loss": 0.4635, "step": 19037 }, { "epoch": 2.97, "grad_norm": 23.357818429175374, "learning_rate": 4.013101811525255e-09, "loss": 0.4779, "step": 19038 }, { "epoch": 2.97, "grad_norm": 22.078226363265927, "learning_rate": 3.9654721258863826e-09, "loss": 0.4067, "step": 19039 }, { "epoch": 2.97, "grad_norm": 24.644678762069717, "learning_rate": 3.918126721621507e-09, "loss": 0.4163, "step": 19040 }, { "epoch": 2.97, "grad_norm": 25.96143597960283, "learning_rate": 3.871065600078439e-09, "loss": 0.4695, "step": 19041 }, { "epoch": 2.97, "grad_norm": 14.725658046460605, "learning_rate": 3.824288762593886e-09, "loss": 0.3881, "step": 19042 }, { "epoch": 2.97, "grad_norm": 15.693133629704562, "learning_rate": 3.777796210499008e-09, "loss": 0.421, "step": 19043 }, { "epoch": 2.97, "grad_norm": 16.176605285607106, "learning_rate": 3.731587945116078e-09, "loss": 0.4373, "step": 19044 }, { "epoch": 2.97, "grad_norm": 26.68550790715355, "learning_rate": 3.6856639677584906e-09, "loss": 0.4516, "step": 19045 }, { "epoch": 2.98, "grad_norm": 22.38465464336222, "learning_rate": 3.640024279734089e-09, "loss": 0.4024, "step": 19046 }, { "epoch": 2.98, "grad_norm": 20.297508200502715, "learning_rate": 3.5946688823396137e-09, "loss": 0.4067, "step": 19047 }, { "epoch": 2.98, "grad_norm": 26.32188164123176, "learning_rate": 3.5495977768640332e-09, "loss": 0.5031, "step": 19048 }, { "epoch": 2.98, "grad_norm": 21.24744857870061, "learning_rate": 3.5048109645907657e-09, "loss": 0.4013, "step": 19049 }, { "epoch": 2.98, "grad_norm": 27.54048088587474, "learning_rate": 3.460308446793237e-09, "loss": 0.4876, "step": 19050 }, { "epoch": 2.98, "grad_norm": 18.753353146495563, "learning_rate": 3.4160902247359904e-09, "loss": 0.4494, "step": 19051 }, { "epoch": 2.98, "grad_norm": 30.05542718618167, "learning_rate": 3.37215629967802e-09, "loss": 0.4343, "step": 19052 }, { "epoch": 2.98, "grad_norm": 23.280379620866015, "learning_rate": 3.3285066728672156e-09, "loss": 0.473, "step": 19053 }, { "epoch": 2.98, "grad_norm": 15.254602517955334, "learning_rate": 3.285141345547027e-09, "loss": 0.3855, "step": 19054 }, { "epoch": 2.98, "grad_norm": 17.312123829063434, "learning_rate": 3.242060318948692e-09, "loss": 0.4466, "step": 19055 }, { "epoch": 2.98, "grad_norm": 22.590060792896477, "learning_rate": 3.1992635942978965e-09, "loss": 0.3997, "step": 19056 }, { "epoch": 2.98, "grad_norm": 30.826135758257536, "learning_rate": 3.1567511728125555e-09, "loss": 0.3865, "step": 19057 }, { "epoch": 2.98, "grad_norm": 17.03789455743406, "learning_rate": 3.114523055700591e-09, "loss": 0.3902, "step": 19058 }, { "epoch": 2.98, "grad_norm": 25.533856886818658, "learning_rate": 3.0725792441643755e-09, "loss": 0.3939, "step": 19059 }, { "epoch": 2.98, "grad_norm": 24.253246856882708, "learning_rate": 3.0309197393962874e-09, "loss": 0.3718, "step": 19060 }, { "epoch": 2.98, "grad_norm": 25.898124468538473, "learning_rate": 2.9895445425798253e-09, "loss": 0.4364, "step": 19061 }, { "epoch": 2.98, "grad_norm": 20.483069362247637, "learning_rate": 2.9484536548929355e-09, "loss": 0.4733, "step": 19062 }, { "epoch": 2.98, "grad_norm": 24.538813588830145, "learning_rate": 2.907647077504683e-09, "loss": 0.4989, "step": 19063 }, { "epoch": 2.98, "grad_norm": 16.220303074303274, "learning_rate": 2.86712481157414e-09, "loss": 0.4245, "step": 19064 }, { "epoch": 2.98, "grad_norm": 18.952393335275733, "learning_rate": 2.826886858253719e-09, "loss": 0.411, "step": 19065 }, { "epoch": 2.98, "grad_norm": 25.645906260133913, "learning_rate": 2.7869332186891697e-09, "loss": 0.4338, "step": 19066 }, { "epoch": 2.98, "grad_norm": 21.984469154633462, "learning_rate": 2.7472638940162498e-09, "loss": 0.4415, "step": 19067 }, { "epoch": 2.98, "grad_norm": 27.13863583892154, "learning_rate": 2.707878885362947e-09, "loss": 0.455, "step": 19068 }, { "epoch": 2.98, "grad_norm": 21.20115647423043, "learning_rate": 2.6687781938483647e-09, "loss": 0.389, "step": 19069 }, { "epoch": 2.98, "grad_norm": 37.3323743692061, "learning_rate": 2.629961820587168e-09, "loss": 0.4341, "step": 19070 }, { "epoch": 2.98, "grad_norm": 17.45510550013736, "learning_rate": 2.5914297666806975e-09, "loss": 0.46, "step": 19071 }, { "epoch": 2.98, "grad_norm": 16.590035204122856, "learning_rate": 2.5531820332247436e-09, "loss": 0.4174, "step": 19072 }, { "epoch": 2.98, "grad_norm": 26.41161283175662, "learning_rate": 2.515218621309545e-09, "loss": 0.4657, "step": 19073 }, { "epoch": 2.98, "grad_norm": 18.175893233524274, "learning_rate": 2.4775395320120187e-09, "loss": 0.4031, "step": 19074 }, { "epoch": 2.98, "grad_norm": 15.164030899003132, "learning_rate": 2.4401447664046398e-09, "loss": 0.3686, "step": 19075 }, { "epoch": 2.98, "grad_norm": 19.19627788989915, "learning_rate": 2.4030343255521115e-09, "loss": 0.3659, "step": 19076 }, { "epoch": 2.98, "grad_norm": 17.318520219648978, "learning_rate": 2.3662082105080365e-09, "loss": 0.436, "step": 19077 }, { "epoch": 2.98, "grad_norm": 19.873689941949166, "learning_rate": 2.3296664223215747e-09, "loss": 0.3835, "step": 19078 }, { "epoch": 2.98, "grad_norm": 25.057739024403343, "learning_rate": 2.2934089620307852e-09, "loss": 0.4465, "step": 19079 }, { "epoch": 2.98, "grad_norm": 26.07033242626989, "learning_rate": 2.257435830665955e-09, "loss": 0.426, "step": 19080 }, { "epoch": 2.98, "grad_norm": 18.889132040771294, "learning_rate": 2.2217470292518196e-09, "loss": 0.4156, "step": 19081 }, { "epoch": 2.98, "grad_norm": 33.06543095474694, "learning_rate": 2.1863425588020125e-09, "loss": 0.4709, "step": 19082 }, { "epoch": 2.98, "grad_norm": 22.210637933547137, "learning_rate": 2.1512224203246167e-09, "loss": 0.4326, "step": 19083 }, { "epoch": 2.98, "grad_norm": 19.713890987628858, "learning_rate": 2.116386614818833e-09, "loss": 0.3843, "step": 19084 }, { "epoch": 2.98, "grad_norm": 18.52326693027949, "learning_rate": 2.081835143272759e-09, "loss": 0.5101, "step": 19085 }, { "epoch": 2.98, "grad_norm": 32.725982118124115, "learning_rate": 2.0475680066711632e-09, "loss": 0.4518, "step": 19086 }, { "epoch": 2.98, "grad_norm": 21.844984637720195, "learning_rate": 2.0135852059888215e-09, "loss": 0.4331, "step": 19087 }, { "epoch": 2.98, "grad_norm": 23.3869577545039, "learning_rate": 1.9798867421905176e-09, "loss": 0.4327, "step": 19088 }, { "epoch": 2.98, "grad_norm": 22.035357402173492, "learning_rate": 1.9464726162365944e-09, "loss": 0.4743, "step": 19089 }, { "epoch": 2.98, "grad_norm": 31.04858755985672, "learning_rate": 1.9133428290751822e-09, "loss": 0.4268, "step": 19090 }, { "epoch": 2.98, "grad_norm": 29.413243408754294, "learning_rate": 1.880497381651081e-09, "loss": 0.4537, "step": 19091 }, { "epoch": 2.98, "grad_norm": 21.444012303361635, "learning_rate": 1.8479362748957674e-09, "loss": 0.4457, "step": 19092 }, { "epoch": 2.98, "grad_norm": 24.13550402604632, "learning_rate": 1.8156595097362783e-09, "loss": 0.43, "step": 19093 }, { "epoch": 2.98, "grad_norm": 14.194137590752707, "learning_rate": 1.783667087090768e-09, "loss": 0.3687, "step": 19094 }, { "epoch": 2.98, "grad_norm": 13.481841177483908, "learning_rate": 1.7519590078696192e-09, "loss": 0.4039, "step": 19095 }, { "epoch": 2.98, "grad_norm": 25.700649642595774, "learning_rate": 1.7205352729732227e-09, "loss": 0.4307, "step": 19096 }, { "epoch": 2.98, "grad_norm": 24.809557723470284, "learning_rate": 1.6893958832964186e-09, "loss": 0.4146, "step": 19097 }, { "epoch": 2.98, "grad_norm": 26.156701973914807, "learning_rate": 1.6585408397240544e-09, "loss": 0.4533, "step": 19098 }, { "epoch": 2.98, "grad_norm": 19.22753737224552, "learning_rate": 1.6279701431343165e-09, "loss": 0.4442, "step": 19099 }, { "epoch": 2.98, "grad_norm": 26.227705283745607, "learning_rate": 1.5976837943953993e-09, "loss": 0.4315, "step": 19100 }, { "epoch": 2.98, "grad_norm": 22.538953271135078, "learning_rate": 1.5676817943699462e-09, "loss": 0.3854, "step": 19101 }, { "epoch": 2.98, "grad_norm": 22.77952149718791, "learning_rate": 1.5379641439106086e-09, "loss": 0.4179, "step": 19102 }, { "epoch": 2.98, "grad_norm": 17.82095976092635, "learning_rate": 1.5085308438633762e-09, "loss": 0.4657, "step": 19103 }, { "epoch": 2.98, "grad_norm": 26.558446492963586, "learning_rate": 1.4793818950642469e-09, "loss": 0.4205, "step": 19104 }, { "epoch": 2.98, "grad_norm": 29.937010316586935, "learning_rate": 1.4505172983425576e-09, "loss": 0.4217, "step": 19105 }, { "epoch": 2.98, "grad_norm": 26.99159863846815, "learning_rate": 1.4219370545187628e-09, "loss": 0.4263, "step": 19106 }, { "epoch": 2.98, "grad_norm": 26.506005887529, "learning_rate": 1.393641164405546e-09, "loss": 0.4655, "step": 19107 }, { "epoch": 2.98, "grad_norm": 25.773095289087827, "learning_rate": 1.3656296288089288e-09, "loss": 0.4774, "step": 19108 }, { "epoch": 2.98, "grad_norm": 16.211623797309347, "learning_rate": 1.3379024485249414e-09, "loss": 0.4244, "step": 19109 }, { "epoch": 2.99, "grad_norm": 21.072265779157753, "learning_rate": 1.310459624341842e-09, "loss": 0.3614, "step": 19110 }, { "epoch": 2.99, "grad_norm": 23.334449939676414, "learning_rate": 1.2833011570401177e-09, "loss": 0.4664, "step": 19111 }, { "epoch": 2.99, "grad_norm": 17.51298214015917, "learning_rate": 1.2564270473924834e-09, "loss": 0.4528, "step": 19112 }, { "epoch": 2.99, "grad_norm": 23.299673938899996, "learning_rate": 1.2298372961627726e-09, "loss": 0.393, "step": 19113 }, { "epoch": 2.99, "grad_norm": 19.474701590143233, "learning_rate": 1.203531904107047e-09, "loss": 0.4197, "step": 19114 }, { "epoch": 2.99, "grad_norm": 14.48891243957197, "learning_rate": 1.1775108719735973e-09, "loss": 0.4491, "step": 19115 }, { "epoch": 2.99, "grad_norm": 17.8203287279255, "learning_rate": 1.1517742005029419e-09, "loss": 0.4563, "step": 19116 }, { "epoch": 2.99, "grad_norm": 24.33407077518308, "learning_rate": 1.1263218904267181e-09, "loss": 0.4634, "step": 19117 }, { "epoch": 2.99, "grad_norm": 23.749761730805847, "learning_rate": 1.101153942468791e-09, "loss": 0.4136, "step": 19118 }, { "epoch": 2.99, "grad_norm": 19.71688178895801, "learning_rate": 1.0762703573452548e-09, "loss": 0.4013, "step": 19119 }, { "epoch": 2.99, "grad_norm": 23.78527307129287, "learning_rate": 1.051671135762211e-09, "loss": 0.4017, "step": 19120 }, { "epoch": 2.99, "grad_norm": 17.272434161483535, "learning_rate": 1.0273562784213209e-09, "loss": 0.45, "step": 19121 }, { "epoch": 2.99, "grad_norm": 31.26487638863152, "learning_rate": 1.0033257860131428e-09, "loss": 0.4873, "step": 19122 }, { "epoch": 2.99, "grad_norm": 28.505163860036593, "learning_rate": 9.795796592204643e-10, "loss": 0.4282, "step": 19123 }, { "epoch": 2.99, "grad_norm": 17.571739471281944, "learning_rate": 9.561178987205211e-10, "loss": 0.4026, "step": 19124 }, { "epoch": 2.99, "grad_norm": 15.528194327419424, "learning_rate": 9.32940505178337e-10, "loss": 0.4177, "step": 19125 }, { "epoch": 2.99, "grad_norm": 31.210650128317532, "learning_rate": 9.100474792544944e-10, "loss": 0.4726, "step": 19126 }, { "epoch": 2.99, "grad_norm": 33.317345870618894, "learning_rate": 8.874388215995844e-10, "loss": 0.4645, "step": 19127 }, { "epoch": 2.99, "grad_norm": 20.593244654948567, "learning_rate": 8.651145328564259e-10, "loss": 0.4757, "step": 19128 }, { "epoch": 2.99, "grad_norm": 21.82547565416537, "learning_rate": 8.430746136600665e-10, "loss": 0.4405, "step": 19129 }, { "epoch": 2.99, "grad_norm": 23.637793451431314, "learning_rate": 8.21319064636672e-10, "loss": 0.4178, "step": 19130 }, { "epoch": 2.99, "grad_norm": 28.063395681827846, "learning_rate": 7.998478864068571e-10, "loss": 0.4538, "step": 19131 }, { "epoch": 2.99, "grad_norm": 22.113518286915223, "learning_rate": 7.786610795790239e-10, "loss": 0.3897, "step": 19132 }, { "epoch": 2.99, "grad_norm": 17.41134479960007, "learning_rate": 7.577586447571339e-10, "loss": 0.4371, "step": 19133 }, { "epoch": 2.99, "grad_norm": 20.617627286972553, "learning_rate": 7.371405825351563e-10, "loss": 0.388, "step": 19134 }, { "epoch": 2.99, "grad_norm": 33.88056126214513, "learning_rate": 7.168068934992889e-10, "loss": 0.5159, "step": 19135 }, { "epoch": 2.99, "grad_norm": 22.62335761705785, "learning_rate": 6.967575782279579e-10, "loss": 0.3901, "step": 19136 }, { "epoch": 2.99, "grad_norm": 17.51772474035553, "learning_rate": 6.769926372918178e-10, "loss": 0.3448, "step": 19137 }, { "epoch": 2.99, "grad_norm": 26.582755048821255, "learning_rate": 6.575120712526417e-10, "loss": 0.4743, "step": 19138 }, { "epoch": 2.99, "grad_norm": 25.05416605983395, "learning_rate": 6.383158806644307e-10, "loss": 0.4437, "step": 19139 }, { "epoch": 2.99, "grad_norm": 27.444541828032182, "learning_rate": 6.194040660723044e-10, "loss": 0.3962, "step": 19140 }, { "epoch": 2.99, "grad_norm": 14.257702662423737, "learning_rate": 6.007766280158312e-10, "loss": 0.3867, "step": 19141 }, { "epoch": 2.99, "grad_norm": 27.69013418579101, "learning_rate": 5.824335670234771e-10, "loss": 0.4224, "step": 19142 }, { "epoch": 2.99, "grad_norm": 20.327416924438346, "learning_rate": 5.643748836170471e-10, "loss": 0.3976, "step": 19143 }, { "epoch": 2.99, "grad_norm": 21.44905646809337, "learning_rate": 5.466005783105743e-10, "loss": 0.4313, "step": 19144 }, { "epoch": 2.99, "grad_norm": 17.317397559300634, "learning_rate": 5.291106516092103e-10, "loss": 0.4331, "step": 19145 }, { "epoch": 2.99, "grad_norm": 26.386520590556934, "learning_rate": 5.119051040092249e-10, "loss": 0.3946, "step": 19146 }, { "epoch": 2.99, "grad_norm": 17.22320143247084, "learning_rate": 4.949839360024467e-10, "loss": 0.4299, "step": 19147 }, { "epoch": 2.99, "grad_norm": 17.268333595438364, "learning_rate": 4.78347148068492e-10, "loss": 0.4069, "step": 19148 }, { "epoch": 2.99, "grad_norm": 14.636166646912105, "learning_rate": 4.6199474068031603e-10, "loss": 0.4332, "step": 19149 }, { "epoch": 2.99, "grad_norm": 37.888977622175, "learning_rate": 4.459267143042123e-10, "loss": 0.538, "step": 19150 }, { "epoch": 2.99, "grad_norm": 18.815274491028674, "learning_rate": 4.301430693964825e-10, "loss": 0.4236, "step": 19151 }, { "epoch": 2.99, "grad_norm": 19.45068957085827, "learning_rate": 4.1464380640565683e-10, "loss": 0.3914, "step": 19152 }, { "epoch": 2.99, "grad_norm": 25.674100075925043, "learning_rate": 3.994289257724937e-10, "loss": 0.3994, "step": 19153 }, { "epoch": 2.99, "grad_norm": 19.993605566397438, "learning_rate": 3.8449842792998016e-10, "loss": 0.4182, "step": 19154 }, { "epoch": 2.99, "grad_norm": 16.46876706298002, "learning_rate": 3.698523133022214e-10, "loss": 0.4588, "step": 19155 }, { "epoch": 2.99, "grad_norm": 17.780245249741547, "learning_rate": 3.5549058230777143e-10, "loss": 0.3533, "step": 19156 }, { "epoch": 2.99, "grad_norm": 20.89951575376462, "learning_rate": 3.4141323535186175e-10, "loss": 0.39, "step": 19157 }, { "epoch": 2.99, "grad_norm": 15.106894907918312, "learning_rate": 3.276202728375033e-10, "loss": 0.4479, "step": 19158 }, { "epoch": 2.99, "grad_norm": 23.51762769795409, "learning_rate": 3.1411169515549457e-10, "loss": 0.3296, "step": 19159 }, { "epoch": 2.99, "grad_norm": 14.507648721817132, "learning_rate": 3.0088750269108293e-10, "loss": 0.413, "step": 19160 }, { "epoch": 2.99, "grad_norm": 22.44856247441304, "learning_rate": 2.8794769581952376e-10, "loss": 0.4144, "step": 19161 }, { "epoch": 2.99, "grad_norm": 26.166302820269554, "learning_rate": 2.7529227490941115e-10, "loss": 0.4075, "step": 19162 }, { "epoch": 2.99, "grad_norm": 20.290964703556106, "learning_rate": 2.629212403193471e-10, "loss": 0.4244, "step": 19163 }, { "epoch": 2.99, "grad_norm": 26.950469954927932, "learning_rate": 2.508345924023825e-10, "loss": 0.4177, "step": 19164 }, { "epoch": 2.99, "grad_norm": 26.53453770813269, "learning_rate": 2.390323315026866e-10, "loss": 0.4609, "step": 19165 }, { "epoch": 2.99, "grad_norm": 25.715568517200367, "learning_rate": 2.2751445795443638e-10, "loss": 0.493, "step": 19166 }, { "epoch": 2.99, "grad_norm": 25.548141875046, "learning_rate": 2.162809720873682e-10, "loss": 0.4288, "step": 19167 }, { "epoch": 2.99, "grad_norm": 15.212979452481388, "learning_rate": 2.0533187421789557e-10, "loss": 0.392, "step": 19168 }, { "epoch": 2.99, "grad_norm": 24.534495238661925, "learning_rate": 1.9466716466021162e-10, "loss": 0.3738, "step": 19169 }, { "epoch": 2.99, "grad_norm": 14.67929715322247, "learning_rate": 1.84286843716297e-10, "loss": 0.4557, "step": 19170 }, { "epoch": 2.99, "grad_norm": 20.541738157376844, "learning_rate": 1.7419091168036085e-10, "loss": 0.3638, "step": 19171 }, { "epoch": 2.99, "grad_norm": 18.0859665938405, "learning_rate": 1.6437936884217132e-10, "loss": 0.4544, "step": 19172 }, { "epoch": 2.99, "grad_norm": 23.817494495637863, "learning_rate": 1.5485221547817396e-10, "loss": 0.4007, "step": 19173 }, { "epoch": 3.0, "grad_norm": 26.901047592987165, "learning_rate": 1.4560945186148366e-10, "loss": 0.3542, "step": 19174 }, { "epoch": 3.0, "grad_norm": 18.6928571618243, "learning_rate": 1.3665107825300284e-10, "loss": 0.4492, "step": 19175 }, { "epoch": 3.0, "grad_norm": 29.834554139735523, "learning_rate": 1.2797709490919296e-10, "loss": 0.4039, "step": 19176 }, { "epoch": 3.0, "grad_norm": 22.011335605512496, "learning_rate": 1.1958750207541337e-10, "loss": 0.3808, "step": 19177 }, { "epoch": 3.0, "grad_norm": 22.482615369557795, "learning_rate": 1.114822999914722e-10, "loss": 0.4353, "step": 19178 }, { "epoch": 3.0, "grad_norm": 23.451599940045742, "learning_rate": 1.0366148888607541e-10, "loss": 0.4419, "step": 19179 }, { "epoch": 3.0, "grad_norm": 18.674575316721064, "learning_rate": 9.612506898348805e-11, "loss": 0.475, "step": 19180 }, { "epoch": 3.0, "grad_norm": 20.415924631877793, "learning_rate": 8.887304049798317e-11, "loss": 0.432, "step": 19181 }, { "epoch": 3.0, "grad_norm": 24.395616023162404, "learning_rate": 8.190540363384181e-11, "loss": 0.4367, "step": 19182 }, { "epoch": 3.0, "grad_norm": 20.142571665203263, "learning_rate": 7.522215859090409e-11, "loss": 0.4019, "step": 19183 }, { "epoch": 3.0, "grad_norm": 30.525650095070272, "learning_rate": 6.882330556012839e-11, "loss": 0.3953, "step": 19184 }, { "epoch": 3.0, "grad_norm": 26.500252992394028, "learning_rate": 6.270884472026062e-11, "loss": 0.4905, "step": 19185 }, { "epoch": 3.0, "grad_norm": 26.868225348266858, "learning_rate": 5.687877624893645e-11, "loss": 0.5009, "step": 19186 }, { "epoch": 3.0, "grad_norm": 16.7091206779695, "learning_rate": 5.133310030935867e-11, "loss": 0.4345, "step": 19187 }, { "epoch": 3.0, "grad_norm": 19.444532200969107, "learning_rate": 4.607181705917896e-11, "loss": 0.4049, "step": 19188 }, { "epoch": 3.0, "grad_norm": 31.746449589275194, "learning_rate": 4.1094926650497855e-11, "loss": 0.4715, "step": 19189 }, { "epoch": 3.0, "grad_norm": 26.348807097256906, "learning_rate": 3.640242922209325e-11, "loss": 0.4398, "step": 19190 }, { "epoch": 3.0, "grad_norm": 27.03494853284541, "learning_rate": 3.199432490941234e-11, "loss": 0.4329, "step": 19191 }, { "epoch": 3.0, "grad_norm": 24.461395597782875, "learning_rate": 2.787061383680012e-11, "loss": 0.4204, "step": 19192 }, { "epoch": 3.0, "grad_norm": 23.763644367276402, "learning_rate": 2.4031296120829993e-11, "loss": 0.4877, "step": 19193 }, { "epoch": 3.0, "grad_norm": 14.940773575296761, "learning_rate": 2.047637187252427e-11, "loss": 0.4944, "step": 19194 }, { "epoch": 3.0, "grad_norm": 30.485828040456667, "learning_rate": 1.720584119180302e-11, "loss": 0.4837, "step": 19195 }, { "epoch": 3.0, "grad_norm": 26.933422551153875, "learning_rate": 1.4219704171924976e-11, "loss": 0.4275, "step": 19196 }, { "epoch": 3.0, "grad_norm": 22.252548970525478, "learning_rate": 1.1517960898377312e-11, "loss": 0.4125, "step": 19197 }, { "epoch": 3.0, "grad_norm": 22.012802411460225, "learning_rate": 9.100611446655194e-12, "loss": 0.4283, "step": 19198 }, { "epoch": 3.0, "grad_norm": 19.45199348039055, "learning_rate": 6.967655886702674e-12, "loss": 0.3943, "step": 19199 }, { "epoch": 3.0, "grad_norm": 30.535730943838594, "learning_rate": 5.119094278471792e-12, "loss": 0.4638, "step": 19200 }, { "epoch": 3.0, "grad_norm": 26.369485826707194, "learning_rate": 3.5549266752532564e-12, "loss": 0.4573, "step": 19201 }, { "epoch": 3.0, "grad_norm": 19.86926674322362, "learning_rate": 2.275153120345763e-12, "loss": 0.4187, "step": 19202 }, { "epoch": 3.0, "grad_norm": 18.529909896802643, "learning_rate": 1.2797736514968962e-12, "loss": 0.4601, "step": 19203 }, { "epoch": 3.0, "grad_norm": 20.335206130392084, "learning_rate": 5.687882964622305e-13, "loss": 0.4302, "step": 19204 }, { "epoch": 3.0, "grad_norm": 21.343195540094797, "learning_rate": 1.4219707522578064e-13, "loss": 0.3673, "step": 19205 }, { "epoch": 3.0, "grad_norm": 26.62800484030171, "learning_rate": 0.0, "loss": 0.4807, "step": 19206 }, { "epoch": 3.0, "step": 19206, "total_flos": 8.250659187302859e+18, "train_loss": 0.6410556696475601, "train_runtime": 20419.4789, "train_samples_per_second": 120.393, "train_steps_per_second": 0.941 } ], "logging_steps": 1.0, "max_steps": 19206, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 50000, "total_flos": 8.250659187302859e+18, "train_batch_size": 2, "trial_name": null, "trial_params": null }