{ "best_metric": 6.35734701, "best_model_checkpoint": "/home/yisiyang/outputland/qwen2-vl-2b-instruct/v7-20241111-031420/checkpoint-2000", "epoch": 6.315789473684211, "eval_steps": 2000, "global_step": 2000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "acc": 0.29797983, "epoch": 0.003157894736842105, "grad_norm": 189.54470825195312, "learning_rate": 0.0, "loss": 6.8490591, "memory(GiB)": 14.42, "step": 1, "train_speed(iter/s)": 0.079303 }, { "acc": 0.30050507, "epoch": 0.015789473684210527, "grad_norm": 130.0618438720703, "learning_rate": 2.4122151757793534e-06, "loss": 5.23786497, "memory(GiB)": 14.42, "step": 5, "train_speed(iter/s)": 0.306791 }, { "acc": 0.35050507, "epoch": 0.031578947368421054, "grad_norm": 32.736263275146484, "learning_rate": 3.4510997050164104e-06, "loss": 3.81935425, "memory(GiB)": 14.42, "step": 10, "train_speed(iter/s)": 0.475545 }, { "acc": 0.37878788, "epoch": 0.04736842105263158, "grad_norm": 41.80891036987305, "learning_rate": 4.05880819719944e-06, "loss": 3.30803146, "memory(GiB)": 14.42, "step": 15, "train_speed(iter/s)": 0.579356 }, { "acc": 0.44040408, "epoch": 0.06315789473684211, "grad_norm": 27.359540939331055, "learning_rate": 4.4899842342534665e-06, "loss": 3.16991081, "memory(GiB)": 14.42, "step": 20, "train_speed(iter/s)": 0.650113 }, { "acc": 0.52828288, "epoch": 0.07894736842105263, "grad_norm": 26.735942840576172, "learning_rate": 4.824430351558707e-06, "loss": 3.03735027, "memory(GiB)": 14.42, "step": 25, "train_speed(iter/s)": 0.703029 }, { "acc": 0.5636364, "epoch": 0.09473684210526316, "grad_norm": 32.92796325683594, "learning_rate": 5.097692726436498e-06, "loss": 2.92974243, "memory(GiB)": 14.42, "step": 30, "train_speed(iter/s)": 0.738904 }, { "acc": 0.66464653, "epoch": 0.11052631578947368, "grad_norm": 25.728958129882812, "learning_rate": 5.328732772382501e-06, "loss": 2.82154236, "memory(GiB)": 14.42, "step": 35, "train_speed(iter/s)": 0.769629 }, { "acc": 0.64444456, "epoch": 0.12631578947368421, "grad_norm": 23.986242294311523, "learning_rate": 5.5288687634905236e-06, "loss": 2.76429024, "memory(GiB)": 14.42, "step": 40, "train_speed(iter/s)": 0.789225 }, { "acc": 0.68080816, "epoch": 0.14210526315789473, "grad_norm": 24.05732536315918, "learning_rate": 5.705401218619526e-06, "loss": 2.64311733, "memory(GiB)": 14.42, "step": 45, "train_speed(iter/s)": 0.806424 }, { "acc": 0.6424243, "epoch": 0.15789473684210525, "grad_norm": 27.014245986938477, "learning_rate": 5.863314880795763e-06, "loss": 2.62187328, "memory(GiB)": 14.42, "step": 50, "train_speed(iter/s)": 0.821589 }, { "acc": 0.65858588, "epoch": 0.1736842105263158, "grad_norm": 22.408689498901367, "learning_rate": 6.006165164335152e-06, "loss": 2.51736145, "memory(GiB)": 14.42, "step": 55, "train_speed(iter/s)": 0.838318 }, { "acc": 0.64848495, "epoch": 0.18947368421052632, "grad_norm": 24.095256805419922, "learning_rate": 6.136577255673552e-06, "loss": 2.47515697, "memory(GiB)": 14.42, "step": 60, "train_speed(iter/s)": 0.853904 }, { "acc": 0.66666675, "epoch": 0.20526315789473684, "grad_norm": 22.594839096069336, "learning_rate": 6.256544750330469e-06, "loss": 2.38148956, "memory(GiB)": 14.42, "step": 65, "train_speed(iter/s)": 0.865441 }, { "acc": 0.6818182, "epoch": 0.22105263157894736, "grad_norm": 22.433921813964844, "learning_rate": 6.367617301619559e-06, "loss": 2.33378811, "memory(GiB)": 14.42, "step": 70, "train_speed(iter/s)": 0.875815 }, { "acc": 0.67373748, "epoch": 0.23684210526315788, "grad_norm": 24.16921043395996, "learning_rate": 6.471023372978793e-06, "loss": 2.24571877, "memory(GiB)": 14.42, "step": 75, "train_speed(iter/s)": 0.884661 }, { "acc": 0.66060615, "epoch": 0.25263157894736843, "grad_norm": 22.719818115234375, "learning_rate": 6.56775329272758e-06, "loss": 2.19932251, "memory(GiB)": 14.42, "step": 80, "train_speed(iter/s)": 0.894036 }, { "acc": 0.67171721, "epoch": 0.26842105263157895, "grad_norm": 22.295177459716797, "learning_rate": 6.6586170853856745e-06, "loss": 2.10522652, "memory(GiB)": 14.42, "step": 85, "train_speed(iter/s)": 0.90264 }, { "acc": 0.65252533, "epoch": 0.28421052631578947, "grad_norm": 25.903900146484375, "learning_rate": 6.744285747856582e-06, "loss": 2.0310627, "memory(GiB)": 14.42, "step": 90, "train_speed(iter/s)": 0.910435 }, { "acc": 0.66565661, "epoch": 0.3, "grad_norm": 22.72349739074707, "learning_rate": 6.825321350816333e-06, "loss": 1.95369396, "memory(GiB)": 14.42, "step": 95, "train_speed(iter/s)": 0.916948 }, { "acc": 0.68080816, "epoch": 0.3157894736842105, "grad_norm": 22.040664672851562, "learning_rate": 6.902199410032821e-06, "loss": 1.89046402, "memory(GiB)": 14.42, "step": 100, "train_speed(iter/s)": 0.923641 }, { "acc": 0.66464653, "epoch": 0.33157894736842103, "grad_norm": 21.79152488708496, "learning_rate": 6.975325793802587e-06, "loss": 1.84504929, "memory(GiB)": 14.42, "step": 105, "train_speed(iter/s)": 0.92652 }, { "acc": 0.67373743, "epoch": 0.3473684210526316, "grad_norm": 22.12148666381836, "learning_rate": 7.045049693572208e-06, "loss": 1.73611717, "memory(GiB)": 14.42, "step": 110, "train_speed(iter/s)": 0.931682 }, { "acc": 0.69191928, "epoch": 0.3631578947368421, "grad_norm": 22.5858154296875, "learning_rate": 7.111673708972303e-06, "loss": 1.65465469, "memory(GiB)": 14.42, "step": 115, "train_speed(iter/s)": 0.937037 }, { "acc": 0.6878788, "epoch": 0.37894736842105264, "grad_norm": 22.869487762451172, "learning_rate": 7.17546178491061e-06, "loss": 1.56676788, "memory(GiB)": 14.42, "step": 120, "train_speed(iter/s)": 0.942049 }, { "acc": 0.68383846, "epoch": 0.39473684210526316, "grad_norm": 22.559728622436523, "learning_rate": 7.236645527338061e-06, "loss": 1.48099957, "memory(GiB)": 14.42, "step": 125, "train_speed(iter/s)": 0.946078 }, { "acc": 0.68787889, "epoch": 0.4105263157894737, "grad_norm": 21.670761108398438, "learning_rate": 7.295429279567525e-06, "loss": 1.41184158, "memory(GiB)": 14.42, "step": 130, "train_speed(iter/s)": 0.948975 }, { "acc": 0.6909091, "epoch": 0.4263157894736842, "grad_norm": 21.114803314208984, "learning_rate": 7.351994240039613e-06, "loss": 1.33319778, "memory(GiB)": 14.42, "step": 135, "train_speed(iter/s)": 0.95101 }, { "acc": 0.6939395, "epoch": 0.4421052631578947, "grad_norm": 20.817777633666992, "learning_rate": 7.406501830856614e-06, "loss": 1.28563147, "memory(GiB)": 14.42, "step": 140, "train_speed(iter/s)": 0.951565 }, { "acc": 0.68585868, "epoch": 0.45789473684210524, "grad_norm": 19.757150650024414, "learning_rate": 7.459096474945029e-06, "loss": 1.21721191, "memory(GiB)": 14.42, "step": 145, "train_speed(iter/s)": 0.952007 }, { "acc": 0.69090915, "epoch": 0.47368421052631576, "grad_norm": 18.436542510986328, "learning_rate": 7.509907902215849e-06, "loss": 1.12313366, "memory(GiB)": 14.42, "step": 150, "train_speed(iter/s)": 0.952888 }, { "acc": 0.68686876, "epoch": 0.48947368421052634, "grad_norm": 18.385173797607422, "learning_rate": 7.559053077443585e-06, "loss": 1.04606466, "memory(GiB)": 14.42, "step": 155, "train_speed(iter/s)": 0.95311 }, { "acc": 0.68080816, "epoch": 0.5052631578947369, "grad_norm": 16.366899490356445, "learning_rate": 7.606637821964637e-06, "loss": 1.02951183, "memory(GiB)": 14.42, "step": 160, "train_speed(iter/s)": 0.953518 }, { "acc": 0.68383846, "epoch": 0.5210526315789473, "grad_norm": 13.738570213317871, "learning_rate": 7.652758185755238e-06, "loss": 0.98384418, "memory(GiB)": 14.42, "step": 165, "train_speed(iter/s)": 0.954063 }, { "acc": 0.66464653, "epoch": 0.5368421052631579, "grad_norm": 11.252880096435547, "learning_rate": 7.697501614622732e-06, "loss": 0.95045271, "memory(GiB)": 14.42, "step": 170, "train_speed(iter/s)": 0.954931 }, { "acc": 0.68484855, "epoch": 0.5526315789473685, "grad_norm": 10.676709175109863, "learning_rate": 7.740947948161855e-06, "loss": 0.88134241, "memory(GiB)": 14.42, "step": 175, "train_speed(iter/s)": 0.956063 }, { "acc": 0.68080816, "epoch": 0.5684210526315789, "grad_norm": 9.484841346740723, "learning_rate": 7.78317027709364e-06, "loss": 0.87568655, "memory(GiB)": 14.42, "step": 180, "train_speed(iter/s)": 0.957191 }, { "acc": 0.68888893, "epoch": 0.5842105263157895, "grad_norm": 10.45778751373291, "learning_rate": 7.824235683113184e-06, "loss": 0.84243193, "memory(GiB)": 14.42, "step": 185, "train_speed(iter/s)": 0.956968 }, { "acc": 0.68686876, "epoch": 0.6, "grad_norm": 8.84788990020752, "learning_rate": 7.86420588005339e-06, "loss": 0.82236614, "memory(GiB)": 14.42, "step": 190, "train_speed(iter/s)": 0.956966 }, { "acc": 0.68686876, "epoch": 0.6157894736842106, "grad_norm": 5.3862409591674805, "learning_rate": 7.903137771750555e-06, "loss": 0.79296165, "memory(GiB)": 14.42, "step": 195, "train_speed(iter/s)": 0.956498 }, { "acc": 0.68989902, "epoch": 0.631578947368421, "grad_norm": 7.4440436363220215, "learning_rate": 7.941083939269876e-06, "loss": 0.78682814, "memory(GiB)": 14.42, "step": 200, "train_speed(iter/s)": 0.956087 }, { "acc": 0.70202026, "epoch": 0.6473684210526316, "grad_norm": 6.23380708694458, "learning_rate": 7.97809306796006e-06, "loss": 0.75604529, "memory(GiB)": 14.42, "step": 205, "train_speed(iter/s)": 0.958917 }, { "acc": 0.70909095, "epoch": 0.6631578947368421, "grad_norm": 7.885345935821533, "learning_rate": 8.014210323039644e-06, "loss": 0.77311206, "memory(GiB)": 14.42, "step": 210, "train_speed(iter/s)": 0.961234 }, { "acc": 0.70202026, "epoch": 0.6789473684210526, "grad_norm": 5.368066310882568, "learning_rate": 8.049477680983675e-06, "loss": 0.77599392, "memory(GiB)": 14.42, "step": 215, "train_speed(iter/s)": 0.963174 }, { "acc": 0.69191928, "epoch": 0.6947368421052632, "grad_norm": 5.842986106872559, "learning_rate": 8.083934222809266e-06, "loss": 0.78730788, "memory(GiB)": 14.42, "step": 220, "train_speed(iter/s)": 0.964927 }, { "acc": 0.7000001, "epoch": 0.7105263157894737, "grad_norm": 4.532631874084473, "learning_rate": 8.11761639439888e-06, "loss": 0.76723304, "memory(GiB)": 14.42, "step": 225, "train_speed(iter/s)": 0.966333 }, { "acc": 0.69696975, "epoch": 0.7263157894736842, "grad_norm": 6.793882846832275, "learning_rate": 8.15055823820936e-06, "loss": 0.77692299, "memory(GiB)": 14.42, "step": 230, "train_speed(iter/s)": 0.967921 }, { "acc": 0.68282833, "epoch": 0.7421052631578947, "grad_norm": 5.946906089782715, "learning_rate": 8.182791600059879e-06, "loss": 0.78708334, "memory(GiB)": 14.42, "step": 235, "train_speed(iter/s)": 0.968657 }, { "acc": 0.69494953, "epoch": 0.7578947368421053, "grad_norm": 5.495345115661621, "learning_rate": 8.214346314147667e-06, "loss": 0.77794123, "memory(GiB)": 14.42, "step": 240, "train_speed(iter/s)": 0.969446 }, { "acc": 0.70202026, "epoch": 0.7736842105263158, "grad_norm": 4.244708061218262, "learning_rate": 8.24525036898565e-06, "loss": 0.74534717, "memory(GiB)": 14.42, "step": 245, "train_speed(iter/s)": 0.971096 }, { "acc": 0.68080816, "epoch": 0.7894736842105263, "grad_norm": 7.3419013023376465, "learning_rate": 8.275530056575115e-06, "loss": 0.79043388, "memory(GiB)": 14.42, "step": 250, "train_speed(iter/s)": 0.971701 }, { "acc": 0.6818182, "epoch": 0.8052631578947368, "grad_norm": 5.537174701690674, "learning_rate": 8.30521010680576e-06, "loss": 0.76065493, "memory(GiB)": 14.42, "step": 255, "train_speed(iter/s)": 0.972913 }, { "acc": 0.69595966, "epoch": 0.8210526315789474, "grad_norm": 6.621566295623779, "learning_rate": 8.334313808804581e-06, "loss": 0.73345542, "memory(GiB)": 14.42, "step": 260, "train_speed(iter/s)": 0.973338 }, { "acc": 0.70101013, "epoch": 0.8368421052631579, "grad_norm": 5.5801239013671875, "learning_rate": 8.36286312072555e-06, "loss": 0.76861334, "memory(GiB)": 14.42, "step": 265, "train_speed(iter/s)": 0.9735 }, { "acc": 0.69797988, "epoch": 0.8526315789473684, "grad_norm": 4.083832263946533, "learning_rate": 8.39087876927667e-06, "loss": 0.76421027, "memory(GiB)": 14.42, "step": 270, "train_speed(iter/s)": 0.972914 }, { "acc": 0.67979808, "epoch": 0.868421052631579, "grad_norm": 8.106366157531738, "learning_rate": 8.418380340114505e-06, "loss": 0.7734849, "memory(GiB)": 14.42, "step": 275, "train_speed(iter/s)": 0.972258 }, { "acc": 0.71010108, "epoch": 0.8842105263157894, "grad_norm": 5.391201972961426, "learning_rate": 8.445386360093671e-06, "loss": 0.75838871, "memory(GiB)": 14.42, "step": 280, "train_speed(iter/s)": 0.973168 }, { "acc": 0.69898996, "epoch": 0.9, "grad_norm": 4.560817241668701, "learning_rate": 8.47191437223642e-06, "loss": 0.72944508, "memory(GiB)": 14.42, "step": 285, "train_speed(iter/s)": 0.974072 }, { "acc": 0.71616168, "epoch": 0.9157894736842105, "grad_norm": 4.5082550048828125, "learning_rate": 8.497981004182087e-06, "loss": 0.72101011, "memory(GiB)": 14.42, "step": 290, "train_speed(iter/s)": 0.974336 }, { "acc": 0.71111121, "epoch": 0.9315789473684211, "grad_norm": 5.37807035446167, "learning_rate": 8.523602030785273e-06, "loss": 0.72204466, "memory(GiB)": 14.42, "step": 295, "train_speed(iter/s)": 0.974602 }, { "acc": 0.69494953, "epoch": 0.9473684210526315, "grad_norm": 7.141894340515137, "learning_rate": 8.548792431452906e-06, "loss": 0.7805851, "memory(GiB)": 14.42, "step": 300, "train_speed(iter/s)": 0.974531 }, { "acc": 0.69595962, "epoch": 0.9631578947368421, "grad_norm": 5.165417194366455, "learning_rate": 8.573566442742008e-06, "loss": 0.77751379, "memory(GiB)": 14.42, "step": 305, "train_speed(iter/s)": 0.973871 }, { "acc": 0.69898996, "epoch": 0.9789473684210527, "grad_norm": 5.182235240936279, "learning_rate": 8.597937606680642e-06, "loss": 0.75329895, "memory(GiB)": 14.42, "step": 310, "train_speed(iter/s)": 0.973247 }, { "acc": 0.71313133, "epoch": 0.9947368421052631, "grad_norm": 4.949851036071777, "learning_rate": 8.621918815222675e-06, "loss": 0.72683144, "memory(GiB)": 14.42, "step": 315, "train_speed(iter/s)": 0.97287 }, { "acc": 0.71616168, "epoch": 1.0105263157894737, "grad_norm": 6.737766742706299, "learning_rate": 8.645522351201692e-06, "loss": 0.76480436, "memory(GiB)": 14.42, "step": 320, "train_speed(iter/s)": 0.970956 }, { "acc": 0.70303035, "epoch": 1.0263157894736843, "grad_norm": 4.715226173400879, "learning_rate": 8.668759926109822e-06, "loss": 0.73685551, "memory(GiB)": 14.42, "step": 325, "train_speed(iter/s)": 0.97194 }, { "acc": 0.69797988, "epoch": 1.0421052631578946, "grad_norm": 7.166709899902344, "learning_rate": 8.691642714992295e-06, "loss": 0.73682857, "memory(GiB)": 14.42, "step": 330, "train_speed(iter/s)": 0.972694 }, { "acc": 0.70404043, "epoch": 1.0578947368421052, "grad_norm": 4.959497928619385, "learning_rate": 8.714181388718075e-06, "loss": 0.7466671, "memory(GiB)": 14.42, "step": 335, "train_speed(iter/s)": 0.972539 }, { "acc": 0.71313133, "epoch": 1.0736842105263158, "grad_norm": 4.548097133636475, "learning_rate": 8.736386143859788e-06, "loss": 0.72707992, "memory(GiB)": 14.42, "step": 340, "train_speed(iter/s)": 0.972893 }, { "acc": 0.69494963, "epoch": 1.0894736842105264, "grad_norm": 5.036413192749023, "learning_rate": 8.75826673039239e-06, "loss": 0.76058822, "memory(GiB)": 14.42, "step": 345, "train_speed(iter/s)": 0.973963 }, { "acc": 0.69899001, "epoch": 1.1052631578947367, "grad_norm": 4.9548420906066895, "learning_rate": 8.77983247739891e-06, "loss": 0.75256853, "memory(GiB)": 14.42, "step": 350, "train_speed(iter/s)": 0.975074 }, { "acc": 0.71010108, "epoch": 1.1210526315789473, "grad_norm": 5.18994140625, "learning_rate": 8.801092316952921e-06, "loss": 0.7360672, "memory(GiB)": 14.42, "step": 355, "train_speed(iter/s)": 0.976375 }, { "acc": 0.71111116, "epoch": 1.1368421052631579, "grad_norm": 3.965578317642212, "learning_rate": 8.822054806330697e-06, "loss": 0.7136857, "memory(GiB)": 14.42, "step": 360, "train_speed(iter/s)": 0.977264 }, { "acc": 0.7060606, "epoch": 1.1526315789473685, "grad_norm": 5.464134216308594, "learning_rate": 8.842728148691392e-06, "loss": 0.7406353, "memory(GiB)": 14.42, "step": 365, "train_speed(iter/s)": 0.978313 }, { "acc": 0.71111121, "epoch": 1.168421052631579, "grad_norm": 4.730813980102539, "learning_rate": 8.863120212350241e-06, "loss": 0.71864109, "memory(GiB)": 14.42, "step": 370, "train_speed(iter/s)": 0.978981 }, { "acc": 0.69898996, "epoch": 1.1842105263157894, "grad_norm": 6.221624851226807, "learning_rate": 8.883238548758146e-06, "loss": 0.73387871, "memory(GiB)": 14.42, "step": 375, "train_speed(iter/s)": 0.979682 }, { "acc": 0.73232336, "epoch": 1.2, "grad_norm": 5.355544567108154, "learning_rate": 8.903090409290447e-06, "loss": 0.70250816, "memory(GiB)": 14.42, "step": 380, "train_speed(iter/s)": 0.980355 }, { "acc": 0.68989906, "epoch": 1.2157894736842105, "grad_norm": 5.793821811676025, "learning_rate": 8.922682760938298e-06, "loss": 0.75479689, "memory(GiB)": 14.42, "step": 385, "train_speed(iter/s)": 0.981586 }, { "acc": 0.69696975, "epoch": 1.231578947368421, "grad_norm": 4.8115434646606445, "learning_rate": 8.942022300987612e-06, "loss": 0.7389276, "memory(GiB)": 14.42, "step": 390, "train_speed(iter/s)": 0.982499 }, { "acc": 0.7181819, "epoch": 1.2473684210526317, "grad_norm": 3.446106195449829, "learning_rate": 8.961115470762944e-06, "loss": 0.71739759, "memory(GiB)": 14.42, "step": 395, "train_speed(iter/s)": 0.983313 }, { "acc": 0.7181819, "epoch": 1.263157894736842, "grad_norm": 5.517762660980225, "learning_rate": 8.979968468506933e-06, "loss": 0.71014919, "memory(GiB)": 14.42, "step": 400, "train_speed(iter/s)": 0.98393 }, { "acc": 0.72121215, "epoch": 1.2789473684210526, "grad_norm": 4.866208076477051, "learning_rate": 8.998587261459699e-06, "loss": 0.70516768, "memory(GiB)": 14.42, "step": 405, "train_speed(iter/s)": 0.984487 }, { "acc": 0.69595966, "epoch": 1.2947368421052632, "grad_norm": 4.031857967376709, "learning_rate": 9.016977597197119e-06, "loss": 0.7197782, "memory(GiB)": 14.42, "step": 410, "train_speed(iter/s)": 0.984938 }, { "acc": 0.71616168, "epoch": 1.3105263157894738, "grad_norm": 5.485307216644287, "learning_rate": 9.035145014281876e-06, "loss": 0.69396477, "memory(GiB)": 14.42, "step": 415, "train_speed(iter/s)": 0.985351 }, { "acc": 0.70303035, "epoch": 1.3263157894736843, "grad_norm": 4.575260162353516, "learning_rate": 9.053094852276701e-06, "loss": 0.71491661, "memory(GiB)": 14.42, "step": 420, "train_speed(iter/s)": 0.98555 }, { "acc": 0.71717176, "epoch": 1.3421052631578947, "grad_norm": 4.316519260406494, "learning_rate": 9.070832261165027e-06, "loss": 0.71715617, "memory(GiB)": 14.42, "step": 425, "train_speed(iter/s)": 0.98596 }, { "acc": 0.69898996, "epoch": 1.3578947368421053, "grad_norm": 4.875655651092529, "learning_rate": 9.088362210220732e-06, "loss": 0.73308363, "memory(GiB)": 14.42, "step": 430, "train_speed(iter/s)": 0.985125 }, { "acc": 0.70505056, "epoch": 1.3736842105263158, "grad_norm": 5.2696852684021, "learning_rate": 9.105689496365117e-06, "loss": 0.74291925, "memory(GiB)": 14.42, "step": 435, "train_speed(iter/s)": 0.98451 }, { "acc": 0.71313133, "epoch": 1.3894736842105262, "grad_norm": 5.138822078704834, "learning_rate": 9.122818752046321e-06, "loss": 0.72327461, "memory(GiB)": 14.42, "step": 440, "train_speed(iter/s)": 0.984452 }, { "acc": 0.72222228, "epoch": 1.4052631578947368, "grad_norm": 3.904170036315918, "learning_rate": 9.13975445267355e-06, "loss": 0.7007791, "memory(GiB)": 14.42, "step": 445, "train_speed(iter/s)": 0.984092 }, { "acc": 0.71414146, "epoch": 1.4210526315789473, "grad_norm": 5.243173599243164, "learning_rate": 9.156500923635937e-06, "loss": 0.69670939, "memory(GiB)": 14.42, "step": 450, "train_speed(iter/s)": 0.984038 }, { "acc": 0.70808086, "epoch": 1.436842105263158, "grad_norm": 5.191768169403076, "learning_rate": 9.173062346933617e-06, "loss": 0.74531546, "memory(GiB)": 14.42, "step": 455, "train_speed(iter/s)": 0.983589 }, { "acc": 0.69797988, "epoch": 1.4526315789473685, "grad_norm": 6.049670696258545, "learning_rate": 9.189442767446416e-06, "loss": 0.70544295, "memory(GiB)": 14.42, "step": 460, "train_speed(iter/s)": 0.983483 }, { "acc": 0.72828288, "epoch": 1.4684210526315788, "grad_norm": 4.956687927246094, "learning_rate": 9.205646098863671e-06, "loss": 0.69190674, "memory(GiB)": 14.42, "step": 465, "train_speed(iter/s)": 0.983203 }, { "acc": 0.71414146, "epoch": 1.4842105263157894, "grad_norm": 4.949287414550781, "learning_rate": 9.221676129296936e-06, "loss": 0.71432433, "memory(GiB)": 14.42, "step": 470, "train_speed(iter/s)": 0.982974 }, { "acc": 0.71010103, "epoch": 1.5, "grad_norm": 4.197946071624756, "learning_rate": 9.237536526595687e-06, "loss": 0.70365272, "memory(GiB)": 14.42, "step": 475, "train_speed(iter/s)": 0.983499 }, { "acc": 0.72222228, "epoch": 1.5157894736842106, "grad_norm": 3.9227776527404785, "learning_rate": 9.253230843384724e-06, "loss": 0.68840942, "memory(GiB)": 14.42, "step": 480, "train_speed(iter/s)": 0.984297 }, { "acc": 0.70101013, "epoch": 1.5315789473684212, "grad_norm": 5.360828876495361, "learning_rate": 9.268762521840534e-06, "loss": 0.71461992, "memory(GiB)": 14.42, "step": 485, "train_speed(iter/s)": 0.985036 }, { "acc": 0.71515155, "epoch": 1.5473684210526315, "grad_norm": 3.924023151397705, "learning_rate": 9.284134898222705e-06, "loss": 0.70787678, "memory(GiB)": 14.42, "step": 490, "train_speed(iter/s)": 0.985206 }, { "acc": 0.72121215, "epoch": 1.563157894736842, "grad_norm": 4.619671821594238, "learning_rate": 9.299351207175325e-06, "loss": 0.70242157, "memory(GiB)": 14.42, "step": 495, "train_speed(iter/s)": 0.984921 }, { "acc": 0.70707073, "epoch": 1.5789473684210527, "grad_norm": 3.776333808898926, "learning_rate": 9.314414585812172e-06, "loss": 0.71011858, "memory(GiB)": 14.42, "step": 500, "train_speed(iter/s)": 0.984704 }, { "acc": 0.71515155, "epoch": 1.594736842105263, "grad_norm": 4.775974273681641, "learning_rate": 9.329328077598717e-06, "loss": 0.72129836, "memory(GiB)": 14.42, "step": 505, "train_speed(iter/s)": 0.984498 }, { "acc": 0.70303035, "epoch": 1.6105263157894738, "grad_norm": 4.790719032287598, "learning_rate": 9.344094636042818e-06, "loss": 0.71535187, "memory(GiB)": 14.42, "step": 510, "train_speed(iter/s)": 0.984973 }, { "acc": 0.70909095, "epoch": 1.6263157894736842, "grad_norm": 5.532032012939453, "learning_rate": 9.358717128205423e-06, "loss": 0.72311149, "memory(GiB)": 14.42, "step": 515, "train_speed(iter/s)": 0.985246 }, { "acc": 0.71313133, "epoch": 1.6421052631578947, "grad_norm": 4.431013107299805, "learning_rate": 9.373198338041638e-06, "loss": 0.70520873, "memory(GiB)": 14.42, "step": 520, "train_speed(iter/s)": 0.985808 }, { "acc": 0.71313133, "epoch": 1.6578947368421053, "grad_norm": 5.026467800140381, "learning_rate": 9.38754096958194e-06, "loss": 0.71683903, "memory(GiB)": 14.42, "step": 525, "train_speed(iter/s)": 0.986072 }, { "acc": 0.71313133, "epoch": 1.6736842105263157, "grad_norm": 4.701731204986572, "learning_rate": 9.401747649962608e-06, "loss": 0.68079748, "memory(GiB)": 14.42, "step": 530, "train_speed(iter/s)": 0.986402 }, { "acc": 0.72020206, "epoch": 1.6894736842105265, "grad_norm": 3.6100666522979736, "learning_rate": 9.415820932313868e-06, "loss": 0.7060638, "memory(GiB)": 14.42, "step": 535, "train_speed(iter/s)": 0.987056 }, { "acc": 0.73535361, "epoch": 1.7052631578947368, "grad_norm": 3.5025079250335693, "learning_rate": 9.429763298513726e-06, "loss": 0.70274849, "memory(GiB)": 14.42, "step": 540, "train_speed(iter/s)": 0.987628 }, { "acc": 0.7272728, "epoch": 1.7210526315789474, "grad_norm": 4.738600254058838, "learning_rate": 9.443577161814856e-06, "loss": 0.69971213, "memory(GiB)": 14.42, "step": 545, "train_speed(iter/s)": 0.988328 }, { "acc": 0.7272728, "epoch": 1.736842105263158, "grad_norm": 4.521397113800049, "learning_rate": 9.45726486935156e-06, "loss": 0.68742881, "memory(GiB)": 14.42, "step": 550, "train_speed(iter/s)": 0.988822 }, { "acc": 0.71111116, "epoch": 1.7526315789473683, "grad_norm": 5.196502685546875, "learning_rate": 9.470828704533272e-06, "loss": 0.69165187, "memory(GiB)": 14.42, "step": 555, "train_speed(iter/s)": 0.989364 }, { "acc": 0.72525253, "epoch": 1.768421052631579, "grad_norm": 5.465752124786377, "learning_rate": 9.484270889330728e-06, "loss": 0.70788536, "memory(GiB)": 14.42, "step": 560, "train_speed(iter/s)": 0.989796 }, { "acc": 0.71414146, "epoch": 1.7842105263157895, "grad_norm": 6.159824848175049, "learning_rate": 9.497593586460533e-06, "loss": 0.70432334, "memory(GiB)": 14.42, "step": 565, "train_speed(iter/s)": 0.990473 }, { "acc": 0.7060607, "epoch": 1.8, "grad_norm": 4.181656837463379, "learning_rate": 9.510798901473478e-06, "loss": 0.70204101, "memory(GiB)": 14.42, "step": 570, "train_speed(iter/s)": 0.991023 }, { "acc": 0.7272728, "epoch": 1.8157894736842106, "grad_norm": 4.163989067077637, "learning_rate": 9.523888884751656e-06, "loss": 0.70399027, "memory(GiB)": 14.42, "step": 575, "train_speed(iter/s)": 0.991508 }, { "acc": 0.71919203, "epoch": 1.831578947368421, "grad_norm": 4.880708694458008, "learning_rate": 9.536865533419144e-06, "loss": 0.67688112, "memory(GiB)": 14.42, "step": 580, "train_speed(iter/s)": 0.991861 }, { "acc": 0.7363637, "epoch": 1.8473684210526315, "grad_norm": 4.401225566864014, "learning_rate": 9.549730793170642e-06, "loss": 0.68770628, "memory(GiB)": 14.42, "step": 585, "train_speed(iter/s)": 0.992325 }, { "acc": 0.7242425, "epoch": 1.8631578947368421, "grad_norm": 4.399166107177734, "learning_rate": 9.562486560022329e-06, "loss": 0.6890409, "memory(GiB)": 14.42, "step": 590, "train_speed(iter/s)": 0.992375 }, { "acc": 0.73939395, "epoch": 1.8789473684210525, "grad_norm": 3.671194314956665, "learning_rate": 9.575134681988824e-06, "loss": 0.6716217, "memory(GiB)": 14.42, "step": 595, "train_speed(iter/s)": 0.992404 }, { "acc": 0.72525253, "epoch": 1.8947368421052633, "grad_norm": 4.786886215209961, "learning_rate": 9.587676960689963e-06, "loss": 0.67677307, "memory(GiB)": 14.42, "step": 600, "train_speed(iter/s)": 0.992267 }, { "acc": 0.7303031, "epoch": 1.9105263157894736, "grad_norm": 4.316333293914795, "learning_rate": 9.60011515289095e-06, "loss": 0.68564234, "memory(GiB)": 14.42, "step": 605, "train_speed(iter/s)": 0.992113 }, { "acc": 0.73131313, "epoch": 1.9263157894736842, "grad_norm": 6.806692123413086, "learning_rate": 9.612450971979065e-06, "loss": 0.70519056, "memory(GiB)": 14.42, "step": 610, "train_speed(iter/s)": 0.991945 }, { "acc": 0.71717176, "epoch": 1.9421052631578948, "grad_norm": 4.798392295837402, "learning_rate": 9.624686089380146e-06, "loss": 0.70453787, "memory(GiB)": 14.42, "step": 615, "train_speed(iter/s)": 0.991941 }, { "acc": 0.72626266, "epoch": 1.9578947368421051, "grad_norm": 5.573512554168701, "learning_rate": 9.6368221359177e-06, "loss": 0.69707575, "memory(GiB)": 14.42, "step": 620, "train_speed(iter/s)": 0.991755 }, { "acc": 0.72525253, "epoch": 1.973684210526316, "grad_norm": 5.02384090423584, "learning_rate": 9.648860703117414e-06, "loss": 0.693221, "memory(GiB)": 14.42, "step": 625, "train_speed(iter/s)": 0.992196 }, { "acc": 0.72323241, "epoch": 1.9894736842105263, "grad_norm": 4.060290336608887, "learning_rate": 9.660803344459732e-06, "loss": 0.71005087, "memory(GiB)": 14.42, "step": 630, "train_speed(iter/s)": 0.992576 }, { "acc": 0.73131313, "epoch": 2.0052631578947366, "grad_norm": 4.20712423324585, "learning_rate": 9.672651576582882e-06, "loss": 0.6809638, "memory(GiB)": 14.42, "step": 635, "train_speed(iter/s)": 0.992308 }, { "acc": 0.74141417, "epoch": 2.0210526315789474, "grad_norm": 4.575295448303223, "learning_rate": 9.68440688043875e-06, "loss": 0.67519088, "memory(GiB)": 14.42, "step": 640, "train_speed(iter/s)": 0.992596 }, { "acc": 0.72828288, "epoch": 2.036842105263158, "grad_norm": 4.49033784866333, "learning_rate": 9.696070702403762e-06, "loss": 0.68399596, "memory(GiB)": 14.42, "step": 645, "train_speed(iter/s)": 0.993133 }, { "acc": 0.71717176, "epoch": 2.0526315789473686, "grad_norm": 6.009328842163086, "learning_rate": 9.707644455346878e-06, "loss": 0.67610931, "memory(GiB)": 14.42, "step": 650, "train_speed(iter/s)": 0.993511 }, { "acc": 0.72020206, "epoch": 2.068421052631579, "grad_norm": 5.208271503448486, "learning_rate": 9.719129519656674e-06, "loss": 0.68497334, "memory(GiB)": 14.42, "step": 655, "train_speed(iter/s)": 0.993788 }, { "acc": 0.73131323, "epoch": 2.0842105263157893, "grad_norm": 4.493729591369629, "learning_rate": 9.730527244229352e-06, "loss": 0.65444393, "memory(GiB)": 14.42, "step": 660, "train_speed(iter/s)": 0.993916 }, { "acc": 0.75050507, "epoch": 2.1, "grad_norm": 3.5913338661193848, "learning_rate": 9.741838947419482e-06, "loss": 0.64511662, "memory(GiB)": 14.42, "step": 665, "train_speed(iter/s)": 0.99447 }, { "acc": 0.73232327, "epoch": 2.1157894736842104, "grad_norm": 4.7521138191223145, "learning_rate": 9.753065917955132e-06, "loss": 0.67699757, "memory(GiB)": 14.42, "step": 670, "train_speed(iter/s)": 0.994865 }, { "acc": 0.71919203, "epoch": 2.1315789473684212, "grad_norm": 3.9687211513519287, "learning_rate": 9.764209415818967e-06, "loss": 0.67539482, "memory(GiB)": 14.42, "step": 675, "train_speed(iter/s)": 0.994992 }, { "acc": 0.7242425, "epoch": 2.1473684210526316, "grad_norm": 4.969510555267334, "learning_rate": 9.775270673096843e-06, "loss": 0.69655457, "memory(GiB)": 14.42, "step": 680, "train_speed(iter/s)": 0.995448 }, { "acc": 0.75050507, "epoch": 2.163157894736842, "grad_norm": 4.0912275314331055, "learning_rate": 9.786250894795326e-06, "loss": 0.66344948, "memory(GiB)": 14.42, "step": 685, "train_speed(iter/s)": 0.995672 }, { "acc": 0.73434348, "epoch": 2.1789473684210527, "grad_norm": 3.628950595855713, "learning_rate": 9.797151259629445e-06, "loss": 0.66907301, "memory(GiB)": 14.42, "step": 690, "train_speed(iter/s)": 0.995844 }, { "acc": 0.74141417, "epoch": 2.194736842105263, "grad_norm": 4.280593395233154, "learning_rate": 9.807972920782061e-06, "loss": 0.65676284, "memory(GiB)": 14.42, "step": 695, "train_speed(iter/s)": 0.996132 }, { "acc": 0.73131323, "epoch": 2.2105263157894735, "grad_norm": 4.347470760345459, "learning_rate": 9.818717006635969e-06, "loss": 0.68502617, "memory(GiB)": 14.42, "step": 700, "train_speed(iter/s)": 0.996579 }, { "acc": 0.75353541, "epoch": 2.2263157894736842, "grad_norm": 4.028265953063965, "learning_rate": 9.829384621479966e-06, "loss": 0.64554987, "memory(GiB)": 14.42, "step": 705, "train_speed(iter/s)": 0.996871 }, { "acc": 0.7333334, "epoch": 2.2421052631578946, "grad_norm": 5.8726959228515625, "learning_rate": 9.839976846189978e-06, "loss": 0.67025905, "memory(GiB)": 14.42, "step": 710, "train_speed(iter/s)": 0.997007 }, { "acc": 0.74040413, "epoch": 2.2578947368421054, "grad_norm": 5.193964004516602, "learning_rate": 9.850494738886266e-06, "loss": 0.6601409, "memory(GiB)": 14.42, "step": 715, "train_speed(iter/s)": 0.996822 }, { "acc": 0.73232327, "epoch": 2.2736842105263158, "grad_norm": 6.3874711990356445, "learning_rate": 9.860939335567754e-06, "loss": 0.65816932, "memory(GiB)": 14.42, "step": 720, "train_speed(iter/s)": 0.996759 }, { "acc": 0.73131323, "epoch": 2.2894736842105265, "grad_norm": 5.662936210632324, "learning_rate": 9.871311650724383e-06, "loss": 0.66208439, "memory(GiB)": 14.42, "step": 725, "train_speed(iter/s)": 0.996521 }, { "acc": 0.7363637, "epoch": 2.305263157894737, "grad_norm": 5.146431922912598, "learning_rate": 9.881612677928449e-06, "loss": 0.65422068, "memory(GiB)": 14.42, "step": 730, "train_speed(iter/s)": 0.996472 }, { "acc": 0.71919203, "epoch": 2.3210526315789473, "grad_norm": 5.648817539215088, "learning_rate": 9.891843390405736e-06, "loss": 0.67373962, "memory(GiB)": 14.42, "step": 735, "train_speed(iter/s)": 0.99626 }, { "acc": 0.7363637, "epoch": 2.336842105263158, "grad_norm": 4.64543342590332, "learning_rate": 9.902004741587298e-06, "loss": 0.66676788, "memory(GiB)": 14.42, "step": 740, "train_speed(iter/s)": 0.996012 }, { "acc": 0.74141421, "epoch": 2.3526315789473684, "grad_norm": 4.293550968170166, "learning_rate": 9.912097665642666e-06, "loss": 0.68388977, "memory(GiB)": 14.42, "step": 745, "train_speed(iter/s)": 0.995772 }, { "acc": 0.73434348, "epoch": 2.3684210526315788, "grad_norm": 4.982192516326904, "learning_rate": 9.922123077995203e-06, "loss": 0.68120513, "memory(GiB)": 14.42, "step": 750, "train_speed(iter/s)": 0.996044 }, { "acc": 0.7212122, "epoch": 2.3842105263157896, "grad_norm": 4.732614040374756, "learning_rate": 9.932081875820369e-06, "loss": 0.67536116, "memory(GiB)": 14.42, "step": 755, "train_speed(iter/s)": 0.996511 }, { "acc": 0.7303031, "epoch": 2.4, "grad_norm": 4.1169867515563965, "learning_rate": 9.941974938527504e-06, "loss": 0.6712615, "memory(GiB)": 14.42, "step": 760, "train_speed(iter/s)": 0.996962 }, { "acc": 0.73131323, "epoch": 2.4157894736842107, "grad_norm": 4.577713966369629, "learning_rate": 9.951803128225848e-06, "loss": 0.69072509, "memory(GiB)": 14.42, "step": 765, "train_speed(iter/s)": 0.997234 }, { "acc": 0.73131323, "epoch": 2.431578947368421, "grad_norm": 4.2971367835998535, "learning_rate": 9.961567290175355e-06, "loss": 0.68363256, "memory(GiB)": 14.42, "step": 770, "train_speed(iter/s)": 0.997844 }, { "acc": 0.74949493, "epoch": 2.4473684210526314, "grad_norm": 5.387909889221191, "learning_rate": 9.97126825322294e-06, "loss": 0.65476675, "memory(GiB)": 14.42, "step": 775, "train_speed(iter/s)": 0.998001 }, { "acc": 0.74747481, "epoch": 2.463157894736842, "grad_norm": 5.511860370635986, "learning_rate": 9.980906830224669e-06, "loss": 0.66419239, "memory(GiB)": 14.42, "step": 780, "train_speed(iter/s)": 0.998258 }, { "acc": 0.75656567, "epoch": 2.4789473684210526, "grad_norm": 4.687104225158691, "learning_rate": 9.990483818454497e-06, "loss": 0.62274342, "memory(GiB)": 14.42, "step": 785, "train_speed(iter/s)": 0.998602 }, { "acc": 0.74444447, "epoch": 2.4947368421052634, "grad_norm": 4.910262107849121, "learning_rate": 1e-05, "loss": 0.6592267, "memory(GiB)": 14.42, "step": 790, "train_speed(iter/s)": 0.998998 }, { "acc": 0.7606061, "epoch": 2.5105263157894737, "grad_norm": 4.4709601402282715, "learning_rate": 9.99999726236902e-06, "loss": 0.65275908, "memory(GiB)": 14.42, "step": 795, "train_speed(iter/s)": 0.999322 }, { "acc": 0.72525253, "epoch": 2.526315789473684, "grad_norm": 4.263107776641846, "learning_rate": 9.999989049479075e-06, "loss": 0.67762294, "memory(GiB)": 14.42, "step": 800, "train_speed(iter/s)": 0.999649 }, { "acc": 0.74747477, "epoch": 2.542105263157895, "grad_norm": 4.175100326538086, "learning_rate": 9.999975361339162e-06, "loss": 0.64345884, "memory(GiB)": 14.42, "step": 805, "train_speed(iter/s)": 0.999975 }, { "acc": 0.75050507, "epoch": 2.557894736842105, "grad_norm": 4.516010284423828, "learning_rate": 9.99995619796427e-06, "loss": 0.64713798, "memory(GiB)": 14.42, "step": 810, "train_speed(iter/s)": 1.000482 }, { "acc": 0.7333334, "epoch": 2.5736842105263156, "grad_norm": 5.067887306213379, "learning_rate": 9.999931559375387e-06, "loss": 0.66570234, "memory(GiB)": 14.42, "step": 815, "train_speed(iter/s)": 1.000429 }, { "acc": 0.7484849, "epoch": 2.5894736842105264, "grad_norm": 5.1078996658325195, "learning_rate": 9.999901445599495e-06, "loss": 0.6631753, "memory(GiB)": 14.42, "step": 820, "train_speed(iter/s)": 1.000908 }, { "acc": 0.74141417, "epoch": 2.6052631578947367, "grad_norm": 3.582270860671997, "learning_rate": 9.999865856669573e-06, "loss": 0.65539184, "memory(GiB)": 14.42, "step": 825, "train_speed(iter/s)": 1.001101 }, { "acc": 0.74646468, "epoch": 2.6210526315789475, "grad_norm": 4.469554901123047, "learning_rate": 9.999824792624602e-06, "loss": 0.66321173, "memory(GiB)": 14.42, "step": 830, "train_speed(iter/s)": 1.001374 }, { "acc": 0.75050507, "epoch": 2.636842105263158, "grad_norm": 4.35246467590332, "learning_rate": 9.999778253509548e-06, "loss": 0.63566313, "memory(GiB)": 14.42, "step": 835, "train_speed(iter/s)": 1.001339 }, { "acc": 0.7636364, "epoch": 2.6526315789473687, "grad_norm": 4.282431125640869, "learning_rate": 9.99972623937538e-06, "loss": 0.64590969, "memory(GiB)": 14.42, "step": 840, "train_speed(iter/s)": 1.000982 }, { "acc": 0.75050507, "epoch": 2.668421052631579, "grad_norm": 4.943569183349609, "learning_rate": 9.999668750279061e-06, "loss": 0.6387764, "memory(GiB)": 14.42, "step": 845, "train_speed(iter/s)": 1.000657 }, { "acc": 0.7454546, "epoch": 2.6842105263157894, "grad_norm": 5.16862154006958, "learning_rate": 9.999605786283555e-06, "loss": 0.65590019, "memory(GiB)": 14.42, "step": 850, "train_speed(iter/s)": 1.000469 }, { "acc": 0.75353541, "epoch": 2.7, "grad_norm": 6.1622633934021, "learning_rate": 9.999537347457813e-06, "loss": 0.63483958, "memory(GiB)": 14.42, "step": 855, "train_speed(iter/s)": 1.000055 }, { "acc": 0.73737373, "epoch": 2.7157894736842105, "grad_norm": 4.183565139770508, "learning_rate": 9.99946343387679e-06, "loss": 0.65940695, "memory(GiB)": 14.42, "step": 860, "train_speed(iter/s)": 1.000002 }, { "acc": 0.72525253, "epoch": 2.731578947368421, "grad_norm": 5.071559906005859, "learning_rate": 9.999384045621432e-06, "loss": 0.65910063, "memory(GiB)": 14.42, "step": 865, "train_speed(iter/s)": 0.999665 }, { "acc": 0.73535361, "epoch": 2.7473684210526317, "grad_norm": 5.395997524261475, "learning_rate": 9.999299182778683e-06, "loss": 0.66515555, "memory(GiB)": 14.42, "step": 870, "train_speed(iter/s)": 0.999492 }, { "acc": 0.7575758, "epoch": 2.763157894736842, "grad_norm": 3.9158406257629395, "learning_rate": 9.99920884544148e-06, "loss": 0.6367353, "memory(GiB)": 14.42, "step": 875, "train_speed(iter/s)": 0.999322 }, { "acc": 0.74949503, "epoch": 2.7789473684210524, "grad_norm": 4.631548881530762, "learning_rate": 9.999113033708756e-06, "loss": 0.64674835, "memory(GiB)": 14.42, "step": 880, "train_speed(iter/s)": 0.9996 }, { "acc": 0.76464653, "epoch": 2.794736842105263, "grad_norm": 4.38846492767334, "learning_rate": 9.999011747685443e-06, "loss": 0.60613623, "memory(GiB)": 14.42, "step": 885, "train_speed(iter/s)": 0.999923 }, { "acc": 0.75858588, "epoch": 2.8105263157894735, "grad_norm": 5.623709201812744, "learning_rate": 9.998904987482465e-06, "loss": 0.61659794, "memory(GiB)": 14.42, "step": 890, "train_speed(iter/s)": 1.000227 }, { "acc": 0.7636364, "epoch": 2.8263157894736843, "grad_norm": 4.804795265197754, "learning_rate": 9.998792753216741e-06, "loss": 0.62756157, "memory(GiB)": 14.42, "step": 895, "train_speed(iter/s)": 1.000273 }, { "acc": 0.7666667, "epoch": 2.8421052631578947, "grad_norm": 5.261493682861328, "learning_rate": 9.998675045011186e-06, "loss": 0.60897455, "memory(GiB)": 14.42, "step": 900, "train_speed(iter/s)": 1.000569 }, { "acc": 0.7454546, "epoch": 2.8578947368421055, "grad_norm": 4.73408317565918, "learning_rate": 9.998551862994711e-06, "loss": 0.62260122, "memory(GiB)": 14.42, "step": 905, "train_speed(iter/s)": 1.000726 }, { "acc": 0.76868696, "epoch": 2.873684210526316, "grad_norm": 5.430922031402588, "learning_rate": 9.998423207302215e-06, "loss": 0.62232122, "memory(GiB)": 14.42, "step": 910, "train_speed(iter/s)": 1.000597 }, { "acc": 0.74646468, "epoch": 2.889473684210526, "grad_norm": 5.249138355255127, "learning_rate": 9.998289078074605e-06, "loss": 0.62488241, "memory(GiB)": 14.42, "step": 915, "train_speed(iter/s)": 1.000763 }, { "acc": 0.7666667, "epoch": 2.905263157894737, "grad_norm": 4.930274963378906, "learning_rate": 9.998149475458766e-06, "loss": 0.61784182, "memory(GiB)": 14.42, "step": 920, "train_speed(iter/s)": 1.000804 }, { "acc": 0.7575758, "epoch": 2.9210526315789473, "grad_norm": 5.189949989318848, "learning_rate": 9.998004399607592e-06, "loss": 0.61897421, "memory(GiB)": 14.42, "step": 925, "train_speed(iter/s)": 1.000775 }, { "acc": 0.76464648, "epoch": 2.9368421052631577, "grad_norm": 5.007906913757324, "learning_rate": 9.997853850679959e-06, "loss": 0.61194296, "memory(GiB)": 14.42, "step": 930, "train_speed(iter/s)": 1.000594 }, { "acc": 0.75353537, "epoch": 2.9526315789473685, "grad_norm": 4.766158103942871, "learning_rate": 9.997697828840749e-06, "loss": 0.64542856, "memory(GiB)": 14.42, "step": 935, "train_speed(iter/s)": 1.000421 }, { "acc": 0.76161613, "epoch": 2.968421052631579, "grad_norm": 4.673051834106445, "learning_rate": 9.997536334260826e-06, "loss": 0.6193069, "memory(GiB)": 14.42, "step": 940, "train_speed(iter/s)": 1.000437 }, { "acc": 0.76262627, "epoch": 2.984210526315789, "grad_norm": 5.344494342803955, "learning_rate": 9.997369367117051e-06, "loss": 0.62504797, "memory(GiB)": 14.42, "step": 945, "train_speed(iter/s)": 1.000649 }, { "acc": 0.7454546, "epoch": 3.0, "grad_norm": 5.629596710205078, "learning_rate": 9.997196927592286e-06, "loss": 0.63563552, "memory(GiB)": 14.42, "step": 950, "train_speed(iter/s)": 1.000471 }, { "acc": 0.7545455, "epoch": 3.0157894736842104, "grad_norm": 5.282176971435547, "learning_rate": 9.997019015875379e-06, "loss": 0.62507687, "memory(GiB)": 14.42, "step": 955, "train_speed(iter/s)": 1.000304 }, { "acc": 0.76060615, "epoch": 3.031578947368421, "grad_norm": 4.46095085144043, "learning_rate": 9.996835632161168e-06, "loss": 0.61408606, "memory(GiB)": 14.42, "step": 960, "train_speed(iter/s)": 1.000621 }, { "acc": 0.75050507, "epoch": 3.0473684210526315, "grad_norm": 6.5806660652160645, "learning_rate": 9.996646776650492e-06, "loss": 0.64194403, "memory(GiB)": 14.42, "step": 965, "train_speed(iter/s)": 1.000733 }, { "acc": 0.7575758, "epoch": 3.0631578947368423, "grad_norm": 4.796421051025391, "learning_rate": 9.996452449550176e-06, "loss": 0.63968344, "memory(GiB)": 14.42, "step": 970, "train_speed(iter/s)": 1.000961 }, { "acc": 0.75151515, "epoch": 3.0789473684210527, "grad_norm": 5.288400650024414, "learning_rate": 9.99625265107304e-06, "loss": 0.61021528, "memory(GiB)": 14.42, "step": 975, "train_speed(iter/s)": 1.001195 }, { "acc": 0.77373743, "epoch": 3.094736842105263, "grad_norm": 4.586065292358398, "learning_rate": 9.996047381437895e-06, "loss": 0.60301132, "memory(GiB)": 14.42, "step": 980, "train_speed(iter/s)": 1.001461 }, { "acc": 0.76565661, "epoch": 3.110526315789474, "grad_norm": 4.384403228759766, "learning_rate": 9.995836640869548e-06, "loss": 0.63142433, "memory(GiB)": 14.42, "step": 985, "train_speed(iter/s)": 1.001819 }, { "acc": 0.75959601, "epoch": 3.126315789473684, "grad_norm": 4.395773887634277, "learning_rate": 9.995620429598791e-06, "loss": 0.63195262, "memory(GiB)": 14.42, "step": 990, "train_speed(iter/s)": 1.001857 }, { "acc": 0.75858588, "epoch": 3.1421052631578945, "grad_norm": 4.790668964385986, "learning_rate": 9.995398747862412e-06, "loss": 0.60636911, "memory(GiB)": 14.42, "step": 995, "train_speed(iter/s)": 1.002024 }, { "acc": 0.77171721, "epoch": 3.1578947368421053, "grad_norm": 4.812241554260254, "learning_rate": 9.995171595903185e-06, "loss": 0.61093168, "memory(GiB)": 14.42, "step": 1000, "train_speed(iter/s)": 1.002015 }, { "acc": 0.77676773, "epoch": 3.1736842105263157, "grad_norm": 7.178610324859619, "learning_rate": 9.994938973969881e-06, "loss": 0.59977636, "memory(GiB)": 14.42, "step": 1005, "train_speed(iter/s)": 1.002083 }, { "acc": 0.74747481, "epoch": 3.1894736842105265, "grad_norm": 4.942941665649414, "learning_rate": 9.99470088231726e-06, "loss": 0.62871065, "memory(GiB)": 14.42, "step": 1010, "train_speed(iter/s)": 1.001901 }, { "acc": 0.77272735, "epoch": 3.205263157894737, "grad_norm": 6.371291160583496, "learning_rate": 9.994457321206067e-06, "loss": 0.63051238, "memory(GiB)": 14.42, "step": 1015, "train_speed(iter/s)": 1.001826 }, { "acc": 0.76565661, "epoch": 3.221052631578947, "grad_norm": 6.595971584320068, "learning_rate": 9.994208290903044e-06, "loss": 0.60574374, "memory(GiB)": 14.42, "step": 1020, "train_speed(iter/s)": 1.001531 }, { "acc": 0.77474747, "epoch": 3.236842105263158, "grad_norm": 6.768241882324219, "learning_rate": 9.993953791680919e-06, "loss": 0.6086751, "memory(GiB)": 14.42, "step": 1025, "train_speed(iter/s)": 1.001264 }, { "acc": 0.76060615, "epoch": 3.2526315789473683, "grad_norm": 5.642907619476318, "learning_rate": 9.993693823818408e-06, "loss": 0.61570983, "memory(GiB)": 14.42, "step": 1030, "train_speed(iter/s)": 1.001062 }, { "acc": 0.7757576, "epoch": 3.268421052631579, "grad_norm": 5.4705810546875, "learning_rate": 9.99342838760022e-06, "loss": 0.57923889, "memory(GiB)": 14.42, "step": 1035, "train_speed(iter/s)": 1.0012 }, { "acc": 0.7787879, "epoch": 3.2842105263157895, "grad_norm": 5.741537570953369, "learning_rate": 9.993157483317049e-06, "loss": 0.59299955, "memory(GiB)": 14.42, "step": 1040, "train_speed(iter/s)": 1.001369 }, { "acc": 0.77373743, "epoch": 3.3, "grad_norm": 6.200732707977295, "learning_rate": 9.99288111126558e-06, "loss": 0.6062952, "memory(GiB)": 14.42, "step": 1045, "train_speed(iter/s)": 1.001563 }, { "acc": 0.77272735, "epoch": 3.3157894736842106, "grad_norm": 5.042914867401123, "learning_rate": 9.992599271748487e-06, "loss": 0.59764023, "memory(GiB)": 14.42, "step": 1050, "train_speed(iter/s)": 1.001803 }, { "acc": 0.76161623, "epoch": 3.331578947368421, "grad_norm": 4.9619293212890625, "learning_rate": 9.992311965074426e-06, "loss": 0.60689812, "memory(GiB)": 14.42, "step": 1055, "train_speed(iter/s)": 1.001739 }, { "acc": 0.77070713, "epoch": 3.3473684210526318, "grad_norm": 4.4396820068359375, "learning_rate": 9.992019191558047e-06, "loss": 0.59743018, "memory(GiB)": 14.42, "step": 1060, "train_speed(iter/s)": 1.001876 }, { "acc": 0.76161623, "epoch": 3.363157894736842, "grad_norm": 4.762928009033203, "learning_rate": 9.991720951519983e-06, "loss": 0.60235682, "memory(GiB)": 14.42, "step": 1065, "train_speed(iter/s)": 1.002076 }, { "acc": 0.79595962, "epoch": 3.3789473684210525, "grad_norm": 6.026463508605957, "learning_rate": 9.991417245286857e-06, "loss": 0.56530128, "memory(GiB)": 14.42, "step": 1070, "train_speed(iter/s)": 1.002259 }, { "acc": 0.7666667, "epoch": 3.3947368421052633, "grad_norm": 5.576995372772217, "learning_rate": 9.991108073191275e-06, "loss": 0.60584636, "memory(GiB)": 14.42, "step": 1075, "train_speed(iter/s)": 1.002387 }, { "acc": 0.76565657, "epoch": 3.4105263157894736, "grad_norm": 5.6912336349487305, "learning_rate": 9.990793435571831e-06, "loss": 0.60695724, "memory(GiB)": 14.42, "step": 1080, "train_speed(iter/s)": 1.002498 }, { "acc": 0.77070708, "epoch": 3.4263157894736844, "grad_norm": 7.098496913909912, "learning_rate": 9.990473332773104e-06, "loss": 0.61024132, "memory(GiB)": 14.42, "step": 1085, "train_speed(iter/s)": 1.002286 }, { "acc": 0.77474747, "epoch": 3.442105263157895, "grad_norm": 5.199113845825195, "learning_rate": 9.990147765145657e-06, "loss": 0.57286983, "memory(GiB)": 14.42, "step": 1090, "train_speed(iter/s)": 1.002033 }, { "acc": 0.77171721, "epoch": 3.457894736842105, "grad_norm": 5.377429485321045, "learning_rate": 9.989816733046042e-06, "loss": 0.59407434, "memory(GiB)": 14.42, "step": 1095, "train_speed(iter/s)": 1.002361 }, { "acc": 0.78585863, "epoch": 3.473684210526316, "grad_norm": 8.064284324645996, "learning_rate": 9.98948023683679e-06, "loss": 0.59351382, "memory(GiB)": 14.42, "step": 1100, "train_speed(iter/s)": 1.00254 }, { "acc": 0.8151516, "epoch": 3.4894736842105263, "grad_norm": 5.125907897949219, "learning_rate": 9.989138276886422e-06, "loss": 0.55623984, "memory(GiB)": 14.42, "step": 1105, "train_speed(iter/s)": 1.002758 }, { "acc": 0.7818182, "epoch": 3.5052631578947366, "grad_norm": 5.503304481506348, "learning_rate": 9.988790853569435e-06, "loss": 0.58769279, "memory(GiB)": 14.42, "step": 1110, "train_speed(iter/s)": 1.002692 }, { "acc": 0.79494948, "epoch": 3.5210526315789474, "grad_norm": 5.440110683441162, "learning_rate": 9.988437967266318e-06, "loss": 0.54681435, "memory(GiB)": 14.42, "step": 1115, "train_speed(iter/s)": 1.002763 }, { "acc": 0.7787879, "epoch": 3.536842105263158, "grad_norm": 6.168666362762451, "learning_rate": 9.98807961836354e-06, "loss": 0.57537584, "memory(GiB)": 14.42, "step": 1120, "train_speed(iter/s)": 1.00306 }, { "acc": 0.78383842, "epoch": 3.5526315789473686, "grad_norm": 6.1219000816345215, "learning_rate": 9.987715807253544e-06, "loss": 0.56294775, "memory(GiB)": 14.42, "step": 1125, "train_speed(iter/s)": 1.003329 }, { "acc": 0.79898987, "epoch": 3.568421052631579, "grad_norm": 7.895573139190674, "learning_rate": 9.987346534334769e-06, "loss": 0.56549187, "memory(GiB)": 14.42, "step": 1130, "train_speed(iter/s)": 1.003635 }, { "acc": 0.76565661, "epoch": 3.5842105263157897, "grad_norm": 7.460114002227783, "learning_rate": 9.986971800011624e-06, "loss": 0.59118419, "memory(GiB)": 14.42, "step": 1135, "train_speed(iter/s)": 1.003749 }, { "acc": 0.78585863, "epoch": 3.6, "grad_norm": 5.666573524475098, "learning_rate": 9.986591604694506e-06, "loss": 0.55221477, "memory(GiB)": 14.42, "step": 1140, "train_speed(iter/s)": 1.004019 }, { "acc": 0.8060606, "epoch": 3.6157894736842104, "grad_norm": 5.389659404754639, "learning_rate": 9.986205948799792e-06, "loss": 0.55691719, "memory(GiB)": 14.42, "step": 1145, "train_speed(iter/s)": 1.004229 }, { "acc": 0.7757576, "epoch": 3.6315789473684212, "grad_norm": 6.799347877502441, "learning_rate": 9.985814832749832e-06, "loss": 0.61102791, "memory(GiB)": 14.42, "step": 1150, "train_speed(iter/s)": 1.004461 }, { "acc": 0.7787879, "epoch": 3.6473684210526316, "grad_norm": 5.257281303405762, "learning_rate": 9.98541825697297e-06, "loss": 0.57405472, "memory(GiB)": 14.42, "step": 1155, "train_speed(iter/s)": 1.004366 }, { "acc": 0.7757576, "epoch": 3.663157894736842, "grad_norm": 5.32186222076416, "learning_rate": 9.98501622190351e-06, "loss": 0.58651361, "memory(GiB)": 14.42, "step": 1160, "train_speed(iter/s)": 1.004375 }, { "acc": 0.79191923, "epoch": 3.6789473684210527, "grad_norm": 6.293529987335205, "learning_rate": 9.984608727981756e-06, "loss": 0.54780645, "memory(GiB)": 14.42, "step": 1165, "train_speed(iter/s)": 1.004371 }, { "acc": 0.78888893, "epoch": 3.694736842105263, "grad_norm": 5.983913898468018, "learning_rate": 9.984195775653973e-06, "loss": 0.56223335, "memory(GiB)": 14.42, "step": 1170, "train_speed(iter/s)": 1.004242 }, { "acc": 0.78585863, "epoch": 3.7105263157894735, "grad_norm": 6.722934722900391, "learning_rate": 9.983777365372412e-06, "loss": 0.56356654, "memory(GiB)": 14.42, "step": 1175, "train_speed(iter/s)": 1.004355 }, { "acc": 0.78080816, "epoch": 3.7263157894736842, "grad_norm": 6.902647495269775, "learning_rate": 9.983353497595303e-06, "loss": 0.56628561, "memory(GiB)": 14.42, "step": 1180, "train_speed(iter/s)": 1.004572 }, { "acc": 0.80505056, "epoch": 3.7421052631578946, "grad_norm": 6.423276901245117, "learning_rate": 9.982924172786847e-06, "loss": 0.51588736, "memory(GiB)": 14.42, "step": 1185, "train_speed(iter/s)": 1.004765 }, { "acc": 0.78888893, "epoch": 3.7578947368421054, "grad_norm": 7.339409828186035, "learning_rate": 9.982489391417225e-06, "loss": 0.55465469, "memory(GiB)": 14.42, "step": 1190, "train_speed(iter/s)": 1.004517 }, { "acc": 0.77979803, "epoch": 3.7736842105263158, "grad_norm": 5.153477191925049, "learning_rate": 9.982049153962591e-06, "loss": 0.57469625, "memory(GiB)": 14.42, "step": 1195, "train_speed(iter/s)": 1.004249 }, { "acc": 0.78989902, "epoch": 3.7894736842105265, "grad_norm": 6.078983306884766, "learning_rate": 9.98160346090508e-06, "loss": 0.56436605, "memory(GiB)": 14.42, "step": 1200, "train_speed(iter/s)": 1.004159 }, { "acc": 0.80505056, "epoch": 3.805263157894737, "grad_norm": 6.977550029754639, "learning_rate": 9.981152312732795e-06, "loss": 0.533461, "memory(GiB)": 14.42, "step": 1205, "train_speed(iter/s)": 1.004298 }, { "acc": 0.80101013, "epoch": 3.8210526315789473, "grad_norm": 5.750452518463135, "learning_rate": 9.980695709939817e-06, "loss": 0.53172665, "memory(GiB)": 14.42, "step": 1210, "train_speed(iter/s)": 1.004419 }, { "acc": 0.79292927, "epoch": 3.836842105263158, "grad_norm": 8.136344909667969, "learning_rate": 9.980233653026204e-06, "loss": 0.56949782, "memory(GiB)": 14.42, "step": 1215, "train_speed(iter/s)": 1.004574 }, { "acc": 0.79898987, "epoch": 3.8526315789473684, "grad_norm": 6.550292015075684, "learning_rate": 9.979766142497977e-06, "loss": 0.57330713, "memory(GiB)": 14.42, "step": 1220, "train_speed(iter/s)": 1.004753 }, { "acc": 0.79292936, "epoch": 3.8684210526315788, "grad_norm": 6.593601703643799, "learning_rate": 9.979293178867138e-06, "loss": 0.55137663, "memory(GiB)": 14.42, "step": 1225, "train_speed(iter/s)": 1.005004 }, { "acc": 0.80505056, "epoch": 3.8842105263157896, "grad_norm": 6.004925727844238, "learning_rate": 9.978814762651657e-06, "loss": 0.522402, "memory(GiB)": 14.42, "step": 1230, "train_speed(iter/s)": 1.005125 }, { "acc": 0.8060606, "epoch": 3.9, "grad_norm": 6.588228702545166, "learning_rate": 9.978330894375483e-06, "loss": 0.53442154, "memory(GiB)": 14.42, "step": 1235, "train_speed(iter/s)": 1.005384 }, { "acc": 0.80707073, "epoch": 3.9157894736842103, "grad_norm": 6.131303787231445, "learning_rate": 9.977841574568525e-06, "loss": 0.55704346, "memory(GiB)": 14.42, "step": 1240, "train_speed(iter/s)": 1.005664 }, { "acc": 0.79191923, "epoch": 3.931578947368421, "grad_norm": 8.15088176727295, "learning_rate": 9.977346803766668e-06, "loss": 0.55681534, "memory(GiB)": 14.42, "step": 1245, "train_speed(iter/s)": 1.005903 }, { "acc": 0.82020206, "epoch": 3.9473684210526314, "grad_norm": 5.325417518615723, "learning_rate": 9.976846582511768e-06, "loss": 0.5185771, "memory(GiB)": 14.42, "step": 1250, "train_speed(iter/s)": 1.00618 }, { "acc": 0.80404043, "epoch": 3.963157894736842, "grad_norm": 5.654501914978027, "learning_rate": 9.976340911351648e-06, "loss": 0.52004037, "memory(GiB)": 14.42, "step": 1255, "train_speed(iter/s)": 1.006248 }, { "acc": 0.79191923, "epoch": 3.9789473684210526, "grad_norm": 6.57743501663208, "learning_rate": 9.975829790840095e-06, "loss": 0.56735611, "memory(GiB)": 14.42, "step": 1260, "train_speed(iter/s)": 1.006459 }, { "acc": 0.79191923, "epoch": 3.9947368421052634, "grad_norm": 5.604462146759033, "learning_rate": 9.975313221536876e-06, "loss": 0.54044685, "memory(GiB)": 14.42, "step": 1265, "train_speed(iter/s)": 1.006423 }, { "acc": 0.79797983, "epoch": 4.010526315789473, "grad_norm": 6.081342697143555, "learning_rate": 9.974791204007714e-06, "loss": 0.55837898, "memory(GiB)": 14.42, "step": 1270, "train_speed(iter/s)": 1.005996 }, { "acc": 0.80808086, "epoch": 4.026315789473684, "grad_norm": 6.739323616027832, "learning_rate": 9.974263738824304e-06, "loss": 0.49736958, "memory(GiB)": 14.42, "step": 1275, "train_speed(iter/s)": 1.005807 }, { "acc": 0.81717176, "epoch": 4.042105263157895, "grad_norm": 6.606033802032471, "learning_rate": 9.973730826564306e-06, "loss": 0.50827093, "memory(GiB)": 14.42, "step": 1280, "train_speed(iter/s)": 1.005844 }, { "acc": 0.79797983, "epoch": 4.057894736842106, "grad_norm": 7.347227096557617, "learning_rate": 9.973192467811341e-06, "loss": 0.5242722, "memory(GiB)": 14.42, "step": 1285, "train_speed(iter/s)": 1.006037 }, { "acc": 0.82323236, "epoch": 4.073684210526316, "grad_norm": 6.304427146911621, "learning_rate": 9.972648663155006e-06, "loss": 0.49846554, "memory(GiB)": 14.42, "step": 1290, "train_speed(iter/s)": 1.006178 }, { "acc": 0.81010103, "epoch": 4.089473684210526, "grad_norm": 7.817327976226807, "learning_rate": 9.97209941319085e-06, "loss": 0.5091485, "memory(GiB)": 14.42, "step": 1295, "train_speed(iter/s)": 1.006371 }, { "acc": 0.82222214, "epoch": 4.105263157894737, "grad_norm": 7.094549179077148, "learning_rate": 9.971544718520392e-06, "loss": 0.49558301, "memory(GiB)": 14.42, "step": 1300, "train_speed(iter/s)": 1.006509 }, { "acc": 0.81313133, "epoch": 4.121052631578947, "grad_norm": 8.023250579833984, "learning_rate": 9.970984579751112e-06, "loss": 0.50098181, "memory(GiB)": 14.42, "step": 1305, "train_speed(iter/s)": 1.006667 }, { "acc": 0.8333334, "epoch": 4.136842105263158, "grad_norm": 7.398085117340088, "learning_rate": 9.970418997496456e-06, "loss": 0.48294721, "memory(GiB)": 14.42, "step": 1310, "train_speed(iter/s)": 1.006569 }, { "acc": 0.8181819, "epoch": 4.152631578947369, "grad_norm": 7.375761032104492, "learning_rate": 9.969847972375823e-06, "loss": 0.50575571, "memory(GiB)": 14.42, "step": 1315, "train_speed(iter/s)": 1.006525 }, { "acc": 0.78888893, "epoch": 4.168421052631579, "grad_norm": 6.804074287414551, "learning_rate": 9.96927150501458e-06, "loss": 0.55558271, "memory(GiB)": 14.42, "step": 1320, "train_speed(iter/s)": 1.006562 }, { "acc": 0.8151516, "epoch": 4.184210526315789, "grad_norm": 6.9089131355285645, "learning_rate": 9.968689596044054e-06, "loss": 0.50905294, "memory(GiB)": 14.42, "step": 1325, "train_speed(iter/s)": 1.006623 }, { "acc": 0.81717176, "epoch": 4.2, "grad_norm": 7.5984086990356445, "learning_rate": 9.968102246101527e-06, "loss": 0.50087404, "memory(GiB)": 14.42, "step": 1330, "train_speed(iter/s)": 1.006669 }, { "acc": 0.81616163, "epoch": 4.215789473684211, "grad_norm": 6.235317707061768, "learning_rate": 9.967509455830244e-06, "loss": 0.48131518, "memory(GiB)": 14.42, "step": 1335, "train_speed(iter/s)": 1.006806 }, { "acc": 0.81717176, "epoch": 4.231578947368421, "grad_norm": 7.801043510437012, "learning_rate": 9.966911225879407e-06, "loss": 0.48789158, "memory(GiB)": 14.42, "step": 1340, "train_speed(iter/s)": 1.006723 }, { "acc": 0.80505056, "epoch": 4.247368421052632, "grad_norm": 7.775477409362793, "learning_rate": 9.966307556904172e-06, "loss": 0.50765548, "memory(GiB)": 14.42, "step": 1345, "train_speed(iter/s)": 1.006702 }, { "acc": 0.8121212, "epoch": 4.2631578947368425, "grad_norm": 7.874118804931641, "learning_rate": 9.965698449565654e-06, "loss": 0.47103491, "memory(GiB)": 14.42, "step": 1350, "train_speed(iter/s)": 1.006579 }, { "acc": 0.82828283, "epoch": 4.278947368421052, "grad_norm": 7.379849433898926, "learning_rate": 9.965083904530928e-06, "loss": 0.48206449, "memory(GiB)": 14.42, "step": 1355, "train_speed(iter/s)": 1.006487 }, { "acc": 0.8242425, "epoch": 4.294736842105263, "grad_norm": 8.313944816589355, "learning_rate": 9.964463922473016e-06, "loss": 0.50168781, "memory(GiB)": 14.42, "step": 1360, "train_speed(iter/s)": 1.006278 }, { "acc": 0.83535366, "epoch": 4.310526315789474, "grad_norm": 6.975119113922119, "learning_rate": 9.963838504070901e-06, "loss": 0.48316975, "memory(GiB)": 14.42, "step": 1365, "train_speed(iter/s)": 1.006081 }, { "acc": 0.82626266, "epoch": 4.326315789473684, "grad_norm": 7.731008052825928, "learning_rate": 9.96320765000952e-06, "loss": 0.50050735, "memory(GiB)": 14.42, "step": 1370, "train_speed(iter/s)": 1.005849 }, { "acc": 0.82828293, "epoch": 4.342105263157895, "grad_norm": 7.532197952270508, "learning_rate": 9.962571360979756e-06, "loss": 0.48054743, "memory(GiB)": 14.42, "step": 1375, "train_speed(iter/s)": 1.005899 }, { "acc": 0.82929296, "epoch": 4.3578947368421055, "grad_norm": 6.518101215362549, "learning_rate": 9.961929637678449e-06, "loss": 0.47691903, "memory(GiB)": 14.42, "step": 1380, "train_speed(iter/s)": 1.006048 }, { "acc": 0.83737373, "epoch": 4.373684210526315, "grad_norm": 6.837509632110596, "learning_rate": 9.961282480808391e-06, "loss": 0.48503637, "memory(GiB)": 14.42, "step": 1385, "train_speed(iter/s)": 1.006208 }, { "acc": 0.79191923, "epoch": 4.389473684210526, "grad_norm": 7.671966552734375, "learning_rate": 9.960629891078325e-06, "loss": 0.53532209, "memory(GiB)": 14.42, "step": 1390, "train_speed(iter/s)": 1.006186 }, { "acc": 0.83131313, "epoch": 4.405263157894737, "grad_norm": 6.7305006980896, "learning_rate": 9.95997186920294e-06, "loss": 0.47744513, "memory(GiB)": 14.42, "step": 1395, "train_speed(iter/s)": 1.006212 }, { "acc": 0.82626266, "epoch": 4.421052631578947, "grad_norm": 6.894629955291748, "learning_rate": 9.959308415902876e-06, "loss": 0.46218028, "memory(GiB)": 14.42, "step": 1400, "train_speed(iter/s)": 1.006345 }, { "acc": 0.8181819, "epoch": 4.436842105263158, "grad_norm": 7.045905113220215, "learning_rate": 9.958639531904725e-06, "loss": 0.46672087, "memory(GiB)": 14.42, "step": 1405, "train_speed(iter/s)": 1.006472 }, { "acc": 0.8303031, "epoch": 4.4526315789473685, "grad_norm": 7.43247652053833, "learning_rate": 9.957965217941023e-06, "loss": 0.47453899, "memory(GiB)": 14.42, "step": 1410, "train_speed(iter/s)": 1.006343 }, { "acc": 0.8363636, "epoch": 4.468421052631579, "grad_norm": 7.158817768096924, "learning_rate": 9.957285474750248e-06, "loss": 0.47886271, "memory(GiB)": 14.42, "step": 1415, "train_speed(iter/s)": 1.00641 }, { "acc": 0.82626266, "epoch": 4.484210526315789, "grad_norm": 6.566321849822998, "learning_rate": 9.956600303076834e-06, "loss": 0.45864029, "memory(GiB)": 14.42, "step": 1420, "train_speed(iter/s)": 1.006614 }, { "acc": 0.83838387, "epoch": 4.5, "grad_norm": 7.215709686279297, "learning_rate": 9.955909703671154e-06, "loss": 0.46511064, "memory(GiB)": 14.42, "step": 1425, "train_speed(iter/s)": 1.006816 }, { "acc": 0.82525253, "epoch": 4.515789473684211, "grad_norm": 8.241830825805664, "learning_rate": 9.955213677289523e-06, "loss": 0.49213314, "memory(GiB)": 14.42, "step": 1430, "train_speed(iter/s)": 1.006992 }, { "acc": 0.85959597, "epoch": 4.531578947368421, "grad_norm": 7.597438335418701, "learning_rate": 9.954512224694207e-06, "loss": 0.42885356, "memory(GiB)": 14.42, "step": 1435, "train_speed(iter/s)": 1.006947 }, { "acc": 0.8333334, "epoch": 4.5473684210526315, "grad_norm": 7.716129779815674, "learning_rate": 9.953805346653407e-06, "loss": 0.48312497, "memory(GiB)": 14.42, "step": 1440, "train_speed(iter/s)": 1.007053 }, { "acc": 0.8272727, "epoch": 4.563157894736842, "grad_norm": 7.242175579071045, "learning_rate": 9.953093043941271e-06, "loss": 0.46601577, "memory(GiB)": 14.42, "step": 1445, "train_speed(iter/s)": 1.006919 }, { "acc": 0.8272728, "epoch": 4.578947368421053, "grad_norm": 8.561416625976562, "learning_rate": 9.952375317337885e-06, "loss": 0.47396536, "memory(GiB)": 14.42, "step": 1450, "train_speed(iter/s)": 1.006824 }, { "acc": 0.82929296, "epoch": 4.594736842105263, "grad_norm": 7.143245220184326, "learning_rate": 9.951652167629274e-06, "loss": 0.47127285, "memory(GiB)": 14.42, "step": 1455, "train_speed(iter/s)": 1.006945 }, { "acc": 0.8303031, "epoch": 4.610526315789474, "grad_norm": 7.557952880859375, "learning_rate": 9.950923595607407e-06, "loss": 0.47242546, "memory(GiB)": 14.42, "step": 1460, "train_speed(iter/s)": 1.006993 }, { "acc": 0.83131313, "epoch": 4.626315789473685, "grad_norm": 6.552892208099365, "learning_rate": 9.950189602070185e-06, "loss": 0.45216031, "memory(GiB)": 14.42, "step": 1465, "train_speed(iter/s)": 1.007145 }, { "acc": 0.87070704, "epoch": 4.6421052631578945, "grad_norm": 6.04880428314209, "learning_rate": 9.949450187821455e-06, "loss": 0.36264579, "memory(GiB)": 14.42, "step": 1470, "train_speed(iter/s)": 1.007325 }, { "acc": 0.85252533, "epoch": 4.657894736842105, "grad_norm": 8.390496253967285, "learning_rate": 9.94870535367099e-06, "loss": 0.42680759, "memory(GiB)": 14.42, "step": 1475, "train_speed(iter/s)": 1.007556 }, { "acc": 0.84343443, "epoch": 4.673684210526316, "grad_norm": 8.083012580871582, "learning_rate": 9.947955100434505e-06, "loss": 0.45011797, "memory(GiB)": 14.42, "step": 1480, "train_speed(iter/s)": 1.007794 }, { "acc": 0.83535357, "epoch": 4.689473684210526, "grad_norm": 7.6660590171813965, "learning_rate": 9.947199428933652e-06, "loss": 0.47048602, "memory(GiB)": 14.42, "step": 1485, "train_speed(iter/s)": 1.008031 }, { "acc": 0.82323227, "epoch": 4.705263157894737, "grad_norm": 7.676632881164551, "learning_rate": 9.94643833999601e-06, "loss": 0.48105483, "memory(GiB)": 14.42, "step": 1490, "train_speed(iter/s)": 1.008291 }, { "acc": 0.83434353, "epoch": 4.721052631578948, "grad_norm": 7.969620704650879, "learning_rate": 9.945671834455097e-06, "loss": 0.46031218, "memory(GiB)": 14.42, "step": 1495, "train_speed(iter/s)": 1.008525 }, { "acc": 0.82020206, "epoch": 4.7368421052631575, "grad_norm": 9.78809928894043, "learning_rate": 9.94489991315036e-06, "loss": 0.49499507, "memory(GiB)": 14.42, "step": 1500, "train_speed(iter/s)": 1.008732 }, { "acc": 0.8303031, "epoch": 4.752631578947368, "grad_norm": 7.448624134063721, "learning_rate": 9.944122576927178e-06, "loss": 0.43737659, "memory(GiB)": 14.42, "step": 1505, "train_speed(iter/s)": 1.00888 }, { "acc": 0.8363636, "epoch": 4.768421052631579, "grad_norm": 7.217111110687256, "learning_rate": 9.943339826636857e-06, "loss": 0.43941054, "memory(GiB)": 14.42, "step": 1510, "train_speed(iter/s)": 1.008827 }, { "acc": 0.84949493, "epoch": 4.784210526315789, "grad_norm": 7.279267311096191, "learning_rate": 9.942551663136639e-06, "loss": 0.41463737, "memory(GiB)": 14.42, "step": 1515, "train_speed(iter/s)": 1.008971 }, { "acc": 0.8363636, "epoch": 4.8, "grad_norm": 8.194135665893555, "learning_rate": 9.941758087289692e-06, "loss": 0.4422389, "memory(GiB)": 14.42, "step": 1520, "train_speed(iter/s)": 1.008925 }, { "acc": 0.82929296, "epoch": 4.815789473684211, "grad_norm": 7.623429775238037, "learning_rate": 9.940959099965104e-06, "loss": 0.47428942, "memory(GiB)": 14.42, "step": 1525, "train_speed(iter/s)": 1.009105 }, { "acc": 0.8484848, "epoch": 4.831578947368421, "grad_norm": 9.262815475463867, "learning_rate": 9.9401547020379e-06, "loss": 0.41575365, "memory(GiB)": 14.42, "step": 1530, "train_speed(iter/s)": 1.009319 }, { "acc": 0.8484848, "epoch": 4.847368421052631, "grad_norm": 8.296936988830566, "learning_rate": 9.939344894389026e-06, "loss": 0.47690134, "memory(GiB)": 14.42, "step": 1535, "train_speed(iter/s)": 1.009534 }, { "acc": 0.84343433, "epoch": 4.863157894736842, "grad_norm": 7.545134544372559, "learning_rate": 9.938529677905353e-06, "loss": 0.44248581, "memory(GiB)": 14.42, "step": 1540, "train_speed(iter/s)": 1.009587 }, { "acc": 0.86868687, "epoch": 4.878947368421053, "grad_norm": 7.578943252563477, "learning_rate": 9.93770905347967e-06, "loss": 0.40284214, "memory(GiB)": 14.42, "step": 1545, "train_speed(iter/s)": 1.009501 }, { "acc": 0.8545455, "epoch": 4.894736842105263, "grad_norm": 7.998993396759033, "learning_rate": 9.9368830220107e-06, "loss": 0.40931978, "memory(GiB)": 14.42, "step": 1550, "train_speed(iter/s)": 1.009094 }, { "acc": 0.83535347, "epoch": 4.910526315789474, "grad_norm": 6.669436454772949, "learning_rate": 9.936051584403078e-06, "loss": 0.43530879, "memory(GiB)": 14.42, "step": 1555, "train_speed(iter/s)": 1.00892 }, { "acc": 0.82929296, "epoch": 4.926315789473684, "grad_norm": 8.15095329284668, "learning_rate": 9.935214741567361e-06, "loss": 0.44833851, "memory(GiB)": 14.42, "step": 1560, "train_speed(iter/s)": 1.008858 }, { "acc": 0.86565657, "epoch": 4.942105263157895, "grad_norm": 10.634395599365234, "learning_rate": 9.934372494420032e-06, "loss": 0.40384045, "memory(GiB)": 14.42, "step": 1565, "train_speed(iter/s)": 1.008755 }, { "acc": 0.85656567, "epoch": 4.957894736842105, "grad_norm": 8.751354217529297, "learning_rate": 9.933524843883484e-06, "loss": 0.40516777, "memory(GiB)": 14.42, "step": 1570, "train_speed(iter/s)": 1.008868 }, { "acc": 0.8575758, "epoch": 4.973684210526316, "grad_norm": 9.245224952697754, "learning_rate": 9.932671790886033e-06, "loss": 0.42211332, "memory(GiB)": 14.42, "step": 1575, "train_speed(iter/s)": 1.008875 }, { "acc": 0.8515152, "epoch": 4.989473684210527, "grad_norm": 8.566627502441406, "learning_rate": 9.93181333636191e-06, "loss": 0.43648109, "memory(GiB)": 14.42, "step": 1580, "train_speed(iter/s)": 1.008726 }, { "acc": 0.87171726, "epoch": 5.005263157894737, "grad_norm": 8.230137825012207, "learning_rate": 9.930949481251263e-06, "loss": 0.38181038, "memory(GiB)": 14.42, "step": 1585, "train_speed(iter/s)": 1.008325 }, { "acc": 0.869697, "epoch": 5.021052631578947, "grad_norm": 8.635025978088379, "learning_rate": 9.930080226500154e-06, "loss": 0.34935093, "memory(GiB)": 14.42, "step": 1590, "train_speed(iter/s)": 1.008136 }, { "acc": 0.85555553, "epoch": 5.036842105263158, "grad_norm": 8.816386222839355, "learning_rate": 9.929205573060552e-06, "loss": 0.4156651, "memory(GiB)": 14.42, "step": 1595, "train_speed(iter/s)": 1.008151 }, { "acc": 0.85959597, "epoch": 5.052631578947368, "grad_norm": 7.993004322052002, "learning_rate": 9.928325521890351e-06, "loss": 0.38215671, "memory(GiB)": 14.42, "step": 1600, "train_speed(iter/s)": 1.008242 }, { "acc": 0.8575758, "epoch": 5.068421052631579, "grad_norm": 9.520785331726074, "learning_rate": 9.927440073953346e-06, "loss": 0.42380819, "memory(GiB)": 14.42, "step": 1605, "train_speed(iter/s)": 1.008391 }, { "acc": 0.83737373, "epoch": 5.08421052631579, "grad_norm": 9.796367645263672, "learning_rate": 9.926549230219245e-06, "loss": 0.42773304, "memory(GiB)": 14.42, "step": 1610, "train_speed(iter/s)": 1.008415 }, { "acc": 0.85858593, "epoch": 5.1, "grad_norm": 8.989090919494629, "learning_rate": 9.925652991663668e-06, "loss": 0.41599984, "memory(GiB)": 14.42, "step": 1615, "train_speed(iter/s)": 1.008499 }, { "acc": 0.86565666, "epoch": 5.11578947368421, "grad_norm": 8.325226783752441, "learning_rate": 9.924751359268142e-06, "loss": 0.37139072, "memory(GiB)": 14.42, "step": 1620, "train_speed(iter/s)": 1.008669 }, { "acc": 0.8787879, "epoch": 5.131578947368421, "grad_norm": 9.19404125213623, "learning_rate": 9.923844334020099e-06, "loss": 0.33956528, "memory(GiB)": 14.42, "step": 1625, "train_speed(iter/s)": 1.008823 }, { "acc": 0.84343433, "epoch": 5.147368421052631, "grad_norm": 6.872360706329346, "learning_rate": 9.922931916912877e-06, "loss": 0.4026576, "memory(GiB)": 14.42, "step": 1630, "train_speed(iter/s)": 1.008627 }, { "acc": 0.8727273, "epoch": 5.163157894736842, "grad_norm": 8.282492637634277, "learning_rate": 9.922014108945724e-06, "loss": 0.35363464, "memory(GiB)": 14.42, "step": 1635, "train_speed(iter/s)": 1.008475 }, { "acc": 0.869697, "epoch": 5.178947368421053, "grad_norm": 9.004985809326172, "learning_rate": 9.921090911123785e-06, "loss": 0.37809303, "memory(GiB)": 14.42, "step": 1640, "train_speed(iter/s)": 1.008427 }, { "acc": 0.86161613, "epoch": 5.1947368421052635, "grad_norm": 10.675838470458984, "learning_rate": 9.920162324458115e-06, "loss": 0.39833202, "memory(GiB)": 14.42, "step": 1645, "train_speed(iter/s)": 1.008556 }, { "acc": 0.86767683, "epoch": 5.2105263157894735, "grad_norm": 9.150327682495117, "learning_rate": 9.919228349965663e-06, "loss": 0.40053234, "memory(GiB)": 14.42, "step": 1650, "train_speed(iter/s)": 1.008788 }, { "acc": 0.87070713, "epoch": 5.226315789473684, "grad_norm": 8.315364837646484, "learning_rate": 9.918288988669284e-06, "loss": 0.38295619, "memory(GiB)": 14.42, "step": 1655, "train_speed(iter/s)": 1.009029 }, { "acc": 0.84949493, "epoch": 5.242105263157895, "grad_norm": 10.003687858581543, "learning_rate": 9.91734424159773e-06, "loss": 0.39980845, "memory(GiB)": 14.42, "step": 1660, "train_speed(iter/s)": 1.009171 }, { "acc": 0.86868687, "epoch": 5.257894736842105, "grad_norm": 6.885008335113525, "learning_rate": 9.916394109785654e-06, "loss": 0.35017097, "memory(GiB)": 14.42, "step": 1665, "train_speed(iter/s)": 1.009269 }, { "acc": 0.869697, "epoch": 5.273684210526316, "grad_norm": 9.228471755981445, "learning_rate": 9.9154385942736e-06, "loss": 0.3592062, "memory(GiB)": 14.42, "step": 1670, "train_speed(iter/s)": 1.00922 }, { "acc": 0.86161613, "epoch": 5.2894736842105265, "grad_norm": 9.5510835647583, "learning_rate": 9.914477696108015e-06, "loss": 0.3855813, "memory(GiB)": 14.42, "step": 1675, "train_speed(iter/s)": 1.008977 }, { "acc": 0.88383846, "epoch": 5.3052631578947365, "grad_norm": 7.740030765533447, "learning_rate": 9.91351141634124e-06, "loss": 0.34937143, "memory(GiB)": 14.42, "step": 1680, "train_speed(iter/s)": 1.008844 }, { "acc": 0.8575758, "epoch": 5.321052631578947, "grad_norm": 9.399805068969727, "learning_rate": 9.912539756031506e-06, "loss": 0.38550258, "memory(GiB)": 14.42, "step": 1685, "train_speed(iter/s)": 1.008777 }, { "acc": 0.86868687, "epoch": 5.336842105263158, "grad_norm": 9.374094009399414, "learning_rate": 9.911562716242933e-06, "loss": 0.37300575, "memory(GiB)": 14.42, "step": 1690, "train_speed(iter/s)": 1.008623 }, { "acc": 0.85353537, "epoch": 5.352631578947369, "grad_norm": 8.98189640045166, "learning_rate": 9.910580298045547e-06, "loss": 0.39055982, "memory(GiB)": 14.42, "step": 1695, "train_speed(iter/s)": 1.008438 }, { "acc": 0.85656567, "epoch": 5.368421052631579, "grad_norm": 7.874349594116211, "learning_rate": 9.90959250251525e-06, "loss": 0.38435912, "memory(GiB)": 14.42, "step": 1700, "train_speed(iter/s)": 1.008244 }, { "acc": 0.87171717, "epoch": 5.38421052631579, "grad_norm": 8.327530860900879, "learning_rate": 9.908599330733839e-06, "loss": 0.35224979, "memory(GiB)": 14.42, "step": 1705, "train_speed(iter/s)": 1.008096 }, { "acc": 0.87474747, "epoch": 5.4, "grad_norm": 9.212203979492188, "learning_rate": 9.907600783788997e-06, "loss": 0.35220504, "memory(GiB)": 14.42, "step": 1710, "train_speed(iter/s)": 1.008077 }, { "acc": 0.869697, "epoch": 5.41578947368421, "grad_norm": 10.622110366821289, "learning_rate": 9.906596862774297e-06, "loss": 0.38187442, "memory(GiB)": 14.42, "step": 1715, "train_speed(iter/s)": 1.008041 }, { "acc": 0.869697, "epoch": 5.431578947368421, "grad_norm": 8.22000503540039, "learning_rate": 9.905587568789192e-06, "loss": 0.35868003, "memory(GiB)": 14.42, "step": 1720, "train_speed(iter/s)": 1.00814 }, { "acc": 0.87171726, "epoch": 5.447368421052632, "grad_norm": 9.036218643188477, "learning_rate": 9.904572902939023e-06, "loss": 0.33650646, "memory(GiB)": 14.42, "step": 1725, "train_speed(iter/s)": 1.008217 }, { "acc": 0.86565657, "epoch": 5.463157894736842, "grad_norm": 8.017346382141113, "learning_rate": 9.903552866335014e-06, "loss": 0.39399965, "memory(GiB)": 14.42, "step": 1730, "train_speed(iter/s)": 1.008266 }, { "acc": 0.8878788, "epoch": 5.478947368421053, "grad_norm": 7.885061740875244, "learning_rate": 9.902527460094271e-06, "loss": 0.32297885, "memory(GiB)": 14.42, "step": 1735, "train_speed(iter/s)": 1.008123 }, { "acc": 0.88383846, "epoch": 5.494736842105263, "grad_norm": 6.989620208740234, "learning_rate": 9.90149668533978e-06, "loss": 0.32285008, "memory(GiB)": 14.42, "step": 1740, "train_speed(iter/s)": 1.007983 }, { "acc": 0.8757576, "epoch": 5.510526315789473, "grad_norm": 10.041963577270508, "learning_rate": 9.900460543200403e-06, "loss": 0.36242232, "memory(GiB)": 14.42, "step": 1745, "train_speed(iter/s)": 1.007954 }, { "acc": 0.86868687, "epoch": 5.526315789473684, "grad_norm": 8.784823417663574, "learning_rate": 9.899419034810887e-06, "loss": 0.3428535, "memory(GiB)": 14.42, "step": 1750, "train_speed(iter/s)": 1.008053 }, { "acc": 0.88585854, "epoch": 5.542105263157895, "grad_norm": 8.738446235656738, "learning_rate": 9.898372161311851e-06, "loss": 0.34290485, "memory(GiB)": 14.42, "step": 1755, "train_speed(iter/s)": 1.00821 }, { "acc": 0.8848485, "epoch": 5.557894736842105, "grad_norm": 7.158604145050049, "learning_rate": 9.897319923849792e-06, "loss": 0.34065032, "memory(GiB)": 14.42, "step": 1760, "train_speed(iter/s)": 1.008285 }, { "acc": 0.86565666, "epoch": 5.573684210526316, "grad_norm": 9.903304100036621, "learning_rate": 9.896262323577079e-06, "loss": 0.39218345, "memory(GiB)": 14.42, "step": 1765, "train_speed(iter/s)": 1.008471 }, { "acc": 0.89595966, "epoch": 5.589473684210526, "grad_norm": 8.26339340209961, "learning_rate": 9.895199361651956e-06, "loss": 0.28524117, "memory(GiB)": 14.42, "step": 1770, "train_speed(iter/s)": 1.008633 }, { "acc": 0.89191914, "epoch": 5.605263157894737, "grad_norm": 8.691139221191406, "learning_rate": 9.894131039238538e-06, "loss": 0.31279945, "memory(GiB)": 14.42, "step": 1775, "train_speed(iter/s)": 1.008776 }, { "acc": 0.87878799, "epoch": 5.621052631578947, "grad_norm": 10.063474655151367, "learning_rate": 9.893057357506812e-06, "loss": 0.35178609, "memory(GiB)": 14.42, "step": 1780, "train_speed(iter/s)": 1.008879 }, { "acc": 0.8848485, "epoch": 5.636842105263158, "grad_norm": 11.359123229980469, "learning_rate": 9.891978317632631e-06, "loss": 0.34196472, "memory(GiB)": 14.42, "step": 1785, "train_speed(iter/s)": 1.009038 }, { "acc": 0.88989897, "epoch": 5.652631578947369, "grad_norm": 9.00007152557373, "learning_rate": 9.890893920797723e-06, "loss": 0.32037597, "memory(GiB)": 14.42, "step": 1790, "train_speed(iter/s)": 1.00922 }, { "acc": 0.88383846, "epoch": 5.668421052631579, "grad_norm": 8.761248588562012, "learning_rate": 9.889804168189673e-06, "loss": 0.33403468, "memory(GiB)": 14.42, "step": 1795, "train_speed(iter/s)": 1.009263 }, { "acc": 0.88282833, "epoch": 5.684210526315789, "grad_norm": 8.742795944213867, "learning_rate": 9.88870906100194e-06, "loss": 0.33450744, "memory(GiB)": 14.42, "step": 1800, "train_speed(iter/s)": 1.00941 }, { "acc": 0.87070713, "epoch": 5.7, "grad_norm": 8.495540618896484, "learning_rate": 9.88760860043384e-06, "loss": 0.36525402, "memory(GiB)": 14.42, "step": 1805, "train_speed(iter/s)": 1.009578 }, { "acc": 0.8818182, "epoch": 5.715789473684211, "grad_norm": 8.872817993164062, "learning_rate": 9.88650278769056e-06, "loss": 0.33951588, "memory(GiB)": 14.42, "step": 1810, "train_speed(iter/s)": 1.009712 }, { "acc": 0.88484859, "epoch": 5.731578947368421, "grad_norm": 9.544350624084473, "learning_rate": 9.885391623983142e-06, "loss": 0.32934637, "memory(GiB)": 14.42, "step": 1815, "train_speed(iter/s)": 1.009819 }, { "acc": 0.89797983, "epoch": 5.747368421052632, "grad_norm": 7.9700117111206055, "learning_rate": 9.884275110528489e-06, "loss": 0.29957988, "memory(GiB)": 14.42, "step": 1820, "train_speed(iter/s)": 1.009811 }, { "acc": 0.87171717, "epoch": 5.7631578947368425, "grad_norm": 9.927080154418945, "learning_rate": 9.883153248549367e-06, "loss": 0.35283065, "memory(GiB)": 14.42, "step": 1825, "train_speed(iter/s)": 1.009745 }, { "acc": 0.8878788, "epoch": 5.778947368421052, "grad_norm": 8.701126098632812, "learning_rate": 9.882026039274394e-06, "loss": 0.30775204, "memory(GiB)": 14.42, "step": 1830, "train_speed(iter/s)": 1.009604 }, { "acc": 0.87979803, "epoch": 5.794736842105263, "grad_norm": 9.421489715576172, "learning_rate": 9.880893483938046e-06, "loss": 0.3247303, "memory(GiB)": 14.42, "step": 1835, "train_speed(iter/s)": 1.009598 }, { "acc": 0.87171717, "epoch": 5.810526315789474, "grad_norm": 11.115700721740723, "learning_rate": 9.879755583780655e-06, "loss": 0.34211159, "memory(GiB)": 14.42, "step": 1840, "train_speed(iter/s)": 1.009748 }, { "acc": 0.89898987, "epoch": 5.826315789473684, "grad_norm": 8.85307502746582, "learning_rate": 9.878612340048409e-06, "loss": 0.28415699, "memory(GiB)": 14.42, "step": 1845, "train_speed(iter/s)": 1.009816 }, { "acc": 0.89898987, "epoch": 5.842105263157895, "grad_norm": 8.639701843261719, "learning_rate": 9.877463753993341e-06, "loss": 0.29079266, "memory(GiB)": 14.42, "step": 1850, "train_speed(iter/s)": 1.009713 }, { "acc": 0.90808086, "epoch": 5.8578947368421055, "grad_norm": 7.4572319984436035, "learning_rate": 9.876309826873343e-06, "loss": 0.27632694, "memory(GiB)": 14.42, "step": 1855, "train_speed(iter/s)": 1.009548 }, { "acc": 0.88888893, "epoch": 5.873684210526315, "grad_norm": 9.328161239624023, "learning_rate": 9.875150559952146e-06, "loss": 0.31321073, "memory(GiB)": 14.42, "step": 1860, "train_speed(iter/s)": 1.009361 }, { "acc": 0.88686876, "epoch": 5.889473684210526, "grad_norm": 8.32144832611084, "learning_rate": 9.873985954499339e-06, "loss": 0.32595744, "memory(GiB)": 14.42, "step": 1865, "train_speed(iter/s)": 1.009233 }, { "acc": 0.89898987, "epoch": 5.905263157894737, "grad_norm": 8.621851921081543, "learning_rate": 9.872816011790353e-06, "loss": 0.28815918, "memory(GiB)": 14.42, "step": 1870, "train_speed(iter/s)": 1.009126 }, { "acc": 0.88282833, "epoch": 5.921052631578947, "grad_norm": 8.001331329345703, "learning_rate": 9.871640733106467e-06, "loss": 0.32237473, "memory(GiB)": 14.42, "step": 1875, "train_speed(iter/s)": 1.009003 }, { "acc": 0.90202026, "epoch": 5.936842105263158, "grad_norm": 9.54694652557373, "learning_rate": 9.870460119734797e-06, "loss": 0.28403633, "memory(GiB)": 14.42, "step": 1880, "train_speed(iter/s)": 1.008879 }, { "acc": 0.89797974, "epoch": 5.9526315789473685, "grad_norm": 8.340572357177734, "learning_rate": 9.869274172968306e-06, "loss": 0.29031484, "memory(GiB)": 14.42, "step": 1885, "train_speed(iter/s)": 1.008727 }, { "acc": 0.89494953, "epoch": 5.968421052631579, "grad_norm": 8.185052871704102, "learning_rate": 9.868082894105802e-06, "loss": 0.27898729, "memory(GiB)": 14.42, "step": 1890, "train_speed(iter/s)": 1.008673 }, { "acc": 0.88080807, "epoch": 5.984210526315789, "grad_norm": 9.495261192321777, "learning_rate": 9.866886284451926e-06, "loss": 0.32206149, "memory(GiB)": 14.42, "step": 1895, "train_speed(iter/s)": 1.00858 }, { "acc": 0.89292936, "epoch": 6.0, "grad_norm": 10.085763931274414, "learning_rate": 9.865684345317158e-06, "loss": 0.32589726, "memory(GiB)": 14.42, "step": 1900, "train_speed(iter/s)": 1.008557 }, { "acc": 0.8939394, "epoch": 6.015789473684211, "grad_norm": 7.984169960021973, "learning_rate": 9.864477078017818e-06, "loss": 0.29888535, "memory(GiB)": 14.42, "step": 1905, "train_speed(iter/s)": 1.008474 }, { "acc": 0.88383846, "epoch": 6.031578947368421, "grad_norm": 8.458024978637695, "learning_rate": 9.863264483876056e-06, "loss": 0.28832068, "memory(GiB)": 14.42, "step": 1910, "train_speed(iter/s)": 1.008555 }, { "acc": 0.91111107, "epoch": 6.0473684210526315, "grad_norm": 8.8190279006958, "learning_rate": 9.862046564219863e-06, "loss": 0.26591012, "memory(GiB)": 14.42, "step": 1915, "train_speed(iter/s)": 1.008617 }, { "acc": 0.9060606, "epoch": 6.063157894736842, "grad_norm": 10.704242706298828, "learning_rate": 9.860823320383057e-06, "loss": 0.30131092, "memory(GiB)": 14.42, "step": 1920, "train_speed(iter/s)": 1.008617 }, { "acc": 0.88989906, "epoch": 6.078947368421052, "grad_norm": 7.7713541984558105, "learning_rate": 9.859594753705287e-06, "loss": 0.31342063, "memory(GiB)": 14.42, "step": 1925, "train_speed(iter/s)": 1.008657 }, { "acc": 0.8969697, "epoch": 6.094736842105263, "grad_norm": 9.331098556518555, "learning_rate": 9.858360865532031e-06, "loss": 0.30202923, "memory(GiB)": 14.42, "step": 1930, "train_speed(iter/s)": 1.008653 }, { "acc": 0.90101013, "epoch": 6.110526315789474, "grad_norm": 8.15951919555664, "learning_rate": 9.857121657214598e-06, "loss": 0.2731153, "memory(GiB)": 14.42, "step": 1935, "train_speed(iter/s)": 1.008746 }, { "acc": 0.9151516, "epoch": 6.126315789473685, "grad_norm": 8.893780708312988, "learning_rate": 9.855877130110123e-06, "loss": 0.23995247, "memory(GiB)": 14.42, "step": 1940, "train_speed(iter/s)": 1.008941 }, { "acc": 0.88686867, "epoch": 6.1421052631578945, "grad_norm": 5.452535629272461, "learning_rate": 9.854627285581564e-06, "loss": 0.28711867, "memory(GiB)": 14.42, "step": 1945, "train_speed(iter/s)": 1.009073 }, { "acc": 0.91111107, "epoch": 6.157894736842105, "grad_norm": 11.006546020507812, "learning_rate": 9.853372124997702e-06, "loss": 0.27313223, "memory(GiB)": 14.42, "step": 1950, "train_speed(iter/s)": 1.009118 }, { "acc": 0.88989906, "epoch": 6.173684210526316, "grad_norm": 7.908746242523193, "learning_rate": 9.852111649733143e-06, "loss": 0.32192535, "memory(GiB)": 14.42, "step": 1955, "train_speed(iter/s)": 1.009218 }, { "acc": 0.89898987, "epoch": 6.189473684210526, "grad_norm": 7.4186692237854, "learning_rate": 9.85084586116831e-06, "loss": 0.28101387, "memory(GiB)": 14.42, "step": 1960, "train_speed(iter/s)": 1.009292 }, { "acc": 0.9, "epoch": 6.205263157894737, "grad_norm": 8.2991943359375, "learning_rate": 9.849574760689445e-06, "loss": 0.30966353, "memory(GiB)": 14.42, "step": 1965, "train_speed(iter/s)": 1.009205 }, { "acc": 0.9181819, "epoch": 6.221052631578948, "grad_norm": 7.403661727905273, "learning_rate": 9.848298349688609e-06, "loss": 0.24571998, "memory(GiB)": 14.42, "step": 1970, "train_speed(iter/s)": 1.009147 }, { "acc": 0.91313133, "epoch": 6.2368421052631575, "grad_norm": 7.864448070526123, "learning_rate": 9.847016629563683e-06, "loss": 0.26097565, "memory(GiB)": 14.42, "step": 1975, "train_speed(iter/s)": 1.00915 }, { "acc": 0.9090909, "epoch": 6.252631578947368, "grad_norm": 10.98455810546875, "learning_rate": 9.845729601718354e-06, "loss": 0.2881371, "memory(GiB)": 14.42, "step": 1980, "train_speed(iter/s)": 1.009002 }, { "acc": 0.9121212, "epoch": 6.268421052631579, "grad_norm": 8.784940719604492, "learning_rate": 9.844437267562127e-06, "loss": 0.24309015, "memory(GiB)": 14.42, "step": 1985, "train_speed(iter/s)": 1.008873 }, { "acc": 0.91616163, "epoch": 6.284210526315789, "grad_norm": 7.331908226013184, "learning_rate": 9.84313962851032e-06, "loss": 0.24364107, "memory(GiB)": 14.42, "step": 1990, "train_speed(iter/s)": 1.008851 }, { "acc": 0.91010113, "epoch": 6.3, "grad_norm": 10.048667907714844, "learning_rate": 9.841836685984052e-06, "loss": 0.24012024, "memory(GiB)": 14.42, "step": 1995, "train_speed(iter/s)": 1.008919 }, { "acc": 0.90505047, "epoch": 6.315789473684211, "grad_norm": 11.079018592834473, "learning_rate": 9.840528441410261e-06, "loss": 0.28295364, "memory(GiB)": 14.42, "step": 2000, "train_speed(iter/s)": 1.009018 }, { "epoch": 6.315789473684211, "eval_acc": 0.5149603288404551, "eval_loss": 6.357347011566162, "eval_runtime": 49.4166, "eval_samples_per_second": 38.449, "eval_steps_per_second": 6.415, "step": 2000 } ], "logging_steps": 5, "max_steps": 15800, "num_input_tokens_seen": 0, "num_train_epochs": 50, "save_steps": 2000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 2.8379717835318886e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }