diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,5732 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 2.0, + "eval_steps": 500, + "global_step": 814, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.002457002457002457, + "grad_norm": 12.175590846877855, + "learning_rate": 0.0, + "loss": 1.2148, + "step": 1 + }, + { + "epoch": 0.004914004914004914, + "grad_norm": 13.112846331209713, + "learning_rate": 5e-08, + "loss": 1.3154, + "step": 2 + }, + { + "epoch": 0.007371007371007371, + "grad_norm": 12.768616794783732, + "learning_rate": 1e-07, + "loss": 1.2865, + "step": 3 + }, + { + "epoch": 0.009828009828009828, + "grad_norm": 12.570106090577944, + "learning_rate": 1.5e-07, + "loss": 1.2459, + "step": 4 + }, + { + "epoch": 0.012285012285012284, + "grad_norm": 12.026844749086528, + "learning_rate": 2e-07, + "loss": 1.2496, + "step": 5 + }, + { + "epoch": 0.014742014742014743, + "grad_norm": 12.166619597781125, + "learning_rate": 2.5e-07, + "loss": 1.2179, + "step": 6 + }, + { + "epoch": 0.0171990171990172, + "grad_norm": 12.450604071787023, + "learning_rate": 3e-07, + "loss": 1.2451, + "step": 7 + }, + { + "epoch": 0.019656019656019656, + "grad_norm": 11.75536378463392, + "learning_rate": 3.5e-07, + "loss": 1.2372, + "step": 8 + }, + { + "epoch": 0.022113022113022112, + "grad_norm": 11.7494451205682, + "learning_rate": 4e-07, + "loss": 1.2174, + "step": 9 + }, + { + "epoch": 0.02457002457002457, + "grad_norm": 11.485164591676915, + "learning_rate": 4.5e-07, + "loss": 1.1803, + "step": 10 + }, + { + "epoch": 0.02702702702702703, + "grad_norm": 11.41671410777608, + "learning_rate": 5e-07, + "loss": 1.1829, + "step": 11 + }, + { + "epoch": 0.029484029484029485, + "grad_norm": 10.737969644794076, + "learning_rate": 5.5e-07, + "loss": 1.1912, + "step": 12 + }, + { + "epoch": 0.03194103194103194, + "grad_norm": 12.0315769410197, + "learning_rate": 6e-07, + "loss": 1.2516, + "step": 13 + }, + { + "epoch": 0.0343980343980344, + "grad_norm": 11.703183966857443, + "learning_rate": 6.5e-07, + "loss": 1.2533, + "step": 14 + }, + { + "epoch": 0.036855036855036855, + "grad_norm": 10.75777994461547, + "learning_rate": 7e-07, + "loss": 1.1763, + "step": 15 + }, + { + "epoch": 0.03931203931203931, + "grad_norm": 9.517567846059583, + "learning_rate": 7.5e-07, + "loss": 1.1474, + "step": 16 + }, + { + "epoch": 0.04176904176904177, + "grad_norm": 9.825919996727293, + "learning_rate": 8e-07, + "loss": 1.1725, + "step": 17 + }, + { + "epoch": 0.044226044226044224, + "grad_norm": 9.370499752545927, + "learning_rate": 8.499999999999999e-07, + "loss": 1.2048, + "step": 18 + }, + { + "epoch": 0.04668304668304668, + "grad_norm": 8.582608952468464, + "learning_rate": 9e-07, + "loss": 1.1428, + "step": 19 + }, + { + "epoch": 0.04914004914004914, + "grad_norm": 7.4140901371154, + "learning_rate": 9.499999999999999e-07, + "loss": 1.1583, + "step": 20 + }, + { + "epoch": 0.051597051597051594, + "grad_norm": 6.890519345341131, + "learning_rate": 1e-06, + "loss": 1.1961, + "step": 21 + }, + { + "epoch": 0.05405405405405406, + "grad_norm": 5.47854198998431, + "learning_rate": 1.05e-06, + "loss": 1.126, + "step": 22 + }, + { + "epoch": 0.056511056511056514, + "grad_norm": 5.350914917259245, + "learning_rate": 1.1e-06, + "loss": 1.1484, + "step": 23 + }, + { + "epoch": 0.05896805896805897, + "grad_norm": 4.114918765551221, + "learning_rate": 1.1499999999999998e-06, + "loss": 1.1328, + "step": 24 + }, + { + "epoch": 0.06142506142506143, + "grad_norm": 3.535376327350945, + "learning_rate": 1.2e-06, + "loss": 1.0889, + "step": 25 + }, + { + "epoch": 0.06388206388206388, + "grad_norm": 3.2480152784573586, + "learning_rate": 1.2499999999999999e-06, + "loss": 1.0437, + "step": 26 + }, + { + "epoch": 0.06633906633906633, + "grad_norm": 3.4795119249436293, + "learning_rate": 1.3e-06, + "loss": 1.0966, + "step": 27 + }, + { + "epoch": 0.0687960687960688, + "grad_norm": 2.8566181371095047, + "learning_rate": 1.35e-06, + "loss": 1.1234, + "step": 28 + }, + { + "epoch": 0.07125307125307126, + "grad_norm": 3.6578709990785048, + "learning_rate": 1.4e-06, + "loss": 1.1265, + "step": 29 + }, + { + "epoch": 0.07371007371007371, + "grad_norm": 3.541143081282278, + "learning_rate": 1.4499999999999999e-06, + "loss": 1.0543, + "step": 30 + }, + { + "epoch": 0.07616707616707617, + "grad_norm": 6.531880288467122, + "learning_rate": 1.5e-06, + "loss": 1.1109, + "step": 31 + }, + { + "epoch": 0.07862407862407862, + "grad_norm": 6.4715106690759, + "learning_rate": 1.55e-06, + "loss": 1.1257, + "step": 32 + }, + { + "epoch": 0.08108108108108109, + "grad_norm": 5.399614723171929, + "learning_rate": 1.6e-06, + "loss": 1.0694, + "step": 33 + }, + { + "epoch": 0.08353808353808354, + "grad_norm": 5.447520160063898, + "learning_rate": 1.6499999999999999e-06, + "loss": 1.047, + "step": 34 + }, + { + "epoch": 0.085995085995086, + "grad_norm": 15.759399235218536, + "learning_rate": 1.6999999999999998e-06, + "loss": 1.0684, + "step": 35 + }, + { + "epoch": 0.08845208845208845, + "grad_norm": 6.138108825882396, + "learning_rate": 1.75e-06, + "loss": 1.0568, + "step": 36 + }, + { + "epoch": 0.09090909090909091, + "grad_norm": 28.96303499581736, + "learning_rate": 1.8e-06, + "loss": 1.0748, + "step": 37 + }, + { + "epoch": 0.09336609336609336, + "grad_norm": 3.400779377353331, + "learning_rate": 1.85e-06, + "loss": 1.0841, + "step": 38 + }, + { + "epoch": 0.09582309582309582, + "grad_norm": 3.2088443423156305, + "learning_rate": 1.8999999999999998e-06, + "loss": 1.0637, + "step": 39 + }, + { + "epoch": 0.09828009828009827, + "grad_norm": 3.650839591876822, + "learning_rate": 1.95e-06, + "loss": 1.065, + "step": 40 + }, + { + "epoch": 0.10073710073710074, + "grad_norm": 2.48874031529166, + "learning_rate": 2e-06, + "loss": 1.0325, + "step": 41 + }, + { + "epoch": 0.10319410319410319, + "grad_norm": 2.4764261500155436, + "learning_rate": 1.9999917626554467e-06, + "loss": 1.0484, + "step": 42 + }, + { + "epoch": 0.10565110565110565, + "grad_norm": 2.0092623603336794, + "learning_rate": 1.9999670507574944e-06, + "loss": 1.0859, + "step": 43 + }, + { + "epoch": 0.10810810810810811, + "grad_norm": 1.915125996872455, + "learning_rate": 1.9999258647132643e-06, + "loss": 1.0423, + "step": 44 + }, + { + "epoch": 0.11056511056511056, + "grad_norm": 1.8459238588876408, + "learning_rate": 1.9998682052012837e-06, + "loss": 1.0686, + "step": 45 + }, + { + "epoch": 0.11302211302211303, + "grad_norm": 2.238673168744122, + "learning_rate": 1.9997940731714745e-06, + "loss": 1.0601, + "step": 46 + }, + { + "epoch": 0.11547911547911548, + "grad_norm": 4.306515287574078, + "learning_rate": 1.9997034698451393e-06, + "loss": 1.0367, + "step": 47 + }, + { + "epoch": 0.11793611793611794, + "grad_norm": 5.689973830908464, + "learning_rate": 1.9995963967149398e-06, + "loss": 1.0542, + "step": 48 + }, + { + "epoch": 0.12039312039312039, + "grad_norm": 2.1880376147294522, + "learning_rate": 1.999472855544872e-06, + "loss": 1.0635, + "step": 49 + }, + { + "epoch": 0.12285012285012285, + "grad_norm": 3.433335591384566, + "learning_rate": 1.999332848370239e-06, + "loss": 1.0323, + "step": 50 + }, + { + "epoch": 0.12530712530712532, + "grad_norm": 1.8169785229927564, + "learning_rate": 1.9991763774976158e-06, + "loss": 1.0218, + "step": 51 + }, + { + "epoch": 0.12776412776412777, + "grad_norm": 1.8670733937364128, + "learning_rate": 1.9990034455048096e-06, + "loss": 1.0405, + "step": 52 + }, + { + "epoch": 0.13022113022113022, + "grad_norm": 2.439448369982022, + "learning_rate": 1.998814055240823e-06, + "loss": 1.0519, + "step": 53 + }, + { + "epoch": 0.13267813267813267, + "grad_norm": 1.5940089652939573, + "learning_rate": 1.9986082098258006e-06, + "loss": 1.0359, + "step": 54 + }, + { + "epoch": 0.13513513513513514, + "grad_norm": 1.7404888899145354, + "learning_rate": 1.9983859126509824e-06, + "loss": 1.0087, + "step": 55 + }, + { + "epoch": 0.1375921375921376, + "grad_norm": 2.2476917865688026, + "learning_rate": 1.998147167378645e-06, + "loss": 1.0138, + "step": 56 + }, + { + "epoch": 0.14004914004914004, + "grad_norm": 1.8002247115537675, + "learning_rate": 1.9978919779420425e-06, + "loss": 1.0488, + "step": 57 + }, + { + "epoch": 0.14250614250614252, + "grad_norm": 1.663251752279236, + "learning_rate": 1.9976203485453413e-06, + "loss": 0.9809, + "step": 58 + }, + { + "epoch": 0.14496314496314497, + "grad_norm": 1.7653797682084729, + "learning_rate": 1.9973322836635515e-06, + "loss": 1.0686, + "step": 59 + }, + { + "epoch": 0.14742014742014742, + "grad_norm": 1.5336222427197481, + "learning_rate": 1.9970277880424526e-06, + "loss": 0.986, + "step": 60 + }, + { + "epoch": 0.14987714987714987, + "grad_norm": 1.4421704567479543, + "learning_rate": 1.9967068666985146e-06, + "loss": 0.9948, + "step": 61 + }, + { + "epoch": 0.15233415233415235, + "grad_norm": 1.4042358001427657, + "learning_rate": 1.996369524918818e-06, + "loss": 0.967, + "step": 62 + }, + { + "epoch": 0.1547911547911548, + "grad_norm": 1.4295825578558987, + "learning_rate": 1.9960157682609633e-06, + "loss": 0.9723, + "step": 63 + }, + { + "epoch": 0.15724815724815724, + "grad_norm": 1.6093180146467703, + "learning_rate": 1.9956456025529804e-06, + "loss": 0.9737, + "step": 64 + }, + { + "epoch": 0.1597051597051597, + "grad_norm": 1.7273275051707897, + "learning_rate": 1.9952590338932358e-06, + "loss": 0.9935, + "step": 65 + }, + { + "epoch": 0.16216216216216217, + "grad_norm": 1.647950655242118, + "learning_rate": 1.994856068650327e-06, + "loss": 1.0509, + "step": 66 + }, + { + "epoch": 0.16461916461916462, + "grad_norm": 1.6052330571582885, + "learning_rate": 1.994436713462982e-06, + "loss": 1.0295, + "step": 67 + }, + { + "epoch": 0.16707616707616707, + "grad_norm": 1.694905424784479, + "learning_rate": 1.994000975239946e-06, + "loss": 1.0256, + "step": 68 + }, + { + "epoch": 0.16953316953316952, + "grad_norm": 1.5039835360361733, + "learning_rate": 1.9935488611598713e-06, + "loss": 1.0347, + "step": 69 + }, + { + "epoch": 0.171990171990172, + "grad_norm": 1.734126157202943, + "learning_rate": 1.9930803786711967e-06, + "loss": 1.023, + "step": 70 + }, + { + "epoch": 0.17444717444717445, + "grad_norm": 1.5242931763881968, + "learning_rate": 1.9925955354920264e-06, + "loss": 1.0406, + "step": 71 + }, + { + "epoch": 0.1769041769041769, + "grad_norm": 1.606370733528827, + "learning_rate": 1.99209433961e-06, + "loss": 1.0222, + "step": 72 + }, + { + "epoch": 0.17936117936117937, + "grad_norm": 2.4762050432839318, + "learning_rate": 1.991576799282164e-06, + "loss": 1.0278, + "step": 73 + }, + { + "epoch": 0.18181818181818182, + "grad_norm": 1.4747727870105878, + "learning_rate": 1.9910429230348343e-06, + "loss": 0.9779, + "step": 74 + }, + { + "epoch": 0.18427518427518427, + "grad_norm": 2.0977776864765167, + "learning_rate": 1.990492719663457e-06, + "loss": 1.0235, + "step": 75 + }, + { + "epoch": 0.18673218673218672, + "grad_norm": 1.9728896794221442, + "learning_rate": 1.9899261982324607e-06, + "loss": 0.9926, + "step": 76 + }, + { + "epoch": 0.1891891891891892, + "grad_norm": 1.7736978075188574, + "learning_rate": 1.98934336807511e-06, + "loss": 1.0354, + "step": 77 + }, + { + "epoch": 0.19164619164619165, + "grad_norm": 2.1716440393079464, + "learning_rate": 1.988744238793351e-06, + "loss": 0.977, + "step": 78 + }, + { + "epoch": 0.1941031941031941, + "grad_norm": 1.599345820298087, + "learning_rate": 1.9881288202576518e-06, + "loss": 0.9685, + "step": 79 + }, + { + "epoch": 0.19656019656019655, + "grad_norm": 1.4582630785882413, + "learning_rate": 1.9874971226068415e-06, + "loss": 0.943, + "step": 80 + }, + { + "epoch": 0.19901719901719903, + "grad_norm": 1.4275025088638074, + "learning_rate": 1.9868491562479424e-06, + "loss": 1.0295, + "step": 81 + }, + { + "epoch": 0.20147420147420148, + "grad_norm": 1.609247365977497, + "learning_rate": 1.9861849318559994e-06, + "loss": 0.9531, + "step": 82 + }, + { + "epoch": 0.20393120393120392, + "grad_norm": 1.711398906589491, + "learning_rate": 1.9855044603739028e-06, + "loss": 1.0302, + "step": 83 + }, + { + "epoch": 0.20638820638820637, + "grad_norm": 1.56099328882278, + "learning_rate": 1.984807753012208e-06, + "loss": 0.9684, + "step": 84 + }, + { + "epoch": 0.20884520884520885, + "grad_norm": 1.5428758027146767, + "learning_rate": 1.9840948212489527e-06, + "loss": 1.0129, + "step": 85 + }, + { + "epoch": 0.2113022113022113, + "grad_norm": 1.9669978126079752, + "learning_rate": 1.983365676829466e-06, + "loss": 0.9976, + "step": 86 + }, + { + "epoch": 0.21375921375921375, + "grad_norm": 1.5144216325008544, + "learning_rate": 1.9826203317661756e-06, + "loss": 0.9967, + "step": 87 + }, + { + "epoch": 0.21621621621621623, + "grad_norm": 2.1395022590018367, + "learning_rate": 1.9818587983384094e-06, + "loss": 0.9693, + "step": 88 + }, + { + "epoch": 0.21867321867321868, + "grad_norm": 1.5091268331439933, + "learning_rate": 1.981081089092194e-06, + "loss": 0.9271, + "step": 89 + }, + { + "epoch": 0.22113022113022113, + "grad_norm": 1.8173026339961775, + "learning_rate": 1.9802872168400476e-06, + "loss": 1.0819, + "step": 90 + }, + { + "epoch": 0.22358722358722358, + "grad_norm": 1.5986438761107704, + "learning_rate": 1.979477194660769e-06, + "loss": 0.9594, + "step": 91 + }, + { + "epoch": 0.22604422604422605, + "grad_norm": 1.5027188037219825, + "learning_rate": 1.978651035899221e-06, + "loss": 0.9964, + "step": 92 + }, + { + "epoch": 0.2285012285012285, + "grad_norm": 1.3856569445993616, + "learning_rate": 1.977808754166113e-06, + "loss": 0.9476, + "step": 93 + }, + { + "epoch": 0.23095823095823095, + "grad_norm": 1.445303089718021, + "learning_rate": 1.976950363337774e-06, + "loss": 0.9664, + "step": 94 + }, + { + "epoch": 0.2334152334152334, + "grad_norm": 1.9267867671090184, + "learning_rate": 1.9760758775559273e-06, + "loss": 1.0446, + "step": 95 + }, + { + "epoch": 0.23587223587223588, + "grad_norm": 1.4687411243380588, + "learning_rate": 1.975185311227453e-06, + "loss": 0.9483, + "step": 96 + }, + { + "epoch": 0.23832923832923833, + "grad_norm": 1.5489115105120785, + "learning_rate": 1.9742786790241546e-06, + "loss": 1.0073, + "step": 97 + }, + { + "epoch": 0.24078624078624078, + "grad_norm": 2.5463794252310676, + "learning_rate": 1.9733559958825166e-06, + "loss": 1.0046, + "step": 98 + }, + { + "epoch": 0.24324324324324326, + "grad_norm": 1.56125807056489, + "learning_rate": 1.9724172770034566e-06, + "loss": 0.971, + "step": 99 + }, + { + "epoch": 0.2457002457002457, + "grad_norm": 1.740701138818393, + "learning_rate": 1.971462537852076e-06, + "loss": 0.9856, + "step": 100 + }, + { + "epoch": 0.24815724815724816, + "grad_norm": 1.4470402917114091, + "learning_rate": 1.970491794157405e-06, + "loss": 0.9897, + "step": 101 + }, + { + "epoch": 0.25061425061425063, + "grad_norm": 1.5706400471340238, + "learning_rate": 1.9695050619121456e-06, + "loss": 0.9328, + "step": 102 + }, + { + "epoch": 0.25307125307125306, + "grad_norm": 1.9773708156391945, + "learning_rate": 1.9685023573724034e-06, + "loss": 0.9719, + "step": 103 + }, + { + "epoch": 0.25552825552825553, + "grad_norm": 1.9195746558098163, + "learning_rate": 1.9674836970574253e-06, + "loss": 0.978, + "step": 104 + }, + { + "epoch": 0.257985257985258, + "grad_norm": 1.8027098570962006, + "learning_rate": 1.966449097749322e-06, + "loss": 1.0023, + "step": 105 + }, + { + "epoch": 0.26044226044226043, + "grad_norm": 1.7805523167150648, + "learning_rate": 1.965398576492796e-06, + "loss": 0.9768, + "step": 106 + }, + { + "epoch": 0.2628992628992629, + "grad_norm": 1.9751836869745634, + "learning_rate": 1.9643321505948585e-06, + "loss": 1.0132, + "step": 107 + }, + { + "epoch": 0.26535626535626533, + "grad_norm": 1.6135714113454362, + "learning_rate": 1.9632498376245444e-06, + "loss": 1.0169, + "step": 108 + }, + { + "epoch": 0.2678132678132678, + "grad_norm": 1.6366575829852479, + "learning_rate": 1.9621516554126237e-06, + "loss": 0.9618, + "step": 109 + }, + { + "epoch": 0.2702702702702703, + "grad_norm": 1.4439243574845881, + "learning_rate": 1.9610376220513063e-06, + "loss": 1.0111, + "step": 110 + }, + { + "epoch": 0.2727272727272727, + "grad_norm": 1.574054550140938, + "learning_rate": 1.9599077558939466e-06, + "loss": 0.9979, + "step": 111 + }, + { + "epoch": 0.2751842751842752, + "grad_norm": 1.9768484700868765, + "learning_rate": 1.958762075554737e-06, + "loss": 1.0266, + "step": 112 + }, + { + "epoch": 0.27764127764127766, + "grad_norm": 1.6821288693476542, + "learning_rate": 1.957600599908406e-06, + "loss": 0.983, + "step": 113 + }, + { + "epoch": 0.2800982800982801, + "grad_norm": 1.6250908693533597, + "learning_rate": 1.956423348089903e-06, + "loss": 0.9735, + "step": 114 + }, + { + "epoch": 0.28255528255528256, + "grad_norm": 1.5158186810772534, + "learning_rate": 1.9552303394940857e-06, + "loss": 0.9767, + "step": 115 + }, + { + "epoch": 0.28501228501228504, + "grad_norm": 1.7591914283817052, + "learning_rate": 1.9540215937754004e-06, + "loss": 0.9515, + "step": 116 + }, + { + "epoch": 0.28746928746928746, + "grad_norm": 1.329980995859771, + "learning_rate": 1.952797130847557e-06, + "loss": 0.944, + "step": 117 + }, + { + "epoch": 0.28992628992628994, + "grad_norm": 1.7122579873222803, + "learning_rate": 1.951556970883201e-06, + "loss": 1.0098, + "step": 118 + }, + { + "epoch": 0.29238329238329236, + "grad_norm": 1.8427163819779844, + "learning_rate": 1.9503011343135825e-06, + "loss": 0.9815, + "step": 119 + }, + { + "epoch": 0.29484029484029484, + "grad_norm": 1.3722808372435795, + "learning_rate": 1.9490296418282184e-06, + "loss": 0.9914, + "step": 120 + }, + { + "epoch": 0.2972972972972973, + "grad_norm": 1.2858110734745247, + "learning_rate": 1.9477425143745523e-06, + "loss": 0.9727, + "step": 121 + }, + { + "epoch": 0.29975429975429974, + "grad_norm": 1.3478569354201173, + "learning_rate": 1.9464397731576094e-06, + "loss": 0.9557, + "step": 122 + }, + { + "epoch": 0.3022113022113022, + "grad_norm": 1.534212309472206, + "learning_rate": 1.945121439639645e-06, + "loss": 0.9995, + "step": 123 + }, + { + "epoch": 0.3046683046683047, + "grad_norm": 1.4583668620611645, + "learning_rate": 1.943787535539795e-06, + "loss": 0.9244, + "step": 124 + }, + { + "epoch": 0.3071253071253071, + "grad_norm": 1.3048849492711914, + "learning_rate": 1.9424380828337143e-06, + "loss": 0.9749, + "step": 125 + }, + { + "epoch": 0.3095823095823096, + "grad_norm": 1.9214890498265684, + "learning_rate": 1.9410731037532167e-06, + "loss": 1.0126, + "step": 126 + }, + { + "epoch": 0.31203931203931207, + "grad_norm": 1.492998355230107, + "learning_rate": 1.9396926207859082e-06, + "loss": 0.9216, + "step": 127 + }, + { + "epoch": 0.3144963144963145, + "grad_norm": 1.529088663498555, + "learning_rate": 1.9382966566748166e-06, + "loss": 1.0266, + "step": 128 + }, + { + "epoch": 0.31695331695331697, + "grad_norm": 1.4961622214991601, + "learning_rate": 1.9368852344180167e-06, + "loss": 0.949, + "step": 129 + }, + { + "epoch": 0.3194103194103194, + "grad_norm": 1.5657840764329438, + "learning_rate": 1.935458377268251e-06, + "loss": 0.9768, + "step": 130 + }, + { + "epoch": 0.32186732186732187, + "grad_norm": 1.458088079896784, + "learning_rate": 1.934016108732548e-06, + "loss": 0.9831, + "step": 131 + }, + { + "epoch": 0.32432432432432434, + "grad_norm": 1.5704249898970843, + "learning_rate": 1.932558452571833e-06, + "loss": 0.9845, + "step": 132 + }, + { + "epoch": 0.32678132678132676, + "grad_norm": 1.366875012537187, + "learning_rate": 1.9310854328005377e-06, + "loss": 0.9751, + "step": 133 + }, + { + "epoch": 0.32923832923832924, + "grad_norm": 1.6682519282454646, + "learning_rate": 1.9295970736862064e-06, + "loss": 0.9278, + "step": 134 + }, + { + "epoch": 0.3316953316953317, + "grad_norm": 1.7010058864881312, + "learning_rate": 1.928093399749091e-06, + "loss": 0.9479, + "step": 135 + }, + { + "epoch": 0.33415233415233414, + "grad_norm": 2.774339563092048, + "learning_rate": 1.926574435761753e-06, + "loss": 0.9216, + "step": 136 + }, + { + "epoch": 0.3366093366093366, + "grad_norm": 1.5996469164653206, + "learning_rate": 1.925040206748652e-06, + "loss": 0.9749, + "step": 137 + }, + { + "epoch": 0.33906633906633904, + "grad_norm": 1.6140642370600065, + "learning_rate": 1.9234907379857334e-06, + "loss": 1.0195, + "step": 138 + }, + { + "epoch": 0.3415233415233415, + "grad_norm": 1.6724211149688903, + "learning_rate": 1.9219260550000143e-06, + "loss": 1.0086, + "step": 139 + }, + { + "epoch": 0.343980343980344, + "grad_norm": 1.6697054307200352, + "learning_rate": 1.920346183569159e-06, + "loss": 1.0177, + "step": 140 + }, + { + "epoch": 0.3464373464373464, + "grad_norm": 1.5507362171936665, + "learning_rate": 1.91875114972106e-06, + "loss": 0.9701, + "step": 141 + }, + { + "epoch": 0.3488943488943489, + "grad_norm": 1.572772556059398, + "learning_rate": 1.9171409797334026e-06, + "loss": 0.9893, + "step": 142 + }, + { + "epoch": 0.35135135135135137, + "grad_norm": 1.4851773686144312, + "learning_rate": 1.9155157001332373e-06, + "loss": 0.9617, + "step": 143 + }, + { + "epoch": 0.3538083538083538, + "grad_norm": 1.3935370189220326, + "learning_rate": 1.91387533769654e-06, + "loss": 0.9647, + "step": 144 + }, + { + "epoch": 0.35626535626535627, + "grad_norm": 2.086032038197472, + "learning_rate": 1.912219919447772e-06, + "loss": 1.0138, + "step": 145 + }, + { + "epoch": 0.35872235872235875, + "grad_norm": 1.8988521674993861, + "learning_rate": 1.9105494726594343e-06, + "loss": 0.9619, + "step": 146 + }, + { + "epoch": 0.36117936117936117, + "grad_norm": 1.7707337422933225, + "learning_rate": 1.9088640248516184e-06, + "loss": 0.9689, + "step": 147 + }, + { + "epoch": 0.36363636363636365, + "grad_norm": 10.58019032383696, + "learning_rate": 1.9071636037915533e-06, + "loss": 0.9886, + "step": 148 + }, + { + "epoch": 0.36609336609336607, + "grad_norm": 1.6326349036691947, + "learning_rate": 1.9054482374931466e-06, + "loss": 0.9928, + "step": 149 + }, + { + "epoch": 0.36855036855036855, + "grad_norm": 1.4602979723959129, + "learning_rate": 1.9037179542165251e-06, + "loss": 0.9451, + "step": 150 + }, + { + "epoch": 0.371007371007371, + "grad_norm": 1.5983113709927816, + "learning_rate": 1.9019727824675682e-06, + "loss": 0.9488, + "step": 151 + }, + { + "epoch": 0.37346437346437344, + "grad_norm": 1.394284340663908, + "learning_rate": 1.9002127509974373e-06, + "loss": 0.9717, + "step": 152 + }, + { + "epoch": 0.3759213759213759, + "grad_norm": 1.856689232868256, + "learning_rate": 1.8984378888021041e-06, + "loss": 0.9286, + "step": 153 + }, + { + "epoch": 0.3783783783783784, + "grad_norm": 1.5404154844189102, + "learning_rate": 1.8966482251218714e-06, + "loss": 0.9625, + "step": 154 + }, + { + "epoch": 0.3808353808353808, + "grad_norm": 1.5417688714414741, + "learning_rate": 1.8948437894408916e-06, + "loss": 0.9314, + "step": 155 + }, + { + "epoch": 0.3832923832923833, + "grad_norm": 1.8457706124864284, + "learning_rate": 1.8930246114866822e-06, + "loss": 1.0147, + "step": 156 + }, + { + "epoch": 0.3857493857493858, + "grad_norm": 1.4971057129868792, + "learning_rate": 1.8911907212296341e-06, + "loss": 0.9309, + "step": 157 + }, + { + "epoch": 0.3882063882063882, + "grad_norm": 1.6347670467700182, + "learning_rate": 1.8893421488825188e-06, + "loss": 0.9621, + "step": 158 + }, + { + "epoch": 0.3906633906633907, + "grad_norm": 1.6324227560172353, + "learning_rate": 1.8874789248999915e-06, + "loss": 0.9772, + "step": 159 + }, + { + "epoch": 0.3931203931203931, + "grad_norm": 1.6267750782874413, + "learning_rate": 1.885601079978088e-06, + "loss": 0.9861, + "step": 160 + }, + { + "epoch": 0.3955773955773956, + "grad_norm": 1.4406136327704948, + "learning_rate": 1.8837086450537192e-06, + "loss": 0.9406, + "step": 161 + }, + { + "epoch": 0.39803439803439805, + "grad_norm": 1.5622857080744104, + "learning_rate": 1.8818016513041624e-06, + "loss": 0.94, + "step": 162 + }, + { + "epoch": 0.4004914004914005, + "grad_norm": 1.7312279395333572, + "learning_rate": 1.8798801301465468e-06, + "loss": 0.96, + "step": 163 + }, + { + "epoch": 0.40294840294840295, + "grad_norm": 1.5736688729195003, + "learning_rate": 1.877944113237336e-06, + "loss": 0.9793, + "step": 164 + }, + { + "epoch": 0.40540540540540543, + "grad_norm": 1.501899237192728, + "learning_rate": 1.8759936324718066e-06, + "loss": 0.9178, + "step": 165 + }, + { + "epoch": 0.40786240786240785, + "grad_norm": 1.3483386139829852, + "learning_rate": 1.874028719983523e-06, + "loss": 0.9593, + "step": 166 + }, + { + "epoch": 0.4103194103194103, + "grad_norm": 1.5326296765065217, + "learning_rate": 1.8720494081438077e-06, + "loss": 0.9794, + "step": 167 + }, + { + "epoch": 0.41277641277641275, + "grad_norm": 1.5759774818522712, + "learning_rate": 1.870055729561207e-06, + "loss": 0.9418, + "step": 168 + }, + { + "epoch": 0.4152334152334152, + "grad_norm": 2.475109331033288, + "learning_rate": 1.8680477170809572e-06, + "loss": 0.9933, + "step": 169 + }, + { + "epoch": 0.4176904176904177, + "grad_norm": 1.4136668445270784, + "learning_rate": 1.8660254037844386e-06, + "loss": 0.9287, + "step": 170 + }, + { + "epoch": 0.4201474201474201, + "grad_norm": 1.7485785437438173, + "learning_rate": 1.863988822988634e-06, + "loss": 0.9749, + "step": 171 + }, + { + "epoch": 0.4226044226044226, + "grad_norm": 1.825009421972696, + "learning_rate": 1.8619380082455796e-06, + "loss": 0.9637, + "step": 172 + }, + { + "epoch": 0.4250614250614251, + "grad_norm": 1.6686645342967068, + "learning_rate": 1.85987299334181e-06, + "loss": 0.9951, + "step": 173 + }, + { + "epoch": 0.4275184275184275, + "grad_norm": 1.9731456114498873, + "learning_rate": 1.857793812297804e-06, + "loss": 0.9426, + "step": 174 + }, + { + "epoch": 0.42997542997543, + "grad_norm": 1.5558317172807916, + "learning_rate": 1.855700499367423e-06, + "loss": 0.9787, + "step": 175 + }, + { + "epoch": 0.43243243243243246, + "grad_norm": 2.098470725111635, + "learning_rate": 1.8535930890373463e-06, + "loss": 0.9936, + "step": 176 + }, + { + "epoch": 0.4348894348894349, + "grad_norm": 1.525785332653322, + "learning_rate": 1.8514716160265044e-06, + "loss": 0.9727, + "step": 177 + }, + { + "epoch": 0.43734643734643736, + "grad_norm": 1.4913555879508107, + "learning_rate": 1.8493361152855056e-06, + "loss": 0.9748, + "step": 178 + }, + { + "epoch": 0.4398034398034398, + "grad_norm": 1.340195087338962, + "learning_rate": 1.84718662199606e-06, + "loss": 0.9706, + "step": 179 + }, + { + "epoch": 0.44226044226044225, + "grad_norm": 2.5585128719755867, + "learning_rate": 1.8450231715704026e-06, + "loss": 0.9467, + "step": 180 + }, + { + "epoch": 0.44471744471744473, + "grad_norm": 1.3066387641923265, + "learning_rate": 1.8428457996507053e-06, + "loss": 0.9194, + "step": 181 + }, + { + "epoch": 0.44717444717444715, + "grad_norm": 1.6465015301936612, + "learning_rate": 1.8406545421084938e-06, + "loss": 1.0045, + "step": 182 + }, + { + "epoch": 0.44963144963144963, + "grad_norm": 1.4122218928192685, + "learning_rate": 1.8384494350440552e-06, + "loss": 0.9596, + "step": 183 + }, + { + "epoch": 0.4520884520884521, + "grad_norm": 1.4364915893717223, + "learning_rate": 1.8362305147858427e-06, + "loss": 0.9375, + "step": 184 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 1.5320964201665865, + "learning_rate": 1.833997817889878e-06, + "loss": 0.98, + "step": 185 + }, + { + "epoch": 0.457002457002457, + "grad_norm": 1.8963928416673594, + "learning_rate": 1.8317513811391477e-06, + "loss": 0.9499, + "step": 186 + }, + { + "epoch": 0.4594594594594595, + "grad_norm": 1.4450513723222074, + "learning_rate": 1.8294912415429993e-06, + "loss": 0.9387, + "step": 187 + }, + { + "epoch": 0.4619164619164619, + "grad_norm": 1.4227834836237356, + "learning_rate": 1.8272174363365298e-06, + "loss": 0.9856, + "step": 188 + }, + { + "epoch": 0.4643734643734644, + "grad_norm": 1.4225544869881186, + "learning_rate": 1.8249300029799732e-06, + "loss": 0.9201, + "step": 189 + }, + { + "epoch": 0.4668304668304668, + "grad_norm": 1.6580547740077536, + "learning_rate": 1.822628979158083e-06, + "loss": 0.9746, + "step": 190 + }, + { + "epoch": 0.4692874692874693, + "grad_norm": 1.6225782675239633, + "learning_rate": 1.8203144027795108e-06, + "loss": 0.9493, + "step": 191 + }, + { + "epoch": 0.47174447174447176, + "grad_norm": 1.8558149480612793, + "learning_rate": 1.8179863119761834e-06, + "loss": 0.9551, + "step": 192 + }, + { + "epoch": 0.4742014742014742, + "grad_norm": 1.5948966144212158, + "learning_rate": 1.815644745102673e-06, + "loss": 0.9799, + "step": 193 + }, + { + "epoch": 0.47665847665847666, + "grad_norm": 1.464419249562776, + "learning_rate": 1.8132897407355654e-06, + "loss": 0.9634, + "step": 194 + }, + { + "epoch": 0.47911547911547914, + "grad_norm": 1.6709063152230734, + "learning_rate": 1.8109213376728257e-06, + "loss": 0.977, + "step": 195 + }, + { + "epoch": 0.48157248157248156, + "grad_norm": 1.4980373204140025, + "learning_rate": 1.8085395749331579e-06, + "loss": 0.9839, + "step": 196 + }, + { + "epoch": 0.48402948402948404, + "grad_norm": 1.4728568491372855, + "learning_rate": 1.8061444917553627e-06, + "loss": 0.9791, + "step": 197 + }, + { + "epoch": 0.4864864864864865, + "grad_norm": 1.3529262817353658, + "learning_rate": 1.803736127597691e-06, + "loss": 0.9534, + "step": 198 + }, + { + "epoch": 0.48894348894348894, + "grad_norm": 1.4726330094781561, + "learning_rate": 1.8013145221371935e-06, + "loss": 0.9397, + "step": 199 + }, + { + "epoch": 0.4914004914004914, + "grad_norm": 1.4468557644109676, + "learning_rate": 1.7988797152690668e-06, + "loss": 0.8993, + "step": 200 + }, + { + "epoch": 0.49385749385749383, + "grad_norm": 1.39576948942371, + "learning_rate": 1.796431747105998e-06, + "loss": 0.945, + "step": 201 + }, + { + "epoch": 0.4963144963144963, + "grad_norm": 1.4248364851440443, + "learning_rate": 1.793970657977501e-06, + "loss": 0.9669, + "step": 202 + }, + { + "epoch": 0.4987714987714988, + "grad_norm": 1.4850571599264226, + "learning_rate": 1.791496488429254e-06, + "loss": 0.9572, + "step": 203 + }, + { + "epoch": 0.5012285012285013, + "grad_norm": 1.4619304758472125, + "learning_rate": 1.7890092792224313e-06, + "loss": 0.9604, + "step": 204 + }, + { + "epoch": 0.5036855036855037, + "grad_norm": 1.6208674632812228, + "learning_rate": 1.786509071333031e-06, + "loss": 0.9577, + "step": 205 + }, + { + "epoch": 0.5061425061425061, + "grad_norm": 1.4904987175582993, + "learning_rate": 1.7839959059512013e-06, + "loss": 0.9039, + "step": 206 + }, + { + "epoch": 0.5085995085995086, + "grad_norm": 1.4636256072722436, + "learning_rate": 1.7814698244805601e-06, + "loss": 0.9481, + "step": 207 + }, + { + "epoch": 0.5110565110565111, + "grad_norm": 1.3573990171789403, + "learning_rate": 1.7789308685375146e-06, + "loss": 0.9531, + "step": 208 + }, + { + "epoch": 0.5135135135135135, + "grad_norm": 1.4176202461890393, + "learning_rate": 1.7763790799505745e-06, + "loss": 0.959, + "step": 209 + }, + { + "epoch": 0.515970515970516, + "grad_norm": 1.5707837841315286, + "learning_rate": 1.773814500759663e-06, + "loss": 0.9798, + "step": 210 + }, + { + "epoch": 0.5184275184275184, + "grad_norm": 1.4706701895149787, + "learning_rate": 1.7712371732154256e-06, + "loss": 0.9928, + "step": 211 + }, + { + "epoch": 0.5208845208845209, + "grad_norm": 2.069991518090132, + "learning_rate": 1.7686471397785318e-06, + "loss": 1.0153, + "step": 212 + }, + { + "epoch": 0.5233415233415234, + "grad_norm": 1.443867260475285, + "learning_rate": 1.766044443118978e-06, + "loss": 0.9452, + "step": 213 + }, + { + "epoch": 0.5257985257985258, + "grad_norm": 1.5151736278472239, + "learning_rate": 1.7634291261153817e-06, + "loss": 0.9114, + "step": 214 + }, + { + "epoch": 0.5282555282555282, + "grad_norm": 1.3847804036493465, + "learning_rate": 1.7608012318542777e-06, + "loss": 0.9144, + "step": 215 + }, + { + "epoch": 0.5307125307125307, + "grad_norm": 1.415085937475802, + "learning_rate": 1.7581608036294075e-06, + "loss": 0.9726, + "step": 216 + }, + { + "epoch": 0.5331695331695332, + "grad_norm": 1.558702474871841, + "learning_rate": 1.7555078849410042e-06, + "loss": 0.9243, + "step": 217 + }, + { + "epoch": 0.5356265356265356, + "grad_norm": 1.826623068304597, + "learning_rate": 1.7528425194950793e-06, + "loss": 1.0106, + "step": 218 + }, + { + "epoch": 0.538083538083538, + "grad_norm": 1.5238015749509666, + "learning_rate": 1.7501647512026994e-06, + "loss": 1.0093, + "step": 219 + }, + { + "epoch": 0.5405405405405406, + "grad_norm": 1.6514187408836531, + "learning_rate": 1.7474746241792646e-06, + "loss": 0.991, + "step": 220 + }, + { + "epoch": 0.542997542997543, + "grad_norm": 1.5401870547381438, + "learning_rate": 1.7447721827437819e-06, + "loss": 0.8904, + "step": 221 + }, + { + "epoch": 0.5454545454545454, + "grad_norm": 1.58085107300811, + "learning_rate": 1.742057471418133e-06, + "loss": 0.907, + "step": 222 + }, + { + "epoch": 0.547911547911548, + "grad_norm": 1.4376088351554372, + "learning_rate": 1.7393305349263432e-06, + "loss": 0.9369, + "step": 223 + }, + { + "epoch": 0.5503685503685504, + "grad_norm": 1.3807320173451696, + "learning_rate": 1.7365914181938437e-06, + "loss": 0.8945, + "step": 224 + }, + { + "epoch": 0.5528255528255528, + "grad_norm": 1.5104969018841128, + "learning_rate": 1.7338401663467309e-06, + "loss": 0.9537, + "step": 225 + }, + { + "epoch": 0.5552825552825553, + "grad_norm": 1.4036248703567802, + "learning_rate": 1.7310768247110232e-06, + "loss": 0.9112, + "step": 226 + }, + { + "epoch": 0.5577395577395577, + "grad_norm": 1.4153113695817048, + "learning_rate": 1.7283014388119158e-06, + "loss": 0.9285, + "step": 227 + }, + { + "epoch": 0.5601965601965602, + "grad_norm": 1.6827361424168579, + "learning_rate": 1.7255140543730282e-06, + "loss": 0.9368, + "step": 228 + }, + { + "epoch": 0.5626535626535627, + "grad_norm": 1.52442672792253, + "learning_rate": 1.7227147173156523e-06, + "loss": 0.9838, + "step": 229 + }, + { + "epoch": 0.5651105651105651, + "grad_norm": 1.8631309960035536, + "learning_rate": 1.7199034737579958e-06, + "loss": 0.9311, + "step": 230 + }, + { + "epoch": 0.5675675675675675, + "grad_norm": 1.38816407656762, + "learning_rate": 1.7170803700144224e-06, + "loss": 0.945, + "step": 231 + }, + { + "epoch": 0.5700245700245701, + "grad_norm": 1.5370320107384547, + "learning_rate": 1.7142454525946889e-06, + "loss": 0.955, + "step": 232 + }, + { + "epoch": 0.5724815724815725, + "grad_norm": 1.5632251869407223, + "learning_rate": 1.7113987682031778e-06, + "loss": 0.9384, + "step": 233 + }, + { + "epoch": 0.5749385749385749, + "grad_norm": 1.602474983632772, + "learning_rate": 1.70854036373813e-06, + "loss": 0.992, + "step": 234 + }, + { + "epoch": 0.5773955773955773, + "grad_norm": 1.3375523516298262, + "learning_rate": 1.70567028629087e-06, + "loss": 0.941, + "step": 235 + }, + { + "epoch": 0.5798525798525799, + "grad_norm": 1.539702405955399, + "learning_rate": 1.7027885831450317e-06, + "loss": 0.9516, + "step": 236 + }, + { + "epoch": 0.5823095823095823, + "grad_norm": 1.6116292160800432, + "learning_rate": 1.6998953017757784e-06, + "loss": 0.992, + "step": 237 + }, + { + "epoch": 0.5847665847665847, + "grad_norm": 2.3507568939715835, + "learning_rate": 1.6969904898490213e-06, + "loss": 0.9165, + "step": 238 + }, + { + "epoch": 0.5872235872235873, + "grad_norm": 2.847402959001731, + "learning_rate": 1.694074195220634e-06, + "loss": 0.9985, + "step": 239 + }, + { + "epoch": 0.5896805896805897, + "grad_norm": 1.5211685594705786, + "learning_rate": 1.6911464659356632e-06, + "loss": 0.9365, + "step": 240 + }, + { + "epoch": 0.5921375921375921, + "grad_norm": 2.034636443003393, + "learning_rate": 1.688207350227539e-06, + "loss": 0.9094, + "step": 241 + }, + { + "epoch": 0.5945945945945946, + "grad_norm": 1.8768465174864484, + "learning_rate": 1.6852568965172791e-06, + "loss": 0.9898, + "step": 242 + }, + { + "epoch": 0.597051597051597, + "grad_norm": 1.7384412195063934, + "learning_rate": 1.6822951534126908e-06, + "loss": 0.9955, + "step": 243 + }, + { + "epoch": 0.5995085995085995, + "grad_norm": 1.4908111245385776, + "learning_rate": 1.6793221697075715e-06, + "loss": 0.9503, + "step": 244 + }, + { + "epoch": 0.601965601965602, + "grad_norm": 1.3104233864174921, + "learning_rate": 1.6763379943809026e-06, + "loss": 0.9087, + "step": 245 + }, + { + "epoch": 0.6044226044226044, + "grad_norm": 1.8378209471753764, + "learning_rate": 1.6733426765960456e-06, + "loss": 0.9295, + "step": 246 + }, + { + "epoch": 0.6068796068796068, + "grad_norm": 1.5513249549838153, + "learning_rate": 1.6703362656999299e-06, + "loss": 0.9622, + "step": 247 + }, + { + "epoch": 0.6093366093366094, + "grad_norm": 1.950619315682982, + "learning_rate": 1.6673188112222395e-06, + "loss": 0.9114, + "step": 248 + }, + { + "epoch": 0.6117936117936118, + "grad_norm": 1.5443899853661804, + "learning_rate": 1.6642903628745993e-06, + "loss": 0.9385, + "step": 249 + }, + { + "epoch": 0.6142506142506142, + "grad_norm": 1.842691343791668, + "learning_rate": 1.6612509705497541e-06, + "loss": 0.9681, + "step": 250 + }, + { + "epoch": 0.6167076167076168, + "grad_norm": 1.4579884817325401, + "learning_rate": 1.6582006843207479e-06, + "loss": 0.9967, + "step": 251 + }, + { + "epoch": 0.6191646191646192, + "grad_norm": 1.602595851064425, + "learning_rate": 1.655139554440098e-06, + "loss": 0.9674, + "step": 252 + }, + { + "epoch": 0.6216216216216216, + "grad_norm": 1.394530716544746, + "learning_rate": 1.652067631338967e-06, + "loss": 0.907, + "step": 253 + }, + { + "epoch": 0.6240786240786241, + "grad_norm": 1.6411505351669575, + "learning_rate": 1.6489849656263334e-06, + "loss": 0.9602, + "step": 254 + }, + { + "epoch": 0.6265356265356266, + "grad_norm": 1.7684281051463475, + "learning_rate": 1.6458916080881565e-06, + "loss": 1.0141, + "step": 255 + }, + { + "epoch": 0.628992628992629, + "grad_norm": 1.771327098736445, + "learning_rate": 1.6427876096865393e-06, + "loss": 0.9917, + "step": 256 + }, + { + "epoch": 0.6314496314496314, + "grad_norm": 1.440535084208374, + "learning_rate": 1.6396730215588912e-06, + "loss": 0.9326, + "step": 257 + }, + { + "epoch": 0.6339066339066339, + "grad_norm": 1.5197592107480333, + "learning_rate": 1.6365478950170833e-06, + "loss": 0.9475, + "step": 258 + }, + { + "epoch": 0.6363636363636364, + "grad_norm": 1.706431633980188, + "learning_rate": 1.6334122815466031e-06, + "loss": 0.9351, + "step": 259 + }, + { + "epoch": 0.6388206388206388, + "grad_norm": 1.323673519490258, + "learning_rate": 1.6302662328057085e-06, + "loss": 0.911, + "step": 260 + }, + { + "epoch": 0.6412776412776413, + "grad_norm": 1.5394197428392233, + "learning_rate": 1.627109800624574e-06, + "loss": 0.9812, + "step": 261 + }, + { + "epoch": 0.6437346437346437, + "grad_norm": 1.4717326057521631, + "learning_rate": 1.6239430370044386e-06, + "loss": 0.8713, + "step": 262 + }, + { + "epoch": 0.6461916461916462, + "grad_norm": 1.5141060178146766, + "learning_rate": 1.6207659941167484e-06, + "loss": 0.934, + "step": 263 + }, + { + "epoch": 0.6486486486486487, + "grad_norm": 1.606823271595554, + "learning_rate": 1.617578724302297e-06, + "loss": 0.9689, + "step": 264 + }, + { + "epoch": 0.6511056511056511, + "grad_norm": 1.436173862475149, + "learning_rate": 1.6143812800703643e-06, + "loss": 0.9524, + "step": 265 + }, + { + "epoch": 0.6535626535626535, + "grad_norm": 1.6794632700616086, + "learning_rate": 1.6111737140978491e-06, + "loss": 0.9377, + "step": 266 + }, + { + "epoch": 0.6560196560196561, + "grad_norm": 1.481170005860829, + "learning_rate": 1.6079560792284045e-06, + "loss": 0.9718, + "step": 267 + }, + { + "epoch": 0.6584766584766585, + "grad_norm": 1.5488863504284036, + "learning_rate": 1.604728428471564e-06, + "loss": 0.9597, + "step": 268 + }, + { + "epoch": 0.6609336609336609, + "grad_norm": 1.5751377061062102, + "learning_rate": 1.6014908150018703e-06, + "loss": 0.9741, + "step": 269 + }, + { + "epoch": 0.6633906633906634, + "grad_norm": 1.8133205336540266, + "learning_rate": 1.598243292157999e-06, + "loss": 0.9234, + "step": 270 + }, + { + "epoch": 0.6658476658476659, + "grad_norm": 2.014347908541205, + "learning_rate": 1.5949859134418797e-06, + "loss": 0.976, + "step": 271 + }, + { + "epoch": 0.6683046683046683, + "grad_norm": 1.3995748329733648, + "learning_rate": 1.5917187325178136e-06, + "loss": 0.9182, + "step": 272 + }, + { + "epoch": 0.6707616707616708, + "grad_norm": 1.536609373517338, + "learning_rate": 1.5884418032115906e-06, + "loss": 0.8921, + "step": 273 + }, + { + "epoch": 0.6732186732186732, + "grad_norm": 1.7529235851864882, + "learning_rate": 1.5851551795096026e-06, + "loss": 0.9083, + "step": 274 + }, + { + "epoch": 0.6756756756756757, + "grad_norm": 1.5263377609284405, + "learning_rate": 1.5818589155579529e-06, + "loss": 0.921, + "step": 275 + }, + { + "epoch": 0.6781326781326781, + "grad_norm": 1.841830762323041, + "learning_rate": 1.5785530656615652e-06, + "loss": 0.943, + "step": 276 + }, + { + "epoch": 0.6805896805896806, + "grad_norm": 1.5135421099319812, + "learning_rate": 1.5752376842832898e-06, + "loss": 0.9315, + "step": 277 + }, + { + "epoch": 0.683046683046683, + "grad_norm": 1.492488292559091, + "learning_rate": 1.571912826043003e-06, + "loss": 0.891, + "step": 278 + }, + { + "epoch": 0.6855036855036855, + "grad_norm": 1.9836354455315945, + "learning_rate": 1.5685785457167113e-06, + "loss": 0.8706, + "step": 279 + }, + { + "epoch": 0.687960687960688, + "grad_norm": 1.5134741374320924, + "learning_rate": 1.565234898235646e-06, + "loss": 0.9786, + "step": 280 + }, + { + "epoch": 0.6904176904176904, + "grad_norm": 1.4859838259301754, + "learning_rate": 1.5618819386853603e-06, + "loss": 0.9595, + "step": 281 + }, + { + "epoch": 0.6928746928746928, + "grad_norm": 1.567447965101693, + "learning_rate": 1.5585197223048202e-06, + "loss": 0.9373, + "step": 282 + }, + { + "epoch": 0.6953316953316954, + "grad_norm": 1.361598883424366, + "learning_rate": 1.5551483044854952e-06, + "loss": 0.9697, + "step": 283 + }, + { + "epoch": 0.6977886977886978, + "grad_norm": 1.6640657059750952, + "learning_rate": 1.551767740770446e-06, + "loss": 0.9244, + "step": 284 + }, + { + "epoch": 0.7002457002457002, + "grad_norm": 1.362536522369026, + "learning_rate": 1.5483780868534083e-06, + "loss": 0.981, + "step": 285 + }, + { + "epoch": 0.7027027027027027, + "grad_norm": 1.6010684699844948, + "learning_rate": 1.544979398577877e-06, + "loss": 1.0169, + "step": 286 + }, + { + "epoch": 0.7051597051597052, + "grad_norm": 1.6122664157541595, + "learning_rate": 1.5415717319361845e-06, + "loss": 0.9507, + "step": 287 + }, + { + "epoch": 0.7076167076167076, + "grad_norm": 1.5044459630841276, + "learning_rate": 1.5381551430685796e-06, + "loss": 0.9431, + "step": 288 + }, + { + "epoch": 0.7100737100737101, + "grad_norm": 1.464490997097453, + "learning_rate": 1.5347296882623016e-06, + "loss": 0.9726, + "step": 289 + }, + { + "epoch": 0.7125307125307125, + "grad_norm": 1.8674476124140849, + "learning_rate": 1.5312954239506533e-06, + "loss": 0.9406, + "step": 290 + }, + { + "epoch": 0.714987714987715, + "grad_norm": 1.3556749625191538, + "learning_rate": 1.5278524067120718e-06, + "loss": 0.8896, + "step": 291 + }, + { + "epoch": 0.7174447174447175, + "grad_norm": 1.4118443972410692, + "learning_rate": 1.5244006932691954e-06, + "loss": 0.9722, + "step": 292 + }, + { + "epoch": 0.7199017199017199, + "grad_norm": 1.6225004312255737, + "learning_rate": 1.5209403404879303e-06, + "loss": 0.929, + "step": 293 + }, + { + "epoch": 0.7223587223587223, + "grad_norm": 1.6022157008964288, + "learning_rate": 1.5174714053765123e-06, + "loss": 0.9497, + "step": 294 + }, + { + "epoch": 0.7248157248157249, + "grad_norm": 1.5891302329335815, + "learning_rate": 1.5139939450845699e-06, + "loss": 0.973, + "step": 295 + }, + { + "epoch": 0.7272727272727273, + "grad_norm": 1.6988844584422054, + "learning_rate": 1.5105080169021789e-06, + "loss": 0.9184, + "step": 296 + }, + { + "epoch": 0.7297297297297297, + "grad_norm": 1.453985001696748, + "learning_rate": 1.5070136782589234e-06, + "loss": 0.9477, + "step": 297 + }, + { + "epoch": 0.7321867321867321, + "grad_norm": 1.604839180071105, + "learning_rate": 1.5035109867229457e-06, + "loss": 0.9409, + "step": 298 + }, + { + "epoch": 0.7346437346437347, + "grad_norm": 1.5773771725228862, + "learning_rate": 1.5e-06, + "loss": 0.937, + "step": 299 + }, + { + "epoch": 0.7371007371007371, + "grad_norm": 1.4272988880705397, + "learning_rate": 1.496480775932501e-06, + "loss": 0.9481, + "step": 300 + }, + { + "epoch": 0.7395577395577395, + "grad_norm": 1.534183187925811, + "learning_rate": 1.492953372498571e-06, + "loss": 0.9111, + "step": 301 + }, + { + "epoch": 0.742014742014742, + "grad_norm": 1.6285985984253928, + "learning_rate": 1.4894178478110854e-06, + "loss": 0.949, + "step": 302 + }, + { + "epoch": 0.7444717444717445, + "grad_norm": 1.521782496026601, + "learning_rate": 1.4858742601167138e-06, + "loss": 0.9405, + "step": 303 + }, + { + "epoch": 0.7469287469287469, + "grad_norm": 1.522930763087887, + "learning_rate": 1.4823226677949622e-06, + "loss": 0.949, + "step": 304 + }, + { + "epoch": 0.7493857493857494, + "grad_norm": 5.354032445168916, + "learning_rate": 1.4787631293572092e-06, + "loss": 0.9708, + "step": 305 + }, + { + "epoch": 0.7518427518427518, + "grad_norm": 1.199402404810751, + "learning_rate": 1.4751957034457446e-06, + "loss": 0.8895, + "step": 306 + }, + { + "epoch": 0.7542997542997543, + "grad_norm": 1.4889622591506924, + "learning_rate": 1.4716204488328005e-06, + "loss": 0.9258, + "step": 307 + }, + { + "epoch": 0.7567567567567568, + "grad_norm": 1.586929861427342, + "learning_rate": 1.4680374244195858e-06, + "loss": 0.9523, + "step": 308 + }, + { + "epoch": 0.7592137592137592, + "grad_norm": 1.7542006852572127, + "learning_rate": 1.464446689235314e-06, + "loss": 0.9671, + "step": 309 + }, + { + "epoch": 0.7616707616707616, + "grad_norm": 1.906886186477315, + "learning_rate": 1.4608483024362302e-06, + "loss": 0.9479, + "step": 310 + }, + { + "epoch": 0.7641277641277642, + "grad_norm": 1.3426576295326547, + "learning_rate": 1.4572423233046385e-06, + "loss": 0.9245, + "step": 311 + }, + { + "epoch": 0.7665847665847666, + "grad_norm": 1.544123812480689, + "learning_rate": 1.453628811247924e-06, + "loss": 0.9833, + "step": 312 + }, + { + "epoch": 0.769041769041769, + "grad_norm": 1.3070207560130556, + "learning_rate": 1.4500078257975745e-06, + "loss": 0.9518, + "step": 313 + }, + { + "epoch": 0.7714987714987716, + "grad_norm": 1.953852840958092, + "learning_rate": 1.4463794266081992e-06, + "loss": 0.8871, + "step": 314 + }, + { + "epoch": 0.773955773955774, + "grad_norm": 1.5767416529599687, + "learning_rate": 1.4427436734565473e-06, + "loss": 0.9921, + "step": 315 + }, + { + "epoch": 0.7764127764127764, + "grad_norm": 1.5600472966901433, + "learning_rate": 1.4391006262405212e-06, + "loss": 0.9068, + "step": 316 + }, + { + "epoch": 0.7788697788697788, + "grad_norm": 1.8524989491255999, + "learning_rate": 1.4354503449781913e-06, + "loss": 0.8956, + "step": 317 + }, + { + "epoch": 0.7813267813267813, + "grad_norm": 1.4701804028885344, + "learning_rate": 1.4317928898068066e-06, + "loss": 0.9547, + "step": 318 + }, + { + "epoch": 0.7837837837837838, + "grad_norm": 1.505304585103799, + "learning_rate": 1.4281283209818039e-06, + "loss": 0.9214, + "step": 319 + }, + { + "epoch": 0.7862407862407862, + "grad_norm": 1.5458422714438642, + "learning_rate": 1.424456698875815e-06, + "loss": 0.9711, + "step": 320 + }, + { + "epoch": 0.7886977886977887, + "grad_norm": 1.8126365786822831, + "learning_rate": 1.4207780839776732e-06, + "loss": 0.9638, + "step": 321 + }, + { + "epoch": 0.7911547911547911, + "grad_norm": 1.3634824425424106, + "learning_rate": 1.4170925368914151e-06, + "loss": 0.925, + "step": 322 + }, + { + "epoch": 0.7936117936117936, + "grad_norm": 1.5523412959213532, + "learning_rate": 1.413400118335283e-06, + "loss": 0.9681, + "step": 323 + }, + { + "epoch": 0.7960687960687961, + "grad_norm": 1.4281815228481956, + "learning_rate": 1.4097008891407245e-06, + "loss": 0.9469, + "step": 324 + }, + { + "epoch": 0.7985257985257985, + "grad_norm": 1.755464439866289, + "learning_rate": 1.4059949102513913e-06, + "loss": 0.9614, + "step": 325 + }, + { + "epoch": 0.800982800982801, + "grad_norm": 1.3755915444522364, + "learning_rate": 1.4022822427221323e-06, + "loss": 0.9344, + "step": 326 + }, + { + "epoch": 0.8034398034398035, + "grad_norm": 1.6217750884199984, + "learning_rate": 1.3985629477179915e-06, + "loss": 0.9499, + "step": 327 + }, + { + "epoch": 0.8058968058968059, + "grad_norm": 1.5965334157340827, + "learning_rate": 1.3948370865131978e-06, + "loss": 0.9433, + "step": 328 + }, + { + "epoch": 0.8083538083538083, + "grad_norm": 1.8371436403022903, + "learning_rate": 1.3911047204901558e-06, + "loss": 0.9899, + "step": 329 + }, + { + "epoch": 0.8108108108108109, + "grad_norm": 1.8839520322523513, + "learning_rate": 1.387365911138436e-06, + "loss": 0.9664, + "step": 330 + }, + { + "epoch": 0.8132678132678133, + "grad_norm": 1.495735998773526, + "learning_rate": 1.3836207200537594e-06, + "loss": 0.9784, + "step": 331 + }, + { + "epoch": 0.8157248157248157, + "grad_norm": 1.4893770976659941, + "learning_rate": 1.3798692089369855e-06, + "loss": 0.9046, + "step": 332 + }, + { + "epoch": 0.8181818181818182, + "grad_norm": 1.5373781226372853, + "learning_rate": 1.3761114395930928e-06, + "loss": 0.9461, + "step": 333 + }, + { + "epoch": 0.8206388206388207, + "grad_norm": 1.497679579290753, + "learning_rate": 1.3723474739301634e-06, + "loss": 0.9109, + "step": 334 + }, + { + "epoch": 0.8230958230958231, + "grad_norm": 1.6333646657141205, + "learning_rate": 1.3685773739583618e-06, + "loss": 0.893, + "step": 335 + }, + { + "epoch": 0.8255528255528255, + "grad_norm": 1.4444606139573457, + "learning_rate": 1.3648012017889121e-06, + "loss": 0.9616, + "step": 336 + }, + { + "epoch": 0.828009828009828, + "grad_norm": 1.633866194720376, + "learning_rate": 1.3610190196330775e-06, + "loss": 0.9316, + "step": 337 + }, + { + "epoch": 0.8304668304668305, + "grad_norm": 1.4965633625059045, + "learning_rate": 1.3572308898011327e-06, + "loss": 0.8995, + "step": 338 + }, + { + "epoch": 0.8329238329238329, + "grad_norm": 1.4593159509482825, + "learning_rate": 1.3534368747013394e-06, + "loss": 0.9416, + "step": 339 + }, + { + "epoch": 0.8353808353808354, + "grad_norm": 1.5060565758495885, + "learning_rate": 1.3496370368389165e-06, + "loss": 0.9385, + "step": 340 + }, + { + "epoch": 0.8378378378378378, + "grad_norm": 1.5228727270675948, + "learning_rate": 1.3458314388150113e-06, + "loss": 0.9249, + "step": 341 + }, + { + "epoch": 0.8402948402948403, + "grad_norm": 1.6137336621038334, + "learning_rate": 1.3420201433256689e-06, + "loss": 0.9549, + "step": 342 + }, + { + "epoch": 0.8427518427518428, + "grad_norm": 1.2744326108891244, + "learning_rate": 1.3382032131607966e-06, + "loss": 0.9275, + "step": 343 + }, + { + "epoch": 0.8452088452088452, + "grad_norm": 1.5480593545772812, + "learning_rate": 1.3343807112031326e-06, + "loss": 0.9342, + "step": 344 + }, + { + "epoch": 0.8476658476658476, + "grad_norm": 1.6690119423341088, + "learning_rate": 1.3305527004272087e-06, + "loss": 0.9375, + "step": 345 + }, + { + "epoch": 0.8501228501228502, + "grad_norm": 1.4907525884430786, + "learning_rate": 1.3267192438983117e-06, + "loss": 0.9183, + "step": 346 + }, + { + "epoch": 0.8525798525798526, + "grad_norm": 1.487251447150466, + "learning_rate": 1.3228804047714462e-06, + "loss": 0.9572, + "step": 347 + }, + { + "epoch": 0.855036855036855, + "grad_norm": 2.022254168777352, + "learning_rate": 1.3190362462902936e-06, + "loss": 0.9431, + "step": 348 + }, + { + "epoch": 0.8574938574938575, + "grad_norm": 1.570248437527878, + "learning_rate": 1.31518683178617e-06, + "loss": 0.8945, + "step": 349 + }, + { + "epoch": 0.85995085995086, + "grad_norm": 1.5017196199325449, + "learning_rate": 1.3113322246769816e-06, + "loss": 0.8841, + "step": 350 + }, + { + "epoch": 0.8624078624078624, + "grad_norm": 1.5003226831400922, + "learning_rate": 1.3074724884661831e-06, + "loss": 0.9785, + "step": 351 + }, + { + "epoch": 0.8648648648648649, + "grad_norm": 1.6237720015743136, + "learning_rate": 1.3036076867417286e-06, + "loss": 0.9942, + "step": 352 + }, + { + "epoch": 0.8673218673218673, + "grad_norm": 1.586236748920305, + "learning_rate": 1.299737883175024e-06, + "loss": 0.9036, + "step": 353 + }, + { + "epoch": 0.8697788697788698, + "grad_norm": 1.6356021082940453, + "learning_rate": 1.2958631415198813e-06, + "loss": 0.9723, + "step": 354 + }, + { + "epoch": 0.8722358722358723, + "grad_norm": 1.5448677133530493, + "learning_rate": 1.2919835256114637e-06, + "loss": 0.9391, + "step": 355 + }, + { + "epoch": 0.8746928746928747, + "grad_norm": 1.6212696290821345, + "learning_rate": 1.2880990993652377e-06, + "loss": 0.9202, + "step": 356 + }, + { + "epoch": 0.8771498771498771, + "grad_norm": 1.5176857112402762, + "learning_rate": 1.2842099267759175e-06, + "loss": 0.9037, + "step": 357 + }, + { + "epoch": 0.8796068796068796, + "grad_norm": 1.3944467856106335, + "learning_rate": 1.2803160719164126e-06, + "loss": 0.8854, + "step": 358 + }, + { + "epoch": 0.8820638820638821, + "grad_norm": 1.4377067419239682, + "learning_rate": 1.2764175989367715e-06, + "loss": 0.905, + "step": 359 + }, + { + "epoch": 0.8845208845208845, + "grad_norm": 1.6123612736739008, + "learning_rate": 1.272514572063124e-06, + "loss": 0.9881, + "step": 360 + }, + { + "epoch": 0.8869778869778869, + "grad_norm": 1.589984246633212, + "learning_rate": 1.2686070555966253e-06, + "loss": 0.9529, + "step": 361 + }, + { + "epoch": 0.8894348894348895, + "grad_norm": 1.4049932596277341, + "learning_rate": 1.2646951139123933e-06, + "loss": 0.8715, + "step": 362 + }, + { + "epoch": 0.8918918918918919, + "grad_norm": 1.7376495814908663, + "learning_rate": 1.260778811458452e-06, + "loss": 0.9551, + "step": 363 + }, + { + "epoch": 0.8943488943488943, + "grad_norm": 1.6356081133105504, + "learning_rate": 1.2568582127546661e-06, + "loss": 0.9357, + "step": 364 + }, + { + "epoch": 0.8968058968058968, + "grad_norm": 1.5620481027616258, + "learning_rate": 1.2529333823916806e-06, + "loss": 0.9724, + "step": 365 + }, + { + "epoch": 0.8992628992628993, + "grad_norm": 1.678396238284589, + "learning_rate": 1.2490043850298557e-06, + "loss": 0.9156, + "step": 366 + }, + { + "epoch": 0.9017199017199017, + "grad_norm": 2.354758429687385, + "learning_rate": 1.2450712853982013e-06, + "loss": 0.8628, + "step": 367 + }, + { + "epoch": 0.9041769041769042, + "grad_norm": 1.600899191967723, + "learning_rate": 1.2411341482933108e-06, + "loss": 0.9431, + "step": 368 + }, + { + "epoch": 0.9066339066339066, + "grad_norm": 1.4058440155556784, + "learning_rate": 1.2371930385782943e-06, + "loss": 0.9446, + "step": 369 + }, + { + "epoch": 0.9090909090909091, + "grad_norm": 1.4613065350370973, + "learning_rate": 1.233248021181709e-06, + "loss": 0.9164, + "step": 370 + }, + { + "epoch": 0.9115479115479116, + "grad_norm": 1.6113469899034865, + "learning_rate": 1.22929916109649e-06, + "loss": 0.9209, + "step": 371 + }, + { + "epoch": 0.914004914004914, + "grad_norm": 1.4540218388463704, + "learning_rate": 1.2253465233788794e-06, + "loss": 0.9865, + "step": 372 + }, + { + "epoch": 0.9164619164619164, + "grad_norm": 1.4002046410996662, + "learning_rate": 1.2213901731473553e-06, + "loss": 0.9229, + "step": 373 + }, + { + "epoch": 0.918918918918919, + "grad_norm": 1.4545211228962136, + "learning_rate": 1.217430175581557e-06, + "loss": 0.9401, + "step": 374 + }, + { + "epoch": 0.9213759213759214, + "grad_norm": 1.4460543335595146, + "learning_rate": 1.2134665959212135e-06, + "loss": 0.9552, + "step": 375 + }, + { + "epoch": 0.9238329238329238, + "grad_norm": 1.5791517432252247, + "learning_rate": 1.209499499465068e-06, + "loss": 0.9547, + "step": 376 + }, + { + "epoch": 0.9262899262899262, + "grad_norm": 1.4523281683604505, + "learning_rate": 1.2055289515698006e-06, + "loss": 0.9669, + "step": 377 + }, + { + "epoch": 0.9287469287469288, + "grad_norm": 1.8584285735562196, + "learning_rate": 1.2015550176489537e-06, + "loss": 0.9434, + "step": 378 + }, + { + "epoch": 0.9312039312039312, + "grad_norm": 1.5086682042761248, + "learning_rate": 1.1975777631718531e-06, + "loss": 0.9377, + "step": 379 + }, + { + "epoch": 0.9336609336609336, + "grad_norm": 1.7455613963553367, + "learning_rate": 1.19359725366253e-06, + "loss": 0.919, + "step": 380 + }, + { + "epoch": 0.9361179361179361, + "grad_norm": 1.5307030954736855, + "learning_rate": 1.1896135546986409e-06, + "loss": 0.9282, + "step": 381 + }, + { + "epoch": 0.9385749385749386, + "grad_norm": 1.5159714020228543, + "learning_rate": 1.1856267319103876e-06, + "loss": 0.9765, + "step": 382 + }, + { + "epoch": 0.941031941031941, + "grad_norm": 1.5839779729377836, + "learning_rate": 1.1816368509794364e-06, + "loss": 0.9793, + "step": 383 + }, + { + "epoch": 0.9434889434889435, + "grad_norm": 1.3472301054596898, + "learning_rate": 1.177643977637835e-06, + "loss": 0.9189, + "step": 384 + }, + { + "epoch": 0.9459459459459459, + "grad_norm": 1.5340710636652473, + "learning_rate": 1.1736481776669305e-06, + "loss": 0.9663, + "step": 385 + }, + { + "epoch": 0.9484029484029484, + "grad_norm": 1.4047565697947708, + "learning_rate": 1.1696495168962846e-06, + "loss": 0.9207, + "step": 386 + }, + { + "epoch": 0.9508599508599509, + "grad_norm": 1.6035760956637677, + "learning_rate": 1.165648061202591e-06, + "loss": 0.9793, + "step": 387 + }, + { + "epoch": 0.9533169533169533, + "grad_norm": 1.4948457768549497, + "learning_rate": 1.1616438765085882e-06, + "loss": 0.9381, + "step": 388 + }, + { + "epoch": 0.9557739557739557, + "grad_norm": 1.5241351339551144, + "learning_rate": 1.1576370287819734e-06, + "loss": 0.931, + "step": 389 + }, + { + "epoch": 0.9582309582309583, + "grad_norm": 1.4995817174577337, + "learning_rate": 1.1536275840343183e-06, + "loss": 0.9594, + "step": 390 + }, + { + "epoch": 0.9606879606879607, + "grad_norm": 1.8781230050093483, + "learning_rate": 1.149615608319978e-06, + "loss": 0.9536, + "step": 391 + }, + { + "epoch": 0.9631449631449631, + "grad_norm": 1.5385341714632899, + "learning_rate": 1.145601167735005e-06, + "loss": 0.9281, + "step": 392 + }, + { + "epoch": 0.9656019656019657, + "grad_norm": 1.6183085918500024, + "learning_rate": 1.1415843284160597e-06, + "loss": 0.9366, + "step": 393 + }, + { + "epoch": 0.9680589680589681, + "grad_norm": 1.5534604794185942, + "learning_rate": 1.137565156539322e-06, + "loss": 0.9316, + "step": 394 + }, + { + "epoch": 0.9705159705159705, + "grad_norm": 1.423811752511575, + "learning_rate": 1.133543718319398e-06, + "loss": 0.8927, + "step": 395 + }, + { + "epoch": 0.972972972972973, + "grad_norm": 1.4578205842089746, + "learning_rate": 1.1295200800082325e-06, + "loss": 0.951, + "step": 396 + }, + { + "epoch": 0.9754299754299754, + "grad_norm": 1.6386231495575438, + "learning_rate": 1.125494307894016e-06, + "loss": 0.8952, + "step": 397 + }, + { + "epoch": 0.9778869778869779, + "grad_norm": 1.5075389802303287, + "learning_rate": 1.1214664683000924e-06, + "loss": 0.8966, + "step": 398 + }, + { + "epoch": 0.9803439803439803, + "grad_norm": 1.631579827773725, + "learning_rate": 1.1174366275838662e-06, + "loss": 0.9411, + "step": 399 + }, + { + "epoch": 0.9828009828009828, + "grad_norm": 1.8249088799770508, + "learning_rate": 1.1134048521357116e-06, + "loss": 0.9232, + "step": 400 + }, + { + "epoch": 0.9852579852579852, + "grad_norm": 1.4813873308103862, + "learning_rate": 1.1093712083778746e-06, + "loss": 0.9625, + "step": 401 + }, + { + "epoch": 0.9877149877149877, + "grad_norm": 1.5996089579510389, + "learning_rate": 1.105335762763382e-06, + "loss": 0.9592, + "step": 402 + }, + { + "epoch": 0.9901719901719902, + "grad_norm": 1.4507338601430246, + "learning_rate": 1.1012985817749462e-06, + "loss": 0.9558, + "step": 403 + }, + { + "epoch": 0.9926289926289926, + "grad_norm": 1.452275573847242, + "learning_rate": 1.097259731923869e-06, + "loss": 0.9582, + "step": 404 + }, + { + "epoch": 0.995085995085995, + "grad_norm": 1.4529835839819856, + "learning_rate": 1.0932192797489458e-06, + "loss": 0.8952, + "step": 405 + }, + { + "epoch": 0.9975429975429976, + "grad_norm": 2.622272111272181, + "learning_rate": 1.0891772918153694e-06, + "loss": 0.918, + "step": 406 + }, + { + "epoch": 1.0, + "grad_norm": 1.5285112861748582, + "learning_rate": 1.0851338347136355e-06, + "loss": 0.877, + "step": 407 + }, + { + "epoch": 1.0024570024570025, + "grad_norm": 1.4118845101420152, + "learning_rate": 1.0810889750584424e-06, + "loss": 0.8647, + "step": 408 + }, + { + "epoch": 1.0049140049140048, + "grad_norm": 1.6747623259684568, + "learning_rate": 1.0770427794875948e-06, + "loss": 0.9671, + "step": 409 + }, + { + "epoch": 1.0073710073710074, + "grad_norm": 1.307864675513283, + "learning_rate": 1.0729953146609075e-06, + "loss": 0.8296, + "step": 410 + }, + { + "epoch": 1.00982800982801, + "grad_norm": 1.4553500313995769, + "learning_rate": 1.0689466472591048e-06, + "loss": 0.937, + "step": 411 + }, + { + "epoch": 1.0122850122850122, + "grad_norm": 1.4315419855901317, + "learning_rate": 1.0648968439827239e-06, + "loss": 0.9042, + "step": 412 + }, + { + "epoch": 1.0147420147420148, + "grad_norm": 1.5739844573571937, + "learning_rate": 1.0608459715510139e-06, + "loss": 0.9011, + "step": 413 + }, + { + "epoch": 1.0171990171990173, + "grad_norm": 1.437687738713176, + "learning_rate": 1.0567940967008395e-06, + "loss": 0.9363, + "step": 414 + }, + { + "epoch": 1.0196560196560196, + "grad_norm": 1.534448861262034, + "learning_rate": 1.052741286185579e-06, + "loss": 0.9539, + "step": 415 + }, + { + "epoch": 1.0221130221130221, + "grad_norm": 1.644806928510893, + "learning_rate": 1.0486876067740252e-06, + "loss": 0.9802, + "step": 416 + }, + { + "epoch": 1.0245700245700247, + "grad_norm": 1.5874571507432869, + "learning_rate": 1.0446331252492864e-06, + "loss": 0.8751, + "step": 417 + }, + { + "epoch": 1.027027027027027, + "grad_norm": 1.5244813435420985, + "learning_rate": 1.0405779084076856e-06, + "loss": 0.8844, + "step": 418 + }, + { + "epoch": 1.0294840294840295, + "grad_norm": 1.5721376461339462, + "learning_rate": 1.036522023057659e-06, + "loss": 0.9593, + "step": 419 + }, + { + "epoch": 1.031941031941032, + "grad_norm": 3.006639344932307, + "learning_rate": 1.0324655360186567e-06, + "loss": 0.9055, + "step": 420 + }, + { + "epoch": 1.0343980343980343, + "grad_norm": 1.3782225003410025, + "learning_rate": 1.0284085141200423e-06, + "loss": 0.9183, + "step": 421 + }, + { + "epoch": 1.0368550368550369, + "grad_norm": 1.376873801999615, + "learning_rate": 1.0243510241999897e-06, + "loss": 0.8767, + "step": 422 + }, + { + "epoch": 1.0393120393120394, + "grad_norm": 1.394373200963149, + "learning_rate": 1.0202931331043839e-06, + "loss": 0.8821, + "step": 423 + }, + { + "epoch": 1.0417690417690417, + "grad_norm": 1.571320978169866, + "learning_rate": 1.0162349076857191e-06, + "loss": 0.8959, + "step": 424 + }, + { + "epoch": 1.0442260442260443, + "grad_norm": 1.5336886338120022, + "learning_rate": 1.0121764148019975e-06, + "loss": 0.9375, + "step": 425 + }, + { + "epoch": 1.0466830466830466, + "grad_norm": 1.4693963607748404, + "learning_rate": 1.0081177213156277e-06, + "loss": 0.8909, + "step": 426 + }, + { + "epoch": 1.049140049140049, + "grad_norm": 1.5830327984938106, + "learning_rate": 1.004058894092323e-06, + "loss": 0.9207, + "step": 427 + }, + { + "epoch": 1.0515970515970516, + "grad_norm": 1.5170501490428927, + "learning_rate": 1e-06, + "loss": 0.9036, + "step": 428 + }, + { + "epoch": 1.054054054054054, + "grad_norm": 1.5692066826348359, + "learning_rate": 9.95941105907677e-07, + "loss": 0.9193, + "step": 429 + }, + { + "epoch": 1.0565110565110565, + "grad_norm": 1.4357790543992526, + "learning_rate": 9.918822786843724e-07, + "loss": 0.9124, + "step": 430 + }, + { + "epoch": 1.058968058968059, + "grad_norm": 1.423440533928829, + "learning_rate": 9.878235851980024e-07, + "loss": 0.9146, + "step": 431 + }, + { + "epoch": 1.0614250614250613, + "grad_norm": 1.590468077725602, + "learning_rate": 9.837650923142808e-07, + "loss": 0.8705, + "step": 432 + }, + { + "epoch": 1.0638820638820639, + "grad_norm": 1.461231262115798, + "learning_rate": 9.797068668956163e-07, + "loss": 0.9438, + "step": 433 + }, + { + "epoch": 1.0663390663390664, + "grad_norm": 1.42661968764142, + "learning_rate": 9.756489758000105e-07, + "loss": 0.8747, + "step": 434 + }, + { + "epoch": 1.0687960687960687, + "grad_norm": 1.4657641576186988, + "learning_rate": 9.715914858799576e-07, + "loss": 0.8939, + "step": 435 + }, + { + "epoch": 1.0712530712530712, + "grad_norm": 1.5022324634860065, + "learning_rate": 9.675344639813434e-07, + "loss": 0.922, + "step": 436 + }, + { + "epoch": 1.0737100737100738, + "grad_norm": 1.424039088022573, + "learning_rate": 9.63477976942341e-07, + "loss": 0.953, + "step": 437 + }, + { + "epoch": 1.076167076167076, + "grad_norm": 1.5965077678963564, + "learning_rate": 9.594220915923147e-07, + "loss": 0.9443, + "step": 438 + }, + { + "epoch": 1.0786240786240786, + "grad_norm": 1.6371319904729476, + "learning_rate": 9.553668747507138e-07, + "loss": 0.9228, + "step": 439 + }, + { + "epoch": 1.0810810810810811, + "grad_norm": 1.8309314739878833, + "learning_rate": 9.51312393225975e-07, + "loss": 0.8753, + "step": 440 + }, + { + "epoch": 1.0835380835380835, + "grad_norm": 1.39319457517673, + "learning_rate": 9.472587138144215e-07, + "loss": 0.8939, + "step": 441 + }, + { + "epoch": 1.085995085995086, + "grad_norm": 1.5070197645260877, + "learning_rate": 9.432059032991607e-07, + "loss": 0.9123, + "step": 442 + }, + { + "epoch": 1.0884520884520885, + "grad_norm": 1.3753531136081873, + "learning_rate": 9.391540284489861e-07, + "loss": 0.9328, + "step": 443 + }, + { + "epoch": 1.0909090909090908, + "grad_norm": 1.4036108581529319, + "learning_rate": 9.351031560172764e-07, + "loss": 0.9138, + "step": 444 + }, + { + "epoch": 1.0933660933660934, + "grad_norm": 1.460079134056615, + "learning_rate": 9.310533527408951e-07, + "loss": 0.8736, + "step": 445 + }, + { + "epoch": 1.095823095823096, + "grad_norm": 2.1567900139463663, + "learning_rate": 9.270046853390924e-07, + "loss": 0.9649, + "step": 446 + }, + { + "epoch": 1.0982800982800982, + "grad_norm": 1.5086116320926348, + "learning_rate": 9.229572205124051e-07, + "loss": 0.8839, + "step": 447 + }, + { + "epoch": 1.1007371007371007, + "grad_norm": 1.4696338643312472, + "learning_rate": 9.189110249415576e-07, + "loss": 0.8971, + "step": 448 + }, + { + "epoch": 1.1031941031941033, + "grad_norm": 1.5799491675072612, + "learning_rate": 9.148661652863642e-07, + "loss": 0.8839, + "step": 449 + }, + { + "epoch": 1.1056511056511056, + "grad_norm": 1.633966629237029, + "learning_rate": 9.108227081846304e-07, + "loss": 0.9053, + "step": 450 + }, + { + "epoch": 1.1081081081081081, + "grad_norm": 1.5227318714591582, + "learning_rate": 9.067807202510542e-07, + "loss": 0.8734, + "step": 451 + }, + { + "epoch": 1.1105651105651106, + "grad_norm": 1.6390683121943366, + "learning_rate": 9.027402680761308e-07, + "loss": 0.8389, + "step": 452 + }, + { + "epoch": 1.113022113022113, + "grad_norm": 1.8480120093100039, + "learning_rate": 8.987014182250537e-07, + "loss": 0.8977, + "step": 453 + }, + { + "epoch": 1.1154791154791155, + "grad_norm": 1.6432113495263132, + "learning_rate": 8.94664237236618e-07, + "loss": 0.8924, + "step": 454 + }, + { + "epoch": 1.117936117936118, + "grad_norm": 1.7383339935167776, + "learning_rate": 8.906287916221258e-07, + "loss": 0.9181, + "step": 455 + }, + { + "epoch": 1.1203931203931203, + "grad_norm": 1.5888604077241295, + "learning_rate": 8.865951478642886e-07, + "loss": 0.9276, + "step": 456 + }, + { + "epoch": 1.1228501228501229, + "grad_norm": 1.4607976669981695, + "learning_rate": 8.825633724161335e-07, + "loss": 0.8802, + "step": 457 + }, + { + "epoch": 1.1253071253071254, + "grad_norm": 1.1981728862199128, + "learning_rate": 8.785335316999078e-07, + "loss": 0.9242, + "step": 458 + }, + { + "epoch": 1.1277641277641277, + "grad_norm": 1.5537897318688183, + "learning_rate": 8.745056921059839e-07, + "loss": 0.9001, + "step": 459 + }, + { + "epoch": 1.1302211302211302, + "grad_norm": 1.475963647033354, + "learning_rate": 8.704799199917672e-07, + "loss": 0.8775, + "step": 460 + }, + { + "epoch": 1.1326781326781328, + "grad_norm": 1.3982302407588918, + "learning_rate": 8.664562816806021e-07, + "loss": 0.8828, + "step": 461 + }, + { + "epoch": 1.135135135135135, + "grad_norm": 1.4892642936338247, + "learning_rate": 8.624348434606781e-07, + "loss": 0.9042, + "step": 462 + }, + { + "epoch": 1.1375921375921376, + "grad_norm": 1.8141339480004706, + "learning_rate": 8.584156715839401e-07, + "loss": 0.8748, + "step": 463 + }, + { + "epoch": 1.1400491400491402, + "grad_norm": 1.641438804529928, + "learning_rate": 8.543988322649953e-07, + "loss": 0.9006, + "step": 464 + }, + { + "epoch": 1.1425061425061425, + "grad_norm": 1.5215620001272283, + "learning_rate": 8.503843916800219e-07, + "loss": 0.9418, + "step": 465 + }, + { + "epoch": 1.144963144963145, + "grad_norm": 1.5095379213125475, + "learning_rate": 8.463724159656815e-07, + "loss": 0.9072, + "step": 466 + }, + { + "epoch": 1.1474201474201475, + "grad_norm": 1.4101920682297036, + "learning_rate": 8.423629712180265e-07, + "loss": 0.8929, + "step": 467 + }, + { + "epoch": 1.1498771498771498, + "grad_norm": 1.5017183949322925, + "learning_rate": 8.38356123491412e-07, + "loss": 0.9246, + "step": 468 + }, + { + "epoch": 1.1523341523341524, + "grad_norm": 1.4487713518305103, + "learning_rate": 8.34351938797409e-07, + "loss": 0.9978, + "step": 469 + }, + { + "epoch": 1.154791154791155, + "grad_norm": 1.9019229859778328, + "learning_rate": 8.303504831037153e-07, + "loss": 0.8963, + "step": 470 + }, + { + "epoch": 1.1572481572481572, + "grad_norm": 1.5553306743565323, + "learning_rate": 8.263518223330696e-07, + "loss": 0.9076, + "step": 471 + }, + { + "epoch": 1.1597051597051597, + "grad_norm": 1.528301521498108, + "learning_rate": 8.22356022362165e-07, + "loss": 0.9005, + "step": 472 + }, + { + "epoch": 1.1621621621621623, + "grad_norm": 1.4246755498283765, + "learning_rate": 8.183631490205635e-07, + "loss": 0.8751, + "step": 473 + }, + { + "epoch": 1.1646191646191646, + "grad_norm": 1.4837658920113785, + "learning_rate": 8.143732680896123e-07, + "loss": 0.8911, + "step": 474 + }, + { + "epoch": 1.1670761670761671, + "grad_norm": 1.3838573607006248, + "learning_rate": 8.103864453013592e-07, + "loss": 0.8752, + "step": 475 + }, + { + "epoch": 1.1695331695331694, + "grad_norm": 1.3956886722399857, + "learning_rate": 8.0640274633747e-07, + "loss": 0.907, + "step": 476 + }, + { + "epoch": 1.171990171990172, + "grad_norm": 1.4606205140333972, + "learning_rate": 8.024222368281468e-07, + "loss": 0.964, + "step": 477 + }, + { + "epoch": 1.1744471744471745, + "grad_norm": 1.438679071763127, + "learning_rate": 7.984449823510466e-07, + "loss": 0.8499, + "step": 478 + }, + { + "epoch": 1.1769041769041768, + "grad_norm": 1.6564632287349166, + "learning_rate": 7.944710484301994e-07, + "loss": 0.892, + "step": 479 + }, + { + "epoch": 1.1793611793611793, + "grad_norm": 1.5772645911948133, + "learning_rate": 7.905005005349319e-07, + "loss": 0.8451, + "step": 480 + }, + { + "epoch": 1.1818181818181819, + "grad_norm": 1.5408140168729385, + "learning_rate": 7.865334040787865e-07, + "loss": 0.9295, + "step": 481 + }, + { + "epoch": 1.1842751842751842, + "grad_norm": 1.5386445563903186, + "learning_rate": 7.825698244184431e-07, + "loss": 0.9029, + "step": 482 + }, + { + "epoch": 1.1867321867321867, + "grad_norm": 1.3266763848904035, + "learning_rate": 7.786098268526447e-07, + "loss": 0.8801, + "step": 483 + }, + { + "epoch": 1.1891891891891893, + "grad_norm": 1.7962161638699905, + "learning_rate": 7.746534766211206e-07, + "loss": 0.8578, + "step": 484 + }, + { + "epoch": 1.1916461916461916, + "grad_norm": 1.477358511769343, + "learning_rate": 7.7070083890351e-07, + "loss": 0.9048, + "step": 485 + }, + { + "epoch": 1.194103194103194, + "grad_norm": 1.4147723236609893, + "learning_rate": 7.667519788182911e-07, + "loss": 0.8898, + "step": 486 + }, + { + "epoch": 1.1965601965601966, + "grad_norm": 1.55983989969961, + "learning_rate": 7.628069614217057e-07, + "loss": 0.9509, + "step": 487 + }, + { + "epoch": 1.199017199017199, + "grad_norm": 1.4383185383531423, + "learning_rate": 7.588658517066892e-07, + "loss": 0.9418, + "step": 488 + }, + { + "epoch": 1.2014742014742015, + "grad_norm": 1.7624942370774765, + "learning_rate": 7.549287146017988e-07, + "loss": 0.8313, + "step": 489 + }, + { + "epoch": 1.203931203931204, + "grad_norm": 1.67001710532923, + "learning_rate": 7.509956149701443e-07, + "loss": 0.8549, + "step": 490 + }, + { + "epoch": 1.2063882063882063, + "grad_norm": 1.6268288582039303, + "learning_rate": 7.470666176083191e-07, + "loss": 0.8649, + "step": 491 + }, + { + "epoch": 1.2088452088452089, + "grad_norm": 1.4853773676445685, + "learning_rate": 7.431417872453339e-07, + "loss": 0.8956, + "step": 492 + }, + { + "epoch": 1.2113022113022114, + "grad_norm": 1.548578088799907, + "learning_rate": 7.39221188541548e-07, + "loss": 0.923, + "step": 493 + }, + { + "epoch": 1.2137592137592137, + "grad_norm": 1.710249233454194, + "learning_rate": 7.353048860876063e-07, + "loss": 0.9336, + "step": 494 + }, + { + "epoch": 1.2162162162162162, + "grad_norm": 1.5916044293544283, + "learning_rate": 7.31392944403375e-07, + "loss": 0.9304, + "step": 495 + }, + { + "epoch": 1.2186732186732188, + "grad_norm": 1.5308549158931122, + "learning_rate": 7.274854279368758e-07, + "loss": 0.9349, + "step": 496 + }, + { + "epoch": 1.221130221130221, + "grad_norm": 1.5229178156260872, + "learning_rate": 7.235824010632283e-07, + "loss": 0.9153, + "step": 497 + }, + { + "epoch": 1.2235872235872236, + "grad_norm": 1.462343882146908, + "learning_rate": 7.196839280835875e-07, + "loss": 0.9376, + "step": 498 + }, + { + "epoch": 1.2260442260442261, + "grad_norm": 1.5273674514727056, + "learning_rate": 7.157900732240827e-07, + "loss": 0.9231, + "step": 499 + }, + { + "epoch": 1.2285012285012284, + "grad_norm": 1.396230424270659, + "learning_rate": 7.119009006347624e-07, + "loss": 0.8868, + "step": 500 + }, + { + "epoch": 1.230958230958231, + "grad_norm": 1.6364859015395088, + "learning_rate": 7.080164743885363e-07, + "loss": 0.9287, + "step": 501 + }, + { + "epoch": 1.2334152334152333, + "grad_norm": 1.5048783187743708, + "learning_rate": 7.041368584801186e-07, + "loss": 0.932, + "step": 502 + }, + { + "epoch": 1.2358722358722358, + "grad_norm": 1.5929336784240413, + "learning_rate": 7.002621168249758e-07, + "loss": 0.9302, + "step": 503 + }, + { + "epoch": 1.2383292383292384, + "grad_norm": 1.5383384042486663, + "learning_rate": 6.963923132582715e-07, + "loss": 0.9059, + "step": 504 + }, + { + "epoch": 1.2407862407862407, + "grad_norm": 1.6790422515254697, + "learning_rate": 6.925275115338167e-07, + "loss": 0.8862, + "step": 505 + }, + { + "epoch": 1.2432432432432432, + "grad_norm": 1.4423572903428208, + "learning_rate": 6.886677753230183e-07, + "loss": 0.9004, + "step": 506 + }, + { + "epoch": 1.2457002457002457, + "grad_norm": 1.5146851789864428, + "learning_rate": 6.848131682138303e-07, + "loss": 0.8873, + "step": 507 + }, + { + "epoch": 1.248157248157248, + "grad_norm": 1.7060116939979457, + "learning_rate": 6.809637537097062e-07, + "loss": 0.9047, + "step": 508 + }, + { + "epoch": 1.2506142506142506, + "grad_norm": 1.3733517118208705, + "learning_rate": 6.77119595228554e-07, + "loss": 0.8742, + "step": 509 + }, + { + "epoch": 1.253071253071253, + "grad_norm": 1.4283683639472458, + "learning_rate": 6.732807561016885e-07, + "loss": 0.8697, + "step": 510 + }, + { + "epoch": 1.2555282555282554, + "grad_norm": 1.4832890039194944, + "learning_rate": 6.694472995727914e-07, + "loss": 0.925, + "step": 511 + }, + { + "epoch": 1.257985257985258, + "grad_norm": 1.4760580473983451, + "learning_rate": 6.656192887968674e-07, + "loss": 0.8959, + "step": 512 + }, + { + "epoch": 1.2604422604422605, + "grad_norm": 1.4517094927145753, + "learning_rate": 6.617967868392036e-07, + "loss": 0.8935, + "step": 513 + }, + { + "epoch": 1.2628992628992628, + "grad_norm": 1.5125319950260128, + "learning_rate": 6.579798566743313e-07, + "loss": 0.8556, + "step": 514 + }, + { + "epoch": 1.2653562653562653, + "grad_norm": 1.4510038145639286, + "learning_rate": 6.541685611849886e-07, + "loss": 0.9068, + "step": 515 + }, + { + "epoch": 1.2678132678132679, + "grad_norm": 1.786431321409276, + "learning_rate": 6.503629631610836e-07, + "loss": 0.8951, + "step": 516 + }, + { + "epoch": 1.2702702702702702, + "grad_norm": 4.646846867306252, + "learning_rate": 6.465631252986608e-07, + "loss": 0.8717, + "step": 517 + }, + { + "epoch": 1.2727272727272727, + "grad_norm": 1.6325223465886023, + "learning_rate": 6.427691101988672e-07, + "loss": 0.9277, + "step": 518 + }, + { + "epoch": 1.2751842751842752, + "grad_norm": 1.46996914036538, + "learning_rate": 6.389809803669225e-07, + "loss": 0.8614, + "step": 519 + }, + { + "epoch": 1.2776412776412776, + "grad_norm": 1.4901572767377598, + "learning_rate": 6.351987982110879e-07, + "loss": 0.8791, + "step": 520 + }, + { + "epoch": 1.28009828009828, + "grad_norm": 1.3956563694145843, + "learning_rate": 6.314226260416383e-07, + "loss": 0.8682, + "step": 521 + }, + { + "epoch": 1.2825552825552826, + "grad_norm": 1.527881875257898, + "learning_rate": 6.276525260698363e-07, + "loss": 0.9153, + "step": 522 + }, + { + "epoch": 1.285012285012285, + "grad_norm": 1.5192904446038689, + "learning_rate": 6.238885604069074e-07, + "loss": 0.9263, + "step": 523 + }, + { + "epoch": 1.2874692874692875, + "grad_norm": 1.5953405324354368, + "learning_rate": 6.201307910630145e-07, + "loss": 0.8788, + "step": 524 + }, + { + "epoch": 1.28992628992629, + "grad_norm": 1.4403593350780215, + "learning_rate": 6.163792799462402e-07, + "loss": 0.9414, + "step": 525 + }, + { + "epoch": 1.2923832923832923, + "grad_norm": 1.4506987352971705, + "learning_rate": 6.126340888615642e-07, + "loss": 0.9307, + "step": 526 + }, + { + "epoch": 1.2948402948402948, + "grad_norm": 1.610182690312099, + "learning_rate": 6.088952795098441e-07, + "loss": 0.9168, + "step": 527 + }, + { + "epoch": 1.2972972972972974, + "grad_norm": 1.5978713160514366, + "learning_rate": 6.051629134868019e-07, + "loss": 0.8393, + "step": 528 + }, + { + "epoch": 1.2997542997542997, + "grad_norm": 1.3547671161607415, + "learning_rate": 6.014370522820085e-07, + "loss": 0.8972, + "step": 529 + }, + { + "epoch": 1.3022113022113022, + "grad_norm": 1.3090769200596557, + "learning_rate": 5.977177572778677e-07, + "loss": 0.8823, + "step": 530 + }, + { + "epoch": 1.3046683046683047, + "grad_norm": 1.4761807072270576, + "learning_rate": 5.940050897486089e-07, + "loss": 0.9733, + "step": 531 + }, + { + "epoch": 1.307125307125307, + "grad_norm": 1.5243972285729563, + "learning_rate": 5.902991108592754e-07, + "loss": 0.8859, + "step": 532 + }, + { + "epoch": 1.3095823095823096, + "grad_norm": 1.764985461344906, + "learning_rate": 5.865998816647171e-07, + "loss": 0.8868, + "step": 533 + }, + { + "epoch": 1.3120393120393121, + "grad_norm": 1.6317180984567103, + "learning_rate": 5.829074631085852e-07, + "loss": 0.9361, + "step": 534 + }, + { + "epoch": 1.3144963144963144, + "grad_norm": 1.4981668702031359, + "learning_rate": 5.792219160223267e-07, + "loss": 0.8472, + "step": 535 + }, + { + "epoch": 1.316953316953317, + "grad_norm": 2.637743576229916, + "learning_rate": 5.755433011241851e-07, + "loss": 0.8736, + "step": 536 + }, + { + "epoch": 1.3194103194103195, + "grad_norm": 1.3474590182260624, + "learning_rate": 5.718716790181965e-07, + "loss": 0.8794, + "step": 537 + }, + { + "epoch": 1.3218673218673218, + "grad_norm": 1.42787784398219, + "learning_rate": 5.682071101931936e-07, + "loss": 0.8738, + "step": 538 + }, + { + "epoch": 1.3243243243243243, + "grad_norm": 1.6067570011968997, + "learning_rate": 5.645496550218089e-07, + "loss": 0.8879, + "step": 539 + }, + { + "epoch": 1.3267813267813269, + "grad_norm": 1.5914444312640619, + "learning_rate": 5.60899373759479e-07, + "loss": 0.8754, + "step": 540 + }, + { + "epoch": 1.3292383292383292, + "grad_norm": 1.4758896244599573, + "learning_rate": 5.572563265434526e-07, + "loss": 0.9307, + "step": 541 + }, + { + "epoch": 1.3316953316953317, + "grad_norm": 1.2626534960911335, + "learning_rate": 5.536205733918007e-07, + "loss": 0.843, + "step": 542 + }, + { + "epoch": 1.3341523341523343, + "grad_norm": 1.701842268857865, + "learning_rate": 5.499921742024257e-07, + "loss": 0.9247, + "step": 543 + }, + { + "epoch": 1.3366093366093366, + "grad_norm": 1.4205425380782735, + "learning_rate": 5.463711887520759e-07, + "loss": 0.8628, + "step": 544 + }, + { + "epoch": 1.339066339066339, + "grad_norm": 1.9140145931124193, + "learning_rate": 5.427576766953614e-07, + "loss": 0.9073, + "step": 545 + }, + { + "epoch": 1.3415233415233416, + "grad_norm": 1.7824661674136197, + "learning_rate": 5.391516975637699e-07, + "loss": 0.9133, + "step": 546 + }, + { + "epoch": 1.343980343980344, + "grad_norm": 1.5617612083029389, + "learning_rate": 5.355533107646858e-07, + "loss": 0.9531, + "step": 547 + }, + { + "epoch": 1.3464373464373465, + "grad_norm": 1.5245643140200376, + "learning_rate": 5.319625755804138e-07, + "loss": 0.8915, + "step": 548 + }, + { + "epoch": 1.348894348894349, + "grad_norm": 1.7286911090374188, + "learning_rate": 5.283795511671994e-07, + "loss": 0.863, + "step": 549 + }, + { + "epoch": 1.3513513513513513, + "grad_norm": 1.6181660068040473, + "learning_rate": 5.248042965542559e-07, + "loss": 0.8852, + "step": 550 + }, + { + "epoch": 1.3538083538083538, + "grad_norm": 1.5075998723303305, + "learning_rate": 5.212368706427912e-07, + "loss": 0.8965, + "step": 551 + }, + { + "epoch": 1.3562653562653564, + "grad_norm": 1.5100541752180852, + "learning_rate": 5.176773322050381e-07, + "loss": 0.8829, + "step": 552 + }, + { + "epoch": 1.3587223587223587, + "grad_norm": 1.863955084942053, + "learning_rate": 5.141257398832862e-07, + "loss": 0.9475, + "step": 553 + }, + { + "epoch": 1.3611793611793612, + "grad_norm": 1.314918195852255, + "learning_rate": 5.105821521889147e-07, + "loss": 0.8643, + "step": 554 + }, + { + "epoch": 1.3636363636363638, + "grad_norm": 1.8350990255330366, + "learning_rate": 5.070466275014287e-07, + "loss": 0.9441, + "step": 555 + }, + { + "epoch": 1.366093366093366, + "grad_norm": 1.5322707436835765, + "learning_rate": 5.03519224067499e-07, + "loss": 0.9033, + "step": 556 + }, + { + "epoch": 1.3685503685503686, + "grad_norm": 1.5974243335350318, + "learning_rate": 5.000000000000002e-07, + "loss": 0.8914, + "step": 557 + }, + { + "epoch": 1.3710073710073711, + "grad_norm": 1.4435566226659817, + "learning_rate": 4.964890132770543e-07, + "loss": 0.878, + "step": 558 + }, + { + "epoch": 1.3734643734643734, + "grad_norm": 1.8477167694746348, + "learning_rate": 4.929863217410767e-07, + "loss": 0.8909, + "step": 559 + }, + { + "epoch": 1.375921375921376, + "grad_norm": 1.4801649083461632, + "learning_rate": 4.894919830978212e-07, + "loss": 0.9079, + "step": 560 + }, + { + "epoch": 1.3783783783783785, + "grad_norm": 1.6811015996059142, + "learning_rate": 4.860060549154301e-07, + "loss": 0.815, + "step": 561 + }, + { + "epoch": 1.3808353808353808, + "grad_norm": 2.3145109947990155, + "learning_rate": 4.825285946234874e-07, + "loss": 0.9305, + "step": 562 + }, + { + "epoch": 1.3832923832923834, + "grad_norm": 1.4759252526730806, + "learning_rate": 4.790596595120698e-07, + "loss": 0.8589, + "step": 563 + }, + { + "epoch": 1.3857493857493859, + "grad_norm": 1.3378195425536465, + "learning_rate": 4.7559930673080475e-07, + "loss": 0.8659, + "step": 564 + }, + { + "epoch": 1.3882063882063882, + "grad_norm": 1.6393333264623262, + "learning_rate": 4.721475932879282e-07, + "loss": 0.8618, + "step": 565 + }, + { + "epoch": 1.3906633906633907, + "grad_norm": 1.6916138016252256, + "learning_rate": 4.6870457604934675e-07, + "loss": 0.9028, + "step": 566 + }, + { + "epoch": 1.393120393120393, + "grad_norm": 1.4225654726940884, + "learning_rate": 4.6527031173769857e-07, + "loss": 0.8765, + "step": 567 + }, + { + "epoch": 1.3955773955773956, + "grad_norm": 1.342379943145193, + "learning_rate": 4.618448569314206e-07, + "loss": 0.8855, + "step": 568 + }, + { + "epoch": 1.398034398034398, + "grad_norm": 2.442614220259753, + "learning_rate": 4.584282680638154e-07, + "loss": 0.8804, + "step": 569 + }, + { + "epoch": 1.4004914004914004, + "grad_norm": 1.5267874267053692, + "learning_rate": 4.550206014221232e-07, + "loss": 0.9668, + "step": 570 + }, + { + "epoch": 1.402948402948403, + "grad_norm": 1.8692131154426705, + "learning_rate": 4.5162191314659194e-07, + "loss": 0.8979, + "step": 571 + }, + { + "epoch": 1.4054054054054055, + "grad_norm": 2.022804110507779, + "learning_rate": 4.48232259229554e-07, + "loss": 0.8396, + "step": 572 + }, + { + "epoch": 1.4078624078624078, + "grad_norm": 1.444700494646088, + "learning_rate": 4.448516955145047e-07, + "loss": 0.8798, + "step": 573 + }, + { + "epoch": 1.4103194103194103, + "grad_norm": 1.5147425793812193, + "learning_rate": 4.414802776951798e-07, + "loss": 0.884, + "step": 574 + }, + { + "epoch": 1.4127764127764126, + "grad_norm": 1.4905275337775719, + "learning_rate": 4.3811806131463946e-07, + "loss": 0.8755, + "step": 575 + }, + { + "epoch": 1.4152334152334152, + "grad_norm": 1.5203731499114503, + "learning_rate": 4.347651017643539e-07, + "loss": 0.9089, + "step": 576 + }, + { + "epoch": 1.4176904176904177, + "grad_norm": 1.4792627181832774, + "learning_rate": 4.314214542832888e-07, + "loss": 0.8995, + "step": 577 + }, + { + "epoch": 1.42014742014742, + "grad_norm": 1.498952396682308, + "learning_rate": 4.280871739569971e-07, + "loss": 0.9, + "step": 578 + }, + { + "epoch": 1.4226044226044225, + "grad_norm": 1.5993242263021854, + "learning_rate": 4.247623157167102e-07, + "loss": 0.9004, + "step": 579 + }, + { + "epoch": 1.425061425061425, + "grad_norm": 1.2909684524904117, + "learning_rate": 4.214469343384346e-07, + "loss": 0.9442, + "step": 580 + }, + { + "epoch": 1.4275184275184274, + "grad_norm": 1.6601643291459118, + "learning_rate": 4.181410844420473e-07, + "loss": 0.8924, + "step": 581 + }, + { + "epoch": 1.42997542997543, + "grad_norm": 1.5456219866657583, + "learning_rate": 4.148448204903977e-07, + "loss": 0.8785, + "step": 582 + }, + { + "epoch": 1.4324324324324325, + "grad_norm": 2.3413636182110906, + "learning_rate": 4.115581967884093e-07, + "loss": 0.8867, + "step": 583 + }, + { + "epoch": 1.4348894348894348, + "grad_norm": 1.600188568539889, + "learning_rate": 4.082812674821865e-07, + "loss": 0.9163, + "step": 584 + }, + { + "epoch": 1.4373464373464373, + "grad_norm": 1.775885867839352, + "learning_rate": 4.0501408655812044e-07, + "loss": 0.94, + "step": 585 + }, + { + "epoch": 1.4398034398034398, + "grad_norm": 1.4853461496779028, + "learning_rate": 4.0175670784200066e-07, + "loss": 0.8828, + "step": 586 + }, + { + "epoch": 1.4422604422604421, + "grad_norm": 1.3437874918142632, + "learning_rate": 3.985091849981297e-07, + "loss": 0.906, + "step": 587 + }, + { + "epoch": 1.4447174447174447, + "grad_norm": 1.8022847542011342, + "learning_rate": 3.952715715284363e-07, + "loss": 0.9078, + "step": 588 + }, + { + "epoch": 1.4471744471744472, + "grad_norm": 1.4544558952772775, + "learning_rate": 3.9204392077159544e-07, + "loss": 0.9007, + "step": 589 + }, + { + "epoch": 1.4496314496314495, + "grad_norm": 1.5362982163739145, + "learning_rate": 3.888262859021507e-07, + "loss": 0.9329, + "step": 590 + }, + { + "epoch": 1.452088452088452, + "grad_norm": 1.35776640743268, + "learning_rate": 3.856187199296358e-07, + "loss": 0.9142, + "step": 591 + }, + { + "epoch": 1.4545454545454546, + "grad_norm": 1.552024766985016, + "learning_rate": 3.8242127569770265e-07, + "loss": 0.8624, + "step": 592 + }, + { + "epoch": 1.457002457002457, + "grad_norm": 1.6837813084688251, + "learning_rate": 3.792340058832515e-07, + "loss": 0.8515, + "step": 593 + }, + { + "epoch": 1.4594594594594594, + "grad_norm": 1.6200568897908438, + "learning_rate": 3.760569629955613e-07, + "loss": 0.9182, + "step": 594 + }, + { + "epoch": 1.461916461916462, + "grad_norm": 1.3613611600691327, + "learning_rate": 3.72890199375426e-07, + "loss": 0.9212, + "step": 595 + }, + { + "epoch": 1.4643734643734643, + "grad_norm": 1.5082364375002426, + "learning_rate": 3.697337671942913e-07, + "loss": 0.9084, + "step": 596 + }, + { + "epoch": 1.4668304668304668, + "grad_norm": 1.6876950054566262, + "learning_rate": 3.6658771845339676e-07, + "loss": 0.9065, + "step": 597 + }, + { + "epoch": 1.4692874692874693, + "grad_norm": 1.5733845918989662, + "learning_rate": 3.634521049829169e-07, + "loss": 0.8815, + "step": 598 + }, + { + "epoch": 1.4717444717444716, + "grad_norm": 1.8147232098247035, + "learning_rate": 3.603269784411089e-07, + "loss": 0.9133, + "step": 599 + }, + { + "epoch": 1.4742014742014742, + "grad_norm": 1.4595323005431404, + "learning_rate": 3.5721239031346063e-07, + "loss": 0.8071, + "step": 600 + }, + { + "epoch": 1.4766584766584767, + "grad_norm": 2.769727304993137, + "learning_rate": 3.541083919118438e-07, + "loss": 0.9423, + "step": 601 + }, + { + "epoch": 1.479115479115479, + "grad_norm": 1.610693658075675, + "learning_rate": 3.5101503437366675e-07, + "loss": 0.8946, + "step": 602 + }, + { + "epoch": 1.4815724815724816, + "grad_norm": 1.6365761390735136, + "learning_rate": 3.479323686610329e-07, + "loss": 0.9079, + "step": 603 + }, + { + "epoch": 1.484029484029484, + "grad_norm": 1.5083166362769518, + "learning_rate": 3.448604455599021e-07, + "loss": 0.8975, + "step": 604 + }, + { + "epoch": 1.4864864864864864, + "grad_norm": 1.60995686456512, + "learning_rate": 3.4179931567925214e-07, + "loss": 0.9498, + "step": 605 + }, + { + "epoch": 1.488943488943489, + "grad_norm": 1.7185280239304535, + "learning_rate": 3.387490294502456e-07, + "loss": 0.906, + "step": 606 + }, + { + "epoch": 1.4914004914004915, + "grad_norm": 1.321610762107921, + "learning_rate": 3.357096371254008e-07, + "loss": 0.8499, + "step": 607 + }, + { + "epoch": 1.4938574938574938, + "grad_norm": 1.2808013142695502, + "learning_rate": 3.3268118877776064e-07, + "loss": 0.8546, + "step": 608 + }, + { + "epoch": 1.4963144963144963, + "grad_norm": 1.3917333892709016, + "learning_rate": 3.296637343000704e-07, + "loss": 0.8776, + "step": 609 + }, + { + "epoch": 1.4987714987714988, + "grad_norm": 1.9760211104870224, + "learning_rate": 3.266573234039541e-07, + "loss": 0.8829, + "step": 610 + }, + { + "epoch": 1.5012285012285012, + "grad_norm": 1.9349300228421435, + "learning_rate": 3.2366200561909717e-07, + "loss": 0.9721, + "step": 611 + }, + { + "epoch": 1.5036855036855037, + "grad_norm": 1.5825092236415208, + "learning_rate": 3.2067783029242865e-07, + "loss": 0.8569, + "step": 612 + }, + { + "epoch": 1.5061425061425062, + "grad_norm": 1.7071421937295232, + "learning_rate": 3.1770484658730887e-07, + "loss": 0.9586, + "step": 613 + }, + { + "epoch": 1.5085995085995085, + "grad_norm": 1.3114653023022718, + "learning_rate": 3.147431034827208e-07, + "loss": 0.8879, + "step": 614 + }, + { + "epoch": 1.511056511056511, + "grad_norm": 1.3660758453242297, + "learning_rate": 3.11792649772461e-07, + "loss": 0.9227, + "step": 615 + }, + { + "epoch": 1.5135135135135136, + "grad_norm": 1.5443322628991898, + "learning_rate": 3.08853534064337e-07, + "loss": 0.8756, + "step": 616 + }, + { + "epoch": 1.515970515970516, + "grad_norm": 1.6205632640066696, + "learning_rate": 3.0592580477936604e-07, + "loss": 0.8631, + "step": 617 + }, + { + "epoch": 1.5184275184275184, + "grad_norm": 1.6750050731142352, + "learning_rate": 3.0300951015097863e-07, + "loss": 0.9308, + "step": 618 + }, + { + "epoch": 1.520884520884521, + "grad_norm": 1.7541208454820925, + "learning_rate": 3.001046982242216e-07, + "loss": 0.9097, + "step": 619 + }, + { + "epoch": 1.5233415233415233, + "grad_norm": 1.9317379819011606, + "learning_rate": 2.972114168549682e-07, + "loss": 0.9212, + "step": 620 + }, + { + "epoch": 1.5257985257985258, + "grad_norm": 1.2650696951039682, + "learning_rate": 2.9432971370913e-07, + "loss": 0.8836, + "step": 621 + }, + { + "epoch": 1.5282555282555284, + "grad_norm": 1.4960196905274168, + "learning_rate": 2.914596362618701e-07, + "loss": 0.8946, + "step": 622 + }, + { + "epoch": 1.5307125307125307, + "grad_norm": 1.507394873293028, + "learning_rate": 2.8860123179682236e-07, + "loss": 0.9141, + "step": 623 + }, + { + "epoch": 1.5331695331695332, + "grad_norm": 1.3335466599194608, + "learning_rate": 2.8575454740531115e-07, + "loss": 0.8667, + "step": 624 + }, + { + "epoch": 1.5356265356265357, + "grad_norm": 1.522699880941422, + "learning_rate": 2.829196299855775e-07, + "loss": 0.9901, + "step": 625 + }, + { + "epoch": 1.538083538083538, + "grad_norm": 1.4978237170833106, + "learning_rate": 2.800965262420043e-07, + "loss": 0.8367, + "step": 626 + }, + { + "epoch": 1.5405405405405406, + "grad_norm": 1.386657482706535, + "learning_rate": 2.7728528268434757e-07, + "loss": 0.8411, + "step": 627 + }, + { + "epoch": 1.542997542997543, + "grad_norm": 1.6884446408530738, + "learning_rate": 2.744859456269718e-07, + "loss": 0.942, + "step": 628 + }, + { + "epoch": 1.5454545454545454, + "grad_norm": 1.5125614736127029, + "learning_rate": 2.716985611880841e-07, + "loss": 0.8738, + "step": 629 + }, + { + "epoch": 1.547911547911548, + "grad_norm": 1.469704160523891, + "learning_rate": 2.689231752889768e-07, + "loss": 0.9165, + "step": 630 + }, + { + "epoch": 1.5503685503685505, + "grad_norm": 1.4870015368902652, + "learning_rate": 2.661598336532692e-07, + "loss": 0.8944, + "step": 631 + }, + { + "epoch": 1.5528255528255528, + "grad_norm": 1.4322815674497984, + "learning_rate": 2.6340858180615645e-07, + "loss": 0.8968, + "step": 632 + }, + { + "epoch": 1.5552825552825553, + "grad_norm": 1.3623692296259504, + "learning_rate": 2.6066946507365696e-07, + "loss": 0.8935, + "step": 633 + }, + { + "epoch": 1.5577395577395579, + "grad_norm": 1.484336106115407, + "learning_rate": 2.5794252858186706e-07, + "loss": 0.8993, + "step": 634 + }, + { + "epoch": 1.5601965601965602, + "grad_norm": 1.4403746721888513, + "learning_rate": 2.552278172562181e-07, + "loss": 0.8937, + "step": 635 + }, + { + "epoch": 1.5626535626535627, + "grad_norm": 1.8086294010381407, + "learning_rate": 2.525253758207353e-07, + "loss": 0.939, + "step": 636 + }, + { + "epoch": 1.5651105651105652, + "grad_norm": 1.4766926006894956, + "learning_rate": 2.4983524879730066e-07, + "loss": 0.9023, + "step": 637 + }, + { + "epoch": 1.5675675675675675, + "grad_norm": 1.3776357069750294, + "learning_rate": 2.471574805049206e-07, + "loss": 0.9233, + "step": 638 + }, + { + "epoch": 1.57002457002457, + "grad_norm": 1.6402577176804487, + "learning_rate": 2.444921150589957e-07, + "loss": 0.9319, + "step": 639 + }, + { + "epoch": 1.5724815724815726, + "grad_norm": 1.633491894306784, + "learning_rate": 2.418391963705927e-07, + "loss": 0.8818, + "step": 640 + }, + { + "epoch": 1.574938574938575, + "grad_norm": 1.5059260563661474, + "learning_rate": 2.391987681457219e-07, + "loss": 0.9138, + "step": 641 + }, + { + "epoch": 1.5773955773955772, + "grad_norm": 1.474101178051224, + "learning_rate": 2.3657087388461815e-07, + "loss": 0.8847, + "step": 642 + }, + { + "epoch": 1.57985257985258, + "grad_norm": 1.5606071866155247, + "learning_rate": 2.339555568810221e-07, + "loss": 0.9451, + "step": 643 + }, + { + "epoch": 1.5823095823095823, + "grad_norm": 1.5959144837669506, + "learning_rate": 2.3135286022146782e-07, + "loss": 0.9202, + "step": 644 + }, + { + "epoch": 1.5847665847665846, + "grad_norm": 1.4739665598722158, + "learning_rate": 2.2876282678457438e-07, + "loss": 0.8617, + "step": 645 + }, + { + "epoch": 1.5872235872235874, + "grad_norm": 1.324486221068965, + "learning_rate": 2.2618549924033704e-07, + "loss": 0.8702, + "step": 646 + }, + { + "epoch": 1.5896805896805897, + "grad_norm": 1.3762365282216713, + "learning_rate": 2.236209200494258e-07, + "loss": 0.9137, + "step": 647 + }, + { + "epoch": 1.592137592137592, + "grad_norm": 1.5081639071286088, + "learning_rate": 2.210691314624854e-07, + "loss": 0.8869, + "step": 648 + }, + { + "epoch": 1.5945945945945947, + "grad_norm": 1.5818213999124708, + "learning_rate": 2.1853017551943987e-07, + "loss": 0.9312, + "step": 649 + }, + { + "epoch": 1.597051597051597, + "grad_norm": 1.7757893297397498, + "learning_rate": 2.160040940487987e-07, + "loss": 0.8902, + "step": 650 + }, + { + "epoch": 1.5995085995085994, + "grad_norm": 1.3784061956740852, + "learning_rate": 2.1349092866696906e-07, + "loss": 0.8984, + "step": 651 + }, + { + "epoch": 1.6019656019656021, + "grad_norm": 1.4232738419586703, + "learning_rate": 2.1099072077756863e-07, + "loss": 0.8676, + "step": 652 + }, + { + "epoch": 1.6044226044226044, + "grad_norm": 3.799736709775996, + "learning_rate": 2.0850351157074597e-07, + "loss": 0.9107, + "step": 653 + }, + { + "epoch": 1.6068796068796067, + "grad_norm": 2.8003735792292055, + "learning_rate": 2.060293420224991e-07, + "loss": 0.8836, + "step": 654 + }, + { + "epoch": 1.6093366093366095, + "grad_norm": 1.320688739165864, + "learning_rate": 2.0356825289400183e-07, + "loss": 0.8912, + "step": 655 + }, + { + "epoch": 1.6117936117936118, + "grad_norm": 1.9634181459213618, + "learning_rate": 2.011202847309329e-07, + "loss": 0.9238, + "step": 656 + }, + { + "epoch": 1.6142506142506141, + "grad_norm": 1.5624806988822288, + "learning_rate": 1.9868547786280665e-07, + "loss": 0.8707, + "step": 657 + }, + { + "epoch": 1.6167076167076169, + "grad_norm": 1.4107049261598619, + "learning_rate": 1.9626387240230891e-07, + "loss": 0.868, + "step": 658 + }, + { + "epoch": 1.6191646191646192, + "grad_norm": 1.492505632847286, + "learning_rate": 1.9385550824463725e-07, + "loss": 0.8894, + "step": 659 + }, + { + "epoch": 1.6216216216216215, + "grad_norm": 1.4382690869482055, + "learning_rate": 1.9146042506684223e-07, + "loss": 0.8978, + "step": 660 + }, + { + "epoch": 1.6240786240786242, + "grad_norm": 1.627212055293953, + "learning_rate": 1.890786623271743e-07, + "loss": 0.871, + "step": 661 + }, + { + "epoch": 1.6265356265356266, + "grad_norm": 1.6843094898216577, + "learning_rate": 1.8671025926443463e-07, + "loss": 0.8952, + "step": 662 + }, + { + "epoch": 1.6289926289926289, + "grad_norm": 1.6654696363977666, + "learning_rate": 1.8435525489732718e-07, + "loss": 0.8591, + "step": 663 + }, + { + "epoch": 1.6314496314496314, + "grad_norm": 1.5188779880666268, + "learning_rate": 1.8201368802381657e-07, + "loss": 0.9323, + "step": 664 + }, + { + "epoch": 1.633906633906634, + "grad_norm": 1.3399754864662434, + "learning_rate": 1.7968559722048903e-07, + "loss": 0.8543, + "step": 665 + }, + { + "epoch": 1.6363636363636362, + "grad_norm": 1.3715522224138303, + "learning_rate": 1.7737102084191703e-07, + "loss": 0.918, + "step": 666 + }, + { + "epoch": 1.6388206388206388, + "grad_norm": 1.6259531621746015, + "learning_rate": 1.7506999702002678e-07, + "loss": 0.8694, + "step": 667 + }, + { + "epoch": 1.6412776412776413, + "grad_norm": 2.365813450212486, + "learning_rate": 1.7278256366347032e-07, + "loss": 0.8874, + "step": 668 + }, + { + "epoch": 1.6437346437346436, + "grad_norm": 1.7862394193970157, + "learning_rate": 1.7050875845700064e-07, + "loss": 0.8803, + "step": 669 + }, + { + "epoch": 1.6461916461916462, + "grad_norm": 1.557725205467593, + "learning_rate": 1.682486188608523e-07, + "loss": 0.8842, + "step": 670 + }, + { + "epoch": 1.6486486486486487, + "grad_norm": 1.6097866192991963, + "learning_rate": 1.6600218211012218e-07, + "loss": 0.9428, + "step": 671 + }, + { + "epoch": 1.651105651105651, + "grad_norm": 1.4644367524635458, + "learning_rate": 1.6376948521415711e-07, + "loss": 0.9066, + "step": 672 + }, + { + "epoch": 1.6535626535626535, + "grad_norm": 1.6645946570893102, + "learning_rate": 1.6155056495594465e-07, + "loss": 0.905, + "step": 673 + }, + { + "epoch": 1.656019656019656, + "grad_norm": 1.3397486338053894, + "learning_rate": 1.5934545789150622e-07, + "loss": 0.8785, + "step": 674 + }, + { + "epoch": 1.6584766584766584, + "grad_norm": 1.3685689789923863, + "learning_rate": 1.5715420034929473e-07, + "loss": 0.9179, + "step": 675 + }, + { + "epoch": 1.660933660933661, + "grad_norm": 1.5445547208764443, + "learning_rate": 1.5497682842959747e-07, + "loss": 0.9187, + "step": 676 + }, + { + "epoch": 1.6633906633906634, + "grad_norm": 1.7915250554930906, + "learning_rate": 1.5281337800393968e-07, + "loss": 0.8996, + "step": 677 + }, + { + "epoch": 1.6658476658476657, + "grad_norm": 1.425206424154818, + "learning_rate": 1.5066388471449453e-07, + "loss": 0.9244, + "step": 678 + }, + { + "epoch": 1.6683046683046683, + "grad_norm": 1.3711937617299463, + "learning_rate": 1.4852838397349542e-07, + "loss": 0.9239, + "step": 679 + }, + { + "epoch": 1.6707616707616708, + "grad_norm": 1.4206125252604587, + "learning_rate": 1.4640691096265357e-07, + "loss": 0.9127, + "step": 680 + }, + { + "epoch": 1.6732186732186731, + "grad_norm": 1.5824939121397916, + "learning_rate": 1.4429950063257712e-07, + "loss": 0.9243, + "step": 681 + }, + { + "epoch": 1.6756756756756757, + "grad_norm": 1.5523209337418313, + "learning_rate": 1.4220618770219604e-07, + "loss": 0.8743, + "step": 682 + }, + { + "epoch": 1.6781326781326782, + "grad_norm": 2.449668270011937, + "learning_rate": 1.4012700665818988e-07, + "loss": 0.9022, + "step": 683 + }, + { + "epoch": 1.6805896805896805, + "grad_norm": 2.797795087084986, + "learning_rate": 1.3806199175442046e-07, + "loss": 0.9127, + "step": 684 + }, + { + "epoch": 1.683046683046683, + "grad_norm": 1.594299579779607, + "learning_rate": 1.3601117701136588e-07, + "loss": 0.8958, + "step": 685 + }, + { + "epoch": 1.6855036855036856, + "grad_norm": 1.3744841196118298, + "learning_rate": 1.3397459621556128e-07, + "loss": 0.8847, + "step": 686 + }, + { + "epoch": 1.6879606879606879, + "grad_norm": 1.4967014666487075, + "learning_rate": 1.319522829190427e-07, + "loss": 0.9375, + "step": 687 + }, + { + "epoch": 1.6904176904176904, + "grad_norm": 1.257226409390929, + "learning_rate": 1.2994427043879275e-07, + "loss": 0.87, + "step": 688 + }, + { + "epoch": 1.692874692874693, + "grad_norm": 1.5489462493350459, + "learning_rate": 1.279505918561923e-07, + "loss": 0.8997, + "step": 689 + }, + { + "epoch": 1.6953316953316953, + "grad_norm": 1.478167106209186, + "learning_rate": 1.2597128001647695e-07, + "loss": 0.88, + "step": 690 + }, + { + "epoch": 1.6977886977886978, + "grad_norm": 1.530749437315356, + "learning_rate": 1.2400636752819337e-07, + "loss": 0.9155, + "step": 691 + }, + { + "epoch": 1.7002457002457003, + "grad_norm": 1.312749085771371, + "learning_rate": 1.2205588676266386e-07, + "loss": 0.9148, + "step": 692 + }, + { + "epoch": 1.7027027027027026, + "grad_norm": 1.2944478484019626, + "learning_rate": 1.2011986985345313e-07, + "loss": 0.9421, + "step": 693 + }, + { + "epoch": 1.7051597051597052, + "grad_norm": 1.9652193892846588, + "learning_rate": 1.1819834869583767e-07, + "loss": 0.8827, + "step": 694 + }, + { + "epoch": 1.7076167076167077, + "grad_norm": 1.4459180494396398, + "learning_rate": 1.1629135494628095e-07, + "loss": 0.9373, + "step": 695 + }, + { + "epoch": 1.71007371007371, + "grad_norm": 1.4835720268158477, + "learning_rate": 1.1439892002191231e-07, + "loss": 0.8919, + "step": 696 + }, + { + "epoch": 1.7125307125307125, + "grad_norm": 1.8195889649889634, + "learning_rate": 1.1252107510000841e-07, + "loss": 0.8234, + "step": 697 + }, + { + "epoch": 1.714987714987715, + "grad_norm": 1.3885033939390314, + "learning_rate": 1.1065785111748117e-07, + "loss": 0.8727, + "step": 698 + }, + { + "epoch": 1.7174447174447174, + "grad_norm": 1.8847414217621874, + "learning_rate": 1.0880927877036606e-07, + "loss": 0.9219, + "step": 699 + }, + { + "epoch": 1.71990171990172, + "grad_norm": 1.4570728152695678, + "learning_rate": 1.0697538851331767e-07, + "loss": 0.8921, + "step": 700 + }, + { + "epoch": 1.7223587223587224, + "grad_norm": 1.6297111465619418, + "learning_rate": 1.0515621055910817e-07, + "loss": 0.8523, + "step": 701 + }, + { + "epoch": 1.7248157248157248, + "grad_norm": 1.3305859076255666, + "learning_rate": 1.0335177487812862e-07, + "loss": 0.8873, + "step": 702 + }, + { + "epoch": 1.7272727272727273, + "grad_norm": 1.4781364488325777, + "learning_rate": 1.0156211119789581e-07, + "loss": 0.9512, + "step": 703 + }, + { + "epoch": 1.7297297297297298, + "grad_norm": 1.3860356549035335, + "learning_rate": 9.978724900256264e-08, + "loss": 0.816, + "step": 704 + }, + { + "epoch": 1.7321867321867321, + "grad_norm": 1.8635382664585312, + "learning_rate": 9.802721753243193e-08, + "loss": 0.8981, + "step": 705 + }, + { + "epoch": 1.7346437346437347, + "grad_norm": 1.7250194554756548, + "learning_rate": 9.628204578347466e-08, + "loss": 0.8479, + "step": 706 + }, + { + "epoch": 1.7371007371007372, + "grad_norm": 1.697724825854203, + "learning_rate": 9.455176250685337e-08, + "loss": 0.8986, + "step": 707 + }, + { + "epoch": 1.7395577395577395, + "grad_norm": 1.6181977387870101, + "learning_rate": 9.283639620844686e-08, + "loss": 0.8641, + "step": 708 + }, + { + "epoch": 1.742014742014742, + "grad_norm": 1.318237926606282, + "learning_rate": 9.113597514838134e-08, + "loss": 0.9309, + "step": 709 + }, + { + "epoch": 1.7444717444717446, + "grad_norm": 2.254843209174964, + "learning_rate": 8.94505273405658e-08, + "loss": 0.94, + "step": 710 + }, + { + "epoch": 1.746928746928747, + "grad_norm": 1.362243307286928, + "learning_rate": 8.778008055222807e-08, + "loss": 0.8757, + "step": 711 + }, + { + "epoch": 1.7493857493857494, + "grad_norm": 1.6314398152119172, + "learning_rate": 8.612466230346016e-08, + "loss": 0.8814, + "step": 712 + }, + { + "epoch": 1.751842751842752, + "grad_norm": 1.754218537191555, + "learning_rate": 8.448429986676298e-08, + "loss": 0.9166, + "step": 713 + }, + { + "epoch": 1.7542997542997543, + "grad_norm": 1.4911185442079653, + "learning_rate": 8.285902026659741e-08, + "loss": 0.9213, + "step": 714 + }, + { + "epoch": 1.7567567567567568, + "grad_norm": 1.4620109120308156, + "learning_rate": 8.124885027894013e-08, + "loss": 0.9285, + "step": 715 + }, + { + "epoch": 1.7592137592137593, + "grad_norm": 1.5197712931486254, + "learning_rate": 7.965381643084068e-08, + "loss": 0.8578, + "step": 716 + }, + { + "epoch": 1.7616707616707616, + "grad_norm": 1.3554180543190764, + "learning_rate": 7.807394499998576e-08, + "loss": 0.8945, + "step": 717 + }, + { + "epoch": 1.7641277641277642, + "grad_norm": 1.7074271145927862, + "learning_rate": 7.650926201426634e-08, + "loss": 0.9014, + "step": 718 + }, + { + "epoch": 1.7665847665847667, + "grad_norm": 1.7305821717380618, + "learning_rate": 7.495979325134805e-08, + "loss": 0.9427, + "step": 719 + }, + { + "epoch": 1.769041769041769, + "grad_norm": 1.6549796883627756, + "learning_rate": 7.342556423824675e-08, + "loss": 0.8929, + "step": 720 + }, + { + "epoch": 1.7714987714987716, + "grad_norm": 1.693588021942915, + "learning_rate": 7.19066002509089e-08, + "loss": 0.9149, + "step": 721 + }, + { + "epoch": 1.773955773955774, + "grad_norm": 1.3883677751773678, + "learning_rate": 7.040292631379386e-08, + "loss": 0.9379, + "step": 722 + }, + { + "epoch": 1.7764127764127764, + "grad_norm": 1.415233971009828, + "learning_rate": 6.891456719946187e-08, + "loss": 0.8786, + "step": 723 + }, + { + "epoch": 1.7788697788697787, + "grad_norm": 1.3887984611992066, + "learning_rate": 6.744154742816721e-08, + "loss": 0.8809, + "step": 724 + }, + { + "epoch": 1.7813267813267815, + "grad_norm": 1.4599586632095534, + "learning_rate": 6.598389126745207e-08, + "loss": 0.8906, + "step": 725 + }, + { + "epoch": 1.7837837837837838, + "grad_norm": 1.3618867212694985, + "learning_rate": 6.454162273174879e-08, + "loss": 0.8875, + "step": 726 + }, + { + "epoch": 1.786240786240786, + "grad_norm": 1.616703409825558, + "learning_rate": 6.311476558198337e-08, + "loss": 0.9186, + "step": 727 + }, + { + "epoch": 1.7886977886977888, + "grad_norm": 1.5643574404054077, + "learning_rate": 6.170334332518323e-08, + "loss": 0.8191, + "step": 728 + }, + { + "epoch": 1.7911547911547911, + "grad_norm": 1.6547417394218757, + "learning_rate": 6.030737921409168e-08, + "loss": 0.9015, + "step": 729 + }, + { + "epoch": 1.7936117936117935, + "grad_norm": 1.3258614962104087, + "learning_rate": 5.892689624678326e-08, + "loss": 0.9003, + "step": 730 + }, + { + "epoch": 1.7960687960687962, + "grad_norm": 1.4886792779962632, + "learning_rate": 5.756191716628556e-08, + "loss": 0.9197, + "step": 731 + }, + { + "epoch": 1.7985257985257985, + "grad_norm": 1.4228121623738594, + "learning_rate": 5.621246446020489e-08, + "loss": 0.9151, + "step": 732 + }, + { + "epoch": 1.8009828009828008, + "grad_norm": 1.4694783098206843, + "learning_rate": 5.487856036035488e-08, + "loss": 0.9478, + "step": 733 + }, + { + "epoch": 1.8034398034398036, + "grad_norm": 1.6835863076974373, + "learning_rate": 5.356022684239059e-08, + "loss": 0.8852, + "step": 734 + }, + { + "epoch": 1.805896805896806, + "grad_norm": 1.468678481999939, + "learning_rate": 5.22574856254474e-08, + "loss": 0.9227, + "step": 735 + }, + { + "epoch": 1.8083538083538082, + "grad_norm": 1.664873807912385, + "learning_rate": 5.0970358171781525e-08, + "loss": 0.9709, + "step": 736 + }, + { + "epoch": 1.810810810810811, + "grad_norm": 1.4869146515566096, + "learning_rate": 4.969886568641757e-08, + "loss": 0.8627, + "step": 737 + }, + { + "epoch": 1.8132678132678133, + "grad_norm": 1.5302504751076207, + "learning_rate": 4.844302911679909e-08, + "loss": 0.8694, + "step": 738 + }, + { + "epoch": 1.8157248157248156, + "grad_norm": 1.3703447874730004, + "learning_rate": 4.720286915244309e-08, + "loss": 0.8388, + "step": 739 + }, + { + "epoch": 1.8181818181818183, + "grad_norm": 1.8909215620115458, + "learning_rate": 4.597840622459936e-08, + "loss": 0.8697, + "step": 740 + }, + { + "epoch": 1.8206388206388207, + "grad_norm": 1.8396393900066417, + "learning_rate": 4.476966050591413e-08, + "loss": 0.9238, + "step": 741 + }, + { + "epoch": 1.823095823095823, + "grad_norm": 1.4266543724530107, + "learning_rate": 4.3576651910097076e-08, + "loss": 0.8901, + "step": 742 + }, + { + "epoch": 1.8255528255528255, + "grad_norm": 1.3873776301248721, + "learning_rate": 4.239940009159415e-08, + "loss": 0.904, + "step": 743 + }, + { + "epoch": 1.828009828009828, + "grad_norm": 1.554746782909149, + "learning_rate": 4.1237924445262864e-08, + "loss": 0.8605, + "step": 744 + }, + { + "epoch": 1.8304668304668303, + "grad_norm": 1.5101859068421868, + "learning_rate": 4.0092244106053386e-08, + "loss": 0.8981, + "step": 745 + }, + { + "epoch": 1.8329238329238329, + "grad_norm": 1.559988660277713, + "learning_rate": 3.896237794869339e-08, + "loss": 0.8617, + "step": 746 + }, + { + "epoch": 1.8353808353808354, + "grad_norm": 1.7088654168690152, + "learning_rate": 3.7848344587376294e-08, + "loss": 0.8696, + "step": 747 + }, + { + "epoch": 1.8378378378378377, + "grad_norm": 1.6151695281653022, + "learning_rate": 3.675016237545536e-08, + "loss": 0.9057, + "step": 748 + }, + { + "epoch": 1.8402948402948403, + "grad_norm": 1.6260854298806076, + "learning_rate": 3.566784940514145e-08, + "loss": 0.8715, + "step": 749 + }, + { + "epoch": 1.8427518427518428, + "grad_norm": 1.5483874580947579, + "learning_rate": 3.460142350720396e-08, + "loss": 0.8928, + "step": 750 + }, + { + "epoch": 1.845208845208845, + "grad_norm": 1.3856178647115844, + "learning_rate": 3.3550902250677914e-08, + "loss": 0.8898, + "step": 751 + }, + { + "epoch": 1.8476658476658476, + "grad_norm": 1.6217388770422025, + "learning_rate": 3.251630294257479e-08, + "loss": 0.8887, + "step": 752 + }, + { + "epoch": 1.8501228501228502, + "grad_norm": 1.7461700477068736, + "learning_rate": 3.1497642627596245e-08, + "loss": 0.9144, + "step": 753 + }, + { + "epoch": 1.8525798525798525, + "grad_norm": 1.3204466111990083, + "learning_rate": 3.0494938087854306e-08, + "loss": 0.9046, + "step": 754 + }, + { + "epoch": 1.855036855036855, + "grad_norm": 1.5311966806895796, + "learning_rate": 2.9508205842594724e-08, + "loss": 0.9614, + "step": 755 + }, + { + "epoch": 1.8574938574938575, + "grad_norm": 1.5618296845089266, + "learning_rate": 2.8537462147924184e-08, + "loss": 0.9499, + "step": 756 + }, + { + "epoch": 1.8599508599508598, + "grad_norm": 1.506551170833026, + "learning_rate": 2.7582722996543405e-08, + "loss": 0.8963, + "step": 757 + }, + { + "epoch": 1.8624078624078624, + "grad_norm": 1.6976856223987622, + "learning_rate": 2.6644004117483355e-08, + "loss": 0.938, + "step": 758 + }, + { + "epoch": 1.864864864864865, + "grad_norm": 1.7152873001996989, + "learning_rate": 2.572132097584523e-08, + "loss": 0.9302, + "step": 759 + }, + { + "epoch": 1.8673218673218672, + "grad_norm": 1.4747780984909165, + "learning_rate": 2.481468877254722e-08, + "loss": 0.8927, + "step": 760 + }, + { + "epoch": 1.8697788697788698, + "grad_norm": 1.6455964184060006, + "learning_rate": 2.3924122444072936e-08, + "loss": 0.92, + "step": 761 + }, + { + "epoch": 1.8722358722358723, + "grad_norm": 1.4135913992521028, + "learning_rate": 2.3049636662225702e-08, + "loss": 0.9095, + "step": 762 + }, + { + "epoch": 1.8746928746928746, + "grad_norm": 1.3283327210815499, + "learning_rate": 2.2191245833886985e-08, + "loss": 0.923, + "step": 763 + }, + { + "epoch": 1.8771498771498771, + "grad_norm": 2.1206540317643316, + "learning_rate": 2.134896410077891e-08, + "loss": 0.8759, + "step": 764 + }, + { + "epoch": 1.8796068796068797, + "grad_norm": 1.4686301829717907, + "learning_rate": 2.052280533923101e-08, + "loss": 0.91, + "step": 765 + }, + { + "epoch": 1.882063882063882, + "grad_norm": 2.2833412802518835, + "learning_rate": 1.9712783159952063e-08, + "loss": 0.8681, + "step": 766 + }, + { + "epoch": 1.8845208845208845, + "grad_norm": 1.5664189962805568, + "learning_rate": 1.8918910907805728e-08, + "loss": 0.8958, + "step": 767 + }, + { + "epoch": 1.886977886977887, + "grad_norm": 1.3597898472460088, + "learning_rate": 1.81412016615905e-08, + "loss": 0.8846, + "step": 768 + }, + { + "epoch": 1.8894348894348894, + "grad_norm": 1.794391013663922, + "learning_rate": 1.737966823382442e-08, + "loss": 0.8704, + "step": 769 + }, + { + "epoch": 1.8918918918918919, + "grad_norm": 1.3876663428773923, + "learning_rate": 1.6634323170533926e-08, + "loss": 0.8826, + "step": 770 + }, + { + "epoch": 1.8943488943488944, + "grad_norm": 1.964992745747424, + "learning_rate": 1.5905178751047134e-08, + "loss": 0.9233, + "step": 771 + }, + { + "epoch": 1.8968058968058967, + "grad_norm": 1.5474022069405389, + "learning_rate": 1.519224698779198e-08, + "loss": 0.9574, + "step": 772 + }, + { + "epoch": 1.8992628992628993, + "grad_norm": 1.367308795693324, + "learning_rate": 1.4495539626097287e-08, + "loss": 0.8783, + "step": 773 + }, + { + "epoch": 1.9017199017199018, + "grad_norm": 1.4446894160381547, + "learning_rate": 1.3815068144000352e-08, + "loss": 0.9278, + "step": 774 + }, + { + "epoch": 1.904176904176904, + "grad_norm": 1.4909159417858395, + "learning_rate": 1.3150843752057438e-08, + "loss": 0.8681, + "step": 775 + }, + { + "epoch": 1.9066339066339066, + "grad_norm": 1.358589936876708, + "learning_rate": 1.2502877393158585e-08, + "loss": 0.8662, + "step": 776 + }, + { + "epoch": 1.9090909090909092, + "grad_norm": 1.4371079365419295, + "learning_rate": 1.1871179742348413e-08, + "loss": 0.8844, + "step": 777 + }, + { + "epoch": 1.9115479115479115, + "grad_norm": 1.5058059815383584, + "learning_rate": 1.1255761206649061e-08, + "loss": 0.8885, + "step": 778 + }, + { + "epoch": 1.914004914004914, + "grad_norm": 1.4300739849238244, + "learning_rate": 1.0656631924889748e-08, + "loss": 0.8997, + "step": 779 + }, + { + "epoch": 1.9164619164619165, + "grad_norm": 1.5658510974559385, + "learning_rate": 1.0073801767539247e-08, + "loss": 0.8983, + "step": 780 + }, + { + "epoch": 1.9189189189189189, + "grad_norm": 1.4965427543973944, + "learning_rate": 9.507280336542911e-09, + "loss": 0.8676, + "step": 781 + }, + { + "epoch": 1.9213759213759214, + "grad_norm": 1.36581135793563, + "learning_rate": 8.957076965165234e-09, + "loss": 0.8792, + "step": 782 + }, + { + "epoch": 1.923832923832924, + "grad_norm": 1.5156402618603095, + "learning_rate": 8.423200717835976e-09, + "loss": 0.8626, + "step": 783 + }, + { + "epoch": 1.9262899262899262, + "grad_norm": 1.3074563333752685, + "learning_rate": 7.905660390000069e-09, + "loss": 0.9315, + "step": 784 + }, + { + "epoch": 1.9287469287469288, + "grad_norm": 1.7609221141327198, + "learning_rate": 7.404464507973607e-09, + "loss": 0.8862, + "step": 785 + }, + { + "epoch": 1.9312039312039313, + "grad_norm": 1.5426817574716647, + "learning_rate": 6.919621328802972e-09, + "loss": 0.9543, + "step": 786 + }, + { + "epoch": 1.9336609336609336, + "grad_norm": 2.3064282757005663, + "learning_rate": 6.4511388401286e-09, + "loss": 0.9583, + "step": 787 + }, + { + "epoch": 1.9361179361179361, + "grad_norm": 1.67978354162131, + "learning_rate": 5.999024760054094e-09, + "loss": 0.9364, + "step": 788 + }, + { + "epoch": 1.9385749385749387, + "grad_norm": 2.269285024101559, + "learning_rate": 5.563286537018319e-09, + "loss": 0.858, + "step": 789 + }, + { + "epoch": 1.941031941031941, + "grad_norm": 1.7349413650625205, + "learning_rate": 5.143931349672837e-09, + "loss": 0.9019, + "step": 790 + }, + { + "epoch": 1.9434889434889435, + "grad_norm": 1.652477600280537, + "learning_rate": 4.7409661067642215e-09, + "loss": 0.9325, + "step": 791 + }, + { + "epoch": 1.945945945945946, + "grad_norm": 1.354819301610776, + "learning_rate": 4.354397447019375e-09, + "loss": 0.9047, + "step": 792 + }, + { + "epoch": 1.9484029484029484, + "grad_norm": 1.615707852796759, + "learning_rate": 3.9842317390369474e-09, + "loss": 0.9033, + "step": 793 + }, + { + "epoch": 1.950859950859951, + "grad_norm": 1.4784764684056093, + "learning_rate": 3.630475081181861e-09, + "loss": 0.8419, + "step": 794 + }, + { + "epoch": 1.9533169533169534, + "grad_norm": 1.316270670020737, + "learning_rate": 3.2931333014850626e-09, + "loss": 0.8478, + "step": 795 + }, + { + "epoch": 1.9557739557739557, + "grad_norm": 1.4172812525176979, + "learning_rate": 2.972211957547377e-09, + "loss": 0.8208, + "step": 796 + }, + { + "epoch": 1.9582309582309583, + "grad_norm": 1.6364175608068232, + "learning_rate": 2.667716336448356e-09, + "loss": 0.9062, + "step": 797 + }, + { + "epoch": 1.9606879606879608, + "grad_norm": 1.6122614160580477, + "learning_rate": 2.379651454658571e-09, + "loss": 0.8939, + "step": 798 + }, + { + "epoch": 1.9631449631449631, + "grad_norm": 1.5173118240679087, + "learning_rate": 2.108022057957348e-09, + "loss": 0.8899, + "step": 799 + }, + { + "epoch": 1.9656019656019657, + "grad_norm": 1.4032962894445904, + "learning_rate": 1.8528326213548273e-09, + "loss": 0.9269, + "step": 800 + }, + { + "epoch": 1.9680589680589682, + "grad_norm": 1.491232169954221, + "learning_rate": 1.6140873490172457e-09, + "loss": 0.887, + "step": 801 + }, + { + "epoch": 1.9705159705159705, + "grad_norm": 1.452706323624227, + "learning_rate": 1.3917901741989923e-09, + "loss": 0.9033, + "step": 802 + }, + { + "epoch": 1.972972972972973, + "grad_norm": 2.9039670137491176, + "learning_rate": 1.1859447591769934e-09, + "loss": 0.8912, + "step": 803 + }, + { + "epoch": 1.9754299754299756, + "grad_norm": 1.426897627539779, + "learning_rate": 9.96554495190205e-10, + "loss": 0.8626, + "step": 804 + }, + { + "epoch": 1.9778869778869779, + "grad_norm": 1.333240805775527, + "learning_rate": 8.236225023844357e-10, + "loss": 0.8773, + "step": 805 + }, + { + "epoch": 1.9803439803439802, + "grad_norm": 1.3907079742810688, + "learning_rate": 6.671516297606094e-10, + "loss": 0.9252, + "step": 806 + }, + { + "epoch": 1.982800982800983, + "grad_norm": 1.698786893046404, + "learning_rate": 5.271444551276927e-10, + "loss": 0.9504, + "step": 807 + }, + { + "epoch": 1.9852579852579852, + "grad_norm": 1.8549325727891992, + "learning_rate": 4.0360328506017226e-10, + "loss": 0.8953, + "step": 808 + }, + { + "epoch": 1.9877149877149876, + "grad_norm": 1.4404540052198962, + "learning_rate": 2.965301548606414e-10, + "loss": 0.8946, + "step": 809 + }, + { + "epoch": 1.9901719901719903, + "grad_norm": 1.2420916620768656, + "learning_rate": 2.0592682852549338e-10, + "loss": 0.8543, + "step": 810 + }, + { + "epoch": 1.9926289926289926, + "grad_norm": 1.5902830913290869, + "learning_rate": 1.317947987163892e-10, + "loss": 0.9287, + "step": 811 + }, + { + "epoch": 1.995085995085995, + "grad_norm": 1.7356553731894724, + "learning_rate": 7.41352867354994e-11, + "loss": 0.9318, + "step": 812 + }, + { + "epoch": 1.9975429975429977, + "grad_norm": 1.3632994827873626, + "learning_rate": 3.294924250540898e-11, + "loss": 0.9022, + "step": 813 + }, + { + "epoch": 2.0, + "grad_norm": 1.5450655192747615, + "learning_rate": 8.237344553241321e-12, + "loss": 0.9407, + "step": 814 + } + ], + "logging_steps": 1, + "max_steps": 814, + "num_input_tokens_seen": 0, + "num_train_epochs": 2, + "save_steps": 407, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 8.602630032143155e+17, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}