zephyr-7b-sft-qlora / trainer_state.json
Aa123564's picture
Model save
d7019e6 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9868043602983362,
"eval_steps": 500,
"global_step": 4300,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.00022948938611589215,
"grad_norm": 0.2832958400249481,
"learning_rate": 4.587155963302753e-07,
"loss": 1.1374,
"step": 1
},
{
"epoch": 0.0011474469305794606,
"grad_norm": 0.3270454704761505,
"learning_rate": 2.2935779816513764e-06,
"loss": 1.1705,
"step": 5
},
{
"epoch": 0.002294893861158921,
"grad_norm": 0.27594515681266785,
"learning_rate": 4.587155963302753e-06,
"loss": 1.1193,
"step": 10
},
{
"epoch": 0.0034423407917383822,
"grad_norm": 0.2871370315551758,
"learning_rate": 6.880733944954129e-06,
"loss": 1.1392,
"step": 15
},
{
"epoch": 0.004589787722317842,
"grad_norm": 0.2703236937522888,
"learning_rate": 9.174311926605506e-06,
"loss": 1.1746,
"step": 20
},
{
"epoch": 0.005737234652897304,
"grad_norm": 0.2527748644351959,
"learning_rate": 1.1467889908256882e-05,
"loss": 1.129,
"step": 25
},
{
"epoch": 0.0068846815834767644,
"grad_norm": 0.1984025090932846,
"learning_rate": 1.3761467889908258e-05,
"loss": 1.0819,
"step": 30
},
{
"epoch": 0.008032128514056224,
"grad_norm": 0.21797247231006622,
"learning_rate": 1.6055045871559634e-05,
"loss": 1.0604,
"step": 35
},
{
"epoch": 0.009179575444635685,
"grad_norm": 0.1990118771791458,
"learning_rate": 1.834862385321101e-05,
"loss": 1.0285,
"step": 40
},
{
"epoch": 0.010327022375215147,
"grad_norm": 0.19321753084659576,
"learning_rate": 2.0642201834862388e-05,
"loss": 1.0389,
"step": 45
},
{
"epoch": 0.011474469305794608,
"grad_norm": 0.18540368974208832,
"learning_rate": 2.2935779816513765e-05,
"loss": 1.0454,
"step": 50
},
{
"epoch": 0.012621916236374068,
"grad_norm": 0.18497976660728455,
"learning_rate": 2.5229357798165138e-05,
"loss": 1.0578,
"step": 55
},
{
"epoch": 0.013769363166953529,
"grad_norm": 0.14606232941150665,
"learning_rate": 2.7522935779816515e-05,
"loss": 1.064,
"step": 60
},
{
"epoch": 0.01491681009753299,
"grad_norm": 0.1588371992111206,
"learning_rate": 2.9816513761467892e-05,
"loss": 1.0231,
"step": 65
},
{
"epoch": 0.01606425702811245,
"grad_norm": 0.1489713490009308,
"learning_rate": 3.211009174311927e-05,
"loss": 0.9939,
"step": 70
},
{
"epoch": 0.01721170395869191,
"grad_norm": 0.1601668745279312,
"learning_rate": 3.4403669724770645e-05,
"loss": 1.04,
"step": 75
},
{
"epoch": 0.01835915088927137,
"grad_norm": 0.17821019887924194,
"learning_rate": 3.669724770642202e-05,
"loss": 1.0147,
"step": 80
},
{
"epoch": 0.019506597819850834,
"grad_norm": 0.15738487243652344,
"learning_rate": 3.89908256880734e-05,
"loss": 1.0404,
"step": 85
},
{
"epoch": 0.020654044750430294,
"grad_norm": 0.16155536472797394,
"learning_rate": 4.1284403669724776e-05,
"loss": 1.0388,
"step": 90
},
{
"epoch": 0.021801491681009755,
"grad_norm": 0.1788385957479477,
"learning_rate": 4.3577981651376146e-05,
"loss": 0.9844,
"step": 95
},
{
"epoch": 0.022948938611589215,
"grad_norm": 0.2024046927690506,
"learning_rate": 4.587155963302753e-05,
"loss": 0.9918,
"step": 100
},
{
"epoch": 0.024096385542168676,
"grad_norm": 0.16764964163303375,
"learning_rate": 4.81651376146789e-05,
"loss": 0.9971,
"step": 105
},
{
"epoch": 0.025243832472748137,
"grad_norm": 0.17267030477523804,
"learning_rate": 5.0458715596330276e-05,
"loss": 1.0152,
"step": 110
},
{
"epoch": 0.026391279403327597,
"grad_norm": 0.1706738919019699,
"learning_rate": 5.2752293577981646e-05,
"loss": 1.0135,
"step": 115
},
{
"epoch": 0.027538726333907058,
"grad_norm": 0.1741032898426056,
"learning_rate": 5.504587155963303e-05,
"loss": 0.9995,
"step": 120
},
{
"epoch": 0.02868617326448652,
"grad_norm": 0.17674313485622406,
"learning_rate": 5.733944954128441e-05,
"loss": 0.9853,
"step": 125
},
{
"epoch": 0.02983362019506598,
"grad_norm": 0.16848498582839966,
"learning_rate": 5.9633027522935784e-05,
"loss": 1.0349,
"step": 130
},
{
"epoch": 0.03098106712564544,
"grad_norm": 0.17360493540763855,
"learning_rate": 6.192660550458716e-05,
"loss": 0.9893,
"step": 135
},
{
"epoch": 0.0321285140562249,
"grad_norm": 0.191062331199646,
"learning_rate": 6.422018348623854e-05,
"loss": 0.9922,
"step": 140
},
{
"epoch": 0.03327596098680436,
"grad_norm": 0.1799137443304062,
"learning_rate": 6.651376146788991e-05,
"loss": 0.9998,
"step": 145
},
{
"epoch": 0.03442340791738382,
"grad_norm": 0.19225718080997467,
"learning_rate": 6.880733944954129e-05,
"loss": 1.0426,
"step": 150
},
{
"epoch": 0.03557085484796328,
"grad_norm": 0.20365940034389496,
"learning_rate": 7.110091743119265e-05,
"loss": 0.9869,
"step": 155
},
{
"epoch": 0.03671830177854274,
"grad_norm": 0.18115347623825073,
"learning_rate": 7.339449541284404e-05,
"loss": 0.9939,
"step": 160
},
{
"epoch": 0.0378657487091222,
"grad_norm": 0.16979213058948517,
"learning_rate": 7.568807339449542e-05,
"loss": 0.9788,
"step": 165
},
{
"epoch": 0.03901319563970167,
"grad_norm": 0.16331952810287476,
"learning_rate": 7.79816513761468e-05,
"loss": 1.0014,
"step": 170
},
{
"epoch": 0.040160642570281124,
"grad_norm": 0.20406360924243927,
"learning_rate": 8.027522935779816e-05,
"loss": 0.988,
"step": 175
},
{
"epoch": 0.04130808950086059,
"grad_norm": 0.17396636307239532,
"learning_rate": 8.256880733944955e-05,
"loss": 0.9756,
"step": 180
},
{
"epoch": 0.042455536431440045,
"grad_norm": 0.19402877986431122,
"learning_rate": 8.486238532110093e-05,
"loss": 1.0017,
"step": 185
},
{
"epoch": 0.04360298336201951,
"grad_norm": 0.16275173425674438,
"learning_rate": 8.715596330275229e-05,
"loss": 0.9881,
"step": 190
},
{
"epoch": 0.04475043029259897,
"grad_norm": 0.16075727343559265,
"learning_rate": 8.944954128440367e-05,
"loss": 0.9818,
"step": 195
},
{
"epoch": 0.04589787722317843,
"grad_norm": 0.17031803727149963,
"learning_rate": 9.174311926605506e-05,
"loss": 0.9813,
"step": 200
},
{
"epoch": 0.04704532415375789,
"grad_norm": 0.16039007902145386,
"learning_rate": 9.403669724770642e-05,
"loss": 0.9655,
"step": 205
},
{
"epoch": 0.04819277108433735,
"grad_norm": 0.17182576656341553,
"learning_rate": 9.63302752293578e-05,
"loss": 1.0034,
"step": 210
},
{
"epoch": 0.04934021801491681,
"grad_norm": 0.16478177905082703,
"learning_rate": 9.862385321100918e-05,
"loss": 0.9993,
"step": 215
},
{
"epoch": 0.05048766494549627,
"grad_norm": 0.16542178392410278,
"learning_rate": 0.00010091743119266055,
"loss": 1.005,
"step": 220
},
{
"epoch": 0.05163511187607573,
"grad_norm": 0.1568165123462677,
"learning_rate": 0.00010321100917431193,
"loss": 1.0148,
"step": 225
},
{
"epoch": 0.052782558806655194,
"grad_norm": 0.15992042422294617,
"learning_rate": 0.00010550458715596329,
"loss": 1.0084,
"step": 230
},
{
"epoch": 0.05393000573723465,
"grad_norm": 0.16262054443359375,
"learning_rate": 0.0001077981651376147,
"loss": 0.9871,
"step": 235
},
{
"epoch": 0.055077452667814115,
"grad_norm": 0.15893304347991943,
"learning_rate": 0.00011009174311926606,
"loss": 0.9609,
"step": 240
},
{
"epoch": 0.05622489959839357,
"grad_norm": 0.1627625972032547,
"learning_rate": 0.00011238532110091744,
"loss": 0.9571,
"step": 245
},
{
"epoch": 0.05737234652897304,
"grad_norm": 0.15538789331912994,
"learning_rate": 0.00011467889908256881,
"loss": 0.9632,
"step": 250
},
{
"epoch": 0.058519793459552494,
"grad_norm": 0.14804205298423767,
"learning_rate": 0.00011697247706422019,
"loss": 1.0001,
"step": 255
},
{
"epoch": 0.05966724039013196,
"grad_norm": 0.15267077088356018,
"learning_rate": 0.00011926605504587157,
"loss": 0.9697,
"step": 260
},
{
"epoch": 0.060814687320711415,
"grad_norm": 0.14952810108661652,
"learning_rate": 0.00012155963302752293,
"loss": 1.0034,
"step": 265
},
{
"epoch": 0.06196213425129088,
"grad_norm": 0.15380804240703583,
"learning_rate": 0.00012385321100917432,
"loss": 0.9614,
"step": 270
},
{
"epoch": 0.06310958118187034,
"grad_norm": 0.144377201795578,
"learning_rate": 0.0001261467889908257,
"loss": 0.965,
"step": 275
},
{
"epoch": 0.0642570281124498,
"grad_norm": 0.13880601525306702,
"learning_rate": 0.00012844036697247707,
"loss": 0.9586,
"step": 280
},
{
"epoch": 0.06540447504302926,
"grad_norm": 0.1473158597946167,
"learning_rate": 0.00013073394495412844,
"loss": 0.994,
"step": 285
},
{
"epoch": 0.06655192197360872,
"grad_norm": 0.14506657421588898,
"learning_rate": 0.00013302752293577983,
"loss": 1.0082,
"step": 290
},
{
"epoch": 0.06769936890418818,
"grad_norm": 0.1545286476612091,
"learning_rate": 0.0001353211009174312,
"loss": 0.9639,
"step": 295
},
{
"epoch": 0.06884681583476764,
"grad_norm": 0.14149315655231476,
"learning_rate": 0.00013761467889908258,
"loss": 0.9682,
"step": 300
},
{
"epoch": 0.0699942627653471,
"grad_norm": 0.1382177472114563,
"learning_rate": 0.00013990825688073395,
"loss": 0.9534,
"step": 305
},
{
"epoch": 0.07114170969592656,
"grad_norm": 0.1412287950515747,
"learning_rate": 0.0001422018348623853,
"loss": 0.9463,
"step": 310
},
{
"epoch": 0.07228915662650602,
"grad_norm": 0.13654330372810364,
"learning_rate": 0.00014449541284403673,
"loss": 0.9851,
"step": 315
},
{
"epoch": 0.07343660355708548,
"grad_norm": 0.13656587898731232,
"learning_rate": 0.0001467889908256881,
"loss": 0.9938,
"step": 320
},
{
"epoch": 0.07458405048766495,
"grad_norm": 0.13453873991966248,
"learning_rate": 0.00014908256880733945,
"loss": 0.9847,
"step": 325
},
{
"epoch": 0.0757314974182444,
"grad_norm": 0.13527260720729828,
"learning_rate": 0.00015137614678899084,
"loss": 0.9436,
"step": 330
},
{
"epoch": 0.07687894434882386,
"grad_norm": 0.13411866128444672,
"learning_rate": 0.0001536697247706422,
"loss": 0.9737,
"step": 335
},
{
"epoch": 0.07802639127940333,
"grad_norm": 0.13352230191230774,
"learning_rate": 0.0001559633027522936,
"loss": 0.977,
"step": 340
},
{
"epoch": 0.07917383820998279,
"grad_norm": 0.14445261657238007,
"learning_rate": 0.00015825688073394496,
"loss": 0.9802,
"step": 345
},
{
"epoch": 0.08032128514056225,
"grad_norm": 0.13070940971374512,
"learning_rate": 0.00016055045871559632,
"loss": 0.9509,
"step": 350
},
{
"epoch": 0.0814687320711417,
"grad_norm": 0.13417434692382812,
"learning_rate": 0.0001628440366972477,
"loss": 1.0253,
"step": 355
},
{
"epoch": 0.08261617900172118,
"grad_norm": 0.15762361884117126,
"learning_rate": 0.0001651376146788991,
"loss": 0.9936,
"step": 360
},
{
"epoch": 0.08376362593230063,
"grad_norm": 0.13492801785469055,
"learning_rate": 0.00016743119266055047,
"loss": 0.9601,
"step": 365
},
{
"epoch": 0.08491107286288009,
"grad_norm": 0.1378079354763031,
"learning_rate": 0.00016972477064220186,
"loss": 1.0292,
"step": 370
},
{
"epoch": 0.08605851979345955,
"grad_norm": 0.1303025484085083,
"learning_rate": 0.00017201834862385322,
"loss": 0.9491,
"step": 375
},
{
"epoch": 0.08720596672403902,
"grad_norm": 0.13195742666721344,
"learning_rate": 0.00017431192660550458,
"loss": 1.0134,
"step": 380
},
{
"epoch": 0.08835341365461848,
"grad_norm": 0.13493070006370544,
"learning_rate": 0.00017660550458715597,
"loss": 0.9563,
"step": 385
},
{
"epoch": 0.08950086058519793,
"grad_norm": 0.12674984335899353,
"learning_rate": 0.00017889908256880734,
"loss": 0.9772,
"step": 390
},
{
"epoch": 0.09064830751577739,
"grad_norm": 0.12925055623054504,
"learning_rate": 0.00018119266055045873,
"loss": 0.9445,
"step": 395
},
{
"epoch": 0.09179575444635686,
"grad_norm": 0.13387279212474823,
"learning_rate": 0.00018348623853211012,
"loss": 0.9724,
"step": 400
},
{
"epoch": 0.09294320137693632,
"grad_norm": 0.13336487114429474,
"learning_rate": 0.00018577981651376148,
"loss": 0.9938,
"step": 405
},
{
"epoch": 0.09409064830751578,
"grad_norm": 0.13736997544765472,
"learning_rate": 0.00018807339449541284,
"loss": 0.9719,
"step": 410
},
{
"epoch": 0.09523809523809523,
"grad_norm": 0.13675136864185333,
"learning_rate": 0.00019036697247706424,
"loss": 0.9648,
"step": 415
},
{
"epoch": 0.0963855421686747,
"grad_norm": 0.13273541629314423,
"learning_rate": 0.0001926605504587156,
"loss": 0.9717,
"step": 420
},
{
"epoch": 0.09753298909925416,
"grad_norm": 0.13400444388389587,
"learning_rate": 0.000194954128440367,
"loss": 0.9819,
"step": 425
},
{
"epoch": 0.09868043602983362,
"grad_norm": 0.1314323991537094,
"learning_rate": 0.00019724770642201835,
"loss": 0.984,
"step": 430
},
{
"epoch": 0.09982788296041308,
"grad_norm": 0.13610613346099854,
"learning_rate": 0.00019954128440366972,
"loss": 0.9681,
"step": 435
},
{
"epoch": 0.10097532989099255,
"grad_norm": 0.13332229852676392,
"learning_rate": 0.00019999948643469536,
"loss": 0.9801,
"step": 440
},
{
"epoch": 0.102122776821572,
"grad_norm": 0.12949495017528534,
"learning_rate": 0.00019999740008468594,
"loss": 0.9846,
"step": 445
},
{
"epoch": 0.10327022375215146,
"grad_norm": 0.13086913526058197,
"learning_rate": 0.00019999370888559804,
"loss": 0.923,
"step": 450
},
{
"epoch": 0.10441767068273092,
"grad_norm": 0.1352614462375641,
"learning_rate": 0.0001999884128966714,
"loss": 0.9459,
"step": 455
},
{
"epoch": 0.10556511761331039,
"grad_norm": 0.13356110453605652,
"learning_rate": 0.00019998151220290082,
"loss": 0.9859,
"step": 460
},
{
"epoch": 0.10671256454388985,
"grad_norm": 0.13900211453437805,
"learning_rate": 0.00019997300691503497,
"loss": 0.9691,
"step": 465
},
{
"epoch": 0.1078600114744693,
"grad_norm": 0.12718099355697632,
"learning_rate": 0.0001999628971695744,
"loss": 0.9618,
"step": 470
},
{
"epoch": 0.10900745840504876,
"grad_norm": 0.12601250410079956,
"learning_rate": 0.00019995118312876944,
"loss": 0.9603,
"step": 475
},
{
"epoch": 0.11015490533562823,
"grad_norm": 0.13364988565444946,
"learning_rate": 0.0001999378649806177,
"loss": 0.9366,
"step": 480
},
{
"epoch": 0.11130235226620769,
"grad_norm": 0.13241152465343475,
"learning_rate": 0.00019992294293886095,
"loss": 0.9565,
"step": 485
},
{
"epoch": 0.11244979919678715,
"grad_norm": 0.1273089498281479,
"learning_rate": 0.00019990641724298156,
"loss": 0.96,
"step": 490
},
{
"epoch": 0.11359724612736662,
"grad_norm": 0.13082469999790192,
"learning_rate": 0.000199888288158199,
"loss": 0.9767,
"step": 495
},
{
"epoch": 0.11474469305794607,
"grad_norm": 0.14526137709617615,
"learning_rate": 0.00019986855597546526,
"loss": 0.9641,
"step": 500
},
{
"epoch": 0.11589213998852553,
"grad_norm": 0.13247860968112946,
"learning_rate": 0.00019984722101146029,
"loss": 0.9676,
"step": 505
},
{
"epoch": 0.11703958691910499,
"grad_norm": 0.14177852869033813,
"learning_rate": 0.000199824283608587,
"loss": 0.9406,
"step": 510
},
{
"epoch": 0.11818703384968446,
"grad_norm": 0.1286071240901947,
"learning_rate": 0.00019979974413496566,
"loss": 0.9689,
"step": 515
},
{
"epoch": 0.11933448078026392,
"grad_norm": 0.1299976259469986,
"learning_rate": 0.00019977360298442803,
"loss": 0.9623,
"step": 520
},
{
"epoch": 0.12048192771084337,
"grad_norm": 0.12755008041858673,
"learning_rate": 0.00019974586057651102,
"loss": 0.9139,
"step": 525
},
{
"epoch": 0.12162937464142283,
"grad_norm": 0.13445819914340973,
"learning_rate": 0.00019971651735644995,
"loss": 0.9513,
"step": 530
},
{
"epoch": 0.1227768215720023,
"grad_norm": 0.13108594715595245,
"learning_rate": 0.00019968557379517152,
"loss": 0.9492,
"step": 535
},
{
"epoch": 0.12392426850258176,
"grad_norm": 0.124118372797966,
"learning_rate": 0.00019965303038928608,
"loss": 0.9411,
"step": 540
},
{
"epoch": 0.12507171543316123,
"grad_norm": 0.12611638009548187,
"learning_rate": 0.00019961888766107972,
"loss": 0.976,
"step": 545
},
{
"epoch": 0.12621916236374067,
"grad_norm": 0.13510450720787048,
"learning_rate": 0.00019958314615850598,
"loss": 0.9632,
"step": 550
},
{
"epoch": 0.12736660929432014,
"grad_norm": 0.12858641147613525,
"learning_rate": 0.00019954580645517697,
"loss": 0.9491,
"step": 555
},
{
"epoch": 0.1285140562248996,
"grad_norm": 0.13497839868068695,
"learning_rate": 0.0001995068691503541,
"loss": 0.9604,
"step": 560
},
{
"epoch": 0.12966150315547906,
"grad_norm": 0.12632782757282257,
"learning_rate": 0.00019946633486893865,
"loss": 0.9363,
"step": 565
},
{
"epoch": 0.13080895008605853,
"grad_norm": 0.12541595101356506,
"learning_rate": 0.00019942420426146153,
"loss": 0.9897,
"step": 570
},
{
"epoch": 0.13195639701663797,
"grad_norm": 0.1276572197675705,
"learning_rate": 0.00019938047800407302,
"loss": 0.9535,
"step": 575
},
{
"epoch": 0.13310384394721744,
"grad_norm": 0.12992243468761444,
"learning_rate": 0.00019933515679853182,
"loss": 0.9879,
"step": 580
},
{
"epoch": 0.1342512908777969,
"grad_norm": 0.1361103504896164,
"learning_rate": 0.0001992882413721937,
"loss": 0.9684,
"step": 585
},
{
"epoch": 0.13539873780837636,
"grad_norm": 0.13442201912403107,
"learning_rate": 0.0001992397324780001,
"loss": 0.9427,
"step": 590
},
{
"epoch": 0.13654618473895583,
"grad_norm": 0.1316874623298645,
"learning_rate": 0.00019918963089446577,
"loss": 0.9795,
"step": 595
},
{
"epoch": 0.13769363166953527,
"grad_norm": 0.12683913111686707,
"learning_rate": 0.00019913793742566647,
"loss": 0.9517,
"step": 600
},
{
"epoch": 0.13884107860011474,
"grad_norm": 0.12805037200450897,
"learning_rate": 0.00019908465290122585,
"loss": 0.9728,
"step": 605
},
{
"epoch": 0.1399885255306942,
"grad_norm": 0.13088442385196686,
"learning_rate": 0.00019902977817630243,
"loss": 0.9695,
"step": 610
},
{
"epoch": 0.14113597246127366,
"grad_norm": 0.1324809193611145,
"learning_rate": 0.00019897331413157548,
"loss": 0.9996,
"step": 615
},
{
"epoch": 0.14228341939185313,
"grad_norm": 0.12211659550666809,
"learning_rate": 0.00019891526167323145,
"loss": 0.989,
"step": 620
},
{
"epoch": 0.1434308663224326,
"grad_norm": 0.13171370327472687,
"learning_rate": 0.0001988556217329488,
"loss": 0.9714,
"step": 625
},
{
"epoch": 0.14457831325301204,
"grad_norm": 0.12398378551006317,
"learning_rate": 0.00019879439526788341,
"loss": 0.9717,
"step": 630
},
{
"epoch": 0.1457257601835915,
"grad_norm": 0.12855181097984314,
"learning_rate": 0.00019873158326065327,
"loss": 0.9635,
"step": 635
},
{
"epoch": 0.14687320711417096,
"grad_norm": 0.13017693161964417,
"learning_rate": 0.00019866718671932249,
"loss": 0.9804,
"step": 640
},
{
"epoch": 0.14802065404475043,
"grad_norm": 0.12396004796028137,
"learning_rate": 0.00019860120667738516,
"loss": 0.9985,
"step": 645
},
{
"epoch": 0.1491681009753299,
"grad_norm": 0.12556292116641998,
"learning_rate": 0.00019853364419374902,
"loss": 0.9498,
"step": 650
},
{
"epoch": 0.15031554790590934,
"grad_norm": 0.12642225623130798,
"learning_rate": 0.00019846450035271808,
"loss": 1.0224,
"step": 655
},
{
"epoch": 0.1514629948364888,
"grad_norm": 0.12983497977256775,
"learning_rate": 0.00019839377626397554,
"loss": 0.9604,
"step": 660
},
{
"epoch": 0.15261044176706828,
"grad_norm": 0.13211478292942047,
"learning_rate": 0.00019832147306256576,
"loss": 0.9864,
"step": 665
},
{
"epoch": 0.15375788869764773,
"grad_norm": 0.1218995749950409,
"learning_rate": 0.00019824759190887622,
"loss": 0.937,
"step": 670
},
{
"epoch": 0.1549053356282272,
"grad_norm": 0.13377314805984497,
"learning_rate": 0.00019817213398861866,
"loss": 0.9704,
"step": 675
},
{
"epoch": 0.15605278255880667,
"grad_norm": 0.12591040134429932,
"learning_rate": 0.0001980951005128104,
"loss": 0.9874,
"step": 680
},
{
"epoch": 0.1572002294893861,
"grad_norm": 0.12856173515319824,
"learning_rate": 0.00019801649271775459,
"loss": 0.9469,
"step": 685
},
{
"epoch": 0.15834767641996558,
"grad_norm": 0.1214781105518341,
"learning_rate": 0.00019793631186502047,
"loss": 0.9739,
"step": 690
},
{
"epoch": 0.15949512335054503,
"grad_norm": 0.12588472664356232,
"learning_rate": 0.00019785455924142318,
"loss": 0.9629,
"step": 695
},
{
"epoch": 0.1606425702811245,
"grad_norm": 0.12644240260124207,
"learning_rate": 0.0001977712361590031,
"loss": 1.0156,
"step": 700
},
{
"epoch": 0.16179001721170397,
"grad_norm": 0.12597164511680603,
"learning_rate": 0.00019768634395500465,
"loss": 0.9728,
"step": 705
},
{
"epoch": 0.1629374641422834,
"grad_norm": 0.12837018072605133,
"learning_rate": 0.00019759988399185505,
"loss": 0.9771,
"step": 710
},
{
"epoch": 0.16408491107286288,
"grad_norm": 0.13782522082328796,
"learning_rate": 0.00019751185765714234,
"loss": 0.9618,
"step": 715
},
{
"epoch": 0.16523235800344235,
"grad_norm": 0.12514767050743103,
"learning_rate": 0.00019742226636359296,
"loss": 0.9594,
"step": 720
},
{
"epoch": 0.1663798049340218,
"grad_norm": 0.1280808001756668,
"learning_rate": 0.00019733111154904943,
"loss": 0.9856,
"step": 725
},
{
"epoch": 0.16752725186460127,
"grad_norm": 0.12929755449295044,
"learning_rate": 0.00019723839467644699,
"loss": 0.9382,
"step": 730
},
{
"epoch": 0.1686746987951807,
"grad_norm": 0.5807048678398132,
"learning_rate": 0.00019714411723379015,
"loss": 0.9602,
"step": 735
},
{
"epoch": 0.16982214572576018,
"grad_norm": 0.13152436912059784,
"learning_rate": 0.0001970482807341289,
"loss": 1.003,
"step": 740
},
{
"epoch": 0.17096959265633965,
"grad_norm": 0.12607312202453613,
"learning_rate": 0.0001969508867155345,
"loss": 0.953,
"step": 745
},
{
"epoch": 0.1721170395869191,
"grad_norm": 0.13689285516738892,
"learning_rate": 0.00019685193674107452,
"loss": 0.974,
"step": 750
},
{
"epoch": 0.17326448651749857,
"grad_norm": 0.1324600875377655,
"learning_rate": 0.00019675143239878805,
"loss": 0.9953,
"step": 755
},
{
"epoch": 0.17441193344807804,
"grad_norm": 0.13404534757137299,
"learning_rate": 0.00019664937530166002,
"loss": 0.9812,
"step": 760
},
{
"epoch": 0.17555938037865748,
"grad_norm": 0.13384674489498138,
"learning_rate": 0.00019654576708759538,
"loss": 0.9632,
"step": 765
},
{
"epoch": 0.17670682730923695,
"grad_norm": 0.1277248114347458,
"learning_rate": 0.00019644060941939286,
"loss": 0.9361,
"step": 770
},
{
"epoch": 0.1778542742398164,
"grad_norm": 0.13291795551776886,
"learning_rate": 0.00019633390398471817,
"loss": 0.9762,
"step": 775
},
{
"epoch": 0.17900172117039587,
"grad_norm": 0.13213355839252472,
"learning_rate": 0.00019622565249607704,
"loss": 0.9474,
"step": 780
},
{
"epoch": 0.18014916810097534,
"grad_norm": 0.12764747440814972,
"learning_rate": 0.0001961158566907877,
"loss": 0.9329,
"step": 785
},
{
"epoch": 0.18129661503155478,
"grad_norm": 0.13760751485824585,
"learning_rate": 0.00019600451833095287,
"loss": 0.9512,
"step": 790
},
{
"epoch": 0.18244406196213425,
"grad_norm": 0.13257338106632233,
"learning_rate": 0.00019589163920343163,
"loss": 0.944,
"step": 795
},
{
"epoch": 0.18359150889271372,
"grad_norm": 0.13907872140407562,
"learning_rate": 0.00019577722111981078,
"loss": 0.9839,
"step": 800
},
{
"epoch": 0.18473895582329317,
"grad_norm": 0.1304904818534851,
"learning_rate": 0.00019566126591637558,
"loss": 0.9763,
"step": 805
},
{
"epoch": 0.18588640275387264,
"grad_norm": 0.13839825987815857,
"learning_rate": 0.0001955437754540805,
"loss": 0.96,
"step": 810
},
{
"epoch": 0.18703384968445208,
"grad_norm": 0.12638252973556519,
"learning_rate": 0.00019542475161851906,
"loss": 0.9421,
"step": 815
},
{
"epoch": 0.18818129661503155,
"grad_norm": 0.12756413221359253,
"learning_rate": 0.00019530419631989392,
"loss": 0.9593,
"step": 820
},
{
"epoch": 0.18932874354561102,
"grad_norm": 0.12539435923099518,
"learning_rate": 0.00019518211149298595,
"loss": 0.9643,
"step": 825
},
{
"epoch": 0.19047619047619047,
"grad_norm": 0.12664757668972015,
"learning_rate": 0.00019505849909712332,
"loss": 0.9575,
"step": 830
},
{
"epoch": 0.19162363740676994,
"grad_norm": 0.12583203613758087,
"learning_rate": 0.00019493336111615003,
"loss": 0.9277,
"step": 835
},
{
"epoch": 0.1927710843373494,
"grad_norm": 0.1292329877614975,
"learning_rate": 0.00019480669955839402,
"loss": 0.9797,
"step": 840
},
{
"epoch": 0.19391853126792885,
"grad_norm": 0.12693488597869873,
"learning_rate": 0.00019467851645663494,
"loss": 0.9749,
"step": 845
},
{
"epoch": 0.19506597819850832,
"grad_norm": 0.13166065514087677,
"learning_rate": 0.00019454881386807163,
"loss": 0.9645,
"step": 850
},
{
"epoch": 0.1962134251290878,
"grad_norm": 0.12855538725852966,
"learning_rate": 0.00019441759387428903,
"loss": 0.9463,
"step": 855
},
{
"epoch": 0.19736087205966724,
"grad_norm": 0.12849357724189758,
"learning_rate": 0.00019428485858122472,
"loss": 0.988,
"step": 860
},
{
"epoch": 0.1985083189902467,
"grad_norm": 0.12299315631389618,
"learning_rate": 0.00019415061011913523,
"loss": 0.9733,
"step": 865
},
{
"epoch": 0.19965576592082615,
"grad_norm": 0.13390155136585236,
"learning_rate": 0.00019401485064256176,
"loss": 0.9502,
"step": 870
},
{
"epoch": 0.20080321285140562,
"grad_norm": 0.13177447021007538,
"learning_rate": 0.0001938775823302957,
"loss": 0.9706,
"step": 875
},
{
"epoch": 0.2019506597819851,
"grad_norm": 0.13233035802841187,
"learning_rate": 0.00019373880738534358,
"loss": 1.0028,
"step": 880
},
{
"epoch": 0.20309810671256454,
"grad_norm": 0.12969984114170074,
"learning_rate": 0.00019359852803489168,
"loss": 0.9423,
"step": 885
},
{
"epoch": 0.204245553643144,
"grad_norm": 0.13504785299301147,
"learning_rate": 0.0001934567465302704,
"loss": 0.9708,
"step": 890
},
{
"epoch": 0.20539300057372348,
"grad_norm": 0.12698131799697876,
"learning_rate": 0.00019331346514691813,
"loss": 0.9548,
"step": 895
},
{
"epoch": 0.20654044750430292,
"grad_norm": 0.1307503879070282,
"learning_rate": 0.00019316868618434455,
"loss": 0.9595,
"step": 900
},
{
"epoch": 0.2076878944348824,
"grad_norm": 0.12671291828155518,
"learning_rate": 0.00019302241196609397,
"loss": 0.9456,
"step": 905
},
{
"epoch": 0.20883534136546184,
"grad_norm": 0.1356162577867508,
"learning_rate": 0.0001928746448397078,
"loss": 0.9805,
"step": 910
},
{
"epoch": 0.2099827882960413,
"grad_norm": 0.1294623613357544,
"learning_rate": 0.00019272538717668715,
"loss": 0.9831,
"step": 915
},
{
"epoch": 0.21113023522662078,
"grad_norm": 0.13363958895206451,
"learning_rate": 0.00019257464137245446,
"loss": 0.9971,
"step": 920
},
{
"epoch": 0.21227768215720022,
"grad_norm": 0.1361180692911148,
"learning_rate": 0.0001924224098463153,
"loss": 0.9393,
"step": 925
},
{
"epoch": 0.2134251290877797,
"grad_norm": 0.1273190975189209,
"learning_rate": 0.00019226869504141943,
"loss": 0.9613,
"step": 930
},
{
"epoch": 0.21457257601835916,
"grad_norm": 0.1308612823486328,
"learning_rate": 0.00019211349942472165,
"loss": 0.9703,
"step": 935
},
{
"epoch": 0.2157200229489386,
"grad_norm": 0.13174669444561005,
"learning_rate": 0.00019195682548694208,
"loss": 0.9693,
"step": 940
},
{
"epoch": 0.21686746987951808,
"grad_norm": 0.1349845975637436,
"learning_rate": 0.00019179867574252638,
"loss": 0.986,
"step": 945
},
{
"epoch": 0.21801491681009752,
"grad_norm": 0.13183628022670746,
"learning_rate": 0.00019163905272960528,
"loss": 0.9804,
"step": 950
},
{
"epoch": 0.219162363740677,
"grad_norm": 0.1328561156988144,
"learning_rate": 0.0001914779590099538,
"loss": 0.9582,
"step": 955
},
{
"epoch": 0.22030981067125646,
"grad_norm": 0.1274995058774948,
"learning_rate": 0.00019131539716895024,
"loss": 0.9245,
"step": 960
},
{
"epoch": 0.2214572576018359,
"grad_norm": 0.13776043057441711,
"learning_rate": 0.00019115136981553464,
"loss": 0.9683,
"step": 965
},
{
"epoch": 0.22260470453241538,
"grad_norm": 0.1258002072572708,
"learning_rate": 0.00019098587958216688,
"loss": 0.968,
"step": 970
},
{
"epoch": 0.22375215146299485,
"grad_norm": 0.13454443216323853,
"learning_rate": 0.00019081892912478456,
"loss": 0.9407,
"step": 975
},
{
"epoch": 0.2248995983935743,
"grad_norm": 0.13280132412910461,
"learning_rate": 0.00019065052112276018,
"loss": 0.9714,
"step": 980
},
{
"epoch": 0.22604704532415376,
"grad_norm": 0.1270975023508072,
"learning_rate": 0.00019048065827885827,
"loss": 0.942,
"step": 985
},
{
"epoch": 0.22719449225473323,
"grad_norm": 0.13367241621017456,
"learning_rate": 0.000190309343319192,
"loss": 0.981,
"step": 990
},
{
"epoch": 0.22834193918531268,
"grad_norm": 0.13318173587322235,
"learning_rate": 0.00019013657899317942,
"loss": 0.9668,
"step": 995
},
{
"epoch": 0.22948938611589215,
"grad_norm": 0.1365250200033188,
"learning_rate": 0.0001899623680734993,
"loss": 0.939,
"step": 1000
},
{
"epoch": 0.2306368330464716,
"grad_norm": 0.13613171875476837,
"learning_rate": 0.00018978671335604665,
"loss": 0.9375,
"step": 1005
},
{
"epoch": 0.23178427997705106,
"grad_norm": 0.12359146773815155,
"learning_rate": 0.00018960961765988792,
"loss": 0.9686,
"step": 1010
},
{
"epoch": 0.23293172690763053,
"grad_norm": 0.12440812587738037,
"learning_rate": 0.00018943108382721562,
"loss": 0.9284,
"step": 1015
},
{
"epoch": 0.23407917383820998,
"grad_norm": 0.13303610682487488,
"learning_rate": 0.00018925111472330283,
"loss": 0.9631,
"step": 1020
},
{
"epoch": 0.23522662076878945,
"grad_norm": 0.13873472809791565,
"learning_rate": 0.00018906971323645713,
"loss": 0.9854,
"step": 1025
},
{
"epoch": 0.23637406769936892,
"grad_norm": 0.13052476942539215,
"learning_rate": 0.00018888688227797432,
"loss": 0.9422,
"step": 1030
},
{
"epoch": 0.23752151462994836,
"grad_norm": 0.1390550136566162,
"learning_rate": 0.00018870262478209163,
"loss": 0.9847,
"step": 1035
},
{
"epoch": 0.23866896156052783,
"grad_norm": 0.13032066822052002,
"learning_rate": 0.00018851694370594069,
"loss": 0.9539,
"step": 1040
},
{
"epoch": 0.23981640849110727,
"grad_norm": 0.13494379818439484,
"learning_rate": 0.00018832984202949996,
"loss": 0.9427,
"step": 1045
},
{
"epoch": 0.24096385542168675,
"grad_norm": 0.13439466059207916,
"learning_rate": 0.00018814132275554713,
"loss": 0.992,
"step": 1050
},
{
"epoch": 0.24211130235226622,
"grad_norm": 0.1290155053138733,
"learning_rate": 0.0001879513889096106,
"loss": 0.9436,
"step": 1055
},
{
"epoch": 0.24325874928284566,
"grad_norm": 0.13683564960956573,
"learning_rate": 0.00018776004353992124,
"loss": 0.9595,
"step": 1060
},
{
"epoch": 0.24440619621342513,
"grad_norm": 0.12635931372642517,
"learning_rate": 0.00018756728971736327,
"loss": 0.9596,
"step": 1065
},
{
"epoch": 0.2455536431440046,
"grad_norm": 0.12940926849842072,
"learning_rate": 0.00018737313053542512,
"loss": 0.976,
"step": 1070
},
{
"epoch": 0.24670109007458405,
"grad_norm": 0.13413465023040771,
"learning_rate": 0.0001871775691101496,
"loss": 0.9374,
"step": 1075
},
{
"epoch": 0.24784853700516352,
"grad_norm": 0.14798474311828613,
"learning_rate": 0.00018698060858008403,
"loss": 0.9331,
"step": 1080
},
{
"epoch": 0.24899598393574296,
"grad_norm": 0.13452093303203583,
"learning_rate": 0.00018678225210622986,
"loss": 0.9893,
"step": 1085
},
{
"epoch": 0.25014343086632246,
"grad_norm": 0.13061241805553436,
"learning_rate": 0.00018658250287199196,
"loss": 0.9793,
"step": 1090
},
{
"epoch": 0.2512908777969019,
"grad_norm": 0.13494077324867249,
"learning_rate": 0.00018638136408312728,
"loss": 0.9793,
"step": 1095
},
{
"epoch": 0.25243832472748134,
"grad_norm": 0.12755267322063446,
"learning_rate": 0.0001861788389676939,
"loss": 0.957,
"step": 1100
},
{
"epoch": 0.2535857716580608,
"grad_norm": 0.13325680792331696,
"learning_rate": 0.00018597493077599867,
"loss": 0.9778,
"step": 1105
},
{
"epoch": 0.2547332185886403,
"grad_norm": 0.13293135166168213,
"learning_rate": 0.00018576964278054544,
"loss": 0.9538,
"step": 1110
},
{
"epoch": 0.25588066551921973,
"grad_norm": 0.13516201078891754,
"learning_rate": 0.00018556297827598242,
"loss": 0.9815,
"step": 1115
},
{
"epoch": 0.2570281124497992,
"grad_norm": 0.12776044011116028,
"learning_rate": 0.00018535494057904915,
"loss": 0.9513,
"step": 1120
},
{
"epoch": 0.25817555938037867,
"grad_norm": 0.12933199107646942,
"learning_rate": 0.00018514553302852356,
"loss": 0.9749,
"step": 1125
},
{
"epoch": 0.2593230063109581,
"grad_norm": 0.13189588487148285,
"learning_rate": 0.00018493475898516813,
"loss": 0.9604,
"step": 1130
},
{
"epoch": 0.26047045324153756,
"grad_norm": 0.12965457141399384,
"learning_rate": 0.00018472262183167614,
"loss": 0.9882,
"step": 1135
},
{
"epoch": 0.26161790017211706,
"grad_norm": 0.13817626237869263,
"learning_rate": 0.00018450912497261723,
"loss": 0.9508,
"step": 1140
},
{
"epoch": 0.2627653471026965,
"grad_norm": 0.12874102592468262,
"learning_rate": 0.00018429427183438288,
"loss": 0.9871,
"step": 1145
},
{
"epoch": 0.26391279403327594,
"grad_norm": 0.12571579217910767,
"learning_rate": 0.00018407806586513134,
"loss": 0.9739,
"step": 1150
},
{
"epoch": 0.26506024096385544,
"grad_norm": 0.13456568121910095,
"learning_rate": 0.00018386051053473232,
"loss": 0.9752,
"step": 1155
},
{
"epoch": 0.2662076878944349,
"grad_norm": 0.1293318271636963,
"learning_rate": 0.00018364160933471134,
"loss": 0.9483,
"step": 1160
},
{
"epoch": 0.26735513482501433,
"grad_norm": 0.13392923772335052,
"learning_rate": 0.0001834213657781936,
"loss": 0.9443,
"step": 1165
},
{
"epoch": 0.2685025817555938,
"grad_norm": 0.1313287764787674,
"learning_rate": 0.00018319978339984767,
"loss": 0.9485,
"step": 1170
},
{
"epoch": 0.26965002868617327,
"grad_norm": 0.13383394479751587,
"learning_rate": 0.0001829768657558288,
"loss": 0.9647,
"step": 1175
},
{
"epoch": 0.2707974756167527,
"grad_norm": 0.12766633927822113,
"learning_rate": 0.00018275261642372175,
"loss": 0.9627,
"step": 1180
},
{
"epoch": 0.2719449225473322,
"grad_norm": 0.12958547472953796,
"learning_rate": 0.0001825270390024834,
"loss": 0.9628,
"step": 1185
},
{
"epoch": 0.27309236947791166,
"grad_norm": 0.1323728859424591,
"learning_rate": 0.00018230013711238513,
"loss": 0.9905,
"step": 1190
},
{
"epoch": 0.2742398164084911,
"grad_norm": 0.13171570003032684,
"learning_rate": 0.00018207191439495438,
"loss": 0.9514,
"step": 1195
},
{
"epoch": 0.27538726333907054,
"grad_norm": 0.13061773777008057,
"learning_rate": 0.00018184237451291665,
"loss": 0.9644,
"step": 1200
},
{
"epoch": 0.27653471026965004,
"grad_norm": 0.1268092542886734,
"learning_rate": 0.00018161152115013637,
"loss": 0.9399,
"step": 1205
},
{
"epoch": 0.2776821572002295,
"grad_norm": 0.12714898586273193,
"learning_rate": 0.00018137935801155794,
"loss": 1.0103,
"step": 1210
},
{
"epoch": 0.27882960413080893,
"grad_norm": 0.16152754426002502,
"learning_rate": 0.0001811458888231462,
"loss": 0.9428,
"step": 1215
},
{
"epoch": 0.2799770510613884,
"grad_norm": 0.13298989832401276,
"learning_rate": 0.0001809111173318267,
"loss": 0.9532,
"step": 1220
},
{
"epoch": 0.28112449799196787,
"grad_norm": 0.1261643022298813,
"learning_rate": 0.00018067504730542551,
"loss": 1.0076,
"step": 1225
},
{
"epoch": 0.2822719449225473,
"grad_norm": 0.13440310955047607,
"learning_rate": 0.0001804376825326088,
"loss": 0.962,
"step": 1230
},
{
"epoch": 0.2834193918531268,
"grad_norm": 0.13871833682060242,
"learning_rate": 0.00018019902682282193,
"loss": 0.9581,
"step": 1235
},
{
"epoch": 0.28456683878370626,
"grad_norm": 0.12701858580112457,
"learning_rate": 0.0001799590840062285,
"loss": 0.9703,
"step": 1240
},
{
"epoch": 0.2857142857142857,
"grad_norm": 0.13460960984230042,
"learning_rate": 0.00017971785793364866,
"loss": 0.9483,
"step": 1245
},
{
"epoch": 0.2868617326448652,
"grad_norm": 0.12756766378879547,
"learning_rate": 0.0001794753524764975,
"loss": 0.9693,
"step": 1250
},
{
"epoch": 0.28800917957544464,
"grad_norm": 0.12904788553714752,
"learning_rate": 0.00017923157152672278,
"loss": 0.9874,
"step": 1255
},
{
"epoch": 0.2891566265060241,
"grad_norm": 0.1294935941696167,
"learning_rate": 0.00017898651899674254,
"loss": 0.9746,
"step": 1260
},
{
"epoch": 0.2903040734366036,
"grad_norm": 0.13519194722175598,
"learning_rate": 0.00017874019881938233,
"loss": 0.9489,
"step": 1265
},
{
"epoch": 0.291451520367183,
"grad_norm": 0.1270766407251358,
"learning_rate": 0.000178492614947812,
"loss": 0.9573,
"step": 1270
},
{
"epoch": 0.29259896729776247,
"grad_norm": 0.1287229061126709,
"learning_rate": 0.00017824377135548236,
"loss": 0.9544,
"step": 1275
},
{
"epoch": 0.2937464142283419,
"grad_norm": 0.1388266533613205,
"learning_rate": 0.00017799367203606128,
"loss": 0.9781,
"step": 1280
},
{
"epoch": 0.2948938611589214,
"grad_norm": 0.1323292851448059,
"learning_rate": 0.00017774232100336982,
"loss": 0.9783,
"step": 1285
},
{
"epoch": 0.29604130808950085,
"grad_norm": 0.12682774662971497,
"learning_rate": 0.00017748972229131757,
"loss": 0.9363,
"step": 1290
},
{
"epoch": 0.2971887550200803,
"grad_norm": 0.12563958764076233,
"learning_rate": 0.000177235879953838,
"loss": 0.972,
"step": 1295
},
{
"epoch": 0.2983362019506598,
"grad_norm": 0.13384099304676056,
"learning_rate": 0.00017698079806482343,
"loss": 0.9991,
"step": 1300
},
{
"epoch": 0.29948364888123924,
"grad_norm": 0.12890778481960297,
"learning_rate": 0.0001767244807180597,
"loss": 0.9437,
"step": 1305
},
{
"epoch": 0.3006310958118187,
"grad_norm": 0.13081791996955872,
"learning_rate": 0.00017646693202716033,
"loss": 0.9583,
"step": 1310
},
{
"epoch": 0.3017785427423982,
"grad_norm": 0.13217350840568542,
"learning_rate": 0.0001762081561255005,
"loss": 0.9599,
"step": 1315
},
{
"epoch": 0.3029259896729776,
"grad_norm": 0.1300850510597229,
"learning_rate": 0.00017594815716615093,
"loss": 0.96,
"step": 1320
},
{
"epoch": 0.30407343660355707,
"grad_norm": 0.13618028163909912,
"learning_rate": 0.000175686939321811,
"loss": 0.9466,
"step": 1325
},
{
"epoch": 0.30522088353413657,
"grad_norm": 0.13294175267219543,
"learning_rate": 0.00017542450678474184,
"loss": 0.9794,
"step": 1330
},
{
"epoch": 0.306368330464716,
"grad_norm": 0.12827391922473907,
"learning_rate": 0.00017516086376669917,
"loss": 0.9388,
"step": 1335
},
{
"epoch": 0.30751577739529545,
"grad_norm": 0.1285424679517746,
"learning_rate": 0.00017489601449886547,
"loss": 0.9533,
"step": 1340
},
{
"epoch": 0.30866322432587495,
"grad_norm": 0.13624240458011627,
"learning_rate": 0.00017462996323178235,
"loss": 0.93,
"step": 1345
},
{
"epoch": 0.3098106712564544,
"grad_norm": 0.13076668977737427,
"learning_rate": 0.00017436271423528206,
"loss": 0.9547,
"step": 1350
},
{
"epoch": 0.31095811818703384,
"grad_norm": 0.12973956763744354,
"learning_rate": 0.0001740942717984192,
"loss": 1.0114,
"step": 1355
},
{
"epoch": 0.31210556511761334,
"grad_norm": 0.13649657368659973,
"learning_rate": 0.00017382464022940182,
"loss": 0.9735,
"step": 1360
},
{
"epoch": 0.3132530120481928,
"grad_norm": 0.13023148477077484,
"learning_rate": 0.00017355382385552206,
"loss": 0.9856,
"step": 1365
},
{
"epoch": 0.3144004589787722,
"grad_norm": 0.1284298449754715,
"learning_rate": 0.0001732818270230871,
"loss": 0.966,
"step": 1370
},
{
"epoch": 0.31554790590935167,
"grad_norm": 0.1332443803548813,
"learning_rate": 0.000173008654097349,
"loss": 0.9429,
"step": 1375
},
{
"epoch": 0.31669535283993117,
"grad_norm": 0.13395999372005463,
"learning_rate": 0.000172734309462435,
"loss": 0.9748,
"step": 1380
},
{
"epoch": 0.3178427997705106,
"grad_norm": 0.12902694940567017,
"learning_rate": 0.00017245879752127692,
"loss": 0.9386,
"step": 1385
},
{
"epoch": 0.31899024670109005,
"grad_norm": 0.12902621924877167,
"learning_rate": 0.0001721821226955405,
"loss": 0.9216,
"step": 1390
},
{
"epoch": 0.32013769363166955,
"grad_norm": 0.1328728049993515,
"learning_rate": 0.00017190428942555463,
"loss": 0.943,
"step": 1395
},
{
"epoch": 0.321285140562249,
"grad_norm": 0.12963657081127167,
"learning_rate": 0.0001716253021702399,
"loss": 0.9254,
"step": 1400
},
{
"epoch": 0.32243258749282844,
"grad_norm": 0.12427925318479538,
"learning_rate": 0.0001713451654070371,
"loss": 0.9665,
"step": 1405
},
{
"epoch": 0.32358003442340794,
"grad_norm": 0.12566518783569336,
"learning_rate": 0.0001710638836318354,
"loss": 0.9778,
"step": 1410
},
{
"epoch": 0.3247274813539874,
"grad_norm": 0.1280212700366974,
"learning_rate": 0.00017078146135890014,
"loss": 0.9761,
"step": 1415
},
{
"epoch": 0.3258749282845668,
"grad_norm": 0.14105786383152008,
"learning_rate": 0.0001704979031208004,
"loss": 0.9658,
"step": 1420
},
{
"epoch": 0.3270223752151463,
"grad_norm": 0.1319873183965683,
"learning_rate": 0.0001702132134683363,
"loss": 0.96,
"step": 1425
},
{
"epoch": 0.32816982214572576,
"grad_norm": 0.12675578892230988,
"learning_rate": 0.00016992739697046586,
"loss": 0.9478,
"step": 1430
},
{
"epoch": 0.3293172690763052,
"grad_norm": 0.1259857565164566,
"learning_rate": 0.00016964045821423178,
"loss": 0.9258,
"step": 1435
},
{
"epoch": 0.3304647160068847,
"grad_norm": 0.1366124451160431,
"learning_rate": 0.00016935240180468775,
"loss": 0.9556,
"step": 1440
},
{
"epoch": 0.33161216293746415,
"grad_norm": 0.13446134328842163,
"learning_rate": 0.00016906323236482465,
"loss": 0.9675,
"step": 1445
},
{
"epoch": 0.3327596098680436,
"grad_norm": 0.13119171559810638,
"learning_rate": 0.00016877295453549614,
"loss": 0.9208,
"step": 1450
},
{
"epoch": 0.33390705679862304,
"grad_norm": 0.13426977396011353,
"learning_rate": 0.00016848157297534453,
"loss": 0.9755,
"step": 1455
},
{
"epoch": 0.33505450372920254,
"grad_norm": 0.1286066770553589,
"learning_rate": 0.0001681890923607256,
"loss": 0.9624,
"step": 1460
},
{
"epoch": 0.336201950659782,
"grad_norm": 0.1409013718366623,
"learning_rate": 0.00016789551738563384,
"loss": 0.9616,
"step": 1465
},
{
"epoch": 0.3373493975903614,
"grad_norm": 0.1301148682832718,
"learning_rate": 0.00016760085276162708,
"loss": 0.9543,
"step": 1470
},
{
"epoch": 0.3384968445209409,
"grad_norm": 0.12759418785572052,
"learning_rate": 0.00016730510321775075,
"loss": 0.9395,
"step": 1475
},
{
"epoch": 0.33964429145152036,
"grad_norm": 0.12628066539764404,
"learning_rate": 0.00016700827350046206,
"loss": 0.951,
"step": 1480
},
{
"epoch": 0.3407917383820998,
"grad_norm": 0.13049326837062836,
"learning_rate": 0.00016671036837355386,
"loss": 0.9897,
"step": 1485
},
{
"epoch": 0.3419391853126793,
"grad_norm": 0.13148042559623718,
"learning_rate": 0.00016641139261807818,
"loss": 0.9477,
"step": 1490
},
{
"epoch": 0.34308663224325875,
"grad_norm": 0.1383514553308487,
"learning_rate": 0.00016611135103226937,
"loss": 0.9836,
"step": 1495
},
{
"epoch": 0.3442340791738382,
"grad_norm": 0.13166822493076324,
"learning_rate": 0.00016581024843146725,
"loss": 0.9857,
"step": 1500
},
{
"epoch": 0.3453815261044177,
"grad_norm": 0.1284855753183365,
"learning_rate": 0.00016550808964803978,
"loss": 0.9568,
"step": 1505
},
{
"epoch": 0.34652897303499713,
"grad_norm": 0.13784931600093842,
"learning_rate": 0.00016520487953130552,
"loss": 0.9829,
"step": 1510
},
{
"epoch": 0.3476764199655766,
"grad_norm": 0.1401350200176239,
"learning_rate": 0.00016490062294745571,
"loss": 0.9468,
"step": 1515
},
{
"epoch": 0.3488238668961561,
"grad_norm": 0.13035733997821808,
"learning_rate": 0.00016459532477947634,
"loss": 0.9681,
"step": 1520
},
{
"epoch": 0.3499713138267355,
"grad_norm": 0.12921980023384094,
"learning_rate": 0.00016428898992706955,
"loss": 0.9839,
"step": 1525
},
{
"epoch": 0.35111876075731496,
"grad_norm": 0.12521222233772278,
"learning_rate": 0.00016398162330657533,
"loss": 0.97,
"step": 1530
},
{
"epoch": 0.35226620768789446,
"grad_norm": 0.13276994228363037,
"learning_rate": 0.0001636732298508922,
"loss": 0.9453,
"step": 1535
},
{
"epoch": 0.3534136546184739,
"grad_norm": 0.1314193159341812,
"learning_rate": 0.0001633638145093984,
"loss": 0.9544,
"step": 1540
},
{
"epoch": 0.35456110154905335,
"grad_norm": 0.13540154695510864,
"learning_rate": 0.00016305338224787235,
"loss": 0.9958,
"step": 1545
},
{
"epoch": 0.3557085484796328,
"grad_norm": 0.13043038547039032,
"learning_rate": 0.0001627419380484128,
"loss": 0.9166,
"step": 1550
},
{
"epoch": 0.3568559954102123,
"grad_norm": 0.13303428888320923,
"learning_rate": 0.00016242948690935912,
"loss": 0.9615,
"step": 1555
},
{
"epoch": 0.35800344234079173,
"grad_norm": 0.13199162483215332,
"learning_rate": 0.00016211603384521083,
"loss": 0.9632,
"step": 1560
},
{
"epoch": 0.3591508892713712,
"grad_norm": 0.1341785341501236,
"learning_rate": 0.00016180158388654742,
"loss": 0.9744,
"step": 1565
},
{
"epoch": 0.3602983362019507,
"grad_norm": 0.13043268024921417,
"learning_rate": 0.00016148614207994735,
"loss": 0.9494,
"step": 1570
},
{
"epoch": 0.3614457831325301,
"grad_norm": 0.13018083572387695,
"learning_rate": 0.00016116971348790712,
"loss": 0.9418,
"step": 1575
},
{
"epoch": 0.36259323006310956,
"grad_norm": 0.1259637027978897,
"learning_rate": 0.0001608523031887601,
"loss": 0.9399,
"step": 1580
},
{
"epoch": 0.36374067699368906,
"grad_norm": 0.13221842050552368,
"learning_rate": 0.00016053391627659505,
"loss": 0.9549,
"step": 1585
},
{
"epoch": 0.3648881239242685,
"grad_norm": 0.14202679693698883,
"learning_rate": 0.0001602145578611742,
"loss": 0.9279,
"step": 1590
},
{
"epoch": 0.36603557085484795,
"grad_norm": 0.13095086812973022,
"learning_rate": 0.00015989423306785142,
"loss": 0.9788,
"step": 1595
},
{
"epoch": 0.36718301778542745,
"grad_norm": 0.12878695130348206,
"learning_rate": 0.00015957294703748982,
"loss": 0.9564,
"step": 1600
},
{
"epoch": 0.3683304647160069,
"grad_norm": 0.12988603115081787,
"learning_rate": 0.00015925070492637944,
"loss": 0.9672,
"step": 1605
},
{
"epoch": 0.36947791164658633,
"grad_norm": 0.13579297065734863,
"learning_rate": 0.0001589275119061542,
"loss": 0.9381,
"step": 1610
},
{
"epoch": 0.37062535857716583,
"grad_norm": 0.12680459022521973,
"learning_rate": 0.00015860337316370916,
"loss": 0.9512,
"step": 1615
},
{
"epoch": 0.3717728055077453,
"grad_norm": 0.12983225286006927,
"learning_rate": 0.0001582782939011173,
"loss": 0.9539,
"step": 1620
},
{
"epoch": 0.3729202524383247,
"grad_norm": 0.13091208040714264,
"learning_rate": 0.00015795227933554568,
"loss": 0.9675,
"step": 1625
},
{
"epoch": 0.37406769936890416,
"grad_norm": 0.13072653114795685,
"learning_rate": 0.00015762533469917216,
"loss": 0.9327,
"step": 1630
},
{
"epoch": 0.37521514629948366,
"grad_norm": 0.13025379180908203,
"learning_rate": 0.00015729746523910113,
"loss": 0.9623,
"step": 1635
},
{
"epoch": 0.3763625932300631,
"grad_norm": 0.14888666570186615,
"learning_rate": 0.00015696867621727942,
"loss": 0.9436,
"step": 1640
},
{
"epoch": 0.37751004016064255,
"grad_norm": 0.1301380842924118,
"learning_rate": 0.00015663897291041175,
"loss": 0.9613,
"step": 1645
},
{
"epoch": 0.37865748709122204,
"grad_norm": 0.12864987552165985,
"learning_rate": 0.00015630836060987624,
"loss": 0.9701,
"step": 1650
},
{
"epoch": 0.3798049340218015,
"grad_norm": 0.13158966600894928,
"learning_rate": 0.00015597684462163923,
"loss": 1.0041,
"step": 1655
},
{
"epoch": 0.38095238095238093,
"grad_norm": 0.12787610292434692,
"learning_rate": 0.0001556444302661704,
"loss": 0.942,
"step": 1660
},
{
"epoch": 0.38209982788296043,
"grad_norm": 0.1329614371061325,
"learning_rate": 0.00015531112287835717,
"loss": 0.9607,
"step": 1665
},
{
"epoch": 0.3832472748135399,
"grad_norm": 0.13437345623970032,
"learning_rate": 0.00015497692780741908,
"loss": 0.9657,
"step": 1670
},
{
"epoch": 0.3843947217441193,
"grad_norm": 0.13434943556785583,
"learning_rate": 0.0001546418504168222,
"loss": 0.9635,
"step": 1675
},
{
"epoch": 0.3855421686746988,
"grad_norm": 0.129192054271698,
"learning_rate": 0.00015430589608419264,
"loss": 0.9527,
"step": 1680
},
{
"epoch": 0.38668961560527826,
"grad_norm": 0.1313447207212448,
"learning_rate": 0.00015396907020123068,
"loss": 0.942,
"step": 1685
},
{
"epoch": 0.3878370625358577,
"grad_norm": 0.13218559324741364,
"learning_rate": 0.00015363137817362392,
"loss": 0.9602,
"step": 1690
},
{
"epoch": 0.3889845094664372,
"grad_norm": 0.14416509866714478,
"learning_rate": 0.00015329282542096064,
"loss": 0.9419,
"step": 1695
},
{
"epoch": 0.39013195639701664,
"grad_norm": 0.14515794813632965,
"learning_rate": 0.00015295341737664285,
"loss": 0.9381,
"step": 1700
},
{
"epoch": 0.3912794033275961,
"grad_norm": 0.13562296330928802,
"learning_rate": 0.000152613159487799,
"loss": 0.9594,
"step": 1705
},
{
"epoch": 0.3924268502581756,
"grad_norm": 0.13049104809761047,
"learning_rate": 0.00015227205721519675,
"loss": 0.9472,
"step": 1710
},
{
"epoch": 0.39357429718875503,
"grad_norm": 0.13842852413654327,
"learning_rate": 0.00015193011603315503,
"loss": 0.995,
"step": 1715
},
{
"epoch": 0.3947217441193345,
"grad_norm": 0.14135673642158508,
"learning_rate": 0.00015158734142945644,
"loss": 0.9992,
"step": 1720
},
{
"epoch": 0.3958691910499139,
"grad_norm": 0.13362735509872437,
"learning_rate": 0.0001512437389052591,
"loss": 0.9279,
"step": 1725
},
{
"epoch": 0.3970166379804934,
"grad_norm": 0.1268196702003479,
"learning_rate": 0.0001508993139750083,
"loss": 0.9369,
"step": 1730
},
{
"epoch": 0.39816408491107286,
"grad_norm": 0.12576265633106232,
"learning_rate": 0.0001505540721663481,
"loss": 0.9777,
"step": 1735
},
{
"epoch": 0.3993115318416523,
"grad_norm": 0.1377476304769516,
"learning_rate": 0.0001502080190200325,
"loss": 0.96,
"step": 1740
},
{
"epoch": 0.4004589787722318,
"grad_norm": 0.12954863905906677,
"learning_rate": 0.00014986116008983664,
"loss": 0.9612,
"step": 1745
},
{
"epoch": 0.40160642570281124,
"grad_norm": 0.13743706047534943,
"learning_rate": 0.00014951350094246762,
"loss": 0.977,
"step": 1750
},
{
"epoch": 0.4027538726333907,
"grad_norm": 0.1272636353969574,
"learning_rate": 0.0001491650471574751,
"loss": 0.9603,
"step": 1755
},
{
"epoch": 0.4039013195639702,
"grad_norm": 0.13247370719909668,
"learning_rate": 0.00014881580432716182,
"loss": 0.9251,
"step": 1760
},
{
"epoch": 0.40504876649454963,
"grad_norm": 0.13983896374702454,
"learning_rate": 0.00014846577805649388,
"loss": 0.9556,
"step": 1765
},
{
"epoch": 0.40619621342512907,
"grad_norm": 0.1289101541042328,
"learning_rate": 0.00014811497396301072,
"loss": 0.9391,
"step": 1770
},
{
"epoch": 0.40734366035570857,
"grad_norm": 0.12569357454776764,
"learning_rate": 0.00014776339767673491,
"loss": 0.9301,
"step": 1775
},
{
"epoch": 0.408491107286288,
"grad_norm": 0.1362708956003189,
"learning_rate": 0.000147411054840082,
"loss": 0.9802,
"step": 1780
},
{
"epoch": 0.40963855421686746,
"grad_norm": 0.13561737537384033,
"learning_rate": 0.00014705795110776974,
"loss": 0.9624,
"step": 1785
},
{
"epoch": 0.41078600114744696,
"grad_norm": 0.13607259094715118,
"learning_rate": 0.0001467040921467275,
"loss": 0.9543,
"step": 1790
},
{
"epoch": 0.4119334480780264,
"grad_norm": 0.13344360888004303,
"learning_rate": 0.00014634948363600518,
"loss": 0.959,
"step": 1795
},
{
"epoch": 0.41308089500860584,
"grad_norm": 0.14161522686481476,
"learning_rate": 0.00014599413126668213,
"loss": 0.9432,
"step": 1800
},
{
"epoch": 0.4142283419391853,
"grad_norm": 0.13088949024677277,
"learning_rate": 0.00014563804074177588,
"loss": 0.9647,
"step": 1805
},
{
"epoch": 0.4153757888697648,
"grad_norm": 0.13591496646404266,
"learning_rate": 0.00014528121777615058,
"loss": 0.9531,
"step": 1810
},
{
"epoch": 0.4165232358003442,
"grad_norm": 0.1339530646800995,
"learning_rate": 0.0001449236680964251,
"loss": 0.9549,
"step": 1815
},
{
"epoch": 0.41767068273092367,
"grad_norm": 0.12666349112987518,
"learning_rate": 0.00014456539744088138,
"loss": 0.9424,
"step": 1820
},
{
"epoch": 0.41881812966150317,
"grad_norm": 0.12611252069473267,
"learning_rate": 0.00014420641155937224,
"loss": 0.9359,
"step": 1825
},
{
"epoch": 0.4199655765920826,
"grad_norm": 0.13051684200763702,
"learning_rate": 0.00014384671621322915,
"loss": 0.9488,
"step": 1830
},
{
"epoch": 0.42111302352266206,
"grad_norm": 0.13898026943206787,
"learning_rate": 0.00014348631717516953,
"loss": 0.9748,
"step": 1835
},
{
"epoch": 0.42226047045324155,
"grad_norm": 0.13012805581092834,
"learning_rate": 0.00014312522022920444,
"loss": 0.9631,
"step": 1840
},
{
"epoch": 0.423407917383821,
"grad_norm": 0.12887822091579437,
"learning_rate": 0.00014276343117054563,
"loss": 0.9532,
"step": 1845
},
{
"epoch": 0.42455536431440044,
"grad_norm": 0.14254815876483917,
"learning_rate": 0.00014240095580551234,
"loss": 1.0004,
"step": 1850
},
{
"epoch": 0.42570281124497994,
"grad_norm": 0.13232548534870148,
"learning_rate": 0.0001420377999514384,
"loss": 0.9644,
"step": 1855
},
{
"epoch": 0.4268502581755594,
"grad_norm": 0.13256525993347168,
"learning_rate": 0.00014167396943657874,
"loss": 0.9671,
"step": 1860
},
{
"epoch": 0.4279977051061388,
"grad_norm": 0.13868069648742676,
"learning_rate": 0.0001413094701000158,
"loss": 0.924,
"step": 1865
},
{
"epoch": 0.4291451520367183,
"grad_norm": 0.13542011380195618,
"learning_rate": 0.00014094430779156597,
"loss": 0.9751,
"step": 1870
},
{
"epoch": 0.43029259896729777,
"grad_norm": 0.15172503888607025,
"learning_rate": 0.00014057848837168547,
"loss": 0.9811,
"step": 1875
},
{
"epoch": 0.4314400458978772,
"grad_norm": 0.13170887529850006,
"learning_rate": 0.00014021201771137663,
"loss": 0.9598,
"step": 1880
},
{
"epoch": 0.4325874928284567,
"grad_norm": 0.13413281738758087,
"learning_rate": 0.00013984490169209333,
"loss": 0.9251,
"step": 1885
},
{
"epoch": 0.43373493975903615,
"grad_norm": 0.13315904140472412,
"learning_rate": 0.00013947714620564678,
"loss": 0.9983,
"step": 1890
},
{
"epoch": 0.4348823866896156,
"grad_norm": 0.131989523768425,
"learning_rate": 0.00013910875715411098,
"loss": 0.9411,
"step": 1895
},
{
"epoch": 0.43602983362019504,
"grad_norm": 0.1380765587091446,
"learning_rate": 0.000138739740449728,
"loss": 0.9292,
"step": 1900
},
{
"epoch": 0.43717728055077454,
"grad_norm": 0.13661834597587585,
"learning_rate": 0.0001383701020148129,
"loss": 0.9477,
"step": 1905
},
{
"epoch": 0.438324727481354,
"grad_norm": 0.13486482203006744,
"learning_rate": 0.000137999847781659,
"loss": 1.0169,
"step": 1910
},
{
"epoch": 0.4394721744119334,
"grad_norm": 0.13214510679244995,
"learning_rate": 0.00013762898369244238,
"loss": 0.9418,
"step": 1915
},
{
"epoch": 0.4406196213425129,
"grad_norm": 0.12652263045310974,
"learning_rate": 0.00013725751569912682,
"loss": 0.9679,
"step": 1920
},
{
"epoch": 0.44176706827309237,
"grad_norm": 0.13453719019889832,
"learning_rate": 0.00013688544976336783,
"loss": 0.9318,
"step": 1925
},
{
"epoch": 0.4429145152036718,
"grad_norm": 0.14246058464050293,
"learning_rate": 0.00013651279185641752,
"loss": 0.9544,
"step": 1930
},
{
"epoch": 0.4440619621342513,
"grad_norm": 0.13308945298194885,
"learning_rate": 0.0001361395479590283,
"loss": 0.9817,
"step": 1935
},
{
"epoch": 0.44520940906483075,
"grad_norm": 0.12958864867687225,
"learning_rate": 0.00013576572406135722,
"loss": 0.9555,
"step": 1940
},
{
"epoch": 0.4463568559954102,
"grad_norm": 0.1342882663011551,
"learning_rate": 0.00013539132616286956,
"loss": 0.9723,
"step": 1945
},
{
"epoch": 0.4475043029259897,
"grad_norm": 0.12933650612831116,
"learning_rate": 0.00013501636027224282,
"loss": 0.9516,
"step": 1950
},
{
"epoch": 0.44865174985656914,
"grad_norm": 0.1319791078567505,
"learning_rate": 0.0001346408324072701,
"loss": 0.9619,
"step": 1955
},
{
"epoch": 0.4497991967871486,
"grad_norm": 0.13945066928863525,
"learning_rate": 0.0001342647485947635,
"loss": 0.9109,
"step": 1960
},
{
"epoch": 0.4509466437177281,
"grad_norm": 0.13085277378559113,
"learning_rate": 0.00013388811487045766,
"loss": 0.968,
"step": 1965
},
{
"epoch": 0.4520940906483075,
"grad_norm": 0.12879082560539246,
"learning_rate": 0.0001335109372789125,
"loss": 0.9263,
"step": 1970
},
{
"epoch": 0.45324153757888697,
"grad_norm": 0.13467159867286682,
"learning_rate": 0.00013313322187341652,
"loss": 0.9529,
"step": 1975
},
{
"epoch": 0.45438898450946646,
"grad_norm": 0.12812353670597076,
"learning_rate": 0.00013275497471588953,
"loss": 0.9455,
"step": 1980
},
{
"epoch": 0.4555364314400459,
"grad_norm": 0.1305771768093109,
"learning_rate": 0.0001323762018767854,
"loss": 0.9627,
"step": 1985
},
{
"epoch": 0.45668387837062535,
"grad_norm": 0.14161469042301178,
"learning_rate": 0.00013199690943499457,
"loss": 0.9981,
"step": 1990
},
{
"epoch": 0.4578313253012048,
"grad_norm": 0.13396026194095612,
"learning_rate": 0.0001316171034777466,
"loss": 0.9786,
"step": 1995
},
{
"epoch": 0.4589787722317843,
"grad_norm": 0.18823152780532837,
"learning_rate": 0.00013123679010051232,
"loss": 0.9294,
"step": 2000
},
{
"epoch": 0.46012621916236374,
"grad_norm": 0.137470543384552,
"learning_rate": 0.00013085597540690618,
"loss": 0.9393,
"step": 2005
},
{
"epoch": 0.4612736660929432,
"grad_norm": 0.1302119791507721,
"learning_rate": 0.00013047466550858812,
"loss": 0.9666,
"step": 2010
},
{
"epoch": 0.4624211130235227,
"grad_norm": 0.13366107642650604,
"learning_rate": 0.00013009286652516575,
"loss": 0.9243,
"step": 2015
},
{
"epoch": 0.4635685599541021,
"grad_norm": 0.14005546271800995,
"learning_rate": 0.00012971058458409576,
"loss": 0.9415,
"step": 2020
},
{
"epoch": 0.46471600688468157,
"grad_norm": 0.1359815001487732,
"learning_rate": 0.00012932782582058584,
"loss": 0.9681,
"step": 2025
},
{
"epoch": 0.46586345381526106,
"grad_norm": 0.12984395027160645,
"learning_rate": 0.00012894459637749627,
"loss": 0.9475,
"step": 2030
},
{
"epoch": 0.4670109007458405,
"grad_norm": 0.12904886901378632,
"learning_rate": 0.0001285609024052411,
"loss": 0.9584,
"step": 2035
},
{
"epoch": 0.46815834767641995,
"grad_norm": 0.13644804060459137,
"learning_rate": 0.00012817675006168963,
"loss": 0.9757,
"step": 2040
},
{
"epoch": 0.46930579460699945,
"grad_norm": 0.1333729475736618,
"learning_rate": 0.00012779214551206746,
"loss": 0.9733,
"step": 2045
},
{
"epoch": 0.4704532415375789,
"grad_norm": 0.128279909491539,
"learning_rate": 0.0001274070949288577,
"loss": 0.9772,
"step": 2050
},
{
"epoch": 0.47160068846815834,
"grad_norm": 0.15798795223236084,
"learning_rate": 0.00012702160449170165,
"loss": 0.9499,
"step": 2055
},
{
"epoch": 0.47274813539873783,
"grad_norm": 0.12967512011528015,
"learning_rate": 0.0001266356803873,
"loss": 0.9532,
"step": 2060
},
{
"epoch": 0.4738955823293173,
"grad_norm": 0.1309663951396942,
"learning_rate": 0.0001262493288093131,
"loss": 0.9517,
"step": 2065
},
{
"epoch": 0.4750430292598967,
"grad_norm": 0.12817169725894928,
"learning_rate": 0.000125862555958262,
"loss": 0.98,
"step": 2070
},
{
"epoch": 0.47619047619047616,
"grad_norm": 0.1262216567993164,
"learning_rate": 0.00012547536804142858,
"loss": 0.9459,
"step": 2075
},
{
"epoch": 0.47733792312105566,
"grad_norm": 0.13227002322673798,
"learning_rate": 0.0001250877712727561,
"loss": 0.9733,
"step": 2080
},
{
"epoch": 0.4784853700516351,
"grad_norm": 0.13167516887187958,
"learning_rate": 0.00012469977187274945,
"loss": 0.9611,
"step": 2085
},
{
"epoch": 0.47963281698221455,
"grad_norm": 0.12410388141870499,
"learning_rate": 0.00012431137606837538,
"loss": 0.9346,
"step": 2090
},
{
"epoch": 0.48078026391279405,
"grad_norm": 0.13433882594108582,
"learning_rate": 0.0001239225900929624,
"loss": 0.9528,
"step": 2095
},
{
"epoch": 0.4819277108433735,
"grad_norm": 0.13042834401130676,
"learning_rate": 0.00012353342018610084,
"loss": 0.958,
"step": 2100
},
{
"epoch": 0.48307515777395293,
"grad_norm": 0.1367100179195404,
"learning_rate": 0.00012314387259354282,
"loss": 0.9665,
"step": 2105
},
{
"epoch": 0.48422260470453243,
"grad_norm": 0.13082900643348694,
"learning_rate": 0.00012275395356710177,
"loss": 0.9706,
"step": 2110
},
{
"epoch": 0.4853700516351119,
"grad_norm": 0.13011516630649567,
"learning_rate": 0.0001223636693645523,
"loss": 1.0014,
"step": 2115
},
{
"epoch": 0.4865174985656913,
"grad_norm": 0.12624427676200867,
"learning_rate": 0.00012197302624952971,
"loss": 0.9607,
"step": 2120
},
{
"epoch": 0.4876649454962708,
"grad_norm": 0.13805273175239563,
"learning_rate": 0.00012158203049142947,
"loss": 0.9245,
"step": 2125
},
{
"epoch": 0.48881239242685026,
"grad_norm": 0.125767320394516,
"learning_rate": 0.00012119068836530644,
"loss": 0.9598,
"step": 2130
},
{
"epoch": 0.4899598393574297,
"grad_norm": 0.12957511842250824,
"learning_rate": 0.00012079900615177449,
"loss": 0.9658,
"step": 2135
},
{
"epoch": 0.4911072862880092,
"grad_norm": 0.13400690257549286,
"learning_rate": 0.00012040699013690543,
"loss": 0.9033,
"step": 2140
},
{
"epoch": 0.49225473321858865,
"grad_norm": 0.1384160965681076,
"learning_rate": 0.00012001464661212827,
"loss": 0.9111,
"step": 2145
},
{
"epoch": 0.4934021801491681,
"grad_norm": 0.13281968235969543,
"learning_rate": 0.0001196219818741281,
"loss": 0.9509,
"step": 2150
},
{
"epoch": 0.4945496270797476,
"grad_norm": 0.12926004827022552,
"learning_rate": 0.00011922900222474523,
"loss": 0.9744,
"step": 2155
},
{
"epoch": 0.49569707401032703,
"grad_norm": 0.13690362870693207,
"learning_rate": 0.00011883571397087387,
"loss": 0.9519,
"step": 2160
},
{
"epoch": 0.4968445209409065,
"grad_norm": 0.13583113253116608,
"learning_rate": 0.00011844212342436112,
"loss": 0.9651,
"step": 2165
},
{
"epoch": 0.4979919678714859,
"grad_norm": 0.13219110667705536,
"learning_rate": 0.00011804823690190538,
"loss": 0.9724,
"step": 2170
},
{
"epoch": 0.4991394148020654,
"grad_norm": 0.13287144899368286,
"learning_rate": 0.00011765406072495528,
"loss": 0.95,
"step": 2175
},
{
"epoch": 0.5002868617326449,
"grad_norm": 0.12971337139606476,
"learning_rate": 0.00011725960121960806,
"loss": 0.933,
"step": 2180
},
{
"epoch": 0.5014343086632244,
"grad_norm": 0.13551461696624756,
"learning_rate": 0.00011686486471650798,
"loss": 0.9417,
"step": 2185
},
{
"epoch": 0.5025817555938038,
"grad_norm": 0.12743504345417023,
"learning_rate": 0.0001164698575507449,
"loss": 0.9396,
"step": 2190
},
{
"epoch": 0.5037292025243832,
"grad_norm": 0.13198213279247284,
"learning_rate": 0.00011607458606175253,
"loss": 0.9139,
"step": 2195
},
{
"epoch": 0.5048766494549627,
"grad_norm": 0.13129118084907532,
"learning_rate": 0.00011567905659320663,
"loss": 0.94,
"step": 2200
},
{
"epoch": 0.5060240963855421,
"grad_norm": 0.13837940990924835,
"learning_rate": 0.00011528327549292326,
"loss": 0.9357,
"step": 2205
},
{
"epoch": 0.5071715433161216,
"grad_norm": 0.13046391308307648,
"learning_rate": 0.00011488724911275694,
"loss": 0.9488,
"step": 2210
},
{
"epoch": 0.5083189902467011,
"grad_norm": 0.12679821252822876,
"learning_rate": 0.00011449098380849858,
"loss": 0.9627,
"step": 2215
},
{
"epoch": 0.5094664371772806,
"grad_norm": 0.12639085948467255,
"learning_rate": 0.00011409448593977363,
"loss": 0.9246,
"step": 2220
},
{
"epoch": 0.51061388410786,
"grad_norm": 0.13501042127609253,
"learning_rate": 0.00011369776186993994,
"loss": 0.9439,
"step": 2225
},
{
"epoch": 0.5117613310384395,
"grad_norm": 0.12938356399536133,
"learning_rate": 0.0001133008179659856,
"loss": 0.9551,
"step": 2230
},
{
"epoch": 0.5129087779690189,
"grad_norm": 0.13302691280841827,
"learning_rate": 0.00011290366059842683,
"loss": 0.9475,
"step": 2235
},
{
"epoch": 0.5140562248995983,
"grad_norm": 0.13080324232578278,
"learning_rate": 0.00011250629614120571,
"loss": 0.9586,
"step": 2240
},
{
"epoch": 0.5152036718301779,
"grad_norm": 0.12532858550548553,
"learning_rate": 0.00011210873097158786,
"loss": 0.9296,
"step": 2245
},
{
"epoch": 0.5163511187607573,
"grad_norm": 0.13357555866241455,
"learning_rate": 0.00011171097147006013,
"loss": 0.9692,
"step": 2250
},
{
"epoch": 0.5174985656913368,
"grad_norm": 0.13192251324653625,
"learning_rate": 0.00011131302402022821,
"loss": 0.9678,
"step": 2255
},
{
"epoch": 0.5186460126219162,
"grad_norm": 0.12825413048267365,
"learning_rate": 0.00011091489500871408,
"loss": 0.9521,
"step": 2260
},
{
"epoch": 0.5197934595524957,
"grad_norm": 0.1338176429271698,
"learning_rate": 0.00011051659082505366,
"loss": 0.9664,
"step": 2265
},
{
"epoch": 0.5209409064830751,
"grad_norm": 0.13753551244735718,
"learning_rate": 0.00011011811786159416,
"loss": 0.9608,
"step": 2270
},
{
"epoch": 0.5220883534136547,
"grad_norm": 0.1290527880191803,
"learning_rate": 0.00010971948251339157,
"loss": 0.9334,
"step": 2275
},
{
"epoch": 0.5232358003442341,
"grad_norm": 0.13250134885311127,
"learning_rate": 0.00010932069117810787,
"loss": 0.9512,
"step": 2280
},
{
"epoch": 0.5243832472748136,
"grad_norm": 0.13204516470432281,
"learning_rate": 0.00010892175025590856,
"loss": 0.9711,
"step": 2285
},
{
"epoch": 0.525530694205393,
"grad_norm": 0.12426961213350296,
"learning_rate": 0.00010852266614935982,
"loss": 0.9634,
"step": 2290
},
{
"epoch": 0.5266781411359724,
"grad_norm": 0.13476422429084778,
"learning_rate": 0.00010812344526332578,
"loss": 0.9345,
"step": 2295
},
{
"epoch": 0.5278255880665519,
"grad_norm": 0.1305239349603653,
"learning_rate": 0.00010772409400486571,
"loss": 0.973,
"step": 2300
},
{
"epoch": 0.5289730349971313,
"grad_norm": 0.12918445467948914,
"learning_rate": 0.00010732461878313125,
"loss": 0.9602,
"step": 2305
},
{
"epoch": 0.5301204819277109,
"grad_norm": 0.13843993842601776,
"learning_rate": 0.00010692502600926348,
"loss": 0.9528,
"step": 2310
},
{
"epoch": 0.5312679288582903,
"grad_norm": 0.1355164349079132,
"learning_rate": 0.00010652532209629011,
"loss": 0.9423,
"step": 2315
},
{
"epoch": 0.5324153757888698,
"grad_norm": 0.1288328468799591,
"learning_rate": 0.00010612551345902244,
"loss": 0.9223,
"step": 2320
},
{
"epoch": 0.5335628227194492,
"grad_norm": 0.13057994842529297,
"learning_rate": 0.00010572560651395258,
"loss": 0.9117,
"step": 2325
},
{
"epoch": 0.5347102696500287,
"grad_norm": 0.12813064455986023,
"learning_rate": 0.0001053256076791503,
"loss": 0.953,
"step": 2330
},
{
"epoch": 0.5358577165806081,
"grad_norm": 0.13439275324344635,
"learning_rate": 0.00010492552337416007,
"loss": 0.9552,
"step": 2335
},
{
"epoch": 0.5370051635111877,
"grad_norm": 0.12726064026355743,
"learning_rate": 0.00010452536001989815,
"loss": 0.9556,
"step": 2340
},
{
"epoch": 0.5381526104417671,
"grad_norm": 0.1329938769340515,
"learning_rate": 0.00010412512403854942,
"loss": 0.9806,
"step": 2345
},
{
"epoch": 0.5393000573723465,
"grad_norm": 0.13323380053043365,
"learning_rate": 0.00010372482185346435,
"loss": 0.9542,
"step": 2350
},
{
"epoch": 0.540447504302926,
"grad_norm": 0.12690098583698273,
"learning_rate": 0.00010332445988905586,
"loss": 0.9443,
"step": 2355
},
{
"epoch": 0.5415949512335054,
"grad_norm": 0.12987449765205383,
"learning_rate": 0.00010292404457069631,
"loss": 0.9617,
"step": 2360
},
{
"epoch": 0.5427423981640849,
"grad_norm": 0.12815946340560913,
"learning_rate": 0.00010252358232461436,
"loss": 0.9741,
"step": 2365
},
{
"epoch": 0.5438898450946644,
"grad_norm": 0.13437722623348236,
"learning_rate": 0.00010212307957779173,
"loss": 0.9798,
"step": 2370
},
{
"epoch": 0.5450372920252439,
"grad_norm": 0.14197693765163422,
"learning_rate": 0.00010172254275786017,
"loss": 0.9112,
"step": 2375
},
{
"epoch": 0.5461847389558233,
"grad_norm": 0.12714983522891998,
"learning_rate": 0.0001013219782929983,
"loss": 0.9103,
"step": 2380
},
{
"epoch": 0.5473321858864028,
"grad_norm": 0.13637425005435944,
"learning_rate": 0.0001009213926118284,
"loss": 0.9564,
"step": 2385
},
{
"epoch": 0.5484796328169822,
"grad_norm": 0.13670361042022705,
"learning_rate": 0.00010052079214331318,
"loss": 0.9295,
"step": 2390
},
{
"epoch": 0.5496270797475616,
"grad_norm": 0.12989582121372223,
"learning_rate": 0.00010012018331665272,
"loss": 0.9428,
"step": 2395
},
{
"epoch": 0.5507745266781411,
"grad_norm": 0.12808671593666077,
"learning_rate": 9.971957256118129e-05,
"loss": 0.9216,
"step": 2400
},
{
"epoch": 0.5519219736087206,
"grad_norm": 0.1358453780412674,
"learning_rate": 9.931896630626402e-05,
"loss": 0.9711,
"step": 2405
},
{
"epoch": 0.5530694205393001,
"grad_norm": 0.1230422705411911,
"learning_rate": 9.891837098119389e-05,
"loss": 0.9282,
"step": 2410
},
{
"epoch": 0.5542168674698795,
"grad_norm": 0.13326068222522736,
"learning_rate": 9.851779301508842e-05,
"loss": 0.9906,
"step": 2415
},
{
"epoch": 0.555364314400459,
"grad_norm": 0.12692315876483917,
"learning_rate": 9.811723883678654e-05,
"loss": 0.9537,
"step": 2420
},
{
"epoch": 0.5565117613310384,
"grad_norm": 0.13576091825962067,
"learning_rate": 9.771671487474546e-05,
"loss": 0.9393,
"step": 2425
},
{
"epoch": 0.5576592082616179,
"grad_norm": 0.12827935814857483,
"learning_rate": 9.731622755693737e-05,
"loss": 0.9014,
"step": 2430
},
{
"epoch": 0.5588066551921974,
"grad_norm": 0.1317823976278305,
"learning_rate": 9.691578331074643e-05,
"loss": 0.9424,
"step": 2435
},
{
"epoch": 0.5599541021227769,
"grad_norm": 0.13604985177516937,
"learning_rate": 9.651538856286551e-05,
"loss": 0.9395,
"step": 2440
},
{
"epoch": 0.5611015490533563,
"grad_norm": 0.1410413533449173,
"learning_rate": 9.611504973919311e-05,
"loss": 0.9711,
"step": 2445
},
{
"epoch": 0.5622489959839357,
"grad_norm": 0.1319737732410431,
"learning_rate": 9.571477326473021e-05,
"loss": 0.944,
"step": 2450
},
{
"epoch": 0.5633964429145152,
"grad_norm": 0.1291595995426178,
"learning_rate": 9.531456556347712e-05,
"loss": 0.9621,
"step": 2455
},
{
"epoch": 0.5645438898450946,
"grad_norm": 0.12568049132823944,
"learning_rate": 9.491443305833043e-05,
"loss": 0.923,
"step": 2460
},
{
"epoch": 0.5656913367756741,
"grad_norm": 0.13689230382442474,
"learning_rate": 9.451438217097994e-05,
"loss": 0.9335,
"step": 2465
},
{
"epoch": 0.5668387837062536,
"grad_norm": 0.12956008315086365,
"learning_rate": 9.411441932180554e-05,
"loss": 0.9757,
"step": 2470
},
{
"epoch": 0.5679862306368331,
"grad_norm": 0.13027647137641907,
"learning_rate": 9.371455092977423e-05,
"loss": 0.9225,
"step": 2475
},
{
"epoch": 0.5691336775674125,
"grad_norm": 0.12894025444984436,
"learning_rate": 9.331478341233706e-05,
"loss": 0.9508,
"step": 2480
},
{
"epoch": 0.570281124497992,
"grad_norm": 0.13095425069332123,
"learning_rate": 9.291512318532614e-05,
"loss": 0.9388,
"step": 2485
},
{
"epoch": 0.5714285714285714,
"grad_norm": 0.1269254833459854,
"learning_rate": 9.251557666285174e-05,
"loss": 0.9682,
"step": 2490
},
{
"epoch": 0.5725760183591508,
"grad_norm": 0.1342654824256897,
"learning_rate": 9.211615025719919e-05,
"loss": 0.9229,
"step": 2495
},
{
"epoch": 0.5737234652897304,
"grad_norm": 0.1338120847940445,
"learning_rate": 9.17168503787262e-05,
"loss": 0.9593,
"step": 2500
},
{
"epoch": 0.5748709122203098,
"grad_norm": 0.1292128562927246,
"learning_rate": 9.131768343575979e-05,
"loss": 0.9353,
"step": 2505
},
{
"epoch": 0.5760183591508893,
"grad_norm": 0.12859606742858887,
"learning_rate": 9.091865583449351e-05,
"loss": 0.9279,
"step": 2510
},
{
"epoch": 0.5771658060814687,
"grad_norm": 0.12240682542324066,
"learning_rate": 9.051977397888464e-05,
"loss": 0.9242,
"step": 2515
},
{
"epoch": 0.5783132530120482,
"grad_norm": 0.13055679202079773,
"learning_rate": 9.01210442705514e-05,
"loss": 0.9374,
"step": 2520
},
{
"epoch": 0.5794606999426276,
"grad_norm": 0.1301288604736328,
"learning_rate": 8.972247310867027e-05,
"loss": 0.9403,
"step": 2525
},
{
"epoch": 0.5806081468732072,
"grad_norm": 0.1294988989830017,
"learning_rate": 8.932406688987309e-05,
"loss": 0.9511,
"step": 2530
},
{
"epoch": 0.5817555938037866,
"grad_norm": 0.1301121711730957,
"learning_rate": 8.892583200814466e-05,
"loss": 0.9553,
"step": 2535
},
{
"epoch": 0.582903040734366,
"grad_norm": 0.12820468842983246,
"learning_rate": 8.852777485471997e-05,
"loss": 0.938,
"step": 2540
},
{
"epoch": 0.5840504876649455,
"grad_norm": 0.12704700231552124,
"learning_rate": 8.81299018179817e-05,
"loss": 0.9188,
"step": 2545
},
{
"epoch": 0.5851979345955249,
"grad_norm": 0.12919333577156067,
"learning_rate": 8.773221928335759e-05,
"loss": 0.9601,
"step": 2550
},
{
"epoch": 0.5863453815261044,
"grad_norm": 0.13542434573173523,
"learning_rate": 8.73347336332181e-05,
"loss": 0.9476,
"step": 2555
},
{
"epoch": 0.5874928284566838,
"grad_norm": 0.12929943203926086,
"learning_rate": 8.693745124677386e-05,
"loss": 0.9397,
"step": 2560
},
{
"epoch": 0.5886402753872634,
"grad_norm": 0.130234032869339,
"learning_rate": 8.654037849997342e-05,
"loss": 0.9313,
"step": 2565
},
{
"epoch": 0.5897877223178428,
"grad_norm": 0.13515686988830566,
"learning_rate": 8.614352176540067e-05,
"loss": 0.9503,
"step": 2570
},
{
"epoch": 0.5909351692484223,
"grad_norm": 0.1329420804977417,
"learning_rate": 8.57468874121729e-05,
"loss": 0.9606,
"step": 2575
},
{
"epoch": 0.5920826161790017,
"grad_norm": 0.1323956400156021,
"learning_rate": 8.535048180583838e-05,
"loss": 0.9512,
"step": 2580
},
{
"epoch": 0.5932300631095812,
"grad_norm": 0.13518303632736206,
"learning_rate": 8.495431130827422e-05,
"loss": 0.9377,
"step": 2585
},
{
"epoch": 0.5943775100401606,
"grad_norm": 0.13215835392475128,
"learning_rate": 8.455838227758432e-05,
"loss": 0.9984,
"step": 2590
},
{
"epoch": 0.5955249569707401,
"grad_norm": 0.13005711138248444,
"learning_rate": 8.416270106799726e-05,
"loss": 0.9556,
"step": 2595
},
{
"epoch": 0.5966724039013196,
"grad_norm": 0.1329335868358612,
"learning_rate": 8.376727402976447e-05,
"loss": 0.9351,
"step": 2600
},
{
"epoch": 0.597819850831899,
"grad_norm": 0.1298844963312149,
"learning_rate": 8.3372107509058e-05,
"loss": 0.9492,
"step": 2605
},
{
"epoch": 0.5989672977624785,
"grad_norm": 0.13211120665073395,
"learning_rate": 8.297720784786906e-05,
"loss": 0.9495,
"step": 2610
},
{
"epoch": 0.6001147446930579,
"grad_norm": 0.12725351750850677,
"learning_rate": 8.2582581383906e-05,
"loss": 0.953,
"step": 2615
},
{
"epoch": 0.6012621916236374,
"grad_norm": 0.12866829335689545,
"learning_rate": 8.218823445049265e-05,
"loss": 0.9536,
"step": 2620
},
{
"epoch": 0.6024096385542169,
"grad_norm": 0.1358039677143097,
"learning_rate": 8.179417337646669e-05,
"loss": 0.9731,
"step": 2625
},
{
"epoch": 0.6035570854847964,
"grad_norm": 0.13068200647830963,
"learning_rate": 8.140040448607804e-05,
"loss": 0.9592,
"step": 2630
},
{
"epoch": 0.6047045324153758,
"grad_norm": 0.1265789270401001,
"learning_rate": 8.100693409888748e-05,
"loss": 0.97,
"step": 2635
},
{
"epoch": 0.6058519793459552,
"grad_norm": 0.13472406566143036,
"learning_rate": 8.061376852966495e-05,
"loss": 0.9648,
"step": 2640
},
{
"epoch": 0.6069994262765347,
"grad_norm": 0.13039621710777283,
"learning_rate": 8.02209140882886e-05,
"loss": 0.926,
"step": 2645
},
{
"epoch": 0.6081468732071141,
"grad_norm": 0.1332240104675293,
"learning_rate": 7.982837707964321e-05,
"loss": 0.9637,
"step": 2650
},
{
"epoch": 0.6092943201376936,
"grad_norm": 0.1377446949481964,
"learning_rate": 7.943616380351913e-05,
"loss": 0.9556,
"step": 2655
},
{
"epoch": 0.6104417670682731,
"grad_norm": 0.12704016268253326,
"learning_rate": 7.904428055451118e-05,
"loss": 0.9323,
"step": 2660
},
{
"epoch": 0.6115892139988526,
"grad_norm": 0.12996916472911835,
"learning_rate": 7.865273362191759e-05,
"loss": 0.9376,
"step": 2665
},
{
"epoch": 0.612736660929432,
"grad_norm": 0.13720089197158813,
"learning_rate": 7.826152928963904e-05,
"loss": 0.9469,
"step": 2670
},
{
"epoch": 0.6138841078600115,
"grad_norm": 0.13306473195552826,
"learning_rate": 7.787067383607796e-05,
"loss": 0.9643,
"step": 2675
},
{
"epoch": 0.6150315547905909,
"grad_norm": 0.1308506578207016,
"learning_rate": 7.748017353403748e-05,
"loss": 0.9476,
"step": 2680
},
{
"epoch": 0.6161790017211703,
"grad_norm": 0.12719397246837616,
"learning_rate": 7.70900346506211e-05,
"loss": 0.949,
"step": 2685
},
{
"epoch": 0.6173264486517499,
"grad_norm": 0.12891539931297302,
"learning_rate": 7.670026344713189e-05,
"loss": 0.9486,
"step": 2690
},
{
"epoch": 0.6184738955823293,
"grad_norm": 0.12883388996124268,
"learning_rate": 7.631086617897203e-05,
"loss": 0.9422,
"step": 2695
},
{
"epoch": 0.6196213425129088,
"grad_norm": 0.128887340426445,
"learning_rate": 7.592184909554245e-05,
"loss": 0.9712,
"step": 2700
},
{
"epoch": 0.6207687894434882,
"grad_norm": 0.12608595192432404,
"learning_rate": 7.553321844014258e-05,
"loss": 0.9674,
"step": 2705
},
{
"epoch": 0.6219162363740677,
"grad_norm": 0.12851373851299286,
"learning_rate": 7.514498044987009e-05,
"loss": 0.9229,
"step": 2710
},
{
"epoch": 0.6230636833046471,
"grad_norm": 0.1275022029876709,
"learning_rate": 7.475714135552074e-05,
"loss": 0.9108,
"step": 2715
},
{
"epoch": 0.6242111302352267,
"grad_norm": 0.13182856142520905,
"learning_rate": 7.43697073814885e-05,
"loss": 0.9536,
"step": 2720
},
{
"epoch": 0.6253585771658061,
"grad_norm": 0.12938585877418518,
"learning_rate": 7.39826847456656e-05,
"loss": 0.9425,
"step": 2725
},
{
"epoch": 0.6265060240963856,
"grad_norm": 0.13789427280426025,
"learning_rate": 7.359607965934274e-05,
"loss": 0.9692,
"step": 2730
},
{
"epoch": 0.627653471026965,
"grad_norm": 0.13062715530395508,
"learning_rate": 7.32098983271094e-05,
"loss": 0.9383,
"step": 2735
},
{
"epoch": 0.6288009179575444,
"grad_norm": 0.13577692210674286,
"learning_rate": 7.282414694675426e-05,
"loss": 0.9281,
"step": 2740
},
{
"epoch": 0.6299483648881239,
"grad_norm": 0.13397711515426636,
"learning_rate": 7.243883170916574e-05,
"loss": 0.9881,
"step": 2745
},
{
"epoch": 0.6310958118187033,
"grad_norm": 0.13154800236225128,
"learning_rate": 7.205395879823271e-05,
"loss": 0.9107,
"step": 2750
},
{
"epoch": 0.6322432587492829,
"grad_norm": 0.12707136571407318,
"learning_rate": 7.166953439074504e-05,
"loss": 0.9728,
"step": 2755
},
{
"epoch": 0.6333907056798623,
"grad_norm": 0.13293033838272095,
"learning_rate": 7.128556465629475e-05,
"loss": 0.9652,
"step": 2760
},
{
"epoch": 0.6345381526104418,
"grad_norm": 0.1299448162317276,
"learning_rate": 7.090205575717678e-05,
"loss": 0.9455,
"step": 2765
},
{
"epoch": 0.6356855995410212,
"grad_norm": 0.12795931100845337,
"learning_rate": 7.051901384829021e-05,
"loss": 0.9289,
"step": 2770
},
{
"epoch": 0.6368330464716007,
"grad_norm": 0.13310518860816956,
"learning_rate": 7.013644507703937e-05,
"loss": 0.9203,
"step": 2775
},
{
"epoch": 0.6379804934021801,
"grad_norm": 0.1404452919960022,
"learning_rate": 6.975435558323532e-05,
"loss": 0.9364,
"step": 2780
},
{
"epoch": 0.6391279403327597,
"grad_norm": 0.1289965659379959,
"learning_rate": 6.937275149899725e-05,
"loss": 0.9313,
"step": 2785
},
{
"epoch": 0.6402753872633391,
"grad_norm": 0.12998373806476593,
"learning_rate": 6.899163894865395e-05,
"loss": 0.9575,
"step": 2790
},
{
"epoch": 0.6414228341939185,
"grad_norm": 0.13112759590148926,
"learning_rate": 6.86110240486457e-05,
"loss": 0.9648,
"step": 2795
},
{
"epoch": 0.642570281124498,
"grad_norm": 0.12994885444641113,
"learning_rate": 6.823091290742602e-05,
"loss": 0.961,
"step": 2800
},
{
"epoch": 0.6437177280550774,
"grad_norm": 0.1322287768125534,
"learning_rate": 6.785131162536374e-05,
"loss": 0.9878,
"step": 2805
},
{
"epoch": 0.6448651749856569,
"grad_norm": 0.12845389544963837,
"learning_rate": 6.747222629464484e-05,
"loss": 0.9455,
"step": 2810
},
{
"epoch": 0.6460126219162364,
"grad_norm": 0.13705497980117798,
"learning_rate": 6.709366299917497e-05,
"loss": 0.9537,
"step": 2815
},
{
"epoch": 0.6471600688468159,
"grad_norm": 0.1286800056695938,
"learning_rate": 6.671562781448166e-05,
"loss": 0.9477,
"step": 2820
},
{
"epoch": 0.6483075157773953,
"grad_norm": 0.13222958147525787,
"learning_rate": 6.633812680761684e-05,
"loss": 0.9348,
"step": 2825
},
{
"epoch": 0.6494549627079748,
"grad_norm": 0.12600712478160858,
"learning_rate": 6.59611660370594e-05,
"loss": 0.9627,
"step": 2830
},
{
"epoch": 0.6506024096385542,
"grad_norm": 0.130609929561615,
"learning_rate": 6.558475155261811e-05,
"loss": 0.9506,
"step": 2835
},
{
"epoch": 0.6517498565691336,
"grad_norm": 0.1283220797777176,
"learning_rate": 6.52088893953344e-05,
"loss": 0.9468,
"step": 2840
},
{
"epoch": 0.6528973034997131,
"grad_norm": 0.13281899690628052,
"learning_rate": 6.48335855973855e-05,
"loss": 0.95,
"step": 2845
},
{
"epoch": 0.6540447504302926,
"grad_norm": 0.12522880733013153,
"learning_rate": 6.445884618198754e-05,
"loss": 0.9691,
"step": 2850
},
{
"epoch": 0.6551921973608721,
"grad_norm": 0.13211198151111603,
"learning_rate": 6.408467716329894e-05,
"loss": 0.9636,
"step": 2855
},
{
"epoch": 0.6563396442914515,
"grad_norm": 0.13233083486557007,
"learning_rate": 6.371108454632391e-05,
"loss": 0.9484,
"step": 2860
},
{
"epoch": 0.657487091222031,
"grad_norm": 0.13322043418884277,
"learning_rate": 6.33380743268159e-05,
"loss": 0.925,
"step": 2865
},
{
"epoch": 0.6586345381526104,
"grad_norm": 0.1293569952249527,
"learning_rate": 6.29656524911817e-05,
"loss": 0.9651,
"step": 2870
},
{
"epoch": 0.6597819850831899,
"grad_norm": 0.15199431777000427,
"learning_rate": 6.259382501638509e-05,
"loss": 0.9831,
"step": 2875
},
{
"epoch": 0.6609294320137694,
"grad_norm": 0.1290358006954193,
"learning_rate": 6.222259786985101e-05,
"loss": 0.9603,
"step": 2880
},
{
"epoch": 0.6620768789443489,
"grad_norm": 0.1347620040178299,
"learning_rate": 6.185197700936982e-05,
"loss": 0.9258,
"step": 2885
},
{
"epoch": 0.6632243258749283,
"grad_norm": 0.13369546830654144,
"learning_rate": 6.14819683830016e-05,
"loss": 0.9471,
"step": 2890
},
{
"epoch": 0.6643717728055077,
"grad_norm": 0.1245836541056633,
"learning_rate": 6.111257792898082e-05,
"loss": 0.9626,
"step": 2895
},
{
"epoch": 0.6655192197360872,
"grad_norm": 0.1284221112728119,
"learning_rate": 6.0743811575620846e-05,
"loss": 0.9219,
"step": 2900
},
{
"epoch": 0.6666666666666666,
"grad_norm": 0.1400459110736847,
"learning_rate": 6.0375675241219e-05,
"loss": 0.9495,
"step": 2905
},
{
"epoch": 0.6678141135972461,
"grad_norm": 0.13489268720149994,
"learning_rate": 6.000817483396148e-05,
"loss": 0.9763,
"step": 2910
},
{
"epoch": 0.6689615605278256,
"grad_norm": 0.13535645604133606,
"learning_rate": 5.96413162518285e-05,
"loss": 0.9263,
"step": 2915
},
{
"epoch": 0.6701090074584051,
"grad_norm": 0.1278652399778366,
"learning_rate": 5.9275105382499694e-05,
"loss": 0.9649,
"step": 2920
},
{
"epoch": 0.6712564543889845,
"grad_norm": 0.13380366563796997,
"learning_rate": 5.890954810325966e-05,
"loss": 0.9073,
"step": 2925
},
{
"epoch": 0.672403901319564,
"grad_norm": 0.12865275144577026,
"learning_rate": 5.854465028090355e-05,
"loss": 0.9642,
"step": 2930
},
{
"epoch": 0.6735513482501434,
"grad_norm": 0.13362818956375122,
"learning_rate": 5.8180417771643006e-05,
"loss": 0.9398,
"step": 2935
},
{
"epoch": 0.6746987951807228,
"grad_norm": 0.1288289725780487,
"learning_rate": 5.781685642101196e-05,
"loss": 0.9917,
"step": 2940
},
{
"epoch": 0.6758462421113024,
"grad_norm": 0.1349886655807495,
"learning_rate": 5.7453972063773184e-05,
"loss": 0.9158,
"step": 2945
},
{
"epoch": 0.6769936890418818,
"grad_norm": 0.12944377958774567,
"learning_rate": 5.7091770523824317e-05,
"loss": 0.9787,
"step": 2950
},
{
"epoch": 0.6781411359724613,
"grad_norm": 0.1282043606042862,
"learning_rate": 5.673025761410462e-05,
"loss": 0.9396,
"step": 2955
},
{
"epoch": 0.6792885829030407,
"grad_norm": 0.12658511102199554,
"learning_rate": 5.636943913650147e-05,
"loss": 0.9507,
"step": 2960
},
{
"epoch": 0.6804360298336202,
"grad_norm": 0.13132619857788086,
"learning_rate": 5.60093208817575e-05,
"loss": 0.9011,
"step": 2965
},
{
"epoch": 0.6815834767641996,
"grad_norm": 0.12832701206207275,
"learning_rate": 5.564990862937744e-05,
"loss": 0.9602,
"step": 2970
},
{
"epoch": 0.6827309236947792,
"grad_norm": 0.13031287491321564,
"learning_rate": 5.5291208147535466e-05,
"loss": 0.934,
"step": 2975
},
{
"epoch": 0.6838783706253586,
"grad_norm": 0.1344255656003952,
"learning_rate": 5.4933225192982586e-05,
"loss": 0.9709,
"step": 2980
},
{
"epoch": 0.685025817555938,
"grad_norm": 0.1263609677553177,
"learning_rate": 5.457596551095441e-05,
"loss": 0.9742,
"step": 2985
},
{
"epoch": 0.6861732644865175,
"grad_norm": 0.13262014091014862,
"learning_rate": 5.421943483507863e-05,
"loss": 0.9751,
"step": 2990
},
{
"epoch": 0.6873207114170969,
"grad_norm": 0.12867961823940277,
"learning_rate": 5.3863638887283364e-05,
"loss": 0.9876,
"step": 2995
},
{
"epoch": 0.6884681583476764,
"grad_norm": 0.12955667078495026,
"learning_rate": 5.350858337770498e-05,
"loss": 0.9609,
"step": 3000
},
{
"epoch": 0.6896156052782558,
"grad_norm": 0.13022245466709137,
"learning_rate": 5.315427400459678e-05,
"loss": 0.9605,
"step": 3005
},
{
"epoch": 0.6907630522088354,
"grad_norm": 0.1346377581357956,
"learning_rate": 5.280071645423726e-05,
"loss": 0.9811,
"step": 3010
},
{
"epoch": 0.6919104991394148,
"grad_norm": 0.1265023946762085,
"learning_rate": 5.244791640083906e-05,
"loss": 0.9523,
"step": 3015
},
{
"epoch": 0.6930579460699943,
"grad_norm": 0.12777413427829742,
"learning_rate": 5.2095879506457736e-05,
"loss": 0.9318,
"step": 3020
},
{
"epoch": 0.6942053930005737,
"grad_norm": 0.12877587974071503,
"learning_rate": 5.174461142090111e-05,
"loss": 0.911,
"step": 3025
},
{
"epoch": 0.6953528399311532,
"grad_norm": 0.1300651729106903,
"learning_rate": 5.139411778163832e-05,
"loss": 0.8949,
"step": 3030
},
{
"epoch": 0.6965002868617326,
"grad_norm": 0.13519862294197083,
"learning_rate": 5.104440421370962e-05,
"loss": 0.9456,
"step": 3035
},
{
"epoch": 0.6976477337923122,
"grad_norm": 0.12986309826374054,
"learning_rate": 5.0695476329635825e-05,
"loss": 0.9451,
"step": 3040
},
{
"epoch": 0.6987951807228916,
"grad_norm": 0.12880775332450867,
"learning_rate": 5.034733972932855e-05,
"loss": 0.9474,
"step": 3045
},
{
"epoch": 0.699942627653471,
"grad_norm": 0.13083116710186005,
"learning_rate": 5.000000000000002e-05,
"loss": 0.9568,
"step": 3050
},
{
"epoch": 0.7010900745840505,
"grad_norm": 0.12505587935447693,
"learning_rate": 4.9653462716073594e-05,
"loss": 0.9103,
"step": 3055
},
{
"epoch": 0.7022375215146299,
"grad_norm": 0.12554022669792175,
"learning_rate": 4.930773343909434e-05,
"loss": 0.9495,
"step": 3060
},
{
"epoch": 0.7033849684452094,
"grad_norm": 0.12707488238811493,
"learning_rate": 4.8962817717639555e-05,
"loss": 0.9487,
"step": 3065
},
{
"epoch": 0.7045324153757889,
"grad_norm": 0.127935528755188,
"learning_rate": 4.8618721087230014e-05,
"loss": 0.9821,
"step": 3070
},
{
"epoch": 0.7056798623063684,
"grad_norm": 0.132330060005188,
"learning_rate": 4.8275449070240854e-05,
"loss": 0.9869,
"step": 3075
},
{
"epoch": 0.7068273092369478,
"grad_norm": 0.1311495006084442,
"learning_rate": 4.793300717581308e-05,
"loss": 0.9842,
"step": 3080
},
{
"epoch": 0.7079747561675273,
"grad_norm": 0.13313139975070953,
"learning_rate": 4.7591400899765234e-05,
"loss": 0.9559,
"step": 3085
},
{
"epoch": 0.7091222030981067,
"grad_norm": 0.13822641968727112,
"learning_rate": 4.7250635724505e-05,
"loss": 0.9324,
"step": 3090
},
{
"epoch": 0.7102696500286861,
"grad_norm": 0.13737522065639496,
"learning_rate": 4.6910717118941286e-05,
"loss": 0.9403,
"step": 3095
},
{
"epoch": 0.7114170969592656,
"grad_norm": 0.12806421518325806,
"learning_rate": 4.6571650538396615e-05,
"loss": 0.9148,
"step": 3100
},
{
"epoch": 0.7125645438898451,
"grad_norm": 0.13433298468589783,
"learning_rate": 4.6233441424519295e-05,
"loss": 0.9628,
"step": 3105
},
{
"epoch": 0.7137119908204246,
"grad_norm": 0.13459208607673645,
"learning_rate": 4.5896095205196356e-05,
"loss": 0.9604,
"step": 3110
},
{
"epoch": 0.714859437751004,
"grad_norm": 0.14309802651405334,
"learning_rate": 4.5559617294466176e-05,
"loss": 0.943,
"step": 3115
},
{
"epoch": 0.7160068846815835,
"grad_norm": 0.1299637407064438,
"learning_rate": 4.522401309243193e-05,
"loss": 0.9355,
"step": 3120
},
{
"epoch": 0.7171543316121629,
"grad_norm": 0.12573669850826263,
"learning_rate": 4.488928798517442e-05,
"loss": 0.9422,
"step": 3125
},
{
"epoch": 0.7183017785427424,
"grad_norm": 0.13579024374485016,
"learning_rate": 4.455544734466624e-05,
"loss": 0.9349,
"step": 3130
},
{
"epoch": 0.7194492254733219,
"grad_norm": 0.12755471467971802,
"learning_rate": 4.422249652868506e-05,
"loss": 0.9367,
"step": 3135
},
{
"epoch": 0.7205966724039014,
"grad_norm": 0.12943492829799652,
"learning_rate": 4.389044088072798e-05,
"loss": 0.9717,
"step": 3140
},
{
"epoch": 0.7217441193344808,
"grad_norm": 0.13614961504936218,
"learning_rate": 4.355928572992547e-05,
"loss": 0.9643,
"step": 3145
},
{
"epoch": 0.7228915662650602,
"grad_norm": 0.12964297831058502,
"learning_rate": 4.322903639095619e-05,
"loss": 0.9397,
"step": 3150
},
{
"epoch": 0.7240390131956397,
"grad_norm": 0.12502126395702362,
"learning_rate": 4.289969816396132e-05,
"loss": 0.9278,
"step": 3155
},
{
"epoch": 0.7251864601262191,
"grad_norm": 0.12723030149936676,
"learning_rate": 4.2571276334459895e-05,
"loss": 0.9721,
"step": 3160
},
{
"epoch": 0.7263339070567987,
"grad_norm": 0.12903666496276855,
"learning_rate": 4.224377617326353e-05,
"loss": 0.9325,
"step": 3165
},
{
"epoch": 0.7274813539873781,
"grad_norm": 0.12750981748104095,
"learning_rate": 4.1917202936392265e-05,
"loss": 0.965,
"step": 3170
},
{
"epoch": 0.7286288009179576,
"grad_norm": 0.1284424364566803,
"learning_rate": 4.15915618649899e-05,
"loss": 0.9264,
"step": 3175
},
{
"epoch": 0.729776247848537,
"grad_norm": 0.12181597203016281,
"learning_rate": 4.126685818524013e-05,
"loss": 0.9283,
"step": 3180
},
{
"epoch": 0.7309236947791165,
"grad_norm": 0.13070735335350037,
"learning_rate": 4.094309710828236e-05,
"loss": 0.9313,
"step": 3185
},
{
"epoch": 0.7320711417096959,
"grad_norm": 0.12752236425876617,
"learning_rate": 4.0620283830128414e-05,
"loss": 0.9366,
"step": 3190
},
{
"epoch": 0.7332185886402753,
"grad_norm": 0.2004683017730713,
"learning_rate": 4.029842353157888e-05,
"loss": 0.9166,
"step": 3195
},
{
"epoch": 0.7343660355708549,
"grad_norm": 0.12685176730155945,
"learning_rate": 3.9977521378140084e-05,
"loss": 0.9164,
"step": 3200
},
{
"epoch": 0.7355134825014343,
"grad_norm": 0.1288214921951294,
"learning_rate": 3.965758251994115e-05,
"loss": 0.9292,
"step": 3205
},
{
"epoch": 0.7366609294320138,
"grad_norm": 0.13338162004947662,
"learning_rate": 3.933861209165146e-05,
"loss": 0.9872,
"step": 3210
},
{
"epoch": 0.7378083763625932,
"grad_norm": 0.12942497432231903,
"learning_rate": 3.9020615212398016e-05,
"loss": 0.9797,
"step": 3215
},
{
"epoch": 0.7389558232931727,
"grad_norm": 0.13232511281967163,
"learning_rate": 3.8703596985683556e-05,
"loss": 0.9463,
"step": 3220
},
{
"epoch": 0.7401032702237521,
"grad_norm": 0.13070105016231537,
"learning_rate": 3.838756249930439e-05,
"loss": 0.9215,
"step": 3225
},
{
"epoch": 0.7412507171543317,
"grad_norm": 0.12770217657089233,
"learning_rate": 3.807251682526902e-05,
"loss": 0.9534,
"step": 3230
},
{
"epoch": 0.7423981640849111,
"grad_norm": 0.1268807053565979,
"learning_rate": 3.775846501971636e-05,
"loss": 0.9179,
"step": 3235
},
{
"epoch": 0.7435456110154905,
"grad_norm": 0.12574085593223572,
"learning_rate": 3.7445412122835077e-05,
"loss": 0.9214,
"step": 3240
},
{
"epoch": 0.74469305794607,
"grad_norm": 0.13617047667503357,
"learning_rate": 3.713336315878224e-05,
"loss": 0.9603,
"step": 3245
},
{
"epoch": 0.7458405048766494,
"grad_norm": 0.1332123875617981,
"learning_rate": 3.6822323135603054e-05,
"loss": 0.9363,
"step": 3250
},
{
"epoch": 0.7469879518072289,
"grad_norm": 0.13324113190174103,
"learning_rate": 3.651229704515018e-05,
"loss": 0.9128,
"step": 3255
},
{
"epoch": 0.7481353987378083,
"grad_norm": 0.1259755641222,
"learning_rate": 3.6203289863003905e-05,
"loss": 0.9431,
"step": 3260
},
{
"epoch": 0.7492828456683879,
"grad_norm": 0.12775766849517822,
"learning_rate": 3.5895306548392005e-05,
"loss": 0.9707,
"step": 3265
},
{
"epoch": 0.7504302925989673,
"grad_norm": 0.130545511841774,
"learning_rate": 3.558835204411044e-05,
"loss": 0.946,
"step": 3270
},
{
"epoch": 0.7515777395295468,
"grad_norm": 0.13085825741291046,
"learning_rate": 3.52824312764438e-05,
"loss": 0.9029,
"step": 3275
},
{
"epoch": 0.7527251864601262,
"grad_norm": 0.12603677809238434,
"learning_rate": 3.497754915508632e-05,
"loss": 0.9256,
"step": 3280
},
{
"epoch": 0.7538726333907056,
"grad_norm": 0.12976112961769104,
"learning_rate": 3.467371057306318e-05,
"loss": 0.9324,
"step": 3285
},
{
"epoch": 0.7550200803212851,
"grad_norm": 0.13229593634605408,
"learning_rate": 3.437092040665183e-05,
"loss": 0.9497,
"step": 3290
},
{
"epoch": 0.7561675272518646,
"grad_norm": 0.13209198415279388,
"learning_rate": 3.406918351530376e-05,
"loss": 0.9345,
"step": 3295
},
{
"epoch": 0.7573149741824441,
"grad_norm": 0.1314823031425476,
"learning_rate": 3.3768504741566664e-05,
"loss": 0.9272,
"step": 3300
},
{
"epoch": 0.7584624211130235,
"grad_norm": 0.13039475679397583,
"learning_rate": 3.346888891100649e-05,
"loss": 0.9358,
"step": 3305
},
{
"epoch": 0.759609868043603,
"grad_norm": 0.12750479578971863,
"learning_rate": 3.3170340832130134e-05,
"loss": 0.9454,
"step": 3310
},
{
"epoch": 0.7607573149741824,
"grad_norm": 0.12613527476787567,
"learning_rate": 3.287286529630832e-05,
"loss": 0.9337,
"step": 3315
},
{
"epoch": 0.7619047619047619,
"grad_norm": 0.12861678004264832,
"learning_rate": 3.2576467077698493e-05,
"loss": 0.9045,
"step": 3320
},
{
"epoch": 0.7630522088353414,
"grad_norm": 0.12556509673595428,
"learning_rate": 3.228115093316848e-05,
"loss": 0.9095,
"step": 3325
},
{
"epoch": 0.7641996557659209,
"grad_norm": 0.13398994505405426,
"learning_rate": 3.198692160221987e-05,
"loss": 0.966,
"step": 3330
},
{
"epoch": 0.7653471026965003,
"grad_norm": 0.13073302805423737,
"learning_rate": 3.169378380691218e-05,
"loss": 0.9506,
"step": 3335
},
{
"epoch": 0.7664945496270797,
"grad_norm": 0.13139608502388,
"learning_rate": 3.140174225178692e-05,
"loss": 0.9285,
"step": 3340
},
{
"epoch": 0.7676419965576592,
"grad_norm": 0.13152596354484558,
"learning_rate": 3.111080162379215e-05,
"loss": 0.9519,
"step": 3345
},
{
"epoch": 0.7687894434882386,
"grad_norm": 0.12944857776165009,
"learning_rate": 3.082096659220722e-05,
"loss": 0.9108,
"step": 3350
},
{
"epoch": 0.7699368904188181,
"grad_norm": 0.12854111194610596,
"learning_rate": 3.0532241808567966e-05,
"loss": 0.9315,
"step": 3355
},
{
"epoch": 0.7710843373493976,
"grad_norm": 0.12585926055908203,
"learning_rate": 3.0244631906591825e-05,
"loss": 0.9585,
"step": 3360
},
{
"epoch": 0.7722317842799771,
"grad_norm": 0.12900350987911224,
"learning_rate": 2.9958141502103722e-05,
"loss": 0.9725,
"step": 3365
},
{
"epoch": 0.7733792312105565,
"grad_norm": 0.13046613335609436,
"learning_rate": 2.9672775192961756e-05,
"loss": 0.9677,
"step": 3370
},
{
"epoch": 0.774526678141136,
"grad_norm": 0.1273292899131775,
"learning_rate": 2.938853755898364e-05,
"loss": 0.9494,
"step": 3375
},
{
"epoch": 0.7756741250717154,
"grad_norm": 0.13062521815299988,
"learning_rate": 2.910543316187301e-05,
"loss": 0.9134,
"step": 3380
},
{
"epoch": 0.7768215720022948,
"grad_norm": 0.12806479632854462,
"learning_rate": 2.882346654514627e-05,
"loss": 0.9374,
"step": 3385
},
{
"epoch": 0.7779690189328744,
"grad_norm": 0.1296064406633377,
"learning_rate": 2.8542642234059725e-05,
"loss": 0.9424,
"step": 3390
},
{
"epoch": 0.7791164658634538,
"grad_norm": 0.12994979321956635,
"learning_rate": 2.826296473553697e-05,
"loss": 0.9855,
"step": 3395
},
{
"epoch": 0.7802639127940333,
"grad_norm": 0.13433627784252167,
"learning_rate": 2.7984438538096392e-05,
"loss": 0.8971,
"step": 3400
},
{
"epoch": 0.7814113597246127,
"grad_norm": 0.13279375433921814,
"learning_rate": 2.7707068111779377e-05,
"loss": 0.9418,
"step": 3405
},
{
"epoch": 0.7825588066551922,
"grad_norm": 0.14554573595523834,
"learning_rate": 2.7430857908078345e-05,
"loss": 0.9549,
"step": 3410
},
{
"epoch": 0.7837062535857716,
"grad_norm": 0.12296421080827713,
"learning_rate": 2.7155812359865517e-05,
"loss": 0.9126,
"step": 3415
},
{
"epoch": 0.7848537005163512,
"grad_norm": 0.12731173634529114,
"learning_rate": 2.6881935881321563e-05,
"loss": 0.9698,
"step": 3420
},
{
"epoch": 0.7860011474469306,
"grad_norm": 0.13245446979999542,
"learning_rate": 2.6609232867864896e-05,
"loss": 0.9658,
"step": 3425
},
{
"epoch": 0.7871485943775101,
"grad_norm": 0.1442078948020935,
"learning_rate": 2.6337707696081094e-05,
"loss": 0.9532,
"step": 3430
},
{
"epoch": 0.7882960413080895,
"grad_norm": 0.12503184378147125,
"learning_rate": 2.606736472365272e-05,
"loss": 0.9534,
"step": 3435
},
{
"epoch": 0.789443488238669,
"grad_norm": 0.13164812326431274,
"learning_rate": 2.5798208289289204e-05,
"loss": 0.9451,
"step": 3440
},
{
"epoch": 0.7905909351692484,
"grad_norm": 0.13405010104179382,
"learning_rate": 2.5530242712657492e-05,
"loss": 0.9547,
"step": 3445
},
{
"epoch": 0.7917383820998278,
"grad_norm": 0.13217557966709137,
"learning_rate": 2.526347229431242e-05,
"loss": 0.9647,
"step": 3450
},
{
"epoch": 0.7928858290304074,
"grad_norm": 0.13415786623954773,
"learning_rate": 2.499790131562797e-05,
"loss": 0.9719,
"step": 3455
},
{
"epoch": 0.7940332759609868,
"grad_norm": 0.12753376364707947,
"learning_rate": 2.4733534038728257e-05,
"loss": 0.9217,
"step": 3460
},
{
"epoch": 0.7951807228915663,
"grad_norm": 0.1291799694299698,
"learning_rate": 2.4470374706419485e-05,
"loss": 0.9598,
"step": 3465
},
{
"epoch": 0.7963281698221457,
"grad_norm": 0.1334388107061386,
"learning_rate": 2.4208427542121504e-05,
"loss": 0.9298,
"step": 3470
},
{
"epoch": 0.7974756167527252,
"grad_norm": 0.1315917819738388,
"learning_rate": 2.394769674980035e-05,
"loss": 0.9623,
"step": 3475
},
{
"epoch": 0.7986230636833046,
"grad_norm": 0.1384865790605545,
"learning_rate": 2.3688186513900455e-05,
"loss": 0.9543,
"step": 3480
},
{
"epoch": 0.7997705106138842,
"grad_norm": 0.12887543439865112,
"learning_rate": 2.34299009992778e-05,
"loss": 0.9742,
"step": 3485
},
{
"epoch": 0.8009179575444636,
"grad_norm": 0.12744690477848053,
"learning_rate": 2.317284435113278e-05,
"loss": 0.954,
"step": 3490
},
{
"epoch": 0.802065404475043,
"grad_norm": 0.12910710275173187,
"learning_rate": 2.2917020694944023e-05,
"loss": 0.9107,
"step": 3495
},
{
"epoch": 0.8032128514056225,
"grad_norm": 0.1282537430524826,
"learning_rate": 2.2662434136401722e-05,
"loss": 0.9795,
"step": 3500
},
{
"epoch": 0.8043602983362019,
"grad_norm": 0.13550768792629242,
"learning_rate": 2.2409088761342235e-05,
"loss": 0.9415,
"step": 3505
},
{
"epoch": 0.8055077452667814,
"grad_norm": 0.12883414328098297,
"learning_rate": 2.215698863568213e-05,
"loss": 0.9445,
"step": 3510
},
{
"epoch": 0.8066551921973609,
"grad_norm": 0.13625557720661163,
"learning_rate": 2.1906137805353212e-05,
"loss": 0.9726,
"step": 3515
},
{
"epoch": 0.8078026391279404,
"grad_norm": 0.12835553288459778,
"learning_rate": 2.1656540296237316e-05,
"loss": 0.9235,
"step": 3520
},
{
"epoch": 0.8089500860585198,
"grad_norm": 0.13299131393432617,
"learning_rate": 2.1408200114101985e-05,
"loss": 0.9173,
"step": 3525
},
{
"epoch": 0.8100975329890993,
"grad_norm": 0.13075728714466095,
"learning_rate": 2.116112124453592e-05,
"loss": 0.9613,
"step": 3530
},
{
"epoch": 0.8112449799196787,
"grad_norm": 0.1347227394580841,
"learning_rate": 2.0915307652885164e-05,
"loss": 0.9611,
"step": 3535
},
{
"epoch": 0.8123924268502581,
"grad_norm": 0.13222533464431763,
"learning_rate": 2.067076328418949e-05,
"loss": 0.9492,
"step": 3540
},
{
"epoch": 0.8135398737808376,
"grad_norm": 0.12829838693141937,
"learning_rate": 2.0427492063118935e-05,
"loss": 0.9843,
"step": 3545
},
{
"epoch": 0.8146873207114171,
"grad_norm": 0.13268935680389404,
"learning_rate": 2.018549789391102e-05,
"loss": 0.9879,
"step": 3550
},
{
"epoch": 0.8158347676419966,
"grad_norm": 0.12710270285606384,
"learning_rate": 1.994478466030787e-05,
"loss": 0.8988,
"step": 3555
},
{
"epoch": 0.816982214572576,
"grad_norm": 0.13554465770721436,
"learning_rate": 1.970535622549401e-05,
"loss": 0.948,
"step": 3560
},
{
"epoch": 0.8181296615031555,
"grad_norm": 0.12602022290229797,
"learning_rate": 1.946721643203443e-05,
"loss": 0.9437,
"step": 3565
},
{
"epoch": 0.8192771084337349,
"grad_norm": 0.1312330663204193,
"learning_rate": 1.923036910181275e-05,
"loss": 0.9544,
"step": 3570
},
{
"epoch": 0.8204245553643144,
"grad_norm": 0.12853941321372986,
"learning_rate": 1.8994818035969975e-05,
"loss": 0.9798,
"step": 3575
},
{
"epoch": 0.8215720022948939,
"grad_norm": 0.13181468844413757,
"learning_rate": 1.8760567014843545e-05,
"loss": 0.944,
"step": 3580
},
{
"epoch": 0.8227194492254734,
"grad_norm": 0.12790626287460327,
"learning_rate": 1.8527619797906494e-05,
"loss": 0.901,
"step": 3585
},
{
"epoch": 0.8238668961560528,
"grad_norm": 0.12571856379508972,
"learning_rate": 1.8295980123707357e-05,
"loss": 0.955,
"step": 3590
},
{
"epoch": 0.8250143430866322,
"grad_norm": 0.1319160759449005,
"learning_rate": 1.8065651709809905e-05,
"loss": 0.9046,
"step": 3595
},
{
"epoch": 0.8261617900172117,
"grad_norm": 0.1291453242301941,
"learning_rate": 1.783663825273372e-05,
"loss": 0.9337,
"step": 3600
},
{
"epoch": 0.8273092369477911,
"grad_norm": 0.1331816017627716,
"learning_rate": 1.7608943427894686e-05,
"loss": 0.9563,
"step": 3605
},
{
"epoch": 0.8284566838783706,
"grad_norm": 0.14010034501552582,
"learning_rate": 1.7382570889546124e-05,
"loss": 0.9518,
"step": 3610
},
{
"epoch": 0.8296041308089501,
"grad_norm": 0.1268058568239212,
"learning_rate": 1.7157524270720036e-05,
"loss": 0.912,
"step": 3615
},
{
"epoch": 0.8307515777395296,
"grad_norm": 0.1255045235157013,
"learning_rate": 1.6933807183168994e-05,
"loss": 0.9146,
"step": 3620
},
{
"epoch": 0.831899024670109,
"grad_norm": 0.14171355962753296,
"learning_rate": 1.6711423217307885e-05,
"loss": 0.9269,
"step": 3625
},
{
"epoch": 0.8330464716006885,
"grad_norm": 0.12760847806930542,
"learning_rate": 1.64903759421566e-05,
"loss": 0.9143,
"step": 3630
},
{
"epoch": 0.8341939185312679,
"grad_norm": 0.1354866921901703,
"learning_rate": 1.627066890528247e-05,
"loss": 0.9634,
"step": 3635
},
{
"epoch": 0.8353413654618473,
"grad_norm": 0.1262219399213791,
"learning_rate": 1.6052305632743592e-05,
"loss": 0.9785,
"step": 3640
},
{
"epoch": 0.8364888123924269,
"grad_norm": 0.1341947317123413,
"learning_rate": 1.583528962903197e-05,
"loss": 0.9685,
"step": 3645
},
{
"epoch": 0.8376362593230063,
"grad_norm": 0.12864305078983307,
"learning_rate": 1.5619624377017537e-05,
"loss": 0.9519,
"step": 3650
},
{
"epoch": 0.8387837062535858,
"grad_norm": 0.15601038932800293,
"learning_rate": 1.540531333789207e-05,
"loss": 0.9302,
"step": 3655
},
{
"epoch": 0.8399311531841652,
"grad_norm": 0.13577106595039368,
"learning_rate": 1.5192359951113755e-05,
"loss": 0.9682,
"step": 3660
},
{
"epoch": 0.8410786001147447,
"grad_norm": 0.1369991898536682,
"learning_rate": 1.4980767634351877e-05,
"loss": 0.9267,
"step": 3665
},
{
"epoch": 0.8422260470453241,
"grad_norm": 0.12964287400245667,
"learning_rate": 1.4770539783432113e-05,
"loss": 0.9279,
"step": 3670
},
{
"epoch": 0.8433734939759037,
"grad_norm": 0.1294836550951004,
"learning_rate": 1.4561679772281877e-05,
"loss": 0.9176,
"step": 3675
},
{
"epoch": 0.8445209409064831,
"grad_norm": 0.1307843178510666,
"learning_rate": 1.4354190952876334e-05,
"loss": 0.944,
"step": 3680
},
{
"epoch": 0.8456683878370626,
"grad_norm": 0.1265002340078354,
"learning_rate": 1.4148076655184373e-05,
"loss": 0.9434,
"step": 3685
},
{
"epoch": 0.846815834767642,
"grad_norm": 0.13340280950069427,
"learning_rate": 1.3943340187115494e-05,
"loss": 0.9606,
"step": 3690
},
{
"epoch": 0.8479632816982214,
"grad_norm": 0.12884503602981567,
"learning_rate": 1.373998483446638e-05,
"loss": 0.9117,
"step": 3695
},
{
"epoch": 0.8491107286288009,
"grad_norm": 0.13453038036823273,
"learning_rate": 1.3538013860868436e-05,
"loss": 0.9357,
"step": 3700
},
{
"epoch": 0.8502581755593803,
"grad_norm": 0.12499819695949554,
"learning_rate": 1.3337430507735205e-05,
"loss": 0.949,
"step": 3705
},
{
"epoch": 0.8514056224899599,
"grad_norm": 0.12760350108146667,
"learning_rate": 1.313823799421051e-05,
"loss": 0.9245,
"step": 3710
},
{
"epoch": 0.8525530694205393,
"grad_norm": 0.12545782327651978,
"learning_rate": 1.2940439517116676e-05,
"loss": 0.9163,
"step": 3715
},
{
"epoch": 0.8537005163511188,
"grad_norm": 0.14146262407302856,
"learning_rate": 1.2744038250903267e-05,
"loss": 0.9344,
"step": 3720
},
{
"epoch": 0.8548479632816982,
"grad_norm": 0.1278744488954544,
"learning_rate": 1.2549037347596115e-05,
"loss": 0.9252,
"step": 3725
},
{
"epoch": 0.8559954102122777,
"grad_norm": 0.1335706263780594,
"learning_rate": 1.2355439936746827e-05,
"loss": 0.9753,
"step": 3730
},
{
"epoch": 0.8571428571428571,
"grad_norm": 0.12698203325271606,
"learning_rate": 1.2163249125382426e-05,
"loss": 0.9356,
"step": 3735
},
{
"epoch": 0.8582903040734366,
"grad_norm": 0.129505917429924,
"learning_rate": 1.1972467997955595e-05,
"loss": 0.9424,
"step": 3740
},
{
"epoch": 0.8594377510040161,
"grad_norm": 0.13229629397392273,
"learning_rate": 1.1783099616295056e-05,
"loss": 0.9503,
"step": 3745
},
{
"epoch": 0.8605851979345955,
"grad_norm": 0.13013221323490143,
"learning_rate": 1.1595147019556607e-05,
"loss": 0.9307,
"step": 3750
},
{
"epoch": 0.861732644865175,
"grad_norm": 0.1345113217830658,
"learning_rate": 1.140861322417417e-05,
"loss": 0.9276,
"step": 3755
},
{
"epoch": 0.8628800917957544,
"grad_norm": 0.1310335397720337,
"learning_rate": 1.1223501223811451e-05,
"loss": 0.9732,
"step": 3760
},
{
"epoch": 0.8640275387263339,
"grad_norm": 0.12777724862098694,
"learning_rate": 1.1039813989313951e-05,
"loss": 0.9675,
"step": 3765
},
{
"epoch": 0.8651749856569134,
"grad_norm": 0.13015612959861755,
"learning_rate": 1.085755446866119e-05,
"loss": 0.9165,
"step": 3770
},
{
"epoch": 0.8663224325874929,
"grad_norm": 0.1256059855222702,
"learning_rate": 1.0676725586919457e-05,
"loss": 0.9824,
"step": 3775
},
{
"epoch": 0.8674698795180723,
"grad_norm": 0.13032986223697662,
"learning_rate": 1.0497330246194848e-05,
"loss": 0.9506,
"step": 3780
},
{
"epoch": 0.8686173264486517,
"grad_norm": 0.12943556904792786,
"learning_rate": 1.031937132558668e-05,
"loss": 0.9085,
"step": 3785
},
{
"epoch": 0.8697647733792312,
"grad_norm": 0.13594165444374084,
"learning_rate": 1.014285168114133e-05,
"loss": 0.9401,
"step": 3790
},
{
"epoch": 0.8709122203098106,
"grad_norm": 0.12672476470470428,
"learning_rate": 9.96777414580633e-06,
"loss": 0.9781,
"step": 3795
},
{
"epoch": 0.8720596672403901,
"grad_norm": 0.13382181525230408,
"learning_rate": 9.794141529384915e-06,
"loss": 0.9456,
"step": 3800
},
{
"epoch": 0.8732071141709696,
"grad_norm": 0.12753736972808838,
"learning_rate": 9.621956618491024e-06,
"loss": 0.9561,
"step": 3805
},
{
"epoch": 0.8743545611015491,
"grad_norm": 0.1312447190284729,
"learning_rate": 9.451222176504414e-06,
"loss": 0.966,
"step": 3810
},
{
"epoch": 0.8755020080321285,
"grad_norm": 0.1302998960018158,
"learning_rate": 9.281940943526491e-06,
"loss": 0.936,
"step": 3815
},
{
"epoch": 0.876649454962708,
"grad_norm": 0.12688572704792023,
"learning_rate": 9.114115636336152e-06,
"loss": 0.9623,
"step": 3820
},
{
"epoch": 0.8777969018932874,
"grad_norm": 0.1292993426322937,
"learning_rate": 8.947748948346357e-06,
"loss": 0.9251,
"step": 3825
},
{
"epoch": 0.8789443488238669,
"grad_norm": 0.12286306172609329,
"learning_rate": 8.782843549560771e-06,
"loss": 0.9388,
"step": 3830
},
{
"epoch": 0.8800917957544464,
"grad_norm": 0.12918008863925934,
"learning_rate": 8.61940208653097e-06,
"loss": 0.9499,
"step": 3835
},
{
"epoch": 0.8812392426850258,
"grad_norm": 0.13565728068351746,
"learning_rate": 8.457427182313937e-06,
"loss": 0.9287,
"step": 3840
},
{
"epoch": 0.8823866896156053,
"grad_norm": 0.137693852186203,
"learning_rate": 8.296921436430071e-06,
"loss": 0.966,
"step": 3845
},
{
"epoch": 0.8835341365461847,
"grad_norm": 0.12873981893062592,
"learning_rate": 8.137887424821277e-06,
"loss": 0.9276,
"step": 3850
},
{
"epoch": 0.8846815834767642,
"grad_norm": 0.13087016344070435,
"learning_rate": 7.980327699809832e-06,
"loss": 0.9506,
"step": 3855
},
{
"epoch": 0.8858290304073436,
"grad_norm": 0.13923919200897217,
"learning_rate": 7.824244790057223e-06,
"loss": 0.9048,
"step": 3860
},
{
"epoch": 0.8869764773379232,
"grad_norm": 0.12863433361053467,
"learning_rate": 7.66964120052377e-06,
"loss": 0.9634,
"step": 3865
},
{
"epoch": 0.8881239242685026,
"grad_norm": 0.12779028713703156,
"learning_rate": 7.516519412428203e-06,
"loss": 0.9463,
"step": 3870
},
{
"epoch": 0.8892713711990821,
"grad_norm": 0.13034799695014954,
"learning_rate": 7.3648818832080745e-06,
"loss": 0.979,
"step": 3875
},
{
"epoch": 0.8904188181296615,
"grad_norm": 0.13406170904636383,
"learning_rate": 7.214731046480094e-06,
"loss": 0.9234,
"step": 3880
},
{
"epoch": 0.891566265060241,
"grad_norm": 0.13078339397907257,
"learning_rate": 7.066069312001289e-06,
"loss": 0.9578,
"step": 3885
},
{
"epoch": 0.8927137119908204,
"grad_norm": 0.13037657737731934,
"learning_rate": 6.918899065630113e-06,
"loss": 0.933,
"step": 3890
},
{
"epoch": 0.8938611589213998,
"grad_norm": 0.12971769273281097,
"learning_rate": 6.773222669288359e-06,
"loss": 0.9478,
"step": 3895
},
{
"epoch": 0.8950086058519794,
"grad_norm": 0.13167798519134521,
"learning_rate": 6.629042460923096e-06,
"loss": 0.9555,
"step": 3900
},
{
"epoch": 0.8961560527825588,
"grad_norm": 0.13504882156848907,
"learning_rate": 6.486360754469234e-06,
"loss": 0.9192,
"step": 3905
},
{
"epoch": 0.8973034997131383,
"grad_norm": 0.1290195882320404,
"learning_rate": 6.345179839812343e-06,
"loss": 0.9263,
"step": 3910
},
{
"epoch": 0.8984509466437177,
"grad_norm": 0.127256378531456,
"learning_rate": 6.205501982751971e-06,
"loss": 0.9283,
"step": 3915
},
{
"epoch": 0.8995983935742972,
"grad_norm": 0.1287613958120346,
"learning_rate": 6.067329424965162e-06,
"loss": 0.9485,
"step": 3920
},
{
"epoch": 0.9007458405048766,
"grad_norm": 0.13389694690704346,
"learning_rate": 5.930664383970641e-06,
"loss": 0.9235,
"step": 3925
},
{
"epoch": 0.9018932874354562,
"grad_norm": 0.13568086922168732,
"learning_rate": 5.795509053093029e-06,
"loss": 0.9484,
"step": 3930
},
{
"epoch": 0.9030407343660356,
"grad_norm": 0.12701798975467682,
"learning_rate": 5.6618656014278406e-06,
"loss": 0.9161,
"step": 3935
},
{
"epoch": 0.904188181296615,
"grad_norm": 0.13050422072410583,
"learning_rate": 5.5297361738065325e-06,
"loss": 0.9688,
"step": 3940
},
{
"epoch": 0.9053356282271945,
"grad_norm": 0.13256116211414337,
"learning_rate": 5.399122890762143e-06,
"loss": 0.9124,
"step": 3945
},
{
"epoch": 0.9064830751577739,
"grad_norm": 0.1276332139968872,
"learning_rate": 5.270027848495207e-06,
"loss": 0.9692,
"step": 3950
},
{
"epoch": 0.9076305220883534,
"grad_norm": 0.140411376953125,
"learning_rate": 5.1424531188402405e-06,
"loss": 0.9739,
"step": 3955
},
{
"epoch": 0.9087779690189329,
"grad_norm": 0.12793396413326263,
"learning_rate": 5.016400749232297e-06,
"loss": 0.9482,
"step": 3960
},
{
"epoch": 0.9099254159495124,
"grad_norm": 0.12200927734375,
"learning_rate": 4.89187276267431e-06,
"loss": 0.95,
"step": 3965
},
{
"epoch": 0.9110728628800918,
"grad_norm": 0.13150322437286377,
"learning_rate": 4.7688711577044354e-06,
"loss": 0.9651,
"step": 3970
},
{
"epoch": 0.9122203098106713,
"grad_norm": 0.13082684576511383,
"learning_rate": 4.647397908364182e-06,
"loss": 0.9028,
"step": 3975
},
{
"epoch": 0.9133677567412507,
"grad_norm": 0.12200944870710373,
"learning_rate": 4.5274549641665105e-06,
"loss": 0.9135,
"step": 3980
},
{
"epoch": 0.9145152036718301,
"grad_norm": 0.1254609227180481,
"learning_rate": 4.40904425006472e-06,
"loss": 0.9437,
"step": 3985
},
{
"epoch": 0.9156626506024096,
"grad_norm": 0.12539255619049072,
"learning_rate": 4.2921676664214535e-06,
"loss": 0.8961,
"step": 3990
},
{
"epoch": 0.9168100975329891,
"grad_norm": 0.12549173831939697,
"learning_rate": 4.176827088978297e-06,
"loss": 0.9377,
"step": 3995
},
{
"epoch": 0.9179575444635686,
"grad_norm": 0.1272364854812622,
"learning_rate": 4.0630243688255185e-06,
"loss": 0.952,
"step": 4000
},
{
"epoch": 0.919104991394148,
"grad_norm": 0.12669628858566284,
"learning_rate": 3.950761332372543e-06,
"loss": 0.9593,
"step": 4005
},
{
"epoch": 0.9202524383247275,
"grad_norm": 0.13297000527381897,
"learning_rate": 3.8400397813185054e-06,
"loss": 0.9559,
"step": 4010
},
{
"epoch": 0.9213998852553069,
"grad_norm": 0.1260041445493698,
"learning_rate": 3.7308614926234165e-06,
"loss": 0.9277,
"step": 4015
},
{
"epoch": 0.9225473321858864,
"grad_norm": 0.12770813703536987,
"learning_rate": 3.6232282184795794e-06,
"loss": 0.9448,
"step": 4020
},
{
"epoch": 0.9236947791164659,
"grad_norm": 0.13022121787071228,
"learning_rate": 3.517141686283498e-06,
"loss": 0.9415,
"step": 4025
},
{
"epoch": 0.9248422260470454,
"grad_norm": 0.12520062923431396,
"learning_rate": 3.412603598608188e-06,
"loss": 0.9453,
"step": 4030
},
{
"epoch": 0.9259896729776248,
"grad_norm": 0.12913860380649567,
"learning_rate": 3.3096156331758e-06,
"loss": 0.9492,
"step": 4035
},
{
"epoch": 0.9271371199082042,
"grad_norm": 0.12940272688865662,
"learning_rate": 3.2081794428307278e-06,
"loss": 0.949,
"step": 4040
},
{
"epoch": 0.9282845668387837,
"grad_norm": 0.12925906479358673,
"learning_rate": 3.1082966555130654e-06,
"loss": 0.9143,
"step": 4045
},
{
"epoch": 0.9294320137693631,
"grad_norm": 0.1461576372385025,
"learning_rate": 3.0099688742324715e-06,
"loss": 0.9194,
"step": 4050
},
{
"epoch": 0.9305794606999426,
"grad_norm": 0.1329105645418167,
"learning_rate": 2.913197677042456e-06,
"loss": 0.9007,
"step": 4055
},
{
"epoch": 0.9317269076305221,
"grad_norm": 0.13174231350421906,
"learning_rate": 2.8179846170150903e-06,
"loss": 0.9247,
"step": 4060
},
{
"epoch": 0.9328743545611016,
"grad_norm": 0.12422098964452744,
"learning_rate": 2.7243312222159924e-06,
"loss": 0.9283,
"step": 4065
},
{
"epoch": 0.934021801491681,
"grad_norm": 0.13059760630130768,
"learning_rate": 2.6322389956799143e-06,
"loss": 0.9783,
"step": 4070
},
{
"epoch": 0.9351692484222605,
"grad_norm": 0.1337437480688095,
"learning_rate": 2.541709415386495e-06,
"loss": 0.9643,
"step": 4075
},
{
"epoch": 0.9363166953528399,
"grad_norm": 0.13037900626659393,
"learning_rate": 2.4527439342366785e-06,
"loss": 0.9529,
"step": 4080
},
{
"epoch": 0.9374641422834193,
"grad_norm": 0.13122588396072388,
"learning_rate": 2.3653439800292556e-06,
"loss": 0.9504,
"step": 4085
},
{
"epoch": 0.9386115892139989,
"grad_norm": 0.12561407685279846,
"learning_rate": 2.2795109554381024e-06,
"loss": 0.9457,
"step": 4090
},
{
"epoch": 0.9397590361445783,
"grad_norm": 0.13426658511161804,
"learning_rate": 2.195246237989479e-06,
"loss": 0.9089,
"step": 4095
},
{
"epoch": 0.9409064830751578,
"grad_norm": 0.12589140236377716,
"learning_rate": 2.1125511800401234e-06,
"loss": 0.8939,
"step": 4100
},
{
"epoch": 0.9420539300057372,
"grad_norm": 0.13555549085140228,
"learning_rate": 2.0314271087554126e-06,
"loss": 0.9647,
"step": 4105
},
{
"epoch": 0.9432013769363167,
"grad_norm": 0.1271609365940094,
"learning_rate": 1.951875326088104e-06,
"loss": 0.9331,
"step": 4110
},
{
"epoch": 0.9443488238668961,
"grad_norm": 0.14045216143131256,
"learning_rate": 1.8738971087574275e-06,
"loss": 0.9597,
"step": 4115
},
{
"epoch": 0.9454962707974757,
"grad_norm": 0.12654034793376923,
"learning_rate": 1.797493708228659e-06,
"loss": 0.9267,
"step": 4120
},
{
"epoch": 0.9466437177280551,
"grad_norm": 0.1343478262424469,
"learning_rate": 1.7226663506929142e-06,
"loss": 0.9373,
"step": 4125
},
{
"epoch": 0.9477911646586346,
"grad_norm": 0.1329105794429779,
"learning_rate": 1.6494162370475852e-06,
"loss": 0.9255,
"step": 4130
},
{
"epoch": 0.948938611589214,
"grad_norm": 0.1293703317642212,
"learning_rate": 1.5777445428770022e-06,
"loss": 0.96,
"step": 4135
},
{
"epoch": 0.9500860585197934,
"grad_norm": 0.13163906335830688,
"learning_rate": 1.5076524184336027e-06,
"loss": 0.924,
"step": 4140
},
{
"epoch": 0.9512335054503729,
"grad_norm": 0.12938107550144196,
"learning_rate": 1.4391409886194474e-06,
"loss": 0.9338,
"step": 4145
},
{
"epoch": 0.9523809523809523,
"grad_norm": 0.13095106184482574,
"learning_rate": 1.3722113529681668e-06,
"loss": 0.9538,
"step": 4150
},
{
"epoch": 0.9535283993115319,
"grad_norm": 0.13413451611995697,
"learning_rate": 1.306864585627332e-06,
"loss": 0.9812,
"step": 4155
},
{
"epoch": 0.9546758462421113,
"grad_norm": 0.1287609338760376,
"learning_rate": 1.2431017353412233e-06,
"loss": 0.9374,
"step": 4160
},
{
"epoch": 0.9558232931726908,
"grad_norm": 0.12516798079013824,
"learning_rate": 1.1809238254339105e-06,
"loss": 0.9607,
"step": 4165
},
{
"epoch": 0.9569707401032702,
"grad_norm": 0.13837283849716187,
"learning_rate": 1.1203318537929996e-06,
"loss": 0.9585,
"step": 4170
},
{
"epoch": 0.9581181870338497,
"grad_norm": 0.13161715865135193,
"learning_rate": 1.0613267928534453e-06,
"loss": 0.9289,
"step": 4175
},
{
"epoch": 0.9592656339644291,
"grad_norm": 0.12652412056922913,
"learning_rate": 1.0039095895820639e-06,
"loss": 0.9477,
"step": 4180
},
{
"epoch": 0.9604130808950087,
"grad_norm": 0.12420324981212616,
"learning_rate": 9.480811654622557e-07,
"loss": 0.9562,
"step": 4185
},
{
"epoch": 0.9615605278255881,
"grad_norm": 0.12834912538528442,
"learning_rate": 8.938424164792736e-07,
"loss": 0.9724,
"step": 4190
},
{
"epoch": 0.9627079747561675,
"grad_norm": 0.12531572580337524,
"learning_rate": 8.411942131058115e-07,
"loss": 0.9464,
"step": 4195
},
{
"epoch": 0.963855421686747,
"grad_norm": 0.12882524728775024,
"learning_rate": 7.90137400288049e-07,
"loss": 0.9373,
"step": 4200
},
{
"epoch": 0.9650028686173264,
"grad_norm": 0.13834188878536224,
"learning_rate": 7.406727974320627e-07,
"loss": 0.937,
"step": 4205
},
{
"epoch": 0.9661503155479059,
"grad_norm": 0.13418422639369965,
"learning_rate": 6.928011983907245e-07,
"loss": 0.9551,
"step": 4210
},
{
"epoch": 0.9672977624784854,
"grad_norm": 0.1264306902885437,
"learning_rate": 6.465233714509245e-07,
"loss": 0.9582,
"step": 4215
},
{
"epoch": 0.9684452094090649,
"grad_norm": 0.1280186027288437,
"learning_rate": 6.01840059321257e-07,
"loss": 0.9523,
"step": 4220
},
{
"epoch": 0.9695926563396443,
"grad_norm": 0.12469854950904846,
"learning_rate": 5.587519791200869e-07,
"loss": 0.9511,
"step": 4225
},
{
"epoch": 0.9707401032702238,
"grad_norm": 0.13365860283374786,
"learning_rate": 5.172598223640468e-07,
"loss": 0.9195,
"step": 4230
},
{
"epoch": 0.9718875502008032,
"grad_norm": 0.1352444291114807,
"learning_rate": 4.773642549569579e-07,
"loss": 0.9651,
"step": 4235
},
{
"epoch": 0.9730349971313826,
"grad_norm": 0.12593533098697662,
"learning_rate": 4.390659171790934e-07,
"loss": 0.8991,
"step": 4240
},
{
"epoch": 0.9741824440619621,
"grad_norm": 0.13229170441627502,
"learning_rate": 4.023654236769647e-07,
"loss": 0.9233,
"step": 4245
},
{
"epoch": 0.9753298909925416,
"grad_norm": 0.13042078912258148,
"learning_rate": 3.672633634534295e-07,
"loss": 0.9276,
"step": 4250
},
{
"epoch": 0.9764773379231211,
"grad_norm": 0.12525303661823273,
"learning_rate": 3.3376029985819903e-07,
"loss": 0.9373,
"step": 4255
},
{
"epoch": 0.9776247848537005,
"grad_norm": 0.12787853181362152,
"learning_rate": 3.0185677057887885e-07,
"loss": 0.9313,
"step": 4260
},
{
"epoch": 0.97877223178428,
"grad_norm": 0.12966030836105347,
"learning_rate": 2.715532876322646e-07,
"loss": 0.9288,
"step": 4265
},
{
"epoch": 0.9799196787148594,
"grad_norm": 0.1332397758960724,
"learning_rate": 2.428503373561708e-07,
"loss": 0.9362,
"step": 4270
},
{
"epoch": 0.9810671256454389,
"grad_norm": 0.13498224318027496,
"learning_rate": 2.1574838040161473e-07,
"loss": 0.9411,
"step": 4275
},
{
"epoch": 0.9822145725760184,
"grad_norm": 0.12744614481925964,
"learning_rate": 1.9024785172541136e-07,
"loss": 0.9343,
"step": 4280
},
{
"epoch": 0.9833620195065979,
"grad_norm": 0.1275186389684677,
"learning_rate": 1.6634916058319018e-07,
"loss": 0.9256,
"step": 4285
},
{
"epoch": 0.9845094664371773,
"grad_norm": 0.12919220328330994,
"learning_rate": 1.4405269052284455e-07,
"loss": 0.9333,
"step": 4290
},
{
"epoch": 0.9856569133677567,
"grad_norm": 0.13401302695274353,
"learning_rate": 1.2335879937839246e-07,
"loss": 0.8878,
"step": 4295
},
{
"epoch": 0.9868043602983362,
"grad_norm": 0.1339481770992279,
"learning_rate": 1.0426781926416996e-07,
"loss": 0.9369,
"step": 4300
},
{
"epoch": 0.99,
"step": 4300,
"total_flos": 1.209382783849123e+19,
"train_loss": 0.0,
"train_runtime": 0.0433,
"train_samples_per_second": 31910.696,
"train_steps_per_second": 993.599
}
],
"logging_steps": 5,
"max_steps": 43,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 100,
"total_flos": 1.209382783849123e+19,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}