ATLAS-8B-Instruct / trainer_state.json
aman-jaglan's picture
Add SFT checkpoint
bd3397c verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 5.0,
"eval_steps": 500,
"global_step": 410,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.012195121951219513,
"grad_norm": 20.605318069458008,
"learning_rate": 0.0,
"loss": 1.7928,
"step": 1
},
{
"epoch": 0.024390243902439025,
"grad_norm": 20.76787567138672,
"learning_rate": 2.439024390243903e-07,
"loss": 1.7386,
"step": 2
},
{
"epoch": 0.036585365853658534,
"grad_norm": 21.81036949157715,
"learning_rate": 4.878048780487805e-07,
"loss": 1.8763,
"step": 3
},
{
"epoch": 0.04878048780487805,
"grad_norm": 20.621498107910156,
"learning_rate": 7.317073170731707e-07,
"loss": 1.8537,
"step": 4
},
{
"epoch": 0.06097560975609756,
"grad_norm": 22.98723793029785,
"learning_rate": 9.75609756097561e-07,
"loss": 1.8113,
"step": 5
},
{
"epoch": 0.07317073170731707,
"grad_norm": 19.314804077148438,
"learning_rate": 1.2195121951219514e-06,
"loss": 1.7677,
"step": 6
},
{
"epoch": 0.08536585365853659,
"grad_norm": 21.158281326293945,
"learning_rate": 1.4634146341463414e-06,
"loss": 1.7847,
"step": 7
},
{
"epoch": 0.0975609756097561,
"grad_norm": 16.294034957885742,
"learning_rate": 1.707317073170732e-06,
"loss": 1.6678,
"step": 8
},
{
"epoch": 0.10975609756097561,
"grad_norm": 16.788780212402344,
"learning_rate": 1.951219512195122e-06,
"loss": 1.6558,
"step": 9
},
{
"epoch": 0.12195121951219512,
"grad_norm": 12.225774765014648,
"learning_rate": 2.1951219512195125e-06,
"loss": 1.2956,
"step": 10
},
{
"epoch": 0.13414634146341464,
"grad_norm": 14.551143646240234,
"learning_rate": 2.4390243902439027e-06,
"loss": 1.5254,
"step": 11
},
{
"epoch": 0.14634146341463414,
"grad_norm": 11.28449535369873,
"learning_rate": 2.682926829268293e-06,
"loss": 1.3579,
"step": 12
},
{
"epoch": 0.15853658536585366,
"grad_norm": 7.676495552062988,
"learning_rate": 2.926829268292683e-06,
"loss": 1.2552,
"step": 13
},
{
"epoch": 0.17073170731707318,
"grad_norm": 6.054831027984619,
"learning_rate": 3.1707317073170736e-06,
"loss": 1.0942,
"step": 14
},
{
"epoch": 0.18292682926829268,
"grad_norm": 6.24427604675293,
"learning_rate": 3.414634146341464e-06,
"loss": 1.1486,
"step": 15
},
{
"epoch": 0.1951219512195122,
"grad_norm": 5.555965900421143,
"learning_rate": 3.6585365853658537e-06,
"loss": 1.0225,
"step": 16
},
{
"epoch": 0.2073170731707317,
"grad_norm": 4.953287124633789,
"learning_rate": 3.902439024390244e-06,
"loss": 1.0188,
"step": 17
},
{
"epoch": 0.21951219512195122,
"grad_norm": 4.212824821472168,
"learning_rate": 4.146341463414634e-06,
"loss": 0.9555,
"step": 18
},
{
"epoch": 0.23170731707317074,
"grad_norm": 4.176329135894775,
"learning_rate": 4.390243902439025e-06,
"loss": 0.9,
"step": 19
},
{
"epoch": 0.24390243902439024,
"grad_norm": 4.0246734619140625,
"learning_rate": 4.634146341463416e-06,
"loss": 0.9001,
"step": 20
},
{
"epoch": 0.25609756097560976,
"grad_norm": 4.022885322570801,
"learning_rate": 4.8780487804878055e-06,
"loss": 0.8557,
"step": 21
},
{
"epoch": 0.2682926829268293,
"grad_norm": 3.9502739906311035,
"learning_rate": 5.121951219512195e-06,
"loss": 0.9313,
"step": 22
},
{
"epoch": 0.2804878048780488,
"grad_norm": 3.4761359691619873,
"learning_rate": 5.365853658536586e-06,
"loss": 0.8559,
"step": 23
},
{
"epoch": 0.2926829268292683,
"grad_norm": 3.896311044692993,
"learning_rate": 5.609756097560977e-06,
"loss": 0.9048,
"step": 24
},
{
"epoch": 0.3048780487804878,
"grad_norm": 3.714123010635376,
"learning_rate": 5.853658536585366e-06,
"loss": 0.7699,
"step": 25
},
{
"epoch": 0.3170731707317073,
"grad_norm": 4.503406524658203,
"learning_rate": 6.0975609756097564e-06,
"loss": 0.8699,
"step": 26
},
{
"epoch": 0.32926829268292684,
"grad_norm": 3.643167734146118,
"learning_rate": 6.341463414634147e-06,
"loss": 0.8047,
"step": 27
},
{
"epoch": 0.34146341463414637,
"grad_norm": 3.93937087059021,
"learning_rate": 6.585365853658538e-06,
"loss": 0.8064,
"step": 28
},
{
"epoch": 0.35365853658536583,
"grad_norm": 3.669752836227417,
"learning_rate": 6.829268292682928e-06,
"loss": 0.7593,
"step": 29
},
{
"epoch": 0.36585365853658536,
"grad_norm": 3.5783209800720215,
"learning_rate": 7.0731707317073175e-06,
"loss": 0.7464,
"step": 30
},
{
"epoch": 0.3780487804878049,
"grad_norm": 3.4129626750946045,
"learning_rate": 7.317073170731707e-06,
"loss": 0.8218,
"step": 31
},
{
"epoch": 0.3902439024390244,
"grad_norm": 3.250596761703491,
"learning_rate": 7.560975609756098e-06,
"loss": 0.8161,
"step": 32
},
{
"epoch": 0.4024390243902439,
"grad_norm": 3.030006170272827,
"learning_rate": 7.804878048780489e-06,
"loss": 0.6851,
"step": 33
},
{
"epoch": 0.4146341463414634,
"grad_norm": 3.556096076965332,
"learning_rate": 8.048780487804879e-06,
"loss": 0.8649,
"step": 34
},
{
"epoch": 0.4268292682926829,
"grad_norm": 3.155592203140259,
"learning_rate": 8.292682926829268e-06,
"loss": 0.7146,
"step": 35
},
{
"epoch": 0.43902439024390244,
"grad_norm": 2.923524856567383,
"learning_rate": 8.536585365853658e-06,
"loss": 0.7535,
"step": 36
},
{
"epoch": 0.45121951219512196,
"grad_norm": 3.1197190284729004,
"learning_rate": 8.78048780487805e-06,
"loss": 0.7267,
"step": 37
},
{
"epoch": 0.4634146341463415,
"grad_norm": 2.902597188949585,
"learning_rate": 9.02439024390244e-06,
"loss": 0.7113,
"step": 38
},
{
"epoch": 0.47560975609756095,
"grad_norm": 3.2583975791931152,
"learning_rate": 9.268292682926831e-06,
"loss": 0.8452,
"step": 39
},
{
"epoch": 0.4878048780487805,
"grad_norm": 3.5036613941192627,
"learning_rate": 9.51219512195122e-06,
"loss": 0.7932,
"step": 40
},
{
"epoch": 0.5,
"grad_norm": 2.883305788040161,
"learning_rate": 9.756097560975611e-06,
"loss": 0.7578,
"step": 41
},
{
"epoch": 0.5121951219512195,
"grad_norm": 2.8983325958251953,
"learning_rate": 1e-05,
"loss": 0.6646,
"step": 42
},
{
"epoch": 0.524390243902439,
"grad_norm": 3.0411853790283203,
"learning_rate": 9.999959340292497e-06,
"loss": 0.743,
"step": 43
},
{
"epoch": 0.5365853658536586,
"grad_norm": 3.015455484390259,
"learning_rate": 9.999837361831269e-06,
"loss": 0.6727,
"step": 44
},
{
"epoch": 0.5487804878048781,
"grad_norm": 3.099972724914551,
"learning_rate": 9.999634066600162e-06,
"loss": 0.7748,
"step": 45
},
{
"epoch": 0.5609756097560976,
"grad_norm": 2.834282875061035,
"learning_rate": 9.999349457905545e-06,
"loss": 0.6954,
"step": 46
},
{
"epoch": 0.573170731707317,
"grad_norm": 3.012594223022461,
"learning_rate": 9.998983540376262e-06,
"loss": 0.8249,
"step": 47
},
{
"epoch": 0.5853658536585366,
"grad_norm": 3.121540069580078,
"learning_rate": 9.99853631996355e-06,
"loss": 0.7512,
"step": 48
},
{
"epoch": 0.5975609756097561,
"grad_norm": 2.814594030380249,
"learning_rate": 9.99800780394095e-06,
"loss": 0.749,
"step": 49
},
{
"epoch": 0.6097560975609756,
"grad_norm": 2.8075897693634033,
"learning_rate": 9.997398000904185e-06,
"loss": 0.7249,
"step": 50
},
{
"epoch": 0.6219512195121951,
"grad_norm": 3.2552330493927,
"learning_rate": 9.996706920771024e-06,
"loss": 0.7802,
"step": 51
},
{
"epoch": 0.6341463414634146,
"grad_norm": 3.095428705215454,
"learning_rate": 9.995934574781108e-06,
"loss": 0.753,
"step": 52
},
{
"epoch": 0.6463414634146342,
"grad_norm": 2.9792091846466064,
"learning_rate": 9.995080975495786e-06,
"loss": 0.7911,
"step": 53
},
{
"epoch": 0.6585365853658537,
"grad_norm": 3.0372695922851562,
"learning_rate": 9.994146136797893e-06,
"loss": 0.7471,
"step": 54
},
{
"epoch": 0.6707317073170732,
"grad_norm": 3.14581036567688,
"learning_rate": 9.993130073891539e-06,
"loss": 0.7912,
"step": 55
},
{
"epoch": 0.6829268292682927,
"grad_norm": 2.859478235244751,
"learning_rate": 9.992032803301852e-06,
"loss": 0.6547,
"step": 56
},
{
"epoch": 0.6951219512195121,
"grad_norm": 2.866575002670288,
"learning_rate": 9.990854342874712e-06,
"loss": 0.7098,
"step": 57
},
{
"epoch": 0.7073170731707317,
"grad_norm": 3.036907434463501,
"learning_rate": 9.98959471177646e-06,
"loss": 0.8274,
"step": 58
},
{
"epoch": 0.7195121951219512,
"grad_norm": 2.837873935699463,
"learning_rate": 9.988253930493592e-06,
"loss": 0.7151,
"step": 59
},
{
"epoch": 0.7317073170731707,
"grad_norm": 2.6678829193115234,
"learning_rate": 9.986832020832416e-06,
"loss": 0.6541,
"step": 60
},
{
"epoch": 0.7439024390243902,
"grad_norm": 2.9930105209350586,
"learning_rate": 9.985329005918702e-06,
"loss": 0.6892,
"step": 61
},
{
"epoch": 0.7560975609756098,
"grad_norm": 2.858548164367676,
"learning_rate": 9.983744910197315e-06,
"loss": 0.6988,
"step": 62
},
{
"epoch": 0.7682926829268293,
"grad_norm": 3.0590319633483887,
"learning_rate": 9.982079759431797e-06,
"loss": 0.6853,
"step": 63
},
{
"epoch": 0.7804878048780488,
"grad_norm": 2.8750498294830322,
"learning_rate": 9.980333580703968e-06,
"loss": 0.7181,
"step": 64
},
{
"epoch": 0.7926829268292683,
"grad_norm": 2.720283031463623,
"learning_rate": 9.978506402413472e-06,
"loss": 0.6339,
"step": 65
},
{
"epoch": 0.8048780487804879,
"grad_norm": 2.936540126800537,
"learning_rate": 9.976598254277324e-06,
"loss": 0.7085,
"step": 66
},
{
"epoch": 0.8170731707317073,
"grad_norm": 2.7820205688476562,
"learning_rate": 9.974609167329425e-06,
"loss": 0.6682,
"step": 67
},
{
"epoch": 0.8292682926829268,
"grad_norm": 2.852302312850952,
"learning_rate": 9.972539173920048e-06,
"loss": 0.7067,
"step": 68
},
{
"epoch": 0.8414634146341463,
"grad_norm": 2.763120651245117,
"learning_rate": 9.970388307715326e-06,
"loss": 0.6512,
"step": 69
},
{
"epoch": 0.8536585365853658,
"grad_norm": 2.834955930709839,
"learning_rate": 9.968156603696696e-06,
"loss": 0.692,
"step": 70
},
{
"epoch": 0.8658536585365854,
"grad_norm": 2.5952882766723633,
"learning_rate": 9.965844098160326e-06,
"loss": 0.6458,
"step": 71
},
{
"epoch": 0.8780487804878049,
"grad_norm": 2.793827533721924,
"learning_rate": 9.963450828716543e-06,
"loss": 0.7312,
"step": 72
},
{
"epoch": 0.8902439024390244,
"grad_norm": 2.7760300636291504,
"learning_rate": 9.960976834289197e-06,
"loss": 0.6733,
"step": 73
},
{
"epoch": 0.9024390243902439,
"grad_norm": 3.0652453899383545,
"learning_rate": 9.958422155115044e-06,
"loss": 0.7255,
"step": 74
},
{
"epoch": 0.9146341463414634,
"grad_norm": 2.7409512996673584,
"learning_rate": 9.955786832743089e-06,
"loss": 0.7146,
"step": 75
},
{
"epoch": 0.926829268292683,
"grad_norm": 2.671405553817749,
"learning_rate": 9.953070910033904e-06,
"loss": 0.7051,
"step": 76
},
{
"epoch": 0.9390243902439024,
"grad_norm": 3.065516233444214,
"learning_rate": 9.95027443115894e-06,
"loss": 0.7027,
"step": 77
},
{
"epoch": 0.9512195121951219,
"grad_norm": 2.724518060684204,
"learning_rate": 9.947397441599801e-06,
"loss": 0.7046,
"step": 78
},
{
"epoch": 0.9634146341463414,
"grad_norm": 2.762394428253174,
"learning_rate": 9.944439988147509e-06,
"loss": 0.6638,
"step": 79
},
{
"epoch": 0.975609756097561,
"grad_norm": 2.7874350547790527,
"learning_rate": 9.941402118901743e-06,
"loss": 0.6985,
"step": 80
},
{
"epoch": 0.9878048780487805,
"grad_norm": 2.785700798034668,
"learning_rate": 9.938283883270051e-06,
"loss": 0.6443,
"step": 81
},
{
"epoch": 1.0,
"grad_norm": 2.859963893890381,
"learning_rate": 9.935085331967054e-06,
"loss": 0.6987,
"step": 82
},
{
"epoch": 1.0121951219512195,
"grad_norm": 2.341641902923584,
"learning_rate": 9.931806517013612e-06,
"loss": 0.4309,
"step": 83
},
{
"epoch": 1.024390243902439,
"grad_norm": 2.2350566387176514,
"learning_rate": 9.928447491735994e-06,
"loss": 0.3769,
"step": 84
},
{
"epoch": 1.0365853658536586,
"grad_norm": 2.750514030456543,
"learning_rate": 9.925008310764988e-06,
"loss": 0.5076,
"step": 85
},
{
"epoch": 1.048780487804878,
"grad_norm": 2.627335548400879,
"learning_rate": 9.921489030035036e-06,
"loss": 0.359,
"step": 86
},
{
"epoch": 1.0609756097560976,
"grad_norm": 2.739978075027466,
"learning_rate": 9.917889706783304e-06,
"loss": 0.4735,
"step": 87
},
{
"epoch": 1.0731707317073171,
"grad_norm": 3.0831549167633057,
"learning_rate": 9.914210399548768e-06,
"loss": 0.5604,
"step": 88
},
{
"epoch": 1.0853658536585367,
"grad_norm": 3.0366146564483643,
"learning_rate": 9.910451168171248e-06,
"loss": 0.3986,
"step": 89
},
{
"epoch": 1.0975609756097562,
"grad_norm": 2.8682730197906494,
"learning_rate": 9.906612073790443e-06,
"loss": 0.4118,
"step": 90
},
{
"epoch": 1.1097560975609757,
"grad_norm": 2.9994473457336426,
"learning_rate": 9.902693178844937e-06,
"loss": 0.4581,
"step": 91
},
{
"epoch": 1.1219512195121952,
"grad_norm": 3.4703030586242676,
"learning_rate": 9.898694547071177e-06,
"loss": 0.5222,
"step": 92
},
{
"epoch": 1.1341463414634148,
"grad_norm": 2.6934309005737305,
"learning_rate": 9.894616243502442e-06,
"loss": 0.3656,
"step": 93
},
{
"epoch": 1.146341463414634,
"grad_norm": 2.379758834838867,
"learning_rate": 9.890458334467784e-06,
"loss": 0.3277,
"step": 94
},
{
"epoch": 1.1585365853658536,
"grad_norm": 2.7950727939605713,
"learning_rate": 9.886220887590953e-06,
"loss": 0.4012,
"step": 95
},
{
"epoch": 1.170731707317073,
"grad_norm": 2.668951988220215,
"learning_rate": 9.881903971789285e-06,
"loss": 0.4384,
"step": 96
},
{
"epoch": 1.1829268292682926,
"grad_norm": 2.785778522491455,
"learning_rate": 9.877507657272596e-06,
"loss": 0.4727,
"step": 97
},
{
"epoch": 1.1951219512195121,
"grad_norm": 2.7798571586608887,
"learning_rate": 9.873032015542027e-06,
"loss": 0.4594,
"step": 98
},
{
"epoch": 1.2073170731707317,
"grad_norm": 2.9862515926361084,
"learning_rate": 9.868477119388897e-06,
"loss": 0.4715,
"step": 99
},
{
"epoch": 1.2195121951219512,
"grad_norm": 2.749171495437622,
"learning_rate": 9.863843042893499e-06,
"loss": 0.4125,
"step": 100
},
{
"epoch": 1.2317073170731707,
"grad_norm": 2.4786319732666016,
"learning_rate": 9.859129861423915e-06,
"loss": 0.4036,
"step": 101
},
{
"epoch": 1.2439024390243902,
"grad_norm": 2.724829912185669,
"learning_rate": 9.854337651634773e-06,
"loss": 0.4688,
"step": 102
},
{
"epoch": 1.2560975609756098,
"grad_norm": 2.5419397354125977,
"learning_rate": 9.849466491466017e-06,
"loss": 0.4276,
"step": 103
},
{
"epoch": 1.2682926829268293,
"grad_norm": 2.508129596710205,
"learning_rate": 9.844516460141622e-06,
"loss": 0.401,
"step": 104
},
{
"epoch": 1.2804878048780488,
"grad_norm": 2.677839756011963,
"learning_rate": 9.839487638168321e-06,
"loss": 0.3839,
"step": 105
},
{
"epoch": 1.2926829268292683,
"grad_norm": 2.811065912246704,
"learning_rate": 9.834380107334284e-06,
"loss": 0.3876,
"step": 106
},
{
"epoch": 1.3048780487804879,
"grad_norm": 2.7741312980651855,
"learning_rate": 9.829193950707798e-06,
"loss": 0.4019,
"step": 107
},
{
"epoch": 1.3170731707317074,
"grad_norm": 2.604609727859497,
"learning_rate": 9.823929252635905e-06,
"loss": 0.3753,
"step": 108
},
{
"epoch": 1.329268292682927,
"grad_norm": 3.5267436504364014,
"learning_rate": 9.818586098743038e-06,
"loss": 0.5063,
"step": 109
},
{
"epoch": 1.3414634146341464,
"grad_norm": 2.785386085510254,
"learning_rate": 9.813164575929628e-06,
"loss": 0.4035,
"step": 110
},
{
"epoch": 1.3536585365853657,
"grad_norm": 2.7874209880828857,
"learning_rate": 9.807664772370689e-06,
"loss": 0.4387,
"step": 111
},
{
"epoch": 1.3658536585365852,
"grad_norm": 2.616459369659424,
"learning_rate": 9.80208677751438e-06,
"loss": 0.4403,
"step": 112
},
{
"epoch": 1.3780487804878048,
"grad_norm": 2.593151092529297,
"learning_rate": 9.79643068208056e-06,
"loss": 0.418,
"step": 113
},
{
"epoch": 1.3902439024390243,
"grad_norm": 2.3522331714630127,
"learning_rate": 9.7906965780593e-06,
"loss": 0.3226,
"step": 114
},
{
"epoch": 1.4024390243902438,
"grad_norm": 2.945878028869629,
"learning_rate": 9.784884558709398e-06,
"loss": 0.4744,
"step": 115
},
{
"epoch": 1.4146341463414633,
"grad_norm": 2.6254990100860596,
"learning_rate": 9.778994718556856e-06,
"loss": 0.3656,
"step": 116
},
{
"epoch": 1.4268292682926829,
"grad_norm": 2.6019349098205566,
"learning_rate": 9.773027153393349e-06,
"loss": 0.3957,
"step": 117
},
{
"epoch": 1.4390243902439024,
"grad_norm": 2.8025217056274414,
"learning_rate": 9.766981960274653e-06,
"loss": 0.4242,
"step": 118
},
{
"epoch": 1.451219512195122,
"grad_norm": 2.747736930847168,
"learning_rate": 9.760859237519087e-06,
"loss": 0.4247,
"step": 119
},
{
"epoch": 1.4634146341463414,
"grad_norm": 2.8022918701171875,
"learning_rate": 9.754659084705893e-06,
"loss": 0.3984,
"step": 120
},
{
"epoch": 1.475609756097561,
"grad_norm": 2.5835225582122803,
"learning_rate": 9.748381602673633e-06,
"loss": 0.4205,
"step": 121
},
{
"epoch": 1.4878048780487805,
"grad_norm": 2.7356934547424316,
"learning_rate": 9.742026893518541e-06,
"loss": 0.4098,
"step": 122
},
{
"epoch": 1.5,
"grad_norm": 2.6171412467956543,
"learning_rate": 9.735595060592861e-06,
"loss": 0.4281,
"step": 123
},
{
"epoch": 1.5121951219512195,
"grad_norm": 2.646216630935669,
"learning_rate": 9.729086208503174e-06,
"loss": 0.4301,
"step": 124
},
{
"epoch": 1.524390243902439,
"grad_norm": 3.031221866607666,
"learning_rate": 9.722500443108687e-06,
"loss": 0.5132,
"step": 125
},
{
"epoch": 1.5365853658536586,
"grad_norm": 2.813753843307495,
"learning_rate": 9.715837871519518e-06,
"loss": 0.464,
"step": 126
},
{
"epoch": 1.548780487804878,
"grad_norm": 2.7644271850585938,
"learning_rate": 9.709098602094952e-06,
"loss": 0.4589,
"step": 127
},
{
"epoch": 1.5609756097560976,
"grad_norm": 2.8581771850585938,
"learning_rate": 9.70228274444168e-06,
"loss": 0.4659,
"step": 128
},
{
"epoch": 1.5731707317073171,
"grad_norm": 2.6003692150115967,
"learning_rate": 9.695390409412011e-06,
"loss": 0.3784,
"step": 129
},
{
"epoch": 1.5853658536585367,
"grad_norm": 2.455249547958374,
"learning_rate": 9.688421709102076e-06,
"loss": 0.4141,
"step": 130
},
{
"epoch": 1.5975609756097562,
"grad_norm": 2.439664363861084,
"learning_rate": 9.681376756850003e-06,
"loss": 0.3995,
"step": 131
},
{
"epoch": 1.6097560975609757,
"grad_norm": 2.6555984020233154,
"learning_rate": 9.67425566723407e-06,
"loss": 0.4611,
"step": 132
},
{
"epoch": 1.6219512195121952,
"grad_norm": 2.4294567108154297,
"learning_rate": 9.667058556070846e-06,
"loss": 0.4316,
"step": 133
},
{
"epoch": 1.6341463414634148,
"grad_norm": 2.5822300910949707,
"learning_rate": 9.659785540413303e-06,
"loss": 0.4274,
"step": 134
},
{
"epoch": 1.6463414634146343,
"grad_norm": 2.7250919342041016,
"learning_rate": 9.652436738548917e-06,
"loss": 0.4271,
"step": 135
},
{
"epoch": 1.6585365853658538,
"grad_norm": 2.6819536685943604,
"learning_rate": 9.645012269997747e-06,
"loss": 0.4141,
"step": 136
},
{
"epoch": 1.6707317073170733,
"grad_norm": 2.830106496810913,
"learning_rate": 9.637512255510475e-06,
"loss": 0.466,
"step": 137
},
{
"epoch": 1.6829268292682928,
"grad_norm": 2.6315557956695557,
"learning_rate": 9.629936817066459e-06,
"loss": 0.4085,
"step": 138
},
{
"epoch": 1.6951219512195121,
"grad_norm": 2.916368246078491,
"learning_rate": 9.622286077871748e-06,
"loss": 0.4728,
"step": 139
},
{
"epoch": 1.7073170731707317,
"grad_norm": 3.0268235206604004,
"learning_rate": 9.614560162357065e-06,
"loss": 0.4548,
"step": 140
},
{
"epoch": 1.7195121951219512,
"grad_norm": 2.8294835090637207,
"learning_rate": 9.606759196175799e-06,
"loss": 0.4145,
"step": 141
},
{
"epoch": 1.7317073170731707,
"grad_norm": 2.861173391342163,
"learning_rate": 9.598883306201949e-06,
"loss": 0.4283,
"step": 142
},
{
"epoch": 1.7439024390243902,
"grad_norm": 2.8794517517089844,
"learning_rate": 9.590932620528068e-06,
"loss": 0.5036,
"step": 143
},
{
"epoch": 1.7560975609756098,
"grad_norm": 2.633896589279175,
"learning_rate": 9.58290726846318e-06,
"loss": 0.4355,
"step": 144
},
{
"epoch": 1.7682926829268293,
"grad_norm": 2.5964772701263428,
"learning_rate": 9.57480738053067e-06,
"loss": 0.443,
"step": 145
},
{
"epoch": 1.7804878048780488,
"grad_norm": 2.5255353450775146,
"learning_rate": 9.566633088466169e-06,
"loss": 0.4135,
"step": 146
},
{
"epoch": 1.7926829268292683,
"grad_norm": 2.3389077186584473,
"learning_rate": 9.558384525215406e-06,
"loss": 0.4233,
"step": 147
},
{
"epoch": 1.8048780487804879,
"grad_norm": 2.570801019668579,
"learning_rate": 9.550061824932047e-06,
"loss": 0.4227,
"step": 148
},
{
"epoch": 1.8170731707317072,
"grad_norm": 2.7482798099517822,
"learning_rate": 9.54166512297552e-06,
"loss": 0.4779,
"step": 149
},
{
"epoch": 1.8292682926829267,
"grad_norm": 3.0880026817321777,
"learning_rate": 9.533194555908796e-06,
"loss": 0.5231,
"step": 150
},
{
"epoch": 1.8414634146341462,
"grad_norm": 2.6744909286499023,
"learning_rate": 9.524650261496195e-06,
"loss": 0.4608,
"step": 151
},
{
"epoch": 1.8536585365853657,
"grad_norm": 2.891713857650757,
"learning_rate": 9.516032378701117e-06,
"loss": 0.473,
"step": 152
},
{
"epoch": 1.8658536585365852,
"grad_norm": 2.547239303588867,
"learning_rate": 9.5073410476838e-06,
"loss": 0.4051,
"step": 153
},
{
"epoch": 1.8780487804878048,
"grad_norm": 2.723076581954956,
"learning_rate": 9.498576409799034e-06,
"loss": 0.4558,
"step": 154
},
{
"epoch": 1.8902439024390243,
"grad_norm": 3.1596052646636963,
"learning_rate": 9.489738607593867e-06,
"loss": 0.4865,
"step": 155
},
{
"epoch": 1.9024390243902438,
"grad_norm": 2.7183949947357178,
"learning_rate": 9.480827784805278e-06,
"loss": 0.5014,
"step": 156
},
{
"epoch": 1.9146341463414633,
"grad_norm": 2.5864574909210205,
"learning_rate": 9.471844086357848e-06,
"loss": 0.4149,
"step": 157
},
{
"epoch": 1.9268292682926829,
"grad_norm": 2.5046157836914062,
"learning_rate": 9.462787658361394e-06,
"loss": 0.3962,
"step": 158
},
{
"epoch": 1.9390243902439024,
"grad_norm": 2.8331422805786133,
"learning_rate": 9.453658648108604e-06,
"loss": 0.3853,
"step": 159
},
{
"epoch": 1.951219512195122,
"grad_norm": 2.512298822402954,
"learning_rate": 9.444457204072632e-06,
"loss": 0.4437,
"step": 160
},
{
"epoch": 1.9634146341463414,
"grad_norm": 2.444852828979492,
"learning_rate": 9.435183475904688e-06,
"loss": 0.3504,
"step": 161
},
{
"epoch": 1.975609756097561,
"grad_norm": 2.8331000804901123,
"learning_rate": 9.425837614431601e-06,
"loss": 0.4716,
"step": 162
},
{
"epoch": 1.9878048780487805,
"grad_norm": 2.661059856414795,
"learning_rate": 9.416419771653368e-06,
"loss": 0.4385,
"step": 163
},
{
"epoch": 2.0,
"grad_norm": 2.646305799484253,
"learning_rate": 9.406930100740686e-06,
"loss": 0.4184,
"step": 164
},
{
"epoch": 2.0121951219512195,
"grad_norm": 2.712597608566284,
"learning_rate": 9.397368756032445e-06,
"loss": 0.2314,
"step": 165
},
{
"epoch": 2.024390243902439,
"grad_norm": 2.586576461791992,
"learning_rate": 9.387735893033244e-06,
"loss": 0.1831,
"step": 166
},
{
"epoch": 2.0365853658536586,
"grad_norm": 2.5278258323669434,
"learning_rate": 9.378031668410836e-06,
"loss": 0.2375,
"step": 167
},
{
"epoch": 2.048780487804878,
"grad_norm": 2.541187047958374,
"learning_rate": 9.368256239993597e-06,
"loss": 0.1981,
"step": 168
},
{
"epoch": 2.0609756097560976,
"grad_norm": 2.764477252960205,
"learning_rate": 9.358409766767946e-06,
"loss": 0.2029,
"step": 169
},
{
"epoch": 2.073170731707317,
"grad_norm": 2.4784131050109863,
"learning_rate": 9.348492408875779e-06,
"loss": 0.1535,
"step": 170
},
{
"epoch": 2.0853658536585367,
"grad_norm": 2.915125846862793,
"learning_rate": 9.338504327611839e-06,
"loss": 0.1598,
"step": 171
},
{
"epoch": 2.097560975609756,
"grad_norm": 2.7254488468170166,
"learning_rate": 9.328445685421113e-06,
"loss": 0.1462,
"step": 172
},
{
"epoch": 2.1097560975609757,
"grad_norm": 2.9409985542297363,
"learning_rate": 9.318316645896182e-06,
"loss": 0.203,
"step": 173
},
{
"epoch": 2.1219512195121952,
"grad_norm": 2.588385820388794,
"learning_rate": 9.308117373774555e-06,
"loss": 0.1795,
"step": 174
},
{
"epoch": 2.1341463414634148,
"grad_norm": 2.7931816577911377,
"learning_rate": 9.297848034936007e-06,
"loss": 0.1993,
"step": 175
},
{
"epoch": 2.1463414634146343,
"grad_norm": 2.3102173805236816,
"learning_rate": 9.287508796399858e-06,
"loss": 0.1839,
"step": 176
},
{
"epoch": 2.158536585365854,
"grad_norm": 2.3756439685821533,
"learning_rate": 9.277099826322277e-06,
"loss": 0.2063,
"step": 177
},
{
"epoch": 2.1707317073170733,
"grad_norm": 2.2752017974853516,
"learning_rate": 9.266621293993534e-06,
"loss": 0.1699,
"step": 178
},
{
"epoch": 2.182926829268293,
"grad_norm": 2.484127998352051,
"learning_rate": 9.256073369835255e-06,
"loss": 0.1763,
"step": 179
},
{
"epoch": 2.1951219512195124,
"grad_norm": 2.3598098754882812,
"learning_rate": 9.245456225397642e-06,
"loss": 0.1677,
"step": 180
},
{
"epoch": 2.207317073170732,
"grad_norm": 2.2330524921417236,
"learning_rate": 9.23477003335669e-06,
"loss": 0.185,
"step": 181
},
{
"epoch": 2.2195121951219514,
"grad_norm": 2.439162492752075,
"learning_rate": 9.224014967511378e-06,
"loss": 0.1582,
"step": 182
},
{
"epoch": 2.231707317073171,
"grad_norm": 2.601541042327881,
"learning_rate": 9.213191202780835e-06,
"loss": 0.1737,
"step": 183
},
{
"epoch": 2.2439024390243905,
"grad_norm": 2.3318488597869873,
"learning_rate": 9.20229891520151e-06,
"loss": 0.1688,
"step": 184
},
{
"epoch": 2.2560975609756095,
"grad_norm": 2.883798122406006,
"learning_rate": 9.191338281924288e-06,
"loss": 0.2094,
"step": 185
},
{
"epoch": 2.2682926829268295,
"grad_norm": 2.4024503231048584,
"learning_rate": 9.180309481211629e-06,
"loss": 0.183,
"step": 186
},
{
"epoch": 2.2804878048780486,
"grad_norm": 2.7932958602905273,
"learning_rate": 9.169212692434658e-06,
"loss": 0.2388,
"step": 187
},
{
"epoch": 2.292682926829268,
"grad_norm": 2.345780372619629,
"learning_rate": 9.158048096070249e-06,
"loss": 0.1698,
"step": 188
},
{
"epoch": 2.3048780487804876,
"grad_norm": 2.3633759021759033,
"learning_rate": 9.14681587369809e-06,
"loss": 0.1797,
"step": 189
},
{
"epoch": 2.317073170731707,
"grad_norm": 2.4073266983032227,
"learning_rate": 9.13551620799773e-06,
"loss": 0.1744,
"step": 190
},
{
"epoch": 2.3292682926829267,
"grad_norm": 2.4266092777252197,
"learning_rate": 9.124149282745614e-06,
"loss": 0.1874,
"step": 191
},
{
"epoch": 2.341463414634146,
"grad_norm": 2.277799129486084,
"learning_rate": 9.112715282812081e-06,
"loss": 0.2014,
"step": 192
},
{
"epoch": 2.3536585365853657,
"grad_norm": 2.5907177925109863,
"learning_rate": 9.101214394158371e-06,
"loss": 0.1911,
"step": 193
},
{
"epoch": 2.3658536585365852,
"grad_norm": 2.6057519912719727,
"learning_rate": 9.089646803833589e-06,
"loss": 0.2172,
"step": 194
},
{
"epoch": 2.3780487804878048,
"grad_norm": 2.3195533752441406,
"learning_rate": 9.078012699971673e-06,
"loss": 0.184,
"step": 195
},
{
"epoch": 2.3902439024390243,
"grad_norm": 2.62652850151062,
"learning_rate": 9.066312271788323e-06,
"loss": 0.2185,
"step": 196
},
{
"epoch": 2.402439024390244,
"grad_norm": 2.2538259029388428,
"learning_rate": 9.054545709577939e-06,
"loss": 0.1797,
"step": 197
},
{
"epoch": 2.4146341463414633,
"grad_norm": 2.573920965194702,
"learning_rate": 9.042713204710509e-06,
"loss": 0.1791,
"step": 198
},
{
"epoch": 2.426829268292683,
"grad_norm": 2.010896921157837,
"learning_rate": 9.030814949628509e-06,
"loss": 0.1471,
"step": 199
},
{
"epoch": 2.4390243902439024,
"grad_norm": 2.5009727478027344,
"learning_rate": 9.018851137843765e-06,
"loss": 0.1805,
"step": 200
},
{
"epoch": 2.451219512195122,
"grad_norm": 2.673194169998169,
"learning_rate": 9.006821963934316e-06,
"loss": 0.2134,
"step": 201
},
{
"epoch": 2.4634146341463414,
"grad_norm": 2.851163387298584,
"learning_rate": 8.994727623541237e-06,
"loss": 0.1902,
"step": 202
},
{
"epoch": 2.475609756097561,
"grad_norm": 3.064375877380371,
"learning_rate": 8.982568313365467e-06,
"loss": 0.2247,
"step": 203
},
{
"epoch": 2.4878048780487805,
"grad_norm": 2.5090184211730957,
"learning_rate": 8.970344231164602e-06,
"loss": 0.2022,
"step": 204
},
{
"epoch": 2.5,
"grad_norm": 2.2384963035583496,
"learning_rate": 8.958055575749685e-06,
"loss": 0.1954,
"step": 205
},
{
"epoch": 2.5121951219512195,
"grad_norm": 2.3855085372924805,
"learning_rate": 8.94570254698197e-06,
"loss": 0.2088,
"step": 206
},
{
"epoch": 2.524390243902439,
"grad_norm": 2.38485050201416,
"learning_rate": 8.933285345769671e-06,
"loss": 0.1926,
"step": 207
},
{
"epoch": 2.5365853658536586,
"grad_norm": 2.5828115940093994,
"learning_rate": 8.920804174064697e-06,
"loss": 0.2452,
"step": 208
},
{
"epoch": 2.548780487804878,
"grad_norm": 2.271554470062256,
"learning_rate": 8.908259234859365e-06,
"loss": 0.1858,
"step": 209
},
{
"epoch": 2.5609756097560976,
"grad_norm": 2.114044189453125,
"learning_rate": 8.895650732183094e-06,
"loss": 0.1766,
"step": 210
},
{
"epoch": 2.573170731707317,
"grad_norm": 2.3854148387908936,
"learning_rate": 8.882978871099104e-06,
"loss": 0.2026,
"step": 211
},
{
"epoch": 2.5853658536585367,
"grad_norm": 2.409749746322632,
"learning_rate": 8.870243857701054e-06,
"loss": 0.2135,
"step": 212
},
{
"epoch": 2.597560975609756,
"grad_norm": 2.269014596939087,
"learning_rate": 8.857445899109716e-06,
"loss": 0.173,
"step": 213
},
{
"epoch": 2.6097560975609757,
"grad_norm": 2.1958768367767334,
"learning_rate": 8.84458520346959e-06,
"loss": 0.1803,
"step": 214
},
{
"epoch": 2.6219512195121952,
"grad_norm": 2.2031567096710205,
"learning_rate": 8.831661979945522e-06,
"loss": 0.1701,
"step": 215
},
{
"epoch": 2.6341463414634148,
"grad_norm": 2.523292303085327,
"learning_rate": 8.818676438719314e-06,
"loss": 0.1988,
"step": 216
},
{
"epoch": 2.6463414634146343,
"grad_norm": 2.597362995147705,
"learning_rate": 8.805628790986284e-06,
"loss": 0.2264,
"step": 217
},
{
"epoch": 2.658536585365854,
"grad_norm": 2.802621603012085,
"learning_rate": 8.792519248951851e-06,
"loss": 0.2293,
"step": 218
},
{
"epoch": 2.6707317073170733,
"grad_norm": 2.707906484603882,
"learning_rate": 8.779348025828071e-06,
"loss": 0.2012,
"step": 219
},
{
"epoch": 2.682926829268293,
"grad_norm": 2.630911350250244,
"learning_rate": 8.766115335830178e-06,
"loss": 0.1975,
"step": 220
},
{
"epoch": 2.6951219512195124,
"grad_norm": 2.492384195327759,
"learning_rate": 8.752821394173092e-06,
"loss": 0.1893,
"step": 221
},
{
"epoch": 2.7073170731707314,
"grad_norm": 2.3401095867156982,
"learning_rate": 8.739466417067926e-06,
"loss": 0.1769,
"step": 222
},
{
"epoch": 2.7195121951219514,
"grad_norm": 2.6099853515625,
"learning_rate": 8.726050621718462e-06,
"loss": 0.1746,
"step": 223
},
{
"epoch": 2.7317073170731705,
"grad_norm": 2.4008710384368896,
"learning_rate": 8.71257422631763e-06,
"loss": 0.222,
"step": 224
},
{
"epoch": 2.7439024390243905,
"grad_norm": 2.5295095443725586,
"learning_rate": 8.699037450043945e-06,
"loss": 0.2196,
"step": 225
},
{
"epoch": 2.7560975609756095,
"grad_norm": 2.4341542720794678,
"learning_rate": 8.685440513057955e-06,
"loss": 0.2019,
"step": 226
},
{
"epoch": 2.7682926829268295,
"grad_norm": 2.379326343536377,
"learning_rate": 8.671783636498652e-06,
"loss": 0.2263,
"step": 227
},
{
"epoch": 2.7804878048780486,
"grad_norm": 2.4653515815734863,
"learning_rate": 8.658067042479877e-06,
"loss": 0.197,
"step": 228
},
{
"epoch": 2.7926829268292686,
"grad_norm": 2.4599173069000244,
"learning_rate": 8.644290954086711e-06,
"loss": 0.1995,
"step": 229
},
{
"epoch": 2.8048780487804876,
"grad_norm": 2.559979200363159,
"learning_rate": 8.630455595371846e-06,
"loss": 0.2138,
"step": 230
},
{
"epoch": 2.817073170731707,
"grad_norm": 2.173933267593384,
"learning_rate": 8.616561191351934e-06,
"loss": 0.1822,
"step": 231
},
{
"epoch": 2.8292682926829267,
"grad_norm": 2.4872312545776367,
"learning_rate": 8.602607968003935e-06,
"loss": 0.1805,
"step": 232
},
{
"epoch": 2.841463414634146,
"grad_norm": 2.255208730697632,
"learning_rate": 8.588596152261447e-06,
"loss": 0.1825,
"step": 233
},
{
"epoch": 2.8536585365853657,
"grad_norm": 2.602861166000366,
"learning_rate": 8.574525972010997e-06,
"loss": 0.2079,
"step": 234
},
{
"epoch": 2.8658536585365852,
"grad_norm": 2.6664066314697266,
"learning_rate": 8.560397656088353e-06,
"loss": 0.1909,
"step": 235
},
{
"epoch": 2.8780487804878048,
"grad_norm": 3.140064001083374,
"learning_rate": 8.546211434274791e-06,
"loss": 0.1985,
"step": 236
},
{
"epoch": 2.8902439024390243,
"grad_norm": 2.759251832962036,
"learning_rate": 8.531967537293365e-06,
"loss": 0.1862,
"step": 237
},
{
"epoch": 2.902439024390244,
"grad_norm": 2.6289727687835693,
"learning_rate": 8.517666196805142e-06,
"loss": 0.207,
"step": 238
},
{
"epoch": 2.9146341463414633,
"grad_norm": 2.671435594558716,
"learning_rate": 8.503307645405461e-06,
"loss": 0.2103,
"step": 239
},
{
"epoch": 2.926829268292683,
"grad_norm": 2.4491989612579346,
"learning_rate": 8.488892116620114e-06,
"loss": 0.2086,
"step": 240
},
{
"epoch": 2.9390243902439024,
"grad_norm": 2.1627562046051025,
"learning_rate": 8.474419844901575e-06,
"loss": 0.1785,
"step": 241
},
{
"epoch": 2.951219512195122,
"grad_norm": 2.683394432067871,
"learning_rate": 8.459891065625184e-06,
"loss": 0.2746,
"step": 242
},
{
"epoch": 2.9634146341463414,
"grad_norm": 2.3977043628692627,
"learning_rate": 8.445306015085301e-06,
"loss": 0.2042,
"step": 243
},
{
"epoch": 2.975609756097561,
"grad_norm": 2.0520613193511963,
"learning_rate": 8.430664930491485e-06,
"loss": 0.1897,
"step": 244
},
{
"epoch": 2.9878048780487805,
"grad_norm": 2.36509370803833,
"learning_rate": 8.415968049964623e-06,
"loss": 0.1859,
"step": 245
},
{
"epoch": 3.0,
"grad_norm": 2.167886257171631,
"learning_rate": 8.401215612533056e-06,
"loss": 0.1665,
"step": 246
},
{
"epoch": 3.0121951219512195,
"grad_norm": 1.8608198165893555,
"learning_rate": 8.386407858128707e-06,
"loss": 0.1037,
"step": 247
},
{
"epoch": 3.024390243902439,
"grad_norm": 1.8207582235336304,
"learning_rate": 8.371545027583154e-06,
"loss": 0.0807,
"step": 248
},
{
"epoch": 3.0365853658536586,
"grad_norm": 1.7909525632858276,
"learning_rate": 8.356627362623742e-06,
"loss": 0.0819,
"step": 249
},
{
"epoch": 3.048780487804878,
"grad_norm": 2.130682945251465,
"learning_rate": 8.341655105869622e-06,
"loss": 0.1154,
"step": 250
},
{
"epoch": 3.0609756097560976,
"grad_norm": 1.9704978466033936,
"learning_rate": 8.326628500827826e-06,
"loss": 0.0959,
"step": 251
},
{
"epoch": 3.073170731707317,
"grad_norm": 2.402252197265625,
"learning_rate": 8.311547791889307e-06,
"loss": 0.1006,
"step": 252
},
{
"epoch": 3.0853658536585367,
"grad_norm": 2.2904582023620605,
"learning_rate": 8.296413224324944e-06,
"loss": 0.0985,
"step": 253
},
{
"epoch": 3.097560975609756,
"grad_norm": 2.511240005493164,
"learning_rate": 8.281225044281578e-06,
"loss": 0.0695,
"step": 254
},
{
"epoch": 3.1097560975609757,
"grad_norm": 2.37315034866333,
"learning_rate": 8.265983498777987e-06,
"loss": 0.086,
"step": 255
},
{
"epoch": 3.1219512195121952,
"grad_norm": 2.4025444984436035,
"learning_rate": 8.25068883570089e-06,
"loss": 0.0877,
"step": 256
},
{
"epoch": 3.1341463414634148,
"grad_norm": 2.855544328689575,
"learning_rate": 8.235341303800892e-06,
"loss": 0.1104,
"step": 257
},
{
"epoch": 3.1463414634146343,
"grad_norm": 2.7334654331207275,
"learning_rate": 8.219941152688459e-06,
"loss": 0.0996,
"step": 258
},
{
"epoch": 3.158536585365854,
"grad_norm": 1.7848544120788574,
"learning_rate": 8.204488632829848e-06,
"loss": 0.0779,
"step": 259
},
{
"epoch": 3.1707317073170733,
"grad_norm": 2.5994298458099365,
"learning_rate": 8.188983995543031e-06,
"loss": 0.1027,
"step": 260
},
{
"epoch": 3.182926829268293,
"grad_norm": 2.1597657203674316,
"learning_rate": 8.173427492993617e-06,
"loss": 0.0974,
"step": 261
},
{
"epoch": 3.1951219512195124,
"grad_norm": 2.6595215797424316,
"learning_rate": 8.157819378190743e-06,
"loss": 0.1053,
"step": 262
},
{
"epoch": 3.207317073170732,
"grad_norm": 1.92975652217865,
"learning_rate": 8.142159904982963e-06,
"loss": 0.1003,
"step": 263
},
{
"epoch": 3.2195121951219514,
"grad_norm": 1.939504861831665,
"learning_rate": 8.126449328054115e-06,
"loss": 0.0948,
"step": 264
},
{
"epoch": 3.231707317073171,
"grad_norm": 2.238565444946289,
"learning_rate": 8.110687902919185e-06,
"loss": 0.1021,
"step": 265
},
{
"epoch": 3.2439024390243905,
"grad_norm": 2.1030704975128174,
"learning_rate": 8.094875885920148e-06,
"loss": 0.0961,
"step": 266
},
{
"epoch": 3.2560975609756095,
"grad_norm": 2.0035948753356934,
"learning_rate": 8.079013534221798e-06,
"loss": 0.0985,
"step": 267
},
{
"epoch": 3.2682926829268295,
"grad_norm": 2.1001100540161133,
"learning_rate": 8.063101105807566e-06,
"loss": 0.1089,
"step": 268
},
{
"epoch": 3.2804878048780486,
"grad_norm": 1.935497760772705,
"learning_rate": 8.047138859475328e-06,
"loss": 0.0882,
"step": 269
},
{
"epoch": 3.292682926829268,
"grad_norm": 2.4864578247070312,
"learning_rate": 8.031127054833192e-06,
"loss": 0.1085,
"step": 270
},
{
"epoch": 3.3048780487804876,
"grad_norm": 1.89180326461792,
"learning_rate": 8.01506595229527e-06,
"loss": 0.1096,
"step": 271
},
{
"epoch": 3.317073170731707,
"grad_norm": 2.166079521179199,
"learning_rate": 7.998955813077457e-06,
"loss": 0.0717,
"step": 272
},
{
"epoch": 3.3292682926829267,
"grad_norm": 2.1305079460144043,
"learning_rate": 7.982796899193177e-06,
"loss": 0.1042,
"step": 273
},
{
"epoch": 3.341463414634146,
"grad_norm": 2.0318334102630615,
"learning_rate": 7.966589473449109e-06,
"loss": 0.0943,
"step": 274
},
{
"epoch": 3.3536585365853657,
"grad_norm": 2.6421074867248535,
"learning_rate": 7.95033379944093e-06,
"loss": 0.1161,
"step": 275
},
{
"epoch": 3.3658536585365852,
"grad_norm": 2.3139538764953613,
"learning_rate": 7.934030141549024e-06,
"loss": 0.1219,
"step": 276
},
{
"epoch": 3.3780487804878048,
"grad_norm": 2.0743587017059326,
"learning_rate": 7.917678764934169e-06,
"loss": 0.1024,
"step": 277
},
{
"epoch": 3.3902439024390243,
"grad_norm": 2.187187671661377,
"learning_rate": 7.901279935533248e-06,
"loss": 0.0864,
"step": 278
},
{
"epoch": 3.402439024390244,
"grad_norm": 1.9640257358551025,
"learning_rate": 7.8848339200549e-06,
"loss": 0.0954,
"step": 279
},
{
"epoch": 3.4146341463414633,
"grad_norm": 2.0996806621551514,
"learning_rate": 7.868340985975195e-06,
"loss": 0.0941,
"step": 280
},
{
"epoch": 3.426829268292683,
"grad_norm": 2.0792341232299805,
"learning_rate": 7.851801401533288e-06,
"loss": 0.0908,
"step": 281
},
{
"epoch": 3.4390243902439024,
"grad_norm": 2.0881197452545166,
"learning_rate": 7.835215435727042e-06,
"loss": 0.1059,
"step": 282
},
{
"epoch": 3.451219512195122,
"grad_norm": 2.6827352046966553,
"learning_rate": 7.818583358308664e-06,
"loss": 0.1316,
"step": 283
},
{
"epoch": 3.4634146341463414,
"grad_norm": 2.0524280071258545,
"learning_rate": 7.801905439780317e-06,
"loss": 0.0957,
"step": 284
},
{
"epoch": 3.475609756097561,
"grad_norm": 2.184852361679077,
"learning_rate": 7.785181951389718e-06,
"loss": 0.1123,
"step": 285
},
{
"epoch": 3.4878048780487805,
"grad_norm": 2.2295339107513428,
"learning_rate": 7.76841316512572e-06,
"loss": 0.1198,
"step": 286
},
{
"epoch": 3.5,
"grad_norm": 2.101522922515869,
"learning_rate": 7.751599353713906e-06,
"loss": 0.0991,
"step": 287
},
{
"epoch": 3.5121951219512195,
"grad_norm": 1.8743051290512085,
"learning_rate": 7.734740790612137e-06,
"loss": 0.0869,
"step": 288
},
{
"epoch": 3.524390243902439,
"grad_norm": 1.9927822351455688,
"learning_rate": 7.717837750006106e-06,
"loss": 0.1094,
"step": 289
},
{
"epoch": 3.5365853658536586,
"grad_norm": 2.079759359359741,
"learning_rate": 7.700890506804895e-06,
"loss": 0.1011,
"step": 290
},
{
"epoch": 3.548780487804878,
"grad_norm": 2.3300259113311768,
"learning_rate": 7.68389933663648e-06,
"loss": 0.1374,
"step": 291
},
{
"epoch": 3.5609756097560976,
"grad_norm": 2.1061301231384277,
"learning_rate": 7.666864515843266e-06,
"loss": 0.1123,
"step": 292
},
{
"epoch": 3.573170731707317,
"grad_norm": 1.9325755834579468,
"learning_rate": 7.649786321477585e-06,
"loss": 0.1052,
"step": 293
},
{
"epoch": 3.5853658536585367,
"grad_norm": 2.3022353649139404,
"learning_rate": 7.632665031297193e-06,
"loss": 0.102,
"step": 294
},
{
"epoch": 3.597560975609756,
"grad_norm": 1.8938615322113037,
"learning_rate": 7.615500923760748e-06,
"loss": 0.1065,
"step": 295
},
{
"epoch": 3.6097560975609757,
"grad_norm": 1.8526796102523804,
"learning_rate": 7.59829427802329e-06,
"loss": 0.0971,
"step": 296
},
{
"epoch": 3.6219512195121952,
"grad_norm": 2.010892391204834,
"learning_rate": 7.581045373931691e-06,
"loss": 0.0932,
"step": 297
},
{
"epoch": 3.6341463414634148,
"grad_norm": 2.140416383743286,
"learning_rate": 7.563754492020108e-06,
"loss": 0.0934,
"step": 298
},
{
"epoch": 3.6463414634146343,
"grad_norm": 1.9991627931594849,
"learning_rate": 7.54642191350542e-06,
"loss": 0.1137,
"step": 299
},
{
"epoch": 3.658536585365854,
"grad_norm": 1.98257577419281,
"learning_rate": 7.5290479202826596e-06,
"loss": 0.1058,
"step": 300
},
{
"epoch": 3.6707317073170733,
"grad_norm": 1.9862565994262695,
"learning_rate": 7.511632794920419e-06,
"loss": 0.0977,
"step": 301
},
{
"epoch": 3.682926829268293,
"grad_norm": 2.034688711166382,
"learning_rate": 7.494176820656258e-06,
"loss": 0.1248,
"step": 302
},
{
"epoch": 3.6951219512195124,
"grad_norm": 1.8107631206512451,
"learning_rate": 7.4766802813921016e-06,
"loss": 0.0888,
"step": 303
},
{
"epoch": 3.7073170731707314,
"grad_norm": 1.7797682285308838,
"learning_rate": 7.4591434616896156e-06,
"loss": 0.0971,
"step": 304
},
{
"epoch": 3.7195121951219514,
"grad_norm": 1.8483872413635254,
"learning_rate": 7.4415666467655835e-06,
"loss": 0.1033,
"step": 305
},
{
"epoch": 3.7317073170731705,
"grad_norm": 1.8434807062149048,
"learning_rate": 7.423950122487269e-06,
"loss": 0.0929,
"step": 306
},
{
"epoch": 3.7439024390243905,
"grad_norm": 2.006572961807251,
"learning_rate": 7.406294175367758e-06,
"loss": 0.1034,
"step": 307
},
{
"epoch": 3.7560975609756095,
"grad_norm": 2.015620708465576,
"learning_rate": 7.388599092561315e-06,
"loss": 0.1091,
"step": 308
},
{
"epoch": 3.7682926829268295,
"grad_norm": 2.08795428276062,
"learning_rate": 7.3708651618586925e-06,
"loss": 0.0908,
"step": 309
},
{
"epoch": 3.7804878048780486,
"grad_norm": 2.066549777984619,
"learning_rate": 7.353092671682464e-06,
"loss": 0.093,
"step": 310
},
{
"epoch": 3.7926829268292686,
"grad_norm": 2.227687120437622,
"learning_rate": 7.335281911082332e-06,
"loss": 0.1042,
"step": 311
},
{
"epoch": 3.8048780487804876,
"grad_norm": 2.5046164989471436,
"learning_rate": 7.317433169730421e-06,
"loss": 0.136,
"step": 312
},
{
"epoch": 3.817073170731707,
"grad_norm": 2.0135955810546875,
"learning_rate": 7.299546737916574e-06,
"loss": 0.0942,
"step": 313
},
{
"epoch": 3.8292682926829267,
"grad_norm": 2.3147573471069336,
"learning_rate": 7.281622906543625e-06,
"loss": 0.11,
"step": 314
},
{
"epoch": 3.841463414634146,
"grad_norm": 2.515584707260132,
"learning_rate": 7.26366196712267e-06,
"loss": 0.1248,
"step": 315
},
{
"epoch": 3.8536585365853657,
"grad_norm": 1.988805890083313,
"learning_rate": 7.245664211768327e-06,
"loss": 0.089,
"step": 316
},
{
"epoch": 3.8658536585365852,
"grad_norm": 2.0414860248565674,
"learning_rate": 7.227629933193983e-06,
"loss": 0.0991,
"step": 317
},
{
"epoch": 3.8780487804878048,
"grad_norm": 1.9820183515548706,
"learning_rate": 7.209559424707034e-06,
"loss": 0.1163,
"step": 318
},
{
"epoch": 3.8902439024390243,
"grad_norm": 1.9290958642959595,
"learning_rate": 7.191452980204119e-06,
"loss": 0.1201,
"step": 319
},
{
"epoch": 3.902439024390244,
"grad_norm": 1.9230592250823975,
"learning_rate": 7.173310894166328e-06,
"loss": 0.1138,
"step": 320
},
{
"epoch": 3.9146341463414633,
"grad_norm": 1.6345875263214111,
"learning_rate": 7.155133461654429e-06,
"loss": 0.0935,
"step": 321
},
{
"epoch": 3.926829268292683,
"grad_norm": 1.9335048198699951,
"learning_rate": 7.136920978304056e-06,
"loss": 0.1031,
"step": 322
},
{
"epoch": 3.9390243902439024,
"grad_norm": 1.7330572605133057,
"learning_rate": 7.118673740320907e-06,
"loss": 0.0945,
"step": 323
},
{
"epoch": 3.951219512195122,
"grad_norm": 1.8825818300247192,
"learning_rate": 7.10039204447593e-06,
"loss": 0.0966,
"step": 324
},
{
"epoch": 3.9634146341463414,
"grad_norm": 2.1690921783447266,
"learning_rate": 7.082076188100483e-06,
"loss": 0.1348,
"step": 325
},
{
"epoch": 3.975609756097561,
"grad_norm": 2.1976025104522705,
"learning_rate": 7.063726469081511e-06,
"loss": 0.1046,
"step": 326
},
{
"epoch": 3.9878048780487805,
"grad_norm": 2.0651566982269287,
"learning_rate": 7.045343185856701e-06,
"loss": 0.0848,
"step": 327
},
{
"epoch": 4.0,
"grad_norm": 2.3218235969543457,
"learning_rate": 7.026926637409615e-06,
"loss": 0.1261,
"step": 328
},
{
"epoch": 4.012195121951219,
"grad_norm": 1.517854928970337,
"learning_rate": 7.008477123264849e-06,
"loss": 0.0424,
"step": 329
},
{
"epoch": 4.024390243902439,
"grad_norm": 1.6785979270935059,
"learning_rate": 6.989994943483136e-06,
"loss": 0.053,
"step": 330
},
{
"epoch": 4.036585365853658,
"grad_norm": 1.0940113067626953,
"learning_rate": 6.971480398656488e-06,
"loss": 0.0347,
"step": 331
},
{
"epoch": 4.048780487804878,
"grad_norm": 1.434532880783081,
"learning_rate": 6.952933789903299e-06,
"loss": 0.0468,
"step": 332
},
{
"epoch": 4.060975609756097,
"grad_norm": 1.7367973327636719,
"learning_rate": 6.93435541886344e-06,
"loss": 0.0439,
"step": 333
},
{
"epoch": 4.073170731707317,
"grad_norm": 1.4013808965682983,
"learning_rate": 6.915745587693365e-06,
"loss": 0.0341,
"step": 334
},
{
"epoch": 4.085365853658536,
"grad_norm": 1.7729628086090088,
"learning_rate": 6.89710459906119e-06,
"loss": 0.0524,
"step": 335
},
{
"epoch": 4.097560975609756,
"grad_norm": 1.8550630807876587,
"learning_rate": 6.878432756141775e-06,
"loss": 0.0559,
"step": 336
},
{
"epoch": 4.109756097560975,
"grad_norm": 1.9048420190811157,
"learning_rate": 6.8597303626117886e-06,
"loss": 0.0567,
"step": 337
},
{
"epoch": 4.121951219512195,
"grad_norm": 2.3313469886779785,
"learning_rate": 6.8409977226447685e-06,
"loss": 0.0589,
"step": 338
},
{
"epoch": 4.134146341463414,
"grad_norm": 1.5067005157470703,
"learning_rate": 6.822235140906183e-06,
"loss": 0.0415,
"step": 339
},
{
"epoch": 4.146341463414634,
"grad_norm": 1.7281876802444458,
"learning_rate": 6.803442922548462e-06,
"loss": 0.0491,
"step": 340
},
{
"epoch": 4.158536585365853,
"grad_norm": 1.7764736413955688,
"learning_rate": 6.784621373206051e-06,
"loss": 0.049,
"step": 341
},
{
"epoch": 4.170731707317073,
"grad_norm": 2.0232222080230713,
"learning_rate": 6.765770798990423e-06,
"loss": 0.0524,
"step": 342
},
{
"epoch": 4.182926829268292,
"grad_norm": 1.9550089836120605,
"learning_rate": 6.746891506485112e-06,
"loss": 0.0526,
"step": 343
},
{
"epoch": 4.195121951219512,
"grad_norm": 2.0394773483276367,
"learning_rate": 6.727983802740723e-06,
"loss": 0.0546,
"step": 344
},
{
"epoch": 4.2073170731707314,
"grad_norm": 1.6590560674667358,
"learning_rate": 6.709047995269939e-06,
"loss": 0.0422,
"step": 345
},
{
"epoch": 4.219512195121951,
"grad_norm": 1.8558006286621094,
"learning_rate": 6.690084392042514e-06,
"loss": 0.0518,
"step": 346
},
{
"epoch": 4.2317073170731705,
"grad_norm": 1.2415188550949097,
"learning_rate": 6.671093301480276e-06,
"loss": 0.0333,
"step": 347
},
{
"epoch": 4.2439024390243905,
"grad_norm": 1.7380534410476685,
"learning_rate": 6.6520750324520965e-06,
"loss": 0.0556,
"step": 348
},
{
"epoch": 4.2560975609756095,
"grad_norm": 1.4161667823791504,
"learning_rate": 6.63302989426888e-06,
"loss": 0.0414,
"step": 349
},
{
"epoch": 4.2682926829268295,
"grad_norm": 1.6313724517822266,
"learning_rate": 6.613958196678525e-06,
"loss": 0.0757,
"step": 350
},
{
"epoch": 4.280487804878049,
"grad_norm": 1.9501330852508545,
"learning_rate": 6.594860249860888e-06,
"loss": 0.0675,
"step": 351
},
{
"epoch": 4.2926829268292686,
"grad_norm": 1.5222731828689575,
"learning_rate": 6.575736364422747e-06,
"loss": 0.0537,
"step": 352
},
{
"epoch": 4.304878048780488,
"grad_norm": 1.367255687713623,
"learning_rate": 6.55658685139273e-06,
"loss": 0.0459,
"step": 353
},
{
"epoch": 4.317073170731708,
"grad_norm": 1.4813297986984253,
"learning_rate": 6.5374120222162815e-06,
"loss": 0.06,
"step": 354
},
{
"epoch": 4.329268292682927,
"grad_norm": 1.5068612098693848,
"learning_rate": 6.518212188750579e-06,
"loss": 0.0514,
"step": 355
},
{
"epoch": 4.341463414634147,
"grad_norm": 1.66206955909729,
"learning_rate": 6.498987663259467e-06,
"loss": 0.0675,
"step": 356
},
{
"epoch": 4.353658536585366,
"grad_norm": 1.4990217685699463,
"learning_rate": 6.479738758408379e-06,
"loss": 0.0695,
"step": 357
},
{
"epoch": 4.365853658536586,
"grad_norm": 1.5749341249465942,
"learning_rate": 6.460465787259251e-06,
"loss": 0.0508,
"step": 358
},
{
"epoch": 4.378048780487805,
"grad_norm": 1.499898076057434,
"learning_rate": 6.44116906326543e-06,
"loss": 0.0591,
"step": 359
},
{
"epoch": 4.390243902439025,
"grad_norm": 1.46736478805542,
"learning_rate": 6.421848900266581e-06,
"loss": 0.05,
"step": 360
},
{
"epoch": 4.402439024390244,
"grad_norm": 1.4807460308074951,
"learning_rate": 6.402505612483569e-06,
"loss": 0.0523,
"step": 361
},
{
"epoch": 4.414634146341464,
"grad_norm": 1.4587833881378174,
"learning_rate": 6.383139514513368e-06,
"loss": 0.0576,
"step": 362
},
{
"epoch": 4.426829268292683,
"grad_norm": 1.4291479587554932,
"learning_rate": 6.363750921323929e-06,
"loss": 0.0479,
"step": 363
},
{
"epoch": 4.439024390243903,
"grad_norm": 1.364157795906067,
"learning_rate": 6.3443401482490615e-06,
"loss": 0.0528,
"step": 364
},
{
"epoch": 4.451219512195122,
"grad_norm": 2.088580369949341,
"learning_rate": 6.32490751098331e-06,
"loss": 0.0605,
"step": 365
},
{
"epoch": 4.463414634146342,
"grad_norm": 1.5994398593902588,
"learning_rate": 6.30545332557681e-06,
"loss": 0.0525,
"step": 366
},
{
"epoch": 4.475609756097561,
"grad_norm": 1.7937228679656982,
"learning_rate": 6.2859779084301584e-06,
"loss": 0.0517,
"step": 367
},
{
"epoch": 4.487804878048781,
"grad_norm": 1.3765718936920166,
"learning_rate": 6.266481576289263e-06,
"loss": 0.041,
"step": 368
},
{
"epoch": 4.5,
"grad_norm": 1.7616742849349976,
"learning_rate": 6.246964646240186e-06,
"loss": 0.0715,
"step": 369
},
{
"epoch": 4.512195121951219,
"grad_norm": 1.496747374534607,
"learning_rate": 6.227427435703997e-06,
"loss": 0.0633,
"step": 370
},
{
"epoch": 4.524390243902439,
"grad_norm": 1.53587007522583,
"learning_rate": 6.207870262431599e-06,
"loss": 0.0557,
"step": 371
},
{
"epoch": 4.536585365853659,
"grad_norm": 1.664995789527893,
"learning_rate": 6.188293444498573e-06,
"loss": 0.0599,
"step": 372
},
{
"epoch": 4.548780487804878,
"grad_norm": 1.8567813634872437,
"learning_rate": 6.1686973002999935e-06,
"loss": 0.0643,
"step": 373
},
{
"epoch": 4.560975609756097,
"grad_norm": 2.01507568359375,
"learning_rate": 6.149082148545258e-06,
"loss": 0.0637,
"step": 374
},
{
"epoch": 4.573170731707317,
"grad_norm": 1.800641417503357,
"learning_rate": 6.129448308252899e-06,
"loss": 0.0587,
"step": 375
},
{
"epoch": 4.585365853658536,
"grad_norm": 2.0126662254333496,
"learning_rate": 6.109796098745398e-06,
"loss": 0.0669,
"step": 376
},
{
"epoch": 4.597560975609756,
"grad_norm": 1.8245577812194824,
"learning_rate": 6.090125839643991e-06,
"loss": 0.0541,
"step": 377
},
{
"epoch": 4.609756097560975,
"grad_norm": 1.3531700372695923,
"learning_rate": 6.070437850863472e-06,
"loss": 0.0445,
"step": 378
},
{
"epoch": 4.621951219512195,
"grad_norm": 1.9308772087097168,
"learning_rate": 6.0507324526069854e-06,
"loss": 0.0608,
"step": 379
},
{
"epoch": 4.634146341463414,
"grad_norm": 1.5027072429656982,
"learning_rate": 6.031009965360824e-06,
"loss": 0.0634,
"step": 380
},
{
"epoch": 4.646341463414634,
"grad_norm": 1.3451308012008667,
"learning_rate": 6.011270709889213e-06,
"loss": 0.0411,
"step": 381
},
{
"epoch": 4.658536585365853,
"grad_norm": 1.618082046508789,
"learning_rate": 5.991515007229093e-06,
"loss": 0.0575,
"step": 382
},
{
"epoch": 4.670731707317073,
"grad_norm": 1.6030172109603882,
"learning_rate": 5.971743178684901e-06,
"loss": 0.0575,
"step": 383
},
{
"epoch": 4.682926829268292,
"grad_norm": 1.582740306854248,
"learning_rate": 5.951955545823342e-06,
"loss": 0.0613,
"step": 384
},
{
"epoch": 4.695121951219512,
"grad_norm": 1.7536263465881348,
"learning_rate": 5.932152430468165e-06,
"loss": 0.052,
"step": 385
},
{
"epoch": 4.7073170731707314,
"grad_norm": 2.1995296478271484,
"learning_rate": 5.912334154694919e-06,
"loss": 0.0629,
"step": 386
},
{
"epoch": 4.719512195121951,
"grad_norm": 1.8581688404083252,
"learning_rate": 5.892501040825721e-06,
"loss": 0.041,
"step": 387
},
{
"epoch": 4.7317073170731705,
"grad_norm": 1.8024824857711792,
"learning_rate": 5.872653411424017e-06,
"loss": 0.0708,
"step": 388
},
{
"epoch": 4.7439024390243905,
"grad_norm": 1.7822990417480469,
"learning_rate": 5.85279158928933e-06,
"loss": 0.0528,
"step": 389
},
{
"epoch": 4.7560975609756095,
"grad_norm": 1.9106731414794922,
"learning_rate": 5.832915897452008e-06,
"loss": 0.0643,
"step": 390
},
{
"epoch": 4.7682926829268295,
"grad_norm": 1.593004584312439,
"learning_rate": 5.813026659167982e-06,
"loss": 0.054,
"step": 391
},
{
"epoch": 4.780487804878049,
"grad_norm": 1.8973208665847778,
"learning_rate": 5.793124197913492e-06,
"loss": 0.0737,
"step": 392
},
{
"epoch": 4.7926829268292686,
"grad_norm": 1.9966886043548584,
"learning_rate": 5.773208837379843e-06,
"loss": 0.0634,
"step": 393
},
{
"epoch": 4.804878048780488,
"grad_norm": 1.5227646827697754,
"learning_rate": 5.753280901468126e-06,
"loss": 0.0496,
"step": 394
},
{
"epoch": 4.817073170731708,
"grad_norm": 1.6435083150863647,
"learning_rate": 5.733340714283959e-06,
"loss": 0.0664,
"step": 395
},
{
"epoch": 4.829268292682927,
"grad_norm": 1.3312773704528809,
"learning_rate": 5.713388600132217e-06,
"loss": 0.0534,
"step": 396
},
{
"epoch": 4.841463414634147,
"grad_norm": 1.868194580078125,
"learning_rate": 5.693424883511748e-06,
"loss": 0.0565,
"step": 397
},
{
"epoch": 4.853658536585366,
"grad_norm": 1.5551823377609253,
"learning_rate": 5.6734498891101005e-06,
"loss": 0.0604,
"step": 398
},
{
"epoch": 4.865853658536586,
"grad_norm": 1.8578870296478271,
"learning_rate": 5.653463941798252e-06,
"loss": 0.0728,
"step": 399
},
{
"epoch": 4.878048780487805,
"grad_norm": 1.5294170379638672,
"learning_rate": 5.633467366625306e-06,
"loss": 0.0637,
"step": 400
},
{
"epoch": 4.890243902439025,
"grad_norm": 1.2593622207641602,
"learning_rate": 5.613460488813225e-06,
"loss": 0.0512,
"step": 401
},
{
"epoch": 4.902439024390244,
"grad_norm": 1.7771371603012085,
"learning_rate": 5.593443633751527e-06,
"loss": 0.0658,
"step": 402
},
{
"epoch": 4.914634146341464,
"grad_norm": 1.5825587511062622,
"learning_rate": 5.573417126992004e-06,
"loss": 0.0671,
"step": 403
},
{
"epoch": 4.926829268292683,
"grad_norm": 1.6244094371795654,
"learning_rate": 5.553381294243413e-06,
"loss": 0.0585,
"step": 404
},
{
"epoch": 4.939024390243903,
"grad_norm": 1.501323938369751,
"learning_rate": 5.5333364613662e-06,
"loss": 0.0578,
"step": 405
},
{
"epoch": 4.951219512195122,
"grad_norm": 1.5930196046829224,
"learning_rate": 5.513282954367179e-06,
"loss": 0.064,
"step": 406
},
{
"epoch": 4.963414634146341,
"grad_norm": 1.4195719957351685,
"learning_rate": 5.493221099394239e-06,
"loss": 0.0443,
"step": 407
},
{
"epoch": 4.975609756097561,
"grad_norm": 1.3484866619110107,
"learning_rate": 5.473151222731044e-06,
"loss": 0.0577,
"step": 408
},
{
"epoch": 4.987804878048781,
"grad_norm": 1.677027940750122,
"learning_rate": 5.453073650791724e-06,
"loss": 0.0604,
"step": 409
},
{
"epoch": 5.0,
"grad_norm": 1.7022733688354492,
"learning_rate": 5.432988710115553e-06,
"loss": 0.0674,
"step": 410
}
],
"logging_steps": 1,
"max_steps": 820,
"num_input_tokens_seen": 0,
"num_train_epochs": 10,
"save_steps": 1,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 7614453848064.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}