|
{ |
|
"best_global_step": null, |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.9908256880733946, |
|
"eval_steps": 98, |
|
"global_step": 972, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0010193679918450561, |
|
"grad_norm": 24.08992576599121, |
|
"learning_rate": 0.0, |
|
"loss": 8.5798, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0020387359836901123, |
|
"grad_norm": 22.663619995117188, |
|
"learning_rate": 2.2624434389140275e-07, |
|
"loss": 8.4222, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.0030581039755351682, |
|
"grad_norm": 25.377544403076172, |
|
"learning_rate": 4.524886877828055e-07, |
|
"loss": 9.392, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.004077471967380225, |
|
"grad_norm": 22.122257232666016, |
|
"learning_rate": 6.787330316742082e-07, |
|
"loss": 8.4193, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.0050968399592252805, |
|
"grad_norm": Infinity, |
|
"learning_rate": 9.04977375565611e-07, |
|
"loss": 8.808, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.0061162079510703364, |
|
"grad_norm": 22.832090377807617, |
|
"learning_rate": 9.04977375565611e-07, |
|
"loss": 10.5002, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.007135575942915392, |
|
"grad_norm": 20.933177947998047, |
|
"learning_rate": 1.1312217194570136e-06, |
|
"loss": 9.5956, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.00815494393476045, |
|
"grad_norm": 20.44132423400879, |
|
"learning_rate": 1.3574660633484164e-06, |
|
"loss": 8.5526, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.009174311926605505, |
|
"grad_norm": 22.630067825317383, |
|
"learning_rate": 1.583710407239819e-06, |
|
"loss": 9.8255, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.010193679918450561, |
|
"grad_norm": 22.625064849853516, |
|
"learning_rate": 1.809954751131222e-06, |
|
"loss": 9.2067, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.011213047910295617, |
|
"grad_norm": 19.427106857299805, |
|
"learning_rate": 2.0361990950226245e-06, |
|
"loss": 8.667, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.012232415902140673, |
|
"grad_norm": 24.354387283325195, |
|
"learning_rate": 2.2624434389140273e-06, |
|
"loss": 10.396, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.013251783893985729, |
|
"grad_norm": 21.99860191345215, |
|
"learning_rate": 2.48868778280543e-06, |
|
"loss": 8.9036, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.014271151885830785, |
|
"grad_norm": 21.275592803955078, |
|
"learning_rate": 2.7149321266968327e-06, |
|
"loss": 7.9307, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.01529051987767584, |
|
"grad_norm": 20.020435333251953, |
|
"learning_rate": 2.9411764705882355e-06, |
|
"loss": 9.0655, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.0163098878695209, |
|
"grad_norm": 20.713603973388672, |
|
"learning_rate": 3.167420814479638e-06, |
|
"loss": 8.6598, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.017329255861365953, |
|
"grad_norm": 22.857194900512695, |
|
"learning_rate": 3.3936651583710405e-06, |
|
"loss": 9.7562, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.01834862385321101, |
|
"grad_norm": 19.031551361083984, |
|
"learning_rate": 3.619909502262444e-06, |
|
"loss": 9.2297, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.019367991845056064, |
|
"grad_norm": 19.30624008178711, |
|
"learning_rate": 3.846153846153847e-06, |
|
"loss": 8.6939, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.020387359836901122, |
|
"grad_norm": 17.09296417236328, |
|
"learning_rate": 4.072398190045249e-06, |
|
"loss": 8.1317, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.021406727828746176, |
|
"grad_norm": 19.199600219726562, |
|
"learning_rate": 4.298642533936651e-06, |
|
"loss": 8.3585, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.022426095820591234, |
|
"grad_norm": 18.50484275817871, |
|
"learning_rate": 4.5248868778280546e-06, |
|
"loss": 8.4533, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.023445463812436288, |
|
"grad_norm": 19.170618057250977, |
|
"learning_rate": 4.751131221719457e-06, |
|
"loss": 9.3014, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.024464831804281346, |
|
"grad_norm": 17.692346572875977, |
|
"learning_rate": 4.97737556561086e-06, |
|
"loss": 8.18, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.0254841997961264, |
|
"grad_norm": 18.87356185913086, |
|
"learning_rate": 5.203619909502263e-06, |
|
"loss": 7.8485, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.026503567787971458, |
|
"grad_norm": 16.432092666625977, |
|
"learning_rate": 5.4298642533936655e-06, |
|
"loss": 8.9669, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.027522935779816515, |
|
"grad_norm": 17.064382553100586, |
|
"learning_rate": 5.656108597285068e-06, |
|
"loss": 10.1397, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.02854230377166157, |
|
"grad_norm": 17.96854591369629, |
|
"learning_rate": 5.882352941176471e-06, |
|
"loss": 10.5216, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.029561671763506627, |
|
"grad_norm": 16.348352432250977, |
|
"learning_rate": 6.108597285067873e-06, |
|
"loss": 7.4782, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.03058103975535168, |
|
"grad_norm": 15.834653854370117, |
|
"learning_rate": 6.334841628959276e-06, |
|
"loss": 8.0439, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.03160040774719674, |
|
"grad_norm": 15.115158081054688, |
|
"learning_rate": 6.5610859728506795e-06, |
|
"loss": 7.1911, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.0326197757390418, |
|
"grad_norm": 17.570573806762695, |
|
"learning_rate": 6.787330316742081e-06, |
|
"loss": 8.5735, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.03363914373088685, |
|
"grad_norm": 15.224530220031738, |
|
"learning_rate": 7.013574660633485e-06, |
|
"loss": 8.3855, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.034658511722731905, |
|
"grad_norm": 16.47282600402832, |
|
"learning_rate": 7.239819004524888e-06, |
|
"loss": 8.4305, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.03567787971457696, |
|
"grad_norm": 16.739215850830078, |
|
"learning_rate": 7.46606334841629e-06, |
|
"loss": 9.4608, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.03669724770642202, |
|
"grad_norm": 13.741637229919434, |
|
"learning_rate": 7.692307692307694e-06, |
|
"loss": 8.1572, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.03771661569826707, |
|
"grad_norm": 14.70285701751709, |
|
"learning_rate": 7.918552036199094e-06, |
|
"loss": 8.1456, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.03873598369011213, |
|
"grad_norm": 11.470185279846191, |
|
"learning_rate": 8.144796380090498e-06, |
|
"loss": 7.3833, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.039755351681957186, |
|
"grad_norm": 13.029812812805176, |
|
"learning_rate": 8.3710407239819e-06, |
|
"loss": 8.8539, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.040774719673802244, |
|
"grad_norm": 12.46716594696045, |
|
"learning_rate": 8.597285067873303e-06, |
|
"loss": 8.9349, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.0417940876656473, |
|
"grad_norm": 12.875706672668457, |
|
"learning_rate": 8.823529411764707e-06, |
|
"loss": 8.1803, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.04281345565749235, |
|
"grad_norm": 12.646770477294922, |
|
"learning_rate": 9.049773755656109e-06, |
|
"loss": 6.7532, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.04383282364933741, |
|
"grad_norm": 13.792744636535645, |
|
"learning_rate": 9.276018099547511e-06, |
|
"loss": 7.127, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.04485219164118247, |
|
"grad_norm": 11.656695365905762, |
|
"learning_rate": 9.502262443438914e-06, |
|
"loss": 7.5565, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.045871559633027525, |
|
"grad_norm": 11.562976837158203, |
|
"learning_rate": 9.728506787330318e-06, |
|
"loss": 7.6078, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.046890927624872576, |
|
"grad_norm": 11.516715049743652, |
|
"learning_rate": 9.95475113122172e-06, |
|
"loss": 8.4153, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.047910295616717634, |
|
"grad_norm": 11.569866180419922, |
|
"learning_rate": 1.0180995475113122e-05, |
|
"loss": 7.1062, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.04892966360856269, |
|
"grad_norm": 11.088666915893555, |
|
"learning_rate": 1.0407239819004526e-05, |
|
"loss": 6.8482, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.04994903160040775, |
|
"grad_norm": 11.396224021911621, |
|
"learning_rate": 1.0633484162895929e-05, |
|
"loss": 7.2262, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.0509683995922528, |
|
"grad_norm": 11.868388175964355, |
|
"learning_rate": 1.0859728506787331e-05, |
|
"loss": 8.0207, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.05198776758409786, |
|
"grad_norm": 10.022957801818848, |
|
"learning_rate": 1.1085972850678733e-05, |
|
"loss": 7.6895, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.053007135575942915, |
|
"grad_norm": 11.007475852966309, |
|
"learning_rate": 1.1312217194570136e-05, |
|
"loss": 7.6185, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.05402650356778797, |
|
"grad_norm": 10.026458740234375, |
|
"learning_rate": 1.153846153846154e-05, |
|
"loss": 8.8153, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.05504587155963303, |
|
"grad_norm": 10.358866691589355, |
|
"learning_rate": 1.1764705882352942e-05, |
|
"loss": 7.7666, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.05606523955147808, |
|
"grad_norm": 10.722491264343262, |
|
"learning_rate": 1.1990950226244344e-05, |
|
"loss": 7.1431, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.05708460754332314, |
|
"grad_norm": 10.623186111450195, |
|
"learning_rate": 1.2217194570135746e-05, |
|
"loss": 6.3969, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.0581039755351682, |
|
"grad_norm": 10.13591480255127, |
|
"learning_rate": 1.244343891402715e-05, |
|
"loss": 8.1643, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.059123343527013254, |
|
"grad_norm": 9.476139068603516, |
|
"learning_rate": 1.2669683257918553e-05, |
|
"loss": 7.1228, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.060142711518858305, |
|
"grad_norm": 8.608465194702148, |
|
"learning_rate": 1.2895927601809957e-05, |
|
"loss": 6.9228, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.06116207951070336, |
|
"grad_norm": 10.69497299194336, |
|
"learning_rate": 1.3122171945701359e-05, |
|
"loss": 10.2251, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.06218144750254842, |
|
"grad_norm": 9.309306144714355, |
|
"learning_rate": 1.3348416289592761e-05, |
|
"loss": 7.1105, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.06320081549439348, |
|
"grad_norm": 9.268863677978516, |
|
"learning_rate": 1.3574660633484162e-05, |
|
"loss": 7.1156, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.06422018348623854, |
|
"grad_norm": 10.207130432128906, |
|
"learning_rate": 1.3800904977375568e-05, |
|
"loss": 6.5522, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.0652395514780836, |
|
"grad_norm": 9.29359245300293, |
|
"learning_rate": 1.402714932126697e-05, |
|
"loss": 6.734, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.06625891946992865, |
|
"grad_norm": 8.38429069519043, |
|
"learning_rate": 1.425339366515837e-05, |
|
"loss": 8.1303, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.0672782874617737, |
|
"grad_norm": 9.689257621765137, |
|
"learning_rate": 1.4479638009049776e-05, |
|
"loss": 7.298, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.06829765545361875, |
|
"grad_norm": 8.886714935302734, |
|
"learning_rate": 1.4705882352941177e-05, |
|
"loss": 6.1227, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.06931702344546381, |
|
"grad_norm": 9.28791332244873, |
|
"learning_rate": 1.493212669683258e-05, |
|
"loss": 6.7938, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.07033639143730887, |
|
"grad_norm": 9.196669578552246, |
|
"learning_rate": 1.5158371040723981e-05, |
|
"loss": 6.4562, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.07135575942915393, |
|
"grad_norm": 10.716215133666992, |
|
"learning_rate": 1.5384615384615387e-05, |
|
"loss": 8.0389, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.07237512742099898, |
|
"grad_norm": 9.852572441101074, |
|
"learning_rate": 1.5610859728506788e-05, |
|
"loss": 8.7218, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.07339449541284404, |
|
"grad_norm": 8.59492301940918, |
|
"learning_rate": 1.583710407239819e-05, |
|
"loss": 6.1906, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.0744138634046891, |
|
"grad_norm": 9.830521583557129, |
|
"learning_rate": 1.6063348416289596e-05, |
|
"loss": 6.7222, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.07543323139653414, |
|
"grad_norm": 9.12816047668457, |
|
"learning_rate": 1.6289592760180996e-05, |
|
"loss": 7.0611, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.0764525993883792, |
|
"grad_norm": 10.391504287719727, |
|
"learning_rate": 1.6515837104072397e-05, |
|
"loss": 7.8241, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.07747196738022426, |
|
"grad_norm": 9.0382719039917, |
|
"learning_rate": 1.67420814479638e-05, |
|
"loss": 6.3791, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.07849133537206932, |
|
"grad_norm": 11.495955467224121, |
|
"learning_rate": 1.6968325791855205e-05, |
|
"loss": 6.8864, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.07951070336391437, |
|
"grad_norm": 9.282613754272461, |
|
"learning_rate": 1.7194570135746606e-05, |
|
"loss": 6.8356, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.08053007135575943, |
|
"grad_norm": 9.06067180633545, |
|
"learning_rate": 1.742081447963801e-05, |
|
"loss": 6.168, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.08154943934760449, |
|
"grad_norm": 10.343846321105957, |
|
"learning_rate": 1.7647058823529414e-05, |
|
"loss": 8.6845, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.08256880733944955, |
|
"grad_norm": 10.185526847839355, |
|
"learning_rate": 1.7873303167420814e-05, |
|
"loss": 5.9739, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.0835881753312946, |
|
"grad_norm": 12.164653778076172, |
|
"learning_rate": 1.8099547511312218e-05, |
|
"loss": 6.2423, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.08460754332313965, |
|
"grad_norm": 10.543149948120117, |
|
"learning_rate": 1.832579185520362e-05, |
|
"loss": 7.6247, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.0856269113149847, |
|
"grad_norm": 10.210731506347656, |
|
"learning_rate": 1.8552036199095023e-05, |
|
"loss": 5.8418, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.08664627930682976, |
|
"grad_norm": 11.613642692565918, |
|
"learning_rate": 1.8778280542986427e-05, |
|
"loss": 7.0948, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.08766564729867482, |
|
"grad_norm": 12.590648651123047, |
|
"learning_rate": 1.9004524886877827e-05, |
|
"loss": 6.7457, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.08868501529051988, |
|
"grad_norm": 12.547815322875977, |
|
"learning_rate": 1.923076923076923e-05, |
|
"loss": 5.6837, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.08970438328236494, |
|
"grad_norm": 14.212437629699707, |
|
"learning_rate": 1.9457013574660635e-05, |
|
"loss": 6.0757, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.09072375127421, |
|
"grad_norm": 14.821358680725098, |
|
"learning_rate": 1.9683257918552036e-05, |
|
"loss": 6.523, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.09174311926605505, |
|
"grad_norm": 14.133096694946289, |
|
"learning_rate": 1.990950226244344e-05, |
|
"loss": 6.2917, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.09276248725790011, |
|
"grad_norm": 14.283154487609863, |
|
"learning_rate": 2.0135746606334844e-05, |
|
"loss": 6.0509, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.09378185524974515, |
|
"grad_norm": 15.914741516113281, |
|
"learning_rate": 2.0361990950226245e-05, |
|
"loss": 6.8298, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.09480122324159021, |
|
"grad_norm": 18.067726135253906, |
|
"learning_rate": 2.058823529411765e-05, |
|
"loss": 7.855, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.09582059123343527, |
|
"grad_norm": 17.288843154907227, |
|
"learning_rate": 2.0814479638009053e-05, |
|
"loss": 6.6372, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.09683995922528033, |
|
"grad_norm": 22.13617515563965, |
|
"learning_rate": 2.1040723981900453e-05, |
|
"loss": 5.7468, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.09785932721712538, |
|
"grad_norm": 22.20960235595703, |
|
"learning_rate": 2.1266968325791857e-05, |
|
"loss": 7.5522, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.09887869520897044, |
|
"grad_norm": 23.28131103515625, |
|
"learning_rate": 2.149321266968326e-05, |
|
"loss": 7.7825, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.0998980632008155, |
|
"grad_norm": 29.695850372314453, |
|
"learning_rate": 2.1719457013574662e-05, |
|
"loss": 8.7452, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.0998980632008155, |
|
"eval_Qnli-dev-1024_cosine_accuracy": 0.6458333333333334, |
|
"eval_Qnli-dev-1024_cosine_accuracy_threshold": 0.995652437210083, |
|
"eval_Qnli-dev-1024_cosine_ap": 0.6274798374964984, |
|
"eval_Qnli-dev-1024_cosine_f1": 0.6518518518518519, |
|
"eval_Qnli-dev-1024_cosine_f1_threshold": 0.9515509605407715, |
|
"eval_Qnli-dev-1024_cosine_mcc": 0.1563007361345257, |
|
"eval_Qnli-dev-1024_cosine_precision": 0.4888888888888889, |
|
"eval_Qnli-dev-1024_cosine_recall": 0.9777777777777777, |
|
"eval_Qnli-dev_cosine_accuracy": 0.7395833333333334, |
|
"eval_Qnli-dev_cosine_accuracy_threshold": 0.8860945701599121, |
|
"eval_Qnli-dev_cosine_ap": 0.7645314494110582, |
|
"eval_Qnli-dev_cosine_f1": 0.7500000000000001, |
|
"eval_Qnli-dev_cosine_f1_threshold": 0.8442017436027527, |
|
"eval_Qnli-dev_cosine_mcc": 0.48653004754089046, |
|
"eval_Qnli-dev_cosine_precision": 0.6610169491525424, |
|
"eval_Qnli-dev_cosine_recall": 0.8666666666666667, |
|
"eval_allNLI--triplets-1024_cosine_accuracy": 0.7291666865348816, |
|
"eval_allNLI-triplets_cosine_accuracy": 0.96875, |
|
"eval_global_dataset_loss": 2.297825574874878, |
|
"eval_global_dataset_runtime": 104.2196, |
|
"eval_global_dataset_samples_per_second": 7.705, |
|
"eval_global_dataset_steps_per_second": 0.163, |
|
"eval_sequential_score": 0.7291666865348816, |
|
"eval_sts-test-1024_pearson_cosine": 0.470983874633109, |
|
"eval_sts-test-1024_spearman_cosine": 0.7146928621162676, |
|
"eval_sts-test_pearson_cosine": 0.904138891044396, |
|
"eval_sts-test_spearman_cosine": 0.9172742489825538, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.10091743119266056, |
|
"grad_norm": 21.82425880432129, |
|
"learning_rate": 2.1945701357466062e-05, |
|
"loss": 6.2322, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.1019367991845056, |
|
"grad_norm": 25.734025955200195, |
|
"learning_rate": 2.2171945701357466e-05, |
|
"loss": 4.8433, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.10295616717635066, |
|
"grad_norm": 28.17144775390625, |
|
"learning_rate": 2.239819004524887e-05, |
|
"loss": 4.8335, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.10397553516819572, |
|
"grad_norm": 27.875871658325195, |
|
"learning_rate": 2.262443438914027e-05, |
|
"loss": 5.138, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.10499490316004077, |
|
"grad_norm": 31.503034591674805, |
|
"learning_rate": 2.2850678733031675e-05, |
|
"loss": 4.8609, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.10601427115188583, |
|
"grad_norm": 26.674440383911133, |
|
"learning_rate": 2.307692307692308e-05, |
|
"loss": 4.6204, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.10703363914373089, |
|
"grad_norm": 25.039222717285156, |
|
"learning_rate": 2.330316742081448e-05, |
|
"loss": 4.3809, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.10805300713557595, |
|
"grad_norm": 26.333913803100586, |
|
"learning_rate": 2.3529411764705884e-05, |
|
"loss": 5.6703, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.109072375127421, |
|
"grad_norm": 23.51517105102539, |
|
"learning_rate": 2.3755656108597284e-05, |
|
"loss": 5.0237, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.11009174311926606, |
|
"grad_norm": 18.25855255126953, |
|
"learning_rate": 2.3981900452488688e-05, |
|
"loss": 4.002, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.1111111111111111, |
|
"grad_norm": 19.852886199951172, |
|
"learning_rate": 2.4208144796380092e-05, |
|
"loss": 5.2532, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.11213047910295616, |
|
"grad_norm": 17.45444107055664, |
|
"learning_rate": 2.4434389140271493e-05, |
|
"loss": 4.1033, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.11314984709480122, |
|
"grad_norm": 14.521421432495117, |
|
"learning_rate": 2.4660633484162897e-05, |
|
"loss": 4.0818, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.11416921508664628, |
|
"grad_norm": 12.525910377502441, |
|
"learning_rate": 2.48868778280543e-05, |
|
"loss": 3.458, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.11518858307849134, |
|
"grad_norm": 14.503193855285645, |
|
"learning_rate": 2.51131221719457e-05, |
|
"loss": 4.3372, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.1162079510703364, |
|
"grad_norm": 14.2279634475708, |
|
"learning_rate": 2.5339366515837106e-05, |
|
"loss": 4.9513, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.11722731906218145, |
|
"grad_norm": 15.238719940185547, |
|
"learning_rate": 2.5565610859728506e-05, |
|
"loss": 5.2602, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.11824668705402651, |
|
"grad_norm": 11.11528491973877, |
|
"learning_rate": 2.5791855203619913e-05, |
|
"loss": 3.1741, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.11926605504587157, |
|
"grad_norm": 12.077157974243164, |
|
"learning_rate": 2.6018099547511314e-05, |
|
"loss": 4.1914, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.12028542303771661, |
|
"grad_norm": 11.872669219970703, |
|
"learning_rate": 2.6244343891402718e-05, |
|
"loss": 2.8383, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.12130479102956167, |
|
"grad_norm": 9.008302688598633, |
|
"learning_rate": 2.647058823529412e-05, |
|
"loss": 3.4165, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.12232415902140673, |
|
"grad_norm": 10.702130317687988, |
|
"learning_rate": 2.6696832579185523e-05, |
|
"loss": 3.5085, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.12334352701325178, |
|
"grad_norm": 10.306276321411133, |
|
"learning_rate": 2.6923076923076923e-05, |
|
"loss": 2.3992, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.12436289500509684, |
|
"grad_norm": 9.035378456115723, |
|
"learning_rate": 2.7149321266968324e-05, |
|
"loss": 2.4849, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.12538226299694188, |
|
"grad_norm": 8.996299743652344, |
|
"learning_rate": 2.737556561085973e-05, |
|
"loss": 2.2839, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.12640163098878696, |
|
"grad_norm": 8.635661125183105, |
|
"learning_rate": 2.7601809954751135e-05, |
|
"loss": 2.567, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.127420998980632, |
|
"grad_norm": 10.015826225280762, |
|
"learning_rate": 2.7828054298642536e-05, |
|
"loss": 4.5119, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.12844036697247707, |
|
"grad_norm": 8.679932594299316, |
|
"learning_rate": 2.805429864253394e-05, |
|
"loss": 2.767, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.12945973496432212, |
|
"grad_norm": 10.05739688873291, |
|
"learning_rate": 2.828054298642534e-05, |
|
"loss": 4.0225, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.1304791029561672, |
|
"grad_norm": 9.361485481262207, |
|
"learning_rate": 2.850678733031674e-05, |
|
"loss": 1.8294, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.13149847094801223, |
|
"grad_norm": 9.865928649902344, |
|
"learning_rate": 2.8733031674208145e-05, |
|
"loss": 4.4174, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.1325178389398573, |
|
"grad_norm": 10.055468559265137, |
|
"learning_rate": 2.8959276018099553e-05, |
|
"loss": 2.0112, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.13353720693170235, |
|
"grad_norm": 9.528116226196289, |
|
"learning_rate": 2.9185520361990953e-05, |
|
"loss": 1.7772, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.1345565749235474, |
|
"grad_norm": 9.870166778564453, |
|
"learning_rate": 2.9411764705882354e-05, |
|
"loss": 3.1912, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.13557594291539246, |
|
"grad_norm": 10.1703462600708, |
|
"learning_rate": 2.9638009049773758e-05, |
|
"loss": 2.4527, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.1365953109072375, |
|
"grad_norm": 7.443604469299316, |
|
"learning_rate": 2.986425339366516e-05, |
|
"loss": 1.6424, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.13761467889908258, |
|
"grad_norm": 10.003544807434082, |
|
"learning_rate": 3.0090497737556562e-05, |
|
"loss": 2.6143, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.13863404689092762, |
|
"grad_norm": 9.352860450744629, |
|
"learning_rate": 3.0316742081447963e-05, |
|
"loss": 2.0498, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.1396534148827727, |
|
"grad_norm": 7.393095970153809, |
|
"learning_rate": 3.0542986425339374e-05, |
|
"loss": 1.962, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.14067278287461774, |
|
"grad_norm": 8.278059959411621, |
|
"learning_rate": 3.0769230769230774e-05, |
|
"loss": 1.789, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.14169215086646278, |
|
"grad_norm": 6.577699184417725, |
|
"learning_rate": 3.0995475113122175e-05, |
|
"loss": 1.459, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.14271151885830785, |
|
"grad_norm": 8.23404312133789, |
|
"learning_rate": 3.1221719457013576e-05, |
|
"loss": 1.2479, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.1437308868501529, |
|
"grad_norm": 9.47106647491455, |
|
"learning_rate": 3.1447963800904976e-05, |
|
"loss": 2.5413, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.14475025484199797, |
|
"grad_norm": 7.330000400543213, |
|
"learning_rate": 3.167420814479638e-05, |
|
"loss": 1.4077, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.145769622833843, |
|
"grad_norm": 9.64534854888916, |
|
"learning_rate": 3.1900452488687784e-05, |
|
"loss": 2.6988, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.14678899082568808, |
|
"grad_norm": 8.404465675354004, |
|
"learning_rate": 3.212669683257919e-05, |
|
"loss": 2.9772, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.14780835881753313, |
|
"grad_norm": 8.019698143005371, |
|
"learning_rate": 3.235294117647059e-05, |
|
"loss": 1.6265, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.1488277268093782, |
|
"grad_norm": 7.635079860687256, |
|
"learning_rate": 3.257918552036199e-05, |
|
"loss": 1.9404, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.14984709480122324, |
|
"grad_norm": 7.929011821746826, |
|
"learning_rate": 3.2805429864253393e-05, |
|
"loss": 1.4251, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.15086646279306828, |
|
"grad_norm": 7.869425296783447, |
|
"learning_rate": 3.3031674208144794e-05, |
|
"loss": 2.6657, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.15188583078491336, |
|
"grad_norm": 8.369176864624023, |
|
"learning_rate": 3.32579185520362e-05, |
|
"loss": 2.1576, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.1529051987767584, |
|
"grad_norm": 9.128487586975098, |
|
"learning_rate": 3.34841628959276e-05, |
|
"loss": 1.53, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.15392456676860347, |
|
"grad_norm": 7.673459529876709, |
|
"learning_rate": 3.371040723981901e-05, |
|
"loss": 1.2642, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.15494393476044852, |
|
"grad_norm": 9.104422569274902, |
|
"learning_rate": 3.393665158371041e-05, |
|
"loss": 2.4846, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.1559633027522936, |
|
"grad_norm": 8.658594131469727, |
|
"learning_rate": 3.416289592760181e-05, |
|
"loss": 1.4979, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.15698267074413863, |
|
"grad_norm": 9.34330940246582, |
|
"learning_rate": 3.438914027149321e-05, |
|
"loss": 1.8149, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.1580020387359837, |
|
"grad_norm": 9.401769638061523, |
|
"learning_rate": 3.461538461538462e-05, |
|
"loss": 1.4693, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.15902140672782875, |
|
"grad_norm": 10.389461517333984, |
|
"learning_rate": 3.484162895927602e-05, |
|
"loss": 2.0114, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.1600407747196738, |
|
"grad_norm": 9.321866989135742, |
|
"learning_rate": 3.506787330316742e-05, |
|
"loss": 1.5511, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.16106014271151886, |
|
"grad_norm": 10.052262306213379, |
|
"learning_rate": 3.529411764705883e-05, |
|
"loss": 1.621, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.1620795107033639, |
|
"grad_norm": 7.535787105560303, |
|
"learning_rate": 3.552036199095023e-05, |
|
"loss": 2.1122, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.16309887869520898, |
|
"grad_norm": 9.70533275604248, |
|
"learning_rate": 3.574660633484163e-05, |
|
"loss": 1.8148, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.16411824668705402, |
|
"grad_norm": 7.81204080581665, |
|
"learning_rate": 3.5972850678733036e-05, |
|
"loss": 1.9861, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.1651376146788991, |
|
"grad_norm": 7.583981513977051, |
|
"learning_rate": 3.6199095022624436e-05, |
|
"loss": 1.3943, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.16615698267074414, |
|
"grad_norm": 8.344895362854004, |
|
"learning_rate": 3.642533936651584e-05, |
|
"loss": 1.7317, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.1671763506625892, |
|
"grad_norm": 7.1097331047058105, |
|
"learning_rate": 3.665158371040724e-05, |
|
"loss": 1.6, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.16819571865443425, |
|
"grad_norm": 7.911113739013672, |
|
"learning_rate": 3.6877828054298645e-05, |
|
"loss": 1.2222, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.1692150866462793, |
|
"grad_norm": 9.282394409179688, |
|
"learning_rate": 3.7104072398190046e-05, |
|
"loss": 1.6152, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.17023445463812437, |
|
"grad_norm": 7.449146270751953, |
|
"learning_rate": 3.733031674208145e-05, |
|
"loss": 1.0374, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.1712538226299694, |
|
"grad_norm": 9.164731979370117, |
|
"learning_rate": 3.7556561085972854e-05, |
|
"loss": 1.2844, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.17227319062181448, |
|
"grad_norm": 6.987304210662842, |
|
"learning_rate": 3.7782805429864254e-05, |
|
"loss": 1.8805, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.17329255861365953, |
|
"grad_norm": 7.447988033294678, |
|
"learning_rate": 3.8009049773755655e-05, |
|
"loss": 1.0972, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.1743119266055046, |
|
"grad_norm": 7.7849321365356445, |
|
"learning_rate": 3.8235294117647055e-05, |
|
"loss": 1.7012, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.17533129459734964, |
|
"grad_norm": 7.341614246368408, |
|
"learning_rate": 3.846153846153846e-05, |
|
"loss": 1.4182, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.1763506625891947, |
|
"grad_norm": 8.514887809753418, |
|
"learning_rate": 3.868778280542987e-05, |
|
"loss": 2.6053, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.17737003058103976, |
|
"grad_norm": 7.384711265563965, |
|
"learning_rate": 3.891402714932127e-05, |
|
"loss": 1.4193, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.1783893985728848, |
|
"grad_norm": 8.553336143493652, |
|
"learning_rate": 3.914027149321267e-05, |
|
"loss": 2.251, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.17940876656472987, |
|
"grad_norm": 8.517749786376953, |
|
"learning_rate": 3.936651583710407e-05, |
|
"loss": 1.9057, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.18042813455657492, |
|
"grad_norm": 8.444558143615723, |
|
"learning_rate": 3.959276018099547e-05, |
|
"loss": 1.1228, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.18144750254842, |
|
"grad_norm": 12.253990173339844, |
|
"learning_rate": 3.981900452488688e-05, |
|
"loss": 4.0905, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.18246687054026503, |
|
"grad_norm": 5.70052433013916, |
|
"learning_rate": 4.004524886877829e-05, |
|
"loss": 0.9007, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.1834862385321101, |
|
"grad_norm": 9.525473594665527, |
|
"learning_rate": 4.027149321266969e-05, |
|
"loss": 2.0665, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.18450560652395515, |
|
"grad_norm": 6.146080493927002, |
|
"learning_rate": 4.049773755656109e-05, |
|
"loss": 1.0946, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.18552497451580022, |
|
"grad_norm": 7.736543655395508, |
|
"learning_rate": 4.072398190045249e-05, |
|
"loss": 1.7479, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.18654434250764526, |
|
"grad_norm": 8.404258728027344, |
|
"learning_rate": 4.095022624434389e-05, |
|
"loss": 2.0877, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.1875637104994903, |
|
"grad_norm": 5.705750942230225, |
|
"learning_rate": 4.11764705882353e-05, |
|
"loss": 0.9239, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.18858307849133538, |
|
"grad_norm": 7.753995895385742, |
|
"learning_rate": 4.14027149321267e-05, |
|
"loss": 1.7865, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.18960244648318042, |
|
"grad_norm": 9.15240478515625, |
|
"learning_rate": 4.1628959276018105e-05, |
|
"loss": 2.1053, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.1906218144750255, |
|
"grad_norm": 7.2251129150390625, |
|
"learning_rate": 4.1855203619909506e-05, |
|
"loss": 1.5273, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.19164118246687054, |
|
"grad_norm": 6.803040981292725, |
|
"learning_rate": 4.2081447963800907e-05, |
|
"loss": 1.8726, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.1926605504587156, |
|
"grad_norm": 5.646162509918213, |
|
"learning_rate": 4.230769230769231e-05, |
|
"loss": 1.4663, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.19367991845056065, |
|
"grad_norm": 7.599930286407471, |
|
"learning_rate": 4.2533936651583714e-05, |
|
"loss": 1.0136, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.1946992864424057, |
|
"grad_norm": 7.882979393005371, |
|
"learning_rate": 4.2760180995475115e-05, |
|
"loss": 1.121, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.19571865443425077, |
|
"grad_norm": 8.919268608093262, |
|
"learning_rate": 4.298642533936652e-05, |
|
"loss": 1.6074, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.1967380224260958, |
|
"grad_norm": 8.914848327636719, |
|
"learning_rate": 4.321266968325792e-05, |
|
"loss": 2.1956, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.19775739041794088, |
|
"grad_norm": 8.603778839111328, |
|
"learning_rate": 4.3438914027149324e-05, |
|
"loss": 1.5425, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.19877675840978593, |
|
"grad_norm": 8.500616073608398, |
|
"learning_rate": 4.3665158371040724e-05, |
|
"loss": 1.4552, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.199796126401631, |
|
"grad_norm": 7.815979957580566, |
|
"learning_rate": 4.3891402714932125e-05, |
|
"loss": 1.2635, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.199796126401631, |
|
"eval_Qnli-dev-1024_cosine_accuracy": 0.7395833333333334, |
|
"eval_Qnli-dev-1024_cosine_accuracy_threshold": 0.8570283651351929, |
|
"eval_Qnli-dev-1024_cosine_ap": 0.7434694144471753, |
|
"eval_Qnli-dev-1024_cosine_f1": 0.7207207207207208, |
|
"eval_Qnli-dev-1024_cosine_f1_threshold": 0.7906914353370667, |
|
"eval_Qnli-dev-1024_cosine_mcc": 0.4081269865567241, |
|
"eval_Qnli-dev-1024_cosine_precision": 0.6060606060606061, |
|
"eval_Qnli-dev-1024_cosine_recall": 0.8888888888888888, |
|
"eval_Qnli-dev_cosine_accuracy": 0.75, |
|
"eval_Qnli-dev_cosine_accuracy_threshold": 0.8287814855575562, |
|
"eval_Qnli-dev_cosine_ap": 0.7646453733471359, |
|
"eval_Qnli-dev_cosine_f1": 0.7378640776699029, |
|
"eval_Qnli-dev_cosine_f1_threshold": 0.7745069265365601, |
|
"eval_Qnli-dev_cosine_mcc": 0.46153029495329345, |
|
"eval_Qnli-dev_cosine_precision": 0.6551724137931034, |
|
"eval_Qnli-dev_cosine_recall": 0.8444444444444444, |
|
"eval_allNLI--triplets-1024_cosine_accuracy": 0.9166666865348816, |
|
"eval_allNLI-triplets_cosine_accuracy": 0.9583333134651184, |
|
"eval_global_dataset_loss": 0.5179261565208435, |
|
"eval_global_dataset_runtime": 104.2216, |
|
"eval_global_dataset_samples_per_second": 7.705, |
|
"eval_global_dataset_steps_per_second": 0.163, |
|
"eval_sequential_score": 0.9166666865348816, |
|
"eval_sts-test-1024_pearson_cosine": 0.8476975008591285, |
|
"eval_sts-test-1024_spearman_cosine": 0.8973182534732806, |
|
"eval_sts-test_pearson_cosine": 0.9039400681490469, |
|
"eval_sts-test_spearman_cosine": 0.9185431775441114, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.20081549439347604, |
|
"grad_norm": 9.34125804901123, |
|
"learning_rate": 4.411764705882353e-05, |
|
"loss": 1.7222, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.2018348623853211, |
|
"grad_norm": 10.679852485656738, |
|
"learning_rate": 4.434389140271493e-05, |
|
"loss": 2.377, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.20285423037716616, |
|
"grad_norm": 7.775190830230713, |
|
"learning_rate": 4.457013574660634e-05, |
|
"loss": 1.5317, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.2038735983690112, |
|
"grad_norm": 6.390950679779053, |
|
"learning_rate": 4.479638009049774e-05, |
|
"loss": 1.0494, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.20489296636085627, |
|
"grad_norm": 9.170794486999512, |
|
"learning_rate": 4.502262443438914e-05, |
|
"loss": 1.7392, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.20591233435270132, |
|
"grad_norm": 7.37787389755249, |
|
"learning_rate": 4.524886877828054e-05, |
|
"loss": 1.2924, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.2069317023445464, |
|
"grad_norm": 6.836249828338623, |
|
"learning_rate": 4.547511312217195e-05, |
|
"loss": 0.9413, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.20795107033639143, |
|
"grad_norm": 9.543895721435547, |
|
"learning_rate": 4.570135746606335e-05, |
|
"loss": 2.1448, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.2089704383282365, |
|
"grad_norm": 7.8430495262146, |
|
"learning_rate": 4.592760180995475e-05, |
|
"loss": 1.0357, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.20998980632008155, |
|
"grad_norm": 9.558221817016602, |
|
"learning_rate": 4.615384615384616e-05, |
|
"loss": 1.3534, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.21100917431192662, |
|
"grad_norm": 5.715826034545898, |
|
"learning_rate": 4.638009049773756e-05, |
|
"loss": 1.0564, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.21202854230377166, |
|
"grad_norm": 8.720932960510254, |
|
"learning_rate": 4.660633484162896e-05, |
|
"loss": 0.9259, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.2130479102956167, |
|
"grad_norm": 9.008890151977539, |
|
"learning_rate": 4.683257918552037e-05, |
|
"loss": 1.3813, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.21406727828746178, |
|
"grad_norm": 7.1262006759643555, |
|
"learning_rate": 4.705882352941177e-05, |
|
"loss": 0.8828, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.21508664627930682, |
|
"grad_norm": 12.986166000366211, |
|
"learning_rate": 4.728506787330317e-05, |
|
"loss": 2.9147, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.2161060142711519, |
|
"grad_norm": 6.804072380065918, |
|
"learning_rate": 4.751131221719457e-05, |
|
"loss": 0.6539, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.21712538226299694, |
|
"grad_norm": 9.138653755187988, |
|
"learning_rate": 4.7737556561085976e-05, |
|
"loss": 1.3092, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.218144750254842, |
|
"grad_norm": 7.303668975830078, |
|
"learning_rate": 4.7963800904977377e-05, |
|
"loss": 1.1562, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.21916411824668705, |
|
"grad_norm": 7.368769645690918, |
|
"learning_rate": 4.8190045248868784e-05, |
|
"loss": 0.9509, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.22018348623853212, |
|
"grad_norm": 5.067785263061523, |
|
"learning_rate": 4.8416289592760185e-05, |
|
"loss": 0.6664, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.22120285423037717, |
|
"grad_norm": 5.643320083618164, |
|
"learning_rate": 4.8642533936651585e-05, |
|
"loss": 1.2315, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.2222222222222222, |
|
"grad_norm": 6.596173286437988, |
|
"learning_rate": 4.8868778280542986e-05, |
|
"loss": 0.9855, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.22324159021406728, |
|
"grad_norm": 6.5434770584106445, |
|
"learning_rate": 4.9095022624434386e-05, |
|
"loss": 0.9258, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.22426095820591233, |
|
"grad_norm": 11.537922859191895, |
|
"learning_rate": 4.9321266968325794e-05, |
|
"loss": 1.6578, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.2252803261977574, |
|
"grad_norm": 7.364137172698975, |
|
"learning_rate": 4.95475113122172e-05, |
|
"loss": 0.9666, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.22629969418960244, |
|
"grad_norm": 8.102925300598145, |
|
"learning_rate": 4.97737556561086e-05, |
|
"loss": 0.9808, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.2273190621814475, |
|
"grad_norm": 10.013775825500488, |
|
"learning_rate": 5e-05, |
|
"loss": 2.4156, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.22833843017329256, |
|
"grad_norm": 7.974793434143066, |
|
"learning_rate": 5.02262443438914e-05, |
|
"loss": 1.1297, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.22935779816513763, |
|
"grad_norm": 7.710846424102783, |
|
"learning_rate": 5.0452488687782804e-05, |
|
"loss": 1.3063, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.23037716615698267, |
|
"grad_norm": 5.633566856384277, |
|
"learning_rate": 5.067873303167421e-05, |
|
"loss": 0.5567, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.23139653414882771, |
|
"grad_norm": 9.50987720489502, |
|
"learning_rate": 5.090497737556561e-05, |
|
"loss": 1.3551, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.2324159021406728, |
|
"grad_norm": 10.309268951416016, |
|
"learning_rate": 5.113122171945701e-05, |
|
"loss": 1.4079, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.23343527013251783, |
|
"grad_norm": 7.812633037567139, |
|
"learning_rate": 5.135746606334841e-05, |
|
"loss": 0.948, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.2344546381243629, |
|
"grad_norm": 8.013436317443848, |
|
"learning_rate": 5.158371040723983e-05, |
|
"loss": 0.9288, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.23547400611620795, |
|
"grad_norm": 7.550686359405518, |
|
"learning_rate": 5.180995475113123e-05, |
|
"loss": 1.0077, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.23649337410805302, |
|
"grad_norm": 7.249583721160889, |
|
"learning_rate": 5.203619909502263e-05, |
|
"loss": 1.0674, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.23751274209989806, |
|
"grad_norm": 7.766678810119629, |
|
"learning_rate": 5.2262443438914036e-05, |
|
"loss": 1.3354, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.23853211009174313, |
|
"grad_norm": 7.417704105377197, |
|
"learning_rate": 5.2488687782805436e-05, |
|
"loss": 1.0076, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.23955147808358818, |
|
"grad_norm": 8.414839744567871, |
|
"learning_rate": 5.271493212669684e-05, |
|
"loss": 0.8814, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.24057084607543322, |
|
"grad_norm": 9.537981986999512, |
|
"learning_rate": 5.294117647058824e-05, |
|
"loss": 1.7839, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.2415902140672783, |
|
"grad_norm": 6.3290886878967285, |
|
"learning_rate": 5.316742081447964e-05, |
|
"loss": 1.2614, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.24260958205912334, |
|
"grad_norm": 8.181835174560547, |
|
"learning_rate": 5.3393665158371045e-05, |
|
"loss": 0.8655, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.2436289500509684, |
|
"grad_norm": 8.01684856414795, |
|
"learning_rate": 5.3619909502262446e-05, |
|
"loss": 1.8418, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.24464831804281345, |
|
"grad_norm": 7.891118049621582, |
|
"learning_rate": 5.384615384615385e-05, |
|
"loss": 1.4465, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.24566768603465852, |
|
"grad_norm": 8.080881118774414, |
|
"learning_rate": 5.407239819004525e-05, |
|
"loss": 0.8695, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.24668705402650357, |
|
"grad_norm": 6.881638050079346, |
|
"learning_rate": 5.429864253393665e-05, |
|
"loss": 0.9695, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.24770642201834864, |
|
"grad_norm": 10.03598690032959, |
|
"learning_rate": 5.4524886877828055e-05, |
|
"loss": 1.3374, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.24872579001019368, |
|
"grad_norm": 7.844127178192139, |
|
"learning_rate": 5.475113122171946e-05, |
|
"loss": 0.6716, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.24974515800203873, |
|
"grad_norm": 8.654071807861328, |
|
"learning_rate": 5.497737556561087e-05, |
|
"loss": 1.032, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.25076452599388377, |
|
"grad_norm": 6.731460094451904, |
|
"learning_rate": 5.520361990950227e-05, |
|
"loss": 0.8033, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.25178389398572887, |
|
"grad_norm": 9.436687469482422, |
|
"learning_rate": 5.542986425339367e-05, |
|
"loss": 0.9257, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.2528032619775739, |
|
"grad_norm": 7.817379474639893, |
|
"learning_rate": 5.565610859728507e-05, |
|
"loss": 0.8311, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.25382262996941896, |
|
"grad_norm": 6.328183650970459, |
|
"learning_rate": 5.588235294117647e-05, |
|
"loss": 0.5609, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.254841997961264, |
|
"grad_norm": 8.576601028442383, |
|
"learning_rate": 5.610859728506788e-05, |
|
"loss": 1.5985, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.2558613659531091, |
|
"grad_norm": 9.092324256896973, |
|
"learning_rate": 5.633484162895928e-05, |
|
"loss": 1.0025, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.25688073394495414, |
|
"grad_norm": 11.906094551086426, |
|
"learning_rate": 5.656108597285068e-05, |
|
"loss": 2.0499, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.2579001019367992, |
|
"grad_norm": 7.968968868255615, |
|
"learning_rate": 5.678733031674208e-05, |
|
"loss": 1.3116, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.25891946992864423, |
|
"grad_norm": 5.355049133300781, |
|
"learning_rate": 5.701357466063348e-05, |
|
"loss": 0.5969, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.2599388379204893, |
|
"grad_norm": 8.151896476745605, |
|
"learning_rate": 5.723981900452488e-05, |
|
"loss": 1.1107, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.2609582059123344, |
|
"grad_norm": 9.651622772216797, |
|
"learning_rate": 5.746606334841629e-05, |
|
"loss": 1.8581, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.2619775739041794, |
|
"grad_norm": 7.1527533531188965, |
|
"learning_rate": 5.769230769230769e-05, |
|
"loss": 0.572, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.26299694189602446, |
|
"grad_norm": 6.141374111175537, |
|
"learning_rate": 5.7918552036199105e-05, |
|
"loss": 0.9267, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.2640163098878695, |
|
"grad_norm": 7.274891376495361, |
|
"learning_rate": 5.8144796380090506e-05, |
|
"loss": 0.6255, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.2650356778797146, |
|
"grad_norm": 5.81080436706543, |
|
"learning_rate": 5.8371040723981906e-05, |
|
"loss": 0.7615, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.26605504587155965, |
|
"grad_norm": 6.9981279373168945, |
|
"learning_rate": 5.859728506787331e-05, |
|
"loss": 0.6026, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.2670744138634047, |
|
"grad_norm": 5.718660831451416, |
|
"learning_rate": 5.882352941176471e-05, |
|
"loss": 0.7263, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.26809378185524974, |
|
"grad_norm": 5.391998767852783, |
|
"learning_rate": 5.9049773755656115e-05, |
|
"loss": 0.4643, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.2691131498470948, |
|
"grad_norm": 6.843007564544678, |
|
"learning_rate": 5.9276018099547516e-05, |
|
"loss": 0.5101, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.2701325178389399, |
|
"grad_norm": 5.087254047393799, |
|
"learning_rate": 5.9502262443438916e-05, |
|
"loss": 0.5562, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.2711518858307849, |
|
"grad_norm": 7.482615947723389, |
|
"learning_rate": 5.972850678733032e-05, |
|
"loss": 1.256, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.27217125382262997, |
|
"grad_norm": 6.911371231079102, |
|
"learning_rate": 5.995475113122172e-05, |
|
"loss": 0.6543, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.273190621814475, |
|
"grad_norm": 7.643139839172363, |
|
"learning_rate": 6.0180995475113125e-05, |
|
"loss": 0.6698, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.2742099898063201, |
|
"grad_norm": 9.08658504486084, |
|
"learning_rate": 6.0407239819004525e-05, |
|
"loss": 1.3843, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.27522935779816515, |
|
"grad_norm": 8.890534400939941, |
|
"learning_rate": 6.0633484162895926e-05, |
|
"loss": 1.1421, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.2762487257900102, |
|
"grad_norm": 9.855698585510254, |
|
"learning_rate": 6.0859728506787327e-05, |
|
"loss": 1.1558, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.27726809378185524, |
|
"grad_norm": 8.32972526550293, |
|
"learning_rate": 6.108597285067875e-05, |
|
"loss": 1.603, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.2782874617737003, |
|
"grad_norm": 8.393510818481445, |
|
"learning_rate": 6.131221719457015e-05, |
|
"loss": 0.7985, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.2793068297655454, |
|
"grad_norm": 7.992040157318115, |
|
"learning_rate": 6.153846153846155e-05, |
|
"loss": 1.3884, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.2803261977573904, |
|
"grad_norm": 8.646651268005371, |
|
"learning_rate": 6.176470588235295e-05, |
|
"loss": 1.0337, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.28134556574923547, |
|
"grad_norm": 7.3104329109191895, |
|
"learning_rate": 6.199095022624435e-05, |
|
"loss": 1.0917, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.2823649337410805, |
|
"grad_norm": 12.030378341674805, |
|
"learning_rate": 6.221719457013575e-05, |
|
"loss": 2.4149, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.28338430173292556, |
|
"grad_norm": 4.781021595001221, |
|
"learning_rate": 6.244343891402715e-05, |
|
"loss": 0.4301, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.28440366972477066, |
|
"grad_norm": 4.352090358734131, |
|
"learning_rate": 6.266968325791855e-05, |
|
"loss": 0.5084, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.2854230377166157, |
|
"grad_norm": 5.88839864730835, |
|
"learning_rate": 6.289592760180995e-05, |
|
"loss": 0.7202, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.28644240570846075, |
|
"grad_norm": 11.228419303894043, |
|
"learning_rate": 6.312217194570135e-05, |
|
"loss": 1.7983, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.2874617737003058, |
|
"grad_norm": 6.119421005249023, |
|
"learning_rate": 6.334841628959275e-05, |
|
"loss": 0.673, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.2884811416921509, |
|
"grad_norm": 6.405134677886963, |
|
"learning_rate": 6.357466063348417e-05, |
|
"loss": 0.6655, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.28950050968399593, |
|
"grad_norm": 6.735506534576416, |
|
"learning_rate": 6.380090497737557e-05, |
|
"loss": 0.9121, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.290519877675841, |
|
"grad_norm": 11.012415885925293, |
|
"learning_rate": 6.402714932126697e-05, |
|
"loss": 1.5978, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.291539245667686, |
|
"grad_norm": 13.007187843322754, |
|
"learning_rate": 6.425339366515838e-05, |
|
"loss": 1.8536, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.29255861365953106, |
|
"grad_norm": 12.273601531982422, |
|
"learning_rate": 6.447963800904978e-05, |
|
"loss": 1.6397, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.29357798165137616, |
|
"grad_norm": 9.6339750289917, |
|
"learning_rate": 6.470588235294118e-05, |
|
"loss": 0.8275, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.2945973496432212, |
|
"grad_norm": 6.717658996582031, |
|
"learning_rate": 6.493212669683258e-05, |
|
"loss": 0.6003, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.29561671763506625, |
|
"grad_norm": 8.443256378173828, |
|
"learning_rate": 6.515837104072399e-05, |
|
"loss": 0.9834, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.2966360856269113, |
|
"grad_norm": 8.823105812072754, |
|
"learning_rate": 6.538461538461539e-05, |
|
"loss": 0.603, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.2976554536187564, |
|
"grad_norm": 6.8099141120910645, |
|
"learning_rate": 6.561085972850679e-05, |
|
"loss": 0.6597, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.29867482161060144, |
|
"grad_norm": 6.705087661743164, |
|
"learning_rate": 6.583710407239819e-05, |
|
"loss": 0.739, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.2996941896024465, |
|
"grad_norm": 7.209024906158447, |
|
"learning_rate": 6.606334841628959e-05, |
|
"loss": 1.2564, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.2996941896024465, |
|
"eval_Qnli-dev-1024_cosine_accuracy": 0.7395833333333334, |
|
"eval_Qnli-dev-1024_cosine_accuracy_threshold": 0.8565528392791748, |
|
"eval_Qnli-dev-1024_cosine_ap": 0.7531377591671699, |
|
"eval_Qnli-dev-1024_cosine_f1": 0.7254901960784313, |
|
"eval_Qnli-dev-1024_cosine_f1_threshold": 0.8200148344039917, |
|
"eval_Qnli-dev-1024_cosine_mcc": 0.43697448216965834, |
|
"eval_Qnli-dev-1024_cosine_precision": 0.6491228070175439, |
|
"eval_Qnli-dev-1024_cosine_recall": 0.8222222222222222, |
|
"eval_Qnli-dev_cosine_accuracy": 0.7395833333333334, |
|
"eval_Qnli-dev_cosine_accuracy_threshold": 0.7719540596008301, |
|
"eval_Qnli-dev_cosine_ap": 0.7588639733888536, |
|
"eval_Qnli-dev_cosine_f1": 0.7454545454545455, |
|
"eval_Qnli-dev_cosine_f1_threshold": 0.7090869545936584, |
|
"eval_Qnli-dev_cosine_mcc": 0.47013467657639685, |
|
"eval_Qnli-dev_cosine_precision": 0.6307692307692307, |
|
"eval_Qnli-dev_cosine_recall": 0.9111111111111111, |
|
"eval_allNLI--triplets-1024_cosine_accuracy": 0.8854166865348816, |
|
"eval_allNLI-triplets_cosine_accuracy": 0.9583333134651184, |
|
"eval_global_dataset_loss": 0.4855804145336151, |
|
"eval_global_dataset_runtime": 104.3189, |
|
"eval_global_dataset_samples_per_second": 7.698, |
|
"eval_global_dataset_steps_per_second": 0.163, |
|
"eval_sequential_score": 0.8854166865348816, |
|
"eval_sts-test-1024_pearson_cosine": 0.8681461030339531, |
|
"eval_sts-test-1024_spearman_cosine": 0.9053809631987397, |
|
"eval_sts-test_pearson_cosine": 0.9051731986667259, |
|
"eval_sts-test_spearman_cosine": 0.920630429781229, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.3007135575942915, |
|
"grad_norm": 7.649487495422363, |
|
"learning_rate": 6.6289592760181e-05, |
|
"loss": 1.4442, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.30173292558613657, |
|
"grad_norm": 7.740142822265625, |
|
"learning_rate": 6.65158371040724e-05, |
|
"loss": 1.4064, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.30275229357798167, |
|
"grad_norm": 8.009271621704102, |
|
"learning_rate": 6.67420814479638e-05, |
|
"loss": 0.8456, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.3037716615698267, |
|
"grad_norm": 5.718809604644775, |
|
"learning_rate": 6.69683257918552e-05, |
|
"loss": 0.7772, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.30479102956167176, |
|
"grad_norm": 7.34658145904541, |
|
"learning_rate": 6.719457013574662e-05, |
|
"loss": 0.7619, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.3058103975535168, |
|
"grad_norm": 8.556058883666992, |
|
"learning_rate": 6.742081447963802e-05, |
|
"loss": 1.002, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.3068297655453619, |
|
"grad_norm": 8.995348930358887, |
|
"learning_rate": 6.764705882352942e-05, |
|
"loss": 1.4774, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.30784913353720694, |
|
"grad_norm": 8.271109580993652, |
|
"learning_rate": 6.787330316742082e-05, |
|
"loss": 0.6814, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.308868501529052, |
|
"grad_norm": 8.896450996398926, |
|
"learning_rate": 6.809954751131222e-05, |
|
"loss": 1.2286, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.30988786952089703, |
|
"grad_norm": 9.076520919799805, |
|
"learning_rate": 6.832579185520362e-05, |
|
"loss": 1.8546, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.3109072375127421, |
|
"grad_norm": 6.780123233795166, |
|
"learning_rate": 6.855203619909502e-05, |
|
"loss": 0.7547, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.3119266055045872, |
|
"grad_norm": 7.728740692138672, |
|
"learning_rate": 6.877828054298642e-05, |
|
"loss": 1.3638, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.3129459734964322, |
|
"grad_norm": 8.941544532775879, |
|
"learning_rate": 6.900452488687784e-05, |
|
"loss": 1.2604, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.31396534148827726, |
|
"grad_norm": 6.70719575881958, |
|
"learning_rate": 6.923076923076924e-05, |
|
"loss": 0.5111, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.3149847094801223, |
|
"grad_norm": 7.599255561828613, |
|
"learning_rate": 6.945701357466064e-05, |
|
"loss": 0.7153, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.3160040774719674, |
|
"grad_norm": 7.323727607727051, |
|
"learning_rate": 6.968325791855204e-05, |
|
"loss": 0.8367, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.31702344546381245, |
|
"grad_norm": 7.314160346984863, |
|
"learning_rate": 6.990950226244344e-05, |
|
"loss": 0.6643, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.3180428134556575, |
|
"grad_norm": 8.265671730041504, |
|
"learning_rate": 7.013574660633484e-05, |
|
"loss": 1.0404, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.31906218144750254, |
|
"grad_norm": 10.820046424865723, |
|
"learning_rate": 7.036199095022625e-05, |
|
"loss": 1.122, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.3200815494393476, |
|
"grad_norm": 7.194378852844238, |
|
"learning_rate": 7.058823529411765e-05, |
|
"loss": 0.7091, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.3211009174311927, |
|
"grad_norm": 7.764474868774414, |
|
"learning_rate": 7.081447963800906e-05, |
|
"loss": 1.07, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.3221202854230377, |
|
"grad_norm": 7.757960796356201, |
|
"learning_rate": 7.104072398190046e-05, |
|
"loss": 0.7246, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.32313965341488277, |
|
"grad_norm": 9.631681442260742, |
|
"learning_rate": 7.126696832579186e-05, |
|
"loss": 0.821, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.3241590214067278, |
|
"grad_norm": 6.478396892547607, |
|
"learning_rate": 7.149321266968326e-05, |
|
"loss": 0.9556, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.3251783893985729, |
|
"grad_norm": 8.858171463012695, |
|
"learning_rate": 7.171945701357467e-05, |
|
"loss": 1.2634, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.32619775739041795, |
|
"grad_norm": 8.02340030670166, |
|
"learning_rate": 7.194570135746607e-05, |
|
"loss": 0.8621, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.327217125382263, |
|
"grad_norm": 8.634239196777344, |
|
"learning_rate": 7.217194570135747e-05, |
|
"loss": 1.4215, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.32823649337410804, |
|
"grad_norm": 8.96740436553955, |
|
"learning_rate": 7.239819004524887e-05, |
|
"loss": 0.8894, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.3292558613659531, |
|
"grad_norm": 4.730165958404541, |
|
"learning_rate": 7.262443438914027e-05, |
|
"loss": 0.4134, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.3302752293577982, |
|
"grad_norm": 6.1243181228637695, |
|
"learning_rate": 7.285067873303167e-05, |
|
"loss": 0.4147, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.3312945973496432, |
|
"grad_norm": 7.8853607177734375, |
|
"learning_rate": 7.307692307692307e-05, |
|
"loss": 0.5721, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.33231396534148827, |
|
"grad_norm": 9.193514823913574, |
|
"learning_rate": 7.330316742081448e-05, |
|
"loss": 0.8541, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.3333333333333333, |
|
"grad_norm": 12.314509391784668, |
|
"learning_rate": 7.352941176470589e-05, |
|
"loss": 2.2959, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.3343527013251784, |
|
"grad_norm": 4.384552955627441, |
|
"learning_rate": 7.375565610859729e-05, |
|
"loss": 0.4452, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.33537206931702346, |
|
"grad_norm": 5.677075386047363, |
|
"learning_rate": 7.398190045248869e-05, |
|
"loss": 0.5008, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.3363914373088685, |
|
"grad_norm": 6.752626419067383, |
|
"learning_rate": 7.420814479638009e-05, |
|
"loss": 0.4106, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.33741080530071355, |
|
"grad_norm": 10.971478462219238, |
|
"learning_rate": 7.44343891402715e-05, |
|
"loss": 0.9237, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.3384301732925586, |
|
"grad_norm": 7.574080944061279, |
|
"learning_rate": 7.46606334841629e-05, |
|
"loss": 0.6275, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.3394495412844037, |
|
"grad_norm": 9.538507461547852, |
|
"learning_rate": 7.48868778280543e-05, |
|
"loss": 1.5184, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.34046890927624873, |
|
"grad_norm": 9.139626502990723, |
|
"learning_rate": 7.511312217194571e-05, |
|
"loss": 1.7865, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.3414882772680938, |
|
"grad_norm": 9.275596618652344, |
|
"learning_rate": 7.533936651583711e-05, |
|
"loss": 1.5947, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.3425076452599388, |
|
"grad_norm": 9.375283241271973, |
|
"learning_rate": 7.556561085972851e-05, |
|
"loss": 1.0249, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.3435270132517839, |
|
"grad_norm": 7.951083660125732, |
|
"learning_rate": 7.579185520361991e-05, |
|
"loss": 1.0227, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.34454638124362896, |
|
"grad_norm": 9.579297065734863, |
|
"learning_rate": 7.601809954751131e-05, |
|
"loss": 1.28, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.345565749235474, |
|
"grad_norm": 5.935997486114502, |
|
"learning_rate": 7.624434389140271e-05, |
|
"loss": 0.798, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.34658511722731905, |
|
"grad_norm": 7.16936731338501, |
|
"learning_rate": 7.647058823529411e-05, |
|
"loss": 1.0408, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.3476044852191641, |
|
"grad_norm": 9.448662757873535, |
|
"learning_rate": 7.669683257918553e-05, |
|
"loss": 0.9732, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.3486238532110092, |
|
"grad_norm": 7.747692584991455, |
|
"learning_rate": 7.692307692307693e-05, |
|
"loss": 0.7588, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.34964322120285424, |
|
"grad_norm": 10.198869705200195, |
|
"learning_rate": 7.714932126696833e-05, |
|
"loss": 0.9615, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.3506625891946993, |
|
"grad_norm": 8.069470405578613, |
|
"learning_rate": 7.737556561085974e-05, |
|
"loss": 0.9895, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.3516819571865443, |
|
"grad_norm": 10.662049293518066, |
|
"learning_rate": 7.760180995475114e-05, |
|
"loss": 1.923, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.3527013251783894, |
|
"grad_norm": 6.53238582611084, |
|
"learning_rate": 7.782805429864254e-05, |
|
"loss": 0.615, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.35372069317023447, |
|
"grad_norm": 11.10132122039795, |
|
"learning_rate": 7.805429864253394e-05, |
|
"loss": 1.4572, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.3547400611620795, |
|
"grad_norm": 7.372711181640625, |
|
"learning_rate": 7.828054298642534e-05, |
|
"loss": 1.0083, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.35575942915392456, |
|
"grad_norm": 7.358077526092529, |
|
"learning_rate": 7.850678733031674e-05, |
|
"loss": 0.922, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.3567787971457696, |
|
"grad_norm": 8.45017147064209, |
|
"learning_rate": 7.873303167420814e-05, |
|
"loss": 1.3767, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.3577981651376147, |
|
"grad_norm": 4.858506679534912, |
|
"learning_rate": 7.895927601809954e-05, |
|
"loss": 0.6378, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.35881753312945974, |
|
"grad_norm": 5.764273643493652, |
|
"learning_rate": 7.918552036199095e-05, |
|
"loss": 0.4063, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.3598369011213048, |
|
"grad_norm": 8.656686782836914, |
|
"learning_rate": 7.941176470588235e-05, |
|
"loss": 1.0834, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.36085626911314983, |
|
"grad_norm": 5.824944496154785, |
|
"learning_rate": 7.963800904977376e-05, |
|
"loss": 0.807, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.36187563710499493, |
|
"grad_norm": 6.73368501663208, |
|
"learning_rate": 7.986425339366516e-05, |
|
"loss": 1.0293, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.36289500509684, |
|
"grad_norm": 5.860096454620361, |
|
"learning_rate": 8.009049773755657e-05, |
|
"loss": 0.4371, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.363914373088685, |
|
"grad_norm": 5.65436315536499, |
|
"learning_rate": 8.031674208144798e-05, |
|
"loss": 0.4334, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.36493374108053006, |
|
"grad_norm": 7.566843509674072, |
|
"learning_rate": 8.054298642533938e-05, |
|
"loss": 0.949, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.3659531090723751, |
|
"grad_norm": 6.286118984222412, |
|
"learning_rate": 8.076923076923078e-05, |
|
"loss": 0.5788, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.3669724770642202, |
|
"grad_norm": 10.212640762329102, |
|
"learning_rate": 8.099547511312218e-05, |
|
"loss": 0.8535, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.36799184505606525, |
|
"grad_norm": 9.267760276794434, |
|
"learning_rate": 8.122171945701358e-05, |
|
"loss": 1.2529, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.3690112130479103, |
|
"grad_norm": 8.794651985168457, |
|
"learning_rate": 8.144796380090498e-05, |
|
"loss": 0.8974, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.37003058103975534, |
|
"grad_norm": 13.3441162109375, |
|
"learning_rate": 8.167420814479638e-05, |
|
"loss": 1.9105, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.37104994903160043, |
|
"grad_norm": 9.258030891418457, |
|
"learning_rate": 8.190045248868778e-05, |
|
"loss": 0.7717, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.3720693170234455, |
|
"grad_norm": 6.051854610443115, |
|
"learning_rate": 8.212669683257918e-05, |
|
"loss": 1.1052, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.3730886850152905, |
|
"grad_norm": 9.53382682800293, |
|
"learning_rate": 8.23529411764706e-05, |
|
"loss": 0.7298, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.37410805300713557, |
|
"grad_norm": 6.723752498626709, |
|
"learning_rate": 8.2579185520362e-05, |
|
"loss": 0.7039, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.3751274209989806, |
|
"grad_norm": 6.844725608825684, |
|
"learning_rate": 8.28054298642534e-05, |
|
"loss": 0.8536, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.3761467889908257, |
|
"grad_norm": 5.233691692352295, |
|
"learning_rate": 8.303167420814481e-05, |
|
"loss": 0.4774, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.37716615698267075, |
|
"grad_norm": 4.231795787811279, |
|
"learning_rate": 8.325791855203621e-05, |
|
"loss": 0.3297, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.3781855249745158, |
|
"grad_norm": 11.760458946228027, |
|
"learning_rate": 8.348416289592761e-05, |
|
"loss": 1.693, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.37920489296636084, |
|
"grad_norm": 10.05996036529541, |
|
"learning_rate": 8.371040723981901e-05, |
|
"loss": 0.853, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.38022426095820594, |
|
"grad_norm": 8.649154663085938, |
|
"learning_rate": 8.393665158371041e-05, |
|
"loss": 0.7242, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.381243628950051, |
|
"grad_norm": 6.6194748878479, |
|
"learning_rate": 8.416289592760181e-05, |
|
"loss": 0.5019, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.382262996941896, |
|
"grad_norm": 8.058365821838379, |
|
"learning_rate": 8.438914027149321e-05, |
|
"loss": 0.6206, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.38328236493374107, |
|
"grad_norm": 6.66504430770874, |
|
"learning_rate": 8.461538461538461e-05, |
|
"loss": 0.4872, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.3843017329255861, |
|
"grad_norm": 5.8679518699646, |
|
"learning_rate": 8.484162895927601e-05, |
|
"loss": 0.4515, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.3853211009174312, |
|
"grad_norm": 9.830297470092773, |
|
"learning_rate": 8.506787330316743e-05, |
|
"loss": 1.4657, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.38634046890927626, |
|
"grad_norm": 8.260361671447754, |
|
"learning_rate": 8.529411764705883e-05, |
|
"loss": 0.8411, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.3873598369011213, |
|
"grad_norm": 8.48035717010498, |
|
"learning_rate": 8.552036199095023e-05, |
|
"loss": 0.7654, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.38837920489296635, |
|
"grad_norm": 7.481667518615723, |
|
"learning_rate": 8.574660633484163e-05, |
|
"loss": 0.5413, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.3893985728848114, |
|
"grad_norm": 5.923032760620117, |
|
"learning_rate": 8.597285067873304e-05, |
|
"loss": 0.4594, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.3904179408766565, |
|
"grad_norm": 11.383003234863281, |
|
"learning_rate": 8.619909502262445e-05, |
|
"loss": 1.2656, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.39143730886850153, |
|
"grad_norm": 9.154252052307129, |
|
"learning_rate": 8.642533936651585e-05, |
|
"loss": 0.6881, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.3924566768603466, |
|
"grad_norm": 8.656584739685059, |
|
"learning_rate": 8.665158371040725e-05, |
|
"loss": 0.8169, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.3934760448521916, |
|
"grad_norm": 9.6775541305542, |
|
"learning_rate": 8.687782805429865e-05, |
|
"loss": 0.937, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.3944954128440367, |
|
"grad_norm": 12.836816787719727, |
|
"learning_rate": 8.710407239819005e-05, |
|
"loss": 2.1343, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.39551478083588176, |
|
"grad_norm": 6.1532487869262695, |
|
"learning_rate": 8.733031674208145e-05, |
|
"loss": 0.3644, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.3965341488277268, |
|
"grad_norm": 6.3952555656433105, |
|
"learning_rate": 8.755656108597285e-05, |
|
"loss": 0.4406, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.39755351681957185, |
|
"grad_norm": 7.005934238433838, |
|
"learning_rate": 8.778280542986425e-05, |
|
"loss": 0.5444, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.3985728848114169, |
|
"grad_norm": 8.97732925415039, |
|
"learning_rate": 8.800904977375566e-05, |
|
"loss": 1.3891, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.399592252803262, |
|
"grad_norm": 6.8778181076049805, |
|
"learning_rate": 8.823529411764706e-05, |
|
"loss": 0.6287, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.399592252803262, |
|
"eval_Qnli-dev-1024_cosine_accuracy": 0.7395833333333334, |
|
"eval_Qnli-dev-1024_cosine_accuracy_threshold": 0.8465664982795715, |
|
"eval_Qnli-dev-1024_cosine_ap": 0.7683064400770494, |
|
"eval_Qnli-dev-1024_cosine_f1": 0.6976744186046511, |
|
"eval_Qnli-dev-1024_cosine_f1_threshold": 0.834477424621582, |
|
"eval_Qnli-dev-1024_cosine_mcc": 0.45496263625850347, |
|
"eval_Qnli-dev-1024_cosine_precision": 0.7317073170731707, |
|
"eval_Qnli-dev-1024_cosine_recall": 0.6666666666666666, |
|
"eval_Qnli-dev_cosine_accuracy": 0.75, |
|
"eval_Qnli-dev_cosine_accuracy_threshold": 0.74493807554245, |
|
"eval_Qnli-dev_cosine_ap": 0.7575725381948821, |
|
"eval_Qnli-dev_cosine_f1": 0.7476635514018692, |
|
"eval_Qnli-dev_cosine_f1_threshold": 0.7015562057495117, |
|
"eval_Qnli-dev_cosine_mcc": 0.47737827504723207, |
|
"eval_Qnli-dev_cosine_precision": 0.6451612903225806, |
|
"eval_Qnli-dev_cosine_recall": 0.8888888888888888, |
|
"eval_allNLI--triplets-1024_cosine_accuracy": 0.9270833134651184, |
|
"eval_allNLI-triplets_cosine_accuracy": 0.9479166865348816, |
|
"eval_global_dataset_loss": 0.3703947365283966, |
|
"eval_global_dataset_runtime": 104.3143, |
|
"eval_global_dataset_samples_per_second": 7.698, |
|
"eval_global_dataset_steps_per_second": 0.163, |
|
"eval_sequential_score": 0.9270833134651184, |
|
"eval_sts-test-1024_pearson_cosine": 0.8782079507952609, |
|
"eval_sts-test-1024_spearman_cosine": 0.9080003485202497, |
|
"eval_sts-test_pearson_cosine": 0.9052799671643099, |
|
"eval_sts-test_spearman_cosine": 0.9200953636370672, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.40061162079510704, |
|
"grad_norm": 7.236085414886475, |
|
"learning_rate": 8.846153846153847e-05, |
|
"loss": 1.066, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.4016309887869521, |
|
"grad_norm": 7.638827323913574, |
|
"learning_rate": 8.868778280542987e-05, |
|
"loss": 1.0406, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.4026503567787971, |
|
"grad_norm": 6.2278876304626465, |
|
"learning_rate": 8.891402714932127e-05, |
|
"loss": 0.819, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.4036697247706422, |
|
"grad_norm": 7.04884147644043, |
|
"learning_rate": 8.914027149321268e-05, |
|
"loss": 0.5826, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.40468909276248727, |
|
"grad_norm": 8.24869441986084, |
|
"learning_rate": 8.936651583710408e-05, |
|
"loss": 0.6355, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.4057084607543323, |
|
"grad_norm": 9.9276704788208, |
|
"learning_rate": 8.959276018099548e-05, |
|
"loss": 0.7566, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.40672782874617736, |
|
"grad_norm": 8.717905044555664, |
|
"learning_rate": 8.981900452488688e-05, |
|
"loss": 0.8174, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.4077471967380224, |
|
"grad_norm": 8.515538215637207, |
|
"learning_rate": 9.004524886877828e-05, |
|
"loss": 0.8905, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.4087665647298675, |
|
"grad_norm": 6.506967067718506, |
|
"learning_rate": 9.027149321266968e-05, |
|
"loss": 0.6646, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 0.40978593272171254, |
|
"grad_norm": 9.33711051940918, |
|
"learning_rate": 9.049773755656108e-05, |
|
"loss": 0.9056, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.4108053007135576, |
|
"grad_norm": 5.124199867248535, |
|
"learning_rate": 9.07239819004525e-05, |
|
"loss": 0.3689, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.41182466870540263, |
|
"grad_norm": 5.597712516784668, |
|
"learning_rate": 9.09502262443439e-05, |
|
"loss": 0.3709, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.41284403669724773, |
|
"grad_norm": 7.897356033325195, |
|
"learning_rate": 9.11764705882353e-05, |
|
"loss": 0.6708, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.4138634046890928, |
|
"grad_norm": 8.37096881866455, |
|
"learning_rate": 9.14027149321267e-05, |
|
"loss": 1.0531, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.4148827726809378, |
|
"grad_norm": 7.530358791351318, |
|
"learning_rate": 9.16289592760181e-05, |
|
"loss": 1.1355, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 0.41590214067278286, |
|
"grad_norm": 10.304217338562012, |
|
"learning_rate": 9.18552036199095e-05, |
|
"loss": 0.8042, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.4169215086646279, |
|
"grad_norm": 7.292766094207764, |
|
"learning_rate": 9.20814479638009e-05, |
|
"loss": 0.3915, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 0.417940876656473, |
|
"grad_norm": 10.453197479248047, |
|
"learning_rate": 9.230769230769232e-05, |
|
"loss": 1.9388, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.41896024464831805, |
|
"grad_norm": 2.7471694946289062, |
|
"learning_rate": 9.253393665158372e-05, |
|
"loss": 0.3044, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 0.4199796126401631, |
|
"grad_norm": 5.923367023468018, |
|
"learning_rate": 9.276018099547512e-05, |
|
"loss": 0.6153, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 0.42099898063200814, |
|
"grad_norm": 8.176202774047852, |
|
"learning_rate": 9.298642533936652e-05, |
|
"loss": 0.9407, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 0.42201834862385323, |
|
"grad_norm": 8.41361141204834, |
|
"learning_rate": 9.321266968325792e-05, |
|
"loss": 0.6876, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.4230377166156983, |
|
"grad_norm": 9.516852378845215, |
|
"learning_rate": 9.343891402714933e-05, |
|
"loss": 0.9694, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.4240570846075433, |
|
"grad_norm": 7.201638698577881, |
|
"learning_rate": 9.366515837104073e-05, |
|
"loss": 0.7868, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 0.42507645259938837, |
|
"grad_norm": 9.961840629577637, |
|
"learning_rate": 9.389140271493213e-05, |
|
"loss": 0.7735, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 0.4260958205912334, |
|
"grad_norm": 10.842241287231445, |
|
"learning_rate": 9.411764705882353e-05, |
|
"loss": 1.1682, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 0.4271151885830785, |
|
"grad_norm": 5.817572116851807, |
|
"learning_rate": 9.434389140271494e-05, |
|
"loss": 0.3465, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 0.42813455657492355, |
|
"grad_norm": 6.870133399963379, |
|
"learning_rate": 9.457013574660634e-05, |
|
"loss": 0.5699, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.4291539245667686, |
|
"grad_norm": 6.472342014312744, |
|
"learning_rate": 9.479638009049774e-05, |
|
"loss": 0.6128, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 0.43017329255861364, |
|
"grad_norm": 6.5723795890808105, |
|
"learning_rate": 9.502262443438914e-05, |
|
"loss": 0.8886, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 0.43119266055045874, |
|
"grad_norm": 6.1384429931640625, |
|
"learning_rate": 9.524886877828054e-05, |
|
"loss": 0.5124, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 0.4322120285423038, |
|
"grad_norm": 6.241471290588379, |
|
"learning_rate": 9.547511312217195e-05, |
|
"loss": 0.4409, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 0.4332313965341488, |
|
"grad_norm": 9.087861061096191, |
|
"learning_rate": 9.570135746606335e-05, |
|
"loss": 0.6368, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.43425076452599387, |
|
"grad_norm": 9.653539657592773, |
|
"learning_rate": 9.592760180995475e-05, |
|
"loss": 0.9874, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 0.4352701325178389, |
|
"grad_norm": 13.366517066955566, |
|
"learning_rate": 9.615384615384617e-05, |
|
"loss": 1.6544, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 0.436289500509684, |
|
"grad_norm": 6.302597522735596, |
|
"learning_rate": 9.638009049773757e-05, |
|
"loss": 0.4561, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 0.43730886850152906, |
|
"grad_norm": 7.133030891418457, |
|
"learning_rate": 9.660633484162897e-05, |
|
"loss": 0.5443, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 0.4383282364933741, |
|
"grad_norm": 6.341556072235107, |
|
"learning_rate": 9.683257918552037e-05, |
|
"loss": 0.5183, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.43934760448521915, |
|
"grad_norm": 10.657116889953613, |
|
"learning_rate": 9.705882352941177e-05, |
|
"loss": 1.1585, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 0.44036697247706424, |
|
"grad_norm": 7.707142353057861, |
|
"learning_rate": 9.728506787330317e-05, |
|
"loss": 1.4285, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.4413863404689093, |
|
"grad_norm": 8.27905559539795, |
|
"learning_rate": 9.751131221719457e-05, |
|
"loss": 1.0638, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 0.44240570846075433, |
|
"grad_norm": 5.601058483123779, |
|
"learning_rate": 9.773755656108597e-05, |
|
"loss": 0.553, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 0.4434250764525994, |
|
"grad_norm": 9.084299087524414, |
|
"learning_rate": 9.796380090497737e-05, |
|
"loss": 1.0009, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.4444444444444444, |
|
"grad_norm": 5.231532573699951, |
|
"learning_rate": 9.819004524886877e-05, |
|
"loss": 0.5211, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 0.4454638124362895, |
|
"grad_norm": 7.0478715896606445, |
|
"learning_rate": 9.841628959276019e-05, |
|
"loss": 0.6483, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 0.44648318042813456, |
|
"grad_norm": 8.44166088104248, |
|
"learning_rate": 9.864253393665159e-05, |
|
"loss": 1.2634, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 0.4475025484199796, |
|
"grad_norm": 7.2984771728515625, |
|
"learning_rate": 9.8868778280543e-05, |
|
"loss": 0.5242, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 0.44852191641182465, |
|
"grad_norm": 9.091867446899414, |
|
"learning_rate": 9.90950226244344e-05, |
|
"loss": 1.3739, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.44954128440366975, |
|
"grad_norm": 8.1068115234375, |
|
"learning_rate": 9.93212669683258e-05, |
|
"loss": 1.0153, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 0.4505606523955148, |
|
"grad_norm": 7.902680397033691, |
|
"learning_rate": 9.95475113122172e-05, |
|
"loss": 0.7174, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 0.45158002038735984, |
|
"grad_norm": 8.784537315368652, |
|
"learning_rate": 9.97737556561086e-05, |
|
"loss": 0.8631, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 0.4525993883792049, |
|
"grad_norm": 8.205148696899414, |
|
"learning_rate": 0.0001, |
|
"loss": 1.0721, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 0.4536187563710499, |
|
"grad_norm": 4.789169788360596, |
|
"learning_rate": 9.999964497873585e-05, |
|
"loss": 0.3682, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.454638124362895, |
|
"grad_norm": 11.335341453552246, |
|
"learning_rate": 9.999857991998499e-05, |
|
"loss": 1.3278, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 0.45565749235474007, |
|
"grad_norm": 8.901962280273438, |
|
"learning_rate": 9.999680483887217e-05, |
|
"loss": 0.665, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 0.4566768603465851, |
|
"grad_norm": 6.525248050689697, |
|
"learning_rate": 9.999431976060504e-05, |
|
"loss": 0.77, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 0.45769622833843016, |
|
"grad_norm": 7.658937931060791, |
|
"learning_rate": 9.999112472047386e-05, |
|
"loss": 0.9903, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 0.45871559633027525, |
|
"grad_norm": 5.406915664672852, |
|
"learning_rate": 9.998721976385087e-05, |
|
"loss": 0.3372, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.4597349643221203, |
|
"grad_norm": 5.920129299163818, |
|
"learning_rate": 9.998260494618979e-05, |
|
"loss": 0.6911, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 0.46075433231396534, |
|
"grad_norm": 7.490262985229492, |
|
"learning_rate": 9.997728033302496e-05, |
|
"loss": 0.505, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 0.4617737003058104, |
|
"grad_norm": 8.21649169921875, |
|
"learning_rate": 9.997124599997043e-05, |
|
"loss": 1.3397, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 0.46279306829765543, |
|
"grad_norm": 5.116532802581787, |
|
"learning_rate": 9.996450203271886e-05, |
|
"loss": 0.2853, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 0.46381243628950053, |
|
"grad_norm": 7.29067325592041, |
|
"learning_rate": 9.995704852704029e-05, |
|
"loss": 1.198, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.4648318042813456, |
|
"grad_norm": 10.033268928527832, |
|
"learning_rate": 9.994888558878086e-05, |
|
"loss": 1.7965, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 0.4658511722731906, |
|
"grad_norm": 5.4102606773376465, |
|
"learning_rate": 9.994001333386125e-05, |
|
"loss": 0.2987, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 0.46687054026503566, |
|
"grad_norm": 8.109895706176758, |
|
"learning_rate": 9.993043188827501e-05, |
|
"loss": 0.6864, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 0.46788990825688076, |
|
"grad_norm": 9.893292427062988, |
|
"learning_rate": 9.992014138808682e-05, |
|
"loss": 0.9016, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 0.4689092762487258, |
|
"grad_norm": 7.73169469833374, |
|
"learning_rate": 9.990914197943053e-05, |
|
"loss": 0.7314, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.46992864424057085, |
|
"grad_norm": 8.335735321044922, |
|
"learning_rate": 9.989743381850711e-05, |
|
"loss": 0.6633, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 0.4709480122324159, |
|
"grad_norm": 8.655631065368652, |
|
"learning_rate": 9.988501707158243e-05, |
|
"loss": 0.9783, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 0.47196738022426094, |
|
"grad_norm": 9.166102409362793, |
|
"learning_rate": 9.987189191498479e-05, |
|
"loss": 1.1307, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 0.47298674821610603, |
|
"grad_norm": 10.597552299499512, |
|
"learning_rate": 9.985805853510262e-05, |
|
"loss": 1.4662, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 0.4740061162079511, |
|
"grad_norm": 10.318975448608398, |
|
"learning_rate": 9.984351712838167e-05, |
|
"loss": 1.4666, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.4750254841997961, |
|
"grad_norm": 7.259106636047363, |
|
"learning_rate": 9.98282679013223e-05, |
|
"loss": 0.5918, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 0.47604485219164117, |
|
"grad_norm": 10.192667961120605, |
|
"learning_rate": 9.981231107047648e-05, |
|
"loss": 1.5836, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 0.47706422018348627, |
|
"grad_norm": 6.506603717803955, |
|
"learning_rate": 9.97956468624448e-05, |
|
"loss": 0.5698, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 0.4780835881753313, |
|
"grad_norm": 5.789127349853516, |
|
"learning_rate": 9.977827551387318e-05, |
|
"loss": 0.4654, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 0.47910295616717635, |
|
"grad_norm": 8.615316390991211, |
|
"learning_rate": 9.976019727144956e-05, |
|
"loss": 0.9522, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.4801223241590214, |
|
"grad_norm": 4.792436599731445, |
|
"learning_rate": 9.974141239190034e-05, |
|
"loss": 0.4748, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 0.48114169215086644, |
|
"grad_norm": 9.958406448364258, |
|
"learning_rate": 9.972192114198677e-05, |
|
"loss": 2.0766, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 0.48216106014271154, |
|
"grad_norm": 4.340735912322998, |
|
"learning_rate": 9.970172379850122e-05, |
|
"loss": 0.7071, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 0.4831804281345566, |
|
"grad_norm": 7.170680999755859, |
|
"learning_rate": 9.968082064826314e-05, |
|
"loss": 0.435, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 0.4841997961264016, |
|
"grad_norm": 3.927189350128174, |
|
"learning_rate": 9.965921198811501e-05, |
|
"loss": 0.4551, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.48521916411824667, |
|
"grad_norm": 10.183062553405762, |
|
"learning_rate": 9.96368981249182e-05, |
|
"loss": 1.1758, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 0.48623853211009177, |
|
"grad_norm": 9.819293022155762, |
|
"learning_rate": 9.961387937554857e-05, |
|
"loss": 0.9995, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 0.4872579001019368, |
|
"grad_norm": 11.188612937927246, |
|
"learning_rate": 9.95901560668919e-05, |
|
"loss": 1.6207, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 0.48827726809378186, |
|
"grad_norm": 7.268994331359863, |
|
"learning_rate": 9.95657285358394e-05, |
|
"loss": 0.6978, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.4892966360856269, |
|
"grad_norm": 5.575627326965332, |
|
"learning_rate": 9.954059712928275e-05, |
|
"loss": 0.4236, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.49031600407747195, |
|
"grad_norm": 9.621591567993164, |
|
"learning_rate": 9.951476220410929e-05, |
|
"loss": 1.8218, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 0.49133537206931704, |
|
"grad_norm": 7.322023391723633, |
|
"learning_rate": 9.948822412719697e-05, |
|
"loss": 0.8749, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 0.4923547400611621, |
|
"grad_norm": 8.407424926757812, |
|
"learning_rate": 9.946098327540902e-05, |
|
"loss": 1.1704, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 0.49337410805300713, |
|
"grad_norm": 4.8855438232421875, |
|
"learning_rate": 9.943304003558873e-05, |
|
"loss": 0.5327, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 0.4943934760448522, |
|
"grad_norm": 8.738515853881836, |
|
"learning_rate": 9.940439480455386e-05, |
|
"loss": 1.2009, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.4954128440366973, |
|
"grad_norm": 7.554356575012207, |
|
"learning_rate": 9.937504798909106e-05, |
|
"loss": 0.5427, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 0.4964322120285423, |
|
"grad_norm": 8.203272819519043, |
|
"learning_rate": 9.934500000595008e-05, |
|
"loss": 0.5893, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 0.49745158002038736, |
|
"grad_norm": 8.477286338806152, |
|
"learning_rate": 9.931425128183782e-05, |
|
"loss": 1.061, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 0.4984709480122324, |
|
"grad_norm": 7.389923095703125, |
|
"learning_rate": 9.928280225341232e-05, |
|
"loss": 0.5465, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 0.49949031600407745, |
|
"grad_norm": 10.051106452941895, |
|
"learning_rate": 9.925065336727654e-05, |
|
"loss": 0.7035, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.49949031600407745, |
|
"eval_Qnli-dev-1024_cosine_accuracy": 0.7291666666666666, |
|
"eval_Qnli-dev-1024_cosine_accuracy_threshold": 0.817620038986206, |
|
"eval_Qnli-dev-1024_cosine_ap": 0.7443202788050278, |
|
"eval_Qnli-dev-1024_cosine_f1": 0.7291666666666667, |
|
"eval_Qnli-dev-1024_cosine_f1_threshold": 0.817620038986206, |
|
"eval_Qnli-dev-1024_cosine_mcc": 0.46405228758169936, |
|
"eval_Qnli-dev-1024_cosine_precision": 0.6862745098039216, |
|
"eval_Qnli-dev-1024_cosine_recall": 0.7777777777777778, |
|
"eval_Qnli-dev_cosine_accuracy": 0.7395833333333334, |
|
"eval_Qnli-dev_cosine_accuracy_threshold": 0.7795548439025879, |
|
"eval_Qnli-dev_cosine_ap": 0.7446338608862075, |
|
"eval_Qnli-dev_cosine_f1": 0.7378640776699029, |
|
"eval_Qnli-dev_cosine_f1_threshold": 0.6985307335853577, |
|
"eval_Qnli-dev_cosine_mcc": 0.46153029495329345, |
|
"eval_Qnli-dev_cosine_precision": 0.6551724137931034, |
|
"eval_Qnli-dev_cosine_recall": 0.8444444444444444, |
|
"eval_allNLI--triplets-1024_cosine_accuracy": 0.90625, |
|
"eval_allNLI-triplets_cosine_accuracy": 0.9375, |
|
"eval_global_dataset_loss": 0.34814590215682983, |
|
"eval_global_dataset_runtime": 104.2751, |
|
"eval_global_dataset_samples_per_second": 7.701, |
|
"eval_global_dataset_steps_per_second": 0.163, |
|
"eval_sequential_score": 0.90625, |
|
"eval_sts-test-1024_pearson_cosine": 0.8574057933500303, |
|
"eval_sts-test-1024_spearman_cosine": 0.8986116241995802, |
|
"eval_sts-test_pearson_cosine": 0.9019111579722014, |
|
"eval_sts-test_spearman_cosine": 0.9181479205822737, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.5005096839959225, |
|
"grad_norm": 10.88537311553955, |
|
"learning_rate": 9.921780507997202e-05, |
|
"loss": 1.6596, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 0.5015290519877675, |
|
"grad_norm": 5.0818891525268555, |
|
"learning_rate": 9.918425785797235e-05, |
|
"loss": 0.4475, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 0.5025484199796126, |
|
"grad_norm": 12.540839195251465, |
|
"learning_rate": 9.915001217767663e-05, |
|
"loss": 2.0803, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 0.5035677879714577, |
|
"grad_norm": 6.171934604644775, |
|
"learning_rate": 9.911506852540267e-05, |
|
"loss": 0.4296, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 0.5045871559633027, |
|
"grad_norm": 9.624109268188477, |
|
"learning_rate": 9.907942739738001e-05, |
|
"loss": 1.5435, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.5056065239551478, |
|
"grad_norm": 6.020090579986572, |
|
"learning_rate": 9.904308929974302e-05, |
|
"loss": 0.6073, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 0.5066258919469928, |
|
"grad_norm": 8.587658882141113, |
|
"learning_rate": 9.900605474852358e-05, |
|
"loss": 1.1774, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 0.5076452599388379, |
|
"grad_norm": 6.535181999206543, |
|
"learning_rate": 9.896832426964382e-05, |
|
"loss": 0.8951, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 0.508664627930683, |
|
"grad_norm": 5.945138454437256, |
|
"learning_rate": 9.892989839890863e-05, |
|
"loss": 0.3775, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 0.509683995922528, |
|
"grad_norm": 7.641120433807373, |
|
"learning_rate": 9.889077768199806e-05, |
|
"loss": 0.8086, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.5107033639143731, |
|
"grad_norm": 8.10549545288086, |
|
"learning_rate": 9.885096267445957e-05, |
|
"loss": 0.3864, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 0.5117227319062182, |
|
"grad_norm": 7.266530990600586, |
|
"learning_rate": 9.881045394170012e-05, |
|
"loss": 0.8865, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 0.5127420998980632, |
|
"grad_norm": 9.056779861450195, |
|
"learning_rate": 9.876925205897818e-05, |
|
"loss": 0.567, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 0.5137614678899083, |
|
"grad_norm": 7.140566349029541, |
|
"learning_rate": 9.872735761139554e-05, |
|
"loss": 0.9304, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 0.5147808358817533, |
|
"grad_norm": 11.422016143798828, |
|
"learning_rate": 9.868477119388896e-05, |
|
"loss": 0.6977, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.5158002038735984, |
|
"grad_norm": 11.155719757080078, |
|
"learning_rate": 9.864149341122181e-05, |
|
"loss": 1.3174, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 0.5168195718654435, |
|
"grad_norm": 8.781103134155273, |
|
"learning_rate": 9.859752487797542e-05, |
|
"loss": 1.2481, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 0.5178389398572885, |
|
"grad_norm": 5.503263473510742, |
|
"learning_rate": 9.855286621854034e-05, |
|
"loss": 0.4894, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 0.5188583078491336, |
|
"grad_norm": 7.503839015960693, |
|
"learning_rate": 9.850751806710753e-05, |
|
"loss": 0.8095, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 0.5198776758409785, |
|
"grad_norm": 5.623706817626953, |
|
"learning_rate": 9.846148106765933e-05, |
|
"loss": 0.388, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.5208970438328236, |
|
"grad_norm": 5.178555965423584, |
|
"learning_rate": 9.841475587396028e-05, |
|
"loss": 0.6725, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 0.5219164118246687, |
|
"grad_norm": 7.296833038330078, |
|
"learning_rate": 9.836734314954785e-05, |
|
"loss": 0.4804, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 0.5229357798165137, |
|
"grad_norm": 8.692532539367676, |
|
"learning_rate": 9.831924356772308e-05, |
|
"loss": 1.2414, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 0.5239551478083588, |
|
"grad_norm": 9.865914344787598, |
|
"learning_rate": 9.827045781154093e-05, |
|
"loss": 1.0319, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 0.5249745158002038, |
|
"grad_norm": 14.857895851135254, |
|
"learning_rate": 9.822098657380065e-05, |
|
"loss": 2.0732, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.5259938837920489, |
|
"grad_norm": 6.85409688949585, |
|
"learning_rate": 9.817083055703587e-05, |
|
"loss": 1.1168, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 0.527013251783894, |
|
"grad_norm": 10.668725967407227, |
|
"learning_rate": 9.811999047350471e-05, |
|
"loss": 0.8056, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 0.528032619775739, |
|
"grad_norm": 6.76224946975708, |
|
"learning_rate": 9.806846704517957e-05, |
|
"loss": 0.5322, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 0.5290519877675841, |
|
"grad_norm": 4.4465789794921875, |
|
"learning_rate": 9.801626100373699e-05, |
|
"loss": 0.4348, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 0.5300713557594292, |
|
"grad_norm": 8.388195991516113, |
|
"learning_rate": 9.796337309054717e-05, |
|
"loss": 0.6316, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.5310907237512742, |
|
"grad_norm": 5.859539031982422, |
|
"learning_rate": 9.790980405666344e-05, |
|
"loss": 0.3212, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 0.5321100917431193, |
|
"grad_norm": 6.299170017242432, |
|
"learning_rate": 9.785555466281162e-05, |
|
"loss": 0.4739, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 0.5331294597349643, |
|
"grad_norm": 9.609426498413086, |
|
"learning_rate": 9.780062567937928e-05, |
|
"loss": 1.1692, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 0.5341488277268094, |
|
"grad_norm": 9.116230964660645, |
|
"learning_rate": 9.774501788640471e-05, |
|
"loss": 0.941, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 0.5351681957186545, |
|
"grad_norm": 5.024673938751221, |
|
"learning_rate": 9.768873207356586e-05, |
|
"loss": 0.3767, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.5361875637104995, |
|
"grad_norm": 7.534763336181641, |
|
"learning_rate": 9.763176904016913e-05, |
|
"loss": 0.5264, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 0.5372069317023446, |
|
"grad_norm": 7.897163391113281, |
|
"learning_rate": 9.757412959513807e-05, |
|
"loss": 0.4345, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 0.5382262996941896, |
|
"grad_norm": 8.391239166259766, |
|
"learning_rate": 9.751581455700181e-05, |
|
"loss": 1.0352, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 0.5392456676860347, |
|
"grad_norm": 6.951046466827393, |
|
"learning_rate": 9.745682475388348e-05, |
|
"loss": 1.1014, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 0.5402650356778798, |
|
"grad_norm": 6.4283671379089355, |
|
"learning_rate": 9.73971610234885e-05, |
|
"loss": 0.7368, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.5412844036697247, |
|
"grad_norm": 7.643414497375488, |
|
"learning_rate": 9.733682421309256e-05, |
|
"loss": 0.5324, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 0.5423037716615698, |
|
"grad_norm": 7.95609188079834, |
|
"learning_rate": 9.727581517952969e-05, |
|
"loss": 0.5351, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 0.5433231396534148, |
|
"grad_norm": 11.28146743774414, |
|
"learning_rate": 9.721413478918007e-05, |
|
"loss": 1.6815, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 0.5443425076452599, |
|
"grad_norm": 8.346885681152344, |
|
"learning_rate": 9.715178391795769e-05, |
|
"loss": 0.8125, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 0.545361875637105, |
|
"grad_norm": 8.147517204284668, |
|
"learning_rate": 9.708876345129797e-05, |
|
"loss": 0.8629, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.54638124362895, |
|
"grad_norm": 10.061439514160156, |
|
"learning_rate": 9.702507428414513e-05, |
|
"loss": 1.3161, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 0.5474006116207951, |
|
"grad_norm": 8.882964134216309, |
|
"learning_rate": 9.696071732093952e-05, |
|
"loss": 1.0465, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 0.5484199796126402, |
|
"grad_norm": 5.954410076141357, |
|
"learning_rate": 9.689569347560475e-05, |
|
"loss": 0.4531, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 0.5494393476044852, |
|
"grad_norm": 10.33085823059082, |
|
"learning_rate": 9.683000367153474e-05, |
|
"loss": 0.5567, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 0.5504587155963303, |
|
"grad_norm": 5.265343189239502, |
|
"learning_rate": 9.676364884158058e-05, |
|
"loss": 0.7093, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.5514780835881753, |
|
"grad_norm": 10.214452743530273, |
|
"learning_rate": 9.66966299280373e-05, |
|
"loss": 1.9339, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 0.5524974515800204, |
|
"grad_norm": 7.001688480377197, |
|
"learning_rate": 9.662894788263044e-05, |
|
"loss": 0.3659, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 0.5535168195718655, |
|
"grad_norm": 6.640339374542236, |
|
"learning_rate": 9.656060366650267e-05, |
|
"loss": 1.0505, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 0.5545361875637105, |
|
"grad_norm": 9.303877830505371, |
|
"learning_rate": 9.649159825019996e-05, |
|
"loss": 0.8766, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 0.5555555555555556, |
|
"grad_norm": 8.21275806427002, |
|
"learning_rate": 9.642193261365791e-05, |
|
"loss": 0.6526, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.5565749235474006, |
|
"grad_norm": 6.97646427154541, |
|
"learning_rate": 9.635160774618782e-05, |
|
"loss": 0.5529, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 0.5575942915392457, |
|
"grad_norm": 6.77686071395874, |
|
"learning_rate": 9.628062464646264e-05, |
|
"loss": 0.4817, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 0.5586136595310908, |
|
"grad_norm": 3.5217092037200928, |
|
"learning_rate": 9.620898432250272e-05, |
|
"loss": 0.4804, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 0.5596330275229358, |
|
"grad_norm": 5.6369476318359375, |
|
"learning_rate": 9.613668779166165e-05, |
|
"loss": 0.4508, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 0.5606523955147809, |
|
"grad_norm": 5.534257888793945, |
|
"learning_rate": 9.606373608061162e-05, |
|
"loss": 0.4339, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.5616717635066258, |
|
"grad_norm": 10.922380447387695, |
|
"learning_rate": 9.5990130225329e-05, |
|
"loss": 0.712, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 0.5626911314984709, |
|
"grad_norm": 6.2288360595703125, |
|
"learning_rate": 9.59158712710795e-05, |
|
"loss": 0.3974, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 0.563710499490316, |
|
"grad_norm": 11.958196640014648, |
|
"learning_rate": 9.58409602724035e-05, |
|
"loss": 1.0016, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 0.564729867482161, |
|
"grad_norm": 8.267114639282227, |
|
"learning_rate": 9.576539829310085e-05, |
|
"loss": 0.5751, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 0.5657492354740061, |
|
"grad_norm": 11.533574104309082, |
|
"learning_rate": 9.568918640621594e-05, |
|
"loss": 1.111, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.5667686034658511, |
|
"grad_norm": 6.519062519073486, |
|
"learning_rate": 9.561232569402239e-05, |
|
"loss": 0.4202, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 0.5677879714576962, |
|
"grad_norm": 9.009593963623047, |
|
"learning_rate": 9.553481724800768e-05, |
|
"loss": 0.7822, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.5688073394495413, |
|
"grad_norm": 12.121257781982422, |
|
"learning_rate": 9.545666216885767e-05, |
|
"loss": 1.3844, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 0.5698267074413863, |
|
"grad_norm": 5.953427314758301, |
|
"learning_rate": 9.537786156644097e-05, |
|
"loss": 0.3881, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 0.5708460754332314, |
|
"grad_norm": 7.334780216217041, |
|
"learning_rate": 9.529841655979315e-05, |
|
"loss": 0.6317, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.5718654434250765, |
|
"grad_norm": 5.987368583679199, |
|
"learning_rate": 9.521832827710088e-05, |
|
"loss": 0.4976, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 0.5728848114169215, |
|
"grad_norm": 3.9462735652923584, |
|
"learning_rate": 9.51375978556859e-05, |
|
"loss": 0.2741, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 0.5739041794087666, |
|
"grad_norm": 6.374652862548828, |
|
"learning_rate": 9.505622644198885e-05, |
|
"loss": 0.6232, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 0.5749235474006116, |
|
"grad_norm": 3.525486707687378, |
|
"learning_rate": 9.497421519155303e-05, |
|
"loss": 0.2083, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 0.5759429153924567, |
|
"grad_norm": 9.60029125213623, |
|
"learning_rate": 9.489156526900795e-05, |
|
"loss": 1.0605, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.5769622833843018, |
|
"grad_norm": 12.22358226776123, |
|
"learning_rate": 9.480827784805278e-05, |
|
"loss": 1.2086, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 0.5779816513761468, |
|
"grad_norm": 4.388841152191162, |
|
"learning_rate": 9.472435411143978e-05, |
|
"loss": 0.2217, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 0.5790010193679919, |
|
"grad_norm": 5.581283092498779, |
|
"learning_rate": 9.463979525095738e-05, |
|
"loss": 0.4215, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 0.5800203873598369, |
|
"grad_norm": 7.996876239776611, |
|
"learning_rate": 9.455460246741331e-05, |
|
"loss": 0.663, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 0.581039755351682, |
|
"grad_norm": 9.21956729888916, |
|
"learning_rate": 9.446877697061757e-05, |
|
"loss": 0.653, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.582059123343527, |
|
"grad_norm": 8.46827220916748, |
|
"learning_rate": 9.43823199793652e-05, |
|
"loss": 0.6895, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 0.583078491335372, |
|
"grad_norm": 9.72203540802002, |
|
"learning_rate": 9.429523272141903e-05, |
|
"loss": 1.1101, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 0.5840978593272171, |
|
"grad_norm": 8.79525089263916, |
|
"learning_rate": 9.420751643349219e-05, |
|
"loss": 1.2991, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 0.5851172273190621, |
|
"grad_norm": 6.719937801361084, |
|
"learning_rate": 9.411917236123059e-05, |
|
"loss": 0.4072, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 0.5861365953109072, |
|
"grad_norm": 8.360040664672852, |
|
"learning_rate": 9.403020175919517e-05, |
|
"loss": 1.169, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.5871559633027523, |
|
"grad_norm": 5.402820587158203, |
|
"learning_rate": 9.394060589084417e-05, |
|
"loss": 0.3374, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 0.5881753312945973, |
|
"grad_norm": 9.037818908691406, |
|
"learning_rate": 9.385038602851515e-05, |
|
"loss": 0.6785, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 0.5891946992864424, |
|
"grad_norm": 9.151761054992676, |
|
"learning_rate": 9.375954345340685e-05, |
|
"loss": 1.2757, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 0.5902140672782875, |
|
"grad_norm": 5.834461212158203, |
|
"learning_rate": 9.366807945556113e-05, |
|
"loss": 0.5899, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 0.5912334352701325, |
|
"grad_norm": 5.722581386566162, |
|
"learning_rate": 9.357599533384453e-05, |
|
"loss": 0.3389, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.5922528032619776, |
|
"grad_norm": 10.132628440856934, |
|
"learning_rate": 9.348329239592995e-05, |
|
"loss": 1.631, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 0.5932721712538226, |
|
"grad_norm": 9.922087669372559, |
|
"learning_rate": 9.338997195827792e-05, |
|
"loss": 1.3975, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 0.5942915392456677, |
|
"grad_norm": 8.382550239562988, |
|
"learning_rate": 9.329603534611806e-05, |
|
"loss": 0.4654, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 0.5953109072375128, |
|
"grad_norm": 8.080007553100586, |
|
"learning_rate": 9.32014838934301e-05, |
|
"loss": 0.56, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 0.5963302752293578, |
|
"grad_norm": 5.616114616394043, |
|
"learning_rate": 9.310631894292518e-05, |
|
"loss": 0.2282, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.5973496432212029, |
|
"grad_norm": 10.813580513000488, |
|
"learning_rate": 9.301054184602647e-05, |
|
"loss": 1.0754, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 0.5983690112130479, |
|
"grad_norm": 8.062788963317871, |
|
"learning_rate": 9.291415396285024e-05, |
|
"loss": 0.4411, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 0.599388379204893, |
|
"grad_norm": 8.6395845413208, |
|
"learning_rate": 9.281715666218643e-05, |
|
"loss": 0.9243, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 0.599388379204893, |
|
"eval_Qnli-dev-1024_cosine_accuracy": 0.71875, |
|
"eval_Qnli-dev-1024_cosine_accuracy_threshold": 0.8436912298202515, |
|
"eval_Qnli-dev-1024_cosine_ap": 0.7587494204458187, |
|
"eval_Qnli-dev-1024_cosine_f1": 0.6875, |
|
"eval_Qnli-dev-1024_cosine_f1_threshold": 0.8028630018234253, |
|
"eval_Qnli-dev-1024_cosine_mcc": 0.3803921568627451, |
|
"eval_Qnli-dev-1024_cosine_precision": 0.6470588235294118, |
|
"eval_Qnli-dev-1024_cosine_recall": 0.7333333333333333, |
|
"eval_Qnli-dev_cosine_accuracy": 0.71875, |
|
"eval_Qnli-dev_cosine_accuracy_threshold": 0.8078321218490601, |
|
"eval_Qnli-dev_cosine_ap": 0.7321739553695406, |
|
"eval_Qnli-dev_cosine_f1": 0.7339449541284404, |
|
"eval_Qnli-dev_cosine_f1_threshold": 0.6781572699546814, |
|
"eval_Qnli-dev_cosine_mcc": 0.4428074427700477, |
|
"eval_Qnli-dev_cosine_precision": 0.625, |
|
"eval_Qnli-dev_cosine_recall": 0.8888888888888888, |
|
"eval_allNLI--triplets-1024_cosine_accuracy": 0.90625, |
|
"eval_allNLI-triplets_cosine_accuracy": 0.9375, |
|
"eval_global_dataset_loss": 0.36118289828300476, |
|
"eval_global_dataset_runtime": 104.3983, |
|
"eval_global_dataset_samples_per_second": 7.692, |
|
"eval_global_dataset_steps_per_second": 0.163, |
|
"eval_sequential_score": 0.90625, |
|
"eval_sts-test-1024_pearson_cosine": 0.8631921152381832, |
|
"eval_sts-test-1024_spearman_cosine": 0.9009700758334896, |
|
"eval_sts-test_pearson_cosine": 0.9009355736320144, |
|
"eval_sts-test_spearman_cosine": 0.9171725695772274, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 0.6004077471967381, |
|
"grad_norm": 6.184821128845215, |
|
"learning_rate": 9.271955132147916e-05, |
|
"loss": 0.3572, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 0.601427115188583, |
|
"grad_norm": 8.318941116333008, |
|
"learning_rate": 9.262133932680733e-05, |
|
"loss": 0.6761, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.6024464831804281, |
|
"grad_norm": 7.54533052444458, |
|
"learning_rate": 9.252252207286479e-05, |
|
"loss": 0.5754, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 0.6034658511722731, |
|
"grad_norm": 4.341547012329102, |
|
"learning_rate": 9.24231009629406e-05, |
|
"loss": 0.3664, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 0.6044852191641182, |
|
"grad_norm": 7.616749286651611, |
|
"learning_rate": 9.232307740889909e-05, |
|
"loss": 0.5391, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 0.6055045871559633, |
|
"grad_norm": 4.843873977661133, |
|
"learning_rate": 9.222245283115979e-05, |
|
"loss": 0.518, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 0.6065239551478083, |
|
"grad_norm": 8.295080184936523, |
|
"learning_rate": 9.21212286586773e-05, |
|
"loss": 0.5263, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.6075433231396534, |
|
"grad_norm": 3.873260736465454, |
|
"learning_rate": 9.201940632892096e-05, |
|
"loss": 0.2995, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 0.6085626911314985, |
|
"grad_norm": 4.403683185577393, |
|
"learning_rate": 9.191698728785448e-05, |
|
"loss": 0.4181, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 0.6095820591233435, |
|
"grad_norm": 7.282264709472656, |
|
"learning_rate": 9.181397298991532e-05, |
|
"loss": 0.5087, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 0.6106014271151886, |
|
"grad_norm": 5.132986068725586, |
|
"learning_rate": 9.171036489799416e-05, |
|
"loss": 0.4344, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 0.6116207951070336, |
|
"grad_norm": 11.096871376037598, |
|
"learning_rate": 9.160616448341403e-05, |
|
"loss": 1.6529, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.6126401630988787, |
|
"grad_norm": 4.306335926055908, |
|
"learning_rate": 9.150137322590944e-05, |
|
"loss": 0.4079, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 0.6136595310907238, |
|
"grad_norm": 5.622674942016602, |
|
"learning_rate": 9.139599261360537e-05, |
|
"loss": 0.3123, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 0.6146788990825688, |
|
"grad_norm": 10.172139167785645, |
|
"learning_rate": 9.129002414299617e-05, |
|
"loss": 1.4398, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 0.6156982670744139, |
|
"grad_norm": 10.175543785095215, |
|
"learning_rate": 9.118346931892423e-05, |
|
"loss": 1.5553, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 0.6167176350662589, |
|
"grad_norm": 7.616044521331787, |
|
"learning_rate": 9.10763296545587e-05, |
|
"loss": 0.7958, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.617737003058104, |
|
"grad_norm": 5.390756607055664, |
|
"learning_rate": 9.096860667137394e-05, |
|
"loss": 0.3815, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 0.6187563710499491, |
|
"grad_norm": 6.750911235809326, |
|
"learning_rate": 9.086030189912794e-05, |
|
"loss": 0.787, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 0.6197757390417941, |
|
"grad_norm": 5.695408344268799, |
|
"learning_rate": 9.075141687584057e-05, |
|
"loss": 0.2352, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 0.6207951070336392, |
|
"grad_norm": 12.017024040222168, |
|
"learning_rate": 9.06419531477718e-05, |
|
"loss": 0.6469, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 0.6218144750254841, |
|
"grad_norm": 9.70870304107666, |
|
"learning_rate": 9.053191226939965e-05, |
|
"loss": 1.0997, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.6228338430173292, |
|
"grad_norm": 8.183333396911621, |
|
"learning_rate": 9.042129580339822e-05, |
|
"loss": 0.8762, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 0.6238532110091743, |
|
"grad_norm": 8.237792015075684, |
|
"learning_rate": 9.031010532061538e-05, |
|
"loss": 0.6259, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 0.6248725790010193, |
|
"grad_norm": 7.553733825683594, |
|
"learning_rate": 9.019834240005058e-05, |
|
"loss": 0.8133, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 0.6258919469928644, |
|
"grad_norm": 8.876506805419922, |
|
"learning_rate": 9.008600862883235e-05, |
|
"loss": 1.2119, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 0.6269113149847095, |
|
"grad_norm": 6.738461971282959, |
|
"learning_rate": 8.997310560219578e-05, |
|
"loss": 0.5068, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.6279306829765545, |
|
"grad_norm": 9.400090217590332, |
|
"learning_rate": 8.985963492345991e-05, |
|
"loss": 0.7723, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 0.6289500509683996, |
|
"grad_norm": 8.690120697021484, |
|
"learning_rate": 8.974559820400486e-05, |
|
"loss": 1.0791, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 0.6299694189602446, |
|
"grad_norm": 5.441365718841553, |
|
"learning_rate": 8.963099706324904e-05, |
|
"loss": 0.348, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 0.6309887869520897, |
|
"grad_norm": 8.373964309692383, |
|
"learning_rate": 8.951583312862616e-05, |
|
"loss": 0.6523, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 0.6320081549439348, |
|
"grad_norm": 8.361169815063477, |
|
"learning_rate": 8.9400108035562e-05, |
|
"loss": 0.6241, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.6330275229357798, |
|
"grad_norm": 8.095520973205566, |
|
"learning_rate": 8.928382342745137e-05, |
|
"loss": 0.5039, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 0.6340468909276249, |
|
"grad_norm": 9.879805564880371, |
|
"learning_rate": 8.916698095563453e-05, |
|
"loss": 1.0113, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 0.6350662589194699, |
|
"grad_norm": 11.630424499511719, |
|
"learning_rate": 8.904958227937406e-05, |
|
"loss": 1.0527, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 0.636085626911315, |
|
"grad_norm": 9.939377784729004, |
|
"learning_rate": 8.893162906583094e-05, |
|
"loss": 1.3893, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 0.6371049949031601, |
|
"grad_norm": 7.852113723754883, |
|
"learning_rate": 8.881312299004117e-05, |
|
"loss": 0.7191, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.6381243628950051, |
|
"grad_norm": 7.134123802185059, |
|
"learning_rate": 8.86940657348918e-05, |
|
"loss": 0.3591, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 0.6391437308868502, |
|
"grad_norm": 5.795046806335449, |
|
"learning_rate": 8.857445899109715e-05, |
|
"loss": 0.9856, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 0.6401630988786952, |
|
"grad_norm": 7.4533610343933105, |
|
"learning_rate": 8.845430445717469e-05, |
|
"loss": 0.7603, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 0.6411824668705403, |
|
"grad_norm": 9.926379203796387, |
|
"learning_rate": 8.8333603839421e-05, |
|
"loss": 1.1553, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 0.6422018348623854, |
|
"grad_norm": 7.032261371612549, |
|
"learning_rate": 8.821235885188754e-05, |
|
"loss": 0.5608, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.6432212028542303, |
|
"grad_norm": 6.283802509307861, |
|
"learning_rate": 8.809057121635624e-05, |
|
"loss": 0.4338, |
|
"step": 631 |
|
}, |
|
{ |
|
"epoch": 0.6442405708460754, |
|
"grad_norm": 2.8640384674072266, |
|
"learning_rate": 8.796824266231511e-05, |
|
"loss": 0.1376, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 0.6452599388379205, |
|
"grad_norm": 7.722833633422852, |
|
"learning_rate": 8.784537492693368e-05, |
|
"loss": 0.6539, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 0.6462793068297655, |
|
"grad_norm": 7.714670658111572, |
|
"learning_rate": 8.772196975503828e-05, |
|
"loss": 0.5017, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 0.6472986748216106, |
|
"grad_norm": 4.0773091316223145, |
|
"learning_rate": 8.759802889908733e-05, |
|
"loss": 0.1888, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.6483180428134556, |
|
"grad_norm": 12.99943733215332, |
|
"learning_rate": 8.747355411914642e-05, |
|
"loss": 1.6077, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 0.6493374108053007, |
|
"grad_norm": 10.86596393585205, |
|
"learning_rate": 8.734854718286327e-05, |
|
"loss": 0.9635, |
|
"step": 637 |
|
}, |
|
{ |
|
"epoch": 0.6503567787971458, |
|
"grad_norm": 9.243484497070312, |
|
"learning_rate": 8.722300986544272e-05, |
|
"loss": 0.9786, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 0.6513761467889908, |
|
"grad_norm": 10.92319393157959, |
|
"learning_rate": 8.709694394962142e-05, |
|
"loss": 0.6728, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 0.6523955147808359, |
|
"grad_norm": 11.628253936767578, |
|
"learning_rate": 8.697035122564266e-05, |
|
"loss": 0.8592, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.6534148827726809, |
|
"grad_norm": 5.602497100830078, |
|
"learning_rate": 8.684323349123075e-05, |
|
"loss": 0.3945, |
|
"step": 641 |
|
}, |
|
{ |
|
"epoch": 0.654434250764526, |
|
"grad_norm": 7.681665420532227, |
|
"learning_rate": 8.671559255156567e-05, |
|
"loss": 0.7486, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 0.6554536187563711, |
|
"grad_norm": 9.017338752746582, |
|
"learning_rate": 8.658743021925733e-05, |
|
"loss": 0.7793, |
|
"step": 643 |
|
}, |
|
{ |
|
"epoch": 0.6564729867482161, |
|
"grad_norm": 5.24987268447876, |
|
"learning_rate": 8.645874831431982e-05, |
|
"loss": 0.4401, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 0.6574923547400612, |
|
"grad_norm": 10.270877838134766, |
|
"learning_rate": 8.632954866414567e-05, |
|
"loss": 0.6189, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.6585117227319062, |
|
"grad_norm": 8.378297805786133, |
|
"learning_rate": 8.619983310347982e-05, |
|
"loss": 0.7339, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 0.6595310907237513, |
|
"grad_norm": 6.045844554901123, |
|
"learning_rate": 8.606960347439355e-05, |
|
"loss": 0.4089, |
|
"step": 647 |
|
}, |
|
{ |
|
"epoch": 0.6605504587155964, |
|
"grad_norm": 10.432483673095703, |
|
"learning_rate": 8.593886162625835e-05, |
|
"loss": 1.1412, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 0.6615698267074414, |
|
"grad_norm": 5.939512729644775, |
|
"learning_rate": 8.580760941571967e-05, |
|
"loss": 0.798, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 0.6625891946992865, |
|
"grad_norm": 12.093332290649414, |
|
"learning_rate": 8.567584870667056e-05, |
|
"loss": 1.0588, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.6636085626911316, |
|
"grad_norm": 8.624043464660645, |
|
"learning_rate": 8.554358137022513e-05, |
|
"loss": 0.9044, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 0.6646279306829765, |
|
"grad_norm": 7.735975742340088, |
|
"learning_rate": 8.54108092846921e-05, |
|
"loss": 0.4464, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 0.6656472986748216, |
|
"grad_norm": 3.8205575942993164, |
|
"learning_rate": 8.527753433554797e-05, |
|
"loss": 0.2756, |
|
"step": 653 |
|
}, |
|
{ |
|
"epoch": 0.6666666666666666, |
|
"grad_norm": 10.537273406982422, |
|
"learning_rate": 8.51437584154104e-05, |
|
"loss": 1.123, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 0.6676860346585117, |
|
"grad_norm": 6.052632808685303, |
|
"learning_rate": 8.500948342401124e-05, |
|
"loss": 0.5377, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.6687054026503568, |
|
"grad_norm": 7.787528991699219, |
|
"learning_rate": 8.48747112681696e-05, |
|
"loss": 0.5164, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 0.6697247706422018, |
|
"grad_norm": 10.115964889526367, |
|
"learning_rate": 8.473944386176469e-05, |
|
"loss": 0.7155, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 0.6707441386340469, |
|
"grad_norm": 6.880122184753418, |
|
"learning_rate": 8.460368312570873e-05, |
|
"loss": 0.4512, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 0.6717635066258919, |
|
"grad_norm": 8.106338500976562, |
|
"learning_rate": 8.446743098791969e-05, |
|
"loss": 0.6199, |
|
"step": 659 |
|
}, |
|
{ |
|
"epoch": 0.672782874617737, |
|
"grad_norm": 11.035154342651367, |
|
"learning_rate": 8.433068938329376e-05, |
|
"loss": 0.6673, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.6738022426095821, |
|
"grad_norm": 4.484703540802002, |
|
"learning_rate": 8.419346025367809e-05, |
|
"loss": 0.5934, |
|
"step": 661 |
|
}, |
|
{ |
|
"epoch": 0.6748216106014271, |
|
"grad_norm": 6.977105140686035, |
|
"learning_rate": 8.4055745547843e-05, |
|
"loss": 0.5034, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 0.6758409785932722, |
|
"grad_norm": 5.447470664978027, |
|
"learning_rate": 8.391754722145449e-05, |
|
"loss": 0.4161, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 0.6768603465851172, |
|
"grad_norm": 13.200489044189453, |
|
"learning_rate": 8.37788672370463e-05, |
|
"loss": 0.9848, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 0.6778797145769623, |
|
"grad_norm": 6.03376579284668, |
|
"learning_rate": 8.36397075639922e-05, |
|
"loss": 0.356, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.6788990825688074, |
|
"grad_norm": 6.075347900390625, |
|
"learning_rate": 8.350007017847788e-05, |
|
"loss": 0.3031, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 0.6799184505606524, |
|
"grad_norm": 5.790109157562256, |
|
"learning_rate": 8.335995706347299e-05, |
|
"loss": 0.254, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 0.6809378185524975, |
|
"grad_norm": 11.979147911071777, |
|
"learning_rate": 8.321937020870296e-05, |
|
"loss": 0.8646, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 0.6819571865443425, |
|
"grad_norm": 9.445723533630371, |
|
"learning_rate": 8.30783116106207e-05, |
|
"loss": 0.7303, |
|
"step": 669 |
|
}, |
|
{ |
|
"epoch": 0.6829765545361876, |
|
"grad_norm": 8.001054763793945, |
|
"learning_rate": 8.293678327237827e-05, |
|
"loss": 0.4105, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.6839959225280327, |
|
"grad_norm": 4.437264919281006, |
|
"learning_rate": 8.279478720379845e-05, |
|
"loss": 0.2874, |
|
"step": 671 |
|
}, |
|
{ |
|
"epoch": 0.6850152905198776, |
|
"grad_norm": 4.547714710235596, |
|
"learning_rate": 8.265232542134622e-05, |
|
"loss": 0.2112, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 0.6860346585117227, |
|
"grad_norm": 7.875749588012695, |
|
"learning_rate": 8.250939994810003e-05, |
|
"loss": 1.0919, |
|
"step": 673 |
|
}, |
|
{ |
|
"epoch": 0.6870540265035678, |
|
"grad_norm": 5.349310874938965, |
|
"learning_rate": 8.236601281372319e-05, |
|
"loss": 0.5927, |
|
"step": 674 |
|
}, |
|
{ |
|
"epoch": 0.6880733944954128, |
|
"grad_norm": 11.490046501159668, |
|
"learning_rate": 8.222216605443496e-05, |
|
"loss": 1.011, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.6890927624872579, |
|
"grad_norm": 7.11298942565918, |
|
"learning_rate": 8.207786171298166e-05, |
|
"loss": 0.5656, |
|
"step": 676 |
|
}, |
|
{ |
|
"epoch": 0.6901121304791029, |
|
"grad_norm": 10.48589038848877, |
|
"learning_rate": 8.193310183860771e-05, |
|
"loss": 0.7199, |
|
"step": 677 |
|
}, |
|
{ |
|
"epoch": 0.691131498470948, |
|
"grad_norm": 9.364179611206055, |
|
"learning_rate": 8.178788848702643e-05, |
|
"loss": 0.7506, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 0.6921508664627931, |
|
"grad_norm": 6.678390026092529, |
|
"learning_rate": 8.164222372039092e-05, |
|
"loss": 0.5386, |
|
"step": 679 |
|
}, |
|
{ |
|
"epoch": 0.6931702344546381, |
|
"grad_norm": 6.151979446411133, |
|
"learning_rate": 8.149610960726479e-05, |
|
"loss": 0.6678, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.6941896024464832, |
|
"grad_norm": 6.415065765380859, |
|
"learning_rate": 8.134954822259271e-05, |
|
"loss": 0.4834, |
|
"step": 681 |
|
}, |
|
{ |
|
"epoch": 0.6952089704383282, |
|
"grad_norm": 4.4640326499938965, |
|
"learning_rate": 8.120254164767101e-05, |
|
"loss": 0.2411, |
|
"step": 682 |
|
}, |
|
{ |
|
"epoch": 0.6962283384301733, |
|
"grad_norm": 6.626987457275391, |
|
"learning_rate": 8.105509197011807e-05, |
|
"loss": 0.5011, |
|
"step": 683 |
|
}, |
|
{ |
|
"epoch": 0.6972477064220184, |
|
"grad_norm": 7.628388404846191, |
|
"learning_rate": 8.090720128384475e-05, |
|
"loss": 0.6573, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 0.6982670744138634, |
|
"grad_norm": 3.4043076038360596, |
|
"learning_rate": 8.075887168902459e-05, |
|
"loss": 0.2798, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.6992864424057085, |
|
"grad_norm": 5.682481288909912, |
|
"learning_rate": 8.061010529206398e-05, |
|
"loss": 0.5887, |
|
"step": 686 |
|
}, |
|
{ |
|
"epoch": 0.6992864424057085, |
|
"eval_Qnli-dev-1024_cosine_accuracy": 0.7291666666666666, |
|
"eval_Qnli-dev-1024_cosine_accuracy_threshold": 0.8446075320243835, |
|
"eval_Qnli-dev-1024_cosine_ap": 0.7501532568375827, |
|
"eval_Qnli-dev-1024_cosine_f1": 0.7207207207207208, |
|
"eval_Qnli-dev-1024_cosine_f1_threshold": 0.756614089012146, |
|
"eval_Qnli-dev-1024_cosine_mcc": 0.4081269865567241, |
|
"eval_Qnli-dev-1024_cosine_precision": 0.6060606060606061, |
|
"eval_Qnli-dev-1024_cosine_recall": 0.8888888888888888, |
|
"eval_Qnli-dev_cosine_accuracy": 0.71875, |
|
"eval_Qnli-dev_cosine_accuracy_threshold": 0.7564685344696045, |
|
"eval_Qnli-dev_cosine_ap": 0.731843650475666, |
|
"eval_Qnli-dev_cosine_f1": 0.7378640776699029, |
|
"eval_Qnli-dev_cosine_f1_threshold": 0.6987220048904419, |
|
"eval_Qnli-dev_cosine_mcc": 0.46153029495329345, |
|
"eval_Qnli-dev_cosine_precision": 0.6551724137931034, |
|
"eval_Qnli-dev_cosine_recall": 0.8444444444444444, |
|
"eval_allNLI--triplets-1024_cosine_accuracy": 0.9270833134651184, |
|
"eval_allNLI-triplets_cosine_accuracy": 0.9583333134651184, |
|
"eval_global_dataset_loss": 0.29955434799194336, |
|
"eval_global_dataset_runtime": 104.3655, |
|
"eval_global_dataset_samples_per_second": 7.694, |
|
"eval_global_dataset_steps_per_second": 0.163, |
|
"eval_sequential_score": 0.9270833134651184, |
|
"eval_sts-test-1024_pearson_cosine": 0.8628190908797548, |
|
"eval_sts-test-1024_spearman_cosine": 0.9062196010289961, |
|
"eval_sts-test_pearson_cosine": 0.9012940791829644, |
|
"eval_sts-test_spearman_cosine": 0.9179642241352577, |
|
"step": 686 |
|
}, |
|
{ |
|
"epoch": 0.7003058103975535, |
|
"grad_norm": 5.198816299438477, |
|
"learning_rate": 8.046090420557231e-05, |
|
"loss": 0.6328, |
|
"step": 687 |
|
}, |
|
{ |
|
"epoch": 0.7013251783893986, |
|
"grad_norm": 7.044593811035156, |
|
"learning_rate": 8.031127054833189e-05, |
|
"loss": 0.4322, |
|
"step": 688 |
|
}, |
|
{ |
|
"epoch": 0.7023445463812437, |
|
"grad_norm": 13.891091346740723, |
|
"learning_rate": 8.016120644526797e-05, |
|
"loss": 1.1473, |
|
"step": 689 |
|
}, |
|
{ |
|
"epoch": 0.7033639143730887, |
|
"grad_norm": 9.329078674316406, |
|
"learning_rate": 8.001071402741842e-05, |
|
"loss": 1.0135, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.7043832823649337, |
|
"grad_norm": 4.1149210929870605, |
|
"learning_rate": 7.985979543190358e-05, |
|
"loss": 0.2655, |
|
"step": 691 |
|
}, |
|
{ |
|
"epoch": 0.7054026503567788, |
|
"grad_norm": 7.722234725952148, |
|
"learning_rate": 7.970845280189586e-05, |
|
"loss": 0.5053, |
|
"step": 692 |
|
}, |
|
{ |
|
"epoch": 0.7064220183486238, |
|
"grad_norm": 6.9180216789245605, |
|
"learning_rate": 7.955668828658937e-05, |
|
"loss": 0.8647, |
|
"step": 693 |
|
}, |
|
{ |
|
"epoch": 0.7074413863404689, |
|
"grad_norm": 5.709589004516602, |
|
"learning_rate": 7.940450404116928e-05, |
|
"loss": 0.4423, |
|
"step": 694 |
|
}, |
|
{ |
|
"epoch": 0.7084607543323139, |
|
"grad_norm": 4.812499523162842, |
|
"learning_rate": 7.925190222678133e-05, |
|
"loss": 0.3673, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.709480122324159, |
|
"grad_norm": 11.944628715515137, |
|
"learning_rate": 7.909888501050109e-05, |
|
"loss": 1.1714, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 0.7104994903160041, |
|
"grad_norm": 7.61957311630249, |
|
"learning_rate": 7.894545456530316e-05, |
|
"loss": 0.8142, |
|
"step": 697 |
|
}, |
|
{ |
|
"epoch": 0.7115188583078491, |
|
"grad_norm": 9.580735206604004, |
|
"learning_rate": 7.879161307003038e-05, |
|
"loss": 0.8027, |
|
"step": 698 |
|
}, |
|
{ |
|
"epoch": 0.7125382262996942, |
|
"grad_norm": 7.831961154937744, |
|
"learning_rate": 7.863736270936284e-05, |
|
"loss": 0.4514, |
|
"step": 699 |
|
}, |
|
{ |
|
"epoch": 0.7135575942915392, |
|
"grad_norm": 9.805893898010254, |
|
"learning_rate": 7.848270567378686e-05, |
|
"loss": 0.8798, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.7145769622833843, |
|
"grad_norm": 8.573545455932617, |
|
"learning_rate": 7.832764415956389e-05, |
|
"loss": 0.7718, |
|
"step": 701 |
|
}, |
|
{ |
|
"epoch": 0.7155963302752294, |
|
"grad_norm": 6.185779571533203, |
|
"learning_rate": 7.817218036869932e-05, |
|
"loss": 0.4094, |
|
"step": 702 |
|
}, |
|
{ |
|
"epoch": 0.7166156982670744, |
|
"grad_norm": 9.415246963500977, |
|
"learning_rate": 7.80163165089112e-05, |
|
"loss": 0.5358, |
|
"step": 703 |
|
}, |
|
{ |
|
"epoch": 0.7176350662589195, |
|
"grad_norm": 5.7925543785095215, |
|
"learning_rate": 7.78600547935989e-05, |
|
"loss": 0.5728, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 0.7186544342507645, |
|
"grad_norm": 8.365612983703613, |
|
"learning_rate": 7.770339744181175e-05, |
|
"loss": 0.4349, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.7196738022426096, |
|
"grad_norm": 11.040353775024414, |
|
"learning_rate": 7.754634667821734e-05, |
|
"loss": 1.0107, |
|
"step": 706 |
|
}, |
|
{ |
|
"epoch": 0.7206931702344547, |
|
"grad_norm": 10.400522232055664, |
|
"learning_rate": 7.73889047330701e-05, |
|
"loss": 1.3393, |
|
"step": 707 |
|
}, |
|
{ |
|
"epoch": 0.7217125382262997, |
|
"grad_norm": 6.314993381500244, |
|
"learning_rate": 7.723107384217958e-05, |
|
"loss": 0.5175, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 0.7227319062181448, |
|
"grad_norm": 7.7337541580200195, |
|
"learning_rate": 7.70728562468787e-05, |
|
"loss": 0.3906, |
|
"step": 709 |
|
}, |
|
{ |
|
"epoch": 0.7237512742099899, |
|
"grad_norm": 8.559732437133789, |
|
"learning_rate": 7.691425419399183e-05, |
|
"loss": 0.726, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.7247706422018348, |
|
"grad_norm": 5.824985504150391, |
|
"learning_rate": 7.675526993580306e-05, |
|
"loss": 0.4299, |
|
"step": 711 |
|
}, |
|
{ |
|
"epoch": 0.72579001019368, |
|
"grad_norm": 9.804418563842773, |
|
"learning_rate": 7.659590573002407e-05, |
|
"loss": 0.7486, |
|
"step": 712 |
|
}, |
|
{ |
|
"epoch": 0.7268093781855249, |
|
"grad_norm": 5.5835957527160645, |
|
"learning_rate": 7.643616383976214e-05, |
|
"loss": 0.3316, |
|
"step": 713 |
|
}, |
|
{ |
|
"epoch": 0.72782874617737, |
|
"grad_norm": 8.719099044799805, |
|
"learning_rate": 7.627604653348796e-05, |
|
"loss": 0.5444, |
|
"step": 714 |
|
}, |
|
{ |
|
"epoch": 0.7288481141692151, |
|
"grad_norm": 7.16873025894165, |
|
"learning_rate": 7.611555608500351e-05, |
|
"loss": 0.4717, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.7298674821610601, |
|
"grad_norm": 8.529095649719238, |
|
"learning_rate": 7.595469477340965e-05, |
|
"loss": 0.5413, |
|
"step": 716 |
|
}, |
|
{ |
|
"epoch": 0.7308868501529052, |
|
"grad_norm": 4.7856245040893555, |
|
"learning_rate": 7.579346488307379e-05, |
|
"loss": 0.2207, |
|
"step": 717 |
|
}, |
|
{ |
|
"epoch": 0.7319062181447502, |
|
"grad_norm": 8.381448745727539, |
|
"learning_rate": 7.563186870359758e-05, |
|
"loss": 0.8042, |
|
"step": 718 |
|
}, |
|
{ |
|
"epoch": 0.7329255861365953, |
|
"grad_norm": 6.099252700805664, |
|
"learning_rate": 7.546990852978415e-05, |
|
"loss": 0.3666, |
|
"step": 719 |
|
}, |
|
{ |
|
"epoch": 0.7339449541284404, |
|
"grad_norm": 6.979067802429199, |
|
"learning_rate": 7.530758666160577e-05, |
|
"loss": 0.4511, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.7349643221202854, |
|
"grad_norm": 8.355476379394531, |
|
"learning_rate": 7.514490540417103e-05, |
|
"loss": 0.782, |
|
"step": 721 |
|
}, |
|
{ |
|
"epoch": 0.7359836901121305, |
|
"grad_norm": 3.4271693229675293, |
|
"learning_rate": 7.498186706769213e-05, |
|
"loss": 0.2947, |
|
"step": 722 |
|
}, |
|
{ |
|
"epoch": 0.7370030581039755, |
|
"grad_norm": 10.753888130187988, |
|
"learning_rate": 7.481847396745215e-05, |
|
"loss": 1.6206, |
|
"step": 723 |
|
}, |
|
{ |
|
"epoch": 0.7380224260958206, |
|
"grad_norm": 10.323583602905273, |
|
"learning_rate": 7.465472842377206e-05, |
|
"loss": 0.6216, |
|
"step": 724 |
|
}, |
|
{ |
|
"epoch": 0.7390417940876657, |
|
"grad_norm": 5.268289566040039, |
|
"learning_rate": 7.449063276197789e-05, |
|
"loss": 0.2819, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.7400611620795107, |
|
"grad_norm": 8.948394775390625, |
|
"learning_rate": 7.432618931236759e-05, |
|
"loss": 0.4797, |
|
"step": 726 |
|
}, |
|
{ |
|
"epoch": 0.7410805300713558, |
|
"grad_norm": 4.78109884262085, |
|
"learning_rate": 7.416140041017802e-05, |
|
"loss": 0.3875, |
|
"step": 727 |
|
}, |
|
{ |
|
"epoch": 0.7420998980632009, |
|
"grad_norm": 7.643434047698975, |
|
"learning_rate": 7.399626839555176e-05, |
|
"loss": 0.6888, |
|
"step": 728 |
|
}, |
|
{ |
|
"epoch": 0.7431192660550459, |
|
"grad_norm": 4.128391742706299, |
|
"learning_rate": 7.383079561350386e-05, |
|
"loss": 0.3023, |
|
"step": 729 |
|
}, |
|
{ |
|
"epoch": 0.744138634046891, |
|
"grad_norm": 8.254578590393066, |
|
"learning_rate": 7.36649844138886e-05, |
|
"loss": 0.6654, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.745158002038736, |
|
"grad_norm": 10.747797966003418, |
|
"learning_rate": 7.3498837151366e-05, |
|
"loss": 0.6517, |
|
"step": 731 |
|
}, |
|
{ |
|
"epoch": 0.746177370030581, |
|
"grad_norm": 6.274332046508789, |
|
"learning_rate": 7.333235618536856e-05, |
|
"loss": 0.4537, |
|
"step": 732 |
|
}, |
|
{ |
|
"epoch": 0.7471967380224261, |
|
"grad_norm": 8.256685256958008, |
|
"learning_rate": 7.316554388006756e-05, |
|
"loss": 0.7224, |
|
"step": 733 |
|
}, |
|
{ |
|
"epoch": 0.7482161060142711, |
|
"grad_norm": 7.657110214233398, |
|
"learning_rate": 7.299840260433965e-05, |
|
"loss": 0.4447, |
|
"step": 734 |
|
}, |
|
{ |
|
"epoch": 0.7492354740061162, |
|
"grad_norm": 6.170997142791748, |
|
"learning_rate": 7.283093473173307e-05, |
|
"loss": 0.4127, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.7502548419979612, |
|
"grad_norm": 5.84876823425293, |
|
"learning_rate": 7.26631426404341e-05, |
|
"loss": 0.3297, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 0.7512742099898063, |
|
"grad_norm": 5.986436367034912, |
|
"learning_rate": 7.249502871323314e-05, |
|
"loss": 0.3664, |
|
"step": 737 |
|
}, |
|
{ |
|
"epoch": 0.7522935779816514, |
|
"grad_norm": 9.613632202148438, |
|
"learning_rate": 7.232659533749092e-05, |
|
"loss": 0.7934, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 0.7533129459734964, |
|
"grad_norm": 5.5741286277771, |
|
"learning_rate": 7.215784490510468e-05, |
|
"loss": 0.4214, |
|
"step": 739 |
|
}, |
|
{ |
|
"epoch": 0.7543323139653415, |
|
"grad_norm": 8.343430519104004, |
|
"learning_rate": 7.198877981247406e-05, |
|
"loss": 0.6174, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.7553516819571865, |
|
"grad_norm": 11.505045890808105, |
|
"learning_rate": 7.18194024604672e-05, |
|
"loss": 0.7011, |
|
"step": 741 |
|
}, |
|
{ |
|
"epoch": 0.7563710499490316, |
|
"grad_norm": 9.192388534545898, |
|
"learning_rate": 7.164971525438657e-05, |
|
"loss": 0.6472, |
|
"step": 742 |
|
}, |
|
{ |
|
"epoch": 0.7573904179408767, |
|
"grad_norm": 10.685009002685547, |
|
"learning_rate": 7.147972060393478e-05, |
|
"loss": 0.9555, |
|
"step": 743 |
|
}, |
|
{ |
|
"epoch": 0.7584097859327217, |
|
"grad_norm": 9.81982421875, |
|
"learning_rate": 7.130942092318051e-05, |
|
"loss": 1.1771, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 0.7594291539245668, |
|
"grad_norm": 7.654698848724365, |
|
"learning_rate": 7.113881863052407e-05, |
|
"loss": 0.6876, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.7604485219164119, |
|
"grad_norm": 10.608144760131836, |
|
"learning_rate": 7.096791614866309e-05, |
|
"loss": 0.6737, |
|
"step": 746 |
|
}, |
|
{ |
|
"epoch": 0.7614678899082569, |
|
"grad_norm": 8.949767112731934, |
|
"learning_rate": 7.079671590455821e-05, |
|
"loss": 0.9648, |
|
"step": 747 |
|
}, |
|
{ |
|
"epoch": 0.762487257900102, |
|
"grad_norm": 5.873875141143799, |
|
"learning_rate": 7.06252203293985e-05, |
|
"loss": 0.3267, |
|
"step": 748 |
|
}, |
|
{ |
|
"epoch": 0.763506625891947, |
|
"grad_norm": 3.814371347427368, |
|
"learning_rate": 7.045343185856701e-05, |
|
"loss": 0.2244, |
|
"step": 749 |
|
}, |
|
{ |
|
"epoch": 0.764525993883792, |
|
"grad_norm": 5.834865570068359, |
|
"learning_rate": 7.028135293160611e-05, |
|
"loss": 0.305, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.7655453618756372, |
|
"grad_norm": 8.765941619873047, |
|
"learning_rate": 7.010898599218296e-05, |
|
"loss": 0.5588, |
|
"step": 751 |
|
}, |
|
{ |
|
"epoch": 0.7665647298674821, |
|
"grad_norm": 8.091228485107422, |
|
"learning_rate": 6.99363334880547e-05, |
|
"loss": 1.0974, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 0.7675840978593272, |
|
"grad_norm": 7.041286468505859, |
|
"learning_rate": 6.976339787103373e-05, |
|
"loss": 0.603, |
|
"step": 753 |
|
}, |
|
{ |
|
"epoch": 0.7686034658511722, |
|
"grad_norm": 6.676450729370117, |
|
"learning_rate": 6.959018159695293e-05, |
|
"loss": 0.6972, |
|
"step": 754 |
|
}, |
|
{ |
|
"epoch": 0.7696228338430173, |
|
"grad_norm": 9.935379981994629, |
|
"learning_rate": 6.94166871256307e-05, |
|
"loss": 0.958, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.7706422018348624, |
|
"grad_norm": 6.536661624908447, |
|
"learning_rate": 6.92429169208361e-05, |
|
"loss": 0.2937, |
|
"step": 756 |
|
}, |
|
{ |
|
"epoch": 0.7716615698267074, |
|
"grad_norm": 5.736427307128906, |
|
"learning_rate": 6.906887345025385e-05, |
|
"loss": 0.3384, |
|
"step": 757 |
|
}, |
|
{ |
|
"epoch": 0.7726809378185525, |
|
"grad_norm": 5.628017425537109, |
|
"learning_rate": 6.88945591854493e-05, |
|
"loss": 0.3321, |
|
"step": 758 |
|
}, |
|
{ |
|
"epoch": 0.7737003058103975, |
|
"grad_norm": 9.1480712890625, |
|
"learning_rate": 6.87199766018332e-05, |
|
"loss": 0.8029, |
|
"step": 759 |
|
}, |
|
{ |
|
"epoch": 0.7747196738022426, |
|
"grad_norm": 7.8731770515441895, |
|
"learning_rate": 6.85451281786268e-05, |
|
"loss": 0.7043, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.7757390417940877, |
|
"grad_norm": 13.733153343200684, |
|
"learning_rate": 6.837001639882641e-05, |
|
"loss": 1.6068, |
|
"step": 761 |
|
}, |
|
{ |
|
"epoch": 0.7767584097859327, |
|
"grad_norm": 9.02813720703125, |
|
"learning_rate": 6.819464374916823e-05, |
|
"loss": 1.1273, |
|
"step": 762 |
|
}, |
|
{ |
|
"epoch": 0.7777777777777778, |
|
"grad_norm": 8.211151123046875, |
|
"learning_rate": 6.801901272009307e-05, |
|
"loss": 0.5429, |
|
"step": 763 |
|
}, |
|
{ |
|
"epoch": 0.7787971457696228, |
|
"grad_norm": 5.243499755859375, |
|
"learning_rate": 6.784312580571091e-05, |
|
"loss": 0.2976, |
|
"step": 764 |
|
}, |
|
{ |
|
"epoch": 0.7798165137614679, |
|
"grad_norm": 11.219100952148438, |
|
"learning_rate": 6.766698550376556e-05, |
|
"loss": 0.9183, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.780835881753313, |
|
"grad_norm": 7.10944938659668, |
|
"learning_rate": 6.749059431559913e-05, |
|
"loss": 0.4734, |
|
"step": 766 |
|
}, |
|
{ |
|
"epoch": 0.781855249745158, |
|
"grad_norm": 7.810965061187744, |
|
"learning_rate": 6.731395474611649e-05, |
|
"loss": 0.5437, |
|
"step": 767 |
|
}, |
|
{ |
|
"epoch": 0.7828746177370031, |
|
"grad_norm": 6.063333034515381, |
|
"learning_rate": 6.71370693037498e-05, |
|
"loss": 0.3382, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 0.7838939857288482, |
|
"grad_norm": 5.784426689147949, |
|
"learning_rate": 6.695994050042277e-05, |
|
"loss": 0.3925, |
|
"step": 769 |
|
}, |
|
{ |
|
"epoch": 0.7849133537206932, |
|
"grad_norm": 7.640711784362793, |
|
"learning_rate": 6.678257085151509e-05, |
|
"loss": 0.4345, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.7859327217125383, |
|
"grad_norm": 9.467418670654297, |
|
"learning_rate": 6.660496287582667e-05, |
|
"loss": 0.9237, |
|
"step": 771 |
|
}, |
|
{ |
|
"epoch": 0.7869520897043832, |
|
"grad_norm": 4.449363708496094, |
|
"learning_rate": 6.642711909554174e-05, |
|
"loss": 0.3875, |
|
"step": 772 |
|
}, |
|
{ |
|
"epoch": 0.7879714576962283, |
|
"grad_norm": 7.483307838439941, |
|
"learning_rate": 6.624904203619333e-05, |
|
"loss": 0.533, |
|
"step": 773 |
|
}, |
|
{ |
|
"epoch": 0.7889908256880734, |
|
"grad_norm": 4.827091693878174, |
|
"learning_rate": 6.607073422662711e-05, |
|
"loss": 0.4211, |
|
"step": 774 |
|
}, |
|
{ |
|
"epoch": 0.7900101936799184, |
|
"grad_norm": 6.135465621948242, |
|
"learning_rate": 6.589219819896565e-05, |
|
"loss": 0.5421, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.7910295616717635, |
|
"grad_norm": 9.622929573059082, |
|
"learning_rate": 6.571343648857242e-05, |
|
"loss": 0.8904, |
|
"step": 776 |
|
}, |
|
{ |
|
"epoch": 0.7920489296636085, |
|
"grad_norm": 5.664134502410889, |
|
"learning_rate": 6.553445163401571e-05, |
|
"loss": 0.4604, |
|
"step": 777 |
|
}, |
|
{ |
|
"epoch": 0.7930682976554536, |
|
"grad_norm": 9.634468078613281, |
|
"learning_rate": 6.535524617703273e-05, |
|
"loss": 0.7431, |
|
"step": 778 |
|
}, |
|
{ |
|
"epoch": 0.7940876656472987, |
|
"grad_norm": 10.855483055114746, |
|
"learning_rate": 6.517582266249336e-05, |
|
"loss": 1.0159, |
|
"step": 779 |
|
}, |
|
{ |
|
"epoch": 0.7951070336391437, |
|
"grad_norm": 9.945262908935547, |
|
"learning_rate": 6.499618363836417e-05, |
|
"loss": 0.6554, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.7961264016309888, |
|
"grad_norm": 7.224388599395752, |
|
"learning_rate": 6.481633165567207e-05, |
|
"loss": 0.8539, |
|
"step": 781 |
|
}, |
|
{ |
|
"epoch": 0.7971457696228338, |
|
"grad_norm": 8.917383193969727, |
|
"learning_rate": 6.463626926846817e-05, |
|
"loss": 0.4543, |
|
"step": 782 |
|
}, |
|
{ |
|
"epoch": 0.7981651376146789, |
|
"grad_norm": 4.411260604858398, |
|
"learning_rate": 6.445599903379154e-05, |
|
"loss": 0.2281, |
|
"step": 783 |
|
}, |
|
{ |
|
"epoch": 0.799184505606524, |
|
"grad_norm": 8.85741138458252, |
|
"learning_rate": 6.427552351163286e-05, |
|
"loss": 1.0334, |
|
"step": 784 |
|
}, |
|
{ |
|
"epoch": 0.799184505606524, |
|
"eval_Qnli-dev-1024_cosine_accuracy": 0.7395833333333334, |
|
"eval_Qnli-dev-1024_cosine_accuracy_threshold": 0.8376107215881348, |
|
"eval_Qnli-dev-1024_cosine_ap": 0.7815698422458957, |
|
"eval_Qnli-dev-1024_cosine_f1": 0.7222222222222222, |
|
"eval_Qnli-dev-1024_cosine_f1_threshold": 0.7544304132461548, |
|
"eval_Qnli-dev-1024_cosine_mcc": 0.41614558708189836, |
|
"eval_Qnli-dev-1024_cosine_precision": 0.6190476190476191, |
|
"eval_Qnli-dev-1024_cosine_recall": 0.8666666666666667, |
|
"eval_Qnli-dev_cosine_accuracy": 0.7291666666666666, |
|
"eval_Qnli-dev_cosine_accuracy_threshold": 0.7388297319412231, |
|
"eval_Qnli-dev_cosine_ap": 0.7636341718424307, |
|
"eval_Qnli-dev_cosine_f1": 0.7450980392156862, |
|
"eval_Qnli-dev_cosine_f1_threshold": 0.695953369140625, |
|
"eval_Qnli-dev_cosine_mcc": 0.4794765594627558, |
|
"eval_Qnli-dev_cosine_precision": 0.6666666666666666, |
|
"eval_Qnli-dev_cosine_recall": 0.8444444444444444, |
|
"eval_allNLI--triplets-1024_cosine_accuracy": 0.9583333134651184, |
|
"eval_allNLI-triplets_cosine_accuracy": 0.9583333134651184, |
|
"eval_global_dataset_loss": 0.3045359253883362, |
|
"eval_global_dataset_runtime": 103.772, |
|
"eval_global_dataset_samples_per_second": 7.738, |
|
"eval_global_dataset_steps_per_second": 0.164, |
|
"eval_sequential_score": 0.9583333134651184, |
|
"eval_sts-test-1024_pearson_cosine": 0.8706480103495355, |
|
"eval_sts-test-1024_spearman_cosine": 0.9094148980677476, |
|
"eval_sts-test_pearson_cosine": 0.9036838203711135, |
|
"eval_sts-test_spearman_cosine": 0.9196077696084266, |
|
"step": 784 |
|
}, |
|
{ |
|
"epoch": 0.800203873598369, |
|
"grad_norm": 10.137900352478027, |
|
"learning_rate": 6.409484526489805e-05, |
|
"loss": 0.9697, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.8012232415902141, |
|
"grad_norm": 8.976780891418457, |
|
"learning_rate": 6.391396685937186e-05, |
|
"loss": 0.7048, |
|
"step": 786 |
|
}, |
|
{ |
|
"epoch": 0.8022426095820592, |
|
"grad_norm": 8.672534942626953, |
|
"learning_rate": 6.373289086368151e-05, |
|
"loss": 0.5263, |
|
"step": 787 |
|
}, |
|
{ |
|
"epoch": 0.8032619775739042, |
|
"grad_norm": 9.115574836730957, |
|
"learning_rate": 6.355161984926019e-05, |
|
"loss": 0.5056, |
|
"step": 788 |
|
}, |
|
{ |
|
"epoch": 0.8042813455657493, |
|
"grad_norm": 5.47214412689209, |
|
"learning_rate": 6.337015639031044e-05, |
|
"loss": 0.3826, |
|
"step": 789 |
|
}, |
|
{ |
|
"epoch": 0.8053007135575942, |
|
"grad_norm": 4.726554870605469, |
|
"learning_rate": 6.318850306376777e-05, |
|
"loss": 0.3029, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.8063200815494393, |
|
"grad_norm": 9.025796890258789, |
|
"learning_rate": 6.300666244926387e-05, |
|
"loss": 0.7712, |
|
"step": 791 |
|
}, |
|
{ |
|
"epoch": 0.8073394495412844, |
|
"grad_norm": 8.51115894317627, |
|
"learning_rate": 6.282463712909018e-05, |
|
"loss": 0.5587, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 0.8083588175331294, |
|
"grad_norm": 4.170646667480469, |
|
"learning_rate": 6.264242968816106e-05, |
|
"loss": 0.2386, |
|
"step": 793 |
|
}, |
|
{ |
|
"epoch": 0.8093781855249745, |
|
"grad_norm": 7.225284576416016, |
|
"learning_rate": 6.246004271397713e-05, |
|
"loss": 0.5662, |
|
"step": 794 |
|
}, |
|
{ |
|
"epoch": 0.8103975535168195, |
|
"grad_norm": 8.109657287597656, |
|
"learning_rate": 6.227747879658859e-05, |
|
"loss": 0.5322, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.8114169215086646, |
|
"grad_norm": 8.729584693908691, |
|
"learning_rate": 6.20947405285583e-05, |
|
"loss": 0.5122, |
|
"step": 796 |
|
}, |
|
{ |
|
"epoch": 0.8124362895005097, |
|
"grad_norm": 6.562040328979492, |
|
"learning_rate": 6.191183050492515e-05, |
|
"loss": 0.5094, |
|
"step": 797 |
|
}, |
|
{ |
|
"epoch": 0.8134556574923547, |
|
"grad_norm": 8.552765846252441, |
|
"learning_rate": 6.172875132316703e-05, |
|
"loss": 0.8412, |
|
"step": 798 |
|
}, |
|
{ |
|
"epoch": 0.8144750254841998, |
|
"grad_norm": 8.517980575561523, |
|
"learning_rate": 6.154550558316405e-05, |
|
"loss": 0.3771, |
|
"step": 799 |
|
}, |
|
{ |
|
"epoch": 0.8154943934760448, |
|
"grad_norm": 9.862586975097656, |
|
"learning_rate": 6.136209588716155e-05, |
|
"loss": 0.626, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.8165137614678899, |
|
"grad_norm": 11.597122192382812, |
|
"learning_rate": 6.117852483973325e-05, |
|
"loss": 0.8902, |
|
"step": 801 |
|
}, |
|
{ |
|
"epoch": 0.817533129459735, |
|
"grad_norm": 4.268974781036377, |
|
"learning_rate": 6.0994795047744144e-05, |
|
"loss": 0.2301, |
|
"step": 802 |
|
}, |
|
{ |
|
"epoch": 0.81855249745158, |
|
"grad_norm": 2.586038112640381, |
|
"learning_rate": 6.081090912031358e-05, |
|
"loss": 0.16, |
|
"step": 803 |
|
}, |
|
{ |
|
"epoch": 0.8195718654434251, |
|
"grad_norm": 6.814731121063232, |
|
"learning_rate": 6.0626869668778085e-05, |
|
"loss": 0.4375, |
|
"step": 804 |
|
}, |
|
{ |
|
"epoch": 0.8205912334352702, |
|
"grad_norm": 9.699979782104492, |
|
"learning_rate": 6.044267930665446e-05, |
|
"loss": 0.9554, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 0.8216106014271152, |
|
"grad_norm": 7.751320838928223, |
|
"learning_rate": 6.025834064960247e-05, |
|
"loss": 0.4906, |
|
"step": 806 |
|
}, |
|
{ |
|
"epoch": 0.8226299694189603, |
|
"grad_norm": 8.852093696594238, |
|
"learning_rate": 6.007385631538787e-05, |
|
"loss": 0.478, |
|
"step": 807 |
|
}, |
|
{ |
|
"epoch": 0.8236493374108053, |
|
"grad_norm": 5.510447025299072, |
|
"learning_rate": 5.988922892384513e-05, |
|
"loss": 0.6057, |
|
"step": 808 |
|
}, |
|
{ |
|
"epoch": 0.8246687054026504, |
|
"grad_norm": 6.745148658752441, |
|
"learning_rate": 5.9704461096840204e-05, |
|
"loss": 0.5003, |
|
"step": 809 |
|
}, |
|
{ |
|
"epoch": 0.8256880733944955, |
|
"grad_norm": 11.509452819824219, |
|
"learning_rate": 5.9519555458233436e-05, |
|
"loss": 1.0844, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.8267074413863404, |
|
"grad_norm": 9.71648120880127, |
|
"learning_rate": 5.933451463384213e-05, |
|
"loss": 1.0267, |
|
"step": 811 |
|
}, |
|
{ |
|
"epoch": 0.8277268093781855, |
|
"grad_norm": 9.810832023620605, |
|
"learning_rate": 5.91493412514034e-05, |
|
"loss": 0.5415, |
|
"step": 812 |
|
}, |
|
{ |
|
"epoch": 0.8287461773700305, |
|
"grad_norm": 5.600392818450928, |
|
"learning_rate": 5.896403794053679e-05, |
|
"loss": 0.3295, |
|
"step": 813 |
|
}, |
|
{ |
|
"epoch": 0.8297655453618756, |
|
"grad_norm": 7.511580944061279, |
|
"learning_rate": 5.877860733270692e-05, |
|
"loss": 0.5511, |
|
"step": 814 |
|
}, |
|
{ |
|
"epoch": 0.8307849133537207, |
|
"grad_norm": 5.374726295471191, |
|
"learning_rate": 5.8593052061186125e-05, |
|
"loss": 0.3234, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 0.8318042813455657, |
|
"grad_norm": 4.7778639793396, |
|
"learning_rate": 5.8407374761017105e-05, |
|
"loss": 0.2917, |
|
"step": 816 |
|
}, |
|
{ |
|
"epoch": 0.8328236493374108, |
|
"grad_norm": 4.155742645263672, |
|
"learning_rate": 5.822157806897548e-05, |
|
"loss": 0.3865, |
|
"step": 817 |
|
}, |
|
{ |
|
"epoch": 0.8338430173292558, |
|
"grad_norm": 5.087594032287598, |
|
"learning_rate": 5.803566462353225e-05, |
|
"loss": 0.2401, |
|
"step": 818 |
|
}, |
|
{ |
|
"epoch": 0.8348623853211009, |
|
"grad_norm": 3.707869529724121, |
|
"learning_rate": 5.7849637064816496e-05, |
|
"loss": 0.1582, |
|
"step": 819 |
|
}, |
|
{ |
|
"epoch": 0.835881753312946, |
|
"grad_norm": 7.63162899017334, |
|
"learning_rate": 5.76634980345778e-05, |
|
"loss": 0.5475, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.836901121304791, |
|
"grad_norm": 5.092942237854004, |
|
"learning_rate": 5.747725017614869e-05, |
|
"loss": 0.3291, |
|
"step": 821 |
|
}, |
|
{ |
|
"epoch": 0.8379204892966361, |
|
"grad_norm": 6.86021089553833, |
|
"learning_rate": 5.72908961344072e-05, |
|
"loss": 0.6867, |
|
"step": 822 |
|
}, |
|
{ |
|
"epoch": 0.8389398572884812, |
|
"grad_norm": 9.336700439453125, |
|
"learning_rate": 5.710443855573919e-05, |
|
"loss": 0.9519, |
|
"step": 823 |
|
}, |
|
{ |
|
"epoch": 0.8399592252803262, |
|
"grad_norm": 6.382976055145264, |
|
"learning_rate": 5.6917880088000894e-05, |
|
"loss": 0.4898, |
|
"step": 824 |
|
}, |
|
{ |
|
"epoch": 0.8409785932721713, |
|
"grad_norm": 8.171992301940918, |
|
"learning_rate": 5.6731223380481257e-05, |
|
"loss": 0.3361, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.8419979612640163, |
|
"grad_norm": 11.304964065551758, |
|
"learning_rate": 5.6544471083864245e-05, |
|
"loss": 1.0131, |
|
"step": 826 |
|
}, |
|
{ |
|
"epoch": 0.8430173292558614, |
|
"grad_norm": 7.883802890777588, |
|
"learning_rate": 5.635762585019136e-05, |
|
"loss": 0.4988, |
|
"step": 827 |
|
}, |
|
{ |
|
"epoch": 0.8440366972477065, |
|
"grad_norm": 5.304625988006592, |
|
"learning_rate": 5.61706903328238e-05, |
|
"loss": 0.2737, |
|
"step": 828 |
|
}, |
|
{ |
|
"epoch": 0.8450560652395515, |
|
"grad_norm": 8.170361518859863, |
|
"learning_rate": 5.598366718640494e-05, |
|
"loss": 0.5214, |
|
"step": 829 |
|
}, |
|
{ |
|
"epoch": 0.8460754332313966, |
|
"grad_norm": 7.193360805511475, |
|
"learning_rate": 5.579655906682255e-05, |
|
"loss": 0.5261, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.8470948012232415, |
|
"grad_norm": 5.908787250518799, |
|
"learning_rate": 5.5609368631171035e-05, |
|
"loss": 0.4337, |
|
"step": 831 |
|
}, |
|
{ |
|
"epoch": 0.8481141692150866, |
|
"grad_norm": 11.470138549804688, |
|
"learning_rate": 5.5422098537713815e-05, |
|
"loss": 1.0523, |
|
"step": 832 |
|
}, |
|
{ |
|
"epoch": 0.8491335372069317, |
|
"grad_norm": 5.7633514404296875, |
|
"learning_rate": 5.52347514458455e-05, |
|
"loss": 0.59, |
|
"step": 833 |
|
}, |
|
{ |
|
"epoch": 0.8501529051987767, |
|
"grad_norm": 9.171930313110352, |
|
"learning_rate": 5.5047330016054154e-05, |
|
"loss": 0.9984, |
|
"step": 834 |
|
}, |
|
{ |
|
"epoch": 0.8511722731906218, |
|
"grad_norm": 7.584822177886963, |
|
"learning_rate": 5.48598369098835e-05, |
|
"loss": 1.0533, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 0.8521916411824668, |
|
"grad_norm": 5.429177761077881, |
|
"learning_rate": 5.4672274789895104e-05, |
|
"loss": 0.266, |
|
"step": 836 |
|
}, |
|
{ |
|
"epoch": 0.8532110091743119, |
|
"grad_norm": 7.292309284210205, |
|
"learning_rate": 5.4484646319630636e-05, |
|
"loss": 0.3497, |
|
"step": 837 |
|
}, |
|
{ |
|
"epoch": 0.854230377166157, |
|
"grad_norm": 7.126836776733398, |
|
"learning_rate": 5.429695416357392e-05, |
|
"loss": 0.5161, |
|
"step": 838 |
|
}, |
|
{ |
|
"epoch": 0.855249745158002, |
|
"grad_norm": 6.357126235961914, |
|
"learning_rate": 5.410920098711323e-05, |
|
"loss": 0.4256, |
|
"step": 839 |
|
}, |
|
{ |
|
"epoch": 0.8562691131498471, |
|
"grad_norm": 6.682480335235596, |
|
"learning_rate": 5.392138945650339e-05, |
|
"loss": 0.3334, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.8572884811416922, |
|
"grad_norm": 6.9180521965026855, |
|
"learning_rate": 5.373352223882787e-05, |
|
"loss": 0.5704, |
|
"step": 841 |
|
}, |
|
{ |
|
"epoch": 0.8583078491335372, |
|
"grad_norm": 6.871384620666504, |
|
"learning_rate": 5.354560200196094e-05, |
|
"loss": 0.3803, |
|
"step": 842 |
|
}, |
|
{ |
|
"epoch": 0.8593272171253823, |
|
"grad_norm": 9.186737060546875, |
|
"learning_rate": 5.335763141452982e-05, |
|
"loss": 0.7648, |
|
"step": 843 |
|
}, |
|
{ |
|
"epoch": 0.8603465851172273, |
|
"grad_norm": 8.700101852416992, |
|
"learning_rate": 5.3169613145876714e-05, |
|
"loss": 0.7548, |
|
"step": 844 |
|
}, |
|
{ |
|
"epoch": 0.8613659531090724, |
|
"grad_norm": 7.032200336456299, |
|
"learning_rate": 5.2981549866020975e-05, |
|
"loss": 0.7275, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 0.8623853211009175, |
|
"grad_norm": 13.48193359375, |
|
"learning_rate": 5.2793444245621146e-05, |
|
"loss": 1.1788, |
|
"step": 846 |
|
}, |
|
{ |
|
"epoch": 0.8634046890927625, |
|
"grad_norm": 9.682479858398438, |
|
"learning_rate": 5.260529895593702e-05, |
|
"loss": 0.7809, |
|
"step": 847 |
|
}, |
|
{ |
|
"epoch": 0.8644240570846076, |
|
"grad_norm": 8.730304718017578, |
|
"learning_rate": 5.241711666879172e-05, |
|
"loss": 0.6487, |
|
"step": 848 |
|
}, |
|
{ |
|
"epoch": 0.8654434250764526, |
|
"grad_norm": 6.570590972900391, |
|
"learning_rate": 5.2228900056533836e-05, |
|
"loss": 0.561, |
|
"step": 849 |
|
}, |
|
{ |
|
"epoch": 0.8664627930682977, |
|
"grad_norm": 8.695535659790039, |
|
"learning_rate": 5.204065179199931e-05, |
|
"loss": 0.5906, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.8674821610601428, |
|
"grad_norm": 5.353935241699219, |
|
"learning_rate": 5.1852374548473614e-05, |
|
"loss": 0.5192, |
|
"step": 851 |
|
}, |
|
{ |
|
"epoch": 0.8685015290519877, |
|
"grad_norm": 10.60522174835205, |
|
"learning_rate": 5.1664070999653766e-05, |
|
"loss": 0.8094, |
|
"step": 852 |
|
}, |
|
{ |
|
"epoch": 0.8695208970438328, |
|
"grad_norm": 3.7188539505004883, |
|
"learning_rate": 5.147574381961032e-05, |
|
"loss": 0.2399, |
|
"step": 853 |
|
}, |
|
{ |
|
"epoch": 0.8705402650356778, |
|
"grad_norm": 5.648993492126465, |
|
"learning_rate": 5.128739568274944e-05, |
|
"loss": 0.4103, |
|
"step": 854 |
|
}, |
|
{ |
|
"epoch": 0.8715596330275229, |
|
"grad_norm": 6.711026668548584, |
|
"learning_rate": 5.109902926377482e-05, |
|
"loss": 0.4969, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 0.872579001019368, |
|
"grad_norm": 5.686347961425781, |
|
"learning_rate": 5.091064723764987e-05, |
|
"loss": 0.37, |
|
"step": 856 |
|
}, |
|
{ |
|
"epoch": 0.873598369011213, |
|
"grad_norm": 4.857931613922119, |
|
"learning_rate": 5.072225227955959e-05, |
|
"loss": 0.4109, |
|
"step": 857 |
|
}, |
|
{ |
|
"epoch": 0.8746177370030581, |
|
"grad_norm": 8.75938606262207, |
|
"learning_rate": 5.053384706487261e-05, |
|
"loss": 0.525, |
|
"step": 858 |
|
}, |
|
{ |
|
"epoch": 0.8756371049949032, |
|
"grad_norm": 5.874378204345703, |
|
"learning_rate": 5.034543426910324e-05, |
|
"loss": 0.5958, |
|
"step": 859 |
|
}, |
|
{ |
|
"epoch": 0.8766564729867482, |
|
"grad_norm": 5.085257530212402, |
|
"learning_rate": 5.0157016567873424e-05, |
|
"loss": 0.4708, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.8776758409785933, |
|
"grad_norm": 7.9917707443237305, |
|
"learning_rate": 4.996859663687479e-05, |
|
"loss": 0.6881, |
|
"step": 861 |
|
}, |
|
{ |
|
"epoch": 0.8786952089704383, |
|
"grad_norm": 8.1506929397583, |
|
"learning_rate": 4.9780177151830634e-05, |
|
"loss": 0.5545, |
|
"step": 862 |
|
}, |
|
{ |
|
"epoch": 0.8797145769622834, |
|
"grad_norm": 9.375650405883789, |
|
"learning_rate": 4.959176078845789e-05, |
|
"loss": 0.645, |
|
"step": 863 |
|
}, |
|
{ |
|
"epoch": 0.8807339449541285, |
|
"grad_norm": 4.8143310546875, |
|
"learning_rate": 4.9403350222429184e-05, |
|
"loss": 0.4112, |
|
"step": 864 |
|
}, |
|
{ |
|
"epoch": 0.8817533129459735, |
|
"grad_norm": 7.862481594085693, |
|
"learning_rate": 4.92149481293348e-05, |
|
"loss": 0.4178, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 0.8827726809378186, |
|
"grad_norm": 5.252464771270752, |
|
"learning_rate": 4.902655718464473e-05, |
|
"loss": 0.2857, |
|
"step": 866 |
|
}, |
|
{ |
|
"epoch": 0.8837920489296636, |
|
"grad_norm": 6.06905460357666, |
|
"learning_rate": 4.883818006367062e-05, |
|
"loss": 0.3374, |
|
"step": 867 |
|
}, |
|
{ |
|
"epoch": 0.8848114169215087, |
|
"grad_norm": 6.810131072998047, |
|
"learning_rate": 4.86498194415278e-05, |
|
"loss": 0.5303, |
|
"step": 868 |
|
}, |
|
{ |
|
"epoch": 0.8858307849133538, |
|
"grad_norm": 7.676322937011719, |
|
"learning_rate": 4.846147799309734e-05, |
|
"loss": 0.7438, |
|
"step": 869 |
|
}, |
|
{ |
|
"epoch": 0.8868501529051988, |
|
"grad_norm": 11.570023536682129, |
|
"learning_rate": 4.8273158392987986e-05, |
|
"loss": 1.0872, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.8878695208970439, |
|
"grad_norm": 6.312341213226318, |
|
"learning_rate": 4.8084863315498234e-05, |
|
"loss": 0.4497, |
|
"step": 871 |
|
}, |
|
{ |
|
"epoch": 0.8888888888888888, |
|
"grad_norm": 7.389033794403076, |
|
"learning_rate": 4.7896595434578356e-05, |
|
"loss": 0.4171, |
|
"step": 872 |
|
}, |
|
{ |
|
"epoch": 0.8899082568807339, |
|
"grad_norm": 8.600625038146973, |
|
"learning_rate": 4.770835742379239e-05, |
|
"loss": 0.4417, |
|
"step": 873 |
|
}, |
|
{ |
|
"epoch": 0.890927624872579, |
|
"grad_norm": 7.350024223327637, |
|
"learning_rate": 4.7520151956280227e-05, |
|
"loss": 0.7023, |
|
"step": 874 |
|
}, |
|
{ |
|
"epoch": 0.891946992864424, |
|
"grad_norm": 12.617684364318848, |
|
"learning_rate": 4.733198170471953e-05, |
|
"loss": 1.0547, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.8929663608562691, |
|
"grad_norm": 5.219171524047852, |
|
"learning_rate": 4.714384934128796e-05, |
|
"loss": 0.3526, |
|
"step": 876 |
|
}, |
|
{ |
|
"epoch": 0.8939857288481141, |
|
"grad_norm": 10.923335075378418, |
|
"learning_rate": 4.6955757537625104e-05, |
|
"loss": 0.7315, |
|
"step": 877 |
|
}, |
|
{ |
|
"epoch": 0.8950050968399592, |
|
"grad_norm": 4.7785325050354, |
|
"learning_rate": 4.6767708964794526e-05, |
|
"loss": 0.4082, |
|
"step": 878 |
|
}, |
|
{ |
|
"epoch": 0.8960244648318043, |
|
"grad_norm": 7.037627696990967, |
|
"learning_rate": 4.6579706293245944e-05, |
|
"loss": 0.8155, |
|
"step": 879 |
|
}, |
|
{ |
|
"epoch": 0.8970438328236493, |
|
"grad_norm": 7.149205207824707, |
|
"learning_rate": 4.6391752192777164e-05, |
|
"loss": 0.5083, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.8980632008154944, |
|
"grad_norm": 5.331564426422119, |
|
"learning_rate": 4.620384933249631e-05, |
|
"loss": 0.655, |
|
"step": 881 |
|
}, |
|
{ |
|
"epoch": 0.8990825688073395, |
|
"grad_norm": 10.019486427307129, |
|
"learning_rate": 4.6016000380783805e-05, |
|
"loss": 0.7207, |
|
"step": 882 |
|
}, |
|
{ |
|
"epoch": 0.8990825688073395, |
|
"eval_Qnli-dev-1024_cosine_accuracy": 0.7708333333333334, |
|
"eval_Qnli-dev-1024_cosine_accuracy_threshold": 0.8360349535942078, |
|
"eval_Qnli-dev-1024_cosine_ap": 0.8011558872452826, |
|
"eval_Qnli-dev-1024_cosine_f1": 0.7250000000000001, |
|
"eval_Qnli-dev-1024_cosine_f1_threshold": 0.8360349535942078, |
|
"eval_Qnli-dev-1024_cosine_mcc": 0.5461802806126049, |
|
"eval_Qnli-dev-1024_cosine_precision": 0.8285714285714286, |
|
"eval_Qnli-dev-1024_cosine_recall": 0.6444444444444445, |
|
"eval_Qnli-dev_cosine_accuracy": 0.7291666666666666, |
|
"eval_Qnli-dev_cosine_accuracy_threshold": 0.7521146535873413, |
|
"eval_Qnli-dev_cosine_ap": 0.7712094779135136, |
|
"eval_Qnli-dev_cosine_f1": 0.7500000000000001, |
|
"eval_Qnli-dev_cosine_f1_threshold": 0.6768573522567749, |
|
"eval_Qnli-dev_cosine_mcc": 0.48653004754089046, |
|
"eval_Qnli-dev_cosine_precision": 0.6610169491525424, |
|
"eval_Qnli-dev_cosine_recall": 0.8666666666666667, |
|
"eval_allNLI--triplets-1024_cosine_accuracy": 0.9479166865348816, |
|
"eval_allNLI-triplets_cosine_accuracy": 0.9583333134651184, |
|
"eval_global_dataset_loss": 0.26387155055999756, |
|
"eval_global_dataset_runtime": 103.9177, |
|
"eval_global_dataset_samples_per_second": 7.727, |
|
"eval_global_dataset_steps_per_second": 0.164, |
|
"eval_sequential_score": 0.9479166865348816, |
|
"eval_sts-test-1024_pearson_cosine": 0.8810824372715894, |
|
"eval_sts-test-1024_spearman_cosine": 0.9117642789427417, |
|
"eval_sts-test_pearson_cosine": 0.9044525796924666, |
|
"eval_sts-test_spearman_cosine": 0.9182572042166309, |
|
"step": 882 |
|
}, |
|
{ |
|
"epoch": 0.9001019367991845, |
|
"grad_norm": 7.724752426147461, |
|
"learning_rate": 4.582820800525455e-05, |
|
"loss": 0.4898, |
|
"step": 883 |
|
}, |
|
{ |
|
"epoch": 0.9011213047910296, |
|
"grad_norm": 9.442131042480469, |
|
"learning_rate": 4.564047487272001e-05, |
|
"loss": 0.5506, |
|
"step": 884 |
|
}, |
|
{ |
|
"epoch": 0.9021406727828746, |
|
"grad_norm": 8.832263946533203, |
|
"learning_rate": 4.5452803649150324e-05, |
|
"loss": 0.6206, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 0.9031600407747197, |
|
"grad_norm": 12.987079620361328, |
|
"learning_rate": 4.5265196999636535e-05, |
|
"loss": 1.9669, |
|
"step": 886 |
|
}, |
|
{ |
|
"epoch": 0.9041794087665648, |
|
"grad_norm": 9.050309181213379, |
|
"learning_rate": 4.5077657588352595e-05, |
|
"loss": 0.8493, |
|
"step": 887 |
|
}, |
|
{ |
|
"epoch": 0.9051987767584098, |
|
"grad_norm": 9.642857551574707, |
|
"learning_rate": 4.489018807851769e-05, |
|
"loss": 0.9698, |
|
"step": 888 |
|
}, |
|
{ |
|
"epoch": 0.9062181447502549, |
|
"grad_norm": 7.444589614868164, |
|
"learning_rate": 4.4702791132358314e-05, |
|
"loss": 0.7322, |
|
"step": 889 |
|
}, |
|
{ |
|
"epoch": 0.9072375127420998, |
|
"grad_norm": 9.99152946472168, |
|
"learning_rate": 4.451546941107046e-05, |
|
"loss": 0.484, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.908256880733945, |
|
"grad_norm": 6.232360363006592, |
|
"learning_rate": 4.432822557478194e-05, |
|
"loss": 0.5604, |
|
"step": 891 |
|
}, |
|
{ |
|
"epoch": 0.90927624872579, |
|
"grad_norm": 3.1541106700897217, |
|
"learning_rate": 4.414106228251446e-05, |
|
"loss": 0.2633, |
|
"step": 892 |
|
}, |
|
{ |
|
"epoch": 0.910295616717635, |
|
"grad_norm": 5.661106109619141, |
|
"learning_rate": 4.3953982192146006e-05, |
|
"loss": 0.2417, |
|
"step": 893 |
|
}, |
|
{ |
|
"epoch": 0.9113149847094801, |
|
"grad_norm": 4.497067451477051, |
|
"learning_rate": 4.3766987960372956e-05, |
|
"loss": 0.4481, |
|
"step": 894 |
|
}, |
|
{ |
|
"epoch": 0.9123343527013251, |
|
"grad_norm": 8.505694389343262, |
|
"learning_rate": 4.358008224267245e-05, |
|
"loss": 0.7402, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 0.9133537206931702, |
|
"grad_norm": 5.820054054260254, |
|
"learning_rate": 4.3393267693264686e-05, |
|
"loss": 0.4897, |
|
"step": 896 |
|
}, |
|
{ |
|
"epoch": 0.9143730886850153, |
|
"grad_norm": 7.943095684051514, |
|
"learning_rate": 4.320654696507511e-05, |
|
"loss": 0.5863, |
|
"step": 897 |
|
}, |
|
{ |
|
"epoch": 0.9153924566768603, |
|
"grad_norm": 10.6437349319458, |
|
"learning_rate": 4.301992270969692e-05, |
|
"loss": 0.7101, |
|
"step": 898 |
|
}, |
|
{ |
|
"epoch": 0.9164118246687054, |
|
"grad_norm": 3.8055593967437744, |
|
"learning_rate": 4.2833397577353284e-05, |
|
"loss": 0.2404, |
|
"step": 899 |
|
}, |
|
{ |
|
"epoch": 0.9174311926605505, |
|
"grad_norm": 8.539854049682617, |
|
"learning_rate": 4.26469742168597e-05, |
|
"loss": 0.5594, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.9184505606523955, |
|
"grad_norm": 5.611748218536377, |
|
"learning_rate": 4.2460655275586494e-05, |
|
"loss": 0.4047, |
|
"step": 901 |
|
}, |
|
{ |
|
"epoch": 0.9194699286442406, |
|
"grad_norm": 4.898343086242676, |
|
"learning_rate": 4.227444339942107e-05, |
|
"loss": 0.4865, |
|
"step": 902 |
|
}, |
|
{ |
|
"epoch": 0.9204892966360856, |
|
"grad_norm": 8.28711986541748, |
|
"learning_rate": 4.208834123273047e-05, |
|
"loss": 0.3909, |
|
"step": 903 |
|
}, |
|
{ |
|
"epoch": 0.9215086646279307, |
|
"grad_norm": 6.98935604095459, |
|
"learning_rate": 4.190235141832375e-05, |
|
"loss": 0.2808, |
|
"step": 904 |
|
}, |
|
{ |
|
"epoch": 0.9225280326197758, |
|
"grad_norm": 9.016980171203613, |
|
"learning_rate": 4.171647659741448e-05, |
|
"loss": 0.7509, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 0.9235474006116208, |
|
"grad_norm": 5.859550476074219, |
|
"learning_rate": 4.153071940958321e-05, |
|
"loss": 0.325, |
|
"step": 906 |
|
}, |
|
{ |
|
"epoch": 0.9245667686034659, |
|
"grad_norm": 7.970040321350098, |
|
"learning_rate": 4.134508249274002e-05, |
|
"loss": 0.5335, |
|
"step": 907 |
|
}, |
|
{ |
|
"epoch": 0.9255861365953109, |
|
"grad_norm": 6.2324981689453125, |
|
"learning_rate": 4.1159568483087e-05, |
|
"loss": 0.6193, |
|
"step": 908 |
|
}, |
|
{ |
|
"epoch": 0.926605504587156, |
|
"grad_norm": 5.227268218994141, |
|
"learning_rate": 4.0974180015080897e-05, |
|
"loss": 0.2974, |
|
"step": 909 |
|
}, |
|
{ |
|
"epoch": 0.9276248725790011, |
|
"grad_norm": 9.293944358825684, |
|
"learning_rate": 4.078891972139564e-05, |
|
"loss": 0.6725, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.928644240570846, |
|
"grad_norm": 10.003561019897461, |
|
"learning_rate": 4.060379023288495e-05, |
|
"loss": 0.8828, |
|
"step": 911 |
|
}, |
|
{ |
|
"epoch": 0.9296636085626911, |
|
"grad_norm": 9.07729721069336, |
|
"learning_rate": 4.0418794178545076e-05, |
|
"loss": 0.8751, |
|
"step": 912 |
|
}, |
|
{ |
|
"epoch": 0.9306829765545361, |
|
"grad_norm": 7.200821876525879, |
|
"learning_rate": 4.023393418547732e-05, |
|
"loss": 0.7019, |
|
"step": 913 |
|
}, |
|
{ |
|
"epoch": 0.9317023445463812, |
|
"grad_norm": 10.154699325561523, |
|
"learning_rate": 4.0049212878850793e-05, |
|
"loss": 0.7131, |
|
"step": 914 |
|
}, |
|
{ |
|
"epoch": 0.9327217125382263, |
|
"grad_norm": 7.271543025970459, |
|
"learning_rate": 3.98646328818652e-05, |
|
"loss": 0.2849, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 0.9337410805300713, |
|
"grad_norm": 9.933566093444824, |
|
"learning_rate": 3.96801968157135e-05, |
|
"loss": 0.8097, |
|
"step": 916 |
|
}, |
|
{ |
|
"epoch": 0.9347604485219164, |
|
"grad_norm": 5.370792865753174, |
|
"learning_rate": 3.949590729954467e-05, |
|
"loss": 0.3447, |
|
"step": 917 |
|
}, |
|
{ |
|
"epoch": 0.9357798165137615, |
|
"grad_norm": 8.846680641174316, |
|
"learning_rate": 3.931176695042664e-05, |
|
"loss": 0.8601, |
|
"step": 918 |
|
}, |
|
{ |
|
"epoch": 0.9367991845056065, |
|
"grad_norm": 5.936051368713379, |
|
"learning_rate": 3.912777838330893e-05, |
|
"loss": 0.4467, |
|
"step": 919 |
|
}, |
|
{ |
|
"epoch": 0.9378185524974516, |
|
"grad_norm": 10.40077018737793, |
|
"learning_rate": 3.8943944210985735e-05, |
|
"loss": 0.8137, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.9388379204892966, |
|
"grad_norm": 7.319591999053955, |
|
"learning_rate": 3.876026704405866e-05, |
|
"loss": 0.4527, |
|
"step": 921 |
|
}, |
|
{ |
|
"epoch": 0.9398572884811417, |
|
"grad_norm": 8.947883605957031, |
|
"learning_rate": 3.8576749490899686e-05, |
|
"loss": 0.7656, |
|
"step": 922 |
|
}, |
|
{ |
|
"epoch": 0.9408766564729868, |
|
"grad_norm": 10.776662826538086, |
|
"learning_rate": 3.839339415761416e-05, |
|
"loss": 1.1218, |
|
"step": 923 |
|
}, |
|
{ |
|
"epoch": 0.9418960244648318, |
|
"grad_norm": 2.9248359203338623, |
|
"learning_rate": 3.821020364800379e-05, |
|
"loss": 0.188, |
|
"step": 924 |
|
}, |
|
{ |
|
"epoch": 0.9429153924566769, |
|
"grad_norm": 9.73752212524414, |
|
"learning_rate": 3.8027180563529616e-05, |
|
"loss": 0.8454, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.9439347604485219, |
|
"grad_norm": 6.643280506134033, |
|
"learning_rate": 3.7844327503275136e-05, |
|
"loss": 0.5368, |
|
"step": 926 |
|
}, |
|
{ |
|
"epoch": 0.944954128440367, |
|
"grad_norm": 9.299040794372559, |
|
"learning_rate": 3.7661647063909294e-05, |
|
"loss": 0.7602, |
|
"step": 927 |
|
}, |
|
{ |
|
"epoch": 0.9459734964322121, |
|
"grad_norm": 6.660792827606201, |
|
"learning_rate": 3.747914183964974e-05, |
|
"loss": 0.4733, |
|
"step": 928 |
|
}, |
|
{ |
|
"epoch": 0.9469928644240571, |
|
"grad_norm": 5.206737995147705, |
|
"learning_rate": 3.729681442222587e-05, |
|
"loss": 0.2305, |
|
"step": 929 |
|
}, |
|
{ |
|
"epoch": 0.9480122324159022, |
|
"grad_norm": 9.746971130371094, |
|
"learning_rate": 3.711466740084211e-05, |
|
"loss": 0.7775, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.9490316004077471, |
|
"grad_norm": 9.825338363647461, |
|
"learning_rate": 3.6932703362141084e-05, |
|
"loss": 0.8859, |
|
"step": 931 |
|
}, |
|
{ |
|
"epoch": 0.9500509683995922, |
|
"grad_norm": 7.335731506347656, |
|
"learning_rate": 3.6750924890166914e-05, |
|
"loss": 0.3918, |
|
"step": 932 |
|
}, |
|
{ |
|
"epoch": 0.9510703363914373, |
|
"grad_norm": 6.4724931716918945, |
|
"learning_rate": 3.656933456632853e-05, |
|
"loss": 0.3842, |
|
"step": 933 |
|
}, |
|
{ |
|
"epoch": 0.9520897043832823, |
|
"grad_norm": 4.886312484741211, |
|
"learning_rate": 3.638793496936296e-05, |
|
"loss": 0.3719, |
|
"step": 934 |
|
}, |
|
{ |
|
"epoch": 0.9531090723751274, |
|
"grad_norm": 8.522834777832031, |
|
"learning_rate": 3.620672867529878e-05, |
|
"loss": 0.8043, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 0.9541284403669725, |
|
"grad_norm": 9.507696151733398, |
|
"learning_rate": 3.602571825741953e-05, |
|
"loss": 0.8282, |
|
"step": 936 |
|
}, |
|
{ |
|
"epoch": 0.9551478083588175, |
|
"grad_norm": 4.895750045776367, |
|
"learning_rate": 3.584490628622705e-05, |
|
"loss": 0.4599, |
|
"step": 937 |
|
}, |
|
{ |
|
"epoch": 0.9561671763506626, |
|
"grad_norm": 7.197470664978027, |
|
"learning_rate": 3.566429532940518e-05, |
|
"loss": 0.649, |
|
"step": 938 |
|
}, |
|
{ |
|
"epoch": 0.9571865443425076, |
|
"grad_norm": 6.60915470123291, |
|
"learning_rate": 3.548388795178307e-05, |
|
"loss": 0.4325, |
|
"step": 939 |
|
}, |
|
{ |
|
"epoch": 0.9582059123343527, |
|
"grad_norm": 10.626359939575195, |
|
"learning_rate": 3.5303686715298955e-05, |
|
"loss": 1.3108, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.9592252803261978, |
|
"grad_norm": 6.316555023193359, |
|
"learning_rate": 3.51236941789637e-05, |
|
"loss": 0.3018, |
|
"step": 941 |
|
}, |
|
{ |
|
"epoch": 0.9602446483180428, |
|
"grad_norm": 7.12025785446167, |
|
"learning_rate": 3.494391289882435e-05, |
|
"loss": 0.6258, |
|
"step": 942 |
|
}, |
|
{ |
|
"epoch": 0.9612640163098879, |
|
"grad_norm": 10.008544921875, |
|
"learning_rate": 3.476434542792805e-05, |
|
"loss": 1.2266, |
|
"step": 943 |
|
}, |
|
{ |
|
"epoch": 0.9622833843017329, |
|
"grad_norm": 8.917716979980469, |
|
"learning_rate": 3.4584994316285604e-05, |
|
"loss": 0.6593, |
|
"step": 944 |
|
}, |
|
{ |
|
"epoch": 0.963302752293578, |
|
"grad_norm": 5.837446689605713, |
|
"learning_rate": 3.4405862110835364e-05, |
|
"loss": 0.3096, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 0.9643221202854231, |
|
"grad_norm": 4.312796115875244, |
|
"learning_rate": 3.422695135540697e-05, |
|
"loss": 0.3436, |
|
"step": 946 |
|
}, |
|
{ |
|
"epoch": 0.9653414882772681, |
|
"grad_norm": 4.772927284240723, |
|
"learning_rate": 3.404826459068536e-05, |
|
"loss": 0.2497, |
|
"step": 947 |
|
}, |
|
{ |
|
"epoch": 0.9663608562691132, |
|
"grad_norm": 3.3676137924194336, |
|
"learning_rate": 3.386980435417457e-05, |
|
"loss": 0.1653, |
|
"step": 948 |
|
}, |
|
{ |
|
"epoch": 0.9673802242609582, |
|
"grad_norm": 6.203863143920898, |
|
"learning_rate": 3.369157318016176e-05, |
|
"loss": 0.469, |
|
"step": 949 |
|
}, |
|
{ |
|
"epoch": 0.9683995922528033, |
|
"grad_norm": 7.628493309020996, |
|
"learning_rate": 3.351357359968117e-05, |
|
"loss": 0.4919, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.9694189602446484, |
|
"grad_norm": 7.940287113189697, |
|
"learning_rate": 3.333580814047826e-05, |
|
"loss": 0.4788, |
|
"step": 951 |
|
}, |
|
{ |
|
"epoch": 0.9704383282364933, |
|
"grad_norm": 6.046499729156494, |
|
"learning_rate": 3.3158279326973766e-05, |
|
"loss": 0.3041, |
|
"step": 952 |
|
}, |
|
{ |
|
"epoch": 0.9714576962283384, |
|
"grad_norm": 4.314492225646973, |
|
"learning_rate": 3.298098968022782e-05, |
|
"loss": 0.3138, |
|
"step": 953 |
|
}, |
|
{ |
|
"epoch": 0.9724770642201835, |
|
"grad_norm": 8.91407585144043, |
|
"learning_rate": 3.2803941717904216e-05, |
|
"loss": 0.7758, |
|
"step": 954 |
|
}, |
|
{ |
|
"epoch": 0.9734964322120285, |
|
"grad_norm": 11.913896560668945, |
|
"learning_rate": 3.26271379542346e-05, |
|
"loss": 0.6974, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 0.9745158002038736, |
|
"grad_norm": 4.831221580505371, |
|
"learning_rate": 3.2450580899982795e-05, |
|
"loss": 0.2964, |
|
"step": 956 |
|
}, |
|
{ |
|
"epoch": 0.9755351681957186, |
|
"grad_norm": 6.116502285003662, |
|
"learning_rate": 3.2274273062409154e-05, |
|
"loss": 0.3473, |
|
"step": 957 |
|
}, |
|
{ |
|
"epoch": 0.9765545361875637, |
|
"grad_norm": 11.75236988067627, |
|
"learning_rate": 3.2098216945234946e-05, |
|
"loss": 0.8905, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 0.9775739041794088, |
|
"grad_norm": 3.468975067138672, |
|
"learning_rate": 3.192241504860675e-05, |
|
"loss": 0.2521, |
|
"step": 959 |
|
}, |
|
{ |
|
"epoch": 0.9785932721712538, |
|
"grad_norm": 7.624709606170654, |
|
"learning_rate": 3.1746869869061063e-05, |
|
"loss": 0.4462, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.9796126401630989, |
|
"grad_norm": 9.019265174865723, |
|
"learning_rate": 3.157158389948871e-05, |
|
"loss": 0.7842, |
|
"step": 961 |
|
}, |
|
{ |
|
"epoch": 0.9806320081549439, |
|
"grad_norm": 4.77131986618042, |
|
"learning_rate": 3.1396559629099574e-05, |
|
"loss": 0.2973, |
|
"step": 962 |
|
}, |
|
{ |
|
"epoch": 0.981651376146789, |
|
"grad_norm": 8.40596866607666, |
|
"learning_rate": 3.122179954338716e-05, |
|
"loss": 0.6026, |
|
"step": 963 |
|
}, |
|
{ |
|
"epoch": 0.9826707441386341, |
|
"grad_norm": 6.705322265625, |
|
"learning_rate": 3.1047306124093335e-05, |
|
"loss": 0.4026, |
|
"step": 964 |
|
}, |
|
{ |
|
"epoch": 0.9836901121304791, |
|
"grad_norm": 10.35732364654541, |
|
"learning_rate": 3.087308184917308e-05, |
|
"loss": 0.9181, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 0.9847094801223242, |
|
"grad_norm": 6.806704998016357, |
|
"learning_rate": 3.069912919275926e-05, |
|
"loss": 0.473, |
|
"step": 966 |
|
}, |
|
{ |
|
"epoch": 0.9857288481141692, |
|
"grad_norm": 10.28345012664795, |
|
"learning_rate": 3.0525450625127575e-05, |
|
"loss": 0.7152, |
|
"step": 967 |
|
}, |
|
{ |
|
"epoch": 0.9867482161060143, |
|
"grad_norm": 11.785171508789062, |
|
"learning_rate": 3.0352048612661416e-05, |
|
"loss": 0.9519, |
|
"step": 968 |
|
}, |
|
{ |
|
"epoch": 0.9877675840978594, |
|
"grad_norm": 8.55274772644043, |
|
"learning_rate": 3.017892561781682e-05, |
|
"loss": 0.5322, |
|
"step": 969 |
|
}, |
|
{ |
|
"epoch": 0.9887869520897044, |
|
"grad_norm": 8.597644805908203, |
|
"learning_rate": 3.0006084099087595e-05, |
|
"loss": 0.8257, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.9898063200815495, |
|
"grad_norm": 6.743808746337891, |
|
"learning_rate": 2.983352651097031e-05, |
|
"loss": 0.5648, |
|
"step": 971 |
|
}, |
|
{ |
|
"epoch": 0.9908256880733946, |
|
"grad_norm": 10.981080055236816, |
|
"learning_rate": 2.9661255303929486e-05, |
|
"loss": 0.909, |
|
"step": 972 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 2943, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 972, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 0.0, |
|
"train_batch_size": 32, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|