|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 3.0, |
|
"eval_steps": 500, |
|
"global_step": 3744, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0008012820512820513, |
|
"grad_norm": 174.17773546874915, |
|
"learning_rate": 1.3333333333333333e-09, |
|
"logits/chosen": -0.416015625, |
|
"logits/rejected": -0.498046875, |
|
"logps/chosen": -520.0, |
|
"logps/rejected": -148.0, |
|
"loss": 0.6914, |
|
"rewards/accuracies": 0.0, |
|
"rewards/chosen": 0.0, |
|
"rewards/margins": 0.0, |
|
"rewards/rejected": 0.0, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.008012820512820512, |
|
"grad_norm": 171.51311699077868, |
|
"learning_rate": 1.3333333333333334e-08, |
|
"logits/chosen": -0.4921875, |
|
"logits/rejected": -0.60546875, |
|
"logps/chosen": -524.0, |
|
"logps/rejected": -148.0, |
|
"loss": 0.7014, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 0.0167236328125, |
|
"rewards/margins": 0.01531982421875, |
|
"rewards/rejected": 0.00141143798828125, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.016025641025641024, |
|
"grad_norm": 135.35847183213565, |
|
"learning_rate": 2.6666666666666667e-08, |
|
"logits/chosen": -0.375, |
|
"logits/rejected": -0.6484375, |
|
"logps/chosen": -486.0, |
|
"logps/rejected": -149.0, |
|
"loss": 0.7055, |
|
"rewards/accuracies": 0.20000000298023224, |
|
"rewards/chosen": -0.030029296875, |
|
"rewards/margins": -0.0263671875, |
|
"rewards/rejected": -0.003753662109375, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.02403846153846154, |
|
"grad_norm": 151.09602882027437, |
|
"learning_rate": 4e-08, |
|
"logits/chosen": -0.3046875, |
|
"logits/rejected": -0.4140625, |
|
"logps/chosen": -472.0, |
|
"logps/rejected": -147.0, |
|
"loss": 0.6877, |
|
"rewards/accuracies": 0.375, |
|
"rewards/chosen": 0.045166015625, |
|
"rewards/margins": 0.050048828125, |
|
"rewards/rejected": -0.0050048828125, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.03205128205128205, |
|
"grad_norm": 139.2330474975386, |
|
"learning_rate": 5.3333333333333334e-08, |
|
"logits/chosen": -0.205078125, |
|
"logits/rejected": -0.416015625, |
|
"logps/chosen": -488.0, |
|
"logps/rejected": -134.0, |
|
"loss": 0.6678, |
|
"rewards/accuracies": 0.32499998807907104, |
|
"rewards/chosen": 0.0250244140625, |
|
"rewards/margins": 0.041259765625, |
|
"rewards/rejected": -0.0162353515625, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.04006410256410257, |
|
"grad_norm": 115.19046941684343, |
|
"learning_rate": 6.666666666666667e-08, |
|
"logits/chosen": -0.45703125, |
|
"logits/rejected": -0.5078125, |
|
"logps/chosen": -532.0, |
|
"logps/rejected": -153.0, |
|
"loss": 0.624, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.1298828125, |
|
"rewards/margins": 0.162109375, |
|
"rewards/rejected": -0.031982421875, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.04807692307692308, |
|
"grad_norm": 121.74476376843914, |
|
"learning_rate": 8e-08, |
|
"logits/chosen": -0.51171875, |
|
"logits/rejected": -0.640625, |
|
"logps/chosen": -516.0, |
|
"logps/rejected": -151.0, |
|
"loss": 0.59, |
|
"rewards/accuracies": 0.800000011920929, |
|
"rewards/chosen": 0.1298828125, |
|
"rewards/margins": 0.1962890625, |
|
"rewards/rejected": -0.06591796875, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.05608974358974359, |
|
"grad_norm": 143.6496465915542, |
|
"learning_rate": 9.333333333333334e-08, |
|
"logits/chosen": -0.5625, |
|
"logits/rejected": -0.5234375, |
|
"logps/chosen": -500.0, |
|
"logps/rejected": -142.0, |
|
"loss": 0.5251, |
|
"rewards/accuracies": 0.925000011920929, |
|
"rewards/chosen": 0.279296875, |
|
"rewards/margins": 0.380859375, |
|
"rewards/rejected": -0.10009765625, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.0641025641025641, |
|
"grad_norm": 88.39739617260635, |
|
"learning_rate": 1.0666666666666667e-07, |
|
"logits/chosen": -0.359375, |
|
"logits/rejected": -0.5390625, |
|
"logps/chosen": -516.0, |
|
"logps/rejected": -143.0, |
|
"loss": 0.4566, |
|
"rewards/accuracies": 0.9750000238418579, |
|
"rewards/chosen": 0.400390625, |
|
"rewards/margins": 0.546875, |
|
"rewards/rejected": -0.1455078125, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.07211538461538461, |
|
"grad_norm": 74.73563635718321, |
|
"learning_rate": 1.2e-07, |
|
"logits/chosen": -0.341796875, |
|
"logits/rejected": -0.3984375, |
|
"logps/chosen": -532.0, |
|
"logps/rejected": -143.0, |
|
"loss": 0.3592, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.65625, |
|
"rewards/margins": 0.890625, |
|
"rewards/rejected": -0.2353515625, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.08012820512820513, |
|
"grad_norm": 50.58405137256872, |
|
"learning_rate": 1.3333333333333334e-07, |
|
"logits/chosen": -0.384765625, |
|
"logits/rejected": -0.52734375, |
|
"logps/chosen": -488.0, |
|
"logps/rejected": -140.0, |
|
"loss": 0.285, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.875, |
|
"rewards/margins": 1.1875, |
|
"rewards/rejected": -0.30859375, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.08814102564102565, |
|
"grad_norm": 32.60717298377466, |
|
"learning_rate": 1.4666666666666666e-07, |
|
"logits/chosen": -0.388671875, |
|
"logits/rejected": -0.453125, |
|
"logps/chosen": -524.0, |
|
"logps/rejected": -143.0, |
|
"loss": 0.1962, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 1.2890625, |
|
"rewards/margins": 1.734375, |
|
"rewards/rejected": -0.447265625, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.09615384615384616, |
|
"grad_norm": 28.73827423954734, |
|
"learning_rate": 1.6e-07, |
|
"logits/chosen": -0.31640625, |
|
"logits/rejected": -0.5, |
|
"logps/chosen": -510.0, |
|
"logps/rejected": -139.0, |
|
"loss": 0.1427, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 1.5234375, |
|
"rewards/margins": 2.203125, |
|
"rewards/rejected": -0.6796875, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.10416666666666667, |
|
"grad_norm": 29.460470106894387, |
|
"learning_rate": 1.7333333333333332e-07, |
|
"logits/chosen": -0.38671875, |
|
"logits/rejected": -0.45703125, |
|
"logps/chosen": -486.0, |
|
"logps/rejected": -153.0, |
|
"loss": 0.1047, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 1.5234375, |
|
"rewards/margins": 2.40625, |
|
"rewards/rejected": -0.89453125, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.11217948717948718, |
|
"grad_norm": 10.689126569457526, |
|
"learning_rate": 1.8666666666666667e-07, |
|
"logits/chosen": -0.390625, |
|
"logits/rejected": -0.5390625, |
|
"logps/chosen": -482.0, |
|
"logps/rejected": -165.0, |
|
"loss": 0.0737, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 1.796875, |
|
"rewards/margins": 3.0625, |
|
"rewards/rejected": -1.265625, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.1201923076923077, |
|
"grad_norm": 8.734640947393299, |
|
"learning_rate": 2e-07, |
|
"logits/chosen": -0.2470703125, |
|
"logits/rejected": -0.15625, |
|
"logps/chosen": -528.0, |
|
"logps/rejected": -151.0, |
|
"loss": 0.0376, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 2.1875, |
|
"rewards/margins": 3.84375, |
|
"rewards/rejected": -1.6484375, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.1282051282051282, |
|
"grad_norm": 6.525920920703583, |
|
"learning_rate": 2.1333333333333334e-07, |
|
"logits/chosen": -0.32421875, |
|
"logits/rejected": -0.482421875, |
|
"logps/chosen": -504.0, |
|
"logps/rejected": -153.0, |
|
"loss": 0.0367, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 2.46875, |
|
"rewards/margins": 4.3125, |
|
"rewards/rejected": -1.859375, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.1362179487179487, |
|
"grad_norm": 15.370405135278967, |
|
"learning_rate": 2.2666666666666663e-07, |
|
"logits/chosen": -0.291015625, |
|
"logits/rejected": -0.515625, |
|
"logps/chosen": -468.0, |
|
"logps/rejected": -153.0, |
|
"loss": 0.0187, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 2.671875, |
|
"rewards/margins": 5.125, |
|
"rewards/rejected": -2.4375, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.14423076923076922, |
|
"grad_norm": 0.8423225813275931, |
|
"learning_rate": 2.4e-07, |
|
"logits/chosen": -0.45703125, |
|
"logits/rejected": -0.41015625, |
|
"logps/chosen": -510.0, |
|
"logps/rejected": -157.0, |
|
"loss": 0.0113, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 2.84375, |
|
"rewards/margins": 5.4375, |
|
"rewards/rejected": -2.59375, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.15224358974358973, |
|
"grad_norm": 2.214593977360192, |
|
"learning_rate": 2.533333333333333e-07, |
|
"logits/chosen": -0.47265625, |
|
"logits/rejected": -0.57421875, |
|
"logps/chosen": -492.0, |
|
"logps/rejected": -171.0, |
|
"loss": 0.0079, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 3.03125, |
|
"rewards/margins": 6.15625, |
|
"rewards/rejected": -3.109375, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.16025641025641027, |
|
"grad_norm": 6.07567342094023, |
|
"learning_rate": 2.6666666666666667e-07, |
|
"logits/chosen": -0.390625, |
|
"logits/rejected": -0.265625, |
|
"logps/chosen": -474.0, |
|
"logps/rejected": -177.0, |
|
"loss": 0.0057, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 3.265625, |
|
"rewards/margins": 7.0, |
|
"rewards/rejected": -3.71875, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.16826923076923078, |
|
"grad_norm": 2.688929236247746, |
|
"learning_rate": 2.8e-07, |
|
"logits/chosen": -0.474609375, |
|
"logits/rejected": -0.2109375, |
|
"logps/chosen": -524.0, |
|
"logps/rejected": -181.0, |
|
"loss": 0.0079, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 3.8125, |
|
"rewards/margins": 7.8125, |
|
"rewards/rejected": -4.0, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.1762820512820513, |
|
"grad_norm": 10.457975237753626, |
|
"learning_rate": 2.933333333333333e-07, |
|
"logits/chosen": -0.392578125, |
|
"logits/rejected": -0.341796875, |
|
"logps/chosen": -450.0, |
|
"logps/rejected": -188.0, |
|
"loss": 0.0079, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 4.0625, |
|
"rewards/margins": 8.375, |
|
"rewards/rejected": -4.3125, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.1842948717948718, |
|
"grad_norm": 3.1954975450295318, |
|
"learning_rate": 3.066666666666666e-07, |
|
"logits/chosen": -0.30859375, |
|
"logits/rejected": -0.1923828125, |
|
"logps/chosen": -468.0, |
|
"logps/rejected": -188.0, |
|
"loss": 0.0038, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 3.890625, |
|
"rewards/margins": 8.0625, |
|
"rewards/rejected": -4.1875, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.19230769230769232, |
|
"grad_norm": 0.928491797463352, |
|
"learning_rate": 3.2e-07, |
|
"logits/chosen": -0.392578125, |
|
"logits/rejected": -0.2119140625, |
|
"logps/chosen": -488.0, |
|
"logps/rejected": -192.0, |
|
"loss": 0.0038, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 4.125, |
|
"rewards/margins": 8.625, |
|
"rewards/rejected": -4.5, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.20032051282051283, |
|
"grad_norm": 0.9581365929864241, |
|
"learning_rate": 3.333333333333333e-07, |
|
"logits/chosen": -0.244140625, |
|
"logits/rejected": -0.1845703125, |
|
"logps/chosen": -466.0, |
|
"logps/rejected": -204.0, |
|
"loss": 0.0054, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 4.09375, |
|
"rewards/margins": 9.125, |
|
"rewards/rejected": -5.0625, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.20833333333333334, |
|
"grad_norm": 0.31677787690134823, |
|
"learning_rate": 3.4666666666666665e-07, |
|
"logits/chosen": -0.2197265625, |
|
"logits/rejected": -0.21875, |
|
"logps/chosen": -460.0, |
|
"logps/rejected": -189.0, |
|
"loss": 0.0009, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 3.640625, |
|
"rewards/margins": 8.4375, |
|
"rewards/rejected": -4.75, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.21634615384615385, |
|
"grad_norm": 0.21999814737171097, |
|
"learning_rate": 3.6e-07, |
|
"logits/chosen": -0.298828125, |
|
"logits/rejected": -0.1708984375, |
|
"logps/chosen": -488.0, |
|
"logps/rejected": -197.0, |
|
"loss": 0.0031, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 4.90625, |
|
"rewards/margins": 9.9375, |
|
"rewards/rejected": -5.0, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.22435897435897437, |
|
"grad_norm": 0.18034977673530014, |
|
"learning_rate": 3.7333333333333334e-07, |
|
"logits/chosen": -0.29296875, |
|
"logits/rejected": 0.078125, |
|
"logps/chosen": -448.0, |
|
"logps/rejected": -192.0, |
|
"loss": 0.002, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 3.875, |
|
"rewards/margins": 8.9375, |
|
"rewards/rejected": -5.03125, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.23237179487179488, |
|
"grad_norm": 0.00815873147574371, |
|
"learning_rate": 3.8666666666666664e-07, |
|
"logits/chosen": -0.1865234375, |
|
"logits/rejected": -0.0712890625, |
|
"logps/chosen": -464.0, |
|
"logps/rejected": -209.0, |
|
"loss": 0.0036, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 5.09375, |
|
"rewards/margins": 10.75, |
|
"rewards/rejected": -5.6875, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.2403846153846154, |
|
"grad_norm": 0.023842090554674955, |
|
"learning_rate": 4e-07, |
|
"logits/chosen": -0.416015625, |
|
"logits/rejected": -0.125, |
|
"logps/chosen": -444.0, |
|
"logps/rejected": -203.0, |
|
"loss": 0.0003, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 4.65625, |
|
"rewards/margins": 10.625, |
|
"rewards/rejected": -5.9375, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.2483974358974359, |
|
"grad_norm": 0.17900978432871079, |
|
"learning_rate": 4.1333333333333333e-07, |
|
"logits/chosen": -0.4140625, |
|
"logits/rejected": -0.1318359375, |
|
"logps/chosen": -466.0, |
|
"logps/rejected": -195.0, |
|
"loss": 0.0015, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 4.75, |
|
"rewards/margins": 10.25, |
|
"rewards/rejected": -5.46875, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.2564102564102564, |
|
"grad_norm": 0.01015318989781134, |
|
"learning_rate": 4.266666666666667e-07, |
|
"logits/chosen": -0.3671875, |
|
"logits/rejected": -0.17578125, |
|
"logps/chosen": -462.0, |
|
"logps/rejected": -202.0, |
|
"loss": 0.0008, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 4.75, |
|
"rewards/margins": 10.8125, |
|
"rewards/rejected": -6.03125, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.2644230769230769, |
|
"grad_norm": 0.16299050612501387, |
|
"learning_rate": 4.3999999999999997e-07, |
|
"logits/chosen": -0.2099609375, |
|
"logits/rejected": -0.130859375, |
|
"logps/chosen": -458.0, |
|
"logps/rejected": -198.0, |
|
"loss": 0.0007, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 4.90625, |
|
"rewards/margins": 11.0, |
|
"rewards/rejected": -6.09375, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.2724358974358974, |
|
"grad_norm": 0.12371498003487358, |
|
"learning_rate": 4.5333333333333326e-07, |
|
"logits/chosen": -0.259765625, |
|
"logits/rejected": -0.03955078125, |
|
"logps/chosen": -450.0, |
|
"logps/rejected": -194.0, |
|
"loss": 0.0003, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 5.21875, |
|
"rewards/margins": 11.875, |
|
"rewards/rejected": -6.65625, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.28044871794871795, |
|
"grad_norm": 0.010975632589342161, |
|
"learning_rate": 4.6666666666666666e-07, |
|
"logits/chosen": -0.384765625, |
|
"logits/rejected": -0.1494140625, |
|
"logps/chosen": -448.0, |
|
"logps/rejected": -210.0, |
|
"loss": 0.0005, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 5.1875, |
|
"rewards/margins": 11.75, |
|
"rewards/rejected": -6.53125, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.28846153846153844, |
|
"grad_norm": 0.0226871218171989, |
|
"learning_rate": 4.8e-07, |
|
"logits/chosen": -0.419921875, |
|
"logits/rejected": -0.11474609375, |
|
"logps/chosen": -446.0, |
|
"logps/rejected": -214.0, |
|
"loss": 0.0009, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 4.5, |
|
"rewards/margins": 11.75, |
|
"rewards/rejected": -7.25, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.296474358974359, |
|
"grad_norm": 0.03224104991667029, |
|
"learning_rate": 4.933333333333333e-07, |
|
"logits/chosen": -0.21875, |
|
"logits/rejected": -0.1201171875, |
|
"logps/chosen": -452.0, |
|
"logps/rejected": -221.0, |
|
"loss": 0.0002, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 5.65625, |
|
"rewards/margins": 12.9375, |
|
"rewards/rejected": -7.28125, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.30448717948717946, |
|
"grad_norm": 0.026704517051017115, |
|
"learning_rate": 4.992579400415554e-07, |
|
"logits/chosen": -0.19921875, |
|
"logits/rejected": -0.2158203125, |
|
"logps/chosen": -444.0, |
|
"logps/rejected": -205.0, |
|
"loss": 0.0002, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 4.71875, |
|
"rewards/margins": 11.4375, |
|
"rewards/rejected": -6.71875, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.3125, |
|
"grad_norm": 0.0005925672776542448, |
|
"learning_rate": 4.97773820124666e-07, |
|
"logits/chosen": -0.1552734375, |
|
"logits/rejected": 0.0019073486328125, |
|
"logps/chosen": -442.0, |
|
"logps/rejected": -206.0, |
|
"loss": 0.0004, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 5.5, |
|
"rewards/margins": 12.5, |
|
"rewards/rejected": -7.0, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.32051282051282054, |
|
"grad_norm": 0.003678184459825849, |
|
"learning_rate": 4.962897002077768e-07, |
|
"logits/chosen": -0.373046875, |
|
"logits/rejected": -0.1220703125, |
|
"logps/chosen": -466.0, |
|
"logps/rejected": -211.0, |
|
"loss": 0.0005, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 5.625, |
|
"rewards/margins": 12.6875, |
|
"rewards/rejected": -7.03125, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.328525641025641, |
|
"grad_norm": 0.19343779207041278, |
|
"learning_rate": 4.948055802908874e-07, |
|
"logits/chosen": -0.3125, |
|
"logits/rejected": -0.1953125, |
|
"logps/chosen": -474.0, |
|
"logps/rejected": -211.0, |
|
"loss": 0.0002, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 5.375, |
|
"rewards/margins": 12.75, |
|
"rewards/rejected": -7.34375, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.33653846153846156, |
|
"grad_norm": 0.009634660344587258, |
|
"learning_rate": 4.933214603739982e-07, |
|
"logits/chosen": -0.2392578125, |
|
"logits/rejected": -0.0849609375, |
|
"logps/chosen": -472.0, |
|
"logps/rejected": -220.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 5.84375, |
|
"rewards/margins": 13.0, |
|
"rewards/rejected": -7.1875, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.34455128205128205, |
|
"grad_norm": 0.05052144148498111, |
|
"learning_rate": 4.918373404571089e-07, |
|
"logits/chosen": -0.1357421875, |
|
"logits/rejected": -0.1162109375, |
|
"logps/chosen": -468.0, |
|
"logps/rejected": -216.0, |
|
"loss": 0.0003, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.375, |
|
"rewards/margins": 13.75, |
|
"rewards/rejected": -7.40625, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.3525641025641026, |
|
"grad_norm": 0.08858979507238661, |
|
"learning_rate": 4.903532205402196e-07, |
|
"logits/chosen": -0.310546875, |
|
"logits/rejected": -0.1572265625, |
|
"logps/chosen": -430.0, |
|
"logps/rejected": -230.0, |
|
"loss": 0.0001, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 5.25, |
|
"rewards/margins": 13.375, |
|
"rewards/rejected": -8.125, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.3605769230769231, |
|
"grad_norm": 0.3135783793799678, |
|
"learning_rate": 4.888691006233304e-07, |
|
"logits/chosen": -0.39453125, |
|
"logits/rejected": -0.01251220703125, |
|
"logps/chosen": -466.0, |
|
"logps/rejected": -224.0, |
|
"loss": 0.0002, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 5.8125, |
|
"rewards/margins": 13.625, |
|
"rewards/rejected": -7.84375, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.3685897435897436, |
|
"grad_norm": 0.11874656728243187, |
|
"learning_rate": 4.873849807064411e-07, |
|
"logits/chosen": -0.27734375, |
|
"logits/rejected": -0.01361083984375, |
|
"logps/chosen": -456.0, |
|
"logps/rejected": -223.0, |
|
"loss": 0.0001, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 5.9375, |
|
"rewards/margins": 13.875, |
|
"rewards/rejected": -7.875, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.3766025641025641, |
|
"grad_norm": 0.0024318317438614843, |
|
"learning_rate": 4.859008607895517e-07, |
|
"logits/chosen": -0.1630859375, |
|
"logits/rejected": 0.1650390625, |
|
"logps/chosen": -450.0, |
|
"logps/rejected": -217.0, |
|
"loss": 0.0001, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.53125, |
|
"rewards/margins": 14.625, |
|
"rewards/rejected": -8.0625, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.38461538461538464, |
|
"grad_norm": 0.004326034777315378, |
|
"learning_rate": 4.844167408726625e-07, |
|
"logits/chosen": -0.53515625, |
|
"logits/rejected": -0.265625, |
|
"logps/chosen": -490.0, |
|
"logps/rejected": -216.0, |
|
"loss": 0.0003, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.5, |
|
"rewards/margins": 14.0, |
|
"rewards/rejected": -7.5, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.3926282051282051, |
|
"grad_norm": 0.23885883299744431, |
|
"learning_rate": 4.829326209557732e-07, |
|
"logits/chosen": -0.423828125, |
|
"logits/rejected": -0.162109375, |
|
"logps/chosen": -472.0, |
|
"logps/rejected": -212.0, |
|
"loss": 0.0001, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.25, |
|
"rewards/margins": 13.9375, |
|
"rewards/rejected": -7.6875, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.40064102564102566, |
|
"grad_norm": 0.0008671673486030223, |
|
"learning_rate": 4.814485010388839e-07, |
|
"logits/chosen": -0.29296875, |
|
"logits/rejected": -0.05322265625, |
|
"logps/chosen": -458.0, |
|
"logps/rejected": -223.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.1875, |
|
"rewards/margins": 14.25, |
|
"rewards/rejected": -8.0625, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.40865384615384615, |
|
"grad_norm": 0.0314369300330135, |
|
"learning_rate": 4.799643811219946e-07, |
|
"logits/chosen": -0.365234375, |
|
"logits/rejected": -0.228515625, |
|
"logps/chosen": -482.0, |
|
"logps/rejected": -198.0, |
|
"loss": 0.0002, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.0625, |
|
"rewards/margins": 13.6875, |
|
"rewards/rejected": -7.625, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.4166666666666667, |
|
"grad_norm": 0.0660602788437808, |
|
"learning_rate": 4.784802612051053e-07, |
|
"logits/chosen": -0.37109375, |
|
"logits/rejected": -0.296875, |
|
"logps/chosen": -468.0, |
|
"logps/rejected": -210.0, |
|
"loss": 0.0001, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.46875, |
|
"rewards/margins": 14.4375, |
|
"rewards/rejected": -7.9375, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.42467948717948717, |
|
"grad_norm": 0.02250458587989884, |
|
"learning_rate": 4.769961412882161e-07, |
|
"logits/chosen": -0.279296875, |
|
"logits/rejected": -0.0625, |
|
"logps/chosen": -452.0, |
|
"logps/rejected": -225.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.09375, |
|
"rewards/margins": 14.0625, |
|
"rewards/rejected": -7.9375, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.4326923076923077, |
|
"grad_norm": 0.0029081003054318715, |
|
"learning_rate": 4.755120213713268e-07, |
|
"logits/chosen": -0.3125, |
|
"logits/rejected": -0.310546875, |
|
"logps/chosen": -454.0, |
|
"logps/rejected": -224.0, |
|
"loss": 0.0001, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 5.8125, |
|
"rewards/margins": 14.125, |
|
"rewards/rejected": -8.3125, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.4407051282051282, |
|
"grad_norm": 0.0015550221403272514, |
|
"learning_rate": 4.740279014544375e-07, |
|
"logits/chosen": -0.259765625, |
|
"logits/rejected": 0.03125, |
|
"logps/chosen": -446.0, |
|
"logps/rejected": -233.0, |
|
"loss": 0.0002, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.375, |
|
"rewards/margins": 14.6875, |
|
"rewards/rejected": -8.3125, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.44871794871794873, |
|
"grad_norm": 0.051495984513767745, |
|
"learning_rate": 4.725437815375482e-07, |
|
"logits/chosen": -0.279296875, |
|
"logits/rejected": -0.025146484375, |
|
"logps/chosen": -444.0, |
|
"logps/rejected": -213.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.59375, |
|
"rewards/margins": 14.5, |
|
"rewards/rejected": -7.90625, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.4567307692307692, |
|
"grad_norm": 0.0010626712949092733, |
|
"learning_rate": 4.710596616206589e-07, |
|
"logits/chosen": -0.1298828125, |
|
"logits/rejected": -0.07177734375, |
|
"logps/chosen": -438.0, |
|
"logps/rejected": -223.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.28125, |
|
"rewards/margins": 14.75, |
|
"rewards/rejected": -8.4375, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.46474358974358976, |
|
"grad_norm": 0.09244610894587413, |
|
"learning_rate": 4.6957554170376963e-07, |
|
"logits/chosen": -0.212890625, |
|
"logits/rejected": 0.002471923828125, |
|
"logps/chosen": -446.0, |
|
"logps/rejected": -223.0, |
|
"loss": 0.0016, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.125, |
|
"rewards/margins": 14.875, |
|
"rewards/rejected": -8.8125, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.47275641025641024, |
|
"grad_norm": 0.0009965192812398677, |
|
"learning_rate": 4.680914217868804e-07, |
|
"logits/chosen": -0.21875, |
|
"logits/rejected": 0.10107421875, |
|
"logps/chosen": -448.0, |
|
"logps/rejected": -234.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.90625, |
|
"rewards/margins": 16.25, |
|
"rewards/rejected": -9.3125, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.4807692307692308, |
|
"grad_norm": 0.00043094256911758435, |
|
"learning_rate": 4.666073018699911e-07, |
|
"logits/chosen": -0.298828125, |
|
"logits/rejected": -0.1708984375, |
|
"logps/chosen": -446.0, |
|
"logps/rejected": -244.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.5, |
|
"rewards/margins": 15.625, |
|
"rewards/rejected": -9.125, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.48878205128205127, |
|
"grad_norm": 0.0019797557895973733, |
|
"learning_rate": 4.6512318195310177e-07, |
|
"logits/chosen": -0.44921875, |
|
"logits/rejected": -0.23828125, |
|
"logps/chosen": -466.0, |
|
"logps/rejected": -229.0, |
|
"loss": 0.0001, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 5.5625, |
|
"rewards/margins": 14.5625, |
|
"rewards/rejected": -9.0, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.4967948717948718, |
|
"grad_norm": 0.003434529146359255, |
|
"learning_rate": 4.636390620362125e-07, |
|
"logits/chosen": -0.1396484375, |
|
"logits/rejected": 0.0458984375, |
|
"logps/chosen": -430.0, |
|
"logps/rejected": -211.0, |
|
"loss": 0.0001, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.6875, |
|
"rewards/margins": 14.5, |
|
"rewards/rejected": -7.78125, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.5048076923076923, |
|
"grad_norm": 0.008343223604426265, |
|
"learning_rate": 4.621549421193232e-07, |
|
"logits/chosen": -0.484375, |
|
"logits/rejected": -0.005218505859375, |
|
"logps/chosen": -484.0, |
|
"logps/rejected": -246.0, |
|
"loss": 0.0003, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.65625, |
|
"rewards/margins": 15.8125, |
|
"rewards/rejected": -9.1875, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.5128205128205128, |
|
"grad_norm": 0.19467872313110163, |
|
"learning_rate": 4.606708222024339e-07, |
|
"logits/chosen": -0.36328125, |
|
"logits/rejected": 0.0078125, |
|
"logps/chosen": -468.0, |
|
"logps/rejected": -224.0, |
|
"loss": 0.0001, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.59375, |
|
"rewards/margins": 16.25, |
|
"rewards/rejected": -9.625, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.5208333333333334, |
|
"grad_norm": 0.00018302103223060052, |
|
"learning_rate": 4.591867022855446e-07, |
|
"logits/chosen": -0.259765625, |
|
"logits/rejected": 0.2080078125, |
|
"logps/chosen": -462.0, |
|
"logps/rejected": -242.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.59375, |
|
"rewards/margins": 16.75, |
|
"rewards/rejected": -10.125, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.5288461538461539, |
|
"grad_norm": 0.0005903642949682872, |
|
"learning_rate": 4.577025823686554e-07, |
|
"logits/chosen": -0.150390625, |
|
"logits/rejected": 0.06689453125, |
|
"logps/chosen": -448.0, |
|
"logps/rejected": -218.0, |
|
"loss": 0.0011, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.25, |
|
"rewards/margins": 15.3125, |
|
"rewards/rejected": -9.0625, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.5368589743589743, |
|
"grad_norm": 0.001573408566067085, |
|
"learning_rate": 4.562184624517661e-07, |
|
"logits/chosen": -0.330078125, |
|
"logits/rejected": -0.087890625, |
|
"logps/chosen": -474.0, |
|
"logps/rejected": -225.0, |
|
"loss": 0.0001, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.75, |
|
"rewards/margins": 14.6875, |
|
"rewards/rejected": -7.9375, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.5448717948717948, |
|
"grad_norm": 0.002613995555658358, |
|
"learning_rate": 4.547343425348768e-07, |
|
"logits/chosen": -0.353515625, |
|
"logits/rejected": -0.016357421875, |
|
"logps/chosen": -442.0, |
|
"logps/rejected": -237.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.4375, |
|
"rewards/margins": 15.75, |
|
"rewards/rejected": -9.3125, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.5528846153846154, |
|
"grad_norm": 0.0009656749108153287, |
|
"learning_rate": 4.5325022261798753e-07, |
|
"logits/chosen": -0.28515625, |
|
"logits/rejected": 0.1845703125, |
|
"logps/chosen": -472.0, |
|
"logps/rejected": -238.0, |
|
"loss": 0.0001, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.96875, |
|
"rewards/margins": 16.625, |
|
"rewards/rejected": -9.6875, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.5608974358974359, |
|
"grad_norm": 0.000259248888359743, |
|
"learning_rate": 4.517661027010982e-07, |
|
"logits/chosen": -0.205078125, |
|
"logits/rejected": 0.1630859375, |
|
"logps/chosen": -428.0, |
|
"logps/rejected": -258.0, |
|
"loss": 0.0001, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.375, |
|
"rewards/margins": 17.625, |
|
"rewards/rejected": -11.25, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.5689102564102564, |
|
"grad_norm": 0.07251165659160795, |
|
"learning_rate": 4.502819827842089e-07, |
|
"logits/chosen": -0.193359375, |
|
"logits/rejected": 0.038818359375, |
|
"logps/chosen": -454.0, |
|
"logps/rejected": -248.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.0625, |
|
"rewards/margins": 17.625, |
|
"rewards/rejected": -11.5625, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.5769230769230769, |
|
"grad_norm": 0.0017128273049482937, |
|
"learning_rate": 4.487978628673196e-07, |
|
"logits/chosen": -0.1416015625, |
|
"logits/rejected": -0.0595703125, |
|
"logps/chosen": -474.0, |
|
"logps/rejected": -256.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.46875, |
|
"rewards/margins": 18.5, |
|
"rewards/rejected": -11.9375, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.5849358974358975, |
|
"grad_norm": 3.5960258938370916e-05, |
|
"learning_rate": 4.473137429504304e-07, |
|
"logits/chosen": -0.2734375, |
|
"logits/rejected": -0.03466796875, |
|
"logps/chosen": -456.0, |
|
"logps/rejected": -278.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.25, |
|
"rewards/margins": 18.125, |
|
"rewards/rejected": -11.875, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.592948717948718, |
|
"grad_norm": 0.0003640457384025675, |
|
"learning_rate": 4.458296230335411e-07, |
|
"logits/chosen": -0.294921875, |
|
"logits/rejected": -0.02880859375, |
|
"logps/chosen": -478.0, |
|
"logps/rejected": -258.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.0625, |
|
"rewards/margins": 19.125, |
|
"rewards/rejected": -12.0625, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.6009615384615384, |
|
"grad_norm": 9.659235251348069e-05, |
|
"learning_rate": 4.443455031166518e-07, |
|
"logits/chosen": -0.1689453125, |
|
"logits/rejected": 0.07666015625, |
|
"logps/chosen": -454.0, |
|
"logps/rejected": -270.0, |
|
"loss": 0.0114, |
|
"rewards/accuracies": 0.9750000238418579, |
|
"rewards/chosen": 6.21875, |
|
"rewards/margins": 18.25, |
|
"rewards/rejected": -11.9375, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.6089743589743589, |
|
"grad_norm": 4.85581499023554e-06, |
|
"learning_rate": 4.4286138319976253e-07, |
|
"logits/chosen": -0.240234375, |
|
"logits/rejected": -0.126953125, |
|
"logps/chosen": -428.0, |
|
"logps/rejected": -266.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.21875, |
|
"rewards/margins": 18.5, |
|
"rewards/rejected": -12.25, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.6169871794871795, |
|
"grad_norm": 0.0007816450843854167, |
|
"learning_rate": 4.4137726328287324e-07, |
|
"logits/chosen": -0.271484375, |
|
"logits/rejected": 0.142578125, |
|
"logps/chosen": -434.0, |
|
"logps/rejected": -262.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.0, |
|
"rewards/margins": 18.75, |
|
"rewards/rejected": -11.75, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.625, |
|
"grad_norm": 0.014442808037563768, |
|
"learning_rate": 4.3989314336598395e-07, |
|
"logits/chosen": -0.09716796875, |
|
"logits/rejected": -0.09423828125, |
|
"logps/chosen": -422.0, |
|
"logps/rejected": -252.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 5.96875, |
|
"rewards/margins": 17.75, |
|
"rewards/rejected": -11.75, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.6330128205128205, |
|
"grad_norm": 0.00014068964421634487, |
|
"learning_rate": 4.384090234490946e-07, |
|
"logits/chosen": -0.10888671875, |
|
"logits/rejected": 0.03662109375, |
|
"logps/chosen": -458.0, |
|
"logps/rejected": -246.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.40625, |
|
"rewards/margins": 18.625, |
|
"rewards/rejected": -11.1875, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.6410256410256411, |
|
"grad_norm": 8.631309050994593e-05, |
|
"learning_rate": 4.369249035322054e-07, |
|
"logits/chosen": -0.15625, |
|
"logits/rejected": 0.11083984375, |
|
"logps/chosen": -456.0, |
|
"logps/rejected": -266.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.90625, |
|
"rewards/margins": 19.5, |
|
"rewards/rejected": -12.5625, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.6490384615384616, |
|
"grad_norm": 2.3946547921903765e-05, |
|
"learning_rate": 4.354407836153161e-07, |
|
"logits/chosen": -0.314453125, |
|
"logits/rejected": 0.0966796875, |
|
"logps/chosen": -454.0, |
|
"logps/rejected": -278.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.65625, |
|
"rewards/margins": 18.875, |
|
"rewards/rejected": -12.25, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.657051282051282, |
|
"grad_norm": 0.0007163877343597987, |
|
"learning_rate": 4.339566636984268e-07, |
|
"logits/chosen": -0.2001953125, |
|
"logits/rejected": 0.0133056640625, |
|
"logps/chosen": -472.0, |
|
"logps/rejected": -254.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.875, |
|
"rewards/margins": 18.75, |
|
"rewards/rejected": -11.8125, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.6650641025641025, |
|
"grad_norm": 0.001989418234757288, |
|
"learning_rate": 4.324725437815375e-07, |
|
"logits/chosen": -0.0167236328125, |
|
"logits/rejected": 0.1552734375, |
|
"logps/chosen": -432.0, |
|
"logps/rejected": -260.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.0625, |
|
"rewards/margins": 17.75, |
|
"rewards/rejected": -11.625, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.6730769230769231, |
|
"grad_norm": 7.672186675245888e-05, |
|
"learning_rate": 4.3098842386464824e-07, |
|
"logits/chosen": -0.357421875, |
|
"logits/rejected": 0.1416015625, |
|
"logps/chosen": -450.0, |
|
"logps/rejected": -276.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.4375, |
|
"rewards/margins": 19.125, |
|
"rewards/rejected": -12.6875, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.6810897435897436, |
|
"grad_norm": 0.002765457862822168, |
|
"learning_rate": 4.2950430394775895e-07, |
|
"logits/chosen": -0.220703125, |
|
"logits/rejected": 0.056884765625, |
|
"logps/chosen": -432.0, |
|
"logps/rejected": -260.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.0, |
|
"rewards/margins": 19.125, |
|
"rewards/rejected": -12.125, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.6891025641025641, |
|
"grad_norm": 0.00010976578527505251, |
|
"learning_rate": 4.280201840308697e-07, |
|
"logits/chosen": -0.07763671875, |
|
"logits/rejected": 0.28515625, |
|
"logps/chosen": -478.0, |
|
"logps/rejected": -256.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.0625, |
|
"rewards/margins": 19.375, |
|
"rewards/rejected": -12.3125, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.6971153846153846, |
|
"grad_norm": 4.134983141010821e-06, |
|
"learning_rate": 4.2653606411398043e-07, |
|
"logits/chosen": -0.1826171875, |
|
"logits/rejected": 0.042724609375, |
|
"logps/chosen": -440.0, |
|
"logps/rejected": -268.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.9375, |
|
"rewards/margins": 19.625, |
|
"rewards/rejected": -12.6875, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.7051282051282052, |
|
"grad_norm": 9.136458698336654e-06, |
|
"learning_rate": 4.2505194419709114e-07, |
|
"logits/chosen": -0.2158203125, |
|
"logits/rejected": -0.12890625, |
|
"logps/chosen": -422.0, |
|
"logps/rejected": -266.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.125, |
|
"rewards/margins": 18.75, |
|
"rewards/rejected": -12.6875, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.7131410256410257, |
|
"grad_norm": 0.007426291357664032, |
|
"learning_rate": 4.235678242802018e-07, |
|
"logits/chosen": -0.2109375, |
|
"logits/rejected": 0.0517578125, |
|
"logps/chosen": -464.0, |
|
"logps/rejected": -251.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.4375, |
|
"rewards/margins": 18.625, |
|
"rewards/rejected": -12.125, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.7211538461538461, |
|
"grad_norm": 3.18757387463839e-06, |
|
"learning_rate": 4.220837043633125e-07, |
|
"logits/chosen": -0.099609375, |
|
"logits/rejected": 0.23046875, |
|
"logps/chosen": -476.0, |
|
"logps/rejected": -268.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.09375, |
|
"rewards/margins": 21.0, |
|
"rewards/rejected": -13.875, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.7291666666666666, |
|
"grad_norm": 0.0027727730490827414, |
|
"learning_rate": 4.2059958444642323e-07, |
|
"logits/chosen": -0.049072265625, |
|
"logits/rejected": 0.11962890625, |
|
"logps/chosen": -444.0, |
|
"logps/rejected": -284.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 5.6875, |
|
"rewards/margins": 18.875, |
|
"rewards/rejected": -13.25, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.7371794871794872, |
|
"grad_norm": 3.5382296968371214e-05, |
|
"learning_rate": 4.1911546452953394e-07, |
|
"logits/chosen": -0.296875, |
|
"logits/rejected": -0.01312255859375, |
|
"logps/chosen": -446.0, |
|
"logps/rejected": -284.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.75, |
|
"rewards/margins": 20.75, |
|
"rewards/rejected": -13.9375, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.7451923076923077, |
|
"grad_norm": 5.85015713116871e-05, |
|
"learning_rate": 4.176313446126447e-07, |
|
"logits/chosen": -0.33984375, |
|
"logits/rejected": 0.04443359375, |
|
"logps/chosen": -432.0, |
|
"logps/rejected": -282.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.375, |
|
"rewards/margins": 20.0, |
|
"rewards/rejected": -13.625, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.7532051282051282, |
|
"grad_norm": 0.0016623747778029196, |
|
"learning_rate": 4.161472246957554e-07, |
|
"logits/chosen": -0.251953125, |
|
"logits/rejected": 0.09521484375, |
|
"logps/chosen": -460.0, |
|
"logps/rejected": -270.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.84375, |
|
"rewards/margins": 20.875, |
|
"rewards/rejected": -14.0, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.7612179487179487, |
|
"grad_norm": 1.321177641149776e-05, |
|
"learning_rate": 4.1466310477886614e-07, |
|
"logits/chosen": -0.1533203125, |
|
"logits/rejected": -0.16796875, |
|
"logps/chosen": -432.0, |
|
"logps/rejected": -276.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.25, |
|
"rewards/margins": 19.625, |
|
"rewards/rejected": -13.375, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.7692307692307693, |
|
"grad_norm": 0.0076723556464482374, |
|
"learning_rate": 4.1317898486197685e-07, |
|
"logits/chosen": -0.19921875, |
|
"logits/rejected": 0.027099609375, |
|
"logps/chosen": -466.0, |
|
"logps/rejected": -276.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.75, |
|
"rewards/margins": 20.5, |
|
"rewards/rejected": -13.75, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.7772435897435898, |
|
"grad_norm": 4.9003066953656295e-05, |
|
"learning_rate": 4.1169486494508756e-07, |
|
"logits/chosen": -0.283203125, |
|
"logits/rejected": 0.031494140625, |
|
"logps/chosen": -434.0, |
|
"logps/rejected": -266.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.96875, |
|
"rewards/margins": 20.25, |
|
"rewards/rejected": -13.3125, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.7852564102564102, |
|
"grad_norm": 2.7494943569823703e-05, |
|
"learning_rate": 4.102107450281982e-07, |
|
"logits/chosen": -0.28515625, |
|
"logits/rejected": -0.01446533203125, |
|
"logps/chosen": -452.0, |
|
"logps/rejected": -280.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.0625, |
|
"rewards/margins": 19.5, |
|
"rewards/rejected": -13.375, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.7932692307692307, |
|
"grad_norm": 1.4548110920381792e-05, |
|
"learning_rate": 4.0872662511130894e-07, |
|
"logits/chosen": -0.1611328125, |
|
"logits/rejected": 0.1318359375, |
|
"logps/chosen": -436.0, |
|
"logps/rejected": -304.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 5.875, |
|
"rewards/margins": 21.125, |
|
"rewards/rejected": -15.1875, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.8012820512820513, |
|
"grad_norm": 0.0015914623500980674, |
|
"learning_rate": 4.072425051944197e-07, |
|
"logits/chosen": -0.408203125, |
|
"logits/rejected": -0.0390625, |
|
"logps/chosen": -460.0, |
|
"logps/rejected": -268.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.1875, |
|
"rewards/margins": 20.75, |
|
"rewards/rejected": -13.5625, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.8092948717948718, |
|
"grad_norm": 8.626603898034529e-06, |
|
"learning_rate": 4.057583852775304e-07, |
|
"logits/chosen": -0.050048828125, |
|
"logits/rejected": 0.0703125, |
|
"logps/chosen": -448.0, |
|
"logps/rejected": -282.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.9375, |
|
"rewards/margins": 21.375, |
|
"rewards/rejected": -14.4375, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.8173076923076923, |
|
"grad_norm": 7.8538196292969e-07, |
|
"learning_rate": 4.0427426536064113e-07, |
|
"logits/chosen": -0.310546875, |
|
"logits/rejected": 0.044677734375, |
|
"logps/chosen": -434.0, |
|
"logps/rejected": -284.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.53125, |
|
"rewards/margins": 20.75, |
|
"rewards/rejected": -14.125, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.8253205128205128, |
|
"grad_norm": 1.4582781033888853e-05, |
|
"learning_rate": 4.0279014544375184e-07, |
|
"logits/chosen": -0.197265625, |
|
"logits/rejected": 0.216796875, |
|
"logps/chosen": -458.0, |
|
"logps/rejected": -292.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.125, |
|
"rewards/margins": 21.0, |
|
"rewards/rejected": -13.8125, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.8333333333333334, |
|
"grad_norm": 0.004174571740946655, |
|
"learning_rate": 4.0130602552686256e-07, |
|
"logits/chosen": -0.27734375, |
|
"logits/rejected": 0.16796875, |
|
"logps/chosen": -460.0, |
|
"logps/rejected": -292.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.21875, |
|
"rewards/margins": 21.0, |
|
"rewards/rejected": -14.75, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.8413461538461539, |
|
"grad_norm": 0.0006261120220267343, |
|
"learning_rate": 3.9982190560997327e-07, |
|
"logits/chosen": -0.2490234375, |
|
"logits/rejected": 0.04296875, |
|
"logps/chosen": -452.0, |
|
"logps/rejected": -290.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.71875, |
|
"rewards/margins": 20.75, |
|
"rewards/rejected": -14.125, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.8493589743589743, |
|
"grad_norm": 1.8893277312540628e-06, |
|
"learning_rate": 3.98337785693084e-07, |
|
"logits/chosen": -0.322265625, |
|
"logits/rejected": 0.10595703125, |
|
"logps/chosen": -446.0, |
|
"logps/rejected": -290.0, |
|
"loss": 0.0001, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 5.90625, |
|
"rewards/margins": 20.25, |
|
"rewards/rejected": -14.3125, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.8573717948717948, |
|
"grad_norm": 2.2415269739272855e-06, |
|
"learning_rate": 3.968536657761947e-07, |
|
"logits/chosen": -0.27734375, |
|
"logits/rejected": -0.06494140625, |
|
"logps/chosen": -448.0, |
|
"logps/rejected": -304.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.65625, |
|
"rewards/margins": 22.0, |
|
"rewards/rejected": -15.375, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.8653846153846154, |
|
"grad_norm": 4.557694360128382e-05, |
|
"learning_rate": 3.953695458593054e-07, |
|
"logits/chosen": -0.2451171875, |
|
"logits/rejected": 0.310546875, |
|
"logps/chosen": -458.0, |
|
"logps/rejected": -292.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.46875, |
|
"rewards/margins": 20.625, |
|
"rewards/rejected": -14.125, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.8733974358974359, |
|
"grad_norm": 6.890981446368462e-08, |
|
"learning_rate": 3.938854259424161e-07, |
|
"logits/chosen": -0.08544921875, |
|
"logits/rejected": 0.134765625, |
|
"logps/chosen": -446.0, |
|
"logps/rejected": -288.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.9375, |
|
"rewards/margins": 21.75, |
|
"rewards/rejected": -14.8125, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.8814102564102564, |
|
"grad_norm": 0.0005901371550357143, |
|
"learning_rate": 3.9240130602552684e-07, |
|
"logits/chosen": -0.208984375, |
|
"logits/rejected": -0.0157470703125, |
|
"logps/chosen": -454.0, |
|
"logps/rejected": -290.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.84375, |
|
"rewards/margins": 21.25, |
|
"rewards/rejected": -14.5, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.8894230769230769, |
|
"grad_norm": 9.373688913667137e-06, |
|
"learning_rate": 3.9091718610863755e-07, |
|
"logits/chosen": -0.259765625, |
|
"logits/rejected": 0.052734375, |
|
"logps/chosen": -454.0, |
|
"logps/rejected": -304.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 5.46875, |
|
"rewards/margins": 20.0, |
|
"rewards/rejected": -14.5625, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.8974358974358975, |
|
"grad_norm": 0.0006236816278666287, |
|
"learning_rate": 3.8943306619174827e-07, |
|
"logits/chosen": -0.224609375, |
|
"logits/rejected": -0.0703125, |
|
"logps/chosen": -440.0, |
|
"logps/rejected": -282.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 5.78125, |
|
"rewards/margins": 20.375, |
|
"rewards/rejected": -14.5625, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.905448717948718, |
|
"grad_norm": 7.77967518016113e-05, |
|
"learning_rate": 3.87948946274859e-07, |
|
"logits/chosen": -0.2021484375, |
|
"logits/rejected": -0.034423828125, |
|
"logps/chosen": -434.0, |
|
"logps/rejected": -300.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.3125, |
|
"rewards/margins": 22.0, |
|
"rewards/rejected": -15.75, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.9134615384615384, |
|
"grad_norm": 0.0007361294866117807, |
|
"learning_rate": 3.8646482635796975e-07, |
|
"logits/chosen": -0.087890625, |
|
"logits/rejected": 0.053466796875, |
|
"logps/chosen": -420.0, |
|
"logps/rejected": -286.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.8125, |
|
"rewards/margins": 21.25, |
|
"rewards/rejected": -14.5, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.9214743589743589, |
|
"grad_norm": 0.03444107459874257, |
|
"learning_rate": 3.8498070644108046e-07, |
|
"logits/chosen": -0.07861328125, |
|
"logits/rejected": 0.2421875, |
|
"logps/chosen": -426.0, |
|
"logps/rejected": -282.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.21875, |
|
"rewards/margins": 22.125, |
|
"rewards/rejected": -14.9375, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.9294871794871795, |
|
"grad_norm": 2.883961656451469e-07, |
|
"learning_rate": 3.834965865241911e-07, |
|
"logits/chosen": -0.287109375, |
|
"logits/rejected": 0.306640625, |
|
"logps/chosen": -446.0, |
|
"logps/rejected": -300.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.03125, |
|
"rewards/margins": 21.5, |
|
"rewards/rejected": -15.5, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.9375, |
|
"grad_norm": 1.7356563406622823e-05, |
|
"learning_rate": 3.8201246660730183e-07, |
|
"logits/chosen": -0.1064453125, |
|
"logits/rejected": 0.1640625, |
|
"logps/chosen": -482.0, |
|
"logps/rejected": -290.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.0625, |
|
"rewards/margins": 22.25, |
|
"rewards/rejected": -15.125, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.9455128205128205, |
|
"grad_norm": 1.268242874979815e-05, |
|
"learning_rate": 3.8052834669041255e-07, |
|
"logits/chosen": -0.3671875, |
|
"logits/rejected": -0.0220947265625, |
|
"logps/chosen": -440.0, |
|
"logps/rejected": -292.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.0, |
|
"rewards/margins": 20.625, |
|
"rewards/rejected": -14.625, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.9535256410256411, |
|
"grad_norm": 0.00012607759481833192, |
|
"learning_rate": 3.7904422677352326e-07, |
|
"logits/chosen": -0.1298828125, |
|
"logits/rejected": 0.2412109375, |
|
"logps/chosen": -430.0, |
|
"logps/rejected": -296.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.71875, |
|
"rewards/margins": 22.125, |
|
"rewards/rejected": -15.4375, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.9615384615384616, |
|
"grad_norm": 1.313583578202776e-06, |
|
"learning_rate": 3.77560106856634e-07, |
|
"logits/chosen": -0.1708984375, |
|
"logits/rejected": -0.1220703125, |
|
"logps/chosen": -450.0, |
|
"logps/rejected": -278.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.5, |
|
"rewards/margins": 21.125, |
|
"rewards/rejected": -14.5625, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.969551282051282, |
|
"grad_norm": 5.541906322587775e-06, |
|
"learning_rate": 3.7607598693974474e-07, |
|
"logits/chosen": -0.32421875, |
|
"logits/rejected": 0.0211181640625, |
|
"logps/chosen": -456.0, |
|
"logps/rejected": -290.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.375, |
|
"rewards/margins": 22.875, |
|
"rewards/rejected": -15.5, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.9775641025641025, |
|
"grad_norm": 0.0009580517831776949, |
|
"learning_rate": 3.7459186702285545e-07, |
|
"logits/chosen": -0.384765625, |
|
"logits/rejected": -0.134765625, |
|
"logps/chosen": -452.0, |
|
"logps/rejected": -294.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.28125, |
|
"rewards/margins": 22.0, |
|
"rewards/rejected": -14.75, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.9855769230769231, |
|
"grad_norm": 1.1763851350595596e-06, |
|
"learning_rate": 3.7310774710596617e-07, |
|
"logits/chosen": -0.158203125, |
|
"logits/rejected": 0.1630859375, |
|
"logps/chosen": -420.0, |
|
"logps/rejected": -298.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.625, |
|
"rewards/margins": 22.75, |
|
"rewards/rejected": -16.125, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.9935897435897436, |
|
"grad_norm": 5.444678378288628e-07, |
|
"learning_rate": 3.716236271890769e-07, |
|
"logits/chosen": -0.302734375, |
|
"logits/rejected": 0.0247802734375, |
|
"logps/chosen": -430.0, |
|
"logps/rejected": -290.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.59375, |
|
"rewards/margins": 22.5, |
|
"rewards/rejected": -15.875, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_logits/chosen": -0.185546875, |
|
"eval_logits/rejected": 0.047119140625, |
|
"eval_logps/chosen": -440.0, |
|
"eval_logps/rejected": -298.0, |
|
"eval_loss": 8.373256719096389e-07, |
|
"eval_rewards/accuracies": 1.0, |
|
"eval_rewards/chosen": 6.53125, |
|
"eval_rewards/margins": 22.0, |
|
"eval_rewards/rejected": -15.375, |
|
"eval_runtime": 25.6054, |
|
"eval_samples_per_second": 7.772, |
|
"eval_steps_per_second": 0.976, |
|
"step": 1248 |
|
}, |
|
{ |
|
"epoch": 1.001602564102564, |
|
"grad_norm": 8.025230375602095e-07, |
|
"learning_rate": 3.7013950727218754e-07, |
|
"logits/chosen": -0.1416015625, |
|
"logits/rejected": 0.2578125, |
|
"logps/chosen": -456.0, |
|
"logps/rejected": -292.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.4375, |
|
"rewards/margins": 21.375, |
|
"rewards/rejected": -15.0, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 1.0096153846153846, |
|
"grad_norm": 2.98858297335877e-07, |
|
"learning_rate": 3.6865538735529826e-07, |
|
"logits/chosen": -0.1318359375, |
|
"logits/rejected": 0.1279296875, |
|
"logps/chosen": -440.0, |
|
"logps/rejected": -294.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.96875, |
|
"rewards/margins": 22.25, |
|
"rewards/rejected": -15.375, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 1.017628205128205, |
|
"grad_norm": 5.973383784003284e-07, |
|
"learning_rate": 3.6717126743840897e-07, |
|
"logits/chosen": -0.10888671875, |
|
"logits/rejected": -0.1533203125, |
|
"logps/chosen": -452.0, |
|
"logps/rejected": -282.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.21875, |
|
"rewards/margins": 22.375, |
|
"rewards/rejected": -15.1875, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 1.0256410256410255, |
|
"grad_norm": 7.232942551732893e-06, |
|
"learning_rate": 3.6568714752151974e-07, |
|
"logits/chosen": -0.140625, |
|
"logits/rejected": 0.03515625, |
|
"logps/chosen": -404.0, |
|
"logps/rejected": -298.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.9375, |
|
"rewards/margins": 22.75, |
|
"rewards/rejected": -15.8125, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 1.0336538461538463, |
|
"grad_norm": 2.3129696720447785e-05, |
|
"learning_rate": 3.6420302760463045e-07, |
|
"logits/chosen": -0.353515625, |
|
"logits/rejected": -0.1630859375, |
|
"logps/chosen": -450.0, |
|
"logps/rejected": -306.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.09375, |
|
"rewards/margins": 24.25, |
|
"rewards/rejected": -17.125, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 1.0416666666666667, |
|
"grad_norm": 1.2196659967369379e-05, |
|
"learning_rate": 3.6271890768774116e-07, |
|
"logits/chosen": -0.216796875, |
|
"logits/rejected": 0.0267333984375, |
|
"logps/chosen": -432.0, |
|
"logps/rejected": -304.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.25, |
|
"rewards/margins": 23.125, |
|
"rewards/rejected": -15.875, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.0496794871794872, |
|
"grad_norm": 1.563732659184636e-05, |
|
"learning_rate": 3.612347877708519e-07, |
|
"logits/chosen": -0.00146484375, |
|
"logits/rejected": 0.384765625, |
|
"logps/chosen": -424.0, |
|
"logps/rejected": -304.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.75, |
|
"rewards/margins": 23.625, |
|
"rewards/rejected": -16.875, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 1.0576923076923077, |
|
"grad_norm": 8.512967205118398e-05, |
|
"learning_rate": 3.597506678539626e-07, |
|
"logits/chosen": -0.322265625, |
|
"logits/rejected": -0.0107421875, |
|
"logps/chosen": -452.0, |
|
"logps/rejected": -308.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.71875, |
|
"rewards/margins": 23.125, |
|
"rewards/rejected": -16.375, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 1.0657051282051282, |
|
"grad_norm": 4.938743218386723e-07, |
|
"learning_rate": 3.582665479370733e-07, |
|
"logits/chosen": -0.255859375, |
|
"logits/rejected": 0.11572265625, |
|
"logps/chosen": -428.0, |
|
"logps/rejected": -300.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.75, |
|
"rewards/margins": 23.0, |
|
"rewards/rejected": -16.25, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 1.0737179487179487, |
|
"grad_norm": 2.718607594817426e-07, |
|
"learning_rate": 3.5678242802018396e-07, |
|
"logits/chosen": -0.59375, |
|
"logits/rejected": -0.0849609375, |
|
"logps/chosen": -488.0, |
|
"logps/rejected": -312.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.4375, |
|
"rewards/margins": 24.25, |
|
"rewards/rejected": -16.875, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 1.0817307692307692, |
|
"grad_norm": 1.8965126469314772e-07, |
|
"learning_rate": 3.5529830810329473e-07, |
|
"logits/chosen": -0.361328125, |
|
"logits/rejected": -0.046142578125, |
|
"logps/chosen": -468.0, |
|
"logps/rejected": -308.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.3125, |
|
"rewards/margins": 23.25, |
|
"rewards/rejected": -15.9375, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 1.0897435897435896, |
|
"grad_norm": 9.791390801402978e-06, |
|
"learning_rate": 3.5381418818640544e-07, |
|
"logits/chosen": -0.2412109375, |
|
"logits/rejected": -0.0517578125, |
|
"logps/chosen": -450.0, |
|
"logps/rejected": -316.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.125, |
|
"rewards/margins": 24.0, |
|
"rewards/rejected": -16.875, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 1.0977564102564104, |
|
"grad_norm": 2.7688965444114683e-07, |
|
"learning_rate": 3.5233006826951616e-07, |
|
"logits/chosen": -0.412109375, |
|
"logits/rejected": 0.06298828125, |
|
"logps/chosen": -436.0, |
|
"logps/rejected": -322.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.09375, |
|
"rewards/margins": 23.0, |
|
"rewards/rejected": -16.875, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 1.1057692307692308, |
|
"grad_norm": 0.0002875733562139178, |
|
"learning_rate": 3.5084594835262687e-07, |
|
"logits/chosen": -0.162109375, |
|
"logits/rejected": 0.12890625, |
|
"logps/chosen": -458.0, |
|
"logps/rejected": -292.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.84375, |
|
"rewards/margins": 23.25, |
|
"rewards/rejected": -16.375, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 1.1137820512820513, |
|
"grad_norm": 8.827404368665739e-05, |
|
"learning_rate": 3.493618284357376e-07, |
|
"logits/chosen": -0.24609375, |
|
"logits/rejected": -0.05322265625, |
|
"logps/chosen": -438.0, |
|
"logps/rejected": -302.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.5625, |
|
"rewards/margins": 23.0, |
|
"rewards/rejected": -16.5, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 1.1217948717948718, |
|
"grad_norm": 5.104603120766749e-07, |
|
"learning_rate": 3.478777085188483e-07, |
|
"logits/chosen": -0.44921875, |
|
"logits/rejected": 0.1083984375, |
|
"logps/chosen": -468.0, |
|
"logps/rejected": -314.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.90625, |
|
"rewards/margins": 24.5, |
|
"rewards/rejected": -17.5, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.1298076923076923, |
|
"grad_norm": 6.98203369126403e-08, |
|
"learning_rate": 3.46393588601959e-07, |
|
"logits/chosen": -0.0233154296875, |
|
"logits/rejected": 0.2265625, |
|
"logps/chosen": -438.0, |
|
"logps/rejected": -304.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.9375, |
|
"rewards/margins": 23.75, |
|
"rewards/rejected": -16.875, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 1.1378205128205128, |
|
"grad_norm": 3.964511109230341e-08, |
|
"learning_rate": 3.449094686850698e-07, |
|
"logits/chosen": -0.1025390625, |
|
"logits/rejected": -0.00738525390625, |
|
"logps/chosen": -432.0, |
|
"logps/rejected": -306.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.65625, |
|
"rewards/margins": 23.625, |
|
"rewards/rejected": -17.0, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 1.1458333333333333, |
|
"grad_norm": 4.514618927821668e-06, |
|
"learning_rate": 3.434253487681805e-07, |
|
"logits/chosen": -0.29296875, |
|
"logits/rejected": 0.1181640625, |
|
"logps/chosen": -446.0, |
|
"logps/rejected": -308.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.34375, |
|
"rewards/margins": 24.25, |
|
"rewards/rejected": -16.875, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 1.1538461538461537, |
|
"grad_norm": 8.477305344127751e-05, |
|
"learning_rate": 3.4194122885129115e-07, |
|
"logits/chosen": -0.1396484375, |
|
"logits/rejected": 0.0181884765625, |
|
"logps/chosen": -444.0, |
|
"logps/rejected": -300.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.78125, |
|
"rewards/margins": 24.0, |
|
"rewards/rejected": -16.25, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 1.1618589743589745, |
|
"grad_norm": 1.14928275960499e-08, |
|
"learning_rate": 3.4045710893440187e-07, |
|
"logits/chosen": -0.14453125, |
|
"logits/rejected": 0.0341796875, |
|
"logps/chosen": -450.0, |
|
"logps/rejected": -290.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.5, |
|
"rewards/margins": 23.625, |
|
"rewards/rejected": -16.125, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 1.169871794871795, |
|
"grad_norm": 7.535912943119137e-08, |
|
"learning_rate": 3.389729890175126e-07, |
|
"logits/chosen": -0.287109375, |
|
"logits/rejected": 0.036376953125, |
|
"logps/chosen": -424.0, |
|
"logps/rejected": -326.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.8125, |
|
"rewards/margins": 24.375, |
|
"rewards/rejected": -17.625, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 1.1778846153846154, |
|
"grad_norm": 5.184146998969776e-06, |
|
"learning_rate": 3.374888691006233e-07, |
|
"logits/chosen": -0.11279296875, |
|
"logits/rejected": 0.171875, |
|
"logps/chosen": -414.0, |
|
"logps/rejected": -324.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.65625, |
|
"rewards/margins": 24.125, |
|
"rewards/rejected": -17.5, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 1.185897435897436, |
|
"grad_norm": 7.433000704694052e-08, |
|
"learning_rate": 3.36004749183734e-07, |
|
"logits/chosen": -0.376953125, |
|
"logits/rejected": 0.10546875, |
|
"logps/chosen": -460.0, |
|
"logps/rejected": -320.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.125, |
|
"rewards/margins": 24.875, |
|
"rewards/rejected": -17.75, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 1.1939102564102564, |
|
"grad_norm": 0.0001254253838129237, |
|
"learning_rate": 3.3452062926684477e-07, |
|
"logits/chosen": -0.353515625, |
|
"logits/rejected": 0.05859375, |
|
"logps/chosen": -442.0, |
|
"logps/rejected": -296.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.09375, |
|
"rewards/margins": 23.75, |
|
"rewards/rejected": -16.75, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 1.2019230769230769, |
|
"grad_norm": 6.90260882071762e-07, |
|
"learning_rate": 3.330365093499555e-07, |
|
"logits/chosen": -0.224609375, |
|
"logits/rejected": 0.06787109375, |
|
"logps/chosen": -426.0, |
|
"logps/rejected": -312.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.46875, |
|
"rewards/margins": 24.0, |
|
"rewards/rejected": -17.625, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.2099358974358974, |
|
"grad_norm": 4.934846926607364e-08, |
|
"learning_rate": 3.315523894330662e-07, |
|
"logits/chosen": -0.03955078125, |
|
"logits/rejected": 0.17578125, |
|
"logps/chosen": -444.0, |
|
"logps/rejected": -314.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.375, |
|
"rewards/margins": 25.125, |
|
"rewards/rejected": -17.75, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 1.217948717948718, |
|
"grad_norm": 9.020471497206749e-08, |
|
"learning_rate": 3.300682695161769e-07, |
|
"logits/chosen": -0.1806640625, |
|
"logits/rejected": 0.212890625, |
|
"logps/chosen": -416.0, |
|
"logps/rejected": -304.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.15625, |
|
"rewards/margins": 24.0, |
|
"rewards/rejected": -16.875, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 1.2259615384615385, |
|
"grad_norm": 5.944028084695953e-08, |
|
"learning_rate": 3.2858414959928757e-07, |
|
"logits/chosen": -0.0177001953125, |
|
"logits/rejected": -0.00482177734375, |
|
"logps/chosen": -412.0, |
|
"logps/rejected": -310.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.6875, |
|
"rewards/margins": 23.5, |
|
"rewards/rejected": -16.875, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 1.233974358974359, |
|
"grad_norm": 1.3278023407952154e-05, |
|
"learning_rate": 3.271000296823983e-07, |
|
"logits/chosen": -0.42578125, |
|
"logits/rejected": -0.0810546875, |
|
"logps/chosen": -480.0, |
|
"logps/rejected": -306.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.59375, |
|
"rewards/margins": 23.5, |
|
"rewards/rejected": -17.0, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 1.2419871794871795, |
|
"grad_norm": 1.8738770850648627e-07, |
|
"learning_rate": 3.25615909765509e-07, |
|
"logits/chosen": -0.19140625, |
|
"logits/rejected": -0.1708984375, |
|
"logps/chosen": -422.0, |
|
"logps/rejected": -320.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.5625, |
|
"rewards/margins": 24.375, |
|
"rewards/rejected": -17.875, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 0.0010963655925342855, |
|
"learning_rate": 3.2413178984861977e-07, |
|
"logits/chosen": -0.3125, |
|
"logits/rejected": 0.0693359375, |
|
"logps/chosen": -448.0, |
|
"logps/rejected": -306.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.75, |
|
"rewards/margins": 23.75, |
|
"rewards/rejected": -17.0, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 1.2580128205128205, |
|
"grad_norm": 8.314207000700108e-08, |
|
"learning_rate": 3.226476699317305e-07, |
|
"logits/chosen": -0.1455078125, |
|
"logits/rejected": -0.18359375, |
|
"logps/chosen": -446.0, |
|
"logps/rejected": -316.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.8125, |
|
"rewards/margins": 24.375, |
|
"rewards/rejected": -17.5, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 1.266025641025641, |
|
"grad_norm": 1.6908890464582575e-05, |
|
"learning_rate": 3.211635500148412e-07, |
|
"logits/chosen": -0.1650390625, |
|
"logits/rejected": 0.0712890625, |
|
"logps/chosen": -428.0, |
|
"logps/rejected": -300.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.0625, |
|
"rewards/margins": 23.625, |
|
"rewards/rejected": -16.5, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 1.2740384615384617, |
|
"grad_norm": 2.9193436149794298e-08, |
|
"learning_rate": 3.196794300979519e-07, |
|
"logits/chosen": -0.1669921875, |
|
"logits/rejected": 0.1171875, |
|
"logps/chosen": -412.0, |
|
"logps/rejected": -316.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.65625, |
|
"rewards/margins": 23.625, |
|
"rewards/rejected": -17.0, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 1.282051282051282, |
|
"grad_norm": 2.0730858792937314e-08, |
|
"learning_rate": 3.181953101810626e-07, |
|
"logits/chosen": -0.10009765625, |
|
"logits/rejected": 0.05908203125, |
|
"logps/chosen": -446.0, |
|
"logps/rejected": -318.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.28125, |
|
"rewards/margins": 25.25, |
|
"rewards/rejected": -18.0, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.2900641025641026, |
|
"grad_norm": 1.2148300703017007e-06, |
|
"learning_rate": 3.1671119026417333e-07, |
|
"logits/chosen": -0.251953125, |
|
"logits/rejected": 0.037841796875, |
|
"logps/chosen": -436.0, |
|
"logps/rejected": -316.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.34375, |
|
"rewards/margins": 24.5, |
|
"rewards/rejected": -17.25, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 1.2980769230769231, |
|
"grad_norm": 3.674133497219966e-05, |
|
"learning_rate": 3.15227070347284e-07, |
|
"logits/chosen": -0.33203125, |
|
"logits/rejected": 0.04833984375, |
|
"logps/chosen": -450.0, |
|
"logps/rejected": -302.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.78125, |
|
"rewards/margins": 24.0, |
|
"rewards/rejected": -17.25, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 1.3060897435897436, |
|
"grad_norm": 3.997496159095626e-06, |
|
"learning_rate": 3.1374295043039476e-07, |
|
"logits/chosen": -0.1005859375, |
|
"logits/rejected": 0.1103515625, |
|
"logps/chosen": -410.0, |
|
"logps/rejected": -328.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.15625, |
|
"rewards/margins": 24.375, |
|
"rewards/rejected": -18.25, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 1.314102564102564, |
|
"grad_norm": 6.582700174080238e-07, |
|
"learning_rate": 3.122588305135055e-07, |
|
"logits/chosen": -0.263671875, |
|
"logits/rejected": -0.11474609375, |
|
"logps/chosen": -442.0, |
|
"logps/rejected": -306.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 5.90625, |
|
"rewards/margins": 23.25, |
|
"rewards/rejected": -17.375, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 1.3221153846153846, |
|
"grad_norm": 2.5302263716994303e-07, |
|
"learning_rate": 3.107747105966162e-07, |
|
"logits/chosen": -0.1640625, |
|
"logits/rejected": -0.021484375, |
|
"logps/chosen": -436.0, |
|
"logps/rejected": -318.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.46875, |
|
"rewards/margins": 24.375, |
|
"rewards/rejected": -17.875, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 1.330128205128205, |
|
"grad_norm": 2.4645003699404177e-07, |
|
"learning_rate": 3.092905906797269e-07, |
|
"logits/chosen": -0.1396484375, |
|
"logits/rejected": 0.12255859375, |
|
"logps/chosen": -430.0, |
|
"logps/rejected": -324.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.3125, |
|
"rewards/margins": 25.625, |
|
"rewards/rejected": -18.25, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 1.3381410256410255, |
|
"grad_norm": 3.576309832920555e-08, |
|
"learning_rate": 3.078064707628376e-07, |
|
"logits/chosen": -0.271484375, |
|
"logits/rejected": -0.04052734375, |
|
"logps/chosen": -430.0, |
|
"logps/rejected": -314.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.9375, |
|
"rewards/margins": 24.625, |
|
"rewards/rejected": -17.75, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 1.3461538461538463, |
|
"grad_norm": 5.918843540732376e-07, |
|
"learning_rate": 3.0632235084594833e-07, |
|
"logits/chosen": -0.4140625, |
|
"logits/rejected": 0.08642578125, |
|
"logps/chosen": -458.0, |
|
"logps/rejected": -318.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.8125, |
|
"rewards/margins": 25.375, |
|
"rewards/rejected": -17.5, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 1.3541666666666667, |
|
"grad_norm": 7.825452845160535e-06, |
|
"learning_rate": 3.048382309290591e-07, |
|
"logits/chosen": -0.166015625, |
|
"logits/rejected": 0.087890625, |
|
"logps/chosen": -412.0, |
|
"logps/rejected": -316.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.1875, |
|
"rewards/margins": 23.875, |
|
"rewards/rejected": -17.625, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 1.3621794871794872, |
|
"grad_norm": 5.042365843359977e-08, |
|
"learning_rate": 3.033541110121698e-07, |
|
"logits/chosen": -0.337890625, |
|
"logits/rejected": 0.01361083984375, |
|
"logps/chosen": -438.0, |
|
"logps/rejected": -316.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.5, |
|
"rewards/margins": 23.5, |
|
"rewards/rejected": -17.0, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.3701923076923077, |
|
"grad_norm": 3.690049413449436e-05, |
|
"learning_rate": 3.018699910952805e-07, |
|
"logits/chosen": -0.10595703125, |
|
"logits/rejected": 0.291015625, |
|
"logps/chosen": -438.0, |
|
"logps/rejected": -314.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.28125, |
|
"rewards/margins": 25.25, |
|
"rewards/rejected": -18.0, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 1.3782051282051282, |
|
"grad_norm": 2.64418779745463e-07, |
|
"learning_rate": 3.003858711783912e-07, |
|
"logits/chosen": -0.169921875, |
|
"logits/rejected": -0.0306396484375, |
|
"logps/chosen": -444.0, |
|
"logps/rejected": -312.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.75, |
|
"rewards/margins": 24.75, |
|
"rewards/rejected": -18.0, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 1.3862179487179487, |
|
"grad_norm": 1.078909519743223e-07, |
|
"learning_rate": 2.989017512615019e-07, |
|
"logits/chosen": -0.169921875, |
|
"logits/rejected": -0.056884765625, |
|
"logps/chosen": -440.0, |
|
"logps/rejected": -318.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.3125, |
|
"rewards/margins": 25.375, |
|
"rewards/rejected": -18.125, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 1.3942307692307692, |
|
"grad_norm": 1.3777240129954046e-06, |
|
"learning_rate": 2.974176313446126e-07, |
|
"logits/chosen": -0.140625, |
|
"logits/rejected": -0.01025390625, |
|
"logps/chosen": -436.0, |
|
"logps/rejected": -312.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.875, |
|
"rewards/margins": 24.375, |
|
"rewards/rejected": -17.5, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 1.4022435897435899, |
|
"grad_norm": 2.6974739166727197e-08, |
|
"learning_rate": 2.959335114277233e-07, |
|
"logits/chosen": -0.291015625, |
|
"logits/rejected": -0.0751953125, |
|
"logps/chosen": -434.0, |
|
"logps/rejected": -330.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.21875, |
|
"rewards/margins": 25.0, |
|
"rewards/rejected": -17.875, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 1.4102564102564101, |
|
"grad_norm": 7.268216612674599e-09, |
|
"learning_rate": 2.944493915108341e-07, |
|
"logits/chosen": -0.203125, |
|
"logits/rejected": 0.1474609375, |
|
"logps/chosen": -424.0, |
|
"logps/rejected": -310.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.71875, |
|
"rewards/margins": 23.875, |
|
"rewards/rejected": -17.125, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 1.4182692307692308, |
|
"grad_norm": 9.327938089710349e-08, |
|
"learning_rate": 2.929652715939448e-07, |
|
"logits/chosen": -0.3046875, |
|
"logits/rejected": 0.11328125, |
|
"logps/chosen": -430.0, |
|
"logps/rejected": -318.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.96875, |
|
"rewards/margins": 25.25, |
|
"rewards/rejected": -18.25, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 1.4262820512820513, |
|
"grad_norm": 6.223578342493452e-08, |
|
"learning_rate": 2.914811516770555e-07, |
|
"logits/chosen": -0.1181640625, |
|
"logits/rejected": 0.050048828125, |
|
"logps/chosen": -426.0, |
|
"logps/rejected": -338.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.875, |
|
"rewards/margins": 25.0, |
|
"rewards/rejected": -18.125, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 1.4342948717948718, |
|
"grad_norm": 2.1178353398260844e-07, |
|
"learning_rate": 2.8999703176016623e-07, |
|
"logits/chosen": -0.23046875, |
|
"logits/rejected": -0.08740234375, |
|
"logps/chosen": -424.0, |
|
"logps/rejected": -304.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.125, |
|
"rewards/margins": 24.625, |
|
"rewards/rejected": -17.5, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 1.4423076923076923, |
|
"grad_norm": 7.927168012405827e-08, |
|
"learning_rate": 2.8851291184327694e-07, |
|
"logits/chosen": -0.1279296875, |
|
"logits/rejected": 0.1484375, |
|
"logps/chosen": -452.0, |
|
"logps/rejected": -318.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.34375, |
|
"rewards/margins": 25.5, |
|
"rewards/rejected": -18.125, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.4503205128205128, |
|
"grad_norm": 4.212381655007542e-07, |
|
"learning_rate": 2.870287919263876e-07, |
|
"logits/chosen": -0.0576171875, |
|
"logits/rejected": 0.068359375, |
|
"logps/chosen": -438.0, |
|
"logps/rejected": -306.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.21875, |
|
"rewards/margins": 25.0, |
|
"rewards/rejected": -17.75, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 1.4583333333333333, |
|
"grad_norm": 6.531321554518771e-08, |
|
"learning_rate": 2.855446720094983e-07, |
|
"logits/chosen": -0.244140625, |
|
"logits/rejected": 0.12158203125, |
|
"logps/chosen": -466.0, |
|
"logps/rejected": -312.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.3125, |
|
"rewards/margins": 25.25, |
|
"rewards/rejected": -17.875, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 1.4663461538461537, |
|
"grad_norm": 1.0125861032777744e-06, |
|
"learning_rate": 2.840605520926091e-07, |
|
"logits/chosen": -0.265625, |
|
"logits/rejected": -0.0238037109375, |
|
"logps/chosen": -434.0, |
|
"logps/rejected": -324.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.03125, |
|
"rewards/margins": 25.5, |
|
"rewards/rejected": -18.5, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 1.4743589743589745, |
|
"grad_norm": 8.280617198333603e-06, |
|
"learning_rate": 2.825764321757198e-07, |
|
"logits/chosen": -0.154296875, |
|
"logits/rejected": 0.208984375, |
|
"logps/chosen": -446.0, |
|
"logps/rejected": -324.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.375, |
|
"rewards/margins": 26.125, |
|
"rewards/rejected": -18.75, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 1.482371794871795, |
|
"grad_norm": 2.5622369627364615e-08, |
|
"learning_rate": 2.810923122588305e-07, |
|
"logits/chosen": -0.33203125, |
|
"logits/rejected": -0.11865234375, |
|
"logps/chosen": -458.0, |
|
"logps/rejected": -308.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.0625, |
|
"rewards/margins": 24.0, |
|
"rewards/rejected": -17.0, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 1.4903846153846154, |
|
"grad_norm": 0.0024471174954825495, |
|
"learning_rate": 2.796081923419412e-07, |
|
"logits/chosen": -0.07421875, |
|
"logits/rejected": 0.1669921875, |
|
"logps/chosen": -440.0, |
|
"logps/rejected": -324.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.78125, |
|
"rewards/margins": 24.875, |
|
"rewards/rejected": -18.125, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 1.498397435897436, |
|
"grad_norm": 9.546331607080248e-06, |
|
"learning_rate": 2.7812407242505194e-07, |
|
"logits/chosen": -0.259765625, |
|
"logits/rejected": -0.134765625, |
|
"logps/chosen": -450.0, |
|
"logps/rejected": -306.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.3125, |
|
"rewards/margins": 24.875, |
|
"rewards/rejected": -17.5, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 1.5064102564102564, |
|
"grad_norm": 2.115531378638584e-06, |
|
"learning_rate": 2.7663995250816265e-07, |
|
"logits/chosen": -0.3203125, |
|
"logits/rejected": 0.0830078125, |
|
"logps/chosen": -446.0, |
|
"logps/rejected": -338.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.3125, |
|
"rewards/margins": 25.0, |
|
"rewards/rejected": -18.75, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 1.5144230769230769, |
|
"grad_norm": 9.236730728695472e-05, |
|
"learning_rate": 2.7515583259127337e-07, |
|
"logits/chosen": -0.1787109375, |
|
"logits/rejected": 0.216796875, |
|
"logps/chosen": -460.0, |
|
"logps/rejected": -310.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 8.0, |
|
"rewards/margins": 25.875, |
|
"rewards/rejected": -17.875, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 1.5224358974358974, |
|
"grad_norm": 5.527728669208339e-05, |
|
"learning_rate": 2.736717126743841e-07, |
|
"logits/chosen": -0.06787109375, |
|
"logits/rejected": 0.0147705078125, |
|
"logps/chosen": -434.0, |
|
"logps/rejected": -326.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.9375, |
|
"rewards/margins": 25.5, |
|
"rewards/rejected": -18.625, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.530448717948718, |
|
"grad_norm": 4.534702413709536e-05, |
|
"learning_rate": 2.721875927574948e-07, |
|
"logits/chosen": -0.1796875, |
|
"logits/rejected": 0.0242919921875, |
|
"logps/chosen": -442.0, |
|
"logps/rejected": -324.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.34375, |
|
"rewards/margins": 23.5, |
|
"rewards/rejected": -17.25, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 1.5384615384615383, |
|
"grad_norm": 0.0011345205138817006, |
|
"learning_rate": 2.707034728406055e-07, |
|
"logits/chosen": -0.291015625, |
|
"logits/rejected": 0.025390625, |
|
"logps/chosen": -444.0, |
|
"logps/rejected": -334.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.4375, |
|
"rewards/margins": 25.625, |
|
"rewards/rejected": -18.125, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 1.546474358974359, |
|
"grad_norm": 5.9306038997837e-09, |
|
"learning_rate": 2.692193529237162e-07, |
|
"logits/chosen": -0.267578125, |
|
"logits/rejected": 0.08349609375, |
|
"logps/chosen": -446.0, |
|
"logps/rejected": -334.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.125, |
|
"rewards/margins": 25.5, |
|
"rewards/rejected": -18.25, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 1.5544871794871795, |
|
"grad_norm": 3.153067733404093e-08, |
|
"learning_rate": 2.6773523300682693e-07, |
|
"logits/chosen": -0.28515625, |
|
"logits/rejected": 0.2119140625, |
|
"logps/chosen": -428.0, |
|
"logps/rejected": -326.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.90625, |
|
"rewards/margins": 25.625, |
|
"rewards/rejected": -18.75, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 1.5625, |
|
"grad_norm": 1.2264334783515192e-05, |
|
"learning_rate": 2.6625111308993765e-07, |
|
"logits/chosen": -0.09228515625, |
|
"logits/rejected": 0.0791015625, |
|
"logps/chosen": -418.0, |
|
"logps/rejected": -326.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.6875, |
|
"rewards/margins": 25.5, |
|
"rewards/rejected": -18.875, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 1.5705128205128205, |
|
"grad_norm": 4.532094438756683e-06, |
|
"learning_rate": 2.6476699317304836e-07, |
|
"logits/chosen": -0.2373046875, |
|
"logits/rejected": 0.10107421875, |
|
"logps/chosen": -442.0, |
|
"logps/rejected": -314.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.40625, |
|
"rewards/margins": 24.5, |
|
"rewards/rejected": -18.125, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 1.578525641025641, |
|
"grad_norm": 1.1139297198561067e-07, |
|
"learning_rate": 2.6328287325615913e-07, |
|
"logits/chosen": -0.328125, |
|
"logits/rejected": 0.0732421875, |
|
"logps/chosen": -484.0, |
|
"logps/rejected": -322.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.15625, |
|
"rewards/margins": 24.5, |
|
"rewards/rejected": -17.375, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 1.5865384615384617, |
|
"grad_norm": 9.958002279011069e-09, |
|
"learning_rate": 2.6179875333926984e-07, |
|
"logits/chosen": -0.1474609375, |
|
"logits/rejected": 0.019287109375, |
|
"logps/chosen": -422.0, |
|
"logps/rejected": -320.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.28125, |
|
"rewards/margins": 25.5, |
|
"rewards/rejected": -18.25, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 1.594551282051282, |
|
"grad_norm": 2.4632944445492358e-08, |
|
"learning_rate": 2.603146334223805e-07, |
|
"logits/chosen": -0.2890625, |
|
"logits/rejected": 0.078125, |
|
"logps/chosen": -436.0, |
|
"logps/rejected": -312.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.59375, |
|
"rewards/margins": 24.25, |
|
"rewards/rejected": -17.625, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 1.6025641025641026, |
|
"grad_norm": 1.0349315078974401e-07, |
|
"learning_rate": 2.588305135054912e-07, |
|
"logits/chosen": -0.0908203125, |
|
"logits/rejected": 0.0947265625, |
|
"logps/chosen": -444.0, |
|
"logps/rejected": -322.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.40625, |
|
"rewards/margins": 26.0, |
|
"rewards/rejected": -18.5, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.6105769230769231, |
|
"grad_norm": 5.060838211776973e-06, |
|
"learning_rate": 2.5734639358860193e-07, |
|
"logits/chosen": -0.103515625, |
|
"logits/rejected": 0.2578125, |
|
"logps/chosen": -434.0, |
|
"logps/rejected": -322.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.875, |
|
"rewards/margins": 25.125, |
|
"rewards/rejected": -18.25, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 1.6185897435897436, |
|
"grad_norm": 1.970124669721806e-07, |
|
"learning_rate": 2.5586227367171264e-07, |
|
"logits/chosen": -0.271484375, |
|
"logits/rejected": -0.0021514892578125, |
|
"logps/chosen": -448.0, |
|
"logps/rejected": -336.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.8125, |
|
"rewards/margins": 25.5, |
|
"rewards/rejected": -18.75, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 1.626602564102564, |
|
"grad_norm": 2.638778874906699e-07, |
|
"learning_rate": 2.5437815375482335e-07, |
|
"logits/chosen": -0.3828125, |
|
"logits/rejected": 0.03515625, |
|
"logps/chosen": -440.0, |
|
"logps/rejected": -314.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.375, |
|
"rewards/margins": 25.5, |
|
"rewards/rejected": -18.0, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 1.6346153846153846, |
|
"grad_norm": 4.5829106318944607e-07, |
|
"learning_rate": 2.528940338379341e-07, |
|
"logits/chosen": -0.11474609375, |
|
"logits/rejected": 0.057373046875, |
|
"logps/chosen": -452.0, |
|
"logps/rejected": -330.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.375, |
|
"rewards/margins": 26.5, |
|
"rewards/rejected": -19.125, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 1.6426282051282053, |
|
"grad_norm": 3.330319068952062e-07, |
|
"learning_rate": 2.5140991392104483e-07, |
|
"logits/chosen": -0.19140625, |
|
"logits/rejected": 0.181640625, |
|
"logps/chosen": -468.0, |
|
"logps/rejected": -318.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.53125, |
|
"rewards/margins": 26.0, |
|
"rewards/rejected": -18.5, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 1.6506410256410255, |
|
"grad_norm": 1.1041670887761836e-08, |
|
"learning_rate": 2.4992579400415555e-07, |
|
"logits/chosen": -0.1357421875, |
|
"logits/rejected": 0.1630859375, |
|
"logps/chosen": -444.0, |
|
"logps/rejected": -318.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.34375, |
|
"rewards/margins": 25.375, |
|
"rewards/rejected": -18.0, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 1.6586538461538463, |
|
"grad_norm": 7.044774804393972e-07, |
|
"learning_rate": 2.4844167408726626e-07, |
|
"logits/chosen": -0.1025390625, |
|
"logits/rejected": 0.0322265625, |
|
"logps/chosen": -436.0, |
|
"logps/rejected": -304.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.78125, |
|
"rewards/margins": 24.5, |
|
"rewards/rejected": -16.625, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 1.6666666666666665, |
|
"grad_norm": 1.2464745380157458e-05, |
|
"learning_rate": 2.469575541703769e-07, |
|
"logits/chosen": -0.291015625, |
|
"logits/rejected": -0.193359375, |
|
"logps/chosen": -446.0, |
|
"logps/rejected": -310.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.1875, |
|
"rewards/margins": 24.0, |
|
"rewards/rejected": -16.875, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 1.6746794871794872, |
|
"grad_norm": 9.643965697234781e-08, |
|
"learning_rate": 2.454734342534877e-07, |
|
"logits/chosen": -0.251953125, |
|
"logits/rejected": 0.1875, |
|
"logps/chosen": -438.0, |
|
"logps/rejected": -312.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.34375, |
|
"rewards/margins": 24.5, |
|
"rewards/rejected": -17.25, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 1.6826923076923077, |
|
"grad_norm": 1.3982353465556387e-07, |
|
"learning_rate": 2.439893143365984e-07, |
|
"logits/chosen": -0.298828125, |
|
"logits/rejected": -0.012939453125, |
|
"logps/chosen": -434.0, |
|
"logps/rejected": -330.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.0625, |
|
"rewards/margins": 24.875, |
|
"rewards/rejected": -17.875, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 1.6907051282051282, |
|
"grad_norm": 1.469833129512893e-06, |
|
"learning_rate": 2.425051944197091e-07, |
|
"logits/chosen": -0.053955078125, |
|
"logits/rejected": 0.240234375, |
|
"logps/chosen": -416.0, |
|
"logps/rejected": -328.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.3125, |
|
"rewards/margins": 25.0, |
|
"rewards/rejected": -17.75, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 1.6987179487179487, |
|
"grad_norm": 1.366913495379827e-05, |
|
"learning_rate": 2.4102107450281983e-07, |
|
"logits/chosen": -0.326171875, |
|
"logits/rejected": -0.1171875, |
|
"logps/chosen": -438.0, |
|
"logps/rejected": -322.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.0625, |
|
"rewards/margins": 23.375, |
|
"rewards/rejected": -16.375, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 1.7067307692307692, |
|
"grad_norm": 4.874059171656603e-06, |
|
"learning_rate": 2.3953695458593054e-07, |
|
"logits/chosen": -0.06591796875, |
|
"logits/rejected": 0.11962890625, |
|
"logps/chosen": -448.0, |
|
"logps/rejected": -308.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.5625, |
|
"rewards/margins": 25.125, |
|
"rewards/rejected": -17.5, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 1.7147435897435899, |
|
"grad_norm": 8.617834517658852e-08, |
|
"learning_rate": 2.3805283466904126e-07, |
|
"logits/chosen": -0.2109375, |
|
"logits/rejected": -0.0947265625, |
|
"logps/chosen": -436.0, |
|
"logps/rejected": -312.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.21875, |
|
"rewards/margins": 25.0, |
|
"rewards/rejected": -17.875, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 1.7227564102564101, |
|
"grad_norm": 2.1143453585488973e-07, |
|
"learning_rate": 2.3656871475215194e-07, |
|
"logits/chosen": -0.255859375, |
|
"logits/rejected": 0.0791015625, |
|
"logps/chosen": -444.0, |
|
"logps/rejected": -332.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.625, |
|
"rewards/margins": 26.125, |
|
"rewards/rejected": -18.5, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 1.7307692307692308, |
|
"grad_norm": 8.1184601282066e-05, |
|
"learning_rate": 2.3508459483526268e-07, |
|
"logits/chosen": -0.29296875, |
|
"logits/rejected": -0.031005859375, |
|
"logps/chosen": -414.0, |
|
"logps/rejected": -332.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.84375, |
|
"rewards/margins": 24.875, |
|
"rewards/rejected": -18.0, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 1.7387820512820513, |
|
"grad_norm": 9.051233071491503e-08, |
|
"learning_rate": 2.336004749183734e-07, |
|
"logits/chosen": -0.2333984375, |
|
"logits/rejected": 0.057861328125, |
|
"logps/chosen": -426.0, |
|
"logps/rejected": -326.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.75, |
|
"rewards/margins": 25.625, |
|
"rewards/rejected": -18.875, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 1.7467948717948718, |
|
"grad_norm": 4.7277968500829626e-05, |
|
"learning_rate": 2.321163550014841e-07, |
|
"logits/chosen": -0.2275390625, |
|
"logits/rejected": 0.162109375, |
|
"logps/chosen": -448.0, |
|
"logps/rejected": -336.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.21875, |
|
"rewards/margins": 25.625, |
|
"rewards/rejected": -18.5, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 1.7548076923076923, |
|
"grad_norm": 0.0038399746408167177, |
|
"learning_rate": 2.3063223508459482e-07, |
|
"logits/chosen": -0.2255859375, |
|
"logits/rejected": 0.09765625, |
|
"logps/chosen": -426.0, |
|
"logps/rejected": -314.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.25, |
|
"rewards/margins": 25.0, |
|
"rewards/rejected": -17.75, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 1.7628205128205128, |
|
"grad_norm": 4.465718484658786e-06, |
|
"learning_rate": 2.2914811516770554e-07, |
|
"logits/chosen": -0.0419921875, |
|
"logits/rejected": 0.1826171875, |
|
"logps/chosen": -420.0, |
|
"logps/rejected": -300.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.1875, |
|
"rewards/margins": 24.625, |
|
"rewards/rejected": -17.5, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 1.7708333333333335, |
|
"grad_norm": 4.7299663657237144e-07, |
|
"learning_rate": 2.2766399525081625e-07, |
|
"logits/chosen": -0.111328125, |
|
"logits/rejected": 0.043212890625, |
|
"logps/chosen": -438.0, |
|
"logps/rejected": -322.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.875, |
|
"rewards/margins": 25.25, |
|
"rewards/rejected": -18.375, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 1.7788461538461537, |
|
"grad_norm": 1.3733679352015224e-05, |
|
"learning_rate": 2.2617987533392696e-07, |
|
"logits/chosen": -0.1640625, |
|
"logits/rejected": 0.09912109375, |
|
"logps/chosen": -426.0, |
|
"logps/rejected": -332.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.4375, |
|
"rewards/margins": 26.125, |
|
"rewards/rejected": -18.75, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 1.7868589743589745, |
|
"grad_norm": 1.2499827254992277e-07, |
|
"learning_rate": 2.246957554170377e-07, |
|
"logits/chosen": -0.38671875, |
|
"logits/rejected": -0.004364013671875, |
|
"logps/chosen": -438.0, |
|
"logps/rejected": -324.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.09375, |
|
"rewards/margins": 25.125, |
|
"rewards/rejected": -18.125, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 1.7948717948717947, |
|
"grad_norm": 0.00019501858175617213, |
|
"learning_rate": 2.232116355001484e-07, |
|
"logits/chosen": -0.33984375, |
|
"logits/rejected": 0.15234375, |
|
"logps/chosen": -438.0, |
|
"logps/rejected": -346.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.78125, |
|
"rewards/margins": 25.125, |
|
"rewards/rejected": -18.375, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 1.8028846153846154, |
|
"grad_norm": 1.1028833635465485e-07, |
|
"learning_rate": 2.217275155832591e-07, |
|
"logits/chosen": -0.12890625, |
|
"logits/rejected": 0.189453125, |
|
"logps/chosen": -446.0, |
|
"logps/rejected": -328.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.0625, |
|
"rewards/margins": 25.75, |
|
"rewards/rejected": -18.625, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 1.810897435897436, |
|
"grad_norm": 5.152650323990336e-07, |
|
"learning_rate": 2.2024339566636982e-07, |
|
"logits/chosen": -0.1591796875, |
|
"logits/rejected": 0.185546875, |
|
"logps/chosen": -442.0, |
|
"logps/rejected": -326.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.46875, |
|
"rewards/margins": 25.375, |
|
"rewards/rejected": -17.875, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 1.8189102564102564, |
|
"grad_norm": 2.022092778793823e-08, |
|
"learning_rate": 2.1875927574948056e-07, |
|
"logits/chosen": -0.1826171875, |
|
"logits/rejected": 0.2294921875, |
|
"logps/chosen": -456.0, |
|
"logps/rejected": -322.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.75, |
|
"rewards/margins": 26.0, |
|
"rewards/rejected": -18.25, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 1.8269230769230769, |
|
"grad_norm": 1.1365155365539883e-08, |
|
"learning_rate": 2.1727515583259127e-07, |
|
"logits/chosen": -0.2431640625, |
|
"logits/rejected": 0.1494140625, |
|
"logps/chosen": -444.0, |
|
"logps/rejected": -340.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.6875, |
|
"rewards/margins": 26.375, |
|
"rewards/rejected": -19.75, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 1.8349358974358974, |
|
"grad_norm": 3.000780091974624e-08, |
|
"learning_rate": 2.1579103591570196e-07, |
|
"logits/chosen": -0.3125, |
|
"logits/rejected": 0.061767578125, |
|
"logps/chosen": -440.0, |
|
"logps/rejected": -318.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.125, |
|
"rewards/margins": 25.0, |
|
"rewards/rejected": -17.875, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 1.842948717948718, |
|
"grad_norm": 2.1981699727737788e-08, |
|
"learning_rate": 2.143069159988127e-07, |
|
"logits/chosen": -0.1953125, |
|
"logits/rejected": 0.0595703125, |
|
"logps/chosen": -420.0, |
|
"logps/rejected": -328.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.59375, |
|
"rewards/margins": 26.0, |
|
"rewards/rejected": -19.375, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 1.8509615384615383, |
|
"grad_norm": 7.691824865782036e-08, |
|
"learning_rate": 2.128227960819234e-07, |
|
"logits/chosen": -0.177734375, |
|
"logits/rejected": -0.049072265625, |
|
"logps/chosen": -444.0, |
|
"logps/rejected": -322.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.15625, |
|
"rewards/margins": 25.75, |
|
"rewards/rejected": -18.625, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 1.858974358974359, |
|
"grad_norm": 0.00039337693411459787, |
|
"learning_rate": 2.1133867616503413e-07, |
|
"logits/chosen": -0.03369140625, |
|
"logits/rejected": 0.042236328125, |
|
"logps/chosen": -420.0, |
|
"logps/rejected": -330.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.25, |
|
"rewards/margins": 25.25, |
|
"rewards/rejected": -18.0, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 1.8669871794871795, |
|
"grad_norm": 8.784582360286491e-08, |
|
"learning_rate": 2.0985455624814487e-07, |
|
"logits/chosen": -0.08203125, |
|
"logits/rejected": 0.173828125, |
|
"logps/chosen": -414.0, |
|
"logps/rejected": -356.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.4375, |
|
"rewards/margins": 26.375, |
|
"rewards/rejected": -20.0, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 1.875, |
|
"grad_norm": 4.716282231379615e-08, |
|
"learning_rate": 2.0837043633125555e-07, |
|
"logits/chosen": -0.1767578125, |
|
"logits/rejected": 0.031982421875, |
|
"logps/chosen": -468.0, |
|
"logps/rejected": -328.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.0625, |
|
"rewards/margins": 26.25, |
|
"rewards/rejected": -19.125, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 1.8830128205128205, |
|
"grad_norm": 6.25939587350738e-07, |
|
"learning_rate": 2.0688631641436627e-07, |
|
"logits/chosen": -0.25, |
|
"logits/rejected": 0.146484375, |
|
"logps/chosen": -438.0, |
|
"logps/rejected": -334.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.90625, |
|
"rewards/margins": 26.5, |
|
"rewards/rejected": -19.625, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 1.891025641025641, |
|
"grad_norm": 2.767633351591515e-07, |
|
"learning_rate": 2.0540219649747698e-07, |
|
"logits/chosen": -0.0869140625, |
|
"logits/rejected": -0.010986328125, |
|
"logps/chosen": -440.0, |
|
"logps/rejected": -324.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.28125, |
|
"rewards/margins": 25.125, |
|
"rewards/rejected": -18.875, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 1.8990384615384617, |
|
"grad_norm": 0.0023401747821045907, |
|
"learning_rate": 2.0391807658058772e-07, |
|
"logits/chosen": -0.09814453125, |
|
"logits/rejected": 0.027099609375, |
|
"logps/chosen": -440.0, |
|
"logps/rejected": -332.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 5.84375, |
|
"rewards/margins": 24.625, |
|
"rewards/rejected": -18.75, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 1.907051282051282, |
|
"grad_norm": 0.008136366428555634, |
|
"learning_rate": 2.024339566636984e-07, |
|
"logits/chosen": -0.2353515625, |
|
"logits/rejected": -0.115234375, |
|
"logps/chosen": -440.0, |
|
"logps/rejected": -336.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.4375, |
|
"rewards/margins": 25.625, |
|
"rewards/rejected": -19.125, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 1.9150641025641026, |
|
"grad_norm": 3.1743064558400314e-06, |
|
"learning_rate": 2.0094983674680912e-07, |
|
"logits/chosen": -0.212890625, |
|
"logits/rejected": 0.007415771484375, |
|
"logps/chosen": -478.0, |
|
"logps/rejected": -316.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.8125, |
|
"rewards/margins": 26.625, |
|
"rewards/rejected": -18.875, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 1.9230769230769231, |
|
"grad_norm": 1.5866116881967017e-10, |
|
"learning_rate": 1.9946571682991986e-07, |
|
"logits/chosen": -0.185546875, |
|
"logits/rejected": 0.08349609375, |
|
"logps/chosen": -438.0, |
|
"logps/rejected": -336.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.3125, |
|
"rewards/margins": 26.75, |
|
"rewards/rejected": -19.375, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 1.9310897435897436, |
|
"grad_norm": 4.737887286793131e-08, |
|
"learning_rate": 1.9798159691303057e-07, |
|
"logits/chosen": -0.2255859375, |
|
"logits/rejected": 0.1083984375, |
|
"logps/chosen": -416.0, |
|
"logps/rejected": -336.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.59375, |
|
"rewards/margins": 26.375, |
|
"rewards/rejected": -19.75, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 1.939102564102564, |
|
"grad_norm": 1.9028302718209288e-08, |
|
"learning_rate": 1.964974769961413e-07, |
|
"logits/chosen": -0.443359375, |
|
"logits/rejected": 0.07275390625, |
|
"logps/chosen": -462.0, |
|
"logps/rejected": -328.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.3125, |
|
"rewards/margins": 26.625, |
|
"rewards/rejected": -19.25, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 1.9471153846153846, |
|
"grad_norm": 1.1498051880290656e-05, |
|
"learning_rate": 1.9501335707925197e-07, |
|
"logits/chosen": -0.06884765625, |
|
"logits/rejected": 0.0245361328125, |
|
"logps/chosen": -448.0, |
|
"logps/rejected": -334.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.625, |
|
"rewards/margins": 27.25, |
|
"rewards/rejected": -19.625, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 1.9551282051282053, |
|
"grad_norm": 3.6749392836711146e-07, |
|
"learning_rate": 1.9352923716236271e-07, |
|
"logits/chosen": -0.31640625, |
|
"logits/rejected": -0.0130615234375, |
|
"logps/chosen": -432.0, |
|
"logps/rejected": -332.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.875, |
|
"rewards/margins": 26.0, |
|
"rewards/rejected": -19.25, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 1.9631410256410255, |
|
"grad_norm": 7.134708932047622e-05, |
|
"learning_rate": 1.9204511724547343e-07, |
|
"logits/chosen": -0.349609375, |
|
"logits/rejected": -0.078125, |
|
"logps/chosen": -462.0, |
|
"logps/rejected": -312.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.1875, |
|
"rewards/margins": 26.25, |
|
"rewards/rejected": -19.0, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 1.9711538461538463, |
|
"grad_norm": 0.0001062037437627698, |
|
"learning_rate": 1.9056099732858414e-07, |
|
"logits/chosen": -0.330078125, |
|
"logits/rejected": -0.1708984375, |
|
"logps/chosen": -448.0, |
|
"logps/rejected": -326.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.8125, |
|
"rewards/margins": 27.25, |
|
"rewards/rejected": -19.5, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 1.9791666666666665, |
|
"grad_norm": 3.204541773390633e-09, |
|
"learning_rate": 1.8907687741169488e-07, |
|
"logits/chosen": -0.18359375, |
|
"logits/rejected": 0.11962890625, |
|
"logps/chosen": -456.0, |
|
"logps/rejected": -344.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 8.1875, |
|
"rewards/margins": 27.375, |
|
"rewards/rejected": -19.125, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 1.9871794871794872, |
|
"grad_norm": 8.745120231147393e-08, |
|
"learning_rate": 1.8759275749480557e-07, |
|
"logits/chosen": -0.2294921875, |
|
"logits/rejected": 0.0341796875, |
|
"logps/chosen": -426.0, |
|
"logps/rejected": -344.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.90625, |
|
"rewards/margins": 27.0, |
|
"rewards/rejected": -20.125, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 1.9951923076923077, |
|
"grad_norm": 1.4961697556815276e-06, |
|
"learning_rate": 1.8610863757791628e-07, |
|
"logits/chosen": -0.2099609375, |
|
"logits/rejected": -0.14453125, |
|
"logps/chosen": -432.0, |
|
"logps/rejected": -326.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.21875, |
|
"rewards/margins": 26.0, |
|
"rewards/rejected": -18.875, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_logits/chosen": -0.1875, |
|
"eval_logits/rejected": 0.060302734375, |
|
"eval_logps/chosen": -442.0, |
|
"eval_logps/rejected": -340.0, |
|
"eval_loss": 3.2709976949263364e-08, |
|
"eval_rewards/accuracies": 1.0, |
|
"eval_rewards/chosen": 6.40625, |
|
"eval_rewards/margins": 26.0, |
|
"eval_rewards/rejected": -19.625, |
|
"eval_runtime": 26.0509, |
|
"eval_samples_per_second": 7.639, |
|
"eval_steps_per_second": 0.96, |
|
"step": 2496 |
|
}, |
|
{ |
|
"epoch": 2.003205128205128, |
|
"grad_norm": 2.6121739610115405e-09, |
|
"learning_rate": 1.84624517661027e-07, |
|
"logits/chosen": -0.052734375, |
|
"logits/rejected": 0.1513671875, |
|
"logps/chosen": -424.0, |
|
"logps/rejected": -352.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.0625, |
|
"rewards/margins": 27.25, |
|
"rewards/rejected": -20.125, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 2.011217948717949, |
|
"grad_norm": 1.6881400833446535e-06, |
|
"learning_rate": 1.8314039774413774e-07, |
|
"logits/chosen": -0.2158203125, |
|
"logits/rejected": -0.298828125, |
|
"logps/chosen": -456.0, |
|
"logps/rejected": -316.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.96875, |
|
"rewards/margins": 26.125, |
|
"rewards/rejected": -19.125, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 2.019230769230769, |
|
"grad_norm": 1.9134457394828338e-07, |
|
"learning_rate": 1.8165627782724842e-07, |
|
"logits/chosen": -0.244140625, |
|
"logits/rejected": 0.2314453125, |
|
"logps/chosen": -436.0, |
|
"logps/rejected": -344.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.71875, |
|
"rewards/margins": 26.375, |
|
"rewards/rejected": -19.625, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 2.02724358974359, |
|
"grad_norm": 1.0600011367880583e-08, |
|
"learning_rate": 1.8017215791035914e-07, |
|
"logits/chosen": 0.0228271484375, |
|
"logits/rejected": 0.142578125, |
|
"logps/chosen": -420.0, |
|
"logps/rejected": -332.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.09375, |
|
"rewards/margins": 27.125, |
|
"rewards/rejected": -20.0, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 2.03525641025641, |
|
"grad_norm": 1.3526185410157084e-08, |
|
"learning_rate": 1.7868803799346988e-07, |
|
"logits/chosen": -0.06298828125, |
|
"logits/rejected": 0.251953125, |
|
"logps/chosen": -462.0, |
|
"logps/rejected": -334.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.75, |
|
"rewards/margins": 27.25, |
|
"rewards/rejected": -19.5, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 2.043269230769231, |
|
"grad_norm": 8.646807088665669e-07, |
|
"learning_rate": 1.772039180765806e-07, |
|
"logits/chosen": -0.212890625, |
|
"logits/rejected": 0.0250244140625, |
|
"logps/chosen": -456.0, |
|
"logps/rejected": -340.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.25, |
|
"rewards/margins": 27.125, |
|
"rewards/rejected": -19.875, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 2.051282051282051, |
|
"grad_norm": 2.088296065357325e-09, |
|
"learning_rate": 1.757197981596913e-07, |
|
"logits/chosen": -0.341796875, |
|
"logits/rejected": 0.0294189453125, |
|
"logps/chosen": -444.0, |
|
"logps/rejected": -336.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.75, |
|
"rewards/margins": 26.125, |
|
"rewards/rejected": -19.375, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 2.059294871794872, |
|
"grad_norm": 5.184571312873049e-06, |
|
"learning_rate": 1.74235678242802e-07, |
|
"logits/chosen": -0.193359375, |
|
"logits/rejected": -0.00186920166015625, |
|
"logps/chosen": -420.0, |
|
"logps/rejected": -322.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.34375, |
|
"rewards/margins": 26.25, |
|
"rewards/rejected": -18.875, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 2.0673076923076925, |
|
"grad_norm": 4.857993190852093e-09, |
|
"learning_rate": 1.7275155832591273e-07, |
|
"logits/chosen": -0.0986328125, |
|
"logits/rejected": 0.20703125, |
|
"logps/chosen": -432.0, |
|
"logps/rejected": -342.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.59375, |
|
"rewards/margins": 27.875, |
|
"rewards/rejected": -20.25, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 2.0753205128205128, |
|
"grad_norm": 9.506659143431453e-08, |
|
"learning_rate": 1.7126743840902344e-07, |
|
"logits/chosen": -0.365234375, |
|
"logits/rejected": 0.115234375, |
|
"logps/chosen": -440.0, |
|
"logps/rejected": -350.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.4375, |
|
"rewards/margins": 27.0, |
|
"rewards/rejected": -20.625, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 2.0833333333333335, |
|
"grad_norm": 1.0482748462454734e-07, |
|
"learning_rate": 1.6978331849213416e-07, |
|
"logits/chosen": -0.3828125, |
|
"logits/rejected": 0.003662109375, |
|
"logps/chosen": -438.0, |
|
"logps/rejected": -342.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.375, |
|
"rewards/margins": 26.625, |
|
"rewards/rejected": -20.25, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 2.0913461538461537, |
|
"grad_norm": 2.8443951275279085e-07, |
|
"learning_rate": 1.6829919857524487e-07, |
|
"logits/chosen": -0.142578125, |
|
"logits/rejected": 0.234375, |
|
"logps/chosen": -438.0, |
|
"logps/rejected": -340.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.65625, |
|
"rewards/margins": 26.0, |
|
"rewards/rejected": -19.375, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 2.0993589743589745, |
|
"grad_norm": 1.1015728545029192e-08, |
|
"learning_rate": 1.6681507865835558e-07, |
|
"logits/chosen": -0.1767578125, |
|
"logits/rejected": 0.059326171875, |
|
"logps/chosen": -442.0, |
|
"logps/rejected": -340.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.0625, |
|
"rewards/margins": 26.75, |
|
"rewards/rejected": -19.75, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 2.1073717948717947, |
|
"grad_norm": 1.9845276726508088e-07, |
|
"learning_rate": 1.653309587414663e-07, |
|
"logits/chosen": -0.318359375, |
|
"logits/rejected": -0.01361083984375, |
|
"logps/chosen": -436.0, |
|
"logps/rejected": -332.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.875, |
|
"rewards/margins": 26.625, |
|
"rewards/rejected": -19.75, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 2.1153846153846154, |
|
"grad_norm": 1.4467524201059651e-08, |
|
"learning_rate": 1.63846838824577e-07, |
|
"logits/chosen": -0.21484375, |
|
"logits/rejected": 0.240234375, |
|
"logps/chosen": -444.0, |
|
"logps/rejected": -318.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.625, |
|
"rewards/margins": 26.625, |
|
"rewards/rejected": -19.0, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 2.123397435897436, |
|
"grad_norm": 3.049897674505671e-08, |
|
"learning_rate": 1.6236271890768775e-07, |
|
"logits/chosen": -0.27734375, |
|
"logits/rejected": 0.1904296875, |
|
"logps/chosen": -454.0, |
|
"logps/rejected": -328.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.65625, |
|
"rewards/margins": 26.875, |
|
"rewards/rejected": -19.25, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 2.1314102564102564, |
|
"grad_norm": 1.3889973894763567e-07, |
|
"learning_rate": 1.6087859899079844e-07, |
|
"logits/chosen": -0.1494140625, |
|
"logits/rejected": 0.10107421875, |
|
"logps/chosen": -434.0, |
|
"logps/rejected": -340.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 8.0625, |
|
"rewards/margins": 28.25, |
|
"rewards/rejected": -20.125, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 2.139423076923077, |
|
"grad_norm": 2.530907693637047e-09, |
|
"learning_rate": 1.5939447907390915e-07, |
|
"logits/chosen": -0.287109375, |
|
"logits/rejected": 0.035400390625, |
|
"logps/chosen": -428.0, |
|
"logps/rejected": -342.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.65625, |
|
"rewards/margins": 26.75, |
|
"rewards/rejected": -20.125, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 2.1474358974358974, |
|
"grad_norm": 1.8385063583999426e-07, |
|
"learning_rate": 1.579103591570199e-07, |
|
"logits/chosen": -0.1689453125, |
|
"logits/rejected": 0.2119140625, |
|
"logps/chosen": -440.0, |
|
"logps/rejected": -342.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.0, |
|
"rewards/margins": 27.25, |
|
"rewards/rejected": -20.25, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 2.155448717948718, |
|
"grad_norm": 1.0222899605629215e-07, |
|
"learning_rate": 1.564262392401306e-07, |
|
"logits/chosen": -0.486328125, |
|
"logits/rejected": 0.146484375, |
|
"logps/chosen": -476.0, |
|
"logps/rejected": -330.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.8125, |
|
"rewards/margins": 26.625, |
|
"rewards/rejected": -19.75, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 2.1634615384615383, |
|
"grad_norm": 2.673009770338465e-09, |
|
"learning_rate": 1.549421193232413e-07, |
|
"logits/chosen": -0.2353515625, |
|
"logits/rejected": -0.058349609375, |
|
"logps/chosen": -464.0, |
|
"logps/rejected": -336.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.65625, |
|
"rewards/margins": 27.5, |
|
"rewards/rejected": -19.75, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 2.171474358974359, |
|
"grad_norm": 3.370161666209054e-09, |
|
"learning_rate": 1.5345799940635203e-07, |
|
"logits/chosen": -0.2275390625, |
|
"logits/rejected": -0.064453125, |
|
"logps/chosen": -402.0, |
|
"logps/rejected": -332.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.34375, |
|
"rewards/margins": 26.125, |
|
"rewards/rejected": -19.75, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 2.1794871794871793, |
|
"grad_norm": 8.240823071933326e-06, |
|
"learning_rate": 1.5197387948946275e-07, |
|
"logits/chosen": -0.162109375, |
|
"logits/rejected": 0.177734375, |
|
"logps/chosen": -424.0, |
|
"logps/rejected": -326.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.65625, |
|
"rewards/margins": 26.125, |
|
"rewards/rejected": -19.375, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 2.1875, |
|
"grad_norm": 4.0628430592445007e-07, |
|
"learning_rate": 1.5048975957257346e-07, |
|
"logits/chosen": -0.0712890625, |
|
"logits/rejected": 0.2080078125, |
|
"logps/chosen": -448.0, |
|
"logps/rejected": -330.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.625, |
|
"rewards/margins": 27.25, |
|
"rewards/rejected": -19.75, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 2.1955128205128207, |
|
"grad_norm": 4.2134652882764516e-07, |
|
"learning_rate": 1.4900563965568417e-07, |
|
"logits/chosen": -0.232421875, |
|
"logits/rejected": 0.050048828125, |
|
"logps/chosen": -438.0, |
|
"logps/rejected": -338.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.6875, |
|
"rewards/margins": 26.5, |
|
"rewards/rejected": -19.75, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 2.203525641025641, |
|
"grad_norm": 2.488951351298977e-08, |
|
"learning_rate": 1.4752151973879489e-07, |
|
"logits/chosen": -0.271484375, |
|
"logits/rejected": 0.005889892578125, |
|
"logps/chosen": -450.0, |
|
"logps/rejected": -332.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.25, |
|
"rewards/margins": 26.75, |
|
"rewards/rejected": -19.5, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 2.2115384615384617, |
|
"grad_norm": 3.4757939124217577e-09, |
|
"learning_rate": 1.460373998219056e-07, |
|
"logits/chosen": -0.21484375, |
|
"logits/rejected": 0.2451171875, |
|
"logps/chosen": -460.0, |
|
"logps/rejected": -350.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.375, |
|
"rewards/margins": 27.25, |
|
"rewards/rejected": -19.875, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 2.219551282051282, |
|
"grad_norm": 2.3051536993458333e-08, |
|
"learning_rate": 1.4455327990501631e-07, |
|
"logits/chosen": -0.10986328125, |
|
"logits/rejected": 0.1728515625, |
|
"logps/chosen": -416.0, |
|
"logps/rejected": -334.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.4375, |
|
"rewards/margins": 27.25, |
|
"rewards/rejected": -19.75, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 2.2275641025641026, |
|
"grad_norm": 2.0745336907653604e-09, |
|
"learning_rate": 1.4306915998812705e-07, |
|
"logits/chosen": -0.197265625, |
|
"logits/rejected": -0.0026092529296875, |
|
"logps/chosen": -466.0, |
|
"logps/rejected": -334.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.90625, |
|
"rewards/margins": 28.0, |
|
"rewards/rejected": -20.125, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 2.235576923076923, |
|
"grad_norm": 3.511851461254038e-10, |
|
"learning_rate": 1.4158504007123777e-07, |
|
"logits/chosen": -0.25390625, |
|
"logits/rejected": 0.046142578125, |
|
"logps/chosen": -460.0, |
|
"logps/rejected": -322.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 8.0625, |
|
"rewards/margins": 27.125, |
|
"rewards/rejected": -19.0, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 2.2435897435897436, |
|
"grad_norm": 1.4333220655307584e-07, |
|
"learning_rate": 1.4010092015434845e-07, |
|
"logits/chosen": -0.10791015625, |
|
"logits/rejected": 0.140625, |
|
"logps/chosen": -444.0, |
|
"logps/rejected": -350.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.59375, |
|
"rewards/margins": 27.125, |
|
"rewards/rejected": -19.5, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 2.251602564102564, |
|
"grad_norm": 4.0120381045337573e-07, |
|
"learning_rate": 1.3861680023745917e-07, |
|
"logits/chosen": -0.287109375, |
|
"logits/rejected": 0.09033203125, |
|
"logps/chosen": -424.0, |
|
"logps/rejected": -344.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.8125, |
|
"rewards/margins": 26.625, |
|
"rewards/rejected": -19.75, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 2.2596153846153846, |
|
"grad_norm": 2.894672064150088e-09, |
|
"learning_rate": 1.371326803205699e-07, |
|
"logits/chosen": -0.35546875, |
|
"logits/rejected": 0.1533203125, |
|
"logps/chosen": -444.0, |
|
"logps/rejected": -340.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.25, |
|
"rewards/margins": 26.875, |
|
"rewards/rejected": -19.625, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 2.2676282051282053, |
|
"grad_norm": 1.8942015867620864e-07, |
|
"learning_rate": 1.3564856040368062e-07, |
|
"logits/chosen": -0.181640625, |
|
"logits/rejected": 0.08349609375, |
|
"logps/chosen": -416.0, |
|
"logps/rejected": -336.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.4375, |
|
"rewards/margins": 27.0, |
|
"rewards/rejected": -19.625, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 2.2756410256410255, |
|
"grad_norm": 6.23116120374719e-08, |
|
"learning_rate": 1.341644404867913e-07, |
|
"logits/chosen": -0.40625, |
|
"logits/rejected": 0.1728515625, |
|
"logps/chosen": -448.0, |
|
"logps/rejected": -322.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.84375, |
|
"rewards/margins": 26.5, |
|
"rewards/rejected": -18.625, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 2.2836538461538463, |
|
"grad_norm": 1.6986049606583425e-08, |
|
"learning_rate": 1.3268032056990205e-07, |
|
"logits/chosen": -0.470703125, |
|
"logits/rejected": -0.208984375, |
|
"logps/chosen": -464.0, |
|
"logps/rejected": -330.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.3125, |
|
"rewards/margins": 26.25, |
|
"rewards/rejected": -19.0, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 2.2916666666666665, |
|
"grad_norm": 5.482877193966729e-07, |
|
"learning_rate": 1.3119620065301276e-07, |
|
"logits/chosen": -0.1689453125, |
|
"logits/rejected": 0.003631591796875, |
|
"logps/chosen": -440.0, |
|
"logps/rejected": -342.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.96875, |
|
"rewards/margins": 27.5, |
|
"rewards/rejected": -20.5, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 2.2996794871794872, |
|
"grad_norm": 1.5967632534444322e-06, |
|
"learning_rate": 1.2971208073612347e-07, |
|
"logits/chosen": -0.212890625, |
|
"logits/rejected": 0.1220703125, |
|
"logps/chosen": -468.0, |
|
"logps/rejected": -320.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.59375, |
|
"rewards/margins": 26.375, |
|
"rewards/rejected": -18.875, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 2.3076923076923075, |
|
"grad_norm": 9.307953613808063e-09, |
|
"learning_rate": 1.282279608192342e-07, |
|
"logits/chosen": -0.1162109375, |
|
"logits/rejected": 0.1455078125, |
|
"logps/chosen": -448.0, |
|
"logps/rejected": -326.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 8.0, |
|
"rewards/margins": 27.0, |
|
"rewards/rejected": -19.0, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 2.315705128205128, |
|
"grad_norm": 8.885637177895448e-07, |
|
"learning_rate": 1.267438409023449e-07, |
|
"logits/chosen": -0.20703125, |
|
"logits/rejected": -0.046630859375, |
|
"logps/chosen": -456.0, |
|
"logps/rejected": -348.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.6875, |
|
"rewards/margins": 28.625, |
|
"rewards/rejected": -21.0, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 2.323717948717949, |
|
"grad_norm": 7.244821343102667e-08, |
|
"learning_rate": 1.2525972098545562e-07, |
|
"logits/chosen": -0.23828125, |
|
"logits/rejected": 0.016845703125, |
|
"logps/chosen": -442.0, |
|
"logps/rejected": -332.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.3125, |
|
"rewards/margins": 25.75, |
|
"rewards/rejected": -18.375, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 2.331730769230769, |
|
"grad_norm": 1.6862089442865096e-08, |
|
"learning_rate": 1.2377560106856633e-07, |
|
"logits/chosen": -0.265625, |
|
"logits/rejected": 0.07763671875, |
|
"logps/chosen": -422.0, |
|
"logps/rejected": -330.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.15625, |
|
"rewards/margins": 26.625, |
|
"rewards/rejected": -19.5, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 2.33974358974359, |
|
"grad_norm": 7.477291169346368e-08, |
|
"learning_rate": 1.2229148115167704e-07, |
|
"logits/chosen": -0.314453125, |
|
"logits/rejected": 0.03515625, |
|
"logps/chosen": -422.0, |
|
"logps/rejected": -342.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.5625, |
|
"rewards/margins": 26.25, |
|
"rewards/rejected": -19.75, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 2.34775641025641, |
|
"grad_norm": 4.820125784216706e-09, |
|
"learning_rate": 1.2080736123478776e-07, |
|
"logits/chosen": -0.345703125, |
|
"logits/rejected": -0.01251220703125, |
|
"logps/chosen": -442.0, |
|
"logps/rejected": -338.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.84375, |
|
"rewards/margins": 26.375, |
|
"rewards/rejected": -19.5, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 2.355769230769231, |
|
"grad_norm": 7.997123136078186e-06, |
|
"learning_rate": 1.1932324131789847e-07, |
|
"logits/chosen": -0.396484375, |
|
"logits/rejected": 0.1181640625, |
|
"logps/chosen": -424.0, |
|
"logps/rejected": -342.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.5, |
|
"rewards/margins": 25.5, |
|
"rewards/rejected": -19.0, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 2.363782051282051, |
|
"grad_norm": 1.129621632217089e-07, |
|
"learning_rate": 1.178391214010092e-07, |
|
"logits/chosen": -0.28515625, |
|
"logits/rejected": -0.0634765625, |
|
"logps/chosen": -432.0, |
|
"logps/rejected": -324.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 8.125, |
|
"rewards/margins": 27.375, |
|
"rewards/rejected": -19.375, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 2.371794871794872, |
|
"grad_norm": 2.148083083277937e-08, |
|
"learning_rate": 1.1635500148411991e-07, |
|
"logits/chosen": -0.28125, |
|
"logits/rejected": 0.0037078857421875, |
|
"logps/chosen": -422.0, |
|
"logps/rejected": -336.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.46875, |
|
"rewards/margins": 27.25, |
|
"rewards/rejected": -19.75, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 2.3798076923076925, |
|
"grad_norm": 4.125937342376267e-09, |
|
"learning_rate": 1.1487088156723062e-07, |
|
"logits/chosen": -0.1494140625, |
|
"logits/rejected": 0.203125, |
|
"logps/chosen": -444.0, |
|
"logps/rejected": -340.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.75, |
|
"rewards/margins": 27.375, |
|
"rewards/rejected": -19.625, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 2.3878205128205128, |
|
"grad_norm": 7.302417174330589e-07, |
|
"learning_rate": 1.1338676165034135e-07, |
|
"logits/chosen": -0.318359375, |
|
"logits/rejected": -0.115234375, |
|
"logps/chosen": -450.0, |
|
"logps/rejected": -332.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.84375, |
|
"rewards/margins": 26.875, |
|
"rewards/rejected": -19.125, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 2.3958333333333335, |
|
"grad_norm": 1.1082703968468988e-09, |
|
"learning_rate": 1.1190264173345205e-07, |
|
"logits/chosen": -0.109375, |
|
"logits/rejected": 0.18359375, |
|
"logps/chosen": -440.0, |
|
"logps/rejected": -318.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.09375, |
|
"rewards/margins": 26.125, |
|
"rewards/rejected": -19.0, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 2.4038461538461537, |
|
"grad_norm": 4.31754340643006e-07, |
|
"learning_rate": 1.1041852181656278e-07, |
|
"logits/chosen": -0.06201171875, |
|
"logits/rejected": 0.310546875, |
|
"logps/chosen": -430.0, |
|
"logps/rejected": -348.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.375, |
|
"rewards/margins": 27.75, |
|
"rewards/rejected": -20.375, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.4118589743589745, |
|
"grad_norm": 1.727993900265872e-08, |
|
"learning_rate": 1.0893440189967348e-07, |
|
"logits/chosen": 0.05859375, |
|
"logits/rejected": 0.265625, |
|
"logps/chosen": -412.0, |
|
"logps/rejected": -350.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.28125, |
|
"rewards/margins": 26.75, |
|
"rewards/rejected": -20.5, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 2.4198717948717947, |
|
"grad_norm": 8.432746597818138e-08, |
|
"learning_rate": 1.074502819827842e-07, |
|
"logits/chosen": -0.1337890625, |
|
"logits/rejected": 0.2373046875, |
|
"logps/chosen": -446.0, |
|
"logps/rejected": -332.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.90625, |
|
"rewards/margins": 27.75, |
|
"rewards/rejected": -19.875, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 2.4278846153846154, |
|
"grad_norm": 2.393685614587526e-09, |
|
"learning_rate": 1.0596616206589493e-07, |
|
"logits/chosen": -0.115234375, |
|
"logits/rejected": -0.044921875, |
|
"logps/chosen": -462.0, |
|
"logps/rejected": -336.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.40625, |
|
"rewards/margins": 27.25, |
|
"rewards/rejected": -19.875, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 2.435897435897436, |
|
"grad_norm": 9.467921708070538e-08, |
|
"learning_rate": 1.0448204214900563e-07, |
|
"logits/chosen": -0.2041015625, |
|
"logits/rejected": 0.2734375, |
|
"logps/chosen": -452.0, |
|
"logps/rejected": -336.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.84375, |
|
"rewards/margins": 26.75, |
|
"rewards/rejected": -19.875, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 2.4439102564102564, |
|
"grad_norm": 5.044263933922816e-06, |
|
"learning_rate": 1.0299792223211636e-07, |
|
"logits/chosen": -0.115234375, |
|
"logits/rejected": 0.049560546875, |
|
"logps/chosen": -406.0, |
|
"logps/rejected": -324.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.1875, |
|
"rewards/margins": 26.375, |
|
"rewards/rejected": -19.25, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 2.451923076923077, |
|
"grad_norm": 9.315786057397964e-09, |
|
"learning_rate": 1.0151380231522706e-07, |
|
"logits/chosen": -0.236328125, |
|
"logits/rejected": 0.12109375, |
|
"logps/chosen": -444.0, |
|
"logps/rejected": -336.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.46875, |
|
"rewards/margins": 27.125, |
|
"rewards/rejected": -19.625, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 2.4599358974358974, |
|
"grad_norm": 9.302550355271312e-07, |
|
"learning_rate": 1.0002968239833778e-07, |
|
"logits/chosen": -0.265625, |
|
"logits/rejected": -0.109375, |
|
"logps/chosen": -448.0, |
|
"logps/rejected": -336.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 5.875, |
|
"rewards/margins": 25.25, |
|
"rewards/rejected": -19.375, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 2.467948717948718, |
|
"grad_norm": 1.6804365204736658e-09, |
|
"learning_rate": 9.854556248144849e-08, |
|
"logits/chosen": -0.2578125, |
|
"logits/rejected": -0.06884765625, |
|
"logps/chosen": -442.0, |
|
"logps/rejected": -360.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.1875, |
|
"rewards/margins": 27.875, |
|
"rewards/rejected": -20.75, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 2.4759615384615383, |
|
"grad_norm": 3.700665777017357e-07, |
|
"learning_rate": 9.706144256455921e-08, |
|
"logits/chosen": -0.1181640625, |
|
"logits/rejected": 0.1630859375, |
|
"logps/chosen": -426.0, |
|
"logps/rejected": -344.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.6875, |
|
"rewards/margins": 26.625, |
|
"rewards/rejected": -20.0, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 2.483974358974359, |
|
"grad_norm": 3.356148054596035e-07, |
|
"learning_rate": 9.557732264766994e-08, |
|
"logits/chosen": -0.0458984375, |
|
"logits/rejected": 0.171875, |
|
"logps/chosen": -436.0, |
|
"logps/rejected": -336.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.53125, |
|
"rewards/margins": 26.25, |
|
"rewards/rejected": -19.75, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 2.4919871794871793, |
|
"grad_norm": 5.502119994839622e-09, |
|
"learning_rate": 9.409320273078064e-08, |
|
"logits/chosen": -0.1611328125, |
|
"logits/rejected": 0.140625, |
|
"logps/chosen": -428.0, |
|
"logps/rejected": -334.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.09375, |
|
"rewards/margins": 26.875, |
|
"rewards/rejected": -19.75, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"grad_norm": 0.0005456427835364934, |
|
"learning_rate": 9.260908281389137e-08, |
|
"logits/chosen": -0.4765625, |
|
"logits/rejected": -0.05029296875, |
|
"logps/chosen": -440.0, |
|
"logps/rejected": -348.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.96875, |
|
"rewards/margins": 26.875, |
|
"rewards/rejected": -20.0, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 2.5080128205128203, |
|
"grad_norm": 2.1176776708636095e-09, |
|
"learning_rate": 9.112496289700207e-08, |
|
"logits/chosen": -0.26171875, |
|
"logits/rejected": 0.154296875, |
|
"logps/chosen": -418.0, |
|
"logps/rejected": -358.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.25, |
|
"rewards/margins": 28.0, |
|
"rewards/rejected": -20.75, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 2.516025641025641, |
|
"grad_norm": 1.1607162397089326e-09, |
|
"learning_rate": 8.964084298011279e-08, |
|
"logits/chosen": -0.19140625, |
|
"logits/rejected": 0.0216064453125, |
|
"logps/chosen": -448.0, |
|
"logps/rejected": -356.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.09375, |
|
"rewards/margins": 28.25, |
|
"rewards/rejected": -21.125, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 2.5240384615384617, |
|
"grad_norm": 9.042045049085206e-09, |
|
"learning_rate": 8.815672306322349e-08, |
|
"logits/chosen": -0.095703125, |
|
"logits/rejected": 0.294921875, |
|
"logps/chosen": -424.0, |
|
"logps/rejected": -354.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.34375, |
|
"rewards/margins": 27.625, |
|
"rewards/rejected": -20.25, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 2.532051282051282, |
|
"grad_norm": 2.4305263518120713e-08, |
|
"learning_rate": 8.667260314633422e-08, |
|
"logits/chosen": -0.34375, |
|
"logits/rejected": 0.0869140625, |
|
"logps/chosen": -456.0, |
|
"logps/rejected": -348.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.09375, |
|
"rewards/margins": 27.25, |
|
"rewards/rejected": -20.25, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 2.5400641025641026, |
|
"grad_norm": 1.022673774580843e-09, |
|
"learning_rate": 8.518848322944495e-08, |
|
"logits/chosen": -0.244140625, |
|
"logits/rejected": 0.06494140625, |
|
"logps/chosen": -462.0, |
|
"logps/rejected": -332.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.5625, |
|
"rewards/margins": 27.25, |
|
"rewards/rejected": -19.75, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 2.5480769230769234, |
|
"grad_norm": 1.8047248402548718e-07, |
|
"learning_rate": 8.370436331255565e-08, |
|
"logits/chosen": -0.169921875, |
|
"logits/rejected": 0.10498046875, |
|
"logps/chosen": -438.0, |
|
"logps/rejected": -336.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.25, |
|
"rewards/margins": 27.25, |
|
"rewards/rejected": -20.0, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 2.5560897435897436, |
|
"grad_norm": 1.503123349138464e-08, |
|
"learning_rate": 8.222024339566637e-08, |
|
"logits/chosen": -0.1748046875, |
|
"logits/rejected": 0.248046875, |
|
"logps/chosen": -446.0, |
|
"logps/rejected": -340.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.84375, |
|
"rewards/margins": 28.0, |
|
"rewards/rejected": -20.125, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 2.564102564102564, |
|
"grad_norm": 3.1730913303972646e-06, |
|
"learning_rate": 8.073612347877707e-08, |
|
"logits/chosen": -0.30859375, |
|
"logits/rejected": 0.1416015625, |
|
"logps/chosen": -470.0, |
|
"logps/rejected": -340.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.21875, |
|
"rewards/margins": 27.25, |
|
"rewards/rejected": -20.0, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 2.5721153846153846, |
|
"grad_norm": 2.0449527362487048e-09, |
|
"learning_rate": 7.92520035618878e-08, |
|
"logits/chosen": -0.30859375, |
|
"logits/rejected": 0.034912109375, |
|
"logps/chosen": -456.0, |
|
"logps/rejected": -344.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.1875, |
|
"rewards/margins": 27.375, |
|
"rewards/rejected": -20.25, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 2.5801282051282053, |
|
"grad_norm": 2.6361775335408106e-08, |
|
"learning_rate": 7.776788364499851e-08, |
|
"logits/chosen": -0.12451171875, |
|
"logits/rejected": -0.031982421875, |
|
"logps/chosen": -426.0, |
|
"logps/rejected": -350.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.84375, |
|
"rewards/margins": 27.375, |
|
"rewards/rejected": -20.5, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 2.5881410256410255, |
|
"grad_norm": 5.643642191509336e-07, |
|
"learning_rate": 7.628376372810923e-08, |
|
"logits/chosen": -0.322265625, |
|
"logits/rejected": -0.01324462890625, |
|
"logps/chosen": -466.0, |
|
"logps/rejected": -358.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.96875, |
|
"rewards/margins": 28.375, |
|
"rewards/rejected": -20.375, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 2.5961538461538463, |
|
"grad_norm": 7.851543088124693e-07, |
|
"learning_rate": 7.479964381121995e-08, |
|
"logits/chosen": -0.357421875, |
|
"logits/rejected": -0.09765625, |
|
"logps/chosen": -444.0, |
|
"logps/rejected": -346.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.9375, |
|
"rewards/margins": 26.875, |
|
"rewards/rejected": -20.0, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 2.6041666666666665, |
|
"grad_norm": 6.947854116251245e-07, |
|
"learning_rate": 7.331552389433065e-08, |
|
"logits/chosen": -0.1279296875, |
|
"logits/rejected": 0.31640625, |
|
"logps/chosen": -440.0, |
|
"logps/rejected": -350.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.65625, |
|
"rewards/margins": 28.25, |
|
"rewards/rejected": -20.5, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 2.6121794871794872, |
|
"grad_norm": 7.407083121469371e-09, |
|
"learning_rate": 7.183140397744138e-08, |
|
"logits/chosen": -0.333984375, |
|
"logits/rejected": 0.09326171875, |
|
"logps/chosen": -440.0, |
|
"logps/rejected": -340.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 5.6875, |
|
"rewards/margins": 25.875, |
|
"rewards/rejected": -20.125, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 2.6201923076923075, |
|
"grad_norm": 4.051007518887843e-09, |
|
"learning_rate": 7.034728406055208e-08, |
|
"logits/chosen": -0.1953125, |
|
"logits/rejected": -0.11865234375, |
|
"logps/chosen": -456.0, |
|
"logps/rejected": -344.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.3125, |
|
"rewards/margins": 27.5, |
|
"rewards/rejected": -20.125, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 2.628205128205128, |
|
"grad_norm": 9.95357533946532e-07, |
|
"learning_rate": 6.886316414366281e-08, |
|
"logits/chosen": -0.11767578125, |
|
"logits/rejected": 0.234375, |
|
"logps/chosen": -432.0, |
|
"logps/rejected": -346.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.125, |
|
"rewards/margins": 26.75, |
|
"rewards/rejected": -19.625, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 2.636217948717949, |
|
"grad_norm": 1.3028765268029341e-06, |
|
"learning_rate": 6.737904422677352e-08, |
|
"logits/chosen": -0.1474609375, |
|
"logits/rejected": -0.06201171875, |
|
"logps/chosen": -452.0, |
|
"logps/rejected": -342.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.6875, |
|
"rewards/margins": 28.125, |
|
"rewards/rejected": -20.375, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 2.644230769230769, |
|
"grad_norm": 1.0559072876683408e-07, |
|
"learning_rate": 6.589492430988424e-08, |
|
"logits/chosen": -0.251953125, |
|
"logits/rejected": 0.0091552734375, |
|
"logps/chosen": -448.0, |
|
"logps/rejected": -336.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.5625, |
|
"rewards/margins": 27.0, |
|
"rewards/rejected": -19.375, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 2.65224358974359, |
|
"grad_norm": 1.8285606675806575e-06, |
|
"learning_rate": 6.441080439299495e-08, |
|
"logits/chosen": -0.34375, |
|
"logits/rejected": 0.0286865234375, |
|
"logps/chosen": -462.0, |
|
"logps/rejected": -350.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.03125, |
|
"rewards/margins": 26.875, |
|
"rewards/rejected": -19.875, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 2.66025641025641, |
|
"grad_norm": 8.537748712200147e-07, |
|
"learning_rate": 6.292668447610566e-08, |
|
"logits/chosen": -0.2470703125, |
|
"logits/rejected": 0.2275390625, |
|
"logps/chosen": -460.0, |
|
"logps/rejected": -336.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.125, |
|
"rewards/margins": 26.0, |
|
"rewards/rejected": -18.875, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 2.668269230769231, |
|
"grad_norm": 6.529274060247193e-08, |
|
"learning_rate": 6.144256455921639e-08, |
|
"logits/chosen": -0.25390625, |
|
"logits/rejected": 0.078125, |
|
"logps/chosen": -440.0, |
|
"logps/rejected": -336.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.09375, |
|
"rewards/margins": 27.0, |
|
"rewards/rejected": -20.0, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 2.676282051282051, |
|
"grad_norm": 4.575112689688311e-08, |
|
"learning_rate": 5.99584446423271e-08, |
|
"logits/chosen": -0.3125, |
|
"logits/rejected": -0.07763671875, |
|
"logps/chosen": -436.0, |
|
"logps/rejected": -320.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.0, |
|
"rewards/margins": 25.875, |
|
"rewards/rejected": -18.875, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 2.684294871794872, |
|
"grad_norm": 1.9762039298590372e-07, |
|
"learning_rate": 5.847432472543781e-08, |
|
"logits/chosen": -0.375, |
|
"logits/rejected": 0.169921875, |
|
"logps/chosen": -430.0, |
|
"logps/rejected": -336.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.8125, |
|
"rewards/margins": 27.0, |
|
"rewards/rejected": -20.25, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 2.6923076923076925, |
|
"grad_norm": 1.1136454045426584e-08, |
|
"learning_rate": 5.699020480854853e-08, |
|
"logits/chosen": -0.322265625, |
|
"logits/rejected": 0.00604248046875, |
|
"logps/chosen": -440.0, |
|
"logps/rejected": -340.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.0, |
|
"rewards/margins": 27.125, |
|
"rewards/rejected": -20.125, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 2.7003205128205128, |
|
"grad_norm": 1.3657969717986824e-05, |
|
"learning_rate": 5.550608489165924e-08, |
|
"logits/chosen": -0.251953125, |
|
"logits/rejected": 0.171875, |
|
"logps/chosen": -434.0, |
|
"logps/rejected": -342.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.53125, |
|
"rewards/margins": 26.375, |
|
"rewards/rejected": -19.875, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 2.7083333333333335, |
|
"grad_norm": 5.489399316332875e-09, |
|
"learning_rate": 5.4021964974769963e-08, |
|
"logits/chosen": -0.31640625, |
|
"logits/rejected": -0.10205078125, |
|
"logps/chosen": -468.0, |
|
"logps/rejected": -342.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.75, |
|
"rewards/margins": 27.75, |
|
"rewards/rejected": -20.0, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 2.7163461538461537, |
|
"grad_norm": 0.0002623014112957254, |
|
"learning_rate": 5.253784505788068e-08, |
|
"logits/chosen": -0.44921875, |
|
"logits/rejected": -0.10009765625, |
|
"logps/chosen": -454.0, |
|
"logps/rejected": -354.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.0625, |
|
"rewards/margins": 27.875, |
|
"rewards/rejected": -20.75, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 2.7243589743589745, |
|
"grad_norm": 2.195529153656888e-08, |
|
"learning_rate": 5.105372514099139e-08, |
|
"logits/chosen": -0.29296875, |
|
"logits/rejected": 0.1591796875, |
|
"logps/chosen": -448.0, |
|
"logps/rejected": -354.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.75, |
|
"rewards/margins": 28.625, |
|
"rewards/rejected": -21.0, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 2.7323717948717947, |
|
"grad_norm": 2.4294081800305804e-06, |
|
"learning_rate": 4.9569605224102104e-08, |
|
"logits/chosen": -0.2158203125, |
|
"logits/rejected": 0.19921875, |
|
"logps/chosen": -432.0, |
|
"logps/rejected": -338.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.53125, |
|
"rewards/margins": 27.625, |
|
"rewards/rejected": -20.125, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 2.7403846153846154, |
|
"grad_norm": 2.446165178575366e-07, |
|
"learning_rate": 4.808548530721282e-08, |
|
"logits/chosen": -0.369140625, |
|
"logits/rejected": 0.154296875, |
|
"logps/chosen": -460.0, |
|
"logps/rejected": -346.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.0, |
|
"rewards/margins": 26.875, |
|
"rewards/rejected": -19.875, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 2.748397435897436, |
|
"grad_norm": 9.050820753016259e-09, |
|
"learning_rate": 4.660136539032353e-08, |
|
"logits/chosen": -0.169921875, |
|
"logits/rejected": -0.1083984375, |
|
"logps/chosen": -434.0, |
|
"logps/rejected": -332.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.9375, |
|
"rewards/margins": 26.0, |
|
"rewards/rejected": -19.125, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 2.7564102564102564, |
|
"grad_norm": 8.70024811580348e-08, |
|
"learning_rate": 4.511724547343425e-08, |
|
"logits/chosen": -0.244140625, |
|
"logits/rejected": 0.275390625, |
|
"logps/chosen": -446.0, |
|
"logps/rejected": -342.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.78125, |
|
"rewards/margins": 28.5, |
|
"rewards/rejected": -20.75, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 2.7644230769230766, |
|
"grad_norm": 1.7819585962761552e-09, |
|
"learning_rate": 4.363312555654497e-08, |
|
"logits/chosen": -0.365234375, |
|
"logits/rejected": 0.0224609375, |
|
"logps/chosen": -466.0, |
|
"logps/rejected": -336.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.8125, |
|
"rewards/margins": 26.625, |
|
"rewards/rejected": -19.75, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 2.7724358974358974, |
|
"grad_norm": 4.096945676829309e-07, |
|
"learning_rate": 4.2149005639655685e-08, |
|
"logits/chosen": -0.18359375, |
|
"logits/rejected": 0.06298828125, |
|
"logps/chosen": -422.0, |
|
"logps/rejected": -350.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.84375, |
|
"rewards/margins": 27.0, |
|
"rewards/rejected": -20.25, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 2.780448717948718, |
|
"grad_norm": 2.206393400089359e-08, |
|
"learning_rate": 4.06648857227664e-08, |
|
"logits/chosen": -0.048095703125, |
|
"logits/rejected": 0.042724609375, |
|
"logps/chosen": -436.0, |
|
"logps/rejected": -336.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 8.0625, |
|
"rewards/margins": 27.875, |
|
"rewards/rejected": -19.75, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 2.7884615384615383, |
|
"grad_norm": 1.3660554306973144e-07, |
|
"learning_rate": 3.918076580587711e-08, |
|
"logits/chosen": -0.234375, |
|
"logits/rejected": 0.2138671875, |
|
"logps/chosen": -448.0, |
|
"logps/rejected": -356.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.28125, |
|
"rewards/margins": 27.625, |
|
"rewards/rejected": -20.375, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 2.796474358974359, |
|
"grad_norm": 1.1879089760749721e-07, |
|
"learning_rate": 3.7696645888987825e-08, |
|
"logits/chosen": -0.255859375, |
|
"logits/rejected": 0.0162353515625, |
|
"logps/chosen": -462.0, |
|
"logps/rejected": -342.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.75, |
|
"rewards/margins": 28.375, |
|
"rewards/rejected": -20.625, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 2.8044871794871797, |
|
"grad_norm": 3.764907615781671e-10, |
|
"learning_rate": 3.621252597209854e-08, |
|
"logits/chosen": -0.11474609375, |
|
"logits/rejected": 0.031982421875, |
|
"logps/chosen": -444.0, |
|
"logps/rejected": -338.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.28125, |
|
"rewards/margins": 27.125, |
|
"rewards/rejected": -19.75, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 2.8125, |
|
"grad_norm": 1.22350859308155e-08, |
|
"learning_rate": 3.4728406055209265e-08, |
|
"logits/chosen": -0.11376953125, |
|
"logits/rejected": 0.1201171875, |
|
"logps/chosen": -422.0, |
|
"logps/rejected": -368.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.0, |
|
"rewards/margins": 27.625, |
|
"rewards/rejected": -20.625, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 2.8205128205128203, |
|
"grad_norm": 9.041162718787802e-08, |
|
"learning_rate": 3.324428613831998e-08, |
|
"logits/chosen": -0.34765625, |
|
"logits/rejected": -0.2578125, |
|
"logps/chosen": -454.0, |
|
"logps/rejected": -340.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.71875, |
|
"rewards/margins": 28.125, |
|
"rewards/rejected": -20.375, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 2.828525641025641, |
|
"grad_norm": 3.1091626777529395e-09, |
|
"learning_rate": 3.176016622143069e-08, |
|
"logits/chosen": -0.2099609375, |
|
"logits/rejected": 0.11865234375, |
|
"logps/chosen": -432.0, |
|
"logps/rejected": -350.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.0, |
|
"rewards/margins": 27.25, |
|
"rewards/rejected": -20.25, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 2.8365384615384617, |
|
"grad_norm": 6.853554510061653e-09, |
|
"learning_rate": 3.0276046304541406e-08, |
|
"logits/chosen": -0.25, |
|
"logits/rejected": 0.23046875, |
|
"logps/chosen": -456.0, |
|
"logps/rejected": -342.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.125, |
|
"rewards/margins": 27.625, |
|
"rewards/rejected": -20.5, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 2.844551282051282, |
|
"grad_norm": 3.2103065888110396e-08, |
|
"learning_rate": 2.879192638765212e-08, |
|
"logits/chosen": -0.162109375, |
|
"logits/rejected": 0.1923828125, |
|
"logps/chosen": -432.0, |
|
"logps/rejected": -330.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.09375, |
|
"rewards/margins": 26.25, |
|
"rewards/rejected": -19.125, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 2.8525641025641026, |
|
"grad_norm": 0.00036624099978641543, |
|
"learning_rate": 2.7307806470762836e-08, |
|
"logits/chosen": -0.25390625, |
|
"logits/rejected": -0.0240478515625, |
|
"logps/chosen": -420.0, |
|
"logps/rejected": -340.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.78125, |
|
"rewards/margins": 27.25, |
|
"rewards/rejected": -20.5, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 2.8605769230769234, |
|
"grad_norm": 1.899113115418092e-08, |
|
"learning_rate": 2.5823686553873553e-08, |
|
"logits/chosen": -0.1298828125, |
|
"logits/rejected": 0.20703125, |
|
"logps/chosen": -432.0, |
|
"logps/rejected": -340.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.46875, |
|
"rewards/margins": 27.875, |
|
"rewards/rejected": -20.5, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 2.8685897435897436, |
|
"grad_norm": 3.5524195529944243e-09, |
|
"learning_rate": 2.4339566636984267e-08, |
|
"logits/chosen": -0.125, |
|
"logits/rejected": 0.05126953125, |
|
"logps/chosen": -432.0, |
|
"logps/rejected": -338.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.09375, |
|
"rewards/margins": 26.625, |
|
"rewards/rejected": -19.625, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 2.876602564102564, |
|
"grad_norm": 1.462546881773056e-07, |
|
"learning_rate": 2.2855446720094983e-08, |
|
"logits/chosen": -0.34765625, |
|
"logits/rejected": 0.123046875, |
|
"logps/chosen": -468.0, |
|
"logps/rejected": -348.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.375, |
|
"rewards/margins": 27.5, |
|
"rewards/rejected": -20.0, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 2.8846153846153846, |
|
"grad_norm": 0.0003947150129901893, |
|
"learning_rate": 2.1371326803205697e-08, |
|
"logits/chosen": -0.259765625, |
|
"logits/rejected": 0.04736328125, |
|
"logps/chosen": -442.0, |
|
"logps/rejected": -348.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.15625, |
|
"rewards/margins": 27.875, |
|
"rewards/rejected": -20.75, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 2.8926282051282053, |
|
"grad_norm": 0.00037491833490326394, |
|
"learning_rate": 1.9887206886316414e-08, |
|
"logits/chosen": -0.2109375, |
|
"logits/rejected": 0.1376953125, |
|
"logps/chosen": -430.0, |
|
"logps/rejected": -348.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.03125, |
|
"rewards/margins": 27.375, |
|
"rewards/rejected": -20.25, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 2.9006410256410255, |
|
"grad_norm": 3.10949403818022e-09, |
|
"learning_rate": 1.840308696942713e-08, |
|
"logits/chosen": -0.09765625, |
|
"logits/rejected": 0.26171875, |
|
"logps/chosen": -470.0, |
|
"logps/rejected": -358.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.5, |
|
"rewards/margins": 28.25, |
|
"rewards/rejected": -20.75, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 2.9086538461538463, |
|
"grad_norm": 1.1737181224667657e-06, |
|
"learning_rate": 1.6918967052537844e-08, |
|
"logits/chosen": -0.032958984375, |
|
"logits/rejected": 0.326171875, |
|
"logps/chosen": -416.0, |
|
"logps/rejected": -354.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.59375, |
|
"rewards/margins": 27.375, |
|
"rewards/rejected": -20.75, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 2.9166666666666665, |
|
"grad_norm": 3.8851775144645865e-08, |
|
"learning_rate": 1.5434847135648558e-08, |
|
"logits/chosen": -0.166015625, |
|
"logits/rejected": 0.04541015625, |
|
"logps/chosen": -398.0, |
|
"logps/rejected": -344.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.40625, |
|
"rewards/margins": 25.875, |
|
"rewards/rejected": -19.5, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 2.9246794871794872, |
|
"grad_norm": 2.74142800862514e-08, |
|
"learning_rate": 1.3950727218759274e-08, |
|
"logits/chosen": -0.390625, |
|
"logits/rejected": -0.06591796875, |
|
"logps/chosen": -434.0, |
|
"logps/rejected": -324.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.71875, |
|
"rewards/margins": 26.875, |
|
"rewards/rejected": -19.125, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 2.9326923076923075, |
|
"grad_norm": 6.044117083165874e-07, |
|
"learning_rate": 1.2466607301869991e-08, |
|
"logits/chosen": -0.0888671875, |
|
"logits/rejected": 0.173828125, |
|
"logps/chosen": -444.0, |
|
"logps/rejected": -322.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.9375, |
|
"rewards/margins": 27.75, |
|
"rewards/rejected": -19.75, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 2.940705128205128, |
|
"grad_norm": 2.4109491364079365e-09, |
|
"learning_rate": 1.0982487384980706e-08, |
|
"logits/chosen": -0.1259765625, |
|
"logits/rejected": -0.052978515625, |
|
"logps/chosen": -448.0, |
|
"logps/rejected": -330.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.6875, |
|
"rewards/margins": 27.25, |
|
"rewards/rejected": -19.625, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 2.948717948717949, |
|
"grad_norm": 9.044790273566575e-08, |
|
"learning_rate": 9.498367468091422e-09, |
|
"logits/chosen": -0.318359375, |
|
"logits/rejected": -0.0888671875, |
|
"logps/chosen": -424.0, |
|
"logps/rejected": -350.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.28125, |
|
"rewards/margins": 27.375, |
|
"rewards/rejected": -20.125, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 2.956730769230769, |
|
"grad_norm": 1.7054162118154904e-09, |
|
"learning_rate": 8.014247551202137e-09, |
|
"logits/chosen": -0.12353515625, |
|
"logits/rejected": -0.212890625, |
|
"logps/chosen": -434.0, |
|
"logps/rejected": -320.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.0625, |
|
"rewards/margins": 26.0, |
|
"rewards/rejected": -19.0, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 2.96474358974359, |
|
"grad_norm": 7.742558826164018e-06, |
|
"learning_rate": 6.530127634312852e-09, |
|
"logits/chosen": -0.408203125, |
|
"logits/rejected": 0.1376953125, |
|
"logps/chosen": -460.0, |
|
"logps/rejected": -342.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.125, |
|
"rewards/margins": 27.125, |
|
"rewards/rejected": -20.0, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 2.97275641025641, |
|
"grad_norm": 1.9824516805849423e-05, |
|
"learning_rate": 5.046007717423567e-09, |
|
"logits/chosen": -0.2158203125, |
|
"logits/rejected": 0.1337890625, |
|
"logps/chosen": -432.0, |
|
"logps/rejected": -316.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 6.78125, |
|
"rewards/margins": 26.125, |
|
"rewards/rejected": -19.25, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 2.980769230769231, |
|
"grad_norm": 2.492236779704375e-06, |
|
"learning_rate": 3.561887800534283e-09, |
|
"logits/chosen": -0.3828125, |
|
"logits/rejected": -0.0247802734375, |
|
"logps/chosen": -432.0, |
|
"logps/rejected": -332.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.03125, |
|
"rewards/margins": 27.25, |
|
"rewards/rejected": -20.125, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 2.988782051282051, |
|
"grad_norm": 1.237021776362628e-09, |
|
"learning_rate": 2.0777678836449987e-09, |
|
"logits/chosen": -0.07373046875, |
|
"logits/rejected": 0.0025482177734375, |
|
"logps/chosen": -444.0, |
|
"logps/rejected": -342.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.5, |
|
"rewards/margins": 27.625, |
|
"rewards/rejected": -20.125, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 2.996794871794872, |
|
"grad_norm": 1.15889015211262e-08, |
|
"learning_rate": 5.936479667557139e-10, |
|
"logits/chosen": 0.111328125, |
|
"logits/rejected": 0.265625, |
|
"logps/chosen": -450.0, |
|
"logps/rejected": -342.0, |
|
"loss": 0.0, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 7.625, |
|
"rewards/margins": 28.25, |
|
"rewards/rejected": -20.5, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_logits/chosen": -0.17578125, |
|
"eval_logits/rejected": 0.07421875, |
|
"eval_logps/chosen": -440.0, |
|
"eval_logps/rejected": -342.0, |
|
"eval_loss": 1.3725887981763663e-08, |
|
"eval_rewards/accuracies": 1.0, |
|
"eval_rewards/chosen": 6.5625, |
|
"eval_rewards/margins": 26.5, |
|
"eval_rewards/rejected": -20.0, |
|
"eval_runtime": 32.6727, |
|
"eval_samples_per_second": 6.091, |
|
"eval_steps_per_second": 0.765, |
|
"step": 3744 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"step": 3744, |
|
"total_flos": 0.0, |
|
"train_loss": 0.016824548338845148, |
|
"train_runtime": 10948.0947, |
|
"train_samples_per_second": 2.736, |
|
"train_steps_per_second": 0.342 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 3744, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 0.0, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|