|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.9999374491774566, |
|
"eval_steps": 500, |
|
"global_step": 7993, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 2.5000000000000004e-07, |
|
"loss": 1.5318, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.25e-06, |
|
"loss": 1.6487, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 2.5e-06, |
|
"loss": 1.6098, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 3.75e-06, |
|
"loss": 1.5929, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 5e-06, |
|
"loss": 1.5514, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 6.25e-06, |
|
"loss": 1.5775, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 7.5e-06, |
|
"loss": 1.5792, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 8.75e-06, |
|
"loss": 1.6037, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1e-05, |
|
"loss": 1.5595, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.125e-05, |
|
"loss": 1.509, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.25e-05, |
|
"loss": 1.4634, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.3750000000000002e-05, |
|
"loss": 1.4445, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.5e-05, |
|
"loss": 1.4019, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.6250000000000002e-05, |
|
"loss": 1.3642, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.75e-05, |
|
"loss": 1.3449, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.8750000000000002e-05, |
|
"loss": 1.3908, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2e-05, |
|
"loss": 1.3045, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.125e-05, |
|
"loss": 1.3436, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.25e-05, |
|
"loss": 1.3136, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.375e-05, |
|
"loss": 1.3701, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.5e-05, |
|
"loss": 1.3698, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.625e-05, |
|
"loss": 1.2973, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.7500000000000004e-05, |
|
"loss": 1.32, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.8749999999999997e-05, |
|
"loss": 1.3294, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3e-05, |
|
"loss": 1.3047, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.125e-05, |
|
"loss": 1.348, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.2500000000000004e-05, |
|
"loss": 1.3091, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.375000000000001e-05, |
|
"loss": 1.3314, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.5e-05, |
|
"loss": 1.333, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.625e-05, |
|
"loss": 1.259, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.7500000000000003e-05, |
|
"loss": 1.3241, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.875e-05, |
|
"loss": 1.2295, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4e-05, |
|
"loss": 1.29, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.125e-05, |
|
"loss": 1.3005, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.25e-05, |
|
"loss": 1.2603, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.375e-05, |
|
"loss": 1.3094, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.5e-05, |
|
"loss": 1.2627, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.6250000000000006e-05, |
|
"loss": 1.3644, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.75e-05, |
|
"loss": 1.307, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.875e-05, |
|
"loss": 1.3347, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 5e-05, |
|
"loss": 1.3205, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 5.125e-05, |
|
"loss": 1.2959, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 5.25e-05, |
|
"loss": 1.3157, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 5.375e-05, |
|
"loss": 1.2894, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 5.500000000000001e-05, |
|
"loss": 1.3265, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 5.6250000000000005e-05, |
|
"loss": 1.2413, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 5.7499999999999995e-05, |
|
"loss": 1.2674, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 5.8750000000000005e-05, |
|
"loss": 1.2743, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 6e-05, |
|
"loss": 1.2951, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 6.125000000000001e-05, |
|
"loss": 1.3442, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 6.25e-05, |
|
"loss": 1.2979, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 6.375e-05, |
|
"loss": 1.3029, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 6.500000000000001e-05, |
|
"loss": 1.2811, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 6.625e-05, |
|
"loss": 1.2649, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 6.750000000000001e-05, |
|
"loss": 1.2804, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 6.875e-05, |
|
"loss": 1.2934, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 7e-05, |
|
"loss": 1.2635, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 7.125000000000001e-05, |
|
"loss": 1.3137, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 7.25e-05, |
|
"loss": 1.2495, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 7.375e-05, |
|
"loss": 1.2988, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 7.500000000000001e-05, |
|
"loss": 1.3132, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 7.625e-05, |
|
"loss": 1.3473, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 7.75e-05, |
|
"loss": 1.2857, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 7.875e-05, |
|
"loss": 1.2912, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8e-05, |
|
"loss": 1.3064, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.125000000000001e-05, |
|
"loss": 1.3107, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.25e-05, |
|
"loss": 1.249, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.375e-05, |
|
"loss": 1.3263, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.5e-05, |
|
"loss": 1.3102, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.625000000000001e-05, |
|
"loss": 1.2893, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.75e-05, |
|
"loss": 1.265, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.875e-05, |
|
"loss": 1.2913, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9e-05, |
|
"loss": 1.2721, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.125e-05, |
|
"loss": 1.2602, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.250000000000001e-05, |
|
"loss": 1.3083, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.375e-05, |
|
"loss": 1.2473, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.5e-05, |
|
"loss": 1.2506, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.625000000000001e-05, |
|
"loss": 1.3117, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.75e-05, |
|
"loss": 1.2814, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.875000000000002e-05, |
|
"loss": 1.3298, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0001, |
|
"loss": 1.2115, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00010125, |
|
"loss": 1.2808, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0001025, |
|
"loss": 1.2638, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00010375000000000001, |
|
"loss": 1.3073, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.000105, |
|
"loss": 1.2952, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00010625000000000001, |
|
"loss": 1.242, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0001075, |
|
"loss": 1.2965, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00010875, |
|
"loss": 1.2791, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00011000000000000002, |
|
"loss": 1.3318, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00011125000000000001, |
|
"loss": 1.3506, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00011250000000000001, |
|
"loss": 1.3047, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00011375, |
|
"loss": 1.3418, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00011499999999999999, |
|
"loss": 1.3019, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00011625000000000002, |
|
"loss": 1.264, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00011750000000000001, |
|
"loss": 1.3025, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00011875, |
|
"loss": 1.2805, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00012, |
|
"loss": 1.2584, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00012124999999999999, |
|
"loss": 1.3043, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00012250000000000002, |
|
"loss": 1.2682, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00012375, |
|
"loss": 1.2965, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.000125, |
|
"loss": 1.3153, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00012625, |
|
"loss": 1.3039, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0001275, |
|
"loss": 1.2954, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00012875, |
|
"loss": 1.2447, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00013000000000000002, |
|
"loss": 1.3793, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00013125000000000002, |
|
"loss": 1.461, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0001325, |
|
"loss": 1.3918, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00013375, |
|
"loss": 1.3351, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00013500000000000003, |
|
"loss": 1.3112, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00013625, |
|
"loss": 1.3414, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0001375, |
|
"loss": 1.3578, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00013875, |
|
"loss": 1.3049, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00014, |
|
"loss": 1.2624, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00014125000000000002, |
|
"loss": 1.3392, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00014250000000000002, |
|
"loss": 1.3567, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00014375, |
|
"loss": 1.285, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.000145, |
|
"loss": 1.2717, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00014625, |
|
"loss": 1.3328, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0001475, |
|
"loss": 1.304, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00014875, |
|
"loss": 1.3277, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00015000000000000001, |
|
"loss": 1.3293, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00015125, |
|
"loss": 1.3282, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0001525, |
|
"loss": 1.254, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00015375000000000002, |
|
"loss": 1.2832, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.000155, |
|
"loss": 1.2347, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00015625, |
|
"loss": 1.3137, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0001575, |
|
"loss": 1.2732, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00015875, |
|
"loss": 1.2789, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00016, |
|
"loss": 1.2878, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00016125000000000002, |
|
"loss": 1.2996, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00016250000000000002, |
|
"loss": 1.3068, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00016375, |
|
"loss": 1.3677, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.000165, |
|
"loss": 1.2717, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00016625000000000003, |
|
"loss": 1.293, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0001675, |
|
"loss": 1.3013, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00016875, |
|
"loss": 1.3059, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00017, |
|
"loss": 1.3053, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00017125, |
|
"loss": 1.322, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00017250000000000002, |
|
"loss": 1.3149, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00017375000000000002, |
|
"loss": 1.3125, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.000175, |
|
"loss": 1.2979, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00017625, |
|
"loss": 1.2662, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0001775, |
|
"loss": 1.2689, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00017875, |
|
"loss": 1.2694, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00018, |
|
"loss": 1.3493, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00018125000000000001, |
|
"loss": 1.2074, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0001825, |
|
"loss": 1.2753, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00018375, |
|
"loss": 1.293, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00018500000000000002, |
|
"loss": 1.2943, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00018625, |
|
"loss": 1.218, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0001875, |
|
"loss": 1.3003, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00018875, |
|
"loss": 1.3075, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00019, |
|
"loss": 1.3206, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00019125000000000001, |
|
"loss": 1.341, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00019250000000000002, |
|
"loss": 1.2731, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00019375000000000002, |
|
"loss": 1.2907, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.000195, |
|
"loss": 1.3551, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00019625, |
|
"loss": 1.2429, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00019750000000000003, |
|
"loss": 1.2952, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00019875, |
|
"loss": 1.3111, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0002, |
|
"loss": 1.2955, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00019999976155431637, |
|
"loss": 1.2993, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0001999990462184026, |
|
"loss": 1.2361, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00019999785399567007, |
|
"loss": 1.3347, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00019999618489180437, |
|
"loss": 1.2602, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00019999403891476531, |
|
"loss": 1.2659, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00019999141607478688, |
|
"loss": 1.2905, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0001999883163843772, |
|
"loss": 1.2445, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00019998473985831844, |
|
"loss": 1.2564, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00019998068651366663, |
|
"loss": 1.3213, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00019997615636975198, |
|
"loss": 1.3097, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00019997114944817824, |
|
"loss": 1.3143, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.000199965665772823, |
|
"loss": 1.3529, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00019995970536983748, |
|
"loss": 1.2879, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0001999532682676463, |
|
"loss": 1.2501, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00019994635449694745, |
|
"loss": 1.2621, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0001999389640907121, |
|
"loss": 1.2709, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0001999310970841845, |
|
"loss": 1.2617, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00019992275351488165, |
|
"loss": 1.3019, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00019991393342259336, |
|
"loss": 1.3013, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0001999046368493819, |
|
"loss": 1.3059, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00019989486383958178, |
|
"loss": 1.2953, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00019988461443979966, |
|
"loss": 1.2599, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0001998738886989141, |
|
"loss": 1.2505, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0001998626866680751, |
|
"loss": 1.2137, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0001998510084007043, |
|
"loss": 1.2453, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00019983885395249434, |
|
"loss": 1.2467, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00019982622338140865, |
|
"loss": 1.3101, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0001998131167476814, |
|
"loss": 1.2404, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00019979953411381697, |
|
"loss": 1.2479, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0001997854755445898, |
|
"loss": 1.2448, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00019977094110704396, |
|
"loss": 1.2714, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0001997559308704929, |
|
"loss": 1.2888, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00019974044490651924, |
|
"loss": 1.2583, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0001997244832889741, |
|
"loss": 1.2803, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00019970804609397713, |
|
"loss": 1.2661, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00019969113339991587, |
|
"loss": 1.2777, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00019967374528744548, |
|
"loss": 1.2551, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0001996558818394884, |
|
"loss": 1.299, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00019963754314123385, |
|
"loss": 1.2646, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00019961872928013749, |
|
"loss": 1.2293, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.000199599440345921, |
|
"loss": 1.2259, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00019957967643057165, |
|
"loss": 1.2242, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00019955943762834188, |
|
"loss": 1.3307, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00019953872403574873, |
|
"loss": 1.2388, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0001995175357515736, |
|
"loss": 1.26, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0001994958728768615, |
|
"loss": 1.2537, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0001994737355149209, |
|
"loss": 1.2528, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00019945112377132293, |
|
"loss": 1.2291, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00019942803775390103, |
|
"loss": 1.2965, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00019940447757275045, |
|
"loss": 1.2564, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00019938044334022763, |
|
"loss": 1.2185, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00019935593517094978, |
|
"loss": 1.2738, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0001993309531817942, |
|
"loss": 1.3172, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00019930549749189793, |
|
"loss": 1.2896, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00019927956822265688, |
|
"loss": 1.271, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0001992531654977255, |
|
"loss": 1.3092, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00019922628944301612, |
|
"loss": 1.2197, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00019919894018669834, |
|
"loss": 1.3077, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0001991711178591984, |
|
"loss": 1.2618, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00019914282259319853, |
|
"loss": 1.2493, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00019911405452363644, |
|
"loss": 1.267, |
|
"step": 1105 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0001990848137877046, |
|
"loss": 1.3094, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00019905510052484955, |
|
"loss": 1.2999, |
|
"step": 1115 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00019902491487677125, |
|
"loss": 1.2972, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00019899425698742244, |
|
"loss": 1.2145, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00019896312700300796, |
|
"loss": 1.2053, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00019893152507198406, |
|
"loss": 1.234, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00019889945134505754, |
|
"loss": 1.2515, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00019886690597518532, |
|
"loss": 1.2671, |
|
"step": 1145 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0001988338891175734, |
|
"loss": 1.3104, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00019880040092967635, |
|
"loss": 1.3029, |
|
"step": 1155 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00019876644157119642, |
|
"loss": 1.2945, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.0001987320112040829, |
|
"loss": 1.2297, |
|
"step": 1165 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.0001986971099925312, |
|
"loss": 1.2432, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00019866173810298223, |
|
"loss": 1.2036, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00019862589570412143, |
|
"loss": 1.2547, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00019858958296687816, |
|
"loss": 1.2464, |
|
"step": 1185 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00019855280006442466, |
|
"loss": 1.2526, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00019851554717217545, |
|
"loss": 1.2482, |
|
"step": 1195 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00019847782446778638, |
|
"loss": 1.3014, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00019843963213115373, |
|
"loss": 1.2236, |
|
"step": 1205 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00019840097034441346, |
|
"loss": 1.2094, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00019836183929194034, |
|
"loss": 1.2747, |
|
"step": 1215 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00019832223916034692, |
|
"loss": 1.2774, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.0001982821701384829, |
|
"loss": 1.2367, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00019824163241743387, |
|
"loss": 1.2305, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.0001982006261905208, |
|
"loss": 1.2255, |
|
"step": 1235 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00019815915165329883, |
|
"loss": 1.3117, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00019811720900355645, |
|
"loss": 1.2853, |
|
"step": 1245 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0001980747984413145, |
|
"loss": 1.2679, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00019803192016882532, |
|
"loss": 1.2815, |
|
"step": 1255 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0001979885743905717, |
|
"loss": 1.2328, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0001979447613132659, |
|
"loss": 1.2808, |
|
"step": 1265 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00019790048114584868, |
|
"loss": 1.3052, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0001978557340994884, |
|
"loss": 1.277, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00019781052038757982, |
|
"loss": 1.2854, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00019776484022574322, |
|
"loss": 1.3023, |
|
"step": 1285 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00019771869383182334, |
|
"loss": 1.267, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0001976720814258884, |
|
"loss": 1.2748, |
|
"step": 1295 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0001976250032302289, |
|
"loss": 1.3018, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00019757745946935668, |
|
"loss": 1.319, |
|
"step": 1305 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00019752945037000388, |
|
"loss": 1.2665, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0001974809761611217, |
|
"loss": 1.2878, |
|
"step": 1315 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0001974320370738795, |
|
"loss": 1.2789, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00019738263334166353, |
|
"loss": 1.2597, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00019733276520007592, |
|
"loss": 1.2327, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0001972824328869336, |
|
"loss": 1.2592, |
|
"step": 1335 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00019723163664226692, |
|
"loss": 1.2907, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00019718037670831887, |
|
"loss": 1.2438, |
|
"step": 1345 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00019712865332954357, |
|
"loss": 1.302, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00019707646675260545, |
|
"loss": 1.314, |
|
"step": 1355 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00019702381722637772, |
|
"loss": 1.3027, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00019697070500194143, |
|
"loss": 1.3051, |
|
"step": 1365 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00019691713033258423, |
|
"loss": 1.2989, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00019686309347379907, |
|
"loss": 1.2897, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00019680859468328306, |
|
"loss": 1.3257, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00019675363422093622, |
|
"loss": 1.3109, |
|
"step": 1385 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0001966982123488603, |
|
"loss": 1.333, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00019664232933135734, |
|
"loss": 1.419, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00019658598543492866, |
|
"loss": 1.3776, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00019652918092827348, |
|
"loss": 1.3408, |
|
"step": 1405 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00019647191608228757, |
|
"loss": 1.4326, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00019641419117006202, |
|
"loss": 1.3389, |
|
"step": 1415 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00019635600646688192, |
|
"loss": 1.3319, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00019629736225022516, |
|
"loss": 1.38, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00019623825879976095, |
|
"loss": 1.3225, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00019617869639734848, |
|
"loss": 1.3411, |
|
"step": 1435 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00019611867532703573, |
|
"loss": 1.3312, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00019605819587505807, |
|
"loss": 1.4009, |
|
"step": 1445 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0001959972583298367, |
|
"loss": 1.3042, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0001959358629819775, |
|
"loss": 1.3582, |
|
"step": 1455 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00019587401012426966, |
|
"loss": 1.3092, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0001958117000516841, |
|
"loss": 1.3469, |
|
"step": 1465 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00019574893306137215, |
|
"loss": 1.4079, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00019568570945266414, |
|
"loss": 1.3684, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0001956220295270681, |
|
"loss": 1.3162, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00019555789358826798, |
|
"loss": 1.349, |
|
"step": 1485 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00019549330194212265, |
|
"loss": 1.3493, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.000195428254896664, |
|
"loss": 1.3412, |
|
"step": 1495 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00019536275276209585, |
|
"loss": 1.4347, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00019529679585079216, |
|
"loss": 1.3623, |
|
"step": 1505 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00019523038447729577, |
|
"loss": 1.3614, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0001951635189583168, |
|
"loss": 1.2906, |
|
"step": 1515 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00019509619961273115, |
|
"loss": 1.3956, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0001950284267615789, |
|
"loss": 1.3891, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.000194960200728063, |
|
"loss": 1.3916, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0001948915218375475, |
|
"loss": 1.3651, |
|
"step": 1535 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00019482239041755607, |
|
"loss": 1.3835, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00019475280679777051, |
|
"loss": 1.4017, |
|
"step": 1545 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0001946827713100291, |
|
"loss": 1.4046, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00019461228428832502, |
|
"loss": 1.3561, |
|
"step": 1555 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00019454134606880478, |
|
"loss": 1.3532, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00019446995698976663, |
|
"loss": 1.3341, |
|
"step": 1565 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00019439811739165895, |
|
"loss": 1.3825, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00019432582761707857, |
|
"loss": 1.3531, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00019425308801076914, |
|
"loss": 1.3161, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00019417989891961965, |
|
"loss": 1.4296, |
|
"step": 1585 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00019410626069266246, |
|
"loss": 1.5025, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.000194032173681072, |
|
"loss": 1.4514, |
|
"step": 1595 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00019395763823816281, |
|
"loss": 1.4209, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00019388265471938798, |
|
"loss": 1.3903, |
|
"step": 1605 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00019380722348233742, |
|
"loss": 1.3339, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.0001937313448867362, |
|
"loss": 1.3525, |
|
"step": 1615 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00019365501929444277, |
|
"loss": 1.4051, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00019357824706944733, |
|
"loss": 1.3953, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00019350102857786995, |
|
"loss": 1.41, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.000193423364187959, |
|
"loss": 1.4805, |
|
"step": 1635 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00019334525427008924, |
|
"loss": 1.4902, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00019326669919676008, |
|
"loss": 1.5222, |
|
"step": 1645 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00019318769934259392, |
|
"loss": 1.4062, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00019310825508433427, |
|
"loss": 1.4796, |
|
"step": 1655 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00019302836680084389, |
|
"loss": 1.4254, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.0001929480348731031, |
|
"loss": 1.5454, |
|
"step": 1665 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.000192867259684208, |
|
"loss": 1.3524, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00019278604161936843, |
|
"loss": 1.4679, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00019270438106590635, |
|
"loss": 1.3318, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00019262227841325386, |
|
"loss": 1.398, |
|
"step": 1685 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00019253973405295146, |
|
"loss": 1.4586, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00019245674837864608, |
|
"loss": 1.4384, |
|
"step": 1695 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00019237332178608925, |
|
"loss": 1.4264, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00019228945467313514, |
|
"loss": 1.3609, |
|
"step": 1705 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00019220514743973878, |
|
"loss": 1.3665, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00019212040048795413, |
|
"loss": 1.3037, |
|
"step": 1715 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00019203521422193204, |
|
"loss": 1.3872, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00019194958904791848, |
|
"loss": 1.2928, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00019186352537425253, |
|
"loss": 1.3462, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00019177702361136437, |
|
"loss": 1.3604, |
|
"step": 1735 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0001916900841717735, |
|
"loss": 1.3701, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00019160270747008658, |
|
"loss": 1.321, |
|
"step": 1745 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00019151489392299552, |
|
"loss": 1.3366, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0001914266439492756, |
|
"loss": 1.2733, |
|
"step": 1755 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00019133795796978328, |
|
"loss": 1.3367, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0001912488364074544, |
|
"loss": 1.3348, |
|
"step": 1765 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00019115927968730193, |
|
"loss": 1.3141, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00019106928823641416, |
|
"loss": 1.3085, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00019097886248395257, |
|
"loss": 1.4148, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00019088800286114977, |
|
"loss": 1.3092, |
|
"step": 1785 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00019079670980130743, |
|
"loss": 1.4045, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00019070498373979428, |
|
"loss": 1.39, |
|
"step": 1795 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00019061282511404404, |
|
"loss": 1.3483, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00019052023436355313, |
|
"loss": 1.3367, |
|
"step": 1805 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00019042721192987893, |
|
"loss": 1.3454, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.0001903337582566374, |
|
"loss": 1.3528, |
|
"step": 1815 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.000190239873789501, |
|
"loss": 1.3359, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00019014555897619665, |
|
"loss": 1.3112, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00019005081426650355, |
|
"loss": 1.3108, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00018995564011225108, |
|
"loss": 1.3212, |
|
"step": 1835 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00018986003696731655, |
|
"loss": 1.3073, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.0001897640052876231, |
|
"loss": 1.3589, |
|
"step": 1845 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00018966754553113751, |
|
"loss": 1.3726, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00018957065815786807, |
|
"loss": 1.3466, |
|
"step": 1855 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00018947334362986223, |
|
"loss": 1.2642, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00018937560241120461, |
|
"loss": 1.3765, |
|
"step": 1865 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.0001892774349680147, |
|
"loss": 1.3399, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00018917884176844444, |
|
"loss": 1.407, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.0001890798232826764, |
|
"loss": 1.2707, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00018898037998292116, |
|
"loss": 1.3525, |
|
"step": 1885 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.0001888805123434152, |
|
"loss": 1.3684, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00018878022084041867, |
|
"loss": 1.3548, |
|
"step": 1895 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00018867950595221312, |
|
"loss": 1.3264, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00018857836815909916, |
|
"loss": 1.3439, |
|
"step": 1905 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00018847680794339418, |
|
"loss": 1.3919, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00018837482578943008, |
|
"loss": 1.3494, |
|
"step": 1915 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00018827242218355097, |
|
"loss": 1.3334, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00018816959761411078, |
|
"loss": 1.3377, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00018806635257147099, |
|
"loss": 1.3578, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00018796268754799834, |
|
"loss": 1.333, |
|
"step": 1935 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00018785860303806233, |
|
"loss": 1.3023, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00018775409953803308, |
|
"loss": 1.38, |
|
"step": 1945 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00018764917754627867, |
|
"loss": 1.315, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00018754383756316306, |
|
"loss": 1.3921, |
|
"step": 1955 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00018743808009104356, |
|
"loss": 1.3223, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0001873319056342684, |
|
"loss": 1.3998, |
|
"step": 1965 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00018722531469917434, |
|
"loss": 1.3031, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00018711830779408447, |
|
"loss": 1.3507, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0001870108854293054, |
|
"loss": 1.3109, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00018690304811712512, |
|
"loss": 1.3071, |
|
"step": 1985 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0001867947963718105, |
|
"loss": 1.3857, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00018668613070960474, |
|
"loss": 1.3404, |
|
"step": 1995 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00018657705164872504, |
|
"loss": 1.3235, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00018646755970935994, |
|
"loss": 1.2963, |
|
"step": 2005 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0001863576554136671, |
|
"loss": 1.3747, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00018624733928577066, |
|
"loss": 1.3332, |
|
"step": 2015 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00018613661185175862, |
|
"loss": 1.2974, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00018602547363968063, |
|
"loss": 1.295, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0001859139251795452, |
|
"loss": 1.2899, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00018580196700331732, |
|
"loss": 1.3351, |
|
"step": 2035 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0001856895996449158, |
|
"loss": 1.2992, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00018557682364021098, |
|
"loss": 1.2951, |
|
"step": 2045 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0001854636395270218, |
|
"loss": 1.3764, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00018535004784511362, |
|
"loss": 1.2796, |
|
"step": 2055 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00018523604913619525, |
|
"loss": 1.2997, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00018512164394391678, |
|
"loss": 1.3098, |
|
"step": 2065 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00018500683281386663, |
|
"loss": 1.3593, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00018489161629356925, |
|
"loss": 1.2587, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00018477599493248222, |
|
"loss": 1.2977, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00018465996928199383, |
|
"loss": 1.3167, |
|
"step": 2085 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0001845435398954204, |
|
"loss": 1.2721, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00018442670732800364, |
|
"loss": 1.3018, |
|
"step": 2095 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00018430947213690795, |
|
"loss": 1.358, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00018419183488121784, |
|
"loss": 1.3212, |
|
"step": 2105 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00018407379612193526, |
|
"loss": 1.2565, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0001839553564219768, |
|
"loss": 1.2673, |
|
"step": 2115 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00018383651634617122, |
|
"loss": 1.3397, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.0001837172764612566, |
|
"loss": 1.3978, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.0001835976373358776, |
|
"loss": 1.3157, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00018347759954058293, |
|
"loss": 1.3174, |
|
"step": 2135 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.0001833571636478224, |
|
"loss": 1.3356, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00018323633023194447, |
|
"loss": 1.2936, |
|
"step": 2145 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00018311509986919324, |
|
"loss": 1.3005, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00018299347313770583, |
|
"loss": 1.2921, |
|
"step": 2155 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00018287145061750962, |
|
"loss": 1.3285, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00018274903289051947, |
|
"loss": 1.32, |
|
"step": 2165 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.000182626220540535, |
|
"loss": 1.301, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00018250301415323765, |
|
"loss": 1.3004, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00018237941431618806, |
|
"loss": 1.3383, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.0001822554216188232, |
|
"loss": 1.2559, |
|
"step": 2185 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00018213103665245357, |
|
"loss": 1.2938, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00018200626001026025, |
|
"loss": 1.3444, |
|
"step": 2195 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00018188109228729234, |
|
"loss": 1.2846, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00018175553408046385, |
|
"loss": 1.2924, |
|
"step": 2205 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0001816295859885511, |
|
"loss": 1.275, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00018150324861218966, |
|
"loss": 1.3059, |
|
"step": 2215 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0001813765225538715, |
|
"loss": 1.3894, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00018124940841794231, |
|
"loss": 1.2676, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00018112190681059842, |
|
"loss": 1.2966, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.000180994018339884, |
|
"loss": 1.3358, |
|
"step": 2235 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00018086574361568812, |
|
"loss": 1.287, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0001807370832497418, |
|
"loss": 1.3234, |
|
"step": 2245 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00018060803785561533, |
|
"loss": 1.3459, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.000180478608048715, |
|
"loss": 1.311, |
|
"step": 2255 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00018034879444628037, |
|
"loss": 1.2919, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00018021859766738133, |
|
"loss": 1.3056, |
|
"step": 2265 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0001800880183329151, |
|
"loss": 1.304, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00017995705706560318, |
|
"loss": 1.3483, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0001798257144899886, |
|
"loss": 1.35, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00017969399123243278, |
|
"loss": 1.3391, |
|
"step": 2285 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00017956188792111251, |
|
"loss": 1.3009, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00017942940518601714, |
|
"loss": 1.3812, |
|
"step": 2295 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00017929654365894535, |
|
"loss": 1.3388, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00017916330397350233, |
|
"loss": 1.3627, |
|
"step": 2305 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0001790296867650966, |
|
"loss": 1.2894, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0001788956926709371, |
|
"loss": 1.2924, |
|
"step": 2315 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0001787613223300301, |
|
"loss": 1.3537, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00017862657638317617, |
|
"loss": 1.3264, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0001784914554729671, |
|
"loss": 1.3742, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00017835596024378284, |
|
"loss": 1.3097, |
|
"step": 2335 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00017822009134178842, |
|
"loss": 1.3545, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00017808384941493094, |
|
"loss": 1.313, |
|
"step": 2345 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0001779472351129363, |
|
"loss": 1.2508, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00017781024908730645, |
|
"loss": 1.341, |
|
"step": 2355 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00017767289199131585, |
|
"loss": 1.3074, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00017753516448000865, |
|
"loss": 1.2625, |
|
"step": 2365 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00017739706721019543, |
|
"loss": 1.3473, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00017725860084045016, |
|
"loss": 1.3443, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.000177119766031107, |
|
"loss": 1.3527, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0001769805634442572, |
|
"loss": 1.326, |
|
"step": 2385 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00017684099374374584, |
|
"loss": 1.3769, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00017670105759516882, |
|
"loss": 1.3567, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00017656075566586948, |
|
"loss": 1.3323, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0001764200886249357, |
|
"loss": 1.3267, |
|
"step": 2405 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0001762790571431964, |
|
"loss": 1.3117, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00017613766189321855, |
|
"loss": 1.3368, |
|
"step": 2415 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0001759959035493039, |
|
"loss": 1.3731, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00017585378278748572, |
|
"loss": 1.3093, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0001757113002855257, |
|
"loss": 1.3548, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00017556845672291055, |
|
"loss": 1.3432, |
|
"step": 2435 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00017542525278084893, |
|
"loss": 1.3199, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00017528168914226802, |
|
"loss": 1.321, |
|
"step": 2445 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00017513776649181052, |
|
"loss": 1.3439, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.000174993485515831, |
|
"loss": 1.2811, |
|
"step": 2455 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.0001748488469023931, |
|
"loss": 1.3078, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.0001747038513412658, |
|
"loss": 1.3093, |
|
"step": 2465 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00017455849952392044, |
|
"loss": 1.303, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00017441279214352727, |
|
"loss": 1.3329, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00017426672989495222, |
|
"loss": 1.3489, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00017412031347475358, |
|
"loss": 1.234, |
|
"step": 2485 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00017397354358117856, |
|
"loss": 1.3548, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00017382642091416016, |
|
"loss": 1.323, |
|
"step": 2495 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.0001736789461753136, |
|
"loss": 1.3244, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00017353112006793331, |
|
"loss": 1.2941, |
|
"step": 2505 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00017338294329698913, |
|
"loss": 1.3132, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00017323441656912333, |
|
"loss": 1.3187, |
|
"step": 2515 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00017308554059264702, |
|
"loss": 1.3282, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00017293631607753692, |
|
"loss": 1.2838, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00017278674373543185, |
|
"loss": 1.2631, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00017263682427962938, |
|
"loss": 1.3203, |
|
"step": 2535 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00017248655842508245, |
|
"loss": 1.3168, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00017233594688839595, |
|
"loss": 1.2831, |
|
"step": 2545 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00017218499038782335, |
|
"loss": 1.7136, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.0001720336896432631, |
|
"loss": 3.4121, |
|
"step": 2555 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00017188204537625543, |
|
"loss": 2.0298, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00017173005830997874, |
|
"loss": 1.6072, |
|
"step": 2565 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00017157772916924624, |
|
"loss": 1.4499, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00017142505868050244, |
|
"loss": 1.5148, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00017127204757181973, |
|
"loss": 1.6338, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00017111869657289485, |
|
"loss": 1.4959, |
|
"step": 2585 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00017096500641504552, |
|
"loss": 1.436, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00017081097783120682, |
|
"loss": 1.5545, |
|
"step": 2595 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00017065661155592777, |
|
"loss": 1.5012, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0001705019083253678, |
|
"loss": 1.4849, |
|
"step": 2605 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00017034686887729326, |
|
"loss": 1.4492, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0001701914939510739, |
|
"loss": 1.4702, |
|
"step": 2615 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0001700357842876793, |
|
"loss": 1.4252, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00016987974062967549, |
|
"loss": 1.4797, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00016972336372122113, |
|
"loss": 1.4073, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0001695666543080642, |
|
"loss": 1.438, |
|
"step": 2635 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0001694096131375384, |
|
"loss": 1.4517, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00016925224095855947, |
|
"loss": 1.5037, |
|
"step": 2645 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00016909453852162178, |
|
"loss": 1.4196, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00016893650657879463, |
|
"loss": 1.4604, |
|
"step": 2655 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0001687781458837187, |
|
"loss": 1.483, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00016861945719160248, |
|
"loss": 1.446, |
|
"step": 2665 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00016846044125921867, |
|
"loss": 1.372, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00016830109884490049, |
|
"loss": 1.41, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00016814143070853813, |
|
"loss": 1.4272, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00016798143761157525, |
|
"loss": 1.5629, |
|
"step": 2685 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00016782112031700502, |
|
"loss": 1.7761, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0001676604795893668, |
|
"loss": 1.8058, |
|
"step": 2695 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00016749951619474243, |
|
"loss": 1.9839, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00016733823090075232, |
|
"loss": 1.7405, |
|
"step": 2705 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0001671766244765522, |
|
"loss": 1.7059, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00016701469769282913, |
|
"loss": 1.7428, |
|
"step": 2715 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00016685245132179794, |
|
"loss": 1.7563, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00016668988613719764, |
|
"loss": 1.8555, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00016652700291428744, |
|
"loss": 2.5574, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00016636380242984346, |
|
"loss": 2.5235, |
|
"step": 2735 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0001662002854621547, |
|
"loss": 2.0932, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00016603645279101952, |
|
"loss": 2.0514, |
|
"step": 2745 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00016587230519774169, |
|
"loss": 2.2654, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00016570784346512694, |
|
"loss": 2.3945, |
|
"step": 2755 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00016554306837747912, |
|
"loss": 2.0086, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.0001653779807205963, |
|
"loss": 2.1143, |
|
"step": 2765 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.0001652125812817674, |
|
"loss": 2.6482, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00016504687084976794, |
|
"loss": 3.0041, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00016488085021485675, |
|
"loss": 3.7007, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00016471452016877188, |
|
"loss": 3.0461, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00016454788150472693, |
|
"loss": 2.5356, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00016438093501740735, |
|
"loss": 2.2243, |
|
"step": 2795 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.0001642136815029665, |
|
"loss": 2.3492, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00016404612175902197, |
|
"loss": 2.6971, |
|
"step": 2805 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00016387825658465167, |
|
"loss": 2.4074, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00016371008678039018, |
|
"loss": 2.2175, |
|
"step": 2815 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00016354161314822475, |
|
"loss": 2.0891, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00016337283649159158, |
|
"loss": 2.0217, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00016320375761537202, |
|
"loss": 2.0209, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00016303437732588858, |
|
"loss": 2.0741, |
|
"step": 2835 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00016286469643090126, |
|
"loss": 1.9469, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0001626947157396036, |
|
"loss": 1.8604, |
|
"step": 2845 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0001625244360626189, |
|
"loss": 1.9221, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0001623538582119961, |
|
"loss": 1.9604, |
|
"step": 2855 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0001621829830012064, |
|
"loss": 2.1039, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0001620118112451388, |
|
"loss": 2.0241, |
|
"step": 2865 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00016184034376009673, |
|
"loss": 2.0757, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00016166858136379376, |
|
"loss": 2.1003, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00016149652487534993, |
|
"loss": 2.0454, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00016132417511528782, |
|
"loss": 2.0411, |
|
"step": 2885 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00016115153290552855, |
|
"loss": 2.0486, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0001609785990693879, |
|
"loss": 1.857, |
|
"step": 2895 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00016080537443157237, |
|
"loss": 1.8482, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00016063185981817537, |
|
"loss": 1.9085, |
|
"step": 2905 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00016045805605667307, |
|
"loss": 1.8315, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0001602839639759206, |
|
"loss": 1.7705, |
|
"step": 2915 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00016010958440614812, |
|
"loss": 1.7341, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0001599349181789567, |
|
"loss": 1.6872, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0001597599661273145, |
|
"loss": 1.6284, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00015958472908555272, |
|
"loss": 1.7347, |
|
"step": 2935 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0001594092078893618, |
|
"loss": 1.6812, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00015923340337578706, |
|
"loss": 1.7171, |
|
"step": 2945 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00015905731638322506, |
|
"loss": 1.7352, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0001588809477514195, |
|
"loss": 1.7436, |
|
"step": 2955 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00015870429832145713, |
|
"loss": 1.8703, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00015852736893576385, |
|
"loss": 1.8567, |
|
"step": 2965 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00015835016043810066, |
|
"loss": 1.7835, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00015817267367355953, |
|
"loss": 1.8888, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00015799490948855956, |
|
"loss": 1.8925, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00015781686873084276, |
|
"loss": 1.899, |
|
"step": 2985 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00015763855224947018, |
|
"loss": 1.7718, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0001574599608948177, |
|
"loss": 1.7478, |
|
"step": 2995 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00015728109551857203, |
|
"loss": 1.753, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00015710195697372684, |
|
"loss": 1.7541, |
|
"step": 3005 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00015692254611457824, |
|
"loss": 1.7187, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00015674286379672121, |
|
"loss": 1.6781, |
|
"step": 3015 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0001565629108770452, |
|
"loss": 1.657, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0001563826882137301, |
|
"loss": 1.6209, |
|
"step": 3025 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00015620219666624233, |
|
"loss": 1.6512, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00015602143709533042, |
|
"loss": 1.7073, |
|
"step": 3035 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00015584041036302118, |
|
"loss": 1.6481, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00015565911733261546, |
|
"loss": 1.7618, |
|
"step": 3045 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00015547755886868408, |
|
"loss": 1.8099, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00015529573583706374, |
|
"loss": 1.7233, |
|
"step": 3055 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00015511364910485268, |
|
"loss": 1.8146, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00015493129954040685, |
|
"loss": 1.8626, |
|
"step": 3065 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00015474868801333558, |
|
"loss": 1.8224, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00015456581539449754, |
|
"loss": 1.6986, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00015438268255599632, |
|
"loss": 1.8571, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00015419929037117675, |
|
"loss": 1.7954, |
|
"step": 3085 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00015401563971462027, |
|
"loss": 1.7512, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00015383173146214102, |
|
"loss": 1.8419, |
|
"step": 3095 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00015364756649078159, |
|
"loss": 1.7217, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00015346314567880878, |
|
"loss": 1.925, |
|
"step": 3105 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00015327846990570958, |
|
"loss": 6.4267, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0001530935400521868, |
|
"loss": 3.3311, |
|
"step": 3115 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0001529083570001549, |
|
"loss": 2.3061, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00015272292163273594, |
|
"loss": 2.5436, |
|
"step": 3125 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0001525372348342551, |
|
"loss": 2.2782, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00015235129749023674, |
|
"loss": 2.1674, |
|
"step": 3135 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00015216511048740005, |
|
"loss": 2.0804, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00015197867471365467, |
|
"loss": 2.119, |
|
"step": 3145 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00015179199105809675, |
|
"loss": 2.0452, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0001516050604110046, |
|
"loss": 2.0013, |
|
"step": 3155 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00015141788366383423, |
|
"loss": 1.9956, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00015123046170921543, |
|
"loss": 1.8451, |
|
"step": 3165 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00015104279544094737, |
|
"loss": 1.9065, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00015085488575399417, |
|
"loss": 1.8467, |
|
"step": 3175 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00015066673354448105, |
|
"loss": 2.0355, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00015047833970968953, |
|
"loss": 2.319, |
|
"step": 3185 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.0001502897051480536, |
|
"loss": 2.4123, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.0001501008307591552, |
|
"loss": 2.3931, |
|
"step": 3195 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00014991171744372, |
|
"loss": 2.2739, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.000149722366103613, |
|
"loss": 2.2935, |
|
"step": 3205 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00014953277764183452, |
|
"loss": 2.1345, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00014934295296251547, |
|
"loss": 2.0865, |
|
"step": 3215 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00014915289297091338, |
|
"loss": 2.095, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00014896259857340795, |
|
"loss": 1.8762, |
|
"step": 3225 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00014877207067749674, |
|
"loss": 2.0121, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00014858131019179082, |
|
"loss": 2.1179, |
|
"step": 3235 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00014839031802601048, |
|
"loss": 2.2072, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00014819909509098094, |
|
"loss": 2.2156, |
|
"step": 3245 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00014800764229862774, |
|
"loss": 2.2413, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00014781596056197282, |
|
"loss": 2.1791, |
|
"step": 3255 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0001476240507951298, |
|
"loss": 2.1746, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0001474319139132998, |
|
"loss": 2.1588, |
|
"step": 3265 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.000147239550832767, |
|
"loss": 2.1877, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00014704696247089437, |
|
"loss": 2.1649, |
|
"step": 3275 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00014685414974611912, |
|
"loss": 2.2178, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00014666111357794853, |
|
"loss": 2.084, |
|
"step": 3285 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00014646785488695545, |
|
"loss": 2.0663, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00014627437459477382, |
|
"loss": 1.9301, |
|
"step": 3295 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00014608067362409453, |
|
"loss": 1.8838, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00014588675289866073, |
|
"loss": 2.0045, |
|
"step": 3305 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00014569261334326363, |
|
"loss": 2.0974, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00014549825588373804, |
|
"loss": 2.1131, |
|
"step": 3315 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00014530368144695784, |
|
"loss": 1.9693, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00014510889096083177, |
|
"loss": 1.9283, |
|
"step": 3325 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00014491388535429887, |
|
"loss": 1.892, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00014471866555732397, |
|
"loss": 1.9198, |
|
"step": 3335 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0001445232325008935, |
|
"loss": 1.821, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0001443275871170108, |
|
"loss": 1.7661, |
|
"step": 3345 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00014413173033869182, |
|
"loss": 1.8564, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00014393566309996064, |
|
"loss": 1.846, |
|
"step": 3355 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00014373938633584495, |
|
"loss": 1.8243, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00014354290098237175, |
|
"loss": 1.8947, |
|
"step": 3365 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00014334620797656274, |
|
"loss": 1.8338, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0001431493082564298, |
|
"loss": 1.9131, |
|
"step": 3375 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00014295220276097076, |
|
"loss": 1.7864, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00014275489243016472, |
|
"loss": 1.839, |
|
"step": 3385 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00014255737820496759, |
|
"loss": 1.785, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00014235966102730765, |
|
"loss": 1.7492, |
|
"step": 3395 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.0001421617418400811, |
|
"loss": 1.767, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.0001419636215871474, |
|
"loss": 1.8241, |
|
"step": 3405 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00014176530121332495, |
|
"loss": 1.8174, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00014156678166438652, |
|
"loss": 1.7312, |
|
"step": 3415 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00014136806388705464, |
|
"loss": 1.7921, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00014116914882899732, |
|
"loss": 1.9462, |
|
"step": 3425 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00014097003743882326, |
|
"loss": 1.9426, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00014077073066607744, |
|
"loss": 1.8433, |
|
"step": 3435 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00014057122946123672, |
|
"loss": 1.8512, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00014037153477570514, |
|
"loss": 1.8083, |
|
"step": 3445 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00014017164756180935, |
|
"loss": 1.7582, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00013997156877279426, |
|
"loss": 1.769, |
|
"step": 3455 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00013977129936281835, |
|
"loss": 1.7905, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.0001395708402869491, |
|
"loss": 1.781, |
|
"step": 3465 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00013937019250115856, |
|
"loss": 1.7939, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00013916935696231872, |
|
"loss": 1.7546, |
|
"step": 3475 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00013896833462819693, |
|
"loss": 1.7585, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0001387671264574513, |
|
"loss": 1.731, |
|
"step": 3485 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00013856573340962627, |
|
"loss": 1.7361, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00013836415644514788, |
|
"loss": 1.7356, |
|
"step": 3495 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00013816239652531928, |
|
"loss": 1.7827, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0001379604546123161, |
|
"loss": 1.7072, |
|
"step": 3505 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0001377583316691819, |
|
"loss": 1.8108, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0001375560286598236, |
|
"loss": 1.7169, |
|
"step": 3515 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00013735354654900665, |
|
"loss": 1.9402, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00013715088630235087, |
|
"loss": 1.7921, |
|
"step": 3525 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00013694804888632545, |
|
"loss": 1.793, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00013674503526824453, |
|
"loss": 1.7958, |
|
"step": 3535 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00013654184641626252, |
|
"loss": 1.7456, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0001363384832993695, |
|
"loss": 1.9234, |
|
"step": 3545 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0001361349468873866, |
|
"loss": 1.8635, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00013593123815096146, |
|
"loss": 1.8212, |
|
"step": 3555 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00013572735806156343, |
|
"loss": 1.8341, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00013552330759147903, |
|
"loss": 1.7364, |
|
"step": 3565 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.0001353190877138074, |
|
"loss": 1.7146, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.0001351146994024554, |
|
"loss": 1.5903, |
|
"step": 3575 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00013491014363213336, |
|
"loss": 1.6464, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00013470542137835004, |
|
"loss": 1.632, |
|
"step": 3585 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00013450053361740814, |
|
"loss": 1.6496, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.0001342954813263998, |
|
"loss": 1.6609, |
|
"step": 3595 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00013409026548320165, |
|
"loss": 1.7025, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00013388488706647032, |
|
"loss": 1.6997, |
|
"step": 3605 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00013367934705563777, |
|
"loss": 1.7278, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00013347364643090655, |
|
"loss": 1.7218, |
|
"step": 3615 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00013326778617324517, |
|
"loss": 1.6857, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.0001330617672643835, |
|
"loss": 1.6586, |
|
"step": 3625 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00013285559068680784, |
|
"loss": 1.6583, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00013264925742375656, |
|
"loss": 1.6609, |
|
"step": 3635 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00013244276845921513, |
|
"loss": 1.6422, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.0001322361247779116, |
|
"loss": 1.6625, |
|
"step": 3645 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.0001320293273653119, |
|
"loss": 1.6048, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00013182237720761497, |
|
"loss": 1.6052, |
|
"step": 3655 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00013161527529174827, |
|
"loss": 1.6401, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.000131408022605363, |
|
"loss": 1.7834, |
|
"step": 3665 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00013120062013682928, |
|
"loss": 1.6526, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.0001309930688752316, |
|
"loss": 1.6975, |
|
"step": 3675 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.000130785369810364, |
|
"loss": 1.6943, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00013057752393272538, |
|
"loss": 1.6672, |
|
"step": 3685 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00013036953223351477, |
|
"loss": 1.6581, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00013016139570462673, |
|
"loss": 1.6109, |
|
"step": 3695 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.0001299531153386463, |
|
"loss": 1.6071, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00012974469212884458, |
|
"loss": 1.5724, |
|
"step": 3705 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00012953612706917387, |
|
"loss": 1.6346, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00012932742115426296, |
|
"loss": 1.5854, |
|
"step": 3715 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00012911857537941233, |
|
"loss": 1.6036, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00012890959074058942, |
|
"loss": 1.552, |
|
"step": 3725 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00012870046823442396, |
|
"loss": 1.6111, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00012849120885820314, |
|
"loss": 1.5881, |
|
"step": 3735 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00012828181360986684, |
|
"loss": 1.5809, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00012807228348800292, |
|
"loss": 1.5749, |
|
"step": 3745 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00012786261949184244, |
|
"loss": 1.5972, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00012765282262125493, |
|
"loss": 1.6511, |
|
"step": 3755 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.0001274428938767435, |
|
"loss": 1.6381, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.0001272328342594403, |
|
"loss": 1.6434, |
|
"step": 3765 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00012702264477110144, |
|
"loss": 1.6911, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00012681232641410244, |
|
"loss": 1.5946, |
|
"step": 3775 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.0001266018801914334, |
|
"loss": 1.5972, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00012639130710669422, |
|
"loss": 1.6137, |
|
"step": 3785 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.0001261806081640897, |
|
"loss": 1.6259, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00012596978436842497, |
|
"loss": 1.5747, |
|
"step": 3795 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.0001257588367251005, |
|
"loss": 1.6364, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00012554776624010736, |
|
"loss": 1.629, |
|
"step": 3805 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00012533657392002254, |
|
"loss": 1.7398, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.0001251252607720039, |
|
"loss": 1.6403, |
|
"step": 3815 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00012491382780378563, |
|
"loss": 1.5895, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.0001247022760236733, |
|
"loss": 1.5988, |
|
"step": 3825 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00012449060644053906, |
|
"loss": 1.651, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00012427882006381697, |
|
"loss": 1.6151, |
|
"step": 3835 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.0001240669179034979, |
|
"loss": 1.6413, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.000123854900970125, |
|
"loss": 1.5878, |
|
"step": 3845 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.0001236427702747887, |
|
"loss": 1.6195, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.000123430526829122, |
|
"loss": 1.5949, |
|
"step": 3855 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00012321817164529547, |
|
"loss": 1.6172, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.0001230057057360128, |
|
"loss": 1.6163, |
|
"step": 3865 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00012279313011450546, |
|
"loss": 1.6464, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00012258044579452828, |
|
"loss": 1.6302, |
|
"step": 3875 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.0001223676537903544, |
|
"loss": 1.5792, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00012215475511677054, |
|
"loss": 1.6559, |
|
"step": 3885 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00012194175078907211, |
|
"loss": 1.6614, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00012172864182305832, |
|
"loss": 1.6347, |
|
"step": 3895 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00012151542923502745, |
|
"loss": 1.6379, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00012130211404177197, |
|
"loss": 1.6567, |
|
"step": 3905 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00012108869726057354, |
|
"loss": 1.6928, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00012087517990919839, |
|
"loss": 1.7283, |
|
"step": 3915 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00012066156300589237, |
|
"loss": 1.8493, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00012044784756937603, |
|
"loss": 1.8739, |
|
"step": 3925 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00012023403461883985, |
|
"loss": 1.7248, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00012002012517393931, |
|
"loss": 1.784, |
|
"step": 3935 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00011980612025479012, |
|
"loss": 1.7133, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.0001195920208819632, |
|
"loss": 1.7509, |
|
"step": 3945 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00011937782807648009, |
|
"loss": 1.723, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00011916354285980767, |
|
"loss": 1.7295, |
|
"step": 3955 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00011894916625385368, |
|
"loss": 1.6931, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00011873469928096169, |
|
"loss": 1.7882, |
|
"step": 3965 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00011852014296390612, |
|
"loss": 1.8715, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00011830549832588755, |
|
"loss": 1.7466, |
|
"step": 3975 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00011809076639052774, |
|
"loss": 1.7847, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00011787594818186474, |
|
"loss": 1.7821, |
|
"step": 3985 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00011766104472434803, |
|
"loss": 1.8205, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00011744605704283369, |
|
"loss": 1.7573, |
|
"step": 3995 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00011723098616257935, |
|
"loss": 1.8366, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00011701583310923953, |
|
"loss": 1.9667, |
|
"step": 4005 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00011680059890886055, |
|
"loss": 1.9864, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00011658528458787568, |
|
"loss": 2.1068, |
|
"step": 4015 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00011636989117310043, |
|
"loss": 2.1032, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.0001161544196917273, |
|
"loss": 1.9108, |
|
"step": 4025 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00011593887117132123, |
|
"loss": 2.0102, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00011572324663981453, |
|
"loss": 2.1442, |
|
"step": 4035 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00011550754712550196, |
|
"loss": 2.2282, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00011529177365703583, |
|
"loss": 2.1359, |
|
"step": 4045 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00011507592726342127, |
|
"loss": 2.0636, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00011486000897401104, |
|
"loss": 2.0083, |
|
"step": 4055 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00011464401981850081, |
|
"loss": 2.219, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00011442796082692426, |
|
"loss": 2.469, |
|
"step": 4065 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00011421183302964806, |
|
"loss": 2.5817, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00011399563745736701, |
|
"loss": 2.3501, |
|
"step": 4075 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00011377937514109913, |
|
"loss": 4.7679, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00011356304711218076, |
|
"loss": 4.1245, |
|
"step": 4085 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00011334665440226156, |
|
"loss": 3.8258, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00011313019804329969, |
|
"loss": 3.1891, |
|
"step": 4095 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00011291367906755683, |
|
"loss": 2.599, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00011269709850759334, |
|
"loss": 2.8603, |
|
"step": 4105 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00011248045739626314, |
|
"loss": 3.3006, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00011226375676670901, |
|
"loss": 4.4655, |
|
"step": 4115 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.0001120469976523576, |
|
"loss": 4.6281, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00011183018108691435, |
|
"loss": 5.3862, |
|
"step": 4125 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00011161330810435877, |
|
"loss": 6.0199, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00011139637973893939, |
|
"loss": 6.226, |
|
"step": 4135 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00011117939702516885, |
|
"loss": 6.3913, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00011096236099781896, |
|
"loss": 6.3674, |
|
"step": 4145 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00011074527269191585, |
|
"loss": 6.3291, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.0001105281331427349, |
|
"loss": 6.144, |
|
"step": 4155 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.0001103109433857958, |
|
"loss": 5.9615, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00011009370445685786, |
|
"loss": 6.0412, |
|
"step": 4165 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.0001098764173919147, |
|
"loss": 5.907, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00010965908322718961, |
|
"loss": 5.6353, |
|
"step": 4175 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00010944170299913048, |
|
"loss": 5.3082, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00010922427774440478, |
|
"loss": 5.1845, |
|
"step": 4185 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00010900680849989485, |
|
"loss": 4.9359, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.0001087892963026927, |
|
"loss": 4.7633, |
|
"step": 4195 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00010857174219009524, |
|
"loss": 4.7494, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00010835414719959927, |
|
"loss": 4.4632, |
|
"step": 4205 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00010813651236889648, |
|
"loss": 4.3406, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00010791883873586863, |
|
"loss": 4.021, |
|
"step": 4215 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00010770112733858242, |
|
"loss": 3.8432, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.0001074833792152848, |
|
"loss": 3.4208, |
|
"step": 4225 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00010726559540439768, |
|
"loss": 3.5933, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00010704777694451332, |
|
"loss": 3.3469, |
|
"step": 4235 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00010682992487438911, |
|
"loss": 4.1818, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.0001066120402329428, |
|
"loss": 6.0256, |
|
"step": 4245 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.0001063941240592474, |
|
"loss": 6.0616, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00010617617739252636, |
|
"loss": 6.3475, |
|
"step": 4255 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.0001059582012721485, |
|
"loss": 6.9718, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00010574019673762314, |
|
"loss": 7.2162, |
|
"step": 4265 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00010552216482859507, |
|
"loss": 7.1021, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00010530410658483966, |
|
"loss": 7.0514, |
|
"step": 4275 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00010508602304625783, |
|
"loss": 6.9977, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00010486791525287114, |
|
"loss": 6.9515, |
|
"step": 4285 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00010464978424481687, |
|
"loss": 6.7638, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.0001044316310623429, |
|
"loss": 6.7592, |
|
"step": 4295 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.000104213456745803, |
|
"loss": 6.7338, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00010399526233565158, |
|
"loss": 6.6529, |
|
"step": 4305 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00010377704887243902, |
|
"loss": 6.6232, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00010355881739680642, |
|
"loss": 6.5477, |
|
"step": 4315 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.0001033405689494809, |
|
"loss": 6.5135, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00010312230457127043, |
|
"loss": 6.2877, |
|
"step": 4325 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00010290402530305901, |
|
"loss": 6.1819, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.0001026857321858016, |
|
"loss": 6.4073, |
|
"step": 4335 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00010246742626051927, |
|
"loss": 6.503, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00010224910856829409, |
|
"loss": 6.3467, |
|
"step": 4345 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.0001020307801502643, |
|
"loss": 5.9681, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.0001018124420476193, |
|
"loss": 5.932, |
|
"step": 4355 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00010159409530159465, |
|
"loss": 5.8538, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00010137574095346712, |
|
"loss": 5.8141, |
|
"step": 4365 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00010115738004454971, |
|
"loss": 5.8956, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00010093901361618681, |
|
"loss": 5.7102, |
|
"step": 4375 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00010072064270974904, |
|
"loss": 5.3527, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00010050226836662839, |
|
"loss": 4.966, |
|
"step": 4385 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00010028389162823326, |
|
"loss": 5.239, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00010006551353598346, |
|
"loss": 5.1448, |
|
"step": 4395 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.984713513130528e-05, |
|
"loss": 4.7699, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.962875745562645e-05, |
|
"loss": 4.1809, |
|
"step": 4405 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.941038155037124e-05, |
|
"loss": 4.1431, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.919200845695555e-05, |
|
"loss": 3.8793, |
|
"step": 4415 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.897363921678175e-05, |
|
"loss": 3.718, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.875527487123397e-05, |
|
"loss": 4.2756, |
|
"step": 4425 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.853691646167285e-05, |
|
"loss": 4.8083, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.83185650294308e-05, |
|
"loss": 4.7415, |
|
"step": 4435 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.810022161580699e-05, |
|
"loss": 4.7713, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.788188726206226e-05, |
|
"loss": 6.1638, |
|
"step": 4445 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.766356300941439e-05, |
|
"loss": 6.6575, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.744524989903278e-05, |
|
"loss": 6.4801, |
|
"step": 4455 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.722694897203387e-05, |
|
"loss": 5.9204, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.70086612694759e-05, |
|
"loss": 5.3692, |
|
"step": 4465 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.679038783235416e-05, |
|
"loss": 4.6888, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.657212970159573e-05, |
|
"loss": 5.3672, |
|
"step": 4475 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.635388791805481e-05, |
|
"loss": 5.7097, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.61356635225077e-05, |
|
"loss": 5.9777, |
|
"step": 4485 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.591745755564761e-05, |
|
"loss": 6.3493, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.569927105808005e-05, |
|
"loss": 6.4762, |
|
"step": 4495 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.54811050703175e-05, |
|
"loss": 6.5824, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.52629606327748e-05, |
|
"loss": 6.3451, |
|
"step": 4505 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.504483878576386e-05, |
|
"loss": 5.9661, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.482674056948902e-05, |
|
"loss": 5.9223, |
|
"step": 4515 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.460866702404178e-05, |
|
"loss": 6.3891, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.439061918939615e-05, |
|
"loss": 6.7532, |
|
"step": 4525 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.417259810540332e-05, |
|
"loss": 6.7318, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.395460481178705e-05, |
|
"loss": 6.6247, |
|
"step": 4535 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.373664034813858e-05, |
|
"loss": 6.534, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.351870575391154e-05, |
|
"loss": 6.4605, |
|
"step": 4545 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.33008020684173e-05, |
|
"loss": 6.3064, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.308293033081964e-05, |
|
"loss": 6.3355, |
|
"step": 4555 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.286509158013012e-05, |
|
"loss": 6.7673, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.264728685520289e-05, |
|
"loss": 6.9996, |
|
"step": 4565 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.242951719472995e-05, |
|
"loss": 7.1712, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.221178363723595e-05, |
|
"loss": 7.1985, |
|
"step": 4575 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.199408722107342e-05, |
|
"loss": 7.1542, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.177642898441784e-05, |
|
"loss": 7.1913, |
|
"step": 4585 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.155880996526248e-05, |
|
"loss": 7.1135, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.134123120141373e-05, |
|
"loss": 7.0798, |
|
"step": 4595 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.112369373048587e-05, |
|
"loss": 7.1231, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.090619858989636e-05, |
|
"loss": 7.0743, |
|
"step": 4605 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.068874681686068e-05, |
|
"loss": 7.0606, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.047133944838766e-05, |
|
"loss": 7.0732, |
|
"step": 4615 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.025397752127424e-05, |
|
"loss": 7.006, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.003666207210064e-05, |
|
"loss": 7.0265, |
|
"step": 4625 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.981939413722553e-05, |
|
"loss": 7.1707, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.960217475278089e-05, |
|
"loss": 7.2962, |
|
"step": 4635 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.938500495466729e-05, |
|
"loss": 7.3594, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.916788577854861e-05, |
|
"loss": 7.3124, |
|
"step": 4645 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.895081825984762e-05, |
|
"loss": 7.2198, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.873380343374047e-05, |
|
"loss": 7.17, |
|
"step": 4655 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.851684233515217e-05, |
|
"loss": 7.1085, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.829993599875149e-05, |
|
"loss": 7.1046, |
|
"step": 4665 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.808308545894595e-05, |
|
"loss": 7.2317, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.786629174987713e-05, |
|
"loss": 7.1664, |
|
"step": 4675 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.764955590541546e-05, |
|
"loss": 7.1815, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.743287895915553e-05, |
|
"loss": 7.1897, |
|
"step": 4685 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.721626194441093e-05, |
|
"loss": 7.2131, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.699970589420957e-05, |
|
"loss": 7.2601, |
|
"step": 4695 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.678321184128847e-05, |
|
"loss": 7.1742, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.656678081808915e-05, |
|
"loss": 7.0768, |
|
"step": 4705 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.635041385675248e-05, |
|
"loss": 7.1501, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.613411198911374e-05, |
|
"loss": 7.0441, |
|
"step": 4715 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.591787624669799e-05, |
|
"loss": 7.0437, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.570170766071468e-05, |
|
"loss": 7.0524, |
|
"step": 4725 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.548560726205325e-05, |
|
"loss": 7.0686, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.526957608127776e-05, |
|
"loss": 7.0504, |
|
"step": 4735 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.505361514862235e-05, |
|
"loss": 6.9478, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.483772549398595e-05, |
|
"loss": 7.0398, |
|
"step": 4745 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.46219081469278e-05, |
|
"loss": 6.955, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.440616413666215e-05, |
|
"loss": 6.9533, |
|
"step": 4755 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.419049449205351e-05, |
|
"loss": 6.9178, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.397490024161191e-05, |
|
"loss": 6.924, |
|
"step": 4765 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.37593824134876e-05, |
|
"loss": 6.9213, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.354394203546658e-05, |
|
"loss": 6.9451, |
|
"step": 4775 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.332858013496536e-05, |
|
"loss": 6.9265, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.31132977390263e-05, |
|
"loss": 6.9345, |
|
"step": 4785 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.289809587431252e-05, |
|
"loss": 6.9697, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.26829755671032e-05, |
|
"loss": 6.9198, |
|
"step": 4795 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.246793784328843e-05, |
|
"loss": 6.9485, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.225298372836458e-05, |
|
"loss": 6.9376, |
|
"step": 4805 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.203811424742934e-05, |
|
"loss": 6.8936, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.182333042517657e-05, |
|
"loss": 6.8373, |
|
"step": 4815 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.16086332858919e-05, |
|
"loss": 6.9144, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.139402385344739e-05, |
|
"loss": 6.8827, |
|
"step": 4825 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.117950315129694e-05, |
|
"loss": 6.9062, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.096507220247122e-05, |
|
"loss": 6.9497, |
|
"step": 4835 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 8.075073202957297e-05, |
|
"loss": 6.8904, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 8.053648365477192e-05, |
|
"loss": 6.9602, |
|
"step": 4845 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 8.032232809980006e-05, |
|
"loss": 6.9755, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 8.01082663859468e-05, |
|
"loss": 7.065, |
|
"step": 4855 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.989429953405391e-05, |
|
"loss": 7.1277, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.968042856451093e-05, |
|
"loss": 7.0898, |
|
"step": 4865 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.946665449724992e-05, |
|
"loss": 7.1136, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.925297835174107e-05, |
|
"loss": 7.0805, |
|
"step": 4875 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.90394011469874e-05, |
|
"loss": 7.2114, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.882592390152022e-05, |
|
"loss": 7.1579, |
|
"step": 4885 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.861254763339405e-05, |
|
"loss": 7.2129, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.839927336018187e-05, |
|
"loss": 7.1796, |
|
"step": 4895 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.818610209897031e-05, |
|
"loss": 7.1316, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.79730348663547e-05, |
|
"loss": 7.1201, |
|
"step": 4905 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.776007267843432e-05, |
|
"loss": 7.0741, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.754721655080735e-05, |
|
"loss": 7.0825, |
|
"step": 4915 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.733446749856643e-05, |
|
"loss": 7.1446, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.712182653629332e-05, |
|
"loss": 7.0858, |
|
"step": 4925 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.690929467805445e-05, |
|
"loss": 7.1693, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.669687293739594e-05, |
|
"loss": 7.1195, |
|
"step": 4935 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.648456232733865e-05, |
|
"loss": 7.0688, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.627236386037363e-05, |
|
"loss": 7.1938, |
|
"step": 4945 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.606027854845696e-05, |
|
"loss": 7.1105, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.584830740300528e-05, |
|
"loss": 7.1628, |
|
"step": 4955 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.563645143489063e-05, |
|
"loss": 7.114, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.542471165443589e-05, |
|
"loss": 7.0888, |
|
"step": 4965 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.52130890714097e-05, |
|
"loss": 7.2031, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.500158469502197e-05, |
|
"loss": 7.4569, |
|
"step": 4975 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.479019953391879e-05, |
|
"loss": 7.5814, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.457893459617769e-05, |
|
"loss": 7.5974, |
|
"step": 4985 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.436779088930302e-05, |
|
"loss": 7.5981, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.41567694202208e-05, |
|
"loss": 7.6413, |
|
"step": 4995 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.394587119527427e-05, |
|
"loss": 7.5388, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.37350972202188e-05, |
|
"loss": 7.4875, |
|
"step": 5005 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.352444850021735e-05, |
|
"loss": 7.4391, |
|
"step": 5010 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.331392603983538e-05, |
|
"loss": 7.4093, |
|
"step": 5015 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.310353084303643e-05, |
|
"loss": 7.4181, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.289326391317697e-05, |
|
"loss": 7.4298, |
|
"step": 5025 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.268312625300186e-05, |
|
"loss": 7.3512, |
|
"step": 5030 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.24731188646395e-05, |
|
"loss": 7.3786, |
|
"step": 5035 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.226324274959691e-05, |
|
"loss": 7.3966, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.20534989087552e-05, |
|
"loss": 7.3913, |
|
"step": 5045 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.184388834236464e-05, |
|
"loss": 7.3248, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.163441205003997e-05, |
|
"loss": 7.3845, |
|
"step": 5055 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.142507103075548e-05, |
|
"loss": 7.3996, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.121586628284048e-05, |
|
"loss": 7.3292, |
|
"step": 5065 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.100679880397429e-05, |
|
"loss": 7.3323, |
|
"step": 5070 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.07978695911817e-05, |
|
"loss": 7.3656, |
|
"step": 5075 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 7.058907964082812e-05, |
|
"loss": 7.3201, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 7.038042994861472e-05, |
|
"loss": 7.355, |
|
"step": 5085 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 7.017192150957397e-05, |
|
"loss": 7.3745, |
|
"step": 5090 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 6.996355531806452e-05, |
|
"loss": 7.3526, |
|
"step": 5095 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 6.97553323677668e-05, |
|
"loss": 7.3823, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 6.954725365167807e-05, |
|
"loss": 7.364, |
|
"step": 5105 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 6.93393201621078e-05, |
|
"loss": 7.3639, |
|
"step": 5110 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 6.913153289067278e-05, |
|
"loss": 7.3759, |
|
"step": 5115 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 6.892389282829265e-05, |
|
"loss": 7.3812, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 6.871640096518491e-05, |
|
"loss": 7.3695, |
|
"step": 5125 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 6.85090582908603e-05, |
|
"loss": 7.3707, |
|
"step": 5130 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 6.830186579411822e-05, |
|
"loss": 7.3587, |
|
"step": 5135 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 6.809482446304172e-05, |
|
"loss": 7.3043, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 6.788793528499307e-05, |
|
"loss": 7.3778, |
|
"step": 5145 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 6.76811992466089e-05, |
|
"loss": 7.337, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 6.747461733379552e-05, |
|
"loss": 7.3518, |
|
"step": 5155 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 6.726819053172423e-05, |
|
"loss": 7.4198, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 6.70619198248267e-05, |
|
"loss": 7.3173, |
|
"step": 5165 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 6.685580619679005e-05, |
|
"loss": 7.292, |
|
"step": 5170 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 6.664985063055238e-05, |
|
"loss": 7.3553, |
|
"step": 5175 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 6.644405410829806e-05, |
|
"loss": 7.3367, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 6.623841761145284e-05, |
|
"loss": 7.3457, |
|
"step": 5185 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 6.603294212067955e-05, |
|
"loss": 7.3096, |
|
"step": 5190 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 6.582762861587294e-05, |
|
"loss": 7.3612, |
|
"step": 5195 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 6.562247807615549e-05, |
|
"loss": 7.3135, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 6.541749147987234e-05, |
|
"loss": 7.3, |
|
"step": 5205 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 6.521266980458696e-05, |
|
"loss": 7.3306, |
|
"step": 5210 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 6.500801402707617e-05, |
|
"loss": 7.3709, |
|
"step": 5215 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 6.480352512332569e-05, |
|
"loss": 7.3217, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 6.459920406852552e-05, |
|
"loss": 7.334, |
|
"step": 5225 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 6.439505183706505e-05, |
|
"loss": 7.3354, |
|
"step": 5230 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 6.419106940252872e-05, |
|
"loss": 7.2666, |
|
"step": 5235 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 6.39872577376911e-05, |
|
"loss": 7.3139, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 6.378361781451245e-05, |
|
"loss": 7.3116, |
|
"step": 5245 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 6.358015060413398e-05, |
|
"loss": 7.279, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 6.337685707687327e-05, |
|
"loss": 7.2851, |
|
"step": 5255 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 6.31737382022196e-05, |
|
"loss": 7.3416, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 6.297079494882927e-05, |
|
"loss": 7.3058, |
|
"step": 5265 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 6.276802828452123e-05, |
|
"loss": 7.2946, |
|
"step": 5270 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 6.256543917627214e-05, |
|
"loss": 7.2861, |
|
"step": 5275 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 6.236302859021205e-05, |
|
"loss": 7.2608, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 6.216079749161946e-05, |
|
"loss": 7.2939, |
|
"step": 5285 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 6.195874684491709e-05, |
|
"loss": 7.3122, |
|
"step": 5290 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 6.175687761366702e-05, |
|
"loss": 7.2933, |
|
"step": 5295 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 6.155519076056619e-05, |
|
"loss": 7.2998, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 6.13536872474418e-05, |
|
"loss": 7.2813, |
|
"step": 5305 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 6.115236803524666e-05, |
|
"loss": 7.2687, |
|
"step": 5310 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 6.0951234084054765e-05, |
|
"loss": 7.2508, |
|
"step": 5315 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 6.075028635305653e-05, |
|
"loss": 7.2309, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 6.054952580055439e-05, |
|
"loss": 7.2651, |
|
"step": 5325 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 6.034895338395801e-05, |
|
"loss": 7.2534, |
|
"step": 5330 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 6.014857005978003e-05, |
|
"loss": 7.2933, |
|
"step": 5335 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.994837678363112e-05, |
|
"loss": 7.2932, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.97483745102158e-05, |
|
"loss": 7.256, |
|
"step": 5345 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.9548564193327636e-05, |
|
"loss": 7.3109, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.9348946785844727e-05, |
|
"loss": 7.2412, |
|
"step": 5355 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.914952323972536e-05, |
|
"loss": 7.2895, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.8950294506003104e-05, |
|
"loss": 7.2577, |
|
"step": 5365 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.8751261534782655e-05, |
|
"loss": 7.2391, |
|
"step": 5370 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.8552425275235035e-05, |
|
"loss": 7.3199, |
|
"step": 5375 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.835378667559323e-05, |
|
"loss": 7.3275, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.8155346683147584e-05, |
|
"loss": 7.3315, |
|
"step": 5385 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.7957106244241263e-05, |
|
"loss": 7.3368, |
|
"step": 5390 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.775906630426581e-05, |
|
"loss": 7.298, |
|
"step": 5395 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 5.75612278076566e-05, |
|
"loss": 7.2765, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 5.7363591697888365e-05, |
|
"loss": 7.2852, |
|
"step": 5405 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 5.716615891747057e-05, |
|
"loss": 7.2814, |
|
"step": 5410 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 5.696893040794327e-05, |
|
"loss": 7.2282, |
|
"step": 5415 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 5.6771907109872015e-05, |
|
"loss": 7.2585, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 5.657508996284402e-05, |
|
"loss": 7.3253, |
|
"step": 5425 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 5.637847990546323e-05, |
|
"loss": 7.2465, |
|
"step": 5430 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 5.618207787534606e-05, |
|
"loss": 7.2002, |
|
"step": 5435 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 5.59858848091168e-05, |
|
"loss": 7.2705, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 5.5789901642403274e-05, |
|
"loss": 7.2102, |
|
"step": 5445 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 5.5594129309832275e-05, |
|
"loss": 7.2682, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 5.539856874502512e-05, |
|
"loss": 7.2422, |
|
"step": 5455 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 5.520322088059334e-05, |
|
"loss": 7.3147, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 5.500808664813399e-05, |
|
"loss": 7.2532, |
|
"step": 5465 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 5.481316697822538e-05, |
|
"loss": 7.2124, |
|
"step": 5470 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 5.461846280042261e-05, |
|
"loss": 7.2201, |
|
"step": 5475 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 5.442397504325307e-05, |
|
"loss": 7.2199, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 5.422970463421204e-05, |
|
"loss": 7.2293, |
|
"step": 5485 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 5.403565249975848e-05, |
|
"loss": 7.219, |
|
"step": 5490 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 5.3841819565310116e-05, |
|
"loss": 7.227, |
|
"step": 5495 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 5.3648206755239494e-05, |
|
"loss": 7.2637, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 5.345481499286948e-05, |
|
"loss": 7.172, |
|
"step": 5505 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 5.326164520046864e-05, |
|
"loss": 7.2098, |
|
"step": 5510 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 5.306869829924705e-05, |
|
"loss": 7.231, |
|
"step": 5515 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 5.28759752093518e-05, |
|
"loss": 7.187, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 5.26834768498627e-05, |
|
"loss": 7.1514, |
|
"step": 5525 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 5.2491204138787745e-05, |
|
"loss": 7.2044, |
|
"step": 5530 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 5.2299157993058986e-05, |
|
"loss": 7.2039, |
|
"step": 5535 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 5.210733932852792e-05, |
|
"loss": 7.2409, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 5.191574905996106e-05, |
|
"loss": 7.1852, |
|
"step": 5545 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 5.172438810103599e-05, |
|
"loss": 7.1771, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 5.153325736433657e-05, |
|
"loss": 7.1748, |
|
"step": 5555 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 5.134235776134878e-05, |
|
"loss": 7.2238, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 5.115169020245629e-05, |
|
"loss": 7.1801, |
|
"step": 5565 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 5.0961255596936366e-05, |
|
"loss": 7.1874, |
|
"step": 5570 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 5.0771054852955036e-05, |
|
"loss": 7.1418, |
|
"step": 5575 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 5.058108887756335e-05, |
|
"loss": 7.1988, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 5.039135857669265e-05, |
|
"loss": 7.176, |
|
"step": 5585 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 5.020186485515024e-05, |
|
"loss": 7.2226, |
|
"step": 5590 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 5.001260861661542e-05, |
|
"loss": 7.1885, |
|
"step": 5595 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.982359076363483e-05, |
|
"loss": 7.181, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.963481219761831e-05, |
|
"loss": 7.1147, |
|
"step": 5605 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.9446273818834476e-05, |
|
"loss": 7.1665, |
|
"step": 5610 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.9257976526406714e-05, |
|
"loss": 7.2306, |
|
"step": 5615 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.9069921218308404e-05, |
|
"loss": 7.1844, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.888210879135917e-05, |
|
"loss": 7.2101, |
|
"step": 5625 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.869454014122025e-05, |
|
"loss": 7.1925, |
|
"step": 5630 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.850721616239034e-05, |
|
"loss": 7.1845, |
|
"step": 5635 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.83201377482013e-05, |
|
"loss": 7.2063, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.8133305790813964e-05, |
|
"loss": 7.1474, |
|
"step": 5645 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.7946721181213806e-05, |
|
"loss": 7.2337, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.776038480920667e-05, |
|
"loss": 7.1697, |
|
"step": 5655 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.757429756341475e-05, |
|
"loss": 7.1528, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.73884603312719e-05, |
|
"loss": 7.1903, |
|
"step": 5665 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.720287399901996e-05, |
|
"loss": 7.1956, |
|
"step": 5670 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.701753945170407e-05, |
|
"loss": 7.1651, |
|
"step": 5675 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.683245757316872e-05, |
|
"loss": 7.1655, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.664762924605339e-05, |
|
"loss": 7.1478, |
|
"step": 5685 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.646305535178841e-05, |
|
"loss": 7.1689, |
|
"step": 5690 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.627873677059076e-05, |
|
"loss": 7.14, |
|
"step": 5695 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.6094674381459804e-05, |
|
"loss": 7.1661, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.591086906217326e-05, |
|
"loss": 7.1725, |
|
"step": 5705 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.5727321689282774e-05, |
|
"loss": 7.164, |
|
"step": 5710 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.5544033138109955e-05, |
|
"loss": 7.2252, |
|
"step": 5715 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 4.536100428274205e-05, |
|
"loss": 7.1618, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 4.517823599602788e-05, |
|
"loss": 7.1502, |
|
"step": 5725 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 4.499572914957364e-05, |
|
"loss": 7.2056, |
|
"step": 5730 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 4.481348461373868e-05, |
|
"loss": 7.1786, |
|
"step": 5735 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 4.463150325763156e-05, |
|
"loss": 7.172, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 4.4449785949105515e-05, |
|
"loss": 7.1585, |
|
"step": 5745 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 4.426833355475482e-05, |
|
"loss": 7.0824, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 4.408714693991023e-05, |
|
"loss": 7.1172, |
|
"step": 5755 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 4.390622696863509e-05, |
|
"loss": 7.1968, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 4.37255745037211e-05, |
|
"loss": 7.1607, |
|
"step": 5765 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 4.354519040668428e-05, |
|
"loss": 7.1378, |
|
"step": 5770 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 4.336507553776082e-05, |
|
"loss": 7.1512, |
|
"step": 5775 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 4.318523075590295e-05, |
|
"loss": 7.1638, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 4.3005656918775005e-05, |
|
"loss": 7.1501, |
|
"step": 5785 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 4.2826354882748974e-05, |
|
"loss": 7.1776, |
|
"step": 5790 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 4.264732550290089e-05, |
|
"loss": 7.1425, |
|
"step": 5795 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 4.2468569633006394e-05, |
|
"loss": 7.183, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 4.229008812553678e-05, |
|
"loss": 7.2219, |
|
"step": 5805 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 4.211188183165494e-05, |
|
"loss": 7.1401, |
|
"step": 5810 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 4.193395160121133e-05, |
|
"loss": 7.1379, |
|
"step": 5815 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 4.175629828273985e-05, |
|
"loss": 7.1859, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 4.15789227234538e-05, |
|
"loss": 7.1229, |
|
"step": 5825 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 4.1401825769242006e-05, |
|
"loss": 7.1255, |
|
"step": 5830 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 4.122500826466443e-05, |
|
"loss": 7.1883, |
|
"step": 5835 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 4.10484710529486e-05, |
|
"loss": 7.1987, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 4.087221497598518e-05, |
|
"loss": 7.1391, |
|
"step": 5845 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 4.06962408743242e-05, |
|
"loss": 7.1025, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 4.052054958717095e-05, |
|
"loss": 7.1166, |
|
"step": 5855 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 4.0345141952382026e-05, |
|
"loss": 7.1652, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 4.017001880646129e-05, |
|
"loss": 7.1493, |
|
"step": 5865 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.999518098455586e-05, |
|
"loss": 7.1326, |
|
"step": 5870 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.98206293204523e-05, |
|
"loss": 7.1932, |
|
"step": 5875 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.964636464657238e-05, |
|
"loss": 7.1522, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.947238779396929e-05, |
|
"loss": 7.1297, |
|
"step": 5885 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.929869959232363e-05, |
|
"loss": 7.2406, |
|
"step": 5890 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.912530086993944e-05, |
|
"loss": 7.1551, |
|
"step": 5895 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.895219245374024e-05, |
|
"loss": 7.1483, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.877937516926511e-05, |
|
"loss": 7.1624, |
|
"step": 5905 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.860684984066485e-05, |
|
"loss": 7.1055, |
|
"step": 5910 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.843461729069773e-05, |
|
"loss": 7.1082, |
|
"step": 5915 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.8262678340726e-05, |
|
"loss": 7.128, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.8091033810711655e-05, |
|
"loss": 7.137, |
|
"step": 5925 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.791968451921263e-05, |
|
"loss": 7.1643, |
|
"step": 5930 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.774863128337887e-05, |
|
"loss": 7.111, |
|
"step": 5935 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.7577874918948586e-05, |
|
"loss": 7.1551, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.740741624024405e-05, |
|
"loss": 7.1342, |
|
"step": 5945 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.723725606016799e-05, |
|
"loss": 7.161, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.706739519019969e-05, |
|
"loss": 7.1229, |
|
"step": 5955 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.689783444039088e-05, |
|
"loss": 7.127, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.67285746193622e-05, |
|
"loss": 7.109, |
|
"step": 5965 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.655961653429911e-05, |
|
"loss": 7.0979, |
|
"step": 5970 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.639096099094813e-05, |
|
"loss": 7.1028, |
|
"step": 5975 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.6222608793612954e-05, |
|
"loss": 7.1263, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.605456074515078e-05, |
|
"loss": 7.1616, |
|
"step": 5985 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.5886817646968105e-05, |
|
"loss": 7.1177, |
|
"step": 5990 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.571938029901739e-05, |
|
"loss": 7.0856, |
|
"step": 5995 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.555224949979288e-05, |
|
"loss": 7.0899, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.538542604632684e-05, |
|
"loss": 7.1465, |
|
"step": 6005 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.521891073418602e-05, |
|
"loss": 7.121, |
|
"step": 6010 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.505270435746752e-05, |
|
"loss": 7.1511, |
|
"step": 6015 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.488680770879523e-05, |
|
"loss": 7.1513, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.4721221579315885e-05, |
|
"loss": 7.1204, |
|
"step": 6025 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.4555946758695554e-05, |
|
"loss": 7.0737, |
|
"step": 6030 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.439098403511546e-05, |
|
"loss": 7.0769, |
|
"step": 6035 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.422633419526868e-05, |
|
"loss": 7.1364, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.4061998024356054e-05, |
|
"loss": 7.1108, |
|
"step": 6045 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.3897976306082614e-05, |
|
"loss": 7.0665, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.3734269822653755e-05, |
|
"loss": 7.0927, |
|
"step": 6055 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.3570879354771554e-05, |
|
"loss": 7.1011, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.340780568163108e-05, |
|
"loss": 7.0809, |
|
"step": 6065 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.3245049580916546e-05, |
|
"loss": 7.1185, |
|
"step": 6070 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.3082611828797816e-05, |
|
"loss": 7.1558, |
|
"step": 6075 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.292049319992647e-05, |
|
"loss": 7.1326, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.275869446743227e-05, |
|
"loss": 7.1077, |
|
"step": 6085 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.259721640291938e-05, |
|
"loss": 7.1089, |
|
"step": 6090 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.243605977646278e-05, |
|
"loss": 7.0998, |
|
"step": 6095 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.227522535660449e-05, |
|
"loss": 7.15, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.211471391034992e-05, |
|
"loss": 7.1657, |
|
"step": 6105 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.195452620316445e-05, |
|
"loss": 7.1311, |
|
"step": 6110 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.179466299896926e-05, |
|
"loss": 7.106, |
|
"step": 6115 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.163512506013827e-05, |
|
"loss": 7.1005, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.1475913147494115e-05, |
|
"loss": 7.1243, |
|
"step": 6125 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.131702802030466e-05, |
|
"loss": 7.1204, |
|
"step": 6130 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.1158470436279364e-05, |
|
"loss": 7.1261, |
|
"step": 6135 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.1000241151565643e-05, |
|
"loss": 7.1328, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.0842340920745316e-05, |
|
"loss": 7.1468, |
|
"step": 6145 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.068477049683091e-05, |
|
"loss": 7.0808, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.052753063126227e-05, |
|
"loss": 7.0821, |
|
"step": 6155 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.03706220739026e-05, |
|
"loss": 7.1532, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.021404557303539e-05, |
|
"loss": 7.1371, |
|
"step": 6165 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.0057801875360404e-05, |
|
"loss": 7.1025, |
|
"step": 6170 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.990189172599035e-05, |
|
"loss": 7.1113, |
|
"step": 6175 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.9746315868447273e-05, |
|
"loss": 7.1125, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.9591075044659e-05, |
|
"loss": 7.1206, |
|
"step": 6185 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.9436169994955643e-05, |
|
"loss": 7.15, |
|
"step": 6190 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.9281601458065945e-05, |
|
"loss": 7.1574, |
|
"step": 6195 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.912737017111403e-05, |
|
"loss": 7.0977, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.8973476869615466e-05, |
|
"loss": 7.1085, |
|
"step": 6205 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.8819922287474222e-05, |
|
"loss": 7.0789, |
|
"step": 6210 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.866670715697879e-05, |
|
"loss": 7.1432, |
|
"step": 6215 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.851383220879892e-05, |
|
"loss": 7.1403, |
|
"step": 6220 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.836129817198203e-05, |
|
"loss": 7.0981, |
|
"step": 6225 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.8209105773949783e-05, |
|
"loss": 7.0925, |
|
"step": 6230 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.8057255740494593e-05, |
|
"loss": 7.091, |
|
"step": 6235 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.790574879577612e-05, |
|
"loss": 7.1134, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.7754585662317968e-05, |
|
"loss": 7.1216, |
|
"step": 6245 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.7603767061004026e-05, |
|
"loss": 7.1437, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.7453293711075212e-05, |
|
"loss": 7.0705, |
|
"step": 6255 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.7303166330125918e-05, |
|
"loss": 7.1788, |
|
"step": 6260 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.715338563410068e-05, |
|
"loss": 7.143, |
|
"step": 6265 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.7003952337290694e-05, |
|
"loss": 7.1132, |
|
"step": 6270 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.685486715233043e-05, |
|
"loss": 7.0928, |
|
"step": 6275 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.670613079019436e-05, |
|
"loss": 7.1189, |
|
"step": 6280 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.6557743960193214e-05, |
|
"loss": 7.0892, |
|
"step": 6285 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.6409707369971094e-05, |
|
"loss": 7.1348, |
|
"step": 6290 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.6262021725501674e-05, |
|
"loss": 7.1308, |
|
"step": 6295 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.611468773108503e-05, |
|
"loss": 7.0822, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.5967706089344258e-05, |
|
"loss": 7.1226, |
|
"step": 6305 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.582107750122219e-05, |
|
"loss": 7.097, |
|
"step": 6310 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.5674802665977814e-05, |
|
"loss": 7.1524, |
|
"step": 6315 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.5528882281183208e-05, |
|
"loss": 7.1398, |
|
"step": 6320 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.5383317042720166e-05, |
|
"loss": 7.1162, |
|
"step": 6325 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.5238107644776642e-05, |
|
"loss": 7.1149, |
|
"step": 6330 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.509325477984379e-05, |
|
"loss": 7.0907, |
|
"step": 6335 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.4948759138712384e-05, |
|
"loss": 7.1303, |
|
"step": 6340 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.4804621410469687e-05, |
|
"loss": 7.1396, |
|
"step": 6345 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.4660842282496033e-05, |
|
"loss": 7.1247, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.4517422440461758e-05, |
|
"loss": 7.0649, |
|
"step": 6355 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.4374362568323638e-05, |
|
"loss": 7.1297, |
|
"step": 6360 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.423166334832183e-05, |
|
"loss": 7.0697, |
|
"step": 6365 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.4089325460976707e-05, |
|
"loss": 7.1298, |
|
"step": 6370 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.394734958508523e-05, |
|
"loss": 7.1407, |
|
"step": 6375 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.380573639771819e-05, |
|
"loss": 7.1568, |
|
"step": 6380 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.3664486574216627e-05, |
|
"loss": 7.1684, |
|
"step": 6385 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.3523600788188748e-05, |
|
"loss": 7.062, |
|
"step": 6390 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.3383079711506682e-05, |
|
"loss": 7.1253, |
|
"step": 6395 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.324292401430339e-05, |
|
"loss": 7.0864, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.3103134364969213e-05, |
|
"loss": 7.1328, |
|
"step": 6405 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.2963711430148926e-05, |
|
"loss": 7.0595, |
|
"step": 6410 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.282465587473852e-05, |
|
"loss": 7.1297, |
|
"step": 6415 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.2685968361881903e-05, |
|
"loss": 7.1528, |
|
"step": 6420 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.2547649552967863e-05, |
|
"loss": 7.0892, |
|
"step": 6425 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.2409700107626842e-05, |
|
"loss": 7.1207, |
|
"step": 6430 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 2.227212068372785e-05, |
|
"loss": 7.0815, |
|
"step": 6435 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 2.2134911937375246e-05, |
|
"loss": 7.0852, |
|
"step": 6440 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 2.1998074522905765e-05, |
|
"loss": 7.126, |
|
"step": 6445 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 2.1861609092885205e-05, |
|
"loss": 7.1435, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 2.1725516298105352e-05, |
|
"loss": 7.1288, |
|
"step": 6455 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 2.1589796787581074e-05, |
|
"loss": 7.0758, |
|
"step": 6460 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 2.1454451208546978e-05, |
|
"loss": 7.1251, |
|
"step": 6465 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 2.1319480206454435e-05, |
|
"loss": 7.1051, |
|
"step": 6470 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 2.118488442496851e-05, |
|
"loss": 7.1061, |
|
"step": 6475 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 2.1050664505964868e-05, |
|
"loss": 7.137, |
|
"step": 6480 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 2.0916821089526683e-05, |
|
"loss": 7.104, |
|
"step": 6485 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 2.0783354813941714e-05, |
|
"loss": 7.1227, |
|
"step": 6490 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 2.0650266315699074e-05, |
|
"loss": 7.0654, |
|
"step": 6495 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 2.0517556229486322e-05, |
|
"loss": 7.1448, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 2.038522518818641e-05, |
|
"loss": 7.0804, |
|
"step": 6505 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 2.025327382287464e-05, |
|
"loss": 7.1061, |
|
"step": 6510 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 2.0121702762815697e-05, |
|
"loss": 7.1668, |
|
"step": 6515 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.9990512635460567e-05, |
|
"loss": 7.0888, |
|
"step": 6520 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.985970406644373e-05, |
|
"loss": 7.0287, |
|
"step": 6525 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.9729277679579837e-05, |
|
"loss": 7.1052, |
|
"step": 6530 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.9599234096861175e-05, |
|
"loss": 7.0935, |
|
"step": 6535 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.946957393845432e-05, |
|
"loss": 7.1134, |
|
"step": 6540 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.9340297822697372e-05, |
|
"loss": 7.0912, |
|
"step": 6545 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.9211406366096974e-05, |
|
"loss": 7.1094, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.908290018332536e-05, |
|
"loss": 7.1309, |
|
"step": 6555 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.89547798872174e-05, |
|
"loss": 7.1019, |
|
"step": 6560 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.8827046088767732e-05, |
|
"loss": 7.1125, |
|
"step": 6565 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.8699699397127857e-05, |
|
"loss": 7.185, |
|
"step": 6570 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.8572740419603085e-05, |
|
"loss": 7.1206, |
|
"step": 6575 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.8446169761649866e-05, |
|
"loss": 7.1546, |
|
"step": 6580 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.831998802687275e-05, |
|
"loss": 7.1544, |
|
"step": 6585 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.81941958170215e-05, |
|
"loss": 7.0619, |
|
"step": 6590 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.8068793731988332e-05, |
|
"loss": 7.092, |
|
"step": 6595 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.7943782369804952e-05, |
|
"loss": 7.1239, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.781916232663975e-05, |
|
"loss": 7.0881, |
|
"step": 6605 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.7694934196794944e-05, |
|
"loss": 7.1331, |
|
"step": 6610 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.7571098572703804e-05, |
|
"loss": 7.1056, |
|
"step": 6615 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.7447656044927697e-05, |
|
"loss": 7.1104, |
|
"step": 6620 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.7324607202153397e-05, |
|
"loss": 7.0552, |
|
"step": 6625 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.72019526311902e-05, |
|
"loss": 7.0812, |
|
"step": 6630 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.7079692916967172e-05, |
|
"loss": 7.0928, |
|
"step": 6635 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.695782864253034e-05, |
|
"loss": 7.1079, |
|
"step": 6640 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.6836360389039906e-05, |
|
"loss": 7.1214, |
|
"step": 6645 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.6715288735767475e-05, |
|
"loss": 7.0516, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.6594614260093288e-05, |
|
"loss": 7.1011, |
|
"step": 6655 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.647433753750357e-05, |
|
"loss": 7.0939, |
|
"step": 6660 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.6354459141587576e-05, |
|
"loss": 7.1032, |
|
"step": 6665 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.6234979644035053e-05, |
|
"loss": 7.1052, |
|
"step": 6670 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.6115899614633388e-05, |
|
"loss": 7.1116, |
|
"step": 6675 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.599721962126497e-05, |
|
"loss": 7.1135, |
|
"step": 6680 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.5878940229904448e-05, |
|
"loss": 7.1303, |
|
"step": 6685 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.5761062004616e-05, |
|
"loss": 7.1007, |
|
"step": 6690 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.5643585507550785e-05, |
|
"loss": 7.1339, |
|
"step": 6695 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.552651129894397e-05, |
|
"loss": 7.1579, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.5409839937112414e-05, |
|
"loss": 7.1171, |
|
"step": 6705 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.5293571978451772e-05, |
|
"loss": 7.1021, |
|
"step": 6710 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.5177707977433876e-05, |
|
"loss": 7.0989, |
|
"step": 6715 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.5062248486604125e-05, |
|
"loss": 7.1189, |
|
"step": 6720 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.494719405657895e-05, |
|
"loss": 7.1078, |
|
"step": 6725 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.4832545236042916e-05, |
|
"loss": 7.0767, |
|
"step": 6730 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.4718302571746345e-05, |
|
"loss": 7.0947, |
|
"step": 6735 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.4604466608502709e-05, |
|
"loss": 7.1365, |
|
"step": 6740 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.4491037889185822e-05, |
|
"loss": 7.1164, |
|
"step": 6745 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.4378016954727492e-05, |
|
"loss": 7.0977, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.4265404344114785e-05, |
|
"loss": 7.0916, |
|
"step": 6755 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.4153200594387527e-05, |
|
"loss": 7.1789, |
|
"step": 6760 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.4041406240635691e-05, |
|
"loss": 7.1367, |
|
"step": 6765 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.3930021815996963e-05, |
|
"loss": 7.073, |
|
"step": 6770 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.3819047851653988e-05, |
|
"loss": 7.1137, |
|
"step": 6775 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.3708484876832006e-05, |
|
"loss": 7.1103, |
|
"step": 6780 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.3598333418796349e-05, |
|
"loss": 7.1446, |
|
"step": 6785 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.3488594002849809e-05, |
|
"loss": 7.0796, |
|
"step": 6790 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.3379267152330165e-05, |
|
"loss": 7.1162, |
|
"step": 6795 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.3270353388607726e-05, |
|
"loss": 7.0939, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.3161853231082855e-05, |
|
"loss": 7.0685, |
|
"step": 6805 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.3053767197183387e-05, |
|
"loss": 7.07, |
|
"step": 6810 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.2946095802362369e-05, |
|
"loss": 7.0702, |
|
"step": 6815 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.2838839560095317e-05, |
|
"loss": 7.0784, |
|
"step": 6820 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.2731998981877991e-05, |
|
"loss": 7.1051, |
|
"step": 6825 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.2625574577223931e-05, |
|
"loss": 7.075, |
|
"step": 6830 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.2519566853661913e-05, |
|
"loss": 7.1234, |
|
"step": 6835 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.2413976316733611e-05, |
|
"loss": 7.109, |
|
"step": 6840 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.2308803469991192e-05, |
|
"loss": 7.1047, |
|
"step": 6845 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.2204048814994883e-05, |
|
"loss": 7.1089, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.2099712851310563e-05, |
|
"loss": 7.1018, |
|
"step": 6855 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.1995796076507482e-05, |
|
"loss": 7.0676, |
|
"step": 6860 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.1892298986155758e-05, |
|
"loss": 7.08, |
|
"step": 6865 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.1789222073824035e-05, |
|
"loss": 7.0252, |
|
"step": 6870 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.1686565831077256e-05, |
|
"loss": 7.0989, |
|
"step": 6875 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.158433074747417e-05, |
|
"loss": 7.1265, |
|
"step": 6880 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.1482517310565066e-05, |
|
"loss": 7.0987, |
|
"step": 6885 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.1381126005889408e-05, |
|
"loss": 7.1037, |
|
"step": 6890 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.1280157316973639e-05, |
|
"loss": 7.1101, |
|
"step": 6895 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.1179611725328632e-05, |
|
"loss": 7.1096, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.10794897104477e-05, |
|
"loss": 7.1102, |
|
"step": 6905 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.0979791749804102e-05, |
|
"loss": 7.1157, |
|
"step": 6910 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.088051831884872e-05, |
|
"loss": 7.0908, |
|
"step": 6915 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.0781669891008061e-05, |
|
"loss": 7.058, |
|
"step": 6920 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.06832469376817e-05, |
|
"loss": 7.0818, |
|
"step": 6925 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.0585249928240226e-05, |
|
"loss": 7.1233, |
|
"step": 6930 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.0487679330022893e-05, |
|
"loss": 7.0973, |
|
"step": 6935 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.0390535608335506e-05, |
|
"loss": 7.0348, |
|
"step": 6940 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.0293819226448031e-05, |
|
"loss": 7.1054, |
|
"step": 6945 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.0197530645592579e-05, |
|
"loss": 7.1117, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.0101670324961098e-05, |
|
"loss": 7.0877, |
|
"step": 6955 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.000623872170311e-05, |
|
"loss": 7.0845, |
|
"step": 6960 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.911236290923765e-06, |
|
"loss": 7.0471, |
|
"step": 6965 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.81666348568142e-06, |
|
"loss": 7.1047, |
|
"step": 6970 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.722520756985632e-06, |
|
"loss": 7.138, |
|
"step": 6975 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.628808553794922e-06, |
|
"loss": 7.1384, |
|
"step": 6980 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.535527323014748e-06, |
|
"loss": 7.0887, |
|
"step": 6985 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.442677509495179e-06, |
|
"loss": 7.0972, |
|
"step": 6990 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 9.350259556029017e-06, |
|
"loss": 7.1319, |
|
"step": 6995 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 9.25827390334949e-06, |
|
"loss": 7.0781, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 9.16672099012822e-06, |
|
"loss": 7.0824, |
|
"step": 7005 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 9.075601252973153e-06, |
|
"loss": 7.1133, |
|
"step": 7010 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 8.984915126426452e-06, |
|
"loss": 7.0799, |
|
"step": 7015 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 8.894663042962426e-06, |
|
"loss": 7.1128, |
|
"step": 7020 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 8.804845432985454e-06, |
|
"loss": 7.0679, |
|
"step": 7025 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 8.715462724827995e-06, |
|
"loss": 7.072, |
|
"step": 7030 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 8.626515344748454e-06, |
|
"loss": 7.1046, |
|
"step": 7035 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 8.5380037169292e-06, |
|
"loss": 7.0422, |
|
"step": 7040 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 8.44992826347456e-06, |
|
"loss": 7.1035, |
|
"step": 7045 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 8.362289404408751e-06, |
|
"loss": 7.1118, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 8.275087557673934e-06, |
|
"loss": 7.1315, |
|
"step": 7055 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 8.188323139128174e-06, |
|
"loss": 7.1291, |
|
"step": 7060 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 8.101996562543546e-06, |
|
"loss": 7.0256, |
|
"step": 7065 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 8.01610823960397e-06, |
|
"loss": 7.0446, |
|
"step": 7070 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 7.930658579903483e-06, |
|
"loss": 7.0618, |
|
"step": 7075 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 7.845647990944138e-06, |
|
"loss": 7.024, |
|
"step": 7080 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 7.761076878134076e-06, |
|
"loss": 7.1422, |
|
"step": 7085 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 7.676945644785638e-06, |
|
"loss": 7.0919, |
|
"step": 7090 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 7.593254692113449e-06, |
|
"loss": 7.0855, |
|
"step": 7095 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 7.510004419232386e-06, |
|
"loss": 7.0651, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 7.427195223155825e-06, |
|
"loss": 7.1134, |
|
"step": 7105 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 7.3448274987937046e-06, |
|
"loss": 7.1094, |
|
"step": 7110 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 7.262901638950548e-06, |
|
"loss": 7.1039, |
|
"step": 7115 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 7.181418034323739e-06, |
|
"loss": 7.0904, |
|
"step": 7120 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 7.100377073501552e-06, |
|
"loss": 7.0765, |
|
"step": 7125 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 7.019779142961325e-06, |
|
"loss": 7.0759, |
|
"step": 7130 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 6.939624627067609e-06, |
|
"loss": 7.0694, |
|
"step": 7135 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 6.859913908070426e-06, |
|
"loss": 7.0478, |
|
"step": 7140 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 6.780647366103276e-06, |
|
"loss": 7.0595, |
|
"step": 7145 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 6.701825379181437e-06, |
|
"loss": 7.122, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 6.62344832320021e-06, |
|
"loss": 7.0699, |
|
"step": 7155 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 6.545516571932964e-06, |
|
"loss": 7.0677, |
|
"step": 7160 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 6.4680304970295095e-06, |
|
"loss": 7.0726, |
|
"step": 7165 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 6.390990468014268e-06, |
|
"loss": 7.1164, |
|
"step": 7170 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 6.314396852284454e-06, |
|
"loss": 7.0973, |
|
"step": 7175 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 6.238250015108427e-06, |
|
"loss": 7.0989, |
|
"step": 7180 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 6.162550319623894e-06, |
|
"loss": 7.0984, |
|
"step": 7185 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 6.087298126836139e-06, |
|
"loss": 7.1048, |
|
"step": 7190 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 6.01249379561637e-06, |
|
"loss": 7.075, |
|
"step": 7195 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 5.938137682700018e-06, |
|
"loss": 7.0979, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 5.8642301426849455e-06, |
|
"loss": 7.0613, |
|
"step": 7205 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 5.790771528029837e-06, |
|
"loss": 7.0866, |
|
"step": 7210 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 5.7177621890524894e-06, |
|
"loss": 7.1399, |
|
"step": 7215 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 5.645202473928113e-06, |
|
"loss": 7.1295, |
|
"step": 7220 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 5.573092728687734e-06, |
|
"loss": 7.0951, |
|
"step": 7225 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 5.501433297216529e-06, |
|
"loss": 7.0653, |
|
"step": 7230 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 5.430224521252136e-06, |
|
"loss": 7.1199, |
|
"step": 7235 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 5.359466740383012e-06, |
|
"loss": 7.0713, |
|
"step": 7240 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 5.2891602920469575e-06, |
|
"loss": 7.079, |
|
"step": 7245 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 5.2193055115293375e-06, |
|
"loss": 7.0955, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 5.149902731961565e-06, |
|
"loss": 7.0858, |
|
"step": 7255 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 5.080952284319496e-06, |
|
"loss": 7.1116, |
|
"step": 7260 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 5.012454497421903e-06, |
|
"loss": 7.0555, |
|
"step": 7265 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.944409697928775e-06, |
|
"loss": 7.1331, |
|
"step": 7270 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.876818210339906e-06, |
|
"loss": 7.0853, |
|
"step": 7275 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.809680356993273e-06, |
|
"loss": 7.1169, |
|
"step": 7280 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.742996458063476e-06, |
|
"loss": 7.0531, |
|
"step": 7285 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.676766831560297e-06, |
|
"loss": 7.0727, |
|
"step": 7290 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.610991793327102e-06, |
|
"loss": 7.1305, |
|
"step": 7295 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.545671657039363e-06, |
|
"loss": 7.1026, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.480806734203169e-06, |
|
"loss": 7.1004, |
|
"step": 7305 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.416397334153766e-06, |
|
"loss": 7.1217, |
|
"step": 7310 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 4.35244376405397e-06, |
|
"loss": 7.1107, |
|
"step": 7315 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 4.288946328892885e-06, |
|
"loss": 7.1448, |
|
"step": 7320 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 4.225905331484292e-06, |
|
"loss": 7.1449, |
|
"step": 7325 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 4.163321072465221e-06, |
|
"loss": 7.1134, |
|
"step": 7330 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 4.101193850294649e-06, |
|
"loss": 7.0643, |
|
"step": 7335 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 4.039523961251912e-06, |
|
"loss": 7.0833, |
|
"step": 7340 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.978311699435399e-06, |
|
"loss": 7.0718, |
|
"step": 7345 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.917557356761081e-06, |
|
"loss": 7.1263, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.857261222961217e-06, |
|
"loss": 7.1026, |
|
"step": 7355 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.7974235855828156e-06, |
|
"loss": 7.1344, |
|
"step": 7360 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.7380447299864342e-06, |
|
"loss": 7.1107, |
|
"step": 7365 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.6791249393446916e-06, |
|
"loss": 7.0813, |
|
"step": 7370 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.6206644946409906e-06, |
|
"loss": 7.0649, |
|
"step": 7375 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.562663674668154e-06, |
|
"loss": 7.1383, |
|
"step": 7380 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.5051227560270683e-06, |
|
"loss": 7.0596, |
|
"step": 7385 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.4480420131254186e-06, |
|
"loss": 7.1122, |
|
"step": 7390 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 3.391421718176324e-06, |
|
"loss": 7.0657, |
|
"step": 7395 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 3.335262141197104e-06, |
|
"loss": 7.141, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 3.279563550007936e-06, |
|
"loss": 7.1023, |
|
"step": 7405 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 3.224326210230566e-06, |
|
"loss": 7.1073, |
|
"step": 7410 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 3.1695503852871234e-06, |
|
"loss": 7.0651, |
|
"step": 7415 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 3.1152363363987856e-06, |
|
"loss": 7.0475, |
|
"step": 7420 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 3.0613843225845486e-06, |
|
"loss": 7.1515, |
|
"step": 7425 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 3.0079946006600247e-06, |
|
"loss": 7.0907, |
|
"step": 7430 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.955067425236202e-06, |
|
"loss": 7.0909, |
|
"step": 7435 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.9026030487181887e-06, |
|
"loss": 7.1178, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.8506017213040802e-06, |
|
"loss": 7.1171, |
|
"step": 7445 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.7990636909837163e-06, |
|
"loss": 7.0874, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.7479892035375154e-06, |
|
"loss": 7.0969, |
|
"step": 7455 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.6973785025352972e-06, |
|
"loss": 7.0708, |
|
"step": 7460 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.6472318293351286e-06, |
|
"loss": 7.111, |
|
"step": 7465 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.5975494230821575e-06, |
|
"loss": 7.0982, |
|
"step": 7470 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.548331520707492e-06, |
|
"loss": 7.1153, |
|
"step": 7475 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.4995783569270793e-06, |
|
"loss": 7.0845, |
|
"step": 7480 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.4512901642405163e-06, |
|
"loss": 7.131, |
|
"step": 7485 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.4034671729300408e-06, |
|
"loss": 7.1348, |
|
"step": 7490 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.356109611059387e-06, |
|
"loss": 7.13, |
|
"step": 7495 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.3092177044726437e-06, |
|
"loss": 7.0617, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.2627916767932854e-06, |
|
"loss": 7.0872, |
|
"step": 7505 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.216831749423054e-06, |
|
"loss": 7.1021, |
|
"step": 7510 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.1713381415408352e-06, |
|
"loss": 7.0934, |
|
"step": 7515 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.1263110701017164e-06, |
|
"loss": 7.051, |
|
"step": 7520 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.0817507498359644e-06, |
|
"loss": 7.1101, |
|
"step": 7525 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.0376573932478384e-06, |
|
"loss": 7.1125, |
|
"step": 7530 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.9940312106147886e-06, |
|
"loss": 7.1313, |
|
"step": 7535 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.9508724099863042e-06, |
|
"loss": 7.1289, |
|
"step": 7540 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.9081811971829787e-06, |
|
"loss": 7.0402, |
|
"step": 7545 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.8659577757955016e-06, |
|
"loss": 7.0607, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.8242023471837788e-06, |
|
"loss": 7.0747, |
|
"step": 7555 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.7829151104757802e-06, |
|
"loss": 7.0816, |
|
"step": 7560 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.7420962625668059e-06, |
|
"loss": 7.0988, |
|
"step": 7565 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.7017459981184202e-06, |
|
"loss": 7.116, |
|
"step": 7570 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.6618645095575425e-06, |
|
"loss": 7.1058, |
|
"step": 7575 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.6224519870755572e-06, |
|
"loss": 7.0819, |
|
"step": 7580 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.5835086186273718e-06, |
|
"loss": 7.1306, |
|
"step": 7585 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.5450345899305496e-06, |
|
"loss": 7.0262, |
|
"step": 7590 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.5070300844644114e-06, |
|
"loss": 7.0972, |
|
"step": 7595 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.4694952834691688e-06, |
|
"loss": 7.0783, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.4324303659450588e-06, |
|
"loss": 7.0594, |
|
"step": 7605 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.395835508651444e-06, |
|
"loss": 7.1183, |
|
"step": 7610 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.359710886106058e-06, |
|
"loss": 7.0961, |
|
"step": 7615 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.3240566705841062e-06, |
|
"loss": 7.0999, |
|
"step": 7620 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.2888730321174657e-06, |
|
"loss": 7.0765, |
|
"step": 7625 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.2541601384938651e-06, |
|
"loss": 7.1213, |
|
"step": 7630 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.2199181552561056e-06, |
|
"loss": 7.1006, |
|
"step": 7635 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.1861472457012301e-06, |
|
"loss": 7.0819, |
|
"step": 7640 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.1528475708798114e-06, |
|
"loss": 7.1316, |
|
"step": 7645 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.1200192895951311e-06, |
|
"loss": 7.0774, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.0876625584024026e-06, |
|
"loss": 7.1259, |
|
"step": 7655 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.055777531608093e-06, |
|
"loss": 7.0931, |
|
"step": 7660 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.0243643612691478e-06, |
|
"loss": 7.1131, |
|
"step": 7665 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 9.93423197192267e-07, |
|
"loss": 7.0825, |
|
"step": 7670 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 9.629541869331848e-07, |
|
"loss": 7.07, |
|
"step": 7675 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 9.329574757959813e-07, |
|
"loss": 7.1022, |
|
"step": 7680 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 9.034332068323825e-07, |
|
"loss": 7.0952, |
|
"step": 7685 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 8.743815208410833e-07, |
|
"loss": 7.089, |
|
"step": 7690 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 8.458025563670702e-07, |
|
"loss": 7.0758, |
|
"step": 7695 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 8.176964497009332e-07, |
|
"loss": 7.1161, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 7.900633348782882e-07, |
|
"loss": 7.1541, |
|
"step": 7705 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 7.629033436790777e-07, |
|
"loss": 7.0598, |
|
"step": 7710 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 7.362166056269493e-07, |
|
"loss": 7.105, |
|
"step": 7715 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 7.100032479886443e-07, |
|
"loss": 7.1209, |
|
"step": 7720 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 6.842633957734213e-07, |
|
"loss": 7.1054, |
|
"step": 7725 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 6.589971717323895e-07, |
|
"loss": 7.1273, |
|
"step": 7730 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 6.342046963580095e-07, |
|
"loss": 7.1387, |
|
"step": 7735 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 6.098860878834489e-07, |
|
"loss": 7.1474, |
|
"step": 7740 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 5.86041462282061e-07, |
|
"loss": 7.0272, |
|
"step": 7745 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 5.626709332667845e-07, |
|
"loss": 7.0742, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 5.397746122896785e-07, |
|
"loss": 7.0726, |
|
"step": 7755 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 5.173526085413105e-07, |
|
"loss": 7.1028, |
|
"step": 7760 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 4.954050289502799e-07, |
|
"loss": 7.0207, |
|
"step": 7765 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 4.7393197818270674e-07, |
|
"loss": 7.0802, |
|
"step": 7770 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 4.529335586417105e-07, |
|
"loss": 7.0851, |
|
"step": 7775 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 4.3240987046694327e-07, |
|
"loss": 7.1154, |
|
"step": 7780 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 4.1236101153409036e-07, |
|
"loss": 7.1015, |
|
"step": 7785 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 3.927870774544484e-07, |
|
"loss": 7.1106, |
|
"step": 7790 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 3.7368816157441476e-07, |
|
"loss": 7.1071, |
|
"step": 7795 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 3.550643549750543e-07, |
|
"loss": 7.1529, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 3.3691574647171096e-07, |
|
"loss": 7.1044, |
|
"step": 7805 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 3.1924242261351935e-07, |
|
"loss": 7.082, |
|
"step": 7810 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 3.020444676830492e-07, |
|
"loss": 7.126, |
|
"step": 7815 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.853219636958504e-07, |
|
"loss": 7.1182, |
|
"step": 7820 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.6907499040011995e-07, |
|
"loss": 7.0536, |
|
"step": 7825 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.533036252762466e-07, |
|
"loss": 7.0932, |
|
"step": 7830 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.3800794353651123e-07, |
|
"loss": 7.0596, |
|
"step": 7835 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.2318801812472035e-07, |
|
"loss": 7.1066, |
|
"step": 7840 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.0884391971579542e-07, |
|
"loss": 7.1337, |
|
"step": 7845 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.949757167155286e-07, |
|
"loss": 7.0695, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.8158347526014973e-07, |
|
"loss": 7.1138, |
|
"step": 7855 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.6866725921613758e-07, |
|
"loss": 7.1164, |
|
"step": 7860 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.5622713017978685e-07, |
|
"loss": 7.0941, |
|
"step": 7865 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.442631474770084e-07, |
|
"loss": 7.0946, |
|
"step": 7870 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.327753681630073e-07, |
|
"loss": 7.0712, |
|
"step": 7875 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.2176384702200505e-07, |
|
"loss": 7.0475, |
|
"step": 7880 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.1122863656698457e-07, |
|
"loss": 7.0865, |
|
"step": 7885 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.01169787039479e-07, |
|
"loss": 7.0835, |
|
"step": 7890 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 9.158734640924981e-08, |
|
"loss": 7.0951, |
|
"step": 7895 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 8.248136037415366e-08, |
|
"loss": 7.0965, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 7.385187235983138e-08, |
|
"loss": 7.0692, |
|
"step": 7905 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 6.569892351956375e-08, |
|
"loss": 7.1096, |
|
"step": 7910 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 5.8022552734082744e-08, |
|
"loss": 7.1295, |
|
"step": 7915 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 5.0822796611327275e-08, |
|
"loss": 7.0962, |
|
"step": 7920 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 4.4099689486298835e-08, |
|
"loss": 7.0861, |
|
"step": 7925 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 3.785326342092832e-08, |
|
"loss": 7.0959, |
|
"step": 7930 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 3.208354820388726e-08, |
|
"loss": 7.0745, |
|
"step": 7935 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.6790571350432393e-08, |
|
"loss": 7.083, |
|
"step": 7940 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.1974358102339054e-08, |
|
"loss": 7.0661, |
|
"step": 7945 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.7634931427679134e-08, |
|
"loss": 7.1285, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 1.3772312020832179e-08, |
|
"loss": 7.1251, |
|
"step": 7955 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 1.0386518302285542e-08, |
|
"loss": 7.123, |
|
"step": 7960 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 7.47756641860109e-09, |
|
"loss": 7.0505, |
|
"step": 7965 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 5.0454702423152754e-09, |
|
"loss": 7.1012, |
|
"step": 7970 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 3.0902413718836288e-09, |
|
"loss": 7.0782, |
|
"step": 7975 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 1.6118891316141416e-09, |
|
"loss": 7.0536, |
|
"step": 7980 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 6.104205716672695e-10, |
|
"loss": 7.0856, |
|
"step": 7985 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 8.584046794490875e-11, |
|
"loss": 7.112, |
|
"step": 7990 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 7.037853717803955, |
|
"eval_runtime": 1231.406, |
|
"eval_samples_per_second": 22.988, |
|
"eval_steps_per_second": 2.874, |
|
"step": 7993 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"step": 7993, |
|
"total_flos": 3821811133317120.0, |
|
"train_loss": 4.183625105175136, |
|
"train_runtime": 33502.9805, |
|
"train_samples_per_second": 7.635, |
|
"train_steps_per_second": 0.239 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 7993, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 100, |
|
"total_flos": 3821811133317120.0, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|