|
{ |
|
"best_metric": 19.615089840756195, |
|
"best_model_checkpoint": "./whisper-medium-tr/checkpoint-1000", |
|
"epoch": 0.17238407171177383, |
|
"eval_steps": 1000, |
|
"global_step": 1000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0034476814342354768, |
|
"grad_norm": 17.02838897705078, |
|
"learning_rate": 1.0000000000000002e-06, |
|
"loss": 1.0558, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.0068953628684709535, |
|
"grad_norm": 8.815332412719727, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 0.5997, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.01034304430270643, |
|
"grad_norm": 9.678966522216797, |
|
"learning_rate": 3e-06, |
|
"loss": 0.2506, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.013790725736941907, |
|
"grad_norm": 7.050973892211914, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 0.1829, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.017238407171177382, |
|
"grad_norm": 12.933165550231934, |
|
"learning_rate": 5e-06, |
|
"loss": 0.1818, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.02068608860541286, |
|
"grad_norm": 8.803665161132812, |
|
"learning_rate": 6e-06, |
|
"loss": 0.1803, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.024133770039648336, |
|
"grad_norm": 7.420976161956787, |
|
"learning_rate": 7e-06, |
|
"loss": 0.1767, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.027581451473883814, |
|
"grad_norm": 8.617654800415039, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 0.2164, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.03102913290811929, |
|
"grad_norm": 13.863837242126465, |
|
"learning_rate": 9e-06, |
|
"loss": 0.2168, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.034476814342354764, |
|
"grad_norm": 9.399210929870605, |
|
"learning_rate": 1e-05, |
|
"loss": 0.1798, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.037924495776590246, |
|
"grad_norm": 7.0017266273498535, |
|
"learning_rate": 9.92857142857143e-06, |
|
"loss": 0.1962, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.04137217721082572, |
|
"grad_norm": 13.431073188781738, |
|
"learning_rate": 9.857142857142859e-06, |
|
"loss": 0.2146, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.044819858645061196, |
|
"grad_norm": 6.425413608551025, |
|
"learning_rate": 9.785714285714286e-06, |
|
"loss": 0.1956, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.04826754007929667, |
|
"grad_norm": 8.163415908813477, |
|
"learning_rate": 9.714285714285715e-06, |
|
"loss": 0.193, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.05171522151353215, |
|
"grad_norm": 12.302197456359863, |
|
"learning_rate": 9.642857142857144e-06, |
|
"loss": 0.2102, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.05516290294776763, |
|
"grad_norm": 10.398489952087402, |
|
"learning_rate": 9.571428571428573e-06, |
|
"loss": 0.221, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.0586105843820031, |
|
"grad_norm": 10.396890640258789, |
|
"learning_rate": 9.5e-06, |
|
"loss": 0.1642, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.06205826581623858, |
|
"grad_norm": 10.882383346557617, |
|
"learning_rate": 9.42857142857143e-06, |
|
"loss": 0.2076, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.06550594725047405, |
|
"grad_norm": 9.351667404174805, |
|
"learning_rate": 9.357142857142859e-06, |
|
"loss": 0.1888, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.06895362868470953, |
|
"grad_norm": 8.273520469665527, |
|
"learning_rate": 9.285714285714288e-06, |
|
"loss": 0.2302, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.072401310118945, |
|
"grad_norm": 9.05259895324707, |
|
"learning_rate": 9.214285714285715e-06, |
|
"loss": 0.2045, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.07584899155318049, |
|
"grad_norm": 10.202899932861328, |
|
"learning_rate": 9.142857142857144e-06, |
|
"loss": 0.2022, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.07929667298741597, |
|
"grad_norm": 7.107177734375, |
|
"learning_rate": 9.071428571428573e-06, |
|
"loss": 0.172, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.08274435442165144, |
|
"grad_norm": 10.326031684875488, |
|
"learning_rate": 9e-06, |
|
"loss": 0.2856, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.08619203585588692, |
|
"grad_norm": 5.693482875823975, |
|
"learning_rate": 8.92857142857143e-06, |
|
"loss": 0.2283, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.08963971729012239, |
|
"grad_norm": 14.126120567321777, |
|
"learning_rate": 8.857142857142858e-06, |
|
"loss": 0.2095, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.09308739872435787, |
|
"grad_norm": 7.242938041687012, |
|
"learning_rate": 8.785714285714286e-06, |
|
"loss": 0.2102, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.09653508015859334, |
|
"grad_norm": 8.650644302368164, |
|
"learning_rate": 8.714285714285715e-06, |
|
"loss": 0.2419, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.09998276159282882, |
|
"grad_norm": 8.036280632019043, |
|
"learning_rate": 8.642857142857144e-06, |
|
"loss": 0.2031, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.1034304430270643, |
|
"grad_norm": 8.81808090209961, |
|
"learning_rate": 8.571428571428571e-06, |
|
"loss": 0.1972, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.10687812446129978, |
|
"grad_norm": 13.098625183105469, |
|
"learning_rate": 8.5e-06, |
|
"loss": 0.2224, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.11032580589553526, |
|
"grad_norm": 6.834613800048828, |
|
"learning_rate": 8.428571428571429e-06, |
|
"loss": 0.1894, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.11377348732977073, |
|
"grad_norm": 11.371912956237793, |
|
"learning_rate": 8.357142857142858e-06, |
|
"loss": 0.2268, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.1172211687640062, |
|
"grad_norm": 11.41600513458252, |
|
"learning_rate": 8.285714285714287e-06, |
|
"loss": 0.1934, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.12066885019824168, |
|
"grad_norm": 9.144827842712402, |
|
"learning_rate": 8.214285714285714e-06, |
|
"loss": 0.1878, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.12411653163247716, |
|
"grad_norm": 3.3032901287078857, |
|
"learning_rate": 8.142857142857143e-06, |
|
"loss": 0.1956, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.12756421306671265, |
|
"grad_norm": 8.833490371704102, |
|
"learning_rate": 8.071428571428572e-06, |
|
"loss": 0.2416, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.1310118945009481, |
|
"grad_norm": 9.268789291381836, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 0.2008, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.1344595759351836, |
|
"grad_norm": 7.276330471038818, |
|
"learning_rate": 7.928571428571429e-06, |
|
"loss": 0.1955, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.13790725736941906, |
|
"grad_norm": 9.219639778137207, |
|
"learning_rate": 7.857142857142858e-06, |
|
"loss": 0.1779, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.14135493880365455, |
|
"grad_norm": 8.820664405822754, |
|
"learning_rate": 7.785714285714287e-06, |
|
"loss": 0.154, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.14480262023789, |
|
"grad_norm": 10.620595932006836, |
|
"learning_rate": 7.714285714285716e-06, |
|
"loss": 0.1912, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.1482503016721255, |
|
"grad_norm": 7.289504051208496, |
|
"learning_rate": 7.642857142857143e-06, |
|
"loss": 0.2045, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.15169798310636098, |
|
"grad_norm": 6.108071804046631, |
|
"learning_rate": 7.571428571428572e-06, |
|
"loss": 0.1861, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.15514566454059645, |
|
"grad_norm": 5.369969367980957, |
|
"learning_rate": 7.500000000000001e-06, |
|
"loss": 0.17, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.15859334597483193, |
|
"grad_norm": 10.743534088134766, |
|
"learning_rate": 7.428571428571429e-06, |
|
"loss": 0.1768, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.1620410274090674, |
|
"grad_norm": 5.289422988891602, |
|
"learning_rate": 7.357142857142858e-06, |
|
"loss": 0.2147, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.16548870884330288, |
|
"grad_norm": 8.861351013183594, |
|
"learning_rate": 7.285714285714286e-06, |
|
"loss": 0.1999, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.16893639027753835, |
|
"grad_norm": 8.204620361328125, |
|
"learning_rate": 7.2142857142857145e-06, |
|
"loss": 0.1898, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.17238407171177383, |
|
"grad_norm": 8.618937492370605, |
|
"learning_rate": 7.1428571428571436e-06, |
|
"loss": 0.1504, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.17238407171177383, |
|
"eval_loss": 0.24223625659942627, |
|
"eval_runtime": 13074.9332, |
|
"eval_samples_per_second": 0.863, |
|
"eval_steps_per_second": 0.108, |
|
"eval_wer": 19.615089840756195, |
|
"step": 1000 |
|
} |
|
], |
|
"logging_steps": 20, |
|
"max_steps": 3000, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 1000, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 8.16483926016e+18, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|