{
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 0.9999305603777515,
  "eval_steps": 1000,
  "global_step": 14400,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 6.943962224845496e-05,
      "grad_norm": 9.8125,
      "learning_rate": 2e-06,
      "loss": 0.7188,
      "step": 1
    },
    {
      "epoch": 0.006943962224845497,
      "grad_norm": 0.1376953125,
      "learning_rate": 0.0002,
      "loss": 0.3907,
      "step": 100
    },
    {
      "epoch": 0.013887924449690994,
      "grad_norm": 0.0849609375,
      "learning_rate": 0.0004,
      "loss": 0.2341,
      "step": 200
    },
    {
      "epoch": 0.02083188667453649,
      "grad_norm": 0.146484375,
      "learning_rate": 0.0006,
      "loss": 0.2181,
      "step": 300
    },
    {
      "epoch": 0.027775848899381988,
      "grad_norm": 0.10546875,
      "learning_rate": 0.0008,
      "loss": 0.197,
      "step": 400
    },
    {
      "epoch": 0.034719811124227486,
      "grad_norm": 0.11376953125,
      "learning_rate": 0.001,
      "loss": 0.1695,
      "step": 500
    },
    {
      "epoch": 0.04166377334907298,
      "grad_norm": 0.0849609375,
      "learning_rate": 0.0012,
      "loss": 0.14,
      "step": 600
    },
    {
      "epoch": 0.04860773557391848,
      "grad_norm": 0.078125,
      "learning_rate": 0.0014,
      "loss": 0.1231,
      "step": 700
    },
    {
      "epoch": 0.055551697798763976,
      "grad_norm": 0.0712890625,
      "learning_rate": 0.0016,
      "loss": 0.1141,
      "step": 800
    },
    {
      "epoch": 0.06249566002360947,
      "grad_norm": 0.06298828125,
      "learning_rate": 0.0018000000000000002,
      "loss": 0.1077,
      "step": 900
    },
    {
      "epoch": 0.06943962224845497,
      "grad_norm": 0.0615234375,
      "learning_rate": 0.002,
      "loss": 0.1045,
      "step": 1000
    },
    {
      "epoch": 0.06943962224845497,
      "eval_covost2-en-de_loss": 1.4858413934707642,
      "eval_covost2-en-de_runtime": 32.1589,
      "eval_covost2-en-de_samples_per_second": 1.99,
      "eval_covost2-en-de_steps_per_second": 0.062,
      "step": 1000
    },
    {
      "epoch": 0.06943962224845497,
      "eval_covost2-zh-en_loss": 2.7152516841888428,
      "eval_covost2-zh-en_runtime": 31.3842,
      "eval_covost2-zh-en_samples_per_second": 2.039,
      "eval_covost2-zh-en_steps_per_second": 0.064,
      "step": 1000
    },
    {
      "epoch": 0.06943962224845497,
      "eval_peoplespeech-clean-transcription_loss": 2.0398874282836914,
      "eval_peoplespeech-clean-transcription_runtime": 32.088,
      "eval_peoplespeech-clean-transcription_samples_per_second": 1.995,
      "eval_peoplespeech-clean-transcription_steps_per_second": 0.062,
      "step": 1000
    },
    {
      "epoch": 0.07638358447330046,
      "grad_norm": 0.053955078125,
      "learning_rate": 0.001999725185109816,
      "loss": 0.101,
      "step": 1100
    },
    {
      "epoch": 0.08332754669814596,
      "grad_norm": 0.0517578125,
      "learning_rate": 0.0019989008914857113,
      "loss": 0.0956,
      "step": 1200
    },
    {
      "epoch": 0.09027150892299146,
      "grad_norm": 0.04443359375,
      "learning_rate": 0.00199752757218401,
      "loss": 0.0936,
      "step": 1300
    },
    {
      "epoch": 0.09721547114783696,
      "grad_norm": 0.0390625,
      "learning_rate": 0.001995605982021898,
      "loss": 0.0917,
      "step": 1400
    },
    {
      "epoch": 0.10415943337268245,
      "grad_norm": 0.0517578125,
      "learning_rate": 0.0019931371771625545,
      "loss": 0.0894,
      "step": 1500
    },
    {
      "epoch": 0.11110339559752795,
      "grad_norm": 0.0419921875,
      "learning_rate": 0.001990122514534651,
      "loss": 0.0868,
      "step": 1600
    },
    {
      "epoch": 0.11804735782237345,
      "grad_norm": 0.039306640625,
      "learning_rate": 0.0019865636510865464,
      "loss": 0.0861,
      "step": 1700
    },
    {
      "epoch": 0.12499132004721894,
      "grad_norm": 0.047119140625,
      "learning_rate": 0.001982462542875576,
      "loss": 0.0854,
      "step": 1800
    },
    {
      "epoch": 0.13193528227206444,
      "grad_norm": 0.0390625,
      "learning_rate": 0.001977821443992945,
      "loss": 0.0837,
      "step": 1900
    },
    {
      "epoch": 0.13887924449690994,
      "grad_norm": 0.04052734375,
      "learning_rate": 0.001972642905324813,
      "loss": 0.0818,
      "step": 2000
    },
    {
      "epoch": 0.13887924449690994,
      "eval_covost2-en-de_loss": 1.4137890338897705,
      "eval_covost2-en-de_runtime": 32.5714,
      "eval_covost2-en-de_samples_per_second": 1.965,
      "eval_covost2-en-de_steps_per_second": 0.061,
      "step": 2000
    },
    {
      "epoch": 0.13887924449690994,
      "eval_covost2-zh-en_loss": 2.667837381362915,
      "eval_covost2-zh-en_runtime": 31.1685,
      "eval_covost2-zh-en_samples_per_second": 2.053,
      "eval_covost2-zh-en_steps_per_second": 0.064,
      "step": 2000
    },
    {
      "epoch": 0.13887924449690994,
      "eval_peoplespeech-clean-transcription_loss": 1.835880160331726,
      "eval_peoplespeech-clean-transcription_runtime": 32.0265,
      "eval_peoplespeech-clean-transcription_samples_per_second": 1.998,
      "eval_peoplespeech-clean-transcription_steps_per_second": 0.062,
      "step": 2000
    },
    {
      "epoch": 0.14582320672175544,
      "grad_norm": 0.039794921875,
      "learning_rate": 0.0019669297731502505,
      "loss": 0.0813,
      "step": 2100
    },
    {
      "epoch": 0.15276716894660092,
      "grad_norm": 0.03515625,
      "learning_rate": 0.00196068518757684,
      "loss": 0.0811,
      "step": 2200
    },
    {
      "epoch": 0.15971113117144642,
      "grad_norm": 0.04443359375,
      "learning_rate": 0.001953912580814779,
      "loss": 0.0793,
      "step": 2300
    },
    {
      "epoch": 0.16665509339629192,
      "grad_norm": 0.037841796875,
      "learning_rate": 0.0019466156752904343,
      "loss": 0.0788,
      "step": 2400
    },
    {
      "epoch": 0.17359905562113742,
      "grad_norm": 0.04052734375,
      "learning_rate": 0.0019387984816003866,
      "loss": 0.0783,
      "step": 2500
    },
    {
      "epoch": 0.18054301784598292,
      "grad_norm": 0.03466796875,
      "learning_rate": 0.0019304652963070869,
      "loss": 0.0772,
      "step": 2600
    },
    {
      "epoch": 0.18748698007082842,
      "grad_norm": 0.036376953125,
      "learning_rate": 0.0019216206995773372,
      "loss": 0.0771,
      "step": 2700
    },
    {
      "epoch": 0.19443094229567393,
      "grad_norm": 0.0400390625,
      "learning_rate": 0.0019122695526648968,
      "loss": 0.0766,
      "step": 2800
    },
    {
      "epoch": 0.2013749045205194,
      "grad_norm": 0.0361328125,
      "learning_rate": 0.0019024169952385887,
      "loss": 0.0753,
      "step": 2900
    },
    {
      "epoch": 0.2083188667453649,
      "grad_norm": 0.03125,
      "learning_rate": 0.0018920684425573864,
      "loss": 0.075,
      "step": 3000
    },
    {
      "epoch": 0.2083188667453649,
      "eval_covost2-en-de_loss": 1.3828500509262085,
      "eval_covost2-en-de_runtime": 32.2972,
      "eval_covost2-en-de_samples_per_second": 1.982,
      "eval_covost2-en-de_steps_per_second": 0.062,
      "step": 3000
    },
    {
      "epoch": 0.2083188667453649,
      "eval_covost2-zh-en_loss": 2.646721839904785,
      "eval_covost2-zh-en_runtime": 31.0128,
      "eval_covost2-zh-en_samples_per_second": 2.064,
      "eval_covost2-zh-en_steps_per_second": 0.064,
      "step": 3000
    },
    {
      "epoch": 0.2083188667453649,
      "eval_peoplespeech-clean-transcription_loss": 1.773127555847168,
      "eval_peoplespeech-clean-transcription_runtime": 32.5192,
      "eval_peoplespeech-clean-transcription_samples_per_second": 1.968,
      "eval_peoplespeech-clean-transcription_steps_per_second": 0.062,
      "step": 3000
    },
    {
      "epoch": 0.2152628289702104,
      "grad_norm": 0.03369140625,
      "learning_rate": 0.0018812295824940284,
      "loss": 0.0743,
      "step": 3100
    },
    {
      "epoch": 0.2222067911950559,
      "grad_norm": 0.034912109375,
      "learning_rate": 0.0018699063724087904,
      "loss": 0.074,
      "step": 3200
    },
    {
      "epoch": 0.2291507534199014,
      "grad_norm": 0.033203125,
      "learning_rate": 0.0018581050358751443,
      "loss": 0.0742,
      "step": 3300
    },
    {
      "epoch": 0.2360947156447469,
      "grad_norm": 0.03857421875,
      "learning_rate": 0.0018458320592590974,
      "loss": 0.0742,
      "step": 3400
    },
    {
      "epoch": 0.24303867786959238,
      "grad_norm": 0.033935546875,
      "learning_rate": 0.0018330941881540914,
      "loss": 0.0728,
      "step": 3500
    },
    {
      "epoch": 0.24998264009443788,
      "grad_norm": 0.031005859375,
      "learning_rate": 0.0018198984236734246,
      "loss": 0.0728,
      "step": 3600
    },
    {
      "epoch": 0.2569266023192834,
      "grad_norm": 0.030029296875,
      "learning_rate": 0.0018062520186022297,
      "loss": 0.0714,
      "step": 3700
    },
    {
      "epoch": 0.2638705645441289,
      "grad_norm": 0.02734375,
      "learning_rate": 0.0017921624734111292,
      "loss": 0.071,
      "step": 3800
    },
    {
      "epoch": 0.2708145267689744,
      "grad_norm": 0.03271484375,
      "learning_rate": 0.001777637532133752,
      "loss": 0.0705,
      "step": 3900
    },
    {
      "epoch": 0.2777584889938199,
      "grad_norm": 0.035400390625,
      "learning_rate": 0.0017626851781103819,
      "loss": 0.0714,
      "step": 4000
    },
    {
      "epoch": 0.2777584889938199,
      "eval_covost2-en-de_loss": 1.3778624534606934,
      "eval_covost2-en-de_runtime": 32.6181,
      "eval_covost2-en-de_samples_per_second": 1.962,
      "eval_covost2-en-de_steps_per_second": 0.061,
      "step": 4000
    },
    {
      "epoch": 0.2777584889938199,
      "eval_covost2-zh-en_loss": 2.6438870429992676,
      "eval_covost2-zh-en_runtime": 31.4603,
      "eval_covost2-zh-en_samples_per_second": 2.034,
      "eval_covost2-zh-en_steps_per_second": 0.064,
      "step": 4000
    },
    {
      "epoch": 0.2777584889938199,
      "eval_peoplespeech-clean-transcription_loss": 1.7361584901809692,
      "eval_peoplespeech-clean-transcription_runtime": 32.442,
      "eval_peoplespeech-clean-transcription_samples_per_second": 1.973,
      "eval_peoplespeech-clean-transcription_steps_per_second": 0.062,
      "step": 4000
    },
    {
      "epoch": 0.2847024512186654,
      "grad_norm": 0.0281982421875,
      "learning_rate": 0.001747313629600077,
      "loss": 0.0713,
      "step": 4100
    },
    {
      "epoch": 0.2916464134435109,
      "grad_norm": 0.028076171875,
      "learning_rate": 0.001731531335263669,
      "loss": 0.0699,
      "step": 4200
    },
    {
      "epoch": 0.2985903756683564,
      "grad_norm": 0.0277099609375,
      "learning_rate": 0.0017153469695201276,
      "loss": 0.0702,
      "step": 4300
    },
    {
      "epoch": 0.30553433789320184,
      "grad_norm": 0.031982421875,
      "learning_rate": 0.0016987694277788418,
      "loss": 0.0692,
      "step": 4400
    },
    {
      "epoch": 0.31247830011804734,
      "grad_norm": 0.02880859375,
      "learning_rate": 0.001681807821550438,
      "loss": 0.0686,
      "step": 4500
    },
    {
      "epoch": 0.31942226234289284,
      "grad_norm": 0.0289306640625,
      "learning_rate": 0.0016644714734388218,
      "loss": 0.0698,
      "step": 4600
    },
    {
      "epoch": 0.32636622456773834,
      "grad_norm": 0.029541015625,
      "learning_rate": 0.0016467699120171987,
      "loss": 0.0683,
      "step": 4700
    },
    {
      "epoch": 0.33331018679258384,
      "grad_norm": 0.034423828125,
      "learning_rate": 0.001628712866590885,
      "loss": 0.0687,
      "step": 4800
    },
    {
      "epoch": 0.34025414901742934,
      "grad_norm": 0.0289306640625,
      "learning_rate": 0.0016103102618497923,
      "loss": 0.0684,
      "step": 4900
    },
    {
      "epoch": 0.34719811124227484,
      "grad_norm": 0.0263671875,
      "learning_rate": 0.0015915722124135226,
      "loss": 0.0681,
      "step": 5000
    },
    {
      "epoch": 0.34719811124227484,
      "eval_covost2-en-de_loss": 1.3711879253387451,
      "eval_covost2-en-de_runtime": 32.6293,
      "eval_covost2-en-de_samples_per_second": 1.961,
      "eval_covost2-en-de_steps_per_second": 0.061,
      "step": 5000
    },
    {
      "epoch": 0.34719811124227484,
      "eval_covost2-zh-en_loss": 2.6346511840820312,
      "eval_covost2-zh-en_runtime": 32.1513,
      "eval_covost2-zh-en_samples_per_second": 1.991,
      "eval_covost2-zh-en_steps_per_second": 0.062,
      "step": 5000
    },
    {
      "epoch": 0.34719811124227484,
      "eval_peoplespeech-clean-transcription_loss": 1.7350472211837769,
      "eval_peoplespeech-clean-transcription_runtime": 32.5813,
      "eval_peoplespeech-clean-transcription_samples_per_second": 1.964,
      "eval_peoplespeech-clean-transcription_steps_per_second": 0.061,
      "step": 5000
    },
    {
      "epoch": 0.35414207346712034,
      "grad_norm": 0.028076171875,
      "learning_rate": 0.001572509017272072,
      "loss": 0.0693,
      "step": 5100
    },
    {
      "epoch": 0.36108603569196585,
      "grad_norm": 0.030517578125,
      "learning_rate": 0.0015531311541251993,
      "loss": 0.0683,
      "step": 5200
    },
    {
      "epoch": 0.36802999791681135,
      "grad_norm": 0.031494140625,
      "learning_rate": 0.0015334492736235703,
      "loss": 0.0677,
      "step": 5300
    },
    {
      "epoch": 0.37497396014165685,
      "grad_norm": 0.0284423828125,
      "learning_rate": 0.0015134741935148419,
      "loss": 0.0669,
      "step": 5400
    },
    {
      "epoch": 0.38191792236650235,
      "grad_norm": 0.0302734375,
      "learning_rate": 0.0014932168926979072,
      "loss": 0.0669,
      "step": 5500
    },
    {
      "epoch": 0.38886188459134785,
      "grad_norm": 0.0260009765625,
      "learning_rate": 0.0014726885051885652,
      "loss": 0.0666,
      "step": 5600
    },
    {
      "epoch": 0.3958058468161933,
      "grad_norm": 0.033447265625,
      "learning_rate": 0.0014519003139999338,
      "loss": 0.0659,
      "step": 5700
    },
    {
      "epoch": 0.4027498090410388,
      "grad_norm": 0.0283203125,
      "learning_rate": 0.0014308637449409706,
      "loss": 0.0653,
      "step": 5800
    },
    {
      "epoch": 0.4096937712658843,
      "grad_norm": 0.0283203125,
      "learning_rate": 0.0014095903603365066,
      "loss": 0.0662,
      "step": 5900
    },
    {
      "epoch": 0.4166377334907298,
      "grad_norm": 0.0267333984375,
      "learning_rate": 0.0013880918526722496,
      "loss": 0.0665,
      "step": 6000
    },
    {
      "epoch": 0.4166377334907298,
      "eval_covost2-en-de_loss": 1.3651559352874756,
      "eval_covost2-en-de_runtime": 32.5621,
      "eval_covost2-en-de_samples_per_second": 1.965,
      "eval_covost2-en-de_steps_per_second": 0.061,
      "step": 6000
    },
    {
      "epoch": 0.4166377334907298,
      "eval_covost2-zh-en_loss": 2.6372551918029785,
      "eval_covost2-zh-en_runtime": 31.028,
      "eval_covost2-zh-en_samples_per_second": 2.063,
      "eval_covost2-zh-en_steps_per_second": 0.064,
      "step": 6000
    },
    {
      "epoch": 0.4166377334907298,
      "eval_peoplespeech-clean-transcription_loss": 1.7209596633911133,
      "eval_peoplespeech-clean-transcription_runtime": 32.6773,
      "eval_peoplespeech-clean-transcription_samples_per_second": 1.959,
      "eval_peoplespeech-clean-transcription_steps_per_second": 0.061,
      "step": 6000
    },
    {
      "epoch": 0.4235816957155753,
      "grad_norm": 0.028564453125,
      "learning_rate": 0.0013663800381682463,
      "loss": 0.0658,
      "step": 6100
    },
    {
      "epoch": 0.4305256579404208,
      "grad_norm": 0.0299072265625,
      "learning_rate": 0.0013444668502843329,
      "loss": 0.0657,
      "step": 6200
    },
    {
      "epoch": 0.4374696201652663,
      "grad_norm": 0.0296630859375,
      "learning_rate": 0.0013223643331611537,
      "loss": 0.0655,
      "step": 6300
    },
    {
      "epoch": 0.4444135823901118,
      "grad_norm": 0.0286865234375,
      "learning_rate": 0.001300084635000341,
      "loss": 0.0654,
      "step": 6400
    },
    {
      "epoch": 0.4513575446149573,
      "grad_norm": 0.028564453125,
      "learning_rate": 0.0012776400013875004,
      "loss": 0.0655,
      "step": 6500
    },
    {
      "epoch": 0.4583015068398028,
      "grad_norm": 0.030029296875,
      "learning_rate": 0.0012550427685616766,
      "loss": 0.0648,
      "step": 6600
    },
    {
      "epoch": 0.4652454690646483,
      "grad_norm": 0.037109375,
      "learning_rate": 0.0012323053566349834,
      "loss": 0.0654,
      "step": 6700
    },
    {
      "epoch": 0.4721894312894938,
      "grad_norm": 0.029296875,
      "learning_rate": 0.0012094402627661448,
      "loss": 0.0643,
      "step": 6800
    },
    {
      "epoch": 0.47913339351433926,
      "grad_norm": 0.030517578125,
      "learning_rate": 0.0011864600542916813,
      "loss": 0.0646,
      "step": 6900
    },
    {
      "epoch": 0.48607735573918476,
      "grad_norm": 0.037353515625,
      "learning_rate": 0.0011633773618185302,
      "loss": 0.0642,
      "step": 7000
    },
    {
      "epoch": 0.48607735573918476,
      "eval_covost2-en-de_loss": 1.3594859838485718,
      "eval_covost2-en-de_runtime": 32.6807,
      "eval_covost2-en-de_samples_per_second": 1.958,
      "eval_covost2-en-de_steps_per_second": 0.061,
      "step": 7000
    },
    {
      "epoch": 0.48607735573918476,
      "eval_covost2-zh-en_loss": 2.626713514328003,
      "eval_covost2-zh-en_runtime": 31.0228,
      "eval_covost2-zh-en_samples_per_second": 2.063,
      "eval_covost2-zh-en_steps_per_second": 0.064,
      "step": 7000
    },
    {
      "epoch": 0.48607735573918476,
      "eval_peoplespeech-clean-transcription_loss": 1.693739652633667,
      "eval_peoplespeech-clean-transcription_runtime": 31.9776,
      "eval_peoplespeech-clean-transcription_samples_per_second": 2.001,
      "eval_peoplespeech-clean-transcription_steps_per_second": 0.063,
      "step": 7000
    },
    {
      "epoch": 0.49302131796403026,
      "grad_norm": 0.02978515625,
      "learning_rate": 0.0011402048722818862,
      "loss": 0.0656,
      "step": 7100
    },
    {
      "epoch": 0.49996528018887576,
      "grad_norm": 0.0281982421875,
      "learning_rate": 0.0011169553219720827,
      "loss": 0.064,
      "step": 7200
    },
    {
      "epoch": 0.5069092424137213,
      "grad_norm": 0.027099609375,
      "learning_rate": 0.001093641489534351,
      "loss": 0.0648,
      "step": 7300
    },
    {
      "epoch": 0.5138532046385668,
      "grad_norm": 0.02783203125,
      "learning_rate": 0.001070276188945293,
      "loss": 0.0642,
      "step": 7400
    },
    {
      "epoch": 0.5207971668634123,
      "grad_norm": 0.0281982421875,
      "learning_rate": 0.00104687226246994,
      "loss": 0.0635,
      "step": 7500
    },
    {
      "epoch": 0.5277411290882578,
      "grad_norm": 0.030517578125,
      "learning_rate": 0.0010234425736032607,
      "loss": 0.0644,
      "step": 7600
    },
    {
      "epoch": 0.5346850913131033,
      "grad_norm": 0.02978515625,
      "learning_rate": 0.001,
      "loss": 0.0645,
      "step": 7700
    },
    {
      "epoch": 0.5416290535379488,
      "grad_norm": 0.0250244140625,
      "learning_rate": 0.0009765574263967396,
      "loss": 0.0635,
      "step": 7800
    },
    {
      "epoch": 0.5485730157627943,
      "grad_norm": 0.033203125,
      "learning_rate": 0.0009531277375300599,
      "loss": 0.0635,
      "step": 7900
    },
    {
      "epoch": 0.5555169779876398,
      "grad_norm": 0.03173828125,
      "learning_rate": 0.0009297238110547074,
      "loss": 0.0624,
      "step": 8000
    },
    {
      "epoch": 0.5555169779876398,
      "eval_covost2-en-de_loss": 1.3578505516052246,
      "eval_covost2-en-de_runtime": 32.6942,
      "eval_covost2-en-de_samples_per_second": 1.958,
      "eval_covost2-en-de_steps_per_second": 0.061,
      "step": 8000
    },
    {
      "epoch": 0.5555169779876398,
      "eval_covost2-zh-en_loss": 2.6352272033691406,
      "eval_covost2-zh-en_runtime": 32.2689,
      "eval_covost2-zh-en_samples_per_second": 1.983,
      "eval_covost2-zh-en_steps_per_second": 0.062,
      "step": 8000
    },
    {
      "epoch": 0.5555169779876398,
      "eval_peoplespeech-clean-transcription_loss": 1.6875255107879639,
      "eval_peoplespeech-clean-transcription_runtime": 34.4533,
      "eval_peoplespeech-clean-transcription_samples_per_second": 1.858,
      "eval_peoplespeech-clean-transcription_steps_per_second": 0.058,
      "step": 8000
    },
    {
      "epoch": 0.5624609402124853,
      "grad_norm": 0.0274658203125,
      "learning_rate": 0.0009063585104656494,
      "loss": 0.064,
      "step": 8100
    },
    {
      "epoch": 0.5694049024373308,
      "grad_norm": 0.033935546875,
      "learning_rate": 0.0008830446780279176,
      "loss": 0.0626,
      "step": 8200
    },
    {
      "epoch": 0.5763488646621763,
      "grad_norm": 0.03955078125,
      "learning_rate": 0.0008597951277181142,
      "loss": 0.0638,
      "step": 8300
    },
    {
      "epoch": 0.5832928268870218,
      "grad_norm": 0.02734375,
      "learning_rate": 0.0008366226381814697,
      "loss": 0.0632,
      "step": 8400
    },
    {
      "epoch": 0.5902367891118673,
      "grad_norm": 0.0269775390625,
      "learning_rate": 0.000813539945708319,
      "loss": 0.0624,
      "step": 8500
    },
    {
      "epoch": 0.5971807513367128,
      "grad_norm": 0.028564453125,
      "learning_rate": 0.0007905597372338558,
      "loss": 0.063,
      "step": 8600
    },
    {
      "epoch": 0.6041247135615583,
      "grad_norm": 0.029541015625,
      "learning_rate": 0.0007676946433650169,
      "loss": 0.0626,
      "step": 8700
    },
    {
      "epoch": 0.6110686757864037,
      "grad_norm": 0.0272216796875,
      "learning_rate": 0.0007449572314383236,
      "loss": 0.0634,
      "step": 8800
    },
    {
      "epoch": 0.6180126380112492,
      "grad_norm": 0.034423828125,
      "learning_rate": 0.0007223599986124993,
      "loss": 0.0625,
      "step": 8900
    },
    {
      "epoch": 0.6249566002360947,
      "grad_norm": 0.03173828125,
      "learning_rate": 0.0006999153649996595,
      "loss": 0.0618,
      "step": 9000
    },
    {
      "epoch": 0.6249566002360947,
      "eval_covost2-en-de_loss": 1.3503832817077637,
      "eval_covost2-en-de_runtime": 32.9543,
      "eval_covost2-en-de_samples_per_second": 1.942,
      "eval_covost2-en-de_steps_per_second": 0.061,
      "step": 9000
    },
    {
      "epoch": 0.6249566002360947,
      "eval_covost2-zh-en_loss": 2.6205379962921143,
      "eval_covost2-zh-en_runtime": 31.3271,
      "eval_covost2-zh-en_samples_per_second": 2.043,
      "eval_covost2-zh-en_steps_per_second": 0.064,
      "step": 9000
    },
    {
      "epoch": 0.6249566002360947,
      "eval_peoplespeech-clean-transcription_loss": 1.685594916343689,
      "eval_peoplespeech-clean-transcription_runtime": 32.7924,
      "eval_peoplespeech-clean-transcription_samples_per_second": 1.952,
      "eval_peoplespeech-clean-transcription_steps_per_second": 0.061,
      "step": 9000
    },
    {
      "epoch": 0.6319005624609402,
      "grad_norm": 0.034912109375,
      "learning_rate": 0.0006776356668388464,
      "loss": 0.0614,
      "step": 9100
    },
    {
      "epoch": 0.6388445246857857,
      "grad_norm": 0.0277099609375,
      "learning_rate": 0.0006555331497156671,
      "loss": 0.0616,
      "step": 9200
    },
    {
      "epoch": 0.6457884869106312,
      "grad_norm": 0.031005859375,
      "learning_rate": 0.0006336199618317538,
      "loss": 0.0614,
      "step": 9300
    },
    {
      "epoch": 0.6527324491354767,
      "grad_norm": 0.0322265625,
      "learning_rate": 0.0006119081473277501,
      "loss": 0.0616,
      "step": 9400
    },
    {
      "epoch": 0.6596764113603222,
      "grad_norm": 0.029296875,
      "learning_rate": 0.0005904096396634935,
      "loss": 0.0609,
      "step": 9500
    },
    {
      "epoch": 0.6666203735851677,
      "grad_norm": 0.0284423828125,
      "learning_rate": 0.0005691362550590297,
      "loss": 0.0609,
      "step": 9600
    },
    {
      "epoch": 0.6735643358100132,
      "grad_norm": 0.0296630859375,
      "learning_rate": 0.0005480996860000663,
      "loss": 0.0611,
      "step": 9700
    },
    {
      "epoch": 0.6805082980348587,
      "grad_norm": 0.031494140625,
      "learning_rate": 0.0005273114948114346,
      "loss": 0.0612,
      "step": 9800
    },
    {
      "epoch": 0.6874522602597042,
      "grad_norm": 0.028564453125,
      "learning_rate": 0.0005067831073020928,
      "loss": 0.0615,
      "step": 9900
    },
    {
      "epoch": 0.6943962224845497,
      "grad_norm": 0.02783203125,
      "learning_rate": 0.00048652580648515787,
      "loss": 0.0621,
      "step": 10000
    },
    {
      "epoch": 0.6943962224845497,
      "eval_covost2-en-de_loss": 1.3505414724349976,
      "eval_covost2-en-de_runtime": 32.9697,
      "eval_covost2-en-de_samples_per_second": 1.941,
      "eval_covost2-en-de_steps_per_second": 0.061,
      "step": 10000
    },
    {
      "epoch": 0.6943962224845497,
      "eval_covost2-zh-en_loss": 2.613917827606201,
      "eval_covost2-zh-en_runtime": 31.7013,
      "eval_covost2-zh-en_samples_per_second": 2.019,
      "eval_covost2-zh-en_steps_per_second": 0.063,
      "step": 10000
    },
    {
      "epoch": 0.6943962224845497,
      "eval_peoplespeech-clean-transcription_loss": 1.6754916906356812,
      "eval_peoplespeech-clean-transcription_runtime": 33.2239,
      "eval_peoplespeech-clean-transcription_samples_per_second": 1.926,
      "eval_peoplespeech-clean-transcription_steps_per_second": 0.06,
      "step": 10000
    },
    {
      "epoch": 0.7013401847093952,
      "grad_norm": 0.0299072265625,
      "learning_rate": 0.0004665507263764299,
      "loss": 0.0614,
      "step": 10100
    },
    {
      "epoch": 0.7082841469342407,
      "grad_norm": 0.0272216796875,
      "learning_rate": 0.0004468688458748006,
      "loss": 0.0616,
      "step": 10200
    },
    {
      "epoch": 0.7152281091590862,
      "grad_norm": 0.03662109375,
      "learning_rate": 0.0004274909827279283,
      "loss": 0.0617,
      "step": 10300
    },
    {
      "epoch": 0.7221720713839317,
      "grad_norm": 0.028564453125,
      "learning_rate": 0.0004084277875864776,
      "loss": 0.0612,
      "step": 10400
    },
    {
      "epoch": 0.7291160336087772,
      "grad_norm": 0.035400390625,
      "learning_rate": 0.00038968973815020803,
      "loss": 0.0615,
      "step": 10500
    },
    {
      "epoch": 0.7360599958336227,
      "grad_norm": 0.0322265625,
      "learning_rate": 0.00037128713340911534,
      "loss": 0.0607,
      "step": 10600
    },
    {
      "epoch": 0.7430039580584682,
      "grad_norm": 0.050537109375,
      "learning_rate": 0.00035323008798280133,
      "loss": 0.0606,
      "step": 10700
    },
    {
      "epoch": 0.7499479202833137,
      "grad_norm": 0.033203125,
      "learning_rate": 0.00033552852656117837,
      "loss": 0.0606,
      "step": 10800
    },
    {
      "epoch": 0.7568918825081592,
      "grad_norm": 0.0308837890625,
      "learning_rate": 0.00031819217844956217,
      "loss": 0.0599,
      "step": 10900
    },
    {
      "epoch": 0.7638358447330047,
      "grad_norm": 0.0283203125,
      "learning_rate": 0.00030123057222115836,
      "loss": 0.061,
      "step": 11000
    },
    {
      "epoch": 0.7638358447330047,
      "eval_covost2-en-de_loss": 1.3495559692382812,
      "eval_covost2-en-de_runtime": 32.3219,
      "eval_covost2-en-de_samples_per_second": 1.98,
      "eval_covost2-en-de_steps_per_second": 0.062,
      "step": 11000
    },
    {
      "epoch": 0.7638358447330047,
      "eval_covost2-zh-en_loss": 2.6154580116271973,
      "eval_covost2-zh-en_runtime": 31.7065,
      "eval_covost2-zh-en_samples_per_second": 2.019,
      "eval_covost2-zh-en_steps_per_second": 0.063,
      "step": 11000
    },
    {
      "epoch": 0.7638358447330047,
      "eval_peoplespeech-clean-transcription_loss": 1.673677682876587,
      "eval_peoplespeech-clean-transcription_runtime": 32.5731,
      "eval_peoplespeech-clean-transcription_samples_per_second": 1.965,
      "eval_peoplespeech-clean-transcription_steps_per_second": 0.061,
      "step": 11000
    },
    {
      "epoch": 0.7707798069578502,
      "grad_norm": 0.029541015625,
      "learning_rate": 0.0002846530304798727,
      "loss": 0.0602,
      "step": 11100
    },
    {
      "epoch": 0.7777237691826957,
      "grad_norm": 0.0294189453125,
      "learning_rate": 0.00026846866473633125,
      "loss": 0.0608,
      "step": 11200
    },
    {
      "epoch": 0.7846677314075411,
      "grad_norm": 0.0264892578125,
      "learning_rate": 0.00025268637039992293,
      "loss": 0.0611,
      "step": 11300
    },
    {
      "epoch": 0.7916116936323866,
      "grad_norm": 0.024658203125,
      "learning_rate": 0.00023731482188961818,
      "loss": 0.0602,
      "step": 11400
    },
    {
      "epoch": 0.7985556558572321,
      "grad_norm": 0.0294189453125,
      "learning_rate": 0.00022236246786624792,
      "loss": 0.0608,
      "step": 11500
    },
    {
      "epoch": 0.8054996180820776,
      "grad_norm": 0.0322265625,
      "learning_rate": 0.00020783752658887068,
      "loss": 0.0602,
      "step": 11600
    },
    {
      "epoch": 0.8124435803069231,
      "grad_norm": 0.035888671875,
      "learning_rate": 0.0001937479813977703,
      "loss": 0.0607,
      "step": 11700
    },
    {
      "epoch": 0.8193875425317686,
      "grad_norm": 0.0262451171875,
      "learning_rate": 0.00018010157632657541,
      "loss": 0.0611,
      "step": 11800
    },
    {
      "epoch": 0.8263315047566141,
      "grad_norm": 0.0296630859375,
      "learning_rate": 0.00016690581184590858,
      "loss": 0.0617,
      "step": 11900
    },
    {
      "epoch": 0.8332754669814596,
      "grad_norm": 0.03564453125,
      "learning_rate": 0.00015416794074090258,
      "loss": 0.0615,
      "step": 12000
    },
    {
      "epoch": 0.8332754669814596,
      "eval_covost2-en-de_loss": 1.3493196964263916,
      "eval_covost2-en-de_runtime": 32.732,
      "eval_covost2-en-de_samples_per_second": 1.955,
      "eval_covost2-en-de_steps_per_second": 0.061,
      "step": 12000
    },
    {
      "epoch": 0.8332754669814596,
      "eval_covost2-zh-en_loss": 2.615288734436035,
      "eval_covost2-zh-en_runtime": 31.3635,
      "eval_covost2-zh-en_samples_per_second": 2.041,
      "eval_covost2-zh-en_steps_per_second": 0.064,
      "step": 12000
    },
    {
      "epoch": 0.8332754669814596,
      "eval_peoplespeech-clean-transcription_loss": 1.6742818355560303,
      "eval_peoplespeech-clean-transcription_runtime": 32.2391,
      "eval_peoplespeech-clean-transcription_samples_per_second": 1.985,
      "eval_peoplespeech-clean-transcription_steps_per_second": 0.062,
      "step": 12000
    },
    {
      "epoch": 0.8402194292063051,
      "grad_norm": 0.033203125,
      "learning_rate": 0.00014189496412485593,
      "loss": 0.061,
      "step": 12100
    },
    {
      "epoch": 0.8471633914311506,
      "grad_norm": 0.0322265625,
      "learning_rate": 0.00013009362759120978,
      "loss": 0.0601,
      "step": 12200
    },
    {
      "epoch": 0.8541073536559961,
      "grad_norm": 0.028076171875,
      "learning_rate": 0.00011877041750597173,
      "loss": 0.0613,
      "step": 12300
    },
    {
      "epoch": 0.8610513158808416,
      "grad_norm": 0.0257568359375,
      "learning_rate": 0.00010793155744261352,
      "loss": 0.0614,
      "step": 12400
    },
    {
      "epoch": 0.8679952781056871,
      "grad_norm": 0.02783203125,
      "learning_rate": 9.758300476141169e-05,
      "loss": 0.0606,
      "step": 12500
    },
    {
      "epoch": 0.8749392403305326,
      "grad_norm": 0.0302734375,
      "learning_rate": 8.773044733510338e-05,
      "loss": 0.0604,
      "step": 12600
    },
    {
      "epoch": 0.8818832025553781,
      "grad_norm": 0.02880859375,
      "learning_rate": 7.837930042266262e-05,
      "loss": 0.0601,
      "step": 12700
    },
    {
      "epoch": 0.8888271647802236,
      "grad_norm": 0.02880859375,
      "learning_rate": 6.953470369291348e-05,
      "loss": 0.0607,
      "step": 12800
    },
    {
      "epoch": 0.8957711270050691,
      "grad_norm": 0.027587890625,
      "learning_rate": 6.120151839961363e-05,
      "loss": 0.0606,
      "step": 12900
    },
    {
      "epoch": 0.9027150892299146,
      "grad_norm": 0.036865234375,
      "learning_rate": 5.338432470956589e-05,
      "loss": 0.0608,
      "step": 13000
    },
    {
      "epoch": 0.9027150892299146,
      "eval_covost2-en-de_loss": 1.3490619659423828,
      "eval_covost2-en-de_runtime": 33.0003,
      "eval_covost2-en-de_samples_per_second": 1.939,
      "eval_covost2-en-de_steps_per_second": 0.061,
      "step": 13000
    },
    {
      "epoch": 0.9027150892299146,
      "eval_covost2-zh-en_loss": 2.615211248397827,
      "eval_covost2-zh-en_runtime": 31.7808,
      "eval_covost2-zh-en_samples_per_second": 2.014,
      "eval_covost2-zh-en_steps_per_second": 0.063,
      "step": 13000
    },
    {
      "epoch": 0.9027150892299146,
      "eval_peoplespeech-clean-transcription_loss": 1.6742777824401855,
      "eval_peoplespeech-clean-transcription_runtime": 32.3145,
      "eval_peoplespeech-clean-transcription_samples_per_second": 1.981,
      "eval_peoplespeech-clean-transcription_steps_per_second": 0.062,
      "step": 13000
    },
    {
      "epoch": 0.9096590514547601,
      "grad_norm": 0.0302734375,
      "learning_rate": 4.6087419185220966e-05,
      "loss": 0.06,
      "step": 13100
    },
    {
      "epoch": 0.9166030136796056,
      "grad_norm": 0.028076171875,
      "learning_rate": 3.931481242315993e-05,
      "loss": 0.0601,
      "step": 13200
    },
    {
      "epoch": 0.9235469759044511,
      "grad_norm": 0.03759765625,
      "learning_rate": 3.307022684974936e-05,
      "loss": 0.0602,
      "step": 13300
    },
    {
      "epoch": 0.9304909381292966,
      "grad_norm": 0.02978515625,
      "learning_rate": 2.7357094675186987e-05,
      "loss": 0.0605,
      "step": 13400
    },
    {
      "epoch": 0.9374349003541421,
      "grad_norm": 0.031005859375,
      "learning_rate": 2.2178556007054874e-05,
      "loss": 0.061,
      "step": 13500
    },
    {
      "epoch": 0.9443788625789876,
      "grad_norm": 0.0289306640625,
      "learning_rate": 1.7537457124423894e-05,
      "loss": 0.0606,
      "step": 13600
    },
    {
      "epoch": 0.9513228248038331,
      "grad_norm": 0.028076171875,
      "learning_rate": 1.3436348913453578e-05,
      "loss": 0.061,
      "step": 13700
    },
    {
      "epoch": 0.9582667870286785,
      "grad_norm": 0.0289306640625,
      "learning_rate": 9.877485465349056e-06,
      "loss": 0.0606,
      "step": 13800
    },
    {
      "epoch": 0.965210749253524,
      "grad_norm": 0.0291748046875,
      "learning_rate": 6.862822837445881e-06,
      "loss": 0.0613,
      "step": 13900
    },
    {
      "epoch": 0.9721547114783695,
      "grad_norm": 0.0302734375,
      "learning_rate": 4.394017978101905e-06,
      "loss": 0.0615,
      "step": 14000
    },
    {
      "epoch": 0.9721547114783695,
      "eval_covost2-en-de_loss": 1.3485227823257446,
      "eval_covost2-en-de_runtime": 32.3946,
      "eval_covost2-en-de_samples_per_second": 1.976,
      "eval_covost2-en-de_steps_per_second": 0.062,
      "step": 14000
    },
    {
      "epoch": 0.9721547114783695,
      "eval_covost2-zh-en_loss": 2.614840269088745,
      "eval_covost2-zh-en_runtime": 31.6743,
      "eval_covost2-zh-en_samples_per_second": 2.021,
      "eval_covost2-zh-en_steps_per_second": 0.063,
      "step": 14000
    },
    {
      "epoch": 0.9721547114783695,
      "eval_peoplespeech-clean-transcription_loss": 1.673971176147461,
      "eval_peoplespeech-clean-transcription_runtime": 32.2005,
      "eval_peoplespeech-clean-transcription_samples_per_second": 1.988,
      "eval_peoplespeech-clean-transcription_steps_per_second": 0.062,
      "step": 14000
    },
    {
      "epoch": 0.979098673703215,
      "grad_norm": 0.027587890625,
      "learning_rate": 2.472427815989886e-06,
      "loss": 0.0607,
      "step": 14100
    },
    {
      "epoch": 0.9860426359280605,
      "grad_norm": 0.031982421875,
      "learning_rate": 1.099108514288627e-06,
      "loss": 0.0608,
      "step": 14200
    },
    {
      "epoch": 0.992986598152906,
      "grad_norm": 0.03466796875,
      "learning_rate": 2.748148901841052e-07,
      "loss": 0.0606,
      "step": 14300
    },
    {
      "epoch": 0.9999305603777515,
      "grad_norm": 0.0277099609375,
      "learning_rate": 0.0,
      "loss": 0.0608,
      "step": 14400
    }
  ],
  "logging_steps": 100,
  "max_steps": 14400,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 1,
  "save_steps": 3600,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": true
      },
      "attributes": {}
    }
  },
  "total_flos": 3.7890136187397734e+18,
  "train_batch_size": 96,
  "trial_name": null,
  "trial_params": null
}