SoccerTwos1-PPO / run_logs /timers.json
haidlir's picture
Initial Commit
a6f0ed7
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": 3.18869686126709,
"min": 3.1670570373535156,
"max": 3.2957441806793213,
"count": 117
},
"SoccerTwos.Policy.Entropy.sum": {
"value": 38570.4765625,
"min": 12105.599609375,
"max": 152054.90625,
"count": 117
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 695.3333333333334,
"min": 433.4166666666667,
"max": 999.0,
"count": 117
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 16688.0,
"min": 4452.0,
"max": 28136.0,
"count": 117
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 1193.2650606209672,
"min": 1190.7322727579492,
"max": 1201.5572309876034,
"count": 108
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 9546.120484967738,
"min": 2381.635987580858,
"max": 16755.748691303368,
"count": 108
},
"SoccerTwos.Step.mean": {
"value": 1169334.0,
"min": 9084.0,
"max": 1169334.0,
"count": 117
},
"SoccerTwos.Step.sum": {
"value": 1169334.0,
"min": 9084.0,
"max": 1169334.0,
"count": 117
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": 0.027071211487054825,
"min": -0.06372575461864471,
"max": 0.027482615783810616,
"count": 117
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": 0.3519257605075836,
"min": -0.8284347653388977,
"max": 0.3535115420818329,
"count": 117
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": 0.026384852826595306,
"min": -0.06369052082300186,
"max": 0.028071589767932892,
"count": 117
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": 0.3430030941963196,
"min": -0.827970564365387,
"max": 0.364930659532547,
"count": 117
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 117
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 117
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": -0.14698461844370916,
"min": -0.5714285714285714,
"max": 0.27497142979076933,
"count": 117
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": -1.910800039768219,
"min": -8.0,
"max": 3.8496000170707703,
"count": 117
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": -0.14698461844370916,
"min": -0.5714285714285714,
"max": 0.27497142979076933,
"count": 117
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": -1.910800039768219,
"min": -8.0,
"max": 3.8496000170707703,
"count": 117
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 117
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 117
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.015360206284094602,
"min": 0.012117091874824838,
"max": 0.024097679225572696,
"count": 54
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.015360206284094602,
"min": 0.012117091874824838,
"max": 0.024097679225572696,
"count": 54
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 0.0034538612274142605,
"min": 0.00034969555053976364,
"max": 0.005978238047100603,
"count": 54
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 0.0034538612274142605,
"min": 0.00034969555053976364,
"max": 0.005978238047100603,
"count": 54
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 0.003469962584010015,
"min": 0.0003521892847250759,
"max": 0.006564258489136895,
"count": 54
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 0.003469962584010015,
"min": 0.0003521892847250759,
"max": 0.006564258489136895,
"count": 54
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 54
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 54
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.20000000000000007,
"min": 0.2,
"max": 0.20000000000000007,
"count": 54
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.20000000000000007,
"min": 0.2,
"max": 0.20000000000000007,
"count": 54
},
"SoccerTwos.Policy.Beta.mean": {
"value": 0.005000000000000001,
"min": 0.005,
"max": 0.005000000000000001,
"count": 54
},
"SoccerTwos.Policy.Beta.sum": {
"value": 0.005000000000000001,
"min": 0.005,
"max": 0.005000000000000001,
"count": 54
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1679736271",
"python_version": "3.9.16 (main, Dec 7 2022, 01:11:51) \n[GCC 9.4.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/poca/SoccerTwos.yaml --env=./training-envs-executables/linux/SoccerTwos.x86_64 --run-id=SoccerTwos --no-graphics",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.11.0+cu102",
"numpy_version": "1.21.2",
"end_time_seconds": "1679738620"
},
"total": 2348.638205398,
"count": 1,
"self": 0.04307650400005514,
"children": {
"run_training.setup": {
"total": 0.11162882000002128,
"count": 1,
"self": 0.11162882000002128
},
"TrainerController.start_learning": {
"total": 2348.483500074,
"count": 1,
"self": 1.5388987449687193,
"children": {
"TrainerController._reset_env": {
"total": 9.502020643000037,
"count": 6,
"self": 9.502020643000037
},
"TrainerController.advance": {
"total": 2337.0845989350314,
"count": 76629,
"self": 1.8646640759989168,
"children": {
"env_step": {
"total": 1861.4849851179915,
"count": 76629,
"self": 1458.597810473985,
"children": {
"SubprocessEnvManager._take_step": {
"total": 401.9372201870403,
"count": 76629,
"self": 12.484379305042978,
"children": {
"TorchPolicy.evaluate": {
"total": 389.4528408819973,
"count": 152102,
"self": 389.4528408819973
}
}
},
"workers": {
"total": 0.9499544569662248,
"count": 76628,
"self": 0.0,
"children": {
"worker_root": {
"total": 2343.819020711023,
"count": 76628,
"is_parallel": true,
"self": 1112.460945891066,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0057090290000019195,
"count": 2,
"is_parallel": true,
"self": 0.0035821259999124777,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0021269030000894418,
"count": 8,
"is_parallel": true,
"self": 0.0021269030000894418
}
}
},
"UnityEnvironment.step": {
"total": 0.07212128599996959,
"count": 1,
"is_parallel": true,
"self": 0.0010296769999627031,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.0008685379999633369,
"count": 1,
"is_parallel": true,
"self": 0.0008685379999633369
},
"communicator.exchange": {
"total": 0.06713946100001067,
"count": 1,
"is_parallel": true,
"self": 0.06713946100001067
},
"steps_from_proto": {
"total": 0.003083610000032877,
"count": 2,
"is_parallel": true,
"self": 0.0005916340001022036,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0024919759999306734,
"count": 8,
"is_parallel": true,
"self": 0.0024919759999306734
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1231.3458499279573,
"count": 76627,
"is_parallel": true,
"self": 73.28550675604083,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 51.22505309598006,
"count": 76627,
"is_parallel": true,
"self": 51.22505309598006
},
"communicator.exchange": {
"total": 881.2321504529953,
"count": 76627,
"is_parallel": true,
"self": 881.2321504529953
},
"steps_from_proto": {
"total": 225.60313962294117,
"count": 153254,
"is_parallel": true,
"self": 41.6531555650144,
"children": {
"_process_rank_one_or_two_observation": {
"total": 183.94998405792677,
"count": 613016,
"is_parallel": true,
"self": 183.94998405792677
}
}
}
}
},
"steps_from_proto": {
"total": 0.012224891999494503,
"count": 10,
"is_parallel": true,
"self": 0.0024052009999877555,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.009819690999506747,
"count": 40,
"is_parallel": true,
"self": 0.009819690999506747
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 473.73494974104085,
"count": 76628,
"self": 15.233939597034862,
"children": {
"process_trajectory": {
"total": 120.74160695800657,
"count": 76628,
"self": 120.02034359700639,
"children": {
"RLTrainer._checkpoint": {
"total": 0.7212633610001831,
"count": 2,
"self": 0.7212633610001831
}
}
},
"_update_policy": {
"total": 337.7594031859994,
"count": 54,
"self": 210.69964387899927,
"children": {
"TorchPOCAOptimizer.update": {
"total": 127.05975930700015,
"count": 1632,
"self": 127.05975930700015
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.3730000318901148e-06,
"count": 1,
"self": 1.3730000318901148e-06
},
"TrainerController._save_models": {
"total": 0.3579803779998656,
"count": 1,
"self": 0.0023122190000322007,
"children": {
"RLTrainer._checkpoint": {
"total": 0.3556681589998334,
"count": 1,
"self": 0.3556681589998334
}
}
}
}
}
}
}