poca-SoccerTwos / run_logs /timers.json
1daniar's picture
First Push
8dad927
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": 3.1866912841796875,
"min": 3.1682472229003906,
"max": 3.2957022190093994,
"count": 98
},
"SoccerTwos.Policy.Entropy.sum": {
"value": 90451.046875,
"min": 8753.3466796875,
"max": 105462.109375,
"count": 98
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 618.2222222222222,
"min": 82.0,
"max": 999.0,
"count": 98
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 22256.0,
"min": 328.0,
"max": 31968.0,
"count": 98
},
"SoccerTwos.Step.mean": {
"value": 999400.0,
"min": 29854.0,
"max": 999400.0,
"count": 98
},
"SoccerTwos.Step.sum": {
"value": 999400.0,
"min": 29854.0,
"max": 999400.0,
"count": 98
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": 0.028475239872932434,
"min": -0.06551233679056168,
"max": 0.028475239872932434,
"count": 98
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": 0.4840790927410126,
"min": -0.9170898795127869,
"max": 0.4840790927410126,
"count": 98
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": 0.02584037370979786,
"min": -0.06547804176807404,
"max": 0.02584037370979786,
"count": 98
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": 0.43928635120391846,
"min": -0.9166770577430725,
"max": 0.43928635120391846,
"count": 98
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 98
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 98
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": -0.32461176199071545,
"min": -0.5,
"max": 0.3692380970432645,
"count": 98
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": -5.518399953842163,
"min": -6.438800048083067,
"max": 7.754000037908554,
"count": 98
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": -0.32461176199071545,
"min": -0.5,
"max": 0.3692380970432645,
"count": 98
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": -5.518399953842163,
"min": -6.438800048083067,
"max": 7.754000037908554,
"count": 98
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 98
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 98
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 1208.4210166434545,
"min": 1198.0257057172744,
"max": 1209.014185855263,
"count": 78
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 14501.052199721455,
"min": 2396.0514114345488,
"max": 24055.55628169224,
"count": 78
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.020749160430083673,
"min": 0.01087481270369608,
"max": 0.022622195289780697,
"count": 45
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.020749160430083673,
"min": 0.01087481270369608,
"max": 0.022622195289780697,
"count": 45
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 0.0055723156857614715,
"min": 3.057377451417172e-06,
"max": 0.0055723156857614715,
"count": 45
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 0.0055723156857614715,
"min": 3.057377451417172e-06,
"max": 0.0055723156857614715,
"count": 45
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 0.005925432623674472,
"min": 3.521252354706424e-06,
"max": 0.005925432623674472,
"count": 45
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 0.005925432623674472,
"min": 3.521252354706424e-06,
"max": 0.005925432623674472,
"count": 45
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 45
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 45
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.20000000000000007,
"min": 0.19999999999999996,
"max": 0.20000000000000007,
"count": 45
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.20000000000000007,
"min": 0.19999999999999996,
"max": 0.20000000000000007,
"count": 45
},
"SoccerTwos.Policy.Beta.mean": {
"value": 0.005000000000000001,
"min": 0.005,
"max": 0.005000000000000001,
"count": 45
},
"SoccerTwos.Policy.Beta.sum": {
"value": 0.005000000000000001,
"min": 0.005,
"max": 0.005000000000000001,
"count": 45
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1690601381",
"python_version": "3.10.6 (main, May 29 2023, 11:10:38) [GCC 11.3.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/poca/SoccerTwos.yaml --env=./training-envs-executables/linux/SoccerTwos --run-id=SoccerTwos --no-graphics --resume",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.11.0+cu102",
"numpy_version": "1.21.2",
"end_time_seconds": "1690603658"
},
"total": 2276.536627766,
"count": 1,
"self": 0.425417648000348,
"children": {
"run_training.setup": {
"total": 0.05573638100008793,
"count": 1,
"self": 0.05573638100008793
},
"TrainerController.start_learning": {
"total": 2276.055473737,
"count": 1,
"self": 1.6762015140311632,
"children": {
"TrainerController._reset_env": {
"total": 5.181112253000265,
"count": 5,
"self": 5.181112253000265
},
"TrainerController.advance": {
"total": 2268.929314434968,
"count": 63379,
"self": 1.8176577400058704,
"children": {
"env_step": {
"total": 1835.2405524999203,
"count": 63379,
"self": 1452.6599075109257,
"children": {
"SubprocessEnvManager._take_step": {
"total": 381.57838352298506,
"count": 63379,
"self": 12.348988554066182,
"children": {
"TorchPolicy.evaluate": {
"total": 369.2293949689189,
"count": 125852,
"self": 369.2293949689189
}
}
},
"workers": {
"total": 1.0022614660095996,
"count": 63379,
"self": 0.0,
"children": {
"worker_root": {
"total": 2270.969339070059,
"count": 63379,
"is_parallel": true,
"self": 1039.3576673331302,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.005149682000137545,
"count": 2,
"is_parallel": true,
"self": 0.0011812910000799093,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.003968391000057636,
"count": 8,
"is_parallel": true,
"self": 0.003968391000057636
}
}
},
"UnityEnvironment.step": {
"total": 0.04324213899997176,
"count": 1,
"is_parallel": true,
"self": 0.0013280710001026819,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.0008874509999259317,
"count": 1,
"is_parallel": true,
"self": 0.0008874509999259317
},
"communicator.exchange": {
"total": 0.03714546599996993,
"count": 1,
"is_parallel": true,
"self": 0.03714546599996993
},
"steps_from_proto": {
"total": 0.0038811509999732152,
"count": 2,
"is_parallel": true,
"self": 0.0007578040001590125,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0031233469998142027,
"count": 8,
"is_parallel": true,
"self": 0.0031233469998142027
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1231.6013744379288,
"count": 63378,
"is_parallel": true,
"self": 75.7722481580563,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 50.12961376796443,
"count": 63378,
"is_parallel": true,
"self": 50.12961376796443
},
"communicator.exchange": {
"total": 853.2457646569906,
"count": 63378,
"is_parallel": true,
"self": 853.2457646569906
},
"steps_from_proto": {
"total": 252.45374785491765,
"count": 126756,
"is_parallel": true,
"self": 41.75554271190401,
"children": {
"_process_rank_one_or_two_observation": {
"total": 210.69820514301364,
"count": 507024,
"is_parallel": true,
"self": 210.69820514301364
}
}
}
}
},
"steps_from_proto": {
"total": 0.010297299000058047,
"count": 8,
"is_parallel": true,
"self": 0.0019517260002430703,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.008345572999814976,
"count": 32,
"is_parallel": true,
"self": 0.008345572999814976
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 431.871104195042,
"count": 63379,
"self": 14.313646950020711,
"children": {
"process_trajectory": {
"total": 106.42276668302168,
"count": 63379,
"self": 105.71568784202157,
"children": {
"RLTrainer._checkpoint": {
"total": 0.7070788410001114,
"count": 2,
"self": 0.7070788410001114
}
}
},
"_update_policy": {
"total": 311.1346905619996,
"count": 45,
"self": 202.38677821800036,
"children": {
"TorchPOCAOptimizer.update": {
"total": 108.74791234399925,
"count": 1362,
"self": 108.74791234399925
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.374000021314714e-06,
"count": 1,
"self": 1.374000021314714e-06
},
"TrainerController._save_models": {
"total": 0.2688441610002883,
"count": 1,
"self": 0.002427545000500686,
"children": {
"RLTrainer._checkpoint": {
"total": 0.2664166159997876,
"count": 1,
"self": 0.2664166159997876
}
}
}
}
}
}
}