poca-SoccerTwos / run_logs /timers.json
mashaal24's picture
First Push
83d26b6 verified
raw
history blame
15.8 kB
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": 3.1975491046905518,
"min": 3.1636340618133545,
"max": 3.29569411277771,
"count": 83
},
"SoccerTwos.Policy.Entropy.sum": {
"value": 60165.0859375,
"min": 22820.60546875,
"max": 119603.46875,
"count": 83
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 999.0,
"min": 446.8333333333333,
"max": 999.0,
"count": 83
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 19980.0,
"min": 13420.0,
"max": 28944.0,
"count": 83
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 1187.1597866913448,
"min": 1182.1128475910111,
"max": 1198.6368125508247,
"count": 58
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 2374.3195733826897,
"min": 2364.2256951820223,
"max": 23894.308557487457,
"count": 58
},
"SoccerTwos.Step.mean": {
"value": 829818.0,
"min": 9492.0,
"max": 829818.0,
"count": 83
},
"SoccerTwos.Step.sum": {
"value": 829818.0,
"min": 9492.0,
"max": 829818.0,
"count": 83
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": 0.009815307334065437,
"min": -0.016077551990747452,
"max": 0.009815307334065437,
"count": 83
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": 0.09815307706594467,
"min": -0.21648471057415009,
"max": 0.1015714779496193,
"count": 83
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": 0.010055381804704666,
"min": -0.014644809998571873,
"max": 0.010055381804704666,
"count": 83
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": 0.10055381804704666,
"min": -0.23140473663806915,
"max": 0.10448700189590454,
"count": 83
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 83
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 83
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": 0.0,
"min": -0.4666666666666667,
"max": 0.20775384169358474,
"count": 83
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": 0.0,
"min": -7.0,
"max": 2.7007999420166016,
"count": 83
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": 0.0,
"min": -0.4666666666666667,
"max": 0.20775384169358474,
"count": 83
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": 0.0,
"min": -7.0,
"max": 2.7007999420166016,
"count": 83
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 83
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 83
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.01743063961524361,
"min": 0.011421181401237846,
"max": 0.021392928244313224,
"count": 38
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.01743063961524361,
"min": 0.011421181401237846,
"max": 0.021392928244313224,
"count": 38
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 0.001036214370590945,
"min": 4.1560988014074e-06,
"max": 0.008760098864634832,
"count": 38
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 0.001036214370590945,
"min": 4.1560988014074e-06,
"max": 0.008760098864634832,
"count": 38
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 0.0010391806029171372,
"min": 5.079699227887128e-06,
"max": 0.008930069782460729,
"count": 38
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 0.0010391806029171372,
"min": 5.079699227887128e-06,
"max": 0.008930069782460729,
"count": 38
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 38
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 38
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.20000000000000007,
"min": 0.2,
"max": 0.20000000000000007,
"count": 38
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.20000000000000007,
"min": 0.2,
"max": 0.20000000000000007,
"count": 38
},
"SoccerTwos.Policy.Beta.mean": {
"value": 0.005000000000000001,
"min": 0.005,
"max": 0.005000000000000001,
"count": 38
},
"SoccerTwos.Policy.Beta.sum": {
"value": 0.005000000000000001,
"min": 0.005,
"max": 0.005000000000000001,
"count": 38
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1723807211",
"python_version": "3.10.12 | packaged by Anaconda, Inc. | (main, Jul 5 2023, 19:01:18) [MSC v.1916 64 bit (AMD64)]",
"command_line_arguments": "C:\\Users\\EED\\miniconda3\\envs\\mlagents\\Scripts\\mlagents-learn ./config/poca/SoccerTwos.yaml --env=./training-envs-executables/SoccerTwos.exe --run-id=SoccerTwos --no-graphics",
"mlagents_version": "1.0.0",
"mlagents_envs_version": "1.0.0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "2.2.0+cpu",
"numpy_version": "1.21.4",
"end_time_seconds": "1723809331"
},
"total": 2120.3000029999994,
"count": 1,
"self": 2.1215368000002854,
"children": {
"run_training.setup": {
"total": 0.1225466999994751,
"count": 1,
"self": 0.1225466999994751
},
"TrainerController.start_learning": {
"total": 2118.0559194999996,
"count": 1,
"self": 1.1239860998603035,
"children": {
"TrainerController._reset_env": {
"total": 25.25254450000284,
"count": 5,
"self": 25.25254450000284
},
"TrainerController.advance": {
"total": 2091.5420100001375,
"count": 54565,
"self": 1.1384467999314438,
"children": {
"env_step": {
"total": 828.285834400076,
"count": 54565,
"self": 594.564090900345,
"children": {
"SubprocessEnvManager._take_step": {
"total": 232.95738159994926,
"count": 54565,
"self": 7.249221400087663,
"children": {
"TorchPolicy.evaluate": {
"total": 225.7081601998616,
"count": 108378,
"self": 225.7081601998616
}
}
},
"workers": {
"total": 0.7643618997817612,
"count": 54565,
"self": 0.0,
"children": {
"worker_root": {
"total": 2087.7030302000867,
"count": 54565,
"is_parallel": true,
"self": 1645.9276910002318,
"children": {
"steps_from_proto": {
"total": 0.005954700000984303,
"count": 10,
"is_parallel": true,
"self": 0.0012953000023117056,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.004659399998672598,
"count": 40,
"is_parallel": true,
"self": 0.004659399998672598
}
}
},
"UnityEnvironment.step": {
"total": 441.7693844998539,
"count": 54565,
"is_parallel": true,
"self": 24.482705599693873,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 23.476010500014127,
"count": 54565,
"is_parallel": true,
"self": 23.476010500014127
},
"communicator.exchange": {
"total": 305.3720151002053,
"count": 54565,
"is_parallel": true,
"self": 305.3720151002053
},
"steps_from_proto": {
"total": 88.4386532999406,
"count": 109130,
"is_parallel": true,
"self": 19.25300659995355,
"children": {
"_process_rank_one_or_two_observation": {
"total": 69.18564669998705,
"count": 436520,
"is_parallel": true,
"self": 69.18564669998705
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 1262.11772880013,
"count": 54565,
"self": 8.93584180011294,
"children": {
"process_trajectory": {
"total": 182.80165380001654,
"count": 54565,
"self": 182.59389220001594,
"children": {
"RLTrainer._checkpoint": {
"total": 0.20776160000059463,
"count": 1,
"self": 0.20776160000059463
}
}
},
"_update_policy": {
"total": 1070.3802332000005,
"count": 39,
"self": 155.56718379997437,
"children": {
"TorchPOCAOptimizer.update": {
"total": 914.8130494000261,
"count": 1156,
"self": 914.8130494000261
}
}
}
}
}
}
},
"trainer_threads": {
"total": 8.999995770864189e-07,
"count": 1,
"self": 8.999995770864189e-07
},
"TrainerController._save_models": {
"total": 0.137377999999444,
"count": 1,
"self": 0.010175799998251023,
"children": {
"RLTrainer._checkpoint": {
"total": 0.12720220000119298,
"count": 1,
"self": 0.12720220000119298
}
}
}
}
}
}
}