SoccerTwos / run_logs /timers.json
yy1636's picture
First Push
b8fe4bc verified
raw
history blame
20.2 kB
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": 1.5221166610717773,
"min": 1.384403109550476,
"max": 3.295748472213745,
"count": 2812
},
"SoccerTwos.Policy.Entropy.sum": {
"value": 31270.365234375,
"min": 14488.515625,
"max": 115618.765625,
"count": 2812
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 50.52040816326531,
"min": 38.458333333333336,
"max": 999.0,
"count": 2812
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 19804.0,
"min": 16108.0,
"max": 23496.0,
"count": 2812
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 1574.8378275970615,
"min": 1189.7606299623117,
"max": 1683.903107630605,
"count": 2786
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 308668.21420902404,
"min": 2384.133845548939,
"max": 422285.02933067916,
"count": 2786
},
"SoccerTwos.Step.mean": {
"value": 28119988.0,
"min": 9596.0,
"max": 28119988.0,
"count": 2812
},
"SoccerTwos.Step.sum": {
"value": 28119988.0,
"min": 9596.0,
"max": 28119988.0,
"count": 2812
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": -0.04123266413807869,
"min": -0.14193083345890045,
"max": 0.1889309287071228,
"count": 2812
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": -8.081602096557617,
"min": -28.670028686523438,
"max": 30.90949249267578,
"count": 2812
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": -0.03751363232731819,
"min": -0.13879628479480743,
"max": 0.1827581524848938,
"count": 2812
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": -7.3526716232299805,
"min": -28.036849975585938,
"max": 31.078296661376953,
"count": 2812
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 2812
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 2812
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": -0.0670489784406156,
"min": -0.75,
"max": 0.4416342888559614,
"count": 2812
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": -13.141599774360657,
"min": -60.755200266838074,
"max": 56.169999837875366,
"count": 2812
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": -0.0670489784406156,
"min": -0.75,
"max": 0.4416342888559614,
"count": 2812
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": -13.141599774360657,
"min": -60.755200266838074,
"max": 56.169999837875366,
"count": 2812
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 2812
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 2812
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.020729929541509286,
"min": 0.009515829482673628,
"max": 0.02383031997208794,
"count": 1361
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.020729929541509286,
"min": 0.009515829482673628,
"max": 0.02383031997208794,
"count": 1361
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 0.10369236866633097,
"min": 1.4918835010272839e-05,
"max": 0.13401054963469505,
"count": 1361
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 0.10369236866633097,
"min": 1.4918835010272839e-05,
"max": 0.13401054963469505,
"count": 1361
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 0.10502697055538496,
"min": 1.6306885057322992e-05,
"max": 0.13664711986978847,
"count": 1361
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 0.10502697055538496,
"min": 1.6306885057322992e-05,
"max": 0.13664711986978847,
"count": 1361
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 1361
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 1361
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.20000000000000007,
"min": 0.20000000000000007,
"max": 0.20000000000000007,
"count": 1361
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.20000000000000007,
"min": 0.20000000000000007,
"max": 0.20000000000000007,
"count": 1361
},
"SoccerTwos.Policy.Beta.mean": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 1361
},
"SoccerTwos.Policy.Beta.sum": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 1361
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1713676045",
"python_version": "3.10.12 (main, Nov 20 2023, 15:14:05) [GCC 11.4.0]",
"command_line_arguments": "/home/robot/.local/bin/mlagents-learn ./config/poca/SoccerTwos.yaml --env=./training-envs-executables/SoccerTwos/SoccerTwos.x86_64 --run-id=SoccerTwos-test-0 --no-graphics",
"mlagents_version": "1.1.0.dev0",
"mlagents_envs_version": "1.1.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "2.2.2+cu121",
"numpy_version": "1.23.5",
"end_time_seconds": "1713703779"
},
"total": 27733.902080582004,
"count": 1,
"self": 0.06665946600332973,
"children": {
"run_training.setup": {
"total": 0.025941288000467466,
"count": 1,
"self": 0.025941288000467466
},
"TrainerController.start_learning": {
"total": 27733.809479828,
"count": 1,
"self": 25.197649379144423,
"children": {
"TrainerController._reset_env": {
"total": 3.2325739870047983,
"count": 141,
"self": 3.2325739870047983
},
"TrainerController.advance": {
"total": 27705.27857127085,
"count": 1945342,
"self": 27.427601072387915,
"children": {
"env_step": {
"total": 21004.1931005967,
"count": 1945342,
"self": 15549.292188255045,
"children": {
"SubprocessEnvManager._take_step": {
"total": 5438.777019470792,
"count": 1945342,
"self": 166.68569100583954,
"children": {
"TorchPolicy.evaluate": {
"total": 5272.091328464952,
"count": 3536941,
"self": 5272.091328464952
}
}
},
"workers": {
"total": 16.123892870864438,
"count": 1945341,
"self": 0.0,
"children": {
"worker_root": {
"total": 27692.80342434056,
"count": 1945341,
"is_parallel": true,
"self": 15027.505136596712,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0014246609989641001,
"count": 2,
"is_parallel": true,
"self": 0.0003357030000188388,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0010889579989452614,
"count": 8,
"is_parallel": true,
"self": 0.0010889579989452614
}
}
},
"UnityEnvironment.step": {
"total": 0.015922061998935533,
"count": 1,
"is_parallel": true,
"self": 0.00035585199839260895,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.00027113100077258423,
"count": 1,
"is_parallel": true,
"self": 0.00027113100077258423
},
"communicator.exchange": {
"total": 0.014303907000794425,
"count": 1,
"is_parallel": true,
"self": 0.014303907000794425
},
"steps_from_proto": {
"total": 0.000991171998975915,
"count": 2,
"is_parallel": true,
"self": 0.0001968040014617145,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0007943679975142004,
"count": 8,
"is_parallel": true,
"self": 0.0007943679975142004
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 12665.141921073857,
"count": 1945340,
"is_parallel": true,
"self": 671.0726459099733,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 411.7606817029191,
"count": 1945340,
"is_parallel": true,
"self": 411.7606817029191
},
"communicator.exchange": {
"total": 9747.83525710191,
"count": 1945340,
"is_parallel": true,
"self": 9747.83525710191
},
"steps_from_proto": {
"total": 1834.4733363590549,
"count": 3890680,
"is_parallel": true,
"self": 342.8023757441897,
"children": {
"_process_rank_one_or_two_observation": {
"total": 1491.6709606148652,
"count": 15562720,
"is_parallel": true,
"self": 1491.6709606148652
}
}
}
}
},
"steps_from_proto": {
"total": 0.15636666999125737,
"count": 280,
"is_parallel": true,
"self": 0.02982434398290934,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.12654232600834803,
"count": 1120,
"is_parallel": true,
"self": 0.12654232600834803
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 6673.6578696017605,
"count": 1945341,
"self": 184.108814187186,
"children": {
"process_trajectory": {
"total": 2964.8615020155776,
"count": 1945341,
"self": 2958.085031615572,
"children": {
"RLTrainer._checkpoint": {
"total": 6.776470400005564,
"count": 56,
"self": 6.776470400005564
}
}
},
"_update_policy": {
"total": 3524.687553398997,
"count": 1361,
"self": 1976.29639970871,
"children": {
"TorchPOCAOptimizer.update": {
"total": 1548.391153690287,
"count": 40830,
"self": 1548.391153690287
}
}
}
}
}
}
},
"trainer_threads": {
"total": 5.729962140321732e-07,
"count": 1,
"self": 5.729962140321732e-07
},
"TrainerController._save_models": {
"total": 0.10068461800256046,
"count": 1,
"self": 0.0008691920011187904,
"children": {
"RLTrainer._checkpoint": {
"total": 0.09981542600144166,
"count": 1,
"self": 0.09981542600144166
}
}
}
}
}
}
}