ppo-PyramidsRND / run_logs /timers.json
tanmayyyj's picture
First Push
627e183
raw
history blame
18.7 kB
{
"name": "root",
"gauges": {
"Pyramids.Policy.Entropy.mean": {
"value": 0.3470213711261749,
"min": 0.33146047592163086,
"max": 1.4657706022262573,
"count": 33
},
"Pyramids.Policy.Entropy.sum": {
"value": 10421.74609375,
"min": 9869.5673828125,
"max": 44465.6171875,
"count": 33
},
"Pyramids.Step.mean": {
"value": 989974.0,
"min": 29922.0,
"max": 989974.0,
"count": 33
},
"Pyramids.Step.sum": {
"value": 989974.0,
"min": 29922.0,
"max": 989974.0,
"count": 33
},
"Pyramids.Policy.ExtrinsicValueEstimate.mean": {
"value": 0.674224317073822,
"min": -0.11279188096523285,
"max": 0.6799185872077942,
"count": 33
},
"Pyramids.Policy.ExtrinsicValueEstimate.sum": {
"value": 196.19927978515625,
"min": -27.182844161987305,
"max": 196.19927978515625,
"count": 33
},
"Pyramids.Policy.RndValueEstimate.mean": {
"value": -0.003972766920924187,
"min": -0.050487518310546875,
"max": 0.4214654564857483,
"count": 33
},
"Pyramids.Policy.RndValueEstimate.sum": {
"value": -1.1560752391815186,
"min": -13.379192352294922,
"max": 99.88731384277344,
"count": 33
},
"Pyramids.Losses.PolicyLoss.mean": {
"value": 0.06674635335682377,
"min": 0.06456962344326346,
"max": 0.07450860194096874,
"count": 33
},
"Pyramids.Losses.PolicyLoss.sum": {
"value": 0.9344489469955328,
"min": 0.5651454042255298,
"max": 1.0894886221560904,
"count": 33
},
"Pyramids.Losses.ValueLoss.mean": {
"value": 0.016521679153230565,
"min": 0.00015447562862011175,
"max": 0.017013443085302712,
"count": 33
},
"Pyramids.Losses.ValueLoss.sum": {
"value": 0.2313035081452279,
"min": 0.0016992319148212292,
"max": 0.23818820319423797,
"count": 33
},
"Pyramids.Policy.LearningRate.mean": {
"value": 1.0183640311257144e-05,
"min": 1.0183640311257144e-05,
"max": 0.000393124501718875,
"count": 33
},
"Pyramids.Policy.LearningRate.sum": {
"value": 0.0001425709643576,
"min": 0.0001425709643576,
"max": 0.0045104706723824006,
"count": 33
},
"Pyramids.Policy.Epsilon.mean": {
"value": 0.10254588571428572,
"min": 0.10254588571428572,
"max": 0.198281125,
"count": 33
},
"Pyramids.Policy.Epsilon.sum": {
"value": 1.4356424,
"min": 1.4356424,
"max": 2.5276176,
"count": 33
},
"Pyramids.Policy.Beta.mean": {
"value": 0.0002643339828571429,
"min": 0.0002643339828571429,
"max": 0.0098282843875,
"count": 33
},
"Pyramids.Policy.Beta.sum": {
"value": 0.0037006757600000005,
"min": 0.0037006757600000005,
"max": 0.11278899824,
"count": 33
},
"Pyramids.Losses.RNDLoss.mean": {
"value": 0.010571293532848358,
"min": 0.010316801257431507,
"max": 0.5025671720504761,
"count": 33
},
"Pyramids.Losses.RNDLoss.sum": {
"value": 0.14799810945987701,
"min": 0.14443521201610565,
"max": 4.020537376403809,
"count": 33
},
"Pyramids.Environment.EpisodeLength.mean": {
"value": 269.34285714285716,
"min": 269.34285714285716,
"max": 999.0,
"count": 33
},
"Pyramids.Environment.EpisodeLength.sum": {
"value": 28281.0,
"min": 16465.0,
"max": 32382.0,
"count": 33
},
"Pyramids.Environment.CumulativeReward.mean": {
"value": 1.6860226385014236,
"min": -0.999962551984936,
"max": 1.6860226385014236,
"count": 33
},
"Pyramids.Environment.CumulativeReward.sum": {
"value": 178.7183996811509,
"min": -31.998801663517952,
"max": 178.7183996811509,
"count": 33
},
"Pyramids.Policy.ExtrinsicReward.mean": {
"value": 1.6860226385014236,
"min": -0.999962551984936,
"max": 1.6860226385014236,
"count": 33
},
"Pyramids.Policy.ExtrinsicReward.sum": {
"value": 178.7183996811509,
"min": -31.998801663517952,
"max": 178.7183996811509,
"count": 33
},
"Pyramids.Policy.RndReward.mean": {
"value": 0.029616010237661167,
"min": 0.029616010237661167,
"max": 10.240952908992767,
"count": 33
},
"Pyramids.Policy.RndReward.sum": {
"value": 3.1392970851920836,
"min": 3.133239044545917,
"max": 174.09619945287704,
"count": 33
},
"Pyramids.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 33
},
"Pyramids.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 33
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1687285473",
"python_version": "3.10.12 (main, Jun 7 2023, 12:45:35) [GCC 9.4.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/PyramidsRND.yaml --env=./training-envs-executables/linux/Pyramids/Pyramids --run-id=Pyramids Training --no-graphics",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.11.0+cu102",
"numpy_version": "1.21.2",
"end_time_seconds": "1687287756"
},
"total": 2283.5389136149997,
"count": 1,
"self": 0.8457689189995108,
"children": {
"run_training.setup": {
"total": 0.06628980000004958,
"count": 1,
"self": 0.06628980000004958
},
"TrainerController.start_learning": {
"total": 2282.626854896,
"count": 1,
"self": 1.393986217065958,
"children": {
"TrainerController._reset_env": {
"total": 4.274403335999978,
"count": 1,
"self": 4.274403335999978
},
"TrainerController.advance": {
"total": 2276.769560419934,
"count": 63989,
"self": 1.3973200408090634,
"children": {
"env_step": {
"total": 1644.5954150750345,
"count": 63989,
"self": 1529.968176437063,
"children": {
"SubprocessEnvManager._take_step": {
"total": 113.7951267239705,
"count": 63989,
"self": 5.021807053010889,
"children": {
"TorchPolicy.evaluate": {
"total": 108.77331967095961,
"count": 62553,
"self": 108.77331967095961
}
}
},
"workers": {
"total": 0.8321119140009614,
"count": 63989,
"self": 0.0,
"children": {
"worker_root": {
"total": 2277.3865459109334,
"count": 63989,
"is_parallel": true,
"self": 864.7544263749496,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0019160439996994683,
"count": 1,
"is_parallel": true,
"self": 0.0005733529997087317,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0013426909999907366,
"count": 8,
"is_parallel": true,
"self": 0.0013426909999907366
}
}
},
"UnityEnvironment.step": {
"total": 0.05089865300033125,
"count": 1,
"is_parallel": true,
"self": 0.0006007300007695449,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.000509100999806833,
"count": 1,
"is_parallel": true,
"self": 0.000509100999806833
},
"communicator.exchange": {
"total": 0.04780261099995187,
"count": 1,
"is_parallel": true,
"self": 0.04780261099995187
},
"steps_from_proto": {
"total": 0.0019862109998030064,
"count": 1,
"is_parallel": true,
"self": 0.0004314769989832712,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0015547340008197352,
"count": 8,
"is_parallel": true,
"self": 0.0015547340008197352
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1412.6321195359837,
"count": 63988,
"is_parallel": true,
"self": 34.476019391085174,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 24.927792617907926,
"count": 63988,
"is_parallel": true,
"self": 24.927792617907926
},
"communicator.exchange": {
"total": 1244.3980112989916,
"count": 63988,
"is_parallel": true,
"self": 1244.3980112989916
},
"steps_from_proto": {
"total": 108.83029622799904,
"count": 63988,
"is_parallel": true,
"self": 21.85862761510998,
"children": {
"_process_rank_one_or_two_observation": {
"total": 86.97166861288906,
"count": 511904,
"is_parallel": true,
"self": 86.97166861288906
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 630.7768253040904,
"count": 63989,
"self": 2.675048261048687,
"children": {
"process_trajectory": {
"total": 114.7818651320431,
"count": 63989,
"self": 114.4813120710428,
"children": {
"RLTrainer._checkpoint": {
"total": 0.30055306100030066,
"count": 2,
"self": 0.30055306100030066
}
}
},
"_update_policy": {
"total": 513.3199119109986,
"count": 447,
"self": 326.7403567330052,
"children": {
"TorchPPOOptimizer.update": {
"total": 186.57955517799337,
"count": 22836,
"self": 186.57955517799337
}
}
}
}
}
}
},
"trainer_threads": {
"total": 2.4000000848900527e-06,
"count": 1,
"self": 2.4000000848900527e-06
},
"TrainerController._save_models": {
"total": 0.18890252300025168,
"count": 1,
"self": 0.003716737000104331,
"children": {
"RLTrainer._checkpoint": {
"total": 0.18518578600014735,
"count": 1,
"self": 0.18518578600014735
}
}
}
}
}
}
}