ppo-Huggy / run_logs /timers.json
javenda's picture
Huggy
d5e4241
{
"name": "root",
"gauges": {
"Huggy.Policy.Entropy.mean": {
"value": 1.409104824066162,
"min": 1.409104824066162,
"max": 1.426338791847229,
"count": 40
},
"Huggy.Policy.Entropy.sum": {
"value": 71829.1171875,
"min": 66677.8125,
"max": 79904.203125,
"count": 40
},
"Huggy.Environment.EpisodeLength.mean": {
"value": 107.83006535947712,
"min": 91.27726432532347,
"max": 383.7153846153846,
"count": 40
},
"Huggy.Environment.EpisodeLength.sum": {
"value": 49494.0,
"min": 48812.0,
"max": 50370.0,
"count": 40
},
"Huggy.Step.mean": {
"value": 1999977.0,
"min": 49890.0,
"max": 1999977.0,
"count": 40
},
"Huggy.Step.sum": {
"value": 1999977.0,
"min": 49890.0,
"max": 1999977.0,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.mean": {
"value": 2.3052117824554443,
"min": 0.14171214401721954,
"max": 2.414104700088501,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.sum": {
"value": 1058.0921630859375,
"min": 18.280866622924805,
"max": 1278.66943359375,
"count": 40
},
"Huggy.Environment.CumulativeReward.mean": {
"value": 3.4680046904061097,
"min": 1.7960675633692926,
"max": 3.879901185429212,
"count": 40
},
"Huggy.Environment.CumulativeReward.sum": {
"value": 1591.8141528964043,
"min": 231.69271567463875,
"max": 2057.1222940683365,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.mean": {
"value": 3.4680046904061097,
"min": 1.7960675633692926,
"max": 3.879901185429212,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.sum": {
"value": 1591.8141528964043,
"min": 231.69271567463875,
"max": 2057.1222940683365,
"count": 40
},
"Huggy.Losses.PolicyLoss.mean": {
"value": 0.015855957536647716,
"min": 0.01359199967046152,
"max": 0.019654141543530082,
"count": 40
},
"Huggy.Losses.PolicyLoss.sum": {
"value": 0.04756787260994315,
"min": 0.02718399934092304,
"max": 0.05749147952689479,
"count": 40
},
"Huggy.Losses.ValueLoss.mean": {
"value": 0.04562835887902313,
"min": 0.0215619460058709,
"max": 0.062098233857088625,
"count": 40
},
"Huggy.Losses.ValueLoss.sum": {
"value": 0.1368850766370694,
"min": 0.0431238920117418,
"max": 0.18629470157126587,
"count": 40
},
"Huggy.Policy.LearningRate.mean": {
"value": 3.2934489022166725e-06,
"min": 3.2934489022166725e-06,
"max": 0.00029534535155155,
"count": 40
},
"Huggy.Policy.LearningRate.sum": {
"value": 9.880346706650018e-06,
"min": 9.880346706650018e-06,
"max": 0.0008438941687019499,
"count": 40
},
"Huggy.Policy.Epsilon.mean": {
"value": 0.1010977833333333,
"min": 0.1010977833333333,
"max": 0.19844845,
"count": 40
},
"Huggy.Policy.Epsilon.sum": {
"value": 0.3032933499999999,
"min": 0.20733220000000005,
"max": 0.5812980499999999,
"count": 40
},
"Huggy.Policy.Beta.mean": {
"value": 6.477938833333344e-05,
"min": 6.477938833333344e-05,
"max": 0.004922577655000001,
"count": 40
},
"Huggy.Policy.Beta.sum": {
"value": 0.00019433816500000033,
"min": 0.00019433816500000033,
"max": 0.014066772694999997,
"count": 40
},
"Huggy.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
},
"Huggy.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1671959515",
"python_version": "3.8.16 (default, Dec 7 2022, 01:12:13) \n[GCC 7.5.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics",
"mlagents_version": "0.29.0.dev0",
"mlagents_envs_version": "0.29.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.8.1+cu102",
"numpy_version": "1.21.6",
"end_time_seconds": "1671961909"
},
"total": 2393.5634094630004,
"count": 1,
"self": 0.4395255230010662,
"children": {
"run_training.setup": {
"total": 0.10978108299991618,
"count": 1,
"self": 0.10978108299991618
},
"TrainerController.start_learning": {
"total": 2393.0141028569997,
"count": 1,
"self": 4.250977300064733,
"children": {
"TrainerController._reset_env": {
"total": 7.506777177999993,
"count": 1,
"self": 7.506777177999993
},
"TrainerController.advance": {
"total": 2381.136416293935,
"count": 231342,
"self": 4.538514965009199,
"children": {
"env_step": {
"total": 1896.1896470619035,
"count": 231342,
"self": 1596.3175394329774,
"children": {
"SubprocessEnvManager._take_step": {
"total": 297.14846167995563,
"count": 231342,
"self": 15.979909401900045,
"children": {
"TorchPolicy.evaluate": {
"total": 281.1685522780556,
"count": 223014,
"self": 69.29229771698692,
"children": {
"TorchPolicy.sample_actions": {
"total": 211.87625456106866,
"count": 223014,
"self": 211.87625456106866
}
}
}
}
},
"workers": {
"total": 2.7236459489706704,
"count": 231342,
"self": 0.0,
"children": {
"worker_root": {
"total": 2384.4803144220537,
"count": 231342,
"is_parallel": true,
"self": 1072.2757925670107,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.002173029999994469,
"count": 1,
"is_parallel": true,
"self": 0.00035237999986748036,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0018206500001269887,
"count": 2,
"is_parallel": true,
"self": 0.0018206500001269887
}
}
},
"UnityEnvironment.step": {
"total": 0.031368645000043216,
"count": 1,
"is_parallel": true,
"self": 0.00028930000007676426,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.0001856410000300457,
"count": 1,
"is_parallel": true,
"self": 0.0001856410000300457
},
"communicator.exchange": {
"total": 0.030012277999958314,
"count": 1,
"is_parallel": true,
"self": 0.030012277999958314
},
"steps_from_proto": {
"total": 0.0008814259999780916,
"count": 1,
"is_parallel": true,
"self": 0.0003188659999295851,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0005625600000485065,
"count": 2,
"is_parallel": true,
"self": 0.0005625600000485065
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1312.204521855043,
"count": 231341,
"is_parallel": true,
"self": 36.156349553181826,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 84.43477471002484,
"count": 231341,
"is_parallel": true,
"self": 84.43477471002484
},
"communicator.exchange": {
"total": 1091.4797398148826,
"count": 231341,
"is_parallel": true,
"self": 1091.4797398148826
},
"steps_from_proto": {
"total": 100.13365777695356,
"count": 231341,
"is_parallel": true,
"self": 43.78459202697161,
"children": {
"_process_rank_one_or_two_observation": {
"total": 56.349065749981946,
"count": 462682,
"is_parallel": true,
"self": 56.349065749981946
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 480.40825426702213,
"count": 231342,
"self": 6.480166209100048,
"children": {
"process_trajectory": {
"total": 154.20592729592295,
"count": 231342,
"self": 153.007204229923,
"children": {
"RLTrainer._checkpoint": {
"total": 1.1987230659999568,
"count": 10,
"self": 1.1987230659999568
}
}
},
"_update_policy": {
"total": 319.72216076199913,
"count": 97,
"self": 266.28918515600947,
"children": {
"TorchPPOOptimizer.update": {
"total": 53.43297560598967,
"count": 2910,
"self": 53.43297560598967
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.094000253942795e-06,
"count": 1,
"self": 1.094000253942795e-06
},
"TrainerController._save_models": {
"total": 0.11993099099981919,
"count": 1,
"self": 0.0024377260001529066,
"children": {
"RLTrainer._checkpoint": {
"total": 0.11749326499966628,
"count": 1,
"self": 0.11749326499966628
}
}
}
}
}
}
}