{ "name": "root", "gauges": { "Huggy.Policy.Entropy.mean": { "value": 1.410301923751831, "min": 1.410301923751831, "max": 1.4296376705169678, "count": 40 }, "Huggy.Policy.Entropy.sum": { "value": 70557.40625, "min": 68387.890625, "max": 77001.578125, "count": 40 }, "Huggy.Environment.EpisodeLength.mean": { "value": 91.93703703703704, "min": 79.56521739130434, "max": 404.55645161290323, "count": 40 }, "Huggy.Environment.EpisodeLength.sum": { "value": 49646.0, "min": 48846.0, "max": 50165.0, "count": 40 }, "Huggy.Step.mean": { "value": 1999913.0, "min": 49698.0, "max": 1999913.0, "count": 40 }, "Huggy.Step.sum": { "value": 1999913.0, "min": 49698.0, "max": 1999913.0, "count": 40 }, "Huggy.Policy.ExtrinsicValueEstimate.mean": { "value": 2.4084649085998535, "min": 0.09907951205968857, "max": 2.45831298828125, "count": 40 }, "Huggy.Policy.ExtrinsicValueEstimate.sum": { "value": 1300.571044921875, "min": 12.186779975891113, "max": 1494.234619140625, "count": 40 }, "Huggy.Environment.CumulativeReward.mean": { "value": 3.779406287935045, "min": 1.8823601509012826, "max": 3.9763208026094223, "count": 40 }, "Huggy.Environment.CumulativeReward.sum": { "value": 2040.8793954849243, "min": 231.53029856085777, "max": 2330.742432832718, "count": 40 }, "Huggy.Policy.ExtrinsicReward.mean": { "value": 3.779406287935045, "min": 1.8823601509012826, "max": 3.9763208026094223, "count": 40 }, "Huggy.Policy.ExtrinsicReward.sum": { "value": 2040.8793954849243, "min": 231.53029856085777, "max": 2330.742432832718, "count": 40 }, "Huggy.Losses.PolicyLoss.mean": { "value": 0.0157028946294708, "min": 0.013045193434421284, "max": 0.020151615798628578, "count": 40 }, "Huggy.Losses.PolicyLoss.sum": { "value": 0.0471086838884124, "min": 0.026090386868842567, "max": 0.05819234631053405, "count": 40 }, "Huggy.Losses.ValueLoss.mean": { "value": 0.04564304662247499, "min": 0.021735196591665347, "max": 0.061515726149082184, "count": 40 }, "Huggy.Losses.ValueLoss.sum": { "value": 0.13692913986742497, "min": 0.043470393183330694, "max": 0.1633564282208681, "count": 40 }, "Huggy.Policy.LearningRate.mean": { "value": 3.3408988863999995e-06, "min": 3.3408988863999995e-06, "max": 0.000295317676560775, "count": 40 }, "Huggy.Policy.LearningRate.sum": { "value": 1.0022696659199998e-05, "min": 1.0022696659199998e-05, "max": 0.00084414136861955, "count": 40 }, "Huggy.Policy.Epsilon.mean": { "value": 0.10111359999999997, "min": 0.10111359999999997, "max": 0.198439225, "count": 40 }, "Huggy.Policy.Epsilon.sum": { "value": 0.3033407999999999, "min": 0.20739824999999995, "max": 0.58138045, "count": 40 }, "Huggy.Policy.Beta.mean": { "value": 6.556864000000001e-05, "min": 6.556864000000001e-05, "max": 0.004922117327499999, "count": 40 }, "Huggy.Policy.Beta.sum": { "value": 0.00019670592000000002, "min": 0.00019670592000000002, "max": 0.014070884455000002, "count": 40 }, "Huggy.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 40 }, "Huggy.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 40 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1690250429", "python_version": "3.10.6 (main, May 29 2023, 11:10:38) [GCC 11.3.0]", "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics", "mlagents_version": "0.31.0.dev0", "mlagents_envs_version": "0.31.0.dev0", "communication_protocol_version": "1.5.0", "pytorch_version": "1.11.0+cu102", "numpy_version": "1.21.2", "end_time_seconds": "1690252908" }, "total": 2478.924074893, "count": 1, "self": 0.7860827260005863, "children": { "run_training.setup": { "total": 0.054666994999934104, "count": 1, "self": 0.054666994999934104 }, "TrainerController.start_learning": { "total": 2478.0833251719996, "count": 1, "self": 4.35886963329267, "children": { "TrainerController._reset_env": { "total": 6.010378813999978, "count": 1, "self": 6.010378813999978 }, "TrainerController.advance": { "total": 2467.5242424857074, "count": 232384, "self": 4.5439013732898275, "children": { "env_step": { "total": 1903.8306120262382, "count": 232384, "self": 1608.4128491693195, "children": { "SubprocessEnvManager._take_step": { "total": 292.5165127110099, "count": 232384, "self": 16.760990450154623, "children": { "TorchPolicy.evaluate": { "total": 275.7555222608553, "count": 223024, "self": 275.7555222608553 } } }, "workers": { "total": 2.9012501459087616, "count": 232384, "self": 0.0, "children": { "worker_root": { "total": 2470.0347394651717, "count": 232384, "is_parallel": true, "self": 1158.6650374882274, "children": { "run_training.setup": { "total": 0.0, "count": 0, "is_parallel": true, "self": 0.0, "children": { "steps_from_proto": { "total": 0.001170564999938506, "count": 1, "is_parallel": true, "self": 0.000323770999784756, "children": { "_process_rank_one_or_two_observation": { "total": 0.00084679400015375, "count": 2, "is_parallel": true, "self": 0.00084679400015375 } } }, "UnityEnvironment.step": { "total": 0.031239979000019957, "count": 1, "is_parallel": true, "self": 0.00038832600012028706, "children": { "UnityEnvironment._generate_step_input": { "total": 0.0002449440003147174, "count": 1, "is_parallel": true, "self": 0.0002449440003147174 }, "communicator.exchange": { "total": 0.028458310999667447, "count": 1, "is_parallel": true, "self": 0.028458310999667447 }, "steps_from_proto": { "total": 0.0021483979999175062, "count": 1, "is_parallel": true, "self": 0.0002473539998391061, "children": { "_process_rank_one_or_two_observation": { "total": 0.0019010440000784001, "count": 2, "is_parallel": true, "self": 0.0019010440000784001 } } } } } } }, "UnityEnvironment.step": { "total": 1311.3697019769443, "count": 232383, "is_parallel": true, "self": 40.59940697097363, "children": { "UnityEnvironment._generate_step_input": { "total": 81.60389747299496, "count": 232383, "is_parallel": true, "self": 81.60389747299496 }, "communicator.exchange": { "total": 1091.8277008720856, "count": 232383, "is_parallel": true, "self": 1091.8277008720856 }, "steps_from_proto": { "total": 97.33869666089004, "count": 232383, "is_parallel": true, "self": 35.42831592991979, "children": { "_process_rank_one_or_two_observation": { "total": 61.910380730970246, "count": 464766, "is_parallel": true, "self": 61.910380730970246 } } } } } } } } } } }, "trainer_advance": { "total": 559.1497290861794, "count": 232384, "self": 6.5021071441115055, "children": { "process_trajectory": { "total": 141.95123428606985, "count": 232384, "self": 140.4046703940703, "children": { "RLTrainer._checkpoint": { "total": 1.54656389199954, "count": 10, "self": 1.54656389199954 } } }, "_update_policy": { "total": 410.696387655998, "count": 97, "self": 350.86794667300774, "children": { "TorchPPOOptimizer.update": { "total": 59.82844098299029, "count": 2910, "self": 59.82844098299029 } } } } } } }, "trainer_threads": { "total": 1.2370001059025526e-06, "count": 1, "self": 1.2370001059025526e-06 }, "TrainerController._save_models": { "total": 0.18983300199943187, "count": 1, "self": 0.0026977469997291337, "children": { "RLTrainer._checkpoint": { "total": 0.18713525499970274, "count": 1, "self": 0.18713525499970274 } } } } } } }