{ "name": "root", "gauges": { "Huggy.Policy.Entropy.mean": { "value": 1.3940504789352417, "min": 1.3940504789352417, "max": 1.4242143630981445, "count": 40 }, "Huggy.Policy.Entropy.sum": { "value": 69756.890625, "min": 69200.578125, "max": 75673.0703125, "count": 40 }, "Huggy.Environment.EpisodeLength.mean": { "value": 68.99859943977592, "min": 68.99859943977592, "max": 411.0983606557377, "count": 40 }, "Huggy.Environment.EpisodeLength.sum": { "value": 49265.0, "min": 49207.0, "max": 50154.0, "count": 40 }, "Huggy.Step.mean": { "value": 1999931.0, "min": 49526.0, "max": 1999931.0, "count": 40 }, "Huggy.Step.sum": { "value": 1999931.0, "min": 49526.0, "max": 1999931.0, "count": 40 }, "Huggy.Policy.ExtrinsicValueEstimate.mean": { "value": 2.4658634662628174, "min": 0.1285676509141922, "max": 2.504225969314575, "count": 40 }, "Huggy.Policy.ExtrinsicValueEstimate.sum": { "value": 1760.62646484375, "min": 15.556686401367188, "max": 1762.1583251953125, "count": 40 }, "Huggy.Environment.CumulativeReward.mean": { "value": 3.828852126494843, "min": 1.7286296090311255, "max": 3.969949254047908, "count": 40 }, "Huggy.Environment.CumulativeReward.sum": { "value": 2733.800418317318, "min": 209.1641826927662, "max": 2770.858525633812, "count": 40 }, "Huggy.Policy.ExtrinsicReward.mean": { "value": 3.828852126494843, "min": 1.7286296090311255, "max": 3.969949254047908, "count": 40 }, "Huggy.Policy.ExtrinsicReward.sum": { "value": 2733.800418317318, "min": 209.1641826927662, "max": 2770.858525633812, "count": 40 }, "Huggy.Losses.PolicyLoss.mean": { "value": 0.01650015048475729, "min": 0.013219569457093409, "max": 0.019359472492942585, "count": 40 }, "Huggy.Losses.PolicyLoss.sum": { "value": 0.04950045145427187, "min": 0.026439138914186817, "max": 0.05435747081549683, "count": 40 }, "Huggy.Losses.ValueLoss.mean": { "value": 0.05773245737784439, "min": 0.02191333652784427, "max": 0.060646917670965195, "count": 40 }, "Huggy.Losses.ValueLoss.sum": { "value": 0.17319737213353317, "min": 0.04382667305568854, "max": 0.17753612870971364, "count": 40 }, "Huggy.Policy.LearningRate.mean": { "value": 3.7824487392166675e-06, "min": 3.7824487392166675e-06, "max": 0.00029528655157114994, "count": 40 }, "Huggy.Policy.LearningRate.sum": { "value": 1.1347346217650003e-05, "min": 1.1347346217650003e-05, "max": 0.0008440686186438, "count": 40 }, "Huggy.Policy.Epsilon.mean": { "value": 0.10126078333333333, "min": 0.10126078333333333, "max": 0.19842885, "count": 40 }, "Huggy.Policy.Epsilon.sum": { "value": 0.30378235, "min": 0.20767645000000004, "max": 0.5813561999999999, "count": 40 }, "Huggy.Policy.Beta.mean": { "value": 7.291308833333334e-05, "min": 7.291308833333334e-05, "max": 0.004921599615000001, "count": 40 }, "Huggy.Policy.Beta.sum": { "value": 0.00021873926500000002, "min": 0.00021873926500000002, "max": 0.014069674379999996, "count": 40 }, "Huggy.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 40 }, "Huggy.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 40 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1690504593", "python_version": "3.10.6 (main, May 29 2023, 11:10:38) [GCC 11.3.0]", "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics", "mlagents_version": "0.31.0.dev0", "mlagents_envs_version": "0.31.0.dev0", "communication_protocol_version": "1.5.0", "pytorch_version": "1.11.0+cu102", "numpy_version": "1.21.2", "end_time_seconds": "1690507201" }, "total": 2608.529170688, "count": 1, "self": 0.4357949149994056, "children": { "run_training.setup": { "total": 0.03461415699985082, "count": 1, "self": 0.03461415699985082 }, "TrainerController.start_learning": { "total": 2608.0587616160005, "count": 1, "self": 4.689323622843858, "children": { "TrainerController._reset_env": { "total": 5.189831120000235, "count": 1, "self": 5.189831120000235 }, "TrainerController.advance": { "total": 2598.0608090951564, "count": 233705, "self": 5.02136477462318, "children": { "env_step": { "total": 2027.2054567147802, "count": 233705, "self": 1717.4056237437403, "children": { "SubprocessEnvManager._take_step": { "total": 306.56229980093485, "count": 233705, "self": 18.08111214291057, "children": { "TorchPolicy.evaluate": { "total": 288.4811876580243, "count": 222908, "self": 288.4811876580243 } } }, "workers": { "total": 3.2375331701050527, "count": 233705, "self": 0.0, "children": { "worker_root": { "total": 2599.686959268829, "count": 233705, "is_parallel": true, "self": 1197.2245663118201, "children": { "run_training.setup": { "total": 0.0, "count": 0, "is_parallel": true, "self": 0.0, "children": { "steps_from_proto": { "total": 0.0011354070002198569, "count": 1, "is_parallel": true, "self": 0.00031257900036507635, "children": { "_process_rank_one_or_two_observation": { "total": 0.0008228279998547805, "count": 2, "is_parallel": true, "self": 0.0008228279998547805 } } }, "UnityEnvironment.step": { "total": 0.03112583500023902, "count": 1, "is_parallel": true, "self": 0.00039741899990985985, "children": { "UnityEnvironment._generate_step_input": { "total": 0.0002480190000824223, "count": 1, "is_parallel": true, "self": 0.0002480190000824223 }, "communicator.exchange": { "total": 0.029702145000101154, "count": 1, "is_parallel": true, "self": 0.029702145000101154 }, "steps_from_proto": { "total": 0.000778252000145585, "count": 1, "is_parallel": true, "self": 0.00021843300010004896, "children": { "_process_rank_one_or_two_observation": { "total": 0.0005598190000455361, "count": 2, "is_parallel": true, "self": 0.0005598190000455361 } } } } } } }, "UnityEnvironment.step": { "total": 1402.462392957009, "count": 233704, "is_parallel": true, "self": 42.86533621156923, "children": { "UnityEnvironment._generate_step_input": { "total": 91.84006307803475, "count": 233704, "is_parallel": true, "self": 91.84006307803475 }, "communicator.exchange": { "total": 1162.8829602061132, "count": 233704, "is_parallel": true, "self": 1162.8829602061132 }, "steps_from_proto": { "total": 104.87403346129167, "count": 233704, "is_parallel": true, "self": 40.12820051344761, "children": { "_process_rank_one_or_two_observation": { "total": 64.74583294784406, "count": 467408, "is_parallel": true, "self": 64.74583294784406 } } } } } } } } } } }, "trainer_advance": { "total": 565.833987605753, "count": 233705, "self": 6.920068913480009, "children": { "process_trajectory": { "total": 153.78612021027539, "count": 233705, "self": 152.54108510627566, "children": { "RLTrainer._checkpoint": { "total": 1.2450351039997258, "count": 10, "self": 1.2450351039997258 } } }, "_update_policy": { "total": 405.1277984819976, "count": 97, "self": 345.7155409989841, "children": { "TorchPPOOptimizer.update": { "total": 59.41225748301349, "count": 2910, "self": 59.41225748301349 } } } } } } }, "trainer_threads": { "total": 1.1640004231594503e-06, "count": 1, "self": 1.1640004231594503e-06 }, "TrainerController._save_models": { "total": 0.1187966139996206, "count": 1, "self": 0.002011046999541577, "children": { "RLTrainer._checkpoint": { "total": 0.11678556700007903, "count": 1, "self": 0.11678556700007903 } } } } } } }