ppo-Pyramids / run_logs /timers.json
YojitShinde's picture
First Push
972c7d9
{
"name": "root",
"gauges": {
"Pyramids.Policy.Entropy.mean": {
"value": 0.5392292141914368,
"min": 0.5392292141914368,
"max": 1.3762880563735962,
"count": 33
},
"Pyramids.Policy.Entropy.sum": {
"value": 16038.8330078125,
"min": 16038.8330078125,
"max": 41751.07421875,
"count": 33
},
"Pyramids.Step.mean": {
"value": 989905.0,
"min": 29952.0,
"max": 989905.0,
"count": 33
},
"Pyramids.Step.sum": {
"value": 989905.0,
"min": 29952.0,
"max": 989905.0,
"count": 33
},
"Pyramids.Policy.ExtrinsicValueEstimate.mean": {
"value": 0.12375917285680771,
"min": -0.10935390740633011,
"max": 0.27559468150138855,
"count": 33
},
"Pyramids.Policy.ExtrinsicValueEstimate.sum": {
"value": 31.063552856445312,
"min": -26.135583877563477,
"max": 65.3159408569336,
"count": 33
},
"Pyramids.Policy.RndValueEstimate.mean": {
"value": 0.03846127912402153,
"min": 0.01337270624935627,
"max": 0.35965967178344727,
"count": 33
},
"Pyramids.Policy.RndValueEstimate.sum": {
"value": 9.653780937194824,
"min": 3.3164310455322266,
"max": 86.31832122802734,
"count": 33
},
"Pyramids.Losses.PolicyLoss.mean": {
"value": 0.07064416863589572,
"min": 0.06549117608561314,
"max": 0.07457250677349035,
"count": 33
},
"Pyramids.Losses.PolicyLoss.sum": {
"value": 1.0596625295384359,
"min": 0.4805440582256716,
"max": 1.0596625295384359,
"count": 33
},
"Pyramids.Losses.ValueLoss.mean": {
"value": 0.00614799391438116,
"min": 0.00011006560505229133,
"max": 0.009394352461179197,
"count": 33
},
"Pyramids.Losses.ValueLoss.sum": {
"value": 0.0922199087157174,
"min": 0.0015409184707320786,
"max": 0.10508711506051494,
"count": 33
},
"Pyramids.Policy.LearningRate.mean": {
"value": 7.4897975034333325e-06,
"min": 7.4897975034333325e-06,
"max": 0.00029515063018788575,
"count": 33
},
"Pyramids.Policy.LearningRate.sum": {
"value": 0.00011234696255149998,
"min": 0.00011234696255149998,
"max": 0.0032558279147240996,
"count": 33
},
"Pyramids.Policy.Epsilon.mean": {
"value": 0.10249656666666668,
"min": 0.10249656666666668,
"max": 0.19838354285714285,
"count": 33
},
"Pyramids.Policy.Epsilon.sum": {
"value": 1.5374485000000002,
"min": 1.3886848,
"max": 2.4434376000000007,
"count": 33
},
"Pyramids.Policy.Beta.mean": {
"value": 0.0002594070099999999,
"min": 0.0002594070099999999,
"max": 0.00983851593142857,
"count": 33
},
"Pyramids.Policy.Beta.sum": {
"value": 0.003891105149999999,
"min": 0.003891105149999999,
"max": 0.10854906241000001,
"count": 33
},
"Pyramids.Losses.RNDLoss.mean": {
"value": 0.01595507189631462,
"min": 0.01577533408999443,
"max": 0.4828554093837738,
"count": 33
},
"Pyramids.Losses.RNDLoss.sum": {
"value": 0.2393260896205902,
"min": 0.22085466980934143,
"max": 3.379987955093384,
"count": 33
},
"Pyramids.Environment.EpisodeLength.mean": {
"value": 742.6511627906976,
"min": 742.6511627906976,
"max": 999.0,
"count": 33
},
"Pyramids.Environment.EpisodeLength.sum": {
"value": 31934.0,
"min": 15984.0,
"max": 32923.0,
"count": 33
},
"Pyramids.Environment.CumulativeReward.mean": {
"value": 0.3268185652965723,
"min": -1.0000000521540642,
"max": 0.45947022897166173,
"count": 33
},
"Pyramids.Environment.CumulativeReward.sum": {
"value": 14.05319830775261,
"min": -31.998401656746864,
"max": 17.000398471951485,
"count": 33
},
"Pyramids.Policy.ExtrinsicReward.mean": {
"value": 0.3268185652965723,
"min": -1.0000000521540642,
"max": 0.45947022897166173,
"count": 33
},
"Pyramids.Policy.ExtrinsicReward.sum": {
"value": 14.05319830775261,
"min": -31.998401656746864,
"max": 17.000398471951485,
"count": 33
},
"Pyramids.Policy.RndReward.mean": {
"value": 0.1215952348659833,
"min": 0.1215952348659833,
"max": 9.277738269418478,
"count": 33
},
"Pyramids.Policy.RndReward.sum": {
"value": 5.228595099237282,
"min": 4.518147401162423,
"max": 148.44381231069565,
"count": 33
},
"Pyramids.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 33
},
"Pyramids.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 33
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1689532314",
"python_version": "3.10.12 (main, Jun 7 2023, 12:45:35) [GCC 9.4.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/PyramidsRND.yaml --env=./training-envs-executables/linux/Pyramids/Pyramids --run-id=Pyramids Training --no-graphics",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.11.0+cu102",
"numpy_version": "1.21.2",
"end_time_seconds": "1689534416"
},
"total": 2101.7249607689996,
"count": 1,
"self": 0.5251996839992898,
"children": {
"run_training.setup": {
"total": 0.04021469799999977,
"count": 1,
"self": 0.04021469799999977
},
"TrainerController.start_learning": {
"total": 2101.1595463870003,
"count": 1,
"self": 1.1981772910171458,
"children": {
"TrainerController._reset_env": {
"total": 4.708219864000057,
"count": 1,
"self": 4.708219864000057
},
"TrainerController.advance": {
"total": 2095.157421332984,
"count": 63191,
"self": 1.252410491955743,
"children": {
"env_step": {
"total": 1460.8306770980244,
"count": 63191,
"self": 1355.29740911211,
"children": {
"SubprocessEnvManager._take_step": {
"total": 104.79497708794725,
"count": 63191,
"self": 4.614229700891883,
"children": {
"TorchPolicy.evaluate": {
"total": 100.18074738705536,
"count": 62561,
"self": 100.18074738705536
}
}
},
"workers": {
"total": 0.7382908979670901,
"count": 63191,
"self": 0.0,
"children": {
"worker_root": {
"total": 2096.631769648029,
"count": 63191,
"is_parallel": true,
"self": 849.4502458101242,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.001735353999720246,
"count": 1,
"is_parallel": true,
"self": 0.0005094829998597561,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.00122587099986049,
"count": 8,
"is_parallel": true,
"self": 0.00122587099986049
}
}
},
"UnityEnvironment.step": {
"total": 0.10125496799992106,
"count": 1,
"is_parallel": true,
"self": 0.0005956850000075065,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.00045543900023403694,
"count": 1,
"is_parallel": true,
"self": 0.00045543900023403694
},
"communicator.exchange": {
"total": 0.09827383399988321,
"count": 1,
"is_parallel": true,
"self": 0.09827383399988321
},
"steps_from_proto": {
"total": 0.0019300099997963116,
"count": 1,
"is_parallel": true,
"self": 0.00035073199887847295,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0015792780009178387,
"count": 8,
"is_parallel": true,
"self": 0.0015792780009178387
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1247.181523837905,
"count": 63190,
"is_parallel": true,
"self": 33.79419163083503,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 23.133726682954148,
"count": 63190,
"is_parallel": true,
"self": 23.133726682954148
},
"communicator.exchange": {
"total": 1085.2138783109976,
"count": 63190,
"is_parallel": true,
"self": 1085.2138783109976
},
"steps_from_proto": {
"total": 105.03972721311811,
"count": 63190,
"is_parallel": true,
"self": 19.831815649127748,
"children": {
"_process_rank_one_or_two_observation": {
"total": 85.20791156399036,
"count": 505520,
"is_parallel": true,
"self": 85.20791156399036
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 633.074333743004,
"count": 63191,
"self": 2.296263444010492,
"children": {
"process_trajectory": {
"total": 105.96893978599837,
"count": 63191,
"self": 105.70650721399852,
"children": {
"RLTrainer._checkpoint": {
"total": 0.26243257199985237,
"count": 2,
"self": 0.26243257199985237
}
}
},
"_update_policy": {
"total": 524.8091305129951,
"count": 441,
"self": 336.94047409405675,
"children": {
"TorchPPOOptimizer.update": {
"total": 187.86865641893837,
"count": 22794,
"self": 187.86865641893837
}
}
}
}
}
}
},
"trainer_threads": {
"total": 8.239994713221677e-07,
"count": 1,
"self": 8.239994713221677e-07
},
"TrainerController._save_models": {
"total": 0.09572707499955868,
"count": 1,
"self": 0.0014443489999393933,
"children": {
"RLTrainer._checkpoint": {
"total": 0.09428272599961929,
"count": 1,
"self": 0.09428272599961929
}
}
}
}
}
}
}