ppo-PyramidsRND / run_logs /timers.json
Facepalm0's picture
First Push
6991074 verified
raw
history blame
18.7 kB
{
"name": "root",
"gauges": {
"Pyramids.Policy.Entropy.mean": {
"value": 0.4732387363910675,
"min": 0.46456775069236755,
"max": 1.5208418369293213,
"count": 33
},
"Pyramids.Policy.Entropy.sum": {
"value": 14454.603515625,
"min": 13899.8671875,
"max": 46136.2578125,
"count": 33
},
"Pyramids.Step.mean": {
"value": 989973.0,
"min": 29934.0,
"max": 989973.0,
"count": 33
},
"Pyramids.Step.sum": {
"value": 989973.0,
"min": 29934.0,
"max": 989973.0,
"count": 33
},
"Pyramids.Policy.ExtrinsicValueEstimate.mean": {
"value": 0.5489568710327148,
"min": -0.18212325870990753,
"max": 0.5831024050712585,
"count": 33
},
"Pyramids.Policy.ExtrinsicValueEstimate.sum": {
"value": 152.61001586914062,
"min": -43.163211822509766,
"max": 162.68557739257812,
"count": 33
},
"Pyramids.Policy.RndValueEstimate.mean": {
"value": -1.2688008546829224,
"min": -1.2688008546829224,
"max": 0.4554428458213806,
"count": 33
},
"Pyramids.Policy.RndValueEstimate.sum": {
"value": -352.7266540527344,
"min": -352.7266540527344,
"max": 107.93995666503906,
"count": 33
},
"Pyramids.Losses.PolicyLoss.mean": {
"value": 0.07093064914818974,
"min": 0.06423239169784614,
"max": 0.07235053450137205,
"count": 33
},
"Pyramids.Losses.PolicyLoss.sum": {
"value": 0.9930290880746564,
"min": 0.5011395681140965,
"max": 1.0711536195691829,
"count": 33
},
"Pyramids.Losses.ValueLoss.mean": {
"value": 0.1653442146810599,
"min": 0.00010283258653561238,
"max": 0.1653442146810599,
"count": 33
},
"Pyramids.Losses.ValueLoss.sum": {
"value": 2.3148190055348383,
"min": 0.001336823624962961,
"max": 2.3148190055348383,
"count": 33
},
"Pyramids.Policy.LearningRate.mean": {
"value": 7.259133294607143e-06,
"min": 7.259133294607143e-06,
"max": 0.00029523368730305713,
"count": 33
},
"Pyramids.Policy.LearningRate.sum": {
"value": 0.0001016278661245,
"min": 0.0001016278661245,
"max": 0.0035093099302300995,
"count": 33
},
"Pyramids.Policy.Epsilon.mean": {
"value": 0.1024196785714286,
"min": 0.1024196785714286,
"max": 0.19841122857142857,
"count": 33
},
"Pyramids.Policy.Epsilon.sum": {
"value": 1.4338755000000003,
"min": 1.3888786,
"max": 2.5697699,
"count": 33
},
"Pyramids.Policy.Beta.mean": {
"value": 0.0002517258892857143,
"min": 0.0002517258892857143,
"max": 0.009841281734285712,
"count": 33
},
"Pyramids.Policy.Beta.sum": {
"value": 0.00352416245,
"min": 0.00352416245,
"max": 0.11700001301,
"count": 33
},
"Pyramids.Losses.RNDLoss.mean": {
"value": 0.011892594397068024,
"min": 0.011687462218105793,
"max": 0.3864750266075134,
"count": 33
},
"Pyramids.Losses.RNDLoss.sum": {
"value": 0.16649632155895233,
"min": 0.16362446546554565,
"max": 2.705325126647949,
"count": 33
},
"Pyramids.Environment.EpisodeLength.mean": {
"value": 370.8641975308642,
"min": 333.7816091954023,
"max": 999.0,
"count": 33
},
"Pyramids.Environment.EpisodeLength.sum": {
"value": 30040.0,
"min": 16733.0,
"max": 32396.0,
"count": 33
},
"Pyramids.Environment.CumulativeReward.mean": {
"value": 1.5807012096967226,
"min": -0.9999548908202879,
"max": 1.6432183772839348,
"count": 33
},
"Pyramids.Environment.CumulativeReward.sum": {
"value": 128.03679798543453,
"min": -31.99360166490078,
"max": 142.95999882370234,
"count": 33
},
"Pyramids.Policy.ExtrinsicReward.mean": {
"value": 1.5807012096967226,
"min": -0.9999548908202879,
"max": 1.6432183772839348,
"count": 33
},
"Pyramids.Policy.ExtrinsicReward.sum": {
"value": 128.03679798543453,
"min": -31.99360166490078,
"max": 142.95999882370234,
"count": 33
},
"Pyramids.Policy.RndReward.mean": {
"value": 0.045621416668758304,
"min": 0.041157531930701055,
"max": 7.6459359693176605,
"count": 33
},
"Pyramids.Policy.RndReward.sum": {
"value": 3.6953347501694225,
"min": 3.580705277970992,
"max": 129.98091147840023,
"count": 33
},
"Pyramids.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 33
},
"Pyramids.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 33
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1707038004",
"python_version": "3.10.12 (main, Nov 20 2023, 15:14:05) [GCC 11.4.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/PyramidsRND.yaml --env=./training-envs-executables/linux/Pyramids/Pyramids --run-id=Pyramids Training --no-graphics --resume",
"mlagents_version": "1.1.0.dev0",
"mlagents_envs_version": "1.1.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "2.2.0+cu121",
"numpy_version": "1.23.5",
"end_time_seconds": "1707040212"
},
"total": 2207.848844487,
"count": 1,
"self": 0.5310461310000392,
"children": {
"run_training.setup": {
"total": 0.08685310600003504,
"count": 1,
"self": 0.08685310600003504
},
"TrainerController.start_learning": {
"total": 2207.23094525,
"count": 1,
"self": 1.3491721259533733,
"children": {
"TrainerController._reset_env": {
"total": 2.953743874999873,
"count": 1,
"self": 2.953743874999873
},
"TrainerController.advance": {
"total": 2202.838695998046,
"count": 63655,
"self": 1.4529343749818509,
"children": {
"env_step": {
"total": 1555.2516809680665,
"count": 63655,
"self": 1422.2122332520785,
"children": {
"SubprocessEnvManager._take_step": {
"total": 132.1737683370361,
"count": 63655,
"self": 4.774521102196104,
"children": {
"TorchPolicy.evaluate": {
"total": 127.39924723484,
"count": 62552,
"self": 127.39924723484
}
}
},
"workers": {
"total": 0.8656793789518815,
"count": 63655,
"self": 0.0,
"children": {
"worker_root": {
"total": 2202.1310441380656,
"count": 63655,
"is_parallel": true,
"self": 897.8805398911154,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0029949600002510124,
"count": 1,
"is_parallel": true,
"self": 0.0009159220007859403,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.002079037999465072,
"count": 8,
"is_parallel": true,
"self": 0.002079037999465072
}
}
},
"UnityEnvironment.step": {
"total": 0.049345730999903026,
"count": 1,
"is_parallel": true,
"self": 0.0005944359995737614,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.0005383470002016111,
"count": 1,
"is_parallel": true,
"self": 0.0005383470002016111
},
"communicator.exchange": {
"total": 0.04651662400010537,
"count": 1,
"is_parallel": true,
"self": 0.04651662400010537
},
"steps_from_proto": {
"total": 0.0016963240000222868,
"count": 1,
"is_parallel": true,
"self": 0.00035250299970357446,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0013438210003187123,
"count": 8,
"is_parallel": true,
"self": 0.0013438210003187123
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1304.2505042469502,
"count": 63654,
"is_parallel": true,
"self": 36.007895732122506,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 24.981384878814424,
"count": 63654,
"is_parallel": true,
"self": 24.981384878814424
},
"communicator.exchange": {
"total": 1141.5022200689905,
"count": 63654,
"is_parallel": true,
"self": 1141.5022200689905
},
"steps_from_proto": {
"total": 101.75900356702277,
"count": 63654,
"is_parallel": true,
"self": 20.32271977162918,
"children": {
"_process_rank_one_or_two_observation": {
"total": 81.43628379539359,
"count": 509232,
"is_parallel": true,
"self": 81.43628379539359
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 646.1340806549979,
"count": 63655,
"self": 2.699905676285198,
"children": {
"process_trajectory": {
"total": 127.89687590672429,
"count": 63655,
"self": 127.68962765272454,
"children": {
"RLTrainer._checkpoint": {
"total": 0.2072482539997509,
"count": 2,
"self": 0.2072482539997509
}
}
},
"_update_policy": {
"total": 515.5372990719884,
"count": 451,
"self": 303.3176037730791,
"children": {
"TorchPPOOptimizer.update": {
"total": 212.2196952989093,
"count": 22812,
"self": 212.2196952989093
}
}
}
}
}
}
},
"trainer_threads": {
"total": 8.600000001024455e-07,
"count": 1,
"self": 8.600000001024455e-07
},
"TrainerController._save_models": {
"total": 0.08933239100042556,
"count": 1,
"self": 0.001975512000171875,
"children": {
"RLTrainer._checkpoint": {
"total": 0.08735687900025368,
"count": 1,
"self": 0.08735687900025368
}
}
}
}
}
}
}