ppo-SnowballTarget / run_logs /training_status.json
siemr's picture
First Push
4b24791 verified
raw
history blame
2.2 kB
{
"SnowballTarget": {
"checkpoints": [
{
"steps": 74992,
"file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-74992.onnx",
"reward": 13.545454545454545,
"creation_time": 1711919548.9971046,
"auxillary_file_paths": [
"results/SnowballTarget1/SnowballTarget/SnowballTarget-74992.pt"
]
},
{
"steps": 149984,
"file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.onnx",
"reward": 21.363636363636363,
"creation_time": 1711919679.771043,
"auxillary_file_paths": [
"results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.pt"
]
},
{
"steps": 224976,
"file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-224976.onnx",
"reward": 25.454545454545453,
"creation_time": 1711919810.9059863,
"auxillary_file_paths": [
"results/SnowballTarget1/SnowballTarget/SnowballTarget-224976.pt"
]
},
{
"steps": 250008,
"file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-250008.onnx",
"reward": 23.681818181818183,
"creation_time": 1711919853.6813004,
"auxillary_file_paths": [
"results/SnowballTarget1/SnowballTarget/SnowballTarget-250008.pt"
]
}
],
"final_checkpoint": {
"steps": 250008,
"file_path": "results/SnowballTarget1/SnowballTarget.onnx",
"reward": 23.681818181818183,
"creation_time": 1711919853.6813004,
"auxillary_file_paths": [
"results/SnowballTarget1/SnowballTarget/SnowballTarget-250008.pt"
]
}
},
"metadata": {
"stats_format_version": "0.3.0",
"mlagents_version": "1.1.0.dev0",
"torch_version": "2.2.1+cu121"
}
}