satcos commited on
Commit
c6cf9a2
1 Parent(s): 166e95b

My trained model

Browse files
PPO_model_v1.zip ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9ef3df4a3722a1888b529f7bf7fe1e3e6a39ea38a83914b6df6bdaa7e9a813b
3
+ size 29447806
PPO_model_v1/_stable_baselines3_version ADDED
@@ -0,0 +1 @@
 
 
1
+ 2.3.2
PPO_model_v1/data ADDED
The diff for this file is too large to render. See raw diff
 
PPO_model_v1/policy.optimizer.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c00f66e642f5c8dbe88596a829504d05729dd7ecdaa5237d82dfcd6d878716b
3
+ size 19518205
PPO_model_v1/policy.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce988d2d967e0cebb8f2b7728388214a34630d8faa0198abdb2a04fd725b6248
3
+ size 9761059
PPO_model_v1/pytorch_variables.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ebdad4b9cfe9cd22a3abadb5623bf7bb1f6eb2e408740245eb3f2044b0adc018
3
+ size 864
PPO_model_v1/system_info.txt ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ - OS: macOS-14.6.1-arm64-i386-64bit Darwin Kernel Version 23.6.0: Mon Jul 29 21:14:30 PDT 2024; root:xnu-10063.141.2~1/RELEASE_ARM64_T6000
2
+ - Python: 3.11.9
3
+ - Stable-Baselines3: 2.3.2
4
+ - PyTorch: 2.3.1
5
+ - GPU Enabled: False
6
+ - Numpy: 1.26.4
7
+ - Cloudpickle: 3.0.0
8
+ - Gymnasium: 0.29.1
9
+ - OpenAI Gym: 0.17.2
README.md ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: stable-baselines3
3
+ tags:
4
+ - CarRacing-v2
5
+ - deep-reinforcement-learning
6
+ - reinforcement-learning
7
+ - stable-baselines3
8
+ model-index:
9
+ - name: PPO
10
+ results:
11
+ - task:
12
+ type: reinforcement-learning
13
+ name: reinforcement-learning
14
+ dataset:
15
+ name: CarRacing-v2
16
+ type: CarRacing-v2
17
+ metrics:
18
+ - type: mean_reward
19
+ value: 850.01 +/- 137.65
20
+ name: mean_reward
21
+ verified: false
22
+ ---
23
+
24
+ # **PPO** Agent playing **CarRacing-v2**
25
+ This is a trained model of a **PPO** agent playing **CarRacing-v2**
26
+ using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3).
27
+
28
+ ## Usage (with Stable-baselines3)
29
+ TODO: Add your code
30
+
31
+
32
+ ```python
33
+ from stable_baselines3 import ...
34
+ from huggingface_sb3 import load_from_hub
35
+
36
+ ...
37
+ ```
config.json ADDED
The diff for this file is too large to render. See raw diff
 
results.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"mean_reward": 850.0146951, "std_reward": 137.65236984414358, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2024-09-19T13:37:19.028795"}