Cryxim commited on
Commit
b14bd8b
·
verified ·
1 Parent(s): 82071f4

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. README.md +1 -1
  2. replay.mp4 +2 -2
  3. sf_log.txt +126 -0
README.md CHANGED
@@ -15,7 +15,7 @@ model-index:
15
  type: doom_health_gathering_supreme
16
  metrics:
17
  - type: mean_reward
18
- value: 8.39 +/- 3.50
19
  name: mean_reward
20
  verified: false
21
  ---
 
15
  type: doom_health_gathering_supreme
16
  metrics:
17
  - type: mean_reward
18
+ value: 8.30 +/- 2.96
19
  name: mean_reward
20
  verified: false
21
  ---
replay.mp4 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4f68904b2fce96562cf5912466be16f4b942d968d6a3696abaffed40ed05fea7
3
- size 15899472
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36c67ce101e64f1dd522e78334ce65d5b847a7be3e0f5c1d706f113a5b21b989
3
+ size 15749513
sf_log.txt CHANGED
@@ -1069,3 +1069,129 @@ main_loop: 1144.5452
1069
  [2024-08-31 18:04:08,906][00204] Avg episode rewards: #0: 17.495, true rewards: #0: 8.395
1070
  [2024-08-31 18:04:08,908][00204] Avg episode reward: 17.495, avg true_objective: 8.395
1071
  [2024-08-31 18:04:55,697][00204] Replay video saved to /content/train_dir/default_experiment/replay.mp4!
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1069
  [2024-08-31 18:04:08,906][00204] Avg episode rewards: #0: 17.495, true rewards: #0: 8.395
1070
  [2024-08-31 18:04:08,908][00204] Avg episode reward: 17.495, avg true_objective: 8.395
1071
  [2024-08-31 18:04:55,697][00204] Replay video saved to /content/train_dir/default_experiment/replay.mp4!
1072
+ [2024-08-31 18:05:02,002][00204] The model has been pushed to https://huggingface.co/Cryxim/rl_course_vizdoom_health_gathering_supreme
1073
+ [2024-08-31 18:07:14,466][00204] Loading existing experiment configuration from /content/train_dir/default_experiment/config.json
1074
+ [2024-08-31 18:07:14,468][00204] Overriding arg 'num_workers' with value 1 passed from command line
1075
+ [2024-08-31 18:07:14,470][00204] Adding new argument 'no_render'=True that is not in the saved config file!
1076
+ [2024-08-31 18:07:14,472][00204] Adding new argument 'save_video'=True that is not in the saved config file!
1077
+ [2024-08-31 18:07:14,474][00204] Adding new argument 'video_frames'=1000000000.0 that is not in the saved config file!
1078
+ [2024-08-31 18:07:14,475][00204] Adding new argument 'video_name'=None that is not in the saved config file!
1079
+ [2024-08-31 18:07:14,477][00204] Adding new argument 'max_num_frames'=100000 that is not in the saved config file!
1080
+ [2024-08-31 18:07:14,478][00204] Adding new argument 'max_num_episodes'=10 that is not in the saved config file!
1081
+ [2024-08-31 18:07:14,479][00204] Adding new argument 'push_to_hub'=True that is not in the saved config file!
1082
+ [2024-08-31 18:07:14,480][00204] Adding new argument 'hf_repository'='Cryxim/rl_course_vizdoom_health_gathering_supreme' that is not in the saved config file!
1083
+ [2024-08-31 18:07:14,481][00204] Adding new argument 'policy_index'=0 that is not in the saved config file!
1084
+ [2024-08-31 18:07:14,482][00204] Adding new argument 'eval_deterministic'=False that is not in the saved config file!
1085
+ [2024-08-31 18:07:14,483][00204] Adding new argument 'train_script'=None that is not in the saved config file!
1086
+ [2024-08-31 18:07:14,484][00204] Adding new argument 'enjoy_script'=None that is not in the saved config file!
1087
+ [2024-08-31 18:07:14,485][00204] Using frameskip 1 and render_action_repeat=4 for evaluation
1088
+ [2024-08-31 18:07:14,494][00204] RunningMeanStd input shape: (3, 72, 128)
1089
+ [2024-08-31 18:07:14,500][00204] RunningMeanStd input shape: (1,)
1090
+ [2024-08-31 18:07:14,513][00204] ConvEncoder: input_channels=3
1091
+ [2024-08-31 18:07:14,548][00204] Conv encoder output size: 512
1092
+ [2024-08-31 18:07:14,549][00204] Policy head output size: 512
1093
+ [2024-08-31 18:07:14,568][00204] Loading state from checkpoint /content/train_dir/default_experiment/checkpoint_p0/checkpoint_000000978_4005888.pth...
1094
+ [2024-08-31 18:07:15,042][00204] Num frames 100...
1095
+ [2024-08-31 18:07:15,157][00204] Num frames 200...
1096
+ [2024-08-31 18:07:15,270][00204] Num frames 300...
1097
+ [2024-08-31 18:07:15,385][00204] Num frames 400...
1098
+ [2024-08-31 18:07:15,496][00204] Num frames 500...
1099
+ [2024-08-31 18:07:15,638][00204] Avg episode rewards: #0: 8.760, true rewards: #0: 5.760
1100
+ [2024-08-31 18:07:15,640][00204] Avg episode reward: 8.760, avg true_objective: 5.760
1101
+ [2024-08-31 18:07:15,676][00204] Num frames 600...
1102
+ [2024-08-31 18:07:15,793][00204] Num frames 700...
1103
+ [2024-08-31 18:07:15,920][00204] Num frames 800...
1104
+ [2024-08-31 18:07:16,037][00204] Num frames 900...
1105
+ [2024-08-31 18:07:16,150][00204] Num frames 1000...
1106
+ [2024-08-31 18:07:16,265][00204] Num frames 1100...
1107
+ [2024-08-31 18:07:16,377][00204] Num frames 1200...
1108
+ [2024-08-31 18:07:16,497][00204] Num frames 1300...
1109
+ [2024-08-31 18:07:16,612][00204] Num frames 1400...
1110
+ [2024-08-31 18:07:16,755][00204] Num frames 1500...
1111
+ [2024-08-31 18:07:16,894][00204] Num frames 1600...
1112
+ [2024-08-31 18:07:16,986][00204] Avg episode rewards: #0: 14.160, true rewards: #0: 8.160
1113
+ [2024-08-31 18:07:16,988][00204] Avg episode reward: 14.160, avg true_objective: 8.160
1114
+ [2024-08-31 18:07:17,109][00204] Num frames 1700...
1115
+ [2024-08-31 18:07:17,268][00204] Num frames 1800...
1116
+ [2024-08-31 18:07:17,425][00204] Num frames 1900...
1117
+ [2024-08-31 18:07:17,576][00204] Num frames 2000...
1118
+ [2024-08-31 18:07:17,734][00204] Num frames 2100...
1119
+ [2024-08-31 18:07:17,908][00204] Avg episode rewards: #0: 11.920, true rewards: #0: 7.253
1120
+ [2024-08-31 18:07:17,910][00204] Avg episode reward: 11.920, avg true_objective: 7.253
1121
+ [2024-08-31 18:07:17,951][00204] Num frames 2200...
1122
+ [2024-08-31 18:07:18,102][00204] Num frames 2300...
1123
+ [2024-08-31 18:07:18,266][00204] Num frames 2400...
1124
+ [2024-08-31 18:07:18,431][00204] Num frames 2500...
1125
+ [2024-08-31 18:07:18,595][00204] Num frames 2600...
1126
+ [2024-08-31 18:07:18,794][00204] Num frames 2700...
1127
+ [2024-08-31 18:07:18,991][00204] Avg episode rewards: #0: 11.210, true rewards: #0: 6.960
1128
+ [2024-08-31 18:07:18,994][00204] Avg episode reward: 11.210, avg true_objective: 6.960
1129
+ [2024-08-31 18:07:19,024][00204] Num frames 2800...
1130
+ [2024-08-31 18:07:19,187][00204] Num frames 2900...
1131
+ [2024-08-31 18:07:19,350][00204] Num frames 3000...
1132
+ [2024-08-31 18:07:19,521][00204] Num frames 3100...
1133
+ [2024-08-31 18:07:19,689][00204] Num frames 3200...
1134
+ [2024-08-31 18:07:19,819][00204] Num frames 3300...
1135
+ [2024-08-31 18:07:19,940][00204] Num frames 3400...
1136
+ [2024-08-31 18:07:20,051][00204] Avg episode rewards: #0: 11.694, true rewards: #0: 6.894
1137
+ [2024-08-31 18:07:20,053][00204] Avg episode reward: 11.694, avg true_objective: 6.894
1138
+ [2024-08-31 18:07:20,117][00204] Num frames 3500...
1139
+ [2024-08-31 18:07:20,232][00204] Num frames 3600...
1140
+ [2024-08-31 18:07:20,349][00204] Num frames 3700...
1141
+ [2024-08-31 18:07:20,464][00204] Num frames 3800...
1142
+ [2024-08-31 18:07:20,589][00204] Num frames 3900...
1143
+ [2024-08-31 18:07:20,713][00204] Num frames 4000...
1144
+ [2024-08-31 18:07:20,839][00204] Num frames 4100...
1145
+ [2024-08-31 18:07:20,988][00204] Avg episode rewards: #0: 12.290, true rewards: #0: 6.957
1146
+ [2024-08-31 18:07:20,990][00204] Avg episode reward: 12.290, avg true_objective: 6.957
1147
+ [2024-08-31 18:07:21,024][00204] Num frames 4200...
1148
+ [2024-08-31 18:07:21,147][00204] Num frames 4300...
1149
+ [2024-08-31 18:07:21,265][00204] Num frames 4400...
1150
+ [2024-08-31 18:07:21,381][00204] Num frames 4500...
1151
+ [2024-08-31 18:07:21,497][00204] Num frames 4600...
1152
+ [2024-08-31 18:07:21,618][00204] Num frames 4700...
1153
+ [2024-08-31 18:07:21,734][00204] Num frames 4800...
1154
+ [2024-08-31 18:07:21,866][00204] Num frames 4900...
1155
+ [2024-08-31 18:07:21,986][00204] Num frames 5000...
1156
+ [2024-08-31 18:07:22,108][00204] Num frames 5100...
1157
+ [2024-08-31 18:07:22,207][00204] Avg episode rewards: #0: 13.049, true rewards: #0: 7.334
1158
+ [2024-08-31 18:07:22,209][00204] Avg episode reward: 13.049, avg true_objective: 7.334
1159
+ [2024-08-31 18:07:22,292][00204] Num frames 5200...
1160
+ [2024-08-31 18:07:22,410][00204] Num frames 5300...
1161
+ [2024-08-31 18:07:22,525][00204] Num frames 5400...
1162
+ [2024-08-31 18:07:22,648][00204] Num frames 5500...
1163
+ [2024-08-31 18:07:22,777][00204] Num frames 5600...
1164
+ [2024-08-31 18:07:22,913][00204] Num frames 5700...
1165
+ [2024-08-31 18:07:23,071][00204] Avg episode rewards: #0: 13.231, true rewards: #0: 7.231
1166
+ [2024-08-31 18:07:23,072][00204] Avg episode reward: 13.231, avg true_objective: 7.231
1167
+ [2024-08-31 18:07:23,093][00204] Num frames 5800...
1168
+ [2024-08-31 18:07:23,208][00204] Num frames 5900...
1169
+ [2024-08-31 18:07:23,329][00204] Num frames 6000...
1170
+ [2024-08-31 18:07:23,449][00204] Num frames 6100...
1171
+ [2024-08-31 18:07:23,565][00204] Num frames 6200...
1172
+ [2024-08-31 18:07:23,687][00204] Num frames 6300...
1173
+ [2024-08-31 18:07:23,803][00204] Num frames 6400...
1174
+ [2024-08-31 18:07:23,938][00204] Num frames 6500...
1175
+ [2024-08-31 18:07:24,057][00204] Num frames 6600...
1176
+ [2024-08-31 18:07:24,172][00204] Num frames 6700...
1177
+ [2024-08-31 18:07:24,283][00204] Avg episode rewards: #0: 14.161, true rewards: #0: 7.494
1178
+ [2024-08-31 18:07:24,284][00204] Avg episode reward: 14.161, avg true_objective: 7.494
1179
+ [2024-08-31 18:07:24,351][00204] Num frames 6800...
1180
+ [2024-08-31 18:07:24,466][00204] Num frames 6900...
1181
+ [2024-08-31 18:07:24,595][00204] Num frames 7000...
1182
+ [2024-08-31 18:07:24,711][00204] Num frames 7100...
1183
+ [2024-08-31 18:07:24,836][00204] Num frames 7200...
1184
+ [2024-08-31 18:07:24,963][00204] Num frames 7300...
1185
+ [2024-08-31 18:07:25,082][00204] Num frames 7400...
1186
+ [2024-08-31 18:07:25,202][00204] Num frames 7500...
1187
+ [2024-08-31 18:07:25,321][00204] Num frames 7600...
1188
+ [2024-08-31 18:07:25,440][00204] Num frames 7700...
1189
+ [2024-08-31 18:07:25,556][00204] Num frames 7800...
1190
+ [2024-08-31 18:07:25,672][00204] Num frames 7900...
1191
+ [2024-08-31 18:07:25,797][00204] Num frames 8000...
1192
+ [2024-08-31 18:07:25,920][00204] Num frames 8100...
1193
+ [2024-08-31 18:07:26,050][00204] Num frames 8200...
1194
+ [2024-08-31 18:07:26,168][00204] Num frames 8300...
1195
+ [2024-08-31 18:07:26,227][00204] Avg episode rewards: #0: 17.001, true rewards: #0: 8.301
1196
+ [2024-08-31 18:07:26,228][00204] Avg episode reward: 17.001, avg true_objective: 8.301
1197
+ [2024-08-31 18:08:13,048][00204] Replay video saved to /content/train_dir/default_experiment/replay.mp4!