Upload . with huggingface_hub
Browse files- README.md +1 -1
- replay.mp4 +2 -2
- sf_log.txt +282 -0
README.md
CHANGED
@@ -15,7 +15,7 @@ model-index:
|
|
15 |
type: doom_health_gathering_supreme
|
16 |
metrics:
|
17 |
- type: mean_reward
|
18 |
-
value:
|
19 |
name: mean_reward
|
20 |
verified: false
|
21 |
---
|
|
|
15 |
type: doom_health_gathering_supreme
|
16 |
metrics:
|
17 |
- type: mean_reward
|
18 |
+
value: 9.48 +/- 4.69
|
19 |
name: mean_reward
|
20 |
verified: false
|
21 |
---
|
replay.mp4
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1b1808132fd6f5e6864ab3695bf2ec741b1bb2be819e19c7e32c48a86b1d0593
|
3 |
+
size 17548925
|
sf_log.txt
CHANGED
@@ -1283,3 +1283,285 @@ main_loop: 1180.8338
|
|
1283 |
[2023-03-11 11:45:53,574][00127] Avg episode rewards: #0: 17.000, true rewards: #0: 8.100
|
1284 |
[2023-03-11 11:45:53,575][00127] Avg episode reward: 17.000, avg true_objective: 8.100
|
1285 |
[2023-03-11 11:46:41,540][00127] Replay video saved to /content/train_dir/default_experiment/replay.mp4!
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1283 |
[2023-03-11 11:45:53,574][00127] Avg episode rewards: #0: 17.000, true rewards: #0: 8.100
|
1284 |
[2023-03-11 11:45:53,575][00127] Avg episode reward: 17.000, avg true_objective: 8.100
|
1285 |
[2023-03-11 11:46:41,540][00127] Replay video saved to /content/train_dir/default_experiment/replay.mp4!
|
1286 |
+
[2023-03-11 11:46:53,587][00127] The model has been pushed to https://huggingface.co/Taratata/rl_course_vizdoom_health_gathering_supreme
|
1287 |
+
[2023-03-11 11:48:24,968][00127] Loading existing experiment configuration from /content/train_dir/default_experiment/config.json
|
1288 |
+
[2023-03-11 11:48:24,970][00127] Overriding arg 'num_workers' with value 1 passed from command line
|
1289 |
+
[2023-03-11 11:48:24,972][00127] Adding new argument 'no_render'=True that is not in the saved config file!
|
1290 |
+
[2023-03-11 11:48:24,974][00127] Adding new argument 'save_video'=True that is not in the saved config file!
|
1291 |
+
[2023-03-11 11:48:24,976][00127] Adding new argument 'video_frames'=1000000000.0 that is not in the saved config file!
|
1292 |
+
[2023-03-11 11:48:24,977][00127] Adding new argument 'video_name'=None that is not in the saved config file!
|
1293 |
+
[2023-03-11 11:48:24,979][00127] Adding new argument 'max_num_frames'=1000000000.0 that is not in the saved config file!
|
1294 |
+
[2023-03-11 11:48:24,981][00127] Adding new argument 'max_num_episodes'=10 that is not in the saved config file!
|
1295 |
+
[2023-03-11 11:48:24,982][00127] Adding new argument 'push_to_hub'=False that is not in the saved config file!
|
1296 |
+
[2023-03-11 11:48:24,983][00127] Adding new argument 'hf_repository'=None that is not in the saved config file!
|
1297 |
+
[2023-03-11 11:48:24,984][00127] Adding new argument 'policy_index'=0 that is not in the saved config file!
|
1298 |
+
[2023-03-11 11:48:24,985][00127] Adding new argument 'eval_deterministic'=False that is not in the saved config file!
|
1299 |
+
[2023-03-11 11:48:24,986][00127] Adding new argument 'train_script'=None that is not in the saved config file!
|
1300 |
+
[2023-03-11 11:48:24,987][00127] Adding new argument 'enjoy_script'=None that is not in the saved config file!
|
1301 |
+
[2023-03-11 11:48:24,988][00127] Using frameskip 1 and render_action_repeat=4 for evaluation
|
1302 |
+
[2023-03-11 11:48:25,005][00127] RunningMeanStd input shape: (3, 72, 128)
|
1303 |
+
[2023-03-11 11:48:25,008][00127] RunningMeanStd input shape: (1,)
|
1304 |
+
[2023-03-11 11:48:25,022][00127] ConvEncoder: input_channels=3
|
1305 |
+
[2023-03-11 11:48:25,062][00127] Conv encoder output size: 512
|
1306 |
+
[2023-03-11 11:48:25,063][00127] Policy head output size: 512
|
1307 |
+
[2023-03-11 11:48:25,082][00127] Loading state from checkpoint /content/train_dir/default_experiment/checkpoint_p0/checkpoint_000000978_4005888.pth...
|
1308 |
+
[2023-03-11 11:48:25,531][00127] Num frames 100...
|
1309 |
+
[2023-03-11 11:48:25,643][00127] Num frames 200...
|
1310 |
+
[2023-03-11 11:48:25,755][00127] Num frames 300...
|
1311 |
+
[2023-03-11 11:48:25,864][00127] Num frames 400...
|
1312 |
+
[2023-03-11 11:48:25,984][00127] Num frames 500...
|
1313 |
+
[2023-03-11 11:48:26,098][00127] Num frames 600...
|
1314 |
+
[2023-03-11 11:48:26,204][00127] Num frames 700...
|
1315 |
+
[2023-03-11 11:48:26,316][00127] Num frames 800...
|
1316 |
+
[2023-03-11 11:48:26,435][00127] Num frames 900...
|
1317 |
+
[2023-03-11 11:48:26,552][00127] Num frames 1000...
|
1318 |
+
[2023-03-11 11:48:26,668][00127] Avg episode rewards: #0: 22.550, true rewards: #0: 10.550
|
1319 |
+
[2023-03-11 11:48:26,671][00127] Avg episode reward: 22.550, avg true_objective: 10.550
|
1320 |
+
[2023-03-11 11:48:26,724][00127] Num frames 1100...
|
1321 |
+
[2023-03-11 11:48:26,841][00127] Num frames 1200...
|
1322 |
+
[2023-03-11 11:48:26,952][00127] Num frames 1300...
|
1323 |
+
[2023-03-11 11:48:27,062][00127] Num frames 1400...
|
1324 |
+
[2023-03-11 11:48:27,171][00127] Num frames 1500...
|
1325 |
+
[2023-03-11 11:48:27,288][00127] Num frames 1600...
|
1326 |
+
[2023-03-11 11:48:27,410][00127] Num frames 1700...
|
1327 |
+
[2023-03-11 11:48:27,525][00127] Num frames 1800...
|
1328 |
+
[2023-03-11 11:48:27,645][00127] Num frames 1900...
|
1329 |
+
[2023-03-11 11:48:27,754][00127] Num frames 2000...
|
1330 |
+
[2023-03-11 11:48:27,864][00127] Num frames 2100...
|
1331 |
+
[2023-03-11 11:48:27,986][00127] Num frames 2200...
|
1332 |
+
[2023-03-11 11:48:28,109][00127] Num frames 2300...
|
1333 |
+
[2023-03-11 11:48:28,220][00127] Avg episode rewards: #0: 27.245, true rewards: #0: 11.745
|
1334 |
+
[2023-03-11 11:48:28,222][00127] Avg episode reward: 27.245, avg true_objective: 11.745
|
1335 |
+
[2023-03-11 11:48:28,284][00127] Num frames 2400...
|
1336 |
+
[2023-03-11 11:48:28,395][00127] Num frames 2500...
|
1337 |
+
[2023-03-11 11:48:28,507][00127] Num frames 2600...
|
1338 |
+
[2023-03-11 11:48:28,624][00127] Num frames 2700...
|
1339 |
+
[2023-03-11 11:48:28,747][00127] Num frames 2800...
|
1340 |
+
[2023-03-11 11:48:28,858][00127] Num frames 2900...
|
1341 |
+
[2023-03-11 11:48:28,968][00127] Num frames 3000...
|
1342 |
+
[2023-03-11 11:48:29,074][00127] Num frames 3100...
|
1343 |
+
[2023-03-11 11:48:29,186][00127] Num frames 3200...
|
1344 |
+
[2023-03-11 11:48:29,301][00127] Num frames 3300...
|
1345 |
+
[2023-03-11 11:48:29,425][00127] Num frames 3400...
|
1346 |
+
[2023-03-11 11:48:29,535][00127] Num frames 3500...
|
1347 |
+
[2023-03-11 11:48:29,652][00127] Num frames 3600...
|
1348 |
+
[2023-03-11 11:48:29,764][00127] Num frames 3700...
|
1349 |
+
[2023-03-11 11:48:29,879][00127] Num frames 3800...
|
1350 |
+
[2023-03-11 11:48:29,989][00127] Num frames 3900...
|
1351 |
+
[2023-03-11 11:48:30,118][00127] Num frames 4000...
|
1352 |
+
[2023-03-11 11:48:30,231][00127] Num frames 4100...
|
1353 |
+
[2023-03-11 11:48:30,340][00127] Num frames 4200...
|
1354 |
+
[2023-03-11 11:48:30,472][00127] Avg episode rewards: #0: 35.563, true rewards: #0: 14.230
|
1355 |
+
[2023-03-11 11:48:30,474][00127] Avg episode reward: 35.563, avg true_objective: 14.230
|
1356 |
+
[2023-03-11 11:48:30,517][00127] Num frames 4300...
|
1357 |
+
[2023-03-11 11:48:30,634][00127] Num frames 4400...
|
1358 |
+
[2023-03-11 11:48:30,753][00127] Num frames 4500...
|
1359 |
+
[2023-03-11 11:48:30,864][00127] Num frames 4600...
|
1360 |
+
[2023-03-11 11:48:30,978][00127] Num frames 4700...
|
1361 |
+
[2023-03-11 11:48:31,088][00127] Num frames 4800...
|
1362 |
+
[2023-03-11 11:48:31,199][00127] Num frames 4900...
|
1363 |
+
[2023-03-11 11:48:31,310][00127] Num frames 5000...
|
1364 |
+
[2023-03-11 11:48:31,408][00127] Avg episode rewards: #0: 30.592, true rewards: #0: 12.592
|
1365 |
+
[2023-03-11 11:48:31,410][00127] Avg episode reward: 30.592, avg true_objective: 12.592
|
1366 |
+
[2023-03-11 11:48:31,481][00127] Num frames 5100...
|
1367 |
+
[2023-03-11 11:48:31,590][00127] Num frames 5200...
|
1368 |
+
[2023-03-11 11:48:31,710][00127] Num frames 5300...
|
1369 |
+
[2023-03-11 11:48:31,822][00127] Num frames 5400...
|
1370 |
+
[2023-03-11 11:48:31,931][00127] Num frames 5500...
|
1371 |
+
[2023-03-11 11:48:32,062][00127] Num frames 5600...
|
1372 |
+
[2023-03-11 11:48:32,169][00127] Num frames 5700...
|
1373 |
+
[2023-03-11 11:48:32,284][00127] Num frames 5800...
|
1374 |
+
[2023-03-11 11:48:32,398][00127] Num frames 5900...
|
1375 |
+
[2023-03-11 11:48:32,513][00127] Num frames 6000...
|
1376 |
+
[2023-03-11 11:48:32,621][00127] Num frames 6100...
|
1377 |
+
[2023-03-11 11:48:32,743][00127] Num frames 6200...
|
1378 |
+
[2023-03-11 11:48:32,862][00127] Num frames 6300...
|
1379 |
+
[2023-03-11 11:48:32,959][00127] Avg episode rewards: #0: 30.476, true rewards: #0: 12.676
|
1380 |
+
[2023-03-11 11:48:32,960][00127] Avg episode reward: 30.476, avg true_objective: 12.676
|
1381 |
+
[2023-03-11 11:48:33,029][00127] Num frames 6400...
|
1382 |
+
[2023-03-11 11:48:33,138][00127] Num frames 6500...
|
1383 |
+
[2023-03-11 11:48:33,295][00127] Num frames 6600...
|
1384 |
+
[2023-03-11 11:48:33,485][00127] Avg episode rewards: #0: 26.150, true rewards: #0: 11.150
|
1385 |
+
[2023-03-11 11:48:33,487][00127] Avg episode reward: 26.150, avg true_objective: 11.150
|
1386 |
+
[2023-03-11 11:48:33,516][00127] Num frames 6700...
|
1387 |
+
[2023-03-11 11:48:33,686][00127] Num frames 6800...
|
1388 |
+
[2023-03-11 11:48:33,842][00127] Num frames 6900...
|
1389 |
+
[2023-03-11 11:48:33,997][00127] Num frames 7000...
|
1390 |
+
[2023-03-11 11:48:34,151][00127] Num frames 7100...
|
1391 |
+
[2023-03-11 11:48:34,303][00127] Num frames 7200...
|
1392 |
+
[2023-03-11 11:48:34,460][00127] Num frames 7300...
|
1393 |
+
[2023-03-11 11:48:34,612][00127] Num frames 7400...
|
1394 |
+
[2023-03-11 11:48:34,786][00127] Num frames 7500...
|
1395 |
+
[2023-03-11 11:48:34,936][00127] Num frames 7600...
|
1396 |
+
[2023-03-11 11:48:35,092][00127] Num frames 7700...
|
1397 |
+
[2023-03-11 11:48:35,266][00127] Avg episode rewards: #0: 26.254, true rewards: #0: 11.111
|
1398 |
+
[2023-03-11 11:48:35,268][00127] Avg episode reward: 26.254, avg true_objective: 11.111
|
1399 |
+
[2023-03-11 11:48:35,311][00127] Num frames 7800...
|
1400 |
+
[2023-03-11 11:48:35,470][00127] Num frames 7900...
|
1401 |
+
[2023-03-11 11:48:35,629][00127] Num frames 8000...
|
1402 |
+
[2023-03-11 11:48:35,788][00127] Num frames 8100...
|
1403 |
+
[2023-03-11 11:48:35,899][00127] Avg episode rewards: #0: 23.542, true rewards: #0: 10.167
|
1404 |
+
[2023-03-11 11:48:35,901][00127] Avg episode reward: 23.542, avg true_objective: 10.167
|
1405 |
+
[2023-03-11 11:48:36,004][00127] Num frames 8200...
|
1406 |
+
[2023-03-11 11:48:36,157][00127] Num frames 8300...
|
1407 |
+
[2023-03-11 11:48:36,312][00127] Num frames 8400...
|
1408 |
+
[2023-03-11 11:48:36,466][00127] Num frames 8500...
|
1409 |
+
[2023-03-11 11:48:36,632][00127] Num frames 8600...
|
1410 |
+
[2023-03-11 11:48:36,761][00127] Num frames 8700...
|
1411 |
+
[2023-03-11 11:48:36,879][00127] Num frames 8800...
|
1412 |
+
[2023-03-11 11:48:36,997][00127] Num frames 8900...
|
1413 |
+
[2023-03-11 11:48:37,115][00127] Num frames 9000...
|
1414 |
+
[2023-03-11 11:48:37,223][00127] Num frames 9100...
|
1415 |
+
[2023-03-11 11:48:37,335][00127] Num frames 9200...
|
1416 |
+
[2023-03-11 11:48:37,492][00127] Avg episode rewards: #0: 23.762, true rewards: #0: 10.318
|
1417 |
+
[2023-03-11 11:48:37,494][00127] Avg episode reward: 23.762, avg true_objective: 10.318
|
1418 |
+
[2023-03-11 11:48:37,514][00127] Num frames 9300...
|
1419 |
+
[2023-03-11 11:48:37,624][00127] Num frames 9400...
|
1420 |
+
[2023-03-11 11:48:37,732][00127] Num frames 9500...
|
1421 |
+
[2023-03-11 11:48:37,849][00127] Num frames 9600...
|
1422 |
+
[2023-03-11 11:48:37,962][00127] Num frames 9700...
|
1423 |
+
[2023-03-11 11:48:38,073][00127] Num frames 9800...
|
1424 |
+
[2023-03-11 11:48:38,183][00127] Num frames 9900...
|
1425 |
+
[2023-03-11 11:48:38,301][00127] Num frames 10000...
|
1426 |
+
[2023-03-11 11:48:38,411][00127] Num frames 10100...
|
1427 |
+
[2023-03-11 11:48:38,520][00127] Num frames 10200...
|
1428 |
+
[2023-03-11 11:48:38,627][00127] Num frames 10300...
|
1429 |
+
[2023-03-11 11:48:38,732][00127] Avg episode rewards: #0: 23.841, true rewards: #0: 10.341
|
1430 |
+
[2023-03-11 11:48:38,733][00127] Avg episode reward: 23.841, avg true_objective: 10.341
|
1431 |
+
[2023-03-11 11:49:39,627][00127] Replay video saved to /content/train_dir/default_experiment/replay.mp4!
|
1432 |
+
[2023-03-11 11:50:02,077][00127] Loading existing experiment configuration from /content/train_dir/default_experiment/config.json
|
1433 |
+
[2023-03-11 11:50:02,079][00127] Overriding arg 'num_workers' with value 1 passed from command line
|
1434 |
+
[2023-03-11 11:50:02,081][00127] Adding new argument 'no_render'=True that is not in the saved config file!
|
1435 |
+
[2023-03-11 11:50:02,083][00127] Adding new argument 'save_video'=True that is not in the saved config file!
|
1436 |
+
[2023-03-11 11:50:02,085][00127] Adding new argument 'video_frames'=1000000000.0 that is not in the saved config file!
|
1437 |
+
[2023-03-11 11:50:02,087][00127] Adding new argument 'video_name'=None that is not in the saved config file!
|
1438 |
+
[2023-03-11 11:50:02,088][00127] Adding new argument 'max_num_frames'=100000 that is not in the saved config file!
|
1439 |
+
[2023-03-11 11:50:02,090][00127] Adding new argument 'max_num_episodes'=10 that is not in the saved config file!
|
1440 |
+
[2023-03-11 11:50:02,091][00127] Adding new argument 'push_to_hub'=True that is not in the saved config file!
|
1441 |
+
[2023-03-11 11:50:02,092][00127] Adding new argument 'hf_repository'='Taratata/rl_course_vizdoom_health_gathering_supreme' that is not in the saved config file!
|
1442 |
+
[2023-03-11 11:50:02,093][00127] Adding new argument 'policy_index'=0 that is not in the saved config file!
|
1443 |
+
[2023-03-11 11:50:02,094][00127] Adding new argument 'eval_deterministic'=False that is not in the saved config file!
|
1444 |
+
[2023-03-11 11:50:02,095][00127] Adding new argument 'train_script'=None that is not in the saved config file!
|
1445 |
+
[2023-03-11 11:50:02,096][00127] Adding new argument 'enjoy_script'=None that is not in the saved config file!
|
1446 |
+
[2023-03-11 11:50:02,097][00127] Using frameskip 1 and render_action_repeat=4 for evaluation
|
1447 |
+
[2023-03-11 11:50:02,117][00127] RunningMeanStd input shape: (3, 72, 128)
|
1448 |
+
[2023-03-11 11:50:02,120][00127] RunningMeanStd input shape: (1,)
|
1449 |
+
[2023-03-11 11:50:02,133][00127] ConvEncoder: input_channels=3
|
1450 |
+
[2023-03-11 11:50:02,168][00127] Conv encoder output size: 512
|
1451 |
+
[2023-03-11 11:50:02,172][00127] Policy head output size: 512
|
1452 |
+
[2023-03-11 11:50:02,192][00127] Loading state from checkpoint /content/train_dir/default_experiment/checkpoint_p0/checkpoint_000000978_4005888.pth...
|
1453 |
+
[2023-03-11 11:50:02,628][00127] Num frames 100...
|
1454 |
+
[2023-03-11 11:50:02,748][00127] Num frames 200...
|
1455 |
+
[2023-03-11 11:50:02,868][00127] Num frames 300...
|
1456 |
+
[2023-03-11 11:50:02,990][00127] Num frames 400...
|
1457 |
+
[2023-03-11 11:50:03,108][00127] Num frames 500...
|
1458 |
+
[2023-03-11 11:50:03,227][00127] Num frames 600...
|
1459 |
+
[2023-03-11 11:50:03,340][00127] Num frames 700...
|
1460 |
+
[2023-03-11 11:50:03,459][00127] Num frames 800...
|
1461 |
+
[2023-03-11 11:50:03,580][00127] Num frames 900...
|
1462 |
+
[2023-03-11 11:50:03,696][00127] Num frames 1000...
|
1463 |
+
[2023-03-11 11:50:03,768][00127] Avg episode rewards: #0: 22.080, true rewards: #0: 10.080
|
1464 |
+
[2023-03-11 11:50:03,770][00127] Avg episode reward: 22.080, avg true_objective: 10.080
|
1465 |
+
[2023-03-11 11:50:03,875][00127] Num frames 1100...
|
1466 |
+
[2023-03-11 11:50:03,999][00127] Num frames 1200...
|
1467 |
+
[2023-03-11 11:50:04,108][00127] Num frames 1300...
|
1468 |
+
[2023-03-11 11:50:04,224][00127] Num frames 1400...
|
1469 |
+
[2023-03-11 11:50:04,339][00127] Num frames 1500...
|
1470 |
+
[2023-03-11 11:50:04,450][00127] Num frames 1600...
|
1471 |
+
[2023-03-11 11:50:04,566][00127] Num frames 1700...
|
1472 |
+
[2023-03-11 11:50:04,685][00127] Num frames 1800...
|
1473 |
+
[2023-03-11 11:50:04,801][00127] Num frames 1900...
|
1474 |
+
[2023-03-11 11:50:04,909][00127] Num frames 2000...
|
1475 |
+
[2023-03-11 11:50:05,022][00127] Num frames 2100...
|
1476 |
+
[2023-03-11 11:50:05,132][00127] Num frames 2200...
|
1477 |
+
[2023-03-11 11:50:05,247][00127] Num frames 2300...
|
1478 |
+
[2023-03-11 11:50:05,357][00127] Num frames 2400...
|
1479 |
+
[2023-03-11 11:50:05,469][00127] Num frames 2500...
|
1480 |
+
[2023-03-11 11:50:05,580][00127] Num frames 2600...
|
1481 |
+
[2023-03-11 11:50:05,695][00127] Num frames 2700...
|
1482 |
+
[2023-03-11 11:50:05,807][00127] Num frames 2800...
|
1483 |
+
[2023-03-11 11:50:05,859][00127] Avg episode rewards: #0: 32.500, true rewards: #0: 14.000
|
1484 |
+
[2023-03-11 11:50:05,861][00127] Avg episode reward: 32.500, avg true_objective: 14.000
|
1485 |
+
[2023-03-11 11:50:05,977][00127] Num frames 2900...
|
1486 |
+
[2023-03-11 11:50:06,085][00127] Num frames 3000...
|
1487 |
+
[2023-03-11 11:50:06,194][00127] Num frames 3100...
|
1488 |
+
[2023-03-11 11:50:06,311][00127] Num frames 3200...
|
1489 |
+
[2023-03-11 11:50:06,423][00127] Num frames 3300...
|
1490 |
+
[2023-03-11 11:50:06,531][00127] Num frames 3400...
|
1491 |
+
[2023-03-11 11:50:06,661][00127] Num frames 3500...
|
1492 |
+
[2023-03-11 11:50:06,776][00127] Num frames 3600...
|
1493 |
+
[2023-03-11 11:50:06,828][00127] Avg episode rewards: #0: 27.333, true rewards: #0: 12.000
|
1494 |
+
[2023-03-11 11:50:06,830][00127] Avg episode reward: 27.333, avg true_objective: 12.000
|
1495 |
+
[2023-03-11 11:50:06,941][00127] Num frames 3700...
|
1496 |
+
[2023-03-11 11:50:07,058][00127] Num frames 3800...
|
1497 |
+
[2023-03-11 11:50:07,163][00127] Num frames 3900...
|
1498 |
+
[2023-03-11 11:50:07,270][00127] Num frames 4000...
|
1499 |
+
[2023-03-11 11:50:07,381][00127] Avg episode rewards: #0: 21.870, true rewards: #0: 10.120
|
1500 |
+
[2023-03-11 11:50:07,383][00127] Avg episode reward: 21.870, avg true_objective: 10.120
|
1501 |
+
[2023-03-11 11:50:07,452][00127] Num frames 4100...
|
1502 |
+
[2023-03-11 11:50:07,565][00127] Num frames 4200...
|
1503 |
+
[2023-03-11 11:50:07,679][00127] Num frames 4300...
|
1504 |
+
[2023-03-11 11:50:07,790][00127] Num frames 4400...
|
1505 |
+
[2023-03-11 11:50:07,908][00127] Num frames 4500...
|
1506 |
+
[2023-03-11 11:50:08,029][00127] Num frames 4600...
|
1507 |
+
[2023-03-11 11:50:08,141][00127] Num frames 4700...
|
1508 |
+
[2023-03-11 11:50:08,257][00127] Num frames 4800...
|
1509 |
+
[2023-03-11 11:50:08,378][00127] Num frames 4900...
|
1510 |
+
[2023-03-11 11:50:08,483][00127] Avg episode rewards: #0: 21.488, true rewards: #0: 9.888
|
1511 |
+
[2023-03-11 11:50:08,484][00127] Avg episode reward: 21.488, avg true_objective: 9.888
|
1512 |
+
[2023-03-11 11:50:08,553][00127] Num frames 5000...
|
1513 |
+
[2023-03-11 11:50:08,665][00127] Num frames 5100...
|
1514 |
+
[2023-03-11 11:50:08,776][00127] Num frames 5200...
|
1515 |
+
[2023-03-11 11:50:08,905][00127] Num frames 5300...
|
1516 |
+
[2023-03-11 11:50:09,033][00127] Num frames 5400...
|
1517 |
+
[2023-03-11 11:50:09,144][00127] Num frames 5500...
|
1518 |
+
[2023-03-11 11:50:09,255][00127] Num frames 5600...
|
1519 |
+
[2023-03-11 11:50:09,398][00127] Avg episode rewards: #0: 20.462, true rewards: #0: 9.462
|
1520 |
+
[2023-03-11 11:50:09,400][00127] Avg episode reward: 20.462, avg true_objective: 9.462
|
1521 |
+
[2023-03-11 11:50:09,440][00127] Num frames 5700...
|
1522 |
+
[2023-03-11 11:50:09,550][00127] Num frames 5800...
|
1523 |
+
[2023-03-11 11:50:09,662][00127] Num frames 5900...
|
1524 |
+
[2023-03-11 11:50:09,791][00127] Num frames 6000...
|
1525 |
+
[2023-03-11 11:50:09,950][00127] Num frames 6100...
|
1526 |
+
[2023-03-11 11:50:10,097][00127] Avg episode rewards: #0: 18.653, true rewards: #0: 8.796
|
1527 |
+
[2023-03-11 11:50:10,099][00127] Avg episode reward: 18.653, avg true_objective: 8.796
|
1528 |
+
[2023-03-11 11:50:10,170][00127] Num frames 6200...
|
1529 |
+
[2023-03-11 11:50:10,324][00127] Num frames 6300...
|
1530 |
+
[2023-03-11 11:50:10,479][00127] Num frames 6400...
|
1531 |
+
[2023-03-11 11:50:10,633][00127] Num frames 6500...
|
1532 |
+
[2023-03-11 11:50:10,710][00127] Avg episode rewards: #0: 16.886, true rewards: #0: 8.136
|
1533 |
+
[2023-03-11 11:50:10,714][00127] Avg episode reward: 16.886, avg true_objective: 8.136
|
1534 |
+
[2023-03-11 11:50:10,861][00127] Num frames 6600...
|
1535 |
+
[2023-03-11 11:50:11,011][00127] Num frames 6700...
|
1536 |
+
[2023-03-11 11:50:11,161][00127] Num frames 6800...
|
1537 |
+
[2023-03-11 11:50:11,320][00127] Num frames 6900...
|
1538 |
+
[2023-03-11 11:50:11,473][00127] Num frames 7000...
|
1539 |
+
[2023-03-11 11:50:11,634][00127] Num frames 7100...
|
1540 |
+
[2023-03-11 11:50:11,794][00127] Num frames 7200...
|
1541 |
+
[2023-03-11 11:50:11,969][00127] Num frames 7300...
|
1542 |
+
[2023-03-11 11:50:12,154][00127] Num frames 7400...
|
1543 |
+
[2023-03-11 11:50:12,328][00127] Num frames 7500...
|
1544 |
+
[2023-03-11 11:50:12,499][00127] Num frames 7600...
|
1545 |
+
[2023-03-11 11:50:12,663][00127] Num frames 7700...
|
1546 |
+
[2023-03-11 11:50:12,823][00127] Num frames 7800...
|
1547 |
+
[2023-03-11 11:50:12,980][00127] Num frames 7900...
|
1548 |
+
[2023-03-11 11:50:13,146][00127] Num frames 8000...
|
1549 |
+
[2023-03-11 11:50:13,323][00127] Avg episode rewards: #0: 19.860, true rewards: #0: 8.971
|
1550 |
+
[2023-03-11 11:50:13,325][00127] Avg episode reward: 19.860, avg true_objective: 8.971
|
1551 |
+
[2023-03-11 11:50:13,358][00127] Num frames 8100...
|
1552 |
+
[2023-03-11 11:50:13,470][00127] Num frames 8200...
|
1553 |
+
[2023-03-11 11:50:13,581][00127] Num frames 8300...
|
1554 |
+
[2023-03-11 11:50:13,695][00127] Num frames 8400...
|
1555 |
+
[2023-03-11 11:50:13,823][00127] Num frames 8500...
|
1556 |
+
[2023-03-11 11:50:13,944][00127] Num frames 8600...
|
1557 |
+
[2023-03-11 11:50:14,054][00127] Num frames 8700...
|
1558 |
+
[2023-03-11 11:50:14,173][00127] Num frames 8800...
|
1559 |
+
[2023-03-11 11:50:14,292][00127] Num frames 8900...
|
1560 |
+
[2023-03-11 11:50:14,404][00127] Num frames 9000...
|
1561 |
+
[2023-03-11 11:50:14,517][00127] Num frames 9100...
|
1562 |
+
[2023-03-11 11:50:14,630][00127] Num frames 9200...
|
1563 |
+
[2023-03-11 11:50:14,756][00127] Num frames 9300...
|
1564 |
+
[2023-03-11 11:50:14,865][00127] Num frames 9400...
|
1565 |
+
[2023-03-11 11:50:15,017][00127] Avg episode rewards: #0: 21.182, true rewards: #0: 9.482
|
1566 |
+
[2023-03-11 11:50:15,018][00127] Avg episode reward: 21.182, avg true_objective: 9.482
|
1567 |
+
[2023-03-11 11:51:10,872][00127] Replay video saved to /content/train_dir/default_experiment/replay.mp4!
|