hungtrab commited on
Commit
4d943fc
·
verified ·
1 Parent(s): 3443a13

Changed HyperParameters Push

Browse files
SnowballTarget.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:348118097b397f0686223be14f667e18ac57a0d2cba996b8068f8c52784f83c0
3
  size 650646
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:390c8895a186c0156463bcfd36ea8c804709938a6edc9a1f98cbaeb754c045f3
3
  size 650646
SnowballTarget/SnowballTarget-1599976.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:105a95fcdab0a8c98123e9aae57d1ae0d6a2650d742a1c0ee118129221fcfe4b
3
+ size 650646
SnowballTarget/SnowballTarget-1599976.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2c0870ac299fc526fb8264a68581392497cfae3bb4c921c72e6eddc64946e47
3
+ size 3850705
SnowballTarget/SnowballTarget-1649992.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0dc3de457d7d8c52afc424abb6e372ce0a3c26500238372ded4bc3115fe7a596
3
+ size 650646
SnowballTarget/SnowballTarget-1649992.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46101c1146ac879b7b4d3d88e89ca894bd19f42c66678941b9cbe871b1e8e9d5
3
+ size 3850705
SnowballTarget/SnowballTarget-1699936.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f8cbc54773ad15042f1718018934e28a080c702b1836461846405b4afa0708d
3
+ size 650646
SnowballTarget/SnowballTarget-1699936.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dedfd76e139d961b3398a682ebe727ad7993886808a999e0cef424be4e29a179
3
+ size 3850705
SnowballTarget/SnowballTarget-1749960.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f94a7866f208fefd1f52d800d2edd300e17912d187122bc47d37cdece2cfcc55
3
+ size 650646
SnowballTarget/SnowballTarget-1749960.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7976bbdb79eb333dad5aad584eaa3e13821f411e67d0db0033b09be19c81c6e5
3
+ size 3850705
SnowballTarget/SnowballTarget-1799984.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e27dd28ee59a32cf0862c97f36e0c406a8aeb3812c7287f1baac73db4a01f9f
3
+ size 650646
SnowballTarget/SnowballTarget-1799984.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf5c5fe59cfbad9dafb23ecdc451e16c2f56c9afd597d31f724a41fce26fa000
3
+ size 3850705
SnowballTarget/SnowballTarget-1849984.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b930aa08dc4b971e5838d54b3c01640b63199c0cce3acb216ba597a509b0de2
3
+ size 650646
SnowballTarget/SnowballTarget-1849984.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9177f9d685b332710400f190b3a531435d2c6279a8a9b712e33116a1c7920129
3
+ size 3850705
SnowballTarget/SnowballTarget-1899944.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9bcfd80ae5b4e4de6c7367e6216cba6bdeb5bb493e8635f2106f3337907b6f2f
3
+ size 650646
SnowballTarget/SnowballTarget-1899944.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88b43144ee506d08da9bf00d13062197da497fc979a7b6a14118cea60a953c65
3
+ size 3850705
SnowballTarget/SnowballTarget-1949968.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f528b24bc65161ce83c341707731e0d2a3656ad94c38633c3973bf1f30589a74
3
+ size 650646
SnowballTarget/SnowballTarget-1949968.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27a11eb6a29b432f034a864d3fa0e66d32f8b94eff12b3ece7bae2b6e8f454c1
3
+ size 3850705
SnowballTarget/SnowballTarget-1999992.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:390c8895a186c0156463bcfd36ea8c804709938a6edc9a1f98cbaeb754c045f3
3
+ size 650646
SnowballTarget/SnowballTarget-1999992.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40649a28ab5b929442f5197306c1c726798eb592046af3a43db565d5075ccf41
3
+ size 3850705
SnowballTarget/SnowballTarget-2000504.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:390c8895a186c0156463bcfd36ea8c804709938a6edc9a1f98cbaeb754c045f3
3
+ size 650646
SnowballTarget/SnowballTarget-2000504.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:586c86918dd1d16fc4fe4ea3daffc4361d84fa3ac97c1631bd56b2e90a87ba1a
3
+ size 3850705
SnowballTarget/checkpoint.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8eb350d7fb4ec20139ff83c77dc67d253943fb75ecfd2080149a1307d4214441
3
  size 3849717
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ced02b0e57ab123d83909078c8eb7e9759b59ccf67ff9ede7af959bf75a5c164
3
  size 3849717
SnowballTarget/events.out.tfevents.1756744183.hungchan-Precision-7560.15823.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b7e7d6bacee5e8f225174f961fb37a0d312b0f0c09935629ab3e0eda91f07f3
3
+ size 184485
config.json CHANGED
@@ -1 +1 @@
1
- {"default_settings": null, "behaviors": {"SnowballTarget": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 1024, "buffer_size": 10240, "learning_rate": 0.0003, "beta": 0.01, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "shared_critic": false, "learning_rate_schedule": "linear", "beta_schedule": "linear", "epsilon_schedule": "linear"}, "checkpoint_interval": 50000, "network_settings": {"normalize": false, "hidden_units": 256, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}}, "init_path": null, "keep_checkpoints": 10, "even_checkpoints": false, "max_steps": 200000, "time_horizon": 64, "summary_freq": 10000, "threaded": false, "self_play": null, "behavioral_cloning": null}}, "env_settings": {"env_path": "./training-envs-executables/linux/SnowballTarget/SnowballTarget", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "timeout_wait": 60, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true, "no_graphics_monitor": false}, "environment_parameters": null, "checkpoint_settings": {"run_id": "SnowballTarget1", "initialize_from": null, "load_model": false, "resume": false, "force": false, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
 
1
+ {"default_settings": null, "behaviors": {"SnowballTarget": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 1024, "buffer_size": 10240, "learning_rate": 0.0003, "beta": 0.01, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 5, "shared_critic": false, "learning_rate_schedule": "linear", "beta_schedule": "linear", "epsilon_schedule": "linear"}, "checkpoint_interval": 50000, "network_settings": {"normalize": false, "hidden_units": 256, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}}, "init_path": null, "keep_checkpoints": 10, "even_checkpoints": false, "max_steps": 2000000, "time_horizon": 64, "summary_freq": 10000, "threaded": false, "self_play": null, "behavioral_cloning": null}}, "env_settings": {"env_path": "./training-envs-executables/linux/SnowballTarget/SnowballTarget", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "timeout_wait": 60, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true, "no_graphics_monitor": false}, "environment_parameters": null, "checkpoint_settings": {"run_id": "SnowballTarget2", "initialize_from": null, "load_model": false, "resume": false, "force": false, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
configuration.yaml CHANGED
@@ -9,7 +9,7 @@ behaviors:
9
  beta: 0.01
10
  epsilon: 0.2
11
  lambd: 0.95
12
- num_epoch: 3
13
  shared_critic: false
14
  learning_rate_schedule: linear
15
  beta_schedule: linear
@@ -38,7 +38,7 @@ behaviors:
38
  init_path: null
39
  keep_checkpoints: 10
40
  even_checkpoints: false
41
- max_steps: 200000
42
  time_horizon: 64
43
  summary_freq: 10000
44
  threaded: false
@@ -66,7 +66,7 @@ engine_settings:
66
  no_graphics_monitor: false
67
  environment_parameters: null
68
  checkpoint_settings:
69
- run_id: SnowballTarget1
70
  initialize_from: null
71
  load_model: false
72
  resume: false
 
9
  beta: 0.01
10
  epsilon: 0.2
11
  lambd: 0.95
12
+ num_epoch: 5
13
  shared_critic: false
14
  learning_rate_schedule: linear
15
  beta_schedule: linear
 
38
  init_path: null
39
  keep_checkpoints: 10
40
  even_checkpoints: false
41
+ max_steps: 2000000
42
  time_horizon: 64
43
  summary_freq: 10000
44
  threaded: false
 
66
  no_graphics_monitor: false
67
  environment_parameters: null
68
  checkpoint_settings:
69
+ run_id: SnowballTarget2
70
  initialize_from: null
71
  load_model: false
72
  resume: false
run_logs/Player-0.log CHANGED
@@ -2,9 +2,6 @@ Mono path[0] = '/media/hungchan/Storage/Works/xai/X AI/Code/Hugging Face RL/5. U
2
  Mono config path = '/media/hungchan/Storage/Works/xai/X AI/Code/Hugging Face RL/5. Unity ML-agent/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/MonoBleedingEdge/etc'
3
  Preloaded 'lib_burst_generated.so'
4
  Preloaded 'libgrpc_csharp_ext.x64.so'
5
- PlayerPrefs - Creating folder: /home/hungchan/.config/unity3d/Hugging Face
6
- PlayerPrefs - Creating folder: /home/hungchan/.config/unity3d/Hugging Face/SnowballTarget
7
- Unable to load player prefs
8
  Initialize engine version: 2021.3.14f1 (eee1884e7226)
9
  [Subsystems] Discovering subsystems at path /media/hungchan/Storage/Works/xai/X AI/Code/Hugging Face RL/5. Unity ML-agent/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/UnitySubsystems
10
  Forcing GfxDevice: Null
@@ -14,7 +11,7 @@ NullGfxDevice:
14
  Renderer: Null Device
15
  Vendor: Unity Technologies
16
  Begin MonoManager ReloadAssembly
17
- - Completed reload, in 0.053 seconds
18
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
19
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
20
  ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
@@ -25,7 +22,7 @@ ERROR: Shader Standard shader is not supported on this GPU (none of subshaders/f
25
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
26
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
27
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
28
- UnloadTime: 0.514438 ms
29
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
30
  requesting resize 84 x 84
31
  Setting up 8 worker threads for Enlighten.
@@ -33,7 +30,7 @@ Memory Statistics:
33
  [ALLOC_TEMP_TLS] TLS Allocator
34
  StackAllocators :
35
  [ALLOC_TEMP_MAIN]
36
- Peak usage frame count: [8.0 KB-16.0 KB]: 5367 frames, [16.0 KB-32.0 KB]: 90 frames, [2.0 MB-4.0 MB]: 1 frames
37
  Initial Block Size 4.0 MB
38
  Current Block Size 4.0 MB
39
  Peak Allocated Bytes 2.0 MB
@@ -219,19 +216,19 @@ Memory Statistics:
219
  Peak Allocated Bytes 0 B
220
  Overflow Count 0
221
  [ALLOC_DEFAULT] Dual Thread Allocator
222
- Peak main deferred allocation count 37
223
  [ALLOC_BUCKET]
224
  Large Block size 4.0 MB
225
  Used Block count 1
226
  Peak Allocated bytes 1.0 MB
227
  [ALLOC_DEFAULT_MAIN]
228
- Peak usage frame count: [4.0 MB-8.0 MB]: 5458 frames
229
  Requested Block Size 16.0 MB
230
- Peak Block count 1
231
- Peak Allocated memory 6.8 MB
232
  Peak Large allocation bytes 0 B
233
  [ALLOC_DEFAULT_THREAD]
234
- Peak usage frame count: [16.0 MB-32.0 MB]: 5458 frames
235
  Requested Block Size 16.0 MB
236
  Peak Block count 1
237
  Peak Allocated memory 17.6 MB
@@ -263,13 +260,13 @@ Memory Statistics:
263
  Used Block count 1
264
  Peak Allocated bytes 1.0 MB
265
  [ALLOC_GFX_MAIN]
266
- Peak usage frame count: [32.0 KB-64.0 KB]: 4045 frames, [64.0 KB-128.0 KB]: 1413 frames
267
  Requested Block Size 16.0 MB
268
  Peak Block count 1
269
  Peak Allocated memory 66.8 KB
270
  Peak Large allocation bytes 0 B
271
  [ALLOC_GFX_THREAD]
272
- Peak usage frame count: [32.0 KB-64.0 KB]: 5458 frames
273
  Requested Block Size 16.0 MB
274
  Peak Block count 1
275
  Peak Allocated memory 39.6 KB
@@ -281,13 +278,13 @@ Memory Statistics:
281
  Used Block count 1
282
  Peak Allocated bytes 1.0 MB
283
  [ALLOC_CACHEOBJECTS_MAIN]
284
- Peak usage frame count: [0.5 MB-1.0 MB]: 5458 frames
285
  Requested Block Size 4.0 MB
286
  Peak Block count 1
287
  Peak Allocated memory 0.6 MB
288
  Peak Large allocation bytes 0 B
289
  [ALLOC_CACHEOBJECTS_THREAD]
290
- Peak usage frame count: [0.5 MB-1.0 MB]: 5457 frames, [2.0 MB-4.0 MB]: 1 frames
291
  Requested Block Size 4.0 MB
292
  Peak Block count 1
293
  Peak Allocated memory 2.2 MB
@@ -299,13 +296,13 @@ Memory Statistics:
299
  Used Block count 1
300
  Peak Allocated bytes 1.0 MB
301
  [ALLOC_TYPETREE_MAIN]
302
- Peak usage frame count: [0-1.0 KB]: 5458 frames
303
  Requested Block Size 2.0 MB
304
  Peak Block count 1
305
  Peak Allocated memory 1.0 KB
306
  Peak Large allocation bytes 0 B
307
  [ALLOC_TYPETREE_THREAD]
308
- Peak usage frame count: [1.0 KB-2.0 KB]: 5458 frames
309
  Requested Block Size 2.0 MB
310
  Peak Block count 1
311
  Peak Allocated memory 1.7 KB
 
2
  Mono config path = '/media/hungchan/Storage/Works/xai/X AI/Code/Hugging Face RL/5. Unity ML-agent/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/MonoBleedingEdge/etc'
3
  Preloaded 'lib_burst_generated.so'
4
  Preloaded 'libgrpc_csharp_ext.x64.so'
 
 
 
5
  Initialize engine version: 2021.3.14f1 (eee1884e7226)
6
  [Subsystems] Discovering subsystems at path /media/hungchan/Storage/Works/xai/X AI/Code/Hugging Face RL/5. Unity ML-agent/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/UnitySubsystems
7
  Forcing GfxDevice: Null
 
11
  Renderer: Null Device
12
  Vendor: Unity Technologies
13
  Begin MonoManager ReloadAssembly
14
+ - Completed reload, in 0.062 seconds
15
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
16
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
17
  ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
 
22
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
23
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
24
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
25
+ UnloadTime: 0.426896 ms
26
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
27
  requesting resize 84 x 84
28
  Setting up 8 worker threads for Enlighten.
 
30
  [ALLOC_TEMP_TLS] TLS Allocator
31
  StackAllocators :
32
  [ALLOC_TEMP_MAIN]
33
+ Peak usage frame count: [8.0 KB-16.0 KB]: 53650 frames, [16.0 KB-32.0 KB]: 909 frames, [2.0 MB-4.0 MB]: 1 frames
34
  Initial Block Size 4.0 MB
35
  Current Block Size 4.0 MB
36
  Peak Allocated Bytes 2.0 MB
 
216
  Peak Allocated Bytes 0 B
217
  Overflow Count 0
218
  [ALLOC_DEFAULT] Dual Thread Allocator
219
+ Peak main deferred allocation count 40
220
  [ALLOC_BUCKET]
221
  Large Block size 4.0 MB
222
  Used Block count 1
223
  Peak Allocated bytes 1.0 MB
224
  [ALLOC_DEFAULT_MAIN]
225
+ Peak usage frame count: [4.0 MB-8.0 MB]: 13187 frames, [8.0 MB-16.0 MB]: 27610 frames, [16.0 MB-32.0 MB]: 13763 frames
226
  Requested Block Size 16.0 MB
227
+ Peak Block count 2
228
+ Peak Allocated memory 24.4 MB
229
  Peak Large allocation bytes 0 B
230
  [ALLOC_DEFAULT_THREAD]
231
+ Peak usage frame count: [16.0 MB-32.0 MB]: 54560 frames
232
  Requested Block Size 16.0 MB
233
  Peak Block count 1
234
  Peak Allocated memory 17.6 MB
 
260
  Used Block count 1
261
  Peak Allocated bytes 1.0 MB
262
  [ALLOC_GFX_MAIN]
263
+ Peak usage frame count: [32.0 KB-64.0 KB]: 51576 frames, [64.0 KB-128.0 KB]: 2984 frames
264
  Requested Block Size 16.0 MB
265
  Peak Block count 1
266
  Peak Allocated memory 66.8 KB
267
  Peak Large allocation bytes 0 B
268
  [ALLOC_GFX_THREAD]
269
+ Peak usage frame count: [32.0 KB-64.0 KB]: 54560 frames
270
  Requested Block Size 16.0 MB
271
  Peak Block count 1
272
  Peak Allocated memory 39.6 KB
 
278
  Used Block count 1
279
  Peak Allocated bytes 1.0 MB
280
  [ALLOC_CACHEOBJECTS_MAIN]
281
+ Peak usage frame count: [0.5 MB-1.0 MB]: 54560 frames
282
  Requested Block Size 4.0 MB
283
  Peak Block count 1
284
  Peak Allocated memory 0.6 MB
285
  Peak Large allocation bytes 0 B
286
  [ALLOC_CACHEOBJECTS_THREAD]
287
+ Peak usage frame count: [0.5 MB-1.0 MB]: 54559 frames, [2.0 MB-4.0 MB]: 1 frames
288
  Requested Block Size 4.0 MB
289
  Peak Block count 1
290
  Peak Allocated memory 2.2 MB
 
296
  Used Block count 1
297
  Peak Allocated bytes 1.0 MB
298
  [ALLOC_TYPETREE_MAIN]
299
+ Peak usage frame count: [0-1.0 KB]: 54560 frames
300
  Requested Block Size 2.0 MB
301
  Peak Block count 1
302
  Peak Allocated memory 1.0 KB
303
  Peak Large allocation bytes 0 B
304
  [ALLOC_TYPETREE_THREAD]
305
+ Peak usage frame count: [1.0 KB-2.0 KB]: 54560 frames
306
  Requested Block Size 2.0 MB
307
  Peak Block count 1
308
  Peak Allocated memory 1.7 KB
run_logs/timers.json CHANGED
@@ -2,213 +2,213 @@
2
  "name": "root",
3
  "gauges": {
4
  "SnowballTarget.Policy.Entropy.mean": {
5
- "value": 2.008164882659912,
6
- "min": 2.008164882659912,
7
- "max": 2.890336751937866,
8
- "count": 20
9
  },
10
  "SnowballTarget.Policy.Entropy.sum": {
11
- "value": 19085.599609375,
12
- "min": 19085.599609375,
13
- "max": 29620.40234375,
14
- "count": 20
15
  },
16
  "SnowballTarget.Step.mean": {
17
- "value": 199984.0,
18
  "min": 9952.0,
19
- "max": 199984.0,
20
- "count": 20
21
  },
22
  "SnowballTarget.Step.sum": {
23
- "value": 199984.0,
24
  "min": 9952.0,
25
- "max": 199984.0,
26
- "count": 20
27
  },
28
  "SnowballTarget.Policy.ExtrinsicValueEstimate.mean": {
29
- "value": 5.566380023956299,
30
- "min": 0.042696863412857056,
31
- "max": 5.566380023956299,
32
- "count": 20
33
  },
34
  "SnowballTarget.Policy.ExtrinsicValueEstimate.sum": {
35
- "value": 1085.444091796875,
36
- "min": 8.283191680908203,
37
- "max": 1085.444091796875,
38
- "count": 20
39
  },
40
  "SnowballTarget.Environment.EpisodeLength.mean": {
41
  "value": 199.0,
42
  "min": 199.0,
43
  "max": 199.0,
44
- "count": 20
45
  },
46
  "SnowballTarget.Environment.EpisodeLength.sum": {
47
- "value": 8756.0,
48
  "min": 8756.0,
49
  "max": 10945.0,
50
- "count": 20
51
  },
52
  "SnowballTarget.Environment.CumulativeReward.mean": {
53
- "value": 14.522727272727273,
54
- "min": 2.6136363636363638,
55
- "max": 14.927272727272728,
56
- "count": 20
57
  },
58
  "SnowballTarget.Environment.CumulativeReward.sum": {
59
- "value": 639.0,
60
- "min": 115.0,
61
- "max": 821.0,
62
- "count": 20
63
  },
64
  "SnowballTarget.Policy.ExtrinsicReward.mean": {
65
- "value": 14.522727272727273,
66
- "min": 2.6136363636363638,
67
- "max": 14.927272727272728,
68
- "count": 20
69
  },
70
  "SnowballTarget.Policy.ExtrinsicReward.sum": {
71
- "value": 639.0,
72
- "min": 115.0,
73
- "max": 821.0,
74
- "count": 20
75
  },
76
  "SnowballTarget.IsTraining.mean": {
77
  "value": 1.0,
78
  "min": 1.0,
79
  "max": 1.0,
80
- "count": 20
81
  },
82
  "SnowballTarget.IsTraining.sum": {
83
  "value": 1.0,
84
  "min": 1.0,
85
  "max": 1.0,
86
- "count": 20
87
  },
88
  "SnowballTarget.Losses.PolicyLoss.mean": {
89
- "value": 0.020801379935195048,
90
- "min": 0.01685354337096214,
91
- "max": 0.029999618922981124,
92
- "count": 19
93
  },
94
  "SnowballTarget.Losses.PolicyLoss.sum": {
95
- "value": 0.020801379935195048,
96
- "min": 0.01685354337096214,
97
- "max": 0.029999618922981124,
98
- "count": 19
99
  },
100
  "SnowballTarget.Losses.ValueLoss.mean": {
101
- "value": 0.355442476272583,
102
- "min": 0.09278392195701599,
103
- "max": 0.3592230121294657,
104
- "count": 19
105
  },
106
  "SnowballTarget.Losses.ValueLoss.sum": {
107
- "value": 0.355442476272583,
108
- "min": 0.09278392195701599,
109
- "max": 0.3592230121294657,
110
- "count": 19
111
  },
112
  "SnowballTarget.Policy.LearningRate.mean": {
113
- "value": 1.9440993520000114e-06,
114
- "min": 1.9440993520000114e-06,
115
- "max": 0.00028363200545599993,
116
- "count": 19
117
  },
118
  "SnowballTarget.Policy.LearningRate.sum": {
119
- "value": 1.9440993520000114e-06,
120
- "min": 1.9440993520000114e-06,
121
- "max": 0.00028363200545599993,
122
- "count": 19
123
  },
124
  "SnowballTarget.Policy.Epsilon.mean": {
125
- "value": 0.10064800000000003,
126
- "min": 0.10064800000000003,
127
- "max": 0.19454399999999994,
128
- "count": 19
129
  },
130
  "SnowballTarget.Policy.Epsilon.sum": {
131
- "value": 0.10064800000000003,
132
- "min": 0.10064800000000003,
133
- "max": 0.19454399999999994,
134
- "count": 19
135
  },
136
  "SnowballTarget.Policy.Beta.mean": {
137
- "value": 7.473520000000039e-05,
138
- "min": 7.473520000000039e-05,
139
- "max": 0.009454945599999997,
140
- "count": 19
141
  },
142
  "SnowballTarget.Policy.Beta.sum": {
143
- "value": 7.473520000000039e-05,
144
- "min": 7.473520000000039e-05,
145
- "max": 0.009454945599999997,
146
- "count": 19
147
  }
148
  },
149
  "metadata": {
150
  "timer_format_version": "0.1.0",
151
- "start_time_seconds": "1756743646",
152
  "python_version": "3.10.12 (main, Jul 5 2023, 18:54:27) [GCC 11.2.0]",
153
- "command_line_arguments": "/home/hungchan/miniconda3/envs/rl_31012/bin/mlagents-learn ./config/ppo/SnowballTarget.yaml --env=./training-envs-executables/linux/SnowballTarget/SnowballTarget --run-id=SnowballTarget1 --no-graphics",
154
  "mlagents_version": "1.2.0.dev0",
155
  "mlagents_envs_version": "1.2.0.dev0",
156
  "communication_protocol_version": "1.5.0",
157
  "pytorch_version": "2.8.0+cu128",
158
  "numpy_version": "1.23.5",
159
- "end_time_seconds": "1756743833"
160
  },
161
- "total": 186.97781900200016,
162
  "count": 1,
163
- "self": 0.21862168200004817,
164
  "children": {
165
  "run_training.setup": {
166
- "total": 0.013576254000099652,
167
  "count": 1,
168
- "self": 0.013576254000099652
169
  },
170
  "TrainerController.start_learning": {
171
- "total": 186.745621066,
172
  "count": 1,
173
- "self": 0.22402318901185936,
174
  "children": {
175
  "TrainerController._reset_env": {
176
- "total": 1.2709940259999257,
177
  "count": 1,
178
- "self": 1.2709940259999257
179
  },
180
  "TrainerController.advance": {
181
- "total": 185.17632726798843,
182
- "count": 18192,
183
- "self": 0.21790039998245447,
184
  "children": {
185
  "env_step": {
186
- "total": 142.81059462600797,
187
- "count": 18192,
188
- "self": 108.78446163200192,
189
  "children": {
190
  "SubprocessEnvManager._take_step": {
191
- "total": 33.88783405701088,
192
- "count": 18192,
193
- "self": 0.6471468669826663,
194
  "children": {
195
  "TorchPolicy.evaluate": {
196
- "total": 33.24068719002821,
197
- "count": 18192,
198
- "self": 33.24068719002821
199
  }
200
  }
201
  },
202
  "workers": {
203
- "total": 0.13829893699517015,
204
- "count": 18192,
205
  "self": 0.0,
206
  "children": {
207
  "worker_root": {
208
- "total": 186.2964024690143,
209
- "count": 18192,
210
  "is_parallel": true,
211
- "self": 91.90503982399446,
212
  "children": {
213
  "run_training.setup": {
214
  "total": 0.0,
@@ -217,48 +217,48 @@
217
  "self": 0.0,
218
  "children": {
219
  "steps_from_proto": {
220
- "total": 0.0012169280000762228,
221
  "count": 1,
222
  "is_parallel": true,
223
- "self": 0.0003886250001414737,
224
  "children": {
225
  "_process_rank_one_or_two_observation": {
226
- "total": 0.0008283029999347491,
227
  "count": 10,
228
  "is_parallel": true,
229
- "self": 0.0008283029999347491
230
  }
231
  }
232
  },
233
  "UnityEnvironment.step": {
234
- "total": 0.017391022000083467,
235
  "count": 1,
236
  "is_parallel": true,
237
- "self": 0.0003446420003001549,
238
  "children": {
239
  "UnityEnvironment._generate_step_input": {
240
- "total": 0.0002456879999499506,
241
  "count": 1,
242
  "is_parallel": true,
243
- "self": 0.0002456879999499506
244
  },
245
  "communicator.exchange": {
246
- "total": 0.015950279999970007,
247
  "count": 1,
248
  "is_parallel": true,
249
- "self": 0.015950279999970007
250
  },
251
  "steps_from_proto": {
252
- "total": 0.0008504119998633541,
253
  "count": 1,
254
  "is_parallel": true,
255
- "self": 0.00018180899996877997,
256
  "children": {
257
  "_process_rank_one_or_two_observation": {
258
- "total": 0.0006686029998945742,
259
  "count": 10,
260
  "is_parallel": true,
261
- "self": 0.0006686029998945742
262
  }
263
  }
264
  }
@@ -267,34 +267,34 @@
267
  }
268
  },
269
  "UnityEnvironment.step": {
270
- "total": 94.39136264501985,
271
- "count": 18191,
272
  "is_parallel": true,
273
- "self": 3.6992152049976994,
274
  "children": {
275
  "UnityEnvironment._generate_step_input": {
276
- "total": 2.0660426019983333,
277
- "count": 18191,
278
  "is_parallel": true,
279
- "self": 2.0660426019983333
280
  },
281
  "communicator.exchange": {
282
- "total": 77.33828555702075,
283
- "count": 18191,
284
  "is_parallel": true,
285
- "self": 77.33828555702075
286
  },
287
  "steps_from_proto": {
288
- "total": 11.287819281003067,
289
- "count": 18191,
290
  "is_parallel": true,
291
- "self": 2.2710576070865045,
292
  "children": {
293
  "_process_rank_one_or_two_observation": {
294
- "total": 9.016761673916562,
295
- "count": 181910,
296
  "is_parallel": true,
297
- "self": 9.016761673916562
298
  }
299
  }
300
  }
@@ -307,31 +307,31 @@
307
  }
308
  },
309
  "trainer_advance": {
310
- "total": 42.147832241998,
311
- "count": 18192,
312
- "self": 0.26467858001501554,
313
  "children": {
314
  "process_trajectory": {
315
- "total": 13.53277166398334,
316
- "count": 18192,
317
- "self": 13.274490549983511,
318
  "children": {
319
  "RLTrainer._checkpoint": {
320
- "total": 0.25828111399982845,
321
- "count": 4,
322
- "self": 0.25828111399982845
323
  }
324
  }
325
  },
326
  "_update_policy": {
327
- "total": 28.350381997999648,
328
- "count": 19,
329
- "self": 22.547381198998664,
330
  "children": {
331
  "TorchPPOOptimizer.update": {
332
- "total": 5.8030007990009835,
333
- "count": 570,
334
- "self": 5.8030007990009835
335
  }
336
  }
337
  }
@@ -340,19 +340,19 @@
340
  }
341
  },
342
  "trainer_threads": {
343
- "total": 1.2429998150764732e-06,
344
  "count": 1,
345
- "self": 1.2429998150764732e-06
346
  },
347
  "TrainerController._save_models": {
348
- "total": 0.07427533999998559,
349
  "count": 1,
350
- "self": 0.00045179900007497054,
351
  "children": {
352
  "RLTrainer._checkpoint": {
353
- "total": 0.07382354099991062,
354
  "count": 1,
355
- "self": 0.07382354099991062
356
  }
357
  }
358
  }
 
2
  "name": "root",
3
  "gauges": {
4
  "SnowballTarget.Policy.Entropy.mean": {
5
+ "value": 0.7899582386016846,
6
+ "min": 0.7675452828407288,
7
+ "max": 2.8903439044952393,
8
+ "count": 200
9
  },
10
  "SnowballTarget.Policy.Entropy.sum": {
11
+ "value": 8133.41015625,
12
+ "min": 7294.75048828125,
13
+ "max": 29570.138671875,
14
+ "count": 200
15
  },
16
  "SnowballTarget.Step.mean": {
17
+ "value": 1999992.0,
18
  "min": 9952.0,
19
+ "max": 1999992.0,
20
+ "count": 200
21
  },
22
  "SnowballTarget.Step.sum": {
23
+ "value": 1999992.0,
24
  "min": 9952.0,
25
+ "max": 1999992.0,
26
+ "count": 200
27
  },
28
  "SnowballTarget.Policy.ExtrinsicValueEstimate.mean": {
29
+ "value": 14.199864387512207,
30
+ "min": -0.09513217210769653,
31
+ "max": 14.263813972473145,
32
+ "count": 200
33
  },
34
  "SnowballTarget.Policy.ExtrinsicValueEstimate.sum": {
35
+ "value": 2910.97216796875,
36
+ "min": -18.45564079284668,
37
+ "max": 2924.081787109375,
38
+ "count": 200
39
  },
40
  "SnowballTarget.Environment.EpisodeLength.mean": {
41
  "value": 199.0,
42
  "min": 199.0,
43
  "max": 199.0,
44
+ "count": 200
45
  },
46
  "SnowballTarget.Environment.EpisodeLength.sum": {
47
+ "value": 10945.0,
48
  "min": 8756.0,
49
  "max": 10945.0,
50
+ "count": 200
51
  },
52
  "SnowballTarget.Environment.CumulativeReward.mean": {
53
+ "value": 27.90909090909091,
54
+ "min": 2.6363636363636362,
55
+ "max": 28.386363636363637,
56
+ "count": 200
57
  },
58
  "SnowballTarget.Environment.CumulativeReward.sum": {
59
+ "value": 1535.0,
60
+ "min": 116.0,
61
+ "max": 1545.0,
62
+ "count": 200
63
  },
64
  "SnowballTarget.Policy.ExtrinsicReward.mean": {
65
+ "value": 27.90909090909091,
66
+ "min": 2.6363636363636362,
67
+ "max": 28.386363636363637,
68
+ "count": 200
69
  },
70
  "SnowballTarget.Policy.ExtrinsicReward.sum": {
71
+ "value": 1535.0,
72
+ "min": 116.0,
73
+ "max": 1545.0,
74
+ "count": 200
75
  },
76
  "SnowballTarget.IsTraining.mean": {
77
  "value": 1.0,
78
  "min": 1.0,
79
  "max": 1.0,
80
+ "count": 200
81
  },
82
  "SnowballTarget.IsTraining.sum": {
83
  "value": 1.0,
84
  "min": 1.0,
85
  "max": 1.0,
86
+ "count": 200
87
  },
88
  "SnowballTarget.Losses.PolicyLoss.mean": {
89
+ "value": 0.023605609668884427,
90
+ "min": 0.015988392693398055,
91
+ "max": 0.029666174179874362,
92
+ "count": 191
93
  },
94
  "SnowballTarget.Losses.PolicyLoss.sum": {
95
+ "value": 0.023605609668884427,
96
+ "min": 0.015988392693398055,
97
+ "max": 0.029666174179874362,
98
+ "count": 191
99
  },
100
  "SnowballTarget.Losses.ValueLoss.mean": {
101
+ "value": 0.15311876356601714,
102
+ "min": 0.0945033897459507,
103
+ "max": 0.3203707575798035,
104
+ "count": 191
105
  },
106
  "SnowballTarget.Losses.ValueLoss.sum": {
107
+ "value": 0.15311876356601714,
108
+ "min": 0.0945033897459507,
109
+ "max": 0.3203707575798035,
110
+ "count": 191
111
  },
112
  "SnowballTarget.Policy.LearningRate.mean": {
113
+ "value": 5.844998051999849e-07,
114
+ "min": 5.844998051999849e-07,
115
+ "max": 0.0002983632005456,
116
+ "count": 191
117
  },
118
  "SnowballTarget.Policy.LearningRate.sum": {
119
+ "value": 5.844998051999849e-07,
120
+ "min": 5.844998051999849e-07,
121
+ "max": 0.0002983632005456,
122
+ "count": 191
123
  },
124
  "SnowballTarget.Policy.Epsilon.mean": {
125
+ "value": 0.1001948,
126
+ "min": 0.1001948,
127
+ "max": 0.1994544,
128
+ "count": 191
129
  },
130
  "SnowballTarget.Policy.Epsilon.sum": {
131
+ "value": 0.1001948,
132
+ "min": 0.1001948,
133
+ "max": 0.1994544,
134
+ "count": 191
135
  },
136
  "SnowballTarget.Policy.Beta.mean": {
137
+ "value": 2.94605199999995e-05,
138
+ "min": 2.94605199999995e-05,
139
+ "max": 0.00994549456,
140
+ "count": 191
141
  },
142
  "SnowballTarget.Policy.Beta.sum": {
143
+ "value": 2.94605199999995e-05,
144
+ "min": 2.94605199999995e-05,
145
+ "max": 0.00994549456,
146
+ "count": 191
147
  }
148
  },
149
  "metadata": {
150
  "timer_format_version": "0.1.0",
151
+ "start_time_seconds": "1756744182",
152
  "python_version": "3.10.12 (main, Jul 5 2023, 18:54:27) [GCC 11.2.0]",
153
+ "command_line_arguments": "/home/hungchan/miniconda3/envs/rl_31012/bin/mlagents-learn ./config/ppo/SnowballTarget.yaml --env=./training-envs-executables/linux/SnowballTarget/SnowballTarget --run-id=SnowballTarget2 --no-graphics",
154
  "mlagents_version": "1.2.0.dev0",
155
  "mlagents_envs_version": "1.2.0.dev0",
156
  "communication_protocol_version": "1.5.0",
157
  "pytorch_version": "2.8.0+cu128",
158
  "numpy_version": "1.23.5",
159
+ "end_time_seconds": "1756746217"
160
  },
161
+ "total": 2034.945576685,
162
  "count": 1,
163
+ "self": 0.22280772400017668,
164
  "children": {
165
  "run_training.setup": {
166
+ "total": 0.014937456999859933,
167
  "count": 1,
168
+ "self": 0.014937456999859933
169
  },
170
  "TrainerController.start_learning": {
171
+ "total": 2034.707831504,
172
  "count": 1,
173
+ "self": 2.090146202893038,
174
  "children": {
175
  "TrainerController._reset_env": {
176
+ "total": 1.483924055999978,
177
  "count": 1,
178
+ "self": 1.483924055999978
179
  },
180
  "TrainerController.advance": {
181
+ "total": 2031.0903070571067,
182
+ "count": 181864,
183
+ "self": 1.9976428490458602,
184
  "children": {
185
  "env_step": {
186
+ "total": 1407.2546968289648,
187
+ "count": 181864,
188
+ "self": 1074.0869557420808,
189
  "children": {
190
  "SubprocessEnvManager._take_step": {
191
+ "total": 331.8149656088565,
192
+ "count": 181864,
193
+ "self": 6.398366203912019,
194
  "children": {
195
  "TorchPolicy.evaluate": {
196
+ "total": 325.41659940494446,
197
+ "count": 181864,
198
+ "self": 325.41659940494446
199
  }
200
  }
201
  },
202
  "workers": {
203
+ "total": 1.3527754780275245,
204
+ "count": 181864,
205
  "self": 0.0,
206
  "children": {
207
  "worker_root": {
208
+ "total": 2031.6684058839683,
209
+ "count": 181864,
210
  "is_parallel": true,
211
+ "self": 1096.1810742638997,
212
  "children": {
213
  "run_training.setup": {
214
  "total": 0.0,
 
217
  "self": 0.0,
218
  "children": {
219
  "steps_from_proto": {
220
+ "total": 0.0009950090000074852,
221
  "count": 1,
222
  "is_parallel": true,
223
+ "self": 0.0002717519998896023,
224
  "children": {
225
  "_process_rank_one_or_two_observation": {
226
+ "total": 0.0007232570001178829,
227
  "count": 10,
228
  "is_parallel": true,
229
+ "self": 0.0007232570001178829
230
  }
231
  }
232
  },
233
  "UnityEnvironment.step": {
234
+ "total": 0.015677181000000928,
235
  "count": 1,
236
  "is_parallel": true,
237
+ "self": 0.00022112700025900267,
238
  "children": {
239
  "UnityEnvironment._generate_step_input": {
240
+ "total": 0.00016523799990864063,
241
  "count": 1,
242
  "is_parallel": true,
243
+ "self": 0.00016523799990864063
244
  },
245
  "communicator.exchange": {
246
+ "total": 0.01464410699986729,
247
  "count": 1,
248
  "is_parallel": true,
249
+ "self": 0.01464410699986729
250
  },
251
  "steps_from_proto": {
252
+ "total": 0.0006467089999659947,
253
  "count": 1,
254
  "is_parallel": true,
255
+ "self": 0.0001443310000013298,
256
  "children": {
257
  "_process_rank_one_or_two_observation": {
258
+ "total": 0.0005023779999646649,
259
  "count": 10,
260
  "is_parallel": true,
261
+ "self": 0.0005023779999646649
262
  }
263
  }
264
  }
 
267
  }
268
  },
269
  "UnityEnvironment.step": {
270
+ "total": 935.4873316200685,
271
+ "count": 181863,
272
  "is_parallel": true,
273
+ "self": 35.60573260907904,
274
  "children": {
275
  "UnityEnvironment._generate_step_input": {
276
+ "total": 20.04231953900762,
277
+ "count": 181863,
278
  "is_parallel": true,
279
+ "self": 20.04231953900762
280
  },
281
  "communicator.exchange": {
282
+ "total": 769.2791198750026,
283
+ "count": 181863,
284
  "is_parallel": true,
285
+ "self": 769.2791198750026
286
  },
287
  "steps_from_proto": {
288
+ "total": 110.56015959697925,
289
+ "count": 181863,
290
  "is_parallel": true,
291
+ "self": 22.2171321491287,
292
  "children": {
293
  "_process_rank_one_or_two_observation": {
294
+ "total": 88.34302744785055,
295
+ "count": 1818630,
296
  "is_parallel": true,
297
+ "self": 88.34302744785055
298
  }
299
  }
300
  }
 
307
  }
308
  },
309
  "trainer_advance": {
310
+ "total": 621.837967379096,
311
+ "count": 181864,
312
+ "self": 2.5286666780179985,
313
  "children": {
314
  "process_trajectory": {
315
+ "total": 135.94426321507808,
316
+ "count": 181864,
317
+ "self": 133.73425045007662,
318
  "children": {
319
  "RLTrainer._checkpoint": {
320
+ "total": 2.2100127650014656,
321
+ "count": 40,
322
+ "self": 2.2100127650014656
323
  }
324
  }
325
  },
326
  "_update_policy": {
327
+ "total": 483.3650374859999,
328
+ "count": 191,
329
+ "self": 388.68595532199015,
330
  "children": {
331
  "TorchPPOOptimizer.update": {
332
+ "total": 94.67908216400974,
333
+ "count": 9550,
334
+ "self": 94.67908216400974
335
  }
336
  }
337
  }
 
340
  }
341
  },
342
  "trainer_threads": {
343
+ "total": 6.63000264466973e-07,
344
  "count": 1,
345
+ "self": 6.63000264466973e-07
346
  },
347
  "TrainerController._save_models": {
348
+ "total": 0.043453525000131776,
349
  "count": 1,
350
+ "self": 0.00045814800023435964,
351
  "children": {
352
  "RLTrainer._checkpoint": {
353
+ "total": 0.04299537699989742,
354
  "count": 1,
355
+ "self": 0.04299537699989742
356
  }
357
  }
358
  }
run_logs/training_status.json CHANGED
@@ -2,58 +2,103 @@
2
  "SnowballTarget": {
3
  "checkpoints": [
4
  {
5
- "steps": 49936,
6
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-49936.onnx",
7
- "reward": 6.181818181818182,
8
- "creation_time": 1756743693.4427552,
9
  "auxillary_file_paths": [
10
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-49936.pt"
11
  ]
12
  },
13
  {
14
- "steps": 99960,
15
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.onnx",
16
- "reward": 8.878787878787879,
17
- "creation_time": 1756743740.8214815,
18
  "auxillary_file_paths": [
19
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.pt"
20
  ]
21
  },
22
  {
23
- "steps": 149984,
24
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.onnx",
25
- "reward": 12.636363636363637,
26
- "creation_time": 1756743786.9100819,
27
  "auxillary_file_paths": [
28
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.pt"
29
  ]
30
  },
31
  {
32
- "steps": 199984,
33
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.onnx",
34
- "reward": null,
35
- "creation_time": 1756743832.776409,
36
  "auxillary_file_paths": [
37
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.pt"
38
  ]
39
  },
40
  {
41
- "steps": 200112,
42
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.onnx",
43
- "reward": null,
44
- "creation_time": 1756743832.8569334,
45
  "auxillary_file_paths": [
46
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
47
  ]
48
  }
49
  ],
50
  "final_checkpoint": {
51
- "steps": 200112,
52
- "file_path": "results/SnowballTarget1/SnowballTarget.onnx",
53
- "reward": null,
54
- "creation_time": 1756743832.8569334,
55
  "auxillary_file_paths": [
56
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
57
  ]
58
  }
59
  },
 
2
  "SnowballTarget": {
3
  "checkpoints": [
4
  {
5
+ "steps": 1599976,
6
+ "file_path": "results/SnowballTarget2/SnowballTarget/SnowballTarget-1599976.onnx",
7
+ "reward": 26.818181818181817,
8
+ "creation_time": 1756745826.9205108,
9
  "auxillary_file_paths": [
10
+ "results/SnowballTarget2/SnowballTarget/SnowballTarget-1599976.pt"
11
  ]
12
  },
13
  {
14
+ "steps": 1649992,
15
+ "file_path": "results/SnowballTarget2/SnowballTarget/SnowballTarget-1649992.onnx",
16
+ "reward": 27.703703703703702,
17
+ "creation_time": 1756745873.9021735,
18
  "auxillary_file_paths": [
19
+ "results/SnowballTarget2/SnowballTarget/SnowballTarget-1649992.pt"
20
  ]
21
  },
22
  {
23
+ "steps": 1699936,
24
+ "file_path": "results/SnowballTarget2/SnowballTarget/SnowballTarget-1699936.onnx",
25
+ "reward": 28.12121212121212,
26
+ "creation_time": 1756745923.30265,
27
  "auxillary_file_paths": [
28
+ "results/SnowballTarget2/SnowballTarget/SnowballTarget-1699936.pt"
29
  ]
30
  },
31
  {
32
+ "steps": 1749960,
33
+ "file_path": "results/SnowballTarget2/SnowballTarget/SnowballTarget-1749960.onnx",
34
+ "reward": 28.318181818181817,
35
+ "creation_time": 1756745972.599262,
36
  "auxillary_file_paths": [
37
+ "results/SnowballTarget2/SnowballTarget/SnowballTarget-1749960.pt"
38
  ]
39
  },
40
  {
41
+ "steps": 1799984,
42
+ "file_path": "results/SnowballTarget2/SnowballTarget/SnowballTarget-1799984.onnx",
43
+ "reward": 26.818181818181817,
44
+ "creation_time": 1756746022.0835266,
45
  "auxillary_file_paths": [
46
+ "results/SnowballTarget2/SnowballTarget/SnowballTarget-1799984.pt"
47
+ ]
48
+ },
49
+ {
50
+ "steps": 1849984,
51
+ "file_path": "results/SnowballTarget2/SnowballTarget/SnowballTarget-1849984.onnx",
52
+ "reward": 28.0,
53
+ "creation_time": 1756746068.6017828,
54
+ "auxillary_file_paths": [
55
+ "results/SnowballTarget2/SnowballTarget/SnowballTarget-1849984.pt"
56
+ ]
57
+ },
58
+ {
59
+ "steps": 1899944,
60
+ "file_path": "results/SnowballTarget2/SnowballTarget/SnowballTarget-1899944.onnx",
61
+ "reward": 27.886363636363637,
62
+ "creation_time": 1756746118.0606852,
63
+ "auxillary_file_paths": [
64
+ "results/SnowballTarget2/SnowballTarget/SnowballTarget-1899944.pt"
65
+ ]
66
+ },
67
+ {
68
+ "steps": 1949968,
69
+ "file_path": "results/SnowballTarget2/SnowballTarget/SnowballTarget-1949968.onnx",
70
+ "reward": 27.757575757575758,
71
+ "creation_time": 1756746167.9905531,
72
+ "auxillary_file_paths": [
73
+ "results/SnowballTarget2/SnowballTarget/SnowballTarget-1949968.pt"
74
+ ]
75
+ },
76
+ {
77
+ "steps": 1999992,
78
+ "file_path": "results/SnowballTarget2/SnowballTarget/SnowballTarget-1999992.onnx",
79
+ "reward": 27.59090909090909,
80
+ "creation_time": 1756746217.2514899,
81
+ "auxillary_file_paths": [
82
+ "results/SnowballTarget2/SnowballTarget/SnowballTarget-1999992.pt"
83
+ ]
84
+ },
85
+ {
86
+ "steps": 2000504,
87
+ "file_path": "results/SnowballTarget2/SnowballTarget/SnowballTarget-2000504.onnx",
88
+ "reward": 27.59090909090909,
89
+ "creation_time": 1756746217.3245153,
90
+ "auxillary_file_paths": [
91
+ "results/SnowballTarget2/SnowballTarget/SnowballTarget-2000504.pt"
92
  ]
93
  }
94
  ],
95
  "final_checkpoint": {
96
+ "steps": 2000504,
97
+ "file_path": "results/SnowballTarget2/SnowballTarget.onnx",
98
+ "reward": 27.59090909090909,
99
+ "creation_time": 1756746217.3245153,
100
  "auxillary_file_paths": [
101
+ "results/SnowballTarget2/SnowballTarget/SnowballTarget-2000504.pt"
102
  ]
103
  }
104
  },