Pdro-ruiz commited on
Commit
3753624
·
verified ·
1 Parent(s): 5cd6c20

First Push

Browse files
SnowballTarget.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:016e7cfbdfd3e643e70162f3233aa3165035ba28331ce69f67416d87e54ea0f0
3
  size 650646
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7fc959ee3f5a7a0163b0acfb3be82fa0681c7234d77380f954907d981e4ac96
3
  size 650646
SnowballTarget/SnowballTarget-149984.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dd79be14b7d653d44748d07c4fa9bee009a43d658bb20f6383dd4c885f804fd2
3
  size 650646
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45c6c9373468ccbfe8071eb18118e6bdcc2d606b140135ed031f431c5dcd1fb0
3
  size 650646
SnowballTarget/SnowballTarget-149984.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6c3c390ed7db9139dbf1e13f3b61af1601daf77f9b26e63f9fc60f522b1e62cd
3
- size 3849115
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c62a727cde7d8b1f9f3567fc248e2f221b51edc39529efa31f27c5d4e3ed307e
3
+ size 3850075
SnowballTarget/SnowballTarget-199984.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:016e7cfbdfd3e643e70162f3233aa3165035ba28331ce69f67416d87e54ea0f0
3
  size 650646
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7fc959ee3f5a7a0163b0acfb3be82fa0681c7234d77380f954907d981e4ac96
3
  size 650646
SnowballTarget/SnowballTarget-199984.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:63124bfff643704ed43d11c4a258957d3464a48fb1f8860c0be8b9756d88a96d
3
- size 3849115
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85a0febdabf116b616bb267561910cb27ec6febdda8eb475561911a886f4bd7e
3
+ size 3850075
SnowballTarget/SnowballTarget-200112.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:016e7cfbdfd3e643e70162f3233aa3165035ba28331ce69f67416d87e54ea0f0
3
  size 650646
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7fc959ee3f5a7a0163b0acfb3be82fa0681c7234d77380f954907d981e4ac96
3
  size 650646
SnowballTarget/SnowballTarget-200112.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ca46b477ead2e9321fe621dc3fc6a08eaec87250ba0fac3bc9ec6fe9c10ca9d0
3
- size 3849115
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:efbb131f4330ba28cb1e64f5e4e288a244533b51fc293c9598afa116a3736d26
3
+ size 3850075
SnowballTarget/SnowballTarget-49936.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eb483a7d018bba682bfa58a059c224bdc5c4674f5d367a9dd203f4beaf41b1d3
3
  size 650646
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21d0ef03adf7429950bab8479a477eb2bb04dd198bab7b3ab91018a6452ed5ca
3
  size 650646
SnowballTarget/SnowballTarget-49936.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:63d75df956f97e3bfaeb041fbea044e654afce31536d5f6515af0d302ea22278
3
- size 3849040
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7093fbc7b77222342b8bc5b048392d5c7de09bf4133025477cb015c7ff3e82a7
3
+ size 3850000
SnowballTarget/SnowballTarget-99960.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5ede1f4f93a8290fcc754f3159fc9d217c05d49985b1bfc38cae46e27c5a981c
3
  size 650646
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b55f67973859d4daa20a7c27b54bb64f316b1526d741fa28bbe3f1dafdda118
3
  size 650646
SnowballTarget/SnowballTarget-99960.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b9bb5472188c50b830b64b59f885c69fcf43f77c1d452c35afd665b06e0b850e
3
- size 3849040
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:976c8b4ac259326c0d7a0bfec85a284da9ad862a94ce791d07a258b9436111d0
3
+ size 3850000
SnowballTarget/checkpoint.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9ca7558bca0cd33c8291516de15f3ddc8b18ee56ad291d8b14c01e31c94fa7ca
3
- size 3848290
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abff4d9bd79af2698e98ca581e210cff62ab2786cf49d6b6c622e636cf3a96fc
3
+ size 3849250
SnowballTarget/events.out.tfevents.1740414182.521d5160cc86.2539.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b406ac84963fd1ad2198688a76b2926c82cfec1fedc365f74a310c397166f060
3
+ size 23525
run_logs/Player-0.log CHANGED
@@ -34,7 +34,7 @@ ALSA lib pcm.c:2664:(snd_pcm_open_noupdate) Unknown PCM default
34
  FMOD failed to initialize the output device.: "Error initializing output device. " (60)
35
  FMOD initialized on nosound output
36
  Begin MonoManager ReloadAssembly
37
- - Completed reload, in 0.126 seconds
38
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
39
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
40
  ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
@@ -45,7 +45,7 @@ ERROR: Shader Standard shader is not supported on this GPU (none of subshaders/f
45
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
46
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
47
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
48
- UnloadTime: 0.811582 ms
49
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
50
  requesting resize 84 x 84
51
  Setting up 1 worker threads for Enlighten.
@@ -169,23 +169,23 @@ Memory Statistics:
169
  Peak Allocated Bytes 0 B
170
  Overflow Count 0
171
  [ALLOC_DEFAULT] Dual Thread Allocator
172
- Peak main deferred allocation count 44
173
  [ALLOC_BUCKET]
174
  Large Block size 4.0 MB
175
  Used Block count 1
176
  Peak Allocated bytes 1.0 MB
177
  [ALLOC_DEFAULT_MAIN]
178
- Peak usage frame count: [4.0 MB-8.0 MB]: 5458 frames
179
  Requested Block Size 16.0 MB
180
  Peak Block count 1
181
- Peak Allocated memory 6.6 MB
182
- Peak Large allocation bytes 0 B
183
  [ALLOC_DEFAULT_THREAD]
184
- Peak usage frame count: [16.0 MB-32.0 MB]: 5458 frames
185
  Requested Block Size 16.0 MB
186
  Peak Block count 1
187
- Peak Allocated memory 17.8 MB
188
- Peak Large allocation bytes 16.0 MB
189
  [ALLOC_TEMP_JOB_1_FRAME]
190
  Initial Block Size 2.0 MB
191
  Used Block Count 1
@@ -213,10 +213,10 @@ Memory Statistics:
213
  Used Block count 1
214
  Peak Allocated bytes 1.0 MB
215
  [ALLOC_GFX_MAIN]
216
- Peak usage frame count: [32.0 KB-64.0 KB]: 4808 frames, [64.0 KB-128.0 KB]: 650 frames
217
  Requested Block Size 16.0 MB
218
  Peak Block count 1
219
- Peak Allocated memory 66.4 KB
220
  Peak Large allocation bytes 0 B
221
  [ALLOC_GFX_THREAD]
222
  Peak usage frame count: [32.0 KB-64.0 KB]: 5458 frames
 
34
  FMOD failed to initialize the output device.: "Error initializing output device. " (60)
35
  FMOD initialized on nosound output
36
  Begin MonoManager ReloadAssembly
37
+ - Completed reload, in 0.091 seconds
38
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
39
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
40
  ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
 
45
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
46
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
47
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
48
+ UnloadTime: 0.690122 ms
49
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
50
  requesting resize 84 x 84
51
  Setting up 1 worker threads for Enlighten.
 
169
  Peak Allocated Bytes 0 B
170
  Overflow Count 0
171
  [ALLOC_DEFAULT] Dual Thread Allocator
172
+ Peak main deferred allocation count 43
173
  [ALLOC_BUCKET]
174
  Large Block size 4.0 MB
175
  Used Block count 1
176
  Peak Allocated bytes 1.0 MB
177
  [ALLOC_DEFAULT_MAIN]
178
+ Peak usage frame count: [16.0 MB-32.0 MB]: 5458 frames
179
  Requested Block Size 16.0 MB
180
  Peak Block count 1
181
+ Peak Allocated memory 23.2 MB
182
+ Peak Large allocation bytes 16.0 MB
183
  [ALLOC_DEFAULT_THREAD]
184
+ Peak usage frame count: [0.5 MB-1.0 MB]: 269 frames, [1.0 MB-2.0 MB]: 5189 frames
185
  Requested Block Size 16.0 MB
186
  Peak Block count 1
187
+ Peak Allocated memory 1.3 MB
188
+ Peak Large allocation bytes 0 B
189
  [ALLOC_TEMP_JOB_1_FRAME]
190
  Initial Block Size 2.0 MB
191
  Used Block Count 1
 
213
  Used Block count 1
214
  Peak Allocated bytes 1.0 MB
215
  [ALLOC_GFX_MAIN]
216
+ Peak usage frame count: [32.0 KB-64.0 KB]: 4384 frames, [64.0 KB-128.0 KB]: 1074 frames
217
  Requested Block Size 16.0 MB
218
  Peak Block count 1
219
+ Peak Allocated memory 67.3 KB
220
  Peak Large allocation bytes 0 B
221
  [ALLOC_GFX_THREAD]
222
  Peak usage frame count: [32.0 KB-64.0 KB]: 5458 frames
run_logs/timers.json CHANGED
@@ -2,15 +2,15 @@
2
  "name": "root",
3
  "gauges": {
4
  "SnowballTarget.Policy.Entropy.mean": {
5
- "value": 0.8904275298118591,
6
- "min": 0.8904275298118591,
7
- "max": 2.856306314468384,
8
  "count": 20
9
  },
10
  "SnowballTarget.Policy.Entropy.sum": {
11
- "value": 8462.623046875,
12
- "min": 8462.623046875,
13
- "max": 29157.173828125,
14
  "count": 20
15
  },
16
  "SnowballTarget.Step.mean": {
@@ -26,39 +26,39 @@
26
  "count": 20
27
  },
28
  "SnowballTarget.Policy.ExtrinsicValueEstimate.mean": {
29
- "value": 12.741615295410156,
30
- "min": 0.33357229828834534,
31
- "max": 12.811731338500977,
32
  "count": 20
33
  },
34
  "SnowballTarget.Policy.ExtrinsicValueEstimate.sum": {
35
- "value": 2484.614990234375,
36
- "min": 64.71302795410156,
37
- "max": 2613.59326171875,
38
  "count": 20
39
  },
40
  "SnowballTarget.Losses.PolicyLoss.mean": {
41
- "value": 0.0717248672913853,
42
- "min": 0.061247457263594055,
43
- "max": 0.07482796639288056,
44
  "count": 20
45
  },
46
  "SnowballTarget.Losses.PolicyLoss.sum": {
47
- "value": 0.2868994691655412,
48
- "min": 0.2684500220265381,
49
- "max": 0.34777520309537924,
50
  "count": 20
51
  },
52
  "SnowballTarget.Losses.ValueLoss.mean": {
53
- "value": 0.21524380293546938,
54
- "min": 0.11826446109825708,
55
- "max": 0.29258550068034844,
56
  "count": 20
57
  },
58
  "SnowballTarget.Losses.ValueLoss.sum": {
59
- "value": 0.8609752117418775,
60
- "min": 0.47305784439302834,
61
- "max": 1.3924340447958778,
62
  "count": 20
63
  },
64
  "SnowballTarget.Policy.LearningRate.mean": {
@@ -110,27 +110,27 @@
110
  "count": 20
111
  },
112
  "SnowballTarget.Environment.CumulativeReward.mean": {
113
- "value": 24.863636363636363,
114
- "min": 3.3181818181818183,
115
- "max": 25.2,
116
  "count": 20
117
  },
118
  "SnowballTarget.Environment.CumulativeReward.sum": {
119
- "value": 1094.0,
120
- "min": 146.0,
121
- "max": 1386.0,
122
  "count": 20
123
  },
124
  "SnowballTarget.Policy.ExtrinsicReward.mean": {
125
- "value": 24.863636363636363,
126
- "min": 3.3181818181818183,
127
- "max": 25.2,
128
  "count": 20
129
  },
130
  "SnowballTarget.Policy.ExtrinsicReward.sum": {
131
- "value": 1094.0,
132
- "min": 146.0,
133
- "max": 1386.0,
134
  "count": 20
135
  },
136
  "SnowballTarget.IsTraining.mean": {
@@ -148,7 +148,7 @@
148
  },
149
  "metadata": {
150
  "timer_format_version": "0.1.0",
151
- "start_time_seconds": "1739559036",
152
  "python_version": "3.10.12 (main, Jul 5 2023, 18:54:27) [GCC 11.2.0]",
153
  "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/SnowballTarget.yaml --env=./training-envs-executables/linux/SnowballTarget/SnowballTarget --run-id=SnowballTarget1 --no-graphics",
154
  "mlagents_version": "1.2.0.dev0",
@@ -156,59 +156,59 @@
156
  "communication_protocol_version": "1.5.0",
157
  "pytorch_version": "2.6.0+cu124",
158
  "numpy_version": "1.23.5",
159
- "end_time_seconds": "1739559590"
160
  },
161
- "total": 553.775966757,
162
  "count": 1,
163
- "self": 0.5968578629999683,
164
  "children": {
165
  "run_training.setup": {
166
- "total": 0.02851163999991968,
167
  "count": 1,
168
- "self": 0.02851163999991968
169
  },
170
  "TrainerController.start_learning": {
171
- "total": 553.1505972540001,
172
  "count": 1,
173
- "self": 0.5905008389963768,
174
  "children": {
175
  "TrainerController._reset_env": {
176
- "total": 3.0744958129999986,
177
  "count": 1,
178
- "self": 3.0744958129999986
179
  },
180
  "TrainerController.advance": {
181
- "total": 549.4032337430036,
182
  "count": 18192,
183
- "self": 0.5896822670050597,
184
  "children": {
185
  "env_step": {
186
- "total": 383.7661822479937,
187
  "count": 18192,
188
- "self": 325.7715936879854,
189
  "children": {
190
  "SubprocessEnvManager._take_step": {
191
- "total": 57.62972968001088,
192
  "count": 18192,
193
- "self": 1.8614130170113867,
194
  "children": {
195
  "TorchPolicy.evaluate": {
196
- "total": 55.768316662999496,
197
  "count": 18192,
198
- "self": 55.768316662999496
199
  }
200
  }
201
  },
202
  "workers": {
203
- "total": 0.3648588799974277,
204
  "count": 18192,
205
  "self": 0.0,
206
  "children": {
207
  "worker_root": {
208
- "total": 551.1256154159803,
209
  "count": 18192,
210
  "is_parallel": true,
211
- "self": 268.53343001597796,
212
  "children": {
213
  "run_training.setup": {
214
  "total": 0.0,
@@ -217,48 +217,48 @@
217
  "self": 0.0,
218
  "children": {
219
  "steps_from_proto": {
220
- "total": 0.006684858000085114,
221
  "count": 1,
222
  "is_parallel": true,
223
- "self": 0.004827687000101832,
224
  "children": {
225
  "_process_rank_one_or_two_observation": {
226
- "total": 0.001857170999983282,
227
  "count": 10,
228
  "is_parallel": true,
229
- "self": 0.001857170999983282
230
  }
231
  }
232
  },
233
  "UnityEnvironment.step": {
234
- "total": 0.04492649300004814,
235
  "count": 1,
236
  "is_parallel": true,
237
- "self": 0.0006925000000137516,
238
  "children": {
239
  "UnityEnvironment._generate_step_input": {
240
- "total": 0.00045358399995620857,
241
  "count": 1,
242
  "is_parallel": true,
243
- "self": 0.00045358399995620857
244
  },
245
  "communicator.exchange": {
246
- "total": 0.041495350999980474,
247
  "count": 1,
248
  "is_parallel": true,
249
- "self": 0.041495350999980474
250
  },
251
  "steps_from_proto": {
252
- "total": 0.002285058000097706,
253
  "count": 1,
254
  "is_parallel": true,
255
- "self": 0.0004296260001410701,
256
  "children": {
257
  "_process_rank_one_or_two_observation": {
258
- "total": 0.001855431999956636,
259
  "count": 10,
260
  "is_parallel": true,
261
- "self": 0.001855431999956636
262
  }
263
  }
264
  }
@@ -267,34 +267,34 @@
267
  }
268
  },
269
  "UnityEnvironment.step": {
270
- "total": 282.59218540000234,
271
  "count": 18191,
272
  "is_parallel": true,
273
- "self": 13.584144121005124,
274
  "children": {
275
  "UnityEnvironment._generate_step_input": {
276
- "total": 7.258736434001889,
277
  "count": 18191,
278
  "is_parallel": true,
279
- "self": 7.258736434001889
280
  },
281
  "communicator.exchange": {
282
- "total": 221.20109826099326,
283
  "count": 18191,
284
  "is_parallel": true,
285
- "self": 221.20109826099326
286
  },
287
  "steps_from_proto": {
288
- "total": 40.54820658400206,
289
  "count": 18191,
290
  "is_parallel": true,
291
- "self": 7.785025289994451,
292
  "children": {
293
  "_process_rank_one_or_two_observation": {
294
- "total": 32.76318129400761,
295
  "count": 181910,
296
  "is_parallel": true,
297
- "self": 32.76318129400761
298
  }
299
  }
300
  }
@@ -307,31 +307,31 @@
307
  }
308
  },
309
  "trainer_advance": {
310
- "total": 165.04736922800487,
311
  "count": 18192,
312
- "self": 0.8107356420080123,
313
  "children": {
314
  "process_trajectory": {
315
- "total": 30.981938301996706,
316
  "count": 18192,
317
- "self": 30.556624780996685,
318
  "children": {
319
  "RLTrainer._checkpoint": {
320
- "total": 0.42531352100002096,
321
  "count": 4,
322
- "self": 0.42531352100002096
323
  }
324
  }
325
  },
326
  "_update_policy": {
327
- "total": 133.25469528400015,
328
  "count": 90,
329
- "self": 48.204229221004084,
330
  "children": {
331
  "TorchPPOOptimizer.update": {
332
- "total": 85.05046606299607,
333
  "count": 4587,
334
- "self": 85.05046606299607
335
  }
336
  }
337
  }
@@ -340,19 +340,19 @@
340
  }
341
  },
342
  "trainer_threads": {
343
- "total": 1.1760000688809669e-06,
344
  "count": 1,
345
- "self": 1.1760000688809669e-06
346
  },
347
  "TrainerController._save_models": {
348
- "total": 0.08236568300003455,
349
  "count": 1,
350
- "self": 0.0012316180000198074,
351
  "children": {
352
  "RLTrainer._checkpoint": {
353
- "total": 0.08113406500001474,
354
  "count": 1,
355
- "self": 0.08113406500001474
356
  }
357
  }
358
  }
 
2
  "name": "root",
3
  "gauges": {
4
  "SnowballTarget.Policy.Entropy.mean": {
5
+ "value": 0.910011351108551,
6
+ "min": 0.8656197190284729,
7
+ "max": 2.8648324012756348,
8
  "count": 20
9
  },
10
  "SnowballTarget.Policy.Entropy.sum": {
11
+ "value": 8648.748046875,
12
+ "min": 8648.748046875,
13
+ "max": 29244.208984375,
14
  "count": 20
15
  },
16
  "SnowballTarget.Step.mean": {
 
26
  "count": 20
27
  },
28
  "SnowballTarget.Policy.ExtrinsicValueEstimate.mean": {
29
+ "value": 12.425869941711426,
30
+ "min": 0.34991899132728577,
31
+ "max": 12.425869941711426,
32
  "count": 20
33
  },
34
  "SnowballTarget.Policy.ExtrinsicValueEstimate.sum": {
35
+ "value": 2423.044677734375,
36
+ "min": 67.88428497314453,
37
+ "max": 2482.50537109375,
38
  "count": 20
39
  },
40
  "SnowballTarget.Losses.PolicyLoss.mean": {
41
+ "value": 0.06616925851823693,
42
+ "min": 0.05926711688896154,
43
+ "max": 0.07734173149672678,
44
  "count": 20
45
  },
46
  "SnowballTarget.Losses.PolicyLoss.sum": {
47
+ "value": 0.26467703407294774,
48
+ "min": 0.26129032025974264,
49
+ "max": 0.3623867884840971,
50
  "count": 20
51
  },
52
  "SnowballTarget.Losses.ValueLoss.mean": {
53
+ "value": 0.19691138393154334,
54
+ "min": 0.12070066804139346,
55
+ "max": 0.29515857141391905,
56
  "count": 20
57
  },
58
  "SnowballTarget.Losses.ValueLoss.sum": {
59
+ "value": 0.7876455357261734,
60
+ "min": 0.4828026721655738,
61
+ "max": 1.4757928570695953,
62
  "count": 20
63
  },
64
  "SnowballTarget.Policy.LearningRate.mean": {
 
110
  "count": 20
111
  },
112
  "SnowballTarget.Environment.CumulativeReward.mean": {
113
+ "value": 24.568181818181817,
114
+ "min": 3.4318181818181817,
115
+ "max": 24.618181818181817,
116
  "count": 20
117
  },
118
  "SnowballTarget.Environment.CumulativeReward.sum": {
119
+ "value": 1081.0,
120
+ "min": 151.0,
121
+ "max": 1354.0,
122
  "count": 20
123
  },
124
  "SnowballTarget.Policy.ExtrinsicReward.mean": {
125
+ "value": 24.568181818181817,
126
+ "min": 3.4318181818181817,
127
+ "max": 24.618181818181817,
128
  "count": 20
129
  },
130
  "SnowballTarget.Policy.ExtrinsicReward.sum": {
131
+ "value": 1081.0,
132
+ "min": 151.0,
133
+ "max": 1354.0,
134
  "count": 20
135
  },
136
  "SnowballTarget.IsTraining.mean": {
 
148
  },
149
  "metadata": {
150
  "timer_format_version": "0.1.0",
151
+ "start_time_seconds": "1740414181",
152
  "python_version": "3.10.12 (main, Jul 5 2023, 18:54:27) [GCC 11.2.0]",
153
  "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/SnowballTarget.yaml --env=./training-envs-executables/linux/SnowballTarget/SnowballTarget --run-id=SnowballTarget1 --no-graphics",
154
  "mlagents_version": "1.2.0.dev0",
 
156
  "communication_protocol_version": "1.5.0",
157
  "pytorch_version": "2.6.0+cu124",
158
  "numpy_version": "1.23.5",
159
+ "end_time_seconds": "1740414603"
160
  },
161
+ "total": 421.24864388800006,
162
  "count": 1,
163
+ "self": 0.48575493300018024,
164
  "children": {
165
  "run_training.setup": {
166
+ "total": 0.022782633999895552,
167
  "count": 1,
168
+ "self": 0.022782633999895552
169
  },
170
  "TrainerController.start_learning": {
171
+ "total": 420.740106321,
172
  "count": 1,
173
+ "self": 0.3420336140151221,
174
  "children": {
175
  "TrainerController._reset_env": {
176
+ "total": 2.9915769550000277,
177
  "count": 1,
178
+ "self": 2.9915769550000277
179
  },
180
  "TrainerController.advance": {
181
+ "total": 417.3181682669848,
182
  "count": 18192,
183
+ "self": 0.3669360199830862,
184
  "children": {
185
  "env_step": {
186
+ "total": 295.2127896729975,
187
  "count": 18192,
188
+ "self": 224.94137538300572,
189
  "children": {
190
  "SubprocessEnvManager._take_step": {
191
+ "total": 70.06671784899618,
192
  "count": 18192,
193
+ "self": 1.2215718040006323,
194
  "children": {
195
  "TorchPolicy.evaluate": {
196
+ "total": 68.84514604499554,
197
  "count": 18192,
198
+ "self": 68.84514604499554
199
  }
200
  }
201
  },
202
  "workers": {
203
+ "total": 0.20469644099557627,
204
  "count": 18192,
205
  "self": 0.0,
206
  "children": {
207
  "worker_root": {
208
+ "total": 419.23537344600663,
209
  "count": 18192,
210
  "is_parallel": true,
211
+ "self": 221.76811015601197,
212
  "children": {
213
  "run_training.setup": {
214
  "total": 0.0,
 
217
  "self": 0.0,
218
  "children": {
219
  "steps_from_proto": {
220
+ "total": 0.005878691000020808,
221
  "count": 1,
222
  "is_parallel": true,
223
+ "self": 0.0043751579999025125,
224
  "children": {
225
  "_process_rank_one_or_two_observation": {
226
+ "total": 0.0015035330001182956,
227
  "count": 10,
228
  "is_parallel": true,
229
+ "self": 0.0015035330001182956
230
  }
231
  }
232
  },
233
  "UnityEnvironment.step": {
234
+ "total": 0.03810243200007335,
235
  "count": 1,
236
  "is_parallel": true,
237
+ "self": 0.0005928630001790225,
238
  "children": {
239
  "UnityEnvironment._generate_step_input": {
240
+ "total": 0.0003768139999920095,
241
  "count": 1,
242
  "is_parallel": true,
243
+ "self": 0.0003768139999920095
244
  },
245
  "communicator.exchange": {
246
+ "total": 0.035276182999950834,
247
  "count": 1,
248
  "is_parallel": true,
249
+ "self": 0.035276182999950834
250
  },
251
  "steps_from_proto": {
252
+ "total": 0.001856571999951484,
253
  "count": 1,
254
  "is_parallel": true,
255
+ "self": 0.00037617899988617864,
256
  "children": {
257
  "_process_rank_one_or_two_observation": {
258
+ "total": 0.0014803930000653054,
259
  "count": 10,
260
  "is_parallel": true,
261
+ "self": 0.0014803930000653054
262
  }
263
  }
264
  }
 
267
  }
268
  },
269
  "UnityEnvironment.step": {
270
+ "total": 197.46726328999466,
271
  "count": 18191,
272
  "is_parallel": true,
273
+ "self": 9.495421679997321,
274
  "children": {
275
  "UnityEnvironment._generate_step_input": {
276
+ "total": 5.2779393760006315,
277
  "count": 18191,
278
  "is_parallel": true,
279
+ "self": 5.2779393760006315
280
  },
281
  "communicator.exchange": {
282
+ "total": 152.50354203899803,
283
  "count": 18191,
284
  "is_parallel": true,
285
+ "self": 152.50354203899803
286
  },
287
  "steps_from_proto": {
288
+ "total": 30.190360194998675,
289
  "count": 18191,
290
  "is_parallel": true,
291
+ "self": 5.320118532977517,
292
  "children": {
293
  "_process_rank_one_or_two_observation": {
294
+ "total": 24.870241662021158,
295
  "count": 181910,
296
  "is_parallel": true,
297
+ "self": 24.870241662021158
298
  }
299
  }
300
  }
 
307
  }
308
  },
309
  "trainer_advance": {
310
+ "total": 121.73844257400424,
311
  "count": 18192,
312
+ "self": 0.41397302000052605,
313
  "children": {
314
  "process_trajectory": {
315
+ "total": 26.338829458004057,
316
  "count": 18192,
317
+ "self": 25.765888316004066,
318
  "children": {
319
  "RLTrainer._checkpoint": {
320
+ "total": 0.5729411419999906,
321
  "count": 4,
322
+ "self": 0.5729411419999906
323
  }
324
  }
325
  },
326
  "_update_policy": {
327
+ "total": 94.98564009599966,
328
  "count": 90,
329
+ "self": 37.724641506003195,
330
  "children": {
331
  "TorchPPOOptimizer.update": {
332
+ "total": 57.26099858999646,
333
  "count": 4587,
334
+ "self": 57.26099858999646
335
  }
336
  }
337
  }
 
340
  }
341
  },
342
  "trainer_threads": {
343
+ "total": 8.450000450466177e-07,
344
  "count": 1,
345
+ "self": 8.450000450466177e-07
346
  },
347
  "TrainerController._save_models": {
348
+ "total": 0.08832663999999113,
349
  "count": 1,
350
+ "self": 0.0008979300000646617,
351
  "children": {
352
  "RLTrainer._checkpoint": {
353
+ "total": 0.08742870999992647,
354
  "count": 1,
355
+ "self": 0.08742870999992647
356
  }
357
  }
358
  }
run_logs/training_status.json CHANGED
@@ -4,8 +4,8 @@
4
  {
5
  "steps": 49936,
6
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-49936.onnx",
7
- "reward": 13.909090909090908,
8
- "creation_time": 1739559176.6895177,
9
  "auxillary_file_paths": [
10
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-49936.pt"
11
  ]
@@ -13,8 +13,8 @@
13
  {
14
  "steps": 99960,
15
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.onnx",
16
- "reward": 22.454545454545453,
17
- "creation_time": 1739559315.8581285,
18
  "auxillary_file_paths": [
19
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.pt"
20
  ]
@@ -22,8 +22,8 @@
22
  {
23
  "steps": 149984,
24
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.onnx",
25
- "reward": 24.818181818181817,
26
- "creation_time": 1739559452.971862,
27
  "auxillary_file_paths": [
28
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.pt"
29
  ]
@@ -31,8 +31,8 @@
31
  {
32
  "steps": 199984,
33
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.onnx",
34
- "reward": 25.363636363636363,
35
- "creation_time": 1739559589.6321197,
36
  "auxillary_file_paths": [
37
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.pt"
38
  ]
@@ -40,8 +40,8 @@
40
  {
41
  "steps": 200112,
42
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.onnx",
43
- "reward": 25.363636363636363,
44
- "creation_time": 1739559589.7287204,
45
  "auxillary_file_paths": [
46
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
47
  ]
@@ -50,8 +50,8 @@
50
  "final_checkpoint": {
51
  "steps": 200112,
52
  "file_path": "results/SnowballTarget1/SnowballTarget.onnx",
53
- "reward": 25.363636363636363,
54
- "creation_time": 1739559589.7287204,
55
  "auxillary_file_paths": [
56
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
57
  ]
 
4
  {
5
  "steps": 49936,
6
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-49936.onnx",
7
+ "reward": 15.818181818181818,
8
+ "creation_time": 1740414289.5383692,
9
  "auxillary_file_paths": [
10
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-49936.pt"
11
  ]
 
13
  {
14
  "steps": 99960,
15
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.onnx",
16
+ "reward": 21.181818181818183,
17
+ "creation_time": 1740414393.7080793,
18
  "auxillary_file_paths": [
19
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.pt"
20
  ]
 
22
  {
23
  "steps": 149984,
24
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.onnx",
25
+ "reward": 21.545454545454547,
26
+ "creation_time": 1740414498.9479148,
27
  "auxillary_file_paths": [
28
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.pt"
29
  ]
 
31
  {
32
  "steps": 199984,
33
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.onnx",
34
+ "reward": 23.545454545454547,
35
+ "creation_time": 1740414602.5646312,
36
  "auxillary_file_paths": [
37
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.pt"
38
  ]
 
40
  {
41
  "steps": 200112,
42
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.onnx",
43
+ "reward": 23.545454545454547,
44
+ "creation_time": 1740414602.6669168,
45
  "auxillary_file_paths": [
46
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
47
  ]
 
50
  "final_checkpoint": {
51
  "steps": 200112,
52
  "file_path": "results/SnowballTarget1/SnowballTarget.onnx",
53
+ "reward": 23.545454545454547,
54
+ "creation_time": 1740414602.6669168,
55
  "auxillary_file_paths": [
56
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
57
  ]