Amshaker commited on
Commit
d45a9e8
·
verified ·
1 Parent(s): 5a77cf9

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -158,3 +158,4 @@ bimamba_stage_a/wandb/offline-run-20260227_135547-lxvmjzq8/run-lxvmjzq8.wandb fi
158
  bimamba_stage_a/wandb/offline-run-20260227_141732-lxvmjzq8/run-lxvmjzq8.wandb filter=lfs diff=lfs merge=lfs -text
159
  bimamba_stage_a/wandb/offline-run-20260301_014655-lxvmjzq8/logs/debug-internal.log filter=lfs diff=lfs merge=lfs -text
160
  bimamba_stage_a/wandb/offline-run-20260301_014655-lxvmjzq8/run-lxvmjzq8.wandb filter=lfs diff=lfs merge=lfs -text
 
 
158
  bimamba_stage_a/wandb/offline-run-20260227_141732-lxvmjzq8/run-lxvmjzq8.wandb filter=lfs diff=lfs merge=lfs -text
159
  bimamba_stage_a/wandb/offline-run-20260301_014655-lxvmjzq8/logs/debug-internal.log filter=lfs diff=lfs merge=lfs -text
160
  bimamba_stage_a/wandb/offline-run-20260301_014655-lxvmjzq8/run-lxvmjzq8.wandb filter=lfs diff=lfs merge=lfs -text
161
+ bimamba_stage_a_128/wandb/offline-run-20260307_233556-cbyk5tna/run-cbyk5tna.wandb filter=lfs diff=lfs merge=lfs -text
bimamba_stage_a_128/checkpoints/epoch0_step2000.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:454eb9f5d2612fa0fc7c393badc4074be2123766abbf3b86ed353196cbe1da5b
3
+ size 3418833602
bimamba_stage_a_128/slurmm1.0-6945.out ADDED
The diff for this file is too large to render. See raw diff
 
bimamba_stage_a_128/wandb/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
 
bimamba_stage_a_128/wandb/debug.log ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2026-03-07 23:35:56,648 INFO MainThread:528584 [wandb_setup.py:_flush():77] Current SDK version is 0.17.9
2
+ 2026-03-07 23:35:56,648 INFO MainThread:528584 [wandb_setup.py:_flush():77] Configure stats pid to 528584
3
+ 2026-03-07 23:35:56,648 INFO MainThread:528584 [wandb_setup.py:_flush():77] Loading settings from /home/x_fahkh/.config/wandb/settings
4
+ 2026-03-07 23:35:56,648 INFO MainThread:528584 [wandb_setup.py:_flush():77] Loading settings from /proj/cvl/users/x_fahkh2/WorldMem_Repro/wandb/settings
5
+ 2026-03-07 23:35:56,648 INFO MainThread:528584 [wandb_setup.py:_flush():77] Loading settings from environment variables: {'disabled': 'true'}
6
+ 2026-03-07 23:35:56,648 INFO MainThread:528584 [wandb_setup.py:_flush():77] Applying setup settings: {'_disable_service': False}
7
+ 2026-03-07 23:35:56,648 WARNING MainThread:528584 [wandb_setup.py:_flush():77] Could not find program at -m main
8
+ 2026-03-07 23:35:56,648 INFO MainThread:528584 [wandb_setup.py:_flush():77] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m main'}
9
+ 2026-03-07 23:35:56,648 INFO MainThread:528584 [wandb_init.py:_log_setup():524] Logging user logs to /proj/cvl/users/x_fahkh2/WorldMem_Repro/checkpoints/bimamba_stage_a_128/wandb/offline-run-20260307_233556-cbyk5tna/logs/debug.log
10
+ 2026-03-07 23:35:56,648 INFO MainThread:528584 [wandb_init.py:_log_setup():525] Logging internal logs to /proj/cvl/users/x_fahkh2/WorldMem_Repro/checkpoints/bimamba_stage_a_128/wandb/offline-run-20260307_233556-cbyk5tna/logs/debug-internal.log
11
+ 2026-03-07 23:35:56,648 INFO MainThread:528584 [wandb_init.py:init():608] calling init triggers
12
+ 2026-03-07 23:35:56,648 INFO MainThread:528584 [wandb_init.py:init():615] wandb.init called with sweep_config: {}
13
+ config: {'experiment': {'debug': '${debug}', 'tasks': ['training'], 'num_nodes': 1, 'training': {'precision': '16-mixed', 'compile': False, 'lr': 0.0002, 'batch_size': 128, 'max_epochs': -1, 'max_steps': 2000, 'max_time': None, 'data': {'num_workers': 4, 'shuffle': True}, 'optim': {'accumulate_grad_batches': 1, 'gradient_clip_val': 1.0}, 'checkpointing': {'every_n_train_steps': 50, 'every_n_epochs': None, 'train_time_interval': None, 'enable_version_counter': False}}, 'validation': {'precision': '16-mixed', 'compile': False, 'batch_size': 4, 'val_every_n_step': 100, 'val_every_n_epoch': None, 'limit_batch': 1, 'inference_mode': False, 'data': {'num_workers': 4, 'shuffle': False}}, 'test': {'precision': '16-mixed', 'compile': False, 'batch_size': 1, 'limit_batch': 1, 'inference_mode': False, 'data': {'num_workers': 4, 'shuffle': False}}, 'logging': {'metrics': None}, '_name': 'exp_video'}, 'dataset': {'debug': '${debug}', 'metadata': 'data/${dataset.name}/metadata.json', 'data_mean': 0.5, 'data_std': 0.5, 'save_dir': '/proj/cvl/users/x_fahkh2/WorldMem_Repro/datasets/minecraft', 'n_frames': 300, 'context_length': 1, 'resolution': 128, 'observation_shape': [3, '${dataset.resolution}', '${dataset.resolution}'], 'external_cond_dim': 0, 'validation_multiplier': 1, 'frame_skip': 1, 'action_cond_dim': 25, '_name': 'video_minecraft_latent', 'precomputed_feature_dir': '/proj/cvl/users/x_fahkh2/WorldMem_Repro/datasets/minecraft/vae_features', 'use_explicit_memory_frames': False, 'n_frames_valid': 700, 'angle_range': 110, 'pos_range': 8, 'customized_validation': True, 'add_timestamp_embedding': True}, 'algorithm': {'debug': '${debug}', 'lr': '${experiment.training.lr}', 'x_shape': '${dataset.observation_shape}', 'frame_stack': 1, 'frame_skip': '${dataset.frame_skip}', 'data_mean': '${dataset.data_mean}', 'data_std': '${dataset.data_std}', 'external_cond_dim': 0, 'context_frames': 100, 'weight_decay': 0.002, 'warmup_steps': 1000, 'optimizer_beta': [0.9, 0.99], 'uncertainty_scale': 1, 'guidance_scale': 0.0, 'chunk_size': 1, 'scheduling_matrix': 'autoregressive', 'noise_level': 'random_all', 'causal': True, 'diffusion': {'objective': 'pred_v', 'beta_schedule': 'sigmoid', 'schedule_fn_kwargs': {}, 'clip_noise': 20.0, 'use_snr': False, 'use_cum_snr': False, 'use_fused_snr': True, 'snr_clip': 5.0, 'cum_snr_decay': 0.96, 'timesteps': 1000, 'sampling_timesteps': 20, 'ddim_sampling_eta': 0.0, 'stabilization_level': 15, 'architecture': {'network_size': 64, 'attn_heads': 4, 'attn_dim_head': 64, 'dim_mults': [1, 2, 4, 8], 'resolution': '${dataset.resolution}', 'attn_resolutions': [16, 32, 64, 128], 'use_init_temporal_attn': True, 'use_linear_attn': True, 'time_emb_type': 'rotary'}}, 'n_frames': '${dataset.n_frames}', 'metadata': '${dataset.metadata}', 'action_cond_dim': 25, 'use_plucker': True, 'memory_condition_length': 0, 'log_video': True, 'use_compressed_causal_memory': False, 'compressed_memory_dim': 256, 'compressed_memory_slots': 64, 'compressed_memory_topk': 4, 'use_mamba_memory_pipeline': True, 'training_stage': 'stage_a_memory_pretrain', 'stage_c_memory_aux_weight': 0.1, 'diff_window_size': 8, 'memory_gap_aux_weight': 0.1, 'memory_gap_aux_weight_init': 0.5, 'memory_gap_aux_weight_final': 0.1, 'memory_gap_aux_decay_fraction': 0.3, 'memory_gap_aux_decay_steps': 0, 'memory_gap_aux_exp_k': 5.0, 'use_precomputed_features': True, 'mamba_latent_channels': 16, 'mamba_model_dim': 256, 'mamba_depth': 4, 'mamba_cond_dim': 256, 'mamba_d_state': 16, 'mamba_d_conv': 4, 'mamba_expand': 2, 'allow_mamba_fallback': False, 'strict_causal_training': True, 'strict_causal_evaluation': True, 'use_oracle_pose_eval': False, 'enable_memory_noise_curriculum': True, 'curriculum_phase_boundaries': [0.2, 0.7], 'curriculum_noise_ranges': [[600, 1000], [200, 900], [0, 400]], 'curriculum_horizons': [50, 100, 200], '_name': 'df_video_mamba3stage', 'require_pose_prediction': False, 'use_memory_attention': False, 'relative_embedding': False, 'memory_retrieval_topk': 32}, 'debug': False, 'wandb': {'entity': 'turlin', 'project': 'worldmem', 'mode': 'online'}, 'resume': None, 'load': None, 'name': 'train_stage_a_mamba', 'customized_load': True, 'seperate_load': True, 'diffusion_model_path': '/proj/cvl/users/x_fahkh2/WorldMem_Repro/checkpoints/diffusion_only.ckpt', 'vae_path': '/proj/cvl/users/x_fahkh2/WorldMem_Repro/checkpoints/vae_only.ckpt', 'output_dir': '/proj/cvl/users/x_fahkh2/WorldMem_Repro/checkpoints/bimamba_stage_a_128/'}
14
+ 2026-03-07 23:35:56,648 INFO MainThread:528584 [wandb_init.py:init():658] starting backend
15
+ 2026-03-07 23:35:56,649 INFO MainThread:528584 [wandb_init.py:init():662] setting up manager
16
+ 2026-03-07 23:35:57,002 INFO MainThread:528584 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2026-03-07 23:35:57,009 INFO MainThread:528584 [wandb_init.py:init():670] backend started and connected
18
+ 2026-03-07 23:35:57,030 INFO MainThread:528584 [wandb_init.py:init():768] updated telemetry
19
+ 2026-03-07 23:35:57,079 INFO MainThread:528584 [wandb_init.py:init():801] communicating run to backend with 90.0 second timeout
20
+ 2026-03-07 23:35:57,084 INFO MainThread:528584 [wandb_init.py:init():852] starting run threads in backend
21
+ 2026-03-07 23:36:04,163 INFO MainThread:528584 [wandb_run.py:_console_start():2465] atexit reg
22
+ 2026-03-07 23:36:04,163 INFO MainThread:528584 [wandb_run.py:_redirect():2311] redirect: wrap_raw
23
+ 2026-03-07 23:36:04,163 INFO MainThread:528584 [wandb_run.py:_redirect():2376] Wrapping output streams.
24
+ 2026-03-07 23:36:04,163 INFO MainThread:528584 [wandb_run.py:_redirect():2401] Redirects installed.
25
+ 2026-03-07 23:36:04,203 INFO MainThread:528584 [wandb_init.py:init():895] run started, returning control to user process
26
+ 2026-03-08 01:27:32,382 WARNING MsgRouterThr:528584 [router.py:message_loop():77] message_loop has been closed
bimamba_stage_a_128/wandb/offline-run-20260307_233556-cbyk5tna/files/conda-environment.yaml ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ name: worldmem
2
+ channels:
3
+ - defaults
4
+ prefix: /home/x_fahkh/.conda/envs/worldmem
bimamba_stage_a_128/wandb/offline-run-20260307_233556-cbyk5tna/files/wandb-metadata.json ADDED
@@ -0,0 +1,1226 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.14.0-611.30.1.el9_7.x86_64-x86_64-with-glibc2.34",
3
+ "python": "3.10.19",
4
+ "heartbeatAt": "2026-03-07T22:35:57.228889",
5
+ "startedAt": "2026-03-07T22:35:56.624358",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [
9
+ "+name=train_stage_a_mamba",
10
+ "algorithm=df_video_mamba3stage",
11
+ "+customized_load=true",
12
+ "+seperate_load=true",
13
+ "+diffusion_model_path=/proj/cvl/users/x_fahkh2/WorldMem_Repro/checkpoints/diffusion_only.ckpt",
14
+ "+vae_path=/proj/cvl/users/x_fahkh2/WorldMem_Repro/checkpoints/vae_only.ckpt",
15
+ "dataset=video_minecraft_latent",
16
+ "dataset.precomputed_feature_dir=/proj/cvl/users/x_fahkh2/WorldMem_Repro/datasets/minecraft/vae_features",
17
+ "dataset.save_dir=/proj/cvl/users/x_fahkh2/WorldMem_Repro/datasets/minecraft",
18
+ "dataset.n_frames=300",
19
+ "+dataset.n_frames_valid=700",
20
+ "+dataset.angle_range=110",
21
+ "+dataset.pos_range=8",
22
+ "+dataset.customized_validation=true",
23
+ "+dataset.add_timestamp_embedding=true",
24
+ "dataset.use_explicit_memory_frames=false",
25
+ "algorithm.training_stage=stage_a_memory_pretrain",
26
+ "algorithm.use_mamba_memory_pipeline=true",
27
+ "algorithm.use_oracle_pose_eval=false",
28
+ "algorithm.use_precomputed_features=true",
29
+ "algorithm.context_frames=100",
30
+ "+algorithm.require_pose_prediction=false",
31
+ "+algorithm.use_memory_attention=false",
32
+ "+algorithm.relative_embedding=false",
33
+ "+algorithm.memory_retrieval_topk=32",
34
+ "experiment.training.batch_size=128",
35
+ "experiment.training.checkpointing.every_n_train_steps=50",
36
+ "experiment.training.max_steps=2000",
37
+ "experiment.validation.val_every_n_step=100",
38
+ "+output_dir=/proj/cvl/users/x_fahkh2/WorldMem_Repro/checkpoints/bimamba_stage_a_128/"
39
+ ],
40
+ "state": "running",
41
+ "program": "-m main",
42
+ "codePathLocal": null,
43
+ "git": {
44
+ "remote": "https://huggingface.co/BonanDing/WorldMem_Repro",
45
+ "commit": "1143c3f563b85228610f25319928d29ff73ae0d3"
46
+ },
47
+ "email": "fahad.khan@liu.se",
48
+ "root": "/proj/cvl/users/x_fahkh2/WorldMem_Repro",
49
+ "host": "node112",
50
+ "username": "x_fahkh",
51
+ "executable": "/proj/cvl/users/x_fahkh2/envs/worldmem/bin/python",
52
+ "cpu_count": 112,
53
+ "cpu_count_logical": 224,
54
+ "cpu_freq": {
55
+ "current": 1082.7314464285714,
56
+ "min": 800.0,
57
+ "max": 3800.0
58
+ },
59
+ "cpu_freq_per_core": [
60
+ {
61
+ "current": 3779.768,
62
+ "min": 800.0,
63
+ "max": 3800.0
64
+ },
65
+ {
66
+ "current": 800.0,
67
+ "min": 800.0,
68
+ "max": 3800.0
69
+ },
70
+ {
71
+ "current": 800.0,
72
+ "min": 800.0,
73
+ "max": 3800.0
74
+ },
75
+ {
76
+ "current": 800.0,
77
+ "min": 800.0,
78
+ "max": 3800.0
79
+ },
80
+ {
81
+ "current": 800.0,
82
+ "min": 800.0,
83
+ "max": 3800.0
84
+ },
85
+ {
86
+ "current": 800.0,
87
+ "min": 800.0,
88
+ "max": 3800.0
89
+ },
90
+ {
91
+ "current": 800.0,
92
+ "min": 800.0,
93
+ "max": 3800.0
94
+ },
95
+ {
96
+ "current": 3800.0,
97
+ "min": 800.0,
98
+ "max": 3800.0
99
+ },
100
+ {
101
+ "current": 3799.961,
102
+ "min": 800.0,
103
+ "max": 3800.0
104
+ },
105
+ {
106
+ "current": 800.0,
107
+ "min": 800.0,
108
+ "max": 3800.0
109
+ },
110
+ {
111
+ "current": 800.0,
112
+ "min": 800.0,
113
+ "max": 3800.0
114
+ },
115
+ {
116
+ "current": 800.0,
117
+ "min": 800.0,
118
+ "max": 3800.0
119
+ },
120
+ {
121
+ "current": 800.0,
122
+ "min": 800.0,
123
+ "max": 3800.0
124
+ },
125
+ {
126
+ "current": 800.0,
127
+ "min": 800.0,
128
+ "max": 3800.0
129
+ },
130
+ {
131
+ "current": 3796.597,
132
+ "min": 800.0,
133
+ "max": 3800.0
134
+ },
135
+ {
136
+ "current": 800.0,
137
+ "min": 800.0,
138
+ "max": 3800.0
139
+ },
140
+ {
141
+ "current": 800.0,
142
+ "min": 800.0,
143
+ "max": 3800.0
144
+ },
145
+ {
146
+ "current": 800.0,
147
+ "min": 800.0,
148
+ "max": 3800.0
149
+ },
150
+ {
151
+ "current": 800.0,
152
+ "min": 800.0,
153
+ "max": 3800.0
154
+ },
155
+ {
156
+ "current": 800.0,
157
+ "min": 800.0,
158
+ "max": 3800.0
159
+ },
160
+ {
161
+ "current": 800.0,
162
+ "min": 800.0,
163
+ "max": 3800.0
164
+ },
165
+ {
166
+ "current": 3788.667,
167
+ "min": 800.0,
168
+ "max": 3800.0
169
+ },
170
+ {
171
+ "current": 800.0,
172
+ "min": 800.0,
173
+ "max": 3800.0
174
+ },
175
+ {
176
+ "current": 3792.641,
177
+ "min": 800.0,
178
+ "max": 3800.0
179
+ },
180
+ {
181
+ "current": 800.0,
182
+ "min": 800.0,
183
+ "max": 3800.0
184
+ },
185
+ {
186
+ "current": 800.0,
187
+ "min": 800.0,
188
+ "max": 3800.0
189
+ },
190
+ {
191
+ "current": 3768.911,
192
+ "min": 800.0,
193
+ "max": 3800.0
194
+ },
195
+ {
196
+ "current": 800.0,
197
+ "min": 800.0,
198
+ "max": 3800.0
199
+ },
200
+ {
201
+ "current": 800.0,
202
+ "min": 800.0,
203
+ "max": 3800.0
204
+ },
205
+ {
206
+ "current": 800.0,
207
+ "min": 800.0,
208
+ "max": 3800.0
209
+ },
210
+ {
211
+ "current": 800.0,
212
+ "min": 800.0,
213
+ "max": 3800.0
214
+ },
215
+ {
216
+ "current": 800.0,
217
+ "min": 800.0,
218
+ "max": 3800.0
219
+ },
220
+ {
221
+ "current": 800.0,
222
+ "min": 800.0,
223
+ "max": 3800.0
224
+ },
225
+ {
226
+ "current": 800.0,
227
+ "min": 800.0,
228
+ "max": 3800.0
229
+ },
230
+ {
231
+ "current": 800.0,
232
+ "min": 800.0,
233
+ "max": 3800.0
234
+ },
235
+ {
236
+ "current": 800.0,
237
+ "min": 800.0,
238
+ "max": 3800.0
239
+ },
240
+ {
241
+ "current": 800.0,
242
+ "min": 800.0,
243
+ "max": 3800.0
244
+ },
245
+ {
246
+ "current": 800.0,
247
+ "min": 800.0,
248
+ "max": 3800.0
249
+ },
250
+ {
251
+ "current": 800.0,
252
+ "min": 800.0,
253
+ "max": 3800.0
254
+ },
255
+ {
256
+ "current": 800.0,
257
+ "min": 800.0,
258
+ "max": 3800.0
259
+ },
260
+ {
261
+ "current": 800.0,
262
+ "min": 800.0,
263
+ "max": 3800.0
264
+ },
265
+ {
266
+ "current": 800.0,
267
+ "min": 800.0,
268
+ "max": 3800.0
269
+ },
270
+ {
271
+ "current": 800.0,
272
+ "min": 800.0,
273
+ "max": 3800.0
274
+ },
275
+ {
276
+ "current": 800.0,
277
+ "min": 800.0,
278
+ "max": 3800.0
279
+ },
280
+ {
281
+ "current": 800.0,
282
+ "min": 800.0,
283
+ "max": 3800.0
284
+ },
285
+ {
286
+ "current": 800.0,
287
+ "min": 800.0,
288
+ "max": 3800.0
289
+ },
290
+ {
291
+ "current": 800.0,
292
+ "min": 800.0,
293
+ "max": 3800.0
294
+ },
295
+ {
296
+ "current": 800.0,
297
+ "min": 800.0,
298
+ "max": 3800.0
299
+ },
300
+ {
301
+ "current": 800.0,
302
+ "min": 800.0,
303
+ "max": 3800.0
304
+ },
305
+ {
306
+ "current": 800.0,
307
+ "min": 800.0,
308
+ "max": 3800.0
309
+ },
310
+ {
311
+ "current": 800.0,
312
+ "min": 800.0,
313
+ "max": 3800.0
314
+ },
315
+ {
316
+ "current": 800.0,
317
+ "min": 800.0,
318
+ "max": 3800.0
319
+ },
320
+ {
321
+ "current": 800.0,
322
+ "min": 800.0,
323
+ "max": 3800.0
324
+ },
325
+ {
326
+ "current": 800.0,
327
+ "min": 800.0,
328
+ "max": 3800.0
329
+ },
330
+ {
331
+ "current": 800.0,
332
+ "min": 800.0,
333
+ "max": 3800.0
334
+ },
335
+ {
336
+ "current": 800.0,
337
+ "min": 800.0,
338
+ "max": 3800.0
339
+ },
340
+ {
341
+ "current": 800.0,
342
+ "min": 800.0,
343
+ "max": 3800.0
344
+ },
345
+ {
346
+ "current": 3800.0,
347
+ "min": 800.0,
348
+ "max": 3800.0
349
+ },
350
+ {
351
+ "current": 3800.0,
352
+ "min": 800.0,
353
+ "max": 3800.0
354
+ },
355
+ {
356
+ "current": 3800.038,
357
+ "min": 800.0,
358
+ "max": 3800.0
359
+ },
360
+ {
361
+ "current": 800.0,
362
+ "min": 800.0,
363
+ "max": 3800.0
364
+ },
365
+ {
366
+ "current": 800.0,
367
+ "min": 800.0,
368
+ "max": 3800.0
369
+ },
370
+ {
371
+ "current": 3800.0,
372
+ "min": 800.0,
373
+ "max": 3800.0
374
+ },
375
+ {
376
+ "current": 800.0,
377
+ "min": 800.0,
378
+ "max": 3800.0
379
+ },
380
+ {
381
+ "current": 3800.0,
382
+ "min": 800.0,
383
+ "max": 3800.0
384
+ },
385
+ {
386
+ "current": 800.0,
387
+ "min": 800.0,
388
+ "max": 3800.0
389
+ },
390
+ {
391
+ "current": 800.0,
392
+ "min": 800.0,
393
+ "max": 3800.0
394
+ },
395
+ {
396
+ "current": 800.0,
397
+ "min": 800.0,
398
+ "max": 3800.0
399
+ },
400
+ {
401
+ "current": 800.0,
402
+ "min": 800.0,
403
+ "max": 3800.0
404
+ },
405
+ {
406
+ "current": 800.0,
407
+ "min": 800.0,
408
+ "max": 3800.0
409
+ },
410
+ {
411
+ "current": 800.0,
412
+ "min": 800.0,
413
+ "max": 3800.0
414
+ },
415
+ {
416
+ "current": 800.0,
417
+ "min": 800.0,
418
+ "max": 3800.0
419
+ },
420
+ {
421
+ "current": 800.0,
422
+ "min": 800.0,
423
+ "max": 3800.0
424
+ },
425
+ {
426
+ "current": 800.0,
427
+ "min": 800.0,
428
+ "max": 3800.0
429
+ },
430
+ {
431
+ "current": 800.0,
432
+ "min": 800.0,
433
+ "max": 3800.0
434
+ },
435
+ {
436
+ "current": 800.0,
437
+ "min": 800.0,
438
+ "max": 3800.0
439
+ },
440
+ {
441
+ "current": 800.0,
442
+ "min": 800.0,
443
+ "max": 3800.0
444
+ },
445
+ {
446
+ "current": 800.0,
447
+ "min": 800.0,
448
+ "max": 3800.0
449
+ },
450
+ {
451
+ "current": 800.0,
452
+ "min": 800.0,
453
+ "max": 3800.0
454
+ },
455
+ {
456
+ "current": 800.0,
457
+ "min": 800.0,
458
+ "max": 3800.0
459
+ },
460
+ {
461
+ "current": 3789.534,
462
+ "min": 800.0,
463
+ "max": 3800.0
464
+ },
465
+ {
466
+ "current": 800.0,
467
+ "min": 800.0,
468
+ "max": 3800.0
469
+ },
470
+ {
471
+ "current": 3799.827,
472
+ "min": 800.0,
473
+ "max": 3800.0
474
+ },
475
+ {
476
+ "current": 800.0,
477
+ "min": 800.0,
478
+ "max": 3800.0
479
+ },
480
+ {
481
+ "current": 800.0,
482
+ "min": 800.0,
483
+ "max": 3800.0
484
+ },
485
+ {
486
+ "current": 800.0,
487
+ "min": 800.0,
488
+ "max": 3800.0
489
+ },
490
+ {
491
+ "current": 3267.328,
492
+ "min": 800.0,
493
+ "max": 3800.0
494
+ },
495
+ {
496
+ "current": 800.0,
497
+ "min": 800.0,
498
+ "max": 3800.0
499
+ },
500
+ {
501
+ "current": 3316.432,
502
+ "min": 800.0,
503
+ "max": 3800.0
504
+ },
505
+ {
506
+ "current": 800.0,
507
+ "min": 800.0,
508
+ "max": 3800.0
509
+ },
510
+ {
511
+ "current": 800.0,
512
+ "min": 800.0,
513
+ "max": 3800.0
514
+ },
515
+ {
516
+ "current": 800.0,
517
+ "min": 800.0,
518
+ "max": 3800.0
519
+ },
520
+ {
521
+ "current": 800.0,
522
+ "min": 800.0,
523
+ "max": 3800.0
524
+ },
525
+ {
526
+ "current": 800.0,
527
+ "min": 800.0,
528
+ "max": 3800.0
529
+ },
530
+ {
531
+ "current": 800.0,
532
+ "min": 800.0,
533
+ "max": 3800.0
534
+ },
535
+ {
536
+ "current": 800.0,
537
+ "min": 800.0,
538
+ "max": 3800.0
539
+ },
540
+ {
541
+ "current": 800.0,
542
+ "min": 800.0,
543
+ "max": 3800.0
544
+ },
545
+ {
546
+ "current": 800.0,
547
+ "min": 800.0,
548
+ "max": 3800.0
549
+ },
550
+ {
551
+ "current": 800.0,
552
+ "min": 800.0,
553
+ "max": 3800.0
554
+ },
555
+ {
556
+ "current": 800.0,
557
+ "min": 800.0,
558
+ "max": 3800.0
559
+ },
560
+ {
561
+ "current": 800.0,
562
+ "min": 800.0,
563
+ "max": 3800.0
564
+ },
565
+ {
566
+ "current": 800.0,
567
+ "min": 800.0,
568
+ "max": 3800.0
569
+ },
570
+ {
571
+ "current": 800.0,
572
+ "min": 800.0,
573
+ "max": 3800.0
574
+ },
575
+ {
576
+ "current": 800.0,
577
+ "min": 800.0,
578
+ "max": 3800.0
579
+ },
580
+ {
581
+ "current": 800.0,
582
+ "min": 800.0,
583
+ "max": 3800.0
584
+ },
585
+ {
586
+ "current": 800.0,
587
+ "min": 800.0,
588
+ "max": 3800.0
589
+ },
590
+ {
591
+ "current": 800.0,
592
+ "min": 800.0,
593
+ "max": 3800.0
594
+ },
595
+ {
596
+ "current": 800.0,
597
+ "min": 800.0,
598
+ "max": 3800.0
599
+ },
600
+ {
601
+ "current": 800.0,
602
+ "min": 800.0,
603
+ "max": 3800.0
604
+ },
605
+ {
606
+ "current": 800.0,
607
+ "min": 800.0,
608
+ "max": 3800.0
609
+ },
610
+ {
611
+ "current": 800.0,
612
+ "min": 800.0,
613
+ "max": 3800.0
614
+ },
615
+ {
616
+ "current": 800.0,
617
+ "min": 800.0,
618
+ "max": 3800.0
619
+ },
620
+ {
621
+ "current": 800.0,
622
+ "min": 800.0,
623
+ "max": 3800.0
624
+ },
625
+ {
626
+ "current": 800.0,
627
+ "min": 800.0,
628
+ "max": 3800.0
629
+ },
630
+ {
631
+ "current": 3800.0,
632
+ "min": 800.0,
633
+ "max": 3800.0
634
+ },
635
+ {
636
+ "current": 800.0,
637
+ "min": 800.0,
638
+ "max": 3800.0
639
+ },
640
+ {
641
+ "current": 800.0,
642
+ "min": 800.0,
643
+ "max": 3800.0
644
+ },
645
+ {
646
+ "current": 800.0,
647
+ "min": 800.0,
648
+ "max": 3800.0
649
+ },
650
+ {
651
+ "current": 800.0,
652
+ "min": 800.0,
653
+ "max": 3800.0
654
+ },
655
+ {
656
+ "current": 800.0,
657
+ "min": 800.0,
658
+ "max": 3800.0
659
+ },
660
+ {
661
+ "current": 800.0,
662
+ "min": 800.0,
663
+ "max": 3800.0
664
+ },
665
+ {
666
+ "current": 800.0,
667
+ "min": 800.0,
668
+ "max": 3800.0
669
+ },
670
+ {
671
+ "current": 800.0,
672
+ "min": 800.0,
673
+ "max": 3800.0
674
+ },
675
+ {
676
+ "current": 800.0,
677
+ "min": 800.0,
678
+ "max": 3800.0
679
+ },
680
+ {
681
+ "current": 800.0,
682
+ "min": 800.0,
683
+ "max": 3800.0
684
+ },
685
+ {
686
+ "current": 800.0,
687
+ "min": 800.0,
688
+ "max": 3800.0
689
+ },
690
+ {
691
+ "current": 3800.0,
692
+ "min": 800.0,
693
+ "max": 3800.0
694
+ },
695
+ {
696
+ "current": 800.0,
697
+ "min": 800.0,
698
+ "max": 3800.0
699
+ },
700
+ {
701
+ "current": 3800.038,
702
+ "min": 800.0,
703
+ "max": 3800.0
704
+ },
705
+ {
706
+ "current": 800.0,
707
+ "min": 800.0,
708
+ "max": 3800.0
709
+ },
710
+ {
711
+ "current": 800.0,
712
+ "min": 800.0,
713
+ "max": 3800.0
714
+ },
715
+ {
716
+ "current": 800.0,
717
+ "min": 800.0,
718
+ "max": 3800.0
719
+ },
720
+ {
721
+ "current": 800.0,
722
+ "min": 800.0,
723
+ "max": 3800.0
724
+ },
725
+ {
726
+ "current": 800.0,
727
+ "min": 800.0,
728
+ "max": 3800.0
729
+ },
730
+ {
731
+ "current": 3795.432,
732
+ "min": 800.0,
733
+ "max": 3800.0
734
+ },
735
+ {
736
+ "current": 800.0,
737
+ "min": 800.0,
738
+ "max": 3800.0
739
+ },
740
+ {
741
+ "current": 3795.627,
742
+ "min": 800.0,
743
+ "max": 3800.0
744
+ },
745
+ {
746
+ "current": 3800.0,
747
+ "min": 800.0,
748
+ "max": 3800.0
749
+ },
750
+ {
751
+ "current": 800.0,
752
+ "min": 800.0,
753
+ "max": 3800.0
754
+ },
755
+ {
756
+ "current": 800.0,
757
+ "min": 800.0,
758
+ "max": 3800.0
759
+ },
760
+ {
761
+ "current": 800.0,
762
+ "min": 800.0,
763
+ "max": 3800.0
764
+ },
765
+ {
766
+ "current": 800.0,
767
+ "min": 800.0,
768
+ "max": 3800.0
769
+ },
770
+ {
771
+ "current": 800.0,
772
+ "min": 800.0,
773
+ "max": 3800.0
774
+ },
775
+ {
776
+ "current": 800.0,
777
+ "min": 800.0,
778
+ "max": 3800.0
779
+ },
780
+ {
781
+ "current": 800.0,
782
+ "min": 800.0,
783
+ "max": 3800.0
784
+ },
785
+ {
786
+ "current": 3397.161,
787
+ "min": 800.0,
788
+ "max": 3800.0
789
+ },
790
+ {
791
+ "current": 800.0,
792
+ "min": 800.0,
793
+ "max": 3800.0
794
+ },
795
+ {
796
+ "current": 800.0,
797
+ "min": 800.0,
798
+ "max": 3800.0
799
+ },
800
+ {
801
+ "current": 3800.0,
802
+ "min": 800.0,
803
+ "max": 3800.0
804
+ },
805
+ {
806
+ "current": 800.0,
807
+ "min": 800.0,
808
+ "max": 3800.0
809
+ },
810
+ {
811
+ "current": 3676.848,
812
+ "min": 800.0,
813
+ "max": 3800.0
814
+ },
815
+ {
816
+ "current": 800.0,
817
+ "min": 800.0,
818
+ "max": 3800.0
819
+ },
820
+ {
821
+ "current": 800.0,
822
+ "min": 800.0,
823
+ "max": 3800.0
824
+ },
825
+ {
826
+ "current": 800.0,
827
+ "min": 800.0,
828
+ "max": 3800.0
829
+ },
830
+ {
831
+ "current": 800.0,
832
+ "min": 800.0,
833
+ "max": 3800.0
834
+ },
835
+ {
836
+ "current": 800.0,
837
+ "min": 800.0,
838
+ "max": 3800.0
839
+ },
840
+ {
841
+ "current": 800.0,
842
+ "min": 800.0,
843
+ "max": 3800.0
844
+ },
845
+ {
846
+ "current": 800.0,
847
+ "min": 800.0,
848
+ "max": 3800.0
849
+ },
850
+ {
851
+ "current": 800.0,
852
+ "min": 800.0,
853
+ "max": 3800.0
854
+ },
855
+ {
856
+ "current": 800.0,
857
+ "min": 800.0,
858
+ "max": 3800.0
859
+ },
860
+ {
861
+ "current": 800.0,
862
+ "min": 800.0,
863
+ "max": 3800.0
864
+ },
865
+ {
866
+ "current": 800.0,
867
+ "min": 800.0,
868
+ "max": 3800.0
869
+ },
870
+ {
871
+ "current": 800.0,
872
+ "min": 800.0,
873
+ "max": 3800.0
874
+ },
875
+ {
876
+ "current": 800.0,
877
+ "min": 800.0,
878
+ "max": 3800.0
879
+ },
880
+ {
881
+ "current": 800.0,
882
+ "min": 800.0,
883
+ "max": 3800.0
884
+ },
885
+ {
886
+ "current": 800.0,
887
+ "min": 800.0,
888
+ "max": 3800.0
889
+ },
890
+ {
891
+ "current": 800.0,
892
+ "min": 800.0,
893
+ "max": 3800.0
894
+ },
895
+ {
896
+ "current": 800.0,
897
+ "min": 800.0,
898
+ "max": 3800.0
899
+ },
900
+ {
901
+ "current": 800.0,
902
+ "min": 800.0,
903
+ "max": 3800.0
904
+ },
905
+ {
906
+ "current": 800.0,
907
+ "min": 800.0,
908
+ "max": 3800.0
909
+ },
910
+ {
911
+ "current": 800.0,
912
+ "min": 800.0,
913
+ "max": 3800.0
914
+ },
915
+ {
916
+ "current": 800.0,
917
+ "min": 800.0,
918
+ "max": 3800.0
919
+ },
920
+ {
921
+ "current": 3793.884,
922
+ "min": 800.0,
923
+ "max": 3800.0
924
+ },
925
+ {
926
+ "current": 800.0,
927
+ "min": 800.0,
928
+ "max": 3800.0
929
+ },
930
+ {
931
+ "current": 800.0,
932
+ "min": 800.0,
933
+ "max": 3800.0
934
+ },
935
+ {
936
+ "current": 800.0,
937
+ "min": 800.0,
938
+ "max": 3800.0
939
+ },
940
+ {
941
+ "current": 800.0,
942
+ "min": 800.0,
943
+ "max": 3800.0
944
+ },
945
+ {
946
+ "current": 800.0,
947
+ "min": 800.0,
948
+ "max": 3800.0
949
+ },
950
+ {
951
+ "current": 800.0,
952
+ "min": 800.0,
953
+ "max": 3800.0
954
+ },
955
+ {
956
+ "current": 800.0,
957
+ "min": 800.0,
958
+ "max": 3800.0
959
+ },
960
+ {
961
+ "current": 800.0,
962
+ "min": 800.0,
963
+ "max": 3800.0
964
+ },
965
+ {
966
+ "current": 800.0,
967
+ "min": 800.0,
968
+ "max": 3800.0
969
+ },
970
+ {
971
+ "current": 800.0,
972
+ "min": 800.0,
973
+ "max": 3800.0
974
+ },
975
+ {
976
+ "current": 800.0,
977
+ "min": 800.0,
978
+ "max": 3800.0
979
+ },
980
+ {
981
+ "current": 800.0,
982
+ "min": 800.0,
983
+ "max": 3800.0
984
+ },
985
+ {
986
+ "current": 800.0,
987
+ "min": 800.0,
988
+ "max": 3800.0
989
+ },
990
+ {
991
+ "current": 800.0,
992
+ "min": 800.0,
993
+ "max": 3800.0
994
+ },
995
+ {
996
+ "current": 800.0,
997
+ "min": 800.0,
998
+ "max": 3800.0
999
+ },
1000
+ {
1001
+ "current": 800.0,
1002
+ "min": 800.0,
1003
+ "max": 3800.0
1004
+ },
1005
+ {
1006
+ "current": 800.0,
1007
+ "min": 800.0,
1008
+ "max": 3800.0
1009
+ },
1010
+ {
1011
+ "current": 800.0,
1012
+ "min": 800.0,
1013
+ "max": 3800.0
1014
+ },
1015
+ {
1016
+ "current": 3800.0,
1017
+ "min": 800.0,
1018
+ "max": 3800.0
1019
+ },
1020
+ {
1021
+ "current": 800.0,
1022
+ "min": 800.0,
1023
+ "max": 3800.0
1024
+ },
1025
+ {
1026
+ "current": 800.0,
1027
+ "min": 800.0,
1028
+ "max": 3800.0
1029
+ },
1030
+ {
1031
+ "current": 3800.0,
1032
+ "min": 800.0,
1033
+ "max": 3800.0
1034
+ },
1035
+ {
1036
+ "current": 800.0,
1037
+ "min": 800.0,
1038
+ "max": 3800.0
1039
+ },
1040
+ {
1041
+ "current": 800.0,
1042
+ "min": 800.0,
1043
+ "max": 3800.0
1044
+ },
1045
+ {
1046
+ "current": 800.0,
1047
+ "min": 800.0,
1048
+ "max": 3800.0
1049
+ },
1050
+ {
1051
+ "current": 800.0,
1052
+ "min": 800.0,
1053
+ "max": 3800.0
1054
+ },
1055
+ {
1056
+ "current": 3800.0,
1057
+ "min": 800.0,
1058
+ "max": 3800.0
1059
+ },
1060
+ {
1061
+ "current": 800.0,
1062
+ "min": 800.0,
1063
+ "max": 3800.0
1064
+ },
1065
+ {
1066
+ "current": 800.0,
1067
+ "min": 800.0,
1068
+ "max": 3800.0
1069
+ },
1070
+ {
1071
+ "current": 800.0,
1072
+ "min": 800.0,
1073
+ "max": 3800.0
1074
+ },
1075
+ {
1076
+ "current": 800.0,
1077
+ "min": 800.0,
1078
+ "max": 3800.0
1079
+ },
1080
+ {
1081
+ "current": 800.0,
1082
+ "min": 800.0,
1083
+ "max": 3800.0
1084
+ },
1085
+ {
1086
+ "current": 800.0,
1087
+ "min": 800.0,
1088
+ "max": 3800.0
1089
+ },
1090
+ {
1091
+ "current": 800.0,
1092
+ "min": 800.0,
1093
+ "max": 3800.0
1094
+ },
1095
+ {
1096
+ "current": 800.0,
1097
+ "min": 800.0,
1098
+ "max": 3800.0
1099
+ },
1100
+ {
1101
+ "current": 800.0,
1102
+ "min": 800.0,
1103
+ "max": 3800.0
1104
+ },
1105
+ {
1106
+ "current": 800.0,
1107
+ "min": 800.0,
1108
+ "max": 3800.0
1109
+ },
1110
+ {
1111
+ "current": 800.0,
1112
+ "min": 800.0,
1113
+ "max": 3800.0
1114
+ },
1115
+ {
1116
+ "current": 800.0,
1117
+ "min": 800.0,
1118
+ "max": 3800.0
1119
+ },
1120
+ {
1121
+ "current": 3572.0,
1122
+ "min": 800.0,
1123
+ "max": 3800.0
1124
+ },
1125
+ {
1126
+ "current": 800.0,
1127
+ "min": 800.0,
1128
+ "max": 3800.0
1129
+ },
1130
+ {
1131
+ "current": 800.0,
1132
+ "min": 800.0,
1133
+ "max": 3800.0
1134
+ },
1135
+ {
1136
+ "current": 800.0,
1137
+ "min": 800.0,
1138
+ "max": 3800.0
1139
+ },
1140
+ {
1141
+ "current": 800.0,
1142
+ "min": 800.0,
1143
+ "max": 3800.0
1144
+ },
1145
+ {
1146
+ "current": 800.0,
1147
+ "min": 800.0,
1148
+ "max": 3800.0
1149
+ },
1150
+ {
1151
+ "current": 800.0,
1152
+ "min": 800.0,
1153
+ "max": 3800.0
1154
+ },
1155
+ {
1156
+ "current": 800.0,
1157
+ "min": 800.0,
1158
+ "max": 3800.0
1159
+ },
1160
+ {
1161
+ "current": 800.0,
1162
+ "min": 800.0,
1163
+ "max": 3800.0
1164
+ },
1165
+ {
1166
+ "current": 800.0,
1167
+ "min": 800.0,
1168
+ "max": 3800.0
1169
+ },
1170
+ {
1171
+ "current": 800.0,
1172
+ "min": 800.0,
1173
+ "max": 3800.0
1174
+ },
1175
+ {
1176
+ "current": 800.0,
1177
+ "min": 800.0,
1178
+ "max": 3800.0
1179
+ }
1180
+ ],
1181
+ "disk": {
1182
+ "/": {
1183
+ "total": 30.58770751953125,
1184
+ "used": 16.829898834228516
1185
+ }
1186
+ },
1187
+ "gpu": "NVIDIA H200",
1188
+ "gpu_count": 8,
1189
+ "gpu_devices": [
1190
+ {
1191
+ "name": "NVIDIA H200",
1192
+ "memory_total": 150754820096
1193
+ },
1194
+ {
1195
+ "name": "NVIDIA H200",
1196
+ "memory_total": 150754820096
1197
+ },
1198
+ {
1199
+ "name": "NVIDIA H200",
1200
+ "memory_total": 150754820096
1201
+ },
1202
+ {
1203
+ "name": "NVIDIA H200",
1204
+ "memory_total": 150754820096
1205
+ },
1206
+ {
1207
+ "name": "NVIDIA H200",
1208
+ "memory_total": 150754820096
1209
+ },
1210
+ {
1211
+ "name": "NVIDIA H200",
1212
+ "memory_total": 150754820096
1213
+ },
1214
+ {
1215
+ "name": "NVIDIA H200",
1216
+ "memory_total": 150754820096
1217
+ },
1218
+ {
1219
+ "name": "NVIDIA H200",
1220
+ "memory_total": 150754820096
1221
+ }
1222
+ ],
1223
+ "memory": {
1224
+ "total": 2015.5592918395996
1225
+ }
1226
+ }
bimamba_stage_a_128/wandb/offline-run-20260307_233556-cbyk5tna/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_wandb": {"runtime": 6693}}
bimamba_stage_a_128/wandb/offline-run-20260307_233556-cbyk5tna/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
 
bimamba_stage_a_128/wandb/offline-run-20260307_233556-cbyk5tna/logs/debug.log ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2026-03-07 23:35:56,648 INFO MainThread:528584 [wandb_setup.py:_flush():77] Current SDK version is 0.17.9
2
+ 2026-03-07 23:35:56,648 INFO MainThread:528584 [wandb_setup.py:_flush():77] Configure stats pid to 528584
3
+ 2026-03-07 23:35:56,648 INFO MainThread:528584 [wandb_setup.py:_flush():77] Loading settings from /home/x_fahkh/.config/wandb/settings
4
+ 2026-03-07 23:35:56,648 INFO MainThread:528584 [wandb_setup.py:_flush():77] Loading settings from /proj/cvl/users/x_fahkh2/WorldMem_Repro/wandb/settings
5
+ 2026-03-07 23:35:56,648 INFO MainThread:528584 [wandb_setup.py:_flush():77] Loading settings from environment variables: {'disabled': 'true'}
6
+ 2026-03-07 23:35:56,648 INFO MainThread:528584 [wandb_setup.py:_flush():77] Applying setup settings: {'_disable_service': False}
7
+ 2026-03-07 23:35:56,648 WARNING MainThread:528584 [wandb_setup.py:_flush():77] Could not find program at -m main
8
+ 2026-03-07 23:35:56,648 INFO MainThread:528584 [wandb_setup.py:_flush():77] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m main'}
9
+ 2026-03-07 23:35:56,648 INFO MainThread:528584 [wandb_init.py:_log_setup():524] Logging user logs to /proj/cvl/users/x_fahkh2/WorldMem_Repro/checkpoints/bimamba_stage_a_128/wandb/offline-run-20260307_233556-cbyk5tna/logs/debug.log
10
+ 2026-03-07 23:35:56,648 INFO MainThread:528584 [wandb_init.py:_log_setup():525] Logging internal logs to /proj/cvl/users/x_fahkh2/WorldMem_Repro/checkpoints/bimamba_stage_a_128/wandb/offline-run-20260307_233556-cbyk5tna/logs/debug-internal.log
11
+ 2026-03-07 23:35:56,648 INFO MainThread:528584 [wandb_init.py:init():608] calling init triggers
12
+ 2026-03-07 23:35:56,648 INFO MainThread:528584 [wandb_init.py:init():615] wandb.init called with sweep_config: {}
13
+ config: {'experiment': {'debug': '${debug}', 'tasks': ['training'], 'num_nodes': 1, 'training': {'precision': '16-mixed', 'compile': False, 'lr': 0.0002, 'batch_size': 128, 'max_epochs': -1, 'max_steps': 2000, 'max_time': None, 'data': {'num_workers': 4, 'shuffle': True}, 'optim': {'accumulate_grad_batches': 1, 'gradient_clip_val': 1.0}, 'checkpointing': {'every_n_train_steps': 50, 'every_n_epochs': None, 'train_time_interval': None, 'enable_version_counter': False}}, 'validation': {'precision': '16-mixed', 'compile': False, 'batch_size': 4, 'val_every_n_step': 100, 'val_every_n_epoch': None, 'limit_batch': 1, 'inference_mode': False, 'data': {'num_workers': 4, 'shuffle': False}}, 'test': {'precision': '16-mixed', 'compile': False, 'batch_size': 1, 'limit_batch': 1, 'inference_mode': False, 'data': {'num_workers': 4, 'shuffle': False}}, 'logging': {'metrics': None}, '_name': 'exp_video'}, 'dataset': {'debug': '${debug}', 'metadata': 'data/${dataset.name}/metadata.json', 'data_mean': 0.5, 'data_std': 0.5, 'save_dir': '/proj/cvl/users/x_fahkh2/WorldMem_Repro/datasets/minecraft', 'n_frames': 300, 'context_length': 1, 'resolution': 128, 'observation_shape': [3, '${dataset.resolution}', '${dataset.resolution}'], 'external_cond_dim': 0, 'validation_multiplier': 1, 'frame_skip': 1, 'action_cond_dim': 25, '_name': 'video_minecraft_latent', 'precomputed_feature_dir': '/proj/cvl/users/x_fahkh2/WorldMem_Repro/datasets/minecraft/vae_features', 'use_explicit_memory_frames': False, 'n_frames_valid': 700, 'angle_range': 110, 'pos_range': 8, 'customized_validation': True, 'add_timestamp_embedding': True}, 'algorithm': {'debug': '${debug}', 'lr': '${experiment.training.lr}', 'x_shape': '${dataset.observation_shape}', 'frame_stack': 1, 'frame_skip': '${dataset.frame_skip}', 'data_mean': '${dataset.data_mean}', 'data_std': '${dataset.data_std}', 'external_cond_dim': 0, 'context_frames': 100, 'weight_decay': 0.002, 'warmup_steps': 1000, 'optimizer_beta': [0.9, 0.99], 'uncertainty_scale': 1, 'guidance_scale': 0.0, 'chunk_size': 1, 'scheduling_matrix': 'autoregressive', 'noise_level': 'random_all', 'causal': True, 'diffusion': {'objective': 'pred_v', 'beta_schedule': 'sigmoid', 'schedule_fn_kwargs': {}, 'clip_noise': 20.0, 'use_snr': False, 'use_cum_snr': False, 'use_fused_snr': True, 'snr_clip': 5.0, 'cum_snr_decay': 0.96, 'timesteps': 1000, 'sampling_timesteps': 20, 'ddim_sampling_eta': 0.0, 'stabilization_level': 15, 'architecture': {'network_size': 64, 'attn_heads': 4, 'attn_dim_head': 64, 'dim_mults': [1, 2, 4, 8], 'resolution': '${dataset.resolution}', 'attn_resolutions': [16, 32, 64, 128], 'use_init_temporal_attn': True, 'use_linear_attn': True, 'time_emb_type': 'rotary'}}, 'n_frames': '${dataset.n_frames}', 'metadata': '${dataset.metadata}', 'action_cond_dim': 25, 'use_plucker': True, 'memory_condition_length': 0, 'log_video': True, 'use_compressed_causal_memory': False, 'compressed_memory_dim': 256, 'compressed_memory_slots': 64, 'compressed_memory_topk': 4, 'use_mamba_memory_pipeline': True, 'training_stage': 'stage_a_memory_pretrain', 'stage_c_memory_aux_weight': 0.1, 'diff_window_size': 8, 'memory_gap_aux_weight': 0.1, 'memory_gap_aux_weight_init': 0.5, 'memory_gap_aux_weight_final': 0.1, 'memory_gap_aux_decay_fraction': 0.3, 'memory_gap_aux_decay_steps': 0, 'memory_gap_aux_exp_k': 5.0, 'use_precomputed_features': True, 'mamba_latent_channels': 16, 'mamba_model_dim': 256, 'mamba_depth': 4, 'mamba_cond_dim': 256, 'mamba_d_state': 16, 'mamba_d_conv': 4, 'mamba_expand': 2, 'allow_mamba_fallback': False, 'strict_causal_training': True, 'strict_causal_evaluation': True, 'use_oracle_pose_eval': False, 'enable_memory_noise_curriculum': True, 'curriculum_phase_boundaries': [0.2, 0.7], 'curriculum_noise_ranges': [[600, 1000], [200, 900], [0, 400]], 'curriculum_horizons': [50, 100, 200], '_name': 'df_video_mamba3stage', 'require_pose_prediction': False, 'use_memory_attention': False, 'relative_embedding': False, 'memory_retrieval_topk': 32}, 'debug': False, 'wandb': {'entity': 'turlin', 'project': 'worldmem', 'mode': 'online'}, 'resume': None, 'load': None, 'name': 'train_stage_a_mamba', 'customized_load': True, 'seperate_load': True, 'diffusion_model_path': '/proj/cvl/users/x_fahkh2/WorldMem_Repro/checkpoints/diffusion_only.ckpt', 'vae_path': '/proj/cvl/users/x_fahkh2/WorldMem_Repro/checkpoints/vae_only.ckpt', 'output_dir': '/proj/cvl/users/x_fahkh2/WorldMem_Repro/checkpoints/bimamba_stage_a_128/'}
14
+ 2026-03-07 23:35:56,648 INFO MainThread:528584 [wandb_init.py:init():658] starting backend
15
+ 2026-03-07 23:35:56,649 INFO MainThread:528584 [wandb_init.py:init():662] setting up manager
16
+ 2026-03-07 23:35:57,002 INFO MainThread:528584 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2026-03-07 23:35:57,009 INFO MainThread:528584 [wandb_init.py:init():670] backend started and connected
18
+ 2026-03-07 23:35:57,030 INFO MainThread:528584 [wandb_init.py:init():768] updated telemetry
19
+ 2026-03-07 23:35:57,079 INFO MainThread:528584 [wandb_init.py:init():801] communicating run to backend with 90.0 second timeout
20
+ 2026-03-07 23:35:57,084 INFO MainThread:528584 [wandb_init.py:init():852] starting run threads in backend
21
+ 2026-03-07 23:36:04,163 INFO MainThread:528584 [wandb_run.py:_console_start():2465] atexit reg
22
+ 2026-03-07 23:36:04,163 INFO MainThread:528584 [wandb_run.py:_redirect():2311] redirect: wrap_raw
23
+ 2026-03-07 23:36:04,163 INFO MainThread:528584 [wandb_run.py:_redirect():2376] Wrapping output streams.
24
+ 2026-03-07 23:36:04,163 INFO MainThread:528584 [wandb_run.py:_redirect():2401] Redirects installed.
25
+ 2026-03-07 23:36:04,203 INFO MainThread:528584 [wandb_init.py:init():895] run started, returning control to user process
26
+ 2026-03-08 01:27:32,382 WARNING MsgRouterThr:528584 [router.py:message_loop():77] message_loop has been closed
bimamba_stage_a_128/wandb/offline-run-20260307_233556-cbyk5tna/run-cbyk5tna.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8bef918963d5729e18bc447ee4bbfbacb95c879101cd84cd5a4c90f7ea14ef1d
3
+ size 2652929
bimamba_stage_a_128/wandb/wandb-resume.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"run_id": "cbyk5tna"}