Knowing commited on
Commit
8f1e9f1
·
verified ·
1 Parent(s): 887442f

Add files using upload-large-folder tool

Browse files
Files changed (50) hide show
  1. .hydra/config.yaml +205 -0
  2. .hydra/hydra.yaml +164 -0
  3. .hydra/overrides.yaml +3 -0
  4. main.log +262 -0
  5. train_ddp_process_1.log +81 -0
  6. train_ddp_process_2.log +81 -0
  7. train_ddp_process_3.log +81 -0
  8. train_ddp_process_4.log +81 -0
  9. train_ddp_process_5.log +81 -0
  10. train_ddp_process_6.log +81 -0
  11. train_ddp_process_7.log +81 -0
  12. wandb/debug-internal.log +7 -0
  13. wandb/debug.log +22 -0
  14. wandb/run-20250916_175907-b2gr1gxl/run-b2gr1gxl.wandb +0 -0
  15. wandb/run-20250916_180036-llem5odx/files/config.yaml +423 -0
  16. wandb/run-20250916_180036-llem5odx/files/output.log +40 -0
  17. wandb/run-20250916_180036-llem5odx/files/requirements.txt +129 -0
  18. wandb/run-20250916_180036-llem5odx/files/wandb-metadata.json +79 -0
  19. wandb/run-20250916_180036-llem5odx/files/wandb-summary.json +1 -0
  20. wandb/run-20250916_180036-llem5odx/logs/debug.log +0 -0
  21. wandb/run-20250916_180036-llem5odx/run-llem5odx.wandb +0 -0
  22. wandb/run-20250916_180413-d4pobtwb/files/output.log +83 -0
  23. wandb/run-20250916_180413-d4pobtwb/files/requirements.txt +129 -0
  24. wandb/run-20250916_180413-d4pobtwb/files/wandb-metadata.json +79 -0
  25. wandb/run-20250916_180413-d4pobtwb/logs/debug-core.log +7 -0
  26. wandb/run-20250916_180413-d4pobtwb/logs/debug-internal.log +7 -0
  27. wandb/run-20250916_180413-d4pobtwb/logs/debug.log +22 -0
  28. wandb/run-20250916_180413-d4pobtwb/run-d4pobtwb.wandb +0 -0
  29. wandb/run-20250916_180427-9sp0azf6/run-9sp0azf6.wandb +0 -0
  30. wandb/run-20250916_180450-bbofdbql/run-bbofdbql.wandb +0 -0
  31. wandb/run-20250916_180500-j938t4jj/run-j938t4jj.wandb +0 -0
  32. wandb/run-20250916_194552-aaciqyuf/files/config.yaml +555 -0
  33. wandb/run-20250916_194552-aaciqyuf/files/media/images/activation_visualization_101_8a45d9dc83c5e5caae0d.png +0 -0
  34. wandb/run-20250916_194552-aaciqyuf/files/media/images/activation_visualization_126_948dc1a285a7709661c2.png +0 -0
  35. wandb/run-20250916_194552-aaciqyuf/files/media/images/activation_visualization_151_b777ce762fa8abbb2f16.png +0 -0
  36. wandb/run-20250916_194552-aaciqyuf/files/media/images/activation_visualization_176_9360fd68ec926bfb2b15.png +0 -0
  37. wandb/run-20250916_194552-aaciqyuf/files/media/images/activation_visualization_201_84928b0fa7964ea26925.png +0 -0
  38. wandb/run-20250916_194552-aaciqyuf/files/media/images/activation_visualization_226_9b48f7ee732928ccde69.png +0 -0
  39. wandb/run-20250916_194552-aaciqyuf/files/media/images/activation_visualization_251_cde359718086fb0623f7.png +0 -0
  40. wandb/run-20250916_194552-aaciqyuf/files/media/images/activation_visualization_26_c73506410e7d6e1d35fd.png +0 -0
  41. wandb/run-20250916_194552-aaciqyuf/files/media/images/activation_visualization_276_cc0deded106bd8c609c8.png +0 -0
  42. wandb/run-20250916_194552-aaciqyuf/files/media/images/activation_visualization_51_e5100843b701f67b204b.png +0 -0
  43. wandb/run-20250916_194552-aaciqyuf/files/media/images/activation_visualization_76_048e6651bc117aff8719.png +0 -0
  44. wandb/run-20250916_194552-aaciqyuf/files/output.log +0 -0
  45. wandb/run-20250916_194552-aaciqyuf/files/requirements.txt +129 -0
  46. wandb/run-20250916_194552-aaciqyuf/files/wandb-metadata.json +79 -0
  47. wandb/run-20250916_194552-aaciqyuf/files/wandb-summary.json +1 -0
  48. wandb/run-20250916_194552-aaciqyuf/logs/debug-core.log +13 -0
  49. wandb/run-20250916_194552-aaciqyuf/logs/debug-internal.log +15 -0
  50. wandb/run-20250916_194552-aaciqyuf/logs/debug.log +364 -0
.hydra/config.yaml ADDED
@@ -0,0 +1,205 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ encoder:
3
+ backbone:
4
+ name: croco
5
+ model: ViTLarge_BaseDecoder
6
+ patch_embed_cls: PatchEmbedDust3R
7
+ asymmetry_decoder: true
8
+ intrinsics_embed_loc: encoder
9
+ intrinsics_embed_degree: 4
10
+ intrinsics_embed_type: token
11
+ name: noposplat
12
+ opacity_mapping:
13
+ initial: 0.0
14
+ final: 0.0
15
+ warm_up: 1
16
+ num_monocular_samples: 32
17
+ num_surfaces: 1
18
+ predict_opacity: false
19
+ gaussians_per_pixel: 1
20
+ gaussian_adapter:
21
+ gaussian_scale_min: 0.5
22
+ gaussian_scale_max: 15.0
23
+ sh_degree: 4
24
+ d_feature: 128
25
+ visualizer:
26
+ num_samples: 8
27
+ min_resolution: 256
28
+ export_ply: false
29
+ apply_bounds_shim: true
30
+ gs_params_head_type: dpt_gs
31
+ pose_free: true
32
+ pretrained_weights: ''
33
+ decoder:
34
+ name: splatting_cuda
35
+ background_color:
36
+ - 0.0
37
+ - 0.0
38
+ - 0.0
39
+ make_scale_invariant: false
40
+ use_gsplat: true
41
+ vggt:
42
+ input_image_shape:
43
+ - 518
44
+ - 518
45
+ head_mode: pcd
46
+ same_head: false
47
+ camera_prompt: false
48
+ num_visual_prompt: 0
49
+ use_extrinsic_token: false
50
+ use_intrinsic_token: false
51
+ detr:
52
+ name: deformable_detr
53
+ features_num_channel:
54
+ - 128
55
+ - 32
56
+ hidden_dim: 32
57
+ dim_down_factor: 1
58
+ dim_feedforward_factor: 4
59
+ enc_layers: 6
60
+ dec_layers: 1
61
+ dropout: 0.1
62
+ num_feature_levels: 2
63
+ dec_n_points: 4
64
+ enc_n_points: 4
65
+ split_threshold: 1.0
66
+ num_queries: 256
67
+ base_HW: 64
68
+ position_embedding: sine
69
+ pre_norm: false
70
+ self_layer_num: 1
71
+ num_level: 3
72
+ debug_mode: 0
73
+ decoder_type: conv
74
+ split_threshold_range:
75
+ - 0.0
76
+ - 2.0
77
+ only_last_level: false
78
+ grad_mode: absgrad
79
+ same_split_head: false
80
+ use_mean_features: true
81
+ loss:
82
+ mse:
83
+ weight: 1.0
84
+ lpips:
85
+ weight: 0.05
86
+ apply_after_step: 0
87
+ wandb:
88
+ project: noposplat
89
+ entity: scene-representation-group
90
+ name: re10k_hpc
91
+ mode: online
92
+ tags:
93
+ - re10k
94
+ - 256x256
95
+ mode: train
96
+ data_loader:
97
+ train:
98
+ num_workers: 16
99
+ persistent_workers: true
100
+ batch_size: 16
101
+ seed: 1234
102
+ test:
103
+ num_workers: 4
104
+ persistent_workers: false
105
+ batch_size: 1
106
+ seed: 2345
107
+ val:
108
+ num_workers: 1
109
+ persistent_workers: true
110
+ batch_size: 1
111
+ seed: 3456
112
+ optimizer:
113
+ lr: 0.0002
114
+ warm_up_steps: 125
115
+ backbone_lr_multiplier: 0.1
116
+ backbone_trainable: GB+H
117
+ accumulate: 1
118
+ checkpointing:
119
+ load: null
120
+ every_n_train_steps: 9375
121
+ save_top_k: 1
122
+ save_weights_only: true
123
+ train:
124
+ depth_mode: null
125
+ extended_visualization: false
126
+ print_log_every_n_steps: 10
127
+ distiller: ''
128
+ distill_max_steps: 1000000
129
+ debug_mode: 0
130
+ camera_consis_loss: 0.0
131
+ camera_loss: 1.0
132
+ aux_loss: 0.0
133
+ one_sample_train: false
134
+ target_scale_mode: 2
135
+ extrinsic_drop: 0.1
136
+ intrinsic_drop: 0.1
137
+ align_corners: false
138
+ backbone_checkpoint: none
139
+ intrinsic_scaling: true
140
+ return_level_gs: true
141
+ train_context: false
142
+ use_pred_point_focal: false
143
+ scaling_detach: true
144
+ split_score_log_scale: true
145
+ split_score_grad_scale: 10000.0
146
+ split_loss: 0.01
147
+ random_split_strategy: score
148
+ test:
149
+ output_path: test
150
+ align_pose: false
151
+ pose_align_steps: 100
152
+ rot_opt_lr: 0.005
153
+ trans_opt_lr: 0.005
154
+ compute_scores: true
155
+ save_image: false
156
+ save_video: false
157
+ save_compare: false
158
+ split_threshold_list:
159
+ - 0.0
160
+ - 0.1
161
+ - 0.2
162
+ - 0.3
163
+ - 0.4
164
+ pred_intrinsic: false
165
+ seed: 111123
166
+ trainer:
167
+ max_steps: 18751
168
+ val_check_interval: 500
169
+ gradient_clip_val: 0.5
170
+ num_nodes: 1
171
+ dataset:
172
+ re10k:
173
+ make_baseline_1: true
174
+ relative_pose: true
175
+ augment: true
176
+ background_color:
177
+ - 0.0
178
+ - 0.0
179
+ - 0.0
180
+ overfit_to_scene: null
181
+ skip_bad_shape: true
182
+ view_sampler:
183
+ name: bounded
184
+ num_target_views: 4
185
+ num_context_views: 2
186
+ min_distance_between_context_views: 45
187
+ max_distance_between_context_views: 90
188
+ min_distance_to_context_views: 0
189
+ warm_up_steps: 9375
190
+ initial_min_distance_between_context_views: 25
191
+ initial_max_distance_between_context_views: 25
192
+ distill_steps: 0
193
+ name: re10k
194
+ roots:
195
+ - datasets/re10k
196
+ input_image_shape:
197
+ - 256
198
+ - 256
199
+ original_image_shape:
200
+ - 360
201
+ - 640
202
+ cameras_are_circular: false
203
+ baseline_min: 0.001
204
+ baseline_max: 10000000000.0
205
+ max_fov: 100.0
.hydra/hydra.yaml ADDED
@@ -0,0 +1,164 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/${wandb.name}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - +experiment=re10k
116
+ - wandb.mode=online
117
+ - wandb.name=re10k_hpc
118
+ job:
119
+ name: main
120
+ chdir: null
121
+ override_dirname: +experiment=re10k,wandb.mode=online,wandb.name=re10k_hpc
122
+ id: ???
123
+ num: ???
124
+ config_name: main
125
+ env_set: {}
126
+ env_copy: []
127
+ config:
128
+ override_dirname:
129
+ kv_sep: '='
130
+ item_sep: ','
131
+ exclude_keys: []
132
+ runtime:
133
+ version: 1.3.2
134
+ version_base: '1.3'
135
+ cwd: /root/highspeedstorage/injae/code/ICLR2026_HCP
136
+ config_sources:
137
+ - path: hydra.conf
138
+ schema: pkg
139
+ provider: hydra
140
+ - path: /root/highspeedstorage/injae/code/ICLR2026_HCP/config
141
+ schema: file
142
+ provider: main
143
+ - path: ''
144
+ schema: structured
145
+ provider: schema
146
+ output_dir: /root/highspeedstorage/injae/code/ICLR2026_HCP/outputs/re10k_hpc
147
+ choices:
148
+ experiment: re10k
149
+ dataset@dataset.re10k: re10k
150
+ dataset/view_sampler_dataset_specific_config@dataset.re10k.view_sampler: bounded_re10k
151
+ dataset/view_sampler@dataset.re10k.view_sampler: bounded
152
+ model/decoder: splatting_cuda
153
+ model/encoder: noposplat
154
+ model/encoder/backbone: croco
155
+ hydra/env: default
156
+ hydra/callbacks: null
157
+ hydra/job_logging: default
158
+ hydra/hydra_logging: default
159
+ hydra/hydra_help: default
160
+ hydra/help: default
161
+ hydra/sweeper: basic
162
+ hydra/launcher: basic
163
+ hydra/output: default
164
+ verbose: false
.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - +experiment=re10k
2
+ - wandb.mode=online
3
+ - wandb.name=re10k_hpc
main.log ADDED
@@ -0,0 +1,262 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-09-16 17:58:02,195][dinov2][INFO] - using MLP layer as FFN
2
+ [2025-09-16 17:58:08,747][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
3
+ warnings.warn(
4
+
5
+ [2025-09-16 17:58:08,748][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
6
+ warnings.warn(msg)
7
+
8
+ [2025-09-16 17:58:50,488][dinov2][INFO] - using MLP layer as FFN
9
+ [2025-09-16 17:58:56,811][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
10
+ warnings.warn(
11
+
12
+ [2025-09-16 17:58:56,812][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
13
+ warnings.warn(msg)
14
+
15
+ [2025-09-16 18:00:28,507][dinov2][INFO] - using MLP layer as FFN
16
+ [2025-09-16 18:00:34,706][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
17
+ warnings.warn(
18
+
19
+ [2025-09-16 18:00:34,708][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
20
+ warnings.warn(msg)
21
+
22
+ [2025-09-16 18:00:39,175][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=279` in the `DataLoader` to improve performance.
23
+
24
+ [2025-09-16 18:00:41,384][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
25
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
26
+ warnings.warn(
27
+
28
+ [2025-09-16 18:01:31,562][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
29
+
30
+ [2025-09-16 18:01:31,563][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
31
+ warnings.warn(
32
+
33
+ [2025-09-16 18:01:31,564][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
34
+ warnings.warn(msg)
35
+
36
+ [2025-09-16 18:01:33,246][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
37
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
38
+
39
+ [2025-09-16 18:02:35,041][dinov2][INFO] - using MLP layer as FFN
40
+ [2025-09-16 18:02:41,051][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
41
+ warnings.warn(
42
+
43
+ [2025-09-16 18:02:41,052][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
44
+ warnings.warn(msg)
45
+
46
+ [2025-09-16 18:04:06,193][dinov2][INFO] - using MLP layer as FFN
47
+ [2025-09-16 18:04:11,843][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
48
+ warnings.warn(
49
+
50
+ [2025-09-16 18:04:11,844][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
51
+ warnings.warn(msg)
52
+
53
+ [2025-09-16 18:04:15,683][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=279` in the `DataLoader` to improve performance.
54
+
55
+ [2025-09-16 18:04:17,971][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
56
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
57
+ warnings.warn(
58
+
59
+ [2025-09-16 18:04:18,161][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
60
+
61
+ [2025-09-16 18:04:18,163][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
62
+ warnings.warn(
63
+
64
+ [2025-09-16 18:04:18,164][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
65
+ warnings.warn(msg)
66
+
67
+ [2025-09-16 18:04:19,828][dinov2][INFO] - using MLP layer as FFN
68
+ [2025-09-16 18:04:19,836][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
69
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
70
+
71
+ [2025-09-16 18:04:25,651][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
72
+ warnings.warn(
73
+
74
+ [2025-09-16 18:04:25,651][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
75
+ warnings.warn(msg)
76
+
77
+ [2025-09-16 18:04:29,494][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=279` in the `DataLoader` to improve performance.
78
+
79
+ [2025-09-16 18:04:42,154][dinov2][INFO] - using MLP layer as FFN
80
+ [2025-09-16 18:04:48,178][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
81
+ warnings.warn(
82
+
83
+ [2025-09-16 18:04:48,179][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
84
+ warnings.warn(msg)
85
+
86
+ [2025-09-16 18:04:52,770][dinov2][INFO] - using MLP layer as FFN
87
+ [2025-09-16 18:04:52,934][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=279` in the `DataLoader` to improve performance.
88
+
89
+ [2025-09-16 18:04:55,463][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
90
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
91
+ warnings.warn(
92
+
93
+ [2025-09-16 18:04:55,703][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
94
+
95
+ [2025-09-16 18:04:55,705][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
96
+ warnings.warn(
97
+
98
+ [2025-09-16 18:04:55,706][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
99
+ warnings.warn(msg)
100
+
101
+ [2025-09-16 18:04:57,296][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
102
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
103
+
104
+ [2025-09-16 18:04:58,729][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
105
+ warnings.warn(
106
+
107
+ [2025-09-16 18:04:58,730][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
108
+ warnings.warn(msg)
109
+
110
+ [2025-09-16 18:05:03,198][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=279` in the `DataLoader` to improve performance.
111
+
112
+ [2025-09-16 18:05:05,392][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
113
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
114
+ warnings.warn(
115
+
116
+ [2025-09-16 18:05:05,581][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
117
+
118
+ [2025-09-16 18:05:05,583][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
119
+ warnings.warn(
120
+
121
+ [2025-09-16 18:05:05,583][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
122
+ warnings.warn(msg)
123
+
124
+ [2025-09-16 18:05:07,261][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
125
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
126
+
127
+ [2025-09-16 18:08:24,727][dinov2][INFO] - using MLP layer as FFN
128
+ [2025-09-16 18:08:30,965][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
129
+ warnings.warn(
130
+
131
+ [2025-09-16 18:08:30,966][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
132
+ warnings.warn(msg)
133
+
134
+ [2025-09-16 18:14:16,577][dinov2][INFO] - using MLP layer as FFN
135
+ [2025-09-16 18:14:22,737][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
136
+ warnings.warn(
137
+
138
+ [2025-09-16 18:14:22,738][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
139
+ warnings.warn(msg)
140
+
141
+ [2025-09-16 18:16:48,200][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=34` in the `DataLoader` to improve performance.
142
+
143
+ [2025-09-16 18:16:48,201][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
144
+ warnings.warn( # warn only once
145
+
146
+ [2025-09-16 18:16:51,169][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
147
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
148
+ warnings.warn(
149
+
150
+ [2025-09-16 18:16:51,447][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
151
+
152
+ [2025-09-16 18:16:51,448][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
153
+ warnings.warn(
154
+
155
+ [2025-09-16 18:16:51,449][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
156
+ warnings.warn(msg)
157
+
158
+ [2025-09-16 18:16:53,384][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
159
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
160
+
161
+ [2025-09-16 18:16:53,735][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/psnr', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
162
+
163
+ [2025-09-16 18:16:53,738][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/lpips', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
164
+
165
+ [2025-09-16 18:16:53,738][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/ssim', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
166
+
167
+ [2025-09-16 18:16:53,739][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/gaussian_num_ratio', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
168
+
169
+ [2025-09-16 18:17:04,137][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
170
+ grad.sizes() = [56, 128, 1, 1], strides() = [128, 1, 128, 128]
171
+ bucket_view.sizes() = [56, 128, 1, 1], strides() = [128, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
172
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
173
+
174
+ [2025-09-16 18:23:12,050][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
175
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
176
+
177
+ [2025-09-16 19:44:06,679][dinov2][INFO] - using MLP layer as FFN
178
+ [2025-09-16 19:44:14,557][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
179
+ warnings.warn(
180
+
181
+ [2025-09-16 19:44:14,557][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
182
+ warnings.warn(msg)
183
+
184
+ [2025-09-16 19:45:55,937][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=34` in the `DataLoader` to improve performance.
185
+
186
+ [2025-09-16 19:45:55,947][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
187
+ warnings.warn( # warn only once
188
+
189
+ [2025-09-16 19:46:04,605][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
190
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
191
+ warnings.warn(
192
+
193
+ [2025-09-16 19:46:05,884][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
194
+
195
+ [2025-09-16 19:46:05,885][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
196
+ warnings.warn(
197
+
198
+ [2025-09-16 19:46:05,886][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
199
+ warnings.warn(msg)
200
+
201
+ [2025-09-16 19:46:07,681][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
202
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
203
+
204
+ [2025-09-16 19:46:08,031][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/psnr', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
205
+
206
+ [2025-09-16 19:46:08,033][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/lpips', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
207
+
208
+ [2025-09-16 19:46:08,033][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/ssim', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
209
+
210
+ [2025-09-16 19:46:08,034][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/gaussian_num_ratio', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
211
+
212
+ [2025-09-16 19:46:18,197][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
213
+ grad.sizes() = [56, 128, 1, 1], strides() = [128, 1, 128, 128]
214
+ bucket_view.sizes() = [56, 128, 1, 1], strides() = [128, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
215
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
216
+
217
+ [2025-09-16 19:52:10,048][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
218
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
219
+
220
+ [2025-09-17 08:15:57,790][dinov2][INFO] - using MLP layer as FFN
221
+ [2025-09-17 08:16:05,563][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
222
+ warnings.warn(
223
+
224
+ [2025-09-17 08:16:05,564][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
225
+ warnings.warn(msg)
226
+
227
+ [2025-09-17 08:17:32,793][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=34` in the `DataLoader` to improve performance.
228
+
229
+ [2025-09-17 08:17:32,795][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
230
+ warnings.warn( # warn only once
231
+
232
+ [2025-09-17 08:17:35,728][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
233
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
234
+ warnings.warn(
235
+
236
+ [2025-09-17 08:17:35,970][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
237
+
238
+ [2025-09-17 08:17:35,971][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
239
+ warnings.warn(
240
+
241
+ [2025-09-17 08:17:35,972][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
242
+ warnings.warn(msg)
243
+
244
+ [2025-09-17 08:17:37,549][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
245
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
246
+
247
+ [2025-09-17 08:17:37,877][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/psnr', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
248
+
249
+ [2025-09-17 08:17:37,879][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/lpips', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
250
+
251
+ [2025-09-17 08:17:37,879][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/ssim', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
252
+
253
+ [2025-09-17 08:17:37,879][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/logger_connector/result.py:434: It is recommended to use `self.log('val/gaussian_num_ratio', ..., sync_dist=True)` when logging on epoch level in distributed setting to accumulate the metric across devices.
254
+
255
+ [2025-09-17 08:17:47,740][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
256
+ grad.sizes() = [56, 128, 1, 1], strides() = [128, 1, 128, 128]
257
+ bucket_view.sizes() = [56, 128, 1, 1], strides() = [128, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
258
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
259
+
260
+ [2025-09-17 08:23:38,721][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
261
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
262
+
train_ddp_process_1.log ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-09-16 17:58:19,330][dinov2][INFO] - using MLP layer as FFN
2
+ [2025-09-16 18:02:49,892][dinov2][INFO] - using MLP layer as FFN
3
+ [2025-09-16 18:03:08,637][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
4
+ warnings.warn(
5
+
6
+ [2025-09-16 18:03:08,638][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
7
+ warnings.warn(msg)
8
+
9
+ [2025-09-16 18:08:41,358][dinov2][INFO] - using MLP layer as FFN
10
+ [2025-09-16 18:09:17,206][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
11
+ warnings.warn(
12
+
13
+ [2025-09-16 18:09:17,206][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
14
+ warnings.warn(msg)
15
+
16
+ [2025-09-16 18:14:31,602][dinov2][INFO] - using MLP layer as FFN
17
+ [2025-09-16 18:14:50,266][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
18
+ warnings.warn(
19
+
20
+ [2025-09-16 18:14:50,267][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
21
+ warnings.warn(msg)
22
+
23
+ [2025-09-16 18:16:48,201][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
24
+ warnings.warn( # warn only once
25
+
26
+ [2025-09-16 18:17:00,737][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
27
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
28
+ warnings.warn(
29
+
30
+ [2025-09-16 18:17:04,136][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
31
+ grad.sizes() = [56, 128, 1, 1], strides() = [128, 1, 128, 128]
32
+ bucket_view.sizes() = [56, 128, 1, 1], strides() = [128, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
33
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
34
+
35
+ [2025-09-16 18:23:12,050][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
36
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
37
+
38
+ [2025-09-16 19:44:23,040][dinov2][INFO] - using MLP layer as FFN
39
+ [2025-09-16 19:44:42,938][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
40
+ warnings.warn(
41
+
42
+ [2025-09-16 19:44:42,939][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
43
+ warnings.warn(msg)
44
+
45
+ [2025-09-16 19:45:55,947][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
46
+ warnings.warn( # warn only once
47
+
48
+ [2025-09-16 19:46:15,532][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
49
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
50
+ warnings.warn(
51
+
52
+ [2025-09-16 19:46:18,197][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
53
+ grad.sizes() = [56, 128, 1, 1], strides() = [128, 1, 128, 128]
54
+ bucket_view.sizes() = [56, 128, 1, 1], strides() = [128, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
55
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
56
+
57
+ [2025-09-16 19:52:10,048][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
58
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
59
+
60
+ [2025-09-17 08:16:14,871][dinov2][INFO] - using MLP layer as FFN
61
+ [2025-09-17 08:16:47,663][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
62
+ warnings.warn(
63
+
64
+ [2025-09-17 08:16:47,664][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
65
+ warnings.warn(msg)
66
+
67
+ [2025-09-17 08:17:32,794][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
68
+ warnings.warn( # warn only once
69
+
70
+ [2025-09-17 08:17:44,025][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
71
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
72
+ warnings.warn(
73
+
74
+ [2025-09-17 08:17:47,739][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
75
+ grad.sizes() = [56, 128, 1, 1], strides() = [128, 1, 128, 128]
76
+ bucket_view.sizes() = [56, 128, 1, 1], strides() = [128, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
77
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
78
+
79
+ [2025-09-17 08:23:38,705][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
80
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
81
+
train_ddp_process_2.log ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-09-16 17:58:18,234][dinov2][INFO] - using MLP layer as FFN
2
+ [2025-09-16 18:02:50,404][dinov2][INFO] - using MLP layer as FFN
3
+ [2025-09-16 18:03:28,788][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
4
+ warnings.warn(
5
+
6
+ [2025-09-16 18:03:28,793][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
7
+ warnings.warn(msg)
8
+
9
+ [2025-09-16 18:08:40,323][dinov2][INFO] - using MLP layer as FFN
10
+ [2025-09-16 18:08:59,994][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
11
+ warnings.warn(
12
+
13
+ [2025-09-16 18:08:59,995][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
14
+ warnings.warn(msg)
15
+
16
+ [2025-09-16 18:14:32,040][dinov2][INFO] - using MLP layer as FFN
17
+ [2025-09-16 18:14:51,075][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
18
+ warnings.warn(
19
+
20
+ [2025-09-16 18:14:51,077][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
21
+ warnings.warn(msg)
22
+
23
+ [2025-09-16 18:16:48,201][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
24
+ warnings.warn( # warn only once
25
+
26
+ [2025-09-16 18:17:00,237][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
27
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
28
+ warnings.warn(
29
+
30
+ [2025-09-16 18:17:04,136][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
31
+ grad.sizes() = [56, 128, 1, 1], strides() = [128, 1, 128, 128]
32
+ bucket_view.sizes() = [56, 128, 1, 1], strides() = [128, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
33
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
34
+
35
+ [2025-09-16 18:23:12,054][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
36
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
37
+
38
+ [2025-09-16 19:44:22,989][dinov2][INFO] - using MLP layer as FFN
39
+ [2025-09-16 19:44:41,440][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
40
+ warnings.warn(
41
+
42
+ [2025-09-16 19:44:41,441][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
43
+ warnings.warn(msg)
44
+
45
+ [2025-09-16 19:45:55,947][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
46
+ warnings.warn( # warn only once
47
+
48
+ [2025-09-16 19:46:14,712][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
49
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
50
+ warnings.warn(
51
+
52
+ [2025-09-16 19:46:18,197][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
53
+ grad.sizes() = [56, 128, 1, 1], strides() = [128, 1, 128, 128]
54
+ bucket_view.sizes() = [56, 128, 1, 1], strides() = [128, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
55
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
56
+
57
+ [2025-09-16 19:52:10,049][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
58
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
59
+
60
+ [2025-09-17 08:16:15,147][dinov2][INFO] - using MLP layer as FFN
61
+ [2025-09-17 08:16:47,824][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
62
+ warnings.warn(
63
+
64
+ [2025-09-17 08:16:47,825][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
65
+ warnings.warn(msg)
66
+
67
+ [2025-09-17 08:17:32,795][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
68
+ warnings.warn( # warn only once
69
+
70
+ [2025-09-17 08:17:44,084][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
71
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
72
+ warnings.warn(
73
+
74
+ [2025-09-17 08:17:47,739][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
75
+ grad.sizes() = [56, 128, 1, 1], strides() = [128, 1, 128, 128]
76
+ bucket_view.sizes() = [56, 128, 1, 1], strides() = [128, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
77
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
78
+
79
+ [2025-09-17 08:23:38,704][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
80
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
81
+
train_ddp_process_3.log ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-09-16 17:58:19,053][dinov2][INFO] - using MLP layer as FFN
2
+ [2025-09-16 18:02:50,349][dinov2][INFO] - using MLP layer as FFN
3
+ [2025-09-16 18:03:29,558][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
4
+ warnings.warn(
5
+
6
+ [2025-09-16 18:03:29,561][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
7
+ warnings.warn(msg)
8
+
9
+ [2025-09-16 18:08:41,428][dinov2][INFO] - using MLP layer as FFN
10
+ [2025-09-16 18:09:17,198][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
11
+ warnings.warn(
12
+
13
+ [2025-09-16 18:09:17,199][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
14
+ warnings.warn(msg)
15
+
16
+ [2025-09-16 18:14:32,923][dinov2][INFO] - using MLP layer as FFN
17
+ [2025-09-16 18:15:09,086][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
18
+ warnings.warn(
19
+
20
+ [2025-09-16 18:15:09,087][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
21
+ warnings.warn(msg)
22
+
23
+ [2025-09-16 18:16:48,201][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
24
+ warnings.warn( # warn only once
25
+
26
+ [2025-09-16 18:17:00,926][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
27
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
28
+ warnings.warn(
29
+
30
+ [2025-09-16 18:17:04,134][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
31
+ grad.sizes() = [56, 128, 1, 1], strides() = [128, 1, 128, 128]
32
+ bucket_view.sizes() = [56, 128, 1, 1], strides() = [128, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
33
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
34
+
35
+ [2025-09-16 18:23:12,047][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
36
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
37
+
38
+ [2025-09-16 19:44:23,779][dinov2][INFO] - using MLP layer as FFN
39
+ [2025-09-16 19:44:55,897][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
40
+ warnings.warn(
41
+
42
+ [2025-09-16 19:44:55,899][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
43
+ warnings.warn(msg)
44
+
45
+ [2025-09-16 19:45:55,947][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
46
+ warnings.warn( # warn only once
47
+
48
+ [2025-09-16 19:46:15,252][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
49
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
50
+ warnings.warn(
51
+
52
+ [2025-09-16 19:46:18,197][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
53
+ grad.sizes() = [56, 128, 1, 1], strides() = [128, 1, 128, 128]
54
+ bucket_view.sizes() = [56, 128, 1, 1], strides() = [128, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
55
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
56
+
57
+ [2025-09-16 19:52:10,048][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
58
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
59
+
60
+ [2025-09-17 08:16:14,263][dinov2][INFO] - using MLP layer as FFN
61
+ [2025-09-17 08:16:34,545][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
62
+ warnings.warn(
63
+
64
+ [2025-09-17 08:16:34,546][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
65
+ warnings.warn(msg)
66
+
67
+ [2025-09-17 08:17:32,795][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
68
+ warnings.warn( # warn only once
69
+
70
+ [2025-09-17 08:17:43,969][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
71
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
72
+ warnings.warn(
73
+
74
+ [2025-09-17 08:17:47,738][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
75
+ grad.sizes() = [56, 128, 1, 1], strides() = [128, 1, 128, 128]
76
+ bucket_view.sizes() = [56, 128, 1, 1], strides() = [128, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
77
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
78
+
79
+ [2025-09-17 08:23:38,705][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
80
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
81
+
train_ddp_process_4.log ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-09-16 17:58:18,473][dinov2][INFO] - using MLP layer as FFN
2
+ [2025-09-16 18:02:50,329][dinov2][INFO] - using MLP layer as FFN
3
+ [2025-09-16 18:03:27,530][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
4
+ warnings.warn(
5
+
6
+ [2025-09-16 18:03:27,531][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
7
+ warnings.warn(msg)
8
+
9
+ [2025-09-16 18:08:41,389][dinov2][INFO] - using MLP layer as FFN
10
+ [2025-09-16 18:09:17,146][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
11
+ warnings.warn(
12
+
13
+ [2025-09-16 18:09:17,147][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
14
+ warnings.warn(msg)
15
+
16
+ [2025-09-16 18:14:32,867][dinov2][INFO] - using MLP layer as FFN
17
+ [2025-09-16 18:15:09,006][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
18
+ warnings.warn(
19
+
20
+ [2025-09-16 18:15:09,007][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
21
+ warnings.warn(msg)
22
+
23
+ [2025-09-16 18:16:48,201][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
24
+ warnings.warn( # warn only once
25
+
26
+ [2025-09-16 18:17:01,226][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
27
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
28
+ warnings.warn(
29
+
30
+ [2025-09-16 18:17:04,135][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
31
+ grad.sizes() = [56, 128, 1, 1], strides() = [128, 1, 128, 128]
32
+ bucket_view.sizes() = [56, 128, 1, 1], strides() = [128, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
33
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
34
+
35
+ [2025-09-16 18:23:12,055][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
36
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
37
+
38
+ [2025-09-16 19:44:23,799][dinov2][INFO] - using MLP layer as FFN
39
+ [2025-09-16 19:44:56,745][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
40
+ warnings.warn(
41
+
42
+ [2025-09-16 19:44:56,746][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
43
+ warnings.warn(msg)
44
+
45
+ [2025-09-16 19:45:55,947][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
46
+ warnings.warn( # warn only once
47
+
48
+ [2025-09-16 19:46:14,882][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
49
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
50
+ warnings.warn(
51
+
52
+ [2025-09-16 19:46:18,197][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
53
+ grad.sizes() = [56, 128, 1, 1], strides() = [128, 1, 128, 128]
54
+ bucket_view.sizes() = [56, 128, 1, 1], strides() = [128, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
55
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
56
+
57
+ [2025-09-16 19:52:10,048][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
58
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
59
+
60
+ [2025-09-17 08:16:14,950][dinov2][INFO] - using MLP layer as FFN
61
+ [2025-09-17 08:16:47,187][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
62
+ warnings.warn(
63
+
64
+ [2025-09-17 08:16:47,188][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
65
+ warnings.warn(msg)
66
+
67
+ [2025-09-17 08:17:32,794][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
68
+ warnings.warn( # warn only once
69
+
70
+ [2025-09-17 08:17:44,542][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
71
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
72
+ warnings.warn(
73
+
74
+ [2025-09-17 08:17:47,739][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
75
+ grad.sizes() = [56, 128, 1, 1], strides() = [128, 1, 128, 128]
76
+ bucket_view.sizes() = [56, 128, 1, 1], strides() = [128, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
77
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
78
+
79
+ [2025-09-17 08:23:38,710][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
80
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
81
+
train_ddp_process_5.log ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-09-16 17:58:18,567][dinov2][INFO] - using MLP layer as FFN
2
+ [2025-09-16 18:02:51,037][dinov2][INFO] - using MLP layer as FFN
3
+ [2025-09-16 18:03:30,313][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
4
+ warnings.warn(
5
+
6
+ [2025-09-16 18:03:30,315][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
7
+ warnings.warn(msg)
8
+
9
+ [2025-09-16 18:08:39,963][dinov2][INFO] - using MLP layer as FFN
10
+ [2025-09-16 18:08:59,630][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
11
+ warnings.warn(
12
+
13
+ [2025-09-16 18:08:59,630][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
14
+ warnings.warn(msg)
15
+
16
+ [2025-09-16 18:14:32,775][dinov2][INFO] - using MLP layer as FFN
17
+ [2025-09-16 18:15:08,948][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
18
+ warnings.warn(
19
+
20
+ [2025-09-16 18:15:08,957][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
21
+ warnings.warn(msg)
22
+
23
+ [2025-09-16 18:16:48,200][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
24
+ warnings.warn( # warn only once
25
+
26
+ [2025-09-16 18:17:00,467][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
27
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
28
+ warnings.warn(
29
+
30
+ [2025-09-16 18:17:04,134][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
31
+ grad.sizes() = [56, 128, 1, 1], strides() = [128, 1, 128, 128]
32
+ bucket_view.sizes() = [56, 128, 1, 1], strides() = [128, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
33
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
34
+
35
+ [2025-09-16 18:23:12,042][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
36
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
37
+
38
+ [2025-09-16 19:44:23,247][dinov2][INFO] - using MLP layer as FFN
39
+ [2025-09-16 19:44:43,991][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
40
+ warnings.warn(
41
+
42
+ [2025-09-16 19:44:43,991][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
43
+ warnings.warn(msg)
44
+
45
+ [2025-09-16 19:45:55,947][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
46
+ warnings.warn( # warn only once
47
+
48
+ [2025-09-16 19:46:15,452][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
49
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
50
+ warnings.warn(
51
+
52
+ [2025-09-16 19:46:18,197][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
53
+ grad.sizes() = [56, 128, 1, 1], strides() = [128, 1, 128, 128]
54
+ bucket_view.sizes() = [56, 128, 1, 1], strides() = [128, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
55
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
56
+
57
+ [2025-09-16 19:52:10,048][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
58
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
59
+
60
+ [2025-09-17 08:16:13,809][dinov2][INFO] - using MLP layer as FFN
61
+ [2025-09-17 08:16:33,675][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
62
+ warnings.warn(
63
+
64
+ [2025-09-17 08:16:33,676][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
65
+ warnings.warn(msg)
66
+
67
+ [2025-09-17 08:17:32,794][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
68
+ warnings.warn( # warn only once
69
+
70
+ [2025-09-17 08:17:44,220][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
71
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
72
+ warnings.warn(
73
+
74
+ [2025-09-17 08:17:47,736][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
75
+ grad.sizes() = [56, 128, 1, 1], strides() = [128, 1, 128, 128]
76
+ bucket_view.sizes() = [56, 128, 1, 1], strides() = [128, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
77
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
78
+
79
+ [2025-09-17 08:23:38,709][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
80
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
81
+
train_ddp_process_6.log ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-09-16 17:58:19,111][dinov2][INFO] - using MLP layer as FFN
2
+ [2025-09-16 18:02:51,004][dinov2][INFO] - using MLP layer as FFN
3
+ [2025-09-16 18:03:30,361][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
4
+ warnings.warn(
5
+
6
+ [2025-09-16 18:03:30,363][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
7
+ warnings.warn(msg)
8
+
9
+ [2025-09-16 18:08:40,776][dinov2][INFO] - using MLP layer as FFN
10
+ [2025-09-16 18:09:13,164][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
11
+ warnings.warn(
12
+
13
+ [2025-09-16 18:09:13,165][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
14
+ warnings.warn(msg)
15
+
16
+ [2025-09-16 18:14:32,593][dinov2][INFO] - using MLP layer as FFN
17
+ [2025-09-16 18:15:06,668][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
18
+ warnings.warn(
19
+
20
+ [2025-09-16 18:15:06,673][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
21
+ warnings.warn(msg)
22
+
23
+ [2025-09-16 18:16:48,201][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
24
+ warnings.warn( # warn only once
25
+
26
+ [2025-09-16 18:17:01,180][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
27
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
28
+ warnings.warn(
29
+
30
+ [2025-09-16 18:17:04,136][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
31
+ grad.sizes() = [56, 128, 1, 1], strides() = [128, 1, 128, 128]
32
+ bucket_view.sizes() = [56, 128, 1, 1], strides() = [128, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
33
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
34
+
35
+ [2025-09-16 18:23:12,046][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
36
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
37
+
38
+ [2025-09-16 19:44:23,838][dinov2][INFO] - using MLP layer as FFN
39
+ [2025-09-16 19:44:56,721][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
40
+ warnings.warn(
41
+
42
+ [2025-09-16 19:44:56,722][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
43
+ warnings.warn(msg)
44
+
45
+ [2025-09-16 19:45:55,947][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
46
+ warnings.warn( # warn only once
47
+
48
+ [2025-09-16 19:46:15,168][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
49
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
50
+ warnings.warn(
51
+
52
+ [2025-09-16 19:46:18,197][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
53
+ grad.sizes() = [56, 128, 1, 1], strides() = [128, 1, 128, 128]
54
+ bucket_view.sizes() = [56, 128, 1, 1], strides() = [128, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
55
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
56
+
57
+ [2025-09-16 19:52:10,048][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
58
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
59
+
60
+ [2025-09-17 08:16:14,202][dinov2][INFO] - using MLP layer as FFN
61
+ [2025-09-17 08:16:34,429][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
62
+ warnings.warn(
63
+
64
+ [2025-09-17 08:16:34,429][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
65
+ warnings.warn(msg)
66
+
67
+ [2025-09-17 08:17:32,794][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
68
+ warnings.warn( # warn only once
69
+
70
+ [2025-09-17 08:17:44,160][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
71
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
72
+ warnings.warn(
73
+
74
+ [2025-09-17 08:17:47,738][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
75
+ grad.sizes() = [56, 128, 1, 1], strides() = [128, 1, 128, 128]
76
+ bucket_view.sizes() = [56, 128, 1, 1], strides() = [128, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
77
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
78
+
79
+ [2025-09-17 08:23:38,716][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
80
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
81
+
train_ddp_process_7.log ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-09-16 17:58:18,319][dinov2][INFO] - using MLP layer as FFN
2
+ [2025-09-16 18:02:51,163][dinov2][INFO] - using MLP layer as FFN
3
+ [2025-09-16 18:03:30,264][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
4
+ warnings.warn(
5
+
6
+ [2025-09-16 18:03:30,265][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
7
+ warnings.warn(msg)
8
+
9
+ [2025-09-16 18:08:40,815][dinov2][INFO] - using MLP layer as FFN
10
+ [2025-09-16 18:09:16,606][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
11
+ warnings.warn(
12
+
13
+ [2025-09-16 18:09:16,607][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
14
+ warnings.warn(msg)
15
+
16
+ [2025-09-16 18:14:32,775][dinov2][INFO] - using MLP layer as FFN
17
+ [2025-09-16 18:15:08,930][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
18
+ warnings.warn(
19
+
20
+ [2025-09-16 18:15:08,931][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
21
+ warnings.warn(msg)
22
+
23
+ [2025-09-16 18:16:48,201][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
24
+ warnings.warn( # warn only once
25
+
26
+ [2025-09-16 18:17:00,823][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
27
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
28
+ warnings.warn(
29
+
30
+ [2025-09-16 18:17:04,135][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
31
+ grad.sizes() = [56, 128, 1, 1], strides() = [128, 1, 128, 128]
32
+ bucket_view.sizes() = [56, 128, 1, 1], strides() = [128, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
33
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
34
+
35
+ [2025-09-16 18:23:12,046][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
36
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
37
+
38
+ [2025-09-16 19:44:23,886][dinov2][INFO] - using MLP layer as FFN
39
+ [2025-09-16 19:44:56,838][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
40
+ warnings.warn(
41
+
42
+ [2025-09-16 19:44:56,839][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
43
+ warnings.warn(msg)
44
+
45
+ [2025-09-16 19:45:55,947][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
46
+ warnings.warn( # warn only once
47
+
48
+ [2025-09-16 19:46:15,492][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
49
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
50
+ warnings.warn(
51
+
52
+ [2025-09-16 19:46:18,197][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
53
+ grad.sizes() = [56, 128, 1, 1], strides() = [128, 1, 128, 128]
54
+ bucket_view.sizes() = [56, 128, 1, 1], strides() = [128, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
55
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
56
+
57
+ [2025-09-16 19:52:10,048][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
58
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
59
+
60
+ [2025-09-17 08:16:14,915][dinov2][INFO] - using MLP layer as FFN
61
+ [2025-09-17 08:16:47,804][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
62
+ warnings.warn(
63
+
64
+ [2025-09-17 08:16:47,805][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
65
+ warnings.warn(msg)
66
+
67
+ [2025-09-17 08:17:32,794][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/distributed/distributed_c10d.py:4631: UserWarning: No device id is provided via `init_process_group` or `barrier `. Using the current device set by the user.
68
+ warnings.warn( # warn only once
69
+
70
+ [2025-09-17 08:17:45,024][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
71
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
72
+ warnings.warn(
73
+
74
+ [2025-09-17 08:17:47,738][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/autograd/graph.py:824: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
75
+ grad.sizes() = [56, 128, 1, 1], strides() = [128, 1, 128, 128]
76
+ bucket_view.sizes() = [56, 128, 1, 1], strides() = [128, 1, 1, 1] (Triggered internally at /pytorch/torch/csrc/distributed/c10d/reducer.cpp:328.)
77
+ return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
78
+
79
+ [2025-09-17 08:23:38,704][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/optim/lr_scheduler.py:198: UserWarning: The epoch parameter in `scheduler.step()` was not necessary and is being deprecated where possible. Please use `scheduler.step()` to step the scheduler. During the deprecation, if epoch is different from None, the closed form is used instead of the new chainable form, where available. Please open an issue if you are unable to replicate your use case: https://github.com/pytorch/pytorch/issues/new/choose.
80
+ warnings.warn(EPOCH_DEPRECATION_WARNING, UserWarning)
81
+
wandb/debug-internal.log ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {"time":"2025-09-17T08:17:30.716918246+08:00","level":"INFO","msg":"stream: starting","core version":"0.19.9","symlink path":"/root/highspeedstorage/injae/code/ICLR2026_HCP/outputs/re10k_hpc/wandb/run-20250917_081730-bcwwe2gj/logs/debug-core.log"}
2
+ {"time":"2025-09-17T08:17:31.176856401+08:00","level":"INFO","msg":"created new stream","id":"bcwwe2gj"}
3
+ {"time":"2025-09-17T08:17:31.177025464+08:00","level":"INFO","msg":"stream: started","id":"bcwwe2gj"}
4
+ {"time":"2025-09-17T08:17:31.177197947+08:00","level":"INFO","msg":"writer: Do: started","stream_id":"bcwwe2gj"}
5
+ {"time":"2025-09-17T08:17:31.177369747+08:00","level":"INFO","msg":"handler: started","stream_id":"bcwwe2gj"}
6
+ {"time":"2025-09-17T08:17:31.177482294+08:00","level":"INFO","msg":"sender: started","stream_id":"bcwwe2gj"}
7
+ {"time":"2025-09-17T08:17:31.468144818+08:00","level":"INFO","msg":"Starting system monitor"}
wandb/debug.log ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-09-17 08:17:30,694 INFO MainThread:31741 [wandb_setup.py:_flush():67] Current SDK version is 0.19.9
2
+ 2025-09-17 08:17:30,694 INFO MainThread:31741 [wandb_setup.py:_flush():67] Configure stats pid to 31741
3
+ 2025-09-17 08:17:30,694 INFO MainThread:31741 [wandb_setup.py:_flush():67] Loading settings from /root/.config/wandb/settings
4
+ 2025-09-17 08:17:30,694 INFO MainThread:31741 [wandb_setup.py:_flush():67] Loading settings from /root/highspeedstorage/injae/code/ICLR2026_HCP/wandb/settings
5
+ 2025-09-17 08:17:30,694 INFO MainThread:31741 [wandb_setup.py:_flush():67] Loading settings from environment variables
6
+ 2025-09-17 08:17:30,694 INFO MainThread:31741 [wandb_init.py:setup_run_log_directory():662] Logging user logs to /root/highspeedstorage/injae/code/ICLR2026_HCP/outputs/re10k_hpc/wandb/run-20250917_081730-bcwwe2gj/logs/debug.log
7
+ 2025-09-17 08:17:30,695 INFO MainThread:31741 [wandb_init.py:setup_run_log_directory():663] Logging internal logs to /root/highspeedstorage/injae/code/ICLR2026_HCP/outputs/re10k_hpc/wandb/run-20250917_081730-bcwwe2gj/logs/debug-internal.log
8
+ 2025-09-17 08:17:30,695 INFO MainThread:31741 [wandb_init.py:init():781] calling init triggers
9
+ 2025-09-17 08:17:30,695 INFO MainThread:31741 [wandb_init.py:init():786] wandb.init called with sweep_config: {}
10
+ config: {'model': {'encoder': {'backbone': {'name': 'croco', 'model': 'ViTLarge_BaseDecoder', 'patch_embed_cls': 'PatchEmbedDust3R', 'asymmetry_decoder': True, 'intrinsics_embed_loc': 'encoder', 'intrinsics_embed_degree': 4, 'intrinsics_embed_type': 'token'}, 'name': 'noposplat', 'opacity_mapping': {'initial': 0.0, 'final': 0.0, 'warm_up': 1}, 'num_monocular_samples': 32, 'num_surfaces': 1, 'predict_opacity': False, 'gaussians_per_pixel': 1, 'gaussian_adapter': {'gaussian_scale_min': 0.5, 'gaussian_scale_max': 15.0, 'sh_degree': 4}, 'd_feature': 128, 'visualizer': {'num_samples': 8, 'min_resolution': 256, 'export_ply': False}, 'apply_bounds_shim': True, 'gs_params_head_type': 'dpt_gs', 'pose_free': True, 'pretrained_weights': ''}, 'decoder': {'name': 'splatting_cuda', 'background_color': [0.0, 0.0, 0.0], 'make_scale_invariant': False, 'use_gsplat': True}, 'vggt': {'input_image_shape': [518, 518], 'head_mode': 'pcd', 'same_head': False, 'camera_prompt': False, 'num_visual_prompt': 0, 'use_extrinsic_token': False, 'use_intrinsic_token': False, 'detr': {'name': 'deformable_detr', 'features_num_channel': [128, 32], 'hidden_dim': 32, 'dim_down_factor': 1, 'dim_feedforward_factor': 4, 'enc_layers': 6, 'dec_layers': 1, 'dropout': 0.1, 'num_feature_levels': 2, 'dec_n_points': 4, 'enc_n_points': 4, 'split_threshold': 1.0, 'num_queries': 256, 'base_HW': 64, 'position_embedding': 'sine', 'pre_norm': False, 'self_layer_num': 1, 'num_level': 3, 'debug_mode': 0, 'decoder_type': 'conv', 'split_threshold_range': [0.0, 2.0], 'only_last_level': False, 'grad_mode': 'absgrad', 'same_split_head': False, 'use_mean_features': True}}}, 'loss': {'mse': {'weight': 1.0}, 'lpips': {'weight': 0.05, 'apply_after_step': 0}}, 'wandb': {'project': 'noposplat', 'entity': 'scene-representation-group', 'name': 're10k_hpc', 'mode': 'online', 'tags': ['re10k', '256x256']}, 'mode': 'train', 'data_loader': {'train': {'num_workers': 16, 'persistent_workers': True, 'batch_size': 16, 'seed': 1234}, 'test': {'num_workers': 4, 'persistent_workers': False, 'batch_size': 1, 'seed': 2345}, 'val': {'num_workers': 1, 'persistent_workers': True, 'batch_size': 1, 'seed': 3456}}, 'optimizer': {'lr': 0.0002, 'warm_up_steps': 125, 'backbone_lr_multiplier': 0.1, 'backbone_trainable': 'GB+H', 'accumulate': 1}, 'checkpointing': {'load': None, 'every_n_train_steps': 9375, 'save_top_k': 1, 'save_weights_only': True}, 'train': {'depth_mode': None, 'extended_visualization': False, 'print_log_every_n_steps': 10, 'distiller': '', 'distill_max_steps': 1000000, 'debug_mode': 0, 'camera_consis_loss': 0.0, 'camera_loss': 1.0, 'aux_loss': 0.0, 'one_sample_train': False, 'target_scale_mode': 2, 'extrinsic_drop': 0.1, 'intrinsic_drop': 0.1, 'align_corners': False, 'backbone_checkpoint': 'none', 'intrinsic_scaling': True, 'return_level_gs': True, 'train_context': False, 'use_pred_point_focal': False, 'scaling_detach': True, 'split_score_log_scale': True, 'split_score_grad_scale': 10000.0, 'split_loss': 0.01, 'random_split_strategy': 'score'}, 'test': {'output_path': 'test', 'align_pose': False, 'pose_align_steps': 100, 'rot_opt_lr': 0.005, 'trans_opt_lr': 0.005, 'compute_scores': True, 'save_image': False, 'save_video': False, 'save_compare': False, 'split_threshold_list': [0.0, 0.1, 0.2, 0.3, 0.4], 'pred_intrinsic': False}, 'seed': 111123, 'trainer': {'max_steps': 18751, 'val_check_interval': 500, 'gradient_clip_val': 0.5, 'num_nodes': 1}, 'dataset': {'re10k': {'make_baseline_1': True, 'relative_pose': True, 'augment': True, 'background_color': [0.0, 0.0, 0.0], 'overfit_to_scene': None, 'skip_bad_shape': True, 'view_sampler': {'name': 'bounded', 'num_target_views': 4, 'num_context_views': 2, 'min_distance_between_context_views': 45, 'max_distance_between_context_views': 90, 'min_distance_to_context_views': 0, 'warm_up_steps': 9375, 'initial_min_distance_between_context_views': 25, 'initial_max_distance_between_context_views': 25, 'distill_steps': 0}, 'name': 're10k', 'roots': ['datasets/re10k'], 'input_image_shape': [256, 256], 'original_image_shape': [360, 640], 'cameras_are_circular': False, 'baseline_min': 0.001, 'baseline_max': 10000000000.0, 'max_fov': 100.0}}, '_wandb': {}}
11
+ 2025-09-17 08:17:30,695 INFO MainThread:31741 [wandb_init.py:init():809] starting backend
12
+ 2025-09-17 08:17:30,695 INFO MainThread:31741 [wandb_init.py:init():813] sending inform_init request
13
+ 2025-09-17 08:17:30,711 INFO MainThread:31741 [backend.py:_multiprocessing_setup():101] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
14
+ 2025-09-17 08:17:30,711 INFO MainThread:31741 [wandb_init.py:init():823] backend started and connected
15
+ 2025-09-17 08:17:30,725 INFO MainThread:31741 [wandb_init.py:init():915] updated telemetry
16
+ 2025-09-17 08:17:30,760 INFO MainThread:31741 [wandb_init.py:init():939] communicating run to backend with 90.0 second timeout
17
+ 2025-09-17 08:17:31,459 INFO MainThread:31741 [wandb_init.py:init():1014] starting run threads in backend
18
+ 2025-09-17 08:17:31,629 INFO MainThread:31741 [wandb_run.py:_console_start():2454] atexit reg
19
+ 2025-09-17 08:17:31,630 INFO MainThread:31741 [wandb_run.py:_redirect():2306] redirect: wrap_raw
20
+ 2025-09-17 08:17:31,632 INFO MainThread:31741 [wandb_run.py:_redirect():2371] Wrapping output streams.
21
+ 2025-09-17 08:17:31,632 INFO MainThread:31741 [wandb_run.py:_redirect():2394] Redirects installed.
22
+ 2025-09-17 08:17:31,635 INFO MainThread:31741 [wandb_init.py:init():1056] run started, returning control to user process
wandb/run-20250916_175907-b2gr1gxl/run-b2gr1gxl.wandb ADDED
Binary file (15.2 kB). View file
 
wandb/run-20250916_180036-llem5odx/files/config.yaml ADDED
@@ -0,0 +1,423 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.19.9
4
+ m:
5
+ - "1": points_context_imgs.height
6
+ "5": 2
7
+ "6":
8
+ - 1
9
+ - 3
10
+ "7": []
11
+ - "1": trainer/global_step
12
+ "6":
13
+ - 3
14
+ "7": []
15
+ - "1": comparison.count
16
+ "5": 2
17
+ "6":
18
+ - 1
19
+ - 3
20
+ "7": []
21
+ - "1": points_context_imgs.format
22
+ "5": 2
23
+ "6":
24
+ - 1
25
+ - 3
26
+ "7": []
27
+ - "1": points_context_imgs._type
28
+ "5": 2
29
+ "6":
30
+ - 1
31
+ - 3
32
+ "7": []
33
+ - "1": attention_map.count
34
+ "5": 2
35
+ "6":
36
+ - 1
37
+ - 3
38
+ "7": []
39
+ - "1": activation_visualization.height
40
+ "5": 2
41
+ "6":
42
+ - 1
43
+ - 3
44
+ "7": []
45
+ - "1": attention_map.captions
46
+ "5": 2
47
+ "6":
48
+ - 1
49
+ - 3
50
+ "7": []
51
+ - "1": comparison.width
52
+ "5": 2
53
+ "6":
54
+ - 1
55
+ - 3
56
+ "7": []
57
+ - "1": activation_visualization.count
58
+ "5": 2
59
+ "6":
60
+ - 1
61
+ - 3
62
+ "7": []
63
+ - "1": activation_visualization.filenames
64
+ "5": 2
65
+ "6":
66
+ - 1
67
+ - 3
68
+ "7": []
69
+ - "1": comparison.filenames
70
+ "5": 2
71
+ "6":
72
+ - 1
73
+ - 3
74
+ "7": []
75
+ - "1": comparison.format
76
+ "5": 2
77
+ "6":
78
+ - 1
79
+ - 3
80
+ "7": []
81
+ - "1": attention_map.filenames
82
+ "5": 2
83
+ "6":
84
+ - 1
85
+ - 3
86
+ "7": []
87
+ - "1": comparison._type
88
+ "5": 2
89
+ "6":
90
+ - 1
91
+ - 3
92
+ "7": []
93
+ - "1": comparison.height
94
+ "5": 2
95
+ "6":
96
+ - 1
97
+ - 3
98
+ "7": []
99
+ - "1": activation_visualization.width
100
+ "5": 2
101
+ "6":
102
+ - 1
103
+ - 3
104
+ "7": []
105
+ - "1": points_context_imgs.captions
106
+ "5": 2
107
+ "6":
108
+ - 1
109
+ - 3
110
+ "7": []
111
+ - "1": points_context_imgs.width
112
+ "5": 2
113
+ "6":
114
+ - 1
115
+ - 3
116
+ "7": []
117
+ - "1": activation_visualization.format
118
+ "5": 2
119
+ "6":
120
+ - 1
121
+ - 3
122
+ "7": []
123
+ - "1": attention_map.width
124
+ "5": 2
125
+ "6":
126
+ - 1
127
+ - 3
128
+ "7": []
129
+ - "1": comparison.captions
130
+ "5": 2
131
+ "6":
132
+ - 1
133
+ - 3
134
+ "7": []
135
+ - "1": points_context_imgs.count
136
+ "5": 2
137
+ "6":
138
+ - 1
139
+ - 3
140
+ "7": []
141
+ - "1": points_context_imgs.filenames
142
+ "5": 2
143
+ "6":
144
+ - 1
145
+ - 3
146
+ "7": []
147
+ - "1": activation_visualization.captions
148
+ "5": 2
149
+ "6":
150
+ - 1
151
+ - 3
152
+ "7": []
153
+ - "1": attention_map.height
154
+ "5": 2
155
+ "6":
156
+ - 1
157
+ - 3
158
+ "7": []
159
+ - "1": attention_map.format
160
+ "5": 2
161
+ "6":
162
+ - 1
163
+ - 3
164
+ "7": []
165
+ - "1": activation_visualization._type
166
+ "5": 2
167
+ "6":
168
+ - 1
169
+ - 3
170
+ "7": []
171
+ - "1": attention_map._type
172
+ "5": 2
173
+ "6":
174
+ - 1
175
+ - 3
176
+ "7": []
177
+ python_version: 3.12.11
178
+ t:
179
+ "1":
180
+ - 1
181
+ - 41
182
+ - 49
183
+ - 50
184
+ - 55
185
+ - 106
186
+ "2":
187
+ - 1
188
+ - 41
189
+ - 49
190
+ - 50
191
+ - 55
192
+ - 106
193
+ "3":
194
+ - 7
195
+ - 13
196
+ - 15
197
+ - 16
198
+ - 23
199
+ - 55
200
+ - 66
201
+ "4": 3.12.11
202
+ "5": 0.19.9
203
+ "8":
204
+ - 5
205
+ "12": 0.19.9
206
+ "13": linux-x86_64
207
+ checkpointing:
208
+ value:
209
+ every_n_train_steps: 9375
210
+ load: null
211
+ save_top_k: 1
212
+ save_weights_only: true
213
+ data_loader:
214
+ value:
215
+ test:
216
+ batch_size: 1
217
+ num_workers: 4
218
+ persistent_workers: false
219
+ seed: 2345
220
+ train:
221
+ batch_size: 16
222
+ num_workers: 16
223
+ persistent_workers: true
224
+ seed: 1234
225
+ val:
226
+ batch_size: 1
227
+ num_workers: 1
228
+ persistent_workers: true
229
+ seed: 3456
230
+ dataset:
231
+ value:
232
+ re10k:
233
+ augment: true
234
+ background_color:
235
+ - 0
236
+ - 0
237
+ - 0
238
+ baseline_max: 1e+10
239
+ baseline_min: 0.001
240
+ cameras_are_circular: false
241
+ input_image_shape:
242
+ - 256
243
+ - 256
244
+ make_baseline_1: true
245
+ max_fov: 100
246
+ name: re10k
247
+ original_image_shape:
248
+ - 360
249
+ - 640
250
+ overfit_to_scene: null
251
+ relative_pose: true
252
+ roots:
253
+ - datasets/re10k
254
+ skip_bad_shape: true
255
+ view_sampler:
256
+ distill_steps: 0
257
+ initial_max_distance_between_context_views: 25
258
+ initial_min_distance_between_context_views: 25
259
+ max_distance_between_context_views: 90
260
+ min_distance_between_context_views: 45
261
+ min_distance_to_context_views: 0
262
+ name: bounded
263
+ num_context_views: 2
264
+ num_target_views: 4
265
+ warm_up_steps: 9375
266
+ loss:
267
+ value:
268
+ lpips:
269
+ apply_after_step: 0
270
+ weight: 0.05
271
+ mse:
272
+ weight: 1
273
+ mode:
274
+ value: train
275
+ model:
276
+ value:
277
+ decoder:
278
+ background_color:
279
+ - 0
280
+ - 0
281
+ - 0
282
+ make_scale_invariant: false
283
+ name: splatting_cuda
284
+ use_gsplat: true
285
+ encoder:
286
+ apply_bounds_shim: true
287
+ backbone:
288
+ asymmetry_decoder: true
289
+ intrinsics_embed_degree: 4
290
+ intrinsics_embed_loc: encoder
291
+ intrinsics_embed_type: token
292
+ model: ViTLarge_BaseDecoder
293
+ name: croco
294
+ patch_embed_cls: PatchEmbedDust3R
295
+ d_feature: 128
296
+ gaussian_adapter:
297
+ gaussian_scale_max: 15
298
+ gaussian_scale_min: 0.5
299
+ sh_degree: 4
300
+ gaussians_per_pixel: 1
301
+ gs_params_head_type: dpt_gs
302
+ name: noposplat
303
+ num_monocular_samples: 32
304
+ num_surfaces: 1
305
+ opacity_mapping:
306
+ final: 0
307
+ initial: 0
308
+ warm_up: 1
309
+ pose_free: true
310
+ predict_opacity: false
311
+ pretrained_weights: ""
312
+ visualizer:
313
+ export_ply: false
314
+ min_resolution: 256
315
+ num_samples: 8
316
+ vggt:
317
+ camera_prompt: false
318
+ detr:
319
+ base_HW: 64
320
+ debug_mode: 0
321
+ dec_layers: 1
322
+ dec_n_points: 4
323
+ decoder_type: conv
324
+ dim_down_factor: 1
325
+ dim_feedforward_factor: 4
326
+ dropout: 0.1
327
+ enc_layers: 6
328
+ enc_n_points: 4
329
+ features_num_channel:
330
+ - 128
331
+ - 32
332
+ grad_mode: absgrad
333
+ hidden_dim: 32
334
+ name: deformable_detr
335
+ num_feature_levels: 2
336
+ num_level: 3
337
+ num_queries: 256
338
+ only_last_level: false
339
+ position_embedding: sine
340
+ pre_norm: false
341
+ same_split_head: false
342
+ self_layer_num: 1
343
+ split_threshold: 1
344
+ split_threshold_range:
345
+ - 0
346
+ - 2
347
+ use_mean_features: true
348
+ head_mode: pcd
349
+ input_image_shape:
350
+ - 518
351
+ - 518
352
+ num_visual_prompt: 0
353
+ same_head: false
354
+ use_extrinsic_token: false
355
+ use_intrinsic_token: false
356
+ optimizer:
357
+ value:
358
+ accumulate: 1
359
+ backbone_lr_multiplier: 0.1
360
+ backbone_trainable: GB+H
361
+ lr: 0.0002
362
+ warm_up_steps: 125
363
+ seed:
364
+ value: 111123
365
+ test:
366
+ value:
367
+ align_pose: false
368
+ compute_scores: true
369
+ output_path: test
370
+ pose_align_steps: 100
371
+ pred_intrinsic: false
372
+ rot_opt_lr: 0.005
373
+ save_compare: false
374
+ save_image: false
375
+ save_video: false
376
+ split_threshold_list:
377
+ - 0
378
+ - 0.1
379
+ - 0.2
380
+ - 0.3
381
+ - 0.4
382
+ trans_opt_lr: 0.005
383
+ train:
384
+ value:
385
+ align_corners: false
386
+ aux_loss: 0
387
+ backbone_checkpoint: none
388
+ camera_consis_loss: 0
389
+ camera_loss: 1
390
+ debug_mode: 0
391
+ depth_mode: null
392
+ distill_max_steps: 1000000
393
+ distiller: ""
394
+ extended_visualization: false
395
+ extrinsic_drop: 0.1
396
+ intrinsic_drop: 0.1
397
+ intrinsic_scaling: true
398
+ one_sample_train: false
399
+ print_log_every_n_steps: 10
400
+ random_split_strategy: score
401
+ return_level_gs: true
402
+ scaling_detach: true
403
+ split_loss: 0.01
404
+ split_score_grad_scale: 10000
405
+ split_score_log_scale: true
406
+ target_scale_mode: 2
407
+ train_context: false
408
+ use_pred_point_focal: false
409
+ trainer:
410
+ value:
411
+ gradient_clip_val: 0.5
412
+ max_steps: 18751
413
+ num_nodes: 1
414
+ val_check_interval: 500
415
+ wandb:
416
+ value:
417
+ entity: scene-representation-group
418
+ mode: online
419
+ name: re10k_hpc
420
+ project: noposplat
421
+ tags:
422
+ - re10k
423
+ - 256x256
wandb/run-20250916_180036-llem5odx/files/output.log ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0]
2
+
3
+ | Name | Type | Params | Mode
4
+ ---------------------------------------------------------
5
+ 0 | vggt | VGGT | 585 M | train
6
+ 1 | decoder | DecoderSplattingCUDA | 0 | train
7
+ 2 | losses | ModuleList | 0 | train
8
+ ---------------------------------------------------------
9
+ 585 M Trainable params
10
+ 0 Non-trainable params
11
+ 585 M Total params
12
+ 2,340.472 Total estimated model params size (MB)
13
+ 773 Modules in train mode
14
+ 980 Modules in eval mode
15
+ Sanity Checking: | | 0/? [00:00<?, ?it/s][2025-09-16 18:00:39,175][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=279` in the `DataLoader` to improve performance.
16
+
17
+ Validation epoch start on rank 0
18
+ Sanity Checking DataLoader 0: 0%| | 0/1 [00:00<?, ?it/s]validation step 0; scene = ['306e2b7785657539']; context = [[48, 73]]
19
+ target intrinsic: tensor(0.8595, device='cuda:0') tensor(0.8597, device='cuda:0')
20
+ pred intrinsic: tensor(0.8780, device='cuda:0') tensor(0.8776, device='cuda:0')
21
+ [2025-09-16 18:00:41,384][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
22
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
23
+ warnings.warn(
24
+
25
+ [2025-09-16 18:01:31,562][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
26
+
27
+ Setting up [LPIPS] perceptual loss: trunk [vgg], v[0.1], spatial [off]
28
+ [2025-09-16 18:01:31,563][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
29
+ warnings.warn(
30
+
31
+ [2025-09-16 18:01:31,564][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
32
+ warnings.warn(msg)
33
+
34
+ Loading model from: /opt/conda/envs/ours/lib/python3.12/site-packages/lpips/weights/v0.1/vgg.pth
35
+ [2025-09-16 18:01:33,246][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
36
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
37
+
38
+ Epoch 0: | | 9/? [00:26<00:00, 0.34it/s, v_num=5odx]
39
+
40
+ Detected KeyboardInterrupt, attempting graceful shutdown ...
wandb/run-20250916_180036-llem5odx/files/requirements.txt ADDED
@@ -0,0 +1,129 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ruff==0.11.4
2
+ PyYAML==6.0.2
3
+ sympy==1.13.3
4
+ fsspec==2024.6.1
5
+ imageio-ffmpeg==0.6.0
6
+ setuptools==78.1.1
7
+ timm==1.0.15
8
+ fonttools==4.57.0
9
+ nvidia-cublas-cu12==12.8.3.14
10
+ nvidia-cuda-nvrtc-cu12==12.8.61
11
+ dacite==1.9.2
12
+ hydra-core==1.3.2
13
+ certifi==2022.12.7
14
+ rich==14.0.0
15
+ filelock==3.13.1
16
+ idna==3.4
17
+ nvidia-cuda-cupti-cu12==12.8.57
18
+ pip==25.2
19
+ wandb==0.19.9
20
+ ninja==1.11.1.4
21
+ networkx==3.3
22
+ GitPython==3.1.44
23
+ moviepy==1.0.3
24
+ markdown-it-py==3.0.0
25
+ lightning==2.5.1
26
+ Jinja2==3.1.4
27
+ nvidia-cusparse-cu12==12.5.7.53
28
+ tabulate==0.9.0
29
+ tifffile==2025.3.30
30
+ nvidia-nvtx-cu12==12.8.55
31
+ jaxtyping==0.2.37
32
+ aiohappyeyeballs==2.6.1
33
+ MarkupSafe==2.1.5
34
+ pydantic_core==2.33.1
35
+ mypy-extensions==1.0.0
36
+ lazy_loader==0.4
37
+ pydantic==2.11.2
38
+ typing-inspection==0.4.0
39
+ pillow==10.4.0
40
+ nvidia-nvjitlink-cu12==12.8.61
41
+ omegaconf==2.3.0
42
+ scikit-image==0.25.2
43
+ jaraco.context==5.3.0
44
+ typing_extensions==4.12.2
45
+ mpmath==1.3.0
46
+ opencv-python==4.11.0.86
47
+ backports.tarfile==1.2.0
48
+ async-timeout==5.0.1
49
+ opt_einsum==3.4.0
50
+ nvidia-cuda-runtime-cu12==12.8.57
51
+ numpy==1.26.4
52
+ torch==2.7.1+cu128
53
+ psutil==7.0.0
54
+ pytorch-lightning==2.5.1
55
+ protobuf==5.29.4
56
+ imageio==2.37.0
57
+ colorama==0.4.6
58
+ attrs==25.3.0
59
+ importlib_metadata==8.0.0
60
+ frozenlist==1.5.0
61
+ beartype==0.19.0
62
+ propcache==0.3.1
63
+ typeguard==4.3.0
64
+ torchaudio==2.7.1+cu128
65
+ decorator==4.4.2
66
+ nvidia-cudnn-cu12==9.7.1.26
67
+ torchvision==0.22.1+cu128
68
+ sentry-sdk==2.25.1
69
+ mdurl==0.1.2
70
+ more-itertools==10.3.0
71
+ einops==0.8.1
72
+ zipp==3.19.2
73
+ antlr4-python3-runtime==4.9.3
74
+ cycler==0.12.1
75
+ nvidia-cufile-cu12==1.13.0.11
76
+ jaraco.collections==5.1.0
77
+ jaraco.text==3.12.1
78
+ torchmetrics==1.8.2
79
+ packaging==24.2
80
+ black==25.1.0
81
+ setproctitle==1.3.5
82
+ Pygments==2.19.2
83
+ tomli==2.2.1
84
+ six==1.17.0
85
+ docker-pycreds==0.4.0
86
+ inflect==7.3.1
87
+ requests==2.28.1
88
+ svg.py==1.6.0
89
+ python-dateutil==2.9.0.post0
90
+ proglog==0.1.11
91
+ gsplat==1.5.3
92
+ e3nn==0.5.6
93
+ pathspec==0.12.1
94
+ aiosignal==1.3.2
95
+ wheel==0.45.1
96
+ multidict==6.3.2
97
+ pyparsing==3.2.3
98
+ plyfile==1.1
99
+ jaraco.functools==4.0.1
100
+ nvidia-nccl-cu12==2.26.2
101
+ triton==3.3.1
102
+ nvidia-cusparselt-cu12==0.6.3
103
+ scikit-video==1.1.11
104
+ huggingface-hub==0.30.1
105
+ lightning-utilities==0.14.3
106
+ wadler_lindig==0.1.4
107
+ nvidia-curand-cu12==10.3.9.55
108
+ matplotlib==3.10.1
109
+ kiwisolver==1.4.8
110
+ tqdm==4.67.1
111
+ safetensors==0.5.3
112
+ scipy==1.15.2
113
+ lpips==0.1.4
114
+ nvidia-cusolver-cu12==11.7.2.55
115
+ click==8.1.8
116
+ charset-normalizer==2.1.1
117
+ platformdirs==4.3.7
118
+ opt-einsum-fx==0.1.4
119
+ annotated-types==0.7.0
120
+ smmap==5.0.2
121
+ gitdb==4.0.12
122
+ nvidia-cufft-cu12==11.3.3.41
123
+ contourpy==1.3.1
124
+ urllib3==1.26.13
125
+ autocommand==2.2.2
126
+ yarl==1.19.0
127
+ aiohttp==3.11.16
128
+ colorspacious==1.1.2
129
+ python-dotenv==1.1.0
wandb/run-20250916_180036-llem5odx/files/wandb-metadata.json ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.0-153-generic-x86_64-with-glibc2.35",
3
+ "python": "CPython 3.12.11",
4
+ "startedAt": "2025-09-16T10:00:36.796366Z",
5
+ "args": [
6
+ "+experiment=re10k",
7
+ "wandb.mode=online",
8
+ "wandb.name=re10k_hpc"
9
+ ],
10
+ "program": "-m src.main",
11
+ "git": {
12
+ "remote": "git@github.com:K-nowing/ICLR2026_HCP.git",
13
+ "commit": "6c0080eae0295d43dd7f18fd71e742d924238f89"
14
+ },
15
+ "email": "dna9041@korea.ac.kr",
16
+ "root": "/root/highspeedstorage/injae/code/ICLR2026_HCP/outputs/re10k_hpc",
17
+ "host": "notebook-72e474dd-92e1-11f0-a6ea-7ab79f1c53cb-0",
18
+ "executable": "/opt/conda/envs/ours/bin/python",
19
+ "cpu_count": 144,
20
+ "cpu_count_logical": 288,
21
+ "gpu": "NVIDIA B200",
22
+ "gpu_count": 8,
23
+ "disk": {
24
+ "/": {
25
+ "total": "1081101176832",
26
+ "used": "211821359104"
27
+ }
28
+ },
29
+ "memory": {
30
+ "total": "2434690400256"
31
+ },
32
+ "cpu": {
33
+ "count": 144,
34
+ "countLogical": 288
35
+ },
36
+ "gpu_nvidia": [
37
+ {
38
+ "name": "NVIDIA B200",
39
+ "memoryTotal": "192265846784",
40
+ "cudaCores": 18944
41
+ },
42
+ {
43
+ "name": "NVIDIA B200",
44
+ "memoryTotal": "192265846784",
45
+ "cudaCores": 18944
46
+ },
47
+ {
48
+ "name": "NVIDIA B200",
49
+ "memoryTotal": "192265846784",
50
+ "cudaCores": 18944
51
+ },
52
+ {
53
+ "name": "NVIDIA B200",
54
+ "memoryTotal": "192265846784",
55
+ "cudaCores": 18944
56
+ },
57
+ {
58
+ "name": "NVIDIA B200",
59
+ "memoryTotal": "192265846784",
60
+ "cudaCores": 18944
61
+ },
62
+ {
63
+ "name": "NVIDIA B200",
64
+ "memoryTotal": "192265846784",
65
+ "cudaCores": 18944
66
+ },
67
+ {
68
+ "name": "NVIDIA B200",
69
+ "memoryTotal": "192265846784",
70
+ "cudaCores": 18944
71
+ },
72
+ {
73
+ "name": "NVIDIA B200",
74
+ "memoryTotal": "192265846784",
75
+ "cudaCores": 18944
76
+ }
77
+ ],
78
+ "cudaVersion": "12.8"
79
+ }
wandb/run-20250916_180036-llem5odx/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation_visualization":{"_type":"images/separated","width":256,"height":256,"format":"png","count":1,"filenames":["media/images/activation_visualization_2_e7083dae4b9eda33b264.png"],"captions":["306e2b7785657539"]},"_step":3,"comparison":{"_type":"images/separated","width":1064,"height":1098,"format":"png","count":1,"filenames":["media/images/comparison_0_621c28fb8c1f6ff0dbb4.png"],"captions":["306e2b7785657539"]},"_timestamp":1.7580168936113427e+09,"trainer/global_step":0,"points_context_imgs":{"height":800,"format":"png","count":1,"filenames":["media/images/points_context_imgs_1_cbb5017831dfea54f21d.png"],"captions":["306e2b7785657539"],"_type":"images/separated","width":536},"attention_map":{"width":800,"height":536,"format":"png","count":1,"filenames":["media/images/attention_map_3_4bf1d5f13f9f4f399591.png"],"captions":["306e2b7785657539"],"_type":"images/separated"},"_wandb":{"runtime":96},"_runtime":56.816239772}
wandb/run-20250916_180036-llem5odx/logs/debug.log ADDED
The diff for this file is too large to render. See raw diff
 
wandb/run-20250916_180036-llem5odx/run-llem5odx.wandb ADDED
Binary file (45.6 kB). View file
 
wandb/run-20250916_180413-d4pobtwb/files/output.log ADDED
@@ -0,0 +1,83 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [1]
2
+
3
+ | Name | Type | Params | Mode
4
+ ---------------------------------------------------------
5
+ 0 | vggt | VGGT | 585 M | train
6
+ 1 | decoder | DecoderSplattingCUDA | 0 | train
7
+ 2 | losses | ModuleList | 0 | train
8
+ ---------------------------------------------------------
9
+ 585 M Trainable params
10
+ 0 Non-trainable params
11
+ 585 M Total params
12
+ 2,340.472 Total estimated model params size (MB)
13
+ 773 Modules in train mode
14
+ 980 Modules in eval mode
15
+ Sanity Checking: | | 0/? [00:00<?, ?it/s][2025-09-16 18:04:15,683][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:425: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=279` in the `DataLoader` to improve performance.
16
+
17
+ Validation epoch start on rank 0
18
+ Sanity Checking DataLoader 0: 0%| | 0/1 [00:00<?, ?it/s]validation step 0; scene = ['306e2b7785657539']; context = [[48, 73]]
19
+ target intrinsic: tensor(0.8595, device='cuda:0') tensor(0.8597, device='cuda:0')
20
+ pred intrinsic: tensor(0.8780, device='cuda:0') tensor(0.8776, device='cuda:0')
21
+ [2025-09-16 18:04:17,971][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/utils/cpp_extension.py:2356: UserWarning: TORCH_CUDA_ARCH_LIST is not set, all archs for visible cards are included for compilation.
22
+ If this is not desired, please set os.environ['TORCH_CUDA_ARCH_LIST'].
23
+ warnings.warn(
24
+
25
+ [2025-09-16 18:04:18,161][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/lightning/pytorch/utilities/data.py:79: Trying to infer the `batch_size` from an ambiguous collection. The batch size we found is 1. To avoid any miscalculations, use `self.log(..., batch_size=batch_size)`.
26
+
27
+ Setting up [LPIPS] perceptual loss: trunk [vgg], v[0.1], spatial [off]
28
+ [2025-09-16 18:04:18,163][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
29
+ warnings.warn(
30
+
31
+ [2025-09-16 18:04:18,164][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=VGG16_Weights.IMAGENET1K_V1`. You can also use `weights=VGG16_Weights.DEFAULT` to get the most up-to-date weights.
32
+ warnings.warn(msg)
33
+
34
+ Loading model from: /opt/conda/envs/ours/lib/python3.12/site-packages/lpips/weights/v0.1/vgg.pth
35
+ [2025-09-16 18:04:19,836][py.warnings][WARNING] - /opt/conda/envs/ours/lib/python3.12/site-packages/torch/functional.py:554: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at /pytorch/aten/src/ATen/native/TensorShape.cpp:4314.)
36
+ return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
37
+
38
+ Epoch 0: | | 0/? [00:00<?, ?it/s]
39
+
40
+ Detected KeyboardInterrupt, attempting graceful shutdown ...
41
+ Exception ignored in: <module 'threading' from '/opt/conda/envs/ours/lib/python3.12/threading.py'>
42
+ Traceback (most recent call last):
43
+ File "/opt/conda/envs/ours/lib/python3.12/threading.py", line 1575, in _shutdown
44
+ def _shutdown():
45
+
46
+ KeyboardInterrupt:
47
+ Exception ignored in atexit callback: <function _start_and_connect_service.<locals>.teardown_atexit at 0x7f90148e3060>
48
+ Traceback (most recent call last):
49
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/service_connection.py", line 94, in teardown_atexit
50
+ conn.teardown(hooks.exit_code)
51
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/service_connection.py", line 226, in teardown
52
+ self._router.join()
53
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/router.py", line 75, in join
54
+ self._thread.join()
55
+ File "/opt/conda/envs/ours/lib/python3.12/threading.py", line 1149, in join
56
+ self._wait_for_tstate_lock()
57
+ File "/opt/conda/envs/ours/lib/python3.12/threading.py", line 1169, in _wait_for_tstate_lock
58
+ if lock.acquire(block, timeout):
59
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
60
+ KeyboardInterrupt:
61
+ Exception ignored in atexit callback: <function _exit_function at 0x7f90d829e160>
62
+ Traceback (most recent call last):
63
+ File "/opt/conda/envs/ours/lib/python3.12/multiprocessing/util.py", line 337, in _exit_function
64
+ _run_finalizers(0)
65
+ File "/opt/conda/envs/ours/lib/python3.12/multiprocessing/util.py", line 303, in _run_finalizers
66
+ finalizer()
67
+ File "/opt/conda/envs/ours/lib/python3.12/multiprocessing/util.py", line 227, in __call__
68
+ res = self._callback(*self._args, **self._kwargs)
69
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
70
+ File "/opt/conda/envs/ours/lib/python3.12/multiprocessing/managers.py", line 675, in _finalize_manager
71
+ process.join(timeout=shutdown_timeout)
72
+ File "/opt/conda/envs/ours/lib/python3.12/multiprocessing/process.py", line 149, in join
73
+ res = self._popen.wait(timeout)
74
+ ^^^^^^^^^^^^^^^^^^^^^^^^^
75
+ File "/opt/conda/envs/ours/lib/python3.12/multiprocessing/popen_fork.py", line 40, in wait
76
+ if not wait([self.sentinel], timeout):
77
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
78
+ File "/opt/conda/envs/ours/lib/python3.12/multiprocessing/connection.py", line 1136, in wait
79
+ ready = selector.select(timeout)
80
+ ^^^^^^^^^^^^^^^^^^^^^^^^
81
+ File "/opt/conda/envs/ours/lib/python3.12/selectors.py", line 415, in select
82
+ fd_event_list = self._selector.poll(timeout)
83
+ ^^^^^^^^^^^^^^KeyboardInterrupt:
wandb/run-20250916_180413-d4pobtwb/files/requirements.txt ADDED
@@ -0,0 +1,129 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ruff==0.11.4
2
+ PyYAML==6.0.2
3
+ sympy==1.13.3
4
+ fsspec==2024.6.1
5
+ imageio-ffmpeg==0.6.0
6
+ setuptools==78.1.1
7
+ timm==1.0.15
8
+ fonttools==4.57.0
9
+ nvidia-cublas-cu12==12.8.3.14
10
+ nvidia-cuda-nvrtc-cu12==12.8.61
11
+ dacite==1.9.2
12
+ hydra-core==1.3.2
13
+ certifi==2022.12.7
14
+ rich==14.0.0
15
+ filelock==3.13.1
16
+ idna==3.4
17
+ nvidia-cuda-cupti-cu12==12.8.57
18
+ pip==25.2
19
+ wandb==0.19.9
20
+ ninja==1.11.1.4
21
+ networkx==3.3
22
+ GitPython==3.1.44
23
+ moviepy==1.0.3
24
+ markdown-it-py==3.0.0
25
+ lightning==2.5.1
26
+ Jinja2==3.1.4
27
+ nvidia-cusparse-cu12==12.5.7.53
28
+ tabulate==0.9.0
29
+ tifffile==2025.3.30
30
+ nvidia-nvtx-cu12==12.8.55
31
+ jaxtyping==0.2.37
32
+ aiohappyeyeballs==2.6.1
33
+ MarkupSafe==2.1.5
34
+ pydantic_core==2.33.1
35
+ mypy-extensions==1.0.0
36
+ lazy_loader==0.4
37
+ pydantic==2.11.2
38
+ typing-inspection==0.4.0
39
+ pillow==10.4.0
40
+ nvidia-nvjitlink-cu12==12.8.61
41
+ omegaconf==2.3.0
42
+ scikit-image==0.25.2
43
+ jaraco.context==5.3.0
44
+ typing_extensions==4.12.2
45
+ mpmath==1.3.0
46
+ opencv-python==4.11.0.86
47
+ backports.tarfile==1.2.0
48
+ async-timeout==5.0.1
49
+ opt_einsum==3.4.0
50
+ nvidia-cuda-runtime-cu12==12.8.57
51
+ numpy==1.26.4
52
+ torch==2.7.1+cu128
53
+ psutil==7.0.0
54
+ pytorch-lightning==2.5.1
55
+ protobuf==5.29.4
56
+ imageio==2.37.0
57
+ colorama==0.4.6
58
+ attrs==25.3.0
59
+ importlib_metadata==8.0.0
60
+ frozenlist==1.5.0
61
+ beartype==0.19.0
62
+ propcache==0.3.1
63
+ typeguard==4.3.0
64
+ torchaudio==2.7.1+cu128
65
+ decorator==4.4.2
66
+ nvidia-cudnn-cu12==9.7.1.26
67
+ torchvision==0.22.1+cu128
68
+ sentry-sdk==2.25.1
69
+ mdurl==0.1.2
70
+ more-itertools==10.3.0
71
+ einops==0.8.1
72
+ zipp==3.19.2
73
+ antlr4-python3-runtime==4.9.3
74
+ cycler==0.12.1
75
+ nvidia-cufile-cu12==1.13.0.11
76
+ jaraco.collections==5.1.0
77
+ jaraco.text==3.12.1
78
+ torchmetrics==1.8.2
79
+ packaging==24.2
80
+ black==25.1.0
81
+ setproctitle==1.3.5
82
+ Pygments==2.19.2
83
+ tomli==2.2.1
84
+ six==1.17.0
85
+ docker-pycreds==0.4.0
86
+ inflect==7.3.1
87
+ requests==2.28.1
88
+ svg.py==1.6.0
89
+ python-dateutil==2.9.0.post0
90
+ proglog==0.1.11
91
+ gsplat==1.5.3
92
+ e3nn==0.5.6
93
+ pathspec==0.12.1
94
+ aiosignal==1.3.2
95
+ wheel==0.45.1
96
+ multidict==6.3.2
97
+ pyparsing==3.2.3
98
+ plyfile==1.1
99
+ jaraco.functools==4.0.1
100
+ nvidia-nccl-cu12==2.26.2
101
+ triton==3.3.1
102
+ nvidia-cusparselt-cu12==0.6.3
103
+ scikit-video==1.1.11
104
+ huggingface-hub==0.30.1
105
+ lightning-utilities==0.14.3
106
+ wadler_lindig==0.1.4
107
+ nvidia-curand-cu12==10.3.9.55
108
+ matplotlib==3.10.1
109
+ kiwisolver==1.4.8
110
+ tqdm==4.67.1
111
+ safetensors==0.5.3
112
+ scipy==1.15.2
113
+ lpips==0.1.4
114
+ nvidia-cusolver-cu12==11.7.2.55
115
+ click==8.1.8
116
+ charset-normalizer==2.1.1
117
+ platformdirs==4.3.7
118
+ opt-einsum-fx==0.1.4
119
+ annotated-types==0.7.0
120
+ smmap==5.0.2
121
+ gitdb==4.0.12
122
+ nvidia-cufft-cu12==11.3.3.41
123
+ contourpy==1.3.1
124
+ urllib3==1.26.13
125
+ autocommand==2.2.2
126
+ yarl==1.19.0
127
+ aiohttp==3.11.16
128
+ colorspacious==1.1.2
129
+ python-dotenv==1.1.0
wandb/run-20250916_180413-d4pobtwb/files/wandb-metadata.json ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.0-153-generic-x86_64-with-glibc2.35",
3
+ "python": "CPython 3.12.11",
4
+ "startedAt": "2025-09-16T10:04:13.466109Z",
5
+ "args": [
6
+ "+experiment=re10k",
7
+ "wandb.mode=online",
8
+ "wandb.name=re10k_hpc"
9
+ ],
10
+ "program": "-m src.main",
11
+ "git": {
12
+ "remote": "git@github.com:K-nowing/ICLR2026_HCP.git",
13
+ "commit": "6c0080eae0295d43dd7f18fd71e742d924238f89"
14
+ },
15
+ "email": "dna9041@korea.ac.kr",
16
+ "root": "/root/highspeedstorage/injae/code/ICLR2026_HCP/outputs/re10k_hpc",
17
+ "host": "notebook-72e474dd-92e1-11f0-a6ea-7ab79f1c53cb-0",
18
+ "executable": "/opt/conda/envs/ours/bin/python",
19
+ "cpu_count": 144,
20
+ "cpu_count_logical": 288,
21
+ "gpu": "NVIDIA B200",
22
+ "gpu_count": 8,
23
+ "disk": {
24
+ "/": {
25
+ "total": "1081101176832",
26
+ "used": "211821613056"
27
+ }
28
+ },
29
+ "memory": {
30
+ "total": "2434690400256"
31
+ },
32
+ "cpu": {
33
+ "count": 144,
34
+ "countLogical": 288
35
+ },
36
+ "gpu_nvidia": [
37
+ {
38
+ "name": "NVIDIA B200",
39
+ "memoryTotal": "192265846784",
40
+ "cudaCores": 18944
41
+ },
42
+ {
43
+ "name": "NVIDIA B200",
44
+ "memoryTotal": "192265846784",
45
+ "cudaCores": 18944
46
+ },
47
+ {
48
+ "name": "NVIDIA B200",
49
+ "memoryTotal": "192265846784",
50
+ "cudaCores": 18944
51
+ },
52
+ {
53
+ "name": "NVIDIA B200",
54
+ "memoryTotal": "192265846784",
55
+ "cudaCores": 18944
56
+ },
57
+ {
58
+ "name": "NVIDIA B200",
59
+ "memoryTotal": "192265846784",
60
+ "cudaCores": 18944
61
+ },
62
+ {
63
+ "name": "NVIDIA B200",
64
+ "memoryTotal": "192265846784",
65
+ "cudaCores": 18944
66
+ },
67
+ {
68
+ "name": "NVIDIA B200",
69
+ "memoryTotal": "192265846784",
70
+ "cudaCores": 18944
71
+ },
72
+ {
73
+ "name": "NVIDIA B200",
74
+ "memoryTotal": "192265846784",
75
+ "cudaCores": 18944
76
+ }
77
+ ],
78
+ "cudaVersion": "12.8"
79
+ }
wandb/run-20250916_180413-d4pobtwb/logs/debug-core.log ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {"time":"2025-09-16T18:04:13.103986093+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpv2hrevc1/port-6788.txt","pid":6788,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false}
2
+ {"time":"2025-09-16T18:04:13.107503769+08:00","level":"INFO","msg":"Will exit if parent process dies.","ppid":6788}
3
+ {"time":"2025-09-16T18:04:13.107294572+08:00","level":"INFO","msg":"server is running","addr":{"IP":"127.0.0.1","Port":37499,"Zone":""}}
4
+ {"time":"2025-09-16T18:04:13.18584671+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"127.0.0.1:42398"}
5
+ {"time":"2025-09-16T18:04:13.469622805+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"d4pobtwb","id":"127.0.0.1:42398"}
6
+ {"time":"2025-09-16T18:04:13.944559071+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"d4pobtwb","id":"127.0.0.1:42398"}
7
+ {"time":"2025-09-16T18:04:27.095445631+08:00","level":"INFO","msg":"Parent process exited, terminating service process."}
wandb/run-20250916_180413-d4pobtwb/logs/debug-internal.log ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {"time":"2025-09-16T18:04:13.471240905+08:00","level":"INFO","msg":"stream: starting","core version":"0.19.9","symlink path":"/root/highspeedstorage/injae/code/ICLR2026_HCP/outputs/re10k_hpc/wandb/run-20250916_180413-d4pobtwb/logs/debug-core.log"}
2
+ {"time":"2025-09-16T18:04:13.944337794+08:00","level":"INFO","msg":"created new stream","id":"d4pobtwb"}
3
+ {"time":"2025-09-16T18:04:13.944517672+08:00","level":"INFO","msg":"stream: started","id":"d4pobtwb"}
4
+ {"time":"2025-09-16T18:04:13.944738149+08:00","level":"INFO","msg":"writer: Do: started","stream_id":"d4pobtwb"}
5
+ {"time":"2025-09-16T18:04:13.944986426+08:00","level":"INFO","msg":"sender: started","stream_id":"d4pobtwb"}
6
+ {"time":"2025-09-16T18:04:13.945237383+08:00","level":"INFO","msg":"handler: started","stream_id":"d4pobtwb"}
7
+ {"time":"2025-09-16T18:04:14.217447222+08:00","level":"INFO","msg":"Starting system monitor"}
wandb/run-20250916_180413-d4pobtwb/logs/debug.log ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-09-16 18:04:13,440 INFO MainThread:6788 [wandb_setup.py:_flush():67] Current SDK version is 0.19.9
2
+ 2025-09-16 18:04:13,440 INFO MainThread:6788 [wandb_setup.py:_flush():67] Configure stats pid to 6788
3
+ 2025-09-16 18:04:13,440 INFO MainThread:6788 [wandb_setup.py:_flush():67] Loading settings from /root/.config/wandb/settings
4
+ 2025-09-16 18:04:13,440 INFO MainThread:6788 [wandb_setup.py:_flush():67] Loading settings from /root/highspeedstorage/injae/code/ICLR2026_HCP/wandb/settings
5
+ 2025-09-16 18:04:13,440 INFO MainThread:6788 [wandb_setup.py:_flush():67] Loading settings from environment variables
6
+ 2025-09-16 18:04:13,441 INFO MainThread:6788 [wandb_init.py:setup_run_log_directory():662] Logging user logs to /root/highspeedstorage/injae/code/ICLR2026_HCP/outputs/re10k_hpc/wandb/run-20250916_180413-d4pobtwb/logs/debug.log
7
+ 2025-09-16 18:04:13,441 INFO MainThread:6788 [wandb_init.py:setup_run_log_directory():663] Logging internal logs to /root/highspeedstorage/injae/code/ICLR2026_HCP/outputs/re10k_hpc/wandb/run-20250916_180413-d4pobtwb/logs/debug-internal.log
8
+ 2025-09-16 18:04:13,441 INFO MainThread:6788 [wandb_init.py:init():781] calling init triggers
9
+ 2025-09-16 18:04:13,442 INFO MainThread:6788 [wandb_init.py:init():786] wandb.init called with sweep_config: {}
10
+ config: {'model': {'encoder': {'backbone': {'name': 'croco', 'model': 'ViTLarge_BaseDecoder', 'patch_embed_cls': 'PatchEmbedDust3R', 'asymmetry_decoder': True, 'intrinsics_embed_loc': 'encoder', 'intrinsics_embed_degree': 4, 'intrinsics_embed_type': 'token'}, 'name': 'noposplat', 'opacity_mapping': {'initial': 0.0, 'final': 0.0, 'warm_up': 1}, 'num_monocular_samples': 32, 'num_surfaces': 1, 'predict_opacity': False, 'gaussians_per_pixel': 1, 'gaussian_adapter': {'gaussian_scale_min': 0.5, 'gaussian_scale_max': 15.0, 'sh_degree': 4}, 'd_feature': 128, 'visualizer': {'num_samples': 8, 'min_resolution': 256, 'export_ply': False}, 'apply_bounds_shim': True, 'gs_params_head_type': 'dpt_gs', 'pose_free': True, 'pretrained_weights': ''}, 'decoder': {'name': 'splatting_cuda', 'background_color': [0.0, 0.0, 0.0], 'make_scale_invariant': False, 'use_gsplat': True}, 'vggt': {'input_image_shape': [518, 518], 'head_mode': 'pcd', 'same_head': False, 'camera_prompt': False, 'num_visual_prompt': 0, 'use_extrinsic_token': False, 'use_intrinsic_token': False, 'detr': {'name': 'deformable_detr', 'features_num_channel': [128, 32], 'hidden_dim': 32, 'dim_down_factor': 1, 'dim_feedforward_factor': 4, 'enc_layers': 6, 'dec_layers': 1, 'dropout': 0.1, 'num_feature_levels': 2, 'dec_n_points': 4, 'enc_n_points': 4, 'split_threshold': 1.0, 'num_queries': 256, 'base_HW': 64, 'position_embedding': 'sine', 'pre_norm': False, 'self_layer_num': 1, 'num_level': 3, 'debug_mode': 0, 'decoder_type': 'conv', 'split_threshold_range': [0.0, 2.0], 'only_last_level': False, 'grad_mode': 'absgrad', 'same_split_head': False, 'use_mean_features': True}}}, 'loss': {'mse': {'weight': 1.0}, 'lpips': {'weight': 0.05, 'apply_after_step': 0}}, 'wandb': {'project': 'noposplat', 'entity': 'scene-representation-group', 'name': 're10k_hpc', 'mode': 'online', 'tags': ['re10k', '256x256']}, 'mode': 'train', 'data_loader': {'train': {'num_workers': 16, 'persistent_workers': True, 'batch_size': 16, 'seed': 1234}, 'test': {'num_workers': 4, 'persistent_workers': False, 'batch_size': 1, 'seed': 2345}, 'val': {'num_workers': 1, 'persistent_workers': True, 'batch_size': 1, 'seed': 3456}}, 'optimizer': {'lr': 0.0002, 'warm_up_steps': 125, 'backbone_lr_multiplier': 0.1, 'backbone_trainable': 'GB+H', 'accumulate': 1}, 'checkpointing': {'load': None, 'every_n_train_steps': 9375, 'save_top_k': 1, 'save_weights_only': True}, 'train': {'depth_mode': None, 'extended_visualization': False, 'print_log_every_n_steps': 10, 'distiller': '', 'distill_max_steps': 1000000, 'debug_mode': 0, 'camera_consis_loss': 0.0, 'camera_loss': 1.0, 'aux_loss': 0.0, 'one_sample_train': False, 'target_scale_mode': 2, 'extrinsic_drop': 0.1, 'intrinsic_drop': 0.1, 'align_corners': False, 'backbone_checkpoint': 'none', 'intrinsic_scaling': True, 'return_level_gs': True, 'train_context': False, 'use_pred_point_focal': False, 'scaling_detach': True, 'split_score_log_scale': True, 'split_score_grad_scale': 10000.0, 'split_loss': 0.01, 'random_split_strategy': 'score'}, 'test': {'output_path': 'test', 'align_pose': False, 'pose_align_steps': 100, 'rot_opt_lr': 0.005, 'trans_opt_lr': 0.005, 'compute_scores': True, 'save_image': False, 'save_video': False, 'save_compare': False, 'split_threshold_list': [0.0, 0.1, 0.2, 0.3, 0.4], 'pred_intrinsic': False}, 'seed': 111123, 'trainer': {'max_steps': 18751, 'val_check_interval': 500, 'gradient_clip_val': 0.5, 'num_nodes': 1}, 'dataset': {'re10k': {'make_baseline_1': True, 'relative_pose': True, 'augment': True, 'background_color': [0.0, 0.0, 0.0], 'overfit_to_scene': None, 'skip_bad_shape': True, 'view_sampler': {'name': 'bounded', 'num_target_views': 4, 'num_context_views': 2, 'min_distance_between_context_views': 45, 'max_distance_between_context_views': 90, 'min_distance_to_context_views': 0, 'warm_up_steps': 9375, 'initial_min_distance_between_context_views': 25, 'initial_max_distance_between_context_views': 25, 'distill_steps': 0}, 'name': 're10k', 'roots': ['datasets/re10k'], 'input_image_shape': [256, 256], 'original_image_shape': [360, 640], 'cameras_are_circular': False, 'baseline_min': 0.001, 'baseline_max': 10000000000.0, 'max_fov': 100.0}}, '_wandb': {}}
11
+ 2025-09-16 18:04:13,442 INFO MainThread:6788 [wandb_init.py:init():809] starting backend
12
+ 2025-09-16 18:04:13,442 INFO MainThread:6788 [wandb_init.py:init():813] sending inform_init request
13
+ 2025-09-16 18:04:13,464 INFO MainThread:6788 [backend.py:_multiprocessing_setup():101] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
14
+ 2025-09-16 18:04:13,464 INFO MainThread:6788 [wandb_init.py:init():823] backend started and connected
15
+ 2025-09-16 18:04:13,488 INFO MainThread:6788 [wandb_init.py:init():915] updated telemetry
16
+ 2025-09-16 18:04:13,535 INFO MainThread:6788 [wandb_init.py:init():939] communicating run to backend with 90.0 second timeout
17
+ 2025-09-16 18:04:14,206 INFO MainThread:6788 [wandb_init.py:init():1014] starting run threads in backend
18
+ 2025-09-16 18:04:14,384 INFO MainThread:6788 [wandb_run.py:_console_start():2454] atexit reg
19
+ 2025-09-16 18:04:14,385 INFO MainThread:6788 [wandb_run.py:_redirect():2306] redirect: wrap_raw
20
+ 2025-09-16 18:04:14,388 INFO MainThread:6788 [wandb_run.py:_redirect():2371] Wrapping output streams.
21
+ 2025-09-16 18:04:14,388 INFO MainThread:6788 [wandb_run.py:_redirect():2394] Redirects installed.
22
+ 2025-09-16 18:04:14,392 INFO MainThread:6788 [wandb_init.py:init():1056] run started, returning control to user process
wandb/run-20250916_180413-d4pobtwb/run-d4pobtwb.wandb ADDED
Binary file (32.8 kB). View file
 
wandb/run-20250916_180427-9sp0azf6/run-9sp0azf6.wandb ADDED
File without changes
wandb/run-20250916_180450-bbofdbql/run-bbofdbql.wandb ADDED
Binary file (97 kB). View file
 
wandb/run-20250916_180500-j938t4jj/run-j938t4jj.wandb ADDED
Binary file (95.8 kB). View file
 
wandb/run-20250916_194552-aaciqyuf/files/config.yaml ADDED
@@ -0,0 +1,555 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.19.9
4
+ m:
5
+ - "1": loss/split_score
6
+ "5": 2
7
+ "6":
8
+ - 1
9
+ - 3
10
+ "7": []
11
+ - "1": trainer/global_step
12
+ "6":
13
+ - 3
14
+ "7": []
15
+ - "1": comparison.height
16
+ "5": 2
17
+ "6":
18
+ - 1
19
+ - 3
20
+ "7": []
21
+ - "1": points_context_imgs.captions
22
+ "5": 2
23
+ "6":
24
+ - 1
25
+ - 3
26
+ "7": []
27
+ - "1": points_context_imgs.format
28
+ "5": 2
29
+ "6":
30
+ - 1
31
+ - 3
32
+ "7": []
33
+ - "1": points_context_imgs.count
34
+ "5": 2
35
+ "6":
36
+ - 1
37
+ - 3
38
+ "7": []
39
+ - "1": activation_visualization.format
40
+ "5": 2
41
+ "6":
42
+ - 1
43
+ - 3
44
+ "7": []
45
+ - "1": lr-AdamW/pg2
46
+ "5": 2
47
+ "6":
48
+ - 1
49
+ - 3
50
+ "7": []
51
+ - "1": loss/aux_1/mse
52
+ "5": 2
53
+ "6":
54
+ - 1
55
+ - 3
56
+ "7": []
57
+ - "1": loss/total
58
+ "5": 2
59
+ "6":
60
+ - 1
61
+ - 3
62
+ "7": []
63
+ - "1": val/ssim
64
+ "5": 2
65
+ "6":
66
+ - 1
67
+ - 3
68
+ "7": []
69
+ - "1": activation_visualization._type
70
+ "5": 2
71
+ "6":
72
+ - 1
73
+ - 3
74
+ "7": []
75
+ - "1": attention_map.filenames
76
+ "5": 2
77
+ "6":
78
+ - 1
79
+ - 3
80
+ "7": []
81
+ - "1": loss/aux_0/mse
82
+ "5": 2
83
+ "6":
84
+ - 1
85
+ - 3
86
+ "7": []
87
+ - "1": comparison.captions
88
+ "5": 2
89
+ "6":
90
+ - 1
91
+ - 3
92
+ "7": []
93
+ - "1": activation_visualization.height
94
+ "5": 2
95
+ "6":
96
+ - 1
97
+ - 3
98
+ "7": []
99
+ - "1": attention_map.count
100
+ "5": 2
101
+ "6":
102
+ - 1
103
+ - 3
104
+ "7": []
105
+ - "1": loss/camera
106
+ "5": 2
107
+ "6":
108
+ - 1
109
+ - 3
110
+ "7": []
111
+ - "1": activation_visualization.filenames
112
+ "5": 2
113
+ "6":
114
+ - 1
115
+ - 3
116
+ "7": []
117
+ - "1": attention_map.width
118
+ "5": 2
119
+ "6":
120
+ - 1
121
+ - 3
122
+ "7": []
123
+ - "1": lr-AdamW/pg2-momentum
124
+ "5": 2
125
+ "6":
126
+ - 1
127
+ - 3
128
+ "7": []
129
+ - "1": val/lpips
130
+ "5": 2
131
+ "6":
132
+ - 1
133
+ - 3
134
+ "7": []
135
+ - "1": comparison.format
136
+ "5": 2
137
+ "6":
138
+ - 1
139
+ - 3
140
+ "7": []
141
+ - "1": comparison._type
142
+ "5": 2
143
+ "6":
144
+ - 1
145
+ - 3
146
+ "7": []
147
+ - "1": points_context_imgs.width
148
+ "5": 2
149
+ "6":
150
+ - 1
151
+ - 3
152
+ "7": []
153
+ - "1": activation_visualization.captions
154
+ "5": 2
155
+ "6":
156
+ - 1
157
+ - 3
158
+ "7": []
159
+ - "1": info/global_step
160
+ "5": 2
161
+ "6":
162
+ - 1
163
+ - 3
164
+ "7": []
165
+ - "1": loss/aux_2/lpips
166
+ "5": 2
167
+ "6":
168
+ - 1
169
+ - 3
170
+ "7": []
171
+ - "1": epoch
172
+ "5": 2
173
+ "6":
174
+ - 1
175
+ - 3
176
+ "7": []
177
+ - "1": points_context_imgs._type
178
+ "5": 2
179
+ "6":
180
+ - 1
181
+ - 3
182
+ "7": []
183
+ - "1": attention_map.captions
184
+ "5": 2
185
+ "6":
186
+ - 1
187
+ - 3
188
+ "7": []
189
+ - "1": attention_map._type
190
+ "5": 2
191
+ "6":
192
+ - 1
193
+ - 3
194
+ "7": []
195
+ - "1": attention_map.format
196
+ "5": 2
197
+ "6":
198
+ - 1
199
+ - 3
200
+ "7": []
201
+ - "1": lr-AdamW/pg1-momentum
202
+ "5": 2
203
+ "6":
204
+ - 1
205
+ - 3
206
+ "7": []
207
+ - "1": loss/aux_1/lpips
208
+ "5": 2
209
+ "6":
210
+ - 1
211
+ - 3
212
+ "7": []
213
+ - "1": train/psnr_probabilistic
214
+ "5": 2
215
+ "6":
216
+ - 1
217
+ - 3
218
+ "7": []
219
+ - "1": loss/aux_0/lpips
220
+ "5": 2
221
+ "6":
222
+ - 1
223
+ - 3
224
+ "7": []
225
+ - "1": comparison.width
226
+ "5": 2
227
+ "6":
228
+ - 1
229
+ - 3
230
+ "7": []
231
+ - "1": comparison.count
232
+ "5": 2
233
+ "6":
234
+ - 1
235
+ - 3
236
+ "7": []
237
+ - "1": comparison.filenames
238
+ "5": 2
239
+ "6":
240
+ - 1
241
+ - 3
242
+ "7": []
243
+ - "1": points_context_imgs.filenames
244
+ "5": 2
245
+ "6":
246
+ - 1
247
+ - 3
248
+ "7": []
249
+ - "1": activation_visualization.width
250
+ "5": 2
251
+ "6":
252
+ - 1
253
+ - 3
254
+ "7": []
255
+ - "1": activation_visualization.count
256
+ "5": 2
257
+ "6":
258
+ - 1
259
+ - 3
260
+ "7": []
261
+ - "1": lr-AdamW/pg1
262
+ "5": 2
263
+ "6":
264
+ - 1
265
+ - 3
266
+ "7": []
267
+ - "1": val/gaussian_num_ratio
268
+ "5": 2
269
+ "6":
270
+ - 1
271
+ - 3
272
+ "7": []
273
+ - "1": val/psnr
274
+ "5": 2
275
+ "6":
276
+ - 1
277
+ - 3
278
+ "7": []
279
+ - "1": points_context_imgs.height
280
+ "5": 2
281
+ "6":
282
+ - 1
283
+ - 3
284
+ "7": []
285
+ - "1": attention_map.height
286
+ "5": 2
287
+ "6":
288
+ - 1
289
+ - 3
290
+ "7": []
291
+ - "1": loss/3dgs_2_0/mse
292
+ "5": 2
293
+ "6":
294
+ - 1
295
+ - 3
296
+ "7": []
297
+ - "1": loss/aux_2/mse
298
+ "5": 2
299
+ "6":
300
+ - 1
301
+ - 3
302
+ "7": []
303
+ - "1": loss/3dgs_2_0/lpips
304
+ "5": 2
305
+ "6":
306
+ - 1
307
+ - 3
308
+ "7": []
309
+ python_version: 3.12.11
310
+ t:
311
+ "1":
312
+ - 1
313
+ - 41
314
+ - 49
315
+ - 50
316
+ - 55
317
+ - 106
318
+ "2":
319
+ - 1
320
+ - 41
321
+ - 49
322
+ - 50
323
+ - 55
324
+ - 106
325
+ "3":
326
+ - 7
327
+ - 13
328
+ - 15
329
+ - 16
330
+ - 23
331
+ - 55
332
+ - 66
333
+ "4": 3.12.11
334
+ "5": 0.19.9
335
+ "8":
336
+ - 5
337
+ "12": 0.19.9
338
+ "13": linux-x86_64
339
+ checkpointing:
340
+ value:
341
+ every_n_train_steps: 9375
342
+ load: null
343
+ save_top_k: 1
344
+ save_weights_only: true
345
+ data_loader:
346
+ value:
347
+ test:
348
+ batch_size: 1
349
+ num_workers: 4
350
+ persistent_workers: false
351
+ seed: 2345
352
+ train:
353
+ batch_size: 16
354
+ num_workers: 16
355
+ persistent_workers: true
356
+ seed: 1234
357
+ val:
358
+ batch_size: 1
359
+ num_workers: 1
360
+ persistent_workers: true
361
+ seed: 3456
362
+ dataset:
363
+ value:
364
+ re10k:
365
+ augment: true
366
+ background_color:
367
+ - 0
368
+ - 0
369
+ - 0
370
+ baseline_max: 1e+10
371
+ baseline_min: 0.001
372
+ cameras_are_circular: false
373
+ input_image_shape:
374
+ - 256
375
+ - 256
376
+ make_baseline_1: true
377
+ max_fov: 100
378
+ name: re10k
379
+ original_image_shape:
380
+ - 360
381
+ - 640
382
+ overfit_to_scene: null
383
+ relative_pose: true
384
+ roots:
385
+ - datasets/re10k
386
+ skip_bad_shape: true
387
+ view_sampler:
388
+ distill_steps: 0
389
+ initial_max_distance_between_context_views: 25
390
+ initial_min_distance_between_context_views: 25
391
+ max_distance_between_context_views: 90
392
+ min_distance_between_context_views: 45
393
+ min_distance_to_context_views: 0
394
+ name: bounded
395
+ num_context_views: 2
396
+ num_target_views: 4
397
+ warm_up_steps: 9375
398
+ loss:
399
+ value:
400
+ lpips:
401
+ apply_after_step: 0
402
+ weight: 0.05
403
+ mse:
404
+ weight: 1
405
+ mode:
406
+ value: train
407
+ model:
408
+ value:
409
+ decoder:
410
+ background_color:
411
+ - 0
412
+ - 0
413
+ - 0
414
+ make_scale_invariant: false
415
+ name: splatting_cuda
416
+ use_gsplat: true
417
+ encoder:
418
+ apply_bounds_shim: true
419
+ backbone:
420
+ asymmetry_decoder: true
421
+ intrinsics_embed_degree: 4
422
+ intrinsics_embed_loc: encoder
423
+ intrinsics_embed_type: token
424
+ model: ViTLarge_BaseDecoder
425
+ name: croco
426
+ patch_embed_cls: PatchEmbedDust3R
427
+ d_feature: 128
428
+ gaussian_adapter:
429
+ gaussian_scale_max: 15
430
+ gaussian_scale_min: 0.5
431
+ sh_degree: 4
432
+ gaussians_per_pixel: 1
433
+ gs_params_head_type: dpt_gs
434
+ name: noposplat
435
+ num_monocular_samples: 32
436
+ num_surfaces: 1
437
+ opacity_mapping:
438
+ final: 0
439
+ initial: 0
440
+ warm_up: 1
441
+ pose_free: true
442
+ predict_opacity: false
443
+ pretrained_weights: ""
444
+ visualizer:
445
+ export_ply: false
446
+ min_resolution: 256
447
+ num_samples: 8
448
+ vggt:
449
+ camera_prompt: false
450
+ detr:
451
+ base_HW: 64
452
+ debug_mode: 0
453
+ dec_layers: 1
454
+ dec_n_points: 4
455
+ decoder_type: conv
456
+ dim_down_factor: 1
457
+ dim_feedforward_factor: 4
458
+ dropout: 0.1
459
+ enc_layers: 6
460
+ enc_n_points: 4
461
+ features_num_channel:
462
+ - 128
463
+ - 32
464
+ grad_mode: absgrad
465
+ hidden_dim: 32
466
+ name: deformable_detr
467
+ num_feature_levels: 2
468
+ num_level: 3
469
+ num_queries: 256
470
+ only_last_level: false
471
+ position_embedding: sine
472
+ pre_norm: false
473
+ same_split_head: false
474
+ self_layer_num: 1
475
+ split_threshold: 1
476
+ split_threshold_range:
477
+ - 0
478
+ - 2
479
+ use_mean_features: true
480
+ head_mode: pcd
481
+ input_image_shape:
482
+ - 518
483
+ - 518
484
+ num_visual_prompt: 0
485
+ same_head: false
486
+ use_extrinsic_token: false
487
+ use_intrinsic_token: false
488
+ optimizer:
489
+ value:
490
+ accumulate: 1
491
+ backbone_lr_multiplier: 0.1
492
+ backbone_trainable: GB+H
493
+ lr: 0.0002
494
+ warm_up_steps: 125
495
+ seed:
496
+ value: 111123
497
+ test:
498
+ value:
499
+ align_pose: false
500
+ compute_scores: true
501
+ output_path: test
502
+ pose_align_steps: 100
503
+ pred_intrinsic: false
504
+ rot_opt_lr: 0.005
505
+ save_compare: false
506
+ save_image: false
507
+ save_video: false
508
+ split_threshold_list:
509
+ - 0
510
+ - 0.1
511
+ - 0.2
512
+ - 0.3
513
+ - 0.4
514
+ trans_opt_lr: 0.005
515
+ train:
516
+ value:
517
+ align_corners: false
518
+ aux_loss: 0
519
+ backbone_checkpoint: none
520
+ camera_consis_loss: 0
521
+ camera_loss: 1
522
+ debug_mode: 0
523
+ depth_mode: null
524
+ distill_max_steps: 1000000
525
+ distiller: ""
526
+ extended_visualization: false
527
+ extrinsic_drop: 0.1
528
+ intrinsic_drop: 0.1
529
+ intrinsic_scaling: true
530
+ one_sample_train: false
531
+ print_log_every_n_steps: 10
532
+ random_split_strategy: score
533
+ return_level_gs: true
534
+ scaling_detach: true
535
+ split_loss: 0.01
536
+ split_score_grad_scale: 10000
537
+ split_score_log_scale: true
538
+ target_scale_mode: 2
539
+ train_context: false
540
+ use_pred_point_focal: false
541
+ trainer:
542
+ value:
543
+ gradient_clip_val: 0.5
544
+ max_steps: 18751
545
+ num_nodes: 1
546
+ val_check_interval: 500
547
+ wandb:
548
+ value:
549
+ entity: scene-representation-group
550
+ mode: online
551
+ name: re10k_hpc
552
+ project: noposplat
553
+ tags:
554
+ - re10k
555
+ - 256x256
wandb/run-20250916_194552-aaciqyuf/files/media/images/activation_visualization_101_8a45d9dc83c5e5caae0d.png ADDED
wandb/run-20250916_194552-aaciqyuf/files/media/images/activation_visualization_126_948dc1a285a7709661c2.png ADDED
wandb/run-20250916_194552-aaciqyuf/files/media/images/activation_visualization_151_b777ce762fa8abbb2f16.png ADDED
wandb/run-20250916_194552-aaciqyuf/files/media/images/activation_visualization_176_9360fd68ec926bfb2b15.png ADDED
wandb/run-20250916_194552-aaciqyuf/files/media/images/activation_visualization_201_84928b0fa7964ea26925.png ADDED
wandb/run-20250916_194552-aaciqyuf/files/media/images/activation_visualization_226_9b48f7ee732928ccde69.png ADDED
wandb/run-20250916_194552-aaciqyuf/files/media/images/activation_visualization_251_cde359718086fb0623f7.png ADDED
wandb/run-20250916_194552-aaciqyuf/files/media/images/activation_visualization_26_c73506410e7d6e1d35fd.png ADDED
wandb/run-20250916_194552-aaciqyuf/files/media/images/activation_visualization_276_cc0deded106bd8c609c8.png ADDED
wandb/run-20250916_194552-aaciqyuf/files/media/images/activation_visualization_51_e5100843b701f67b204b.png ADDED
wandb/run-20250916_194552-aaciqyuf/files/media/images/activation_visualization_76_048e6651bc117aff8719.png ADDED
wandb/run-20250916_194552-aaciqyuf/files/output.log ADDED
The diff for this file is too large to render. See raw diff
 
wandb/run-20250916_194552-aaciqyuf/files/requirements.txt ADDED
@@ -0,0 +1,129 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ruff==0.11.4
2
+ PyYAML==6.0.2
3
+ sympy==1.13.3
4
+ fsspec==2024.6.1
5
+ imageio-ffmpeg==0.6.0
6
+ setuptools==78.1.1
7
+ timm==1.0.15
8
+ fonttools==4.57.0
9
+ nvidia-cublas-cu12==12.8.3.14
10
+ nvidia-cuda-nvrtc-cu12==12.8.61
11
+ dacite==1.9.2
12
+ hydra-core==1.3.2
13
+ certifi==2022.12.7
14
+ rich==14.0.0
15
+ filelock==3.13.1
16
+ idna==3.4
17
+ nvidia-cuda-cupti-cu12==12.8.57
18
+ pip==25.2
19
+ wandb==0.19.9
20
+ ninja==1.11.1.4
21
+ networkx==3.3
22
+ GitPython==3.1.44
23
+ moviepy==1.0.3
24
+ markdown-it-py==3.0.0
25
+ lightning==2.5.1
26
+ Jinja2==3.1.4
27
+ nvidia-cusparse-cu12==12.5.7.53
28
+ tabulate==0.9.0
29
+ tifffile==2025.3.30
30
+ nvidia-nvtx-cu12==12.8.55
31
+ jaxtyping==0.2.37
32
+ aiohappyeyeballs==2.6.1
33
+ MarkupSafe==2.1.5
34
+ pydantic_core==2.33.1
35
+ mypy-extensions==1.0.0
36
+ lazy_loader==0.4
37
+ pydantic==2.11.2
38
+ typing-inspection==0.4.0
39
+ pillow==10.4.0
40
+ nvidia-nvjitlink-cu12==12.8.61
41
+ omegaconf==2.3.0
42
+ scikit-image==0.25.2
43
+ jaraco.context==5.3.0
44
+ typing_extensions==4.12.2
45
+ mpmath==1.3.0
46
+ opencv-python==4.11.0.86
47
+ backports.tarfile==1.2.0
48
+ async-timeout==5.0.1
49
+ opt_einsum==3.4.0
50
+ nvidia-cuda-runtime-cu12==12.8.57
51
+ numpy==1.26.4
52
+ torch==2.7.1+cu128
53
+ psutil==7.0.0
54
+ pytorch-lightning==2.5.1
55
+ protobuf==5.29.4
56
+ imageio==2.37.0
57
+ colorama==0.4.6
58
+ attrs==25.3.0
59
+ importlib_metadata==8.0.0
60
+ frozenlist==1.5.0
61
+ beartype==0.19.0
62
+ propcache==0.3.1
63
+ typeguard==4.3.0
64
+ torchaudio==2.7.1+cu128
65
+ decorator==4.4.2
66
+ nvidia-cudnn-cu12==9.7.1.26
67
+ torchvision==0.22.1+cu128
68
+ sentry-sdk==2.25.1
69
+ mdurl==0.1.2
70
+ more-itertools==10.3.0
71
+ einops==0.8.1
72
+ zipp==3.19.2
73
+ antlr4-python3-runtime==4.9.3
74
+ cycler==0.12.1
75
+ nvidia-cufile-cu12==1.13.0.11
76
+ jaraco.collections==5.1.0
77
+ jaraco.text==3.12.1
78
+ torchmetrics==1.8.2
79
+ packaging==24.2
80
+ black==25.1.0
81
+ setproctitle==1.3.5
82
+ Pygments==2.19.2
83
+ tomli==2.2.1
84
+ six==1.17.0
85
+ docker-pycreds==0.4.0
86
+ inflect==7.3.1
87
+ requests==2.28.1
88
+ svg.py==1.6.0
89
+ python-dateutil==2.9.0.post0
90
+ proglog==0.1.11
91
+ gsplat==1.5.3
92
+ e3nn==0.5.6
93
+ pathspec==0.12.1
94
+ aiosignal==1.3.2
95
+ wheel==0.45.1
96
+ multidict==6.3.2
97
+ pyparsing==3.2.3
98
+ plyfile==1.1
99
+ jaraco.functools==4.0.1
100
+ nvidia-nccl-cu12==2.26.2
101
+ triton==3.3.1
102
+ nvidia-cusparselt-cu12==0.6.3
103
+ scikit-video==1.1.11
104
+ huggingface-hub==0.30.1
105
+ lightning-utilities==0.14.3
106
+ wadler_lindig==0.1.4
107
+ nvidia-curand-cu12==10.3.9.55
108
+ matplotlib==3.10.1
109
+ kiwisolver==1.4.8
110
+ tqdm==4.67.1
111
+ safetensors==0.5.3
112
+ scipy==1.15.2
113
+ lpips==0.1.4
114
+ nvidia-cusolver-cu12==11.7.2.55
115
+ click==8.1.8
116
+ charset-normalizer==2.1.1
117
+ platformdirs==4.3.7
118
+ opt-einsum-fx==0.1.4
119
+ annotated-types==0.7.0
120
+ smmap==5.0.2
121
+ gitdb==4.0.12
122
+ nvidia-cufft-cu12==11.3.3.41
123
+ contourpy==1.3.1
124
+ urllib3==1.26.13
125
+ autocommand==2.2.2
126
+ yarl==1.19.0
127
+ aiohttp==3.11.16
128
+ colorspacious==1.1.2
129
+ python-dotenv==1.1.0
wandb/run-20250916_194552-aaciqyuf/files/wandb-metadata.json ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.0-153-generic-x86_64-with-glibc2.35",
3
+ "python": "CPython 3.12.11",
4
+ "startedAt": "2025-09-16T11:45:52.656293Z",
5
+ "args": [
6
+ "+experiment=re10k",
7
+ "wandb.mode=online",
8
+ "wandb.name=re10k_hpc"
9
+ ],
10
+ "program": "-m src.main",
11
+ "git": {
12
+ "remote": "git@github.com:K-nowing/ICLR2026_HCP.git",
13
+ "commit": "6c0080eae0295d43dd7f18fd71e742d924238f89"
14
+ },
15
+ "email": "dna9041@korea.ac.kr",
16
+ "root": "/root/highspeedstorage/injae/code/ICLR2026_HCP/outputs/re10k_hpc",
17
+ "host": "notebook-72e474dd-92e1-11f0-a6ea-7ab79f1c53cb-0",
18
+ "executable": "/opt/conda/envs/ours/bin/python",
19
+ "cpu_count": 144,
20
+ "cpu_count_logical": 288,
21
+ "gpu": "NVIDIA B200",
22
+ "gpu_count": 8,
23
+ "disk": {
24
+ "/": {
25
+ "total": "1081101176832",
26
+ "used": "208657956864"
27
+ }
28
+ },
29
+ "memory": {
30
+ "total": "2434690371584"
31
+ },
32
+ "cpu": {
33
+ "count": 144,
34
+ "countLogical": 288
35
+ },
36
+ "gpu_nvidia": [
37
+ {
38
+ "name": "NVIDIA B200",
39
+ "memoryTotal": "192265846784",
40
+ "cudaCores": 18944
41
+ },
42
+ {
43
+ "name": "NVIDIA B200",
44
+ "memoryTotal": "192265846784",
45
+ "cudaCores": 18944
46
+ },
47
+ {
48
+ "name": "NVIDIA B200",
49
+ "memoryTotal": "192265846784",
50
+ "cudaCores": 18944
51
+ },
52
+ {
53
+ "name": "NVIDIA B200",
54
+ "memoryTotal": "192265846784",
55
+ "cudaCores": 18944
56
+ },
57
+ {
58
+ "name": "NVIDIA B200",
59
+ "memoryTotal": "192265846784",
60
+ "cudaCores": 18944
61
+ },
62
+ {
63
+ "name": "NVIDIA B200",
64
+ "memoryTotal": "192265846784",
65
+ "cudaCores": 18944
66
+ },
67
+ {
68
+ "name": "NVIDIA B200",
69
+ "memoryTotal": "192265846784",
70
+ "cudaCores": 18944
71
+ },
72
+ {
73
+ "name": "NVIDIA B200",
74
+ "memoryTotal": "192265846784",
75
+ "cudaCores": 18944
76
+ }
77
+ ],
78
+ "cudaVersion": "12.8"
79
+ }
wandb/run-20250916_194552-aaciqyuf/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lr-AdamW/pg2":2e-05,"loss/aux_0/mse":0.0034407840576022863,"attention_map":{"count":1,"filenames":["media/images/attention_map_277_0b32a0de1e86a80b9b35.png"],"captions":["651a7f83ed093001"],"_type":"images/separated","width":800,"height":536,"format":"png"},"val/gaussian_num_ratio":0.5851516723632812,"loss/total":0.03648084029555321,"lr-AdamW/pg2-momentum":0.9,"loss/aux_0/lpips":0.007770626340061426,"loss/aux_2/mse":0.0024059175048023462,"loss/aux_1/lpips":0.005286840721964836,"val/psnr":27.459697723388672,"_wandb":{"runtime":15987},"trainer/global_step":5599,"loss/aux_2/lpips":0.004525118973106146,"lr-AdamW/pg1-momentum":0.9,"loss/split_score":0.002340342616662383,"loss/3dgs_2_0/lpips":0.005342466291040182,"points_context_imgs":{"format":"png","count":1,"filenames":["media/images/points_context_imgs_275_a06eaa47ddd5bc5fb373.png"],"captions":["651a7f83ed093001"],"_type":"images/separated","width":536,"height":800},"activation_visualization":{"width":256,"height":256,"format":"png","count":1,"filenames":["media/images/activation_visualization_276_cc0deded106bd8c609c8.png"],"captions":["651a7f83ed093001"],"_type":"images/separated"},"_timestamp":1.7580390648492198e+09,"loss/aux_1/mse":0.0026043567340821028,"val/ssim":0.8060880303382874,"loss/camera":0.00019968389824498445,"epoch":0,"train/psnr_probabilistic":26.956830978393555,"val/lpips":0.12397965788841248,"info/global_step":5600,"comparison":{"captions":["651a7f83ed093001"],"_type":"images/separated","width":1064,"height":1098,"format":"png","count":1,"filenames":["media/images/comparison_274_cbe0488a19d085054063.png"]},"_step":282,"lr-AdamW/pg1":0.00016472919392187862,"_runtime":15912.194091688,"loss/3dgs_2_0/mse":0.0025647059082984924}
wandb/run-20250916_194552-aaciqyuf/logs/debug-core.log ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2025-09-16T19:45:52.307943975+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpg2d9js4m/port-217.txt","pid":217,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false}
2
+ {"time":"2025-09-16T19:45:52.319896951+08:00","level":"INFO","msg":"Will exit if parent process dies.","ppid":217}
3
+ {"time":"2025-09-16T19:45:52.326518926+08:00","level":"INFO","msg":"server is running","addr":{"IP":"127.0.0.1","Port":37777,"Zone":""}}
4
+ {"time":"2025-09-16T19:45:52.440206332+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"127.0.0.1:46444"}
5
+ {"time":"2025-09-16T19:45:52.665531521+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"aaciqyuf","id":"127.0.0.1:46444"}
6
+ {"time":"2025-09-16T19:45:54.002875103+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"aaciqyuf","id":"127.0.0.1:46444"}
7
+ {"time":"2025-09-17T00:12:20.364469246+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"127.0.0.1:46444"}
8
+ {"time":"2025-09-17T00:12:20.364629241+08:00","level":"INFO","msg":"connection: closing","id":"127.0.0.1:46444"}
9
+ {"time":"2025-09-17T00:12:20.364808859+08:00","level":"INFO","msg":"server is shutting down"}
10
+ {"time":"2025-09-17T00:12:20.364931191+08:00","level":"INFO","msg":"connection: closed successfully","id":"127.0.0.1:46444"}
11
+ {"time":"2025-09-17T00:12:20.884683539+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"127.0.0.1:46444"}
12
+ {"time":"2025-09-17T00:12:20.884731379+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"127.0.0.1:46444"}
13
+ {"time":"2025-09-17T00:12:20.884751901+08:00","level":"INFO","msg":"server is closed"}
wandb/run-20250916_194552-aaciqyuf/logs/debug-internal.log ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2025-09-16T19:45:52.670660357+08:00","level":"INFO","msg":"stream: starting","core version":"0.19.9","symlink path":"/root/highspeedstorage/injae/code/ICLR2026_HCP/outputs/re10k_hpc/wandb/run-20250916_194552-aaciqyuf/logs/debug-core.log"}
2
+ {"time":"2025-09-16T19:45:54.002750276+08:00","level":"INFO","msg":"created new stream","id":"aaciqyuf"}
3
+ {"time":"2025-09-16T19:45:54.002860648+08:00","level":"INFO","msg":"stream: started","id":"aaciqyuf"}
4
+ {"time":"2025-09-16T19:45:54.003038437+08:00","level":"INFO","msg":"writer: Do: started","stream_id":"aaciqyuf"}
5
+ {"time":"2025-09-16T19:45:54.003259991+08:00","level":"INFO","msg":"sender: started","stream_id":"aaciqyuf"}
6
+ {"time":"2025-09-16T19:45:54.003399136+08:00","level":"INFO","msg":"handler: started","stream_id":"aaciqyuf"}
7
+ {"time":"2025-09-16T19:45:54.411678515+08:00","level":"INFO","msg":"Starting system monitor"}
8
+ {"time":"2025-09-17T00:12:20.36477451+08:00","level":"INFO","msg":"stream: closing","id":"aaciqyuf"}
9
+ {"time":"2025-09-17T00:12:20.364855166+08:00","level":"INFO","msg":"Stopping system monitor"}
10
+ {"time":"2025-09-17T00:12:20.36503602+08:00","level":"INFO","msg":"Stopped system monitor"}
11
+ {"time":"2025-09-17T00:12:20.743361444+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
12
+ {"time":"2025-09-17T00:12:20.884041581+08:00","level":"INFO","msg":"handler: closed","stream_id":"aaciqyuf"}
13
+ {"time":"2025-09-17T00:12:20.884178301+08:00","level":"INFO","msg":"sender: closed","stream_id":"aaciqyuf"}
14
+ {"time":"2025-09-17T00:12:20.884170005+08:00","level":"INFO","msg":"writer: Close: closed","stream_id":"aaciqyuf"}
15
+ {"time":"2025-09-17T00:12:20.884479503+08:00","level":"INFO","msg":"stream: closed","id":"aaciqyuf"}
wandb/run-20250916_194552-aaciqyuf/logs/debug.log ADDED
@@ -0,0 +1,364 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-09-16 19:45:52,629 INFO MainThread:217 [wandb_setup.py:_flush():67] Current SDK version is 0.19.9
2
+ 2025-09-16 19:45:52,629 INFO MainThread:217 [wandb_setup.py:_flush():67] Configure stats pid to 217
3
+ 2025-09-16 19:45:52,629 INFO MainThread:217 [wandb_setup.py:_flush():67] Loading settings from /root/.config/wandb/settings
4
+ 2025-09-16 19:45:52,629 INFO MainThread:217 [wandb_setup.py:_flush():67] Loading settings from /root/highspeedstorage/injae/code/ICLR2026_HCP/wandb/settings
5
+ 2025-09-16 19:45:52,629 INFO MainThread:217 [wandb_setup.py:_flush():67] Loading settings from environment variables
6
+ 2025-09-16 19:45:52,630 INFO MainThread:217 [wandb_init.py:setup_run_log_directory():662] Logging user logs to /root/highspeedstorage/injae/code/ICLR2026_HCP/outputs/re10k_hpc/wandb/run-20250916_194552-aaciqyuf/logs/debug.log
7
+ 2025-09-16 19:45:52,630 INFO MainThread:217 [wandb_init.py:setup_run_log_directory():663] Logging internal logs to /root/highspeedstorage/injae/code/ICLR2026_HCP/outputs/re10k_hpc/wandb/run-20250916_194552-aaciqyuf/logs/debug-internal.log
8
+ 2025-09-16 19:45:52,630 INFO MainThread:217 [wandb_init.py:init():781] calling init triggers
9
+ 2025-09-16 19:45:52,631 INFO MainThread:217 [wandb_init.py:init():786] wandb.init called with sweep_config: {}
10
+ config: {'model': {'encoder': {'backbone': {'name': 'croco', 'model': 'ViTLarge_BaseDecoder', 'patch_embed_cls': 'PatchEmbedDust3R', 'asymmetry_decoder': True, 'intrinsics_embed_loc': 'encoder', 'intrinsics_embed_degree': 4, 'intrinsics_embed_type': 'token'}, 'name': 'noposplat', 'opacity_mapping': {'initial': 0.0, 'final': 0.0, 'warm_up': 1}, 'num_monocular_samples': 32, 'num_surfaces': 1, 'predict_opacity': False, 'gaussians_per_pixel': 1, 'gaussian_adapter': {'gaussian_scale_min': 0.5, 'gaussian_scale_max': 15.0, 'sh_degree': 4}, 'd_feature': 128, 'visualizer': {'num_samples': 8, 'min_resolution': 256, 'export_ply': False}, 'apply_bounds_shim': True, 'gs_params_head_type': 'dpt_gs', 'pose_free': True, 'pretrained_weights': ''}, 'decoder': {'name': 'splatting_cuda', 'background_color': [0.0, 0.0, 0.0], 'make_scale_invariant': False, 'use_gsplat': True}, 'vggt': {'input_image_shape': [518, 518], 'head_mode': 'pcd', 'same_head': False, 'camera_prompt': False, 'num_visual_prompt': 0, 'use_extrinsic_token': False, 'use_intrinsic_token': False, 'detr': {'name': 'deformable_detr', 'features_num_channel': [128, 32], 'hidden_dim': 32, 'dim_down_factor': 1, 'dim_feedforward_factor': 4, 'enc_layers': 6, 'dec_layers': 1, 'dropout': 0.1, 'num_feature_levels': 2, 'dec_n_points': 4, 'enc_n_points': 4, 'split_threshold': 1.0, 'num_queries': 256, 'base_HW': 64, 'position_embedding': 'sine', 'pre_norm': False, 'self_layer_num': 1, 'num_level': 3, 'debug_mode': 0, 'decoder_type': 'conv', 'split_threshold_range': [0.0, 2.0], 'only_last_level': False, 'grad_mode': 'absgrad', 'same_split_head': False, 'use_mean_features': True}}}, 'loss': {'mse': {'weight': 1.0}, 'lpips': {'weight': 0.05, 'apply_after_step': 0}}, 'wandb': {'project': 'noposplat', 'entity': 'scene-representation-group', 'name': 're10k_hpc', 'mode': 'online', 'tags': ['re10k', '256x256']}, 'mode': 'train', 'data_loader': {'train': {'num_workers': 16, 'persistent_workers': True, 'batch_size': 16, 'seed': 1234}, 'test': {'num_workers': 4, 'persistent_workers': False, 'batch_size': 1, 'seed': 2345}, 'val': {'num_workers': 1, 'persistent_workers': True, 'batch_size': 1, 'seed': 3456}}, 'optimizer': {'lr': 0.0002, 'warm_up_steps': 125, 'backbone_lr_multiplier': 0.1, 'backbone_trainable': 'GB+H', 'accumulate': 1}, 'checkpointing': {'load': None, 'every_n_train_steps': 9375, 'save_top_k': 1, 'save_weights_only': True}, 'train': {'depth_mode': None, 'extended_visualization': False, 'print_log_every_n_steps': 10, 'distiller': '', 'distill_max_steps': 1000000, 'debug_mode': 0, 'camera_consis_loss': 0.0, 'camera_loss': 1.0, 'aux_loss': 0.0, 'one_sample_train': False, 'target_scale_mode': 2, 'extrinsic_drop': 0.1, 'intrinsic_drop': 0.1, 'align_corners': False, 'backbone_checkpoint': 'none', 'intrinsic_scaling': True, 'return_level_gs': True, 'train_context': False, 'use_pred_point_focal': False, 'scaling_detach': True, 'split_score_log_scale': True, 'split_score_grad_scale': 10000.0, 'split_loss': 0.01, 'random_split_strategy': 'score'}, 'test': {'output_path': 'test', 'align_pose': False, 'pose_align_steps': 100, 'rot_opt_lr': 0.005, 'trans_opt_lr': 0.005, 'compute_scores': True, 'save_image': False, 'save_video': False, 'save_compare': False, 'split_threshold_list': [0.0, 0.1, 0.2, 0.3, 0.4], 'pred_intrinsic': False}, 'seed': 111123, 'trainer': {'max_steps': 18751, 'val_check_interval': 500, 'gradient_clip_val': 0.5, 'num_nodes': 1}, 'dataset': {'re10k': {'make_baseline_1': True, 'relative_pose': True, 'augment': True, 'background_color': [0.0, 0.0, 0.0], 'overfit_to_scene': None, 'skip_bad_shape': True, 'view_sampler': {'name': 'bounded', 'num_target_views': 4, 'num_context_views': 2, 'min_distance_between_context_views': 45, 'max_distance_between_context_views': 90, 'min_distance_to_context_views': 0, 'warm_up_steps': 9375, 'initial_min_distance_between_context_views': 25, 'initial_max_distance_between_context_views': 25, 'distill_steps': 0}, 'name': 're10k', 'roots': ['datasets/re10k'], 'input_image_shape': [256, 256], 'original_image_shape': [360, 640], 'cameras_are_circular': False, 'baseline_min': 0.001, 'baseline_max': 10000000000.0, 'max_fov': 100.0}}, '_wandb': {}}
11
+ 2025-09-16 19:45:52,631 INFO MainThread:217 [wandb_init.py:init():809] starting backend
12
+ 2025-09-16 19:45:52,631 INFO MainThread:217 [wandb_init.py:init():813] sending inform_init request
13
+ 2025-09-16 19:45:52,655 INFO MainThread:217 [backend.py:_multiprocessing_setup():101] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
14
+ 2025-09-16 19:45:52,655 INFO MainThread:217 [wandb_init.py:init():823] backend started and connected
15
+ 2025-09-16 19:45:52,668 INFO MainThread:217 [wandb_init.py:init():915] updated telemetry
16
+ 2025-09-16 19:45:52,704 INFO MainThread:217 [wandb_init.py:init():939] communicating run to backend with 90.0 second timeout
17
+ 2025-09-16 19:45:54,358 INFO MainThread:217 [wandb_init.py:init():1014] starting run threads in backend
18
+ 2025-09-16 19:45:54,813 INFO MainThread:217 [wandb_run.py:_console_start():2454] atexit reg
19
+ 2025-09-16 19:45:54,814 INFO MainThread:217 [wandb_run.py:_redirect():2306] redirect: wrap_raw
20
+ 2025-09-16 19:45:54,814 INFO MainThread:217 [wandb_run.py:_redirect():2371] Wrapping output streams.
21
+ 2025-09-16 19:45:54,814 INFO MainThread:217 [wandb_run.py:_redirect():2394] Redirects installed.
22
+ 2025-09-16 19:45:54,833 INFO MainThread:217 [wandb_init.py:init():1056] run started, returning control to user process
23
+ 2025-09-17 00:12:20,362 INFO MsgRouterThr:217 [mailbox.py:close():129] [no run ID] Closing mailbox, abandoning 1 handles.
24
+ 2025-09-17 00:12:20,841 ERROR MainThread:217 [redirect.py:_on_write():661] [no run ID] error in stderr callback
25
+ Traceback (most recent call last):
26
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/redirect.py", line 659, in _on_write
27
+ cb(written_data)
28
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 2381, in <lambda>
29
+ lambda data: self._console_raw_callback("stderr", data),
30
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
31
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 387, in wrapper
32
+ return func(self, *args, **kwargs)
33
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^
34
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 435, in wrapper_fn
35
+ return func(self, *args, **kwargs)
36
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^
37
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 1518, in _console_raw_callback
38
+ self._backend.interface.publish_output_raw(name, data)
39
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface.py", line 761, in publish_output_raw
40
+ self._publish_output_raw(o)
41
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface_shared.py", line 38, in _publish_output_raw
42
+ self._publish(rec)
43
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface_sock.py", line 39, in _publish
44
+ self._sock_client.send_record_publish(record)
45
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 174, in send_record_publish
46
+ self.send_server_request(server_req)
47
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 154, in send_server_request
48
+ self._send_message(msg)
49
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 151, in _send_message
50
+ self._sendall_with_error_handle(header + data)
51
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 130, in _sendall_with_error_handle
52
+ sent = self._sock.send(data)
53
+ ^^^^^^^^^^^^^^^^^^^^^
54
+ BrokenPipeError: [Errno 32] Broken pipe
55
+ 2025-09-17 00:12:20,882 ERROR MainThread:217 [redirect.py:_on_write():661] [no run ID] error in stderr callback
56
+ Traceback (most recent call last):
57
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/redirect.py", line 659, in _on_write
58
+ cb(written_data)
59
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 2381, in <lambda>
60
+ lambda data: self._console_raw_callback("stderr", data),
61
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
62
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 387, in wrapper
63
+ return func(self, *args, **kwargs)
64
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^
65
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 435, in wrapper_fn
66
+ return func(self, *args, **kwargs)
67
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^
68
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 1518, in _console_raw_callback
69
+ self._backend.interface.publish_output_raw(name, data)
70
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface.py", line 761, in publish_output_raw
71
+ self._publish_output_raw(o)
72
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface_shared.py", line 38, in _publish_output_raw
73
+ self._publish(rec)
74
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface_sock.py", line 39, in _publish
75
+ self._sock_client.send_record_publish(record)
76
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 174, in send_record_publish
77
+ self.send_server_request(server_req)
78
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 154, in send_server_request
79
+ self._send_message(msg)
80
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 151, in _send_message
81
+ self._sendall_with_error_handle(header + data)
82
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 130, in _sendall_with_error_handle
83
+ sent = self._sock.send(data)
84
+ ^^^^^^^^^^^^^^^^^^^^^
85
+ BrokenPipeError: [Errno 32] Broken pipe
86
+ 2025-09-17 00:12:20,882 ERROR MainThread:217 [redirect.py:_on_write():661] [no run ID] error in stderr callback
87
+ Traceback (most recent call last):
88
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/redirect.py", line 659, in _on_write
89
+ cb(written_data)
90
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 2381, in <lambda>
91
+ lambda data: self._console_raw_callback("stderr", data),
92
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
93
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 387, in wrapper
94
+ return func(self, *args, **kwargs)
95
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^
96
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 435, in wrapper_fn
97
+ return func(self, *args, **kwargs)
98
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^
99
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 1518, in _console_raw_callback
100
+ self._backend.interface.publish_output_raw(name, data)
101
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface.py", line 761, in publish_output_raw
102
+ self._publish_output_raw(o)
103
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface_shared.py", line 38, in _publish_output_raw
104
+ self._publish(rec)
105
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface_sock.py", line 39, in _publish
106
+ self._sock_client.send_record_publish(record)
107
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 174, in send_record_publish
108
+ self.send_server_request(server_req)
109
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 154, in send_server_request
110
+ self._send_message(msg)
111
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 151, in _send_message
112
+ self._sendall_with_error_handle(header + data)
113
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 130, in _sendall_with_error_handle
114
+ sent = self._sock.send(data)
115
+ ^^^^^^^^^^^^^^^^^^^^^
116
+ BrokenPipeError: [Errno 32] Broken pipe
117
+ 2025-09-17 00:12:20,882 ERROR MainThread:217 [redirect.py:_on_write():661] [no run ID] error in stderr callback
118
+ Traceback (most recent call last):
119
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/redirect.py", line 659, in _on_write
120
+ cb(written_data)
121
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 2381, in <lambda>
122
+ lambda data: self._console_raw_callback("stderr", data),
123
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
124
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 387, in wrapper
125
+ return func(self, *args, **kwargs)
126
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^
127
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 435, in wrapper_fn
128
+ return func(self, *args, **kwargs)
129
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^
130
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 1518, in _console_raw_callback
131
+ self._backend.interface.publish_output_raw(name, data)
132
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface.py", line 761, in publish_output_raw
133
+ self._publish_output_raw(o)
134
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface_shared.py", line 38, in _publish_output_raw
135
+ self._publish(rec)
136
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface_sock.py", line 39, in _publish
137
+ self._sock_client.send_record_publish(record)
138
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 174, in send_record_publish
139
+ self.send_server_request(server_req)
140
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 154, in send_server_request
141
+ self._send_message(msg)
142
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 151, in _send_message
143
+ self._sendall_with_error_handle(header + data)
144
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 130, in _sendall_with_error_handle
145
+ sent = self._sock.send(data)
146
+ ^^^^^^^^^^^^^^^^^^^^^
147
+ BrokenPipeError: [Errno 32] Broken pipe
148
+ 2025-09-17 00:12:20,883 ERROR MainThread:217 [redirect.py:_on_write():661] [no run ID] error in stderr callback
149
+ Traceback (most recent call last):
150
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/redirect.py", line 659, in _on_write
151
+ cb(written_data)
152
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 2381, in <lambda>
153
+ lambda data: self._console_raw_callback("stderr", data),
154
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
155
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 387, in wrapper
156
+ return func(self, *args, **kwargs)
157
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^
158
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 435, in wrapper_fn
159
+ return func(self, *args, **kwargs)
160
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^
161
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 1518, in _console_raw_callback
162
+ self._backend.interface.publish_output_raw(name, data)
163
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface.py", line 761, in publish_output_raw
164
+ self._publish_output_raw(o)
165
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface_shared.py", line 38, in _publish_output_raw
166
+ self._publish(rec)
167
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface_sock.py", line 39, in _publish
168
+ self._sock_client.send_record_publish(record)
169
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 174, in send_record_publish
170
+ self.send_server_request(server_req)
171
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 154, in send_server_request
172
+ self._send_message(msg)
173
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 151, in _send_message
174
+ self._sendall_with_error_handle(header + data)
175
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 130, in _sendall_with_error_handle
176
+ sent = self._sock.send(data)
177
+ ^^^^^^^^^^^^^^^^^^^^^
178
+ BrokenPipeError: [Errno 32] Broken pipe
179
+ 2025-09-17 00:12:20,886 ERROR MainThread:217 [redirect.py:_on_write():661] [no run ID] error in stderr callback
180
+ Traceback (most recent call last):
181
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/redirect.py", line 659, in _on_write
182
+ cb(written_data)
183
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 2381, in <lambda>
184
+ lambda data: self._console_raw_callback("stderr", data),
185
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
186
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 387, in wrapper
187
+ return func(self, *args, **kwargs)
188
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^
189
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 435, in wrapper_fn
190
+ return func(self, *args, **kwargs)
191
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^
192
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 1518, in _console_raw_callback
193
+ self._backend.interface.publish_output_raw(name, data)
194
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface.py", line 761, in publish_output_raw
195
+ self._publish_output_raw(o)
196
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface_shared.py", line 38, in _publish_output_raw
197
+ self._publish(rec)
198
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface_sock.py", line 39, in _publish
199
+ self._sock_client.send_record_publish(record)
200
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 174, in send_record_publish
201
+ self.send_server_request(server_req)
202
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 154, in send_server_request
203
+ self._send_message(msg)
204
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 151, in _send_message
205
+ self._sendall_with_error_handle(header + data)
206
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 130, in _sendall_with_error_handle
207
+ sent = self._sock.send(data)
208
+ ^^^^^^^^^^^^^^^^^^^^^
209
+ BrokenPipeError: [Errno 32] Broken pipe
210
+ 2025-09-17 00:12:20,887 ERROR MainThread:217 [redirect.py:_on_write():661] [no run ID] error in stderr callback
211
+ Traceback (most recent call last):
212
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/redirect.py", line 659, in _on_write
213
+ cb(written_data)
214
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 2381, in <lambda>
215
+ lambda data: self._console_raw_callback("stderr", data),
216
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
217
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 387, in wrapper
218
+ return func(self, *args, **kwargs)
219
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^
220
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 435, in wrapper_fn
221
+ return func(self, *args, **kwargs)
222
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^
223
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 1518, in _console_raw_callback
224
+ self._backend.interface.publish_output_raw(name, data)
225
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface.py", line 761, in publish_output_raw
226
+ self._publish_output_raw(o)
227
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface_shared.py", line 38, in _publish_output_raw
228
+ self._publish(rec)
229
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface_sock.py", line 39, in _publish
230
+ self._sock_client.send_record_publish(record)
231
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 174, in send_record_publish
232
+ self.send_server_request(server_req)
233
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 154, in send_server_request
234
+ self._send_message(msg)
235
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 151, in _send_message
236
+ self._sendall_with_error_handle(header + data)
237
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 130, in _sendall_with_error_handle
238
+ sent = self._sock.send(data)
239
+ ^^^^^^^^^^^^^^^^^^^^^
240
+ BrokenPipeError: [Errno 32] Broken pipe
241
+ 2025-09-17 00:12:20,887 ERROR MainThread:217 [redirect.py:_on_write():661] [no run ID] error in stderr callback
242
+ Traceback (most recent call last):
243
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/redirect.py", line 659, in _on_write
244
+ cb(written_data)
245
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 2381, in <lambda>
246
+ lambda data: self._console_raw_callback("stderr", data),
247
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
248
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 387, in wrapper
249
+ return func(self, *args, **kwargs)
250
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^
251
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 435, in wrapper_fn
252
+ return func(self, *args, **kwargs)
253
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^
254
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 1518, in _console_raw_callback
255
+ self._backend.interface.publish_output_raw(name, data)
256
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface.py", line 761, in publish_output_raw
257
+ self._publish_output_raw(o)
258
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface_shared.py", line 38, in _publish_output_raw
259
+ self._publish(rec)
260
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface_sock.py", line 39, in _publish
261
+ self._sock_client.send_record_publish(record)
262
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 174, in send_record_publish
263
+ self.send_server_request(server_req)
264
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 154, in send_server_request
265
+ self._send_message(msg)
266
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 151, in _send_message
267
+ self._sendall_with_error_handle(header + data)
268
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 130, in _sendall_with_error_handle
269
+ sent = self._sock.send(data)
270
+ ^^^^^^^^^^^^^^^^^^^^^
271
+ BrokenPipeError: [Errno 32] Broken pipe
272
+ 2025-09-17 00:12:20,887 ERROR MainThread:217 [redirect.py:_on_write():661] [no run ID] error in stderr callback
273
+ Traceback (most recent call last):
274
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/redirect.py", line 659, in _on_write
275
+ cb(written_data)
276
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 2381, in <lambda>
277
+ lambda data: self._console_raw_callback("stderr", data),
278
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
279
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 387, in wrapper
280
+ return func(self, *args, **kwargs)
281
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^
282
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 435, in wrapper_fn
283
+ return func(self, *args, **kwargs)
284
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^
285
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 1518, in _console_raw_callback
286
+ self._backend.interface.publish_output_raw(name, data)
287
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface.py", line 761, in publish_output_raw
288
+ self._publish_output_raw(o)
289
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface_shared.py", line 38, in _publish_output_raw
290
+ self._publish(rec)
291
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface_sock.py", line 39, in _publish
292
+ self._sock_client.send_record_publish(record)
293
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 174, in send_record_publish
294
+ self.send_server_request(server_req)
295
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 154, in send_server_request
296
+ self._send_message(msg)
297
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 151, in _send_message
298
+ self._sendall_with_error_handle(header + data)
299
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 130, in _sendall_with_error_handle
300
+ sent = self._sock.send(data)
301
+ ^^^^^^^^^^^^^^^^^^^^^
302
+ BrokenPipeError: [Errno 32] Broken pipe
303
+ 2025-09-17 00:12:20,888 ERROR MainThread:217 [redirect.py:_on_write():661] [no run ID] error in stderr callback
304
+ Traceback (most recent call last):
305
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/redirect.py", line 659, in _on_write
306
+ cb(written_data)
307
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 2381, in <lambda>
308
+ lambda data: self._console_raw_callback("stderr", data),
309
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
310
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 387, in wrapper
311
+ return func(self, *args, **kwargs)
312
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^
313
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 435, in wrapper_fn
314
+ return func(self, *args, **kwargs)
315
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^
316
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 1518, in _console_raw_callback
317
+ self._backend.interface.publish_output_raw(name, data)
318
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface.py", line 761, in publish_output_raw
319
+ self._publish_output_raw(o)
320
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface_shared.py", line 38, in _publish_output_raw
321
+ self._publish(rec)
322
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface_sock.py", line 39, in _publish
323
+ self._sock_client.send_record_publish(record)
324
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 174, in send_record_publish
325
+ self.send_server_request(server_req)
326
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 154, in send_server_request
327
+ self._send_message(msg)
328
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 151, in _send_message
329
+ self._sendall_with_error_handle(header + data)
330
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 130, in _sendall_with_error_handle
331
+ sent = self._sock.send(data)
332
+ ^^^^^^^^^^^^^^^^^^^^^
333
+ BrokenPipeError: [Errno 32] Broken pipe
334
+ 2025-09-17 00:12:21,715 ERROR child-process-observer:217 [redirect.py:_on_write():661] [no run ID] error in stderr callback
335
+ Traceback (most recent call last):
336
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/redirect.py", line 659, in _on_write
337
+ cb(written_data)
338
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 2381, in <lambda>
339
+ lambda data: self._console_raw_callback("stderr", data),
340
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
341
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 387, in wrapper
342
+ return func(self, *args, **kwargs)
343
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^
344
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 435, in wrapper_fn
345
+ return func(self, *args, **kwargs)
346
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^
347
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/wandb_run.py", line 1518, in _console_raw_callback
348
+ self._backend.interface.publish_output_raw(name, data)
349
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface.py", line 761, in publish_output_raw
350
+ self._publish_output_raw(o)
351
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface_shared.py", line 38, in _publish_output_raw
352
+ self._publish(rec)
353
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/interface/interface_sock.py", line 39, in _publish
354
+ self._sock_client.send_record_publish(record)
355
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 174, in send_record_publish
356
+ self.send_server_request(server_req)
357
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 154, in send_server_request
358
+ self._send_message(msg)
359
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 151, in _send_message
360
+ self._sendall_with_error_handle(header + data)
361
+ File "/opt/conda/envs/ours/lib/python3.12/site-packages/wandb/sdk/lib/sock_client.py", line 130, in _sendall_with_error_handle
362
+ sent = self._sock.send(data)
363
+ ^^^^^^^^^^^^^^^^^^^^^
364
+ BrokenPipeError: [Errno 32] Broken pipe