rogermt commited on
Commit
80b1d4b
·
verified ·
1 Parent(s): a365009

config.yaml: CIFAR sinkhorn batch 128→32 for T4, pool batches 2500→10000 to compensate

Browse files
Files changed (1) hide show
  1. config.yaml +8 -3
config.yaml CHANGED
@@ -143,16 +143,21 @@ experiment_cifar10:
143
  use_scale_shift_norm: true
144
 
145
  # Sinkhorn gradient flow (Phase 1)
 
 
 
 
146
  sinkhorn:
147
  blur: 1.0
148
  scaling: 0.85
149
  eta: 1.0
150
  num_steps: 5
151
- batch_size: 128
152
 
153
- # Trajectory pool (Appendix E.2: 128 batch * 2500 batches * 5 steps ~ 45GB)
 
154
  pool:
155
- num_batches: 2500
156
  storage_limit_gb: 45
157
 
158
  # Velocity field matching training (NSGF model)
 
143
  use_scale_shift_norm: true
144
 
145
  # Sinkhorn gradient flow (Phase 1)
146
+ # NOTE: batch_size reduced from paper's 128 to 32 for T4 16GB VRAM.
147
+ # Sinkhorn on 3072-dim flattened vectors (3x32x32) with tensorized backend
148
+ # uses O(N^2 * D) memory. 128 samples OOMs on T4; 32 fits comfortably.
149
+ # Compensate by increasing pool batches (32 * 10000 = 320K ≈ 128 * 2500).
150
  sinkhorn:
151
  blur: 1.0
152
  scaling: 0.85
153
  eta: 1.0
154
  num_steps: 5
155
+ batch_size: 32
156
 
157
+ # Trajectory pool adjusted for smaller Sinkhorn batch
158
+ # 32 batch * 10000 batches * 5 steps = 1.6M entries (same order as paper)
159
  pool:
160
+ num_batches: 10000
161
  storage_limit_gb: 45
162
 
163
  # Velocity field matching training (NSGF model)