config.yaml: CIFAR sinkhorn batch 128→32 for T4, pool batches 2500→10000 to compensate
Browse files- config.yaml +8 -3
config.yaml
CHANGED
|
@@ -143,16 +143,21 @@ experiment_cifar10:
|
|
| 143 |
use_scale_shift_norm: true
|
| 144 |
|
| 145 |
# Sinkhorn gradient flow (Phase 1)
|
|
|
|
|
|
|
|
|
|
|
|
|
| 146 |
sinkhorn:
|
| 147 |
blur: 1.0
|
| 148 |
scaling: 0.85
|
| 149 |
eta: 1.0
|
| 150 |
num_steps: 5
|
| 151 |
-
batch_size:
|
| 152 |
|
| 153 |
-
# Trajectory pool
|
|
|
|
| 154 |
pool:
|
| 155 |
-
num_batches:
|
| 156 |
storage_limit_gb: 45
|
| 157 |
|
| 158 |
# Velocity field matching training (NSGF model)
|
|
|
|
| 143 |
use_scale_shift_norm: true
|
| 144 |
|
| 145 |
# Sinkhorn gradient flow (Phase 1)
|
| 146 |
+
# NOTE: batch_size reduced from paper's 128 to 32 for T4 16GB VRAM.
|
| 147 |
+
# Sinkhorn on 3072-dim flattened vectors (3x32x32) with tensorized backend
|
| 148 |
+
# uses O(N^2 * D) memory. 128 samples OOMs on T4; 32 fits comfortably.
|
| 149 |
+
# Compensate by increasing pool batches (32 * 10000 = 320K ≈ 128 * 2500).
|
| 150 |
sinkhorn:
|
| 151 |
blur: 1.0
|
| 152 |
scaling: 0.85
|
| 153 |
eta: 1.0
|
| 154 |
num_steps: 5
|
| 155 |
+
batch_size: 32
|
| 156 |
|
| 157 |
+
# Trajectory pool — adjusted for smaller Sinkhorn batch
|
| 158 |
+
# 32 batch * 10000 batches * 5 steps = 1.6M entries (same order as paper)
|
| 159 |
pool:
|
| 160 |
+
num_batches: 10000
|
| 161 |
storage_limit_gb: 45
|
| 162 |
|
| 163 |
# Velocity field matching training (NSGF model)
|