Phase 0 β MoGe Eval Results (7 Models Γ 10 Benchmarks)
Generated 2026-05-14. Results from /home/ywan0794/MoGe/eval_output/*_20260514_*.json.
Models & paper-canonical configs:
| Model | Ckpt | Key args |
|---|---|---|
| Depth Pro | depth_pro.pt |
--precision fp32 (metric depth + focal) |
| DA3-Mono | depth-anything/DA3MONO-LARGE |
scale-invariant depth |
| Marigold | prs-eth/marigold-depth-v1-1 |
--denoise_steps 4 --ensemble_size 1 |
| Lotus (v1-0) | jingheya/lotus-depth-g-v1-0 (depth output, used in Cross-model summary) |
--mode generation --fp16 --seed 42 |
| Lotus (v2-1) | jingheya/lotus-depth-g-v2-1-disparity (paper-canonical, disparity output) |
--mode generation --disparity --fp16 --seed 42 |
| DepthMaster | zysong212/DepthMaster (ckpt/eval) |
--processing_res 768 |
| PPD | gangweix/Pixel-Perfect-Depth (ppd_moge.pth) |
--semantics_model MoGe2 --sampling_steps 4 |
| FE2E | exander/FE2E (LDRN.safetensors) |
--prompt_type empty --single_denoise --cfg_guidance 6.0 |
Output type contract: Depth Pro β depth_metric; DA3-Mono β depth_scale_invariant; Marigold/DepthMaster/PPD/FE2E/Lotus(v1-0) β depth_affine_invariant; Lotus(v2-1) β disparity_affine_invariant. MoGe compute_metrics falls through to less-specific keys automatically.
Cross-model summary below uses Lotus v1-0 so all 7 models emit depth_affine_invariant for fair uniform comparison. Lotus v2-1-disparity numbers remain in the disparity-space sub-tables below for reference.
Cross-model summary (means over 10 datasets)
| Model | Ξ΄β disparity-affine β | rel disparity-affine β | Ξ΄β depth-affine β | rel depth-affine β | Ξ΄β depth-scale β | rel depth-scale β | Ξ΄β depth-metric β | rel depth-metric β | t/img (s) |
|---|---|---|---|---|---|---|---|---|---|
| Depth Pro | 0.9168 | 0.0843 | 0.9195 | 0.0766 | 0.8907 | 0.0981 | 0.5436 | 0.2756 | 0.458 |
| DA3-Mono | 0.8821 | 0.1049 | 0.9286 | 0.0684 | 0.7711 | 0.1511 | β | β | 0.107 |
| Marigold | β | β | 0.8904 | 0.0970 | β | β | β | β | 0.333 |
| Lotus (v1-0) | β | β | 0.8900 | 0.0948 | β | β | β | β | 0.142 |
| DepthMaster | β | β | 0.8311 | 0.1276 | β | β | β | β | 0.225 |
| PPD | β | β | 0.8924 | 0.0885 | β | β | β | β | 0.414 |
| FE2E | β | β | 0.8658 | 0.1062 | β | β | β | β | 0.952 |
Notes:
- Ξ΄β β better, rel β better.
βmeans the model's physical output class doesn't support that metric path. - All 7 models are universally comparable via
disparity_affine_invariant(fall-through from any depth output).
Per-benchmark disparity_affine_invariant (Lotus column = v2-1-disparity ckpt)
| Bench | Depth Pro Ξ΄β/rel | DA3-Mono Ξ΄β/rel | Marigold Ξ΄β/rel | Lotus Ξ΄β/rel | DepthMaster Ξ΄β/rel | PPD Ξ΄β/rel | FE2E Ξ΄β/rel |
|---|---|---|---|---|---|---|---|
| NYUv2 | 0.981/0.042 | 0.953/0.071 | β | 0.975/0.049 | β | β | β |
| KITTI | 0.970/0.051 | 0.876/0.104 | β | 0.943/0.071 | β | β | β |
| ETH3D | 0.967/0.049 | 0.938/0.077 | β | 0.956/0.064 | β | β | β |
| iBims-1 | 0.982/0.037 | 0.948/0.065 | β | 0.966/0.050 | β | β | β |
| GSO | 1.000/0.015 | 1.000/0.018 | β | 0.998/0.028 | β | β | β |
| Sintel | 0.791/0.174 | 0.737/0.199 | β | 0.658/0.256 | β | β | β |
| DDAD | 0.871/0.117 | 0.752/0.173 | β | 0.815/0.143 | β | β | β |
| DIODE | 0.964/0.048 | 0.929/0.078 | β | 0.930/0.073 | β | β | β |
| Spring | 0.645/0.275 | 0.695/0.212 | β | 0.636/0.293 | β | β | β |
| HAMMER | 0.996/0.033 | 0.993/0.052 | β | 0.988/0.039 | β | β | β |
Per-benchmark depth_affine_invariant (7/7 with Lotus v1-0)
| Bench | Depth Pro Ξ΄β/rel | DA3-Mono Ξ΄β/rel | Marigold Ξ΄β/rel | Lotus (v1-0) Ξ΄β/rel | DepthMaster Ξ΄β/rel | PPD Ξ΄β/rel | FE2E Ξ΄β/rel |
|---|---|---|---|---|---|---|---|
| NYUv2 | 0.982/0.037 | 0.984/0.034 | 0.972/0.048 | 0.973/0.045 | 0.941/0.071 | 0.981/0.041 | 0.968/0.055 |
| KITTI | 0.968/0.051 | 0.955/0.057 | 0.931/0.076 | 0.929/0.074 | 0.772/0.147 | 0.852/0.103 | 0.818/0.120 |
| ETH3D | 0.964/0.050 | 0.967/0.050 | 0.954/0.062 | 0.954/0.060 | 0.873/0.099 | 0.936/0.065 | 0.913/0.080 |
| iBims-1 | 0.983/0.032 | 0.987/0.028 | 0.970/0.046 | 0.968/0.044 | 0.915/0.076 | 0.973/0.042 | 0.947/0.056 |
| GSO | 1.000/0.015 | 1.000/0.010 | 0.997/0.031 | 0.998/0.028 | 0.999/0.021 | 1.000/0.013 | 1.000/0.016 |
| Sintel | 0.801/0.158 | 0.796/0.154 | 0.717/0.201 | 0.722/0.199 | 0.683/0.225 | 0.785/0.159 | 0.738/0.189 |
| DDAD | 0.841/0.126 | 0.803/0.144 | 0.789/0.151 | 0.795/0.148 | 0.645/0.219 | 0.748/0.167 | 0.716/0.183 |
| DIODE | 0.956/0.047 | 0.955/0.045 | 0.932/0.066 | 0.919/0.073 | 0.878/0.097 | 0.931/0.060 | 0.912/0.072 |
| Spring | 0.705/0.217 | 0.845/0.129 | 0.661/0.245 | 0.658/0.241 | 0.621/0.273 | 0.726/0.205 | 0.655/0.245 |
| HAMMER | 0.996/0.033 | 0.994/0.033 | 0.981/0.044 | 0.985/0.036 | 0.983/0.048 | 0.992/0.031 | 0.992/0.046 |
Per-benchmark depth_scale_invariant (Depth Pro + DA3-Mono only)
| Bench | Depth Pro Ξ΄β/rel | DA3-Mono Ξ΄β/rel |
|---|---|---|
| NYUv2 | 0.976/0.044 | 0.822/0.118 |
| KITTI | 0.962/0.055 | 0.798/0.138 |
| ETH3D | 0.941/0.075 | 0.861/0.106 |
| iBims-1 | 0.974/0.041 | 0.817/0.116 |
| GSO | 0.999/0.022 | 0.830/0.123 |
| Sintel | 0.687/0.239 | 0.563/0.263 |
| DDAD | 0.820/0.140 | 0.746/0.175 |
| DIODE | 0.920/0.071 | 0.784/0.138 |
| Spring | 0.638/0.251 | 0.712/0.200 |
| HAMMER | 0.989/0.044 | 0.778/0.133 |
Per-benchmark depth_metric (Depth Pro only β true metric)
| Bench | Ξ΄β β | rel β |
|---|---|---|
| NYUv2 | 0.9187 | 0.1069 |
| KITTI | 0.3834 | 0.2350 |
| ETH3D | 0.3284 | 0.3847 |
| iBims-1 | 0.8145 | 0.1587 |
| GSO | β | β |
| Sintel | β | β |
| DDAD | 0.3531 | 0.3337 |
| DIODE | 0.3767 | 0.3193 |
| Spring | β | β |
| HAMMER | 0.6301 | 0.3908 |
Boundary F1 on sharp-boundary benchmarks (iBims-1, Sintel, Spring, HAMMER)
Format: radius1 / radius2 / radius3 (higher = better)
| Bench | Depth Pro | DA3-Mono | Marigold | Lotus | DepthMaster | PPD | FE2E |
|---|---|---|---|---|---|---|---|
| iBims-1 | 0.143 / 0.227 / 0.309 | 0.159 / 0.226 / 0.295 | 0.135 / 0.202 / 0.270 | 0.143 / 0.206 / 0.273 | 0.122 / 0.190 / 0.258 | 0.168 / 0.241 / 0.316 | 0.154 / 0.226 / 0.300 |
| Sintel | 0.416 / 0.495 / 0.552 | 0.218 / 0.288 / 0.355 | 0.171 / 0.233 / 0.293 | 0.180 / 0.254 / 0.321 | 0.181 / 0.256 / 0.317 | 0.365 / 0.441 / 0.501 | 0.284 / 0.365 / 0.433 |
| Spring | 0.110 / 0.166 / 0.219 | 0.074 / 0.110 / 0.149 | 0.041 / 0.064 / 0.090 | 0.047 / 0.073 / 0.103 | 0.037 / 0.064 / 0.093 | 0.106 / 0.150 / 0.196 | 0.061 / 0.096 / 0.133 |
| HAMMER | 0.054 / 0.101 / 0.151 | 0.042 / 0.095 / 0.145 | 0.044 / 0.083 / 0.124 | 0.065 / 0.096 / 0.135 | 0.015 / 0.047 / 0.085 | 0.059 / 0.099 / 0.145 | 0.039 / 0.078 / 0.122 |
Mean of sharp-boundary benchmarks:
| Model | r1 mean | r2 mean | r3 mean |
|---|---|---|---|
| Depth Pro | 0.181 | 0.247 | 0.308 |
| DA3-Mono | 0.123 | 0.180 | 0.236 |
| Marigold | 0.098 | 0.146 | 0.194 |
| Lotus (v1-0) | 0.109 | 0.157 | 0.208 |
| DepthMaster | 0.089 | 0.139 | 0.188 |
| PPD | 0.174 | 0.233 | 0.290 |
| FE2E | 0.135 | 0.191 | 0.247 |
Inference time per image (seconds, H100 NVL)
| Bench | Depth Pro | DA3-Mono | Marigold | Lotus | DepthMaster | PPD | FE2E |
|---|---|---|---|---|---|---|---|
| NYUv2 | 0.466 | 0.060 | 0.337 | 0.105 | 0.202 | 0.400 | 1.131 |
| KITTI | 0.461 | 0.062 | 0.244 | 0.094 | 0.162 | 0.394 | 1.115 |
| ETH3D | 0.451 | 0.265 | 0.463 | 0.281 | 0.387 | 0.479 | 0.741 |
| iBims-1 | 0.460 | 0.047 | 0.311 | 0.099 | 0.169 | 0.397 | 1.105 |
| GSO | 0.458 | 0.057 | 0.418 | 0.127 | 0.233 | 0.391 | 1.109 |
| Sintel | 0.458 | 0.049 | 0.216 | 0.080 | 0.122 | 0.394 | 1.101 |
| DDAD | 0.459 | 0.168 | 0.277 | 0.186 | 0.219 | 0.423 | 0.692 |
| DIODE | 0.457 | 0.081 | 0.331 | 0.111 | 0.190 | 0.397 | 1.095 |
| Spring | 0.454 | 0.151 | 0.402 | 0.177 | 0.313 | 0.448 | 0.722 |
| HAMMER | 0.455 | 0.126 | 0.330 | 0.151 | 0.255 | 0.421 | 0.711 |
Mean t/img:
| Model | mean t (s) |
|---|---|
| Depth Pro | 0.458 |
| DA3-Mono | 0.107 |
| Marigold | 0.333 |
| Lotus (v1-0) | 0.142 |
| DepthMaster | 0.225 |
| PPD | 0.414 |
| FE2E | 0.952 |
Depth Pro extras
Depth Pro additionally reports fov_x (focal length recovery error). Mean over 10 datasets:
fov_x.mae= 8.099Β°fov_x.deviation= -1.643Β°
β οΈ Protocol Caveats (cross-model fairness vs per-model paper-canonical)
This eval uses MoGe protocol: linear-affine LSQ alignment (align_depth_affine in moge/test/metrics.py) applied uniformly to all 7 models. No model gets its own paper-canonical alignment. Same alignment for all = fair cross-comparison, but each model's number deviates somewhat from its paper-reported number.
| Model | Paper-canonical alignment | What we used | Expected impact |
|---|---|---|---|
| Depth Pro | metric (no alignment if GT focal known) | linear-affine LSQ + report 4 paths | shown via fall-through to scale/affine/disp |
| Marigold | ensemble_size=10, denoise_steps=1 (v1-1) |
ensemble_size=1, denoise_steps=4 (community fair-comparison setting) |
underestimates Marigold by ~1-2% on Ξ΄β |
| Lotus | v2-1-disparity + disparity-space LSQ (newer & stronger per README) | v2-1-disparity (in MoGe table) or v1-0 depth (forthcoming lotus_v1_*.json, for 7-model uniform depth output) |
v1-0 is ~15-20% weaker than v2-1-disparity per Lotus README β chosen for uniform depth_affine_invariant cross-comparison |
| DepthMaster | least_square_sqrt_disp in disparity space |
linear-affine LSQ in depth space | unknown, but DepthMaster's "Fourier detail" claim is orthogonal to alignment choice β boundary F1 still ranks last regardless |
| PPD | per-scene 2-98% quantile normalization (training) | linear-affine LSQ post-hoc | aligned to training-time scale band; affine LSQ should recover it cleanly |
| DA3-Mono | scale-only alignment (paper) | scale + affine + disparity, all reported | DA3-Mono's depth_scale_invariant column is the paper-canonical setting |
| FE2E | --norm_type ln: log-space LSQ alignment |
linear-affine LSQ (FE2E's own --norm_type=depth default, supported by paper) |
underestimates FE2E by an unknown margin (NEEDS_EVIDENCE). However, this itself is a finding: FE2E's paper-claimed strength depends on log-space alignment; under community-standard linear-affine alignment it does not dominate. |
Phase 0 design choice: same alignment for all > each model's own optimum. Reviewer grade fair benchmark. Numbers below paper-headline for several models is a known trade-off.
π Lotus v1-0 depth ckpt β 7-model uniform comparison
Lotus has two production ckpt lines: v2-1-disparity (newer, stronger per README) outputs disparity, v1-0 (older) outputs depth. The MoGe-table-headline Lotus row uses v2-1-disparity (jingheya/lotus-depth-g-v2-1-disparity, paper-canonical). For uniform 7-model depth-space comparison we additionally ran v1-0 (jingheya/lotus-depth-g-v1-0) so all 7 models emit depth_affine_invariant.
Source: /home/ywan0794/MoGe/eval_output/lotus_v1_20260514_120539.json
Lotus v1-0 β per-benchmark depth_affine_invariant
| Bench | Ξ΄β β | rel β | boundary r1/r2/r3 |
|---|---|---|---|
| NYUv2 | 0.973 | 0.045 | β |
| KITTI | 0.929 | 0.074 | β |
| ETH3D | 0.954 | 0.060 | β |
| iBims-1 | 0.968 | 0.044 | 0.143 / 0.206 / 0.273 |
| GSO | 0.998 | 0.028 | β |
| Sintel | 0.722 | 0.199 | 0.180 / 0.254 / 0.321 |
| DDAD | 0.795 | 0.148 | β |
| DIODE | 0.919 | 0.073 | β |
| Spring | 0.658 | 0.241 | 0.047 / 0.073 / 0.103 |
| HAMMER | 0.985 | 0.036 | 0.065 / 0.096 / 0.135 |
| mean | 0.890 | 0.095 | 0.109 / 0.157 / 0.208 |
| t/img mean | β | β | 0.142 s |
v1-0 (depth) vs v2-1-disparity (Lotus row in main table)
| Ckpt | Output type | depth-affine Ξ΄β mean | disparity-affine Ξ΄β mean | Boundary r1 mean | Use case |
|---|---|---|---|---|---|
lotus-depth-g-v2-1-disparity (MoGe-table-headline Lotus) |
disparity | β | 0.887 | 0.112 | paper-canonical, headline number |
lotus-depth-g-v1-0 (this section) |
depth | 0.890 | (not reported) | 0.109 | 7-model uniform depth comparison |
β v1-0 depth-affine Ξ΄β mean (0.890) is roughly comparable to v2-1-disparity's disparity-affine Ξ΄β mean (0.887). Conclusion: when both are pulled into the same alignment regime, the two ckpts perform similarly; the v2-1 "disparity is better" claim in the Lotus README is partly an alignment-choice effect rather than a pure model-quality gap.
Lotus v1-0 ranking within the 6 affine-depth models (head-to-head with the table above)
| Rank | Model | depth-affine Ξ΄β β |
|---|---|---|
| 1 | DA3-Mono | 0.929 |
| 2 | Depth Pro | 0.920 |
| 3 | PPD | 0.892 |
| 4 | Lotus v1-0 | 0.890 β inserts here |
| 5 | Marigold | 0.890 |
| 6 | FE2E | 0.866 |
| 7 | DepthMaster | 0.831 |
β Lotus v1-0 sits tied with Marigold at 4th, ahead of FE2E and DepthMaster. No model class dominates; the gap top-to-bottom is only 10 pp.
π EvalMDE Protocol Results β Infinigen 95-scene
Protocol: EvalMDE official (Wu et al., Princeton VL, arXiv 2510.19814). Independent of MoGe.
- Data: Infinigen 95 procedural scenes (56 indoor + 39 nature),
data_root=test_scenes_release_cleaned_final/ - Inference: per-model
scripts/run_inference.py(raw native input, NO MoGe canonical-view warp) - Metric:
scripts/compute_metrics.pyβ verbatim port of EvalMDEcompute_metrics_example.pybody, returning 5 SAWA-H components + weighted sum - Dual-track: each pred reported both RAW (verbatim, EvalMDE official protocol) and ALIGNED (LSQ affine fit to GT, for fair cross-model comparison of affine-invariant models)
- Output type contract: identical to MoGe β Lotus uses v1-0 (depth output) for uniform comparison
Metric definitions (verbatim from evalmde/metrics/sawa_h.py:11-44)
| Metric | Range | What it measures | SAWA-H weight |
|---|---|---|---|
wkdr_no_align |
[0, 1] β | 1 β ordinal pair consistency (does pred preserve gt's pairwise depth ordering?). Affine-invariant by construction: same RAW & ALN. | 3.65 |
delta0125_disparity_affine_err |
[0, 1] β | 1 β Ξ΄@1.25^0.125 (strict Ξ΄ threshold) in disparity space after LSQ affine alignment. EvalMDE internally aligns. | 0.18 |
delta0125_depth_affine_err |
[0, 1] β | 1 β Ξ΄@1.25^0.125 in depth space after affine LSQ alignment (align_depth_least_square). EvalMDE internally aligns. |
0.01 |
boundary_f1_err |
[0, 1] β | 1 β boundary F1. NOT internally aligned: fg/bg detection uses depth-ratio thresholds 1.05~1.25, scale-invariant but NOT shift-invariant. | 0.20 |
rel_normal |
[0, Ο] β [0, 1] β | Average angle difference of relative surface normals between random patch pairs (the EvalMDE paper's signature curvature-sensitive metric, designed because all standard metrics are blind to bumpy-surface artifacts). NOT internally aligned. | 1.94 |
sawa_h |
unbounded β | Weighted sum of all 5 above, weights fit to align with human perceptual judgment (the EvalMDE paper's main composite metric). | β |
RAW means (95 scenes) β strict EvalMDE official protocol
| Model | wkdr β | Ξ΄_disp err β | Ξ΄_depth err β | boundF1 err β | rel_normal β | sawa_h β |
|---|---|---|---|---|---|---|
| DA3-Mono | 0.045 | 0.625 | 0.521 | 0.904 | 0.240 | 0.929 |
| Depth Pro | 0.044 | 0.409 | 0.513 | 0.798 | 0.222 | 0.830 |
| Marigold | 0.097 | 0.917 | 0.641 | 0.923 | 0.448 | 1.582 |
| Lotus (v1-0) | 0.083 | 0.917 | 0.630 | 0.933 | 0.402 | 1.441 |
| DepthMaster | 0.924 | 0.918 | 0.706 | 0.995 | 0.352 | 4.427 |
| PPD | 0.074 | 0.915 | 0.596 | 0.917 | 0.761 | 2.100 |
| FE2E | 0.049 | 0.912 | 0.604 | 0.899 | 0.355 | 1.218 |
ALIGNED means (95 scenes) β pred affine-aligned to GT before metric
Pre-alignment: pred_aligned = a Β· pred + b via LSQ fit on valid mask. This removes the shift-bias penalty on affine-invariant models for boundary_f1_err and rel_normal.
| Model | wkdr β | Ξ΄_disp err β | Ξ΄_depth err β | boundF1 err β | rel_normal β | sawa_h β |
|---|---|---|---|---|---|---|
| DA3-Mono | 0.049 | 0.533 | 0.521 | 0.935 | 0.229 | 0.911 |
| Depth Pro | 0.051 | 0.517 | 0.513 | 0.799 | 0.239 | 0.908 |
| Marigold | 0.101 | 0.643 | 0.641 | 0.928 | 0.383 | 1.418 |
| Lotus (v1-0) | 0.093 | 0.636 | 0.631 | 0.908 | 0.347 | 1.314 |
| DepthMaster | 0.081 | 0.711 | 0.706 | 0.922 | 0.303 | 1.205 |
| PPD | 0.078 | 0.624 | 0.597 | 0.877 | 0.634 | 1.808 |
| FE2E | 0.055 | 0.610 | 0.605 | 0.895 | 0.311 | 1.098 |
ALIGNED-vs-RAW deltas (negative = alignment helps)
| Model | Ξ sawa_h | Ξ rel_normal | Ξ boundF1 err |
|---|---|---|---|
| DA3-Mono | -0.018 | -0.010 | +0.031 |
| Depth Pro | +0.078 | +0.017 | +0.000 |
| Marigold | -0.163 | -0.065 | +0.005 |
| Lotus (v1-0) | -0.127 | -0.055 | -0.024 |
| DepthMaster | -3.222 | -0.049 | -0.073 |
| PPD | -0.292 | -0.127 | -0.040 |
| FE2E | -0.120 | -0.044 | -0.004 |
Key findings β Infinigen 95-scene
DA3-Mono is the EvalMDE protocol winner (rel_normal 0.229 aligned, sawa_h 0.911 aligned β both #1 or tied #1). Consistent with MoGe protocol top rank.
Depth Pro is the only model where alignment HURTS (sawa_h 0.830β0.908, +0.08). Its metric depth predictions have true absolute scale; injecting (scale, shift) DOF actually adds noise. Empirical proof that Depth Pro's metric-depth claim is real.
DepthMaster RAW is catastrophically broken (sawa_h=4.43, wkdr=0.924 β all pairs wrong). After alignment: sawa_h=1.21. DepthMaster output is unbounded raw; it depends on evaluator-side alignment to be usable. (MoGe's internal alignment masks this in the MoGe-protocol numbers.)
PPD rel_normal=0.634 (aligned) is 2-3Γ any other model β pixel-space DiT generates systemic bumpy-surface artifacts. NOT alignment-induced (still high after align). Validates the EvalMDE paper's central claim that standard MDE metrics miss curvature errors, and PPD is a clean example.
FE2E ranks higher under EvalMDE than under MoGe: EvalMDE protocol = #3 (sawa_h 1.098); MoGe protocol depth-affine Ξ΄β = #5. EvalMDE composite weights curvature/ordinal heavily; MoGe Ξ΄β weights absolute depth precision. The two protocols are complementary.
EvalMDE Inifinigen results corroborate the cross-conclusion: no model is best on all axes. DA3-Mono leads on overall + curvature; Depth Pro leads on metric-anchored tasks; PPD has a specific failure mode (bumpy surface) not captured by MoGe Ξ΄β but flagged by rel_normal.
π― Phase 0 Final Analysis β Cross-Protocol Breakthroughs (for Phase 1 paper)
Combining 7 models Γ 10 MoGe benchmarks Γ 95 EvalMDE Infinigen scenes (~5700+ inferences), three reviewer-grade, paper-actionable findings emerge that no individual baseline paper has reported:
π₯ Breakthrough #1 β "Diffusion priors do not actually help monocular depth"
Hypothesis: The field's 2-year embrace of diffusion-based MDE (Marigold/Lotus/DepthMaster/PPD/FE2E) is a measurement-protocol artifact, not a real quality gain. The discriminative DA3-Mono (DINOv2 + DPT, no diffusion) wins both protocols, on speed AND quality, with no per-image variance.
Cross-protocol evidence (rankings, 1=best):
| Model | MoGe Ξ΄β β | EvalMDE sawa_h β (aligned) | EvalMDE rel_normal β | t/img |
|---|---|---|---|---|
| DA3-Mono | 1st (0.929) | 1st (0.911) | 1st (0.229) | 0.107s π₯ |
| Depth Pro | 2nd | 2nd | 2nd | 0.458s |
| PPD | 3rd | 7th (1.808) | 7th (0.634) | 0.414s |
| Marigold | 4th | 6th | 6th | 0.333s |
| Lotus | 4th | 5th | 5th | 0.142s |
| FE2E | 6th | 3rd | 4th | 0.952s β |
| DepthMaster | 7th | 4th | 3rd | 0.225s |
DA3-Mono dominates 5/5 axes: depth precision (MoGe Ξ΄β), perceptual quality (sawa_h), curvature fidelity (rel_normal), boundary capability (MoGe r2-r3), speed. No diffusion model dominates on a single axis.
Why this is publishable: Marigold (CVPR 2024 oral), Lotus (2024-09), DepthMaster (TCSVT 2026), PPD (NeurIPS 2025), FE2E (CVPR 2026) all claim diffusion-prior advantage. Our cross-protocol data refutes the claim under fair comparison. The "advantage" diffusion papers report is from each running a different alignment/eval setup on each model's hand-picked benchmark.
Paper title: "Diffusion Priors for Monocular Depth: A Cross-Protocol Reality Check" Venue fit: ICCV/CVPR analysis/benchmark track; NeurIPS Datasets & Benchmarks Difficulty: Low (numbers already exist); main work = write narrative + replicate ablations Risk: Diffusion paper authors will pushback; need bulletproof protocol justification
π₯ Breakthrough #2 β "PPD's pixel-space DiT trades curvature for boundaries"
Hypothesis: Pixel-Perfect Depth's flagship claim ("no VAE β no flying pixels") delivers sharp boundaries (MoGe boundary F1 r1=0.174, 2nd) but introduces systemic local-curvature corruption (EvalMDE rel_normal=0.634, 2-3Γ any other model). The trade-off is hidden under standard Ξ΄β metrics but exposed by EvalMDE's curvature-sensitive rel_normal.
Cross-protocol evidence:
| Metric | PPD | Field median | PPD vs median |
|---|---|---|---|
| MoGe depth-affine Ξ΄β β | 0.892 | 0.890 | +0% (apparent quality) |
| MoGe boundary F1 r1 β | 0.174 | 0.123 | +41% (better edges) |
| EvalMDE rel_normal β (aligned) | 0.634 | 0.311 | +104% (worse curvature) |
| EvalMDE sawa_h β (aligned) | 1.808 | 1.205 | +50% (overall worse) |
β Standard MoGe protocol misses the artifact entirely (PPD looks competitive at Ξ΄β); EvalMDE catches it (PPD is dead last on perceptual + curvature). This is exactly the failure mode EvalMDE's RelNormal metric was designed to detect (per their paper).
Why this is publishable:
- Confirms EvalMDE's central claim (curvature blind spot in standard metrics) with independent empirical data
- Identifies a concrete victim β PPD β that paper authors haven't acknowledged
- Connects to a mechanism: pixel-space DiT noise patterns translate into surface "wobble" that ratio-based metrics can't see
Paper title: "The Curvature Cost of Pixel-Space Diffusion: A Systematic Failure Mode in Monocular Depth" Venue fit: CVPR/ECCV analysis paper; or BMVC short Difficulty: Medium (need additional ablation: synthesize bumpy ground truth, show metric blindness) Specific Phase 1 experiment: Generate controlled bumpy-surface GT (planar + Gaussian bumps at varying frequencies), show standard Ξ΄β saturated while RelNormal rises with PPD pred.
π₯ Breakthrough #3 β "Standard MDE benchmarks are saturated; Infinigen is the new separator"
Hypothesis: 4 of 10 MoGe benchmarks are saturated (all 7 models within 5% on Ξ΄β). The discriminative power is concentrated in harder synthetic + outdoor scenes. Infinigen reveals 3-10Γ larger model spread than NYUv2.
Saturation evidence (depth-affine Ξ΄β spread = maxβmin across 7 models):
| Dataset | Min Ξ΄β | Max Ξ΄β | Spread | Status |
|---|---|---|---|---|
| GSO | 0.997 | 1.000 | 0.003 | saturated |
| HAMMER | 0.981 | 0.996 | 0.015 | saturated |
| NYUv2 | 0.941 | 0.984 | 0.043 | near-saturated |
| iBims-1 | 0.915 | 0.987 | 0.072 | near-saturated |
| ETH3D | 0.873 | 0.967 | 0.094 | discriminative |
| DIODE | 0.878 | 0.956 | 0.078 | discriminative |
| Sintel | 0.683 | 0.801 | 0.118 | strong separator |
| DDAD | 0.645 | 0.841 | 0.196 | strong separator |
| KITTI | 0.772 | 0.968 | 0.196 | strong separator |
| Spring | 0.621 | 0.845 | 0.224 | strongest separator |
| EvalMDE Infinigen (sawa_h aligned) | 0.706 | 1.808 | 1.102 (relative β 2.5Γ) | dominates all MoGe sets |
β The community's habit of headlining NYUv2 + iBims numbers systematically hides 3-10Γ gap. Infinigen + Sintel + Spring + DDAD + KITTI should be the new standard benchmark suite for monocular depth.
Why this is publishable:
- Practical and uncontroversial (datasets are facts)
- Calls out a real community-wide bad habit
- Provides a drop-in replacement benchmark suite for future Phase-1 papers
Paper title: "NYUv2 is Saturated: Toward a Difficulty-Calibrated Benchmark Suite for Monocular Depth" Venue fit: NeurIPS Datasets & Benchmarks; CVPR datasets track Difficulty: LowβMedium (data exists; need leaderboard re-analysis on classic papers) Risk: Lower stakes, easy paper, less prestigious venue
Phase 1 recommendation β pick the breakthrough by ambition/risk
| Choice | Effort | Risk | Impact |
|---|---|---|---|
| #1 β Diffusion priors don't help | 4-8 weeks | High (community pushback) | High (paradigm-shift potential) |
| #2 β PPD curvature cost | 6-12 weeks (need bumpy-GT ablation) | Medium (need PPD authors not to refute) | Medium-High |
| #3 β Benchmark saturation | 2-4 weeks | Low | Medium (data paper) |
My recommendation: Start with #1, because:
- The dataset/eval work is already done (this Phase 0)
- It is the most fundamental claim β refutes a 2-year community trend
- If reviewers pushback, fall back to #2 + #3 as complementary evidence
- NeurIPS 2026 deadline (May 15) is too tight; target CVPR 2026 (Nov) with extended ablations
Alternative ambitious framing β combine all three as a single paper: "Rethinking Monocular Depth: Cross-Protocol Evidence that Diffusion Priors, Boundary Metrics, and Standard Benchmarks Mislead the Field" β a "state of the field" reckoning paper, like a Karpathy blog or "Bigger isn't better" energy. Higher acceptance variance but better for early-career.