MoGe / baselines /rae_depth.py
zeyuren2002's picture
Add files using upload-large-folder tool
da3ed5b verified
import os
import sys
from typing import *
import math
import click
import torch
import torch.nn.functional as F
from moge.test.baseline import MGEBaselineInterface
class Baseline(MGEBaselineInterface):
def __init__(self, repo_path: str, config_path: str, checkpoint_path: str,
image_size: int, num_steps: int, use_fp16: bool, device: str = 'cuda:0'):
super().__init__()
repo_path = os.path.abspath(repo_path)
src_path = os.path.join(repo_path, 'src')
if src_path not in sys.path:
sys.path.insert(0, src_path)
from omegaconf import OmegaConf
from stage2.transport import create_transport, Sampler
from utils.model_utils import instantiate_from_config
# Load config
full_cfg = OmegaConf.load(config_path)
rae_config = full_cfg.get("stage_1", None)
model_config = full_cfg.get("stage_2", None)
transport_config = full_cfg.get("transport", {})
sampler_config = full_cfg.get("sampler", {})
misc_config = full_cfg.get("misc", {})
transport_cfg = OmegaConf.to_container(transport_config, resolve=True) if transport_config else {}
sampler_cfg = OmegaConf.to_container(sampler_config, resolve=True) if sampler_config else {}
misc = OmegaConf.to_container(misc_config, resolve=True) if misc_config else {}
latent_size = tuple(int(dim) for dim in misc.get("latent_size", (768, 32, 32)))
shift_dim = misc.get("time_dist_shift_dim", math.prod(latent_size))
shift_base = misc.get("time_dist_shift_base", 4096)
time_dist_shift = math.sqrt(shift_dim / shift_base)
# Load RAE (DepthRAE)
rae = instantiate_from_config(rae_config).to(device)
rae.eval()
# Load Stage-2 model
model = instantiate_from_config(model_config).to(device)
# Load checkpoint
checkpoint = torch.load(checkpoint_path, map_location='cpu', weights_only=False)
if 'ema' in checkpoint:
state_dict = checkpoint['ema']
elif 'model' in checkpoint:
state_dict = checkpoint['model']
else:
state_dict = checkpoint
model.load_state_dict(state_dict)
model.eval()
# Create transport sampler
transport_params = dict(transport_cfg.get("params", {}))
transport_params.pop("time_dist_shift", None)
transport = create_transport(**transport_params, time_dist_shift=time_dist_shift)
transport_sampler = Sampler(transport)
sampler_mode = sampler_cfg.get("mode", "ODE").upper()
sampler_params = dict(sampler_cfg.get("params", {}))
sampler_params['num_steps'] = num_steps
if sampler_mode == "ODE":
eval_sampler = transport_sampler.sample_ode(**sampler_params)
else:
eval_sampler = transport_sampler.sample_sde(**sampler_params)
self.rae = rae
self.model = model
self.eval_sampler = eval_sampler
self.latent_size = latent_size
self.image_size = image_size
self.device = torch.device(device)
self.use_fp16 = use_fp16
@click.command()
@click.option('--repo', 'repo_path', type=str, default='/home/ywan0794/RAE')
@click.option('--rae_config', 'config_path', type=str, required=True)
@click.option('--checkpoint', 'checkpoint_path', type=str, required=True)
@click.option('--image_size', type=int, default=512)
@click.option('--num_steps', type=int, default=2)
@click.option('--fp16', 'use_fp16', is_flag=True)
@click.option('--device', type=str, default='cuda:0')
@staticmethod
def load(repo_path, config_path, checkpoint_path, image_size, num_steps, use_fp16, device):
return Baseline(repo_path, config_path, checkpoint_path, image_size, num_steps, use_fp16, device)
def _predict_depth(self, image: torch.FloatTensor):
original_height, original_width = image.shape[-2:]
if image.ndim == 3:
image = image.unsqueeze(0)
omit_batch_dim = True
else:
omit_batch_dim = False
b = image.shape[0]
# Resize to model input size
image_resized = F.interpolate(
image, size=(self.image_size, self.image_size),
mode='bilinear', align_corners=False, antialias=True,
)
# Encode RGB
z_rgb = self.rae.encode(image_resized)
# Sample depth from noise
z_noise = torch.randn(b, *self.latent_size, device=self.device)
y = torch.zeros(b, dtype=torch.long, device=self.device)
# Marigold-style: concat z_rgb with xt before passing to model
def model_fn(xt, t, y):
x_input = torch.cat([xt, z_rgb], dim=1)
return self.model(x_input, t, y)
# Run diffusion sampling
z_pred = self.eval_sampler(z_noise, model_fn, y=y)[-1]
# Decode to depth (pass z_rgb for conditioning)
depth_pred = self.rae.decode(z_pred.float(), z_rgb)
depth_pred = depth_pred.mean(dim=1) # (B, H, W)
# Resize back to original size
depth_pred = F.interpolate(
depth_pred.unsqueeze(1), size=(original_height, original_width),
mode='bilinear', align_corners=False,
)[:, 0]
if omit_batch_dim:
depth_pred = depth_pred.squeeze(0)
return depth_pred
@torch.inference_mode()
def infer(self, image: torch.FloatTensor, intrinsics: Optional[torch.FloatTensor] = None):
depth_pred = self._predict_depth(image)
return {
'depth_affine_invariant': depth_pred,
}
@torch.inference_mode()
def infer_for_evaluation(self, image: torch.FloatTensor, intrinsics: torch.FloatTensor = None):
with torch.cuda.amp.autocast(enabled=self.use_fp16, dtype=torch.float16):
depth_pred = self._predict_depth(image)
return {
'depth_affine_invariant': depth_pred,
}