-
Compose and Conquer: Diffusion-Based 3D Depth Aware Composable Image Synthesis
Paper • 2401.09048 • Published • 10 -
Improving fine-grained understanding in image-text pre-training
Paper • 2401.09865 • Published • 18 -
Depth Anything: Unleashing the Power of Large-Scale Unlabeled Data
Paper • 2401.10891 • Published • 62 -
Scaling Up to Excellence: Practicing Model Scaling for Photo-Realistic Image Restoration In the Wild
Paper • 2401.13627 • Published • 78
Collections
Discover the best community collections!
Collections including paper arxiv:2511.10555
-
Arbitrary-steps Image Super-resolution via Diffusion Inversion
Paper • 2412.09013 • Published • 13 -
Deep Researcher with Test-Time Diffusion
Paper • 2507.16075 • Published • 68 -
nablaNABLA: Neighborhood Adaptive Block-Level Attention
Paper • 2507.13546 • Published • 126 -
Yume: An Interactive World Generation Model
Paper • 2507.17744 • Published • 92
-
EVA-CLIP-18B: Scaling CLIP to 18 Billion Parameters
Paper • 2402.04252 • Published • 30 -
Vision Superalignment: Weak-to-Strong Generalization for Vision Foundation Models
Paper • 2402.03749 • Published • 15 -
ScreenAI: A Vision-Language Model for UI and Infographics Understanding
Paper • 2402.04615 • Published • 44 -
EfficientViT-SAM: Accelerated Segment Anything Model Without Performance Loss
Paper • 2402.05008 • Published • 23
-
Boosting Latent Diffusion Models via Disentangled Representation Alignment
Paper • 2601.05823 • Published • 17 -
VQRAE: Representation Quantization Autoencoders for Multimodal Understanding, Generation and Reconstruction
Paper • 2511.23386 • Published • 16 -
Visual Generation Tuning
Paper • 2511.23469 • Published • 16 -
A Style is Worth One Code: Unlocking Code-to-Style Image Generation with Discrete Style Space
Paper • 2511.10555 • Published • 63
-
LoRACLR: Contrastive Adaptation for Customization of Diffusion Models
Paper • 2412.09622 • Published • 8 -
AnyDressing: Customizable Multi-Garment Virtual Dressing via Latent Diffusion Models
Paper • 2412.04146 • Published • 23 -
Learning Flow Fields in Attention for Controllable Person Image Generation
Paper • 2412.08486 • Published • 36 -
LoRA.rar: Learning to Merge LoRAs via Hypernetworks for Subject-Style Conditioned Image Generation
Paper • 2412.05148 • Published • 12
-
Compose and Conquer: Diffusion-Based 3D Depth Aware Composable Image Synthesis
Paper • 2401.09048 • Published • 10 -
Improving fine-grained understanding in image-text pre-training
Paper • 2401.09865 • Published • 18 -
Depth Anything: Unleashing the Power of Large-Scale Unlabeled Data
Paper • 2401.10891 • Published • 62 -
Scaling Up to Excellence: Practicing Model Scaling for Photo-Realistic Image Restoration In the Wild
Paper • 2401.13627 • Published • 78
-
Boosting Latent Diffusion Models via Disentangled Representation Alignment
Paper • 2601.05823 • Published • 17 -
VQRAE: Representation Quantization Autoencoders for Multimodal Understanding, Generation and Reconstruction
Paper • 2511.23386 • Published • 16 -
Visual Generation Tuning
Paper • 2511.23469 • Published • 16 -
A Style is Worth One Code: Unlocking Code-to-Style Image Generation with Discrete Style Space
Paper • 2511.10555 • Published • 63
-
Arbitrary-steps Image Super-resolution via Diffusion Inversion
Paper • 2412.09013 • Published • 13 -
Deep Researcher with Test-Time Diffusion
Paper • 2507.16075 • Published • 68 -
nablaNABLA: Neighborhood Adaptive Block-Level Attention
Paper • 2507.13546 • Published • 126 -
Yume: An Interactive World Generation Model
Paper • 2507.17744 • Published • 92
-
LoRACLR: Contrastive Adaptation for Customization of Diffusion Models
Paper • 2412.09622 • Published • 8 -
AnyDressing: Customizable Multi-Garment Virtual Dressing via Latent Diffusion Models
Paper • 2412.04146 • Published • 23 -
Learning Flow Fields in Attention for Controllable Person Image Generation
Paper • 2412.08486 • Published • 36 -
LoRA.rar: Learning to Merge LoRAs via Hypernetworks for Subject-Style Conditioned Image Generation
Paper • 2412.05148 • Published • 12
-
EVA-CLIP-18B: Scaling CLIP to 18 Billion Parameters
Paper • 2402.04252 • Published • 30 -
Vision Superalignment: Weak-to-Strong Generalization for Vision Foundation Models
Paper • 2402.03749 • Published • 15 -
ScreenAI: A Vision-Language Model for UI and Infographics Understanding
Paper • 2402.04615 • Published • 44 -
EfficientViT-SAM: Accelerated Segment Anything Model Without Performance Loss
Paper • 2402.05008 • Published • 23