Text To Image
updated
Improving Text-to-Image Consistency via Automatic Prompt Optimization
Paper
• 2403.17804
• Published • 19
Be Yourself: Bounded Attention for Multi-Subject Text-to-Image
Generation
Paper
• 2403.16990
• Published • 25
Getting it Right: Improving Spatial Consistency in Text-to-Image Models
Paper
• 2404.01197
• Published • 31
Condition-Aware Neural Network for Controlled Image Generation
Paper
• 2404.01143
• Published • 13
TextCraftor: Your Text Encoder Can be Image Quality Controller
Paper
• 2403.18978
• Published • 15
ObjectDrop: Bootstrapping Counterfactuals for Photorealistic Object
Removal and Insertion
Paper
• 2403.18818
• Published • 28
FlexEdit: Flexible and Controllable Diffusion-based Object-centric Image
Editing
Paper
• 2403.18605
• Published • 11
InstantStyle: Free Lunch towards Style-Preserving in Text-to-Image
Generation
Paper
• 2404.02733
• Published • 22
ByteEdit: Boost, Comply and Accelerate Generative Image Editing
Paper
• 2404.04860
• Published • 25
BeyondScene: Higher-Resolution Human-Centric Scene Generation With
Pretrained Diffusion
Paper
• 2404.04544
• Published • 23
SwapAnything: Enabling Arbitrary Object Swapping in Personalized Visual
Editing
Paper
• 2404.05717
• Published • 26
MoMA: Multimodal LLM Adapter for Fast Personalized Image Generation
Paper
• 2404.05674
• Published • 15
ControlNet++: Improving Conditional Controls with Efficient Consistency
Feedback
Paper
• 2404.07987
• Published • 48
Ferret-v2: An Improved Baseline for Referring and Grounding with Large
Language Models
Paper
• 2404.07973
• Published • 32
BitsFusion: 1.99 bits Weight Quantization of Diffusion Model
Paper
• 2406.04333
• Published • 38
An Image is Worth 32 Tokens for Reconstruction and Generation
Paper
• 2406.07550
• Published • 60
An Image is Worth More Than 16x16 Patches: Exploring Transformers on
Individual Pixels
Paper
• 2406.09415
• Published • 51
EMMA: Your Text-to-Image Diffusion Model Can Secretly Accept Multi-Modal
Prompts
Paper
• 2406.09162
• Published • 14