| { |
| "1706.03762": { |
| "arxivId": "1706.03762", |
| "title": "Attention is All you Need" |
| }, |
| "1405.0312": { |
| "arxivId": "1405.0312", |
| "title": "Microsoft COCO: Common Objects in Context" |
| }, |
| "2005.14165": { |
| "arxivId": "2005.14165", |
| "title": "Language Models are Few-Shot Learners" |
| }, |
| "2103.00020": { |
| "arxivId": "2103.00020", |
| "title": "Learning Transferable Visual Models From Natural Language Supervision" |
| }, |
| "2006.11239": { |
| "arxivId": "2006.11239", |
| "title": "Denoising Diffusion Probabilistic Models" |
| }, |
| "2112.10752": { |
| "arxivId": "2112.10752", |
| "title": "High-Resolution Image Synthesis with Latent Diffusion Models" |
| }, |
| "2005.12872": { |
| "arxivId": "2005.12872", |
| "title": "End-to-End Object Detection with Transformers" |
| }, |
| "2203.02155": { |
| "arxivId": "2203.02155", |
| "title": "Training language models to follow instructions with human feedback" |
| }, |
| "2302.13971": { |
| "arxivId": "2302.13971", |
| "title": "LLaMA: Open and Efficient Foundation Language Models" |
| }, |
| "2303.08774": { |
| "arxivId": "2303.08774", |
| "title": "GPT-4 Technical Report" |
| }, |
| "2307.09288": { |
| "arxivId": "2307.09288", |
| "title": "Llama 2: Open Foundation and Fine-Tuned Chat Models" |
| }, |
| "2201.11903": { |
| "arxivId": "2201.11903", |
| "title": "Chain of Thought Prompting Elicits Reasoning in Large Language Models" |
| }, |
| "1412.2306": { |
| "arxivId": "1412.2306", |
| "title": "Deep visual-semantic alignments for generating image descriptions" |
| }, |
| "1602.07332": { |
| "arxivId": "1602.07332", |
| "title": "Visual Genome: Connecting Language and Vision Using Crowdsourced Dense Image Annotations" |
| }, |
| "1505.00468": { |
| "arxivId": "1505.00468", |
| "title": "VQA: Visual Question Answering" |
| }, |
| "2304.02643": { |
| "arxivId": "2304.02643", |
| "title": "Segment Anything" |
| }, |
| "1411.5726": { |
| "arxivId": "1411.5726", |
| "title": "CIDEr: Consensus-based image description evaluation" |
| }, |
| "1707.07998": { |
| "arxivId": "1707.07998", |
| "title": "Bottom-Up and Top-Down Attention for Image Captioning and Visual Question Answering" |
| }, |
| "2205.11916": { |
| "arxivId": "2205.11916", |
| "title": "Large Language Models are Zero-Shot Reasoners" |
| }, |
| "2109.01652": { |
| "arxivId": "2109.01652", |
| "title": "Finetuned Language Models Are Zero-Shot Learners" |
| }, |
| "2201.12086": { |
| "arxivId": "2201.12086", |
| "title": "BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation" |
| }, |
| "2301.12597": { |
| "arxivId": "2301.12597", |
| "title": "BLIP-2: Bootstrapping Language-Image Pre-training with Frozen Image Encoders and Large Language Models" |
| }, |
| "2210.11416": { |
| "arxivId": "2210.11416", |
| "title": "Scaling Instruction-Finetuned Language Models" |
| }, |
| "2204.14198": { |
| "arxivId": "2204.14198", |
| "title": "Flamingo: a Visual Language Model for Few-Shot Learning" |
| }, |
| "2304.08485": { |
| "arxivId": "2304.08485", |
| "title": "Visual Instruction Tuning" |
| }, |
| "2210.08402": { |
| "arxivId": "2210.08402", |
| "title": "LAION-5B: An open large-scale dataset for training next generation image-text models" |
| }, |
| "2305.18290": { |
| "arxivId": "2305.18290", |
| "title": "Direct Preference Optimization: Your Language Model is Secretly a Reward Model" |
| }, |
| "1505.04870": { |
| "arxivId": "1505.04870", |
| "title": "Flickr30k Entities: Collecting Region-to-Phrase Correspondences for Richer Image-to-Sentence Models" |
| }, |
| "2303.18223": { |
| "arxivId": "2303.18223", |
| "title": "A Survey of Large Language Models" |
| }, |
| "2304.07193": { |
| "arxivId": "2304.07193", |
| "title": "DINOv2: Learning Robust Visual Features without Supervision" |
| }, |
| "2101.03961": { |
| "arxivId": "2101.03961", |
| "title": "Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficient Sparsity" |
| }, |
| "2212.10560": { |
| "arxivId": "2212.10560", |
| "title": "Self-Instruct: Aligning Language Models with Self-Generated Instructions" |
| }, |
| "2107.07651": { |
| "arxivId": "2107.07651", |
| "title": "Align before Fuse: Vision and Language Representation Learning with Momentum Distillation" |
| }, |
| "2009.01325": { |
| "arxivId": "2009.01325", |
| "title": "Learning to summarize from human feedback" |
| }, |
| "2110.08207": { |
| "arxivId": "2110.08207", |
| "title": "Multitask Prompted Training Enables Zero-Shot Task Generalization" |
| }, |
| "2310.03744": { |
| "arxivId": "2310.03744", |
| "title": "Improved Baselines with Visual Instruction Tuning" |
| }, |
| "2305.06500": { |
| "arxivId": "2305.06500", |
| "title": "InstructBLIP: Towards General-purpose Vision-Language Models with Instruction Tuning" |
| }, |
| "2304.10592": { |
| "arxivId": "2304.10592", |
| "title": "MiniGPT-4: Enhancing Vision-Language Understanding with Advanced Large Language Models" |
| }, |
| "1312.6211": { |
| "arxivId": "1312.6211", |
| "title": "An Empirical Investigation of Catastrophic Forgeting in Gradient-Based Neural Networks" |
| }, |
| "1909.08593": { |
| "arxivId": "1909.08593", |
| "title": "Fine-Tuning Language Models from Human Preferences" |
| }, |
| "2302.04761": { |
| "arxivId": "2302.04761", |
| "title": "Toolformer: Language Models Can Teach Themselves to Use Tools" |
| }, |
| "2303.03378": { |
| "arxivId": "2303.03378", |
| "title": "PaLM-E: An Embodied Multimodal Language Model" |
| }, |
| "2111.02114": { |
| "arxivId": "2111.02114", |
| "title": "LAION-400M: Open Dataset of CLIP-Filtered 400 Million Image-Text Pairs" |
| }, |
| "2112.09332": { |
| "arxivId": "2112.09332", |
| "title": "WebGPT: Browser-assisted question-answering with human feedback" |
| }, |
| "2203.03605": { |
| "arxivId": "2203.03605", |
| "title": "DINO: DETR with Improved DeNoising Anchor Boxes for End-to-End Object Detection" |
| }, |
| "2104.08786": { |
| "arxivId": "2104.08786", |
| "title": "Fantastically Ordered Prompts and Where to Find Them: Overcoming Few-Shot Prompt Order Sensitivity" |
| }, |
| "2102.08981": { |
| "arxivId": "2102.08981", |
| "title": "Conceptual 12M: Pushing Web-Scale Image-Text Pre-Training To Recognize Long-Tail Visual Concepts" |
| }, |
| "2309.16609": { |
| "arxivId": "2309.16609", |
| "title": "Qwen Technical Report" |
| }, |
| "2205.10625": { |
| "arxivId": "2205.10625", |
| "title": "Least-to-Most Prompting Enables Complex Reasoning in Large Language Models" |
| }, |
| "1811.10830": { |
| "arxivId": "1811.10830", |
| "title": "From Recognition to Cognition: Visual Commonsense Reasoning" |
| }, |
| "2202.03052": { |
| "arxivId": "2202.03052", |
| "title": "OFA: Unifying Architectures, Tasks, and Modalities Through a Simple Sequence-to-Sequence Learning Framework" |
| }, |
| "1906.10770": { |
| "arxivId": "1906.10770", |
| "title": "Deep Modular Co-Attention Networks for Visual Question Answering" |
| }, |
| "1709.05522": { |
| "arxivId": "1709.05522", |
| "title": "AISHELL-1: An open-source Mandarin speech corpus and a speech recognition baseline" |
| }, |
| "2209.09513": { |
| "arxivId": "2209.09513", |
| "title": "Learn to Explain: Multimodal Reasoning via Thought Chains for Science Question Answering" |
| }, |
| "2304.14178": { |
| "arxivId": "2304.14178", |
| "title": "mPLUG-Owl: Modularization Empowers Large Language Models with Multimodality" |
| }, |
| "2108.10904": { |
| "arxivId": "2108.10904", |
| "title": "SimVLM: Simple Visual Language Model Pretraining with Weak Supervision" |
| }, |
| "2308.12966": { |
| "arxivId": "2308.12966", |
| "title": "Qwen-VL: A Frontier Large Vision-Language Model with Versatile Abilities" |
| }, |
| "2303.17580": { |
| "arxivId": "2303.17580", |
| "title": "HuggingGPT: Solving AI Tasks with ChatGPT and its Friends in Hugging Face" |
| }, |
| "2106.13884": { |
| "arxivId": "2106.13884", |
| "title": "Multimodal Few-Shot Learning with Frozen Language Models" |
| }, |
| "2303.16199": { |
| "arxivId": "2303.16199", |
| "title": "LLaMA-Adapter: Efficient Fine-tuning of Language Models with Zero-init Attention" |
| }, |
| "2305.05665": { |
| "arxivId": "2305.05665", |
| "title": "ImageBind One Embedding Space to Bind Them All" |
| }, |
| "2306.02858": { |
| "arxivId": "2306.02858", |
| "title": "Video-LLaMA: An Instruction-tuned Audio-Visual Language Model for Video Understanding" |
| }, |
| "2303.04671": { |
| "arxivId": "2303.04671", |
| "title": "Visual ChatGPT: Talking, Drawing and Editing with Visual Foundation Models" |
| }, |
| "2307.06281": { |
| "arxivId": "2307.06281", |
| "title": "MMBench: Is Your Multi-modal Model an All-around Player?" |
| }, |
| "2211.07636": { |
| "arxivId": "2211.07636", |
| "title": "EVA: Exploring the Limits of Masked Visual Representation Learning at Scale" |
| }, |
| "2204.00598": { |
| "arxivId": "2204.00598", |
| "title": "Socratic Models: Composing Zero-Shot Multimodal Reasoning with Language" |
| }, |
| "2212.07143": { |
| "arxivId": "2212.07143", |
| "title": "Reproducible Scaling Laws for Contrastive Language-Image Learning" |
| }, |
| "2306.14824": { |
| "arxivId": "2306.14824", |
| "title": "Kosmos-2: Grounding Multimodal Large Language Models to the World" |
| }, |
| "2102.02779": { |
| "arxivId": "2102.02779", |
| "title": "Unifying Vision-and-Language Tasks via Text Generation" |
| }, |
| "2304.03277": { |
| "arxivId": "2304.03277", |
| "title": "Instruction Tuning with GPT-4" |
| }, |
| "2306.13394": { |
| "arxivId": "2306.13394", |
| "title": "MME: A Comprehensive Evaluation Benchmark for Multimodal Large Language Models" |
| }, |
| "2304.15010": { |
| "arxivId": "2304.15010", |
| "title": "LLaMA-Adapter V2: Parameter-Efficient Visual Instruction Model" |
| }, |
| "2305.03726": { |
| "arxivId": "2305.03726", |
| "title": "Otter: A Multi-Modal Model with In-Context Instruction Tuning" |
| }, |
| "2309.17421": { |
| "arxivId": "2309.17421", |
| "title": "The Dawn of LMMs: Preliminary Explorations with GPT-4V(ision)" |
| }, |
| "2306.15195": { |
| "arxivId": "2306.15195", |
| "title": "Shikra: Unleashing Multimodal LLM's Referential Dialogue Magic" |
| }, |
| "2210.03493": { |
| "arxivId": "2210.03493", |
| "title": "Automatic Chain of Thought Prompting in Large Language Models" |
| }, |
| "2305.10355": { |
| "arxivId": "2305.10355", |
| "title": "Evaluating Object Hallucination in Large Vision-Language Models" |
| }, |
| "1812.08658": { |
| "arxivId": "1812.08658", |
| "title": "nocaps: novel object captioning at scale" |
| }, |
| "2306.00890": { |
| "arxivId": "2306.00890", |
| "title": "LLaVA-Med: Training a Large Language-and-Vision Assistant for Biomedicine in One Day" |
| }, |
| "2311.16502": { |
| "arxivId": "2311.16502", |
| "title": "MMMU: A Massive Multi-Discipline Multimodal Understanding and Reasoning Benchmark for Expert AGI" |
| }, |
| "2308.02490": { |
| "arxivId": "2308.02490", |
| "title": "MM-Vet: Evaluating Large Multimodal Models for Integrated Capabilities" |
| }, |
| "2211.10435": { |
| "arxivId": "2211.10435", |
| "title": "PAL: Program-aided Language Models" |
| }, |
| "1812.05252": { |
| "arxivId": "1812.05252", |
| "title": "Dynamic Fusion With Intra- and Inter-Modality Attention Flow for Visual Question Answering" |
| }, |
| "2109.05014": { |
| "arxivId": "2109.05014", |
| "title": "An Empirical Study of GPT-3 for Few-Shot Knowledge-Based VQA" |
| }, |
| "2305.11175": { |
| "arxivId": "2305.11175", |
| "title": "VisionLLM: Large Language Model is also an Open-Ended Decoder for Vision-Centric Tasks" |
| }, |
| "2305.06355": { |
| "arxivId": "2305.06355", |
| "title": "VideoChat: Chat-Centric Video Understanding" |
| }, |
| "2311.12793": { |
| "arxivId": "2311.12793", |
| "title": "ShareGPT4V: Improving Large Multi-Modal Models with Better Captions" |
| }, |
| "2307.16125": { |
| "arxivId": "2307.16125", |
| "title": "SEED-Bench: Benchmarking Multimodal LLMs with Generative Comprehension" |
| }, |
| "1809.02156": { |
| "arxivId": "1809.02156", |
| "title": "Object Hallucination in Image Captioning" |
| }, |
| "2306.05424": { |
| "arxivId": "2306.05424", |
| "title": "Video-ChatGPT: Towards Detailed Video Understanding via Large Vision and Language Models" |
| }, |
| "2303.11381": { |
| "arxivId": "2303.11381", |
| "title": "MM-REACT: Prompting ChatGPT for Multimodal Reasoning and Action" |
| }, |
| "2303.15389": { |
| "arxivId": "2303.15389", |
| "title": "EVA-CLIP: Improved Training Techniques for CLIP at Scale" |
| }, |
| "2211.11559": { |
| "arxivId": "2211.11559", |
| "title": "Visual Programming: Compositional visual reasoning without training" |
| }, |
| "1901.06706": { |
| "arxivId": "1901.06706", |
| "title": "Visual Entailment: A Novel Task for Fine-Grained Image Understanding" |
| }, |
| "2311.03079": { |
| "arxivId": "2311.03079", |
| "title": "CogVLM: Visual Expert for Pretrained Language Models" |
| }, |
| "2308.01390": { |
| "arxivId": "2308.01390", |
| "title": "OpenFlamingo: An Open-Source Framework for Training Large Autoregressive Vision-Language Models" |
| }, |
| "2302.00923": { |
| "arxivId": "2302.00923", |
| "title": "Multimodal Chain-of-Thought Reasoning in Language Models" |
| }, |
| "2309.05519": { |
| "arxivId": "2309.05519", |
| "title": "NExT-GPT: Any-to-Any Multimodal LLM" |
| }, |
| "1808.10583": { |
| "arxivId": "1808.10583", |
| "title": "AISHELL-2: Transforming Mandarin ASR Research Into Industrial Scale" |
| }, |
| "2310.02255": { |
| "arxivId": "2310.02255", |
| "title": "MathVista: Evaluating Mathematical Reasoning of Foundation Models in Visual Contexts" |
| }, |
| "2212.12017": { |
| "arxivId": "2212.12017", |
| "title": "OPT-IML: Scaling Language Model Instruction Meta Learning through the Lens of Generalization" |
| }, |
| "2304.09842": { |
| "arxivId": "2304.09842", |
| "title": "Chameleon: Plug-and-Play Compositional Reasoning with Large Language Models" |
| }, |
| "2308.00692": { |
| "arxivId": "2308.00692", |
| "title": "LISA: Reasoning Segmentation via Large Language Model" |
| }, |
| "2305.16355": { |
| "arxivId": "2305.16355", |
| "title": "PandaGPT: One Model To Instruction-Follow Them All" |
| }, |
| "1510.01431": { |
| "arxivId": "1510.01431", |
| "title": "SentiCap: Generating Image Descriptions with Sentiments" |
| }, |
| "2305.04790": { |
| "arxivId": "2305.04790", |
| "title": "MultiModal-GPT: A Vision and Language Model for Dialogue with Humans" |
| }, |
| "2305.11000": { |
| "arxivId": "2305.11000", |
| "title": "SpeechGPT: Empowering Large Language Models with Intrinsic Cross-Modal Conversational Abilities" |
| }, |
| "2209.14610": { |
| "arxivId": "2209.14610", |
| "title": "Dynamic Prompt Learning via Policy Gradient for Semi-structured Mathematical Reasoning" |
| }, |
| "2310.07704": { |
| "arxivId": "2310.07704", |
| "title": "Ferret: Refer and Ground Anything Anywhere at Any Granularity" |
| }, |
| "2306.12925": { |
| "arxivId": "2306.12925", |
| "title": "AudioPaLM: A Large Language Model That Can Speak and Listen" |
| }, |
| "2309.14525": { |
| "arxivId": "2309.14525", |
| "title": "Aligning Large Multimodal Models with Factually Augmented RLHF" |
| }, |
| "2306.05425": { |
| "arxivId": "2306.05425", |
| "title": "MIMIC-IT: Multi-Modal In-Context Instruction Tuning" |
| }, |
| "2307.03601": { |
| "arxivId": "2307.03601", |
| "title": "GPT4RoI: Instruction Tuning Large Language Model on Region-of-Interest" |
| }, |
| "2102.09542": { |
| "arxivId": "2102.09542", |
| "title": "Slake: A Semantically-Labeled Knowledge-Enhanced Dataset For Medical Visual Question Answering" |
| }, |
| "2307.12981": { |
| "arxivId": "2307.12981", |
| "title": "3D-LLM: Injecting the 3D World into Large Language Models" |
| }, |
| "2305.18752": { |
| "arxivId": "2305.18752", |
| "title": "GPT4Tools: Teaching Large Language Model to Use Tools via Self-instruction" |
| }, |
| "2311.06607": { |
| "arxivId": "2311.06607", |
| "title": "Monkey: Image Resolution and Text Label are Important Things for Large Multi-Modal Models" |
| }, |
| "2306.14565": { |
| "arxivId": "2306.14565", |
| "title": "Mitigating Hallucination in Large Multi-Modal Models via Robust Instruction Tuning" |
| }, |
| "2305.15021": { |
| "arxivId": "2305.15021", |
| "title": "EmbodiedGPT: Vision-Language Pre-Training via Embodied Chain of Thought" |
| }, |
| "2311.07575": { |
| "arxivId": "2311.07575", |
| "title": "SPHINX: The Joint Mixing of Weights, Tasks, and Visual Embeddings for Multi-modal Large Language Models" |
| }, |
| "2307.15189": { |
| "arxivId": "2307.15189", |
| "title": "Med-Flamingo: a Multimodal Medical Few-shot Learner" |
| }, |
| "2303.17395": { |
| "arxivId": "2303.17395", |
| "title": "WavCaps: A ChatGPT-Assisted Weakly-Labelled Audio Captioning Dataset for Audio-Language Multimodal Research" |
| }, |
| "1711.06475": { |
| "arxivId": "1711.06475", |
| "title": "AI Challenger : A Large-scale Dataset for Going Deeper in Image Understanding" |
| }, |
| "2003.10286": { |
| "arxivId": "2003.10286", |
| "title": "PathVQA: 30000+ Questions for Medical Visual Question Answering" |
| }, |
| "2303.02151": { |
| "arxivId": "2303.02151", |
| "title": "Prompt, Generate, Then Cache: Cascade of Foundation Models Makes Strong Few-Shot Learners" |
| }, |
| "2403.09611": { |
| "arxivId": "2403.09611", |
| "title": "MM1: Methods, Analysis & Insights from Multimodal LLM Pre-training" |
| }, |
| "2205.12255": { |
| "arxivId": "2205.12255", |
| "title": "TALM: Tool Augmented Language Models" |
| }, |
| "2306.06687": { |
| "arxivId": "2306.06687", |
| "title": "LAMM: Language-Assisted Multi-Modal Instruction-Tuning Dataset, Framework, and Benchmark" |
| }, |
| "2306.09093": { |
| "arxivId": "2306.09093", |
| "title": "Macaw-LLM: Multi-Modal Language Modeling with Image, Audio, Video, and Text Integration" |
| }, |
| "2306.09265": { |
| "arxivId": "2306.09265", |
| "title": "LVLM-eHub: A Comprehensive Evaluation Benchmark for Large Vision-Language Models" |
| }, |
| "2305.11834": { |
| "arxivId": "2305.11834", |
| "title": "Pengi: An Audio Language Model for Audio Tasks" |
| }, |
| "2309.07915": { |
| "arxivId": "2309.07915", |
| "title": "MMICL: Empowering Vision-language Model with Multi-Modal In-Context Learning" |
| }, |
| "2311.16922": { |
| "arxivId": "2311.16922", |
| "title": "Mitigating Object Hallucinations in Large Vision-Language Models through Visual Contrastive Decoding" |
| }, |
| "2307.05222": { |
| "arxivId": "2307.05222", |
| "title": "Generative Pretraining in Multimodality" |
| }, |
| "2310.00754": { |
| "arxivId": "2310.00754", |
| "title": "Analyzing and Mitigating Object Hallucination in Large Vision-Language Models" |
| }, |
| "2312.13771": { |
| "arxivId": "2312.13771", |
| "title": "AppAgent: Multimodal Agents as Smartphone Users" |
| }, |
| "2306.04387": { |
| "arxivId": "2306.04387", |
| "title": "M3IT: A Large-Scale Dataset towards Multi-Modal Multilingual Instruction Tuning" |
| }, |
| "2212.10773": { |
| "arxivId": "2212.10773", |
| "title": "MultiInstruct: Improving Multi-Modal Zero-Shot Learning via Instruction Tuning" |
| }, |
| "2305.16934": { |
| "arxivId": "2305.16934", |
| "title": "On Evaluating Adversarial Robustness of Large Vision-Language Models" |
| }, |
| "2401.15947": { |
| "arxivId": "2401.15947", |
| "title": "MoE-LLaVA: Mixture of Experts for Large Vision-Language Models" |
| }, |
| "2312.00849": { |
| "arxivId": "2312.00849", |
| "title": "RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-Grained Correctional Human Feedback" |
| }, |
| "2311.03356": { |
| "arxivId": "2311.03356", |
| "title": "GLaMM: Pixel Grounding Large Multimodal Model" |
| }, |
| "2309.03905": { |
| "arxivId": "2309.03905", |
| "title": "ImageBind-LLM: Multi-modality Instruction Tuning" |
| }, |
| "2308.16911": { |
| "arxivId": "2308.16911", |
| "title": "PointLLM: Empowering Large Language Models to Understand Point Clouds" |
| }, |
| "2305.10415": { |
| "arxivId": "2305.10415", |
| "title": "PMC-VQA: Visual Instruction Tuning for Medical Visual Question Answering" |
| }, |
| "2310.14566": { |
| "arxivId": "2310.14566", |
| "title": "HallusionBench: You See What You Think? Or You Think What You See? An Image-Context Reasoning Benchmark Challenging for GPT-4V(ision), LLaVA-1.5, and Other Multi-modality Models" |
| }, |
| "2212.10846": { |
| "arxivId": "2212.10846", |
| "title": "From Images to Textual Prompts: Zero-shot Visual Question Answering with Frozen Large Language Models" |
| }, |
| "2305.04160": { |
| "arxivId": "2305.04160", |
| "title": "X-LLM: Bootstrapping Advanced Large Language Models by Treating Multi-Modalities as Foreign Languages" |
| }, |
| "2303.06594": { |
| "arxivId": "2303.06594", |
| "title": "ChatGPT Asks, BLIP-2 Answers: Automatic Questioning Towards Enriched Visual Descriptions" |
| }, |
| "2307.02499": { |
| "arxivId": "2307.02499", |
| "title": "mPLUG-DocOwl: Modularized Multimodal Large Language Model for Document Understanding" |
| }, |
| "2305.14167": { |
| "arxivId": "2305.14167", |
| "title": "DetGPT: Detect What You Need via Reasoning" |
| }, |
| "2211.11682": { |
| "arxivId": "2211.11682", |
| "title": "PointCLIP V2: Adapting CLIP for Powerful 3D Open-world Learning" |
| }, |
| "2309.16058": { |
| "arxivId": "2309.16058", |
| "title": "AnyMAL: An Efficient and Scalable Any-Modality Augmented Language Model" |
| }, |
| "2402.11684": { |
| "arxivId": "2402.11684", |
| "title": "ALLaVA: Harnessing GPT4V-Synthesized Data for Lite Vision-Language Models" |
| }, |
| "2211.16198": { |
| "arxivId": "2211.16198", |
| "title": "SuS-X: Training-Free Name-Only Transfer of Vision-Language Models" |
| }, |
| "2310.16045": { |
| "arxivId": "2310.16045", |
| "title": "Woodpecker: Hallucination Correction for Multimodal Large Language Models" |
| }, |
| "2311.07574": { |
| "arxivId": "2311.07574", |
| "title": "To See is to Believe: Prompting GPT-4V for Better Visual Instruction Tuning" |
| }, |
| "2307.14539": { |
| "arxivId": "2307.14539", |
| "title": "Jailbreak in pieces: Compositional Adversarial Attacks on Multi-Modal Language Models" |
| }, |
| "2305.15023": { |
| "arxivId": "2305.15023", |
| "title": "Cheap and Quick: Efficient Vision-Language Instruction Tuning for Large Language Models" |
| }, |
| "2305.02677": { |
| "arxivId": "2305.02677", |
| "title": "Caption Anything: Interactive Image Description with Diverse Multimodal Controls" |
| }, |
| "2311.07397": { |
| "arxivId": "2311.07397", |
| "title": "An LLM-free Multi-dimensional Benchmark for MLLMs Hallucination Evaluation" |
| }, |
| "2311.05332": { |
| "arxivId": "2311.05332", |
| "title": "On the Road with GPT-4V(ision): Early Explorations of Visual-Language Model on Autonomous Driving" |
| }, |
| "2307.02469": { |
| "arxivId": "2307.02469", |
| "title": "What Matters in Training a GPT4-Style Language Model with Multimodal Inputs?" |
| }, |
| "2402.03766": { |
| "arxivId": "2402.03766", |
| "title": "MobileVLM V2: Faster and Stronger Baseline for Vision Language Model" |
| }, |
| "2312.14135": { |
| "arxivId": "2312.14135", |
| "title": "V*: Guided Visual Search as a Core Mechanism in Multimodal LLMs" |
| }, |
| "2202.06767": { |
| "arxivId": "2202.06767", |
| "title": "Wukong: A 100 Million Large-scale Chinese Cross-modal Pre-training Benchmark" |
| }, |
| "2403.12895": { |
| "arxivId": "2403.12895", |
| "title": "mPLUG-DocOwl 1.5: Unified Structure Learning for OCR-free Document Understanding" |
| }, |
| "2311.12871": { |
| "arxivId": "2311.12871", |
| "title": "An Embodied Generalist Agent in 3D World" |
| }, |
| "2310.16436": { |
| "arxivId": "2310.16436", |
| "title": "DDCoT: Duty-Distinct Chain-of-Thought Prompting for Multimodal Reasoning in Language Models" |
| }, |
| "2402.12226": { |
| "arxivId": "2402.12226", |
| "title": "AnyGPT: Unified Multimodal LLM with Discrete Sequence Modeling" |
| }, |
| "2310.05126": { |
| "arxivId": "2310.05126", |
| "title": "UReader: Universal OCR-free Visually-situated Language Understanding with Multimodal Large Language Model" |
| }, |
| "2308.15126": { |
| "arxivId": "2308.15126", |
| "title": "Evaluation and Analysis of Hallucination in Large Vision-Language Models" |
| }, |
| "2401.16158": { |
| "arxivId": "2401.16158", |
| "title": "Mobile-Agent: Autonomous Multi-Modal Mobile Device Agent with Visual Perception" |
| }, |
| "2403.04473": { |
| "arxivId": "2403.04473", |
| "title": "TextMonkey: An OCR-Free Large Multimodal Model for Understanding Document" |
| }, |
| "2309.09971": { |
| "arxivId": "2309.09971", |
| "title": "MindAgent: Emergent Gaming Interaction" |
| }, |
| "2308.12067": { |
| "arxivId": "2308.12067", |
| "title": "InstructionGPT-4: A 200-Instruction Paradigm for Fine-Tuning MiniGPT-4" |
| }, |
| "2312.12436": { |
| "arxivId": "2312.12436", |
| "title": "A Challenger to GPT-4V? Early Explorations of Gemini in Visual Expertise" |
| }, |
| "2312.10665": { |
| "arxivId": "2312.10665", |
| "title": "Silkie: Preference Distillation for Large Visual Language Models" |
| }, |
| "2312.10032": { |
| "arxivId": "2312.10032", |
| "title": "Osprey: Pixel Understanding with Visual Instruction Tuning" |
| }, |
| "2305.16103": { |
| "arxivId": "2305.16103", |
| "title": "ChatBridge: Bridging Modalities with Large Language Model as a Language Catalyst" |
| }, |
| "2305.14705": { |
| "arxivId": "2305.14705", |
| "title": "Mixture-of-Experts Meets Instruction Tuning: A Winning Combination for Large Language Models" |
| }, |
| "2310.01779": { |
| "arxivId": "2310.01779", |
| "title": "HallE-Switch: Rethinking and Controlling Object Existence Hallucinations in Large Vision Language Models for Detailed Caption" |
| }, |
| "2305.14985": { |
| "arxivId": "2305.14985", |
| "title": "IdealGPT: Iteratively Decomposing Vision and Language Reasoning via Large Language Models" |
| }, |
| "2311.18651": { |
| "arxivId": "2311.18651", |
| "title": "LL3DA: Visual Interactive Instruction Tuning for Omni-3D Understanding, Reasoning, and Planning" |
| }, |
| "2308.12038": { |
| "arxivId": "2308.12038", |
| "title": "Large Multilingual Models Pivot Zero-Shot Multimodal Learning across Languages" |
| }, |
| "2311.16103": { |
| "arxivId": "2311.16103", |
| "title": "Video-Bench: A Comprehensive Benchmark and Toolkit for Evaluating Video-based Large Language Models" |
| }, |
| "2310.00582": { |
| "arxivId": "2310.00582", |
| "title": "Pink: Unveiling the Power of Referential Comprehension for Multi-modal LLMs" |
| }, |
| "2312.06968": { |
| "arxivId": "2312.06968", |
| "title": "Hallucination Augmented Contrastive Learning for Multimodal Large Language Model" |
| }, |
| "2309.09958": { |
| "arxivId": "2309.09958", |
| "title": "An Empirical Study of Scaling Instruct-Tuned Large Multimodal Models" |
| }, |
| "2305.02317": { |
| "arxivId": "2305.02317", |
| "title": "Visual Chain of Thought: Bridging Logical Gaps with Multimodal Infillings" |
| }, |
| "2311.01477": { |
| "arxivId": "2311.01477", |
| "title": "FAITHSCORE: Evaluating Hallucinations in Large Vision-Language Models" |
| }, |
| "2309.15564": { |
| "arxivId": "2309.15564", |
| "title": "Jointly Training Large Autoregressive Multimodal Models" |
| }, |
| "2304.07919": { |
| "arxivId": "2304.07919", |
| "title": "Chain of Thought Prompt Tuning in Vision Language Models" |
| }, |
| "2401.12915": { |
| "arxivId": "2401.12915", |
| "title": "Red Teaming Visual Language Models" |
| }, |
| "2311.18248": { |
| "arxivId": "2311.18248", |
| "title": "mPLUG-PaperOwl: Scientific Diagram Analysis with the Multimodal Large Language Model" |
| }, |
| "2312.02153": { |
| "arxivId": "2312.02153", |
| "title": "Aligning and Prompting Everything All at Once for Universal Visual Perception" |
| }, |
| "2311.01487": { |
| "arxivId": "2311.01487", |
| "title": "What Makes for Good Visual Instructions? Synthesizing Complex Visual Reasoning Instructions for Visual Instruction Tuning" |
| }, |
| "2308.07891": { |
| "arxivId": "2308.07891", |
| "title": "Link-Context Learning for Multimodal LLMs" |
| }, |
| "2401.06395": { |
| "arxivId": "2401.06395", |
| "title": "ModaVerse: Efficiently Transforming Modalities with LLMs" |
| }, |
| "2312.07553": { |
| "arxivId": "2312.07553", |
| "title": "Hijacking Context in Large Multi-modal Models" |
| }, |
| "2312.02520": { |
| "arxivId": "2312.02520", |
| "title": "Towards More Unified In-Context Visual Understanding" |
| }, |
| "2305.13903": { |
| "arxivId": "2305.13903", |
| "title": "Let's Think Frame by Frame: Evaluating Video Chain of Thought with Video Infilling and Prediction" |
| } |
| } |