The newest flagship non-reasoning model series.
AI & ML interests
None defined yet.
Recent Activity
View all activity
Papers
Zooming without Zooming: Region-to-Image Distillation for Fine-Grained Multimodal Perception
UI-Venus-1.5 Technical Report
Ming is the multi-modal series of any-to-any models developed by Ant Ling team.
-
inclusionAI/Ming-flash-omni-2.0
Any-to-Any β’ Updated β’ 5.81k β’ 258 -
inclusionAI/Ming-omni-tts-16.8B-A3B
Text-to-Speech β’ 18B β’ Updated β’ 646 β’ 30 -
inclusionAI/Ming-omni-tts-0.5B
Text-to-Speech β’ 2B β’ Updated β’ 4.74k β’ 34 -
inclusionAI/Ming-omni-tts-tokenizer-12Hz
Audio-to-Audio β’ 0.8B β’ Updated β’ 10 β’ 7
-
Zooming without Zooming: Region-to-Image Distillation for Fine-Grained Multimodal Perception
Paper β’ 2602.11858 β’ Published β’ 62 -
inclusionAI/ZwZ-4B
Image-Text-to-Text β’ 5B β’ Updated β’ 350 β’ 31 -
inclusionAI/ZwZ-8B
Image-Text-to-Text β’ 9B β’ Updated β’ 10.1k β’ 44 -
inclusionAI/ZwZ-RL-VQA
Viewer β’ Updated β’ 74k β’ 1.62k β’ 10
-
LLaDA2.0: Scaling Up Diffusion Language Models to 100B
Paper β’ 2512.15745 β’ Published β’ 88 -
inclusionAI/LLaDA2.0-flash
Text Generation β’ 103B β’ Updated β’ 882 β’ 68 -
inclusionAI/LLaDA2.0-mini
Text Generation β’ 16B β’ Updated β’ 78.4k β’ 63 -
inclusionAI/LLaDA2.0-flash-preview
Text Generation β’ 103B β’ Updated β’ 32 β’ 68
Ring is a reasoning MoE LLM provided and open-sourced by InclusionAI, derived from Ling.
The Agent Runtime for Self-Improvement
-
UI-Venus-1.5 Technical Report
Paper β’ 2602.09082 β’ Published β’ 157 -
inclusionAI/UI-Venus-1.5-30B-A3B
Image-Text-to-Text β’ 31B β’ Updated β’ 3.18k β’ 25 -
inclusionAI/UI-Venus-1.5-8B
Image-Text-to-Text β’ 9B β’ Updated β’ 4k β’ 25 -
inclusionAI/UI-Venus-1.5-2B
Image-Text-to-Text β’ 2B β’ Updated β’ 1.79k β’ 35
-
Ming-Omni: A Unified Multimodal Model for Perception and Generation
Paper β’ 2506.09344 β’ Published β’ 31 -
inclusionAI/Ming-Lite-Omni
Any-to-Any β’ 19B β’ Updated β’ 81 β’ 198 -
inclusionAI/Ming-Lite-Omni-1.5
Any-to-Any β’ Updated β’ 246 β’ 85 -
inclusionAI/Ming-UniAudio-16B-A3B
Any-to-Any β’ 18B β’ Updated β’ 100 β’ 78
A collection of TwinFlow-accelerated diffusion models
GroveMoE is an open-source family of large language models developed by the AGI Center, Ant Research Institute.
-
inclusionAI/Ling-lite-1.5-2507
Text Generation β’ 17B β’ Updated β’ 41 β’ 76 -
inclusionAI/Ling-lite-1.5-2506
Text Generation β’ 17B β’ Updated β’ 48 β’ 52 -
inclusionAI/Ling-lite-1.5
Text Generation β’ 17B β’ Updated β’ 20.8k β’ 57 -
inclusionAI/Ling-lite-base-1.5
Text Generation β’ 17B β’ Updated β’ 35 β’ 33
AReaL-boba-2
The newest flagship non-reasoning model series.
Ming is the multi-modal series of any-to-any models developed by Ant Ling team.
-
inclusionAI/Ming-flash-omni-2.0
Any-to-Any β’ Updated β’ 5.81k β’ 258 -
inclusionAI/Ming-omni-tts-16.8B-A3B
Text-to-Speech β’ 18B β’ Updated β’ 646 β’ 30 -
inclusionAI/Ming-omni-tts-0.5B
Text-to-Speech β’ 2B β’ Updated β’ 4.74k β’ 34 -
inclusionAI/Ming-omni-tts-tokenizer-12Hz
Audio-to-Audio β’ 0.8B β’ Updated β’ 10 β’ 7
-
Zooming without Zooming: Region-to-Image Distillation for Fine-Grained Multimodal Perception
Paper β’ 2602.11858 β’ Published β’ 62 -
inclusionAI/ZwZ-4B
Image-Text-to-Text β’ 5B β’ Updated β’ 350 β’ 31 -
inclusionAI/ZwZ-8B
Image-Text-to-Text β’ 9B β’ Updated β’ 10.1k β’ 44 -
inclusionAI/ZwZ-RL-VQA
Viewer β’ Updated β’ 74k β’ 1.62k β’ 10
-
LLaDA2.0: Scaling Up Diffusion Language Models to 100B
Paper β’ 2512.15745 β’ Published β’ 88 -
inclusionAI/LLaDA2.0-flash
Text Generation β’ 103B β’ Updated β’ 882 β’ 68 -
inclusionAI/LLaDA2.0-mini
Text Generation β’ 16B β’ Updated β’ 78.4k β’ 63 -
inclusionAI/LLaDA2.0-flash-preview
Text Generation β’ 103B β’ Updated β’ 32 β’ 68
A collection of TwinFlow-accelerated diffusion models
Ring is a reasoning MoE LLM provided and open-sourced by InclusionAI, derived from Ling.
The Agent Runtime for Self-Improvement
GroveMoE is an open-source family of large language models developed by the AGI Center, Ant Research Institute.
-
UI-Venus-1.5 Technical Report
Paper β’ 2602.09082 β’ Published β’ 157 -
inclusionAI/UI-Venus-1.5-30B-A3B
Image-Text-to-Text β’ 31B β’ Updated β’ 3.18k β’ 25 -
inclusionAI/UI-Venus-1.5-8B
Image-Text-to-Text β’ 9B β’ Updated β’ 4k β’ 25 -
inclusionAI/UI-Venus-1.5-2B
Image-Text-to-Text β’ 2B β’ Updated β’ 1.79k β’ 35
-
inclusionAI/Ling-lite-1.5-2507
Text Generation β’ 17B β’ Updated β’ 41 β’ 76 -
inclusionAI/Ling-lite-1.5-2506
Text Generation β’ 17B β’ Updated β’ 48 β’ 52 -
inclusionAI/Ling-lite-1.5
Text Generation β’ 17B β’ Updated β’ 20.8k β’ 57 -
inclusionAI/Ling-lite-base-1.5
Text Generation β’ 17B β’ Updated β’ 35 β’ 33
AReaL-boba-2
-
Ming-Omni: A Unified Multimodal Model for Perception and Generation
Paper β’ 2506.09344 β’ Published β’ 31 -
inclusionAI/Ming-Lite-Omni
Any-to-Any β’ 19B β’ Updated β’ 81 β’ 198 -
inclusionAI/Ming-Lite-Omni-1.5
Any-to-Any β’ Updated β’ 246 β’ 85 -
inclusionAI/Ming-UniAudio-16B-A3B
Any-to-Any β’ 18B β’ Updated β’ 100 β’ 78