Papers
arxiv:2406.18587
Nomic Embed Vision: Expanding the Latent Space
Published on Jun 6, 2024
Authors:
Abstract
Nomic-embed-vision and nomic-embed-text form a unified latent space for high-performance vision, language, and multimodal tasks.
AI-generated summary
This technical report describes the training of nomic-embed-vision, a highly performant, open-code, open-weights image embedding model that shares the same latent space as nomic-embed-text. Together, nomic-embed-vision and nomic-embed-text form the first unified latent space to achieve high performance across vision, language, and multimodal tasks.
Get this paper in your agent:
hf papers read 2406.18587 Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash Models citing this paper 2
Datasets citing this paper 0
No dataset linking this paper
Cite arxiv.org/abs/2406.18587 in a dataset README.md to link it from this page.