Papers
arxiv:2406.18587

Nomic Embed Vision: Expanding the Latent Space

Published on Jun 6, 2024
Authors:
,

Abstract

Nomic-embed-vision and nomic-embed-text form a unified latent space for high-performance vision, language, and multimodal tasks.

AI-generated summary

This technical report describes the training of nomic-embed-vision, a highly performant, open-code, open-weights image embedding model that shares the same latent space as nomic-embed-text. Together, nomic-embed-vision and nomic-embed-text form the first unified latent space to achieve high performance across vision, language, and multimodal tasks.

Community

Sign up or log in to comment

Get this paper in your agent:

hf papers read 2406.18587
Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash

Models citing this paper 2

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2406.18587 in a dataset README.md to link it from this page.

Spaces citing this paper 16

Collections including this paper 1