Jina CLIP v2 is a state-of-the-art multilingual and multimodal (text-image) embedding model. It excels in both cross-modal (text-to-image, image-to-text) and unimodal (text-to-text) retrieval tasks within a single vector space. It supports 100 languages with a focus on 30 (including English, Spanish, Chinese, Arabic, and more) and flexible embedding generation through Matryoshka Representation Learning (MRL), and it allows for shortened vector lengths via the dimensions
parameter.
jina-clip-v2
Lorem Ipsum
Was this page helpful?