Jina CLIP v2 is a state-of-the-art multilingual and multimodal (text-image) embedding model, It excels in both cross-modal (text-to-image, image-to-text) and unimodal (text-to-text) retrieval tasks within a single vector space. It supports 100 languages with a focus on 30 (including English, Spanish, Chinese, Arabic, and more), it supports flexible embedding generation through Matryoshka Representation Learning (MRL) and allows for shortened vector lengths via the dimensions
parameter.
jina-clip-v2
Lorem Ipsum
Was this page helpful?