Abstract: We introduce Jina Embeddings V3, a 570-million-parameter text embedding model that excels in long-context (up to 8192 tokens) and multilingual text retrieval tasks. The model incorporates task-specific Low-Rank Adaptation (LoRA) modules for high-quality embeddings specialized for retrieval, clustering, classification, and text matching. On the MTEB benchmark, Jina Embeddings V3 outperforms other embedding models of similar size.
Loading