jina-embeddings-v5-text
Collection
Our 5th-gen embeddings: two lightweight multilingual models with SOTA performance in retrieval, matching, clustering, and classification. • 29 items • Updated • 38
MLX port of jina-embeddings-v5-text-nano-text-matching for efficient inference on Apple Silicon. For the full-size model, see jina-embeddings-v5-text-small-text-matching-mlx.
Elastic Inference Service | ArXiv | Blog
| Feature | Value |
|---|---|
| Parameters | 239M |
| Supported Task | text-matching |
| Max Sequence Length | 8192 |
| Embedding Dimension | 768 |
| Pooling Strategy | Last-token pooling |
| Base Model | EuroBERT/EuroBERT-210m |
| Format | MLX float16 safetensors |
For training details and evaluation results, see our technical report.
pip install mlx tokenizers
git clone https://huggingface.co/jinaai/jina-embeddings-v5-text-nano-text-matching-mlx
cd jina-embeddings-v5-text-nano-text-matching-mlx
import mlx.core as mx
from tokenizers import Tokenizer
from model import JinaEmbeddingModel
import json
with open("config.json") as f:
config = json.load(f)
model = JinaEmbeddingModel(config)
weights = mx.load("model.safetensors")
model.load_weights(list(weights.items()))
tokenizer = Tokenizer.from_file("tokenizer.json")
texts = [
"غروب جميل على الشاطئ", # Arabic
"海滩上美丽的日落", # Chinese
"Un beau coucher de soleil sur la plage", # French
"Ein wunderschöner Sonnenuntergang am Strand", # German
]
embeddings = model.encode(texts, tokenizer, task_type="text-matching")
jina-embeddings-v5-text-nano is licensed under CC BY-NC 4.0. For commercial use, please contact sales@jina.ai.
If you find jina-embeddings-v5-text-nano useful in your research, please cite the following paper:
@article{akram2026jina,
title={jina-embeddings-v5-text: Task-Targeted Embedding Distillation},
author={Mohammad Kalim Akram and Saba Sturua and Nastia Havriushenko and Quentin Herreros and Michael G{\"u}nther and Maximilian Werk and Han Xiao},
journal={arXiv preprint arXiv:2602.15547},
year={2026}
}
Quantized
Base model
EuroBERT/EuroBERT-210m