Massive Text Embedding Benchmark across 8 task categories
6 results indexed across 1 metric. Shaded row marks current SOTA; ties broken by submission date.
| # | Model | Org | Submitted | Paper / code | avg-score |
|---|---|---|---|---|---|
| 01 | NV-Embed-v2OSS | NVIDIA | Sep 2024 | NV-Embed: Improved Techniques for Training LLMs as Gener… | 72.31 |
| 02 | GTE-Qwen2-7B-instructOSS | Alibaba | Jun 2024 | arxiv | 72.05 |
| 03 | voyage-3-large | Voyage AI | Jan 2025 | arxiv | 70.32 |
| 04 | E5-Mistral-7B-instructOSS | Microsoft | Jan 2024 | Improving Text Embeddings with Large Language Models | 66.63 |
| 05 | jina-embeddings-v3OSS | Jina AI | Sep 2024 | jina-embeddings-v3: Multilingual Embeddings With Task Lo… | 65.18 |
| 06 | text-embedding-3-large | OpenAI | Jan 2024 | arxiv | 64.60 |
Every paper below corresponds to at least one row in the leaderboard above. Click through for the arXiv preprint and, when available, the reference implementation.
Submit a checkpoint and a reproduction script. We will run it, publish the score, and — if it takes the top — annotate the step on the progress chart with your name.