Mixedbread

Reranking Models

The Mixedbread rerank family is a collection of state-of-the-art, open-source reranking models designed to significantly enhance search accuracy across various domains. These models can be seamlessly integrated into existing search systems, offering best-in-class performance and easy implementation for improved user satisfaction in search results.

Boost your search with our crispy reranking models! The Mixedbread rerank family offers state-of-the-art performance across a large variety of domains and can be easily integrated into your existing search stack.

What's new in the Mixedbread rerank family?

We recently finished baking a fresh set of rerank models, the mxbai-rerank-v2 series. These models feature reinforcement learning training, multilingual support for 100+ languages, and extended context handling up to 32K tokens. After receiving a wave of interest from the community, we're now happy to provide access to the models with the highest demand via our API:

ModelStatusContext Length (tokens)Description
up to 32kDelivers the highest accuracy, performance and supports multiple languages
API unavailableup to 32kStrikes balance between size and performance with better accuracy
512Delivers the highest accuracy and performance
API unavailable512Strikes balance between size and performance
API unavailable512Focuses on capacity-efficiency while retaining performance

Why Mixedbread rerank?

Not only are the Mixedbread rerank models powerful and fully open-source, they're also extremely easy to integrate into your current search stack. All you need to do is give the original search query as well as your search system's output to our reranking models, and they will tremendously boost your search accuracy - your users will love it!

Training Methodology

Our v2 models were built using a three-step reinforcement learning process:

  1. GRPO (Guided Reinforcement Prompt Optimization) - Teaching the model to output clear relevance scores
  2. Contrastive Learning - Developing fine-grained understanding of query-document relationships
  3. Preference Learning - Tuning the model to prioritize the most relevant documents

This layered approach yields a richer query understanding whether you're reordering text results, code snippets, or product listings.

Performance Benchmarks

We evaluated our models by letting them perform the reranking step on the top 100 lexical search results on a subset of the BEIR benchmark, a commonly used collection of evaluation datasets. Specifically, we used the NDCG@10 metric, which measures the overall relevance of the search results compared to the order in which they are ranked by the model, and the accuracy@3 metric, which measures the likelihood of a highly relevant search result appearing in the top 3 results - in our opinion, this is the most important metric to anticipate user satisfaction.

For illustrative purposes, we also included classic keyword search and a current full semantic search model in the evaluation. The results make us confident that our models show best-in-class performance in their size category:

Comparison of overall relevance scores between the Mixedbread rerank family and other models

Comparison of overall relevance scores between the Mixedbread rerank family and other models

ModelBEIR Accuracy
mxbai-rerank-base-v255.57
mxbai-rerank-large-v257.49
Comparison of accuracy scores between the Mixedbread rerank family and other models

Latency Comparison

Below is latency per query (seconds) on the NFC dataset, tested on an A100 (80GB) GPU:

ModelLatency (s)
mixedbread-ai/mxbai-rerank-xsmall-v10.32
mixedbread-ai/mxbai-rerank-base-v20.67
mixedbread-ai/mxbai-rerank-base-v10.76
mixedbread-ai/mxbai-rerank-large-v20.89
mixedbread-ai/mxbai-rerank-large-v12.24
BAAI/bge-reranker-v2-m33.05
BAAI/bge-reranker-v2-gemma7.20

Our 1.5B model is 8x faster than bge-reranker-v2-gemma while delivering higher accuracy.

Our reranking models excel at numerous specialized tasks:

  • Code and SQL Snippets: Perfect for developer docs or internal codebases
  • LLM Tool Selection: Identify the right function from thousands of definitions
  • E-Commerce: Combine product metadata to push the most relevant items to the top
  • Multilingual Applications: Support for 100+ languages enables global use cases

Why should you use our API?

To get started, you can easily use our open-source version of the models. However, the models provided through the API are trained on new data every month. This ensures that the models understand ongoing developments in the world and can identify the most relevant information for any questions they might be asked without a knowledge cutoff. Naturally, our quality control ensures that the models' performance always remains at least similar to previous versions.

Last updated on

On this page