Loading

Semantic reranking

Elastic Stack Serverless

Warning

This functionality is in technical preview and may be changed or removed in a future release. Elastic will work to fix any issues, but features in technical preview are not subject to the support SLA of official GA features.

Tip

This overview focuses more on the high-level concepts and use cases for semantic re-ranking. For full implementation details on how to set up and use semantic re-ranking in Elasticsearch, see the reference documentation in the Search API docs.

Re-rankers improve the relevance of results from earlier-stage retrieval mechanisms. Semantic re-rankers use machine learning models to reorder search results based on their semantic similarity to a query.

Semantic re-ranking requires relatively large and complex machine learning models and operates in real-time in response to queries. This technique makes sense on a small top-k result set, as one the of the final steps in a pipeline. This is a powerful technique for improving search relevance that works equally well with keyword, semantic, or hybrid retrieval algorithms.

The next sections provide more details on the benefits, use cases, and model types used for semantic re-ranking. The final sections include a practical, high-level overview of how to implement semantic re-ranking in Elasticsearch and links to the full reference documentation.

Semantic re-ranking enables a variety of use cases:

  • Lexical (BM25) retrieval results re-ranking

    • Out-of-the-box semantic search by adding a simple API call to any lexical/BM25 retrieval pipeline.
    • Adds semantic search capabilities on top of existing indices without reindexing, perfect for quick improvements.
    • Ideal for environments with complex existing indices.
  • Semantic retrieval results re-ranking

    • Improves results from semantic retrievers using ELSER sparse vector embeddings or dense vector embeddings by using more powerful models.
    • Adds a refinement layer on top of hybrid retrieval with reciprocal rank fusion (RRF).
  • General applications

    • Supports automatic and transparent chunking, eliminating the need for pre-chunking at index time.
    • Provides explicit control over document relevance in retrieval-augmented generation (RAG) uses cases or other scenarios involving language model (LLM) inputs.

Now that we’ve outlined the value of semantic re-ranking, we’ll explore the specific models that power this process and how they differ.

At a high level, two model types are used for semantic re-ranking: cross-encoders and bi-encoders.

Note

In this version, Elasticsearch only supports cross-encoders for semantic re-ranking.

  • A cross-encoder model can be thought of as a more powerful, all-in-one solution, because it generates query-aware document representations. It takes the query and document texts as a single, concatenated input.

  • A bi-encoder model takes as input either document or query text. Documents and query embeddings are computed separately, so they aren’t aware of each other.

    • To compute a ranking score, an external operation is required. This typically involves computing dot-product or cosine similarity between the query and document embeddings.

In brief, cross-encoders provide high accuracy but are more resource-intensive. Bi-encoders are faster and more cost-effective but less precise.

In future versions, Elasticsearch will also support bi-encoders. If you’re interested in a more detailed analysis of the practical differences between cross-encoders and bi-encoders, untoggle the next section.

In Elasticsearch, semantic re-rankers are implemented using the Elasticsearch Inference API and a retriever.

To use semantic re-ranking in Elasticsearch, you need to:

  1. Select and configure a re-ranking model. You have the following options:

    1. Use the Elastic Rerank cross-encoder model via the inference API’s Elasticsearch service.

    2. Use the Cohere Rerank inference endpoint to create a rerank endpoint.

    3. Use the Google Vertex AI inference endpoint to create a rerank endpoint.

    4. Upload a model to Elasticsearch from Hugging Face with Eland. You’ll need to use the text_similarity NLP task type when loading the model using Eland. Then set up an Elasticsearch service inference endpoint with the rerank endpoint type.

      Refer to the Elastic NLP model reference for a list of third party text similarity models supported by Elasticsearch for semantic re-ranking.

  2. Create a rerank endpoint using the Elasticsearch Inference API. The Inference API creates an inference endpoint and configures your chosen machine learning model to perform the re-ranking task.

  3. Define a text_similarity_reranker retriever in your search request. The retriever syntax makes it simple to configure both the retrieval and re-ranking of search results in a single API call.