PromisingData & RetrievalNo changeMarch 2026 Backfill

Strong signal and real results. Worth committing a pilot to.

Contextual Retrieval

A pre-processing step that reduces retrieval failures by up to 67% when combined with hybrid search and reranking — at roughly $1.02 per million document tokens with prompt caching.

RAG·Context

anthropic.com

Our Take

What It Is

Contextual Retrieval adds a pre-processing step to your RAG pipeline. Before embedding each document chunk, an LLM generates a short context snippet explaining where the chunk fits in the original document. This contextualized chunk then gets embedded and indexed. The technique was published by Anthropic in September 2024 and has become a standard practice in production RAG stacks.

Why It Matters

The numbers are compelling: contextual embeddings alone reduce retrieval failures by 35%. Add contextual BM25 and you reach 49%. Add reranking on top and you hit 67% fewer retrieval failures. These aren't marginal gains — they represent the difference between a RAG system that occasionally hallucinates and one that reliably finds the right context.

With Anthropic's prompt caching (cache reads at 10% of input cost), the contextualization costs approximately $1.02 per million document tokens. That's a one-time indexing cost for a significant, permanent improvement in retrieval quality.

Key Developments

  • Late 2025: AWS Bedrock Knowledge Bases adds native Contextual Retrieval support via Anthropic models.
  • 2025-2026: Production stacks stabilise around document parsing + semantic chunking + contextual enrichment.
  • 2025: Together AI publishes implementation guide for Contextual RAG using open-source models.

What to Watch

The signal for movement to Proven is whether other providers build contextual retrieval into their managed offerings (like AWS Bedrock did). Also watch for re-indexing cost optimisation — currently, updating a document requires re-contextualizing all its chunks, which is expensive for frequently changing corpora.

Strengths

  • 49-67% retrieval failure reduction: Validated across multiple embedding models when combined with hybrid search and reranking.
  • Model-agnostic approach: Works with any embedding model. No need to retrain or fine-tune embeddings.
  • Cost-effective with prompt caching: ~$1.02 per million tokens for one-time indexing with Anthropic's caching.
  • Composable with existing infrastructure: Adds a pre-processing step without changing vector stores or retrieval logic.

Considerations

  • Upfront indexing cost: Every chunk requires an LLM call. For large corpora, this is significant compute even with caching.
  • Tied to Anthropic's ecosystem: Designed and benchmarked with Claude. Prompt caching savings only apply on Anthropic's API.
  • Chunk strategy sensitivity: Performance depends heavily on chunk size and boundary decisions. The technique amplifies your chunking approach.
  • Not a managed service: Published technique without official SDK or reference implementation. Implementation quality varies.