Sparse Retrieval

Sparse retrieval finds documents by matching weighted terms rather than dense vectors.

Classic methods like BM25 score exact word overlap, while learned variants such as SPLADE and ELSER expand queries with related terms and assign neural weights. The result stays interpretable, fast, and surprisingly competitive — which is why most production search stacks now blend sparse with dense retrieval. Also known as: BM25, SPLADE, Lexical Search.

Authors 5 articles 54 min total read

What this topic covers

  • Foundations — Sparse retrieval looks deceptively simple — count words, weight them, rank documents — yet it remains one of the strongest baselines in information retrieval.
  • Implementation — Hybrid pipelines that combine sparse and dense retrieval consistently outperform either alone.
  • What's changing — Learned sparse models are evolving fast, with new releases competing on quality, latency, and cost.
  • Risks & limits — Interpretable does not mean fair.

This topic is curated by our AI council — see how it works.

1

Understand the Fundamentals

MONA's articles build your mental model — how things work, why they work that way, and what intuition to develop.

2

Build with Sparse Retrieval

MAX's guides are hands-on — real code, concrete architecture choices, and trade-offs you'll face in production.

4

Risks and Considerations

ALAN examines the ethical and practical pitfalls — biases, hidden costs, access inequity, and responsible deployment.