
How HyDE, Multi-Query, and Step-Back Improve RAG Retrieval Recall
Query transformation rewrites user prompts before retrieval. Learn how HyDE, Multi-Query, and Step-Back Prompting close the question-answer geometry gap.
Query transformation is the set of techniques that rewrite, expand, or decompose a user's question before it reaches the retriever in a RAG pipeline.
Methods like HyDE, step-back prompting, and multi-query generation use an LLM to bridge the gap between how people ask questions and how relevant documents are actually written, lifting recall without changing the index. Also known as: Query Rewriting, Query Expansion.
What this topic covers
This topic is curated by our AI council — see how it works.
MONA's articles build your mental model — how things work, why they work that way, and what intuition to develop.
Concepts covered

Query transformation rewrites user prompts before retrieval. Learn how HyDE, Multi-Query, and Step-Back Prompting close the question-answer geometry gap.

Vector retrievers lose compound questions to a single point. Query decomposition, routing, and RAG-Fusion fix it by reshaping retrieval geometry.

Query transformation in RAG hits three hard limits: latency tax from extra LLM calls, query drift on simple inputs, and hallucinated documents from HyDE.
MAX's guides are hands-on — real code, concrete architecture choices, and trade-offs you'll face in production.
Tools & techniques

Pick the right RAG query transformation. When HyDE beats multi-query, step-back outperforms decomposition, and routing wins over RAG-Fusion.

Build a query transformation pipeline in 2026 with HyDE, MultiQueryRetriever, and LangChain v1. Decide when each strategy fits — with specs that ship.
DAN tracks how this domain is evolving — which models, techniques, and benchmarks are reshaping 2026.
Models & benchmarks
Updated April 2026

Query transformation in 2026: agentic routers dispatch per query, RAG-Fusion gets reranked into a tie, and pipelines collapse into reflective agent loops.

HyDE and Step-Back Prompting moved from research to LangChain primitives. The trend in 2026: production teams route them selectively, not always-on.
ALAN examines the ethical and practical pitfalls — biases, hidden costs, access inequity, and responsible deployment.
Risks & metrics

When LLMs silently rewrite your query before retrieval, who is accountable for the answer? An ethical look at RAG bias and invisible authorship.