The Concept
Self-RAG trains models to make three decisions autonomously: (1) Assess whether they have enough information before answering — triggering retrieval only when needed. (2) Evaluate the quality of retrieved results before using them. (3) Critique their own outputs for faithfulness to the retrieved context.
Why It Matters
Traditional RAG retrieves for every query, even when the model already knows the answer. Self-RAG eliminates unnecessary retrievals, reducing latency and cost. When it does retrieve, it evaluates quality before injecting into context — preventing low-quality chunks from diluting the model’s attention.
Key insight: Self-RAG adds metacognition to retrieval. The model doesn’t just retrieve and generate — it reasons about whether retrieval is needed, whether the results are good enough, and whether its answer is faithful to the evidence.