RAG vs fine-tuning for enterprise knowledge
Every enterprise AI project eventually asks: should we fine-tune the model on our data, or use retrieval-augmented generation? The answer, almost always, is RAG first — fine-tune only where retrieval hits a wall. Here's the decision framework we apply.
The short version
- Use RAG when the answer needs to cite a source, the knowledge changes, or coverage must be auditable.
- Fine-tune when the behaviour (style, format, domain vocabulary, structured output) needs to change, not the facts.
- Do both when you need a model that speaks your domain fluently and grounds every claim in current documents.
Why RAG wins on knowledge
Three reasons, in order of importance:
- Freshness. Fine-tuning freezes knowledge at the training snapshot. If your policy, pricing, or product catalogue updates weekly, fine-tuned knowledge goes stale weekly. Retrieval pulls from a live index.
- Auditability. Regulators, legal teams, and security reviewers ask where an answer came from. "The model learned it" is not an acceptable answer. RAG gives you citations; fine-tuning gives you a black box. See Anthropic's work on citations for why this matters for production systems.
- Coverage control. With RAG, you decide exactly what documents are in-scope. With fine-tuning, the model's general-web knowledge bleeds in — sometimes helpfully, often not.
Where RAG struggles
Retrieval breaks down on three patterns:
- Style and format mimicry. If you need outputs that sound like your legal team's voice or fit a rigid JSON schema, pushing that into the prompt every time is slow and error-prone. Fine-tuning is cheaper at scale.
- Domain vocabulary. Highly specialised terminology (clinical, legal, industrial) where the base model doesn't know what tokens mean in context. Fine-tuning shifts the model's priors; retrieval alone won't.
- Multi-step reasoning over many documents. RAG is one-shot by default. For tasks that need structured reasoning across 20+ chunks, you're writing an agent with retrieval as a tool, not just "doing RAG."
The hybrid pattern
For most enterprise builds we ship, the architecture is:
- Retrieval layer for all facts — current policy, product data, customer history, pricing.
- Light fine-tuning or instruction-tuning for output format, style, and domain vocabulary. Often an adapter (LoRA) rather than a full fine-tune — see Hu et al., 2021.
- Evaluation harness that checks both: retrieval quality (recall@k, citation accuracy) and generation quality (format adherence, factuality against retrieved chunks).
A common anti-pattern
"Let's just fine-tune on our docs." We see this pitch often. It mostly fails, for the same reason every time: the base model was trained on trillions of tokens; a 50MB corpus of internal docs can't overwrite that prior reliably. What it can do is shift style and format. If the team pushes back with "but we want the model to know our product" — RAG is almost always the better answer, with a retrieval index over those same docs.
When to revisit
Re-evaluate the RAG-vs-fine-tune call when:
- Retrieval latency is a bottleneck (> 500ms and climbing as corpus grows).
- Output format adherence drops below 95% even with strong prompting.
- You're spending more context-window budget on retrieved chunks than on reasoning.
The right answer is rarely pure. Most production systems are 80% retrieval, 20% fine-tune, and 100% measured.
