Retrieval-Augmented Generation (RAG)
RAG is a technique that enhances large language model responses by retrieving relevant information from external knowledge sources before generating an answer, reducing hallucinations and keeping outputs grounded in factual data.
What is Retrieval-Augmented Generation?
How RAG Works
Why RAG Matters for Business
Practical Applications
Related Terms
Explore further
FAQ
Frequently asked questions
RAG retrieves external information at query time and feeds it to the model, whereas fine-tuning modifies the model's weights by training it on new data. RAG is better for frequently changing knowledge and is more cost-effective, while fine-tuning is suited for teaching the model new behaviours or specialised language patterns.
RAG can work with virtually any text-based data source including PDFs, web pages, databases, internal wikis, emails, and structured documents. With multimodal models, RAG pipelines can also incorporate images, tables, and other non-text formats.
RAG significantly reduces hallucinations by grounding responses in retrieved data, but it does not eliminate them entirely. The model can still misinterpret retrieved content or generate unsupported inferences. Proper chunking strategies, retrieval quality checks, and guardrails are important for minimising remaining risks.
Need help implementing this?
Our team can help you apply these concepts to your business. Book a free strategy call.