Retrieval-Augmented Generation
← Back to Large Language Models
Enhance LLM responses by retrieving relevant documents and injecting them into the context. Reduces hallucination, enables access to private/recent data, and provides attributable answers.
Key Properties
Architecture
- User query → embed query
- Search vector database for similar documents
- Inject retrieved documents into LLM prompt
- LLM generates answer grounded in retrieved context
Related
- Embeddings (encode documents and queries)
- Foundation Models (generate answers from context)