1 Improving LLM accuracy
This chapter covers
- Large Language Models
- Limitations of LLMs
- Shortcomings of continuously finetuning a model
- Retrieval augmented generation (RAG)
- Combining structured and unstructured data to support all types of questions
Large Language Models (LLMs) have shown impressive abilities across a variety of domains, but they have significant limitations that affect their utility, particularly when tasked with generating accurate and up-to-date information. One widely adopted approach to addressing these limitations is Retrieval-Augmented Generation (RAG), a workflow that combines an LLM with an external knowledge base to deliver accurate and current responses. By pulling data from trusted sources at runtime, RAG can significantly reduce, though not completely eliminate hallucinations, one of the most persistent challenges with LLMs. In addition, RAG allows systems to seamlessly bridge general knowledge with niche, domain-specific information that may not be well-represented in the pretraining data of the model. Despite these advantages, RAG implementations have often focused solely on unstructured data, overlooking the potential of structured sources like Knowledge Graphs (KGs).