Large Language Models (LLMs) have revolutionized natural language processing tasks, yet they often suffer from "hallucination” the confident generation of factually incorrect information. Retrieval-Augmented Generation (RAG) has emerged as a promising technique to mitigate hallucinations by grounding model responses in external documents. This article explores the underlying causes of hallucinations in LLMs, the mechanisms and architectures of RAG systems, their effectiveness in reducing hallucinations, and ongoing challenges. We conclude with a discussion of future directions for integrating retrieval mechanisms more seamlessly into generative architecture.
Zhangyin FengXiaocheng FengDezhi ZhaoMaojin YangBing Qin
Dena Abu LailaQais Al-Na'amnehMahmoud AljawarnehWalid DhifallahKhaled Saleh Maabreh
Jiawei ChenHongyu LinXianpei HanLe Sun