可靠性
模块化设计
水准点(测量)
计算机科学
数据科学
领域(数学分析)
语言模型
人工智能
程序设计语言
政治学
大地测量学
数学
数学分析
法学
地理
作者
Yunfan Gao,Yun Xiong,Xinyu Gao,K. Jia,Jeng‐Shyang Pan,Yuhai Bi,Danhui Yi,Jiawei Sun,Haofen Wang
出处
期刊:Cornell University - arXiv
日期:2023-01-01
被引量:22
标识
DOI:10.48550/arxiv.2312.10997
摘要
Large Language Models (LLMs) showcase impressive capabilities but encounter challenges like hallucination, outdated knowledge, and non-transparent, untraceable reasoning processes. Retrieval-Augmented Generation (RAG) has emerged as a promising solution by incorporating knowledge from external databases. This enhances the accuracy and credibility of the generation, particularly for knowledge-intensive tasks, and allows for continuous knowledge updates and integration of domain-specific information. RAG synergistically merges LLMs' intrinsic knowledge with the vast, dynamic repositories of external databases. This comprehensive review paper offers a detailed examination of the progression of RAG paradigms, encompassing the Naive RAG, the Advanced RAG, and the Modular RAG. It meticulously scrutinizes the tripartite foundation of RAG frameworks, which includes the retrieval, the generation and the augmentation techniques. The paper highlights the state-of-the-art technologies embedded in each of these critical components, providing a profound understanding of the advancements in RAG systems. Furthermore, this paper introduces up-to-date evaluation framework and benchmark. At the end, this article delineates the challenges currently faced and points out prospective avenues for research and development.
科研通智能强力驱动
Strongly Powered by AbleSci AI