Retrieval-Augmented Generation for LargeLanguage Models: A Survey

标题:Retrieval-Augmented Generation for Large Language Models: A Survey

作者:Yunfan Gaoa , Yun Xiongb , Xinyu Gaob , Kangxiang Jiab , Jinliu Panb , Yuxi Bic , Yi Daia , Jiawei Suna , Meng Wangc , and Haofen Wang

  1. By referencing external knowledge, RAG effectively reduces the problem of generating factually incorrect content. Its integration into LLMs has resulted in widespread adoption, establishing RAG as a key technology in advancing chatbots and enhancing the suitability of LLMs for real-world applications

  2. The RAG research paradigm is continuously evolving, and we categorize it into three stages: Naive RAG, Advanced RAG, and Modular RAG

  3. The Naive RAG:

Indexing starts with the cleaning and extraction of raw data

Retrieval. Upon receipt of a user query, the RAG system employs the same encoding model utilized during the indexing phase to transform the query into a vector representation.

Generation. The posed query and selected documents are synthesized into a coherent prompt to which a large language model is tasked with formulating a response.

Advanced RAG introduces specific improvements to overcome the limitations of Naive RAG. Focusing on enhancing retrieval quality, it employs pre-retrieval and post-retrieval strategies.

Pre-retrieval process. In this stage, the primary focus is on optimizing the indexing structure and the original query. The goal of optimizing indexing is to enhance the quality of the content being indexed.

Post-Retrieval Process. Once relevant context is retrieved, it's crucial to integrate it effectively with the query

  1. Innovations such as the Rewrite-Retrieve-Read [7]model leverage the LLM's capabilities to refine retrieval queries through a rewriting module and a LM-feedback mechanism to update rewriting model

  2. RAG is often compared with Fine-tuning (FT) and prompt engineering. Each method has distinct characteristics as illustrated in Figure 4.

  3. In the context of RAG, it is crucial to efficiently retrieve relevant documents from the data source. There are several key issues involved, such as the retrieval source, retrieval granularity, pre-processing of the retrieval, and selection of the corresponding embedding model.

相关推荐
结冰架构2 小时前
【AI提示词】Emoji风格排版艺术与设计哲学
大数据·人工智能·ai·提示词
wang_yb3 小时前
当决策树遇上脏数据:连续值与缺失值的解决方案
ai·databook
SmallFatMan7 小时前
智能客服系统中的意图识别与分类技术详解
大数据·人工智能·ai·数据挖掘·ai编程
YXWik614 小时前
coze生成流程图和思维导图工作流
ai·coze
CCEEHM20 小时前
不花钱也能玩GPT-4o,国内可用
ai
By北阳1 天前
2025年AI开发学习路线
人工智能·学习·ai·aigc
wang_yb1 天前
决策树剪枝:平衡模型复杂性与泛化能力
ai·databook
风尘浪子1 天前
基于大模型的 RAG 核心开发——详细介绍 DeepSeek R1 本地化部署流程
ai·rag·ai人工智能·deepseek·本地化部署
Elastic 中国社区官方博客1 天前
将 CrewAI 与 Elasticsearch 结合使用
大数据·人工智能·elasticsearch·机器学习·搜索引擎·ai·全文检索
令狐少侠20111 天前
python之安装PaddlePaddle和PaddleX解析pdf表格
python·ai·pdf·paddlepaddle