Knowledge intensive agents

IF 14.8
AI Open Pub Date : 2026-01-01 Epub Date: 2026-02-05 DOI:10.1016/j.aiopen.2026.02.002
Zhenghao Liu , Pengcheng Huang , Zhipeng Xu , Xinze Li , Shuliang Liu , Chunyi Peng , Haidong Xin , Yukun Yan , Shuo Wang , Xu Han , Zhiyuan Liu , Maosong Sun , Yu Gu , Ge Yu
{"title":"Knowledge intensive agents","authors":"Zhenghao Liu ,&nbsp;Pengcheng Huang ,&nbsp;Zhipeng Xu ,&nbsp;Xinze Li ,&nbsp;Shuliang Liu ,&nbsp;Chunyi Peng ,&nbsp;Haidong Xin ,&nbsp;Yukun Yan ,&nbsp;Shuo Wang ,&nbsp;Xu Han ,&nbsp;Zhiyuan Liu ,&nbsp;Maosong Sun ,&nbsp;Yu Gu ,&nbsp;Ge Yu","doi":"10.1016/j.aiopen.2026.02.002","DOIUrl":null,"url":null,"abstract":"<div><div>Large Language Models (LLMs) have exhibited impressive capabilities in reasoning and language understanding. However, their reliance on memorized knowledge and tendency to generate hallucinated content limit their reliability in real-world applications. Retrieval-Augmented Generation (RAG) mitigates these issues by integrating a retrieval module that supplements LLMs with relevant external knowledge. This paradigm bridges parametric memory and explicit retrieval, offering a principled way to ground generation in factual evidence. Despite substantial progress, most prior work has focused on optimizing isolated components, either retrieval or generation, while overlooking the agentic perspective, in which LLMs act as autonomous agents capable of actively acquiring and strategically utilizing knowledge. In this perspectives paper, we argue for reinterpreting RAG as a collaborative knowledge process among agents with distinct yet complementary roles. We categorize knowledge-intensive agents into two primary roles: knowledge acquisition (e.g., routing, query reformulation) and knowledge utilization (e.g., knowledge refinement, response generation). From this viewpoint, RAG becomes a dynamic system in which knowledge is continuously transmitted, transformed, and aligned across agent roles. To fully realize this paradigm, we advocate a joint optimization framework for knowledge-intensive agents within RAG systems. This framework explicitly models the dynamics of knowledge flow in multi-agent settings, aligning knowledge supply with knowledge demand through LLM-driven data synthesis, feedback, and evaluation. By fostering adaptive and targeted knowledge exchange, the framework mitigates conflicts between parametric and retrieved knowledge, thereby enhancing both coherence and factuality. We argue that this multi-agent joint optimization paradigm improves RAG systems in scalability, reliability, and adaptability, unlocking the potential for next-generation knowledge-intensive LLMs that reason, retrieve, and collaborate across deep retrieval processes and diverse vertical domains.</div></div>","PeriodicalId":100068,"journal":{"name":"AI Open","volume":"7 ","pages":"Pages 18-44"},"PeriodicalIF":14.8000,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"AI Open","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2666651026000033","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2026/2/5 0:00:00","PubModel":"Epub","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Large Language Models (LLMs) have exhibited impressive capabilities in reasoning and language understanding. However, their reliance on memorized knowledge and tendency to generate hallucinated content limit their reliability in real-world applications. Retrieval-Augmented Generation (RAG) mitigates these issues by integrating a retrieval module that supplements LLMs with relevant external knowledge. This paradigm bridges parametric memory and explicit retrieval, offering a principled way to ground generation in factual evidence. Despite substantial progress, most prior work has focused on optimizing isolated components, either retrieval or generation, while overlooking the agentic perspective, in which LLMs act as autonomous agents capable of actively acquiring and strategically utilizing knowledge. In this perspectives paper, we argue for reinterpreting RAG as a collaborative knowledge process among agents with distinct yet complementary roles. We categorize knowledge-intensive agents into two primary roles: knowledge acquisition (e.g., routing, query reformulation) and knowledge utilization (e.g., knowledge refinement, response generation). From this viewpoint, RAG becomes a dynamic system in which knowledge is continuously transmitted, transformed, and aligned across agent roles. To fully realize this paradigm, we advocate a joint optimization framework for knowledge-intensive agents within RAG systems. This framework explicitly models the dynamics of knowledge flow in multi-agent settings, aligning knowledge supply with knowledge demand through LLM-driven data synthesis, feedback, and evaluation. By fostering adaptive and targeted knowledge exchange, the framework mitigates conflicts between parametric and retrieved knowledge, thereby enhancing both coherence and factuality. We argue that this multi-agent joint optimization paradigm improves RAG systems in scalability, reliability, and adaptability, unlocking the potential for next-generation knowledge-intensive LLMs that reason, retrieve, and collaborate across deep retrieval processes and diverse vertical domains.
知识密集型代理
大型语言模型(llm)在推理和语言理解方面表现出了令人印象深刻的能力。然而,它们对记忆知识的依赖和产生幻觉内容的倾向限制了它们在实际应用中的可靠性。检索增强生成(RAG)通过集成检索模块来缓解这些问题,该模块为llm补充了相关的外部知识。这种范式连接了参数记忆和显式检索,为事实证据的生成提供了一种原则性的方法。尽管取得了实质性的进展,但大多数先前的工作都集中在优化孤立的组件,无论是检索还是生成,而忽略了代理的角度,其中llm作为能够主动获取和战略性地利用知识的自主代理。在这篇观点论文中,我们主张将RAG重新解释为具有不同但互补角色的代理之间的协作知识过程。我们将知识密集型代理分为两个主要角色:知识获取(例如,路由,查询重新表述)和知识利用(例如,知识提炼,响应生成)。从这个角度来看,RAG成为一个动态系统,在这个系统中,知识在各个代理角色之间不断地传递、转换和对齐。为了充分实现这一范式,我们提倡为RAG系统中的知识密集型代理建立一个联合优化框架。该框架明确地为多智能体设置中的知识流动动态建模,通过llm驱动的数据合成、反馈和评估,使知识供给与知识需求保持一致。通过促进适应性和针对性的知识交换,该框架减轻了参数知识和检索知识之间的冲突,从而增强了一致性和真实性。我们认为,这种多智能体联合优化范例提高了RAG系统的可扩展性、可靠性和适应性,释放了下一代知识密集型llm的潜力,这些llm可以跨深度检索过程和不同的垂直领域进行推理、检索和协作。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
45.00
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信
小红书