迈向可还原不确定性建模以实现可靠大型语言模型智能体
📄 中文摘要
大型语言模型(LLM)的不确定性量化(UQ)是日常LLM应用安全防护的关键组成部分。然而,尽管LLM智能体越来越多地部署在高度复杂的任务中,大多数UQ研究仍集中在单轮问答上。UQ研究需要转向具有交互式智能体的现实设置,并且需要一个新的、有原则的智能体UQ框架。提出了一种首次通用的智能体UQ公式,该公式涵盖了现有UQ设置的广泛类别。在该公式下,发现以往工作隐式地将LLM UQ视为不确定性累积过程,这种观点在开放世界中的交互式智能体面前失效。因此,提出了一种新颖的P-UQ框架,该框架将智能体的不确定性视为可还原的,而非简单累积。该框架通过引入可还原不确定性概念,为LLM智能体在复杂、开放世界环境中的可靠性提供了新的视角和解决方案,旨在提升LLM智能体在实际应用中的安全性和鲁棒性。
📄 English Summary
Towards Reducible Uncertainty Modeling for Reliable Large Language Model Agents
Uncertainty quantification (UQ) for large language models (LLMs) is a foundational element for ensuring the safety and reliability of daily LLM applications. Despite the increasing deployment of LLM agents in highly complex tasks, the majority of UQ research remains confined to single-turn question-answering scenarios. A critical shift is needed in UQ research towards realistic settings involving interactive agents, necessitating a principled framework for agent UQ. This paper introduces the first general formulation of agent UQ, which effectively subsumes a broad spectrum of existing UQ setups. Within this formulation, prior works are shown to implicitly treat LLM UQ as an uncertainty accumulation process. This perspective, however, proves inadequate and breaks down when applied to interactive agents operating in open-world environments. In contrast, a novel P-UQ framework is proposed, which conceptualizes agent uncertainty as reducible rather than merely accumulative. This framework offers a fresh perspective and innovative solutions for enhancing the reliability of LLM agents in complex, open-world contexts by introducing the concept of reducible uncertainty. The aim is to significantly improve the safety and robustness of LLM agents in practical applications, moving beyond the limitations of traditional UQ approaches.
Powered by Cloudflare Workers + Payload CMS + Claude 3.5
数据源: OpenAI, Google AI, DeepMind, AWS ML Blog, HuggingFace 等