AgentArk:将多智能体智能蒸馏到单个大型语言模型智能体中

📄 中文摘要

多智能体系统在复杂任务中展现出卓越能力,但其高昂的计算成本和推理延迟限制了实际应用。AgentArk提出了一种新颖的蒸馏框架,旨在将多智能体系统的集体智能压缩到单个大型语言模型(LLM)智能体中。该框架包含三个核心模块:首先,通过多智能体协作生成高质量的轨迹数据,作为蒸馏的教师信号;其次,设计了多视角蒸馏策略,包括行为克隆、思维链蒸馏和规划蒸馏,以全面捕捉多智能体的决策过程和推理能力;最后,引入了自适应蒸馏机制,根据学生智能体的学习进度动态调整蒸馏权重。实验结果表明,AgentArk显著提升了单一LLM智能体在多个复杂任务上的性能,使其在保持高效率的同时,逼近甚至超越了原始多智能体系统的表现。AgentArk为构建高效、强大的LLM智能体提供了新途径,有望推动其在资源受限环境下的广泛应用。

📄 English Summary

AgentArk: Distilling Multi-Agent Intelligence into a Single LLM Agent

Multi-agent systems demonstrate remarkable capabilities in complex tasks, yet their high computational cost and inference latency hinder practical deployment. AgentArk proposes a novel distillation framework to compress the collective intelligence of multi-agent systems into a single Large Language Model (LLM) agent. This framework comprises three core modules: First, it generates high-quality trajectory data through multi-agent collaboration, serving as the teacher signal for distillation. Second, it designs a multi-faceted distillation strategy, encompassing behavior cloning, chain-of-thought distillation, and planning distillation, to comprehensively capture the multi-agent decision-making processes and reasoning abilities. Finally, an adaptive distillation mechanism is introduced to dynamically adjust distillation weights based on the student agent's learning progress. Experimental results show that AgentArk significantly enhances the performance of a single LLM agent across various complex tasks, enabling it to approach or even surpass the original multi-agent system's performance while maintaining high efficiency. AgentArk offers a new avenue for constructing efficient and powerful LLM agents, promising to promote their widespread application in resource-constrained environments.

Powered by Cloudflare Workers + Payload CMS + Claude 3.5

数据源: OpenAI, Google AI, DeepMind, AWS ML Blog, HuggingFace 等