NextMem:面向基于大语言模型的代理的潜在事实记忆
📄 中文摘要
记忆对于基于大语言模型的代理在未来决策中保留过去观察至关重要,其中事实记忆是其基础部分。然而,现有构建事实记忆的方法面临多种限制。文本方法在上下文和索引上负担沉重,而参数化方法则遭遇灾难性遗忘和高成本。为了解决这些挑战,提出了NextMem,一个潜在事实记忆框架,利用自回归自编码器高效构建潜在记忆,同时确保准确重构。为优化效果,提出了两阶段训练过程,包括自回归重构对齐和渐进式潜在替代。
📄 English Summary
NextMem: Towards Latent Factual Memory for LLM-based Agents
Memory is essential for LLM-based agents to retain past observations for future decision-making, with factual memory being a foundational component. Existing methods for constructing factual memory have several limitations. Textual methods impose significant context and indexing burdens, while parametric approaches suffer from catastrophic forgetting and high costs. To address these challenges, NextMem is introduced as a latent factual memory framework that employs an autoregressive autoencoder to efficiently build latent memory while ensuring accurate reconstruction. A two-stage training process is proposed for better optimization, which includes autoregressive reconstruction alignment and progressive latent substitution.
Powered by Cloudflare Workers + Payload CMS + Claude 3.5
数据源: OpenAI, Google AI, DeepMind, AWS ML Blog, HuggingFace 等