潜在思想调优:通过融合信息在潜在标记中架起上下文与推理的桥梁

📄 中文摘要

在大语言模型(LLMs)中,显式的思维链(CoT)赋予了模型强大的推理能力,但要求模型在文本标记中逐步表达每个中间步骤,这限制了模型思想在离散词汇空间中的表现。最近,连续潜在空间中的推理作为一种有前景的替代方案出现,使得推理更加稳健,计算更加灵活,超越了离散标记的限制。然而,当前的潜在范式常常面临特征崩溃和不稳定性的问题,这主要源于在反复使用隐藏状态作为输入嵌入时的分布不匹配,或依赖辅助模型时的对齐问题。为了解决这些问题,提出了潜在思想调优(LT-Tuning)框架,旨在重新定义潜在空间中的推理过程。

📄 English Summary

Latent Thoughts Tuning: Bridging Context and Reasoning with Fused Information in Latent Tokens

The research proposes Latent Thoughts Tuning (LT-Tuning), a framework designed to bridge context and reasoning in large language models (LLMs) by utilizing fused information in latent tokens. While explicit Chain-of-Thought (CoT) enhances reasoning capabilities, it confines model thoughts to discrete vocabulary, limiting flexibility. Recent advancements in reasoning within continuous latent spaces offer a promising alternative, allowing for more robust inference and computation. However, existing latent paradigms often encounter issues such as feature collapse and instability due to distribution mismatches when repeatedly using hidden states as input embeddings, as well as alignment challenges when relying on assistant models. LT-Tuning aims to address these issues by redefining the reasoning process in latent spaces, enhancing the overall performance of LLMs.

Powered by Cloudflare Workers + Payload CMS + Claude 3.5

数据源: OpenAI, Google AI, DeepMind, AWS ML Blog, HuggingFace 等