📄 中文摘要
当前,人工智能模型的上下文窗口不断扩大,达到数百万个标记,令人印象深刻。然而,填充这些窗口的内容却很少被关注。现在的做法是将所有信息都放入上下文中,包括文档、聊天记录和工具定义,期望模型能够识别出重要信息。然而,增大的窗口并不能解决上下文问题,反而可能使问题更加复杂。大量无结构的上下文信息比起少量精准的信息更难以提取有用信号,模型在噪声中寻找信号的难度加大,导致效果不佳。
📄 English Summary
First Principles of AI Context
The context windows of AI models are expanding dramatically, reaching millions of tokens, which is indeed impressive. However, the question of what content fills these windows is often overlooked. Currently, the approach is to dump everything into the context—documents, chat histories, and tool definitions—hoping the model will discern what is important. However, larger windows do not solve the context problem; they merely provide more room for error. A million tokens of unfocused, unstructured context is less effective than ten thousand tokens of the right context, as the model has to work harder to find the signal amidst the noise, leading to poorer performance.
Powered by Cloudflare Workers + Payload CMS + Claude 3.5
数据源: OpenAI, Google AI, DeepMind, AWS ML Blog, HuggingFace 等