📄 中文摘要
灾难性遗忘是对大型语言模型(LLMs)进行细调时面临的主要挑战,尤其是在狭窄的任务特定数据上进行细调时,常常会降低其一般知识和推理能力。提出了一种轻量级自我增强例程SA-SFT,该例程允许LLM在细调之前生成自我对话,并将生成的自我创作数据与任务数据混合,而无需修改优化或训练计划。SA-SFT在不需要外部数据或额外调优的情况下,始终有效地减轻了灾难性遗忘,同时提高了领域内的表现。在50个评估场景中,该方法保持了与原始模型相当的性能,并在40个案例中取得了最佳结果,超越了常见的基准模型。
📄 English Summary
Talking to Yourself: Defying Forgetting in Large Language Models
Catastrophic forgetting poses a significant challenge when fine-tuning large language models (LLMs) on narrow, task-specific data, often degrading their general knowledge and reasoning capabilities. This study introduces SA-SFT, a lightweight self-augmentation routine where an LLM generates self-dialogues prior to fine-tuning. The resulting self-authored data is mixed with task data without altering optimization or training schedules. SA-SFT effectively mitigates catastrophic forgetting while enhancing in-domain performance, requiring no external data or additional tuning. Across 50 evaluation scenarios, it maintains performance comparable to the original model and achieves the best results in 40 cases, outperforming common baselines.
Powered by Cloudflare Workers + Payload CMS + Claude 3.5
数据源: OpenAI, Google AI, DeepMind, AWS ML Blog, HuggingFace 等