📄 中文摘要
随着微调的 LoRA 模块在开放预训练模型中的广泛应用,研究者们对自适应合并 LoRA 的方法产生了浓厚的兴趣,以提升模型性能。这些方法通常涉及从一个池中选择 LoRA,并根据特定任务的数据集调整合并系数。尽管自适应合并方法在某些场景中已显示出性能提升,但过去的研究并未尝试从 Hugging Face Hub 等模型库中回收“野外”找到的 LoRA。为填补这一空白,研究考虑从近 1000 个用户贡献的 LoRA 中进行回收,这些 LoRA 是基于 Llama 3.1 8B-Instruct 语言模型训练的。实证研究涵盖了一系列自适应和非自适应的合并方法,旨在评估其在实际应用中的有效性。
📄 English Summary
The Appeal and Reality of Recycling LoRAs with Adaptive Merging
The widespread availability of fine-tuned LoRA modules for open pre-trained models has sparked interest in methods for adaptively merging LoRAs to enhance performance. These methods typically involve selecting LoRAs from a pool and tuning merging coefficients based on a task-specific dataset. While adaptive merging techniques have shown improvements in certain settings, no prior work has attempted to recycle LoRAs found 'in the wild' from model repositories like the Hugging Face Hub. To address this gap, the study investigates recycling from a pool of nearly 1,000 user-contributed LoRAs trained from the Llama 3.1 8B-Instruct language model. The empirical study includes a variety of adaptive and non-adaptive merging methods to evaluate their effectiveness in practical applications.
Powered by Cloudflare Workers + Payload CMS + Claude 3.5
数据源: OpenAI, Google AI, DeepMind, AWS ML Blog, HuggingFace 等