FedEMA-Distill:基于指数移动平均的知识蒸馏方法以增强联邦学习的鲁棒性
📄 中文摘要
联邦学习(FL)在客户端数据异构且非独立同分布(non-IID)时,尤其是当某些客户端表现出对抗性行为时,常常会出现性能下降,导致客户端漂移、收敛速度缓慢以及高通信开销。提出了一种名为FedEMA-Distill的服务器端程序,该程序结合了全局模型的指数移动平均(EMA)和来自客户端上传的预测logits的集成知识蒸馏,这些logits是在一个小型公共代理数据集上评估的。客户端进行标准的本地训练,仅上传压缩后的logits,并且可以使用不同的模型架构,因此无需对客户端软件进行更改,同时支持设备间的模型异构性。在CIFAR-10和CIFAR-100数据集上的实验表明,该方法有效提高了联邦学习的鲁棒性和效率。
📄 English Summary
FedEMA-Distill: Exponential Moving Average Guided Knowledge Distillation for Robust Federated Learning
Federated learning (FL) often suffers from performance degradation when clients hold heterogeneous non-Independent and Identically Distributed (non-IID) data, particularly when some clients exhibit adversarial behavior, leading to client drift, slow convergence, and high communication overhead. This study proposes FedEMA-Distill, a server-side procedure that combines an exponential moving average (EMA) of the global model with ensemble knowledge distillation from client-uploaded prediction logits evaluated on a small public proxy dataset. Clients perform standard local training, upload only compressed logits, and may utilize different model architectures, thus requiring no changes to client-side software while still supporting model heterogeneity across devices. Experiments conducted on CIFAR-10 and CIFAR-100 datasets demonstrate that this method effectively enhances the robustness and efficiency of federated learning.
Powered by Cloudflare Workers + Payload CMS + Claude 3.5
数据源: OpenAI, Google AI, DeepMind, AWS ML Blog, HuggingFace 等