DecHW: 利用二阶信息的异构去中心化联邦学习

📄 中文摘要

去中心化联邦学习(DFL)是一种无服务器的协作机器学习范式,其中设备直接与相邻设备协作交换模型信息,以学习一个泛化模型。然而,由于个体经验的差异和设备交互水平的不同,导致设备之间存在数据和模型初始化异构性。这些异构性使得本地模型参数之间存在差异,当采用基于一阶梯度的方法时,这会阻碍收敛并导致性能下降。为了解决这个问题,DecHW(Decentralized Heterogeneous Federated Learning with Second-Order Information)被提出。DecHW利用二阶信息来更有效地处理异构性。

📄 English Summary

DecHW: Heterogeneous Decentralized Federated Learning Exploiting Second-Order Information

Decentralized Federated Learning (DFL) is a serverless collaborative machine learning paradigm where devices directly collaborate with neighboring devices to exchange model information for learning a generalized model. However, variations in individual experiences and different levels of device interactions lead to data and model initialization heterogeneities across devices. Such heterogeneities result in discrepancies in local model parameters, which hinder convergence and degrade performance when first-order gradient-based methods are employed. To address this issue, DecHW (Decentralized Heterogeneous Federated Learning with Second-Order Information) is proposed. DecHW leverages second-order information to more effectively handle heterogeneity. Specifically, it introduces an adaptive learning rate adjustment mechanism based on the Hessian matrix, enabling each device to dynamically adjust its model update step size according to the curvature information of its local data distribution. This approach better captures non-linear relationships between data from different devices, leading to faster convergence and higher model accuracy in heterogeneous environments. Furthermore, DecHW introduces a novel neighbor aggregation strategy that not only considers parameter differences among neighbor models but also incorporates a weight allocation mechanism based on second-order information, allowing neighbors with higher information gain to contribute more significantly to the global model update. Experimental results demonstrate that DecHW achieves significant performance improvements compared to existing decentralized federated learning methods across various heterogeneous datasets and network topologies, particularly in terms of convergence speed and final model accuracy. This method provides an effective solution for addressing data and model heterogeneity in decentralized federated learning, opening new avenues for future research.

Powered by Cloudflare Workers + Payload CMS + Claude 3.5

数据源: OpenAI, Google AI, DeepMind, AWS ML Blog, HuggingFace 等