Author Login Chief Editor Login Reviewer Login Editor Login Remote Office

Computer Engineering ›› 2026, Vol. 52 ›› Issue (4): 103-110. doi: 10.19678/j.issn.1000-3428.0070254

• Computational Intelligence and Pattern Recognition • Previous Articles    

Local Momentum Accelerated Based Non-IID Federated Learning Method

YIN Hengjie1,2, ZHENG Keqing1, KE Jiannan1, DONG Yunquan1   

  1. 1. School of Electronics and Information Engineering, Nanjing University of Information Science and Technology, Nanjing 210044, Jiangsu, China;
    2. Tianchang Research Institute, Nanjing University of Information Science and Technology, Chuzhou 239300, Anhui, China
  • Received:2024-08-14 Revised:2024-10-08 Published:2024-12-11

基于本地动量加速的非独立同分布联邦学习方法

尹恒杰1,2, 郑克清1, 柯建楠1, 董云泉1   

  1. 1. 南京信息工程大学电子与信息工程学院, 江苏 南京 210044;
    2. 南京信息工程大学天长研究院, 安徽 滁州 239300
  • 作者简介:尹恒杰,男,硕士研究生,主研方向为联邦学习,E-mail:2607307739@qq.com;郑克清、柯建楠,硕士研究生;董云泉,教授、博士。
  • 基金资助:
    国家自然科学基金面上项目(62071237)。

Abstract: Federated Learning (FL), a distributed machine learning technology, has achieved significant results in privacy protection. However, in practical applications, client drift phenomena occur because of the Non-Independent and Identically Distributed (Non-IID) nature of data sources, leading to slow model convergence and performance degradation. To address this issue, this study proposes a Federated Local Momentum accelerated learning (FedLM) algorithm combined with the attention mechanism. FedLM introduces a global momentum term into local model updates, utilizing the global gradient information from previous rounds to smooth the current update process and correct the divergence of parameter update directions among heterogeneous clients, thereby reducing gradient oscillations and alleviating data heterogeneity issues. The attention mechanism dynamically adjusts the weight of each client in the global model update to improve the quality of the aggregation model. Experimental results show that FedLM achieves significantly better accuracy and stability than existing federated learning algorithms such as SCAFFOLD, FedCM, and Moon in image classification tasks with different levels of data heterogeneity, model structures, and datasets.

Key words: Federated Learning(FL), data heterogeneity, attention mechanism, client drift, local momentum

摘要: 联邦学习(FL)作为一种分布式机器学习技术,在隐私保护方面取得了显著成果。然而,在实际应用中,由于数据源的非独立同分布(Non-IID)性,导致客户端漂移现象,从而引发模型收敛缓慢和性能下降问题。为此,提出一种结合注意力机制的联邦本地动量加速学习(FedLM)算法。FedLM在本地模型更新中引入全局动量项,利用前几轮的全局梯度信息来平滑当前的更新过程,修正异构客户端的参数更新方向分歧,从而减少梯度震荡,缓解数据异构性问题。注意力机制则通过动态调整各客户端在全局模型更新中的权重,以提升聚合模型的质量。实验结果表明,在不同数据异构程度、不同模型结构以及不同数据集的图像分类任务中,FedLM的准确率和稳定性均显著优于现有的SCAFFOLD、FedCM、Moon等联邦学习算法。

关键词: 联邦学习, 数据异构, 注意力机制, 客户端漂移, 本地动量

CLC Number: