通信学报 ›› 2023, Vol. 44 ›› Issue (5): 79-93.doi: 10.11959/j.issn.1000-436x.2023072

• 学术论文 • 上一篇    下一篇

冗余数据去除的联邦学习高效通信方法

李开菊1,2, 许强3, 王豪1,4   

  1. 1 重庆邮电大学计算机科学与技术学院,重庆 400065
    2 重庆大学计算机学院,重庆 400044
    3 香港城市大学电机工程系,香港 999077
    4 旅游多源数据感知与决策技术文化和旅游部重点实验室,重庆 400065
  • 修回日期:2023-02-04 出版日期:2023-05-25 发布日期:2023-05-01
  • 作者简介:李开菊(1992- ),女,土家族,湖北恩施人,重庆大学博士生,主要研究方向为联邦学习、隐私保护
    许强(1992- ),男,江西赣州人,博士,香港城市大学在站博士后,主要研究方向为视频安全、图像处理等
    王豪(1990- ),男,河南驻马店人,博士,重庆邮电大学副教授,主要研究方向为联邦学习、隐私保护
  • 基金资助:
    国家自然科学基金资助项目(42001398);重庆市自然科学基金资助项目(cstc2020jcyj-msxmX0635);重庆市博士后研究项目特别资助项目(2021XM3009);中国博士后基金资助项目(2021M693929)

Communication-efficient federated learning method via redundant data elimination

Kaiju LI1,2, Qiang XU3, Hao WANG1,4   

  1. 1 School of Computer Science and Technology, Chongqing University of Posts and Telecommunications, Chongqing 400065, China
    2 College of Computer Science, Chongqing University, Chongqing 400044, China
    3 Department of Electrical Engineering, City University of Hong Kong, Hong Kong 999077, China
    4 Key Laboratory of Tourism Multisource Data Perception and Decision, Ministry of Culture and Tourism, Chongqing University of Posts and Telecommunications, Chongqing 400065, China
  • Revised:2023-02-04 Online:2023-05-25 Published:2023-05-01
  • Supported by:
    The National Natural Science Foundation of China(42001398);The Natural Science Foundation of Chongqing(cstc2020jcyj-msxmX0635);Chongqing Postdoctoral Research Program Special Funding(2021XM3009);China Postdoctoral Foundation(2021M693929)

摘要:

为了应对终端设备网络带宽受限对联邦学习通信效率的影响,高效地传输本地模型更新以完成模型聚合,提出了一种冗余数据去除的联邦学习高效通信方法。该方法通过分析冗余更新参数产生的本质原因,根据联邦学习中数据非独立同分布特性和模型分布式训练特点,给出新的核心数据集敏感度和损失函数容忍度定义,提出联邦核心数据集构建算法。此外,为了适配所提取的核心数据,设计了分布式自适应模型演化机制,在每次训练迭代前动态调整训练模型的结构和大小,在减少终端与云服务器通信比特数传输的同时,保证了训练模型的准确率。仿真实验表明,与目前最优的方法相比,所提方法减少了17%的通信比特数,且只有0.5%的模型准确率降低。

关键词: 联邦学习, 通信效率, 核心数据, 模型演化, 准确率

Abstract:

To address the influence of limited network bandwidth of edge devices on the communication efficiency of federated learning, and efficiently transmit local model update to complete model aggregation, a communication-efficient federated learning method via redundant data elimination was proposed.The essential reasons for generation of redundant update parameters and according to non-IID properties and model distributed training features of FL were analyzed, a novel sensitivity and loss function tolerance definitions for coreset was given, and a novel federated coreset construction algorithm was proposed.Furthermore, to fit the extracted coreset, a novel distributed adaptive sparse network model evolution mechanism was designed to dynamically adjust the structure and the training model size before each global training iteration, which reduced the number of communication bits between edge devices and the server while also guarantees the training model accuracy.Experimental results show that the proposed method achieves 17% reduction in communication bits transmission while only 0.5% degradation in model accuracy compared with state-of-the-art method.

Key words: federated learning, communication efficiency, coreset, model evolution, accuracy

中图分类号: 

No Suggested Reading articles found!