Journal of Computer Applications ›› 2022, Vol. 42 ›› Issue (2): 333-342.DOI: 10.11772/j.issn.1001-9081.2021020232

• Artificial intelligence •     Next Articles

Survey of communication overhead of federated learning

Xinyuan QIU1,2, Zecong YE1,2, Xiaolong CUI2,3(), Zhiqiang GAO2   

  1. 1.Postgraduate Brigade,Engineering University of PAP,Xi’an Shaanxi 710086,China
    2.Anti-Terrorism Command Information Engineering Research Team,Engineering University of PAP,Xi’an Shaanxi 710086,China
    3.Urumqi Campus of Engineering University of PAP,Urumqi Xinjiang 830049,China
  • Received:2021-02-09 Revised:2021-04-13 Accepted:2021-04-20 Online:2022-02-11 Published:2022-02-10
  • Contact: Xiaolong CUI
  • About author:QIU Xinyuan, born in 1999, M. S. candidate. Her research interests include federated learning, deep learning.
    YE Zecong, born in 1997, M. S. candidate. His research interests include model compressing, object detection.
    CUI Xiaolong, born in 1973, Ph. D., professor. His research interests include command information system, big data analysis.
    GAO Zhiqiang, born in 1989, Ph. D., lecturer. His research interests include federated learning, edge intelligence.
  • Supported by:
    National Natural Science Foundation of China(U1603261);Fundamental Research Funds for Engineering University of PAP(WJY202124)

联邦学习通信开销研究综述

邱鑫源1,2, 叶泽聪1,2, 崔翛龙2,3(), 高志强2   

  1. 1.武警工程大学 研究生大队, 西安 710086
    2.武警工程大学 反恐指挥信息工程研究团队, 西安 710086
    3.武警工程大学乌鲁木齐校区, 乌鲁木齐 830049
  • 通讯作者: 崔翛龙
  • 作者简介:邱鑫源(1999—),女,江西南昌人,硕士研究生,主要研究方向:联邦学习、深度学习;
    叶泽聪(1997—),男,广东东莞人,硕士研究生,主要研究方向:模型压缩、目标检测;
    崔翛龙(1973—),男,安徽长丰人,教授,博士,主要研究方向:指挥信息系统、大数据分析;
    高志强(1989—),男,河北故城人,讲师,博士,主要研究方向:联邦学习、边缘智能。
  • 基金资助:
    国家自然科学基金资助项目(U1603261);武警工程大学基础研究基金资助项目(WJY202124)

Abstract:

To solve the irreconcilable contradiction between data sharing demands and requirements of privacy protection, federated learning was proposed. As a distributed machine learning, federated learning has a large number of model parameters needed to be exchanged between the participants and the central server, resulting in higher communication overhead. At the same time, federated learning is increasingly deployed on mobile devices with limited communication bandwidth and limited power, and the limited network bandwidth and the sharply raising client amount will make the communication bottleneck worse. For the communication bottleneck problem of federated learning, the basic workflow of federated learning was analyzed at first, and then from the perspective of methodology, three mainstream types of methods based on frequency reduction of model updating, model compression and client selection respectively as well as special methods such as model partition were introduced, and a deep comparative analysis of specific optimization schemes was carried out. Finally, the development trends of federated learning communication overhead technology research were summarized and prospected.

Key words: federated learning, communication overhead, model compression, parallel computing, client selection strategy

摘要:

为了解决数据共享需求与隐私保护要求之间不可调和的矛盾,联邦学习应运而生。联邦学习作为一种分布式机器学习,其中的参与方与中央服务器之间需要不断交换大量模型参数,而这造成了较大通信开销;同时,联邦学习越来越多地部署在通信带宽有限、电量有限的移动设备上,而有限的网络带宽和激增的客户端数量会使通信瓶颈加剧。针对联邦学习的通信瓶颈问题,首先分析联邦学习的基本工作流程;然后从方法论的角度出发,详细介绍基于降低模型更新频率、模型压缩、客户端选择的三类主流方法和模型划分等特殊方法,并对具体优化方案进行深入的对比分析;最后,对联邦学习通信开销技术研究的发展趋势进行了总结和展望。

关键词: 联邦学习, 通信开销, 模型压缩, 并行计算, 客户端选择策略

CLC Number: