Journal of Computer Applications ›› 2026, Vol. 46 ›› Issue (1): 10-20.DOI: 10.11772/j.issn.1001-9081.2025010115

• Artificial intelligence • Previous Articles     Next Articles

Personalized federated learning method based on model pre-assignment and self-distillation

Kejia ZHANG1, Zhijun FANG1,2(), Nanrun ZHOU1, Zhicai SHI3   

  1. 1.School of Electronic and Electrical Engineering,Shanghai University of Engineering Science,Shanghai 201620,China
    2.School of Computer Science and Technology,Donghua University,Shanghai 201620,China
    3.School of Information Engineering,Shanghai Zhongqiao Vocational and Technical University,Shanghai 201514,China
  • Received:2025-02-07 Revised:2025-03-14 Accepted:2025-03-19 Online:2026-01-10 Published:2026-01-10
  • Contact: Zhijun FANG
  • About author:ZHANG Kejia, born in 2001, M. S. candidate. His research interests include federated learning.
    ZHOU Nanrun, born in 1976, Ph. D., professor. His research interests include information and communication engineering, cyberspace security.
    SHI Zhicai, born in 1964, Ph. D., professor. His research interests include embedded system, computer network, information security.
  • Supported by:
    Scientific and Technological Innovation 2030-Major Project of “New Generation of Artificial Intelligence”(2020AAA0109300);Shanghai Municipal Science and Technology Commission Local College Capacity Building Project(23010501800)

基于模型预分配与自蒸馏的个性化联邦学习方法

张珂嘉1, 方志军1,2(), 周南润1, 史志才3   

  1. 1.上海工程技术大学 电子电气工程学院,上海 201620
    2.东华大学 计算机科学与技术学院,上海 201620
    3.上海中侨职业技术大学 信息工程学院,上海 201514
  • 通讯作者: 方志军
  • 作者简介:张珂嘉(2001—),男,河南新乡人,硕士研究生,主要研究方向:联邦学习
    周南润(1976—),男,江西吉安人,教授,博士,主要研究方向:信息与通信工程、网络空间安全
    史志才(1964—),男,吉林磐石人,教授,博士,主要研究方向:嵌入式系统、计算机网络、信息安全。
  • 基金资助:
    科技创新2030—“新一代人工智能”重大项目(2020AAA0109300);上海市科委地方院校能力建设项目(23010501800)

Abstract:

Federated Learning (FL) is a distributed machine learning method that utilizes distributed data for model training while ensuring data privacy. However, it performs poorly in scenarios with highly heterogeneous data distributions. Personalized Federated Learning (PFL) addresses this challenge by providing personalized models for each client. However, the previous PFL algorithms primarily focus on optimizing client local models, while ignoring optimization of server global model. Consequently, server computational resources are not utilized fully. To overcome these limitations,FedPASD, a PFL method based on model pre-assignment and self-distillation, was proposed. FedPASD was operated in both server-side and client-side aspects. On server-side, client models for the next round were pre-assigned targetedly, which not only enhanced model personalization performance, but also utilized server computational resources effectively. On client-side, models were trained hierarchically and fine-tuned using self-distillation to better adapt to local data distribution characteristics. Experimental results on three datasets, CIFAR-10,Fashion-MNIST, and CIFAR-100 of comparing FedPASD with classic algorithms such as FedCP (Federated Conditional Policy),FedPAC (Personalization with feature Alignment and classifier Collaboration), and FedALA (Federated learning with Adaptive Local Aggregation) as baselines demonstrate that FedPASD achieves higher test accuracy than those of baseline algorithms under various data heterogeneity settings. Specifically,FedPASD achieves a test accuracy improvement of 29.05 to 29.22 percentage points over traditional FL algorithms and outperforms the PFL algorithms by 1.11 to 20.99 percentage points on CIFAR-100 dataset; on CIFAR-10 dataset,FedPASD achieves a maximum accuracy of 88.54%.

Key words: Federated Learning (FL), data heterogeneity, Personalized Federated Learning (PFL), model pre-assignment, self-distillation

摘要:

联邦学习(FL)是一种分布式机器学习方法,即利用分布式数据在训练模型的同时保护数据隐私。然而,它在高度异构的数据分布情况时表现不佳。个性化联邦学习(PFL)通过为每个客户端提供个性化模型来解决这一问题。然而,以往的PFL算法主要侧重于客户端本地模型的优化,忽略了服务器端全局模型的优化,导致服务器计算资源没有得到充分利用。针对上述局限性,提出基于模型预分配(PA)与自蒸馏(SD)的PFL方法FedPASD。FedPASD从服务器端和客户端2方面入手:在服务器端,对下一轮客户端模型有针对性地预先分配,这样不仅能提高模型的个性化性能,还能有效利用服务器的计算能力;在客户端,经过分层训练,并通过模型自蒸馏微调使模型更好地适应本地数据分布的特点。在3个数据集CIFAR-10、Fashion-MNIST和CIFAR-100上,将FedPASD与FedCP (Federated Conditional Policy)、FedPAC (Personalization with feature Alignment and classifier Collaboration)和FedALA (Federated learning with Adaptive Local Aggregation)等作为基准的典型算法进行对比实验的结果表明:FedPASD在不同异构设置下的测试准确率都高于基准算法。具体而言,FedPASD在CIFAR-100数据集上,客户端数量为50,参与率为50%的实验设置中,测试准确率较传统FL算法提升了29.05~29.22个百分点,较PFL算法提升了1.11~20.99个百分点;在CIFAR-10数据集上最高可达88.54%测试准确率。

关键词: 联邦学习, 数据异构, 个性化联邦学习, 模型预分配, 自蒸馏

CLC Number: