北京邮电大学学报

  • EI核心期刊

北京邮电大学学报

• •    

基于知识蒸馏的个性化联邦多任务学习算法

孙艳华,王子航,刘畅,王朱伟,李萌   

  1. 北京工业大学
  • 收稿日期:2024-04-20 修回日期:2024-06-03 发布日期:2024-11-22
  • 通讯作者: 孙艳华
  • 基金资助:
    国家自然科学基金;北京市自然科学基金资助项目

Personalized Federated Multi Task Learning Algorithm Based on Knowledge Distillation

  • Received:2024-04-20 Revised:2024-06-03 Published:2024-11-22

摘要: 为了应对联邦学习(Federated Learning)框架中参与者拥有异构模型和异构数据的情况,同时考虑了参与者的信任问题,本文提出了一种基于知识蒸馏的个性化联邦多任务学习(pFedMK)算法,每个客户端除了学习全局目标任务外,还需要同时学习基于本地和其他客户端信息的个性化目标任务,采用了两级蒸馏方案,即全局蒸馏和相互蒸馏并考虑了客户端信誉问题。首先,每个客户端会在公共数据集上进行训练并计算得到自己的软预测值,中心服务器会根据客户的软预测值与前一轮信誉加权,更新当前每个客户的信誉值并根据信誉值大小选出除自己外信誉值最大的k位客户构成一个联盟,通过合作博弈的夏普利值(Shapley Value)计算出合理的聚合系数,再将聚合的模型知识利用知识蒸馏方式传输到本地模型完成全局蒸馏。然后,利用相互蒸馏,以分布的点对点方法,实现了在异构模型下,每个客户端可以从其它客户端学习的目标。通过在两个数据集上进行训练仿真并与其他同类型算法对比得出pFedMK算法可以改进系统性能,提升个性化精度。

关键词: 联邦学习, 个性化, 异构, 知识蒸馏

Abstract: In order to address the situation where participants in the Federated Learning framework have heterogeneous models and data, while also considering the issue of trust among participants, this paper proposes a personalized Federated Multi-Task Knowledge Distillation (pFedMK) algorithm. Each client not only learns the global target task but also needs to learn personalized target tasks based on local and information from other clients. The algorithm employs a two-level distillation scheme, including global distillation and mutual distillation, taking into account client reputation issues. Initially, each client trains to obtain its own soft predictions on a public dataset. The central server updates the reputation value of each client based on their soft predictions and previous round's reputation weighting, selecting the k clients with the highest reputation values (excluding itself) to form an alliance. The Shapley Value from cooperative games is calculated to determine a reasonable aggregation coefficient, and the aggregated model knowledge is then transferred to local models through knowledge distillation to complete global distillation. Subsequently, through mutual distillation using a distributed point-to-point method, under a heterogeneous model architecture, each client can learn targets from other clients. By conducting training simulations on two datasets and comparing with other similar algorithms, it is concluded that the pFedMK algorithm can improve system performance and enhance personalized accuracy.

Key words: 联邦学习, 个性化, 异构, 知识蒸馏

中图分类号: