论文标题
FedClassavg:针对异质神经网络的个性化联合学习的本地代表性学习
FedClassAvg: Local Representation Learning for Personalized Federated Learning on Heterogeneous Neural Networks
论文作者
论文摘要
个性化的联合学习旨在允许众多客户在不交流私人数据的情况下以沟通效率的方式参加协作培训的同时培训个性化模型。但是,许多个性化的联合学习算法都认为客户具有相同的神经网络体系结构,而用于异质模型的神经网络架构仍在研究中。在这项研究中,我们提出了一种新型的个性化联合学习方法,称为联邦分类器平均(FedClassavg)。用于监督学习任务的深层神经网络由提取器和分类器层组成。 FedClassavg汇总分类器权重作为特征空间的决策边界的协议,以便没有独立且分布相同(非IID)数据的客户可以了解稀缺标签。此外,还采用本地功能表示学习来稳定决策边界并提高客户的本地功能提取功能。尽管现有方法需要收集辅助数据或模型权重以生成同行,但FedClassavg仅要求客户与几个完全连接的层进行通信,这是高度沟通效率的。此外,FedClassavg不需要额外的优化问题,例如知识转移,这需要大量的计算开销。我们通过广泛的实验评估了FedClassavg,并证明它的表现优于异质个性化联合学习任务的当前最新算法。
Personalized federated learning is aimed at allowing numerous clients to train personalized models while participating in collaborative training in a communication-efficient manner without exchanging private data. However, many personalized federated learning algorithms assume that clients have the same neural network architecture, and those for heterogeneous models remain understudied. In this study, we propose a novel personalized federated learning method called federated classifier averaging (FedClassAvg). Deep neural networks for supervised learning tasks consist of feature extractor and classifier layers. FedClassAvg aggregates classifier weights as an agreement on decision boundaries on feature spaces so that clients with not independently and identically distributed (non-iid) data can learn about scarce labels. In addition, local feature representation learning is applied to stabilize the decision boundaries and improve the local feature extraction capabilities for clients. While the existing methods require the collection of auxiliary data or model weights to generate a counterpart, FedClassAvg only requires clients to communicate with a couple of fully connected layers, which is highly communication-efficient. Moreover, FedClassAvg does not require extra optimization problems such as knowledge transfer, which requires intensive computation overhead. We evaluated FedClassAvg through extensive experiments and demonstrated it outperforms the current state-of-the-art algorithms on heterogeneous personalized federated learning tasks.