论文标题
UVEQFED:联合学习的通用矢量量化
UVeQFed: Universal Vector Quantization for Federated Learning
论文作者
论文摘要
传统的深度学习模型是使用从最终设备或用户收集的标签数据样本在集中式服务器上培训的。这些数据样本通常包括私人信息,用户可能不愿意分享这些信息。联合学习(FL)是一种新兴的方法,可以培训此类学习模型,而无需用户共享其可能的私人标签数据。在FL中,每个用户在本地训练其学习模型的副本。然后,服务器收集单个更新并将它们汇总到全局模型中。在这种方法中出现的一个主要挑战是每个用户需要在吞吐量有限的上行链路通道上有效传输其学习的模型。在这项工作中,我们使用量化理论中的工具来应对这一挑战。特别是,我们确定了与对速率受限的通道传达经过训练的模型相关的独特特征,并为此类设置提出了适当的量化方案,称为FL(UVEQFED)的通用向量量化。我们表明,将通用矢量量化方法与FL相结合会产生一个分散的训练系统,在该系统中,受过训练的模型的压缩仅诱导最小失真。然后,我们理论上分析了失真,表明它随着用户数量的增长而消失。我们还表征了使用传统联邦平均方法与UVEQFED训练的模型的收敛性,该模型与模型相结合,从而最小化损失函数。我们的数值结果表明,根据量化的量和准确性的失真和所得聚合模型的准确性,uveqfed在先前提出的方法上的收益。
Traditional deep learning models are trained at a centralized server using labeled data samples collected from end devices or users. Such data samples often include private information, which the users may not be willing to share. Federated learning (FL) is an emerging approach to train such learning models without requiring the users to share their possibly private labeled data. In FL, each user trains its copy of the learning model locally. The server then collects the individual updates and aggregates them into a global model. A major challenge that arises in this method is the need of each user to efficiently transmit its learned model over the throughput limited uplink channel. In this work, we tackle this challenge using tools from quantization theory. In particular, we identify the unique characteristics associated with conveying trained models over rate-constrained channels, and propose a suitable quantization scheme for such settings, referred to as universal vector quantization for FL (UVeQFed). We show that combining universal vector quantization methods with FL yields a decentralized training system in which the compression of the trained models induces only a minimum distortion. We then theoretically analyze the distortion, showing that it vanishes as the number of users grows. We also characterize the convergence of models trained with the traditional federated averaging method combined with UVeQFed to the model which minimizes the loss function. Our numerical results demonstrate the gains of UVeQFed over previously proposed methods in terms of both distortion induced in quantization and accuracy of the resulting aggregated model.