论文标题

无线网络中可变位的联合学习的性能优化

Performance Optimization for Variable Bitwidth Federated Learning in Wireless Networks

论文作者

Wang, Sihua, Chen, Mingzhe, Brinton, Christopher G., Yin, Changchuan, Saad, Walid, Cui, Shuguang

论文摘要

本文考虑通过模型量化提高联邦学习(FL)的无线通信和计算效率。在提出的Bitwidth FL方案中,Edge设备将其本地FL模型参数的量化版本训练并传输到协调服务器,该版本将它们汇总到量化的全局模型中并同步设备。目的是共同确定用于本地FL模型量化的位宽度以及每次迭代中参与FL训练的设备集。我们将其作为一个优化问题,旨在最大程度地减少量化FL的训练损失,而在每卷工具采样预算和延迟要求下。但是,没有(i)对量化如何影响全局ML性能的具体理解以及(ii)服务器有效构建此过程估计的能力的具体理解很难解决。为了应对第一个挑战,我们通过分析表征有限的无线资源和诱导量化错误如何影响所提出的FL方法的性能。我们的结果量化了两个连续迭代之间FL训练损失的改善如何取决于设备的选择和量化方案以及所学模型固有的几个参数。然后,我们表明,FL训练过程可以描述为马尔可夫决策过程,并提出一种基于模型的增强学习(RL)方法,以优化迭代术的动作选择。与无模型RL相比,此基于模型的RL方法利用FL训练过程的数学表征来发现有效的设备选择和量化方案,而无需强加其他设备通信开销。仿真结果表明,提出的FL算法可以减少收敛时间。

This paper considers improving wireless communication and computation efficiency in federated learning (FL) via model quantization. In the proposed bitwidth FL scheme, edge devices train and transmit quantized versions of their local FL model parameters to a coordinating server, which aggregates them into a quantized global model and synchronizes the devices. The goal is to jointly determine the bitwidths employed for local FL model quantization and the set of devices participating in FL training at each iteration. We pose this as an optimization problem that aims to minimize the training loss of quantized FL under a per-iteration device sampling budget and delay requirement. However, the formulated problem is difficult to solve without (i) a concrete understanding of how quantization impacts global ML performance and (ii) the ability of the server to construct estimates of this process efficiently. To address the first challenge, we analytically characterize how limited wireless resources and induced quantization errors affect the performance of the proposed FL method. Our results quantify how the improvement of FL training loss between two consecutive iterations depends on the device selection and quantization scheme as well as on several parameters inherent to the model being learned. Then, we show that the FL training process can be described as a Markov decision process and propose a model-based reinforcement learning (RL) method to optimize action selection over iterations. Compared to model-free RL, this model-based RL approach leverages the derived mathematical characterization of the FL training process to discover an effective device selection and quantization scheme without imposing additional device communication overhead. Simulation results show that the proposed FL algorithm can reduce the convergence time.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源