论文标题
沟通有效的联合学习,并具有补偿的重叠式填充
Communication-Efficient Federated Learning with Compensated Overlap-FedAvg
论文作者
论文摘要
每天通过新兴的物联网(IoT)生成数据的PB,但是由于对数据和隐私泄漏的逮捕,最终只能收集和用于机器学习的目的(ML)目的,这严重阻碍了ML的增长。为了减轻这个问题,建议通过在群集中进行数据集共享,建议由多个客户的合并数据进行联合学习。然而,联邦学习引入了大规模的通信开销,因为每个时期中的同步数据与模型的大小相同,从而导致沟通效率低。因此,提出了主要关注交流弹和数据压缩的变体方法,以减少联邦学习的交流开销。在本文中,我们提出了Ryplap-Fedavg,该框架与模型培训阶段与模型上传和下载阶段相似,以便后一个阶段可以被以前的阶段完全覆盖。与Vanilla FedAvg相比,通过层次计算策略,数据补偿机制和Nesterov加速梯度〜(nag)算法进一步开发了重叠式填充。此外,重叠 - fedavg与许多其他压缩方法是正交的,因此可以将它们一起应用在一起以最大程度地利用集群的利用率。此外,提供了理论分析,以证明所提出的重叠式 - 罚款框架的收敛性。使用多个模型和数据集的常规任务和经常性任务进行的广泛实验也表明,所提出的重叠式FEDAVG框架大大促进了联合学习过程。
Petabytes of data are generated each day by emerging Internet of Things (IoT), but only few of them can be finally collected and used for Machine Learning (ML) purposes due to the apprehension of data & privacy leakage, which seriously retarding ML's growth. To alleviate this problem, Federated learning is proposed to perform model training by multiple clients' combined data without the dataset sharing within the cluster. Nevertheless, federated learning introduces massive communication overhead as the synchronized data in each epoch is of the same size as the model, and thereby leading to a low communication efficiency. Consequently, variant methods mainly focusing on the communication rounds reduction and data compression are proposed to reduce the communication overhead of federated learning. In this paper, we propose Overlap-FedAvg, a framework that parallels the model training phase with model uploading & downloading phase, so that the latter phase can be totally covered by the former phase. Compared to vanilla FedAvg, Overlap-FedAvg is further developed with a hierarchical computing strategy, a data compensation mechanism and a nesterov accelerated gradients~(NAG) algorithm. Besides, Overlap-FedAvg is orthogonal to many other compression methods so that they can be applied together to maximize the utilization of the cluster. Furthermore, the theoretical analysis is provided to prove the convergence of the proposed Overlap-FedAvg framework. Extensive experiments on both conventional and recurrent tasks with multiple models and datasets also demonstrate that the proposed Overlap-FedAvg framework substantially boosts the federated learning process.