论文标题
Hidenseek:通过服务器端修剪和签名SuperMask的联合彩票票
HideNseek: Federated Lottery Ticket via Server-side Pruning and Sign Supermask
论文作者
论文摘要
联合学习仅通过将本地模型更新传输到中央服务器来减轻分布式学习的隐私风险。但是,它面临着挑战,包括客户数据集的统计异质性以及客户设备的资源限制,这严重影响了培训性能和用户体验。先前的工作通过将个性化与包括量化和修剪在内的模型压缩方案相结合来解决这些挑战。但是,修剪是数据依赖性的,因此必须在客户端进行,这需要相当大的计算成本。此外,修剪通常会在\ {0,1 \} $中训练二进制超级携带$ \,这大大限制了模型容量,但没有计算益处。因此,培训需要高计算成本,并且需要很长时间才能收敛,而模型性能则无法获得回报。在这项工作中,我们提出了Hidenseek,该HIDENSEK在初始化时采用单次数据不足的修剪,以根据权重的突触显着性获得子网。然后,每个客户端在\ { - 1,+1 \} $中优化一个符号super-mask $ \乘以未经修复的权重,以允许更快的收敛速度与最先进的压缩率相同。来自三个数据集的经验结果表明,与最新的Hidenseek相比,Hidenseek将推断精度提高了40.6 \%,同时将通信成本和培训时间分别缩短了39.7%\%和46.8%。
Federated learning alleviates the privacy risk in distributed learning by transmitting only the local model updates to the central server. However, it faces challenges including statistical heterogeneity of clients' datasets and resource constraints of client devices, which severely impact the training performance and user experience. Prior works have tackled these challenges by combining personalization with model compression schemes including quantization and pruning. However, the pruning is data-dependent and thus must be done on the client side which requires considerable computation cost. Moreover, the pruning normally trains a binary supermask $\in \{0, 1\}$ which significantly limits the model capacity yet with no computation benefit. Consequently, the training requires high computation cost and a long time to converge while the model performance does not pay off. In this work, we propose HideNseek which employs one-shot data-agnostic pruning at initialization to get a subnetwork based on weights' synaptic saliency. Each client then optimizes a sign supermask $\in \{-1, +1\}$ multiplied by the unpruned weights to allow faster convergence with the same compression rates as state-of-the-art. Empirical results from three datasets demonstrate that compared to state-of-the-art, HideNseek improves inferences accuracies by up to 40.6\% while reducing the communication cost and training time by up to 39.7\% and 46.8\% respectively.