论文标题
通过随机签到的隐私放大
Privacy Amplification via Random Check-Ins
论文作者
论文摘要
私有随机梯度下降(DP-SGD)构成了许多用于学习敏感数据的应用中的基本构建块。两种标准方法,通过次采样进行隐私放大以及通过改组来扩大隐私放大,允许在DP-SGD中添加较低的噪声,而不是通过Na \“ıve计划。这两种方法中的一个关键假设是,数据集中的元素可以统一地分配或在本文中均匀地分配出来,而我们可以将数据逐渐分发,而我们可以在本过程中进行分配。在联合学习(FL)中进行DP-SGD等迭代方法,其中数据在许多设备之间分布(我们的主要贡献)。需要服务器的沟通,甚至是对我们知识的知识,这是针对分布式学习框架量身定制的第一个隐私放大,并且在途中,它可能会更广泛地适用。用户少。
Differentially Private Stochastic Gradient Descent (DP-SGD) forms a fundamental building block in many applications for learning over sensitive data. Two standard approaches, privacy amplification by subsampling, and privacy amplification by shuffling, permit adding lower noise in DP-SGD than via na\"ıve schemes. A key assumption in both these approaches is that the elements in the data set can be uniformly sampled, or be uniformly permuted -- constraints that may become prohibitive when the data is processed in a decentralized or distributed fashion. In this paper, we focus on conducting iterative methods like DP-SGD in the setting of federated learning (FL) wherein the data is distributed among many devices (clients). Our main contribution is the \emph{random check-in} distributed protocol, which crucially relies only on randomized participation decisions made locally and independently by each client. It has privacy/accuracy trade-offs similar to privacy amplification by subsampling/shuffling. However, our method does not require server-initiated communication, or even knowledge of the population size. To our knowledge, this is the first privacy amplification tailored for a distributed learning framework, and it may have broader applicability beyond FL. Along the way, we extend privacy amplification by shuffling to incorporate $(ε,δ)$-DP local randomizers, and exponentially improve its guarantees. In practical regimes, this improvement allows for similar privacy and utility using data from an order of magnitude fewer users.