论文标题

通过随机签到的隐私放大

Privacy Amplification via Random Check-Ins

论文作者

Balle, Borja, Kairouz, Peter, McMahan, H. Brendan, Thakkar, Om, Thakurta, Abhradeep

论文摘要

私有随机梯度下降(DP-SGD)构成了许多用于学习敏感数据的应用中的基本构建块。两种标准方法,通过次采样进行隐私放大以及通过改组来扩大隐私放大,允许在DP-SGD中添加较低的噪声,而不是通过Na \“ıve计划。这两种方法中的一个关键假设是,数据集中的元素可以统一地分配或在本文中均匀地分配出来,而我们可以将数据逐渐分发,而我们可以在本过程中进行分配。在联合学习(FL)中进行DP-SGD等迭代方法,其中数据在许多设备之间分布(我们的主要贡献)。需要服务器的沟通,甚至是对我们知识的知识,这是针对分布式学习框架量身定制的第一个隐私放大,并且在途中,它可能会更广泛地适用。用户少。

Differentially Private Stochastic Gradient Descent (DP-SGD) forms a fundamental building block in many applications for learning over sensitive data. Two standard approaches, privacy amplification by subsampling, and privacy amplification by shuffling, permit adding lower noise in DP-SGD than via na\"ıve schemes. A key assumption in both these approaches is that the elements in the data set can be uniformly sampled, or be uniformly permuted -- constraints that may become prohibitive when the data is processed in a decentralized or distributed fashion. In this paper, we focus on conducting iterative methods like DP-SGD in the setting of federated learning (FL) wherein the data is distributed among many devices (clients). Our main contribution is the \emph{random check-in} distributed protocol, which crucially relies only on randomized participation decisions made locally and independently by each client. It has privacy/accuracy trade-offs similar to privacy amplification by subsampling/shuffling. However, our method does not require server-initiated communication, or even knowledge of the population size. To our knowledge, this is the first privacy amplification tailored for a distributed learning framework, and it may have broader applicability beyond FL. Along the way, we extend privacy amplification by shuffling to incorporate $(ε,δ)$-DP local randomizers, and exponentially improve its guarantees. In practical regimes, this improvement allows for similar privacy and utility using data from an order of magnitude fewer users.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源