论文标题
安全和隐私保护深度学习
Security and Privacy Preserving Deep Learning
论文作者
论文摘要
大规模收集用户数据的商业公司一直是这种趋势的主要受益者,因为深度学习技术的成功与可用于培训的数据量成正比。深度学习所需的大量数据收集出现了明显的隐私问题。用户的个人,高度敏感的数据(例如照片和语音录音)无限期地由收集它的公司保存。用户既不能删除它,也不能限制其使用的目的。因此,如今,数据隐私一直是政府和公司的一个非常重要的问题。这引起了一个非常有趣的挑战,因为一方面,我们正在越来越多地推动高质量的模型和可访问的数据,但另一方面,我们需要使数据免受故意和意外泄漏的影响。数据越私密的是,它的限制越多,这意味着一些最重要的社会问题无法使用机器学习解决,因为研究人员无法访问适当的培训数据。但是,通过学习如何保护隐私的机器学习,我们可以在解决许多社会问题(例如治愈疾病等)方面产生巨大的影响。深层神经网络在记住有关培训数据的信息时易受各种推理攻击的影响。在本章中,我们介绍了差异隐私,该隐私确保了各种统计分析不会损害隐私和联合学习,请培训机器学习模型对我们无法访问的数据进行培训。
Commercial companies that collect user data on a large scale have been the main beneficiaries of this trend since the success of deep learning techniques is directly proportional to the amount of data available for training. Massive data collection required for deep learning presents obvious privacy issues. Users personal, highly sensitive data such as photos and voice recordings are kept indefinitely by the companies that collect it. Users can neither delete it nor restrict the purposes for which it is used. So, data privacy has been a very important concern for governments and companies these days. It gives rise to a very interesting challenge since on the one hand, we are pushing further and further for high-quality models and accessible data, but on the other hand, we need to keep data safe from both intentional and accidental leakage. The more personal the data is it is more restricted it means some of the most important social issues cannot be addressed using machine learning because researchers do not have access to proper training data. But by learning how to machine learning that protects privacy we can make a huge difference in solving many social issues like curing disease etc. Deep neural networks are susceptible to various inference attacks as they remember information about their training data. In this chapter, we introduce differential privacy, which ensures that different kinds of statistical analyses dont compromise privacy and federated learning, training a machine learning model on a data to which we do not have access to.