论文标题
尾巴的攻击:是的,您真的可以后门联邦学习
Attack of the Tails: Yes, You Really Can Backdoor Federated Learning
论文作者
论文摘要
由于其分散的性质,联邦学习(FL)在训练过程中以后门的形式以对抗性攻击。后门的目的是破坏受过特定子任务的训练模型的性能(例如,通过将绿色汽车分类为青蛙)。文献中已经引入了一系列FL后门攻击,也引入了防御它们的方法,目前,FL Systems是否可以量身定制以防止后门进行量身定制。在这项工作中,我们提供了相反的证据。我们首先确定,在一般情况下,对后门的鲁棒性意味着对对抗性示例的鲁棒性本身就是一个主要的开放问题。此外,在FL模型中检测后门的存在是不可能假设一阶orac或多项式时间的。我们将我们的理论结果与新的后门攻击家族融为一体,我们称之为边缘案例后门。边缘后门迫使模型错误地分类在看似简单的输入上,但是这些输入不太可能成为培训或测试数据的一部分,即它们生活在输入分布的尾巴上。我们解释了这些边缘案例后门如何导致不良的失败,并可能对公平产生严重的影响,并在对手的一边进行仔细的调整,可以将它们插入一系列机器学习任务(例如,图像分类,OCR,OCR,OCR,文本预测,情感分析)。
Due to its decentralized nature, Federated Learning (FL) lends itself to adversarial attacks in the form of backdoors during training. The goal of a backdoor is to corrupt the performance of the trained model on specific sub-tasks (e.g., by classifying green cars as frogs). A range of FL backdoor attacks have been introduced in the literature, but also methods to defend against them, and it is currently an open question whether FL systems can be tailored to be robust against backdoors. In this work, we provide evidence to the contrary. We first establish that, in the general case, robustness to backdoors implies model robustness to adversarial examples, a major open problem in itself. Furthermore, detecting the presence of a backdoor in a FL model is unlikely assuming first order oracles or polynomial time. We couple our theoretical results with a new family of backdoor attacks, which we refer to as edge-case backdoors. An edge-case backdoor forces a model to misclassify on seemingly easy inputs that are however unlikely to be part of the training, or test data, i.e., they live on the tail of the input distribution. We explain how these edge-case backdoors can lead to unsavory failures and may have serious repercussions on fairness, and exhibit that with careful tuning at the side of the adversary, one can insert them across a range of machine learning tasks (e.g., image classification, OCR, text prediction, sentiment analysis).