论文标题

了解现实世界中对Android应用中深度学习模型的威胁

Understanding Real-world Threats to Deep Learning Models in Android Apps

论文作者

Deng, Zizhuang, Chen, Kai, Meng, Guozhu, Zhang, Xiaodong, Xu, Ke, Cheng, Yao

论文摘要

深度学习(DL)以其出色的表现而闻名,在许多应用程序中都广泛使用,同时也引起了对模型的各种威胁。一个主要威胁是对抗攻击。研究人员已经深入研究了这一威胁,并提出了数十种创建对抗性例子(AES)的方法。但是,大多数方法仅在有限模型和数据集(例如MNIST,CIFAR-10)上进行评估。因此,攻击现实世界DL模型的有效性并不清楚。在本文中,我们对现实世界DNN模型的对抗攻击进行了首次系统研究,并提供了名为RWM的现实世界模型数据集。特别是,我们设计了一套方法,以使当前的AE生成算法适应不同的现实世界DL模型,包括从Android应用程序中自动提取DL模型,从而在应用程序中捕获DL模型的输入和输出,从而生成AES并通过观察应用程序的执行来验证它们。对于Black-Box DL模型,我们设计了一种基于语义的方法来构建合适的数据集,并在执行基于转移的攻击时将其用于训练替代模型。在分析了从62,583个现实世界应用中收集的245个DL模型之后,我们有一个独特的机会来了解现实世界DL模型与当代AE生成算法之间的差距。令我们惊讶的是,当前的AE生成算法只能直接攻击6.53%的模型。从我们的方法中受益,成功率升级到47.35%。

Famous for its superior performance, deep learning (DL) has been popularly used within many applications, which also at the same time attracts various threats to the models. One primary threat is from adversarial attacks. Researchers have intensively studied this threat for several years and proposed dozens of approaches to create adversarial examples (AEs). But most of the approaches are only evaluated on limited models and datasets (e.g., MNIST, CIFAR-10). Thus, the effectiveness of attacking real-world DL models is not quite clear. In this paper, we perform the first systematic study of adversarial attacks on real-world DNN models and provide a real-world model dataset named RWM. Particularly, we design a suite of approaches to adapt current AE generation algorithms to the diverse real-world DL models, including automatically extracting DL models from Android apps, capturing the inputs and outputs of the DL models in apps, generating AEs and validating them by observing the apps' execution. For black-box DL models, we design a semantic-based approach to build suitable datasets and use them for training substitute models when performing transfer-based attacks. After analyzing 245 DL models collected from 62,583 real-world apps, we have a unique opportunity to understand the gap between real-world DL models and contemporary AE generation algorithms. To our surprise, the current AE generation algorithms can only directly attack 6.53% of the models. Benefiting from our approach, the success rate upgrades to 47.35%.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源