论文标题
对基于机器学习的视网膜病变分类的白盒会员资格攻击
White-box Membership Attack Against Machine Learning Based Retinopathy Classification
论文作者
论文摘要
机器学习的进步(ML)大大改善了医学成像中基于AI的诊断辅助系统。但是,基于收集特定于个人的医学数据会引起几个安全问题,尤其是在隐私方面。即使像医院这样的图像所有者在其信息系统层面上设置了严格的隐私保护规定,但对其图像进行培训的模型仍然具有披露潜力。攻击者可以访问训练有素的模型为:1)白色框:访问模型体系结构和参数; 2)黑匣子:他只能通过适当的接口用自己的输入来查询模型。现有的攻击方法包括:特征估计攻击(FEA),会员推理攻击(MIA),模型记忆攻击(MMA)和识别攻击(IA)。在这项工作中,我们将重点放在MIA上,该模型已经过训练,可以从视网膜图像中检测糖尿病性视网膜病变。糖尿病性视网膜病是一种疾病,可导致患有糖尿病患者的视力丧失和失明。 MIA是确定数据样本是否来自训练有素的ML模型的训练数据集的过程。从隐私的角度来看,在我们的用例中,将糖尿病性视网膜病变分类模型与具有处置图像的合作伙伴以及患者的识别剂一起提供,推断数据样本的成员资格状态可以帮助声明患者是否对该模型的培训做出了贡献。
The advances in machine learning (ML) have greatly improved AI-based diagnosis aid systems in medical imaging. However, being based on collecting medical data specific to individuals induces several security issues, especially in terms of privacy. Even though the owner of the images like a hospital put in place strict privacy protection provisions at the level of its information system, the model trained over his images still holds disclosure potential. The trained model may be accessible to an attacker as: 1) White-box: accessing to the model architecture and parameters; 2) Black box: where he can only query the model with his own inputs through an appropriate interface. Existing attack methods include: feature estimation attacks (FEA), membership inference attack (MIA), model memorization attack (MMA) and identification attacks (IA). In this work we focus on MIA against a model that has been trained to detect diabetic retinopathy from retinal images. Diabetic retinopathy is a condition that can cause vision loss and blindness in the people who have diabetes. MIA is the process of determining whether a data sample comes from the training data set of a trained ML model or not. From a privacy perspective in our use case where a diabetic retinopathy classification model is given to partners that have at their disposal images along with patients' identifiers, inferring the membership status of a data sample can help to state if a patient has contributed or not to the training of the model.