论文标题
熟:神经网络的公平验证
Fairify: Fairness Verification of Neural Networks
论文作者
论文摘要
机器学习的公平性(ML)软件已成为最近的主要问题。尽管最近关于测试和改善公平性的研究表明了对现实世界软件的影响,但在实践中提供公平保证仍然缺乏。由于模型的复杂决策过程,ML模型的认证是具有挑战性的。在本文中,我们提出了一种基于SMT的方法,以验证神经网络(NN)模型中的个人公平属性。个人公平可确保任何两个类似的人都能得到类似的待遇,无论其受保护的属性,例如种族,性别,年龄。由于NN中的全局检查和非线性计算节点,验证这种公平性属性很难。我们提出了声音方法,以使开发人员可以进行个人公平验证。关键的想法是,当考虑输入域的较小部分时,NN中的许多神经元始终保持不活跃。因此,Pairhify利用白框在生产中访问模型,然后应用基于正式的基于分析的修剪。我们的方法采用输入分区,然后对每个分区进行修剪,以提供公平认证或反例。我们杠杆间隔算术和神经元的激活启发式,以根据需要进行修剪。我们评估了从四个不同来源收集的25个现实世界神经网络的公平化,并证明了基线和密切相关的工作的有效性,可伸缩性和性能。 Fairify还可以根据NN的域和大小配置。我们对问题的新颖表述可以通过放松和反例来回答有针对性的验证查询,这些查询具有实际的影响。
Fairness of machine learning (ML) software has become a major concern in the recent past. Although recent research on testing and improving fairness have demonstrated impact on real-world software, providing fairness guarantee in practice is still lacking. Certification of ML models is challenging because of the complex decision-making process of the models. In this paper, we proposed Fairify, an SMT-based approach to verify individual fairness property in neural network (NN) models. Individual fairness ensures that any two similar individuals get similar treatment irrespective of their protected attributes e.g., race, sex, age. Verifying this fairness property is hard because of the global checking and non-linear computation nodes in NN. We proposed sound approach to make individual fairness verification tractable for the developers. The key idea is that many neurons in the NN always remain inactive when a smaller part of the input domain is considered. So, Fairify leverages whitebox access to the models in production and then apply formal analysis based pruning. Our approach adopts input partitioning and then prunes the NN for each partition to provide fairness certification or counterexample. We leveraged interval arithmetic and activation heuristic of the neurons to perform the pruning as necessary. We evaluated Fairify on 25 real-world neural networks collected from four different sources, and demonstrated the effectiveness, scalability and performance over baseline and closely related work. Fairify is also configurable based on the domain and size of the NN. Our novel formulation of the problem can answer targeted verification queries with relaxations and counterexamples, which have practical implications.