论文标题
机器学习中的公平性:调查
Fairness in Machine Learning: A Survey
论文作者
论文摘要
随着机器学习技术越来越多地用于影响公民的环境中,公司和研究人员都需要确信他们对这些方法的应用将不会产生意外的社会影响,例如对性别,种族和/或残疾人的偏见。关于减轻偏见和促进公平性的方法有重要的文献,但是该地区很复杂,很难渗透到该领域的新移民。本文旨在概述不同的思想和方法,以减轻(社会)偏见并增加机器学习文献中的公平性。它将方法组织到了被广泛接受的预处理,进行内部处理和后处理方法的框架中,并将其分类为进一步的11方法领域。尽管许多文献都强调了二进制分类,但还提供了有关回归,推荐系统,无监督学习和自然语言处理的公平性的讨论,并提供了许多当前可用的开源库。本文总结了公开研究的四个困境,总结了公开挑战。
As Machine Learning technologies become increasingly used in contexts that affect citizens, companies as well as researchers need to be confident that their application of these methods will not have unexpected social implications, such as bias towards gender, ethnicity, and/or people with disabilities. There is significant literature on approaches to mitigate bias and promote fairness, yet the area is complex and hard to penetrate for newcomers to the domain. This article seeks to provide an overview of the different schools of thought and approaches to mitigating (social) biases and increase fairness in the Machine Learning literature. It organises approaches into the widely accepted framework of pre-processing, in-processing, and post-processing methods, subcategorizing into a further 11 method areas. Although much of the literature emphasizes binary classification, a discussion of fairness in regression, recommender systems, unsupervised learning, and natural language processing is also provided along with a selection of currently available open source libraries. The article concludes by summarising open challenges articulated as four dilemmas for fairness research.