论文标题

用于个性化解释机器学习的信息理论方法

An Information-Theoretic Approach to Personalized Explainable Machine Learning

论文作者

Jung, Alexander, Nardelli, Pedro H. J.

论文摘要

自动决策在我们的日常生活中常规使用。推荐系统决定哪些作业,电影或其他用户资料对我们来说可能很有趣。拼写检查器可以帮助我们充分利用语言。欺诈检测系统决定是否应更仔细地验证信用卡交易。这些决策系统中的许多使用机器学习方法,这些方法适合复杂模型与大量数据集。机器学习(ML)方法的成功部署到许多(关键)应用领域至关重要的取决于其解释性。确实,人类有强烈的渴望获得解释,从而解决了对经验现象的不确定性,例如从ML方法获得的预测和决策。可解释的ML具有挑战性,因为必须针对背景不同的个体用户量身定制(个性化)解释。一些用户可能已经接受了ML的大学级教育,而其他用户可能没有在线性代数上接受正式培训。线性回归很少,对于第一组而言,可能是完全可以解释的,但后者可能被视为黑框。我们为预测和用户知识提出了一个简单的概率模型。该模型允许使用信息理论研究可解释的ML。这里的解释被认为是减少预测产生的“惊喜”的任务。给定用户背景,我们通过说明和预测之间的条件互信息来量化解释的效果。

Automated decision making is used routinely throughout our everyday life. Recommender systems decide which jobs, movies, or other user profiles might be interesting to us. Spell checkers help us to make good use of language. Fraud detection systems decide if a credit card transactions should be verified more closely. Many of these decision making systems use machine learning methods that fit complex models to massive datasets. The successful deployment of machine learning (ML) methods to many (critical) application domains crucially depends on its explainability. Indeed, humans have a strong desire to get explanations that resolve the uncertainty about experienced phenomena like the predictions and decisions obtained from ML methods. Explainable ML is challenging since explanations must be tailored (personalized) to individual users with varying backgrounds. Some users might have received university-level education in ML, while other users might have no formal training in linear algebra. Linear regression with few features might be perfectly interpretable for the first group but might be considered a black-box by the latter. We propose a simple probabilistic model for the predictions and user knowledge. This model allows to study explainable ML using information theory. Explaining is here considered as the task of reducing the "surprise" incurred by a prediction. We quantify the effect of an explanation by the conditional mutual information between the explanation and prediction, given the user background.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源