论文标题

关于沙普利的信用分配,可解释性

On Shapley Credit Allocation for Interpretability

论文作者

Basu, Debraj

论文摘要

我们强调在解释学习模型的决策时提出正确问题的重要性。我们讨论了Janzing等人的理论机制的自然扩展。 al。 2020,哪个回答了一个问题:“为什么我的模型预测一个人患有癌症?”为了回答一个更涉及的问题,“是什么原因导致我的模型预测一个人患有癌症?”虽然前者量化了变量对模型的直接影响,但后者也解释了间接影响,从而提供有意义的见解。我们提出了三个广泛的解释类别:观察性,特定于模型和因果关系本身具有重要意义。此外,本文通过编织不同的解释本质以及不同的措施作为沙普利对称的特征函数来量化特征相关性。除了模型的广泛期望值外,我们还讨论了统计不确定性和分散的衡量标准,作为内容丰富的候选者,以及它们为每个数据点生成解释的优点,其中一些是在此上下文中首次使用的。这些度量不仅可用于研究变量对模型输出的影响,而且对模型的预测性能有用,为此,我们提出了第一次使用的相关特征函数。

We emphasize the importance of asking the right question when interpreting the decisions of a learning model. We discuss a natural extension of the theoretical machinery from Janzing et. al. 2020, which answers the question "Why did my model predict a person has cancer?" for answering a more involved question, "What caused my model to predict a person has cancer?" While the former quantifies the direct effects of variables on the model, the latter also accounts for indirect effects, thereby providing meaningful insights wherever human beings can reason in terms of cause and effect. We propose three broad categories for interpretations: observational, model-specific and causal each of which are significant in their own right. Furthermore, this paper quantifies feature relevance by weaving different natures of interpretations together with different measures as characteristic functions for Shapley symmetrization. Besides the widely used expected value of the model, we also discuss measures of statistical uncertainty and dispersion as informative candidates, and their merits in generating explanations for each data point, some of which are used in this context for the first time. These measures are not only useful for studying the influence of variables on the model output, but also on the predictive performance of the model, and for that we propose relevant characteristic functions that are also used for the first time.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源