论文标题
关于外形解释的障碍
On the Tractability of SHAP Explanations
论文作者
论文摘要
Shap解释是可解释AI的流行功能贡献机制。他们使用游戏理论概念来衡量单个特征对机器学习模型预测的影响。尽管学术界和行业都有很大的兴趣,但尚不清楚是否可以有效地计算出通用机器学习模型的塑性解释。在本文中,我们确定了在三个重要设置中计算形状解释的复杂性。首先,我们考虑了完全物质化的数据分布,并表明计算形状解释的复杂性与计算模型期望值的复杂性相同。这种完整的设置通常用于简化塑造计算,但我们的结果表明,对于常用模型,例如逻辑回归,计算可能是棘手的。超越了完全造成的分布,我们表明计算形状的解释已经在非常简单的设置中很棘手:计算对天真贝叶斯分布的琐碎分类器的塑性解释。最后,我们表明,即使在经验分布上计算的形状也是#p-hard。
SHAP explanations are a popular feature-attribution mechanism for explainable AI. They use game-theoretic notions to measure the influence of individual features on the prediction of a machine learning model. Despite a lot of recent interest from both academia and industry, it is not known whether SHAP explanations of common machine learning models can be computed efficiently. In this paper, we establish the complexity of computing the SHAP explanation in three important settings. First, we consider fully-factorized data distributions, and show that the complexity of computing the SHAP explanation is the same as the complexity of computing the expected value of the model. This fully-factorized setting is often used to simplify the SHAP computation, yet our results show that the computation can be intractable for commonly used models such as logistic regression. Going beyond fully-factorized distributions, we show that computing SHAP explanations is already intractable for a very simple setting: computing SHAP explanations of trivial classifiers over naive Bayes distributions. Finally, we show that even computing SHAP over the empirical distribution is #P-hard.