论文标题

解释深度神经网络

Explaining Deep Neural Networks

论文作者

Camburu, Oana-Maria

论文摘要

由于它们在计算机视觉,自然语言处理和语音识别等各个领域的革命性成功,深层神经网络越来越受欢迎。但是,这些模型的决策过程通常无法解释用户。在医疗保健,金融或法律等各种领域中,了解人工智能系统做出的决定的原因至关重要。因此,最近探索了一些解释神经模型的方向。在本文中,我研究了解释深神网络的两个主要方向。第一个方向由基于特征的事后解释方法组成,即旨在解释已经训练和固定模型(事后)的方法,并提供了有关输入特征的解释,例如图像的文本和超类的标记(基于功能)。第二个方向由产生自然语言解释的自称神经模型组成,即具有内置模块的模型,该模型为模型的预测产生解释。

Deep neural networks are becoming more and more popular due to their revolutionary success in diverse areas, such as computer vision, natural language processing, and speech recognition. However, the decision-making processes of these models are generally not interpretable to users. In various domains, such as healthcare, finance, or law, it is critical to know the reasons behind a decision made by an artificial intelligence system. Therefore, several directions for explaining neural models have recently been explored. In this thesis, I investigate two major directions for explaining deep neural networks. The first direction consists of feature-based post-hoc explanatory methods, that is, methods that aim to explain an already trained and fixed model (post-hoc), and that provide explanations in terms of input features, such as tokens for text and superpixels for images (feature-based). The second direction consists of self-explanatory neural models that generate natural language explanations, that is, models that have a built-in module that generates explanations for the predictions of the model.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源