论文标题

评估神经机器翻译的解释方法

Evaluating Explanation Methods for Neural Machine Translation

论文作者

Li, Jierui, Liu, Lemao, Li, Huayang, Li, Guanlin, Huang, Guoping, Shi, Shuming

论文摘要

最近,许多努力都致力于解释黑盒NMT模型,但是对指标进行评估解释方法的进展很少。单词对齐错误率可以用作与人类理解相匹配的指标,但是,它无法测量与任何源词不一致的目标词的解释方法。因此,本文最初尝试从替代角度评估解释方法。为此,它提出了基于忠诚度的原则度量,该指标就NMT模型的预测行为而言。由于该指标的确切计算是棘手的,因此我们采用有效的方法作为近似值。在六项标准翻译任务上,我们根据所提出的指标对几种解释方法进行了定量评估,并在实验中揭示了这些解释方法的一些宝贵发现。

Recently many efforts have been devoted to interpreting the black-box NMT models, but little progress has been made on metrics to evaluate explanation methods. Word Alignment Error Rate can be used as such a metric that matches human understanding, however, it can not measure explanation methods on those target words that are not aligned to any source word. This paper thereby makes an initial attempt to evaluate explanation methods from an alternative viewpoint. To this end, it proposes a principled metric based on fidelity in regard to the predictive behavior of the NMT model. As the exact computation for this metric is intractable, we employ an efficient approach as its approximation. On six standard translation tasks, we quantitatively evaluate several explanation methods in terms of the proposed metric and we reveal some valuable findings for these explanation methods in our experiments.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源