论文标题
如何评估解释性? - 三个标准的案例
How to Evaluate Explainability? -- A Case for Three Criteria
论文作者
论文摘要
软件系统的复杂性日益增加以及我们社会中软件支持的决策的影响激发了对安全,可靠和公平的软件的需求。解释性已被确定为实现这些品质的一种手段。它被认为是对系统质量产生重大影响的新兴非功能要求(NFR)。但是,为了开发可解释的系统,我们需要了解系统何时满足该NFR。为此,需要适当的评估方法。但是,该领域挤满了评估方法,并且尚无共识是“正确的”。更少,甚至不同意应评估哪些标准。在本视觉论文中,我们将为有关系统应提供的信息提供三个此类质量标准的多学科动机:可理解性,忠诚度和可评估性。我们的目的是推动有关这些标准的讨论,以便为他们提供适当的评估方法。
The increasing complexity of software systems and the influence of software-supported decisions in our society have sparked the need for software that is safe, reliable, and fair. Explainability has been identified as a means to achieve these qualities. It is recognized as an emerging non-functional requirement (NFR) that has a significant impact on system quality. However, in order to develop explainable systems, we need to understand when a system satisfies this NFR. To this end, appropriate evaluation methods are required. However, the field is crowded with evaluation methods, and there is no consensus on which are the "right" ones. Much less, there is not even agreement on which criteria should be evaluated. In this vision paper, we will provide a multidisciplinary motivation for three such quality criteria concerning the information that systems should provide: comprehensibility, fidelity, and assessability. Our aim is to to fuel the discussion regarding these criteria, such that adequate evaluation methods for them will be conceived.