论文标题
事实检查没有足够的证据
Fact Checking with Insufficient Evidence
论文作者
论文摘要
自动化事实检查(FC)流程取决于从外部来源获得的信息。在这项工作中,我们认为,只有在有足够的证据时才进行真实性预测至关重要。为此,我们是第一个通过引入新任务并用三个主要贡献来推进它来研究哪些信息FC模型对哪些信息进行充分考虑的人。首先,我们使用一种新的保留流利度的方法对任务进行了深入的经验分析,以省略组成和句子级别的证据中的信息。我们确定模型何时基于三个具有不同变压器体系结构和三个FC数据集的训练有素的模型来考虑FC的剩余证据(in)。其次,我们问注释者,省略的证据对于FC是否很重要,导致了一个新颖的诊断数据集,足够的足够,用于省略证据。我们发现,当省略副词修饰符时,模型在检测缺失的证据方面最不成功(准确性21%),而对于省略的日期修饰符(63%的精度)最容易。最后,我们提出了一种新型的数据增强策略,用于通过采用建议的遗漏方法与三训练相结合,以对对比的缺失证据进行对比。它可以提高证据充分预测的性能,最高17.8 F1得分,进而将FC的性能提高了2.6 F1分数。
Automating the fact checking (FC) process relies on information obtained from external sources. In this work, we posit that it is crucial for FC models to make veracity predictions only when there is sufficient evidence and otherwise indicate when it is not enough. To this end, we are the first to study what information FC models consider sufficient by introducing a novel task and advancing it with three main contributions. First, we conduct an in-depth empirical analysis of the task with a new fluency-preserving method for omitting information from the evidence at the constituent and sentence level. We identify when models consider the remaining evidence (in)sufficient for FC, based on three trained models with different Transformer architectures and three FC datasets. Second, we ask annotators whether the omitted evidence was important for FC, resulting in a novel diagnostic dataset, SufficientFacts, for FC with omitted evidence. We find that models are least successful in detecting missing evidence when adverbial modifiers are omitted (21% accuracy), whereas it is easiest for omitted date modifiers (63% accuracy). Finally, we propose a novel data augmentation strategy for contrastive self-learning of missing evidence by employing the proposed omission method combined with tri-training. It improves performance for Evidence Sufficiency Prediction by up to 17.8 F1 score, which in turn improves FC performance by up to 2.6 F1 score.