论文标题
在复杂的医疗保健问题上进行知识提取的可解释的多步推理
Interpretable Multi-Step Reasoning with Knowledge Extraction on Complex Healthcare Question Answering
论文作者
论文摘要
Healthcare问答援助旨在提供客户保健信息,这些信息广泛出现在网络和移动互联网中。这些问题通常需要帮助才能具有熟练的医疗背景知识以及知识的推理能力。最近,已经提出了一个涉及复杂医疗保健推理的挑战,即HeadQA数据集,其中包含授权公共医疗专业考试的多项选择问题。与大多数关注语言理解的质量检查任务不同,HeadQA不仅需要更深入的推理,不仅涉及知识提取,而且还需要具有医疗保健知识的复杂推理。对于当前的质量检查系统,这些问题是最具挑战性的问题,而最新方法的当前性能比随机猜测要好得多。为了解决这项具有挑战性的任务,我们提出了具有知识提取框架(Murke)的多步推理。拟议的框架首先将医疗保健知识作为大型语料库的支持文件提取。为了找到推理链并选择正确的答案,在选择支持文档,使用支持文档重新计算查询表示之间的穆克迭代,并使用Intailment模型为每个选择获得每个选择的需要。重新制度模块利用选定的文件来保留可解释性的证据。此外,我们正在努力充分利用现成的预培训模型。训练有素的模型具有较低的训练重量,可以通过有限的培训样品轻松适应医疗保健任务。从实验结果和消融研究中,我们的系统能够在HEADQA数据集上胜过几个强大的基线。
Healthcare question answering assistance aims to provide customer healthcare information, which widely appears in both Web and mobile Internet. The questions usually require the assistance to have proficient healthcare background knowledge as well as the reasoning ability on the knowledge. Recently a challenge involving complex healthcare reasoning, HeadQA dataset, has been proposed, which contains multiple-choice questions authorized for the public healthcare specialization exam. Unlike most other QA tasks that focus on linguistic understanding, HeadQA requires deeper reasoning involving not only knowledge extraction, but also complex reasoning with healthcare knowledge. These questions are the most challenging for current QA systems, and the current performance of the state-of-the-art method is slightly better than a random guess. In order to solve this challenging task, we present a Multi-step reasoning with Knowledge extraction framework (MurKe). The proposed framework first extracts the healthcare knowledge as supporting documents from the large corpus. In order to find the reasoning chain and choose the correct answer, MurKe iterates between selecting the supporting documents, reformulating the query representation using the supporting documents and getting entailment score for each choice using the entailment model. The reformulation module leverages selected documents for missing evidence, which maintains interpretability. Moreover, we are striving to make full use of off-the-shelf pre-trained models. With less trainable weight, the pre-trained model can easily adapt to healthcare tasks with limited training samples. From the experimental results and ablation study, our system is able to outperform several strong baselines on the HeadQA dataset.