论文标题
填补多转向对话的话语和说话者意识的代表的空白
Filling the Gap of Utterance-aware and Speaker-aware Representation for Multi-turn Dialogue
论文作者
论文摘要
多转话的对话由两个或更多不同的扬声器角色的多种话语组成。因此,在模型中应该很好地捕获了话语和说话者意识的线索。但是,在现有的基于检索的多转化对话建模中,作为编码器作为编码器的预训练的语言模型(PRLMS)代表了对话,通过采用成对对话历史记录和整个候选人响应,整体上,关于话语相互关系或扬声器角色的层次结构信息在此类表示中耦合在此类表示方面。在这项工作中,我们提出了一个新颖的模型,通过建模在对话历史中需要建模有效的话语和说话者意识的表述来填补这一空白。详细说明,我们通过掩盖了基于变压器的PRLM中的机制来使上下文化的单词表示形式,使每个单词仅关注当前话语,其他话语,两个说话者角色(即发件人的话语和接收者的话语)中的单词。实验结果表明,我们的方法在四个公共基准数据集中大大提高了强元基线,并且在以前的方法上实现了各种新的最新性能。进行了一系列消融研究,以证明我们方法的有效性。
A multi-turn dialogue is composed of multiple utterances from two or more different speaker roles. Thus utterance- and speaker-aware clues are supposed to be well captured in models. However, in the existing retrieval-based multi-turn dialogue modeling, the pre-trained language models (PrLMs) as encoder represent the dialogues coarsely by taking the pairwise dialogue history and candidate response as a whole, the hierarchical information on either utterance interrelation or speaker roles coupled in such representations is not well addressed. In this work, we propose a novel model to fill such a gap by modeling the effective utterance-aware and speaker-aware representations entailed in a dialogue history. In detail, we decouple the contextualized word representations by masking mechanisms in Transformer-based PrLM, making each word only focus on the words in current utterance, other utterances, two speaker roles (i.e., utterances of sender and utterances of receiver), respectively. Experimental results show that our method boosts the strong ELECTRA baseline substantially in four public benchmark datasets, and achieves various new state-of-the-art performance over previous methods. A series of ablation studies are conducted to demonstrate the effectiveness of our method.