论文标题
低资源自然语言理解的有效转移学习
Effective Transfer Learning for Low-Resource Natural Language Understanding
论文作者
论文摘要
自然语言理解(NLU)是机器对人类语言进行语义解码的任务。 NLU模型在很大程度上依赖大型培训数据来确保良好的性能。但是,大量的语言和领域的数据资源和域专家很少。当很少甚至零培训样本可用时,有必要克服数据稀缺挑战。在本论文中,我们专注于开发跨语性和跨域方法来解决低资源问题。首先,我们建议通过关注与任务相关的关键字,增强模型的鲁棒性并正式化表示形式来提高模型的跨语性能力。我们发现,仅专注于关键字,可以轻松地改善低资源语言的表示形式。其次,我们提出了跨语义适应的订单减少建模方法,并发现建模部分单词顺序而不是整个顺序可以提高模型的鲁棒性,以针对语言和任务知识之间的单词顺序差异和任务知识之间的单词差异转移到低资源的语言。第三,我们建议利用与域相关的不同级别的跨域适应性培训中数据的额外掩盖,并发现更具挑战性的预训练可以更好地解决任务知识转移中的领域差异问题。最后,我们介绍了一个粗到精细的框架,教练和跨语言和跨域解析框架x2parser。 Coach将表示过程分解为粗粒和细粒度学习,而X2Parser将层次任务结构简化为扁平化的任务结构。我们观察到,简化任务结构使表示形式对低资源语言和域更有效。
Natural language understanding (NLU) is the task of semantic decoding of human languages by machines. NLU models rely heavily on large training data to ensure good performance. However, substantial languages and domains have very few data resources and domain experts. It is necessary to overcome the data scarcity challenge, when very few or even zero training samples are available. In this thesis, we focus on developing cross-lingual and cross-domain methods to tackle the low-resource issues. First, we propose to improve the model's cross-lingual ability by focusing on the task-related keywords, enhancing the model's robustness and regularizing the representations. We find that the representations for low-resource languages can be easily and greatly improved by focusing on just the keywords. Second, we present Order-Reduced Modeling methods for the cross-lingual adaptation, and find that modeling partial word orders instead of the whole sequence can improve the robustness of the model against word order differences between languages and task knowledge transfer to low-resource languages. Third, we propose to leverage different levels of domain-related corpora and additional masking of data in the pre-training for the cross-domain adaptation, and discover that more challenging pre-training can better address the domain discrepancy issue in the task knowledge transfer. Finally, we introduce a coarse-to-fine framework, Coach, and a cross-lingual and cross-domain parsing framework, X2Parser. Coach decomposes the representation learning process into a coarse-grained and a fine-grained feature learning, and X2Parser simplifies the hierarchical task structures into flattened ones. We observe that simplifying task structures makes the representation learning more effective for low-resource languages and domains.