论文标题

从任务说明中学习的鲁棒性

Robustness of Learning from Task Instructions

论文作者

Gu, Jiasheng, Zhao, Hongyu, Xu, Hanzi, Nie, Liangyu, Mei, Hongyuan, Yin, Wenpeng

论文摘要

传统的监督学习主要是在单个任务上工作的,并且需要针对大量特定任务的示例进行培训。由于准备特定于任务的示例集是昂贵的,因此这种范式严重阻碍了任务概括的发展。为了构建一个可以快速,可以轻松地概括为新任务的系统,最近已将任务指令作为新兴的监督趋势。这些说明使模型具有任务的定义,并允许模型根据说明和输入输出适当的答案。但是,任务说明通常以不同的形式表达,可以从两个线程中解释:首先,某些说明是简短的句子,是面向提示的语言模型(PLM),例如提示,而其他说明则是段落,并且是面向人类的,例如亚马逊MTURK中的段落;其次,不同的最终用户很可能用不同的文本表达式的说明来解释相同的任务。用于任务概括的强大系统应该能够处理任何新任务,而不管说明的可变性如何。 但是,在处理指令驱动的任务概括方面的系统鲁棒性仍未开发。当(i)操纵新任务的指示,(ii)从不同级别的简洁性措施中解释或(iii)时,这项工作会调查系统的鲁棒性。据我们所知,这是第一项系统地研究PLM在通过具有不同变异因素的指令监督下的鲁棒性的鲁棒性。

Traditional supervised learning mostly works on individual tasks and requires training on a large set of task-specific examples. This paradigm seriously hinders the development of task generalization since preparing a task-specific example set is costly. To build a system that can quickly and easily generalize to new tasks, task instructions have been adopted as an emerging trend of supervision recently. These instructions give the model the definition of the task and allow the model to output the appropriate answer based on the instructions and inputs. However, task instructions are often expressed in different forms, which can be interpreted from two threads: first, some instructions are short sentences and are pretrained language model (PLM) oriented, such as prompts, while other instructions are paragraphs and are human-oriented, such as those in Amazon MTurk; second, different end-users very likely explain the same task with instructions of different textual expressions. A robust system for task generalization should be able to handle any new tasks regardless of the variability of instructions. However, the system robustness in dealing with instruction-driven task generalization is still unexplored. This work investigates the system robustness when the instructions of new tasks are (i) manipulated, (ii) paraphrased, or (iii) from different levels of conciseness. To our knowledge, this is the first work that systematically studies how robust a PLM is when it is supervised by instructions with different factors of variability.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源