论文标题
中级任务转移学习具有验证的模型,以了解自然语言的理解:何时以及为什么起作用?
Intermediate-Task Transfer Learning with Pretrained Models for Natural Language Understanding: When and Why Does It Work?
论文作者
论文摘要
尽管伯特(Bert)等经过验证的模型已经在自然语言理解任务中显示出很大的收益,但通过对数据富的中间任务进行进一步培训模型,可以改善其性能,然后将其仔细调整到目标任务上。但是,在何时以及为什么中级培训对给定的目标任务有益的何时以及为何尚不清楚。为了调查这一点,我们对预计的罗伯塔模型进行了一项大规模研究,其中有110个中级目标组合。我们进一步评估了所有训练有素的模型,其中25个探测任务旨在揭示驱动转移的特定技能。我们观察到,需要高级推理和推理能力的中间任务往往效果最好。我们还观察到目标任务绩效与诸如Coreference解决方案之类的高级能力密切相关。但是,我们无法观察到探测和目标任务性能之间的更详细的相关性,从而强调了在宽覆盖探测基准上进行进一步工作的必要性。我们还观察到证据表明,在训练训练期间学习的知识的忘记可能会限制我们的分析,从而强调需要在这些环境中进一步进行转移学习方法的工作。
While pretrained models such as BERT have shown large gains across natural language understanding tasks, their performance can be improved by further training the model on a data-rich intermediate task, before fine-tuning it on a target task. However, it is still poorly understood when and why intermediate-task training is beneficial for a given target task. To investigate this, we perform a large-scale study on the pretrained RoBERTa model with 110 intermediate-target task combinations. We further evaluate all trained models with 25 probing tasks meant to reveal the specific skills that drive transfer. We observe that intermediate tasks requiring high-level inference and reasoning abilities tend to work best. We also observe that target task performance is strongly correlated with higher-level abilities such as coreference resolution. However, we fail to observe more granular correlations between probing and target task performance, highlighting the need for further work on broad-coverage probing benchmarks. We also observe evidence that the forgetting of knowledge learned during pretraining may limit our analysis, highlighting the need for further work on transfer learning methods in these settings.