论文标题

元元自我进行稳健学习和弱监督

Meta Self-Refinement for Robust Learning with Weak Supervision

论文作者

Zhu, Dawei, Shen, Xiaoyu, Hedderich, Michael A., Klakow, Dietrich

论文摘要

在弱监督下培训深层神经网络(DNNS)引起了越来越多的研究注意力,因为它可以大大降低注释成本。但是,来自弱监督的标签可能很吵,DNN的高容量使它们能够轻松地过度贴合标签噪声,从而导致概括不佳。最近的方法利用自我训练来建立抗噪音的模型,在这种模型中,在弱监督下接受培训的教师用于为教学学生提供高度自信的标签。然而,源自此类框架的教师可能已经安装了大量噪音,因此具有很高的信心产生错误的伪标记,导致严重的错误传播。在这项工作中,我们提出了一种抗噪声的学习框架META自我限制(MSR),以有效地抵抗弱监督的标签噪声。我们鼓励老师不依靠接受嘈杂标签的固定老师,而是鼓励老师精炼其伪标签。在每个训练步骤中,MSR在当前的迷你批次上执行元梯度下降,以最大程度地提高学生的表现。对八个NLP基准测试的广泛实验表明,在所有设置中,MSR对标签噪声都具有鲁棒性,并且胜过最先进的方法的准确性高达11.4%,而F1得分的标签差异为9.26%。

Training deep neural networks (DNNs) under weak supervision has attracted increasing research attention as it can significantly reduce the annotation cost. However, labels from weak supervision can be noisy, and the high capacity of DNNs enables them to easily overfit the label noise, resulting in poor generalization. Recent methods leverage self-training to build noise-resistant models, in which a teacher trained under weak supervision is used to provide highly confident labels for teaching the students. Nevertheless, the teacher derived from such frameworks may have fitted a substantial amount of noise and therefore produce incorrect pseudo-labels with high confidence, leading to severe error propagation. In this work, we propose Meta Self-Refinement (MSR), a noise-resistant learning framework, to effectively combat label noise from weak supervision. Instead of relying on a fixed teacher trained with noisy labels, we encourage the teacher to refine its pseudo-labels. At each training step, MSR performs a meta gradient descent on the current mini-batch to maximize the student performance on a clean validation set. Extensive experimentation on eight NLP benchmarks demonstrates that MSR is robust against label noise in all settings and outperforms state-of-the-art methods by up to 11.4% in accuracy and 9.26% in F1 score.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源