论文标题

DUNST:半监督可控文本生成的双嘈杂自训练

DuNST: Dual Noisy Self Training for Semi-Supervised Controllable Text Generation

论文作者

Feng, Yuxi, Yi, Xiaoyuan, Wang, Xiting, Lakshmanan, Laks V. S., Xie, Xing

论文摘要

自我训练(ST)在语言理解中再次繁荣发展,通过在标记的数据不足时增强预训练的语言模型的微调。但是,将ST纳入可控制的语言生成中仍然具有挑战性。由于仅由自我生成的伪文本增强,生成模型过分强调了对以前学到的空间的开发,并遭受了受约束的概括边界的困扰。我们重新访问St,并提出了一种新颖的方法,以减轻这个问题。邓斯特(Dunst)用共享的变异自动编码器共同对文本生成和分类进行了分类,并通过两种灵活的噪声来破坏生成的伪文本以干扰空间。通过这种方式,我们的模型可以从给定标签和伪标签中构建和利用可用未标记文本的伪文本,这些文本在ST过程中逐渐完善。从理论上讲,我们证明邓斯特可以被认为是对潜在的真实文本空间的增强探索,从而保证了提高性能。对三个可控生成任务进行的实验表明,邓斯特可以显着提高控制精度,同时与几种强大的基线保持可比的产生流利度和多样性。

Self-training (ST) has prospered again in language understanding by augmenting the fine-tuning of pre-trained language models when labeled data is insufficient. However, it remains challenging to incorporate ST into attribute-controllable language generation. Augmented by only self-generated pseudo text, generation models over-emphasize exploitation of the previously learned space, suffering from a constrained generalization boundary. We revisit ST and propose a novel method, DuNST to alleviate this problem. DuNST jointly models text generation and classification with a shared Variational AutoEncoder and corrupts the generated pseudo text by two kinds of flexible noise to disturb the space. In this way, our model could construct and utilize both pseudo text from given labels and pseudo labels from available unlabeled text, which are gradually refined during the ST process. We theoretically demonstrate that DuNST can be regarded as enhancing exploration towards the potential real text space, providing a guarantee of improved performance. Experiments on three controllable generation tasks show that DuNST could significantly boost control accuracy while maintaining comparable generation fluency and diversity against several strong baselines.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源