论文标题
添加基于鉴别器的过滤器以改善无条件的文本生成
Adding A Filter Based on The Discriminator to Improve Unconditional Text Generation
论文作者
论文摘要
接受最大似然估计(MLE)训练的自回归语言模型(ALM)被广泛用于无条件文本生成中。由于暴露偏见,生成的文本仍然遭受低质量和多样性的困扰。从统计上讲,这是真实文本和生成文本之间的差异。一些研究表明,歧视者可以检测到这种差异。由于鉴别器比发电机可以编码更多的信息,因此鉴别器具有改善生成器的潜力。为了减轻曝光偏差,生成对抗网络(GAN)使用歧视器直接更新发电机的参数,但通过精确评估它们而失败。失败的关键原因是鉴别器输入和ALM输入之间的差异。我们通过添加一个与鉴别器相同的滤波器来提出一种新型机制。首先,鉴别器检测到差异信号并通过直接过滤(或通过学习)。然后,我们使用过滤器使用基于采样的方法拒绝一些生成的样品。因此,对原始生成分布进行修订以减少差异。实验了两个基于RNN的施舍和基于变压器的施舍。通过三个指标,我们的机制始终优于两个基准数据集的施舍和各种gans。
The autoregressive language model (ALM) trained with maximum likelihood estimation (MLE) is widely used in unconditional text generation. Due to exposure bias, the generated texts still suffer from low quality and diversity. This presents statistically as a discrepancy between the real text and generated text. Some research shows a discriminator can detect this discrepancy. Because the discriminator can encode more information than the generator, discriminator has the potentiality to improve generator. To alleviate the exposure bias, generative adversarial networks (GAN) use the discriminator to update the generator's parameters directly, but they fail by being evaluated precisely. A critical reason for the failure is the difference between the discriminator input and the ALM input. We propose a novel mechanism by adding a filter which has the same input as the discriminator. First, discriminator detects the discrepancy signals and passes to filter directly (or by learning). Then, we use the filter to reject some generated samples with a sampling-based method. Thus, the original generative distribution is revised to reduce the discrepancy. Two ALMs, RNN-based and Transformer-based, are experimented. Evaluated precisely by three metrics, our mechanism consistently outperforms the ALMs and all kinds of GANs across two benchmark data sets.