论文标题
与文本生成的比较歧视的自我反向学习
Self-Adversarial Learning with Comparative Discrimination for Text Generation
论文作者
论文摘要
用于文本生成的常规生成对抗网络(GAN)往往会出现奖励稀疏性和模式崩溃的问题,这些问题影响了生成样品的质量和多样性。为了解决这些问题,我们提出了一种新颖的自我逆转学习(SAL)范式,以提高甘斯在文本生成中的性能。与使用二进制分类器作为鉴别器来预测样品是真实还是生成的标准gan相反,SAL采用了比较歧视器,该比较歧视器是一个成对的分类器,用于比较一对样品之间的文本质量。在培训期间,SAL奖励发电机当前生成的句子比以前生成的样品更好。这种自我改进的奖励机制使模型可以更轻松地获得信用,并避免对有限数量的真实样品崩溃,这不仅有助于减轻奖励稀疏问题,而且还降低了模式崩溃的风险。文本生成基准数据集的实验表明,我们提出的方法显着提高了质量和多样性,并且与先前的文本生成剂相比,性能更稳定。
Conventional Generative Adversarial Networks (GANs) for text generation tend to have issues of reward sparsity and mode collapse that affect the quality and diversity of generated samples. To address the issues, we propose a novel self-adversarial learning (SAL) paradigm for improving GANs' performance in text generation. In contrast to standard GANs that use a binary classifier as its discriminator to predict whether a sample is real or generated, SAL employs a comparative discriminator which is a pairwise classifier for comparing the text quality between a pair of samples. During training, SAL rewards the generator when its currently generated sentence is found to be better than its previously generated samples. This self-improvement reward mechanism allows the model to receive credits more easily and avoid collapsing towards the limited number of real samples, which not only helps alleviate the reward sparsity issue but also reduces the risk of mode collapse. Experiments on text generation benchmark datasets show that our proposed approach substantially improves both the quality and the diversity, and yields more stable performance compared to the previous GANs for text generation.