Skip to yearly menu bar Skip to main content


Poster

RelGAN: Relational Generative Adversarial Networks for Text Generation

Weili Nie · Nina Narodytska · Ankit B Patel

Great Hall BC #30

Keywords: [ multiple embedded representations ] [ gumbel-softmax relaxation ] [ relational memory ] [ text generation ] [ relgan ]


Abstract:

Generative adversarial networks (GANs) have achieved great success at generating realistic images. However, the text generation still remains a challenging task for modern GAN architectures. In this work, we propose RelGAN, a new GAN architecture for text generation, consisting of three main components: a relational memory based generator for the long-distance dependency modeling, the Gumbel-Softmax relaxation for training GANs on discrete data, and multiple embedded representations in the discriminator to provide a more informative signal for the generator updates. Our experiments show that RelGAN outperforms current state-of-the-art models in terms of sample quality and diversity, and we also reveal via ablation studies that each component of RelGAN contributes critically to its performance improvements. Moreover, a key advantage of our method, that distinguishes it from other GANs, is the ability to control the trade-off between sample quality and diversity via the use of a single adjustable parameter. Finally, RelGAN is the first architecture that makes GANs with Gumbel-Softmax relaxation succeed in generating realistic text.

Live content is unavailable. Log in and register to view live content