/SeqGAN

Implementation of Sequence Generative Adversarial Nets with Policy Gradient

Primary LanguagePython

SeqGAN

Requirements:

  • Tensorflow (r0.12)
  • Cuda (7.5+)
  • nltk python package

Introduction

Apply Generative Adversarial Nets to generating sequences of discrete tokens.

The illustration of SeqGAN. Left: D is trained over the real data and the generated data by G. Right: G is trained by policy gradient where the final reward signal is provided by D and is passed back to the intermediate action value via Monte Carlo search.

The research paper SeqGAN: Sequence Generative Adversarial Nets with Policy Gradient has been accepted at the Thirty-First AAAI Conference on Artificial Intelligence (AAAI-17).

We provide example codes to repeat the synthetic data experiments with oracle evaluation mechanisms. Move to MLE_SeqGAN folder and run

python pretrain_experiment.py

will start maximum likelihood training with default parameters. In the same folder, run

python sequence_gan.py

will start SeqGAN training. After installing nltk python package, move to pg_bleu folder and run

python pg_bleu.py

will start policy gradient algorithm with BLEU score (PG-BLEU), where the final reward for MC search comes
from a predefined score function instead of a CNN classifier. Finally, move to schedule_sampling folder and run

python schedule_sampling.py

will launch SS algorithm with default parameters.

Note: this code is based on the previous work by ofirnachum. Many thanks to ofirnachum.

After running the experiments, the learning curve should be like this: