/WarpGAN

(CVPR 2019 Oral) Style Transfer with Geometric Deformation

Primary LanguagePythonMIT LicenseMIT

WarpGAN: Automatic Caricature Generation

By Yichun Shi, Debayan Deb and Anil K. Jain

A tensorflow implementation of WarpGAN, a fully automatic network that can generate caricatures given an input face photo. Besides transferring rich texture styles, WarpGAN learns to automatically predict a set of control points that can warp the photo into a caricature, while preserving identity. We introduce an identity-preserving adversarial loss that aids the discriminator to distinguish between different subjects. Moreover, WarpGAN allows customization of the generated caricatures by controlling the exaggeration extent and the visual styles.

Tensorflow release

Currently this repo is compatible with Tensorflow r1.9.

News

Date Update
2019-04-10 Testing Code
2019-04-07 Training Code
2019-04-05 Initial Code Upload

Citation

@article{warpgan,
  title = {WarpGAN: Automatic Caricature Generation},
  author = {Shi, Yichun, Deb, Debayan and Jain, Anil K.},
  booktitle = {CVPR},
  year = {2019}
}

Usage

Note: In this section, we assume that you are always in the directory $WARPGAN_ROOT/

Preprocessing

  1. Download the original images of WebCaricature dataset and unzip them into data/WebCaricature/OriginalImages. Rename the images by running
    python data/rename.py
  2. Then, normalize all the faces by running the following code:
    python align/align_dataset.py data/landmarks.txt data/webcaricacture_aligned_256 --scale 0.7
    The command will normalize all the photos and caricatures using the landmark points pre-defined in the WebCaricature protocol (we use only 5 landmarks). Notice that during deployment, we will use MTCNN to detect the face landmarks for images not in the dataset.

Training

  1. Before training, you need to download the discriminator model to initialize the parameters of the disrcimanator, which is pre-trained as an identity classifier. Unzip the files under pretrained/discriminator_casia_256/.

  2. The configuration files for training are saved under config/ folder, where you can define the dataset prefix, training list, model file and other hyper-parameters. Use the following command to run the default training configuration:

    python train.py config/default.py

    The command will create an folder under log/default/ which saves all the checkpoints, test samples and summaries. The model directory is named as the time you start training.

Testing

  • Run the test code in the following format:
    python test.py /path/to/model/dir /path/to/input/image /prefix/of/output/image
  • For example, if you want to use the pre-trained model, download the model and unzip it into pretrained/warpgan_pretrained. Then, run the following command to generate 5 images for captain marvel of different random styles:
    python test.py pretrained/warpgan_pretrained \
    data/example/CaptainMarvel.jpg \
    result/CaptainMarvel \
    --num_styles 5
  • You can also change the warping extent by using the --scale argument. For example, the following command doubles the displacement of the warpping control points:
    python test.py pretrained/warpgan_pretrained \
    data/example/CaptainMarvel.jpg \
    result/CaptainMarvel \
    --num_styles 5 --scale 2.0

Pre-trained Model

Discriminator Initializaiton:

Google Drive

WarpGAN:

Google Drive