/Pytorch_Style_Swap

Unofficial Pytorch(1.0+) implementation of paper [Fast Patch-based Style Transfer of Arbitrary Style](https://arxiv.org/abs/1612.04337).

Primary LanguagePythonGNU General Public License v3.0GPL-3.0

Pytorch_Style_Swap

Unofficial Pytorch(1.0+) implementation of paper Fast Patch-based Style Transfer of Arbitrary Style.

Original torch implementation from the author can be found here.

This repository provides a pre-trained model for you to generate your own image given content image and style image. Also, you can download the training dataset or prepare your own dataset to train the model from scratch.

If you have any question, please feel free to contact me. (Language in English/Japanese/Chinese will be ok!)

Notice

I propose a structure-emphasized multimodal style transfer(SEMST), feel free to use it here.


Requirements

  • Python 3.7
  • PyTorch 1.0+
  • TorchVision
  • Pillow

Anaconda environment recommended here!

  • GPU environment (For the calculation of style-swap)

Usage


test

  1. Clone this repository

    git clone https://github.com/irasin/Pytorch_Style_Swap
    cd Pytorch_Style_Swap
  2. Prepare your content image and style image. I provide some in the content and style and you can try to use them easily. Notice that they may be too large to tranfer because the style_swap.py really consume a lot of gpu memory, So I'm not sure all the image provided can be transferred.

  3. Generate the output image. A transferred output image and a content_output_pair image and a NST_demo_like image will be generated.

    python test -c content_image_path -s style_image_path
    usage: test.py [-h] 
                   [--content CONTENT] 
                   [--style STYLE]
                   [--output_name OUTPUT_NAME] 
                   [--patch_size PATCH_SIZE]
                   [--gpu GPU] 
                   [--model_state_path MODEL_STATE_PATH]
    
    

    If output_name is not given, it will use the combination of content image name and style image name.


train

  1. Download COCO (as content dataset)and Wikiart (as style dataset) and unzip them, rename them as content and style respectively (recommended).

  2. Modify the argument in the train.py such as the path of directory, epoch, learning_rate or you can add your own training code.

  3. Train the model using gpu.

  4. python train.py
    usage: train.py [-h] 
                    [--batch_size BATCH_SIZE] 
                    [--epoch EPOCH]
                    [--patch_size PATCH_SIZE] 
                    [--gpu GPU]
                    [--learning_rate LEARNING_RATE] 
                    [--tv_weight TV_WEIGHT]
                    [--snapshot_interval SNAPSHOT_INTERVAL]
                    [--train_content_dir TRAIN_CONTENT_DIR]
                    [--train_style_dir TRAIN_STYLE_DIR]
                    [--test_content_dir TEST_CONTENT_DIR]
                    [--test_style_dir TEST_STYLE_DIR] 
                    [--save_dir SAVE_DIR]
    

Result

Some results of content image and my cat (called Sora) will be shown here.

image image image image image image

My Opinion

The style-swap is implemented by a serious of convolutional operation.I am a beginner of Pytorch, so I afraid that my implementation technique are too poor so that it caused style-swap consuming too much gpu memory. I will be very appreciated if you can improve this implementation.Feel free to give me a PR.Thanks.

Also, as you may know, Adain and WCT are more powerful than style-swap, check them if you are interested.