Preprint | Project Page | Google Colab
Official Pytorch implementation of the preprint paper "Castle in the Sky: Dynamic Sky Replacement and Harmonization in Videos", in arXiv:2010.11800.
We propose a vision-based method for video sky replacement and harmonization, which can automatically generate realistic and dramatic sky backgrounds in videos with controllable styles. Different from previous sky editing methods that either focus on static photos or require inertial measurement units integrated in smartphones on shooting videos, our method is purely vision-based, without any requirements on the capturing devices, and can be well applied to either online or offline processing scenarios. Our method runs in real-time and is free of user interactions. We decompose this artistic creation process into a couple of proxy tasks including sky matting, motion estimation, and image blending. Experiments are conducted on videos diversely captured in the wild by handheld smartphones and dash cameras, and show high fidelity and good generalization of our method in both visual quality and lighting/motion dynamics.
In this repository, we implement the complete training/testing pipeline of our paper based on Pytorch and provide several demo videos that can be used for reproduce the results reported in our paper. With the code, you can also try on your own data by following the instructions below.
Our code is partially adapted from the project pytorch-CycleGAN-and-pix2pix, and the project Python-Video-Stab.
SkyAR by Zhengxia Zou is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.
See Requirements.txt.
- Clone this repo:
git clone https://github.com/jiupinjia/SkyAR.git
cd SkyAR
- Download the pretrained sky matting model from Google Drive, and unzip into the repo directory.
unzip checkpoints_G_coord_resnet50.zip
python skymagic.py --path ./config/config-canyon-district9ship.json
python skymagic.py --path ./config/config-annarbor-supermoon.json
If you want to try on your own data, or want a different blending style, you can config the .json files in the ./config directory. The following we give a simple example of how the parameters are defined.
{
"net_G": "coord_resnet50",
"ckptdir": "./checkpoints_G_coord_resnet50",
"input_mode": "video",
"datadir": "./test_videos/annarbor.mp4",
"skybox": "floatingcastle.jpg",
"in_size_w": 384,
"in_size_h": 384,
"out_size_w": 845,
"out_size_h": 480,
"skybox_cernter_crop": 0.5,
"auto_light_matching": false,
"relighting_factor": 0.8,
"recoloring_factor": 0.5,
"halo_effect": true,
"output_dir": "./eval_output",
"save_jpgs": false
}
Here we also provide a minimal working example of the inference runtime of our method. Check out this link and see your result on Colab.
Please note that if you want to train your own model, you need to download the complete CVPRW20-SkyOpt dataset. We only uploaded a very small part of it due to the limited space of the repository. The mini-dataset we included in this repository is only used as an example to show how the structure of the file directory is organized.
unzip datasets.zip
python train.py \
--dataset cvprw2020-ade20K-defg \
--checkpoint_dir checkpoints \
--vis_dir val_out \
--in_size 384 \
--max_num_epochs 200 \
--lr 1e-4 \
--batch_size 8 \
--net_G coord_resnet50
If you use this code for your research, please cite our paper:
@inproceedings{zou2020skyar,
title={Castle in the Sky: Dynamic Sky Replacement and Harmonization in Videos},
author={Zhengxia Zou},
year={2020},
journal={arXiv preprint arXiv:2010.11800},
}