This repository aims to use LSeg (Li et al., ICLR 2022) with StyleGAN2 (Karras et al., CVPR 2020)
conda create -n lseg -y python=3.8
conda activate lseg
conda install -y cudatoolkit=11.1 -c nvidia
pip install torch==1.8.0+cu111 torchvision==0.9.0+cu111 torchaudio==0.8.0 -f https://download.pytorch.org/whl/torch_stable.html
conda install -y libxcrypt gxx_linux-64=7 cxx-compiler ninja cudatoolkit-dev -c conda-forge
pip install click requests tqdm opencv-python-headless matplotlib regex ftfy timm==0.9.16
- This environment is tested on RTX3060ti
- Download LSeg weights from Pretrained weights
- Put the downloaded weights under folder checkpoints as checkpoints/lseg.ckpt
- Note that above weights containts model weights from Originral Pretrained Weights from Official LSeg Github.
- Directly utilizing the author provided weights needs complicated environment setup.
- I just loaded the author provided weights and save the model state dict only.
- Dataset Version
-
Prepare dataset (FFHQ / LSUN-CHURCH / LSUN-BEDROOM / AFHQ-CAT / AFHQ-WILD) following instructions in official StyleGAN2-ADA github repository, StyleGAN2-ADA Dataset Preparation
-
I only tested datasets with resolution 256x256px.
-
You may use LSUN-CAT pre-trained models with AFHQ-CAT settings.
- StyleGAN2 generated samples
-
Prepare pre-trained StyleGAN2 model weights from FFHQ256x256 or LSUN-CHURCH256x256
-
I tested pre-trained models with 256x256 px resolutions from above two links. For AFHQ-CAT, AFHQ-WILD, I used manually trained models.
-
You may use LSUN-CAT pre-trained models with AFHQ-CAT settings.
- Dataset Version
- Fill in --dataset_path in generate_cams.sh
python generate_cams_dataset.py --dataset ffhq --dataset_path /path/to/ffhq_dataset --save_path segmentation/ffhq --mask
- StyleGAN2 generated samples
- Fill in --network in generate_cams_stylegan2.sh
python generate_cams_dataset_stylegan2.py --dataset ffhq --network /path/to/stylegan2-ffhq-network --save_path segmentation_stylegan2/ffhq --mask
@inproceedings{
li2022languagedriven,
title={Language-driven Semantic Segmentation},
author={Boyi Li and Kilian Q Weinberger and Serge Belongie and Vladlen Koltun and Rene Ranftl},
booktitle={International Conference on Learning Representations},
year={2022},
url={https://openreview.net/forum?id=RriDjddCLN}
}
@inproceedings{Karras2020ada,
title = {Training Generative Adversarial Networks with Limited Data},
author = {Tero Karras and Miika Aittala and Janne Hellsten and Samuli Laine and Jaakko Lehtinen and Timo Aila},
booktitle = {Proc. NeurIPS},
year = {2020}
}
This repository heavily depends on LSeg, StyleGAN2-ADA, CLIP