/SEA-T2F

Multi-caption Text-to-Face Synthesis: Database and Algorithm

Primary LanguagePython

SEA-T2F

Multi-caption Text-to-Face Synthesis: Database and Algorithm

Requirements: python=3.7, pytorch=1.1.0, torchvision=0.3.0

  1. Download data:

Download the CelebAText-HQ from https://drive.google.com/drive/folders/1IAb_iy6-soEGQWhbgu6cQODsIUJZpahC?usp=sharing, and extract them into data/CelebAText-HQ.

Download the Multi-modal CelebA-HQ from https://github.com/IIGROUP/Multi-Modal-CelebA-HQ-Dataset and extract them into data/Multi-modal.

Download the CelebAMask-HQ from https://github.com/switchablenorms/CelebAMask-HQ and copy CelebAimg into data/CelebAText-HQ or data/Multi-modal CelebA-HQ.

  1. Download pretrain model:

Download the pretriain model of CelebAText-HQ from https://drive.google.com/drive/folders/1XufcOo_I09h86ZR2M4UJ8WTVR7ARF87d, and extract them into /DAMSMEencoders/CelebAText-HQ.

Download the pretriain model of Multi-modal CelebA-HQ from https://drive.google.com/drive/folders/1FN4q7xD1jKvXeG3Pd6wqieUlr52bVoIX, and extract them into /DAMSMEencoders/CelebAText-HQ.

  1. Download generative model:

Download the generative model of CelebAText-HQ from https://drive.google.com/drive/folders/1XufcOo_I09h86ZR2M4UJ8WTVR7ARF87d, and copy it into /models.

Download the generative model of Multi-modal CelebA-HQ from https://drive.google.com/drive/folders/1FN4q7xD1jKvXeG3Pd6wqieUlr52bVoIX, and copy it into /models.

  1. Train:

python main.py --cfg cfg/train_face.yml --gpu 7

  1. Evaluation:

python main.py --cfg cfg/eval_face.yml --gpu 7