Official code for ACM MM 2022 paper 'Progressive Limb-Aware Virtual Try-On'
We propose a novel progressive limb-aware virtual try-on framework named PL-VTON. PL-VTON consists of Multi-attribute Clothing Warping (MCW), Human Parsing Estimator (HPE), and Limb-aware Texture Fusion (LTF), which produces stable clothing deformation and handles the texture retention well in the final try-on result.
IEEE Transactions on Multimedia 2023 paper by us (follow-up research): https://github.com/xyhanHIT/PL-VTONv2
python 3.7
torch 1.9.0+cu111
torchvision 0.10.0+cu111
For the dataset, please refer to VITON.
-
Download the checkpoints from here.
-
Get VITON dataset.
-
Run the "test.py".
python test.py
Note that the results of our pretrained model are guaranteed in VITON dataset only.
The use of this code is restricted to non-commercial research and educational purposes.
If you use our code or models, please cite with:
@inproceedings{han2022progressive,
title={Progressive Limb-Aware Virtual Try-On},
author={Han, Xiaoyu and Zhang, Shengping and Liu, Qinglin and Li, Zonglin and Wang, Chenyang},
booktitle={Proceedings of the 30th ACM International Conference on Multimedia},
pages={2420--2429},
year={2022}
}
@article{zhang2023limb,
title={Limb-aware virtual try-on network with progressive clothing warping},
author={Zhang, Shengping and Han, Xiaoyu and Zhang, Weigang and Lan, Xiangyuan and Yao, Hongxun and Huang, Qingming},
journal={IEEE Transactions on Multimedia},
year={2023},
publisher={IEEE}
}
@inproceedings{han2018viton,
title={VITON: An Image-Based Virtual Try-On Network},
author={Han, Xintong and Wu, Zuxuan and Wu, Zhe and Yu, Ruichi and Davis, Larry S},
booktitle={Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition},
pages={7543--7552},
year={2018}
}
If you use the offered baseline results in your research, please cite with:
@inproceedings{wang2018toward,
title={Toward Characteristic-Preserving Image-Based Virtual Try-On Network},
author={Wang, Bochao and Zheng, Huabin and Liang, Xiaodan and Chen, Yimin and Lin, Liang and Yang, Meng},
booktitle={Proceedings of the European Conference on Computer Vision},
pages={589--604},
year={2018}
}
@inproceedings{minar2020cp,
title={CP-VTON+: Clothing Shape and Texture Preserving Image-Based Virtual Try-On},
author={Minar, Matiur Rahman and Tuan, Thai Thanh and Ahn, Heejune and Rosin, Paul and Lai, Yu-Kun},
booktitle={Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops},
year={2020}
}
@inproceedings{yang2020towards,
title={Towards Photo-Realistic Virtual Try-On by Adaptively Generating-Preserving Image Content},
author={Yang, Han and Zhang, Ruimao and Guo, Xiaobao and Liu, Wei and Zuo, Wangmeng and Luo, Ping},
booktitle={Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition},
pages={7850--7859},
year={2020}
}
@inproceedings{ge2021parser,
title={Parser-Free Virtual Try-On via Distilling Appearance Flows},
author={Ge, Yuying and Song, Yibing and Zhang, Ruimao and Ge, Chongjian and Liu, Wei and Luo, Ping},
booktitle={Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition},
pages={8485--8493},
year={2021}
}