Official implementation of Improving Zero-Shot Generalization for CLIP with Synthesized Prompts.
This paper has been accepted by ICCV 2023.
Create a conda environment and install dependencies:
conda create -n ship python=3.9
conda activate ship
pip install -r requirements.txt
# Install the according versions of torch and torchvision
conda install pytorch torchvision cudatoolkit
Follow DATASET.md to install ImageNet and other 10 datasets referring to CoOp.
The running configurations can be modified in coop-configs/dataset.yaml
, including shot numbers, visual encoders, and hyperparamters.
For ImageNet dataset:
CUDA_VISIBLE_DEVICES=0 python main_imagenet_coop_vae.py --config configs/imagenet.yaml
For other 10 datasets:
CUDA_VISIBLE_DEVICES=0 python main_coop_vae.py --config configs/dataset.yaml
This repo benefits from CLIP, CoOp and Tip-Adapter. Thanks for their wonderful works.
@inproceedings{wang2023improving,
title={Improving Zero-Shot Generalization for CLIP with Synthesized Prompts},
author={Zhengbo Wang and Jian Liang and Ran He and Nan Xu and Zilei Wang and Tieniu Tan},
booktitle={IEEE/CVF International Conference on Computer Vision (ICCV)},
year={2023}
}
If you have any question, feel free to contact zhengbowang@mail.ustc.edu.cn.