/SHIP

Official code for ICCV 2023 paper, "Improving Zero-Shot Generalization for CLIP with Synthesized Prompts"

Primary LanguagePython

Improving Zero-Shot Generalization for CLIP with Synthesized Prompts

Official implementation of Improving Zero-Shot Generalization for CLIP with Synthesized Prompts.

This paper has been accepted by ICCV 2023.

Requirements

Installation

Create a conda environment and install dependencies:

conda create -n ship python=3.9
conda activate ship

pip install -r requirements.txt

# Install the according versions of torch and torchvision
conda install pytorch torchvision cudatoolkit

Dataset

Follow DATASET.md to install ImageNet and other 10 datasets referring to CoOp.

Get Started

Configs

The running configurations can be modified in coop-configs/dataset.yaml, including shot numbers, visual encoders, and hyperparamters.

Running

For ImageNet dataset:

CUDA_VISIBLE_DEVICES=0 python main_imagenet_coop_vae.py --config configs/imagenet.yaml

For other 10 datasets:

CUDA_VISIBLE_DEVICES=0 python main_coop_vae.py --config configs/dataset.yaml

Acknowledgement

This repo benefits from CLIP, CoOp and Tip-Adapter. Thanks for their wonderful works.

Citation

@inproceedings{wang2023improving,
  title={Improving Zero-Shot Generalization for CLIP with Synthesized Prompts},
  author={Zhengbo Wang and Jian Liang and Ran He and Nan Xu and Zilei Wang and Tieniu Tan},
  booktitle={IEEE/CVF International Conference on Computer Vision (ICCV)},
  year={2023}
}

Contact

If you have any question, feel free to contact zhengbowang@mail.ustc.edu.cn.