/ILM-VP

[CVPR23] "Understanding and Improving Visual Prompting: A Label-Mapping Perspective" by Aochuan Chen, Yuguang Yao, Pin-Yu Chen, Yihua Zhang, and Sijia Liu

Primary LanguagePythonMIT LicenseMIT

Iterative-Label-Mapping Visual Prompting (ILM-VP)

License: MIT

[CVPR23] "Understanding and Improving Visual Prompting: A Label-Mapping Perspective"

In this work, we revisit and advance visual prompting (VP), an input prompting technique for vision tasks. VP can reprogram a fixed, pre-trained source model to accomplish downstream tasks in the target domain by simply incorporating universal prompts (in terms of input perturbation patterns) into downstream data points. Yet, it remains elusive why VP stays effective even given a ruleless label mapping (LM) between the source classes and the target classes. Inspired by the above, we ask: How is LM interrelated with VP? And how to exploit such a relationship to improve its accuracy on target tasks? We peer into the influence of LM on VP and provide an affirmative answer that a better 'quality' of LM (assessed by mapping precision and explanation) can consistently improve the effectiveness of VP. This is in contrast to the prior art where the factor of LM was missing. To optimize LM, we propose a new VP framework, termed ILM-VP (iterative label mapping-based visual prompting), which automatically re-maps the source labels to the target labels and progressively improves the target task accuracy of VP. Further, when using a contrastive language-image pretrained (CLIP) model, we propose to integrate an LM process to assist the text prompt selection of CLIP and to improve the target task accuracy. Extensive experiments demonstrate that our proposal significantly outperforms state-of-the-art VP methods. As highlighted below, we show that when reprogramming an ImageNet-pretrained ResNet-18 to 13 target tasks, our method outperforms baselines by a substantial margin, e.g., 7.9% and 6.7% accuracy improvements in transfer learning to the target Flowers102 and CIFAR100 datasets. Besides, our proposal on CLIP-based VP provides 13.7% and 7.1% accuracy improvements on Flowers102 and DTD respectively.

Overview

What is in this repository?

We provide the training code for our ILM-VP method and the baselines both on ResNets and CLIP.

Dependencies

Run pip3 install -r requirements.txt.

For path configurations, modify cfg.py according to your need.

Datasets

You can find datasets here.

Put dataset under data_path in cfg.py. (e.g., data_path/flowers102, data_path/ucf101, ...)

For Flowers102, DTD, UCF101, Food101, EuroSAT, OxfordPets, StanfordCars and SUN397, we use datasets splitted in CoOp. For ABIDE, we use download code in acerta-abide. For other datasets we use official ones provided by pytorch.

Generate Prompts

VP on CNN:

ILM-VP: python experiments/cnn/ilm_vp.py --network resnet18 --dataset flowers102

FLM-VP: python experiments/cnn/flm_vp.py --network resnet18 --dataset flowers102

RLM-VP: python experiments/cnn/rlm_vp.py --network resnet18 --dataset flowers102

TP on CLIP:

ILM-TP-VP: python experiments/clip/ilm_tp_vp.py --dataset flowers102

SINGLE-TP-VP: python experiments/clip/single_tp_vp.py --dataset flowers102

Contributor

Aochuan Chen

Citation

@InProceedings{Chen_2023_CVPR,
    author    = {Chen, Aochuan and Yao, Yuguang and Chen, Pin-Yu and Zhang, Yihua and Liu, Sijia},
    title     = {Understanding and Improving Visual Prompting: A Label-Mapping Perspective},
    booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
    month     = {June},
    year      = {2023},
    pages     = {19133-19143}
}