/large_scale_ood

MOS: Towards Scaling Out-of-distribution Detection for Large Semantic Space

Primary LanguagePythonApache License 2.0Apache-2.0

MOS: Towards Scaling Out-of-distribution Detection for Large Semantic Space

This is the source code for our paper MOS: Towards Scaling Out-of-distribution Detection for Large Semantic Space by Rui Huang and Sharon Li. Code is modified from Google BiT, ODIN, Outlier Exposure, deep Mahalanobis detector and Robust OOD Detection.

This is a group-based OOD detection framework that is effective for large-scale image classification. Our key idea is to decompose the large semantic space into smaller groups with similar concepts, which allows simplifying the decision boundary and reducing the uncertainty space between in- vs. out-of-distribution data.

model_architecture

Usage

1. Dataset Preparation

In-distribution dataset

Please download ImageNet-1k and place the training data and validation data in ./dataset/id_data/ILSVRC-2012/train and ./dataset/id_data/ILSVRC-2012/val, respectively.

Out-of-distribution dataset

We have curated 4 OOD datasets from iNaturalist, SUN, Places, and Textures, and de-duplicated concepts overlapped with ImageNet-1k.

For iNaturalist, SUN, and Places, we have sampled 10,000 images from the selected concepts for each dataset, which can be download via the following links:

wget http://pages.cs.wisc.edu/~huangrui/imagenet_ood_dataset/iNaturalist.tar.gz
wget http://pages.cs.wisc.edu/~huangrui/imagenet_ood_dataset/SUN.tar.gz
wget http://pages.cs.wisc.edu/~huangrui/imagenet_ood_dataset/Places.tar.gz

For Textures, we use the entire dataset, which can be downloaded from their original website.

Please put all downloaded OOD datasets into ./dataset/ood_data/. For more details about these OOD datasets, please check out our paper.

2. Pre-trained Model Preparation

Please download the BiT-S pre-trained model families and put them into the folder ./bit_pretrained_models. The backbone used in our paper for main results is BiT-S-R101x1.

3. Group-softmax/Flat-softmax Model Finetuning

For group-softmax finetuning (MOS), please run:

bash ./scripts/finetune_group_softmax.sh

For flat-softmax finetuning (baselines), please run:

bash ./scripts/finetune_flat_softmax.sh

4. OOD Detection Evaluation

To reproduce our MOS results, please run:

bash ./scripts/test_mos.sh iNaturalist(/SUN/Places/Textures)

To reproduce baseline approaches, please run:

bash ./scripts/test_baselines.sh MSP(/ODIN/Energy/Mahalanobis/KL_Div) iNaturalist(/SUN/Places/Textures)

Note: before testing Mahalanobis, make sure you have tuned and saved its hyperparameters first by running:

bash ./scripts/tune_mahalanobis.sh

Our Fine-tuned Model

To facilitate the reproduction of the results reported in our paper, we also provide our group-softmax finetuned model and flat-softmax finetuned model, which can be downloaded via the following links:

wget http://pages.cs.wisc.edu/~huangrui/finetuned_model/BiT-S-R101x1-group-finetune.pth.tar
wget http://pages.cs.wisc.edu/~huangrui/finetuned_model/BiT-S-R101x1-flat-finetune.pth.tar

After downloading the provided models, you can skip Step 3 and set --model_path in scripts in Step 4 accordingly.

OOD Detection Results

MOS achieves state-of-the-art performance averaged on the 4 OOD datasets.

results

Citation

If you use our codebase or OOD datasets, please cite our work:

@inproceedings{huang2021mos,
  title={MOS: Towards Scaling Out-of-distribution Detection for Large Semantic Space},
  author={Huang, Rui and Li, Yixuan},
  booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
  year={2021}
}