/relabel_imagenet

Primary LanguagePythonApache License 2.0Apache-2.0

Re-labeling ImageNet: from Single to Multi-Labels, from Global to Localized Labels

Official PyTorch implementation of Re-labeling ImageNet | Paper | Pretrained Models

Sangdoo Yun, Seong Joon Oh, Byeongho Heo, Dongyoon Han, Junsuk Choe, Sanghyuk Chun

NAVER AI LAB

teaser

Abstract

ImageNet has been arguably the most popular image classification benchmark, but it is also the one with a significant level of label noise. Recent studies have shown that many samples contain multiple classes, despite being assumed to be a single-label benchmark. They have thus proposed to turn ImageNet evaluation into a multi-label task, with exhaustive multi-label annotations per image. However, they have not fixed the training set, presumably because of a formidable annotation cost. We argue that the mismatch between single-label annotations and effectively multi-label images is equally, if not more, problematic in the training setup, where random crops are applied. With the single-label annotations, a random crop of an image may contain an entirely different object from the ground truth, introducing noisy or even incorrect supervision during training. We thus re-label the ImageNet training set with multi-labels. We address the annotation cost barrier by letting a strong image classifier, trained on an extra source of data, generate the multi-labels. We utilize the pixel-wise multi-label predictions before the final pooling layer, in order to exploit the additional location-specific supervision signals. Training on the re-labeled samples results in improved model performances across the board. ResNet-50 attains the top-1 classification accuracy of 78.9% on ImageNet with our localized multi-labels, which can be further boosted to 80.2% with the CutMix regularization. We show that the models trained with localized multi-labels also outperforms the baselines on transfer learning to object detection and instance segmentation tasks, and various robustness benchmarks.

Updates

  • Jan 14, 2021: Initial upload

Installation

Dependancies

Our implementations are tested on the following libraries with Python 3.6.9 and CUDA 10.1.

torch: 1.6.0
torchvision: 0.7.0
apex: 0.1
munch: 2.5.0
adamp: 0.2.0
fire: 0.3.1
pyyaml: 5.3.1

Install PyTorch and Torchvision as following official PyTorch instruction

conda install pytorch==1.6.0 torchvision==0.7.0 cudatoolkit=10.1 -c pytorch

Install other dependencies using the following command.

pip install -r requirements.txt

Dataset

  1. Download the original ImageNet dataset and place it in --data.data_path directory.
  2. Download relabel_imagenet.tar (12GB) and extract it in --data.relabel.path directory.

The ReLabel's structure is the same as ImageNet (ILSVRC) dataset's.

root
 └─── n04355338
          └───── n04355338_22023.pt
          └───── n04355338_2663.pt
          └───── ...
 └─── n03786901
          └───── n03786901_5410.pt
          └───── ...
 └─── ...

How to Run

Train a model with ReLabel

  • We used four NVIDIA V100 GPUs.
  • Train ResNet-50 with ReLabel to obtain 78.9% top-1 accuracy.
python main.py --config-file-path configs/relabel_train_resnet50.yaml \
               --data.relabel.path 'path/to/relabel'
  • Train ResNet-50 with ReLabel+CutMix to obtain 80.2% top-1 accuracy.
python main.py --config-file-path configs/relabel_train_resnet50_cutmix.yaml \
               --data.relabel.path 'path/to/relabel'

Test a pretrained model

python main.py --config-file-path configs/relabel_train_resnet50.yaml \
               --util.test_only True \
               --util.test_weight_file 'path/to/checkpoint'

Arguments

  • --config-file-path: Path to config file (.yaml) which overrides configs/base.yaml.
  • --model.arch: Network architecture provided by torchvision (Default: resnet50)
  • --data.data_path: Path to the original ImageNet dataset (Default: '/home/data/ILSVRC2015/')
  • --data.relabel.use: Use ReLabel ground truth (Default: True)
  • --data.relabel.path: Path to ReLabel annotations
  • --data.cutmix.prob: Probability to apply CutMix regularizer (Default: 0.0)
  • --data.cutmix.beta: Hyper-parameter for CutMix regularizer (Default: 0.0)
  • --data.cutmix.off_epoch: How many last epochs to train without CutMix regularizer (Default: 5)
  • --optim.optimizer: Name of optimizer. sgd and AdamP optimizers are available (Default: sgd)
  • --optim.epochs: Total epochs to train (Default: 100)
  • --optim.start_epoch: Starting epoch for restarts
  • --optim.batch_size: Mini-batch size of all GPUs (Default: 512)
  • --optim.lr.init: Initial learning rate
  • --optim.lr.decay_type: Type of learning rate decay (Default: cos)
  • --optim.wd.base: Weight decay
  • --optim.warmup.lr: Initial warmup learning rate (Default: 0.0)
  • --optim.warmup.epochs: Warmup epochs (Default: 5)
  • --optim.nesterov: Enable Nesterov momentum (Default: True)
  • --optim.momentum: Momentum parameter (Default: 0.9)
  • --optim.label_smoothing: Label smoothing parameter (Default: 0.0)
  • --util.test_only: Only evaluate model without training (Default: False)
  • --util.test_weight_file: Path to the model file to be evaluated

Performances and Pretrained Models

  • ResNet-50 results on single-label and multi-label evalutation for ImageNet classficiation
Method ImageNet ImageNetV2 ReaL Shankar et al.
ResNet-50 77.5 79.0 83.6 85.3
ResNet-50 + Label smoothing (e=0.1) 78.0 79.5 84.0 84.7
ResNet-50 + Label cleaning 78.1 79.1 83.6 85.2
ResNet-50 + ReLabel 78.9 80.5 85.0 86.1
  • ReLabel on various architectures
Method ImageNet GT ReLabel GT
ResNet-18 71.7 72.5 (+0.8) [model_file]
ResNet-50 77.5 78.9 (+1.4) [model_file]
ResNet-101 78.1 80.7 (+2.6) [model_file]
ReXNet (x1.0) 77.9 78.4 (+0.5)
EfficientNet-B0 77.4 78.0 (+0.6)
EfficientNet-B1 79.2 80.3 (+1.1)
EfficientNet-B2 80.3 81.0 (+0.7)
EfficientNet-B3 81.7 82.5 (+0.8)
  • ReLabel with additional tricks
Method ImageNet
ResNet-18 71.7
ResNet-18 + ReLabel 72.5 (+0.8) [model_file]
ResNet-18 + ReLabel + CutMix 73.2 (+1.5) [model_file]
ResNet-50 77.5
ResNet-50 + ReLabel 78.9 (+1.4) [model_file]
ResNet-50 + ReLabel + CutMix 80.2 (+2.7) [model_file]
ResNet-50 + ReLabel + CutMix + ImageNet-21K 81.2 (+3.7) [model_file]
ResNet-101 78.1
ResNet-101 + ReLabel 80.7 (+2.6) [model_file]
ResNet-101 + ReLabel + CutMix 81.6 (+3.5) [model_file]

License

Copyright 2021-present NAVER Corp.

Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at

    http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.

Citation

@article{yun2021relabel,
    title={Re-labeling ImageNet: from Single to Multi-Labels, from Global to Localized Labels},
    author={Yun, Sangdoo and Oh, Seong Joon and Heo, Byeongho and Han, Dongyoon and Choe, Junsuk and Chun, Sanghyuk},
    year={2021},
    journal={arXiv preprint arXiv:2101.05022},
}