SNIPER: Efficient Multi-Scale Training
SNIPER is an efficient multi-scale training approach for instance-level recognition tasks like object detection and instance-level segmentataion. Instead of processing all pixels in an image pyramid, SNIPER selectively processes context regions around the ground-truth objects (a.k.a chips). This significantly speeds up multi-scale training as it operates on low-resolution chips. Due to its memeory efficient design, SNIPER can benefit from Batch Normalization during training and it makes larger batch-sizes possible for instance-level recognition tasks on a single GPU. Hence, we do not need to synchronize batch-normalization statistics across GPUs and we can train object detectors similar to the way we do image classification!
SNIPER is described in the following paper:
SNIPER: Efficient Multi-Scale Training Bharat Singh*, Mahyar Najibi*, and Larry S. Davis (* denotes equal contribution) arXiv preprint arXiv:1805.09300, 2018.
Features
- Train with a batch size of 160 images with a ResNet-101 backbone on 8 V100 GPUs
- NO PYTHON LAYERS (Every layer is optimized for large batch sizes in CUDA/C++)
- HALF PRECISION TRAINING with no loss in accuracy
- 5 Images/second during inference on a single V100 GPU, 47.6/68.5 on COCO without training on segmentation masks
- The R-FCN-3K branch is also powered by SNIPER. Now 21% better than YOLO-9000 on ImageNetDet. This branch also supports on-the-fly training (in seconds) with very few samples (no bounding boxes needed!)
- Train on OpenImagesV4 (14x bigger than COCO) with ResNet-101 in 3 days on a p3.x16.large AWS instance!
Results
Here are the coco test-dev results for SNIPER trained with this repository on the coco trainval set and using only the bounding box annotations.
pre-trained dataset | network structure | mAP | mAP@0.5 | mAP@0.75 | mAP@S | mAP@M | mAP@L | |
---|---|---|---|---|---|---|---|---|
SNIPER | ImageNet | ResNet101 | 46.5 | 67.5 | 52.2 | 30.0 | 49.4 | 58.4 |
SNIPER | OpenImages | ResNet101 | 47.8 | 68.2 | 53.6 | 31.5 | 50.4 | 59.8 |
You can download the OpenImages pre-trained model and the SNIPER detector by running bash scripts/download_pretrained_models.sh
and bash scripts/download_sniper_detector.sh
respectively.
License
SNIPER is released under Apache license. See LICENSE for details.
Citing
@article{sniper2018,
title={{SNIPER}: Efficient Multi-Scale Training},
author={Singh, Bharat and Najibi, Mahyar and Davis, Larry S},
journal={arXiv preprint arXiv:1805.09300},
year={2018}
}
@article{analysissnip2017,
title={An analysis of scale invariance in object detection-snip},
author={Singh, Bharat and Davis, Larry S},
journal={CVPR},
year={2018}
}
Contents
- Installation
- Running the demo
- Training a model with SNIPER
- Evaluting a trained model
- Other methods and branches in this repo (SSH face, R-FCN-3K, open-images)
Installation
- Clone the repository:
git clone --recursive https://github.com/mahyarnajibi/SNIPER.git
- Compile the provided mxnet fork in the repository:
cd SNIPER-mxnet
mkdir build
cd build
cmake ..
make
- Compile the C++ files in the lib directory. The following script compiles them all:
bash scripts/compile.sh
- Add mxnet to the
PYTHONPATH
:
export PYTHONPATH=SNIPER-mxnet/python:$PYTHONPATH
- Install the required python packages:
pip install -r requirements.txt
Running the demo
For running the demo, you need to download the provided SNIPER model. The following script downloads the SNIPER model and extracts it into the default location:
bash download_sniper_detector.sh
After downloading the model, the following command would run the SNIPER detector with the default configs on the provided sample image:
python demo.py
If everything goes well, the sample detections would be saved as data/demo/demo_detections.png
.
You can also run the detector on an arbitary image by providing its path to the script:
python demo.py --im_path [PATH to the image]
Training a model
For training SNIPER on COCO, you would need to download the pre-trained models, the pre-computed proposals used for negative chip mining (you can also use any other set of proposals), and configure the dataset as described below.
Downloading pre-trained models
Running the following script downloads and extracts the pre-trained models into the default path (data/pretrained_model
):
bash download_pretrained_models.sh
Downloading pre-computed proposals for negative chip mining
Running the following script downloads and extract the pre-computed proposals into the default path (data/proposals
):
bash download_sniper_neg_props.sh
Configuring the COCO dataset
Please follow the official COCO dataset website to download the dataset. After downloading the dataset you should have the following directory structure:
data
|--datasets
|--coco
|--annotations
|--images
To train a model with SNIPER and default parameters you can call the following script:
python main_train.py
The default settings can be overwritten by passing a configuration file (see the configs
folder for example configuration files).
The path to the configuration file can be passed as an argument to the above script using the --cfg
flag .
Please note that the default config file has the same settings used to train the released models. If you are using a GPU with less amount of memory, please consider reducing the training batch size (by setting TRAIN.BATCH_IMAGES
in the config file). Also, multi-processing is used to process the data. For lower amounts of memory, you may need to reduce the number of processes and number of threads according to your system (by setting TRAIN.NUM_PROCESS
and TRAIN.NUM_THREAD
).
Evaluating a trained model
Evaluating the provided SNIPER models
The repository provides a set of pre-trained SNIPER models which can be downloaded by running the following script:
bash download_sniper_detector.sh
This script downloads the model weights and extracts them into the expected directory. To evaluate these models on coco test-dev with the default configuration, you can run the following script:
python main_test.py
The default settings can be overwritten by passing the path to a configuration file with the --cfg
flag
(See the configs
folder for examples).
Please note that the evaluation is performed in a multi-image per batch and parallel model forward setting. In case of lower GPU memory, please consider reducing the batch size for different scales (by setting TEST.BATCH_IMAGES
) or reducing the number of parallel jobs (by setting TEST.CONCURRENT_JOBS
in the config file).
Evaluating a model trained with this repository
For evaluating a model trained with this repository, you can run the following script by passing the same configuration file used during the training.
The test settings can be set by updating the TEST
section of the configuration file (See the configs
folder for examples).
python main_test.py --cfg [PATH TO THE CONFIG FILE USED FOR TRAINING]
By default, this would produce a json
file containing the detections on the test-dev
which can be zipped and uploaded to the COCO evaluation server.
Other methods and branches in this repo (SSH face, R-FCN-3K, open-images)
R-FCN-3K
This repo also contains the R-FCN-3k detector.
Please switch to the R-FCN-3k branch for specific instructions.
SSH Face Detector (Comming Soon)
The SSH face detector would be added to this repository soon. In the meanwhile, you use the code available at the original SSH repository.
OpenImagesV4
This repo also contains modules to train on the open-images dataset. Please switch to the openimages2 branch for specific instructions.