This is the implementation of our ECCV2022 paper: Towards-Accurate-Binary-Neural-Networks-via-Modeling-Contextual-Dependencies
Existing Binary Neural Networks (BNNs) mainly operate on local convolutions with binarization function. However, such simple bit operations lack the ability of modeling contextual dependencies, which is critical for learning discriminative deep representations in vision models. In this work, we tackle this issue by presenting new designs of binary neural modules, which enables BNNs to learn effective contextual dependencies. First, we propose a binary multi-layer perceptron (MLP) block as an alternative to binary convolution blocks to directly model contextual dependencies. Both short-range and long-range feature dependencies are modeled by binary MLPs, where the former provides local inductive bias and the latter breaks limited receptive field in binary convolutions. A sparse contextual interaction is achieved with the long-short range binary MLP block. Second, we compute the contextual dynamic embeddings to determine the binarization thresholds in general binary convolutional blocks. Armed with our binary MLP blocks and improved binary convolution, we build the BNNs with explicit Contextual Dependency modeling, termed as BCDNet. On the standard ImageNet-1K classification benchmark, the BCDNet achieves 72.3% Top-1 accuracy and outperforms leading binary methods by a large margin. In particular, the proposed BCDNet exceeds the state-of-the-art ReActNet-A by 2.9% Top-1 accuracy with similar operations, demonstrating the effectiveness of our binary designs.
The performance of BCDNets on ImageNet-1K is:
- slurm server
- python3, pytorch, torchvision ...
- Download ImageNet dataset
(1) Step1: binarizing activations
- Change directory
cd model_zoo_exp/BCDNet/a_3f
- run
bash run.sh
(2) Step2: binarizing weights and activations
- Change directory
cd model_zoo_exp/BCDNet/a_3
- run
bash run.sh
@article{xing2022towards,
title={Towards Accurate Binary Neural Networks via Modeling Contextual Dependencies},
author={Xing, Xingrun and Li, Yangguang and Li, Wei and Ding, Wenrui and Jiang, Yalong and Wang, Yufeng and Shao, Jing and Liu, Chunlei and Liu, Xianglong},
journal={arXiv preprint arXiv:2209.01404},
year={2022}
}