Code for NEUROCOMPUTING 2020 paper. "RGB-D Salient Object Detection via Cross-Modal Joint Feature Extraction and Low-Bound Fusion Loss".
- caffe
- PIL
- numpy
git clone https://github.com/Xinxin-Zhu/CJLB.git
- test
Download related dataset link, and put the test model link, in the "/models". Meanwhile, you need to set relevant path correctly.
cd Test
python RGBD_test.py
- train
The whole network training process includes two stages. In the first stage, a VGG16 model pre-trained link, on ImageNet is used to initialize the parameters of RGB and depth saliency prediction streams respectively, and the two independent streams are trained until convergence.
cd Train
python run.py ../models/vgg16_RGB-Depth_pre_train.caffemodel RGBNet_train.prototxt(or DepthNet_train.prototxt)
In the second stage, the whole network is initialized by the weights of the two streams link, , and the final model is obtained through further joint training.
cd Train
python run.py ../models/RGBDNet_pre_train.caffemodel RGBDNet_train.prototxt
The following are the evaluation results of the model on five RGB-D datasets.
Datasets\EvaluationMetrics | F-measure | E-measure | S-measure | MAE |
---|---|---|---|---|
NLPR | 0.887 | 0.949 | 0.906 | 0.033 |
NJUD | 0.877 | 0.923 | 0.883 | 0.056 |
STEREO | 0.872 | 0.927 | 0.880 | 0.055 |
LFSD | 0.807 | 0.856 | 0.832 | 0.106 |
DES | 0.898 | 0.962 | 0.910 | 0.030 |
- Tips: The results of the paper shall prevail. Because of the randomness of the training process, the results fluctuated slightly.
| NJUD | | NLPR | | STEREO | | LFSD | | DES |
- Note: The extraction code is XING.
- The web link is here.
@article{Zhu_2020_NC,
title={RGB-D Salient Object Detection via Cross-Modal Joint Feature Extraction and Low-Bound Fusion Loss},
author={Zhu, Xinxin and Li, Yi and Fu, Huazhu and Fan, Xiaoting and Shi, Yanan and Lei, Jianjun},
journal={Neurocomputing},
year={2020}
}
If you have any questions, please contact us ( xxzhu@163.com or mlzhang1998@gmail.com ).