Now in experimental release, suggestions welcome
M.J. Sun, Z.Q. Zhou, Q. H. Hu, and Z. Wang, SG-FCN: A Motion and Memory-Based Deep Learning Model for Video Saliency Detection, IEEE Transactions on Cybernetics.
@ARTICLE{8365810,
author={M. Sun and Z. Zhou and Q. Hu and Z. Wang and J. Jiang},
journal={IEEE Transactions on Cybernetics},
title={SG-FCN: A Motion and Memory-Based Deep Learning Model for Video Saliency Detection},
year={2018},
volume={},
number={},
pages={1-12},
keywords={Computational modeling;Saliency detection;Predictive models;Feature extraction;Video sequences;Visualization;Training;Eye fixation detection;fully convolutional neural networks;video saliency},
doi={10.1109/TCYB.2018.2832053},
ISSN={2168-2267},
month={},}
Note: We are very grateful to the source code provided by ConsistentViSal and the role of VSOD in promoting this work.
Which can be cited by:
W. Wang, J. Shen, and L. Shao,
Consistent video saliency using local gradient flow optimization and global refinement,
IEEE Trans. on Image Processing, 24(11):4185-4196, 2015
W. Wang, J. Shen, and L. Shao,
Video salient object detection via fully convolutional networks,
IEEE Trans. on Image Processing, 27(1):38-49, 2018
Structure of model SGF(E). As shown in the flowchart, the input data is a tensor of h × w × 4. At the top of the model, we add an Eltwise layer with function SUM [big map(i), boundary map(i)] before Sigmoid function.
details of the OPB algorithm can be found in './OPB/' Note: The OPB algorithm draws on ConsistentViSal, which can be accessed and referenced by
W. Wang, J. Shen, and L. Shao,
Consistent video saliency using local gradient flow optimization and global refinement,
IEEE Trans. on Image Processing, 24(11):4185-4196, 2015
Please first download and install caffe. caffe
The model weights trained on HOLLYWOOD2 and UCF-Sports datasets can be downloaded from
Baidu Wangpan: https://pan.baidu.com/s/1bgu80UOKJOXN2OvhAx_2sw
password: lfi8
Please put the download 'caffe' folder under the main branch, and put './models/' under the folder './caffe/', then run main.m.
In order to better eliminate the checkerboard effect, we make an adjustment to the parameters of the deconvolution layer.
Our results are sightly improved over the original scores reported in our paper:
dataset | CC | SIM | NSS | EMD | AUC |
---|---|---|---|---|---|
HOLLYWOOD2 | 0.6181 | 0.5161 | 1.5158 | 0.8984 | 0.8936 |
UCF-Sports | 0.5985 | 0.4524 | 1.4582 | 0.8301 | 0.9065 |
If you find our method useful in your research, please consider citing:
M. Sun, Z. Zhou, Q. Hu, Z. Wang, and J. Jiang, “Sg-fcn: A motion and memory-based deep learning model for video saliency detection,” IEEE Transactions on Cybernetics, vol. PP, no. 99, pp. 1–12, 2018.
Welcome to our LAB:
https://zhengwangtju.github.io/
http://www.escience.cn/people/sunmeijun/index.html
Any questions, please contact me via ziqizhou@tju.edu.cn