Abstract
In this paper, we propose a saliency detection model for RGB-D images based on the contrasting features of colour and depth with a generative mixture model. The depth feature map is extracted based on superpixel contrast computation with spatial priors. We model the depth saliency map by approximating the density of depth-based contrast features using a Gaussian distribution. Similar to the depth saliency computation, the colour saliency map is computed using a Gaussian distribution based on multi-scale contrasts in superpixels by exploiting low-level cues. By assuming that colour- and depth-based contrast features are conditionally independent, given the classes, a discriminative mixed-membership naive Bayes (DMNB) model is used to calculate the final saliency map from the depth saliency and colour saliency probabilities by applying Bayes’ theorem. The Gaussian distribution parameter can be estimated in the DMNB model by using a variational inference-based expectation maximization algorithm. The experimental results on a recent eye tracking database show that the proposed model performs better than other existing models.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Achanta, R., Shaji, A., Smith, K., Lucchi, A., Fua, P., Süsstrunk, S.: Slic superpixels compared to state-of-the-art superpixel methods. IEEE Trans. Pattern Anal. Mach. Intell. 34, 2274–2282 (2012)
Blei, D.M., Jordan, M.I.: Variational inference for Dirichlet process mixtures. Bayesian Anal. 1, 121–143 (2006)
Borji, A., Itti, L.: State-of-the-art in visual attention modelling. IEEE Trans. Pattern Anal. Mach. Intell. 35, 185–207 (2013)
Borji, A., Cheng, M., Jiang, H., Li, J.: Salient object detection: a benchmark. IEEE Trans. Image Process. 24, 5706–5722 (2015)
Desingh, K., Madhava, K.K., Rajan, D., Jawahar, C.V.: Depth really matters: improving visual salient region detection with depth. In: BMVC (2013)
Fang, Y., Wang, J., Narwaria, M., Le Callet, L., Lin, W.: Saliency detection for stereoscopic images. IEEE Trans. Image Process. 23, 2625–2636 (2014)
Fang, Y., Lin, W., Fang, Z., Lei, J., Le Callet, P., Yuan, F.: Learning visual saliency for stereoscopic images. In: ICME (2014)
Ju, R., Ge, L., Geng, W., Ren, T., Wu, G.: Depth saliency based on anisotropic centre-surround difference. In: ICIP (2014)
Kim, H., Lee, S., Bovik, C.A.: Saliency prediction on stereoscopic videos. IEEE Trans. Image Process. 23, 1476–1490 (2014)
Lang, C., Ngugen, T.V., Katti, H., Yadati, K., Kankanhalli, M., Yan, S.: Depth matters: influence of depth cues on visual saliency. In: Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C. (eds.) ECCV 2012. LNCS, vol. 7573, pp. 101–115. Springer, Heidelberg (2012). doi:10.1007/978-3-642-33709-3_8
Le Callet, P., Niebur, E.: Visual attention and applications in multimedia technology. Proc. IEEE 101, 2058–2067 (2013)
Peng, H., Li, B., Xiong, W., Hu, W., Ji, R.: RGBD salient object detection: a benchmark and algorithms. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8691, pp. 92–109. Springer, Heidelberg (2014). doi:10.1007/978-3-319-10578-9_7
Radhakrishna, A., Sheila, H., Francisco, E., Sabine, S.: Frequency-tuned salient region detection. In: CVPR (2009)
Ren, J., Gong, X., Yu, L., Zhou, W.: Exploiting global priors for RGB-D saliency detection. In: CVPRW (2015)
Shan, H., Banerjee, A., Oza, N.C.: Discriminative mixed-membership models. In: ICDM (2009)
Wang, J., Fang, Y., Narwaria, M., Lin, W., Le Callet, P.: Stereoscopic image retargeting based on 3D saliency detection. In: ICASSP (2014)
Wu, P., Duan, L., Kong, L.: RGB-D salient object detection via feature fusion and multi-scale enhancement. In: Zha, H., Chen, X., Wang, L., Miao, Q. (eds.) CCCV 2015. CCIS, vol. 547, pp. 359–368. Springer, Heidelberg (2015). doi:10.1007/978-3-662-48570-5_35
Yang, C., Zhang, L., Lu, H., Ruan, X., Yang, M.H.: Saliency detection via graph based manifold ranking. In: CVPR (2013)
Zhang, Y., Jiang, G., Yu, M., Chen, K.: Stereoscopic visual attention model for 3D video. In: Boll, S., Tian, Q., Zhang, L., Zhang, Z., Chen, Y.-P.P. (eds.) MMM 2010. LNCS, vol. 5916, pp. 314–324. Springer, Heidelberg (2010). doi:10.1007/978-3-642-11301-7_33
Zhu, L., Cao, Z., Fang, Z., Xiao, Y., Wu, J., Deng, H., Liu, J.: Selective features for RGB-D saliency. In: CAC (2015)
Acknowledgement
This work was supported in part by the Beijing Academy of Science and Technology Youth Backbone Training Plan (2015–16) and Innovation Group Plan of Beijing Academy of Science and Technology (IG201506N).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2017 Springer International Publishing AG
About this paper
Cite this paper
Wang, ST., Zhou, Z., Qu, HB., Li, B. (2017). Visual Saliency Detection for RGB-D Images with Generative Model. In: Lai, SH., Lepetit, V., Nishino, K., Sato, Y. (eds) Computer Vision – ACCV 2016. ACCV 2016. Lecture Notes in Computer Science(), vol 10115. Springer, Cham. https://doi.org/10.1007/978-3-319-54193-8_2
Download citation
DOI: https://doi.org/10.1007/978-3-319-54193-8_2
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-54192-1
Online ISBN: 978-3-319-54193-8
eBook Packages: Computer ScienceComputer Science (R0)