Abstract
Saliency modeling has played an important part in computer vision studies over the past 30 years. Many state-of-the-art models adopted complex mathematical and machine learning theories. In this paper, a simple and effective visual attention model is proposed. We find that a single fixed template is enough for saliency map generation; this idea is inspired by the receptive field of the human visual system. All that is needed is to convolve the input image with this template with additional post-processing. Experiments show that our model is extremely fast and performs better than state-of-the-art models in human eye fixation prediction.
Similar content being viewed by others
References
Rodieck, R. W. (1965). Quantitative analysis of cat retinal ganglion cell response to visual stimuli. Vision Research, 5(12), 583–601.
Vinje, W. E., & Gallant, J. L. (2000). Sparse coding and decorrelation in primary visual cortex during natural vision. Science, 287(5456), 1273–1276.
Treisman, A. M., & Gelade, G. (1980). A feature-integration theory of attention. Cognitive Psychology, 12(1), 97–136.
Koch, C., & Ullman, S. (1987). Shifts in selective visual attention: Towards the underlying neural circuitry. In Matters of intelligence (pp. 115–141). Netherlands: Springer.
Itti, L., Koch, C., & Niebur, E. (1998). A model of saliency-based visual attention for rapid scene analysis. IEEE Transactions on Pattern Analysis & Machine Intelligence, 11, 1254–1259.
Hou, X., & Zhang, L. (2007, June). Saliency detection: A spectral residual approach. In Computer Vision and Pattern Recognition, 2007. CVPR’07. IEEE Conference on (pp. 1–8). IEEE.
Bruce, N. D., & Tsotsos, J. K. (2009). Saliency, attention, and visual search: an information theoretic approach. Journal of Vision, 9(3), 5.
Judd, T., Ehinger, K., Durand, F., & Torralba, A. (2009, September). Learning to predict where humans look. In Computer Vision, 2009 I.E. 12th international conference on (pp. 2106–2113). IEEE.
Tong, Y., Konik, H., Cheikh, F., & Tremeau, A. (2010). Full reference image quality assessment based on saliency map analysis. Journal of Imaging Science and Technology, 54(3), 30503-1.
Yubing, T., Konik, H., & Tremeau, A. (2010, July). Color face-tuned salient detection for image quality assessment. In Visual Information Processing (EUVIP), 2010 2nd European Workshop on (pp. 253–260). IEEE.
Yubing, T., Cheikh, F. A., Guraya, F. F. E., Konik, H., & Trémeau, A. (2011). A spatiotemporal saliency model for video surveillance. Cognitive Computation, 3(1), 241–263.
Guraya, F. F. E., Cheikh, F. A., Tremeau, A., Tong, Y., & Konik, H. (2010, August). Predictive saliency maps for surveillance videos. In Distributed Computing and Applications to Business Engineering and Science (DCABES), 2010 Ninth International Symposium on (pp. 508–513). IEEE.
Tong, Y., Konik, H., Cheikh, F. A., Guraya, F. F. E., & Tremeau, A. (2010, July). Multi-feature based visual saliency detection in surveillance video. In Visual Communications and Image Processing 2010 (pp. 774404–774404). International Society for Optics and Photonics.
Cheng, M. M., Zhang, Z., Lin, W. Y., & Torr, P. (2014). BING: Binarized normed gradients for objectness estimation at 300fps. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 3286–3293).
Zhao, Q., Liu, Z., & Yin, B. (2014, September). Cracking BING and Beyond. In BMVC.
Rantalankila, P., Kannala, J., & Rahtu, E. (2014). Generating object segmentation proposals using global and local search. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 2417–2424).
Zitnick, C. L., & Dollár, P. (2014). Edge boxes: Locating object proposals from edges. In Computer Vision–ECCV 2014 (pp. 391–405). Springer International Publishing.
Krähenbühl, P., & Koltun, V. (2014). Geodesic object proposals. In Computer Vision–ECCV 2014 (pp. 725–739). Springer International Publishing.
Perazzi, F., Krähenbühl, P., Pritch, Y., & Hornung, A. (2012, June). Saliency filters: Contrast based filtering for salient region detection. In Computer Vision and Pattern Recognition (CVPR), 2012 I.E. Conference on (pp. 733–740). IEEE.
Zhang, L., Tong, M. H., Marks, T. K., Shan, H., & Cottrell, G. W. (2008). SUN: a Bayesian framework for saliency using natural statistics. Journal of Vision, 8(7), 32.
Harel, J., Koch, C., & Perona, P. (2006). Graph-based visual saliency. In Advances in neural information processing systems (pp. 545–552).
Li, J., Levine, M. D., An, X., Xu, X., & He, H. (2013). Visual saliency based on scale-space analysis in the frequency domain. Pattern Analysis and Machine Intelligence, IEEE Transactions on, 35(4), 996–1010.
Borji, A., & Itti, L. (2012, June). Exploiting local and global patch rarities for saliency detection. In Computer Vision and Pattern Recognition (CVPR), 2012 I.E. Conference on (pp. 478–485). IEEE.
Acknowledgments
This work is supported by the Doctoral Dissertation Innovation Fund of Huazhong University of Science and Technology.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Tang, H., Chen, C. & Bie, Y. Prediction of Human Eye Fixation by a Single Filter. J Sign Process Syst 87, 197–202 (2017). https://doi.org/10.1007/s11265-016-1131-8
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11265-016-1131-8