Abstract
The computer model of visual attention derives an interest or saliency map from an input image in a process that encompasses several data combination steps. While several combination strategies are possible, not all perform equally well. This paper compares main cue combination strategies by measuring the performance of the considered models with respect to human eye movements. Six main combination methods are compared in experiments involving the viewing of 40 images by 20 observers. Similarity is evaluated qualitatively by visual tests and quantitatively by use of a similarity score. The study provides insight into the map combination mechanisms and proposes in this respect an overall optimal strategy for a computer saliency model.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Itti, L., Koch, C., Niebur, E.: A model of saliency-based visual attention for rapid scene analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI) 20(11), 1254–1259 (1998)
Ouerhani, N., Hugli, H.: Real-time visual attention on a massively parallel SIMD architecture. International Journal of Real Time Imaging 9(3), 189–196 (2003)
Ouerhani, N., Hugli, H.: MAPS: Multiscale attention-based presegmentation of color images. In: Griffin, L.D., Lillholm, M. (eds.) Scale-Space 2003. LNCS, vol. 2695, pp. 537–549. Springer, Heidelberg (2003)
Walther, D., Rutishauser, U., Koch, C., Perona, P.: Selective visual attention enables learning and recognition of multiple objects in cluttered scenes. Computer Vision and Image Understanding 100(1-2), 41–63 (2005)
Koch, C., Ullman, S.: Shifts in selective visual attention: Towards the underlying neural circuitry. Human Neurobiology 4, 219–227 (1985)
Itti, L., Koch, C.: A comparison of feature combination strategies for saliency-based visual attention systems. In: SPIE Human Vision and Electronic Imaging IV (HVEI’99), vol. 3644, pp. 373–382 (1999)
Le Meur, O., Le Callet, P., Barba, D., Thoreau, D.: A coherent computational approach to model bottom-up visual attention. PAMI 28(5) (2006)
Ouerhani, N., Bur, A., Hügli, H.: Linear vs. nonlinear feature combination for saliency computation: A comparison with human vision. In: Franke, K., Müller, K.-R., Nickolay, B., Schäfer, R. (eds.) Pattern Recognition. LNCS, vol. 4174, pp. 314–323. Springer, Heidelberg (2006)
Ouerhani, N., Jost, T., Bur, A., Hugli, H.: Cue normalization schemes in saliency-based visual attention models. In: International Cognitive Vision Workshop, Graz, Austria (2006)
Parkhurst, D., Law, K., Niebur, E.: Modeling the role of salience in the allocation of overt visual attention. Vision Research 42(1), 107–123 (2002)
Ouerhani, N., von Wartburg, R., Hugli, H., Mueri, R.: Empirical validation of the saliency-based model of visual attention. Electronic Letters on Computer Vision and Image Analysis (ELCVIA) 3(1), 13–24 (2004)
Itti, L.: Quantitative modeling of perceptual salience at human eye position. Visual Cognition, in press (2005)
Author information
Authors and Affiliations
Editor information
Rights and permissions
Copyright information
© 2007 Springer Berlin Heidelberg
About this paper
Cite this paper
Bur, A., Hügli, H. (2007). Optimal Cue Combination for Saliency Computation: A Comparison with Human Vision. In: Mira, J., Álvarez, J.R. (eds) Nature Inspired Problem-Solving Methods in Knowledge Engineering. IWINAC 2007. Lecture Notes in Computer Science, vol 4528. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-73055-2_13
Download citation
DOI: https://doi.org/10.1007/978-3-540-73055-2_13
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-73054-5
Online ISBN: 978-3-540-73055-2
eBook Packages: Computer ScienceComputer Science (R0)