Abstract
In this paper, we present a new spatiotemporal visual attention system. Typical feature integration model is expanded to incorporate motion in our suggested system, and is able to respond to motion stimulus by employing motion fields map as one of temporal features. Proposed system is based on bottom-up approach of human visual attention, but the main difference lies in its temporal feature extraction method, and integration method of multiple spatial and temporal features. Spatial features are integrated into spatial saliency map by weighted combination method. Temporal feature is extracted by SIFT and is analyzed and reorganized into temporal saliency map. Finally, dynamic fusion technique applied to make one spatiotemporal saliency map. To evaluate the performance of the system, we tested with various kinds of real video sequences. We also compared our system with several previous systems to validate the performance of the system.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Zhai, Y., Shah, C.: Visual Attention Detection in Video Sequences Using Spatiotemporal Cues. In: Proceedings of the 14th Annual ACM International Conference on Multimedia, pp. 815–824. ACM, New York (2006)
Lowe, D.G.: Distinctive Image Features from Scale-Invariant Key points. International Journal of Computer Vision 60(2), 91–110 (2004)
Itti, L., Koch, C.: A saliency-based search mechanism for overt and covert shifts of visual attention. Vision Research 40(10-12), 1489–1506 (2000)
Cheoi, K.: Visual attention system based on bottom-up theory of human visual attention. Doctorate Thesis of Yonsei University (2002)
Lee, J.: Selective Visual Attention System Based on Motion Information for Active Vision System. Master Thesis of Korea University (2008)
Itti, L., Koch, C.: Computational Modeling of visual attention. Nature Reviews Neuroscience 2(3), 194–203 (2001)
Treisman, A.M., Gelade, G.: A Feature-Integration Theory of Attention. Cognitive Psychology 12, 97–136 (1980)
Itti, L., Koch, C., Niebur, E.: A Model of Saliency-Based Visual Attention for Rapid Scene Analysis. IEEE Transaction on Pattern Analysis and Machine Intelligence 20(11) (1998)
Yu, Y., Mann, G., Gosine, R.G.: An Object-based Visual Attention Model for Robots. In: IEEE International Conference on Robotics and Automation Pasadena, CA, USA, May 19-23 (2008)
Xiao, J., Cai, C., Ding, M., Zhou, C.: Application of a Novel Target Region Extraction Model Based on Object-accumulated Visual Attention Mechanism. In: Fourth International Conference on Natural Computation, pp. 116–120 (2008)
Begum, M., Karray, F., Mann, G.K.I., Gosine, R.G.: A Probabilistic Model of Overt Visual Attention for Cognitive Robots. IEEE Transactions on Systems Man and Cybernetics-Part B: Cybernetics, 1305–1318 (2010)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2011 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Choi, B.G., Cheoi, K.J. (2011). Development of a Biologically Inspired Real-Time Spatiotemporal Visual Attention System. In: Nguyen, N.T., Kim, CG., Janiak, A. (eds) Intelligent Information and Database Systems. ACIIDS 2011. Lecture Notes in Computer Science(), vol 6591. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-20039-7_42
Download citation
DOI: https://doi.org/10.1007/978-3-642-20039-7_42
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-20038-0
Online ISBN: 978-3-642-20039-7
eBook Packages: Computer ScienceComputer Science (R0)