Abstract
A method is presented for extracting object information from an image sequence taken by a static monocular camera. The method was developed towards a low computational complexity in order to be used in real-time surveillance applications. Our approach makes use of both intensity and edge information of each frame and works efficiently in an indoor environment. It consists of two major parts: background processing and foreground extraction. The background estimation and updating makes the object detection robust to environment changes like illumination changes and camera jitter. The fusion of intensity and edge information allows a more precise estimation of the position of the different foreground objects in a video sequence. The result obtained are quite reliable, under a variety of environmental conditions.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Sebe, I.O.: Object-Tracking I|Using Multiple Constraints, EE392J. Digital Video Processing, Standford University (2002)
Toyama, K., Krumm, J., Brumitt, B., Meyers, B.: Wallflower: Principles and Practice of Background Maintenance. Microsoft Research Redmond (1999)
Jung, Y.-K., Lee, K.-W., Ho, Y.-S.: Content-Based Event Retrieval Using Semantic Scene Interpretation for Automated Traffic Surveillance. IEEE Transactions on Intelligent Transpotation Systems 2(3) (September 2001)
Cucchiara, R., Grana, C., Piccardi, M., Prati, A.: Detecting Moving Objects, Ghosts, and Shadows in Video Streams. IEEE Transactions on Pattern Analysis and Machine Intelligence 25(10) (October 2003)
Cheung, S.-C.S., Kamath, C.: Robust techniques for background subtraction in urban traffic video, Center for Applied Scientiffic Computing Lawrence Livermore National Laboratory, Livermore, Canada
Liedtke, C.E.: Mustererkennung Vorlesungs Manuskript. Universitaet Hannover, Hannover (1999)
Matrox Image Libraty 7.5 User Guide, Matrox Electronic Systems Ltd. (2003)
Jarbi, S., Rosenfeld, A.: Detection and Location of People. In: Video Images Using Adaptive Fusion of Color and Edge Information. Department of Computer Science, George Mason University, Fairfax
Leykin, A., Cutzu, F.: Differences of edge properties in photographs and paintings. Dept. of Computer Science. Indiana University, Bloomington, USA
Xu, F., Fujimura, K.: Human Detection Using Depth and Gray Images. Honda Research Institute USA. Ohio State University (2003)
CAVIAR project (Context Aware Vision using Image-based Active Recognition ), http://homepages.inf.ed.ac.uk/rbf/CAVIAR
OPTAG project (Improving airport Efficiency, Security and Passenger Flow by Enhanced Passenger Monitoring FP6-2002-Aero No.502858)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2005 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Markopoulos, N., Zervakis, M. (2005). Design of a Hybrid Object Detection Scheme for Video Sequences. In: Blanc-Talon, J., Philips, W., Popescu, D., Scheunders, P. (eds) Advanced Concepts for Intelligent Vision Systems. ACIVS 2005. Lecture Notes in Computer Science, vol 3708. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11558484_32
Download citation
DOI: https://doi.org/10.1007/11558484_32
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-29032-2
Online ISBN: 978-3-540-32046-3
eBook Packages: Computer ScienceComputer Science (R0)