Abstract
Current visual surveillance system usually includes multiple cameras to monitor the activities of targets over a large area. An important issue for the guard or user using the system is to understand a series of events occurring in the environment, for example to track a target walking across multiple cameras. Opposite to the traditional systems switching the camera view from one to another directly, we propose a novel approach to egocentric view transition, which synthesizes the virtual views during the period of switching cameras, to ease the mental effort for users to understand the events. An important property of our system is that it can be applied to the situations of where the view fields of transition cameras are not close enough or even exclusive. Such situations have never been taken into consideration in the state-of-the-art view transition techniques, to our best knowledge.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Chen, K.W., Lai, C.C., Hung, Y.P., Chen, C.S.: An Adaptive Learning Method for Target Tracking across Multiple Cameras. In: CVPR (2008)
Debevec, P.E., Taylor, C.J., Malik, J.: Modeling and Rendering Architecture from Photographs: A Hybrid Geometry- and Image-Based Approach. In: SIGGRAPH (1996)
Finke, R.A.: Principles of Mental Imagery. MIT Press, Cambridge (1989)
Girgensohn, A., Kimber, D., Vaughan, J., Yang, T., Shipman, F., Turner, T., Rieffel, E., Wilcox, L., Chen, F., Dunnigan, T.: DOTS: Support for Effective Video Surveillance. In: MULTIMEDIA (2007)
Hsiao, C.H., Huang, W.C., Chen, K.W., Chang, L.W., Hung, Y.P.: Generating Pictorial-Based Representation of Mental Image for Video Monitoring. In: IUI (2009)
Horprasert, T., Harwood, D., Davi, L.: A Statistical Approach for Real-Time Robust Background Subtraction and Shadow Detection. In: FRAME-RATE Workshop (1999)
Haan, G., Scheuer, J., Vries, R., Post, F.H.: Egocentric Navigation for Video Surveillance in 3D Virtual Environments. In: IEEE Symposium on 3D User Interfaces (2009)
Hartley, R.I., Zisserman, A.: Multiple View Geometry, 2nd edn. Cambridge University Press, Cambridge (2004)
Katkere, A., Moezzi, S., Kuramura, D.Y., Kelly, P., Jain, R.: Towards Video-Based Immersive Environments. Multimedia System 5(2), 69–85 (1997)
Kanade, T., Narayanan, P., and Rander, P.: Virtualized Reality: Concept and Early Results. Tech. Rep. CMU-CS-95-153 (1995)
Levenberg, K.: A method for the solution of certain problems in least squares. Quarterly Applied Math. 2, 164–168 (1944)
Lei, B., Hendriks, E.: Real-Time Multi-Step View Reconstruction for a Virtual Teleconference System. EURASIP J. Appl. Signal Process 2002(10), 1067–1088 (2002)
Neumann, U., You, S., Hu, J., Jiang, B., Lee, J.: Augmented Virtual Environment (AVE): Dynamic Fusion of Imagery and 3d Models. In: IEEE Virtual Reality (2003)
Ohta, Y., Kitahara, I., Kameda, Y., Ishikawa, H., Koyama, T.: Live 3D Video in Soccer Stadium. IJCV 75(1), 173–187 (2007)
Palmer, S.: Vision Science: Photons to Phenomenology. MIT Press, Cambridge (1999)
Reeves, W.T.: Particle Systems - a Technique for Modeling a Class of Fuzzy Objects. ACM Transactions on Graphics 2, 91–108 (1983)
Sawhney, H.S., Arpa, A., Kumar, R., Samarasekera, S., Aggarwal, M., Hsu, S., Nister, D., Hanna, K.: Video Flashlights: Read Time Rendering of Multiple Videos for Immersive Model Visualization. In: EGRW (2002)
Seitz, S., Dyer, C.: View Morphing. In: SIGGRAPH (1996)
Stauffer, C., Grimson, W.E.L.: Learning Patterns of Activity using Real-Time Tracking. IEEE Transactions on PAMI 22(8), 747–757 (2000)
Segal, M., Korobkin, C., Widenfelt, R., Foran, J., Haeberli, P.: Fast Shadows and Lighting Effects Using Texture Mapping. In: SIGGRAPH (1992)
Snavely, N., Seitz, S.M., Szeliski, R.: Photo Tourism: Exploring Photo Collections in 3d. In: SIGGRAPH (2006)
Thorndyke, P., Hayes-Roth, B.: Differences in Spatial Knowledge Acquired from Maps and Navigation. Cognitive Psychology 14(4), 560–589 (1982)
Welch, G., Bishop, G.: An introduction to the kalman filter. Chapel Hill, NC, USA, Tech. Rep (1995)
Wang, Y., Krum, D.M., Coelho, E.M., Bowman, D.A.: Contextualized Videos: Combining Videos with Environment Models to Support Situational Understanding. IEEE TVCG 13(6), 1568–1575 (2007)
Zhang, Z.: A Flexible New Technique for Camera Calibration. IEEE Transactions on PAMI 22, 1330–1334 (2000)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2011 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Chen, KW., Lee, PJ., Hung, YP. (2011). Egocentric View Transition for Video Monitoring in a Distributed Camera Network. In: Lee, KT., Tsai, WH., Liao, HY.M., Chen, T., Hsieh, JW., Tseng, CC. (eds) Advances in Multimedia Modeling. MMM 2011. Lecture Notes in Computer Science, vol 6523. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-17832-0_17
Download citation
DOI: https://doi.org/10.1007/978-3-642-17832-0_17
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-17831-3
Online ISBN: 978-3-642-17832-0
eBook Packages: Computer ScienceComputer Science (R0)