Abstract
This paper addresses active image capturing and dynamic scene visualization by Cooperative Distributed Vision (CDV, in short). The concept of CDV was proposed by our five years project starting from 1996. From a practical point of view, the goal of CDV is summarized as follows: Embed in the real world a group of network-connected Observation Stations (real time video image processor with active camera(s)) and mobile robots with vision. And realize 1) wide-area dynamic scene understanding and 2) versatile scene visualization. Applications of CDV include real time wide-area surveillance, remote conference and lecturing systems, interactive 3D TV and intelligent TV studio, navigation of (non-intelligent) mobile robots and disabled people, cooperative mobile robots, and so on. In this paper, we first define the framework of CDV and give a brief retrospective view of the computer vision research to show the background of CDV. Then we present technical research results so far obtained: 1) fixed viewpoint pan-tilt-zoom camera for wide-area active imaging, 2) moving object detection and tracking for reactive image acquisition, 3) multi-viewpoints object imaging by cooperative observation stations, and 4) scenario-based cooperative camera-work planning for dynamic scene visualization. Prototype systems demonstrate the effectiveness and practical utilities of the proposed methods.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Matsuyama, T.: Cooperative Distributed Vision-Dynamic Integration of Visual Perception, Action, and Communication-, Proc. of Image Understanding Workshop, Monterey CA, 1998.11
Aloimonos, Y. (ed.): Special Issue on Purposive, Qualitative, Active Vision, CVGIP: Image Understanding, Vol.56 No.1, 1992.
Aloimonos, Y. (ed.): Active Perception, Lawrence Erlbaum Associates Publisher, 1993
Yagi Y. and Yachida M.: Real-Time Generation of Environmental Map and Obstacle Avoidance Using Omnidirectional Image Sensor with Conic Mirror, Prof. of CVPR, 160–165, 1991.
Yamazawa K., Yagi Y. and Yachida M.: Obstacle Detection with Omnidirectional Image Sensor HyperOmni Vision, Proc. of ICRA, pp.1062–1067, 1995.
Peri V. N. and Nayar S. K.: Generation of Perspective and Panoramic Video from Omnidirectional Video, Proc. of IUW, pp. 243–245, 1997.
Murray, D. and Basu, A.: Motion Tracking with an Active Camera, IEEE Trans. of PAMI, Vol. 16, No. 5, pp. 449–459, 1994.
Wada T. and Matsuyama T.: Appearance Sphere: Background Model for Pan-Tilt-Zoom Camera, Proc. of ICPR, Vol. A, pp. 718–722, 1996.
Lavest, J.M., Delherm, C., Peuchot, B, and Daucher, N.: Implicit Reconstruction by Zooming, Computer Vision and Image Understanding, Vol.66, No.3, pp.301–315, 1997.
Hall R.: Hybrid Techniques for Rapid Image Synthesis, in Image Rendering Tricks (Whitted T. and Cook R. eds.), Course Notes 16 for SIGGRAPH’86, 1986.
Greene N.: Environment Mapping and Other Applications of World Projections, CGA, 6(11), pp. 21–29, 1986.
Chen S.E.: QuickTime VR-An Image-Based Approach to Virtual Environment Navigation, Proc. of SIGGRAPH’95, pp. 29–38, 1995.
Nakai, H.: Robust Object Detection Using A-Posteriori Probability, Tech. Rep. of IPSJ, SIG-CV90-1, 1994 (in Japanese).
Grimson, E.: A Forest of Sensors, Proc. of VSAM Workshop, 1997.
Davis, L.: Visual Surveillance and Monitoring, Proc. of VSAM Workshop, 1997.
Habe, H., Ohya, T., and Matsuyama, T.: A Robust Background Subtraction Method for Non-Stationary Scenes, Proc. of MIRU’98, Vol.1, pp.467–472, 1998 (in Japanese).
Yamaashi, K., Cooperstock, J.R., Narine, T., and Buxton, W.: Beating the Limitations of Camera-Monitor Mediated Telepresence with Extra Eyes, Proc. of CHI, pp.50–57, 1996.
Hiura, S. and Matsuyama, T.: Depth Measurement by the Multi-Focus Camera, Proc. of CVPR, pp.953–959, 1998
Mikoshi, Y.: 3D Image Measurement Based on Planes, Master Thesis, Kyoto University, 1998 (in Japanese)
Arijon, D.: Grammar of the Film Language, Focal Press Ltd., London, 1976
He, L., Cohen, M.F., and Salesin, D.H.: The Virtual Cinematographer: A Paradigm for Automatic Real-Time Camera Control and Directing, SIGGRAPH’96, pp.217–224, 1996.
Christianson,D.B., Anderson, S.E., He, L., Weld, D.S., Cohen, M.F., and Salesin, D.H.: Declarative Camera Control for Automatic Cinematography, Proceedings of AAAI’ 96, pp.148–155, 1996.
Mase, K., Pinhanez, C.S., and Bobick, A.F.: Scripting Method Based on Temporal Intervals for Designing Interactive Systems, Trans. of IPSJ, Vol.39, No.5, pp.1403–1413, 1998 (in Japanese).
Allen, J.F.: Towards a General Theory of Action and Time, Artificial Intelligence, Vol.23, pp.123–154, 1984.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 1999 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Matsuyama, T., Wada, T., Tokai, S. (1999). Active Image Capturing and Dynamic Scene Visualization by Cooperative Distributed Vision. In: Nishio, S., Kishino, F. (eds) Advanced Multimedia Content Processing. AMCP 1998. Lecture Notes in Computer Science, vol 1554. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-48962-2_18
Download citation
DOI: https://doi.org/10.1007/3-540-48962-2_18
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-65762-0
Online ISBN: 978-3-540-48962-7
eBook Packages: Springer Book Archive