Abstract
Models of real-world objects and actions for use in graphics, Virtual and augmented reality and related fields can only be obtained automatically through the use of visual data and particularly video. This paper introduces a new framework for integrating multiple synchronized videos of an object or a scene captured by a calibrated configuration of cameras into a unified space-time description. Given videos of a rigidly-moving object, it is a first step for dynamic descriptions and model building to recover the three-dimensional (3D) motion of the object which consists of a rotation and a translation at each instant. In this paper we extend the structure-from-motion framework for moving cameras in a static world to static multi-camera configurations consisting of large numbers of cameras observing a moving world. This new framework integrates the information from all the cameras into a single model. Using all the given motion constraints at once, it does not suffer from the intrinsic ambiguities of single-camera approaches. We show this by arguments based on the distribution of camera locations in space. Finally, we present a new algorithm that implements this new framework and computes the motion and structure of a rigidly moving object observed by 50 calibrated cameras.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
K. Daniilidis and M. Spetsakis. Understanding noise sensitivity in structure from motion. In Y. Aloimonos, editor, Visual Navigation: From Biological Systems to Unmanned Ground Vehicles, chapter 4. Lawrence Erlbaum, 1996.
C. Fermüller and Y. Aloimonos. Ämbiguity in structure from motion: Sphere versus plane. Int. Journal of Computer Vision, 28(2)137–154, 1998.
K. Hanna. Direct multi-resolution estimation of ego-motion and structure from motion. In MOTION91, pages 156–162. 1991.
R. Hartley and A. Zisserman. Multiple View Geometry. Cambridge University Press, 2000.
B.K.P. Horn and E.J. Weldon, Jr.Direct methods for recovering motion. Int. Journal of Computer Vision. 2:51–76, 1988.
R. Koch, M. Pollefeys, B. Heigl, L. VanGool, and H. Niemann. Calibration of hand-held camera sequences for plenoptic modeling. In ICCV99, pages 585-591, 1999.
S. J. Maybank. Theory of Reconstruction from Image Motion. Springer, 1993.
P. Rander, P. Narayanan, and T. Kanade.Recovery of dynamic scene structure from multiple image sequences..In MSFIIS96, pages 305-312, 1996.
S. Seitz and C. Dyer. Photorealistic scene reconstruction by voxel coloring. IJCV, 35(2)1–23, November 1999.
L. Tai and R. Jain. 3d video generation with multiple perspective camera views. In ICIP97, pages I:9-xx, 1997.
S. Weik, J. Wmgbermühle, and W. Niem. Automatic creation of flexible antropomorphic models for 3d videoconferencing. In Proceedings of Computer Graphics International CGI, Hanover, Germany, June 1998.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2000 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Neumann, J., Fermüller, C., Aloimonos, Y. (2000). A New Framework for Multi-camera Structure from Motion. In: Sommer, G., Krüger, N., Perwass, C. (eds) Mustererkennung 2000. Informatik aktuell. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-59802-9_10
Download citation
DOI: https://doi.org/10.1007/978-3-642-59802-9_10
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-67886-1
Online ISBN: 978-3-642-59802-9
eBook Packages: Springer Book Archive