Skip to main content

Appearance-Based Virtual-View Generation for Fly Through in a Real Dynamic Scene

  • Conference paper
Data Visualization 2000

Part of the book series: Eurographics ((EUROGRAPH))

  • 268 Accesses

Abstract

We present appearance-based Virtual view generation which allows viewers to fly through a real dynamic scene. The scene is captured by synchronized multiple cameras. Arbitrary views are generated by interpolating two original camera-view images near the given viewpoint. The quality of the generated synthetic view is determined by the precision, consistency and density of correspondences between the two images. All or most of previous work that uses interpolation extracts the correspondences from these two images. However, not only is it difficult to do so reliably (the task requires a good stereo algorithm), but also the two images alone sometimes do not have enough information, due to problems such as occlusion. Instead, we take advantage of the fact that we have many views, from which we can extract much more reliable and comprehensive 3D geometry of the scene as a 3D model. The dense and precise correspondences between the two images, to be used for interpolation, are derived from this constructed 3D model. Our method of 3D modeling from multiple images uses the Multiple Baseline Stereo method and Shape from Silhouette method.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Chen, S., Williams, L.:View Interpolation for Image Synthesis. Proc. of SIGGRAPH’ 93. (1982) 279-288

    Google Scholar 

  2. Debevec, P., Taylor, C., Malik, J.: Modeling and Rendering Architecture from Photographs: A Hybrid Geometry and Image-Based Approach. Proc. of SIGGRAPH’ 96. (1996)

    Google Scholar 

  3. Avidan S., Shashua, A.: Novel View Synthesis by Cascading Trilinear Tensors. IEEE TVCG.Vol.4. No.4 (1998) 293–306

    Google Scholar 

  4. Potmesil, M.: Generating Octree Models of 3D Objects from Their Silhouettes in a Sequence of Images. Computer Vision, Graphics and Image Processing. 40 (1987) 277–283

    Article  Google Scholar 

  5. Gortier, S.J., Grzeszczuk, R., Szeliski, R., Cohen, M.F. The Lumigraph. Proc. of SIGGRAPH’96. (1996)

    Google Scholar 

  6. Chein, C.H., Aggarawal, J.K.: Identification of 3D Objects from Multiple Silhouettes using Quadtrees / Octrees. Computer Vision, Graphics and Image Processing. 36 (1986) 100–113

    Article  Google Scholar 

  7. Kanade, T., Rander, P.W., Narayanan, P.J.: Virtualized Reality: Constructing Virtual Worlds from Real Scenes. IEEE Multimedia. Vol.4. No.1 (1997)

    Google Scholar 

  8. Kanade, T., Saito, H., Vedula, S.: The 3D Room: Digitizing Time-Varying 3D Events by Synchronized Multiple Video Streams. CMU-RI-TR-98-34 (1998)

    Google Scholar 

  9. Katayama, A., Tanaka, K., Oshino, T., Tamura, H.:A Viewpoint Dependent Stereoscopic Display Using Interpolation of Multi-Viewpoint Images. SPIE Proc. Vol.2409. Stereoscopic Displays and Virtual Reality Systems II (1995) 11-20.

    Google Scholar 

  10. Levoy, M., Hanrahan, P.:Light Field Rendering. Proc. of SIGGRAPH’96 (1996)

    Google Scholar 

  11. Okutomi, M., Kanade, T.: A Multiple-Baseline Stereo.IEEE Trans, on PAMI. Vol. 15. No.4 (1993) 353–363

    Article  Google Scholar 

  12. Narayanan, P.J., Rander, P.W., Kanade, T.: Constructing Virtual Worlds Using Dense Stereo. Proc. ICCV’98 (1998)

    Google Scholar 

  13. Seitz, S.M., Dyler, C.R: View Morphing. Proc. of SIGGRAPH’96 (1996) 21-30

    Google Scholar 

  14. Tsai, R.: A Versatile Camera Calibration Technique for High-Accuracy 3D Machine Vision Metrology Using Off-The-Shelf TV Cameras and Lenses. IEEE J. of Robotics and Automation RA-3. 4. (1987) 323–344

    Article  Google Scholar 

  15. Vedula, S., Rander, P.W., Saito, H., Kanade, T.:Modeling, Combining and Rendering Dynamic Real-World Events from Image Sequences. Proc. 4th Conf. Virtual Systems and Multimedia. Vol.1 (1998) 326–332

    Google Scholar 

  16. Wheeler, M.D., Sato,Y., Ikeuchi, K.: Consensus Surfaces for Modeling 3D Objects from Multiple Range Images. DARPA Image Understanding Workshop (1997)

    Google Scholar 

  17. Faugeras, O., Laveau, S., Robert, L., Csurka, G., Zeller, C.: 3-D Reconstruction of Urban Scenes from Sequences of Images. INRIA Technical Report. No.2572 (1995

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2000 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Baba, S., Saito, H., Vedula, S., Cheung, K.M., Kanade, T. (2000). Appearance-Based Virtual-View Generation for Fly Through in a Real Dynamic Scene. In: de Leeuw, W.C., van Liere, R. (eds) Data Visualization 2000. Eurographics. Springer, Vienna. https://doi.org/10.1007/978-3-7091-6783-0_18

Download citation

  • DOI: https://doi.org/10.1007/978-3-7091-6783-0_18

  • Publisher Name: Springer, Vienna

  • Print ISBN: 978-3-211-83515-9

  • Online ISBN: 978-3-7091-6783-0

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics