Abstract
Projected texture based 3D sensing modalities are being increasingly used for a variety of 3D computer vision applications. However, these sensing modalities, exemplified by the Microsoft Kinect Sensor, suffer from severe drawbacks that hamper the quality of the range estimate output from the sensor. It is well known that the quality of reconstruction of the 3D projected texture for range estimation is a function of the material properties of objects in the image. Objects colored black, yellow or deep red often do not reflect the texture in a manner suitable for the detector to estimate the range values. Furthermore, shiny or highly reflective objects can also scatter the projected texture patterns. Objects with skewed surface orientation, occlusions, object self-shadows and intra-object mutual shadows, transparency and other factors also create problems with projected texture reconstruction. In order to alleviate these concerns, depth interpolation techniques have been used in the past. These techniques, however, create loss of depth structures crucial for segmentation and detection processes. In order to alleviate these concerns, we present a novel MRF based color- depth fusion algorithm which uses information from the RGB sensor of the Kinect and couples it with the depth content to produce fine structure, high fidelity depth maps. This algorithm can be implemented in hardware on the Kinect device, thereby improving the depth resolution, fidelity of the sensor while eliminating range errors and shadows.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Diebel, J., Thrun, S.: An application of Markov random fields to range sensing. In: Neural Information Processing Systems, NIPS 2005 (2005)
Dolson, J., et al.: Upsampling range data in dynamic environments. In: IEEE Computer Vision and Pattern Recognition, CVPR 2010 (2010)
Varadarajan, K.M., Vincze, M.: Real time depth diffusion for 3D Surface Reconstruction. In: International Conference on Image Processing, ICIP 2010 (2010)
Boykov, Y., Kolmogorov, V.: An Experimental Comparison of Min-Cut/Max-Flow Algorithms for Energy Minimization in Vision. IEEE Transactions on Pattern Analysis and Machine Intelligence 26(9), 1124–1137 (2004)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2013 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Varadarajan, K.M., Vincze, M. (2013). MRF Guided Anisotropic Depth Diffusion for Kinect Range Image Enhancement. In: Park, JI., Kim, J. (eds) Computer Vision - ACCV 2012 Workshops. ACCV 2012. Lecture Notes in Computer Science, vol 7729. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-37484-5_19
Download citation
DOI: https://doi.org/10.1007/978-3-642-37484-5_19
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-37483-8
Online ISBN: 978-3-642-37484-5
eBook Packages: Computer ScienceComputer Science (R0)