skip to main content
10.1145/3297097.3297098acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicraiConference Proceedingsconference-collections
research-article

Depth Image Super-Resolution with Semantic and RGB Images Using CNN

Published:17 November 2018Publication History

ABSTRACT

Depth images acquired by consumer depth sensors, such as Kinect and ToF, usually are of low resolution and insufficient quality. This limits the application of these depth sensors. Therefore, depth map enhancement is essential to its application. Most existing depth map super-resolution methods employ an RGB image of the same scene in the depth image as a guidance to up-sample the depth map. However, due to part of edges in RGB image do not occurrence in depth image, such as texture in RGB image, most existing methods will introduce a problem of texture-copy in these areas. To address this problem, we propose an approach that introduce semantic information of RGB image. On the other hand, existing methods rely on various kinds of explicit filter construction or hand-designed objective function. It is thus difficult to understand, improve, and accelerate them in a coherent framework. In this paper we use a learning-based approach to construct a joint filter based on Convolutional Neural Networks. In contrast to existing methods that consider only the RGB guidance image, our method can suppress the texture-copy problem. We validate the effectiveness of the proposed method through extensive comparisons with state-of-the-art methods on NYU v2 dataset. Experiment results show that our method suppress the texture-copy problem.

References

  1. Aodha, O. M., Campbell, N. D. F., Nair, A., and Brostow, G. J. 2012. Patch based synthesis for single depth image super-resolution. In: European Conference on Computer Vision. pp. 71--84. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Chan, D., Buisman, H., Theobalt, C., and Thrun, S. 2008. A noise-aware filter for real-time depth up-sampling. In: The Workshop on Multi-Camera and Multi-Modal Sensor Fusion Algorithms and Applications.Google ScholarGoogle Scholar
  3. Diebel, J. and Thrun, S. 2005. An application of markov random fields to range sensing. Advances in Neural Information Processing Systems. pp. 291--298. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Dong, C., Chen, C. L., He, K., and Tang, X. 2016. Image super-resolution using deep convolutional networks. IEEE Transactions on Pattern Analysis and Machine Intelligence. 38(2), 295--307. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Ferstl, D., Reinbacher, C., Ranftl, R., Ruether, M., and Bischof, H. 2014. Image guided depth up-samjpling using anisotropic total generalized variation. In: IEEE International Conference on Computer Vision. pp. 993--1000. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Ferstl, D., Ruther, M., and Bischof, H. 2015. Variational depth super-resolution using example-based edge representations. In: IEEE International Conference on Computer Vision. pp. 513--521. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Gregor, K. and Lecun, Y. 2010. Learning fast approximations of sparse coding. In: International Conference on International Conference on Machine Learning. pp. 399--406. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Ham, B., Cho, M., and Ponce, J. 2015. Robust image filtering using joint static and dynamic guidance. In: Computer Vision and Pattern Recognition. pp. 4823--4831.Google ScholarGoogle Scholar
  9. He, K., Sun, J., and Tang, X. 2010. Guided image filtering. In: European Conference on Computer Vision. pp. 1--14.Google ScholarGoogle Scholar
  10. Hirschmuller, H. and Scharstein, D. 2007. Evaluation of cost functions for stereo matching. In: Computer Vision and Pattern Recognition, 2007. CVPR '07. IEEE Conference on. pp. 1--8.Google ScholarGoogle Scholar
  11. Kopf, J., Cohen, M. F., Lischinski, D., Uyttendaele, M. 2007. Joint bilateral up-sampling. In: Acm Siggraph. p. 96. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Li, Y., Huang, J. B., Ahuja, N., and Yang, M. H. 2016. Deep Joint Image Filtering. Springer International Publishing.Google ScholarGoogle Scholar
  13. Liu, M. Y., Tuzel, O., and Taguchi, Y. 2013. Joint geodesic up-sampling of depth images. In: Computer Vision and Pattern Recognition. pp. 169--176. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Mandal, S., Bhavsar, A., and Sao, A. K. 2014. Hierarchical example-based range-image super-resolution with edge-preservation. pp. 3867--3871.Google ScholarGoogle Scholar
  15. Osendorfer, C., Soyer, H., and Smagt, P.V. D. 2014. Image super-resolution with fast approximate convolutional sparse coding. Lecture Notes in Computer Science 8836, 250--257.Google ScholarGoogle ScholarCross RefCross Ref
  16. Park, J., Kim, H., Tai, Y.W., Brown, M. S., and Kweon, I. 2011. High quality depth map up-sampling for 3d-tof cameras. In: IEEE International Conference on Computer Vision. pp. 1623--1630. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Scharstein, D. and Pal, C. 2007. Learning conditional random fields for stereo. In: Computer Vision and Pattern Recognition, 2007. CVPR '07. IEEE Conference on. pp. 1--8.Google ScholarGoogle Scholar
  18. Silberman, N., Hoiem, D., Kohli, P., and Fergus, R. 2012. Indoor Segmentation and Support Inference from RGBD Images. Springer Berlin Heidelberg.Google ScholarGoogle Scholar
  19. Tomasi, C. and Manduchi, R. 1998. Bilateral filtering for gray and color images iccv. Proc ICCV p. 839. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. Vedaldi, A. and Lenc, K. 2014. Matconvnet: Convolutional neural networks for matlab pp. 689--692. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. Wang, Y., Ortega, A., Tian, D., and Vetro, A. 2014. A graph-based joint bilateral approach for depth enhancement pp. 885--889.Google ScholarGoogle Scholar
  22. Wang, Z., Liu, D., Yang, J., Han, W., and Huang, T. 2016. Deep networks for image super-resolution with sparse prior. In: IEEE International Conference on Computer Vision. pp. 370--378. Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. Yang, Q., Yang, R., Davis, J., and Nister, D. 2007. Spatial-depth super resolution for range images. In: IEEE Conference on Computer Vision and Pattern Recognition. pp. 1--8.Google ScholarGoogle Scholar

Index Terms

  1. Depth Image Super-Resolution with Semantic and RGB Images Using CNN

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in
      • Published in

        cover image ACM Other conferences
        ICRAI '18: Proceedings of the 4th International Conference on Robotics and Artificial Intelligence
        November 2018
        109 pages
        ISBN:9781450365840
        DOI:10.1145/3297097

        Copyright © 2018 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 17 November 2018

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article
        • Research
        • Refereed limited
      • Article Metrics

        • Downloads (Last 12 months)1
        • Downloads (Last 6 weeks)0

        Other Metrics

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader