skip to main content
research-article
Public Access

DeepPCD: Enabling AutoCompletion of Indoor Point Clouds with Deep Learning

Published:07 July 2022Publication History
Skip Abstract Section

Abstract

3D Point Cloud Data (PCD) is an efficient machine representation for surrounding environments and has been used in many applications. But the measured PCD is often incomplete and sparse due to the sensor occlusion and poor lighting conditions. To automatically reconstruct complete PCD from the incomplete ones, we propose DeepPCD, a deep-learning-based system that reconstructs both geometric and color information for large indoor environments. For geometric reconstruction, DeepPCD uses a novel patch based technique that splits the PCD into multiple parts, approximates, extends, and independently reconstructs the parts by 3D planes, and then merges and refines them. For color reconstruction, DeepPCD uses a conditional Generative Adversarial Network to infer the missing color of the geometrically reconstructed PCD by using the color feature extracted from incomplete color PCD. We experimentally evaluate DeepPCD with several real PCD collected from large, diverse indoor environments and explore the feasibility of PCD autocompletion in various ubiquitous sensing applications.

References

  1. B.-S. Kim, P. Kohli, and S. Savarese, "3D Scene Understanding by Voxel-CRF," in IEEE International Conference on Computer Vision (ICCV), 2013.Google ScholarGoogle Scholar
  2. A. Vincent, "A 3D Perception System for the Mobile Robot Hilare," in IEEE International Conference on Robotics and Automation, 1986.Google ScholarGoogle Scholar
  3. T. Vieville, E. Clergue, R. Enciso, and H. Mathieu, "Experimenting with 3D Vision on a Robotic Head," Robotics and Autonomous Systems, vol. 14, no. 1, 1995.Google ScholarGoogle Scholar
  4. H. Zhang, G. Wang, Z. Lei, and J.-N. Hwang, "Eye in the Sky: Drone-Based Object Tracking and 3D Localization," in ACM International Conference on Multimedia, 2019.Google ScholarGoogle Scholar
  5. Y. Zeng, Y. Hu, S. Liu, J. Ye, Y. Han, X. Li, and N. Sun, "RT3D: Real-Time 3-D Vehicle Detection in LiDAR Point Cloud for Autonomous Driving," IEEE Robotics and Automation Letters, vol. 3, no. 4, 2018.Google ScholarGoogle Scholar
  6. K. C. Kwan and H. Fu, "Mobi3DSketch: 3D Sketching in Mobile AR," in ACM CHI, 2019.Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Y. Zhou and O. Tuzel, "VoxelNet: End-to-End Learning for Point Cloud Based 3D Object Detection," in IEEE/CVF CVPR, 2018.Google ScholarGoogle Scholar
  8. C. Gotsman, X. Gu, and A. Sheffer, "Fundamentals of Spherical Parameterization for 3D Meshes," in ACM SIGGRAPH, 2003.Google ScholarGoogle Scholar
  9. L. Linsen, "Point Cloud Representation," Karlsruhe Institute of Technology, Germany, Tech. Rep., 2001.Google ScholarGoogle Scholar
  10. X. Li, S. Du, G. Li, and H. Li, "Integrate Point-Cloud Segmentation with 3D LiDAR Scan-Matching for Mobile Robot Localization and Mapping," MDPI Sensors, vol. 20, no. 1, 2020.Google ScholarGoogle Scholar
  11. R. Radkowski, "Object Tracking with a Range Camera for Augmented Reality Assembly Assistance," Journal of Computing and Information Science in Engineering, vol. 16, no. 1, 2016.Google ScholarGoogle ScholarCross RefCross Ref
  12. Q. Wang and M.-K. Kim, "Applications of 3D Point Cloud Data in the Construction Industry: A Fifteen-Year Review from 2004 to 2018," Advanced Engineering Informatics, vol. 39, 2019.Google ScholarGoogle Scholar
  13. S. B. Walsh, D. J. Borello, B. Guldur, and J. F. Hajjar, "Data Processing of Point Clouds for Object Detection for Structural Engineering Applications," Computer-Aided Civil and Infrastructure Engineering, vol. 28, no. 7, 2013.Google ScholarGoogle Scholar
  14. D. A. White, "LIDAR, Point Clouds, and Their Archaeological Applications," in Mapping Archaeological Landscapes from Space, 2013.Google ScholarGoogle Scholar
  15. M. Labbe and F. Michaud, "RTAB-Map as an Open-Source Lidar and Visual Simultaneous Localization and Mapping Library for Large-Scale and Long-Term Online Operation," Journal of Field Robotics, vol. 36, no. 2, 2019.Google ScholarGoogle Scholar
  16. Y. Xiao and Y. Taguchi and V. R. Kamat, "Coupling Point Cloud Completion and Surface Connectivity Relation Inference for 3D Modeling of Indoor Building Environments," Journal of Computing in Civil Engineering, vol. 32, no. 5, 2018.Google ScholarGoogle Scholar
  17. A. Dai, C. Ruizhongtai Q., and M. Niessner, "Shape Completion using 3D-Encoder-Predictor CNNs and Shape Synthesis," in IEEE/CVF CVPR, 2017.Google ScholarGoogle Scholar
  18. W. Yuan, T. Khot, D. Held, C. Mertz, and M. Hebert, "PCN: Point Completion Network," in International Conference on 3D Vision (3DV), 2018.Google ScholarGoogle Scholar
  19. H. Xie, H. Yao, S. Zhou, J. Mao, S. Zhang, and W. Sun, "GRNet: Gridding Residual Network for Dense Point Cloud Completion," in European Conference on Computer Vision: Computer Vision - ECCV, 2020.Google ScholarGoogle Scholar
  20. M. Liu, L. Sheng, S. Yang, J. Shao, and S.-M. Hu, "Morphing and Sampling Network for Dense Point Cloud Completion," AAAI Conference on Artificial Intelligence, vol. 34, no. 07, 2020.Google ScholarGoogle Scholar
  21. A. X. Chang, T. Funkhouser, L. Guibas, P. Hanrahan, Q. Huang, Z. Li, S. Savarese, M. Savva, S. Song, H. Su, J. Xiao, L. Yi, and F. Yu, "ShapeNet: An Information-Rich 3D Model Repository," https://shapenet.org/, 2015.Google ScholarGoogle Scholar
  22. L. P. Tchapmi, V. Kosaraju, H. Rezatofighi, I. Reid, and S. Savarese, "TopNet: Structural Point Cloud Decoder," in IEEE/CVF CVPR, 2019.Google ScholarGoogle Scholar
  23. Y. Yang, C. Feng, Y. Shen, and D. Tian, "FoldingNet: Point Cloud Auto-Encoder via Deep Grid Deformation," in IEEE/CVF CVPR, 2018.Google ScholarGoogle Scholar
  24. C. R. Qi, L. Yi, H. Su, and L. J. Guibas, "PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space," in Advances in Neural Information Processing Systems (NIPS), 2017.Google ScholarGoogle Scholar
  25. A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly, J. Uszkoreit, and N. Houlsby, "An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale," in International Conference on Learning Representations (ICLR), 2021.Google ScholarGoogle Scholar
  26. H. Fan, H. Su, and L. Guibas, "A Point Set Generation Network for 3D Object Reconstruction from a Single Image," in IEEE/CVF CVPR, 2017.Google ScholarGoogle Scholar
  27. P. Xiang, X. Wen, Y.-S. Liu, Y.-P. Cao, P. Wan, W. Zheng, and Z. Han, "SnowflakeNet: Point Cloud Completion by Snowflake Point Deconvolution with Skip-Transformer," in Proceedings of the IEEE International Conference on Computer Vision (ICCV), 2021.Google ScholarGoogle Scholar
  28. S. Sur, "DeepPCD Project," 2022. [Online]. Available: https://syrex.cse.sc.edu/research/ubiquitous-sensing/deeppcd/Google ScholarGoogle Scholar
  29. H. Sobreira, C. M. Costa, I. Sousa, L. Rocha, J. Lima, P. C. Farias, P. Costa, and A. P. Moreira, "Map-matching algorithms for robot self-localization: A comparison between perfect match, iterative closest point and normal distributions transform," J. Intell. Robotics Syst., vol. 93, 2019.Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. Z. Liu, H. Tang, Y. Lin, and S. Han, "Point-Voxel CNN for Efficient 3D Deep Learning," in Advances in Neural Information Processing Systems (NIPS), 2019.Google ScholarGoogle Scholar
  31. W. Shi and R. Rajkumar, "Point-GNN: Graph Neural Network for 3D Object Detection in a Point Cloud," in IEEE/CVF CVPR, 2020.Google ScholarGoogle Scholar
  32. S. I. Kabanikhin, "Definitions and Examples of Inverse and Ill-posed Problems," Journal of Inverse and Ill-posed Problems, vol. 16, no. 4, 2008.Google ScholarGoogle ScholarCross RefCross Ref
  33. Y. Lecun, L. Bottou, Y. Bengio, and P. Haffner, "Gradient-based Learning Applied to Document Recognition," Proceedings of the IEEE, vol. 86, no. 11, 1998.Google ScholarGoogle Scholar
  34. Y. Liu, B. Fan, S. Xiang, and C. Pan, "Relation-Shape Convolutional Neural Network for Point Cloud Analysis," in IEEE/CVF CVPR, 2019.Google ScholarGoogle Scholar
  35. S. Ji, W. Xu, M. Yang, and K. Yu, "3D Convolutional Neural Networks for Human Action Recognition," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 35, no. 1, 2013.Google ScholarGoogle Scholar
  36. A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. Kaiser, and I. Polosukhin, "Attention is All You Need," in Advances in Neural Information Processing Systems (NIPS), 2017.Google ScholarGoogle Scholar
  37. X. Cao and K. Nagao, "Point Cloud Colorization Based on Densely Annotated 3D Shape Dataset," in MMM, 2019.Google ScholarGoogle Scholar
  38. J. Liu, S. Dai, and X. Li, "PCCN: Point Cloud Colorization Network," in IEEE International Conference on Image Processing (ICIP), 2019.Google ScholarGoogle Scholar
  39. T. Shinohara, H. Xiu, and M. Matsuoka, "Point2Color: 3D Point Cloud Colorization Using a Conditional Generative Network and Differentiable Rendering for Airborne LiDAR," in IEEE/CVF CVPR Workshops, 2021.Google ScholarGoogle Scholar
  40. M. Mirza and S. Osindero, "Conditional Generative Adversarial Nets," 2014. [Online]. Available: https://arxiv.org/abs/1411.1784Google ScholarGoogle Scholar
  41. I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, "Generative Adversarial Nets," in Advances in Neural Information Processing Systems (NIPS), 2014.Google ScholarGoogle ScholarDigital LibraryDigital Library
  42. AsusTek Computer Inc., "Zenfone AR: Go Beyond Reality," 2021. [Online]. Available: https://www.asus.com/us/Phone/ZenFone-ARZS571KL/Google ScholarGoogle Scholar
  43. IntRoLab, "Real-Time Appearance-Based Mappingy," 2021. [Online]. Available: http://introlab.github.io/rtabmap/Google ScholarGoogle Scholar
  44. P. Cignoni, M. Callieri, M. Corsini, M. Dellepiane, F. Ganovelli, and G. Ranzuglia, "MeshLab: an Open-Source Mesh Processing Tool," in Eurographics Italian Chapter Conference, 2008.Google ScholarGoogle Scholar
  45. I. Armeni, O. Sener, A. R. Zamir, H. Jiang, I. Brilakis, M. Fischer, and S. Savarese, "3D Semantic Parsing of Large-Scale Indoor Spaces," in IEEE/CVF CVPR, 2016.Google ScholarGoogle Scholar
  46. R. Bridson, "Fast Poisson Disk Sampling in Arbitrary Dimensions," in ACM SIGGRAPH 2007 Sketches, 2007.Google ScholarGoogle Scholar
  47. Open-Source, "PyTroch," 2021. [Online]. Available: https://pytorch.org/Google ScholarGoogle Scholar
  48. NVIDIA, "RTX A6000," 2021. [Online]. Available: https://www.nvidia.com/en-us/design-visualization/rtx-a6000/Google ScholarGoogle Scholar
  49. L. Li, F. Yang, H. Zhu, D. Li, Y. Li, and L. Tang, "An Improved RANSAC for 3D Point Cloud Plane Segmentation Based on Normal Distribution Transformation Cells," Remote Sensing, vol. 9, no. 5, 2017.Google ScholarGoogle Scholar
  50. Geodan, "Generate Synthetic Points to Fill Holes in Point Clouds," 2020. [Online]. Available: https://github.com/Geodan/fill-holespointcloudGoogle ScholarGoogle Scholar
  51. Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli, "Image Quality Assessment: From Error Visibility to Structural Similarity," IEEE Transactions on Image Processing, vol. 13, no. 4, 2004.Google ScholarGoogle Scholar
  52. Q. Yang, H. Chen, Z. Ma, Y. Xu, R. Tang, and J. Sun, "Predicting the perceptual quality of point cloud: A 3d-to-2d projection-based exploration," IEEE Transactions on Multimedia, vol. 23, 2021.Google ScholarGoogle Scholar
  53. M. A. Fischler and R. C. Bolles, "Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography," Commun. ACM, vol. 24, no. 6, 1981.Google ScholarGoogle ScholarDigital LibraryDigital Library
  54. M. Pauly, N. J. Mitra, J. Wallner, H. Pottmann, and L. J. Guibas, "Discovering Structural Regularity in 3D Geometry," in ACM SIGGRAPH, 2008.Google ScholarGoogle Scholar
  55. P. Speciale, M. R. Oswald, A. Cohen, and M. Pollefeys, "A Symmetry Prior for Convex Variational 3D Reconstruction," in European Conference on Computer Vision: Computer Vision - ECCV, 2016.Google ScholarGoogle Scholar
  56. Y. Li, A. Dai, L. Guibas, and M. NieBner, "Database-Assisted Object Retrieval for Real-Time 3D Reconstruction," Computer Graphics Forum, 2015.Google ScholarGoogle Scholar
  57. V. G. Kim, W. Li, N. J. Mitra, S. Chaudhuri, S. DiVerdi, and T. Funkhouser, "Learning Part-Based Templates from Large Collections of 3D Shapes," ACM Transactions on Graphics, vol. 32, no. 70, 2013.Google ScholarGoogle Scholar
  58. X. Zhang, Y. Feng, S. Li, C. Zou, H. Wan, X. Zhao, Y. Guo, and Y. Gao, "View-Guided Point Cloud Completion," in IEEE/CVF CVPR, 2021.Google ScholarGoogle Scholar
  59. X. Wang, J. Ang, Marcelo H, and G. H. Lee, "Point Cloud Completion by Learning Shape Priors," in IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2020.Google ScholarGoogle Scholar
  60. A. Dai, D. Ritchie, M. Bokeloh, S. Reed, J. Sturm, and M. Niebner, "ScanComplete: Large-Scale Scene Completion and Semantic Segmentation for 3D Scans," in IEEE/CVF CVPR, 2018.Google ScholarGoogle Scholar
  61. Open-Source, "Deploy Machine Learning Models on Mobile and IoT Devices," 2022. [Online]. Available: https://www.tensorflow.org/liteGoogle ScholarGoogle Scholar
  62. Open-Source, "PyTorch Mobile," 2022. [Online]. Available: https://pytorch.org/mobile/home/Google ScholarGoogle Scholar

Index Terms

  1. DeepPCD: Enabling AutoCompletion of Indoor Point Clouds with Deep Learning

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in

      Full Access

      • Published in

        cover image Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies
        Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies  Volume 6, Issue 2
        July 2022
        1551 pages
        EISSN:2474-9567
        DOI:10.1145/3547347
        Issue’s Table of Contents

        Copyright © 2022 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 7 July 2022
        Published in imwut Volume 6, Issue 2

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article
        • Research
        • Refereed
      • Article Metrics

        • Downloads (Last 12 months)116
        • Downloads (Last 6 weeks)18

        Other Metrics

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader