skip to main content
10.1145/3425577.3425587acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicccvConference Proceedingsconference-collections
research-article

An Efficient and Dynamical Way for Local Feature Extraction on Point Cloud

Published:23 January 2021Publication History

ABSTRACT

This paper presents a flexible module that utilizes the 3D position attention mechanism to extract contextual features from local regions of point cloud. The key point is to create an effective representation of local features. Due to the irregularity of point cloud, previous algorithms for point cloud processing have not fully explored how to enhance the extraction of local features. Inspired by the position attention mechanism in the 2D image segmentation algorithm, we propose a Point Attention Graph (PAG) module that can be used to improve the fusion of local features and make it better and faster. The PAG module uses the point attention mechanism to adaptively calculate the interaction between all nodes of the local graph. It can efficiently definite the relations of local points to enhance the performance of feature extraction both in accuracy and time efficiency, especially compared with some related models like PointWEB. Experiments show that our method can be effectively applied to semantic segmentation datasets.

References

  1. Table 2: Runtime comparison (4×4096 points of 5 cycles on S3DIS He, Kaiming, "Deep residual learning for image recognition." Proceedings of the IEEE conference on computer vision and pattern recognition. 2016.Google ScholarGoogle Scholar
  2. Ronneberger, Olaf, Philipp Fischer, and Thomas Brox. "U-net: Convolutional networks for biomedical image segmentation." International Conference on Medical image computing and computer-assisted intervention. Springer, Cham, 2015.Google ScholarGoogle Scholar
  3. Simonyan, Karen, and Andrew Zisserman. "Very deep convolutional networks for large-scale image recognition." arXiv preprint arXiv:1409.1556 (2014).Google ScholarGoogle Scholar
  4. Qi, Charles R., "Pointnet: Deep learning on point sets for 3d classification and segmentation." Proceedings of the IEEE conference on computer vision and pattern recognition. 2017.Google ScholarGoogle Scholar
  5. Qi, Charles Ruizhongtai, "Pointnet++: Deep hierarchical feature learning on point sets in a metric space." Advances in neural information processing systems. 2017.Google ScholarGoogle Scholar
  6. Jiang, Mingyang, "Pointsift: A sift-like network module for 3d point cloud semantic segmentation." arXiv preprint arXiv:1807.00652 (2018).Google ScholarGoogle Scholar
  7. Wang, Yue, "Dynamic graph cnn for learning on point clouds." ACM Transactions on Graphics (TOG) 38.5 (2019): 1-12.Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Zhao, Hengshuang, "PointWeb: Enhancing local neighborhood features for point cloud processing." Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2019.Google ScholarGoogle Scholar
  9. You, Yurong, "Pseudo-lidar++: Accurate depth for 3d object detection in autonomous driving." arXiv preprint arXiv:1906.06310 (2019).Google ScholarGoogle Scholar
  10. Wang, Chen, "Densefusion: 6d object pose estimation by iterative dense fusion." Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2019.Google ScholarGoogle Scholar
  11. Qi, Charles R., "Frustum pointnets for 3d object detection from rgb-d data." Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2018.Google ScholarGoogle Scholar
  12. Peng, Sida, "Pvnet: Pixel-wise voting network for 6dof pose estimation." Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2019.Google ScholarGoogle Scholar
  13. Joo, Hanbyul, Tomas Simon, and Yaser Sheikh. "Total capture: A 3d deformation model for tracking faces, hands, and bodies." Proceedings of the IEEE conference on computer vision and pattern recognition. 2018.Google ScholarGoogle Scholar
  14. Huang, Haibin, "Learning local shape descriptors from part correspondences with multiview convolutional networks." ACM Transactions on Graphics (TOG) 37.1 (2017): 1-14.Google ScholarGoogle Scholar
  15. Zhou, Yin, and Oncel Tuzel. "Voxelnet: End-to-end learning for point cloud based 3d object detection." Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2018.Google ScholarGoogle Scholar
  16. Liu, Baoyuan, "Sparse convolutional neural networks." Proceedings of the IEEE conference on computer vision and pattern recognition. 2015.Google ScholarGoogle Scholar
  17. Li, Yangyan, "Pointcnn: Convolution on x-transformed points." Advances in neural information processing systems. 2018.Google ScholarGoogle Scholar
  18. Landrieu, Loic, and Martin Simonovsky. "Large-scale point cloud semantic segmentation with superpoint graphs." Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2018.Google ScholarGoogle Scholar
  19. Wu, Wenxuan, Zhongang Qi, and Li Fuxin. "Pointconv: Deep convolutional networks on 3d point clouds." Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2019.Google ScholarGoogle Scholar
  20. Fu, Jun, "Dual attention network for scene segmentation." Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2019.Google ScholarGoogle Scholar
  21. Zhang, Han, "Self-attention generative adversarial networks." arXiv preprint arXiv:1805.08318 (2018).Google ScholarGoogle Scholar
  22. Wang, X., He, J., Ma, L.: Exploiting local and global structure for point cloud se- mantic segmentation with contextual point representations. In: Advances in Neural Information Processing Systems. pp. 4573–4583 (2019)Google ScholarGoogle Scholar
  23. Armeni, I., Sener, O., Zamir, A.R., Jiang, H., Brilakis, I., Fischer, M., Savarese, S.: 3d semantic parsing of large-scale indoor spaces. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. pp. 1534–1543 (2016)Google ScholarGoogle ScholarCross RefCross Ref
  24. Tchapmi, L., Choy, C., Armeni, I., Gwak, J., Savarese, S.: Segcloud: Semantic segmentation of 3d point clouds. In: 2017 International Conference on 3D Vision (3DV). pp. 537–547. IEEE (2017)Google ScholarGoogle Scholar
  25. Wang, S., Suo, S., Ma, W.C., Pokrovsky, A., Urtasun, R.: Deep parametric continuous convolutional neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. pp. 2589–2597 (2018)Google ScholarGoogle ScholarCross RefCross Ref

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image ACM Other conferences
    ICCCV '20: Proceedings of the 3rd International Conference on Control and Computer Vision
    August 2020
    114 pages
    ISBN:9781450388023
    DOI:10.1145/3425577

    Copyright © 2020 ACM

    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    • Published: 23 January 2021

    Permissions

    Request permissions about this article.

    Request Permissions

    Check for updates

    Qualifiers

    • research-article
    • Research
    • Refereed limited
  • Article Metrics

    • Downloads (Last 12 months)4
    • Downloads (Last 6 weeks)0

    Other Metrics

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format .

View HTML Format