skip to main content
10.1145/3523150.3523177acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicmlscConference Proceedingsconference-collections
research-article

Progressive Multimodal Shape Generation via Contextual Part Reasoning

Published:13 April 2022Publication History

ABSTRACT

We present a contextual generative network for 3D shapes based on a conditional variational autoencoder, which learns a subspace of plausible complementary parts in the context of a partial shape. With the learned part subspace prior, which encodes bi-part spatial relations and geometry descriptions, a shape is generated via iterative “next part reasoning”, where a next part is sampled conditioned on a partial shape. Furthermore, our conditional subspace allows not just one, but a set of reasonable next parts to be generated, which adds controllability (e.g., via user selection) to the generative process. Our core idea of reasoning about next parts via conditional modeling offers a new way of understanding shape structures via part correlation modeling. Evaluations show the effectiveness of our approach and also the diversity of the generated shapes.

References

  1. Panos Achlioptas, Olga Diamanti, Ioannis Mitliagkas, and Leonidas Guibas. 2018. Learning Representations and Generative Models for 3D Point Clouds. arXiv preprint arXiv:1707.02392(2018).Google ScholarGoogle Scholar
  2. Angel X. Chang, Thomas Funkhouser, Leonidas J. Guibas, Pat Hanrahan, Qixing Huang, Zimo Li, Silvio Savarese, Manolis Savva, Shuran Song, Hao Su, Jianxiong Xiao, Li Yi, and Fisher Yu. 2015. ShapeNet: An Information-Rich 3D Model Repository. arXiv:1512.03012 [cs.GR] (2015).Google ScholarGoogle Scholar
  3. Siddhartha Chaudhuri, Daniel Ritchie, Jiajun Wu, Kai Xu, and Hao Zhang. 2020. Learning generative models of 3D structures. In Computer Graphics Forum, Vol. 39. Wiley Online Library, 643–666.Google ScholarGoogle Scholar
  4. Zhiqin Chen and Hao Zhang. 2019. Learning Implicit Fields for Generative Shape Modeling. In CVPR.Google ScholarGoogle Scholar
  5. Haoqiang Fan, Hao Su, and Leonidas Guibas. 2016. A point set generation network for 3D object reconstruction from a single image. arXiv preprint arXiv:1612.00603(2016).Google ScholarGoogle Scholar
  6. Yutong Feng, Yifan Feng, Haoxuan You, Xibin Zhao, and Yue Gao. 2019. MeshNet: Mesh neural network for 3D shape representation. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 33. 8279–8286.Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Rohit Girdhar, David F Fouhey, Mikel Rodriguez, and Abhinav Gupta. 2016. Learning a predictable and generative vector representation for objects. In European Conference on Computer Vision. Springer, 484–499.Google ScholarGoogle ScholarCross RefCross Ref
  8. Thibault Groueix, Matthew Fisher, Vladimir G Kim, Bryan C Russell, and Mathieu Aubry. 2018. A Papier-Mâché Approach to Learning 3D Surface Generation. In Proc. CVPR. 216–224.Google ScholarGoogle ScholarCross RefCross Ref
  9. Rana Hanocka, Amir Hertz, Noa Fish, Raja Giryes, Shachar Fleishman, and Daniel Cohen-Or. 2019. MeshCNN: a network with an edge. ACM Transactions on Graphics (TOG) 38, 4 (2019), 1–12.Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Jun Li, Chengjie Niu, and Kai Xu. 2019. Learning Part Generation and Assembly for Structure-aware Shape Synthesis. arXiv preprint arXiv:1906.06693(2019).Google ScholarGoogle Scholar
  11. Jun Li, Kai Xu, Siddhartha Chaudhuri, Ersin Yumer, Hao Zhang, and Leonidas Guibas. 2017. GRASS: Generative Recursive Autoencoders for Shape Structures. arXiv preprint arXiv:1705.02090(2017).Google ScholarGoogle Scholar
  12. Yangyan Li, Rui Bu, Mingchao Sun, Wei Wu, Xinhan Di, and Baoquan Chen. 2018. Pointcnn: Convolution on x-transformed points. In Advances in neural information processing systems. 820–830.Google ScholarGoogle Scholar
  13. Lars Mescheder, Michael Oechsle, Michael Niemeyer, Sebastian Nowozin, and Andreas Geiger. 2019. Occupancy Networks: Learning 3D Reconstruction in Function Space. In Proceedings IEEE Conf. on Computer Vision and Pattern Recognition (CVPR).Google ScholarGoogle ScholarCross RefCross Ref
  14. Kaichun Mo, Paul Guerrero, Li Yi, Hao Su, Peter Wonka, Niloy Mitra, and Leonidas J Guibas. 2019. StructureNet: Hierarchical Graph Networks for 3D Shape Generation. ACM Trans. on Graph. (SIGGRAPH Asia)(2019).Google ScholarGoogle Scholar
  15. Kaichun Mo, Shilin Zhu, Angel X. Chang, Li Yi, Subarna Tripathi, Leonidas J. Guibas, and Hao Su. 2019. PartNet: A Large-Scale Benchmark for Fine-Grained and Hierarchical Part-Level 3D Object Understanding. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR).Google ScholarGoogle ScholarCross RefCross Ref
  16. Jeong Joon Park, Peter Florence, Julian Straub, Richard Newcombe, and Steven Lovegrove. 2019. DeepSDF: Learning Continuous Signed Distance Functions for Shape Representation. In CVPR.Google ScholarGoogle Scholar
  17. Gernot Riegler, Ali Osman Ulusoy, and Andreas Geiger. 2017. Octnet: Learning deep 3d representations at high resolutions. In Proc. CVPR, Vol. 3.Google ScholarGoogle ScholarCross RefCross Ref
  18. Nadav Schor, Oren Katzir, Hao Zhang, and Daniel Cohen-Or. 2018. CompoNet: Learning to Generate the Unseen by Part Synthesis and Composition. arXiv preprint arXiv:1811.07441(2018).Google ScholarGoogle Scholar
  19. Amir Arsalan Soltani, Haibin Huang, Jiajun Wu, Tejas D. Kulkarni, and Joshua B. Tenenbaum. 2017. Synthesizing 3D Shapes via Modeling Multi-View Depth Maps and Silhouettes with Deep Generative Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 1511–1519.Google ScholarGoogle ScholarCross RefCross Ref
  20. Hao Su, Charles Ruizhongtai Qi, Kaichun Mo, and Leonidas J. Guibas. 2017. PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation. In Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR). to appear.Google ScholarGoogle Scholar
  21. Jiapeng Tang, Xiaoguang Han, Junyi Pan, Kui Jia, and Xin Tong. 2019. A skeleton-bridged deep learning approach for generating meshes of complex topologies from single rgb images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 4541–4550.Google ScholarGoogle ScholarCross RefCross Ref
  22. Hao Wang, Nadav Schor, Ruizhen Hu, Haibin Huang, Daniel Cohen-Or, and Hui Huang. 2018. Global-to-Local Generative Model for 3D Shapes. ACM Transactions on Graphics (Proc. SIGGRAPH ASIA) 37, 6 (2018), 214:1—214:10.Google ScholarGoogle Scholar
  23. Nanyang Wang, Yinda Zhang, Zhuwen Li, Yanwei Fu, Wei Liu, and Yu-Gang Jiang. 2018. Pixel2mesh: Generating 3d mesh models from single rgb images. In ECCV. 52–67.Google ScholarGoogle Scholar
  24. Peng-Shuai Wang, Yang Liu, Yu-Xiao Guo, Chun-Yu Sun, and Xin Tong. 2017. O-CNN: Octree-based Convolutional Neural Networks for 3D Shape Analysis. ACM Transactions on Graphics (SIGGRAPH) 36, 4 (2017).Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Jiajun Wu, Chengkai Zhang, Tianfan Xue, Bill Freeman, and Josh Tenenbaum. 2016. Learning a probabilistic latent space of object shapes via 3d generative-adversarial modeling. In Advances in Neural Information Processing Systems. 82–90.Google ScholarGoogle Scholar
  26. Rundi Wu, Yixin Zhuang, Kai Xu, Hao Zhang, and Baoquan Chen. 2020. PQ-NET: A Generative Part Seq2Seq Network for 3D Shapes. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).Google ScholarGoogle ScholarCross RefCross Ref
  27. Guandao Yang, Xun Huang, Zekun Hao, Ming-Yu Liu, Serge Belongie, and Bharath Hariharan. 2019. PointFlow: 3D Point Cloud Generation with Continuous Normalizing Flows. arXiv (2019).Google ScholarGoogle Scholar
  28. Yin Zhou and Oncel Tuzel. 2018. Voxelnet: End-to-end learning for point cloud based 3d object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 4490–4499.Google ScholarGoogle ScholarCross RefCross Ref
  29. Chuhang Zou, Ersin Yumer, Jimei Yang, Duygu Ceylan, and Derek Hoiem. 2017. 3D-PRNN: Generating Shape Primitives with Recurrent Neural Networks. 2017 IEEE International Conference on Computer Vision (ICCV) (Oct 2017). https://doi.org/10.1109/iccv.2017.103Google ScholarGoogle ScholarCross RefCross Ref

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image ACM Other conferences
    ICMLSC '22: Proceedings of the 2022 6th International Conference on Machine Learning and Soft Computing
    January 2022
    185 pages
    ISBN:9781450387477
    DOI:10.1145/3523150

    Copyright © 2022 ACM

    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    • Published: 13 April 2022

    Permissions

    Request permissions about this article.

    Request Permissions

    Check for updates

    Qualifiers

    • research-article
    • Research
    • Refereed limited
  • Article Metrics

    • Downloads (Last 12 months)23
    • Downloads (Last 6 weeks)0

    Other Metrics

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format .

View HTML Format