Skip to main content

Predicting Multiple Pregrasping Poses by Combining Deep Convolutional Neural Networks with Mixture Density Networks

  • Conference paper
  • First Online:
Neural Information Processing (ICONIP 2016)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 9949))

Included in the following conference series:

Abstract

In this paper, we propose a deep neural network to predict the pregrasp poses of a three-dimensional (3D) object. Specifically, a single RGB-D image is used to determine multiple pregrasp position of three fingers of the robotic hand for various poses of known or unknown objects. Multiple pregrasping pose prediction typically involves the use of complex multi-valued functions where standard regression models fail. To this end, we propose a deep neural network containing a variant of the traditional deep convolutional neural network as well as a mixture density network. Furthermore, in order to overcome the difficulty of learning with insufficient data in the first part of the proposed network, we develop a supervised learning technique to pretrain the variant of the convolutional neural network.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Sahbani, A., El-Khoury, S., Bidaud, P.: An overview of 3D object grasp synthesis algorithms. Robot. Auton. Syst. 60(3), 326–336 (2012)

    Article  Google Scholar 

  2. Bohg, J., Morales, A., Asfour, T., Kragic, D.: Data-driven grasp synthesis — a survey. IEEE Trans. Robot. 30(2), 289–309 (2014)

    Article  Google Scholar 

  3. Pinto, L., Gupta, A., Supersizing self-supervision: learning to grasp from 50k tries and 700 robot hours. arXiv:1509.06825 (2015)

  4. Levine, S., Peter, P., Alex, K., Deirdre, Q.: Learning hand-eye coordination for robotic grasping with deep learning, large-scale data collection. arXiv:1603.02199 (2016)

  5. Levine, S., Finn, C., Darrell, T., Abbeel, P.: End-to-end training of deep visuomotor policies. JMLR 17(39), 1–40 (2016)

    MathSciNet  MATH  Google Scholar 

  6. Han, W., Levine, S., Abbeel, P.: Learning compound multi-step controllers under unknown dynamics. In: Intelligent Robots and Systems (IROS), pp. 6435–6442 (2015)

    Google Scholar 

  7. Krizhevsky, A., Sutskever, I., Hinton, G.: ImageNet classification with deep convolutional neural networks. In: NIPS, pp. 1097–1105 (2012)

    Google Scholar 

  8. Noh, H., Hong, S., Han, B.: Learning deconvolution network for semantic segmentation. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1520–1528 (2015)

    Google Scholar 

  9. Xu, K., Ba, J., Kiros, R., Courville, A., Salakhutdinov, R., Zemel, R., Bengio, Y. Show Attend, tell: Neural image caption generation with visual attention. ICML, vol. 37 of JMLR Proceedings, pp. 2048–2057, 2015

    Google Scholar 

  10. Bishop, C.: Mixture density networks. Technical Report NCRG/94/004, Neural Computing Research Group, Aston University (1994)

    Google Scholar 

  11. Detry, R., Baseski, E., Popovic, M., Touati, Y., Kruger, N., Kroemer, O., Piater, J.: Learning object-specific grasp affordance densities. In: IEEE International Conference on Development and Learning (ICDL), pp. 1–7 (2009)

    Google Scholar 

  12. Paolini, R., Rodriguez, A., Srinivasa, S.S., Mason, M.T.: A data-driven statistical framework for post-grasp manipulation. Int. J. Robot. Res. 33(4), 600–615 (2014)

    Article  Google Scholar 

  13. Lenz, I., Lee, H., Saxena, A.: Deep learning for detecting robotic grasps. Int. J. Robot. Res. 34(4–5), 705–724 (2015)

    Article  Google Scholar 

  14. Finn, C., Tan, X. Y., Duan, Y., Darrell, T., Levine, S., Abbeel, P.: Deep spatial autoencoders for visuomotor learning. In: IEEE International Conference on Robotics and Automation (2016)

    Google Scholar 

Download references

Acknowledgement

This work was supported by the <Technology Innovation Industrial Program> funded by the Ministry of Trade, (MI, South Korea) [10048320, Technology Innovation Program], by the National Research Foundation of Korea grant funded by the Korea Government (MEST) (NRF-MIAXA003- 2010-0029744). All correspondences should be addressed to I.H. Suh.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Il Hong Suh .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2016 Springer International Publishing AG

About this paper

Cite this paper

Moon, S., Park, Y., Suh, I.H. (2016). Predicting Multiple Pregrasping Poses by Combining Deep Convolutional Neural Networks with Mixture Density Networks. In: Hirose, A., Ozawa, S., Doya, K., Ikeda, K., Lee, M., Liu, D. (eds) Neural Information Processing. ICONIP 2016. Lecture Notes in Computer Science(), vol 9949. Springer, Cham. https://doi.org/10.1007/978-3-319-46675-0_64

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-46675-0_64

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-46674-3

  • Online ISBN: 978-3-319-46675-0

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics