Abstract
In this paper, we propose a deep neural network to predict the pregrasp poses of a three-dimensional (3D) object. Specifically, a single RGB-D image is used to determine multiple pregrasp position of three fingers of the robotic hand for various poses of known or unknown objects. Multiple pregrasping pose prediction typically involves the use of complex multi-valued functions where standard regression models fail. To this end, we propose a deep neural network containing a variant of the traditional deep convolutional neural network as well as a mixture density network. Furthermore, in order to overcome the difficulty of learning with insufficient data in the first part of the proposed network, we develop a supervised learning technique to pretrain the variant of the convolutional neural network.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Sahbani, A., El-Khoury, S., Bidaud, P.: An overview of 3D object grasp synthesis algorithms. Robot. Auton. Syst. 60(3), 326–336 (2012)
Bohg, J., Morales, A., Asfour, T., Kragic, D.: Data-driven grasp synthesis — a survey. IEEE Trans. Robot. 30(2), 289–309 (2014)
Pinto, L., Gupta, A., Supersizing self-supervision: learning to grasp from 50k tries and 700 robot hours. arXiv:1509.06825 (2015)
Levine, S., Peter, P., Alex, K., Deirdre, Q.: Learning hand-eye coordination for robotic grasping with deep learning, large-scale data collection. arXiv:1603.02199 (2016)
Levine, S., Finn, C., Darrell, T., Abbeel, P.: End-to-end training of deep visuomotor policies. JMLR 17(39), 1–40 (2016)
Han, W., Levine, S., Abbeel, P.: Learning compound multi-step controllers under unknown dynamics. In: Intelligent Robots and Systems (IROS), pp. 6435–6442 (2015)
Krizhevsky, A., Sutskever, I., Hinton, G.: ImageNet classification with deep convolutional neural networks. In: NIPS, pp. 1097–1105 (2012)
Noh, H., Hong, S., Han, B.: Learning deconvolution network for semantic segmentation. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1520–1528 (2015)
Xu, K., Ba, J., Kiros, R., Courville, A., Salakhutdinov, R., Zemel, R., Bengio, Y. Show Attend, tell: Neural image caption generation with visual attention. ICML, vol. 37 of JMLR Proceedings, pp. 2048–2057, 2015
Bishop, C.: Mixture density networks. Technical Report NCRG/94/004, Neural Computing Research Group, Aston University (1994)
Detry, R., Baseski, E., Popovic, M., Touati, Y., Kruger, N., Kroemer, O., Piater, J.: Learning object-specific grasp affordance densities. In: IEEE International Conference on Development and Learning (ICDL), pp. 1–7 (2009)
Paolini, R., Rodriguez, A., Srinivasa, S.S., Mason, M.T.: A data-driven statistical framework for post-grasp manipulation. Int. J. Robot. Res. 33(4), 600–615 (2014)
Lenz, I., Lee, H., Saxena, A.: Deep learning for detecting robotic grasps. Int. J. Robot. Res. 34(4–5), 705–724 (2015)
Finn, C., Tan, X. Y., Duan, Y., Darrell, T., Levine, S., Abbeel, P.: Deep spatial autoencoders for visuomotor learning. In: IEEE International Conference on Robotics and Automation (2016)
Acknowledgement
This work was supported by the <Technology Innovation Industrial Program> funded by the Ministry of Trade, (MI, South Korea) [10048320, Technology Innovation Program], by the National Research Foundation of Korea grant funded by the Korea Government (MEST) (NRF-MIAXA003- 2010-0029744). All correspondences should be addressed to I.H. Suh.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer International Publishing AG
About this paper
Cite this paper
Moon, S., Park, Y., Suh, I.H. (2016). Predicting Multiple Pregrasping Poses by Combining Deep Convolutional Neural Networks with Mixture Density Networks. In: Hirose, A., Ozawa, S., Doya, K., Ikeda, K., Lee, M., Liu, D. (eds) Neural Information Processing. ICONIP 2016. Lecture Notes in Computer Science(), vol 9949. Springer, Cham. https://doi.org/10.1007/978-3-319-46675-0_64
Download citation
DOI: https://doi.org/10.1007/978-3-319-46675-0_64
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-46674-3
Online ISBN: 978-3-319-46675-0
eBook Packages: Computer ScienceComputer Science (R0)