Abstract
Collaboration with a human partner is a challenging task expected of intelligent robots. To realize this, robots need the ability to share a particular goal with a human and dynamically infer whether the goal state is changed by the human. In this paper, we propose a neural network-based computational framework with a gradient-based optimization of the goal state that enables robots to achieve this ability. The proposed framework consists of convolutional variational autoencoders (ConvVAEs) and a recurrent neural network (RNN) with a long short-term memory (LSTM) architecture that learns to map a given goal image for collaboration to visuomotor predictions. More specifically, visual and goal feature states are first extracted by the encoder of the respective ConvVAEs. Visual feature and motor predictions are then generated by the LSTM based on their current state and are conditioned according to the extracted goal feature state. During collaboration after the learning process, the goal feature state is optimized by gradient descent to minimize errors between the predicted and actual visual feature states. This enables the robot to dynamically infer situational (goal) changes of the human partner from visual observations alone. The proposed framework is evaluated by conducting experiments on a human–robot collaboration task involving object assembly. Experimental results demonstrate that a robot equipped with the proposed framework can collaborate with a human partner through dynamic goal inference even when the situation is ambiguous.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Maeda, G., Ewerton, M., Lioutikov, R., Ben Amor, H., Peters, J., Neumann, G.: Learning interaction for collaborative tasks with probabilistic movement primitives. In: 2014 IEEE-RAS International Conference on Humanoid Robots, vol. 2015, pp. 527–534. IEEE, February 2014
Murata, S., Li, Y., Arie, H., Ogata, T., Sugano, S.: Learning to achieve different levels of adaptability for human-robot collaboration utilizing a neuro-dynamical system. IEEE Trans. Cogn. Dev. Syst. 10(3), 712–725 (2018)
Tani, J.: Learning to generate articulated behavior through the bottom-up and the top-down interaction processes. Neural Netw. 16(1), 11–23 (2003)
Murata, S., Yamashita, Y., Arie, H., Ogata, T., Sugano, S., Tani, J.: Learning to perceive the world as probabilistic or deterministic via interaction with others: a neuro-robotics experiment. IEEE Trans. Neural Netw. Learn. Syst. 28(4), 830–848 (2017)
Rao, R.P., Ballard, D.H.: Predictive coding in the visual cortex: a functional interpretation of some extra-classical receptive-field effects. Nat. Neurosci. 2(1), 79–87 (1999)
Clark, A.: Whatever next? predictive brains, situated agents, and the future of cognitive science. Behav. Brain Sci. 36(3), 181–204 (2013)
Hafner, D., et al.: Learning latent dynamics for planning from pixels. In: Chaudhuri, K., Salakhutdinov, R. (eds.) Proceedings of the 36th International Conference on Machine Learning, PMLR, pp. 2555–2565 (2019)
Srinivas, A., Jabri, A., Abbeel, P., Levine, S., Finn, C.: Universal planning networks. In Dy, J., Krause, A. (eds.) Proceedings of the 35th International Conference on Machine Learning, PMLR (2018)
Hinton, G.E., Salakhutdinov, R.R.: Reducing the dimensionality of data with neural networks. Sci. (New York, N.Y.) 313(5786), 504–507 (2006)
Noda, K., Arie, H., Suga, Y., Ogata, T.: Multimodal integration learning of robot behavior using deep neural networks. Robot. Auton. Syst. 62(6), 721–736 (2014)
Takahashi, K., Ogata, T., Tjandra, H., Yamaguchi, Y., Sugano, S.: Tool-body assimilation model based on body babbling and neurodynamical system. Math. Probl. Eng. 2015, 1–15 (2015)
Kingma, D.P., Welling, M.: Auto-encoding variational Bayes. In: Salakhutdinov, K.C.R. (eds.) Proceedings of the 2nd International Conference on Learning Representations (ICLR), pp. 1–14 (2014)
Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. In: Proceedings of the 3rd International Conference on Learning Representations (ICLR), pp. 1–15 (2015)
Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)
Acknowledgement
This work was supported in part by JST CREST (JPMJCR15E3), JSPS KAKENHI (JP16H05878), and the Research Institute for Science and Engineering, Waseda University.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Murata, S., Masuda, W., Chen, J., Arie, H., Ogata, T., Sugano, S. (2019). Achieving Human–Robot Collaboration with Dynamic Goal Inference by Gradient Descent. In: Gedeon, T., Wong, K., Lee, M. (eds) Neural Information Processing. ICONIP 2019. Lecture Notes in Computer Science(), vol 11954. Springer, Cham. https://doi.org/10.1007/978-3-030-36711-4_49
Download citation
DOI: https://doi.org/10.1007/978-3-030-36711-4_49
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-36710-7
Online ISBN: 978-3-030-36711-4
eBook Packages: Computer ScienceComputer Science (R0)