Abstract
Single Robot Hose Transport (SRHT) is a limit case of Linked Multicomponent Robotic Systems (L-MCRS), when one robot moves the tip of a hose to a desired position, while the other hose extreme is attached to a source position. Reinforcement Learning (RL) algorithms have been applied to learn autonomously the robot control with success. However, RL algorithms produce large and intractable data structures. This paper addresses the problem by learning an Extreme Learning Machine (ELM) from the state-action value Q-table, obtaining very relevant data reduction. In this paper we evaluate empirically a classification strategy to formulate ELM learning to provide approximations to the Q-table, obtaining very promising results.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Duro, R., GraƱa, M., de Lope, J.: On the potential contributions of hybrid intelligent approaches to multicomponen robotic system development. Inf. Sci. 180(14), 2635ā2648 (2010)
LĆ³pez-Guede, J.M., GraƱa, M., Zulueta, E.: On distributed cooperative control for the manipulation of a hose by a multirobot system. In: Corchado, E., Abraham, A., Pedrycz, W. (eds.) HAIS 2008. LNCS (LNAI), vol. 5271, pp. 673ā679. Springer, Heidelberg (2008)
Fernandez-Gauna, B., Lopez-Guede, J., Zulueta, E., GraƱa, M.: Learning hose transport control with Q-Learning. Neural Netw. World 20(7), 913ā923 (2010)
Fernandez-Gauna, B., Lopez-Guede, J.M., Zulueta, E., Echegoyen, Z., GraƱa, M.: Basic results and experiments on robotic multi-agent system for hose deployment and transportation. Int. J. Artif. Intell. 6(S11), 183ā202 (2011)
Sutton, R., Barto, A.: Reinforcement Learning: An Introduction. MIT Press, Cambridge (1998)
Watkins, C., Dayan, P.: Q-Learning. Mach. Learn. 8(3ā4), 279ā292 (1992)
Lopez-Guede, J.M., Fernandez-Gauna, B., GraƱa, M., Zulueta, E.: Improving the control of single robot hose transport. Cybern. Syst. 43(4), 261ā275 (2012)
Bellman, R.: A Markovian decision process. Indiana Univ. Math. J. 6, 679ā684 (1957)
Tijms, H.: Discrete Time Markov Decision Processes, pp. 233ā277. Wiley, Hoboken (2004)
Watkins, C.: Learning from delayed rewards. In: Ph.D. dissertation, University of Cambridge, England (1989)
Watkins, C., Dayan, P.: Technical note: Q-Learning. Mach. Learn. 8, 279ā292 (1992)
Huang, D.H.W.G.-B., Lan, Y.: Extreme learning machines: a survey. Int. J. Mach. Learn. Cybern. 2(2), 107ā122 (2011)
Huang, G.-B., Zhu, Q.-Y., Siew, C.-K.: Extreme learning machine: theory and applications. Neurocomputing 70(1ā3), 489ā501 (2006)
Widrow, B., Lehr, M.: 30 years of adaptive neural networks: perceptron, madaline, and backpropagation. Proc. IEEE 78(9), 1415ā1442 (1990)
Lopez-Guede, J.M., GraƱa, M., Zulueta, E., Barambones, O.: Economical implementation of control loops for multi-robot systems. In: Kƶppen, M., Kasabov, N., Coghill, G. (eds.) ICONIP 2008, Part I. LNCS, vol. 5506, pp. 1053ā1059. Springer, Heidelberg (2009). Revised Selected Papers, Part I
Lopez-Guede, J.M., Fernandez-Gauna, B., GraƱa, M., Zulueta, E.: Multi-robot systems control implementation. In: Robot Learning, pp. 137ā150. SCIYO (2010)
Acknowledgements
The research was supported by the Computational Intelligence Group, funded by the Basque Government with grant IT874-13.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
Ā© 2016 Springer International Publishing Switzerland
About this paper
Cite this paper
Lopez-Guede, J.M., Garmendia, A., GraƱa, M. (2016). Knowledge Modeling by ELM in RL for SRHT Problem. In: MartĆnez-Ćlvarez, F., Troncoso, A., QuintiĆ”n, H., Corchado, E. (eds) Hybrid Artificial Intelligent Systems. HAIS 2016. Lecture Notes in Computer Science(), vol 9648. Springer, Cham. https://doi.org/10.1007/978-3-319-32034-2_27
Download citation
DOI: https://doi.org/10.1007/978-3-319-32034-2_27
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-32033-5
Online ISBN: 978-3-319-32034-2
eBook Packages: Computer ScienceComputer Science (R0)