Abstract
We present an algorithm for learning planning actions for waypoint simulations, a crucial subtask for robotics, gaming, and transportation agents that must perform locomotion behavior. Our algorithm is capable of learning operator’s symbolic literals and continuous effects even under noisy training data. It accepts as input a set of preprocessed positive and negative simulation-generated examples. It identifies symbolic preconditions using a MAX-SAT constraint solver and learns numeric preconditions and effects as continuous functions of numeric state variables by fitting a logistic regression model. We test the correctness of the learned operators by solving test problems and running the resulting plans on the simulator.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Bettstetter, C., Wagner, C.: The spatial node distribution of the random waypoint mobility model. In: WMAN, vol. 11, pp. 41–58 (2002)
Tan, C.H., Ang, J.H., Tan, K.C., Tay, A.: Online adaptive controller for simulated car racing. In: 2008 IEEE Congress on Evolutionary Computation (IEEE World Congress on Computational Intelligence), pp. 2239–2245. IEEE (2008)
Bruce, J., Veloso, M.M.: Real-time randomized path planning for robot navigation. In: Robot Soccer World Cup, pp. 288–295. Springer, Heidelberg (2002)
Arora, A., Fiorino, H., Pellier, D., Métivier, M., Pesty, S.: A review of learning planning action models. Knowl. Eng. Rev. 33 (2018)
Segura-Muros, J.Á., Pérez, R., Fernández-Olivares, J.: Learning numerical action models from noisy and partially observable states by means of inductive rule learning techniques. In: KEPS 2018, vol. 46 (2018)
Balac, N., Gaines, D.M., Fisher, D.: Learning action models for navigation in noisy environments. In: ICML Workshop on Machine Learning of Spatial Knowledge, Stanford, July 2000
Yang, Q., Wu, K., Jiang, Y.: Learning action models from plan examples using weighted MAX-SAT. Artif. Intell. 171(2–3), 107–143 (2007)
Gil, Y.: Learning by experimentation: incremental refinement of incomplete planning domains. In: Machine Learning Proceedings 1994, pp. 87–95. Morgan Kaufmann (1994)
Benson, S.S.: Learning action models for reactive autonomous agents. Doctoral dissertation, Stanford University (1996)
Lavrac, N., Dzeroski, S.: Inductive logic programming. In: WLP, pp. 146–160 (1994)
Nilsson, N.: Teleo-reactive programs for agent control. J. Artif. Intell. Res. 1, 139–158 (1993)
Pasula, H.M., Zettlemoyer, L.S., Kaelbling, L.P.: Learning symbolic models of stochastic domains. J. Artif. Intell. Res. 29, 309–352 (2007)
Lindsay, A., Gregory, P.: Discovering Numeric Constraints for Planning Domain Models. In: KEPS 2018, vol. 62 (2018)
Walsh, T.J., Littman, M.L.: Efficient learning of action schemas and web-service descriptions. In: AAAI-2008, vol. 8, pp. 714–719 (2008)
Ansótegui, C., et al.: Improving SAT-based weighted MaxSAT solvers. In: International Conference on Principles and Practice of Constraint Programming, pp. 86–101. Springer, Heidelberg (2012)
Hosmer Jr., D.W., Lemeshow, S., Sturdivant, R.X.: Applied Logistic Regression, vol. 398. Wiley, Hoboken (2013)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Fine-Morris, M., Auslander, B., Muños-Avila, H., Gupta, K. (2021). Learning Actions with Symbolic Literals and Continuous Effects for a Waypoint Navigation Simulation. In: Arai, K., Kapoor, S., Bhatia, R. (eds) Intelligent Systems and Applications. IntelliSys 2020. Advances in Intelligent Systems and Computing, vol 1250. Springer, Cham. https://doi.org/10.1007/978-3-030-55180-3_7
Download citation
DOI: https://doi.org/10.1007/978-3-030-55180-3_7
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-55179-7
Online ISBN: 978-3-030-55180-3
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)