Abstract
Motion planning is a hot topic in robotics, and the sampling-based algorithms have gained their popularities in research areas. However, these methods are still not suitable for real-world motion planning problems, because it is computationally expensive to completely explore the high-dimensional configuration space (C-space) of robots. Inspired by the related works on learning from demonstration, we propose a novel motion planning method named teaching roadmaps, which can take advantage of the optimal teaching data and quickly find a new path in the similar scenarios. The theoretical analysis and our experiments indicated that our approach is probabilistically complete, and it can find a feasible path faster than other sampling-based methods in similar environments.
Similar content being viewed by others
References
Kavraki LE, Svestka P, Latombe J-C, Overmars MH (1996) Probabilistic roadmaps for path planning in high-dimensional configuration spaces. IEEE Trans Robot Autom 12:566–580
LaValle SM, Kuffner JJ Jr (2000) Rapidly-exploring random trees: progress and prospects. Algorithmic Comput Robot New Dir 5:293–308
Edwards S, Lewis C (2012) Ros-industrial: applying the robot operating system (ros) to industrial applications. In: IEEE international conference on robotics and automation, ECHORD workshop, 2012
Calinon S, Guenter F, Billard A (2007) On learning, representing, and generalizing a task in a humanoid robot. IEEE Trans Syst Man and Cybern Part B (Cybern) 37:286–298
Schaal S (2006) Dynamic movement primitives-a framework for motor control in humans and humanoid robotics. In: Adaptive motion of animals and machines. Springer, pp. 261-280
Bohlin R, Kavraki LE (2000) Path planning using lazy PRM. In: Proceedings. ICRA’00. IEEE international conference on robotics and automation, 2000. IEEE, pp 521–528
Hsu D, Latombe J-C, Motwani R (1997) Path planning in expansive configuration spaces. In: Proceedings., 1997 IEEE international conference on robotics and automation, 1997. IEEE, pp 2719–2726
Sánchez G, Latombe J-C (2003) A single-query bi-directional probabilistic roadmap planner with lazy collision checking. Robot Res 6:403–417
Kuffner JJ, LaValle SM (2000) RRT-Connect: an efficient approach to single-query path planning. In: Proceedings. ICRA’00. IEEE international conference on robotics and automation, 2000. IEEE, pp 995–1001
Şucan IA, Kavraki LE (2009) Kinodynamic motion planning by interior-exterior cell exploration. In: Algorithmic foundation of robotics, vol 8. Springer, pp 449–464
Elbanhawi M, Simic M (2014) Sampling-based robot motion planning: a review. IEEE Access 2:56–77
Jaillet L, Cortés J, Siméon T (2010) Sampling-based path planning on configuration-space costmaps. IEEE Trans Rob 26:635–646
Karaman S, Frazzoli E (2011) Sampling-based algorithms for optimal motion planning. Int J Robot Res 30:846–894
Gammell JD, Srinivasa SS, Barfoot TD (2015) Batch informed trees (BIT*): sampling-based optimal planning via the heuristically guided search of implicit random geometric graphs. In: 2015 IEEE International Conference on robotics and automation (ICRA). IEEE, pp 3067–3074
Shin SY, Kim C (2015) Human-like motion generation and control for humanoid’s dual arm object manipulation. IEEE Trans Ind Electron 62:2265–2276
Calinon S, Sardellitti I, Caldwell DG (2010) Learning-based control strategy for safe human-robot interaction exploiting task and robot redundancies. In: 2010 IEEE/RSJ international conference on intelligent robots and systems (IROS). IEEE, pp 249–254
Rai A, Meier F, Ijspeert A, Schaal S (2014) Learning coupling terms for obstacle avoidance. In: 2014 14th IEEE-RAS international conference on humanoid robots (humanoids). IEEE, pp 512–518
Paxton C, Hager GD, Bascetta L (2015) An incremental approach to learning generalizable robot tasks from human demonstration. In: 2015 IEEE international conference on robotics and automation (ICRA). IEEE, pp 5616–5621
Jiang X, Kallmann M (2007) Learning humanoid reaching tasks in dynamic environments. In: IROS 2007. IEEE/RSJ international conference on intelligent robots and systems, 2007. IEEE, pp 1148–1153
Rosell J, Suárez R, Pérez A (2013) Path planning for grasping operations using an adaptive PCA-based sampling method. Auton Robot 35:27–36
García N, Rosell J, Suárez R (2015) Motion planning using first-order synergies. In: 2015 IEEE/RSJ international conference on intelligent robots and systems (IROS). IEEE, pp 2058–2063
Phillips M, Cohen BJ, Chitta S, Likhachev M (2012) E-graphs: bootstrapping planning with experience graphs. In: Robotics: science and systems, 2012
Phillips M, Hwang V, Chitta S, Likhachev M (2016) Learning to plan for constrained manipulation from demonstrations. Auton Robots 40:109–124
Rodriguez A, Laio A (2014) Clustering by fast search and find of density peaks. Science 344:1492–1496
Dempster AP, Laird NM, Rubin DB (1977) Maximum likelihood from incomplete data via the EM algorithm. J R Stat Soc Ser B (Methodol) 39(1):1–38
Kavraki LE, Kolountzakis MN, Latombe J-C (1998) Analysis of probabilistic roadmaps for path planning. IEEE Trans Robot Autom 14:166–171
Ladd A, Kavraki LE (2002) Generalizing the analysis of PRM. In: Proceedings. ICRA’02. IEEE international conference on robotics and automation, 2002. IEEE, pp 2120–2125
Sucan IA, Moll M, Kavraki LE (2012) The open motion planning library. IEEE Robot Autom Mag 19:72–82
Sucan IA, Chitta S (2013) Moveit! http://moveit.ros.org. Accessed 21 Jan 2018
Pan J, Chitta S, Manocha D (2012) FCL: a general purpose library for collision and proximity queries. In: 2012 IEEE international conference on robotics and automation (ICRA). IEEE, pp 3859–3866
Frank M, Leitner J, Stollenga M, Förster A, Schmidhuber J (2014) Curiosity driven reinforcement learning for motion planning on humanoids. Front Neurorobot 7:25–40
Dragan AD, Lee KC, Srinivasa SS (2013) Legibility and predictability of robot motion. In: 2013 8th ACM/IEEE international conference on human–robot interaction (HRI). IEEE, pp 301–308
Mnih V, Kavukcuoglu K, Silver D, Rusu AA, Veness J, Bellemare MG, Graves A, Riedmiller M, Fidjeland AK, Ostrovski G (2015) Human-level control through deep reinforcement learning. Nature 518:529–533
Acknowledgements
This work was partially supported by the National Natural Science Foundation of China under Grant No. 61673261 and YASKAWA Electric Corporation.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Qiu, Q., Cao, Q. Motion planning in semistructured environments with teaching roadmaps. Intel Serv Robotics 13, 331–342 (2020). https://doi.org/10.1007/s11370-020-00316-9
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11370-020-00316-9