skip to main content
10.1145/3371382.3377436acmconferencesArticle/Chapter ViewAbstractPublication PageshriConference Proceedingsconference-collections
abstract

Detecting Hypothesis Space Misspecification in Robot Learning from Human Input

Published:01 April 2020Publication History

ABSTRACT

Learning from human input has enabled autonomous agents to perform increasingly more complex tasks that are otherwise difficult to carry out automatically. To this end, recent works have studied how robots can incorporate such input - like demonstrations or corrections - into objective functions describing the desired behaviors. While these methods have shown progress in a variety of settings, from semi-autonomous driving, to household robotics, to automated airplane control, they all suffer from the same crucial drawback: they implicitly assume that the person's intentions can always be captured by the robot's hypothesis space. We call attention to the fact that this assumption is often unrealistic, as no model can completely account for every single possible situation ahead of time. When the robot's hypothesis space is misspecified, human input can be unhelpful - or even detrimental - to the way the robot is performing its tasks. Our work tackles this issue by proposing that the robot should first explicitly reason about how well its hypothesis space can explain human inputs, then use that situational confidence to inform how it should incorporate them.

References

  1. Pieter Abbeel and Andrew Y Ng. 2004. Apprenticeship learning via inverse reinforcement learning. In Machine Learning (ICML), International Conference on. ACM.Google ScholarGoogle Scholar
  2. Andrea Bajcsy, Dylan P. Losey, Marcia Kilchenman O'Malley, and Anca D. Dragan. 2017. Learning Robot Objectives from Physical Human Interaction. In CoRL .Google ScholarGoogle Scholar
  3. Chris L Baker, Joshua B Tenenbaum, and Rebecca R Saxe. 2007. Goal inference as inverse planning. In Proceedings of the Annual Meeting of the Cognitive Science Society, Vol. 29.Google ScholarGoogle Scholar
  4. Andreea Bobu, Andrea Bajcsy, Jaime F. Fisac, Sampada Deglurkar, and Anca D. Dragan. 2019. Quantifying Hypothesis Space Misspecification in Learning from Human-Robot Demonstrations and Physical Corrections. (2019). To appear in Transactions on Robotics.Google ScholarGoogle Scholar
  5. Andreea Bobu, Andrea Bajcsy, Jaime F. Fisac, and Anca D. Dragan. 2018. Learning under Misspecified Objective Spaces. In 2nd Annual Conference on Robot Learning, CoRL 2018, Zü rich, Switzerland, 29--31 October 2018, Proceedings . 796--805.Google ScholarGoogle Scholar
  6. Paul Christiano, Jan Leike, Tom B. Brown, Miljan Martic, Shane Legg, and Dario Amodei. 2017. Deep reinforcement learning from human preferences. (06 2017).Google ScholarGoogle Scholar
  7. Justin Fu, Avi Singh, Dibya Ghosh, Larry Yang, and Sergey Levine. [n.d.]. Variational Inverse Control with Events: A General Framework for Data-Driven Reward Definition. arXiv preprint , Vol. arXiv:1805.11686 ( [n.,d.]).Google ScholarGoogle Scholar
  8. Ashesh Jain, Shikhar Sharma, Thorsten Joachims, and Ashutosh Saxena. 2015. Learning preferences for manipulation tasks from online coactive feedback. The International Journal of Robotics Research , Vol. 34, 10 (2015), 1296--1313.Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Shervin Javdani, Siddhartha S Srinivasa, and J Andrew Bagnell. 2015. Shared autonomy via hindsight optimization. arXiv preprint arXiv:1503.07619 (2015).Google ScholarGoogle Scholar
  10. John Von Neumann and Oskar Morgenstern. 1945. Theory of games and economic behavior .Princeton University Press Princeton, NJ.Google ScholarGoogle Scholar

Index Terms

  1. Detecting Hypothesis Space Misspecification in Robot Learning from Human Input

                Recommendations

                Comments

                Login options

                Check if you have access through your login credentials or your institution to get full access on this article.

                Sign in
                • Published in

                  cover image ACM Conferences
                  HRI '20: Companion of the 2020 ACM/IEEE International Conference on Human-Robot Interaction
                  March 2020
                  702 pages
                  ISBN:9781450370578
                  DOI:10.1145/3371382

                  Copyright © 2020 Owner/Author

                  Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

                  Publisher

                  Association for Computing Machinery

                  New York, NY, United States

                  Publication History

                  • Published: 1 April 2020

                  Check for updates

                  Qualifiers

                  • abstract

                  Acceptance Rates

                  Overall Acceptance Rate192of519submissions,37%
                • Article Metrics

                  • Downloads (Last 12 months)16
                  • Downloads (Last 6 weeks)1

                  Other Metrics

                PDF Format

                View or Download as a PDF file.

                PDF

                eReader

                View online with eReader.

                eReader