Incorporating decision-theoretic planning in a robot architecture

https://doi.org/10.1016/S0921-8890(02)00326-3Get rights and content

Abstract

The goal of robotics research is to design a robot to fulfill a variety of tasks in the real world. Inherent in the real world is a high degree of uncertainty about the robot’s behavior and about the world. We introduce a robot task architecture, DTRC, that generates plans with actions that incorporate costs and uncertain effects, and states that yield rewards.

The use of a decision-theoretic planner in a robot task architecture is demonstrated on the mobile robot domain of miniature golf. The miniature golf domain shows the application of decision-theoretic planning in an inherently uncertain domain, and demonstrates that by using decision-theoretic planning as the reasoning method in a robot task architecture, accommodation for uncertain information plays a direct role in the reasoning process.

Introduction

Robot task architectures represent an assemblage of algorithms that interpret user goals and sensor readings, in order to determine robot motor actions that will reach the user-specified goals. Many current approaches to robot task architectures use symbolic planning to perform high-level reasoning. This reasoning generates a sequence of actions for problems that transition the robot agent from an initial condition to a goal condition. In classical symbolic planning problems, the agent is given all relevant information about the world and executed actions are both deterministic and assumed to always succeed.

The use of a symbolic planner in a robot task architecture overlooks the inherent uncertainty in robotic domains. The uncertainty stems from the existence of multiple possible action effects, unreliable sensors, and incomplete domain information. Instead of having the planner address these issues, current task architectures choose to have other parts of the architecture handle problems arising from uncertainties or to create domain-specific and robot-specific workarounds that allow the robot to complete the given task.

To meet the goal for robots to complete tasks in the real world, the robot must deal with the uncertainty associated with actions, sensors and information. Additionally, actions in the real world have associated costs. To perform this reasoning within a robot architecture, we have developed the Decision-Theoretic Robot Controller (DTRC).

Our robot task control architecture contains three main elements: the DT-Graphplan planner, the robot skills, and the execution monitor. The DT-Graphplan planner is a decision-theoretic planner that generates a satisfying plan for domains with incomplete information, stochastic action effects, state reward conditions, and action costs. DT-Graphplan performs high-level reasoning over the domain, generating action sequences to transition the agent from the initial condition to the goal condition. The robot skills are a collection of low-level base building blocks that handle motor and sensor control and integration, essentially the actions the planner uses to assemble a plan. The execution monitor communicates between DT-Graphplan and the robot behaviors by maintaining current domain information for the planner and activating behaviors based on the plan provided. DTRC’s execution monitor also detects plan failure and triggers replanning from the current condition.

We illustrate DTRC on a robot miniature golf domain. For robot miniature golf, the task the robot must address is to get the golf ball into the cup with the fewest number of strokes. The course has a number of stationary and non-stationary obstacles to avoid. The robot has three methods of moving the ball, and each action has a different stroke penalty (cost) and outcome. Our results are generated from an implementation of this domain using a Pioneer 1 robot. This domain demonstrates the application of decision theory in a robot architecture, shows how changes in the domain result in changes in the preference of plans, and describes how the robot architecture sustains a probabilistic representation of the domain state and generates initial conditions for replanning.

This work introduces the DT-Graphplan planner and our robot control architecture, illustrating the application of the planner to imprecise and uncertain domains such as robotics. Section 2 briefly covers related robot control architecture work. Section 3 discusses the interactions of the elements of the task control architecture, and design of DT-Graphplan to handle decision theory instead of using a strictly Bayesian method. Section 4 contains the results of applying the architecture to the task of robot miniature golf, and the required elements to convert to the robot soccer domain and how altering the domain description changes the overall behavior of the robot. We finish with concluding remarks and possible extensions of this work.

Section snippets

Related work

The standard robot task architecture consists of a planner, a sequencer, and a set of robot skills. The robot architectures divide the entire robotic planning process into planning, sequencing, and basic skills so that the robot can reason about actuator or control errors while also dealing with a changing set of interacting goals. In the basic sense, we abstract the planning phase away from the robot skill level to increase planning speed and success.

Two of the best-known control architectures

The robot architecture

The DTRC consists of DT-Graphplan as the high-level reasoning system, a set of robot skills as the low behavioral level, and an execution monitor which passes information between the two. These three elements provide a robust method for handling uncertain robotic environments. This section discusses the execution monitor, the skills layer, and how these two elements interact with DT-Graphplan.

Fig. 1 shows an overview of the robot architecture. The user supplies domain information to the

Robot miniature golf

Integration of the systems of the DTRC architecture is illustrated using the application of a real robot on the robot miniature golf domain. Through this example we provide details of the DTRC implementation, component interaction, and handling of plan failures. The domain does not demonstrate maintenance goals, or replanning based on receiving new goals while executing a plan.

For the robot miniature golf domain, the robot is an Activmedia Pioneer 1 with a gripper and camera. A neon orange

Conclusions and future work

The goal of robotics-related research is for robots to complete complex tasks in the real world. However, in the real world there is a great deal of uncertainty associated with actions, sensors, and environment information. This paper discusses DTRC, a robot control architecture, capable of interpreting uncertainty associated with real world domains and generating acceptable plans of action. DTRC consists of a low-level robot skill layer, an execution monitor, and DT-Graphplan. In DTRC

Gilbert L. Peterson recently completed his Ph.D. in Computer Science from the University of Texas at Arlington. His research interests include robotics, artificial intelligence, planning, and machine learning. Dr. Peterson received his B.S. in Architecture from the University of Texas at Arlington, and an M.S. in Computer Science from the University of Texas at Arlington.

References (12)

  • M. Asada, A. Birk, E. Pagello, M. Fujita, I. Noda, S. Tadokoro, P. Duhaut, P. Stone, M. Veloso, T. Balch, H. Kitano, B....
  • R.S. Aylett, A.M. Coddington, D.P. Barnes, R.A. Ghanea-Hercock, What does a planner need to know about execution? in:...
  • T. Balch, Learning roles: behavioral diversity in robot teams, in: Proceedings of the AAAI Workshop on Multiagent...
  • A.L. Blum, M.L. Furst, Fast planning through planning graph analysis, in: Proceedings of the International Joint...
  • A.L. Blum, J.C. Langford, Probabilistic planning in the graphplan framework, in: Proceedings of the Fifth European...
  • R.P. Bonasso, D. Kortenkamp, Using a layered control architecture to alleviate planning with incomplete information,...
There are more references available in the full text version of this article.

Cited by (9)

View all citing articles on Scopus

Gilbert L. Peterson recently completed his Ph.D. in Computer Science from the University of Texas at Arlington. His research interests include robotics, artificial intelligence, planning, and machine learning. Dr. Peterson received his B.S. in Architecture from the University of Texas at Arlington, and an M.S. in Computer Science from the University of Texas at Arlington.

Diane J. Cook is currently a Professor in the Computer Science and Engineering Department at the University of Texas at Arlington. Her research interests include artificial intelligence, machine learning, data mining, robotics, and parallel algorithms for artificial intelligence. She has published over 120 papers in these areas. Dr. Cook received her B.S. from Wheaton College in 1985, and her M.S. and Ph.D. from the University of Illinois in 1987 and 1990, respectively.

View full text