Abstract
During the planning process, a planner may have many options for refinements to perform on the plan being developed. The planner's efficiency depends on how it chooses which refinement to do next. Recent studies have shown that several versions of the popular “least commitment” plan refinement strategy are often outperformed by a fewest alternatives first (FAF) strategy that chooses to refine the plan element that has the smallest number of alternative refinement options.
In this paper, we examine the FAF strategy in more detail, to try to gain a better understanding of how well it performs and why. We present the following results:
-
•
A refinement planner's search space is an AND/OR graph, and the planner “serializes” this graph by mapping it into an equivalent state-space graph. Different plan refinement strategies produce different serializations of the AND/OR graph.
-
•
The sizes of different serializations of the AND/OR graph can differ by an exponential amount. A planner whose refinement strategy produces a small serialization is likely to be more efficient than a planner whose refinement strategy produces a large serialization.
-
•
The FAF heuristic can be computed in constant time, and in our experimental studies it usually produced an optimal or near-optimal serialization. This suggests that using FAF (or some similar heuristic) is preferable to trying to guarantee an optimal serialization (which we conjecture is a computationally intractible problem).
This research was supported in part by grants from NSF (IRI-9306580 and EEC 94-02384), ONR (N00014-J-91-1451), AFOSR (F49620-93-1-0065), the ARPA/Rome Laboratory Planning Initiative (F30602-93-C-0039), the ARPA 13 Initiative (N00014-94-10907), the ARL (DAAH049610297), and ARPA contract DABT-95-00037. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the view of the funders.
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
[Barret and Weld, 1994] Anthony Barret and Daniel Weld. Partial-order planning: Evaluating possible efficiency gains. Artificial Intelligence 67(1), pp. 71–112
[Bitner and Reingold, 1975] James Bitner and Edward Reingold. Backtrack Programming Techniques. CACM 18(11), pp. 651–656.
W. Clocksin and C. Mellish. Programming in PROLOG. Springer-Verlag.
[Currie and Tate, 1991] Ken Currie and Austin Tate. O-plan: the open planning architecture. Artificial Intelligence 52, pp. 49–86.
Kutluhan Erol. HTN planning: Formalization, analysis, and implementation. Ph.D. dissertation, Computer Science Dept., U. of Maryland.
[Gupta and Nau, 1992] Naresh Gupta and Dana Nau. On the complexity of blocksworld planning. Artificial Intelligence 56:2-3, pp. 223–254.
[Gerevini and Schubert, 1996] Alfonso Gerevini and Lenhart Schubert. Accelerating Partial-Order Planners: Some Techniques for Effective Search Control and Pruning. Journal of Artificial Intelligence Research 5, pp. 95–137.
David Joslin and Martha Pollack. Least-cost flaw repair: A plan refinement strategy for partial-order planning. In Proceedings of the Twelfth National Conference on Artificial Intelligence, pp. 1004–1009.
David Joslin and Martha Pollack. Is “early commitment” in plan generation ever a good idea? In Proc. Thirteenth National Conference on Artificial Intelligence, pp. 1188–1193.
[Kambhampati et al., 1995] Subbarao Kambhampati, Craig Knoblock, and Qiang Yang. Planning as refinement search: A unified framework for evaluating design tradeoffs in partial-order planning. Artificial Intelligence 76, pp. 167–238.
Vipin Kumar. Algorithms for constraint-satisfaction problems: A survey. AI Magazine, pp. 32–44.
Sean Luke. A Fast Probabilistic Tree Generation Algorithm. Unpublished manuscript.
J. S. Penberthy and Daniel Weld. UCPOP: A sound, complete, partial order planner for ADL. Proc. KR-92.
[Purdom, 1983] Paul W. Purdom. Search Rearrangement Backtracking and Polynomial Average Time. Artificial Intelligence 21, pp. 117–133.
[Purdom and Brown, 1983] Paul W. Purdom and Cynthia A. Brown. An Analysis of Backtracking with Search Rearragement. SIAM J. Computing 12(4), pp.717–733
Earl Sacerdoti. A Structure for Plans and Behavior. American Elsevier Publishing Company.
[Stefik, 1981] Mark Stefik. Planning with constraints (MOLGEN: part 1). Artificial Intelligence 16, pp. 111–140.
S. J. J. Smith, D. S. Nau, and T. A. Throop. Total-order multiagent task-network planning for control bridge. AAAI-96, pp. 108-113.
[Stockman, 1979] G. Stockman. A minimax algorithm better than alpha-beta? Artificial Intelligence 12(2), pp. 179–96.
[Tsuneto et al., 1996] Reiko Tsuneto, Kutluhan Erol, James Hendler, and Dana Nau. Commitment strategies in hierarchical task network planning. In Proc. Thirteenth National Conference on Artificial Intelligence, pp. 536–542.
[Veloso and Stone, 1995] Manuela Veloso and Peter Stone. FLECS: Planning with a flexible commitment strategy. JAIR 3, pp. 25–52.
Author information
Authors and Affiliations
Editor information
Rights and permissions
Copyright information
© 1997 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Tsuneto, R., Nau, D., Hendler, J. (1997). Plan-refinement strategies and search-space size. In: Steel, S., Alami, R. (eds) Recent Advances in AI Planning. ECP 1997. Lecture Notes in Computer Science, vol 1348. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-63912-8_103
Download citation
DOI: https://doi.org/10.1007/3-540-63912-8_103
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-63912-1
Online ISBN: 978-3-540-69665-0
eBook Packages: Springer Book Archive