PALO: a probabilistic hill-climbing algorithm

https://doi.org/10.1016/0004-3702(95)00040-2Get rights and content
Under an Elsevier user license
open archive

Abstract

Many learning systems search through a space of possible performance elements, seeking an element whose expected utility, over the distribution of problems, is high. As the task of finding the globally optimal element is often intractable, many practical learning systems instead hill-climb to a local optimum. Unfortunately, even this is problematic as the learner typically does not know the underlying distribution of problems, which it needs to determine an element's expected utility. This paper addresses the task of approximating this hill-climbing search when the utility function can only be estimated by sampling. We present a general algorithm, palo, that returns an element that is, with provably high probability, essentially a local optimum. We then demonstrate the generality of this algorithm by presenting three distinct applications that respectively find an element whose efficiency, accuracy or completeness is nearly optimal. These results suggest approaches to solving the utility problem from explanation-based learning, the multiple extension problem from nonmonotonic reasoning and the tractability/completeness tradeoff problem from knowledge representation.

Keywords

Computational learning theory
Hill-climbing
Speed-up learning
Utility problem
Knowledge compilation
Theory revision
Prioritized default theories

Cited by (0)

This paper expands the short article, “Probabilistic hill-climbing: theory and applications” that was awarded the “Artificial Intelligence Journal Best Paper Award” at the Ninth Canadian Conference on Artificial Intelligence (CSCSI-92), in Vancouver, in May 1992.