The maximum entropy method applied to stationary density computation

https://doi.org/10.1016/j.amc.2006.07.052Get rights and content

Abstract

The maximum entropy method (maxent) is widely used in the context of the moment problem which appears naturally in many branches of physics and engineering; it is used to numerically recover the density with least bias from finitely many known moments. We introduce the basic idea behind this method and apply this method to approximating fixed densities of Markov operators in stochastic analysis and Frobenius–Perron operators in ergodic theory of chaotic dynamics.

Introduction

The concept of entropy was first introduced by Clausius into thermodynamics in the middle of the nineteenth century, and later used in a different form by L. Boltzmann in his pioneering work on the kinetic theory of gases in 1866 [9]. It is a measure of the amount of information required to specify the state of a thermodynamic system. The famous Second Law of Thermodynamics says that in an isolated system, the (thermodynamic) entropy never decreases.

A modern concept of entropy was established in information theory by C.E. Shannon in 1948. The Shannon entropy defined for all finite sample spaces with events w1, w2,  , wn with probabilities p1, p2,  , pn (discrete information sources) isH(p1,p2,,pn)=-i=1npilogpiwhich is a measure of uncertainty of an information source.

Based on the Shannon entropy, the Kolmogorov–Sinai–Ornstein entropy defined for measurable transformations of probability spaces was successfully used in solving the problem of isomorphism of dynamical systems. And in 1965 R. Adler, A. Konhein, and M. McAndrew introduced the topological entropy for continuous transformations of compact Hausdorff spaces (see e.g., [5]).

The principle of maximum entropy was introduced in the context of information theory in 1957 by Jaynes. Since his seminal paper [8], the maximum entropy method has been widely used in many areas of science and technology. Basically this numerical scheme is used to numerically recover the required density function with least bias among all the possible candidates which satisfy given constraints. The maximum entropy method has diverse applications in physics and engineering (see [12], [13] and references therein).

In this paper, we will introduce the basic idea of the maximum entropy method and present some of the recent progresses of the method in solving some type of operator equations. In Section 2 we give the formation of the method and study their properties. Another formulation based on the Galerkin projection will also be presented. In Section 3 we apply the method to solve the Markov operator fixed density problem which is important in the stochastic analysis of deterministic dynamical systems. Some error analysis and numerical experiments are contained in Section 4. We conclude in Section 5.

Section snippets

The Boltzmann entropy and the maximum entropy method

Let (X, Σ, μ) be a probability measure space. A nonnegative function in L1  L1(X) such that f1X|f|(x)dμ(x)=1 is called a density. The set of all densities is denoted by D. Letη(u)=-uloguu>0,0u=0.

Definition 2.1

If f  0, then the (Boltzmann) entropy of f is defined asH(f)=Xη(f(x))dμ(x)=-Xf(x)logf(x)dμ(x).

Some basic properties of the entropy are [2], [9]:

  • (i)

    H(f) is either finite or −∞.

  • (ii)

    H : {f  0 : f  L1}  [−∞, ∞) is a proper, upper semicontinuous concave function, strictly concave on its domain that consists of all

The maximum entropy method for Markov operator equations

The maximum entropy idea has been successfully applied to solving Fredholm integral equations [11], computing absolutely continuous invariant measures [4], and estimating the Lyapunov exponents [6]. In this section we use the same approach to solve Markov operator equations. A linear operator P : L1  L1 is called a Markov operator if PD  D. Markov operators describe the evolution of probability densities of dynamical systems. A special subclass of Markov operators is that of Frobenius–Perron

Numerical results

We use the maximum entropy method with a high precision Gaussian quadrature to calculate the stationary density f of a Markov operator. First we apply our method to a Markov operator P : L1(0, 1)  L1(0, 1) with a stochastic kernelPf(x)=01yexyey-1f(y)dy.The unique stationary density f of P isf(x)=ex-1ax,where the constant a01(ex-1)/xdx1.317902151. In implementing our algorithm we needed to find Pxn explicitly. Since P is an integral operator, its dual operator P is given byPg(y)=01yexyey-1g

Conclusions

The maximum entropy method gives an alternative means to numerically determine with good accuracy some desired density of some physical process with even small number of known moments. In most examples, it is comparable or even better than the famous Ulam method for computing invariant measures [10].

References (13)

There are more references available in the full text version of this article.

Cited by (9)

View all citing articles on Scopus
View full text