Elsevier

Information Sciences

Volume 282, 20 October 2014, Pages 167-179
Information Sciences

Neural-network-based robust optimal control design for a class of uncertain nonlinear systems via adaptive dynamic programming

https://doi.org/10.1016/j.ins.2014.05.050Get rights and content

Abstract

In this paper, the neural-network-based robust optimal control design for a class of uncertain nonlinear systems via adaptive dynamic programming approach is investigated. First, the robust controller of the original uncertain system is derived by adding a feedback gain to the optimal controller of the nominal system. It is also shown that this robust controller can achieve optimality under a specified cost function, which serves as the basic idea of the robust optimal control design. Then, a critic network is constructed to solve the Hamilton–Jacobi–Bellman equation corresponding to the nominal system, where an additional stabilizing term is introduced to verify the stability. The uniform ultimate boundedness of the closed-loop system is also proved by using the Lyapunov approach. Moreover, the obtained results are extended to solve decentralized optimal control problem of continuous-time nonlinear interconnected large-scale systems. Finally, two simulation examples are presented to illustrate the effectiveness of the established control scheme.

Introduction

In practical control systems, model uncertainties arise frequently and can severely degrade the closed-loop system performance. Hence, the problem of designing robust controller for nonlinear systems with uncertainties has drawn considerable attention in recent literature [43], [15], [31]. Lin et al. [15] showed that the robust control problem can be solved by studying the optimal control problem of the corresponding nominal system, but the detailed procedure was not presented. In [31], the authors developed an iterative algorithm for online design of robust control for a class of continuous-time nonlinear systems. However, the optimality of the robust controller with respect to a specified cost function was not discussed. In [43], the authors addressed the problem of designing robust tracking controls for a class of uncertain nonholonomic systems actuated by brushed direct current motors, while the research was not related with the optimality.

The starting point of the obtained strategy of this paper is optimal control. The nonlinear optimal control problem always requires to solve the Hamilton–Jacobi–Bellman (HJB) equation. Though dynamic programming has been a conventional method in solving optimization and optimal control problems, it often suffers from the curse of dimensionality, which was primarily due to the backward-in-time approach. To avoid the difficulty, based on function approximators, such as neural networks, adaptive/approximate dynamic programming (ADP) was proposed by Werbos [35] as a method to solve optimal control problems forward-in-time. Recently, the study on ADP and related fields have gained much attention from various scholars [1], [2], [3], [4], [5], [6], [7], [8], [9], [10], [12], [13], [14], [16], [17], [18], [19], [20], [21], [22], [23], [24], [25], [28], [29], [30], [32], [33], [34], [36], [37], [38], [40], [41], [42], [44], [45], [46]. Lewis and Vrabie [13] stated that the ADP technique is closely related to the field of reinforcement learning. As is known to all, policy iteration is one of the basic algorithms of reinforcement learning. In addition, the initial admissible control is necessary when employing the policy iteration algorithm. However, in many situations, it is difficult to find the initial admissible control.

To the best of our knowledge, there are few results on robust optimal control of uncertain nonlinear systems based on ADP, not to mention the decentralized optimal control of large-scale systems. This is the motivation of our research. Actually, it is the first time that the robust optimal control scheme for a class of uncertain nonlinear systems via ADP technique and without using an initial admissible control is established. To begin with, the optimal controller of the nominal system is designed. It can be proved that the modification of optimal control law is in fact the robust controller of the original uncertain system, which also achieves optimality under the definition of a cost function. Then, a critic network is constructed for solving the HJB equation corresponding to the nominal system. In addition, inspired by the work of [5], [24], an additional stabilizing term is introduced to verify the stability, which relaxes the need for an initial stabilizing control. The uniform ultimate boundedness (UUB) of the closed-loop system is also proved via the Lyapunov approach. Furthermore, the aforementioned results are extended to deal with the decentralized optimal control for a class of continuous-time nonlinear interconnected systems. At last, two simulation examples are given to show the effectiveness of the robust optimal control scheme.

Section snippets

Problem statement and preliminaries

In this paper, we study the continuous-time uncertain nonlinear systems given byẋ(t)=f(x(t))+g(x(t))(u¯(t)+d¯(x(t))),where x(t)Rn is the state vector and u¯(t)Rm is the control vector, f(·) and g(·) are differentiable in their arguments with f(0)=0, and d¯(x) is the unknown nonlinear perturbation. Let x(0)=x0 be the initial state. We assume that d¯(0)=0, so that x=0 is an equilibrium of system (1). As in many other literature, for the nominal systemẋ(t)=f(x(t))+g(x(t))u(t),we also assume

Robust optimal control design of uncertain nonlinear systems

In this section, for establishing the robust stabilizing control strategy of system (1), we modify the optimal control law (8) of system (2) by proportionally increasing a feedback gain, i.e.,u¯(x)=ζu(x)=-12ζR-1gT(x)J(x).Now, we present the following lemma to indicate that the optimal control has infinite gain margin.

Lemma 1

For system (2), the feedback control given by (12) ensures that the closed-loop system is asymptotically stable for all ζ1/2.

Proof

We show that J(x) is a Lyapunov function. In light

Optimal control design via ADP approach and the stability proof

According to the universal approximation property of neural networks, J(x) can be reconstructed by a single-layer neural network on a compact set Ω asJ(x)=ωcTσc(x)+εc(x),where ωcRl is the ideal weight, σc(x)Rl is the activation function, l is the number of neurons in the hidden layer, and εc(x) is the approximation error. Then, we haveJ(x)=(σc(x))Tωc+εc(x).Based on (24), the Lyapunov Eq. (4) becomes0=dM2(x)+uT(x)Ru(x)+ωcTσc(x)+(εc(x))T(f(x)+g(x)u(x)).In light of [28], [4], [5], in

Decentralized optimal control design of nonlinear interconnected systems

Large-scale systems are common in engineering area when doing research on complex dynamical systems that can be partitioned into a set of interconnected subsystems. The decentralized control is one of the effective design approaches and has attracted a great amount of interest due to its advantages in easier implementation and lower dimensionality [17], [10], [26], [27], [39]. In this section, we generalize the aforementioned results to decentralized optimal control for a class of

Simulation studies

Two examples are provided in this section to demonstrate the effectiveness of the robust optimal control strategy.

Example 1

Consider the following continuous-time nonlinear system:ẋ=-0.5x1+x2(1+0.5x22)-0.8(x1+x2)+0.5x2(1-0.3x22)+0-0.6(u¯+d¯(x)),where x=[x1,x2]TR2 and u¯R are the state and control variables, respectively. The term d¯(x)=δ1x2cos(δ2x1+δ3x2) reflects the uncertainty of the controlled plant, where δ1, δ2, and δ3 are unknown parameters with δ1[-1,1],δ2[-5,5], and δ3[-3,3]. We set R=I and

Conclusion

A novel robust optimal control scheme for a class of uncertain nonlinear systems via ADP approach is developed in this paper. It is proved that the robust controller of the original uncertain system achieves optimality under a specified cost function. During the implementation process, a critic network is constructed to solve the HJB equation of the nominal system and an additional stabilizing term is introduced to verify the stability. The obtained results are also extended to design the

References (46)

  • D. Wang et al.

    Neuro-optimal control for a class of unknown nonlinear dynamic systems using SN-DHP technique

    Neurocomputing

    (2013)
  • D. Wang et al.

    Finite-horizon neuro-optimal tracking control for a class of discrete-time nonlinear systems using adaptive dynamic programming approach

    Neurocomputing

    (2012)
  • D. Wang et al.

    Optimal control of unknown nonaffine nonlinear discrete-time systems based on adaptive dynamic programming

    Automatica

    (2012)
  • H.N. Wu et al.

    Simultaneous policy update algorithms for learning the solution of linear continuous-time H state feedback control

    Inform. Sci.

    (2013)
  • D. Xu et al.

    Decentralized measurement feedback stabilization of large-scale systems via control vector Lyapunov functions

    Syst. Control Lett.

    (2013)
  • X. Xu et al.

    Reinforcement learning algorithms with function approximation: recent advances and applications

    Inform. Sci.

    (2014)
  • H.M. Yen et al.

    Design of a robust neural network-based tracking controller for a class of electrically driven nonholonomic mechanical systems

    Inform. Sci.

    (2013)
  • A. Al-Tamimi et al.

    Discrete-time nonlinear HJB solution using approximate dynamic programming: convergence proof

    IEEE Trans. Syst. Man Cybernet. – Part B: Cybernet.

    (2008)
  • D.P. Bertsekas et al.

    Missile defense and interceptor allocation by neuro-dynamic programming

    IEEE Trans. Syst. Man Cybernet. – Part A: Syst. Humans

    (2000)
  • T. Dierks, S. Jagannathan, Optimal control of affine nonlinear continuous-time systems, in: Proceedings of the American...
  • T. Dierks et al.

    Online optimal control of affine nonlinear discrete-time systems with unknown internal dynamics by using time-based policy update

    IEEE Trans. Neural Netw. Learn. Syst.

    (2012)
  • J. Fu et al.

    Adaptive learning and control for MIMO system based on adaptive dynamic programming

    IEEE Trans. Neural Netw.

    (2011)
  • A. Heydari et al.

    Finite-horizon control-constrained nonlinear optimal control using single network adaptive critics

    IEEE Trans. Neural Netw. Learn. Syst.

    (2013)
  • Cited by (139)

    View all citing articles on Scopus

    This work was supported in part by the National Natural Science Foundation of China under Grants 61034002, 61233001, 61273140, 61304086, and 61374105, in part by Beijing Natural Science Foundation under Grant 4132078, and in part by the Early Career Development Award of SKLMCCS.

    View full text