A Study on State Constraint Optimal Control Problems

Abstract

In this paper, we use partial differential equations to deal with constraint optimal control problems. We construct extremal flows by differential-algebraic equation to approximate the optimal objective value of constraint optimal control problems. We prove a convergent theorem for an approximation approach to the optimal objective value of a state-constraint optimal control problem. Subject AreasMathematics

Share and Cite:

Zhu, J.H. (2024) A Study on State Constraint Optimal Control Problems. Open Access Library Journal, 11, 1-9. doi: 10.4236/oalib.1112066.

1. Introduction

To solve an optimal control problem, it is a rather common procedure to use a direct discretization approach to exact solution for the problem ([1]-[3]). One usually expects a desired error between a numerical solution and the infimum of the original problem. However, a direct discretization method is capable of solving some constrained optimal control problem efficiently from an engineering point of view, but there is no theoretical foundation in terms of a convergence result for these methods as yet. For this issue, the authors ([3]) have given an analysis in detail. We deal with state constrained optimal control problems considered in paper by a new mathematical method. The main purpose of this paper is to introduce an extremal flow by a partial differential equation for an constraint optimal control problem to ensure the convergence process before implementing a numerical process.

We consider two constraint optimal control problems as follows:

1)

(1.1)

where β>0 is given and the cost function P( x ) is continuously differentiable on R n . The matrices A R n×n , B R n×m and the vector c R n are given in the linear control system in (1.1). Where β>0 is given and the cost function P( x ) is continuously differentiable on R n . The matrices A R n×n , B R n×m and the vector C R n are given in the linear control system in (1.1).

2)

(1.2)

where M>0 is given and the cost function P( x ) is continuously differentiable on R n . The matrices A R n×n , B R n×n and the vector c R n are given in the linear control system in (1.2).

We suppose that the matrices A,B and the vectors c,a appearing in these problems above are the same. But when we consider the problem by means of , we assume m=n for .

The rest of the paper is organized as follows. In Section 2, to deal with the problem , we present a partial differential equation by rewriting HJB equation. Then we create an extremal flow by a differential-algebraic equation to give an optimal feedback control for the problem . In Section 3, we prove a convergent theorem for an approximation approach to the optimal objective value of the problem by optimal objective values of a series of problem .

2. On Optimal Control Problem with a Partial Differential Equation

In this section, we deal with the optimal control problem by a partial differential equation. In the following, the positive number β is fixed. For given x R n , define a set

S 1 ( x )={ u: u T u c T x β }. (2.1)

Note that if c T x>0 then S 1 ( x )= . In the following, we assume that

S 1 ( x ). (2.2)

We consider the Hamilton-Jacobi-Bellman equation for optimal control theory ([2]) as follows:

v t ( t,x )+ v x T ( t,x )Ax+ min u S 1 ( x ) { v x T ( t,x )Bu }=0,v( T,x )=P( x ). (2.3)

For given ( λ,x ) R m × R n , define a function

H( λ,x ):= min u S 1 ( x ) { λ T u }, (2.4)

then given ( t,x ) R 1 × R n , for λ= B T v x ( t,x ) , we have

H( B T v x ( t,x ),x )= min u S 1 ( x ) { v x T ( t,x )Bu }, (2.5)

By (2.5), we can rewrite the Hamilton-Jacobi-Bellman equation in (2.3) as a partial differential equation as follows ([4] [5]):

v t ( t,x )+ v x T ( t,x )Ax+H( B T v x ( t,x ),x )=0,v( T,x )=P( x ). (2.6)

We will deal with the optimal control problem in (1.1) by the partial differential equation in (2.6).

Given a pair ( x,u ) R n × R m , by the definition of S 1 ( x ) , the relationship u S 1 ( x ) means

u T u c T x β .

Then by primary optimization, for a nonzero vector λ R m ,

min u S 1 ( x ) { λ T u }= min u T u β 1 ( c T x ) { λ T u }= β 1 ( c T x ) λ . (2.7)

By (2.4), (2.5), (2.7), we have,

H( B T v x ( t,x ),x )= min u S 1 ( x ) { v x T ( t,x )Bu }= β 1 ( c T x ) B T v x ( t,x ) . (2.8)

Remark 2.1. By (2.8), we can rewrite the partial differential equation in (2.6) as

v t ( t,x )+ v x T ( t,x )Ax β 1 ( c T x ) B T v x ( t,x ) =0,v( T,x )=P( x ). (2.9)

Let v( t,x ) be a solution of the partial differential equation in (2.9). If for c T x0 v x ( t,x ) is continuous, then H( B T v x ( t,x ),x )= β 1 ( c T x ) B T v x ( t,x ) is continuous on { x: c T x0 } . By the viscosity solution of optimal control theory, there is a viscosity approximation to the solution of the above partial differential equation ([6]-[9]).

Definition 2.1. For a solution v( t,x ) of the partial differential equation in (2.9), we call ( x ^ ( ), u ^ ( ) ) an extremal flow related to v( t,x ) if it is a solution of the following differential-algebraic equation:

{ x ^ ˙ ( t )=A x ^ ( t )+B u ^ ( t ), v x T ( t, x ^ ( t ) )B u ^ ( t )= β 1 c T x ^ ( t ) B T v x ( t, x ^ ( t ) ) , x[ 0,T ], x ^ ( 0 )=a, (2.10)

Theorem 2.1. Let v( t,x ) be a solution of the partial differential equation in (2.9) and ( x ^ ( ), u ^ ( ) ) be an extremal flow defined by (2.10). Then, u ^ ( ) is an optimal control of the problem , P( x ^ ( T ) ) is the optimal value of the problem and P( x ^ ( T ) )=v( 0,a ) .

Proof. By (2.10), we have

v x T ( t, x ^ ( t ) )B u ^ ( t )= β 1 c T x ^ ( t ) B T v x ( t, x ^ ( t ) ) . (2.11)

By (2.9), (2.10), (2.11), we have, for t[ 0,T ] ,

d dt v( t, x ^ ( t ) )= v t ( t, x ^ ( t ) )+ v x T ( t, x ^ ( t ) )A x ^ ( t )+ v x T ( t, x ^ ( t ) )B u ^ ( t ) = v t ( t, x ^ ( t ) )+ v x T ( t, x ^ ( t ) )A x ^ ( t ) β 1 c T x ^ ( t ) B T v x ( t, x ^ ( t ) ) =0. (2.12)

Integrating the above equality with respect to t from 0 to T, noting v( T, x ^ ( T ) )=P( x ^ ( T ) ), x ^ ( 0 )=a , we have

0= 0 T d dt v( t, x ^ ( t ) )dt =P( x ^ ( T ) )v( 0,a ). (2.13)

Let ( x( ),u( ) ) be an arbitrary admissible pair of the control system in the problem . We have, for t[ 0,T ] ,

c T x( t )+β u T ( t )u( t )0, (2.14)

which implies u( t )S( x( t ) ) . Thus, by (2.5), (2.7) with λ= B T v x ( t,x( t ) ) for a t[ 0,T ] , we have

β 1 c T x( t ) B T v x ( t,x( t ) ) v x T ( t,x( t ) )Bu( t ). (2.15)

Then for each t[ 0,T ] , by the partial deferential equation in (2.9), also noting that ( x( ),u( ) ) is an arbitrary admissible pair of the control system in the problem , we have

0= v t ( t,x( t ) )+ v x T ( t,x( t ) )Ax( t ) β 1 c T x( t ) B T v x ( t,x( t ) ) v t ( t,x( t ) )+ v x T ( t,x( t ) )Ax( t )+ v x T ( t,x( t ) )Bu( t ) = v t ( t,x( t ) )+ v x T ( t,x( t ) ) dx( t ) dt = d dt v( t,x( t ) ). (2.16)

Integrating the above inequality over [ 0,T ] , noting v( T,x( T ) )=P( x( T ) ) , x( 0 )=a , by (2.16), we have

0 0 T d dt v( t,x( t ) )dt =P( x( T ) )v( 0,a ). (2.17)

By (2.13), (2.17), we have

P( x ^ ( T ) )P( x( T ) ). (2.18)

By (2.18), we see that P( x ^ ( T ) ) is the optimal value of the problem and u ^ ( ) is an optimal control to the problem . But by (4.8), we have P( x ^ ( T ) )=v( 0,a ) . The theorem has been proved.

Let ( x ^ ( ), u ^ ( ) ) be an extremal flow. By the definition of extremal flow in Definition 2.1, we have

v x T ( t, x ^ ( t ) )B u ^ ( t )= β 1 c T x ^ ( t ) B T v x ( t, x ^ ( t ) ) ,( t[ 0,T ] ). (2.19)

We have reached the following result.

Theorem 2.2. Let v( t,x ) be a solution of the partial differential equation in (2.9) and ( x ^ ( ), u ^ ( ) ) be an extremal flow. If, for some t[ 0,T ] , B T v x ( t, x ^ ( t ) )0 , then

u ^ ( t )= β 1 c T x ^ ( t ) B T v x ( t, x ^ ( t ) ) B T v x ( t, x ^ ( t ) ) . (2.20)

Proof. Since B T v x ( t, x ^ ( t ) )0 , we see by (2.19) that the vector u ^ ( t ) is in an opposite direction to B T v x ( t, x ^ ( t ) ) . It follows from the inner product on the left side of (2.19) that u ^ ( t ) = β 1 c T x ^ ( t ) . Then we have

u ^ ( t )= β 1 c T x ^ ( t ) B T v x ( t, x ^ ( t ) ) B T v x ( t, x ^ ( t ) ) .

The theorem has been proved.

Remark 2.2. Let v( t,x ) be a solution of the partial differential equation in (2.9) and ( x ^ ( ), u ^ ( ) ) be an extremal flow. By (2.4) we see that if λ=0 , then H( λ,x )=0 . If for some t[ 0,T ] , B T v x ( t, x ^ ( t ) )=0 , then in (2.19) let u ^ ( t )=0 . Thus we have

u ^ ( t )={ β 1 c T x( t ) B T v x ( t, x ^ ( t ) ) B T v x ( t, x ^ ( t ) ) , B T v x ( t, x ^ ( t ) )0, 0, B T v x ( t, x ^ ( t ) )=0, (2.21)

which is an optimal feedback control to the problem ( 1 ) .

Example 2.1. Given a function P( x ) C 1 ( R 1 ) , consider the following optimal problem:

{ minP( x( 1 ) ) s.t. x ˙ ( t )=u( t ),( 0 )=1, x( t )+ u 2 ( t )0,x( t ) R 1 ,u( t ) R 1 ,t[ 0,1 ]. (2.22)

where, for the data corresponding to problem , we see that A=0 , B=1 , c=1 , β=1 .

For x R 1 , x0 , define

S( x )={ u R 1 : u 2 x }. (2.23)

We have

min uS( x ) λu= x | λ |. (2.24)

The corresponding partial differential equation in (2.9) is

v t ( t,x ) x | v x ( t,x ) |=0, (2.25)

with the boundary condition v( 1,x )=P( x ) . A numerical method can be used to find a viscosity solution to the partial differential equation in (2.25) ([10]).

By Theorem 2.2, we can find an extremal flow by the following differential equation

x ^ ˙ ( t )= u ^ ( t ),a.e.t[ 0,1 ], (2.26)

where

u ^ ( t )={ x ^ ( t ) , v x ( t, x ^ ( t ) )>0, x ^ ( t ) , v x ( t, x ^ ( t ) )<0, 0, v x ( t, x ^ ( t ) )=0, (2.27)

which is an optimal feedback control to the problem in (2.22) and can be obtained numerically ([11]) as follows.

Let w( t,x ) be a solution of the partial differential equation in (2.25). We may find a discrete solution of (2.26) with the feedback control presented in (2.27) in the following algorithm. For given positive integer N, let i=1,2,,N and t 0 =0 , x 0 =1 . Solve the following discrete equation:

T( x i x i1 )=N u i1 , (2.28)

u i1 ={ x i1 , v x ( t i1 , x i1 )>0, x i1 , v x ( t i1 , x i1 )<0, 0, v x ( t i1 , x i1 )=0. (2.29)

3. A Convergent Result on the Optimal Value of the Problem

In this section for the problems we assume that m=n and the matrix B is invertible. We consider the following constraint optimal control problem:

(3.1)

where the cost function P( x ) is continuously differentiable on R n . The matrices A R n×n , B R n×n and the vector c R n are given in the linear control system in (3.1).

In the following, for a given positive number β , the optimal value of the problem is denoted by V β and the optimal value of the problem is denoted by V ^ .

Lemma 3.1. 1) For every given β>0 , V β V ^ . 2) If β 1 β 2 >0 , then V β 1 V β 2 .

Proof: Firstly, let ( x( ),u( ) ) be an admissible pair of the problem . Note that the matrices A,B and the vector c,a appearing in and are the same. It follows from the fact c T x( t )+β u T ( t )u( t )0 , t[ 0,T ] that c T x( t )0 , t[ 0,T ] . Thus ( x( ),u( ) ) is also an admissible pair of the problem . Consequently, V β V ^ .

Secondly, let ( x( ),u( ) ) be an admissible pair of the problem with β 1 . Note that the matrices A,B and the vector c,a appearing in do not depend on the parameter β . Noting β 1 β 2 >0 , it follows from the fact c T x( t )+ β 1 u T ( t )u( t )0 , t[ 0,T ] that c T x( t )+ β 2 u T ( t )u( t )0 , t[ 0,T ] . Thus, ( x( ),u( ) ) is also an admissible pair of the problem with β 2 . Consequently, V β 1 V β 2 . The lemma is proved.

By Lemma 3.1, if β k ,k=1,2, is a decrease series of positive numbers satisfying β k 0 when k+ , then there exists a number V * such that

lim k+ V β k = V * V ^ . (3.2)

In the following we need to show that V * = V ^ .

Theorem 3.1. Let β k ,k=1,2, be a decrease series of positive numbers satisfying β k 0 when k+ . Let ( x ^ , u ^ ) be an optimal pair of the problem satisfying c T a<0 . Then

lim k+ V β k = V ^ . (3.3)

Proof: Given ϵ>0 . Noting c T x ^ ( T )0 , inside a sufficiently small ball which is centered at x ^ ( T ) we can find a point x ¯ such that c T x ¯ <0 such that

| P( x ¯ )P( x ^ ( T ) ) |<ϵ, (3.4)

also noting that P( x ) is continuous.

Since c T a<0 , c T x ¯ <0 and the set { x: c T x0 } is convex, we can define a linear function denoted by x ¯ ( t ) such that the line { x ¯ ( t ),[ 0,T ] } is contained in { x: c T x<0 } satisfying x ¯ ( 0 )=a , x ¯ ( T )= x ¯ , i.e.

x ¯ ( t )=( 1 t T )a+ t T x ¯ ,t[ 0,T ].

Noting that the matrix B is convertible, now we get a control u ¯ ( t )= B 1 { x ¯ ˙ ( t )A x ¯ ( t ) } , t[ 0,T ] , i.e.

x ¯ ˙ ( t )=A x ¯ ( t )+B u ¯ ( t ),t[ 0,T ], x ¯ ( 0 )=a. (3.5)

We see that the linear function c T x ¯ ( t ) is continuous. It follows c T x ¯ ( t )<0 , t[ 0,T ] that there is a negative number γ such that

max t[ 0,T ] c T x ¯ ( t )γ<0. (3.6)

On the other hand, since u ¯ ( t )= B 1 { x ¯ ˙ ( t )A x ¯ ( t ) } , t[ 0,T ] is bounded, we have a positive number L such that

max t[ 0,T ] u ¯ T ( t ) u ¯ ( t )L. (3.7)

Then, by (3.6), (3.7), noting β k 0 , for a sufficiently large positive integer K, in the decrease series { β k ,k=1,2, } , for k>K , the positive number β k >0 is sufficiently small such that

c T x ¯ ( t )+ β k u ¯ T ( t ) u ¯ ( t )<γ+ β k L<0,t[ 0,T ]. (3.8)

Thus by (3.7), (3.8), ( x ¯ ( t ), u ¯ ( t ) ),t[ 0,T ] is an admissible pair of the problem with respect to β k , k>K . Noting P( x ^ ( T ) )= V ^ , P( x ¯ ( T ) ) V β k , k>K and x ¯ ( T )= x ¯ , by Lemma 3.1 and (3.4), for k>K , we have

0 V β k V ^ = V β k P( x ^ ( T ) )P( x ¯ ( T ) )P( x ^ ( T ) )=P( x ¯ )P( x ^ ( T ) )<ϵ. (3.9)

Since is an arbitrary positive number, we have proved

lim k+ V β k = V ^ .

The theorem has been proved.

4. Conclusion

It is well-known that in general, it is hard to find a state constrained optimal control problem. In this paper, we study an auxiliary optimal control problem for an approximation approach to the optimal objective value of a state constrained optimal control problem. We also suggest a computational approach to deal with this process by partial differential equations. By more works in future, we may consider more state constrained optimal control problems by this mathematical method.

Conflicts of Interest

The author declares no conflicts of interest.

Conflicts of Interest

The author declares no conflicts of interest.

References

[1] Pontryagin, L.S. (1964) The Mathematical Theory of Optimal Processes. Pergamon Press.
[2] Sontag, E.D. (1998) Mathematical Control Theory: Deterministic Finite Dimensional Systems. 2nd Edition, Springer.
[3] Martens, B. and Gerdts, M. (2020) Convergence Analysis for Approximations of Optimal Control Problems Subject to Higher Index Differential-Algebraic Equations and Mixed Control-State Constraints. SIAM Journal on Control and Optimization, 58, 1-33.
https://doi.org/10.1137/18m1219382
[4] Zhu, J. (2017) A Feedback Optimal Control by Hamilton-Jacobi-Bellman Equation. European Journal of Control, 37, 70-74.
https://doi.org/10.1016/j.ejcon.2017.05.007
[5] Zhu, J. (2018) Singular Optimal Control by Minimizer Flows. European Journal of Control, 42, 32-37.
https://doi.org/10.1016/j.ejcon.2018.03.001
[6] Zhu, J. (2021) A Computational Approach to Non-Smooth Optimization by Diffusion Equations. Journal of Computational and Applied Mathematics, 384, Article ID: 113166.
https://doi.org/10.1016/j.cam.2020.113166
[7] Crandall, M.G. and Lions, P. (1983) Viscosity Solutions of Hamilton-Jacobi Equations. Transactions of the American Mathematical Society, 277, 1-42.
https://doi.org/10.1090/s0002-9947-1983-0690039-8
[8] Bardi, M. and Capuzzo-Dolcetta, I. (1997) Optimal Control and Viscosity Solutions of Hamilton-Jacobi-Bellman Equations. Birkhauser.
[9] Fleming, W.H. (1969) The Cauchy Problem for a Nonlinear First Order Partial Differential Equation. Journal of Differential Equations, 5, 515-530.
https://doi.org/10.1016/0022-0396(69)90091-6
[10] Zhu, J. (2021) A Computational Approach to Optimal Control Problems Subject to Mixed Control-State Constraints. International Journal of Control, 96, 41-47.
https://doi.org/10.1080/00207179.2021.1978556
[11] Burden, R.L. and Faires, J.D. (1989) Numerical Analysis. 4th Edition, Weder and Schmidt.

Copyright © 2024 by authors and Scientific Research Publishing Inc.

Creative Commons License

This work and the related PDF file are licensed under a Creative Commons Attribution 4.0 International License.