1. Introduction
In general, it is hard to obtain an analytic solution for nonlinear optimal control problem under state constraint. To solve a constraint optimal control problem, it is rather common to use a direct discretization approach to exact solution for the problem [1] [2]. Mathematically, the direct discretization methods should be supported by the investigation of convergence properties for the solutions of discretized problems approximating to the solution of the continuous problem. One usually expects a desired error between a numerical value and the optimal objective value of the original problem. However, for many engineering applications, the direct discretization methods may be useful to deal with some constrained optimal control problems efficiently, but so far for these methods one still expects more researches on theoretical foundation of convergence results [2]. The purpose of this paper is to provide a convergence result for a nonlinear optimal control problem under state constraint.
In this paper, according to the traditional optimal control theory, an admissible control is measurable and bounded on the interval
such that the ordinary differential equation in the control problem has a unique solution.
We consider three nonlinear optimal control problems under state constraint as follows.
(1.1)
where the cost function
is continuously differentiable on
and
is a convex function on
. For this problem the matrix functions
on
are smooth and the vector
satisfying
is given in the control system in (1.1).
(1.2)
where the cost function
is continuously differentiable on
and
is a convex function on
. For this problem the matrix functions
on
are smooth and the vector
satisfying
is given in the control system in (1.2).
(1.3)
where the parameter
is given and the cost function
is continuously differentiable on
and
is a convex function on
. For this problem the matrix functions
on
are smooth and the vector
satisfying
is given in the control system in (1.3).
Main assumption: In this paper, we assume that the sets of admissible control of all problems concerned in this paper are not empty.
Remark 1.1. Through out the paper by optimal value of the problem we mean the infimum of the cost functional, i.e.
. On the other hand, noting that
is continuously differentiable and that the matrix functions
on
are smooth, by functional analysis we see that the cost functional
is continuous on admissible control space.
The rest of the paper is organized as follows. In Section 2, to deal with the problem
, we present a partial differential equation by rewriting the Hamilton-Jacobi-Bellman equation. Then we create an extremal flow by a differential-algebraic equation to compute the optimal value of the problem
. We prove a conver-gence theorem for an approximation approach to the optimal value of the problem
by a series of optimal values of the problem
with different parameters in Section 3. In Section 4, we provide a convergence result for the optimal value of the problem
by proving that the problem
and the problem
have the same optimal value. We give a proof of Theorem 2.1 in Section 5 and a conclusion in Section 6.
2. A Study on Optimal Control Problem Subject to Mixed
Control-State Constraint
In this section, we deal with the optimal control problem
by a partial differential equation. In the following, the positive number
is fixed. For given
, define a set
(2.1)
Note that if
then
. In the following we assume that
(2.2)
We consider the Hamilton-Jacobi-Bellman equation as follows:
(2.3)
For given
with
, define a function
(2.4)
then for
and
, we have
(2.5)
By (2.5), we can rewrite the Hamilton-Jacobi-Bellman equation in (2.3) with global optimization to obtain the following partial differential equation [3]:
(2.6)
We will solve the optimal control problem
in (1.3) by the partial dif-ferential equation in (2.6).
Given a pair
, satisfying
, i.e.
. For
, in the following let
denote the global minimizer of
, i.e. . We need the following lemma to study the expression of
. For given
and
, we then define auxiliary function
. Again let
denote the global minimizer of
, i.e. . We see that, given
satisfying
, for
,
, we have
. By primary optimi-zation theory we have the following lemma.
Lemma 2.1. Given
and
. If
, then
. On the other hand, if
, then
,
.
For given
such that
, denoting
by
and denoting
by
, by Lemma 2.1, we see that, if
, then
(2.7)
and if
,then
(2.8)
Remark 2.1. By Lemma 2.1 we see that
is continuous with respect to
. We can get a viscosity solution of the partial differential equation in (2.6) [4]-[6]. Then the Hamilton-Jacobi-Bellman equation in (2.3) can be solved for a numerical solution [7].
Definition 2.1. For a solution
of the partial differential equation in (2.6), we call
an extremal flow related to
if it is a solution of the following differential-algebraic equation:
(2.9)
(2.10)
By the same way in [7], we can prove the following theorem.
Theorem 2.1. Let
be a solution of the partial differential equation in (2.6) and
be an extremal flow defined by (2.9), (2.10). Then,
is an optimal control of the problem
, and
is the optimal value of the problem
.
Theorem 2.2. If the continuously differentiable function
is a solution of the partial differential equation in (2.6) on
and
is the function defined in (2.7), (2.8), then
is an optimal feedback control of the problem
.
Proof: Since
is continuous, by (2.7), (2.8), we see that
is continuous on
. By classical theory of ordinary differential equation we see that the equation
(2.11)
has a solution on
. Let the solution of the ODE in (2.11) be denoted by
and let
be denoted by
. By lemma 2.1 and (2.7),(2.8) we see that
(2.12)
Noting (2.11), (2.12), by Definition 2.1, the pair
is an extremal flow related to
. It follows from Theorem 2.1 that
is an optimal feedback control of the problem
.
3. An Approximation Approach to the Optimal Value of
Problem 
In this section we show a convergent result for an approximation approach to the optimal value of problem
which is restated as follows.
(3.1)
where the cost function
is continuously differentiable on
and
is a convex function on
. In this problem the matrix functions
on
are continuously differentiable and the vector
such that
are given in the control system in (3.1).
In the following, for a given positive number
, the optimal value of problem
is denoted by
and the optimal value of problem
is denoted by
.
Lemma 3.1. (i). For each given number
,
. (ii). If
, then
.
Proof: Firstly, let
be an admissible pair of the problem
. Note that the functions
and the vector
appearing in
and
are the same. It follows from the fact
,
that
,
. Thus
is also an admissible pair of the problem
. Consequently,
.
Secondly, let
be an admissible pair of the problem
with the parameter
. Note that functions
and the vector
appearing in
do not depend on different parameter
. Noting
, it follows from the fact
,
that
,
. Thus
is also an admissible pair of the problem
with the parameter
. Consequently,
. The lemma is proved.
Theorem 3.1. For given
there exists a positive number
such that
(3.2)
Proof: Given
. Let
be an admissible pair of the problem
such that
(3.3)
noting that
is the infimum of
for the problem
.
Noting that
is continuous and , we see that there is a
such that
(3.4)
Noting that the admissible control is bounded on
, there is a number
such that
(3.5)
By (3.4), (3.5), there is a
, such that
(3.6)
Thus
is an admissible pair of both problem
and problem
with the parameter
. Then we have
(3.7)
By Lemma 3.1 and (3.3),(3.7) we have
(3.8)
Therefore (3.2) is true and the theorem has been proved.
Corollary 3.1. Let
be a decrease sequence of positive numbers satisfying
when
. Then
(3.9)
Proof: By Lemma 3.1, noting (3.3), (3.6) in the proof of Theorem 3.1, for each
, there is an admissible pair
of the problem
satisfying
(3.10)
noting that
is the infimum of
for the problem
.
Noting that
is continuous and
, we see that there is a
such that
(3.11)
and noting that the admissible control is bounded on
, there is a number
such that
(3.12)
By (3.11), (3.12), there is a
, such that
(3.13)
Thus
is an admissible pair of both problem
and problem
with the parameter
. Then we have
(3.14)
This process (3.10)-(3.14) begins from
. But by Lemma 3.1 we see that if a positive number
is got to satisfy (3.13) then for every positive number
less than
the process (3.10)-(3.14) still works. For
, we choose
as in (3.10)-(3.14). After the step
has been done, in the next
step we choose
such that
(3.15)
Then we see that in this way the positive sequence
is strictly decreasing and tends to zero when
. By the deductive process the same as (3.8) in the proof of Theorem 3.1, or by Lemma 3.1 and (3.10), (3.14), we have for each
,
(3.15)
Therefore we have
(3.16)
with the positive sequence
being strictly decreasing and tending to zero when
. The Corollary 3.1 has been proved.
4. On the Optimal Value of Problem
.
In this section we deal with the problem
:
(4.1)
In the following, the optimal value of the problem
is denoted by
. Recall that in Section 4 the optimal value of the problem
is denoted by
. In the following lemma, noting that
, we define two sets of admissible control as follows:
Lemma 4.1. Under the notations above, we have
(4.2)
Proof: Let
be the solution of the linear equation
corresponding to an admissible control
. It is clear that
. Con-sequently,
(4.4)
The lemma is proved.
In the following lemma we should recall that, in the first section of this paper, we have assumed that the admissible control set
is not empty.
Lemma 4.2. Let
. Then for any admis-sible control
such that
,
, we have, for
,
(4.5)
Proof: Let
and
be the trajectories of the linear system
,
corresponding to
and
respec-tively. Noting that
is convex, we have, for
,
(4.6)
also noting that
and
,
. The lemma is proved.
Theorem 4.1. Let the notations of be as in Lemma 4.1. Then
(4.7)
Proof: Let . We show (4.7) by induc-tion. In the initial step, for given
, we have an admissible control
satisfying ,
and
(4.8)
noting that
is the infimum of
for the problem
.
Recalling Remark 1.1, noting that each admissible control is bounded, the func-tion
in the cost functional for the concerned problems is continuously differentiable and the cost functional
is continuous on the control space. By Lemma 4.2, there exists a number
, such that the control
(4.9)
satisfying
(4.10)
and
(4.11)
Thus
is also an admissible control for the problem
. Then by Lemma 4.1 and (4.10) we have
(4.12)
consequently,
(4.13)
Next the same as in the previous step we have an admissible control which is denoted by
satisfying ,
such that
noting that
is the infimum of
for the problem
. As in the previous step above, we have a number
, such that the control
(4.14)
satisfying
(4.15)
and
(4.16)
Thus
is also an admissible control for the problem
. Then by Lemma 4.1 and (4.15) we have
(4.17)
consequently,
(4.18)
Similar to the process from (4.13) to (4.18), by induction, in this way, for
, when
, it is in the initial step, we have , and when staying in the
-th step, we have,
(4.19)
Thus for each positive integer
, we have
Let
, we have
(4.20)
Therefore (4.7) is true and the theorem has been proved. By Theorem 4.1 and Corollary 3.1, we have the following convergence result.
Corollary 4.1. Let
be a decrease sequence of positive numbers satisfying
when
. Then
5. A proof of Theorem 2.1
By (2.9), (2.10), we have, for
,
(5.1)
Integrating the above equality with respect to
from 0 to
, noting that
,
, we have
(5.2)
and
(5.3)
Now let
be an arbitrary admissible pair of the control system in the problem
. We have, for
,
(5.4)
which implies
. Thus, by (2.5) with
for a
, we have
(5.5)
Then for each
, by the partial deferential equation in (2.6), also noting that
is an arbitrary admissible pair of the control system in the problem
, we have, by (5.5),
(5.6)
Integrating the above inequality over
, noting
,
, by (5.6), we have
(5.7)
By (5.3), (5.7), we have
(5.8)
By (5.8), we see that
is an optimal control for the problem
and
is the optimal value of the problem
.
The theorem has been proved.
6. Conclusion
In this paper, we study nonlinear optimal control problem under state constraint. Firstly we deal with a nonlinear optimal control problem subject to mixed control-state constraint. We try to create a partial differential equation by Hamilton-Jacobi-Bellman equation with global optimization. Then we provide a convergence result for an approximation approach to the optimal value of a nonlinear optimal control problem under state constraint.
Conflicts of Interest
The author declares no conflicts of interest.