1. Introduction
To solve an optimal control problem, it is a rather common procedure to use a direct discretization approach to exact solution for the problem ([1]-[3]). One usually expects a desired error between a numerical solution and the infimum of the original problem. However, a direct discretization method is capable of solving some constrained optimal control problem efficiently from an engineering point of view, but there is no theoretical foundation in terms of a convergence result for these methods as yet. For this issue, the authors ([3]) have given an analysis in detail. We deal with state constrained optimal control problems considered in paper by a new mathematical method. The main purpose of this paper is to introduce an extremal flow by a partial differential equation for an constraint optimal control problem to ensure the convergence process before implementing a numerical process.
We consider two constraint optimal control problems as follows:
1)
(1.1)
where
is given and the cost function
is continuously differentiable on
. The matrices
,
and the vector
are given in the linear control system in (1.1). Where
is given and the cost function
is continuously differentiable on
. The matrices
,
and the vector
are given in the linear control system in (1.1).
2)
(1.2)
where
is given and the cost function
is continuously differentiable on
. The matrices
,
and the vector
are given in the linear control system in (1.2).
We suppose that the matrices
and the vectors
appearing in these problems above are the same. But when we consider the problem by means of , we assume
for .
The rest of the paper is organized as follows. In Section 2, to deal with the problem , we present a partial differential equation by rewriting HJB equation. Then we create an extremal flow by a differential-algebraic equation to give an optimal feedback control for the problem . In Section 3, we prove a convergent theorem for an approximation approach to the optimal objective value of the problem by optimal objective values of a series of problem .
2. On Optimal Control Problem with a Partial Differential Equation
In this section, we deal with the optimal control problem by a partial differential equation. In the following, the positive number
is fixed. For given
, define a set
(2.1)
Note that if
then
. In the following, we assume that
(2.2)
We consider the Hamilton-Jacobi-Bellman equation for optimal control theory ([2]) as follows:
(2.3)
For given
, define a function
(2.4)
then given
, for
, we have
(2.5)
By (2.5), we can rewrite the Hamilton-Jacobi-Bellman equation in (2.3) as a partial differential equation as follows ([4] [5]):
(2.6)
We will deal with the optimal control problem in (1.1) by the partial differential equation in (2.6).
Given a pair
, by the definition of
, the relationship
means
Then by primary optimization, for a nonzero vector
,
(2.7)
By (2.4), (2.5), (2.7), we have,
(2.8)
Remark 2.1. By (2.8), we can rewrite the partial differential equation in (2.6) as
(2.9)
Let
be a solution of the partial differential equation in (2.9). If for
is continuous, then
is continuous on
. By the viscosity solution of optimal control theory, there is a viscosity approximation to the solution of the above partial differential equation ([6]-[9]).
Definition 2.1. For a solution
of the partial differential equation in (2.9), we call
an extremal flow related to
if it is a solution of the following differential-algebraic equation:
(2.10)
Theorem 2.1. Let
be a solution of the partial differential equation in (2.9) and
be an extremal flow defined by (2.10). Then,
is an optimal control of the problem ,
is the optimal value of the problem and
.
Proof. By (2.10), we have
(2.11)
By (2.9), (2.10), (2.11), we have, for
,
(2.12)
Integrating the above equality with respect to t from 0 to T, noting
, we have
(2.13)
Let
be an arbitrary admissible pair of the control system in the problem . We have, for
,
(2.14)
which implies
. Thus, by (2.5), (2.7) with
for a
, we have
(2.15)
Then for each
, by the partial deferential equation in (2.9), also noting that
is an arbitrary admissible pair of the control system in the problem , we have
(2.16)
Integrating the above inequality over
, noting
,
, by (2.16), we have
(2.17)
By (2.13), (2.17), we have
(2.18)
By (2.18), we see that
is the optimal value of the problem and
is an optimal control to the problem . But by (4.8), we have
. The theorem has been proved.
Let
be an extremal flow. By the definition of extremal flow in Definition 2.1, we have
(2.19)
We have reached the following result.
Theorem 2.2. Let
be a solution of the partial differential equation in (2.9) and
be an extremal flow. If, for some
, , then
(2.20)
Proof. Since , we see by (2.19) that the vector
is in an opposite direction to . It follows from the inner product on the left side of (2.19) that . Then we have
The theorem has been proved.
Remark 2.2. Let
be a solution of the partial differential equation in (2.9) and
be an extremal flow. By (2.4) we see that if
, then
. If for some
, , then in (2.19) let
. Thus we have
(2.21)
which is an optimal feedback control to the problem
.
Example 2.1. Given a function
, consider the following optimal problem:
(2.22)
where, for the data corresponding to problem , we see that
,
,
,
.
For
,
, define
(2.23)
We have
(2.24)
The corresponding partial differential equation in (2.9) is
(2.25)
with the boundary condition
. A numerical method can be used to find a viscosity solution to the partial differential equation in (2.25) ([10]).
By Theorem 2.2, we can find an extremal flow by the following differential equation
(2.26)
where
(2.27)
which is an optimal feedback control to the problem in (2.22) and can be obtained numerically ([11]) as follows.
Let
be a solution of the partial differential equation in (2.25). We may find a discrete solution of (2.26) with the feedback control presented in (2.27) in the following algorithm. For given positive integer N, let
and
,
. Solve the following discrete equation:
(2.28)
(2.29)
3. A Convergent Result on the Optimal Value of the Problem
In this section for the problems we assume that
and the matrix B is invertible. We consider the following constraint optimal control problem:
(3.1)
where the cost function
is continuously differentiable on
. The matrices
,
and the vector
are given in the linear control system in (3.1).
In the following, for a given positive number
, the optimal value of the problem is denoted by
and the optimal value of the problem is denoted by
.
Lemma 3.1. 1) For every given
,
. 2) If
, then
.
Proof: Firstly, let
be an admissible pair of the problem . Note that the matrices
and the vector
appearing in and are the same. It follows from the fact
,
that
,
. Thus
is also an admissible pair of the problem . Consequently,
.
Secondly, let
be an admissible pair of the problem with
. Note that the matrices
and the vector
appearing in do not depend on the parameter
. Noting
, it follows from the fact
,
that
,
. Thus,
is also an admissible pair of the problem with
. Consequently,
. The lemma is proved.
By Lemma 3.1, if
is a decrease series of positive numbers satisfying
when
, then there exists a number
such that
(3.2)
In the following we need to show that .
Theorem 3.1. Let
be a decrease series of positive numbers satisfying
when
. Let
be an optimal pair of the problem satisfying
. Then
(3.3)
Proof: Given
. Noting , inside a sufficiently small ball which is centered at
we can find a point
such that such that
(3.4)
also noting that
is continuous.
Since
, and the set
is convex, we can define a linear function denoted by
such that the line
is contained in
satisfying
,
, i.e.
Noting that the matrix B is convertible, now we get a control
,
, i.e.
(3.5)
We see that the linear function is continuous. It follows ,
that there is a negative number
such that
(3.6)
On the other hand, since
,
is bounded, we have a positive number L such that
(3.7)
Then, by (3.6), (3.7), noting
, for a sufficiently large positive integer K, in the decrease series
, for
, the positive number
is sufficiently small such that
(3.8)
Thus by (3.7), (3.8), is an admissible pair of the problem with respect to
,
. Noting
,
,
and
, by Lemma 3.1 and (3.4), for
, we have
(3.9)
Since is an arbitrary positive number, we have proved
The theorem has been proved.
4. Conclusion
It is well-known that in general, it is hard to find a state constrained optimal control problem. In this paper, we study an auxiliary optimal control problem for an approximation approach to the optimal objective value of a state constrained optimal control problem. We also suggest a computational approach to deal with this process by partial differential equations. By more works in future, we may consider more state constrained optimal control problems by this mathematical method.
Conflicts of Interest
The author declares no conflicts of interest.