1]\orgdivSchool of Mathematics, \orgnameJilin University, \orgaddress \cityChangchun, \postcode130012, \stateJilin Province, \countryChina
Abstract
In this paper, the optimal control for discrete-time systems driven by fractional noises is studied. A stochastic maximum principle is obtained by introducing a backward stochastic difference equation contains both fractional noises and the constructed white noises. The solution of the backward stochastic difference equations is also investigated. As an application, the linear quadratic case is considered to illustrate the main results.
1 Introduction
From last century, stochastic maximum principle (SMP) and backward stochastic differential equations (BSDEs) are studied popularly to deal with optimal control problems of stochastic systems. For SMP, some original works refer to [1, 2, 3]. Right up to 1990, Peng [4] obtain the stochastic maximum principle for a general control system, where the control domain could be non-convex and the diffusion term contains control process. Then SMP for many different control problems are investigated, such as near-optimal control [5], doubly stochastic control systems [6, 7], mean-field optimal control [8, 9] and delayed control problems [10, 11]. For BSDEs, as the adjoint equation in SMP, it is originated from Bismut [12]. In 1990, Pardoux and Peng [13] prove the existence and uniqueness of the solution of nonlinear BSDEs. Then some related early works refer to [14, 15].
For SMP for discrete-time systems, the original work is by Lin and Zhang [16], a type of backward stochastic difference equations (BSEs) is introduced as adjoint equation.
Based on this work, much progress has been made by
researchers. Discrete-time stochastic games are studied by Wu and Zhang [17]. Dong et al. obtain the SMP for discrete-time systems with mean-field [18] and delay [19]. Ji and Zhang investigate the infinite horizon recursive optimal control and infinite horizon BSE. But as we known, there are poorly works consider the discrete-time control problems with fractional noises or other “colored noises”.
In this paper, motivated by stochastic optimal control of continuous systems driven by fractional Brownian motion (FBM), we concern the discrete-time optimal control for systems driven by fractional noises.
Let be a fixed constant, which is called Hurst parameter. The -dimensional FBM is a continuous, mean 0 Gaussian process with the covariance
|
|
|
. Moreover, the FBM could be generated by a standard Brownian motion through
|
|
|
for some explicit functions .
For control problem of continuous systems driven by FBM, Han et al. [20] firstly obtain the maximum principle for general systems. Some other related works refer to [21, 22, 23].
For our problem, the state equation is
|
|
|
to minimize the cost function
|
|
|
Here is called a fractional noise describe by the increment of a FBM.
To obtain the stochastic maximum principle, the following BSE is introduced as adjoint equation:
|
|
|
(1.5) |
where is a Gaussian white noises constructed by .
A difficulty is that it is hard to estimate the -norm for and the variation equation appears in section 3, because of the dependence of and its coefficient. We deal with this with more complex calculations, then we prove the uniqueness of the state SE and adjoint BSE, and obtain the maximum principle.
The rest of this paper is organized as follows. In section 2, we introduce the BSE driven by both fractional noise and white noise, and prove the existence and the uniqueness of the solution of this type of BSE.
In section 3, we obtain the stochastic maximum principle by proving the existence and uniqueness of the state SE and showing the convergence of the variation equation.
In section 4, the linear quadratic case is investigated to illustrate the main results.
2 Backward stochastic difference equations
Let be a filtered probability space, be a sub -algebra. is a sequence of fractional noises described by the increment of a -dimensional fractional Brownian motion, namely, . Define the filtration by .
Denote by , or for simplify, the set of all -measurable random variables taking values in , such that . Denote by , or for simplify, the set of all -adapted process such that
|
|
|
Then we consider the following BSE ,
|
|
|
(2.3) |
where for some , and , which will be defined in the following lemma, is a sequence of independent Gaussian random variables generated by . We also assume the following conditions for :
-
(H2.1)
and are adapted processes: for all .
-
(H2.2)
There are some constants , such that
|
|
|
|
|
|
|
|
-
(H2.3)
There exists a -measurable functions , such that and for all .
Then we show the construction and properties of .
Lemma 1.
Let is given by , where is given by
|
|
|
(2.4) |
Then is a sequence of independent Gaussian random variables such that is -measurable but independent with .
Proof.
From the properties of fractional Brownian motion, we know
|
|
|
where . So that there exists invertible matrix such that
|
|
|
(2.5) |
Notice that is not unique, we choose the lower triangular one since is -measurable. But it is hard to obtain directly, we determine at first by checking satisfy equality (2.4).
Let , rewrite (2.5) as
|
|
|
or
|
|
|
Since is -measurable and independent with , for .
Then let for , we obtain the induction formula for :
|
|
|
which equally to (2.4).
∎
Then we show the solvability of BSE (2.3).
Theorem 1.
Assume that assumptions (H2.1)-(H2.3) hold. Then BSE (2.3) has a unique solution.
Proof.
We first define by constructing
|
|
|
Through the Lipschitz’s condition, Hölder’s inequality and the fact , we have
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
(2.6) |
for . Then it is a straightforward result that , so that is a square integrable martingale. Similar to formula (2.5) of , there exists a unique adapted process , such that
|
|
|
Rewrite the above equality as
|
|
|
which shows . Multiply on both side and take , we obtain by
|
|
|
Similar to formula (2), we have . Up to now, we obtain the unique pair .
Then by induction, for all , if for some . Define
|
|
|
for . Then
we show :
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Let and
|
|
|
which implies
|
|
|
Choosing , we obtain the unique process pair .
∎
3 A maximum principle
In this section, we study the optimal control problem for discrete-time systems driven by fractional noises. Let be a filtered probability space, be a sub -algebra and be the filtration defined by . The state equation is
|
|
|
(3.3) |
with the cost function
|
|
|
(3.4) |
Here is independent with ,
and are measurable functions on with values in and , respectively.
and be measurable functions on and , respectively, with values in .
Denote by the set of progressively measurable process
u taking values in a given closed-convex set and satisfying . The problem is to find an optimal control to minimized the cost function, i.e.,
|
|
|
To simplify the notation without losing the generality, we assume that . We give the following assumptions:
-
(H3.1)
and are adapted processes: for all .
-
(H3.2)
are differential w.r.t and there exists some constants , such that
|
|
|
|
|
|
|
|
for
-
(H3.3)
are differential w.r.t and
|
|
|
-
(H3.4)
for all .
Then we show the solvability of the state equation.
Lemma 2.
Assume that assumptions (H3.1)-(H3.4) hold and , then SE (3.3) has a unique solution .
Proof.
Since , by assumptions (3.1) and (3.2), we have
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
for some and .
Then, by induction, if , we show :
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Thus, the solution of (3.3) in .
Uniqueness. Let and be two solutions of SE (3.3), so that . If , we have
|
|
|
|
|
|
|
|
|
|
|
|
for . Then we conclude , which shows the solution is unique.
∎
For any and , let
|
|
|
Denote
|
|
|
|
|
|
for .
Define the variation equation by
|
|
|
Then we have the following convergence result.
Lemma 3.
Let , be the corresponding state equation to , , respectively. Then we have
|
|
|
(3.5) |
and
|
|
|
(3.6) |
Proof.
The proof of (3.5) follows lemma 2, if , then
|
|
|
|
|
|
|
|
|
|
|
|
So that
|
|
|
Denote , it follows
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
where
|
|
|
for . Then
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Since , by induction, we have
|
|
|
which complete the proof of (3.6).
∎
Then we give the stochastic maximum principle for control system (3.3), (3.4).
Theorem 2.
Let assumptions (3.1)-(3.3) hold, be the optimal control and the corresponding state process. Let be the solution to the following BSE:
|
|
|
(3.11) |
Then the following inequality holds:
|
|
|
a.s., for all . Here and are given by lemma 1.
Proof.
According to the assumptions (H3.1)-(H3.2) and remark 3, it is easy to check BSE (3.11) satisfies (H2.1)-(H2.3), so that BSE (3.11) has a unique solution.
Trough lemma 3, it is easy to show the directional derivative of is
|
|
|
(3.12) |
Consider
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
(3.13) |
Notice that
|
|
|
and
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
(3.14) |
so take summation and expectation of equation (3), we have
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
(3.15) |
since .
∎
Substitute equation (3) to (3.12), it follows
|
|
|
|
|
|
|
|
(3.16) |
Notice that, similar to equation (3),
|
|
|
and
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
through equation (3) and the fact
, we conclude
|
|
|
By the arbitrary of , we have
|
|
|
for all , which implies
|
|
|