Linear Quadratic Control of Backward Stochastic Differential Equation with Partial Information ††thanks: This work supported by the National Natural Science Foundations of China under Grants 61821004, 61633015, 61877062, and 61977043.
Abstract: In this paper, we study an optimal control problem of linear backward stochastic differential equation (BSDE) with quadratic cost functional under partial information. This problem is solved completely and explicitly by using a stochastic maximum principle and a decoupling technique. By using the maximum principle, a stochastic Hamiltonian system, which is a forward-backward stochastic differential equation (FBSDE) with filtering, is obtained. By decoupling the stochastic Hamiltonian system, three Riccati equations, a BSDE with filtering, and a stochastic differential equation (SDE) with filtering are derived. We then get an optimal control with a feedback representation. An explicit formula for the corresponding optimal cost is also established. As illustrative examples, we consider two special scalar-valued control problems and give some numerical simulations.
Keywords: Linear quadratic optimal control; backward stochastic differential equation; filtering; Ricatti equation; feedback representation.
Mathematics Subject Classification: 93E20, 60H10
1 Introduction
A BSDE is an Itô SDE for which a random terminal rather than an initial condition on state has been specified. Bismut [1] first introduced a linear BSDE, which is an adjoint equation of stochastic optimal control problem. Pardoux and Peng [2] extended the linear BSDE to a general case. Since then, there has been considerable attention on related topics and their applications among researchers in mathematical finance and stochastic optimal control. See for example, El Karoui et al. [3], Ma and Yong [4], Kohlmann and Zhou [5].
Since BSDE stems from stochastic control theory, it is very natural and appealing to investigate optimal control problem of BSDE. Moreover, controlled BSDE is expected to have wide and important applications in various fields, especially in mathematical finance. In financial investment, a European contingent claim , which is a random variable, can be thought as a contract to be guaranteed at maturity . Peng [6] and Dokuchaev and Zhou [7] derived local and global stochastic maximum principles of optimality for BSDEs, respectively. Linear quadratic (LQ) optimal control problems of BSDEs have also been investigated. Lim and Zhou [8] discussed an LQ control problem of BSDE with a general setting and gave a feedback representation of the optimal control. Li et al. [9] extended the results in [8] to the case with mean-field term. Huang et al. [10] and Du et al. [11] considered LQ backward mean-field games. Du and Wu [12] concerned a stackelberg game for mean field linear BSDE with quadratic cost functionals.
In this paper, we investigate an LQ control problem of BSDE with partial information, which will be referred as a stochastic backward LQ control problem. We are devoted to deriving the optimal control with a feedback representation and establishing an explicit formula for the corresponding optimal cost. Note that the mentioned papers above are concentrated on the complete information case. The motivation of studying stochastic control problems with partial information arises naturally from the area of financial economics. In a portfolio and consumption problem, let denote the flow of information generated by all market noises. In reality, the information available to an agent maybe less than the one produced by the market noises, that is, , where is the information available to the agent. There are considerable literatures on related topics [13, 14, 15, 16, 17, 18]. In particular, Huang et al. [15] derived a necessary condition for optimality of BSDE with partial information and applied their results to two classes of LQ problems. Wang et al. [16] and Wang et al. [17] concerned LQ problems with partially observable information driven by FBSDE and mean field FBSDE, respectively. LQ non-zero sum stochastic differential game of BSDE is considered in Wang et al. [18]. They obtained feedback Nash equilibrium points by FBSDE and Riccati equation under asymmetric information.
Our work distinguishes itself from existing literatures in the following aspects. (i) Both the generator of dynamic system and the cost functional contain diffusion terms and . Moreover, our results are obtained under some usual conditions (see Assumptions and in section 2). In the literatures on this topic, diffusion terms and are usually assumed not to be contained into the generator (see [15], [17]), or there are some additional conditions to ensure the solvability of Riccati equation (see [16], [18]). (ii) Sufficient and necessary conditions of optimality are established, which provide an expression for optimal control via the solution of stochastic Hamiltonian system. (iii) Explicit representations for optimal control in terms of three Riccati equations, a BSDE with filtering, and an SDE with filtering are obtained, as well as the associated optimal cost. The derivation of associated Riccati equations is extremely different from [8] and [9], since the stochastic Hamiltonian system is an FBSDE with filtering. Moreover, the uniqueness and existence of solution to BSDE (3.5) is first obtained, which is important in deriving explicit representations for optimal control and associated optimal cost. (iv) Last but not least, we consider two special scalar-valued control problems of BSDEs with partial information. In the case of , we obtain explicit solutions of the stochastic Hamiltonian system, as well as related Riccati equations. In the case of , we give some numerical simulations to illustrate our theoretical results.
The rest of this paper is organized as follows. In Section 2, we formulate the stochastic backward LQ control problem and give some preliminary results. Section 3 aims to decouple the associated stochastic Hamiltonian system and derive some Riccati equations. In Section 4, we give explicit representations of optimal control and the associated optimal cost. Section 5 is devoted to solving two special scalar-valued control problems and giving some numerical simulations. Finally, we conclude this paper.
2 Preliminaries
Let be a complete filtered probability space and let be a fixed time horizon. Let be a -valued standard Wiener process, defined on . is a natural filtration of augmented by all -null sets. Let be the filtration generated by a stochastic process . Let be the set of all matrices and be the set of all symmetric matrices. For a matrix , let be its transpose. The inner product on is defined by with an induced norm . In particular, we denote by () the set of all (uniformly) positive definite matrices. For any Euclidean space , we adopt the following notations:
is an -measurable random variable, };
is a bounded function};
is an -adapted stochastic process,
is an -adapted stochastic process and has continuous paths, .
Consider a controlled linear BSDE
(2.1) |
where and , valued in , is a control process.
Introduce an admissible control set
adapted,
Any is called an admissible control.
Assumption : The coefficients of dynamic system satisfy
Under Assumption , dynamic system (2.1) admits a unique solution pair , which is called the corresponding state process, for any (see Pardoux and Peng [2], Yong and Zhou [19]). We introduce a quadratic cost functional
(2.2) |
Assumption : The weighting matrices in cost functional satisfy
Our stochastic backward LQ control problem can be stated as follows.
Problem BLQ. Find a such that
(2.3) |
Any satisfying (2.3) is called an optimal control, and the state process is called an optimal state process. Under Assumptions and , Problem BLQ is uniquely solvable for any terminal state (see Li et al. [9]). We suppressed the time argument in the sequel of this paper wherever necessary, for the sake of notation simplicity. The following theorem is a necessary condition of optimality, which is easy to be obtained from Theorem 3.1 in Huang et al. [15].
Theorem 2.1.
Under Assumptions , if is an optimal control of Problem BLQ and is the corresponding optimal state process, then
(2.4) |
admits a unique solution such that
According to the above analysis, we end up with a stochastic Hamiltonian system
(2.5) |
This is a coupled FBSDE with filtering. Note that the coupling comes from the last equation in (2.5), which is also called a stationarity condition. We point out that in our setting, the stationarity condition involves a conditional expectation, which makes the decoupling of this stochastic Hamiltonian system different and difficult. We now prove the sufficiency of the above result.
Theorem 2.2.
Let Assumption hold. If is an adapted solution to stochastic Hamiltonian system (2.5), then is an optimal control.
Proof.
For any , let be the corresponding state process. Let satisfies
According to the existence and uniqueness of solution to BSDE, we have . With the notation, we derive
where
It is easy to see that under Assumption . Further,
Thus, we have
Then, is an optimal control. ∎
3 Decoupling stochastic Hamiltonian system (2.5)
In this section, we use the decoupling method for general FBSDE introduced in [4] to solve stochastic Hamiltonian system (2.5), which is an FBSDE with filtering. Different from the results in [8], we obtain three Riccati equations, an BSDE with filtering and an SDE with filtering. For simplicity of notation, we denote . To be precise, we assume that
(3.1) |
where is a differential and deterministic matrix-valued function with a terminal condition , and is a stochastic process satisfying the BSDE
for -adapted processes , and . According to Theorem 2.1 in Wang et al. [20] (see also Theorem 5.7 in Xiong [21] and Theorem 8.1 in Liptser and Shiryayev [22] ), we have
Applying Itô formula to (3.1), we get
This implies
(3.2) |
Assuming that is invertible, we have
(3.3) |
Substituting (3.1) and (3.3) into the first equation in (3.2), we obtain
Then satisfies a Riccati equation
(3.4) |
and satisfies a BSDE
(3.5) |
Riccati equation (3.4) admits a unique solution under Assumptions (see [9, 8]). Note that (3.5) is a BSDE with filtering, for which the solvability has not been given in literatures before. We will specified this problem in Section 4. In order to give the optimal control with a feedback representation, we conjecture that
(3.6) |
where and are differential and deterministic matrix-valued functions with initial conditions and , respectively; is a stochastic process satisfying an SDE
where , and are -adapted processes. Note that
where . Hence,
Applying Itô formula to (3.6), we obtain
It yields
Assuming that is invertible, we arrive at
Further, it follows from (3.3) and (3.6) that
Introduce
(3.7) |
(3.8) |
and
(3.9) |
There is a unique solution to Riccati equation (3.7), since Assumptions and hold (see Yong and Zhou [19]). Corollary 4.6 in Lim and Zhou [8] implies that (3.8) admits a unique solution . Once , , and the solution of (3.5) are known, the solvability of (3.9) will be obtained immediately.
4 Explicit representations of optimal control and optimal cost
Now we would like to give explicit formulas of optimal control and associated optimal cost in terms of Riccati equations (3.4), (3.7), (3.8), BSDE (3.5) and SDE (3.9). We first prove that (3.5) admits a unique solution. Consider a BSDE
(4.1) |
where .
We assume that
Assumption : There exists a constant , such that, -a.s., for all , , ,
Assumption : .
Lemma 4.1.
Let Assumptions and hold. For any , BSDE (4.1) admits a unique solution .
Proof.
We first introduce a norm on , which is equivalent to the canonical norm
The parameter will be specified later. For any , the following BSDE
admits a unique solution . We then introduce a mapping : by
For any , , we denote , , and . Applying Itô formula to and taking conditional expectations, we get
Thus we have
Taking , we arrive at
which implies . That is, is a contraction on , endowed with the norm . According to the contraction mapping theorem, we know that there is a unique fixed point , such that , which is exactly the solution of (4.1). We now proceed to prove that . Using Jensen inequality, Hölder’s inequality and Burkholder-Davis-Gundy’s inequality yields
Therefore, we obtain . ∎
Remark 4.1.
We have the following theorem which specifies the solvability of stochastic Hamiltonian system (2.5) and gives some relations between the forward component and the backward components.
Theorem 4.1.
Proof.
Consider the following SDE with filtering
(4.2) |
where and are solutions of (3.4) and (3.5), respectively. According to Theorem 2.1 in Wang et al. [20], we get
(4.3) |
From the theory of linear SDE, (4.3) has a unique solution . Then it follows that (4.2) also admits a unique solution . We define
By using Itô formula, satisfies
with an initial condition . Defining
It is obvious that is a solution to stochastic Hamiltonian system (2.5).
We now turn to prove the uniqueness. Suppose that equation (2.5) admits two solutions and , respectively. Let . Thus satisfies
Applying Itô formula to , we obtain
We adopt the same procedure as in the proof of Theorem 2.2. Since satisfy Assumption , it follows that
Recalling is uniformly positive, it yields
With the equality, satisfies
(4.4) |
It is easy to see that (4.4) admits a unique solution . Then
Hence it follows from the uniqueness of solution that . The proof is completed. ∎
To summarize the above analysis, we establish the following main result.
Theorem 4.2.
Let Assumptions hold and let be given. Let , , be the solutions of Riccati equations (3.4), (3.7) and (3.8), respectively. Let and be the solutions of (3.5) and (3.9), respectively. Then the BSDE with filtering
admits a unique solution . By defining
the 5-tuple is an adapted solution to FBSDE (2.5) and is an optimal control of Problem BLQ. The corresponding optimal cost is
(4.5) | ||||
where is the solution of
Proof.
Applying Itô formula to , we have
With the equality, we derive
Recalling that satisfies (3.5) and applying Itô formula to , we have
We obtain
Then our claims follow. ∎
Remark 4.2.
5 One-dimensional case
In this section, we consider two scalar-valued backward LQ problems with partial information and give more detailed analyses. In the case of , we work out an explicit control problem and show the detailed procedure to obtain the feedback representation of optimal control using our theoretical results. In the case of , we give some numerical simulations to illustrate our theoretical results, since we can not obtain explicit solutions of related stochastic Hamiltonian system and Riccati equation.
5.1 Special case:
Under Assumptions and , let all the coefficients of (2.1) and (2.2) are constants, and
In this case, (2.1) is given by
The cost functional takes the form of
Then Problem BLQ is stated as follows.
Problem BLQA. Find a such that
where the admissible control set is given by
-adapted,
The corresponding stochastic Hamiltonian system reads
(5.1) |
We introduce
and
It is easy to see that
and
(5.2) |
Taking in (5.2), we have
Then it follows from (5.1) and (5.2) that
which admits a unique solution
(5.3) |
with
Further,
(5.4) |
Theorem 2.2 implies that given by (5.4) is an optimal control of Problem BLQA.
In the following, we aim to derive a feedback representation of . For this end, we introduce
(5.5) |
(5.6) |
and
Solving (5.5) and (5.6), we get
and
respectively.
According to Theorem 2.1 in Wang et al. [20], we have
where
Similarly, we derive
where
Then Theorem 4.2 implies that (5.4) admits a feedback representation below
where satisfies
(5.7) |
The corresponding optimal cost is
5.2 Special case:
In this case, (2.1) is written as
Cost functional (2.2) takes the form of
Then Problem BLQ is formulated as follows.
Problem BLQB. Find a such that
The corresponding stochastic Hamiltonian system reads
According to Theorem 4.2, the optimal control is
with
The corresponding Riccati equations are
(5.8) |
Equations (3.5) and (3.9) are reduced to
(5.9) |
and
respectively.

Note that it is hard to obtain a more explicit expression of due to the complexity of (5.8) and (5.9). In the following, we hope to give numerical solutions for this case with certain particular coefficients. Let and . Applying Runge-Kutta method, we generate the dynamic simulations of and , shown in Figure 1.



It seems that there is no existing literature on numerical methods of equation (5.9), which is a BSDE with filtering. Using Theorem 2.1 in Wang et al. [20] again, we get
Applying the numerical method introduced in Ma et al. [24], we generate the dynamic simulations of and , shown in Fig. 2. For more information about numerical methods for BSDEs, please refer to Peng and Xu [25], Zhao et al. [26] and the references therein. The simulation of is also shown in Figure 2.
6 Conclusion
We investigate an LQ control problem of BSDE with partial information, where both the generator of dynamic system and the cost functional contain diffusion terms and . This problem is solved completely and explicitly under some standard conditions. An feedback representation of optimal control and an explicit formula of corresponding optimal cost are given in terms of three Riccati equations, a BSDE with filtering and an SDE with filtering. Moreover, we work out two special scalar-valued control problems to illustrate our theoretical results.
Note that the coefficients in the generator of state equation and the weighting matrices in the cost functional are deterministic. If the coefficients are random, there will be an essential difficulty in solving the case. Since is no longer true if is an -adapted stochastic process. We will investigate the stochastic case in future.
References
- [1] J. M. Bismut, An introductory approach to duality in optimal stochastic control, SIAM Rev. 20 (1978) 62-78.
- [2] E. Pardoux, S. Peng, Adapted solution of a backward stochastic differential equation, Syst. Control Lett. 14 (1) (1990) 55-61.
- [3] N. El. Karoui, S. Peng, M. C. Quenez, Backward stochastic differential equations in finance, Math. Finance 7 (1997) 1-71.
- [4] J. Ma, J. Yong, Forward-backward stochastic differential equations and their applications, Lecture Notes in Math, Springer-Verlag, New York, 1999.
- [5] M. Kohlmann, X. Zhou, Relationship between backward stochastic differential equations and stochastic controls: a linear-quadratic approach, SIAM J. Control Optim. 38 (5) (2000) 1392-1407.
- [6] S. Peng, Backward stochastic differential equations and applications to optimal control, Appl. Math. Optim. 27 (2) (1993) 125-144.
- [7] N. G. Dokuchaev, X. Zhou, Stochastic controls with terminal contingent conditions, J. Math. Anal. Appl. 238 (1999) 143-165.
- [8] A. E. B. Lim, X. Zhou, Linear-quadratic control of backward stochastic differential equations, SIAM J. Control Optim. 40 (2) (2001) 450-474.
- [9] X. Li, J. Sun, J. Xiong, Linear quadratic optimal control problems for mean-field backward stochastic differential equations, Appl. Math. Optim. 80 (2019) 223-250.
- [10] J. Huang, S. Wang, Z. Wu, Backward mean-field Linear-quadratic-gaussian (LQG) games: full and partial information, IEEE Trans. Automat. Control 60 (12) (2016) 3784-3796.
- [11] K. Du, J. Huang, Z. Wu, Linear quadratic mean-field-game of backward stochastic differential systems, Math. Control Relat. Fields 8 (2018) 653-678.
- [12] K. Du, Z. Wu, Linear-quadratic stackelberg game for mean-field backward stochastic differential system and application, Math. Probl. Eng. 17 (2019) 1-17.
- [13] Y. Hu, B. Øksendal, Partial information linear quadratic control for jump diffusions, SIAM J. Control Optim. 47 (4) (2008) 1744-1761.
- [14] Z. Wu, A maximum principle for partially observed optimal control of forward-backward stochastic control systems, Sci. China Infor. Sci. 53 (11) (2010) 2205-2214.
- [15] J. Huang, G. Wang, J. Xiong, A maximum principle for partial information backward stochastic control problems with applications, SIAM J. Control Optim. 48 (4) (2009) 2106-2117.
- [16] G. Wang, Z. Wu, J. Xiong, A linear-quadratic optimal control problem of forward-backward stochastic differential equations with partial information, IEEE Trans. Automat. Control 60 (11) (2015) 2904-2916.
- [17] G. Wang, H. Xiao, G. Xing, An optimal control problem for mean-field forward-backward stochastic differential equation with noisy observation, Automatica 86 (2017) 104-109.
- [18] G. Wang, H. Xiao, J. Xiong, A kind of LQ non-zero sum differential game of backward stochastic differential equation with asymmetric information, Automatica 97 (2018) 346-352.
- [19] J. Yong, X. Zhou, Stochastic Controls: Hamiltonian Systems and HJB Equations, Springer-Verlag, New York, 1999.
- [20] G. Wang, Z. Wu, J. Xiong, An Introduction to Optimal Control of FBSDE with Incomplete Information, Springer-Verlag, New York, 2018.
- [21] J. Xiong, An Introduction to Stochastic Filtering Theory, Oxford University Press, London, 2008.
- [22] R. S. Liptser, A. N. Shiryayev, Statistics of Random Processes, Springer-Verlag, New York, 1977.
- [23] P. Huang, G. Wang, H. Zhang, A partial information linear-quadratic optimal control problem of backward stochastic differential equation with its applications, Sci. China Infor. Sci. 63 (9) (2020) 1-14.
- [24] J. Ma, P. Protter, J. S. Martin, S. Torres, Numerical method for backward stochastic differential equations, Ann. Appl. Probab. 12 (2002) 302-316.
- [25] S. Peng, M. Xu, Numerical algorithms for backward stochastic differential equations with 1-d brownian motion: Convergence and simulations, ESAIM Math. Model. Numer. Anal. 45 (2011) 335-360.
- [26] W. Zhao, L. Chen, S. Peng, A new kind of accurate numerical method for backward stochastic differential equations, SIAM J. Sci. Comput. 28 (4) (2006) 1563-1581.