Next Article in Journal
Non-Linear Dynamic Movements of CNT/Graphene/Aluminum Oxide and Copper/Silver/Cobalt Ferrite Solid Particles in a Magnetized and Suction-Based Internally Heated Surface: Sensitivity and Response Surface Optimization
Next Article in Special Issue
Optimality of a Network Monitoring Agent and Validation in a Real Probe
Previous Article in Journal
Spheres and Tori as Elliptic Linear Weingarten Surfaces
Previous Article in Special Issue
A Queueing Model for Traffic Flow Control in the Road Intersection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Closed-Loop Solvability of Stochastic Linear-Quadratic Optimal Control Problems with Poisson Jumps

School of Mathematics, Shandong University, Jinan 250100, China
*
Author to whom correspondence should be addressed.
Mathematics 2022, 10(21), 4062; https://doi.org/10.3390/math10214062
Submission received: 30 September 2022 / Revised: 27 October 2022 / Accepted: 27 October 2022 / Published: 1 November 2022
(This article belongs to the Special Issue Stochastic Control Systems: Theory and Applications)

Abstract

:
The stochastic linear–quadratic optimal control problem with Poisson jumps is addressed in this paper. The coefficients in the state equation and the weighting matrices in the cost functional are all deterministic but are allowed to be indefinite. The notion of closed-loop strategies is introduced, and the sufficient and necessary conditions for the closed-loop solvability are given. The optimal closed-loop strategy is characterized by a Riccati integral–differential equation and a backward stochastic differential equation with Poisson jumps. A simple example is given to demonstrate the effectiveness of the main result.

1. Introduction

The stochastic linear–quadratic (SLQ) optimal control problem plays an extremely important role in modern control theory and methodology, because of its elegant structure of solutions and wide applications in engineering, finance, networks, etc. More importantly, SLQ optimal control problems can also reasonably approximate some nonlinear stochastic optimal control problems. In the literature for SLQ optimal control problems, refer to Wonham [1], Bismut [2], Bensoussan [3], Peng [4], Chen et al. [5], Chen and Zhou [6], Chen and Yong [7], Ait Rami et al. [8], Tang [9], Yu [10], Tang [11], Sun et al. [12], Sun and Yong [13], and Sun et al. [14], for journal papers and Davis [15], Anderson and Moore [16], Yong and Zhou [17], and Sun and Yong [18] for monographs.
In the above work, stochastic systems are modelled by Brownian motions. However, in reality, Brownian noises are usually inadequate in a mathematical modeling sense. For example, it is particularly appropriate to use stochastic systems with Poisson jumps or Lévy jumps to describe the large fluctuations in the stock market (Merton [19], Kou [20], Cont and Tankov [21], Oksendal and Sulem [22], Lim [23], Hanson [24]). Moreover, from a mathematical point of view, there exist essential differences between stochastic systems with and without jumps.
SLQ optimal control problems with Poisson jumps (SLQP optimal control problems) are also researched by many authors. Tang and Hou [25] studied an optimal control problem of partially observed linear–quadratic stochastic systems with a Poisson process and obtained an explicit solution of this problem by the partially observed maximum principle. Wu and Wang [26] studied a kind of SLQP optimal control problem, the explicit form of optimal controls is obtained by the solutions to a forward–backward stochastic differential equation with Poisson jumps (FBSDEP) and a generalized Riccati equation system. Hu and Oksendal [27] studied an SLQP optimal control problem with partial information. Meng [28] considered an SLQP optimal control problem with random coefficients. The state feedback representation was obtained for the open-loop optimal control by a matrix-valued backward stochastic Riccati equations with jumps (BSREJ), and the solvability of it in a special case was discussed. The solvability of BSREJ in the general case was studied in Zhang et al. [29]. Note that Li et al. [30] gave the concept of relax compensator, which is used to describe indefinite BSREJ, then they investigated the solvability of BSREJ and gave the optimal control. Moon and Chung [31] studied the indefinite SLQP optimal control problem with random coefficients by a completion of squares approach.
Our interest in this paper is the closed-loop solvability of the SLQP optimal control problem, which, as far as we know, is not researched in the literature. In 2014, the notions of open-loop and closed-loop solvabilities for SLQ optimal control problems were introduced in Sun and Yong [32], where they concentrated on the LQ zero-sum stochastic differential game, in which SLQ optimal control problem is a special case when there is only one player/controller is considered. Sun et al. [12] further gave more detailed necessary and sufficient conditions of the open-loop and closed-loop solvability for SLQ optimal control problems. Sun and Yong [13] studied the open-loop and closed-loop solvability of SLQ optimal control problems in the infinite horizon and showed that open-loop and closed-loop solvabilities are equivalent in the infinite horizon. For more details and complete content, please also refer to their book [18]. Li et al. [33] studied the SLQ optimal control problem of mean-field type and gave the characterization of the closed-loop optimal strategy. Lv [34,35] researched the closed-loop solvabilities of SLQ optimal control problems for systems governed by stochastic evolution equations (SEEs) and SEEs of mean-field type, respectively. Tang et al. [36] studied the open-loop and closed-loop solvability for indefinite SLQP optimal control problem of mean-field type and its application in finance. Sun et al. [14] considered the indefinite SLQ optimal control problem with random coefficients and investigated the closed-loop representation of open-loop optimal controls.
Our work differs from the existing results in the following respects. (1) We consider an SLQP optimal control problem with deterministic coefficients in a general framework (problem (SLQP) in Section 2), where the weighting matrices in the cost functional are allowed to be indefinite. Moreover, cross-product terms in the control and state processes are present in the cost functional. Non-homogenous terms also appear in the controlled state equation and cost functional.The model considered in this paper is a nontrivial generalization of those in [12,32]. (2) Characterization of the closed-loop solvability for the SLQP optimal control problem is obtained, via the Riccati integral–differential equation (RIDE). For the SLQ optimal control problem without Poisson jumps, Sun et al. [12] first found two matrix-valued SDE of X ( · ) , Y ( · ) , then they applied Itô’s formula to find X 1 ( · ) . The solution to the related Riccati equation is defined as P ( · ) = Y ( · ) X 1 ( · ) . However, this method fails in our SLQP optimal control problem, as the Poisson jumps appear in the controlled system and difficulty is encountered. In detail, when we take the inverse of the matrix-valued SDEP as that of the matrix-valued SDE in [12] and apply Itô–Wentzell’s formula (Oksendal and Zhang [37]), terms such as F ( e ) + G ( e ) Θ ¯ + 1 1 will appear in X 1 ( · ) . Because we do not have any restrictions on the coefficients in our system and Θ ¯ ( · ) is the closed-loop optimal strategy that we are going to seek, there is no reason to arbitrarily presume that F ( e ) + G ( e ) Θ ¯ + 1 is invertible. From Lv [34,35], we overcome this difficulty by transforming the original problem (SLQP) into a problem of solving the open-loop optimal control of problem (SLQP) Θ ¯ 0 in Section 2. Thus, a Lyapunov integral–differential Equation (25) is given first and then the RIDE (29) is obtained. Note that the technique used in this paper is also different from that in Tang et al. [36], where a matrix minimum principle by Athans [38] was used when dealing with the SLQP problem of mean-field type.
The rest of this paper is organized as follows. Section 2 begins with the preparation work, including giving some basic knowledge and presenting the formulation of the SLQP optimal control problem. In Section 3, characterizations of closed-loop solvability of SLQP optimal control problems are presented and the concrete proofs are given. Section 4 gives an example to demonstrate the effectiveness of the main result. Finally, in Section 5, some concluding remarks are given.

2. Problem Formulation and Preliminaries

First, let us introduce some notations that will be used throughout this paper.
Let T > 0 be a constant, and [ 0 , T ] is a finite time duration. Let R n × m be the collection of all ( n × m ) matrices, and S n × n be the collection of all ( n × n ) symmetric matrices. We let I be the identity matrix with a suitable size. We use · , · to denote inner products in possibly different Hilbert spaces and | · | to denote the norm induced by these inner products. Let M and R ( M ) be the transpose and range of a matrix M, respectively. For M , N S n × n , M N (respectively, M > N ) implies that M N is a positive semi-definite matrix (respectively, positive definite matrix). Let M denote the pseudo-inverse of a matrix M R m × n . If the inverse M 1 of M R n × n exists, then the pseudo-inverse is equal to the inverse. See Penrose [39] for the definition and some basic properties of the pseudo-inverse.
For any Banach space H (for example, H = R n , R n × m , S n × n ) and t [ 0 , T ) , let L p ( t , T ; H ) ( 1 p ) be the space of all L p -integrable functions valued from H on [ t , T ] , C ( [ t , T ] ; H ) be the space of all continuous functions valued from H on [ t , T ] , and L ( t , T ; H ) be the space of Lebesgue measurable, essentially bounded functions from [ t , T ] into H.
Let ( Ω , F , F , P ) be a completed filtered probability space, where F = { F t } t 0 is filtration generated by the following two mutually independent stochastic processes and augmented by all the P -null sets in F :
  • A standard one-dimensional Brownian motion W = { W ( t ) ; 0 t < } .
  • A Poisson random measure N defined on E × R + , where E R l is a nonempty open set and its Borel field is B ( E ) . The compensator of N is N ^ ( d e d t ) = π ( d e ) d t , satisfying π ( A ) < , such that N ˜ ( A × [ 0 , t ] ) = ( N N ^ ) ( A × [ 0 , t ] ) t 0 is a martingale for any A B ( E ) ; π is assumed to be a σ -finite measure on ( E , B ( E ) ) and is called the characteristic measure.
For t [ 0 , T ) , we introduce some notation for spaces of random variables and stochastic processes:
L F s 2 ( Ω ; H ) = ξ : Ω H | ξ is F s measurable   random   variable , s . t . , E | ξ | 2 < , s [ t , T ] , L F 2 ( t , T ; H ) = f : [ t , T ] × Ω H | f is F progressively   measurable , s . t . , E t T | f ( s ) | 2 d s < , L F , p 2 ( t , T ; H ) = f : [ t , T ] × Ω H | f is F predictable, , s . t . , E t T | f ( s ) | 2 d s < , F p ( t , T ; H ) = f : [ t , T ] × E H | f satisfies sup t s T , e E | f ( s , e ) | < , F F , p 2 ( t , T ; H ) = f : Ω × [ t , T ] × E H | f is F predictable , s . t . , E t T E | f ( · , s , e ) | 2 π ( d e ) d s < .
We consider the following controlled linear SDEP on [ t , T ] :
d X ( s ) = A ( s ) X ( s ) + B ( s ) u ( s ) + b ( s ) d s + C ( s ) X ( s ) + D ( s ) u ( s ) + σ ( s ) d W ( s ) + E F ( s , e ) X ( s ) + G ( s , e ) u ( s ) + f ( s , e ) N ˜ ( d e d s ) , s [ t , T ] , X ( t ) = x ,
where t [ 0 , T ) is the initial time and x L F t 2 ( Ω ; R n ) is the given initial state; A ( · ) , B ( · ) , C ( · ) , D ( · ) are given deterministic matrix-valued functions of proper dimensions, and F ( · , · ) , G ( · , · ) are independent of ω . The expressions b ( · ) , σ ( · ) are F -progressively measurable processes and f ( · , · ) is also random; u ( · ) is the control process. We define the admissible control set:
U [ t , T ] = { u : [ t , T ] × Ω R m | u i s F p r e d i c t a b l e , E t T | u ( s ) | 2 d s < } .
The control process u ( · ) U [ t , T ] is called an admissible control.
Then we define the cost functional:
J ( t , x ; u ( · ) ) = E { t T [ Q ( s ) X ( s ) , X ( s ) + 2 S ( s ) X ( s ) , u ( s ) + R ( s ) u ( s ) , u ( s ) + 2 q ( s ) , X ( s ) + 2 ρ ( s ) , u ( s ) ] d s + H X ( T ) , X ( T ) + 2 g , X ( T ) } ,
where H is a symmetric matrix and Q ( · ) , S ( · ) , and R ( · ) are deterministic matrix-valued functions of proper dimensions that satisfy Q ( · ) = Q ( · ) , R ( · ) = R ( · ) . The expression g is an F T -measurable random variable, q ( · ) is an F -progressively measurable process, and ρ ( · ) is an F -predictable process.
In order to make the given (1) and (3) meaningful, we adopt some assumptions for coefficients as follows.
Hypothesis 1.
The coefficients of the state Equation (1) satisfy the following:
A ( · ) L ( t , T ; R n × n ) , B ( · ) L ( t , T ; R n × m ) , b ( · ) L F 2 ( t , T ; R n ) , C ( · ) L ( t , T ; R n × n ) , D ( · ) L ( t , T ; R n × m ) , σ ( · ) L F 2 ( t , T ; R n ) , F ( · , · ) F p ( t , T ; R n × n ) , G ( · , · ) F p ( t , T ; R n × m ) , f ( · , · ) F F , p 2 ( t , T ; R n ) .
Hypothesis 2.
The weighting coefficients of the cost functional (3) satisfy the following:
Q ( · ) L ( t , T ; S n × n ) , S ( · ) L ( t , T ; R m × n ) , R ( · ) L ( t , T ; S m × m ) , q ( · ) L F 2 ( t , T ; R n ) , ρ ( · ) L F , p 2 ( t , T ; R m ) , g L F T 2 ( Ω ; R n ) , H S n × n .
For simplicity, we denote the above Hypothesis 1 and Hypothesis 2 as (H1) and (H2), respectively.
Under (H1) and (H2), for any given x L F t 2 ( Ω ; R n ) and u ( · ) U [ t , T ] L F , p 2 ( t , T ; R m ) , state Equation (1) admits a unique adapted solution X ( · ) L F 2 ( t , T ; R n ) and the cost functional is well-defined. Therefore, the following problem is meaningful.
Problem 1. (SLQP). For given initial pair ( t , x ) [ 0 , T ] × L F t 2 ( Ω ; R n ) , find a u ¯ ( · ) U [ t , T ] such that
J ( t , x ; u ¯ ( · ) ) = inf u ( · ) U [ t , T ] J ( t , x ; u ( · ) ) V ( t , x ) .
Any u ¯ ( · ) U [ 0 , T ] satisfying (4) is called an open-loop optimal control of problem (SLQP) for ( t , x ) , the corresponding X ¯ ( · ) X ( · ; t , x , u ¯ ( · ) ) is called an open-loop optimal state, and ( X ¯ ( · ) , u ¯ ( · ) ) is called an open-loop optimal pair. The map V ( · , · ) is called the value function of problem (SLQP).
In particular, when b ( · ) , σ ( · ) , f ( · , · ) , q ( · ) , ρ ( · ) , and g are all zero, we refer to the above problem as the problem (SLQP) 0 .
Definition 1.
For given initial pair ( t , x ) [ 0 , T ] × L F t 2 ( Ω ; R n ) , if there exists a (unique) u ¯ ( · ) U [ t , T ] such that (4) holds, then we say that problem (SLQP) is (uniquely) open-loop solvable for ( t , x ) .
Next, take Θ ( · ) L 2 ( t , T ; R m × n ) Q [ t , T ] and v ( · ) U [ t , T ] . For given initial pair ( t , x ) [ 0 , T ) × L F t 2 ( Ω ; R n ) , let us consider the following equation (some time variables are usually omitted):
d X = ( A + B Θ ) X + B v + b d s + ( C + D Θ ) X + D v + σ d W + E ( F ( e ) + G ( e ) Θ ) X + G ( e ) v + f ( e ) N ˜ ( d e d s ) , X ( t ) = x ,
which admits a unique solution X ( · ) X ( · ; t , x , Θ ( · ) , v ( · ) ) , depending on the Θ ( · ) and v ( · ) ; ( Θ ( · ) , v ( · ) ) Q [ t , T ] × U [ t , T ] is called a closed-loop strategy and the above Equation (5) is called a closed-loop system of the original state Equation (1) under ( Θ ( · ) , v ( · ) ) . We point out that ( Θ ( · ) , v ( · ) ) is independent of the initial state x L F t 2 ( Ω ; R n ) . With the above solution X ( · ) , we define
J ( t , x ; Θ ( · ) X ( · ) + v ( · ) ) = E { t T [ Q S S R X Θ X + v , X Θ X + v + 2 q ρ , X Θ X + v ] d s + H X ( T ) , X ( T ) + 2 g , X ( T ) } = E { t T [ ( Q + Θ S + S Θ + Θ R Θ ) X , X + 2 ( S + R Θ ) X , v + R v , v + 2 q + Θ ρ , X + 2 ρ , v ] d s + H X ( T ) , X ( T ) + 2 g , X ( T ) } .
We now introduce the following definition.
Definition 2.
If a closed-loop strategy pair ( Θ ¯ ( · ) , v ¯ ( · ) ) Q [ t , T ] × U [ t , T ] satisfies the following inequality
J ( t , x ; Θ ¯ ( · ) X ¯ ( · ) + v ¯ ( · ) ) J ( t , x ; Θ ( · ) X ( · ) + v ( · ) ) , x L F t 2 ( Ω ; R n ) , Θ ( · ) Q [ t , T ] , v ( · ) U [ t , T ] ,
where X ¯ ( · ) X ( · ; t , x , Θ ¯ ( · ) , v ¯ ( · ) ) on the left and X ( · ) X ( · ; t , x , Θ ( · ) , v ( · ) ) on the right, then ( Θ ¯ ( · ) , v ¯ ( · ) ) is called a closed-loop optimal strategy of problem (SLQP) on [ t , T ] , and we say that problem (SLQP) is closed-loop solvable on [ t , T ] .
We emphasize that the pair ( Θ ¯ ( · ) , v ¯ ( · ) ) is required to be independent of the initial state x L F t 2 ( Ω ; R n ) . We have the following equivalence theorem.
Theorem 1.
Let (H1) and (H2) hold and let ( Θ ¯ ( · ) , v ¯ ( · ) ) Q [ t , T ] × U [ t , T ] . Then the following statements are equivalent:
 (i)
( Θ ¯ ( · ) , v ¯ ( · ) ) is a closed-loop optimal strategy of problem (SLQP) on [ t , T ] .
 (ii)
For any given x L F t 2 ( Ω ; R n ) and v ( · ) U [ t , T ] ,
J ( t , x ; Θ ¯ ( · ) X ¯ ( · ) + v ¯ ( · ) ) J ( t , x ; Θ ¯ ( · ) X ( · ) + v ( · ) ) ,
where X ( · ) X ( · ; t , x , Θ ¯ ( · ) , v ( · ) ) on the right.
 (iii)
For any given x L F t 2 ( Ω ; R n ) and u ( · ) U [ t , T ] ,
J ( t , x ; Θ ¯ ( · ) X ¯ ( · ) + v ¯ ( · ) ) J ( t , x ; u ( · ) ) .
Proof. 
(i) ⇒ (ii). From the definition of closed-loop optimal strategy, it can be proved. (ii) ⇒ (iii). For given x L F t 2 ( Ω ; R n ) and u ( · ) U [ t , T ] , X ( · ) is the adapted solution to the following SDEP:
d X ( s ) = A ( s ) X ( s ) + B ( s ) u ( s ) + b ( s ) d s + C ( s ) X ( s ) + D ( s ) u ( s ) + σ ( s ) d W ( s ) + E F ( s , e ) X ( s ) + G ( s , e ) u ( s ) + f ( s , e ) N ˜ ( d e d s ) , s [ t , T ] , X ( t ) = x .
Taking v ( · ) = u ( · ) Θ ¯ ( · ) X ( · ) , it is easy to see that v ( · ) U [ t , T ] . Thus
u ( · ) = Θ ¯ ( · ) X ( · ) + v ( · ) .
From the existence and uniqueness of the adapt solution to SDEP, we obtain that
X ( · ) = X ( · ; t , x , u ( · ) ) = X ( · ; t , x , Θ ¯ ( · ) , v ( · ) ) X ˜ ( · ) .
Therefore, by (ii), one has
J ( t , x ; Θ ¯ ( · ) X ¯ ( · ) + v ¯ ( · ) ) J ( t , x ; Θ ¯ ( · ) X ˜ ( · ) + v ( · ) ) = J ( t , x ; u ( · ) ) ,
proving (iii). (iii) ⇒ (i). For a given x L F t 2 ( Ω ; R n ) , u ( · ) U [ t , T ] and ( Θ ( · ) , v ( · ) ) Q [ t , T ] × U [ t , T ] , let X ( · ) X ( · ; t , x , Θ ( · ) , v ( · ) ) . Taking
u ( · ) = Θ ( · ) X ( · ) + v ( · ) ,
from the existence and uniqueness of the adapt solution of SDEP, we know that
X ( · ; t , x , Θ ( · ) , v ( · ) ) = X ( · ; t , x , u ( · ) ) .
Therefore, by (iii), we have
J ( t , x ; Θ ¯ ( · ) X ¯ ( · ) + v ¯ ( · ) ) J ( t , x ; u ( · ) ) = J ( t , x ; Θ ( · ) X ( · ) + v ( · ) ) .
For the closed-loop optimal strategy ( Θ ¯ ( · ) , v ¯ ( · ) ) of problem (SLQP) on [ t , T ] and the corresponding closed-loop optimal state X ¯ ( · ) = X ¯ ( · ; t , x , Θ ¯ ( · ) , v ¯ ( · ) ) , we can define the outcome
u ¯ ( · ) = Θ ¯ ( · ) X ¯ ( · ) + v ¯ ( · ) U [ t , T ] .
From the third part of Theorem 1, we see that for any given initial state x L F t 2 ( Ω ; R n ) and u ( · ) U [ t , T ] , u ¯ ( · ) is an open-loop optimal control of problem (SLQP) for x. Therefore, if problem (SLQP) is closed-loop solvable on [ t , T ] , it must also be open-loop solvable, and the outcome of the closed-loop optimal strategy is the open-loop optimal control for any x L F t 2 ( Ω ; R n ) .
The following result is concerned with open-loop solvability of problem (SLQP) for given initial state.
Proposition 1.
Let (H1) and (H2) hold. For given initial pair ( t , x ) [ 0 , T ] × L F t 2 ( Ω ; R n ) , a control u ¯ ( · ) is an open-loop optimal control of problem (SLQP) if and only if the following hold:
 (i)
The stationarity condition holds:
B Y ¯ + D Z ¯ + E G ( e ) K ¯ ( e ) π ( d e ) + S X ¯ + R u ¯ + ρ = 0 , a . e . , P a . s . ,
where ( X ¯ ( · ) , Y ¯ ( · ) , Z ¯ ( · ) , K ¯ ( · , · ) ) L F 2 ( t , T ; R n ) × L F 2 ( t , T ; R n ) × L F , p 2 ( t , T ; R n ) × F F , p 2 ( t , T ; R n ) is the adapted solution to the following FBSDEP:
d X ¯ = A X ¯ + B u ¯ + b d s + C X ¯ + D u ¯ + σ d W + E F ( e ) X ¯ + G ( e ) u ¯ + f ( e ) N ˜ ( d e d s ) , d Y ¯ = A Y ¯ + C Z ¯ + E F ( e ) K ¯ ( e ) π ( d e ) + Q X ¯ + S u ¯ + q d s + Z ¯ d W + E K ¯ ( e ) N ˜ ( d e d s ) , X ¯ ( t ) = x , Y ¯ ( T ) = H X ¯ ( T ) + g .
 (ii)
The convexity condition holds: For any u ( · ) U [ t , T ] ,
E H x 0 ( T ) , x 0 ( T ) + t T Q x 0 , x 0 + 2 S x 0 , u + R u , u d s 0 ,
where x 0 ( · ) L F 2 ( t , T ; R n ) is the adapted solution to the following SDEP:
d x 0 ( s ) = A ( s ) x 0 ( s ) + B ( s ) u ( s ) d s + C ( s ) x 0 ( s ) + D ( s ) u ( s ) d W ( s ) + E F ( s , e ) x 0 ( s ) + G ( s , e ) u ( s ) N ˜ ( d e d s ) , s [ t , T ] , x 0 ( t ) = 0 .
Proof. 
For any u ( · ) U [ t , T ] and ϵ R , let u ϵ ( · ) = u ¯ ( · ) + ϵ u ( · ) , thus X ϵ ( · ) X ( · ; t , x , u ¯ ( · ) + ϵ u ( · ) ) is the corresponding state that satisfies
d X ϵ = A X ϵ + B ( u ¯ + ϵ u ) + b d s + C X ϵ + D ( u ¯ + ϵ u ) + σ d W + E F ( e ) X ϵ + G ( e ) ( u ¯ + ϵ u ) + f ( e ) N ˜ ( d e d s ) , X ϵ ( t ) = x .
Then X ϵ ( · ) X ( · ) ϵ satisfies the following SDEP:
d X ϵ X ϵ = A X ϵ X ϵ + B u d s + C X ϵ X ϵ + D u d W + E F ( e ) X ϵ X ϵ + G ( e ) u N ˜ ( d e d s ) , X ϵ X ϵ ( t ) = 0 .
From the existence and uniqueness of the solution to SDEP, we know that x 0 X ϵ X ϵ . Then
J ( t , x ; u ¯ ( · ) + ϵ u ( · ) ) J ( t , x ; u ¯ ( · ) ) = 2 ϵ E { H X ¯ ( T ) + g , x 0 ( T ) + t T [ Q x 0 , X ¯ + S X ¯ , u + S x 0 , u ¯ + R u ¯ , u + q , x 0 + ρ , u ] d s } + ϵ 2 E { H x 0 ( T ) , x 0 ( T ) + t T Q x 0 , x 0 + 2 S x 0 , u + R u , u d s .
Applying Itô’s formula to Y ¯ ( · ) , x 0 ( · ) , we get
J ( t , x ; u ¯ ( · ) + ϵ u ( · ) ) J ( t , x ; u ¯ ( · ) ) = 2 ϵ E t T B Y ¯ + D Z ¯ + E G ( e ) K ¯ ( e ) π ( d e ) + S X ¯ + R u ¯ + ρ , u d s + ϵ 2 E { H x 0 ( T ) , x 0 ( T ) + t T Q x 0 , x 0 + 2 S x 0 , u + R u , u d s .
Therefore, ( X ¯ ( · ) , u ¯ ( · ) ) is an open-loop optimal pair of problem (SLQP) if and only if (9) and (11) hold. □
On the other hand, from the second part of Theorem 1, we can see that ( Θ ¯ ( · ) , v ¯ ( · ) ) being a closed-loop optimal strategy of problem (SLQP) is equivalent to v ¯ ( · ) being an open-loop optimal control of the SLQP optimal control problems (5) and (6) with Θ ( · ) = Θ ¯ ( · ) , which we denote by problem (SLQP) Θ ¯ . In particular, when b ( · ) , σ ( · ) , f ( · , · ) , q ( · ) , ρ ( · ) and g are all zero, we refer to it as problem (SLQP) Θ ¯ 0 . Similar to Proposition 1, we can give the following result.
Proposition 2.
Let (H1) and (H2) hold. For any given x L F t 2 ( Ω ; R n ) , v ¯ ( · ) is an open-loop optimal control of problem (SLQP) Θ ¯ if and only if the following stationarity condition holds:
B Y ¯ + D Z ¯ + E G ( e ) K ¯ ( e ) π ( d e ) + ( S + R Θ ¯ ) X ¯ + R v ¯ + ρ = 0 , a . e . , P a . s . ,
where ( X ¯ ( · ) , Y ¯ ( · ) , Z ¯ ( · ) , K ¯ ( · , · ) ) L F 2 ( t , T ; R n ) × L F 2 ( t , T ; R n ) × L F , p 2 ( t , T ; R n ) × F F , p 2 ( t , T ; R n ) is the adapted solution to the following FBSDEP:
d X ¯ = ( A + B Θ ¯ ) X ¯ + B v ¯ + b d s + ( C + D Θ ¯ ) X ¯ + D v ¯ + σ d W + E ( F ( e ) + G ( e ) Θ ¯ ) X ¯ + G ( e ) v ¯ + f ( e ) N ˜ ( d e d s ) , d Y ¯ = [ ( A + B Θ ¯ ) Y ¯ + ( C + D Θ ¯ ) Z ¯ + E ( F ( e ) + G ( e ) Θ ¯ ) K ¯ ( e ) π ( d e ) + ( Q + S Θ ¯ + Θ ¯ S + Θ ¯ R Θ ¯ ) X ¯ + ( S + R Θ ¯ ) v ¯ + q + Θ ¯ ρ ] d s + Z ¯ d W + E K ¯ ( e ) N ˜ ( d e d s ) , X ¯ ( t ) = x , Y ¯ ( T ) = H X ¯ ( T ) + g ,
and the following convexity condition holds: For any v ( · ) U [ t , T ] ,
E { t T [ ( Q + Θ ¯ S + S Θ ¯ + Θ ¯ R Θ ¯ ) X , X + 2 ( S + R Θ ¯ ) X , v + R v , v ] d s + H X ( T ) , X ( T ) } 0 ,
where X ( · ) L F 2 ( t , T ; R n ) is the adapted solution to the following SDEP:
d X = ( A + B Θ ¯ ) X + B v d s + ( C + D Θ ¯ ) X + D v d W + E ( F ( e ) + G ( e ) Θ ¯ ) X + G ( e ) v N ˜ ( d e d s ) , X ( t ) = 0 .

3. Main Results

In this section, we will study the necessary and sufficient conditions for problem (SLQP) to be closed-loop solvable.
Making use of (13), we may rewrite the BSDEP in (14) and obtain
d X ¯ = ( A + B Θ ¯ ) X ¯ + B v ¯ + b d s + ( C + D Θ ¯ ) X ¯ + D v ¯ + σ d W + E ( F ( e ) + G ( e ) Θ ¯ ) X ¯ + G ( e ) v ¯ + f ( e ) N ˜ ( d e d s ) , d Y ¯ = A Y ¯ + C Z ¯ + E F ( e ) K ¯ ( e ) π ( d e ) + ( Q + S Θ ¯ ) X ¯ + S v ¯ + q d s + Z ¯ d W + E K ¯ ( e ) N ˜ ( d e d s ) , s [ t , T ] , X ¯ ( t ) = x , Y ¯ ( T ) = H X ¯ ( T ) + g .
We first have the following result.
Theorem 2.
Let (H1) and (H2) hold. If ( Θ ¯ ( · ) , v ¯ ( · ) ) Q [ t , T ] × U [ t , T ] is an optimal closed-loop strategy of problem (SLQP) on [ t , T ] , then ( Θ ¯ ( · ) , 0 ) is an optimal closed-loop strategy of problem (SLQP) 0 on [ t , T ] .
Proof. 
By Proposition 2, we can see that ( Θ ¯ ( · ) , v ¯ ( · ) ) is an optimal closed-loop strategy of problem (SLQP) on [ t , T ] if and only if for any x L F t 2 ( Ω ; R n ) , the stationarity condition holds:
B Y ¯ + D Z ¯ + E G ( e ) K ¯ ( e ) π ( d e ) + ( S + R Θ ¯ ) X ¯ + R v ¯ + ρ = 0 , a . e . , P a . s . ,
where ( X ¯ ( · ) , Y ¯ ( · ) , Z ¯ ( · ) , K ¯ ( · , · ) ) is the adapted solution to the following FBSDEP:
d X ¯ = ( A + B Θ ¯ ) X ¯ + B v ¯ + b d s + ( C + D Θ ¯ ) X ¯ + D v ¯ + σ d W + E ( F ( e ) + G ( e ) Θ ¯ ) X ¯ + G ( e ) v ¯ + f ( e ) N ˜ ( d e d s ) , d Y ¯ = [ ( A + B Θ ¯ ) Y ¯ + ( C + D Θ ¯ ) Z ¯ + E ( F ( e ) + G ( e ) Θ ¯ ) K ¯ ( e ) π ( d e ) + ( Q + S Θ ¯ + Θ ¯ S + Θ ¯ R Θ ¯ ) X ¯ + ( S + R Θ ¯ ) v ¯ + q + Θ ¯ ρ ] d s + Z ¯ d W + E K ¯ ( e ) N ˜ ( d e d s ) , X ¯ ( t ) = x , Y ¯ ( T ) = H X ¯ ( T ) + g ,
and the following convexity condition holds: For any v ( · ) U [ t , T ] ,
E { t T [ ( Q + Θ ¯ S + S Θ ¯ + Θ ¯ R Θ ¯ ) X 0 , X 0 + 2 ( S + R Θ ¯ ) X 0 , v + R v , v ] d s + H X 0 ( T ) , X 0 ( T ) } 0 ,
where X 0 ( · ) is the adapted solution to the following SDEP:
d X 0 = ( A + B Θ ¯ ) X 0 + B v d s + ( C + D Θ ¯ ) X 0 + D v d W + E ( F ( e ) + G ( e ) Θ ¯ ) X 0 + G ( e ) v N ˜ ( d e d s ) , s [ t , T ] , X 0 ( t ) = 0 .
Since the FBSDEP (17) admits a solution for each x L F t 2 ( Ω ; R n ) and ( Θ ¯ ( · ) , v ¯ ( · ) ) is independent of x, by subtracting solutions corresponding x and 0, the latter from the former, we see that, for any x L F t 2 ( Ω ; R n ) , the following FBSDEP:
d X = ( A + B Θ ¯ ) X d s + ( C + D Θ ¯ ) X d W + E ( F ( e ) + G ( e ) Θ ¯ ) X N ˜ ( d e d s ) , d Y = [ ( A + B Θ ¯ ) Y + ( C + D Θ ¯ ) Z + E ( F ( e ) + G ( e ) Θ ¯ ) K ( e ) π ( d e ) + ( Q + S Θ ¯ + Θ ¯ S + Θ ¯ R Θ ¯ ) X ] d s + Z d W + E K ( e ) N ˜ ( d e d s ) , X ( t ) = x , Y ( T ) = H X ( T ) ,
and in this time ( X ( · ) , Y ( · ) , Z ( · ) , K ( · , · ) ) satisfies
B Y + D Z + E G ( e ) K ( e ) π ( d e ) + ( S + R Θ ¯ ) X = 0 , a . e . , P a . s . .
It follows, again from Theorem 1 and Proposition 2, that ( Θ ¯ ( · ) , 0 ) is an optimal closed-loop strategy of problem (SLQP) 0 on [ t , T ] . □
To summarize the relationship between problem (SLQP), problem (SLQP) 0 , problem (SLQP) Θ ¯ , and problem (SLQP) Θ ¯ 0 , we plot the following diagram in Figure 1:
It is clear that when we want to study the necessary conditions for the closed-loop solvability of problem (SLQP), we can transform the original problem into the open-loop solvability of problem (SLQP) Θ ¯ 0 where the open-loop optimal control is v ¯ ( · ) 0 . Thus from Proposition 1, we can know that the optimal system of problem (SLQP) Θ ¯ 0 is the following FBSDEP:
d X ¯ = ( A + B Θ ¯ ) X ¯ d s + ( C + D Θ ¯ ) X ¯ d W + E ( F ( e ) + G ( e ) Θ ¯ ) X ¯ N ˜ ( d e d s ) , d Y ¯ = A Y ¯ + C Z ¯ + E F ( e ) K ¯ ( e ) π ( d e ) + ( Q + S Θ ¯ ) X ¯ d s + Z ¯ d W + E K ¯ ( e ) N ˜ ( d e d s ) , X ¯ ( t ) = x , Y ¯ ( T ) = H X ¯ ( T ) ,
and in this time ( X ¯ ( · ) , Y ¯ ( · ) , Z ¯ ( · ) , K ¯ ( · , · ) ) satisfies
B Y ¯ + D Z ¯ + E G ( e ) K ¯ ( e ) π ( d e ) + ( S + R Θ ¯ ) X ¯ = 0 , a . e . , P a . s . .
In the light of Y ¯ ( T ) = H X ¯ ( T ) , we assume that
Y ¯ ( · ) = P ( · ) X ¯ ( · ) ,
where P ( · ) : [ t , T ] S n × n is a matrix-valued differential function satisfying P ( T ) = H . Applying Itô–Wentzell’s formula to (21), we have
d Y ¯ = P ˙ X ¯ d s + P d X ¯ = P ˙ X ¯ + P ( A + B Θ ¯ ) X ¯ d s + P ( C + D Θ ¯ ) X ¯ d W + E P ( F ( e ) + G ( e ) Θ ¯ ) X ¯ N ˜ ( d e d s ) .
Comparing the diffusion coefficient of the above equation and of the second equation in (19), we can see
Z ¯ = P ( C + D Θ ¯ ) X ¯ , K ¯ ( e ) = P ( F ( e ) + G ( e ) Θ ¯ ) X ¯ .
Plugging (21) and (23) into (20), we obtain
B P + D P ( C + D Θ ¯ ) + E G ( e ) P ( F ( e ) + G ( e ) Θ ¯ ) π ( d e ) + ( S + R Θ ¯ ) X ¯ = 0 .
Now comparing the drift coefficient of (22) and of the second equation in (19), noting that (21), (23) and (24), we have
0 = [ P ˙ + P ( A + B Θ ¯ ) + A P + C P ( C + D Θ ¯ ) + E F ( e ) P ( F ( e ) + G ( e ) Θ ¯ ) π ( d e ) + Q + S Θ ¯ ] X ¯ = [ P ˙ + P ( A + B Θ ¯ ) + A P + C P ( C + D Θ ¯ ) + Q + S Θ ¯ + E F ( e ) P ( F ( e ) + G ( e ) Θ ¯ ) π ( d e ) + Θ ¯ ( B P + D P ( C + D Θ ¯ ) + S + R Θ ¯ + E G ( e ) P ( F ( e ) + G ( e ) Θ ¯ ) π ( d e ) ) ] X ¯ = [ P ˙ + P ( A + B Θ ¯ ) + ( A + B Θ ¯ ) P + ( C + D Θ ¯ ) P ( C + D Θ ¯ ) + Q + S Θ ¯ + Θ ¯ S + Θ ¯ R Θ ¯ + E ( F ( e ) + G ( e ) Θ ¯ ) P ( F ( e ) + G ( e ) Θ ¯ ) π ( d e ) ] X ¯ .
Thus, we let P ( · ) satisfy the following Lyapunov integral–differential equation:
0 = P ˙ + P ( A + B Θ ¯ ) + ( A + B Θ ¯ ) P + ( C + D Θ ¯ ) P ( C + D Θ ¯ ) + Q + S Θ ¯ + Θ ¯ S + Θ ¯ R Θ ¯ + E ( F ( e ) + G ( e ) Θ ¯ ) P ( F ( e ) + G ( e ) Θ ¯ ) π ( d e ) , P ( T ) = H .
Proposition 3.
Let P ( · ) be the solution to (25). Then for any s [ t , T ] , we have
R + D P D + E G ( e ) P G ( e ) π ( d e ) 0 , 0 = R + D P D + E G ( e ) P G ( e ) π ( d e ) Θ ¯ + B P + D P C + S + E G ( e ) P F ( e ) π ( d e ) .
Proof. 
Let us consider problem (SLQP) Θ ¯ 0 . For any v ( · ) U [ t , T ] , the state equation and the cost functional are:
d X = ( A + B Θ ¯ ) X + B v d s + ( C + D Θ ¯ ) X + D v d W + E ( F ( e ) + G ( e ) Θ ¯ ) X + G ( e ) v N ˜ ( d e d s ) , X ( t ) = x ,
and
J ˜ ( t , x ; v ( · ) ) = E { t T [ Q + Θ ¯ S + S Θ ¯ + Θ ¯ R Θ ¯ X , X + 2 S + R Θ ¯ X , v + R v , v ] d s + H X ( T ) , X ( T ) } .
Applying Itô–Wentzell’s formula to P X ( · ) , X ( · ) , we get
d P X , X = P ( A + B Θ ¯ ) X ( A + B Θ ¯ ) P X ( C + D Θ ¯ ) P ( C + D Θ ¯ ) X E ( F ( e ) + G ( e ) Θ ¯ ) P ( F ( e ) + G ( e ) Θ ¯ ) X π ( d e ) ( Q + Θ ¯ S + S Θ ¯ + Θ ¯ R Θ ¯ ) X , X d s + 2 P ( A + B Θ ¯ ) X + B v , X d s + P ( C + D Θ ¯ ) X + D v , ( C + D Θ ¯ ) X + D v d s + E P ( F ( e ) + G ( e ) Θ ¯ ) X + G ( e ) v , ( F ( e ) + G ( e ) Θ ¯ ) X + G ( e ) v π ( d e ) d s + [ ] d W + E [ ] N ˜ ( d e d s ) = { Q + Θ ¯ S + S Θ ¯ + Θ ¯ R Θ ¯ X , X + 2 B P + D P ( C + D Θ ¯ ) X , v + E 2 G ( e ) P F ( e ) + G ( e ) Θ ¯ X + G ( e ) P G ( e ) v , v π ( d e ) + D P D v , v } d s + [ ] d W + E [ ] N ˜ ( d e d s ) .
Thus, we have
E H X ( T ) , X ( T ) E P ( t ) x , x = E t T { Q + Θ ¯ S + S Θ ¯ + Θ ¯ R Θ ¯ X , X + D P D + E G ( e ) P G ( e ) π ( d e ) v , v + 2 B P + D P ( C + D Θ ¯ ) + E G ( e ) P F ( e ) + G ( e ) Θ ¯ π ( d e ) X , v } d s .
Putting the above equation into the cost functional, we have
J ˜ ( t , x ; v ( · ) ) = E P ( t ) x , x + E t T { R + D P D + E G ( e ) P G ( e ) π ( d e ) v , v + 2 [ B P + D P C + S + E G ( e ) P F ( e ) π ( d e ) + R + D P D + E G ( e ) P G ( e ) π ( d e ) Θ ¯ ] X , v } d s .
From the previous analysis, we know that v ¯ ( · ) = 0 is the open-loop optimal control for problem (SLQP) Θ ¯ 0 , i.e.,
J ˜ ( t , x ; v ( · ) ) J ˜ ( t , x ; 0 ) = E P ( t ) x , x ,
then
0 J ˜ ( t , x ; v ( · ) ) E P ( t ) x , x = E t T { R + D P D + E G ( e ) P G ( e ) π ( d e ) v , v + 2 [ B P + D P C + S + E G ( e ) P F ( e ) π ( d e ) + R + D P D + E G ( e ) P G ( e ) π ( d e ) Θ ¯ ] X , v } d s .
For simplicity, we give the following notations:
R : = R + D P D + E G ( e ) P G ( e ) π ( d e ) , L : = B P + D P C + S + E G ( e ) P F ( e ) π ( d e ) ,
thus,
E t T R v , v + 2 L + R Θ ¯ X , v d s 0 , v ( · ) U [ t , T ] .
Choose the initial pair ( t , x ) = ( 0 , 0 ) and v ( · ) = v 0 I [ r , r + h ] ( · ) , 0 r < r + h T , v 0 R m . In this time, v ( · ) is a deterministic function, thus
E 0 T R v , v + 2 L + R Θ ¯ X , v d s = 0 T R v , v + 2 L + R Θ ¯ E X , v d s = r r + h R v 0 , v 0 + 2 L + R Θ ¯ E X , v 0 d s ,
where E X ( · ) is solution to the following ordinary differential equation (ODE):
d E X ( s ) = A ( s ) + B ( s ) Θ ¯ ( s ) E X ( s ) + B ( s ) v 0 I [ r , r + h ] ( s ) d s , s [ 0 , T ] , E X ( 0 ) = 0 .
Then
E X ( t ) = 0 t B ( s ) v 0 I [ r , r + h ] ( s ) e 0 s A ( r ) + B ( r ) Θ ¯ ( r ) d r d s · e 0 t A ( r ) + B ( r ) Θ ¯ ( r ) d r , t [ 0 , T ] .
It is easy to see that
E X ( t ) 0 , a s h 0 , t [ 0 , T ] ,
thus
lim h 0 1 h r r + h R ( s ) v 0 , v 0 + 2 L ( s ) + R ( s ) Θ ¯ ( s ) E X ( s ) , v 0 d s 0 .
We know
R ( r ) v 0 , v 0 0 , r [ 0 , T ] , v 0 R m R ( r ) 0 , r [ 0 , T ] .
Thus, the first inequality in (26) is obtained, and now let us prove the second equation. We take the initial pair ( t , x ) [ 0 , T ) × R n and v ( · ) = v 0 n I [ r , r + h ] ( · ) , v 0 R m , t r < r + h T . In this time,
E t T R v , v + 2 L + R Θ ¯ X , v d s = E r r + h R v 0 n , v 0 n + 2 L + R Θ ¯ X , v 0 n d s = r r + h 1 n 2 R v 0 , v 0 + 2 n L + R Θ ¯ E X , v 0 d s ,
where E X ( · ) is solution to the following ODE:
d E X ( s ) = A ( s ) + B ( s ) Θ ¯ ( s ) E X ( s ) + B ( s ) v 0 n I [ r , r + h ] ( s ) d s , s [ t , T ] , E X ( t ) = E x .
As we know, this is an inhomogeneous linear ODE, and we get for s [ t , T ] ,
E X ( s ) = E x + t s B ( y ) v 0 n I [ r , r + h ] ( y ) e t y A ( z ) + B ( z ) Θ ¯ ( z ) d z d y · e t s A ( z ) + B ( z ) Θ ¯ ( z ) d z .
It is easy to see
E X ( s ) E x · e t s A ( z ) + B ( z ) Θ ¯ ( z ) d z E x · Ψ ( s ) , a s n .
Thus
lim n r r + h n · 1 n 2 R v 0 , v 0 + 2 n L ( s ) + R ( s ) Θ ¯ ( s ) E X ( s ) , v 0 d s = lim n r r + h · 1 n R v 0 , v 0 + 2 L ( s ) + R ( s ) Θ ¯ ( s ) E X ( s ) , v 0 d s = 2 r r + h L ( s ) + R ( s ) Θ ¯ ( s ) Ψ ( s ) E x , v 0 d s 0 .
Since v 0 R m is arbitrary,
r r + h L ( s ) + R ( s ) Θ ¯ ( s ) Ψ ( s ) E x , v 0 d s = 0 .
Dividing both sides by h and let h 0 , we have
L ( r ) + R ( r ) Θ ¯ ( r ) Ψ ( r ) E x , v 0 = 0 , r [ t , T ] .
Since v 0 R m is arbitrary,
L ( r ) + R ( r ) Θ ¯ ( r ) = 0 , r [ t , T ] .
This is the second equation in (26). □
The following theorem is the main result in this paper, which characterizes the closed-loop solvability of problem (SLQP).
Theorem 3.
Let (H1) and (H2) hold. Problem (SLQP) admits an optimal closed-loop strategy ( Θ ¯ ( · ) , v ¯ ( · ) ) Q [ t , T ] × U [ t , T ] if and only if the following RIDE:
0 = P ˙ + P A + A P + C P C + E F ( e ) P F ( e ) π ( d e ) + Q B P + D P C + E G ( e ) P F ( e ) π ( d e ) + S ( R + D P D + E G ( e ) P G ( e ) π ( d e ) ) B P + D P C + E G ( e ) P F ( e ) π ( d e ) + S , R B P + D P C + E G ( e ) P F ( e ) π ( d e ) + S R R + D P D + E G ( e ) P G ( e ) π ( d e ) , R + D P D + E G ( e ) P G ( e ) π ( d e ) 0 , P ( T ) = H ,
admits a solution P ( · ) C ( [ t , T ] ; S n × n ) such that
R + D P D + E G ( e ) P G ( e ) π ( d e ) B P + D P C + E G ( e ) P F ( e ) π ( d e ) + S L 2 ( t , T ; R m × n ) ,
and the following BSDEP:
d η = { [ A B R + D P D + E G ( e ) P G ( e ) π ( d e ) × B P + D P C + E G ( e ) P F ( e ) π ( d e ) + S ] η + [ C D R + D P D + E G ( e ) P G ( e ) π ( d e ) × B P + D P C + E G ( e ) P F ( e ) π ( d e ) + S ] ( ζ + P σ ) + E [ F ( e ) G ( e ) R + D P D + E G ( e ) P G ( e ) π ( d e ) × B P + D P C + E G ( e ) P F ( e ) π ( d e ) + S ] ψ ( e ) + P f ( e ) π ( d e ) + P b + q B P + D P C + E G ( e ) P F ( e ) π ( d e ) + S × R + D P D + E G ( e ) P G ( e ) π ( d e ) ρ } d s + ζ d W + E ψ ( e ) N ˜ ( d e d s ) , B η + D ( ζ + P σ ) + E G ( e ) ψ ( e ) + P f ( e ) π ( d e ) + ρ R R + D P D + E G ( e ) P G ( e ) π ( d e ) , a . e . , P a . s . , η ( T ) = g ,
admits a unique solution ( η ( · ) , ζ ( · ) , ψ ( · ) ) , which satisfies
R + D P D + E G ( e ) P G ( e ) π ( d e ) × B η + D ( ζ + P σ ) + E G ( e ) ψ ( e ) + P f ( e ) π ( d e ) + ρ L F , p 2 ( t , T ; R m ) .
In this case, the closed-loop optimal strategy ( Θ ¯ ( · ) , v ¯ ( · ) ) of problem (SLQP) admits the following representation:
Θ ¯ = R + D P D + E G ( e ) P G ( e ) π ( d e ) × B P + D P C + E G ( e ) P F ( e ) π ( d e ) + S + [ I R + D P D + E G ( e ) P G ( e ) π ( d e ) × R + D P D + E G ( e ) P G ( e ) π ( d e ) ] θ , v ¯ = R + D P D + E G ( e ) P G ( e ) π ( d e ) × B η + D ( ζ + P σ ) + E G ( e ) ψ ( e ) + P f ( e ) π ( d e ) + ρ + [ I R + D P D + E G ( e ) P G ( e ) π ( d e ) × R + D P D + E G ( e ) P G ( e ) π ( d e ) ] v ,
for some θ ( · ) Q [ t , T ] , v ( · ) U [ t , T ] . Further, the value function V ( · , · ) is given by
V ( t , x ) inf u ( · ) U [ t , T ] J ( t , x ; u ( · ) ) = E { P ( t ) x , x + 2 η ( t ) , x + t T [ 2 η , b + 2 ζ , σ + 2 E ψ ( e ) , f ( e ) π ( d e ) + P σ , σ + E P f ( e ) , f ( e ) π ( d e ) | R + D P D + E G ( e ) P G ( e ) π ( d e ) 1 2 × B η + D ( ζ + P σ ) + E G ( e ) ψ ( e ) + P f ( e ) π ( d e ) + ρ | 2 ] d s } .
Proof. 
We first prove the necessity. Let ( Θ ¯ ( · ) , v ¯ ( · ) ) Q [ t , T ] × U [ t , T ] be a closed-loop optimal strategy of problem (SLQP) over [ t , T ] . The second equation of (26) implies
R B P + D P C + E G ( e ) P F ( e ) π ( d e ) + S R R + D P D + E G ( e ) P G ( e ) π ( d e ) , a . e . .
Denoting R : = R + D P D + E G ( e ) P G ( e ) π ( d e ) , since
R B P + D P C + E G ( e ) P F ( e ) π ( d e ) + S = R R Θ ¯ ,
and R R is an orthogonal projection, we see that (30) holds and
Θ ¯ = R B P + D P C + E G ( e ) P F ( e ) π ( d e ) + S + I R R θ ,
for some θ ( · ) Q [ t , T ] . Consequently,
P B + C P D + E F ( e ) P G ( e ) π ( d e ) + S Θ ¯ = Θ ¯ R Θ ¯ = Θ ¯ R R B P + D P C + E G ( e ) P F ( e ) π ( d e ) + S = B P + D P C + E G ( e ) P F ( e ) π ( d e ) + S R × B P + D P C + E G ( e ) P F ( e ) π ( d e ) + S .
Plugging the above into the Lyapunov Equation (25), we obtain the RIDE in (29).
To determine v ¯ ( · ) , we define
η = Y ¯ P X ¯ , ζ = Z ¯ P ( C + D Θ ¯ ) X ¯ P D v ¯ P σ , ψ ( e ) = K ¯ ( e ) P F ( e ) + G ( e ) Θ ¯ X ¯ P G ( e ) v ¯ P f ( e ) ,
where ( X ¯ ( · ) , Y ¯ ( · ) , Z ¯ ( · ) , K ¯ ( · , · ) ) is the adapted solution to the FBSDEP (17). Then
d η = d Y ¯ P ˙ X ¯ d s P d X ¯ = { A Y ¯ C Z ¯ E F ( e ) K ¯ ( e ) π ( d e ) ( Q + S Θ ¯ ) X ¯ S v ¯ q + A P X ¯ + P A X ¯ + C P C X ¯ + P B + C P D + E F ( e ) P G ( e ) π ( d e ) + S Θ ¯ X ¯ + Q X ¯ + E F ( e ) P F ( e ) X ¯ π ( d e ) P ( A + B Θ ¯ ) X ¯ P B v ¯ P b } d s + Z ¯ P ( C + D Θ ¯ ) X ¯ P D v ¯ P σ d W + E K ¯ ( e ) P F ( e ) + G ( e ) Θ ¯ X ¯ P G ( e ) v ¯ P f ( e ) N ˜ ( d e d s ) = { A ( P X ¯ + η ) + C ζ + P ( C + D Θ ¯ ) X ¯ + P D v ¯ + P σ + E F ( e ) [ ψ ( e ) + P F ( e ) + G ( e ) Θ ¯ X ¯ + P G ( e ) v ¯ + P f ( e ) ] π ( d e ) + ( Q + S Θ ¯ ) X ¯ + S v ¯ + q A P X ¯ P A X ¯ C P C X ¯ P B + C P D + E F ( e ) P G ( e ) π ( d e ) + S Θ ¯ X ¯ Q X ¯ E F ( e ) P F ( e ) X ¯ π ( d e ) + P ( A + B Θ ¯ ) X ¯ + P B v ¯ + P b } d s + ζ d W + E ψ ( e ) N ˜ ( d e d s ) = { A η + C ζ + E F ( e ) ψ ( e ) π ( d e ) + ( P B + C P D + E F ( e ) P G ( e ) π ( d e ) + S ) v ¯ + P b + C P σ + E F ( e ) P f ( e ) π ( d e ) + q } d s + ζ d W + E ψ ( e ) N ˜ ( d e d s ) .
According to the stationarity condition (16), we have
0 = B Y ¯ + D Z ¯ + E G ( e ) K ¯ ( e ) π ( d e ) + ( S + R Θ ¯ ) X ¯ + R v ¯ + ρ = B ( P X ¯ + η ) + D ζ + P ( C + D Θ ¯ ) X ¯ + P D v ¯ + P σ + ( S + R Θ ¯ ) X ¯ + R v ¯ + ρ + E G ( e ) ψ ( e ) + P F ( e ) + G ( e ) Θ ¯ X ¯ + P G ( e ) v ¯ + P f ( e ) π ( d e ) = B η + D ζ + E G ( e ) ψ ( e ) π ( d e ) + D P σ + ρ + E G ( e ) P f ( e ) π ( d e ) + R + D P D + E G ( e ) P G ( e ) π ( d e ) v ¯ + [ B P + D P C + E G ( e ) P F ( e ) π ( d e ) + S + R + D P D + E G ( e ) P G ( e ) π ( d e ) Θ ¯ ] X ¯ = B η + D ζ + E G ( e ) ψ ( e ) π ( d e ) + D P σ + ρ + E G ( e ) P f ( e ) π ( d e ) + R + D P D + E G ( e ) P G ( e ) π ( d e ) v ¯ .
Hence,
B η + D ζ + E G ( e ) ψ ( e ) π ( d e ) + D P σ + ρ + E G ( e ) P f ( e ) π ( d e ) R R + D P D + E G ( e ) P G ( e ) π ( d e ) , a . e . , P a . s . .
Since
R B η + D ζ + E G ( e ) ψ ( e ) π ( d e ) + D P σ + ρ + E G ( e ) P f ( e ) π ( d e ) = R R v ¯
and R R is an orthogonal projection, we see that (32) holds and
v ¯ = R B η + D ( ζ + P σ ) + E G ( e ) ψ ( e ) + P f ( e ) π ( d e ) + ρ + ( I R R ) v ,
for some v ( · ) L F , p 2 ( t , T ; R m ) . Consequently,
P B + C P D + E F ( e ) P G ( e ) π ( d e ) + S v ¯ = P B + C P D + E F ( e ) P G ( e ) π ( d e ) + S × R B η + D ( ζ + P σ ) + E G ( e ) ψ ( e ) + P f ( e ) π ( d e ) + ρ + P B + C P D + E F ( e ) P G ( e ) π ( d e ) + S ( I R R ) v = P B + C P D + E F ( e ) P G ( e ) π ( d e ) + S R [ B η + D ( ζ + P σ ) + E G ( e ) ψ ( e ) + P f ( e ) π ( d e ) + ρ ] Θ ¯ R ( I R R ) v = P B + C P D + E F ( e ) P G ( e ) π ( d e ) + S × R B η + D ( ζ + P σ ) + E G ( e ) ψ ( e ) + P f ( e ) π ( d e ) + ρ .
Therefore, ( η ( · ) , ζ ( · ) , ψ ( · , · ) ) is the unique solution to the following BSDEP:
d η = { A η + C ζ + E F ( e ) ψ ( e ) π ( d e ) + P B + C P D + E F ( e ) P G ( e ) π ( d e ) + S v ¯ + P b + C P σ + E F ( e ) P f ( e ) π ( d e ) + q } d s + ζ d W + E ψ ( e ) N ˜ ( d e d s ) = { A η + C ζ + E F ( e ) ψ ( e ) π ( d e ) + P b + C P σ + E F ( e ) P f ( e ) π ( d e ) + q P B + C P D + E F ( e ) P G ( e ) π ( d e ) + S R [ B η + D ( ζ + P σ ) + E G ( e ) ψ ( e ) + P f ( e ) π ( d e ) + ρ ] } d s + ζ d W + E ψ ( e ) N ˜ ( d e d s ) , η ( T ) = g .
To prove the sufficiency, we take any u ( · ) U [ t , T ] , and let X ( · ) X ( · ; t , x , u ( · ) ) be the corresponding state process. Applying Itô–Wentzell’s formula to P X , X and η , X , we have
d P X , X = P ( A + B Θ ¯ ) X ( A + B Θ ¯ ) P X ( C + D Θ ¯ ) P ( C + D Θ ¯ ) X E ( F ( e ) + G ( e ) Θ ¯ ) P ( F ( e ) + G ( e ) Θ ¯ ) X π ( d e ) ( Q + Θ ¯ S + S Θ ¯ + Θ ¯ R Θ ¯ ) X , X d s + 2 P A X + P B u + P b , X d s + P ( C X + D u + σ ) , C X + D u + σ d s + E P F ( e ) X + G ( e ) u + f ( e ) , F ( e ) X + G ( e ) u + f ( e ) π ( d e ) d s + [ ] d W + E [ ] N ˜ ( d e d s ) = { [ ( P B + S ) Θ ¯ + Θ ¯ ( B P + S ) + ( C + D Θ ¯ ) P ( C + D Θ ¯ ) + E ( F ( e ) + G ( e ) Θ ¯ ) P ( F ( e ) + G ( e ) Θ ¯ ) π ( d e ) + Q + Θ ¯ R Θ ¯ ] X , X + 2 P ( B u + b ) , X + P ( C X + D u + σ ) , C X + D u + σ + E P F ( e ) X + G ( e ) u + f ( e ) , F ( e ) X + G ( e ) u + f ( e ) π ( d e ) } d s + [ ] d W + E [ ] N ˜ ( d e d s ) ,
and
d η , X = { A η C ζ E F ( e ) ψ ( e ) π ( d e ) P B + C P D + E F ( e ) P G ( e ) π ( d e ) + S v ¯ P b C P σ E F ( e ) P f ( e ) π ( d e ) q , X + η , A X + B u + b + ζ , C X + D u + σ + E ψ ( e ) , F ( e ) X + G ( e ) u + f ( e ) π ( d e ) } d s + [ ] d W + E [ ] N ˜ ( d e d s ) = { P B + C P D + E F ( e ) P G ( e ) π ( d e ) + S v ¯ + P b + C P σ + E F ( e ) P f ( e ) π ( d e ) + q , X + η , B u + b + ζ , D u + σ + E ψ ( e ) , G ( e ) u + f ( e ) π ( d e ) } d s + [ ] d W + E [ ] N ˜ ( d e d s ) .
Thus,
J ( t , x ; u ( · ) ) = E { H X ( T ) , X ( T ) + 2 g , X ( T ) + t T Q X , X + 2 S X , u + R u , u + 2 q , X + 2 ρ , u d s } = E { P ( t ) x , x + 2 η ( t ) , x + t T [ Q X , X + 2 S X , u + R u , u + 2 q , X + 2 ρ , u 2 B P X , Θ ¯ X C P C X , x 2 D P C X , Θ ¯ X D P D Θ ¯ X , Θ ¯ X Q X , X 2 S X , Θ ¯ X R Θ ¯ X , Θ ¯ X + 2 B P X , u + 2 P b , X + C P C X , X + D P C X , u + 2 P C X , σ + D P X C , u + D P D u , u + 2 P D u , σ + P σ , σ 2 P B + C P D + E F ( e ) P G ( e ) π ( d e ) + S v ¯ , X 2 C P σ + P b + E F ( e ) P f ( e ) π ( d e ) + q , X + 2 B η , u + 2 η , b + 2 D ζ , u + 2 ζ , σ E F ( e ) P F ( e ) X , X π ( d e ) 2 E G ( e ) P F ( e ) X , Θ ¯ X π ( d e ) E G ( e ) P G ( e ) Θ ¯ X , Θ ¯ X π ( d e ) + E F ( e ) P F ( e ) X , X π ( d e ) + 2 E G ( e ) P F ( e ) X , u π ( d e ) + 2 E F ( e ) P f ( e ) , X π ( d e ) + E G ( e ) P G ( e ) u , u π ( d e ) + 2 E G ( e ) P f ( e ) , u π ( d e ) + E P f ( e ) , f ( e ) π ( d e ) + 2 E G ( e ) ψ ( e ) , u π ( d e ) + 2 E ψ ( e ) , f ( e ) π ( d e ) ] d s }
= E { P ( t ) x , x + 2 η ( t ) , x + t T [ R u , u R Θ ¯ X , Θ ¯ X + 2 B P + D P C + E G ( e ) P F ( e ) π ( d e ) + S , u Θ ¯ X v ¯ + 2 B η + D ζ + E G ( e ) ψ ( e ) π ( d e ) + D P σ + ρ + E G ( e ) P f ( e ) π ( d e ) , u + P σ , σ + E P f ( e ) , f ( e ) π ( d e ) + 2 η , b + 2 ζ , σ + 2 E ψ ( e ) , f ( e ) π ( d e ) ] d s } = E { P ( t ) x , x + 2 η ( t ) , x + t T R u Θ ¯ X v ¯ , u Θ ¯ X v ¯ d s + t T [ P σ , σ + E P f ( e ) , f ( e ) π ( d e ) + 2 η , b + 2 ζ , σ + 2 E ψ ( e ) , f ( e ) π ( d e ) R B η + D ζ + E G ( e ) ψ ( e ) π ( d e ) + D P σ + ρ + E G ( e ) P f ( e ) π ( d e ) , B η + D ζ + E G ( e ) ψ ( e ) π ( d e ) + D P σ + ρ + E G ( e ) P f ( e ) π ( d e ) ] d s } = J ( t , x ; Θ ¯ ( · ) X ¯ ( · ) + v ¯ ( · ) ) + t T R u Θ ¯ X v ¯ , u Θ ¯ X v ¯ d s .
Hence,
J ( t , x ; Θ ¯ ( · ) X ¯ ( · ) + v ¯ ( · ) ) J ( t , x ; u ( · ) ) , u ( · ) U [ t , T ] ,
if and only if
R + D P D + E G ( e ) P G ( e ) π ( d e ) 0 , a . e . .

4. Example

In this section, we will give a simple example. Consider the following controlled linear SDE with Poisson jumps:
d X ( s ) = A ( s ) X ( s ) + B ( s ) u ( s ) d s + C ( s ) X ( s ) + D ( s ) u ( s ) d W ( s ) + E G ( s , e ) u ( s ) N ˜ ( d e d s ) , s [ t , T ] , X ( t ) = x ,
and the cost functional is defined to be
J ( t , x ; u ( · ) ) = E t T Q X , X + 2 S X , u + R u , u d s + H X ( T ) , X ( T ) .
Then, for any closed-loop strategy pair ( Θ ( · ) , v ( · ) ) Q [ t , T ] × U [ t , T ] , we get the closed-loop system:
d X = ( A + B Θ ) X + B v d s + ( C + D Θ ) X + D v d W + E G ( e ) Θ X + G ( e ) v N ˜ ( d e d s ) , s [ t , T ] , X ( t ) = x ,
and the closed-loop cost functional
J t , x ; Θ ( · ) X ( · ) + v ( · ) = E { H X ( T ) , X ( T ) + t T ( Q + Θ S + S Θ + Θ R Θ ) X , X + 2 ( S + R Θ ) X , v + R v , v d s } .
In this case, ( 0 , 0 , 0 ) is the unique adaptive solution to BSDEP (31). Then, by Theorem 3, we obtain the closed-loop optimal strategy ( Θ ¯ ( · ) , v ¯ ( · ) ) admits the following representation:
Θ ¯ = R + D P D + E G ( e ) P G ( e ) π ( d e ) B P + D P C + S + [ I R + D P D + E G ( e ) P G ( e ) π ( d e ) × R + D P D + E G ( e ) P G ( e ) π ( d e ) ] θ , v ¯ = 0 ,
for some θ ( · ) Q [ t , T ] , where P ( · ) C ( [ t , T ] ; S n × n ) is the solution to the following RIDE:
0 = P ˙ + P A + A P + C P C + Q B P + D P C + S R + D P D + E G ( e ) P G ( e ) π ( d e ) B P + D P C + S , R B P + D P C + S R R + D P D + E G ( e ) P G ( e ) π ( d e ) , R + D P D + E G ( e ) P G ( e ) π ( d e ) 0 , P ( T ) = H ,
such that
R + D P D + E G ( e ) P G ( e ) π ( d e ) B P + D P C + S L 2 ( t , T ; R m × n ) .
Further, the value function V ( · , · ) is given by
V ( t , x ) inf u ( · ) U [ t , T ] J ( t , x ; u ( · ) ) = E P ( t ) x , x .
In this time, the closed-loop optimal control of this problem is u ¯ ( · ) = Θ ¯ ( · ) X ¯ ( · ) , where X ¯ ( · ) admits the following repesentation:
X ¯ ( s ) = x exp { t s A + B Θ ¯ C + D Θ ¯ 2 2 E G ( e ) Θ ¯ π ( d e ) d r + t s C + D Θ ¯ d W } · t < r s 1 + E G ( e ) Θ ¯ N ( d e , r ) , s [ t , T ] .
We further make the following assumptions:
Hypothesis 3.
The coefficients of the drift and diffusion terms of (37) are time-invariant constants, and the coefficient of the jump diffusion term G ( · ) depends only on the variable e, that is, G ( · ) is independent of t.
Hypothesis 4.
The Poisson process N has jumps of unit size, i.e., E = { 1 } .
Hypothesis 5.
The weighting coefficients of the cost functional (38) satisfy the following:
R 0 , Q S R 1 S 0 , H 0 ,
Under (H4), the compensated Poisson process N ˜ ( [ 0 , t ] ) = ( N N ^ ) ( [ 0 , t ] ) t 0 is a martingale, where N ^ ( d e d t ) = π ( d e ) d t = π d t and π > 0 is the intensity of the Poisson process N. Then
E G ( s , e ) P ( s ) G ( s , e ) π ( d e ) d s π G ( s ) P ( s ) G ( s ) d s , s [ t , T ] .
Under (H4)–(H5), the equation of RIDE (42) can be reduced to the following:
0 = P ˙ + P A + A P + C P C + Q B P + D P C + S R + D P D + π G P G 1 B P + D P C + S , P ( T ) = H ,
where A, B, C, D, G, Q, R, S, and H are all constants.
When we assume that the coefficients are A = 2 , B = 3 , C = 1 , D = 3 , G = 3 , Q = 4 , R = 2 , S = 1 , H = 3 , and π = 1 , respectively, the solution P ( · ) of the above Riccati equation can be shown in the following Figure 2.

5. Concluding Remarks

In this paper, we have investigated the closed-loop solvability of problem (SLQP): the stochastic linear–quadratic optimal control problem with Poisson jumps. We transform the initial problem into a new and simple problem and obtain a Riccati integral–differential equation first. Then we get the characterization of its closed-loop solvability, i.e., the solvabilities of the RIDE and a BSDEP (Theorem 3). We also give a simple example to prove the effectiveness of the main result.
Motivated by [13,14,32,40,41], characterization of the closed-loop solvability for SLQP optimal control problem in infinite horizon, with random coefficients, and of the closed-loop saddle points, Nash equilibria for SLQ zero-sum, nonzero-sum differential games, respectively, are our future research topics. The closed-loop solvability for Stackelberg stochastic LQ differential game is recently studied by Li and Shi [42,43], and we are also interested in that for Stackelberg stochastic LQ differential game of mean-field type. We will consider them in the near future.

Author Contributions

Conceptualization, Z.L. and J.S.; methodology, Z.L. and J.S.; validation, Z.L. and J.S.; formal analysis, Z.L. and J.S.; investigation, Z.L. and J.S.; resources, Z.L. and J.S.; writing—original draft preparation, Z.L.; writing—review and editing, J.S.; visualization, Z.L. and J.S.; supervision, J.S.; funding acquisition, J.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by National Natural Science Foundations of China (Grant Nos. 11971266, 11831010), and Shandong Provincial Natural Science Foundations (Grant Nos. ZR2020ZD24, ZR2019ZD42).

Data Availability Statement

Not applicable.

Acknowledgments

We would like to express our thanks to the anonymous referees and editors for their valuable comments and advantageous suggestions to improve the quality of this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wonham, W.M. On a matrix Riccati equation of stochastic control. SIAM J. Control 1968, 6, 681–697. [Google Scholar] [CrossRef]
  2. Bismut, J. Linear quadratic optimal stochastic control with random coefficients. SIAM J. Control Optim. 1976, 14, 419–444. [Google Scholar] [CrossRef]
  3. Bensoussan, A. Lectures on stochastic control. In Nonlinear Filtering and Stochastic Control; Springer: Berlin/Heidelberg, Germany, 1981. [Google Scholar]
  4. Peng, S.G. Stochastic Hamilton-Jacobi-Bellman equations. SIAM J. Control Optim. 1992, 30, 284–304. [Google Scholar] [CrossRef]
  5. Chen, S.P.; Li, X.J.; Zhou, X.Y. Stochastic linear quadratic regulators with indefinite control weight costs. SIAM J. Control Optim. 1998, 36, 1685–1702. [Google Scholar] [CrossRef]
  6. Chen, S.P.; Zhou, X.Y. Stochastic linear quadratic regulators with indefinite control weight costs. II. SIAM J. Control Optim. 2000, 39, 1065–1081. [Google Scholar] [CrossRef] [Green Version]
  7. Chen, S.P.; Yong, J.M. Stochastic linear quadratic optimal control problems. Appl. Math. Optim. 2001, 43, 21–45. [Google Scholar] [CrossRef]
  8. Ait Rami, M.; Moore, J.B.; Zhou, X.Y. Indefinite stochastic linear quadratic control and generalized differential Riccati equation. SIAM J. Control Optim. 2001, 40, 1296–1311. [Google Scholar] [CrossRef] [Green Version]
  9. Tang, S.J. General linear quadratic optimal stochastic control problems with random coefficients: Linear stochastic Hamilton systems and backward stochastic Riccati equations. SIAM J. Control Optim. 2003, 42, 53–75. [Google Scholar] [CrossRef]
  10. Yu, Z.Y. Equivalent cost functionals and stochastic linear quadratic optimal control problems. ESAIM Control Optim. Calc. Var. 2013, 19, 78–90. [Google Scholar] [CrossRef] [Green Version]
  11. Tang, S.J. Dynamic programming for general linear quadratic optimal stochastic control with random coefficients. SIAM J. Control Optim. 2015, 53, 1082–1106. [Google Scholar] [CrossRef]
  12. Sun, J.R.; Li, X.; Yong, J.M. Open-loop and closed-loop solvabilities for stochastic linear quadratic optimal control problems. SIAM J. Control Optim. 2016, 54, 2274–2308. [Google Scholar] [CrossRef] [Green Version]
  13. Sun, J.R.; Yong, J.M. Stochastic linear quadratic optimal control problems in infinite horizon. Appl. Math. Optim. 2018, 78, 145–183. [Google Scholar] [CrossRef]
  14. Sun, J.R.; Xiong, J.; Yong, J.M. Indefinite stochastic linear-quadratic optimal control problems with random coefficients: Closed-loop representation of open-loop optimal controls. Anna. Appl. Probab. 2021, 31, 460–499. [Google Scholar] [CrossRef]
  15. Davis, M.H.A. Linear Estimation and Stochastic Control; Chapman and Hall: London, UK, 1977. [Google Scholar]
  16. Anderson, B.D.O.; Moore, J.B. Optimal Control: Linear Quadratic Methods; Prentice Hall: Englewood Cliffs, NJ, USA, 1989. [Google Scholar]
  17. Yong, J.M.; Zhou, X.Y. Stochastic Controls: Hamiltonian Systems and HJB Equations; Springer: New York, NY, USA, 1999. [Google Scholar]
  18. Sun, J.R.; Yong, J.M. Stochastic Linear-Quadratic Optimal Control Theory: Open-Loop and Closed-Loop Solutions; Springer Briefs in Mathematics: Cham, Switzerland, 2020. [Google Scholar]
  19. Merton, R.C. Option pricing when underlying stock returns are discontinuous. J. Finan. Econom. 1976, 3, 125–144. [Google Scholar] [CrossRef] [Green Version]
  20. Kou, S.G. A jump-diffusion model for option pricing. Manag. Sci. 2002, 48, 1086–1101. [Google Scholar] [CrossRef] [Green Version]
  21. Cont, R.; Tankov, P. Financial Modelling with Jump Processes; Chapmam Hall/CRC: New York, NY, USA, 2004. [Google Scholar]
  22. Oksendal, B.; Sulem, A. Applied Stochastic Control of Jump Diffusions; Springer: Berlin/Heidelberg, Germany, 2005. [Google Scholar]
  23. Lim, A.E.B. Mean-variance hedging when there are jumps. SIAM J. Control Optim. 2005, 44, 1893–1922. [Google Scholar] [CrossRef] [Green Version]
  24. Hanson, F.B. Applied Stochastic Processes and Control for Jump-Diffusions: Modeling, Analysis and Computation; SIAM: Philadelphia, PA, USA, 2007. [Google Scholar]
  25. Tang, S.J.; Hou, S.H. Optimal control of point processes with noisy observations: The maximum principle. Appl. Math. Optim. 2002, 45, 185–212. [Google Scholar] [CrossRef]
  26. Wu, Z.; Wang, X.R. FBSDE with poisson process and its application to linear quadratic stochastic optimal control problem with random jumps. Acta Autom. Sin. 2003, 29, 821–826. [Google Scholar]
  27. Hu, Y.Z.; Oksendal, B. Partial information linear quadratic control for jump diffusions. SIAM J. Control Optim. 2008, 47, 1744–1761. [Google Scholar] [CrossRef] [Green Version]
  28. Meng, Q.X. General linear quadratic optimal stochastic control problem driven by a brownian motion and Poisson random martingale measure with random coefficients. Stoch. Anal. Appl. 2014, 32, 88–109. [Google Scholar] [CrossRef] [Green Version]
  29. Zhang, F.; Dong, Y.C.; Meng, Q.X. Backward stochatic Riccati equation with jumps associated with stochastic linear quadratic optimal control with jumps and random coefficients. SIAM J. Control Optim. 2020, 58, 393–424. [Google Scholar] [CrossRef]
  30. Li, N.; Wu, Z.; Yu, Z.Y. Indefinite stochastic linear-quadratic optimal control problems with random jumps and related stochastic Riccati equations. Sci. China Math. 2018, 61, 563–576. [Google Scholar] [CrossRef]
  31. Moon, J.; Chung, J.H. Indefinite linear-quadratic stochastic control problem for jump-diffusion models with random coefficients: A completion of squares approach. Mathematics 2021, 9, 2918. [Google Scholar] [CrossRef]
  32. Sun, J.R.; Yong, J.M. Linear quadratic stochastic differential games: Open-loop and closed-loop saddle points. SIAM J. Control Optim. 2014, 52, 4082–4121. [Google Scholar] [CrossRef] [Green Version]
  33. Li, X.; Sun, J.R.; Yong, J.M. Mean-field stochastic linear quadratic optimal control problems: Closed-loop solvability. Probab. Uncertain. Quant. Risk 2016, 1, 24. [Google Scholar] [CrossRef] [Green Version]
  34. Lv, Q. Well-posedness of stochastic Riccati equations and closed-loop solvability for stochastic linear quadratic optimal control problems. J. Differ. Equ. 2019, 267, 180–227. [Google Scholar]
  35. Lv, Q. Stochastic linear quadratic optimal control problems for mean-field stochastic evolution equations. ESAIM Control Optim. Calc. Var. 2020, 26, 127. [Google Scholar]
  36. Tang, C.; Li, X.Q.; Huang, T.M. Solvability for indefinite mean-field stochastic linear quadratic optimal control with random jumps and its applications. Optim. Control Appl. Meth. 2020, 41, 2320–2348. [Google Scholar] [CrossRef]
  37. Oksendal, B.; Zhang, T.S. The Itô-Wentzell formula and forward stochastic differential equations driven by Poisson random measures. Osaka J. Math. 2007, 44, 207–230. [Google Scholar]
  38. Athans, H. The matrix minmum principle. Inform. Control 1968, 11, 592–606. [Google Scholar] [CrossRef] [Green Version]
  39. Penrose, R. A generalized inverse of matrices. Proc. Camb. Philos. Soc. 1955, 52, 17–19. [Google Scholar] [CrossRef]
  40. Sun, J.R.; Yong, J.M. Linear quadratic stocahastic two-person nonzero-sum differential games: Open-loop and closed-loop Nash equilibria. Stochastic Process. Appl. 2019, 129, 381–418. [Google Scholar] [CrossRef] [Green Version]
  41. Li, X.; Shi, J.T.; Yong, J.M. Mean-field linear-quadratic stochastic differential games in an infinite horizon. ESAIM Control Optim. Calc. Var. 2021, 21, 78. [Google Scholar] [CrossRef]
  42. Li, Z.X.; Shi, J.T. Linear quadratic Stackelberg stochastic differential games: Closed-loop solvability. In Proceedings of the 40th Chinese Control Conference, Shanghai China, 26–28 July 2021; pp. 1063–1070. Available online: https://arxiv.org/abs/2012.14599 (accessed on 29 September 2022).
  43. Li, Z.X.; Shi, J.T. Linear quadratic leader-follower stochastic differential games: Closed loop solvability. J. Syst. Sci. Complex. 2021; accepted. [Google Scholar]
Figure 1. The relationship among problems.
Figure 1. The relationship among problems.
Mathematics 10 04062 g001
Figure 2. Riccati equation.
Figure 2. Riccati equation.
Mathematics 10 04062 g002
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, Z.; Shi, J. Closed-Loop Solvability of Stochastic Linear-Quadratic Optimal Control Problems with Poisson Jumps. Mathematics 2022, 10, 4062. https://doi.org/10.3390/math10214062

AMA Style

Li Z, Shi J. Closed-Loop Solvability of Stochastic Linear-Quadratic Optimal Control Problems with Poisson Jumps. Mathematics. 2022; 10(21):4062. https://doi.org/10.3390/math10214062

Chicago/Turabian Style

Li, Zixuan, and Jingtao Shi. 2022. "Closed-Loop Solvability of Stochastic Linear-Quadratic Optimal Control Problems with Poisson Jumps" Mathematics 10, no. 21: 4062. https://doi.org/10.3390/math10214062

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop