Next Article in Journal
A Combined Interval Type-2 Fuzzy MCDM Framework for the Resilient Supplier Selection Problem
Previous Article in Journal
Nonsingular Terminal Sliding Mode Control Based on Adaptive Barrier Function for nth-Order Perturbed Nonlinear Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Full Information H2 Control of Borel-Measurable Markov Jump Systems with Multiplicative Noises

College of Mathematics and Systems Science, Shandong University of Science and Technology, Qingdao 266590, China
*
Author to whom correspondence should be addressed.
Mathematics 2022, 10(1), 37; https://doi.org/10.3390/math10010037
Submission received: 6 December 2021 / Revised: 21 December 2021 / Accepted: 22 December 2021 / Published: 23 December 2021
(This article belongs to the Topic Dynamical Systems: Theory and Applications)

Abstract

:
This paper addresses an H 2 optimal control problem for a class of discrete-time stochastic systems with Markov jump parameter and multiplicative noises. The involved Markov jump parameter is a uniform ergodic Markov chain taking values in a Borel-measurable set. In the presence of exogenous white noise disturbance, Gramian characterization is derived for the H 2 norm, which quantifies the stationary variance of output response for the considered systems. Moreover, under the condition that full information of the system state is accessible to measurement, an H 2 dynamic optimal control problem is shown to be solved by a zero-order stabilizing feedback controller, which can be represented in terms of the stabilizing solution to a set of coupled stochastic algebraic Riccati equations. Finally, an iterative algorithm is provided to get the approximate solution of the obtained Riccati equations, and a numerical example illustrates the effectiveness of the proposed algorithm.

1. Introduction

Markov jump systems belong to a class of multimodal stochastic dynamical models with regime switching governed by a Markov chain, and have found wide-spread applications ranging from manipulator robot [1], biology of viruses [2], portfolio selection [3] to transmission control of wire/wireless network [4]. According to the number of states included in the state space of Markov chain, finite and infinite Markov jump systems are classified and investigated, respectively. After a half century’s research, the control theory for finite Markov jump systems has reached a remarkable degree of maturity, such as stability analysis [5], observability and detectability [6], linear quadratic (LQ) optimal control [7], filter design [8] and finite-time control [9]. When the state space of Markov chain is expanded to an infinite set, there will arise some properties essentially different from the finite case. For example, it was pointed out in [10] that asymptotic mean square stability, stochastic L 2 -stability and exponential mean square stability are no more equivalent for countably infinite Markov jump systems, while these notions were proven in [11] to be equivalent for finite Markov jump systems. The technical difficulty for handling infinite Markov jumps was also exhibited in [12] which is concerned with the stability of a class of stochastic time-delay differential dynamic systems with infinite Markovian switchings. Besides the intrinsic theoretical merits, it is more often appropriate to characterize the real scenarios by infinite state space. An illustrative example is presented in [13], where the change of atmospheric conditions is modeled as a Markov chain taking values in a Borel-measurable set. By now, an increasing interest was attracted to the control issues of Markov jump systems with Markov chain taking values in a Borel set [14,15,16]. However, compared to the existing literature of finite Markov jump systems, there remain many gaps in the study of countably-infinite/Borel-measurable Markov jump systems, which deserves more attention.
H 2 control, based on the seminal work of state-space description [17], was at the core of robust control theory. In contrast to H control (another popular robust control approach), H 2 control aims to minimize the perturbation influence on the output response caused by the exogenous additive white noise with known distribution; while H control method deals with the disturbance attenuation problem in the presence of random disturbance whose statistical law is unclear but total energy is finite. For finite Markov jump systems, H 2 control problem was elaborately addressed in [18,19,20] and the references therein. In a recent paper [21], an H 2 control problem was studied for discrete-time Markov jump systems where the Markov chain has a Borel state space. Based on the accessible information of output variable, the optimal H 2 controller is obtained via an optimal filter, which produces a separation principle for the concerned dynamics. Different from the problem formulation of [21], this study focuses on a class of discrete-time stochastic systems subject to both Borel-measurable Markov jump parameter and multiplicative noises. The importance for studying these type of systems lies in the following two aspects. On one hand, it was recognized that the multiplicative noise perfectly depicts the stochastic fluctuation of physical parameter caused by uncertain environment; on the other hand, Markov chain with a Borel-measurable state space can provide substantial benefit for real applications, e.g., the networked control systems analyzed in [22]. Besides, the information of system state is assumed to be fully accessible to measurement, and the optimal H 2 controller will be selected among the admissible control set with a prescribed dynamic structure but unfixed dimensions to minimize the H 2 norm of resulting closed-loop system.
The main contribution of this paper is as follows. Firstly, we present the characterization of H 2 norm in terms of controllability and observability Gramians, which quantify the stationary variance of the perturbed output response. This result can be regarded as an extension of Proposition 3.1 [21] to the considered systems. It is remarkable that this formula is not only necessary to handle the considered H 2 optimal control problem, but also paves the way for further studying H 2 filter about the concerned systems. Different from the method used in [21], our technique takes full advantage of the internal stability and avoids the analysis of the asymptotic tendency about the covariance of the augmented state variable, which is generally very challenging, especially taking into account that the considered model is more complex than that of [21]. Secondly, among all n c -dimensional dynamic stabilizing controllers, the optimal H 2 control strategy is achieved by a zero-order controller with feedback gain determined by the stabilizing solution to a set of coupled stochastic algebraic Riccati equations. This control design allows an off-line computation and hence can be readily realized in practice. Particularly, the obtained Riccati equations involve integrals over a continuous interval, which are completely different from the Riccati equations associated to finite Markov jump systems. Hence, an iterative algorithm is proposed to seek the numerical solution for these coupled Riccati equations, which can approximate the exact value of real solution according to any prescribed accuracy.
The rest of this paper is organized as follows. In Section 2, we give some preliminaries and then derive the Gramian representation of H 2 norm for the considered Borel-measurable Markov jump systems. Section 3 proceeds with the discussion of H 2 optimal control problem. In Section 4, a numerical algorithm is proposed for solving the coupled stochastic algebraic Riccati equations. Section 5 ends this paper with a concluding remark.
The notations adopted in this paper are standard. R n : n-dimensional real Euclidean space; R m × n : the normed linear space of all m by n real matrices; · : the Euclidean norm of R n or the operator norm of R m × n ; A : the transpose of a matrix (or vector) A; T r ( A ) : the trace of a square matrix A; S n : the set of n × n real symmetric matrices; A > 0 ( 0 ) : A is a positive (semi-)definite symmetric matrix; I n : the n × n identity matrix; δ ( · ) : the Kronecker functional; Z + : = { 0 , 1 , 2 , } .

2. H 2 Norm and Gramian

On a complete probability space ( Ω , F , P ) , we consider the following discrete-time Markov jump system with multiplicative noises:
G 0 x ( t + 1 ) = A 0 ( r t ) x ( t ) + k = 1 d A k ( r t ) x ( t ) w k ( t ) + B v ( r t ) v ( t ) , z ( t ) = C ( r t ) x ( t ) , x ( 0 ) R n , t Z + ,
where x ( t ) R n , z ( t ) R n z and v ( t ) R n v represent the system state, measurement output and exogenous random disturbance, respectively. The multiplicative noise { w ( t ) | w ( t ) = ( w 1 ( t ) , , w d ( t ) ) , t Z + } is a stationary process satisfying E w ( t ) = 0 and E [ w ( t ) w ( s ) ] = I d δ ( t s ) . In (1), { r t } t Z + is a Markov chain taking values in a Borel set S and having a transition probability kernel G ( · , · ) with respect to a measure μ defined on S . For any given B σ ( S ) (Borel σ -algebra of S ), there exists a uniform bounded probability density g ( · , · ) such that
G ( r t , B ) = P ( r t + 1 B | r t ) = B g ( r t , s ) μ ( d s ) .
Moreover, the initial distribution of Markov chain is described by P ( r 0 B ) = B ν ( s ) μ ( d s ) where ν > 0 is absolutely integrable with respect to μ . Thus, the probability density function of { r t } t Z + is formulated as follows:
π ( 0 , ) = ν ( ) , π ( t + 1 , ) = S g ( s , ) π ( t , s ) μ ( d s ) , t Z + .
In the subsequent discussion, there is some constant k > 1 such that g ( s , ) < k for arbitrary s , S , and { r t } t Z + is a uniform ergodic Markov chain [23]. The disturbance signal { v ( t ) } t Z + R n v is a sequence of zero mean white noises satisfying E v ( t ) = 0 and E [ v ( t ) v ( s ) ] = I n v δ ( t s ) .
Throughout this paper, we assume that three stochastic processes { r t } t Z + , { w ( t ) } t Z + and { v ( t ) } t Z + are mutually independent. All the coefficients of system (1) belong to H m × n with suitable dimensions m and n, where H m × n is a Banach space { H ( · ) : S R m × n | ess sup S H ( ) < } equipped with the norm H = ess sup S H ( ) . Besides, H 1 m × n is also a Banach space defined by { H ( · ) : S R m × n | S H ( ) μ ( d ) < } equipped with the norm H 1 : = S H ( ) μ ( d ) . Let F k be the σ -algebra σ { r t , w ( s ) , v ( s ) | 0 t k , 0 s k 1 } . When k = 0 , we set F 0 = σ { r 0 } .
To make the presentation more concise, the following linear operators will be used:
E ( X ) ( ) = S g ( , s ) X ( s ) μ ( d s ) , L ( Y ) ( ) = k = 0 d S g ( s , ) A k ( s ) Y ( s ) A k ( s ) μ ( d s ) , T ( X ) ( ) = k = 0 d A k ( ) E ( X ) ( ) A k ( ) ,
where X H n and Y H 1 n . It can be verified that L maps H 1 n to H 1 n , while E and T map H n to H n . Moreover, L and T are adjoint with respect to the following bilinear operator:
X , Y = S T r [ X ( ) Y ( ) ] μ ( d ) , X H n , Y H 1 n .
Definition 1
([16]). The following discrete-time stochastic system with Borel-measurable Markov jump parameter (denoted by ( A ; P ) ):
x ( t + 1 ) = A 0 ( r t ) x ( t ) + k = 1 d A k ( r t ) x ( t ) w k ( t ) , t Z +
is said to be strongly exponentially mean square stable (SEMSS) if r ( L ) < 1 , where r ( L ) indicates the spectral radius of the operator L .
Dince all the coefficients of L are real matrices, by the Krein–Rutman theorem [24], there must exist a real eigenvalue λ of L and a corresponding real eigenvector 0 X S n H 1 n such that r ( L ) = λ and L ( X ) = λ X .
The following Lyapunov-type stability theorem follows directly from Theorems 2.4, 2.5 and 2.6 of [25].
Lemma 1.
If ( A ; P ) is SEMSS, then
 (i) 
for any 0 M H n , the following linear equation admits a unique 0 U H n such that
U ( ) = T ( U ) ( ) + M ( ) , S ;
 (ii) 
for any 0 N H 1 n H n , the following linear equation admits a unique 0 V H 1 n H n such that
V ( ) = L ( U ) ( ) + N ( ) , S .
Lemma 2.
Let x 0 ( t ) be the state trajectory of system (1) corresponding to the initial state x ( 0 ) = 0 , then for any S and t Z + , there holds
E [ x 0 ( t + 1 ) x 0 ( t + 1 ) g ( r t , ) ] = S g ( s , ) k = 0 d A k ( s ) E [ x 0 ( t ) x 0 ( t ) g ( r t 1 , ) ] A k ( s ) μ ( d s ) + S π ( t , s ) g ( s , ) B v ( s ) B v ( s ) μ ( d s ) .
Proof. 
Set F ˜ t = F t σ ( w ( t ) ) σ ( v ( t ) ) , then it is obvious that F t F ˜ t . In terms of the state equation of (1), it can be computed that
E [ x 0 ( t + 1 ) x 0 ( t + 1 ) g ( r t , ) ] = E { E [ x 0 ( t + 1 ) x 0 ( t + 1 ) g ( r t , ) | F ˜ t 1 ] } = E { E A 0 ( r t ) x 0 ( t ) + k = 1 d A k ( r t ) x 0 ( t ) w k ( t ) + B v ( r t ) v ( t ) [ A 0 ( r t ) x 0 ( t ) + k = 1 d A k ( r t ) x 0 ( t ) w k ( t ) + B v ( r t ) v ( t ) ] g ( r t , ) | F ˜ t 1 } .
Because x 0 ( t ) , w k ( t ) and v ( t ) are all F ˜ t -measurable and mutually independent, it follows from (9) that
E [ x 0 ( t + 1 ) x 0 ( t + 1 ) g ( r t , ) ] = E E k = 0 d A k ( r t ) x 0 ( t ) x 0 ( t ) g ( r t , ) A k ( r t ) + B v ( r t ) B v ( r t ) g ( r t , ) | F ˜ t 1 = E S k = 0 d A k ( s ) x 0 ( t ) x 0 ( t ) g ( s , ) A k ( s ) + B v ( s ) B v ( s ) g ( s , ) g ( r t 1 , s ) μ ( d s ) .
By Fubini’s theorem, Equation (10) can be rewritten as
E [ x 0 ( t + 1 ) x 0 ( t + 1 ) g ( r t , ) ] = S g ( s , ) k = 0 d A k ( s ) E [ x 0 ( t ) x 0 ( t ) g ( r t 1 , s ) ] A k ( s ) + B v ( s ) B v ( s ) g ( s , ) E [ g ( r t 1 , s ) ] μ ( d s ) .
According to the formula (7) of [26], there stands E [ g ( r t 1 , s ) ] = π ( t , s ) , which together with (11) justifies the validity of (8). The proof is ended.    □
Remark 1.
If we denote Y ˜ ( t , ) = E [ x 0 ( t ) x 0 ( t ) g ( r t 1 , ) ] and X ˜ ( t , ) = S π ( t , s ) g ( s , ) B v ( s ) B v ( s ) μ ( d s ) , then Lemma 2 implies that Y ˜ ( t , ) satisfies the following Lyapunov equation:
Y ˜ ( t + 1 , ) = L ( Y ˜ ( t ) ) ( ) + X ˜ ( t , ) , Y ˜ ( 0 , ) = 0 S n , t Z + , S .
By induction, it can be derived that the solution to (12) can be represented as Y ˜ ( t ) = k = 0 t 1 L t k 1 ( X ˜ ( k ) ) for t 1 .
Next, we present the main result of this section, which led to the Gramian characterization for H 2 norm of system (1).
Theorem 1.
If ( A ; P ) is SEMSS, then the following equations admit a unique pair of solutions 0 U H n and 0 Y H n H 1 :
U ( ) = T ( U ) ( ) + C ( ) C ( ) , ( observability Gramian ) Y ( ) = L ( Y ) ( ) + X ( ) , ( controllability Gramian ) , t Z + , S ,
where
X ( ) = S π ( s ) g ( s , ) B v ( s ) B v ( s ) μ ( d s ) , π ( s ) : = lim k π ( k , s ) .
Furthermore, for the output response z ( t ) of system (1) corresponding to arbitrary initial state x ( 0 ) = x 0 R n , we have
lim t E z ( t ) 2 = S T r [ C ( s ) Y ( s ) C ( s ) ] μ ( d s )
= S S π ( s ) g ( s , ) T r [ B v ( s ) U ( ) B v ( s ) ] μ ( d s ) μ ( d ) .
Proof. 
Above all, since { r t } t Z + is a uniform ergodic Markov chain, by Theorem 16.2.1 of [23], there exists an invariant probability density π with the property π ( s ) = lim k π ( k , s ) and π ( ) = S π ( s ) g ( s , ) μ ( d s ) . Hence, X is well defined in (14). Further, the existence and uniqueness of the solutions to (13) are guaranteed by Lemma 1. To proceed, it can be derived from the linearity of system (1) that
z ( t ) = C ( r t ) x ( t ) = C ( r t ) x ( t ; 0 , v ) + C ( r t ) x ( t ; x 0 , 0 ) : = z ( t ; 0 , v ) + z ( t ; x 0 , 0 ) ,
where z ( t ; 0 , v ) is the zero-initial-state output response influenced by the exogenous disturbance v and z ( t ; x 0 , 0 ) is the unforced output response arising from the initial state x 0 . Noticing that x ( t ; 0 , v ) is just x 0 ( t ) in Lemma 2, we have
E z ( t ; 0 , v ) 2 = E { T r [ z ( t ; 0 , v ) z ( t ; 0 , v ) ] } = E { T r [ C ( r t ) x 0 ( t ) x 0 ( t ) C ( r t ) ] } = E { E { T r [ C ( r t ) x 0 ( t ) x 0 ( t ) C ( r t ) ] | F ˜ t 1 } } = E S T r [ C ( s ) x 0 ( t ) x 0 ( t ) C ( s ) g ( r t 1 , s ) ] μ ( d s ) = S T r { C ( s ) E [ x 0 ( t ) x 0 ( t ) g ( r t 1 , s ) ] C ( s ) } μ ( d s ) .
By Remark 1, it follows from (18) that
E z ( t ; 0 , v ) 2 = S T r [ C ( s ) Y ˜ ( t , s ) C ( s ) ] μ ( d s ) = S T r C ( s ) k = 0 t 1 L t k 1 ( X ˜ ( k ) ) ( s ) C ( s ) μ ( d s ) = S T r C ( s ) k = 0 t 1 L t k 1 ( X ) ( s ) C ( s ) μ ( d s ) + S T r C ( s ) k = 0 t 1 L t k 1 ( X ˜ ( k ) X ) ( s ) C ( s ) μ ( d s ) ,
where X is given by (14). Let Y ( ) = t = 0 L t ( X ) ( ) , then it can be verified that Y satisfies the second equation of (13). Now, taking the limit t in the first term of the last equality of (19), we get
lim t S T r C ( s ) k = 0 t 1 L t k 1 ( X ) ( s ) C ( s ) μ ( d s ) = S T r C ( s ) lim t k = 0 t 1 L t k 1 ( X ) ( s ) C ( s ) μ ( d s ) = S T r C ( s ) lim t τ = 0 t 1 L τ ( X ) ( s ) C ( s ) μ ( d s ) = S T r [ C ( s ) Y ( s ) C ( s ) ] μ ( d s ) .
By the assumption that ( A ; P ) is SEMSS (i.e., r ( L ) < 1 ), we can infer from [27] that
C ( s ) k = 0 t 1 L t k 1 ( X ˜ ( k ) X ) ( s ) C ( s ) ζ k = 0 t 1 L t k 1 ( X ˜ ( k ) X ) ( s ) ζ k = 0 t 1 λ α t k 1 X ˜ ( k ) X ,
where ζ = C ( r t ) C ( r t ) > 0 , λ > 0 and 0 < α < 1 . In view of π ( s ) = lim k π ( k , s ) , we have lim k X ˜ ( k ) X = 0 . Thus, taking the limit t in the second term of the last equality of (19) leads to that
lim t S T r C ( s ) k = 0 t 1 L t k 1 ( X ˜ ( k ) X ) ( s ) C ( s ) μ ( d s ) = 0 .
Combining (19), (20) and (22), we can conclude that
lim t E z ( t ; 0 , v ) 2 = S T r [ C ( s ) Y ( s ) C ( s ) ] μ ( d s ) .
Moreover, because ( A ; P ) is SEMSS, it is deduced from (17) that there exist γ > 0 and 0 < q < 1 such that
E z ( t ) z ( t ; 0 , v ) 2 = E z ( t ; x 0 , 0 ) 2 γ q t x 0 2 0 ( t ) ,
which implies that
lim t E z ( t ) 2 = lim t E z ( t ; 0 , v ) 2 = S T r [ C ( s ) Y ( s ) C ( s ) ] μ ( d s ) .
Hence, Equation (15) is validated.
Next, let us show (16). To this end, we make use of the bilinear operator (4) to reformulate (15) as follows:
lim t E z ( t ) 2 = S T r [ C ( s ) Y ( s ) C ( s ) ] μ ( d s ) = C C , Y = U T ( U ) , Y = U , Y T ( U ) , Y = U , Y U , L ( Y ) = U , Y L ( Y ) = U , X = S S π ( s ) g ( s , ) T r [ B v ( s ) U ( ) B v ( s ) ] μ ( d s ) μ ( d ) ,
where the third and seventh equalities follow from the first and second equations of (13), respectively. The proof is completed.    □
Remark 2.
In the literature of H 2 control theory, the stationary variance of output response caused by the exogenous white noise can serve as an H 2 norm of the perturbed stochastic dynamics. That is,
G 0 2 2 = lim t E z ( t ) 2 .
Hence, Theorem 1 supplies a Gramian characterization for the H 2 norm of system (1) in the presence of additive white noise disturbance v, and makes it possible to explore H 2 optimal control problem for the considered systems.

3. H 2 Optimal Control

In this section, we are concentrated on an H 2 optimal control problem for the following controlled discrete-time Borel-measurable Markov jump systems:
x ( t + 1 ) = A 0 ( r t ) x ( t ) + G 0 ( r t ) u ( t ) + k = 1 d [ A k ( r t ) x ( t ) + G k ( r t ) u ( t ) ] w k ( t ) + B v ( r t ) v ( t ) , z ( t ) = C ( r t ) x ( t ) D ( r t ) u ( t ) , x ( 0 ) R n , t Z + ,
where u ( t ) R n u is the control input. Without loss of generality, we assume that D ( ) D ( ) > ε I n u ( S ) for some ε > 0 . In what follows, let us consider the following type of n c -dimensional dynamic controllers:
G x ^ ( t + 1 ) = A ^ 0 ( r t ) x ^ ( t ) + G ^ 0 ( r t ) u ^ ( t ) + k = 1 d [ A ^ k ( r t ) x ^ ( t ) + G ^ k ( r t ) u ^ ( t ) ] w k ( t ) , y ( t ) = C ^ ( r t ) x ^ ( t ) + F ( r t ) u ^ ( t ) , x ^ ( 0 ) R n c , t Z + .
Through taking u ( t ) = y ( t ) and u ^ ( t ) = x ( t ) , the dynamic controller (29) is incorporated in the perturbed system (28) and the following augmented system is obtained:
G c x c ( t + 1 ) = A 0 c ( r t ) x c ( t ) + k = 1 d A k c ( r t ) x c ( t ) w k ( t ) + B c ( r t ) v ( t ) , z c ( t ) = C c ( r t ) x c ( t ) , t Z + .
In (30), the closed-loop state and the coefficients are given by
x c ( t ) = x ( t ) x ^ ( t ) , A k c ( r t ) = A k ( r t ) + G k ( r t ) F ( r t ) G k ( r t ) C ^ ( r t ) B ^ k ( r t ) A ^ k ( r t ) , B c ( r t ) = B v ( r t ) 0 , C c ( r t ) = C ( r t ) 0 D ( r t ) F ( r t ) D ( r t ) C ^ ( r t ) .
If system (30) is SEMSS when v ( t ) 0 , we will call (29) a stabilizing dynamic controller. In the sequel, the set of all stabilizing controller with the form (29) will be denoted as K . In the special case of n c = 0 , G turns out to be a zero-order dynamic controller, i.e., a state-feedback controller u ( t ) = F ( r t ) x ( t ) .
Our objective is to find a stabilizing controller in the admissible control set K such that the H 2 norm of (30) is minimized. Below, we introduce a standard notion which will be used in the subsequent discussion.
Definition 2.
For the following system:
x ( t + 1 ) = A 0 ( r t ) x ( t ) + G 0 ( r t ) u ( t ) + k = 1 d [ A k ( r t ) x ( t ) + G k ( r t ) u ( t ) ] w k ( t ) , t Z + ,
if there is u ( t ) = F ( r t ) x ( t ) such that the resulting closed-loop system of (32) is SEMSS, then we say ( A , G ) is stochastically stabilizable and F ( r t ) R n u × n is called a stabilizing feedback gain. Moreover, if there is H ( r t ) R n × n z such that the following system is SEMSS:
x ( t + 1 ) = [ A 0 ( r t ) + H ( r t ) C ( r t ) ] x ( t ) + k = 1 d A k ( r t ) x ( t ) w k ( t ) , z ( t ) = C ( r t ) x ( t ) , t Z + ,
then we say ( C , A ) is stochastically detectable.
The following result is an extension of Proposition 3.3 ([21]), which tackles the existence and uniqueness of the stabilizing solution to a set of coupled stochastic algebraic Riccati equations. The detailed proof can be presented by following the same line as [21].
Lemma 3.
If ( A , G ) is stochastically stabilizable and ( C , A ) is stochastically detectable, then the following set of coupled stochastic algebraic Riccati equations admit a unique solution 0 P H n :
P ( ) = T ( P ) ( ) + C ( ) C ( ) [ Φ ( P ) ( ) ] [ Π ( P ) ( ) ] 1 [ Φ ( P ) ( ) ] , Π ( P ) ( ) = k = 0 d G k ( ) E ( P ) ( ) G k ( ) + D ( ) D ( ) > 0 , Φ ( P ) ( ) = k = 0 d G k ( ) E ( P ) ( ) A k ( ) , S .
Moreover, F ( r t ) = [ Π ( P ) ( r t ) ] 1 Φ ( P ) ( r t ) is a stabilizing feedback gain (i.e., P is a stabilizing solution to (34)).
Now, we are prepared to give the main result of this section, which provides an H 2 optimal control design for the considered systems.
Theorem 2.
If ( A , G ) is stochastically stabilizable and ( C , A ) is stochastically detectable, then the H 2 optimal controller in the set K is given by
u * ( t ) = F * ( r t ) x ( t ) = [ Π ( P ) ( r t ) ] 1 [ Φ ( P ) ( r t ) ] x ( t ) ,
where P is the stabilizing solution of (34). Moreover, the minimal H 2 norm of G c is represented as
min G K G c 2 2 = G u * 2 2 = S S π ( s ) g ( s , ) T r [ B v ( s ) P ( ) B v ( s ) ] μ ( d s ) μ ( d ) ,
where G u * denotes the system (28) driven by u * ( t ) = F * ( r t ) x ( t ) .
Proof. 
By Lemma 3, Equation (34) admits a unique stabilizing solution 0 P H n and u * ( t ) is a stabilizing state-feedback controller. Furthermore, we can rewrite (34) as follows:
P ( ) = k = 0 d [ A k ( ) + G k ( ) F * ( ) ] E ( P ) ( ) [ A k ( ) + G k ( ) F * ( ) ] + C ˜ ( ) C ˜ ( ) , S ,
where F * ( ) is determined by (35) and C ˜ ( ) = [ C ( ) F * ( ) D ( ) ] . Via simple calculations, it can be verified that the solution P of (37) is indeed the observability Gramian of G u * . From Theorem 1 and Remark 2, we derive that
G u * 2 2 = S S π ( s ) g ( s , ) T r [ B v ( s ) P ( ) B v ( s ) ] μ ( d s ) μ ( d ) .
Next, it remains to show that u * ( t ) achieves the minimal H 2 norm among the set K . To this end, for any dynamic controller G K , the corresponding closed-loop system is given by (30). It is easy to get the observability Gramian equation for (30), which is given as follows:
U c ( ) = k = 0 d [ A k c ( ) ] E ( U c ) ( ) [ A k c ( ) ] + C c ( ) C c ( ) , S .
Making use of Theorem 1 and Remark 2 again, we have
G c 2 2 = S S π ( s ) g ( s , ) T r [ B c ( s ) U c ( ) B c ( s ) ] μ ( d s ) μ ( d ) .
Now, according to the structure of A k c ( ) , we can separate U c ( ) into four blocks as follows:
U c ( ) = U 11 c ( ) U 12 c ( ) U 12 c ( ) U 22 c ( ) .
Then, by combing (37) and (39), it can be verified that the following block matrix
U ¯ ( ) : = U 11 c ( ) U 12 c ( ) U 12 c ( ) U 22 c ( ) P ( ) 0 0 0 ( S )
satisfies the following Lyapunov equation:
U ¯ ( ) = k = 0 d [ A k c ( ) ] E ( U ¯ ) ( ) [ A k c ( ) ] + [ F * ( ) F ( ) C ^ ( ) ] Π ( P ) ( ) [ F * ( ) F ( ) C ^ ( ) ] ,
where F ( ) and C ^ ( ) are the coefficients of (29). Because G K is a stabilizing controller, we have that system (30) is SEMSS when v ( t ) 0 . Hence, by Lemma 1 (i), we can derive from Π ( P ) ( ) > 0 (see (34)) that the above Lyapunov Equation (43) admits a unique solution 0 U ¯ ( ) H 2 n . Therefore, from (40) and (42), it follows that
G c 2 2 = S S π ( s ) g ( s , ) T r [ B c ( s ) U ¯ ( ) B c ( s ) ] μ ( d s ) μ ( d ) + S S π ( s ) g ( s , ) T r [ B v ( s ) P ( ) B v ( s ) ] μ ( d s ) μ ( d ) S S π ( s ) g ( s , ) T r [ B v ( s ) P ( ) B v ( s ) ] μ ( d s ) μ ( d ) = G u * 2 2 ,
which completes the proof.    □
Remark 3.
The H 2 optimal control design proposed in Theorem 2 can be regarded as a generalization of Theorem 7.6 of [25]. More specifically, if the state space of { r t } t Z + reduces to a finite set, then the above result produces an H 2 optimal controller for discrete-time stochastic systems with finite Markov jump parameters and multiplicative noises, as considered in [25].

4. Numerical Algorithm

From the preceding analysis, the applicability of H 2 optimal controller depends on how to get the solution of coupled stochastic algebraic Riccati Equation (34). Based on the grid of state space S and the iterative algorithm proposed in [28], we can present the following (Algorithm 1) procedure to compute the approximate solution of (34).
Algorithm 1: Numerical algorithm for solving (34)
Input:  The number of nodes on the grid of S and the selected accuracy ϵ > 0 ;
1:
Submitting the coefficients on node i to the iterative algorithm of [28];
2:
Computing the numerical solution of (34) on node i;
3:
Counting the number of iterations on node i;
Output:  Numerical solution of (34) on each node and the number of iterations on each node.
Example 1.
Consider a thermal receiver with dynamics given by (28) with d = 1 . Let r t be a Markov chain with the state space S = { { 1 } × [ 0 , 1 ] , { 2 } × [ 0 , 1 ] } , which stands for two atmospheric conditions (sunny and cloudy). Given r t = { i , } ( i = 1 , 2 and [ 0 , 1 ] ), ℓ is supposed to have a Beta distribution on [ 0 , 1 ] with parameters α = β = 2 , i.e., B e ( 2 , 2 ) . By (2), the transition probability of Markov chain r t is represented as follows:
p i j = P r t + 1 = ( j , [ 0 , 1 ] ) | r t = ( i , ) = G ( ( i , ) , ( j , [ 0 , 1 ] ) ) = 0 1 p i j f ( s ) d s , i , j = 1 , 2 ,
where f ( s ) = 6 s ( 1 s ) ( s [ 0 , 1 ] ) is the probability density function of B e ( 2 , 2 ) . Obviously, Equation (45) implies that
g ( ( i , ) , ( j , s ) ) = p i j ( 6 s 6 s 2 ) , s [ 0 , 1 ] , i , j = 1 , 2 .
Thus, the coupled stochastic algebraic Riccati Equation (34) can be written as follows:
P ( i , ) = k = 0 d A k ( i , ) j = 1 2 0 1 p i j ( 6 s 6 s 2 ) P ( j , s ) d s A k ( i , ) + C ( i , ) C ( i , ) [ Φ ( P ) ( i , ) ] [ Π ( P ) ( i , ) ] 1 [ Φ ( P ) ( i , ) ] , Π ( P ) ( i , ) = k = 0 d G k ( i , ) j = 1 2 0 1 p i j ( 6 s 6 s 2 ) P ( j , s ) d s G k ( i , ) + D ( i , ) D ( i , ) > 0 , Φ ( P ) ( i , ) = k = 0 d G k ( i , ) j = 1 2 0 1 p i j ( 6 s 6 s 2 ) P ( j , s ) d s A k ( i , ) , ( i , ) S .
Now, we divide the interval [ 0 , 1 ] into 100 equal segments and the state space becomes:
S i = { { i } × [ 0 , 1 n ] , , { i } × [ m 1 n , m n ] , , { i } × [ n 1 n , 1 ] } , i = 1 , 2 , n = 100 .
Further, let us utilize the Trapezoidal Rule to compute the approximate value of the integral in (47):
0 1 p i j ( 6 s 6 s 2 ) P ( j , s ) d s p i j 6 [ 0 100 ( 0 100 ) 2 ] P ( j , n 100 ) · 1 200 + 2 n = 1 99 p i j 6 [ n 100 ( n 100 ) 2 ] P ( j , n 100 ) · 1 200 + p i j 6 [ 100 100 ( 100 100 ) 2 ] P ( j , n 100 ) · 1 200 = n = 1 99 p i j 6 [ n 100 ( n 100 ) 2 ] P ( j , n 100 ) · 1 100 : = E i j .
Hence, for the node ( i , m ) ( i = 1 , 2 ; m = m 100 , m = 0 , 1 , , 100 ) , the numerical solution of (34) can be approximately expressed by:
P ( i , m ) = k = 0 d A k ( i , m ) j = 1 2 E i j A k ( i , m ) + C ( i , m ) C ( i , m ) [ Φ ( P ) ( i , m ) ] [ Π ( P ) ( i , m ) ] 1 [ Φ ( P ) ( i , m ) ] , Π ( P ) ( i , m ) = k = 0 d G k ( i , m ) j = 1 2 E i j G k ( i , m ) + D ( i , m ) D ( i , m ) > 0 , Φ ( P ) ( i , m ) = k = 0 d G k ( i , m ) j = 1 2 E i j A k ( i , m ) .
At this stage, we have transformed (47) into a set of discrete-time algebraic Riccati equations as considered in [28]. By applying the iterative algorithm of [28], we can calculate the approximate solution of P ( i , m ) for every node.
Here, we set p 11 = 0.9 , p 12 = 0.1 , p 21 = 0.7 and p 22 = 0.3 . The coefficients of (28) are determined by
A j ( i , ) = A j ( i , 1 ) + [ A j ( i , 2 ) A j ( i , 1 ) ] , G j ( i , ) = G j ( i , 1 ) + [ G j ( i , 2 ) G j ( i , 1 ) ] , B v ( i , ) = B v ( i , 1 ) + [ B v ( i , 2 ) B v ( i , 1 ) ] , C ( i , ) 0.1884 , D ( i , ) 1 , ( i , j = 1 , 2 )
with the following parameters:
A 0 ( 1 , 1 ) = 1.9 , A 0 ( 1 , 2 ) = 1.7 , A 0 ( 2 , 1 ) = 1.2 , A 0 ( 2 , 2 ) = 1.4 , B v ( 1 , 1 ) = 1.2 , B v ( 1 , 2 ) = 0.9 , B v ( 2 , 1 ) = 1.1 , B v ( 2 , 2 ) = 0.7 , A 1 ( 1 , 1 ) = 1.4 , A 1 ( 1 , 2 ) = 1.3 , A 1 ( 2 , 1 ) = 1.6 , A 1 ( 2 , 2 ) = 1.8 , G 1 ( 1 , 1 ) = 1.3 , G 1 ( 1 , 2 ) = 1.4 , G 1 ( 2 , 1 ) = 1.3 , G 1 ( 2 , 2 ) = 1.4 , G 0 ( 1 , 1 ) = 1.1 , G 0 ( 1 , 2 ) = 1.4 , G 0 ( 2 , 1 ) = 1.0 , G 0 ( 2 , 2 ) = 0 .
For a given accuracy ϵ = 1 × 10 5 , by applying Algorithm 1, we have calculated (via Matlab) the numerical solution of (34) on each node, which is displayed in Figure 1.

5. Conclusions

In this paper, we studied the H 2 optimal control problem for Borel-measurable Markov jump systems with multiplicative noises. H 2 norm is introduced by the stationary variance of perturbed output response, which can be quantified by controllability and observability Gramians of the considered systems. Moreover, by means of the stabilizing solution to a set of coupled stochastic algebraic Riccati equations, H 2 optimal controller is obtained. The current study will yield some interesting open topics. For example, when the information of system state is only partially accessible to measurement, as considered in [21], how to settle the H 2 optimal control problem for the considered systems. This issue no doubt deserves a further research.

Author Contributions

H.M. studied the characterization of H 2 norm and H 2 optimal control design; Y.W. contributed to the numerical algorithm. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Natural Science Foundation of Shandong Province (No. ZR2016FM16), the Scientific Research Foundation of Shandong University of Science and Technology for Recruited Talents (No. 2016RCJJ031) and Elite Project of Shandong University of Science and Technology (No. 107).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Siqueira, A.A.G.; Terra, M.H. A fault-tolerant manipulator robot based on H2, H and mixed H2/H Markovian controls. IEEE Trans. Mechatron. 2009, 14, 257–263. [Google Scholar] [CrossRef]
  2. Sazonov, I.; Grebennikov, D.; Meyerhans, A.; Bocharov, G. Markov chain-based stochastic modelling of HIV-1 life cycle in a CD4 T Cell. Mathematics 2021, 9, 2025. [Google Scholar] [CrossRef]
  3. Yin, G.; Zhou, X.Y. Markowitz’s mean-variance portfolio selection with regime switching: From discrete-time models to their continuous-time limits. IEEE Trans. Autom. Control 2004, 49, 349–360. [Google Scholar] [CrossRef]
  4. Borisov, A.; Bosov, A.; Miller, G.; Sokolov, I. Partial diffusion Markov model of heterogeneous TCP link: Optimiation with incomplete information. Mathematics 2021, 9, 1632. [Google Scholar] [CrossRef]
  5. Li, Z.Y.; Zhou, B.; Wang, Y.; Duan, G.R. On eigenvalue sets and convergence rate of Itô stochastic systems with Markovian switching. IEEE Trans. Autom. Control 2011, 56, 1118–1124. [Google Scholar] [CrossRef]
  6. Shen, L.; Sun, J.; Wu, Q. Observability and detectability of discrete-time stochastic systems with Markovian jump. Syst. Control Lett. 2013, 62, 37–42. [Google Scholar] [CrossRef]
  7. Dragan, V.; Morozan, T. The linear quadratic optimization problems of a class of linear stochastic systems with multiplicative white noise and Markovian jumping. IEEE Trans. Autom. Control 2004, 49, 665–675. [Google Scholar] [CrossRef]
  8. Dong, H.; Wang, Z.; Ho, D.W.C.; Gao, H. Robust H filtering for Markovian jump systems with randomly occurring nonlinearities and sensor saturation: The finite horizon case. IEEE Trans. Signal Process. 2011, 59, 3048–3057. [Google Scholar] [CrossRef]
  9. Tian, Y.; Yan, H.; Zhang, H.; Yang, S.X.; Li, Z. Obeserver-based finite-time control of nonliear semi-Markovian jump systems with saturation constraint. IEEE Trans. Syst. Man Cybern. Syst. 2021, 51, 6639–6649. [Google Scholar] [CrossRef]
  10. Ma, H.; Jia, Y. Stability analysis for stochastic differential equations with infinite Markovian switchings. J. Math. Appl. 2016, 435, 593–605. [Google Scholar] [CrossRef]
  11. Hosoe, Y.; Hagivara, T. Equivalent stability notions, Lyapunov inequality, and its application in discrete-time linear systems with stochastic dyanmics determined by an i.i.d. process. IEEE Trans. Autom. Control 2019, 64, 4764–4771. [Google Scholar] [CrossRef] [Green Version]
  12. Song, R.; Zhu, Q. Stability of linear stochastic delay differential equations with infinte Markovian switchings. Int. J. Robust Nonlinear Control 2018, 28, 825–837. [Google Scholar] [CrossRef]
  13. Costa, O.L.V.; Figueiredo, D.Z. Stochastic stability of jump discrete-time linear systems with Markov chain in a general Borel space. IEEE Trans. Autom. Control 2014, 59, 223–227. [Google Scholar] [CrossRef]
  14. Li, C.; Chen, M.; Lam, J.; Mao, X. On exponential almost sure stability of random jump systems. IEEE Trans. Autom. Control 2012, 57, 3064–3077. [Google Scholar]
  15. Costa, O.L.V.; Figueiredo, D.Z. LQ control of discrete-time jump systems with Markov chain in a general Borel space. IEEE Trans. Autom. Control 2015, 60, 2530–2535. [Google Scholar] [CrossRef]
  16. Ma, H.; Cui, Y.; Wang, Y. H control of discrete-time stochastic systems with borel-measurable Markov jumps. IEEE Access 2020, 8, 49812–49822. [Google Scholar] [CrossRef]
  17. Doyle, J.; Glover, K.; Khargonekar, P.; Francis, B. State space solution to standard H2 and H control problems. IEEE Trans. Autom. Control 1989, 34, 831–847. [Google Scholar] [CrossRef]
  18. Costa, O.L.V.; Do Val, J.B.R.; Geromel, J.C. A Convex Programming Approach to H2 Control of Discrete-Time Markovian Jump Linear Systems. Int. J. Control 1997, 66, 557–579. [Google Scholar] [CrossRef]
  19. Dragan, V.; Morozan, T.; Stoica, A.M. H2 Optimal Control for Linear Stochastic Systems. Automatica 2004, 40, 1103–1113. [Google Scholar] [CrossRef]
  20. Ma, H.; Jia, Y.M. H2 control of discrete-time periodic systems with Markovian jumps and multiplicative noise. Int. J. Control 2013, 86, 1837–1849. [Google Scholar] [CrossRef]
  21. Figueiredo, D.Z.; Costa, O.L.V. H2 control and the separation principle for discrete-time jump systems with the Markov chain in a general state space. Int. J. Syst. Sci. 2017, 48, 2728–2741. [Google Scholar] [CrossRef]
  22. Kordonis, I.; Papavassilopoulos, G.P. On stability and LQ control of MJLS with a Markov chain with general state space. IEEE Trans. Autom. Control 2014, 59, 535–540. [Google Scholar] [CrossRef]
  23. Meyn, S.; Tweedie, R.L. Markov Chains and Stochastic Stability, 2nd ed.; Cambridge University Press: New York, NY, USA, 2009. [Google Scholar]
  24. Damm, T. Rational Matrix Equations in Stochastic Control; Springer: Berlin, Germany, 2004. [Google Scholar]
  25. Dragan, V.; Morozan, T.; Stoica, A.M. Mathematical Methods in Robust Control of Discrete-Time Linear Stochastic Systems; Springer: New York, NY, USA, 2010. [Google Scholar]
  26. Costa, O.L.V.; Figueiredo, D.Z. Quadratic control with partial information for discrete-time jump systems with the Markov chain in a general Borel space. Automatica 2016, 66, 73–84. [Google Scholar]
  27. Kubrusly, C.S. Mean square stability for discrete bounded linear systems in Hilbert space. Siam J. Control Optim. 1985, 23, 19–29. [Google Scholar] [CrossRef]
  28. Dragan, V.; Morozan, T.; Stoica, A.M. Iterative algorithm to compute the maximal and stabilising solution of a general class of discrete-time Riccati equation. Int. J. Control 2010, 83, 837–847. [Google Scholar]
Figure 1. Numerical solution and iterations on each node.
Figure 1. Numerical solution and iterations on each node.
Mathematics 10 00037 g001
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ma, H.; Wang, Y. Full Information H2 Control of Borel-Measurable Markov Jump Systems with Multiplicative Noises. Mathematics 2022, 10, 37. https://doi.org/10.3390/math10010037

AMA Style

Ma H, Wang Y. Full Information H2 Control of Borel-Measurable Markov Jump Systems with Multiplicative Noises. Mathematics. 2022; 10(1):37. https://doi.org/10.3390/math10010037

Chicago/Turabian Style

Ma, Hongji, and Yang Wang. 2022. "Full Information H2 Control of Borel-Measurable Markov Jump Systems with Multiplicative Noises" Mathematics 10, no. 1: 37. https://doi.org/10.3390/math10010037

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop