Next Article in Journal
Statistical Modeling of Shadows in SAR Imagery
Next Article in Special Issue
Numerical Linear Algebra for the Two-Dimensional Bertozzi–Esedoglu–Gillette–Cahn–Hilliard Equation in Image Inpainting
Previous Article in Journal
Passive Stabilization of Static Output Feedback of Disturbed Nonlinear Stochastic System
Previous Article in Special Issue
Schur Complement-Based Infinity Norm Bound for the Inverse of Dashnic-Zusmanovich Type Matrices
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On the Iterative Methods for the Solution of Three Types of Nonlinear Matrix Equations

1
Faculty of Economics and Business Administration, Sofia University “St.Kl.Ohridski”, 1000 Sofia, Bulgaria
2
College of Mathematics and Systems Science, Shandong University of Science and Technology, Qingdao 266590, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Mathematics 2023, 11(21), 4436; https://doi.org/10.3390/math11214436
Submission received: 12 September 2023 / Revised: 8 October 2023 / Accepted: 23 October 2023 / Published: 26 October 2023
(This article belongs to the Special Issue Numerical Analysis and Matrix Computations: Theory and Applications)

Abstract

:
In this paper, we investigate the iterative methods for the solution of different types of nonlinear matrix equations. More specifically, we consider iterative methods for the minimal nonnegative solution of a set of Riccati equations, a nonnegative solution of a quadratic matrix equation, and the maximal positive definite solution of the equation X + A X 1 A = Q . We study the recent iterative methods for computing the solution to the above specific type of equations and propose more effective modifications of these iterative methods. In addition, we make comments and comparisons of the existing methods and show the effectiveness of our methods by illustration examples.

1. Introduction

Nonlinear matrix equations are commonly used in many fields of scientific and engineering computing. Research on the existence and properties of the solution to the matrix equations, as well as the corresponding numerical methods, has important theoretical significance and practical value. In this paper, we focus on the iterative methods for the solution of different types of nonlinear matrix equations. More specifically, we consider the iterative methods for computing the minimal nonnegative solution of a set of Riccati equations, a nonnegative solution of the quadratic matrix equation, and the maximal positive definite solution of the equation X + A X 1 A = Q . Iterative methods for solving the nonlinear matrix equation, which avoid the calculation of the inverse matrix at each iteration step, have gained wide popularity [1,2]. Without commenting on the reliability of this approach, this is an efficient approach that speeds up convergence. Users of similar methods should be aware of the possibility that some of these methods may lose accuracy during the calculations and may not reach the result. Moreover, we study the recent iterative methods for computing the above specific type of equations and propose more effective modifications of these iterative methods.
The investigated equations can be encountered in various applied tasks, for example, in the solution of problems for stability analysis  [3,4]. There have been many published papers on the field of matrix iterative schemes and their applications. We cite some of them related to our investigation  [5,6,7,8].
We will exploit a class of nonnegative matrices for the first two equations. Some notations are made throughout this paper. A matrix is nonnegative if all entries are either greater than zero or equal to zero. The set of real r × n matrices is denoted as R r × n . The notations I or I r are used for an unit r × r matrix. We need an elementwise order relation. The inequality A B ( A > B ) for A = ( a i j ) , B = ( b i j ) means that a i j b i j ( a i j > b i j ) for all indexes i and j. A matrix A = ( a i j ) R p × p is said to be a Z-matrix if it has non-positive off-diagonal elements. A  Z-matrix Q has the presentation Q = γ I P , with P being a nonnegative matrix. Each M-matrix is a Z-matrix if γ ρ ( P ) , where ρ ( P ) is the spectral radius of P. A Z-matrix Q is called a non-singular M-matrix if γ > ρ ( P ) ; otherwise, it is a singular M-matrix. For the third equations, we will exploit a class of Hermitian matrices. If the matrix Q is positive definite, we write Q 0 or Q 0 for positive semidefinite. Therefore, P Q means the matrix P Q is positive semidefinite.

2. Numerical Methods for the Solution of a Set of Riccati Equations

In this section, we investigate different iterative methods to compute the minimal nonnegative solution of a set of matrix Riccati equations, where the matrix coefficients of each equation are associated with an M-matrix. Our investigation follows the ideas of Bai and coauthors in [9], Ma and Lu [10], Guan and Lu [11], Guan [12], and Ivanov and Yang [13]. In fact, we propose a new modification of the alternate linear implicit method introduced in [14] and modified in [15]. We propose a different iterative method to compute the minimal nonnegative solution and derive convergence properties of the new iteration. We apply some properties of M-matrices in the proof and show that the new iteration method is faster than Newton’s method investigated in [16] by Liu, Zhang, and Luo.
Consider a set of nonsymmetric coupled Riccati equations (SNCRE) associated with M-matrices:
M i ( X 1 , , X q ) : = X i C i X i X i D i A i X i + B i + j i e i j X j = 0 ,
i = 1 , , q , , which is introduced in [14]. The coefficients of matrix X i are A i = ( a k p i ) R m × m , B i R m × n , C i R n × m , D i = ( d k p i ) R n × n . Let ( X 1 , , X s ) be a solution of the set of Equation (1) with X i R m × n , i = 1 , , q . Entries of E = ( e i j ) are nonnegative constants.
The couple of matrices ( X ˜ 1 , , X ˜ q ) is the minimal nonnegative solution to (1) if X ˜ i X i , i = 1 , , q (elementwise order) for any nonnegative solution ( X 1 , , X q ) to (1).
Zhang and Tan [14] have investigated the inexact Newton method and the alternate linear implicit method (ALI) to compute the minimal nonnegative solution of the SNCRE (1). They have proved the convergence properties of these iterations. We define the ALI iterative method with initial matrices X i ( 0 ) = 0 R n × n ( m = n ) . The method uses positive constants γ i , i = 1 , , q , which are computed via ((31), [14]):
γ i = max { m a x j a j j i , m a x j d j j i } .
for k = 0 , 1 , 2 , : Y i ( k ) ( γ i I n + D i C i X i ( k ) ) = ( γ i I n A i ) X i ( k ) + B i + j i e i j X j ( k ) , ( γ i I n + A i Y i ( k ) C i ) X i k + 1 = Y i ( k ) ( γ i I n D i ) + B i + j i e i j Y j ( k ) .
Iteration (3) computes two inverse matrices at each iteration step. In order to avoid the computation of inverse matrices at each step, we have proposed a modification, as in [15]:
X i ( 0 ) = 0 , i = 1 , 2 , q , k = 0 , 1 , 2 , : Y i ( k ) ( γ i I n + D i ) = ( γ i I n A i + X i ( k ) C i ) X i ( k ) + B i + j i e i j X j ( k ) , ( γ i I n + A i ) X i k + 1 = Y i ( k ) ( γ i I n D i + C i Y i ( k ) ) + B i + j i e i j Y j ( k ) .
Iteration (4) and its convergence properties are derived in [15]. Here, the computations of the inverse matrices of ( γ i I n + D i ) and ( γ i I n + A i ) are executed in the beginning of the iterative process, i.e., it operates only one time. This fact significantly reduces the computational cost throughout the iteration process, which is confirmed by the numerical experiments executed in [15].

2.1. Newton Method and Its Modifications

Authors Liu, Zhang, and Luo [16] investigated the Newton method to compute the positive minimal solution to the set of Riccati Equation (1).
X i ( 0 ) = 0 , i = 1 , 2 , q , k = 0 , 1 , 2 , : ( A i X i ( k ) C i ) X i ( k + 1 ) + X i ( k + 1 ) ( D i C i X i ( k ) ) = B i + + j i e i j X j ( k ) C i X i ( k ) C i .
Together with (5), the following modifications are studied by the same authors:
X i ( 0 ) = 0 , i = 1 , 2 , q , k = 0 , 1 , 2 , : ( A i X i ( k ) C i ) X i ( k + 1 ) + X i ( k + 1 ) ( D i C i X i ( k ) ) = B i + + j < i e i j X j ( k + 1 ) + j > i e i j X j ( k ) C i X i ( k ) C i .
and
X i ( 0 ) = 0 , i = 1 , 2 , q , k = 0 , 1 , 2 , : ( A i X i ( k ) C i ) X i ( k + 1 ) + X i ( k + 1 ) ( D i C i X i ( k ) ) = B i + + j < i e i j ( ω X j ( k + 1 ) + ( 1 ω ) X j ( k ) ) + j > i e i j X j ( k ) C i X i ( k ) C i .
The proof of the convergence for (5) is derived by Liu, Zhang, and Luo in [16], whereas the iterations (6) and (7) are used by them as an empirical experiment. The idea of applying approximation X i ( k + 1 ) for computing X j ( k + 1 ) , j > i , as in (6), is an effective one.

2.2. Our New Iteration Scheme and Convergence Proof

Here, we propose the following iteration strategy to compute the minimal nonnegative solution to (1):
X i ( 0 ) = 0 , i = 1 , 2 , q γ i as in ( 2 ) , i = 1 , 2 , q k = 0 , 1 , 2 , : , 0 ω Y i ( k ) ( γ i I + D i ) = ( γ i I A i + X i ( k ) C i ) X i ( k ) + B i + j < i e i j [ ω Y j ( k ) + ( 1 ω ) X j ( k ) ] + j > i e i j X j ( k ) , ( γ i I + A i ) X i ( k + 1 ) = Y i ( k ) ( γ i I D i + C i Y i ( k ) ) + B i + j < i e i j [ ω X j ( k + 1 ) + ( 1 ω ) Y j ( k ) ] + j > i e i j Y j ( k ) .
We derive several matrix identities for matrices obtained by iteration (8) in the lemma.
Lemma 1.
The matrix sequences { X i ( k ) , Y i ( k ) } k = 0 are constructed by iteration (8) with initial values X i ( 0 ) = 0 , i = 1 , 2 , . . q . The following matrix identities are satisfied for k = 0 , 1 , , :
( i ) ( Y i ( k ) X i ( k ) ) ( γ i I + D i ) = ( X i ( k ) Y i ( k 1 ) ) ( γ i I D i ) + X i ( k ) C i ( X i ( k ) Y i ( k 1 ) ) + ( X i ( k ) Y i ( k 1 ) ) C i Y i ( k 1 ) + j < i e i j [ ω ( X j ( k + 1 ) X j ( k ) ) + ( 1 ω ) ( Y j ( k ) X j ( k ) ) ] + j > i e i j ( X j ( k ) Y j ( k ) ) , ( i i ) ( γ i I + A i ) ( X i ( k + 1 ) Y i ( k ) ) = ( γ i I A i ) ( Y i ( k ) X i ( k ) ) + Y i ( k ) C i ( Y i ( k ) X i ( k ) ) + ( Y i ( k ) X i ( k ) ) C i X i ( k ) + j < i e i j [ ω ( X j ( k + 1 ) Y j ( k ) ) + ( 1 ω ) ( Y j ( k ) X j ( k ) ) ] + j > i e i j ( Y j ( k ) X j ( k ) ) ,
where I is an identity n × n matrix.
Moreover, if  ( X ˜ 1 , , X ˜ q ) is an exact nonnegative solution of M i ( X 1 , , X q ) = 0 , the subsequent identities can be verified:
( i i i ) ( X ˜ i Y i ( k ) ) ( γ i I + D i ) = ( γ i I A i ) ( X ˜ i X i ( k ) ) + X i ( k ) C i ( X ˜ i X i ( k ) ) + ( X ˜ i X i ( k ) ) C i X ˜ i + j < i e i j [ ω ( X ˜ j Y j ( k ) ) + ( 1 ω ) ( X ˜ j X j ( k ) ) ] + j > i e i j ( X ˜ j X j ( k ) ) ( i v ) ( γ i I + A i ) ( X ˜ i X i ( k + 1 ) ) = ( X ˜ i Y i ( k ) ) ( γ i I D i ) + ( X ˜ i Y i ( k ) ) C i X ˜ i + Y i ( k ) C i ( X ˜ i Y i ( k ) ) + j < i e i j [ ω ( X ˜ i X j ( k + 1 ) ) + ( 1 ω ) ( X ˜ i Y j ( k ) ) ] + j > i e i j ( X ˜ i Y j ( k ) )
Proof. 
The proof is completed by direct calculations and matrices manipulations. We rewrite Equation (8) for X i ( k ) and consider the difference Y i ( k ) ( γ I + A i ) ( γ I 2 n + D i ) X i ( k ) . After some matrix calculations, we obtain the matrix identity (i). Subtracting matrix equations in (8), we derive (ii).    □
We prove the convergence of the matrix sequence generated by (8).
Theorem 1.
Suppose the matrix coefficients A i , D i of (1) are Z-matrices and B i , C i , ( i = 1 , , q ) are nonnegative. There must exist positive scalars γ i such that ( γ i I + A i ) and ( γ i I + D i ) are nonsingular M-matrices.
If there exits a nonnegative solution to set of matrix Equation (1), then the matrix sequences { X i ( k ) , Y i ( k ) } k = 0 i = 1 , , q generated by (8) satisfy the following monotonicity property:
(i) X ^ i X i ( k + 1 ) Y i ( k ) X i ( k ) for i = 1 , , q , k = 0 ,   1 , for an exact nonnegative solution ( X ^ 1 , , X ^ q ) of (1). Moreover, the same matrix sequences converge to the nonnegative minimal solution of (1).
(ii) Moreover, if  A i X ^ i C i and D i C i X ^ i , i = 1 , , q are nonsingular M-matrices, then A i X ˜ i C i and D i C i X ˜ i , i = 1 , , q are nonsingular M-matrices, i.e., matrices A i + X ˜ i C i and D i + C i X ˜ i , i = 1 , , q are c-stable.
Proof. 
Under assumptions that we have ( γ i I + A i ) 1 0 and ( γ i I + D i ) 1 0 , i = 1 , , q . Apply recurrence Equation (8) with X 1 ( 0 ) = = X q ( 0 ) = 0 and γ i computed by (2).
For Y 1 ( 0 ) , we have Y 1 ( 0 ) ( γ 1 I + D 1 ) = B 1 0 and Y 1 ( 0 ) = B 1 ( γ i I + D 1 ) 1 0 . For Y 2 ( 0 ) , we have Y 2 ( 0 ) ( γ 2 I + D 2 ) = B 2 + e 21 ω Y 1 ( 0 ) 0 . Thus, Y 2 ( 0 ) 0 . Therefore Y i ( 0 ) 0 , and Y i ( 0 ) X i ( 0 ) = 0 , i = 1 , , q .
Construct matrix sequences { X i ( k ) , Y i ( k ) } k = 0 i = 1 , , q by (8) and exploit the facts γ i I D i 0 and γ i I A i 0 , i = 1 , , q .
We assume that the inequalities are true X i ( p ) Y i ( p 1 ) X i ( p 1 ) 0 for some integer p.
Next, we prove that X i ( p + 1 ) Y i ( p ) X i ( p ) 0 , i = 1 , , q .
Taking into account of Lemma 1(i), we get:
( Y i ( p ) X i ( p ) ) = F i ( p ) ( γ i I + D i ) 1 0 ,
because
F i ( p ) : = ( X i ( p ) Y i ( p 1 ) ) ( γ i I D i ) + X i ( p ) C i ( X i ( p ) Y i ( p 1 ) ) + ( X i ( p ) Y i ( p 1 ) ) C i Y i ( p 1 ) + j < i e i j [ ω ( X j ( p + 1 ) X j ( p ) ) + ( 1 ω ) ( Y j ( p ) X j ( p ) ) ] + j > i e i j ( X j ( p ) Y j ( p ) ) 0 .
Note that:
ω ( X j ( p + 1 ) X j ( p ) ) + ( 1 ω ) ( Y j ( p ) X j ( p ) ) = ( Y j ( p ) X j ( p ) ) + ω ( X j ( p + 1 ) Y j ( p ) ) 0 .
Thus, ω ( X j ( p + 1 ) X j ( p ) ) + ( 1 ω ) ( Y j ( p ) X j ( p ) ) 0 for positive ω and all j.
Therefore, Y i ( p ) X i ( p ) 0 , i = 1 , , q .
Taking account of Lemma 1(ii), we have:
( X i ( p + 1 ) Y i ( p ) ) = ( γ i I + A i ) 1 G i ( p ) ,
where
G i ( p ) = ( γ i I A i ) ( Y i ( p ) X i ( p ) ) + Y i ( p ) C i ( Y i ( p ) X i ( p ) ) + ( Y i ( p ) X i ( p ) ) C i X i ( p ) + j < i e i j [ ω ( X j ( p + 1 ) Y j ( p ) ) + ( 1 ω ) ( Y j ( p ) X j ( p ) ) ] + j > i e i j ( Y j ( p ) X j ( p ) ) 0 .
Thus, X i ( p + 1 ) Y i ( p ) 0 , i = 1 , , q .
We conclude that the matrix sequences { X i ( k ) , Y i ( k ) } k = 0 are monotone increasing. We have to prove that they are bonded above. Consider any exact nonnegative solution ( X ^ 1 , , X ^ q ) of (1). We shall prove that the solution is an upper bound of the matrix sequences.
For k = 0 , we have X ^ i X i ( 0 ) = 0 . We compute Y i ( 0 ) , i = 1 , q , and by (Lemma 1(iii)):
X ^ i Y i ( 0 ) = ( Q i ( 0 ) + S i ( 0 ) ) ( γ i I + D i ) 1 ,
where
Q i ( 0 ) = ( γ i I A i ) ( X ^ i X i ( 0 ) ) + X i ( 0 ) C i ( X ^ i X i ( 0 ) ) + ( X ^ i X i ( 0 ) ) C i X ^ i ,
and
S i ( 0 ) = j < i e i j [ ω ( X ˜ j Y j ( 0 ) ) + ( 1 ω ) ( X ˜ j X j ( 0 ) ) ] + j > i e i j ( X ˜ j X j ( 0 ) ) .
Note that Q i ( 0 ) 0 , i = 1 , , q .
Moreover, for i = 1 we have X ^ 1 Y 1 ( 0 ) 0 , because:
S 1 ( 0 ) = j > 1 e 1 j ( X ^ j X j ( 0 ) ) 0 .
For i = 2 , we obtain:
S 2 ( 0 ) = e 21 [ ω ( X ^ 1 Y 1 ( 0 ) ) + ( 1 ω ) ( X ^ 1 X 1 ( 0 ) ) ] + j > 2 e 2 j ( X ^ j X j ( 0 ) ) 0 .
Thus, X ^ 2 Y 2 ( 0 ) 0 . We conclude that X ^ j Y j ( 0 ) 0 j = 1 , , q .
Thus:
X ^ j Y j ( 0 ) X j ( 0 ) 0 , j = 1 , , q .
We will prove:
X ^ j X j ( 1 ) , j = 1 , , q .
From Lemma 1(iv), for  k = 0 , we obtain:
X ^ i X i ( k + 1 ) = ( γ i I + A i ) 1 ( G Y i ( 0 ) + L i ( 0 ) ) ,
where
G Y i ( 0 ) = ( X ˜ i Y i ( 0 ) ) ( γ i I D i ) + ( X ^ i Y i ( 0 ) ) C i X ^ i + Y i ( 0 ) C i ( X ^ i Y i ( 0 ) ) ,
which is a nonnegative matrix. For the matrix L i ( 0 ) , write:
L i ( 0 ) = j < i e i j [ ω ( X ^ j X j ( 1 ) ) + ( 1 ω ) ( X ^ j Y j ( 0 ) ) ] + j > i e i j ( X ^ j Y j ( 0 ) ) .
For i = 1 , we obtain:
L 1 ( 0 ) = j > 1 e 1 j ( X ^ j Y j ( 0 ) ) 0 ,
and thus X ^ 1 X 1 ( 1 ) 0 . For i = 2 , write:
L 2 ( 0 ) = e 21 [ ω ( X ˜ 1 X 1 ( 1 ) ) + ( 1 ω ) ( X ^ 1 Y 1 ( 0 ) ) ] + j > 2 e 2 j ( X ^ j Y j ( 0 ) ) 0 ,
which leads to X ^ 2 X 2 ( 1 ) 0 .
Consequently, we infer X ^ j X j ( 1 ) 0 , j = 1 , , q .
Assume:
X ^ j X j ( k ) Y j ( k 1 ) 0 , j = 1 , , q .
With similar reasoning, we derive the inequalities:
X ^ j X j ( k + 1 ) Y j ( k ) 0 , j = 1 , , q .
Both matrix sequences are monotone increasing in the elemenwise order and bounded by the above. They converge to same limit ( P ˜ 1 , , P ˜ q ) . Going to the limits in Equation (8), one concludes that ( P ˜ 1 , , P ˜ q ) is a nonnegative solution of (1).
Suppose there is another solution ( S ˜ 1 , , S ˜ q ) with S ˜ j P ˜ j . The last inequalities lead us to a contradiction with the inequalities S ˜ j P ˜ j . Therefore, the solution ( S ˜ 1 , , S ˜ q ) is the minimal one.
Furthermore, we shall prove point (ii) of the theorem. Matrices A i X ^ i C i and D i C i X ^ i , i = 1 , , q are nonsingular M-matrices for an upper nonnegative limit ( X ^ 1 , , X ^ q ) for nonnegative solutions of (1). According to properties of M-matrices, we conclude that A i X ˜ i C i is a nonsingular M-matrix for i = 1 , , q and, moreover, A i + X ˜ i C i , i = 1 , , s is c-stable for the minimal nonnegative solution ( X ˜ 1 , , X ˜ q ) of (1).    □
Remark 1.
The existence of a nonnegative solution for the set of matrix Equation (1) is commented by [16]. Two assumptions are necessary in [14], which involve the existence of a nonnegative matrices sequence Z 1 , , Z q such that M i ( Z 1 , , Z q ) 0 . However, we drop this condition in our investigation. We derive a direct convergence proof for iteration (8) based on Lemma 1.
Remark 2.
We use the parameter ω, the values of which are bigger than 2 in (8) in order to speed up the rate of convergence for (8) comparing with the case ω = 1 . We denote W j , ω = ω Y j ( k ) + ( 1 ω ) X j ( k ) and V j , ω = ω X j ( k + 1 ) + ( 1 ω ) Y j ( k ) . For  ω > 2 , we have W j , ω W j , ω = 1 = ω Y j ( k ) + ( 1 ω ) X j ( k ) Y j ( k ) = ( ω 1 ) Y j ( k ) + ( 1 ω ) X j ( k ) 0 . That means W j , ω > 2 W j , ω = 1 0 . Analogously, the inequality V j , ω > 2 V j , ω = 1 0 is true for all values of j. We expect that iteration (8) for ω 2 makes a smaller number of iteration steps than the case ω = 1 . We shall track this fact in numerical experiments.
The above remarks allow choosing ω > 1 , and confirm that the choice preserves the monotony of the matrix sequences { X i ( k ) , Y i ( k ) } k = 0 i = 1 , , q .

2.3. Numerical Experiments

We provide numerical experiments to compute the minimal nonnegative solution to (1). We compare the results of iterations (5)–(7) with the results of the proposed new iterations (8). All experiments are performed in MATLAB (version R2018b) on a personal computer. The iterations stop when the current iterative step satisfies R E S i 10 × 10 12 , where R E S i is defined as [14]:
R E S i : = M i ( X 1 ( k ) , , X q ( k ) ) M i ( X 1 ( 0 ) , , X q ( 0 ) ) ,
i = 1 , , q .
In the experiments, we choose the parameters γ i , as defined in (2). We take X 1 ( 0 ) = = X q ( 0 ) = 0 for all examples and all iterative methods. Thus, M i ( X 1 ( 0 ) , , X q ( 0 ) ) = B i .
Example 1.
A set of n × n matrix coefficients for different values of n are tested. The matrices A i , D i , i = 1 , 2 , 3 are introduced following the Matlab terminology:
A 1 = A 2 = A 3 = zeros(n,n);
For i = 1 : n , A 1 ( i , i ) = 4 ; A 2 ( i , i ) = 3 ; A 3 ( i , i ) = 2 ; end
For i = 1 : n 1 , A 1 ( i , i + 1 ) = 0.5 ; A 1 (i + 1,i) = 0.03 ; end
For i = 1 : n 2 , A 1 ( i , i + 2 ) = 0.25 ; A 1 ( i + 2 , i ) = 0.9 ; end
A 1 ( 1 , n ) = 0.05 ;   A 1 ( n , 1 ) = 0.4 ;
A 2 = A 1 ; A 2 ( 1 , n ) = 0.8 ;   A 2 ( n , 1 ) = 0.06 ;
A 3 = A 1 ; A 3 ( 1 , n ) = 0.7 ;   A 3 ( n , 1 ) = 0.09 ;
B 1 = B 2 = B 3 = 0.75 I n , B 2 = B 1 , B 3 = B 1 , C 1 = 0.92 I n , C 2 = C 1 , C 3 = C 1 , where I n is an identity matrix order n.
E = ( e i j ) = 0.0661 0.4512 0.8887 0.4965 0.3156 0.8780 0.6542 0.8914 0.1947 ,
The results from the experiments are presented at Table 1. A hundred runs are executed for each example for n = 12 and  n = 24 . Ten runs are executed for n = 48 . In this case, iterations (5) and (6) are very slowly (in the used computer), whereas iteration (8) is fastest.
Example 2.
A set of n × n matrix examples with the matrix coefficients for different values of n are tested.
The matrices A i , D i , i = 1 , 2 , 4 are introduced following the Matlab terminology:
A 1 = g a l l e r y (‘tridiag’ , n , 0 , 1 , 1 ) ; A 1 = f u l l ( A 1 ) ;
A 2 = g a l l e r y (‘tridiag’ , n , 0 , 2 , 1 ) ; A 2 = f u l l ( A 2 ) ;
A 3 = g a l l e r y (‘tridiag’ , n , 0 , 3 , 1 ) ; A 3 = f u l l ( A 3 ) ;
A 4 = g a l l e r y (‘tridiag’ , n , 0 , 4 , 1 ) ; A 4 = f u l l ( A 4 ) ;
D 1 = g a l l e r y (‘tridiag’ , n , 0 , 2 , 1 ) ; D 1 = f u l l ( D 1 ) ;
D 2 = g a l l e r y (‘tridiag’ , n , 0 , 4 , 1 ) ; D 2 = f u l l ( D 2 ) ;
D 3 = g a l l e r y (‘tridiag’ , n , 0 , 6 , 1 ) ; D 3 = f u l l ( D 3 ) ;
D 4 = g a l l e r y (‘tridiag’ , n , 0 , 8 , 1 ) ; D 4 = f u l l ( D 4 ) ;
B 1 = 0.5 e y e ( n , n ) ; B 2 = B 1 ; B 3 = B 1 ; B 4 = B 1 ;
C 1 = 0.2 e y e ( n , n ) ; C 2 = C 1 ; C 3 = C 1 ; C 4 = C 1 ;
E = r a n d ( 4 ) ;
The results from the experiments are presented in Table 2.
The experiments with the above examples show the effectiveness of the proposed iteration Formula (8). Moreover, the high value of ω speeds up the convergence.

3. Numerical Method for the Maximal Solutions of Specifical Nonlinear Matrix Equations

Consider the iterative solution to the following nonlinear matrix equations:
X + A X 1 A = Q ,
M Y 2 + N Y + P = 0 ,
investigated in [1,2,17]. Numerical methods on the specific solutions of the above matrix equations (maximal positive definite and minimal nonnegative) are investigated and some families of iterative formulas are proposed in [1,2,17]. However, comments and improvements of the proposed iteration schemes are provided to improve and accelerate the convergence. In this section, we will focus on the problem of how to accelerate the numerical solution of the above nonlinear matrix equations. The main tricks in the iterative methods proposed in these publications are to avoid the computation of an inverse matrix at each iteration step.
In general cases, the matrix A may be a real or complex square matrix. The notation A denotes a complex conjugate operation.

3.1. Iterative Solution of X + A X 1 A = Q

We firstly list several known algorithms for computing the maximal solution of X + A X 1 A = Q and compare their computational behavior.
Algorithm 1 follows iterative Formula (2.2) and the corresponding algorithm from [1].
Algorithm 1 For matrix equation X + A X 1 A = Q
1:
Introduce matrix coefficients A , Q = I and a small positive number t o l .
Take X 0 = Y 0 = I (the identity matrix).
2:
Y k + 1 = I + Y k ( 3 I + X k 2 X k Y k ) ,
X k + 1 = I A Y k + 1 A ,
3:
Stop if X k + 1 + A X k + 1 1 A I     t o l . Otherwise, k : = k + 1 go to 2.
end
Algorithm 2 follows iterative Formula (3.3) and the corresponding algorithm from [2].
Algorithm 2 For matrix equation X + A X 1 A = Q
1:
Introduce matrix coefficients A , Q = I and a small positive number t o l . Choose p = 1 , m = 1 , q 1 = 1 for Equation (1.9) [2].
Take X 0 = Y 0 = I (the identity matrix).
2:
E k = X k Y k ,
Y k + 1 = 2 5 I + 12 5 Y k + 1 5 ( E k + E k ) 7 5 Y k E k ) ,
X k + 1 = I A Y k + 1 A ,
R e s k = X k + 1 + A X k + 1 1 A I t o l .
3:
If  R e s k t o l then stop. Otherwise, k : = k + 1 go to 2.
end
In addition, we apply iterative Formula (3) from [18] to compute the same solution. The iteration (3) is:
X k + 1 = I A X k 1 A , X 0 = α I , 0.5 α 1 , k = 0 , 1 , .
Here, we apply Algorithms 1 and 2 and iteration (9) to compute the maximal positive definite solution to X + A X 1 A = I . We use t o l = 10 16 in examples. The computations are performed on a computer Intel(R) Core(TM) i7-1065G7 CPU @ 1.30 GHz via Matlab R2018b.
Example 3.
Consider the Example 3.1 introduced in [1]. The matrix is:
A = 1 40 2 1 3 4 7 6 5 9 4 8 10 6 3 5 2 8 .
We have executed 100 runs with all algorithms. Algorithm 1 makes 26 iteration steps for 0.0276 s. Algorithm 2 makes 21 iteration steps for 0.0238 s. The computer realization of iteration (9) performs 21 iteration steps for 0.0186 s. The performance results of the three algorithms are comparable and show their applicability.
Example 4.
The example is considered in [2] as Example 4.1.
A = 0.37 0.13 0.12 0.30 0.34 0.12 0.11 0.17 0.29 .
We have executed 100 runs with all algorithms. Algorithm 1 needs 81 iteration steps for 0.0754 s. Algorithm 2 needs 111 iteration steps for 0.1005 s. Iteration (9) performs 124 iteration steps for 0.0942 s. All three algorithms are working effectively for this example. The computational time is almost the same.
Example 5.
The example is introduced by Guo and Lancaster in [19] with:
A = 0.2 0.2 0.1 0.2 0.15 0.15 0.1 0.15 0.25 ,
We have executed 100 runs with all algorithms using two different values of t o l . We take t o l = 10 4 . Algorithm 1 needs 48 iteration steps to compute the solution for 0.05 s. Algorithm 2 needs 59 iteration steps for 0.0586 s. Iteration (9) applies only three iteration steps for 0.0064 s with α = 0.5 . Further on, we take t o l = 10 8 . Algorithm 1 needs 4714 iteration steps to compute the solution for 4.1306 s (for 100 runs). Algorithm 2 needs 5893 iteration steps for 5.7824 s (for 100 runs). However, iteration (9) has done only five iteration steps for 0.0101 s (for 100 runs) with α = 0.5 . Thus, iteration (9) is superior than Algorithms 1 and 2 when the maximal solution is computed in this example.
Example 6.
The example is firstly considered in [20] and, next, is investigated in [18]. The matrix A is defined:
A = A ˜ 2 A ˜ , A ˜ = 0.1 0.15 0.2598076 0.15 0.2125 0.0649519 0.2598076 0.0649519 0.1375 .
Algorithms 1 and 2 do not converge for this example. Iteration (9) with α = 0.5 converges to the maximal solution after 11 iteration steps for t o l = 10 7 . The maximal solution X ˜ is:
X ˜ = 0.500000082310064 0.000000016964994 0.000000002309095 0.000000016964994 0.729639588876686 0.132582448109853 0.000000002309095 0.132582448109853 0.576546597071862 .
The results of the experiments in this section show that the introduced iterative method (9) in [18] is effective and comparable to the iterative methods introduced in [1,2], and even better. Iterative method (9) uses the choice of an initial approximation depending on the value of α . How to make the choice of α can be read in [18]. Algorithms 1 and 2 avoid the computation of the inverse matrix, but this is not always reliable, as can be seen from the examples discussed in this section. Thus, we have to be careful where the inverse free algorithm is applied.

3.2. Numerical Method for the Solution of M Y 2 + N Y + P = 0

In this section, we study square matrix equation M Y 2 + N Y + P = 0 , where M , N , P are real matrix coefficients. Different iterative methods are analyzed in  [17]. The authors of [17] have investigated a family of iterative methods for finding the minimal nonnegative solution to M Y 2 + N Y + P = 0 . Their conclusion shows that Algorithms 1 and 6 defined in [17] are able to find the corresponding solution with the given accuracy. We will present these two algorithms and propose their modifications to improve their computational behavior, i.e., we will propose new modifications of both algorithms to make them more effective in the computational aspects.
We describe Algorithm 1 proposed in [17] as Algorithm  3 here.
Algorithm 3 Algorithm  1 [17]
1:
Input n × n matrices M, N, P.
2:
We take Y 0 and α > 0 .
3:
Compute V M = α M , and  W M = ( 1 α ) M .
Note that M = V M + W M .
4:
Compute Y r + 1 from
( V M Y r + N + R ) Y r + 1 = ( R W M Y r ) Y r P .
5:
If  M Y r 2 + N Y r + P < t o l then stop.
Now, we introduce our modifications to the above algorithms. The aim of the modifications is to use a diagonal matrix W M = ξ I n . Then, the matrix multiplication W M Y r can be realized as ξ Y r in Matlab. Taking W M as a diagonal matrix, we preserve the properties of Theorem 2.4 proved by Erfanifar and Hajarian [17]. Thus, the matrix V M Y r + 1 + N is an M-matrix and the matrix sequence { Y r } is monotone increasing and then converges to the minimal nonnegative solution. Moreover, applying a diagonal form for the matrix W M , we avoid a matrix multiplication and replace it with a matrix multiplication by a number.
Compare the results from Algorithms 3 and 4 by Example 7.
Algorithm 4 Our modification to Algorithm 3
1:
Input n × n matrices M, N, P.
2:
Take Y 0 = 0 and α > 0 , R = α I n .
3:
Compute V M = M + R , and  W M = R , and  N N = N + R .
Note that M = V M + W M .
4:
Compute Y r + 1 using the equation
( V M Y r + N + R ) Y r + 1 = ( R W M Y r ) Y r P .
             4.1: Compute i m = i n v ( V M Y 0 + N N ) . (Remark Y 0 = Y r .)
             4.2: Compute t Q = ( R + α Y 0 ) Y 0 P .
             4.3: Compute Y 0 = i m t Q ; . (Remark Y 0 = Y r + 1 here).
             4.4: If  n o r m ( ( M Y 0 + N ) Y 0 + P ) t o l then stop. Ortherwise, r = r + 1 and go to Step 4.2.
5:
The computed solution is Y 0 .
Example 7
(Example 4.1, [17]). For s × s matrix coefficients M = ( m i j ) , P = ( p i j ) , N = ( n i j ) , we have:
m i i = 1.5 , i = 1 , s ; m i , i + 1 = 8 , m i + 1 , i = 5 , i = 1 , s 1 ; p i i = 0.5 , i = 1 , s ; p i , i + 1 = 0.8 , p i + 1 , i = 1.5 , i = 1 , s 1 ; n i i = 45 , i = 1 , s , n i , i + 1 = 6 , n i + 1 , i = 4 , i = 1 , s 1 , n 11 = n s s = 18 .
Introducing a vector-row of size s of units, i.e.,  e = ( 1 , , 1 ) , we compute e m a t = 0.1 e e , and  M = M - e m a t .
Based on the matrices M , N , P , we compute a nonnegative solution of matrix Equation (1) with Algorithms 3 and 4 with the stop criterion with t o l = 10 14 and compare numbers of iteration steps (It) and CPU time for 1000 runs for each value of s. The results are listed in Table 3.
Further on, we describe Algorithm 6 introduced in [17].
Applying the same approach, we obtain a modification of Algorithm 5.
Algorithm 5 Algorithm  6 [17]
1:
Input n × n matrices M, N, P.
2:
We take Y 0 and α > 0 , β > 0 .
3:
Compute V M = α M , and  W M = ( 1 α ) M .
V N = α N , and  W N = ( 1 β ) N .
Note that M = V M + W M and N = V N + W N .
4:
Compute Z r , Y r + 1 from
( V M Y r + V N + R ) Z r = ( R W M Y r W N ) Y r P ,
( W M Z r + V N + S ) Y r + 1 = ( S V M Z r W N ) Z r P .
5:
If  M Y r 2 + N Y r + P < t o l then stop. Ortherwise, r = r + 1 and go to Step 4.
We have performed experiments with Algorithms 4 and 5 for Example 7. The tol value is t o l = 10 14 and 1000 runs for each value of s are played. The results can be found in Table 4.
Algorithm 6 Our modification of Algorithm 5
1:
Input n × n matrices M, N, P.
2:
We take Y 0 and α > 0 , β > 0 , R = α I n , S = β I n .
3:
Compute V M = M + R , and  W M = R , and  V N = β N , W N = ( 1 β ) N and N N = V N + R , N M = V N + S .
Note that M = V M + W M and N = V N + W N .
4:
Compute Z r , Y r + 1 from matrix equations:
( V M Y r + V N + R ) Z r = ( R W M Y r W N ) Y r P ,
( W M Z r + V N + S ) Y r + 1 = ( S V M Z r W N ) Z r P .
            4.1: Compute i m = i n v ( V M Y 0 + N N ) . (Remark Y 0 = Y r .)
            4.2: Compute t Q = ( R + α Y 0 W N ) Y 0 P .
            4.3: Compute Z 0 = i m t Q ; . (Remark Z 0 = Z r here).
            4.4: Compute i m = i n v ( N M α Z 0 ) .
            4.5: Compute t Q = ( S V M X 0 W N ) Y 0 P .
            4.6: Compute Y 0 = i m t Q ; . (Remark Y 0 = Y r + 1 here).
            4.7: If  n o r m ( ( M Y 0 + N ) Y 0 + P ) t o l then stop. Ortherwise, r = r + 1 and go to Step 4.1.
5:
The computed solution is Y 0 .
Comparing Table 3 and Table 4, we conclude that Algorithm 5 is faster than Algorithm 3, and Algorithm 6 is faster than Algorithm 4. Algorithm 6 is faster than the remaining algorithms. The approach to divide the given iteration from Algorithm 3 in two parts, as it is shown in Algorithm 5, is more effective than the original one.

4. Conclusions

In this paper, we have studied numerical methods for three computational tasks: (a) to compute the minimal nonnegative solution of a set of Riccati equations, (b) to compute the maximal positive definite solution of the equation X + A X 1 A = Q , and (c) to compute the minimal nonnegative solution to the quadratic matrix equation M Y 2 + N Y + P = 0 . We have considered the existing iterative methods and have proposed their improvements to accelerate the convergence process. We have performed several numerical experiments for each task where to show the effectiveness of the proposed modifications.
Moreover, as a weakness of the iterative methods for solving task (b), we note that the application of the inverse free approach, where the computation of an inverse matrix is avoided, will save the cost of computation. However, the use of this approach is limited. This fact is confirmed by experiments in Section 3.1. In recent years, this approach has been widely used in the analysis of iterative solutions of matrix equations. The effectiveness of this approach will be investigated in our future work more deeply.

Author Contributions

Conceptualization, I.G.I. and H.Y.; methodology, I.G.I. and H.Y.; software, I.G.I. and H.Y.; validation, I.G.I. and H.Y.; formal analysis, I.G.I. and H.Y.; investigation, I.G.I. and H.Y.; supervision, I.G.I. and H.Y.; writing—original draft and improved manuscript, I.G.I. and H.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Not applicable.

Acknowledgments

Thanks to the reviewers for their useful comments, remarks, and constructive recommendations, which have increased the value of this manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Erfanifar, R.; Sayevand, K.; Esmaeili, H. A novel iterative method for the solution of a nonlinear matrix equation. Appl. Numer. Math. 2020, 153, 503–518. [Google Scholar] [CrossRef]
  2. Erfanifar, R.; Sayevand, K.; Hajarian, M. An efficient inversion-free method for solving the nonlinear matrix equation XP + j = 1 m Aj*XqjAj = Q. J. Frankl. Inst. 2022, 359, 3071–3089. [Google Scholar]
  3. Yang, Q.; Wang, X.; Cheng, X.; Du, B.; Zhao, Y. Positive periodic solution for neutral-type integral differential equation arising in epidemic model. Mathematics 2023, 11, 2701. [Google Scholar] [CrossRef]
  4. Fan, L.; Zhu, Q.; Zheng, W.X. Stability analysis of switched stochastic nonlinear systems with state-dependent delay. IEEE Trans. Autom. Control 2023, 1–8. [Google Scholar] [CrossRef]
  5. Erfanifar, R.; Sayevand, K.; Hajarian, M. Convergence analysis of Newton method without inversion for solving discrete algebraic Riccati equations. J. Frankl. Inst. 2022, 359, 7540–7561. [Google Scholar] [CrossRef]
  6. Erfanifar, R.; Sayevand, K.; Hajarian, M. Solving system of nonlinear matrix equations over Hermitian positive definite matrices. Linear Multilinear Algebra 2023, 71, 597–630. [Google Scholar] [CrossRef]
  7. Hasanov, V.I.; Ali, A.A. On convergence of three iterative methods for solving of the matrix equation X + AX−1A + BX−1B = Q. Comput. Appl. Math. 2017, 36, 79–87. [Google Scholar] [CrossRef]
  8. El-Sayed, S.M.; Ivanov, I.G.; Petkov, M.G. A new modification of the Rojo method for solving symmetric circulant five-diagonal systems of linear equations. Comput. Math. Appl. 1998, 35, 35–44. [Google Scholar] [CrossRef]
  9. Bai, Z.-Z.; Guo, X.-X.; Xu, S.-F. Alternately linearized implicit iteration methods for the minimal nonnegative solutions of the nonsymmetric algebraic Riccati equations. Numer. Linear Algebra Appl. 2006, 13, 655–674. [Google Scholar]
  10. Ma, C.; Lu, H. Numerical Study on Nonsymmetric Algebraic Riccati Equations. Mediterr. J. Math. 2016, 13, 4961–4973. [Google Scholar] [CrossRef]
  11. Guan, J.; Lu, L. New alternately linearized implicit iteration for M-matrix algebraic Riccati equations. J. Math. Study 2017, 50, 54–64. [Google Scholar]
  12. Guan, J. Modified alternately linearized implicit iteration method for M-matrix algebraic Riccati equations. Appl. Math. Comput. 2019, 347, 442–448. [Google Scholar]
  13. Ivanov, I.; Yang, H. An effective approach to solve a nonsymmetric algebraic Riccati equation. Innov. Model. Anal. J. Res. 2021, 6, 7–14. [Google Scholar]
  14. Zhang, J.; Tan, F. Numerical methods for the minimal non-negative solution of the non-symmetric coupled algebraic Riccati equation. Asian J. Control 2021, 23, 374–386. [Google Scholar] [CrossRef]
  15. Ivanov, I. Iterative computing the minimal solution of the coupled nonlinear matrix equations in terms of nonnegative matrices. Ann. Acad. Rom. Sci. Ser. Math. Appl. 2020, 12, 226–237. [Google Scholar] [CrossRef]
  16. Liu, J.; Zhang, J.; Luo, F. Newton’s method for the positive solution of the coupled algebraic Riccati equation applied to automatic control. Comput. Appl. Math. 2020, 39, 113. [Google Scholar] [CrossRef]
  17. Erfanifar, R.; Hajarian, M. Weight splitting iteration methods to solve quadratic nonlinear matrix equation MY2 + NY + P = 0. J. Frankl. Inst. 2023, 360, 1904–1928. [Google Scholar]
  18. Ivanov, I.G.; Hasanov, V.I.; Uhlig, F. Improved methods and starting values to solve the matrix equations X ± AX−1A = I iteratively. Math. Comput. 2005, 74, 263–278. [Google Scholar]
  19. Guo, C.-H.; Lancaster, P. Iterative Solution of Two Matrix Equations. Math. Comput. 1999, 68, 1589–1603. [Google Scholar] [CrossRef]
  20. Zhan, X. Computing the extremal positive definite solutions of a matrix equation. SIAMJ Sci. Comput. 1996, 17, 1167–1174. [Google Scholar] [CrossRef]
Table 1. Example 1 with (5)–(8).
Table 1. Example 1 with (5)–(8).
(5)(6)(7), ω = 1 . 2 (8), ω = 2 . 5
n I t CPU I t CPU I t CPU I t CPU
12344.3 s192.57 s182.44 s250.10 s
2438128.7 s2181.9 s1970.7 s280.33 s
10 runs
4822323.0 s22281 s20220.8 s330.23 s
Table 2. Example 2 for 10 runs with (5)–(8).
Table 2. Example 2 for 10 runs with (5)–(8).
(5)(6)(7), ω = 1 . 2 (8), ω = 2 . 5
n I t CPU I t CPU I t CPU I t CPU
12310.63 s170.32 s140.3 s170.03 s
243210.7 s164.62 s164.6 s190.03 s
4828406.2 s25386.6 s19291.9 s290.24 s
96slow convergence350.98 s
Table 3. Example 7 for 1000 runs with Algorithms 3 and 4.
Table 3. Example 7 for 1000 runs with Algorithms 3 and 4.
Algorithm 3Algorithm 4
s ( α ) I t CPU Time I t CPU Time
Seconds Seconds
10 (0.6)140.20130.19
20 (0.6)140.48130.42
30 (0.6)140.80130.74
40 (0.6)141.45131.41
50 (0.6)144.35133.87
60 (0.6)145.56135.25
70 (0.6)158.60148.22
80 (0.7)no convergence147.52
80 (0.9)158.37158.12
t o l = 10 13
90 (0.6)1410.51139.43
100 (0.6)1414.11313.71
Table 4. Example 7 for 1000 runs with Algorithms 5 and 6.
Table 4. Example 7 for 1000 runs with Algorithms 5 and 6.
Algorithm 5Algorithm 6Algorithm 6
α = 0.8 ,   β = 0.9 α = β = 0.94 α = 0.8 ;   β = 0.95
n ( α , β ) I t CPU Time I t CPU Time I t CPU Time
Seconds Seconds Seconds
1070.1560.1460.12
2070.3460.3160.28
3070.5960.5160.48
4071.0560.8760.85
5073.3261.9261.82
6073.1162.3862.46
7074.2563.5663.62
t o l = 10 13
8075.3264.7064.44
9075.6265.6065.87
100710.2668.3068.24
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ivanov, I.G.; Yang, H. On the Iterative Methods for the Solution of Three Types of Nonlinear Matrix Equations. Mathematics 2023, 11, 4436. https://doi.org/10.3390/math11214436

AMA Style

Ivanov IG, Yang H. On the Iterative Methods for the Solution of Three Types of Nonlinear Matrix Equations. Mathematics. 2023; 11(21):4436. https://doi.org/10.3390/math11214436

Chicago/Turabian Style

Ivanov, Ivan G., and Hongli Yang. 2023. "On the Iterative Methods for the Solution of Three Types of Nonlinear Matrix Equations" Mathematics 11, no. 21: 4436. https://doi.org/10.3390/math11214436

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop