Next Article in Journal
The Convergence Rate on a Quadrature of a Fourier Integral with Symmetrical Jacobi Weight for an Analytical Function
Next Article in Special Issue
Finding Solutions to the Yang–Baxter-like Matrix Equation for Diagonalizable Coefficient Matrix
Previous Article in Journal
Analyzing the Jacobi Stability of Lü’s Circuit System
Previous Article in Special Issue
An Efficient Method for Split Quaternion Matrix Equation XAf(X)B = C
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Improved Convergence Theorem of the Newton-Based AOR Method for Generalized Absolute Value Equations

School of Mathematical Sciences, South China Normal University, Guangzhou 510631, China
*
Author to whom correspondence should be addressed.
Symmetry 2022, 14(6), 1249; https://doi.org/10.3390/sym14061249
Submission received: 7 May 2022 / Revised: 8 June 2022 / Accepted: 13 June 2022 / Published: 16 June 2022
(This article belongs to the Special Issue Tensors and Matrices in Symmetry with Applications)

Abstract

:
For solving the large sparse generalized absolute value equations, recently a Newton-based accelerated over-relaxation (NAOR) method was investigated. In this paper, we widen the convergence regions for the parameters and establish a new convergence theorem of the NAOR method when the system matrix is an H + -matrix. Numerical examples demonstrate that the NAOR method has a better convergence performance when the parameters are taken according to the proposed convergence theorem.

1. Introduction

Consider the generalized absolute value equation (GAVE)
A x B | x | = b ,
where A , B R n × n , b R n and | x | denotes the component-wise absolute value of the vector x. When B = I with I denoting the identity matrix, the GAVE becomes the absolute value equation A x | x | = b . The GAVE was introduced by Rohn [1] and further investigated in [2,3,4]. Many problems of scientific computing and engineering applications, such as interval linear equations [5] and linear complementarity problems [6,7,8] can be equivalently transformed into the GAVE.
There exist many efficient numerical methods for the GAVE; e.g., see [9,10,11,12,13,14,15,16,17,18], and references therein. Recently, Zhou et al. proposed a Newton-based matrix splitting (NMS) method by using the matrix technique [19]. Based on the different matrix splittings, the NMS method provides a general framework of Newton-based matrix splitting methods. As a special case of the NMS method, the Newton-based accelerated over-relaxation (NAOR) method is given as follows.
Method 1 (The NAOR method [19]). Let A = M N be a splitting of the matrix A with
M : = 1 α ( D A β L ) , N : = 1 α ( ( 1 α ) D A + ( α β ) L + α U ) , α , β R ,
where D A , L , and U are the diagonal, the strictly lower-triangular, and the strictly upper-triangular matrices of A, respectively. Assume that x ( 0 ) R n is an arbitrary initial guess. For k = 0 , 1 , 2 , until the iteration sequence { x ( k ) } k = 0 is convergent, computing x ( k + 1 ) by
x ( k + 1 ) = x ( k ) ( M + Ω ) 1 ( A x ( k ) B | x ( k ) | b ) = ( M + Ω ) 1 ( ( N + Ω ) x ( k ) + B | x ( k ) | + b ) ,
where M + Ω is invertible and Ω is a given matrix.
For α = β 0 , α = β = 1 and α = 1 , β = 0 , the NAOR method reduces the Newton-based successive over-relaxation (NSOR) method, the Newton-based Gauss–Seidel (NGS) method and the Newton-based Jacobi (NJ) method, respectively.
From Theorem 4.5 in [19], one can observe that the associated convergence conditions have not taken into account the choice of Ω and the upper bounds of α , β are smaller than 2. However, our numerical tests reveal that better performance of the NAOR method can be obtained for α , β larger than 2. This motivates us to look for some new convergence conditions of the NAOR method, which have the following properties:
(i)
Based on the choice of Ω , widen the convergence regions for the parameters α and β such that their optimal values can be included inside;
(ii)
With the above new convergence conditions, the NAOR method can obtain better convergence performance.
The rest of this paper is organized as follows. In Section 2, we introduce some notations, necessary definitions, and auxiliary results. In Section 3, an improved convergence theorem for the NAOR method is proved. Two numerical experiments and some concluding remarks are given in Section 4 and Section 5, respectively.

2. Preliminaries

Some notations, definitions, and basic results are given as follows, which can be found in [20,21].
For two real m × n matrices A = ( a i j ) and B = ( b i j ) , we write A B ( A > B ) if a i j b i j ( a i j > b i j ) for all 1 i m and 1 j n . A matrix A = ( a i j ) R m × n is said to be nonnegative (positive) if the entries satisfy a i j 0 ( a i j > 0 ) for all 1 i m and 1 j n . Let | A | = ( | a i j | ) R m × n be the absolute of the matrix A, and A T be its transpose. These notations apply to vectors in R n as well.
Let A be a square matrix and ρ ( A ) be its spectral radius. Its comparison matrix A = ( a i j ) is defined by a i j = | a i j | if i = j and a i j = | a i j | if i j . The matrix A is called a Z-matrix if all of its off-diagonal entries are non-positive, an M-matrix if it is a Z-matrix with A 1 0 , and an H-matrix if its comparison matrix A is an M-matrix. An H-matrix with positive diagonal entries is called an H + -matrix. If A is an M-matrix and B is a Z-matrix, then B A implies that B is an M-matrix.
Lemma 1
([20]). Let A R n × n be an H-matrix, then A is nonsingular and | A 1 | A 1 .
Lemma 2
([22]). Let A R n × n be a Z-matrix. Then the following statements are equivalent:
(i)
A is an M-matrix;
(ii)
If the representation A = M N satisfies that M 1 0 , N 0 , then ρ ( M 1 N ) < 1 .

3. Improved Convergence Theorem

In this section, we will establish a new convergence theorem of the NAOR method.
Theorem 1.
Let A , B R n × n and A be an H + -matrix with A = D A L U satisfying ρ : = ρ ( D A 1 ( | L | + | U | + | B | ) < 1 , where D A , L and U are the diagonal, the strictly lower-triangular and the strictly upper-triangular matrices of A, respectively. Assume that Ω is a positive diagonal matrix, then for any initial vector,
(i)
the NAOR method is convergent provided that α and β satisfy
0 < α 1 , or α > 1 with Ω α 1 α D A
and
0 β α with α 0 , or α < β < ( ρ + 1 ) α 2 ρ ,
respectively.
(ii)
The NSOR method is convergent for α satisfying (3);
(iii)
Both the NGS method and the NJ method are convergent.
Proof. 
We only need to verify the validity of ( i ) since the others are special cases. Let
M Ω : = M + Ω , N Ω : = | N + Ω | + | B | and A Ω : = M Ω N Ω .
where M and N are given as in (2). From Corollary 4.1 in [19], one only needs to prove
ρ ( | ( M + Ω ) 1 | N Ω ) < 1 ,
which is a sufficient convergence condition for the NAOR method.
Firstly, we will prove that the Z-matrix A Ω is an M-matrix. By substituting (2), the matrices M Ω and N Ω can be rewritten as
M Ω = Ω + 1 α D A β α | L | , N Ω = | Ω + 1 α α D A | + | α β | α | L | + | U | + | B | .
A simple computation gives that A Ω = M ˜ N ˜ with
M ˜ = Ω + 1 α D A | Ω + 1 α α D A | , N ˜ = β + | α β | α | L | + | U | + | B | .
From (3), it follows that M ˜ = D A . Obviously, M ˜ 1 > 0 and N ˜ 0 . By Lemma 2, A Ω is an M-matrix if and only if ρ ( M ˜ 1 N ˜ ) < 1 . For this, we distinguish the following two cases according to (4).
Case 1: The parameter β satisfies that 0 β α with α 0 .
For this case, it holds that N ˜ = | L | + | U | + | B | . Combining (3) and the assumption ρ < 1 , we obtain
ρ ( M ˜ 1 N ˜ ) = ρ ( D A 1 N ˜ ) = ρ < 1 .
Case 2: The parameter β satisfies that α < β < ( ρ + 1 ) α 2 ρ .
For this case, it follows that
N ˜ = 2 β α α | L | + | U | + | B | 2 β α α ( | L | + | U | + | B | ) .
Applying the monotonicity of the spectral radius of the nonnegative matrix and (3) gives
ρ ( M ˜ 1 N ˜ ) = ρ ( D A 1 N ˜ ) 2 β α α ρ ( D A 1 ( | L | + | U | + | B | ) ) = 2 β α α ρ ,
which, together with the assumption β < α ( ρ + 1 ) 2 ρ , implies that ρ ( M ˜ 1 N ˜ ) < 1 .
From Cases 1–2, we have proved that A Ω is an M-matrix for α and β satisfying (3) and (4). Together with the fact that M Ω A Ω , it follows that M Ω is an M-matrix, i.e., M + Ω is an H + -matrix. This implies that
M Ω 1 0 and | ( M + Ω ) 1 | M + Ω 1 = M Ω 1 ,
where the second inequality of (7) follows from Lemma 1. Note that M Ω 1 0 and N Ω 0 . From Lemma 2, we immediately obtain that ρ ( M Ω 1 N Ω ) < 1 . Moreover, the second inequality of (7) gives
ρ ( | ( M + Ω ) 1 | N Ω ) ρ ( M Ω 1 N Ω ) ,
from which the convergence condition (5) holds. This proof is completed. □
Remark 1.
Comparing Theorem 1 with Theorem 4.5 in [19], we give the following remarks.
(i)
In the proofs of the above two theorems, the key is to prove that A Ω is an M-matrix. In Theorem 4.5 in [19], A Ω is defined as A Ω = M Ω N ^ Ω , where M Ω is given by (6) and
N ^ Ω = Ω + | 1 α | α D A + | α β | α | L | + | U | + | B | .
Obviously, A Ω is independent of Ω. From (8) and (6), it holds that N ^ Ω N Ω . Different from [19], we set A Ω = M Ω N Ω , which depends on Ω. It is easy to show that M Ω N Ω is an M-matrix if M Ω N ^ Ω is an M-matrix, but not vice versa. This implies Theorem 1 may weaken the convergence conditions of the NAOR method.
(ii)
In Theorem 4.5 of [19], the convergence conditions on the parameters can be rewritten as
0 < β α < 2 1 + ρ or α < β < m i n ( ρ + 1 ) α 2 ρ , 2 ( 1 ρ ) α 2 ρ .
Comparing (9) with (3), (4), we see easily that Theorem 1 gives a wider convergence region for α and β than Theorem 4.5 in [19]. Thus, the NAOR method may have better performance by choosing the appropriate values of α and β according to (3) and (4). This means that Theorem 1 improves Theorem 4.5 in [19].

4. Numerical Results

In this section, we use the first two examples in [19] to examine the effectiveness of Theorem 1 for the NSOR method from three aspects: the number of iteration steps (denoted by ‘IT’), the elapsed CPU time in seconds (denoted by ‘CPU’), and the norm of the relative residual vectors (denoted by ‘RES’). Here, ‘RES’ is defined as
RES = A x ( k ) B | x ( k ) | b 2 / b 2 ,
where x ( k ) is the k-th approximate solution to the GAVE. All numerical experiments were performed in a MATLAB environment with double machine precision. In our tests, all initial vectors were chosen as
x ( 0 ) = ( 1 , 0 , 1 , 0 , , 1 , 0 , ) T R n
and all iterations were terminated once RES < 10 6 .
Next, we tested two special linear complementarity problems (LCP), i.e., Examples 1 and 2 in [19], which was also given in [23,24]. As an important application of the GAVE, the LCP( q , M ) is to find the vector z, such that
M z + q 0 , z 0 and z T ( M z + q ) = 0 ,
with M R n × n , q R n . Set
A = M + I , B = M I and x = 1 2 ( M I ) z + q ,
the LCP( q , M ) is equivalently transformed into the GAVE.
For simplicity, we use notations Tridiag ( · ) and tridiag ( · ) to denote the associated block-tridiagonal and tridiagonal matrices, respectively.
Example 1
(Zhou et al. [19], Example 1). Let m be a positive integer and n = m 2 . Consider the LCP( q , M ) with M = M ^ + μ I and q = M z * R n , where
M ^ = Tridiag ( I , S , I ) R n × n , S = tridiag ( 1 , 4 , 1 ) R m × m
and z * = ( 1.2 , 1.2 , , 1.2 , ) T R n is the unique solution of the LCP( q , M ).
Example 2
(Zhou et al. [19], Example 2). Let m be a positive integer and n = m 2 . Consider the LCP( q , M ) with M = M ^ + μ I and q = M z * R n , where
M ^ = Tridiag ( 1.5 I , S , 0.5 I ) R n × n , S = tridiag ( 1.5 , 4 , 0.5 ) R m × m
and z * = ( 1.2 , 1.2 , , 1.2 , ) T R n is the unique solution of the LCP( q , M ).
For Examples 1 and 2 with μ = 1 , since both the system matrices are M-matrices, the associated LCPs have the unique solutions by [7]. Moreover, their equivalent GAVEs (1) have also the unique solutions by (10).
According to Theorem 1, the NSOR method will converge to the unique solution x * for Ω and α satisfying the conditions (3). In the numerical tests, we set Ω = D A , which naturally satisfies (3). From this case, the NSOR method converges for α > 0 .
To further obtain the suitable range of α , the NSOR method is applied for the different scale problems with the changed α . The test results are demonstrated in Figure 1. From Figure 1, one can obverse that the iteration steps depend on the values of α but are nearly independent of the sizes of the test problems. In particular, the iteration steps tend to stabilize when α > 10 for Example 1 and α > 50 for Example 2. On the other hand, we observe that the NSOR method can attain its minimum iteration steps with α near to 5 for Example 1 and 50 for Example 2, respectively. This means that the convergence regions for α by Theorem 4.5 in [19] are too small to contain its optimal values. Therefore, we will take α ( 0 , 5 ) for Example 1 and ( 0 , 50 ) for Example 2 in the numerical test, respectively.
The numerical results of the NSOR method with different sizes are given in Table 1, where α ˜ e x p and α e x p are obtained experimentally from the different convergence regions for α by minimizing the associated iteration steps. Based on the above analysis, α ˜ e x p is chosen in the interval α ( 0 , 5 ) for Example 1 and α ( 0 , 50 ) for Example 2, respectively. For comparison, we also list the results in [19], where Ω = M ^ and α e x p are taken from the interval ( 0 , 2 ) .
From Figure 1 and Table 1, we have the following observations and remarks:
(1)
In terms of both the CPU times and the iteration steps, the NSOR method with Ω = D A and α ˜ e x p is always superior to the one with Ω = M ^ and α e x p . This means that the proposed convergence theorem improves Theorem 4.5 in [19] by taking the suitable values of α in a wider convergence region. In particular, the former has much less CPU time than the latter since the matrix Ω is taken as D A instead of tridiagonal matrix M ^ in [19].
(2)
The iteration steps are nearly independent of the scale of the test problems. Hence, a strategy of choosing α ˜ e x p or α e x p involves testing the small-scale problems and using them to the larger-scale problems.

5. Concluding Remarks

In this paper, we investigated the new convergence conditions of the NAOR method for solving the GAVE. By considering the relationship between Ω and the parameter α , we widened the convergence regions for the two parameters. Numerical results show that the NAOR method can obtain faster convergence when the appropriate parameters are chosen according to Theorem 1 instead of Theorem 4.5 in [19].
In numerical tests, the NSOR method can attain the convergence rate nearly independent of the scale of problems. However, it is difficult to prove independence in theory. Moreover, determining optimal parameters is still a challenging problem. We will further study these interesting topics in the future.

Author Contributions

Investigation, R.C. and X.P.; writing—original draft preparation, R.C.; writing—review and editing, X.P. and W.Y.; supervision, X.P. All authors have read and agreed to the published version of the manuscript.

Funding

The work was supported by Guangdong Basic and Applied Basic Research Foundation, no. 2021A1515012032, 2022A1515011123, the National Natural Science Foundation of China under grant nos. 12071159, U1811464, 11801097 and Foundation of Department of Education of Guangdong Province No. 2020ZDZX3004.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets generated during and/or analysed during the current study are available from the corresponding author on reasonable request.

Acknowledgments

The authors would like to thank the anonymous referees for their valuable comments and suggestions, which greatly improved the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Rohn, J. A theorem of the alternative for the equation Ax + B|x| = b. Linear Multilinear Algebra 2004, 52, 421–426. [Google Scholar] [CrossRef] [Green Version]
  2. Mangasarian, O.L.; Meyer, R.R. Absolute value equations. Linear Algebra Appl. 2006, 419, 359–367. [Google Scholar] [CrossRef] [Green Version]
  3. Mangasarian, O.L. Absolute value programming. Comput. Optim. Appl. 2007, 36, 43–53. [Google Scholar] [CrossRef]
  4. Rohn, J. On unique solvability of the absolute value equation. Optim. Lett. 2009, 3, 603–606. [Google Scholar] [CrossRef]
  5. Rohn, J. System of linear interval equations. Linear Algebra Appl. 1989, 126, 39–78. [Google Scholar] [CrossRef] [Green Version]
  6. Cottle, R.W.; Dantzig, G. Complementary pivot theory of mathematical programming. Linear Algebra Appl. 1968, 1, 103–125. [Google Scholar] [CrossRef] [Green Version]
  7. Cottle, R.W.; Pang, J.-S.; Stone, R.E. The Linear Complementarity Problem; Academic Press: San Diego, CA, USA, 1992. [Google Scholar]
  8. Ferris, M.C.; Pang, J.-S. Engineering and economic applications of complementarity problems. SIAM Rev. 1997, 39, 669–713. [Google Scholar] [CrossRef] [Green Version]
  9. Mangasarian, O.L. A generalized Newton method for absolute value equations. Optim. Lett. 2009, 3, 101–108. [Google Scholar] [CrossRef]
  10. Hu, S.-L.; Huang, Z.-H.; Zhang, Q. A generalized Newton method for absolute value equations associated with second order cones. J. Comput. Appl. Math. 2011, 235, 1490–1501. [Google Scholar] [CrossRef]
  11. Rohn, J.; Hooshyarbakhsh, V.; Farhadsefat, R. An iterative method for solving absolute value equations and sufficient conditions for unique solvability. Optim. Lett. 2014, 8, 35–44. [Google Scholar] [CrossRef]
  12. Salkuyeh, D.K. The Picard-HSS iteration method for absolute value equations. Optim. Lett. 2014, 8, 2191–2202. [Google Scholar] [CrossRef]
  13. Li, C.-X. A modified generalized Newton method for absolute value equations. J. Optim. Theory. Appl. 2016, 170, 1055–1059. [Google Scholar] [CrossRef]
  14. Wang, A.; Cao, Y.; Chen, J.-X. Modified Newton-type iteration methods for generalized absolute value equations. J. Optim. Theory Appl. 2019, 181, 216–230. [Google Scholar] [CrossRef]
  15. Ke, Y.-F.; Ma, C.-F. SOR-like iteration method for solving absolute value equations. Appl. Math. Comput. 2017, 311, 195–202. [Google Scholar] [CrossRef]
  16. Guo, P.; Wu, S.-L.; Li, C.-X. On the SOR-like iteration method for solving absolute value equations. Appl. Math. Lett. 2019, 97, 107–113. [Google Scholar] [CrossRef]
  17. Dong, X.; Shao, X.-H.; Shen, H.-L. A new SOR-like method for solving absolute value equations. Appl. Numer. Math. 2020, 156, 410–421. [Google Scholar] [CrossRef]
  18. Huang, B.-H.; Li, W. A modified SOR-like method for absolute value equations associated with second order cones. J. Comput. Appl. Math. 2022, 400, 113745. [Google Scholar] [CrossRef]
  19. Zhou, H.-Y.; Wu, S.-L.; Li, C.-X. Newton-based matrix splitting method for generalized absolute value equation. J. Comput. Appl. Math. 2021, 394, 113578. [Google Scholar] [CrossRef]
  20. Berman, A.; Plemmons, R.J. Nonnegative Matrix in the Mathematical Sciences; Academic Press: New York, NY, USA, 1979. [Google Scholar]
  21. Bai, Z.-Z. On the convergence of the multisplitting methods for the linear complementarity problem. SIAM J. Matrix Anal. Appl. 1999, 21, 67–78. [Google Scholar] [CrossRef]
  22. Plemmons, R.J. M-matrix characterizations. I-Nonsingular M-Matrices. Linear Algebra Appl. 1977, 18, 175–188. [Google Scholar] [CrossRef] [Green Version]
  23. Dong, J.-L.; Jiang, M.-Q. A modified modulus method for symmetric positive-definite linear complementarity problems. Numer. Linear Algebra Appl. 2009, 16, 129–143. [Google Scholar] [CrossRef]
  24. Bai, Z.-Z. Modulus-based matrix splitting iteration methods for linear complementarity problems. Numer. Linear Algebra Appl. 2010, 17, 917–933. [Google Scholar] [CrossRef]
Figure 1. The iteration steps of the NSOR method with Ω = D A and the different α : (ac) for Example 1 and (df) for Example 2 with n = m 2 .
Figure 1. The iteration steps of the NSOR method with Ω = D A and the different α : (ac) for Example 1 and (df) for Example 2 with n = m 2 .
Symmetry 14 01249 g001
Table 1. Numerical results of the NSOR method for Examples 1 and 2 with μ = −1.
Table 1. Numerical results of the NSOR method for Examples 1 and 2 with μ = −1.
n360049006400810010,000
Example 1 Ω = M ^ α e x p 1.331.331.321.311.3
IT5453535353
CPU0.15500.16540.21360.25690.3517
RES 8.60 × 10 7 9.65 × 10 7 9.27 × 10 7 8.92 × 10 7 8.57 × 10 7
Ω = D A α ˜ e x p 4.024.824.604.464.34
IT2928282828
CPU0.01240.01670.02180.02840.0367
RES 9.99 × 10 7 9.99 × 10 7 9.99 × 10 7 9.96 × 10 7 9.99 × 10 7
Example 2 Ω = M ^ α e x p 1.11.081.081.081.08
IT9798969595
CPU0.20300.36160.51890.58630.7690
RES 9.67 × 10 7 9.14 × 10 7 9.88 × 10 7 8.81 × 10 7 9.01 × 10 7
Ω = D A α ˜ e x p 43.4239.9638.2339.6041.50
IT7473727170
CPU0.03140.04450.05890.07790.0968
RES 1.00 × 10 6 1.00 × 10 6 1.00 × 10 6 1.00 × 10 6 1.00 × 10 6
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chen, R.; Peng, X.; Yu, W. An Improved Convergence Theorem of the Newton-Based AOR Method for Generalized Absolute Value Equations. Symmetry 2022, 14, 1249. https://doi.org/10.3390/sym14061249

AMA Style

Chen R, Peng X, Yu W. An Improved Convergence Theorem of the Newton-Based AOR Method for Generalized Absolute Value Equations. Symmetry. 2022; 14(6):1249. https://doi.org/10.3390/sym14061249

Chicago/Turabian Style

Chen, Raojie, Xiaofei Peng, and Wensong Yu. 2022. "An Improved Convergence Theorem of the Newton-Based AOR Method for Generalized Absolute Value Equations" Symmetry 14, no. 6: 1249. https://doi.org/10.3390/sym14061249

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop