Next Article in Journal
An Aggregation Rule Based on the Binomial Distribution
Previous Article in Journal
A Fully Coupled Thermomechanical Phase Field Method for Modeling Cracks with Frictional Contact
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Pattern-Multiplicative Average of Nonnegative Matrices: When a Constrained Minimization Problem Requires Versatile Optimization Tools

by
Vladimir Yu. Protasov
1,2,3,
Tatyana I. Zaitseva
3,4 and
Dmitrii O. Logofet
5,*
1
Faculty DISIM, University of L’Aquila, 67100 L’Aquila, Italy
2
Faculty of Computer Science of National Research, University Higher School of Economics, 109028 Moscow, Russia
3
Department of Mechanics and Mathematics, Moscow State University, 119992 Moscow, Russia
4
Moscow Center for Fundamental and Applied Mathematics, 119992 Moscow, Russia
5
Laboratory of Mathematical Ecology, A.M. Obukhov Institute of Atmospheric Physics, Russian Academy of Sciences, 119017 Moscow, Russia
*
Author to whom correspondence should be addressed.
Mathematics 2022, 10(23), 4417; https://doi.org/10.3390/math10234417
Submission received: 22 October 2022 / Revised: 10 November 2022 / Accepted: 13 November 2022 / Published: 23 November 2022
(This article belongs to the Section Mathematical Biology)

Abstract

:
Given several nonnegative matrices with a single pattern of allocation among their zero/nonzero elements, the average matrix should have the same pattern as well. This is the first tenet of the pattern-multiplicative average (PMA) concept, while the second one suggests the multiplicative nature of averaging. The concept of PMA was motivated in a number of application fields, of which we consider the matrix population models and illustrate solving the PMA problem with several sets of model matrices calibrated in particular botanic case studies. The patterns of those matrices are typically nontrivial (they contain both zero and nonzero elements), the PMA problem thus having no exact solution for a fundamental reason (an overdetermined system of algebraic equations). Therefore, searching for the approximate solution reduces to a constrained minimization problem for the approximation error, the loss function in optimization terms. We consider two alternative types of the loss function and present a general algorithm of searching the optimal solution: basin-hopping global search, then local descents by the method of conjugate gradients or that of penalty functions. Theoretical disadvantages and practical limitations of both loss functions are discussed and illustrated with a number of practical examples.

1. Introduction

The concept of pattern-multiplicative average (PMA), or pattern-geometric mean, was proposed with regard to matrix population models (MPMs) for the dynamics of discrete-structured populations [1] in order to summarize the outcome of monitoring a local population of a biological species over several years and to calculate the ensuing measure for assessing the population viability.

1.1. Matrix Population Models

The MPM is represented by a system of difference equations,
x(t + 1) = L(t) x(t), t = 0, 1, 2, …
for the vector of population structure, x(t) ∈ + n , with a nonnegative n × n matrix, L(t), called the population projection matrix (PPM) [1]. Each component of x(t) is the (absolute or relative) number of individuals in the corresponding status-specific group at the observation year t, while the elements of L, called vital rates (ibidem), bear information about the rates of demographic processes in the population.
The pattern of matrix L shows the allocation of zero/nonzero elements in the matrix. It corresponds to the associated directed graph [2] called the life cycle graph (LCG) [1] as it represents graphically the knowledge of life histories involved into the model in combination with the way the population structure is observed in the field or laboratory (Figure 1 gives an example). If matrix L is positive or L = 0, then its pattern is trivial. When the LCG is strongly connected [2], it signifies a certain integrity of the individuals’ life history and provides for the PPM being irreducible [3,4]. However, the LCG ceases to be strong when it includes post-reproductive stages (a further, more complicated sample follows).
In theoretical layouts and practical applications, it is convenient to consider the PPM as the sum,
L = T + F,
of its parts that are responsible for the transitions (T) between individual statuses and population recruitment (F) [1,6]. In particular, if we number the graph nodes in Figure 1 from left to right, then we have
L = T + F =   [ 0 0 0 0 0 d e 0 0 f h 0 0 0 0 0 0 0 k 0 0 l 0 m 0 ] + [ 0 0 0 0 a 0 0 0 0 0 0 0 b 0 c 0 0 0 0 0 0 0 0 0 0 ]
associated with the LCG shown in Figure 1. Matrix T is always column substochastic, i.e., its column sums do not exceed 1 according to their biological sense.
“When calibrated reliably from data, matrix L(t) gives rise to a rich repertoire of qualitative properties and quantitative indices characterizing the population under study at the place where and the time when the data were mined” ([7], p. 2/15). In particular, the dominant eigenvalue, λ1(L) > 0, of matrix L, which coincides with ρ(L) (the spectral radius of L) and exists by the classical Perron–Frobenius theorem for nonnegative matrices [3,8,9], “gives a quantitative measure of how the local population is adapted to its environment [9], thus serving as an efficient tool of comparative demography [1] and enabling a forecast of population viability. This ability ensues from the dynamics of trajectory x(t) as t tends to +∞ when (a primitive [4]) L(t) = L does not change with time” ([7], p. 2/15). In formal terms, we have
x ( t )   ~   λ 1 t   x * ,     x ( 0 )     + n ,
where x* is a positive eigenvector corresponding to λ1, with a norm depending on x(0) [1]. Thus,
x ( t )     { 0 , if   λ 1 < 1 ; x * , if   λ 1 = 1 ; , if   λ 1 > 1 ,
and the location of λ1 relative to 1 may serve as a forecast of population viability if we believe that the vital rates do not change with t.
In practice, however, they do change quantitatively (yet retaining their single pattern) with time when t > 2. Each pair of consecutive years of observation generate a particular annual PPM L(t) obeying Equation (1) [1,5,10], and we have a finite set, L(0), L(1), …, L(M − 1), of M annual PPMs as a result of M + 1 observation years. Different PPMs generate different or even controversial (Table 3 in [9]) forecasts of population viability, and these motivate a task to average the set of PPMs in order to summarize all the years of observation.

1.2. Pattern-Multiplicative Average of Several PPMs

The logic behind averaging may be implicit or explicit and vary in complexity from the ordinariness of the arithmetic mean [11,12] through the weighted mean values of matrix elements [13] to the PMA concept [14] explained below.
Given a vector x(0) at the initial year of observation and a vector x(M) at the final year, it follows from (1) that
x(M) = L(M − 1) L(M − 2) … L(1) L(0) x(0),
i.e., the product of M annual PPMs (in the chronological order) transforms x(0) exactly to x(M). The logic of PMA suggests that the average matrix G should do exactly the same when raised to the Mth power. This is supposedly true for any observed vectors, whereby we conclude that
GM = L(M − 1) L(M − 2) … L(1) L(0).
One more natural constraint on G consists in the pattern of G coinciding with that of the PPMs to be averaged.
Definition 1.
Let L(0), L(1), …, L(M − 1) be nonnegative square matrices with a single nontrivial pattern. Matrix G = G{L(0), L(1), …, L(M − 1)} is called the pattern-multiplicative average (or pattern-geometric average) of the M given matrices if it has the same pattern and obeys the averaging Equation (7).
Note that the vector-matrix Equation (7) with an n × n matrix G is equivalent to a system of n2 scalar algebraic equations, and the question is how many unknown elements of G the system has to contain. When the pattern of G is trivially complete, i.e., when all the nonnegative matrices L(t) are actually positive, the number of unknowns equals n2 as well, so that Equation (7) may have a nontrivial exact solution (Table 3 in [15]). However, when the pattern of L(t) is nontrivial, matching a real LCG, the number (k) of unknown positive entries in G is less than n2, so that system (7) becomes an overdetermined system of algebraic equations (see, e.g., [16]). There is no reason to consider the remaining (n2k) equations as a linear combination of the former k ones, so that the overdetermined system (7) is inconsistent and has no exact solutions.
The task to average M given PPMs can therefore be accomplished as an approximate solution to system (7), with a logical requirement to minimize the approximation error, which is measured in some reasonable way.
Definition 2.
Let L(0), L(1), …, L(M − 1) be nonnegative square matrices with a single nontrivial pattern. Matrix G = G{L(0), L(1), …, L(M − 1)} is called an approximate pattern-multiplicative average (APMA) of the M given matrices if it has the same pattern and represents a solution to a constrained minimization problem for the approximation error in solving system (7).
In what follows, we use the same notation G for the APMA matrix.

1.3. Minimization Problem for the Approximation Error

Equation (7) is obviously equivalent to
GML(M − 1)L(M − 2) L(1) L (0) = 0,
so that a norm of the left-hand side can be considered as the approximation error when G is an approximate solution rather than the exact one. When the matrix size (n × n) and the number of cofactors (M) are not too high, e.g., n = 5, M = 7 in Figure 1 and Equation (3), the APMA problem can be solved numerically by means of a modern software system such as MATLAB® with an acceptable accuracy (Tables 3 and 4 and Appendix A in [5]).
However, the standard MATLAB tools return very rough estimates for higher dimensions n. This is seen, for example, in the sample calculations with n = 11 and M = 12. This phenomenon is quite natural since the error function is highly non-convex, in which case the optimization problem is notoriously hard. The number of local minima can grow exponentially with n. Therefore, the standard computer optimization software suffers when n is relatively large. In this case, one needs to choose special mathematical tools suitable for a particular problem. We find those tools as combinations of modern algorithms for local and global optimization.

2. Objects and Methods

2.1. A. albana Annual PPMs

Note that annual PPMs (3) contain 10 nontrivial elements a, b, …, l, m (called vital rates [1]), and if we consider them as a row vector, then the 12 annual PPMs of A. albana can be presented as a single table (Table 1). Those rates take on the form of rational numbers according to the way they were mined in the field [5].
We consider solving the APMA problem for the first 7 PPMs (2009–2015) as Example 1 and for all the 12 ones as Example 4. Example 2 will refer to the seven matrices modified in some elements as marked in Table 2 with bold.

2.2. Calamagrostis epigeios PPM and Its Perturbations

C. epigeios is a long-rhizome perennial graminoid actively colonizing open areas in the temperate zone [17]. Its biology is well studied, and the life cycle graph was constructed in terms of both ontogenetic stage and chronological age (Figure 2). In contrast to Figure 1, the C. epigeios digraph is not strongly connected, but it contains what was called the reproductive core of the LCG, the maximal connected subgraph [13].
The construction was based on the data gained by excavating a whole local colony of the plants [17]. Those data enabled calibrating a single PPM alone. The entries of the transition part were determined in a unique way as rational numbers, while those of the 5 × 5 reproductive-core submatrix remained uncertain:
L = [ a / 43 b / 7 0 c / 12 d / 3 0 e / 3 0 0 0 0 23 / 43 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 f / 43 0 0 0 g / 3 0 h / 3 0 0 0 0 10 / 43 0 0 7 / 12 0 0 0 0 0 0 0 k / 43 0 0 0 0 0 0 0 0 0 0 2 / 43 0 0 1 / 12 0 0 0 0 0 0 0 0 0 0 0 0 0 2 / 3 0 0 0 0 1 / 43 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 2 / 7 0 0 2 / 3 0 0 0 0 0 0 ] ,  
restricted with certain linear equalities (“reproductive uncertainty” [17], p. 387):
{ a + b + c + d + e = 163 , f + g + h = 22 ,
and certain linear inequalities:
a + f + kc + gd + hbe,
ensued from the expert knowledge [17]. Here, the 9 reproduction rates a, b,…, h, k ≥ 0 take on integer values only in accordance with empirical data [17]: after the Colony had been excavated and all the mother-daughter rhizome links counted, the 9 uncertain parameters became certain (ibidem, Table 3) and resulted in λ1(L) = 3.4266 (ibidem) signifying rapid growth of the young Colony.
To illustrate the concept of PMA we have to obtain a finite set of such matrices, and we have obtained it by means of perturbing the nine rates artificially to the values presented in Table 3 (“years” 2 to 9) together with the corresponding values of λ1(L).

2.3. Mathematical Problem Formulation

We solve the following APMA problem:
{ f ( X ) = X M L M L 1 F m i n X , m i n k + L k ( i , j ) X ( i , j ) m a x k L k ( i , j ) ,    i , j = 1 , , n ; m i n + ( x , y ) = { m i n ( x , y )   i f   x · y 0 , m a x ( x , y )   i f   x · y = 0 ,
where the n × n matrix X has a prescribed pattern (the allocation of nonzero elements) and the expression
X F = x 11 2 + + x n n 2
denotes the Frobenius norm of the matrix. Additional linear constraints on the elements of X (both equalities and inequalities) can also be incorporated, such as conditions (10)–(11) and those of the transition part in (3) being substochastic.
Instead of minimizing the “loss function” (in the terminology of optimization theory) f(X) = X M L M L M 1 L 2 L 1 F , we may consider minimizing the loss function Φ(X) = f2(X) that is calculated as the sum of all the squared elements in the difference of two matrices. The minimum has to be reached at the same matrix X.

2.4. Method 1 to Solve the APMA Problem

We use the basin-hopping method of global optimization suggested in [18,19,20]. This is an iterative stochastic method reduced to repeating three key steps. First, the coordinates are randomly perturbed, and a method of local optimization is then launched from the perturbed starting point. Finally, based on the obtained minimum, it is determined whether the coordinate perturbation has turned out to be successful and has to be saved. The base version of the basin-hopping method uses the Metropolis criterion algorithm [21], although modifications are also possible.
To find a local minimum, we use the trust-region method (TRM) suggested in [22,23] as the algorithm of local optimization. It first determines a trust region around the best current solution, and a quadratic model of approximation is then used within this region. The region size is updated at each step, i.e., the size increases if the approximation of the objective function is rather good; otherwise, it decreases. Trust-region methods are quite popular in solving various nonlinear optimization problems due to their applicability to the ill-conditioned ones and to their convergence and robustness properties; see [24] and references therein.
TRM methods support both the equality- and inequality-type constraints, and they guarantee the constraints to hold after optimization (in contrast to other local methods such as SLSQP or L-BFGS [25]). However, the TRM may sometimes be slower, and the alternative methods of local optimization could also be useful in some cases.

2.5. Method 2 to Solve the APMA Problem

The second approach originates from the class of penalty methods. In these methods, an additional term is introduced as a measure of constraint violations. In our case, we consider the column substochastic constraints and solve the following problem:
{ X M L M L 1 2 2 + ε ( X ) m i n X ; m i n k + L k ( i , j ) X ( i , j ) m a x k L k ( i , j ) ,    i , j = 1 , , n ,
where ε(X) consists of several summands, each one corresponding to one of the inequalities. If an inequality is violated, the corresponding summand equals the squared magnitude of violation multiplied with a constant. For example, when the inequality is X(0, 0) + X(0, 1) ≤ 1, the summand equals C (X(0, 0) + X(0, 1) − 1) 2 if X(0, 0) + X(0, 1) > 1.
Constant C should be chosen manually as its value depends on the current state of the algorithm. It is a popular strategy to change the value of C in the course of the algorithm because smaller Cs are better at the beginning phase, as they accelerate the convergence of the gradient descent. However, larger values of C are desirable in the end of the algorithm since they guarantee the feasibility of constraints.
We have seen in experiments that it is better not to put any box constraints in the penalty summands, but rather to diversify them as the bounds for the variables.
The basin-hopping algorithm described above (Section 2.4) can be used as a global optimization method. If we do not have any additional constraints of equality type, then the choice of the underlying local methods becomes wider. For example, it is possible to use the L-BFGS algorithm [26], which is faster than the TRM. If, however, we do have additional equality constraints, the TRM shall still be used.
We suggest another problem setting in the next section, where we minimize another loss function, the same methods still being applicable there. Both Method 1 and Method 2 work more stably when the derivatives of their loss functions are given explicitly. They are calculated for the both loss functions in the next Section.

3. New Loss Function of Approximation

Introduced in Section 2.3 as a measure of approximation error, the loss function Φ(X) has several disadvantages. First, it measures the absolute error and changes with the normalization of matrices. A suitable normalization can be done by the division by the norm of the matrix, or by the spectral radius product of the product LML1. However, normalization cannot exclude other shortcomings, either. Some of them are explained below.
(1)
Loss function Φ behaves differently in the approximations from above and from below. This phenomenon already occurs in the dimension n = 1. In the one-dimensional case, we denote the “matrices” (numbers) with small letters. Let x be an approximation of the multiplicative mean of numbers lM, …, l1. Denote q = (lMl1)1/M; then Φ(x) = |xMqM|2. Normalize all the lj such that q = 1 and note a key difference between the cases of x < 1 and x > 1, namely,
Φ ( x ) {   as   M   if   x > 1 ,   < 1   for   all   M if   x < 1 .
Let, for example, M = 10. When x = 1.1, we have Φ(x) = |x10 − 1|2 = 2.54. On the other hand, for x = 0.1, we have Φ(x) = |0.110 − 1|2 = 0.9999…, hence the loss function Φ has to choose the approximation x = 0.1 in spite of it being much worse.
(2)
The value of Φ may increase with M and even tend to infinity. In the example above, if x = 1.1, we have Φ(x) → ∞ as M → ∞. In real practice, however, the situation should be the opposite: the greater M, the sharper will be the approximation to the average value.
(3)
When ρ(X), the spectral radius of X is small, and the value of Φ can be very small and unable to distinguish the quality of approximation. In the example above, when M = 100, the approximations x = 0.5 and x = 0.1 are practically not comparable because Φ(x) < 10−30 in both cases.
The following idea helps avoid all of these shortcomings. In the case of n = 1, if q denotes the actual value of the multiplicative mean and x our approximation, then
| x q | 2 = | x M q M | 2 | q M 1 + q M 2 x + + q x M 2 + x M 1 | 2
Thus, the numerator Φ(x) is divided by a specially chosen denominator. If q is unknown, the powers qk can be replaced by the products lMlMk+1. The same principle can be applied in our problem, when we deal with matrices instead of numbers. We begin with lM since this is chronologically the last measurement and it should give the main contribution to measuring the quality of approximation when the application context suggests such a priority. Translating this idea to the general dimension n, we replace Φ with the new loss function S(X) = s(X)2:
S ( X ) = Φ ( X ) Ψ ( X ) ,
where Ψ is defined as
Ψ ( X ) = j = 0 M 1 L M L j + 2 2
We derive the derivative of Ψ by applying Proposition 1 (see Appendix A):
S ( X ) = Ψ Φ Φ Ψ Ψ 2 ,
where Φ = Φ(X), Ψ = Ψ(X) are defined in (A.1) and the derivatives Φ’, Ψ ’ are derived by the formulas from Proposition 1. Note that Φ’, Ψ ’ are n × n matrices, while Φ and Ψ are nonnegative numbers, Ψ ≠ 0. That is why the division in (18) is well defined and Ψ commutes with Φ’.

4. Results

In this section, we report on the numerical results obtained by the methods presented in Section 2.4, Section 2.5 and Section 3 and applied to the Examples listed in Section 2.1, Section 2.2 and Section 2.3. We consider settings with both the loss function Φ(X) = f(X)2 (12) and S(X) = s(X)2 (16), both being calculated in each example for the purpose of comparison (Table 4, Table 5, Table 6 and Table 7).
Technical parameters are as follows: in Method 1, we have used 20 iterations of the basin-hopping algorithm; in Method 2, the first 80 iterations have been made with C = 40, and the next 20 with C = 2040.
Next, we demonstrate the numerical results for Example 2, an artificial example to analyze the sensitivity of our method to the substochasticity constraint. The numerical experiments (see Table 5) show that the results may change after imposing that the solution must be a column substochastic matrix.
The results for Example 3 turn out to be the same and independent of the linear inequalities. Using Method 2 makes no sense in this case as the linear equalities (10) suggest using the TRM. Note that without linear inequalities, Method 2 is the same as Method 1.

5. Discussion

To realize our goal of finding the multiplicative mean of several PPMs, we note first that this problem may have an exact solution. For example, if the matrix product is diagonalizable, then the multiplicative mean is a diagonal matrix too in the appropriate basis, with the elements being the complex roots of the corresponding power from the diagonal elements of the original matrix product. The solution is not unique because of the non-uniqueness of the complex roots. Taking any of the solutions and transferring back to the original basis, we get a multiplicative mean. This mean, however, may not be a nonnegative matrix. Even if it is, it may not satisfy our imposed constraints. This argument, in addition to that one in Section 1.2 concerning the lack of exact solutions to Equation (7), substantiates the approximation approach to the problem of pattern-multiplicative average.
We search for a matrix satisfying all the constrains and being close to the unknown pattern-multiplicative average, G, by means of some loss function. This leads to optimization problem (13), where the loss function is the Frobenius norm. The Frobenius norm is the most popular one in matrix analysis since it is actually the Euclidean norm in the space of matrices equipped with the standard scalar product. However, this choice of the distance has several disadvantages, as noted in Section 3. We introduce a modified loss function and obtain another optimization problem. Both problems are not easy to solve as they are highly non-convex, and hence may possess many local extrema, and because of a large number of variables. Non-convexity implies all the known optimization methods to converge only to local extrema. Therefore, we use the following strategy: first we get a rough approximation to the optimum by using the global optimization tools, then we improve the solution by searching for a close point of local minimum (hoping that this local minimum is close to the global one). Among all the global optimization methods we choose the basing-hopping method since it successfully approximates the optimal value for problems with linear constraints even in relatively high dimensions. The next step, i.e., improving the solution with a close local minimum, can be realized by various popular algorithms: the Newton method, gradient methods, the Frank–Wolf algorithm, the trust region method (TRM), etc. Most of them need a relatively small Lipschitz constant for the gradient, which is not always the case (see the explicit formulas of the gradient in Appendix A). That is why the TRM is the most efficient algorithm for our problem.
Despite theoretical disadvantages of the loss function Φ noted in Section 3, our practical calculations have revealed only minor improvements that the alternative S brings with any method of the optimal search.
It is easy to see that the arithmetic mean of all the nine perturbed λ1s (Table 3) equals 3.4034. Neither the arithmetic mean of the annual λ1s, nor λ1 of the arithmetic average of the annual PPMs can even serve as a crude estimate of the λ1(G).
Note that we do not pretend to exhaust here the theme of “versatile optimization tools”. For example, a penalty method that can reduce a non-convex problem to an infinite-dimensional optimal control problem [27] or the generalized subdiffenential technique based on the sequential optimality conditions [28] look fairly perspective, too. Some non-convex optimization problems related to the matrix calculus also appear naturally in the approximation problems, for example, the barycentric rational polynomial approximation [29] and the generalized weak greedy approximation [30].
Coming back to the logic of PMA, consider the equation
GM = ProdL
equivalent to (8) if ProdL denotes the product of M matrices in its left-hand side. Since λ1(G M) = λ1(G )M, it follows from (19) that the exact value of λ1(G),
λ1(G) = [λ1(ProdL)]1/M,
is quite calculable from the data and equal to 3.4017 in Example 3 (M = 9), to 0.893087 in Example 4 (M = 12). It should not be surprising therefore that the corresponding optimized values of ρ(X) = 3.41... (Table 6) and = 0.88… (Table 7) are close to the exact ones above. Not surprising would also be an alternative formulation of the APMA problem suggesting the difference between λ1s as the loss function. Optimization along this line or even that in combination with the former loss function is worthy of further research efforts.
Table 6. Results for Example 3.
Table 6. Results for Example 3.
Method f ( X ^ ) s ( X ^ ) ρ ( X ^ ) X ^
Method 1 with Φ(X)8244.6215520.02923.41182.60260.490103.41041.322900.9206
0.5349000000
0000000
0.25580003.333300.3333
0.2326000.5833000
0.0233000000
0.0465000.0833000
Method 1 with S(X)8257.1675550.0291573.41072.59530.489103.42871.3500.928
0.5349000000
0000000
0.25580003.333200.3334
0.2326000.5833000
0.0237000000
0.0465000.0833000
Table 7. Results for Example 4.
Table 7. Results for Example 4.
Method f ( X ^ ) s ( X ^ ) ρ ( X ^ ) X ^
Method 1 with Φ(X)0.0458980.0071490.887600002.7782
0. 15080003.0007
0.11250.3106000.8163
000.31050.74530
0000.15150
Method 2 with Φ(X)0.051560.0082830.886300003.3354
0.31080003
0.05910.31650.147400.6754
000.25770.76820
0000.12570
Method 1 with S(X)0.0461950.007080.88800002.6526
0.00410003.3033
0.14360.28180.000100.8541
000.31930.73780
0000.15880

6. Conclusions

We solve the problem of finding a matrix of a special pattern which is close, in a certain sense, to the multiplicative mean of several given matrices. The closeness is measured by a loss function, which is usually non-convex and may have a large Lipschitz constant of the derivative. This makes the problem of finding the optimal matrix, i.e., of minimizing the loss function, very difficult. We have elaborated a two-phase method for this problem. The first phase realizes a rough search of the absolute minimum, and this is done by the basing-hopping method of global optimization. The second phase provides for an improved solution at a point of local minimum in a vicinity of our global solution, and this is done by the trust region method. In numerical examples, we demonstrate the efficiency of our strategy even for relatively high dimensions. Implementation of these ideas in the problem of pattern-multiplicative average is not direct and requires the overcoming of some technical obstacles, which have nevertheless been bypassed in this study. While the traditional approach to global optimization reduces to searching for the best solution among the local ones found preliminary, the nontraditional inverted logic may be successful when the direct one fails.

Author Contributions

Conceptualization, V.Y.P. and D.O.L.; methodology, V.Y.P. and D.O.L.; software, D.O.L. and T.I.Z.; validation, V.Y.P. and D.O.L.; formal analysis, V.Y.P.; investigation, D.O.L.; resources, V.Y.P. and D.O.L.; data curation, D.O.L. and T.I.Z.; writing—original draft preparation, D.O.L.; visualization, D.O.L.; supervision, V.Y.P.; project administration, D.O.L.; funding acquisition, D.O.L. All authors have read and agreed to the published version of the manuscript.

Funding

DL was supported by the Russian Scientific Foundation, grant number [22–24–00628]. The work of VP was done within the framework of the HSE University Basic Research Program and funded by the Russian Academic Excellence Project ‘5-100′. The work of TZ was supported by the Foundation for Advancement of Theoretical Physics and Mathematics “BASIS”.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Acknowledgments

Programming and calculations were implemented in Python, 3.10.7. Comments and suggestions by wo anonymous reviewers have helped us improve the manuscript quality.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Derivatives of matrix functions.
We consider two functions Φ and Ψ defined on Mn, the set of real n × n matrices, as follows:
Φ ( X ) = X M L M L M 1 L 2 L 1 F 2 ,   Ψ ( X ) = j = 0 M 1 L M L j + 2 2
(in the latter sum, the term with j = M − 1 is equal to XM−1). The norm is Frobenius and defined by the scalar product (X, Y) = tr(XT Y). We need to find the derivatives of Φ and Ψ in the space Mn. The derivative is understood in the standard way as a linear functional on Mn that can be identified with an element of Mn, i.e., with a matrix A = Φ′(X) ∈ Mn such that
Φ(X + H) = Φ(X) + (A, H) + o(H),
the same being true for the function Ψ, too.
Let
P = P ( X ) = X M L M L 1 ,   Q = Q ( X ) = j = 0 M 1 L M L j + 2 X j
Proposition 1.
The derivatives of Φ and Ψ at an arbitrary point XMn are
Φ ( X ) = 2 j = 0 M 1 ( X T ) j P ( X T ) M j 1 ,
Ψ ( X ) = 2 j = 0 M 1 ( X T ) k L j + 2 T L M T Q ( X T ) j k 1 .
Proof. 
To prove (A3), we write down the expression for Φ(X + H) − Φ(X) (refusing to bold to simplify lay-outs) and collect the linear terms:
( X + H ) M = X M + j = 0 M 2 X j H X M j 1 + O ( H 2 ) .
Therefore,
Φ ( X + H ) = = ( X + H ) M L M L 1 2 = tr [ ( X + H ) M L M L 1 ] T · [ ( X + H ) M L M L 1 ] = = tr [ P ( X ) + j = 0 M 1 X j H X M j 1 + O ( H 2 ) ] T [ P ( X ) + j = 0 M 1 X j H X M j 1 + O ( H 2 ) ] = = tr P T P + 2 tr P T [ j = 0 M 1 X j H X M j 1 + O ( H 2 ) ]
(taking into account that the matrix and its transpose have the same trace).
Since tr PTP = Φ(X), it follows that
Φ   ( X + H ) Φ   ( X ) = 2 j = 0 M 1 tr P T X j H X M j 1 + O ( H 2 ) .
Thus, the linear part of the variation in Φ at the point X is equal to
2 j = 0 M 1 tr P T X j H X M j 1 .
This is a linear functional in H, albeit, to find the derivative A = Φ′(X), we have to write it in the scalar product form (A, H). Using the trace invariance with respect to cyclic permutations of the matrix product, we obtain
tr P T X j H X M j 1 = tr X M j 1 P T X j H = tr [ ( X T ) j P ( X T ) M j 1 ] T H = ( ( X T ) j P ( X T ) M j 1 , H ) .
Therefore, the linear part of the variation in Φ is equal to
( 2 j = 0 M 1 ( X T ) j P ( X T ) M j 1 ) P ,
which proves (A3).
To prove (A4), we write down the increment of Ψ as
Ψ ( X + H ) Ψ ( X ) = 2 tr j = 0 M 1 k = 0 j 1 L M L j + 2 X k H X j k 1 Q T P + O ( H ) 2 = 2 tr j = 0 M 1 k = 0 j 1 X j k 1 Q T L M L j + 2 X k H + O ( H ) 2 = 2 tr [ j = 0 M 1 k = 0 j 1 ( X T ) k L j + 2 T L M T Q ( X T ) j k 1 ] T H + O ( H ) 2 =   ( 2 j = 0 M 1 k = 0 j 1 ( X T ) k L j + 2 T L M T Q ( X T ) j k 1 ,   H ) + O ( H ) 2 ,  
which completes the proof of (A4). □

References

  1. Caswell, H. Matrix Population Models: Construction, Analysis and Interpretation, 2nd ed.; Sinauer Associates: Sunderland, MA, USA, 2001. [Google Scholar]
  2. Harary, F.; Norman, R.Z.; Cartwright, D. Structural Models: An Introduction to the Theory of Directed Graphs; John Wiley: New York, NY, USA, 1965. [Google Scholar]
  3. Gantmacher, F.R. Matrix Theory; Chelsea Publ.: New York, NY, USA, 1959. [Google Scholar]
  4. Horn, R.A.; Johnson, C.R. Matrix Analysis; Cambridge University Press: Cambridge, UK, 1990. [Google Scholar]
  5. Logofet, D.O.; Kazantseva, E.S.; Belova, I.N.; Onipchenko, V.G. How long does a short-lived perennial live? A modelling approach. Biol. Bul. Rev. 2018, 8, 406–420. [Google Scholar] [CrossRef]
  6. Cushing, J.M.; Yicang, Z. The net reproductive value and stability in matrix population models. Nat. Res. Model. 1994, 8, 297–333. [Google Scholar] [CrossRef]
  7. Logofet, D.O.; Razzhevaikin, V.N. Potential-growth indicators revisited: Higher generality and wider merit of indication. Mathematics 2021, 9, 1649. [Google Scholar] [CrossRef]
  8. Seneta, E. Non-Negative Matrices; Wiley: New York, NY, USA, 1973. [Google Scholar]
  9. Berman, A.; Plemmons, R.J. Nonnegative Matrices in the Mathematical Sciences; Academic: New York, NY, USA, 1979. [Google Scholar]
  10. Logofet, D.O.; Kazantseva, E.S.; Belova, I.N.; Onipchenko, V.G. Backward prediction confirms the conclusion on local plant population viability. Biol. Bul. Rev. 2021, 11, 406–420. [Google Scholar] [CrossRef]
  11. Klimas, C.A.; Cropper, W.P., Jr.; Kainera, K.A.; de Oliveira Wadt, L.H. Viability of combined timber and non-timber harvests for one species: A Carapa guianensis case study. Ecol. Model. 2012, 246, 147–156. [Google Scholar] [CrossRef]
  12. Logofet, D.O. Projection matrices in variable environments: λ1 in theory and practice. Ecol. Model. 2013, 251, 307–311. [Google Scholar] [CrossRef]
  13. Maslov, A.A.; Logofet, D.O. Joint population dynamics of Vaccinium myrtillus and V. vitis-idaea in the protected postfire Cladina-Vaccinium pine forest. Markov model with averaged transition probabilities. Biol. Bul. Rev. 2021, 11, 406–420. [Google Scholar] [CrossRef]
  14. Logofet, D.O. Averaging the population projection matrices: Heuristics against uncertainty and nonexistence. Ecol. Complex. 2018, 33, 66–74. [Google Scholar] [CrossRef]
  15. Logofet, D.O.; Maslov, A.A. Analyzing the fine-scale dynamics of two dominant species in a Polytrichum–Myrtillus pine forest. II. An inhomogeneous Markov chain and averaged indices. Biol. Bul. Rev. 2019, 9, 62–72. [Google Scholar] [CrossRef]
  16. Anton, H.; Rorres, C. Elementary Linear Algebra, 9th ed.; Wiley: Hoboken, NJ, USA, 2005. [Google Scholar]
  17. Logofet, D.O.; Ulanova, N.G.; Belova, I.N. From uncertainty to an exact number: Developing a method to estimate the fitness of a clonal species with polyvariant ontogeny. Biol. Bull. Rev. 2017, 7, 387–402. [Google Scholar] [CrossRef]
  18. Wales, D.J.; Doye, J.P.K. Global optimization by basin-hopping and the lowest energy structures of Lennard-Jones clusters containing up to 110 atoms. J. Phys. Chem. A 1997, 101, 5111–5116. [Google Scholar] [CrossRef] [Green Version]
  19. Conn, A.R.; Gould, N.I.M.; Toint, P.L. Trust Region Methods; MOS-SIAM Series on Optimization; Society for Industrial and Applied Mathematics: Philadelphia, PA, USA, 2000. [Google Scholar] [CrossRef]
  20. Vinkó, T.; Gelle, K. Basin Hopping Networks of continuous global optimization problems. Cent. Eur. J. Oper Res. 2017, 25, 985–1006. [Google Scholar] [CrossRef] [Green Version]
  21. Robert, C.P.; Casella, G. Monte Carlo Statistical Methods, 2nd ed.; Springer: New York, NY, USA, 2004. [Google Scholar] [CrossRef] [Green Version]
  22. Sorensen, D.C. Newton’s method with a model trust region modification. SIAM J. Numer. Anal. 1982, 19, 409–426. [Google Scholar] [CrossRef] [Green Version]
  23. Yuan, Y. A review of trust region algorithms for optimization. In Proceedings of the 4th International Congress on Industrial & Applied Mathematics (ICIAM 99), Edinburgh, UK, 14 December 2000; pp. 271–282. [Google Scholar]
  24. Sun, W.; Yuan, Y.-X. Optimization Theory and Methods: Nonlinear Programming; Springer: New York, NY, USA, 2006; Available online: https://books.google.ru/books?id=o0BYHLhhPJMC (accessed on 21 October 2022).
  25. Jorge Nocedal, J.; Wright, S.J. Numerical Optimization; Springer: New York, NY, USA, 2006. [Google Scholar]
  26. Liu, D.C.; Nocedal, J. On the limited memory BFGS method for large scale optimization. Math. Program. 1989, 45, 503–528. [Google Scholar] [CrossRef] [Green Version]
  27. Hammoudi, A.; Benharrat, M. An exact penalty method for constrained optimal control problems. Rend. Circ. Mat. Palermo Ser. 2 2021, 70, 275–293. [Google Scholar] [CrossRef]
  28. Moustaid, M.B.; Rikouane, A.; Dali, I.; Laghdir, M. Sequential approximate weak optimality conditions for multiobjective fractional programming problems via sequential calculus rules for the Brøndsted-Rockafellar approximate subdifferential. Rend. Circ. Mat. Palermo Ser. 2 2022, 71, 737–754. [Google Scholar] [CrossRef]
  29. Li, J. Linear barycentric rational collocation method for solving biharmonic equation. Demonstr. Math. 2022, 55, 587–603. [Google Scholar] [CrossRef]
  30. Valiullin, A.R.; Valiullin, A.R.; Solodov, A.P. Sharp sufficient condition for the convergence of greedy expansions with errors in coefficient computation. Demonstr. Math. 2022, 55, 254–264. [Google Scholar] [CrossRef]
Figure 1. Life cycle graph for a local population of Androsace albana, an alpine short-lived perennial species, observed once a year. Ontogenetic stage notations: pl, seedlings; j, juvenile plants; im, immature plants; v, adult vegetative plants; g, generative plants, the stages being distinguishable in the field. Solid arrows indicate transitions occurring for one year (no transition, in particular); dashed arrows correspond to the annual population recruitment [5].
Figure 1. Life cycle graph for a local population of Androsace albana, an alpine short-lived perennial species, observed once a year. Ontogenetic stage notations: pl, seedlings; j, juvenile plants; im, immature plants; v, adult vegetative plants; g, generative plants, the stages being distinguishable in the field. Solid arrows indicate transitions occurring for one year (no transition, in particular); dashed arrows correspond to the annual population recruitment [5].
Mathematics 10 04417 g001
Figure 2. LCG for Calamagrostis epigeios according to the excavation data of a 3-year-old colony in 2015: v denotes the stage of adult vegetative plants; gi (i = 1, 2, 3), generative plans with i generative shoots; ss and s, subsenile and senile stages, respectively. Solid arrows indicate the ontogenetic transitions that have occurred for 1 year: numbers at the arrows indicate proportions of the tufts outgoing from an initial status in 2014 and reaching the ingoing status by 2015; the number inside the vertex is the ordinal number the corresponding component has in the vector of population structure; gray background highlights the reproductive core of the LCG (Figure 4I in [17]).
Figure 2. LCG for Calamagrostis epigeios according to the excavation data of a 3-year-old colony in 2015: v denotes the stage of adult vegetative plants; gi (i = 1, 2, 3), generative plans with i generative shoots; ss and s, subsenile and senile stages, respectively. Solid arrows indicate the ontogenetic transitions that have occurred for 1 year: numbers at the arrows indicate proportions of the tufts outgoing from an initial status in 2014 and reaching the ingoing status by 2015; the number inside the vertex is the ordinal number the corresponding component has in the vector of population structure; gray background highlights the reproductive core of the LCG (Figure 4I in [17]).
Mathematics 10 04417 g002
Table 1. Vital rates of A. albana as the entries of 12 annual PPMs (extracted from Table 3 in [10]).
Table 1. Vital rates of A. albana as the entries of 12 annual PPMs (extracted from Table 3 in [10]).
Yearabcdefhklm
200930/1340/133/138/372/3722/11028/297/9919/351/35
201019/131/10/114/304/3022/4817/5534/5523/261/26
201149/185/125/11/196/1935/4521/4310/4348/577/57
201219/4136/41/41/4910/4945/8639/8728/8745/586/58
201316/698/62/602/1916/13714/956/9564/733/73
20144/319/30/302/162/986/344/3416/504/50
201510/429/40/4 0010/193/105/1017/201/10
20163/28/2002/105/295/138/1320/221/22
201712/123/1003/32/88/122/1221/282/28
201813/238/2001/121/230/131/1322/231/23
20198/29/901/32/32/34/42/419/195/19
20208/29/201/32/32/34/41/1319/195/15
Table 2. Example 2: modified Example 1 for testing the column substochastic property.
Table 2. Example 2: modified Example 1 for testing the column substochastic property.
“Year”abcdefhklm
130/134/133/138/372/3722/11028/297/999/3511/35
219/131/10/114/304/3022/4817/5534/5525/261/26
349/185/125/11/196/1935/4521/4310/4348/577/57
419/4136/41/41/4910/4945/8639/8728/8745/588/58
516/698/62/602/1916/13714/956/9564/733/73
64/319/30/302/162/986/344/3416/5026/50
710/429/40/4 0010/193/105/1017/201/10
Table 3. Example 3: Nonzero elements of the 9 Calamagrostis epigeios PPMs to be averaged.
Table 3. Example 3: Nonzero elements of the 9 Calamagrostis epigeios PPMs to be averaged.
“Year”abcdefghkλ1
111733832129113.4266
210025452128223.3093
312033622147153.4198
4125430221010223.4707
5117141311010213.4124
611923732137233.4050
712333412119213.4281
810654543129123.3740
9111345311010213.3847
Table 4. Results for Example 1.
Table 4. Results for Example 1.
Method f ( X ^ ) s ( X ^ ) ρ ( X ^ ) X ^
Method 1 with Φ(X)0.021090.0023740.858500003.3309
0.4530007.8767
0.02880.29360.147400
000.17260.75890
0000.10340
Method 2 with Φ(X)0.0210890.0023740.858500003.3309
0.45330007.8757
0.02870.229360.147400
000.17260.75890
0000.10340
Method 1 with S(X)0.0211760.0023790.858400003.3348
0.43220007.9666
0.03630.28970.148500.0022
000.17280.75870
0000.10340
Table 5. Results for Example 2.
Table 5. Results for Example 2.
Method f ( X ^ ) s ( X ^ ) ρ ( X ^ ) X ^
Method 1 with Φ(X)2.8657250.016890.8952000029.9759
000085
0.13830.02040.147400.1929
000.10950.320
0000.520
Method 2 with Φ(X)2.8657250.0168900.8952000029.9759
000085
0.13830.02040.147400.1929
000.10950.320
0000.520
Method 1 with S(X)2.8744380.0168510.8968000030.7974
0.000200084.9995
0.10330.03440.147400.2694
000.10530.32010
0000.520
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Protasov, V.Y.; Zaitseva, T.I.; Logofet, D.O. Pattern-Multiplicative Average of Nonnegative Matrices: When a Constrained Minimization Problem Requires Versatile Optimization Tools. Mathematics 2022, 10, 4417. https://doi.org/10.3390/math10234417

AMA Style

Protasov VY, Zaitseva TI, Logofet DO. Pattern-Multiplicative Average of Nonnegative Matrices: When a Constrained Minimization Problem Requires Versatile Optimization Tools. Mathematics. 2022; 10(23):4417. https://doi.org/10.3390/math10234417

Chicago/Turabian Style

Protasov, Vladimir Yu., Tatyana I. Zaitseva, and Dmitrii O. Logofet. 2022. "Pattern-Multiplicative Average of Nonnegative Matrices: When a Constrained Minimization Problem Requires Versatile Optimization Tools" Mathematics 10, no. 23: 4417. https://doi.org/10.3390/math10234417

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop