Next Article in Journal
One-Step Clustering with Adaptively Local Kernels and a Neighborhood Kernel
Next Article in Special Issue
Approximation of Brownian Motion on Simple Graphs
Previous Article in Journal
A Quantum-Resistant Blockchain System: A Comparative Analysis
Previous Article in Special Issue
Superconvergent Nyström Method Based on Spline Quasi-Interpolants for Nonlinear Urysohn Integral Equations
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Algebraic Solution of Tropical Best Approximation Problems

Faculty of Mathematics and Mechanics, St. Petersburg State University, Universitetskaya Emb. 7/9, 199034 St. Petersburg, Russia
Mathematics 2023, 11(18), 3949; https://doi.org/10.3390/math11183949
Submission received: 14 August 2023 / Revised: 9 September 2023 / Accepted: 14 September 2023 / Published: 17 September 2023
(This article belongs to the Special Issue Approximation Theory and Applications)

Abstract

:
We introduce new discrete best approximation problems, formulated and solved in the framework of tropical algebra, which deals with semirings and semifields with idempotent addition. Given a set of samples, each consisting of the input and output of an unknown function defined on an idempotent semifield, the problem is to find a best approximation of the function, by tropical Puiseux polynomial and rational functions. A new solution approach is proposed, which involves the reduction of the problem of polynomial approximation to the best approximate solution of a tropical linear vector equation with an unknown vector on one side (a one-sided equation). We derive a best approximate solution to the one-sided equation, and we evaluate the inherent approximation error in a direct analytical form. Furthermore, we reduce the rational approximation problem to the best approximate solution of an equation with unknown vectors on both sides (a two-sided equation). A best approximate solution to the two-sided equation is obtained in numerical form, by using an iterative alternating algorithm. To illustrate the new technique developed, we solve example approximation problems in terms of a real semifield, where addition is defined as maximum and multiplication as arithmetic addition (max-plus algebra), which corresponds to the best Chebyshev approximation by piecewise linear functions.

1. Introduction

The discrete approximation problem—where an unknown real function y = f ( x ) is approximated, given a set of samples ( x i , y i ) of function values y i = f ( x i ) at some points x i —is of primary interest in data analysis. Let F( x ; θ ) be an approximating function that depends on the vector θ of the unknown parameters to be determined. Consider a best approximate solution to the problem, which is defined in the sense of a distance function, d, used for estimating the approximation error, to find
θ * = arg min θ max i d ( F ( x i ; θ ) , y i ) .
The minimax optimization problem at (1) is commonly formulated in terms of the distance between two reals a and b, given by the absolute value d ( a , b ) = | a b | , which leads to the Chebyshev metric to measure the error. The problem is then referred to as the best Chebyshev approximation problem [1,2,3], and it has drawn much attention over the centuries since Laplace’s classical work [4] (Book 3, Chapter V, §39).
In many Chebyshev approximation problems, the approximating function F ( x ; θ ) is taken as linear, with respect to parameters. These problems appear in the solution of overdetermined systems of linear equations [5,6,7], regression analysis when the distribution of errors has bounded support [6,8,9,10] and in other applications in science and engineering.
Under the above linearity assumption, the problems can be reformulated as linear programs and then numerically solved by computational procedures and techniques available in linear programming. Examples of linear programming and other algorithmic solutions to the best Chebyshev approximation problem can be found in early works [5,6,8,11,12,13,14,15] and in more recent publications [16,17,18,19,20].
If the approximating function F ( x ; θ ) is assumed to be nonlinear in its parameters, the solution of the best Chebyshev approximation problem becomes more difficult, since it involves the minimization of a complicated nonconvex and nondifferentiable function. In this case, the problems are handled by appropriate methods and techniques of nonlinear programming [21], which may use various gradient and gradient-free optimization schemes. As typical solutions, one can consider iterative computational procedures that solve an auxiliary linear [22,23,24,25] or quadratic [26,27,28] programming problem at each iteration.
In this paper, we offer a new treatment of the best approximation problem, where the problem is put in the framework of tropical algebra, to develop new approximation models and solutions. Tropical algebra deals with the theory and methods of semirings and semifields with idempotent addition [29,30,31,32,33,34,35] and is being increasingly applied in a range of fields, from algebraic geometry to decision making. An example of the tropical semifield, which is usually referred to as max-plus algebra, is an extended set of reals, where the addition is defined as maximum and the multiplication as arithmetic addition.
Application of tropical algebra allows one to represent various nonlinear problems as linear ones in the tropical algebra setting, which simplifies the solution of the problems and facilitates the interpretation of results. In many cases, the solution of problems in the framework of tropical algebra can lead to direct analytical results that describe all solutions in an explicit compact vector form.
Models and methods of tropical algebra are used in approximation problems in various settings, including problems of best approximation, with respect to the Hilbert projective metric [36,37] and the Chebyshev metric [34]. The known results mainly concern best approximation of a vector by elements of a tropical linear space (tropical linear regression), defined in the framework of max-plus algebra. The discrete best approximation problems, where given data are to be fitted with tropical polynomial and rational functions, present another promising direction of tropical approximation. These problems have appeared in a few recent works on neural networks and machine learning (see, e.g., [38,39]) and require further development of related theory and solution techniques.
We consider approximation problem (1), reformulated to approximate functions defined on a tropical semifield (a semiring with idempotent addition and invertible multiplication). The approximation error is defined in the form of a generalized metric on the tropical vector space over the semifield. As the approximating function, we take tropical analogues of Puiseux polynomial and rational functions. We observe that in the case of max-plus algebra, the generalized metric coincides with the Chebyshev metric, whereas the Puiseux polynomial and rational functions are real convex and nonconvex piecewise linear functions. This leads to a natural application of the problem under consideration to the best Chebyshev piecewise linear approximation of real functions.
In conventional algebra, the Puiseux polynomials extend the notion of standard polynomials, by allowing rational exponents. Tropical Puiseux polynomials are introduced in the same way, where addition, multiplication and exponentiation are defined in terms of an idempotent semifield. These polynomials arise as both an effective instrument and the key object of analysis in a range of research contexts, from optimization problems in operations research to tropical algebraic geometry [32,35,40,41,42,43]. Thus, the development of approximation techniques using tropical Puiseux polynomial and rational functions can be considered to be of benefit to both tropical algebra and its applications.
The key ingredient of the proposed approximation technique is to turn the best discrete approximation problem into that of solving tropical linear vector equations with an unknown vector on one side of the equation (one-sided equations) or unknown vectors on both sides (two-sided equations). We reduce the approximation by tropical polynomials to finding a best approximate solution of a one-sided vector equation, and the approximation by tropical rational functions to a best approximate solution of a two-sided equation. We handle the one-sided equation by applying the results in [44,45], which offer a direct explicit analytical solution to the problem. To obtain a best approximate solution of the two-sided equation, we use the iterative alternating algorithm proposed in [46].
As the main contribution of the study, we consider the development of a new analytical framework for the best approximation of functions in the tropical algebra setting. We propose a new formulation of the approximation problem, find a strong connection between the approximation problem and tropical linear vector equations and, finally, provide a workable solution to the problem. We demonstrate the applicability of the proposed technique, with examples of best approximation problems formulated in the max-plus algebra setting. Both the problems of approximating a convex function by tropical polynomials and a nonconvex function by tropical rational functions are considered. We obtain the best approximate solutions to the problems and give graphical illustrations.
Among the advantages of the proposed approach over the other solutions to the discrete best approximation problems in the tropical algebra setting is the formulation and solution of the problem in general terms of an arbitrary tropical semifield. This allows one to handle in a unified way approximation problems defined in terms of semifields other than max-plus algebra. Moreover, the new approach uses Puiseux polynomial and rational functions, which considerably extends the allowed classes of approximating functions and, hence, makes the approach more flexible and adaptable to a wider range of practical problems. The proposed solution of the rational approximation problem, which is based on a new efficient alternating algorithm for solving two-sided vector equations, and which seems to have no analogue in the literature, is one of the main advantages of the approach.
The rest of the paper proceeds as follows. In Section 2, we present the basic definitions and notation of tropical algebra, which are used throughout the paper. Section 3 describes the solution of one- and two-sided tropical linear vector equations. In Section 4, we offer our main results, which provide solutions to discrete best approximation problems with tropical polynomial and rational approximating functions. Section 5 includes numerical examples of solutions for approximation problems formulated in terms of max-plus algebra. In Section 6, we give some concluding remarks and discuss further research directions.

2. Definitions, Notation and Preliminary Results

We start with key definitions, notation and preliminary results that describe a formal framework for the development of solutions of approximation problems in the tropical algebra setting in what follows. Further information on tropical (idempotent) algebra and its applications can be found in a range of publications, including [29,30,31,32,33,34,35].

2.1. Idempotent Semifield

Let X be a non-empty set that is equipped with binary operations ⊕ (addition) and ⊗ (multiplication), and contains distinct elements 0 (zero) and 1 (unit). Assume that ( X , , 0 ) is an idempotent commutative monoid, ( X \ { 0 } , , 1 ) is an Abelian group and multiplication ⊗ distributes over addition ⊕. We consider the algebraic system ( X , , , 0 , 1 ) , which is commonly referred to as a tropical (idempotent) semifield.
The semifield has addition, which is idempotent: for each x X the equality x x = x holds, and multiplication that is invertible: for each x 0 , there exists x 1 , such that x x 1 = 1 (here and hereafter the multiplication sign ⊗ is omitted, to save writing). Moreover, the semifield is considered complete, in the sense that the sums of infinite number of terms are well-defined and multiplication distributes over the infinite sums.
The integer powers are defined in terms of multiplication ⊗, to represent repeated multiplication. It is assumed that the equation x p = a has a unique solution x for any a X and integer p > 0 , which provides powers with rational exponents and, thus, turns the idempotent semifield into an algebraically closed (radicable) semifield.
Idempotent addition induces a partial order by the rule: x y if and only if x y = y . This partial order is assumed to extend to a total order, which makes the semifield linearly ordered. Since the semifield is complete, each non-empty set bounded from below has its greatest lower bound with respect to this order.
A typical example of the idempotent semifield under consideration is the real semifield R max , + = ( R { } , max , + , , 0 ) , which is known as max-plus algebra. In this semifield, addition is defined as max, multiplication as +, zero as and unit as 0. The power x y coincides with the arithmetic product x × y defined for any x , y R . The inverse x 1 of any x R corresponds to the usual opposite number x . The order relation, which is associated with idempotent addition, agrees with the standard linear order on R .
Another example is the semifield R max = ( R + , max , × , 0 , 1 ) , known as max-algebra, where R + = { x R | x 0 } . It has the operations = max and = × and the neutral elements 0 = 0 and 1 = 1 . The notations of powers and inverses have the standard interpretation. The order induced by addition is the natural linear order on R + .
In conclusion, we note that the algebraic expressions in what follows are thought of as defined in terms of an idempotent semifield unless otherwise indicated.

2.2. Algebra of Matrices and Vectors

Consider matrices and vectors over the idempotent semifield. The set of matrices of M rows and N columns with entries in X is denoted X M × N . A matrix that has all entries equal to 0 is the zero matrix. If a matrix has no zero entries, it is called regular.
Addition, multiplication and scalar multiplication of matrices follow the standard entrywise rules, where the arithmetic addition and multiplication are replaced by ⊕ and ⊗.
A matrix with a single column (row) is a column (row) vector. All vectors below are column vectors unless transposed. The set of column vectors that consist of N elements is denoted by X N . A vector that has all elements equal to 0 is the zero vector denoted by 0 . If a vector has no zero elements, it is called regular.
For any nonzero column vector x = ( x i ) , the multiplicative conjugate is defined as the row vector x = ( x i ) , with elements x i = x i 1 if x i 0 , and x i = 0 otherwise.

2.3. Tropical Vector Space

Let a 1 , , a N X M be a set of vectors and b = x 1 a 1 x N a N be a linear combination of these vectors with coefficients x 1 , , x N X . The set of linear combinations A = { x 1 a 1 x N a N | x 1 , , x N X } , which is closed under vector addition and scalar multiplication, is a tropical (finite-dimensional) vector space generated by a 1 , , a N .
For any vector a = ( a i ) , the support of a is given by supp ( a ) = { i | a i 0 , 1 i M } . Let us take two nonzero vectors a = ( a i ) and b = ( b i ) , such that supp ( a ) = supp ( b ) , and let us define a distance function as follows:
d ( a , b ) = i supp ( a ) a i b i 1 a i 1 b i = b a a b .
When supp ( a ) supp ( b ) , we set the function to a value that is greater than any x X (an undefined value) and write d ( a , b ) = . If a = b = 0 , then we assume d ( a , b ) = 1 .
Suppose the function d is defined in the framework of the semifield R max , + (max-plus algebra), where 1 = 0 . In this context, it coincides for all a , b R M with the Chebyshev metric that is given in terms of the standard algebra as follows:
d ( a , b ) = max 1 i M | a i b i | = max 1 i M max ( a i b i , b i a i ) .
In the case of the arbitrary idempotent semifield X , we can interpret the distance function d as a generalized metric that takes values in the subset [ 1 , ) of X .

2.4. Tropical Puiseux Polynomials and Rational Functions

A tropical Puiseux monomial in one variable x over X is a power x p with an exponent p P , where P Q is a finite set of rationals. A tropical Puiseux polynomial is defined as a linear combination of monomials x p with given nonzero coefficients θ p X in the form
P ( x ) = p P θ p x p , x 0 .
Since all coefficients satisfy the condition θ p > 0 and, hence, are tropically positive, the function P can also be considered as a tropical posynomial.
Suppose that the polynomial P consists of N monomials, each given by its exponent p. Next, we assume that the monomials are ordered in increasing exponents p and then consecutively numbered, starting from 1 to N. We use the numbers i = 1 , , N to relabel the coefficients as θ i and the exponents as p i . Moreover, we can assume that p i p j for all i j (since, otherwise, one of the monomials a i x p i and a j x p i is always dominated by the other and can be omitted).
As a result, the tropical Puiseux polynomial can be written as
P ( x ) = j = 1 N θ j x p j , x 0 ,
where θ j 0 and p j Q for all j = 1 , , N , and p 1 < < p N .
We note that, in the context of the max-plus algebra (the semifield R max , + ), the polynomial is represented, in terms of the usual operations, as
P ( x ) = max 1 j N ( p j x + θ j ) ,
and, thus, defines a piecewise-linear convex function on R . In the framework of R max , the conventional form of the polynomial becomes
P ( x ) = max 1 j N θ j × x p j ,
which in general specifies a nonlinear nonconvex function on R + .
Now, consider two polynomials:
P ( x ) = j = 1 N θ j x p j , Q ( x ) = k = 1 L σ k x q k , x 0 ,
and note that Q ( x ) 0 , since we assume that σ k 0 for all k.
A tropical rational function is the tropical quotient
R ( x ) = P ( x ) Q ( x ) .
If we interpret the rational function in terms of max-plus algebra, it is represented in the standard form as
R ( x ) = P ( x ) Q ( x ) = max 1 j N ( p j x + θ j ) max 1 k L ( q k x + σ k ) ,
which is a difference of convex functions (a DC-function). We observe that any arbitrary continuous function can be represented as the difference of two convex functions [47].

3. Best Approximate Solution of Vector Equations

We are now concerned with the solution of linear vector equations, and we describe procedures to find best approximate solutions, in the sense of the distance function given by (2). We examine the equations
A x = b , A x = B y ,
where A and B are given matrices, b is a given vector and x and y are unknown vectors.
The first equation, which has an unknown vector on one side, is normally referred to as a one-sided equation. The second equation, with unknowns on both sides, is two-sided.
To solve the equations, we first evaluate distances between vectors in tropical vector spaces, and we then apply the result, to derive the best approximate solutions.

3.1. Solution of One-Sided Equation

Consider a tropical vector space A generated by nonzero vectors a 1 , , a N X M . Any vector a A can be written as a linear combination a = x 1 a 1 x N a N , where x 1 , , x N X , and can then be represented in the form of the matrix-vector product as a = A x , with the matrix A = ( a 1 , , a N ) and vector x = ( x 1 , , x N ) T .
We examine the distance from a vector b to the vector space A , which is given by
d ( A , b ) = min a A d ( a , b ) = min x X n d ( A x , b ) .
In the case when vector b is regular, we can replace the minimum over all x X on the right-hand side by the minimum taken over regular x (see, e.g., [44]) and write
d ( A , b ) = min x > 0 d ( A x , b ) = min x > 0 ( b A x ( A x ) b ) .
It follows from the definition of d that the equality d ( A , b ) = 1 indicates the condition b A , whereas the inequality d ( A , b ) > 1 corresponds to the condition b A .
The next result is obtained in [44] (see also [45]):
Lemma 1. 
Let A be a regular matrix, b a regular vector and denote Δ = ( A ( b A ) ) b . Then,
min x > 0 d ( A x , b ) = Δ ,
where the minimum is achieved at x = Δ ( b A ) .
The above lemma demonstrates that the distance from a vector b to the vector space A generated by the columns of a matrix A is calculated as d ( A , b ) = Δ . Moreover, the vector given by y = Δ A ( b A ) is a vector of A , which is closest to the vector b .
Given a matrix A X M × N and vector b X M , consider the problem of finding regular vectors x X N that solve the one-sided equation
A x = b .
Observing that the equation can have no solution (be inconsistent), we turn to finding a best approximate solution, in the sense of the metric d. Vector x * is a best approximate solution of Equation (3) if it satisfies for all vectors x the inequality
d ( A x * , b ) d ( A x , b ) .
We note that vector A x * can be considered as a projection of b onto the space A .
A solution to Equation (3) can be obtained, based on results in [44,45], as follows.
Theorem 1. 
Let A be a regular matrix, b a regular vector and denote Δ = ( A ( b A ) ) b . Then:
1. 
The best approximate solution of Equation (3), in the sense of the distance function d, is given by
x * = Δ ( b A ) .
2. 
If Δ = 1 , the equation has exact solutions; the vector x * = ( b A ) is the maximal solution.
Since the value of Δ shows the minimum deviation between both sides of (3), measured on the scale of the function d, it also specifies the approximation error.

3.2. Solution of Two-Sided Equation

Let A be a vector space generated by vectors a 1 , , a N X M , and let B be a vector space generated by b 1 , , b L X M . With the matrices A = ( a 1 , , a N ) and B = ( b 1 , , b L ) , we define the distance between the spaces as
d ( A , B ) = min a A , b B d ( a , b ) = min x > 0 , y > 0 d ( A x , B y ) = min x > 0 ( ( B y ) A x ( A x ) B y ) .
The equality d ( A , B ) = 1 indicates that the vector spaces A and B have non-empty intersection. If the inequality d ( A , B ) > 1 holds, then these spaces have no common point. In this case, the value of the function d shows the minimum distance between the spaces.
Suppose that A X M × N and B X M × L are given regular matrices. Consider the problem of finding regular vectors x X N and y X L that solve the two-sided equation,
A x = B y .
A pair of vectors ( x * , y * ) is a best approximate solution of Equation (4) if, for all vectors x and y , the following inequality holds:
d ( A x * , B y * ) d ( A x , B y ) .
We note that the problem of finding a best approximate solution of Equation (4) is close to the best proximity point problems introduced as an extension of the classical fixed point problem. For further details on the related theory and references, one can see [48].
To obtain a best approximate solution, we apply the procedure proposed in [46], which constructs a sequence of vectors from the spaces A and B generated by the columns of the matrices A and B . The vectors are taken alternatively from both spaces, in such a way that after selecting a vector in one space the next vector is found in the other space, to minimize the distance to the former vector. The vectors in the spaces A and B are given by coefficients in their decompositions as linear combinations of columns in A and B .
Let x 0 X N be a regular vector and a 0 = A x 0 A . We apply Theorem 1, to find the minimum distance from the vector a 0 to the vectors in B as follows:
d ( a 0 , B ) = Δ 0 , Δ 0 = ( B ( ( A x 0 ) B ) ) A x 0 .
This minimum distance is provided by a vector b 1 B that is calculated as
b 1 = B y 1 , y 1 = Δ 0 ( ( A x 0 ) B ) .
We evaluate the minimum distance from b 1 to the vectors in A , which is given by
d ( b 1 , A ) = Δ 1 , Δ 1 = ( A ( ( B y 1 ) A ) ) B y 1 ,
and attained at a vector a 2 A , such that
a 2 = A x 2 , x 2 = Δ 1 ( ( B y 1 ) A ) .
Furthermore, we evaluate the distance d ( a 2 , B ) , by calculating Δ 2 , and we then use this distance to obtain vectors y 3 and b 3 . Next, we evaluate Δ 3 and find vectors x 4 and a 4 .
We continue the alternating procedure, to form a sequence of vectors a 0 , b 1 , a 2 , b 3 , a 4 , that are taken alternatively from the spaces A and B , to minimize the distance between successive vectors. A sequence of pairs ( x 0 , y 1 ) , ( x 2 , y 3 ) , is also generated, which provides successive approximations to the best approximate solution of Equation (4).
Consider the sequence Δ 0 , Δ 1 , Δ 2 , , which is bounded from below, since Δ i 1 for all i = 0 , 1 , 2 , . One can verify (see [46]) that Δ i + 1 Δ i and, hence, the sequence is non-increasing. As a result, this sequence converges to a limit Δ * 1 , defined as the greatest lower bound of its elements. Observing that the underlying semifield is complete, we conclude that this limit exists and coincides with the minimum element of the sequence.
We note that each element of the sequence specifies the squared distance between a vector of one of the spaces A and B and the nearest vector in the other space. If the equality Δ i = 1 is satisfied for some i, then the spaces have non-empty intersection, and Equation (4) has exact regular solutions. In the case of even i, the intersection contains the vector a i = A x i , whereas the pair of vectors ( x i , y i + 1 ) is a solution. If i is odd, the intersection contains b i = B y i , whereas the pair ( x i + 1 , y i ) is a solution.
If the inequality Δ * > 1 holds, then the spaces A and B do not intersect, and Equation (4) does not have exact solutions. In such an event, the procedure stops as soon as a repeated element occurs in any of the sequences x 0 , x 2 , or y 1 , y 3 , .
The above alternating procedure can be represented as the following algorithm.
Algorithm 1: Solution of two-sided equation A x = B x
  • Input regular matrices A , B and regular vector x 0 ; set i = 0 .
  • Calculate
    Δ i = ( B ( ( A x i ) B ) ) A x i , y i + 1 = Δ i ( ( A x i ) B ) .
  • If Δ i = 1 or y i + 1 = y j for some j < i , then set
    Δ * = Δ i , x * = x i , y * = y i + 1 ,
    and stop; otherwise, set i = i + 1 .
  • Calculate
    Δ i = ( A ( ( B y i ) A ) ) B y i , x i + 1 = Δ i ( ( B y i ) A ) .
  • If Δ i = 1 or x i + 1 = x j for some j < i , then set
    Δ * = Δ i , x * = x i + 1 , y * = y i ,
    and stop; otherwise, set i = i + 1 .
  • Go to step 2.
Suppose that upon completion of the algorithm, we obtain Δ * = 1 . This result means that Equation (4) has exact regular solutions, which include the obtained pair of vectors ( x * , y * ) . If we have Δ * > 1 , then the equation has no exact solution. The value of Δ * specifies the approximation error attained at a best approximate solution given by ( x * , y * ) .

4. Tropical Discrete Best Approximation of Functions

We are now in a position to propose new algebraic techniques to solve the data approximation problems of finding an unknown but definite function y = f ( x ) from finitely many samples ( x i , y i ) in the tropical algebra setting. The purpose is to determine a functional model that can effectively approximate the underlying relation between x i and y i .
As approximating functions, the techniques use tropical polynomial and rational functions. The approximation problems are reduced to evaluating the best approximate solutions of vector equations obtained from the sample data.

4.1. Tropical Polynomial Approximation

Suppose that samples x i , y i X are given for i = 1 , , M , where y i is assumed to be an output obtained from an input x i by some function f : X X . Let us consider the problem of approximating the function f by a polynomial, using these sample data.
For each j = 1 , , N , let p j Q denote a fixed number, and θ j X denote an unknown parameter. Define the approximating polynomial as
P ( x ) = j = 1 N θ j x p j = θ 1 x p 1 θ N x p N .
The problem consists in the determination of the parameters that make the equations
θ 1 x i p 1 θ N x i p N = y i i = 1 , , M ,
hold exactly or approximately, by minimizing the deviation between both sides.
We introduce the matrix-vector notation,
X = x 1 p 1 x 1 p N x M p 1 x M p N , y = y 1 , y M , θ = θ 1 θ N ,
and we combine the above scalar equations into the one-sided vector equation,
X θ = y ,
where X and y are a known matrix and vector, and θ is an unknown vector of parameters.
To make the equation fit the sample data, we need to find a best approximate solution, in the sense of a minimum distance between the vectors X θ and y . With the distance function d, the problem is to obtain a vector θ * = ( θ 1 * , , θ N * ) T that is given by
θ * = arg min θ d ( X θ , y ) .
We solve the problem by applying Theorem 1, which provides a best approximation solution in terms of the metric d. We replace the matrix A by X , and the vectors b and x by y and θ . First, we calculate the squared best approximation error:
Δ * = ( X ( y X ) ) y .
Next, we obtain the vector of parameters that achieves this error:
θ * = Δ * ( y X ) .
The best approximating polynomial is then given by
P * ( x ) = θ 1 * x p 1 θ N * x p N .

4.2. Tropical Rational Approximation

As before, let x i , y i X be given samples for i = 1 , , M . Furthermore, let p j , q k Q be fixed numbers and θ j , σ k X be unknown parameters for j = 1 , , N and k = 1 , , L .
Denote numerator and denominator polynomials by
P ( x ) = j = 1 N θ j x p j , Q ( x ) = k = 1 L σ k x q k ,
and define an approximating rational function as
R ( x ) = P ( x ) Q ( x ) = θ 1 x p 1 θ N x p N σ 1 x q 1 σ L x q L .
We now consider the problem of evaluating the parameters that achieve the best approximation of the equations
θ 1 x i p 1 θ N x i p N σ 1 x i q 1 σ L x i q L = y i , i = 1 , , M .
We rewrite the equations as
θ 1 x i p 1 θ N x i p N = y i ( σ 1 x i q 1 σ L x i q L ) , i = 1 , , M .
To represent the problem in vector form, we use the matrix-vector notation
X = x 1 p 1 x 1 p N x M p 1 x M p N , Y = y 1 0 0 y M , Z = x 1 q 1 x 1 q L x M q 1 x M q L , θ = θ 1 θ N , σ = σ 1 σ L ,
where X , Y and Z are known data matrices, and where θ and σ are unknown vectors of parameters.
With this notation, the scalar equations take the form of the two-sided vector equation
X θ = Y Z σ .
Let us verify that the approximation error for the last vector equation coincides with that of the system (5). We define the vector w = ( w i ) with elements w i = ( X θ ) i ( Z σ ) i 1 , and we examine the distance d ( w , y ) that represents the approximation error for (5). By using the scalar representation at (2), we immediately obtain
d ( w , y ) = i = 1 M ( ( X θ ) i ( Z σ ) i 1 y i 1 ( X θ ) i 1 ( Z σ ) i y i ) = d ( X θ , Y Z σ ) .
We now seek for parameters that provide a minimum distance between X θ and Y Z σ . We need to find vectors θ * = ( θ 1 * , , θ N * ) T and σ * = ( σ 1 * , , σ L * ) T , such that
( θ * , σ * ) = arg min θ , σ d ( X θ , Y Z σ ) .
To solve the problem, we apply Algorithm 1, where the matrices A and B are replaced by X and Y Z , and the vectors x and y by θ and σ . The algorithm yields the squared minimum distance Δ * and vectors θ * and σ * that attain the minimum.
The best approximate rational function is given by
R * ( x ) = θ 1 * x p 1 θ N * x p N σ 1 * x q 1 σ L * x q L .
To conclude this section, we note that the proposed approximation techniques assume that the number of terms (monomials) and their related degrees in the approximating polynomials P ( x ) and Q ( x ) are fixed in advance. For some approximation problems, these characteristics of polynomials can be determined based on additional information about the nature and properties of the phenomena under study. However, in many cases there is no prior information on the characteristics, which makes the selection of approximating polynomials a nontrivial challenge that needs to be addressed in future research.
As a first simple but reasonable step to meet the challenge, a Monte Carlo random sampling technique can be implemented, to search for the class of optimal polynomials. We use a random sampling technique to determine the degrees of monomials when solving illustrative example problems in the next section.

5. Numerical Examples

In this section, we demonstrate how tropical approximation techniques can be used to solve discrete approximation problems with sample data evaluated from simple convex and nonconvex functions. To provide transparent graphical illustrations, we offer examples, in terms of the semifield R max , + (max-plus algebra), where both polynomials and rational functions can be represented in Cartesian coordinates as piecewise linear functions.
We consider approximation problems in which approximating polynomials have a fixed number of monomials, while the degrees of these monomials are not given in advance. We propose a two-level solution procedure that combines direct random search, to select a class of approximating polynomials or rational functions at the upper level, with best approximation using approximants from the selected class at the lower level.
To solve polynomial approximation problems, given a number of monomials in the polynomial, we implement uniform random sampling over a feasible set of degrees. Each sample sets the degree of each monomial to a fixed value and, hence, we define a class of polynomials with specified degrees and arbitrary coefficients. Then, we apply Theorem 1, to find coefficients of a best approximate polynomial in the class, and we evaluate the related approximation error, to allow the search of a sample that produces the minimum error.
The solution of rational approximation problems includes independent random sampling for both polynomials in the numerator and denominator of the approximating rational function. The coefficients of the polynomials and corresponding approximation error are obtained by solving the approximation problem, with the use of Algorithm 1.
To reduce the number of classes of polynomials, to examine and simplify the random search in numerical experiments, we consider only polynomials with integer degrees.

5.1. Discrete Approximation of Convex Function

We consider a function defined in terms of standard algebra as
f ( x ) = x 2 3 x 1 / 3 + 5 / 2 , 0 x 2 .
We suppose that there are M = 21 samples ( x i , y i ) available, obtained with f ( x ) on the interval [ 0 , 2 ] at x i = ( i 1 ) / 10 , where i = 1 , , M . The problem is to find an approximate tropical polynomial that best describes the relationship between x i and y i for all i.
Observing that the function f ( x ) is convex on the interval considered, we start with the approximation of f ( x ) by polynomials. We first consider polynomials of N = 5 terms,
P ( x ) = θ 1 x p 1 θ 5 x p 5 ,
where the vector of degrees p = ( p 1 , , p 5 ) T is not fixed in advance and, thus, its elements are produced by random sampling from the discrete uniform distribution over [ 15 , 5 ] .
For each obtained sample vector of degrees, we evaluate the vector of coefficients θ = ( θ 1 , , θ 5 ) T , to achieve the minimum approximation error corresponding to the sample. The combination of sample degrees and their related coefficients that attains the minimum error over all samples is taken, to form a best approximate polynomial.
After examining 10,000 sample vectors of exponents, we arrive at the minimum squared error Δ * = 0.1360 . We obtain the following vectors of coefficients and degrees:
θ * T = 2.5680 0.9176 0.4320 1.6281 3.2413 , p * T = 14 1 1 2 3 .
The obtained polynomial represented in the conventional form is given by
P * ( x ) = max ( 2.5680 15 x , 0.9176 x , 0.4320 + x , 1.6281 + 2 x , 3.2413 + 3 x ) .
A graphical illustration of the solution is presented in Figure 1.
Next, we use polynomials with N = 7 terms, in the form
P ( x ) = θ 1 x p 1 θ 7 x p 7 .
Approximation by these polynomials, combined with a random search based on 10,000 samples, yields the minimum Δ * = 0.0481 and the vectors
θ * T = 2.5240 1.4096 0.8736 0.3503 0.4760 1.6720 3.2853 , p * T = 15 3 1 0 1 2 3 .
The obtained solution takes the form of the function
P * ( x ) = max ( 2.5240 15 x , 1.4096 3 x , 0.8736 x , 0.3503 , 0.4760 + x , 1.6720 + 2 x , 3.2853 + 3 x ) ,
which is demonstrated in Figure 2.
We observe that, as expected, the approximation by polynomials with the number of terms N increases from 5 to 7 results in a reduction in approximation error, which decreases the squared error Δ * from 0.1360 to 0.0481 .

5.2. Discrete Approximation of Nonconvex Function

Once again, we suppose that M = 21 samples ( x i , y i ) are given, which are obtained at the points x i = ( i 1 ) / 10 for i = 1 , , M , from the function
g ( x ) = 3 ( x 1 ) 2 sin ( x ) + 1 / 4 , 0 x 2 .
This function is nonconvex, which makes it appropriate to approximate the sample data by a tropical rational function defined as the ratio of polynomials:
R ( x ) = P ( x ) / Q ( x ) .
We take numerator and denominator polynomials with N = 4 and L = 2 terms to be
P ( x ) = θ 1 x p 1 θ 4 x p 4 , Q ( x ) = σ 1 x q 1 σ 2 x q 2 .
Since the vectors p = ( p 1 , , p 4 ) T and q = ( q 1 , q 2 ) T are not known, we obtain their components by random sampling from the discrete uniform distribution over [ 10 , 10 ] . For each obtained pair of vectors, we employ the procedure of rational approximation described in the previous section.
Solving the approximation problem for 10,000 sample vector pairs results in the minimum squared error Δ * = 0.1395 . The optimal vectors of coefficients, together with corresponding vectors of degrees, are found in the following form:
θ * T = 11.4688 9.6386 5.3845 10.5453 , σ * T = 11.1281 10.3861 , p * T = 3 2 1 2 , q * T = 5 2 .
In the ordinary notation, the approximating function is written as
R * ( x ) = P * ( x ) Q * ( x ) ,
where
P * ( x ) = max ( 3.4753 3 x , 2.7409 2 x , 1.0110 + x , 2.6014 + 2 x ) , Q * ( x ) = max ( 3.2525 5 x , 2.4211 2 x ) .
A graphical illustration is given in Figure 3.
To obtain a more accurate solution, we consider an approximating rational function that has numerator and denominator polynomials with N = 6 and L = 4 terms given by
P ( x ) = θ 1 x p 1 θ 6 x p 6 , Q ( x ) = σ 1 x q 1 σ 4 x q 4 .
After random sampling of 10,000 vector pairs, to define degrees of polynomials involved, and solving the corresponding rational approximation problems, we obtain a solution with squared error Δ * = 0.0701 and the following vectors:
θ * T = 6.9455 6.0860 4.9978 3.7461 0.7639 2.6361 , σ * T = 6.6880 6.2962 5.8009 2.4211 , p * T = 3 2 1 0 2 4 , q * T = 5 3 2 0 .
For the approximating function in the usual form R * ( x ) = P * ( x ) Q * ( x ) , we obtain
P * ( x ) = max ( 6.9455 3 x , 6.0860 2 x , 4.9978 x , 3.7461 , 0.7639 + 2 x , 2.6361 + 4 x ) , Q * ( x ) = max ( 6.6880 5 x , 6.2962 3 x , 5.8009 2 x , 2.4211 ) .
The solution obtained is illustrated in Figure 4.

6. Conclusions

In this paper, we introduced new discrete best approximation problems given in the tropical algebra setting. The problems were formulated to approximate an unknown function on a tropical (idempotent) semifield by tropical Puiseux polynomial and rational functions. We proposed an approach to solving both polynomial and rational approximation problems, by finding the best approximate solutions of the corresponding tropical linear vector equations. Specifically, the approximation by polynomials was reduced to a one-sided vector equation, for which we provided a direct analytical solution. The rational approximation led to a two-sided equation, which we solved using an iterative alternating algorithm.
To illustrate the solution approach, we gave examples of solving approximation problems in terms of max-plus algebra. If defined in the context of max-plus algebra, Puiseux polynomials take the form of convex piecewise linear functions, whereas rational functions take the form of nonconvex piecewise linear functions. For the sake of simplicity, we used approximating polynomials with integer exponents that, in general, provide less flexibility of fitting data than polynomials with rational exponents. Nevertheless, the examples clearly demonstrated the applicability of the technique, at least to solve problems of discrete best Chebyshev approximation of real functions.
Moreover, we observed that the solution procedures proposed to solve the approximation problems did not change when the Puiseux polynomials were replaced by polynomials that may have had real exponents (if defined in the underlying semifield) and not just rational ones. This extended the approach, to allow generalized tropical Puiseux polynomial and rational functions to also be used as approximating functions.
We believe the main purpose of the paper is to concentrate on the key component of the new approximation approach and, therefore, we have left for future research many issues that may arise in implementation. Specifically, for approximation problems in the context of max-plus algebra, it is important to study how the approximate solution and its error depend on the number of monomials in the approximating polynomial (rational) functions and on the values of their exponents. The results of the study could help in the development of optimization methods to find exponents that minimize the approximation error, which presents another line of future research. Finally, implementation of the technique—in terms of other tropical semifields, including max-algebra—and the analysis of related results are also of considerable interest.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The author is very grateful to the anonymous reviewers for their valuable comments and suggestions, which have been incorporated in the revised manuscript.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Harter, H.L. The method of least squares and some alternatives: Part IV. Int. Stat. Rev. 1975, 43, 125–190. [Google Scholar] [CrossRef]
  2. Mhaskar, H.N.; Pai, D.V. Fundamentals of Approximation Theory; Narosa Publishing House: New Delhi, India, 2000. [Google Scholar]
  3. Steffens, K.G. The History of Approximation Theory; Birkhäuser: Boston, MA, USA, 2006. [Google Scholar] [CrossRef]
  4. de Laplace, P.S. Mécanique Céleste; Bowditch, N., Ed.; Hillard, Gray, Littlè, and Wilkins: Boston, MA, USA, 1832; Volume 2. [Google Scholar]
  5. Tewarson, R.P. On minimax solutions of linear equations. Comput. J. 1972, 15, 277–279. [Google Scholar] [CrossRef]
  6. Appa, G.; Smith, C. On L1 and Chebyshev estimation. Math. Program. 1973, 5, 73–87. [Google Scholar] [CrossRef]
  7. Pĭnar, M.Ç. Overdetermined systems of linear equations. In Encyclopedia of Optimization, 2nd ed.; Floudas, C.A., Pardalos, P.M., Eds.; Springer: Boston, MA, USA, 2009; pp. 2893–2896. [Google Scholar] [CrossRef]
  8. Rabinowitz, P. Applications of linear programming to numerical analysis. SIAM Rev. 1968, 10, 121–159. [Google Scholar] [CrossRef]
  9. Hand, M.L. Aspects of Linear Regression Estimation under the Criterion of Minimizing the Maximum Absolute Residual. Ph.D. Thesis, Iowa State University, Ames, IA, USA, 1978. [Google Scholar]
  10. Kennedy, W.J.; Gentle, J.E. Statistical Computing. In Statistics: Textbooks and Monographs; Dekker: New York, NY, USA, 1980; Volume 33. [Google Scholar]
  11. Wagner, H.M. Linear programming techniques for regression analysis. J. Am. Statist. Assoc. 1959, 54, 206–212. [Google Scholar] [CrossRef]
  12. Stiefel, E. Note on Jordan elimination, linear programming and Tchebycheff approximation. Numer. Math. 1960, 2, 1–17. [Google Scholar] [CrossRef]
  13. Watson, G.A. On the best linear one-sided Chebyshev approximation. J. Approx. Theory 1973, 7, 48–58. [Google Scholar] [CrossRef]
  14. Sposito, V.A. Minimizing the maximum absolute deviation. ACM SIGMAP Bull. 1976, 20, 51–53. [Google Scholar] [CrossRef]
  15. Armstrong, R.D.; Kung, D.S. Algorithm AS 135: Min-max estimates for a linear multiple regression problem. J. R. Stat. Soc. Ser. C Appl. Stat. 1979, 28, 93–100. [Google Scholar] [CrossRef]
  16. Kim, B.Y. Algorithm for the constrained Chebyshev estimation in linear regression. Commun. Stat. Appl. Methods 2000, 7, 47–54. [Google Scholar]
  17. Boyd, S.; Vandenberghe, L. Convex Optimization; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar]
  18. Castillo, E.; Mínguez, R.; Castillo, C.; Cofiño, A.S. Dealing with the multiplicity of solutions of the 1 and regression models. Eur. J. Oper. Res. 2008, 188, 460–484. [Google Scholar] [CrossRef]
  19. Ene, A.; Vladu, A. Improved convergence for 1 and regression via iteratively reweighted least squares. In Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; Chaudhuri, K., Salakhutdinov, R., Eds.; PMLR: Cambridge, MA, USA, 2019; Volume 97, pp. 1794–1801. [Google Scholar]
  20. Krivulin, N. Using parameter elimination to solve discrete linear Chebyshev approximation problems. Mathematics 2020, 8, 2210. [Google Scholar] [CrossRef]
  21. Conn, A.R.; Li, Y. The Computational Structure and Characterization of Nonlinear Discrete Chebyshev Problem; Technical Report 88-956; Department of Computer Science, Cornell University: Ithaca, NY, USA, 1988. [Google Scholar]
  22. Osborne, M.R.; Watson, G.A. An algorithm for minimax approximation in the nonlinear case. Comput. J. 1969, 12, 63–68. [Google Scholar] [CrossRef]
  23. Madsen, K. Minimax solution of nonlinear equations without calculating derivatives. In Nondifferentiable Optimization; Balinski, M.L., Wolfe, P., Eds.; Mathematical Programming Studies; Springer: Berlin/Heidelberg, Germany, 1975; Volume 4, pp. 110–126. [Google Scholar] [CrossRef]
  24. Anderson, D.H.; Osborne, M.R. Discrete, nonlinear approximation problems in polyhedral norms. Numer. Math. 1977, 28, 143–156. [Google Scholar] [CrossRef]
  25. Charalambous, C.; Conn, A.R. An efficient method to solve the minimax problem directly. SIAM J. Numer. Anal. 1978, 15, 162–187. [Google Scholar] [CrossRef]
  26. Hald, J.; Madsen, K. Combined LP and quasi-Newton methods for minimax optimization. Math. Program. 1981, 20, 49–62. [Google Scholar] [CrossRef]
  27. Han, S.P. Variable metric methods for minimizing a class of nondifferentiable functions. Math. Program. 1981, 20, 1–13. [Google Scholar] [CrossRef]
  28. Fletcher, R. A model algorithm for composite nondifferentiable optimization problems. In Nondifferential and Variational Techniques in Optimization; Sorensen, D.C., Wets, R.J.B., Eds.; Mathematical Programming Studies; Springer: Berlin/Heidelberg, Germany, 1982; Volume 17, pp. 67–76. [Google Scholar] [CrossRef]
  29. Kolokoltsov, V.N.; Maslov, V.P. Idempotent Analysis and Its Applications. In Mathematics and Its Applications; Springer: Dordrecht, The Netherlands, 1997; Volume 401. [Google Scholar] [CrossRef]
  30. Golan, J.S. Semirings and Affine Equations Over Them. In Mathematics and Its Applications; Springer: Dordrecht, The Netherlands, 2003; Volume 556. [Google Scholar] [CrossRef]
  31. Heidergott, B.; Olsder, G.J.; van der Woude, J. Max Plus at Work; Princeton Series in Applied Mathematics; Princeton University Press: Princeton, NJ, USA, 2006. [Google Scholar]
  32. Itenberg, I.; Mikhalkin, G.; Shustin, E. Tropical Algebraic Geometry. In Oberwolfach Seminars; Birkhäuser: Basel, Switzerland, 2007; Volume 35. [Google Scholar] [CrossRef]
  33. Gondran, M.; Minoux, M. Graphs, Dioids and Semirings. In Operations Research/ Computer Science Interfaces; Springer: New York, NY, USA, 2008; Volume 41. [Google Scholar] [CrossRef]
  34. Butkovič, P. Max-Linear Systems; Springer Monographs in Mathematics; Springer: London, UK, 2010. [Google Scholar] [CrossRef]
  35. Maclagan, D.; Sturmfels, B. Introduction to Tropical Geometry. In Graduate Studies in Mathematics; AMS: Providence, RI, USA, 2015; Volume 161. [Google Scholar] [CrossRef]
  36. Akian, M.; Gaubert, S.; Niţică, V.; Singer, I. Best approximation in max-plus semimodules. Linear Algebra Appl. 2011, 435, 3261–3296. [Google Scholar] [CrossRef]
  37. Saadi, O. Zero-Sum Repeated Games: Accelerated Algorithms and Tropical Best-Approximation. Ph.D. Thesis, Institut Polytechnique de Paris, Paris, France, 2021. [Google Scholar]
  38. Zhang, L.; Naitzat, G.; Lim, L.H. Tropical geometry of deep neural networks. In Proceedings of the 35th International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; Dy, J., Krause, A., Eds.; PMLR: Cambridge, MA, USA, 2018; Volume 80, pp. 5824–5832. [Google Scholar]
  39. Maragos, P.; Charisopoulos, V.; Theodosis, E. Tropical geometry and machine learning. Proc. IEEE 2021, 109, 728–755. [Google Scholar] [CrossRef]
  40. Markwig, T. A field of generalised Puiseux series for tropical geometry. Rend. Sem. Mat. Univ. Politec. Torino 2010, 68, 79–82. [Google Scholar]
  41. Hampe, S.; Joswig, M. Tropical computations in polymake. In Algorithmic and Experimental Methods in Algebra, Geometry, and Number Theory; Böckle, G., Decker, W., Malle, G., Eds.; Springer: Cham, Switzerland, 2017; pp. 361–385. [Google Scholar] [CrossRef]
  42. Grigoriev, D. Tropical Newton-Puiseux polynomials. In Proceedings of the Computer Algebra in Scientific Computing; Lecture Notes in Computer Science; Gerdt, V.P., Koepf, W., Seiler, W.M., Vorozhtsov, E.V., Eds.; Springer: Cham, Switzerland, 2018; Volume 11077, pp. 177–186. [Google Scholar] [CrossRef]
  43. Krivulin, N. Algebraic solution of tropical polynomial optimization problems. Mathematics 2021, 9, 2472. [Google Scholar] [CrossRef]
  44. Krivulin, N.K. On solution of a class of linear vector equations in idempotent algebra. Vestn.-St.-Peterbg. Univ. Seriya 2009, 10, 64–77. (In Russian) [Google Scholar]
  45. Krivulin, N. Solution of linear equations and inequalities in idempotent vector spaces. Int. J. Appl. Math. Inform. 2013, 7, 14–23. [Google Scholar]
  46. Krivulin, N.K. On the solution of a two-sided vector equation in tropical algebra. Vestnik St. Petersburg Univ. Math. 2023, 56, 172–181. [Google Scholar] [CrossRef]
  47. Tuy, H.D.C. Programming. In Encyclopedia of Optimization; Floudas, C.A., Pardalos, P.M., Eds.; Springer: Boston, MA, USA, 2009; pp. 607–612. [Google Scholar] [CrossRef]
  48. Sadiq Basha, S. Best proximity point theorems. J. Approx. Theory 2011, 163, 1772–1781. [Google Scholar] [CrossRef]
Figure 1. Approximation of the function f ( x ) by a polynomial P * ( x ) with N = 5 terms.
Figure 1. Approximation of the function f ( x ) by a polynomial P * ( x ) with N = 5 terms.
Mathematics 11 03949 g001
Figure 2. Approximation of the function f ( x ) by a polynomial P * ( x ) with N = 7 terms.
Figure 2. Approximation of the function f ( x ) by a polynomial P * ( x ) with N = 7 terms.
Mathematics 11 03949 g002
Figure 3. Approximation of function g ( x ) by a rational function R * ( x ) , with N = 4 and L = 2 .
Figure 3. Approximation of function g ( x ) by a rational function R * ( x ) , with N = 4 and L = 2 .
Mathematics 11 03949 g003
Figure 4. Approximation of function g ( x ) by a rational function R * ( x ) , with N = 6 and L = 4 .
Figure 4. Approximation of function g ( x ) by a rational function R * ( x ) , with N = 6 and L = 4 .
Mathematics 11 03949 g004
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Krivulin, N. Algebraic Solution of Tropical Best Approximation Problems. Mathematics 2023, 11, 3949. https://doi.org/10.3390/math11183949

AMA Style

Krivulin N. Algebraic Solution of Tropical Best Approximation Problems. Mathematics. 2023; 11(18):3949. https://doi.org/10.3390/math11183949

Chicago/Turabian Style

Krivulin, Nikolai. 2023. "Algebraic Solution of Tropical Best Approximation Problems" Mathematics 11, no. 18: 3949. https://doi.org/10.3390/math11183949

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop