Next Article in Journal
Mining High Utility Itemsets Based on Pattern Growth without Candidate Generation
Next Article in Special Issue
Multiple Comparisons for Exponential Median Lifetimes with the Control Based on Doubly Censored Samples
Previous Article in Journal
Mathematical Approach to Improve the Thermoeconomics of a Humidification Dehumidification Solar Desalination System
Previous Article in Special Issue
Complex Ginzburg–Landau Equation with Generalized Finite Differences
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Towards a Vector Field Based Approach to the Proper Generalized Decomposition (PGD)

1
ESI International Chair@CEU-UCH, Departamento de Matemáticas, Física y Ciencias Tecnológicas, Universidad Cardenal Herrera-CEU, CEU Universities San Bartolomé 55, 46115 Alfara del Patriarca, Spain
2
Departamento de Ingeniería Mecánica y Construcción, Universitat Jaume I, Avd. Vicent Sos Baynat s/n, 12071 Castellón, Spain
3
Departamento de Ingeniería Mecánica y de Materiales, Universitat Politècnica de València Camino de Vera, s/n, 46022 Valencia, Spain
*
Author to whom correspondence should be addressed.
Mathematics 2021, 9(1), 34; https://doi.org/10.3390/math9010034
Submission received: 11 October 2020 / Revised: 13 December 2020 / Accepted: 21 December 2020 / Published: 25 December 2020
(This article belongs to the Special Issue Applications of Partial Differential Equations in Engineering)

Abstract

:
A novel algorithm called the Proper Generalized Decomposition (PGD) is widely used by the engineering community to compute the solution of high dimensional problems. However, it is well-known that the bottleneck of its practical implementation focuses on the computation of the so-called best rank-one approximation. Motivated by this fact, we are going to discuss some of the geometrical aspects of the best rank-one approximation procedure. More precisely, our main result is to construct explicitly a vector field over a low-dimensional vector space and to prove that we can identify its stationary points with the critical points of the best rank-one optimization problem. To obtain this result, we endow the set of tensors with fixed rank-one with an explicit geometric structure.

1. Introduction

Many problems in science and engineering are hard to compute due their numerical complexity. Moreover, in the analysis of complex systems under a real time constraint, the evaluation of all possible scenarios appears as a necessity [1]. Despite the improvements in techniques used in high dimensional problems, some challenging questions remain unresolved due to the efficiency of our computers. However, a novel technique called Proper Generalized Decomposition (PGD) [2,3] has been developed to provide an answer of these difficult tasks. It was initially proposed to compute, in a separated representation framework, the variational solution of partial differential equations (PDE) defined over a tensor product space [4]. It is possible to distinguish two different benefits. The first one is the possibility of managing high dimensional problems, and the second is the possibility to include the model’s parameters as extra-coordinates. This last fact gives a powerful strategy to deal with classical problems because the PGD framework facilitates an efficient design and a real-time decision-making [5,6].
This novel technique allows for computing the whole set of solutions of a parametrized problem. The strategy is to include in an equivalent non-parametrized problem all possible parameter values as extra-coordinates. The name of this particular PGD based approach is Progressive Variational Vademecum [1], and it can be implemented offline. As a consequence, the PGD based approach opens the possibility of solving problems in industry with a different strategy not envisioned until now.
The mathematical analysis of the PGD was given by Falcó and Nouy, Ref. [4] in a Hilbert space framework and in [7] for a more general setting.
The Greedy Rank One Algorithm (GROA) [8] used to solve high-dimensional linear systems (with a full rank matrix) is the procedure of choice in the engineering community to implement the PGD. It is an iterative method made up of two steps that we can cyclically repeat until convergence. The first one consists of computing the minimal residual of the linear system over the set of tensors with bounded rank-one. In the second step, we use this optimal rank-one solution to update the residual. In the following, we return to the first step. Ammar et al. [8] propose an Alternating Least Squares (ALS) Algorithm for the practical implementation of the first step of the GROA. In the aforementioned paper, the authors justify the choice of the ALS showing that its convergence to a critical point of the optimization problem (not necessarily an optimal one) is assured under very weak conditions. In addition, the convergence has been studied by El Hamidi, Osman and Jazar [9] in the framework of Sobolev tensor spaces.
In this work, we want to study the optimization problem of the first step the GROA. To this end, even though our problem is not convex, we will take into account the relationship between the underlying convex optimization problem and the behaviour of its associated gradient flow. It is well-known that the vector field constructed by using a convex functional (for example related with a convex minimization problem defined over a finite dimensional vector space), has a gradient flow that provides a dynamical system with a unique stationary point. Moreover, it can be shown that it is a sink and its stable manifold coincides with the whole domain of the convex functional. This fact motivates the classical paradigm about the convergence of gradient-based numerical optimization algorithms.
In this work, we want to adapt the above paradigm in the framework of the GROA [8]. The main goal is to find a vector field in a low-dimensional vector space related to the gradient flow of a convex functional defined over the set of tensors of fixed rank one. The idea is to use this vector field to characterize the behaviour of the solutions of the non-convex optimization problem associated with the first step of GROA. To achieve this, we will prove that the set of critical points of the optimization problem over the set of tensors with fixed rank-one can be identified with the set of stationary points of that vector field. In order to construct it, we will proceed as follows. First, we will endow the set tensors of fixed rank-one with an explicit structure of smooth manifold. Second, by the help of this geometric structure, we will explicitly construct a vector field over a low dimensional vector space related with the first step of GROA. Finally, we will show that the set of stationary points of this vector field coincides with the set of critical points of the optimization problem associated with the PGD algorithm. In consequence, this vector field allows us to explain the dynamical behaviour around each of its stationary points. Moreover, we can get explicit information, in a neighbourhood of each of these stationary points, about the structure of its stable and unstable manifolds. In our opinion, a more precise knowledge of these invariant sets can help us develop better and more efficient PGD approaches.
The paper is organised as follows. Section 2 provides some preliminary definitions and results used along this paper. Section 3 shows a geometric approach to the PGD. In Section 4, the characterization of the smooth manifold of the set of tensors of fixed rank-one is given. After that, Section 5 shows the first order optimality conditions for the PGD which is the main result of this paper. Finally, Section 6 provides some conclusions of the work.

2. Preliminary Definitions and Results

First of all, we introduce some notation used along this paper. We denote by R N × M the set of N × M -matrices and by A T the transpose of a given matrix A . As usual, we use
x , y = x T y = y T x
to denote the Euclidean inner product in R N , and its corresponding 2-norm, by  x 2 = x , x 1 / 2 . Let I N be the N × N -identity matrix and when the dimension is clear from the context, we simply denote it by I . Given a sequence { u j } j = 0 R N , we say that a vector u R N can be written as
u = j = 0 u j
if and only if
lim n j = 0 n u j = u
in the · 2 -topology. Now, we recall the definition and some properties of the Kronecker product. The Kronecker product of A R N 1 × N 1 and B R N 2 × N 2 , written A B , is the tensor algebraic operation defined as
A B = A 1 , 1 B A 1 , 2 B A 1 , N 1 B A 2 , 1 B A 2 , 2 B A 2 , N 1 B A N 1 , 1 B A N 1 , 2 B A N 1 , N 1 B R N 1 N 2 × N 1 N 2 .
To conclude, we list some of the well-known properties of the Kronecker product (see, for example, [10] or [11]).
1.
A ( B C ) = ( A B ) C .
2.
( A + B ) ( C + D ) = ( A C ) + ( B C ) + ( A D ) + ( B D ) .
3.
A B C D = ( A C ) ( B D ) .
4.
( A B ) 1 = A 1 B 1 .
5.
( A B ) T = A T B T .
6.
If A and B are banded, then A B is banded.
7.
If A and B are symmetric, then A B is symmetric.
8.
If A and B are definite positive, then A B is definite positive.
The concept of separated representation was introduced by Beylkin and Mohlenkamp in [12], and it is related to the problem of constructing the approximate solutions of some classes of problems in high-dimensional spaces by means of a separable function. In particular, for a given map
u : [ 0 , 1 ] d R d R ,
we say that it has a separable representation if
u ( x 1 , , x d ) = j = 1 u 1 ( j ) ( x 1 ) u d ( j ) ( x d )
Now, consider a mesh of [ 0 , 1 ] in the x k -variable given by N k -mesh points, 1 k d , then we can write a discrete version of (1) by
u ( x i 1 , , x i d ) = j = 1 u 1 ( j ) ( x i 1 ) u d ( j ) ( x i d ) ,
where 1 i k N k for 1 k d . Observe that, for each 1 k d , if x k j R N k denotes the vector with components u k ( j ) ( x i k ) for 1 i k N k , then (2) is equivalent to
u = j = 1 x 1 j x d j .
We point out that (3) is a useful expression to implement numerical algorithms using the Matlab and Octave function kron.
Suppose that, for given a linear Partial Differential Equation, and after a discretization by means of Finite Elements, we need to solve the linear system:
A u = f ,
where A is a ( N 1 N d ) × ( N 1 N d ) -dimensional invertible matrix, for some N 1 , , N d N , that is, A GL ( N 1 N d ) . Then, from all said above, a low rank approximation
A 1 f u n = j = 1 n x 1 j x d j
with sufficient approximation exists, for some n 1 and where x i j R N i for i = 1 , 2 , , d  and j = 1 , 2 , , n . Moreover, we would show that
lim n A 1 f u n 2 = 0 ,
that is,
A 1 f = j = 1 x 1 j x d j .
Thus, in a first approach to solve it, we would like to determine vectors x 1 j , , x d j for j = 1 , 2 , , n that minimizes
f A j = 1 n x 1 j x d j 2 ,
or, in short
argmin rank u n f A u 2 ,
by using the notation introduced in [13].
The Proper Generalised Decomposition (PGD in short) appears when we consider solving the linear Equation (4) as an optimization problem as follows. For each fixed A GL ( N 1 N 2 N d ) and f R N 1 N 2 N d , we define a map
J A , f : R N 1 N d R , J A , f ( u ) = 1 2 f A u 2 2 ,
hence
argmin u R N 1 N 2 N d J A , f ( u ) = { A 1 f }
holds. The goal is to use (5) to approximate the solution of (4). To this end, for each n N , we define the set
S n = { x R N 1 N d : rank x n } ,
introduced in [13], in the following way. Given x R N 1 N d , we say that x S 1 = S 1 ( N 1 , N 2 , , N d ) if x = x 1 x 2 x d , where x i R N i , for i = 1 , , d . For n 2 , we define inductively S n = S n ( N 1 , N 2 , , N d ) = S n 1 + S 1 , that is,
S n = x : x = i = 1 k x ( i ) , x ( i ) S 1 for 1 i k n .
Note that S n S n + 1 for all n 1 .
Unfortunately, from Proposition 4.1 (a) of [13], we have that the set S n is not necessarily (or even usually) closed for each n 2 . However, from Proposition 4.2 of [13], it follows that S 1 is a closed set in any norm-topology. This fact implies (see Lemma 1 in [8]), that given A GL ( N 1 N 2 N d ) , then, for every f R N 1 N d , we have that the set
C ( A , f ) : = argmin x S 1 J A , f ( x ) .
This allows for considering the following iterative scheme. Let u 0 = y 0 = 0 , and, for each n 1 , take
r n 1 = f A u n 1 ,
u n = u n 1 + y n where y n C ( A , r n 1 ) .
Note that, given A GL ( N 1 N 2 N d ) and f R N 1 N d , we can construct for each n , by using (8) and (9), a vector
u n = j = 1 n y n S n \ S n 1 .
Here, we assume that y j 0 for 1 j n , that is, rank u n = n . Since u n A 1 f , we define the rank for A 1 f obtained by the Greedy Rank-One Update Algorithms (8) and (9) as
rank G ( A 1 f ) = if { j 1 : y j = 0 } = , min { j 1 : y j = 0 } 1 otherwise .
The following theorem (see Theorem 1 in [8]) gives the convergence of the Greedy Rank-One Update Approximation for solving linear systems with full rank matrix.
Theorem 1.
Let f R N 1 N 2 N d and A GL ( N 1 N 2 N d ) . Then, by using the iterative scheme (8) and (9), we obtain that the sequence { r n 2 } n = 0 rank G ( A 1 f ) , is strictly decreasing and
A 1 f = lim n u n = j = 0 rank G ( A 1 f ) y j .
Moreover, the rate of convergence is given by
r n 2 r 0 2 = j = 1 n sin θ j
for 1 n rank G ( A 1 f ) where
θ j = arccos r j 1 , A y j r j 1 2 A y j 2 ( 0 , π / 2 )
for 1 j n .
From (10), we obtain that, if rank G ( A 1 f ) < , then r n 2 = 0 for all n > rank G ( A 1 f ) . Thus, the  above theorem allows for us to construct a procedure, which we give in the pseudo-code form in Algorithm 1, under the assumption that we have a numerical method in order to find a y solving (7) (see the step 5 in Algorithm 1) and that we introduce below.
Algorithm 1 Greedy Rank-One Update
1: procedure GROU ( f , A , ε , tol , rank _ max )
2:     r 0 = f
3:     u = 0
4:    for i = 0 , 1 , 2 , , rank _ max do
5:         y = procedure ( min x S 1 J A , r i ( x ) )
6:         r i + 1 = r i A y
7:         u u + y
8:        if r i + 1 2 < ε or | r i + 1 2 r i 2 | < tol then goto 13
9:        end if
10:    end for
11:    return u and r rank _ max 2 .
12:    break
13:    return u and r i + 1 2
14: end procedure

3. A Geometric Approach to the PGD

In this section, to study the procedure given in the line 5 of the Algorithm 1, we introduce a smooth manifold. To this end, introduce first the set of tensors of fixed rank-one in the tensor space R N 1 N d = j = 1 d R N j defined as
M N 1 N d = u j = 1 d R N j : u = λ j = 1 d u j , λ R * , u j R N j \ { 0 } , 1 j d
where R * = R \ { 0 } . Observe that the set
S 1 = { 0 } M N 1 N d .
Then, our first result is the following theorem of the alternative.
Theorem 2.
Let A GL ( N 1 N 2 N d ) and f R N 1 N 2 N d . Either C ( A , f ) M N 1 N d , or 0 C ( A , f ) but not both.
Proof. 
Assume that 0 C ( A , f ) and that there exists u C ( A , f ) M N 1 N d . Since we can write
J A , f ( u ) = 1 2 f 2 2 f T A u + 1 2 A u 2 2
and J A , f ( 0 ) = J A , f ( u ) = 1 2 f 2 2 J A , f ( u ) holds for all u M N 1 N d , we have
f T A u = 1 2 A u 2 2 > 0 .
Now, consider the map f : R R defined as
f ( λ ) : = J A , f ( λ u ) = 1 2 f 2 2 λ f T A u + λ 2 2 A u 2 2 = 1 2 f 2 2 f T A u λ + f T A u λ 2 .
Then, f ( 0 ) = J A , f ( 0 ) = J A , f ( u ) = f ( 1 ) f ( λ ) for all λ R holds. Observe that
f ( λ ) = f T A u + 2 f T A u λ
and hence the map f has a global minimum for λ = 1 2 , a contradiction. □
The main consequence of this result is the following. It says that the output of the procedure given in step 5 in Algorithm 1 always remains in the set M N 1 N d before it gives us the final output.
Corollary 1.
Let f R N 1 N 2 N d and A GL ( N 1 N 2 N d ) such that A 1 f S 1 . Then, rank G ( A 1 ) = i 1 for some i > 1 if and only if C ( A , r i ) = { 0 } , and hence C ( A , r j ) M N 1 N d for all j < i 1 .
As a consequence of the above corollary, the situation of interest is when given A GL ( N 1 N 2 N d ) and f R N 1 N 2 N d , we have that C ( A , f ) M N 1 N d . Thus, in order to study the vectors in C ( A , f ) , we need to characterize the structure of the critical points of the map u f A u restricted to the set M N 1 N d . To see this in the next section, we provide to M N 1 N d of a structure of smooth manifold.

4. The Set of Tensors of Fixed Rank-One as a Smooth Manifold

Along this paper, we will consider a manifold as a pair ( M , A ) , where M is a subset of some finite-dimensional vector space V and A is an atlas representing the local coordinate system of M . We recall the definition of an atlas associated with a set M .
Definition 1.
Let M be a set. An atlas of class C p ( p 0 ) or analytic on M is a family of charts with some indexing set A , namely { ( U α , φ α ) : α A } , having the following properties (see [14]):
AT1 
{ U α } α A is a covering of M , that is, U α M for all α A and α A U α = M .
AT2 
For each α A , ( U α , φ α ) stands for a bijection φ α : U α W α of U α onto an open set W α of a finite dimensional normed space ( X α , · α ) , and for any α and β the set φ α ( U α U β ) is open in X α .
AT3 
Finally, if we let U α U β = U α , β and φ α ( U α , β ) = U α , β , the transition mapping φ β φ α 1 : U α , β U β , α is a diffeomorphism of class C p ( p 0 ) or analytic.
Observe that the condition of an open covering is not used, see [14]. Moreover, in AT2, we do not require that the normed spaces to be the same for all indices α , or even to be isomorphic. If X α is linearly isomorphic to some finite dimensional normed space X for all α , we have the following definition.
Definition 2.
Let M be a set and X be a finite dimensional normed space. We say that M is a C p (respectively, analytic) manifold modelled on X if there exists an atlas of class C p (respectively, analytic) over M with X α linearly isomorphic to X for all α A .
Since different atlases can give the same manifold, we say that two atlases are compatible if each chart of one atlas is compatible with the charts of the other atlas in the sense of AT3. One verifies that the relation of compatibility between atlases is an equivalence relation.
Definition 3.
An equivalence class of atlases of class C p on M , also denoted by A , is said to define a structure of a C p -manifold on M , and hence we say that ( M , A ) is a finite dimensional manifold. In a similar way, if an equivalence class of atlases is given by analytic maps, then we say that ( M , A ) is an analytic finite dimensional manifold.
For each u = λ j = 1 d u j M N 1 N d , we construct a local chart as follows. Let be
span { u j } = { v j R N j : v j T · u j = 0 } ,
the orthogonal complement of the linear space span { u j } for 1 j d . Let us consider the set
U u : = span { u 1 } × × span { u d } × R * ,
which is an open and dense set of the finite-dimensional vector space
X u : = span { u 1 } × × span { u d } × R .
Observe that the vector space X u is linearly isomorphic to the vector space R N 1 1 × × R N d 1 × R for all u M N 1 N d .
Now, we introduce the set
U ( u ) : = u M N 1 N d : u = β j = 1 d ( u j + w j ) , w j span { u j } , 1 j d β R *
in M N 1 N d for which we can construct a natural bijection:
φ u : U ( u ) U u , u ( w 1 , , w d , β ) u = β j = 1 d ( u j + w j ) .
Then, we can state the following result.
Theorem 3.
The set A N 1 N d = { ( U ( u ) , φ u ) : u M N 1 N d } is an atlas for M N 1 N d and hence ( M N 1 N d , A N 1 N d ) is a C -manifold modelled on
R N 1 1 × × R N d 1 × R R log 2 ( 2 N 1 N d ) d + 1 .
Proof. 
Clearly, AT1 holds. To prove AT2 and AT3, let us consider u , u M N 1 N d be such that U ( u ) U ( u ) . Without loss of generality we may assume that u = λ j = 1 d u i and u = λ j = 1 d u i , where u i i , 2 = u i i , 2 = 1 for 1 i d . Then, for each z U ( u ) U ( u ) , there exists a unique ( w 1 , , w d , β ) U u and a unique ( w 1 , , w d , β ) U u such that
z = β j = 1 d ( u j + w j ) = β j = 1 d ( u j + w j ) .
Since
span { u j + w j } = span { u j + w j }
holds, there exists a unique λ j R * such that
u j + w j = λ j ( u j + w j ) for 1 j d .
Thus, multiplying (13) on the left side by u j T and, by using that u j T u j = 1 and u j T w j = 0 , we obtain
λ j = 1 u j T u j + u j T w j for 1 j d .
Hence,
w j = u j + w j u j T u j + u j T w j u j
defines a C -function from the open set V j = { w j span { u j } : u j T u j + u j T w j 0 } to span { u j } for each 1 j d . Moreover,
1 + w j j , 2 2 = u j + w j u j T u j + u j T w j j , 2 2
holds for 1 j d . Observe that z can be written as
z = β j = 1 d ( u j + w j ) = β j = 1 d 1 + w j j , 2 2 j = 1 d z j ,
where
z j : = 1 1 + w j j , 2 2 u j + 1 1 + w j j , 2 2 w j
has norm one for 1 j d . In addition,
z = β j = 1 d 1 + w j j , 2 2 j = 1 d z j ,
where
z j : = 1 1 + w j j , 2 2 u j + 1 1 + w j j , 2 2 w j
has norm one for 1 j d . Thus,
β = β j = 1 d 1 + w j j , 2 2 j = 1 d 1 + w j j , 2 2 = β j = 1 d 1 + w j j , 2 2 j = 1 d u j + w j u j T u j + u j T w j j , 2
clearly defines a C -function from the open set V 1 × × V d × R * U u to U u . Finally, we conclude that
φ u ( U ( u ) U ( u ) ) = V 1 × × V d × R * U u ,
the map
φ u φ u 1 : φ u ( U ( u ) U ( u ) ) φ u ( U ( u ) U ( u ) )
is given by
( w 1 , , w d , β ) u 1 + w 1 u 1 T u 1 + u 1 T w 1 u 1 , , u d + w d u d T u d + u d T w d u d , β j = 1 d 1 + w j j , 2 2 j = 1 d u j + w j u j T u j + u j T w j j , 2 ,
and it is C . This follows AT2, AT3 and concludes the proof of the theorem. □
The construction of M N 1 N d as an algebraic variety is well-known (see, for example, [15]). More recently, in [16], a structure of smooth manifold is given, in the framework of Banach spaces, for the set of tensors of fixed rank-one. Following [16], it can be shown that the manifold M N 1 N d is also a principal bundle as follows. Consider the Grassmann manifold of one-dimensional, subspaces of R N j , denoted by G 1 ( R N j ) , for 1 j d and define the surjective map
π : M N 1 N d G 1 ( R N 1 ) × × G 1 ( R N d ) , u = λ j = 1 d u j ( span { u j } ) j = 1 d
Then, for each u = λ j = 1 d u j M N 1 N d , it holds that
π 1 ( π ( u ) ) = j = 1 d span { u j } \ { 0 } = span { u } \ { 0 } R * .
Consequently, M N 1 N d is also a principal bundle with base space G 1 ( R N 1 ) × × G 1 ( R N d ) and fibre R * . It allows for decomposing the tangent space at u M N 1 N d , denoted T u M N 1 N d into the vertical and horizontal spaces:
T u M N 1 N d = X u = H u M N 1 N d + V u M N 1 N d ,
where
H u M N 1 N d . : = span { u 1 } × × span { u d }
and V u M N 1 N d : = R .

5. On the First Order Optimality Conditions for the PGD

The goal of this section is given A GL ( N 1 N 2 N d ) and f R N 1 N 2 N d characterize the points in the manifold M N 1 N d satisfying the first order optimality conditions of the problem
min z M N 1 N d J A , f ( z ) .
Recall that the map J A , f is defined in the whole ambient space R N 1 R N d . We will denote its derivative at u R N 1 R N d by J A , f ( u ) = ( f A u ) T , which is a bounded linear map from R N 1 R N d to R . From Theorem 3, we known that M N 1 N d is a C -manifold and hence it allows us to write the constrained map J A , f | M N 1 N d as follows. Since
J A , f : R N 1 R N d R ,
and M N 1 N d R N 1 R N d , we can take into account the standard inclusion map
i : M N 1 N d R N 1 R N d , z z
in order to write J A , f | M N 1 N d = ( J A , f i ) .
Definition 4.
We say that u M N 1 N d is a critical point for J A , f in M N 1 N d if
T u J A , f ( v ) : = [ D ( J A , f φ u 1 ) ( φ u ( u ) ) ] ( v ) = 0
holds for all v T u M N 1 N d . Clearly, if u is an extremal point for J A , f in M N 1 N d , then it is also a critical point for J A , f in M N 1 N d .
Observe that we can write
J A , f φ u 1 = J A , f ( i φ u 1 ) ,
where on the left side of the equality, we consider J A , f over M N 1 N d , whereas, in the right one, J A , f is considered defined over the whole space. Thus, by using the chain rule, we have
D ( J A , f φ u 1 ) ( φ u ( u ) ) = D ( J A , f ( i φ u 1 ) ) ( φ u ( u ) ) = J A , f ( u ) D ( i φ u 1 ) ( φ u ( u ) ) ,
that is,
T u J A , f = J A , f ( u ) T u i = ( f A u ) T T u i .
In order to compute (15), we consider first the standard inclusion map i that in local coordinates in a neighbourhood of u = λ j = 1 d u j M N 1 N d looks like
( i φ u 1 ) ( w 1 , , w d , β ) = β j = 1 w ( u j + w j ) .
Here, ( i φ u 1 ) : U u R N 1 R N d . Hence, its derivative as a morphism (a map between manifolds)
T u i = D ( i φ u 1 ) ( 0 , , 0 , λ ) : T u M N 1 N d R N 1 R N d ,
is given by
T u i ( β , w 1 , , w d ) = j = 1 d w j u [ j ] + β λ u ,
where
w j u [ j ] : = λ u 1 u j 1 w j u j + 1 u d .
for 1 j d . From (15), we have that u * = λ * j = 1 d u i * is a critical point for J A , f in M N 1 N d if and only if
( f A u * ) T T u * i ( w 1 , , w d , β ) = 0 , holds for all ( w 1 , , w d , β ) T u * M N 1 N d ,
that is, it is equivalent to state that
( f A u * ) T j = 1 d w j u [ j ] * + β λ u * = 0 , holds for all ( w 1 , , w d , β ) T u * M N 1 N d .
Now, we will prove the following result that characterizes the set of critical points for J A , f in M N 1 N d as the stationary points of a vector field in R N 1 × × R N d , and it is the main result of this paper.
Theorem 4.
Given A GL ( N 1 N 2 N d ) and f R N 1 N 2 N d . Then, u * = j = 1 d u i * M N 1 N d is a critical point for J A , f in M N 1 N d if and only if X A , f ( u 1 * , , u d * ) = 0 , where X A , f is a vector field in M N 1 N d given by
X A , f ( u 1 , , u d ) = ( y 1 , , y d ) .
Here,
y j : = ( u 1 u j 1 i d R N j u j + 1 u d ) T ( f A ( u 1 u d ) )
for 1 j d .
Proof. 
Since u * = j = 1 d u i * M N 1 N d is a critical point for J A , f in M N 1 N d if and only if (17) holds for ( w 1 , , w d , β ) T u * M N 1 N d . Then, (17) is equivalent to prove that
( f A u * ) T u * = 0 ,
and
( f A u * ) T ( w j u [ j ] * ) = 0 holds for all w j span { u j } and 1 j d .
Since w j u [ j ] * = u * when w j = u j * for 1 j d , we conclude that u * = j = 1 d u i * M N 1 N d is a critical point for J A , f in M N 1 N d if and only if
( f A u * ) T ( w j u [ j ] * ) = 0 holds for all w j R N j and 1 j d .
Writing the equality in (20) as
( f A u 1 * u d * ) T ( u 1 * u j 1 * i d R N j u j + 1 * u d * ) w j = 0 ,
we conclude that it is equivalent to state that
( f A u 1 * u d * ) T ( u 1 * u j 1 * i d R N j u j + 1 * u d * ) = 0 T
for 1 j d . This concludes the proof of theorem. □
Remark 1.
From (17), we can deduce that u * = j = 1 d u i * is a critical point for J A , f in M N 1 N d if and only if
( f A u * ) T j = 1 d w j u [ j ] * = 0 holds for all ( w 1 , , w d ) R N 1 × × R N d .
Statement (21) was first introduced in [2,3] as a step to enrich the approximation basis in the PGD algorithm. From the proof of Theorem 4, we have that (21) is also equivalent to (18) and (19).
There are several strategies in the literature that can be used to solve X A , f ( u ) = 0 . The first one, closely related to the one used in [2,3], is to find a fixed point of the map
F A , f ( u 1 , , u d ) : = X A , f ( u 1 , , u d ) ( u 1 , , u d )
One of the most popular numerical strategies to compute an approximated value of u * such that X A , f ( u * ) = 0 is founded under the following argument. Since the y 1 is equal to
( i d R N 1 u 2 u d ) T ( f A u 1 u d ) = ( i d R N 1 u 2 u d ) T f ( i d R N 1 u 2 u d ) T A ( u 1 u d ) = ( i d R N 1 u 2 u d ) T f ( i d R N 1 u 2 u d ) T A ( i d R N 1 u 2 u d ) u 1 ,
we can choose randomly some vectors u j R N j for 2 j d and then try to solve the linear system
( i d R N 1 u 2 u d ) T A ( i d R N 1 u 2 u d ) u 1 = ( i d R N 1 u 2 u d ) T f .
By using Least Squares, we can compute a u 1 , and next we can proceed in a similar way iteratively with each of the other components y 2 , , y d as follows. After the step i 1 , we know that u 1 , u 2 , , u i . Then, by choosing randomly u i + 1 , u d , we can solve, by using Least Squares, the linear system
( u 1 u i i d R N i u i + 1 u d ) T A ( i d R N 1 u 2 u d ) u i = ( u 1 u i i d R N i u i + 1 u d ) T f ,
to compute u i . We can proceed cyclically until f A ( u 1 u d ) < ε holds. This strategy is known by the name of the Alternating Least Squares (ALS).
From (17), we have that u * is a critical point of J A , f in M N 1 N d if and only if
( f A u * ) T u * i ( T u * M N 1 N d )
holds, that is, the residual at u * is orthogonal to linear subspace T u * i ( T u * M N 1 N d ) . Observe that w j u [ j ] V j ( u ) , where
V j ( u ) : = span { u 1 } span { u j 1 } span { u j } span { u j + 1 } span { u d }
is an N j 1 -dimensional subspace of R N 1 R N d for 1 j d . Then, the next result explicitly describes the linear subspace T u * i ( T u * M N 1 N d ) .
Proposition 1.
The linear map T u i : T u M N 1 N d R N 1 R N d is injective. Moreover,
T u i ( T u M N 1 N d ) = j = 1 d V j ( u ) span { u } T u M N 1 N d .
Proof. 
Assume that
T u i ( β , w 1 , , w d ) = j = 1 d w j u [ j ] + β λ u = 0 .
Since u ( w j u [ j ] ) for 1 j d and ( w i u [ j ] ) ( w j u [ j ] ) for all i j , then
β = 0 and w i u [ j ] = 0 for 1 j d .
Since u [ j ] 0 , then w j = 0 for 1 j d . This follows the first statement. To prove the second one, we remark that the inclusion
T u i ( T u M N 1 N d ) j = 1 d V j ( u ) span { u }
between both subspaces is trivial. Clearly,
dim T u i ( T u M N 1 N d ) = dim j = 1 d V j ( u ) span { u }
holds and then the second statement is also proved. □

6. Conclusions

In this paper, a geometric upgrade of the PGD algorithm is proposed. To this end, we endow a set of tensors of fixed rank-one with a smooth manifold structure. This construction provides the solutions of a non-convex optimization problem by using a set of stationary points of a vector field. This new perspective allows for the characterization of the behaviour of the solutions of the Greedy Rank One Algorithm which could not be achieved with the ALS strategy used in [8].

Author Contributions

Investigation, writing and revision: A.F., L.H., N.M., M.C.M. and E.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the GVA/2019/124 grant from Generalitat Valenciana and by the RTI2018-093521-B-C32 grant from the Ministerio de Ciencia, Innovación y Universidades.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Falcó, A.; Montés, N.; Chinesta, F.; Hilario, L.; Mora, M.C. On the existence of a Progressive Variational Vademecum based on the Proper Generalized Decomposition for a Class of Elliptic Parametrized Problems. J. Comput. Appl. Math. 2018, 330, 1093–1107. [Google Scholar] [CrossRef]
  2. Ammar, A.; Mokdad, B.; Chinesta, F.; Keunings, R. A new family of solvers for some classes of multidimensional partial differential equations encountered in Kinetic Theory modelling Complex Fluids. J. Non-Newton. Fluid Mech. 2006, 139, 153–176. [Google Scholar] [CrossRef] [Green Version]
  3. Ammar, A.; Mokdad, B.; Chinesta, F.; Keunings, R. A new family of solvers for some classes of multidimensional partial differential equations encountered in Kinetic Theory modelling Complex Fluids. Part II: Transient Simulations using Space-Time Separated Representations. J. Non–Newton. Fluid Mech. 2007, 144, 98–121. [Google Scholar] [CrossRef] [Green Version]
  4. Falcó, A.; Nouy, A. A Proper Generalized Decomposition for the solution of elliptic problems in abstract form by using a functional Eckart- Young approach. J. Math. Anal. Appl. 2011, 376, 469–480. [Google Scholar] [CrossRef] [Green Version]
  5. Chinesta, F.; Leygue, A.; Bordeu, F.; Aguado, J.V.; Cueto, E.; González, D.; Alfaro, I.; Ammar, A.; Huerta, A. PGD-Based computational vademecum for efficient Design, Optimization and Control. Arch. Comput. Methods Eng. 2013, 20, 31–49. [Google Scholar] [CrossRef]
  6. Chinesta, F.; Ladevéze, P. Separated Representations and PGD-Based Model Reduction; Springer International Centre for Mechanical Sciences: Cham, Switzerland, 2014. [Google Scholar]
  7. Falcó, A.; Nouy, A. Proper Generalized Decomposition for Nonlinear Convex Problems in Tensor Banach Spaces. Numer. Math. 2012, 121, 503–530. [Google Scholar] [CrossRef] [Green Version]
  8. Ammar, A.; Chinesta, F.; Falcó, A. On the convergence of a greedy rank-one update algorithm for a class of linear systems. Arch. Comput. Methods Eng. 2010, 17, 473–486. [Google Scholar] [CrossRef]
  9. El Hamidi, A.; Osman, H.; Jazar, M. On the convergence of alternating minimization methods in variational PGD. Comput. Optim. Appl. 2017, 68, 455–472. [Google Scholar] [CrossRef]
  10. Graham, A. Kronecker Products and Matrix Calculus with Applications; John Wiley: Hoboken, NJ, USA, 1981. [Google Scholar]
  11. Van Loan, C.F. The ubiquitous Kronecker product. J. Comput. Appl. Math. 2000, 123, 85–100. [Google Scholar] [CrossRef] [Green Version]
  12. Beylkin, G.; Mohlenkamp, M.J. Algorithms for Numerical Analysis in High Dimensions. SIAM J. Sci. Comput. 2005, 26, 2133–2159. [Google Scholar] [CrossRef] [Green Version]
  13. de Silva, V.; Lim, L.-H. Tensor Rank and Ill-posedness of the best low-rank approximation problem. SIAM J. Matrix Anal. Appl. 2008, 30, 1084–1127. [Google Scholar] [CrossRef]
  14. Lang, S. Differential and Riemannian Manifolds; Graduate Texts in Mathematics; Springer: Berlin/Heidelberg, Germany, 1995; Volume 160. [Google Scholar]
  15. Landsberg, J.M. Tensor: Geometry and Applications. Graduate Studies in Mathematics; American Mathematical Society: Providence, Rhode Island, 2012; Volume 128. [Google Scholar]
  16. Falcó, A.; Hackbusch, W.; Nouy, A. On the Dirac-Frenkel Variational Principal on Tensor Banach Spaces. Found. Comput. Math. 2019, 19, 159–204. [Google Scholar] [CrossRef] [Green Version]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Falcó, A.; Hilario, L.; Montés, N.; Mora, M.C.; Nadal, E. Towards a Vector Field Based Approach to the Proper Generalized Decomposition (PGD). Mathematics 2021, 9, 34. https://doi.org/10.3390/math9010034

AMA Style

Falcó A, Hilario L, Montés N, Mora MC, Nadal E. Towards a Vector Field Based Approach to the Proper Generalized Decomposition (PGD). Mathematics. 2021; 9(1):34. https://doi.org/10.3390/math9010034

Chicago/Turabian Style

Falcó, Antonio, Lucía Hilario, Nicolás Montés, Marta C. Mora, and Enrique Nadal. 2021. "Towards a Vector Field Based Approach to the Proper Generalized Decomposition (PGD)" Mathematics 9, no. 1: 34. https://doi.org/10.3390/math9010034

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop