Next Article in Journal
Rain Rendering and Construction of Rain Vehicle Color-24 Dataset
Next Article in Special Issue
An Evolving Fuzzy Neural Network Based on Or-Type Logic Neurons for Identifying and Extracting Knowledge in Auction Fraud
Previous Article in Journal
New Generalized Contractions by Employing Two Control Functions and Coupled Fixed-Point Theorems with Applications
Previous Article in Special Issue
Describing Linguistic Vagueness of Evaluative Expressions Using Fuzzy Natural Logic and Linguistic Constraints
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Preimage Problem Inspired by the F-Transform

by
Jiří Janeček
1,*,† and
Irina Perfilieva
2,*,†
1
Department of Mathematics, Faculty of Science, University of Ostrava, 30. dubna 22, 701 03 Ostrava, Czech Republic
2
Institute for Research and Applications of Fuzzy Modeling, University of Ostrava, 30. dubna 22, 701 03 Ostrava, Czech Republic
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Mathematics 2022, 10(17), 3209; https://doi.org/10.3390/math10173209
Submission received: 10 June 2022 / Revised: 23 August 2022 / Accepted: 1 September 2022 / Published: 5 September 2022
(This article belongs to the Special Issue Fuzzy Natural Logic in IFSA-EUSFLAT 2021)

Abstract

:
In this article, we focus on discrete data processing. We propose to use the concept of closeness, which is less restrictive than a metric, to describe a certain relationship between objects. We establish a fuzzy partition of a given set of objects in a way that admits a closeness space to emerge. The fuzzy (F-) transform is a tool that maps objects with common characteristics to the same discrete image—the direct F-transform. We are interested in the inverse (preimage) problem: How can we describe the class of all functions mapped onto the same direct F-transform? In this manuscript, we focus on this preimage problem, formulated accordingly. Its solution is presented from three different points of view and shows which functions belong to the same class determined by a given image (by the direct F-transform). Conditions under which a solution to the preimage problem is given by the inverse F-transform over the same fuzzy partition, or by transforming a given image using a new system of basic functions, are formulated. The developed theory contributes to a better understanding of ill-posed problems that are typical for machine learning. The appendix contains illustrative numerical examples.

1. Introduction

In this paper, we are focused on spaces with closeness that are characterized using weighted graphs, as in [1], or local neighborhoods of elements as in [2]. Formally, closeness can be described by the graph adjacency matrix that contains weights—a higher value of weight is assigned to the edge that connects closer objects. Those matrices were always square as they contained mutual closeness values of all objects in the dataset.
In our previous research [1,3], the notion of closeness was used in the new approach to dimensionality reduction for which fuzzy transform (F-transform for short, [4]) proved to be useful. Otherwise, the concept of closeness can be observed in multiple contexts and under different names, where it serves auxiliary purposes. For example, in [5], the phase space of a dynamical system is described by a network (graph) where each point represents one state of the system and closeness-describing weights (between two points) are equal to the frequency with which a transition occurred between the two states. There exists a related concept of closeness centrality (aggregation of closeness-describing weights with respect to all neighbors except oneself computed by arithmetic or harmonic mean) that describes the data density around one particular point in a network, used, e.g., in [6]. In the area of image processing (image compression, image segmentation, image retrieval, e.g., [7], etc.), the related concept of proximity space is widely used. It originated from [8] and then evolved—currently, it is used, e.g., to describe the similarity between pixels or patches.
In this research, we consider closeness determined by a fuzzy partition of a universe of discourse. This particular space structure is used in the theory of F-transforms [4,9]. The notion of F-transform was introduced in [4], where we explained modelling with fuzzy IF-THEN rules as a specific transformation. From this point of view, F-transform bridges fuzzy modelling and the theory of linear (in particular, integral) transforms. The generalization to a higher degree version was proposed in [9]. Generally speaking, F-transform performs a transformation of the original universe of functions into a universe of their skeleton models (vectors or matrices of F-transform components), for which further computations are easier. In [4], the approximation property of F-transform was described, and in [10], the effect of the shapes of basic functions on the approximation quality was demonstrated. F-transform has many other useful properties and great potential in various applications, such as special numerical methods as well as solutions to ordinary and partial differential equations with fuzzy initial conditions, mining dependencies from numerical data, signal processing, compression and decompression of images [11] and image fusion.
Among the recent applications of F-transform, we refer to an image classification problem, in which data is scarce [12], a numerical solution to fuzzy integral equations [13] and improving the JPEG compression algorithm in the cases where a high ratio compression is required [11].
Similarly to F-transform, which, in its direct phase, aggregates a large number of functional values of all points into a small number of components associated with selected nodes, we consider that within our (possibly large) dataset, there are several selected points (nodes) with known closeness values with respect to all the other points and that closeness among other points is undefined, which gives rise to a rectangular closeness matrix that describes the closeness space. Nodes can be thought of as the most prominent data points because the dataset can be sparsely represented as a collection of their neighborhoods.
Following the methodology of F-transform, we aim to demonstrate that a partial knowledge about the mutual position of all points in the dataset is sufficient for a successful retrieval of a discrete function (up to non-significant differences that distinguish particular functions from each other within the same equivalence class) from its F-transform components. Thus, we present a not just theoretical tool for demarcating the set of such functions that can be mutually replaced without losing an important feature in the space.
This allows us to express similarity between functions defined on spaces with closeness. In specific cases, a representative function is given by the inverse F-transform. Moreover, we contribute to the general theory of F-transforms in aspects described in this section. The task of reconstructing an element of an input space (or finding an approximate solution if it does not exist) based on (a lower-dimensional representation of) an element of a feature space has been studied in various machine learning domains, e.g., in kernel-based methods (e.g., [14] where the nonlinear dimensionality reduction is performed using the kernel trick on an input image and the task is to recover the image from its denoised version in the feature space, or [15] where algebraic tools based on the simple relationship of distances in the input space and in the feature space induced by a kernel are used to find the preimage of a feature vector) and graph-based methods (where the task is to recover structured data from a single point of a feature space, e.g., to find a representative graph based on features of a found data cluster). In these applications, the concept of preimage problem was already used. This paper is focused on finding all functions that share a set of features (given by the F-transform components) utilizing the above-described concept. We do not consider that these features might be noisy. On the other hand, in [16], we used a similar initial setting of the input space and we assumed that the input signal (function placed in this space) is noisy in a certain sense. The signal was processed in the same manner (based on closeness) and we proposed how to denoise the signal by finding the appropriate closeness parameters. The inverse F-transform is known to reduce the noise of the input signal (provided a suitable fuzzy partition) as was in the continuous case shown, e.g., in [17].
As explained in [18], preimage problems are useful not only in (pattern) denoising and Kernel Dependency Estimation but also in signal compression, where the kernel technique serves as an encoder and the preimage technique as a decoder. Moreover, in [18], a technique to learn the preimage without the need to solve a difficult nonlinear optimization problem is presented. Additionally, in [15], the authors consider that a noisy pattern is mapped to a feature space using the Kernel Principal Component Analysis and then the approximate preimage is found using their technique. To summarize the main stream of preimage problems in machine learning, we can say that a nonlinear optimization, nonlinear iteration, or learning method is used to find a function (the space of all possible functions comprises an input space), s.t. its image in a specified direct mapping has the minimal squared distance from the given point in a feature space. The existence of a precise solution is considered to be a coincidence. The purposes of computing the preimage in machine learning are various (to denoise a pattern, reconstruct a signal, compress an image, find a representative example of a data cluster, or learn a general mapping between the input space and the feature space) but all of them are highly application-oriented, and, hence, the initial settings are adjusted accordingly. This limits their transferability to another task.
In contrast, our approach pays close attention to the structure of the input space and this structure is quite flexible. We are focused on finding the precise solution (we ensure that it always exists) to the special case of the preimage problem. Using the assumption of the finiteness of the universe and endowing the input space with a fuzzy partition and the corresponding closeness, we solve it by the means of linear algebra. The consequent mapping forms a special case of direct mappings mentioned above. Since our theoretical work is not aimed at solving any specific task, we compute the whole class (usually infinite) from which we do not need to choose a particular solution (all of them are equivalent).
In Section 2, we give the details about the notion of closeness considered in this article. After that, the fuzzy partition and transform are recalled and modified for the discrete case. We show that a certain fuzzy partitioned space (space with a fuzzy partition) is a special case of the space with closeness. Section 3 discusses the formulation of the main topic of the paper—called preimage problem—that takes place in the space introduced in Section 2. Preimage problem belongs to basic problems in algebra (e.g., in [19]) where it is considered between certain two structured spaces (we propose to use the space with closeness and the space of scaled F-transform components). The solution to this problem is described in three different ways. Theorem 1 is based on a commonly used solution to the set of algebraic equations. In Section 4, we examine the conditions when the inverse F-transform forms the solution to this problem. In Section 5, we propose a technique consisting of checking whether a solution to the preimage problem can be obtained by a certain transformation of the given element of the space of scaled F-transform components given by a new set of fuzzy partition units. Appendix A provides the numerical examples related to four preceding sections, except for the Conclusions (Section 6).

Relationship of Closeness, Metric and Similarity

A space with closeness ( X , w ) is more general than a metric space ( X , ρ ) as they share the axioms of symmetry and non-negativity but metric is more restrictive—it requires two additional axioms ( ρ ( x , y ) = 0 iff x = y and the triangle inequality ρ ( x , y ) ρ ( x , z ) + ρ ( y , z ) for all x , y , z X ). As closeness is a relaxed version of a reciprocal distance metric, it is more suitable for the description of data with graph structure (where the triangle inequality is generally non-enforceable). Another example of a context where closeness is more suitable is formed by the data that is assumed to lie on a topological manifold as there is no straightforward way how to establish a metric there. Similarly to metric, closeness encodes mutual relations between data points.
There is a close connection between closeness and similarity measures—each of these concepts applies in different contexts. An algebraic, fuzzy, or probabilistic background is usually assumed as a similarity requirement. There is no standard axiomatization of its definition. Similarly to closeness, a similarity value is higher for more similar objects (that are, e.g., more correlated, have a higher intersection, smaller cross-entropy, or belong to the same cluster), its values can be negative, it can be non-symmetric, etc. Similarity spaces emerging in various applied fields are even more general than spaces with closeness. A solution to a particular type of problem is often associated with a certain type of similarity measure.
This explains why we introduce the notion of closeness and why we give it a preference over that of metric (too-restrictive axioms) and similarity (too-loose axiomatization). This good trade-off allows us to express various established concepts, e.g., of graph theory (edge weights used, e.g., in the dimensionality reduction technique of Laplacian eigenmaps [2]), fuzzy logic (values of biresiduum of truth values of two formulae), clustering problems (class membership degrees in, e.g., KNN clustering algorithm), etc., in terms of closeness and incorporate a theoretical apparatus upon that. Therefore, as the closeness values can be extracted from the data in various ways, the closeness space creates a platform that enables both data-driven and model-driven approaches to be utilized.

2. Preliminaries

In this section, we describe the mathematical background by giving some basic definitions and properties. Throughout the manuscript, all settings are discrete.

2.1. Closeness Space

We are interested in the notion of closeness on finite sets that, as described below, agrees with [1,2]. To emphasize its general applicability, we express it using the language of classical set theory. We show that it can be also expressed in the languages of fuzzy set theory and graph theory.
Definition 1 (Closeness).
Let X = { x i | i = 1 , , L } be a set, L N , then closeness on X is any symmetric, non-negative function w : X 2 R , where symmetry means x , y X : w ( x , y ) = w ( y , x ) and X 2 = X × X = { ( x , y ) | x , y X } is the Cartesian product of X with itself. The pair ( X , w ) is called a space with closeness or simply a closeness space.
We provide the following semantic interpretation of closeness: objects that are closer in a certain sense (as explained in the Introduction, in some cases, metric cannot be defined) have a greater closeness value (with respect to natural order on R ) than the less-close ones. Non-close objects have a closeness value equal to zero, and, vice versa, if the closeness value of a pair of objects is greater, we consider these objects closer in this sense than a pair of objects with a smaller closeness value. The closeness value hence quantifies a certain quality of mutual alikeness between objects.
Example 1.
Let X = { 0 , 10 , 20 , 30 , 40 , 50 } and for all x i , x j X : w ( x i , x j ) = 1 1 + | x i x j | , then w is closeness on X and ( X , w ) is a space with closeness.
Example 2.
Let ρ be a metric and ( X , ρ ) a metric space. Let moreover, X X and for all x i , x j X , w 1 ( x i , x j ) = 1 1 + ρ ( x i , x j ) and w 2 ( x i , x j ) = e ρ ( x i , x j ) hold true. Then, ( X , w 1 ) and ( X , w 2 ) are closeness spaces.
Definition 2 ((Weakly) Reflexive Closeness).
Let X = { x i | i = 1 , , L } be a set, L N , and w : X 2 R be closeness on X. Then, w is weakly reflexive closeness on X if there exists a positive real number w ¯ , s.t. x , y X : w ( x , y ) w ( x , x ) = w ¯ . If, moreover, w ¯ = 1 , w is called reflexive closeness on X. The pair ( X , w ) is then a space with (weakly) reflexive closeness or simply a (weakly) reflexive closeness space.
Remark 1.
In this paper, we consider only finite closeness domain X, hence there always exists a constant m = max { w ( x i , x j ) | ( x i , x j ) X 2 } , in case 1 < m < + , it would be more appropriate to define strong reflexivity by w ¯ = m but later, we work with closeness w : X 2 [ 0 , 1 ] .
Example 2 contains two reflexive closeness-defining functions. Examples of weakly reflexive closenesses derived from an arbitrary metric ρ include 1 2 + ρ ( x i , x j ) and 1 2 e ρ ( x i , x j ) , both with w ¯ = 1 2 .
Let ( X , w ) be a space with closeness where X = { x i | i = 1 , , L } is a set, L N . Let us denote
x i , x j X : w i j = w ( x i , x j ) ,
and consider a matrix that contains the closeness values
W = ( w i j ) i , j = 1 L R L × L ,
then the matrix W is called the matrix of closeness on X or simply the closeness matrix. Functional values of closeness w for the elements of X 2 can be arranged into a matrix W—the matrix of closeness on X.

2.2. Fuzzy Set, Relation and Partition

Definition 3 (Fuzzy Set).
Let X be a non-empty universe and K : X [ 0 , 1 ] be a function. Fuzzy set K is the set of pairs { ( x , K ( x ) ) | x X } . The function K is a membership function and its functional value K ( x ) is called the membership degree of the element x to K. Fuzzy set is conventionally identified with its membership function. We call it fuzzy subset of X which is denoted by K X . We say that the point x X is covered by the fuzzy set K if K ( x ) > 0 .
Remark 2.
Note that if K : X { 0 , 1 } , then K is a characteristic function of an ordinary set. The membership function K : X [ 0 , 1 ] is a generalization of the characteristic function of the set K X given by a function χ K : X { 0 , 1 } . If the universe of a (fuzzy) set is clear, we can talk about (fuzzy) set K without referring to it.
Definition 4 (Fuzzy Relation).
Let V and X be non-empty sets, then binary fuzzy relation r is any fuzzy subset of V × X , i.e., r V × X .
Hence, the binary fuzzy relation r is identified with a membership function r : V × X [ 0 , 1 ] .
Definition 5 (Convex Fuzzy Set).
Let X R and K X . Then K is a convex fuzzy set, if x c , x d , x e X , x c < x d < x e : K ( x d ) min { K ( x c ) , K ( x e ) } .
Following [4], we recall the notion of fuzzy partition and modify it for a discrete case. Consecutively, we set fuzzy transform in Section 2.6.
Definition 6 (Fuzzy Partition of a Real Interval with Nodes).
Let [ y 0 , y l + 1 ] be a real interval and y 1 , , y l R , s.t. y 0 < y 1 < y 2 < < y l < y l + 1 and 0 < l < + . Then the fuzzy partition of [ y 0 , y l + 1 ] is given by a set of fuzzy sets { A 1 , , A l } , if:
1. 
i = 1 , , l : A i : [ y 0 , y l + 1 ] [ 0 , 1 ] ,
2. 
for all i = 1 , , l : A i is continuous with respect to topologies on [ y 0 , y l + 1 ] , [ 0 , 1 ] R induced by the natural topology on R ,
3. 
i = 1 , , l : A i ( x ) > 0 iff x ( y i 1 , y i + 1 ) ,
4. 
for all i = 1 , , l : A i strictly increases in ( y i 1 , y i ] ,
5. 
for all i = 1 , , l : A i strictly decreases in [ y i , y i + 1 ) .
Elements of the set { A 1 , , A l } are called basic functions (fuzzy partition units). Based on condition 3., A 1 , , A l are associated with points y 1 , , y l , respectively. These points are called nodes of the fuzzy partition.
In Definition 6, conditions 3–5 ensure that each basic function is a convex fuzzy set. Moreover, condition 3 ensures that each basic function covers at least one point. This might not hold in the case of a discrete universe which motivated the following definition.
Definition 7 (Sufficient Density with respect to Fuzzy Partition).
Let X be a set, s.t. X [ y 0 , y l + 1 ] R , and 0 < l < + . Let P = { A 1 , , A l } be a fuzzy partition of the interval [ y 0 , y l + 1 ] . Then X is sufficiently dense with respect to P , if i = 1 , , l x X : A i ( x ) > 0 .
Definition 7 states that a set is sufficiently dense with respect to a given fuzzy partition if every basic function covers at least one point of that set. This notion allows us to define the fuzzy partition of a discrete set.
Definition 8 (Fuzzy Partition of a Discrete Subset of a Real Interval with Nodes).
Let X = { x 1 , , x L } be a set, s.t. X [ y 0 , y l + 1 ] R , and 0 < l L < + . Let P = { A 1 , , A l } be a fuzzy partition of the interval [ y 0 , y l + 1 ] with nodes in Y = { y 1 , , y l } X and X be sufficiently dense with respect to P . Then P = { A 1 , , A l } is the fuzzy partition of X with nodes in Y, if i = 1 , , l : A i = A i | X , i.e., if each basic function is replaced by its restriction on X.
Remark 3.
Definition 8 can be extended in the converse direction: let X = { x 1 , , x L } be a set, s.t. X [ y 0 , y l + 1 ] R , Y = { y 1 , , y l } X and 0 < l L < + . Let P = { A 1 , , A l } , s.t. i = 1 , , l : A i X & A i ( y i ) > 0 . Then P will be also called the fuzzy partition of X with nodes in Y, if there exists a fuzzy partition P = { A 1 , , A l } with nodes in Y, s.t. X is sufficiently dense with respect to P and for each i = 1 , , l , A i is a continuous extension of A i on [ y 0 , y l + 1 ] , i.e., A i [ y 0 , y l + 1 ] , A i is continuous and A i = A i | X .
Below, we show how the space with closeness can be represented by a fuzzy partitioned space and by a weighted graph-structured space that, satisfying certain conditions, are examples of closeness spaces.

2.3. Closeness as Fuzzy Relation and Closeness Given by a Fuzzy Partition

Using the language of fuzzy set theory, we can say that any symmetric fuzzy relation w X 2 , where symmetry means x , y X : w ( x , y ) = w ( y , x ) , is closeness on X.
Note that in the context of fuzzy relations, it holds that for weakly reflexive closeness, we have w ¯ 1 . For w ¯ = 1 , we obtain a special case called reflexive closeness on X.
Lemma 1 (Closeness Given by a Fuzzy Partition – 1).
Let { A 1 , , A l } be a fuzzy partition associated with nodes y 1 , , y l of the set { x 1 , , x L } [ y 0 , y l + 1 ] , s.t. y 0 , , y l + 1 , x 1 , , x L R , 0 < l = L < + and y 0 < y 1 = x 1 < y 2 = x 2 < < y l = x L < y l + 1 . For all i , j = 1 , , l , let it hold that
A i ( y j ) = A j ( y i ) ,
then the function w : { x 1 , , x L } 2 R given by i , j = 1 , , L : w ( x i , x j ) = A i ( x j ) is closeness on { x 1 , , x L } .
Proof. 
The function w is obviously real, bivariate and non-negative. Using the property (3), we have i , j = 1 , , L : w ( x i , x j ) = A i ( x j ) = A i ( y j ) = A j ( y i ) = A j ( x i ) = w ( x j , x i ) , hence w is also symmetric. □
Lemma 1 shows that a fuzzy partition of a subset of a real interval, s.t. its every point is a node and satisfying condition (3), determines closeness on this set.

2.4. Closeness as Weighted Adjacency of Graph Vertices

In this subsection, we show that closeness can be expressed using the language of graph theory. We do not recall the basic concepts of this theory (details can be found, e.g., in [20]). Generally, closeness between any two objects x i and x j in X is described by the values (weights) w ( x i , x j ) of a real, non-negative, symmetric bivariate function w.
Let G ( X , E , W ) be a weighted graph where the set of vertices corresponds to objects in X, the set of edges E X 2 and weights of the edges are given by the adjacency matrix W and in accordance with (2). The objects are non-close if and only if the corresponding vertices are not connected by a direct edge, i.e., there is a fictional edge with zero weight between them. We denote this disconnectedness or non-closeness by which is a classical (meaning non-fuzzy) symmetric binary relation on X, i.e., X 2 :
x i x j w ( x i , x j ) = 0 .
In the other words, the value of closeness determines the existence of an edge.
By defining the values of closeness { w ( x i , x j ) | x i , x j X } , i.e., by defining the values of entries of the closeness matrix, we simultaneously define the values of entries of the adjacency matrix W which uniquely determines a weighted graph G ( X , E , W ) by (1), (2) and (4). That is why the notion of closeness can be also established from the perspective of weighted graphs: let X = { x i | i = 1 , , L } be a set of vertices, L N , then, following the convention given by (4) as E X 2 , any evaluation of the set of all edges X 2 described by a symmetric, non-negative function w : X 2 R is closeness on X.
Below, we join the theory of closeness with a certain type of fuzzy partition to describe the space this paper deals with.

2.5. Initial Assumptions

Let us have a finite set of points on R :
X = { x i | i = 1 , , L } ,
and its non-empty subset
Y X , 0 < | Y | = l L = | X | < + .
Let { A t | t Y } be a fuzzy partition of X with nodes in Y (in the sense of Remark 3), s.t. with each node t Y , we associate a basic function A t : X [ 0 , 1 ] . Let, moreover, (3) be satisfied, i.e., t , s Y : A t ( s ) = A s ( t ) , and let all points in X be covered by at least one basic function, i.e.,
t Y { x X | A t ( x ) > 0 } = X .
Let us define node degrees by
t Y : d t t = x X A t ( x ) .
Following condition 3 in Definition 6, we know that A t ( t ) > 0 , t Y , which implies that each node degree d t t ( 0 , ) . Moreover, based on Definition 7, it means that the set Y is sufficiently dense with respect to the fuzzy partition { A t | t Y } ; hence, the same holds for the universe X.
It is worth noting that instead of working with the closeness on the whole universe X, given by a function on X 2 , we work only with its restriction on Y × X (the values of closeness on ( X Y ) × X are undefined, as explained in the Introduction). In the other words, closeness w : X 2 R is a partial function (partial function f : A B maps each element of a set A to at most one element of the set B, it is a functional binary relation with carrier set being a subset of A × B ) given by
x i , x j X : w ( x i , x j ) = A t ( x j ) t = x i Y undefined otherwise .
This description justifies the selection of nodes within the universe (Semantic description of nodes is also explained in the Introduction).
Even though we will still call the total function w : Y × X R (total function f : A B maps each element of a set A to exactly one element of the set B, it is a classical function) closeness on X and denote the corresponding closeness space as ( X , w ) .
Lemma 2 (Closeness Given by a Fuzzy Partition—2).
Let { A 1 , , A l } be a fuzzy partition associated with nodes y 1 , , y l of the set { x 1 , , x L } [ y 0 , y l + 1 ] , s.t. y 0 , , y l + 1 , x 1 , , x L R , y 0 < y 1 < y 2 < < y l < y l + 1 , y 0 < x 1 < x 2 < < x L < y l + 1 , { y 1 , , y l } { x 1 , , x L } , 0 < l L < + and let condition (3) hold. Let us denote
i = 1 , , L : κ ( i ) = j x i = y j l + 1 y { y 1 , , y l } : x i = y
(note that in case x i = y j , we have y j = y κ ( i ) ). Then the function w : { x 1 , , x L } 2 R given by w ( x i , x j ) = A κ ( i ) ( x j ) , where κ ( i ) l , is closeness on { x 1 , , x L } .
Proof. 
The function w is obviously real, bivariate and non-negative. Using the property (3), we have i , j = 1 , , L , κ ( i ) , κ ( j ) l : w ( x i , x j ) = A κ ( i ) ( x j ) = A κ ( i ) ( y κ ( j ) ) = A κ ( j ) ( y κ ( i ) ) = A κ ( j ) ( x i ) = w ( x j , x i ) , hence w is also symmetric. □
Lemma 2 shows that a fuzzy partition of a subset of a real interval, s.t. it contains all nodes and satisfies condition (3), determines closeness on this set. We will follow Lemma 2, and hence closeness is given by:
w ( t , x ) = A t ( x ) , where t Y , x X .
By that, the closeness space ( X , w ) is specified. The values of closeness w : Y × X R are inserted in the closeness matrix W R l × L and given by
t Y , x X : W ( t , x ) = w ( t , x ) .
It means that by fixing a set of basic functions, we determine values of the closeness-describing function w : Y × X R .
It follows from (4), (6) and (8) that for every point x of the universe X there exists a node t Y , s.t. w ( t , x ) > 0 . It means that every point is connected with at least one node.
Section 2.1 and Section 2.2 covered the basic theory of closeness and the theory of fuzzy partitions, respectively. In the following Section 2.6, we recall the theory of fuzzy transform that we need to describe a fundamental property (the F-transform identity) of the space with closeness that will be used in Section 3 to describe the preimage problem.

2.6. Fuzzy Transform

In Section 2.2, we defined the fuzzy partition and its units, basic functions. Below, we proceed by the establishment of fuzzy transform. The closeness space ( X , w ) serves as a ground on which F-transform is established as its key element, a fuzzy partition, is selected on condition (3) that allows the creation of ( X , w ) based on (8) in Section 2.5.
Definition 9
(Direct F-Transform [4]).Let X be a set, Y X [ y 0 , y l + 1 ] R , 0 < | Y | = l | X | = L < + , P = { A t : X [ 0 , 1 ] | t Y } be a fuzzy partition of X with nodes in Y, s.t. y 0 < x 1 < x 2 < < x L < y l + 1 , and satisfying (3), i.e., t , s Y : A t ( s ) = A s ( t ) . Let d t t be a node degree given by (7) and u : X R be a function where i = 1 , , L : u ( x i ) = u i . Then the direct discrete F-transform of u with respect to P is a vector F [ u ] R l of F-transform components defined by
t Y : F [ u ] t = i = 1 L u i A t ( x i ) i = 1 L A t ( x i ) = i = 1 L u i A t ( x i ) d t t .
Remark 4.
Note that in [4], the direct F-transform is defined only with respect to a fuzzy partition (without closeness) and in that case, the requirement that P satisfies (3) is omitted. This paper, however, assumes a closeness background. That is why for every node t Y , the value
d t t = i = 1 L w t i ,
is also called a node degree which fully agrees with (7). For the corresponding space with closeness ( X , w ) where the closeness values are denoted according to (1), we equivalently define the direct F-transform of u as a vector F [ u ] R l of F-transform components defined by
t Y : F [ u ] t = i = 1 L u i w t i i = 1 L w t i = i = 1 L u i w t i d t t ,
which fully agrees with (10).
Let F be a real vector-valued operator that acts on the space of all real functions on the universe X and gives us a real vector in R l given by (10) that can be uniquely identified with a function on the set of nodes Y, an element of a functional space on Y, s.t. t Y : F [ u ] ( t ) = F [ u ] t . So,
F : u F [ u ] , where u : X R and F [ u ] : Y R .
Then Definition 9 describes F-transform as a vector of F-transform components, i.e., as an image of the operator F.
Lemma 3 (F-Transform Identity).
Let X be a set, Y X , 0 < | Y | = l | X | = L < + , P = { A t : X [ 0 , 1 ] | t Y } be a fuzzy partition of X with nodes Y satisfying the assumptions of Lemma 2 and ( X , w ) be a corresponding space with closeness where W R l × L is the closeness matrix given by (8) and (9), and D R l × l be a diagonal scaling matrix with diagonal elements ( d t t ) t Y given by (7) and (11). Let F [ u ] R l be the vector of F-transform components of the function u : X R with respect to P and u R L be the vector form of the function u : X R given by i = 1 , , L : u i = u ( x i ) . Then the following holds:
D F [ u ] = W u .
Equation (13) is called the F-transform identity.
Lemma 3 characterizes the operator F in a matrix form utilizing the fact that both functional spaces contain discrete functions; hence, they can be represented as vectors (matrices of finite order):
F : R L R l , u F [ u ] = D 1 W u .
There is a historical reason for representing F-transform as a mapping.
Definition 10 (Inverse F-Transform).
Let X be a set, Y X , 0 < | Y | = l | X | = L < + , P = { A t : X [ 0 , 1 ] | t Y } be a fuzzy partition of X with nodes Y satisfying the assumptions of Lemma 2 and ( X , w ) be a corresponding space with closeness where W R l × L is the closeness matrix given by (8) and (9). Let W t be its t-th row and F [ u ] be the direct F-transform of a function u : X R in accordance with Definition 9 and (12). Then the inverse F-transform of the function u is the vector
F ^ [ u ] = W F [ u ] = t Y F [ u ] t W t R L .
Definition 10 agrees with the standard case given in [4] where it is shown that the inverse F-transform approximates the original function defined on the fuzzy partitioned space (the function is sparsely represented by the direct F-transform). Note that it is called “inverse” as it transforms the vector of direct F-transform components back to the original space. It is not an inverse mapping, though.

3. Preimage Problem

In this section, we describe a form of preimage problem (see Definition 11 below) that we would like to solve in order to describe a similarity between functions. The inverse F-transform of a function u provides a vector F ^ [ u ] that approximates the vector form of u. Therefore, we were motivated to solve the problem of how to find the class of functions that have the same inverse F-transform but decided to solve a more specific problem described in the following paragraph.
Let us have a fixed mapping (based on the F-transform identity in Lemma 3) that maps functions on vectors: u D F [ u ] = W u . We would like to characterize the similarity between functions in the sense that two functions are similar if their images (scaled vectors of direct F-transform components) coincide. In that case, their inverse F-transforms coincide as well ( D F [ u ] = D F [ u ¯ ] F ^ [ u ] = F ^ [ u ¯ ] ). We use the singular value decomposition (formalized, e.g., in [21] and commonly abbreviated as SVD) of the closeness matrix W to specify the problem and its solution.
We assume that the universe X, its subset Y of nodes and the closeness matrix W R l × L given by (8) and (9), where 0 < | Y | = l | X | = L < + , are fixed. Recall that the node degrees d t t ’s are given by (7) and form the diagonal of the matrix D R l × l . In the following text, we will respect the aforegiven assumptions.
Moreover, for simplicity, we will use the language of linear algebra in the following two subsections. That is why the clarification of the correspondence between discrete functions (functional spaces) and vectors (vector spaces) follows.
From the functional viewpoint, let A be the space of all functions u : X R and let B be the space of such functions f : Y R that assign to each node t Y the value of the corresponding F-transform component multiplied by the corresponding node degree, i.e.,
f B u : X R t Y : f ( t ) = d t t F [ u ] t .
From the viewpoint of the linear algebra, recalling that | X | = L , A = R L is a vector space, its elements are later denoted by u or v. Space A contains all vector representations of functions on X. As each element of A corresponds to one function on the universe X, we can further identify X, given by (5), with the set of its indices { 1 , , L } . Hence, the correspondence between the function u : X R and the vector u R l is given by the following equation:
x i X : u ( x i ) = u i .
Following (8) and (9), recall that each row of the closeness matrix W R l × L defines one basic function. Then, based on the F-transform identity, B = range W (range, or column space, of the matrix W R l × L is a linear subspace of R l spanned by all columns of W) is a vector subspace of R l containing the scaled images of the operator F that are uniquely determined by W. In the other words, the set of all vectors (elements) of B, f R l , is defined by
f B u A : f = D F [ u ] = W u .
Hence, the correspondence between the function f : Y R (satisfying t Y : f ( t ) = d t t F [ u ] t ) and the vector f range W is given by the following equation:
t Y : f ( t ) = f t .
Note that, for convenience, the vectors in R l are indexed based on the indices of nodes in R L , so their indices need not form an arithmetic sequence. The same convention will be also respected for the rows of the matrix W.
In the following text, we will also respect the notation of the vector spaces A and B.

3.1. Problem Formulation

In this subsection, we describe the preimage problem in terms of a mapping between two linear vector spaces.
Definition 11 (Preimage Problem).
Let ( X , w ) be a closeness space with a closeness matrix W R l × L given by t Y , x X : w ( t , x ) = W ( t , x ) , where Y is a non-empty subset of X and 0 < l L < + . Let A = R L , B = range W R l and let an element b B be given. Let the direct mapping W : A B be given by the closeness matrix W, i.e., v A : W ( v ) = W v . Then the preimage problem is to find the set { v A | W v = b } . This set is the preimage of b.
Although the solution to the preimage problem is a set, we will call any of its elements a solution to the preimage problem. By choosing the element b B , we want to utilize the closeness on X to induce an equivalence relation on A to describe the similarity between vectors.
The closeness matrix W generally exists for any closeness (based on (1) and (2)) on a finite set. It can be also defined for the special case that originates in a certain fuzzy partition (based on (8) and (9)). Hence, the preimage problem described in Definition 11 is a general formulation of the problem. By solving the general one (described by the closeness matrix), we simultaneously solve the problem specified by F-transform.
Assume that we found an SVD (see [21] to recall this technique) of the closeness matrix W in the form of the product of three matrices P S Z , where P R l × l and Z R L × L are orthogonal (these matrices are not determined uniquely) and S R l × L is diagonal ( i , j : i j s i j = 0 ) with so-called singular values i = 1 , , l : σ i = s i i on its diagonal, s.t. σ 1 σ 2 σ k > σ k + 1 = = σ l = 0 where k = dim ( range W ) = rank W is the number of linearly independent rows of W ( 0 < k l ). Any complex-valued matrix can be represented in an SVD form where the singular values are invariant under different choices of P and Z.
Lemma 4.
Let W R l × L be a closeness matrix and P S Z be its SVD as above, then
range W = range P S Z = i = 1 k c i p i | c i R ,
where p i is the i-th column of the matrix P.
Lemma 4 states that range W is spanned by the left singular vectors ( p i , i = 1 , , k ) corresponding to positive singular values of W.
Proof of Lemma 4.
Let y range W , i.e.,
y = j = 1 L d j W j ,
where for j = 1 , , L , d j R and W j is the j-th column of W. For any r = 1 , , l , we have
y r = j = 1 L d j w r j = j = 1 L d j ( P S Z ) r j = j = 1 L d j i = 1 k p r i σ i z j i = i = 1 k p r i j = 1 L d j σ i z j i .
Let
S = i = 1 k c i p i | c i R .
If for each r = 1 , , k , we set
c i = j = 1 L d j σ i z j i ,
this proves that y S because the value c i does not depend on r. Hence, range W S . As both of these sets are k-dimensional subspaces of R l , they must coincide. □

3.2. Problem Solution

Firstly, we discuss conditions when the preimage problem has a solution and when this solution is unique.
As Y , the matrix W has a positive number of rows and, hence, a positive number of columns, so range W and a vector b can be always selected. Following Definition 11, we were given a vector b B . This means that the solution always exists. The solution is always unique—it is the set of all vectors v A that are mapped on b.
Let us now discuss when the solution to the preimage problem is formed by a set with exactly one element. Following the well-known Fredholm alternative, if the homogeneous system W v 0 = 0 has a nontrivial solution v 0 A , then the set has more than one element. If W v 0 = 0 has only the trivial solution v 0 = 0 , then there is only one vector v solving the problem. Details can be found in Section 3.2.3.
Secondly, we characterize the solution from three perspectives: the first one is a useful characterization (Lemmas 5 and 6), the latter two (Theorem 1 together with Corollary 1, and Lemma 9) describe what the solution looks like.

3.2.1. Weighted Arithmetic Mean

Definition 12 (Weighted Arithmetic Mean of a Vector).
With respect to the real weights ( w i ) i = 1 L that satisfy i = 1 L w i 0 , the weighted arithmetic mean of a vector v 1 , , v L R L is a real number
i = 1 L v i w i i = 1 L w i .
The equation of the preimage problem, W v = b , is a set of linear equations W t v = b t where W t is a row of the matrix W corresponding to the index t Y , and we solve
b t = i = 1 L v i w t i where t Y .
We have l = | Y | equations in the form (16) of L l variables v 1 , , v L R , from which stems the following lemma.
Lemma 5.
Any vector v A is a solution to the preimage problem W v = b if and only if for every index t Y , its weighted arithmetic mean with respect to closeness-defining weights ( w t i ) i = 1 L is equal to b t d t t , where b t is the t-th component of the vector b and d t t is the node degree given by (11).
Proof. 
Dividing both sides of (16) by d t t = i = 1 L w t i , the claim is obvious. □
Lemma 5 states that with respect to every row of W, the weighted arithmetic means of the solutions to the preimage problem are equal. By that, Lemma 5 characterizes the set of all solutions to the preimage problem W v = b for a given vector b B . Hence, the similarity we are looking for is determined by the weighted mean of vectors. Moreover, it demarcates the solution as the equivalence class of vectors (discrete functions):
Lemma 6.
Let us have the equivalence relation on A given by: for all v , v ¯ A , we have
v v ¯ t Y : i = 1 L v i w t i i = 1 L w t i = i = 1 L v ¯ i w t i i = 1 L w t i .
Let b = W u for some u A (i.e., b B ), then the solution to the preimage problem W v = b is the set { v A | v u } .
Proof. 
This lemma is a direct consequence of Lemma 5 where for any v u , we have
t Y : i = 1 L v i w t i i = 1 L w t i = i = 1 L u i w t i i = 1 L w t i = b t d t t ,
and because u is a solution to the preimage problem, any v u is a solution.
Conversely, consider a vector v * u , s.t. W v * = b . Then
t * Y : i = 1 L v i * w t i i = 1 L w t i i = 1 L u i w t i i = 1 L w t i = b t d t t ,
which, based on (16), contradicts W v * = b . This contradiction proves that the equivalence class contains exactly all solutions to the preimage problem. □

3.2.2. Pseudoinverses

Decomposing W into its SVD by W = P S Z gives the preimage problem in the form P S Z v = b where P R l × l and Z R L × L are orthogonal and S R l × L is diagonal (i.e., i j : s i j = 0 ) with singular values of W, σ t = s t t , t Y , on its diagonal. The equation at hand is thus
t Y : q X r Y σ r p t r z t q v q = b t ,
where z t q = z q t is the element of the matrix Z corresponding to the index t in its q-th column.
Theorem 1 (Explicit Solution).
Let W R l × L be the closeness matrix given by (8) and (9), its SVD W = P S Z be given by matrices P R l × l , S R l × L and Z R L × L , s.t. rows of P, columns of P and rows of S are indexed by Y, Y + = { t Y | σ t > 0 } . Let, moreover, b B . Then, an explicit solution to the matrix equation W v = b with respect to the given SVD is the vector
v = t Y + b p t σ t z t A ,
where p t is the t-th column of the matrix P and z t is the t-th column of the matrix Z.
Proof. 
Let 1 t denote the vector containing all zeros but 1 on the t-th coordinate and ( σ t 1 ) t denote the vector containing all zeros but σ t 1 on the t-th coordinate. Following the orthogonality of P, we have:
W v = P S Z t Y + b p t σ t z t = t Y + b p t σ t P S Z z t = t Y + b p t σ t P S 1 t = t Y + b p t P S ( σ t 1 ) t = t Y + b p t P 1 t = t Y + b p t p t = b .
Remark 5.
Note that the matrices P and Z in Theorem 1 are not determined uniquely, and so there is generally more than one vector v A solving W v = b . Following Lemma 6, the solution to the corresponding preimage problem is given by the set { v A | v v } formed by all vectors equivalent with the explicit solution v.
Another general description of the solution is expressed using the notion of right inverse matrix, also called (right) pseudoinverse. Any matrix W 1 R L × l satisfying W W 1 = I where I R l × l is the identity matrix, is the right inverse matrix of W R l × L ; if we demanded that W 1 W were symmetric, W 1 would be unique, otherwise there can be multiple right inverse matrices.
Lemma 7.
Let W 1 R L × l be any right inverse matrix of the closeness matrix W and b B , then the vector v = W 1 b is a solution to the preimage problem W v = b with respect to the given right inverse matrix W 1 .
Proof. 
W v = W W 1 b = I b = b . □
The following corollary is a special case of Theorem 1 written in a matrix form.
Corollary 1.
Let W R l × L be the closeness matrix and rank W = l . If W = P S Z is its fixed SVD, then its right inverse is obtained as W 1 = Z S + P where S + R L × l is the right inverse of S, i.e., a diagonal matrix with inverted singular values of S on its diagonal. Let b B , then v = Z S + P b is a solution to the preimage problem W v = b .
Proof. 
Following the orthogonality of matrices P and Z, we have W W 1 = W Z S + P = P S Z Z S + P = P S S + P = P P = I , hence, W v = W W 1 b = b . □
Lemma 8.
If a matrix W R l × L , 0 < l L , has linearly independent rows, then W 1 = W ( W W ) 1 is its right inverse.
Proof. 
Following SVD of W used in the Corollary 1, full rank of W ensures the existence of ( W W ) 1 . Then, W W 1 = W W ( W W ) 1 = I . □
Note that the condition that no singular value of W is zero is equivalent to the condition in Lemma 8 that W has linearly independent rows.
Example A1 in the Appendix A illustrates the preimage problem and shows one vector that solves it.

3.2.3. Affine Subspace

The solution to the preimage problem does not form a linear subspace because, e.g., W ( v + v ¯ ) = 2 b which is generally not equal to b. It is a linear subspace of A if and only if b = 0 , i.e., if and only if it is equal to null W (right nullspace, or kernel, of a matrix W R l × L contains all vectors a R L , s.t. W a = 0 , it is a linear subspace of R L ). In the other words, if b = 0 , then the preimage problem, W v = b , becomes a homogeneous system of linear equations, making its solution be a linear subspace of A.
Coming back to the general case where the vector b B is arbitrary, we obtain that the solution to the preimage problem forms an affine subspace of A.
Lemma 9.
Let b B and v A be a particular solution to the preimage problem W v = b , then the set of all vectors that solve this problem, forms an affine subspace v + v 0 | v 0 null W of A.
Proof. 
For the vector v , it holds that W v = b and for any vector v 0 null W , it holds that W v 0 = 0 . Hence, for any element a = v + v 0 of the set A = v + v 0 | v 0 null W , it holds that W a = W v + W v 0 = b + 0 = b , so every element of A is a solution to the preimage problem.
Conversely, consider a vector a A , s.t. W a = b and a A . Then W ( a v ) = W a W v = b b = 0 , so the vector a v null W . As the vector a can be expressed as v + a v , we see that a A . This contradiction proves that the set A contains exactly all solutions to the preimage problem. Moreover, it proves that the set A is unique, i.e., if we express the solution to the preimage problem as B = v ¯ + v 0 ¯ | v 0 ¯ null W for any vector v ¯ A , s.t. W v ¯ = b , then A = B .
As null W A , A is an affine subspace of A with the displacement vector v . □
Using the SVD of W as above, we can express the solution to the preimage problem by right singular vectors corresponding to zero singular values of W as
v + i = k + 1 L c i z i | c i R .
The dimension of that affine subspace is then dim ( null W ) which is zero (and hence exists only one vector solving the preimage problem) if and only if k = l = L , i.e., the closeness matrix W is square and regular. If W is rectangular, we have infinitely many vectors a A solving the preimage problem.
Corollary 2.
Any weighted mean of elements of the set of all vectors solving the preimage problem with respect to weights with nonzero sum is again an element of this set.
Proof. 
Let n N and i = 1 , , n : W v i = b , i.e., each v i A be a solution to the preimage problem. Let c 1 , , c n R , s.t. i = 1 n c i 0 . Then it holds that
W i = 1 n c i v i i = 1 n c i = i = 1 n c i W v i i = 1 n c i = i = 1 n c i b i = 1 n c i = b ,
which proves the claim. □
In this subsection, we described the solution to the preimage problem from three different perspectives: using weighted means of functional values (determining an equivalence class of mutually similar functions), using right inverses of the closeness matrix (that can be obtained also by SVD) and using the notion of affine subspace (its displacement vectors can be also obtained by SVD of W, based on Theorem 1).
In the next section, we show the connection between the inverse F-transform and a solution to the preimage problem as the rows of W are formed by the basic functions of the fuzzy partition.

4. Inverse F-Transform

In this section, we omit the assumption (6) stating every point of the universe X must be covered by at least one basic function. Admitting this degenerated case allows us to express a useful characterization of the case when the inverse F-transform, given by Definition 10, provides a solution to the preimage problem.
Lemma 10.
Let the assumptions of Lemma 3 be satisfied. If
W W = D ,
then the inverse F-transform v = F ^ [ u ] given by Definition 10 is a solution to the preimage problem according to Definition 11, D F [ v ] = W v = W u = D F [ u ] = b for any vector u A .
Proof. 
We see that the inverse F-transform is a solution to the preimage problem if and only if for any u A , it holds:
W u = D F [ u ] = b = W v = W F ^ [ u ] = W W F [ u ] = W W D 1 W u ,
which holds if
W W D 1 = I ,
or equivalently if (17) holds true. □
Corollary 3.
Condition (17) holds true if both of the following equations are fulfilled:
1. 
t Y : d t t = i = 1 L w t i = i = 1 L w t i 2 ,
2. 
t , s Y , t s : d t s = 0 = i = 1 L w t i w s i .
Recalling that w t i = W ( t , x i ) = A t ( x i ) [ 0 , 1 ] , we infer that condition 1 is equivalent to stating that t Y , x i X : w t i { 0 , 1 } , i.e., only simple-minded assignment of closeness is possible to be used. Condition 2 is equivalent to stating that every column of the matrix W contains at most one non-zero value—let us denote the system of all such matrices as W . Putting these two conditions together, we obtain the following corollary.
Corollary 4.
If the matrix W W contains only values 0 and 1, then the inverse F-transform v = F ^ [ u ] given by Definition 10 is a solution to the preimage problem.
This property is demonstrated in Examples A2 and A3 in the Appendix A.

5. New Set of Basic Functions

This section was inspired by the paper [17] where one of the main goals consists in finding conditions under which a noisy signal must be sampled so that it can be reconstructed from its F-transform components. The authors showed that a fuzzy partition adjoint to the partition (with the same nodes) that determined the F-transform components, ensures that the associated inverse F-transform provides the best approximation of the original, continuous function in a certain space.
We analyzed a possible connection between discrete fuzzy partitions on the closeness space and decided to address the following question: can we find a solution to the preimage problem W v = b by creating a new set of basic functions, s.t. they determine a linear transformation of the vector b that would solve the problem? If so, what are the conditions that this new set of basic functions must satisfy?
Throughout this section, assume that the sets X = { 1 , , L } , Y and { A t : X [ 0 , 1 ] | t Y } , the matrices W and D and the vector b are related as in Lemma 3 and fixed.
We are looking for a new set of basic functions (a new fuzzy partition) { B t : X R | t Y } , written in rows of a newly created matrix M R l × L , i.e.,
t Y , x X : m t x = B t ( x ) ,
s.t. the vector v = M b solves the preimage problem.
Therefore, we require that W M b = b and from this, we deduce properties of the matrix M and express them in terms of basic functions A t ’s and B t ’s. Let for an arbitrary matrix N, N t denote its t-th row, N t denote its t-th column and N t s denote its entry n t s .
Hence, we require that
t Y : ( W M ) t b = s Y ( W M ) t s b s = s Y W t ( M ) s b s = b t .
Recalling that W t x = A t ( x ) and ( M ) x s = M s x = B s ( x ) , for t , s Y , x X , we have
t Y : s Y b s x X A t ( x ) B s ( x ) = b t .
Substituting the constants
c t s = x X A t ( x ) B s ( x ) , ( t , s ) Y × Y ,
in (19), we obtain that
t Y : s Y b s c t s = b t .
In the other words, by creating the matrix C = W M R l × l from the constants c t s , ( t , s ) Y × Y , we solve C b = b . As B can be the whole space R l , we generally demand that C is an identity matrix I (in the special case of rank W < l , this requirement would be unnecessarily strong but leads again to a good choice of new basic functions B t ’s).
This means that for all ( t , s ) Y × Y , we demand
1.
t = s : c t s = x X A t ( x ) B s ( x ) = 1 , and
2.
t s : c t s = x X A t ( x ) B s ( x ) = 0 .
This observation motivates the following definition.
Definition 13 (Compatible Set of Basic Functions).
Let the assumptions of Lemma 3 be satisfied. We say that the set { B t : X R | t Y } is a compatible set of basic functions with respect to { A t : X R | t Y } in the closeness space ( X , w ) , if
t Y : x X A t ( x ) B t ( x ) = 1 ,
and
t , s Y , t s : x X A t ( x ) B s ( x ) = 0 .
Theorem 2 (Solution Determined by a Compatible Set of Basic Functions).
Let the assumptions of Lemma 3 be satisfied and let { B t : X R | t Y } be a compatible set of basic functions with respect to { A t : X R | t Y } in the closeness space ( X , w ) . Let b B and a matrix M R l × L be defined by M t x = B t ( x ) , t Y , x X , then the vector v = M b is a solution to the preimage problem W v = b .
Proof. 
W v = W M b R l and for each of its components, we have
t Y : W t v = ( W M ) t b = s Y ( W M ) t s b s = s Y W t ( M ) s b s = s Y x X W t x M s x b s = s Y x X A t ( x ) B s ( x ) b s = b t .
This proves that W v = b . □
If we can find a set of new basic functions B t ’s compatible with A t ’s, then the vector v = M b can be described as an inverse F-transform of an unknown vector u A , s.t. b = D F [ u ] , with respect to { B t : X R | t Y } written in the rows of M.
In other words, the direct mapping between the vector spaces A and B is given by W ( W : A B ) and any M gives an element of the induced inverse relation, i.e., ( b , M b ) W 1 where W 1 B × A is the inverse relation to the mapping W. This leads to multiple solutions forming a subset of the affine subspace described in Lemma 9.
Lemma 11.
Let the assumptions of Lemma 3 be satisfied. If rank W = l , then there always exists a compatible set of basic functions with respect to { A t : X R | t Y } in the closeness space ( X , w ) .
Proof. 
Following Lemma 8, W 1 = W ( W W ) 1 is a right inverse to W. By setting M = W 1 , we found a compatible set of basic functions { B t : X R | t Y } given by (18). □
To conclude, we found the inverse-F-transform-like procedure associated with the system of new basic functions that, applied on b, gives a solution to the preimage problem. Examples A4 and A5 in the Appendix A illustrate the proposition of this section.

6. Conclusions

In this paper, we discussed the preimage problem in the space where the relationship between objects is determined by closeness. We showed that any metric can be transformed into closeness and therefore, the latter is weaker than the former. We interlaced closeness with fuzzy partition and characterized both by the closeness matrix.
We expressed similarity between functions based on their images (coinciding with vectors of scaled F-transform components) computed using the closeness matrix. By that (and by setting the basic structure of the space without metric, or norm), we contributed to the mathematical theory in the field of functional analysis. We formulated the preimage problem using the language of matrix calculus. The preimage problem solution is given by (i) a weighted arithmetic mean, (ii) any right inverse of the closeness matrix or (iii) any element of a certain affine subspace. Singular value decomposition was applied to describe the problem and its solution.
We defined the notion of compatible set of basic functions and found conditions under which the inverse F-transform with respect to the given and compatible set of basic functions forms a solution to the preimage problem.
Theoretical results were illustrated by numerical examples. They demonstrate, e.g., that requiring reflexivity of closeness can be counterproductive.
The future research will be focused on imposing further conditions on both collections of basic functions ( A t ’s and B t ’s) to reveal stronger connections between the spaces A and B.

Author Contributions

Investigation, J.J. and I.P.; resources, J.J. and I.P.; writing—original draft preparation, J.J.; writing—review and editing, J.J. and I.P.; supervision, I.P. All authors have read and agreed to the published version of the manuscript.

Funding

The support of the project SGS20/PřF-MF/2022 of the University of Ostrava is greatly appreciated.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following symbols and abbreviations are used in this manuscript:
wcloseness, closeness-describing bivariate weight function
( X , w ) closeness space with the carrier set (universe) X and closeness w
×Cartesian product
Wcloseness matrix storing weights w i j = w ( x i , x j ) describing closeness between
x i and x j
fuzzy subset of a given universe
A t basic function associated with the node t Y
l = | Y | number of nodes within the universe X
L = | X | number of all points of the universe X
d t t degree of the node t Y , diagonal element of the scaling (degree) matrix D
F [ u ] = F [ u ] t t Y F-transform of the function u, vector of F-transform components
F ^ [ u ] inverse F-transform of the function u
· transpose of ·
W t t-indexed row of the closeness matrix W
Aspace of all real functions on the universe X or all corresponding
L-dimensional vectors
B = range W space of all real l-dimensional vectors that belong to the column space of the
closeness matrix W
P, S, ZSVD matrices: m. of left singular vectors, m. with singular values σ t on its
diagonal, m. of right singular vectors, respectively
equivalence relation between vectors in A
W 1 right inverse matrix of the closeness matrix W
W system of all real l × L matrices containing at most one non-zero entry in each column
F-transformfuzzy transform
SVDsingular value decomposition
s.t.such that

Appendix A

This section contains numerical examples related to the previous five sections (excluding Section 6) to increase their legibility.
Example A1 relates to Section 3.2, explains the notations of matrices and shows the detailed computation of the F-transform components which will be omitted in other examples. It illustrates the preimage problem and shows one vector that solves it.
Example A1.
Let a set of nodes Y = { 3 , 4 , 6 } be a selected subset of X = { 1 , 2 , 3 , 4 , 5 , 6 , 7 , 8 } , the universe, and
t Y : A t ( x ) = x t 3 + 1 x [ t 3 , t ] X t x 3 + 1 x [ t , t + 3 ] X 0 x X [ t 3 , t + 3 ]
be triangular basic functions on X associated with elements of Y that determine the closeness-describing (weight) matrix
W = 1 3 2 3 1 2 3 1 3 0 0 0 0 1 3 2 3 1 2 3 1 3 0 0 0 0 0 1 3 2 3 1 2 3 1 3 ,
and the scaling (node degree) matrix
D = 3 0 0 0 3 0 0 0 3 .
Note that these two matrices, W and the derived D, are independent of the functions acting on X. Let u = x 2 be a particular function on X which has a vector form
u = 1 4 9 16 25 36 49 64 .
The function u is described by its values u i i = 1 L = u ( x i ) x i X that represent registered values of a certain variable (physical, chemical, etc.) measured at all points (on all objects) of the universe (data set) X. Another possible interpretation of u consists in assuming that there is a continuous signal that spreads over a certain medium during a time frame described by an interval [ y 0 , y l + 1 ] X and we are able to register that signal only in discrete time steps x i i = 1 L . The components of the direct F-transform of the function u corresponding to all A t ’s ( t Y ) are:
F [ u ] 1 = 1 · 1 3 + 4 · 2 3 + 9 · 1 + 16 · 2 3 + 25 · 1 3 : 1 3 + 2 3 + 1 + 2 3 + 1 3 = 31 3
F [ u ] 2 = 4 · 1 3 + 9 · 2 3 + 16 · 1 + 25 · 2 3 + 36 · 1 3 : 1 3 + 2 3 + 1 + 2 3 + 1 3 = 52 3
F [ u ] 3 = 16 · 1 3 + 25 · 2 3 + 36 · 1 + 49 · 2 3 + 64 · 1 3 : 1 3 + 2 3 + 1 + 2 3 + 1 3 = 112 3 .
Writing them in the vector form, we obtain
F [ u ] = 31 3 52 3 112 3 .
Obviously, it holds that D F [ u ] = W u , i.e.,
3 0 0 0 3 0 0 0 3 · 31 3 52 3 112 3 = 1 3 2 3 1 2 3 1 3 0 0 0 0 1 3 2 3 1 2 3 1 3 0 0 0 0 0 1 3 2 3 1 2 3 1 3 · 1 4 9 16 25 36 49 64 = 31 52 112 .
The preimage problem given by the vector we ended up with (this ensures that b B ) then has the form
b = D F [ u ] = 31 52 112 = 1 3 2 3 1 2 3 1 3 0 0 0 0 1 3 2 3 1 2 3 1 3 0 0 0 0 0 1 3 2 3 1 2 3 1 3 · v = W v .
We will use SVD of W to find one of the right inverse matrices that constitute a solution.
For one particular SVD decomposition, we obtain an explicit solution
v = Z S + P b 10.563025 8.92437 7.285714 6.403361 29.92437 53.445378 43.764706 21.882353 .
To verify it, we compute
W v 31 52 112 ,
and comparing it to b we see that we found a solution to the preimage problem.
In Example A1, we obtained the same particular solution while using the explicit formula from Theorem 1 as well as while using the right inverse given by SVD of W (Corollary 1) and the right inverse given by Lemma 8. As SVD is widely used to compute pseudoinverses, this was not too surprising and our software is no exception.
Examples A2 and A3 relate to Section 4 and show that under certain conditions, the inverse F-transform is one of the possible solutions to the preimage problem. The former example illustrates the degenerated case where not all points of the universe are covered by basic functions (are not connected with at least one node), the latter example uses non-convex basic functions.
Example A2.
Let Y = { 3 , 4 , 6 } be a selected subset of nodes of the set X = { 1 , 2 , 3 , 4 , 5 , 6 , 7 , 8 } ,
t Y : A t ( x ) = 1 t = x 0 t x
be singleton basic functions on X associated with elements of Y that determine the closeness matrix
W = 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 ,
and the scaling matrix
D = 1 0 0 0 1 0 0 0 1 .
Let u be a particular function on X with a vector form
u = 0 0 9 16 0 36 0 0 .
The components of the direct F-transform of u corresponding to all A t ’s ( t Y ) form the vector:
F [ u ] = 9 16 36 .
Obviously, it holds that D F [ u ] = W u , i.e.,
1 0 0 0 1 0 0 0 1 · 9 16 36 = 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 · 0 0 9 16 0 36 0 0 = 9 16 36 .
The preimage problem given by the vector we ended up with then has the form
b = D F [ u ] = 9 16 36 = 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 · v = W v .
The inverse F-transform of u with respect to W has the form
F ^ [ u ] = W F [ u ] = 0 0 9 16 0 36 0 0 .
Since this vector is equal to u, we obtain that W F ^ [ u ] = W u , so F ^ [ u ] is a solution to the preimage problem. Note that W is such that W W = I = D .
Example A3.
Let Y = { 3 , 4 , 6 } be a selected subset of nodes of the set X = { 1 , 2 , 3 , 4 , 5 , 6 , 7 , 8 } and let the closeness be determined by the matrix
W = 1 1 1 0 0 0 0 0 0 0 0 1 1 0 0 0 0 0 0 0 0 1 1 1 ,
and the corresponding scaling matrix
D = 3 0 0 0 2 0 0 0 3 .
Let u be a particular function on X with a vector form
u = 0 0 9 16 0 36 0 0 .
The components of the direct F-transform of u corresponding to all A t ’s ( t Y ) form the vector:
F [ u ] = 3 8 12 .
Obviously, it holds that D F [ u ] = W u , i.e.,
3 0 0 0 2 0 0 0 3 · 3 8 12 = 1 1 1 0 0 0 0 0 0 0 0 1 1 0 0 0 0 0 0 0 0 1 1 1 · 0 0 9 16 0 36 0 0 = 9 16 36 .
The preimage problem given by the vector we ended up with then has the form
b = D F [ u ] = 9 16 36 = 1 1 1 0 0 0 0 0 0 0 0 1 1 0 0 0 0 0 0 0 0 1 1 1 · v = W v .
The inverse F-transform of u with respect to W has the form
F ^ [ u ] = W F [ u ] = 3 3 3 8 8 12 12 12 .
We see that
W F ^ [ u ] = 1 1 1 0 0 0 0 0 0 0 0 1 1 0 0 0 0 0 0 0 0 1 1 1 · 3 3 3 8 8 12 12 12 = 9 16 36 = W u ,
so F ^ [ u ] is a solution to the preimage problem. Note that again W is such that W W = D .
Examples A4 and A5 relate to Section 5 and illustrate that creating a compatible set of basic functions leads to a solution to the preimage problem. The latter example shows a drawback of requiring reflexivity of closeness.
Example A4.
Let X = { 1 , 2 , 3 , 4 , 5 , 6 , 7 , 8 , 9 } be the universe with a selected subset Y = { 2 , 5 , 8 } of nodes and let the closeness be determined by the matrix
W = 1 2 3 4 1 2 0 0 0 0 0 0 0 0 0 1 2 3 4 1 2 0 0 0 0 0 0 0 0 0 1 2 3 4 1 2 ,
and the corresponding scaling matrix
D = 7 4 0 0 0 7 4 0 0 0 7 4 .
Let u = x 2 be a particular function on X which has a vector form
u = 1 4 9 16 25 36 49 64 81 .
The components of the direct F-transform of the function u corresponding to all A t ’s ( t Y ) form the vector:
F [ u ] = 32 7 179 7 452 7 .
Obviously, it holds that D F [ u ] = W u , i.e.,
7 4 0 0 0 7 4 0 0 0 7 4 · 32 7 179 7 452 7 = 1 2 3 4 1 2 0 0 0 0 0 0 0 0 0 1 2 3 4 1 2 0 0 0 0 0 0 0 0 0 1 2 3 4 1 2 · 1 4 9 16 25 36 49 64 81 = 8 179 4 113 .
The preimage problem given by the vector we ended up with then has the form
b = 8 179 4 113 = 1 2 3 4 1 2 0 0 0 0 0 0 0 0 0 1 2 3 4 1 2 0 0 0 0 0 0 0 0 0 1 2 3 4 1 2 · v = W v .
Let us consider, e.g., the following matrix M satisfying W M = I :
M = 1 2 2 3 1 2 0 0 0 0 0 0 0 0 0 1 2 2 3 1 2 0 0 0 0 0 0 0 0 0 1 2 2 3 1 2 .
Then the vector
v = M b = 1 2 2 3 1 2 0 0 0 0 0 0 0 0 0 1 2 2 3 1 2 0 0 0 0 0 0 0 0 0 1 2 2 3 1 2 · 8 179 4 113 = 4 16 3 4 179 8 179 6 179 8 113 2 226 3 113 2 ,
is a solution to the preimage problem that can be verified by
W v = 1 2 3 4 1 2 0 0 0 0 0 0 0 0 0 1 2 3 4 1 2 0 0 0 0 0 0 0 0 0 1 2 3 4 1 2 · 4 16 3 4 179 8 179 6 179 8 113 2 226 3 113 2 = 8 179 4 113 = b .
Moreover, it demonstrates the fact that the sets of basic functions A t ’s (given by the rows of W) and B t ’s (given by the rows of M), t Y , are compatible.
If we assumed that for each t Y it holds A t ( t ) = 1 (reflexive closeness), the compatible set of convex basic functions satisfying that also for each t Y it holds B t ( t ) = 1 , would be uniquely formed by singletons (degenerated case). This is illustrated by Example A5.
Example A5.
Let X = { 1 , 2 , 3 , 4 , 5 , 6 , 7 , 8 , 9 } be the universe with the selected subset Y = { 2 , 5 , 8 } of nodes and let the closeness be determined by the matrix
W = 1 2 1 1 2 0 0 0 0 0 0 0 0 0 1 2 1 1 2 0 0 0 0 0 0 0 0 0 1 2 1 1 2 ,
and the corresponding scaling matrix
D = 2 0 0 0 2 0 0 0 2 .
Let u = x 2 be a particular function on X which has a vector form
u = 1 4 9 16 25 36 49 64 81 .
The components of the direct F-transform of the function u corresponding to all A t s ( t Y ) form the vector:
F [ u ] = 9 2 51 2 129 2 .
Obviously, it holds that D F [ u ] = W u , i.e.,
2 0 0 0 2 0 0 0 2 · 9 2 51 2 129 2 = 1 2 1 1 2 0 0 0 0 0 0 0 0 0 1 2 1 1 2 0 0 0 0 0 0 0 0 0 1 2 1 1 2 · 1 4 9 16 25 36 49 64 81 = 9 51 129 .
The preimage problem given by the vector we ended up with then has the form
b = 9 51 129 = 1 2 1 1 2 0 0 0 0 0 0 0 0 0 1 2 1 1 2 0 0 0 0 0 0 0 0 0 1 2 1 1 2 · v = W v .
Let us consider the following matrix M satisfying W M = I (if we want each of its rows to form a convex fuzzy set with M t [ t ] = 1 , then this is the only option):
M = 0 1 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 1 0 .
Note that if for any s Y { t } it holds that A t ( s ) > 0 , then no compatible matrix M exists.
The vector
v = M b = 0 1 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 1 0 · 9 51 129 = 0 9 0 0 51 0 0 129 0 ,
is a solution to the preimage problem that can be verified by
W v = 1 2 1 1 2 0 0 0 0 0 0 0 0 0 1 2 1 1 2 0 0 0 0 0 0 0 0 0 1 2 1 1 2 · 0 9 0 0 51 0 0 129 0 = 9 51 129 = b .
This again demonstrates the fact that the sets of basic functions A t ’s (given by the rows of W) and B t ’s (given by the rows of M), t Y , are compatible. More importantly, it demonstrates that requiring A t ( t ) = 1 and B t ( t ) = 1 for all node indices t Y can lead to an empty set of compatible convex basic functions.

References

  1. Janeček, J.; Perfilieva, I. Three Methods of Data Analysis in a Space with Closeness. In Developments of Artificial Intelligence Technologies in Computation and Robotics: Proceedings of the 14th International FLINS Conference (FLINS 2020), Cologne, Germany, 18–21 August 2020; World Scientific: Singapore, 2020; pp. 947–955. [Google Scholar]
  2. Belkin, M.; Niyogi, P. Laplacian Eigenmaps for Dimensionality Reduction and Data Representation. Neural Comput. 2003, 15, 1373–1396. [Google Scholar] [CrossRef]
  3. Janeček, J.; Perfilieva, I. F-transform and Dimensionality Reduction: Common and Different. In International Summer School on Aggregation Operators 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 267–278. [Google Scholar]
  4. Perfilieva, I. Fuzzy Transforms: Theory and Applications. Fuzzy Sets Syst. 2006, 157, 993–1023. [Google Scholar] [CrossRef]
  5. St. Luce, S.; Sayama, H. Analysis and Visualization of High-Dimensional Dynamical Systems’ Phase Space Using a Network-Based Approach. Complexity 2022, 2022, 3937475. [Google Scholar] [CrossRef]
  6. Schneider, N.; Wurm, G.; Teiser, J.; Klahr, H.; Carpenter, V. Dense Particle Clouds in Laboratory Experiments in Context of Drafting and Streaming Instability. Astrophys. J. 2019, 872, 3. [Google Scholar] [CrossRef]
  7. Wang, J.; Wang, L.; Liu, X.; Ren, Y.; Yuan, Y. Color-Based Image Retrieval Using Proximity Space Theory. Algorithms 2018, 11, 115. [Google Scholar] [CrossRef]
  8. Riesz, F. Stetigkeitsbegriff und abstrakte Mengenlehre. In Proceedings of the International Congress of Mathematicians, Rome, Italy, 6–11 April 1908; Volume 2, pp. 18–24. [Google Scholar]
  9. Perfilieva, I.; Daňková, M.; Bede, B. Towards a Higher Degree F-Transform. Fuzzy Sets Syst. 2011, 180, 3–19. [Google Scholar] [CrossRef]
  10. Stefanini, L. Fuzzy Transform with Parametric LU-Fuzzy Partitions. In Computational Intelligence in Decision and Control; World Scientific: Singapore, 2008; pp. 399–404. [Google Scholar]
  11. Perfilieva, I.; Hurtik, P. The F-Transform Preprocessing for JPEG Strong Compression of High-Resolution Images. Inf. Sci. 2021, 550, 221–238. [Google Scholar] [CrossRef]
  12. Hurtik, P.; Molek, V.; Perfilieva, I. Novel Dimensionality Reduction Approach for Unsupervised Learning on Small Datasets. Pattern Recognit. 2020, 103, 107291. [Google Scholar] [CrossRef]
  13. Zakeri, K.A.; Ziari, S.; Araghi, M.A.F.; Perfilieva, I. Efficient Numerical Solution to a Bivariate Nonlinear Fuzzy Fredholm Integral Equation. IEEE Trans. Fuzzy Syst. 2019, 29, 442–454. [Google Scholar] [CrossRef]
  14. Honeine, P.; Richard, C. Preimage Problem in Kernel-based Machine Learning. IEEE Signal Process. Mag. 2011, 28, 77–88. [Google Scholar] [CrossRef] [Green Version]
  15. Kwok, J.Y.; Tsang, I.H. The Pre-Image Problem in Kernel Methods. IEEE Trans. Neural Netw. 2004, 15, 1517–1525. [Google Scholar] [CrossRef] [PubMed]
  16. Janeček, J.; Perfilieva, I. Noise Reduction as an Inverse Problem in F-Transform Modelling. In Proceedings of the International Conference on Information Processing and Management of Uncertainty in Knowledge-Based Systems 2022, Milan, Italy, 11–15 July 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 405–417. [Google Scholar] [CrossRef]
  17. Perfilieva, I.; Holčapek, M.; Kreinovich, V. A New Reconstruction from the F-Transform Components. Fuzzy Sets Syst. 2016, 288, 3–25. [Google Scholar] [CrossRef]
  18. Bakır, G.H.; Weston, J.; Schölkopf, B. Learning to Find Pre-Images. Adv. Neural Inf. Process. Syst. 2004, 16, 449–456. [Google Scholar]
  19. Blyth, T.S. Set Theory and Abstract Algebra; Longmans Mathematical Texts; Longman Publishing Group: Harlow, UK, 1975. [Google Scholar]
  20. Knauer, U.; Knauer, K. Algebraic Graph Theory: Morphisms, Monoids and Matrices; De Gruyter: Berlin, Germany, 2019. [Google Scholar]
  21. Golub, G.H.; Van Loan, C.F. Matrix Computations; Johns Hopkins University Press: Baltimore, MD, USA, 1996. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Janeček, J.; Perfilieva, I. Preimage Problem Inspired by the F-Transform. Mathematics 2022, 10, 3209. https://doi.org/10.3390/math10173209

AMA Style

Janeček J, Perfilieva I. Preimage Problem Inspired by the F-Transform. Mathematics. 2022; 10(17):3209. https://doi.org/10.3390/math10173209

Chicago/Turabian Style

Janeček, Jiří, and Irina Perfilieva. 2022. "Preimage Problem Inspired by the F-Transform" Mathematics 10, no. 17: 3209. https://doi.org/10.3390/math10173209

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop