Next Article in Journal
Improving Spiking Neural Network Performance with Auxiliary Learning
Previous Article in Journal
Behavior-Aware Pedestrian Trajectory Prediction in Ego-Centric Camera Views with Spatio-Temporal Ego-Motion Estimation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Identifying the Regions of a Space with the Self-Parameterized Recursively Assessed Decomposition Algorithm (SPRADA)

LISSI Laboratory EA 3956, Sénart-FB Institute of Technology, Campus of Sénart, University of Paris-Est Créteil, 77567 Lieusaint, France
*
Authors to whom correspondence should be addressed.
Mach. Learn. Knowl. Extr. 2023, 5(3), 979-1009; https://doi.org/10.3390/make5030051
Submission received: 11 July 2023 / Revised: 31 July 2023 / Accepted: 1 August 2023 / Published: 4 August 2023

Abstract

:
This paper introduces a non-parametric methodology based on classical unsupervised clustering techniques to automatically identify the main regions of a space, without requiring the objective number of clusters, so as to identify the major regular states of unknown industrial systems. Indeed, useful knowledge on real industrial processes entails the identification of their regular states, and their historically encountered anomalies. Since both should form compact and salient groups of data, unsupervised clustering generally performs this task fairly accurately; however, this often requires the number of clusters upstream, knowledge which is rarely available. As such, the proposed algorithm operates a first partitioning of the space, then it estimates the integrity of the clusters, and splits them again and again until every cluster obtains an acceptable integrity; finally, a step of merging based on the clusters’ empirical distributions is performed to refine the partitioning. Applied to real industrial data obtained in the scope of a European project, this methodology proved able to automatically identify the main regular states of the system. Results show the robustness of the proposed approach in the fully-automatic and non-parametric identification of the main regions of a space, knowledge which is useful to industrial anomaly detection and behavioral modeling.

1. Introduction

The fourth industrial revolution, known as Industry 4.0, proposes to rethink the way the industrial sector is managed and controlled [1], by extending the areas of the Internet of Things (IoT) [2] and Artificial Intelligence (AI) [3] to traditional industry. Now, large, specialized machines are being replaced by smaller, possibly versatile production units, whilst the supply chain trades its conveyor belts for small, highly mobile robots [4]. The addition of intelligence to a hyper-connected system is expected to improve the autonomy of the units [5], and global efficiency more broadly [6], by making them react automatically in case of need (e.g., in case of disruption, change in the production schedule, etc.).
The most noticeable aspect of Industry 4.0 is the Industrial Internet of Things (IIoT), a topic which consists of connecting many systems within a wireless network, and investigating how to make them communicate. This can be achieved by using classical IoT tools (self-sufficient systems on chip, low-energy communication protocols, etc.) [7], or through a dedicated interface, for instance by using a Cyber-Physical System (CPS) [8].
Moreover, having many free-moving, possibly autonomous robots performing their tasks in parallel makes the investigation of solutions to manage them all at the same time a necessity. Such tools may allow them to collaborate [9], to decide which path to take to reach a place without injuring any system or people [10], to avoid collisions [11], etc.
Additionally, another major challenge of Industry 4.0 is the virtualization of systems, called digital twin, which aims to build an accurate and faithful representation of them [12]. Such tool can serve many purposes: modeling for anomaly detection [13], simulation for predictive maintenance [14], pure modeling and prediction for simulation [15], etc.
Actually, all these areas of research are different sides of the same challenge: the deep understanding of the industrial systems, so as to ameliorate their management and control. Indeed, intelligence often relies on information [16], for instance in the form of knowledge, which may be derived from the system’s processes, and especially their time series [17]. With real industrial systems, it may be automatically extracted from the processes’ historical data or human expertise by data-driven, machine learning-based approaches [18]. In that context, the extraction of knowledge seems to constitute the very first step of the emergence of a true intelligence [19]. Such knowledge can take many shapes: smart scheduling of the production orders to improve efficiency [20], modeling and prediction of time series for simulation [21], behavioral modeling to better understand and predict systems [22], etc.
In industry, in the scope of the predictive maintenance [23], a major challenge is the detection and diagnosis of anomalies [24], which, if not handled and corrected rapidly, may cause a more severe disruption or a true failure thereafter [25]. In industrial contexts, anomaly detection is often performed by machine learning or deep learning [26], especially using supervised algorithms, i.e., data must be labeled in some fashion upstream; however, obtaining such prior knowledge is often difficult, tedious and costly, since it typically requires manual investigation. On the contrary, data mining deals with the automatic extraction of knowledge from raw, unlabeled data, which makes it appealing as a possible candidate for anomaly detection [27]. In particular, unsupervised clustering may be greatly useful in such a context; clustering is commonly used in exploratory data analysis to find the main regions of a space [28], and can be applied to industrial contexts to isolate both the anomalies and the regular states of real unknown systems [29]. Indeed, this technique of statistical data analysis aims to regroup data sharing similarities with each other, whilst isolating the so-build groups from one another; with real system’s feature space, both the regular states and the historically-encountered anomalies should form compact groups, likely isolated from one another: clustering appears to be a relevant tool to separate the different groups automatically. Once the groups are isolated, they can easily be tagged as either anomalies or true regular states (that one may call the regular system’s “behaviors”), and can later on be used to feed any supervised classifier or predictive tool [30].
Although unsupervised clustering performs the task of separating the different groups of data within a space well, most of the related techniques are parametric: they require some prior information to correctly set the parameters before they can be performed [31]. Correctly setting them is essential to obtain a meaningful partitioning of the space [32]. Particularly, a parameter which is often required is the objective number of clusters to build: it greatly impacts the accuracy of the partitioning, and the relevance of the groups. Nevertheless, it is generally hard to set without accessing any prior knowledge on the system, thus, manual expertise is typically performed upstream [33]. Some clustering algorithms do not require that knowledge directly, by building the clusters one after the other until every data has been processed. However, such algorithms generally trade the number of clusters to be built by another parameter, which typically takes the form of a threshold, which determines the number of clusters issued, and thus the meaningfulness of the partitioning.
In order to accurately separate the different groups, one should feed the clustering algorithms with the correct objective number of clusters; nonetheless, there is no true solution to estimate that value without upstream manual processing. Additionally, the grain of the desired partitioning (i.e., comprising only the regular states, or also the anomalies within smaller groups) will affect that value: it is difficult to answer the question “how many behaviors and anomalies are contained within this N-month long dataset?” with absolute certainty, and especially in a fully automatic and blind fashion.
For that reason, this paper introduces a new unsupervised clustering methodology, specifically crafted to answer the above-mentioned difficulty of setting the objective number of clusters. This method consists of recursively splitting a space according to a split criterion: the space is partitioned a first time, then each cluster is analyzed using an integrity criterion, and, if it obtained a bad score, the cluster is itself split, and so on until the integrity of every issued cluster is acceptable. Once all the clusters are built, a final, small step of cluster merging is operated to fuse the smallest groups (especially that with one single data) into the largest ones. This method is self-parameterized, and is therefore named SPRADA, standing for Self-Parameterized Recursively Assessed Decomposition Algorithm. Notice that the objective number of clusters still matters, but considerably less, since if this value is not correctly set, then a dedicated step of clustering will be applied to compensate it.
The proposed method is applied to real industrial data, obtained in the context of an Industry 4.0-oriented European project; it proved to be accurate in identifying the main regions of the feature space, while also isolating the historical anomalies. These observations make the SPRADA method appealing for future work in the scope of predictive maintenance, alongside the development and deployment of Industry 4.0 solutions.
The manuscript is composed as follows: Section 2 is a detailed review of the literature on anomaly detection, the techniques to estimate the number of clusters of a dataset and the clustering techniques which do not require it. Then, the tools necessary for the proposed methodology are presented in Section 3, and the proposed SPRADA method is introduced and detailed in Section 4, and it is applied to a real industrial dataset in Section 5 to assess its accuracy and representativeness, and the findings are discussed in Section 6. Finally, Section 7 concludes this study.

2. State-of-the-Art

First, a short review of recent works about anomaly detection in industrial contexts is worth exploring to set the background of this study. To that end, several data-driven techniques were used, mostly based on deep learning, and especially on Convolutional Neural Networks (CNNs). For instance, Ref. [34] proposed to feed labeled anomalies of real solar plants into a CNN to learn and predict them; the study was then extended in [35] to many more Artificial Neural Networks (ANNs). Similarly, Ref. [36] learned the modes of a real hydro-power turbine by using Variational Modal Decomposition (VCM) and used CNN-oriented auto-encoders to detect unusual behaviors. Moreover, Ref. [37] used Root Cause Analysis (RCA) to identify the root causes of defects in a real steel industry hot strip mill, and then semi-automatically labeled the identified events to feed them in more classical machine learning-based ANNs. Eventually, some works used the real-time data collected by a Supervisory Control and Data Acquisition (SCADA) system to detect possible measurement intrusion, by using a classical Long Short-Term Memory (LSTM) [38], or by using supervised classification to identify both normal and abnormal activities, and finally learn the possible intrusions using supervised machine learning-based approaches [39].
All these techniques perform well, achieving promising results in the identification and prediction of anomalies, but they are computationally demanding, require a significant amount of data (a typical limitation in deep learning), and are supervised. Therefore, they are not suited to the fully-automatic and blind identification of the distinct anomalies—and behaviors—of industrial systems. That being said, the proposed SPRADA methodology may serve as a first fully-automated and non-parametric pre-processing tool to identify the historical anomalies, which may later on be fed to the works mentioned in the previous paragraph, whence the necessity of such a tool, and the correct estimation of the true number of groups (both anomalies and behaviors) in real industrial processes’ datasets.
Several attempts were made to estimate the “optimal” number of clusters to build to obtain the most representative partitioning as possible. Actually, the optimal value is not always that which optimizes an integrity criterion (e.g., cluster’s variance, density, etc.), since, with most quantifiers, the best partitioning method would be that where every single datapoint is contained in a unique cluster (for instance, the variance of every cluster would be zero); nevertheless, this partitioning would be positively useless, for it brings no information on the system’s feature space. Therefore, the techniques aiming to estimate the optimal number of clusters endeavor to provide a value which optimizes the integrity criterion of the clusters, whilst also maximizing the information brought by the partitioning (its meaningfulness).
Most of the techniques operate in an experimental fashion, by operating clustering several times, then evaluating the relevance of the partition (according to any criterion), and eventually estimating the optimal number of clusters to build as the value which provides a good trade-off between the accuracy and the size/number of clusters.
Among the techniques of the literature, probably the most widespread, albeit aging, is the Elbow method [40], which operates a given clustering algorithm several times, each time with a different value for the number of clusters, plots the explained variance of the clusters against the objective number of clusters, and eventually provides as optimal value that at the curve’s knee (or “elbow”), i.e., the point where the slope significantly changes. For instance, the Elbow method was used in [41] to automatically categorize the headlines from one thousand websites, and led to the identification of 7–8 major types of news, a pertinent piece of information to ease the research for readers. Nonetheless, the problem with this method is twofold: first, the curve’s knee is only an estimate, not an optimum, and a debate about the relevance of this point is discussed [42]; then, this method is essentially empirical, and not efficient, since the clustering algorithm must be executed several times to draw the curve whose knee must be found. Therefore, the Elbow method is essentially a sophisticated fashion to select the best partitioning among several of them, and not truly a way to provide an estimate for the optimal number of clusters.
Similarly, some methods trade the explained variance to more elaborated indicators, while keeping the same methodology than the Elbow method (i.e., finding the number of clusters which best maximizes the integrity of the clusters, while also providing a meaningful and representative partition of the feature space) [43]. Among others, some greatly popular criteria are the silhouette coefficients, which are integrity indicators providing, for every datapoint, a score indicating how close it is from its same cluster’s neighbors and how distant it is from the other clusters’ data. This approach gave birth to a dedicated clustering method, named silhouette clustering [44], which proposes to converge toward an optimal clustering (in the sense of the silhouette coefficients) by exhaustively and iteratively swapping the clusters’ barycenters and any other data to see if any improvement is obtained by this swap [45]. An example of application can be found in [43], who simulated a dataset composed of a certain number of noisy Gaussian distributions, and endeavored to retrieve the number of groups using the silhouettes-based version of the Elbow method; they obtained very good results, but the main problem with this method is its computational heaviness [46], since the silhouette coefficients themselves take a long time to compute ( O ( N 2 ) , with N the number of data), and the comparisons to every pair barycenter-data ( O ( k N ) , with k the number of clusters) are numerous. Therefore, the complexity of the algorithm is in O ( k N 3 i ) , with i the number of iterations before convergence, which does not make it well-suited to large datasets such as industrial ones.
In addition, another well-known method is the Gap statistic [47]. This exploratory analysis technique relies on a simulated ground truth, obtained by generating a reference dataset under the null hypothesis, i.e., with the same statistical characteristics than the original one. Then, both the original and simulated datasets are clustered with a same algorithm, with the same value set for the number of clusters, and the average dispersion (for instance the mean distance between any possible pair of points) of the clusters for both partitions are compared. This procedure is repeated for several values as number of clusters, and that leading to the lowest difference between the mean dispersion of both partitions is chosen as the optimal number of clusters to build. An example of their application is presented in [48], where the authors aimed to retrieve the number of groups in real reference datasets using the Gap statistic method; they obtained mitigated results though, mostly due to the parameters to set: when correctly setting the parameters, they were able to retrieve the correct number of groups, but, when the parameters slightly differed from their optimal values, they obtained poorer results, mostly due to close and overlapping groups. Nonetheless, although being deemed reliable [49], this technique suffers from two major drawbacks: its computational heaviness, and its empirical approach. Since several values of the number of clusters must be tested, it could be considered a more sophisticated method to select the best partitioning among others, rather than a true way to estimate the number of clusters in a dataset.
Finally, in order to estimate the number of clusters, the authors of [50] proposed the ECD test, a semi-empirical approach which operates in a split-and-merge fashion. To that end, the authors began by creating a cluster dataset (only once) using a large value for the number of clusters, resulting in possibly very small groups, which are later characterized by their Empirical Cumulative Distributions (ECDs), which are finally compared using the Hausdorff distance and linked to one another using region-growing clustering. By comparing the ECDs of the different groups, the idea is to regroup the clusters in the same region of the feature space (which are assumed to have similar features, and therefore close ECDs), which are eventually merged to empirically estimate the borders of the distinct regions of the space. The ECD test aims to estimate the true number of regions of the feature space, value which serves as an estimate for the optimal number of clusters. This technique has the advantage of performing the clustering only once, is computationally efficient, since it relies on simple comparisons (the ECDs and the Hausdorff distance), and proved to be accurate in the identification of the borders of the feature space of a real industrial dataset.
All these works aim to estimate the “optimal“ number of clusters for a clustering algorithm to build, either by following a semi-empirical approach or by a simpler selection of the best partitioning among a set; they are summarized in Table 1 to provide an overview of their advantages and limitations. However, some unsupervised clustering techniques are specifically crafted to perform without this parameter, in order to avoid the intrinsic difficulty one may encounter in estimating it.
Probably the most common category of unsupervised clustering techniques which build the clusters automatically is Hierarchical Clustering [51,52]. Such algorithms typically regroup data, or groups of data, by proximity or similarity in a hierarchical fashion. For instance, with a given dataset, the distance between every pair of data is computed, and the data of the pair with the lowest distance are linked together, forming the very first cluster; then, the procedure is repeated, but this time replacing the two previous data by their mean. As a consequence, all data are linked to each other in a hierarchical fashion, and the clusters can be obtained by stopping the linkage procedure somewhere. Knowing where to cut the link between the data/groups is the major drawback of hierarchical clustering, since it relies on a threshold (cut when two data/groups are at a distance higher than this threshold). In practice, Hierarchical Clustering was used in [53] to extract the manufacturing bottlenecks of a real-world production system from the event logs only, allowing the operators to identify the corresponding bottlenecks in the real system; it was also used by [54] in a more Industry 4.0-oriented context to propose a hybrid cut criterion in the scope of the design of a Decision Support System (DSS) for 3D printers.
Another relevant unsupervised clustering algorithm which does not require the number of clusters to build is Region Growing [55]. The general approach of such techniques is to draw a point from the database, aggregate its neighbors to it, thus forming a true cluster, and repeat the aggregation with any data contained within; once there is no more data close enough to any data in the cluster, it is considered fully built and another one is started following the exact same procedure, starting from any yet-unprocessed data. Region growing often achieves good results, but it is highly sensitive to initialization, and a threshold is required to characterize what “neighborhood” means, i.e., what is the maximal distance between two datapoints to consider them neighbors. It is popular in the segmentation of regularly spaced data, in particular images [56] and point clouds [57], since the data representing a same object are often similar, and a sudden change in values likely means an actual change of objects; for instance, when operating in the color space of an image, if two neighboring pixels belong to a same object, they should have approximately the same colors, but if they represent different objects, their colors should greatly differ. Nonetheless, with industrial processes, data are generally evolving continuously: sudden changes are rare, except when a shutdown or a power-on order is executed; as such, region growing may not be the most appropriate solution for real, dynamic industrial processes.
Another very popular algorithm is the Density-Based Spatial Clustering of Applications with Noise (DBSCAN) [58] algorithm, which operates somehow by region growing, but with a dedicated initialization procedure. Indeed, according to this method, the density of a cluster is defined as the number of data contained within a given topology ball. Therefore, the algorithm searches for local groups of data whose density meets a given parameter (e.g., if a datapoint has enough neighbors at a distance below this parameter, then it is assimilated as the cluster’s seed), regroups all of them to form a “proto-cluster”, before expanding it step-by-step by considering the neighborhood of the added data. Among other applications, DBSCAN was used in [59] to detect and segment the planar objects of a 3D point cloud, and it obtained good results on both academic and real datasets. It was also used in more industrial-related contexts, such as in [60], in which they compared it to the K-Means in the automatic detection of the different types of tasks performed by automated guided vehicles; DBSCAN proved able to better pinpoint the real tasks historically executed by the vehicles, knowledge which may serve for anomaly detection as a possible future refinement.
DBSCAN generates good results, but has the disadvantage of relying on a threshold as the first clusters’ radius (delineating the neighborhood of the seeds), and can only use one value for the density. To compensate that limitation, another algorithm, the Ordering Points To Identify the Clustering Structure (OPTICS) [61] replaces the fixed value for the neighborhood by a maximal value, thus every cluster smaller than this maximal threshold is considered (resulting in possibly varying densities). The counterpart is that this leads more to hierarchically connected groups of data rather than a true partitioning. Moreover, even though the impact of the threshold is leveled, it still must be set. As examples, OPTICS was used in [62] to localize the defects of the rolling parts of real systems, obtaining fairly encouraging results, and was also applied to Wireless Sensor Networks to detect the outliers, to serve as a semi-automatic pre-processing tool [63].
All these techniques of unsupervised clustering do not require the objective number of clusters to perform; nonetheless, they all have their own drawbacks. Some are highly sensitive to initialization, some others trade the number of clusters of the more traditional clustering methods for a neighborhood threshold, which sill needs to be provided upstream, and conditions the number of clusters issued, and some others are more a way to build a network of clusters, whose inter-distances would be represented by the length of the edges, rather than a true partitioning. The main methods of unsupervised clustering introduced in this section are summarized in Table 2, where their advantages and limitations are shortly mentioned to provide an overview of these methods.
This detailed literature review shows that many methods exist to, on the one hand, estimate the optimal number of groups present in a given dataset, and, on the other hand, to propose clustering algorithms which would overcome that limitation, by operating in an iterative fashion. Nonetheless, most of the already-existing techniques are either a sophisticated manner to select the best partitioning among others (e.g., the Elbow method), are computationally heavy (e.g., the silhouette coefficients), or replace that knowledge by another as of a threshold. As such, there currently exists no reliable and non-parametric solution to automatically identify and extract the true regular states (the behaviors) of real unknown systems, whence the proposal of the SPRADA method.

3. Required Already-Existing Materials for the Proposed Methodology

The proposed methodology, the SPRADA method, aims to identify the main region of an unknown feature space, by using a recursive and self-assessed approach. The proposed solution does not introduce new concepts sensu stricto, for it is essentially a methodology, i.e., how to use already-existing tools to carry out the proposed behavioral identification. As a consequence, in order to carry out this study, some concepts must be introduced, in particular unsupervised clustering and the metrics to quantify the integrity of the clusters. These topics are the matter of the two following subsections.
In all the following, d refers to any mathematical distance (e.g., Euclidean), the data instances are represented by x, and the database is denoted as D and is defined by (1),
D = { x i } i [ [ 1 , N ] ] ,
with N = | D | the cardinal (size) of the dataset, the final set of clusters is defined by (2),
C = { C k } k [ [ 1 , K ] ] ,
where K is the total number of clusters, with any cluster C k typically defined by (3),
C k = x D | argmin k [ [ 1 , K ] ] d ( x , p k ) = k ,
where p k is the representative of cluster C k , outputted by the learning stage.

3.1. Unsupervised Clustering

Unsupervised clustering is the task which consists in regrouping data sharing similarities with each other, while also separating those which do not. For instance, the data contained within a same cluster should be no more distant than a given value, but two data from two distinct clusters should be more distant. Notice that any mathematical distance may be used, possibly leading to small variations in the results; typically, it is the Euclidean distance which is used. Additionally, the notion of “closeness”, or “similarity“, depends on the space where clustering is being operated; typically, it is the feature space which is chosen for that purpose. In the case of a real industrial system, the feature space may be that spanned by the basis formed by all the sensors of the system.
Probably the most well-known unsupervised clustering algorithm is the K-Means (KM) [64] algorithm. Albeit naive, it is also the most intuitive fashion to regroup data. It starts by drawing K points from the database, the barycenters, then it associates any point of the database to the closest barycenter to form the groups, the clusters. Once they are built, the barycenters are updated as the clusters’ means, and the procedure is repeated. The learning stops when any criterion is achieved, for instance a maximal number of iterations, or once the barycenters change very little from a round to the next. However, the K-Means has two limitations: it is not computationally efficient, and is only able to separate linearly separable groups (at least, in an Euclidean space).
An accurate alternative to the K-Means is the Self-Organizing Maps (SOMs) [65] algorithm, which operate similarly, but adds a notion of neighborhood to the learning. Indeed, there are still K points drawn first and assimilated to the clusters’ barycenters, but now they are virtually linked to each other in the form of a grid. Then, the learning procedure consists of statistically learning from the dataset. To that end, a datapoint is randomly drawn, and every barycenter is attracted to it, by applying a learning function updating the position of the barycenters, to bring them closer to the data drawn. The update is the strongest for the closest barycenter, and its strength diminishes as the farther away the barycenters get, and the less connected to the first one they are (i.e., the update is stronger for the direct neighbors of the closest barycenter, is weaker for the neighbors of the neighbors, and so on). Although aging, the SOMs are still very popular for their accuracy and efficiency: for instance, they found practical applications in [66] to evaluate the quality of groundwater, and in [67] to track the metaloids contained in soils. The SOMs are very accurate, are computationally efficient (due to their statistical approach), and proved able to distinguish most regular states from real industrial datasets [30], which makes them appealing for the identification of the main regions of industrial processes with the SPRADA method.
The SOMs are very general, work well in many cases, but some local adjustments may increase their accuracy and representativeness: Geographical SOM (Geo-SOM) to take into account spatial dependency [68], Time-Adaptive SOM (TASOM) to use time-adaptive parameters [69], Growing SOM (GSOM) to use a self-growing map [70], etc.
A useful alternative may be the Bi-level SOM (BSOM), proposed in [71]. This variant was originally crafted to identify the true regular states of a real industrial system, by somehow averaging several SOMs. Indeed, the main drawback of any statistical approach is that they are often sensitive to initialization: the very first points drawn will greatly affect the results, and the meaningfulness of the final partitioning; however, it is often very difficult to judge the quality of a space partition without any prior knowledge, thus there is no real way to decide which SOM is the most representative, and even if there is any. Nonetheless, in general, two SOMs will result in close partitioning, or, at least, some motif should appear, but each may be correct to some extent if no ground truth or any knowledge is made available, especially in a blind, Data Mining-oriented context. To answer that question, a BSOM aims to take benefit from the self-organization capability of the SOMs to average several maps as one, by generating a set of several SOMs, and then projecting all of them into a final one, somehow representing a statistically averaged map. In real situations, the BSOMs proved to be more accurate than the original SOMs and the K-Means to automatically identify the regular states of a real industrial system (its behaviors) [71].
The three main methods discussed in this subsection are summarized in Table 3 to provide an overview on their strengths and weaknesses. These three unsupervised clustering algorithms can all be used in the proposed SPRADA methodology, and their respective accuracy will be discussed in Section 5.4. The advantage of using fairly simple algorithms (compared to that of Table 2) is that they can be performed several times.

3.2. Quantification of the Integrity of the Clusters

Once unsupervised clustering has been applied, the issued groups can be characterized in some fashion. Several metrics may be considered to that purpose: data dispersion, cluster compactness, span, density, etc. Whilst the traditional mathematical moments (mean, standard deviation, etc.) provide good indicators, they are generally not sufficient to correctly quantify the quality of the groups; additionally, they do not take into account the possible relationship between the clusters themselves. Actually, two aspects are worth being considered: how the data contained within a given cluster are organized between each other, and how the clusters are positioned in relation to each other. Indeed, if the data contained within the same cluster are very compact (e.g., low standard deviation), but that the cluster is close to another one, itself compact, therefore, both may be merged. On the contrary, if a cluster comprises a lot of data, but that are widely scattered, the cluster may be worth being split into several more compact pieces. The best case would be to obtain highly compact groups (low standard deviation of the data contained within each), but also highly distant from each other; for instance, the maximal distance between two data of any cluster would be ε , and the minimal distance between two data of two different clusters would be several times greater, such as 10 ε . In such a case, the clusters would be easily distinguishable from one another; however, this rarely happens with real systems, the clusters are never as compact, nor as distant from each other, and overlaps often occur.
A very well-known metric used to estimate the quality of a partition is the silhouette coefficients (or just silhouettes) [72]; they are now quite old, but are still in use by many works, such as by silhouette clustering. The silhouettes are normalized scores (ranging from 1 to + 1 ) which indicate if a datapoint is close to its same cluster’s neighbors, while also being distant from the data of the other clusters. To do so, for every datapoint x, the metric computes the mean distance avg, defined by (4), to each of its neighbors, then, for every other cluster, it computes the mean distance dis, defined by (5), between x and every datapoint of that cluster, and it finally compares the two values by subtracting both and by dividing by the minimum of both to normalize the score, as expressed by (6). The closer to + 1 , the closer the data to its neighbors and the more distant from the other clusters, and vice-versa for the closer to 1 . The silhouettes generally provide a good indicator of the integrity of the whole partitioning, but take a long time to compute (complexity in O ( N 2 ) ), and there are too many coefficients to consider them all, thus, usually, only statistics of them are used (such as the average of the silhouettes of the data comprised by a cluster). Nonetheless, the computational complexity of the silhouette coefficients makes them ill-suited to very large datasets, such as industrial ones.
x i C k , avg ( x i ) = 1 | C k | 1 x j C k { x i } d ( x i , x j )
x i C k , dis ( x i ) = min k [ [ 1 , K ] ] k k 1 | C k | x j C k d ( x i , x j )
x i C k , sil ( x i ) = dis ( x i ) avg ( x i ) max { dis ( x i ) , avg ( x i ) }
Another relevant metric to estimate the integrity of the clusters is the Average Standard Deviation (AvStd), which was the original metric used in the T-DTS [73]. It is a simple and relatively unrepresentative measure of data scattering, even though it was sufficient to test and assess the recursive decomposition principle of the method. The Average Standard Deviation is aptly named: it is defined as the average of all the standard deviations of a dataset, computed along every dimension. Indeed, when dealing with a n-dimensional space, the traditional standard deviation does not really make sense, since it is typically computed with scalar values, not vectors; nonetheless, with real systems, especially when using machine-learning-based tools, the input data are very often vectors. Therefore, the AvStd computes the standard deviation σ ( i ) , defined by (7), of the data of a given set (e.g., the database D or any cluster C k ) along every dimension i of the space, with m the dataset’s mean (or m k for cluster C k ), and eventually returns the mean of all these values, as expressed by (8). As such, the AvStd is a fair indicator of data compactness.
k [ [ 1 , K ] ] , i [ [ 1 , n ] ] , σ k ( i ) = 1 | C k | x C k m k ( i ) x ( i ) 2
k [ [ 1 , K ] ] , AvStd k = σ ¯ k = 1 n i = 1 n σ k ( i )
Eventually, in order to compensate the simplicity of the AvStd and the computational heaviness of the silhouette coefficients, another metric was proposed as a replacement, the Hyper-Density (HyDensity) [74] metric. This metric redefines the concept of density in computer science; indeed, many assimilate the density of a group of data to the number of items contained within a given radius from the barycenter. This works fine in 2- or 3-dimensional spaces, but less when the dimensions increase. Therefore, Hyper-Density proposes to use the theory of the hyper-volumes to estimate the n-dimensional volume that a group of n-dimensional data occupy, and divides this estimate by the actual number of items contained within. This is close to the Physics’ specific mass, thus, to obtain a true density, this value is eventually normalized by a reference, such as the maximal values found among all the clusters. Applied to both academic and real datasets, the Hyper-Density proved to be more accurate than the AvStd in the identification of the issuing clusters. Several hyper-volumes may be considered to estimate the volume of a group of n-dimensional data, but the hyper-sphere appears to be the most democratic one; according to [75], the hyper-volume V ( n ) of a hyper-sphere of radius r is given by (9),
V ( n ) r = r n . π n / 2 Γ n 2 + 1 ,
where Γ is the Gamma function. The radius r of the hyper-sphere may be estimated as half the span of the dataset S of mean m, or simply by the maximal distance from any point and that mean, as expressed by (10).
r S = max x S d ( x , m )
The specific mass ρ S of S is obtained by dividing the number N S of data contained within by the volume V ( n ) of the hyper-sphere containing all of them, as denoted by (11),
ρ S = N V ( n ) ( r S ) .
The final Hyper-Density is eventually given by dividing the specif mass ρ S by a reference ρ r e f , as expressed by (12),
ρ S = ρ S ρ r e f .
In the case of a set of clusters C , the reference ρ r e f may be chosen as the maximal value found among all the clusters, as given by (13),
ρ r e f = max C k C ρ C k = max C k C ρ k .
The three quantification metrics discussed in this section are summarized in Table 4. In this study, the AvStd and the Hyper-Density will both be used as integrity indicators of the clusters for the recursive decomposition, and the silhouettes will be used to compare the partitionings obtained when using different clustering algorithms and quantifiers.

4. Proposed Solution: The SPRADA Methodology

The proposed methodology can be summarized as an optimizing decomposition of a space, by recursively analyzing the integrity of the clusters and deciding whether or not a new stage of partitioning could improve the results. That idea is intuitive, and was investigated about two decades ago, resulting in a multi-modeling method called the Tree-like Divide To Simplify (T-DTS) [73], whose first aim was to reduce the complexity of real, dynamic and nonlinear processes by locally splitting their feature space into smaller and more homogeneous regions [76], which was later extended in [77] to the modeling of such systems by locally modeling any of their identified regions.
This method was first crafted to approximate functions, or to model complex systems, by recursively decomposing a database, then building local models upon each cluster, and eventually connecting all of them in some fashion to provide a global model. However, it was not oriented toward anomaly detection nor behavior identification; with Industry 4.0, these two topics are very important to understand the industrial systems in depth, reason why a smart, automatized and recursive decomposition method can be of great help.
Therefore, the proposed solution, the SPRADA method, takes the recursive decomposition principle of the T-DTS, but extends its capability to the identification of the regular states (the behaviors) of real industrial processes, possibly in the scope of anomaly detection and diagnosis. As a methodology, any unsupervised clustering method may be used, and any cluster’s integrity indicator as well; nonetheless, with respect to the observations made in Section 3, the Bi-Level Self-Organizing Maps and the Hyper-Density will be used as default settings, for they both proved to be well-suited to industrial contexts in previous works. Additionally, since the main drawback of the T-DTS is its habit of issuing many groups, sometimes comprising very few data (and even a single one), groups possibly usable for modeling but less for anomaly detection (since such groups are rarely meaningful, and are hard to classify), the SPRADA method includes a merging stage based on the empirical distributions of the clusters (inspired by the ECD test), to proceed in a split-and-merge fashion. This last stage aims to clarify the contours of the clusters, which are expected to represent true industrial anomalies or behaviors.

4.1. Recursive Decomposition Principle

The proposed SPRADA method relies on a recursive decomposition of a database. To that end, a database D is first split into a set of several clusters using any unsupervised clustering, and the integrity of every of them is then computed using any characterizing metric. Then, every cluster whose integrity is judged as low is split, and the integrity of the so-built groups is analyzed, and a new step of split is once again applied if needed. This procedure is repeated until the integrity of any cluster is good enough, or stops when a group contains only one single data. As a consequence, the split process resembles a tree, whose root is the full database, and the leaves are the clusters. The general operating principle of the recursive decomposition is depicted in Figure 1.
The last concern is to choose the value to which set the split criterion to. To answer that question, to assume the “Self-Parameterized” property of the proposed method, one may make a weak assumption: at least one region of the feature space is salient and is correctly isolated by clustering. From experience, this is often true, even with real industrial data: there is almost always at least one state which stands out, for instance the steady state, since the system should behave this way most of the time, thus the database should comprise a lot of data pertaining to it. Therefore, since this assumption is very often true, this cluster may serve as a reference, since it should have the best integrity among all, and can as such provide an idea of what “good integrity“ means with the current database (whichever may it be). As a consequence, as a split criterion, one may consider using a percentage of that maximal value, thus providing a dynamic and self-adaptive split criterion. A good value may be around 10% in general, but it depends on the inner scattering of the database.
The pseudo-code of the recursive decomposition is given as Algorithm 1 to help the reader understand the different steps introduced until now. In this example, it is the Hyper-Density which is used as split criterion, but it may be replaced by any other metric; one should only pay attention to the “sense” of the criterion: it can be the higher or the lower, depending on the metric. For the Hyper-Density and the silhouettes, the closer to + 1 , the better, but, for the AvStd, the closer to 0, the better. Therefore, the word “higher” in the if statement on line 3 may be replaced my the word “lower” if needed.
The recursive decomposition principle is fairly intuitive, and has the advantage of greatly diminishing the impact of a wrong value set for the objective of clusters, if any. For instance, assuming the database actually features K clear regions, and that one partitions it using the K-Means such that K < K , thus some regions will be merged (best case), or even wrongly split (worst case). Nevertheless, assuming that merging two dissimilar regions with each other will lead to a unique cluster with a low integrity is not a so strong assumption; as a consequence, the problematic cluster would be detected and split to cluster it (more) correctly. However, this may lead to numerous small clusters, and the two regions considered may be split into many pieces, leading to inconsistent groups, likely poorly meaningful. That being said, all the data of a given region should have very similar features (especially compared to that of the other regions), thus, if that region is decomposed into several groups, they should all be very similar, and thus it should be fairly possible to compare them, and then to merge them to reform the original region.
Algorithm 1. Pseudo-code of the SPRADA’s recursive decomposition principle
Make 05 00051 i001

4.2. Merging of the Clusters

Recursive split has the bad trend of generating numerous clusters, which may comprise possibly very few data; as a consequence, it may be hard to use the results as they are, with no further refinement. Therefore, the proposed SPRADA method includes a dedicated step of merging, in order to rebuild the regions of the feature space in a split-and-merge fashion. Indeed, the appearance of very small clusters is generally due to the split of a larger, not highly compact nor consistent regular state (behavior); in such a case, i.e., when a same behavior is split into several clusters, their respective data should have close features, at least to some extent (especially compared to that of the other behaviors), and may therefore be compared and regrouped fairly easily.
To that end, the SPRADA method will use the merging procedure introduced in the ECD test [50], which proved to be accurate when comparing (and merging) clusters obtained with real industrial datasets. The ECD test computes the Empirical Cumulative Distributions (ECDs) of every cluster, along every dimension of the space, then it computes the mean of the ECDs along these different dimensions, and eventually compares the mean ECDs with each other using the Modified Hausdorff Distance (MHD) [78]. That metric was chosen for it is computationally efficient and provides good results to compare two curves, and since an ECD is somehow the empirical probability of an experimental realization of a continuous random variable, the MHD is a very good candidate to compare two of them. Finally, the ECD test proposes to link the different clusters by using a region growing clustering to rebuild the main regions of the feature space, since this approach does not require the objective number of clusters to form.
For a cluster C k = { x n } n [ [ 1 , N k ] ] , with N k = | C k | , the empirical probability P k of a real value x is the proportion of data whose value is lower than x, as expressed by (14),
x R , P k ( x ) = 1 | C k | x n C k 1 x n x ,
where 1 x n x is the indicator function defined by (15),
x n , x R , 1 x n x = 1 , if x n x 0 , else .
The ECD of cluster C k is defined as the empirical probabilities computed for every real number, as given by (16). Notice that, in fact, an ECD is generally discretized, and is therefore only computed for some values of the domain of the cluster’s data, spaced by a constant step for instance.
E C D k = P k ( x ) x R
Finally, the Modified Hausdorff Distance of two discretized sets (such as curves) P = { p i } i [ [ 1 , N P ] ] and Q = { q j } j [ [ 1 , N Q ] ] is given by (17),
d h P , Q = max h ( P , Q ) , h ( Q , P ) ,
where h is the mean of the minimal distances from every point of a set and any point of the second set, as expressed by (18). One may notice that h ( P , Q ) h ( Q , P ) , since h is not symmetric, whence the computation of both in d h ,
h ( P , Q ) = 1 | P | x i P min x j Q d ( x i , x j ) .
Once all the ECDs have been obtained, and the MHDs of every pair of them have been computed, the clusters can be compared to each other to know if they are close or distant in the feature space, with respect to the empirical distribution of their respective data. Although this comparison may be performed in an expert system fashion, it makes more sense to delegate the work to machine learning instead, for instance by using region growing clustering, which would compare the ECDs to each other and link the clusters to each other with respect to the closeness of their ECDs, represented by a low MHD.

4.3. Proposed Methodology for the SPRADA Method

The flowchart of the overall proposed SPRADA method is depicted in Figure 2. A few words may be worth being said to clarify some specificities of the study in Section 5.
First, any unsupervised clustering algorithm may be used, but the BSOMs are good candidates, for they are accurate, and perform well with real industrial data, especially in the identification of the regular states of industrial systems [71]. The grid’s size still must be set, but the recursive decomposition approach diminishes the importance of that value; as such, a default 3 × 3 map is more than enough to cover almost any cases. Notice that several clustering algorithms may also be used in the different stages of the decomposition (for instance, a 3 × 3 BSOM for the initial partitioning, then a K-Means, with K set to 2 or 3 for the recursive decomposition of the clusters to limit the number of groups generated).
Second, any metric may be used as split criterion, but the Hyper-Density seems to be a good choice, for it proved to be accurate when characterizing real industrial processes [79]. The silhouettes may also be considered, but they are long to compute, thus they may not be adapted to large industrial datasets. Notice that the Hyper-Density does not take into account the relationship between the clusters, but the recursive approach of the proposed decomposition overcomes that limit. Additionally, the split criterion may be set to 10% of the maximal Hyper-Density found among the clusters obtained at the initial partitioning.
Third, for the final merging by region growing, a similar threshold may be obtained by searching for the maximal MHD between every pair of ECDs (representing the most distant clusters), and using a ratio of this value.
Eventually, the final set of refined clusters should be representative of the main regions of the feature space, obtained in a split-and-merge fashion. Such an approach aims to obtain clear groups’ contours, may the groups be the regular states of the system (behaviors), or true anomalies historically encountered. These groups bring a greatly useful knowledge on the system, to automatically define what are its normal ways to behave, and what an anomaly is for that given system (indeed, the definition of anomaly may greatly vary from a system to another, and even from a process to another).

5. Application to a Real Industrial Dataset

Now that the SPRADA method is introduced and explained, it must be assessed over real industrial data. First, the database used to conduct this study will be presented, then the proposed methodology will be applied to it, and the obtained results will be discussed and compared to more standard unsupervised clustering techniques from literature.

5.1. Presentation of the Database

First, it should be mentioned that this work takes place within the European project HyperCOG, which aims to study the Industry 4.0-orientated cognitive plant. To that end, some of the partners develop a CPS to connect the production units and management tools with each other, whilst the others work on the automatic extraction of knowledge so as to build some form of intelligence which may help operators manage and control their systems in a semi-supervised fashion. One important topic of the project is anomaly diagnosis, i.e., detecting, identifying and investigating how to correct such anomalies. As discussed in Section 1, this can be performed using dedicated tools oriented toward (automatic) anomaly diagnosis, but also by identifying the regular behaviors of the system, since an anomaly should be any differing state, whence the proposal of the SPRADA methodology.
One of the industrialists involved in the project HyperCOG is Solvay, a chemistry plant located in La Rochelle, France, specialized in rare earth specialty products. In order to test and assess the solutions developed during the project, they provided the partners with several datasets recorded in real situations. One of them is a two-month long record of fourteen key sensors, operating at a frequency of one data every five minutes, for a total of 17,568 samples. Over that period, the processes worked correctly most of the time, although some local events occurred nonetheless (e.g., short maintenance or true anomalies). Notice that the data of two of these sensors were almost the same than that of two other sensors, which have therefore been removed to avoid redundancy. Additionally, the sensors have been normalized between 0 and 1, and their names and dimensions have been randomized for confidentiality concerns. The corresponding dataset is depicted in Figure 3.

5.2. Average Number of Clusters Issued by the Recursive Decomposition

First of all, since the unsupervised clustering algorithms considered in this study are stochastic, they may issue a different number of clusters when executed several times, and so does the overall SPRADA method, which is, by extent, itself stochastic. As a consequence, the results may change from a try to another when applying the proposed methodology, depending on the number of clusters issued. To remain as objective as possible, typical clusterings will be considered during the whole following study. Nevertheless, the concept of “typical“ is debatable: to propose a definition, a certain number of tries may be performed, and the average number of clusters of all of them may be computed, which may serve as a “typical” number of clusters issued by the recursive decomposition when using a specific unsupervised clustering method and a given quantification metric.
To estimate the number of tries to perform to obtain reliable results, one may use the Cochran equation, which provides such a number η with a given confidence interval, as defined by (19),
η = p ( 1 p ) Z m 2 ,
where m is the margin error, p is the part of the population with the desired characteristic, and Z is the standard score (Z score), directly yielded by the objective confidence level. By setting the margin error m to 3% and the confidence level to 95% (whose corresponding Z score is 1.96), and by considering the most random value for p of 0.5 (with m and Z set, p = 0.5 is the global maximum of η , thus the less meaningful case, since it means that no information is available on the population), the Cochran equation yields η 1067 .
As such, the recursive decomposition of the SPRADA method is performed 1067 times, with every pair of clustering method and quantification metric, and the average number of clusters μ and the corresponding standard deviation σ are computed, and summarized within Table 5. The K-Means are performed with K set to 2 in order to operate a binary decomposition. The SOMs use 3 × 3 grids, since any real system is crafted to behave in a handful of different ways (steady state, power-on and shutdown procedures, and some transient states in between), thus a set of ten objective clusters is generally more than enough, as pinpointed in [30]. The BSOMs are performed using ten SOMs of size 3 × 3 , which are generally enough to provide a reliable and meaningful averaged map [71].
A couple of remarks can be drawn from the table. First, the K-Means-based version of the recursive decomposition provides the lowest average number of clusters, due to the binary decomposition (few clusters contain very few data), and to the intrinsic operating principle of the K-Means: the data are gathered with each other with respect to their proximity in the feature space, thus the compactness of the clusters is likely fairly high, whatever the homogeneity of the data. On the contrary, the SOMs provide the largest number of clusters, due to the fairly large size of their grids (nine clusters); indeed, having as many clusters refines the partitioning, but also leads to some clusters containing very few data, whence the larger number of clusters. Eventually, due to the averaging approach of the BSOMs, the number of clusters issued is with no surprise lower than with the SOMs.
Notice also that the standard deviation is the largest for the K-Means, indicating that this algorithm can provide a highly varying number of clusters from an execution to another, making the results poorly reproducible. On the contrary, the SOMs have a slightly lower standard deviation due to the better homogeneity of the data contained within the clusters, and the averaging approach of the BSOMs makes them fairly more consistent from an execution to another (reason why the BSOM approach was originally crafted).
Finally, one may also notice that the average numbers of clusters are similar when considering both quantifying metrics, but the standard deviation is slightly lower with the average standard deviation. Indeed, the main drawback of the Hyper-Density is its sensitivity to outliers, and since the recursive approach of the SPRADA method generally issues small clusters, which can more or less themselves be assimilated to outliers, the Hyper-Density initiates a new round of recursive split.

5.3. Application of the SPRADA Method to the Real Industrial Dataset

In order to assess the proposed methodology, it is applied to the industrial dataset of Figure 3. In this subsection, in order to illustrate the operating principle of the SPRADA method, only the Self-Organizing Maps will be used for clustering, and the Hyper-Density for the quantification of the integrity of the clusters. The impact of the clustering algorithm and the quantification metric will be discussed in Section 5.4.
In this subsection, the SPRADA method is performed in the feature space spanned by the twelve sensors of the database, using all of them during every stage of clustering and quantification. The clustering method used throughout this study is a 3 × 3 SOM, and the quantification metric is Hyper-Density with a hyper-sphere used as delimiting shape.
First, the recursive decomposition principle is illustrated by Sensor 12, as depicted in Figure 4; indeed, even though all the sensors are used, only one is depicted for the sake of conciseness. In the figure, the different stages of the recursive decomposition are represented, one row for every time a split occurs (except top row), i.e., when a cluster obtained a low Hyper-Density score; here, six splits occurred. For every row but the top, on the left column is the current state of the partitioning, with the issuing cluster in red, and on the right the clusters generated by the SOM after the split are represented; for every of these rows, the title of the left graph indicates the depth of the recursion, and the title of the right graph indicates which cluster is split. For instance, the three graphs from rows two to four are linked to each other: “Recursion 0” and “Cluster 0” of the second row indicate it is the cluster 0 of the “Initial Clustering” (right hand corner graph) which is split, and whose subclusters are represented on the right; then on the row below, “Recursion 1” means that it is one of these clusters (“Cluster 3”) which is split, and on the fourth row, “Recursion 2” means that is one of the clusters (“Cluster 6”) of “Recursion 1” which is split.
The recursion decomposition principle is efficient, but has the drawback of generating some clusters comprising very few data, especially as the recursion depth increases. For instance, on the forth row of Figure 4, which represents a recursion of depth two, the cluster split (in red in the left graph) is fairly small, thus splitting it generates several very small clusters (right graph), which may be assimilated as outliers there. However, as small as they may be, they may also be part of true regular behaviors (or anomalies), whence the second stage of the SPRADA method, i.e., the clusters merging.
To provide an idea on the possible results of the complete recursive decomposition, Figure 5 depicts the outputted set of clusters for the twelve sensors of the database. This partitioning corresponds to the six splits of Figure 4, for a total of fifty clusters.
This figure is barely explainable, for they are too many clusters (the reason why a special focus on a single sensor will be made in the next subsection); nonetheless, the will behind this is to show that the recursive decomposition tends to produce many clusters, sometimes comprising many data (as those represented by the blue, orange, green and red crosses on the figure), but many other comprise very few; as a consequence, it is generally not possible to make sense of the clusters as they currently are, for they do not really represent anything in their present format. Nevertheless, the recursive decomposition aims to produce local, compact clusters, which may be some pieces of true, wider regions of the feature space, which may be reconstituted by merging the clusters with each other, whence the second stage of clusters merging of the SPRADA method, based on the ECD test.
To illustrate the cluster merging process, the ECD test is applied to the clusters of Figure 5, with the coefficient ε set to 66% (cf. Algorithm 1). The results of the merging are depicted in Figure 6; the process eventually generated four clusters, two of which are very small (the green and red crosses, containing two and four datapoints, respectively), and two are large regions (the blue and orange crosses, containing 10,413 and 7149 datapoints, respectively). After having discussed the results with the experts of the industrial partner, it appears that the blue cluster distinctly corresponds to the state when the systems are running (turn-on procedure and steady state), whilst the orange cluster corresponds to the state when the systems are stopped (power-off procedure and rest mode).
Finally, with its split-and-merge approach, the SPRADA method recursively split the feature space of a real dynamic industrial system into many clusters, sometimes large, sometimes small, and then merged all of them to identify the two main regions of the feature space (the steady and rest states). These results are encouraging, since the SPRADA method is automatized (unsupervised clustering), and requires no parameter to be set (except the coefficient ε ). Notice that a similar study was carried out in [30], but using the K-Means and SOMs, which achieved poorer results (although acceptable), which underlines the usefulness of an automatized, split-and-merged-based approach such as SPRADA.

5.4. Impact of the Clustering Methods and of the Quantification Metrics

The previous section illustrated the operating principle of the SPRADA method, and showed that it works fairly well to identify the major regions of the feature space of a real industrial system. Nevertheless, only one clustering method (SOM) and one quantification metric (Hyper-Density) were considered. Therefore, this present section focuses on the study of the impact of different clustering methods, and of different quantification metrics as well. To that end, all the methods introduced in Section 3.1 and Section 3.2 will be considered, i.e., the K-Means, the SOMs and the BSOMs for clustering, and the AvStd and the Hyper-Density for the quantification, for a total of six variants. Notice that only the recursive decomposition stage should be directly affected by these changes, the merging process remains the same. As such, the SPRADA method will be applied to the dataset of Figure 3, with each of these six variants. Moreover, even though the twelve sensors are used to carry out this study, only Sensor 12 will be depicted for the sake of clarity and readability.
Notice that, since the number of clusters may differ from one test to another due to the stochastic property of unsupervised clustering, the results presented in the following have been selected so as to be compliant with Table 5. In practice, this means that, for a given pair of clustering methods and quantification metrics, the SPRADA method has been performed several times until it obtained a partitioning whose number of clusters is close to the typical number of groups one may expect when considering that pair: it is that partitioning that is presented and discussed thereafter. This is performed on purpose to ensure an objective comparison of the different variants; nonetheless, since the SPRADA methodology is very general, it works the same with any partitioning.
The results of the recursive decomposition process are depicted in Figure 7, and that of the merging process are depicted in Figure 8; notice that the position of the graphs are consistent across the figures, e.g., the clusters represented on the left hand corner of Figure 8 are the merged version of the clusters represented on the left hand corner of Figure 7. On both figures, the two graphs of the same row correspond to the same clustering technique (K-Means at the top, SOM in the middle, and BSOM at the bottom), and the three graphs of the same column correspond to the same quantification metric (Hyper-Density for the left column, ant AvStd for the right one). Additionally, the graphs on the same row share the same ordinates, and those on the same column share the same abscissa. Additionally, every cluster uses a unique pair of symbol-color for representation (all the data represented with a same scheme belong to a same cluster), and the number of clusters is indicated in the bottom right corner of every graph. The symbol-color schemes are not consistent across the graphs (they do not represent the same regions of the feature space).
In all of the following, in order to simplify notations, the abbreviation “KM” is used to designate the K-Means, and “HyDensity” for the Hyper-Density. Additionally, any variant is referred to as a pair “clustering-metric”, e.g., “KM-HyDensity” to designate the variant using K-Means as the clustering algorithm, and Hyper-Density as the quantification metric.
Regarding Figure 7, a noticeable fact is the fairly low impact of the quantification metric on the results: for the same clustering method, the main clusters (i.e., the largest ones) exist in the same region of the feature space. Nonetheless, the major difference is that some large regions have been isolated as lone clusters with a metric, but have been split into many pieces with the other one (e.g., the region covered by the blue cluster of the KM-HyDensity variant is split into the orange, green and red clusters of the KM-AvStd variant). In general, Hyper-Density led to slightly more clusters, but several contain few data, and led to a significantly better identification of the main regions of the space.
Nonetheless, having a large number of (even) small clusters may have a beneficial impact on the results when dealing with a split-and-merge approach, for the main regions can be identified correctly, but also include some outliers, or data belonging to another region. Cutting them off the major part of the original cluster, therefore issuing a possibly very small cluster, may increase the total number of clusters, and increase the scattering of the groups, but this also makes a possible fine-grained merging operation feasible.
The results of the merging process (Figure 8) tend to confirm that statement: no matter the clustering algorithm used, Hyper-Density led to the identification of the two major regions of the feature space (the steady and shutdown states), whilst the results are a little more mitigated when using the AvStd, leading to some cluster’s splits here and there. Notice that the ε coefficient was set to the same value, except for the BSOMs, for which it was a little increased to compensate the averaging property of that method. In any case, whatever may be the actual number of clusters obtained, another very important remark is that, for a given quantification metric, the clusters are about the same for the three clustering methods considered: this can be easily observed on the left column of the figure, corresponding to Hyper-Density, but it is also true for the right column, where the identified groups are very close to one another, give or take a few refinements (e.g., the red cluster of KM-AvStd is merged with the green one in SOM-AvStd). Notice that the final clusters may change a little depending of the value set for ε , which is sensitive to the quantification metric; indeed, since it represents the percentage of the best value found among all the clusters, if the quantification metric is accurate, the values may greatly differ from each other, thus the coefficient ε may be slightly lower, and the other way around.
These two statements are very important: both metrics led to the identification of about the same regions of the feature space, and, provided a given quantification metric and a coefficient ε , the refined versions of the clusters (i.e., after the merging process) are fairly close, no matter the clustering method used. This means that the SPRADA method is highly resilient and reliable: with its split-and-merge approach, it will likely always provide close results, no matter the clustering algorithm used nor their parameters set (i.e., K for the K-Means, the grid’s size for the SOMs, and the number of maps for the BSOMs).
Finally, the mean silhouette coefficients of the merged versions of the clusters are summarized into Table 6 in order to quantify that qualitative comparison. As a reminder, there is one silhouette per data, thus the values provided are the averages of the coefficients of all the data of any given cluster, referred to as “cltX” in the table; additionally, there is no such coefficient for a single datapoint, thus the value provided for the clusters containing only one data are designated by a “X”. The results of the table underline the remarks drawn above: all the partitionings have obtained fairly good scores, indicating that the clusters are fairly compact and isolated from each other. Additionally, except BSOM-AvStd, they obtained close scores (around 0.70): neither the clustering method nor the quantification metric greatly affected them, indicating that the split-and-merge-oriented operating principle of the SPRADA method is resilient to both, and can therefore be easily parameterized.
Nevertheless, the fact that the BSOM-AvStd obtained a lower score seems to indicate that a BSOM may not be the most appropriate solution for the SPRADA method. Indeed, it was originally crafted to average several maps, in order to diminish the number of outlying clusters, and obtain more compact groups; however, that leads to a fewer number of clusters (compared to the SOMs), and thus affects the split-and-merge approach, for there are less to merge, whence the results were slightly worse than with the two other clustering techniques.
As a conclusion, in this section, the SPRADA method proved to be accurate in the identification of the major regions of the feature space, especially the steady and shutdown states, even though finer identification may have been possible by adjusting the coefficient ε , which is the only true parameter to be set here, any others being more some local adjustments to speed up performances or slightly refine the results.

5.5. Comparison to Traditional Clustering Algorithms

The last study to assess the proposed SPRADA methodology is to compare the obtained results to those issued by more standard unsupervised clustering algorithms, in particular those discussed in Section 2 and Section 3.1. To that end, the three algorithms used during the recursive decomposition stage, namely the K-Means, SOMs and BSOMs, will be executed with an arbitrary objective number of clusters, and two other techniques which do not require such a value, namely DBSCAN and OPTICS, will be used to serve as references, for they are the closest to the proposed methodology. Notice that, among the five methods mentioned, the true closest one is OPTICS, for it truly requires no specific parameter to operate (the former three need the number of clusters, and DBSCAN needs the maximal distance to delineate its neighborhood).
These five algorithms are applied to the dataset presented in Section 5.1 (cf. Figure 3); the feature space where the algorithms operate is the twelve-dimensional space spanned by the twelve sensors. Their respective results are represented in Figure 9, where every method is performed twice (corresponding to the two graphs of a same row), each time with a different value set to the main parameter. The main parameters are the objective number of clusters for the three first algorithms (K for the K-Means, and the grid size for the SOMs and BSOMs), the maximal distance ε delineating the neighborhood of a data for DBSCAN, and the minimal number of samples a cluster must comprise for OPTICS. Notice that this very last parameter can apply to every method, but it was chosen specifically for OPTICS since it has no true other parameter to vary to affect the results.
For the K-Means, SOMs and BSOMs, the number of clusters was set to nine to emulate the case of the identification of the main regions of the feature space (the steady and shutdown states, the turn-on and power-off procedures, one or two anomalies, plus a couple of additional transient states) and to twenty-five to emulate the case when this parameter is set with a very too large value. For DBSCAN, the coefficient ε was set to 0.75 and 0.15 for arbitrary reasons, only drawn by experimentation; several values were tested, and two were selected: one leading to a favorable case (0.75, whose results are really meaningful with respect to the study carried out in Section 5.3) and the other to an unfavorable case (0.25, with a large number of clusters, not always meaningful), to emphasize the impact of a wrong value set to ε . Similar reasoning led to set the minimal number of samples a cluster must comprise for OPTICS, with a more favorable case (250) and an unfavorable one (50), chosen in a purely experimental fashion.
Among the ten partitionings obtained, only that generated by DBSCAN with ε set to 0.75 gave very good results, close to that of the SPRADA method, as discussed in Section 5.3. Indeed, this partition of the space (left graph of the fourth row of Figure 9) contains five clusters, but three are very small; the two largest ones, in blue and orange, are almost the same than that of Figure 6, representing the steady and the shutdown states, respectively.
Otherwise, none of the three methods requiring the number of clusters led to a good partitioning, either when using nine or twenty-five clusters. Nonetheless, the 3 × 3 BSOM obtained better results than the five others: the steady state was correctly identified, in blue, but the shutdown state was split in two, in orange and green; the five additional clusters comprise very few data, which are just outliers here. Notice that this was the original reason why the BSOMs were crafted, i.e., to provide a fast and efficient clustering algorithm able to delineate the main regions of a feature space by averaging several experimental partitionings. On the contrary, neither the SOMs nor the K-Means were able to correctly identified the main regions, no matter the objective number of clusters used.
Finally, the last method, OPTICS, yet deemed the most accurate one among the five approaches, obtained very poor results when the minimal number of samples is fairly low (50), and obtained mitigated results when this value increased (250). Notice that, in practice, the minimal number of samples is one for the K-Means, zero for the SOMs and BSOMs and two for DBSCAN. These results seem to indicate that OPTICS is positively not suited to the identification of the major regions of an industrial feature space. Notice also that this clustering algorithm is the heaviest one among the five others, plus the SPRADA method, which reinforces that OPTICS may not be the best solution for real industrial processes.
Additionally, the silhouette coefficients of the five graphs of the left column of Figure 9 have been computed and are gathered within Table 7. Only these five ones were considered for the sake of readability; moreover, the partitionings on the right column were obtained with intentionally erroneous values set to the main parameters, thus their silhouettes will very likely be worse. The results of the table emphasize the qualitative conclusions drawn above: the best partitioning is unquestionably DBSCAN, with a mean of 0.75, which is very close to that obtained when using the SPRADA method (and is even slightly better, cf. Table 6), then the BSOM with a good mean nonetheless, then the SOM and the K-Means, and eventually OPTICS, which even obtained a negative mean, indicating that the data of clt1 are closer to other clusters than their own neighbors of clt1.
This comparative study aimed to compare the results of the SPRADA method to that obtained when using more standard unsupervised clustering algorithms. With respect to these observations, only two other methods obtained good results: the 3 × 3 BSOM, and, especially, DBSCAN with ε set to 0.75. The latter one obtained very close results to that of the SPRADA method, indicating that the proposed methodology is resilient and accurate in the identification of the regions of a space, especially that of a real industrial use-case.
The main strength of the SPRADA method is its capability of being very generic, if not to say universal, since it can use any clustering algorithm and quantification metric, thus it can be adapted to the situations where a clustering technique performs better than another. For instance, region-growing clustering is very well-suited to the segmentation of images, thus the SPRADA method may use such a clustering algorithm to propose a split-and-merge variant of it. Additionally, it requires no true parameter to be set manually, except a dynamic parameter ε , which is automatically set by the algorithm, even though the user can still have control over it to possibly refine the results. All these specificities make the proposed SPRADA method very promising for future work on behavioral identification and anomaly detection and diagnosis in real industrial systems.

6. Discussion of the Findings

The SPRADA method has been crafted to help data scientists choose an unsupervised clustering algorithm to automatically identify the main regions of a space, while overcoming the necessity of fine-tuning parameters manually, since it requires almost no parameters to set, except the threshold ε , which might be seen as the grain of the partitioning.
Actually, several parameters are still necessary: the clustering algorithm, the quantification metric, the inner parameters of both, etc., but it was shown that, following the proposed recursive decomposition and the clusters merging stage, results were similar, whatever were the parameters used, and especially the objective number of clusters. Indeed, with respect to Figure 8, the main regions identified by the six variants tested were very close, even though the K-Means used only two clusters ( K = 2 ), whilst the SOMs and BOMs used nine ( 3 × 3 maps). As such, the initial objective of the proposed methodology has been correctly achieved, i.e., overcoming the necessity of setting that specific parameter before processing without having to use non-efficient and computationally heavy methods, such as the Elbow method or its silhouettes-based variant.
Furthermore, even though the SPRADA method is very generic and should work with any dataset, it was applied to real industrial data, provided by Solvay, one of the industrial partners involved in the European project HyperCOG, and proved able to correctly and accurately identify the major regions of the feature space spanned by the basis consisting of the different sensors of the processes under study, in particular the steady and shutdown states (thus, the two extreme states). This knowledge is greatly valuable in the context of predictive maintenance and anomaly detection, forecasting and diagnosis. Indeed, automatically extracting the true regular states (the behaviors) of a system is essential to build a sort of ground truth to which the unusual and/or abnormal groups of data, possibly corresponding to true anomalies, can be compared to.
Additionally, it proved to be more accurate than almost all the other more standard unsupervised clustering methods tested, and has the great advantage of not requiring external parameters to be manually set. However, the good results of DBSCAN mitigate that observation, since the SPRADA method cannot pretend to be the by far best solution, even though it is actually the case with every other method tested. Nonetheless, DBSCAN has its own limitations: it consumes a lot of memory, is fairly computationally heavy, performs well only when the data density is consistent, and still requires a meta-parameter, as of the neighborhood ε . On the contrary, the SPRADA method is a methodology which can use any clustering algorithm, making it very general, and more or less universal, is very fast and efficient, for it uses very simple tools (the SOMs, which are very fast and efficient, the ECDs, which are simple logical comparisons, and the MHDs, which are simple summations), and does not truly require such a parameter. Additionally, the good results obtained with the SPRADA method on real industrial data seem to indicate that this methodology is, among other uses, very well-suited to the identification of the main regions of space, even with real industrial data.
Although the SPRADA method outputted very promising results in the scope of the automatic behavioral identification of unknown industrial systems, this study presents three main limitations. First, it has been assessed on only one dataset: it should be worth testing it on other ones, and not limited to the chemical industry. Second, the impact of the parameters has been greatly diminished, and the method proved to be resilient to most of them, resulting in very close partitionings, but there is one which still needs to be set, as of the threshold ε ; it is dynamic, and the exact value does not really matter, for the methodology is resilient to it, but it still has to be set nonetheless, and will condition the grain of the partitioning. Third, the SPRADA method is well-suited to the identification of large and compact groups of data (the regular states), but not really to the isolation of smaller groups, such as, for instance, the transient states or the anomalies (even though this can nevertheless be performed by adjusting the coefficient ε ).
Nonetheless, to compensate that last limitation, a possibility could be to exploit the capability of the SPRADA method for the identification of the behaviors of industrial systems and the capability of a more classical clustering technique (such as the SOMs) to isolate more groups of data. Indeed, the SPRADA method performs well in the identification of the major regions of a space, but poorly in the identification of the smallest ones; yet, the anomalies rarely form large groups, since they generally do not last long and, as such, are often regrouped within scattered groups. On the contrary, unsupervised clustering, with a fairly high number of clusters set as objective, such as a 4 × 4 SOM (16 clusters), is more prone to the identification of smaller groups, which could correspond to anomalies. However, a SOM performing with the wrong number of clusters has the trend of cutting the major regions in several pieces, thus it is difficult to automatically distinguish the anomalies from the behaviors. Therefore, the idea would be to compare these groups to that issued by the SPRADA method, which could allow an automatic classification of the clusters issued by the SOM as either part of a behavior or a true anomaly. The authors are currently investigating such a solution, but the results are not published yet.
Moreover, the next work that the authors are currently focusing on is the modeling and prediction of the dynamic processed one may find in real industrial systems, a tool which may be of great help in an Industry 4.0-oriented context. To that end, the idea may consist in using a highly reliable and accurate unsupervised clustering method to automatically identify and isolate the real behaviors of the system (i.e., its historically-encountered regular ways to behave), then locally modeling each of these behaviors separately, and finally connecting all of them in some fashion into a kind of neural network. The overall modeling (and/or forecasting) tool would therefore be a complex combination of the different behaviors of the system, automatically identified and modeled using unsupervised, data-driven and machine learning-based approaches, as shown in Figure 10.

7. Conclusions

In the context of the identification of the regular states (behaviors) of real, dynamic and nonlinear industrial processes, this paper introduced the Self-Parameterized Recursively Assessed Decomposition Algorithm (SPRADA), a recursive methodology inspired by the decomposition principle used by the Tree-Like Divide To Simplify modeling algorithm. A space is first clustered using any clustering algorithm (e.g., the K-Means or the SOMs), and the “quality” of every cluster is evaluated by using any quantification metric (e.g., the Average Standard Deviation or the Hyper-Density). If the quality of a cluster is judged as low, thus that group is split by applying any clustering algorithm. This process is repeated for every cluster obtained, including those generated by the split of another one, until all of them obtain a good quality score (or that they contain only one data).
Recursive decomposition allows splitting a space into many pieces, but often generates numerous groups, possibly comprising very few data. To compensate that, and to allow a true identification of the regions of a given space, a second stage of clusters merging is applied, based on the Empirical Cumulative Distributions (ECD), following the approach proposed by the ECD test. To do so, the ECD of every cluster is computed, then they are all compared to each other, and the clusters are eventually linked by region growing clustering, using the Modified Hausdorff Distance between the ECD curves to evaluate the (dis-)similarity between them.
The proposed methodology was assessed on real industrial data provided by an industrial partner, Solvay, in the scope of the European project HyperCOG, and proved able to identify the two main regions of the feature space (steady and rest states) in a fully automatic fashion, without requiring manual parameters to set, except the grain ε of the merging. This study showed the strength, and universality, of the SPRADA method to detect and isolate the main regions of a space, by using a split-and-merge approach, without the need of setting many parameters than none knows about. As a consequence, the unsupervised, data-driven and self-parameterized properties of that clustering methodology make it appealing for more elaborated pieces of work, such as, for instance, the modeling and prediction of complex, dynamic industrial systems.
Additionally, the study showed the resistance of the method to parameters: neither the clustering method, their inner parameters (especially the number of groups to build), nor the quantification metric greatly affected the final results. Only the coefficient ε (a dynamic threshold) affects the merging process, and conditions not so much the accuracy of the results as the number of regions identified, which can be seen as somehow the grain of the partitioning. For instance, with a high value set for ε , the results showed that only the two major states were identified, whilst sub-behaviors (e.g., transient states) were also isolated when using a lower value for ε .

Author Contributions

Conceptualization, D.M., K.M., V.A. and A.C; methodology, D.M., K.M., V.A and A.C; software, D.M.; validation, D.M.; formal analysis, D.M. and K.M.; investigation, D.M. and K.M.; resources, K.M.; data curation, D.M.; writing—original draft preparation, D.M. and K.M; writing—review and editing, D.M., K.M., and V.A. and A.C.; visualization, D.M.; supervision, K.M.; project administration, K.M. and V.A.; funding acquisition, K.M. All authors have read and agreed to the published version of the manuscript.

Funding

This paper took place within the Hyperconnected architecture for highly cognitive production plants (HyperCOG) project. The project has received funding from the European Union’s Horizon 2020 research and innovation program under grant agreement No 869886. Website: www.hypercog.eu (accessed on 26 July 2023).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Private industrial data; not publicly available.

Acknowledgments

The authors thank Marc Legros, Senior Engineer at Solvay, for his help and useful discussions concerning the data collection and understanding. Authors also wish to express their gratitude to EU 2020 program for supporting the presented works.

Conflicts of Interest

The industrial partners involved in the HyperCOG project have scrutiny rights due to the confidential nature of their data; they helped with data understanding and did not intervene concerning the content of this article (strict advisors). The conclusions presented in this article were drawn without their help. The authors declare no personal conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial Intelligence
(I)IoT(Industrial) Internet of Things
CPSCyber-Physical System
KMK-Means
SOMSelf-Organizing Map
BSOMBi-level Self-Organizing Map
T-DTSTree-like Divide to Simplify
AvStdAverage Standard deviation
HyDensityHyper-Density
ECDEmpirical Cumulative Distribution
CLTCluster

References

  1. Jafari, N.; Azarian, M.; Yu, H. Moving from Industry 4.0 to Industry 5.0: What Are the Implications for Smart Logistics? Logistics 2022, 6, 26. [Google Scholar] [CrossRef]
  2. Rahim, M.A.; Rahman, M.A.; Rahman, M.; Asyhari, A.T.; Bhuiyan, M.Z.A.; Ramasamy, D. Evolution of IoT-enabled connectivity and applications in automotive industry: A review. Veh. Commun. 2021, 27, 100285. [Google Scholar] [CrossRef]
  3. Khosravani, M.R.; Nasiri, S.; Reinicke, T. Intelligent knowledge-based system to improve injection molding process. J. Ind. Inf. Integr. 2022, 25, 100275. [Google Scholar] [CrossRef]
  4. Arents, J.; Greitans, M. Smart Industrial Robot Control Trends, Challenges and Opportunities within Manufacturing. Appl. Sci. 2022, 12, 937. [Google Scholar] [CrossRef]
  5. Ahmed, I.; Jeon, G.; Piccialli, F. From Artificial Intelligence to Explainable Artificial Intelligence in Industry 4.0: A Survey on What, How, and Where. IEEE Trans. Ind. Inform. 2022, 18, 5031–5042. [Google Scholar] [CrossRef]
  6. Paiola, M.; Schiavone, F.; Khvatova, T.; Grandinetti, R. Prior knowledge, industry 4.0 and digital servitization. An inductive framework. Technol. Forecast. Soc. Chang. 2021, 171, 120963. [Google Scholar] [CrossRef]
  7. Mercan, S.; Cain, L.; Akkaya, K.; Cebe, M.; Uluagac, S.; Alonso, M.; Cobanoglu, C. Improving the service industry with hyper-connectivity: IoT in hospitality. Int. J. Contemp. Hosp. Manag. 2021, 33, 243–262. [Google Scholar] [CrossRef]
  8. Huertos, F.J.; Masenlle, M.; Chicote, B.; Ayuso, M. Hyperconnected Architecture for High Cognitive Production Plants. Procedia CIRP 2021, 104, 1692–1697. [Google Scholar] [CrossRef]
  9. Bai, Y.; Lv, Y.; Zhang, J. Smart mobile robot fleet management based on hierarchical multi-agent deep Q network towards intelligent manufacturing. Eng. Appl. Artif. Intell. 2023, 124, 106534. [Google Scholar] [CrossRef]
  10. Nonoyama, K.; Liu, Z.; Fujiwara, T.; Alam, M.M.; Nishi, T. Energy-Efficient Robot Configuration and Motion Planning Using Genetic Algorithm and Particle Swarm Optimization. Energies 2022, 15, 2074. [Google Scholar] [CrossRef]
  11. Hentout, A.; Maoudj, A.; Aouache, M. A review of the literature on fuzzy-logic approaches for collision-free path planning of manipulator robots. Artif. Intell. Rev. 2023, 56, 3369–3444. [Google Scholar] [CrossRef]
  12. Liu, S.; Bao, J.; Zheng, P. A review of digital twin-driven machining: From digitization to intellectualization. J. Manuf. Syst. 2023, 67, 361–378. [Google Scholar] [CrossRef]
  13. Calvo-Bascones, P.; Voisin, A.; Do, P.; Sanz-Bobi, M.A. A collaborative network of digital twins for anomaly detection applications of complex systems. Snitch Digital Twin concept. Comput. Ind. 2023, 144, 103767. [Google Scholar] [CrossRef]
  14. van Dinter, R.; Tekinerdogan, B.; Catal, C. Reference architecture for digital twin-based predictive maintenance systems. Comput. Ind. Eng. 2023, 177, 109099. [Google Scholar] [CrossRef]
  15. Stavropoulos, P.; Papacharalampopoulos, A.; Sabatakakis, K.; Mourtzis, D. Metamodelling of Manufacturing Processes and Automation Workflows towards Designing and Operating Digital Twins. Appl. Sci. 2023, 13, 1945. [Google Scholar] [CrossRef]
  16. Gao, Y.; Li, X.; Wang, X.V.; Wang, L.; Gao, L. A Review on Recent Advances in Vision-based Defect Recognition towards Industrial Intelligence. J. Manuf. Syst. 2022, 62, 753–766. [Google Scholar] [CrossRef]
  17. Liu, M.; Li, X.; Li, J.; Liu, Y.; Zhou, B.; Bao, J. A knowledge graph-based data representation approach for IIoT-enabled cognitive manufacturing. Adv. Eng. Inform. 2022, 51, 101515. [Google Scholar] [CrossRef]
  18. Ullah, A.S. What is knowledge in Industry 4.0? Eng. Rep. 2020, 2, e12217. [Google Scholar] [CrossRef]
  19. Zhou, G.; Zhang, C.; Li, Z.; Ding, K.; Wang, C. Knowledge-driven digital twin manufacturing cell towards intelligent manufacturing. Int. J. Prod. Res. 2020, 58, 1034–1051. [Google Scholar] [CrossRef]
  20. Krämer, K.; van Elst, L.; Arteaga, A. Traveling Salesman Problem: A Case Study of a Scheduling Problem in a Steelmaking Plant. In Proceedings of the IN4PL 2022: 3rd International Conference on Innovative Intelligent Industrial Production and Logistics, Valletta, Malta, 24–26 October 2022; pp. 291–296. [Google Scholar] [CrossRef]
  21. Hoffmann, M.; Kotzur, L.; Stolten, D.; Robinius, M. A Review on Time Series Aggregation Methods for Energy System Models. Energies 2020, 13, 641. [Google Scholar] [CrossRef] [Green Version]
  22. Molinié, D.; Madani, K.; Amarger, V. Behavioral Modeling of Real Dynamic Processes in an Industry 4.0-Oriented Context. In Proceedings of the DATA 2023: 12th International Conference on Data Science, Technology and Applications, Lisbon, Portugal, 11–13 July 2023; pp. 510–517. [Google Scholar] [CrossRef]
  23. Zonta, T.; da Costa, C.A.; da Rosa Righi, R.; de Lima, M.J.; da Trindade, E.S.; Li, G.P. Predictive maintenance in the Industry 4.0: A systematic literature review. Comput. Ind. Eng. 2020, 150, 106889. [Google Scholar] [CrossRef]
  24. Abdallah, M.; Joung, B.G.; Lee, W.J.; Mousoulis, C.; Raghunathan, N.; Shakouri, A.; Sutherland, J.W.; Bagchi, S. Anomaly Detection and Inter-Sensor Transfer Learning on Smart Manufacturing Datasets. Sensors 2023, 23, 486. [Google Scholar] [CrossRef] [PubMed]
  25. Huang, J.; You, J.X.; Liu, H.C.; Song, M.S. Failure mode and effect analysis improvement: A systematic literature review and future research agenda. Reliab. Eng. Syst. Saf. 2020, 199, 106885. [Google Scholar] [CrossRef]
  26. Dalzochio, J.; Kunst, R.; Pignaton, E.; Binotto, A.; Sanyal, S.; Favilla, J.; Barbosa, J. Machine learning and reasoning for predictive maintenance in Industry 4.0: Current status and challenges. Comput. Ind. 2020, 123, 103298. [Google Scholar] [CrossRef]
  27. El Baz, J.; Cherrafi, A.; Benabdellah, A.C.; Zekhnini, K.; Beka Be Nguema, J.N.; Derrouiche, R. Environmental Supply Chain Risk Management for Industry 4.0: A Data Mining Framework and Research Agenda. Systems 2023, 11, 46. [Google Scholar] [CrossRef]
  28. Govender, P.; Sivakumar, V. Application of k-means and hierarchical clustering techniques for analysis of air pollution: A review (1980–2019). Atmospheric Pollut. Res. 2020, 11, 40–56. [Google Scholar] [CrossRef]
  29. Benabdellah, A.C.; Benghabrit, A.; Bouhaddou, I. A survey of clustering algorithms for an industrial context. Procedia Comput. Sci. 2019, 148, 291–302. [Google Scholar] [CrossRef]
  30. Molinié, D.; Madani, K.; Amarger, V. Clustering at the Disposal of Industry 4.0: Automatic Extraction of Plant Behaviors. Sensors 2022, 22, 2939. [Google Scholar] [CrossRef]
  31. Molinié, D.; Madani, K.; Amarger, C. Identifying the Behaviors of an Industrial Plant: Application to Industry 4.0. In Proceedings of the IDAACS’2021: 11th International Conference on Intelligent Data Acquisition and Advanced Computing Systems: Technology and Applications, Cracow, Poland, 22–25 September 2021; Volume 2, pp. 802–807. [Google Scholar] [CrossRef]
  32. Madhulatha, T.S. An Overview on Clustering Methods. arXiv 2012, arXiv:1205.1117. [Google Scholar] [CrossRef]
  33. Schubert, E. Stop Using the Elbow Criterion for K-Means and How to Choose the Number of Clusters Instead. ACM SIGKDD Explor. Newsl. 2023, 25, 36–42. [Google Scholar] [CrossRef]
  34. Ruiz-Moreno, S.; Gallego, A.J.; Sanchez, A.J.; Camacho, E.F. Deep Learning-Based Fault Detection and Isolation in Solar Plants for Highly Dynamic Days. In Proceedings of the ICCAD’22: 6th International Conference on Control, Automation and Diagnosis, Lisbon, Portugal, 13–15 July 2022. [Google Scholar] [CrossRef]
  35. Ruiz-Moreno, S.; Gallego, A.J.; Sanchez, A.J.; Camacho, E.F. A cascade neural network methodology for fault detection and diagnosis in solar thermal plants. Renew. Energy 2023, 211, 76–86. [Google Scholar] [CrossRef]
  36. Wang, H.; Liu, X.; Ma, L.; Zhang, Y. Anomaly detection for hydropower turbine unit based on variational modal decomposition and deep autoencoder. Energy Rep. 2021, 7, 938–946, 2nd International Conference on Energy Engineering and Power Systems (EEPS). [Google Scholar] [CrossRef]
  37. Latham, S.; Giannetti, C. Root Cause Classification of Temperature-related Failure Modes in a Hot Strip Mill. In Proceedings of the IN4PL 2022: 3rd International Conference on Innovative Intelligent Industrial Production and Logistics, Valletta, Malta, 24–26 October 2022; pp. 36–45. [Google Scholar] [CrossRef]
  38. Chen, H.; Liu, H.; Chu, X.; Liu, Q.; Xue, D. Anomaly detection and critical SCADA parameters identification for wind turbines based on LSTM-AE neural network. Renew. Energy 2021, 172, 829–840. [Google Scholar] [CrossRef]
  39. Mokhtari, S.; Abbaspour, A.; Yen, K.K.; Sargolzaei, A. A Machine Learning Approach for Anomaly Detection in Industrial Control Systems Based on Measurement Data. Electronics 2021, 10, 407. [Google Scholar] [CrossRef]
  40. Shi, C.; Wei, B.; Wei, S.; Wang, W.; Liu, H.; Liu, J. A quantitative discriminant method of elbow point for the optimal number of clusters in clustering algorithm. EURASIP J. Wirel. Commun. Netw. 2021, 2021, 31. [Google Scholar] [CrossRef]
  41. Marutho, D.; Hendra Handaka, S.; Wijaya, E.; Muljono. The Determination of Cluster Number at k-Mean Using Elbow Method and Purity Evaluation on Headline News. In Proceedings of the iSemantic 2018: International Seminar on Application for Technology of Information and Communication, Semarang, Indonesia, 21–22 September 2018; pp. 533–538. [Google Scholar] [CrossRef]
  42. Ketchen, D.J.; Shook, C.L. The Application of Cluster Analysis in Strategic Management Research: An Analysis and Critique. Strateg. Manag. J. 1996, 17, 441–458. [Google Scholar] [CrossRef]
  43. Amorim, R.; Hennig, C. Recovering the number of clusters in data sets with noise features using feature rescaling factors. Inf. Sci. 2015, 324, 126–145. [Google Scholar] [CrossRef] [Green Version]
  44. Van der Laan, M.; Pollard, K.; Bryan, J. A new partitioning around medoids algorithm. J. Stat. Comput. Simul. 2003, 73, 575–584. [Google Scholar] [CrossRef] [Green Version]
  45. Lenssen, L.; Schubert, E. Clustering by Direct Optimization of the Medoid Silhouette. In Similarity Search and Applications; Skopal, T., Falchi, F., Lokoč, J., Sapino, M.L., Bartolini, I., Patella, M., Eds.; Springer International Publishing: Cham, Switzerland, 2022; pp. 190–204. [Google Scholar] [CrossRef]
  46. Bagirov, A.M.; Aliguliyev, R.M.; Sultanova, N. Finding compact and well-separated clusters: Clustering using silhouette coefficients. Pattern Recognit. 2023, 135, 109144. [Google Scholar] [CrossRef]
  47. Tibshirani, R.; Walther, G.; Hastie, T. Estimating the number of clusters in a data set via the gap statistic. J. R. Stat. Soc. Ser. B Stat. Methodol. 2001, 63, 411–423. [Google Scholar] [CrossRef]
  48. Mohajer, M.; Englmeier, K.H.; Schmid, V. A comparison of Gap statistic definitions with and without logarithm function. arXiv 2011, arXiv:1103.4767. [Google Scholar] [CrossRef]
  49. Cui, H.; Chang, Y.; Zhang, H.; Mi, X.; Kang, B. Determine the number of unknown targets in the open world from the perspective of bidirectional analysis using Gap statistic and Isolation forest. Inf. Sci. 2023, 623, 832–856. [Google Scholar] [CrossRef]
  50. Molinié, D.; Madani, K. ECD Test: An Empirical Way based on the Cumulative Distributions to Evaluate the Number of Clusters for Unsupervised Clustering. In Proceedings of the IN4PL 2022: 3rd International Conference on Innovative Intelligent Industrial Production and Logistics, Valletta, Malta, 24–26 October 2022; pp. 279–290. [Google Scholar] [CrossRef]
  51. Murtagh, F.; Contreras, P. Algorithms for hierarchical clustering: An overview. Wire Data Min. Knowl. Discov. 2012, 2, 86–97. [Google Scholar] [CrossRef]
  52. Murtagh, F.; Contreras, P. Algorithms for hierarchical clustering: An overview, II. Wire Data Min. Knowl. Discov. 2017, 7, e1219. [Google Scholar] [CrossRef] [Green Version]
  53. Subramaniyan, M.; Skoogh, A.; Muhammad, A.S.; Bokrantz, J.; Johansson, B.; Roser, C. A generic hierarchical clustering approach for detecting bottlenecks in manufacturing. J. Manuf. Syst. 2020, 55, 143–158. [Google Scholar] [CrossRef]
  54. Karna, A.; Gibert, K. Automatic identification of the number of clusters in hierarchical clustering. Neural Comput. Appl. 2022, 34, 119–134. [Google Scholar] [CrossRef]
  55. Buyssens, P.; Gardin, I.; Ruan, S.; Elmoataz, A. Eikonal-based region growing for efficient clustering. Image Vis. Comput. 2014, 32, 1045–1054. [Google Scholar] [CrossRef] [Green Version]
  56. Wu, Q.; Castleman, K.R. Chapter Seven—Image Segmentation. In Microscope Image Processing, 2nd ed.; Merchant, F.A., Castleman, K.R., Eds.; Academic Press: Cambridge, MA, USA, 2023; pp. 119–152. [Google Scholar] [CrossRef]
  57. Poux, F.; Mattes, C.; Selman, Z.; Kobbelt, L. Automatic region-growing system for the segmentation of large point clouds. Autom. Constr. 2022, 138, 104250. [Google Scholar] [CrossRef]
  58. Ouyang, T.; Shen, X. Online structural clustering based on DBSCAN extension with granular descriptors. Inf. Sci. 2022, 607, 688–704. [Google Scholar] [CrossRef]
  59. Chen, H.; Liang, M.; Liu, W.; Wang, W.; Liu, P.X. An approach to boundary detection for 3D point clouds based on DBSCAN clustering. Pattern Recognit. 2022, 124, 108431. [Google Scholar] [CrossRef]
  60. Steclik, T.; Cupek, R.; Drewniak, M. Automatic grouping of production data in Industry 4.0: The use case of internal logistics systems based on Automated Guided Vehicles. J. Comput. Sci. 2022, 62, 101693. [Google Scholar] [CrossRef]
  61. Liang, L.; Qian, L. Clustering Algorithm for Running Test Data Sequence of Computer Integrated System. In Proceedings of the ICOSEC 2022: 3rd International Conference on Smart Electronics and Communication, Tamil Nadu, India, 20–22 October 2022; pp. 886–889. [Google Scholar] [CrossRef]
  62. Hamouche, K.; Rasolofondraibe, L.; Chiementin, X.; Felkaoui, A. Localization of defects in rolling element bearings by dynamic classification based on meta-analysis of indicators: Supervised real-time OPTICS method. Arab. J. Sci. Eng. 2022, 47, 9151–9171. [Google Scholar] [CrossRef]
  63. Al Samara, M.; Bennis, I.; Abouaissa, A.; Lorenz, P. Complete outlier detection and classification framework for WSNs based on OPTICS. J. Netw. Comput. Appl. 2023, 211, 103563. [Google Scholar] [CrossRef]
  64. Lloyd, S.P. Least squares quantization in PCM. IEEE Trans. Inf. Theory 1982, 28, 129–136. [Google Scholar] [CrossRef] [Green Version]
  65. Kohonen, T. Self-organized formation of topologically correct feature maps. Biol. Cybern. 1982, 43, 59–69. [Google Scholar] [CrossRef]
  66. Gholami, V.; Khaleghi, M.R.; Pirasteh, S.; Booij, M.J. Comparison of Self-Organizing Map, Artificial Neural Network, and Co-Active Neuro-Fuzzy Inference System Methods in Simulating Groundwater Quality: Geospatial Artificial Intelligence. Water Resour. Manag. 2022, 36, 451–469. [Google Scholar] [CrossRef]
  67. Gao, L.; Zhang, W.; Liu, Q.; Lin, X.; Huang, Y.; Zhang, X. Machine learning based on the graph convolutional self-organizing map method increases the accuracy of pollution source identification: A case study of trace metal(loid)s in soils of Jiangmen City, south China. Ecotoxicol. Environ. Saf. 2023, 250, 114467. [Google Scholar] [CrossRef] [PubMed]
  68. Chon, T.S.; Jang, Y.H.; Jung, N.; Lee, K.; Kwak, G.S.; Kim, D.H.; Sim, K.S.; Lee, J.E.; Min, J.H.; Park, Y.S. Spatial patterning of benthic macroinvertebrate communities using Geo-self-organizing map (Geo-SOM): A case study in the Nakdong River, South Korea. Ecol. Inform. 2023, 76, 102148. [Google Scholar] [CrossRef]
  69. Hazaveh, S.H.; Bayandour, A.; Khalili, A.; Barkhordary, A.; Farzamnia, A.; Moung, E.G. Impulsive Noise Suppression Methods Based on Time Adaptive Self-Organizing Map. Energies 2023, 16, 2034. [Google Scholar] [CrossRef]
  70. Macshane, J.; Ahmadinia, A. AI Assisted Trail Map Generation based on Public GPS Data. In Proceedings of the SIEDS 2023: Systems and Information Engineering Design Symposium (SIEDS), Rome, Italy, 13–14 July 2023; pp. 37–42. [Google Scholar] [CrossRef]
  71. Molinié, D.; Madani, K. BSOM: A Two-Level Clustering Method Based on the Efficient Self-Organizing Maps. In Proceedings of the ICCAD’22: 6th International Conference on Control, Automation and Diagnosis, Lisbon, Portugal, 13–15 July 2022; pp. 1–6. [Google Scholar] [CrossRef]
  72. Rousseeuw, P. Silhouettes: A Graphical Aid to the Interpretation and Validation of Cluster Analysis. J. Comput. Appl. Math. 1987, 20, 53–65. [Google Scholar] [CrossRef] [Green Version]
  73. Rybnik, M. Contribution to the Modelling and the Exploitation of Hybrid Multiple Neural Networks Systems: Application to Intelligent Processing of Information. Ph.D. Thesis, University Paris-Est XII, Paris, France, 2004. [Google Scholar]
  74. Molinié, D.; Madani, K.; Chebira, A. HyDensity: A Hyper-Volume-Based Density Metric for Automatic Cluster Evaluation. In Innovative Intelligent Industrial Production and Logistics; Smirnov, A., Panetto, H., Madani, K., Eds.; Springer Nature: Cham, Switzerland, 2023; pp. 50–69. [Google Scholar] [CrossRef]
  75. Lawrence, A.E. The Volume of an n-Dimensional Hypersphere; University of Loughborough: Loughborough, UK, 2001. [Google Scholar]
  76. Madani, K.; Chebira, A.; Rybnik, M. Data Driven Multiple Neural Network Models Generator Based on a Tree-like Scheduler. In Computational Methods in Neural Modeling; Mira, J., Álvarez, J.R., Eds.; Springer: Berlin/Heidelberg, Germany, 2003; pp. 382–389. [Google Scholar] [CrossRef]
  77. Madani, K.; Chebira, A.; Rybnik, M.; Bouyoucef, E.k. Tree-Like Multiple Neural Network Models Generator with a Complexity Estimation Based Decomposer. In Proceedings of the IDAACS’2005: 3rd IEEE Intelligent Data Acquisition and Advanced Computing Systems: Technology and Applications, Sofia, Bulgaria, 5–7 September 2005; pp. 60–65. [Google Scholar] [CrossRef]
  78. Dubuisson, M.; Jain, A. A modified Hausdorff distance for object matching. In Proceedings of the ICPR94: 12th International Conference on Pattern Recognition, Jerusalem, Israel, 9–13 October 1994; IEEE Computer Society: Los Alamitos, CA, USA, 1994; Volume 2, pp. 566–568. [Google Scholar] [CrossRef]
  79. Molinié, D.; Madani, K. Characterizing N-Dimension Data Clusters: A Density-based Metric for Compactness and Homogeneity Evaluation. In Proceedings of the IN4PL 2021: 2nd International Conference on Innovative Intelligent Industrial Production and Logistics, Virtual Event, 25–27 October 2021; pp. 13–24. [Google Scholar] [CrossRef]
Figure 1. Recursive decomposition of a database into local subdatabases (SB). Term “DU” stands for “Decomposition Unit“, i.e., representing any clustering algorithm.
Figure 1. Recursive decomposition of a database into local subdatabases (SB). Term “DU” stands for “Decomposition Unit“, i.e., representing any clustering algorithm.
Make 05 00051 g001
Figure 2. Flowchart of the proposed SPRADA method.
Figure 2. Flowchart of the proposed SPRADA method.
Make 05 00051 g002
Figure 3. The real industrial dataset (Solvay) used to carry out the study.
Figure 3. The real industrial dataset (Solvay) used to carry out the study.
Make 05 00051 g003
Figure 4. The recursive principle of the SPRADA method, represented on Sensor 12. At the top row, the original dataset and the initial partitioning (SOM) on the left and right graphs, respectively. For every row below, on the left is the current state of the split, with the cluster to be split in red and the others in blue, and on the right, the corresponding cluster split by a SOM (one color per cluster).
Figure 4. The recursive principle of the SPRADA method, represented on Sensor 12. At the top row, the original dataset and the initial partitioning (SOM) on the left and right graphs, respectively. For every row below, on the left is the current state of the split, with the cluster to be split in red and the others in blue, and on the right, the corresponding cluster split by a SOM (one color per cluster).
Make 05 00051 g004
Figure 5. The fifty clusters generated by the recursive decomposition of the SPRADA method when applied to the twelve sensors of the dataset. Every pair color-symbol represents a unique cluster.
Figure 5. The fifty clusters generated by the recursive decomposition of the SPRADA method when applied to the twelve sensors of the dataset. Every pair color-symbol represents a unique cluster.
Make 05 00051 g005
Figure 6. The four clusters obtained after the final merging stage of the SPRADA method. Every cluster is represented by a unique color.
Figure 6. The four clusters obtained after the final merging stage of the SPRADA method. Every cluster is represented by a unique color.
Make 05 00051 g006
Figure 7. The clusters generated by the recursive decomposition with every of the six variants studied. Clustering techniques on the rows (from top to bottom: K-Means, SOMs and BSOMs), and quantification metrics on the columns (from left to right: Hyper-Density and AvStd).
Figure 7. The clusters generated by the recursive decomposition with every of the six variants studied. Clustering techniques on the rows (from top to bottom: K-Means, SOMs and BSOMs), and quantification metrics on the columns (from left to right: Hyper-Density and AvStd).
Make 05 00051 g007
Figure 8. The clusters obtained after applying the ECD-based merging process of the SPRADA method to the clusters of Figure 7 (consistent positioning of the graphs across the two figures).
Figure 8. The clusters obtained after applying the ECD-based merging process of the SPRADA method to the clusters of Figure 7 (consistent positioning of the graphs across the two figures).
Make 05 00051 g008
Figure 9. The clusters of Sensor 12 obtained with the five more standard clustering algorithms. Every row corresponds to a same method, performed with two different values set to the main parameter.
Figure 9. The clusters of Sensor 12 obtained with the five more standard clustering algorithms. Every row corresponds to a same method, performed with two different values set to the main parameter.
Make 05 00051 g009
Figure 10. The future behaviors-based multi-model to be investigated, built, crafted and assessed.
Figure 10. The future behaviors-based multi-model to be investigated, built, crafted and assessed.
Make 05 00051 g010
Table 1. Recapitulation of the methods to estimate the number of clusters for unsupervised clustering.
Table 1. Recapitulation of the methods to estimate the number of clusters for unsupervised clustering.
NameAdvantagesLimitations
Elbow method [40]Reference method; statistical analysis of the information provided by the clusters.Curve’s elbow not always easy to find; requires to execute the clustering algorithm several times.
Silhouettes Elbow [43]Silhouette coefficients accurate and representative; provides a good estimate of the number of clusters.Idem Elbow method + silhouette coefficients very long and expensive to compute.
Gap statistic [47]Very good estimates when correctly parameterized.Computationally heavy; requires to execute the clustering algorithm several times.
ECD test [50]Simple and efficient; good results with industrial processes; requires to execute clustering only once.The threshold conditions the grain of the merging, thus the number of regions identified.
Table 2. Recapitulation of the algorithms which do not require the number of clusters as parameter.
Table 2. Recapitulation of the algorithms which do not require the number of clusters as parameter.
NameAcronymAdvantagesLimitations
Hierarchical Clustering [51,52]HCSimple and intuitive; provides a good representation of the data distribution.Not a true space partitioning; requires the threshold of where to cut the links.
Region Growing [55]RGSimple and efficient; performs well in image segmentation.Sensitive to initialization; needs the maximal distance delineating the neighborhood of data.
Density-Based Spatial Clustering of Applications with Noise [58]DBSCANIntuitive and generic; provides good results in real situations.Idem RG + slightly higher complexity and fixed density threshold.
Ordering Points To Identify The Clustering Structure [61]OPTICSImprovement of DBSCAN with a dynamic density threshold.Closer to a hierarchically ordered set of clusters than a true partitioning.
Table 3. Recapitulation of the unsupervised clustering methods used in this study.
Table 3. Recapitulation of the unsupervised clustering methods used in this study.
Clustering MethodsAcronymAdvantagesDrawbacks
K-Means [64]KMSimple and intuitive; often used as a reference algorithm.Not efficient and unable to separate nonlinear databases.
Self-Organizing Maps [65]SOMsFast, efficient and generally deal properly with nonlinearities.Tends to split the main behaviors of industrial processes in pieces.
Bi-Level Self-Organizing Maps [71]BSOMsGood results when identifying the behaviors of industrial processes.Anomalies often missed (drowned out in the main behaviors).
Table 4. Recapitulation of the cluster’s integrity estimators.
Table 4. Recapitulation of the cluster’s integrity estimators.
NameAcronymAdvantagesLimitations
Silhouette coefficients [72]SCsRepresentative of how compact and distant from each other the clusters are.High complexity in O ( N 2 ) , not suited to repetitive executions on large datasets.
Average Standard Deviation [73]AvStdVery simple and fairly representative of the degree of scattering of the data.More a compactness indicator; not very representative of the integrity of clusters.
Hyper-Density [74]HyDensityEfficient and intuitive; representative of cluster’s density (in the sense of Physics).Sensitive to outliers; the hyper-volume’s borders might be complex to estimate.
Table 5. Average numbers of clusters issued by the recursive decomposition using the K-Means, the SOMs, and the BSOMs, and using the Hyper-Density and the AvStd as integrity estimators, computed over 1067 iterations. Symbols μ and σ stand for the mean and the standard deviation, respectively.
Table 5. Average numbers of clusters issued by the recursive decomposition using the K-Means, the SOMs, and the BSOMs, and using the Hyper-Density and the AvStd as integrity estimators, computed over 1067 iterations. Symbols μ and σ stand for the mean and the standard deviation, respectively.
AlgorithmsParametersHyper-DensityAvStd
K-Means K = 2 μ = 14.1 / σ = 9.2 μ = 13.7 / σ = 6.9
SOMs 3 × 3 -map μ = 33.0 / σ = 7.8 μ = 33.1 / σ = 5.5
BSOMs103 ×   3 -maps μ = 26.8 / σ = 3.7 μ = 26.0 / σ = 3.2
Table 6. The silhouette coefficients of the clusters of the six variants under study, after the merging process (cf. Figure 8). The K-Means is referred to as “KM”, and the Hyper-Density as “HyDensity”; the symbol “X” means no value (i.e., when a cluster contains only one data).
Table 6. The silhouette coefficients of the clusters of the six variants under study, after the merging process (cf. Figure 8). The K-Means is referred to as “KM”, and the Hyper-Density as “HyDensity”; the symbol “X” means no value (i.e., when a cluster contains only one data).
Variant/
Cluster
clt1clt2clt3clt4clt5clt6clt7clt8clt9Mean
KM-HyDensity0.610.850.760.010.96X 0.75
KM-AvStd0.880.510.560.460.560.070.790.290.800.73
SOM-HyDensity0.870.590.550.250.23 0.75
SOM-AvStd0.850.470.470.920.96X 0.69
BSOM-HyDensity0.860.510.170.700.36 0.72
BSOM-AvStd0.470.510.520.430.100.600.540.96X0.47
Table 7. The silhouette coefficients obtained with the five standard unsupervised clustering algorithms considered, in the most favorable case for the main parameters (left column of Figure 9).
Table 7. The silhouette coefficients obtained with the five standard unsupervised clustering algorithms considered, in the most favorable case for the main parameters (left column of Figure 9).
Variant/
Cluster
clt1clt2clt3clt4clt5clt6clt7clt8clt9Mean
K-Means0.600.180.430.32−0.030.330.120.190.720.33
SOM0.340.340.630.280.330.770.150.160.110.40
BSOM0.800.540.400.020.110.130.450.18 0.68
DBSCAN0.870.590.590.60−0.04 0.75
OPTICS−0.540.900.970.87 −0.35
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Molinié, D.; Madani, K.; Amarger, V.; Chebira, A. Identifying the Regions of a Space with the Self-Parameterized Recursively Assessed Decomposition Algorithm (SPRADA). Mach. Learn. Knowl. Extr. 2023, 5, 979-1009. https://doi.org/10.3390/make5030051

AMA Style

Molinié D, Madani K, Amarger V, Chebira A. Identifying the Regions of a Space with the Self-Parameterized Recursively Assessed Decomposition Algorithm (SPRADA). Machine Learning and Knowledge Extraction. 2023; 5(3):979-1009. https://doi.org/10.3390/make5030051

Chicago/Turabian Style

Molinié, Dylan, Kurosh Madani, Véronique Amarger, and Abdennasser Chebira. 2023. "Identifying the Regions of a Space with the Self-Parameterized Recursively Assessed Decomposition Algorithm (SPRADA)" Machine Learning and Knowledge Extraction 5, no. 3: 979-1009. https://doi.org/10.3390/make5030051

Article Metrics

Back to TopTop