Next Article in Journal
Metapopulation Modeling of Socioeconomic Vulnerability of Sahelian Populations to Climate Variability: Case of Tougou, Village in Northern Burkina Faso
Next Article in Special Issue
Estimating the Individual Treatment Effect with Different Treatment Group Sizes
Previous Article in Journal
Asymmetric Graph Contrastive Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Exploratory Matching Model Search Algorithm (EMMSA) for Causal Analysis: Application to the Cardboard Industry

by
Richard Aviles-Lopez
1,*,
Juan de Dios Luna del Castillo
2 and
Miguel Ángel Montero-Alonso
2,*
1
Department of Computer Science and A.I., University of Granada, 18071 Granada, Spain
2
Department of Statistics and Operational Research, University of Granada, 18071 Granada, Spain
*
Authors to whom correspondence should be addressed.
Mathematics 2023, 11(21), 4506; https://doi.org/10.3390/math11214506
Submission received: 23 September 2023 / Revised: 16 October 2023 / Accepted: 28 October 2023 / Published: 31 October 2023
(This article belongs to the Special Issue Computational Methods and Machine Learning for Causal Inference)

Abstract

:
This paper aims to present a methodology for the application of matching methods in industry to measure causal effect size. Matching methods allow us to obtain treatment and control samples with their covariates as similar as possible. The matching techniques used are nearest, optimal, full, coarsened exact matching (CEM), and genetic. These methods have been widely used in medical, psychological, and economic sciences. The proposed methodology provides two algorithms to execute these methods and to conduct an exhaustive search for the best models. It uses three conditions to ensure, as far as possible, the balance of all covariates, the maximum number of units in the treatment and control groups, and the most significant causal effect sizes. These techniques are applied in the carton board industry, where the causal variable is downtime, and the outcome variable is waste generated. A dataset from the carton board industry is used, and the results are contrasted with an expert in this process. Meta-analysis techniques are used to integrate the results of different comparative studies, which could help to determine and prioritize where to reduce waste. Two machines were found to generate more waste in terms of standardized measures whose values are 0.52 and 0.53, representing 48.60 and 36.79 linear meters (LM) on average for each production order with a total downtime of more than 3000 s. In general, for all machines, the maximum average wastage for each production order is 24.98 LM and its confidence interval is [13.40;36.23] LM. The main contribution of this work is the use of causal methodology to estimate the effect of downtime on waste in an industry. Particularly relevant is the contribution of an algorithm that aims to obtain the best matching model for this application. Its advantages and disadvantages are evaluated, and future areas of research are outlined. We believe that this methodology can be applied to other industries and fields of knowledge.

1. Introduction

The cardboard industry has been growing steadily, attributable to increased corrugated box packaging needs for export and e-commerce from its versatility to protect, store, and transport [1]. The corrugated cardboard is made up of two sheets of paper (internal and external) and, in the middle, the corrugated paper that provides the energy-absorbing capacity [2]. The production volume is expected to reach 700 million metric tons by 2040, with a market size bordering USD 220 billion by 2028 [3]. Consequently, more waste will be generated which must be controlled because of its environmental, energy, and economic burdens [4].
One of the sources of waste in this industry is the printing subsystems, composed of several printing machines forming production lines (printing lines), where each includes several stages or modules [1]. Each machine may have different characteristics depending on the number of colors to be printed, the volume to be produced, and the post-printing stages it includes.
The production process of an order starts with a setup stoppage, and multiple stoppages may occur due to non-programmed or programmed operations. These stoppages are registered in real time by alarm monitoring software. The waste can arise from intrinsic reasons of the product design that can be calculated for each sheet, but we are interested in the waste sheets due to stoppages in the printing line. Thus, some of these stoppages can generate more waste due to effects on unscheduled maintenance, staffing, supply of materials, resource planning, subsystem failures, or fault propagations, among others [5].
Thus, determining the waste volume produced by each stoppage is impractical, complex, and difficult to quantify. The strategy followed has been to address all sources of waste using, for example, total productive maintenance [6], but which machine generates the most waste due to downtime? In which time interval does the most waste occur? Where to start? This generates more uncertainty because it is very complex to determine which stoppages are more important than others in almost half a million stoppage records.
However, if the waste is significant, the machine will likely stop, so somehow there should be a relationship between downtime and waste. For this reason, from a causal perspective, a study is required to determine the magnitude of the effect produced by the downtime of all these causes on waste.
Causal research in the manufacturing industry has been concentrated on the alarm root cause analysis, presenting multiple data-driven methods, such as those oriented to time-series data (Granger causality or entropy transfer), as well as methods oriented towards cross-sectional data, such as machine learning and probabilistic graphical models [7]. However, the former cannot be applied because the actual available data for downtimes only records the date but not the time when it occurs. Machine learning makes excellent predictions, but it does not properly determine the causes or adequately measure the causal effect size [8,9].
Therefore, the objective of this paper is to present a methodology to measure the causal effect size of the total downtime (seconds) of production orders on waste (LM) using production observational data. In addition, we will study its strengths and weaknesses. We believe that this methodology can be extended to other industries.
Addressing this problem is challenging due to the multiple sources of noise that occur in an industry: data distributions are non-normal, relationships are non-linear, and many variables exhibit multicollinearity [10]. Another problem is to find a parametric model that adequately measures the effect size. In addition, one of the challenges is to preprocess the data in such a way as to eliminate as much bias as possible [11].
We will use the matching methods to balance observational data and choose the best matching models that properly measure the effect size [12]. However, traditionally, it has been recommended to use several matching models to choose the best models, this being a manual search process. To our knowledge, no specific methodology exists for this purpose, nor have matching methods been used in the industry. The main contributions of this research are:
  • The presentation of the exploratory matching model search algorithm (EMMSA) for causal analysis.
  • The exploratory matching algorithm (EMA) for building matching models for each machine and treatment.
  • The homologous model selection algorithm (HMSA) to choose the best matching models and select the one that detects the maximum effect.
  • The application of meta-analysis to integrate the findings from the different machines and treatments.
To accomplish this, we will use a real dataset from a cardboard factory that is not fully automated and facing organizational difficulties and limitations, which increases the dataset’s variability. For these data, we pose a single research question: What is the effect size of total downtime on the waste generated on a printing line?
This article is organized into five sections. Section 2 summarizes the main causal methods used in the industry, the potential outcome framework, multivariate matching techniques, and meta-analysis models. Section 3 presents a description of the data, and how they were obtained. Section 4 presents the proposed industry application methodology and describes the exploratory matching algorithms and the homologous model selection algorithm. Section 5 presents and discusses the results. Section 6 concludes the study.

2. Literature Review

2.1. Causality Techniques Applied to the Manufacturing Process

Almizahad et al. synthesized the different methods of root cause analysis for both time series and cross-sectional data [7]. For time series, two approaches are available depending on whether they use multivariate Granger causality (MGC) [12] or transfer entropy (TE) methods.
Granger causality (G-causality) is a concept of direct causality used to establish whether the past values of a time series help to improve the prediction of another time series [13]. This method makes several assumptions, such as that the cause occurs before its effect, the time series values are continuous and Gaussian, the causal effect is linear, the time series is stationary [14], there is a discrete sampling frequency, and unmeasured confounding variables are not present [15]. These assumptions are unlikely to be met in practice [15]. MGC extends bivariate Granger causality to multiple variables and its advantage is that it has a low computational cost, is easier to implement, and works well with little data [16]. Principal component analysis is traditionally used as a feature selection method [7]. The main applications of MGC in root cause analysis are indicated in Table A1.
TE measures the amount of information transferred between two time series [17]. The TE methods were developed by establishing probability density functions (PDF) for each variable using kernel estimators because they offer robustness and accuracy [18]. A casual measure is established comparing the transfer entropy between two variables in one and another direction, and the significance level (threshold) of this measure is defined using Monte Carlo simulations of surrogated data [18]. Surrogated data can be generated by randomly shuffled data, or by the interactive amplitude-adjusted Fourier transform [19]. The main applications of MGC in root cause analysis are indicated in Table A2.
TE methods have the advantage of detecting linear or nonlinear causal relationships but have a high computational cost due to the application of their significance test based on Monte Carlo simulations. In addition, they require a large amount of seasonal data. They also require a method to select model parameters due to the use of PDF methods based on kernel estimators. Other disadvantages of TE and MGC are the inability to detect spurious relationships and to determine the existence of hidden (unmeasured) variables.
In cross-sectional data, Bayesian networks have been widely used. A Bayesian network (BN) is a directed acyclic graph (DAG) in which the nodes represent the random variables of the system under study, and the edges indicate the incidence of one node on another [20]. However, they have the disadvantage that they do not allow cycles, which are very present in industrial control processes. They are also not as accurate because the learning algorithms find equivalent Markov classes with multiple linking possibilities, which makes them difficult to interpret [21].
All these methods are data-driven, only detect directionality, and do not measure the causal effect size. To achieve this, they require the elimination of random noise that could affect the measurement of the effect size. In addition, they must control for bias introduced, and none of these approaches have addressed this yet.
As far as we know, no relevant records of matching methods were obtained in the fields of production and logistics. Therefore, it can be deduced that matching techniques have not been used to measure causal effect size in the industrial sector.

2.2. Potential Outcome Framework

To make inferences about the causal effect of a treatment, the potential outcome model was formalized by Donald Rubin [22]. A causal effect is the value comparison (difference or ratio) of the potential outcome (Y) when it is exposed to two alternative causal states or alternative treatments [22,23]. Alternative treatments can be binary (1: treatment and 0: control) or multi-treatment, and they should be defined into mutually exclusive alternative states [24]. In randomized trials, the way “each individual” is chosen to receive the treatment condition is known as the allocation mechanism [25], but in observational studies, it is impossible to control this aspect.
The individual causal effect is defined by the contrast between alternative states (treatment and control) for the same individual in the population of interest. The aggregate causal effects (average treatment effect—ATE) estimate the mean of the individual effects [26] which are shown in Equations (1) and (2). These causal effects can be defined over any subset of the population. E[.] signifies the expectation operator of probability theory.
ATE = E[Y1] − E[Y0] = E[Y1 − Y0]
A T E = 1 N i = 1 N E Y i 1 Y i 0 |   X i     = 1 N i = 1 N ( μ 1 Y i μ 0 Y i ) ,
where E[Y1] is the expected value of potential outcome for the treatment group; E[Y0] is the expected value of the potential outcome for the control group; E[Y1 − Y0] is the expected value of the difference of the two expectations equivalent to the mean treatment effect; E Y i 1 Y i 0 |   X i   is the expected value of the difference in the individual potential outcome, given their covariables Xi; μ 1 Y i is the mean individual potential outcome of the treatment group; μ 0 Y i is the mean individual potential outcome of the control group; and N is the size of the control or treatment groups.
In observational studies, the potential outcome of an alternative treatment that is not known is called the counterfactual. Thus, each observed potential outcome variable only reveals half of the information contained because it is impossible to observe both values for the same individual, either for ethical or practical reasons [23], and so one can never calculate the individual causal effect. This challenge is known as the fundamental problem of causal inference [26].
Because of this impossibility, the average treatment effect on the treated (ATT) is used more often (Equation (3)), where n is the size of the groups, m is the number of treatments, and T j is the weighting of each treatment.
A T T = 1   j = 1 m T j i = 1 n T j E Y i 1 Y i 0 |   X i   = 1 j = 1 m T j i = 1 n T j   μ 1 Y i μ 0 Y i
When the causal effects are constant in each i, the ATT is equal to the ATE. However, another group-level causal effect is the causal hazard rate (or causal risk ratio), which is often used in epidemiology and health sciences [27].
C a u s a l   H a z a r d   R a t e = E Y 1 E Y 0 = P Y 1 = 1 P   Y 0 = 1

2.3. Multivariable Matching Methods Overview

Estimating the causal effect in observational studies is a critical task as the researcher has no control over the treatment allocation mechanism. Therefore, its covariates may differ between treatment and control individuals (unbalanced) [28]. The balance refers to the similarity of the marginal covariate distribution in both groups and thus avoids the model dependence or imprecise inference [29].
The Model dependence is the variation in causal effect estimation at point xi among plausible alternative models that fit the data reasonably well [30]. However, when there is no model dependency, the functional form does not matter because the result will be approximately the same [31].
Thus, observational studies require a design phase to resemble randomized experiments, using background information (covariates), create more balanced treatment and control samples, and support more robust and credible inferences [25,29,32].
Multivariate matching is a set of nonparametric methods applied before parametric analysis to eliminate the difference in the covariates (Xi) affecting the causal variable (Ti) without accessing the outcome variable (Yi) [31]. It improves the balance with a minimum of loose observations in the process, allows for controlling the influence of confounding factors, and obtains more precise estimates of the causal effect with reduced bias and variance [28].
In order not to bias the analysis, the outcomes should not be available, and thus, by comparing several matched designs, the search for a satisfactory design becomes transparent [33].
The fundamental rule to avoid selection bias is that our preprocessed data set should include a subset of records from the observed sample whose treatment and control groups are unrelated and which have the same baseline characteristics [31], such that:
p ~ X i     T = 1 ) = p ~ X i     T = 0 )
Matching techniques only require that their distributions (treatment and control) be as close as possible, although the unobserved covariate bias is always possible [29]. In this way, Ti is independent of Xi, and the dependence on the functional form in the parametric analysis will be eliminated [31].
A first matching attempt was to use one-to-one exact matching, a treated unit with a control unit that contains both identical values of Xi, discarding all units that do not match. Exact matching uses one or more control units that match each treated unit, which improves efficiency [28].
According to [30], they suggest removing observations from the control group that are outside the “convex hull” of the treatment group. That is, units in the control group that are outside the range of the treatment variable are discarded. For this reason, to avoid extrapolation bias, it is necessary to check the common support and verify whether the control units are in the convex hull of the treatment group [31].
Thus, if the number of observations of the matched sample N is not reduced so much, matching will frequently reduce the bias and the variance of the estimates of the parametric analysis [31]. If NC >> NT, then more control units can be matched with each treatment unit, which improves the efficiency of estimation, and the bias may be reduced [34]. If NT >> NC, then each unit control may be matched with several treated units, which is known as matching with replacement [35].
The propensity score is a measure that summarizes all Xi variables, and it is defined as the conditional probability of unit i receiving treatment T, given their covariates Xi [36]. The propensity score is formulated as e(Xi) in the following equation:
e   X i = Pr ( T = 1   |   X i )
It is commonly calculated using a logit model (or random forests). Some matching techniques are based on a proximity metric, either distance or similarity. It may also use a threshold to ensure that the difference between the values of the metrics of the matched treatment and control units is less than or equal to it [37]. This threshold is known as a caliper.
Nearest-neighbor matching calculates the closeness metric of each observation based on the values of its covariates and then it matches the closest control and treatment units to each other [38]. Two versions are available for matching without substitution: greedy matching or optimal matching. The first matches a treated unit with the control unit that has the most similar propensity score without minimizing the overall distance, and the second includes minimizing the absolute distance between the matched units [39].
However, propensity scores attempt to balance the covariates but two units with the same value score may be very different. So, another approach is to additionally use the Mahalanobis distance within the calibrators of the propensity score, thus adding a layer of protection, but will not normally generate optimal pair matching [40].
Optimal matching solves an optimal assignment problem to find the closest treatment and control units that meet certain requirements or constraints on the balance of covariates [41]. Optimal matching is often used with the RELAX IV algorithm to solve the minimum cost flow problem in a network [42].
Full matching creates matching subsets containing one treated unit with one or more control units, or one control unit with one or more treated units [43]. Optimal full matching minimizes a weighted average of the estimated distance measure between pairs matched within each subclass, and it is available in the matchit package [41,44].
Genetic matching implements a more general distance measure based on the Mahalanobis distance, which incorporates a matrix of weights to be optimized over several generations. Optimization is carried out by minimizing a loss function over the matched sample at each iteration, identifying the corresponding weights until it converges asymptotically to the optimal solution [45].
Multiple measures of balance provide a more complete image of balance and the biases [46]. One such measure is the standardized mean difference (SMD) between treatment and control groups before and after matching. The SMD of each variable can be visualized with an SMD chart. If the absolute value of SMD is greater than 0.1, it indicates evidence of imbalance between the treatment and control groups. The SMD is not dependent on the unit of measure, and it is expressed for continuous covariates with the following equation [47]:
S M D = X 1 ¯ X 0 ¯ s 1 2 + s 0 2 2
For binary covariables, it is calculated using the proportions ( p i ^ ) of their values, both in the treated group and the control group [48].
S M D = p 1 ^ p 0 ^ [ p 1 1 p 1 ^ + p 0 ^ ( 1 p 0 ^ ) ] ^ 2
The variance ratio (VR) compares the variances of the two matched groups. If this indicator is equal to 1, it means that the balance is good, and less than 2 is generally acceptable [31]. The hypothesis test and its corresponding p-value should be used with caution. The significance test can detect spurious statistically significant differences [28].
The prognostic score is defined as the predicted probability of an outcome under control conditions [49]. It has been demonstrated to be an excellent indicator of the degree of bias reduction, and it is more efficient than the covariate mean differences and significance tests [46], even when the prognostic scoring model is misspecified. For this, a regression model is fitted to predict the outcome using the control group covariates. Thus, using the same model, the outcome of both control and treatment group units is predicted. The absolute SMD (ASMD) of both groups is then compared [50].
Even if optimally matched records were found, it is possible that they do not coincide in their unobserved variables, and therefore, there may be a bias in the treatment assignment due to these factors [51,52]. Sensitivity analysis in observational studies is a mathematical calculation of the amount of unobserved covariate bias needed to change the conclusions of the study [33]. This mathematical calculation varies an underlying assumption to a statistical procedure to determine what magnitude of deviation from that assumption would be necessary to alter the conclusion [33]. The Rosenbaum sensitivity test for Hodges—Lehmann point estimate calculates the odds of differential assignment to treatment due to unobserved factors for a range of gamma values. It is an unconfounded estimate that reflects “the uncertainty of the hidden bias due to failure to control for unobserved covariates” [52]. Better observational study design decisions result in better insensitivity to large unmeasured biases [51].
Cohen’s [53] effect size measure (d) is defined as the mean difference between the treatment and control groups divided by the population standard deviation. There are other more precise measures with various corrections of the standard deviation depending on the size of the groups and the (non)similarity of variances. However, the square root of the mean of the variances of both groups is frequently used as the population standard deviation. Thus, the effect size is similar to a Z-score and its formula is the same as the SMD. The Z-score is a standardized measure that allows us to make comparisons between different populations. In addition, Cohen defined three values for the value of d for the corresponding small, medium, and large effect sizes (0.2, 0.5, and 0.8, respectively) [53].
The proximity metrics are described in Table A3. The R libraries used in this paper are shown in Table A4.
In practice, many researchers have recommended building different matching models and choosing the one with the best fit of their covariates [33,54,55]. However, we have reviewed the literature (as of 30 September 2023), and, to our knowledge, only individual approaches exist for each specific problem, but we have not found generic methods to identify these models, which represents a knowledge gap.
In the absence of a general methodology to determine the best matching models, we developed in this paper a novel methodology to find them. For this purpose, we have run the most popular matching methods, providing balancing criteria and priorities to facilitate the identification and selection of the best models. The proposed methodology is presented in Section 4.3.

2.4. Statistical Models for Meta-Analysis

Meta-analysis is a statistical technique for integrating the findings from different comparative studies [56].
Two statistical models are most commonly used: the fixed-effect model and the random-effect model. The first assumes that a common effect size exists for each study and that their differences are due to sampling errors [57]. The second assumes that a distribution of true effects exists with different effect sizes for each study, and it focuses on calculating the parameters of this distribution [57]. The difference among studies can be due to background factors as well as the design or execution of the study. The random effects model nomenclature of DerSimonianLaird and their formulas are shown in Table A5 and Table A6, respectively [56].

3. Data

The production process we are concerned with has been summarized in the Introduction because it is the stochastic process that generates the available data.
The cardboard industry has two main production areas: corrugating and printing. The printing subsystem is composed of several printing lines. Each printing line is made up of several sequential modules or stages, for example, feeder, printer, die cutter, slotter, folder, gluer, stacker, or binder. These printing lines are also referred to as machines. The characteristics of each machine will depend mainly on the maximum number of colors it can print, the production volume, and the post-printing stages that compose it.
The industry works in three rotating shifts, 24 h a day, 7 days a week. The work teams are roughly similar in productivity and efficiency. During the production process of an order, there may be one or more stoppages, starting with the setup stoppage, and then, there may be zero or n stoppages due to scheduled or unscheduled operations. At the end of the production of an order, the wasted units are counted.
Data on production orders and downtime are stored in the organization’s ERP (enterprise resource planning) and alarm monitoring software, respectively.
The process begins when a production order is created in the sales department at the client’s request. Then, during the week before delivery, the scheduler assigns a date, shift, and machine on which the order will be produced, as well as the raw material and personnel requirements. When the assigned day arrives, the production order enters the queue of the machine, according to the allocated position.
Before the machine starts processing the order, the raw material and personnel are provisioned, and the machine is guaranteed to be operational. The production order starts with the setup stoppage where the cliché of each color is loaded, color shade is adjusted, and print accuracy adjustments are made. Each subsequent stoppage may be related to the modules of the printing line.
We want to determine the effect of the total downtime (seconds) of a production order on its waste (linear meters). Preliminary data that could help answer the causal question were selected by eliminating incomplete orders (due to cancellations), inconsistent or flawed records, and production lines with insufficient records (less than 1000 records).
New variables are calculated from the existing variables. Additionally, aggregate variables are generated by production orders. All available variables that could be associated with the outcome are described in Table 1. The dataset contains 59,676 records of the production orders collected over three years (2019 to 2021).
A more complex data set containing the individual stoppages of the different stages will not be used for the effect size calculations because it increases the number of variables and increases the complexity of the matching. However, we will address this data set in another study, for which we will need the collaboration of an expert in the process.

4. Methodology

To address how to use matching methods to measure the effect of a causal variable on the outcome variable, we propose the following steps outlined in Figure 1, which are described in the subsequent sections.

4.1. Causal Questions and Causal Variable

To conduct a causal analysis of observational data, a causal question must first be established to guide the whole process [27]. For this, it is necessary to distinguish between causal and non-causal questions. Causal questions deal with the mechanisms of data generation or predictions of how a system will behave after an intervention [58].
Therefore, for the selection of the causal variable, the phenomenon is first analyzed in search of variables that allow an intervention, to produce the expected value in the outcome variable [23,59].
Then, to formulate the causal question, both the causal and outcome variables must be explicit in the question with the corresponding units of measurement. For example, “What is the effect size of the causal variable (units) on the outcome variable (units) in the system?”

4.2. Preparation of Dataset for the Matching Algorithms

4.2.1. Exploring Feature Importance

It is advisable to create a regression model (e.g., light gradient boosting [60]) to determine the importance of the predictor variables on the outcome variable. In addition, we could examine the order of importance of variables to check that the chosen causal variable is included. If not, it is advisable to review the causal question again by choosing an important variable on which an intervention is possible [27].
In the industrial environment, production order variables are strongly associated with the output variable (e.g., the higher the number of order units is, the higher the number of waste units is). They are not causal variables, but associative, because they are set by the client, and, therefore, we cannot modify them; i.e., they do not allow interventions.
The causal variable is not always the one with the highest level of association with the outcome variable. Some causal variables may have a low level of association. Likewise, causal variables exist that are key and trigger the increase (or decrease) in the outcome variable, but their level of association could be close to zero, so they might not be considered causal variables. Therefore, the causal variable may or may not be strongly associated with the outcome variable.
Thus, causal analysis is an adjusted analysis, because the covariates, whether causal or associative, are controlled to rule out their effects on the outcome variable so that the size of the causal effect can be measured directly with the minimum possible bias and variance.
Therefore, in observational studies, the causal variable is not always required to be a necessary and sufficient condition to yield the effect on the outcome variable. In most cases, it is not sufficient, or it may be sufficient but not necessary, because other variables may be causing the effect.

4.2.2. Treatment Interval Design

A first approach consists of measuring the effect of a causal variable that groups together many causes affecting an outcome variable. To analyze it, the causal variable is segmented into intervals, depending on the amount of data, to measure the effect produced by each of them. From this result, we could determine the intervals with the greatest impact on the outcome variable to prioritize the search for its causes in the production process. This is because problems are often multi-causal, which is common in the manufacturing environment.
Each treatment (interval) is numbered from 0 to N, in ascending order of the causal variable. A control treatment is considered when the value of the causal variable is zero, and the treatment number assigned will be 0. Each treatment (from 1 to 10) will be contrasted with the control treatment (0). Then, the effect of treatment n will be the mean of the difference in potential outcomes between the two groups, treatment n and treatment 0.
But how many intervals should be defined? One alternative is to use a scatter diagram of the two variables (cause and outcome) and manually define the intervals of interest. Another alternative is to define the different intervals using percentiles and adjust their values, considering, as far as possible, at least 100 records per treatment to be able to detect the effect size with the minimum possible bias.

4.2.3. Selection of Systems (Machines)

Matching algorithms require sufficient records to function properly. For this reason, we scanned for the number of records for each production line and discarded those with insufficient records; for example, for 10 treatments, at least 1000 records are required for the printing line.

4.3. Exploratory Matching Algorithm

To assemble the samples as randomized experiments using the covariates prior to the causal variable, several popular matching algorithms are used, such as nearest, optimal, full, genetic, and coarsened exact matching (Section 2.3).
Algorithm 1 shows the exploratory matching algorithm (EMA). It begins by defining the formulas for matching and random forest (RF) models, as well as defining the lists of treatments, the matching methods, and the distance measures used. Then, it proceeds to set up four nested loops for the different machines, treatments, methods, and distances to be used in the matching methods. This algorithm is very brief as it does not contain the R-specific programming details but illustrates how to proceed. Before the end of each iteration, different tests are performed to collect the balancing information for subsequent analysis.
All model and test summaries are saved in a text file. Another algorithm in Python extracts the interest variables for its posterior analysis which are saved in a CSV file. These variables are machine; treatment; matching algorithm; distance; binning; control units; treatment units; matched control units; matched treatment units; variance of treatment group; and variance of control group. From the predictions of the RF model, the explained variance and the size effect (the mean difference for each group) are obtained, as are the following: the number of non-balanced variables whose standard mean differences are greater than 0.1; the number of non-balanced variables whose variance ratios are greater than 2; gamma and upper bound of the Rosenbaum sensitivity test when the upper bound is greater than 0.05; the F-statistic, p-value, and ratio of variances of the F-test; the t, p-value, confidence interval, and mean in group C and T of the Welch two-sample t-test; and the unconfounded estimate of the Rosenbaum sensitivity test for HodgesLehmann.
Algorithm 1. Exploratory matching algorithm (EMA).
Mathematics 11 04506 i001

4.4. Homologous Model Selection Algorithm (HMSA)

The CSV file of the matching results contains the relevant variables that we could use to select the matching models that best balance our data for each machinetreatment. Then, HMSA processes this file, calculates the Z-Score, and proceeds to find the register of matching results with the lowest number of unbalanced variables and with the highest number of matched units. To accomplish this, we define two conditions to validate models. These are responsible for assigning balancing priorities to each matching model. A third condition is set to search for models with higher priority in both condition 1 and condition 2, and then, at the machinetreatment level, the mean Z-score is calculated and those models whose Z-score is greater than 95% of the mean are selected. These models selected are considered homologous due to their similar characteristics. Next, the selection of the better matching method is chosen from homologous models for each machinetreatment. Algorithm 2 summarizes the HMSA algorithm with its conditions and priorities.
Algorithm 2. Homologous model selection algorithm (HMSA).
Mathematics 11 04506 i002

4.4.1. Condition 1—Unbalanced Variables

The first condition has three priorities. Priority 1 is assigned to models that have zero unbalanced variables in both SMD and VR. Otherwise, priority 2 is assigned if zero unbalanced variables exist in SMD and up to 20% unbalanced variables exist in VR, or zero unbalanced variables exist in VR and up to 20% unbalanced variables exist in SMD. If the two previous cases do not occur, priority 3 will be assigned if the models present up to 20% of unbalanced variables both in SMD and VR. If the above three cases do not occur, priority 9 will be assigned to models with an excessive number of unbalanced variables.

4.4.2. Condition 2—Matched Units

The second condition also has three priorities. Priority 1 is assigned to models whose matched units in the treatment and control groups are equal to the maximum number of treatment and control units, respectively, i.e., no treatment units have been lost, and the number of control-matched units is at least the same as the number of treatment units. Otherwise, priority 2 is assigned to models with a maximum number of matched units in the treatment group and with matched units in the control group greater than or equal to 90% of the control units, or vice versa. If the above two cases do not occur, priority 3 is assigned to models with a number of matched units greater than or equal to 90% of units in both the control and treatment groups. If the above three cases do not occur, priority 9 is assigned to indicate the loss of a significant number of treatment or control units.

4.4.3. Condition 3—Homologous and Selected Models

Matching models whose priorities of condition 1 and condition 2 are less than or equal to 3 are found, thus discarding all models that do not present a satisfactory balance. If no records remain after filtering for a machinetreatment, it is assumed that a satisfactory model does not exist. Then, the highest priority models of condition 1 (minimum value) are chosen. From these registers, the highest priority registers of condition 2 (minimum value) are chosen. Then, for each machinetreatment, the models whose Z-score exceeds 95% of the mean Z-score are considered homologous models because they have similar characteristics and represent the best models that capture the largest effects. From the homologous models, the model with the maximum Z-score is selected as the best model for the respective machinetreatment.
In addition, two values are obtained as effect size for each machinetreatment, and the first one will be obtained from the selected model (Z-score), and the second one will be the mean of the effect size of the homologous models with their respective confidence intervals.
The results of homologous models, including the selected model, are exported to Excel for visualization and analysis. The EMA is implemented using the R libraries listed in Table A4. The HMSA is implemented in Python. Both algorithms are available for consultation upon reasonable request by any reader. However, it will not be possible to establish comparisons of this methodology with other existing methodologies because such comparisons do not exist, as already indicated in the Literature Review section.

4.5. Validation of Results with the Expert

The results obtained from the quantitative models will be validated with the qualitative knowledge of the expert in this process. In addition, the results are contrasted with a data analysis of the outcome variable.

4.6. Meta-Analysis

Meta-analysis techniques are used to make comparisons of the effect of the causal variable on the outcome variable in a subsystem. For this purpose, the “meta” library in R was used to develop another algorithm that integrates these results and determines, in a subsystem, the contribution of each treatment to the outcome variable. Similarly, for each treatment, a meta-analysis is performed to determine the contribution of the causal effect in each subsystem to the outcome variable. Since these are standardized measures, it allows these comparisons to be made.

5. Results and Discussion

5.1. Causal Questions and Causal Variable

The causal variable chosen in the Introduction to formulate the causal question is the total downtime (MSUMALL) of a production order. This variable can be intervened in since it can be changed by using different strategies. The main causes of these stoppages can be due to inefficient processes, operator errors, lean inventory of materials, or inefficient machine maintenance. By addressing these causes, downtimes can be reduced, and at the same time, waste can also be reduced as it is likely that the machine will be stopped when it produces defective units. Intuitively, the downtime could be directly proportional to the error severity.
The outcome variable is WASTELM since we are concerned about its behavior and intend to control it through MSUMALL to reduce costs and increase profitability.
For this reason, these two variables are crucial in the analysis of productivity, and it becomes valuable to detect the time intervals in which waste is higher. Thus, with an appropriate intervention according to the error type, it will be possible to change the causal variable to the lowest possible intervals (treatments) where the waste is smaller.

5.2. Exploring the Feature Importance

A regression was performed using the LightGBM library. The most satisfactory model was found using a hyperparameter grid, and its values are shown in Figure 2, which illustrates that among the important features associated with the outcome variable (WASTELM) are (i) OVERPROCLM, which indicates the linear meters of overproduction, either by a small safety percentage to cover possible defective units not observed by the naked eye, or by replacement of non-conforming units; (ii) M2LAM, which is a product specification; (iii) SCHEDLM, which is determined in planning, and its value is slightly higher than the linear meters specified by the client in the production order; (iv) ORDEREDLM, which represents the linear meters of the production order; (v) PRODTIME, which depends on the downtimes of the current order; (vi) ASLEEPTIME, which depends on the order date and its delivery date; (vii) FINALLAM, which represents the linear meters of production (without waste) shipped to the client; (viii) MSUMALL, which depends on the downtime of all stoppages after MSETUP; (ix) MSETUP, which depends on the operator and the order complexity; and (x) QUEUETTIME, which depends on the downtime of previous orders. However, OVERPROCLM and FINALLAM do not directly cause waste but are its effects.
Therefore, the causes of waste could somehow be reflected in the total downtime, which will be confirmed by this study. However, it is previously necessary to segment MSUMALL to generate the treatment variables (TMSUMALL). Thus, with the proposed methodology, we will be able to identify for each machine the time interval where the greatest waste is generated in order to subsequently identify with the expert the possible causes that explain it.
Then, the covariates that we should control for matching algorithms are all those that occur prior to the treatment variable (TMSUMALL), such as MT2LAM, ORDEREDLM, SCHEDLM, QUEUETIME, ASLEEPTIME, TEST, and MSETUP. The covariate TEST is included because it is important to control the carton strength, as it is somehow an indicator of the thickness and would help to obtain more similarly matched units. MSETUP is considered because it occurs before the stoppages to be analyzed.
Equation (9) shows the matching formula that will be used to build the models controlling the TMSUMALL variable with the variables that precede it and which are not causes of WASTEML.
TMSUMALL ~ MT2LAM + ORDEREDLM + SCHEDLM + QUEUETIME + ASLEEPTIME + TEST + MSETUP
PRODCTIME, OVERPROCLM, PRODLM, and FINALLM are involved a posteriori and will be included in the random forest regression model to measure the effect size of each model. Equation (10) shows the formula for building random forest regression models in which participate all predictors variables. In the end, after selecting the best models, it will be checked whether the effect size is similar between the simple mean of the two groups and the mean of the difference of the predictions of the RF models in both groups.
WASTELM ~ PRICELAM + MT2LAM + ORDEREDLM + SCHEDLM + FINALLM + QUEUETIME+ ASLEEPTIME + PRODCTIME + MSETUP + MSUMALL + OVERPROCLM + TEST

5.3. System Selection and Treatment Interval Construction

The treatment variable was generated by splitting the MSUMALL variable, considering all machines with more than one thousand records (machines with an insufficient number of records were excluded). The treatments correspond to each decile, considering all the records for their calculation to obtain the same analysis interval for all the machines. Records whose MSUMALL is zero are considered the treatment group (0). Table A7 shows a summary of the number of records and the treatment interval (in seconds) for each machine. Figure A1 shows the MSUMALL density function for each machine. Figure A2 shows the WASTELM density function for each machine.

5.4. EMA and HMSA Algorithms

Algorithm 1 (EMA), which executes the different matching methods with the various distance metrics selected, was run and 7691 matching models were obtained. Then, Algorithm 2 (HMSA), which performs an exhaustive search for the best models to be considered as homologous, was run, and 2644 models (34%) fulfilling conditions 1 and 2 were obtained, this being the first filter.
The results of the application of Algorithm 2 (HMSA) are summarized in Table 2. It shows a summary of all models found that meet condition 1 (rows) and condition 2 (columns) showing their respective priorities.
In addition, for each priority of condition 1, there can be two rows to indicate the fulfillment of condition 3. If it is 1, it indicates the number of models that fulfill condition 3, which we will call homologous models, or 0 indicates the number of models that do not fulfill this condition, and therefore, they are consequently discarded. Condition 3 acts as the second filter.
For example, the HMSA algorithm found 839 models from the genetic method that met condition 1—priority 1 and condition 2—priority 1, of which 400 models were considered homologous because they also met condition 3, and 439 were discarded. In addition, it found 183 genetic models that fulfilled condition1—priority 2 and condition 2—priority 1, of which 4 models were considered homologous because they also met condition 3 and 179 were discarded.
The bottom part of Table 2 shows a summary of models filtered by HMSA. For example, it found 1093 genetic models that met conditions 1 and 2 (II), and they represent 41.3% (III) of the 2644 models that met both conditions (1 and 2) for any priorities. In addition, it found 404 homologous models (IV), and they represent 37% of the total genetic models (V), 56% of the models found (VI) from all homologous models (723), and 15.3% of the total models (VII) fulfilling conditions 1 and 2. Finally, it selects for each machinetreatment 39 genetic models that represent 56% of the 70 models selected from all methods.
Therefore, 723 homologous matching models fulfilling conditions 1, 2, and 3 were obtained (IV), which corresponds to 27% of the models (V) that meet conditions 1 and 2 (2644).
In other words, only 9% of all generated models (7691) fulfill the three expected conditions of having almost all variables balanced (condition 1), the maximum number of treatment and control units (condition 2), and an effect size greater than 95% of the mean (condition 3). The remaining percentage of models is discarded because the models have a priority value greater than 1 in conditions 1 and 2 or because they do not meet condition 3, which acts as a second filter.
The results of the application of Algorithm 2 (HMSA) are plotted by methods (Figure 3) and distances (Figure 4). Thus, Figure 3a and Table 2 (V and VI) show that most of the homologous matching models are selected by the genetic (56%), nearest (22%), and optimal (15%) methods. Figure 3a shows that the most selected distances in the homologous models are scaled Euclidean (33%), GAM (15%), GBM (15%), Mahalanobis (12%), and GLM (10%).
Methods and distances by machine: Figure 3b shows that the genetic, nearest, and optimal methods find homologous models on all machines. Fullmatch does not find homologous models in M3, and CEM only finds them in M1. More models are obtained in M4 because it has a larger number of records, which could facilitate matching.
Figure 4b shows that the genetic method obtains an almost similar number of models for each distance measure, whether it uses a caliper or not. Figure 4c shows that the scaled Euclidean and Mahalanobis distances have most of the homologous models across all machines.
Therefore, the genetic method is the one that generates more homologous models due to its iterative process to automatically evaluate the balance and optimize the search for a better solution. For this, it uses a generalized version of the Mahalanobis distance that incorporates a matrix of weights to be adjusted in each iteration to minimize the loss function. At each iteration, it produces a new generation and converges asymptotically [45]. However, the disadvantage of the genetic method is the mean execution time with any distance, as shown in Figure 4f, and the Euclidean, GBM, and Bart distances are the most time-consuming.
Methods and distances by treatment: Figure 3c shows that genetic, nearest, optimal, and fullmatch methods find homologous models on all treatments. Figure 4d shows that the scaled Euclidean, GAM, GLM, and Mahalanobis distances have the most homologous models across all treatments.
Therefore, the genetic method with Mahalanobis and scaled Euclidean distances ensures finding homologous models for all machines and all treatments, with or without caliper, as shown in Figure 3b,c and Figure 4b–d.
The run time was approximately 140 h on an Intel Core I7 Fourth Generation computer with 12 Gb RAM (1600 Mhz) and an SSD hard disk (560 Mb/s). The matching library used is matchit. For the genetic method, the maximum number of generations is 100, the maximum unchanging generational is 4, and the convergence tolerance is 0.001. More details about this algorithm can be found in [45].

5.5. Effect Size

Finally, HMSA chooses one model the among homologous models for each machinetreatment. The model chosen is the one with the highest Z-score to indicate the largest effect captured from the corresponding homologous models. Then, HMSA calculates the mean effect size of these homologous models and their confidence interval (CI).
The results of the selected models for each machinetreatment are shown in Table A8, and their effects are plotted in Figure 5a,b. Almost all selected models obtained priority 1 in conditions 1 and 2, indicating that all variables were balanced, with no loss of treatment units, and with the maximum number of control units (or at least equal to the number of treatment units). The exceptions appear in machine M1, treatments 8, 9, and 10, in which there was an unbalanced variable. Furthermore, it is observed in Table A8 that the mean of the effects calculated as difference of means (EffMD), and the values of the differences of the means of the predictions using RF models (EfffRF) are approximately similar, which confirms the low dependence of the models.
In addition, the fullmatch method detects the largest effects because it incorporates almost all of the control units, increasing the variability of mean differences and capturing larger Z-scores. However, it does not find balanced models in all treatments, which are covered by the models of the other methods.
Therefore, the genetic, fullmatch, optimal, and nearest methods guarantee to find models with the highest Z-score on each machine and treatment, as also is shown in Table 2 (VIII y IX). In addition, the scaled Euclidean, Mahalanobis, GAM, robust Mahalanobis, and Euclidean distances are used in 80% of the selected models, as shown in Figure 6.
Table 3 summarizes the Z-score greater than 0.30 and the mean effect of homologous models greater than 20 LM. Then, we can determine that, in standardized terms, the three higher Z-scores are in M4-T10, M5-T10, and M6-T10. In linear meters, wastes greater than 20 linear meters are generated in M1-T10, M4-T10, M5-T10, and M6-T10. Therefore, treatment 10 is the one that generates the most waste on machines M1, M4, M5, and M6.

5.6. Overall Effect Size

To find out which machine generates more waste in statistical terms, we integrate the different studies of each machine and treatment using the meta-analysis techniques described in Section 2.4 and Section 4.6. For this purpose, the meta-library has been used and, for the sake of brevity, we only present the M6 machine and treatment 10, which is where more waste is generated in standardized terms, as shown in Figure 7a,b.
The results of the meta-analysis are summarized in Table 4. It shows the random effect (RandEff) among studies, as well as their maximum effects found. Columns I2 and τ2 indicate the among-study heterogeneity, measured as the percentage of the total variance (I2) that is explained by the between-study variance (τ2). The “Max” column indicates, for each machine (treatment), the treatment (machine), method, and distance that produce the maximum mean effect measured in SMD or EffAvg LM, to which their respective confidence intervals are appended (95%-CI and 95%-CI LM).
τ2 is the variance of the random effects and has a very relevant value because it indicates the importance of each machine for the differences in waste. For example, 0.0271 is an important variance concerning the other variances. This indicates great variability exists between machines, and some machines have an inherent effect on waste. Furthermore, we did not eliminate or correct for this variability in the matching process. Therefore, there must be other variables in the M5 machine that have more influence on the waste.
In Table 4, the column “EffAvg LM” is the mean of the maximum effects in linear meters corresponding to the SMD value. The “All Mchs” row is weighted average of the effects of the treatments of each machine in which the maximum total wastage is reached. Thus, the maximum average waste for each production order that has stoppages after MSETUP is 24.98 LM with a confidence interval of [13.40;36.23].
In addition, we observe that the selected models of machines M2 and M3 present similarity since their I2 is 0%. The same occurs with treatments 2 and 4. In these cases, their mean random effects are approximately equal in normalized terms.
In contrast, it is observed that the selected models of machines M4, M5, and M6 present greater differences, which indicates the heterogeneity of the studies, and therefore, their means and CIs are different. The same is true for treatments 6 to 10. In addition, it is observed that the fullmatch method detects the largest effects in 71% of the machines and 90% of the treatments.

5.7. Advantages and Disadvantages of This Methodology

An advantage of this methodology is that it measures the causal effect of downtime on waste. In addition, this methodology could be used to study with the experts the causes that reduce productivity and overall equipment effectiveness (OEE). Therefore, it is important to address the downtimes to increase machine availability, avoid their deterioration and improve the performance, efficiency, and quality of the printing line. This way of studying the impact of downtime on waste is justified because it addresses two crucial issues at the same time: waste as a function of downtime, which will help reduce production costs and improve the profitability of the organization.
The process followed to obtain the models to which the required strong restrictions apply means that, in percentage terms, this number of models is small, although in absolute numbers it is still a very large number of resulting models.
For this reason, the three HMSA conditions are so restrictive that models could be obtained whose causal effect sizes, for the same machinetreatment, have slight variations.
Condition 3 (effect size greater than 95% of the mean) selects models that capture large effects, so it is possible that other more moderate effects are being neglected in this situation.
However, the choice to stay with models that meet all the constraints and are therefore strongly consolidated gives value to these models as it allows us to detect the variability caused by waste due to downtime in very extreme situations.
If, with these restrictions, homologous models are not found in some machinetreatment, it means that no reliable models were found to measure the corresponding effect size.
The main limitation of this study is that it does not measure the effect of the stoppages of each stage of the production line of each machine, which will be addressed in a future study, but it serves the purpose of an exploratory search for the main causes using the matching models.
Another disadvantage is that the effect size is calculated assuming that the standard deviations are different, and the group sizes are similar (Equation (7)). For this reason, in cases where the sample size and the standard deviation of each group are markedly different, it will be necessary to correct the standard deviation of the population. This may occur mainly with fullmatch, which considers more control units.
In any case, the differences detected when the standard deviation correction is performed would produce effects analogous or slightly underestimated to the uncorrected effect measurements. Equation (7) has more power to measure the effect because it has more type 1 errors. Therefore, the measured effects of the fullmatch models, using Equation (7), provide sufficient reliability to be considered.

6. Conclusions

In this work, the EMMSA methodology has been developed to measure, for each machine, the size of the causal effect of downtime on waste. As mentioned in the Introduction, being able to measure this effect could contribute to studying, quantitatively and causally, one of the main problems of the manufacturing industry, since it affects productivity and profitability.
This methodology allows a rigorous comparison between the different models generated by the matching methods used to perform the causal analysis.
For this purpose, the EMMSA is carried out with two developed algorithms: Algorithms 1 (EMA) and 2 (HMSA). Both have the advantage of performing an exploratory search for the best models to ensure, as far as possible, comparative groups with all covariates balanced, the maximum number of treatment and control units, and an effect size greater than 95% of the mean.
In the applied case, the machines and treatments where most waste is generated are shown in Table 3. Using meta-analysis methods, it was determined that M5 and M6 machines and treatment 10 generate the most waste in terms of standardized measurements. As such, the meta-analysis method helps us to prioritize, at the machine and treatment levels, where waste should be primarily addressed.
In future studies, the EMMSA methodology can be extended to more than ten covariates, include binary variables, and use variables with very different probability distributions. These extensions may require more criteria or priors to improve the model selection process.
The EMMSA methodology could be adapted for applications in other industries involving production processes in which waste is a fundamental part of the cost. It could also be extended to other fields of knowledge.

Author Contributions

Choice of article topic: R.A.-L. and J.d.D.L.d.C.; Development and review of model comparison methodology: R.A.-L. and J.d.D.L.d.C.; Data collection: R.A.-L.; Editing, cleaning, and adjustment of data: R.A.-L.; Development and review of algorithm application software: R.A.-L.; Analysis of results: R.A.-L., J.d.D.L.d.C. and M.Á.M.-A.; Drafting and review of article: R.A.-L., J.d.D.L.d.C. and M.Á.M.-A.; Final editing and review of article: R.A.-L. and J.d.D.L.d.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The datasets generated and analyzed during the current study are available from the corresponding author on reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Appendix A.1. Causality Methods in Time Series

Table A1. Multivariable Granger causality methods for time series.
Table A1. Multivariable Granger causality methods for time series.
MethodRelationsDataReference
Auto Regressive ModelLinearStationary [16]
Gaussian Process
Regression
Non-linearNon-stationary data[16]
Spectral Granger CausalityNon-linearOscillating variables: Fourier transformation[61]
Grouping Multivariate Granger CausalityNon-linearOscillating variables: Multivariate non-linear chirp mode decomposition[62]
Table A2. Granger causality methods for time series.
Table A2. Granger causality methods for time series.
MethodDescriptionReference
Direct Transfer
Entropy (DTE)
Detects spurious causalities and direct (indirect) paths between univariate time series. Requires stationary data.[19]
(Direct) Transfer Zero Entropy (T0E)Like DTE but does not assume well-defined PDF. Detects direct or indirect causalities for multivariate data. Does not require stationary data or a large sample.[63]
Trend Transfer
Entropy (TTE)
Causal analysis based on the time-series trends reducing computational burden.[64]
Symbolic dynamic based normalized Transfer Entropy (SDNTE)Fast and efficient to root cause diagnostic in real time using xD-Markov machine. Accepts only stationary variables.[65]
Normalized (Direct) Transfer EntropyIt assumes a random delay in the occurrence of the alarms and mutual independence between them.[66]

Appendix A.2. Distances and R Libraries

Table A3. Proximity metrics.
Table A3. Proximity metrics.
DistanceDescriptionReference
Canberra d p , q = i = 1 n p i q i p i q i [67]
Euclidean and Scaled Euclidean d p , q = i = 1 n p i q i 2
GAMGeneralized additive model distance is a form of parametric logistic regression. It belongs to the family of generalized linear models. It assumes that outcome variables depends linearly on unknown smooth functions of its predictor variables.[68]
GLMThe generalized linear model distance is a generalization of ordinary linear regression using a link function between predictors and outcome and for each variable a variance function of its predictors. It assumes that each predictor is generated by a probability distribution family (normal, binomial, exponential, poisson, etc.).[69]
Mahalanobis and Robust Mahalanobis d p , q ,   θ = p q T S 1 p q
Where, p and q are vectors of size n, θ is the probability distribution over Rn, and S is the positive definite covariance matrix.
[70]
Manhattan (Taxicab) d p , q = p i q i
Hermann Minkowski (1864–1909)
Maximum d p , q = m a x i p i q i
Pafnuty Chebyshev (1821–1894)
Minkowski d p , q = i = 1 n p i q i s s
s = 1, Manhattan distance; s = 2, Euclidean distance;
s = , Chebyshev distance. Hermann Minkowski (1864–1909)
Table A4. R Packages used.
Table A4. R Packages used.
PackageDescriptionReference
matchItIt implements a wide variety of matching algorithms such as nearest Mahalanobis, optimal matching, fullmatch, genetic, and coarsened exact matching.[44]
optmatchFinds the optimal match and support variable number of controls and full matching. It uses auction algorithm implemented with RELAX IV Fortran code [42] to solve the minimum cost flow problem. It has academic license.[41]
sensitivitymultSensitivity analysis for normal distributions using M-Test.[51]
carSensitivity analysis for no normal distributions using Levene test.[71]
cobaltGenerates balance tables and plots for covariates of matched groups[72]
rboundsImplements sensitivity analysis methods documented in [29] and calculates point estimates from HodgesLehmann and Wilcoxon tests.[73]
tableoneIt summarizes the variables (continuous or categorical) of the matching results, thus facilitating their reading. In addition, it performs statistical tests and calculates the standardized mean difference.[74]
lightgbmOptimized machine learning algorithm using gradient boosting decision tree (GBDT) implemented with gradient-based one-side sampling (GOSS) and exclusive feature bundling (EFB)[60]
randomForestA classification and regression algorithm that combines randomized decision trees and aggregates their predictions by averages.[75]
metaImplements methods for meta-analysis of random-effect models and common-effect models.[76]
Table A5. DerSimonianLaird nomenclature.
Table A5. DerSimonianLaird nomenclature.
VariableNomenclature of Random Effect Model for DerSimonian-Laird
kNumber of comparative studies
i1, …, k
nTi, nCiNumber of records in treatment and control groups
rTi, rCiProportion of records with some event in treatment and control groups
ѲiTrue treatment effect in the i-th study
eiSampling error in the i-th study
siStandard deviation of the i-th study
Observed
treatment effect
Yi = True treatment effect + sampling error
Yi = Ѳi + ei
Variance of ei σ i 2
The sampling variance capturing intra-study variance and the sampling size. It is usually unknown, but it is estimated from study data ( s i 2 ) .
ѲiѲi = μ + δi
μOverall treatment effect of study population
δiδi = Ѳi − μ
It is the deviation of the study from overall treatment effect.
Variance of δiτ2
It is the between-study variance and constitutes both the degree of variation in the true treatment effect across studies and the degree of variation in biased assessments of the treatment effect of individual studies. It is estimated from study data (t2).
τ2 = 0Homogeneity among the true treatment effects. Therefore, Ѳi = μ
τ2 > 0 reflects treatment effects heterogeneity.
yiThe observed treatment effect for i-th study. It is a random variable.
y1,…, yk are values of random variable from a distribution with overall means μ; true treatment effect Ѳi;
and variances τ 2 + σ 1 2 ,   ,   τ 2 + σ k 2 ;   τ 2   >   0 ;   and   σ 1 2 > 0.
Variance of each observed treatment
effect
τ2 +  σ i 2
Inter-study variance = τ2.
Intra-study variance = σ i 2 ; it is approximately equal to s i 2 .
Table A6. DerSimonianLaird Formulas.
Table A6. DerSimonianLaird Formulas.
Formulas for the Estimation of the Overall Population Treatment Effect μEquation
Estimate of the initial weighting w i 0 w i 0 = 1 s i 2 (A1)
Estimate   of   initial   y w ( 0 )   y w 0 = i = 1 k w i 0 y i i = 1 k w i 0 (A2)
Estimate Q statistics Q = i = 1 k w i 0 y i     y w ( 0 ) 2 (A3)
Estimate   t 2 t 2 = Q     k 1 i = 1 k w i 0       i = 1 k w i 0 2 i = 1 k w i 0 (A4)
Estimate second step weighting, wi w i = 1 t 2 + s i 2 (A5)
Estimate m w m w = i = 1 k w i y i i = 1 k w i (A6)
Estimate τ2, weighted method of moments τ 2   i = 1 k w i y i     m w 2           i = 1 k w i s i 2         i = 1 k w i 2 s i 2 i = 1 k w i i = 1 k w i       i = 1 k w i 2 i = 1 k w i (A7)
Estimate final weighting Wi W i = 1 τ 2 + s i 2 (A8)
Estimate μ M W = i = 1 k W i y i i = 1 k W i (A9)
Estimate   Standard   error   μ W S t a n d a r   e r r o r   o f   μ W = S E M W =   1   i = 1 k W i 1 2 (A10)
Confidence Interval μCI =  M W   ±   1.96 SE M W (A11)

Appendix B

Treatment Intervals, Density Functions, and Selected Models

Table A7. Downtime intervals of machines for each treatment (MSUMALL).
Table A7. Downtime intervals of machines for each treatment (MSUMALL).
TratMinMaxM1M2M3M4M5M6M7Total
0--3521 978 3008 7318 1831 3087 4039 23,782
159 291 166 62 54 990 50 331 173 1826
2292 441 150 68 80 897 138 317 210 1860
3442 580 142 76 109 779 129 436 209 1880
4581 722 152 77 85 715 127 493 233 1882
5723 908 187 91 96 693 133 355 308 1863
6909 1160 219 112 103 673 171 349 243 1870
71161 1497 222 106 107 661 165 370 240 1871
81498 2010 252 136 119 605 178 389 202 1881
92011 3005 274 154 124 531 200 371 214 1868
103006 18,113 334 237 162 406 231 407 195 1972
Total 5619 2097 4047 14,268 3353 6905 6266 42,555
% Control units62.7%46.6%74.3%51.3%54.6%44.7%64.5%55.9%
% Treatment units37.3%53.4%25.7%48.7%45.4%55.3%35.5%44.1%
Figure A1. MSUMALL density function for each machine.
Figure A1. MSUMALL density function for each machine.
Mathematics 11 04506 g0a1
Figure A2. WASTELM density function for each machine.
Figure A2. WASTELM density function for each machine.
Mathematics 11 04506 g0a2
Table A8. Selected models: balance information and causal effect size.
Table A8. Selected models: balance information and causal effect size.
MCHTreatMethodDistanceC1C2CUTUNSMDNRVCMUTMURVWtpvalEffMDEffRFCIhighCIlowZScore
M11Optmatchmahalanobis113521166001661660.840.0119.8019.4221.7317.880.27
M12FullMatchscaled_euclidean1135211500035211500.820.0216.3516.1917.0915.620.20
M13Geneticbart113521142001421420.640.1116.0715.6017.6114.540.19
M14Geneticgam113521152001521520.910.2511.5311.27 0.13
M15Geneticgam113521187001871870.920.358.007.2610.005.990.10
M16Geneticgam113521219002192191.020.1213.0513.7814.0612.030.15
M17Geneticgam113521222002222221.420.2410.289.87 0.11
M18Geneticgbm213521252102522520.760.1411.9712.8712.8911.050.13
M19FullMatchscaled_euclidean2135212741035212740.610.0024.3824.5226.2122.550.30
M110FullMatchmahalanobis2135213341035213340.580.0034.2534.1137.5730.940.41
M21Geneticrobust_mahalanobis11978620062621.820.377.878.3814.341.410.16
M22Optmatchmahalanobis11978680068680.480.1415.1315.5617.6712.590.26
M23FullMatchrobust_mahalanobis119787600978761.860.089.7810.3911.408.150.18
M24Geneticgbm11978770077771.840.1215.6716.5519.4211.910.25
M25Geneticgbm11978910091911.420.2914.4514.1627.970.930.16
M26Geneticgbm11978112001121120.690.3811.6312.0711.6311.630.12
M27Nearestscaled_euclidean11978106001061060.700.645.345.829.621.060.07
M28Geneticscaled_euclidean11978136001361360.540.0219.7020.5721.2218.190.29
M29Optmatchscaled_euclidean11978154001541541.140.427.546.378.586.490.09
M210Geneticscaled_euclidean11978237002372370.770.415.845.6811.220.450.08
M31Geneticgam113007540054541.310.349.399.3515.912.870.18
M32Geneticmahalanobis113007800080800.410.0120.9922.1423.4618.520.44
M33Geneticscaled_euclidean113007109001091091.390.535.805.079.671.930.09
M34Geneticbart113007850085850.450.0020.5520.6923.4217.670.45
M35Geneticgam113007950095950.730.227.477.328.536.400.18
M36Geneticgam113007103001031031.880.237.717.959.146.280.17
M37Geneticgam113007106001061060.570.1413.7113.1015.3812.050.21
M38Geneticgam113007119001191190.610.277.056.278.285.820.14
M39Nearestscaled_euclidean113007124001241240.660.1412.6511.9313.4611.830.19
M310Geneticglm113007162001621620.710.1411.8412.3413.0410.630.16
M41Geneticeuclidean117318990009909900.650.104.905.095.294.520.07
M42Nearestscaled_euclidean117318897008978970.660.0011.6511.7512.4210.890.16
M43Geneticmahalanobis117318779007797790.780.0013.6813.7813.9813.370.17
M44Geneticgbm117318715007157150.850.018.538.679.058.000.15
M45Optmatchscaled_euclidean117318693006936930.620.0013.8914.0314.4613.330.19
M46Geneticbart117318673006736730.610.0017.8318.2918.5517.110.23
M47Nearestrobust_mahalanobis117318661006616610.650.0020.1920.3821.1719.200.25
M48Geneticscaled_euclidean117318605006056050.490.0022.5822.6923.2821.890.26
M49Geneticeuclidean117318531005315310.570.0029.2829.0130.1228.440.36
M410Geneticeuclidean117318406004064060.490.0048.6049.2750.2346.980.52
M51Geneticrobust_mahalanobis111831500050501.640.0724.8125.6528.3621.260.37
M52Nearestmahalanobis111831138001381380.700.0315.6415.6017.5513.740.27
M53Geneticmahalanobis111831129001291290.760.0710.8610.4412.149.580.23
M54Geneticglm111831127001271270.390.0219.8220.0921.4118.230.30
M55Geneticrobust_mahalanobis111831133001331330.790.0715.6215.9617.3313.910.23
M56Geneticmahalanobis111831171001711710.740.079.999.8414.865.120.20
M57Optmatchglm111831165001651650.630.0126.1726.3226.7725.560.31
M58Geneticgam111831178001781780.700.0115.5615.6917.6113.500.26
M59FullMatcheuclidean1118312000018312000.820.0025.2224.9627.3823.060.37
M510FullMatchscaled_euclidean1118312310018312310.480.0036.7937.4245.7027.880.53
M61FullMatchscaled_euclidean1130873310030873310.840.008.068.119.416.720.24
M62FullMatchscaled_euclidean1130873170030873170.840.054.364.247.231.480.12
M63FullMatchscaled_euclidean1130874360030874360.610.0012.5612.7713.4311.690.25
M64FullMatchmahalanobis1130874930030874930.700.007.717.7512.802.620.21
M65FullMatchscaled_euclidean1130873550030873550.770.007.217.239.524.910.24
M66FullMatcheuclidean1130873490030873490.580.0014.8714.9317.8011.940.34
M67FullMatchscaled_euclidean1130873700030873700.700.0016.2816.4823.389.190.42
M68FullMatchmahalanobis1130873890030873890.510.0019.7519.9625.7913.720.44
M69Geneticbart113087371003713710.740.0114.9714.8716.0713.880.20
M610FullMatcheuclidean1130874070030874070.460.0034.6935.6134.6934.690.46
M71Geneticrobust_mahalanobis114039173001731730.790.276.596.489.094.100.12
M72Nearestmahalanobis114039210002102100.980.366.877.058.155.590.09
M73Nearesteuclidean114039209002092090.820.068.518.5811.465.570.18
M74Optmatchmahalanobis114039233002332330.720.0512.1412.7414.0610.230.18
M75Nearestmahalanobis114039308003083080.650.0214.7414.4715.8513.640.19
M76Optmatchglm114039243002432430.680.237.117.717.946.290.11
M77Geneticscaled_euclidean114039240002402401.160.147.727.818.786.660.14
M78Optmatchrobust_mahalanobis114039202002022020.590.266.496.569.953.020.11
M79Geneticgam114039214002142141.180.0513.0413.0314.6411.430.19
M710Geneticbart114039195001951950.620.0414.9915.7716.6813.300.21
C1:Priority of condition 1; C2: Priority of condition 2; CU: Control units; TU: Treatment units; NSMD: Non-balanced variables by SMS; NRV: Non-balanced variables by RV; CMU: Control-matched units; TMU: Treatment-matched units; RV: Variance ratio; Wtpval: p-value of Welch test; EffMD: Mean effect size of treatment of homologous models; EffRF: Effect size of RF model of treatment selected; CIhigh: Confidence interval high value for machinetreatment of homologous models; CIlow: Confidence interval low value for machinetreatment of homologous models; HLUCE: HodgesLehmann unconfounded estimate; ZScore: Z-score of selected model.

References

  1. Alfieri, A. Workload Simulation and Optimisation in Multi-Criteria Hybrid Flowshop Scheduling: A Case Study. Int. J. Prod. Res. 2009, 47, 5129–5145. [Google Scholar] [CrossRef]
  2. Li, X.; Wang, J.; Huang, C.; Gao, D.; Lu, G.; Lu, L.; Wang, Z. Mathematical Models for Predicting the Quasi-Static Stress Characteristics of Corrugated Paperboard with Sinusoidal Core along the Longitudinal Compression. Int. J. Mech. Sci. 2018, 149, 136–149. [Google Scholar] [CrossRef]
  3. Research and Markets. In Paperboard Packaging Market—Growth, Trends, COVID-19 Impact, and Forecasts (2023–2028); Mordor Intelligence: Dublin, Ireland, 2023; Available online: https://www.researchandmarkets.com/reports/4536057/paperboard-packaging-market-growth-trends (accessed on 22 September 2023).
  4. Farghaly, A.; Roux, S.L.; Peu, P.; Dabert, P.; Tawfik, A. Effect of Starvation Period on Microbial Community Producing Hydrogen from Paperboard Mill Wastewater Using Anaerobic Baffled Reactor. Environ. Technol. 2019, 40, 2389–2399. [Google Scholar] [CrossRef] [PubMed]
  5. Kirwan, M.J. Handbook of Paper and Paperboard Packaging Technology; Wiley-Blackwell: Chichester, UK, 2013; ISBN 9780470670668. [Google Scholar]
  6. Kot, S.; Grondys, K. Total Productive Maintenance in Enterprise Operations Support Processes. Appl. Mech. Mater. 2013, 309, 324–331. [Google Scholar] [CrossRef]
  7. Seyed Alinezhad, H.; Roohi, M.H.; Chen, T. A Review of Alarm Root Cause Analysis in Process Industries: Common Methods, Recent Research Status and Challenges. Chem. Eng. Res. Des. 2022, 188, 846–860. [Google Scholar] [CrossRef]
  8. Vuković, M.; Thalmann, S. Causal Discovery in Manufacturing: A Structured Literature Review. J. Manuf. Mater. Process. 2022, 6, 10. [Google Scholar] [CrossRef]
  9. Kaddour, J.; Lynch, A.; Liu, Q.; Kusner, M.J.; Silva, R. Causal Machine Learning: A Survey and Open Problems. arXiv 2022, arXiv:2206.15475. [Google Scholar]
  10. Hagedorn, C.; Huegle, J.; Schlosser, R. Understanding Unforeseen Production Downtimes in Manufacturing Processes Using Log Data-Driven Causal Reasoning. J. Intell. Manuf. 2022, 33, 2027–2043. [Google Scholar] [CrossRef]
  11. Choudhury, M.A.A.S. Plantwide Oscillations Diagnosis-Current State and Future Directions. Asia-Pac. J. Chem. Eng. 2011, 6, 484–496. [Google Scholar] [CrossRef]
  12. Stuart, E.A. Matching Methods for Causal Inference: A Review and a Look Forward. Stat. Sci. 2010, 25, 1–21. [Google Scholar] [CrossRef] [PubMed]
  13. Granger, C.W.J. Investigating Causal Relations by Econometric Models and Cross-Spectral Methods. Econometrica 1969, 37, 424–438. [Google Scholar] [CrossRef]
  14. Barnett, L.; Barrett, A.B.; Seth, A.K. Granger Causality and Transfer Entropy Are Equivalent for Gaussian Variables. Phys. Rev. Lett. 2009, 103, 238701. [Google Scholar] [CrossRef] [PubMed]
  15. Shojaie, A.; Fox, E.B. Annual Review of Statistics and Its Application Granger Causality: A Review and Recent Advances. Annu. Rev. Stat. Appl. 2022, 9, 289–319. [Google Scholar] [CrossRef] [PubMed]
  16. Chen, H.S.; Yan, Z.; Yao, Y.; Huang, T.B.; Wong, Y.S. Systematic Procedure for Granger-Causality-Based Root Cause Diagnosis of Chemical Process Faults. Ind. Eng. Chem. Res. 2018, 57, 9500–9512. [Google Scholar] [CrossRef]
  17. Schreiber, T. Measuring Information Transfer. Phys. Rev. Lett. 2000, 85, 461. [Google Scholar] [CrossRef]
  18. Bauer, M.; Cox, J.W.; Caveness, M.H.; Downs, J.J.; Thornhill, N.F. Finding the Direction of Disturbance Propagation in a Chemical Process Using Transfer Entropy. IEEE Trans. Control Syst. Technol. 2007, 15, 12–21. [Google Scholar] [CrossRef]
  19. Duan, P.; Yang, F.; Chen, T.; Shah, S.L. Direct Causality Detection via the Transfer Entropy Approach. IEEE Trans. Control Syst. Technol. 2013, 21, 2052–2066. [Google Scholar] [CrossRef]
  20. Pearl, J.; Russell, S. Bayesian Networks. Handb. Brain Theory Neural Netw. 2003, 2, 157–160. [Google Scholar] [CrossRef]
  21. Glymour, C.; Zhang, K.; Spirtes, P. Review of Causal Discovery Methods Based on Graphical Models. Front. Genet. 2019, 10, 524. [Google Scholar] [CrossRef]
  22. Rubin, D.B. Formal Modes of Statistical Inference for Causal Effects. J. Stat. Plan. Inference 1990, 25, 279–292. [Google Scholar] [CrossRef]
  23. Rubin, D.B. Estimating Causal Effects of Treatments in Randomized and Nonrandomized Studies. J. Educ. Psychol. Am. Psychol. Assoc. 1974, 66, 688. [Google Scholar] [CrossRef]
  24. Morgan, S.L.; Winship, C. Counterfactuals and Causal Inference. In Methods and Principles for Social Research, 2nd ed.; Cambridge University Press: New York, NY, USA, 2015; ISBN 9781107065079. [Google Scholar]
  25. Imbens, G.W.; Rubin, D.B. Causal Inference for Statistics, Social, and Biomedical Sciences: An Introduction; Cambridge University Press: New York, NY, USA, 2015; ISBN 0521885884/9780521885881. [Google Scholar]
  26. Holland, P.W. Statistics and Causal Inference. J. Am. Stat. Assoc. 1986, 81, 945–960. [Google Scholar] [CrossRef]
  27. Hernán, M.A.; Robins, J.M. Estimating Causal Effects from Epidemiological Data. J. Epidemiol. Community Health 2006, 60, 578–586. [Google Scholar] [CrossRef]
  28. Imai, K.; King, G.; Stuart, E.A. Misunderstandings between Experimentalists and Observationalists about Causal Inference. J. R. Statist. Soc. A 2008, 171, 481–502. [Google Scholar] [CrossRef]
  29. Rosenbaum, P.R. Design of Observational Studies; Springer: New York, NY, 2010; ISBN 0387947256. [Google Scholar]
  30. King, G.; Zeng, L. The Dangers of Extreme Counterfactuals. Political Anal. 2006, 14, 131–159. [Google Scholar] [CrossRef]
  31. Ho, D.E.; Imai, K.; King, G.; Stuart, E.A. Matching as Nonparametric Preprocessing for Reducing Model Dependence in Parametric Causal Inference. Political Anal. 2007, 15, 199–236. [Google Scholar] [CrossRef]
  32. Morgan, S.L. Handbook of Causal Analysis for Social Research; Springer: Dordrecht, The Netherlands, 2013; ISBN 9789400760943/99789400760943. [Google Scholar]
  33. Rosenbaum, P.R. Modern Algorithms for Matching in Observational Studies. Annu. Rev. Stat. Appl. 2020, 7, 143–176. [Google Scholar] [CrossRef]
  34. Smith, H.L. Matching with Multiple Controls to Estimate Treatment Effect in Observational Studies. Sociol. Methodol. 1997, 325–353. [Google Scholar] [CrossRef]
  35. Dehejia, R.H.; Wahba, S. Causal Effects in Nonexperimental Studies: Reevaluating the Evaluation of Training Programs. J. Am. Stat. Assoc. 1999, 94, 1053–1062. [Google Scholar] [CrossRef]
  36. Rosenbaum, P.R.; Rubin, D.B. The Central Role of the Propensity Score in Observational Studies for Causal Effects. Biometrica 1983, 70, 41–55. [Google Scholar] [CrossRef]
  37. Cochran, W.G.; Rubin, D.B. Controlling Bias in Observational Studies: A Review. In Matched Sampling for Causal Effects; Cambridge University Press: New York, NY, USA, 2006; pp. 30–58. ISBN 9780511810725. [Google Scholar]
  38. Muja, M.; Lowe, D.G. Fast Approximate Nearest Neighbors with Automatic Algorithm Configuration. VISAPP 2009, 1, 331–340. [Google Scholar]
  39. Gu, X.S.; Rosenbaum, P.R. Comparison of Multivariate Matching Methods: Structures, Distances, and Algorithms. J. Comput. Graph. Stat. 1993, 2, 405–420. [Google Scholar] [CrossRef]
  40. Rosenbaum, P.R.; Rubin, D.B. Constructing a Control Group Using Multivariate Matched Sampling Methods That Incorporate the Propensity Score. Am. Stat. 1985, 39, 33–38. [Google Scholar] [CrossRef]
  41. Hansen, B.B.; Olsen Klopfer, S. Optimal Full Matching and Related Designs via Network Flows. J. Comput. Graph. Stat. 2006, 15, 609–627. [Google Scholar] [CrossRef]
  42. Bertsekas, D.P.; Tseng, P. RELAXATION METHODS FOR MINIMUM COST ORDINARY AND GENERALIZED NETWORK FLOW PROBLEMS. Oper. Res. 1988, 36, 93–114. [Google Scholar] [CrossRef]
  43. Hansen, B.B. Full Matching in an Observational Study of Coaching for the SAT. J. Am. Stat. Assoc. 2004, 99, 609–618. [Google Scholar] [CrossRef]
  44. Stuart, E.A.; King, G.; Imai, K.; Ho, D. MatchIt: Nonparametric Preprocessing for Parametric Causal Inference. J. Stat. Softw. 2011, 42, 1–28. [Google Scholar] [CrossRef]
  45. Diamond, A.; Sekhon, J.S. Genetic Matching for Estimating Causal Effects: A General Multivariate Matching Method for Achieving Balance in Observational Studies. Rev. Econ. Stat. 2013, 95, 932–945. [Google Scholar] [CrossRef]
  46. Stuart, E.A.; Lee, B.K.; Leacy, F.P. Prognostic Score-Based Balance Measures Can Be a Useful Diagnostic for Propensity Score Methods in Comparative Effectiveness Research. J. Clin. Epidemiol. 2013, 66, S84–S90.e1. [Google Scholar] [CrossRef]
  47. Flury, B.K.; Riedwyl, H. Standard Distance in Univariate and Multivariate Analysis. Am. Stat. 1986, 40, 249–251. [Google Scholar] [CrossRef]
  48. Austin, P.C. Balance Diagnostics for Comparing the Distribution of Baseline Covariates between Treatment Groups in Propensity-Score Matched Samples. Stat. Med. 2009, 28, 3083–3107. [Google Scholar] [CrossRef] [PubMed]
  49. Hansen, B.B. The Prognostic Analogue of the Propensity Score. Biometrika 2008, 95, 481–488. [Google Scholar] [CrossRef]
  50. Zhang, Z.; Kim, H.J.; Lonjon, G.; Zhu, Y. Balance Diagnostics after Propensity Score Matching. Ann. Transl. Med. 2019, 7, 16. [Google Scholar] [CrossRef] [PubMed]
  51. Rosenbaum, P.R. Sensitivity Analyses Informed by Tests for Bias in Observational Studies. Biometrics 2021, 79, 475–487. [Google Scholar] [CrossRef]
  52. Rosenbaum, P.R. Hodges–Lehmann Point Estimates of Treatment Effect in Observational Studies. J. Am. Stat. Assoc. 1993, 88, 1250–1253. [Google Scholar] [CrossRef]
  53. Cohen, J. Statistical Power Analysis for the Behavioral Sciences, 2nd ed.; Lawrence Erlbaum Associates: New York, NY, USA, 1988; ISBN 0805802835. [Google Scholar]
  54. Kallus, N. Generalized Optimal Matching Methods for Causal Inference. J. Mach. Learn. Res. 2020, 21, 1–54. Available online: https://jmlr.org/papers/v21/19-120.html (accessed on 22 September 2023).
  55. Zhao, S.; van Dyk, D.A.; Imai, K. Propensity Score-Based Methods for Causal Inference in Observational Studies with Non-Binary Treatments. Stat. Methods Med. Res. 2020, 29, 709–727. [Google Scholar] [CrossRef]
  56. DerSimonian, R.; Kacker, R. Random-Effects Model for Meta-Analysis of Clinical Trials: An Update. Contemp. Clin. Trials 2007, 28, 105–114. [Google Scholar] [CrossRef]
  57. Borenstein, M.; Hedges, L.V.; Higgins, J.P.T.; Rothstein, H.R. A Basic Introduction to Fixed-Effect and Random-Effects Models for Meta-Analysis. Res. Synth. Methods 2010, 1, 97–111. [Google Scholar] [CrossRef]
  58. Maathuis, M.H.; Nandy, P. A Review of Some Recent Advances in Causal Inference. arXiv 2015. [Google Scholar] [CrossRef]
  59. Pearl, J. Causality: Models, Reasoning and Inference, 2nd ed.; Cambridge University Press: New York, NY, USA, 2009; ISBN 052189560X/9780521895606. [Google Scholar]
  60. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.-Y. LightGBM: A Highly Efficient Gradient Boosting Decision Tree. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  61. Yuan, T.; Qin, S.J. Root Cause Diagnosis of Plant-Wide Oscillations Using Granger Causality. J. Process. Control. 2014, 24, 450–459. [Google Scholar] [CrossRef]
  62. Chen, Q.; Lang, X.; Lu, S.; ur Rehman, N.; Xie, L.; Su, H. Detection and Root Cause Analysis of Multiple Plant-Wide Oscillations Using Multivariate Nonlinear Chirp Mode Decomposition and Multivariate Granger Causality. Comput. Chem. Eng. 2021, 147, 107231. [Google Scholar] [CrossRef]
  63. Duan, P.; Yang, F.; Shah, S.L.; Chen, T. Transfer Zero-Entropy and Its Application for Capturing Cause and Effect Relationship Between Variables. IEEE Trans. Control. Syst. Technol. 2015, 23, 855–867. [Google Scholar] [CrossRef]
  64. Guo, C.; Yang, F.; Yu, W. A Causality Capturing Method for Diagnosis based on Transfer Entropy by Analyzing Trends of Time Series. IFAC-PapersOnLine 2015, 48, 778–783. [Google Scholar] [CrossRef]
  65. Rashidi, B.; Singh, D.S.; Zhao, Q. Data-Driven Root-Cause Fault Diagnosis for Multivariate Non-Linear Processes. Control. Eng. Pract. 2018, 70, 134–147. [Google Scholar] [CrossRef]
  66. Hu, W.; Wang, J.; Chen, T.; Shah, S.L. Cause-Effect Analysis of Industrial Alarm Variables Using Transfer Entropies. Control. Eng. Pract. 2017, 64, 205–214. [Google Scholar] [CrossRef]
  67. Lance, G.N.; Williams, W.T. Mixed-Data Classificatory Programs I—Agglomerative Systems. Aust. Comput. J. 1967, 1, 15–20. [Google Scholar]
  68. Hastle, T.; Tibshirani, R. Generalized Additive Models; Some Applications. J. Am. Stat. Assoc. 1987, 82, 371–386. [Google Scholar] [CrossRef]
  69. Nelder, J.A.; Wedderburn, R.W.M. Generalized Linear Models. J. R. Stat. Soc. Ser. A 1972, 135, 370–384. [Google Scholar] [CrossRef]
  70. Mahalanobis, P.C. On the Generalized Distance in Statistics. Sankhyā Indian J. Stat. Ser. A 2018, 80-A, S1–S7. Available online: https://www.jstor.org/stable/e48513082 (accessed on 22 September 2023).
  71. Fox, J.; Weisberg, S. An R Companion to Applied Regression, 3rd ed.; SAGE Publications, Inc.: Thousand Oaks, CA, USA, 2019; ISBN 9781544336473. [Google Scholar]
  72. Greifer, N. Covariate Balance Tables and Plots: A Guide to the Cobalt Package. 2020. Available online: https://cran.r-project.org/web/packages/cobalt/index.html (accessed on 22 September 2023).
  73. Keele, L.J. Perform Rosenbaum Bounds Sensitivity Tests for Matched and Unmatched Data. R Package ‘rbounds’. 2022. Available online: https://cran.r-project.org/web/packages/rbounds/rbounds.pdf (accessed on 22 September 2023).
  74. Yoshida, K.; Bohn, J.; Yoshida, M. Package ‘tableone’. R Foundation for Statistical Computing. 2020. Available online: https://github.com/kaz-yos/tableone (accessed on 22 September 2023).
  75. Liaw, A.; Wiener, M. Classification and Regression by randomForest. R News 2002, 2, 18–22. Available online: https://journal.r-project.org/articles/RN-2002-022/RN-2002-022.pdf (accessed on 22 September 2023).
  76. Cohen, Y.; Cohen, J.Y. Statistics and Data with R. An Applied Approach through Examples; John Wiley & Sons: Chichester, UK, 2008; ISBN 0470721898. [Google Scholar]
Figure 1. Exploratory matching model search algorithm (EMMSA) for causal inference.
Figure 1. Exploratory matching model search algorithm (EMMSA) for causal inference.
Mathematics 11 04506 g001
Figure 2. Feature importance for WASTELM. Note: Parameters for LightGBM Regression. max_depth = 5, num_leaves = 28, num_iterations = 200, early_stopping_rounds = 40, and learning_rate = 0.46.
Figure 2. Feature importance for WASTELM. Note: Parameters for LightGBM Regression. max_depth = 5, num_leaves = 28, num_iterations = 200, early_stopping_rounds = 40, and learning_rate = 0.46.
Mathematics 11 04506 g002
Figure 3. Homologous models by methods, machine, and treatment.
Figure 3. Homologous models by methods, machine, and treatment.
Mathematics 11 04506 g003
Figure 4. Homologous models by distances, methods, machine, and treatment.
Figure 4. Homologous models by distances, methods, machine, and treatment.
Mathematics 11 04506 g004
Figure 5. Mean effect size of homologous models and Z-scores of selected models.
Figure 5. Mean effect size of homologous models and Z-scores of selected models.
Mathematics 11 04506 g005
Figure 6. Distances used in selected models.
Figure 6. Distances used in selected models.
Mathematics 11 04506 g006
Figure 7. Meta-analysis for (a) machine 6 and (b) treatment 10.
Figure 7. Meta-analysis for (a) machine 6 and (b) treatment 10.
Mathematics 11 04506 g007
Table 1. Dataset variables description.
Table 1. Dataset variables description.
VariableDescriptionDatatype/UnitOperation
MCHPrint subsystems (machine)Categoric-
PRICELAMSheet priceeuros-
MT2LAMSheet aream2-
TESTBox compression testN-
ORDEREDLMLinear meters of an orderLM-
SCHEDLMPlanned linear metersLM-
FINALLMFinal linear meters to send to clientLM-
OVERPROCLMLinear meters of overprocessingLM-
ASLEEPTIMETime of the production order since it was created until it enters the machine queueminutesCalculated
QUEUETIMEQueue time of a production orderminutesCalculated
PRODCTIMEOrder production timeminutesCalculated
MSETUPStoppage time for setting the order on the machineseconds-
MSUMALLSum of the time of all stoppages after setup of a production order (causal variable)secondsCalculated and aggregated
TMSUMALLTreatment variable for causal variable1, …, 10Calculated
WASTELMLinear meters of wasteLMCalculated
Table 2. Model number resulting of the application of HMSA by conditions and priorities.
Table 2. Model number resulting of the application of HMSA by conditions and priorities.
Condition 2: Methods and Priorities
Condition 1
    Condition 3: Homol.
CEMFullMatchGeneticNearestOptmatch
21231121Total Cond1%
Total
Priority 1176805713083931886189187571%
           017538571304391588685116862%
           1142 400160 10470738%
Priority 2247217411831553210062424%
           019691741179153329860897%
           153 42 2163%
Priority 3 42471368201455%
           0 4247136820145100%
I. Total by priorities2001567617510935091263092644100%
II. Total by methods20040710936353092644
III. % Models by meth.7.6%15.4%41.3%24.0%11.7%100%
IV. Homol. models645404162106723
V. % Homol. models (II)3.0%11.1%37.0%25.5%34.3%27.3%
VI. % Homol. models0.8%6%55.9%22%15%100.0%
VII. % Homol. Models of total 0.2%1.7%15.3%6.1%4.0%27.3%
VIII. Selected models015398870
IX. % Selected models by meth.0.0%21.4%55.7%11.4%11.4%100.0%
Note: Condition 1 and 2 can have up to three priorities each one. Condition 3 can be 0 (for discarded models) or 1 (for homologous models). Total values are shown in bold, either by conditions, priority, methods, homologous models or selected models.
Table 3. Selected models: Z-scores and EffMeans greater than 0.30 and 20 LM, respectively.
Table 3. Selected models: Z-scores and EffMeans greater than 0.30 and 20 LM, respectively.
Selected Models (Z-Score > 0.30)Seleted Models (EffMean > 20 LM)
MCHTreatEffMDZ-ScoreMCHTreatEffMDZ-Score
M1T924.380.30M1T924.380.30
M1T1034.250.41M1T1034.250.41
M3T220.990.44M3T220.990.44
M3T420.550.45M3T420.550.45
----M4T720.190.25
----M4T822.580.26
M4T929.280.44M4T929.280.44
M4T1048.600.52M4T1048.600.52
M5T124.810.37M5T124.810.37
M5T726.170.31M5T726.170.31
M5T925.220.37M5T925.220.37
M5T1036.790.53M5T1036.790.53
M6T614.870.34----
M6T716.280.42----
M6T819.750.44----
M6T1034.690.46M6T1034.690.46
Table 4. Meta-analysis for each machine and treatment with its maximum effect.
Table 4. Meta-analysis for each machine and treatment with its maximum effect.
Meta-Analysis for Each MachineMaximum Effect for Each Machine
MCHRandEff95%-CII2τ2Max (Treat)SMD95%-CIEffAvg LM95%-CI LM
M10.18[0.09;0.26]37%0.0067T10-Full-Mah0.41[0.25;0.56]34.25[12.53;55.97]
M20.11[0.03;0.19]0%0.0001T8-Near-Mah0.26[0.02;0.50]19.70[18.18;21.22]
M30.07[−0.01;0.16]0%0.0000T3-Full-Calip0.30[−0.02;0.61]5.80[5.99;5.61]
M40.20[0.13;0.28]76%0.0108T10-Full-Glm0.47[0.33;0.61]48.60[14.48;82.72]
M50.23[0.10;0.36]70%0.0271T10-Full-Mah0.53[0.34;0.72]36.79[13.19;60.39]
M60.32[0.24;0.40]65%0.0103T9-Full-Gam0.52[0.37;0.67]14.97[4.32;25.62]
M70.07[−0.01;0.15]44%0.0072T5-Near-Mah0.19[0.04;0.35]14.74[12.41;17.07]
All Mchs. 0.38 24.98[13.40;36.23]
Meta-Analysis for Each TreatmentMaximum Effect for Each Treatment
TreatRandEff95%-CII2τ2MaxSMD95%-CIEffAvg LM95%-CI LM
10.11[0.01;0.21]40%0.0055M1-Opt-Mah0.27[0.05;0.49]19.80[16.13;23.47]
20.13[0.07;0.20]0%0.0000M3-Full-Calip0.30[−0.02;0.61]20.99 [21.69;20.29]
30.16[0.09;0.22]5%0.0001M6-Full-Mah0.25[0.11;0.38]12.56 [6.53;18.59]
40.14[0.07;0.20]0%0.0001M6-Full-Mah0.21[0.09;0.34]7.71 [4.77;10.65]
50.16[0.10;0.22]10%0.0001M6-Full-Calip0.24[0.10;0.39]7.21 [4.51;9.91]
60.13[0.01;0.26]67%0.0181M6-Full-Glm0.35[0.20;0.50]14.87 [6.37;23.37]
70.18[0.03;0.33]79%0.0313M6-Full-Calip0.42[0.28;0.57]16.28 [5.81;26.75]
80.21[0.10;0.31]60%0.0120M6-Full-Mah0.44[0.29;0.58]19.75 [6.28;33.22]
90.22[0.02;0.42]87%0.0544M6-Full-Gam0.52[0.37;0.67]14.97 [4.32;25.62]
100.34[0.22;0.46]69%0.0186M5-Full-Mah0.53[0.34;0.72]34.69[12.44;56.94]
All Treat. 16.88[7.99;25.77]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Aviles-Lopez, R.; Luna del Castillo, J.d.D.; Montero-Alonso, M.Á. Exploratory Matching Model Search Algorithm (EMMSA) for Causal Analysis: Application to the Cardboard Industry. Mathematics 2023, 11, 4506. https://doi.org/10.3390/math11214506

AMA Style

Aviles-Lopez R, Luna del Castillo JdD, Montero-Alonso MÁ. Exploratory Matching Model Search Algorithm (EMMSA) for Causal Analysis: Application to the Cardboard Industry. Mathematics. 2023; 11(21):4506. https://doi.org/10.3390/math11214506

Chicago/Turabian Style

Aviles-Lopez, Richard, Juan de Dios Luna del Castillo, and Miguel Ángel Montero-Alonso. 2023. "Exploratory Matching Model Search Algorithm (EMMSA) for Causal Analysis: Application to the Cardboard Industry" Mathematics 11, no. 21: 4506. https://doi.org/10.3390/math11214506

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop