Next Article in Journal
Interleaving Shifted Versions of a PN-Sequence
Next Article in Special Issue
Deep Neural Network for Gender-Based Violence Detection on Twitter Messages
Previous Article in Journal
Boscovich Fuzzy Regression Line
Previous Article in Special Issue
A Review of Tracking and Trajectory Prediction Methods for Autonomous Driving
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Solving Regression Problems with Intelligent Machine Learner for Engineering Informatics

1
Department of Civil and Construction Engineering, National Taiwan University of Science and Technology, Taipei City 106335, Taiwan
2
Department of Civil Engineering, University of Architecture Ho Chi Minh City (UAH), Ho Chi Minh City 700000, Vietnam
3
Department of Information Management, National Central University, Taoyuan City 320317, Taiwan
*
Author to whom correspondence should be addressed.
Mathematics 2021, 9(6), 686; https://doi.org/10.3390/math9060686
Submission received: 18 February 2021 / Revised: 17 March 2021 / Accepted: 19 March 2021 / Published: 23 March 2021

Abstract

:
Machine learning techniques have been used to develop many regression models to make predictions based on experience and historical data. They might be used singly or in ensembles. Single models are either classification or regression models that use one technique, while ensemble models combine various single models. To construct or find the best model is very complex and time-consuming, so this study develops a new platform, called intelligent Machine Learner (iML), to automatically build popular models and identify the best one. The iML platform is benchmarked with WEKA by analyzing publicly available datasets. After that, four industrial experiments are conducted to evaluate the performance of iML. In all cases, the best models determined by iML are superior to prior studies in terms of accuracy and computation time. Thus, the iML is a powerful and efficient tool for solving regression problems in engineering informatics.

1. Introduction

Machine Learning (ML)-based methods for building prediction models have attracted abundant scientific attention and are extensively used in industrial engineering [1,2,3], design optimization of electromagnetic devices, and other areas [4,5]. The ML-based methods have been confirmed to be effective for solving real-world engineering problems [6,7,8]. Various supervised ML techniques (e.g., artificial neural network, support vector machine, classification and regression tree, linear (ridge) regression, and logistic regression) are typically used individually to construct single models and ensemble models [9,10]. To construct a series of models and identify the best one among these ML techniques, users need a comprehensive knowledge of ML and spend a significant effort building advanced models.
The primary objective of this research is to develop a user-friendly and powerful ML platform, called intelligent Machine Learner (iML), to help its users to solve real-world engineering problems with a shorter training time and greater accuracy than before. The iML can automatically build and scan all regression models, and then identify the best one. Novice users with no experience of ML can easily use this system. Briefly, the iML (1) helps users to make prediction model easily; (2) provides an overview of the parameter settings for the purpose of making objective choices; and (3) yields clear performance indicators, facilitating reading and understanding of the results, on which decisions can be based.
Four experiments were carried out to evaluate the performance of iML and were compared with previous studies. In the first experiment, empirical data concerning enterprise resource planning (ERP) for software projects by a leading Taiwan software provider over the last five years were collected and analyzed [1]. The datasets in the other three experiments were published on the UCI website [11,12,13]. Specifically, the purpose of the second experiment was to train a regression model of comparing the performance of CPU processors by using some characteristics as input. The third experiment involved forecasting the demand supporting structured productivity and high levels of customer service, and the fourth experiment involved estimating the total bikes rented per day.
The rest of this paper is organized as follows. Section 2 reviews application of machine learning techniques in various disciplines. Section 3 presents the proposed methodology and iML framework. Section 4 introduces the evaluation metrics to measure accuracy of the developed system. Section 5 demonstrates iML’s interface. Section 6 shows benchmarks between iML and WEKA (a free, open source program). Section 7 exhibits the applicability of iML in numerical experiments. Section 8 draws conclusions, and provides managerial implications and suggestions for future research.

2. Literature Review

Numerous researchers in various fields, such as ecology [14,15], materials properties [16,17,18], water resource [19], energy management [20], and decision support [21,22], use data-mining techniques to solve regression problems, and especially project-related problems [23,24]. Artificial neural network (ANN), support vector machine/regression (SVM/SVR) classification and regression tree (CART), linear ridge regression (LRR), and logistic regression (LgR) are the most commonly used methods for this purpose and are all considered to be among the best machine learning techniques [25,26,27]. Similarly, four popular ensemble models, including voting, bagging, stacking and tiering [28,29,30], can be built based on the meta-combination rules of aforementioned single models.
Chou (2009) [31] developed a generalized linear model-based expert system for estimating the cost of transportation projects. Dandikas et al. (2018) [32] assessed the advantages and disadvantages of regression models for predicting potential of biomethane. The results indicated that the regression method could predict variations in the methane yield and could be used to rank substrates for production quality. However, least squares-based regression usually leads to overfitting a model, failure to find unique solutions, and issues dealing with multicollinearity among the predictors [33], so ridge regression, another type of regularized regression, is favorably integrated in this study to avoid the above problems. Additionally, Sentas and Angelis (2006) [34] investigated the possibility of using some machine learning methods for estimating categorical missing values in software cost databases. They concluded that multinomial logistic regression was the best for imputation owing to its superior accuracy.
The general regression neural network was originally designed chiefly to solve regression problems [24,35]. Caputo and Pelagagge (2008) [36] compared the ANN with the parametric methods for estimating the cost of manufacturing large, complex-shaped pressure vessels in engineer-to-order manufacturing systems. Their comparison demonstrated that the ANN was more effective than the parametric models, presumably because of its better mapping capabilities. Rocabruno-Valdés et al. (2015) [37] developed models based on ANN for predicting the density, dynamic viscosity, and cetane number of methyl esters and biodiesel. Similarly, Ganesan et al. (2015) [38] used ANN to predict the performance and exhaust emissions of a diesel electricity generator.
SVM was originally developed by Vapnik (1999) for classification (SVM) and regression (SVR) [39,40]. Jing et al. (2018) [41] used SVM to classify air balancing, which is a key element for heating, ventilating, air-conditioning (HAVC), and variable air volume (VAV) system installation, and is useful for improving the energy efficiency by minimizing unnecessary fresh air to the air-conditioned zones. The results demonstrated that SVM achieved 4.6% of relative error value and is a promising approach for air balancing. García-Floriano et al. (2018) [42] used SVR to model software maintenance (SM) effort prediction. The SVR model was superior to regression, neural networks, association rules and decision trees, with 95% confidence level.
The classification and regression tree method (CART), introduced by Breiman et al. (2017) [43], is an effective method to solve classification and regression problems [42]. Choi and Seo (2018) [44] predicted the fecal coliform in the North Han River, South Korea by CART models, the test results showed the total correct classification rates of the four models ranged from 83.7% to 93.0%. Ru et al. (2016) [45] used the CART model to predict cadmium enrichment levels in reclaimed coastal soils. The results showed that cadmium enrichment levels had an accuracy of 78.0%. Similarly, Li (2006) [16] used CART to predict materials properties and behavior. Chou et al. (2014, 2017) [26,46] utilized the CART method to modeling steel pitting risk and corrosion rate and forecasting project dispute resolutions.
In addition to the aforementioned single models, Elish (2013) [47] used voting ensemble for estimating software development effort. The ensemble model outperformed all the single models in terms of Mean Magnitude of Relative Error (MMRE), and achieved competitive percentage of observations whose Magnitude of Relative Error (MRE) is less than 0.25 (PRED (25)) and recently proposed Evaluation Function (EF) results. Wang at el. (2018) demonstrated that ensemble bagging tree (EBT) model could accurately predict hourly building energy usage with MAPE ranging from 2.97% to 4.63% [48]. Comparing to the conventional single prediction model, EBT is superior in prediction accuracy and stability. However, it requires more computation time and is short of interpretability owing to its sophisticated model structure.
Chen et al. (2019) [49] showed that the stacking model outperformed the individual models, achieving the highest R 2 of 0.85, followed by XGBoost (0.84), AdaBoost (0.84) and random forest (0.82). For the estimation of hourly PM2.5 in China, the stacking model exhibited relatively high stability, with R 2 ranging from 0.79 to 0.92. Basant at el. (2016) [50] proposed a three-tier quantitative structure-activity relationship (QSAR) model. This model can be used for the screening of chemicals for future drug design and development process and safety assessment of the chemicals. In comparison with previously studies, the QSAR models on the same endpoint property showed the encouraging statistical quality of the proposed models.
According to the reviewed literature, various machine learning platforms have been developed for the past decades, such as the Scikit-Learn Python libraries, Google’s TensorFlow, WEKA and Microsoft Research’s CNTK. Users can find it easy to use a machine learning tool and/or framework to solve numerous problems as per their needs [51]. ML-based approaches have been confirmed to be effective in providing decisive information. Since there is no best model suitable to predict all problems (the “No Free Lunch” theorem [52,53]), a comprehensive comparison of single and ensemble models embedded within an efficient forecasting platform for solving real-world engineering problems is imperatively needed. The iML platform proposed in this study can efficiently address this issue.

3. Applied Machine Learning

3.1. Classification and Regression Model

3.1.1. Artificial Neural Network (ANN)

Neural networks (or artificial neural networks) comprise information-processing units, which are similar to the neurons in the human brain, except that a neural network is composed of artificial neurons (Figure 1) [54]. Particular, back-propagation networks (BPNNs) are widely used, and are known to be the most effective network models [55,56].
Equation (1) uses sigmoid function to activate each neuron in a hidden output layer, and the Scaled Conjugate Gradient Algorithm is used to calculate the weights of the network. BPNNs will be trained until the stopping criteria is reached by default settings in MATLAB.
n e t k = w k j O j   and   y k = f ( n e t k ) = 1 1 + e n e t k
where n e t k is the activation of the k th neuron; j is the set of neurons in the preceding layer; w k j is the weight of the connection between neuron k and neuron j; O j is the output neuron j; and y k is the sigmoid or logistic transfer function.

3.1.2. Support Vector Machine (SVM) and Support Vector Regression (SVR)

Developed by Cortes and Vapnik (1995) [57], SVM is used for binary classification problems. The SVM was created based on decision hyper-planes that determine decision boundaries in an input space or a high-dimensional feature space [40,58]. Binary classification can only classify samples into negative and positive while multi-class classification problems are complex (Figure 2). In this study, One Against All (OAA) is used to solve multiple classification problems.
The OAA-SVM constructs m SVM models for m -class classification problems, and the i t h SVM model is trained based on the dataset of the i t h class which includes a positive class and a negative class. In training, a set of l data points ( x i , y i ) i = 1 l , where x i R n the input data, and y i ( 1 , 2 , , m ) is the class label of x i ; the i t h SVM model is solved using the following optimization problem equation [59].
min w i , b , ξ J ( w i , b , ξ ) = 1 2 ( w i ) T w i + C i = 1 l ξ j i
subject   to :   { ( w i ) T φ ( x j ) + b i 1 ξ j i , y j = i , ( w i ) T φ ( x j ) + b i 1 + ξ j i , y j i ,   ξ j i 0 ,   j = 1 , , l .
When the SVM models have been solved, the class label of example x is predicted as follows:
y ( x ) = arg max i = 1 m ( ( w i ) T φ ( x ) + b i )
where i is the i th SVM model; w i is a vector normal to the hyper-plane; b i is a bias,   φ ( x ) is a nonlinear function that maps x to a high-dimension feature space, ξ i is the error in misclassification, and C 0 is a constant that specifies the trade-off between the classification margin and the cost of misclassification.
To train the SVM model, radial basic function (RBF) kernel maps samples non-linearly into a feature space with more dimensions. In this study, the RBF kernel is used as SVM kernel function.
K ( x i , x j ) = exp ( x i x j 2 2 σ 2 )
where σ is a positive parameter that controls the radius of RBF kernel function.
Support vector regression (SVR) [40] is one version of SVM. SVR computes a linear regression function for the new higher-dimensional feature space using ε -insensitive loss while simultaneously reducing model complexity of the model by minimizing w 2 . This process can be implemented by introducing (non-negative) slack variables ξ i , ξ i * to measure the deviation in training samples outside the ε -insensitive zone. The SVR can be formulated as the minimization of the following equation:
min w , b , ξ J ( w , b , ξ ) = 1 2 ( w ) T w + C i = 1 l ( ξ i + ξ i * )
subject   to :   { y i f ( x i , w ) ε + ξ i * f ( x i , w ) y i ε + ξ i ξ i * , ξ i   0 ,   i = 1 , , n
When SVR model has been solved, the value of example x is predicted as follows.
f ( x ) = ( α i α i * ) K ( x i , x ) + b
where K ( x i , x ) is the kernel function and α i * , α i are Lagrange multipliers in the dual function.

3.1.3. Classification and Regression Tree (CART)

Classification and regression tree technique is described as a tree on which each internal (non-leaf) node represents a test of an attribute, each branch represents the test result, and each leaf (or terminal) node has a class label and class result (Figure 3) [60]. The tree is “trimmed” until total error is minimized to optimize the predictive accuracy of the tree by minimizing the number of branches. The training CART is constructed through the Gini index. The formulas are as follows.
g ( t ) = j i p ( j | t ) p ( i | t )
p ( j | t ) = p ( j , t ) p ( t )
p ( j , t ) = p ( j ) N j ( t ) N j
p ( t ) = j p ( j , t )
G i n i   i n d e x = 1 p ( j , t ) 2
where i and j are the categorical variables in each item; N j ( t ) is the recorded number of nodes t in category j ; and N j . is the recorded number of the root nodes in category j ; and p ( j ) is the prior probability value for category j .

3.1.4. Linear Ridge Regression (LRR) and Logistic Regression (LgR)

Statistical models of the relationship between dependent variables (response variables) and independent variables (explanatory variables) are developed using linear regression (Figure 4). The general formula for multiple regression models is as follows.
y = f ( x ) = β o + j = 1 n β j x j + ε
where y is a dependent variable; β o is a constant; β j is a regression coefficient ( j = 1 , 2 , , n ), and ε is an error term.
Linear ridge regression (LRR) is a regularization technique that can be used together with generic regression algorithms to model highly correlated data [61,62]. Least squares method is a powerful technique for training the LRR model, which denotes β to minimize the Residual Sum Squares (RSS)-function. Therefore, the cost function is presented as below.
C o s t ( β ) = R S S ( β ) = i = 1 l ( y y ) 2 + λ ( j = 1 n β j 2 )
y = β 0 + β j x j
where λ is a pre-chosen constant, which is the product of a penalty term and the squared norm in the β vector of regression method, and y is the predicted values.
Statistician David Cox developed logistic regression in 1958 [63]. An explanation of logistic regression begins with an explanation of the standard logistic function. Equation (17) mathematically represents the logistic regression model.
p ( x ) = 1 1 + e ( β o + j = 1 n β j x j )
where p ( x ) is the probability that the dependent variable equals a “success” or “case” rather than a failure or non-case. β o and β j are found by minimizing cost function defined in Equation (18).
C o s t ( β ) = ( i = 1 l ( y i ln ( p ( x i ) ) + ( 1 y i ) ln ( 1 p ( x i ) ) ) ) +   λ 2 j = 1 n β j 2
where y i is the observed outcome of case x i , having 0 or 1 as possible values [64]

3.2. Ensemble Regression Model

In this study, several ensemble schemes, including voting, bagging, stacking, and tiering were investigated using the input data and described as below.
  • Voting: The voting ensemble model combines the outputs of the single models using a meta-rule. The mean of the output values is used in this study. According to the adopted ML models, 11 voting models are trained in this study, including (1) ANN + SVR, (2) ANN + CART, (3) ANN + LRR, (4) SVR + CART, (5) SVR + LRR, (6) CART + LRR, (7) ANN + SVR + CART, (8) ANN + CART + LRR, (9) ANN + CART + LRR, (10) SVR + CART + LRR, (11) ANN + SVR + CART + LRR. Figure 5a presents the voting ensemble model.
  • Bagging: The bagging ensemble model duplicates samples at random, and each regression model predicts values from the samples independently. The meta-rule is applied to all of the outputs in this study. Bagging ensemble model is depicted at Figure 5b.
  • Stacking: The stacking ensemble model is a two-stage model, and Figure 5c describes the principle of the model. In stage 1, each single model predicts one output value. Then, these outputs are used as inputs to train a model by these machine learning techniques again to make a meta-prediction in stage 2. There are four stacking models herein, including ANN (ANN, SVR, CART, LRR); SVR (ANN, SVR, CART, LRR); CART (ANN, SVR, CART, LRR); LRR (ANN, SVR, CART, LRR).
  • Tiering: Figure 5d illustrates the tiering ensemble model. There are two tiers inside a tiering ensemble model in this study. The first tier is to classify data into k classes on the basis of T value [18]. Machine learning technique in the first tier for classifying data needs to be identified. After classifying the data, the regression machine learning is used to train each data (Sub Data) of each class (second tier) to predict results. In the iML, we developed three types of models, including 2-class, 3-class, and 4-class. The equation for calculating T value is:
    T = y m a x + y m i n k
    where T is standard value, k is the number of classes, and y m a x and y m i n are the maximum and minimum of actual values, respectively.

3.3. K-Fold Cross Validation

K-fold cross validation is used to compare two or more prediction models. This method randomly divides a sample into a training sample and a test sample by splitting into K subsets. K-1 subsets are selected to train the model while the other is used to test, and this training process is repeated K times (Figure 6). To compare models, the average of performance results (e.g., RMSE, and MAPE) is computed. Kohavi (1995) stated that K = 10 provides analytical validity, computational efficiency, and optimal deviation [65]. Thus, K = 10 is used in this study. Performance metrics will be explained in details Section 4.

3.4. Intelligent Machine Learner Framework

Figure 7 presents the structure of iML. In stage 1 (data preprocessing), the data is classified distinctly for particular use in the Tiering ensemble model. Meanwhile, all data is divided into two main data groups, namely, learning data and test data, and the learning data is duplicated for training ensemble models.
At the next stage, all retrieved data is automatically used for training models, which include single models (ANN, SVR, LRR, and CART), and ensemble models (voting, bagging, stacking, and tiering). Notably, the tiering ensemble model needs to employ a classification technique to assign a class label to the original input at the first tier. A corresponding regression model for the particular class is then adopted at the second tier to obtain the predictive value [17,26].
Finally, in stage 3 (find the best model), the predictive performances of all the models learned (trained) in stage 2 using test dataset are compared to identify the best models. Section 4 describes the performance evaluation metrics in detail.

4. Mathematical Formulas for Performance Measures

To measure the performance of classification models, the accuracy, precision, sensitivity, specificity and the area under the curve (AUC) are calculated. For the regression models, five-performance measures, (i.e., correlation coefficient (R), mean absolute error (MAE), mean absolute percentage error (MAPE), root mean squared error (RMSE), and total error rate (TER)) are calculated. Table 1 presents a confusion matrix and Table 2 exhibits those performance measures [17,66].
In Table 2, MAE is the mean absolute difference between the prediction and the actual value. MAPE represents the mean percentage error between prediction and actual value, the smaller value of MAPE, the better prediction result achieved by the model. The MAPE is the index typically used to evaluate the accuracy of prediction models. RMSE represents the dispersion of errors by a prediction model. The statistical index that shows the linear correlation between two variables is denoted as R. Lastly, TER is the total difference of predicted and actual values [17].
The goal is to identify the model that yields the lowest error of test data. To obtain a comprehensive performance measure, the five statistical measures (RMSE, MAE, MAPE, 1-R, and TER) were combined into a synthesis index (SI) using Equation (20). Based on the SI values, the best model is identified.
SI = 1 m p i = 1 m p ( P i P m i n , i P m a x , i P m i n , i )
where m p = number of performance measures; P i = i th performance measure; and P m i n , i and P m a x , i are the maximum and minimum of i th measure. The SI range is 0–1; the SI value close to 0 indicates a better accuracy of the predictive model.

5. Design and Implementation of iML Interface

The iML was developed in MATLAB R2016a on a PC with an Intel Core i5-750 CPU, a clock speed of 3.4 GHz, and 8 GB of RAM, running Windows 10. Figure 8 presents a user-friendly interface for iML. First, users select models on setting-parameters board and set the parameters for the chosen models, which will be trained and analyzed. Next, users choose whether to test with either “K-Fold Validation” or “Percentage Split” before uploading the data. Notably, if “Percentage Split” is selected, the user only has to input percentage value of learning data. Then, users click on the “Run” button to train the model. Finally, the “Make Report” function is to create a report containing performance metrics of all selected models and the identified best model. Figure 9 displays a snapshot of report file in notepad.

6. Benchmarks between iML and WEKA

6.1. Publicly Available Datasets

Table 3 shows the publicly available datasets from the UCI Machine Learning Repository (https://archive.ics.uci.edu/mL/index.php; accessed 1 March 2021). The iML is benchmarked with WEKA (a free, open source program) using hold-out validation and K-fold cross-validation on the target datasets. All algorithm parameters are set default for both iML and WEKA platforms.

6.2. Hold-Out Validation

In this test, datasets are randomly partitioned into 80% and 20% for learning and test, respectively. Table 4, Table 5, Table 6, Table 7 and Table 8 show the one-time performance results on these five datasets. A model with a normalized SI value of 0.000 is the best prediction model among all the models tested by iML and WEKA. Notably, the best model can be automatically identified by iML with “one-click”. To train models with WEKA, the users need to build each model individually. Moreover, iML gives better test results of single, voting and bagging models than those of WEKA. Based on the benchmark results, iML is effective to find the best model in the hold-out validation.

6.3. K-Fold Cross-Validation

Tenfold cross-validation is used to evaluate the generalized performance of WEKA and iML. Table 9, Table 10, Table 11, Table 12 and Table 13 show the average performance measures of five datasets, respectively. Similarly, iML identifies better models in single, voting, and bagging schemes than those trained by WEKA. The best model for each dataset is automatically determined by iML. Therefore, iML is a powerful tool to find the best model in the cross-fold validation.

6.4. Discussion

Single, voting, bagging, and stacking models are compared using WEKA and iML, except for the tiering method, which is not available in WEKA. Additionally, unlike manual construction of individual models in WEKA interface, iML can automatically build and identify the best model for the imported datasets. Hold-out validation and tenfold cross-validation are used to evaluate the performance results (R, MAE, RMSE, and MAPE) in each scheme (single, voting, bagging, and stacking). The analytical results of either validation show that most of the models trained by iML are superior to those trained by WEKA using the same datasets. Hence, iML is an effective platform to solve regression problems.

7. Numerical Experiments

This section validates iML by using various industrial datasets, including (1) enterprise resource planning data [1], (2) CPU computer performance data [12], (3) customer data for a logistics company [13], and (4) daily data bike rentals [11]. Table 14 presents the initial parameter settings for these problems.

7.1. Enterprise Resource Planning Software Development Effort

Enterprise Resource Planning (ERP) data for 182 software projects of a leading Taiwan software provider over the last five years was collected, analyzed, and tested with K-fold cross validation.

7.1.1. Variable Selection

Experienced in-house project managers were interviewed to identify factors that affect the ERP software development effort (SDE). There are 182 samples and 17 attributes, and Table 15 summarizes the descriptive statistical data in details. The input and output attributes are defined by Chou el at. (2012) [1].

7.1.2. iML Results

iML automatically trains the models and calculates the performance values. Then, it compares the SI values ( SI local   and   SI globlal ) among the selected modeling type (singe, voting ensemble, bagging ensemble, stacking ensemble and tiering ensemble). Table 16 presents the detailed results of iML and Figure 10 plots the RMSE of best models for the studied case. Both SI local and SI global values of bagging ANN ensemble are equal to zero, which indicate that the bagging ANN ensemble is the best model in terms of prediction accuracy.
Three models (single, voting, and bagging) provided better results in terms of R (0.94 to 0.99) than the tiering and stacking ensemble models, which had the R values of 0.58 to 0.95. Among these three best models, in terms of MAPE, the bagging model exhibited the best balance of MAPE results from learning and test data (21.45% and 19.50%, respectively). The single and voting models depicted un-balanced MAPEs for training and test data (19.91% and 30.65% for the single model; 16.83% and 33.90% for the voting model). Thus, the bagging model was the best model to predict ERP.
The first experiment indicates that, the iML not only identifies the best model, but also reports the performance values of all the training models. Chou et al. (2012) obtained training and testing MAPEs of 26.8% and 27.3%, and RMSEs of 234.0157 h and 97.2667 h using Evolutionary Support Vector Machine Inference Model (ESIM) [1]. The iML yields the bagging ensemble model with MAPEs of 21.45% and 19.50%, and RMSEs of 70.28hr and 65.58 h for the same training and test data, respectively. As a result, the iML is effective to find the best model among the popular regression models.

7.2. Experiments on Industrial Datasets

Three additional experiments were performed to evaluate iML. To ensure a fair comparison, 70 % of the data was used for learning whereas the remaining 30% was utilized for testing.

7.2.1. Performance of CPU Processors

This experiment is about the comparison of performance of CPU processors. The data for this experiment was taken from Maurya and Gupta (2015) [12]. This dataset contained 209 samples with a total of 6 attributes (Table 17). The descriptions of the attributes are as follows: X1: Machine cycle time in nanoseconds (integer, input); X2: Minimum main memory in kilobytes (integer, input); X3: Maximum main memory in kilobytes (integer, input); X4: Cache memory in kilobytes (integer, input); X5: Minimum channels in units (integer, input); X6: Maximum channels in units (integer, input); and Y: Estimated relative performance (integer, output).

7.2.2. Daily Demand Forecasting Orders

This experiment is about the daily demand forecasting orders. The data used in this experiment was taken from Ferreira et al. (2016) [13]. Table 18 shows a statistical analysis of the data. There were 60 samples with 12 attributes, including X1: Week of the month (first week, second, third or fourth week of month, input); X2: Day of the week (Monday to Friday, input); X3: Urgent orders (integer, input); X4: Non-urgent orders (integer, input); X5: Type A orders (integer, input); X6: Type B orders (integer, input); X7: Orders of type C (integer, input); X8: Orders from the tax sector (integer, input); X9: Orders from the traffic controller sector (integer, input); X10: Orders from the banking sector 1 (integer, input); X11: Orders from the banking sector 2 (integer, input); X12: Banking orders 3 (integer, input); and Y: Total orders (integer, output).

7.2.3. Total Hourly-Shared Bike Rental per Days

The experiment is about the total hourly-shared bike rental per days. The data was adopted from Fanaee-T and Gama (2014) [11], and statistically analyzed in Table 19. In total, there were 731 samples and 11 attributes, defined as follows: X1: Season (1: spring, 2: summer, 3: fall, 4: winter, input); X2: Month (1 to 12, input); X3: Year (0:2011, 1:2012, input); X4: Weather day is holiday or not (input); X5: Day of the week (input); X6: Working day if day is neither weekend nor holiday is 1, otherwise is 0 (input); X7: Weather condition (1: Clear, Few clouds, partly cloudy; 2: Mist + Cloudy, Mist + Broken clouds, Mist + Few clouds, Mist; 3: Light Snow, Light Rain + Thunderstorm + Scattered clouds, Light Rain + Scattered clouds; 4: Heavy Rain + Ice Pallets + Thunderstorm + Mist, Snow + Fog, input); X8: Normalized temperature in Celsius. The values are divided to 41 (max) (input); X9: Normalized feeling temperature in Celsius. The values are divided to 50 (max) (input); X10: Normalized humidity. The values are divided to 100 (max) (input); X11: Normalized wind speed. The values are divided to 67 (max) (input); and Y: Count of total rental bikes including both casual and registered (output).
In this study, to calculate MAPE, the output was normalized and 0.1 was added to prevent a zero value.
y i = y i y m i n y m a x y m i n + 0.1
where y i , y m i n , and y m a x are actual value, minimum and maximum of actual value, respectively.

7.2.4. Performance Results

Table 20 presents the performance results of all models for the three additional datasets. Using the same dataset in the experiment No. 2, Maurya and Gupta (2015) [12] trained ANN models with the maximum R-learn and R-test values of 0.98146 and 0.98662, respectively. Meanwhile, the iML identifies the ANN single model as the best model with R-learn and R-test values of 0.99990 and 0.99629, respectively. The iML gives out a slightly better model than those of the previous research in this numerical experiment.
In the experiment No. 3, Ferreira et al. (2016) had an analytical result of MAPE 3.45% and iML confirms ANN single model as the best model, with MAPE values for learning and test of 0.023% and 0.093%, respectively [13]. The stacking ANN ensemble also performs well with the MAPEs for the learning and test data by 0.026% and 0.010%, respectively.
Finally, in the experiment No. 4, iML achieves R-learn and R-test values of 0.97660 and 0.94790, with bagging ANN as the best model. In contrast, Fanaee-T and Gama (2014) obtained a maximum R value of 0.91990 [11].
As shown in the above numerical experiments, iML trains and identifies the best models which are better than those in the previous studies.

8. Conclusions and Future Work

This study develops an iML platform to efficiently operate data-mining techniques. The iML is designed to be user-friendly, so users can get the results with only “One-Click”. The numerical experiments have demonstrated that iML is a powerful soft computing to identify the best prediction model by automating comparison among diverse machine learning techniques.
To benchmark the effectiveness of iML with WEKA, five datasets collected from the UCI Machine Learning Repository were analyzed via hold-out validation and tenfold cross validation. The performance results indicate that iML can find a more accurate model than that of WEKA in the publicly available datasets. The best prediction model identified by iML is also the best model among all the models trained by iML and WEKA. Notably, iML requires minimal effort from the users to build single, voting, bagging, and stacking models in comparison with WEKA.
Four industrial experiments were carried out to validate the performance of iML. The first experiment involved training a model for prediction of ERP development effort, in which iML yielded an RMSE for learning data with 70.28 h and for testing data with 65.58 h, by using the bagging ANN ensemble (best model). In contrast, Chou et al. (2012) [1] obtained training and testing RMSE values of 234.0157 h and 97.2667 h, respectively.
In the second experiment on performance of CPU processors, iML yielded 0.99990 for R-learning and 0.99629 for R-testing, which are better than those reported in Maurya and Gupta (2015) [12], and confirmed that single ANN was the best model. In the third experiment of daily demand forecasting orders, iML achieved MAPE values of 0.026% (learning) and 0.010% (testing). The results are as excellent as those obtained in Ferreira et al. (2016) [13]. In the fourth experiment for total hourly-shared bike rental, R-learning and R-testing values of 0.97660 and 0.94790 were reached using iML. The test performance was 6% better than that obtained by Fanaee-T and Gama (2014) [11]. In addition to the enhanced prediction performance, the iML possesses ability to determine the best models on the basis of multiple evaluation metrics.
In conclusion, the iML is a powerful and promising prediction platform for solving diverse engineering problems. Since the iML platform can only deal with regression problems, future research should upgrade iML for solving complex classification and time series problems by automatically presenting the alternative models for practical use in engineering applications, as well as adding some other advanced ML methods (such as deep learning models). Moreover, metaheuristic optimization algorithms could be integrated with the iML to help the users finetune the hyperparameters of chosen machine learning models.

Author Contributions

Conceptualization, J.-S.C.; data curation, D.-N.T.; formal analysis, J.-S.C. and D.-N.T.; funding acquisition, J.-S.C.; investigation, J.-S.C., D.-N.T. and C.-F.T.; methodology, J.-S.C. and C.-F.T.; project administration, J.-S.C.; resources, J.-S.C. and C.-F.T.; software, D.-N.T.; supervision, J.-S.C.; validation, J.-S.C., D.-N.T. and C.-F.T.; visualization, J.-S.C. and D.-N.T.; writing—original draft, J.-S.C., D.-N.T. and C.-F.T.; writing—review and editing, J.-S.C. and D.-N.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Ministry of Science and Technology, Taiwan, under grants 108-2221-E-011-003-MY3 and 107-2221-E-011-035-MY3.

Data Availability Statement

The data that support the findings of this study are available from the UCI Machine Learning Repository or corresponding author upon reasonable request.

Acknowledgments

The authors would like to thank the Ministry of Science and Technology, Taiwan, for financially supporting this research.

Conflicts of Interest

The authors declare that they have no conflict of interest.

References

  1. Chou, J.-S.; Cheng, M.-Y.; Wu, Y.-W.; Wu, C.-C. Forecasting enterprise resource planning software effort using evolutionary support vector machine inference model. Int. J. Proj. Manag. 2012, 30, 967–977. [Google Scholar] [CrossRef]
  2. Pham, A.-D.; Ngo, N.-T.; Nguyen, Q.-T.; Truong, N.-S. Hybrid machine learning for predicting strength of sustainable concrete. Soft Comput. 2020. [Google Scholar] [CrossRef]
  3. Cheng, M.-Y.; Chou, J.-S.; Cao, M.-T. Nature-inspired metaheuristic multivariate adaptive regression splines for predicting refrigeration system performance. Soft Comput. 2015, 21, 477–489. [Google Scholar] [CrossRef]
  4. Li, Y.; Lei, G.; Bramerdorfer, G.; Peng, S.; Sun, X.; Zhu, J. Machine Learning for Design Optimization of Electromagnetic Devices: Recent Developments and Future Directions. Appl. Sci. 2021, 11, 1627. [Google Scholar] [CrossRef]
  5. Piersanti, S.; Orlandi, A.; Paulis, F.d. Electromagnetic Absorbing Materials Design by Optimization Using a Machine Learning Approach. IEEE Trans. Electromagn. Compat. 2018, 1–8. [Google Scholar] [CrossRef]
  6. Chou, J.S.; Pham, A.D. Smart artificial firefly colony algorithm-based support vector regression for enhanced forecasting in civil engineering. Comput.-Aided Civ. Infrastruct. Eng. 2015, 30, 715–732. [Google Scholar] [CrossRef]
  7. Cheng, M.-Y.; Prayogo, D.; Wu, Y.-W. A self-tuning least squares support vector machine for estimating the pavement rutting behavior of asphalt mixtures. Soft Comput. 2019, 23, 7755–7768. [Google Scholar] [CrossRef]
  8. Al-Ali, H.; Cuzzocrea, A.; Damiani, E.; Mizouni, R.; Tello, G. A composite machine-learning-based framework for supporting low-level event logs to high-level business process model activities mappings enhanced by flexible BPMN model translation. Soft Comput. 2019. [Google Scholar] [CrossRef]
  9. López, J.; Maldonado, S.; Carrasco, M. A novel multi-class SVM model using second-order cone constraints. Appl. Intell. 2016, 44, 457–469. [Google Scholar] [CrossRef]
  10. Bogawar, P.S.; Bhoyar, K.K. An improved multiclass support vector machine classifier using reduced hyper-plane with skewed binary tree. Appl. Intell. 2018, 48, 4382–4391. [Google Scholar] [CrossRef]
  11. Fanaee-T, H.; Gama, J. Event labeling combining ensemble detectors and background knowledge. Prog. Artif. Intell. 2014, 2, 113–127. [Google Scholar] [CrossRef] [Green Version]
  12. Maurya, V.; Gupta, S.C. Comparative Analysis of Processors Performance Using ANN. In Proceedings of the 2015 5th International Conference on IT Convergence and Security (ICITCS), Kuala Lumpur, Malaysia, 24–27 August 2015; pp. 1–5. [Google Scholar]
  13. Ferreira, R.P.; Martiniano, A.; Ferreira, A.; Ferreira, A.; Sassi, R.J. Study on Daily Demand Forecasting Orders using Artificial Neural Network. IEEE Lat. Am. Trans. 2016, 14, 1519–1525. [Google Scholar] [CrossRef]
  14. De’ath, G.; Fabricius, K.E. Classification and regression trees: A powerful yet simple technique for ecological data analysis. Ecology 2000, 81, 3178–3192. [Google Scholar] [CrossRef]
  15. Li, H.; Wen, G. Modeling reverse thinking for machine learning. Soft Comput. 2020, 24, 1483–1496. [Google Scholar] [CrossRef] [Green Version]
  16. Li, Y. Predicting materials properties and behavior using classification and regression trees. Mater. Sci. Eng. A 2006, 433, 261–268. [Google Scholar] [CrossRef]
  17. Chou, J.-S.; Yang, K.-H.; Lin, J.-Y. Peak Shear Strength of Discrete Fiber-Reinforced Soils Computed by Machine Learning and Metaensemble Methods. J. Comput. Civ. Eng. 2016, 30, 04016036. [Google Scholar] [CrossRef] [Green Version]
  18. Qi, C.; Tang, X. Slope stability prediction using integrated metaheuristic and machine learning approaches: A comparative study. Comput. Ind. Eng. 2018, 118, 112–122. [Google Scholar] [CrossRef]
  19. Chou, J.-S.; Ho, C.-C.; Hoang, H.-S. Determining quality of water in reservoir using machine learning. Ecol. Inform. 2018, 44, 57–75. [Google Scholar] [CrossRef]
  20. Chou, J.-S.; Bui, D.-K. Modeling heating and cooling loads by artificial intelligence for energy-efficient building design. Energy Build. 2014, 82, 437–446. [Google Scholar] [CrossRef]
  21. Alkahtani, M.; Choudhary, A.; De, A.; Harding, J.A. A decision support system based on ontology and data mining to improve design using warranty data. Comput. Ind. Eng. 2018. [Google Scholar] [CrossRef] [Green Version]
  22. Daras, G.; Agard, B.; Penz, B. A spatial data pre-processing tool to improve the quality of the analysis and to reduce preparation duration. Comput. Ind. Eng. 2018, 119, 219–232. [Google Scholar] [CrossRef]
  23. Chou, J.-S.; Tsai, C.-F. Preliminary cost estimates for thin-film transistor liquid–crystal display inspection and repair equipment: A hybrid hierarchical approach. Comput. Ind. Eng. 2012, 62, 661–669. [Google Scholar] [CrossRef]
  24. Chen, T. An ANN approach for modeling the multisource yield learning process with semiconductor manufacturing as an example. Comput. Ind. Eng. 2017, 103, 98–104. [Google Scholar] [CrossRef]
  25. Wu, X.; Kumar, V.; Quinlan, J.R.; Ghosh, J.; Yang, Q.; Motoda, H.; McLachlan, G.J.; Ng, A.; Liu, B.; Philip, S.Y. Top 10 algorithms in data mining. Knowl. Inf. Syst. 2008, 14, 1–37. [Google Scholar] [CrossRef] [Green Version]
  26. Chou, J.-S.; Ngo, N.-T.; Chong, W.K. The use of artificial intelligence combiners for modeling steel pitting risk and corrosion rate. Eng. Appl. Artif. Intell. 2017, 65, 471–483. [Google Scholar] [CrossRef]
  27. Das, D.; Pratihar, D.K.; Roy, G.G.; Pal, A.R. Phenomenological model-based study on electron beam welding process, and input-output modeling using neural networks trained by back-propagation algorithm, genetic algorithms, particle swarm optimization algorithm and bat algorithm. Appl. Intell. 2018, 48, 2698–2718. [Google Scholar] [CrossRef]
  28. Tewari, S.; Dwivedi, U.D. Ensemble-based big data analytics of lithofacies for automatic development of petroleum reservoirs. Comput. Ind. Eng. 2018. [Google Scholar] [CrossRef]
  29. Priore, P.; Ponte, B.; Puente, J.; Gómez, A. Learning-based scheduling of flexible manufacturing systems using ensemble methods. Comput. Ind. Eng. 2018, 126, 282–291. [Google Scholar] [CrossRef] [Green Version]
  30. Fang, K.; Jiang, Y.; Song, M. Customer profitability forecasting using Big Data analytics: A case study of the insurance industry. Comput. Ind. Eng. 2016, 101, 554–564. [Google Scholar] [CrossRef]
  31. Chou, J.-S. Generalized linear model-based expert system for estimating the cost of transportation projects. Expert Syst. Appl. 2009, 36, 4253–4267. [Google Scholar] [CrossRef]
  32. Dandikas, V.; Heuwinkel, H.; Lichti, F.; Drewes, J.E.; Koch, K. Predicting methane yield by linear regression models: A validation study for grassland biomass. Bioresour. Technol. 2018, 265, 372–379. [Google Scholar] [CrossRef] [PubMed]
  33. Ngo, S.H.; Kemény, S.; Deák, A. Performance of the ridge regression method as applied to complex linear and nonlinear models. Chemom. Intell. Lab. Syst. 2003, 67, 69–78. [Google Scholar] [CrossRef]
  34. Sentas, P.; Angelis, L. Categorical missing data imputation for software cost estimation by multinomial logistic regression. J. Syst. Softw. 2006, 79, 404–414. [Google Scholar] [CrossRef]
  35. Slowik, A. Application of an Adaptive Differential Evolution Algorithm With Multiple Trial Vectors to Artificial Neural Network Training. IEEE Trans. Ind. Electron. 2011, 58, 3160–3167. [Google Scholar] [CrossRef]
  36. Caputo, A.C.; Pelagagge, P.M. Parametric and neural methods for cost estimation of process vessels. Int. J. Prod. Econ. 2008, 112, 934–954. [Google Scholar] [CrossRef]
  37. Rocabruno-Valdés, C.I.; Ramírez-Verduzco, L.F.; Hernández, J.A. Artificial neural network models to predict density, dynamic viscosity, and cetane number of biodiesel. Fuel 2015, 147, 9–17. [Google Scholar] [CrossRef]
  38. Ganesan, P.; Rajakarunakaran, S.; Thirugnanasambandam, M.; Devaraj, D. Artificial neural network model to predict the diesel electric generator performance and exhaust emissions. Energy 2015, 83, 115–124. [Google Scholar] [CrossRef]
  39. Vapnik, V.N. An overview of statistical learning theory. IEEE Trans. Neural Netw. 1999, 10, 988–999. [Google Scholar] [CrossRef] [Green Version]
  40. Vapnik, V. The Nature of Statistical Learning Theory, 2nd ed.; Springer: New York, NY, USA, 2013. [Google Scholar]
  41. Jing, G.; Cai, W.; Chen, H.; Zhai, D.; Cui, C.; Yin, X. An air balancing method using support vector machine for a ventilation system. Build. Environ. 2018, 143, 487–495. [Google Scholar] [CrossRef]
  42. García-Floriano, A.; López-Martín, C.; Yáñez-Márquez, C.; Abran, A. Support vector regression for predicting software enhancement effort. Inf. Softw. Technol. 2018, 97, 99–109. [Google Scholar] [CrossRef]
  43. Breiman, L.; Friedman, J.; Stone, C.J.; Olshen, R.A. Classification and Regression Trees; Routledge: New York, NY, USA, 2017; p. 368. [Google Scholar] [CrossRef]
  44. Choi, S.Y.; Seo, I.W. Prediction of fecal coliform using logistic regression and tree-based classification models in the North Han River, South Korea. J. Hydro-Environ. Res. 2018, 21, 96–108. [Google Scholar] [CrossRef]
  45. Ru, F.; Yin, A.; Jin, J.; Zhang, X.; Yang, X.; Zhang, M.; Gao, C. Prediction of cadmium enrichment in reclaimed coastal soils by classification and regression tree. Estuar. Coast. Shelf Sci. 2016, 177, 1–7. [Google Scholar] [CrossRef]
  46. Chou, J.-S.; Tsai, C.-F.; Pham, A.-D.; Lu, Y.-H. Machine learning in concrete strength simulations: Multi-nation data analytics. Constr. Build. Mater. 2014, 73, 771–780. [Google Scholar] [CrossRef]
  47. Elish, M.O. Assessment of voting ensemble for estimating software development effort. In Proceedings of the 2013 IEEE Symposium on Computational Intelligence and Data Mining (CIDM), Singapore, 16–19 April 2013; pp. 316–321. [Google Scholar]
  48. Wang, Z.; Wang, Y.; Srinivasan, R.S. A novel ensemble learning approach to support building energy use prediction. Energy Build. 2018, 159, 109–122. [Google Scholar] [CrossRef]
  49. Chen, J.; Yin, J.; Zang, L.; Zhang, T.; Zhao, M. Stacking machine learning model for estimating hourly PM2.5 in China based on Himawari 8 aerosol optical depth data. Sci. Total Environ. 2019, 697, 134021. [Google Scholar] [CrossRef] [PubMed]
  50. Basant, N.; Gupta, S.; Singh, K.P. A three-tier QSAR modeling strategy for estimating eye irritation potential of diverse chemicals in rabbit for regulatory purposes. Regul. Toxicol. Pharmacol. 2016, 77, 282–291. [Google Scholar] [CrossRef]
  51. Lee, K.M.; Yoo, J.; Kim, S.-W.; Lee, J.-H.; Hong, J. Autonomic machine learning platform. Int. J. Inf. Manag. 2019, 49, 491–501. [Google Scholar] [CrossRef]
  52. Wolpert, D.H.; Macready, W.G. No free lunch theorems for optimization. IEEE Trans. Evol. Comput. 1997, 1, 67–82. [Google Scholar] [CrossRef] [Green Version]
  53. Wolpert, D.H.; Macready, W.G. No Free Lunch Theorems for Search; Technical Report SFI-TR-95-02-010; Santa Fe Institute: Santa Fe, NM, USA, 1995. [Google Scholar]
  54. Cheng, D.; Shi, Y.; Gwee, B.; Toh, K.; Lin, T. A Hierarchical Multiclassifier System for Automated Analysis of Delayered IC Images. IEEE Intell. Syst. 2019, 34, 36–43. [Google Scholar] [CrossRef]
  55. Basheer, I.A.; Hajmeer, M. Artificial neural networks: Fundamentals, computing, design, and application. J. Microbiol. Methods 2000, 43, 3–31. [Google Scholar] [CrossRef]
  56. Jain, A.K.; Jianchang, M.; Mohiuddin, K.M. Artificial neural networks: A tutorial. Computer 1996, 29, 31–44. [Google Scholar] [CrossRef] [Green Version]
  57. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  58. Chamasemani, F.F.; Singh, Y.P. Multi-class Support Vector Machine (SVM) Classifiers—An Application in Hypothyroid Detection and Classification. In Proceedings of the 2011 Sixth International Conference on Bio-Inspired Computing: Theories and Applications, Penang, Malaysia, 27–29 September 2011; pp. 351–356. [Google Scholar]
  59. Yang, X.; Yu, Q.; He, L.; Guo, T. The one-against-all partition based binary tree support vector machine algorithms for multi-class classification. Neurocomputing 2013, 113, 1–7. [Google Scholar] [CrossRef]
  60. Tuv, E.; Runger, G.C. Scoring levels of categorical variables with heterogeneous data. IEEE Intell. Syst. 2004, 19, 14–19. [Google Scholar] [CrossRef]
  61. Chiang, W.; Liu, X.; Zhang, T.; Yang, B. A Study of Exact Ridge Regression for Big Data. In Proceedings of the 2018 IEEE International Conference on Big Data (Big Data), Seattle, WA, USA, 10–13 December 2018; pp. 3821–3830. [Google Scholar]
  62. Marquardt, D.W.; Snee, R.D. Ridge Regression in Practice. Am. Stat. 1975, 29, 3–20. [Google Scholar] [CrossRef]
  63. Cox, D.R. The regression analysis of binary sequences. J. R. Stat. Society. Ser. B 1958, 20, 215–242. [Google Scholar] [CrossRef]
  64. Jiang, F.; Guan, Z.; Li, Z.; Wang, X. A method of predicting visual detectability of low-velocity impact damage in composite structures based on logistic regression model. Chin. J. Aeronaut. 2021, 34, 296–308. [Google Scholar] [CrossRef]
  65. Kohavi, R. A study of cross-validation and bootstrap for accuracy estimation and model selection. In Proceedings of the International Joint Conference on Artificial Intelligence 1995, Montreal, QC, Canada, 20–25 August 1995; pp. 1137–1143. [Google Scholar]
  66. Chou, J.; Truong, D.; Le, T. Interval Forecasting of Financial Time Series by Accelerated Particle Swarm-Optimized Multi-Output Machine Learning System. IEEE Access 2020, 8, 14798–14808. [Google Scholar] [CrossRef]
  67. Yeh, I.-C. Analysis of Strength of Concrete Using Design of Experiments and Neural Networks. J. Mater. Civ. Eng. 2006, 18, 597–604. [Google Scholar] [CrossRef]
  68. Yeh, I.C.; Hsu, T.-K. Building real estate valuation models with comparative approach through case-based reasoning. Appl. Soft Comput. 2018, 65, 260–271. [Google Scholar] [CrossRef]
  69. Tsanas, A.; Xifara, A. Accurate quantitative estimation of energy performance of residential buildings using statistical machine learning tools. Energy Build. 2012, 49, 560–567. [Google Scholar] [CrossRef]
  70. Lau, K.; López, R. A Neural Networks Approach to Aerofoil Noise Prediction; International Center for Numerical Methods in Engineering: Barcelona, Spain, 2009. [Google Scholar]
Figure 1. Artificial neural network (ANN) model.
Figure 1. Artificial neural network (ANN) model.
Mathematics 09 00686 g001
Figure 2. Support Vector Machine (SVM) and Support Vector Regression (SVR) models.
Figure 2. Support Vector Machine (SVM) and Support Vector Regression (SVR) models.
Mathematics 09 00686 g002
Figure 3. The classification and regression tree (CART) model.
Figure 3. The classification and regression tree (CART) model.
Mathematics 09 00686 g003
Figure 4. Linear Ridge Regression (LRR) and Logistic Regression (LgR) models.
Figure 4. Linear Ridge Regression (LRR) and Logistic Regression (LgR) models.
Mathematics 09 00686 g004
Figure 5. Ensemble models.
Figure 5. Ensemble models.
Mathematics 09 00686 g005
Figure 6. K-fold cross-validation method.
Figure 6. K-fold cross-validation method.
Mathematics 09 00686 g006
Figure 7. Intelligent machine leaner framework.
Figure 7. Intelligent machine leaner framework.
Mathematics 09 00686 g007
Figure 8. Snapshot of intelligent Machine Learner (iML) interface.
Figure 8. Snapshot of intelligent Machine Learner (iML) interface.
Mathematics 09 00686 g008
Figure 9. Snapshot of report file.
Figure 9. Snapshot of report file.
Mathematics 09 00686 g009
Figure 10. Root mean square errors of best models.
Figure 10. Root mean square errors of best models.
Mathematics 09 00686 g010
Table 1. Confusion matrix.
Table 1. Confusion matrix.
Actual Class
PositiveNegative
Predicted classPositiveTrue positiveFalse Negative
NegativeFalse positiveTrue negative
Table 2. Mathematical formulas for performance measures.
Table 2. Mathematical formulas for performance measures.
MeasureFormulaMeasureFormula
Accuracy Accuracy = tp + tn tp + fp + tn + fn Mean absolute error MAE = 1 n i = 1 n | y i y i |
Precision Precision = tp tp + fp Mean absolute percentage error MAPE = 1 n | y i y i y i |
Sensitivity Sensitivity = tp tp + fn Root mean square error RMSE = 1 n i = 1 n ( y i y i ) 2
Specificity Specificity = tn tn + fp Correlation coefficient R = n y i . y i ( y i ) ( y i ) n ( y i 2 ) ( y i ) 2 n ( y i 2 ) ( y i ) 2
Area under the curve AUC = 1 2 [ ( tp tp + fn ) + ( tn tn + fp ) ] Total error rate TER = | i = 1 n y i i = 1 n y i | i = 1 n y i
tp is the true positives (number of correctly recognized class examples); tn is the true negatives (number of correctly recognized examples that do not belong to the class); fp is the number of false positives (number of examples that were incorrectly assigned to a class); fn is the number of false negatives (number of examples that were not assigned to a class); y i is actual value; y i is predicted value; n is sample size.
Table 3. Characteristic of data from UCI Machine Learning Repository.
Table 3. Characteristic of data from UCI Machine Learning Repository.
UCI Data SetNo. of SamplesNo. of AttributesOutput Information
Concrete Compressive Strength (Yeh (2006) [67])10308Concrete compressive strength (MPa)
Real estate valuation (Yeh and Hsu (2018) [68])4146Y = house price of unit area (10,000 New Taiwan Dollar/Ping, where Ping is a local unit, 1 Ping = 3.3 m squared)
Energy efficiency (Tsanas and Xifara (2012) [69])7688y1 Heating Load (kW)
y2 Cooling Load (kW)
Airfoil Self-Noise (Lau and López (2009) [70])15035Scaled sound pressure level (dB).
Table 4. Test results by WEKA and iML on concrete compressive strength dataset via hold-out validation.
Table 4. Test results by WEKA and iML on concrete compressive strength dataset via hold-out validation.
ModelWEKASI (Ranking)iMLSI (Ranking)
RRMSE (MPa)MAE (MPa)MAPE (%)RRMSE (MPa)MAE (MPa)MAPE (%)
I. SingleCART ANN
0.9276.5465.17018.7700.142 (7)0.9465.3023.72812.6730.023 (3)
II. VotingANN + CART ANN + CART
0.9366.2024.93019.0900.124 (6)0.9564.7713.55012.7230.000 (1)
III. BaggingCART ANN
0.9605.0443.98315.1300.032 (4)0.9515.0563.64712.2490.010 (2)
IV. Stacking(*) CART (*) LRR
0.9395.9864.79217.5200.104 (5)0.44414.82911.77956.7751.000 (8)
Note: (*) is (ANN + SVR + CART + LRR); bold value denotes the best overall performance.
Table 5. Test results by WEKA and iML on real estate dataset via hold-out validation.
Table 5. Test results by WEKA and iML on real estate dataset via hold-out validation.
ModelWEKASI (Ranking)iMLSI (Ranking)
RRMSE (U)MAE (U)MAPE (%)RRMSE (U)MAE (U)MAPE (%)
I. SingleCART ANN
0.74010.7625.88213.2100.321 (6)0.8716.6304.91213.5910.049 (3)
II. VotingANN + CART + LRR ANN + CART
0.74511.0545.90812.7800.327 (7)0.8776.6154.73912.8670.030 (2)
III. BaggingCART CART
0.77010.3215.28111.7600.246 (4)0.8846.4854.38112.3050.000 (1)
IV. Stacking(*) CART (*) ANN
0.74410.7485.77412.9400.311 (5)0.39112.63810.41133.8071.000 (8)
Note: (*) is (ANN + SVR + CART + LRR); bold value denotes the best overall performance; U: house price of unit area (10,000 New Taiwan Dollar/Ping, where Ping is a local unit, 1 Ping = 3.3 m squared).
Table 6. Test results by WEKA and iML on energy efficiency data set (Heating load) via hold-out validation.
Table 6. Test results by WEKA and iML on energy efficiency data set (Heating load) via hold-out validation.
ModelWEKASI (Ranking)iMLSI (Ranking)
RRMSE (kW)MAE (kW)MAPE (%)RRMSE (kW)MAE (kW)MAPE (%)
I. SingleCART ANN
0.9960.9140.6463.3000.418 (6)0.9990.4880.3541.7000.046 (3)
II. VotingANN + CART ANN + CART
0.9960.9290.7293.8200.449 (7)0.9990.4950.3361.6170.045 (2)
III. BaggingCART ANN
0.9970.8700.6193.2100.354 (5)0.9990.4260.3111.5190.000 (1)
IV. Stacking(*) CART (*) LRR
0.9980.7540.5242.4800.231 (4)0.9983.4543.22617.6581.000 (8)
Note: (*) is (ANN + SVR + CART + LRR); bold value denotes the best overall performance.
Table 7. Test results by WEKA and iML on energy efficiency dataset (Cooling load) via hold-out validation.
Table 7. Test results by WEKA and iML on energy efficiency dataset (Cooling load) via hold-out validation.
ModelWEKASI (Ranking)iMLSI (Ranking)
RRMSE (kW)MAE (kW)MAPE (%)RRMSE (kW)MAE (kW)MAPE (%)
I. SingleCART ANN
0.9861.5241.0063.9000.320 (6)0.9921.2310.8843.5770.038 (2)
II. VotingANN + CART ANN + CART
0.9871.5041.0644.3300.293 (4)0.9881.5090.9823.5440.222 (3)
III. BaggingCART ANN
0.9861.5651.0464.0300.358 (7)0.9931.1770.8093.1650.000 (1)
IV. Stacking(*) SVR (*) LRR
0.9861.5370.9793.7000.314 (5)0.9894.2903.76217.3051.000 (8)
Note: (*) is (ANN + SVR + CART + LRR); bold value denotes the best overall performance.
Table 8. Test results by WEKA and iML on airfoil self-noise dataset via hold-out validation.
Table 8. Test results by WEKA and iML on airfoil self-noise dataset via hold-out validation.
ModelWEKASI (Ranking)iMLSI (Ranking)
RRMSE (dB)MAE (dB)MAPE (%)RRMSE (dB)MAE (dB)MAPE (%)
I. SingleCART ANN
0.8983.1852.3391.8800.502 (6)0.9532.1491.5771.2590.044 (2)
II. VotingANN + CART ANN + CART
0.8933.4712.6492.1000.591 (7)0.9522.1631.6331.3010.058 (3)
III. BaggingCART ANN
0.9222.9022.1351.7100.332 (4)0.9582.0311.4941.1940.000 (1)
IV. Stacking(*) CART (*) LRR
0.9053.0822.2711.8200.450 (5)0.9527.0505.6484.6131.000 (8)
Note: (*) is (ANN + SVR + CART + LRR); bold value denotes the best overall performance.
Table 9. Performance of WEKA and iML on concrete compressive strength dataset via tenfold cross-validation.
Table 9. Performance of WEKA and iML on concrete compressive strength dataset via tenfold cross-validation.
ModelWEKASI (Ranking)iMLSI (Ranking)
RRMSE (MPa)MAE (MPa)MAPE (%)RRMSE (MPa)MAE (MPa)MAPE (%)
I. SingleCART ANN
0.9236.4344.81015.5100.228 (5)0.9465.4114.00313.8660.154 (3)
II. VotingANN + CART ANN + CART
0.9176.8235.21317.2300.265 (7)0.9554.9033.50612.3970.111 (2)
III. BaggingCART CART
0.9326.0824.59815.0300.205 (4)0.9803.3592.4328.3560.000 (1)
IV. Stacking(*) SVR (*) ANN
0.9246.4364.85215.5300.229 (6)0.61314.38110.86744.7591.000 (8)
Note: (*) is (ANN + SVR + CART + LRR); bold value denotes the best overall performance.
Table 10. Performance of WEKA and iML on real estate valuation dataset via tenfold cross-validation.
Table 10. Performance of WEKA and iML on real estate valuation dataset via tenfold cross-validation.
ModelWEKASI (Ranking)iMLSI (Ranking)
RRMSE (U)MAE (U)MAPE (%)RRMSE (U)MAE (U)MAPE (%)
I. SingleCART ANN
0.8078.0215.19715.2700.314 (5)0.8138.0115.38814.9910.315 (6)
II. VotingSVR + CART ANN + CART + LRR
0.8058.0915.19815.0900.315 (7)0.8217.8785.37615.1160.308 (4)
III. BaggingCART CART
0.8287.6375.01714.9300.280 (2)0.9254.7743.2018.9740.000 (1)
IV. Stacking(*) SVR (*) ANN
0.8197.8234.96914.4400.284 (3)0.43212.3099.52632.2671.000 (8)
Note: (*) is (ANN + SVR + CART + LRR); bold value denotes the best overall performance; U: house price of unit area (10,000 New Taiwan Dollar/Ping, where Ping is a local unit, 1 Ping = 3.3 m squared).
Table 11. Performance of WEKA and iML on energy efficiency dataset (Heating load) via tenfold cross-validation.
Table 11. Performance of WEKA and iML on energy efficiency dataset (Heating load) via tenfold cross-validation.
ModelWEKASI (Ranking)iMLSI (Ranking)
RRMSE (kW)MAE (kW)MAPE (%)RRMSE (kW)MAE (kW)MAPE (%)
I. SingleCART ANN
0.9951.0460.7123.2000.459 (7)0.9990.4840.3601.7220.049 (2)
II. VotingANN + CART ANN + CART
0.9970.8530.6413.1900.309 (4)0.9990.4970.3521.6020.053 (3)
III. BaggingCART ANN
0.9970.9150.6332.8900.324 (5)0.9990.3840.2911.4090.000 (1)
IV. Stacking(*) SVR (*) LRR
0.9960.8720.6392.9900.337 (6)0.9983.5223.22618.1811.000 (8)
Note: (*) is (ANN + SVR + CART + LRR); bold value denotes the best overall performance.
Table 12. Performance of WEKA and iML on energy efficiency dataset (Cooling load) via tenfold cross-validation.
Table 12. Performance of WEKA and iML on energy efficiency dataset (Cooling load) via tenfold cross-validation.
ModelWEKASI (Ranking)iMLSI (Ranking)
RRMSE (kW)MAE (kW)MAPE (%)RRMSE (kW)MAE (kW)MAPE (%)
I. SingleCART ANN
0.9821.8121.1834.1600.460 (5)0.9931.1400.7993.1610.150 (2)
II. VotingANN + CART ANN + CART
0.9821.8311.2764.7700.491 (7)0.9891.4150.9003.2060.250 (3)
III. BaggingCART ANN
0.9831.7851.1604.0700.444 (4)0.9970.8080.5562.1290.000 (1)
IV. Stacking(*) SVR (*) LRR
0.9821.8271.1954.2100.465 (6)0.9894.1083.61917.2531.000 (8)
Note: (*) is (ANN + SVR + CART + LRR); bold value denotes the best overall performance.
Table 13. Performance of WEKA and iML on airfoil self-noise dataset via tenfold cross-validation.
Table 13. Performance of WEKA and iML on airfoil self-noise dataset via tenfold cross-validation.
ModelWEKASI (Ranking)iMLSI (Ranking)
RRMSE (dB)MAE (dB)MAPE (%)RRMSE (dB)MAE (dB)MAPE (%)
I. SingleCART ANN
0.8773.3142.3811.9100.497 (5)0.9462.2391.6601.3310.152 (2)
II. VotingANN + CART ANN + CART
0.8513.6852.7472.2200.641 (7)0.9462.2461.6641.3340.152 (3)
III. BaggingCART CART
0.9112.9062.1601.7300.352 (4)0.9711.7271.2711.0230.000 (1)
IV. Stacking(*) LRR (*) LRR
0.8743.3742.4941.9900.525 (6)0.9466.8945.5874.5621.000 (8)
Note: (*) is (ANN + SVR + CART + LRR); bold value denotes the best overall performance.
Table 14. Parameter setting.
Table 14. Parameter setting.
ExperimentModelANNSVR and SVMLRR and LgRCART
Number of
Hidden Node
CSigmaEpsilonLambdaMin Leaf
1Single regression model30 7.3 × 10 6 45.67 1.0 × 10 5 1.0 × 10 8 1
Single classification model3041,7033.67- 1.0 × 10 5 1
Ensemble regression model30 7.3 × 10 6 45.67 1.0 × 10 5 1.0 × 10 8 1
2Single regression model30 7.3 × 10 6 20.03 1.0 × 10 5 1.0 × 10 8 1
Single classification model3042003.40- 1.0 × 10 5 1
Ensemble regression model30 7.3 × 10 6 45.67 1.0 × 10 5 1.0 × 10 8 1
3Single regression model30 7.3 × 10 6 30.00 1.0 × 10 5 1.0 × 10 8 1
Single classification model2041,7033.67- 1.0 × 10 5 1
Ensemble regression model30 7.3 × 10 6 30.00 1.0 × 10 5 1.0 × 10 8 1
4Single regression model15 7.3 × 10 6 45.67 1.0 × 10 5 1.0 × 10 8 1
Single classification model1541,7033.67- 1.0 × 10 5 1
Ensemble regression model15 7.3 × 10 6 45.67 1.0 × 10 5 1.0 × 10 8 1
Table 15. Variables and descriptive statistics for predicting enterprise resource planning (ERP) software development effort.
Table 15. Variables and descriptive statistics for predicting enterprise resource planning (ERP) software development effort.
VariableMin.Max.MeanStandardData
DeviationType
Y: Software development effort42694258.55394.69Numerical
(person-hour)
X1: Program type entry01Dummy variableBoolean
X2: Program type report01Dummy variableBoolean
X3: Program type batch01Dummy variableBoolean
X4: Program type query01Dummy variableBoolean
X5: Program type transaction00Referential categoryBoolean
X6: Number of programs18816.7319.12Numerical
X7: Number of zooms02028100.22255.40Numerical
X8: Number of columns in form33216397.75548.06Numerical
X9: Number of actions01645288.44339.61Numerical
X10: Number of signature tasks0150.391.77Numerical
X11: Number of batch serial numbers0110.311.50Numerical
X12: Number of multi-angle trade tasks0220.552.66Numerical
X13: Number of multi-unit tasks0211.103.41Numerical
X14: Number of reference calls052813.9649.92Numerical
X15: Number of confirmed tasks0211.503.99Numerical
X16: Number of post tasks0120.231.33Numerical
X17: Number of industry type tasks0210.802.97Numerical
Table 16. Performances of predictive models for ERP software development effort.
Table 16. Performances of predictive models for ERP software development effort.
No.ModelLearnTestSI and Ranking
RMSEMAEMAPERTERRMSEMAEMAPERTERSIlocalSIglobal
ISingle
1ANN68.8124.9219.91%0.981.49%115.2461.8530.65%0.959.86%0.00 (1)0.13 (2)
2SVR0.000.000.00%1.000.00%361.88255.35611.63%Inf36.81%0.89 (4)
3CART86.8940.3819.56%0.970.00%196.48107.0247.29%0.8512.77%0.15 (2)
4LRR250.21221.15617.02%0.8448.77%255.43227.80647.08%0.7863.59%0.72 (3)
IIVoting
1(*)73.3062.90157.96%0.9912.39%185.39139.02321.77%0.9423.90%0.34 (5)
2ANN + CART + LRR97.7483.87210.62%0.9816.52%139.70110.32227.32%0.9422.01%0.21 (2)
3ANN + SVR + CART39.9519.3311.22%0.990.50%181.63117.20215.13%0.9414.45%0.22 (3)
4ANN + SVR + LRR87.6676.90207.61%0.9916.52%202.91158.85419.21%0.9430.90%0.46 (7)
5SVR + CART + LRR93.1780.47208.60%0.9916.26%236.10178.73427.29%0.8932.97%0.60 (10)
6ANN + CART59.9228.9916.83%0.990.75%123.4368.4633.90%0.9410.10%0.01 (1)0.15 (3)
7ANN + LRR131.49115.35311.42%0.9724.78%152.57129.46326.55%0.9432.67%0.34 (6)
8ANN + SVR34.4012.469.96%1.000.75%199.24135.42307.89%0.9518.21%0.31 (4)
9SVR + CART43.4520.199.78%0.990.00%254.27166.51320.26%0.8521.09%0.56 (9)
10CART + LRR139.76120.71312.91%0.9624.38%190.91151.58337.64%0.8932.09%0.48 (8)
11SVR + LRR125.10110.57308.51%0.9724.38%287.41230.03626.83%0.7847.93%1.00 (11)
IIIBagging
1ANN70.2833.4821.45%0.982.51%65.5840.5119.50%0.995.59%0.00 (1)0.00 (1)
2SVR174.5692.31231.82%0.913.98%162.3899.0287.79%0.8711.69%0.47 (3)
3CART123.0552.4525.94%0.942.19%127.2179.9720.11%0.968.58%0.21 (2)
4LRR249.00222.81666.29%0.8859.61%309.18257.31269.67%0.7110.90%0.97 (4)
IVStacking
1(*) ANN0.070.020.03%1.000.00%361.54255.19611.55%0.7136.77%0.80 (2)
2(*) SVR0.000.000.00%1.000.00%361.76255.53612.78%NaN36.76%1.00 (4)
3(*) CART52.0017.405.69%0.990.00%360.24252.61593.46%NaN34.86%0.81 (3)
4(*) LRR132.7397.48187.26%0.9523.62%289.20206.32494.18%0.6234.11%0.03 (1)1.00 (5)
VTiering
12-Class (**)317.0471.3824.45%0.5820.48%176.7765.4223.57%0.7916.13%0.00 (1)0.31 (4)
23-Class (***)383.02115.4826.68%0.3036.54%278.46111.7426.94%0.5128.64%0.54 (2)
34-Class (****)414.30151.3229.96%0.1049.11%347.05147.2229.76%0.2643.00%1.00 (3)
Note: (*) is (ANN + SVR + CART + LRR); (**) SVM-(ANN, SVR); (***) CART-(ANN, SVR, SVR); (****) CART-(CART, SVR, SVR, SVR); (No.): Ranking.
Table 17. Descriptive statistics for CPU processors.
Table 17. Descriptive statistics for CPU processors.
Statistic ValueInputOutput
X1X2X3X4X5X6Y
Min17646400015
Max150032,00064,000256521761238
Mean203.822867.9811,796.225.214.718.2799.33
Std.260.263878.7411,726.640.636.8226154.76
Table 18. Variables and descriptive statistics for daily demand forecasting orders.
Table 18. Variables and descriptive statistics for daily demand forecasting orders.
Statistic ValueInputOutput
X1X2X3X4X5X6X7X8X9X10X11X12Y
Min1243.6577.3721.8325.1374.37011,992345216,4117679129.41
Max56435.30223.27118.18267.34302.4586571,772210,508188,41173,839616.45
Mean--172.55118.9252.11109.23139.5377.444,504.446,640.879,401.523,114.6300.87
Std.--69.5127.1718.8350.7441.44186.512,197.945,220.740,504.413,14889.6
Table 19. Variables and descriptive statistics for total hourly-shared bike rental per days.
Table 19. Variables and descriptive statistics for total hourly-shared bike rental per days.
Statistic ValueInputOutput
X1X2X3X4X5X6X7X8X9X10X11Y
Min10100010.0590.07900.02222
Max411216130.8620.8410.9730.5078714
Mean---0.0292.9970.6841.3950.4950.4740.6280.194504.35
Std.---0.1672.0050.4650.5450.1830.1630.1420.0771937.21
Table 20. Performance results of three additional numerical experiments.
Table 20. Performance results of three additional numerical experiments.
No.ModelLearnTest
RMSEMAEMAPERTERRMSEMAEMAPERTER
2Single2.4620.5691.015%1.0000.236%8.7383.6833.775%0.9962.730%
Voting17.5094.9652.808%0.9950.118%13.0875.1735.685%0.9890.921%
Bagging40.1278.8932.360%0.9813.015%13.4843.9302.489%0.9963.795%
Stacking40.4289.0303.389%0.9730.000%64.78243.615104.992%0.8429.852%
Tiering-2class163.33826.9473.229%0.38324.818%18.1965.8224.629%0.9863.220%
Tiering-3class167.09329.7843.856%0.34027.446%76.40612.6985.268%0.63913.094%
Tiering-4class182.57244.7227.970%0.11241.332%91.70121.0898.342%0.42224.368%
3Single0.3490.0800.023%1.0000.021%0.3170.2310.093%1.0000.042%
Voting17.41710.7543.089%0.9850.010%12.16210.1573.993%0.9510.867%
Bagging0.9170.3990.110%1.0000.020%0.2960.2210.087%1.0000.074%
Stacking0.3380.0900.026%1.0000.014%0.3350.2510.101%1.0000.042%
Tiering-2class169.29663.58014.294%−0.39921.483%214.67486.71116.747%−0.70427.688%
Tiering-3class273.303212.04762.384%−0.66471.449%295.065223.20962.186%−0.57071.054%
Tiering-4class329.001312.39797.619%−0.30499.023%51.16445.12218.684%0.70611.339%
4Single0.0460.0306.670%0.97910.450%0.1050.07314.120%0.8830.550%
Voting0.0520.0377.850%0.9748.000%0.0800.05610.750%0.9290.260%
Bagging0.0490.0347.150%0.9776.930%0.0690.0468.870%0.9480.190%
Stacking0.0050.0030.700%1.00021.430%0.2140.16938.680%0.0000.086%
Tiering-2class0.5800.43257.620%−0.58258.900%0.5890.45161.410%−0.57068.290%
Tiering-3class0.6390.56884.420%−0.68064.600%0.6460.56582.110%−0.71790.260%
Tiering-4class0.6480.59692.370%−0.51365.200%0.6520.58487.400%−0.63094.300%
Note: No. 2: CPU experiment dataset; No. 3: Customer experiment dataset; No. 4: Rental bike experiment dataset; the bold denotes the best model in each experiment.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chou, J.-S.; Truong, D.-N.; Tsai, C.-F. Solving Regression Problems with Intelligent Machine Learner for Engineering Informatics. Mathematics 2021, 9, 686. https://doi.org/10.3390/math9060686

AMA Style

Chou J-S, Truong D-N, Tsai C-F. Solving Regression Problems with Intelligent Machine Learner for Engineering Informatics. Mathematics. 2021; 9(6):686. https://doi.org/10.3390/math9060686

Chicago/Turabian Style

Chou, Jui-Sheng, Dinh-Nhat Truong, and Chih-Fong Tsai. 2021. "Solving Regression Problems with Intelligent Machine Learner for Engineering Informatics" Mathematics 9, no. 6: 686. https://doi.org/10.3390/math9060686

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop