Next Article in Journal
Sudden Sensorineural Hearing Loss in the COVID-19 Pandemic: A Systematic Review and Meta-Analysis
Next Article in Special Issue
An Ensemble of Transfer Learning Models for the Prediction of Skin Cancers with Conditional Generative Adversarial Networks
Previous Article in Journal
Pathogenic Copy Number Variations Involved in the Genetic Etiology of Syndromic and Non-Syndromic Intellectual Disability—Data from a Romanian Cohort
Previous Article in Special Issue
A Novel Framework for Abnormal Risk Classification over Fetal Nuchal Translucency Using Adaptive Stochastic Gradient Descent Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Proposed Framework for Early Prediction of Schistosomiasis

1
Department of Computer Engineering, University of Engineering and Technology, Lahore 54890, Pakistan
2
School of Information and Physical Sciences, The University of Newcastle, Newcastle, NSW 2308, Australia
3
Department of Data Science, University of the Punjab, Lahore 54890, Pakistan
4
Department of Computer Science and Information Technology, Virtual University of Pakistan, Lahore 54000, Pakistan
5
Department of ICT and Natural Sciences, Norwegian University of Science and Technology, 7034 Trondheim, Norway
6
Department of Information Systems, College of Computer and Information Science, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
*
Authors to whom correspondence should be addressed.
Diagnostics 2022, 12(12), 3138; https://doi.org/10.3390/diagnostics12123138
Submission received: 7 November 2022 / Revised: 8 December 2022 / Accepted: 8 December 2022 / Published: 12 December 2022
(This article belongs to the Special Issue Deep Disease Detection and Diagnosis Models)

Abstract

:
Schistosomiasis is a neglected tropical disease that continues to be a leading cause of illness and mortality around the globe. The causing parasites are affixed to the skin through defiled water and enter the human body. Failure to diagnose Schistosomiasis can result in various medical complications, such as ascites, portal hypertension, esophageal varices, splenomegaly, and growth retardation. Early prediction and identification of risk factors may aid in treating disease before it becomes incurable. We aimed to create a framework by incorporating the most significant features to predict Schistosomiasis using machine learning techniques. A dataset of advanced Schistosomiasis has been employed containing recovery and death cases. A total data of 4316 individuals containing recovery and death cases were included in this research. The dataset contains demographics, socioeconomic, and clinical factors with lab reports. Data preprocessing techniques (missing values imputation, outlier removal, data normalisation, and data transformation) have also been employed for better results. Feature selection techniques, including correlation-based feature selection, Information gain, gain ratio, ReliefF, and OneR, have been utilised to minimise a large number of features. Data resampling algorithms, including Random undersampling, Random oversampling, Cluster Centroid, Near miss, and SMOTE, are applied to address the data imbalance problem. We applied four machine learning algorithms to construct the model: Gradient Boosting, Light Gradient Boosting, Extreme Gradient Boosting and CatBoost. The performance of the proposed framework has been evaluated based on Accuracy, Precision, Recall and F1-Score. The results of our proposed framework stated that the CatBoost model showed the best performance with the highest accuracy of (87.1%) compared with Gradient Boosting (86%), Light Gradient Boosting (86.7%) and Extreme Gradient Boosting (86.9%). Our proposed framework will assist doctors and healthcare professionals in the early diagnosis of Schistosomiasis.

1. Introduction

Schistosomiasis, snail fever or bilharzia, is one of the most lethal and contagious among several ignored tropical diseases of the world. This parasitic disease affixed to the skin through defiled water enters the human body. It moves into human veins, where parasites lay eggs which becomes the reason for two stages (chronic and acute) of Schistosomiasis [1,2]. The most severe type of late-stage Schistosomiasis japonica in Asia is advanced Schistosomiasis. Hepatosplenomegaly, portal hypertension, gastroesophageal varices, periportal and liver cirrhosis, and other significant liver parenchyma damage are all caused by parasites trapped in eggs [3,4]. Advanced Schistosomiasis patients have a high rate of fatality or disability. World health organisation (WHO) reported that millions of people got infected [5]. There are over 30,000 advanced Schistosomiasis patients in China, and the number of cases has grown in recent years [6].
The quantity and size of medical data are increasing continuously. Because of this enormous volume of information, the prognosis of diseases, analysis based on statistics, extraction of disguised information, and identification of symptoms causing the disease are now possible in the digital era [7,8]. The significant data mining and machine learning techniques can help in the early prediction of Schistosomiasis, which can be very helpful for decision-makers and researchers [9]. However, dealing with a huge volume of data such as images, historical data, and genes-related expressions has been the main problem in data mining and machine learning applications which requires high computational time and storage capability that leads to poor performances. Feature selection plays an important role in solving the issue of huge data. It can be applied to reduce the dimension, irrelevance, and redundant data features, shorten the training time and improve performance [10].
A comprehensive model is not only beneficial to health professionals, but it may also assist people in diagnosing which factors may harm their health. The fundamental objective for early prediction of Schistosomiasis is to use medical data to assist medical practitioners in disease diagnosis, reduce mortality, and improve recovery. Statistical approaches are used in several risk prediction models, including K-nearest neighbour, discriminant analysis, and logistic regression [11]. Classification and regression models were also used to forecast risk variables with the progress of machine learning and artificial intelligence technology [12]. Machine learning and pattern recognition algorithms have demonstrated promising performance for high-dimensional and imbalanced clinical datasets. A dataset can be imbalanced when one class is over-represented among others. This concern is especially fundamental in real-world datasets where it’s expensive to misclassify models from the minority class, such as detection of fake calls, diagnosis of diseases like Schistosomiasis, data recovery, and text categorisation. Machine learning algorithms are a collection of approaches that automatically discover patterns from huge, complex data as an alternative to traditional prediction methods [13,14,15,16].
Data mining approaches were integrated with various critical parameters, such as clinical, socioeconomic and geographical factors, to diagnose Schistosomiasis. Various computational approaches are commonly used to uncover the link between various illnesses and patient clinical characteristics. Furthermore, machine learning techniques to find significant features from complicated datasets demonstrate their value. On the other hand, schistosomiasis pathogenesis is a complex process influenced by various factors such as demography, socioeconomic situation, and healthcare infrastructure. Therefore, an appropriate analytical framework has been required to combine this multidimensional, unstructured, and collinear data. This study aims to create a framework based on machine learning models for early Schistosomiasis prediction. We employed various computational methods, machine learning models and feature selection techniques to predict Schistosomiasis. The main contributions of our work are:
  • A latest and imbalanced dataset related to Schistosomiasis has been employed for this research.
  • Preprocessing like missing values imputation, data transformation, feature selection techniques correlation. Information gain, gain ratio, ReliefF, and OneR minimise the features for prediction and resampling techniques, Random undersampling and oversampling, Cluster Centroid, Near miss, and SMOTE have been employed to carter the problems in the dataset.
  • Advanced machine learning techniques have been employed, which outperformed state-of-the-art methods.
  • Our proposed framework can aid medical and healthcare practitioners in the early identification and improved treatment of Schistosomiasis.
The remainder of this work is structured as follows. Section 2 reviews the literature, including past work on Schistosomiasis and traditional statistical approaches applied to high-dimensional medical data. Section 3 describes the methodology that contains dataset description, data preprocessing, feature selection, resampling, data transformation approaches, and modelling. Section 4 details the framework’s experimental setup and outcomes. Section 5 discusses the ramifications of the results in terms of their relevance. Section 6 finishes the report by highlighting future work in this field.

2. Literature Survey

Li et al. divided Schistosomiasis cases into two groups: favourable and poor prognoses [1]. The cases in which improved patient health are referred to as favourable prognoses, while cases of perishing and death are classified as poor prognoses. Various machine learning models were employed for advanced Schistosomiasis prediction in which ANN outperformed the other models. Holing et al. [17] proposed a method for predicting 1-year unfavourable prognosis for advanced Schistosomiasis. Demographics, clinical factors, medical exams, and test results were used to choose candidate predictors. To build one-year prognostic model, five machine learning techniques were used: Logistic Regression (LR), Decision Tree (DT), Random Forest (RF), Artificial Neural Network (ANN), and Extreme Gradient Boosting (XGBoost). The model’s performance was assessed using the area under the receiver operating characteristic curve (AUROC). XGBoost has outperformed other Machine learning models based on performance.
Computational techniques to estimate the amount of Schistosomiasis susceptibility and vulnerability, the researchers Olanloye et al. [18] compared support vector machine (SVM) models, i.e., Linear, Quadratic, Cubic, Fine Gaussian, and Medium Gaussian and Coarse Gaussian. Receiver Operating Character (ROC) and Parallel Coordinate Plots (PCP) were used to evaluate the experiments in terms of accuracy, processing speed, and execution time. Finally, it was determined that Medium Gaussian is the best among the six models. Asarnow et al. [19] applied the Asarnow-Singh algorithm to a set of images and extracted features by defining a threshold to identify Schistosomiasis mansoni on images with infected foreground and parasitic background areas. Effective results were obtained by using SVM for training and testing images. Data was gathered through a network of wireless sensors, and Kasse et al. [20] developed a system based on IoT monitoring that can help control and predict disease. For disease detection and transmission, multiple data mining algorithms were applied. The SVM detects irregularity better than other models. Chicco et al. [21] collected a dataset of 324 patients, of which 96 were mesothelioma patients. The imbalanced class problem arises because of the ratio between the positive and negative instances. The perceptron-based neural network was used to check the effectiveness of ANN. RF feature selection (RFFS) was used to investigate the most relevant features due to the best diagnosis predictive results.
According to the previous literature, the problem of class imbalance with medical datasets exists, and there are a couple of articles regarding this matter. Subsequently, significant writing looked at clinical science and different fields of science. Different methods to solve class imbalance were described [22]. Various techniques to solve this problem are categorised into different levels: data, algorithmic, cost-sensitive, feature selection and ensemble level. Understanding which preprocessing strategy should be chosen according to the problem is challenging. The dataset is considered to be an imbalanced dataset in which instances between both classes are not equal. Therefore, the class imbalance is mostly removed at the data level, and multiple under and oversampling-based approaches can be used.
Under-sampling was not preferred to deal with class imbalance by some scholars [23] [24], while others do not prefer oversampling [25]. On the other hand, other studies [26] show that undersampling, especially multiple undersampling, performs better. In the dataset, data preprocessing techniques could remove missing values, redundancy, and high scarcity. Hyper-parameters can be optimised to increase the performance of classifiers, or they can be improved by categorising and expelling less related features [27].

3. Methodology

The detail of the proposed framework has been discussed in this section. Firstly, a dataset was obtained and then standardised to remove any biases among the various features. Then, several feature selection approaches are used to extract significant features from high-dimensional data [28,29]. In addition, several resampling approaches have been employed to address the data imbalance problem. After data preprocessing, various machine learning models were used, and the results were compared, as shown in Figure 1.

3.1. Dataset

A health record of Schistosomiasis has been analysed from the disease database compiled by the (HISPC) Hubei Institute of Schistosomiasis Prevention and Control, China. The patients with advanced Schistosomiasis gave information by surveying sociodemographic and epidemiological factors in Hubei. The (WS261–2006) National diagnostic criteria were used to assess Schistosomiasis. Advanced Schistosomiasis treatment varies from patient to patient depending upon disease conditions. If the patient’s condition is stable for six months, then praziquantel (PZD) treatment can be utilised. The dataset has been categorised into two groups containing 4136 individuals, both males and females. The 1232 cases in which improved patient health are referred to as favourable prognoses, coded as 0, while 2904 death cases are classified as poor prognoses, coded as 1. 18 features are recorded for each participant, as shown in Table 1 [1]. Data was gathered on socioeconomic, demographic, hospitalisation expenditures, clinical characteristics, and surgical procedures [30]. It was found that the area more exposed to water or lake containing marshy lands has a prevalence of Schistosomiasis, as the data also provides evidence.

3.2. Missing Data Imputation

An incomplete dataset for classification and prediction can reduce its effectiveness. The dataset containing large training data was more desirable for prediction because models would train better as more data was given to the model for training. The removal of missing values may increase the model’s performance. A dataset contains a high percentage of empty or blank cells, and it was unclear whether they should be replaced with zeros or remain as missing values. The methods for dealing with the imputation problem use mean attribute values from the original set [31]. The missing values in the dataset are replaced by the class’s median, as presented in Equations (1) and (2).
For odd data elements = n + 1 2 th term
For even data elements = n 2 th term + n 2 + 1 th term 2

3.3. Data Normalization

Data normalisation is a technique in which values are scaled and shifted to make new values [32]. We scaled the data ranging between 0 and 1, known as Z-score normalisation. Data normalisation is useful because all the features are in different forms or units. To normalise attributes in our dataset, Z-score normalisation has been used. It uses mean and standard deviation to normalise, as shown in Equation (3).
X i = X i E ¯ std ( E )
where X i Z-score is a normalised value and X i is the row E value of ith column or mean value. The std E and E are calculated using Equations (4) and (5).
std E = 1 ( n 1 ) i = 1 n ( X i E ¯ ) 2
E ¯ = 1 n i = 1 n X i

3.4. Data Transformation

A process to convert or change data format to another format, structure, or value [33]. Our dataset contains the numerical features which have been converted into categorical values. After the literature survey, continuous features are converted to Boolean and continuous values [34,35], as shown in Table 1.

3.5. Feature Selection

An n-dimensional vector x is commonly used to represent data, with each element x i of x representing a feature. The dataset has multiple features, but not all of them will likely have a favourable effect on the target class [36]. Based on the input x and output y , the features using a scoring function s f calculate filter scores where the feature is denoted by f. The filters are faster in processing and resistant to overfitting as compared to wrapper methods. Information gain [37], correlation-based technique [38], ReliefF [39], OneR [40], and gain ratio are the filters used in the scope of our study.

3.5.1. Information Gain

The Information Gain (IG) approach is based on the information theory notion of entropy. This method ranks features (or input variables) according to the rate at which a variable reduces the target class’s uncertainty (entropy). Shannon’s entropy E s measures how much uncertainty there is in a distribution and is calculated using Equation (6).
E s = j = 1 m P j log 2 P
P j is the probability for outcome j , and m is the number of alternative outcomes. To retrieve the information for all data with c number of classes, we employed Shannon’s entropy, as shown in Equation (7).
Infogain C = j = 1 c P c ( b i ) log 2 P c b i
P c b i  denotes the observed probability for each P c class. Data C has now been divided into L pieces C 1 , C 2 , , C L by feature y . Equation (8) denoted the information for a portion  C k .
Info C k y = j = 1 c P C k y b i log 2 P C k y b i
Info C k y = 0 if  P j is zero then lim P j 0 log 2 P j = 0 . Due to feature y , the value of the information gain is calculated using Equation (9).
InfoGain Y = Infogain C j = 1 m C K C Info C k y
where C and C K represent the number of instances in C and C k . Features with more distinct values are favoured by information gain. Considering Equation (9), more distinct characteristics do not affect the Infogain ( C ), but they do have an impact on j = 1 c P c ( b i ) log 2 P c ( b i ) . For example, patient identification is a feature where each class has exactly one instance, with C id = 0 for each patient Id , maximising the amount of information gained. To put it another way, patient Id is a good feature for the class but ineffective for generalisation. Quinlan et al. [41] proposed that information gain may be normalised using the information split technique, as presented in Equation (10).
SplitInfogain Y = j = 1 m F j F log 2 F j F
F is the dataset, which is divided into F j datasets by feature Y. The information gain ratio is the normalised information gain measure, as shown in Equation (11).
Gainratio = InfoGain ( Y ) SplitInfogain ( Y )

3.5.2. Correlation-Based Feature Selection (CFS)

The link between predictive factors and target variables, often called class, is the foundation of CFS. CFS identifies a selection of traits more correlated with the class than one another. As a result, CFS finds and removes redundant, irrelevant, and noisy features, leaving us with a subset of useful features. The single evaluation function is defined as in Equation (12).
Eval A f = f b cf ¯ f + f ( f 1 ) b ff ¯
where Eval A f is the merit of a feature subset A with f features, b cf is the average value of all feature-class correlations defined in Equation (13).
b cf ¯ = j = 1 f b cfj F
b cfj indicates the correlation between feature j and the class, and b ff is the average of feature correlations, with b f j f k indicating the correlation degree between features j and k as defined in Equations (14) and (15).
b ff ¯ = j = 1 f k = 1 j 1 b f j f k f p
f p = f f 1 2
f p is the amount of feature pairs in subset A, and an information-theoretic measure can be used to calculate the correlation matrix.
b yz = yz m σ y σ z
Here, m in Equation (16) shows instances of dataset whereas σ x σ z shows the standard deviation of x and z , respectively. A weighted Pearson’s Correlation is calculated when one characteristic is linear, and the other is categorically shown in Equations (17) and (18).
b y z = j = 1 w p Y = y j b Y c j Z
b YZ = j = 1 w k = 1 t p ( Y = y j , Z = z j ) ) b Y cj Z cj
The likelihood that Y takes the jth value in the full training set is weighted in each of the estimated correlations.

3.5.3. ReliefF

When dealing with real-time and noisy data, the reliefF algorithm is favourable. The reliefF algorithm selects the instances s j at random and then determines the k nearest neighbours in the similar and dissimilar classes [42]. These instances are referred to as T i , R i for nearest hit and miss. The Manhattan distance is commonly used to distinguish between T i and R i occurrences. Using T i , s j and R i for the updating, the quality estimation, Q E . of all the attributes E is updated. If the values of the instances T i and s j are the same, the attribute E is divided into two instances with the same classes, which is necessary to reduce the Q E and if values of both instances are different with E is divided into different classes of two instances then Q E maximises. The entire procedure is performed n times, with n being a user-defined value. The Q E is updated in this procedure by applying Equations (19)–(21).
Q E = Q E + ( T ¯ + R ¯ ) / n
T ¯ = i = 1 l A ( E , s j , T i ) / l
R ¯ = B bl ( s j ) [ ( P ( B ) 1 P ( bl s j ) ) i = 1 l A ( E , s j , R i ( B ) ) ] / l
P ( B ) stands for the previous class, and A stands for the distance between the examples s j , B stands for classes and bl s j stands for a class of jth sample. The eight most significant features that have been selected through various feature selection techniques are presented in Table 2, and the score values of all features are shown in Table 3.

3.6. Data Resampling

A dataset is imbalanced when the number of instances in one class does not equal other classes. The resampling methods are widely used to balance the dataset. Usually, the model trained on a balanced dataset achieves high results. Most medical-related datasets are imbalanced, which needs resampling approaches to balance the dataset, so multiple undersampling and oversampling techniques have been used in this study.

3.6.1. Random Undersampling

In this approach, most class instances are randomly removed from the training set, so the ratio between both classes would be the same. The problem with this technique is that one would not know which information has been eliminated. The information which was very important for the study may be removed using Equation (22) [43].
R i = S i S i { 0 , 1 }
where S i represents the total number of values with the label i . Undersampling is denoted by S 0 and 0 means from class 0. R 0 R 1 Show that one class has more values than the other, and by combining both, we get Equation (23).
R ( 0 ) R ( 1 ) = S 0 S S 1 S = 1   R ( 0 ) R ( 1 )
When the dataset is rebalanced properly, this case can be applied S 0 = S 1 S 1 but when we have a scenario where the classes are balanced to a certain factor, inequality remains valid. On the contrary, when random points were selected for removal, the conditional data distribution between the classes was assumed to be unchanged, as shown in Equation (24).
R 0 = R 0

3.6.2. Near Miss

Zhang and Mani [44] proposed a method to prevent information loss while utilising the undersampling technique. Instead of resampling the minority class, this strategy undersampling the majority class and making it equal to the minority class. This method uses the average distance between two points, a particular point, and the furthest points in the other class. NearMiss-1, 2, and 3 are the distinct versions. In NearMiss-1, we must choose a proportion of the major class size, and close to minor class points implies choosing the shortest average distance between the three minor class points. We chose the least average distance to the three farthest points from the minority class in NearMiss-2, which implies we chose the percentage of the majority class size closest to all points of the minority class. Select the nearest majority class points for each minor class in NearMiss-3. In this study, the NearMiss-1 method has been used.

3.6.3. Cluster Centroid

Cluster Centroids were proposed to minimise the loss of information in the majority class [45]. By calculating the ratio between both classes, Cluster Centroids substitutes the majority of samples with clusters of centroids using the K-means method. The loss of information has been minimised by making groups based on similarity. The K-means algorithm was used to get the clusters when applied to the data based on the level of undersampling. Then, the majority of samples from the clusters were replaced by the cluster centroids, which contain a different representation of the majority class at the centre from K-Means.

3.6.4. Synthetic Minority Oversampling Technique

Chawla et al. [46] introduced the SMOTE as it is one of the easier and most successful to address the problem of class imbalance. It creates new minority class instances synthetically rather than repeating duplicates of minority class instances [47]. Class overlapping is a critical aspect that makes learning a good classifier hypothesis for an unbalanced dataset difficult. It over-samples by using k nearest neighbours from minority samples, calculating the difference, multiplying the difference by a random number, and adding it to the feature vector, as shown in Equation (25).
y new = y i + y i y i α
y i is K-nearest neighbour of y i and α € [0; 1] is a random number. Borderline-SMOTE is a further variant of SMOTE that solves its limitation. The cases closer to the borderline (i.e., majority class examples) are more difficult to categorise correctly. This approach prioritises these examples to increase oversampling performance. It is similar to SMOTE in that it calculates synthetic examples based on the class with minimum samples on the borderline. For example, borderline-SMOTE [48] prioritises locating samples on the class boundaries, i.e., borderline samples, and then oversampling them to improve prediction classification techniques are predicated on learning these boundaries during training. The algorithm is largely based on SMOTE.
  • For each sample in the minority class y i Ɛ S min . Get the collection of k-nearest neighbours, S Knn .
  • Determine the number of nearest neighbours who are members of the majority class S Knn S maj for each y i . Select observations such that in Equation (26):
    n 2 S Knn S maj < n
  • After that, the observations are run through the standard SMOTE method to generate synthetic points that solely include the minority and majority classes.
Clustering, filtering, and oversampling are the three steps in K mean SMOTE [49] proposed by Douzas et al. [50]. The input samples are grouped into k groups using k-means clustering, then choosing clusters for oversampling at the filtering stage, keeping the high proportion of samples from minority classes. The amount of synthetic samples to be generated is then dispersed, with more samples being assigned to clusters where minority samples are sparsely distributed then, to achieve the required ratio of samples SMOTE is used.
Each cluster’s proportion of minority and majority instances is used to identify clusters for oversampling. Oversampling is selected by default for any cluster containing at least 50% minority samples; by changing the imbalance ratio threshold, we can change the behaviour, a k-means SMOTE hyperparameter that defaults to 1. The ratio is defined as majoritycount + 1 minoritycount + 1 of imbalanced instances. Cluster selection becomes more selective, requiring a higher ratio of minority samples to be selected when there is a criterion for an imbalance ratio. Lowering the barrier, on the other hand, loosens the selection criteria, making it possible to choose clusters with a bigger majority share and assigned sample weights between 0 and 1. They are calculated as follows:
  • Euclidean distance was calculated for selected clusters (s), ignoring the majority class. The mean distance of each cluster is obtained by adding all non-diagonal entries of the distance matrix and dividing them with non-diagonal elements.
  • Divide the number of minority instances in each cluster by the average minority distance raised to the power of the number of features n to get density, i.e., density s = minoritycount s averageMinorityDistance s n .
  • To acquire a measure of sparsity, invert the density measure: sparsity s = 1 density s .
  • We can calculate the cluster’s sampling weight by dividing the cluster’s sparsity factor by the sum of all cluster’s sparsity factors.
Resulting total sampling Weights = 1 , i.e., samplingWeights s x X is the total number of samples generated in that cluster that was oversampled using SMOTE. The hyperparameter k nearest neighbours, or KNN, in SMOTE, determines how many surrounding minority samples of a the point b are randomly chosen. When a cluster includes fewer than knn + 1 minority samples, the value of KNN may need to be modified downward, depending on the SMOTE implementation. After SMOTE has been employed, both classes, minority and majority ones are the same. Figure 2 depicts the dataset before and after resampling. Table 4 presents the distribution of the dataset before and after resampling.

3.7. Gradient Boosting

Gradient Boosting (GB) is an iterative machine learning-based approach for solving classification problems. This method is based on an ensemble learning model trained using the previous iteration’s mistakes as input. By fitting new learners, GB corrects misclassified data to the ensemble residual, which is the gap between the goal outputs and the ensemble’s current predictions. GB aims to maximise the ensemble’s prediction power while minimising bias. The benefit of Boosting is that it significantly impacts accuracy, but it comes at the expense of being slower to train because each learner is trained linearly.
It is a Boosting-like algorithm [51] and has a dataset D t = a i , b i 1 M for training, the goal of GB is to find an approximation, A ^ , of the function A a , which maps instances a to their output values b , by minimising the expected value of a given loss function, L f b , A a . GB builds an additive approximation of A a as a weighted sum of functions calculated via Equation (27).
A n a = A n 1 a + w n g n a
where w n is the weight of n t h function and g n a are the parts of ensemble models. First, we obtain a constant approximation of A a as shown in Equation (28).
A 0 a = argmin β i = 1 M Lf ( b i , β )
Models are minimised as in Equation (29). However, instead of solving the optimisation problem directly, each g n can be seen as a greedy step in a gradient descent optimisation for A . For that, each model, A is trained on a new dataset D t = a i , pr ni i = 1 M , where the pseudo-residuals pr ni , are calculated via Equation (29).
( w n ,   g n a ) = argmin w , g i = 1 M Lf ( b i , A n 1 a i + wg ( a i ) )
After that, a line search optimisation problem is used to calculate the value of w n . If the iterative procedure is not adequately regularised, this approach may suffer from over-fitting. If the model g n fully matches the pseudo-residuals for some loss functions (e.g., quadratic loss), the pseudo-residuals become zero in the following iteration, and the process finishes prematurely. Several regularisation hyper-parameters are examined to govern the additive process of GB. The natural way to regularise GB is to apply shrinkage to reduce each gradient descent step A n a = A n 1 a + w n g n a with u = 0 , 1.0 . Typically, the value of u is set to 0.1. Furthermore, more regularisation may be obtained by reducing the complexity of the trained models. We can restrict the depth of decision trees or the minimum number of occurrences required to divide a node in the case of decision trees.
In contrast to a random forest, the default settings for these hyper-parameters in GB are designed to severely limit the expressive potential of the trees (for example, the depth is often limited to 3 to 5). Finally, various versions of GB incorporate hyper-parameters that randomise the base learners, such as random subsampling without substitution. These hyper-parameters can increase the generalisation of the ensemble [52]—Algorithm 1 lists all the steps of gradient boosting.
Algorithms 1 Gradient Boosting
1.
Initialising the constant
A 0 a = argmin β i = 1 M Lf ( b i , β )
2.
A For-Do loop ( i = 1 M )
3.
At the training point, calculate the gradient. The new base learner was fitted to the target value to find the best gradient step. Then, update the estimate function
( w n , g n a ) = argmin w , g i = 1 M Lf ( b i , A n 1 a i + wg ( a i ) )
4.
ii+1
5.
Loop end
6.
Return
XGBoost [53] is a highly scalable DT ensemble method based on GB. XGBoost, like GB, minimises a loss function because it is only interested in DT as a base classifier shown in Equation (30).
Lf xgb = i = 1 M Lf ( b i , A a i ) + n = 1 N ( g n )
g = γ Lt + 1 2 π v 2
With the leaves of tree Lt in Equation (31), and the output scores v , the split criterion can be incorporated into decision trees’ loss function, resulting in a pre-pruning approach. In XGBoost, shrinkage is an additional regularisation hyper-parameter that reduces the step size in the additive expansion. Finally, other tactics, such as tree depth, limit tree complexity, reduce storage space and increase training speed. To train individual trees, random subsamples and column subsampling at the tree and tree node levels are among the randomisation approaches used in XGBoost. Furthermore, first and second-order gradients were obtained by constructing a function and sending it through the "objective" hyper-parameter.
XGBoost, in particular, focuses on minimising computation cost by determining the optimal split of DT building algorithms. In most split discovery algorithms, all feasible candidate splits are enumerated, and the One with the biggest gain has been chosen. A regular scan of each ranked attribute is required to discover the optimal split for each node. XGBoost employs a compressed column-based structure containing data in a pre-sorted form to prevent sorting the data in each node. XGBoost uses an approach based on data percentiles, in which just a sample of candidate splits is examined, and their gain is calculated using aggregated statistics. This study uses the parameters used to tune XGBoost: learning rate, minimum loss (gamma), max_dep, sample at each split, and subsampling rate.
Light Gradient Boosting (LGBoost) [54] executes GB and suggests several variations focusing on a computationally efficient approach, similar to XGBoost, based on histogram features and also provides hyper-parameters that allow being used in a range of scenarios: This works on both GPU and CPU, as simple GB, and has several options. Gradient-based One-Side Sampling (GOSS) and Exclusive Feature Bundling are two new features proposed by LGBoost. GOSS is a subsampling strategy for creating the training sets for the ensemble’s base trees. This strategy, like AdaBoost, tries to increase the relevance of cases, referred to as instances with a higher gradient.
The training consist of gradients x and a sample fraction y when the GOSS option is enabled, when computing the information gain, the instances are weighted by 1 x / y for the change in distribution. Many sparse features were combined with the Exclusive Feature Bundling (EFB) method into a single feature. GOSS and EFB both improve training speed. The parameters tuned for LGBoost are learning rate, max_num_leaves, highest_gradient, lowest_gradient, and feature_fraction_node.

CatBoost

CatBoost [55] is a GB library that reduces prediction shifts during training. This distribution shift is the difference between A a | a with a i being a training instance and A a | a for a test instance a . GB minimises gradients during training using the same instances for both estimates and models. CatBoost proposes that the gradients be estimated in their training set by a series of models which do not contain that occurrence. CatBoost does this by introducing a random permutation into the training cases. CatBoost’s logic is to create i = 1 , , N base models for each of the M Boosting iterations. The gradient of the i + 1 instance for the ( m + 1 ) t h Boosting iteration is estimated using the i t h model of the m t h iteration, trained on permutation first i instance. This method is repeated with j other random permutations in order to be independent of the first random permutation [56,57,58].
CatBoost is implemented so that a single model per iteration handles all permutations and models. Symmetric trees are formed by extending all leaf nodes level-wise with the same split condition to serve as the foundation models. The catBoost algorithm handles categorical features by replacing them with the numeric feature, which assesses the predicted goal value for each category. This numeric feature should preferably be generated using a different dataset to minimise overfitting the training data. This, however, is not always achievable. CatBoost proposes a strategy for computing this additional feature identical to the One used to generate the models. That is the information of i n s t a n c e s < i is utilised to determine the feature value of instance i for a given random permutation of the instances. The feature value acquired for each instance is then averaged after many variations. Computing target statistics for categorical features, like EFB in the LGBoost model, is a preprocessing approach. CatBoost is a large library with many features like GPU training, different parameters according to scenarios, and Boosting standards.

4. Results

The model has been trained on training data and then uses the test set to assess the model’s generalisation with a 70:30 split approach. When evaluating the performance of classifiers, performance assessment is crucial. A confusion matrix has been used to evaluate the performance of a classifier. The confusion matrix is a multidimensional table that depicts successfully or incorrectly predicted times for each class in real and projected costs for all times. Any predictive modelling assignment necessitates the evaluation of models. It is considerably more important in predictive modelling, whereas the performance and variety of classifiers must be assessed comprehensively. Each of the assessment measures is based on one of four classes: true positive (TP), true negative (TN), false positive (FP), and false negative (FN). The experiments and results show that our model achieved state-of-the-art performance. We evaluate our model by comparing it with other machine-learning models. The results are compared in terms of accuracy, precision, recall, and F1-Score, which can be calculated using parameters which are defined as [59,60,61,62]:
  • True Positive (TP): It identifies how much the data instances are identified as recovery cases.
  • True Negative (TN): It identifies how many data instances are categorised as death cases.
  • False Positive (FP): It identifies how much the data instances are incorrectly categorised as recovery cases.
  • False Negative (FN): It Identifies how many data instances are incorrectly categorised as death cases;
Accuracy is the ratio of all correctly predicted samples to the summation of total predictions, as shown in Equation (32).
Accuracy = TP + TN TP + TN + FP + FN
Precision: It identifies if the positive predictions are correctly determined and is the ratio of TP to the summation of TP and FP, as shown in Equation (33).
Precision = TP TP + FP
Recall: It identifies total relevant results correctly predicted by the model and is the ratio of TP to the summation of TP and FN, as shown in Equation (34).
Recall = TP TP + FN
F1-Score: It is characterised as the harmonic mean of the model’s precision and recall and is a way to integrate the model’s precision and recall. It is stated in Equation (35).
F 1 - Score = 2 × Precision × Recall Precision + Recall
Before resampling, different feature selection techniques have also been used to reduce the data dimensionality. We have applied the five feature selection techniques: Correlation, Information gain, Gain ratio, ReliefF and OneR. We have selected eight features by using these feature selection techniques. Each feature selection approach scores all features; hence each technique ranks the features. Which enables the selection of the most important characteristics k using a feature selection approach. j J where J be a set contain integers and (j ∈ [1, 2, …, n − 1, n]. Each j specifies the amount of the experiment’s best features that must be assessed. In the best-case scenario, J includes all numbers in the range [1, 2, …, n − 1, n]. Because feature selection strives to choose the best features with respect to scorer. The feature selection technique’s assessment depends on the score of the underlying feature selection approach, and for each method, an experiment was run numerous times. Furthermore, we utilised undersampling and oversampling techniques to balance the dataset.
The GB algorithm we used with the combination of hyperparameter tuning, which are learning-rate = 0.1, the number of tress (n-estimators = 100), max-depth = 3, min-samples-split = 2, min-sample-leaf = 1 and subsample = 1. The combination of the parameter used by LGBoost is max-depth = 20, num-of–iterations or estimator = 100, num-leaves = 31, and sub-sample = 1. The other parameter, like L1, and L2 regularisation or the alpha value, was 0. For XGBoost, we have set the parameter’s value as learning rate = 0.1, n-estimator = 100, subsample = 1 and max-depth = 3. When tuning XGBoost, we have to take care of mostly 3 parameters: number-of-tree, trees-depth and stepsize to achieve better results. The parameters used for the CatBoost algorithm are: iteration = 1000, leaf-estimation-iterations = 100, depth = 7, L2 regularisation = 5, learning-rate = 0.03, and other parameters like bagging-temperature, random-strength, leaf-estimation, eval-metric, bootstrap-type and loss-function are optional. As our model was a binary class, parameters like loss-function and eval-metric were not mentioned, but in the case of multiclass need to mention it. Table 5 shows the result of the random undersampling technique with multiple feature selection techniques. The GB model performed poorly, with an accuracy of 68.1% in the case of ReliefF and 68.6% with the OneR feature selections technique [63,64,65]. The XGBOOST model performed very well for correlation and information gain feature selection techniques with an accuracy of 82.8%. In contrast, the CatBoost model performed very well for the Gain ratio feature selection technique leading in the case of recall with a value of 82.7%.
Table 6 and Table 7 show the results of Cluster Centroid and Near Miss techniques. The GB model performed well in the case of the Gain ratio with an accuracy of 78.9%, shown in Table 6, and the CatBoost model performed well overall in both tables with an accuracy of 80% and 78.9%. Still, the results obtained from both undersampling techniques were less than the previous study done by Li G et al. [1]. The CatBoost performed very well classifying both classes when undersampling techniques were used. However, in the case of the undersampling, the accuracy decreased to 78.9% when clustering centroid techniques were used and 79.2 for Near miss, which was poor because, on an imbalanced dataset, Li G et al. [1] achieved an accuracy of 80%. Even though the accuracy achieved by CatBoost using random undersampling techniques was 82.7%, greater than Li G et al. [1], we cannot neglect oversampling techniques. As in Figure 3, a graph plot shows the highest result to feature selection and resampling techniques.
Table 8 contains the result of oversampling techniques to feature selection techniques. Table 8 shows the result of the random oversampling technique for multiple feature selection techniques. All OneR and ReliefF models performed very poorly, with the highest accuracy of 68.3% and 73.3%, whereas they performed well in the Gain ratio with an accuracy of 82.9%. Both XGBoost and CatBoost models obtained the highest result of 82.9 % concerning correlation and Information gain feature selection techniques.
Table 9 shows the outcome of the SMOTE oversampling approach in comparison to numerous feature selection strategies. In terms of multiple feature selection strategies, the results produced using these resampling techniques have been less than those obtained using random oversampling techniques. In the OneR and ReliefF tests, models concerning those parameters underperformed with values of 75.9% and 71.0% but well in the Gain Ratio test with a value of 82.9 by LGBoost. Regarding correlation and information gain feature selection strategies, both the XGBoost and CatBoost models came on top, but XGBoost took the lead with a value of 83.3%.
Table 10 presents the results of an oversampling strategy K-means SMOTE and various feature selection procedures. The result obtained using this resampling approach was the best of all the resampling techniques in terms of multiple feature selection strategies. Utilising this resampling method, the results obtained from models employing feature selection techniques, Information gain, ReliefF, and OneR were unsatisfactory; however, models using Gain ratio and Correlation feature selection strategies did extremely well with a value of 87.1%. In addition, the CatBoost model outperformed other models.
Summarising the K means SMOTE approach when utilised, the results were more efficient when using the correlation-based feature selection technique. The table, as shown above, summarises the performance evaluation of several classifiers using a selected feature subset from each feature selection approach and method. The CatBoost approach’s performance was also compared to that of typical machine learning models, and it was found that the CatBoost method outperformed traditional machine learning models. Table 5 shows the findings, which show that the CatBoost model produced the greatest results, with an accuracy of 87.1 percent, also shown in Figure 4.

5. Discussion

Machine learning and data mining approaches can automatically discover complex patterns in the healthcare domain, such as COVID-19 [66,67], Skin Cancer [59], Breast Cancer [16], Malignant Mesothelioma [68,69,70], and Cervical Cancer [7], researchers are motivated to use these techniques in the early prediction of Schistosomiasis. In recent years, many machine learning methods, including LR, DT, RF, and ANN, have been widely applied in disease detection and prediction [59]. Before resampling, different feature selection techniques have also been used to reduce the data dimensionality. We have applied the five feature selection techniques: Correlation, Information gain, Gain ratio, ReliefF and OneR. We have selected eight features by performing multiple experiments, as explained above. Furthermore, we utilised undersampling and oversampling techniques to balance the dataset. We have used four algorithms, GB, LGBoost, XGBoost and CatBoost, with a combination of various hyperparameters discussed in the result section.
The results of the random undersampling approach with several feature selection strategies are shown in Table 5. The GB model fared badly, with an accuracy of 68.1% for ReliefF and 68.6% for OneR feature choices. The XGBOOST model fared extremely well for correlation and information gain feature selection approaches, with an accuracy of 82.8%, while the CatBoost model did very well for the Gain ratio feature selection technique, with an accuracy of 82.7%. The outcomes of the Cluster Centroid and Near Miss approaches are shown in Table 6 and Table 7. The GB model fared well in the Gain ratio instance, with an accuracy of 78.9% in Table 6, while the CatBoost model did well overall in both tables, with an accuracy of 80% and 78.9%.
Nonetheless, the findings from both undersampling strategies were lower than the prior study by Li G et al. [1]. When undersampling techniques were utilised, the CatBoost performed admirably in categorising both groups. However, when undersampling was utilised, the accuracy dropped to 78.9% when clustering centroid approaches were applied, and 79.2% for Near miss, which was disappointing considering Li G et al. [1] obtained an accuracy of 80% on an unbalanced dataset. Even though CatBoost’s accuracy utilising random undersampling strategies was 82.7%, which was higher than Li G et al. [1] as shown in a graph plot, Figure 3, displays the best outcome for feature selection and resampling procedures.
The results of the random oversampling approach for different feature selection strategies are shown in Table 8. All models fared badly with OneR and ReliefF techniques, with the greatest accuracy of 68.3% and 73.3%, respectively. However, they did well in the Gain ratio, with an accuracy of 82.9%. In terms of correlation and information gain feature selection strategies, both the XGBoost and CatBoost models achieved 82.9%. Table 9 compares the SMOTE oversampling strategy to various feature selection algorithms. The results achieved via these resampling procedures were lower than those acquired by random oversampling techniques. Models considering such parameters underperformed at 71.0% in the OneR and ReliefF tests. XGBoost takes the lead in correlation and information gain feature selection algorithms, with a value of 83.3%. Table 10 shows the outcomes of a K-means SMOTE oversampling technique and several feature selection procedures. In terms of various feature selection strategies, the result of this resampling methodology was the best of all resampling procedures. Using this resampling strategy, the results from models utilising feature selection procedures, Information gain, ReliefF, and OneR were poor; however, models adopting Gain ratio and Correlation feature selection strategies performed very well, with a value of 87.1%. Other models were outperformed by the CatBoost model. When the K-means SMOTE strategy was applied, the results were more efficient with the correlation-based feature selection technique. Table 10 summarises the performance evaluation of classifiers. The performance of the CatBoost technique was also compared to that of standard machine learning models, and it was discovered that the CatBoost method beat traditional machine learning models. The outcomes are provided in Table 10, which reveals that the CatBoost model gave the best results, with an accuracy of 87.1 %, as shown in Figure 4.
CatBoost was more accurate than other methods in our research on different medical conditions. In the CatBoost model, we discovered that ascitic fluid volume was the most useful predictor. The most prevalent sign of advanced hepatic illness is ascites. The ascitic subtype of advanced Schistosomiasis is the most dangerous, accounting for 65–90 percent of cases [60]. Granulomatous inflammation can be caused by venous blockage and portal hypertension and leads to continuous fibrosis and a drop in plasma colloid osmotic pressure [61]. Ascites are the most common consequence of liver injury, and their severity directly influences the overall prognosis. In advanced Schistosomiasis patients, the occurrence of severe ascites is one of the best indicators of a high level of impairment [62]. This study has demonstrated that CatBoost is a novel and promising modelling framework for the unfavourable prognosis of advanced Schistosomiasis. A comparison between our model and the previous studies is shown in Table 11.

6. Conclusions

Machine learning methods have been used in various fields in combination with unbalanced approaches. This study aims to use supervised learning algorithms to train multiple Schistosomiasis disease prediction systems. Because an unbalanced dataset is important for improving the model’s performance in classification challenges, several resampling strategies were utilised to balance the dataset. We began by employing exploratory data analysis approaches, such as data standardisation, to analyse the dataset. Then, we conducted several tests to validate the performance of feature selection strategies against one another and narrow down the characteristics that may properly diagnose the disease. These variables can be addressed early in the disease to the point where it threatens human life. We began by comparing the outcomes of several GB models, such as LGBoost, XGBoost, and CatBoost, against each other and earlier studies. The CatBoost model has a greater prediction accuracy rate than typical machine learning-based models. Using the K-means SMOTE resampling approach, the CatBoost method had the greatest accuracy of 87.1%. We will utilise more datasets linked to this disease in the future to include the features categorised as important by the dataset collector but not included in this dataset. Finally, we will release this framework on the web so medical professionals can take benefit from it.

Author Contributions

Data curation, Z.A., M.F.H., T.M.A., K.S., I.A.H. and S.L.; Formal analysis, Z.A., M.F.H., K.S., I.A.H., and S.B.; Funding acquisition, I.A.H., K.S., S.B. and M.A.; Investigation, Z.A., M.F.H., T.M.A. and K.S.; Methodology, Z.A., M.F.H., T.M.A., K.S., I.A.H. and S.B.; Project administration, K.S., I.A.H. and S.L.; Software, Z.A., M.F.H., T.M.A., K.S., I.A.H. and M.A.; Supervision, K.S., I.A.H. and S.L.; Validation, Z.A., M.F.H., K.S., I.A.H., S.B., A.K., M.A. and S.L.; Visualization, Z.A., M.F.H., K.S., T.M.A., I.A.H., S.B., A.K., M.A. and S.L.; Writing—original draft, Z.A., M.F.H., T.M.A., K.S., I.A.H., A.K. and S.L.; Writing—review & editing, Z.A., M.F.H., T.M.A., K.S., I.A.H., S.B., M.A., A.K. and S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2023R195), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data will be available upon request.

Acknowledgments

This research was funded by Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2023R195), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, G.; Zhou, X.; Liu, J.; Chen, Y.; Zhang, H.; Chen, Y.; Liu, J.; Jiang, H.; Yang, J.; Nie, S. Comparison of three data mining models for prediction of advanced schistosomiasis prognosis in the Hubei province. PLoS Negl. Trop. Dis. 2018, 12, e0006262. [Google Scholar] [CrossRef] [PubMed]
  2. Fusco, T.; Bi, Y.; Wang, H.; Browne, F. Data mining and machine learning approaches for prediction modelling of schistosomiasis disease vectors: Epidemic disease prediction modelling. Int. J. Mach. Learn. Cybern. 2020, 11, 1159–1178. [Google Scholar] [CrossRef] [PubMed]
  3. Olveda, D.U.; Olveda, R.M.; Jan Montes, C.; Chy, D.; Modesto Abellera III, J.B.; Cuajunco, D.; Lam, A.K.; McManus, D.P.; Li, Y.; Ross, A.G.; et al. Clinical management of advanced Schistosomiasis: A case of portal vein thrombosis-induced splenomegaly requiring surgery. Case Rep. 2014, 2014, bcr2014203897. [Google Scholar] [CrossRef] [PubMed]
  4. Huang, L.H.; Qiu, Y.W.; Hua, H.Y.; Niu, X.H.; Wu, P.F.; Wu, H.Y.; Zhu, H.Y.; Yang, X.J.; Yao, S.Z.; Li, Y.G. The efficacy and safety of entecavir in patients with advanced Schistosomiasis co-infected with hepatitis B virus. Int. J. Infect. Dis. 2013, 17, e606–e609. [Google Scholar] [CrossRef]
  5. Schistosomiasis (Bilharzia). Available online: https://www.who.int/health-topics/schistosomiasis#tab=tab_1 (accessed on 1 March 2022).
  6. Zhang, L.J.; Xu, Z.M.; Dang, H.; Li, Y.L.; Lü, S.; Xu, J.; Li, S.Z.; Zhou, X.N. Endemic status of schistosomiasis in People’s Republic of China in 2019. Zhongguo Xue Xi Chong Bing Fang Zhi Za Zhi 2020, 32, 551–558. [Google Scholar] [CrossRef]
  7. Alam, T.M.; Milhan, M.; Khan, A.; Iqbal, M.A.; Wahab, A.; Mushtaq, M. Cervical Cancer Prediction through Different Screening Methods Using Data Mining. IJACSA Int. J. Adv. Comput. Sci. Appl. 2019, 10. Available online: https://www.ijacsa.thesai.org (accessed on 1 March 2022). [CrossRef]
  8. Osakunor, D.N.M.; Woolhouse, M.E.J.; Mutapi, F. Paediatric schistosomiasis: What we know and what we need to know. PLoS Negl. Trop. Dis. 2018, 12, e0006144. [Google Scholar] [CrossRef]
  9. Ashour, A.S.; Hawas, A.R.; Guo, Y. Comparative study of multiclass classification methods on light microscopic images for hepatic schistosomiasis fibrosis diagnosis. Health Inf. Sci. Syst. 2018, 6, 7. [Google Scholar] [CrossRef]
  10. Zhao, M.; Zhang, Z.; Chow, T.W.S. Trace ratio criterion based generalised discriminative learning for semi-supervised dimensionality reduction. Pattern Recognit. 2012, 45, 1482–1499. [Google Scholar] [CrossRef]
  11. Baig, T.I.; Khan, Y.D.; Alam, T.M.; Biswal, B.; Aljuaid, H.; Gillani, D.Q. ILipo-PseAAC: Identification of lipoylation sites using statistical moments and general PseAAC. Comput. Mater. Contin. 2022, 71, 215–230. [Google Scholar]
  12. Tariq, A.; Awan, M.J.; Alshudukhi, J.; Alam, T.M.; Alhamazani, K.T.; Meraf, Z. Software Measurement by Using Artificial Intelligence. J. Nanomater. 2022, 2022, 7283171. [Google Scholar] [CrossRef]
  13. Alam, T.K.; Shaukat, M.; Mushtaq, M.; Ali, Y.; Khushi, M. Corporate Bankruptcy Prediction: An Approach Towards Better Corporate World. Available online: https://academic.oup.com/comjnl/article-abstract/64/11/1731/5856206 (accessed on 14 April 2022).
  14. Alam, T.K.; Shaukat, I.; Hameed, S.; Li, J.; Khushi, M. An Investigation of Credit Card Default Prediction in the Imbalanced Datasets. Available online: https://ieeexplore.ieee.org/abstract/document/9239944/ (accessed on 14 April 2022).
  15. Baig, T.I.; Alam, T.M.; Anjum, T.; Naseer, S.; Wahab, A.; Imtiaz, M.; Raza, M.M. Classification of Human Face: Asian and Non-Asian People. In Proceedings of the 2019 International Conference on Innovative Computing (ICIC), Lahore, Pakistan, 1–2 November 2019; pp. 1–6. [Google Scholar] [CrossRef]
  16. Ghani, M.U.; Alam, T.M.; Jaskani, F.H. Comparison of Classification Models for Early Prediction of Breast Cancer. In Proceedings of the 2019 International Conference on Innovative Computing (ICIC), Lahore, Pakistan, 1–2 November 2019; pp. 1–6. [Google Scholar] [CrossRef]
  17. Jiang, H.; Deng, W.; Zhou, J.; Ren, G.; Cai, X.; Li, S.; Hu, B.; Li, C.; Shi, Y.; Zhang, N.; et al. Machine learning algorithms to predict the 1 year unfavourable prognosis for advanced Schistosomiasis. Int. J. Parasitol. 2021, 51, 959–965. [Google Scholar] [CrossRef]
  18. Olanloye, O.D.; Olasunkanmi, O.; Oduntan, O.E. Comparison of Support Vector Machine Models in the Classification of Susceptibility to Schistosomiasis. Balk. J. Electr. Comput. Eng. 2020, 8, 266–271. [Google Scholar] [CrossRef]
  19. Asarnow, D.; Singh, R. Determining Dose-Response Characteristics of Molecular Perturbations in Whole-Organism Assays Using Biological Imaging and Machine Learning. In Proceedings of the 2018 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Madrid, Spain, 3–6 December 2018; pp. 283–290. [Google Scholar] [CrossRef]
  20. Kasse, B.; Gueye, B.; Diallo, M.; Santatra, F.; Elbiaze, H. IoT based Schistosomiasis Monitoring for More Efficient Disease Prediction and Control Model. In Proceedings of the 2019 IEEE Sensors Applications Symposium (SAS), Sophia Antipolis, France, 11–13 March 2019. [Google Scholar] [CrossRef]
  21. Chicco, D.; Rovelli, C. Computational prediction of diagnosis and feature selection on mesothelioma patient health records. PLoS ONE 2019, 14, e0208737. [Google Scholar] [CrossRef]
  22. He, H.; Garcia, E.A. Learning from imbalanced data. IEEE Trans. Knowl. Data Eng. 2009, 21, 1263–1284. [Google Scholar] [CrossRef]
  23. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic Minority Over-sampling Technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  24. Mazurowski, M.A.; Habas, P.A.; Zurada, J.M.; Lo, J.Y.; Baker, J.A.; Tourassi, G.D. Training neural network classifiers for medical decision making: The effects of imbalanced datasets on classification performance. Neural. Netw. 2008, 21, 427–436. [Google Scholar] [CrossRef] [PubMed]
  25. García-Pedrajas, N.; Ortiz-Boyer, D.; García-Pedrajas, M.D.; Fyfe, C. Class imbalance methods for translation initiation site recognition. In Proceedings of the International Conference on Industrial, Engineering and Other Applications of Applied Intelligent Systems, Cordoba, Spain, 1–4 June 2010; pp. 327–336. [Google Scholar] [CrossRef]
  26. Blagus, R.; Lusa, L. Class prediction for high-dimensional class-imbalanced data. BMC Bioinform. 2010, 11, 523. [Google Scholar] [CrossRef]
  27. Xia, S.; Xue, J.B.; Zhang, X.; Hu, H.H.; Abe, E.M.; Rollinson, D.; Bergquist, R.; Zhou, Y.; Li, S.Z.; Zhou, X.N. Pattern analysis of schistosomiasis prevalence by exploring predictive modeling in Jiangling County, Hubei Province, China. Infect. Dis. Poverty 2017, 6, 91. [Google Scholar] [CrossRef]
  28. Ali, Y.; Farooq, A.; Alam, T.M.; Farooq, M.S.; Awan, M.J.; Baig, T.I. Detection of Schistosomiasis Factors Using Association Rule Mining. IEEE Access 2019, 7, 186108–186114. [Google Scholar] [CrossRef]
  29. Wrable, M.; Kulinkina, A.V.; Liss, A.; Koch, M.; Cruz, M.S.; Biritwum, N.-K.; Ofosu, A.; Gute, D.M.; Kosinski, K.C.; Naumova, E.N. The use of remotely sensed environmental parameters for spatial and temporal schistosomiasis prediction across climate zones in Ghana. Environ. Monit. Assess. 2019, 191, 301. [Google Scholar] [CrossRef] [PubMed]
  30. Gong, Y.-F.; Zhu, L.-Q.; Li, Y.-L.; Zhang, L.-J.; Xue, J.-B.; Xia, S.; Lv, S.; Xu, J.; Li, S.-Z. Identification of the high-risk area for schistosomiasis transmission in China based on information value and machine learning: A newly data-driven modeling attempt. Infect. Dis. Poverty 2021, 10, 88. [Google Scholar] [CrossRef] [PubMed]
  31. Van Buuren, S. Flexible Imputation of Missing. 2018. Available online: https://scholar.google.com/scholar?hl=en&as_sdt=0%2C5&q=Van+Buuren%2C+S.+%282018%29+Flexible+Imputation+of+Missing+Data.+Chapman+and+Hall%2FCRC&btnG= (accessed on 11 April 2022).
  32. Patro, S.G.K.; Sahu, K.K. Normalisation: A preprocessing stage. arXiv 2015, arXiv:1503.06462. [Google Scholar]
  33. Fan, Q.; Zhu, C.J.; Xiao, J.Y.; Wang, B.H.; Yin, L.; Xu, X.L.; Rong, F. An application of apriori algorithm in SEER breast cancer data. In Proceedings of the 2010 International Conference on Artificial Intelligence and Computational Intelligence, Sanya, China, 23–24 October 2010. [Google Scholar]
  34. Shaukat, K.; Luo, S.; Abbas, N.; Mahboob Alam, T.; Ehtesham Tahir, M.; Hameed, I.A. An analysis of blessed Friday sale at a retail store using classification models. In Proceedings of the 4th International Conference on Software Engineering and Information Management (ICSIM 2021), Yokohama, Japan, 16–18 January 2021. [Google Scholar]
  35. Joseph, J.; Badrinath, P.; Basran, G.S.; Sahn, S.A. Is albumin gradient or fluid to serum albumin ratio better than the pleural fluid lactate dehydroginase in the diagnostic of separation of pleural effusion? BMC Pulm. Med. 2002, 2, 1. [Google Scholar] [CrossRef] [PubMed]
  36. Alam, T.M.; Mushtaq, M.; Shaukat, K.; Hameed, I.A.; Sarwar, M.U.; Luo, S. A Novel Method for Performance Measurement of Public Educational Institutions Using Machine Learning Models. Appl. Sci. 2021, 11, 9296. [Google Scholar] [CrossRef]
  37. Azhagusundari, B.; Thanamani, A.S. Feature selection based on information gain. Int. J. Innov. Technol. Explor. Eng. 2013, 2, 18–21. [Google Scholar]
  38. Hall, M. Correlation-Based Feature Selection for Machine Learning. Doctoral Dissertation, The University of Waikato, Hamilton, New Zealand, 1999. [Google Scholar]
  39. Wang, Y.; Makedon, F. Application of Relief-F feature filtering algorithm to selecting informative genes for cancer classification using microarray data. In Proceedings of the 2004 IEEE Computational Systems Bioinformatics Conference, 2004. CSB 2004, Stanford, CA, USA, 19 August 2004. [Google Scholar]
  40. Mejía-Lavalle, M.; Sucar, E.; Arroyo, G. Feature selection with a perceptron neural net. In Proceedings of the International Workshop on Feature Selection for Data Mining, Hyderabad, India, 21 June 2010. [Google Scholar]
  41. Wu, X.; Kumar, V.; Quinlan, J.R.; Ghosh, J.; Yang, Q.; Motoda, H.; McLachlan, G.J.; Ng, A.; Liu, B.; Yu, P.S.; et al. Top 10 algorithms in data mining. Knowl. Inf. Syst. 2008, 14, 1–37. [Google Scholar] [CrossRef]
  42. Praveena, H.D.; Subhas, C.; Naidu, K.R. Automatic epileptic seizure recognition using reliefF feature selection and long short term memory classifier. J. Ambient Intell. Humaniz. Comput. 2021, 12, 6151–6167. [Google Scholar] [CrossRef]
  43. Fernández, A.; Garcia, S.; Herrera, F.; Chawla, N.V. SMOTE for learning from imbalanced data: Progress and challenges, marking the 15-year anniversary. J. Artif. Intell. Res. 2018, 61, 863–905. [Google Scholar] [CrossRef]
  44. Mani, I.; Zhang, I. kNN approach to unbalanced data distributions: A case study involving information extraction. In Proceedings of the Workshop on Learning from Imbalanced Datasets, Mclean, VA, USA, 21 August 2003. [Google Scholar]
  45. Yen, S.J.; Lee, Y.S. Cluster-based under-sampling approaches for imbalanced data distributions. Expert Syst. Appl. 2009, 36, 5718–5727. [Google Scholar] [CrossRef]
  46. Nasir, A.; Shaukat, K.; Iqbal Khan, K.; A. Hameed, I.; Alam, T.M.; Luo, S. Trends and Directions of Financial Technology (Fintech) in Society and Environment: A Bibliometric Study. Appl. Sci. 2021, 11, 10353. [Google Scholar] [CrossRef]
  47. Khushi, M.; Shaukat, K.; Alam, T.M.; Hameed, I.A.; Uddin, S.; Luo, S.; Yang, X.; Reyes, M.C. A Comparative Performance Analysis of Data Resampling Methods on Imbalance Medical Data. IEEE Access 2021, 9, 109960–109975. [Google Scholar] [CrossRef]
  48. Han, H.; Wang, W.Y.; Mao, B.H. Borderline-SMOTE: A new over-sampling method in imbalanced data sets learning. In Proceedings of the International Conference on Intelligent Computing, Hefei, China, 23–26 August 2005; Springer: Berlin/Heidelberg, Germany, 2005; pp. 878–887. [Google Scholar]
  49. Last, F.; Douzas, G.; Bacao, F. Oversampling for Imbalanced Learning Based on K-Means and SMOTE. Inf. Sci. 2017, 465, 1–20. [Google Scholar] [CrossRef]
  50. Douzas, G.; Bacao, F.; Sciences, F.L.-I. Improving imbalanced learning through a heuristic oversampling method based on k-means and SMOTE. Inf. Sci. 2018, 465, 1–20. [Google Scholar] [CrossRef]
  51. Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  52. Friedman, J.H. Stochastic gradient boosting. Comput. Stat. Data Anal. 2002, 38, 367–378. [Google Scholar] [CrossRef]
  53. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd Acm Sigkdd International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016. [Google Scholar]
  54. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.Y. LightGBM: A highly efficient gradient boosting decision tree. Adv. Neural Inf. Process. Syst. 2017, 30, 3147–3155. [Google Scholar]
  55. Dorogush, A.V.; Ershov, V.; Gulin, A. CatBoost: Gradient boosting with categorical features support. arXiv 2018, arXiv:1810.11363. [Google Scholar]
  56. Shaukat, K.; Luo, S.; Varadharajan, V.; Hameed, I.A.; Xu, M. A Survey on Machine Learning Techniques for Cyber Security in the Last Decade. IEEE Access 2020, 8, 222310–222354. [Google Scholar] [CrossRef]
  57. Shaukat, K.; Luo, S.; Varadharajan, V.; Hameed, I.A.; Chen, S.; Liu, D.; Li, J. Performance Comparison and Current Challenges of Using Machine Learning Techniques in Cybersecurity. Energies 2020, 13, 2509. [Google Scholar] [CrossRef]
  58. Shaukat, K.; Luo, S.; Varadharajan, V. A novel method for improving the robustness of deep learning-based malware detectors against adversarial attacks. Eng. Appl. Artif. Intell. 2022, 116, 105461. [Google Scholar] [CrossRef]
  59. Nasir, A.; Shaukat, K.; Khan, K.I.; Hameed, I.A.; Alam, T.M.; Luo, S. What is core and what future holds for blockchain technologies and cryptocurrencies: A bibliometric analysis. IEEE Access 2020, 9, 989–1004. [Google Scholar] [CrossRef]
  60. Ibrar, M.; Muhammad, A.H.; Kamran, S.; Talha, M.A.; Khaldoon, S.K.; Ibrahim, A.H.; Hanan, A.; Suhuai, L. A Machine Learning-Based Model for Stability Prediction of Decentralized Power Grid Linked with Renewable Energy Resources. Wirel. Commun. Mobile Comput. 2022, 2022, 2697303. [Google Scholar] [CrossRef]
  61. Batool, D.; Shahbaz, M.; Shahzad Asif, H.; Shaukat, K.; Alam, T.M.; Hameed, I.A.; Ramzan, Z.; Waheed, A.; Aljuaid, H.; Luo, S. A Hybrid Approach to Tea Crop Yield Prediction Using Simulation Models and Machine Learning. Plants 2022, 11, 1925. [Google Scholar] [CrossRef] [PubMed]
  62. Alam, T.M.; Shaukat, K.; Hameed, I.A.; Khan, W.A.; Sarwar, M.U.; Iqbal, F.; Luo, S. A novel framework for prognostic factors identification of malignant mesothelioma through association rule mining. Biomed. Signal Process. Control. 2021, 68, 102726. [Google Scholar] [CrossRef]
  63. Shaukat, K.; Luo, S.; Chen, S.; Liu, D. Cyber threat detection using machine learning techniques: A performance evaluation perspective. In Proceedings of the 2020 International Conference on Cyber Warfare and Security (ICCWS), Norfolk, VI, USA, 20–21 October 2020; pp. 1–6. [Google Scholar]
  64. Shaukat, K.; Masood, N.; Khushi, M. A Novel Approach to Data Extraction on Hyperlinked Webpages. Appl. Sci. 2019, 9, 5102. [Google Scholar] [CrossRef]
  65. Kumar, M.R.; Vekkot, S.; Lalitha, S.; Gupta, D.; Govindraj, V.J.; Shaukat, K.; Alotaibi, Y.A.; Zakariah, M. Dementia Detection from Speech Using Machine Learning and Deep Learning Architectures. Sensors 2022, 22, 9311. [Google Scholar] [CrossRef]
  66. Alam, T.M.; Shaukat, K.; Khelifi, A.; Khan, W.A.; Raza, H.M.E.; Idrees, M.; Luo, S.; Hameed, I.A. Disease Diagnosis System Using IoT Empowered with Fuzzy Inference System. Comput. Mater. Contin. 2022, 70, 5305–5319. [Google Scholar] [CrossRef]
  67. Alam, T.M.; Shaukat, K.; Mahboob, H.; Sarwar, M.U.; Iqbal, F.; Nasir, A.; Hameed, I.A.; Luo, S. A Machine Learning Approach for Identification of Malignant Mesothelioma Etiological Factors in an Imbalanced Dataset. Comput. J. 2021, 65, 1740–1751. [Google Scholar] [CrossRef]
  68. Shaukat, K.; Rubab, A.; Shehzadi, I.; Iqbal, R. A socio-technological analysis of cyber crime and cyber security in Pakistan. Transylv. Rev. 2017, 1, 84. [Google Scholar]
  69. Shabbir, S.; Asif, M.S.; Alam, T.M.; Ramzan, Z. Early Prediction of Malignant Mesothelioma: An Approach Towards Non-invasive Method. Curr. Bioinform. 2021, 16, 1257–1277. [Google Scholar] [CrossRef]
  70. Latif, M.Z.; Shaukat, K.; Luo, S.; Hameed, I.A.; Iqbal, F.; Alam, T.M. Risk Factors Identification of Malignant Mesothelioma: A Data Mining Based Approach. In Proceedings of the 2020 International Conference on Electrical, Communication, and Computer Engineering (ICECCE), Istanbul, Turkey, 12–13 June 2020; pp. 1–6. [Google Scholar] [CrossRef]
Figure 1. Proposed Methodology.
Figure 1. Proposed Methodology.
Diagnostics 12 03138 g001
Figure 2. Visualisation of data before and after resampling: (a) Imbalanced dataset; (b) Balanced Dataset.
Figure 2. Visualisation of data before and after resampling: (a) Imbalanced dataset; (b) Balanced Dataset.
Diagnostics 12 03138 g002
Figure 3. Result of CatBoost models to feature selection undersampling technique.
Figure 3. Result of CatBoost models to feature selection undersampling technique.
Diagnostics 12 03138 g003
Figure 4. Result of CatBoost models to Oversampling and feature selection technique.
Figure 4. Result of CatBoost models to Oversampling and feature selection technique.
Diagnostics 12 03138 g004
Table 1. Data Transformation.
Table 1. Data Transformation.
NameDescriptionRange
Occupation
-
Farmer
-
Other
Categorical
Annual Income
-
Less than 10,000 = 0
-
Equal or greater than 10,000 = 1
Boolean
Height
Weight
BMI>18.5—Underweight0Categorical
18.5–24.5 Normal1
25.0–29.5 overweight2
≤30.0 Obese3
Viability/Development (disease)
-
Normal = 1
-
Poor = 2
-
Supernormal =3
Categorical
Nourishment
(nutrient level inside the body)
-
Well =1
-
Medium = 2
-
Bad = 3
-
Cachexia/severe = 4
Categorical
Diagnostic Evidence 1
-
Blood Test = 1
-
Stool test = 2
-
Blood + stool = 3
-
Proctoscopy = 4
Categorical
Diagnostic Evidence 2
-
Ascites (Grade 1, 2, 3) = 1–3
-
Non-Ascites = 4
Categorical
Prior Treatment
-
Yes =1
-
No = 0
Boolean
History of Splenectomy
-
Yes = 1
-
No = 0
Boolean
History of ascites
-
Yes = 1
-
No = 0
Boolean
Other Disease
-
No = 0
-
Cardiovascular = 1
-
Digestive = 2
-
Neuropsychiatric = 3
-
Respiratory = 4
-
Urinary = 5
-
Other = 6
-
Cardiovascular + Digestive = 12
-
Digestive + Urinary = 25
-
Digestive + other = 26
-
Urinary + other = 56
Categorical
Extent of Ascites
-
Mild to moderate = 0
-
Severe = 1
Boolean
Clinical Classification
-
Ascites = 1
-
Megalosplenia = 2
-
Colonic tumoroid proliferation = 3
-
Dwarfism and others = 4
Categorical
Type of treatment the patient
-
New patient (Schistosomiasis newly diagnosed)
-
Retreated patient (Improved after treatment and will be treated again next year)
-
Relapsed patient (clinically cured but need further treatment for disease recurrence)
Categorical
Means of Treatment
-
Medical
-
Surgical
-
Medical + Surgical
Categorical
Cost of Treatment
-
Less than 5876.9 yuan
-
Greater or equal to 5876.9 yuan
Categorical
Table 2. Selected features from multiple feature selection techniques.
Table 2. Selected features from multiple feature selection techniques.
OneRInformation GainReliefFGain RatioCorrelation
ViabilityCost of treatmentDiagnostic Evidence 1The extent of ascitesCost of treatment
Means of treatmentAnnual IncomeBMIViabilityAnnual Income
The extent of ascitesClinical classificationClinical classificationClinical classificationClinical classification
BMIHistory of splenectomyType of treating patientsMeans of treatmentHistory of splenectomy
Diagnostic Evidence 2ViabilityHistory of ascitesCost of treatmentViability
Diagnostic Evidence 1The extent of ascitesDiagnostic Evidence 2Annual IncomeMeans of treatment
NourishmentMeans of treatmentCost of treatmentHistory of splenectomyThe extent of ascites
Annual IncomeDiagnostic Evidence 1History of splenectomyOccupationDiagnostic Evidence 2
Table 3. Features Score.
Table 3. Features Score.
FeaturesScore
Cost of treatment0.07295
Annual Income0.04749
Clinical classification0.00604
History of splenectomy−0.05814
Viability−0.05815
Means of treatment−0.06133
The extent of ascites−0.07251
Diagnostic Evidence 2−0.07253
Diagnostic Evidence 1−0.07254
Type of treating patients−0.07434
Occupation−0.11227
BMI−0.11531
Nourishment−0.16426
History of ascites−0.17023
Prior treatment−0.20311
Other diseases−0.27979
Table 4. Dataset Resampling.
Table 4. Dataset Resampling.
Total InstancesStatusImbalanced Dataset InstanceAfter UndersamplingAfter Oversampling
4136Recovery123212322904
Death290412322904
Table 5. Random undersampling to each feature selection technique.
Table 5. Random undersampling to each feature selection technique.
PerformanceClassifierCorrelationInfo GainReliefFOneRGain Ratio
Accuracy (%)GB82.282.768.168.682.2
LGBoost82.282.267.464.881.8
XGBoost82.882.86766.482.4
CatBoost82.782.767.766.482.8
Precision (%)GB82.582.567.768.682.1
LGBoost82.182.167.166.381.1
XGBoost82.682.666.667.482.2
CatBoost82.582.567.467.182.6
Recall (%)GB82.582.567.468.882.1
LGBoost82.182.166.665.981.8
XGBoost82.682.666.367.282.3
CatBoost82.582.566.767.182.7
F1-Score (%)GB82.582.567.568.582.1
LGBoost82.182.166.864.781.7
XGBoost82.682.666.466.482.2
CatBoost82.582.466.866.482.6
Table 6. Results of Cluster Centroid to each feature selection technique.
Table 6. Results of Cluster Centroid to each feature selection technique.
PerformanceClassifierCorrelationInfo gainReliefFOneRGain Ratio
Accuracy (%)GB76.474.868.564.178.9
LGBoost78.177.968.763.678.6
XGBoost78.577.967.963.979.4
CatBoost78.778.16964.780
Precision (%)GB76.374.668.263.778.7
LGBoost77.977.768.563.978.4
XGBoost78.477.767.664.279.3
CatBoost78.777.968.76579.9
Recall (%)GB76.474.468.363.478.4
LGBoost77.777.968.66478.3
XGBoost7877.967.664.379.1
CatBoost78.37868.865.179.5
F1-Score (%)GB76.374.568.363.478.7
LGBoost77.877.868.563.678.4
XGBoost78.177.867.763.979.1
CatBoost78.477.9.68.864.779.7
Table 7. Results of Near Miss with respect to each feature selection technique.
Table 7. Results of Near Miss with respect to each feature selection technique.
PerformanceClassifierCorrelationInfo GainReliefFOneRGain Ratio
Accuracy (%)GB78.678.664.571.878.6
LGBoost78.178.168.372.177.5
XGBoost78.778.768.272.478.1
CatBoost78.978.968.671.878.1
Precision (%)GB78.978.966.771.778.9
LGBoost78.178.168.372.177.7
XGBoost79.179.168.172.478.3
CatBoost79.279.268.571.778.3
Recall (%)GB77.877.865.971.977.8
LGBoost77.477.468.472.476.8
XGBoost77.877.968.372.677.3
CatBoost78.178.168.771.977.3
F1-Score (%)GB78.178.164.471.778.1
LGBoost77.677.668.272.177
XGBoost78.278.268.172.377.5
CatBoost 78.378.368.571.777.5
Table 8. Results of models with respect to Random Oversampling and feature selection technique.
Table 8. Results of models with respect to Random Oversampling and feature selection technique.
PerformanceClassifierCorrelationInfo GainReliefFOneRGain Ratio
Accuracy (%)GB82.682.671.66882
LGBoost83.483.473.367.682.9
XGBoost83.483.47368.382.9
CatBoost83.783.772.967.882.9
Precision (%)GB82.682.672.16882
LGBoost83.483.474.167.982.9
XGBoost83.583.573.668.482.9
CatBoost83.783.773.167.982.9
Recall (%)GB82.682.671.56882
LGBoost83.483.473.267.782.9
XGBoost83.483.47368.382.8
CatBoost83.783.772.967.882.9
F1-Score (%)GB82.682.671.46882
LGBoost83.483.47367.582.9
XGBoost83.483.472.868.282.8
CatBoost83.683.672.867.882.9
Table 9. Results of the models with respect to SMOTE and feature selection techniques.
Table 9. Results of the models with respect to SMOTE and feature selection techniques.
PerformanceClassifierCorrelationInfo GainReliefFOneRGain Ratio
Accuracy (%)GB82.482.474.869.782.2
LGBoost83.183.176.170.982..9
XGBoost83.383.375.870.782.8
CatBoost82.982.975.971.082.7
Precision (%)GB82.482.475.069.782.2
LGBoost83.283.276.471.582.9
XGBoost83.383.376.171.382.4
CatBoost82.982.976.271.882.7
Recall (%)GB82.482.474.769.782.2
LGBoost83.183.176.170.982.9
XGBoost83.383.375.870.882.8
CatBoost82.982.975.871.182.7
F1-Score (%)GB82.482.474.769.782.2
LGBoost83.183.176.070.782.9
XGBoost83.383.375.770.682.8
CatBoost82.982.975.870.882.7
Table 10. Results of the models with respect to K means SMOTE and feature selection techniques.
Table 10. Results of the models with respect to K means SMOTE and feature selection techniques.
PerformanceClassifierCorrelationInfo GainReliefFOneRGain Ratio
Accuracy (%)GB86.076.974.877.686.2
LGBoost86.779.276.177.986.4
XGBoost86.979.176.078.086.6
CatBoost87.179.177.078.486.8
Precision (%)GB86.077.074.878.086.4
LGBoost87.079.376.877.986.6
XGBoost87.279.176.678.086.9
CatBoost87.179.377.778.487.0
Recall (%)GB85.977.074.877.686.2
LGBoost86.779.276.177.986.3
XGBoost86.979.176.078.086.6
CatBoost87.179.177.078.486.7
F1-Score (%)GB85.976.974.877.586.2
LGBoost86.779.275.977.986.3
XGBoost86.979.175.978.086.6
CatBoost87.179.176.878.486.7
Table 11. Comparison of previous studies.
Table 11. Comparison of previous studies.
ReferenceYearObjectiveApproachUndersampling TechniqueOversampling TechniqueOutcome
[27]2017To check the pattern analysis of SchistosomiasisdiseaseK-mean algorithm, clusteringN.A
[1]2018To detect advanced Schistosomiasis in the people of Hubei provinceANN, DT, LR80%
[21]2019To detect and predict the diagnosis of disease on an imbalanced datasetPNN, RF, one rule, and DT82%
[28]2019Detection of disease factorAssociation Rule MiningN.A
[29]2019Relationship between climate change and disease factorK- means clusteringN.A
[18]2020Comparison to check the level of susceptibility of SchistosomiasisSVM, Gaussian Methods76.6–94%
[17]20211 year prognosis for advanced SchistosomiasisLR, RF, DT, ANN, XGBoost79%
[30]2021Identify the high-risk areas of SchistosomiasisLR, RF, GB73–87%
Our study2022Early prediction of SchistosomiasisGB, LGBoost, XGBoost, CatBoost87.1%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ali, Z.; Hayat, M.F.; Shaukat, K.; Alam, T.M.; Hameed, I.A.; Luo, S.; Basheer, S.; Ayadi, M.; Ksibi, A. A Proposed Framework for Early Prediction of Schistosomiasis. Diagnostics 2022, 12, 3138. https://doi.org/10.3390/diagnostics12123138

AMA Style

Ali Z, Hayat MF, Shaukat K, Alam TM, Hameed IA, Luo S, Basheer S, Ayadi M, Ksibi A. A Proposed Framework for Early Prediction of Schistosomiasis. Diagnostics. 2022; 12(12):3138. https://doi.org/10.3390/diagnostics12123138

Chicago/Turabian Style

Ali, Zain, Muhammad Faisal Hayat, Kamran Shaukat, Talha Mahboob Alam, Ibrahim A. Hameed, Suhuai Luo, Shakila Basheer, Manel Ayadi, and Amel Ksibi. 2022. "A Proposed Framework for Early Prediction of Schistosomiasis" Diagnostics 12, no. 12: 3138. https://doi.org/10.3390/diagnostics12123138

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop