Next Article in Journal
Video Capsule Endoscopy Plays an Important Role in the Management of Crohn’s Disease
Next Article in Special Issue
Usefulness of Heat Map Explanations for Deep-Learning-Based Electrocardiogram Analysis
Previous Article in Journal
The Prevalence of Multidrug-Resistant Enterobacteriaceae among Neonates in Kuwait
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Polycystic Ovary Syndrome Detection Machine Learning Model Based on Optimized Feature Selection and Explainable Artificial Intelligence

1
Department of Information Technology, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
2
Machine Learning and Information Retrieval Department, Faculty of Artificial Intelligence, Kafrelsheiksh University, Kafrelsheiksh 13518, Egypt
3
Faculty of Information Technology, Applied Science Private University, Amman 11937, Jordan
4
Department of Information Systems, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
5
Faculty of Computers and Informations, South Valley University, Qena 83523, Egypt
6
Faculty of Computer Science and Engineering, Galala University, Suez 435611, Egypt
7
Information Systems Department, Faculty of Computers and Artificial Intelligence, Benha University, Banha 13518, Egypt
8
Faculty of Computers and Artificial Intelligence, South Valley University, Hurghada 84511, Egypt
*
Authors to whom correspondence should be addressed.
Diagnostics 2023, 13(8), 1506; https://doi.org/10.3390/diagnostics13081506
Submission received: 18 March 2023 / Revised: 13 April 2023 / Accepted: 15 April 2023 / Published: 21 April 2023

Abstract

:
Polycystic ovary syndrome (PCOS) has been classified as a severe health problem common among women globally. Early detection and treatment of PCOS reduce the possibility of long-term complications, such as increasing the chances of developing type 2 diabetes and gestational diabetes. Therefore, effective and early PCOS diagnosis will help the healthcare systems to reduce the disease’s problems and complications. Machine learning (ML) and ensemble learning have recently shown promising results in medical diagnostics. The main goal of our research is to provide model explanations to ensure efficiency, effectiveness, and trust in the developed model through local and global explanations. Feature selection methods with different types of ML models (logistic regression (LR), random forest (RF), decision tree (DT), naive Bayes (NB), support vector machine (SVM), k-nearest neighbor (KNN), xgboost, and Adaboost algorithm to get optimal feature selection and best model. Stacking ML models that combine the best base ML models with meta-learner are proposed to improve performance. Bayesian optimization is used to optimize ML models. Combining SMOTE (Synthetic Minority Oversampling Techniques) and ENN (Edited Nearest Neighbour) solves the class imbalance. The experimental results were made using a benchmark PCOS dataset with two ratios splitting 70:30 and 80:20. The result showed that the Stacking ML with REF feature selection recorded the highest accuracy at 100 compared to other models.

1. Introduction

Polycystic ovary syndrome (PCOS) affects pregnant women and current mothers. PCOS affects the health of women by causing hormonal imbalances and metabolism problems. It is a disease that primarily affects women’s fertility, as 5 to 10% of females suffer from this disease in their childbearing years (15–45) [1]. It is a hormonal disorder that causes problems with the ovaries. In the normal state, the ovaries produce hormones (chemicals that control the functions of the body), namely estrogen (female hormone) and androgens (male hormones), for normal health [2]. In affected women, the hormones are imbalanced, with higher androgens or less estrogen than normal. This causes lumps (fluid-filled sacs) to grow on the ovaries. These lumps gradually enlarge and then obstruct the ovulation process. This disruption of ovulation in women with PCOS reduces their chances of becoming pregnant [3]. Women with PCOS are more likely to develop diabetes, heart disease, high blood pressure, endometrial thickness, sleep apnea, depression, anxiety, eating disorders, and endometrial cancer [4]. In addition to genetic factors, environmental factors may also contribute to PCOS development. In addition to early diagnosis, treatment, and weight loss, long-term complications can be reduced [5].
Artificial intelligence (AI) has revolutionized the detection and treatment of diseases, specifically PCOS [6]. AI-based technologies such as machine learning (ML) algorithms and deep learning networks (DL) have enabled the development of automated systems for the accurate and reliable detection of heart disease [7,8]. AI-based methods can identify patterns in medical data, such as hormone levels, to distinguish PCOS patients from those without the disorder. This improved accuracy could lead to earlier, more accurate diagnoses and better overall outcomes for PCOS patients. Furthermore, AI-based systems can be used to monitor patients over time, providing clinicians with valuable insights into potential treatments and enabling more precise interventions. In short, AI-driven technology has the potential to revolutionize PCOS detection and treatment, providing more effective and efficient care for those suffering from the condition.
Feature selection [9] reduces the number of input variables when developing a predictive model. The goal of feature selection approaches in ML [10] is to find the best features to build effective models of the studied phenomena. It involves automatically selecting features for your ML model pertinent to the problem you are attempting to solve. We accomplish this by adding or removing significant features without altering them. It assists in minimizing the amount of noise in our data and the quantity of our input data. A hybrid model combines two or more different models or strategies to address a challenge or accomplish a goal [11]. A hybrid model can be used in machine learning to combine various algorithms or strategies to increase a model’s performance and accuracy [12]. For instance, a hybrid model can improve both the accuracy and efficiency of neural networks and decision trees by combining both qualities. Combining statistical and rule-based models is another illustration of a hybrid model.
In ML, there is always a tradeoff between the complexity and performance of the developed model. A simple model (i.e., linear regression) could be more interoperable and provide a more understandable explanation than complex ML and DL models [13]. Therefore, providing a clear explanation of such a complex model is a significant point in increasing trust in the developed model. Explainability is motivated by the lack of model transparency of complex (black-box) models that lack model trust [14]. Explainable AI (XAI) techniques improve model predictions’ understanding, interpretability, and reliability. Explainability has two primary levels, i.e., local explainability and global explainability. Global explainability explains the final decision at the level of all data points. It provides casual analysis in terms of global fidelity. It only explained the instance level with the importance of such a level [15]. Local allegiance could explain in terms of all samples. It provides a more accurate explanation.
Our research aims to provide model explanations to ensure efficiency, effectiveness, and trust in the developed model through local and global explanations. Feature selection methods with different types of ML models and proposed are proposed to predict PCOS. The following are the main contributions of the suggested work:
  • A combination of SMOTE (Synthetic Minority Oversampling Techniques) and ENN (Edited Nearest Neighbour) solves the class imbalance.
  • Applying feature selection (FS) to reduce data dimensionality and select the optimal feature set.
  • Applying Bayesian Optimization with cross-validation to optimize ML algorithms and enhance accuracy.
  • Proposing stacking ML and comparing it with different ML models using evaluation methods, including accuracy (Acc), precision (P), recall (R), F1 score (F1), and area under the receiver operating characteristic (ROC) (AUC) curve.
  • Increasing the model trust by clearly explaining the final prediction using global and local explainability terms.
This paper is organized as follows, Section 2 summarizes the related work in the PCOS domain. The dataset and proposed methodology can be described in Section 3. Section 4 shows the results. Section 5 shows the discussion, including a comparison with related work and model explainability. The paper concludes in Section 6.

2. Related Work

The authors applied ML models to PCOS from Kaggle to predict PCOS. For example, in [16], the authors applied gradient boosting, RF, LR, and a hybrid RFLR model that integrated RF with LR with a univariate feature selection (UFS) algorithm from the PCOS dataset. They split the dataset using holdout and cross-validation methods to train and test models. The result showed that RFLR with UFS achieved the highest performance.
In [17], the authors reduced the number of features using Principal Component Analysis (PCA). They applied NB, KNN, LR, RF, and SVM with selected features to predict PCOS. The result showed that RF achieved the highest accuracy. In [6], the authors used correlation feature selection methodology to select a subset of features from the database. They applied different ML models: SVM, LR, RF, DT, KNN, Quadratic Discriminant Analysis (QDA), Linear Discriminant Analysis (LDA), GB, AdaBoost (AB), XGBoost (XB), and CatBoost, and obtained the optimal model based on correlation thresholds. The result showed that RF was the optimal model.
In [18], the authors compared different models, i.e., CNN, ANN, SVM, DT, and KNN, and applied feature selection methods to diagnose PCOS. RF achieved the best-performing model. In [19], the authors utilized Pearson correlation to determine the best features. The applied SVM, RF, and XG boost multi-layer perceptron with selected features to detect the accuracy rate of their SVM have the highest rate. In [20], the authors proposed a hybrid feature selection approach using filters and wrappers to reduce the number of features. Furthermore, they applied different ML models with selected features to predict PCOS. SVM achieved the highest accuracy.
In [21], they applied SVM, LR, NB, and KNN to detect whether a woman was suffering from PCOS. They used chi-square feature selection methods to select the top 30 features. The accuracy of RF has achieved the highest rate. In [16], the authors used RF, DT, SVM, LR, KNN, XGBRF, and CatBoost Classifier to detect whether a woman was suffering from PCOS. The result showed that CatBoost recorded the highest accuracy.
In [22], the authors used Gini importance to select features. They applied different ML models: KNN, DT, SVM, LR, and NB, to detect PCOS. Based on the accuracy, DT recorded the highest rate. In [23], the authors applied CatBoost, RF, LR, NB, DT, SVM, and DT. Furthermore, they compared their outcomes in terms of the evaluation matrix. CatBoost has the highest accuracy in predicting whether a woman should seek medical help for PCOS. In [24], the authors applied Chi-Square, ANOVA, and Mutual Information to identify insignificant features from the data. They used selected features to detect PCOS by applying SVM, LR, DT, NB, XGBRF, RF, and CatBoost. The CatBoost classifier performed with the best accuracy.
In [25], the authors used ML models: LR, DT, RF, SVM, NB, KNN, AdaBoost, XGBoost, and Extratrees and DL and proposed multi-stacking ML to predict PCOS. They used Explainable AI (XAI) techniques to make model predictions understandable, interpretable, and trustworthy. The result showed that multi-stacking ML recorded the best performance.

3. Methodology

We applied different ML models: SVM, NB, LR, KNN, RF, DT, XGboost, and AdaBoost, with FS methods to predict PCOS. We proposed Stacking ML models that combine the best ML models. Figure 1 shows the phases of prediction PCOS.

3.1. Database Description

We used the PCOS dataset from Kaggle [26], which includes 541 instances and 41 attributes. There are 178 instances of the positive class (1) and 363 instances of the negative class (3). The dataset has a mismatched distribution of classes. The dataset includes two files—we merged two files: PCOS_infertility and PCOS_data_without_infertility, and deleted redundant columns. Table 1 shows the details of the database features.

3.2. Data Processing

This stage aims to enhance the quality of the utilized dataset, as it include several missing values and outliers. Medical datasets commonly suffer from such issues due to various causes, including device failure, network loss, irregular time recording, etc. Unfortunately, Several ML models are sensitive to outliers; most cannot handle missing values. Data preprocessing include filling in missing data and data encoding.

Filling Missing Values

Many statistical approaches exist to deal with missing data, but it mainly depends on how much data are missing and the importance of the feature missing [27]. When the fraction of the missing data is between 5% and 10%, traditional statistical approaches, such as mean, max, and mode, work exceptionally well. When the fraction of missing values is 20–50%, sophisticated approaches, such as hot-deck [28] and expectation maximization [29], are appropriate. To ensure data reliability in our used data, we choose to remove features with more than 30% missing values. Features with missing values that are less than 30% are imputed using feature means. Remove columns that include many null values: BMI, FSH/LH, and Waist:Hip Ratio. Furthermore, we drop Sl. No, Patient File No. Columns. Filling NA values with the median of that feature: Marraige Status (Yrs), II beta-HCG (mIU/mL), AMH (ng/mL), and Fast food (Y/N).

3.3. Data Encoding

Categorical and numeric features are combined in the utilized dataset. Numeric features perform better with ML and DL than categorical ones, unfortunately. Therefore, we encoded all categorical features using the label encoder module of the Scikit-learn library.

3.4. Sampling Data

We used SMOTEENN to re-sampling data. The SMOTE-ENN method combines the SMOTE and ENN techniques. SMOTE is an oversampling method, and ENN is an edited closest neighbor undersampling method (ENN). In the ENN approach, the observation and its KNN are removed from the dataset if the majority class of the observation’s KNN and the observation’s class are different. Due to this, information about the minority class in the majority class is lost. By doing this, the bias towards the majority class is lessened, which enhances the performance of machine learning models [30].

3.5. Feature Selection Techniques

An optimal feature subset is determined by feature selection (FS), which removes irrelevant features to increase learning accuracy [31,32]. The feature subset is chosen from the original feature set based on feature relevance and redundancy. As shown in Figure 2, FS is categorized into three main types according to the interaction with the utilized model: filter approach, wrapper approach, and embedded approach. The following subsection details the different approaches of FS. Our study used one method of each type, such as mutual information-based, REF, and tree based (RF).

3.5.1. Filter Approach

The filter approach utilizes statistical tests to score all features and select the best, independent of the learning algorithm a mutual information-based [33], correlation coefficient [34], and the Chi-square test (Chi2) [7].

3.5.2. Wrapper Approach

The wrapper approach mainly depends on the performance of the learning algorithm. The chosen feature subsets estimate the model performance. According to the model performance, the algorithm adds or removes features until the optimal feature subset is reached. It is more computationally expensive than the filter approach because it repeats the learning and evaluation process. However, it is considered to be more accurate and efficient than the filter approach. The best feature subset is mainly chosen based on the classifier performance. Sequential feature selection [35] and recursive feature elimination (RFE) [36] is an example of this approach.

3.5.3. Embedded Approach

The third method is the embedded approach. This approach uses both ensemble and hybrid learning to make FS. It works by choosing the best features during the learning process. Selecting the optimum feature subset chosen during the training process takes advantage of enhancing computational cost. Since it depends on a collective decision, its performance is better than the filter and wrapper approach regarding computational cost and classification accuracy. Several techniques developed in terms of embedded FS include tree based (RF) and Relevant Sample-Feature Machine (RSFM).

3.6. Splitting Dataset

The PCOS was split into two sets using a stratified sampling method, i.e., a ratio of 80% training set and 20% testing sets and a ratio of 70% training set and 30% testing set. Training sets are used to train and optimize models; the testing sets are used to evaluate models.

3.7. Models Optimization and Training

Bayesian Optimization (BO) is used to optimize diffident ML models using training sets and cross-validation.

3.7.1. ML Models

We used different ML models, namely logistic regression (LR) [37], random forest (RF) [38], decision tree (DT), naive Bayes (NB) [39], support vector machine (SVM) [40], k-nearest neighbor (KNN) [41], Xgboost [42], and the Adaboost algorithm [43].

3.7.2. Bayesian Optimization

Hyperparameter optimization techniques aim to find the optimum hyperparameter that gives the best performance on a validation set [44]. It can be represented with the following Equation [44]:
x = arg min c X f ( x )
where x is the optimum hyperparameter list that will give the best performance, f ( x ) is the objective that needs to be minimized, such as the error rate evaluated based on the validation set, and c represents any value in the x domain [44].
Using uniform hyperparameter optimization such as grid search and random search gives enhanced performance over a manual search. It starts with a list of values for each hyperparameter and runs a train-predict-evaluate loop. The problem with this approach is that it is completely uniform and does not consider the previous evaluation. Therefore, it could take significant time to evaluate bad hyperparameters. In contrast, BO considers past performance when building a probability model of the objective function [45].
This model is known as a “surrogate” that could represent (p(Y|X). This model works by finding the next list of hyperparameters that perform best according to the surrogate function.

3.8. Stacking Machine Learning

The ensemble model builds on combining decisions from several models to improve the model’s overall performance. This approach enhances performance over a single model [46,47]. Bagging, boosting, and staking are the most popular ensemble techniques. Stacking is an ensemble technique that combines different classifications through a meta-classifier [48]. The base model (base classifiers) is trained on the dataset, after which it meta-learns the features that are out of the base classifiers. Therefore, stacking is considered to be one of the more sophisticated heterogeneous classifiers. The architecture of the stacking model includes two or more base models called base–learning, and level-2 is the meta-learning layer that combined the base model’s prediction. Figure 1 shows the general architecture of the stacking ensemble model.

3.9. Evaluating Models

As illustrated in Equations (1)–(4), the models are evaluated using four methods: accuracy, precision, recall, and F-score, where TP indicates true positive, TN indicates true negative, FP indicates false positive, and FN indicates false negative:
A c c u r a c y = T P + T N T P + F P + T N + F N .
P r e c i s i o n = TP TP + FP
R e c a l l = T P T P + F N
F-score = 2 · p r e c i s i o n · r e c a l l p r e c i s i o n + r e c a l l
Furthermore, the models evaluated by the ROC [49] curve is a graphical representation of the performance of a binary classification model. FPR is shown at different classification thresholds. A positive TPR represents the percentage of positive cases that are correctly classified as positive, while a negative FPR represents a percentage of negative cases that are incorrectly classified as positive; AUC (Area Under the Curve) [49] is a measure used to evaluate the performance of a binary rating model, and measures the area under the receiver operating characteristic (ROC) curve, with different rating thresholds, i.e., TPR versus FPR.

4. Experimental Results

4.1. Experiment Setup

This section presents and discusses the experimental results. Scikit-learn was used to develop the ML models. Google Colab was used to conduct the experiments. Furthermore, the stacking ML models were compared with different ML models based on various feature selection methods (RFE, tree based, and mutual_info). The performance of the models is recorded with two ratios of 20:80 and 30:70 training and testing sets.

4.2. Feature Selection Methods

These experiments investigate the essential features of feature selection methods applied to the PCOS dataset.

4.2.1. Scores of Selected Features by Mutual_Info

After applying mutual_info to the dataset, the score of each feature is shown in Figure 3. We can see that FL_R has the highest score at 0.33584, and FN_L has the second-highest score at 0.317744. Beta_I, MS, and Cycle have approximately scores of 0.1447437, 0.143511, and 0.1412555, respectively. Vit_D3, PRL, RE, HL, Waist, and TSH have the lowest scores. Aborptions and Pregnant have zero sores. We selected the 30 highest features for applying ML models.

4.2.2. Importance of Selected Features by Tree Based

Figure 4 shows importance of features that are selected features by based tree. FL_R has the highest importance at 0.189997, and FN_L has the second-highest score at 0.176050. CL and AMH have approximate importance of 0.067357 and 0.06720, respectively. RE, Pregnant, HL, and Pimples have the lowest score. We selected the 30 highest features to apply to the ML models.

4.2.3. Ranking of Selected Features by RFE

Figure 5 shows REF’s ranking of features; 30 top features have a ranking of 1, such as Age, Weight, Height, BG, PR, HB, Cycle, CL, and MS. The worst features are aborptions and BP_Diastolic, which have a ranking of 5.

4.3. Performance of the Classifiers with Selected Features Using Splitting 80:20

This subsection presents the experimental results of selected features by mutual_info, RFE, and tree based, which are used to train and evaluate the various classifiers with 80:30 splitting. These results are summarized in Table 2. The different classifiers’ AUC values and ROC curves are shown in Figure 6.
Overall, Stacking ML with RFE achieved the highest ACC, PRE, REC, F1, and AUC. For Info_mun, Stacking ML combined the best models to obtain the final prediction and improve performance by 1%, with the highest AUC, ACC, PRE, REC, and F1 of 99, 98.48, 98.41, 98.42, and 98.81, respectively, compared to other models. XGB demonstrated the second-best performance. As we can observe, NB and KNN performed similarly (ACC = 95.24, PRE = 95.56, REC = 95.24 and F1 = 95.14, AUC = 92.86).
For RFE, Stacking ML combined the output of the best models: RF, NB, XGB, and AdaBoost, to obtain the final prediction and improve performance by 1.5 with the highest AUC, ACC, PRE, REC, and F1 of 100, 100, 98.41, 100, and 100, respectively. As we can observe, RF, NB, XGB, and AdaBoost demonstrate the second-best performance (ACC = 98.41, PRE = 98.45, REC = 98.41, and F1 = 98.40, AUC = 97.62). KNN recorded the lowest performance (ACC = 93.65, PRE = 93.99, REC = 93.65 and F1 = 93.72, AUC = 94.05).
For tree based, Stacking ML combined the best models, i.e., RF, SVM, XGB, and AdaBoost, to obtain the final prediction and improve performance by 1.5 with the highest ACC, PRE, REC, F1, and AUC of 97.41, 97.45, 97.41, 97.4, and 97.62, respectively. As we can observe, RF, SVM, XGB, and AdaBoost demonstrate the second-best performance: (ACC = 96.83, PRE = 96.83, REC = 96.83, F1 = 96.83, AUC = 96.43). KNN recorded the lowest performance (ACC = 93.65, PRE = 93.65, REC = 93.65, F1 = 93.65, AUC = 92.86).

4.4. Performance of the Classifiers with Selected Features Using Splitting 70:30

This subsection presents experimental results of selected features by mutual_info, RFE, and tree based are used to train and evaluate the various classifiers with 70:20 splitting. These results are summarized in Table 3. The different classifiers’ AUC values and ROC curves are also shown in Figure 7. Overall, Stacking ML with RFE achieved the highest ACC, PRE, REC, F1, and AUC.
For mutual_info, Stacking ML combined the best models to obtain the final prediction and improve performance by 1%, with the highest ACC, PRE, REC, and F1 of 96.81, 96.81, 96.81, and 96.80, respectively, compared to other models. AdaBoost obtained the second-best performance for ACC, PRE, REC, and F1, i.e., 95.74, 95.74, 95.74, and 95.74, respectively. NB registered the lowest ACC, PRE, REC, and F1, i.e., 79.79, 82.15, 79.79, and 80.24, respectively.
For RFE, Stacking ML combined the best models to obtain the final prediction and improve performance by 1%, with the highest ACC, PRE, REC, and F1 of 98.87, 98.00, 98.87, and 98.89, respectively, compared to other models. AdaBoost obtained the second-best performance for ACC, PRE, REC, and F1, i.e., 96.81, 96.86, 96.81, and 96.82, respectively. NB registered the lowest ACC, PRE, REC, and F1, i.e., 85.11, 85.39, 85.11, and 85.21, respectively.
For tree based, Stacking ML combined the best models to obtain the final prediction and improved performance by 1%, with the highest ACC, PRE, REC, and F1 of 97.81, 97.81, 97.81, and 97.8, respectively, compared to other models. AdaBoost and SVM obtained the second-best performance for ACC, PRE, REC, and F1, i.e., 96.81, 96.81, 96.81, and 96.80, respectively. NB registered the lowest ACC, PRE, REC, and F1, i.e., 87.23, 87.89, 87.23, and 87.40, respectively.

5. Discussion

A summary of the experimental results is presented in this section. Additionally, we discuss which model is best for each method of selecting features. The proposed model is also compared to previous studies. Furthermore, model explainability is discussed.

5.1. The Best Models

Overall, Stacking ML with RFE achieved the highest ACC, PRE, REC, F1, and AUC. Figure 8 shows the best models for each of the 20–80 feature selection methods. Stacking ML with RFE achieves the highest percentages of different evaluation metrics at 100. Stacking ML with tree based recorded has the lowest ACC, PRE, REC, and F1 performance at 97.41, 97.45, 97.41, and 97.4, respectively.
Figure 9 shows the best models for each of the 20–80 feature selection methods. Stacking ML with RFE achieved the highest percentages of different evaluation metrics at ACC, PRE, REC, and F1 at 98.87, 98.00, 98.87, and 98.89, respectively. Stacking ML with mutual_info recorded the lowest performance of ACC, PRE, REC, F1, and AUC at 96.81, 96.81, 96.8, and 96.42, respectively.

5.2. Comparison with Previous Studies

Table 4 compared previous studies and the proposed model. We can see that our work achieved the highest ACC compared to other studies. In [16], the authors proposed an RFLR hybrid model, applied it with UFS, and achieved an ACC of 91.01. In [17], PCA with RF recorded an ACC of 89.02. In [6], RF with correlation recorded an ACC of 92.4. In [19], SVM with Pearson correlation recorded an ACC of 91.6. In [20], ACC was 91.6 SVM with hybrid feature selection. In [21], RF with chi-square recorded an ACC of 90.9. In [22], DT with Gini importance recorded an ACC of 92.59. In [25], multi-stack of ML recorded an ACC of 98.

5.3. Model Explainability

Explainability has two primary levels, i.e., local explainability and global explainability. Global explainability explains the final decision at the level of all data points. It provides casual analysis in terms of global fidelity. It only explained the instance level with the importance of such a level [15]. Local fidelity could explain in terms of all samples. It provides a more accurate explanation. To identify causality and description of the best model (Stacking ML with RFE), in this section, we describe the final decision of the output in terms of global explainability (at the level of a dataset) and local explainability (instance level).

5.3.1. Global Explainability

Figure 10a shows the bar plot of the feature importance of each feature with the developed model; in other words, it displays the collective contribution of the features and the less critical features. Figure 10b shows the cohort plot, which divides the total test data into two groups according to the most affected features in all data. As shown in Figure 4, the total data are divided into two main groups according to the number of follicles. The total samples were divided into two groups according to the optimal threshold. Follicles number = 6.5. The bar plot shows that the most affected reason that the instance belongs to the hormonal disorder class is that FL_R (SHAP = 0.09), cycle (SHAP = +0.14), and age (SHAP = 0.07). To provide more information, Figure 5 shows the heat map that shows the importance of the variables in terms of a horizontal bar that shows the rank of the variables from highest to lowest. This importance explains the global interoperability of the developed model. Our developed model depends on FN_L, cycle, and FN_R as the three most important features that affect the overall decision. Sample 32 has a high prediction, which means that FN_L has a significant effect on the prediction. The heatmap in Figure 11 shows the number of instances in the test data in the x_axis, and the curve of F(x) above the plot shows the model prediction for the cases. The observations are also arranged in a way that colors are collected together.

5.3.2. Local Explainability

Many methods are utilized to explain the prediction at the instance level. In this section, we provide several methods to explain the model in terms of instances, such as force plot, water full plot, and summary plot. First, the waterfall plot clarifies why the instance receives the developed value prediction. Figure 12 shows the prediction of the first observation. The prediction for the first observation was 1; this value ranges from (0.665 + 0.2 + 0.09 + 0.6 − 0.03 + 0.01 + 0.01). The values that bedside the variable name refer to the value of the instance feature FN_l = 9, age = 24, etc., and the number in the arrows shows how these features positively or negatively contribute to the decision as shown in Figure 12. The same is true in Figure 12b, which shows the prediction for observation 2 and the values for each feature, and how each feature contributes to the final decision. Second, the force plot explains the key factors. As shown in Figure 13a,b, the plot states that the final prediction of the observation’s higher score led the model to predict 1. In the figure, the bold score was 0.87 for that observation. This means that the observation is highly correlated with class 1. Features with a red color in the horizontal line represent the features that push the model towards a high score, while blue represents features that make the model move towards a low score. Features significantly impacting the final prediction are closer to the dividing boundary between the red and blue areas. All features are sorted from more important to less in the horizontal line in red and blue areas. The same is true for observation 2. The total score was 0. It belongs to class 0 vit_D, and PRG is the feature that pushes the prediction to a high score, and all other features, such as FN_L, FN_R, age, etc., push the prediction to a low score. Figure 14 shows the collective force plot for the developed model. A collective force plot treats the same way as an individual force plot; rotate all samples 90 degrees and add them together.

6. Conclusions

The main objective of our paper is to provide an early detection model for PCOS. The early detection of PCOS reduces the possibility of long-term complications. Several ML utilized to build the proposed stacking ensemble ML model. It combines diverse ML models (LR, RF, DT, NB, SVM, KNN, xgboost, and Adaboostare) at the base learner level with RF at the meta-learner level is proposed to improve the performance of a single ML. The following steps apply to build the proposed model: (1) SMOTEENN is applied to the PCOS dataset to solve the class imbalance; (2) Feature selection methods (RFE, tree bases, and mutual info) are applied to select the optimal subset of features; (3) Bayesian optimization finds the optimum hyperparameter that performs best on a validation set; (4) Data are split using two ratios, 70:30 and 80:20; (5) The stacking ensemble model is built with several ML in the base learner level and RF meta learner. The result showed that the Stacking ML with REF feature selection recorded the highest performance at 100 compared to other models with 80:20. It achieved the highest percentages of different evaluation Metrics at ACC, PRE, REC, and F1 at 98.87, 98, 98.87, and 98.89, respectively. To ensure model trust, efficiency, and effectiveness, our research also provides model explanations both at the model level (global explanation) and at the instance level (local explanation).

Author Contributions

Methodology, N.E.-R. and H.S.; Writing—original draft, N.E.-R., S.F. and S.E.-S.; Writing—review & editing, H.E., N.E.-R., I.M., M.A.A., S.F., S.E.-S. and H.S.; Visualization, H.S. All authors have read and agreed to the published version of the manuscript.

Funding

Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2023R330), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Data Availability Statement

The direct link in the dataset citations will take you to all of the datasets that were utilized to support the study’s assertions.

Acknowledgments

The authors would like to thank Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2023R330), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Conflicts of Interest

All authors declare that they have no conflict of interest.

References

  1. Escobar-Morreale, H.F. Polycystic ovary syndrome: Definition, aetiology, diagnosis and treatment. Nat. Rev. Endocrinol. 2018, 14, 270–284. [Google Scholar] [CrossRef] [PubMed]
  2. Norman, R.J.; Dewailly, D.; Legro, R.S.; Hickey, T.E. Polycystic ovary syndrome. Lancet 2007, 370, 685–697. [Google Scholar] [CrossRef] [PubMed]
  3. McCartney, C.R.; Marshall, J.C. Polycystic ovary syndrome. N. Engl. J. Med. 2016, 375, 54–64. [Google Scholar] [CrossRef] [PubMed]
  4. Barber, T.M.; Franks, S. Obesity and polycystic ovary syndrome. Clin. Endocrinol. 2021, 95, 531–541. [Google Scholar] [CrossRef] [PubMed]
  5. Azziz, R. Polycystic ovary syndrome. Obstet. Gynecol. 2018, 132, 321–336. [Google Scholar] [CrossRef]
  6. Tiwari, S.; Kane, L.; Koundal, D.; Jain, A.; Alhudhaif, A.; Polat, K.; Zaguia, A.; Alenezi, F.; Althubiti, S.A. SPOSDS: A smart Polycystic Ovary Syndrome diagnostic system using machine learning. Expert Syst. Appl. 2022, 203, 117592. [Google Scholar] [CrossRef]
  7. Almulihi, A.; Saleh, H.; Hussien, A.M.; Mostafa, S.; El-Sappagh, S.; Alnowaiser, K.; Ali, A.A.; Refaat Hassan, M. Ensemble Learning Based on Hybrid Deep Learning Model for Heart Disease Early Prediction. Diagnostics 2022, 12, 3215. [Google Scholar] [CrossRef]
  8. Elmannai, H.; Saleh, H.; Algarni, A.D.; Mashal, I.; Kwak, K.S.; El-Sappagh, S.; Mostafa, S. Diagnosis Myocardial Infarction Based on Stacking Ensemble of Convolutional Neural Network. Electronics 2022, 11, 3976. [Google Scholar] [CrossRef]
  9. Venkatesh, B.; Anuradha, J. A review of feature selection and its methods. Cybern. Inf. Technol. 2019, 19, 3–26. [Google Scholar] [CrossRef]
  10. Cai, J.; Luo, J.; Wang, S.; Yang, S. Feature selection in machine learning: A new perspective. Neurocomputing 2018, 300, 70–79. [Google Scholar] [CrossRef]
  11. Sarkar, B.K. Hybrid model for prediction of heart disease. Soft Comput. 2020, 24, 1903–1925. [Google Scholar] [CrossRef]
  12. Thomas, N.; Kavitha, A. Prediction of polycystic ovarian syndrome with clinical dataset using a novel hybrid data mining classification technique. Int. J. Adv. Res. Eng. Technol. 2020, 11, 1872–1881. [Google Scholar]
  13. Rudin, C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell. 2019, 1, 206–215. [Google Scholar] [CrossRef] [PubMed]
  14. El-Sappagh, S.; Alonso, J.M.; Islam, S.; Sultan, A.M.; Kwak, K.S. A multilayer multimodal detection and prediction model based on explainable artificial intelligence for Alzheimer’s disease. Sci. Rep. 2021, 11, 2660. [Google Scholar] [CrossRef] [PubMed]
  15. Lee, H.; Yune, S.; Mansouri, M.; Kim, M.; Tajmir, S.H.; Guerrier, C.E.; Ebert, S.A.; Pomerantz, S.R.; Romero, J.M.; Kamalian, S.; et al. An explainable deep-learning algorithm for the detection of acute intracranial haemorrhage from small datasets. Nat. Biomed. Eng. 2019, 3, 173–182. [Google Scholar] [CrossRef] [PubMed]
  16. Bharati, S.; Podder, P.; Mondal, M.R.H. Diagnosis of polycystic ovary syndrome using machine learning algorithms. In Proceedings of the 2020 IEEE Region 10 Symposium (TENSYMP), Dhaka, Bangladesh, 5–7 June 2020; pp. 1486–1489. [Google Scholar]
  17. Denny, A.; Raj, A.; Ashok, A.; Ram, C.M.; George, R. i-hope: Detection and prediction system for polycystic ovary syndrome (pcos) using machine learning techniques. In Proceedings of the TENCON 2019—2019 IEEE Region 10 Conference (TENCON), Kochi, India, 17–20 October 2019; pp. 673–678. [Google Scholar]
  18. Anda, D.; Iyamah, E. Comparative Analysis of Artificial Intelligence in the Diagnosis of Polycystic Ovary Syndrome. Available online: https://www.researchgate.net/publication/366320486_Comparative_Analysis_of_Artificial_Intelligence_in_the_Diagnosis_of_Polycystic_Ovary_Syndrome (accessed on 17 March 2023).
  19. Bhardwaj, P.; Tiwari, P. Manoeuvre of Machine Learning Algorithms in Healthcare Sector with Application to Polycystic Ovarian Syndrome Diagnosis. In Proceedings of Academia-Industry Consortium for Data Science: AICDS 2020; Springer: New York, NY, USA, 2022; pp. 71–84. [Google Scholar]
  20. Adla, Y.A.A.; Raydan, D.G.; Charaf, M.Z.J.; Saad, R.A.; Nasreddine, J.; Diab, M.O. Automated detection of polycystic ovary syndrome using machine learning techniques. In Proceedings of the 2021 Sixth International Conference on Advances in Biomedical Engineering (ICABME), Werdanyeh, Lebanon, 7–9 October 2021; pp. 208–212. [Google Scholar]
  21. Thakre, V.; Vedpathak, S.; Thakre, K.; Sonawani, S. PCOcare: PCOS detection and prediction using machine learning algorithms. Biosci. Biotechnol. Res. Commun. 2020, 13, 240–244. [Google Scholar] [CrossRef]
  22. Chauhan, P.; Patil, P.; Rane, N.; Raundale, P.; Kanakia, H. Comparative analysis of machine learning algorithms for prediction of pcos. In Proceedings of the 2021 International Conference on Communication information and Computing Technology (ICCICT), Mumbai, India, 25–27 June 2021; pp. 1–7. [Google Scholar]
  23. Rathod, Y.; Komare, A.; Ajgaonkar, R.; Chindarkar, S.; Nagare, G.; Punjabi, N.; Karpate, Y. Predictive Analysis of Polycystic Ovarian Syndrome using CatBoost Algorithm. In Proceedings of the 2022 IEEE Region 10 Symposium (TENSYMP), Mumbai, India, 1–3 July 2022; pp. 1–6. [Google Scholar]
  24. Aggarwal, N.; Shukla, U.; Saxena, G.J.; Kumar, M.; Bafila, A.S.; Singh, S.; Pundir, A. An Improved Technique for Risk Prediction of Polycystic Ovary Syndrome (PCOS) Using Feature Selection and Machine Learning. In Computational Intelligence: Select Proceedings of InCITe 2022; Springer: New York, NY, USA, 2023; pp. 597–606. [Google Scholar]
  25. Khanna, V.V.; Chadaga, K.; Sampathila, N.; Prabhu, S.; Bhandage, V.; Hegde, G.K. A Distinctive Explainable Machine Learning Framework for Detection of Polycystic Ovary Syndrome. Appl. Syst. Innov. 2023, 6, 32. [Google Scholar] [CrossRef]
  26. Polycystic Ovary Syndrome (PCOS). 2023. Available online: https://www.kaggle.com/datasets/prasoonkottarathil/polycystic-ovary-syndrome-pcos (accessed on 17 March 2023).
  27. Mahdhaoui, A.; Chetouani, M.; Cassel, R.S.; Saint-Georges, C.; Parlato, E.; Laznik, M.C.; Apicella, F.; Muratori, F.; Maestro, S.; Cohen, D. Computerized home video detection for motherese may help to study impaired interaction between infants who become autistic and their parents. Int. J. Methods Psychiatr. Res. 2011, 20, e6–e18. [Google Scholar] [CrossRef]
  28. Joenssen, D.; Bankhofer, U. Hot Deck Methods for Imputing Missing Data Hot Deck Methods for Imputing Missing Data the Effects of Limiting Donor Usage. 13 July 2012. Available online: https://www.semanticscholar.org/paper/Hot-Deck-Methods-for-Imputing-Missing-Data-The-of-Joenssen-Bankhofer/853253faf9d7ee66a4ebd749659c463cdc475f7c (accessed on 17 March 2023).
  29. Moon, T.K. The expectation-maximization algorithm. IEEE Signal Process. Mag. 1996, 13, 47–60. [Google Scholar] [CrossRef]
  30. Cho, E.; Chang, T.W.; Hwang, G. Data preprocessing combination to improve the performance of quality classification in the manufacturing process. Electronics 2022, 11, 477. [Google Scholar] [CrossRef]
  31. Gu, Q.; Li, Z.; Han, J. Generalized fisher score for feature selection. arXiv 2012, arXiv:1202.3725. [Google Scholar]
  32. Lin, X.; Li, C.; Zhang, Y.; Su, B.; Fan, M.; Wei, H. Selecting feature subsets based on SVM-RFE and the overlapping ratio with applications in bioinformatics. Molecules 2017, 23, 52. [Google Scholar] [CrossRef] [PubMed]
  33. Huang, J.; Cai, Y.; Xu, X. A filter approach to feature selection based on mutual information. In Proceedings of the 2006 5th IEEE International Conference on Cognitive Informatics, Beijing, China, 17–19 July 2006; Volume 1, pp. 84–89. [Google Scholar]
  34. He, Y.; Yu, H.; Yu, R.; Song, J.; Lian, H.; He, J.; Yuan, J. A correlation-based feature selection algorithm for operating data of nuclear power plants. Sci. Technol. Nucl. Install. 2021, 2021, 9994340. [Google Scholar] [CrossRef]
  35. Bateni, M.; Chen, L.; Fahrbach, M.; Fu, G.; Mirrokni, V.; Yasuda, T. Sequential Attention for Feature Selection. arXiv 2022, arXiv:2209.14881. [Google Scholar]
  36. Socher, R.; Perelygin, A.; Wu, J.; Chuang, J.; Manning, C.D.; Ng, A.Y.; Potts, C. Recursive deep models for semantic compositionality over a sentiment treebank. In Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, Seattle, WA, USA, 18–21 October 2013; pp. 1631–1642. [Google Scholar]
  37. LaValley, M.P. Logistic regression. Circulation 2008, 117, 2395–2399. [Google Scholar] [CrossRef] [PubMed]
  38. Rigatti, S.J. Random forest. J. Insur. Med. 2017, 47, 31–39. [Google Scholar] [CrossRef] [PubMed]
  39. Webb, G.I.; Keogh, E.; Miikkulainen, R. Naïve Bayes. Encycl. Mach. Learn. 2010, 15, 713–714. [Google Scholar]
  40. Suthaharan, S.; Suthaharan, S. Support vector machine. In Machine Learning Models and Algorithms for Big Data Classification: Thinking with Examples for Effective Learning; Spring: Berlin/Heidelberg, Germany, 2016; pp. 207–235. [Google Scholar]
  41. Peterson, L.E. K-nearest neighbor. Scholarpedia 2009, 4, 1883. [Google Scholar] [CrossRef]
  42. Chen, T.; He, T.; Benesty, M.; Khotilovich, V.; Tang, Y.; Cho, H.; Chen, K.; Mitchell, R.; Cano, I.; Zhou, T.; et al. Xgboost: Extreme Gradient Boosting. R Package Version 0.4-2. 2015, Volume 1, pp. 1–4. Available online: https://scholar.google.com/scholar?oi=bibs&cluster=11444560539169478279&btnI=1&hl=en (accessed on 17 March 2023).
  43. Schapire, R.E. Explaining adaboost. In Empirical Inference: Festschrift in Honor of Vladimir N. Vapnik; Spring: Berlin/Heidelberg, Germany, 2013; pp. 37–52. [Google Scholar]
  44. Wu, J.; Chen, X.Y.; Zhang, H.; Xiong, L.D.; Lei, H.; Deng, S.H. Hyperparameter optimization for machine learning models based on Bayesian optimization. J. Electron. Sci. Technol. 2019, 17, 26–40. [Google Scholar]
  45. Snoek, J.; Larochelle, H.; Adams, R.P. Practical bayesian optimization of machine learning algorithms. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; Volume 25. [Google Scholar]
  46. El-Rashidy, N.; Abuhmed, T.; Alarabi, L.; El-Bakry, H.M.; Abdelrazek, S.; Ali, F.; El-Sappagh, S. Sepsis prediction in intensive care unit based on genetic feature optimization and stacked deep ensemble learning. In Neural Computing and Applications; Spring: Berlin/Heidelberg, Germany, 2022; pp. 1–30. [Google Scholar]
  47. Saleh, H.; Mostafa, S.; Alharbi, A.; El-Sappagh, S.; Alkhalifah, T. Heterogeneous ensemble deep learning model for enhanced Arabic sentiment analysis. Sensors 2022, 22, 3707. [Google Scholar] [CrossRef]
  48. El-Rashidy, N.; El-Sappagh, S.; Abuhmed, T.; Abdelrazek, S.; El-Bakry, H.M. Intensive care unit mortality prediction: An improved patient-specific stacking ensemble model. IEEE Access 2020, 8, 133541–133564. [Google Scholar] [CrossRef]
  49. Narkhede, S. Understanding auc-roc curve. Towards Data Sci. 2018, 26, 220–227. [Google Scholar]
Figure 1. The phases of prediction PCOS.
Figure 1. The phases of prediction PCOS.
Diagnostics 13 01506 g001
Figure 2. The different types of feature selection methods.
Figure 2. The different types of feature selection methods.
Diagnostics 13 01506 g002
Figure 3. Scores of selected features by mutual_info.
Figure 3. Scores of selected features by mutual_info.
Diagnostics 13 01506 g003
Figure 4. Sores of selected features by based tree.
Figure 4. Sores of selected features by based tree.
Diagnostics 13 01506 g004
Figure 5. Ranking of the selected features by RFE.
Figure 5. Ranking of the selected features by RFE.
Diagnostics 13 01506 g005
Figure 6. ROC curves of splitting 80:20.
Figure 6. ROC curves of splitting 80:20.
Diagnostics 13 01506 g006
Figure 7. ROC curves of splitting 70:30.
Figure 7. ROC curves of splitting 70:30.
Diagnostics 13 01506 g007
Figure 8. The best models for 80:20 splitting.
Figure 8. The best models for 80:20 splitting.
Diagnostics 13 01506 g008
Figure 9. The best models for 70:30 splitting.
Figure 9. The best models for 70:30 splitting.
Diagnostics 13 01506 g009
Figure 10. Global explainability of the developed model: (a) bar plot (b) Cohort plot.
Figure 10. Global explainability of the developed model: (a) bar plot (b) Cohort plot.
Diagnostics 13 01506 g010
Figure 11. Heatmap of the developed model.
Figure 11. Heatmap of the developed model.
Diagnostics 13 01506 g011
Figure 12. Full water plot for the first and second observation: (a) first observation; (b) second observation.
Figure 12. Full water plot for the first and second observation: (a) first observation; (b) second observation.
Diagnostics 13 01506 g012
Figure 13. Individual fore plot for several instances according to the developed model (a) for observation 1 and (b) for observation 2.
Figure 13. Individual fore plot for several instances according to the developed model (a) for observation 1 and (b) for observation 2.
Diagnostics 13 01506 g013
Figure 14. Collective force plot for the developed model.
Figure 14. Collective force plot for the developed model.
Diagnostics 13 01506 g014
Table 1. Dataset features description details.
Table 1. Dataset features description details.
Feature NameAbbDescription
Patient File No. Patient file number (unique identifier)
Polycystic Ovary SyndromePCOSClass label (determine if the patient has this syndrome or not)
AgeAGEPatient’s age in years
WeightWEIGHTPatient’s weight in KG
HeightHEIGHTPatient’s height in CM
Body Mass IndexBMIBody mass index of the patient (height/weight)
Blood GroupBGPatients belong to which blood group
(A+, A−, B+, B−, O+, O−, AB+, AB−)
Pulse RatePRHeartbeat per minute
Respiration RatesRRRespiration rates per minute
HemoglobinHBNumber of red blood cells in patient’s body
CycleCYCLELength of the menstrual cycle
Cycle LengthCLNumber of days of a cycle
Marriage StatusMSNumber of years since marriage
PregnantPPregnant status
No. of AbortionsABNo. of abortions
I Beta-HCGBETA_IAmount of human chorionic gonadotropin
Beta Healthy Singleton PregnancyBETA_IIBeta HCG level is indication of 100 mIU/ml
about 16 days after ovulation,
Follicle-Stimulating HormoneFSHAttributes ranging from 0.3 to 10.0 mIU/mL
indicate if are still menstruating or have undergone menopause
Luteinizing HormoneLHChemical agitator that stimulates the reproductive system
Follicle-Stimulating Hormone/
Luteinizing Hormone
FSH/LHRatio of FSH and LH
Hip SizeHIPSize of hip in inches
Waist SizeWAISTSize of waist in inches
Waist-Hip RatioHIP_RATIOWaist size proportion to hip
Thyroid-Stimulating HormoneTSHAmount of TSH in the blood
Anti-Mullerian HormoneAMHPlays a key role in developing a baby’s sex organs
while in the womb
Prolactin levelsPRLProlactin levels in women’s bodies
Vitamin DVIT_D3Vitamin D levels
Progesterone LevelsPRGProgesterone levels
Random Blood SugarRBSValue of random blood sugar (RBS) test
Weight GainWGTest to check if the patient gains weight
Hair GrowthHGTest to check if a patient has hair growth
Skin DarkeningSDTest to check the appearance of darkness in skin
Hair LossHLTest to check hair loss
PimplesPIMPLESPimple issues
Fast FoodFFCheck if fast food part of the diet
Reg.ExerciseRECheck if patient exercises on a regular basis
Blood Pressure SystolicBP_ SYSTOLIC:Amount of pressure in the arteries when the heart is contracting
Blood Pressure DiastolicBP_ DiastolicAmount of pressure in the arteries while
the heart is resting in between heart beats
Follicle No.FNFollicle number in the left side
Table 2. Performance of the classifiers with selected features using splitting 80:20.
Table 2. Performance of the classifiers with selected features using splitting 80:20.
Feature Selection MethodsModelsACCPRERECF1
mutual_infoRF96.8396.8396.8396.83
LR96.8396.9796.8396.78
DT95.2495.3495.2495.26
NB95.2495.5695.2495.14
KNN95.2495.5695.2495.14
SVM96.8397.1096.8396.86
XGB98.1298.1098.1298.12
AdaBoost95.2495.3495.2495.26
Stacking ML98.4198.4898.4198.42
RFERF98.4198.4598.4198.40
LR96.8396.9796.8396.78
DT93.6593.9993.6593.72
NB98.4198.4598.4198.40
KNN95.2495.5695.2495.14
SVM96.8396.8396.8396.83
XGB98.4198.4598.4198.40
AdaBoost98.4198.4598.4198.40
Stacking ML100100100100
Tree basedRF96.8396.8396.8396.83
LR92.0692.0292.0692.01
DT93.6593.9993.6593.72
NB95.2495.2395.2495.21
KNN93.6593.6593.6593.65
SVM96.8396.8396.8396.83
XGB96.8396.8396.8396.83
AdaBoost96.8396.8396.8396.83
Stacking ML97.4197.4597.4197.4
Table 3. Performance of the classifiers with selected features using splitting 70:30.
Table 3. Performance of the classifiers with selected features using splitting 70:30.
Feature Selection MethodsModelsACCPRERECF1
mutual_infoRF94.6894.6694.6894.66
LR90.4390.3990.4390.30
DT92.5592.5892.5592.46
NB79.7982.1579.7980.24
KNN84.0485.0184.0484.29
SVM91.4991.4491.4991.42
XGB93.6293.6193.6293.56
AdaBoost95.7495.7495.7495.74
Stacking ML96.8196.8196.8196.80
RFERF95.7495.7795.7495.71
LR93.6293.6193.6293.56
DT92.5592.8392.5592.38
NB85.1185.3985.1185.21
KNN89.3689.2889.3689.27
SVM96.8196.8196.8196.80
XGB94.6894.6694.6894.66
AdaBoost96.8196.8696.8196.82
Stacking ML98.8798.0098.8798.89
Tree basedRF95.8195.8195.8195.80
LR93.6293.6293.6293.62
DT92.5592.5892.5592.46
NB87.2387.8987.2387.40
KNN80.8582.8380.8581.25
SVM96.8196.8196.8196.80
XGB92.5592.6392.5592.58
AdaBoost96.8196.8196.8196.80
Stacking ML97.8197.8197.8197.8
Table 4. Comparison with previous studies.
Table 4. Comparison with previous studies.
PapersMethodsAccuracy
 [16]RFLR with UFS91.01
 [17]RF with PCA89.02
 [6]RF with correlation92.4
 [18]RF96
 [19]SVM with Pearson correlation93
 [20]SVM with hybrid feature selection91.6
 [21]RF with chi square90.9
 [22]DT with Gini importance92.59
 [25]multi-stack of ML98
Our workStacking ML with RFE100
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Elmannai, H.; El-Rashidy, N.; Mashal, I.; Alohali, M.A.; Farag, S.; El-Sappagh, S.; Saleh, H. Polycystic Ovary Syndrome Detection Machine Learning Model Based on Optimized Feature Selection and Explainable Artificial Intelligence. Diagnostics 2023, 13, 1506. https://doi.org/10.3390/diagnostics13081506

AMA Style

Elmannai H, El-Rashidy N, Mashal I, Alohali MA, Farag S, El-Sappagh S, Saleh H. Polycystic Ovary Syndrome Detection Machine Learning Model Based on Optimized Feature Selection and Explainable Artificial Intelligence. Diagnostics. 2023; 13(8):1506. https://doi.org/10.3390/diagnostics13081506

Chicago/Turabian Style

Elmannai, Hela, Nora El-Rashidy, Ibrahim Mashal, Manal Abdullah Alohali, Sara Farag, Shaker El-Sappagh, and Hager Saleh. 2023. "Polycystic Ovary Syndrome Detection Machine Learning Model Based on Optimized Feature Selection and Explainable Artificial Intelligence" Diagnostics 13, no. 8: 1506. https://doi.org/10.3390/diagnostics13081506

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop