Next Article in Journal
The Relationship between On-Ice and Off-Ice Performance in Elite Male Adolescent Ice Hockey Players—An Observation Study
Next Article in Special Issue
Pathomics and Deep Learning Classification of a Heterogeneous Fluorescence Histology Image Dataset
Previous Article in Journal
Damage Evolution Analysis in a “Spaghetti” Bridge Model Using the Acoustic Emission Technique
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Classification of Shoulder X-ray Images with Deep Learning Ensemble Models

1
Department of Electrical and Electronics Engineering, Faculty of Engineering, Gazi University, TR 06570 Ankara, Turkey
2
Department of Orthopaedics and Traumatology, Faculty of Medicine, Gazi University, TR 06570 Ankara, Turkey
3
Department of Radiology, Faculty of Medicine, Gazi University, TR 06570 Ankara, Turkey
*
Author to whom correspondence should be addressed.
Appl. Sci. 2021, 11(6), 2723; https://doi.org/10.3390/app11062723
Submission received: 2 March 2021 / Revised: 12 March 2021 / Accepted: 14 March 2021 / Published: 18 March 2021
(This article belongs to the Special Issue Computer Aided Diagnosis)

Abstract

:
Fractures occur in the shoulder area, which has a wider range of motion than other joints in the body, for various reasons. To diagnose these fractures, data gathered from X-radiation (X-ray), magnetic resonance imaging (MRI), or computed tomography (CT) are used. This study aims to help physicians by classifying shoulder images taken from X-ray devices as fracture/non-fracture with artificial intelligence. For this purpose, the performances of 26 deep learning-based pre-trained models in the detection of shoulder fractures were evaluated on the musculoskeletal radiographs (MURA) dataset, and two ensemble learning models (EL1 and EL2) were developed. The pre-trained models used are ResNet, ResNeXt, DenseNet, VGG, Inception, MobileNet, and their spinal fully connected (Spinal FC) versions. In the EL1 and EL2 models developed using pre-trained models with the best performance, test accuracy was 0.8455, 0.8472, Cohen’s kappa was 0.6907, 0.6942 and the area that was related with fracture class under the receiver operating characteristic (ROC) curve (AUC) was 0.8862, 0.8695. As a result of 28 different classifications in total, the highest test accuracy and Cohen’s kappa values were obtained in the EL2 model, and the highest AUC value was obtained in the EL1 model.

1. Introduction

Having a wider and more varied range of movement than the other joints in the body, the shoulder has a flexible structure. The fractures in the shoulder may result from incidents such as dislocation of the shoulder and engaging in contact sports and motor vehicle accidents. The shoulder bone mainly consists of three different bones: the upper arm bone, named the ‘humerus’, the shoulder blade, named the ‘scapula’, and the collarbone, named the ‘clavicle’. Figure 1 shows the anatomic structure of the shoulder bone. The image in this figure was taken from the MURA dataset used in this study, and the markings thereon were placed by the physicians at Gazi University. As shown in Figure 1, the upper end of the humerus has a ball-like shape that connects with the scapula, called the glenoid. The types of shoulder fractures vary depending on age. While most fractures in children occur in the clavicle bone, the most common fracture in adults occur on the top part of the humerus, i.e., the proximal humerus. The types of shoulder bone fractures are divided into three categories in general: clavicle fractures, which are the most common shoulder fracture, frequently the result of a fall, scapula fractures, which rarely occur, and resulting fractures, which occur as cracks in the upper part of the arm in individuals over 65 years of age. The images from X-ray devices are primarily used for imaging of the shoulder bone for diagnosis and treatment of such fractures, while MRI or CT devices may also be used when required [1].
Deep learning classification procedures of shoulder bone X-ray images were carried out in the study. The main contributions of this study are as follows:
  • The most suitable model for the classification of shoulder bone X-ray images as a fracture or non-fracture is determined.
  • An approach that can be used in similar studies is developed via new ensemble learning models.
  • The study can assist physicians who are not experts in the field in the classification stage, especially in cases of shoulder fractures, which are frequently encountered in the emergency departments of hospitals.
  • The method suggested in the study contributes to the literature with two different ensemble approaches.
  • With the first model proposed in the study, a performance study was conducted with transfer learning for the MURA dataset, which is widely used in X-ray studies. Thus, three models that give the best classification results have been combined into a single model, and the classification performance has been increased. The developed model can be used to classify many medical X-ray images.
  • With the second model, it is determined which model finds which class best by looking at the success of finding the classes on the models in the dataset. It is ensured that the model decides the prediction of that class. Thus, regardless of the dataset studied, a similar decision system can be designed with models that find the classes in a given dataset, and a higher performance can be achieved compared to a single model.
  • The proposed ensemble model approaches can be applied and generalized by performing similar preprocessing steps in other X-ray biomedical datasets. In addition, the proposed method can be easily used in studies with transfer learning.

2. Related Works

In this study, the classification of X-ray images of shoulder fractures in the MURA dataset was carried out into the categories of ‘fracture’ or “non-fracture’ using built CNN-based deep learning models, new models thereof adapted to SpinalNet, and models developed based on ensemble learning. Therefore, two main topics have been taken into account while examining the literature regarding the study. Upon examination of the studies conducted thus far, these topics are as follows:
  • what the studies conducted using the MURA dataset are, and why this dataset is preferred in this study;
  • what kinds of studies have been conducted on the classification of shoulder bone images, and what kinds of innovations may be set forth based on the efficiency of deep learning models used in the classification procedures carried out in this study.

2.1. Studies Conducted Using the MURA Dataset

The MURA dataset was first introduced to the literature in a paper published in the OpenReview platform, announced in the conference on “Medical Imaging with Deep Learning” held in Amsterdam in 2018. Following this publication, this dataset was made publicly available for academic studies in a competition called “Bone X-Ray Deep Learning Competition” by the Machine Learning group of the Stanford University. Being one of the largest public radiographic image datasets, MURA contains a total of 40,561 X-ray images in png format for the following parts of the body, labeled as either normal or abnormal (fracture): elbow, finger, forearm, hand, humerus, shoulder, and wrist. In this classification study conducted by Rajpurkar et al. using DenseNet-169 on this dataset, the AUC score representing the area under the overall Receiver Operator Characteristics (ROC) curve was 0.929, and the overall Cohen’s kappa score was 0.705 [2]. Following this first study in which this dataset was introduced to the literature, there have been various studies that have been conducted and published using all or a part of the dataset. These studies are as follows: an average precision (AP) value of 62.04% was achieved in the fracture detection procedure using the proposed deep CNN model after the fractures were marked by the physicians on the arm X-ray images in the MURA dataset by Guan et al. [3]. The following classification accuracies were achieved by Galal et al. using a very small part of the elbow X-ray images in the MURA dataset (56 images for the train dataset, 24 images for the test): 97% with the support vector machine (SVM), 91.6% with random forest (RF), and 91.6% with naive Bayes [4]. The classification by Liang and Gu with the multi-scale CNN and graph convolution network (GCN) they proposed using the entire MURA dataset achieved an overall Cohen’s kappa score of 0.836. [5]. The overall Cohen’s kappa score achieved by Saif et al. in classification performed using the whole MURA dataset with the proposed capsule network architecture was 0.80115 [6]. In the classification carried out by Cheng et al. using the dataset containing hip bone images and the whole MURA dataset, the highest accuracy achieved was 86.53% for humerus images with the proposed adversarial policy gradient augmentation (APGA) [7]. The highest accuracy achieved in the classification performed by Pelka et al. using the whole MURA dataset was 79.85% with the InceptionV3 model [8]. The AUC score was 0.88 in the classification carried out by Varma et al. on a Lower Extremity Radiographs Dataset (LERA) containing foot, knee, ankle, and hip data for an ImageNet and DenseNet161 model pre-trained with MURA [9]. In the classification performed by Harini et al. on images of the finger, wrist, and shoulder in the MURA dataset using five different CNN-based built deep learning methods, the highest accuracy was 56.30% on the wrist data with DenseNet169 [10]. The overall accuracy achieved for MURA with the proposed iterative fusion CNN (IFCNN) method is 73.4% in classification carried out by Fang et al. using the optical coherence tomography (OCT) dataset and the entire MURA dataset [11]. The overall Cohen’s kappa score achieved was 0.717 with the proposed deep CNN-based ensemble model in the classification carried out by Mondol et al. on elbow, finger, humerus, and wrist data in the MURA dataset [12]. In the type classification performed by Pradhan et al. using the whole MURA dataset, 91.37% accuracy was achieved with the proposed deep CNN model [13]. In the classification carried out by Shao and Wang using the entire MURA dataset, a two-stage method was developed, and the highest accuracy achieved was in humerus images with 88.5% for the SENet154 model, while the highest accuracy for the DenseNet201 model was achieved again on humerus images with 90.94% [14].
Although there are many different publicly available datasets on bone fractures in addition to the MURA dataset in the literature, the main reason for using only this dataset in this study is that MURA is one of the largest datasets for both normal (negative, non-fracture) and abnormal (positive, fracture) groups compared to the other public datasets. It is observed upon examination of the studies conducted via this dataset that classification and/or fracture detection procedures are carried out using the dataset in whole or in part. Only the X-ray images for shoulder bone in the MURA dataset were used in this study. The reason why only this type of dataset was used for classification despite the availability of seven different types of datasets and the reason why such a dataset contains shoulder data is that the shoulder dataset is the most balanced type in terms of distribution of the amount of data provided for both training and validation. Another reason for classification for a single type only is to be able to develop the most optimum model for fracture and non-fracture images on this type, using and developing as many different deep learning models as possible.

2.2. Classification Studies Carried Out on the Shoulder Bone

While some of the studies mentioned under the previous title with the MURA dataset include classification of shoulder bone images, there are also studies in the literature on this type of classification exclusively. In classification carried out by Chung et al. with the ResNet152 model on a total of 1891 patients, including 1376 proximal humeral fracture cases of four different types, a 96% accuracy was achieved [15]. In the CNN model proposed by Sezers using a total of 219 shoulder MR images, a classification was made in three different groups as normal, edematous, and Hill–Sachs lesions with a 98.43% accuracy [16]. The highest accuracy achieved in the classification carried out by Urban et al. on 597 X-ray images of shoulders with implants was 80.4% in the NASNet model pre-trained with ImageNet [17]. In the classification performed by Sezer et al. using a total of 219 shoulder MR images grouped as 91 edematous, 49 Hill–Sachs lesions, and 79 normal, an 88% success rate was achieved with a kernel-based SVM, and 94% success was achieved with extreme learning machines [18]. A 94.74% accuracy was achieved with the proposed CapsNet model in the classification performed by Sezers on a total of 1006 shoulder MR images, grouped as 316 normal, 311 degenerated, and 379 torn [19]. Some other important studies on medical data classification and machine/deep learning approach in the literature are as follows: An accuracy of 93.5% was achieved in cardiac arrhythmia classification procedures with long short-term memory (LSTM) by Khan and Kim [20]. ResNet18 and GoogLeNet models were used by Storey et al. to detect wrist bone abnormality [21]. As a result of the classification made by Yin et al. for kidney disease with the multi-instance deep learning method, an accuracy of 88.6% was obtained [22]. As a result of the detection of musculoskeletal abnormalities performed by Dias, the highest accuracy was obtained as 81.98% with the SqueezeNet model [23]. An accuracy of 98.20% was achieved by Khan and Kim using spark machine learning and conv-autoencoder to detect and classify unpredictable malicious attacks [24]. Transcriptional co-activators have been identified by Kegelman et al. as regulators of bone fracture healing [25]. The bone fracture process performed by Sharma et al. with machine learning and digital geometry achieved 92% classification accuracy [26].
It is observed upon examination of the recent classification studies on shoulder bone in the literature that the normal and abnormal (fracture, edematous, Hill–Sachs lesion, degenerated, or torn) images obtained from CT, MRI, or X-ray devices are classified by not only the traditional machine learning methods such as the SVM but also by deep learning-based methods such as ResNet, NasNet, and CapsNet. In this study, the classification was performed on normal and abnormal shoulder bone X-ray (fracture) images in the MURA dataset. In the performance thereof, as a practice different from the literature, built CNN-based deep learning models (ResNet, ResNeXt, DenseNet, VGG, InceptionV3, and MobileNetV2) with a varying structure and number of layers were used by transfer learning with pre-trained ImageNet by also adding SpinalNet to the classification layers. Based on the results of classification achieved by the procedures performed herein, another classification was performed with ensemble learning. The main reason for applying transfer learning on built CNN models, adding SpinalNet, and applying ensemble learning is to contribute to the efficiency of deep learning models proposed in the classification of shoulder bone X-ray images.
Section 3 discusses the CNN-based built deep learning models, such as ResNet, DenseNet, VGG, and InceptionV3, used in classification as well as their adaptations with Spinal FC and the proposed ensemble learning models. Section 4 explains the open source dataset, including the shoulder bone X-ray images, data augmentation, and data pre-processing procedures applied thereon, a table containing the accuracy, recall, F1-score, and Cohen’s kappa scores achieved by classification models, and the results of classification, including a confusion matrix, ROC curves, and AUC scores. The last section interprets the results achieved by classification models and discusses the contribution of this study to the literature and improvements that can be applied in future work.

3. Methods

A number of different deep-learning-based methods have been used and developed to classify X-ray images of the shoulder bone in png format and with three channels as normal and abnormal in this study. Firstly, classification was performed with ResNet (34,50,101,152), ResNeXt (50,101), DenseNet (169,201), VGG (13,16,19), InceptionV3, and MobileNetV2, which are CNN-based deep learning models with different structures and layers that are publicly available. Subsequently, new classification networks were established by replacement of the classification layer of each model used herein with the SpinalNet FC layer. Based on the results obtained herein, new ensemble learning models specific to this study were developed in order to further increase the classification accuracy. Built deep learning models, newly developed models with SpinalNet, and details of the ensemble learning models used for classification are described in the following subsections.

3.1. Classification Models Based on CNNs for Shoulder Bone X-ray Images

The CNN-based deep learning models currently available were used to classify X-ray images of the shoulder bone. In training the models, training was firstly carried out with data in the training section of the dataset without pre-training, i.e., with a random weight as the initial weight. However, the required level could not be reached in training the network or at the end of classification by this application. Therefore, the transfer learning method was applied.
In this transfer learning method, two versions of each available deep learning model pre-trained with ImageNet data, i.e., the existing weights in the pre-trained models, were used. ImageNet is a dataset and benchmark that contains millions of images and hundreds of object categories, on which the procedures of image classification, single-object localization, and object detection are mainly performed [27]. The shoulder bone X-ray images dataset used in this study was used for input of the deep learning models, which are specified in Figure 2 and were trained with ImageNet data. The structure comprised of 1000 classes in the last layer of these models was made to have two classes in order to classify bone images specific to our study as normal and abnormal. Following the abovementioned procedures, classification was carried out by training the network with the shoulder dataset. Moreover, the last layer of 13 built deep learning models were fine-tuned with SpinalNet/Spinal FC; with the newly acquired models, shoulder bone X-ray images were trained and classified.
The currently available ResNet, ResNeXt, DenseNet, VGG, InceptionV3, and MobileNetV2 deep learning models, which can be used in classification, along with versions with different numbers of layers were some of the models used in the classification of shoulder bone X-ray images in this study. The required method-based details regarding these models and SpinalNet are described in the following subsections.

3.1.1. ResNet

ResNet is a CNN-based deep learning model including more than one building blocks of residual learning depending on the number of layers [28]. In part (a) of the Figure 3, there are k number of building blocks with 3 × 3 conv for ResNet34, and, in part (b), there are m number of building blocks with 1 × 1 and 3 × 3 conv and n number of building blocks with 1 × 1 conv for ResNet50, 101, and 152, respectively.
In this study, four ResNet models with 34, 50, 101, and 152 layers, respectively, were used in classification. The models were adapted, and the structure is shown in Figure 4.
ResNeXt is a deep learning model for deep neural networks that reduces the number of parameters in ResNet. With this model, cardinality, which is an additional dimension for the width and depth of ResNet, defining the size of the set of transformations, is used [29]. The structure of the ResNext block is shown in Figure 5. In this structure, there is a block structure with a cardinality value of 32, with r number of 1 × 1 and 3 × 3 conv and s number of 1 × 1 conv, respectively.
Two ResNeXt models with 50 and 101 layers, respectively, were used in classification in this study. The models were adapted, and their structure is shown in Figure 6.

3.1.2. DenseNet

In a dense convolutional neural network, known as DenseNet, each layer affects every other layer in a feed-forward fashion [30]. A DenseNet block consists of five layers. The first four are dense layers, and the last is a transition layer. If the growth rate value is (k) for each layer, it is 4 for this dense block. In the transition layer, there is a 2 × 2 average pool with a 1 × 1 conv and a stride of 2. In the dense layer, there are 1 × 1 and 3 × 3 convs with a stride of 1.
In this study, two DenseNet models with 169 and 201 layers, respectively, and a growth rate (k) of 32 were used in classification. These models were adapted, and their structure is shown in Figure 7.

3.1.3. VGG

VGG is a CNN-based deep learning model with very small (3 × 3) convolution filters [31]. In this study, three VGG models with 13, 16, and 19 layers, respectively, were used for the classification procedure. The structure of conv blocks was comprised of a combination of 2, 3, or 4 consecutive 3 × 3 conv (with k, m or n filters) layers in Figure 8, and these VGG-based models, which were used and adapted from a combination thereof, are shown in Figure 9.

3.1.4. InceptionV3

InceptionV3 is a CNN-based deep learning model that includes convolution factorization, which is called the inception module [32]. The structure is shown in Figure 10.
In this study, the InceptionV3 model was used in classification. This model was adapted, and its structure is shown in Figure 11.

3.1.5. MobileNetV2

MobileNetV2 is a deep learning model that includes residual bottleneck layers [33]. The structure of the convolution block of this model is shown in Figure 12.
The MobileNetV2 model was used in classification in this study. The structure of this MobileNetV2-based model, which contains a total of 17 MobileNet bottleneck blocks, comprised of 13 bottlenecks with a stride value of 1, and 4 bottlenecks with a stride value of 2, is shown in Figure 13.

3.2. SpinalNet

In SpinalNet, the structure of the hidden layers is different from a normal neural network model. In a neural network, the hidden layers receive inputs in the first layer and then transfer the intermediate outputs to the next layer. However, the hidden layer in SpinalNet enables the previous layer to receive a certain part of its inputs and outputs. Therefore, the number of incoming weights in the hidden layer is lower than in normal neural networks [34].
The classification layer of the 13 above-mentioned CNN models based on ResNet, ResNeXt, DenseNet, VGG, InceptionV3, and MobileNetV2, used in classification, was adapted with SpinalNet/Spinal FC, and classification was also carried out with Spinal FC versions of these models. This allowed observation of the effect of SpinalNet on the classification of shoulder bone X-ray images. The details regarding this effect are described in Section 4. Moreover, the layer widths in Spinal FC of each classification model are as Table 1.

3.3. Proposed Classification Models Based on EL for Shoulder Bone X-ray Images

Two ensemble models were established to further increase the accuracy of classification by examining the classification results of shoulder bone X-ray images performed with the CNN-based models and their SpinalNet versions used in the study, mentioned in the previous subsections. While choosing the sub-models required for ensemble learning, the accuracy rates in the classification results of each model, the confusion matrix scores, and the (normal/abnormal) AUC scores for each class were taken into consideration. The details of these two ensemble-based classification models are explained in the following subsections.

3.3.1. EL1 (ResNext50 with Spinal FC, DenseNet169 with Standard FC, and DenseNet201 with Spinal FC)

The EL1 (Ensemble learning-1) model was developed using a combination of ResNext50 with Spinal FC, DenseNet169 with Standard FC, and DenseNet201 with Spinal FC. A schematic diagram of the developed model is shown in Figure 14.
It is observed upon examination of the figure above of the EL1 model proposed for classification of shoulder bone X-ray images as normal/abnormal (fracture) that three different CNN-based models are used. The “data” specified in this figure represent the entire dataset of shoulder bone X-ray images, “Model I” is the ResNeXt50 model with SpinalNet FC, “Model II” is the DenseNet169 model with a Standard FC layer, “Model III” is the DenseNet201 model with SpinalNet FC, “Outputs 1–3” are the outputs of these three models, and “Prediction” is the normal/abnormal (fracture) class types. In selection of the three models specified in this ensemble model, the parameters in the training outputs of the 26 CNN models presented in the previous title were taken into consideration. These parameters can be stated as Cohen’s kappa score, AUC, and test accuracy. The steps of the procedure and a block diagram in Figure 15 for the proposed EL1 ensemble model are provided as follows:
  • Step 1: The last layers of the three pre-trained sub-models in the EL1 ensemble model are adjusted as the Identity layer.
  • Step 2: A final layer with 80 outputs for ResNext50 with Spinal FC, 1664 for DenseNet169 with Standard FC, and 960 for DenseNet201 with Spinal FC is achieved.
  • Step 3: These outputs are combined to form a single linear layer and a hidden layer with 2704 outputs.
  • Step 4: This hidden layer is connected to the classifying layer connected to the sigmoid activation function, the output of which is 2.
  • Step 5: The network established as a result of these procedures is re-trained, providing results.

3.3.2. EL2 (ResNet34 with Spinal FC, DenseNet169 with Standard FC, DenseNet201 with Spinal FC, and ResNext50 with Spinal FC)

The EL2 (Ensemble learning-2) model was developed using a combination of ResNet34 with Spinal FC, DenseNet169 with Standard FC, DenseNet201 with Spinal FC, and a sub-ensemble model. A schematic diagram of the developed model is provided in Figure 16. It is observed upon examination of the figure that the EL2 model, differently from the EL1 model, consists of three single models and one sub-ensemble model and has a different structure. The “X-ray Images” specified in the figure refers to the dataset containing shoulder bone images, “Model I” to the ResNet34 model with Spinal FC, “Model II” to the DenseNet201 model with Spinal FC, “Model III” to the sub-ensemble model, “Model IV” to the DenseNet169 model with Standard FC, and “Predictions I–IV” refer to the outputs of the classification performed with these four models. The parts of the procedure for the proposed EL2 ensemble model are as follows:
  • Part 1: In the “Input” section, there is a shoulder X-ray image dataset that has been subjected to certain image processing techniques.
  • Part 2: In the “Classification Network” section, the models that establish our ensemble model are defined. There are three single sub-models and a sub-ensemble model that constitute our ensemble model therein. Our sub-ensemble model is an architecture trained by connecting the predicted outputs of ResNet34 with Spinal FC, DenseNet201, ResNeXt50, and DenseNet169 with Standard FC to a linear layer with eight inputs and two outputs. The evaluation of 26 models as single was effective in the selection of these four models.
  • Part 3: In the “Prediction” sections, there are normal/abnormal (fracture) class type outputs achieved as a result of the classification performed in the previous section.
  • Part 4: In the “Main Check” section, there is a main check mechanism that plays a role in determining the class of the input image. Therein, Models I and II suggest classifying the input image as abnormal in the final classification, while Models III and IV suggest classifying the input image as normal in the final classification. In cases where suggestions are not available, the classification is carried out with the sub-ensemble model. In the selection of the referred models for each class, Confusion Matrix and Recall parameters previously obtained for the 26 CNN models were taken into consideration.
  • Part 5: In the “Sub Check” section, there is a supplementary check mechanism under the main check mechanism. The aim here is to use the classification result of a model (Model IV for Class 1 (abnormal) and Model II for Class 0 (normal)) other than the two models referred to in the main check section as a supplement for the final classification process.
  • Part 6: In the “Final Prediction” section, the final output determined as a result of the check mechanisms is achieved.
The structure of the EL2 model is explained in Algorithm 1.
Algorithm 1 EL2
Input: Shoulder bone X-ray images Dataset = test_dataset
Process:
  for image in test_dataset:
    pred_1 = Model_I(image)
    pred_2 = Model_II(image)
    pred_3 = Model_III(image)
    pred_4 = Model_IV(image)
    if(pred_1 and pred_2==1):
      if pred_4==1:
        final_pred = 1:
      if pred_4==0:
        final_pred = 0:
    elif(pred_3 and pred_4==0):
      if pred_2==1:
        final_pred = 1:
      if pred_2==0:
        final_pred = 0:
    else:
      final_pred = pred_3
Output: final_pred
The block diagram of the proposed EL2 ensemble model is shown in Figure 17.

4. Experiments

In this study, CNN-based deep learning (DL) models were used to classify shoulder bone X-ray images as abnormal (indicating fracture). X-ray images of the shoulder bone in the open source MURA dataset were used. Data pre-processing/augmentation was performed first. Following this, classification was carried out using 13 different CNN models, SpinalNet versions of these models, and two ensemble learning models. The proposed classification models for normal/abnormal (fracture) shoulder bone X-ray images are shown in Figure 18.

4.1. Dataset of Shoulder Bone X-ray Images

The MURA dataset is one of the largest among the published open-source radiography datasets. It contains finger, elbow, wrist, hand, forearm, humerus, and shoulder bone X-ray images [2]. In this study, only the shoulder bone X-ray images within the MURA dataset were used mainly because it is the most balanced type in the MURA dataset in terms of distribution of the amount of data provided for both training and validation. This balanced distribution is presented in Figure 19. A balanced dataset can otherwise be obtained with data augmentation or synthetic data generation to avoid problems that may arise when working with an imbalanced dataset. Although the MURA dataset is an open-source dataset, only the training and validation datasets are publicly available. The classification models used in the first study and in studies conducted in competition conducted with the MURA dataset were tested using test data that are not publicly available. Due to the confidential nature of the test data and the inability to conduct testing with these data as other studies have, the validation data was used as test data in this study.
These images, which initially had different resolutions and three channels, were first pre-processed and then converted to 320 × 320 × 3 pixels before being used for the deep learning model. The size 320 × 320 × 3 was chosen because it is the resolution most compatible with other studies using this dataset. The data formats are png, and no alterations were made in terms of the format type. The image type of the shoulder bone X-ray images, the number of training images, the number of test images, and the original and new image sizes are provided in the Table 2.
The table above shows that the image sizes were originally different but later converted to 320 × 320 × 3 in size. All images are in png format. The original quantity of normal images was 4211 in the training section and 285 in the test section, and that of abnormal images was 4168 in the training section and 278 in the test section.

4.2. Data Augmentation for Shoulder Bone X-ray Images

Since the results obtained in deep learning, particularly in classification studies, are very sensitive to the dataset, and since an increase in the amount of data in the training stage of the network has a positive effect on the training of the network, the quantity of shoulder bone X-ray images in the MURA dataset used in this study was increased by data augmentation. For this procedure, new images were obtained by rotating the images to the right or to the left to a maximum of 10 degrees.

4.3. Data Pre-Processing for Shoulder Bone X-ray Images

There is both noise and a dark background in the images observed in this study, which may adversely affect the classification and/or fracture detection processes. Various pre-processing steps were applied to the dataset to eliminate the abovementioned adverse effects as much as possible and to obtain more accurate results in the classification process. These pre-processing steps are listed below:
  • Detection of the Corresponding Area: Most of the X-ray images in the used dataset were insufficient in terms of semantic information in relation to the image size. In order to eliminate such insufficiency, the images were first converted to gray-scale and then subjected to double thresholding and to an adaptive threshold value determined using Otsu’s thresholding value method. In the gray-scale images, the within-class variance value corresponding to all possible threshold values for the two color classes assumed as background and foreground was calculated. The threshold value that made this variance the smallest was the optimal threshold value. This method is known as Otsu’s thresholding value method [35]. Subsequently, the edge in the thresholded image was determined using edge detection methods. After this process, the original image was cropped based on the calculated values.
  • CLAHE Transformation: In the next step, the contrast-limited adaptive histogram equalization (CLAHE) transformation in the OpenCV library was used. In the transformation used, the input image was divided into parts by the user, with each part containing a histogram within itself. The histogram of each part was then adjusted based on the histogram cropping limit entered by the user, and all parts were finally brought together to obtain a Clahe-transformed version of the input image [36,37]. New outputs were achieved by contrast equalization of the cropped images by the CLAHE method.
  • Normalization and Standardization: In the last step, the images were normalized and standardized using the image-net values.
For each class (normal/abnormal (fracture)), an original shoulder bone X-ray image and an image subjected to the abovementioned pre-processing steps are provided as examples in Figure 20.
In the figure above, section a shows an original version of a normal (negative) shoulder bone X-ray image; section b shows a cropped and CLAHE pre-processed version of this normal image; section c shows an original version of an abnormal (positive, fractured) shoulder bone X-ray image; section d shows a cropped and CLAHE pre-processed version of this abnormal image.

4.4. Classification Results

Online servers were used as hardware in the process of classification of shoulder bone X-ray images in this study. All classification processes were carried out on Google Colab with nVIDIA Tesla T4 16GB GDDR6 graphics cards. In addition, the torchvision, seaborn, scikitlearn, matplotlib, and torch packages were used. The program codes written using the PyTorch deep learning library are publicly available at https://github.com/fatihuysal88/shoulder-c (accessed on 1 March 2021). A learning rate of 0.0001, an epoch number of 40, the Adam optimizer, and the cross-entropy loss function were used in all classification procedures performed with the ResNet, ResNeXt, DenseNet, VGG, InceptionV3, and MobileNetV2 deep learning models. While the same learning rate, epoch number, optimizer, and loss function were used in the versions with SpinalNet of these models, Spinal FC was used in the classification layer. The learning rate value initially used with these models was not fixed, and it was decreased 10 times every 10 epochs to increase network learning success. The same parameters were used in the ensemble learning models.

4.4.1. Evaluation Metrics

In order to evaluate the results obtained in classification problems accurately and completely, the following must be obtained for each class: accuracy percentage, confusion matrix, precision, recall, F1-score from each classification model, ROC curves, and AUC scores. The accuracy percentage refers to what percentage of test data was correctly classified. The confusion matrix provides a table of the current status in the dataset and the number of correct and incorrect predictions in our classification model. This table contains true positive (TP), true negative (TN), false positive (FP), and false negative (FN) values. Precision refers to values predicted as positive. Recall is defined as the true positive rate. F1-score is the harmonic mean of precision and recall values. Cohen’s kappa coefficient is a statistical method used to measure the reliability of agreement between two raters. The p0 value used in this calculation refers to the result of accuracy, and the pe value refers to the probability of random agreement. The following Table 3, Table 4, Table 5 and Table 6 contain information on what TP, TN, FP, FN, and the confusion matrix refer to and how the accuracy, precision, recall, F1-score, Cohen’s-kappa score, ROC curve, and AUC scores are calculated.

4.4.2. Classification Results of 13 CNN-Based Deep Learning Models with Standard FC/Spinal FC

The training accuracy, test accuracy, precision, recall, F1-score, and Cohen’s kappa scores obtained as the result of the classification performed with each model are presented in the following figures and Table 7, Table 8, Table 9, Table 10, Table 11 and Table 12.
Figure 21 shows that the highest training accuracy in classification is achieved by VGG13 among models with Standard FC and by InceptionV3 among models with Spinal FC. The test accuracy in Figure 22, the precision in Figure 23, the recall in Figure 24, the F1-score in Figure 25, and Cohen’s kappa score in Figure 26 show that the highest scores are achieved by DenseNet169 among models with Standard FC and by DenseNet201 among models with Spinal FC.
The confusion matrix values and AUC scores achieved as a result of 26 classification procedures, performed with DenseNet, Inception, MobileNet, ResNet, ResNeXt, and VGG models with Standard FC and Spinal FC with different numbers of layers, are provided in Table 13.
The table shows that the highest classification accuracy (TP value) out of the 278 Class 1 (positive, abnormal, fracture) images in the test data is achieved by the VGG13 model with Spinal FC, with 231 images. Moreover, out of 285 images in Class 0 (negative, normal) in the test data, the highest classification accuracy (TN value) was achieved by the VGG13 model with Standard FC, with 258 images. The highest AUC score in Class 1 is 0.8797, and the highest score in Class 0 is 0.8809, and both were achieved by the DenseNet169 model with Standard FC.
In addition to the 26 classification models that were used for the classification of shoulder bone X-ray images, two different ensemble models were developed, which further improved the classification results.

4.4.3. Classification Results of Our Ensemble Models

ResNeXt50 with Spinal FC, DenseNet169 with Standard FC, and DenseNet201 with Spinal FC were used in EL1:
  • ResNeXt50 with Spinal FC was selected because the AUC score achieved by detecting fracture images is the second highest, with 0.8783, after DenseNet169 with Standard FC.
  • DenseNet169 with Standard FC was selected because its classification had the highest test accuracy, Cohen’s kappa score, and AUC score among all models used in this study.
  • DenseNet201 with Spinal FC was selected because the test accuracy and Cohen’s kappa score were the second highest after DenseNet169 with Standard FC among all models and the highest among models with Spinal FC.
The classification results of the EL1 model are provided in the following figures and tables. In Table 14, ResNeXt50b represents the ResNeXt50 model with Spinal FC, DenseNet169a represents the DenseNet169 model with Standard FC, and DenseNet201b represents the DenseNet201 model with Spinal FC.
The table above shows that, with the EL1 model, the test accuracy is 0.8455 and Cohen’s kappa score is 0.6907. Therefore, the highest classification result obtained with DenseNet169 with Standard FC given in the previous section was exceeded.
The confusion matrix specified in Figure 27 was a result of the classification performed using the EL1 model. The following results are concluded as per the figure: TN: 249, FP: 36, FN: 51, and TP: 227.
The ROC curve representing the change in the TP rate and FP rate of the EL1 model entails an AUC score of 0.8862 for Class 0 (normal) and Class 1 (abnormal, fracture) in Figure 28. This AUC score is also higher than the AUC achieved with ResNet169 with Spinal FC in the previous section.
Three different single classification models and a sub-ensemble model were used in the EL2 model: ResNet34 with Spinal FC, DenseNet201, DenseNet169 with Standard FC, and sub-ensemble models. The TP and TN values in the confusion matrix scores and recall scores from the 26 different classification models used in the first stage were taken into consideration when choosing the three single models used in the EL2 model and developing the sub-ensemble model. The results of the classification carried out with the EL2 model are provided in Table 15, where ResNet34b represents the ResNet34 model with Spinal FC, DenseNet169a represents the DenseNet169 model with Standard FC, DenseNet201b represents the DenseNet201 model with Spinal FC, and SubEnsemble represents the developed sub-ensemble model.
The table shows that Cohen’s kappa is 0.6942, and the test accuracy score is 0.8472. Therefore, the EL2 model outperforms the 26 CNN-based models and the EL1 model.
Figure 29 presents the confusion matrix scores achieved as a result of the classification carried out with the EL2 model: TN: 248, FP: 36, FN: 49 and TP: 229. For comparison, the TP in the EL1 model was 227.
The ROC curves and AUC scores achieved as a result of the classification performed with the EL2 model are provided in Figure 30. The AUC scores are 0.8698 for Class 0 (normal, negative) and 0.8695 for Class 1 (abnormal, positive, fracture). Although these AUC scores appear to be slightly decreased compared to the EL1 model, since the test accuracy, Cohen’s kappa scores, and the number of detected shoulder bone fracture images (TP) in the EL2 model are higher than the EL1 model, the EL2 model overall provides better results.
The test accuracy values obtained in classification studies vary depending on the amount and type of test data. In other classification studies on shoulder images in the MURA dataset, different approaches are used to obtain test data. Therefore, the comparison of classification results of the 28 models used in this study will be more appropriate. Comparison with the classification results obtained in other studies may be misleading due to the differences in the test data. Considering this, Table 16 of other important shoulder bone classification studies available in the literature, regarding the dataset, amount, method used, and classification accuracies.

5. Conclusions and Future Work

The aim of this study was to find the most optimal model for classifying shoulder bone X-ray images as normal or abnormal (fracture). In order to achieve this aim, the first objective was to classify fracture images using 26 classification models, i.e., Standard FC and Spinal FC versions of 13 different CNN-based models. Based on the results obtained therein, two different ensemble models (EL1 and EL2) were developed and used for classification procedures in order to further increase classification accuracy and Cohen’s kappa score. Among the 28 different classifications models, the highest test accuracy and Cohen’s kappa score were achieved in the EL2 model, and the highest AUC score was achieved in the EL1 model. In the EL1 model, the fully connected layers of the models trained on the dataset were combined and turned into a single fully connected layer. This new layer was linked to the classification layer. In this way, in the last training part performed in the EL1 model, the weights in the new fully connected layer were updated, and the best configuration of the three models was obtained. For this reason, the EL1 model can be used on different datasets because it updates the weights of the new FC layer for the best configuration on any dataset. In the EL2 model, the sub-models consulted in the decision mechanism consist of models that show the best performance on the dataset. For this reason, if one is working on their own dataset, they should determine their sub-models according to the performance results obtained with transfer learning. The reason for using transfer learning is the further improvement of classifications made with models that have not been pre-trained. With this approach, a serious increase in the classification problem was observed. Ensemble learning was mainly used to further improve the classification results obtained with 26 different deep learning models and to bring new approaches to the literature in this field. The contribution of the study to the literature is as follows.
  • In similar studies, binary classification is mostly performed. However, while there are mainly two classes (normal/abnormal) in this study, differently from the literature, multi-class classification was carried out, in order to determine the most compatible models to be used in ensemble models, developed by evaluating the outputs of each class of the 26 classification models initially used. This allowed the best results in this study to be achieved with ensemble models.
  • This is the first time that Spinal FC, which, compared to the Standard FC, has a lower number of weights in the hidden layer, was used in many models (Inception, ResNeXt, and MobileNet). Moreover, SpinalNet was used on medical images for the first time, and it had a positive effect on more than half of the classification results.
  • A unique structure is introduced, since the reliability of the detection of classes was used as a basis when designing the EL2 model, which further improves classification results.
The aim of this study, focused on the detection of shoulder bone fractures in X-ray images, is to help physicians diagnose fractures in the shoulder and apply the required treatment. Following this study, a real-time mobile application that detects fractures and/or fracture areas in shoulder bones should be developed, specifically to help physicians performing emergency services.

Author Contributions

Conceptualization, F.U., F.H., and O.P.; methodology, F.U., F.H., and O.P.; software, F.U., F.H., and O.P.; validation, F.U., F.H., and O.P.; formal analysis, F.U., F.H., and O.P.; investigation, F.U., F.H., and O.P.; resources, F.U., F.H., and O.P.; data curation, F.U., F.H., and O.P.; writing—original draft preparation, F.U., F.H., and O.P.; writing—review and editing, F.U., F.H., O.P., T.T., and N.T.; visualization, F.U., F.H., and O.P.; supervision, F.U., F.H., and O.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data used in this study are available at https://stanfordmlgroup.github.io/competitions/mura (accessed on 1 September 2020).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shoulder Fracture 2014. Available online: https://www.assh.org/handcare/condition/shoulder-fracture (accessed on 1 September 2020).
  2. Rajpurkar, P.; Irvin, J.; Bagul, A.; Ding, D.; Duan, T.; Mehta, H.; Yang, B.; Zhu, K.; Laird, D.; Ball, R.L.; et al. MURA Dataset: Towards Radiologist-Level Abnormality Detection in Musculoskeletal Radiographs. In Proceedings of the 1st Conference on Medical Imaging with Deep Learning, Amsterdam, The Netherlands, 4–6 June 2018. [Google Scholar]
  3. Guan, B.; Zhang, G.; Yao, J.; Wang, X.; Wang, M. Arm fracture detection in X-rays based on improved deep convolutional neural network. Comput. Electr. Eng. 2020, 81, 1–11. [Google Scholar] [CrossRef]
  4. Galal, A.; Hisham, F.; Mohamed, M.; Hassan, S.; Ghanim, T.; Nabil, A. Automatic Recognition of Elbow Musculoskeletal Disorders using Cloud Application. In Proceedings of the 2019 8th International Conference on Software and Information Engineering, Cairo, Egypt, 9–12 April 2019. [Google Scholar]
  5. Liang, S.; Gu, Y. Towards Robust and Accurate Detection of Abnormalities in Musculoskeletal Radiographs with a Multi-Network Model. Sensors 2020, 20, 3153. [Google Scholar] [CrossRef] [PubMed]
  6. Saif, A.F.M.; Shahnaz, C.; Zhu, W.P.; Ahmad, M.O. Abnormality Detection in Musculoskeletal Radiographs Using Capsule Network. IEEE Access 2019, 7, 81494–81503. [Google Scholar] [CrossRef]
  7. Cheng, K.; Iriondo, C.; Calivá, F.; Krogue, J.; Majumdar, S.; Pedoia, V. Adversarial Policy Gradient for Deep Learning Image Augmentation. In Proceedings of the 22nd International Conference on Medical Image Computing and Computer Assisted Intervention, Shenzhen, China, 13–17 October 2019. [Google Scholar]
  8. Pelka, O.; Nensa, F.; Friedrich, C.M. Branding-Fusion of Meta Data and Musculoskeletal Radiographs for Multi-modal Diagnostic Recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshop, Seoul, Korea, 27–28 October 2019. [Google Scholar]
  9. Varma, M.; Lu, M.; Gardner, R.; Dunnmon, J.; Khandwala, N.; Rajpurkar, P.; Long, J.; Beaulieu, C.; Shpanskaya, K.; Fei-Fei, L.; et al. Automated abnormality detection in lower extremity radiographs using deep learning. Nat. Mach. Intell. 2019, 1, 578–583. [Google Scholar] [CrossRef]
  10. Harini, N.; Ramji, B.; Sriram, S.; Sowmya, V.; Soman, K.P. Musculoskeletal radiographs classification using deep learning. In Deep Learning for Data Analytics: Foundations, Biomedical Applications and Challenges, 1st ed.; Das, H., Pradhan, C., Dey, N., Eds.; Academic Press: London, UK, 2020; pp. 79–98. [Google Scholar]
  11. Fang, L.; Jin, Y.; Huang, L.; Guo, S.; Zhao, G.; Chen, X. Iterative fusion convolutional neural networks for classification of optical coherence tomography images. J. Vis. Commun. Image Represent. 2019, 59, 327–333. [Google Scholar] [CrossRef]
  12. Mondol, T.C.; Iqbal, H.; Hashem, M. Deep CNN-Based Ensemble CADx Model for Musculoskeletal Abnormality Detection from Radiographs. In Proceedings of the 2019 5th International Conference on Advances in Electrical Engineering, Dhaka, Bangladesh, 26–28 September 2019. [Google Scholar]
  13. Pradhan, N.; Dhaka, V.S.; Chaudhary, H. Classification of Human Bones Using Deep Convolutional Neural Network. In Proceedings of the IOP Conference Series: Materials Science and Engineering, International Conference on Startup Ventures: Technology Developments and Future Strategies, Rajasthan, India, 8–9 October 2019. [Google Scholar]
  14. Shao, Y.; Wang, X. A Two Stage Method for Abnormality Diagnosis of Musculoskeletal Radiographs. In Proceedings of the International Conference on Pattern Recognition and Artificial Intelligence, Zhongshan, China, 19–23 October 2020. [Google Scholar]
  15. Chung, S.W.; Han, S.S.; Lee, J.W.; Oh, K.S.; Kim, N.R.; Yoon, J.P.; Kim, J.Y.; Moon, S.H.; Kwon, J.; Lee, H.J.; et al. Automated detection and classification of the proximal humerus fracture by using deep learning algorithm. Acta Orthop. 2018, 89, 468–473. [Google Scholar] [CrossRef] [Green Version]
  16. Sezer, A.; Sezer, H.B. Convolutional neural network based diagnosis of bone pathologies of proximal humerus. Neurocomputing 2020, 392, 124–131. [Google Scholar] [CrossRef]
  17. Urban, G.; Porhemmat, S.; Stark, M.; Feeley, B.; Okada, K.; Baldi, P. Classifying shoulder implants in X-ray images using deep learning. Comput. Struct. Biotechnol. J. 2020, 18, 967–972. [Google Scholar] [CrossRef]
  18. Sezer, A.; Sigirci, I.O.; Sezer, H.B. Shoulder lesion classification using shape and texture features via composite kernel. In Proceedings of the 25th Signal Processing and Communications Applications Conference (SIU), Antalya, Turkey, 15–18 May 2017. [Google Scholar]
  19. Sezer, A.; Sezer, H.B. Capsule network-based classification of rotator cuffpathologies from MRI. Comput. Electr. Eng. 2019, 80, 106480. [Google Scholar] [CrossRef]
  20. Khan, M.A.; Kim, Y. Cardiac Arrhythmia Disease Classification Using LSTM Deep Learning Approach. Comput. Mater. Contin. 2021, 67, 427–443. [Google Scholar] [CrossRef]
  21. Storey, O.; Wei, B.; Zhang, L.; Mtope, F.R.F. Adaptive bone abnormality detection in medical imagery using deep neural networks. In Proceedings of the 14th International FLINS Conference, Cologne, Germany, 18–21 August 2020. [Google Scholar]
  22. Yin, S.; Peng, Q.; Li, H.; Zhang, Z.; You, X.; Liu, H.; Fischer, K.; Furth, S.L.; Tasian, G.E.; Fan, Y. Multi-instance Deep Learning with Graph Convolutional Neural Networks for Diagnosis of Kidney Diseases Using Ultrasound Imaging. In Proceedings of the International Workshop on Uncertainty for Safe Utilization of Machine Learning in Medical Imaging, Lima, Peru, 17 October 2019. [Google Scholar]
  23. Dias, D.D.A. Musculoskeletal Abnormality Detection on X-ray Using Transfer Learning. Master’s Thesis, Pompeu Fabra University, Barcelona, Spain, July 2019. [Google Scholar]
  24. Khan, M.A.; Kim, J. Toward Developing Efficient Conv-AE-Based Intrusion Detection System Using Heterogeneous Dataset. Electronics 2020, 9, 1771. [Google Scholar] [CrossRef]
  25. Kegelman, C.D.; Nijsure, M.P.; Moharrer, Y.; Pearson, H.B.; Dawahare, J.H.; Jordan, K.M.; Qin, L.; Boerckel, J.D. YAP and TAZ Promote Periosteal Osteoblast Precursor Expansion and Differentiation for Fracture Repair. J. Bone Miner. Res. 2020, 36, 143–157. [Google Scholar] [CrossRef] [PubMed]
  26. Sharma, A.; Mishra, A.; Bansal, A.; Bansal, A. Bone Fractured Detection Using Machine Learning and Digital Geometry. In Mobile Radio Communications and 5G Networks, Lecture Notes in Networks and Systems, 1st ed.; Marriwala, N., Tripathi, C.C., Kumar, D., Jain, S., Eds.; Springer: Singapore, 2021; Volume 140, pp. 369–376. [Google Scholar]
  27. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  28. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  29. Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated Residual Transformations for Deep Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  30. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  31. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  32. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  33. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  34. Kabir, H.M.D.; Abdar, M.; Jalali, S.M.J.; Khosravi, A.; Atiya, A.F.; Nahavandi, S.; Srinivasan, D. SpinalNet: Deep neural networkwith gradual input. arXiv 2020, arXiv:2007.03347. [Google Scholar]
  35. Otsu, N. A Threshold Selection Method from Gray-Level Histograms. IEEE Trans. Syst. Man Cybern. 1979, 9, 62–66. [Google Scholar] [CrossRef] [Green Version]
  36. Pizer, S.M.; Johnston, R.E.; Ericksen, J.P.; Yankaskas, B.C.; Muller, K.E. Contrast-limited adaptive histogram equalization: Speed and effectiveness. In Proceedings of the First Conference on Visualization in Biomedical Computing, Atlanta, GA, USA, 22–25 May 1990. [Google Scholar]
  37. Zuiderveld, K. Contrast limited adaptive histogram equalization. In Graphics Gems IV; Heckbert, P.S., Ed.; Academic Press Professional: San Diego, CA, USA, 1994; pp. 474–485. [Google Scholar]
Figure 1. The anatomy of the shoulder bone.
Figure 1. The anatomy of the shoulder bone.
Applsci 11 02723 g001
Figure 2. Standard FC/Spinal FC CNN-based deep learning methods subject to transfer learning.
Figure 2. Standard FC/Spinal FC CNN-based deep learning methods subject to transfer learning.
Applsci 11 02723 g002
Figure 3. ResNet building block. (a) BB-I, k; (b) BB-II, m, n [28].
Figure 3. ResNet building block. (a) BB-I, k; (b) BB-II, m, n [28].
Applsci 11 02723 g003
Figure 4. The structure of the adapted ResNet-based models. (a) ResNet34; (b) ResNet50; (c) ResNet101; (d) ResNet152.3.1.2. ResNeXt [28].
Figure 4. The structure of the adapted ResNet-based models. (a) ResNet34; (b) ResNet50; (c) ResNet101; (d) ResNet152.3.1.2. ResNeXt [28].
Applsci 11 02723 g004aApplsci 11 02723 g004b
Figure 5. ResNeXt building block (BB-III, r, s, C = 32) [29].
Figure 5. ResNeXt building block (BB-III, r, s, C = 32) [29].
Applsci 11 02723 g005
Figure 6. The structure of the adapted ResNeXt-based models. (a) ResNeXt50 (32 × 4d); (b) ResNeXt101 (32 × 8d) [29].
Figure 6. The structure of the adapted ResNeXt-based models. (a) ResNeXt50 (32 × 4d); (b) ResNeXt101 (32 × 8d) [29].
Applsci 11 02723 g006
Figure 7. The structure of the adapted DenseNet-based models. (a) DenseNet169; (b) DenseNet201 [30].
Figure 7. The structure of the adapted DenseNet-based models. (a) DenseNet169; (b) DenseNet201 [30].
Applsci 11 02723 g007
Figure 8. The conv blocks of 2, 3, and 4 3 × 3 conv layers (CB-I, II, and III, respectively). (a) CB-I, k; (b) CB-II, m; (c) CB-III, n [31].
Figure 8. The conv blocks of 2, 3, and 4 3 × 3 conv layers (CB-I, II, and III, respectively). (a) CB-I, k; (b) CB-II, m; (c) CB-III, n [31].
Applsci 11 02723 g008
Figure 9. The structure of the adapted VGG-based models. (a) VGG13; (b) VGG16; (c) VGG19 [31].
Figure 9. The structure of the adapted VGG-based models. (a) VGG13; (b) VGG16; (c) VGG19 [31].
Applsci 11 02723 g009
Figure 10. The inception blocks containing convolution factorization (IB-A, B, C, D, and E). (a) IB-A; (b) IB-B; (c) IB-C; (d) IB-D; (e) IB-E [32].
Figure 10. The inception blocks containing convolution factorization (IB-A, B, C, D, and E). (a) IB-A; (b) IB-B; (c) IB-C; (d) IB-D; (e) IB-E [32].
Applsci 11 02723 g010
Figure 11. The structure of the adapted InceptionV3-based model [32].
Figure 11. The structure of the adapted InceptionV3-based model [32].
Applsci 11 02723 g011
Figure 12. MobileNet bottleneck blocks where stride = 1 and stride = 2, respectively (MB-I and -II). (a) MB-I; (b) MB-II [33].
Figure 12. MobileNet bottleneck blocks where stride = 1 and stride = 2, respectively (MB-I and -II). (a) MB-I; (b) MB-II [33].
Applsci 11 02723 g012
Figure 13. The structure of the adapted MobileNetV2-based model [33].
Figure 13. The structure of the adapted MobileNetV2-based model [33].
Applsci 11 02723 g013
Figure 14. The proposed EL1 classification model.
Figure 14. The proposed EL1 classification model.
Applsci 11 02723 g014
Figure 15. A block diagram of the proposed EL1 ensemble model.
Figure 15. A block diagram of the proposed EL1 ensemble model.
Applsci 11 02723 g015
Figure 16. The proposed EL2 classification model.
Figure 16. The proposed EL2 classification model.
Applsci 11 02723 g016
Figure 17. A block diagram of the proposed EL2 ensemble model.
Figure 17. A block diagram of the proposed EL2 ensemble model.
Applsci 11 02723 g017
Figure 18. The proposed models for classification of shoulder bone X-ray images.
Figure 18. The proposed models for classification of shoulder bone X-ray images.
Applsci 11 02723 g018
Figure 19. The shoulder bone X-ray dataset [2].
Figure 19. The shoulder bone X-ray dataset [2].
Applsci 11 02723 g019
Figure 20. Original and pre-processed X-ray images of normal/abnormal (fractured) shoulder bone. (a) original normal; (b) pre-processed normal; (c) original abnormal; (d) pre-processed abnormal [2].
Figure 20. Original and pre-processed X-ray images of normal/abnormal (fractured) shoulder bone. (a) original normal; (b) pre-processed normal; (c) original abnormal; (d) pre-processed abnormal [2].
Applsci 11 02723 g020
Figure 21. Training accuracy results of classification models.
Figure 21. Training accuracy results of classification models.
Applsci 11 02723 g021
Figure 22. Test accuracy results of classification models.
Figure 22. Test accuracy results of classification models.
Applsci 11 02723 g022
Figure 23. Precision results of classification models.
Figure 23. Precision results of classification models.
Applsci 11 02723 g023
Figure 24. Recall results of classification models.
Figure 24. Recall results of classification models.
Applsci 11 02723 g024
Figure 25. F1-score results of classification models.
Figure 25. F1-score results of classification models.
Applsci 11 02723 g025
Figure 26. Cohen’s kappa results of classification models.
Figure 26. Cohen’s kappa results of classification models.
Applsci 11 02723 g026
Figure 27. Confusion matrix results of EL1.
Figure 27. Confusion matrix results of EL1.
Applsci 11 02723 g027
Figure 28. ROC curves and AUC results of EL1.
Figure 28. ROC curves and AUC results of EL1.
Applsci 11 02723 g028
Figure 29. Confusion matrix results of EL2.
Figure 29. Confusion matrix results of EL2.
Applsci 11 02723 g029
Figure 30. ROC curves and AUC results of EL2.
Figure 30. ROC curves and AUC results of EL2.
Applsci 11 02723 g030
Table 1. Layer with values of Spinal FCs used in classification models.
Table 1. Layer with values of Spinal FCs used in classification models.
ModelsSpinal FC Layer WidthModelsSpinal FC Layer Width
ResNet34256ResNeXt5020
ResNet50128ResNeXt101128
ResNet101,1521024DenseNet169, 201240
VGG13256MobileNetV2320
VGG16,19512InceptionV320
Table 2. Details of the shoulder bone X-ray images used in the study [2].
Table 2. Details of the shoulder bone X-ray images used in the study [2].
Shoulder Bone
X-ray Images
Image TypesTrain
Dataset
Test
Dataset
Org.
Image Size
New Image
Size
Class 0:
Normal (Negative) Bone X-ray Images
png,
3-ch.
4211 285 various 320 × 320 × 3
Class 1:
Abnormal (Positive)
Bone X-ray Images
png,
3-ch.
4168 278 various 320 × 320 × 3
Totalpng,
3-ch.
8379563various320 × 320 × 3
Table 3. TP, TN, FP, and FN calculations.
Table 3. TP, TN, FP, and FN calculations.
Label ValuePrediction Value
TPpositivepositive
TNnegativenegative
FPpositivenegative
FNnegativepositive
Table 4. Confusion matrix.
Table 4. Confusion matrix.
Predicted Class NegativePredicted Class Positive
Actual Class: NegativeTNFP
Actual Class: PositiveFNTP
Table 5. Calculation of Cohen’s kappa score parameters.
Table 5. Calculation of Cohen’s kappa score parameters.
P0(TP + TN)/(TP + TN + FP + FN)
Ppositive(TP + FP)(TP + FN)/(TP + TN + FP + FN)2
Pnegative(FN + TN)(FP + TN)/(TP + TN + FP + FN)2
PePpositive + Pnegative
Table 6. Evaluation metrics.
Table 6. Evaluation metrics.
Confusion MatrixTP, TN, FP, FN
Training/Testing Accuracy(TP + TN)/(TP + TN + FP + FN)
PrecisionTP/(TP + FP)
RecallTP/(TP + FN)
F1-score2TP/(2TP + FP + FN)
Cohen’s kappa score(p0 − pe)/(1 − pe)
ROC curveTP rate—FP rate change
AUC scoresArea under the ROC curve
Table 7. Training accuracy results of classification models.
Table 7. Training accuracy results of classification models.
ModelsStandart
FC
Spinal
Net
ModelsStandart
FC
Spinal
Net
DenseNet1690.88830.8666ResNet1010.84190.8225
DenseNet2010.89340.8694ResNet1520.8450.8346
InceptionV30.88820.8914ResNeXt500.87070.8654
MobileNetV20.86470.8845ResNeXt1010.85390.8303
ResNet340.86730.8517VGG130.93890.8507
ResNet500.84890.8395VGG160.86980.812
VGG190.90550.8011
Table 8. Test accuracy results of classification models.
Table 8. Test accuracy results of classification models.
ModelsStandart
FC
Spinal
Net
ModelsStandart
FC
Spinal
Net
DenseNet1690.84190.8152ResNet1010.8170.8188
DenseNet2010.82060.8294ResNet1520.81170.817
InceptionV30.82590.817ResNeXt500.8170.8241
MobileNetV20.82410.8099ResNeXt1010.82060.8082
ResNet340.81880.8206VGG130.77970.8223
ResNet500.81880.8081VGG160.7850.801
VGG190.7850.8046
Table 9. Precision results of classification models.
Table 9. Precision results of classification models.
ModelsStandart
FC
Spinal
Net
ModelsStandart
FC
Spinal
Net
DenseNet1690.8450.815ResNet1010.8150.82
DenseNet2010.820.83ResNet1520.8150.82
InceptionV30.8250.82ResNeXt500.820.825
MobileNetV20.830.81ResNeXt1010.820.815
ResNet340.8250.82VGG130.80.825
ResNet500.8150.81VGG160.790.805
VGG190.7850.81
Table 10. Recall results of classification models.
Table 10. Recall results of classification models.
ModelsStandart
FC
Spinal
Net
ModelsStandart
FC
Spinal
Net
DenseNet1690.840.815ResNet1010.8150.82
DenseNet2010.8150.83ResNet1520.810.815
InceptionV30.8250.815ResNeXt500.8150.825
MobileNetV20.820.81ResNeXt1010.8150.815
ResNet340.8150.82VGG130.780.825
ResNet500.820.81VGG160.7850.8
VGG190.7850.805
Table 11. F1-score results of classification models.
Table 11. F1-score results of classification models.
ModelsStandart
FC
Spinal
Net
ModelsStandart
FC
Spinal
Net
DenseNet1690.840.815ResNet1010.820.82
DenseNet2010.820.83ResNet1520.810.815
InceptionV30.8250.815ResNeXt500.8150.825
MobileNetV20.820.81ResNeXt1010.820.815
ResNet340.820.82VGG130.7750.82
ResNet500.8150.805VGG160.7850.8
VGG190.7850.805
Table 12. Cohen’s kappa results of classification models.
Table 12. Cohen’s kappa results of classification models.
ModelsStandart
FC
Spinal
Net
ModelsStandart
FC
Spinal
Net
DenseNet1690.68340.6302ResNet1010.6340.6372
DenseNet2010.6410.6588ResNet1520.62310.6332
InceptionV30.65140.6338ResNeXt500.63380.648
MobileNetV20.64780.6195ResNeXt1010.6410.6267
ResNet340.63720.6411VGG130.5580.6442
ResNet500.63750.6161VGG160.56950.6014
VGG190.56980.6085
Table 13. Confusion matrix values and AUC scores of 26 classification models (a: with Standard FC, b: with Spinal FC).
Table 13. Confusion matrix values and AUC scores of 26 classification models (a: with Standard FC, b: with Spinal FC).
ModelsTPFPFNTNClass0: AUCClass1: AUC
DenseNet169a22233562520.88090.8797
DenseNet169b21844602410.86020.8598
DenseNet201a22548532370.85840.8653
DenseNet201b22846502390.87270.8724
InceptionV3a21838602470.87540.8707
InceptionV3b22045582400.85820.8585
MobileNetV2a21536632490.87770.8378
MobileNetV2b21645622400.86330.861
ResNet34a21539632460.87050.8767
ResNet34b22851502340.86170.8619
ResNet50a22448542370.87150.8662
ResNet50b21949592360.85880.8584
ResNet101a22853502320.86830.8703
ResNet101b21640622450.86090.861
ResNet152a21745612400.86480.8701
ResNet152b22045582400.85970.8606
ResNeXt50a22146572390.86440.8699
ResNeXt50b22344552410.87890.8783
ResNeXt101a22548532370.87720.8765
ResNeXt101b21946592390.86520.8561
VGG13a18127972580.84060.8415
VGG13b23135652500.87050.8737
VGG16a20346752390.85170.8523
VGG16b20438742470.85420.857
VGG19a21255662300.83740.8502
VGG19b20537732480.8580.8539
Table 14. Classification results of EL1.
Table 14. Classification results of EL1.
ModelsTest Acc.Pre.RecallF1-ScoreCohen’s Kappa
ResNeXt50b0.82410.8250.8250.8250.648
DenseNet169a 0.84190.8450.840.840.6834
DenseNet201b 0.82940.830.830.830.6588
EL10.84550.86310.81650.84550.6907
Table 15. Classification results of EL2.
Table 15. Classification results of EL2.
ModelsTest Acc.Pre.RecallF1-scoreCohen’s Kappa
ResNet34b0.82060.820.820.820.6411
DenseNet169a0.84190.8150.840.840.6834
DenseNet201b0.82940.830.830.830.6588
SubEnsemble0.84010.840.840.840.6799
EL20.84720.850.8450.8450.6942
Table 16. Comparison with previous studies.
Table 16. Comparison with previous studies.
StudiesDatasetBest MethodClassification Type and Test Accuracy
Our studyMURA Shoulder, 563 validation/test images, open dataEL2Fracture/normal: 0.8472
Liang and Gu [5]MURA dataset, 194 validation imagesGCNFracture/normal: 0.9112
Saif et al. [6]MURA dataset, various test images, %50 test data, test quantity not specifiedCapsule NetworkFracture/normal: 0.9208
Chung et al. [15]1376 CT imagesResNet152Normal/four different type: 0.96
Urban et al. [17]597 X-ray imagesNASNetImplant: 0.804
Sezers [16]219 MR imagesCNNNormal/edematous/Hill–Sachs lesions: 0.9843
Sezers [18]219 MR imagesExtreme learning machinesNormal/edematous/Hill–Sachs lesions: 0.94
Sezers [19]1006 MR imagesCapsNetNormal/degenerated/torn: 0.9474
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Uysal, F.; Hardalaç, F.; Peker, O.; Tolunay, T.; Tokgöz, N. Classification of Shoulder X-ray Images with Deep Learning Ensemble Models. Appl. Sci. 2021, 11, 2723. https://doi.org/10.3390/app11062723

AMA Style

Uysal F, Hardalaç F, Peker O, Tolunay T, Tokgöz N. Classification of Shoulder X-ray Images with Deep Learning Ensemble Models. Applied Sciences. 2021; 11(6):2723. https://doi.org/10.3390/app11062723

Chicago/Turabian Style

Uysal, Fatih, Fırat Hardalaç, Ozan Peker, Tolga Tolunay, and Nil Tokgöz. 2021. "Classification of Shoulder X-ray Images with Deep Learning Ensemble Models" Applied Sciences 11, no. 6: 2723. https://doi.org/10.3390/app11062723

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop