Next Article in Journal
COVID-19 Infection during Pregnancy: Risk of Vertical Transmission, Fetal, and Neonatal Outcomes
Next Article in Special Issue
Opinion Piece: Patient-Specific Implants May Be the Next Big Thing in Spinal Surgery
Previous Article in Journal
Targeted Approaches to T-Cell Lymphoma
Previous Article in Special Issue
The Orthopedic-Vascular Multidisciplinary Approach Improves Patient Safety in Surgery for Musculoskeletal Tumors: A Large-Volume Center Experience
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Artificial Intelligence-Based Recognition of Different Types of Shoulder Implants in X-ray Scans Based on Dense Residual Ensemble-Network for Personalized Medicine

Division of Electronics and Electrical Engineering, Dongguk University, 30 Pildong-ro 1-gil, Jung-gu, Seoul 04620, Korea
*
Author to whom correspondence should be addressed.
J. Pers. Med. 2021, 11(6), 482; https://doi.org/10.3390/jpm11060482
Submission received: 31 March 2021 / Revised: 15 May 2021 / Accepted: 24 May 2021 / Published: 27 May 2021
(This article belongs to the Special Issue Patient-Specific Implants in Musculoskeletal (Orthopedic) Surgery)

Abstract

:
Re-operations and revisions are often performed in patients who have undergone total shoulder arthroplasty (TSA) and reverse total shoulder arthroplasty (RTSA). This necessitates an accurate recognition of the implant model and manufacturer to set the correct apparatus and procedure according to the patient’s anatomy as personalized medicine. Owing to unavailability and ambiguity in the medical data of a patient, expert surgeons identify the implants through a visual comparison of X-ray images. False steps cause heedlessness, morbidity, extra monetary weight, and a waste of time. Despite significant advancements in pattern recognition and deep learning in the medical field, extremely limited research has been conducted on classifying shoulder implants. To overcome these problems, we propose a robust deep learning-based framework comprised of an ensemble of convolutional neural networks (CNNs) to classify shoulder implants in X-ray images of different patients. Through our rotational invariant augmentation, the size of the training dataset is increased 36-fold. The modified ResNet and DenseNet are then combined deeply to form a dense residual ensemble-network (DRE-Net). To evaluate DRE-Net, experiments were executed on a 10-fold cross-validation on the openly available shoulder implant X-ray dataset. The experimental results showed that DRE-Net achieved an accuracy, F1-score, precision, and recall of 85.92%, 84.69%, 85.33%, and 84.11%, respectively, which were higher than those of the state-of-the-art methods. Moreover, we confirmed the generalization capability of our network by testing it in an open-world configuration, and the effectiveness of rotational invariant augmentation.

1. Introduction

The human shoulder is the most mobile joint of the body. The shoulder may be damaged owing to severe fractures or injuries to the upper arm or severe joint infection. Shoulder surgery is needed when damage to the shoulder joint progresses to such an extent that non-operative procedures cannot resolve the issue or the joint movement causes severe pain. According to the Agency for Healthcare Research and Quality, 53,000 Americans undergo shoulder replacement surgery each year [1]. Total shoulder arthroplasty (TSA) and reverse total shoulder arthroplasty (RTSA) [2] are medical procedures for treating arthritic shoulder joints. With this treatment, a prosthesis is used to repair the damaged joint of the shoulder to re-establish movement and reduce pain. TSA and RTSA are critical for shoulder pain in osteoarthritis. Proper preoperative preparation can help avoid many complications in the revision of TSA and RTSA.
One key surgical step that helps avoid more common complications is identifying prostheses to properly position them. As the morphology of the human shoulder varies from person to person, prostheses are comprised of fixtures and superstructures that can vary by their model, structure, and manufacturer. Therefore, the “one size fits all” idea is not suitable for the treatment of shoulder arthroplasty. Therefore, selecting the correct prostheses model from the right manufacturer for the right patient is very important as personalized medicine. Designing a framework for automatic selection of suitable prostheses for a patient would allow the surgeons to conduct prior and more effective decision-making.
There are many different combinations of device characteristics and surgical approaches, and surgeons often deal with a small number of implants at a time to maximize their expertise with the technology [3]. With a lack of comparable data, surgeons choose which from the few implants they currently offer are appropriate solutions for each patient, rather than choosing from the whole range of alternatives available on the market. However, in some clinical situations, surgeons may believe that only one device is the best option. Older patients, for example, are unlikely to gain additional benefits from a newer implant, but they are at higher risk of surgical problems than younger patients if revision is required [4]. In such cases, selecting a particular implant is crucial. Owing to the limited experience of surgeons with limited implants models, this makes them difficult to work in such situations. Moreover, implants are not identified by medical doctors due to incoherence in documentation and global limitations relating to access to such records, in particular by outside hospital systems [5]. With time, some models of former implants have been halted and their production cut off, whereas new models that differ somewhat from the prior models are being introduced by manufacturers. Moreover, the inclinations of doctors toward certain prostheses change over time. In an investigation carried out by arthroplasty surgeons, 88% of surgeons indicated that implant identification is a critical obstacle to the treatment of an arthroplasty patient [6]. Different prosthesis models require different systems and equipment for replacement and repair, and accurate identification of the model is mandatory. Failure to identify the correct model before surgery results in a waste of healthcare resources, time, and the health of the patient. In some situations, the manufacture and model of the implant might be obscure to surgeons and patients, for example when the original medical procedure is performed outside of the county, and the patients are unable to access their medical records. Over 40% of patients in institutions other than their original arthroplasty are less likely to access outside medical records in a timely manner [5]. As for other reasons why the prosthesis model and manufacturer are unknown, the first original surgery might be performed numerous years before the subsequent surgery, and the patient’s medical information might become lost or unclear. In these cases, medical experts identify a prosthesis through a visual comparison of X-ray images and an implant atlas [7]. This task is tedious, time-consuming, dependent on the surgeon’s experience, and an erroneous recognition can have certain consequences. Therefore, there is a need for an automated method for the identification of prostheses to aid surgeons with pre-operative planning and to save time and medical costs. However, high intra-class variabilities and low inter-class variabilities in shoulder implants appear in X-ray images, as shown in Figure 1, which makes this research extremely challenging.
Despite significant advancements in pattern recognition and deep learning (DL) in the medical field, there has been extremely limited research conducted on classifying shoulder implants. To address these issues, we propose a robust deep-learning-based framework comprising an ensemble of convolutional neural networks (CNNs) to classify shoulder implants in X-ray images. Compared to previous studies, our research is novel in the following five ways.
  • To effectively identify shoulder implants, we propose a dense residual ensemble-network (DRE-Net) comprising two CNN models and a shallow concatenation network (SCN). Our network achieves a higher accuracy compared with state-of-the-art studies.
  • We propose a rotational invariant augmentation (RIA) to tackle the overfitting problem.
  • To check the generalization capability of our network, the proposed DRE-Net is analyzed in different configuration modes of open and closed worlds.
  • We analyzed the impact of end-to-end and sequential training of DRE-Net on the testing accuracy of shoulder implant images.
  • Our model is publicly available [8] for a fair comparison by other researchers.
The remainder of this paper proceeds as follows. In Section 2, related studies on the classification of different prostheses are described. Section 3 details our proposed classification framework for shoulder implants. In Section 4, the experimental setup and results are presented. Finally, the discussion and conclusions are presented in Section 5 and Section 6, respectively.

2. Related Works

Previous studies on implant recognition have classified handcrafted feature-based and deep feature-based methods. Prior to the approach of DL strategies, previous studies have considered handcrafted feature-based methods for implant identification [9,10,11].
DL models have recently contributed pivotal additions in different clinical areas [12,13], including lesion classification [14,15], lesion detection [16,17,18], and lesion segmentation [19,20,21,22]. DL also affects every clinical specialty, including orthopedic surgery [23,24]. Plain film radiographs have been subjected to highly developed DL methods for identification of the elbow, wrist, ankle, and humerus; classification of the hip fracture types and proximal humerus; detection of the presence of arthroplasty and its type; detection of aseptic loosening; and staging the severity of knee osteoarthritis; among other applications [25,26,27,28,29,30,31]. In [32], a DL system was proposed to classify the knee implants of three datasets. The authors used variants of the residual network (ResNet) for different datasets and conducted a classification of two manufacturers and two models. Their network is trained to recognize only two classes, which limits its generalizability. In [33], the authors achieved 99% accuracy by using an artificial intelligence-based DL model to classify knee implants from four manufacturers. In [34], the authors used the visual geometry group (VGG)-16 and VGG-19 models by applying transfer learning to classify dental implants in panoramic X-ray images. Transfer learning with pre-trained networks is effective for learning richer features from large datasets to a small dataset to achieve a high level of accuracy. They manually segmented the panoramic images, and their network was unable to detect the uncropped panoramic image.
In [35], the authors used different CNN models, including SqueezeNet [36], GoogLeNet [37], ResNet-18 [38], MobileNet-v2 [39], and ResNet-50 [38] for the classification of dental implants in X-ray images. They used transfer learning with these pre-trained networks and achieved an accuracy of 90%–97%. In [40], they used a dense convolutional network (DenseNet)-201 [41] CNN with transfer learning to classify three total hip replacement prosthesis models in X-ray images with 100% accuracy. They implemented DenseNet-201 using two different weight initialization methods: (1) a random Gaussian distribution and (2) pre-trained weights of a CNN on the ImageNet database [42]. They demonstrated that a pretrained CNN cannot learn to identify the implant design in X-ray images well. DL also plays a vital role in the detection and classification of bone fractures [27,43]. However, this study was limited to a binary classification of broken and unbroken bones. In [44], a computer-assisted diagnosis (CAD) system based on a hierarchical CNN was designed for the classification of different types of fractures in X-ray images. However, in the case of some classes, the accuracy does not meet the expectations of physicians, and the system still needs to be improved for the classification of subclasses. A deep learning-based study was conducted on the classification of shoulder implants by four manufacturers, where the authors presented comparisons of DL models with different classifiers [45]. Nevertheless, the experiments were only conducted for a closed-world problem. They used the transfer-learning method and did not involve an open-world setting to address real-world problems. In [46], DL was used for the binary classification of shoulder implant models. They used a transfer learning approach and fine-tuned ResNet-18 for binary classification of the existence of arthroplasty implants. Similarly, they used the same approach to distinguish between TSA and RTSA. Finally, they used five fine-tuned models based on ResNet-152 to classify the five TSA models in a binary fashion. However, there is a possibility for an image to be labeled for multiple classes using this method.
To overcome these problems, we propose DRE-Net comprised of two deep CNNs and an SCN to classify shoulder implants in X-ray images. We considered a total of four different classes by manufacturers of 597 unidentified patients related to shoulder implants. We propose a deep feature-based framework for the accurate identification of shoulder implants to ease surgeons. We also address the open-world configuration and found that our model has the capability of generalizability and is therefore applicable to real-world problems.
Table 1 shows comparisons of the strengths and weaknesses of previous studies and our approach for the recognition of implants in X-ray images.

3. Proposed Methods

3.1. Overview of Proposed Method

Figure 2 shows the overall procedure of our proposed method of shoulder implant classification. During the training phase, input images of 224 × 224 × 3 were augmented using the proposed RIA. This technique artificially increases the number of training datasets by the in-plane rotation of each image from 0° to 360° with an interval of 10°. In this way, in addition to the original image, we obtained 36 augmented images from one input. Training is then applied with the proposed DRE-Net, including a modified ResNet-50, a modified DenseNet-201, and an SCN for feature concatenation. During the testing phase, an image is input into the trained DRE-Net, and the final classification of the shoulder implant is conducted based on the output of DRE-Net. Detailed explanations of the proposed RIA and DRE-Net are presented in Section 3.2 and Section 3.3, respectively.

3.2. Rotational Invariant Augmentation (RIA)

The performance of a deep CNN on a dataset, including a small number of images, usually suffers from many different problems, such as an overfitting and a lack of generality. To address this issue, data augmentation has been proposed. Data augmentation includes setting up strategies that upgrade the size and worth of the training dataset with an end goal in which better DL models can be assembled utilizing such strategies [47]. Therefore, we augmented our training dataset based on the in-plane rotation. As a reason for using the in-plane rotation scheme, our dataset consists of implanted shoulder prostheses with rod-like shapes that are easily in-plane rotated in the captured X-ray images, as shown in Figure 1. Data augmentation by an in-plane rotation is applied on each image by rotating the image based on an image center of between 0° and 360°, with an interval of 10°. In this way, we obtained each image with 36 postures at different angles. Figure 3 shows the RIA samples of one image from the Cofield class.

3.3. Classification of Shoulder Implants by DRE-Net

In machine learning, ensemble strategies merge various learning algorithms to achieve a preferable performance over any of the constituent models alone [48,49]. In the general frameworks of image classification, the main element is the optimum representation of the visual details or features. Based on this, we propose DRE-Net for the classification of shoulder implants, as shown in Figure 4. In the first stage of DRE-Net, an input image of 224 × 224 × 3 is input to two CNNs of modified ResNet-50 and DenseNet-201, which are modified by removing the fully connected layer (FCL) to extract the optimum features. Explanations of the first stage based on modified ResNet-50 and DenseNet-201 are presented in Section 3.3.1 and Section 3.3.2, respectively. In the second stage of DRE-Net, the SCN obtains two feature vectors (f1 and f2 of Figure 4) from the first-stage networks. These features are then concatenated and passed through the FCL and SoftMax layers to classify the shoulder implant into one of the four manufacturers. Detailed explanations of our developed SCN are presented in Section 3.3.3.

3.3.1. Feature Extraction Using Modified ResNet-50

Deep CNNs have demonstrated extreme power in representation learning because they learn the features on a pre-training task and transmit effective knowledge to the target tasks. AlexNet [50], VGG, GoogLeNet, ResNet, and DenseNet are commonly used deep CNNs for transfer learning. The experiments showed that constructing a deep network by copying layers from a learned shallow model leads to a high training error owing to vanishing gradient problems [38]. The residual network has an identity shortcut connection that skips some layers and therefore assists in shielding the network from vanishing gradient issues and improving the performance by deepening the network. Residual nets [38] were first placed in the ImageNet competition [51] for classification, localization, detection, and scoring the first position in common objects in context (COCO) competition for detection and segmentation. In our work, a state-of-the-art deep learning model of ResNet-50 pre-trained on the ImageNet dataset [42] was modified to extract the features for the classification of shoulder implant images.
As shown in Table 2, an image with a resolution of 224 × 224 × 3 was given as an input to the first layer labeled “Image Input.” The second layer labeled “Conv 1” was comprised of 64 filters of 7 × 7 × 3 , which exploits the input image. The convolution layer is a max-pooling layer, which reduced the dimensions of the feature map to a pixel resolution of 56 × 56 × 64 . Following the max-pooling layer, the layers were grouped into four residual blocks. Each residual block was comprised of two layers of a 1 × 1 convolution and one layer of a 3 × 3 convolution. The first group of layers labeled “Conv 2_x” were comprised of three residual blocks, which processed the feature map and down-sampled it to a pixel resolution of 56 × 56 × 256 . The output feature map of “Conv 2_x” was processed by the second group of layers labeled “Conv 3_x.” This group contained four residual blocks and output a feature map with a pixel resolution of 28 × 28 × 512 . Similarly, the third group of layers, labeled “Conv 4_x,” contained six residual blocks. It processed the feature map of “Conv 3_x” and generated a feature map with a pixel resolution of 14 × 14 × 1024 . The last group of layers labeled “Conv 5_x” contained three residual blocks. It processed the feature map of the previous layer and produced a 7 × 7 × 2048 sized feature map. Finally, the last average pooling layer named “Average Pooling” was applied with a filter size of 7 × 7 pixels and obtained a spatial feature vector f 1 of 1 × 1 × 2048 . The last three layers of ResNet, labeled “FCL,” “SoftMax,” and “Classification Output” were removed in our modified ResNet to enhance the training convergence and extract only features not considering the classification.

3.3.2. Feature Extraction Using Modified DenseNet-201

With the rapid advancement of CNNs, they are becoming deeper, and the problem of a vanishing gradient has emerged. One solution to this problem is to introduce skip connections between layers, as in the ResNet model. These skip connections guarantee an efficient data stream among the layers in the network. To ensure the stream of maximum information among layers, all layers are associated legitimately with one another, and each layer acquires extra inputs from prior layers and gives its feature map to every single ensuing layer in the DenseNet model [41]. In our work, a state-of-the-art DenseNet-201 pre-trained on the ImageNet dataset [42] was modified to derive the features and classify the shoulder implant images.
As shown in Table 3, an image with a pixel resolution of 224 × 224 × 3 was given as an input to the first input layer called an “Image Input.” The second layer, named “Conv 1,” was comprised of 64 filters of 7 × 7 × 3 , which exploited the input image. Following the convolution layer was a max-pooling layer, which reduced the dimensions of the feature map to 56 × 56 × 64 pixels. The layers were then grouped into four dense blocks. Each dense block included a three-sequential composite function with a convolution of 3 × 3 , a rectified linear unit (ReLU) [52], and batch normalization (BN) [53]. The first group of layers, labeled “DenseBlock_1”, which were comprised of six dense blocks, processed the feature map and down-sampled it to a pixel resolution of 28 × 28 × 128 . The output feature map of “DenseBlock_1” was processed by the second group of layers, labeled “DenseBlock_2.” This group contained 12 dense blocks and output a feature map with a pixel resolution of 14 × 14 × 256 . Similarly, the third group of layers, labeled “DenseBlock_3,” contained 48 dense blocks and processed the feature map of “DenseBlock_2.” It down-sampled the features, and generated a feature map with a pixel resolution of 7 × 7 × 896 . The last group of layers, labeled “DenseBlock_4”, contained 32 dense blocks, processed the feature map of the previous layer, and produced a feature map with a pixel resolution of 7 × 7 × 1920 . Although the architecture contains dense blocks with various filters, the dimensions inside the blocks are equivalent. For compactness of the model and down-sampling of the representations, the transition layer was applied between dense blocks, which comprise the convolution and pooling functions. Finally, the last average pooling layer, named “Average Pooling,” was applied using a filter with a pixel resolution of 7 × 7 , and obtained a spatial feature vector f 2 with a pixel resolution of 1 × 1 × 1920 . The last three layers of DenseNet, named “FCL,” “SoftMax,” and “Classification Output” were removed to enhance the training convergence and extract only features not considering the classification. The feature vector f 2 with 1920 dimensions was concatenated using the 2048-dimension feature vector f 1 of ResNet-50 in an SCN, and the final classification was made based on the output of the SCN, as shown in Figure 4.

3.3.3. Feature Concatenation and Final Classification by SCN

After extracting the feature vectors from each CNN of the first-stage networks, we further ensembled them to obtain a concatenated feature map using the proposed SCN, as shown in Figure 4. The efficiency of the ensemble learning model was substantially improved. The ensemble model allowed the true objective function to be best approximated within the space of the hypothesis, and the overall performance could be improved using various CNN features [54,55]. We propose an SCN that concatenates two sets of features into a longer feature vector. Table 4 presents the architecture of the SCN. The first layer of the SCN, called “Concat,” takes the inputs from two networks of the first stage with different dimensions and concatenates them. In detail, the feature map f 1 with pixel dimensions of 1 × 1 × 2048 by modified ResNet is concatenated with f 2 with pixel dimensions 1 × 1 × 1920 by modified DenseNet. The Concat layer of the SCN provides a feature map f with a pixel size of 1 × 1 × 3968 . It then passes through the FCL. The FCL includes a limited number of neurons, taking data from one vector and returning data from another. In general, considering the jth node of the ith layer, we can obtain the following equation:
z i = l = 1 n i 1 ( w j , l [ i ] a l [ i 1 ] + b j [ i ] )
where in Equation (1), a [ i 1 ] is the output of the previous layer with dimensions ( n H [ i 1 ] × n W [ i 1 ] × n C [ i 1 ] ) and is given as input to the FCL by flattening the tensor to a 1D vector with dimensions of ( n H [ i 1 ] × n W [ i 1 ] × n C [ i 1 ] , 1 ) [56]. The learned parameters at the l t h layers are weights w j , l with n l 1 × n l parameters, and bias b j with n l parameters. In addition, n H ,   n W , and n C represent the height, width, and number of channels, respectively, whereas the final output of the FCL is z i . Subsequently, the SoftMax layer is executed. It computes the results of the FCL using the SoftMax function, which compresses the vector z of arbitrary K real numbers to a normalized vector of K real number probabilities, as a probability distribution ranging between zero and 1 with a probability equivalent to 1 [56]. The SoftMax function is as follows:
f ( z ) i = e z i j K e z j
where in Equation (2), K is the number of output classes, and the output f ( z ) i is the probability for each class. These probabilities are obtained by taking the exponential of each neuron (value) for its class, that is, e z i , and dividing by the sum of all exponentials. The denominator part acts as a normalization term to make the sum of all output values equal to 1. Finally, the classification layer computes the final probabilities to determine the class for the image.

3.4. Classification Configuration

In our DRE-Net-based classification of shoulder implants, we designed two configurations of closed-world and open-world configurations. The detailed explanations are as follows: for the closed-world configuration, data from the same class are used for both training and testing. In detail, we applied a 10-fold cross-validation. Therefore, 90% of the data of each class were used for training, and the remaining 10% of the data of the same class were used for testing. This procedure was iterated 10 times, and the average accuracy of 10 trials was obtained as the final classification accuracy. Because the output classes of training and testing were the same, the final classification was made based on the output of DRE-Net, as shown in Figure 5.
For the open-world configuration, data from the same class are not used for both training and testing, which means that the classes of training and testing data are completely different, as in general content-based image retrieval systems [57]. We conducted a 2-fold cross-validation considering four output classes. Therefore, the data of classes 1 and 2 were used for training, and the remaining data of classes 3 and 4 were used for testing in the first trial. In the second trial, the training and testing data were exchanged with each other, and the same procedure was repeated. The average accuracy of the two trials was obtained as the final accuracy of classification. Because the output classes of training and testing are different, the final classification cannot be made based on the output of DRE-Net, as in the close-world configuration shown in Figure 5. Instead, the feature vector (1 × 3968) of one testing image is extracted from the first layer (the concatenation layer of Figure 4 and Table 4) of the SCN with trained DRE-Net, and the best matching class is determined based on the L2-norm distance (Euclidean distance) between the extracted feature vector and mean vector of the testing classes, as shown in Figure 6. The open-world configuration can reflect the real scenario better than the closed-world configuration, because the data of the untrained class can be obtained in the medical field, as a new manufacturer appears. In this scenario, there is no need to retrain the whole network for all the previous and new classes. Only a reference mean feature vector of the new class (extracted from our network) and its corresponding label (assigned by the medical professional) need to be registered. Then, the model can also work for all the data samples of the new class. In detail, when a new implant model needs to be recognized in a testing phase, the feature vector (1 × 3968) of the image of the new implant model can be extracted from the first layer (the concatenation layer of Figure 4 and Table 4) of the SCN with DRE-Net without additional training. Then, the best matching class can be determined based on the L2-norm distance (Euclidean distance) between the extracted feature vector and the set of reference mean feature vectors.

4. Experimental Setups and Results

4.1. Dataset and Experimental Setups

The dataset used in our research was collected from two different sources comprised of 597 X-ray images of shoulder implant prostheses. This is an open medical dataset that can be used for research purposes. The dataset consists of shoulder prosthesis images of 16 different models from 4 different manufacturers, which were collected from individual manufacturers, surgeons, and the University of Washington [11,45]. One image was captured from each patient in the dataset. The 597 X-ray images of implants are the sum of 83, 294, 71, and 149 of the four manufacturers, Cofield, Depuy, Tornier, and Zimmer, respectively. Figure 7 shows representatives from the dataset, including actual class labels. As shown in Figure 1, the dataset is challenging owing to (1) a high intra-class variance resulting from the various models of the same manufacturer, (2) a small inter-class variance from all X-ray scans of the implants being generally indistinguishable, and (3) a class imbalance. The intra-class variance and class imbalance problems were solved by increasing the dataset size using RIA with sufficient training.
Following the size of the input layer of our model, we resized all images of each class to spatial dimensions with a pixel resolution of 224 × 224 × 3 in a portable network graphics (PNG) file format. For the closed-world configuration, we randomly divided the dataset into 10 folds for a cross-validation, as described in Section 3.4. The number of images for the training dataset is not uniform for all classes, and this imbalance problem of the classes degrades the classification performance [58]. To eliminate this issue, we expanded the size of the training dataset by using RIA, but did not perform this augmentation with the testing dataset. Table 5 shows the detailed explanations of the 10-fold cross-validation of the training and testing datasets for the closed-world configuration. C1, C2, C3, and C4 represent the class Cofield, Depuy, Tornier, and Zimmer. We analyzed the performance of state-of-the-art methods using the same experimental protocols. In addition, state-of-the-art methods were also analyzed with online data augmentation and RIA to optimize the results.
A desktop system with the following specifications was used for all experiments in our work: 3.50 GHz Intel® (Santa Clara, CA, USA) Core™ i7–3770K central processing unit [59] with 16 GB RAM, and an NVIDIA (Santa Clara, CA, USA) GeForce GTX 1070 graphics card [60]. A deep learning toolbox with MATLAB R2019b (MathWorks, Inc., Natick, MA, USA) [61] was used on the Windows 10 operating system to implement our RIA algorithm and DRE-Net.

4.2. Training of CNN Model

For training DRE-Net, the cross-entropy loss function was used as follows [62]:
C E = i K t i l o g ( f ( z ) i )
where in Equation (3), f ( z ) i is the probability for each class, which is defined in Equation (2). Cross entropy is simply the negative log of f ( z ) i for the true label class t i . For the true label class, t i becomes 1, whereas it becomes zero for all other classes.
Prior to training the CNNs, all of the dataset images were resized to 224 × 224 × 3 pixels. We trained different CNNs involving VGG-16, VGG-19, ResNet-18, ResNet-50, NASNet, DenseNet-201, and our deep DRE-Net for comparison. All CNNs were trained using the stochastic gradient descent (SGD) algorithm [63]. SGD is an optimization method that applies a backpropagation algorithm. The main goal of SGD is to find the optimum parameters for the model based on a mini-batch using the derivative of the loss function. SGD updates parameters, such as the weights and biases for each training instance and label. During the training of the CNN, the loss between the actual label and predicted label is calculated, and the SGD updates the parameters based on the loss function. Owing to the problems of class imbalance and the limited size of the dataset, the dataset was augmented using the proposed RIA. Owing to the small dataset, the filter weights of the first-stage networks of the modified DenseNet and ResNet were initialized using the parameters of pre-trained DenseNet-201 and ResNet-50 along with the ImageNet dataset, respectively. Transfer learning with our training data was then conducted using these CNN models. Transfer learning with pre-trained networks is effective for learning richer features from large datasets to a small dataset to achieve high accuracy. The details of the training parameters for the modified DenseNet, ResNet, and DRE-Net are listed in Table 6. The explanations of these parameters are given in [64]. In our research, we compared the accuracies from sequential training, by which modified DenseNet, ResNet, and SCN were separately trained, and the accuracies from end-to-end training, by which DRE-Net including modified DenseNet, ResNet, and SCN were trained at the same time. The training parameters of the two training cases are presented in Table 6.
The graphs of the training losses and the accuracies through both sequential and end-to-end training are visualized according to the number of epochs, as shown in Figure 8. All networks were converged by increasing the accuracy to 100% while decreasing the loss to 0%, which shows that all networks were successfully trained well. However, the convergence time in terms of loss of the end-to-end training was longer than that of the modified DenseNet, ResNet, and DRE-Net when applying sequential training. In our experiments, we selected 25% of the data as a validation subset and the remaining 75% of the data as a training subset from the training data. We provide the validation losses and accuracies of the proposed SCN (Figure 8c) which shows the better testing accuracies than DRE-Net (end-to-end training) (Figure 8d). Even with the model of training accuracies at 100% (Figure 8c), we could obtain the high validation accuracy and low validation loss as shown in Figure 8e, which confirms the optimal convergence of the proposed network without causing overfitting problem with training data.

4.3. Testing and Performance Analysis

We used four qualitative evaluation metrics to assess the performance of our classification network: the accuracy, F1-score, precision, and recall. These metrics are commonly used to evaluate classification frameworks [65] and are calculated as follows:
Accuracy = 1 K k = 1 K TP k + TN k TP k + TN k + FP k + FN k
F1-score = 2 × Precision × Recall Precision + Recall
Precision = 1 K k = 1 K TP k TP k + FP k
Recall = 1 K k = 1 K TP k TP k + TN k
where K represents the total number of classes, which is equivalent to 4 in our study; T P k is the number of true positives of class k, which represents the correctly predicted image from class k ; and F P k represents the number of false positives of class k, which represents the incorrect prediction of another class into class k. In addition, T N k represents the number of true negatives of class k, and is the result in which the other class (except for class k) is correctly predicted by the model. Finally, F N k represents the number of false negatives of class k, which occurs when class k is incorrectly predicted into another class using the model.

4.3.1. Ablation Studies

We studied ablation studies to check the performance and contribution of each component to the overall framework. As the first ablation study, we compared the accuracies of our SCN in Figure 4 with those of the principal component analysis (PCA) + K-NN classifier. A PCA [66] followed by a K-NN [67] was utilized as a post-processing stage to generate the uncorrelated features and scale down the dimensions of the feature vector. The main purpose of applying a PCA is to analyze the discrimination of the selected features (i.e., whether features are distinctive or redundant). From the concatenation layer of a SCN, shown in Figure 4, 1 × 3968 features are projected into the eigenspace to obtain 3968 eigenvectors and eigenvalues of the training samples. As shown in Figure 9, different eigenvectors are selected to evaluate the PCA for computing the eigenvector (λ), which shows the best performance. As shown in Figure 9, the maximum average performance of λ = 10 was found among all eigenvectors with the training data. Then, the PCA features of the testing samples at λ = 10 were calculated and used as an input to the K-NN classifier. Detailed comparative classification results are shown in Table 7. Although the PCA can reduce the number of dimensions from 1 × 3968 to 1 × 10 , the classification performance was not higher than that without the PCA-based classification framework (our SCN), as shown in Table 7. This indicates that the high-dimensional features extracted by our deep DRE-Net are already diversified.
Table 8 shows the second ablation study of the shoulder implant classification. As shown in this table, DenseNet-201 and ResNet-50 without the proposed RIA showed lower accuracies by DenseNet-201 and ResNet-50 with RIA. However, the proposed DRE-Net, including DenseNet-201, ResNet-50, and SCN, showed the highest accuracies. The diversity of individually trained ensembles has been reported to be advantageous [68]. Therefore, we compared the results of DRE-Net using sequential and end-to-end training. The results in Table 8 suggest that ensembles of the models benefit from independent training (sequential training). End-to-end training showed a lower performance than sequential training, and the reason for this is that we used high-capacity models, and the ensemble of these models in end-to-end training shows a “model dominance” effect. Table 8 shows that there is a small difference between the results of DRE-Net (end-to-end) and ResNet-50 + RIA compared to those of DenseNet-201 + RIA. That is because DRE-Net (end-to-end) has “model dominance” effect of ResNet-50 + RIA.
Figure 10a–c present the classification performances of the second-best (DenseNet-201 + RIA) and third-best approaches (ResNet-50 + RIA) and our model (DRE-Net (sequential training)) from Table 8 in terms of a confusion matrix. The diagonal values of each table in Figure 10 show the average recall of each class. As shown in Figure 10, our model outperforms both DenseNet-201 + RIA and ResNet-50 + RIA. The reason why class 4 shows lower accuracies by our model than with the other classes is that the data of class 4 have a higher interclass similarity with those of class 2, as explained in Section 5.

4.3.2. Comparison of Proposed DRE-Net with the Subjective Evaluation

To highlight the significance of the proposed deep learning method, we additionally performed a subjective evaluation experiment considering the same experimental setup (same testing data samples and 10-fold cross validation). The graphical user interface (GUI) of the experimental protocol was designed in MATLAB R2019b (MathWorks, Inc., Natick, MA, USA) [61], as shown in Figure 11. In detail, a total of 10 individuals (without medical training) participated in this subjective evaluation and visually predicted the class label of all testing data samples one by one for each fold. The demographic details of these participants and subjective performance are given in Table 9. Participants (80% male and 20% female) from three different countries, including 50% from South Korea, 40% from Pakistan, and 10% from Iran took part in this subjective evaluation. All information for experiments was given to participants in advance. Each participant could observe both a set of random training samples of each manufacturer of Figure 11a, and one-fold testing images which is the 10% of the data of Figure 11b at the same time. In this way, each testing-fold samples were provided to each person to perform 10-fold cross validation. The group evaluated all of the testing images of each fold, and assigned the appropriate label to each sample of Figure 11b by visually comparing the training set as shown in Figure 11a. The average time calculated for the evaluation of one participant was about twenty minutes. Once all individuals had completed the evaluation, the average performance of each fold was calculated as shown in Table 9. Finally, we obtained the average performance (as confusion matrix, average accuracy, F1-score, precision, and recall) of this subjective evaluation and compared them with the proposed DRE-Net as presented in Figure 12 and Table 10. It can be observed that our proposed DRE-Net shows the superior performance over subjective evaluation with average performance gains of 33.67%, 35.15%, 36.47%, and 33.83% in terms of accuracy, F1-score, precision, and recall, respectively.
In addition, as shown in Figure 12a, the correct classification accuracy by human subject with Cofield data (C1) was 63.86% which was much lower than that by our proposed method of 84.34%. These results confirm that it is visually difficult to discriminate the data of C1 from Figure 1a, and we can tell that there exist the differences among those intra models.

4.3.3. Comparisons of Proposed DRE-Net with the State-of-The-Art Methods

The performances of various state-of-the-art methods [38,41,45,46,69,70] were compared with those of our approach. Table 11 shows the performance comparisons by the state-of-the-art methods and the proposed method without data augmentation, and ResNet-50 [38] outperformed the other methods. In this case, all methods were compared without a data augmentation for a fair comparison. Table 12 shows the performance comparisons by the state-of-the-art methods and the proposed method with data augmentation (through a random in-plane rotation and translation), which shows higher accuracies than those listed in Table 11. The results in most cases show that ResNet-50 [38] and DenseNet-201 [41] outperformed the other methods. In this case, all methods were compared with the data augmentation (random in-plane rotation and translation) for fair comparisons. However, our proposed model does not produce state-of-the-art results with this augmentation technique, as shown in Table 12. This demonstrates that different augmentation techniques have different impacts on the neural networks.
As can be seen in Table 13, when the performances are compared between the state-of-the-art methods and the proposed method with RIA, a 4.18% performance gain was shown in the average accuracy of DenseNet-201 with ResNet-50. In addition, NASNet exhibited a 1.34% performance decrease in terms of the average accuracy with ResNet-50. Among all methods applied, our approach (DRE-Net (sequential training)) outperforms all other state-of-the-art methods. In this case, all methods were compared with RIA for a fair comparison. In addition, we can confirm that the accuracies of Table 13 are higher than those of Table 11 and Table 12 in most cases. For fair comparison, the weights of the CNN models were pre-trained on the ImageNet dataset, and transfer learning was performed again with our training data in all experiments presented in Table 11, Table 12 and Table 13.
We evaluated the deep models using a 10-fold cross-validation and calculated the mean scores. To verify that the difference between mean scores was statistically significant, a t-test [71] was conducted. This test is based on a null hypothesis (H), which states that the performances of our model and the other approaches are not expected to be different (i.e., H = 0). The T-test is carried out to verify the substantial disparity between our model and the second-best [41] and third-best [38] baseline models in Table 13. Our sample size was small and increased the complexity of the statistical analysis. In detail, as the sample size decreases, the chance that every measured mean value is the same as the real total mean value decreases and the degree of uncertainty about the true value of the mean increases. Therefore, we conducted a t-test by combining 10-fold cross-validation values of the accuracy, F1-score, precision, and recall. The null hypothesis is rejected when there is less than a 5% chance of validity. The results in Table 14 show that the p-values calculated by the second- and third-best methods with our model are 0.03 (<0.05) and 7.84 × 10−9 (<0.001%), respectively, which demonstrates the effective distinction between our model and the other approaches. The p-value (0.03) for the second-best model shows that the null hypothesis is rejected at a 97% confidence level and shows a significant difference between our approach and the second-best model. In the case of the third-best model, the p-value (7.84 × 10−9) indicates a significant difference between our approach and the third-best model, and the null hypothesis is rejected at a 99% confidence level.

5. Discussions

In this study, we implemented two spatial feature extraction networks using a densely connected convolution network and a residual neural network. In the first stage, our proposed model envisages the spatially extracted features of both networks, which eventually leads to better results compared to other state-of-the-art classification networks. In the second stage, the proposed SCN further processes the spatial features, and therefore, ideal spatial features are extracted to achieve the best result. The architecture of the modified DenseNet model is shown in Table 3, and shows various dense blocks and transition layers used to exploit the optimal spatial features of the input image and achieve superior outcomes over other CNN models.
In this section, we generate class activation maps to illustrate the performance of the achievements of the modified DenseNet. Figure 13 shows the discriminative image regions used by the modified DenseNet to identify the class. The activation maps calculated for each dense block are represented using a pseudo color scheme [72]. The left column in Figure 13 shows the input images of four classes (C1–C4) given to DenseNet to learn its features, and it can be seen that the activation maps (F1, F2, …, F5) become salient after processing through each dense block. Finally, we can obtain class-specific regions (activation map F5) that provide the specific visual pattern for each class, which ensures that DenseNet learns the features well. Similarly, we generated class activation maps to illustrate the performance of the modified ResNet. The architecture of the modified ResNet model is listed in Table 2, and shows various residual blocks used to exploit the optimal spatial features of the input image. Figure 14 shows the discriminative image regions used by the modified ResNet to identify the class. The left column in Figure 14 shows the input images of four classes (C1–C4) given to ResNet to learn its features and activation maps calculated by each residual block, which are represented by a pseudo color scheme [72]. The activation maps (F1, F2, …, F5) become prominent after processing through the residual blocks. Ultimately, we can obtain class-specific regions (activation map F5) that provide a specific visual pattern for each class. However, as shown in Figure 14, the activation map for class 4 (Zimmer) does not clearly match visually distinct patterns. For a fair comparison between first-stage networks, we used the same input images of different classes to generate activation maps in Figure 13 and Figure 14. The activation maps for class 4 generated by DenseNet and ResNet are quite different. The activation map generated by DenseNet for class 4 is the representation of its visually discriminated region, as shown in the last row of Figure 13, whereas that generated by ResNet for the same class shows a deviation from the discriminated region, as shown in the last row of Figure 14. This indicates that ResNet made predictions not on the head of the implants, which is a discriminated part, but on the background. Therefore, ResNet does not make a decision well for class 4 to learn the features. Moreover, as shown in Figure 10, the confusion matrix of the first-stage networks shows that ResNet has 5.37% less average recall than DenseNet for class 4. In addition, the activation map generated by ResNet-50 for class 3, as shown in the third row of Figure 14, is not focused on the head and is larger than that generated by DenseNet-201, shown in the third row of Figure 13. Therefore, the recall of ResNet-50 is much lower than that of DenseNet-201, as indicated in Figure 10. A similar analysis can be made for class 1. The activation map generated by ResNet-50 for class 1, as shown in the first row of Figure 14, does not accurately exist in the head area compared to that by DenseNet-201, as shown in the first row of Figure 13. Therefore, the recall of ResNet-50 is lower than that of DenseNet-201, as shown in Figure 10.
Finally, the final class activation maps (F5) of the first-stage networks are processed by the proposed SCN for final classification after passing through their respective average pooling layers. A class activation map for the second-stage network cannot be generated. The reason for this is that, in the second stage network, the feature vector is 1 × 1 × 4, and it lacks the visual information. Moreover, the ability of visual object detection by convolution layer was lost when FCL was used for classification in the second stage network. The fundamental difference between the SCN and first-stage networks is the processing of the feature maps. DenseNet and ResNet extract and process the feature maps of an image independently, whereas SCN combines the connectivity of both networks and processes their feature maps. In this way, an optimal representation of the spatial features is generated, which ultimately leads to a better performance in the classification of various types of shoulder prostheses.
We also computed the performance of our proposed network for an open-world configuration. For the open-world configuration, we conducted two-fold experiments by splitting the datasets into two halves, as explained in Section 3.4. The first half was used for training, while the other half was used for testing. Similar to the closed-world configuration, the training dataset in the open-world configuration is augmented using RIA. The main step in the open-world setup is to judge the real class label of the query image by calculating its similarity score with the class mean features. Thus, the Euclidean distance can be used to predict a class label for the query image. Owing to the limited number of classes (i.e., 4), we used two-fold cross-validation. Table 15 displays the details of the two-fold cross-validation of the training and testing datasets for the open-world configuration. Table 16 shows the experimental results of our proposed model, and the second- and third-best models are shown in Table 13 for the open-world configuration. There is a 0.72% performance gain in the average accuracy of our model over the second-best model and 2.4% over the third-best model.
In this section, we also measured the performance of the proposed network in terms of confusion matrices considering open-world setting, as shown in Figure 15. In the 1st fold-A and -B, Tornier (C3), Zimmer (C4) (Figure 15b) and Cofield (C1), Depuy (C2) (Figure 15a) are used in testing, respectively. Similarly, in the 2nd fold-A and -B, Depuy (C2), Zimmer (C4) (Figure 15c) and Cofield (C1), Tornier (C3) (Figure 15d) are used in testing, respectively. As shown in these figures, the average value of correct classification ((84.01 + 51.68)/2(%)) with the testing of C2 and C4 (Figure 15c) is lower than those with the testing of C1 and C2 (Figure 15a) and C1 and C3 (Figure 15d). However, it is higher than that with the testing of C3 and C4 (Figure 15b). These results mean that the similarity between C2 and C4 does not give much effect on testing by open-world configuration compared to that by closed-world configuration. That is because the number of classes in the testing of open-world configuration (two classes) is half of that of closed-world configuration (four classes), which increases the inter-distance between two classes and consequently reduces the effect of similarity of C2 and C4 on testing accuracy of open-world configuration. In the open-world configuration mode, which is more complicated and challenging than the closed-world configuration mode, our model performs the best and is likely applicable to real-world problems as well.
We analyzed the false-positive and false-negative cases of our classifier and found that the reasons for the erroneous classification are structural similarities of the prostheses and the limited size of the dataset. For example, in Figure 10, the confusion matrix of our proposed model shows a lower average recall of class 4 (Zimmer) than that of the other classes. This is because the size (the number of images) of class 4 is two times less than that of class 2 (Depuy) with a high inter-class similarity between them, as can be seen in Figure 16. However, we maintain the sizes of the classes using RIA, although the class imbalance problem remains. It should be considered that the class imbalance problem is still an open issue [73], and thus various solutions are not guaranteed to be optimal. In addition, we analyzed the two-fold experiments for the open-world configuration owing to the limited number of classes. We plan to increase the number of folds in the future by increasing the number of classes. We trained two separate CNNs to extract the features and ensemble them using an SCN. This approach increases the training time owing to the large number of parameters required but makes the model more robust.

6. Conclusions

In this study, we proposed the use of DRE-Net by combining features for shoulder implant classification in X-ray images based on two independent models: modified ResNet-50 and DenseNet-201. This framework automatically detects the prostheses by the manufacturer and aids the surgeons to fit it in the patient’s body by their anatomy as personalized medicine. We analyzed the application of different deep learning models for the classification of shoulder implants by the manufacturer, and compared them with the ensemble of two deep learning models. The ensemble of models using the proposed SCN minimizes the weaknesses of each individually and takes advantage of the strengths of both. To further improve the efficiency of the classification, we proposed the application of RIA and increased the results by 8.87%. We discovered that independent (sequential) training of ensemble models shows better performance than end-to-end training. Although the dataset is relatively small, we obtained the optimum results for shoulder implant classification by integrating transfer learning, ensemble learning, feature concatenation, and RIA. We also examined our model for an open-world configuration and achieved the best results compared to the other deep models, which demonstrates the generalizability of our approach. As reported in previous research [11,45,46], the usage of computer-based algorithms can do better to identify shoulder arthroplasty implants compared to medical experts, which can reduce the risk of delayed operations, perioperative morbidity, and overuse of resources due to lack of correct identification of shoulder arthroplasty implants. Based on these motivations, previous research [11,45,46] has also studied the computer-based algorithms for the identification of shoulder implants. This study is helpful for personalized shoulder arthroscopy and researchers working on X-ray image-based implant recognition.
In the future, we plan to upgrade the results and reduce the training time of the proposed technique by establishing a custom-built model. We also plan to extend this work by adding additional manufacturers and classifying shoulder implants using the models. In addition, the class imbalance problem and increased number of classes for open-world configurations will also researched in the future.

Author Contributions

H.S., M.O. and K.R.P. designed the overall framework. In addition, H.S. and K.R.P. wrote and revised the paper. C.P., T.M. and A.H. helped the comparative analysis and experiments. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Research Foundation of Korea (NRF) funded by the Ministry of Science and ICT (MSIT) through the Basic Science Research Program (NRF-2020R1A2C1006179), in part by the MSIT, Korea, under the ITRC (Information Technology Research Center) support program (IITP-2021-2020-0-01789) supervised by the IITP (Institute for Information & Communications Technology Planning & Evaluation), and in part by the NRF funded by the MSIT through the Basic Science Research Program (NRF-2019R1A2C1083813).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. OrthoInfo, AAOS. Shoulder Joint Replacement. Available online: https://www.orthoinfo.org/en/treatment/shoulder-joint-replacement/ (accessed on 5 February 2021).
  2. Wicha, M.; Tomczyk-Warunek, A.; Jarecki, J.; Dubiel, A. Total Shoulder Arthroplasty, an Overview, Indications and Prosthetic Options. Wiad. Lek. 2020, 73, 1870–1873. [Google Scholar] [CrossRef]
  3. Burns, L.R.; Housman, M.G.; Booth, R.E.J.; Koenig, A. Implant Vendors and Hospitals: Competing Influences over Product Choice by Orthopedic Surgeons. Health Care Manag. Rev. 2009, 34, 2–18. [Google Scholar] [CrossRef]
  4. Mahomed, N.N.; Barrett, J.A.; Katz, J.N.; Phillips, C.B.; Losina, E.; Lew, R.A.; Guadagnoli, E.; Harris, W.H.; Poss, R.; Baron, J.A. Rates and Outcomes of Primary and Revision Total Hip Replacement in the United States Medicare Population. JBJS 2003, 85, 27–32. [Google Scholar] [CrossRef]
  5. Dy, C.J.; Bozic, K.J.; Padgett, D.E.; Pan, T.J.; Marx, R.G.; Lyman, S. Is Changing Hospitals for Revision Total Joint Arthroplasty Associated with More Complications? Clin. Orthop. Relat. Res. 2014, 472, 2006–2015. [Google Scholar] [CrossRef] [Green Version]
  6. Wilson, N.A.; Jehn, M.; York, S.; Davis, C.M. Revision Total Hip and Knee Arthroplasty Implant Identification: Implications for Use of Unique Device Identification 2012 AAHKS Member Survey Results. J. Arthroplast. 2014, 29, 251–255. [Google Scholar] [CrossRef]
  7. Branovacki, G. Ortho Atlas: Hip Arthroplasty U.S. Femoral Implants 1938–2008; Ortho Atlas Publishing: Chicago, IL, USA, 2008. [Google Scholar]
  8. CNN Models for Shoulder Implants Recognition with Algorithms. Available online: http://dm.dgu.edu/link.html (accessed on 6 February 2021).
  9. Bredow, J.; Wenk, B.; Westphal, R.; Wahl, F.; Budde, S.; Eysel, P.; Oppermann, J. Software-Based Matching of X-ray Images and 3D Models of Knee Prostheses. Technol. Health Care 2014, 22, 895–900. [Google Scholar] [CrossRef] [Green Version]
  10. Morais, P.; Queirós, S.; Moreira, A.H.J.; Ferreira, A.; Ferreira, E.; Duque, D.; Rodrigues, N.F.; Vilaça, J.L. Computer-Aided Recognition of Dental Implants in X-ray Images. In Proceedings of the SPIE 9414, Medical Imaging: Computer-Aided Diagnosis, Orlando, FL, USA, 20 March 2015; Volume 9414, p. 94142E. [Google Scholar]
  11. Stark, M.B.C.G. Automatic Detection and Segmentation of Shoulder Implants in X-ray Images. Master Thesis, San Francisco State University, San Francisco, CA, USA, 2018. [Google Scholar]
  12. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; van der Laak, J.A.W.M.; van Ginneken, B.; Sánchez, C.I. A Survey on Deep Learning in Medical Image Analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Koteluk, O.; Wartecki, A.; Mazurek, S.; Kołodziejczak, I.; Mackiewicz, A. How Do Machines Learn? Artificial Intelligence as a New Era in Medicine. J. Pers. Med. 2021, 11, 32. [Google Scholar] [CrossRef]
  14. Owais, M.; Arsalan, M.; Choi, J.; Mahmood, T.; Park, K.R. Artificial Intelligence-Based Classification of Multiple Gastrointestinal Diseases Using Endoscopy Videos for Clinical Diagnosis. J. Clin. Med. 2019, 8, 986. [Google Scholar] [CrossRef] [Green Version]
  15. Owais, M.; Arsalan, M.; Choi, J.; Park, K.R. Effective Diagnosis and Treatment through Content-Based Medical Image Retrieval (CBMIR) by Using Artificial Intelligence. J. Clin. Med. 2019, 8, 462. [Google Scholar] [CrossRef] [Green Version]
  16. Mahmood, T.; Arsalan, M.; Owais, M.; Lee, M.B.; Park, K.R. Artificial Intelligence-Based Mitosis Detection in Breast Cancer Histopathology Images Using Faster R-CNN and Deep CNNs. J. Clin. Med. 2020, 9, 749. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Suh, Y.J.; Jung, J.; Cho, B.-J. Automated Breast Cancer Detection in Digital Mammograms of Various Densities via Deep Learning. J. Clin. Med. 2020, 10, 211. [Google Scholar] [CrossRef]
  18. Arsalan, M.; Baek, N.R.; Owais, M.; Mahmood, T.; Park, K.R. Deep Learning-Based Detection of Pigment Signs for Analysis and Diagnosis of Retinitis Pigmentosa. Sensors 2020, 20, 3454. [Google Scholar] [CrossRef]
  19. Arsalan, M.; Owais, M.; Mahmood, T.; Cho, S.W.; Park, K.R. Aiding the Diagnosis of Diabetic and Hypertensive Retinopathy Using Artificial Intelligence-Based Semantic Segmentation. J. Clin. Med. 2019, 8, 1446. [Google Scholar] [CrossRef] [Green Version]
  20. Arsalan, M.; Kim, D.S.; Owais, M.; Park, K.R. OR-Skip-Net: Outer Residual Skip Network for Skin Segmentation in Non-Ideal Situations. Expert Syst. Appl. 2020, 141, 112922. [Google Scholar] [CrossRef]
  21. Arsalan, M.; Owais, M.; Mahmood, T.; Choi, J.; Park, K.R. Artificial Intelligence-Based Diagnosis of Cardiac and Related Diseases. J. Clin. Med. 2020, 9, 871. [Google Scholar] [CrossRef] [Green Version]
  22. Arsalan, M.; Kim, D.S.; Lee, M.B.; Owais, M.; Park, K.R. FRED-Net: Fully Residual Encoder–Decoder Network for Accurate Iris Segmentation. Expert Syst. Appl. 2019, 122, 217–241. [Google Scholar] [CrossRef]
  23. Olczak, J.; Fahlberg, N.; Maki, A.; Razavian, A.S.; Jilert, A.; Stark, A.; Sköldenberg, O.; Gordon, M. Artificial Intelligence for Analyzing Orthopedic Trauma Radiographs: Deep Learning Algorithms-Are They on Par with Humans for Diagnosing Fractures? Acta Orthop. 2017, 88, 581–586. [Google Scholar] [CrossRef] [Green Version]
  24. Bini, S.A. Artificial Intelligence, Machine Learning, Deep Learning, and Cognitive Computing: What Do These Terms Mean and How Will They Impact Health Care? J. Arthroplast. 2018, 33, 2358–2361. [Google Scholar] [CrossRef]
  25. Tiulpin, A.; Thevenot, J.; Rahtu, E.; Lehenkari, P.; Saarakkala, S. Automatic Knee Osteoarthritis Diagnosis from Plain Radiographs: A Deep Learning-Based Approach. Sci. Rep. 2018, 8, 1727. [Google Scholar] [CrossRef]
  26. Chung, S.W.; Han, S.S.; Lee, J.W.; Oh, K.-S.; Kim, N.R.; Yoon, J.P.; Kim, J.Y.; Moon, S.H.; Kwon, J.; Lee, H.-J.; et al. Automated Detection and Classification of the Proximal Humerus Fracture by Using Deep Learning Algorithm. Acta Orthop. 2018, 89, 468–473. [Google Scholar] [CrossRef] [Green Version]
  27. Lindsey, R.; Daluiski, A.; Chopra, S.; Lachapelle, A.; Mozer, M.; Sicular, S.; Hanel, D.; Gardner, M.; Gupta, A.; Hotchkiss, R.; et al. Deep Neural Network Improves Fracture Detection by Clinicians. Proc. Natl. Acad. Sci. USA 2018, 115, 11591–11596. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Kitamura, G.; Chung, C.Y.; Moore, B.E. Ankle Fracture Detection Utilizing a Convolutional Neural Network Ensemble Implemented with a Small Sample, De Novo Training, and Multiview Incorporation. J. Digit. Imaging 2019, 32, 672–677. [Google Scholar] [CrossRef] [Green Version]
  29. Rayan, J.C.; Reddy, N.; Kan, J.H.; Zhang, W.; Annapragada, A. Binomial Classification of Pediatric Elbow Fractures Using a Deep Learning Multiview Approach Emulating Radiologist Decision Making. Radiol. Artif. Intell. 2019, 1, e180015. [Google Scholar] [CrossRef]
  30. Borjali, A.; Chen, A.F.; Muratoglu, O.K.; Morid, M.A.; Varadarajan, K.M. Detecting Mechanical Loosening of Total Hip Replacement Implant from Plain Radiograph Using Deep Convolutional Neural Network. arXiv 2019, arXiv:1912.00943v1. Available online: https://arxiv.org/abs/1912.00943 (accessed on 7 February 2021).
  31. Krogue, J.D.; Cheng, K.V.; Hwang, K.M.; Toogood, P.; Meinberg, E.G.; Geiger, E.J.; Zaid, M.; McGill, K.C.; Patel, R.; Sohn, J.H.; et al. Automatic Hip Fracture Identification and Functional Subclassification with Deep Learning. Radiol. Artif. Intell. 2020, 2, e190023. [Google Scholar] [CrossRef] [Green Version]
  32. Yi, P.H.; Wei, J.; Kim, T.K.; Sair, H.I.; Hui, F.K.; Hager, G.D.; Fritz, J.; Oni, J.K. Automated Detection & Classification of Knee Arthroplasty Using Deep Learning. Knee 2020, 27, 535–542. [Google Scholar] [CrossRef]
  33. Karnuta, J.M.; Luu, B.C.; Roth, A.L.; Haeberle, H.S.; Chen, A.F.; Iorio, R.; Schaffer, J.L.; Mont, M.A.; Patterson, B.M.; Krebs, V.E.; et al. Artificial Intelligence to Identify Arthroplasty Implants from Radiographs of the Knee. J. Arthroplast. 2021, 36, 935–940. [Google Scholar] [CrossRef]
  34. Sukegawa, S.; Yoshii, K.; Hara, T.; Yamashita, K.; Nakano, K.; Yamamoto, N.; Nagatsuka, H.; Furuki, Y. Deep Neural Networks for Dental Implant System Classification. Biomolecules 2020, 10, 984. [Google Scholar] [CrossRef]
  35. Kim, J.-E.; Nam, N.-E.; Shim, J.-S.; Jung, Y.-H.; Cho, B.-H.; Hwang, J.J. Transfer Learning via Deep Neural Networks for Implant Fixture System Classification Using Periapical Radiographs. J. Clin. Med. 2020, 9, 1117. [Google Scholar] [CrossRef] [Green Version]
  36. Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-Level Accuracy with 50× Fewer Parameters and <0.5 MB Model Size. arXiv 2016, arXiv:1602.07360v4. Available online: https://arxiv.org/abs/1602.07360v4?source=post_page (accessed on 9 February 2021).
  37. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going Deeper with Convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar] [CrossRef] [Green Version]
  38. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  39. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar] [CrossRef] [Green Version]
  40. Borjali, A.; Chen, A.F.; Muratoglu, O.K.; Morid, M.A.; Varadarajan, K.M. Detecting Total Hip Replacement Prosthesis Design on Plain Radiographs Using Deep Convolutional Neural Network. J. Orthop. Res. 2020, 38, 1465–1471. [Google Scholar] [CrossRef] [Green Version]
  41. Huang, G.; Liu, Z.; van der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar] [CrossRef] [Green Version]
  42. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Fei-Fei, L. ImageNet: A Large-scale Hierarchical Image Database. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  43. Pranata, Y.D.; Wang, K.-C.; Wang, J.-C.; Idram, I.; Lai, J.-Y.; Liu, J.-W.; Hsieh, I.-H. Deep Learning and SURF for Automated Classification and Detection of Calcaneus Fractures in CT Images. Comput. Meth. Programs Biomed. 2019, 171, 27–37. [Google Scholar] [CrossRef]
  44. Tanzi, L.; Vezzetti, E.; Moreno, R.; Aprato, A.; Audisio, A.; Massè, A. Hierarchical Fracture Classification of Proximal Femur X-ray Images Using a Multistage Deep Learning Approach. Eur. J. Radiol. 2020, 133, 109373. [Google Scholar] [CrossRef]
  45. Urban, G.; Porhemmat, S.; Stark, M.; Feeley, B.; Okada, K.; Baldi, P. Classifying Shoulder Implants in X-ray Images Using Deep Learning. Comp. Struct. Biotechnol. J. 2020, 18, 967–972. [Google Scholar] [CrossRef] [PubMed]
  46. Yi, P.H.; Kim, T.K.; Wei, J.; Li, X.; Hager, G.D.; Sair, H.I.; Fritz, J. Automated Detection and Classification of Shoulder Arthroplasty Models Using Deep Learning. Skelet. Radiol. 2020, 49, 1623–1632. [Google Scholar] [CrossRef] [PubMed]
  47. Shorten, C.; Khoshgoftaar, T.M. A Survey on Image Data Augmentation for Deep Learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
  48. Polikar, R. Ensemble Based Systems in Decision Making. IEEE Circuits Syst. Mag. 2006, 6, 21–45. [Google Scholar] [CrossRef]
  49. Rokach, L. Ensemble-Based Classifiers. Artif. Intell. Rev. 2010, 33, 1–39. [Google Scholar] [CrossRef]
  50. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Proceedings of the 25th International Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–8 December 2012; pp. 1097–1105. [Google Scholar] [CrossRef]
  51. ImageNet. Available online: http://www.image-net.org/ (accessed on 7 February 2021).
  52. Glorot, X.; Bordes, A.; Bengio, Y. Deep Sparse Rectifier Neural Networks. In Proceedings of the 14th International Conference on Artificial Intelligence and Statistics, Fort Lauderdale, FL, USA, 11–13 April 2011; pp. 315–323. [Google Scholar]
  53. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. In Proceedings of the International Conference on Machine Learning, Lille, France, 7–9 July 2015; pp. 448–456. [Google Scholar]
  54. Zheng, L.; Zhao, Y.; Wang, S.; Wang, J.; Tian, Q. Good Practice in CNN Feature Transfer. arXiv 2016, arXiv:1604.00133v1. Available online: https://arxiv.org/abs/1604.00133 (accessed on 9 February 2021).
  55. Kawahara, J.; Hamarneh, G. Multi-Resolution-Tract CNN with Hybrid Pretrained and Skin-Lesion Trained Layers. In Proceedings of the International Workshop on Machine Learning in Medical Imaging, Athens, Greece, 17 October 2016; pp. 164–171. [Google Scholar]
  56. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016; Volume 1. [Google Scholar]
  57. Heesch, D. A Survey of Browsing Models for Content Based Image Retrieval. Multimed. Tools Appl. 2008, 40, 261–284. [Google Scholar] [CrossRef]
  58. Buda, M.; Maki, A.; Mazurowski, M.A. A Systematic Study of the Class Imbalance Problem in Convolutional Neural Networks. Neural Netw. 2018, 106, 249–259. [Google Scholar] [CrossRef] [Green Version]
  59. Intel® Core™ i7-3770K Processor. Available online: https://ark.intel.com/content/www/us/en/ark/products/65523/intel-core-i7-3770k-processor-8m-cache-up-to-3-90-ghz.html (accessed on 1 December 2020).
  60. GeForce GTX 1070. Available online: https://www.geforce.com/hardware/desktop-gpus/geforce-gtx-1070/specifications (accessed on 1 December 2020).
  61. Deep Learning Toolbox. Available online: https://www.mathworks.com/products/deep-learning.html (accessed on 1 December 2020).
  62. Heaton, J. Artificial Intelligence for Humans, Volume 3—Deep Learning and Neural Networks; Heaton Research, Inc.: St. Louis, MO, USA, 2013; ISBN 978-1-5057-1434-0. [Google Scholar]
  63. Ruder, S. An Overview of Gradient Descent Optimization Algorithms. arXiv 2017, arXiv:1609.04747v2. Available online: https://arxiv.org/abs/1609.04747 (accessed on 9 February 2021).
  64. Options for Training Deep Learning Neural Network. Available online: https://www.mathworks.com/help/deeplearning/ref/trainingoptions.html (accessed on 13 February 2021).
  65. Hossin, M. A Review on Evaluation Metrics for Data Classification Evaluations. Int. J. Data Min. Knowl. Manag. Process 2015, 5, 1–11. [Google Scholar] [CrossRef]
  66. Abdi, H.; Williams, L.J. Principal Component Analysis. WIREs Comput. Stat. 2010, 2, 433–459. [Google Scholar] [CrossRef]
  67. Cover, T.; Hart, P. Nearest Neighbor Pattern Classification. IEEE Trans. Inf. Theory 1967, 13, 21–27. [Google Scholar] [CrossRef]
  68. Kuncheva, L.I.; Whitaker, C.J. Measures of Diversity in Classifier Ensembles and Their Relationship with the Ensemble Accuracy. Mach. Learn. 2003, 51, 181–207. [Google Scholar] [CrossRef]
  69. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2015, arXiv:1409.1556v6. Available online: https://arxiv.org/abs/1409.1556 (accessed on 11 February 2021).
  70. Zoph, B.; Vasudevan, V.; Shlens, J.; Le, Q.V. Learning Transferable Architectures for Scalable Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8697–8710. [Google Scholar] [CrossRef] [Green Version]
  71. Livingston, E.H. Who Was Student and Why Do We Care so Much about His T-Test? J. Surg. Res. 2004, 118, 58–65. [Google Scholar] [CrossRef]
  72. Zhou, B.; Khosla, A.; Lapedriza, A.; Oliva, A.; Torralba, A. Learning Deep Features for Discriminative Localization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2921–2929. [Google Scholar]
  73. Krawczyk, B. Learning from Imbalanced Data: Open Challenges and Future Directions. Prog. Artif. Intell. 2016, 5, 221–232. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Examples showing high intra-class variabilities and low inter-class variabilities. Examples showing (a) high intra-class variability of one manufacturer (Cofield) and (b) low inter-class variability. In (b), upper-left, upper-right, lower-left, and lower-right images show the cases of four manufacturers of Cofield, Depuy, Tornier, and Zimmer, respectively.
Figure 1. Examples showing high intra-class variabilities and low inter-class variabilities. Examples showing (a) high intra-class variability of one manufacturer (Cofield) and (b) low inter-class variability. In (b), upper-left, upper-right, lower-left, and lower-right images show the cases of four manufacturers of Cofield, Depuy, Tornier, and Zimmer, respectively.
Jpm 11 00482 g001
Figure 2. Overall procedure of the proposed method.
Figure 2. Overall procedure of the proposed method.
Jpm 11 00482 g002
Figure 3. Examples of rotational invariant augmentation (RIA).
Figure 3. Examples of rotational invariant augmentation (RIA).
Jpm 11 00482 g003
Figure 4. Diagram of our proposed DRE-Net for feature extraction and classification.
Figure 4. Diagram of our proposed DRE-Net for feature extraction and classification.
Jpm 11 00482 g004
Figure 5. Diagram of closed-world configuration for classification.
Figure 5. Diagram of closed-world configuration for classification.
Jpm 11 00482 g005
Figure 6. Diagram of open-world configuration for feature extraction and classification.
Figure 6. Diagram of open-world configuration for feature extraction and classification.
Jpm 11 00482 g006
Figure 7. Examples of the dataset: shoulder implants of four different manufacturers: (a) Cofield, (b) Depuy, (c) Tornier, and (d) Zimmer.
Figure 7. Examples of the dataset: shoulder implants of four different manufacturers: (a) Cofield, (b) Depuy, (c) Tornier, and (d) Zimmer.
Jpm 11 00482 g007
Figure 8. Plots for training losses and training accuracies: sequential training of (a) modified DenseNet-201, (b) modified ResNet-50, (c) SCN, (d) DRE-Net (end-to-end training), and (e) plots for validation losses and validation accuracies of SCN of (c).
Figure 8. Plots for training losses and training accuracies: sequential training of (a) modified DenseNet-201, (b) modified ResNet-50, (c) SCN, (d) DRE-Net (end-to-end training), and (e) plots for validation losses and validation accuracies of SCN of (c).
Jpm 11 00482 g008
Figure 9. PCA-based performance for different numbers of eigenvectors (λ = 1, 2, 3,…,100).
Figure 9. PCA-based performance for different numbers of eigenvectors (λ = 1, 2, 3,…,100).
Jpm 11 00482 g009
Figure 10. Confusion matrix of (a) DenseNet-201 + RIA, (b) ResNet-50 + RIA, and (c) DRE-Net (sequential training). C1–C4 indicate the classes of four manufacturers of Cofield, Depuy, Tornier, and Zimmer, respectively (unit: %).
Figure 10. Confusion matrix of (a) DenseNet-201 + RIA, (b) ResNet-50 + RIA, and (c) DRE-Net (sequential training). C1–C4 indicate the classes of four manufacturers of Cofield, Depuy, Tornier, and Zimmer, respectively (unit: %).
Jpm 11 00482 g010
Figure 11. Graphical user interface used for subjective evaluation (a) random training samples of each class, which are shown to user during subjective evaluation, (b) interface showing all the testing data samples to user one by one for subjective class prediction.
Figure 11. Graphical user interface used for subjective evaluation (a) random training samples of each class, which are shown to user during subjective evaluation, (b) interface showing all the testing data samples to user one by one for subjective class prediction.
Jpm 11 00482 g011
Figure 12. Performance comparison in terms of confusion matrices of (a) subjective method and (b) the proposed DRE-Net (sequential training). C1–C4 indicate the classes of four manufacturers of Cofield, Depuy, Tornier, and Zimmer, respectively (unit: %).
Figure 12. Performance comparison in terms of confusion matrices of (a) subjective method and (b) the proposed DRE-Net (sequential training). C1–C4 indicate the classes of four manufacturers of Cofield, Depuy, Tornier, and Zimmer, respectively (unit: %).
Jpm 11 00482 g012
Figure 13. Class activation maps for given inputs of four classes (C1–C4), which are extracted from modified DenseNet-201 of Table 3. C1–C4 indicate the classes of four manufacturers, Cofield, Depuy, Tornier, and Zimmer, respectively.
Figure 13. Class activation maps for given inputs of four classes (C1–C4), which are extracted from modified DenseNet-201 of Table 3. C1–C4 indicate the classes of four manufacturers, Cofield, Depuy, Tornier, and Zimmer, respectively.
Jpm 11 00482 g013
Figure 14. Class activation maps for given inputs of four classes (C1–C4), which are extracted from modified ResNet-50 of Table 2. C1–C4 indicate the classes of the four manufacturers, Cofield, Depuy, Tornier, and Zimmer, respectively.
Figure 14. Class activation maps for given inputs of four classes (C1–C4), which are extracted from modified ResNet-50 of Table 2. C1–C4 indicate the classes of the four manufacturers, Cofield, Depuy, Tornier, and Zimmer, respectively.
Jpm 11 00482 g014
Figure 15. Performance of the proposed network considering open-world setting as confusion matrices (a) 1st fold-A (using C1 and C2 in testing), (b) 1st fold-B (using C3 and C4 in testing), (c) 2nd fold-A (using C2 and C4 in testing), and (d) 2nd fold-B (using C1 and C3 in testing).
Figure 15. Performance of the proposed network considering open-world setting as confusion matrices (a) 1st fold-A (using C1 and C2 in testing), (b) 1st fold-B (using C3 and C4 in testing), (c) 2nd fold-A (using C2 and C4 in testing), and (d) 2nd fold-B (using C1 and C3 in testing).
Jpm 11 00482 g015aJpm 11 00482 g015b
Figure 16. The high inter-class similarity between the two classes: (a) class 2 (Depuy) and (b) class 4 (Zimmer).
Figure 16. The high inter-class similarity between the two classes: (a) class 2 (Depuy) and (b) class 4 (Zimmer).
Jpm 11 00482 g016
Table 1. Comparisons between our proposed and previous methods for implant recognition in X-ray images.
Table 1. Comparisons between our proposed and previous methods for implant recognition in X-ray images.
CategoryTypeMethods# ClassesResultsStrengthWeakness
Handcrafted feature-basedKneeTemplate matching [9]170% to 90% accuracies- Uses a simple image processing technique including Sobel operator, binarization, and template matching
- Computationally efficient
Requires 3D CAD models for template generation of implants
DentalActive contours +
K-nearest neighborhood (K-NN) [10]
1191% of the known implants are recognized- Optimal initial location of the contour can be selected by their method
- Uses simple machine learning algorithm for classification
- K-NN classifier is time-consuming for large numbers of features
- Because of the large number of dental implant models, their approach returns a set of possible candidate results for identifying new implants and needs a user interaction to verify the candidate result
ShoulderHough transform + histogram equalization + mean shift filter [11]477% precision, and 64% F-measure- Uses conventional image processing schemes involving bilateral filter, mean shift filter, and a median blur filter
- Develops a pre-processing tool for training a classifier
Segmentation performance is dependent on the growing approach of seed region
Deep feature-basedKneePre-trained CNN [32]2100% sensitivity, and 100% specificity- High classification performance
- Precisely determines the presence of total knee arthroplasty (TKA)
- Accurately classifies the TKA and unicompartmental knee arthroplasty (UKA)
Classification is performed in a binary fashion (the presence of implant)
Pre-trained CNN [33]999% accuracy, 95% sensitivity, and 99% specificityHigh classification performancePre-processing is needed
and computationally expensive
DentalPre-trained CNN [34]1193.5% accuracy, 91.6% F-measuresHigh average classification accuracy with a small dataset of panoramasVGG network can be replaced with the state-of-the-art networks
Pre-trained CNN [35]496% to 97% accuraciesHigh classification performance and computationally efficientTheir method is unable to detect several implants simultaneously
HipPre-trained CNN [40]3100% accuracyHigh classification performance- Requires high processing power for extensive training
- Uses only one post-surgery anteroposterior (AP) X-ray per patient
ShoulderPre-trained CNN [46]295% sensitivity, and 90% specificity to classify TSA and RTSA- High accuracy to detect the existence of shoulder arthroplasty
- High sensitivity to classify TSA and RTSA
Classification is performed in a binary fashion (the presence of implant)
Pre-trained CNN [45]480.4% accuracy, 80% precision, 75% recall, and 76% F1-score- First deep learning based-approach to classify the manufacturers of shoulder implants
- Higher classification accuracy than non-deep learning-based methods
- Accuracies are needed to be enhanced
- Performance was measured only by closed-world configuration
DRE-Net (Proposed)485.92% accuracy, 84.69% F1-score, 85.33% precision, and 84.11% recall - High classification accuracy
- Applicable to real-world problems by considering both closed-world and open-world configurations
Requires more training time
Table 2. Layer configuration details of modified ResNet-50.
Table 2. Layer configuration details of modified ResNet-50.
Layers NameOutput Feature Map SizeKernel SizeNumber of Iterations
Image Input 224 × 224 × 3 --
Conv 1 112 × 112 × 64 7 × 7   c o n v 1
Max Pooling 56 × 56 × 64 3 × 3   m a x   p o o l 1
Conv 2_x 56 × 56 × 256 1 × 1   c o n v
3 × 3   c o n v
1 × 1   c o n v
3
Conv 3_x 28 × 28 × 512 1 × 1   c o n v
3 × 3   c o n v
1 × 1   c o n v
4
Conv 4_x 14 × 14 × 1024 1 × 1   c o n v
3 × 3   c o n v
1 × 1   c o n v
6
Conv 5_x 7 × 7 × 2048 1 × 1   c o n v
3 × 3   c o n v
1 × 1   c o n v
3
Average Pooling 1 × 1 × 2048 7 × 7   a v g   p o o l 1
Table 3. Layer configuration details of modified DenseNet-201.
Table 3. Layer configuration details of modified DenseNet-201.
Layer NameOutput Feature Map SizeKernel SizeNumber of Iterations
Image Input 224 × 224 × 3 --
Conv 1 112 × 112 × 64 7 × 7   c o n v 1
Max Pooling 56 × 56 × 64 3 × 3   m a x   p o o l 1
DenseBlock_1 56 × 56 × 256 1 × 1   c o n v 3 × 3   c o n v 6
Transition Layer 28 × 28 × 128 1 × 1   c o n v   2 × 2     a v g   p o o l 1
DenseBlock_2 28 × 28 × 512 1 × 1   c o n v 3 × 3   c o n v 12
Transition Layer 14 × 14 × 256 1 × 1   c o n v 2 × 2   a v g   p o o l 1
DenseBlock_3 14 × 14 × 1792 1 × 1   c o n v 3 × 3   c o n v 48
Transition Layer 7 × 7 × 896 1 × 1   c o n v 2 × 2   a v g   p o o l 1
DenseBlock_4 7 × 7 × 1920 1 × 1   c o n v 3 × 3   c o n v 32
Average Pooling 1 × 1 × 1920 7 × 7   a v g   p o o l 1
Table 4. Layer configuration details of SCN.
Table 4. Layer configuration details of SCN.
Layers NameOutput Feature Map SizeKernel SizeNumber of Iterations
Concat 1 × 1 × 3968 -1
Fully Connected 1 × 1 × 4 -1
SoftMax 1 × 1 × 4 -1
Classification 4 -1
Table 5. Summary of 10-fold cross-validation of training and testing data for closed-world configuration (unit: images).
Table 5. Summary of 10-fold cross-validation of training and testing data for closed-world configuration (unit: images).
ValidationTrainingTestingTotal
OriginalAugmentedC1C2C3C4
1st fold53819,36882971519,965
2nd fold53619,29693071519,893
3rd fold53819,36882971519,965
4th fold53719,33283071519,929
5th fold53619,29692981519,893
6th fold53919,40482971420,001
7th fold53719,33283071519,929
8th fold53819,36882971519,965
9th fold53619,29693071519,893
10th fold53819,36882971519,965
Table 6. Parameters for network training.
Table 6. Parameters for network training.
MethodsNumber of EpochsMini-Batch SizeLearning RateMomentum TermL2-RegularizationLearning Rate Drop Factor
Sequential trainingModified
DenseNet-201
13100.0010.90.00010.1
Modified
ResNet-50
13100.0010.90.00010.1
SCN9100.0010.90.00010.1
End-to-end trainingDRE-Net7100.0010.90.00010.1
Table 7. Performance comparisons of our proposed SCN using a PCA and a K-NN (unit: %).
Table 7. Performance comparisons of our proposed SCN using a PCA and a K-NN (unit: %).
FoldPerformance without a PCA (our SCN)Performance with PCA (λ = 10) + K-NN
AccuracyF1-ScoreRecallPrecisionAccuracyF1-ScoreRecallPrecision
10-Fold
Average
85.9284.6984.1185.3357.9448.0440.6060.17
Table 8. Performance comparisons of each sub-network and proposed DRE-Net by end-to-end or sequential training (unit: %).
Table 8. Performance comparisons of each sub-network and proposed DRE-Net by end-to-end or sequential training (unit: %).
MethodsAccuracyF1-ScorePrecisionRecall
ResNet-50 [38]66.7062.0264.6759.83
DenseNet-201 [41]55.7647.5549.7345.73
ResNet-50 + RIA80.5778.0279.2176.95
DenseNet-201 + RIA84.7583.7685.2182.42
DRE-Net (end-to-end)81.5579.1280.7777.66
DRE-Net (sequential)85.9284.6985.3384.11
Table 9. Demographic details of different subjects and their subjective evaluation results.
Table 9. Demographic details of different subjects and their subjective evaluation results.
Demographic DetailsSubjective Performance (%)
Participant IndexAgeNationalitySexAccuracyF1-ScorePrecisionRecall
128PakistanMale57.6353.4053.5153.29
228PakistanMale55.7448.8649.2348.49
323South KoreaMale50.8555.3555.5155.19
432PakistanMale48.3348.4346.9050.06
527South KoreaMale50.8245.5845.5145.66
629South KoreaMale55.1745.6745.1346.23
742IranFemale58.3354.8752.9256.96
827South KoreaFemale45.7642.8341.8443.88
932PakistanMale52.4646.7746.6346.90
1028South KoreaMale47.4653.6851.4756.09
Table 10. Average performance comparison (10-folds) between subjective evaluation and the proposed DRE-Net (unit: %).
Table 10. Average performance comparison (10-folds) between subjective evaluation and the proposed DRE-Net (unit: %).
MethodsAccuracyF1-ScorePrecisionRecall
Subjective Method52.2549.5448.8650.28
DRE-Net (sequential)85.9284.6985.3384.11
Table 11. Performance comparisons of state-of-the-art methods and the proposed approach without data augmentation. Averages from a 10-fold cross-validation are shown (unit: %).
Table 11. Performance comparisons of state-of-the-art methods and the proposed approach without data augmentation. Averages from a 10-fold cross-validation are shown (unit: %).
MethodsAccuracy F1-ScorePrecisionRecall
VGG-16 [45,69]58.70455445
VGG-19 [45,69]63.60546153
ResNet-18 [38,46]66.1360.8664.2558.13
ResNet-50 [38]66.7062.0264.6759.83
NASNet [45,70]64.50546252
DenseNet-201 [41]55.7647.5549.7345.73
Proposed58.1050.8251.7849.96
Table 12. Performance comparisons of the state-of-the-art methods and proposed approach with data augmentation by random in-plane rotation and translation. Averages from a 10-fold cross-validation are shown (unit: %).
Table 12. Performance comparisons of the state-of-the-art methods and proposed approach with data augmentation by random in-plane rotation and translation. Averages from a 10-fold cross-validation are shown (unit: %).
MethodsAccuracyF1-ScorePrecisionRecall
VGG-16 [45,69]74697268
VGG-19 [45,69]76.20707569
ResNet-18 [38,46]70.8265.9368.0264.38
ResNet-50 [38]80.5677.6679.4976.02
NASNet [45,70]80.40768075
DenseNet-201 [41]80.5777.6079.0576.32
Proposed77.0574.8076.9373.07
Table 13. Performance comparisons of the state-of-the-art methods and proposed approach with RIA. Averages from a 10-fold cross-validation are shown (unit: %).
Table 13. Performance comparisons of the state-of-the-art methods and proposed approach with RIA. Averages from a 10-fold cross-validation are shown (unit: %).
MethodsAccuracyF1-ScorePrecisionRecall
VGG-16 [45,69] 68.8566.9066.8267.22
VGG-19 [45,69]66.5463.5463.8163.35
ResNet-18 [38,46]77.4174.6776.6073.05
ResNet-50 [38]80.5778.0279.2176.95
NASNet [45,70]79.2376.2877.2575.44
DenseNet-201 [41]84.7583.7685.2182.42
Proposed85.9284.6985.3384.11
Table 14. The t-test analysis results between our model and the second-best and third-best models.
Table 14. The t-test analysis results between our model and the second-best and third-best models.
Comparisonsp-ValueConfidence Level
ProposedSecond-best0.0397%
ProposedThird-best7.84 × 10−999%
Table 15. Summary of two-fold cross-validation of training and testing data for open-world configuration (unit: images).
Table 15. Summary of two-fold cross-validation of training and testing data for open-world configuration (unit: images).
ValidationTrainingTesting
ClassesOriginalAugmentedClassesOriginalTotal
1st fold-ACofield,
Depuy
37713,572Tornier,
Zimmer
22014,169
1st fold-BTornier,
Zimmer
2207920Cofield,
Depuy
3778517
2nd fold-ATornier,
Cofield
1545544Zimmer,
Depuy
4433585
2nd fold-BZimmer,
Depuy
44315,948Tornier,
Cofield
15416,545
Table 16. Comparison of our proposed model with the second- and third-best models of Table 13 for open-world configuration (unit: %).
Table 16. Comparison of our proposed model with the second- and third-best models of Table 13 for open-world configuration (unit: %).
CNN ModelAccuracyF1-ScorePrecisionRecall
ResNet-50 [38]74.9667.1467.7866.51
DenseNet-201 [41]76.6471.3170.6472.05
Proposed77.3670.8571.2270.49
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Sultan, H.; Owais, M.; Park, C.; Mahmood, T.; Haider, A.; Park, K.R. Artificial Intelligence-Based Recognition of Different Types of Shoulder Implants in X-ray Scans Based on Dense Residual Ensemble-Network for Personalized Medicine. J. Pers. Med. 2021, 11, 482. https://doi.org/10.3390/jpm11060482

AMA Style

Sultan H, Owais M, Park C, Mahmood T, Haider A, Park KR. Artificial Intelligence-Based Recognition of Different Types of Shoulder Implants in X-ray Scans Based on Dense Residual Ensemble-Network for Personalized Medicine. Journal of Personalized Medicine. 2021; 11(6):482. https://doi.org/10.3390/jpm11060482

Chicago/Turabian Style

Sultan, Haseeb, Muhammad Owais, Chanhum Park, Tahir Mahmood, Adnan Haider, and Kang Ryoung Park. 2021. "Artificial Intelligence-Based Recognition of Different Types of Shoulder Implants in X-ray Scans Based on Dense Residual Ensemble-Network for Personalized Medicine" Journal of Personalized Medicine 11, no. 6: 482. https://doi.org/10.3390/jpm11060482

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop