Next Article in Journal
Phytochemistry, Pharmacology and Mode of Action of the Anti-Bacterial Artemisia Plants
Next Article in Special Issue
Surgical Phase Recognition in Inguinal Hernia Repair—AI-Based Confirmatory Baseline and Exploration of Competitive Models
Previous Article in Journal
Gender Differences in Lower Extremity Stiffness during a Single-Leg Landing Motion in Badminton
Previous Article in Special Issue
Surgical Gesture Recognition in Laparoscopic Tasks Based on the Transformer Network and Self-Supervised Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine Learning for Detecting Total Knee Arthroplasty Implant Loosening on Plain Radiographs

Department of Orthopaedic Surgery, Seoul St. Mary’s Hospital, College of Medicine, The Catholic University of Korea, 222, Banpo-daero, Seocho-gu, Seoul 06591, Republic of Korea
*
Author to whom correspondence should be addressed.
Bioengineering 2023, 10(6), 632; https://doi.org/10.3390/bioengineering10060632
Submission received: 4 May 2023 / Revised: 15 May 2023 / Accepted: 22 May 2023 / Published: 23 May 2023
(This article belongs to the Special Issue Artificial Intelligence in Surgery)

Abstract

:
(1) Background: The purpose of this study was to investigate whether the loosening of total knee arthroplasty (TKA) implants could be detected accurately on plain radiographs using a deep convolution neural network (CNN). (2) Methods: We analyzed data for 100 patients who underwent revision TKA due to prosthetic loosening at a single institution from 2012 to 2020. We extracted 100 patients who underwent primary TKA without loosening through a propensity score, matching for age, gender, body mass index, operation side, and American Society of Anesthesiologists class. Transfer learning was used to prepare a detection model using a pre-trained Visual Geometry Group (VGG) 19. For transfer learning, two methods were used. First, the fully connected layer was removed, and a new fully connected layer was added to construct a new model. The convolutional layer was frozen without training, and only the fully connected layer was trained (transfer learning model 1). Second, a new model was constructed by adding a fully connected layer and varying the range of freezing for the convolutional layer (transfer learning model 2). (3) Results: The transfer learning model 1 gradually increased in accuracy and ultimately reached 87.5%. After processing through the confusion matrix, the sensitivity was 90% and the specificity was 100%. Transfer learning model 2, which was trained on the convolutional layer, gradually increased in accuracy and ultimately reached 97.5%, which represented a better improvement than for model 1. Processing through the confusion matrix affirmed that the sensitivity was 100% and the specificity was 97.5%. (4) Conclusions: The CNN algorithm, through transfer learning, shows high accuracy for detecting the loosening of TKA implants on plain radiographs.

1. Introduction

Total knee arthroplasty (TKA) is a validated and effective surgical treatment for the functional recovery of and significant improvements in the quality of life of patients with end-stage knee osteoarthritis (OA) [1,2,3,4,5,6,7,8]. TKA is performed in over 1 million cases in the United States annually and is expected to continue to increase in frequency [9]. TKA has been shown to have a 15-year survival rate of 95% or greater due to the continuous development of technology [10]. However, due to increases in the frequency of arthroplasty and in life expectancy, the revision burden is also expected to continue to increase [9,11,12].
TKA failure may occur due to various causes, the most common being periprosthetic loosening [13,14]. Precise detection of prosthetic loosening of TKA implants is challenging even for experienced clinicians and may not always be feasible, potentially delaying diagnosis. Multiple imaging tools are used for diagnosis, including plain radiographs, scintigraphy, arthrograms, fluorodeoxyglucose-positron emission tomography (FDG-PET), and magnetic resonance imaging (MRI) [15,16]. However, except for plain radiographs, the available imaging tools are invasive and expensive and do not actually confer significant benefits in terms of cost effectiveness compared to plain radiographs [17]. The concordance rate also varies as readings are often performed by clinicians who are not experts [18,19].
Due to recent developments, machine learning has been widely used in orthopedic surgery. Especially in the field of artificial joints, machine learning is being studied a lot. For patients, it is used as a decision aid for diagnosis, the severity and complexity of OA, and the appropriate indication of artificial joint surgery, and machine learning is also used for implant size, positioning or alignment, and ligament balancing. In particular, machine learning is showing excellent performance in diagnosis through images [20,21]. Machine learning techniques are used to diagnose lung disease and breast cancer on radiographs and to support diagnoses in hospitals [22,23]. Machine learning can be used to detect prosthetic loosening after arthroplasty in orthopedic surgery, but the studies are still lacking [24,25,26,27,28]. Therefore, the purpose of this study was to investigate whether the prosthetic loosening of TKA implants could be detected precisely on plain radiographs using a deep convolution neural network (CNN).

2. Materials and Methods

We included patients who underwent revision TKA with prosthetic loosening at a single institution from 2012 to 2020. After obtaining Institutional Review Board (IRB) approval, we collected image data for patients diagnosed with prosthetic loosening after revision TKA during the study period. Preoperative radiographs of arthroplasty surgery for each patient were also collected. In revision TKA, anteroposterior (AP) radiographs were collected at the original resolution.
Images of a total of 100 revision TKA patients were collected and categorized as “loosened”. Cases classified as loosened were defined only when revision TKA was performed because TKA implant loosening was demonstrated in the intraoperative field as well as preoperatively. In order to collect information on well-fixed TKA for comparison, we conducted a survey of primary TKA patients during this period and performed propensity score matching (one-to-one) to minimize selection bias. One hundred patients who underwent primary TKA without loosening were extracted through propensity score matching for age, gender, body mass index, operation side, and American Society of Anesthesiologists (ASA) class. These patients were labeled as “fixed”.
In order to better visualize loosening around the knee implant, we obtained the original X-ray image of each TKA including only the knee implant periphery. The image was separated using the cropping technique through numpy in Python by focusing the image around the implant–bone interface, which was recognized as the most important position, to better confirm loosening around the implant. We compared AP knee implant X-ray images of loosened and fixed groups after resizing the images to 224 × 224 pixels. Since the size of the original image was different for each image, it was modified and there was no significant change from the original image. The image data set was too small to train the deep learning model, so we used an augmentation technique to balance the training image classes and increase the size of the training set to avoid overfitting. Rotation, width shift, height shift, zoom, and flip techniques were used. Rotation was allowed up to 360 degrees, and the width, height shift, and zoom range were set to 0.2 [29,30]. Transfer learning was used to create a model that detects TKA loosening. Transfer learning is a strategy for transferring the knowledge extracted by a neural network from specific data to solve a problem and applying it to a new task, including new and usually insufficient data to train neural networks from the outset [31,32].
In this study, we used the CNN model VGG 19 net, which is widely used in image classification algorithms because it has an easy-to-use structure and good performance [32]. VGG19, proposed by Simonyan and Zisserman [32], consists of 16 convolutional layers and 3 fully connected 19 layers to classify images into 1000 object categories. The model was trained using approximately 1.2 million images, including 100,000 images for testing and 50,000 images for validation. This is a very popular method for image classification because it uses multiple 3 × 3 filters for each convolutional layer. The architecture of VGG19 includes 16 convolutional layers for feature extraction and 3 layers for classification. The layers used for feature extraction are divided into 5 groups, with max pooling layers following each set. A 224 × 224 pixel image is input to this model, and the model outputs the labels of the objects in the image [32] (Figure 1).
We performed transfer learning with fine-tuning to reuse the model and improve accuracy. We implemented the VGG 19 CNN algorithm by replacing the fully connected layer (classifier) with 7 layers as follows: global average pooling, batch normalization, dropout (0.5), fully connected neural network (dense layer) with 256, batch normalization, dropout (0.5) layer to further reduce overfitting, and 1 output neuron for binary classification. For transfer learning models, cross-entropy was used as the loss function, and rectified linear unit (ReLU) was used as the activation function. The final output node used the Softmax activation function to classify the highest output for each class as the correct answer class [33,34,35,36].
The performance of the algorithm was analyzed while gradually changing the freezing part of the end layer. We used two methods for transfer learning. First, the fully connected layer was removed, and a new fully connected layer was added to construct a new model. The convolutional layer was frozen without training, and only the fully connected layer was trained (transfer learning model 1). Second, a new model was constructed by adding a fully connected layer and varying the range of freezing for the convolutional layer (transfer learning model 1). In this study, a fine-tuned model was designed by freezing up to convolutional block 4 and re-training from convolutional block 5 to the new fully connected layer [33,34,35,36] (Figure 2).
Patients in our study were divided into training (80%) and test (20%) groups. The model was trained on training patients and tested on test patients. The best models were identified by patient-rated performance. We then evaluated test patients using this best model to determine performance with an independent data set.

3. Statistical Analysis

Data were analyzed for both groups and are presented as mean and standard deviation. Chi-square tests were applied to identify significant differences between categorical variables, and we used the Wilcoxon signed-rank test to assess the significance of differences between continuous variables. In the case of propensity score matching, fixed cases were four times more than loose cases, so fixed cases could be adjusted 1:1 according to loosened cases. For each subsequent change to the model, starting with the first model tuned by transfer learning with fine-tuning of the CNN architecture, the model was tested on a test patient. Accuracy in this context is defined as the agreement of the model predictions to known values defined as intraoperative findings of “fixed” or “loosened” implants. Statistical analysis was performed using SPSS® for Windows v21.0, with p < 0.05 indicating statistical significance.

4. Results

A total of 126 patients underwent revision TKA from 2010 to 2020. Twenty-six patients were excluded, including seven due to polyethylene wear, three due to fracture, and sixteen due to infection, with no loosening confirmed. In 100 patients, implant loosening of tibia or femur was confirmed on imaging and surgical records. The average age of the 100 patients was 70.4 years, and 80.0% were female. The average BMI was 26.3, and 37.0% underwent revision TKA on the left side. A control pool was recruited, including 399 patients who underwent primary TKA in 2020, and a control group of 100 patients was extracted through propensity score matching. In the control group, the average age was 70.9 years, 80.0% were female, the average BMI was 26.5, and 37.0% underwent revision TKA on the left side. There were no differences in demographic data between the two groups (all p > 0.05) (Table 1). Before propensity score matching, there was a significant difference in gender and operation side between the two groups, but after propensity score matching, both gender and operation side showed the same results between both groups.
Using 500 epochs, the transfer learning model 1 gradually increased in accuracy and ultimately reached 87.5%. The loss rate gradually decreased to 0.2527 (Figure 3). After assessing the results in the validation set through the confidence matrix, the accuracy was 87.5%, the sensitivity was 90%, the specificity was 100%, the positive predictive value was 100%, and the negative predictive value was 80% (Table 2).
Transfer learning model 2, which was trained on the convolutional layer, gradually increased in accuracy and reached 97.5%, which was a larger improvement than with model 1. The loss rate was 0.0307, confirming improvement compared to model 1 (Figure 4). After checking the results in the validation set through the confidence matrix, the accuracy was 97.5%, the sensitivity was 100%, the specificity was 95%, the positive predictive value was 95.2%, and the negative predictive value was 100%. Transfer learning model 2 exhibited a better sensitivity to detect loosening cases compared to model 1 (Table 2, Figure 5).
Actual fixed and loosened cases were evaluated using transfer learning model 2, and we confirmed that they were accurately judged (Figure 6).

5. Discussion

Because the final diagnosis of prosthetic loosening of TKA implants remains challenging, particularly in the early stages [17], interest in the use of artificial intelligence-based algorithms as a diagnostic tool is increasing [24,25,26,27,28]. The accurate detection of prosthetic loosening of TKA is difficult even for experienced surgeons [18,19]. In this study, the accuracy of detection for prosthetic loosening of TKA was confirmed using a transfer learning model based on VGG 19, a pre-trained CNN model. Two transfer learning model techniques were used with different ranges of freezing, and 97.5% accuracy was achieved using only images with fine-tuning.
The main purpose of TKA is to relieve the patient’s pain and restore function for a long time [37]. It is well known as the most effective and satisfactory surgical treatment in the field of orthopedic surgery and has a very good survival rate [38,39,40,41,42]. The 15-year survival rate is 93.0%; the 20-year survival rate is 90.1%; and the 25-year survival rate is reported to be 82.3% [37]. The result of such excellent long-term survival rates was made possible by the development of artificial joint materials and continuous technology [39]. Unfortunately, artificial joints always have a risk of failure due to various reasons [14]. There are various causes of TKA failure, including loosening, infection, instability, and persistent pain. Among them, the most common cause of TKA failure is loosening caused by osteolysis [14]. Demand for detecting implant loosening has been continuously made [43]. However, it is still difficult to detect implant loosening even though various methods have been used. Classen et al. analyzed aseptic loosening of TKA by bone scintigraphy and reported a sensitivity of 76%, specificity of 83%, positive predictive value of 93%, and a negative predictive value of 56% [43]. Sterner et al. evaluated the loosening of TKA using positron emission tomography and showed a large difference with sensitivity 100% and specificity 56% [44]. Mayer Wagner et al. reported that the detection of TKA loosening using positron emission tomography was less accurate than the detection of THA loosening [45]. The detection of TKA loosening by positron emission tomography showed a sensitivity of 56% for aseptic loosening and 14% for septic loosening [45]. As such, it is difficult to detect TKA loosening, even with more advanced imaging techniques than plain radiographs, and it is common to show large differences in sensitivity, specificity, and accuracy [43,44,45]. In diagnosing implant loosening, many studies have been conducted on many expensive imaging techniques, including positron emission tomography or bone scintigraphy [43,44,45]. However, when compared with a simple radiograph, it did not show a significant improvement in its diagnostic accuracy [43,44,45]. As an alternative to these limitations, interest in diagnostic techniques that have constant and relatively high accuracy in evaluating TKA loosening has increased, and machine learning is receiving great attention as an alternative [24,25,26,27,28]. Machine learning is an innovative method that can improve the diagnosis accuracy without being expensive, exposed to radiation, or invasive [24,25,26,28].
Among machine learning, deep learning is a more advanced and complex form of machine learning that mimics the neural connections in the brain using artificial neural networks (ANNs) arranged and organized in a hierarchical structure [46]. CNNs are another type of deep learning used in computer vision tasks, including medical image analysis, because they show excellent ability for image discrimination and processing [46]. Deep learning analysis technology for evaluating medical images is being applied to radiographic, computed tomography, ultrasound, MR, and fluoroscopic imaging, and significant results have been reported for the diagnosis of diseases of the chest, heart, brain, and breast [47]. In orthopedic surgery, image analysis studies using CNN algorithm structures have been conducted [48], such as measuring bone age using radiographic images [49] and recognizing fractures [50]. To improve the prediction accuracy of models, it is necessary to precede the construction with learning of high-quality images correctly classified by experts and a large training image dataset [51].
However, there are practical difficulties in securing data from large imaging sets. In addition, large-capacity learning data require a considerable learning time, even if high-performance hardware is used [35]. The method proposed to solve this problem is a transfer learning model, which increases the accuracy of the model by adding training data suitable for the study to an existing neural network trained with large amounts of more general data [52]. CNN-based algorithms such as Alexnet [53], VGGNet [32], GoogLeNet [54], and Residual Neural Network (ResNet) [55], proposed through the ImageNet image classification challenge competition, are used for transfer learning. VGGNet is a convolutional neural network with 16 deeply stacked convolution layers and is widely used as a deep learning transfer model in image analysis fields because of its high performance [32]. In this study, the VGG-19 deep learning algorithm, derived from VGGNet, was applied through transfer learning, and an accuracy greater than 95% was achieved.
In general, transfer learning using fine-tuning is a multi-step process. First, the fully connected layer at the end of the network is removed. Second, the fully connected layer is replaced with freshly initialized layers. Third, the earlier convolutional layers are frozen earlier in the network to ensure that previous robust features learned by the CNN are not destroyed. Model training is initiated only for the fully connected layer. Additionally, some/all of the convolutional layers in the network are unfrozen, and a second pass of training can be used to improve functionality [35,36,56]. If the size of the data set is large, it is recommended to use a transfer learning method that trains the entire model. However, the data set is not large in most cases when medical images are used, so most transfer learning models in the medical field only train the fully connected layer [57]. Shah et al. [28] used a transfer learning model for the detection of implant loosening in total hip arthroplasty (THA) and TKA, changing only the last linear layer representing the output with training performed for the entire model. This transfer learning method yielded an accuracy of about 80%. Lau et al. [25] used Xception, a pre-trained model for detecting TKA implant loosening, and observed greater than 96% accuracy through a transfer learning model that trains the whole structure. However, when a pre-trained model is used for medical images with small datasets and low data similarity, it is preferable to only train a part of the model [35,36,56]. In our study, only the fully connected layer was changed to a new fully connected layer, and one model that only trained the fully connected layer and another model that partially trained the convolutional layer were used. In VGG architecture, fine-tuning only the top block results in the best performance compared to the transfer learning model, which relearns the entire model [57]. Such tuning was used with a transfer learning model for detecting prosthetic loosening of TKA in this study for the first time. Our method of training up to the last convolutional layer resulted in the best accuracy.
Our results, including the accuracy, sensitivity, and specificity of evaluating TKA implant loosening using only X-ray images, are not inferior to those of other studies. Previously, Lau et al. [25] analyzed TKA loosening using an image-based machine learning model on 440 knee radiograph images, including 206 with loosening and 234 without loosening. They used the Xception model without fine-tuning and conducted 5000 epochs, resulting in an accuracy of 96.3%, a sensitivity of 96.1%, a specificity of 90.9%, a positive predictive value of 92.4%, a negative predictive value of 95.2%, and an AUC of 93.5%. The accuracy evaluated by the senior orthopedic specialist was 89.09–94.54%. Shah et al. [28] evaluated a loosening detection model using 217 fixed TKA and 137 loosened TKA X-rays. Among the CNN models, ResNet, AlexNet, Inception, and DenseNet were used. The image modification method was used for the transfer learning model, and changes in the layer were not performed. The resulting accuracy was 70.8% when using raw images and untrained models, 73.3% with segmentation and cropping tools, and greater than 80% when using pre-trained models using large datasets. When only the image was used, the accuracy was less than 90%; when clinical information was added, the accuracy was greater than 90%. Compared to THA, the accuracy of detecting TKA loosening was lower. In our study, the transfer learning model 2 showed 97.5% accuracy in the training set and greater than 95% accuracy, sensitivity, and specificity in the test set.
Our study has limitations. First, we only included patients who underwent revision TKA due to loosening. The model’s performance cannot be judged for those whose loosening has not yet been recognized because the model only judged the confirmed loosening of TKA implants on plain radiographs. Therefore, it is questionable whether our model would be effective for detecting loosening during the early stages. It is thought that it can be the basis for developing a model that can recognize loosening at an early stage. Second, preprocessing, which resizes and crops images to substitute for transfer learning models, can reduce image resolution, which can have a significant impact on the model’s performance. Third, since the patients were collected retrospectively, prospective X-ray follow-up of patients is necessary to further validate our results. Fourth, the pre-trained model used in this study was evaluated using only VGGNet [32]. There are several pre-trained models using Imagenets, and a more advanced prosthetic loosening detection model using transfer learning may be established through additional evaluation [28]. Fifth, image augmentation techniques or preprocessing are necessary to overcome reduced image resolution due to the size of images. This may affect performance, including the accuracy of the transfer learning model [29,30]. Sixth, as in other studies, additional patient demographic data and clinical information could further improve the accuracy of the prosthetic loosening model [25,28]. However, the purpose of this study was to detect loosening using only images. Seventh, most of the patients included in the study were women, as the majority of patients undergoing TKA in Asia are women [58,59,60,61]. Gender differences may affect the performance of the model, and further studies that include additional male patients are needed. Finally, the number of images used in this study was insufficient. In general, more than 1000 images are necessary to train a machine learning model [62]. To overcome this, we used image augmentation technology to achieve greater than 95% accuracy. However, it is necessary to include more image data in future studies.

6. Conclusions

The application of the CNN algorithm through transfer learning results in high accuracy for detecting prosthetic loosening of TKA implants by studying plain radiographs. This method may be utilized as an auxiliary tool for diagnosing prosthetic loosening of TKA implants, but it is judged that additional research is needed.

Author Contributions

Y.I. had full access to all the data in the study and takes responsibility for the integrity of the data and the accuracy of the data analysis. Concept and design: M.-S.K. and Y.I.; Acquisition, analysis, or interpretation of data: M.-S.K., R.-K.C., S.-C.Y. and J.-H.H.; Drafting of the manuscript: M.-S.K. and Y.I; Critical revision of the manuscript for important intellectual content: All authors; Administrative, technical, or material support: M.-S.K., R.-K.C., S.-C.Y. and J.-H.H.; Supervision: Y.I. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (2021R1I1A1A01059558).

Institutional Review Board Statement

This study was approved by the institutional review board at Seoul St. Mary’s Hospital (KC21RISI0151).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study. Written informed consent has been obtained from the patients to publish this paper.

Data Availability Statement

The data presented in this study are available in the main article.

Conflicts of Interest

There is no relevant conflict of interest for all authors. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Carr, A.J.; Robertsson, O.; Graves, S.; Price, A.J.; Arden, N.K.; Judge, A.; Beard, D.J. Knee replacement. Lancet 2012, 379, 1331–1340. [Google Scholar] [CrossRef] [PubMed]
  2. Jang, S.; Shin, W.C.; Song, M.K.; Han, H.S.; Lee, M.C.; Ro, D.H. Which orally administered antithrombotic agent is most effective for preventing venous thromboembolism after total knee arthroplasty? A propensity score-matching analysis. Knee Surg. Relat. Res. 2021, 33, 10. [Google Scholar] [CrossRef] [PubMed]
  3. Kulshrestha, V.; Sood, M.; Kumar, S.; Sood, N.; Kumar, P.; Padhi, P.P. Does Risk Mitigation Reduce 90-Day Complications in Patients Undergoing Total Knee Arthroplasty?: A Cohort Study. Clin. Orthop. Surg. 2022, 14, 56–68. [Google Scholar] [CrossRef]
  4. Lee, J.K.; Lee, K.B.; Kim, J.I.; Park, G.T.; Cho, Y.C. Risk factors for deep vein thrombosis even using low-molecular-weight heparin after total knee arthroplasty. Knee Surg. Relat. Res. 2021, 33, 29. [Google Scholar] [CrossRef]
  5. Lee, J.M.; Ha, C.; Jung, K.; Choi, W. Clinical Results after Design Modification of Lospa Total Knee Arthroplasty System: Comparison between Posterior-Stabilized (PS) and PS Plus Types. Clin. Orthop. Surg. 2022, 14, 236–243. [Google Scholar] [CrossRef] [PubMed]
  6. Patrick, N.J.; Man, L.L.C.; Wai-Wang, C.; Tim-Yun, O.M.; Wing, C.K.; Hing, C.K.; Yin, C.K.; Ki-Wai, H.K. No difference in long-term functional outcomes or survivorship after total knee arthroplasty with or without computer navigation: A 17-year survivorship analysis. Knee Surg. Relat. Res. 2021, 33, 30. [Google Scholar] [CrossRef]
  7. Song, S.J.; Kim, K.I.; Suh, D.U.; Park, C.H. Comparison of Patellofemoral-Specific Clinical and Radiographic Results after Total Knee Arthroplasty Using a Patellofemoral Design-Modified Prosthesis and Its Predecessor. Clin. Orthop. Surg. 2021, 13, 175–184. [Google Scholar] [CrossRef]
  8. Takamura, D.; Iwata, K.; Sueyoshi, T.; Yasuda, T.; Moriyama, H. Relationship between early physical activity after total knee arthroplasty and postoperative physical function: Are these related? Knee Surg. Relat. Res. 2021, 33, 35. [Google Scholar] [CrossRef]
  9. Kurtz, S.; Ong, K.; Lau, E.; Mowat, F.; Halpern, M. Projections of primary and revision hip and knee arthroplasty in the United States from 2005 to 2030. J. Bone Jt. Surg. Am. 2007, 89, 780–785. [Google Scholar] [CrossRef]
  10. Ranawat, C.S.; Flynn, W.F., Jr.; Deshmukh, R.G. Impact of modern technique on long-term results of total condylar knee arthroplasty. Clin. Orthop. Relat. Res. 1994, 309, 131–135. [Google Scholar]
  11. Bozic, K.J.; Kurtz, S.M.; Lau, E.; Ong, K.; Chiu, V.; Vail, T.P.; Rubash, H.E.; Berry, D.J. The epidemiology of revision total knee arthroplasty in the United States. Clin. Orthop. Relat. Res. 2010, 468, 45–51. [Google Scholar] [CrossRef]
  12. Cram, P.; Lu, X.; Kates, S.L.; Singh, J.A.; Li, Y.; Wolf, B.R. Total knee arthroplasty volume, utilization, and outcomes among Medicare beneficiaries, 1991–2010. JAMA 2012, 308, 1227–1236. [Google Scholar] [CrossRef] [PubMed]
  13. Geary, M.B.; Macknet, D.M.; Ransone, M.P.; Odum, S.D.; Springer, B.D. Why Do Revision Total Knee Arthroplasties Fail? A Single-Center Review of 1632 Revision Total Knees Comparing Historic and Modern Cohorts. J. Arthroplasty 2020, 35, 2938–2943. [Google Scholar] [CrossRef]
  14. Na, B.R.; Kwak, W.K.; Lee, N.H.; Song, E.K.; Seon, J.K. Trend Shift in the Cause of Revision Total Knee Arthroplasty over 17 Years. Clin. Orthop. Surg. 2023, 15, 219–226. [Google Scholar] [CrossRef] [PubMed]
  15. French, T.H.; Russell, N.; Pillai, A. The diagnostic accuracy of radionuclide arthrography for prosthetic loosening in hip and knee arthroplasty. BioMed Res. Int. 2013, 2013, 693436. [Google Scholar] [CrossRef]
  16. Signore, A.; Sconfienza, L.M.; Borens, O.; Glaudemans, A.; Cassar-Pullicino, V.; Trampuz, A.; Winkler, H.; Gheysens, O.; Vanhoenacker, F.; Petrosillo, N.; et al. Consensus document for the diagnosis of prosthetic joint infections: A joint paper by the EANM, EBJIS, and ESR (with ESCMID endorsement). Eur. J. Nucl. Med. Mol. Imaging 2019, 46, 971–988. [Google Scholar] [CrossRef] [PubMed]
  17. Barnsley, L.; Barnsley, L. Detection of aseptic loosening in total knee replacements: A systematic review and meta-analysis. Skelet. Radiol. 2019, 48, 1565–1572. [Google Scholar] [CrossRef]
  18. Khalily, C.; Whiteside, L.A. Predictive value of early radiographic findings in cementless total hip arthroplasty femoral components: An 8- to 12-year follow-up. J. Arthroplasty 1998, 13, 768–773. [Google Scholar] [CrossRef] [PubMed]
  19. Smith, T.O.; Williams, T.H.; Samuel, A.; Ogonda, L.; Wimhurst, J.A. Reliability of the radiological assessments of radiolucency and loosening in total hip arthroplasty using PACS. HIP Int. 2011, 21, 577–582. [Google Scholar] [CrossRef]
  20. Jamshidi, A.; Pelletier, J.P.; Martel-Pelletier, J. Machine-learning-based patient-specific prediction models for knee osteoarthritis. Nat. Rev. Rheumatol. 2019, 15, 49–60. [Google Scholar] [CrossRef]
  21. Rodríguez-Merchán, E.C. The current role of the virtual elements of artificial intelligence in total knee arthroplasty. EFORT Open Rev. 2022, 7, 491–497. [Google Scholar] [CrossRef]
  22. Hanis, T.M.; Islam, M.A.; Musa, K.I. Diagnostic Accuracy of Machine Learning Models on Mammography in Breast Cancer Classification: A Meta-Analysis. Diagnostics 2022, 12, 1643. [Google Scholar] [CrossRef] [PubMed]
  23. Kuo, K.M.; Talley, P.C.; Chang, C.S. The accuracy of machine learning approaches using non-image data for the prediction of COVID-19: A meta-analysis. Int. J. Med. Inform. 2022, 164, 104791. [Google Scholar] [CrossRef] [PubMed]
  24. Borjali, A.; Chen, A.F.; Muratoglu, O.K.; Morid, M.A.; Varadarajan, K.M. Detecting mechanical loosening of total hip replacement implant from plain radiograph using deep convolutional neural network. arXiv 2019, arXiv:1912.00943. [Google Scholar]
  25. Lau, L.C.M.; Chui, E.C.S.; Man, G.C.W.; Xin, Y.; Ho, K.K.W.; Mak, K.K.K.; Ong, M.T.Y.; Law, S.W.; Cheung, W.H.; Yung, P.S.H. A novel image-based machine learning model with superior accuracy and predictability for knee arthroplasty loosening detection and clinical decision making. J. Orthop. Transl. 2022, 36, 177–183. [Google Scholar] [CrossRef] [PubMed]
  26. Loppini, M.; Gambaro, F.M.; Chiappetta, K.; Grappiolo, G.; Bianchi, A.M.; Corino, V.D.A. Automatic Identification of Failure in Hip Replacement: An Artificial Intelligence Approach. Bioengineering 2022, 9, 288. [Google Scholar] [CrossRef]
  27. Rahman, T.; Khandakar, A.; Islam, K.R.; Soliman, M.M.; Islam, M.T.; Elsayed, A.; Qiblawey, Y.; Mahmud, S.; Rahman, A.; Musharavati, F. HipXNet: Deep Learning Approaches to Detect Aseptic Loos-Ening of Hip Implants Using X-ray Images. IEEE Access 2022, 10, 53359–53373. [Google Scholar] [CrossRef]
  28. Shah, R.F.; Bini, S.A.; Martinez, A.M.; Pedoia, V.; Vail, T.P. Incremental inputs improve the automated detection of implant loosening using machine-learning algorithms. Bone Jt. J. 2020, 102-B, 101–106. [Google Scholar] [CrossRef]
  29. Chowdhury, M.E.; Rahman, T.; Khandakar, A.; Al-Madeed, S.; Zughaier, S.M.; Doi, S.A.; Hassen, H.; Islam, M.T. An early warning tool for predicting mortality risk of COVID-19 patients using machine learning. Cogn. Comput. 2021, 1–16. [Google Scholar] [CrossRef]
  30. Rahman, T.; Khandakar, A.; Qiblawey, Y.; Tahir, A.; Kiranyaz, S.; Kashem, S.B.A.; Islam, M.T.; Al Maadeed, S.; Zughaier, S.M.; Khan, M.S. Exploring the effect of image enhancement techniques on COVID-19 detection using chest X-ray images. Comput. Biol. Med. 2021, 132, 104319. [Google Scholar] [CrossRef]
  31. Lee, K.S.; Jung, S.K.; Ryu, J.J.; Shin, S.W.; Choi, J. Evaluation of Transfer Learning with Deep Convolutional Neural Networks for Screening Osteoporosis in Dental Panoramic Radiographs. J. Clin. Med. 2020, 9, 392. [Google Scholar] [CrossRef] [PubMed]
  32. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  33. Hosny, K.M.; Kassem, M.A.; Foaud, M.M. Classification of skin lesions using transfer learning and augmentation with Alex-net. PLoS ONE 2019, 14, e0217293. [Google Scholar] [CrossRef]
  34. Khan, N.M.; Abraham, N.; Hon, M. Transfer learning with intelligent training data selection for prediction of Alzheimer’s disease. IEEE Access 2019, 7, 72726–72735. [Google Scholar] [CrossRef]
  35. Tajbakhsh, N.; Shin, J.Y.; Gurudu, S.R.; Hurst, R.T.; Kendall, C.B.; Gotway, M.B.; Liang, J. Convolutional neural networks for medical image analysis: Full training or fine tuning? IEEE Trans. Med. Imaging 2016, 35, 1299–1312. [Google Scholar] [CrossRef]
  36. Yosinski, J.; Clune, J.; Bengio, Y.; Lipson, H. How transferable are features in deep neural networks? Adv. Neural Inf. Process. Syst. 2014, 27, 3320–3328. [Google Scholar] [CrossRef]
  37. Evans, J.T.; Walker, R.W.; Evans, J.P.; Blom, A.W.; Sayers, A.; Whitehouse, M.R. How long does a knee replacement last? A systematic review and meta-analysis of case series and national registry reports with more than 15 years of follow-up. Lancet 2019, 393, 655–663. [Google Scholar] [CrossRef]
  38. Bieganowski, T.; Buchalter, D.B.; Singh, V.; Mercuri, J.J.; Aggarwal, V.K.; Rozell, J.C.; Schwarzkopf, R. Bone loss in aseptic revision total knee arthroplasty: Management and outcomes. Knee Surg. Relat. Res. 2022, 34, 30. [Google Scholar] [CrossRef]
  39. Gupta, P.; Czerwonka, N.; Desai, S.S.; deMeireles, A.J.; Trofa, D.P.; Neuwirth, A.L. The current utilization of the patient-reported outcome measurement information system (PROMIS) in isolated or combined total knee arthroplasty populations. Knee Surg. Relat. Res. 2023, 35, 3. [Google Scholar] [CrossRef]
  40. Khanasuk, Y.; Ngarmukos, S.; Tanavalee, A. Does the intramedullary femoral canal plug reduce blood loss during total knee arthroplasty? Knee Surg. Relat. Res. 2022, 34, 31. [Google Scholar] [CrossRef] [PubMed]
  41. Liu, D.W.; Martinez Martos, S.; Dai, Y.; Beller, E.M. The femoral intercondylar notch is an accurate landmark for the resection depth of the distal femur in total knee arthroplasty. Knee Surg. Relat. Res. 2022, 34, 32. [Google Scholar] [CrossRef] [PubMed]
  42. Shon, O.J.; Kim, G.B. Does the degree of intraoperatively identified cartilage loss affect the outcomes of primary total knee arthroplasty without patella resurfacing? A prospective comparative cohort study. Knee Surg. Relat. Res. 2022, 34, 36. [Google Scholar] [CrossRef]
  43. Claassen, L.; Ettinger, M.; Plaass, C.; Daniilidis, K.; Calliess, T.; Ezechieli, M. Diagnostic value of bone scintigraphy for aseptic loosening after total knee arthroplasty. Technol. Health Care 2014, 22, 767–773. [Google Scholar] [CrossRef] [PubMed]
  44. Sterner, T.; Pink, R.; Freudenberg, L.; Jentzen, T.; Quitmann, H.; Bockisch, A.; Löer, F. The role of [18F]fluoride positron emission tomography in the early detection of aseptic loosening of total knee arthroplasty. Int. J. Surg. 2007, 5, 99–104. [Google Scholar] [CrossRef]
  45. Mayer-Wagner, S.; Mayer, W.; Maegerlein, S.; Linke, R.; Jansson, V.; Müller, P.E. Use of 18F-FDG-PET in the diagnosis of endoprosthetic loosening of knee and hip implants. Arch. Orthop. Trauma Surg. 2010, 130, 1231–1238. [Google Scholar] [CrossRef]
  46. Soffer, S.; Ben-Cohen, A.; Shimon, O.; Amitai, M.M.; Greenspan, H.; Klang, E. Convolutional Neural Networks for Radiologic Images: A Radiologist’s Guide. Radiology 2019, 290, 590–606. [Google Scholar] [CrossRef]
  47. Hong, J.-Y.; Park, S.H.; Jung, Y.-J. Artificial intelligence based medical imaging: An Overview. J. Radiol. Sci. Technol. 2020, 43, 195–208. [Google Scholar]
  48. Lalehzarian, S.P.; Gowd, A.K.; Liu, J.N. Machine learning in orthopaedic surgery. World J. Orthop. 2021, 12, 685–699. [Google Scholar] [CrossRef] [PubMed]
  49. Halabi, S.S.; Prevedello, L.M.; Kalpathy-Cramer, J.; Mamonov, A.B.; Bilbily, A.; Cicero, M.; Pan, I.; Pereira, L.A.; Sousa, R.T.; Abdala, N.; et al. The RSNA Pediatric Bone Age Machine Learning Challenge. Radiology 2019, 290, 498–503. [Google Scholar] [CrossRef]
  50. Meena, T.; Roy, S. Bone Fracture Detection Using Deep Supervised Learning from Radiological Images: A Paradigm Shift. Diagnostics 2022, 12, 2420. [Google Scholar] [CrossRef]
  51. Saeed, S.U.; Fu, Y.; Stavrinides, V.; Baum, Z.M.C.; Yang, Q.; Rusu, M.; Fan, R.E.; Sonn, G.A.; Noble, J.A.; Barratt, D.C.; et al. Image quality assessment for machine learning tasks using meta-reinforcement learning. Med. Image Anal. 2022, 78, 102427. [Google Scholar] [CrossRef] [PubMed]
  52. Ng, H.-W.; Nguyen, V.D.; Vonikakis, V.; Winkler, S. Deep learning for emotion recognition on small datasets using transfer learning. In Proceedings of the 2015 ACM on International Conference on Multimodal Interaction, Seattle, WA, USA, 9–13 November 2015; pp. 443–449. [Google Scholar] [CrossRef]
  53. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  54. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar] [CrossRef]
  55. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
  56. Vrbančič, G.; Podgorelec, V. Transfer learning with adaptive fine-tuning. IEEE Access 2020, 8, 196197–196211. [Google Scholar] [CrossRef]
  57. Kandel, I.; Castelli, M. How deeply to fine-tune a convolutional neural network: A case study using a histopathology dataset. Appl. Sci. 2020, 10, 3359. [Google Scholar] [CrossRef]
  58. Chang, M.J.; Ro, D.H.; Kim, T.W.; Lee, Y.S.; Han, H.S.; Chang, C.B.; Kang, S.B.; Lee, M.C. Worse outcome of debridement, antibiotics, and implant retention in acute hematogenous infections than in postsurgical infections after total knee arthroplasty: A multicenter study. Knee Surg. Relat. Res. 2022, 34, 38. [Google Scholar] [CrossRef] [PubMed]
  59. Choi, B.S.; Kim, J.M.; Han, H.S. Decision-making factors and their thresholds for total knee arthroplasty in lateral tibiofemoral osteoarthritis patients: A retrospective cohort study. Knee Surg. Relat. Res. 2022, 34, 41. [Google Scholar] [CrossRef] [PubMed]
  60. Lee, Y.M.; Kim, G.W.; Lee, C.Y.; Song, E.K.; Seon, J.K. No Difference in Clinical Outcomes and Survivorship for Robotic, Navigational, and Conventional Primary Total Knee Arthroplasty with a Minimum Follow-up of 10 Years. Clin. Orthop. Surg. 2023, 15, 82–91. [Google Scholar] [CrossRef]
  61. Ryu, J.J.; Kim, Y.H.; Choi, C.H. The additional tibial stem extension is not mandatory for the stability of 5 mm metal block augmented tibial prosthesis construct in primary total knee arthroplasty: 5-year minimum follow-up results. Knee Surg. Relat. Res. 2023, 35, 5. [Google Scholar] [CrossRef]
  62. Cho, J.; Lee, K.; Shin, E.; Choy, G.; Do, S. How much data is needed to train a medical image deep learning system to achieve necessary high accuracy? arXiv 2015, arXiv:1511.06348. [Google Scholar]
Figure 1. Visual Geometry Group (VGG) 19 architecture.
Figure 1. Visual Geometry Group (VGG) 19 architecture.
Bioengineering 10 00632 g001
Figure 2. Different transfer learning techniques. (A) In the first technique, in which the source weights are fixed and the original fully connected layers would be replaced by new fully connected layers to suit the target dataset, the convolutional layer is frozen without training, and only the fully connected layer is trained. (B) In the second technique, a new model was constructed by adding the last fully connected layer and varying the range of freezing for the convolutional layer.
Figure 2. Different transfer learning techniques. (A) In the first technique, in which the source weights are fixed and the original fully connected layers would be replaced by new fully connected layers to suit the target dataset, the convolutional layer is frozen without training, and only the fully connected layer is trained. (B) In the second technique, a new model was constructed by adding the last fully connected layer and varying the range of freezing for the convolutional layer.
Bioengineering 10 00632 g002
Figure 3. Accuracy and loss curves of transfer learning model 1.
Figure 3. Accuracy and loss curves of transfer learning model 1.
Bioengineering 10 00632 g003
Figure 4. Accuracy and loss curves of transfer learning model 2.
Figure 4. Accuracy and loss curves of transfer learning model 2.
Bioengineering 10 00632 g004
Figure 5. Confusion matrix for control and loosened TKA implants using (A) transfer learning model 1 and (B) transfer learning model 2.
Figure 5. Confusion matrix for control and loosened TKA implants using (A) transfer learning model 1 and (B) transfer learning model 2.
Bioengineering 10 00632 g005
Figure 6. Prediction of fixed and loosened total knee arthroplasty implants.
Figure 6. Prediction of fixed and loosened total knee arthroplasty implants.
Bioengineering 10 00632 g006
Table 1. Patient characteristics before and after propensity score matching (PSM).
Table 1. Patient characteristics before and after propensity score matching (PSM).
Fixed (n = 399)Loosened (n = 100)p-Value
Demographics before PSM
Age (years) *69.7 ± 6.870.4 ± 8.30.430
Gender (female, %)353 (88.5%)80 (80.0%)0.032
BMI (kg/m2)26.1 ± 3.426.3 ± 3.40.430
Operation side (left, %)203 (50.9%)37 (37.0%)0.013
ASA grade 0.073
145 (11.3%)8 (8.0%)
2347 (87.2%)87 (87.0%)
36 (1.5%)5 (5.0%)
Demographics after PSM
Age (years) *70.9 ± 6.770.4 ± 8.30.603
Gender (female, %)80 (80.0%)80 (80.0%)1.000
BMI (kg/m2)26.5 ± 3.626.3 ± 3.40.649
Operation side (left, %)37 (37.0%)37 (37.0%)1.000
ASA grade 0.238
17 (7.0%)8 (7.0%)
292 (92.0%)87 (87.0%)
31 (1.0%)5 (5.0%)
* The values are presented as mean and standard deviation. BMI, body mass index; ASA, American Society of Anesthesiologists.
Table 2. Image-based machine learning model performance using the test set.
Table 2. Image-based machine learning model performance using the test set.
Performance CriteriaTransfer Learning Model 1Transfer Learning Model 2
Accuracy87.5%97.5%
Sensitivity75.0%100%
Specificity100%95.0%
Positive predictive value100%95.2%
Negative predictive value80.0%100%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kim, M.-S.; Cho, R.-K.; Yang, S.-C.; Hur, J.-H.; In, Y. Machine Learning for Detecting Total Knee Arthroplasty Implant Loosening on Plain Radiographs. Bioengineering 2023, 10, 632. https://doi.org/10.3390/bioengineering10060632

AMA Style

Kim M-S, Cho R-K, Yang S-C, Hur J-H, In Y. Machine Learning for Detecting Total Knee Arthroplasty Implant Loosening on Plain Radiographs. Bioengineering. 2023; 10(6):632. https://doi.org/10.3390/bioengineering10060632

Chicago/Turabian Style

Kim, Man-Soo, Ryu-Kyoung Cho, Sung-Cheol Yang, Jae-Hyeong Hur, and Yong In. 2023. "Machine Learning for Detecting Total Knee Arthroplasty Implant Loosening on Plain Radiographs" Bioengineering 10, no. 6: 632. https://doi.org/10.3390/bioengineering10060632

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop