Next Article in Journal
Targeting IRAK4 with Emavusertib in Lymphoma Models with Secondary Resistance to PI3K and BTK Inhibitors
Next Article in Special Issue
A 3-miRNA Risk Scoring Signature in Early Diabetic Retinopathy
Previous Article in Journal
Evaluation of the Performance of ACR TI-RADS Also Considering Those Nodules with No Indication of FNAC: A Single-Center Experience
Previous Article in Special Issue
Genome-Wide Pleiotropy Study Identifies Association of PDGFB with Age-Related Macular Degeneration and COVID-19 Infection Outcomes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

iERM: An Interpretable Deep Learning System to Classify Epiretinal Membrane for Different Optical Coherence Tomography Devices: A Multi-Center Analysis

1
Department of Ophthalmology, The Second Affiliated Hospital of Zhejiang University, College of Medicine, Hangzhou 310009, China
2
School of Mechanical, Electrical and Information Engineering, Shandong University, Weihai 264209, China
3
Department of Ophthalmology, Kagoshima University Graduate School of Medical and Dental Sciences, Kagoshima 890-8520, Japan
4
Ophthalmology and Visual Sciences, Khoo Teck Puat Hospital, National Healthcare Group, Singapore 768828, Singapore
*
Author to whom correspondence should be addressed.
J. Clin. Med. 2023, 12(2), 400; https://doi.org/10.3390/jcm12020400
Submission received: 14 November 2022 / Revised: 29 December 2022 / Accepted: 3 January 2023 / Published: 4 January 2023
(This article belongs to the Special Issue State-of-the-Art Research on Vitreoretinal Disorders)

Abstract

:
Background: Epiretinal membranes (ERM) have been found to be common among individuals >50 years old. However, the severity grading assessment for ERM based on optical coherence tomography (OCT) images has remained a challenge due to lacking reliable and interpretable analysis methods. Thus, this study aimed to develop a two-stage deep learning (DL) system named iERM to provide accurate automatic grading of ERM for clinical practice. Methods: The iERM was trained based on human segmentation of key features to improve classification performance and simultaneously provide interpretability to the classification results. We developed and tested iERM using a total of 4547 OCT B-Scans of four different commercial OCT devices that were collected from nine international medical centers. Results: As per the results, the integrated network effectively improved the grading performance by 1–5.9% compared with the traditional classification DL model and achieved high accuracy scores of 82.9%, 87.0%, and 79.4% in the internal test dataset and two external test datasets, respectively. This is comparable to retinal specialists whose average accuracy scores are 87.8% and 79.4% in two external test datasets. Conclusion: This study proved to be a benchmark method to improve the performance and enhance the interpretability of the traditional DL model with the implementation of segmentation based on prior human knowledge. It may have the potential to provide precise guidance for ERM diagnosis and treatment.

1. Introduction

The epiretinal membrane (ERM) is a pre-retinal proliferative disease, which can result in blurred vision and central vision loss [1,2]. In the 20-year follow-up Beaver Dam Eye Study, the prevalence of ERM was 34.1% [3]. Due to older age being the most common risk factor for ERM, it is likely that the prevalence of ERM will further increase with an aging population. This brings consequential social and economic burdens [4,5].
Optical coherence tomography (OCT), particularly spectral-domain OCT, has been the gold standard for diagnosing ERM in clinical practice due to its higher sensitivity in evaluating ERM compared to other retinal examinations [6,7,8]. OCT can reveal not only the presence of ERM but also the specific severity stage that can assist in guiding disease management [6,9,10]. However, the accurate fine-grading assessment of ERM based on OCT images requires reliable and interpretable analysis methods [11]. The large-scale interpretation of OCT images is deemed to be labor-intensive and time-consuming.
Over the last few years, several studies have shown remarkable results by developing deep learning (DL) systems for automatic retinal disease diagnosis based on OCT images [12,13,14,15]. In the field of ERM diagnosis, those studies prove the ability of DL algorithms to accurately detect ERM from OCT images [16,17,18]. They have the potential to simplify or accelerate OCT image interpretation in clinical practice.
However, several crucial challenges remain. First, no studies to date have tested these algorithms on their classification of ERM into precise severity grading, which could be a guide for referral and treatment. Second, the main drawback of the DL classification system is that it might be difficult for clinicians to trust, thus hindering its practical application. Further effort is required to enhance its interpretability. Third, the OCT data from most research are only obtained from a single commercial device in a single medical center without external datasets to test the generalizability of DL algorithms in a real-world setting.
Thus, to address these gaps, this study developed an interpretable DL system based on multi-center OCT data for ERM severity grading; this was named iERM. It was tested in two independent external datasets. It may have the potential to provide precise guidance in ERM diagnosis and treatment.

2. Materials and Methods

The workflow of the iERM system is demonstrated in Figure 1A, which included data collection, image labeling, development of the segmentation model, and development of the classification model.

2.1. Data Collection

This was a multi-center, retrospective, cross-sectional study. In total, 4547 OCT B-scans (2606 with ERM and 1968 normal) from four different commercial OCT devices from nine international medical centers were collected for this study.
The primary dataset was collected from 1593 eyes of 1046 ERM patients, who underwent a spectral-domain OCT (Heidelberg, Germany) at the Eye Center at the Second Affiliated Hospital of Zhejiang University, from April 2017 to August 2020 (ZJU dataset). All OCT B-scans were central fovea cross sections. Subjects who had other retinopathies other than ERM that can affect the macula or had overt media opacity were excluded. This study was approved by the Medical Ethics Committee of the Second Affiliated Hospital, Zhejiang University, and complied with the Declaration of Helsinki. The demographic and statistical information of all subjects in the ZJU dataset is shown in Table 1. The other ten datasets followed the same data collection as the ZJU dataset, and each OCT image was collected from different subjects. The examples of OCT data collected from different commercial devices are illustrated in Figure 1B.
This retrospective study is a sub-analysis of data from a clinical study (A New Technique for Retinal Disease Treatment, ClinicalTrials.gov identifier: NCT04718532). Ethical approval for this study was obtained from the Ethics Committee for ZJU-2 (No Y2020–1027). The research adhered to the tenets of the Declaration of Helsinki and the Health Insurance Portability and Accessibility Act.

2.2. Image Labeling

In this study, an OCT-based six-stage ERM classification standard was utilized, and the typical case of each stage is shown in Figure 2A [19]. The normal retina was classified in a normal category. In stage 1 of the epiretinal membrane, the mild ERM brought negligible anatomic disruption, while the central fovea had a normal morphology. The outer nuclear layer thickness in the central fovea was markedly thickened, and the central fovea lost its normal morphology in stage 2 of the epiretinal membrane. In stage 3 of the epiretinal membrane, the foveal depression was absent with a continuous ectopic inner foveal layer crossing through the foveal area anomalously. All retinal layers were still identifiable. In stage 4 of the epiretinal membrane, severe retinal thickening and extremely distorted retinal layers were observed. A continuous ectopic inner foveal layer crossing through the foveal area was also identified. Stage 5 of the epiretinal membrane was defined as epiretinal membrane-induced tractional lamellar macular holes or macular pseudo-holes. The label and distribution of all the datasets are shown in Table 1.
For classification, four experienced retinal specialists were recruited to identify the ERM stages of the OCT images independently without communication. The images for the three or four specialists had the same label. The labels of questionable images were confirmed again until full consensus was reached. The label and distribution of all the datasets are shown in Table 2.
For the segmentation work, a five-level stratification of specific layers, namely Layers 1 to 5 was designed according to the key feature of disease progression. The first layer is defined as the morphological curve of the upper boundary of the retina, which is the most distinguishing feature. The second layer was the nerve fiber layer, the ganglion cell layer, and the inner plexiform layer. The inner nuclear layer is the third layer, whose shape is crucial for distinguishing stages 2–4. The fourth layer was the outer plexiform layer and outer nuclear layer, while the last layer was from the external limiting membrane to the pigment epithelium layer. Two experienced retinal specialists were recruited to label the boundaries of these layers. In total, 300 OCT images (50 images for each stage) of the ZJU dataset were labeled to develop the segmentation model.

2.3. Data Processing

There were 5 different image specifications of images generated by different OCT devices, namely Heidelberg 1, Heidelberg 2, Crrius, Zeiss, and Nidek. The specifications of images generated by each OCT device are shown in Table S1. In terms of image mode, Zeiss was a Pseudo-color image, and other images were grayscale images. In the grayscale images, Heidelberg1, Crrius, and Nidek had low grayscale values in the background region and high grayscale values in the retinal region, while Heidelberg 2 is just the opposite. The first step of data processing is to uniformly convert all image modes into grayscale images with low grayscale values in the background region and high grayscale values in the retinal region. The image mode conversion of the Zeiss image and Heidelberg 2 image is shown in Figure S1.
After unifying the image mode, the images of these different devices retained the problems of inconsistent gray value distribution intervals and inconsistent sizes. The gray value distribution histograms are shown in Figure S2. Aiming at the inconsistency of the gray value distribution interval, random changes in brightness, saturation, and contrast were used to enrich the gray value distribution so that the network could be used on images with different gray value distributions. For the problem of inconsistent size, the linear interpolation of the image was unified to the size of 224 × 224, which not only unified the size of the multi-machine images but also conformed to the input size of most classification networks and reduced the computational consumption.
The background area in OCT images, including part of the vitreous chamber and the region below the pigment epithelium layer, was removed to reduce noise. First, a U-Net network was established to distinguish the retina from the background area; then, the maximum connected domain was used to remove the wrong segmentation fragments, and finally, the maximum peripheral rectangle of the retina region was identified [20]. The data augmentation was performed by randomly adjusting the brightness, contrast, and saturation of OCT images and randomly flipping the images in the horizontal direction. Flipping in the vertical direction and rotation were abnegated since they did not follow the natural retinal structure.

2.4. Development of the DL System

(1)
Segmentation model
A U-Net network was selected as the segmentation model for retinal layer segmentation, which is the most classic and extensively used model in medical image segmentation [20]. U-Net was cascaded by a down-sampling pathway and an up-sampling pathway. The encoder in the down-sampling pathway extracted features from the input image and extracted high-dimensional features from low-dimensional features, while the decoder in the up-sampling pathway restored the high-dimensional feature to the low-dimensional feature, and finally generated the segmentation result consistent with the size of the input image. To pursue more accurate segmentation, skipping the connection into the U-Net to fuse low-dimensional detailed information with high-dimensional abstract information was utilized.
The proposed first layer was to detect the upper boundary of the retina within a width of three pixels to enhance the morphological information. The remaining four layers and background were segmented via the U-Net network. The dataset contained 300 images with the annotations described above and were divided for training, verification, and testing according to the ratio of 7:1:2. The RMSProp optimization algorithm was used with a weight decay of 10−8 and a momentum of 0.9 to update the network weights, and a learning rate scheduler with decay of 0.5 per 10 steps was used. The batch size was set to 8, and the maximum epoch was 50. At the end of each epoch, the current network weight was verified, and the weights of the best results were saved.
(2)
Classification Model
Three classification models were compared in this study, and the model with the best results was selected as the backbone network of the proposed two-stage system [21,22,23]. The Xian 1 dataset, the Japanese dataset, and part of the ZJU dataset were divided into two independent external test sets and the internal test set without training. The remaining datasets were combined like the internal dataset to develop the DL model. The ratio for training and verification was 7:1. The specific details in terms of the usage of image data for the classification task is shown in Tables S2 and S3. The SGD optimization algorithm was used with a momentum of 0.9. The initial learning rate was set to 0.001, and a learning rate scheduler with a decay of 0.5 every 30 steps was used. The batch size was set to 32, and the maximum epoch was 80. At the end of each epoch, the current network weight was verified, and the weights of the best results were saved.
(3)
Two-stage System
In the first step of this system, the OCT image was input into the segmentation model and output as a segmentation map. In the second step, the original OCT image and corresponding segmentation map were jointly input into the classification model. The channel concatenation method for merging was utilized. First, the three channels of the original OCT image were standardized, and the values of five channels of the corresponding segmentation map were mapped 0–1 using softmax. The value range of the original image and the segmentation results were consistent after that, which avoided bias with the classification network. Then, the input channel was modified for the first convolutional layer of the classification network to eight to receive three channels simultaneously for the original OCT image and five channels of the corresponding segmentation map as input.

2.5. Statistical Analysis

To achieve a comprehensive and objective assessment of our segmentation network, the following metrics are calculated compared with the ground truth: Dice coefficient (Dice), precision, recall, intersection over union (IoU), and average surface distance (ASD). To evaluate the performance of our classification systems, we generated the receiver operating characteristic (ROC) curve by plotting the true positive rate versus the false positive rate. Compared to the ground truth label generated by clinical ophthalmologists, the precision, recall, F1-score, the area under the curve (AUC) of the ROC curve, and total accuracy were calculated.

2.6. Model Visualization

To further investigate the classification model, a class activation map for the last convolutional layer of the classification model was generated [24]. For the class predicted by the model for each image, the gradient of the last convolutional layer features for that class was identified to find the region that the classification model focuses on. Based on the regions of interest for each image, a more detailed picture of what information the classification model was classifying was generated for the image.

3. Results

3.1. The Performance of the Segmentation Network

The segmentation network has exhibited great performance compared with the ground truth labeled by retinal specialists, and the specific results of the segmentation task are demonstrated in Figure 2B and Table S4. In layers 1–5, Dice scores were 0.997, 0.955, 0.896, 0.962, and 0.958, respectively. The ASD of the layers 1–5 were 0.014, 0.119, 0.308, 0.119, and 0.915, respectively. The mean Dice and ASD were 0.953 and 0.130, respectively, which demonstrated consistency with the ground truth in total and in the boundary.

3.2. The Comparison of Three Classification Networks

The performances of the three typical classification networks were compared in the internal dataset. The ResNet-34 architecture was noted to have the best results with a total accuracy of 81.9%, which was selected to be the backbone of the proposed two-stage network. In stage-specific performance, the areas under the curves (AUCs) of each category were 0.998, 0.924, 0.909, 0.904, 0.970, and 0.983. The detailed results are shown in Table S5.

3.3. The Performance of iERM

The proposed two-stage architecture had a better performance compared with the traditional single-stage classification architecture. For the internal dataset, the AUCs of each category were increased to 0.999, 0.904, 0.925, 0.928, 0.974, and 0.990 with a total accuracy of 82.9% (Table S6). For two external datasets, the total accuracy scores were improved from 84.0% to 87.0% in the Xian 1 dataset and from 73.5% to 79.4% to 79.4% in the Japanese dataset, respectively, with the assistance of segmentation (Table 3 and Table 4). The final AUCs of each category were 0.998, 0.826, 0.906, 0.958, 0.943, and 0.985 for the Xian 1 dataset and 0.999, 0.826, 0.906, 0.958, 0.943, and 0.985 for the Japanese dataset. The performance of fine-grained classification can be advanced by segmentation according to clinical prior knowledge. The receiver operating characteristic (ROC) curves and confusion matrices (CM) of the proposed two-stage system from two external datasets are illustrated in Figure 3 and Figure 4, respectively.

3.4. The Comparison of iERM and Human

Furthermore, four retinal ophthalmologists and four medical students were recruited to identify the stage of ERM for two external test datasets. The average total accuracy scores of four retinal ophthalmologists and four medical students were calculated and compared to the two-stage system. For the Xian 1 dataset, the average total accuracy scores of four ophthalmologists and four medical students were 87.8% and 79.4%, respectively (Figure 1E). For the Japanese dataset, those data were 81.2% and 54.1%. More detailed results are demonstrated in Figures S3 and S4. In conclusion, the DL system’s ability to diagnose and classify ERM was comparable to that of retinal ophthalmologists and performed better than medical students.

4. Discussion

This study has shown potential for AI-assisted fine-grained clinical grading. This study established an effective fine-grained grading system, that is, iERM, to directly guide ERM management. The proposed two-stage architecture, which implemented clinical prior knowledge segmentation as an auxiliary module, proved that it could improve the performance of the traditional classification model. This fine-grading system was evaluated with OCT image data collected from four different commercial OCT devices in multiple international medical centers, and its robustness and applicability were tested in two dependent external datasets. Furthermore, an ERM grading AI interactive interface was developed to show the progression of ERM and corresponding management recommendations applicable in clinical situations for patients.
This study substantially extends existing research and other studies in four aspects. First, ERM was classified into five severity grading subgroups by iERM. This subgroup categorization of ERM could clinically guide ERM management, including screening, referral, re-examination frequency, and surgical decision. According to previous research, decreasing best-corrected visual acuity of ERM patients had an association with the progression to more serious stages graded in OCT images, particularly progression from stage 2 to stage 3 [19]. The fine-grade ERM scheme could also provide a more objective and elaborate reference for clinical practices, rather than just detecting the presence of ERM. For example, patients with stage 5 ERM identified by iERM would be recommended for a referral to specialized retinal clinics regarding vitreoretinal surgery as soon as possible. Furthermore, this study established an iERM system with a user convenient interface (Figure S5). Patients or doctors can upload OCT reports and receive the analyzed results, including automatic diagnosis and management recommendations from iERM. Therefore, this AI system could contribute to the promotion of ophthalmological telemedicine. Additionally, it could improve access to medical services and reduce medical costs, particularly for distant regions. It could also alleviate ophthalmologists’ workload and reduce false-negative diagnosis.
Second, this proposed method integrated the segmentation network, with the classification network substantially increasing the results of automatic fine-grained clinical grading of ERM. Fine-grained clinical grading is a challenge for DL classification due to small inter-class variation and large intra-class variations [25]. To detect normal OCT images in the internal dataset, the Xian 1 dataset, and the Japanese dataset, the precisions were 98%, 100%, and 89.6%, with AUCs of 0.999, 0.998, and 0.999, respectively. This result showed the remarkable ability of iERM to distinguish normal OCT images from ERM images, which is effective and practical in terms of ERM screening. For fine-grained grading, ERM severity was based on OCT images. The similarity of specific grades was quite high compared to junior ophthalmologists or ophthalmologists, who might not be able to tell the difference due to lack of experience. The single-stage classification network achieved total accuracy scores of 81.9%, 84%, and 73.5% for the internal dataset, the Xian 1 dataset, the Japanese dataset, respectively. In Figure 1D, with the assistance of a segmentation network during two-stage architecture, total accuracy scores were noted to improve to 82.9%, 87.0%, and 79.4%, respectively.
The final novelty of this study was the implementation of segmentation with prior clinical knowledge that demonstrated the interpretability of this DL model. The transparency of the DL is an underlying issue, which could impede the application of AI in clinical practice (22, 23). Previous DL studies have integrated the segmentation network with the classification network to tackle the “black box” problem [26,27,28,29]. Jeffrey et al. first trained a DL network to create a segmentation map of 15 lesions or elements and then utilized it as input for second stage classification [26]. The segmentation map provided an interpretable representation for doctors, which was particularly practical in difficult and ambiguous cases. Xu et al. established a hierarchical DL system with prior human knowledge to solve this problem [27]. Extraction and use of anatomical characteristics of the fundus images comprehensively simulated the diagnostic thinking of human experts. In this study, the specific layer of segmentation was designed according to the clinical ERM severity grade standard. The key features of grading ERM progression were the morphological changes of retinal tissues displayed in OCT images. From a retinal specialists’ perspective, clinical diagnosis of ERM based on OCT images was also a two-stage process. First, it requires discovery and analysis of the morphological changes of specific key features and then synthesis of all information to make a diagnosis decision. This study’s proposed method adopted this two-stage clinical diagnosis strategy and was proven to be effective. It is more interpretable for ophthalmologists than the conventional DL classification systems.
The last main contribution of this work is the application of diverse datasets obtained from multiple OCT devices of multiple international medical institutes. The generalizability of DL algorithms has been a concern in multi-center studies [30]. In Figure 1B, the presentation of OCT images was diverse from various OCT devices or various capture patterns of the same OCT device. Though those OCT images displayed similar lesions of the same disease, the different grayscale information resulted in the heterogeneity of the data, which could also create an impediment for the applicability of DL algorithms in a real-world setting. In this study, the DL algorithm was correlated with OCT data from four common commercial devices and tested it on two independent external datasets. This work attempts to bridge the gap of data heterogeneity.
This current study had several limitations. First, the dataset size of OCT images from the other OCT device was relatively small, compared to the Heidelberg device. Thus, large-scale studies should be performed, and an external dataset collected from diverse OCT devices should be created to further test the generalizability of this DL system in the future. Second, the image number of each sub-category was imbalanced due to various prevalence rates of specific stages. Even though this study attempted to adopt some preprocessing methods, the data imbalance could also influence the performance. Finally, the interpretability seemed still insufficient to be applied in clinical practice. It may hinder trust from healthcare providers and result in an important challenge for the integration of AI systems in clinical settings. Joint efforts should be made by healthcare providers and AI developers in future work.

5. Conclusions

This study proved a benchmark method to improve the performance and enhance the interpretability of the traditional DL model with the implementation of segmentation based on prior human knowledge. This study successfully proposed a two-stage DL framework for the fine-grading assessment of ERM progression and achieved advanced results comparable to retinal specialists based on multi-source data, which contributed to efficient clinical ERM management and automatic diagnosis of retinal diseases.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/jcm12020400/s1, Figure S1. The image mode conversion of the Zeiss image and Heidelberg 2 image. (A) Zeiss OCT image (before image mode conversion). (B) Zeiss OCT image (after image mode conversion). (C) Heidelberg 2 OCT image (before image mode conversion). (D) Heidelberg 2 OCT image (after image mode conversion). Figure S2. The gray value distribution histograms of different devices. (A) Heidelberg 1. (B) Heidelberg 2. (C) Crrius. (D) Zeiss. (E) Nidek. Figure S3. Comparison of iERM with retinal ophthalmologists and medical students on the Xian 1 dataset. (A–D) The confusion matrices of four retinal ophthalmologists. (E) the accuracy scores of iERM, four retinal ophthalmologists, and four medical students. Figure S4. Comparison of iERM with retinal ophthalmologists and medical students on the Japan dataset. (A–D) The confusion matrices of four retinal ophthalmologists. (E) the accuracy scores of iERM, four retinal ophthalmologists, and four medical students. Figure S5. The workflow of our user interface. (A) The heat map of each stage. (B) The corresponding recommendation of each stage. Table S1. The results of the segmentation task. Dice: the Dice coefficient. IoU: mean intersection over union. Table S2. The results of three traditional classification models in the internal dataset. Table S3. The results of iERM in the internal dataset compared with those of the traditional classification model. Table S4.The results of the segmentation task. Dice: the Dice coefficient. IoU: mean intersection over union. Table S5.The results of three traditional classification models in the internal dataset. Table S6. The results of iERM in the internal dataset compared with those of the traditional classification model.

Author Contributions

Conceptualization, K.J., Y.Y., Y.W. and J.Y.; Data curation, Y.Y., M.C., X.L., H.T., T.-H.Y. and N.G.S.; Formal analysis, S.W. and X.L.; Investigation, K.J. and T.-H.Y.; Methodology, C.Y. and M.C.; Software, S.W. and C.Y.; Supervision, J.Y.; Validation, S.W., M.C., X.L., H.T., N.G.S. and Y.W.; Visualization, C.Y.; Writing—original draft, K.J. and Y.Y.; Writing—review and editing, K.J. and J.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was financially supported by the National Natural Science Foundation Regional Innovation and Development Joint Fund (U20A20386), the National Key Research and Development Program of China (grant number 2019YFC0118400), Zhejiang Provincial Key Research and Development Plan (grant number 2019C03020), the Natural Science Foundation of China (grant number 81670888), and the Clinical Medical Research Center for Eye Diseases of Zhejiang Province (2021E50007).

Institutional Review Board Statement

This retrospective study is a sub-analysis of data from a clinical study (A New Technique for Retinal Disease Treatment, ClinicalTrials.gov identifier: NCT04718532). Ethical approval for this study was obtained from the Ethics Committee for ZJU-2 (No Y2020–1027). The research adhered to the tenets of the Declaration of Helsinki and the Health Insurance Portability and Accessibility Act.

Informed Consent Statement

Written informed consent was not obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article. As the study was a retrospective review and analysis of fully anonymized retinal fundus images, the medical ethics committee declared it exempt from informed consent.

Data Availability Statement

The de-identified individual participant data can be requested from the corresponding author, who will evaluate such requests on a case-by-case basis.

Acknowledgments

We thank the Second Affiliated Hospital of Xi’an Jiaotong University, Ningbo First Hospital, Jinhua Municipal Central Hospital, Taizhou Hospital of Zhejiang Province, the First Affiliated Hospital of Dali University, and the First Affiliated Hospital of Anhui Medical University for their contributions to data collection.

Conflicts of Interest

The authors have declared that no competing interests exist. None of the authors have any financial/conflicting interests to disclose.

Abbreviations

ERM, Epiretinal membrane; OCT, Optical coherence tomography; DL, Deep learning; IoU, Intersection over union; ASD, Average surface distance; ROC, Receiver operating characteristic curve; AUC, Area under the curve; CM, Confusion matrices.

References

  1. Fraser-Bell, S.; Guzowski, M.; Rochtchina, E.; Wang, J.J.; Mitchell, P. Five-year cumulative incidence and progression of epiretinal membranes: The Blue Mountains Eye Study. Ophthalmology 2003, 110, 34–40. [Google Scholar] [CrossRef] [PubMed]
  2. Carpineto, P.; Ciciarelli, V.; Borrelli, E.; Aharrh-Gnama, A.; Mastropasqua, R. Epiretinal Membrane in Eyes with Vitreomacular Traction. Retina 2019, 39, 1061–1065. [Google Scholar] [CrossRef] [PubMed]
  3. Meuer, S.M.; Myers, C.E.; Klein, B.E.; Swift, M.K.; Huang, Y.; Gangaputra, S.; Pak, J.W.; Danis, R.P.; Klein, R. The Epidemiology of Vitreoretinal Interface Abnormalities as Detected by Spectral-Domain Optical Coherence Tomography. Ophthalmology 2014, 122, 787–795. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Cheung, N.; Tan, S.-P.; Lee, S.Y.; Cheung, G.C.M.; Tan, G.; Kumar, N.; Cheng, C.-Y.; Wong, T.Y. Prevalence and risk factors for epiretinal membrane: The Singapore Epidemiology of Eye Disease study. Br. J. Ophthalmol. 2017, 101, 371–376. [Google Scholar] [CrossRef] [PubMed]
  5. Flaxman, S.R.; Bourne, R.R.A.; Resnikoff, S.; Ackland, P.; Braithwaite, T.; Cicinelli, M.V.; Das, A.; Jonas, J.B.; Keeffe, J.; Kempen, J.H.; et al. Global causes of blindness and distance vision impairment 1990–2020: A systematic review and meta-analysis. Lancet Glob. Health 2017, 5, e1221–e1234. [Google Scholar] [CrossRef] [Green Version]
  6. Do, D.V.; Cho, M.; Nguyen, Q.D.; Shah, S.M.; Handa, J.T.; Campochiaro, P.A.; Zimmer-Galler, I.; Sung, J.U.; Haller, J.A. The impact of optical coherence tomography on surgical decision making in epiretinal membrane and vitreomacular traction. Trans. Am. Ophthalmol. Soc. 2006, 104, 161–166. [Google Scholar] [CrossRef]
  7. Theodossiadis, G.P.; Grigoropoulos, V.G.; Theodoropoulou, S.; Datseris, I.; Theodossiadis, P.G. Spontaneous Resolution of Vitreomacular Traction Demonstrated by Spectral-Domain Optical Coherence Tomography. Am. J. Ophthalmol. 2014, 157, 842–851.e1. [Google Scholar] [CrossRef]
  8. Delyfer, M.; Legout, P.; Le Goff, M.; Blaizeau, M.; Rougier, M.; Schweitzer, C.; Korobelnik, J.; Delcourt, C. Prevalence of epiretinal membranes in the ageing population using retinal colour images and SD-OCT: The Alienor Study. Acta Ophthalmol. 2020, 98, e830–e838. [Google Scholar] [CrossRef]
  9. Konidaris, V.; Androudi, S.; Alexandridis, A.; Dastiridou, A.; Brazitikos, P. Optical coherence tomography-guided classification of epiretinal membranes. Int. Ophthalmol. 2014, 35, 495–501. [Google Scholar] [CrossRef]
  10. Watanabe, K.; Tsunoda, K.; Mizuno, Y.; Akiyama, K.; Noda, T. Outer Retinal Morphology and Visual Function in Patients with Idiopathic Epiretinal Membrane. JAMA Ophthalmol. 2013, 131, 172–177. [Google Scholar] [CrossRef]
  11. Zheng, S.; Choo, J.; Chen, J.; Joshi, S.; Sun, Z. The barriers to global eye care equity and the role of digital innovations. Adv. Ophthalmol. Pr. Res. 2021, 1, 100021. [Google Scholar] [CrossRef]
  12. Kermany, D.S.; Goldbaum, M.; Cai, W.; Valentim, C.C.S.; Liang, H.; Baxter, S.L.; McKeown, A.; Yang, G.; Wu, X.; Yan, F.; et al. Identifying Medical Diagnoses and Treatable Diseases by Image-Based Deep Learning. Cell 2018, 172, 1122–1131.e9. [Google Scholar] [CrossRef] [PubMed]
  13. Hwang, D.-K.; Hsu, C.-C.; Chang, K.-J.; Chao, D.; Sun, C.-H.; Jheng, Y.-C.; Yarmishyn, A.A.; Wu, J.-C.; Tsai, C.-Y.; Wang, M.-L.; et al. Artificial intelligence-based decision-making for age-related macular degeneration. Theranostics 2019, 9, 232–245. [Google Scholar] [CrossRef] [PubMed]
  14. Yim, J.; Chopra, R.; Spitz, T.; Winkens, J.; Obika, A.; Kelly, C.; Askham, H.; Lukic, M.; Huemer, J.; Fasler, K.; et al. Predicting conversion to wet age-related macular degeneration using deep learning. Nat. Med. 2020, 26, 892–899. [Google Scholar] [CrossRef] [PubMed]
  15. Zhang, G.; Fu, D.J.; Liefers, B.; Faes, L.; Glinton, S.; Wagner, S.; Struyven, R.; Pontikos, N.; Keane, P.A.; Balaskas, K. Clinically relevant deep learning for detection and quantification of geographic atrophy from optical coherence tomography: A model development and external validation study. Lancet Digit. Health 2021, 3, e665–e675. [Google Scholar] [CrossRef] [PubMed]
  16. Lu, W.; Tong, Y.; Yu, Y.; Xing, Y.; Chen, C.; Shen, Y. Deep Learning-Based Automated Classification of Multi-Categorical Abnormalities from Optical Coherence Tomography Images. Transl. Vis. Sci. Technol. 2018, 7, 41. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Sonobe, T.; Tabuchi, H.; Ohsugi, H.; Masumoto, H.; Ishitobi, N.; Morita, S.; Enno, H.; Nagasato, D. Comparison between support vector machine and deep learning, machine-learning technologies for detecting epiretinal membrane using 3D-OCT. Int. Ophthalmol. 2018, 39, 1871–1877. [Google Scholar] [CrossRef]
  18. Lo, Y.-C.; Lin, K.-H.; Bair, H.; Sheu, W.H.-H.; Chang, C.-S.; Shen, Y.-C.; Hung, C.-L. Epiretinal Membrane Detection at the Ophthalmologist Level using Deep Learning of Optical Coherence Tomography. Sci. Rep. 2020, 10, 8424. [Google Scholar] [CrossRef]
  19. Govetto, A.; Lalane, R.A., III; Sarraf, D.; Figueroa, M.S.; Hubschman, J.P. Insights Into Epiretinal Membranes: Presence of Ectopic Inner Foveal Layers and a New Optical Coherence Tomography Staging Scheme. Am. J. Ophthalmol. 2017, 175, 99–113. [Google Scholar] [CrossRef]
  20. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention, PT III, Nagoya, Japan, 22–26 September 2013; pp. 234–241. [Google Scholar]
  21. He, K.M.; Zhang, X.Y.; Ren, S.Q.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  22. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  23. Tan, M.X.; Le, Q.V. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. In Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  24. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-based Localization. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  25. Wei, X.-S.; Song, Y.-Z.; Mac Aodha, O.; Wu, J.; Peng, Y.; Tang, J.; Yang, J.; Belongie, S. Fine-Grained Image Analysis With Deep Learning: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 8927–8948. [Google Scholar] [CrossRef]
  26. De Fauw, J.; Ledsam, J.R.; Romera-Paredes, B.; Nikolov, S.; Tomasev, N.; Blackwell, S.; Askham, H.; Glorot, X.; O’Donoghue, B.; Visentin, D.; et al. Clinically applicable deep learning for diagnosis and referral in retinal disease. Nat. Med. 2018, 24, 1342–1350. [Google Scholar] [CrossRef] [PubMed]
  27. Xu, Y.; Hu, M.; Liu, H.; Yang, H.; Wang, H.; Lu, S.; Liang, T.; Li, X.; Xu, M.; Li, L.; et al. A hierarchical deep learning approach with transparency and interpretability based on small samples for glaucoma diagnosis. Npj Digit. Med. 2021, 4, 48. [Google Scholar] [CrossRef] [PubMed]
  28. Zhou, Y.; Wang, B.; Huang, L.; Cui, S.; Shao, L. A Benchmark for Studying Diabetic Retinopathy: Segmentation, Grading, and Transferability. IEEE Trans. Med. Imaging 2021, 40, 818–828. [Google Scholar] [CrossRef] [PubMed]
  29. Sussillo, D.; Barak, O. Opening the Black Box: Low-Dimensional Dynamics in High-Dimensional Recurrent Neural Networks. Neural Comput. 2013, 25, 626–649. [Google Scholar] [CrossRef] [PubMed]
  30. He, J.; Baxter, S.L.; Xu, J.; Xu, J.; Zhou, X.; Zhang, K. The practical implementation of artificial intelligence technologies in medicine. Nat. Med. 2019, 25, 30–36. [Google Scholar] [CrossRef]
Figure 1. Overview of iERM for ERM severity grading through OCT images. (A) Workflow of the iERM system. (B) The examples of OCT data collected from different commercial devices. (C) A brief demonstration of the proposed two-stage system and the schematic diagram of deep learning networks. (D) The performance of ERM severity grading on the internal dataset and two external datasets was improved by the two-stage system, compared to the traditional classification model. (E) Comparison of the average total accuracy scores of four retinal ophthalmologists and four medical students in two external datasets with those corresponding to the two-stage system and ResNet-34.
Figure 1. Overview of iERM for ERM severity grading through OCT images. (A) Workflow of the iERM system. (B) The examples of OCT data collected from different commercial devices. (C) A brief demonstration of the proposed two-stage system and the schematic diagram of deep learning networks. (D) The performance of ERM severity grading on the internal dataset and two external datasets was improved by the two-stage system, compared to the traditional classification model. (E) Comparison of the average total accuracy scores of four retinal ophthalmologists and four medical students in two external datasets with those corresponding to the two-stage system and ResNet-34.
Jcm 12 00400 g001
Figure 2. The demonstration of classification standard and segmentation results. (A) The typical case of each stage and their corresponding segmentation maps. (B) The Dice coefficient (Dice) scores and mean intersection over union (IoU) of layers 1–5. Ori: the original OCT images. Seg: the segmentation maps.
Figure 2. The demonstration of classification standard and segmentation results. (A) The typical case of each stage and their corresponding segmentation maps. (B) The Dice coefficient (Dice) scores and mean intersection over union (IoU) of layers 1–5. Ori: the original OCT images. Seg: the segmentation maps.
Jcm 12 00400 g002
Figure 3. The results of iERM in the Xian 1 dataset. (A) The receiver operating characteristic (ROC) curve of iERM. (B) The confusion matrices (CM) of iERM. (C) Comparison of iERM with the traditional classification model on the precision of each stage. (D) Comparison of iERM with the traditional classification model for the recall of each stage.
Figure 3. The results of iERM in the Xian 1 dataset. (A) The receiver operating characteristic (ROC) curve of iERM. (B) The confusion matrices (CM) of iERM. (C) Comparison of iERM with the traditional classification model on the precision of each stage. (D) Comparison of iERM with the traditional classification model for the recall of each stage.
Jcm 12 00400 g003
Figure 4. The results of iERM in the Japanese dataset. (A) The receiver operating characteristic (ROC) curve of iERM. (B) The confusion matrices (CM) of iERM. (C) Comparison of iERM with the traditional classification model on the precision of each stage. (D) Comparison of iERM with the traditional classification model on the recall of each stage.
Figure 4. The results of iERM in the Japanese dataset. (A) The receiver operating characteristic (ROC) curve of iERM. (B) The confusion matrices (CM) of iERM. (C) Comparison of iERM with the traditional classification model on the precision of each stage. (D) Comparison of iERM with the traditional classification model on the recall of each stage.
Jcm 12 00400 g004
Table 1. The demographics and statistics information of patients in the ZJU dataset.
Table 1. The demographics and statistics information of patients in the ZJU dataset.
Age (Year)Male Sex (%)OD/OSOCT Images
Normal61.52 ± 7.9044.60.961674
Stage 163.12 ± 9.1349.40.98288
Stage 264.35 ± 7.9844.11.02440
Stage 365.52 ± 8.7147.60.96741
Stage 465.14 ± 6.3445.91.03185
Stage 565.47 ± 7.8352.01.03229
Total63.29 ± 8.2946.00.983557
Table 2. The data distribution of all datasets.
Table 2. The data distribution of all datasets.
DatasetOCT DeviceImage NumberNormalStage 1Stage 2Stage 3Stage 4Stage 5
ZJUHeidelberg35571674288440741185229
Xian 1Heidelberg2701689156297
NingboHeidelberg841710163533
JinhuaHeidelberg126401211421110
DaliHeidelberg12102669386
AnhuiHeidelberg190101332
JapanHeidelberg11761533747
Singapore 1Heidelberg71851719517
TaizhouNidek220541210
Xian 2Zeiss1090161049268
Singapore 2Cirrus780211223814
Table 3. The results of iERM in the Xian 1 dataset compared to the traditional classification model.
Table 3. The results of iERM in the Xian 1 dataset compared to the traditional classification model.
Xian 1NormalStage 1Stage 2Stage 3Stage 4Stage 5
without segmentation
Precision (%)98.136.433.380.350.083.3
Recall (%)93.588.933.379.033.371.4
F1-score (%)95.751.633.379.740.076.9
AUC (%)99.486.381.797.189.797.1
Accuracy (%) 84.0
with segmentation
Precision (%)100.030.464.384.460.066.7
Recall (%)94.077.860.087.133.357.1
F1-score (%)96.943.762.185.742.961.5
AUC (%)99.882.690.695.894.398.5
Accuracy (%) 87.0
Table 4. The results of iERM in the Japanese dataset compared to the traditional classification model.
Table 4. The results of iERM in the Japanese dataset compared to the traditional classification model.
JapanNormalStage 1Stage 2Stage 3Stage 4Stage 5
without segmentation
Precision (%)98.333.35.971.40.080.0
Recall (%)96.740.033.354.10.057.1
F1-score (%)97.536.410.061.50.066.7
AUC (%)96.572.383.390.984.596.0
Accuracy (%) 73.5
with segmentation
Precision (%)89.633.325.082.1100.0100.0
Recall (%)98.460.066.762.225.057.1
F1-score (%)93.742.936.670.840.072.7
AUC (%)99.978.795.990.288.598.1
Accuracy (%) 79.4
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jin, K.; Yan, Y.; Wang, S.; Yang, C.; Chen, M.; Liu, X.; Terasaki, H.; Yeo, T.-H.; Singh, N.G.; Wang, Y.; et al. iERM: An Interpretable Deep Learning System to Classify Epiretinal Membrane for Different Optical Coherence Tomography Devices: A Multi-Center Analysis. J. Clin. Med. 2023, 12, 400. https://doi.org/10.3390/jcm12020400

AMA Style

Jin K, Yan Y, Wang S, Yang C, Chen M, Liu X, Terasaki H, Yeo T-H, Singh NG, Wang Y, et al. iERM: An Interpretable Deep Learning System to Classify Epiretinal Membrane for Different Optical Coherence Tomography Devices: A Multi-Center Analysis. Journal of Clinical Medicine. 2023; 12(2):400. https://doi.org/10.3390/jcm12020400

Chicago/Turabian Style

Jin, Kai, Yan Yan, Shuai Wang, Ce Yang, Menglu Chen, Xindi Liu, Hiroto Terasaki, Tun-Hang Yeo, Neha Gulab Singh, Yao Wang, and et al. 2023. "iERM: An Interpretable Deep Learning System to Classify Epiretinal Membrane for Different Optical Coherence Tomography Devices: A Multi-Center Analysis" Journal of Clinical Medicine 12, no. 2: 400. https://doi.org/10.3390/jcm12020400

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop