Next Article in Journal
Exploring the Optimal Timing of Endoscopic Ultrasound Performance Post-Acute Idiopathic Pancreatitis
Next Article in Special Issue
Dose Descriptors and Assessment of Risk of Exposure-Induced Death in Patients Undergoing COVID-19 Related Chest Computed Tomography
Previous Article in Journal
Thousands of Women’s Lives Depend on the Improvement of Poland’s Cervical Cancer Screening and Prevention Education as Well as Better Networking Strategies Amongst Cervical Cancer Facilities
Previous Article in Special Issue
COVI3D: Automatic COVID-19 CT Image-Based Classification and Visualization Platform Utilizing Virtual and Augmented Reality Technologies
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

COVID-19 Infection Segmentation and Severity Assessment Using a Self-Supervised Learning Approach

1
School of Computer Science and Technology, Wuhan University of Science and Technology, Wuhan 430065, China
2
Hubei Province Key Laboratory of Intelligent Information Processing and Real-Time Industrial System, Wuhan 430065, China
3
Wuhan First Hospital, Wuhan 430030, China
4
Department of Health Administration and Policy, College of Health and Human Services, George Mason University, Fairfax, VA 22030, USA
*
Authors to whom correspondence should be addressed.
Diagnostics 2022, 12(8), 1805; https://doi.org/10.3390/diagnostics12081805
Submission received: 29 June 2022 / Revised: 14 July 2022 / Accepted: 23 July 2022 / Published: 26 July 2022
(This article belongs to the Special Issue The Role of CT in 2019 Novel Coronavirus Pneumonia (COVID-19))

Abstract

:
Background: Automated segmentation of COVID-19 infection lesions and the assessment of the severity of the infections are critical in COVID-19 diagnosis and treatment. Based on a large amount of annotated data, deep learning approaches have been widely used in COVID-19 medical image analysis. However, the number of medical image samples is generally huge, and it is challenging to obtain enough annotated medical images for training a deep CNN model. Methods: To address these challenges, we propose a novel self-supervised deep learning method for automated segmentation of COVID-19 infection lesions and assessing the severity of infection, which can reduce the dependence on the annotation of the training samples. In the proposed method, first, many unlabeled data are used to pre-train an encoder-decoder model to learn rotation-dependent and rotation-invariant features. Then, a small amount of labeled data is used to fine-tune the pre-trained encoder-decoder for COVID-19 severity classification and lesion segmentation. Results: The proposed methods were tested on two public COVID-19 CT datasets and one self-built dataset. Accuracy, precision, recall, and F1-score were used to measure classification performance and Dice coefficient was used to measure segmentation performance. For COVID-19 severity classification, the proposed method outperformed other unsupervised feature learning methods by about 7.16% in accuracy. For segmentation, when the amount of labeled data was 100%, the Dice value of the proposed method was 5.58% higher than that of U-Net.; in 70% of the cases, our method was 8.02% higher than U-Net; in 30% of the cases, our method was 11.88% higher than U-Net; and in 10% of the cases, our method was 16.88% higher than U-Net. Conclusions: The proposed method provides better classification and segmentation performance under limited labeled data than other methods.

1. Introduction

Currently, the novel coronavirus disease 2019 (COVID-19) is spreading rapidly around the world, seriously affecting people’s daily life. The gold standard for detecting COVID-19 is reverse transcription-polymerase chain reaction (RT-PCR) [1], which uses the combination of RNA reverse transcription and polymerase chain reaction to detect viral RNA fragments. However, RT-PCR tests consume time to obtain the results and have relatively low sensitivity, which does not meet the appeal of detecting positive cases as quickly as possible to separate the persons with positive tests. Another solution for rapid screening is to use medical imaging, such as chest X-ray images or computed tomography (CT) scanners [2]. In addition, some methods have also been proposed for the treatment of COVID-19 [3,4]. In medical image analysis, many image-based artificial intelligence approaches have been developed to help combat the disease, including automatic diagnosis, segmentation, and prognosis. However, most previous image-based studies focused on identifying COVID-19 patients. Moreover, approximately 80% of patients with COVID-19 have only mild to moderate symptoms, and only 20% have severe symptoms [5,6]. Therefore, there is no doubt that automatic assessment of COVID-19 severity is of great significance for clinical diagnosis. At the same time, the segmentation of new coronary pneumonia lesions in lung CT images can also assist doctors in diagnosing the disease.
In recent years, deep learning has achieved great success in medical image analysis. The convolutional neural network’s (CNN) success is mainly due to its ability to extract highly representative features. However, it relies heavily on many high-quality manual annotations. There are issues when manual annotations are used: expensive acquisition costs and patient data privacy. To alleviate the challenge of complex medical image annotation, several solutions were provided in the past:
  • Transfer learning [7] was used to fine-tune the pre-trained model for the target task. However, transfer learning can be impractical and inefficient because the predefined model’s architecture is not as accurate as transfer learning from medical images because of the differences in image features.
  • Semi-supervised learning [8] was proposed to exploit an unlimited amount of unlabeled data to improve performance. However, it usually assumes that labeled data is sufficient to cover the data distribution, with smoothness assumptions, low-density separation assumptions, and popular assumptions, such as adversarial learning [9] and noisy labels [10].
  • Active learning [11] was developed to select the most representative samples that are annotated without unannotated data but only focus on saving manual labor.
  • Self-Supervised Learning [12], also known as unsupervised visual representation learning, can be achieved by providing a strategy to pre-train the neural network with unlabeled data, draw the prior knowledge distribution of the image itself, and then fine-tune the downstream tasks with limited annotation to transfer the knowledge learned during the pre-training process to the downstream tasks and improve the downstream tasks.
In this paper, we developed a new SSL approach. Our new approach first pre-trains a convolutional neural network on a pretext task with a large unannotated dataset. It then uses a small set of annotated data to fine-tune the pre-trained network for a specific target task. The pre-task uses a neural network to deeply mine useful information from the unlabeled raw data to improve the performance of subsequent target tasks when the training data is limited.
Various pretext tasks have been proposed, including grayscale image colorization prediction [13], puzzle prediction [14], object motion and depth estimation [15], rotation prediction [16], etc. Self-supervision in the image domain can be divided into generative self-supervised learning and discriminative self-supervised learning. Variational auto-encoder (VAE) [17] and generative adversarial network (GAN) [18] are generative self-supervised learning, while contrastive learning (CL) is a typical discriminative self-supervised learning. Contrastive learning defines a contrastive prediction task, which namely attempts to maximize the similarity between the features in different augmented views of the same image through a contrastive loss while maximizing the distance between the features in different images. Studies have shown that representations learned through contrastive learning gradually approach learning through strong supervision to the characterization, even when only limited data or small-scale datasets are available. Various types of self-supervised methods have been developed for a variety of medical applications, such as object recognition for spine MRI [19], cardiac MR image segmentation [20], lung lobe segmentation and nodule detection [21], intracerebral hemorrhage classification, brain tumors segmentation [22,23], etc.
Inspired by the rotation-dependent and rotation-invariant features [24,25,26,27], we formulate a rotation prediction task and instance discrimination task to learn rotation-dependent features through rotation transformations and identify instances without considering image rotations. We observed that the blood vessels in the pneumonia image are often interlaced with ground glass or consolidation, which is very sensitive to the direction. The blood vessels interspersed by the lesion have a specific direction, as shown in Figure 1, and the CT image was rotated by 90°, 180°, and 270°. Afterwards, the apparent orientation of these structures will be changed. Therefore, learning the predicted rotation helps to learn the deep features of pneumonia images, which is beneficial for self-supervised feature learning while improving the accuracy of downstream diagnostic tasks. However, CNNs with pooling layers are less sensitive to some spatial transformations, such as positional changes, for segmentation tasks. Furthermore, such spatially invariant features are advantageous for classification problems but a disadvantage for segmentation problems. So, in addition to learning salient features, we also propose a multi-view instance discrimination task to learn rotation-invariant features. As shown in Figure 2, the multi-view instance discrimination task aims to learn a feature representation such as the representation of the transformed version of the input image but different from other images, where transformations include random cropping, color dithering, random erasing, and binarization. By formulating collaborative learning tasks, namely rotation prediction and multi-view instance discrimination, we propose the network to discover the discriminative structure of pneumonia CT images and explore for diagnosis and lesion segmentation of COVID-19 severity. Finally, after appending a decoder with an encoder, we pre-train the encoder-decoder architecture to obtain proper initializations for downstream segmentation tasks.
In summary, our main contributions are three-fold.
  • We propose a novel SSL framework that provides an encoder for classification and an encoder-decoder for segmentation for severity classification and lesion segmentation from COVID-19 CT images.
  • We formulate a collaborative learning task that splits features into learning rotation-dependent and invariant representations, which can learn rotation-discriminative features from pneumonia images and rotation-independent features. When comparing prediction tasks, we investigate different data augmentation operations in the definition. The results show that random cropping, color dithering, random erasing, and binarization combined are helpful for the pretext task of CT images to learn discriminative feature representations.
  • Experiments are conducted on an aggregated dataset consisting of three COVID-19 CT image datasets. The results show that our self-supervised method achieved better results than other state-of-the-art methods. Our method has demonstrated superior performance in COVID-19 severity diagnosis and lesion segmentation. Our method can reduce the performance gaps caused by the number of annotated datasets (10% vs. 100% annotations) with limited annotated datasets.

2. Related Works

2.1. Deep Learning for COIVD-19 Diagnosis Based on Classification and Segmentation Tasks

Deep learning-based methods have been widely used in the automatic diagnosis and lesion segmentation of COVID-19 [28,29,30]. For example, Qian et al. [31] proposed a multi-task multi-slice deep learning system (M3 Lung-Sys) for multi-type lung pneumonia screening from CT images. The system consists of only two 2D CNN networks. The former is designed to find feature representations from rich CT slices, and the latter can recover temporal information through feature refinement and aggregation between different slices. In addition to linking COVID-19 with health, H1N1 and CAP cases are distinguished, and the system can also locate the area of related lesions. Zhang et al. [32] proposed a detection and diagnosis model based on EfficientNet, using the EfficientNet-B0 network to pre-train ImageNet to extract features. An anomaly detection module and a confidence score prediction module was proposed to classify COVID-19 and non-COVID-19 patients. Liu et al. [33] proposed a new weakly supervised segmentation method for COVID-19 infection in CT slices, which only requires scribble supervision and is enhanced by uncertainty-aware self-integration and transformation consistency techniques. Wu et al. [34] proposed a sequential region generation network (SRGNet) to detect and segment the lesion areas of COVID-19 jointly. SRGNet can use the supervised segmentation information and then outputs multi-scale segmentation predictions and generates high-quality lesion area proposals on the predicted segmentation map. At the same time, the detection results, in turn, refine the segmentation map through the post-processing process, which significantly improves the segmentation accuracy.
In the application of COVID-19 classification and segmentation, a few works were devoted to the severity assessment of COVID-19. For example, He et al. [5] proposed a synergistic learning framework for automated severity assessment of COVID-19 in 3D CT images by jointly performing lung lobe segmentation and multi-instance classification. Goncharov et al. [35] proposed a new convolutional neural network model. ResNet50 [36] was trained as a classification network in their model, and the distances between the voxels were finally used to segment the lesion area based on U-Net [37]. The proportion of the lesion area in the lung area corresponds to a different severity degree. However, these works are based on supervised or weakly supervised learning, which adopts a large amount of labeled data for training, requiring a lot of time. Unlike previous work, this paper proposes a self-supervised method, which makes full use of a large amount of unlabeled data and reduces the labeling work of doctors.

2.2. Self-Supervised Learning

Self-supervised learning is a widely studied area [38,39], and Doersch et al. [40] proposed a framework to learn visual features by predicting the relative positions of two patches from the same image. Another representative method for relative position prediction is the jigsaw puzzle proposed by Noroozi [14]. This work requires a deep learning network to rearrange the positions of nine patches cropped from the same image. In addition, colorization [41] can also be formulated as a pretext task for pre-trained neural networks. Recently, models with SSL ideas began to be gradually applied in the field of medical images, and pretext tasks used in medical images include Rubik’s Cube and Rubik’s Cube + Recovery, anatomical location prediction, and reconstructing a part of an image. For example, Zhuang et al. [22] proposed a self-supervised learning framework for 3D medical images. It proposes a new pretext task-Rubik’s Cube Restoration, i.e., cube rearrangement, and cube rotation, to pre-train a 3D neural network that learns translation and rotation invariant features from raw data for subsequent intracerebral hemorrhage classification and brain tumor segmentation task. Zhu et al. [23] further improved this method. They proposed a new pretext task, adding a masking recognition pretext task based on the two pretext tasks of Zhuang [22], which increased the difficulty of recovering the Rubik’s Cube. Deep learning networks are encouraged to utilize more spatial information and produce more robust feature representations. Bai et al. [20] formulated an anatomical location prediction pretext task to learn self-supervised features for cardiac MR image segmentation. Liang et al. [42] proposed a novel self-supervised learning strategy based on context recovery, which allows the model to learn a priori visual features by reconstructing the original image, which is then applied to many medical image tasks. Recently, contrastive learning methods based on instance discrimination tasks have achieved state-of-the-art performance in SSL. The main idea of contrastive learning methods is to make representations of different views of the same representations (“negative pairs”) of the views of the image that are separated. For example, Dosovitskiy et al. [43] proposed to use SoftMax embeddings with classifier weights to compute feature similarity. However, it prevents the explicit comparison of features, which leads to limited efficiency and discriminability. Wu et al. [44] developed a memory bank to memorize the features of each instance. Ye [45] computes positive set attributes based on “real” instance features rather than classifier weights or the memory bank. There are also some self-supervisions in COVID-19 learning work. Chen et al. [46] proposed a self-supervised learning method. They used contrastive learning to train an encoder that could capture expressive feature representations on a large publicly available lung dataset. A prototype network is used for the classification of the task. We propose a new collaborative method to learn complementary information from different pretext tasks, i.e., rotation-related features, and rotation-invariant features, to better use unlabeled data and improve model performance in generalization ability on downstream tasks.

3. Materials and Methods

The overall architecture of the proposed self-supervised method for assessing the severity of COVID-19 is shown in Figure 2. The proposed self-supervised method comprises four modules: image transform enhancement module, rebuild task module, instance discrimination task module, and rotation prediction task module.
This paper randomly selects m slices from the training dataset S = { x i } i = 1 N . For each image slice x i , two new images, x ^ i and x ˜ i , are generated from it by applying two of the four image enhancement techniques. The enhanced image of slides x 1 and x 2 are shown in Figure 2. Then, these images are rotated at 0°, 90°, 180°, and 270°to further generate 8 rotated images (denoted by x i ). After that, each rotated image is assigned a corresponding rotation label 0, 1, 2, and 3. On this basis, a feature embedding network F ( ; θ ) is proposed, which maps the input x i to a high-dimensional feature vector f i and then is decoupled into f i ( d ) and f i ( r ) . These two decoupled features are co-optimized by the multi-view instance discrimination and rotation prediction tasks. Finally, the features learned from the rotation prediction task are utilized for COVID-19 severity classification. The following subsection will introduce the rotation prediction task, multi-view instance discrimination task, reconstruction task, and encoder-decoder details in detail.

3.1. Rotation Prediction Task

To discover the salient features of COVID-19 CT images, we use the rotation prediction task module to learn rotation-related features. The input x i is the input to the encoder, and we represent the output of the last convolutional layer as a feature f i . The input x i is rotated to get x i , y ; the output f i feature should be f i , y , y { 0 , 1 , 2 , 3 } . To simplify the description, here, f i is used instead of f i , y . Then, to reduce the feature dimension and obtain a high-level representation, f i is followed by a module, which is FC, BN, and ReLU. Then, f i is decoupled into f i ( d ) and f i ( r ) along with the channel layer, respectively. Finally, a fully connected layer is denoted as F c ( ; θ c ) , inputs f i ( r ) , and after a Softmax operation, outputs four probability values corresponding to 0, 1, 2, and 3. The rotation prediction loss is expressed as:
r = 1 4 N i = 1 N y = 0 3 l ( F c ( f i , y ( r ) ; θ c ) , y ) .
where is the cross-entropy loss for the classification task and is the rotation label.

3.2. Multi-View Instance Discrimination Task

To learn rotation and transformation invariant representations for COVID-19 lesion segmentation, we proposed a multi-view instance discrimination task. As shown in Figure 2, x ˜ i , y , x ^ i , y represent different data-augmented views of x i . The key assumption of the multi-view instance discrimination task is that good features are shared among the various views of the same image. Therefore, the key is that different data-augmented views (front-facing) of a single image should remain invariant in the embedding space, while different data-augmented views from different patient’s images (negative pair) should be discrepant. After obtaining the decoupling feature f i ( d ) , this paper first uses l 2 to normalize to make f i ( d ) 2 = 1 . For simplicity, in this section, this paper uses f i to represent f i ( d ) , positive pairs are denoted as ( f ^ i , y , f ˜ i , k ) , y and k are denoted rotation labels, y , k { 0 , 1 , 2 , 3 } , and negative pairs are denoted as ( f ^ i , y , f ˜ j , k ) , i j , as shown in Figure 2. For each image x i , the augmented sample should be classified into class i , and other images from x ^ i , y cannot be classified into class i. For X, the probability of being classified as class i is:
P ( i | x ^ i , y ) =   exp   ( k = 0 3 f ˜ i , y T f ^ i , k / τ ) j = 1 m k = 0 3 f ˜ j , k T f ^ i , y   e x p  
In the above formula, τ is the temperature parameter, and by default, τ is set to 0.1. f ˜ i , y T f ^ i , k represents the cosine similarity between positive pairs, and f ˜ j , k T f ^ i , y represents the cosine similarity between negative pairs. Embedding through Softmax in Equation (2) function, the network pushes “negative pairs” from different samples away and pulls “positive pairs” from the same sample closer. The goal is to minimize the contrastive loss, as described below:
d = i y l o g P ( i | x ^ i , y ) i j i y l o g { ( 1 P ( i | x ˜ j , y ) ) }
P ( i | x ^ i , y ) represents the probability that x ^ i , y is classified into class i , and 1 P ( i | x ˜ j , y ) represents the probability that x ˜ j , y is not classified into class i .

3.3. Reconstruction Task

To obtain a suitable initialization for the downstream segmentation task, this paper formulates a reconstruction pretext task using the encoder-decoder structure. The loss function is defined as:
rec   = 1 m I D S ( E ( x i ) ) x i 2
E ( ) is the encoder, S ( ) is the decoder, and 2 is the L 2 regular.

3.4. Loss Function

In summary, we combine the encoder rotation prediction loss, the multi-view discriminative loss, and the encoder-decoder auxiliary reconstruction loss, taking their weighted sum as the final total loss function, and jointly optimize the model. The final loss function is:
= λ 1 d + λ 2 r + λ 3 rec
where λ 1 ,   λ 2 , and   λ 3 are the weights, which represent the importance ratio of each task. In the experiment, we set λ 1 = 1.5 ,   λ 2 = 1 , and   λ 3 = 3 .

3.5. Encoder-Decoder Architecture

We take U-Net as the basic encoder-decoder network for self-supervised pre-training. As shown in Figure 2, the encoder is mainly composed of convolutional, activation, and pooling layers. The encoder extracts the effective features of the input image and gradually reduces the spatial dimension of the input data, using max-pooling after the output of the last convolutional layer in the encoder. Then, the features are flattened into a vector and passed through the FC layer, BN, and ReLU, in turn, to reduce the feature dimension to 256.After that, we split and learn the rotation prediction task and the multi-view discrimination task. A layer fully connected to output channel 4 is applied to generate probabilities for each rotation type, while a normalization layer is used to compute the cosine similarity between the features. The decoder mainly consists of convolutional, activation, and up-sampling layers.
The decoder gradually recovers the size and spatial dimension of the image through an up-sampling layer. There is information stitching between the encoder and decoder for fusing low-level and high-level semantic features to help the decoder better recover image details. The detailed network architecture is shown in Table 1.

4. Experiments and Results

4.1. Datasets

We conducted experiments on two public COVID-19 datasets and a self-built COVID-19 dataset to evaluate the effectiveness of our method.

4.1.1. Self-Built Dataset (3D-COVID)

The 3D COVID-19 CT dataset contained 2722 3D chest CT slices from 146 patients with confirmed COVID-19 (i.e., positive RT-PCR test). We collected the 3D CT dataset from Wuhan First Hospital. The image size varies from 512 × 512 × 233 to 512 × 512 × 395 , and the spatial resolution is 0.78125 × 0.78125 × 1.25 mm. Diagnosis based on these images is a very challenging task. In the work of He et al. [5], the severity of patients was divided into severe and non-severe. In this paper, the severity is defined the same as [47] and is divided into mild, moderate, and severe, as follows: mild is defined as containing less than 3 GGO lesions less than 3 cm in size; moderate is defined as; and severe is defined as the lesion area exceeding 50% of the entire lung field, as shown in Figure 3. Therefore, the task of severity assessment was modeled as a three-category problem. Since the labeling work is very time-consuming, to verify the effectiveness of SSL, 600 slices from the dataset were annotated at the pixel level. A radiologist completed all annotations. In this paper, the image intensity values of all slices were truncated to the HU range [ 800 , 100 ] to remove irrelevant details. To improve the efficiency of network training, all slices were resized to 256 × 256 after data augmentation.

4.1.2. Lesion Segmentation Dataset (COVID19-Seg)

The second dataset was from Radiopedia (http://medicalsegmentation.com/covid19/, accessed on 1 February 2022). The first version of this dataset contained 100 axial CT images of 40 patients with COVID-19. The size of the original CT images and all ground truth masks was 512 × 512. The second version of the dataset was expanded to 829 images (from nine patients), of which 373 were labeled as COVID-19, and the rest were marked as normal. In this paper, only 373 CT images marked with COVID-19 were used in the second part. The two versions were merged for a total of 473 samples. The dataset provided three image-level lesion labels: ground-glass, consolidation, and pleural effusion, and only ground-glass and consolidation ground-truth segmentation masks were used in this paper. Image-level severity labeling was done by doctors, and 100 images were annotated in this dataset.

4.1.3. Lesion Segmentation Dataset (CC-COVID)

The dataset was a large CT dataset from the Chinese Consortium for Chest CT Image Investigation (CC-CCII), which included a total of 617,775 CT images from 4154 patients. These included 752 NCP patients, 797 common pneumonia patients, and 697 normal control patients. This paper adopted 800 CT slices of size 512 × 512 from 100 COVID-19 patients. After data augmentation, the original slices were resized to 256 × 256 as the input of the network in this paper. We randomly selected 100 CT slices labeled, and the rest were unlabeled data.

4.2. Experimental Details

The CT images from the three different datasets were combined into one dataset, denoted as X all , and the labeled datasets are denoted as X l . In the pretext task stage, all unlabeled datasets X d o w n were used, and then the labeled images from X d o w n were used to fine-tune the pre-processing and train the network for classification and segmentation tasks. To verify the effect of self-supervised learning, this paper fine-tunes with different numbers of labeled datasets X d o w n s , where s { 10 % , 30 % , 70 % ,   100 % } , represents the ratio X d o w n s X d o w n . The training, validation, and testing datasets in the fine-tuning stage were performed according to 7:2:1, respectively. The training, validation, and testing datasets do not contain images from the same patients. The number of sample distributions for the labeled datasets is shown in Table 2. In each training iteration, m images were randomly selected, and two random data augmentations were applied to the selected images to generate 2 m images. Then, each image was rotated by 0°, 90°, 180°, and 270°, thus 8 m images were generated. The final batch size was 8 m. In this experiment, m was set to 16. Data augmentation is a technique widely used in deep learning. As shown in Figure 2, four augmentation operations were used in this paper: random cropping, color dithering, random erasing, and binarization. Random clipping was set to center clipping and adjusted to 256 × 256. Color dithering adjusted the brightness, saturation, contrast, and hue of the image. In our experiment, the range of brightness dithering was set to 0.7, saturation was set to 0.7, contrast was set to 0.4, and hue was set to 0.5. The minimum ratio of the erased area to the input image in random erasing was set to 0.1, and the maximum ratio was set to 0.2. The threshold for binarization was set to 125. The algorithm in this paper was implemented in Python and Pytorch, and the environment was Intel (R) Core (TM) i7 3.40 G CPU, NVIDIA 2080Ti graphics card, 32 G DDR4 RAM. This paper adjusted the image resolution to 256 × 256, and the Adam optimizer was used for network optimization. The learning rate was set to 0.0001, and the weight decay was set to 0.1.

4.3. Assessment of Performance

We employed accuracy, precision, recall, and F1-score to measure classification performance and Dice coefficient to measure segmentation performance. The definitions are as follows:
Accuracy = T P + T N T P + T N + F P + F N
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1   S c o r e = 2 × P r e c i s i o n × R e c a l l   P r e c i s i o n + R e c a l l  
D i c e = 2 | A B | | A | + | B |
where T P , T N , F P , and   F N refer to true positives, true negatives, false positives, and false negatives, respectively; A is the set of segmentation result pixels; and B is the set of actual dataset label pixels.

4.4. Classification Results of COVID-19 Severity

The method was compared with some advanced supervised learning methods and self-supervised learning methods. The results are shown in Table 3. For the supervised learning models, this paper was compared with VGG18 [48], ResNet50 [36], and DenseNet121 [49]; we modified the output channels of the last fully connected layer of these models to 3 and trained the models using cross-entropy loss for three-classification with all labeled data. All methods are applied to the same testing database. The paper also compared the proposed algorithm with other self-supervised methods, including Rotation prediction method [50], instance discrimination method [44], Moco V1 [51], and SimCLR [52]. For the rotation prediction method [27], this paper modified the output channel of the last fully connected layer of ResNet18 to 4 and a cross-entropy loss was used to train the network to make predictions for four rotation types. For instance, in discriminant methods [44], Moco V1 [51], and SimCLR [52], we employed the experimental configuration and open-source code in the above literature on all non-implemented labeled data. This paper also trained 1000 epochs on the joint dataset. For simplicity, a KNN classifier was performed on all self-supervised feature learning methods to evaluate the final performance of the classification. It can be seen that the proposed method in this paper was better than other state-of-the-art unsupervised feature learning methods. The proposed method outperformed other unsupervised feature learning methods by about 7.16% in accuracy. It is worth noting that compared with the supervised learning methods trained from scratch, our self-supervised method was pretrained using many unlabeled datasets and fine-tuned only using a small number of labeled datasets. Compared with the supervised learning method trained from scratch, the accuracy of our method achieved 18.85% higher than DenseNet, which had the highest accuracy among all supervised learning methods compared in the experiments. The lowest method was also 5.69% higher than the DenseNet with the highest accuracy in supervised learning, which further proves the effectiveness of the self-supervised learning method in this paper.

4.5. Segmentation Results of COVID-19 Lesions

The pre-trained encoder-decoder model is transferred to the downstream lesion segmentation task, and the labeled data set was used for fine-tuning training. The labeled dataset was divided into four proportions, 10%, 30%, 70%, and 100%, respectively. These four sets of data were used to fine-tune the encoder-decoder model to segment the lesion area in the CT images of COVID-19. In addition, each group also experimented in a supervised learning environment. The models under supervised learning adopted the classic segmentation models U-Net and U-Net++ [53] and were trained from scratch using labeled data. The experiment results are shown in Table 4. It can be seen that the greater the amount of data involved in supervised learning training, the smaller the gap between the segmentation accuracy of the proposed method and that of fully supervised learning was. When the amount of data involved in optimizing supervised learning gradually decreased, the performance of the fully supervised learning model gradually declined. The decline rate increased with the reduction of labeled training samples. Although the method used in this paper was only slightly better than other supervised learning in the case of a large amount of data, when the amount of labeled data was reduced to a certain extent, its advantage over supervised learning was getting bigger and bigger. When the amount of labeled data was 100%, as shown in Figure 4, the Dice value of this method was 5.58% higher than that of U-Net.; in 70% of cases, our method is 8.02% higher than U-Net; in 30% of cases, our method is 11.88% higher than U-Net; and in 10% of cases, our method is 16.88%. higher than U-Net. Figure 5 shows the visualization of segmentation results achieved by three different methods on different amounts of annotated data. Under-segmentation and wrong-segmentation results were produced in this case. These results further show the advantages of our method, and the comparative segmentation effects prove that fine-tuning with less data is better than other supervised learning methods.

4.6. Ablation Experiment

In this section, extensive ablation studies were conducted to demonstrate the importance of some settings in the model. Several issues were investigated: (1) The effect of data augmentation on the performance of self-supervised learning; (2) The impact of pretext tasks; (3) The impact of the amount of unlabeled data.

4.6.1. Data Augmentation Analysis

We evaluated the importance of augmentation operations by removing them from the transformation set or applying them individually to determine the effect of data augmentation. First, we performed experiments using the model without augmentation. As shown in Table 5, data augmentation had a significant impact on model performance, and a single augmentation operation was not enough to learn discriminative representations. Nonetheless, binarization alone performed much better than other operations, indicating that it enabled the encoder to extract more critical features. We believe the reason lies in that the ground-glass and solid-change gray values in CT images are similar, and more obvious global and local features can be obtained after binarization. It can be seen that the absence of any one of the four augmentation operations led to a decrease in performance and applying a composite data augmentation combination of the four augmentation operations improved the performance of self-supervised learning representations significantly.

4.6.2. Influence of Pretext Tasks

Three pretext tasks were co-trained to analyze the effect of the balance parameters ( λ 1 , λ 2 , λ 3 ) between the loss functions of each task in the model. When λ 1 = 0 , λ 2 = 0 , λ 3 = 0 , the pretext task was not trained. As the values of λ 1 , λ 2 , and   λ 3 increased, the task become more and more important in network training. The experimental settings were the same as those in Section 3.4. When λ 1 = 1 , λ 2 = 0 , only the instance discrimination task was trained. When λ 1 = 0 , λ 2 = 1 , only rotation prediction task was trained. Due to the irreplaceability of the reconstruction task, this paper set λ 3 to 3. It can be seen from Table 6 that when only the rotation prediction task was trained, we achieved very limited performance, with an accuracy of 78.42%. However, only using instance discrimination task training obtained a better result, with an accuracy of 85.67%. The result indicated that the instance discrimination task, i.e., the contrastive learning method, outperformed other pretext tasks. Through the collaborative training of these two tasks, our method can achieve better classification results. When λ 1 = 1.5 , λ 2 = 1 , the model achieved the best results with 95.49% accuracy, 93.66% precision, 86.98% recall, and 90.19% F1 score.

4.6.3. The Impact of Unlabeled Sample Data Volume

We pre-trained a self-supervised model using the 3D-COVID private dataset and fine-tuned it on the lesion segmentation datasets (COVID19-seg and CC-COVID) to analyze the impact of the amount of unlabeled data, respectively. We conducted comparative experiments with all datasets and pre-trained using only the unlabeled data of the 3D-COVID dataset. The experimental settings were consistent with Section 4.4 and Section 4.5. As can be seen from Table 7, compared with the models pre-trained with the three unlabeled datasets, the model was pre-trained only with the 3D-COVID unlabeled dataset. The features learned by the model were limited, and the classification and segmentation effects on downstream tasks were general. This also shows that if more unlabeled CT images of pneumonia are available, the performance of the proposed self-supervised method can be further improved, and the generalization ability is also stronger.

5. Discussion

COVID-19 was recognized as a pandemic by the World Health Organization on 11 March 2020. The variants it produces are still affecting human health so far. Improving the efficiency of COVID-19 identification is an urgent problem for researchers. Although deep learning has been well developed, the pneumonia image data that can be used for training requires precise human annotation, which means a lot of labor and cost. We may have found a solution to this problem. In this paper, we proposed a self-supervised learning framework, which is used for automatic severity assessment of COVID-19 and lung lesion segmentation in chest CT images. The proposed framework trains the network in a self-supervised manner without relying on a large amount of labeled data. We find and verify that the pretext tasks of decision rotation and multi-instance instance discrimination enable neural networks to learn image features label-free. The results presented in Figure 4 demonstrate the outstanding advantages of our method in the case of limited labeled data. With only 10% annotation, our Dice value is even 17.73% higher than that of U-Net. Perhaps the reason why our methods work is that they help to identify differences in the inherent structure of the lung and the characteristics of the lesions. Rotation prediction helps discover the discriminative structure of CT images by learning rotation-dependent features. Multi-view instance discrimination helps explore rotation-invariant features for COVID-19 lesion segmentation.
Experimental results on an aggregated dataset consisting of three COVID-19 CT image datasets show that our self-supervised method achieved better results than other state-of-the-art methods. We demonstrated a rotation-oriented collaborative pretext task in an AI-based system by using four performance evaluation metrics including (i) severity figure and (ii) dice curve. Figure 5 shows visual binary mask overlays, where red represents the output of the AI model and green represents the foreground (white) region. Due to the pre-training on many unlabeled datasets, there were limited annotations. In the data set scenario, the method in this paper has great advantages compared with supervised learning methods. In the conventional segmentation tasks, researchers focus on a rigorous fit of the model to ensure accurate labeling and prediction outputs, whereas our approach focuses on alleviating the reliance on a large number of markers. Compared with enhancing the model effect by pre-training with general classification datasets, the features learned by our method are more suitable for medical scenarios because our training data are all from medical images. Although our method achieves impressive results, it still suffers from some limitations. The features learned by the model are independent in 3D, and the connections between each lung slice are not counted. This may reduce the performance of the algorithm. Furthermore, it still relies on marking some very precise annotations, which still has a non-negligible amount of work. We also possibly combine learning-based algorithm with traditional segmentation algorithms to improve the performance [54,55]. Anyway, the development of artificial intelligence (AI)-based solutions specific to COVID-19 identification and severity quantification could provide a fast, efficient, and reliable alternative. This approach complements traditional medical diagnostic strategies and accelerates research in image analysis while reducing the burden on physicians.

6. Conclusions

This article proposes a self-supervised learning method for lesion segmentation and disease severity grading for COVID-19. The main idea of this algorithm is to learn visual features from many unlabeled images by developing joint rotation-oriented tasks, namely rotation prediction tasks and multi-view instance discrimination tasks. The rotation prediction task helps to discover the discriminative structure of COVID-19 CT images by learning rotation-related features, while the multi-view instance discrimination task helps to explore rotation-invariant features in COVID-19 images. These two features, namely rotation-dependent features and rotation-invariant features are obtained by co-training two pre-tasks. Experimental results on three datasets show that our method outperforms state-of-the-art SSL methods. Due to the availability of a large amount of unlabeled data, our method can exceed the baseline of pure segmentation tasks and is very close to the baseline of severity classification, showing that our method can be used in clinical potential benefits.
The future research direction of this paper is to design a more efficient excuse learning method and make full use of unlabeled data to solve the problem of scarcity of labeled medical image datasets.

Author Contributions

Data curation, Y.S. and X.L.; Investigation, J.L.; Methodology, J.L., Y.S., X.L. and J.T.; Project administration, J.L.; Software, Y.S. and X.L.; Supervision, J.L. and J.T.; Writing—original draft, J.L., Y.S. and X.L.; Writing—review & editing, Y.S. and J.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Xu, X.; Jiang, X.; Ma, C.; Du, P.; Li, X.; Lv, S.; Yu, L.; Ni, Q.; Chen, Y.; Su, J. A deep learning system to screen novel coronavirus disease 2019 pneumonia. Engineering 2020, 6, 1122–1129. [Google Scholar] [CrossRef]
  2. Rahman, T.; Al-Ishaq, F.A.; Al-Mohannadi, F.S.; Mubarak, R.S.; Al-Hitmi, M.H.; Islam, K.R.; Khandakar, A.; Hssain, A.A.; Al-Madeed, S.; Zughaier, S.M. Mortality prediction utilizing blood biomarkers to predict the severity of COVID-19 using machine learning technique. Diagnostics 2021, 11, 1582. [Google Scholar] [CrossRef]
  3. Baratella, E.; Ruaro, B.; Marrocchio, C.; Starvaggi, N.; Salton, F.; Giudici, F.; Quaia, E.; Confalonieri, M.; Cova, M.A. Interstitial lung disease at high resolution CT after SARS-CoV-2-related acute respiratory distress syndrome according to pulmonary segmental anatomy. J. Clin. Med. 2021, 10, 3985. [Google Scholar] [CrossRef]
  4. Ruaro, B.; Confalonieri, P.; Pozzan, R.; Tavano, S.; Mondini, L.; Baratella, E.; Pagnin, A.; Lerda, S.; Geri, P.; Biolo, M. Severe COVID-19 ARDS Treated by Bronchoalveolar Lavage with Diluted Exogenous Pulmonary Surfactant as Salvage Therapy: In Pursuit of the Holy Grail? J. Clin. Med. 2022, 11, 3577. [Google Scholar] [CrossRef]
  5. He, K.; Zhao, W.; Xie, X.; Ji, W.; Liu, M.; Tang, Z.; Shi, Y.; Shi, F.; Gao, Y.; Liu, J. Synergistic learning of lung lobe segmentation and hierarchical multi-instance classification for automated severity assessment of COVID-19 in CT images. Pattern Recognit. 2021, 113, 107828. [Google Scholar] [CrossRef]
  6. Qiblawey, Y.; Tahir, A.; Chowdhury, M.E.; Khandakar, A.; Kiranyaz, S.; Rahman, T.; Ibtehaz, N.; Mahmud, S.; Maadeed, S.A.; Musharavati, F. Detection and severity classification of COVID-19 in CT images using deep learning. Diagnostics 2021, 11, 893. [Google Scholar] [CrossRef]
  7. He, K.; Girshick, R.; Dollár, P. Rethinking imagenet pre-training. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 4918–4927. [Google Scholar]
  8. Li, X.; Yu, L.; Chen, H.; Fu, C.-W.; Xing, L.; Heng, P.-A. Transformation-consistent self-ensembling model for semisupervised medical image segmentation. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 523–534. [Google Scholar] [CrossRef]
  9. Madani, A.; Moradi, M.; Karargyris, A.; Syeda-Mahmood, T. Semi-supervised learning with generative adversarial networks for chest X-ray classification with ability of data domain adaptation. In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, DC, USA, 4–7 April 2018; pp. 1038–1042. [Google Scholar]
  10. Yu, L.; Wang, S.; Li, X.; Fu, C.-W.; Heng, P.-A. Uncertainty-aware self-ensembling model for semi-supervised 3D left atrium segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019; pp. 605–613. [Google Scholar]
  11. Yang, L.; Zhang, Y.; Chen, J.; Zhang, S.; Chen, D.Z. Suggestive annotation: A deep active learning framework for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Quebec City, QC, Canada, 11–13 September 2017; Springer: Cham, Switzerland, 2017; pp. 399–407. [Google Scholar]
  12. Zhai, X.; Oliver, A.; Kolesnikov, A.; Beyer, L. S4l: Self-supervised semi-supervised learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 1476–1485. [Google Scholar]
  13. Zhang, R.; Isola, P.; Efros, A.A. Colorful image colorization. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Cham, Switzerland, 2016; pp. 649–666. [Google Scholar]
  14. Noroozi, M.; Favaro, P. Unsupervised learning of visual representations by solving jigsaw puzzles. In Proceedings of European Conference on Computer Vision; Springer: Cham, Switzerland, 2016; pp. 69–84. [Google Scholar]
  15. Dai, Q.; Patil, V.; Hecker, S.; Dai, D.; Van Gool, L.; Schindler, K. Self-supervised object motion and depth estimation from video. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 1004–1005. [Google Scholar]
  16. Feng, Z.; Xu, C.; Tao, D. Self-supervised representation learning by rotation feature decoupling. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 10364–10374. [Google Scholar]
  17. Kingma, D.P.; Welling, M. Auto-encoding variational bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  18. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. arXiv 2014, arXiv:1406.2661. [Google Scholar]
  19. Jamaludin, A.; Kadir, T.; Zisserman, A. Self-supervised learning for spinal MRIs. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Springer: Berlin/Heidelberg, Germany, 2017; pp. 294–302. [Google Scholar]
  20. Bai, W.; Chen, C.; Tarroni, G.; Duan, J.; Guitton, F.; Petersen, S.E.; Guo, Y.; Matthews, P.M.; Rueckert, D. Self-supervised learning for cardiac mr image segmentation by anatomical position prediction. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019; Springer: Cham, Switzerland, 2019; pp. 541–549. [Google Scholar]
  21. Tajbakhsh, N.; Hu, Y.; Cao, J.; Yan, X.; Xiao, Y.; Lu, Y.; Liang, J.; Terzopoulos, D.; Ding, X. Surrogate supervision for medical image analysis: Effective deep learning from limited quantities of labeled data. In Proceedings of the 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019), Venice, Italy, 8–11 April 2019; pp. 1251–1255. [Google Scholar]
  22. Zhuang, X.; Li, Y.; Hu, Y.; Ma, K.; Yang, Y.; Zheng, Y. Self-supervised feature learning for 3d medical images by playing a rubik’s cube. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019; Springer: Cham, Switzerland, 2019; pp. 420–428. [Google Scholar]
  23. Zhu, J.; Li, Y.; Hu, Y.; Ma, K.; Zhou, S.K.; Zheng, Y. Rubik’s cube+: A self-supervised feature learning framework for 3d medical image analysis. Med. Image Anal. 2020, 64, 101746. [Google Scholar] [CrossRef]
  24. Cheng, G.; Zhou, P.; Han, J. Rifd-cnn: Rotation-invariant and fisher discriminative convolutional neural networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2884–2893. [Google Scholar]
  25. Cheng, G.; Zhou, P.; Han, J. Learning rotation-invariant convolutional neural networks for object detection in VHR optical remote sensing images. IEEE Trans. Geosci. Remote Sens. 2016, 54, 7405–7415. [Google Scholar] [CrossRef]
  26. Cheng, G.; Han, J.; Zhou, P.; Xu, D. Learning rotation-invariant and fisher discriminative convolutional neural networks for object detection. IEEE Trans. Image Processing 2018, 28, 265–278. [Google Scholar] [CrossRef] [PubMed]
  27. Li, X.; Hu, X.; Qi, X.; Yu, L.; Zhao, W.; Heng, P.-A.; Xing, L. Rotation-oriented Collaborative Self-supervised Learning for Retinal Disease Diagnosis. IEEE Trans. Med. Imaging 2021, 40, 2284–2294. [Google Scholar] [CrossRef] [PubMed]
  28. Aslan, M.F.; Sabanci, K.; Durdu, A.; Unlersen, M.F. COVID-19 diagnosis using state-of-the-art CNN architecture features and Bayesian Optimization. Comput. Biol. Med. 2022, 2022, 105244. [Google Scholar] [CrossRef]
  29. Aslan, M.F.; Unlersen, M.F.; Sabanci, K.; Durdu, A. CNN-based transfer learning–BiLSTM network: A novel approach for COVID-19 infection detection. Appl. Soft Comput. 2021, 98, 106912. [Google Scholar] [CrossRef]
  30. Zhao, C.; Xu, Y.; He, Z.; Tang, J.; Zhang, Y.; Han, J.; Shi, Y.; Zhou, W. Lung segmentation and automatic detection of COVID-19 using radiomic features from chest CT images. Pattern Recognit. 2021, 119, 108071. [Google Scholar] [CrossRef]
  31. Qian, X.; Fu, H.; Shi, W.; Chen, T.; Fu, Y.; Shan, F.; Xue, X. M3 Lung-Sys: A Deep Learning System for Multi-Class Lung Pneumonia Screening From CT Imaging. IEEE J. Biomed. Health Inform. 2020, 24, 3539–3550. [Google Scholar] [CrossRef]
  32. Zhang, J.; Xie, Y.; Pang, G.; Liao, Z.; Verjans, J.; Li, W.; Sun, Z.; He, J.; Li, Y.; Shen, C. Viral pneumonia screening on chest x-rays using Confidence-Aware anomaly detection. IEEE Trans. Med. Imaging 2020, 40, 879–890. [Google Scholar] [CrossRef]
  33. Liu, X.; Yuan, Q.; Gao, Y.; He, K.; Wang, S.; Tang, X.; Tang, J.; Shen, D. Weakly Supervised Segmentation of COVID19 Infection with Scribble Annotation on CT Images. Pattern Recognit. 2021, 122, 108341. [Google Scholar] [CrossRef]
  34. Wu, J.; Xu, H.; Zhang, S.; Li, X.; Chen, J.; Zheng, J.; Gao, Y.; Tian, Y.; Liang, Y.; Ji, R. Joint segmentation and detection of COVID-19 via a sequential region generation network. Pattern Recognit. 2021, 118, 108006. [Google Scholar] [CrossRef]
  35. Goncharov, M.; Pisov, M.; Shevtsov, A.; Shirokikh, B.; Kurmukov, A.; Blokhin, I.; Chernina, V.; Solovev, A.; Gombolevskiy, V.; Morozov, S. CT-based COVID-19 triage: Deep multitask learning improves joint identification and severity quantification. Med. Image Anal. 2021, 71, 102054. [Google Scholar] [CrossRef] [PubMed]
  36. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  37. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  38. Liu, X.; Yu, A.; Wei, X.; Pan, Z.; Tang, J. Multimodal MR image synthesis using gradient prior and adversarial learning. IEEE J. Sel. Top. Signal Processing 2020, 14, 1176–1188. [Google Scholar] [CrossRef]
  39. Mu, N.; Wang, H.; Zhang, Y.; Jiang, J.; Tang, J. Progressive global perception and local polishing network for lung infection segmentation of COVID-19 CT images. Pattern Recognit. 2021, 120, 108168. [Google Scholar] [CrossRef]
  40. Doersch, C.; Gupta, A.; Efros, A.A. Unsupervised visual representation learning by context prediction. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1422–1430. [Google Scholar]
  41. Larsson, G.; Maire, M.; Shakhnarovich, G. Colorization as a proxy task for visual understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6874–6883. [Google Scholar]
  42. Chen, L.; Bentley, P.; Mori, K.; Misawa, K.; Fujiwara, M.; Rueckert, D. Self-supervised learning for medical image analysis using image context restoration. Med. Image Anal. 2019, 58, 101539. [Google Scholar] [CrossRef] [PubMed]
  43. Dosovitskiy, A.; Fischer, P.; Springenberg, J.T.; Riedmiller, M.; Brox, T. Discriminative unsupervised feature learning with exemplar convolutional neural networks. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 1734–1747. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Wu, Z.; Xiong, Y.; Yu, S.X.; Lin, D. Unsupervised feature learning via non-parametric instance discrimination. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 3733–3742. [Google Scholar]
  45. Ye, M.; Zhang, X.; Yuen, P.C.; Chang, S.-F. Unsupervised embedding learning via invariant and spreading instance feature. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 6210–6219. [Google Scholar]
  46. Chen, X.; Yao, L.; Zhou, T.; Dong, J.; Zhang, Y. Momentum contrastive learning for few-shot COVID-19 diagnosis from chest CT images. Pattern Recognit. 2021, 113, 107826. [Google Scholar] [CrossRef]
  47. Zhang, K.; Liu, X.; Shen, J.; Li, Z.; Sang, Y.; Wu, X.; Zha, Y.; Liang, W.; Wang, C.; Wang, K. Clinically applicable AI system for accurate diagnosis, quantitative measurements, and prognosis of COVID-19 pneumonia using computed tomography. Cell 2020, 181, 1423–1433.e11. [Google Scholar] [CrossRef] [PubMed]
  48. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  49. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  50. Gidaris, S.; Singh, P.; Komodakis, N. Unsupervised representation learning by predicting image rotations. arXiv 2018, arXiv:1803.07728. [Google Scholar]
  51. He, K.; Fan, H.; Wu, Y.; Xie, S.; Girshick, R. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 9729–9738. [Google Scholar]
  52. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the International Conference on Machine Learning, Vienna, Austria, 12–18 July 2020; pp. 1597–1607. [Google Scholar]
  53. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. Unet++: A nested u-net architecture for medical image segmentation. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Springer: Berlin/Heidelberg, Germany, 2018; pp. 3–11. [Google Scholar]
  54. Tang, J.; Millington, S.; Acton, S.T.; Crandall, J.; Hurwitz, S. Ankle Cartilage Surface Segmentation using directional Gradient Vector Flow Snakes. In Proceedings of the IEEE International Conference on Image Processing (ICIP 2004), Singapore, 24–27 October 2004; pp. 2745–2748. [Google Scholar]
  55. Tang, J.; Guo, S.; Sun, Q.; Deng, Y.; Zhou, D. Speckle reducing bilateral filter for cattle follicle segmentation. BMC Genom. 2010, 11, 1–9. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Example of blood vessel orientation and rotation in CT images of COVID-19 patients.
Figure 1. Example of blood vessel orientation and rotation in CT images of COVID-19 patients.
Diagnostics 12 01805 g001
Figure 2. The framework of the proposed method.
Figure 2. The framework of the proposed method.
Diagnostics 12 01805 g002
Figure 3. Three examples of CT images of different severity.
Figure 3. Three examples of CT images of different severity.
Diagnostics 12 01805 g003
Figure 4. Comparison of lesion segmentation results under different labeled data volumes.
Figure 4. Comparison of lesion segmentation results under different labeled data volumes.
Diagnostics 12 01805 g004
Figure 5. Comparison of Lesion Segmentation Results by Different Methods.
Figure 5. Comparison of Lesion Segmentation Results by Different Methods.
Diagnostics 12 01805 g005
Table 1. Encoder-Decoder Network Architecture.
Table 1. Encoder-Decoder Network Architecture.
Module NameNumFloorInputThe Amount of Data
Encoder block 12{conv, batchnorm, ReLU}2D CT slices34 K
pooling layer 11max-poolingEncoder block 1-
Encoder block 22{conv, batchnorm, ReLU}pooling layer 168 K
pooling layer 21max-poolingEncoder block 2-
Encoder block 32{conv, batchnorm, ReLU}pooling layer 268 K
pooling layer 31max-poolingEncoder block 3-
Encoder block 42{conv, batchnorm, ReLU}pooling layer 368 K
pooling layer 41max-poolingEncoder block 4-
Encoder block 52{conv, batchnorm, ReLU}pooling layer 42465 K
Decoder-block 41{up-sample, conv, batchnorm, ReLU, concat}Encoder block 5358 K
2{conv, batchnorm, ReLU}Decoder-block 4
Decoder-block 31{up-sample, conv, batchnorm, ReLU, concat}Decoder-block 4137 K
2{conv, batchnorm, ReLU}Decoder-block 3
Decoder-block 21{up-sample, conv, batchnorm, ReLU, concat}Decoder-block 3137 K
2{conv, batchnorm, ReLU}Decoder-block 2
Decoder-block 11{up-sample, conv, batchnorm, ReLU, concat}Decoder-block 2137 K
1{conv, batchnorm, ReLU}Decoder-block 1
1 × 1 Conv block11 × 1 convDecoder-block 10.25 K
Table 2. Annotate the number of samples of different severity in the dataset.
Table 2. Annotate the number of samples of different severity in the dataset.
Datasets3D-COVIDCOVID19-SegCC-COVID
Slight2006047
Medium2503029
Severe1501024
Sum600100100
Table 3. Comparison of Different Methods in COVID-19 Severity Classification Experiment.
Table 3. Comparison of Different Methods in COVID-19 Severity Classification Experiment.
MethodAccuracyPrecisionRecallF1 Score
SupervisedVGG19 [48]73.2168.6063.2665.82
ResNet50 [36]75.8973.1969.5871.34
DenseNet121 [49]76.6478.5172.9575.62
Self-supervised Rotation [50]82.3378.3965.0071.06
Wu [44]84.6988.6374.3480.85
Moco V1 [51]88.3382.4970.5376.04
SimCLR [52]84.2179.1771.8875.34
Ours95.4993.6686.9890.19
Table 4. Comparison of Dice coefficients of three segmentation methods under different labeled data volumes.
Table 4. Comparison of Dice coefficients of three segmentation methods under different labeled data volumes.
LabelsMethodDice %
10%Ours74.94
U-Net58.06
U-Net++57.21
30%Ours78.32
U-Net66.44
U-Net++66.27
70%Ours82.11
U-Net74.09
U-Net++76.32
100%Ours84.91
U-Net79.33
U-Net++81.16
Table 5. Comparison of Dice Coefficients for Different Enhanced Combinations.
Table 5. Comparison of Dice Coefficients for Different Enhanced Combinations.
CropColor JitterRandom ErasingBinarizationDice
77.12
79.44
78.79
77.89
79.22
80.83
81.66
81.19
82.91
84.91
Table 6. Comparison of models under different parameters.
Table 6. Comparison of models under different parameters.
AccuracyPrecisionDiceF1 Score
λ1 = 1, λ2 = 0 85.6783.9576.5180.05
λ1 = 1, λ2 = 0.587.6984.4280.9882.66
λ1 = λ2 = 1 89.0387.1282.3784.67
λ1 = 1.5, λ2 = 1 95.4993.6686.9890.19
λ1 = 2, λ2 = 1 92.2188.5581.2384.73
λ1 = 0, λ2 = 1 78.4276.9769.3872.98
Table 7. Comparison of pre-training experiments with different data volumes.
Table 7. Comparison of pre-training experiments with different data volumes.
Pre-TrainingAccuracyPrecisionRecallF1 ScoreDice
All95.4993.6686.9890.1984.91
Dataset 183.5279.1978.3078.7475.26
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Song, Y.; Liu, J.; Liu, X.; Tang, J. COVID-19 Infection Segmentation and Severity Assessment Using a Self-Supervised Learning Approach. Diagnostics 2022, 12, 1805. https://doi.org/10.3390/diagnostics12081805

AMA Style

Song Y, Liu J, Liu X, Tang J. COVID-19 Infection Segmentation and Severity Assessment Using a Self-Supervised Learning Approach. Diagnostics. 2022; 12(8):1805. https://doi.org/10.3390/diagnostics12081805

Chicago/Turabian Style

Song, Yao, Jun Liu, Xinghua Liu, and Jinshan Tang. 2022. "COVID-19 Infection Segmentation and Severity Assessment Using a Self-Supervised Learning Approach" Diagnostics 12, no. 8: 1805. https://doi.org/10.3390/diagnostics12081805

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop