Next Article in Journal
Long Short-Term Memory Neural Networks for Modeling Dynamical Processes and Predictive Control: A Hybrid Physics-Informed Approach
Previous Article in Journal
Analysis of Musculoskeletal Biomechanics of Lower Limbs of Drivers in Pedal-Operation States
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A High-Resolution Network with Strip Attention for Retinal Vessel Segmentation

1
Research Center of Fluid Machinery Engineering & Technology, Jiangsu University, Zhenjiang 212013, China
2
Department of Mechanical Engineering, Texas Tech University, Lubbock, TX 79411, USA
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(21), 8899; https://doi.org/10.3390/s23218899
Submission received: 26 September 2023 / Revised: 27 October 2023 / Accepted: 30 October 2023 / Published: 1 November 2023
(This article belongs to the Section Biomedical Sensors)

Abstract

:
Accurate segmentation of retinal vessels is an essential prerequisite for the subsequent analysis of fundus images. Recently, a number of methods based on deep learning have been proposed and shown to demonstrate promising segmentation performance, especially U-Net and its variants. However, tiny vessels and low-contrast vessels are hard to detect due to the issues of a loss of spatial details caused by consecutive down-sample operations and inadequate fusion of multi-level features caused by vanilla skip connections. To address these issues and enhance the segmentation precision of retinal vessels, we propose a novel high-resolution network with strip attention. Instead of the U-Net-shaped architecture, the proposed network follows an HRNet-shaped architecture as the basic network, learning high-resolution representations throughout the training process. In addition, a strip attention module including a horizontal attention mechanism and a vertical attention mechanism is designed to obtain long-range dependencies in the horizontal and vertical directions by calculating the similarity between each pixel and all pixels in the same row and the same column, respectively. For effective multi-layer feature fusion, we incorporate the strip attention module into the basic network to dynamically guide adjacent hierarchical features. Experimental results on the DRIVE and STARE datasets show that the proposed method can extract more tiny vessels and low-contrast vessels compared with existing mainstream methods, achieving accuracies of 96.16% and 97.08% and sensitivities of 82.68% and 89.36%, respectively. The proposed method has the potential to aid in the analysis of fundus images.

1. Introduction

Fundus imaging is a non-invasive, reproducible, and inexpensive method that shows retinal vessels and pathology [1]. In the medical domain, the morphological changes of retinal vessels, e.g., vessel diameter, branch angles, and branch lengths, can be used as clinical indicators for the detection and diagnosis of diabetes, hypertension, atherosclerosis, and other diseases [2]. In addition, the retinal vascular tree can serve as a unique identifier for identification systems in the social security domain, due to the unique morphology of this feature in individuals [3]. Retinal vessel segmentation is the process of determining from fundus images whether each pixel is a vessel or non-vessel pixel and is the preliminary step in objectively assessing retinal vasculature and quantitatively interpreting the morphometrics. Nevertheless, manual approaches to retinal vessel segmentation by trained experts are expensive, time-consuming, and laborious, especially in screening a large number of people. Furthermore, manual segmentation cannot guarantee segmentation performance because the results often vary from expert to expert due to their subjective segmentation. Therefore, the development of an automatic and high-precision method for retinal vessel segmentation is highly demanded. However, the retinal vascular tree presents an extremely complicated morphological structure and has many tiny vessels with a width of fewer than ten pixels or even one pixel, and are, therefore, generally difficult to distinguish from the background. Similarly, owing to uneven illumination and lesion regions, the contrast between blood vessels and non-vascular structures is relatively low. Because of these problems, it remains a challenging task to accurately segment retinal vessels from fundus images, especially tiny vessels and low-contrast vessels.
In 1989, Chaudhuri et al. became the first to deal with the problem of automatically segmenting retinal vessels [4]. Following this work, many methods have been proposed for retinal vessel segmentation, spurred by developments in digital image processing technology in recent decades [5]. Early studies based on various hand-crafted features, e.g., shape [6], color [7], and edge [4], usually exhibit low accuracy and poor robustness due to the features being shallow and insufficiently expressing semantic-rich information. Recently, deep learning methods, especially deep convolutional neural networks (DCNNs), have achieved superior results for many computer vision tasks, e.g., image classification [8], object detection [9], human pose estimation [10], and semantic segmentation [11]. Compared with conventional methods, DCNNs are able to automatically learn richer representations from raw input data and demonstrate superior segmentation performance [12]. In particular, Long et al. [13] proposed a novel end-to-end and pixel-to-pixel semantic segmentation network, called FCN, which introduces the most basic framework for natural image segmentation: the encoder–decoder structure. However, unlike the large number of natural image datasets available, the number of medical image datasets is relatively small because they are difficult to collect due to patient privacy and ethical issues. In this regard, Ronneberger et al. [14] proposed U-Net, an improvement on FCN, that could be trained with only a few images and still predict precise results. U-Net is a breakthrough advancement in deep learning in the field of medical image segmentation. In addition to its encoder–decoder structure, the success of U-Net is largely attributed to the skip connection between the encoder sub-network and the decoder sub-network, which combines multi-level features at different stages. As a general rule, the low-level features of shallow layers have abundant spatial details but lack sufficient semantic information, while the high-level features of deep layers have semantic-rich information but lose spatial details. It is an intuitive method that adopts the skip connection to fuse the spatial details of the encoder sub-network and the semantic information of the decoder sub-network.
Even though U-Net and its variants have achieved state-of-the-art results on many medical image segmentation tasks including kidney segmentation, pancreas segmentation, and liver segmentation, it is still not good enough to efficiently and effectively segment retinal vessels. In general, there are two main limitations. Firstly, consecutive down-sample operations in the encoder sub-network result in the loss of spatial information of tiny vessels and vessel edge information, and the final segmentation map cannot recover this lost information through skip-connections and up-sample operations in the decoder sub-network. Clinically, tiny vessels consisting of only several pixels provide an indispensable reference for the diagnosis of diseases like neovascular diseases. Therefore, we should pay more attention to the tiny vessels than to the thick vessels. Secondly, there exists a certain semantic gap between low-level features and high-level features in fundus images, especially in low-contrast regions. The vanilla skip connection introduces too much irrelevant redundant information, harming retinal vessel segmentation performance, especially with low-contrast vessels. It is essential to intelligently enhance vessel representations while suppressing background noise.
To tackle these problems and improve the segmentation precision of retinal vessels, we propose a novel network based on HRNet [15] and the self-attention mechanism [16]. The proposed network is implemented based on an HRNet-shaped architecture and integrates our newly designed strip attention module. The main contributions in this paper are summarized as follows:
(1) The proposed network retains high-resolution representations instead of employing the encoder–decoder architecture, and accordingly, learns high-resolution representations throughout the training process;
(2) A strip attention module is proposed to adaptively and effectively fuse multi-level features. The module can calculate the similarity between each pixel and all pixels in the same row and the same column, respectively, thereby selectively emphasizing salient features useful for a specific task while suppressing redundant irrelevant information;
(3) We empirically conduct experiments on the DRIVE and STARE datasets, achieving state-of-the-art results.
The rest of this paper is organized as follows. In Section 2, we introduce a brief literature review of the related works. Section 3 explains the proposed method in detail. Section 4 describes the details of the experiments. Section 5 presents the results and analysis. Discussions and conclusions of our work are given in Section 6 and Section 7, respectively.

2. Related Works

2.1. Deep Convolutional Neural Networks

To date, many segmentation networks based on the fully convolutional network (FCN) with the encoder–decoder (high-to-low and low-to-high in series) architecture have been proposed in the field of semantic segmentation. Among them, U-Net [14] and its variants have achieved remarkable performance in medical image segmentation including retinal vessel segmentation. For instance, DUNet [17] replaced standard convolutions with deformable convolutions because of the high complexity of the structures of retinal vessels. Zhang et al. [18] introduced new edge-aware flows into U-Net to make predictive outcomes more sensitive to vessel edge information. For multi-source vessel image segmentation, Yin et al. [19] designed a deep fusion network, called DF-Net, which is composed of multi-scale fusion, feature fusion, and classifier fusion. Li et al. [20] proposed a multi-task symmetric network, called GDF-Net, which consists of three typical U-Net-shaped sub-networks consisting of a global segmentation network branch, a detail enhancement network branch, and a fusion network branch. As an alternative to the encoder–decoder architecture, Guo [21] put forward a low-to-high segmentation architecture, called CSGNet, which first obtains low-resolution representations, and then learns high-resolution representations with the help of low-resolution representations. Recently, some studies [15,22] have demonstrated that learning high-resolution representations throughout the training process can preserve spatial details of tiny vessels and vessel edge information, which is beneficial to segmenting tiny vessels and locating vessel boundaries. A representative method is HRNet, which was originally proposed for human pose estimation and used for other position-sensitive vision tasks [15]. HRNet maintains high resolutions from input data to final outcomes without the requirement of restoring high resolutions and generates semantic-rich high-resolution representations via repeatedly exchanging information from multi-resolution features. Motivated by HRNet, Lin et al. [22] proposed a novel high-resolution representation network with a multi-path scale, called MPS-Net. In MPS-Net, there are three paths with different resolutions, in which the main path maintains high resolutions throughout the entire process, while two branch paths with low-resolution representations are added to the main path in parallel. In this paper, an HRNet-shaped network is used as the basic network due to its superior performance on retinal vessel segmentation.

2.2. Self-Attention Modules

Generally speaking, humans can analyze and understand complex scenes naturally and effectively. Motivated by this observation, attention mechanisms [23,24] were introduced into deep learning in order to dynamically adjust the weight of feature maps. In particular, Vaswani et al. [16] proposed a self-attention mechanism with the aim of acquiring the long-range dependencies of timing signals, which facilitates machine translation and natural language processing. Then, Wang et al. [25] introduced the self-attention mechanism into computer vision to obtain long-range dependencies via non-local operations. Based on the self-attention mechanism, Fu et al. [26] presented DANet for scene segmentation, which includes a position-attention module to focus on the relationship in the spatial dimension and a channel-attention module to pay attention to the interdependencies in channel dimensions. However, the self-attention mechanism needs to generate a huge attention matrix, whose complexity is 𝒪 H × W × H × W , where H × W denotes the resolution of the input feature map, which seriously limits its practical applicability. Therefore, several variants of the self-attention mechanism have been proposed to reduce computational complexity. For instance, Huang et al. [27] viewed the self-attention operation as a graph convolution and utilized several sparsely connected graphs instead of the densely connected graph generated by the original self-attention mechanism. To do so, Huang et al. introduced a criss-cross attention module, whose weight is H + W 1 not H × W , reducing the computational complexity from 𝒪 H × W × H × W to 𝒪 H × W × H + W 1 . In addition, Li et al. [28] regarded the self-attention mechanism in terms of an expectation–maximization manner to obtain a much more compact set of bases, reducing the computational complexity from 𝒪 H × W × H × W to 𝒪 H × W × K , where K represents the number of the compact bases. Li et al. [29] designed a lightweight dual-direction attention block, generating the attention matrix with computational complexity of 𝒪 H × W via horizontal and vertical pooling operations. However, these existing variants are insufficient for retinal vessel segmentation, as they fail to focus on the characteristics of vessel structures. In this paper, an efficient strip attention module is proposed with better alignment with structures.

3. Methodology

3.1. Overview of the Network Architecture

As shown in Figure 1, the proposed network is implemented based on an HRNet-shaped architecture rather than the architecture of high-to-low and low-to-high in series and consists of three streams including a high-resolution main stream, a medium-resolution branch stream, and a low-resolution branch stream. In the main stream, the resolution is H × W from input data to final outcomes, and the depth (the number of channels) remains C , changed by 3 × 3 convolutional layers. To obtain a larger receptive field and richer semantic information, the two branch streams are gradually conducted into the main stream in parallel. The resolutions of the medium-resolution stream and the low-resolution stream are H / 2 × W / 2 and H / 4 × W / 4 , respectively, and the corresponding depths are 2 C and 4 C , respectively. Considering that retinal vessels vary greatly in terms of width, from one pixel to twenty pixels, multi-level fusion is repeated several times to exchange information across multi-resolution representations. What is more, to overcome the semantic gap issue of multi-level features and further enhance the fusion effect, we replace the vanilla skip connection with our proposed strip attention module.
Stacking small kernel convolutional layers can lead to a large receptive field, for example, two stacking 3 × 3 convolutional layers yield a receptive field of a standard 5 × 5 convolutional layer. Based on this characteristic, we employ deeply stacking 3 × 3 convolutional layers to model the representations of the proposed network, which significantly decreases the number of parameters and GPU memory to ease the training process. Meanwhile, we introduce batch normalization and nonlinear activation ReLU after every convolutional layer to reduce the change in the distributions of internal nodes of the proposed network and accelerate the training process by a significant margin. At the end of the proposed network, we utilize a convolutional layer and a sigmoid layer to obtain the final segmentation map.

3.2. Strip Attention Module

The original skip connection is used to combine multi-level feature maps with different scale information directly via concatenation or addition, but it is a native method that cannot efficiently and effectively merge representations. To deal with this issue, the self-attention mechanism may be a feasible solution. Taking a 2-D image as an example, the whole process of the self-attention mechanism [16,25,30] is as follows. Given a feature map X C × H × W , where C , H , and W represent the depth, height, and width, respectively, the self-attention mechanism firstly feeds the map into a linear projection and a reshape operation to generate three branches: Q , K , and V , where Q , K , V C × N , N = H × W . Then, the self-attention mechanism can be formulated as
A = Soft m a x Q K T
Z = Re s h a p e V A T
where is matrix multiplication; K T is the transpose of K ; A T is the transpose of the self-attention matrix A N × N ; and Z C × H × W is the output.
As shown in the process of the self-attention mechanism, a self-attention matrix is generated to calculate the similarity between each pixel and all pixels. As the self-attention matrix has a quadratic complexity of 𝒪( N 2 ) and practically suffers from high computational and memory costs, the application of self-attention mechanisms is limited in real-time applications. At the same time, we note that vessels are thin and elongated in structure, and as such, the self-attention matrix may not be the perfect solution. That is to say, we need to pay more attention to the similarity of vessel pixels rather than all pixels. Therefore, we introduce a strip attention module based on the self-attention mechanism, as shown in Figure 2. The strip attention module consists of a horizontal attention mechanism (see Figure 2a) and a vertical attention mechanism (see Figure 2b) to obtain long-range dependencies in the horizontal and vertical directions, respectively. To be specific, the output feature map Z 1 C × H × W of the horizontal attention mechanism and the output feature map Z 2 C × H × W of the vertical attention mechanism are both fed into a 1 × 1 convolutional layer followed by batch normalization and nonlinear activation ReLU. Finally, we perform a broadcast element-wise addition with the input feature map X C × H × W to obtain the final output feature map Z C × H × W , an approach that is similar to residual learning [31], to alleviate the notorious problem of vanishing/exploding gradients and aid the training process.
Z = X + λ 1 × Z 1 + λ 2 × Z 2
where λ 1 and λ 2 are both learnable parameters initialized as 0 to gradually assign more weight.
(a) Horizontal attention mechanism: As shown in Figure 2a, given a feature map X C × H × W , we first employ three 1 × 1 convolutional layers followed by batch normalization and nonlinear activation ReLU on X to generate three branches: Q W , K W , and V W , where Q W , K W , V W C × H × W . Then, we reshape Q W and K W to R W × C H , where C H = C × H . Next, we perform a matrix multiplication between Q W and the transpose of K W , and apply a softmax function to calculate the horizontal attention matrix A W W × W :
A W = Soft m a x Q W K W T A W i , j = e x p Q W i K W T j i = 1 W e x p Q W i K W T j
where is matrix multiplication, ● is element-wise multiplication, and A W i , j is an element of A W that measures the i t h position’s impact on the j t h position in the same row. The higher the value of A W i , j , the more similar i t h and j t h are in the same row, and vice versa. Meanwhile, we reshape V W to R C H × W . Finally, we perform a matrix multiplication between V W and the transpose of A W , and reshape the result to C × H × W , which is the output feature map of the horizontal attention mechanism.
(b) Vertical attention mechanism: As shown in Figure 2b, similar to the horizontal attention mechanism, we first feed the input feature map X into three 1 × 1 convolutional layers followed by batch normalization and the nonlinear activation ReLU to yield three branches: Q H , K H , and V H , where Q H , K H , V H C × H × W . Then, we reshape Q H and K H to R H × C W , where C W = C × W . Next, we perform a matrix multiplication between Q H and the transpose of K H , and apply a softmax function to count the vertical attention matrix A H H × H :
A H = Soft m a x Q H K H T A H i , j = e x p Q H i K H T j i = 1 W e x p Q H i K H T j
where is matrix multiplication, ● is element-wise multiplication, and A H i , j is an element of A H that measures the i t h position’s impact on the j t h position in the same column. The higher the value of A H i , j , the more similar i t h and j t h are in the same column, and vice versa. Meanwhile, we reshape V H to R C W × H . Finally, we perform a matrix multiplication between V H and the transpose of V H , and reshape the outcome to C × H × W , which is the output feature map of the vertical attention mechanism.

3.3. Loss Function

Due to the advantage of the binary cross-entropy loss L B C E , it is widely used in semantic segmentation and defined as follows:
L B C E = 1 N i = 1 N y i l o g p i + 1 y i l o g 1 p i
where y i denotes the true value of the position i t h , p i denotes the corresponding prediction value of the model, and N denotes the number of total pixels. As shown in Equation (6), the binary cross-entropy loss is the average loss of total pixels, and the contribution of each pixel is equal. We believe that the segmentation results would deviate towards the background pixels if we directly use the binary cross-entropy loss, due to the class imbalance problem that the number of the vessel pixels is small while the background pixels account for a large proportion of the entire fundus image. To alleviate the above problem, we introduce the Dice loss [32] as a regular term. The Dice loss function L D i c e is defined as follows:
L D i c e = 1 2 i = 1 N y i p i + ε i = 1 N y i + i = 1 N p i + ε
where y i denotes the true value of the position i t h , p i denotes the corresponding prediction value of the model, N denotes the number of total pixels, and ε denotes a small constant. ε is set to 1 × 10−5 in our experiments. The Dice loss function is based on the Dice coefficient, which can deal with the problem of imbalance between positive and negative samples. Therefore, we adopt a joint loss function L J o int to optimize the training process, which is calculated as follows:
L J o int = L B C E + L D i c e

4. Experiments

4.1. Datasets

The proposed method is trained and tested on the DRIVE and STARE datasets, which are public and standard datasets for retinal vessel segmentation. As shown in Table 1, as the image formats of DRIVE and STARE are different, we convert all images to PNG images to facilitate the analysis. Given that the field of view (FoV) masks are not given on STARE, we manually generate the corresponding masks as described in [17] for the consistency of experimental results.
DRIVE: The DRIVE (Digital Retinal Images for Vessel Extraction) dataset [33], obtained from a diabetic retinopathy screening program in the Netherlands, contains 40 images with a resolution of 565 × 584 , and is the most widely used public dataset to measure the performance of retinal vessel segmentation. There are 7 images with signs of mild early diabetic retinopathy and 33 images without any sign of diabetic retinopathy. We adopt the official data division, utilizing 20 images for model construction and the other 20 images for evaluation.
STARE: The STARE (Structured Analysis of the Retina) dataset [34] consists of 20 images, with half the images containing varying degrees of pathology. As pathological regions may obscure or confuse the blood vessel appearance in varying portions of the image, STARE is an extremely challenging dataset. The plane resolution of the fundus image is 700 × 605 . Since no official split between the training set and test set is provided on the STARE dataset, we employ leave-one-out cross-validation [17,19], whereby every image is iteratively tested by the segmentation model trained by the remaining 19 images.

4.2. Fundus Image Preprocessing and Patch Extraction

Considering the low contrast of fundus images, we employ some preprocessing strategies. First, we transform the fundus images from RGB to gray. Following this, image normalization, contrast-limited adaptive histogram equalization (CLAHE) [35], and gamma transformation are sequentially utilized. As shown in Figure 3, the image quality is dramatically ameliorated and the contrast between vessel pixels and non-vessel pixels is effectively enhanced to a certain extent.
Because of the small number of images in the DRIVE and STARE datasets, we utilize a random crop operation after image preprocessing for data augmentation in order to mitigate the overfitting problem in the training process. Taking computational complexity and surrounding local features into account, we empirically set the plane resolution of the patch to 48 × 48 (see Figure 4). During the training stage, 1000 patches are randomly selected from each fundus image from the DRIVE and STARE datasets, in which 10% of the patches are used as the validation set for selecting the optimal weight of the training model. Therefore, the DRIVE dataset has 18,000 patches as the training set and 2000 patches as the validation set; while the STARE dataset has 17,100 patches as the training set and 1900 patches as the validation set.

4.3. Implementation Details

The implementation of the proposed network is based on the PyTorch platform (torch == 2.0.1 + cu118), and all experiments are built with an NVIDIA RTX 2080Ti graphics card (produced by NVIDIA in Santa Clara, CA, USA). We set the batch size to 64 according to the memory of the GPU card and the size of the model. We employ the initialization method as introduced in [31] to initialize the weights of all layers. During the training stage, we utilize the Adam algorithm as an optimization method, in which the initial learning rate and the weight decay are set to 1 × 10−3 and 1 × 10−4, respectively. In addition, we halve the current learning rate if the validation loss does not decrease for 5 consecutive epochs. If the validation loss does not decrease for 20 consecutive epochs, the model stops training.
Figure 5 shows loss curves for the proposed network on the DRIVE and STARE datasets in the training process, which show that the proposed network is easy to train. As shown in Figure 5, the optimum model is selected based on the minimum validation loss.

4.4. Evaluation Criteria

To comprehensively evaluate the proposed method, we quantitatively analyze the experimental results in terms of accuracy (ACC), sensitivity (SE), specificity (SP), and F1-score (F1), which have the following formulas:
A C C = T P + T N T P + T N + F P + F N
S E = T P T P + F N
S P = T N F P + T N
F 1 = 2 T P 2 T P + F P + F N
where TP (true positive) denotes a vessel pixel with correct classification; TN (true negative) denotes a non-vessel pixel with correct classification; FP (false positive) denotes that a non-vessel pixel is incorrectly predicted to be a vessel pixel; and FN (false negative) denotes that a vessel pixel is incorrectly predicted to be a non-vessel pixel. Additionally, AUC is also employed, which denotes the area under the receiver operating characteristic curve. Note that we only calculate the pixels inside the FoV.

5. Results and Analysis

We systematically compare the segmentation performance of the proposed network with U-Net and the baseline via a series of experiments on the DRIVE and SATRE datasets. It should be noted that the baseline is obtained by removing the strip attention module in the proposed network rather than by using the original HRNet by Wang et al. [15].

5.1. Experiments on DRIVE

Table 2 shows the comparison results of U-Net, the baseline, and the proposed network in terms of ACC, SE, SP, F1, and AUC on the DRIVE dataset. As shown in Table 2, the baseline obtains higher values compared with U-Net on key indicators such as ACC, SE, and AUC, while SP and F1 of the baseline are slightly worse than U-Net. In particular, the baseline achieves a significant improvement of 6.79% on SE, from 75.37% to 82.16%. As can be seen from the third and fourth rows of Table 2, the proposed method with the help of the strip attention module surpasses the baseline in terms of all metrics. Meanwhile, the ACC, SE, SP, F1, and AUC of the proposed method are increased by 0.85%, 7.31%, 0.27%, 2.85%, and 1.09%, respectively, compared with U-Net.
We also employed ROC curves to further evaluate the capability of different models on the DRIVE dataset, as shown in Figure 6. The closer the ROC curve is to the top-left border, the more accurate a model is. We can clearly observe that the proposed network outperforms other models on the DRIVE dataset, as shown by the AUC score.
In addition to the quantitative analysis, a qualitative analysis was also performed on the DRIVE dataset. Figure 7 shows the overall segmentation results of the three models on the DRIVE dataset, which include the entire retinal vessel structures. In Figure 7, the first row of images is from a healthy person while the second row contains fundus images with lesion areas. In order to further observe the details of the segmentation results, Figure 8 is presented to show the local magnification view of Figure 7. As shown in Figure 7 and Figure 8, all three methods can accurately segment thick vessels from fundus images. However, U-Net struggles to accurately detect tiny vessels consisting of only a few pixels and low-contrast vessels, especially in lesion areas. In contrast, the baseline can detect some tiny vessels, while, with the help of the strip attention module, the proposed model is able to distinguish more tiny vessels and deal with the lesion areas.

5.2. Experiments on STARE

Table 3 gives the comparison experiment results on the STARE dataset. It can be seen from Table 3 that U-Net achieves the highest SP score with 98.67%, which means that U-Net can better classify non-vessel pixels compared with the baseline and the proposed method. As the STARE dataset contains many lesion regions with some non-vessels that are extremely similar to vessel pixels, the baseline and proposed networks incorrectly predicted these pixels as vessel pixels because they pay too much attention to vessels due to maintaining high resolutions. For retinal vessel segmentation, SE is able to more objectively evaluate the segmentation performance than SP due to the highly imbalanced ratio between vessel pixels and non-vessel pixels. As shown in Table 3, it is worth noting that the baseline achieves a significant improvement of 11.35% in SE score compared with U-Net, and the proposed method obtains a significant improvement of 4.74% in SE score compared with the baseline. As for F1, the proposed method has the best value, increasing from 80.89% of U-Net and 84.23% of the baseline to 98.47%. At the same time, the ACC and AUC of the baseline are slightly increased by 0.36% and 0.78%, respectively, compared with U-Net, while slightly decreased by 0.44% and 0.79%, respectively, compared with the proposed method.
As shown in Figure 9, we also assess the three models on the STARE dataset by using ROC curves, which show that the segmentation performance of the proposed network is superior to that of its competitors.
Figure 10 shows the overall segmentation results of the three models on the STARE dataset, in which the first row is fundus images with plaques and the second row is fundus images with hemorrhages. As shown in Figure 10, plaques in the fundus image result in obvious differences between adjacent regions, while hemorrhage pixels are extremely similar to vessel pixels. These factors make retinal vessels, especially low-contrast capillaries, hard to segment. In order to more intuitively observe the segmentation performance, Figure 11 visualizes challenging patches selected in Figure 10 for experimental analysis. We observe that U-Net produces too many false positive pixels to serve as a general benchmark for retinal vessel segmentation. In contrast, based on learning high-resolution representations, the baseline focuses more on vessel edge presentation and surpasses U-Net. Meanwhile, the proposed model has a clear background and fewer artifacts compared with the other two methods, indicating that the proposed model achieves desirable segmentation results on tiny vessels and low-contrast vessels.

5.3. Cross-Training Experiments

Compared with unsupervised methods, a major shortcoming of deep learning methods is that they are extremely reliant on the quality of training data, such that the segmentation performance may deteriorate if a well-trained model is applied to another dataset. In this section, we perform cross-training experiments (testing a model trained on one dataset on another dataset) on the DRIVE and STARE datasets to evaluate the extendibility and generalization of the proposed method. Table 4 shows the experimental results of the proposed method, other deep learning methods including those of Jin et al. [17] and Li et al. [20] and other unsupervised methods including those of Fathi et al. [36] and Zhao et al. [37]. Although the proposed method reduces the segmentation performance to a certain extent (see Table 2 and Table 3), it is still satisfactory compared with other comparison methods. When using the well-trained model on the STARE dataset to test the DRIVE dataset, the proposed model obtains the highest SE score of 81.47% among all comparison methods and achieves a significant improvement of 3.27% in SE score compared with the second-highest result. In addition, it also achieves state-of-the-art performance in terms of ACC, SP, and AUC. When we train the model on the DRIVE dataset and then test the STARE dataset, our proposed model achieves the top performance in terms of ACC, SE, SP, and AUC, with scores of 96.23%, 84.52%, 98.17%, and 97.92%, respectively. Possible explanations for the slight performance difference might be that the DRIVE dataset mainly contains tiny vessels and clearer images, while the STARE dataset mainly contains thick vessels and more pathological images.

6. Discussion

6.1. Comparison with Other Methods

Recent years have seen a renewal of interest in image segmentation based on DCNNs. In particular, U-Net has shown a promising segmentation performance on medical images. However, there are still some shortcomings with U-Net for retinal vessel segmentation, such as the loss of detail caused by consecutive down-sample operations and insufficient fusion caused by vanilla skip connections. Inspired by the architecture of HRNet and the self-attention mechanism, in this paper, we propose a novel method to deal with misleading diagnoses caused by the above problems. To show the superiority of the proposed method, we compare it with other proposed deep learning methods on the DRIVE and STARE datasets, as shown in Table 5 and Table 6. Note that [38] does not report F1 scores.
In Table 5, the best indicators of ACC, SE, SP, F1, and AUC achieved by methods reported in the literature are 96.22%, 83.61%, 98.52%, 83.15%, and 98.59%, respectively, while the performance indicators of our proposed method are 96.16%, 82.68%, 98.47%, 84.27%, and 98.64%, respectively. It is worth emphasizing that the proposed method achieves a new state-of-the-art performance in terms of F1 and AUC on the DRIVE dataset, which proves the high ability of the model to distinguish between vessels and non-vessels. To be specific, the proposed method outperforms high-resolution networks, such as [21,22,39]. Although the proposed method is slightly worse than MPS-Net [22] in terms of SE, the proposed method achieves higher scores on other metrics including ACC, SP, F1, and AUC. Meantime, the proposed method also has superior performance compared to the U-Net-shaped models with an attention mechanism detailed in references [20,38,40,41,42,43]. To be specific, the F1 score of GDF-Net [20] is over 1% lower than that of the proposed method. In addition, GDF-Net has shortcomings in the training process due to it consisting of three typical U-Net-shaped sub-networks.
As can be seen in Table 6, among all comparison methods, the best indicators of ACC, SE, SP, F1, and AUC achieve scores of 97.10%, 85.66%, 99.57%, 85.42%, and 99.10%, respectively, while the ACC, SE, SP, F1, and AUC scores obtained by the proposed method are 97.08%, 89.36%, 98.47%, 87.85%, and 99.21%, respectively. At the same time, the proposed method obtains the best results in terms of SE, F1, and AUC compared the best method among all comparison methods. In particular, it is worth noting that the proposed method achieves a significant improvement of 3.70% in SE score and 2.43% in F1 score compared with the second-highest result, indicating that the proposed method can correctly predict more vessel pixels. Although GDF-Net [20], recently proposed by Li et al., achieves the highest SP score, its scores are lower than those of the proposed method in terms of the other evaluation metrics. Meanwhile, the proposed method achieves higher scores in all indicators compared with MAGF-Net [43], which also supports the superiority of the proposed method.

6.2. Advantages

Retinal vessels contain abundant biological information and can be used as a feature for computer-aided systems. However, fundus images usually present some challenging elements such as uneven illumination, weak texture, inconsistent vascular thickness, and complicated vascular structure, which readily result in pseudo results for retinal vessel segmentation. In addition, the complicated pathological manifestations in fundus images pose a great challenge to the accurate segmentation of retinal vessels from lesion regions. In order to improve the segmentation precision of retinal vessels, a novel HRNet-shaped network with strip attention modules is proposed. Instead of the encoder–decoder architecture used by models such as U-Net, the proposed network adopts a high-resolution network as the baseline network, maintaining high-resolution representations throughout the whole process, which contributes to obtaining spatial details of tiny vessels and vessel edge information. On this basis, a strip attention module is proposed to reduce the semantic gap issue between multi-level features and enhance the feature fusion process, highlighting salient features while suppressing irrelevant information. The strip attention module consists of a horizontal attention mechanism and a vertical attention mechanism, which can capture long-range dependencies in the horizontal and vertical directions. The advantage of the strip attention module is twofold. First, the strip attention module has better alignment with the shape of retinal vessels compared to the original self-attention module. Second, unlike the original self-attention matrix of H × W weights, the attention matrix of the horizontal attention and vertical attention mechanisms has W weights and H weights, respectively. The higher the resolution of the input feature map, the lower the computation and memory costs of the strip attention module compared to the original self-attention module.
To test the segmentation performance of the proposed method on retinal vessels, especially low-contrast capillaries, we performed a series of experiments on the DRIVE and STARE datasets. Quantitative and qualitative analyses show that the proposed method can accurately and effectively segment retinal vessels from fundus images, even outperforming other existing methods to some extent. Even in lesion regions, the segmentation results of the proposed method are very close to the ground truth. Extensive cross-training experiments are used to further assess the extendibility and generalization of the proposed method. The promising segmentation performance reveals that the proposed model has the potential to be deployed on portable devices.

6.3. Limitations

Although the proposed method has achieved remarkable performance on retinal vessel segmentation, it still possesses some limitations. First, we only take the horizontal and vertical directions into account in the strip attention module, which may result in some fractured vessels, as vessels are oriented in many directions. Second, with the proposed method, the average processing time (including inference and I/O) for one fundus image from the DRIVE and STARE datasets is almost 1.86 s and 2.47 s, respectively. In addition, the process of patch extraction and prediction reconstruction is also memory intensive. In general, two main aspects need to be addressed. One is that even though we utilize random crop operation to augment the number of the DRIVE and STARE datasets, this does not actually increase the number of images in the datasets. In the follow-up study, we will explore Generative Adversarial Networks (GANs) to augment the datasets artificially. Another aspect that needs to be addressed is that the resolution of the DRIVE and STARE datasets is far from that of current fundus image acquisition systems. If the input feature maps have very high resolution, the memory and operations of the proposed network will increase dramatically, mainly because of the strip attention module. Therefore, we will continue this research by focusing on a more lightweight model based on self-attention in the future.

7. Conclusions

In this paper, we have presented a novel end-to-end and pixel-to-pixel segmentation network for retinal vessel segmentation. The proposed network is mainly implemented based on HRNet-shaped architecture with the addition of a strip attention module, which can maintain high-resolution representations through the whole process to obtain local spatial details and promote the fusion of features at different levels to obtain global semantic information. Extensive experiments on the DRIVE and STARE datasets demonstrate that the proposed method achieves promising segmentation performance over existing mainstream methods.

Author Contributions

Conceptualization, Z.Y. and Z.H.; methodology, Z.Y. and Y.L.; software, Z.Y.; validation, Z.Y. and Y.L.; formal analysis, T.J. and Z.H.; writing—original draft preparation, Z.Y.; writing—review and editing, Y.L., T.J. and Z.H.; super-vision, Z.H. and L.Z.; project administration, Z.Y. and L.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The DRIVE dataset used in the experiments can be publicly obtained from the Grand Challenge website at https://drive.grand-challenge.org/DRIVE/ (accessed on 10 September 2023). The STARE dataset used in the experiments can be publicly obtained from the following website: https://cecas.clemson.edu/~ahoover/stare/ (accessed on 10 September 2023).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liew, G.; Wang, J.J.; Mitchell, P.; Wong, T.Y. Retinal vascular imaging: A new tool in microvascular disease research. Circ. Cardiovasc. Imaging 2008, 1, 156–161. [Google Scholar] [CrossRef] [PubMed]
  2. Grogan, A.; Barclay, K.; Colville, D.; Hodgson, L.; Savige, J. Retinal small vessel dilatation in the systemic inflammatory response to surgery. Sci. Rep. 2022, 12, 13291. [Google Scholar] [CrossRef] [PubMed]
  3. Aleem, S.; Sheng, B.; Li, P.; Yang, P.; Feng, D.D. Fast and accurate retinal identification system: Using retinal blood vasculature landmarks. Trans. Ind. Inform. 2018, 15, 4099–4110. [Google Scholar] [CrossRef]
  4. Chaudhuri, S.; Chatterjee, S.; Katz, N.; Nelson, M.; Goldbaum, M. Detection of blood vessels in retinal images using two-dimensional matched filters. IEEE Trans. Med. Imaging 1989, 8, 263–269. [Google Scholar] [CrossRef]
  5. Khandouzi, A.; Ariafar, A.; Mashayekhpour, Z.; Pazira, M.; Baleghi, Y. Retinal vessel segmentation, a review of classic and deep methods. Ann. Biomed. Eng. 2022, 50, 1292–1314. [Google Scholar] [CrossRef] [PubMed]
  6. Vlachos, M.; Dermatas, E. Multi-scale retinal vessel segmentation using line tracking. Comput. Med. Imaging Graph. 2010, 34, 213–227. [Google Scholar] [CrossRef] [PubMed]
  7. Li, Q.; You, J.; Zhang, D. Vessel segmentation and width estimation in retinal images using multiscale production of matched filter responses. Expert Syst. Appl. 2012, 39, 7600–7610. [Google Scholar] [CrossRef]
  8. Wang, S.; Han, Y.; Chen, J.; Pan, Y.; Cao, Y.; Meng, H. Weed classification of remote sensing by UAV in ecological irrigation areas based on deep learning. J. Drain. Irrig. Mach. Eng. 2018, 36, 1137–1141. [Google Scholar]
  9. Xu, Y.; Wen, D.; Zhou, J.; Fan, X.; Liu, Y. Identification method of cotton seedlings and weeds in Xinjiang based on faster R-CNN. J. Drain. Irrig. Mach. Eng. 2021, 39, 602–607. [Google Scholar]
  10. Chen, J.; Zhan, Y. Action recognition based on multiple time scale two-stream CNN and confidence fusion in video. J. Jiangsu Univ. Nat. Sci. Ed. 2021, 42, 318–324. [Google Scholar]
  11. Fan, Y.; Shi, L.; Su, W.; Yan, H. Lane detection algorithm based on PINet + RESA network. J. Jiangsu Univ. Nat. Sci. Ed. 2023, 44, 373–378. [Google Scholar]
  12. Galdran, A.; Anjos, A.; Dolz, J.; Chakor, H.; Lombaert, H.; Ayed, I.B. State-of-the-art retinal vessel segmentation with minimalistic models. Sci. Rep. 2022, 12, 6174. [Google Scholar] [CrossRef] [PubMed]
  13. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  14. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  15. Wang, J.; Sun, K.; Cheng, T.; Jiang, B.; Deng, C.; Zhao, Y.; Liu, D.; Mu, Y.; Tan, M.; Wang, X.; et al. Deep high-resolution representation learning for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 3349–3364. [Google Scholar] [CrossRef] [PubMed]
  16. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2017; pp. 5998–6008. [Google Scholar]
  17. Jin, Q.; Meng, Z.; Pham, T.D.; Chen, Q.; Wei, L.; Su, R. DUNet: A deformable network for retinal vessel segmentation. Knowl. Based Syst. 2019, 178, 149–162. [Google Scholar] [CrossRef]
  18. Zhang, Y.; Fang, J.; Chen, Y.; Jia, L. Edge-aware U-net with gated convolution for retinal vessel segmentation. Biomed. Signal Process. Control 2022, 73, 103472. [Google Scholar] [CrossRef]
  19. Yin, P.; Cai, H.; Wu, Q. DF-Net: Deep fusion network for multi-source vessel segmentation. Inf. Fusion 2022, 78, 199–208. [Google Scholar] [CrossRef]
  20. Li, J.; Gao, G.; Yang, L.; Liu, Y. GDF-Net: A multi-task symmetrical network for retinal vessel segmentation. Biomed. Signal Process. Control 2023, 81, 104426. [Google Scholar] [CrossRef]
  21. Guo, S. CSGNet: Cascade semantic guided net for retinal vessel segmentation. Biomed. Signal Process. Control 2022, 78, 103930. [Google Scholar] [CrossRef]
  22. Lin, Z.; Huang, J.; Chen, Y.; Zhang, X.; Zhao, W.; Li, Y.; Lu, L.; Zhan, M.; Jiang, X.; Liang, X. A high resolution representation network with multi-path scale for retinal vessel segmentation. Comput. Methods Programs Biomed. 2021, 208, 106206. [Google Scholar] [CrossRef]
  23. Guo, M.; Xu, T.; Liu, J.; Liu, Z.; Jiang, P.; Mu, T.; Zhang, S.; Martin, R.R.; Cheng, M.; Hu, S. Attention mechanisms in computer vision: A survey. Comput. Vis. Media 2022, 8, 331–368. [Google Scholar] [CrossRef]
  24. Corbetta, M.; Shulman, G.L. Control of goal-directed and stimulus-driven attention in the brain. Nat. Rev. Neurosci. 2002, 3, 201–215.1. [Google Scholar] [CrossRef] [PubMed]
  25. Wang, X.; Girshick, R.; Gupta, A.; He, K. Non-local neural networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7794–7803. [Google Scholar]
  26. Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual attention network for scene segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3146–3154. [Google Scholar]
  27. Huang, Z.; Wang, X.; Huang, L.; Huang, C.; Wei, Y.; Liu, W. Ccnet: Criss-cross attention for semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 603–612. [Google Scholar]
  28. Li, X.; Zhong, Z.; Wu, J.; Yang, Y.; Lin, Z.; Liu, H. Expectation-maximization attention networks for semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9167–9176. [Google Scholar]
  29. Li, K.; Qi, X.; Luo, Y.; Yao, Z.; Zhou, X.; Sun, M. Accurate retinal vessel segmentation in color fundus images via fully attention-based networks. IEEE J. Biomed. Health Inform. 2020, 25, 2071–2081. [Google Scholar] [CrossRef] [PubMed]
  30. Wang, M.; Zhu, W.; Shi, F.; Sun, J.; Chen, H.; Yu, K.; Zhou, Y.; Peng, Y.; Chen, Z.; Chen, X. MsTGANet: Automatic drusen segmentation from retinal OCT images. IEEE Trans. Med. Imaging 2021, 41, 394–406. [Google Scholar] [CrossRef]
  31. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  32. Milletari, F.; Navab, N.; Ahmadi, S.A. V-net: Fully convolutional neural networks for volumetric medical image segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar]
  33. Staal, J.; Abràmoff, M.D.; Niemeijer, M.; Viergever, M.A.; Van Ginneken, B. Ridge-based vessel segmentation in color images of the retina. IEEE Trans. Med. Imaging 2004, 23, 501–509. [Google Scholar] [CrossRef] [PubMed]
  34. Hoover, A.D.; Kouznetsova, V.; Goldbaum, M. Locating blood vessels in retinal images by piecewise threshold probing of a matched filter response. IEEE Trans. Med. Imaging 2000, 19, 203–210. [Google Scholar] [CrossRef] [PubMed]
  35. Pizer, S.M.; Amburn, E.P.; Austin, J.D.; Cromartie, R.; Geselowitz, A.; Greer, T.; ter Haar Romeny, B.; Zimmerman, J.B.; Zuiderveld, K. Adaptive histogram equalization and its variations. Comput. Vis. Graph. Image Process. 1987, 39, 355–368. [Google Scholar] [CrossRef]
  36. Fathi, A.; Naghsh-Nilchi, A.R. Automatic wavelet-based retinal blood vessels segmentation and vessel diameter estimation. Biomed. Signal Process. Control. 2013, 8, 71–80. [Google Scholar] [CrossRef]
  37. Zhao, Y.; Zhao, J.; Yang, J.; Liu, Y.; Zhao, Y.; Zheng, Y.; Xia, L.; Wang, Y. Saliency driven vasculature segmentation with infinite perimeter active contour model. Neurocomputing 2017, 259, 201–209. [Google Scholar] [CrossRef]
  38. Li, X.; Jiang, Y.; Li, M.; Yin, S. Lightweight attention convolutional neural network for retinal vessel image segmentation. IEEE Trans. Ind. Inform. 2020, 17, 1958–1967. [Google Scholar] [CrossRef]
  39. Guo, S. Fundus image segmentation via hierarchical feature learning. Comput. Biol. Med. 2021, 138, 104928. [Google Scholar] [CrossRef]
  40. Wang, B.; Wang, S.; Qiu, S.; Wei, W.; Wang, H.; He, H. CSU-Net: A context spatial U-Net for accurate blood vessel segmentation in fundus images. IEEE J. Biomed. Health Inform. 2020, 25, 1128–1138. [Google Scholar] [CrossRef] [PubMed]
  41. Ye, Y.; Pan, C.; Wu, Y.; Wang, S.; Xia, Y. MFI-Net: Multiscale feature interaction network for retinal vessel segmentation. IEEE J. Biomed. Health Inform. 2022, 26, 4551–4562.1. [Google Scholar] [CrossRef] [PubMed]
  42. Yang, X.; Li, Z.; Guo, Y.; Zhou, D. DCU-net: A deformable convolutional neural network based on cascade U-net for retinal vessel segmentation. Multimed. Tools Appl. 2022, 81, 15593–15607. [Google Scholar] [CrossRef]
  43. Li, J.; Gao, G.; Liu, Y.; Yang, L. MAGF-Net: A multiscale attention-guided fusion network for retinal vessel segmentation. Measurement 2023, 206, 112316. [Google Scholar] [CrossRef]
  44. Alvarado-Carrillo, D.E.; Dalmau-Cedeño, O.S. Width attention based convolutional neural network for retinal vessel segmentation. Expert Syst. Appl. 2022, 209, 118313. [Google Scholar] [CrossRef]
  45. Zhang, Y.; He, M.; Chen, Z.; Hu, K.; Li, X.; Gao, X. Bridge-Net: Context-involved U-net with patch-based loss weight mapping for retinal blood Vessel segmentation. Expert Syst. Appl. 2022, 195, 116526. [Google Scholar] [CrossRef]
Figure 1. The architecture of the proposed network. (⊕ is broadcast element-wise addition).
Figure 1. The architecture of the proposed network. (⊕ is broadcast element-wise addition).
Sensors 23 08899 g001
Figure 2. Illustration of the strip attention module. (a) Horizontal attention mechanism; (b) Vertical attention mechanism. (⊗ is matrix multiplication, and ⊕ is broadcast element-wise addition).
Figure 2. Illustration of the strip attention module. (a) Horizontal attention mechanism; (b) Vertical attention mechanism. (⊗ is matrix multiplication, and ⊕ is broadcast element-wise addition).
Sensors 23 08899 g002
Figure 3. Preprocessing results. (a) Original image; (b) Grayed image; (c) Normalized image; (d) Image after CLAHE operation; (e) Image after Gamma correction.
Figure 3. Preprocessing results. (a) Original image; (b) Grayed image; (c) Normalized image; (d) Image after CLAHE operation; (e) Image after Gamma correction.
Sensors 23 08899 g003
Figure 4. Typical 48 × 48 patches. (a) Original image patches; (b) Input image patches; (c) Corresponding label patches.
Figure 4. Typical 48 × 48 patches. (a) Original image patches; (b) Input image patches; (c) Corresponding label patches.
Sensors 23 08899 g004
Figure 5. Loss curves. (a) DRIVE; (b) STARE.
Figure 5. Loss curves. (a) DRIVE; (b) STARE.
Sensors 23 08899 g005
Figure 6. ROC curves of different models on DRIVE.
Figure 6. ROC curves of different models on DRIVE.
Sensors 23 08899 g006
Figure 7. Some typical visualization results for different methods on DRIVE. (a) Original image; (b) Ground truth; (c) U-Net; (d) Baseline; (e) Proposed.
Figure 7. Some typical visualization results for different methods on DRIVE. (a) Original image; (b) Ground truth; (c) U-Net; (d) Baseline; (e) Proposed.
Sensors 23 08899 g007
Figure 8. Visualization results of details on DRIVE. (a) Original image; (b) Ground truth; (c) U-Net; (d) Baseline; (e) Proposed.
Figure 8. Visualization results of details on DRIVE. (a) Original image; (b) Ground truth; (c) U-Net; (d) Baseline; (e) Proposed.
Sensors 23 08899 g008
Figure 9. ROC curves of different models on STARE.
Figure 9. ROC curves of different models on STARE.
Sensors 23 08899 g009
Figure 10. Some typical visualization results for different methods on STARE. (a) Original image; (b) Ground truth; (c) U-Net; (d) Baseline; (e) Proposed.
Figure 10. Some typical visualization results for different methods on STARE. (a) Original image; (b) Ground truth; (c) U-Net; (d) Baseline; (e) Proposed.
Sensors 23 08899 g010
Figure 11. Visualization results of details on STARE. (a) Original image; (b) Ground truth; (c) U-Net; (d) Baseline; (e) Proposed.
Figure 11. Visualization results of details on STARE. (a) Original image; (b) Ground truth; (c) U-Net; (d) Baseline; (e) Proposed.
Sensors 23 08899 g011
Table 1. Description of datasets.
Table 1. Description of datasets.
DatasetQuantityTrain-Test SplitResolutionFormatFoV Mask
DRIVE4020–20 565 × 584 .tiff
STARE20leave-one-out 700 × 605 .ppm×
Table 2. Performance of models tested on DRIVE.
Table 2. Performance of models tested on DRIVE.
MethodACC (%)SE (%)SP (%)F1 (%)AUC (%)
U-Net95.3175.3798.2081.4297.55
Baseline95.7582.1698.1681.3698.21
Proposed96.1682.6898.4784.2798.64
Table 3. Performance of models tested on STARE.
Table 3. Performance of models tested on STARE.
MethodACC (%)SE (%)SP (%)F1 (%)AUC (%)
U-Net96.2873.2798.6780.8997.64
Baseline96.6484.6298.4884.2398.42
Proposed97.0889.3698.4787.8599.21
Table 4. Comparisons of the cross-training evaluation.
Table 4. Comparisons of the cross-training evaluation.
DatasetMethodACC (%)SE (%)SP (%)AUC (%)
DRIVE
(trained on STARE)
Jin et al. [17]94.8165.0599.1497.18
Li et al. [20]95.9372.8999.0297.15
Fathi et al. [36]95.8177.6897.5995.16
Zhao et al. [37]95.7078.2097.9088.60
Proposed95.6881.4798.0696.17
STARE
(trained on DRIVE)
Jin et al. [17]94.7470.0097.5995.71
Li et al. [20]95.8870.8997.9596.85
Fathi et al. [36]95.9180.6197.1796.80
Zhao et al. [37]95.6078.9097.8088.50
Proposed96.2384.5298.1797.92
Table 5. Performance comparison with existing methods on DRIVE.
Table 5. Performance comparison with existing methods on DRIVE.
MethodYearACC (%)SE (%)SP (%)F1 (%)AUC (%)
GDF-Net [20]202396.2282.9198.5283.0298.59
CSGNet [21]202295.7679.4398.1483.1098.23
MPS-Net [22]202195.6383.6197.4082.8798.05
Li et al. [38]202195.6879.2198.10-98.06
HHNet [39]202195.7579.9398.0683.0698.22
CSU-Net [40]202195.6580.7097.8282.5198.01
MFI-Net [41]202295.8181.7097.9083.1598.36
DCU-Net [42]202295.6881.1597.8082.7298.10
MAGF-Net [43]202395.7882.6297.8383.0798.19
Proposed202396.1682.6898.4784.2798.64
Table 6. Performance comparison with existing methods on STARE.
Table 6. Performance comparison with existing methods on STARE.
MethodYearACC (%)SE (%)SP (%)F1 (%)AUC (%)
GDF-Net [20]202396.5376.1699.5780.2298.89
CSGNet [21]202297.1083.5798.6285.4299.10
MPS-Net [22]202196.8985.6698.1984.9198.73
Li et al. [38]202196.7883.5298.23 98.75
CSU-Net [40]202197.0284.3298.4585.1698.25
MFI-Net [41]202296.8782.2098.5483.9698.97
MAGF-Net [43]202396.4980.9398.4483.6498.98
WA-Net [44]202296.6577.6798.7781.7698.65
Bridge-Net [45]202296.6880.0298.6482.8999.01
Proposed202397.0889.3698.4787.8599.21
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ye, Z.; Liu, Y.; Jing, T.; He, Z.; Zhou, L. A High-Resolution Network with Strip Attention for Retinal Vessel Segmentation. Sensors 2023, 23, 8899. https://doi.org/10.3390/s23218899

AMA Style

Ye Z, Liu Y, Jing T, He Z, Zhou L. A High-Resolution Network with Strip Attention for Retinal Vessel Segmentation. Sensors. 2023; 23(21):8899. https://doi.org/10.3390/s23218899

Chicago/Turabian Style

Ye, Zhipin, Yingqian Liu, Teng Jing, Zhaoming He, and Ling Zhou. 2023. "A High-Resolution Network with Strip Attention for Retinal Vessel Segmentation" Sensors 23, no. 21: 8899. https://doi.org/10.3390/s23218899

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop