Next Article in Journal
A Comparison of Two Forward Head Posture Corrective Approaches in Elderly with Chronic Non-Specific Neck Pain: A Randomized Controlled Study
Previous Article in Journal
Redo Surgical Aortic Valve Replacement versus Valve-In-Valve Transcatheter Aortic Valve Implantation: A Systematic Review and Reconstructed Time-To-Event Meta-Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

HMNet: Hierarchical Multi-Scale Brain Tumor Segmentation Network

1
School of Microelectronicss, Tianjin University, Tianjin 300072, China
2
School of Electrical and Information Engineering, Tianjin University, Tianjin 300072, China
3
Department of Cardiopulmonary Bypass, Chest Hospital, Tianjin University, Tianjin 300072, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
J. Clin. Med. 2023, 12(2), 538; https://doi.org/10.3390/jcm12020538
Submission received: 2 December 2022 / Revised: 30 December 2022 / Accepted: 4 January 2023 / Published: 9 January 2023
(This article belongs to the Section Clinical Guidelines)

Abstract

:
An accurate and efficient automatic brain tumor segmentation algorithm is important for clinical practice. In recent years, there has been much interest in automatic segmentation algorithms that use convolutional neural networks. In this paper, we propose a novel hierarchical multi-scale segmentation network (HMNet), which contains a high-resolution branch and parallel multi-resolution branches. The high-resolution branch can keep track of the brain tumor’s spatial details, and the multi-resolution feature exchange and fusion allow the network’s receptive fields to adapt to brain tumors of different shapes and sizes. In particular, to overcome the large computational overhead caused by expensive 3D convolution, we propose a lightweight conditional channel weighting block to reduce GPU memory and improve the efficiency of HMNet. We also propose a lightweight multi-resolution feature fusion (LMRF) module to further reduce model complexity and reduce the redundancy of the feature maps. We run tests on the BraTS 2020 dataset to determine how well the proposed network would work. The dice similarity coefficients of HMNet for ET, WT, and TC are 0.781, 0.901, and 0.823, respectively. Many comparative experiments on the BraTS 2020 dataset and other two datasets show that our proposed HMNet has achieved satisfactory performance compared with the SOTA approaches.

1. Introduction

Glioma is the most common primary brain tumor, accounting for 70% of malignant primary brain tumors in adults. The World Health Organization (WHO) classifies brain tumors into grades I–IV [1], with different survival times for different grades; the higher the grade of a brain tumor, the more aggressive it is, and the shorter the survival time of patients [2], so the early diagnosis and treatment of brain tumors are crucial. Physicians usually use magnetic resonance imaging (MRI) [3] as the main basis for the clinical diagnosis of brain tumors. The four common MRI modalities of the brain include: T1-weighted (T1), contrast-enhanced T1-weighted (T1c), T2-weighted (T2), and fluid-attenuated inversion recovery (FLAIR) [4]. However, brain tumors are characterized by different morphologies and variable locations [5], and manual segmentation requires doctors to rely on their professional knowledge and work experience to analyze the condition, so human factors such as fatigue, memory, and lack of work experience may make the diagnostic results wrong during the segmentation process. Medical imaging and computer-aided diagnosis (MICAD) technology has been made possible by the growing connection between computer science and medicine. Combining MICAD technology with computer vision to allow doctors to see the results of MRI image segmentation and use them as a reference can speed up diagnostics and make them much more accurate.
Traditional brain tumor segmentation algorithms include threshold-based, clustering-based, atlas-based, and supervision-based algorithms [6,7,8,9,10]. Although the computational complexity of brain tumor image segmentation algorithms based on traditional machine learning is low, manually designed features cannot effectively utilize the rich information of MRI brain images. Moreover, such algorithms usually need to compute many features to ensure accuracy, which slows down the computation speed and increases the memory overhead.
In recent years, convolutional neural networks (CNNs) have been successfully applied in the field of medical images [11]. Unlike traditional machine learning methods, CNNs use massive amounts of data to learn representative features automatically. This way, they do not have to go through the complex process of extracting features. CNNs have been applied successfully in medical image classification [12], detection [13], segmentation [14,15], and other image processing tasks [16]. Many studies on brain tumor segmentation have used CNN-based methods, which have greatly improved the speed and accuracy of diagnosing brain tumors. Havaei et al. [17] designed a cascaded dual-path CNN structure. Convolution kernels of different sizes were used to learn local and global features of images simultaneously; then local features and global features were fused to obtain more context information. Salehi et al. [18] proposed an auto-context-based CNN algorithm to obtain the local features and global features using an approach based on three parallel two-dimensional convolutional paths in the axial, coronal, and sagittal directions. However, these algorithms are subject to necessary post-processing as they are executed based on image blocks. Ronneberger et al. [19] proposed the encoding-decoding network U-net, which increases the number of upsampling layers compared with the traditional FCN [20] and gradually restores the image details and image resolution, and the feature map concatenation of the encoding path and the decoding path can maintain complete contextual features. U-net has become the most popular basic network in medical image processing tasks. Researchers have improved U-net and produced variants of higher accuracy [21,22]. Marcinkiewicz et al. [23] proposed an improved structure of 2D U-Net, where MR images of three modalities are input into three independent encoders. Then, the feature channels are fused and fed into the encoders. Although 2D CNN-based brain tumor segmentation methods are efficient, 3D MRI image continuity and complete 3D spatial contextual information are difficult to capture. The 3D U-net [24] network extended the U-net architecture by replacing all 2D operations with their 3D operations. Mehta et al. [25] proposed a 3D U-net brain tumor segmentation network using 3D convolution. Myronenko et al. [26,27] reconstructed the input image by adding variational auto-coding branches to regularize the shared encoder and achieved first place in the BraTS 2018 and BraTS 2019 competitions, respectively. Attention mechanisms are often introduced in brain tumor segmentation networks to make the model more focused on tumor-related regions and improve segmentation performance [28,29,30,31]. Cao et al. [15] used a novel multi-branch 3D shuffle attention module as the attention layer in the encoder, which grouped along the channel dimension and divided the feature maps into small feature maps.
Although the segmentation networks based on encoding and decoding structures can achieve satisfactory performance, continuous upsampling does not bring back image detail information from multiple downsampling, cannot bring back spatial information well, and does not pay much attention to edges. The extracted feature approach inevitably produces a fuzzy feature mapping after multiple convolutions and loses some important details, such as the boundaries of the brain tumor. Sun et al. [32] proposed a high-resolution network (HRNet). HRNet has been successfully applied in many tasks, such as human pose estimation [33] and semantic segmentation [34]. Compared with U-Net, it can maintain accurate spatial feature information due to the high-resolution branch. However, a large number of repetitive inter-fusion operations between multiple stages increase computational complexity, and intensive feature fusion computes a large amount of redundant feature information.
Although 3D convolution has good segmentation performance, it significantly increases the number of model training parameters, which results in a larger computational overhead. Nowadays, in edge computing systems, model efficiency is also an important aspect to be considered. With the rapid development of CNN technology, many lightweight networks [35,36,37] have been proposed to further reduce the model complexity and the memory occupied. Ma et al. [38] utilized three operations, channel split, depthwise separable convolutions, and channel shuffle, to enable the model to greatly reduce the computational effort of model while maintaining accuracy. Chen et al. [39] exploited the advantage of separable 3D convolutions to reduce the network complexity but with much lower segmentation accuracy. To balance network efficiency and accuracy in 3D MRI brain tumor segmentation, Chen et al. [40] proposed a new 3D expanded multi-fiber network (DMFNet) that uses efficient group convolution on top of multi-fiber units and obtains a multi-scale image representation for segmentation by introducing weighted 3D expanded convolution operations, which achieved precise segmentation while reducing the number of network parameters. However, the problem of the lack of exchange of channel information was not solved. Zhou et al. [6] proposed an efficient 3D residual neural network (ERV-Net) by first utilizing the computationally efficient network 3D ShuffleNetV2 as an encoder to reduce GPU memory and improve the segmentation efficiency of the network, and then introducing a decoder with residual blocks to avoid degradation.
The challenge of this study is to design a 3D brain tumor segmentation network with high accuracy and efficiency. The networks based on the U-net network ignore the gap problem between low-level visual features and high-level semantic features, have limited ability to reconstruct spatial information of brain tumors, and pay little attention to boundaries. The purpose of this paper is to achieve fast and accurate segmentation of brain tumors using a high-resolution-based network.
In this paper, we propose a hierarchical multi-scale brain tumor segmentation network (HMNet) to segment brain tumors with high efficiency and accuracy. Firstly, the network contains a high-resolution branch to maintain accurate spatial feature representation and multi-resolution branches to acquire multi-scale receptive fields and multi-scale features, which helps to segment brain tumors of different sizes and shapes. Secondly, to reduce the network parameters and computation to improve the efficiency of brain tumor segmentation, we exploit a 3D shuffle block [38] for feature extraction, then propose a lightweight conditional channel weighting (LCC) block to improve it. Thirdly, we propose a lightweight multi-resolution feature fusion (LMRF) module to further reduce model complexity and reduce feature redundancy. Finally, we assess our experimental results using the BraTS 2018, BraTS 2019 and BraTS 2020 datasets.
The contributions of this paper can be summarized as follows.
  • We propose a hierarchical multi-scale brain tumor segmentation network (HMNet). HMNet uses a parallel multi-resolution feature extraction network with a high-resolution branch and parallel multi-resolution branches to figure out what brain tumors look like.
  • We propose a lightweight conditional channel weighting (LCC) block based on a 3D shuffle block [38] to overcome the computational problem caused by traditional 3D convolution and enhance useful features.
  • We propose a lightweight multi-resolution fusion (LMRF) module to solve the problem that the traditional method of downsampling and upsampling has a lot of computational complexity.
The remaining sections of this article are structured as follows. Section 2 describes our methodology in depth. Section 3 contains the appropriate experimental data and analyses. Section 4 discusses the effectiveness of our proposed network. Section 5 summarizes the work of this paper.

2. Method

According to the above motivation, a hierarchical multi-scale brain tumor segmentation network (HMNet) is proposed. The architecture of HMNet is shown in Figure 1 and Table 1. HMNet includes three important components: (1) a backbone based on the parallel multi-resolution feature extraction network (PMRNet). It can maintain high-resolution feature extraction, and multi-resolution features are exchanged and fused to adapt the receptive field to brain tumors of different sizes and shapes; (2) the lightweight conditional channel weighting (LCC) block. It can overcome the large computational overhead by introducing depth-wise separable convolution. Then, we take advantage of multi-resolution and spatial information with a cross-resolution weighting unit (CWU) and spatial weighting unit (SWU) to improve network performance; and (3) the lightweight multi-resolution fusion (LMRF) module. It can overcome the large computational overhead caused by the fusion layers and reduce the redundancy of the feature maps.

2.1. Parallel Multi-Resolution Feature Extraction Network (PMRNet)

Inspired by the high-resolution network (HRNet) [32], a parallel multi-resolution feature extraction network (PMRNet) is proposed, as shown in Figure 1.
The PMRNet consists of four stages (Stage 1–Stage 4). Each new stage (except Stage 4) adds a new branch to the previous one, which has half the resolution and twice the number of channels of the previously added branch. Every stage consists of a lightweight conditional channel weighting (LCC) block to fully exploit multi-resolution features and a lightweight multi-resolution fusion (LMRF) module to aggregate rich multi-resolution contextual information. The feature maps with different resolutions are generated in parallel by PMRNet, which can maintain high-resolution feature extraction, and multi-resolution features are exchanged and fused to adapt the receptive field to brain tumors of different sizes and shapes. In this paper, we designed four subnetworks with different resolutions, 64 × 64 × 64, 32 × 32 × 32, 16 × 16 × 16, and 8 × 8 × 8 from top to bottom.
The PMRNet has three advantages. Firstly, the PMRNet can preserve the details of brain tumors due to the high-resolution branch. Next, the multi-resolution branches can provide different receptive fields for brain tumors with different sizes. Moreover, the exchange and fusion of the multi-resolution features can extract more abundant tumor features.

2.2. Lightweight Conditional Channel Weighting (LCC) Block

The 3D brain tumor segmentation task is usually time-consuming and GPU memory-consuming, so we propose an efficient unit, a lightweight conditional channel weighting (LCC) block, as shown in Figure 2 and Table 2. We extended the shuffle block [38] to the 3D version and improved it to extract features in PMRNet. First, the input channels were split into two branches by the operation called “Channel Split”. Then, the main branch ran a channel weighting unit (CWU), 3 × 3 × 3 depthwise separable convolution (DWConv), and spatial weighting unit (SWU). The CWU and SWU replace the costly traditional 1 × 1 × 1 convolution in the 3D shuffle block [38] to overcome the problem that the heavily-used 1 × 1 × 1 convolution becomes computational. The other branch was mapped as constant. The output feature maps of the two branches were merged by channel-wise concatenation, and a channel shuffle was performed on the merged feature maps. This operation takes a channel-wise concatenation approach. The input feature map was first divided into several subgroups by channel and handed over to different convolution kernels for group convolutions. Then, different subgroups were randomly extracted for channel shuffle into a new feature map so that the input feature information from different groups was fused by the next group convolution. This can enhance the exchange of information between channel groups and ensure that the input and output channels are fully correlated with each other.
As shown in Figure 3, we used a cross-resolution weighting unit (CWU) instead of the first 1 × 1 × 1 convolution block in the shuffle block [38]. Firstly, the input feature map of the parallel multi-resolution branch was subjected to the adaptive average pooling (AAP) operation to make its resolution equal to the minimum resolution of the branch and concatenated in the channel direction. Global average pooling (GAP) was performed to compress the spatial features of the input feature map, and the compressed feature maps were subjected to a one-dimensional convolution operation with k kernel size to learn the importance of different channels. The value of k varied with the number of channels (C). The relationship between them can be written as [41]:
k = | log 2 ( C ) 2 + 1 2 | .
Then, the weight maps obtained by the sigmoid operation were upsampled in order to make the resolution of the weight maps consistent with the resolution of the feature maps of their corresponding branches, so the weight of a particular position in a branch fused with the multi-resolution feature.
We also introduced a spatial weighting unit (SWU). For each resolution in PMRNet, we computed the spatial weights, which are homogeneous to spatial positions. The process is implemented as:
X GAP Conv 1 × 1 × 1 ReLU Conv 1 × 1 × 1 sigmoid W ,
where X denotes the input feature map, and W denotes the output weight map. The global average pooling (GAP) operator gathers the spatial information of brain tumors from all positions. By weighting the channels with the spatial weights, each element in the output channels receives the contribution from all the positions of all the input channels.

2.3. Lightweight Multi-Resolution Fusion (LMRF)

The fusion layer between stages in PMRNet is used for information interaction between multi-resolution branches. The traditional method is to use 3 × 3 × 3 convolutions with stride = 2 to downsample to reduce the resolution and use the nearest neighbor upsampling to increase the resolution, then use 1 × 1 × 1 convolution to implement dimensionality reduction. In order to avoid the high computational complexity problem caused by the fusion layer, we propose a lightweight multi-resolution fusion (LMRF) module. The structure of the LRMF module is shown in Table 3. We use a depthwise separable convolution (DWC) of 3 × 3 × 3 with stride = 2 to downsample. To solve the computationally intensive problem of a 1 × 1 × 1 convolution when increasing the resolution, we proposed a method that is similar to the Ghost module [42] to implement the same function as a 1 × 1 × 1 convolution. The structure is shown in Figure 4, firstly using a 1 × 1 × 1 convolution to generate a feature map with the number of channels as half of the number of output channels, and then the feature map of the required number of channels is generated by DWC and channel-wise concatenation. This method can achieve the same function as a 1 × 1 × 1 convolution and reduce feature redundancy.

3. Results

3.1. Dataset and Pre-Processing

We evaluate our HMNet on three datasets [43,44,45]: (1) the BraTS 2018 dataset. It includes a training set of 285 samples and a validation set of 66 samples. (2) The BraTS 2019 dataset. It includes a training set of 335 samples and a validation set of 125 samples. (3) The BraTS 2020 dataset. It includes a training set of 369 samples and a validation set of 125 samples. Every sample in the training set includes four MRI modality images and a ground truth label, as shown in Figure 5. The size of the MRI images is 240 × 240 × 155 mm3, with the spacings among voxels are 1 × 1 × 1 mm3. The ground truth label includes enhancing tumors (label 4), peritumor edema (label 2), and necrotic and non-enhancing tumor cores (label 1). The validation results need to be uploaded to the online platform (https://ipp.cbica.upenn.edu/, accessed on 1 December 2022) for validation. The segmentation results can be divided into three classes: ET (label 1), WT (labels 1, 2 and 4), and TC (labels 1 and 4).
To facilitate model training, we preprocessed the brain MRI images to address the noise and grayscale inhomogeneity. First, we standardized MRI brain regions to reduce the influence of intensity inhomogeneity. The strategy of data augmentation was used to expand the training set in order to prevent the overfitting problem: (1) random mirror flips in the axial, coronal, and sagittal directions with a probability of 0.5; (2) random rotations in the range of (−10°, 10°).

3.2. Evaluation Metrics

The performance of HMNet can be evaluated quantitatively using computational complexity and segmentation accuracy. The Dice similarity coefficient (Dice) and the 95% quantile of Hausdorff distance (Hausdorff95) are used to evaluate the segmentation accuracy. Dice is a measurement of the similarity between the brain tumor segmentation result and the ground truth. Hausdorff distance (HD) is used to measure the distance between the brain tumor segmentation regions and the ground truth. Dice and HD are defined as [6]:
Dice = 2 T P F P + 2 T P + F N
HD = d P , T = max sup p P inf t T d p , t , sup t T inf p P d p , t
where F P , T P , and F N denote the number of false positive, true positive, and false negative voxels, respectively. P denotes the pixel set of the predicted tumor region, and p denotes the pixel in the set. T denotes the pixel set of the ground truth, and t denotes the pixel in the set. d ( p , t ) is the distance between pixel points. s u p denotes the supremum of tumor region, and i n f denotes the infimum of tumor region. The larger the Dice is, the better the segmentation result of the network is. The smaller the Hausdorff95 is, the closer the distance between the predicted brain tumor boundary and the ground truth boundary is, and the higher the segmentation quality is. Parameters (Params) and floating point of operations (Flops) were used to evaluate the computational complexity of the network. Params represents the spatial complexity of the network, and Flops represents the time complexity of the network. They are expressed as:
Params = k d × k h × k w × C i n × C o u t
Flops = 2 × k d × k h × k w × C i n × C o u t × d × h × w
where k d , k h and k w denote the depth, height and width of the convolution kernel, respectively. C i n and C o u t denote the number of input and output channels, respectively. d, h and w denote the depth, height and width of the image, respectively.

3.3. Implementation Details

In this paper, the experimental code was carried out in Python 3.6, the server environment was Ubuntu 16.04, the CPU was Intel Core i9-9900X (3.5HGz), the graphics card was an Nvidia GTX2080Ti with 11 GB, and all the experimental networks were built using a PyTorch framework. The experimental parameters were set as follows: the network input image size was 128 × 128 × 128, the batch size was 2, and the maximum number of training epochs was 500. The Adam optimizer was used for the automatic optimization of the network, and the initial learning rate was 10 4 . During training, we divided the training set of every dataset into five parts, four of which were used as the local training set, and the other one was used as the local validation set. The code of this paper is available at https://github.com/jia-604/HMNet, accessed on 1 December 2022.

3.4. Comparison with Non-Lightweight Brain Tumor Segmentation Networks

To verify the performance of the proposed network, we conducted comparison experiments between the proposed network and current non-lightweight brain tumor segmentation networks on the BraTS 2020 dataset. The selected non-lightweight comparison networks are the 3D U-net [24], the NoNew-Net [46], the cascaded 3D densely connected U-network-based brain tumor segmentation network proposed by Ghaffari et al. [47], the Transformer-based brain tumor segmentation network proposed by Wang et al. [48], and the dResU-Net proposed by Raza et al. [49].
The results in Table 4 show that the proposed HMNet has the advantage of low model complexity compared to the non-lightweight network. Compared with the traditional 3D U-net, the number of parameters of our network was reduced to a fifth, and the amount of computation was reduced by 1540.1 G. The Dice of HMNet were 7.5%, 3.9%, and 8.6% higher than 3D U-net in ET, WT, and CT, respectively, which means that our network is more accurate than 3D U-net for brain tumor segmentation. Figure 6 shows more obviously that our network greatly improved segmentation accuracy compared to 3D U-net. It can be observed in Figure 7 that our network can extract more detailed information than 3D U-net. Table 4 shows that our network can achieve excellent segmentation results in WT and TC compared with the other non-lightweight network. Therefore, the HMNet is competitive in non-lightweight networks.

3.5. Comparison with Lightweight Brain Tumor Segmentation Networks

To verify the efficiency of our proposed network, we compared the performance of HMNet and other high-performing lightweight networks. Table 5 and Figure 7 display the experimental results. We retrained the network in Table 5 on the BraTS 2020 dataset.
Table 5 shows that the HMNet uses only about a quarter of the parameters of 3D ESP-Net, while the Dice of ET, WT, and TC are improved greatly, increasing by 9.1%, 3.0%, and 3.7%, respectively. Compared with DMF-Net, HMNet has 79.4% fewer parameters, while the Dice coefficients of WT and TC have improved by 0.2% and 0.6%, respectively. HDC-Net is currently the lightest brain tumor segmentation network, but the network is so light that it leads to poor segmentation metrics. Compared with HDC-Net, the Dice coefficients of ET, WT, and TC of the segmentation results of HMNet improved by 1.4%, 0.50%, and 2.60%, respectively, and the Hausdorff distances of ET and TC were shortened by 11.1 mm and 7.1 mm, respectively. As we can see in Figure 7, our network is more accurate than other networks in segmenting brain tumor boundaries. The above analysis shows that the proposed network is competitive in lightweight networks, and the proposed improvement direction is feasible.

3.6. Analysis about the Baseline

We designed a baseline network in which the PMRNet uses an ordinary bottleneck block in HRNet [32] for feature extraction and used ordinary upsampling and downsampling for multi-resolution fusion. In order to evaluate the effect of the number of Bottleneck blocks on the network segmentation performance, we set the number of bottleneck blocks as 1, 2, 3, and 4 in the same experimental environment, corresponding to network 1, network 2, network 3, and network 4, respectively. It can be seen from Table 6 and Figure 8 that as the number of blocks increased, the number of parameters and computation increased, but the improvement in segmentation accuracy was not significant, and even the accuracy of TC decreased. Considering both the network complexity and segmentation accuracy, we finally set the number of feature extraction blocks for each branch of PMRNet to 1.
In order to evaluate the impact of the input size of PMRNet and the Skip operation on the segmentation results, we designed a series of experiments, the results of which are shown in Table 7. The experimental results show that the PMRNet input size of 64 × 64 × 64 obtained better segmentation results, and the Skip operation improved the segmentation Dice of ET regions significantly. There was no decreasing trend in the average Dice of the three tumor regions, which proves the effectiveness of the Skip operation.

3.7. Ablation Experiments

We conducted ablation experiments to evaluate the effectiveness of each module in HMNet on the BraTS 2020 dataset. The results of the ablation experiments are shown in Table 8 and Figure 9. We first used the shuffle block to replace the ordinary bottleneck block as Network 2 in Table 8, which reduced the network parameters by 86% and the computation by 30.2 G. The Dice coefficient of the ET was reduced by 1.1%, but the Dice coefficients of the WT and CT were 0.3% and 0.6% higher, respectively. Therefore, the shuffle block can guarantee no decrease in accuracy based on greatly reducing the network complexity. To evaluate the effectiveness of the LCC block, we used the LCC block to replace the shuffle block as Network 4 in Table 8, which further reduced the parameters and computation while improving the segmentation accuracy of ET, WT, and CT by 0.9%, 0.2%, and 0.2% due to the lightweight conditional channel weighting block we designed. From the segmentation results of Network 3 and Network 6, we can obtain that our LMRF module reduced the number of network parameters by about 1M, reduced the computational effort by about 3G, and made improvements in all three brain tumor regions to some extent. The Dice coefficients of our HMNet for the three brain tumor regions were 0.781, 0.901, and 0.823, respectively. The visualization of our segmentation results obtained using the network with different methods is given in Figure 9. The results of HMNet are closer to the ground truth and have better segmentation results than the baseline, which further proves the effectiveness of our improved method in our network.

3.8. Comparative Experiments on the BraTS 2018 and BraTS 2019 Datasets

In order to evaluate the generalized performance of the HMNet, we compared the segmentation performance of the HMNet and other advanced networks on the BraTS 2018 and BraTS 2019 datasets.
The results on the BraTS 2018 dataset are shown in Table 9. The experiment results of 3D U-net, 3D ESP-Net, and DMF-Net are obtained by retraining the networks. The HMNet was more lightweight than the 3D U-net. The Dice coefficients of the HMNet for ET, WT, and TC were 0.786, 0.901, and 0.843, respectively. Compared to the 3D ESP-Net and DMF-Net, our proposed network had fewer parameters but higher segmentation accuracy. The Dice coefficients were higher than other networks. The Dice coefficients of HMNet were 0.9%, 0.6%, and 4.6% higher than the latest network proposed by Akbar et al. [28] in ET, WT, and CT, respectively, and the Hausdorff95 distances for ET, WT, and TC were shorter by 1.20 mm, 4.40 mm, and 1.02 mm. The HMNet is more competitive than other lightweight networks and non-lightweight networks.
Table 10 displays the experiment results on the BraTS 2019 dataset. The results of 3D U-Net, 3D ESP-Net, DMF-Net, and HDC-Net are obtained by retraining the networks. As we can see, the Dice coefficients of the HMNet for ET, WT and TC were 0.772, 0.899 and 0.830, respectively. Compared to 3D ESP-Net, the HMNet was lighter, and the Dice was higher. Compared to DMF-Net and HDC-Net, the Dice coefficients of the HMNet for ET and WT were about the same but higher for TC. Compared to the latest network proposed by Wang et al. [48], the Dice for ET was lower by 1.7%, but the Dice for TC was higher by 1.1%. In general, the network we proposed is more competitive than other networks.

4. Discussion

In this paper, we proposed a lightweight brain tumor segmentation network HMNet. Specifically, the shape, size and location of brain tumors vary greatly among patients, and noise is introduced during MRI scanning, resulting in blurred boundaries between tumor regions, making it difficult to segment brain tumors accurately. In addition, the networks based on 3D convolutions are computationally intensive, so these networks are not efficient enough for clinical practice. To balance segmentation accuracy and network complexity, we designed a hierarchical multi-scale brain tumor segmentation network.
We ran tests on the BraTS 2020 dataset to verify the effectiveness of the HMNet. The dice similarity coefficients of HMNet for ET, WT, and TC are 0.781, 0.901, and 0.823, respectively. Extensive experiments on the BraTS 2018, BraTS 2019, and BraTS 2020 datasets show that the proposed network has achieved satisfactory performance compared with the SOTA approaches.
Compared with the baseline network in Table 8, the parameters and flops of HMNet are significantly reduced while improving segmentation accuracy. The reason for this is that we introduce the depthwise separable convolution to replace the traditional 3 × 3 × 3 convolutions in the LCC block and LMRF module and replace the costly 1 × 1 × 1 convolution in the 3D shuffle block with CWU and SWU to overcome the problem of the heavily-used 1 × 1 × 1 convolution. HMNet has higher segmentation accuracy than the baseline network because the CWU and SWU make the network more focused on the areas associated with the tumor. In addition, the LMRF module and LCC block adopt the idea of feature reuse, where shallow feature information is directly used by deeper layers to prevent network degradation or overfitting. This can reduce the computational effort while reducing feature redundancy and improving the segmentation accuracy of the network.
Compared with other non-lightweight brain tumor segmentation networks, our proposed HMNet has better segmentation performance and lower model complexity. This is mainly because the PMRNet maintains the high-resolution features during feature propagation and ensures that the features contain accurate detailed information. However, the networks based on U-net lose some image detail information from multiple downsampling, especially at the edges of brain tumors. Moreover, our proposed HMNet is competitive with other lightweight brain tumor segmentation networks. Some lightweight networks such as DMF-Net and HDC-Net are based on U-net, so they may lose some important information. Furthermore, HDC-Net is too light to segment brain tumor boundaries accurately. Overall, our network achieves a better balance of accuracy and complexity.

5. Conclusions

In this paper, we propose a hierarchical multi-scale brain tumor segmentation network (HMNet), which consists of a high-resolution master branch with the highest resolution and three branches from high to low resolution, in which we can maintain the high resolution of the input image throughout the path based on parallel high-resolution feature extraction, thus preserving spatial details and performing more accurate tumor region segmentation for brain tumors, especially for fuzzy boundaries. Our network has the following advantages: (1) the multi-scale branches can provide different sensory fields for brain tumors of different sizes, which can effectively segment brain tumors of different shapes, sizes, and locations; (2) the network exchanges and fuses multi-scale features during feature propagation, which makes the extracted rich brain tumor feature; and (3) the introduction of a lightweight conditional channel weighting (LCC) block instead of the original 3D convolution and the use of the lightweight multi-resolution fusion (LMRF) module for feature fusion can greatly reduce the number of parameters and computation and improve computational efficiency. The effectiveness of the network is verified by a series of ablation experiments conducted on the BraTS 2020 dataset. Finally, the comparison with current state-of-the-art networks on the BraTS 2018 and BraTS 2019 datasets demonstrates the competitiveness of the HMNet.

Author Contributions

Methodology and writing-original draft preparation, S.J. and W.N.; writing-review and editing, W.N. and Q.L.; resources, S.J.; data curation, R.Z. and S.J.; supervision, W.N., M.J.A. and T.W.; project administration, W.N.; funding acquisition, Q.L. All authors have read and agreed to the published version of this manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under Grant No.61471263, the National Natural Science Foundation of China under Grant No.62272337, the Natural Science Foundation of Tianjin, China, under Grant 16JCZDJC31100, and Tianjin University Innovation Foundation under Grant 2021XZC-0024.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data will be made available on request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wu, J.; Xu, Q.; Shen, Y.; Chen, W.; Xu, K.; Qi, X.R. Swin Transformer Improves the IDH Mutation Status Prediction of Gliomas Free of MRI-Based Tumor Segmentation. J. Clin. Med. 2022, 11, 4625. [Google Scholar] [CrossRef] [PubMed]
  2. Bauer, S.; Wiest, R.; Nolte, L.P.; Reyes, M. A survey of MRI-based medical image analysis for brain tumor studies. Phys. Med. Biol. 2013, 58, R97. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Weizman, L.; Eldar, Y.C.; Ben Bashat, D. Reference-based MRI. Med. Phys. 2016, 43, 5357–5369. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Zhang, Y.; Lu, Y.; Chen, W.; Chang, Y.; Gu, H.; Yu, B. MSMANet: A multi-scale mesh aggregation network for brain tumor segmentation. Appl. Soft Comput. 2021, 110, 107733. [Google Scholar] [CrossRef]
  5. Kamnitsas, K.; Ledig, C.; Newcombe, V.F.; Simpson, J.P.; Kane, A.D.; Menon, D.K.; Rueckert, D.; Glocker, B. Efficient multi-scale 3D CNN with fully connected CRF for accurate brain lesion segmentation. Med. Image Anal. 2017, 36, 61–78. [Google Scholar] [CrossRef]
  6. Zhou, X.; Li, X.; Hu, K.; Zhang, Y.; Chen, Z.; Gao, X. ERV-Net: An efficient 3D residual neural network for brain tumor segmentation. Expert Syst. Appl. 2021, 170, 114566. [Google Scholar] [CrossRef]
  7. Balafar, M. Fuzzy C-mean based brain MRI segmentation algorithms. Artif. Intell. Rev. 2014, 41, 441–449. [Google Scholar] [CrossRef]
  8. Bach Cuadra, M.; Duay, V.; Thiran, J.P. Atlas-based segmentation. In Handbook of Biomedical Imaging; Springer: Berlin/Heidelberg, Germany, 2015; pp. 221–244. [Google Scholar]
  9. Khalid, N.E.A.; Ibrahim, S.; Haniff, P. MRI Brain Abnormalities Segmentation using K-Nearest Neighbors(k-NN). Int. J. Comput. Sci. Eng. 2011, 3, 980–990. [Google Scholar]
  10. Kharrat, A.; Gasmi, K.; Messaoud, M.B.; Benamrane, N.; Abid, M. A hybrid approach for automatic classification of brain MRI using genetic algorithm and support vector machine. Leonardo J. Sci. 2010, 17, 71–82. [Google Scholar]
  11. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; Van Der Laak, J.A.; Van Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [Green Version]
  12. Mehnatkesh, H.; Jalali, S.M.J.; Khosravi, A.; Nahavandi, S. An intelligent driven deep residual learning framework for brain tumor classification using MRI images. Expert Syst. Appl. 2022, 213, 119087. [Google Scholar] [CrossRef]
  13. Wu, Y.; Du, R.; Feng, J.; Qi, S.; Pang, H.; Xia, S.; Qian, W. Deep CNN for COPD identification by Multi-View snapshot integration of 3D airway tree and lung field. Biomed. Signal Process. Control 2023, 79, 104162. [Google Scholar] [CrossRef]
  14. Mazumdar, I.; Mukherjee, J. Fully Automatic MRI Brain Tumor Segmentation Using Efficient Spatial Attention Convolutional Networks with Composite Loss. Neurocomputing 2022, 500, 243–254. [Google Scholar] [CrossRef]
  15. Cao, Y.; Zhou, W.; Zang, M.; An, D.; Feng, Y.; Yu, B. MBANet: A 3D convolutional neural network with multi-branch attention for brain tumor segmentation from MRI images. Biomed. Signal Process. Control 2023, 80, 104296. [Google Scholar] [CrossRef]
  16. Gao, F.; Yoon, H.; Wu, T.; Chu, X. A feature transfer enabled multi-task deep learning model on medical imaging. Expert Syst. Appl. 2020, 143, 112957. [Google Scholar] [CrossRef] [Green Version]
  17. Havaei, M.; Davy, A.; Warde-Farley, D.; Biard, A.; Courville, A.; Bengio, Y.; Pal, C.; Jodoin, P.M.; Larochelle, H. Brain tumor segmentation with deep neural networks. Med. Image Anal. 2017, 35, 18–31. [Google Scholar] [CrossRef] [Green Version]
  18. Salehi, S.S.M.; Erdogmus, D.; Gholipour, A. Auto-context convolutional neural network (auto-net) for brain extraction in magnetic resonance imaging. IEEE Trans. Med. Imaging 2017, 36, 2319–2330. [Google Scholar] [CrossRef]
  19. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  20. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  21. Yang, J.; Faraji, M.; Basu, A. Robust segmentation of arterial walls in intravascular ultrasound images using Dual Path U-Net. Ultrasonics 2019, 96, 24–33. [Google Scholar] [CrossRef]
  22. Tulsani, A.; Kumar, P.; Pathan, S. Automated segmentation of optic disc and optic cup for glaucoma assessment using improved UNET++ architecture. Biocybern. Biomed. Eng. 2021, 41, 819–832. [Google Scholar] [CrossRef]
  23. Marcinkiewicz, M.; Nalepa, J.; Lorenzo, P.R.; Dudzik, W.; Mrukwa, G. Segmenting brain tumors from MRI using cascaded multi-modal U-Nets. In Proceedings of the International MICCAI Brainlesion Workshop, Grenade, Spain, 16–20 September 2018; pp. 13–24. [Google Scholar]
  24. Çiçek, Ö.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-Net: Learning dense volumetric segmentation from sparse annotation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Athens, Greece, 17–21 October 2016; pp. 424–432. [Google Scholar]
  25. Mehta, R.; Arbel, T. 3D U-Net for brain tumour segmentation. In Proceedings of the International MICCAI Brainlesion Workshop, Grenade, Spain, 16–20 September 2018; pp. 254–266. [Google Scholar]
  26. Myronenko, A. 3D MRI brain tumor segmentation using autoencoder regularization. In Proceedings of the International MICCAI Brainlesion Workshop, Grenade, Spain, 16–20 September 2018; pp. 311–320. [Google Scholar]
  27. Jiang, Z.; Ding, C.; Liu, M.; Tao, D. Two-stage cascaded u-net: 1st place solution to brats challenge 2019 segmentation task. In Proceedings of the International MICCAI Brainlesion Workshop, Shenzhen, China, 17 October 2019; pp. 231–241. [Google Scholar]
  28. Akbar, A.S.; Fatichah, C.; Suciati, N. Single level UNet3D with multipath residual attention block for brain tumor segmentation. J. King Saud-Univ.-Comput. Inf. Sci. 2022, 34, 3247–3258. [Google Scholar] [CrossRef]
  29. Chang, Y.; Zheng, Z.; Sun, Y.; Zhao, M.; Lu, Y.; Zhang, Y. Dpafnet: A residual dual-path attention-fusion convolutional neural network for multimodal brain tumor segmentation. Biomed. Signal Process. Control 2023, 79, 104037. [Google Scholar] [CrossRef]
  30. Xu, W.; Yang, H.; Zhang, M.; Cao, Z.; Pan, X.; Liu, W. Brain tumor segmentation with corner attention and high-dimensional perceptual loss. Biomed. Signal Process. Control 2022, 73, 103438. [Google Scholar] [CrossRef]
  31. Huang, Z.; Zhao, Y.; Liu, Y.; Song, G. GCAUNet: A group cross-channel attention residual UNet for slice based brain tumor segmentation. Biomed. Signal Process. Control 2021, 70, 102958. [Google Scholar] [CrossRef]
  32. Sun, K.; Xiao, B.; Liu, D.; Wang, J. Deep high-resolution representation learning for human pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5693–5703. [Google Scholar]
  33. Luo, Y.; Ou, Z.; Wan, T.; Guo, J.M. FastNet: Fast high-resolution network for human pose estimation. Image Vis. Comput. 2022, 119, 104390. [Google Scholar] [CrossRef]
  34. Wu, H.; Liang, C.; Liu, M.; Wen, Z. Optimized HRNet for image semantic segmentation. Expert Syst. Appl. 2021, 174, 114532. [Google Scholar] [CrossRef]
  35. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  36. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4510–4520. [Google Scholar]
  37. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar]
  38. Ma, N.; Zhang, X.; Zheng, H.T.; Sun, J. Shufflenet v2: Practical guidelines for efficient cnn architecture design. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 116–131. [Google Scholar]
  39. Chen, W.; Liu, B.; Peng, S.; Sun, J.; Qiao, X. S3D-UNet: Separable 3D U-Net for brain tumor segmentation. In Proceedings of the International MICCAI Brainlesion Workshop, Grenade, Spain, 16–20 September 2018; pp. 358–368. [Google Scholar]
  40. Chen, C.; Liu, X.; Ding, M.; Zheng, J.; Li, J. 3D dilated multi-fiber network for real-time brain tumor segmentation in MRI. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019; pp. 184–192. [Google Scholar]
  41. Wang, Q.; Wu, B.; Zhu, P.F.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020; pp. 11531–11539. [Google Scholar]
  42. Han, K.; Wang, Y.; Tian, Q.; Guo, J.; Xu, C.; Xu, C. Ghostnet: More features from cheap operations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 1580–1589. [Google Scholar]
  43. Menze, B.H.; Jakab, A.; Bauer, S.; Kalpathy-Cramer, J.; Farahani, K.; Kirby, J.; Burren, Y.; Porz, N.; Slotboom, J.; Wiest, R.; et al. The multimodal brain tumor image segmentation benchmark (BRATS). IEEE Trans. Med. Imaging 2014, 34, 1993–2024. [Google Scholar] [CrossRef]
  44. Bakas, S.; Akbari, H.; Sotiras, A.; Bilello, M.; Rozycki, M.; Kirby, J.S.; Freymann, J.B.; Farahani, K.; Davatzikos, C. Advancing the cancer genome atlas glioma MRI collections with expert segmentation labels and radiomic features. Sci. Data 2017, 4, 170117. [Google Scholar] [CrossRef] [Green Version]
  45. Bakas, S.; Reyes, M.; Jakab, A.; Bauer, S.; Rempfler, M.; Crimi, A.; Shinohara, R.T.; Berger, C.; Ha, S.M.; Rozycki, M.; et al. Identifying the best machine learning algorithms for brain tumor segmentation, progression assessment, and overall survival prediction in the BRATS challenge. arXiv 2018, arXiv:1811.02629. [Google Scholar]
  46. Brügger, R.; Baumgartner, C.F.; Konukoglu, E. A partially reversible U-Net for memory-efficient volumetric image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019; pp. 429–437. [Google Scholar]
  47. Ghaffari, M.; Sowmya, A.; Oliver, R. Automated brain tumour segmentation using cascaded 3d densely-connected u-net. In Proceedings of the International MICCAI Brainlesion Workshop, Lima, Peru, 4 October 2020; pp. 481–491. [Google Scholar]
  48. Wang, W.; Chen, C.; Ding, M.; Yu, H.; Zha, S.; Li, J. Transbts: Multimodal brain tumor segmentation using transformer. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Online, 27 September–1 October 2021; pp. 109–119. [Google Scholar]
  49. Raza, R.; Ijaz Bajwa, U.; Mehmood, Y.; Waqas Anwar, M.; Hassan Jamal, M. dResU-Net: 3D deep residual U-Net based brain tumor segmentation from multimodal MRI. Biomed. Signal Process. Control 2023, 79, 103861. [Google Scholar] [CrossRef]
  50. Nuechterlein, N.; Mehta, S. 3D-ESPNet with pyramidal refinement for volumetric brain tumor image segmentation. In Proceedings of the International MICCAI Brainlesion Workshop, Grenade, Spain, 16–20 September 2018; pp. 245–253. [Google Scholar]
  51. Luo, Z.; Jia, Z.; Yuan, Z.; Peng, J. HDC-Net: Hierarchical decoupled convolution network for brain tumor segmentation. IEEE J. Biomed. Health Inform. 2020, 25, 737–745. [Google Scholar] [CrossRef]
Figure 1. The architecture of our proposed HMNet. x denotes the input resolution, and C denotes the channel number of feature maps. We design a backbone network (PMRNet), which contains a high-resolution branch and parallel multi-resolution branches. The high-resolution branch can maintain the spatial details of the brain tumor. Meanwhile, multi-resolution feature exchange and fusion enables the receptive fields of the network to adapt to brain tumors with different sizes and shapes. In PMRNet, we propose a lightweight conditional channel weighting (LCC) block and a lightweight multi-resolution fusion (LMRF) module to overcome the large computational overhead caused by expensive 3D convolution and the fusion layers.
Figure 1. The architecture of our proposed HMNet. x denotes the input resolution, and C denotes the channel number of feature maps. We design a backbone network (PMRNet), which contains a high-resolution branch and parallel multi-resolution branches. The high-resolution branch can maintain the spatial details of the brain tumor. Meanwhile, multi-resolution feature exchange and fusion enables the receptive fields of the network to adapt to brain tumors with different sizes and shapes. In PMRNet, we propose a lightweight conditional channel weighting (LCC) block and a lightweight multi-resolution fusion (LMRF) module to overcome the large computational overhead caused by expensive 3D convolution and the fusion layers.
Jcm 12 00538 g001
Figure 2. The structure of LCC block. First, the input channels are split into two branches by a channel-splitting operation. The main branch runs a channel weighting unit (CWU), 3 × 3 × 3 depthwise separable convolution (DWConv), and spatial weighting unit (SWU). The other branch is mapped as constant.
Figure 2. The structure of LCC block. First, the input channels are split into two branches by a channel-splitting operation. The main branch runs a channel weighting unit (CWU), 3 × 3 × 3 depthwise separable convolution (DWConv), and spatial weighting unit (SWU). The other branch is mapped as constant.
Jcm 12 00538 g002
Figure 3. The structure of CWU. The input feature map of the parallel multi-resolution branch is subjected to the adaptive average pooling (AAP) operation to make its resolution equal to the minimum resolution of the branch and concatenated in the channel direction. Then, global average pooling (GAP) is performed to compress the spatial features of the input feature map, and the compressed feature maps are subjected to a one-dimensional convolution operation to learn the importance of different channels.
Figure 3. The structure of CWU. The input feature map of the parallel multi-resolution branch is subjected to the adaptive average pooling (AAP) operation to make its resolution equal to the minimum resolution of the branch and concatenated in the channel direction. Then, global average pooling (GAP) is performed to compress the spatial features of the input feature map, and the compressed feature maps are subjected to a one-dimensional convolution operation to learn the importance of different channels.
Jcm 12 00538 g003
Figure 4. The dimensionality reduction method in LMRF module. This method can achieve the same function as 1 × 1 × 1 convolution and reduce feature redundancy.
Figure 4. The dimensionality reduction method in LMRF module. This method can achieve the same function as 1 × 1 × 1 convolution and reduce feature redundancy.
Jcm 12 00538 g004
Figure 5. Visualization of a sample from the BraTS 2020 dataset. From left to right, there are Flair, T1, T1c and T2, respectively. The fifth image is the ground truth label. Label 1 is the necrotic and non-enhancing area in red. Label 2 is the edema area in green. Label 4 is the enhancing tumor area in yellow.
Figure 5. Visualization of a sample from the BraTS 2020 dataset. From left to right, there are Flair, T1, T1c and T2, respectively. The fifth image is the ground truth label. Label 1 is the necrotic and non-enhancing area in red. Label 2 is the edema area in green. Label 4 is the enhancing tumor area in yellow.
Jcm 12 00538 g005
Figure 6. Boxplots of the Dice and Hausdorff95 distances of our proposed HMNet and other networks on the BraTS 2020 validation set. I, II, III, and IV stand for 3D U-net, DMF-Net, HDC-Net, and HMNet, respectively.
Figure 6. Boxplots of the Dice and Hausdorff95 distances of our proposed HMNet and other networks on the BraTS 2020 validation set. I, II, III, and IV stand for 3D U-net, DMF-Net, HDC-Net, and HMNet, respectively.
Jcm 12 00538 g006
Figure 7. Visualization of segmentation results. Each row represents a different sample. (a) is the modality of Flair. (bf) are the segmentation results of 3D Unet, DMF-Net, HDC-Net, HMNet, and ground truth. The red area is necrosis and non-enhancing, the yellow area is enhancing tumor, and the green area is edema.
Figure 7. Visualization of segmentation results. Each row represents a different sample. (a) is the modality of Flair. (bf) are the segmentation results of 3D Unet, DMF-Net, HDC-Net, HMNet, and ground truth. The red area is necrosis and non-enhancing, the yellow area is enhancing tumor, and the green area is edema.
Jcm 12 00538 g007
Figure 8. The Dice coefficients of the baseline with different num_block on the BraTS 2020 validation set.
Figure 8. The Dice coefficients of the baseline with different num_block on the BraTS 2020 validation set.
Jcm 12 00538 g008
Figure 9. Visualization of segmentation results. Each row represents a different sample. (a) is the modality of Flair. (bg) are the segmentation results of Network 1 to Network 5 in Table 8 and the ground truth. The yellow area is the enhancing tumor area. The red area is the necrosis and non-enhancing area. The green area is the edema area.
Figure 9. Visualization of segmentation results. Each row represents a different sample. (a) is the modality of Flair. (bg) are the segmentation results of Network 1 to Network 5 in Table 8 and the ground truth. The yellow area is the enhancing tumor area. The red area is the necrosis and non-enhancing area. The green area is the edema area.
Jcm 12 00538 g009
Table 1. The network structure of HMNet, where Conv 3 × 3 × 3 Block stands for 3 × 3 × 3 convolution, and Group Normalization, ReLU, and Concat stand for channel-wise concatenation.
Table 1. The network structure of HMNet, where Conv 3 × 3 × 3 Block stands for 3 × 3 × 3 convolution, and Group Normalization, ReLU, and Concat stand for channel-wise concatenation.
LayerOutput SizeOperationResolution BranchOutput_Channels
Input128 × 128 × 128 4
Stem128 × 128 × 128Conv 3 × 3 × 3 Block132
64 × 64 × 64Conv 3 × 3 × 3 Block132
Stage164 × 64 × 64LCC block1, 1/232, 64
LMRF
Stage264 × 64 × 64LCC block1, 1/2, 1/432, 64, 128
LMRF
Stage364 × 64 × 64LCC block1, 1/2, 1/4, 1/832, 64, 128, 256
LMRF
Stage464 × 64 × 64LCC block1, 1/2, 1/4, 1/832, 64, 128, 256
LMRF
Skip128 × 128 × 128Element-wise sum 32
Last Stage64 × 64 × 64Upsample, Concat 480
64 × 64 × 64Conv 3 × 3 × 3 Block32
128 × 128 × 128Upsample32
128 × 128 × 128Conv 3 × 3 × 3 Block3
Table 2. The structure of LCC block. Take the first LCC block in Stage 2 as an example.
Table 2. The structure of LCC block. Take the first LCC block in Stage 2 as an example.
Input SizeInput ChannelsDetailsOutput SizeOutput Channels
64 × 64 × 6432Channel Split64 × 64 × 6416, 16
64 × 64 × 6416Adaptive average pooling32 × 32 × 3216
32 × 32 × 3264Concat32 × 32 × 3280
32 × 32 × 3216
32 × 32 × 3280Conv 1 × 11 × 1 × 180
1 × 1 × 180Dimension extension, Upsample64 × 64 × 6416
64 × 64 × 6416DWConv 3 × 3 × 364 × 64 × 6416
64 × 64 × 6416Global Average Pooling1 × 1 × 116
1 × 1 × 116Conv 1 × 1 × 1, Conv 1 × 1 × 164 × 64 × 6416
64 × 64 × 6416Sigmoid, Dot product64 × 64 × 6416
64 × 64 × 6416Concat64 × 64 × 6432
64 × 64 × 6416
64 × 64 × 6432Channel Shuffle64 × 64 × 6432
Table 3. The structure of LRMF module.
Table 3. The structure of LRMF module.
MethodInput SizeInput ChannelsDetailsOutput SizeOutput Channels
Downsample32 × 32 × 3264DWConv 3 × 3 × 316 × 16 × 16128
Upsample32 × 32 × 3264Upsample64 × 64 × 6464
64 × 64 × 6464Conv 1 × 1 × 164 × 64 × 6416
64 × 64 × 6416DWConv 3 × 3 × 364 × 64 × 6416
64 × 64 × 6416Concat64 × 64 × 6432
64 × 64 × 6416
Table 4. Comparison with other non-lightweight brain tumor segmentation networks on the BraTS 2020 validation set. A (-) denotes that the results are not reported.
Table 4. Comparison with other non-lightweight brain tumor segmentation networks on the BraTS 2020 validation set. A (-) denotes that the results are not reported.
NetworkParams/MFlops/GDiceHausdorff95/mm
ETWTTCETWTTC
3D U-net [24]16.211669.50.7060.8620.73739.77412.72019.106
Robin et al. [46]12.42296.820.7680.8910.81938.3516.3207.345
Ghaffari et al. [47]--0.7800.9000.8207.7105.1406.640
Wang et al. [48]--0.7870.9000.81717.9474.9649.769
Raza et al. [49]--0.8000.8600.822---
HMNet (ours)0.80129.40.7810.9010.82321.3405.9547.055
Table 5. Comparison with other lightweight brain tumor segmentation networks on the BraTS 2020 validation set.
Table 5. Comparison with other lightweight brain tumor segmentation networks on the BraTS 2020 validation set.
NetworkParams/MFlops/GDiceHausdorff95/mm
ETWTTCETWTTC
3D ESP-Net [50]3.3676.510.6900.8710.78631.2997.10014.617
DMF-Net [40]3.8827.040.7830.8990.81726.2555.09911.515
HDC-Net [51]0.2925.60.7670.8960.79732.4355.47614.203
HMNet (ours)0.80129.40.7810.9010.82321.3405.9547.055
Table 6. The evaluation metrics of baseline using different num_block.
Table 6. The evaluation metrics of baseline using different num_block.
NetworkNum_BlockParams/MFlops/GDice
ETWTTC
1×16.87164.40.7740.8950.804
2×27.29168.50.7680.8940.805
3×37.74173.80.7740.8920.817
4×48.17178.50.7780.8990.800
Table 7. The evaluation metrics of baseline using different methods. Here, baseline_W32 and baseline_W64 represent networks with PMRNet input size of 32 × 32 × 32 and 64 × 64 × 64, respectively, and without Skip represents the removal of the Skip operation. Avg_Dice refers to the average Dice of the three tumor regions.
Table 7. The evaluation metrics of baseline using different methods. Here, baseline_W32 and baseline_W64 represent networks with PMRNet input size of 32 × 32 × 32 and 64 × 64 × 64, respectively, and without Skip represents the removal of the Skip operation. Avg_Dice refers to the average Dice of the three tumor regions.
MethodAvg_DiceDice
ETWTTC
baseline_W32(without Skip)0.8100.7200.8960.815
baseline_W320.8120.7380.8780.820
baseline_W64(without Skip)0.8210.7480.8960.821
baseline_W640.8240.7740.8950.804
Table 8. The evaluation metrics for networks using different modules.
Table 8. The evaluation metrics for networks using different modules.
NetworkMethodParam/MFlops/GDiceHausdorff95/mm
ETWTTCETWTTC
1baseline6.87164.40.7740.8950.80421.6807.85414.138
2baseline+Shuffle0.96134.20.7630.8980.81022.9275.98913.457
3baseline+Shuffle+LMRF0.85130.30.7800.8970.82027.0716.5367.539
4baseline+LCC0.92133.30.7720.9000.81227.0616.02410.329
5HMNet(ours)0.80129.40.7810.9010.82321.3405.9547.055
Table 9. Comparison with other brain tumor segmentation networks on the BraTS 2018 validation set. Here, (-) denotes that the results are not reported.
Table 9. Comparison with other brain tumor segmentation networks on the BraTS 2018 validation set. Here, (-) denotes that the results are not reported.
NetworkParams/MFlops/GDiceHausdorff95/mm
ETWTTCETWTTC
3D U-net16.211669.50.7590.8850.7176.04017.10011.620
3D ESP-Net [50]3.3676.510.7370.8830.8145.3025.4637.853
DMF-Net [40]3.8827.040.7810.8990.8353.3854.8617.743
Akbar et al. [28]--0.7770.8950.7973.909.138.67
HMNet (ours)0.80129.40.7860.9010.8432.6994.7277.731
Table 10. Comparison with other brain tumor segmentation networks on the BraTS 2019 validation set. Here, (-) denotes that the results are not reported.
Table 10. Comparison with other brain tumor segmentation networks on the BraTS 2019 validation set. Here, (-) denotes that the results are not reported.
NetworkParams/MFlops/GDiceHausdorff95/mm
ETWTTCETWTTC
3D U-net16.211669.50.7370.8940.8076.4112.3210.44
3D ESP-Net [50]3.3676.510.6630.8710.7866.847.429.74
DMF-Net [40]3.8827.040.7760.9000.8152.994.646.22
HDC-Net [51]0.2925.60.7730.8930.8184.196.747.98
Wang et al. [48]--0.7890.9000.8193.735.646.04
HMNet (ours)0.80129.40.7720.8990.8304.015.216.57
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, R.; Jia, S.; Adamu, M.J.; Nie, W.; Li, Q.; Wu, T. HMNet: Hierarchical Multi-Scale Brain Tumor Segmentation Network. J. Clin. Med. 2023, 12, 538. https://doi.org/10.3390/jcm12020538

AMA Style

Zhang R, Jia S, Adamu MJ, Nie W, Li Q, Wu T. HMNet: Hierarchical Multi-Scale Brain Tumor Segmentation Network. Journal of Clinical Medicine. 2023; 12(2):538. https://doi.org/10.3390/jcm12020538

Chicago/Turabian Style

Zhang, Ruifeng, Shasha Jia, Mohammed Jajere Adamu, Weizhi Nie, Qiang Li, and Ting Wu. 2023. "HMNet: Hierarchical Multi-Scale Brain Tumor Segmentation Network" Journal of Clinical Medicine 12, no. 2: 538. https://doi.org/10.3390/jcm12020538

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop