Next Article in Journal
Can Perfusion-Based Brain Tissue Oxygenation MRI Support the Understanding of Cerebral Abscesses In Vivo?
Previous Article in Journal
Quantitative Study on the Breast Density and the Volume of the Mammary Gland According to the Patient’s Age and Breast Quadrant
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Segmentation of Brain Tumor Using a 3D Generative Adversarial Network

by
Behnam Kiani Kalejahi
1,
Saeed Meshgini
1,* and
Sebelan Danishvar
2,*
1
Department of Biomedical Engineering, Faculty of Electrical and Computer Engineering, University of Tabriz, Tabriz 385Q+246, Iran
2
Department of Electronic and Computer Engineering, Brunel University, London UB8 3PH, UK
*
Authors to whom correspondence should be addressed.
Diagnostics 2023, 13(21), 3344; https://doi.org/10.3390/diagnostics13213344
Submission received: 8 September 2023 / Revised: 15 October 2023 / Accepted: 16 October 2023 / Published: 30 October 2023
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
Images of brain tumors may only show up in a small subset of scans, so important details may be missed. Further, because labeling is typically a labor-intensive and time-consuming task, there are typically only a small number of medical imaging datasets available for analysis. The focus of this research is on the MRI images of the human brain, and an attempt has been made to propose a method for the accurate segmentation of these images to identify the correct location of tumors. In this study, GAN is utilized as a classification network to detect and segment of 3D MRI images. The 3D GAN network model provides dense connectivity, followed by rapid network convergence and improved information extraction. Mutual training in a generative adversarial network can bring the segmentation results closer to the labeled data to improve image segmentation. The BraTS 2021 dataset of 3D images was used to compare two experimental models.

1. Introduction

Various imaging methods in medicine are designed for specific purposes, among which the most widely used are angiography (using X-rays), computer tomography (CT) scans (using X-rays), sonography (using ultrasound waves), magnetic resonance imaging (MRI) (using radio waves and magnetic amplification), and radiology (using X-rays). Each of the mentioned imaging techniques is specific to certain tissues in the body. The MRI technique provides clinicians with crucial information on the type, size, shape, and location of brain tumors without exposing the patient to dangerous radiation. Meningioma and glioma are fatal brain tumors that can be discovered via magnetic resonance imaging studies [1]. If tumors are not diagnosed in their early stages, they can be seriously dangerous in the future. T1, T1c, T2, and FLAIR MRI sequences provide detailed information about brain tumors:
(1)
T1-weighted scans that distinguish healthy tissues from those with tumors.
(2)
T2-weighted scans to delineate the tumor area, which creates a bright image area.
(3)
T1-c scans use a contrast agent that builds up at the edge of the tumor and gives a bright signal.
(4)
The water molecule suppression signal is used in FLAIR scans.
Glioma is the most common kind of brain cancer in humans. The World Health Organization (WHO) divides tumors into four types [1]. Low-level tumors such as meningioma are classified as grade one and two cancers, whereas gliomas are classified as grade three and four cancers. Meningiomas account for around 20% of all brain tumors. This type of tumor has a spherical shape and grows at a slower rate. Even though meningioma is a low-risk tumor with a modest growth rate, it can cause considerable harm to the patient if not treated early. Since lesions are typically tiny and have varying color, shape, and texture alterations, interpreting MRI images to detect brain tumors is a time-consuming, difficult, and delicate process. Neurologists and surgeons sometimes struggle to make the right call. Noisy images and exhausted doctors can also cause misinterpretations of images. Analysis with the help of computer algorithms is one of the most promising methods for facing such problems in MRI images. In the meantime, deep learning (DL) architectures are prominent as a method and work well in this field.

1.1. Brain Tumor Segmentation

The present categorization of methods for brain tumor segmentation can be classified into many groups according to unique conceptual frameworks. The categorization of brain tumor segmentation methods is commonly based on the level of human involvement, resulting in three distinct categories: manual, semi-automated, and fully automatic [2]. Proficient manual brain tumor segmentation requires professionals to possess a comprehensive understanding of picture interpretation, brain tumor characteristics, and relevant disciplines such as anatomy. The manual segmentation of brain tumors is the process of manually delineating the boundaries of the tumor and assigning distinct labels to the various regions of the anatomical components. To date, manual segmentation has been extensively utilized in clinical trials. Due to the association of numerous images with the progression of brain tumors, professionals may encounter challenges when manually segmenting various regions of brain tumors, as this process is prone to errors, time-consuming, and yields unsatisfactory results. Consequently, the utilization of more sophisticated segmentation techniques, such as semi-automatic and fully automatic segmentation methods, can effectively address this issue. The process of semi-automated brain tumor segmentation requires the utilization of specialized knowledge, human involvement, and computational tools. In the context of brain tumor diagnosis, semi-automated procedures involve the input of parameters by an expert, the analysis of visual data, and the provision of feedback for software computations. The method category is comprised of three distinct components: initial processing, feedback, and assessment. While semi-automated algorithms for brain tumor segmentation have demonstrated superior performance compared to manual segmentation, it is important to note that discrepancies in findings can arise due to variations among various specialists and over different time points. Consequently, there has been a development of techniques aimed at achieving the fully automated segmentation of brain tumors.

1.2. Deep Learning

Deep learning is a subfield within the broader discipline of Machine Learning (ML) that is employed to model complex problems and abstract notions. Deep learning (DL) facilitates the training of models that consist of numerous layers of processing, known as deep neural networks, which have the ability to acquire abstract representations of data. The assessment of conceptual qualities as nonlinear functions of low-level features is facilitated by the multilayer nature of DL networks. Convolutional Neural Networks (CNN), Restricted Boltzmann Machines (RBM), Deep Belief Networks (DBN), Deep Auto-Encoders (DAE), Recurrent Neural Networks (RNN), and their derivatives, including Long Short-term Memory (LSTM), are considered to be very valuable Deep Neural Networks (DNNs). These networks are frequently employed to execute extensive operations for a multitude of objectives. The important difference between conventional machine learning (ML) and deep learning (DL) algorithms is due to the aspect of feature engineering. Traditional machine learning (ML) algorithms perform classification tasks by utilizing a predefined set of characteristics. In contrast, deep learning (DL) techniques possess the ability to automatically extract features, resulting in higher accuracy compared to conventional ML models. The effectiveness of these models in addressing large-scale data issues surpasses that of shallow machine learning techniques due to the increased depth of the processing layers [2]. The study applied various machine learning (ML) methodologies, such as the random forest algorithm and Support Vector Machines (SVM), to automate the process of identifying and segmenting lesions using MRI data. Deep learning methods such as Restricted Boltzmann Machines (RBM), Denoising Autoencoders (DAE), Convolutional Neural Networks (CNN), and Long Short-Term Memory (LSTM) have recently been increasingly prominent in the field of medical picture analysis. The Generative Adversarial Network (GAN) is a widely recognized and very efficient deep learning model.

1.3. GAN Network

GAN networks are generative models that generate high-quality data from a small collection of training datasets. The GAN network is made up of two components: the generator and the discriminator. The generator attempts to learn the model built from the data and, as a result, generates graphics from random noise inputs. The discriminator is a CNN network that attempts to discriminate between actual data (training data) and data generated by the generator, and it assesses the likelihood of a mistake using this approach. Two of the most well-known GAN architectures are the semi-supervised GAN and Laplacian pyramid GAN. Since GAN networks are trained using only non-lesion images, it is expected that the error probability assigned by the network to images with lesions will be significantly different from that assigned to non-lesion images. The general structure of a standard GAN network with a generator and discriminator is shown in Figure 1. In general, GANs have two uses in medical imaging. The first use is concerned with generation, which can aid in understanding the underlying structure of training data and learning to produce new pictures. This aspect of GAN networks holds great promise for coping with data shortages while protecting patient privacy. The second application emphasizes the differentiating feature. Due to the necessity for vast volumes of training data, image creation methods have proliferated in the field of deep learning. Recent research has demonstrated that GAN networks may be utilized for a variety of applications, including image-to-image translation [3] and unsupervised representation learning. GANs have also been demonstrated for unsupervised data domain pattern matching in multimodal medical imaging, indicating their potential for usage with limited medical imaging data.
This research is organized into the following sections: Section 2 of this paper presents an overview of the relevant research and studies conducted on the architecture of Generative Adversarial Networks (GANs), covering various methodologies and technical foundations. The configuration of the suggested framework is detailed in Section 3, while the evaluation of the test results will be discussed in Section 4. Section 5 is the Concluding Reflections on the Proposed Research.

2. Review of Previous Works

The generation of labeled data to train a deep network for brain tumor segmentation in MRI scans is time-consuming and requires the contribution of experienced radiologists.

2.1. CNN and FCN Networks Methods

Dong et al. demonstrated an FCN network for detecting and segmenting brain tumors. Some instances of low-grade glioma were less successful. The U-Net hybrid pyramid model was proposed by Kong et al. [4]. The U-Net method examines both general and regional data. Alex et al. [5] proposed utilizing an FCN network to segregate brain tumors. To eliminate false-positives, they used a 23-layer voxel-based classification using a forward path and linked component analysis. Hawai et al. [6] improved brain tumor segmentation on BraTS. Their network was two-way and cascaded. Two-way architecture considers two receiving portions’ local and global properties. Wang et al. [7] employed convolutional neural networks in a cascade architecture to automatically segment brain tumors. Hossein et al. [8] employed deep cascade neural networks to segregate brain tumors. Hyunwoo et al. [9] presented a decoder–encoder network for semantic segmentation that aggregates high-resolution data into low-level salient edges.
In 2015, Ronberger et al. [10], inspired by the DL approach, proposed a U-Net architecture upon fully convolution networks and used mirror concepts to predict border pixels. The basic idea behind this network is to augment a decreasing route with regular convolutional layers, where merge operators are substituted by upsampling operators. As a result, these layers improve the output resolution. For precise localization, the contraction path’s high-resolution characteristics are mixed with the upsampled output. Then, a successive convolution layer can learn to provide a more accurate output based on the information. In this architecture, the left side is the contraction side, the center is the bridge or bottleneck, and the right side is the expansion side. The three large gray arrows show how the learned feature mapping is copied from the contraction path to the expansion path. This transfer of learned features is the key point of U-Net, which gives the segmented image.
Inspired by U-Net, Miltri et al. [11] presented V-Net for brain tumor segmentation using MRI data. This network has a shrinking path with convolution layers and 5 × 5 × 5 volume kernels for downsampling and a dilation path for deconvolution to increase data size. This helps segmentation. V-Net is an extension of U-Net’s 3D architecture, and each convolution and deconvolution layer are volumetric.
Low- and high-grade gliomas have different architectures. Post-processing volume threshold limitations eliminated the inaccurate cluster categorization error. During training and assessment, the following methods clustered medical scans. SegNet was created by Drozdahl et al. Instead of long connections, they used identity mappings (short-hop connections) [12]. These identity maps allow for a deeper vanishing gradient CNN network and quick learning by recovering lost spatial information during dimensionality reduction. The researchers designed a deeper network with a small core. Pereira et al. [13] constructed a 33-kernel DNN. This model’s fewer weights minimized overfitting. Kamenisas et al. [14] reported a fully linked 3D neural network for brain lesion segmentation. Cui et al. [15] employed a deep cascade neural network to segregate brain tumors. Lin et al. [16] optimized segmentation using fully connected conditional dense learning and CNN. Zhao et al. [17] integrated FCN networks to improve tumor segmentation. Badrinarayanan et al. [18] proposed SegNet, an encoder–decoder architecture. In this architecture, maximal aggregation indices contain all the information in floating precision without feature maps, and the accompanying encoder upscales the input feature maps. Raza, R. et al. [19]’s proposed model is a hybrid of the deep residual network and U-Net model (dResU-Net). The residual network is used as an encoder in the proposed architecture with the decoder of the U-Net model used to handle the issue of the vanishing gradient. The proposed model is designed to take advantage of low-level and high-level features simultaneously in order to make the prediction. Yousef, R et al. [20] introduced Bridged U-Net-ASPP-EVO which exploits Atrous Spatial Pyramid Pooling to enhance capturing multi-scale information to help in segmenting different tumor sizes, Evolving Normalization layers, squeeze and excitation residual blocks, and the max-average pooling for downsampling in two variants.

2.2. GAN Networks Methods

As a result of the preceding, a model that takes all the images as the input and utilizes local and global features to balance the data’s class imbalance is required. To produce high-quality results when analyzing medical images, methods based on GAN networks are frequently used. The following studies, each of which has advantages and disadvantages that will be discussed below, are related to and are like the subject of this study. Different architectures characterized low- and high-grade gliomas, post-processing volume thresholds eliminated cluster categorization errors, and cluster medical scans were carried out during training and assessment. Zhu et al. [21] demonstrated unpaired image training for style transfer. This approach removes fog from photos, adds and removes particles, and more. Ref. [22] proposes a two-stage GAN network (ToStaGAN) to improve brain tumor segmentation. ToStaGAN is a two-stage GAN that uses high-level semantic knowledge to improve brain tumor segmentation. In ToStaGAN, the U-Net network is adopted as the “coarse” generating network in the first step, and a U-shaped text auto-encoder (ConEnDer) is offered as an “excellent” production network in the second stage. This approach improves semantic segmentation by raising numerous parameters. ConEnDer can collect more diversified and abstract characteristics, and by using FEM and coarse prediction map features from the previous stage, it may produce optimal segmentation results. Two-stage production is superior to one-stage. ToStaGAN outperforms the rival segmentation methods. Ref. [23] offers a brain tumor segmentation technique based on GAN. A 3D U-Net is used for segmentation and a classification network is used for discrimination in the GAN architecture. Both integrate multidimensional context information using 3D convolutions. The U-Net 3D model’s extensive connectivity improves network convergence and accuracy. Adversarial training improves segmentation results by bringing them closer to the labeled data, allowing the network to segment unexpectedly tiny tumor subregions and reliably identify each voxel. The suggested method precisely segments brain tumors. In addition, [24] creates the RescueNet network architecture by combining residual and mirror principles. CycleGAN [21], which is used for image-to-image pattern transfer, is improved by this design. For improved brain MRI segmentation, this technique employs less training data and a non-pairwise strategy. RescueNet uses unpaired adversarial training to better the overall tumor, then the central portion of the brain MRI. Preparing vast amounts of labeled data for deep network training is time-consuming and labor-intensive. To train the proposed network without paired data, unpaired training is utilized. In [25], the unsupervised computer modeling of symmetrical alterations in normal brains is used to separate brain cancers from MRI pictures without symmetry. SD-GAN is an unsupervised brain tumor segmentation algorithm. The SD-GAN model learned non-linear mappings between left and right brain pictures to describe normal brain variability (symmetry). The trained SD-GAN was then used to recreate normal brains and segment brain tumors based on asymmetry. Two benchmark datasets were used to test SD-GAN. SD-GAN outperformed new unsupervised segmentation algorithms and was comparable to supervised U-Net. This study showed that symmetric alterations (intrinsic anatomical changes) can be predicted using unannotated normal MRI data and employed in tumor segmentation. Ref. [26] compares rapid MRI algorithms that leverage GAN using anatomical data. This is done to prove that rapid MRI is generalizable and trustworthy, and to look ahead.
In [27], a GAN-based MS-GAN semantic segmentation framework is suggested to localize MS lesions in multimodal brain MRI. It comprises a multimodal encoder–decoder converter G and numerous classifiers D. Multiple input methods are employed to create the generator, which bypasses the encoder’s location information to the decoder to reduce network parameters and improve localization performance. The converter combines multimodal imaging data with multipath encoding and cross-combination. An additional category-related constraint is presented for GAN model adversarial training to reduce convergence in category-based image-to-image translation challenges. In total, 126 relapse MS individuals were evaluated. Performance was compared to other semantic segmentation models including patch-based deep learning. This method segmented more accurately than the advanced techniques. In [28], a new automatic data augmentation approach using GAN networks for reinforcement learning is described, allowing a machine learning-based method to learn current annotation instances more efficiently. This architecture uses a coarse-to-fine generator to generate generic enhanced data from training sets. In trials, the methodology improves MRI pictures by 3.5% compared to typical enhancement methods using the BRATS15 dataset. Using a collaborative segmentation network improves the BRATS15 performance. Ref. [29] proposes an end-to-end GAN-based technique for brain tumor segmentation. This method combines generative and discriminative models and uses GAN instead of conditional random fields for high-order smoothing (CRF). The proposed technique was evaluated in the BRATS 2015 database, and it was found that using GAN enhances the network performance. This method separates a patient’s brain tumor image in 10.8 s less than the existing CNN-based methods. In [30], it is evaluated if synthesizing brain MRI data with GANs allows the employment of a DL model for segmenting T1-weighted, post-contrast T1-weighted, FLAIR, and T2-weighted brain lesions. In this study, 2011–2019 brain MRI data were gathered and simulated without T1-weighted and FLAIR images. Two GANs were trained, verified, and tested on 210 glioblastomas (BRATS 2017) to create T1-weighted pictures from postcontrast T1 images and FLAIR images from T2-weighted images. MSE and SSI were used to evaluate image quality (SSI). A dice similarity coefficient was used to compare segmentations from produced and original MRI scans (DSC). GANs were verified on GBMs and CNS lymphomas to test generalizability. Mann–Whitney, Friedman, and Dunn tests were used. When MRI sequences are absent, GAN-generated brain MRI scans can be fed into deep learning models. In [31], a deep learning approach for classifying MRI tumors is described. A deep neural network is trained as a discriminator in a GAN’s convolutional layers on distinct MRI datasets to extract robust features and understand the visual structure. After replacing completely linked layers, the deep network is taught to recognize three types of tumors. There are six layers in all, with 1.7 million weight parameters. Pre-training a GAN as a discriminant, along with data augmentation (image rotation and mirroring), reduces over-training on a small dataset. This approach was used on 3064 T1-CE MRI scans from 233 people. A neural network in a GAN is initially trained as a discriminator in its convolutional layers on multiple MRI datasets to extract robust features and grasp the structure of MRI pictures. After replacing completely linked layers, the deep network is taught to recognize three types of tumors. There are six layers in all, with 1.7 million weight parameters. Pre-training as a GAN discriminant combined with data augmentation and deletion reduces over-training on a small dataset.
In [32], Kiani Kalejahi, B. et al. used the possibility of producing multiple-sequence MR images by the application of auxiliary classifier-generating adversarial networks (ACGANs). At the beginning, a deep neural network is trained to function as a discriminator in GAN data sets consisting of MRI images in order to extract the features and also learn the structure of the MR images in their annular layers, and then the layers that are already fully connected are removed.

3. Suggested Method

A neural network is proposed and developed based on the electrochemical activities of neurons. The task of these networks is to model human neural activities in learning knowledge and apply the knowledge obtained in similar situations. Various models have been proposed for modeling neural structures, among which structures like CNN have performed better in modeling the activity of the human visual system in relation to the visual cortex of the brain. The proposed method of this research attempted to design and improve the optimal structure of this combination of these two types of networks for the segmentation of brain MRI images by utilizing innovations related to CNN and GAN networks in various topics of machine vision over the last few years.
This approach develops a GAN framework where one neural network (the generator) tries to generate a model that resembles real neuron activity. In contrast, the other network (the discriminator) tries to distinguish between real and generated data. In addition, it defines a loss function that incorporates both the traditional GAN loss (adversarial loss) and a loss related to the generated model to the real data and fine-tunes the network architecture and hyperparameters based on the performance of a validation set.

3.1. Data Preprocessing

The term “data preprocessing” is used to describe the process of preparing data for use in training a network. Image resizing, or scaling volumetric images to a specified dimension, is a useful preprocessing technique in this situation. As datasets might contain information on a wide range of dimensions, it is vital to standardize the data across all of them. So, resampling the images to get a consistent image resolution is important. This easy technique can also assist in keeping the number of calculations from increasing excessively when large images are delivered over a network. The term “normalization” refers to a crucial step in the pre-training phase of neural network image processing. Normalization is used to adapt measurements to a standard scale. When the inputs to a neural network have significantly different scales (the pixel values of the pictures are of varying sizes), normalization is necessary. Standard normalization procedures include min–max [12], which shifts the input range to a new interval (often 0–1), and z-score [13], which transforms each input sample so that its mean is zero and its variance is one. The use of a decimal system of measurement is highlighted [14]. The min–max technique is used to standardize the raw data in this research.

3.2. Structure of the Proposed Method

A fully convolutional network (FCN) is combined with a generative adversarial network (GAN) in the suggested method. The proposed GAN network design has two major nodes: a generator and a discriminator.
  • Generator: A segmentation network that automatically creates segmented images of brain tumors. This is referred to as the algorithm’s core module.
  • Discriminator: the classification network validates the input, which indirectly improves segmentation accuracy.
  • Loss function: adversarial training coordination between generator and discriminator, combination of conventional segmentation and adversarial training optimization objectives.
Once the network is randomly set up, its parameters stay the same. The generative network used for segmentation is then repeated to create a generative network that does not need adversarial training. The parameters of the generator network are then modified, and the discriminator network is repeated. The alternative loop iteration method can generate segmentation results while enhancing classification accuracy.
As previously stated, the suggested 3D CNN is based on the FCN approach and is employed in two structures: the generator and the discriminator. FCN networks such as [15], 3D U-Net [16], and V-Net [11] have seen substantial growth in medical image segmentation applications in recent years. The FC layer(s) in a traditional CNN are totally deleted in the FCN method, and convolution layer(s) are employed instead. In general, the selection of the FCN network for segmentation applications is a very difficult and complex task. There are some networks in the literature (such as [17] and [18]) that are used for this application. These networks, however, do not provide the necessary interdependence at the ground level. Adding hop connections between the upsampling and downsampling pathways is how the U-Net architecture [10] gets around this issue. The word “mirroring” has come to describe this type of alteration, and gradient fading is another issue with the training of deep networks. Residual Network Architecture [7] proposes using homogeneous mappings, or shortcut paths, to solve the issue of gradient fading during back propagation and learning network parameters.
A distinct Fully Convolutional Network (FCN) architecture has been developed for the purpose of segmenting brain tumors. This architecture is specifically designed for the two components of the proposed Generative Adversarial Network (GAN), namely the generator and discriminator. The suggested Generative Adversarial Network (GAN) incorporates the ideas of mirroring and residual. Figure 2 illustrates the comprehensive architecture of the proposed Generative Adversarial Network (GAN). The model consists of two distinct sub-networks, namely the generator and the differentiator, each with its own individual framework. In order to address the issue of image generation, the traditional GAN network architecture is employed, wherein the generator component is utilized to generate synthetic images through the algorithmic process. In order to address the image segmentation problem, the GAN network generator model is employed by replacing the conventional generator with the image segmentation network. The generative network is capable of autonomously segmenting images that are classified as unreal. Furthermore, the discriminator can serve the purpose of segregation and classification. Numerous sophisticated models are built upon Convolutional Neural Networks (CNNs) that are specifically designed for the purpose of picture classification. Nevertheless, there are fundamental differences between dense prediction applications, such as semantic segmentation, and picture classification. In this study, we employ a novel convolutional module known as extended convolution [33], specifically designed for the purpose of dense prediction. The module offered utilizes an extended convolution technique to gather multiscale conceptual information while minimizing any loss in resolution. Extended convolutions enable a significant expansion of the affected region, exponentially increasing its scope, while maintaining high resolution and comprehensive coverage. The study conducted by Yu et al. [33] provided evidence of the module’s ability to improve the performance of semantic segmentation systems. The utilization of extended convolution has been implemented for all convolutional layers in this study based on the provided justification and empirical observations.

3.2.1. Generator Structure

The design provided for the generating structure in our research involves the utilization of a three-dimensional fully convolutional network (3D FCN) to perform the multi-class segmentation of magnetic resonance imaging (MRI) images showing brain tumors. The FCN architecture we propose integrates the U-Net model with the Residual Network. The suggested Fully Convolutional Network (FCN) is designed to accept three-dimensional (3D) Magnetic Resonance Imaging (MRI) pictures of varying dimensions as its input. The segmentation map generated offers the probability distribution of the voxel classes at each spatial position. The quantity of channels depicted on the map is equivalent to the quantity of labels present. The network depicted in Figure 2 consists of two distinct pathways: the Downsampling path, sometimes referred to as the straw sampling road, and the Upsampling path, also known as the rising sampling path. The convolution operation consists of a series of sequential layers, including volume convolution, normalization, and activation. In addition, the construction of feature jump links is implemented to enhance the segmentation feature maps’ quality by facilitating the transition from the contraction path to the expansion path.
The generator network under consideration comprises three distinct building blocks, namely a convolutional block, a residual block, and an output block. The composition of each CB block consists of a three-dimensional extended convolution layer, a group normalization (GN) layer, and an exponential linear unit (ELU) activation function. In contrast, every database block consists of a three-dimensional extended deconvolution (Deconv) layer, a group normalization (GN) layer, and an exponential linear unit (ELU) activation function. In order to achieve feature compression sampling, the integration of residual blocks is implemented at intermediate phases of the contraction process. The aforementioned blocks are designed to tackle the issue of gradient fading that arises in the process of back-propagation. Within the framework of the expansion pipeline, it is observed that an intermediate stage exists whereby an upsampling layer is implemented prior to the subsequent appending layer. The layer serves as a connection between the expansion path feature maps and the contraction path feature maps, as their quantities may vary. Additionally, the incorporation of this topology adds a sense of symmetry to the network under consideration. To address the problem of resolution loss, a concatenation layer is included between the symmetric feature maps obtained from the contraction path and the output of each Deconv layer in the expansion path. The information pertaining to the quantity and sizes of feature maps, which provides a more detailed understanding of the network’s intended functionality, is shown in Figure 3.
The generative network generates a segmentation volume that matches the dimensions of the input volume. The number of final convolutional features in the generated volume is determined by the desired number of classes for prediction, achieved through the use of 1 × 1 × 1 convolutions. It is worth mentioning that the study of volumetric images necessitates the utilization of 3D operations, such as convolution, deconvolution, and fusion layers. These operations are presently supported by some machine learning frameworks, such as TensorFlow. The aim of the subsequent sections is to provide clear definitions of these notions in order to enhance the comprehension of the proposed network.

3.2.2. Discriminating Structure

The utilization of the 3D FCN network in the GAN network involves the categorization of images as a discriminator, which in turn generates the generative model. The primary purpose of this network is to differentiate real input from fake input. The discriminator segment inputs a pair of images consisting of the initial segmentation images and the images segmented by the generator segment. Based on the generated segmented images, it is apparent that this pair of images may be classified into two separate groups:
(1)
Experts manually label original images and segmented images. In other words, these categories are segmentation masks and represent the real value and the category with the value.
(2)
The generator automatically labels both the original image and the segmented image with the generated value. The category is 0.

3.2.3. Convolutional Neural Networks

The convolutional layer is used as the primary component within the Convolutional Neural Network (CNN) architecture, responsible for conducting convolutions between numerous filters, commonly referred to as “kernels,” and the input image. The filters execute the convolution process by traversing the input with respect to its dimensions. The result derived from this process is sometimes referred to as a “feature map” or an “activation map.” In a more precise style, the convolutional layers of a neural network receive the unprocessed image as the input in order to compute the resulting signal. This enables the deep learning model to circumvent the need for manual techniques in the segmentation of brain tumors within MRI brain images. The resolution of intricate abstract learning patterns can be effectively achieved by the utilization of non-linear algorithms and varying degrees. The objective of utilizing this layer is to extract features from input images. This is achieved by convolving the pixels of the input image with a small region known as the receptive field, which exhibits local connectivity. Indeed, it might be posited that this stratum imparts a sense of profundity to the visual representation. The hyperparameters of the system consist of a collection of trainable filters, which encompass the dimensions (F) and step size (S) of each filter. The term “step size” (S) denotes the quantity of pixels by which the window is displaced following each action. In the majority of instances, SF is regarded as being significant. It is noteworthy to emphasize that the convolution process can be executed in one-dimensional, two-dimensional, and three-dimensional formats. In the network under consideration, the utilization of a 3D convolution operation is necessitated by the inherent three-dimensional structure of the input images. As previously stated, a specialized convolutional module known as extended convolution has been employed for conducting 3D convolution operations. One notable advantage of this method in comparison to the traditional convolution operation is its ability to provide a broader perspective with a reduced number of parameters. This enables the extraction of a greater amount of content information from the input, resulting in enhanced efficiency. Next, the difference between extended convolution and standard convolution will be explained.
(1)
Let F   :   Z 2 R be a discrete function;
(2)
Let Ω r = [ r , r ] 2 Z 2 ;
(3)
k   :   Ω r R is a discrete filter with size ( 2 r + 1 ) 2 .
If it is assumed that these conditions are true, then the discrete convolution encoder can be defined as a relation:
F k ( p ) = s + t = p F s k ( t )
Now, the generalized state of this operator can be defined as relation:
F l k ( p ) = s + l t = p F s k ( t )
where l is the expansion coefficient and l is the expanded convolution operation. It is worth noting that the standard convolution operation is a special case of extended convolution operation when l is equal to one.
This operator is based on the notion that extended convolution expands the region of impact exponentially without sacrificing coverage or image resolution.
If F 0 , F 1 , , F n 1   :   Z 2 R are discrete functions and k 0 , k 1 , ,   k n 2 :   Ω 1 R is a discrete filter with a size of 3 × 3, then the filters are applied by expanding exponentially for i from 0 to n − 2 in the form of the relation:
F i + 1 = F i 2 t k i
Define the receptive field of an element p in F i + 1 as the set of elements in F 0 that modify the value of F i + 1 ( p ) . Let the size of the receptive field of p in F i + 1 be the number of these elements. It is easy to see that the size of the receptive field of each element in F i + 1   in ( 2 i + 2 1 ) × ( 2 i + 2 1 ) . In fact, the area of influence of the square root increases exponentially. More precisely, it can be stated that the number of parameters associated with each layer is the same and the influence area grows exponentially, while the number of parameters grows linearly. This is shown in Figure 4.

3.2.4. Pooling Layer

The next phase is to use the features acquired from the convolutional layer for the purpose of classification. There are no discernible benefits associated with employing a classifier for the purpose of categorizing an extensive array of features. Furthermore, it is possible for the classifier to experience the issue of overfitting. The integration layer is employed as an approach to address this issue, effectively reducing the spatial dimensions and partially mitigating the computational burden associated with network training. It is noteworthy to mention that the merging procedure is executed individually on every image channel. There are two types of integration: maximum integration and mean integration. Peak integration involves determining the highest value inside a certain area by employing an integration window, often a 2 × 2 filter. The practice of average pooling involves computing the mean value within a specified region, and it was initially implemented in the LeNet convolutional neural network. In practical applications, it has been observed that max pooling tends to yield superior results compared to mean pooling due to its ability to retain the discovered characteristics. As a result, the suggested network, which encompasses both the generator and discriminator, employs a strategy of maximum integration.

3.2.5. Group Normalization Layer (GN)

Normalization is a widely employed technique in deep neural networks, which has the potential to enhance the generalizability of the network and facilitate the convergence of the cost function. Normalization is a common practice that is typically carried out subsequent to a fully connected layer, prior to a convolutional layer, or prior to a nonlinear layer. This feature facilitates an accelerated pace of learning and mitigates the model’s dependence on its original configuration. Batch normalization (BN) is a fundamental and valuable technique in the advancement of deep learning, enabling the training of various networks. On the contrary, normalization in relation to the category dimension gives rise to challenges. As the size of the cluster shrinks, there is a significant increase in the BN error. This increase can be attributed to the inaccurate estimation of cluster statistics. As a consequence of this constraint, Bayesian Networks (BNs) are not suitable for training larger models or transferring data to computer vision applications that need the classification of small classes due to their reduced memory utilization. Consequently, the GN approach [21] was chosen as a viable alternative to the BN in this thesis. It is important to acknowledge that this particular layer was only utilized during the productive phase. The Gaussian normalization (GN) technique partitions the channels into several groups and calculates the average and variance for the normalization of each group. The accuracy of GN computations is consistent across a diverse range of batch sizes, indicating that they are not influenced by the size of the batch being processed. The performance of GN is comparable to that of BN when employing standard batch sizes, and it surpasses alternative normalization approaches, such as instance normalization (IN) [34] and layer normalization (LN) [22]. The experimental findings indicate that GN has the potential to effectively substitute BN in various tasks.
Normalization procedures, such as Batch Normalization (BN), Instance Normalization (IN), Layer Normalization (LN), and Group Normalization (GN), typically involve the computation of relational equivalents:
x ^ i = 1 σ i ( x i μ i )
where x is the feature calculated by a layer and i is the index. For example, i in a two-dimensional image is a four-dimensional vector as i = i N , i C , i H , i W to index the features, where N is the number of samples, C is the number of channels, H is the spatial length, and W is the spatial width. µ and σ represent the mean and standard deviation (std), respectively, and are calculated using the relations:
μ i = 1 m k S i x k
and
σ i = 1 m k S i ( x k μ i ) 2 + ϵ
ϵ is a small constant to maintain numerical stability, S i is the set of pixels whose mean and standard deviation are calculated, and m is the size of the set. Most of the normalization methods use the relation in common, but they differ in using different pixel sets (in other words, S i ) to estimate µ and σ; in other words, the number of their estimated probabilities is different. This is shown in Figure 5. It is worth mentioning that in this form, GN has 2 groups, and each group has 3 channels.
Intuitively, GN uses the set S i to calculate µ and σ. The set S i has relational equivalents:
S i = k k N = i N , k C C / G = i C C / G
where G is a default hyperparameter indicating the number of groups, C / G indicates the number of channels in each group, the ⌊. ⌋ function is correct, and k C C / G = i C C / G indicates that indices i and k are in the same group of channels. This, of course, is assuming that each group of channels is stored sequentially along the C axis. GN calculates µ and σ values along H and W and a group of channels (in other words, C / G ).

3.2.6. Spectral Normalization (SN)

The instability of the training process is one of the issues in the field of GAN networks. As a result, we applied a unique weight normalization method termed SN [23] to stabilize the discriminator training in this thesis. Among the benefits of this method are its low computational complexity and ease of implementation. On benchmark datasets such as CIFAR-10, STL-10, and ILSVRC2012, the performance of SN has been tested, and it has been experimentally demonstrated that spectrally normalized GANs can produce images of higher or equal quality than previously trained stabilization methods. A simple discriminator based on a neural network with input x can be thought of as the following relationship:
f ( x , θ ) = W L + 1 a L ( W L ( a L 1 W L 1 ( a 1 W 1 x ) ) )
where θ = { W 1 , , W L , W L + 1 } , the set of learning parameters is W l R d l × d l 1 , and W L + 1 R 1 × d l and a 1 are the non-linear activation function of the drives.
Here, the bias is omitted for the simplicity of the calculations. The final output of the differentiator can be considered as the relation:
D ( x , θ ) = A f ( x , θ )
where A is the activation function associated with the user’s preferred divergence in the distance measurement. The conventional GAN formula is defined as the relationship:
min G max D V ( G , D )
where the minimum and maximum values of G and D are determined by assessing the sets of generating and discriminating functions. Goodfellow et al. [9] have proposed the conventional relationship V (G, D) in the form of equation:
E x ~ q d a t a log D ( x ) + E x ~ p G [ log ( 1 D x ) ]
where q d a t a is the distribution of the data to be learnt and p G is the distribution of the generator to be trained using min–max adversarial optimization. The assertion of activation function assumes that the value of A in D is a continuous function in [0, 1]. (For example, the sigmoid function.) The optimal discriminant V (G, D) for a given generator G has the interfaces of the type.
D G x : = q d a t a ( x ) / ( q d a t a + p G ( x ) )
According to the DL literature, the performance space from which the discriminators are chosen influences the GAN performance. Some studies (such as [24,25,26]) support the necessity of Lipschitz constant continuity in ensuring statistical boundedness. The optimum discriminator of GANs, for example, is produced in the conventional formula as the relation:
D G x = q d a t a x q d a t a + p G x = s i g m o i d ( f x )
f x is defined as equation:
f x = log q d a t a ( x ) log p G ( x )
Also, its derivative is calculated as equation:
x f x = 1 q d a t a ( x ) x q d a t a ( x ) 1 p G ( x ) x p G ( x )
This derivative can be infinite or even uncomputable. This prompts us to apply some regularization conditions to the derivative of f(x).
Refs. [25,26] suggest strategies to regulate the Lipschitz constant of the differentiator by adding a systematic term to input samples x. In SN, we followed in their footsteps and sought for the discriminant D from continuous K-Lipschitz functions:
arg   max f l i p K V ( G , D )
where f l i p is the smallest value of M in relation (17) for each x and x with soft L2.
f x f ( x ) / x x M
Since input-based systematization approaches offer relatively simple formulae based on examples, they suffer from the difficulty that systematization cannot be applied to the space beyond the data generator and distribution without the use of heuristics. The SN technique circumvents this issue by normalizing the weight matrices, as proposed by Yoshida et al. [27]. Following is an explanation of the SN technique in further depth.
The SN method controls the Lipschitz constant of the discriminant function f by limiting the spectral softness of each layer (in other words, g   : h i n h o u t ). The Lipschitz smooth (in other words, g L i p ) is equal to s u p h σ ( g h ) , where σ(A) is the spectral smooth of the matrix A (the L2-soft of the matrix A)
σ ( A ) = max h : h 0 A h 2 h 2 = max f 2 1 A h 2
This is equivalent to the largest single value of A. Therefore, for a linear layer g(h) = Wh, the smooth is obtained using g L i p = s u p h σ g h = s u p h σ W = σ W . If the Lipschitz smoothness of the activation function a l l i p is equal to 1, we can use the kg inequality to see the following bound on f l i p :
f L i p h L W L + 1 h L L i p . a L L i p . h L 1 W L h L 1 L i p a 1 L i p . h 0 W 1 h 0 L i p = l = 1 L + 1 h l 1 W l h l 1 L i p = l = 1 L + 1 σ W l
The SN method normalizes the spectral smoothness of the weight matrix W so that it satisfies the constraint σ(W) = 1 for the Lipschitz constant:
W ¯ S N ( W ) = W / σ W
If we normalize each W l using the relation (20), we can obtain the inequality (), and also we can find that σ ( W ¯ S N ( W ) ) is equal to 1 to find that f L i p from the upper bound to 1 is limited. The key difference between the SN technique and the soft spectral regularization approach [27] is that the soft spectral approach penalizes the objective function by including an explicit regularization component. This approach differs from the SN method in that it does not attempt to adjust the spectral smoothness to a certain value. In addition, in SN, when the derivative of the normalized cost function is reorganized and the objective function is rewritten, the cost function is augmented with a regularization function depending on the sample data. On the other hand, spectral soft segmentation, such as L2 soft and Lasso segmentation, imposes the independent segmentation of the sample data on the cost function.

3.2.7. Loss Function

Assuming that the network projections consist of two volumes with the identical resolution as the original input data, it frequently appears that the region of interest occupies a relatively small portion of the volumetric medical images, such as the 3D MRI images employed in this investigation. In this study, the region of interest undergoes processing via a SoftMax layer, which generates the probability of each voxel belonging to either the foreground or background. This occurrence often leads to the learning process getting encased in the local minima of the loss function, which in turn produces a network that exhibits highly biased predictions compared to the background. Consequently, the primary subject matter is typically obstructed or only partially discernible. Previous studies employed loss functions that incorporated sample reweighting, wherein greater emphasis was placed on foreground instances compared to background instances throughout the learning process. Mi-lhtari et al. [11] introduced a novel objective function that utilizes the Dice coefficient, a metric that ranges from 0 to 1, as its basis. The term “D” is used to denote the correlation or association between two binary volumes:
D = 2 i N p i g i i N p i 2 + i N g i 2
where collection is performed on N voxels, a predicted binary segmentation volume p i G , and a binary mask (label) volume g i G . This Dice formula can be distinguished by creating a gradient:
D p j = 2 g j ( i N p i 2 + i N g i 2 ) 2 p j i N p i g i ( i N p i 2 + i N g i 2 ) 2
The prediction is made using voxel j. Using this connection, it is not essential to add weights to samples of different classes in order to obtain a reasonable balance of foreground and background voxels, and the experimentally observed results are generated. One of the binary Dice loss function’s key shortcomings is its inability to be used for multi-class segmentation. The mean of the binary Dice function is used as the relation to derive a unique measure for multiclass segmentation:
D m e a n ( p , g ) = 1 L 1 L 2 i ( g l i + p l i ) i ( g l i + p l i )
where { g l i } i X ,   l L and { p l i } i X ,   l L are the sets of label probability vectors for all real and predicted voxels, respectively. Sudre et al. [29] have proposed a generalized form (GDL) of this function. The GDL form has a relationship of the form:
D G D L ( p , g ) = 2 l α l i m i n ( p l i + g l i ) l α l i m i n ( p l i + g l i )
where the coefficient { α l } l L allows to weigh the participation rate of each class. Since the proposed GAN network includes two independent generators and discriminator structures, we have used two loss functions. The discriminant cost function (LD) consists of two parts. The first part is the sum of the L2 error of the discriminator output, D (.,.), between the original image x and the corresponding mask y calculated with a tensor of ones. The second part, the L2 error of the discriminator output between the original image and the prediction of the corresponding segmentation y ^ , is calculated by the generator with a tensor of zeros; in other words, we have a relation as:
L D = L 2 D x , y , 1 + L 2 [ D x , y ^ , 0 ]
On the other hand, the productive cost function (LG) consists of two parts. The first part is the total L2 error of the discriminator output between the original image x and the corresponding segmentation prediction y ^ with a tensor of ones. The second part, the GDL error [29], is calculated between the mask image and the output of the generator. In other words, we have interfaces in the form of:
L G = L 2 D x , y ^ , 1 + G D L y , y ^

4. Experiments and Results

4.1. Dataset

The approach used in this research involved the utilization of the brain tumor segmentation challenge dataset (BraTs 2021). The dataset was publicly released during the MICCAI conference and was subsequently utilized in a competitive setting, allowing the participating groups to evaluate their novel methodologies using this dataset. The BraTs 2021 dataset comprises three-dimensional magnetic resonance imaging scans of individuals. The Hold-Out strategy was employed to partition the dataset into two distinct subsets, namely the training set and the test set. The training set constituted 80% of the dataset, while the remaining 20% was allocated to the test set. Figure 6 shows the instances inside the dataset, along with a visual representation of their corresponding actual states in a three-dimensional space.
The representation of three-dimensional (3D) images on paper is often challenging, thus necessitating the display of the two-dimensional (2D) parts of these images. Figure 7 is an instance from the dataset, whereby three cuts have been made along the three axes of the image. The leftmost column displays the Flair channel images of the sample. The middle column exhibits the equivalent slices extracted from the real-state image. Lastly, the rightmost column portrays the amalgamation of the two columns, revealing three distinct tumor locations on the brain image, each represented by a different color, having undergone a transformation. The provided photos depict the inflamed region of the tumor in a green hue, the necrotic tissue of the tumor in a blue hue, and the actively proliferating core of the tumor in a red hue.

Data Set Structure

The BraTS2021 dataset files consist of patients with the highest grade of glioma brain tumor and patients with the lowest grade of glioma brain tumor. There is a separate folder for each patient. The number of patients in the training data set is 1251, and the validation data set includes 219 cases. For each patient, 3D images are available to the researchers; there are four images related to different imaging channels (Flair, T2, T1c, and T1), and one image of a real situation in which different parts of the brain tumor are identified by medical experts.

4.2. Evaluation Criteria

In general, the goal of neural networks is to provide a highly accurate estimate, and in fact, each neural network is an estimator. The closer the value estimated by the network is to the ideal value, the more successful the network is. The type of the estimated value is determined according to the type of ideal labels in the data set. In this research, our ideal labels are three-dimensional images that we call true state images; these images are used in the form of tensors. As a result, the output of our network will be a 3D tensor.
To evaluate an estimator, statistical evaluation criteria are used. The type of statistical criterion is determined according to the nature of the network output and must be chosen in such a way that it can correctly measure the behavior of the estimator to be evaluated.
Our goal in this research is to find the tumor areas in the 3D real-state images. These areas in the real-state tensor are very limited in number because the healthy brain tissues in the real-state image are marked with a zero label and occupy a very large volume of this image. As a result, the estimation of the elements that have a zero label has no value and considering these textures in the evaluation of the results will lead to wrong conclusions. Therefore, the Dice criterion is used in medical image segmentation methods to obtain estimation accuracy because this criterion does not value true negative samples.
Since the proposed GAN network includes two independent generator and discriminator structures, we have used two loss functions. The discriminant cost function (LD) consists of two parts. In other words, we have interfaces in the form:
L D = L 2 D x , y , 1 + L 2 [ D x , y ^ , 0 ]
On the other hand, the productive cost function (LG) consists of two parts, and we have a relation of:
L G = L 2 D x , y ^ , 1 + G D L y , y ^

4.3. Implementation

Deep neural networks have the capability to undergo training utilizing a central processing unit (CPU). However, due to the substantial computational demands associated with these networks, the training process executed on a CPU tends to be somewhat sluggish and time intensive. The utilization of graphics processing units (GPUs) is prevalent in the training process of the deepest neural networks due to this underlying rationale. The training and testing of the proposed network were conducted on a computer system with 128 GB of RAM, an Intel® Xeon (R) E5-2683 v4 @ 2.10 GHz 52 CPU, and a PNY NVIDIA Quadro P6000 PCI-E Pro 24 GB GPU. Each model took approximately 5–6 days to run. While Google Colab offers the opportunity to utilize the graphics processing unit (GPU) at no cost, it is worth noting that this research also employed its Pro Plus version. However, it is important to acknowledge that the utilization of Google Colab is accompanied by certain obstacles that notably impact the duration of network training. The network was trained using the cross-software framework implemented on the TensorFlow platform. TensorFlow is an open-source software framework that was initially developed by Google and has since been collaboratively enhanced by a global community of contributors. It is primarily implemented using the C programming language. During the implementation phase, the suggested architecture was trained using two different approaches.
The initial approach involved utilizing images with a higher count of convolutional layers following each normalization. Subsequently, in the subsequent training phase, these layers were omitted in both architectures. Additionally, data augmentation techniques were employed throughout the experiments.
In each training iteration, a systematic process is followed where all the training samples are inputted into the neural network. Subsequently, the network generates an output and evaluates the estimation error. The optimizer function then adjusts the network’s variable parameters. The network’s training duration, when employing data augmentation, averaged approximately 25 min for each training iteration. The training process, which encompassed both methods, required 500 training rounds to achieve the necessary response. Without the inclusion of data augmentation techniques, the full training process lasted around 4–5 days. The network configurations are effectively presented in Table 1.

4.4. Results

In this study, following the necessary preprocessing procedures specific to magnetic resonance imaging, the obtained data were utilized to generate image segmentation and classification. These outputs were subsequently employed as the input to a proposed network model, aiming to extract either a feature or category. Classifications are employed for categorization purposes. Several studies have demonstrated the potential impact of utilizing deep learning models in conjunction with neural imaging data [33,34]. Additionally, a group of researchers has presented a collection of brain tumor cases [25,29]. The findings presented by the experts indicate a notable enhancement in the classification outcomes. Consequently, it is anticipated that this study will contribute to the advancement of classification results in the domain of brain tumor detection and classification, employing diverse imaging data and deep learning models. Following the training of the proposed network on the majority of the BraTS dataset, we employed various data augmentation techniques to enrich the data and enhance the performance of the network. The subsequent sections will elaborate on the results obtained from these tests, while the final section will include a comparative analysis between the network’s highest accuracy and that of prior research. The initial network configuration is outlined in Table 1. Figure 8 and Figure 9 show a visual depiction of the network’s output and its corresponding steady-state image in three dimensions.
Table 2 and Table 3 summarize the average loss values and accuracy of the two techniques for the overall tumor, the tumor core, and the active tumor core, based on the execution of 1000 batches and 500 IPOCs. Table 3 shows the improved performance due to the second model’s usage of data augmentation and network design.

4.5. Comparison of Results

To assess the quality of the suggested approaches, we compared the findings generated from the proposed model with other comparable studies on the BraTS dataset, as shown in Table 4.
Table 3 shows the average accuracy of the two techniques for the three regions of the entire tumor, tumor core, and active tumor core, based on the execution of 1000 batches and 500 IPOCs. As seen in the table, the second model’s usage of data augmentation and network design produced superior outcomes.

5. Conclusions

Recent research in the field of medical imaging, specifically focusing on MRI scans, aims to explore various methodologies and their use for the precise extraction of tumor location and classification. This study aimed to implement measures to enhance the following within this field of study. Based on the experimental results of this study, it is recommended that the following methodologies be employed to enhance the precision and caliber of upcoming research attempts. The 3D GAN model that has been proposed exhibits potential for utilization in the upcoming BraTS2021 challenges through multiple avenues. These include training the model to enhance the optimization of the BraTS2021 evaluation criteria, as opposed to relying solely on the criteria provided within the dataset. Additionally, optimizing the proposed architecture can be achieved through the application of algorithmic optimization techniques for hyperparameters. Furthermore, the integration of various data augmentation methods can be explored to enhance the performance of the trained model. The investigation of the multi-task learning approach presents an additional path for inquiry, as it has the potential to be combined with the proposed architecture in order to enhance precision. To enhance the detection and differentiation of tumors, this network has the potential to utilize data from several sources, thereby constructing a more comprehensive representation. The proposed architectural framework can also be assessed using alternative imaging modalities, such as CT scans, and segmentation software designed for general purposes. In conclusion, it can be asserted that the proposed 3D GAN model has the potential to be employed for both image segmentation and image data augmentation. The proposed model has the potential to be employed in conjunction with a 3D noise-to-image GAN model, such as those referenced in citations [38,39,40]. This combination allows for the integration of authentic segmentation outputs, which may then be transformed into realistic MR volumes. The simultaneous utilization of both generative adversarial networks (GANs) shows prospects for the quick production of high-fidelity magnetic resonance (MR) images.

Author Contributions

Writing—original draft, B.K.K.; Supervision, S.M.; Project administration, S.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Used dataset is available in: https://www.med.upenn.edu/cbica/brats2021/ and prepared model is available in: https://github.com/hamyadkiani/3D-GAN accessed on 7 September 2023.

Conflicts of Interest

The Title of manuscript: “Segmentation of Brain Tumor Using a 3D Generative Adversarial Network “it is declared that all co-authors have confirmed the contents of the manuscript and there is no financial interest to report.

References

  1. Gould, S.; Fulton, R.; Koller, D. Decomposing a scene into geometric and semantically consistent regions. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 29 September–2 October 2009; pp. 1–8. [Google Scholar] [CrossRef]
  2. Yousefi-Azar, M.; Varadharajan, V.; Hamey, L.; Tupakula, U. Autoencoder-based feature learning for cyber security applications. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; pp. 3854–3861. [Google Scholar] [CrossRef]
  3. Hatcher, W.G.; Yu, W. A Survey of Deep Learning: Platforms, Applications and Emerging Research Trends. IEEE Access 2018, 6, 24411–24432. [Google Scholar] [CrossRef]
  4. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2323. [Google Scholar] [CrossRef]
  5. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  6. Sutskever, I. Training Recurrent Neural Networks. Ph.D. Thesis, University of Toronto, Toronto, ON, Canada, 2013; p. 101. [Google Scholar]
  7. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
  8. Ulku, I.; Akagündüz, E. A Survey on Deep Learning-based Architectures for Semantic Segmentation on 2D Images. Appl. Artif. Intell. 2022, 36, 2032924. [Google Scholar] [CrossRef]
  9. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Networks. arXiv 2014. [Google Scholar] [CrossRef]
  10. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In MICCAI 2015: Medical Image Computing and Computer-Assisted Intervention; Lecture Notes in Computer Science (Including subseries Lecture Notes in Artificial Intelligence (LNAI) and Lecture Notes in Bioinformatics (LNBI)); Springer: Cham, Switzerland, 2015; Volume 9351, pp. 234–241. [Google Scholar] [CrossRef]
  11. Milletari, F.; Navab, N.; Ahmadi, S.A. V-Net: Fully convolutional neural networks for volumetric medical image segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016. [Google Scholar] [CrossRef]
  12. Gopal, S.; Patro, K.; Sahu, K.K. Normalization: A Preprocessing Stage. Int. Adv. Res. J. Sci. Eng. Technol. IARJSET 2015, 2, 20–22. [Google Scholar] [CrossRef]
  13. Cheadle, C.; Vawter, M.P.; Freed, W.J.; Becker, K.G. Analysis of Microarray Data Using Z Score Transformation. J. Mol. Diagn. 2003, 5, 73–81. [Google Scholar] [CrossRef]
  14. Shalabi, L.A.; Shaaban, Z.; Kasasbeh, B. Data Mining: A Preprocessing Engine. J. Comput. Sci. 2006, 2, 735–739. [Google Scholar] [CrossRef]
  15. Kayalibay, B.; Jensen, G.; van der Smagt, P. CNN-Based Segmentation of Medical Imaging Data. arXiv 2017. [Google Scholar] [CrossRef]
  16. Çiçek, Ö.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-net: Learning dense volumetric segmentation from sparse annotation. In MICCAI 2016: Medical Image Computing and Computer-Assisted Intervention; Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence (LNAI) and Lecture Notes in Bioinformatics (LNBI)); Springer: Cham, Switzerland, 2016; Volume 9901, pp. 424–432. [Google Scholar] [CrossRef]
  17. Noh, H.; Hong, S. Learning Deconvolution Network for Semantic Segmentation. Available online: http://openaccess.thecvf.com/content_iccv_2015/html/Noh_Learning_Deconvolution_Network_ICCV_2015_paper.html (accessed on 5 July 2022).
  18. Badrinarayanan, V.; Kendall, A.; Cipolla, R. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  19. Raza, R.; Bajwa, U.I.; Mehmood, Y.; Anwar, M.W.; Jamal, M.H. dResU-Net: 3D deep residual U-Net based brain tumor segmentation from multimodal MRI. Biomed. Signal Process. Control 2023, 79, 103861. [Google Scholar] [CrossRef]
  20. Yousef, R.; Khan, S.; Gupta, G.; Albahlal, B.M.; Alajlan, S.A.; Ali, A. Bridged-U-Net-ASPP-EVO and Deep Learning Optimization for Brain Tumor Segmentation. Diagnostics 2023, 13, 2633. [Google Scholar] [CrossRef]
  21. Wu, Y.; He, K. Group Normalization. Int. J. Comput. Vis. 2020, 128, 742–755. [Google Scholar] [CrossRef]
  22. Ba, J.; Kiros, J.; Hinton, G. Layer Normalization. In Proceedings of the NIPS 2016—Deep Learning Symposium, Barcelona, Spain, 8 December 2016; Available online: https://openreview.net/forum?id=BJLa_ZC9 (accessed on 3 July 2022).
  23. Miyato, T.; Kataoka, T.; Koyama, M.; Yoshida, Y. Spectral normalization for generative adversarial networks. arXiv 2018. [Google Scholar] [CrossRef]
  24. Uehara, M.; Sato, I.; Suzuki, K.N. Generative adversarial nets from a density ratio estimation perspective. arXiv 2016. [Google Scholar] [CrossRef]
  25. Qi, G.J. Loss-Sensitive Generative Adversarial Networks on Lipschitz Densities. Int. J. Comput. Vis. 2020, 128, 1118–1140. [Google Scholar] [CrossRef]
  26. Gulrajani, I.; Ahmed, F.; Arjovsky, M.; Dumoulin, V.; Courville, A. Improved Training of Wasserstein GANs. In Advances in Neural Information Processing Systems 30 (NIPS 2017); Curran Associates: Red Hook, NY, USA, 2017; pp. 5768–5778. [Google Scholar]
  27. Yoshida, Y.T.M. Spectral norm regularization for improving the generalizability of deep learning. arXiv 2017. [Google Scholar] [CrossRef]
  28. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn Res. 2014, 15, 1929–1958. [Google Scholar]
  29. Sudre, C.H.; Li, W.; Vercauteren, T.; Ourselin, S.; Cardoso, M.J. Generalised Dice overlap as a deep learning loss function for highly unbalanced segmentations. In DLMIA 2017, ML-CDS 2017: Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Springer: Cham, Switzerland, 2017. [Google Scholar] [CrossRef]
  30. Kingma, D.P.; Ba, J.L. Adam: A Method for Stochastic Optimization. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  31. Bowles, C.; Chen, L.; Guerrero, R.; Bentley, P.; Gunn, R.; Hammers, A.; Dickie, D.A.; Hernández, M.V.; Wardlaw, J.; Rueckert, D. GAN Augmentation: Augmenting Training Data using Generative Adversarial Networks. arXiv 2018. [Google Scholar] [CrossRef]
  32. Kiani Kalejahi, B.; Meshgini, S.; Danishvar, S. Brain tumor segmentation by auxiliary classifier generative adversarial network. Signal Image Video Process. 2023, 17, 3339–3345. [Google Scholar] [CrossRef]
  33. Yu, F.; Koltun, V. Multi-scale context aggregation by dilated convolutions. In Proceedings of the 4th International Conference on Learning Representations, ICLR 2016, San Juan, PR, USA, 2–4 May 2016. [Google Scholar]
  34. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the 32nd International Conference on Machine Learning (ICML 2015), Lille, France, 6–11 July 2015; Volume 1, pp. 448–456. [Google Scholar]
  35. Kamnitsas, K.; Ledig, C.; Newcombe, V.F.J.; Simpson, J.P.; Kane, A.D.; Menon, D.K.; Rueckert, D.; Glocker, B. Efficient multi-scale 3D CNN with fully connected CRF for accurate brain lesion segmentation. Med. Image Anal. 2017, 36, 61–78. [Google Scholar] [CrossRef] [PubMed]
  36. Beers, A.; Chang, K.; Brown, J.; Sartor, E.; Mammen, C.P.; Gerstner, E.; Rosen, B.; Kalpathy-Cramer, J. Sequential 3D U-Nets for Biologically Informed Brain Tumor Segmentation. arXiv 2017, arXiv:1709.02967. [Google Scholar]
  37. Shen, H.; Zhang, J.; Zheng, W. Efficient symmetry-driven fully convolutional network for multimodal brain tumor segmentation. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017. [Google Scholar]
  38. Eklund, A. Feeding the zombies: Synthesizing brain volumes using a 3D progressive growing GAN. arXiv 2019. [Google Scholar] [CrossRef]
  39. Kwon, G.; Han, C.; Kim, D.S. Generation of 3D Brain MRI Using Auto-Encoding Generative Adversarial Networks. arXiv 2019. [Google Scholar] [CrossRef]
  40. Mutepfe, F.; Kalejahi, B.K.; Meshgini, S.; Danishvar, S. Generative adversarial network image synthesis method for skin lesion generation and classification. J. Med. Signals Sens. 2021, 11, 237. [Google Scholar]
Figure 1. General structure of GAN network.
Figure 1. General structure of GAN network.
Diagnostics 13 03344 g001
Figure 2. The general structure of the proposed GAN network. Direct diffusion is shown using blue lines and reverse diffusion is shown using black dotted lines.
Figure 2. The general structure of the proposed GAN network. Direct diffusion is shown using blue lines and reverse diffusion is shown using black dotted lines.
Diagnostics 13 03344 g002
Figure 3. Suggested 3D-GAN network generator portion architecture’s general structure.
Figure 3. Suggested 3D-GAN network generator portion architecture’s general structure.
Diagnostics 13 03344 g003
Figure 4. Systematic expansion allows for the exponential extension of the impact area while maintaining resolution and coverage. (a) Expansion with a factor of one, where F_1 is formed by F_0; each element in F_1 has a 3 × 3 impact area. (b) Expansion with a factor of two, where F_2 is formed by F_1 and each element in F_2 has a 7 × 7 impact area. (c) Expansion with a factor of four, where F_3 is formed by F_2; each element in F_3 has a 15 × 15 impact area [33].
Figure 4. Systematic expansion allows for the exponential extension of the impact area while maintaining resolution and coverage. (a) Expansion with a factor of one, where F_1 is formed by F_0; each element in F_1 has a 3 × 3 impact area. (b) Expansion with a factor of two, where F_2 is formed by F_1 and each element in F_2 has a 7 × 7 impact area. (c) Expansion with a factor of four, where F_3 is formed by F_2; each element in F_3 has a 15 × 15 impact area [33].
Diagnostics 13 03344 g004
Figure 5. The effectiveness of the BN, LN, IN, and GN normalizing procedures. Each image shows a feature mapping tensor with N representing the number of samples, C representing the number of channels, and (H and W) representing the spatial length and breadth. Blue pixels are normalized with the same mean and variance, which are derived by adding their values [21].
Figure 5. The effectiveness of the BN, LN, IN, and GN normalizing procedures. Each image shows a feature mapping tensor with N representing the number of samples, C representing the number of channels, and (H and W) representing the spatial length and breadth. Blue pixels are normalized with the same mean and variance, which are derived by adding their values [21].
Diagnostics 13 03344 g005
Figure 6. Samples in the dataset.
Figure 6. Samples in the dataset.
Diagnostics 13 03344 g006
Figure 7. Images of tumor swelling area with three different colors.
Figure 7. Images of tumor swelling area with three different colors.
Diagnostics 13 03344 g007
Figure 8. A 3D view of the network output image and the corresponding real-state image.
Figure 8. A 3D view of the network output image and the corresponding real-state image.
Diagnostics 13 03344 g008
Figure 9. Shows a two-dimensional cross-section of the brain image of one of the existing samples. In this figure, we can see the real status image next to the network output image, and the segmented areas of the brain tumor are marked with three colors. In this way, the swelling area of the tumor is shown in green, the dead tissue of the tumor is shown in blue, and the active core of the tumor is shown in red.
Figure 9. Shows a two-dimensional cross-section of the brain image of one of the existing samples. In this figure, we can see the real status image next to the network output image, and the segmented areas of the brain tumor are marked with three colors. In this way, the swelling area of the tumor is shown in green, the dead tissue of the tumor is shown in blue, and the active core of the tumor is shown in red.
Diagnostics 13 03344 g009
Table 1. Network settings.
Table 1. Network settings.
Learning Rate1
Optimizer functionAdam
Chance of accidental deletion30%
CostWeighted dice
Table 2. Loss values of the proposed methods.
Table 2. Loss values of the proposed methods.
lossdice_lossdisc_lossloss_valdice_loss_valdisc_loss_val
First Model 2.5476 0.3975 0.5602 2.0057 0.3969 0.0213
Second Model 1.4375 0.2069 0.4030 1.4636 0.2126 0.4004
Table 3. Proposed methods accuracy on BraTS2021 dataset.
Table 3. Proposed methods accuracy on BraTS2021 dataset.
the Active Tumor CoreTumor NucleusWhole TumorNetwork
0.820.820.88Model 1
0.880.860.94Model 2
Table 4. Comparison of accuracy criteria in different networks on BraTS2021 dataset.
Table 4. Comparison of accuracy criteria in different networks on BraTS2021 dataset.
the Active Tumor CoreTumor NucleusWhole TumorNetwork
0.630.670.85Ref [35]
0.610.720.85Ref [15]
0.730.730.88Ref [36]
0.640.750.89Ref [37]
0.840.860.82Ref [22]
0.840.850.91Ref [23]-v1
0.830.860.92Ref [23]-v2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kalejahi, B.K.; Meshgini, S.; Danishvar, S. Segmentation of Brain Tumor Using a 3D Generative Adversarial Network. Diagnostics 2023, 13, 3344. https://doi.org/10.3390/diagnostics13213344

AMA Style

Kalejahi BK, Meshgini S, Danishvar S. Segmentation of Brain Tumor Using a 3D Generative Adversarial Network. Diagnostics. 2023; 13(21):3344. https://doi.org/10.3390/diagnostics13213344

Chicago/Turabian Style

Kalejahi, Behnam Kiani, Saeed Meshgini, and Sebelan Danishvar. 2023. "Segmentation of Brain Tumor Using a 3D Generative Adversarial Network" Diagnostics 13, no. 21: 3344. https://doi.org/10.3390/diagnostics13213344

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop