Next Article in Journal
Novel Probe for Thermally Controlled Raman Spectroscopy Using Online IR Sensing and Emissivity Measurements
Next Article in Special Issue
Automated Precancerous Lesion Screening Using an Instance Segmentation Technique for Improving Accuracy
Previous Article in Journal
Vision-Based Learning from Demonstration System for Robot Arms
Previous Article in Special Issue
Brain Magnetic Resonance Imaging Classification Using Deep Learning Architectures with Gender and Age
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Scale Attention Convolutional Network for Masson Stained Bile Duct Segmentation from Liver Pathology Images

1
Institute of Computer and Communication Engineering, National Cheng Kung University, Tainan City 701, Taiwan
2
Division of Hematology and Oncology, Department of Internal Medicine, E-Da Hospital, Kaohsiung 824, Taiwan
3
Department of Pathology, National Cheng Kung University Hospital, Tainan City 704, Taiwan
4
Kaohsiung Veterans General Hospital, Kaohsiung 813414, Taiwan
5
School of Medicine, I-Shou University, Kaohsiung 824, Taiwan
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(7), 2679; https://doi.org/10.3390/s22072679
Submission received: 25 February 2022 / Revised: 24 March 2022 / Accepted: 29 March 2022 / Published: 31 March 2022
(This article belongs to the Special Issue Artificial Intelligence-Based Applications in Medical Imaging)

Abstract

:
In clinical practice, the Ishak Score system would be adopted to perform the evaluation of the grading and staging of hepatitis according to whether portal areas have fibrous expansion, bridging with other portal areas, or bridging with central veins. Based on these staging criteria, it is necessary to identify portal areas and central veins when performing the Ishak Score staging. The bile ducts have variant types and are very difficult to be detected under a single magnification, hence pathologists must observe bile ducts at different magnifications to obtain sufficient information. This pathologic examinations in routine clinical practice, however, would result in the labor intensive and expensive examination process. Therefore, the automatic quantitative analysis for pathologic examinations has had an increased demand and attracted significant attention recently. A multi-scale inputs of attention convolutional network is proposed in this study to simulate pathologists’ examination procedure for observing bile ducts under different magnifications in liver biopsy. The proposed multi-scale attention network integrates cell-level information and adjacent structural feature information for bile duct segmentation. In addition, the attention mechanism of proposed model enables the network to focus the segmentation task on the input of high magnification, reducing the influence from low magnification input, but still helps to provide wider field of surrounding information. In comparison with existing models, including FCN, U-Net, SegNet, DeepLabv3 and DeepLabv3-plus, the experimental results demonstrated that the proposed model improved the segmentation performance on Masson bile duct segmentation task with 72.5% IOU and 84.1% F1-score.

1. Introduction

According to the Global Hepatitis Report, viral hepatitis led to 1.34 million deaths in 2015 [1]. The WHO have also estimated that 788,000 people die from liver cancer per year, and viral hepatitis (such as hepatitis B and C) is the primary cause leading to hepatocellular carcinoma and cirrhosis. Statistics showed that the prevalence rate of liver disease in Asian had a higher proportion than all the other region [2,3].
In clinical diagnosis, liver disease is generally performed by liver biopsy, since hepatology images can provide information at the cellular level. Generally, the grading and staging of liver disease are evaluated by the Ishak Score system [4]. When scoring a liver biopsy, the pathologists have to observe the characteristics of the entire liver biopsy under a microscope or a digital image. The staging criteria of the Ishak Score include periportal or periseptal interface hepatitis, confluent necrosis, portal inflammation, focal lytic necrosis, apoptosis, focal inflammation, and architectural changes such as fibrosis or cirrhosis. Different standards require pathologists to diagnose in different magnifications to obtain an overall view information or a detailed cellular information. For instance, the fibrosis staging is assigned from 0 to 6, corresponding from normal to cirrhosis, and is observed using Masson stained liver biopsy (see Table 1). The staging is based on the findings that whether portal areas have fibrous expansion, bridging with other portal areas, or bridging with central veins, or even cirrhosis.
According to the above application situation, in order to identify the fibrosis staging, it is necessary to first distinguish the central veins between portal areas. Generally, portal areas and central veins have different characteristics, and the most important is that bile ducts and arteries only appear in portal areas, as shown in Figure 1. Therefore, the findings of bile ducts and arteries in portal areas could solve the above distinction problem. Clinically, the bile ducts are considered as one of the criteria to identify portal areas. In addition, the detection of bile duct can also help diagnose bile duct cancer or vanishing bile duct syndrome. Therefore, in this study, a bile duct segmentation method is proposed to assist doctors and pathologists not only in staging the fibrosis score, but also in finding other bile duct diseases.
Automatic bile duct segmentation in liver pathology images, however, is challenging. First of all, since the color of the stained tissue gradually fades over time, and the staining protocols from different hospitals do not follow the same criteria, it is difficult to control the quality of the whole slide image (WSI) from liver biopsy. Secondly, even from the same liver biopsy image, different scanners can reveal color differences. The blurred area caused by the scanners and the fold region caused by the sectioning procedure also affect the display of the WSI. Most important of all, the bile ducts in different liver biopsies show variant types. The bile ducts from fibrous expansion, for instance, are thinner than normal bile ducts. The area and the number of bile ducts also differ in different portal areas. A larger portal area normally exists bile ducts clustering, or a huge bile duct with small ductules neighboring. Therefore, these factors increase the difficulty of automatic segmentation.
Accordingly, a multi-scale convolutional network architecture for the automatic segmentation of bile ducts in Masson stained from liver pathology images was proposed in this study. Three main concepts have been developed in the proposed model. First, the proposed method obtains features from both high-magnification and low-magnification patches in an attention method. The high-magnification view provides the microscopic feature but cannot provide enough macroscopic information, while low-magnification provides a larger view to solve the ambiguous features but loses the microscopic information. Hence, the multi-scale attention network solves the above problem by integrating low-magnification features and high-magnification features. Secondly, the integration is processed in an attention-grabbing way in order to avoid information received from low-magnification overly affecting the final segmentation result. Since the high-magnification patch is the main target to be segmented, the low-magnification features should be an assistant for the segmentation task. The patches extract from different magnifications are all generated in the same size, and the higher magnification patches are located in the central region of the lower magnification patches. Finally, the decoder structure in proposed model is to recover the lost information from low-level feature maps in encoder layer.
In order to verify the segmentation performance of the above design concepts in proposed model, two types of experiments were conducted in this study, including segmentation task compared with some existing models and the comparison of internal structural differences in proposed model. The experimental results demonstrated that the proposed model could improve the segmentation performance based on the above design concepts. From the measurement values and visual segmentation results, the contributions of the proposed model on bile duct segmentation from liver WSI would be addressed as follows:
  • Both cell-level and adjacent structural feature information are integrated to simulate pathologists’ examination procedure for observing bile ducts under different magnifications in liver WSI;
  • The attention method in proposed model would bring low-magnification input as an assistant help, enhancing the feature maps from high-magnification;
  • The decoder structure of proposed model could recover the lost information from low-level feature maps in encoder layer, resulting in the obviously smooth segmentation boundaries;
  • In clinical practice, the automatic analysis for pathologic examinations would contribute to the evaluation of the grading and staging of hepatitis based on Ishak Score system.

2. Related Works

Although there is still no research on the issue of automatic bile duct segmentation from liver pathology images, the segmentation and classification tasks focus on the medical filed have been increasing year by year, showing the appearance of novel methods in recent years significantly help in medical applications. Convolutional neural network is the main reason conduces to the development of image processing field. AlexNet [5], a classical convolutional neural network proposed by Krizhevsky et al., reached a new achievement in the ImageNet Large Scale Visual Recognition Challenge (LSVRC) in 2012, and is therefore leading to a great interest in convolutional networks among researchers. Compared to earlier image processing methods, convolutional neural networks have been improved more effectively, and therefore be applied in a large number of different fields including image classification [5,6], semantic segmentation [7], object detection [8], object generation [9], human pose estimation [10], diagnosis of diseases using EEG signals [11,12], and also an amount of applications in medical field.

2.1. Semantic Segmentation

The fully convolutional network (FCN) [7] is the first convolutional neural network structure applied to the semantic segmentation task and reached outstanding achievement. FCN provides an end-to-end training and pixel-wise predictions. Since the fully connected layer disrupt spatial information, the FCN architecture replaces all the fully connected layer by convolutional layer. The integration between different layers also helps segment the boundaries sharply. The FCN establish the foundation of semantic segmentation in deep learning. Accordingly, the U-net [13] architecture was proposed with a symmetric U-shaped encoder-decoder structure. In decoder, U-Net concatenates the layers from encoder at the same spatial resolution to recover lost information caused by pooling or convolution with stride 2, and gradually upsampled layer by layer to obtain finer boundaries. SegNet [14] also adopted a symmetric encoder-decoder structure similar to U-Net. In the encoder structure, SegNet stores the indices of the largest values in the max pooling layer. Then in the decoder, gradually upsample the layer and fills the value from the indices stored from corresponding positions in encoder. In recent years, DeepLab series [15,16,17,18] have been the standard for semantic segmentation network. The atrous spatial pyramid pooling (ASPP) is proposed to obtain the features at different receptive field to handle the problem of segmenting objects at multiple scales, and thus be widely applied in other segmentation networks. The comparison of above related works and proposed model is highlighted in Table 2.

2.2. Medical Image Task

In recent years, automatic analysis and diagnosis has attracted great interest in the medical field [19,20,21,22,23,24,25,26,27]. Due to the significant success in convolutional networks, deep learning has also been applied to medical images. Wang et al. [28] proposed an approach to combining both CNN and handcrafted features for mitosis detection, the dimensionality of handcrafted features was reduced with principal component analysis (PCA), and the features from two methods were classified by random forest classifiers. In lung cancer analysis, Cui et al. [29] applied the U-Net architecture for cancer cell detection, and an architecture based on VGG-Net with global average pooling to classify cells into different risk groups. In glioma research, Kurc et al. [30] applied a series of deep learning networks from digital pathology, the Mask-RCNN network with non-maximum suppression was adopted to nuclei segmentation in brain tissue images. Furthermore, the classification of brain cancer cases is applied in combining the predictions of both radiology image and pathology image, where a 3D CNN network is adopted as the radiology classification model and a DenseNet network pretrained on ImageNet is adopted as the histopathology classification model.
The multi-scale networks have also been applied in medical challenging tasks, since medical images is hard to predict in a single magnification. The information of different magnifications helps networks to obtain both local microscopic information and larger macroscopic information, and further help improve the prediction. In Liu et al. [31], a multi-scale approach applied with 40× magnification patches and lower magnification patches is proposed to classify breast cancer patches. Both inputs adopted Inception-v3 as the feature extraction model, and a fully connected layer is applied after two feature maps to combine features from different magnification information. Song et al. [32] also adopted a multi-scale approach to accurately segment cervical cytoplasm and nuclei. The multi-scale convolutional network is composed of multistage trainable architecture and each stage includes convolution layer, nonlinearity layer, and feature pooling layer. The proposed network is explored to extract scale invariant features, and segment regions centered at each pixel. Huang et al., and Sayıcı et al., both applied multi-scale magnification network to solve hepatocellular carcinoma classification task in H&E stained [33,34]. Huang et al., proposed a central region crop method to solve the alignment problem of local information in two different inputs to improve the performance, whereas Sayıcı et al., proposed an attention method to reweight the influences from different magnifications with softmax.
In this paper, a multi-scale input concept is adopted to obtain different fields of view information. The proposed multi-scale attention network integrates cell-level information and adjacent structural feature information for bile duct segmentation. In addition, the attention mechanism enables the network to focus the segmentation task on the input of high magnification, reducing the influence from low magnification input, but still helps to provide wider field of surrounding information. Our experimental result demonstrated that the multi-scale patches could improve the segmentation performance on Masson bile duct segmentation task, and our approach would achieve better performance of IOU and F1-score compared to other networks.

3. Materials and Methods

3.1. Multi-Magnification Patches

Clinically, pathologists and doctors examine the whole slide image on different magnifications to provide more spatial information, such as obtaining local information around the target object from high-magnification, or obtaining a global view of the entire tissue from low-magnification. Therefore, our study follows the same approach to feed images in multi-magnification, helping the model accurately segment bile ducts from additional local information.
In multi-magnification patches generation, an image pyramid building by applying bilinear interpolation is adopted to down sample the whole slide images (WSI) layer by layer from the highest magnification. After that, a sliding window fixed with size 256 × 256 moves from the whole slide image in different pyramid level to generate multi-magnification patches. Moreover, each patch generated in different pyramid level should center at the same location, higher magnification patches will locate in the central region of lower magnification patches. The process of multi-magnification patches generation is shown in Figure 2 and Figure 3.
While high-magnification patches provide detailed information from the object, low-magnification patches can obtain local information around the target to assist in the bile duct segmentation. However, the choosing of the two magnifications is important. Since bile ducts often be visible in a high-magnification level, a too much low-magnification input with useless background information will mislead the model to a wrong prediction. Therefore, the selection of input magnification has a significant influence on the model’s ability. In order to avoid the loss from resized patches in low-magnifications, in our experiment, we select the highest 40× magnification patches as the main input, meanwhile select 20× magnification patches as the secondary input to increase additional information.

3.2. Multi-Scale Attention Convolutional Network

Multi-scale Attention Convolutional Network (MACN) is an image segmentation architecture. The structure of the network is composed of three main concepts: (a) feature extraction from high-magnification and low-magnification images respectively by separate convolutional networks, (b) integration of the two magnification feature maps in an attention method, and (c) the decoder structure to recover the lost information from low-level feature maps in encoder layer.
According to the first concept of the present structure, the extraction applies two parallel convolutional neural networks, respectively. In this paper, the ResNet-101 [35] pretrained from ILSVRC-2012-CLS image classification dataset is utilized as the feature extraction network to reduce training time in our segmentation task. In order to further increase receptive field and obtain features from different scale of bile ducts, dense atrous spatial pyramid pooling block (DenseASPP) [36] is considered to be suitable framework, and hence concatenates after ResNet-101 block.
In feature map integration, the alignment between two different magnification is important. Since the spatial information from different magnification feature map does not align on the same location, straightly doing convolution operation on multi-magnification feature maps can leads to the spatial information disruption. The central crop method is adopted to improve this problem, spatially-constrained integration phase to align the location of each feature map element between low magnification feature maps and high magnification feature maps, as shown in Figure 4.
Since the high-magnification input is the target prediction, an attention method is proposed to prompt low-magnification input as an assistant help, enhancing the feature maps from high-magnification. After central crop and bilinear resize, the low-magnification input pass through a sigmoid function to converge weights between [0, 1]. The useful information in the nodes of low-magnification feature maps will activate and approach 1, indicating the corresponding positions in low-magnification have significant local information. Afterwards, an element-wised multiplication in the two feature maps is generated to produce an attention feature map. In the attention map, the value of the nodes in low-magnification feature map approaching 1 preserve the value corresponding to the same location from high-magnification feature map. Concretely, the multiplication retains the value of the nodes when values in both two magnification patches are activated, indicating that the location in both two patches exists useful information. Furthermore, since the activated nodes from low-magnification feature map multiply with the nodes from high-magnification feature map that are not activated will still remain low value, the attention method can sufficiently control the influence from low-magnification input, and therefore can be considered as an enhancement. Lastly, an element-wised addition between attention feature map and high-magnification feature map is adopted to enforce the values of high-magnification feature map. The attention method can be defined as:
Y ( x ) = ( 1 + σ ( L ( x ) )   ) H ( x )
where Y(x) is the output, σ is a sigmoid function, L(x) and H(x) are low-magnification and high-magnification feature maps, respectively. Figure 5 and Figure 6 shows the whole process of the attention method.
In order to refine the segmentation results especially along object boundaries. The decoder structure similar to [13] and [18] is also applied. The decoder module that gradually recovers the spatial information obtains sharper boundaries. In detail, the features after attention method are bilinearly upsampled by a factor of 2 and then concatenated with the corresponding low-level features from the ResNet-101 network, then apply convolution layers to merge the features.

3.3. Focal Loss

In this study, we change the traditional loss method from cross-entropy to focal loss [37], which get a great success in object detection. The cross-entropy loss can be defined as:
C E ( p t ) = y l o g ( p t )
where p t is the probability predicted from model, and y is the ground truth.
Focal loss is extended from cross entropy to focus the loss calculation in challenging data, the equation can be defined as:
F L ( p t ) = y ( 1 p t ) r l o g ( p t )
where r is a hyperparameter default as 2.
The equation of focal loss will reduce the influence of easy prediction object. For instance, if a prediction of probability from model is 0.9 and the label is 1, it indicates the pixel is easy to predict, and thus decrease its influence on the model weights update, lowering 100× loss compared with cross-entropy ( r = 2). The challenging task will significantly affect the loss. Focal loss adopted in medical images also have a big influence since medical images are various, so we changed the traditional cross-entropy method to focal loss to help correct prediction.

3.4. Basic Model Operations

In this section, the basic model operations such as activation function or batch normalization employed in our proposed models are described more specifically in the following section parts.

3.4.1. Activation Function

The activation function is usually an abstraction concept inspired by biological neural network, representing the activation of cells. In artificial neural networks, we generally add activation functions to simulate the operation of biological neurons and generate non-linear transformations to increase model complexity. The Rectified Linear Unit (ReLU) [38] has been widely used as an activation function in neural network structure, providing the sparsity and preventing model from vanishing gradient problems compared to other activation functions, like sigmoid or hyperbolic tangent. The Rectified Linear Unit (ReLU) function can be defined as:
ReLU ( x ) = max   ( 0 , x )
where x is the summation of weights from previous layers. In our model, we apply ReLU as the default activation function after the convolution layer and the fully connected layer.
The SoftMax function is a generalization of the logistic function that “squashes” the value of a vector into range [0, 1], and all the values add up to 1. The output of SoftMax function is used to represent a probability distribution in probability theory. That is to say, the output can be considered as the individual probability of different input values in the vector, making SoftMax function widely used in the final output layer of neural networks. The predicted probability of the j_th class from SoftMax function can be defined as:
P ( y = j | x ) = e x T w j k = 1 K e x T w k
where K is the total number of the class, and x is a vector of output layer in neural network.

3.4.2. Batch Normalization

Batch normalization [39] is a hidden layer feature normalization approach. The approach id applied to avoid covariate shift problem at the hidden layers. Covariate shift problem means that the distribution of the different data features has a large variance in the training phase. If the network goes deeper, the problem will be spread in the deep layers. The batch normalization forces the mean of the batch data features be 0 and the standard deviation of the batch data features be 1, and it makes the training more stable. The batch normalization equation B N r , β ( x i ) can be defined as:
μ β = 1 m i = 1 m x i
σ β 2 = 1 m i = 1 m ( x i   μ β ) 2  
x ^ l = x i   μ β   σ β 2 + ε
B N γ ,     β ( x i ) = γ x ^ l + β
where m is batch size, x i is i t h data in batch, μ β is the mean of the batch, σ β 2 is the standard deviation of the batch, x ^ l is the i t h normalized value in the batch, ε is a small constant for numerical stability and the γ and β are trainable parameters for distribution scale and shift.

4. Results

4.1. Masson Stained Dataset

The dataset used in this study was obtained from the Department of Pathology, National Cheng Kung University Hospital approved by the Institutional Review Board and diagnosed by experienced pathologist. The scanners currently serving in the hospital is captured as a WSI at 40× magnification using an Aperio AT2–Digital Whole Slide Scanner (Leica Biosystems Imaging, Inc., Wetzlar, Hesse, Germany). In the semantic segmentation of the WSIs, the dataset contains 60 whole slide images from cases with diagnosis confirmed by specialist. The training set includes 32 cases and the testing set contains 28 cases. In the experiments, after pyramid patch generation, a total of 215,607 patches were generated with each patch resolution fixed at 256 × 256. However, most of these patches are segmented from hepatocyte areas, which might result in an imbalance for our training and thus influence the prediction capability. To reduce the proportion of hepatocyte patches in total dataset, patches that include fibrosis area, artery, vein and bile duct was retained. With this strategy, the total patches were reduced to 40,120 in the dataset and 23,822 and 17,298 patches were included as training and testing set, respectively. The bile duct pixels were labeled as foreground, and others as the background.

4.2. Evaluation

In the experiment, the proposed segmentation network was compared to recently segmentation networks. The experiment was designed to evaluate the performance of each segmentation network in the histopathological image dataset.
In this study, Precision (P), Recall (R), and F1-score were used to quantify the segmentation performance. The aim of the proposed method is to segment the bile duct from each patch. To this end, a pixel segmentation point is considered as true-positive (TP) if the point is located within bile duct. In contract, a pixel segmentation point is considered as false-positive (FP) if the point is in other areas such as fibrosis, artery, vein and hepatocyte. The TP, FP, TN and FN were formulated in Table 3.
According to the above definition, equations that include true-negative (TN) were excluded because most of the pixels are predicted as negative. Hence, the main criteria for evaluation are defined as follows:
Precision   ( P ) = TP TP + FP
Recall ( R ) = TP TP + FN
F 1   score   = 2 R P R + P
IOU   = GroundTruth     DetectionResult GroundTruth     DetectionResult = TP TP + FP + FN
Table 4 shows the performance between the proposed network and other segmentation networks. The performance of precision and F1-score and IOU in our proposed network are higher than others in testing cases, while recall value is similar to DeepLabv3 and DeepLabv3-plus. The results suggest that capability of these networks to identify bile duct is similar. However, our proposed network has an advantage of diminishing false positive predictions.
In addition to the above experiment of proposed and other segmentation networks, comparison of internal structural differences in proposed model was also conducted, including the performance between attention mechanism and concatenation, crop/without crop the central region in the feature map integration, integration the feature map after DenseASPP block or after ASPP block, model with/without decoder, and the performance of loss methods between cross-entropy and focal loss.

5. Discussion

5.1. Visual Experiment Results

The visual segmentation results of existing and proposed models are shown in Figure 7, Figure 8, Figure 9, Figure 10, Figure 11 and Figure 12. As shown in Figure 7, our proposed network correctly segmented the clearly visible bile ducts and significantly reduced incorrect predictions in hepatocyte area in comparison with other networks. As for the case with small bile ductules, the visual segmentation results in Figure 8 illustrated that the proposed model could accurately segment the small bile ductules with reduced incorrect predictions in hepatocyte area. The importance of decoder structure can be illustrated in Figure 9 in which the object boundaries are smoother in our network. In addition, the proposed network also demonstrated the ability to find small ductules stained similar to hepatocyte area in Figure 9. Figure 10 demonstrated a difficult case of segmentation task in which cells in arteries are much similar to real bile ducts. From the visual segmentation results in Figure 10, all other compared models gave erroneous predictions on these arteries, but our network had fewer incorrect predictions than others. In Figure 11, compared to other networks, our network segmented the big bile duct and bile ductules near hepatocyte area more completely. Figure 12 shows a special case of bile duct. The type of the bile duct in Figure 12 is between the hepatocyte area and the classic bile ducts, and thus is difficult to be segmented correctly. From the visual segmentation results in Figure 12, our model achieved a smoother segmentation result. Since the types of bile ducts and ductules vary between different cases, the segmentation results of FCN, U-Net, SegNet are fragmented and incomplete, while the DeepLab series and the proposed network perform more completely. Furthermore, the proposed network had a better ability to reduce segmentation errors in hepatocyte areas, arteries and fibrosis areas.
As for the comparison of internal structural differences in proposed model, Table 5 shows that in the model without crop, the local information of low magnification input will disrupt spatial information, resulting in increasing false positive predictions in wrong area and declining the precision value. Similarly, simply concatenating two different magnification would increase more unnecessary information from high magnification input. Hence, the model with attention method to focus the information on high magnification could significantly improve the performance, as shown in Table 6.
Table 7 shows the comparison between the performance of model with DenseASPP block and with ASPP block. The performance in both methods is similar, but the model with DenseASPP block would perform better owing to the increase of complexity and receptive field. Table 8 shows the performance between model with and without decoder. Although the advantage of decoder structure does not show a significant influence on the values of evaluation criterion compared to the structure without decoder, it does obviously smooth the segmentation boundaries that could be observed in the visible results.
Lastly, the performance comparison of focal loss with different gamma values is shown in Table 9. Since gamma value definitely decreases the total loss value, the higher gamma value will get lower performance in which total loss is too low to modify model weights.

5.2. Limitations of the Study in Digital Pathologyl Image Analysis

The clinical diagnosis of liver diseases is generally performed with liver biopsy since pathologists could examine the liver biopsy at different magnifications under microscope or digital image for more detailed and precise information at both structural and cellular levels. This labor intensive and expensive examination process, however, would waste lots of pathologists’ working time. In order to reduce the intensive labor costs and potential human diagnose error, automatic diagnosis and assistance for pathologists are extremely important. Thus, the proposed model could provide efficient and precise segmentation results for pathologists to distinguish bile ducts in WSI.
Owing to that the labeling job for digital pathological image is time consuming and the digital pathological image has features of huge file size (over 10 GB in average), it is quite difficult to obtain open dataset of digital pathological images labeled by pathologists. Thus, to obtain more precise segmentation results it would be necessary to increase the labeled cases and/or to perform post-processing mechanism to reduce the false-positive rate and keep the IoU score. In spite of these limitations of image segmentation task on digital pathological images, it would be great helpful assistance in clinical practice when adopting suitable and advanced AI model in digital pathology examinations to provide more efficient and precise segmentation results for pathologists.

6. Conclusions

This paper presents a multi-scale deep convolution neural network with integration of different magnification feature maps in an attention method. Three main concepts have been developed in our model: (a) feature extraction from high-magnification and low-magnification images respectively by separate convolutional networks, (b) integration of the two magnification feature maps in an attention method, and (c) the decoder structure to recover the lost information from low-level feature maps in encoder layer. As shown in our results, multi-scale deep attention neural network demonstrates the multi-scale patches that could improve the segmentation performance on Masson bile duct segmentation task. Furthermore, 72.5% IOU and 84.1% F1-score were achieved in our approach, better than other networks reported earlier.
Although the labeling job for digital pathological image is time consuming, we are still ongoing to increase the labeled case to improve the accuracy of the proposed model. In addition, a new post-processing mechanism would be tried to reduce the false-positive rate and keep the IoU score. The post-processing mechanism in development is to provide confirmed patches and candidate for doctors to quickly filter the candidate patches.

Author Contributions

Conceptualization, C.-H.S. and P.-C.C.; methodology, C.-H.S., P.-C.C., S.-F.L., H.-W.T., T.-L.Y. and Y.-C.S.; software, C.-H.S.; validation, C.-H.S., P.-C.C. and H.-W.T.; formal analysis, C.-H.S., P.-C.C. and H.-W.T.; investigation, C.-H.S.; resources, S.-F.L., H.-W.T., T.-L.Y. and Y.-C.S.; data curation, S.-F.L., H.-W.T., T.-L.Y. and Y.-C.S.; writing—original draft preparation, C.-H.S.; writing—review and editing, P.-C.C. and Y.-C.S.; visualization, C.-H.S.; supervision, P.-C.C.; project administration, P.-C.C.; funding acquisition, P.-C.C., S.-F.L. and Y.-C.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Ministry of Science and Technology (MOST), Taiwan (Grant number: MOST 110-2634-F-006-022), and by E-DA hospital (Grant number: EDAHP110045).

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Institutional Review Board of National Cheng Kung University Hospital (A-ER-106-233 on 27 July 2021).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Not applicable.

Acknowledgments

We thank to National Center for High-performance Computing (NCHC) for providing computational and storage resources. We thank to National Core Facility for Biopharmaceuticals (MST 109-2740-B-492-001) for support and National Center for High-performance Computing (NCHC) of National Applied Research Laboratories (NARLabs) in Taiwan for providing computational and storage resources.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. World Health Organization. Global Hepatitis Report 2017; World Health Organization: Geneva, Switzerland, 2017. [Google Scholar]
  2. Asrani, S.K.; Devarbhavi, H.; Eaton, J.; Kamath, P.S. Burden of liver diseases in the world. J. Hepatol. 2019, 70, 151–171. [Google Scholar] [CrossRef]
  3. Villanueva, A. Hepatocellular Carcinoma. N. Engl. J. Med. 2019, 380, 1450–1462. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Ishak, K.; Baptista, A.; Bianchi, L.; Callea, F.; De Groote, J.; Gudat, F.; Denk, H.; Desmet, V.; Korb, G.; Macsween, R.N.; et al. Histological grading and staging of chronic hepatitis. J. Hepatol. 1995, 22, 696–699. [Google Scholar] [CrossRef]
  5. Krizhevsky, A.; Ilya, S.; Geoffrey, E.H. Imagenet classification with deep convolutional neural networks. In Proceedings of the 26th Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–8 December 2012. [Google Scholar]
  6. Simonyan, K.; Andrew, Z. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  7. Long, J.; Evan, S.; Trevor, D. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  8. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
  9. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 2, 2672–2680. [Google Scholar]
  10. Toshev, A.; Christian, S. Deeppose: Human pose estimation via deep neural networks. IEEE Conf. Comput. Vis. Pattern Recognit. arXiv 2014, arXiv:1312.4659. [Google Scholar]
  11. Shoeibi, A.; Sadeghi, D.; Moridian, P.; Ghassemi, N.; Heras, J.; Alizadehsani, R.; Khadem, A.; Kong, Y.; Nahavandi, S.; Zhang, Y.; et al. Automatic Diagnosis of Schizophrenia in EEG Signals Using CNN-LSTM Models. Front. Neuroinform. 2021, 15, 777977. [Google Scholar] [CrossRef]
  12. Shoeibi, A.; Ghassemi, N.; Khodatars, M.; Moridian, P.; Alizadehsani, R.; Zare, A.; Khosravi, A.; Subasi, A.; Acharya, U.R.; Gorriz, J.M. Detection of epileptic seizures on EEG signals using ANFIS classifier, autoencoders and fuzzy entropies. Biomed. Signal. Process. Control. 2022, 73, 103417. [Google Scholar] [CrossRef]
  13. Ronneberger, O.; Philipp, F.; Thomas, B. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Cham, Switzerland. [Google Scholar]
  14. Badrinarayanan, V.; Alex, K.; Roberto, C. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  15. Chen, L.-C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Semantic image segmentation with deep convolutional nets and fully connected crfs. arXiv 2014, arXiv:1412.7062. [Google Scholar]
  16. Chen, L.-C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Chen, L.-C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
  18. Chen, L.-C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018. [Google Scholar]
  19. Haris, K.; Efstratiadis, S.N.; Maglaveras, N.; Katsaggelos, A.K. Hybrid image segmentation using watersheds and fast region merging. IEEE Trans. Image Process. 1998, 7, 1684–1699. [Google Scholar] [CrossRef] [Green Version]
  20. Wu, H.S.; Xu, R.; Harpaz, N.; Burstein, D.; Gil, J. Segmentation of microscopic images of small intestinal glands with directional 2-d filters. Anal. Quant. Cytol. Histol. 2005, 27, 291. [Google Scholar]
  21. Matula, P.; Kumar, A.; Wörz, I.; Erfle, H.; Bartenschlager, R.; Eils, R.; Rohr, K. Single-cell-based image analysis of high-throughput cell array screens for quantification of viral infection. Cytom. Part A J. Int. Soc. Adv. Cytom. 2009, 75, 309–318. [Google Scholar] [CrossRef]
  22. Simsek, A.C.; Tosun, A.B.; Aykanat, C.; Sokmensuer, C.; Gunduz-Demir, C. Multilevel segmentation of histopathological images using cooccurrence of tissue objects. IEEE Trans. Biomed. Eng. 2012, 59, 1681–1690. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Veta, M.; van Diest, P.J.; Kornegoor, R.; Huisman, A.; Viergever, M.A.; Pluim, J.P.W. Automatic nuclei segmentation in H&E stained breast cancer histopathology images. PLoS ONE 2013, 8, e70221. [Google Scholar]
  24. Irshad, H.; Veillard, A.; Roux, L.; Racoceanu, D. Methods for nuclei detection, segmentation, and classification in digital histopathology: A review—Current status and future potential. IEEE Rev. Biomed. Eng. 2013, 7, 97–114. [Google Scholar] [CrossRef] [PubMed]
  25. Kong, J.; Cooper, L.A.D.; Wang, F.; Gao, J.; Teodoro, G.; Scarpace, L.; Mikkelsen, T.; Schniederjan, M.J.; Moreno, C.S.; Saltz, J.H.; et al. Machine-based morphologic analysis of glioblastoma using whole-slide pathology images uncovers clinically relevant molecular correlates. PLoS ONE 2013, 8, e81049. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Webster, J.D.; Dunstan, R.W. Whole-slide imaging and automated image analysis: Considerations and opportunities in the practice of pathology. Vet. Pathol. 2014, 51, 211–223. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Xing, F.; Lin, Y. Robust nucleus/cell detection and segmentation in digital pathology and microscopy images: A comprehensive review. IEEE Rev. Biomed. Eng. 2016, 9, 234–263. [Google Scholar] [CrossRef] [PubMed]
  28. Wang, H.; Cruz-Roa, A.; Basavanhally, A.; Gilmore, H.; Shih, N.; Feldman, M.; Tomaszewski, J.; Gonzalez, F.; Madabhushi, A. Mitosis detection in breast cancer pathology images by combining handcrafted and convolutional neural network features. J. Med. Imaging 2014, 1, 034003. [Google Scholar] [CrossRef] [PubMed]
  29. Cui, L.; Li, H.; Hui, W.; Chen, S.; Yang, L.; Kang, Y.; Bo, Q.; Feng, J. A deep learning-based framework for lung cancer survival analysis with biomarker interpretation. BMC Bioinform. 2020, 21, 1–14. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  30. Kurc, T.; Bakas, S.; Ren, X.; Bagari, A.; Momeni, A.; Huang, Y.; Zhang, L.; Kumar, A.; Thibault, M.; Qi, Q. Segmentation and Classification in Digital Pathology for Glioma Research: Challenges and Deep Learning Approaches. Front. Neurosci. 2020, 14, 27. [Google Scholar] [CrossRef] [Green Version]
  31. Liu, Y.; Gadepalli, K.; Norouzi, M.; Dahl, G.E.; Kohlberger, T.; Boyko, A.; Venugopalan, S.; Timofeev, A.; Nelson, P.Q.; Corrado, G.S.; et al. Detecting cancer metastases on gigapixel pathology images. arXiv 2017, arXiv:1703.02442. [Google Scholar]
  32. Song, Y.; Zhang, L.; Chen, S.; Ni, D.; Lei, B.; Wang, T. Accurate segmentation of cervical cytoplasm and nuclei based on multiscale convolutional network and graph partitioning. IEEE Trans. Biomed. Eng. 2015, 62, 2421–2433. [Google Scholar] [CrossRef] [PubMed]
  33. Huang, W.C.; Chung, P.C.; Tsai, H.W.; Chow, N.H.; Juang, Y.Z.; Tsai, H.H.; Lin, S.H.; Wang, C.H. Automatic HCC detection using convolutional network with multi-magnification input images. In Proceedings of the IEEE International Conference on Artificial Intelligence Circuits and Systems, Hsinchu, Taiwan, 18–20 March 2019. [Google Scholar]
  34. Sayıcı, M.B.; Rikiya, Y.; Jeanne, S. Analysis of Multi Field of View Cnn and Attention Cnn on H&E Stained Whole-Slide Images on Hepatocellular Carcinoma. arXiv 2020, arXiv:2002.04836. [Google Scholar]
  35. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference On Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  36. Yang, M.; Yu, K.; Zhang, C.; Li, Z.; Yang, K. Denseaspp for semantic segmentation in street scenes. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  37. Lin, T.; Goyal, P.; Girshick, R.; He, K.; Dollar, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017. [Google Scholar]
  38. Vinod, N.; Hinton, G.E. Rectified linear units improve restricted boltzmann machines. In Proceedings of the 27th International Conference on Machine Learning, Haifa, Isreal, 21–24 June 2010. [Google Scholar]
  39. Sergey, I.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv 2015, arXiv:1502.03167. [Google Scholar]
Figure 1. Structure of portal area in Masson stain.
Figure 1. Structure of portal area in Masson stain.
Sensors 22 02679 g001
Figure 2. Multi-magnification patches generated by sliding window.
Figure 2. Multi-magnification patches generated by sliding window.
Sensors 22 02679 g002
Figure 3. Patches from different magnifications. The higher magnification patch is located in the central region of the lower magnification patch.
Figure 3. Patches from different magnifications. The higher magnification patch is located in the central region of the lower magnification patch.
Sensors 22 02679 g003
Figure 4. The central crop method to align different magnification feature maps.
Figure 4. The central crop method to align different magnification feature maps.
Sensors 22 02679 g004
Figure 5. An attention method to enhance the high-magnification feature map.
Figure 5. An attention method to enhance the high-magnification feature map.
Sensors 22 02679 g005
Figure 6. The attention method visualization.
Figure 6. The attention method visualization.
Sensors 22 02679 g006
Figure 7. Segmentation results for clearly visible bile ducts: original (a), GT (b), FCN (c), SegNet (d), U-Net (e), Deeplabv3 (f), Deeplabv3-plus (g), and the proposed network (h).
Figure 7. Segmentation results for clearly visible bile ducts: original (a), GT (b), FCN (c), SegNet (d), U-Net (e), Deeplabv3 (f), Deeplabv3-plus (g), and the proposed network (h).
Sensors 22 02679 g007
Figure 8. Segmentation results for small bile ducts: original (a), GT (b), FCN (c), SegNet (d), U-Net (e), Deeplabv3 (f), Deeplabv3-plus (g), and the proposed network (h).
Figure 8. Segmentation results for small bile ducts: original (a), GT (b), FCN (c), SegNet (d), U-Net (e), Deeplabv3 (f), Deeplabv3-plus (g), and the proposed network (h).
Sensors 22 02679 g008
Figure 9. Segmentation results for small ductules stained similar to hepatocyte area: original (a), GT (b), FCN (c), SegNet (d), U-Net (e), Deeplabv3 (f), Deeplabv3-plus (g), and the proposed network (h).
Figure 9. Segmentation results for small ductules stained similar to hepatocyte area: original (a), GT (b), FCN (c), SegNet (d), U-Net (e), Deeplabv3 (f), Deeplabv3-plus (g), and the proposed network (h).
Sensors 22 02679 g009
Figure 10. Segmentation results for the case in which cells in arteries are much similar to real bile ducts: original (a), GT (b), FCN (c), SegNet (d), U-Net (e), Deeplabv3 (f), Deeplabv3-plus (g), and the proposed network (h).
Figure 10. Segmentation results for the case in which cells in arteries are much similar to real bile ducts: original (a), GT (b), FCN (c), SegNet (d), U-Net (e), Deeplabv3 (f), Deeplabv3-plus (g), and the proposed network (h).
Sensors 22 02679 g010
Figure 11. Segmentation results for big bile ducts and bile ductules near hepatocyte area: original (a), GT (b), FCN (c), SegNet (d), U-Net (e), Deeplabv3 (f), Deeplabv3-plus (g), and the proposed network (h).
Figure 11. Segmentation results for big bile ducts and bile ductules near hepatocyte area: original (a), GT (b), FCN (c), SegNet (d), U-Net (e), Deeplabv3 (f), Deeplabv3-plus (g), and the proposed network (h).
Sensors 22 02679 g011
Figure 12. Segmentation results for the case in which the bile duct is between the hepatocyte area and the classic bile ducts: original (a), GT (b), FCN (c), SegNet (d), U-Net (e), Deeplabv3 (f), Deeplabv3-plus (g), and the proposed network (h).
Figure 12. Segmentation results for the case in which the bile duct is between the hepatocyte area and the classic bile ducts: original (a), GT (b), FCN (c), SegNet (d), U-Net (e), Deeplabv3 (f), Deeplabv3-plus (g), and the proposed network (h).
Sensors 22 02679 g012
Table 1. Ishak fibrosis staging.
Table 1. Ishak fibrosis staging.
StatusScore
No fibrosis0
Fibrous expansion of some portal areas, with or without short fibrous septa1
Fibrous expansion of most portal areas, with or without short fibrous septa2
Fibrous expansion of most portal areas with occasional portal to portal (P-P) bridging3
Fibrous expansion of portal areas with marked bridging; portal to portal (P-P) as well as portal to central (P-C)4
Marked bridging (P-P and/or P-C) with occasional nodules (incomplete cirrhosis)5
Cirrhosis probable or definite6
Table 2. Comparison of related works and proposed model.
Table 2. Comparison of related works and proposed model.
ComparisonStrengthsWeaknesses
Model
FCNfast learning and inferencesegmentation performance
U-netfiner segmentation boundary; training with very few imageshigh computational resources
SegNetlow computational memoryhigh computational time
DeepLabv3augmenting ASPP for better performancepoor segmentation boundary
DeepLabv3-plusfiner segmentation boundaryhigh computational resources
proposedfiner segmentation boundary; high segmentation performancehigh computational resources
Table 3. TP, FP, TN and FN performance metrics.
Table 3. TP, FP, TN and FN performance metrics.
Ground TruthPositiveNegative
Prediction
PositiveTrue positive (TP)False positive (FP)
NegativeFalse negative (FN)True negative (TN)
Table 4. Performance between semantic segmentation networks.
Table 4. Performance between semantic segmentation networks.
EvaluationPrecisionRecallF1-ScoreIOU
Model
FCN [7]0.8180.7700.7930.657
SegNet [14]0.6370.6780.6570.489
U-net [13]0.5900.7230.6510.482
DeepLabv3 [17]0.7690.8570.8100.681
DeepLabv3-plus [18]0.7750.8550.8130.685
Proposed0.8240.8580.8410.725
Table 5. Performance between model with crop and without crop.
Table 5. Performance between model with crop and without crop.
EvaluationPrecisionRecallF1-ScoreIOU
Method
With crop (Proposed)0.8240.8580.8410.725
Without crop0.7600.8660.8100.680
Table 6. Performance between model with attention and with concatenation.
Table 6. Performance between model with attention and with concatenation.
EvaluationPrecisionRecallF1-ScoreIOU
Method
Attention (Proposed)0.8240.8580.8410.725
Concatenation0.7720.8500.8090.680
Table 7. Performance between model with DenseASPP and with ASPP.
Table 7. Performance between model with DenseASPP and with ASPP.
EvaluationPrecisionRecallF1-ScoreIOU
Method
DenseASPP (Proposed)0.8240.8580.8410.725
ASPP0.8380.8380.8380.722
Table 8. Performance between model with decoder and without decoder.
Table 8. Performance between model with decoder and without decoder.
EvaluationPrecisionRecallF1-ScoreIOU
Method
With decoder (Proposed)0.8240.8580.8410.725
Without decoder0.8220.8520.8370.720
Table 9. Performance of focal loss between different gamma values.
Table 9. Performance of focal loss between different gamma values.
EvaluationPrecisionRecallF1-ScoreIOU
Method
gamma=0 (Cross Entropy)0.8060.8520.8290.707
gamma = 1 (Proposed)0.8240.8580.8410.725
gamma = 20.8230.8400.8320.712
gamma = 30.8070.8370.8220.698
gamma = 40.7750.8390.8060.675
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Su, C.-H.; Chung, P.-C.; Lin, S.-F.; Tsai, H.-W.; Yang, T.-L.; Su, Y.-C. Multi-Scale Attention Convolutional Network for Masson Stained Bile Duct Segmentation from Liver Pathology Images. Sensors 2022, 22, 2679. https://doi.org/10.3390/s22072679

AMA Style

Su C-H, Chung P-C, Lin S-F, Tsai H-W, Yang T-L, Su Y-C. Multi-Scale Attention Convolutional Network for Masson Stained Bile Duct Segmentation from Liver Pathology Images. Sensors. 2022; 22(7):2679. https://doi.org/10.3390/s22072679

Chicago/Turabian Style

Su, Chun-Han, Pau-Choo Chung, Sheng-Fung Lin, Hung-Wen Tsai, Tsung-Lung Yang, and Yu-Chieh Su. 2022. "Multi-Scale Attention Convolutional Network for Masson Stained Bile Duct Segmentation from Liver Pathology Images" Sensors 22, no. 7: 2679. https://doi.org/10.3390/s22072679

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop