Next Article in Journal
The Dislocation- and Cracking-Mediated Deformation of Single Asperity GaAs during Plowing Using Molecular Dynamics Simulation
Next Article in Special Issue
Monolayer MoS2-Based Flexible and Highly Sensitive Pressure Sensor with Wide Sensing Range
Previous Article in Journal
Quality Study and Numerical Simulation Analysis of Solid–Liquid Two-Phase Magnetic Fluid Polishing Seven-Order Variable-Diameter Pipe
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Classification Framework of Tongue Feature Based on Convolutional Neural Networks

Key Laboratory of Instrumentation Science & Dynamic Measurement, North University of China, Taiyuan 030051, China
*
Author to whom correspondence should be addressed.
Micromachines 2022, 13(4), 501; https://doi.org/10.3390/mi13040501
Submission received: 1 March 2022 / Revised: 21 March 2022 / Accepted: 22 March 2022 / Published: 24 March 2022
(This article belongs to the Special Issue Flexible and Wearable Sensors)

Abstract

:
Tongue diagnosis is an important part of the diagnostic process in traditional Chinese medicine (TCM). It primarily relies on the expertise and experience of TCM practitioners in identifying tongue features, which are subjective and unstable. We proposed a tongue feature classification framework based on convolutional neural networks to reduce the differences in diagnoses among TCM practitioners. Initially, we used our self-designed instrument to capture 482 tongue photos and created 11 data sets based on different features. Then, the tongue segmentation task was completed using an upgraded facial landmark detection method and UNET. Finally, we used ResNet34 as the backbone to extract features from the tongue photos and classify them. Experimental results show that our framework has excellent results with an overall accuracy of over 86 percent and is particularly sensitive to the corresponding feature regions, and thus it could assist TCM practitioners in making more accurate diagnoses.

1. Introduction

Traditional Chinese medicine (TCM) has received increasing attention and acknowledgment from medical experts since the World Health Organization (WHO) included it in the latest global medical guidelines [1,2]. Tongue diagnosis is a key criterion for TCM diagnosis; it is a noninvasive and convenient approach to assess human health [3,4]. However, tongue features are varied and intertwined, and subtle differences in features may correspond to completely different diseases. TCM practitioners rely on their expertise and previous experience to identify tongue features, which are susceptible to environmental factors, resulting in unstable and inaccurate diagnoses. Therefore, building a system that can objectively classify tongue features is urgently required.
Generally, tongue feature classification refers to identifying the types of different tongue features and assisting TCM practitioners to confirm diagnoses. Accurate classification is difficult due to the minor differences within the inner class. In addition, this classification mainly extracts and classifies the feature vector of the tongue body images, rather than the entire tongue regions. However, the tongue image captured by the camera has other information including lips, teeth, and even facial images. Image segmentation should be used to eliminate useless information, which can increase the accuracy of the classification and reduce the amount of calculation. Therefore, objectification and automated tongue diagnosis are mainly composed of two tasks; tongue image segmentation and tongue feature classification. Segmenting the tongue from the image is a prerequisite for tongue feature classification [5].
For tongue segmentation, traditional methods are generally used, including the region growing method, threshold method, watershed transformation method, edge detection, and snake model [6,7,8,9,10,11]. Nevertheless, the color around the tongue is similar to the color of the tongue’s body, and the edge contour is relatively fuzzy. The application of traditional methods to achieve the effect of tongue image segmentation is not ideal. In recent years, more studies on using semantic segmentation based on deep convolutional neural networks (CNN) for tongue segmentation have been conducted, and the effect is better than some traditional image segmentation methods [12,13]. However, most training data sets used only contain normal tongue images, thereby reducing the clinical practicability and robustness of the segmentation algorithm.
In the research of tongue classification, many scholars digitize tongue features and classify them according to particular threshold standards [14,15]. Nevertheless, the accuracy of recognition is unsatisfactory. A number of scholars have also developed classification models for various tongue features using CNN. They use convolution operations to extract high-level semantic features and support vector machine (SVM), softmax, or other classifiers for classification. These approaches have been shown to be effective in the recognition of various features of tongue texture, tongue coating, and tongue shape [3,16,17,18]. These studies, however, only focus on a few features and do not systematically obtain all tongue features for analysis based on the theory of TCM tongue diagnosis; thus, the results lack clinical significance.
In this paper, we use the CNN approach to deal with these limitations. The main contribution and innovation of this paper is the optimization of the usability, efficiency, and interpretability of the tongue diagnosis classification method. First, our method is to classify 11 tongue features according to TCM theory, which can identify more comprehensive tongue features and is consistent with TCM theory to assist practitioners in making more standard diagnoses. Therefore, a clinical and standard data set with 11 tongue features was constructed due to the lack of a relevant open-source data set. In contrast, most previous studies have classified one or a few features, which are not applicable to clinical use. Second, we propose a tongue region extraction method, which can convert the original image of the tongue into a smaller pixel image before segmentation and classification to improve the efficiency and accuracy of segmentation and classification. Finally, we use GradCAM to visualize the decisions of the classification network, enabling the user to know where anomalous features appear and to better understand the basis of the classification. The proposed method is described in detail in Section 2, and the experimental results are presented in Section 3. Section 4 concludes this study.

2. Related Works

The main challenge in the domain of tongue diagnosis classification systems is to digitize the diagnostic experience of TCM practitioners in order to allow models to make as clinically meaningful and accurate diagnoses as TCM practitioners do. Such models are generally composed of two parts: tongue segmentation and classification of tongue features.
In the field of tongue segmentation, Li et al. [19] transformed the tongue image into HSI color space and used thresholding to obtain the initial region of the tongue body, and then used the tongue root and upper lip gap region to remove the irrelevant region to achieve the final segmentation. Shi et al. [20] combined the geometrical Snake model with the parameterized GVFSnake model to establish the C2G2FSnake tongue segmentation model. They introduced color space information as control information to update the external force parameters to control the accuracy and velocity of the curve, and then applied the gradient vector flow field (GVF) to obtain the results. However, these methods based on traditional image processing techniques will occasionally judge similarly colored regions around the tongue as tongue pixels as well, and the segmentation edge is rough. The segmentation effect of the traditional method is not ideal. In [12,13,21,22], semantic segmentation methods such as FCN, SegNet, and ResNet were applied to tongue segmentation. These methods using CNN for feature extraction can precisely distinguish the difference between tongue pixels and irrelevant pixels, and the contour of the segmentation result is smooth. Also, these methods do not require manual assistance at all. However, the data sets referenced by these methods are normal tongue images and lack clinicality. Therefore, this paper uses semantic segmentation based on CNN for the tongue segmentation and uses clinical tongue images as the data set for training, which is more clinically relevant.
In the field of tongue feature classification, Zhang et al. [14] extracted the tongue measurements, distances, areas, and their proportions from the tongue image and classified the tongue shape using a decision tree. Bo et al. [15] corrected the tongue deflection by applying three geometric criteria and then classified tongue shapes by analytic hierarchy process according to seven geometric features defined by various measurements of length, area, and angle of the tongue. These methods are used to convert specific tongue features into digital quantities by corresponding image processing, and then classifiers such as decision tree or SVM are used to achieve feature classification. However, the classification accuracy of these methods is not ideal, and it can only classify a few specific features, which is not very practical. Li et al. [3] proposed a method for the recognition of tooth marks. The method first generates suspicious regions based on concavity information, then extracts the depth features within the regions by CNN, and finally uses multiple-instance SVM to complete the final classification. Tang et al. [16] selected suspected rotten-greasy tongue coating patches firstly, and then used ResNet to extract features of each patch to complete the classification of the curdy or greasy coating by MI-SVM. These tongue feature classification methods based on deep learning can automatically extract more detailed and relevant features by CNN for different tongue features with higher accuracy. However, these classification methods currently classify only a few tongue features and cannot effectively assist physicians in clinical diagnosis. In addition, interpretability is very important for medical diagnosis results, but these methods are to draw direct conclusions without explainable process. Therefore, we built a classification model for 11 tongue features using CNN and visualized the decisions for the classification using GradCAM.

3. Methods

A tongue feature classification method is proposed, and the entire framework detail is shown in Figure 1. To extract the tongue region from the original input images, the recognition algorithm with 68 facial landmarks in the dlib library is upgraded. Then, the tongue images are segmented by the UNET [23]. Lastly, we build 11 tongue classification models by using the Residual Network (ResNet) [24] according to the different features. In addition, the Grad-CAM [25] is used to provide visual explanations for model decisions.

3.1. Data Sets

The tongue images in the data set were collected with consistent conditions and high resolution to ensure the objectivity and effectiveness of the classification. The tongue images of the data set, which contains 482 photos with the pixel size of 3264 × 2448, were collected by a custom designed tongue diagnosis instrument (Figure 2). The tongue image was then classified by the expert TCM practitioners from Shanxi University of Chinese Medicine’s Affiliated Hospital who have received extensive training and have normal or corrected vision. Eleven datasets were created based on different features, including six tongue body features and five tongue coating features (Table 1).

3.2. Extraction of the Tongue Region

The original image captured by the tongue diagnostic instrument is a photograph containing the entire face as well as the edges of the instrument. The extraction of the tongue region can improve the efficiency of tongue segmentation by accurately labeling the tongue segmentation and reducing the amount of image segmentation computation. However, no detector for images of sticking out tongues is available at present. In this stage, we upgraded the facial landmark recognition algorithm in the dlib library to extract the tongue region. The steps were as follows:
Step 1: Load the original image into the facial detector and obtain a matrix containing pixel coordinate information of all 68 facial landmarks. Extract the coordinates of each landmark using traversal and label them with sequential numbers. Then, record the coordinates of the 49th, 55th, and 9th of the 68 landmarks, which are the landmarks on the sides of the mouth and the bottom of the chin respectively.
Step 2: Calculate the bounding box of the tongue area according to the coordinate difference ratio in the result of step 1. The height, width, and center point coordinates of the bounding box are calculated as follows:
height =   y 3 y 1 + y 2 2 + x 2 x 1 2
width = 2 × ( x 2 x 1 )
[ x center y center ] = [ x 1 + x 2 2 y 3 + x 2 x 1 4 height 2 ]
where (x1, y1) and (x2, y2) are the coordinates of the landmarks on the sides of the mouth, and (x3, y3) are the coordinates of the landmark on the bottom of the chin.
Step 3: Cut out the tongue region from the original image (Figure 3b) according to the bounding box.
We also built a data set for tongue segmentation after completing the task of extracting the tongue region of the original images. We manually labeled the edges of the tongue with the labelme software (Figure 3c) and generated corresponding contour images (Figure 3d). Tongue feature annotations were included with each contour image, and the data set was divided into the training set, validation set, and test set for the training of the segmentation model.

3.3. Segmentation of Tongue Images

As stated in Section 1, tongue segmentation can improve the effect of tongue feature classification by eliminating invalid information from the images. In this stage, we selected the UNET [23] based on deep CNN to complete the task of tongue segmentation. This method judges the category of each pixel in the image to obtain the tongue contour.
  • Structure: The structure of UNET can be divided into two parts (Figure 4). The first half of the UNET is the backbone feature extraction network. VGG16 [26] was chosen for feature extraction, which is a stack of convolution and maximum pooling operations. A feature layer with a new scale can be acquired after each pooling, resulting in five feature layers with distinct scales. The up-sampling part takes up the second half. The five feature layers were merged through the up-convolution method to produce an effective feature layer that contains all the features. The category of each pixel can be predicted according to the last obtained effective feature layer.
  • Training: The data set for tongue segmentation was divided into the training set, validation set, and test set according to 8:1:1. The data were enhanced prior to training by random rotation and horizontal flipping of the image, as well as normalization. The loss included cross-entropy loss and dice loss. Adam algorithm was applied for optimization. Then, we used the official weight of the UNET network in the ImageNet data set as the initial weight for transfer learning. A total of 160 rounds were used to train the network. The weights of the backbone network were frozen in the first 80 rounds for rough training, and the learning rate was 1 × 10−4. The global network was trained with a learning rate of 1 × 10−5 in the last 80 rounds for fine training.
  • Image Processing: The segmented contour images were processed by grayscale. Through observation, the generated gray image had a single gray level, with black pixels in the outer circle. Therefore, the grayscale image could be used as a mask to perform AND operate on the original tongue region image to realize the separation of the tongue. Then, we appended corresponding labels to the segmented tongue images to create classification data sets, which were also divided into the training, validation, and test sets for the training of the classification models.

3.4. Classification

We used the ResNet-34 [24] networks to classify tongue features.
  • Structure: Residual Network (ResNet-34) is a deep CNN with 34 layers, including 16 residual blocks, each with two layers (Figure 5). The last layer is an FC layer for tongue feature classification. The residual network increases the depth of the network through the connection of multiple residual blocks, while also avoiding the problem of gradient disappearance or gradient explosion.
  • Training: The data sets for tongue feature classification are divided into the training set, validation set, and test set according to 6:2:2. The data are enhanced and normalized in the same way as the segmentation network preprocessing. The official model trained by ImageNet is used for initialization. The training is terminated after 160 rounds at a learning rate of 1 × 10−4. The models are trained separately for 11 different tongue feature data sets. At the end of the network, the output layer is adjusted accordingly to the number of internal categories of the different features, and the final classification judgment is made using the argmax function.

3.5. Feature Visualization

In this stage, we used Grad-CAM [21] to produce visual explanations for decisions from tongue feature classification models.
TCM practitioners perform tongue diagnoses by identifying corresponding features in various positions on the tongue. Therefore, whether the corresponding feature area contributes more in tongue feature classification models should be determined. The GradCAM algorithm is a method for visualizing the feature maps in CNNs. The algorithm assigns importance values to each neuron through the gradient information of the back propagation of the network, and then generates a heat map by linear weighted summation to obtain the regions that the model focuses on. The tongue feature classification model can be divided into two parts: feature extraction and classifier. The classifier only classifies based on the extracted features, while the last convolutional layer of the feature extraction part is the one with the richest semantic information of tongue features. Therefore, the feature map of this layer is used for visualization. The specific realization of visualization of the corresponding regions of tongue features is as follows:
Step 1: Calculating the gradient of the classification score of the tongue feature to the last convolutional layer in conv5_x of ResNet-34. Then, similar to global average pooling, each pixel value is averaged in each channel dimension to obtain the neuron importance weight. The formula is as follows:
α k c = 1 Z i j y c A ij k
where Z represents the number of pixels in the feature map, and A ij k represents the pixel value at position ( i , j ) of the k -th feature map.
Step 2: the neuron importance weight is multiplied by each channel, added, and finally rectified by ReLU to obtain the heat map. The formula is as follows:
L Grad CAM c =   ReLU ( k α k c A k )

4. Results

In this section, we present the experimental results of the proposed method. The metrics pixel accuracy (PA) and mean intersection over union (MIoU) are used to evaluate the effect of tongue image segmentation, and the accuracy (Acc) and F1-Score are adopted to evaluate the final classification results. The calculation formulas of PA, MIoU, Acc, and F1-Score are as follows:
PA = i = 0 k p ii i = 0 k j = 0 k p ij
MIoU = 1 k + 1 i = 0 k p ii j = 0 k p ij + j = 0 k p ji p ii
Acc = TP + TN TP + FN + FP + TN
F 1 score = 1 c k = 1 c 2 p k × r k p k + r k
where k represents the number of categories other than the background, and p ij represents the number of pixels belonging to the i category, predicted to be the j category pixels. Therefore, p ii represents true positive (TP), and p ij and p ji are false positive (FP) and false negative (FN) respectively; p k represents precision, and r k represents recall rate.

4.1. Results of Tongue Image Segmentation

We have compared the GrabCut [27] algorithm, which is based on edge detection of image processing, with the UNET segmentation model after the same processing to verify the feasibility of the segmentation model. The effect comparison is shown in Figure 6, and the values of PA and MIoU corresponding to the segmentation results are shown in Table 2. The GrabCut algorithm has more isolated noises in the segmentation. On the contrary, the UNET algorithm based on deep CNN has a better tongue image segmentation effect.

4.2. Results of Tongue Feature Classification

To verify the feasibility of the classification models, we test them on the test data sets. Figure 7 visualizes the result of tongue image classification. Table 3 shows the specific result parameters of the models for identifying the different categories of tongue features. The data show that the recognition effect is ideal, with an overall recognition accuracy of 86.14% and an overall F1-Score of 80.06%.

4.3. Visualization of the Indicator Regions of Tongue Feature Classification

The Grad-CAM algorithm is used to process the final convolution layer of the classification model. Figure 8 shows the results of part of the feature classification model. The red area in the heat map has a greater impact on recognition, whereas the blue area has less impact. The observation shows that the classification model is more sensitive to the corresponding characteristic regions, confirming that it has high practical clinical value.

5. Conclusions

We have proposed a tongue feature classification method based on CNN. The proposed method has three stages. First, the upgraded dlib facial 68 landmark algorithm is used to extract the tongue region. Second, the UNET network is used to accurately segment the tongue image. Finally, the ResNet network is used as the backbone feature extraction network to achieve the final classification of the characteristics of tongue texture and tongue coating and visualize the decisions using GradCAM. The experiment shows that the method has good accuracy for tongue feature recognition, with an overall accuracy rate of over 86%, and it is more sensitive to the corresponding feature regions. Compared with other approaches, our method identifies more comprehensive features and is more in line with TCM theory, which is suitable for clinical tongue diagnosis. When patients register at the hospital or have remote treatment, the approach can complete the classification of tongue features in advance, saving the time of TCM practitioners and thus allowing more patients to receive treatment. In addition, the GradCAM algorithm is applied to obtain a visible reference position of the network’s decisions, which can help users to know the location of their tongue abnormalities and help novice practitioners to gain experience. However, the approach is not applicable to children with small faces or those who have difficulty extending their tongues because the tongue diagnostic instrument cannot perform accurate tongue acquisition in these individuals.
In addition, the proposed method can be found to have some deviations in the evaluation indexes when performing the classification of tongue color features. We believe there are two reasons for the phenomenon. One is because there is little variation in the inter-class features of tongue color, the confidence levels of these classes are similar and not high enough in the final fully connected layer of the network. The method selects the class with the highest confidence level by the argmax function, which leads to insufficient accuracy. Another is that the number of samples within the class of the data set for tongue color features is unevenly distributed, making the classification of these features less accurate. Therefore, the method still needs to be improved for the network model and data sets afterwards.

Author Contributions

Conceptualization, Z.Z. and J.L.; methodology, J.L. and X.Z.; software, J.L. and X.Z.; validation, Y.Z., Y.M., and J.L.; investigation, J.Z.; resources, X.C.; data curation, B.L. and J.L.; writing—original draft preparation, J.L.; writing—review and editing, Z.Z. and J.L.; visualization, J.L.; supervision, Z.Z.; project administration, C.X. and Z.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the National Natural Science Foundation of China (62001430); the Shanxi Province Traditional Chinese Medicine Strong Province Special Project (204-110457); the Scientific & Technological Innovation Programs of Higher Education Institutions in Shanxi (2020L0275).

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to this data being supplied by Key Laboratory of Instrumentation Science & Dynamic Measurement (North University of China), Ministry of Education, and so cannot be made freely available.

Acknowledgments

The authors gratefully acknowledge the medical technical supports by Jiuzhang Men and his team at the Shanxi University of Chinese Medicine’s Affiliated Hospital.

Conflicts of Interest

The authors declare no conflict of interest.

Ethical Statement

All subjects gave their informed consent for inclusion before they participated in the study. The study was conducted in accordance with the Declaration of Helsinki, and the protocol was approved by the Biological and Medical Ethics Committee of North University of China (110457).

References

  1. Cyranoski, D. Why Chinese medicine is heading for clinics around the world. Nature 2018, 561, 448–450. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Xu, Q.; Tang, W.; Teng, F.; Peng, W.; Zhang, Y.; Li, W.; Wen, C.; Guo, J. Intelligent Syndrome Differentiation of Traditional Chinese Medicine by ANN: A Case Study of Chronic Obstructive Pulmonary Disease. IEEE Access 2019, 7, 76167–76175. [Google Scholar] [CrossRef]
  3. Li, X.; Zhang, Y.; Cui, Q.; Yi, X.; Zhang, Y. Tooth-Marked Tongue Recognition Using Multiple Instance Learning and CNN Features. IEEE Trans. Cybern. 2018, 49, 380–387. [Google Scholar] [CrossRef] [PubMed]
  4. Zhou, H.; Hu, G.; Zhang, X. Constitution Identification of Tongue Image Based on CNN. In Proceedings of the 2018 11th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics, CISP-BMEI 2018, Beijing, China, 13–15 October 2018; IEEE: Beijing, China, 2019; pp. 1–5. [Google Scholar]
  5. Jiang, L.; Xu, W.; Chen, J. Digital imaging system for physiological analysis by tongue colour inspection. In Proceedings of the 2008 3rd IEEE Conference on Industrial Electronics and Applications, Singapore, 3–5 June 2008; IEEE: Singapore, 2008; pp. 1833–1836. [Google Scholar]
  6. Li, W.; Hu, S.; Wang, S.; Xu, S. Towards the objectification of tongue diagnosis: Automatic segmentation of tongue image. In Proceedings of the 2009 35th Annual Conference of IEEE Industrial Electronics, Porto, Portugal, 3–5 November 2009; IEEE: Porto, Portugal, 2010; pp. 2121–2124. [Google Scholar]
  7. Ning, J.; Zhang, D.; Wu, C.; Feng, Y. Automatic tongue image segmentation based on gradient vector flow and region merging. Neural Comput. Appl. 2012, 21, 1819–1826. [Google Scholar] [CrossRef]
  8. Zhu, M.; Du, J. A Novel Approach for Color Tongue Image Extraction Based on Random Walk Algorithm. Appl. Mech. Mater. 2013, 462–463, 338–342. [Google Scholar] [CrossRef]
  9. Li, C.; Wang, D.; Liu, Y.; Gao, X.; Shang, H. A novel automatic tongue image segmentation algorithm: Color enhancement method based on L*a*b* color space. In Proceedings of the 2015 IEEE International Conference on Bioinformatics and Biomedicine, BIBM 2015, Washington, DC, USA, 9–12 November 2015; IEEE: Washington, DC, USA, 2015; pp. 990–993. [Google Scholar]
  10. Xu, W.; Kanawong, R.; Xu, D.; Li, S.; Ma, T.; Zhang, G.; Duan, Y. An automatic tongue detection and segmentation framework for computer-aided tongue image analysis. In Proceedings of the 2011 IEEE 13th International Conference on e-Health Networking, Applications and Services, Columbia, MO, USA, 13–15 June 2011; IEEE: Columbia, MO, USA, 2011; pp. 189–192. [Google Scholar]
  11. Wu, J.; Zhang, Y.; Bai, J. Tongue Area Extraction in Tongue Diagnosis of Traditional Chinese Medicine. In Proceedings of the 2005 IEEE Engineering in Medicine and Biology 27th Annual Conference, Shanghai, China, 17–18 January 2006; IEEE: Shanghai, China, 2006; pp. 4955–4957. [Google Scholar]
  12. Lin, B.; Xie, J.; Li, C.; Qu, Y. Deeptongue: Tongue Segmentation via Resnet. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2018, Calgary, AB, Canada, 15–20 April 2018; IEEE: Calgary, AB, Canada, 2018; pp. 1035–1039. [Google Scholar]
  13. Xue, Y.; Li, X.; Wu, P.; Li, J.; Wang, L.; Tong, W. Automated Tongue Segmentation in Chinese Medicine Based on Deep Learning. In Proceedings of the 25th International Conference on Neural Information Processing, ICONIP 2018, Siem Reap, Cambodia, 13–16 December 2018; Cheng, L., Leung, A., Ozawa, S., Eds.; Springer: Cham, Switzerland; Siem Reap, Cambodia, 2018; pp. 542–553. [Google Scholar]
  14. Zhang, B.; Zhang, H. Significant Geometry Features in Tongue Image Analysis. Evid.-Based Complement. Alternat. Med. 2015, 2015, 897580. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Huang, B.; Wu, J.; Zhang, D.; Li, N. Tongue shape classification by geometric features. Inform. Sci. 2010, 180, 312–324. [Google Scholar] [CrossRef]
  16. Tang, Y.; Sun, Y.; Chiang, J.; Li, X. Research on Multiple-Instance Learning for Tongue Coating Classification. IEEE Access 2021, 9, 66361–66370. [Google Scholar] [CrossRef]
  17. Xu, Q.; Zeng, Y.; Tang, W.; Peng, W.; Xia, T.; Li, Z.; Teng, F.; Li, W.; Guo, J. Multi-Task Joint Learning Model for Segmenting and Classifying Tongue Images Using a Deep Neural Network. IEEE J. Biomed. Health 2020, 24, 2481–2489. [Google Scholar] [CrossRef] [PubMed]
  18. Wang, X.; Liu, J.; Wu, C.; Liu, J.; Li, Q.; Chen, Y.; Wang, X.; Chen, X.; Pang, X.; Chang, B.; et al. Artificial intelligence in tongue diagnosis: Using deep convolutional neural network for recognizing unhealthy tongue with tooth-mark. Comput. Struct. Biotechnol. J. 2020, 18, 973–980. [Google Scholar] [CrossRef] [PubMed]
  19. Li, Z.; Yu, Z.; Liu, W.; Zhang, Z. Tongue Image Segmentation via Color Decomposition and Thresholding. In Proceedings of the 2017 4th International Conference on Information Science and Control Engineering, ICISCE 2017, Changsha, China, 21–23 July 2017; IEEE: Changsha, China, 2017; pp. 752–755. [Google Scholar]
  20. Shi, M.; Li, G.; Li, F. C2G2FSnake: Automatic tongue image segmentation utilizing prior knowledge. Sci. China Inf. Sci. 2013, 56, 1–14. [Google Scholar] [CrossRef]
  21. Qu, P.; Zhang, H.; Zhuo, L.; Zhang, J.; Chen, G. Automatic Tongue Image Segmentation for Traditional Chinese Medicine Using Deep Neural Network. In Proceedings of the Intelligent Computing Theories and Application, ICIC 2017, Liverpool, UK, 20 July 2017; Springer: Cham, Switzerland; Liverpool, UK, 2017; pp. 247–259. [Google Scholar]
  22. Zhou, C.; Fan, H.; Li, C. TongueNet: Accurate Localization and Segmentation for Tongue Images using Deep Neural Networks. IEEE Access 2019, 7, 1. [Google Scholar] [CrossRef]
  23. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. arXiv 2015, arXiv:1505.04597. [Google Scholar]
  24. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. arXiv 2015, arXiv:1512.03385. [Google Scholar]
  25. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-based Localization. Int. J. Comput. Vis. 2020, 128, 336–359. [Google Scholar] [CrossRef] [Green Version]
  26. Karen, S.; Andrew, Z. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  27. Rother, C.; Kolmogorov, V.; Blake, A. GrabCut: Interactive foreground extraction using iterated graph cut. ACM Trans. Graphic. 2004, 23, 309–314. [Google Scholar] [CrossRef]
Figure 1. Overview of our framework.
Figure 1. Overview of our framework.
Micromachines 13 00501 g001
Figure 2. Tongue image acquisition with the tongue diagnosis instrument.
Figure 2. Tongue image acquisition with the tongue diagnosis instrument.
Micromachines 13 00501 g002
Figure 3. (a) Original image with 68 facial landmarks. (b) Tongue region image. (c) Edge annotation of tongue. (d) Tongue contour image.
Figure 3. (a) Original image with 68 facial landmarks. (b) Tongue region image. (c) Edge annotation of tongue. (d) Tongue contour image.
Micromachines 13 00501 g003
Figure 4. Structure of UNET.
Figure 4. Structure of UNET.
Micromachines 13 00501 g004
Figure 5. ResNet-34 Structure.
Figure 5. ResNet-34 Structure.
Micromachines 13 00501 g005
Figure 6. Segmentation effect comparison.
Figure 6. Segmentation effect comparison.
Micromachines 13 00501 g006
Figure 7. Tongue feature recognition results.
Figure 7. Tongue feature recognition results.
Micromachines 13 00501 g007
Figure 8. Heat map for different feature recognition.
Figure 8. Heat map for different feature recognition.
Micromachines 13 00501 g008
Table 1. Tongue Features.
Table 1. Tongue Features.
FeaturesInner-Class
Tongue colorPale tongue, Light red tongue, Light cyanosed tongue, Red tongue, Deep red tongue, Cyanosed tongue, Ashen tongue, Red tongue borders and tip
Rough and tender tongueNormal, Rough tongue, Tender tongue
Puffy and thin tongueNormal, Puffy tongue, Swollen tongue, Thin tongue
Spots and prickles tongueNormal, Spots and prickles tongue
Fissured tongueNormal, Fissured tongue
Tooth-marked tongueNormal, Tooth-marked tongue
Tongue coating colorWhite coating, Yellow coating, Grayish black coating
Thin and thick coatingThin coating, Thick coating
Moist and dry coatingMoist coating, Slippery coating, Dry coating
Curdy and greasy coatingNormal, Greasy coating, Curdy coating
Peeled coatingNormal, Peeled coating
Table 2. GrabCut and UNET segmentation results.
Table 2. GrabCut and UNET segmentation results.
MethodPAMIoU
GrabCut79.96%66.26%
UNET98.54%97.14%
Table 3. Tongue feature recognition result parameters.
Table 3. Tongue feature recognition result parameters.
FeatureAccF1-Score
Tongue color62.4%55.2%
Rough and tender tongue91.6%83.6%
Puffy and thin tongue86.3%74.4%
Spots and prickles tongue83.3%76.5%
Fissured tongue87.5%82.9%
Tooth-marked tongue86.7%84.0%
Tongue coating color87.5%86.5%
Thin and thick coating89.5%89.2%
Moist and dry coating87.4%67.0%
Curdy and greasy coating86.3%87.2%
Peeled coating98.9%94.2%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, J.; Zhang, Z.; Zhu, X.; Zhao, Y.; Ma, Y.; Zang, J.; Li, B.; Cao, X.; Xue, C. Automatic Classification Framework of Tongue Feature Based on Convolutional Neural Networks. Micromachines 2022, 13, 501. https://doi.org/10.3390/mi13040501

AMA Style

Li J, Zhang Z, Zhu X, Zhao Y, Ma Y, Zang J, Li B, Cao X, Xue C. Automatic Classification Framework of Tongue Feature Based on Convolutional Neural Networks. Micromachines. 2022; 13(4):501. https://doi.org/10.3390/mi13040501

Chicago/Turabian Style

Li, Jiawei, Zhidong Zhang, Xiaolong Zhu, Yunlong Zhao, Yuhang Ma, Junbin Zang, Bo Li, Xiyuan Cao, and Chenyang Xue. 2022. "Automatic Classification Framework of Tongue Feature Based on Convolutional Neural Networks" Micromachines 13, no. 4: 501. https://doi.org/10.3390/mi13040501

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop