4.7 Article

A multi-step approach for tongue image classification in patients with diabetes

Journal

COMPUTERS IN BIOLOGY AND MEDICINE
Volume 149, Issue -, Pages -

Publisher

PERGAMON-ELSEVIER SCIENCE LTD
DOI: 10.1016/j.compbiomed.2022.105935

Keywords

Diabetes; Tongue image; Vector quantized variational autoencoder; K -means; Vision transformer; Machine learning; Deep learning

Funding

  1. National Key Research and Development Program of China [2017YFC1703300, 2017YFC1703301]
  2. National Natural Science Foundation of China [81873235, 81973750, 82004258, 81904094, 82104736, 82104738]
  3. Shanghai Science and Technology Commission [21010504400]
  4. Shanghai Municipal Health Commission [2020JQ003, BWS17J028]

Ask authors/readers for more resources

This study achieved fine classification of diabetic population based on tongue images data, providing diagnostic basis for individualized treatment plans and promoting objective and standardized development of TCM diagnosis. The proposed method combined unsupervised, self-supervised, and supervised learning, achieving state-of-the-art results and offering new ideas for promoting the standardization of TCM diagnosis.
Background: In China, diabetes is a common, high-incidence chronic disease. Diabetes has become a severe public health problem. However, the current diagnosis and treatment methods are difficult to control the progress of diabetes. Traditional Chinese Medicine (TCM) has become an option for the treatment of diabetes due to its low cost, good curative effect, and good accessibility.Objective: Based on the tongue images data to realize the fine classification of the diabetic population, provide a diagnostic basis for the formulation of individualized treatment plans for diabetes, ensure the accuracy and consistency of the TCM diagnosis, and promote the objective and standardized development of TCM diagnosis. Methods: We use the TFDA-1 tongue examination instrument to collect the tongue images of the subjects. Tongue Diagnosis Analysis System (TDAS) is used to extract the TDAS features of the tongue images. Vector Quantized Variational Autoencoder (VQ-VAE) extracts VQ-VAE features from tongue images. Based on VQ-VAE features, K -means clustering tongue images. TDAS features are used to describe the differences between clusters. Vision Transformer (ViT) combined with Grad-weighted Class Activation Mapping (Grad-CAM) is used to verify the clustering results and calculate positioning diagnostic information.Results: Based on VQ-VAE features, K-means divides the diabetic population into 4 clusters with clear boundaries. The silhouette, calinski harabasz, and davies bouldin scores are 0.391, 673.256, and 0.809, respectively. Cluster 1 had the highest Tongue Body L (TB-L) and Tongue Coating L (TC-L) and the lowest Tongue Coating Angular second moment (TC-ASM), with a pale red tongue and white coating. Cluster 2 had the highest TC-b with a yellow tongue coating. Cluster 3 had the highest TB-a with a red tongue. Group 4 had the lowest TB-L, TC-L, and TB-b and the highest Per-all with a purple tongue and the largest tongue coating area. ViT verifies the clustering results of K-means, the highest Top-1 Classification Accuracy (CA) is 87.8%, and the average CA is 84.4%.Conclusions: The study organically combined unsupervised learning, self-supervised learning, and supervised learning and designed a complete diabetic tongue image classification method. This method does not rely on human intervention, makes decisions based entirely on tongue image data, and achieves state-of-the-art results. Our research will help TCM deeply participate in the individualized treatment of diabetes and provide new ideas for promoting the standardization of TCM diagnosis.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available