4.5 Article

Multiview multimodal network for breast cancer diagnosis in contrast-enhanced spectral mammography images

Publisher

SPRINGER HEIDELBERG
DOI: 10.1007/s11548-021-02391-4

Keywords

Contrast-enhanced spectral mammography; Classification; Breast cancer; Multiview; Multimodal

Funding

  1. National Natural Science Foundation of China [81871508, 61773246]
  2. Major Program of Shandong Province Natural Science Foundation [ZR2018ZB0419]
  3. Taishan Scholar Program of Shandong Province of China [TSHW201502038]

Ask authors/readers for more resources

The study proposed a deep learning classification model that combines multiple features of CESM. The experimental results indicate that our method is more precise than the state-of-the-art methods and produces accurate results for the classification of CESM images.
Purpose CESM (contrast-enhanced spectral mammography) is an efficient tool for detecting breast cancer because of its image characteristics. However, among most deep learning-based methods for breast cancer classification, few models can integrate both its multiview and multimodal features. To effectively utilize the image features of CESM and thus help physicians to improve the accuracy of diagnosis, we propose a multiview multimodal network (MVMM-Net). Methods The experiment is carried out to evaluate the in-house CESM images dataset taken from 95 patients aged 21-74 years with 760 images. The framework consists of three main stages: the input of the model, image feature extraction, and image classification. The first stage is to preprocess the CESM to utilize its multiview and multimodal features effectively. In the feature extraction stage, a deep learning-based network is used to extract CESM images features. The last stage is to integrate different features for classification using the MVMM-Net model. Results According to the experimental results, the proposed method based on the Res2Net50 framework achieves an accuracy of 96.591%, sensitivity of 96.396%, specificity of 96.350%, precision of 96.833%, F1_score of 0.966, and AUC of 0.966 on the test set. Comparative experiments illustrate that the classification performance of the model can be improved by using multiview multimodal features. Conclusion We proposed a deep learning classification model that combines multiple features of CESM. The results of the experiment indicate that our method is more precise than the state-of-the-art methods and produces accurate results for the classification of CESM images.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available