4.7 Article

Multi-task learning for segmentation and classification of tumors in 3D automated breast ultrasound images☆

Journal

MEDICAL IMAGE ANALYSIS
Volume 70, Issue -, Pages -

Publisher

ELSEVIER
DOI: 10.1016/j.media.2020.101918

Keywords

AUBS image; Segmentation; Classification; Multi-task learning; Joint training

Funding

  1. National Natural Science Foundation of China [61872030]
  2. Major Science and Technology Innovation Project of Shandong Province [2019TSLH0206]
  3. China Scholarship Council [201907090085]

Ask authors/readers for more resources

A novel multi-task learning framework for joint segmentation and classification of tumors in 3D automated breast ultrasound (ABUS) images is proposed in this paper, consisting of an encoder-decoder network for segmentation and a light-weight multi-scale network for classification. The framework uses an iterative training strategy to refine feature maps with the help of probability maps and experiments show that it outperforms single-task learning counterparts.
Tumor classification and segmentation are two important tasks for computer-aided diagnosis (CAD) using 3D automated breast ultrasound (ABUS) images. However, they are challenging due to the significant shape variation of breast tumors and the fuzzy nature of ultrasound images (e.g., low contrast and signal to noise ratio). Considering the correlation between tumor classification and segmentation, we argue that learning these two tasks jointly is able to improve the outcomes of both tasks. In this paper, we propose a novel multi-task learning framework for joint segmentation and classification of tumors in ABUS images. The proposed framework consists of two sub-networks: an encoder-decoder network for segmentation and a light-weight multi-scale network for classification. To account for the fuzzy boundaries of tumors in ABUS images, our framework uses an iterative training strategy to refine feature maps with the help of probability maps obtained from previous iterations. Experimental results based on a clinical dataset of 170 3D ABUS volumes collected from 107 patients indicate that the proposed multi-task framework improves tumor segmentation and classification over the single-task learning counterparts. (c) 2020 Elsevier B.V. All rights reserved.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available