4.7 Article

Diverse data augmentation for learning image segmentation with cross-modality annotations

期刊

MEDICAL IMAGE ANALYSIS
卷 71, 期 -, 页码 -

出版社

ELSEVIER
DOI: 10.1016/j.media.2021.102060

关键词

Disentangled representation learning; Data augmentation; Generative adversarial learning; Medical image segmentation

资金

  1. NIH/NIDCR [R01 DE022676, R01 DE027251, R01 DE021863]

向作者/读者索取更多资源

The lack of annotated data is a major challenge in building reliable image segmentation models. In this paper, a diverse data augmentation generative adversarial network (DDA-GAN) is introduced to train a segmentation model for an unannotated target image domain by borrowing information from an annotated source image domain. The method effectively combines features from different domains to improve segmentation quality through data augmentation and model training.
The dearth of annotated data is a major hurdle in building reliable image segmentation models. Man-ual annotation of medical images is tedious, time-consuming, and significantly variable across imaging modalities. The need for annotation can be ameliorated by leveraging an annotation-rich source modality in learning a segmentation model for an annotation-poor target modality. In this paper, we introduce a diverse data augmentation generative adversarial network (DDA-GAN) to train a segmentation model for an unannotated target image domain by borrowing information from an annotated source image domain. This is achieved by generating diverse augmented data for the target domain by one-to-many source-to-target translation. The DDA-GAN uses unpaired images from the source and target domains and is an end-to-end convolutional neural network that (i) explicitly disentangles domain-invariant structural fea-tures related to segmentation from domain-specific appearance features, (ii) combines structural features from the source domain with appearance features randomly sampled from the target domain for data augmentation, and (iii) train the segmentation model with the augmented data in the target domain and the annotations from the source domain. The effectiveness of our method is demonstrated both qualita-tively and quantitatively in comparison with the state of the art for segmentation of craniomaxillofacial bony structures via MRI and cardiac substructures via CT.(c) 2021 Elsevier B.V. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据