4.6 Article

Dynamic-Pix2Pix: Medical image segmentation by injecting noise to cGAN for modeling input and target domain joint distributions with limited training data

期刊

出版社

ELSEVIER SCI LTD
DOI: 10.1016/j.bspc.2023.104877

关键词

Deep neural networks; Pix2Pix; Image translation; Domain adaptation; Domain generalization; Dynamic neural networks

向作者/读者索取更多资源

This study aims to improve the performance of conditional Generative Adversarial Networks (cGANs) in translating images by learning the target domain distribution from limited data with the help of noise input. The proposed method achieves better model generalization and comparable results compared to state-of-the-art methods.
Learning to translate images from a source to a target domain with applications such as converting simple line drawing to oil painting has attracted significant attention. The quality of translated images is directly related to two crucial issues. First, the consistency of the output distribution with that of the target is essential. Second, the generated output should have a high correlation with the input. Conditional Generative Adversarial Networks, cGANs, are the most common models for translating images. The performance of a cGAN drops when we use a limited training dataset. In this work, we study the obstacles that prevent a U-shaped model from learning the target domain distribution from limited data by using noise as input. This study helps to increase the Pix2Pix (a form of cGAN) target distribution modeling ability from limited data with the help of dynamic neural network theory. Our model has two learning cycles. The model learns the correlation between input and ground truth in the first cycle. Then, the model's architecture is refined in the second cycle to learn the target distribution from noise input. These processes are executed in each iteration of the training procedure. Helping the cGAN learn the target distribution from noise input results in a better model generalization during the test time and allows the model to fit almost perfectly to the target domain distribution. As a result, our model surpasses the Pix2Pix model in segmenting HC18 and Montgomery's chest x-ray images. Both qualitative and Dice scores show the superiority of our model. Although our proposed method does not use thousand of additional data for pretraining, it produces comparable results for the in and out-domain generalization compared to the state-of-the-art methods.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据