4.7 Article

3D multi-modality Transformer-GAN for high-quality PET reconstruction

期刊

MEDICAL IMAGE ANALYSIS
卷 91, 期 -, 页码 -

出版社

ELSEVIER
DOI: 10.1016/j.media.2023.102983

关键词

Positron emission tomography (PET); Transformer; Multi-modality; Generative adversarial network (GAN); PET reconstruction

向作者/读者索取更多资源

This paper proposes a method for high-quality SPET reconstruction using low-dose PET images and T1 acquisitions from MRI. It extracts local spatial features from different modalities using separate CNN encoders and integrates these features effectively using a multimodal feature integration module. It further uses a Transformer-based encoder to extract global semantic information and a CNN decoder to transform the encoded features into SPET images. Additionally, a patch-based discriminator and an edge-aware loss are applied to retain edge detail information in the reconstructed SPET images. Experimental results demonstrate that the proposed method outperforms current state-of-the-art methods in reconstructing high-quality SPET images.
Positron emission tomography (PET) scans can reveal abnormal metabolic activities of cells and provide favorable information for clinical patient diagnosis. Generally, standard-dose PET (SPET) images contain more diagnostic information than low-dose PET (LPET) images but higher-dose scans can also bring higher potential radiation risks. To reduce the radiation risk while acquiring high-quality PET images, in this paper, we propose a 3D multi-modality edge-aware Transformer-GAN for high-quality SPET reconstruction using the corresponding LPET images and T1 acquisitions from magnetic resonance imaging (T1-MRI). Specifically, to fully excavate the metabolic distributions in LPET and anatomical structural information in T1-MRI, we first use two separate CNN-based encoders to extract local spatial features from the two modalities, respectively, and design a multimodal feature integration module to effectively integrate the two kinds of features given the diverse contributions of features at different locations. Then, as CNNs can describe local spatial information well but have difficulty in modeling long-range dependencies in images, we further apply a Transformer-based encoder to extract global semantic information in the input images and use a CNN decoder to transform the encoded features into SPET images. Finally, a patch-based discriminator is applied to ensure the similarity of patch-wise data distribution between the reconstructed and real images. Considering the importance of edge information in anatomical structures for clinical disease diagnosis, besides voxel-level estimation error and adversarial loss, we also introduce an edge-aware loss to retain more edge detail information in the reconstructed SPET images. Experiments on the phantom dataset and clinical dataset validate that our proposed method can effectively reconstruct high-quality SPET images and outperform current state-of-the-art methods in terms of qualitative and quantitative metrics.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据