4.7 Article

3D multi-modality Transformer-GAN for high-quality PET reconstruction

Journal

MEDICAL IMAGE ANALYSIS
Volume 91, Issue -, Pages -

Publisher

ELSEVIER
DOI: 10.1016/j.media.2023.102983

Keywords

Positron emission tomography (PET); Transformer; Multi-modality; Generative adversarial network (GAN); PET reconstruction

Ask authors/readers for more resources

This paper proposes a method for high-quality SPET reconstruction using low-dose PET images and T1 acquisitions from MRI. It extracts local spatial features from different modalities using separate CNN encoders and integrates these features effectively using a multimodal feature integration module. It further uses a Transformer-based encoder to extract global semantic information and a CNN decoder to transform the encoded features into SPET images. Additionally, a patch-based discriminator and an edge-aware loss are applied to retain edge detail information in the reconstructed SPET images. Experimental results demonstrate that the proposed method outperforms current state-of-the-art methods in reconstructing high-quality SPET images.
Positron emission tomography (PET) scans can reveal abnormal metabolic activities of cells and provide favorable information for clinical patient diagnosis. Generally, standard-dose PET (SPET) images contain more diagnostic information than low-dose PET (LPET) images but higher-dose scans can also bring higher potential radiation risks. To reduce the radiation risk while acquiring high-quality PET images, in this paper, we propose a 3D multi-modality edge-aware Transformer-GAN for high-quality SPET reconstruction using the corresponding LPET images and T1 acquisitions from magnetic resonance imaging (T1-MRI). Specifically, to fully excavate the metabolic distributions in LPET and anatomical structural information in T1-MRI, we first use two separate CNN-based encoders to extract local spatial features from the two modalities, respectively, and design a multimodal feature integration module to effectively integrate the two kinds of features given the diverse contributions of features at different locations. Then, as CNNs can describe local spatial information well but have difficulty in modeling long-range dependencies in images, we further apply a Transformer-based encoder to extract global semantic information in the input images and use a CNN decoder to transform the encoded features into SPET images. Finally, a patch-based discriminator is applied to ensure the similarity of patch-wise data distribution between the reconstructed and real images. Considering the importance of edge information in anatomical structures for clinical disease diagnosis, besides voxel-level estimation error and adversarial loss, we also introduce an edge-aware loss to retain more edge detail information in the reconstructed SPET images. Experiments on the phantom dataset and clinical dataset validate that our proposed method can effectively reconstruct high-quality SPET images and outperform current state-of-the-art methods in terms of qualitative and quantitative metrics.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available