4.6 Article

Multimodal Medical Image Fusion Using Stacked Auto-encoder in NSCT Domain

Journal

JOURNAL OF DIGITAL IMAGING
Volume 35, Issue 5, Pages 1308-1325

Publisher

SPRINGER
DOI: 10.1007/s10278-021-00554-y

Keywords

Deep Learning; Image fusion; Medical image modalities; Non-subsampled contourlet transform (NSCT); Stacked sparse auto-encoder (SSAE)

Ask authors/readers for more resources

Medical image fusion is a process of merging important information from different modality images to create a more informative fused image. Deep learning methods have achieved significant breakthroughs in the field of image fusion. This paper proposes a medical image fusion method based on stacked sparse auto-encoder (SSAE) and non-subsampled contourlet transform (NSCT), which has been evaluated on various medical image modalities.
Medical image fusion is a process that aims to merge the important information from images with different modalities of the same organ of the human body to create a more informative fused image. In recent years, deep learning (DL) methods have achieved significant breakthroughs in the field of image fusion because of their great efficiency. The DL methods in image fusion have become an active topic due to their high feature extraction and data representation ability. In this work, stacked sparse auto-encoder (SSAE), a general category of deep neural networks, is exploited in medical image fusion. The SSAE is an efficient technique for unsupervised feature extraction. It has high capability of complex data representation. The proposed fusion method is carried as follows. Firstly, the source images are decomposed into low- and high-frequency coefficient sub-bands with the non-subsampled contourlet transform (NSCT). The NSCT is a flexible multi-scale decomposition technique, and it is superior to traditional decomposition techniques in several aspects. After that, the SSAE is implemented for feature extraction to obtain a sparse and deep representation from high-frequency coefficients. Then, the spatial frequencies are computed for the obtained features to be used for high-frequency coefficient fusion. After that, a maximum-based fusion rule is applied to fuse the low-frequency sub-band coefficients. The final integrated image is acquired by applying the inverse NSCT. The proposed method has been applied and assessed on various groups of medical image modalities. Experimental results prove that the proposed method could effectively merge the multimodal medical images, while preserving the detail information, perfectly.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available