4.7 Article

STFuse: Infrared and Visible Image Fusion via Semisupervised Transfer Learning

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TNNLS.2023.3328060

Keywords

Cross-task knowledge; deep learning (DL); image fusion; semisupervised transfer learning

Ask authors/readers for more resources

This paper proposes a semisupervised transfer learning-based method called STFuse for infrared and visible image fusion (IVIF). By borrowing supervised knowledge from multifocus image fusion (MFIF) task and using a guidance loss, the method effectively utilizes cross-task knowledge to alleviate the limitation of the lack of ground truth. The design of a cross-feature enhancement module further enhances visual quality, statistical metrics, and docking performance with high-level vision tasks.
Infrared and visible image fusion (IVIF) aims to obtain an image that contains complementary information about the source images. However, it is challenging to define complementary information between source images in the lack of ground truth and without borrowing prior knowledge. Therefore, we propose a semisupervised transfer learning-based method for IVIF, termed STFuse, which aims to transfer knowledge from an informative source domain to a target domain, thus breaking the above limitations. The critical aspect of our method is to borrow supervised knowledge from the multifocus image fusion (MFIF) task and to filter out task-specific attribute knowledge by using a guidance loss L-g, which motivates its cross-task use in IVIF tasks. Using this cross-task knowledge effectively alleviates the limitation of the lack of ground truth on fusion performance, and the complementary expression ability under the constraint of supervised knowledge is more instructive than prior knowledge. Moreover, we designed a cross-feature enhancement module (CEM) that utilizes self-attention and mutual-attention features to guide each branch to refine features and then facilitate the integration of cross-modal complementary features. Extensive experiments demonstrate that our method has good advantages in terms of visual quality and statistical metrics, as well as the docking of high-level vision tasks, compared with other state-of-the-art methods.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available