4.7 Article

SuperFusion: A Versatile Image Registration and Fusion Network with Semantic Awareness

Journal

IEEE-CAA JOURNAL OF AUTOMATICA SINICA
Volume 9, Issue 12, Pages 2121-2137

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/JAS.2022.106082

Keywords

Global spatial attention; image fusion; image registration; mutual promotion; semantic awareness

Funding

  1. National Natural Science Foundation of China [62276192, 62075169, 62061160370]
  2. Key Research and Development Program of Hubei Province [2020BAB113]

Ask authors/readers for more resources

This study proposes a novel image registration and fusion method called SuperFusion, which integrates image registration, image fusion, and semantic requirements of high-level vision tasks into a single framework. By utilizing registration network, fusion network, and semantic constraint, SuperFusion overcomes the artifacts caused by slight shifts or deformations in input images and achieves superior performance in high-level vision tasks.
Image fusion aims to integrate complementary information in source images to synthesize a fused image comprehensively characterizing the imaging scene. However, existing image fusion algorithms are only applicable to strictly aligned source images and cause severe artifacts in the fusion results when input images have slight shifts or deformations. In addition, the fusion results typically only have good visual effect, but neglect the semantic requirements of high-level vision tasks. This study incorporates image registration, image fusion, and semantic requirements of high-level vision tasks into a single framework and proposes a novel image registration and fusion method, named SuperFusion. Specifically, we design a registration network to estimate bidirectional deformation fields to rectify geometric distortions of input images under the supervision of both photometric and end-point constraints. The registration and fusion are combined in a symmetric scheme, in which while mutual promotion can be achieved by optimizing the naive fusion loss, it is further enhanced by the mono-modal consistent constraint on symmetric fusion outputs. In addition, the image fusion network is equipped with the global spatial attention mechanism to achieve adaptive feature integration. Moreover, the semantic constraint based on the pre-trained segmentation model and Lovasz-Softmax loss is deployed to guide the fusion network to focus more on the semantic requirements of high-level vision tasks. Extensive experiments on image registration, image fusion, and semantic segmentation tasks demonstrate the superiority of our SuperFusion compared to the state-of-the-art alternatives. The source code and pre-trained model are publicly available at https://github.com/Linfeng-Tang/SuperFusion.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available