4.7 Article

Multi-Interactive Dual-Decoder for RGB-Thermal Salient Object Detection

Journal

IEEE TRANSACTIONS ON IMAGE PROCESSING
Volume 30, Issue -, Pages 5678-5691

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TIP.2021.3087412

Keywords

Decoding; Feature extraction; Object detection; Semantics; Fuses; Task analysis; Streaming media; Salient object detection; information fusion; dual-decoder; multiple interactions

Funding

  1. Natural Science Foundation of Anhui Higher Education Institution of China [KJ2020A0033, KJ2020A0061, KJ2019A0005]
  2. National Natural Science Foundation of China [61976003]
  3. University Synergy Innovation Program of Anhui Province [GXXT-2020-051]
  4. Key Project of Research and Development of Anhui Province [201904b11020037]
  5. Nature Science Research Project of Anhui province [1908085MF185]
  6. China Postdoctoral Science Foundation [2020M681989]

Ask authors/readers for more resources

In this study, a multi-interactive dual-decoder is proposed to mine and model the multi-type interactions for accurate RGBT SOD. The method performs well on public datasets and can handle challenging scenarios even in the presence of invalid modality.
RGB-thermal salient object detection (SOD) aims to segment the common prominent regions of visible image and corresponding thermal infrared image that we call it RGBT SOD. Existing methods don't fully explore and exploit the potentials of complementarity of different modalities and multi-type cues of image contents, which play a vital role in achieving accurate results. In this paper, we propose a multi-interactive dual-decoder to mine and model the multi-type interactions for accurate RGBT SOD. In specific, we first encode two modalities into multi-level multi-modal feature representations. Then, we design a novel dual-decoder to conduct the interactions of multi-level features, two modalities and global contexts. With these interactions, our method works well in diversely challenging scenarios even in the presence of invalid modality. Finally, we carry out extensive experiments on public RGBT and RGBD SOD datasets, and the results show that the proposed method achieves the outstanding performance against state-of-the-art algorithms. The source code has been released at: https://github.com/lz118/Multi-interactive-Dual-decoder.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available