4.6 Article

Global-and-Local Collaborative Learning for Co-Salient Object Detection

Journal

IEEE TRANSACTIONS ON CYBERNETICS
Volume 53, Issue 3, Pages 1920-1931

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TCYB.2022.3169431

Keywords

Semantics; Task analysis; Feature extraction; Convolution; Object detection; Computational modeling; Collaborative work; 3-D convolution; co-salient object detection (CoSOD); global correspondence modeling (GCM); local correspondence modeling (LCM)

Ask authors/readers for more resources

This article proposes a global-and-local collaborative learning architecture (GLNet) to effectively extract interimage correspondence in co-salient object detection. The GLNet utilizes global and local correspondence modeling, pairwise correlation transformation, and correspondence aggregation to enhance the comprehensive interimage collaboration cues. The evaluation results demonstrate the superiority of GLNet over state-of-the-art competitors.
The goal of co-salient object detection (CoSOD) is to discover salient objects that commonly appear in a query group containing two or more relevant images. Therefore, how to effectively extract interimage correspondence is crucial for the CoSOD task. In this article, we propose a global-and-local collaborative learning (GLNet) architecture, which includes a global correspondence modeling (GCM) and a local correspondence modeling (LCM) to capture the comprehensive interimage corresponding relationship among different images from the global and local perspectives. First, we treat different images as different time slices and use 3-D convolution to integrate all intrafeatures intuitively, which can more fully extract the global group semantics. Second, we design a pairwise correlation transformation (PCT) to explore similarity correspondence between pairwise images and combine the multiple local pairwise correspondences to generate the local interimage relationship. Third, the interimage relationships of the GCM and LCM are integrated through a global-and-local correspondence aggregation (GLA) module to explore more comprehensive interimage collaboration cues. Finally, the intra and inter features are adaptively integrated by an intra-and-inter weighting fusion (AEWF) module to learn co-saliency features and predict the co-saliency map. The proposed GLNet is evaluated on three prevailing CoSOD benchmark datasets, demonstrating that our model trained on a small dataset (about 3k images) still outperforms 11 state-of-the-art competitors trained on some large datasets (about 8k-200k images).

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available