4.7 Article

Deep Coattention-Based Comparator for Relative Representation Learning in Person Re-Identification

Journal

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TNNLS.2020.2979190

Keywords

Feature extraction; Fuses; Visualization; Learning systems; Benchmark testing; Australia; Neural networks; Attention models; coattention; person re-identification (re-ID); relative representations

Funding

  1. NSFC [61725203, 61732008]
  2. National Key Research and Development Program of China [2018YFB0804200]

Ask authors/readers for more resources

In this article, a deep co-attention-based comparator (DCC) is introduced to fuse codependent representations of paired images to correlate the best relevant parts and produce their relative representations accordingly. The proposed approach mimics human foveation to detect distinct regions concurrently across images and alternately attends to fuse them into the similarity learning. Extensive experiments show that the comparator achieves state-of-the-art results in benchmark data sets, with 1.2 and 2.5 points gain in mean average precision (mAP) on DukeMTMC-reID and Market-1501, respectively.
Person re-identification (re-ID) favors discriminative representations over unseen shots to recognize identities in disjoint camera views. Effective methods are developed via pair-wise similarity learning to detect a fixed set of region features, which can be mapped to compute the similarity value. However, relevant parts of each image are detected independently without referring to the correlation on the other image. Also, region-based methods spatially position local features for their aligned similarities. In this article, we introduce the deep coattention-based comparator (DCC) to fuse codependent representations of paired images so as to correlate the best relevant parts and produce their relative representations accordingly. The proposed approach mimics the human foveation to detect the distinct regions concurrently across images and alternatively attends to fuse them into the similarity learning. Our comparator is capable of learning representations relative to a test shot and well-suited to reidentifying pedestrians in surveillance. We perform extensive experiments to provide the insights and demonstrate the state of the arts achieved by our method in benchmark data sets: 1.2 and 2.5 points gain in mean average precision (mAP) on DukeMTMC-reID and Market-1501, respectively.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available