4.7 Article

Deep Coupled Metric Learning for Cross-Modal Matching

期刊

IEEE TRANSACTIONS ON MULTIMEDIA
卷 19, 期 6, 页码 1234-1244

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TMM.2016.2646180

关键词

Coupled learning; cross-modal matching; deep model; metric learning; multimedia retrieval

资金

  1. National Key Research and Development Program of China [2016YFB1001001]
  2. National Natural Science Foundation of China [61672306, 61225008, 61572271, 61527808, 61373074, 61373090]
  3. National 1000 Young Talents Plan Program
  4. National Basic Research Program of China [2014CB349304]
  5. Ministry of Education of China [20120002110033]
  6. Tsinghua University Initiative Scientific Research Program

向作者/读者索取更多资源

In this paper, we propose a new deep coupled metric learning (DCML) method for cross-modal matching, which aims to match samples captured from two different modalities (e.g., texts versus images, visible versus near infrared images). Unlike existing cross-modal matching methods which learn a linear common space to reduce the modality gap, our DCML designs two feedforward neural networks which learn two sets of hierarchical nonlinear transformations (one set for each modality) to nonlinearly map samples from different modalities into a shared latent feature subspace, under which the intraclass variation is minimized and the interclass variation is maximized, and the difference of each data pair captured from two modalities of the same class is minimized, respectively. Experimental results on four different cross-modal matching datasets validate the efficacy of the proposed approach.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据