4.7 Article

Online deep hashing for both uni-modal and cross-modal retrieval

Journal

INFORMATION SCIENCES
Volume 608, Issue -, Pages 1480-1502

Publisher

ELSEVIER SCIENCE INC
DOI: 10.1016/j.ins.2022.07.039

Keywords

Online hashing; Cross-modal retrieval; Uni-modal retrieval; Online learning; Knowledge distillation; Deep learning

Funding

  1. Natural Science Foundation of Chongqing, China [cstc2019jcyjzdxmX0011]
  2. National Natural Science Foundation of China [62076044, BYJS202013, 20202BAB202017]
  3. Chongqing University of Posts and Telecommunications [61966002]
  4. Natural Science Foundation of Jiangxi Province [61962003]

Ask authors/readers for more resources

A novel method called ODHUC is proposed for both uni-modal and cross-modal retrieval. It adopts an online deep hashing approach to continuously learn hash codes by sampling and updating the model. ODHUC also avoids forgetting old knowledge through knowledge distillation. Experimental results demonstrate that ODHUC outperforms other methods.
The batch-based hash learning paradigm for uni-modal or cross-modal retrieval has made great progress in recent decades. However, methods that are based on this paradigm can-not adapt to the scenario of incoming new data streams continuously; in addition, they are inefficient in terms of training time and memory cost in big data searches because they have to accumulate all the database data and new incoming data before training. Although a few online hash retrieval methods have been proposed to address these issues in recent years, they are based on shallow models, and none of them can perform uni-modal and cross-modal retrieval in one framework. To this end, we propose a novel method, namely, Online Deep Hashing for both Uni-modal and Cross-modal retrieval (ODHUC). For online deep hashing, ODHUC first trains image and text neural networks with image and text fundamental databases, and their hash codes are learned. When new data arrive, ODHUC samples images and texts from the newly arriving data and fundamental databases to update the image and text networks and learn the hash codes of the new data, the aim of which is to keep the similarity of the hash codes of the incremental and old data consistent with the supervised semantic similarity. ODHUC is also trained with sampled old data by knowledge distillation to learn new knowledge and avoid catastrophic forget-ting of old knowledge. In this way, ODHUC achieves online deep hash learning continually without performance decline. For two categories of retrieval tasks, ODHUC learns image and text deep hash functions in two stages to align image and text features; it generates image hash codes first and then utilizes them to supervise the text hash code learning. Thus, ODHUC can perform both uni-modal and cross-modal retrieval tasks in one frame-work. Extensive experiments on three real-world datasets demonstrate that ODHUC out-performs several state-of-the-art methods.(c) 2022 Elsevier Inc. All rights reserved.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available