Home > Research > Publications & Outputs > Sequential Discrete Hashing for Scalable Cross-...

Associated organisational unit

Electronic data

  • CSDH_TIP_submit

    Rights statement: ©2017 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE.

    Accepted author manuscript, 1.15 MB, PDF document

    Available under license: CC BY-NC: Creative Commons Attribution-NonCommercial 4.0 International License

View graph of relations

Sequential Discrete Hashing for Scalable Cross-modality Similarity Retrieval

Research output: Contribution to Journal/MagazineJournal articlepeer-review

Published
  • Li Liu
  • Zijia Lin
  • Ling Shao
  • Fumin Shen
  • Guiguang Ding
  • Jungong Han
Close
<mark>Journal publication date</mark>01/2017
<mark>Journal</mark>IEEE Transactions on Image Processing
Issue number1
Volume26
Number of pages12
Pages (from-to)107-118
Publication StatusPublished
Early online date20/10/16
<mark>Original language</mark>English

Abstract

With the dramatic development of the Internet, how to exploit large-scale retrieval techniques for multimodal web data has become one of the most popular but challenging problems in computer vision and multimedia. Recently, hashing methods are used for fast nearest neighbor search in large-scale data spaces, by embedding high-dimensional feature descriptors into a similarity preserving Hamming space with a low dimension. Inspired by this, in this paper, we introduce a novel supervised cross-modality hashing framework, which can generate unified binary codes for instances represented in different modalities. Particularly, in the learning phase, each bit of a code can be sequentially learned with a discrete optimization scheme that jointly minimizes its empirical loss based on a boosting strategy. In a bitwise manner, hash functions are then learned for each modality, mapping the corresponding representations into unified hash codes. We regard this approach as cross-modality sequential discrete hashing (CSDH), which can effectively reduce the quantization errors arisen in the oversimplified rounding-off step and thus lead to high-quality binary codes. In the test phase, a simple fusion scheme is utilized to generate a unified hash code for final retrieval by merging the predicted hashing results of an unseen instance from different modalities. The proposed CSDH has been systematically evaluated on three standard data sets: Wiki, MIRFlickr, and NUS-WIDE, and the results show that our method significantly outperforms the state-of-the-art multimodality hashing techniques.