Sequential discrete hashing for scalable cross-modality similarity retrieval

Li Liu, Zijia Lin, Ling Shao, Fumin Shen, Guiguang Ding, Jungong Han

Research output: Contribution to journalArticlepeer-review

68 Citations (Scopus)
25 Downloads (Pure)

Abstract

With the dramatic development of the Internet, how to exploit large-scale retrieval techniques for multimodal web data has become one of the most popular but challenging problems in computer vision and multimedia. Recently, hashing methods are used for fast nearest neighbor search in large-scale data spaces, by embedding high-dimensional feature descriptors into a similarity preserving Hamming space with a low dimension. Inspired by this, in this paper, we introduce a novel supervised cross-modality hashing framework, which can generate unified binary codes for instances represented in different modalities. Particularly, in the learning phase, each bit of a code can be sequentially learned with a discrete optimization scheme that jointly minimizes its empirical loss based on a boosting strategy. In a bitwise manner, hash functions are then learned for each modality, mapping the corresponding representations into unified hash codes. We regard this approach as cross-modality sequential discrete hashing (CSDH), which can effectively reduce the quantization errors arisen in the oversimplified rounding-off step and thus lead to high-quality binary codes. In the test phase, a simple fusion scheme is utilized to generate a unified hash code for final retrieval by merging the predicted hashing results of an unseen instance from different modalities. The proposed CSDH has been systematically evaluated on three standard data sets: Wiki, MIRFlickr, and NUS-WIDE, and the results show that our method significantly outperforms the state-of-the-art multimodality hashing techniques.
Original languageEnglish
Pages (from-to)107-118
Number of pages12
JournalIEEE Transactions on Image Processing
Volume26
Issue number1
Early online date20 Oct 2016
DOIs
Publication statusPublished - 1 Jan 2017

Cite this