TY - JOUR
T1 - Sequential Discrete Hashing for Scalable Cross-modality Similarity Retrieval
AU - Liu, Li
AU - Lin, Zijia
AU - Shao, Ling
AU - Shen, Fumin
AU - Ding, Guiguang
AU - Han, Jungong
PY - 2017/1
Y1 - 2017/1
N2 - With the dramatic development of the Internet, how to exploit large-scale retrieval techniques for multimodal web data has become one of the most popular but challenging problems in computer vision and multimedia. Recently, hashing methods are used for fast nearest neighbor search in large-scale data spaces, by embedding high-dimensional feature descriptors into a similarity-preserving Hamming space with a low dimension. Inspired by this, in this paper, we introduce a novel supervised cross-modality hashing framework which can generate unified binary codes for instances represented in different modalities. Particularly, in the learning phase, each bit of a code can be sequentially learned with a discrete optimization scheme that jointly minimizes its empirical loss based on a boosting strategy. In a bitwise manner, hash functions are then learned for each modality, mapping the corresponding representations into unified hash codes. We regard this approach as Cross-modality Sequential Discrete Hashing (CSDH) which can effectively reduce the quantization errors arisen in the the oversimplified roundingoff step and thus lead to high-quality binary codes. In the test phase, a simple fusion scheme is utilized to generate a unified hash code for final retrieval by merging the predicted hashing results of an unseen instance from different modalities. The proposed CSDH has been systematically evaluated on three standard datasets: Wiki, MIRFlickr and NUS-WIDE, and the results show that our method significantly outperforms the state-of-the-art multi-modality hashing techniques.
AB - With the dramatic development of the Internet, how to exploit large-scale retrieval techniques for multimodal web data has become one of the most popular but challenging problems in computer vision and multimedia. Recently, hashing methods are used for fast nearest neighbor search in large-scale data spaces, by embedding high-dimensional feature descriptors into a similarity-preserving Hamming space with a low dimension. Inspired by this, in this paper, we introduce a novel supervised cross-modality hashing framework which can generate unified binary codes for instances represented in different modalities. Particularly, in the learning phase, each bit of a code can be sequentially learned with a discrete optimization scheme that jointly minimizes its empirical loss based on a boosting strategy. In a bitwise manner, hash functions are then learned for each modality, mapping the corresponding representations into unified hash codes. We regard this approach as Cross-modality Sequential Discrete Hashing (CSDH) which can effectively reduce the quantization errors arisen in the the oversimplified roundingoff step and thus lead to high-quality binary codes. In the test phase, a simple fusion scheme is utilized to generate a unified hash code for final retrieval by merging the predicted hashing results of an unseen instance from different modalities. The proposed CSDH has been systematically evaluated on three standard datasets: Wiki, MIRFlickr and NUS-WIDE, and the results show that our method significantly outperforms the state-of-the-art multi-modality hashing techniques.
KW - Cross-modality Retrieval
KW - Hashing
KW - Discrete Optimization
KW - Bitwise
KW - Unified Hash Code
U2 - 10.1109/TIP.2016.2619262
DO - 10.1109/TIP.2016.2619262
M3 - Article
VL - 26
SP - 107
EP - 118
JO - IEEE Transactions on Image Processing
JF - IEEE Transactions on Image Processing
SN - 1057-7149
IS - 1
ER -