Binary Set Embedding for Cross-Modal Retrieval

Mengyang Yu, Li Liu, Ling Shao

Research output: Contribution to journalArticlepeer-review

22 Citations (Scopus)


Cross-modal retrieval is such a challenging topic that traditional global representations would fail to bridge the semantic gap between images and texts to a satisfactory level. Using local features from images and words from documents directly can be more robust for the scenario with large intraclass variations and small interclass discrepancies. In this paper, we propose a novel unsupervised binary coding algorithm called binary set embedding (BSE) to obtain meaningful hash codes for local features from the image domain and words from text domain. Understanding image features with the word vectors learned from the human language instead of the provided documents from data sets, BSE can map samples into a common Hamming space effectively and efficiently where each sample is represented by the sets of local feature descriptors from image and text domains. In particular, BSE explores relationship among local features in both feature level and image (text) level, which can balance the sensitivity of each other. Furthermore, a recursive orthogonalization procedure is applied to reduce the redundancy of codes. Extensive experiments demonstrate the superior performance of BSE compared with state-of-the-art cross-modal hashing methods using either image or text queries.
Original languageEnglish
Pages (from-to)2899-2910
JournalIEEE Transactions on Neural Networks and Learning Systems
Issue number12
Publication statusPublished - 27 Sept 2017


Dive into the research topics of 'Binary Set Embedding for Cross-Modal Retrieval'. Together they form a unique fingerprint.

Cite this