TY - GEN
T1 - Dual Self-Paced Cross-Modal Hashing
AU - Sun, Yuan
AU - Dai, Jian
AU - Ren, Zhenwen
AU - Chen, Yingke
AU - Peng, Dezhong
AU - Hu, Peng
PY - 2024/3/25
Y1 - 2024/3/25
N2 - Cross-modal hashing (CMH) is an efficient technique to retrieve relevant data across different modalities, such as images, texts, and videos, which has attracted more and more attention due to its low storage cost and fast query speed. Although existing CMH methods achieve remarkable processes, almost all of them treat all samples of varying difficulty levels without discrimination, thus leaving them vulnerable to noise or outliers. Based on this observation, we reveal and study dual difficulty levels implied in cross-modal hashing learning, i.e., instance-level and feature-level difficulty. To address this problem, we propose a novel Dual Self-Paced Cross-Modal Hashing (DSCMH) that mimics human cognitive learning to learn hashing from 'easy' to 'hard' in both instance and feature levels, thereby embracing robustness against noise/outliers. Specifically, our DSCMH assigns weights to each instance and feature to measure their difficulty or reliability, and then uses these weights to automatically filter out the noisy and irrelevant data points in the original space. By gradually increasing the weights during training, our method can focus on more instances and features from 'easy' to 'hard' in training, thus mitigating the adverse effects of noise or outliers. Extensive experiments are conducted on three widely-used benchmark datasets to demonstrate the effectiveness and robustness of the proposed DSCMH over 12 state-of-the-art CMH methods.
AB - Cross-modal hashing (CMH) is an efficient technique to retrieve relevant data across different modalities, such as images, texts, and videos, which has attracted more and more attention due to its low storage cost and fast query speed. Although existing CMH methods achieve remarkable processes, almost all of them treat all samples of varying difficulty levels without discrimination, thus leaving them vulnerable to noise or outliers. Based on this observation, we reveal and study dual difficulty levels implied in cross-modal hashing learning, i.e., instance-level and feature-level difficulty. To address this problem, we propose a novel Dual Self-Paced Cross-Modal Hashing (DSCMH) that mimics human cognitive learning to learn hashing from 'easy' to 'hard' in both instance and feature levels, thereby embracing robustness against noise/outliers. Specifically, our DSCMH assigns weights to each instance and feature to measure their difficulty or reliability, and then uses these weights to automatically filter out the noisy and irrelevant data points in the original space. By gradually increasing the weights during training, our method can focus on more instances and features from 'easy' to 'hard' in training, thus mitigating the adverse effects of noise or outliers. Extensive experiments are conducted on three widely-used benchmark datasets to demonstrate the effectiveness and robustness of the proposed DSCMH over 12 state-of-the-art CMH methods.
UR - http://www.scopus.com/inward/record.url?scp=85189612562&partnerID=8YFLogxK
UR - https://aaai.org/aaai-conference/
U2 - 10.1609/aaai.v38i14.29441
DO - 10.1609/aaai.v38i14.29441
M3 - Conference contribution
AN - SCOPUS:85189612562
SN - 1577358872
SN - 9781577358879
VL - 38
T3 - Proceedings of the AAAI Conference on Artificial Intelligence
SP - 15184
EP - 15192
BT - Proceedings of the 38th AAAI Conference on Artificial Intelligence
A2 - Wooldridge, Michael
A2 - Dy, Jennifer
A2 - Natarajan, Sriraam
PB - AAAI Press/International Joint Conferences on Artificial Intelligence
CY - Washington, DC
T2 - 38th AAAI Conference on Artificial Intelligence, AAAI 2024
Y2 - 20 February 2024 through 27 February 2024
ER -