We found a match
Your institution may have rights to this item. Sign in to continue.
- Title
Deep semantic hashing with dual attention for cross-modal retrieval.
- Authors
Wu, Jiagao; Weng, Weiwei; Fu, Junxia; Liu, Linfeng; Hu, Bin
- Abstract
With the explosive growth of multimodal data, cross-modal retrieval has drawn increasing research interests. Hashing-based methods have made great advancements in cross-modal retrieval due to the benefits of low storage cost and fast query speed. However, there still exists a crucial challenge to improve the accuracy of cross-modal retrieval due to the heterogeneity gap between modalities. To further tackle this problem, in this paper, we propose a new two-staged cross-modal retrieval method, called Deep Semantic Hashing with Dual Attention (DSHDA). In the first stage of DSHDA, a Semantic Label Network (SeLabNet) is designed to extract label semantic features and hash codes by training the multi-label annotations, which can make the learning of different modalities in a common semantic space and bridge the modality gap effectively. In the second stage of DSHDA, we propose a deep neural network to simultaneously integrate feature and hash code learning for each modality into the same framework, the training of the framework is guided by the label semantic features and hash codes generated from SeLabNet to maximize the cross-modal semantic relevance. Moreover, dual attention mechanisms are used in our neural networks: (1) Lo-attention is used to extract the local key information of each modality and improve the quality of modality features. (2) Co-attention is used to strengthen the relationship between different modalities to produce more consistent and accurate hash codes. Extensive experiments on two real datasets with image-text modalities demonstrate the superiority of the proposed method in cross-modal retrieval tasks.
- Subjects
AUTOMATED storage retrieval systems; LABEL design; COMPUTER programming education; MACHINE learning
- Publication
Neural Computing & Applications, 2022, Vol 34, Issue 7, p5397
- ISSN
0941-0643
- Publication type
Article
- DOI
10.1007/s00521-021-06696-y