Skip to content
/ DSAH Public

Deep Semantic-Alignment Hashing(ICMR2020, Oral)

Notifications You must be signed in to change notification settings

idejie/DSAH

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

DSAH

Source code of our ICMR 2020 paper "Deep Semantic-Alignment Hashing for Unsupervised Cross-Modal Retrieval"

Introduction

Deep hashing methods have achieved tremendous success in cross-modal retrieval, due to its low storage consumption and fast retrieval speed. In real cross-modal retrieval applications, it's hard to obtain label information. Recently, increasing attention has been paid to unsupervised cross-modal hashing. However, existing methods fail to exploit the intrinsic connections between images and their corresponding descriptions or tags (text modality). In this paper, we propose a novel Deep Semantic-Alignment Hashing (DSAH) for unsupervised cross-modal retrieval, which sufficiently utilizes the co-occurred image-text pairs. DSAH explores the similarity information of different modalities and we elaborately design a semantic-alignment loss function, which elegantly aligns the similarities between features with those between hash codes. Moreover, to further bridge the modality gap, we innovatively propose to reconstruct features of one modality with hash codes of the other one. Extensive experiments on three cross-modal retrieval datasets demonstrate that DSAH achieves the state-of-the-art performance. Framework

Requirements

  • Python: 3.x
  • other dependencies: env.yaml

Run

  • Update the setting.py with your data_dir. And change the value EVAL, for train setting it with False
  • run the train.py
    python train.py
  • For test, set the value EVAL with True. And the model will load the checkpoint/DATASET_CODEBIT_bit_best_epoch.pth

Datasets

For datasets, we follow Deep Cross-Modal Hashing's Github (Jiang, CVPR 2017). You can download these datasets from:

Citation

If you find this code useful, please cite our paper:

@inproceedings{10.1145/3372278.3390673,
author = {Yang, Dejie and Wu, Dayan and Zhang, Wanqian and Zhang, Haisu and Li, Bo and Wang, Weiping},
title = {Deep Semantic-Alignment Hashing for Unsupervised Cross-Modal Retrieval},
year = {2020},
isbn = {9781450370875},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
url = {https://doi.org/10.1145/3372278.3390673},
doi = {10.1145/3372278.3390673},
booktitle = {Proceedings of the 2020 International Conference on Multimedia Retrieval},
pages = {44–52},
numpages = {9},
keywords = {cross-modal hashing, cross-media retrieval, semantic-alignment},
location = {Dublin, Ireland},
series = {ICMR ’20}
}

All rights are reserved by the authors.

References

About

Deep Semantic-Alignment Hashing(ICMR2020, Oral)

Resources

Stars

Watchers

Forks

Languages