SCH-GAN: Semi-Supervised Cross-Modal Hashing by Generative Adversarial Network

被引:92
|
作者
Zhang, Jian [1 ]
Peng, Yuxin [1 ]
Yuan, Mingkuan [1 ]
机构
[1] Peking Univ, Inst Comp Sci & Technol, Beijing 100871, Peoples R China
基金
中国国家自然科学基金;
关键词
Semantics; Data models; Correlation; Generative adversarial networks; Training data; Predictive models; Gallium nitride; Cross-modal hashing; generative adversarial network (GAN); semi-supervised; CODES;
D O I
10.1109/TCYB.2018.2868826
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Cross-modal hashing maps heterogeneous multimedia data into a common Hamming space to realize fast and flexible cross-modal retrieval. Supervised cross-modal hashing methods have achieved considerable progress by incorporating semantic side information. However, they heavily rely on large-scale labeled cross-modal training data which are hard to obtain, since multiple modalities are involved. They also ignore the rich information contained in the large amount of unlabeled data across different modalities, which can help to model the correlations between different modalities. To address these problems, in this paper, we propose a novel semi-supervised cross-modal hashing approach by generative adversarial network (SCH-GAN). The main contributions can be summarized as follows: 1) we propose a novel generative adversarial network for cross-modal hashing, in which the generative model tries to select margin examples of one modality from unlabeled data when given a query of another modality (e.g., giving a text query to retrieve images and vice versa). The discriminative model tries to distinguish the selected examples and true positive examples of the query. These two models play a minimax game so that the generative model can promote the hashing performance of the discriminative model and 2) we propose a reinforcement learning-based algorithm to drive the training of proposed SCH-GAN. The generative model takes the correlation score predicted by discriminative model as a reward, and tries to select the examples close to the margin to promote a discriminative model. Extensive experiments verify the effectiveness of our proposed approach, compared with nine state-of-the-art methods on three widely used datasets.
引用
收藏
页码:489 / 502
页数:14
相关论文
共 50 条
  • [31] Self-Supervised Adversarial Hashing Networks for Cross-Modal Retrieval
    Li, Chao
    Deng, Cheng
    Li, Ning
    Liu, Wei
    Gao, Xinbo
    Tao, Dacheng
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 4242 - 4251
  • [32] A semi-supervised framework with generative adversarial network for pansharpening
    Yu-Xuan Wang
    Ting-Zhu Huang
    Ran Ran
    Rui Wen
    Liang-Jian Deng
    Signal, Image and Video Processing, 2025, 19 (6)
  • [33] BL-GAN: Semi-Supervised Bug Localization via Generative Adversarial Network
    Zhu, Ziye
    Tong, Hanghang
    Wang, Yu
    Li, Yun
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (11) : 11112 - 11125
  • [34] Semi-supervised Deep Quantization for Cross-modal Search
    Wang, Xin
    Zhu, Wenwu
    Liu, Chenghao
    PROCEEDINGS OF THE 27TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA (MM'19), 2019, : 1730 - 1739
  • [35] Semi-Supervised Cross-Modal Retrieval With Label Prediction
    Mandal, Devraj
    Rao, Pramod
    Biswas, Soma
    IEEE TRANSACTIONS ON MULTIMEDIA, 2020, 22 (09) : 2345 - 2353
  • [36] Enhancing Semi-Supervised Learning with Cross-Modal Knowledge
    Zhu, Hui
    Lu, Yongchun
    Wang, Hongbin
    Zhou, Xunyi
    Ma, Qin
    Liu, Yanhong
    Jiang, Ning
    Wei, Xin
    Zeng, Linchengxi
    Zhao, Xiaofang
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 4456 - 4465
  • [37] R2GAN: Cross-modal Recipe Retrieval with Generative Adversarial Network
    Zhu, Bin
    Ngo, Chong-Wah
    Chen, Jingjing
    Hao, Yanbin
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 11469 - 11478
  • [38] DA-GAN: Dual Attention Generative Adversarial Network for Cross-Modal Retrieval
    Cai, Liewu
    Zhu, Lei
    Zhang, Hongyan
    Zhu, Xinghui
    FUTURE INTERNET, 2022, 14 (02)
  • [39] Supervised Hierarchical Cross-Modal Hashing
    Sun, Changchang
    Song, Xuemeng
    Feng, Fuli
    Zhao, Wayne Xin
    Zhang, Hao
    Nie, Liqiang
    PROCEEDINGS OF THE 42ND INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '19), 2019, : 725 - 734
  • [40] Weakly Supervised Cross-Modal Hashing
    Liu, Xuanwu
    Yu, Guoxian
    Domeniconi, Carlotta
    Wang, Jun
    Xiao, Guoqiang
    Guo, Maozu
    IEEE TRANSACTIONS ON BIG DATA, 2022, 8 (02) : 552 - 563