Semi-supervised cross-modal representation learning with GAN-based Asymmetric Transfer Network

被引:1
|
作者
Zhang, Lei [1 ,2 ]
Chen, Leiting [1 ,2 ,3 ]
Ou, Weihua [4 ]
Zhou, Chuan [1 ,2 ]
机构
[1] Univ Elect Sci & Technol China, Sch Comp Sci & Engn, Chengdu, Peoples R China
[2] Univ Elect Sci & Technol China, Digital Media Technol Key Lab Sichuan Prov, Chengdu, Peoples R China
[3] Inst Elect & Informat Engn UESTC Guangdong, Dongguan, Peoples R China
[4] Guizhou Normal Univ, Sch Big Data & Comp Sci, Guiyang, Peoples R China
基金
中国国家自然科学基金;
关键词
Cross-modal retrieval; Modality gap; Generative adversarial network;
D O I
10.1016/j.jvcir.2020.102899
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we proposed a semi-supervised common representation learning method with GAN-based Asymmetric Transfer Network (GATN) for cross modality retrieval. GATN utilizes the asymmetric pipeline to guarantee the semantic consistency and adopt (Generative Adversarial Network) GAN to fit the distributions of different modalities. Specifically, the common representation learning across modalities includes two stages: (1) the first stage, GATN trains source mapping network to learn the semantic representation of text modality by supervised method; and (2) the second stage, GAN-based unsupervised modality transfer method is proposed to guide the training of target mapping network, which includes generative network (target mapping network) and discriminative network. Experimental results on three widely-used benchmarks show that GATN have achieved better performance comparing with several existing state-of-the-art methods.
引用
收藏
页数:9
相关论文
共 50 条
  • [41] The Network Representation Learning Algorithm Based on Semi-Supervised Random Walk
    Liu, Dong
    Li, Qinpeng
    Ru, Yan
    Zhang, Jun
    IEEE ACCESS, 2020, 8 : 222956 - 222965
  • [42] Coupled feature selection based semi-supervised modality-dependent cross-modal retrieval
    En Yu
    Jiande Sun
    Li Wang
    Wenbo Wan
    Huaxiang Zhang
    Multimedia Tools and Applications, 2019, 78 : 28931 - 28951
  • [43] X-ModalNet: A semi-supervised deep cross-modal network for classification of remote sensing data
    Hong, Danfeng
    Yokoya, Naoto
    Xia, Gui-Song
    Chanussot, Jocelyn
    Zhu, Xiao Xiang
    ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2020, 167 : 12 - 23
  • [44] Coupled feature selection based semi-supervised modality-dependent cross-modal retrieval
    Yu, En
    Sun, Jiande
    Wang, Li
    Wan, Wenbo
    Zhang, Huaxiang
    MULTIMEDIA TOOLS AND APPLICATIONS, 2019, 78 (20) : 28931 - 28951
  • [45] SEMI-SUPERVISED SEMANTIC-PRESERVING HASHING FOR EFFICIENT CROSS-MODAL RETRIEVAL
    Wang, Xingzhi
    Liu, Xin
    Hu, Zhikai
    Wang, Nannan
    Fan, Wentao
    Du, Ji-Xiang
    2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2019, : 1006 - 1011
  • [46] Semantic supervised learning based Cross-Modal Retrieval
    Li, Zhuoyi
    Fu, Hao
    Gu, Guanghua
    PROCEEDINGS OF THE ACM TURING AWARD CELEBRATION CONFERENCE-CHINA 2024, ACM-TURC 2024, 2024, : 207 - 209
  • [47] APPLICATION OF GAN IN SEMI-SUPERVISED LEARNING
    Lai, Zhilong
    Li, Jianping
    Wang, Qingsong
    Guo, Longyuan
    2019 16TH INTERNATIONAL COMPUTER CONFERENCE ON WAVELET ACTIVE MEDIA TECHNOLOGY AND INFORMATION PROCESSING (ICWAMTIP), 2019, : 254 - 257
  • [48] Duplicate Image Representation Based on Semi-Supervised Learning
    Chen, Ming
    Yan, Jinghua
    Gao, Tieliang
    Li, Yuhua
    Ma, Huan
    INTERNATIONAL JOURNAL OF GRID AND HIGH PERFORMANCE COMPUTING, 2022, 14 (01)
  • [49] Learning Semi-Supervised Representation Towards a Unified Optimization Framework for Semi-Supervised Learning
    Li, Chun-Guang
    Lin, Zhouchen
    Zhang, Honggang
    Guo, Jun
    2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, : 2767 - 2775
  • [50] Intra-class low-rank regularization for supervised and semi-supervised cross-modal retrieval
    Kang, Peipei
    Lin, Zehang
    Yang, Zhenguo
    Fang, Xiaozhao
    Bronstein, Alexander M.
    Li, Qing
    Liu, Wenyin
    APPLIED INTELLIGENCE, 2022, 52 (01) : 33 - 54