Adversarial Modality Alignment Network for Cross-Modal Molecule Retrieval

被引:7
|
作者
Zhao W. [1 ,2 ]
Zhou D. [3 ]
Cao B. [1 ]
Zhang K. [2 ]
Chen J. [2 ]
机构
[1] Hunan University of Science and Technology, School of Computer Science and Engineering, Xiangtan
[2] Swinburne University of Technology, Department of Computing Technologies, Melbourne, 3122, VIC
[3] Guangdong University of Foreign Studies, School of Information Science and Technology, Guangzhou
来源
关键词
Cross-modal molecule retrieval (Text2Mol); graph transformer network (GTN); modality alignment; molecule representation;
D O I
10.1109/TAI.2023.3254518
中图分类号
学科分类号
摘要
The cross-modal molecule retrieval (Text2Mol) task aims to bridge the semantic gap between molecules and natural language descriptions. A solution to this nontrivial problem relies on a graph convolutional network (GCN) and cross-modal attention with contrastive learning for reasonable results. However, there exist the following issues. First, the cross-modal attention mechanism is only in favor of text representations and cannot provide helpful information for molecule representations. Second, the GCN-based molecule encoder ignores edge features and the importance of various substructures of a molecule. Finally, the retrieval learning loss function is rather simplistic. This article further investigates the Text2Mol problem and proposes a novel adversarial modality alignment network (AMAN) based method to sufficiently learn both description and molecule information. Our method utilizes a SciBERT as a text encoder and a graph transformer network as a molecule encoder to generate multimodal representations. Then, an adversarial network is used to align these modalities interactively. Meanwhile, a triplet loss function is leveraged to perform retrieval learning and further enhance the modality alignment. Experiments on the ChEBI-20 dataset show the effectiveness of our AMAN method compared with baselines. © 2020 IEEE.
引用
收藏
页码:278 / 289
页数:11
相关论文
共 50 条
  • [21] Token Embeddings Alignment for Cross-Modal Retrieval
    Xie, Chen-Wei
    Wu, Jianmin
    Zheng, Yun
    Pan, Pan
    Hua, Xian-Sheng
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 4555 - 4563
  • [22] Robust cross-modal retrieval with alignment refurbishment
    Guo, Jinyi
    Ding, Jieyu
    FRONTIERS OF INFORMATION TECHNOLOGY & ELECTRONIC ENGINEERING, 2023, 24 (10) : 1403 - 1415
  • [23] Adequate alignment and interaction for cross-modal retrieval
    Mingkang WANG
    Min MENG
    Jigang LIU
    Jigang WU
    虚拟现实与智能硬件(中英文), 2023, 5 (06) : 509 - 522
  • [24] Modality-specific adaptive scaling and attention network for cross-modal retrieval
    Ke, Xiao
    Chen, Baitao
    Cai, Yuhang
    Liu, Hao
    Guo, Wenzhong
    Chen, Weibin
    NEUROCOMPUTING, 2025, 612
  • [25] Adversarial Tri-Fusion Hashing Network for Imbalanced Cross-Modal Retrieval
    Liu, Xin
    Cheung, Yiu-ming
    Hu, Zhikai
    He, Yi
    Zhong, Bineng
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2021, 5 (04): : 607 - 619
  • [26] Adversarial cross-modal retrieval based on dictionary learning
    Shang, Fei
    Zhang, Huaxiang
    Zhu, Lei
    Sun, Jiande
    NEUROCOMPUTING, 2019, 355 : 93 - 104
  • [27] Adversarial Cross-Modal Retrieval Based on Association Constraint
    Guo Q.
    Qian Y.
    Liang X.
    Moshi Shibie yu Rengong Zhineng/Pattern Recognition and Artificial Intelligence, 2021, 34 (01): : 68 - 76
  • [28] Representation separation adversarial networks for cross-modal retrieval
    Deng, Jiaxin
    Ou, Weihua
    Gou, Jianping
    Song, Heping
    Wang, Anzhi
    Xu, Xing
    WIRELESS NETWORKS, 2024, 30 (05) : 3469 - 3481
  • [29] Adaptive Adversarial Learning based cross-modal retrieval
    Li, Zhuoyi
    Lu, Huibin
    Fu, Hao
    Wang, Zhongrui
    Gu, Guanghun
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 123
  • [30] Discrete Fusion Adversarial Hashing for cross-modal retrieval
    Li, Jing
    Yu, En
    Ma, Jianhua
    Chang, Xiaojun
    Zhang, Huaxiang
    Sun, Jiande
    KNOWLEDGE-BASED SYSTEMS, 2022, 253