GNDAN: Graph Navigated Dual Attention Network for Zero-Shot Learning

被引:25
|
作者
Chen, Shiming [1 ]
Hong, Ziming [1 ]
Xie, Guosen [2 ]
Peng, Qinmu [1 ]
You, Xinge [1 ]
Ding, Weiping [3 ]
Shao, Ling [4 ]
机构
[1] Huazhong Univ Sci & Technol, Sch Elect Informat & Commun, Wuhan 430074, Peoples R China
[2] Nanjing Univ Sci & Technol, Sch Comp Sci & Engn, Nanjing, Peoples R China
[3] Nantong Univ, Sch Informat Sci & Technol, Nantong 226019, Peoples R China
[4] Saudi Data & Artificial Intelligence Author SDAIA, Natl Ctr Artificial Intelligence NCAI, Riyadh, Saudi Arabia
基金
中国国家自然科学基金;
关键词
Semantics; Visualization; Feature extraction; Task analysis; Knowledge transfer; Navigation; Learning systems; Attribute-based region features; graph attention network (GAT); graph neural network (GNN); zero-shot learning (ZSL);
D O I
10.1109/TNNLS.2022.3155602
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Zero-shot learning (ZSL) tackles the unseen class recognition problem by transferring semantic knowledge from seen classes to unseen ones. Typically, to guarantee desirable knowledge transfer, a direct embedding is adopted for associating the visual and semantic domains in ZSL. However, most existing ZSL methods focus on learning the embedding from implicit global features or image regions to the semantic space. Thus, they fail to: 1) exploit the appearance relationship priors between various local regions in a single image, which corresponds to the semantic information and 2) learn cooperative global and local features jointly for discriminative feature representations. In this article, we propose the novel graph navigated dual attention network (GNDAN) for ZSL to address these drawbacks. GNDAN employs a region-guided attention network (RAN) and a region-guided graph attention network (RGAT) to jointly learn a discriminative local embedding and incorporate global context for exploiting explicit global embeddings under the guidance of a graph. Specifically, RAN uses soft spatial attention to discover discriminative regions for generating local embeddings. Meanwhile, RGAT employs an attribute-based attention to obtain attribute-based region features, where each attribute focuses on the most relevant image regions. Motivated by the graph neural network (GNN), which is beneficial for structural relationship representations, RGAT further leverages a graph attention network to exploit the relationships between the attribute-based region features for explicit global embedding representations. Based on the self-calibration mechanism, the joint visual embedding learned is matched with the semantic embedding to form the final prediction. Extensive experiments on three benchmark datasets demonstrate that the proposed GNDAN achieves superior performances to the state-of-the-art methods. Our code and trained models are available at https://github.com/shiming-chen/GNDAN.
引用
收藏
页码:4516 / 4529
页数:14
相关论文
共 50 条
  • [1] Attribute Propagation Network for Graph Zero-Shot Learning
    Liu, Lu
    Zhou, Tianyi
    Long, Guodong
    Jiang, Jing
    Zhang, Chengqi
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 4868 - 4875
  • [2] HAPZSL: A hybrid attention prototype network for knowledge graph zero-shot relational learning
    Li, Xuewei
    Ma, Jinming
    Yu, Jian
    Xu, Tianyi
    Zhao, Mankun
    Liu, Hongwei
    Yu, Mei
    Yu, Ruiguo
    NEUROCOMPUTING, 2022, 508 : 324 - 336
  • [3] Dual Part Discovery Network for Zero-Shot Learning
    Ge, Jiannan
    Xie, Hongtao
    Min, Shaobo
    Li, Pandeng
    Zhang, Yongdong
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 3244 - 3252
  • [4] Dual triplet network for image zero-shot learning
    Ji, Zhong
    Wang, Hai
    Pang, Yanwei
    Shao, Ling
    NEUROCOMPUTING, 2020, 373 : 90 - 97
  • [5] Dual-verification network for zero-shot learning
    Zhang, Haofeng
    Long, Yang
    Yang, Wankou
    Shao, Ling
    INFORMATION SCIENCES, 2019, 470 : 43 - 57
  • [6] Attention-Based Graph Convolutional Network for Zero-Shot Learning with Pre-Training
    Wu, Xuefei
    Liu, Mingjiang
    Xin, Bo
    Zhu, Zhangqing
    Wang, Gang
    MATHEMATICAL PROBLEMS IN ENGINEERING, 2021, 2021
  • [7] Learning Multipart Attention Neural Network for Zero-Shot Classification
    Meng, Min
    Wei, Jie
    Wu, Jigang
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2022, 14 (02) : 414 - 423
  • [8] Dual Prototype Contrastive Network for Generalized Zero-Shot Learning
    Jiang, Huajie
    Li, Zhengxian
    Hu, Yongli
    Yin, Baocai
    Yang, Jian
    van den Hengel, Anton
    Yang, Ming-Hsuan
    Qi, Yuankai
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2025, 35 (02) : 1111 - 1122
  • [9] Dual Expert Distillation Network for Generalized Zero-Shot Learning
    Rao, Zhijie
    Guo, Jingcai
    Lu, Xiaocheng
    Liang, Jingming
    Zhang, Jie
    Wang, Haozhao
    Wei, Kang
    Cao, Xiaofeng
    PROCEEDINGS OF THE THIRTY-THIRD INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2024, 2024, : 4833 - 4841
  • [10] Dual-focus transfer network for zero-shot learning
    Jia, Zhen
    Zhang, Zhang
    Shan, Caifeng
    Wang, Liang
    Tan, Tieniu
    NEUROCOMPUTING, 2023, 541