Visual-textual sentiment classification with bi-directional multi-level attention networks

被引:49
|
作者
Xu, Jie [1 ]
Huang, Feiran [2 ,3 ,4 ]
Zhang, Xiaoming [5 ]
Wang, Senzhang [6 ]
Li, Chaozhuo [1 ]
Li, Zhoujun [1 ]
He, Yueying [7 ]
机构
[1] Beihang Univ, Sch Comp Sci & Engn, State Key Lab Software Dev Environm, Beijing 100191, Peoples R China
[2] Jinan Univ, Coll Cyber Secur, Guangzhou 510632, Guangdong, Peoples R China
[3] Jinan Univ, Coll Informat Sci & Technol, Guangzhou 510632, Guangdong, Peoples R China
[4] Guangdong Key Lab Data Secur & Privacy Preserving, Guangzhou 510632, Guangdong, Peoples R China
[5] Beihang Univ, Sch Cyber Sci & Technol, Beijing 100191, Peoples R China
[6] Nanjing Univ Aeronaut & Astronaut, Sch Comp Sci & Technol, Nanjing 210016, Jiangsu, Peoples R China
[7] Coordinat Ctr China, Natl Comp Network Emergency Response Tech Team, Beijing 100029, Peoples R China
基金
北京市自然科学基金; 中国国家自然科学基金;
关键词
Multi-modal; Social image; Attention model; Sentiment analysis;
D O I
10.1016/j.knosys.2019.04.018
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Social network has become an inseparable part of our daily lives and thus the automatic sentiment analysis on social media content is of great significance to identify people's viewpoints, attitudes, and emotions on the social websites. Most existing works have concentrated on the sentiment analysis of single modality such as image or text, which cannot handle the social media content with multiple modalities including both image and text. Although some works tried to conduct multi modal sentiment analysis, the complicated correlations between the two modalities have not been fully explored. In this paper, we propose a novel Bi-Directional Multi-Level Attention (BDMLA) model to exploit the complementary and comprehensive information between the image modality and text modality for joint visual-textual sentiment classification. Specifically, to highlight the emotional regions and words in the image-text pair, visual attention network and semantic attention network are proposed respectively. The visual attention network makes region features of the image interact with multiple semantic levels of text (word, phrase, and sentence) to obtain the attended visual features. The semantic attention network makes semantic features of the text interact with multiple visual levels of image (global and local) to obtain the attended semantic features. Then, the attended visual and semantic features from the two attention networks are unified into a holistic framework to conduct visual-textual sentiment classification. Proof-of-concept experiments conducted on three real-world datasets verify the effectiveness of our model. (C) 2019 Elsevier B.V. All rights reserved.
引用
收藏
页码:61 / 73
页数:13
相关论文
共 50 条
  • [21] Bi-directional LSTM with multi-scale dense attention mechanism for hyperspectral image classification
    Gao, Jinxiong
    Gao, Xiumei
    Wu, Nan
    Yang, Hongye
    MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (17) : 24003 - 24020
  • [22] Multi-level, uni-directional AC-DC converters, a cost effective alternative to bi-directional converters
    Carlton, D
    Dunford, WG
    PESC 2001: 32ND ANNUAL POWER ELECTRONICS SPECIALISTS CONFERENCE, VOLS 1-4, CONFERENCE PROCEEDINGS, 2001, : 1911 - 1916
  • [23] Temporal Textual Localization in Video via Adversarial Bi-Directional Interaction Networks
    Zhang, Zijian
    Zhao, Zhou
    Zhang, Zhu
    Lin, Zhijie
    Wang, Qi
    Hong, Richang
    IEEE TRANSACTIONS ON MULTIMEDIA, 2021, 23 : 3306 - 3317
  • [24] Visual Relation Detection with Multi-Level Attention
    Zheng, Sipeng
    Chen, Shizhe
    Jin, Qin
    PROCEEDINGS OF THE 27TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA (MM'19), 2019, : 121 - 129
  • [25] Droop Control for Bi-Directional DC-DC Converters Used in Multi-Level Virtual Conductors
    Takahashi, Tomohiro
    Miyamoto, Masakazu
    Dousoky, Gamal M.
    Shoyama, Masahito
    2015 IEEE INTERNATIONAL TELECOMMUNICATIONS ENERGY CONFERENCE (INTELEC), 2015,
  • [26] Electricity demand error corrections with attention bi-directional neural networks
    Ghimire, Sujan
    Deo, Ravinesh C.
    Casillas-Perez, David
    Salcedo-Sanz, Sancho
    ENERGY, 2024, 291
  • [27] Multi-Level Attention Map Network for Multimodal Sentiment Analysis
    Xue, Xiaojun
    Zhang, Chunxia
    Niu, Zhendong
    Wu, Xindong
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (05) : 5105 - 5118
  • [28] Exploiting bi-directional deep neural networks for multi-domain sentiment analysis using capsule network
    Ghorbanali, Alireza
    Sohrabi, Mohammad Karim
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 82 (15) : 22943 - 22960
  • [29] Exploiting bi-directional deep neural networks for multi-domain sentiment analysis using capsule network
    Alireza Ghorbanali
    Mohammad Karim Sohrabi
    Multimedia Tools and Applications, 2023, 82 : 22943 - 22960
  • [30] Multi-level textual-visual alignment and fusion network for multimodal aspect-based sentiment analysis
    Li, You
    Ding, Han
    Lin, Yuming
    Feng, Xinyu
    Chang, Liang
    ARTIFICIAL INTELLIGENCE REVIEW, 2024, 57 (04)