Aspect Based Sentiment Analysis on Multimodal Data: A Transformer and Low-Rank Fusion Approach

被引:0
|
作者
Jin, Meilin [1 ]
Shao, Lianhe [1 ]
Wang, Xihan [1 ]
Yan, Qianqian [1 ]
Chu, Zhulu [1 ]
Luo, Tongtong [1 ]
Tang, Jiacheng [1 ]
Gao, Quanli [1 ]
机构
[1] Xian Polytech Univ, Sch Comp Sci, Xian, Peoples R China
关键词
multimodal sentiment analysis; cross-modal attention mechanism; low-rank fusion;
D O I
10.1109/CCAI61966.2024.10603022
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Video has become the main way for users to share their daily lives and express their opinions, most of which reflect their emotional information. Sentiment analysis of videos can be used to understand user behavior and thus provide better improvements and services. However, due to various reasons, most of the existing sentiment analysis is based on unimodal, and the accuracy of unimodal-based sentiment analysis is low and prone to ambiguity. Some existing multimodal sentiment analysis methods lack effective interaction between modalities. This paper proposes a model called Aspect Based Sentiment Analysis on Multimodal Data: A Transformer and Low-Rank Fusion Approach (ABSA-TLRF). Specifically, ABSA-TLRF utilizes a cross-modal alignment module based on the cross-modal attention mechanism and an efficient method based on low-rank fusion to effectively integrate information within and between modalities. This achieves global-local information interaction, ultimately yielding more accurate emotion fusion results. The model performs sentiment classification, enabling high-level multimodal sentiment analysis. Experimental results indicate that our model outperforms several state-of-the-art methods on three commonly used datasets. Our research suggests that combining a cross-modal alignment module based on the cross-modal attention mechanism and an efficient method based on low-rank fusion can enhance our understanding of multimodal content, thereby improving sentiment analysis performance.
引用
收藏
页码:332 / 338
页数:7
相关论文
共 50 条
  • [21] Multimodal Medical Image Fusion Based on Parameter Adaptive PCNN and Latent Low-rank Representation
    WANG Wenyan
    ZHOU Xianchun
    YANG Liangjian
    Instrumentation, 2023, 10 (01) : 45 - 58
  • [22] Fusion of Hyperspectral and LiDAR Data Using Sparse and Low-Rank Component Analysis
    Rasti, Behnood
    Ghamisi, Pedram
    Plaza, Javier
    Plaza, Antonio
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2017, 55 (11): : 6354 - 6365
  • [23] Multimodal Aspect-Based Sentiment Classification with Knowledge-Injected Transformer
    Xu, Zenan
    Su, Qinliang
    Xiao, Junxi
    2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 1379 - 1384
  • [24] A Unimodal Reinforced Transformer With Time Squeeze Fusion for Multimodal Sentiment Analysis
    He, Jiaxuan
    Mai, Sijie
    Hu, Haifeng
    IEEE SIGNAL PROCESSING LETTERS, 2021, 28 : 992 - 996
  • [25] TETFN: A text enhanced transformer fusion network for multimodal sentiment analysis
    Wang, Di
    Guo, Xutong
    Tian, Yumin
    Liu, Jinhui
    He, LiHuo
    Luo, Xuemei
    PATTERN RECOGNITION, 2023, 136
  • [26] Multimodal Sentiment Analysis via Low-Rank Tensor Attention Network with Unimodal Self-Supervised Learning
    Pan, Jie (panjie@sdnu.edu.cn), 1600, Institute of Electrical and Electronics Engineers Inc.
  • [27] Research on Multimodal Aspect-Based Sentiment Analysis Based on Image Caption and Multimodal Aspect Extraction
    Huang, Peng
    Tao, Jun
    Su, Tengrong
    Zhang, Xiaoqing
    2023 35TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2023, : 5415 - 5418
  • [28] Review of Multimodal Sensor Data Fusion in Sentiment Analysis
    Jin, Yelei
    Gulanbaier, Tuerhong
    Mairidan, Wushouer
    Computer Engineering and Applications, 2023, 59 (23) : 1 - 14
  • [29] TCMT: Target-oriented Cross Modal Transformer for Multimodal Aspect-Based Sentiment Analysis
    Zou, Wang
    Sun, Xia
    Wu, Wenhuan
    Lu, Qiang
    Zhao, Xiaodi
    Bo, Qirong
    Yan, Jianqiang
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 264
  • [30] A conditioned joint-modality attention fusion approach for multimodal aspect-level sentiment analysis
    Yang, Ying
    Qian, Xinyu
    Zhang, Lingfeng
    Tang, Si
    Zhao, Qinna
    JOURNAL OF SUPERCOMPUTING, 2024, 80 (15): : 21676 - 21698