Aspect Based Sentiment Analysis on Multimodal Data: A Transformer and Low-Rank Fusion Approach

被引:0
|
作者
Jin, Meilin [1 ]
Shao, Lianhe [1 ]
Wang, Xihan [1 ]
Yan, Qianqian [1 ]
Chu, Zhulu [1 ]
Luo, Tongtong [1 ]
Tang, Jiacheng [1 ]
Gao, Quanli [1 ]
机构
[1] Xian Polytech Univ, Sch Comp Sci, Xian, Peoples R China
关键词
multimodal sentiment analysis; cross-modal attention mechanism; low-rank fusion;
D O I
10.1109/CCAI61966.2024.10603022
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Video has become the main way for users to share their daily lives and express their opinions, most of which reflect their emotional information. Sentiment analysis of videos can be used to understand user behavior and thus provide better improvements and services. However, due to various reasons, most of the existing sentiment analysis is based on unimodal, and the accuracy of unimodal-based sentiment analysis is low and prone to ambiguity. Some existing multimodal sentiment analysis methods lack effective interaction between modalities. This paper proposes a model called Aspect Based Sentiment Analysis on Multimodal Data: A Transformer and Low-Rank Fusion Approach (ABSA-TLRF). Specifically, ABSA-TLRF utilizes a cross-modal alignment module based on the cross-modal attention mechanism and an efficient method based on low-rank fusion to effectively integrate information within and between modalities. This achieves global-local information interaction, ultimately yielding more accurate emotion fusion results. The model performs sentiment classification, enabling high-level multimodal sentiment analysis. Experimental results indicate that our model outperforms several state-of-the-art methods on three commonly used datasets. Our research suggests that combining a cross-modal alignment module based on the cross-modal attention mechanism and an efficient method based on low-rank fusion can enhance our understanding of multimodal content, thereby improving sentiment analysis performance.
引用
收藏
页码:332 / 338
页数:7
相关论文
共 50 条
  • [41] Multimodal Sentiment Analysis Based on Composite Hierarchical Fusion
    Lei, Yu
    Qu, Keshuai
    Zhao, Yifan
    Han, Qing
    Wang, Xuguang
    COMPUTER JOURNAL, 2024, 67 (06): : 2230 - 2245
  • [42] Multimodal sentiment analysis based on fusion methods: A survey
    Zhu, Linan
    Zhu, Zhechao
    Zhang, Chenwei
    Xu, Yifei
    Kong, Xiangjie
    INFORMATION FUSION, 2023, 95 : 306 - 325
  • [43] Multimodal sparse and low-rank subspace clustering
    Abavisani, Mahdi
    Patel, Vishal M.
    INFORMATION FUSION, 2018, 39 : 168 - 177
  • [44] Survey of Sentiment Analysis Algorithms Based on Multimodal Fusion
    Guo, Xu
    Mairidan, Wushouer
    Gulanbaier, Tuerhong
    Computer Engineering and Applications, 2024, 60 (02) : 1 - 18
  • [45] AMIFN: Aspect-guided multi-view interactions and fusion network for multimodal aspect-based sentiment analysis°
    Yang, Juan
    Xu, Mengya
    Xiao, Yali
    Du, Xu
    NEUROCOMPUTING, 2024, 573
  • [46] AMIFN: Aspect-guided multi-view interactions and fusion network for multimodal aspect-based sentiment analysis
    Yang, Juan
    Xu, Mengya
    Xiao, Yali
    Du, Xu
    Neurocomputing, 2024, 573
  • [47] An Approach for Detecting Band Data in Smart Grid Based on Low-Rank Multi-View Analysis
    Li Y.-P.
    Peng W.-L.
    Men K.
    Wu J.-Y.
    Dianzi Keji Daxue Xuebao/Journal of the University of Electronic Science and Technology of China, 2019, 48 (03): : 361 - 365
  • [48] MMTF-DES: A fusion of multimodal transformer models for desire, emotion, and sentiment analysis of social media data
    Aziz, Abdul
    Chowdhury, Nihad Karim
    Kabir, Muhammad Ashad
    Chy, Abu Nowshed
    Siddique, Md. Jawad
    NEUROCOMPUTING, 2025, 623
  • [49] LORTSAR: Low-Rank Transformer for Skeleton-Based Action Recognition
    Oraki, Soroush
    Zhuang, Harry
    Liang, Jie
    ADVANCES IN VISUAL COMPUTING, ISVC 2024, PT I, 2025, 15046 : 196 - 207
  • [50] Dynamic Low-rank Estimation for Transformer-based Language Models
    Huai, Ting
    Lie, Xiao
    Gao, Shangqian
    Hsu, Yenchang
    Shen, Yilin
    Jin, Hongxia
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 9275 - 9287