Momentum Distillation Improves Multimodal Sentiment Analysis

被引:2
|
作者
Li, Siqi [1 ]
Deng, Weihong [1 ]
Hu, Jiani [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Sch Artificial Intelligence, Beijing 100876, Peoples R China
关键词
Multimodal sentiment analysis; Sarcasm detection; Momentum distillation;
D O I
10.1007/978-3-031-18907-4_33
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the development of computer technology, the Internet floods with abundant multimodal data. For better understanding users' feelings, multimodal sentiment analysis and sarcasm detection have become popular research topics. However, previous studies did not take noise into account when designing models. In this paper, based on designing a novel architecture, we also introduce a momentum distillation method to improve the model's performance from noisy data. Specifically, we propose the Transformer-Based Network with Momentum Distillation (TBNMD). For model architecture, we first encode different modalities to obtain hidden representations. Then we use a multimodal interaction module to obtain text-guided image features and image-guided text features. After that, we use a multimodal fusion module to obtain the fusion features. For momentum distillation, it is a self-distillation method. During the training process, the teacher model generates semantically similar samples as additional supervision of the student model. Experimental results on five publicly available datasets demonstrate the effectiveness of our method.
引用
收藏
页码:423 / 435
页数:13
相关论文
共 50 条
  • [1] Multi-Task Momentum Distillation for Multimodal Sentiment Analysis
    Lin, Ronghao
    Hu, Haifeng
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2024, 15 (02) : 549 - 565
  • [2] Adaptive Modality Distillation for Separable Multimodal Sentiment Analysis
    Peng, Wei
    Zhao, Guoying
    Hong, Xiaopeng
    IEEE INTELLIGENT SYSTEMS, 2021, 36 (03) : 82 - 89
  • [3] Correlation-Decoupled Knowledge Distillation for Multimodal Sentiment Analysis with Incomplete Modalities
    Li, Mingcheng
    Yang, Dingkang
    Zhao, Xiao
    Wang, Shuaibing
    Wang, Yan
    Yang, Kun
    Sun, Mingyang
    Kou, Dongliang
    Qian, Ziyun
    Zhang, Lihua
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 12458 - 12468
  • [4] A Unified Self-Distillation Framework for Multimodal Sentiment Analysis with Uncertain Missing Modalities
    Li, Mingcheng
    Yang, Dingkang
    Lei, Yuxuan
    Wang, Shunli
    Wang, Shuaibing
    Su, Liuzhen
    Yang, Kun
    Wang, Yuzheng
    Sun, Mingyang
    Zhang, Lihua
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 9, 2024, : 10074 - 10082
  • [5] MOCOLNet: A Momentum Contrastive Learning Network for Multimodal Aspect-Level Sentiment Analysis
    Mu, Jie
    Nie, Feiping
    Wang, Wei
    Xu, Jian
    Zhang, Jing
    Liu, Han
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (12) : 8787 - 8800
  • [6] VisdaNet: Visual Distillation and Attention Network for Multimodal Sentiment Classification
    Hou, Shangwu
    Tuerhong, Gulanbaier
    Wushouer, Mairidan
    SENSORS, 2023, 23 (02)
  • [7] Sentiment Interaction Distillation Network for Image Sentiment Analysis
    Wu, Lifang
    Deng, Sinuo
    Zhang, Heng
    Shi, Ge
    APPLIED SCIENCES-BASEL, 2022, 12 (07):
  • [8] A survey of multimodal sentiment analysis
    Soleymani, Mohammad
    Garcia, David
    Jou, Brendan
    Schuller, Bjoern
    Chang, Shih-Fu
    Pantic, Maja
    IMAGE AND VISION COMPUTING, 2017, 65 : 3 - 14
  • [9] A Survey on Multimodal Sentiment Analysis
    Zhang Y.
    Rong L.
    Song D.
    Zhang P.
    Moshi Shibie yu Rengong Zhineng/Pattern Recognition and Artificial Intelligence, 2020, 33 (05): : 426 - 438
  • [10] Benchmarking Multimodal Sentiment Analysis
    Cambria, Erik
    Hazarika, Devamanyu
    Poria, Soujanya
    Hussain, Amir
    Subramanyam, R. B. V.
    COMPUTATIONAL LINGUISTICS AND INTELLIGENT TEXT PROCESSING, CICLING 2017, PT II, 2018, 10762 : 166 - 179