EmoComicNet: A multi-task model for comic emotion recognition

被引:4
|
作者
Dutta, Arpita [1 ,2 ]
Biswas, Samit [1 ]
Das, Amit Kumar [1 ]
机构
[1] Indian Inst Engn Science&Technol, Dept Comp Science&Technol, Howrah 711103, West Bengal, India
[2] Techno Main, Artificial Intelligence & Machine Learning, Dept Comp Sci & Engn, Kolkata 700091, West Bengal, India
关键词
Comic analysis; Multi-modal emotion recognition; Document image processing; Deep learning; Multi-task learning;
D O I
10.1016/j.patcog.2024.110261
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The emotion and sentiment associated with comic scenes can provide potential information for inferring the context of comic stories, which is an essential pre -requisite for developing comics' automatic content understanding tools. Here, we address this open area of comic research by exploiting the multi -modal nature of comics. The general assumptions for multi -modal sentiment analysis methods are that both image and text modalities are always present at the test phase. However, this assumption is not always satisfied for comics since comic characters' facial expressions, gestures, etc., are not always clearly visible. Also, the dialogues between comic characters are often challenging to comprehend the underlying context. To deal with these constraints of comic emotion analysis, we propose a multi -task -based framework, namely EmoComicNet, to fuse multi -modal information (i.e., both image and text) if it is available. However, the proposed EmoComicNet is designed to perform even when any modality is weak or completely missing. The proposed method potentially improves the overall performance. Besides, EmoComicNet can also deal with the problem of weak or absent modality during the training phase.
引用
收藏
页数:11
相关论文
共 50 条
  • [21] Inconsistency-Based Multi-Task Cooperative Learning for Emotion Recognition
    Xu, Yifan
    Cui, Yuqi
    Jiang, Xue
    Yin, Yingjie
    Ding, Jingting
    Li, Liang
    Wu, Dongrui
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2022, 13 (04) : 2017 - 2027
  • [22] An Emotion Type Informed Multi-Task Model for Emotion Cause Pair Extraction
    Chen, Zhe
    Zhang, Ming
    Palade, Vasile
    Wang, Liya
    Zhang, Junchi
    Feng, Ying
    IEEE ACCESS, 2024, 12 : 15662 - 15674
  • [23] Multi-EmoNet: A Novel Multi-Task Neural Network for Driver Emotion Recognition
    Cui, Yaodong
    Ma, Yintao
    Li, Wenbo
    Bian, Ning
    Li, Guofa
    Cao, Dongpu
    IFAC PAPERSONLINE, 2020, 53 (05): : 650 - 655
  • [24] Logistic Regression Based Multi-task, Multi-kernel Learning for Emotion Recognition
    He, Xinrun
    Huang, Jian
    Zeng, Zhigang
    2021 6TH IEEE INTERNATIONAL CONFERENCE ON ADVANCED ROBOTICS AND MECHATRONICS (ICARM 2021), 2021, : 572 - 577
  • [25] MMATERIC: Multi-Task Learning and Multi-Fusion for AudioText Emotion Recognition in Conversation
    Liang, Xingwei
    Zou, You
    Zhuang, Xinnan
    Yang, Jie
    Niu, Taiyu
    Xu, Ruifeng
    ELECTRONICS, 2023, 12 (07)
  • [26] Multi-task Recurrent Model for Speech and Speaker Recognition
    Tang, Zhiyuan
    Li, Lantian
    Wang, Dong
    2016 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA), 2016,
  • [27] MTLSER: Multi-task learning enhanced speech emotion recognition with pre-trained acoustic model
    Chen, Zengzhao
    Liu, Chuan
    Wang, Zhifeng
    Zhao, Chuanxu
    Lin, Mengting
    Zheng, Qiuyu
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 273
  • [28] Coarse-to-Fine Speech Emotion Recognition Based on Multi-Task Learning
    Zhao Huijuan
    Ye Ning
    Wang Ruchuan
    Journal of Signal Processing Systems, 2021, 93 : 299 - 308
  • [29] Hybrid Multi-Task Learning for End-To-End Multimodal Emotion Recognition
    Chen, Junjie
    Li, Yongwei
    Zhao, Ziping
    Liu, Xuefei
    Wen, Zhengqi
    Tao, Jianhua
    2023 ASIA PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE, APSIPA ASC, 2023, : 1966 - 1971
  • [30] Speech Emotion Recognition using Decomposed Speech via Multi-task Learning
    Hsu, Jia-Hao
    Wu, Chung-Hsien
    Wei, Yu-Hung
    INTERSPEECH 2023, 2023, : 4553 - 4557