A Method for Generating Explanations of Offensive Memes Based on Multimodal Large Language Models

被引:0
|
作者
Lin M. [1 ,2 ]
Dai C. [1 ,2 ]
Guo T. [1 ]
机构
[1] Institute of Information Engineering, Chinese Academy of Sciences, Beijing
[2] School of Cyber Security, University of Chinese Academy of Sciences, Beijing
关键词
data augment; explanation generation; instruction fine-tuning; multi-modal LLM; offensive meme;
D O I
10.7544/issn1000-1239.202330960
中图分类号
学科分类号
摘要
With the advancement of 5G technology, offensive speech has increasingly proliferated across social networks in the form of multimodal memes. Consequently, the detection and interpretive generation of offensive memes play a crucial role in enhancing content moderation effectiveness and maintaining a harmonious and healthy public discourse environment. Existing studies on the interpretive generation of offensive memes focus solely on the targets and content of offense, neglecting the societal background knowledge and metaphorical expressions embedded in memes. This oversight limits the ability to comprehensively and accurately interpret the meaning of offensive memes, thus constraining the applicability of interpretations. To address this challenge, we propose a method based on multimodal large language model for generating interpretations of offensive memes. By augmenting elements such as offense targets, the content of the offense, and metaphor recognition into the instruction tuning data, we can effectively improve the multimodal large model’s proficiency in interpretively generating offensive memes through instruction tuning. The experimental outcomes validate three key strengths of our method: first, it achieves a notable 19% enhancement in the BERTScore evaluation metric over baseline models; second, it incorporates comprehensive background knowledge pertinent to offensive metaphorical expressions within its interpretations; third, it exhibits strong generalization capabilities when handling previously unseen meme data. © 2024 Science Press. All rights reserved.
引用
收藏
页码:1206 / 1217
页数:11
相关论文
共 43 条
  • [1] Songlin Hu, Jun Zhao, Jie Tang, Et al., Preface to the special issue on fake information detection[J], Journal of Computer Research and Development, 58, 7, pp. 1351-1352, (2021)
  • [2] Kiela D, Firooz H, Mohan A, Et al., The hateful memes challenge: Detecting hate speech in multimodal memes[J], Advances in Neural Information Processing Systems, 33, pp. 2611-2624, (2020)
  • [3] Zhang Linhao, Jin Li, Sun Xian, Et al., TOT: Topology-aware optimal transport for multimodal hate detection, Proc of the AAAI Conf on Artificial Intelligence, 37, 4, pp. 4884-4892, (2023)
  • [4] Cao Rui, Hee MS, Kuek A, Et al., Pro-Cap: Leveraging a frozen vision-language model for hateful meme detection[C], Proc of the 31st ACM Int Conf on Multimedia, pp. 5244-5252, (2023)
  • [5] Hee M S, Chong W H, Lee R K W., Decoding the underlying meaning of multimodal hateful memes, Proc of the 32nd Int Joint Conf on Artificial Intelligence, pp. 5995-6003, (2023)
  • [6] Sharma S, Agarwal S, Suresh T, Et al., What do you MEME? Generating explanations for visual semantic role labelling in memes[C], Proc of the AAAI Conf on Artificial Intelligence, 37, 8, pp. 9763-9771, (2023)
  • [7] Scott K., Memes as multimodal metaphors: A relevance theory analysis[J], Pragmatics & Cognition, 28, 2, (2021)
  • [8] Pramanick S, Sharma S, Dimitrov D, Et al., MOMENTA: A multimodal framework for detecting harmful memes and their targets[C], Findings of the Association for Computational Linguistics: EMNLP 2021, pp. 4439-4455, (2021)
  • [9] Zhu Ron, Enhance multimodal Transformer with external label and in-domain pretrain: Hateful meme challenge winning solution, (2020)
  • [10] Yang Chuanpeng, Zhu Fuqing, Liu Guihua, Et al., Multimodal hate speech detection via cross-domain knowledge transfer, Proc of the 30th ACM Int Conf on Multimedia, pp. 4505-4514, (2022)