Performance Evaluation of Pre-trained Models in Sarcasm Detection Task

被引:2
|
作者
Wang, Haiyang [1 ]
Song, Xin [1 ]
Zhou, Bin [1 ]
Wang, Ye [1 ]
Gao, Liqun [1 ]
Jia, Yan [1 ]
机构
[1] Natl Univ Def Technol, Changsha, Peoples R China
关键词
Sarcasm detection; Pre-trained models; Natural language processing;
D O I
10.1007/978-3-030-91560-5_5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Sarcasm is a widespread phenomenon in social media such as Twitter or Instagram. As a critical task of Natural Language Processing (NLP), sarcasm detection plays an important role in many domains of semantic analysis, such as stance detection and sentiment analysis. Recently, pre-trained models (PTMs) on large unlabelled corpora have shown excellent performance in various tasks of NLP. PTMs have learned universal language representations and can help researchers avoid training a model from scratch. The goal of our paper is to evaluate the performance of various PTMs in the sarcasm detection task. We evaluate and analyse the performance of several representative PTMs on four well-known sarcasm detection datasets. The experimental results indicate that RoBERTa outperforms other PTMs and it is also better than the best baseline in three datasets. DistilBERT is the best choice for sarcasm detection task when computing resources are limited. However, XLNet may not be suitable for sarcasm detection task. In addition, we implement detailed grid search for four hyperparameters to investigate their impact on PTMs. The results show that learning rate is the most important hyperparameter. Furthermore, we also conduct error analysis by means of several sarcastic sentences to explore the reasons of detection failures, which provides instructive ideas for future research.
引用
收藏
页码:67 / 75
页数:9
相关论文
共 50 条
  • [1] Performance Evaluation of CNN and Pre-trained Models for Malware Classification
    Omar Habibi
    Mohammed Chemmakha
    Mohamed Lazaar
    Arabian Journal for Science and Engineering, 2023, 48 : 10355 - 10369
  • [2] Performance Evaluation of CNN and Pre-trained Models for Malware Classification
    Habibi, Omar
    Chemmakha, Mohammed
    Lazaar, Mohamed
    ARABIAN JOURNAL FOR SCIENCE AND ENGINEERING, 2023, 48 (08) : 10355 - 10369
  • [3] Evaluation of Pre-Trained CNN Models for Geographic Fake Image Detection
    Fezza, Sid Ahmed
    Ouis, Mohammed Yasser
    Kaddar, Bachir
    Hamidouche, Wassim
    Hadid, Abdenour
    2022 IEEE 24TH INTERNATIONAL WORKSHOP ON MULTIMEDIA SIGNAL PROCESSING (MMSP), 2022,
  • [4] Performance Evaluation of Pre-Trained CNN Models for Visual Saliency Prediction
    Ghariba, Bashir
    Shehata, Mohamed S.
    McGuire, Peter
    2020 IEEE CANADIAN CONFERENCE ON ELECTRICAL AND COMPUTER ENGINEERING (CCECE), 2020,
  • [5] Pre-trained Lightweight Deep Learning Models for Surgical Instrument Detection: Performance Evaluation for Edge Inference
    Ahmed, Md Sabbir
    Giordano, Stefano
    IEEE CONFERENCE ON GLOBAL COMMUNICATIONS, GLOBECOM, 2023, : 3873 - 3878
  • [6] Task Arithmetic in the Tangent Space: Improved Editing of Pre-Trained Models
    Ortiz-Jimenez, Guillermo
    Favero, Alessandro
    Frossard, Pascal
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [7] Evaluation and optimisation of pre-trained CNN models for asphalt pavement crack detection and classification
    Matarneh, Sandra
    Elghaish, Faris
    Rahimian, Farzad Pour
    Abdellatef, Essam
    Abrishami, Sepehr
    AUTOMATION IN CONSTRUCTION, 2024, 160
  • [8] Adapting Pre-trained Language Models to Rumor Detection on Twitter
    Slimi, Hamda
    Bounhas, Ibrahim
    Slimani, Yahya
    JOURNAL OF UNIVERSAL COMPUTER SCIENCE, 2021, 27 (10) : 1128 - 1148
  • [9] Using hybrid pre-trained models for breast cancer detection
    Zarif, Sameh
    Abdulkader, Hatem
    Elaraby, Ibrahim
    Alharbi, Abdullah
    Elkilani, Wail S.
    Plawiak, Pawel
    PLOS ONE, 2024, 19 (01):
  • [10] Multi-task Learning Based Online Dialogic Instruction Detection with Pre-trained Language Models
    Hao, Yang
    Li, Hang
    Ding, Wenbiao
    Wu, Zhongqin
    Tang, Jiliang
    Luckin, Rose
    Liu, Zitao
    ARTIFICIAL INTELLIGENCE IN EDUCATION (AIED 2021), PT II, 2021, 12749 : 183 - 189