Calibration of Transformer-Based Models for Identifying Stress and Depression in Social Media

被引:23
|
作者
Ilias, Loukas [1 ]
Mouzakitis, Spiros [1 ]
Askounis, Dimitris [1 ]
机构
[1] Natl Tech Univ Athens, Decis Support Syst Lab, Schoolof Elect & Comp Engn, Athens 15780, Greece
关键词
~Calibration; depression; emotion; mental health; stress; transformers; EMOTION;
D O I
10.1109/TCSS.2023.3283009
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
In today's fast-paced world, the rates of stress and depression present a surge. People use social media for expressing their thoughts and feelings through posts. Therefore, social media provide assistance for the early detection of mental health conditions. Existing methods mainly introduce feature extraction approaches and train shallow machine learning (ML) classifiers. For addressing the need of creating a large feature set and obtaining better performance, other research studies use deep neural networks or language models based on transformers. Despite the fact that transformer-based models achieve noticeable improvements, they cannot often capture rich factual knowledge. Although there have been proposed a number of studies aiming to enhance the pretrained transformer-based models with extra information or additional modalities, no prior work has exploited these modifications for detecting stress and depression through social media. In addition, although the reliability of a machine learning (ML) model's confidence in its predictions is critical for high-risk applications, there is no prior work taken into consideration the model calibration. To resolve the above issues, we present the first study in the task of depression and stress detection in social media, which injects extra-linguistic information in transformer-based models, namely, bidirectional encoder representations from transformers (BERT) and MentalBERT. Specifically, the proposed approach employs a multimodal adaptation gate for creating the combined embeddings, which are given as input to a BERT (or MentalBERT) model. For taking into account the model calibration, we apply label smoothing. We test our proposed approaches in three publicly available datasets and demonstrate that the integration of linguistic features into transformer-based models presents a surge in performance. Also, the usage of label smoothing contributes to both the improvement of the model's performance and the calibration of the model. We finally perform a linguistic analysis of the posts and show differences in language between stressful and nonstressful texts, as well as depressive and nondepressive posts.
引用
收藏
页码:1979 / 1990
页数:12
相关论文
共 50 条
  • [41] Characterization of MPC-based Private Inference for Transformer-based Models
    Wang, Yongqin
    Edward, G.
    Xiong, Wenjie
    Lefaudeux, Benjamin
    Knott, Brian
    Annavaram, Murali
    Lee, Hsien-Hsin S.
    2022 IEEE INTERNATIONAL SYMPOSIUM ON PERFORMANCE ANALYSIS OF SYSTEMS AND SOFTWARE (ISPASS 2022), 2022, : 187 - 197
  • [42] AMMU: A survey of transformer-based biomedical pretrained language models
    Kalyan, Katikapalli Subramanyam
    Rajasekharan, Ajit
    Sangeetha, Sivanesan
    JOURNAL OF BIOMEDICAL INFORMATICS, 2022, 126
  • [43] Transformer-based language models for mental health issues: A survey
    Greco, Candida M.
    Simeri, Andrea
    Tagarelli, Andrea
    Zumpano, Ester
    PATTERN RECOGNITION LETTERS, 2023, 167 : 204 - 211
  • [44] Transformer-Based Models for the Automatic Indexing of Scientific Documents in French
    Angel Gonzalez, Jose
    Buscaldi, Davide
    Sanchis, Emilio
    Hurtado, Lluis-F
    NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS (NLDB 2022), 2022, 13286 : 60 - 72
  • [45] Detecting Sarcasm in Conversation Context Using Transformer-Based Models
    Avvaru, Adithya
    Vobilisetty, Sanath
    Mamidi, Radhika
    FIGURATIVE LANGUAGE PROCESSING, 2020, : 98 - 103
  • [46] ORCA: A Distributed Serving System for Transformer-Based Generative Models
    Yu, Gyeong-In
    Jeong, Joo Seong
    Kim, Geon-Woo
    Kim, Soojeong
    Chun, Byung-Gon
    PROCEEDINGS OF THE 16TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, OSDI 2022, 2022, : 521 - 538
  • [47] Pre-trained transformer-based language models for Sundanese
    Wilson Wongso
    Henry Lucky
    Derwin Suhartono
    Journal of Big Data, 9
  • [48] Classification and recognition of gesture EEG signals with Transformer-Based models
    Qu, Yan
    Li, Congsheng
    Jiang, Haoyu
    2024 3RD INTERNATIONAL CONFERENCE ON ROBOTICS, ARTIFICIAL INTELLIGENCE AND INTELLIGENT CONTROL, RAIIC 2024, 2024, : 415 - 418
  • [49] Pulsar candidate identification using advanced transformer-based models
    Cao, Jie
    Xu, Tingting
    Deng, Linhua
    Zhou, Xueliang
    Li, Shangxi
    Liu, Yuxia
    Zhou, Weihong
    CHINESE JOURNAL OF PHYSICS, 2024, 90 : 121 - 133
  • [50] Pre-trained transformer-based language models for Sundanese
    Wongso, Wilson
    Lucky, Henry
    Suhartono, Derwin
    JOURNAL OF BIG DATA, 2022, 9 (01)