Calibration of Transformer-Based Models for Identifying Stress and Depression in Social Media

被引:23
|
作者
Ilias, Loukas [1 ]
Mouzakitis, Spiros [1 ]
Askounis, Dimitris [1 ]
机构
[1] Natl Tech Univ Athens, Decis Support Syst Lab, Schoolof Elect & Comp Engn, Athens 15780, Greece
关键词
~Calibration; depression; emotion; mental health; stress; transformers; EMOTION;
D O I
10.1109/TCSS.2023.3283009
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
In today's fast-paced world, the rates of stress and depression present a surge. People use social media for expressing their thoughts and feelings through posts. Therefore, social media provide assistance for the early detection of mental health conditions. Existing methods mainly introduce feature extraction approaches and train shallow machine learning (ML) classifiers. For addressing the need of creating a large feature set and obtaining better performance, other research studies use deep neural networks or language models based on transformers. Despite the fact that transformer-based models achieve noticeable improvements, they cannot often capture rich factual knowledge. Although there have been proposed a number of studies aiming to enhance the pretrained transformer-based models with extra information or additional modalities, no prior work has exploited these modifications for detecting stress and depression through social media. In addition, although the reliability of a machine learning (ML) model's confidence in its predictions is critical for high-risk applications, there is no prior work taken into consideration the model calibration. To resolve the above issues, we present the first study in the task of depression and stress detection in social media, which injects extra-linguistic information in transformer-based models, namely, bidirectional encoder representations from transformers (BERT) and MentalBERT. Specifically, the proposed approach employs a multimodal adaptation gate for creating the combined embeddings, which are given as input to a BERT (or MentalBERT) model. For taking into account the model calibration, we apply label smoothing. We test our proposed approaches in three publicly available datasets and demonstrate that the integration of linguistic features into transformer-based models presents a surge in performance. Also, the usage of label smoothing contributes to both the improvement of the model's performance and the calibration of the model. We finally perform a linguistic analysis of the posts and show differences in language between stressful and nonstressful texts, as well as depressive and nondepressive posts.
引用
收藏
页码:1979 / 1990
页数:12
相关论文
共 50 条
  • [21] Transformer-based Context-aware Sarcasm Detection in Conversation Threads from Social Media
    Dong, Xiangjue
    Li, Changmao
    Choi, Jinho D.
    FIGURATIVE LANGUAGE PROCESSING, 2020, : 276 - 280
  • [22] Position-context additive transformer-based model for classifying text data on social media
    Abd-Elaziz, M. M.
    El-Rashidy, Nora
    Abou Elfetouh, Ahmed
    El-Bakry, Hazem M.
    SCIENTIFIC REPORTS, 2025, 15 (01):
  • [23] Ouroboros: On Accelerating Training of Transformer-Based Language Models
    Yang, Qian
    Huo, Zhouyuan
    Wang, Wenlin
    Huang, Heng
    Carin, Lawrence
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [24] Transformer-based Models for Arabic Online Handwriting Recognition
    Alwajih, Fakhraddin
    Badr, Eman
    Abdou, Sherif
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2022, 13 (05) : 898 - 905
  • [25] Blockwise compression of transformer-based models without retraining
    Dong, Gaochen
    Chen, W.
    NEURAL NETWORKS, 2024, 171 : 423 - 428
  • [26] Transformer-Based Language Models for Software Vulnerability Detection
    Thapa, Chandra
    Jang, Seung Ick
    Ahmed, Muhammad Ejaz
    Camtepe, Seyit
    Pieprzyk, Josef
    Nepal, Surya
    PROCEEDINGS OF THE 38TH ANNUAL COMPUTER SECURITY APPLICATIONS CONFERENCE, ACSAC 2022, 2022, : 481 - 496
  • [27] Transformer-Based Federated Learning Models for Recommendation Systems
    Reddy, M. Sujaykumar
    Karnati, Hemanth
    Sundari, L. Mohana
    IEEE ACCESS, 2024, 12 : 109596 - 109607
  • [28] A Comparison of Transformer-Based Language Models on NLP Benchmarks
    Greco, Candida Maria
    Tagarelli, Andrea
    Zumpano, Ester
    NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS (NLDB 2022), 2022, 13286 : 490 - 501
  • [29] RadBERT: Adapting Transformer-based Language Models to Radiology
    Yan, An
    McAuley, Julian
    Lu, Xing
    Du, Jiang
    Chang, Eric Y.
    Gentili, Amilcare
    Hsu, Chun-Nan
    RADIOLOGY-ARTIFICIAL INTELLIGENCE, 2022, 4 (04)
  • [30] Semantics of Multiword Expressions in Transformer-Based Models: A Survey
    Miletic, Filip
    Walde, Sabine Schulte Im
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2024, 12 : 593 - 612