FinSoSent: Advancing Financial Market Sentiment Analysis through Pretrained Large Language Models

被引:1
|
作者
Delgadillo, Josiel [1 ]
Kinyua, Johnson [2 ]
Mutigwe, Charles [3 ]
机构
[1] Univ Penn, Sch Engn & Appl Sci, Philadelphia, PA 19104 USA
[2] Penn State Univ, Coll Informat Sci & Technol, Philadelphia, PA 19104 USA
[3] Western New England Univ, Coll Business, Springfield, MA 01119 USA
关键词
BERT; financial markets; Twitter/X; StockTwits; sentiment analysis; LLM; social media;
D O I
10.3390/bdcc8080087
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Predicting the directions of financial markets has been performed using a variety of approaches, and the large volume of unstructured data generated by traders and other stakeholders on social media microblog platforms provides unique opportunities for analyzing financial markets using additional perspectives. Pretrained large language models (LLMs) have demonstrated very good performance on a variety of sentiment analysis tasks in different domains. However, it is known that sentiment analysis is a very domain-dependent NLP task that requires knowledge of the domain ontology, and this is particularly the case with the financial domain, which uses its own unique vocabulary. Recent developments in NLP and deep learning including LLMs have made it possible to generate actionable financial sentiments using multiple sources including financial news, company fundamentals, technical indicators, as well social media microblogs posted on platforms such as StockTwits and X (formerly Twitter). We developed a financial social media sentiment analyzer (FinSoSent), which is a domain-specific large language model for the financial domain that was pretrained on financial news articles and fine-tuned and tested using several financial social media corpora. We conducted a large number of experiments using different learning rates, epochs, and batch sizes to yield the best performing model. Our model outperforms current state-of-the-art FSA models based on over 860 experiments, demonstrating the efficacy and effectiveness of FinSoSent. We also conducted experiments using ensemble models comprising FinSoSent and the other current state-of-the-art FSA models used in this research, and a slight performance improvement was obtained based on majority voting. Based on the results obtained across all models in these experiments, the significance of this study is that it highlights the fact that, despite the recent advances of LLMs, sentiment analysis even in domain-specific contexts remains a difficult research problem.
引用
收藏
页数:21
相关论文
共 50 条
  • [21] Application of Pretrained Large Language Models in Embodied Artificial Intelligence
    Kovalev, A. K.
    Panov, A. I.
    DOKLADY MATHEMATICS, 2022, 106 (SUPPL 1) : S85 - S90
  • [22] Advancing Cyber Incident Timeline Analysis Through Retrieval-Augmented Generation and Large Language Models
    Loumachi, Fatma Yasmine
    Ghanem, Mohamed Chahine
    Ferrag, Mohamed Amine
    COMPUTERS, 2025, 14 (02)
  • [23] Sentiment analysis of online responses in the performing arts with large language models
    Seong, Baekryun
    Song, Kyungwoo
    HELIYON, 2023, 9 (12)
  • [24] Revisiting Sentiment Analysis for Software Engineering in the Era of Large Language Models
    Zhang, Ting
    Irsan, Ivana clairine
    Thung, Ferdian
    Lo, David
    ACM TRANSACTIONS ON SOFTWARE ENGINEERING AND METHODOLOGY, 2025, 34 (03)
  • [25] Syntax-Enhanced Pretrained Language Models for Aspect-Level Sentiment Classification
    Yuan, Li
    Wang, Jin
    Lee, Lung-Hao
    Zhang, Xuejie
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2024,
  • [26] Injecting User Identity Into Pretrained Language Models for Document-Level Sentiment Classification
    Cao, Xinlei
    Yu, Jinyang
    Zhuang, Yan
    IEEE ACCESS, 2022, 10 : 30157 - 30167
  • [27] Trend Analysis Through Large Language Models
    Alzapiedi, Lucas
    Bihl, Trevor
    IEEE NATIONAL AEROSPACE AND ELECTRONICS CONFERENCE, NAECON 2024, 2024, : 370 - 374
  • [28] Robust Transfer Learning with Pretrained Language Models through Adapters
    Han, Wenjuan
    Pang, Bo
    Wu, Yingnian
    ACL-IJCNLP 2021: THE 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 2, 2021, : 854 - 861
  • [29] Leveraging hierarchical language models for aspect-based sentiment analysis on financial data
    Lengkeek, Matteo
    Knaap, Finn van der
    Frasincar, Flavius
    INFORMATION PROCESSING & MANAGEMENT, 2023, 60 (05)
  • [30] Discovering sentiment insights: streamlining tourism review analysis with Large Language Models
    Guidotti, Dario
    Pandolfo, Laura
    Pulina, Luca
    INFORMATION TECHNOLOGY & TOURISM, 2025, : 227 - 261