TopicBERT: A Topic-Enhanced Neural Language Model Fine-Tuned for Sentiment Classification

被引:23
|
作者
Zhou, Yuxiang [1 ]
Liao, Lejian [1 ]
Gao, Yang [1 ]
Wang, Rui [2 ]
Huang, Heyan [1 ]
机构
[1] Beijing Inst Technol, Fac Comp Sci, Beijing 100081, Peoples R China
[2] Nanjing Univ Posts & Telecommun, Fac Comp Sci, Nanjing 210023, Peoples R China
基金
中国国家自然科学基金;
关键词
Task analysis; Bit error rate; Semantics; Predictive models; Training; Context modeling; Social networking (online); Bidirectional encoder representations from transformers (BERT); pretrained neural language model; sentiment classification; topic-enhanced neural network;
D O I
10.1109/TNNLS.2021.3094987
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Sentiment classification is a form of data analytics where people's feelings and attitudes toward a topic are mined from data. This tantalizing power to ``predict the zeitgeist'' means that sentiment classification has long attracted interest, but with mixed results. However, the recent development of the BERT framework and its pretrained neural language models is seeing new-found success for sentiment classification. BERT models are trained to capture word-level information via mask language modeling and sentence-level contexts via next sentence prediction tasks. Out of the box, they are adequate models for some natural language processing tasks. However, most models are further fine-tuned with domain-specific information to increase accuracy and usefulness. Motivated by the idea that a further fine-tuning step would improve the performance for downstream sentiment classification tasks, we developed TopicBERT--a BERT model fine-tuned to recognize topics at the corpus level in addition to the word and sentence levels. TopicBERT comprises two variants: TopicBERT-ATP (aspect topic prediction), which captures topic information via an auxiliary training task, and TopicBERT-TA, where topic representation is directly injected into a topic augmentation layer for sentiment classification. With TopicBERT-ATP, the topics are predetermined by an LDA mechanism and collapsed Gibbs sampling. With TopicBERT-TA, the topics can change dynamically during the training. Experimental results show that both approaches deliver the state-of-the-art performance in two different domains with SemEval 2014 Task 4. However, in a test of methods, direct augmentation outperforms further training. Comprehensive analyses in the form of ablation, parameter, and complexity studies accompany the results.
引用
收藏
页码:380 / 393
页数:14
相关论文
共 50 条
  • [41] Transformer based Contextual Model for Sentiment Analysis of Customer Reviews: A Fine-tuned BERT A Sequence Learning BERT Model for Sentiment Analysis
    Durairaj, Ashok Kumar
    Chinnalagu, Anandan
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2021, 12 (11) : 474 - 480
  • [42] Accelerating the Classification of NOVA Food Processing Levels Using a Fine-Tuned Language Model: A Multi-Country Study
    Hu, Guanlan
    Flexner, Nadia
    Tiscornia, Maria Victoria
    L'Abbe, Mary R.
    NUTRIENTS, 2023, 15 (19)
  • [43] Comparing Fine-Tuned Transformers and Large Language Models for Sales Call Classification: A Case Study
    Eisenstadt, Roy
    Asi, Abedelkader
    Ronen, Royi
    PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, : 5240 - 5241
  • [44] Sentiment Analysis of Song Dynasty Classical Poetry Using Fine-Tuned Large Language Models: A Study with LLMs
    Ihnaini, Baha
    Sun, Weiyi
    Cai, Yingchao
    Xu, Zhijun
    Sangi, Rashid
    2024 7TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND BIG DATA, ICAIBD 2024, 2024, : 590 - 597
  • [45] Learning document representation via topic-enhanced LSTM model
    Zhang, Wenyue
    Li, Yang
    Wang, Suge
    KNOWLEDGE-BASED SYSTEMS, 2019, 174 : 194 - 204
  • [46] Topic-Enhanced Capsule Network for Multi-Label Emotion Classification
    Fei, Hao
    Ji, Donghong
    Zhang, Yue
    Ren, Yafeng
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2020, 28 : 1839 - 1848
  • [47] Enhancing Solution Diversity in Arithmetic Problems using Fine-Tuned AI Language Model
    Lee, Chang-Yu
    Lai, I-Wei
    2024 11TH INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS-TAIWAN, ICCE-TAIWAN 2024, 2024, : 515 - 516
  • [48] Improving RAG Quality for Large Language Models with Topic-Enhanced Reranking
    Ampazis, Nicholas
    ARTIFICIAL INTELLIGENCE APPLICATIONS AND INNOVATIONS, PT II, AIAI 2024, 2024, 712 : 74 - 87
  • [49] Exploring Generalizability of a fine-tuned Large Language Model for Impression Generation in PET Reports
    Yousefirizi, F.
    Wang, L.
    Gowdy, C.
    Shariftabrizi, A.
    Harsini, S.
    Ahamed, S.
    Sabouri, M.
    Mollaheydar, E.
    Rahmim, A.
    EUROPEAN JOURNAL OF NUCLEAR MEDICINE AND MOLECULAR IMAGING, 2024, 51 : S785 - S785
  • [50] Effectiveness of Fine-tuned BERT Model in Classification of Helpful and Unhelpful Online Customer Reviews
    Muhammad Bilal
    Abdulwahab Ali Almazroi
    Electronic Commerce Research, 2023, 23 : 2737 - 2757