A Study on Performance Enhancement by Integrating Neural Topic Attention with Transformer-Based Language Model

被引:1
|
作者
Um, Taehum [1 ]
Kim, Namhyoung [1 ]
机构
[1] Gachon Univ, Dept Appl Stat, 1342 Seongnam Daero, Seongnam 13120, South Korea
来源
APPLIED SCIENCES-BASEL | 2024年 / 14卷 / 17期
基金
新加坡国家研究基金会;
关键词
natural language processing; neural topic model; ELECTRA; ALBERT; multi-classification;
D O I
10.3390/app14177898
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
As an extension of the transformer architecture, the BERT model has introduced a new paradigm for natural language processing, achieving impressive results in various downstream tasks. However, high-performance BERT-based models-such as ELECTRA, ALBERT, and RoBERTa-suffer from limitations such as poor continuous learning capability and insufficient understanding of domain-specific documents. To address these issues, we propose the use of an attention mechanism to combine BERT-based models with neural topic models. Unlike traditional stochastic topic modeling, neural topic modeling employs artificial neural networks to learn topic representations. Furthermore, neural topic models can be integrated with other neural models and trained to identify latent variables in documents, thereby enabling BERT-based models to sufficiently comprehend the contexts of specific fields. We conducted experiments on three datasets-Movie Review Dataset (MRD), 20Newsgroups, and YELP-to evaluate our model's performance. Compared to the vanilla model, the proposed model achieved an accuracy improvement of 1-2% for the ALBERT model in multiclassification tasks across all three datasets, while the ELECTRA model showed an accuracy improvement of less than 1%.
引用
收藏
页数:14
相关论文
共 50 条
  • [21] TRANSFORMER-BASED STREAMING ASR WITH CUMULATIVE ATTENTION
    Li, Mohan
    Zhang, Shucong
    Zorila, Catalin
    Doddipatla, Rama
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 8272 - 8276
  • [22] The Case for Translation-Invariant Self-Attention in Transformer-Based Language Models
    Wennberg, Ulme
    Henter, Gustav Eje
    ACL-IJCNLP 2021: THE 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 2, 2021, : 130 - 140
  • [23] Korean Sign Language Recognition Using Transformer-Based Deep Neural Network
    Shin, Jungpil
    Musa Miah, Abu Saleh
    Hasan, Md. Al Mehedi
    Hirooka, Koki
    Suzuki, Kota
    Lee, Hyoun-Sup
    Jang, Si-Woong
    APPLIED SCIENCES-BASEL, 2023, 13 (05):
  • [24] A Framework for Accelerating Transformer-Based Language Model on ReRAM-Based Architecture
    Kang, Myeonggu
    Shin, Hyein
    Kim, Lee-Sup
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2022, 41 (09) : 3026 - 3039
  • [25] A Transformer-Based Math Language Model for Handwritten Math Expression Recognition
    Huy Quang Ung
    Cuong Tuan Nguyen
    Hung Tuan Nguyen
    Thanh-Nghia Truong
    Nakagawa, Masaki
    DOCUMENT ANALYSIS AND RECOGNITION, ICDAR 2021, PT II, 2021, 12917 : 403 - 415
  • [26] Hybrid transformer-based model for mammogram classification by integrating prior and current images
    Jeny, Afsana Ahsan
    Hamzehei, Sahand
    Jin, Annie
    Baker, Stephen Andrew
    Van Rathe, Tucker
    Bai, Jun
    Yang, Clifford
    Nabavi, Sheida
    MEDICAL PHYSICS, 2025,
  • [27] An Improved Transformer-Based Neural Machine Translation Strategy: Interacting-Head Attention
    Li, Dongxing
    Luo, Zuying
    COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2022, 2022
  • [28] A novel transformer-based neural network model for tool wear estimation
    Liu, Hui
    Liu, Zhenyu
    Jia, Weiqiang
    Lin, Xianke
    Zhang, Shuo
    MEASUREMENT SCIENCE AND TECHNOLOGY, 2020, 31 (06)
  • [29] Transformer-Based Topic Modeling for Urdu Translations of the Holy Quran
    Zafar, Amna
    Wasim, Muhammad
    Zulfiqar, Shaista
    Waheed, Talha
    Siddique, Abubakar
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2024, 23 (10)
  • [30] Transformer-Based Music Language Modelling and Transcription
    Zonios, Christos
    Pavlopoulos, John
    Likas, Aristidis
    PROCEEDINGS OF THE 12TH HELLENIC CONFERENCE ON ARTIFICIAL INTELLIGENCE, SETN 2022, 2022,