Domain Adaptation of Transformer-Based Models Using Unlabeled Data for Relevance and Polarity Classification of German Customer Feedback

被引:0
|
作者
Idrissi-Yaghir A. [1 ,3 ]
Schäfer H. [1 ,2 ]
Bauer N. [1 ]
Friedrich C.M. [1 ,3 ]
机构
[1] Department of Computer Science, University of Applied Sciences and Arts Dortmund (FHDO), Emil-Figge Str. 42, Dortmund
[2] Institute for Transfusion Medicine, University Hospital Essen, Hufelandstraße 55, Essen
[3] Institute for Medical Informatics, Biometry and Epidemiology (IMIBE), University Hospital Essen, Hufelandstraße 55, Essen
关键词
Domain adaptation; Sentiment analysis; Text classification; Transformer-based models;
D O I
10.1007/s42979-022-01563-6
中图分类号
学科分类号
摘要
Understanding customer feedback is becoming a necessity for companies to identify problems and improve their products and services. Text classification and sentiment analysis can play a major role in analyzing this data by using a variety of machine and deep learning approaches. In this work, different transformer-based models are utilized to explore how efficient these models are when working with a German customer feedback dataset. In addition, these pre-trained models are further analyzed to determine if adapting them to a specific domain using unlabeled data can yield better results than off-the-shelf pre-trained models. To evaluate the models, two downstream tasks from the GermEval 2017 are considered. The experimental results show that transformer-based models can reach significant improvements compared to a fastText baseline and outperform the published scores and previous models. For the subtask Relevance Classification, the best models achieve a micro-averaged F1-Score of 96.1 % on the first test set and 95.9 % on the second one, and a score of 85.1 % and 85.3 % for the subtask Polarity Classification. © 2023, The Author(s).
引用
收藏
相关论文
共 22 条
  • [1] TRANSFORMER-BASED DOMAIN ADAPTATION FOR EVENT DATA CLASSIFICATION
    Zhao, Junwei
    Zhang, Shiliang
    Huang, Tiejun
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 4673 - 4677
  • [2] Localizing in-domain adaptation of transformer-based biomedical language models
    Buonocore, Tommaso Mario
    Crema, Claudio
    Redolfi, Alberto
    Bellazzi, Riccardo
    Parimbelli, Enea
    JOURNAL OF BIOMEDICAL INFORMATICS, 2023, 144
  • [3] Transformer-Based Multi-Source Domain Adaptation Without Source Data
    Li, Gang
    Wu, Chao
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [4] MI-CAT: A transformer-based domain adaptation network for motor imagery classification
    Zhang, Dongxue
    Li, Huiying
    Xie, Jingmeng
    NEURAL NETWORKS, 2023, 165 : 451 - 462
  • [5] Empirical Study of Tweets Topic Classification Using Transformer-Based Language Models
    Mandal, Ranju
    Chen, Jinyan
    Becken, Susanne
    Stantic, Bela
    INTELLIGENT INFORMATION AND DATABASE SYSTEMS, ACIIDS 2021, 2021, 12672 : 340 - 350
  • [6] Enhancing Address Data Integrity using Transformer-Based Language Models
    Kurklu, Omer Faruk
    Akagiunduz, Erdem
    32ND IEEE SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE, SIU 2024, 2024,
  • [7] Metric Learning Using Labeled and Unlabeled Data for Semi-Supervised/Domain Adaptation Classification
    Benisty, Hadas
    Crammer, Koby
    2014 IEEE 28TH CONVENTION OF ELECTRICAL & ELECTRONICS ENGINEERS IN ISRAEL (IEEEI), 2014,
  • [8] Classification of Highly Divergent Viruses from DNA/RNA Sequence Using Transformer-Based Models
    Sadad, Tariq
    Aurangzeb, Raja Atif
    Imran
    Safran, Mejdl
    Alfarhood, Sultan
    Kim, Jungsuk
    BIOMEDICINES, 2023, 11 (05)
  • [9] Failure Mode Classification for Rolling Element Bearings Using Time-Domain Transformer-Based Encoder
    Vu, Minh Tri
    Hiraga, Motoaki
    Miura, Nanako
    Masuda, Arata
    SENSORS, 2024, 24 (12)
  • [10] Automated genre-based multi-domain sentiment lexicon adaptation using unlabeled data
    Sanagar, Swati
    Gupta, Deepa
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2020, 38 (05) : 6223 - 6234