BERT-based chinese text classification for emergency management with a novel loss function

被引:22
|
作者
Wang, Zhongju [1 ,2 ]
Wang, Long [1 ,2 ,3 ]
Huang, Chao [1 ,2 ]
Sun, Shutong [4 ]
Luo, Xiong [1 ,2 ]
机构
[1] Univ Sci & Technol Beijing, Sch Comp & Commun Engn, Beijing 100083, Peoples R China
[2] Beijing Key Lab Knowledge Engn Mat Sci, Beijing 100083, Peoples R China
[3] Univ Sci & Technol Beijing, Shunde Grad Sch, Foshan, Peoples R China
[4] Univ Sci & Technol Beijing, Sch Automat & Elect Engn, Beijing 100083, Peoples R China
关键词
Natural language processing; Deep learning; Text classification; Emergency management; SMOTE; DRIVEN;
D O I
10.1007/s10489-022-03946-x
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper proposes an automatic Chinese text categorization method for solving the emergency event report classification problem. Since the bidirectional encoder representations from transformers (BERT) has achieved great success in the natural language processing domain, it is employed to derive emergency text features in this study. To overcome the data imbalance problem in the distribution of emergency event categories, a novel loss function is proposed to improve the performance of the BERT-based model. Meanwhile, in order to avoid the negative impacts of the extreme learning rate, the Adabound optimization algorithm that achieves a gradual smooth transition from Adam optimizer to stochastic gradient descent optimizer is employed to learn the parameters of the model. The feasibility and competitiveness of the proposed method are validated on both imbalanced and balanced datasets. Furthermore, the generic BERT, BERT ensemble LSTM-BERT (BERT-LB), Attention-based BiLSTM fused CNN with gating mechanism (ABLG-CNN), TextRCNN, Att-BLSTM, and DPCNN are used as benchmarks on these two datasets. Meanwhile, sampling methods, including random sampling, ADASYN, synthetic minority over-sampling techniques (SMOTE), and Borderline-SMOTE, are employed to verify the performance of the proposed loss function on the imbalance dataset. Compared with benchmarking methods, the proposed method has achieved the best performance in terms of accuracy, weighted average precision, weighted average recall, and weighted average F1 values. Therefore, it is promising to employ the proposed method for real applications in smart emergency management systems.
引用
收藏
页码:10417 / 10428
页数:12
相关论文
共 50 条
  • [1] BERT-based chinese text classification for emergency management with a novel loss function
    Zhongju Wang
    Long Wang
    Chao Huang
    Shutong Sun
    Xiong Luo
    Applied Intelligence, 2023, 53 : 10417 - 10428
  • [2] BAE: BERT-based Adversarial Examples for Text Classification
    Garg, Siddhant
    Ramakrishnan, Goutham
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 6174 - 6181
  • [3] Arabic Synonym BERT-based Adversarial Examples for Text Classification
    Alshahrani, Norah
    Alshahrani, Saied
    Wali, Esma
    Matthews, Jeanna
    PROCEEDINGS OF THE 18TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: STUDENT RESEARCH WORKSHOP, 2024, : 137 - 147
  • [4] Improving BERT-Based Text Classification With Auxiliary Sentence and Domain Knowledge
    Yu, Shanshan
    Su, Jindian
    Luo, Da
    IEEE ACCESS, 2019, 7 : 176600 - 176612
  • [5] Improving Bert-Based Model for Medical Text Classification with an Optimization Algorithm
    Gasmi, Karim
    ADVANCES IN COMPUTATIONAL COLLECTIVE INTELLIGENCE, ICCCI 2022, 2022, 1653 : 101 - 111
  • [6] Short-Text Classification Detector: A Bert-Based Mental Approach
    Hu, Yongjun
    Ding, Jia
    Dou, Zixin
    Chang, Huiyou
    COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2022, 2022
  • [7] Three-Branch BERT-Based Text Classification Network for Gastroscopy Diagnosis Text
    Wang Z.
    Zheng X.
    Zhang J.
    Zhang M.
    International Journal of Crowd Science, 2024, 8 (01) : 56 - 63
  • [8] A Fine-Tuned BERT-Based Transfer Learning Approach for Text Classification
    Qasim, Rukhma
    Bangyal, Waqas Haider
    Alqarni, Mohammed A.
    Almazroi, Abdulwahab Ali
    JOURNAL OF HEALTHCARE ENGINEERING, 2022, 2022
  • [9] BBAEG: Towards BERT-based Biomedical Adversarial Example Generation for Text Classification
    Mondal, Ishani
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 5378 - 5384
  • [10] A Fine-Tuned BERT-Based Transfer Learning Approach for Text Classification
    Qasim, Rukhma
    Bangyal, Waqas Haider
    Alqarni, Mohammed A. A.
    Almazroi, Abdulwahab Ali
    JOURNAL OF HEALTHCARE ENGINEERING, 2022, 2022