Chinese cyber-violent Speech Detection and Analysis Based on Pre-trained Model

被引:0
|
作者
Zhou, Sunrui [1 ]
机构
[1] Shanghai Univ, Shanghai, Peoples R China
关键词
Chinese cyber-violent speech; BERT; Hanyu Pinyin; Emotion;
D O I
10.1145/3670105.3670179
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Cyber-violent speech is prevalent on Chinese social platforms today, and traditional manual moderation by platform administrators is no longer effective in detecting and analyzing it. Therefore, the use of artificial intelligence technologies like natural language processing for automated detection on the Internet is an essential requirement to promptly prevent the spread of cyber-violent speech. Due to the covert and diverse nature of cyber-violent speech, existing models have shown unsatisfactory performance in detecting implicitly expressed violent speech. This paper proposes a violence speech detection method based on BERT and Hanyu Pinyin and emotion assistance, and its effectiveness and advancement are validated on multiple datasets. Subsequently, the experimental results are analyzed to summarize the characteristics of Chinese violent speech, facilitating further development in violence speech detection efforts in the future.
引用
收藏
页码:443 / 447
页数:5
相关论文
共 50 条
  • [41] Hyperbolic Pre-Trained Language Model
    Chen, Weize
    Han, Xu
    Lin, Yankai
    He, Kaichen
    Xie, Ruobing
    Zhou, Jie
    Liu, Zhiyuan
    Sun, Maosong
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 3101 - 3112
  • [42] GAIT analysis based on GENDER detection using pre-trained models and tune parameters
    Vora C.
    Katkar V.
    Lunagaria M.
    Discover Artificial Intelligence, 2024, 4 (01):
  • [43] Embedding Articulatory Constraints for Low-resource Speech Recognition Based on Large Pre-trained Model
    Lee, Jaeyoung
    Mimura, Masato
    Kawahara, Tatsuya
    INTERSPEECH 2023, 2023, : 1394 - 1398
  • [44] COVID-HateBERT: a Pre-trained Language Model for COVID-19 related Hate Speech Detection
    Li, Mingqi
    Liao, Song
    Okpala, Ebuka
    Tong, Max
    Costello, Matthew
    Cheng, Long
    Hu, Hongxin
    Luo, Feng
    20TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2021), 2021, : 233 - 238
  • [45] AraXLNet: pre-trained language model for sentiment analysis of Arabic
    Alduailej, Alhanouf
    Alothaim, Abdulrahman
    JOURNAL OF BIG DATA, 2022, 9 (01)
  • [46] Combining multiple pre-trained models for hate speech detection in Bengali, Marathi, and Hindi
    Nandi, Arpan
    Sarkar, Kamal
    Mallick, Arjun
    De, Arkadeep
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (32) : 77733 - 77757
  • [47] An Integrated Chinese Malicious Webpages Detection Method Based on Pre-trained Language Models and Feature Fusion
    Jiang, Yanting
    Wu, Di
    Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 2022, 13579 LNCS : 155 - 167
  • [48] Research on machine reading comprehension based on pre-trained model
    Chen G.
    Yao R.
    Zhou H.
    Li T.
    Yang W.
    International Journal of Reasoning-based Intelligent Systems, 2022, 14 (04) : 240 - 246
  • [49] A survey of text classification based on pre-trained language model
    Wu, Yujia
    Wan, Jun
    NEUROCOMPUTING, 2025, 616
  • [50] Pre-Trained Language Model Ensemble for Arabic Fake News Detection
    Al-Zahrani, Lama
    Al-Yahya, Maha
    MATHEMATICS, 2024, 12 (18)