Fine-Tuning Transformer Models Using Transfer Learning for Multilingual Threatening Text Identification

被引:6
|
作者
Rehan, Muhammad [1 ]
Malik, Muhammad Shahid Iqbal [2 ]
Jamjoom, Mona Mamdouh [3 ]
机构
[1] Capital Univ Sci & Technol, Dept Comp Sci, Islamabad 44000, Pakistan
[2] Natl Res Univ Higher Sch Econ, Dept Comp Sci, Moscow 109028, Russia
[3] Princess Nourah Bint Abdulrahman Univ, Coll Comp & Informat Sci, Dept Comp Sci, Riyadh 11671, Saudi Arabia
关键词
Multi-lingual; Urdu; XLM-RoBERTa; threatening text; fine-tunning; MuRIL; LANGUAGE DETECTION;
D O I
10.1109/ACCESS.2023.3320062
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Threatening content detection on social media has recently gained attention. There is very limited work regarding threatening content detection in low-resource languages, especially in Urdu. Furthermore, previous work explored only mono-lingual approaches, and multi-lingual threatening content detection was not studied. This research addressed the task of Multi-lingual Threatening Content Detection (MTCD) in Urdu and English languages by exploiting transfer learning methodology with fine-tuning techniques. To address the multi-lingual task, we investigated two methodologies: 1) Joint multi-lingual, and 2) Joint-translated method. The former approach employs the concept of building a universal classifier for different languages whereas the latter approach applies the translation process to transform the text into one language and then perform classification. We explore the Multilingual Representations for Indian Languages (MuRIL) and Robustly Optimized BERT Pre-Training Approach (RoBERTa) with fine-tuning that already demonstrated state-of-the-art in capturing the contextual and semantic characteristics within the text. For hyper-parameters, manual search and grid search strategies are utilized to find the optimum values. Various experiments are performed on bi-lingual English and Urdu datasets and findings revealed that the proposed methodology outperformed the baselines and showed benchmark performance. The RoBERTa model achieved the highest performance by demonstrating 92% accuracy and 90% macro F1-score with the joint multi-lingual approach.
引用
收藏
页码:106503 / 106515
页数:13
相关论文
共 50 条
  • [41] AdaFilter: Adaptive Filter Fine-Tuning for Deep Transfer Learning
    Guo, Yunhui
    Li, Yandong
    Wang, Liqiang
    Rosing, Tajana
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 4060 - 4066
  • [42] Fine-Tuning Multimodal Transformer Models for Generating Actions in Virtual and Real Environments
    Staroverov, Aleksei
    Gorodetsky, Andrey S.
    Krishtopik, Andrei S.
    Izmesteva, Uliana A.
    Yudin, Dmitry A.
    Kovalev, Alexey K.
    Panov, Aleksandr I.
    IEEE ACCESS, 2023, 11 : 130548 - 130559
  • [43] Scaling Federated Learning for Fine-Tuning of Large Language Models
    Hilmkil, Agrin
    Callh, Sebastian
    Barbieri, Matteo
    Sutfeld, Leon Rene
    Zec, Edvin Listo
    Mogren, Olof
    NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS (NLDB 2021), 2021, 12801 : 15 - 23
  • [44] Using Optimal Transport as Alignment Objective for fine-tuning Multilingual Contextualized Embeddings
    Alqahtani, Sawsan
    Lalwani, Garima
    Zhang, Yi
    Romeo, Salvatore
    Mansour, Saab
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 3904 - 3919
  • [45] Diagnosis of Salivary Gland Tumors Using Transfer Learning with Fine-Tuning and Gradual Unfreezing
    Cheng, Ping-Chia
    Chiang, Hui-Hua Kenny
    DIAGNOSTICS, 2023, 13 (21)
  • [46] Fine-tuning Transformer-based MT Using Syntactic Guides
    Makwisai, Jinnawat
    Boonkwan, Prachya
    Usanavasin, Sasiporn
    Kaothanthong, Natsuda
    Okumura, Manabu
    2023 18TH INTERNATIONAL JOINT SYMPOSIUM ON ARTIFICIAL INTELLIGENCE AND NATURAL LANGUAGE PROCESSING, ISAI-NLP, 2023,
  • [47] Facial Expression Recognition using Transfer Learning and Fine-tuning Strategies: A Comparative Study
    Abdulsattar, Nadia Shamsulddin
    Hussain, Mohammed Nasser
    PROCEEDING OF THE 2ND 2022 INTERNATIONAL CONFERENCE ON COMPUTER SCIENCE AND SOFTWARE ENGINEERING (CSASE 2022), 2022, : 101 - 106
  • [48] Detecting Exams Fraud Using Transfer Learning and Fine-Tuning for ResNet50
    Luong, Huong Hoang
    Khanh, Toan Tran
    Ngoc, Minh Doan
    Kha, Minh Ho
    Duy, Khang Thuong
    Anh, Tho Tieu
    FUTURE DATA AND SECURITY ENGINEERING. BIG DATA, SECURITY AND PRIVACY, SMART CITY AND INDUSTRY 4.0 APPLICATIONS, FDSE 2022, 2022, 1688 : 747 - 754
  • [49] Enhanced Discriminative Fine-Tuning of Large Language Models for Chinese Text Classification
    Song, Jinwang
    Zan, Hongying
    Zhang, Kunli
    2024 INTERNATIONAL CONFERENCE ON ASIAN LANGUAGE PROCESSING, IALP 2024, 2024, : 168 - 174
  • [50] Fine-tuning transformers: Vocabulary transfer
    Mosin, Vladislav
    Samenko, Igor
    Kozlovskii, Borislav
    Tikhonov, Alexey
    Yamshchikov, Ivan P.
    ARTIFICIAL INTELLIGENCE, 2023, 317