Limitations of Large Language Models in Propaganda Detection Task

被引:0
|
作者
Szwoch, Joanna [1 ]
Staszkow, Mateusz [2 ]
Rzepka, Rafal [3 ]
Araki, Kenji [3 ]
机构
[1] Hokkaido Univ, Grad Sch Informat Sci & Technol, Sapporo 0600808, Japan
[2] Mateusz Staszkow Software Dev, PL-01234 Warsaw, Poland
[3] Hokkaido Univ, Fac Informat Sci & Technol, Sapporo 0600808, Japan
来源
APPLIED SCIENCES-BASEL | 2024年 / 14卷 / 10期
关键词
propaganda detection; media bias; online news analysis; propaganda in online news; propaganda techniques; FAKE NEWS; MEDIA;
D O I
10.3390/app14104330
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Propaganda in the digital era is often associated with online news. In this study, we focused on the use of large language models and their detection of propaganda techniques in the electronic press to investigate whether it is a noteworthy replacement for human annotators. We prepared prompts for generative pre-trained transformer models to find spans in news articles where propaganda techniques appear and name them. Our study was divided into three experiments on different datasets-two based on an annotated SemEval2020 Task 11 corpora and one on an unannotated subset of the Polish Online News Corpus, which we claim to be an even bigger challenge as an example of an under-resourced language. Reproduction of the results of the first experiment resulted in a higher recall of 64.53% than the original run, and the highest precision of 81.82% was achieved for gpt-4-1106-preview CoT. None of our attempts outperformed the baseline F1 score. One of the attempts with gpt-4-0125-preview on original SemEval2020 Task 11 achieved an almost 20% F1 score, but it was below the baseline, which oscillated around 50%. Part of our work that was dedicated to Polish articles showed that gpt-4-0125-preview had a 74% accuracy in the binary detection of propaganda techniques and 69% in propaganda technique classification. The results for SemEval2020 show that the outputs of generative models tend to be unpredictable and are hardly reproducible for propaganda detection. For the time being, these are unreliable methods for this task, but we believe they can help to generate more training data.
引用
收藏
页数:22
相关论文
共 50 条
  • [1] Limitations of large language models in medical applications
    Deng, Jiawen
    Zubair, Areeba
    Park, Ye-Jean
    POSTGRADUATE MEDICAL JOURNAL, 2023, 99 (1178) : 1298 - 1299
  • [2] Fundamental Limitations of Alignment in Large Language Models
    Wolf, Yotam
    Wies, Noam
    Avnery, Oshri
    Levine, Yoav
    Shashua, Amnon
    arXiv, 2023,
  • [3] Teaching the Limitations of Large Language Models in Medical School
    Gunawardene, Araliya N.
    Schmuter, Gabriella
    JOURNAL OF SURGICAL EDUCATION, 2024, 81 (05) : 625 - 625
  • [4] Semantic anomaly detection with large language models
    Amine Elhafsi
    Rohan Sinha
    Christopher Agia
    Edward Schmerling
    Issa A. D. Nesnas
    Marco Pavone
    Autonomous Robots, 2023, 47 : 1035 - 1055
  • [5] Semantic anomaly detection with large language models
    Elhafsi, Amine
    Sinha, Rohan
    Agia, Christopher
    Schmerling, Edward
    Nesnas, Issa A. D.
    Pavone, Marco
    AUTONOMOUS ROBOTS, 2023, 47 (08) : 1035 - 1055
  • [6] Finetuning Large Language Models for Vulnerability Detection
    Shestov, Aleksei
    Levichev, Rodion
    Mussabayev, Ravil
    Maslov, Evgeny
    Zadorozhny, Pavel
    Cheshkov, Anton
    Mussabayev, Rustam
    Toleu, Alymzhan
    Tolegen, Gulmira
    Krassovitskiy, Alexander
    IEEE ACCESS, 2025, 13 : 38889 - 38900
  • [7] Detection avoidance techniques for large language models
    Schneider, Sinclair
    Steuber, Florian
    Schneider, Joao A. G.
    Rodosek, Gabi Dreo
    DATA & POLICY, 2025, 7
  • [8] Task and Motion Planning with Large Language Models for Object Rearrangement
    Ding, Yan
    Zhang, Xiaohan
    Paxton, Chris
    Zhang, Shiqi
    2023 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, IROS, 2023, : 2086 - 2092
  • [9] Large Language Models as Commonsense Knowledge for Large-Scale Task Planning
    Zhao, Zirui
    Lee, Wee Sun
    Hsu, David
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [10] Beyond the limitations of any imaginable mechanism: Large language models and psycholinguistics
    Houghton, Conor
    Kazanina, Nina
    Sukumaran, Priyanka
    BEHAVIORAL AND BRAIN SCIENCES, 2023, 46