Automatic Text Classification With Large Language Models: A Review of <monospace>openai</monospace> for Zero- and Few-Shot Classification

被引:0
|
作者
Anglin, Kylie L. [1 ]
Ventura, Claudia [1 ]
机构
[1] Univ Connecticut, Storrs, CT 06269 USA
关键词
large language models; LLMs; artificial intelligence; <monospace>openai</monospace>; educational measurement;
D O I
10.3102/10769986241279927
中图分类号
G40 [教育学];
学科分类号
040101 ; 120403 ;
摘要
While natural language documents, such as intervention transcripts and participant writing samples, can provide highly nuanced insights into educational and psychological constructs, researchers often find these materials difficult and expensive to analyze. Recent developments in machine learning, however, have allowed social scientists to harness the power of artificial intelligence for complex data categorization tasks. One approach, supervised learning, supports high-performance categorization yet still requires a large, hand-labeled training corpus, which can be costly. An alternative approach-zero- and few-shot classification with pretrained large language models-offers a cheaper, compelling alternative. This article considers the application of zero-shot and few-shot classification in educational research. We provide an overview of large language models, a step-by-step tutorial on using the Python openai package for zero-shot and few-shot classification, and a discussion of relevant research considerations for social scientists.<br />
引用
收藏
页数:23
相关论文
共 50 条
  • [21] A Review of Few-Shot and Zero-Shot Learning for Node Classification in Social Networks
    Chen, Junyang
    Mi, Rui
    Wang, Huan
    Wu, Huisi
    Mo, Jiqian
    Guo, Jingcai
    Lai, Zhihui
    Zhang, Liangjie
    Leung, Victor C. M.
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2024,
  • [22] Large Language Models-aided Literature Reviews: A Study on Few-Shot Relevance Classification
    Giobergia, Flavio
    Koudounas, Alkis
    Baralis, Elena
    2024 IEEE 18TH INTERNATIONAL CONFERENCE ON APPLICATION OF INFORMATION AND COMMUNICATION TECHNOLOGIES, AICT 2024, 2024,
  • [24] Investigating Prompt Learning for Chinese Few-Shot Text Classification with Pre-Trained Language Models
    Song, Chengyu
    Shao, Taihua
    Lin, Kejing
    Liu, Dengfeng
    Wang, Siyuan
    Chen, Honghui
    APPLIED SCIENCES-BASEL, 2022, 12 (21):
  • [25] A Neural Few-Shot Text Classification Reality Check
    Dopierre, Thomas
    Gravier, Christophe
    Logerais, Wilfried
    16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 935 - 943
  • [26] Retrieval-Augmented Few-shot Text Classification
    Yu, Guoxin
    Liu, Lemao
    Jiang, Haiyun
    Shi, Shuming
    Ao, Xiang
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 6721 - 6735
  • [27] Distinct Label Representations for Few-Shot Text Classification
    Ohashi, Sora
    Takayama, Junya
    Kajiwara, Tomoyuki
    Arase, Yuki
    ACL-IJCNLP 2021: THE 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 2, 2021, : 831 - 836
  • [28] Few-Shot Text Classification with External Knowledge Expansion
    Guan, Jian
    Xu, Rui
    Ya, Jing
    Tang, Qiu
    Xue, Jidong
    Zhang, Ni
    2021 5TH INTERNATIONAL CONFERENCE ON INNOVATION IN ARTIFICIAL INTELLIGENCE (ICIAI 2021), 2021, : 184 - 189
  • [29] Zero- and few-shot prompting of generative large language models provides weak assessment of risk of bias in clinical trials
    Suster, Simon
    Baldwin, Timothy
    Verspoor, Karin
    RESEARCH SYNTHESIS METHODS, 2024, 15 (06) : 988 - 1000
  • [30] Hierarchical Verbalizer for Few-Shot Hierarchical Text Classification
    Ji, Ke
    Lian, Yixin
    Gao, Jingsheng
    Wang, Baoyuan
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 2918 - 2933