ZARA: Improving Few-Shot Self-Rationalization for Small Language Models

被引:0
|
作者
Chen, Wei-Lin [1 ]
Yen, An-Zi [2 ]
Wu, Cheng-Kuang [1 ]
Huang, Hen-Hsen [3 ]
Chen, Hsin-Hsi [1 ]
机构
[1] Natl Taiwan Univ, Taipei, Taiwan
[2] Natl Yang Ming Chiao Tung Univ, Taipei, Taiwan
[3] Acad Sinica, Taipei, Taiwan
关键词
ERROR;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Language models (LMs) that jointly generate end-task answers as well as free-text rationales are known as self-rationalization models. Recent works demonstrate great performance gain for self-rationalization by few-shot prompting LMs with rationale-augmented exemplars. However, the ability to benefit from explanations only emerges with large-scale LMs, which have poor accessibility. In this work, we explore the less-studied setting of leveraging explanations for small LMs to improve few-shot self-rationalization. We first revisit the relationship between rationales and answers. Inspired by the implicit mental process of how human beings assess explanations, we present a novel approach, Zero-shot Augmentation of Rationale-Answer pairs (ZARA), to automatically construct pseudo-parallel data for self-training by reducing the problem of plausibility judgement to natural language inference. Experimental results show ZARA achieves SOTA performance on the FEB benchmark, for both the task accuracy and the explanation metric. In addition, we conduct human and quantitative evaluation validating ZARA's ability to automatically identify plausible and accurate rationale-answer pairs.(1)
引用
收藏
页码:4682 / 4693
页数:12
相关论文
共 50 条
  • [21] WAVPROMPT: Towards Few-Shot Spoken Language Understanding with Frozen Language Models
    Gao, Heting
    Ni, Junrui
    Qian, Kaizhi
    Zhang, Yang
    Chang, Shiyu
    Hasegawa-Johnson, Mark
    INTERSPEECH 2022, 2022, : 2738 - 2742
  • [22] Fairness-guided Few-shot Prompting for Large Language Models
    Ma, Huan
    Zhang, Changqing
    Bian, Yatao
    Liu, Lemao
    Zhang, Zhirui
    Zhao, Peilin
    Zhang, Shu
    Fu, Huazhu
    Hu, Qinghua
    Wu, Bingzhe
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [23] LLaFS: When Large Language Models Meet Few-Shot Segmentation
    Zhu, Lanyun
    Chen, Tianrun
    Ji, Deyi
    Ye, Jieping
    Liu, Jun
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2024, 2024, : 3065 - 3075
  • [24] Political Bias of Large Language Models in Few-Shot News Summarization
    Onishi, Takeshi
    Caverlee, James
    ADVANCES IN BIAS AND FAIRNESS IN INFORMATION RETRIEVAL, BIAS 2024, 2025, 2227 : 32 - 45
  • [25] Adapting Language-Audio Models as Few-Shot Audio Learners
    Liang, Jinhua
    Liu, Xubo
    Liu, Haohe
    Phan, Huy
    Benetos, Emmanouil
    Plumbley, Mark D.
    Wang, Wenwu
    INTERSPEECH 2023, 2023, : 276 - 280
  • [26] Automated Few-shot Classification with Instruction-Finetuned Language Models
    Aly, Rami
    Shi, Xingjian
    Lin, Kaixiang
    Zhang, Aston
    Wilson, Andrew Gordon
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 2414 - 2432
  • [27] Refactoring Programs Using Large Language Models with Few-Shot Examples
    Shirafuji, Atsushi
    Oda, Yusuke
    Suzuki, Jun
    Morishita, Makoto
    Watanobe, Yutaka
    PROCEEDINGS OF THE 2023 30TH ASIA-PACIFIC SOFTWARE ENGINEERING CONFERENCE, APSEC 2023, 2023, : 151 - 160
  • [28] Black Box Few-Shot Adaptation for Vision-Language models
    Ouali, Yassine
    Bulat, Adrian
    Matinez, Brais
    Tzimiropoulos, Georgios
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 15488 - 15500
  • [29] TabLLM: Few-shot Classification of Tabular Data with Large Language Models
    Hegselmann, Stefan
    Buendia, Alejandro
    Lang, Hunter
    Agrawal, Monica
    Jiang, Xiaoyi
    Sontag, David
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 206, 2023, 206
  • [30] Prompt Programming for Large Language Models: Beyond the Few-Shot Paradigm
    Reynolds, Laria
    McDonell, Kyle
    EXTENDED ABSTRACTS OF THE 2021 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS (CHI'21), 2021,