On Conditional and Compositional Language Model Differentiable Prompting

被引:0
|
作者
Pilault, Jonathan [1 ]
Liu, Can [2 ]
Bansal, Mohit [3 ]
Dreyer, Markus [2 ]
机构
[1] Polytech Montreal, Mila Quebec AI Inst, Montreal, PQ, Canada
[2] Amazon Alexa, Seattle, WA USA
[3] Univ North Carolina Chapel Hill, Chapel Hill, NC USA
来源
PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023 | 2023年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Prompts have been shown to be an effective method to adapt a frozen Pretrained Language Model (PLM) to perform well on downstream tasks. Prompts can be represented by a human-engineered word sequence or by a learned continuous embedding. In this work, we investigate conditional and compositional differentiable prompting. We propose a new model, Prompt Production System (PROPS), which learns to transform task instructions or input metadata, into continuous prompts that elicit task-specific outputs from the PLM. Our model uses a modular network structure based on our neural formulation of Production Systems, which allows the model to learn discrete rules - neural functions that learn to specialize in transforming particular prompt input patterns, making it suitable for compositional transfer learning and few-shot learning. We present extensive empirical and theoretical analysis and show that PROPS consistently surpasses other PLM adaptation techniques, and often improves upon fully fine-tuned models, on compositional generalization tasks, controllable summarization and multilingual translation, while needing fewer trainable parameters.
引用
收藏
页码:4136 / 4144
页数:9
相关论文
共 50 条
  • [21] Establishing best practices in large language model research: an application to repeat prompting
    Gallo, Robert J.
    Baiocchi, Michael
    Savage, Thomas R.
    Chen, Jonathan H.
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2024,
  • [22] Dialogue State Tracking with a Language Model using Schema-Driven Prompting
    Lee, Chia-Hsuan
    Cheng, Hao
    Ostendorf, Mari
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 4937 - 4949
  • [23] Conditional Mixture Path Guiding for Differentiable Rendering
    Fan, Zhimin
    Shi, Pengcheng
    Guo, Mufan
    Fu, Ruoyu
    Guo, Yanwen
    Guo, Jie
    ACM TRANSACTIONS ON GRAPHICS, 2024, 43 (04):
  • [24] Pre-service Language Teachers' Task-specific Large Language Model Prompting Practices
    Moorhouse, Benjamin Luke
    Ho, Tsz Ying
    Wu, Chenze
    Wan, Yuwei
    RELC JOURNAL, 2025,
  • [25] PHALM: Building a Knowledge Graph from Scratch by Prompting Humans and a Language Model
    Ide, Tatsuya
    Murata, Eiki
    Kawahara, Daisuke
    Yamazaki, Takato
    Li, Shengzhe
    Shinzato, Kenta
    Sato, Toshinori
    arXiv, 2023,
  • [26] Cancer Treatment Information Differences by Bilingual Prompting in Large Language Model Chatbots
    Mora, J.
    Chen, S.
    Mak, R. H.
    Bitterman, D. S.
    INTERNATIONAL JOURNAL OF RADIATION ONCOLOGY BIOLOGY PHYSICS, 2024, 120 (02): : E645 - E646
  • [27] Compositional Prompting for Anti-Forgetting in Domain Incremental Learning
    Liu, Zichen
    Peng, Yuxin
    Zhou, Jiahuan
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, 132 (12) : 5783 - 5800
  • [28] Compositional Chain-of-Thought Prompting for Large Multimodal Models
    Mitra, Chancharik
    Huang, Brandon
    Darrell, Trevor
    Herzig, Roei
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 14420 - 14431
  • [29] Prompting Language Models for Linguistic Structure
    Blevins, Terra
    Gonen, Hila
    Zettlemoyer, Luke
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 6649 - 6663
  • [30] A Simple Differentiable Programming Language
    Abadi, Martin
    Plotkin, Gordon D.
    PROCEEDINGS OF THE ACM ON PROGRAMMING LANGUAGES-PACMPL, 2020, 4 (POPL):