MultiAICL: Multi-task Tuning for Augmented In-Context Learning in Text Style Transfer

被引:0
|
作者
Zhu, Linan [1 ]
Zhou, Zehai [1 ]
Chen, Xiangfan [1 ]
Guo, Xiaolei [1 ]
Kong, Xiangjie [1 ]
机构
[1] Zhejiang Univ Technol, Hangzhou, Zhejiang, Peoples R China
基金
中国国家自然科学基金;
关键词
In-Context Learning; Text Style Transfer; Large Language Models;
D O I
10.1007/978-981-97-9437-9_5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In-context learning (ICL) enhances the performance of large language models (LLMs) across various natural language process (NLP) tasks by simply demonstrating a few-shot of examples or instructions during inference. However, ICL still encounters significant challenges on the text style transfer (TST) tasks, which require high levels of model reasoning. The existing ICL ability has not been further developed because LLMs lack the process of training and learning in context. To address these issues, we introduce Multi-Task Tuning for Augmented In-Context Learning (MultiAICL), a framework designed to enhance model ICL ability by simulating LLM's supervised fine-tuning steps. MultiAICL contains three main components: firstly, we construct example instructions for multiple tasks from the text corpus, where these examples are in the form of text-label pairs; secondly, we propose the Multi-Task Tuning (MTT) module, which tunes the model by randomly combining example instructions; and thirdly, we design the Augmented In-context Learning (AICL) module, which incorporates different tasks into example templates to infer the model. MultiAICL improves the ICL ability of LLMs while maintaining their generalization across multiple tasks, thus encouraging models to generate high-quality text. Extensive experiments show that MultiAICL achieves excellent results on all 6 TST tasks, even outperforming larger LLMs. The code and data are available at https://github.com/fuz999/NLPCC-2024- MultiAICL.
引用
收藏
页码:55 / 66
页数:12
相关论文
共 50 条
  • [11] MCapsNet: Capsule Network for Text with Multi-Task Learning
    Xiao, Liqiang
    Zhang, Honglun
    Chen, Wenqing
    Wang, Yongkun
    Jin, Yaohui
    2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 4565 - 4574
  • [12] A Multi-Task Learning Framework for Abstractive Text Summarization
    Lu, Yao
    Liu, Linqing
    Jiang, Zhile
    Yang, Min
    Goebel, Randy
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 9987 - 9988
  • [13] Adaptive multi-task learning for speech to text translation
    Feng, Xin
    Zhao, Yue
    Zong, Wei
    Xu, Xiaona
    EURASIP JOURNAL ON AUDIO SPEECH AND MUSIC PROCESSING, 2024, 2024 (01):
  • [14] In-Context Learning Creates Task Vectors
    Hendel, Roee
    Geva, Mor
    Globerson, Amir
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 9318 - 9333
  • [15] A multi-task transfer learning method with dictionary learning
    Zheng, Xin
    Lin, Luyue
    Liu, Bo
    Xiao, Yanshan
    Xiong, Xiaoming
    KNOWLEDGE-BASED SYSTEMS, 2020, 191
  • [16] Multilingual multi-task quantum transfer learning
    Buonaiuto, Giuseppe
    Guarasci, Raffaele
    De Pietro, Giuseppe
    Esposito, Massimo
    QUANTUM MACHINE INTELLIGENCE, 2025, 7 (01)
  • [17] Retrieval-style In-context Learning for Few-shot Hierarchical Text Classification
    Chen, Huiyao
    Zhao, Yu
    Chen, Zulong
    Wang, Mengjia
    Li, Liangyue
    Zhang, Meishan
    Zhang, Min
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2024, 12 : 1214 - 1231
  • [18] Multi-Task Transfer Matters During Instruction-Tuning
    Mueller, David
    Dredze, Mark
    Andrews, Nicholas
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 14880 - 14891
  • [19] Multi-task gradient descent for multi-task learning
    Lu Bai
    Yew-Soon Ong
    Tiantian He
    Abhishek Gupta
    Memetic Computing, 2020, 12 : 355 - 369
  • [20] Multi-task gradient descent for multi-task learning
    Bai, Lu
    Ong, Yew-Soon
    He, Tiantian
    Gupta, Abhishek
    MEMETIC COMPUTING, 2020, 12 (04) : 355 - 369