Compositional Zero-Shot Domain Transfer with Text-to-Text Models

被引:1
|
作者
Liu, Fangyu [1 ]
Liu, Qianchu [2 ]
Bannur, Shruthi [2 ]
Perez-Garcia, Fernando [2 ]
Usuyama, Naoto [3 ]
Zhang, Sheng [3 ]
Naumann, Tristan [3 ]
Nori, Aditya [2 ]
Poon, Hoifung [3 ]
Alvarez-Valle, Javier [2 ]
Oktay, Ozan [2 ]
Hyland, Stephanie L. [2 ]
机构
[1] Univ Cambridge, Cambridge, England
[2] Microsoft Hlth Futures, Cambridge, England
[3] Microsoft Hlth Futures, Redmond, WA USA
关键词
721.1 Computer Theory; Includes Computational Logic; Automata Theory; Switching Theory; Programming Theory - 723.2 Data Processing and Image Processing - 723.4 Artificial Intelligence;
D O I
10.1162/tacl_a_00585
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Label scarcity is a bottleneck for improving task performance in specialized domains. We propose a novel compositional transfer learning framework (DoT51) for zero-shot domain transfer. Without access to in-domain labels, DoT5 jointly learns domain knowledge (from masked language modelling of unlabelled in-domain free text) and task knowledge (from task training on more readily available general-domain data) in a multi-task manner. To improve the transferability of task training, we design a strategy named NLGU: We simultaneously train natural language generation (NLG) for in-domain label-to-data generation, which enables data augmentation for self-finetuning and natural language understanding (NLU) for label prediction. We evaluate DoT5 on the biomedical domain and the resource-lean subdomain of radiology, focusing on natural language inference, text summarization, and embedding learning. DoT5 demonstrates the effectiveness of compositional transfer learning through multi-task learning. In particular, DoT5 outperforms the current state-of-the-art in zero-shot transfer by over 7 absolute points in accuracy on RadNLI. We validate DoT5 with ablations and a case study demonstrating its ability to solve challenging NLI examples requiring in-domain expertise.
引用
收藏
页码:1097 / 1113
页数:17
相关论文
共 50 条
  • [41] Zero-shot Topical Text Classification with LLMs - an Experimental Study
    Gretz, Shai
    Halfon, Alon
    Shnayderman, Ilya
    Toledo-Ronen, Orith
    Dankin, Lena
    Katsis, Yannis
    Arviv, Ofir
    Katz, Yoav
    Slonim, Noam
    Ein-Dor, Liat
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 9647 - 9676
  • [42] Improved Zero-Shot Classification by Adapting VLMs with Text Descriptions
    Saha, Oindrila
    Van Horn, Grant
    Maji, Subhransu
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 17542 - 17552
  • [43] Evaluation of Transfer Learning for Polish with a Text-to-Text Model
    Chrabrowa, Aleksandra
    Dragan, Lukasz
    Grzegorczyk, Karol
    Kajtoch, Dariusz
    Koszowski, Mikolaj
    Mroczkowski, Robert
    Rybak, Piotr
    LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 4374 - 4394
  • [44] Zero-Shot Text Classification with Semantically Extended Textual Entailment
    Liu, Tengfei
    Hu, Yongli
    Chen, Puman
    Sun, Yanfeng
    Yin, Baocai
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [45] Neural Pipeline for Zero-Shot Data-to-Text Generation
    Kasner, Zdenek
    Dusek, Ondrej
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 3914 - 3932
  • [46] Zero-Shot Text-to-SQL Learning with Auxiliary Task
    Chang, Shuaichen
    Liu, Pengfei
    Tang, Yun
    Huang, Jing
    He, Xiaodong
    Zhou, Bowen
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 7488 - 7495
  • [47] Text2Light. Zero-Shot Text-Driven HDR Panorama Generation
    Chen, Zhaoxi
    Wang, Guangcong
    Liu, Ziwei
    ACM TRANSACTIONS ON GRAPHICS, 2022, 41 (06):
  • [48] Distractor Generation Through Text-to-Text Transformer Models
    de-Fitero-Dominguez, David
    Garcia-Lopez, Eva
    Garcia-Cabot, Antonio
    del-Hoyo-Gabaldon, Jesus-Angel
    Moreno-Cediel, Antonio
    IEEE ACCESS, 2024, 12 : 25580 - 25589
  • [49] Space-Time Diffusion Features for Zero-Shot Text-Driven Motion Transfer
    Yatim, Danah
    Fridman, Rafail
    Bar-Tal, Omer
    Kasten, Yoni
    Dekel, Tali
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2024, 2024, : 8466 - 8476
  • [50] Zero-Shot Compositional Concept Learning
    Xu, Guangyue
    Kordjamshidi, Parisa
    Chai, Joyce Y.
    1ST WORKSHOP ON META LEARNING AND ITS APPLICATIONS TO NATURAL LANGUAGE PROCESSING (METANLP 2021), 2021, : 19 - 27