Compositional Zero-Shot Domain Transfer with Text-to-Text Models

被引:1
|
作者
Liu, Fangyu [1 ]
Liu, Qianchu [2 ]
Bannur, Shruthi [2 ]
Perez-Garcia, Fernando [2 ]
Usuyama, Naoto [3 ]
Zhang, Sheng [3 ]
Naumann, Tristan [3 ]
Nori, Aditya [2 ]
Poon, Hoifung [3 ]
Alvarez-Valle, Javier [2 ]
Oktay, Ozan [2 ]
Hyland, Stephanie L. [2 ]
机构
[1] Univ Cambridge, Cambridge, England
[2] Microsoft Hlth Futures, Cambridge, England
[3] Microsoft Hlth Futures, Redmond, WA USA
关键词
721.1 Computer Theory; Includes Computational Logic; Automata Theory; Switching Theory; Programming Theory - 723.2 Data Processing and Image Processing - 723.4 Artificial Intelligence;
D O I
10.1162/tacl_a_00585
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Label scarcity is a bottleneck for improving task performance in specialized domains. We propose a novel compositional transfer learning framework (DoT51) for zero-shot domain transfer. Without access to in-domain labels, DoT5 jointly learns domain knowledge (from masked language modelling of unlabelled in-domain free text) and task knowledge (from task training on more readily available general-domain data) in a multi-task manner. To improve the transferability of task training, we design a strategy named NLGU: We simultaneously train natural language generation (NLG) for in-domain label-to-data generation, which enables data augmentation for self-finetuning and natural language understanding (NLU) for label prediction. We evaluate DoT5 on the biomedical domain and the resource-lean subdomain of radiology, focusing on natural language inference, text summarization, and embedding learning. DoT5 demonstrates the effectiveness of compositional transfer learning through multi-task learning. In particular, DoT5 outperforms the current state-of-the-art in zero-shot transfer by over 7 absolute points in accuracy on RadNLI. We validate DoT5 with ablations and a case study demonstrating its ability to solve challenging NLI examples requiring in-domain expertise.
引用
收藏
页码:1097 / 1113
页数:17
相关论文
共 50 条
  • [21] Generalized Zero-Shot Text Classification for ICD Coding
    Song, Congzheng
    Zhang, Shanghang
    Sadoughi, Najmeh
    Xie, Pengtao
    Xing, Eric
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 4018 - 4024
  • [22] PRESENT: Zero-Shot Text-to-Prosody Control
    Lam, Perry
    Zhang, Huayun
    Chen, Nancy F.
    Sisman, Berrak
    Herremans, Dorien
    IEEE SIGNAL PROCESSING LETTERS, 2025, 32 : 776 - 780
  • [23] Homograph Disambiguation with Text-to-Text Transfer Transformer
    Rezackova, Marketa
    Tihelka, Daniel
    Matousek, Jindrich
    INTERSPEECH 2024, 2024, : 2785 - 2789
  • [24] Learning to Speak from Text: Zero-Shot Multilingual Text-to-Speech with Unsupervised Text Pretraining
    Saeki, Takaaki
    Maiti, Soumi
    Li, Xinjian
    Watanabe, Shinji
    Takamichi, Shinnosuke
    Saruwatari, Hiroshi
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 5179 - 5187
  • [25] Task guided representation learning using compositional models for zero-shot domain adaptation
    Liu, Shuang
    Ozay, Mete
    NEURAL NETWORKS, 2023, 165 : 370 - 380
  • [26] Zero-Shot Text-to-Speech for Text-Based Insertion in Audio Narration
    Tang, Chuanxin
    Luo, Chong
    Zhao, Zhiyuan
    Yin, Dacheng
    Zhao, Yucheng
    Zeng, Wenjun
    INTERSPEECH 2021, 2021, : 3600 - 3604
  • [27] Modular Speech-to-Text Translation for Zero-Shot Cross-Modal Transfer
    Duquenne, Paul-Ambroise
    Schwenk, Holger
    Sagot, Benoit
    INTERSPEECH 2023, 2023, : 32 - 36
  • [28] Zero-Shot Text Recognition Combining Transfer Guide and Bidirectional Cycle Structure GAN
    Zhang G.
    Long B.
    Lu F.
    Moshi Shibie yu Rengong Zhineng/Pattern Recognition and Artificial Intelligence, 2020, 33 (12): : 1083 - 1096
  • [29] Adapting Pretrained Text-to-Text Models for Long Text Sequences
    Xiong, Wenhan
    Gupta, Anchit
    Toshniwal, Shubham
    Mehdad, Yashar
    Yih, Wen-tau
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 5566 - 5578
  • [30] Zero-Shot Contrastive Loss for Text-Guided Diffusion Image Style Transfer
    Yang, Serin
    Hwang, Hyunmin
    Ye, Jong Chul
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 22816 - 22825