Parameter-efficient feature-based transfer for paraphrase identification

被引:1
|
作者
Liu, Xiaodong [1 ]
Rzepka, Rafal [2 ]
Araki, Kenji [2 ]
机构
[1] Hokkaido Univ, Grad Sch Informat Sci & Technol, Sapporo, Hokkaido, Japan
[2] Hokkaido Univ, Fac Informat Sci & Technol, Sapporo, Hokkaido, Japan
关键词
Parameter-efficient feature-based transfer; Paraphrase identification; Natural language inference; Semantic textual similarity; Continual learning;
D O I
10.1017/S135132492200050X
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
There are many types of approaches for Paraphrase Identification (PI), an NLP task of determining whether a sentence pair has equivalent semantics. Traditional approaches mainly consist of unsupervised learning and feature engineering, which are computationally inexpensive. However, their task performance is moderate nowadays. To seek a method that can preserve the low computational costs of traditional approaches but yield better task performance, we take an investigation into neural network-based transfer learning approaches. We discover that by improving the usage of parameters efficiently for feature-based transfer, our research goal can be accomplished. Regarding the improvement, we propose a pre-trained task-specific architecture. The fixed parameters of the pre-trained architecture can be shared by multiple classifiers with small additional parameters. As a result, the computational cost left involving parameter update is only generated from classifier-tuning: the features output from the architecture combined with lexical overlap features are fed into a single classifier for tuning. Furthermore, the pre-trained task-specific architecture can be applied to natural language inference and semantic textual similarity tasks as well. Such technical novelty leads to slight consumption of computational and memory resources for each task and is also conducive to power-efficient continual learning. The experimental results show that our proposed method is competitive with adapter-BERT (a parameter-efficient fine-tuning approach) over some tasks while consuming only 16% trainable parameters and saving 69-96% time for parameter update.
引用
收藏
页码:1066 / 1096
页数:31
相关论文
共 50 条
  • [1] Parameter-Efficient Transfer Learning for NLP
    Houlsby, Neil
    Giurgiu, Andrei
    Jastrzebski, Stanislaw
    Morrone, Bruna
    de laroussilhe, Quentin
    Gesmundo, Andrea
    Attariyan, Mona
    Gelly, Sylvain
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [2] Parameter-Efficient Transfer Learning with Diff Pruning
    Guo, Demi
    Rush, Alexander M.
    Kim, Yoon
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 4884 - 4896
  • [3] Abnormal Action Detection Based on Parameter-Efficient Transfer Learning in Laboratory Scenarios
    Liu, Changyu
    Huang, Hao
    Huang, Guogang
    Wu, Chunyin
    Liang, Yingqi
    CMC-COMPUTERS MATERIALS & CONTINUA, 2024, 80 (03): : 4219 - 4242
  • [4] Conditional Adapters: Parameter-efficient Transfer Learning with Fast Inference
    Lei, Tao
    Bai, Junwen
    Brahma, Siddhartha
    Ainslie, Joshua
    Lee, Kenton
    Zhou, Yanqi
    Du, Nan
    Zhao, Vincent Y.
    Wu, Yuexin
    Li, Bo
    Zhang, Yu
    Chang, Ming-Wei
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [5] Parameter-Efficient Transfer Learning for Medical Visual Question Answering
    Liu, Jiaxiang
    Hu, Tianxiang
    Zhang, Yan
    Feng, Yang
    Hao, Jin
    Lv, Junhui
    Liu, Zuozhu
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2024, 8 (04): : 2816 - 2826
  • [6] Plant leaf disease identification by parameter-efficient transformer with adapter
    Xu, Xingshi
    Yang, Guangyuan
    Wang, Yunfei
    Shang, Yuying
    Hua, Zhixin
    Wang, Zheng
    Song, Huaibo
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 138
  • [7] AiRs: Adapter in Remote Sensing for Parameter-Efficient Transfer Learning
    Hu, Leiyi
    Yu, Hongfeng
    Lu, Wanxuan
    Yin, Dongshuo
    Sun, Xian
    Fu, Kun
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2024, 62 : 1 - 18
  • [8] Parameter-Efficient Masking Networks
    Bai, Yue
    Wang, Huan
    Ma, Xu
    Zhang, Yitian
    Tao, Zhiqiang
    Fu, Yun
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [9] PreAdapter: Sparse Adaptive Parameter-efficient Transfer Learning for Language Models
    Mao, Chenyang
    Jin, Xiaoxiao
    Yue, Dengfeng
    Leng, Tuo
    2024 7TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND BIG DATA, ICAIBD 2024, 2024, : 218 - 225
  • [10] Parameter-Efficient Transfer from Sequential Behaviors for User Modeling and Recommendation
    Yuan, Fajie
    He, Xiangnan
    Karatzoglou, Alexandros
    Zhang, Liguang
    PROCEEDINGS OF THE 43RD INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '20), 2020, : 1469 - 1478