Using Pre-trained Language Model to Enhance Active Learning for Sentence Matching

被引:0
|
作者
Bai, Guirong [1 ,2 ]
He, Shizhu [1 ,2 ]
Liu, Kang [1 ,2 ]
Zhao, Jun [1 ,2 ]
机构
[1] Chinese Acad Sci, Natl Lab Pattern Recognit, Inst Automat, 95 Zhongguancun East Rd, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Sch Artificial Intelligence, 95 Zhongguancun East Rd, Beijing 100190, Peoples R China
基金
中国国家自然科学基金;
关键词
Sentence matching; active learning; pre-trained language model;
D O I
10.1145/3480937
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Active learning is an effective method to substantially alleviate the problem of expensive annotation cost for data-driven models. Recently, pre-trained language models have been demonstrated to be powerful for learning language representations. In this article, we demonstrate that the pre-trained language model can also utilize its learned textual characteristics to enrich criteria of active learning. Specifically, we provide extra textual criteria with the pre-trained language model to measure instances, including noise, coverage, and diversity. With these extra textual criteria, we can select more efficient instances for annotation and obtain better results. We conduct experiments on both English and Chinese sentence matching datasets. The experimental results show that the proposed active learning approach can be enhanced by the pre-trained language model and obtain better performance.
引用
收藏
页数:19
相关论文
共 50 条
  • [31] Kurdish Sign Language Recognition Using Pre-Trained Deep Learning Models
    Alsaud, Ali A.
    Yousif, Raghad Z.
    Aziz, Marwan. M.
    Kareem, Shahab W.
    Maho, Amer J.
    JOURNAL OF ELECTRICAL SYSTEMS, 2024, 20 (06) : 1334 - 1344
  • [32] Misspelling Correction with Pre-trained Contextual Language Model
    Hu, Yifei
    Ting, Xiaonan
    Ko, Youlim
    Rayz, Julia Taylor
    PROCEEDINGS OF 2020 IEEE 19TH INTERNATIONAL CONFERENCE ON COGNITIVE INFORMATICS & COGNITIVE COMPUTING (ICCI*CC 2020), 2020, : 144 - 149
  • [33] CLIP-Llama: A New Approach for Scene Text Recognition with a Pre-Trained Vision-Language Model and a Pre-Trained Language Model
    Zhao, Xiaoqing
    Xu, Miaomiao
    Silamu, Wushour
    Li, Yanbing
    SENSORS, 2024, 24 (22)
  • [34] Emotional Paraphrasing Using Pre-trained Language Models
    Casas, Jacky
    Torche, Samuel
    Daher, Karl
    Mugellini, Elena
    Abou Khaled, Omar
    2021 9TH INTERNATIONAL CONFERENCE ON AFFECTIVE COMPUTING AND INTELLIGENT INTERACTION WORKSHOPS AND DEMOS (ACIIW), 2021,
  • [35] Hockey activity recognition using pre-trained deep learning model
    Rangasamy, Keerthana
    As'ari, Muhammad Amir
    Rahmad, Nur Azmina
    Ghazali, Nurul Fathiah
    ICT EXPRESS, 2020, 6 (03): : 170 - 174
  • [36] Online Active Model Selection for Pre-trained Classifiers
    Karimi, Mohammad Reza
    Guerel, Nezihe Merve
    Karlas, Bojan
    Rausch, Johannes
    Zhang, Ce
    Krause, Andreas
    24TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS (AISTATS), 2021, 130 : 307 - +
  • [37] CommitBERT: Commit Message Generation Using Pre-Trained Programming Language Model
    Jung, Tae-Hwan
    NLP4PROG 2021: THE 1ST WORKSHOP ON NATURAL LANGUAGE PROCESSING FOR PROGRAMMING (NLP4PROG 2021), 2021, : 26 - 33
  • [38] Meta Distant Transfer Learning for Pre-trained Language Models
    Wang, Chengyu
    Pan, Haojie
    Qiu, Minghui
    Yang, Fei
    Huang, Jun
    Zhang, Yin
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 9742 - 9752
  • [39] CommitBERT: Commit message generation using pre-trained programming language model
    Jung, Tae-Hwan
    arXiv, 2021,
  • [40] Classifying informative tweets using feature enhanced pre-trained language model
    Yandrapati, Prakash Babu
    Eswari, R.
    SOCIAL NETWORK ANALYSIS AND MINING, 2024, 14 (01)