Pre-trained Transformer-based Classification for Automated Patentability Examination

被引:3
|
作者
Lo, Hao-Cheng [1 ]
Chu, Jung-Mei [2 ]
机构
[1] Natl Taiwan Univ, Dept Psychol, Taipei, Taiwan
[2] Natl Taiwan Univ, Grad Inst Networking & Multimedia, Taipei, Taiwan
关键词
Patentability; Multi-label Classification; Pre-trained Transformers; Natural Language Processing;
D O I
10.1109/CSDE53843.2021.9718474
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Patentability examination, which means checking whether claims of a patent application meet the requirements for being patentable, is highly reliant on experts' arduous endeavors entailing domain knowledge. Therefore, automated patentability examination would be the immediate priority, though underappreciated. In this work, being the first to cast deep-learning light on automated patentability examination, we formulate this task as a multi-label text classification problem, which is challenging due to learning cross-sectional characteristics of abstract requirements (labels) from text content replete with inventive terms. To address this problem, we fine-tune downstream multi-label classification models over pre-trained transformer variants (BERT-Base/Large, RoBERTa-Base/Large, and XLNet) in light of their state-of-the-art achievements on many tasks. On a large USPTO patent database, we assess the performance of our models and find the model outperforming others based on the metrics, namely micro-precision, micro-recall, and micro-F1.
引用
收藏
页数:5
相关论文
共 50 条
  • [31] Generative Pre-Trained Transformer for Cardiac Abnormality Detection
    Gaudilliere, Pierre Louis
    Sigurthorsdottir, Halla
    Aguet, Clementine
    Van Zaen, Jerome
    Lemay, Mathieu
    Delgado-Gonzalo, Ricard
    2021 COMPUTING IN CARDIOLOGY (CINC), 2021,
  • [32] OMPGPT: A Generative Pre-trained Transformer Model for OpenMP
    Chen, Le
    Bhattacharjee, Arijit
    Ahmed, Nesreen
    Hasabnis, Niranjan
    Oren, Gal
    Vo, Vy
    Jannesari, Ali
    EURO-PAR 2024: PARALLEL PROCESSING, PT I, EURO-PAR 2024, 2024, 14801 : 121 - 134
  • [33] On the effect of dropping layers of pre-trained transformer models
    Sajjad, Hassan
    Dalvi, Fahim
    Durrani, Nadir
    Nakov, Preslav
    COMPUTER SPEECH AND LANGUAGE, 2022, 77
  • [34] Aspect-Based API Review Classification: How Far Can Pre-Trained Transformer Model Go?
    Yang, Chengran
    Xu, Bowen
    Khan, Junaed Younus
    Uddin, Gias
    Han, Donggyun
    Yang, Zhou
    Lo, David
    2022 IEEE INTERNATIONAL CONFERENCE ON SOFTWARE ANALYSIS, EVOLUTION AND REENGINEERING (SANER 2022), 2022, : 385 - 395
  • [35] An Application of pre-Trained CNN for Image Classification
    Abdullah
    Hasan, Mohammad S.
    2017 20TH INTERNATIONAL CONFERENCE OF COMPUTER AND INFORMATION TECHNOLOGY (ICCIT), 2017,
  • [36] Patent classification with pre-trained Bert model
    Kahraman, Selen Yuecesoy
    Durmusoglu, Alptekin
    Dereli, Tuerkay
    JOURNAL OF THE FACULTY OF ENGINEERING AND ARCHITECTURE OF GAZI UNIVERSITY, 2024, 39 (04): : 2485 - 2496
  • [37] UAPT: an underwater acoustic target recognition method based on pre-trained Transformer
    Tang, Jun
    Ma, Enxue
    Qu, Yang
    Gao, Wenbo
    Zhang, Yuchen
    Gan, Lin
    MULTIMEDIA SYSTEMS, 2025, 31 (01)
  • [38] Transformer based contextualization of pre-trained word embeddings for irony detection in Twitter
    Angel Gonzalez, Jose
    Hurtado, Lluis-F
    Pla, Ferran
    INFORMATION PROCESSING & MANAGEMENT, 2020, 57 (04)
  • [39] Photo-based Carbohydrates Counting using Pre-trained Transformer Models
    Contreras, Ivan
    Guso, Marti
    Beneyto, Aleix
    Vehi, Josep
    IFAC PAPERSONLINE, 2023, 56 (02): : 11533 - 11538
  • [40] SMILESynergy: Anticancer drug synergy prediction based on Transformer pre-trained model
    Zhang L.
    Qin Y.
    Chen M.
    Shengwu Yixue Gongchengxue Zazhi/Journal of Biomedical Engineering, 2023, 40 (03): : 544 - 551