Pre-trained Transformer-based Classification for Automated Patentability Examination

被引:3
|
作者
Lo, Hao-Cheng [1 ]
Chu, Jung-Mei [2 ]
机构
[1] Natl Taiwan Univ, Dept Psychol, Taipei, Taiwan
[2] Natl Taiwan Univ, Grad Inst Networking & Multimedia, Taipei, Taiwan
关键词
Patentability; Multi-label Classification; Pre-trained Transformers; Natural Language Processing;
D O I
10.1109/CSDE53843.2021.9718474
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Patentability examination, which means checking whether claims of a patent application meet the requirements for being patentable, is highly reliant on experts' arduous endeavors entailing domain knowledge. Therefore, automated patentability examination would be the immediate priority, though underappreciated. In this work, being the first to cast deep-learning light on automated patentability examination, we formulate this task as a multi-label text classification problem, which is challenging due to learning cross-sectional characteristics of abstract requirements (labels) from text content replete with inventive terms. To address this problem, we fine-tune downstream multi-label classification models over pre-trained transformer variants (BERT-Base/Large, RoBERTa-Base/Large, and XLNet) in light of their state-of-the-art achievements on many tasks. On a large USPTO patent database, we assess the performance of our models and find the model outperforming others based on the metrics, namely micro-precision, micro-recall, and micro-F1.
引用
收藏
页数:5
相关论文
共 50 条
  • [21] Integrally Pre-Trained Transformer Pyramid Networks
    Tian, Yunjie
    Xie, Lingxi
    Wang, Zhaozhi
    Wei, Longhui
    Zhang, Xiaopeng
    Jiao, Jianbin
    Wang, Yaowei
    Tian, Qi
    Ye, Qixiang
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 18610 - 18620
  • [22] Automatic Question Generation using RNN-based and Pre-trained Transformer-based Models in Low Resource Indonesian Language
    Vincentio, Karissa
    Suhartono, Derwin
    INFORMATICA-AN INTERNATIONAL JOURNAL OF COMPUTING AND INFORMATICS, 2022, 46 (07): : 103 - 118
  • [23] Framing and BERTology: A Data-Centric Approach to Integration of Linguistic Features into Transformer-Based Pre-trained Language Models
    Avetisyan, Hayastan
    Safikhani, Parisa
    Broneske, David
    INTELLIGENT SYSTEMS AND APPLICATIONS, VOL 4, INTELLISYS 2023, 2024, 825 : 81 - 90
  • [24] Simple and Effective Multimodal Learning Based on Pre-Trained Transformer Models
    Miyazawa, Kazuki
    Kyuragi, Yuta
    Nagai, Takayuki
    IEEE ACCESS, 2022, 10 : 29821 - 29833
  • [25] Selection of pre-trained weights for transfer learning in automated cytomegalovirus retinitis classification
    Choopong, Pitipol
    Kusakunniran, Worapan
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [26] A survey of text classification based on pre-trained language model
    Wu, Yujia
    Wan, Jun
    NEUROCOMPUTING, 2025, 616
  • [27] Speech Topic Classification Based on Pre-trained and Graph Networks
    Niu, Fangjing
    Cao, Tengfei
    Hu, Ying
    Huang, Hao
    He, Liang
    2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 1721 - 1726
  • [28] Underwater Image Enhancement Using Pre-trained Transformer
    Boudiaf, Abderrahmene
    Guo, Yuhang
    Ghimire, Adarsh
    Werghi, Naoufel
    De Masi, Giulia
    Javed, Sajid
    Dias, Jorge
    IMAGE ANALYSIS AND PROCESSING, ICIAP 2022, PT III, 2022, 13233 : 480 - 488
  • [29] Detecting Syntactic Change with Pre-trained Transformer Models
    Hou, Liwen
    Smith, David A.
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 3564 - 3574
  • [30] A Robust Approach to Fine-tune Pre-trained Transformer-based models for Text Summarization through Latent Space Compression
    Falaki, Ala Alam
    Gras, Robin
    2022 21ST IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS, ICMLA, 2022, : 160 - 167