Pre-trained Transformer-based Classification for Automated Patentability Examination

被引:3
|
作者
Lo, Hao-Cheng [1 ]
Chu, Jung-Mei [2 ]
机构
[1] Natl Taiwan Univ, Dept Psychol, Taipei, Taiwan
[2] Natl Taiwan Univ, Grad Inst Networking & Multimedia, Taipei, Taiwan
关键词
Patentability; Multi-label Classification; Pre-trained Transformers; Natural Language Processing;
D O I
10.1109/CSDE53843.2021.9718474
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Patentability examination, which means checking whether claims of a patent application meet the requirements for being patentable, is highly reliant on experts' arduous endeavors entailing domain knowledge. Therefore, automated patentability examination would be the immediate priority, though underappreciated. In this work, being the first to cast deep-learning light on automated patentability examination, we formulate this task as a multi-label text classification problem, which is challenging due to learning cross-sectional characteristics of abstract requirements (labels) from text content replete with inventive terms. To address this problem, we fine-tune downstream multi-label classification models over pre-trained transformer variants (BERT-Base/Large, RoBERTa-Base/Large, and XLNet) in light of their state-of-the-art achievements on many tasks. On a large USPTO patent database, we assess the performance of our models and find the model outperforming others based on the metrics, namely micro-precision, micro-recall, and micro-F1.
引用
收藏
页数:5
相关论文
共 50 条
  • [1] Towards a Transformer-Based Pre-trained Model for IoT Traffic Classification
    Bazaluk, Bruna
    Hamdan, Mosab
    Ghaleb, Mustafa
    Gismalla, Mohammed S. M.
    da Silva, Flavio S. Correa
    Batista, Daniel Macedo
    PROCEEDINGS OF 2024 IEEE/IFIP NETWORK OPERATIONS AND MANAGEMENT SYMPOSIUM, NOMS 2024, 2024,
  • [2] A survey of transformer-based multimodal pre-trained modals
    Han, Xue
    Wang, Yi-Tong
    Feng, Jun-Lan
    Deng, Chao
    Chen, Zhan-Heng
    Huang, Yu-An
    Su, Hui
    Hu, Lun
    Hu, Peng-Wei
    NEUROCOMPUTING, 2023, 515 : 89 - 106
  • [3] Pre-trained transformer-based language models for Sundanese
    Wilson Wongso
    Henry Lucky
    Derwin Suhartono
    Journal of Big Data, 9
  • [4] Pre-trained transformer-based language models for Sundanese
    Wongso, Wilson
    Lucky, Henry
    Suhartono, Derwin
    JOURNAL OF BIG DATA, 2022, 9 (01)
  • [5] TrOCR: Transformer-Based Optical Character Recognition with Pre-trained Models
    Li, Minghao
    Lv, Tengchao
    Chen, Jingye
    Cui, Lei
    Lu, Yijuan
    Florencio, Dinei
    Zhang, Cha
    Li, Zhoujun
    Wei, Furu
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 11, 2023, : 13094 - 13102
  • [6] Pre-Trained Transformer-Based Models for Text Classification Using Low-Resourced Ewe Language
    Agbesi, Victor Kwaku
    Chen, Wenyu
    Yussif, Sophyani Banaamwini
    Hossin, Md Altab
    Ukwuoma, Chiagoziem C.
    Kuadey, Noble A.
    Agbesi, Colin Collinson
    Samee, Nagwan Abdel
    Jamjoom, Mona M.
    Al-antari, Mugahed A.
    SYSTEMS, 2024, 12 (01):
  • [7] Multi-task Active Learning for Pre-trained Transformer-based Models
    Rotman, Guy
    Reichart, Roi
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2022, 10 : 1209 - 1228
  • [8] Commit-Level Software Change Intent Classification Using a Pre-Trained Transformer-Based Code Model
    Hericko, Tjasa
    Sumak, Bostjan
    Karakatic, Saso
    MATHEMATICS, 2024, 12 (07)
  • [9] Generative Pre-Trained Transformer-Based Reinforcement Learning for Testing Web Application Firewalls
    Liang, Hongliang
    Li, Xiangyu
    Xiao, Da
    Liu, Jie
    Zhou, Yanjie
    Wang, Aibo
    Li, Jin
    IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2024, 21 (01) : 309 - 324
  • [10] A Survey of Controllable Text Generation Using Transformer-based Pre-trained Language Models
    Zhang, Hanqing
    Song, Haolin
    Li, Shaoyu
    Zhou, Ming
    Song, Dawei
    ACM COMPUTING SURVEYS, 2024, 56 (03)