A novel approach to measuring the scope of patent claims based on probabilities obtained from (large) language models

被引:0
|
作者
Ragot, Sebastien [1 ]
机构
[1] E Blum & Co Ltd, Patent & Trademark Attorneys VSP, Vorderberg 11, CH-8044 Zurich, Switzerland
关键词
Patent scope; Patent value; Patent claims; Language models; Large language models; GPT; Information theory; Self-information;
D O I
10.1016/j.wpi.2024.102321
中图分类号
G25 [图书馆学、图书馆事业]; G35 [情报学、情报工作];
学科分类号
1205 ; 120501 ;
摘要
This work proposes to measure the scope of a patent claim as the reciprocal of self-information contained in this claim. Self-information is calculated based on a probability of occurrence of the claim, where this probability is obtained from a language model. Grounded in information theory, this approach is based on the assumption that an unlikely concept is more informative than a usual concept, insofar as it is more surprising. In turn, the more surprising the information required to define the claim, the narrower its scope. Seven language models are considered, ranging from simplest models (each word or character has an identical probability) to intermediate models (based on average word or character frequencies), to large language models (LLMs) such as GPT2 and davinci-002. Remarkably, when using the simplest language models to compute the probabilities, the scope becomes proportional to the reciprocal of the number of words or characters involved in the claim, a metric already used in previous works. Application is made to multiple series of patent claims directed to distinct inventions, where each series consists of a set of claims having a gradually decreasing scope. The performance of the language models is then assessed through several ad hoc tests. The LLMs outperform models based on word and character frequencies, which themselves outdo the simplest models based on word or character counts. Interestingly, however, the character count appears to be a more reliable indicator than the word count.
引用
收藏
页数:29
相关论文
共 50 条
  • [21] Needs Companion: A Novel Approach to Continuous User Needs Sensing Using Virtual Agents and Large Language Models
    Nakata, Takuya
    Nakamura, Masahide
    Chen, Sinan
    Saiki, Sachio
    SENSORS, 2024, 24 (21)
  • [22] Evaluating large language models for criterion-based grading from agreement to consistency
    Zhang, Da-Wei
    Boey, Melissa
    Tan, Yan Yu
    Jia, Alexis Hoh Sheng
    NPJ SCIENCE OF LEARNING, 2024, 9 (01)
  • [23] Industrial Internet of Things With Large Language Models (LLMs): An Intelligence-Based Reinforcement Learning Approach
    Ren, Yuzheng
    Zhang, Haijun
    Yu, Fei Richard
    Li, Wei
    Zhao, Pincan
    He, Ying
    IEEE TRANSACTIONS ON MOBILE COMPUTING, 2025, 24 (05) : 4136 - 4152
  • [24] A Hybrid Approach to Dimensional Aspect-Based Sentiment Analysis Using BERT and Large Language Models
    Zhang, Yice
    Xu, Hongling
    Zhang, Delong
    Xu, Ruifeng
    ELECTRONICS, 2024, 13 (18)
  • [25] A Systematic Approach to Prompting Large Language Models for Automated Feature Extraction from Cardiovascular Imaging Reports
    Goldfinger, Shir
    Mackay, Emily
    Chan, Trevor
    Eswar, Vikram
    Grasfield, Rachel
    Yan, Vivian
    Barreto, David
    Pouch, Alison
    CIRCULATION, 2024, 150
  • [26] A Congruence-based Approach to Active Automata Learning from Neural Language Models
    Mayr, Franz
    Yovine, Sergio
    Carrasco, Matias
    Pan, Federico
    Vilensky, Federico
    INTERNATIONAL CONFERENCE ON GRAMMATICAL INFERENCE, VOL 217, 2023, 217 : 250 - 264
  • [27] Building a hospitable and reliable dialogue system for android robots: a scenario-based approach with large language models
    Yamazaki, Takato
    Yoshikawa, Katsumasa
    Kawamoto, Toshiki
    Mizumoto, Tomoya
    Ohagi, Masaya
    Sato, Toshinori
    ADVANCED ROBOTICS, 2023, 37 (21) : 1364 - 1381
  • [28] A Novel Approach for Mixed-Methods Research Using Large Language Models: A Report Using Patients' Perspectives on Barriers to Arthroplasty
    Mannstadt, Insa
    Goodman, Susan M.
    Rajan, Mangala
    Young, Sarah R.
    Wang, Fei
    Navarro-Millan, Iris
    Mehta, Bella
    ACR OPEN RHEUMATOLOGY, 2024, 6 (06) : 375 - 379
  • [29] Grammar-obeying program synthesis: A novel approach using large language models and many-objective genetic programming
    Tao, Ning
    Ventresque, Anthony
    Nallur, Vivek
    Saber, Takfarinas
    COMPUTER STANDARDS & INTERFACES, 2025, 92
  • [30] A novel approach for efficient fatigue based shape optimisation of large scale finite element analysis models
    Ozturk, Umud Esat
    INTERNATIONAL JOURNAL OF VEHICLE DESIGN, 2017, 75 (1-4) : 139 - 153