A novel approach to measuring the scope of patent claims based on probabilities obtained from (large) language models

被引:0
|
作者
Ragot, Sebastien [1 ]
机构
[1] E Blum & Co Ltd, Patent & Trademark Attorneys VSP, Vorderberg 11, CH-8044 Zurich, Switzerland
关键词
Patent scope; Patent value; Patent claims; Language models; Large language models; GPT; Information theory; Self-information;
D O I
10.1016/j.wpi.2024.102321
中图分类号
G25 [图书馆学、图书馆事业]; G35 [情报学、情报工作];
学科分类号
1205 ; 120501 ;
摘要
This work proposes to measure the scope of a patent claim as the reciprocal of self-information contained in this claim. Self-information is calculated based on a probability of occurrence of the claim, where this probability is obtained from a language model. Grounded in information theory, this approach is based on the assumption that an unlikely concept is more informative than a usual concept, insofar as it is more surprising. In turn, the more surprising the information required to define the claim, the narrower its scope. Seven language models are considered, ranging from simplest models (each word or character has an identical probability) to intermediate models (based on average word or character frequencies), to large language models (LLMs) such as GPT2 and davinci-002. Remarkably, when using the simplest language models to compute the probabilities, the scope becomes proportional to the reciprocal of the number of words or characters involved in the claim, a metric already used in previous works. Application is made to multiple series of patent claims directed to distinct inventions, where each series consists of a set of claims having a gradually decreasing scope. The performance of the language models is then assessed through several ad hoc tests. The LLMs outperform models based on word and character frequencies, which themselves outdo the simplest models based on word or character counts. Interestingly, however, the character count appears to be a more reliable indicator than the word count.
引用
收藏
页数:29
相关论文
共 50 条
  • [41] Explainable cognitive decline detection in free dialogues with a Machine Learning approach based on pre-trained Large Language Models
    de Arriba-Perez, Francisco
    Garcia-Mendez, Silvia
    Otero-Mosquera, Javier
    Gonzalez-Castano, Francisco J.
    APPLIED INTELLIGENCE, 2024, : 12613 - 12628
  • [42] Learning Pattern-Based Extractors from Natural Language and Knowledge Graphs: Applying Large Language Models to Wikipedia and Linked Open Data
    Ringwald, Celian
    THIRTY-EIGTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 21, 2024, : 23411 - 23412
  • [43] Using Large Language Models to Detect Depression From User-Generated Diary Text Data as a Novel Approach in Digital Mental Health Screening: Instrument Validation Study
    Shin, Daun
    Kim, Hyoseung
    Lee, Seunghwan
    Cho, Younhee
    Jung, Whanbo
    JOURNAL OF MEDICAL INTERNET RESEARCH, 2024, 26
  • [44] An AI approach to unraveling treatment response in pancreatic cancer: Insights from the COMPASS trial leveraging large language models (LLMs)
    Geraci, Joseph
    Qorri, Bessi
    Tsay, Mike
    Cumbaa, Christian
    Leonchyk, Paul
    Alphs, Larry
    Pani, Luca
    CANCER RESEARCH, 2024, 84 (17)
  • [45] A Novel Approach to Measuring Urban Waterlogging Depth from Images Based on Mask Region-Based Convolutional Neural Network
    Huang, Jing
    Kang, Jinle
    Wang, Huimin
    Wang, Zhigiang
    Qiu, Tian
    SUSTAINABILITY, 2020, 12 (05)
  • [46] Prediction of water saturation in tight sandstone reservoirs from well log data based on the large language models (LLMs)
    Wu, Juan
    Luo, Renze
    Lei, Canru
    Yin, Jiang
    Chen, Xingting
    Natural Gas Industry, 44 (09): : 77 - 87
  • [47] Harnessing prompt-based large language models for disaster monitoring and automated reporting from social media feedback
    Cantini, Riccardo
    Cosentino, Cristian
    Marozzo, Fabrizio
    Talia, Domenico
    Trunfio, Paolo
    ONLINE SOCIAL NETWORKS AND MEDIA, 2025, 45
  • [48] AdaShield: Safeguarding Multimodal Large Language Models from Structure-Based Attack via Adaptive Shield Prompting
    Wang, Yu
    Liu, Xiaogeng
    Li, Yu
    Chen, Muhao
    Xiao, Chaowei
    COMPUTER VISION - ECCV 2024, PT XX, 2025, 15078 : 77 - 94
  • [49] Enhancing Skeleton-Based Action Recognition With Language Descriptions From Pre-Trained Large Multimodal Models
    He, Tian
    Chen, Yang
    Gao, Xu
    Wang, Ling
    Hu, Ting
    Cheng, Hong
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2025, 35 (03) : 2118 - 2132
  • [50] NLP4ReF: Requirements Classification and Forecasting: From Model-Based Design to Large Language Models
    Peer, Jordan
    Mordecai, Yaniv
    Reich, Yoram
    2024 IEEE AEROSPACE CONFERENCE, 2024,