Do large language models "understand" their knowledge?

被引:0
|
作者
Venkatasubramanian, Venkat [1 ]
机构
[1] Columbia Univ, Dept Chem Engn, Complex Resilient Intelligent Syst Lab, New York, NY 10027 USA
关键词
Knowledge representation; LLM; Industrial revolution 4.0; LKM; Transformers; PROCESS FAULT-DETECTION; QUANTITATIVE MODEL; PART I; FRAMEWORK; DESIGN; SYSTEM;
D O I
10.1002/aic.18661
中图分类号
TQ [化学工业];
学科分类号
0817 ;
摘要
Large language models (LLMs) are often criticized for lacking true "understanding" and the ability to "reason" with their knowledge, being seen merely as autocomplete engines. I suggest that this assessment might be missing a nuanced insight. LLMs do develop a kind of empirical "understanding" that is "geometry"-like, which is adequate for many applications. However, this "geometric" understanding, built from incomplete and noisy data, makes them unreliable, difficult to generalize, and lacking in inference capabilities and explanations. To overcome these limitations, LLMs should be integrated with an "algebraic" representation of knowledge that includes symbolic AI elements used in expert systems. This integration aims to create large knowledge models (LKMs) grounded in first principles that can reason and explain, mimicking human expert capabilities. Furthermore, we need a conceptual breakthrough, such as the transformation from Newtonian mechanics to statistical mechanics, to create a new science of LLMs.
引用
收藏
页数:10
相关论文
共 50 条
  • [21] Do Pretrained Language Models Indeed Understand Software Engineering Tasks?
    Li, Yao
    Zhang, Tao
    Luo, Xiapu
    Cai, Haipeng
    Fang, Sen
    Yuan, Dawei
    IEEE TRANSACTIONS ON SOFTWARE ENGINEERING, 2023, 49 (10) : 4639 - 4655
  • [22] Quo Vadis ChatGPT? From large language models to Large Knowledge Models
    Venkatasubramanian, Venkat
    Chakraborty, Arijit
    COMPUTERS & CHEMICAL ENGINEERING, 2025, 192
  • [23] Large Language Models to Understand Reasons for Anticoagulation Nonprescription in Atrial Fibrillation
    Somani, Sulaiman
    Kim, Dale
    Perez, Eduardo
    Ngo, Summer
    Hernandez-Boussard, Tina
    Rodriguez, Fatima
    CIRCULATION, 2024, 150
  • [24] From Sentence Embeddings to Large Language Models to Detect and Understand Wordplay
    Dsilva, Ryan Rony
    EXPERIMENTAL IR MEETS MULTILINGUALITY, MULTIMODALITY, AND INTERACTION, PT I, CLEF 2024, 2024, 14958 : 205 - 214
  • [25] Do Multimodal Large Language Models and Humans Ground Language Similarly?
    Jones, Cameron R.
    Bergen, Benjamin
    Trott, Sean
    COMPUTATIONAL LINGUISTICS, 2024, 50 (04) : 1415 - 1440
  • [26] Benchmarking Biomedical Relation Knowledge in Large Language Models
    Zhang, Fenghui
    Yang, Kuo
    Zhao, Chenqian
    Li, Haixu
    Dong, Xin
    Tian, Haoyu
    Zhou, Xuezhong
    BIOINFORMATICS RESEARCH AND APPLICATIONS, PT II, ISBRA 2024, 2024, 14955 : 482 - 495
  • [27] Updating knowledge in Large Language Models: an Empirical Evaluation
    Marinelli, Alberto Roberto
    Carta, Antonio
    Passaro, Lucia C.
    IEEE CONFERENCE ON EVOLVING AND ADAPTIVE INTELLIGENT SYSTEMS 2024, IEEE EAIS 2024, 2024, : 289 - 296
  • [28] ALCUNA: Large Language Models Meet New Knowledge
    Yin, Xunjian
    Huang, Baizhou
    Wan, Xiaojun
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 1397 - 1414
  • [29] Systematic Assessment of Factual Knowledge in Large Language Models
    Luo, Linhao
    Thuy-Trang Vu
    Phung, Dinh
    Haffari, Gholamreza
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 13272 - 13286
  • [30] Knowledge of cultural moral norms in large language models
    Ramezani, Aida
    Xu, Yang
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 428 - 446