Do large language models "understand" their knowledge?

被引:0
|
作者
Venkatasubramanian, Venkat [1 ]
机构
[1] Columbia Univ, Dept Chem Engn, Complex Resilient Intelligent Syst Lab, New York, NY 10027 USA
关键词
Knowledge representation; LLM; Industrial revolution 4.0; LKM; Transformers; PROCESS FAULT-DETECTION; QUANTITATIVE MODEL; PART I; FRAMEWORK; DESIGN; SYSTEM;
D O I
10.1002/aic.18661
中图分类号
TQ [化学工业];
学科分类号
0817 ;
摘要
Large language models (LLMs) are often criticized for lacking true "understanding" and the ability to "reason" with their knowledge, being seen merely as autocomplete engines. I suggest that this assessment might be missing a nuanced insight. LLMs do develop a kind of empirical "understanding" that is "geometry"-like, which is adequate for many applications. However, this "geometric" understanding, built from incomplete and noisy data, makes them unreliable, difficult to generalize, and lacking in inference capabilities and explanations. To overcome these limitations, LLMs should be integrated with an "algebraic" representation of knowledge that includes symbolic AI elements used in expert systems. This integration aims to create large knowledge models (LKMs) grounded in first principles that can reason and explain, mimicking human expert capabilities. Furthermore, we need a conceptual breakthrough, such as the transformation from Newtonian mechanics to statistical mechanics, to create a new science of LLMs.
引用
收藏
页数:10
相关论文
共 50 条
  • [31] Poisoning medical knowledge using large language models
    Yang, Junwei
    Xu, Hanwen
    Mirzoyan, Srbuhi
    Chen, Tong
    Liu, Zixuan
    Liu, Zequn
    Ju, Wei
    Liu, Luchen
    Xiao, Zhiping
    Zhang, Ming
    Wang, Sheng
    NATURE MACHINE INTELLIGENCE, 2024, 6 (10) : 1156 - 1168
  • [32] Large language models: What could they do for neurology?
    Lajoie, Guillaume
    JOURNAL OF THE NEUROLOGICAL SCIENCES, 2023, 455
  • [33] Do Large Language Models Bias Human Evaluations?
    O'Leary, Daniel E.
    IEEE INTELLIGENT SYSTEMS, 2024, 39 (04) : 83 - 87
  • [34] SKILL: Structured Knowledge Infusion for Large Language Models
    Moiseev, Fedor
    Dong, Zhe
    Alfonseca, Enrique
    Jaggi, Martin
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 1581 - 1588
  • [35] Detoxifying Large Language Models via Knowledge Editing
    Wang, Mengru
    Zhang, Ningyu
    Xu, Ziwen
    Xi, Zekun
    Deng, Shumin
    Yao, Yunzhi
    Zhang, Qishen
    Yang, Linyi
    Wang, Jindong
    Chen, Huajun
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 3093 - 3118
  • [36] Knowledge Graph Treatments for Hallucinating Large Language Models
    Collarana, Diego
    Busch, Moritz
    Lange, Christoph
    ERCIM NEWS, 2024, (136): : 35 - 36
  • [37] Unifying Large Language Models and Knowledge Graphs: A Roadmap
    Pan, Shirui
    Luo, Linhao
    Wang, Yufei
    Chen, Chen
    Wang, Jiapu
    Wu, Xindong
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (07) : 3580 - 3599
  • [38] Do Language Models Understand Morality? Towards a Robust Detection of Moral Content
    Bulla, Luana
    Gangemi, Aldo
    Mongiovi, Misael
    VALUE ENGINEERING IN ARTIFICIAL INTELLIGENCE, VALE 2023, 2024, 14520 : 98 - 113
  • [39] People Use their Knowledge of Common Events to Understand Language, and Do So as Quickly as Possible
    McRae, Ken
    Matsuki, Kazunaga
    LANGUAGE AND LINGUISTICS COMPASS, 2009, 3 (06): : 1417 - 1429
  • [40] DO YOU UNDERSTAND THE LANGUAGE OF BEHAVIOR
    WEISBORD, MR
    VOLTA REVIEW, 1964, 66 (08) : 610 - 614