Large Language Models Demonstrate the Potential of Statistical Learning in Language

被引:36
|
作者
Contreras Kallens, Pablo [1 ]
Kristensen-McLachlan, Ross Deans [2 ,3 ,4 ]
Christiansen, Morten H. [1 ,3 ,4 ,5 ,6 ]
机构
[1] Cornell Univ, Dept Psychol, Ithaca, NY USA
[2] Aarhus Univ, Ctr Humanities Comp, Aarhus, Denmark
[3] Aarhus Univ, Interacting Minds Ctr, Aarhus, Denmark
[4] Aarhus Univ, Sch Commun & Culture, Aarhus, Denmark
[5] Haskins Labs Inc, New Haven, CT USA
[6] Cornell Univ, Dept Psychol, 228 Uris Hall, Ithaca, NY 14853 USA
关键词
Large language models; Artificial intelligence; Language acquisition; Statistical learning; Grammar; Innateness; Linguistic experience; PRINCIPLES;
D O I
10.1111/cogs.13256
中图分类号
B84 [心理学];
学科分类号
04 ; 0402 ;
摘要
To what degree can language be acquired from linguistic input alone? This question has vexed scholars for millennia and is still a major focus of debate in the cognitive science of language. The complexity of human language has hampered progress because studies of language-especially those involving computational modeling-have only been able to deal with small fragments of our linguistic skills. We suggest that the most recent generation of Large Language Models (LLMs) might finally provide the computational tools to determine empirically how much of the human language ability can be acquired from linguistic experience. LLMs are sophisticated deep learning architectures trained on vast amounts of natural language data, enabling them to perform an impressive range of linguistic tasks. We argue that, despite their clear semantic and pragmatic limitations, LLMs have already demonstrated that human-like grammatical language can be acquired without the need for a built-in grammar. Thus, while there is still much to learn about how humans acquire and use language, LLMs provide full-fledged computational models for cognitive scientists to empirically evaluate just how far statistical learning might take us in explaining the full complexity of human language.
引用
收藏
页数:6
相关论文
共 50 条
  • [21] Learning Video Representations from Large Language Models
    Zhao, Yue
    Misra, Ishan
    Krahenbuhl, Philipp
    Girdhar, Rohit
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 6586 - 6597
  • [22] Forward Learning of Large Language Models by Consumer Devices
    Pau, Danilo Pietro
    Aymone, Fabrizio Maria
    ELECTRONICS, 2024, 13 (02)
  • [23] Large Language Models
    Vargas, Diego Collarana
    Katsamanis, Nassos
    ERCIM NEWS, 2024, (136): : 12 - 13
  • [24] Large Language Models
    Cerf, Vinton G.
    COMMUNICATIONS OF THE ACM, 2023, 66 (08) : 7 - 7
  • [25] Unveiling the potential of large language models in generating semantic and cross-language clones
    Roy, Palash R.
    Alam, Ajmain I.
    Al-omari, Farouq
    Roy, Banani
    Roy, Chanchal K.
    Schneider, Kevin A.
    2023 IEEE 17TH INTERNATIONAL WORKSHOP ON SOFTWARE CLONES, IWSC 2023, 2023, : 22 - 28
  • [26] Statistical learning of language: Theory, validity, and predictions of a statistical learning account of language acquisition
    Erickson, Lucy C.
    Thiessen, Erik D.
    DEVELOPMENTAL REVIEW, 2015, 37 : 66 - 108
  • [27] The Importance of Understanding Language in Large Language Models
    Youssef, Alaa
    Stein, Samantha
    Clapp, Justin
    Magnus, David
    AMERICAN JOURNAL OF BIOETHICS, 2023, 23 (10): : 6 - 7
  • [28] Dissociating language and thought in large language models
    Mahowald, Kyle
    Ivanova, Anna A.
    Blank, Idan A.
    Kanwisher, Nancy
    Tenenbaum, Joshua B.
    Fedorenko, Evelina
    TRENDS IN COGNITIVE SCIENCES, 2024, 28 (06) : 517 - 540
  • [29] Large Language Models in der WissenschaftLarge language models in science
    Karl-Friedrich Kowalewski
    Severin Rodler
    Die Urologie, 2024, 63 (9) : 860 - 866
  • [30] Large Language Models Are Zero-Shot Fuzzers: Fuzzing Deep-Learning Libraries via Large Language Models
    Deng, Yinlin
    Xia, Chunqiu Steven
    Peng, Haoran
    Yang, Chenyuan
    Zhan, Lingming
    PROCEEDINGS OF THE 32ND ACM SIGSOFT INTERNATIONAL SYMPOSIUM ON SOFTWARE TESTING AND ANALYSIS, ISSTA 2023, 2023, : 423 - 435