Shared computational principles for language processing in humans and deep language models

被引:157
|
作者
Goldstein, Ariel [1 ,2 ,3 ]
Zada, Zaid [1 ,2 ]
Buchnik, Eliav [3 ]
Schain, Mariano [3 ]
Price, Amy [1 ,2 ]
Aubrey, Bobbi [1 ,2 ,4 ]
Nastase, Samuel A. [1 ,2 ]
Feder, Amir [3 ]
Emanuel, Dotan [3 ]
Cohen, Alon [3 ]
Jansen, Aren [3 ]
Gazula, Harshvardhan [1 ,2 ]
Choe, Gina [1 ,2 ,4 ]
Rao, Aditi [1 ,2 ,4 ]
Kim, Catherine [1 ,2 ,4 ]
Casto, Colton [1 ,2 ]
Fanda, Lora [4 ]
Doyle, Werner [4 ]
Friedman, Daniel [4 ]
Dugan, Patricia [4 ]
Melloni, Lucia [5 ]
Reichart, Roi [6 ]
Devore, Sasha [4 ]
Flinker, Adeen [4 ]
Hasenfratz, Liat [1 ,2 ]
Levy, Omer [7 ]
Hassidim, Avinatan [3 ]
Brenner, Michael [3 ,8 ]
Matias, Yossi [3 ]
Norman, Kenneth A. [1 ,2 ]
Devinsky, Orrin [4 ]
Hasson, Uri [1 ,2 ,3 ]
机构
[1] Princeton Univ, Dept Psychol, Princeton, NJ 08544 USA
[2] Princeton Univ, Neurosci Inst, Princeton, NJ 08544 USA
[3] Google Res, Mountain View, CA 94043 USA
[4] NYU, Grossman Sch Med, New York, NY USA
[5] Max Planck Inst Empir Aesthet, Frankfurt, Germany
[6] Technion Israel Inst Technol, Fac Ind Engn & Management, Haifa, Israel
[7] Tel Aviv Univ, Blavatnik Sch Comp Sci, Tel Aviv, Israel
[8] Harvard Univ, Sch Engn & Appl Sci, Cambridge, MA 02138 USA
基金
美国国家卫生研究院;
关键词
WORDS; TIMESCALES; COMPONENT; FUTURE;
D O I
10.1038/s41593-022-01026-4
中图分类号
Q189 [神经科学];
学科分类号
071006 ;
摘要
Departing from traditional linguistic models, advances in deep learning have resulted in a new type of predictive (autoregressive) deep language models (DLMs). Using a self-supervised next-word prediction task, these models generate appropriate linguistic responses in a given context. In the current study, nine participants listened to a 30-min podcast while their brain responses were recorded using electrocorticography (ECoG). We provide empirical evidence that the human brain and autoregressive DLMs share three fundamental computational principles as they process the same natural narrative: (1) both are engaged in continuous next-word prediction before word onset; (2) both match their pre-onset predictions to the incoming word to calculate post-onset surprise; (3) both rely on contextual embeddings to represent words in natural contexts. Together, our findings suggest that autoregressive DLMs provide a new and biologically feasible computational framework for studying the neural basis of language. Deep language models have revolutionized natural language processing. The paper discovers three computational principles shared between deep language models and the human brain, which can transform our understanding of the neural basis of language.
引用
收藏
页码:369 / +
页数:28
相关论文
共 50 条
  • [31] Models and modularity in language processing
    Dell, GS
    PROCEEDINGS OF THE EIGHTEENTH ANNUAL CONFERENCE OF THE COGNITIVE SCIENCE SOCIETY, 1996, : 47 - 47
  • [32] Neurocomputational Models of Language Processing
    Hale, John T.
    Campanelli, Luca
    Li, Jixing
    Bhattasali, Shohini
    Pallier, Christophe
    Brennan, Jonathan R.
    ANNUAL REVIEW OF LINGUISTICS, 2022, 8 : 427 - 446
  • [33] Natural Language Processing and Deep Learning for Blended Learning as an Aspect of Computational Linguistics
    Pikhart, Marcel
    CROSS REALITY AND DATA SCIENCE IN ENGINEERING, 2021, 1231 : 842 - 848
  • [34] Speech and language processing: An introduction to natural language processing, computational linguistics, and speech recognition
    Teller, V
    COMPUTATIONAL LINGUISTICS, 2000, 26 (04) : 638 - 641
  • [35] Fast compression and optimization of deep learning models for natural language processing
    Pietron, Marcin
    Karwatowski, Michal
    Wielgosz, Maciej
    Duda, Jerzy
    2019 SEVENTH INTERNATIONAL SYMPOSIUM ON COMPUTING AND NETWORKING WORKSHOPS (CANDARW 2019), 2019, : 162 - 168
  • [36] An introduction to Deep Learning in Natural Language Processing: Models, techniques, and tools
    Lauriola, Ivano
    Lavelli, Alberto
    Aiolli, Fabio
    NEUROCOMPUTING, 2022, 470 : 443 - 456
  • [37] Principles and Interactive Tools for Evaluating and Improving the Behavior of Natural Language Processing models
    Wu, Tongshuang
    EXTENDED ABSTRACTS OF THE 2021 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS (CHI'21), 2021,
  • [38] Development of language resources for natural language processing in deep level
    Zhang, Yujie
    Kuroda, Kow
    Izumi, Emi
    Nozawa, Hajime
    Journal of the National Institute of Information and Communications Technology, 2007, 54 (03): : 53 - 61
  • [40] COMPUTATIONAL MODELS OF NATURAL-LANGUAGE PROCESSING - BARA,BG, GUIDA,G
    REYNOLDS, PC
    COMPUTERS AND THE HUMANITIES, 1988, 22 (04): : 319 - 320