The Grammar-Learning Trajectories of Neural Language Models

被引:0
|
作者
Choshen, Leshem [1 ]
Hacohen, Guy [1 ,2 ]
Weinshall, Daphna [1 ]
Abend, Omri [1 ]
机构
[1] Hebrew Univ Jerusalem, Dept Comp Sci, Jerusalem, Israel
[2] Hebrew Univ Jerusalem, Dept Brain Sci, Jerusalem, Israel
基金
以色列科学基金会;
关键词
ACQUISITION; PERCEPTION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The learning trajectories of linguistic phenomena in humans provide insight into linguistic representation, beyond what can be gleaned from inspecting the behavior of an adult speaker. To apply a similar approach to analyze neural language models (NLM), it is first necessary to establish that different models are similar enough in the generalizations they make. In this paper, we show that NLMs with different initialization, architecture, and training data acquire linguistic phenomena in a similar order, despite their different end performance. These findings suggest that there is some mutual inductive bias that underlies these models' learning of linguistic phenomena. Taking inspiration from psycholinguistics, we argue that studying this inductive bias is an opportunity to study the linguistic representation implicit in NLMs. Leveraging these findings, we compare the relative performance on different phenomena at varying learning stages with simpler reference models. Results suggest that NLMs exhibit consistent "developmental" stages. Moreover, we find the learning trajectory to be approximately one-dimensional: given an NLM with a certain overall performance, it is possible to predict what linguistic generalizations it has already acquired. Initial analysis of these stages presents phenomena clusters (notably morphological ones), whose performance progresses in unison, suggesting a potential link between the generalizations behind them.
引用
收藏
页码:8281 / 8297
页数:17
相关论文
共 50 条
  • [1] Grammar Induction with Neural Language Models: An Unusual Replication
    Htut, Phu Mon
    Cho, Kyunghyun
    Bowman, Samuel R.
    2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 4998 - 5003
  • [2] Opposition logic and neural network models in artificial grammar learning
    Vokey, JR
    Higham, PA
    CONSCIOUSNESS AND COGNITION, 2004, 13 (03) : 565 - 578
  • [3] Neural network models of category learning and language
    Cangelosi, A
    BRAIN AND COGNITION, 2003, 53 (02) : 106 - 107
  • [4] Efficient transfer learning for neural network language models
    Skryzalin, Jacek
    Link, Hamilton
    Wendt, Jeremy
    Field, Richard
    Richter, Samuel N.
    2018 IEEE/ACM INTERNATIONAL CONFERENCE ON ADVANCES IN SOCIAL NETWORKS ANALYSIS AND MINING (ASONAM), 2018, : 897 - 902
  • [5] Models of artificial grammar learning
    Kinder, Annette
    INTERNATIONAL JOURNAL OF PSYCHOLOGY, 2008, 43 (3-4) : 583 - 583
  • [6] SCIENTIFIC GRAMMAR IN LANGUAGE-LEARNING
    EISENBERG, P
    ARGUMENT, 1976, 18 (JAN-F): : 9 - 23
  • [7] Grammar learning for spoken language understanding
    Wang, YY
    Acero, A
    ASRU 2001: IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING, CONFERENCE PROCEEDINGS, 2001, : 292 - 295
  • [8] Neural oscillatory mechanisms during novel grammar learning underlying language analytical abilities
    Kepinska, Olga
    Pereda, Ernesto
    Caspers, Johanneke
    Schiller, Niels O.
    BRAIN AND LANGUAGE, 2017, 175 : 99 - 110
  • [9] Transductive Learning of Neural Language Models for Syntactic and Semantic Analysis
    Ouchi, Hiroki
    Suzuki, Jun
    Inui, Kentaro
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 3665 - 3671
  • [10] Incrementally Learning the Hierarchical Softmax Function for Neural Language Models
    Peng, Hao
    Li, Jianxin
    Song, Yangqiu
    Liu, Yaopeng
    THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 3267 - 3273