An Information Theoretic Approach to Symbolic Learning in Synthetic Languages

被引:2
|
作者
Back, Andrew D. [1 ]
Wiles, Janet [1 ]
机构
[1] Univ Queensland, Sch Informat Technol & Elect Engn, Brisbane, Qld 4072, Australia
关键词
information theoretic models; synthetic language; entropy; Zipf-Mandelbrot-Li law; language models; behavior prediction; NUMERICAL COMPUTATION; MAXIMUM-LIKELIHOOD; ENTROPY; SPEECH; LAW; RECOGNITION; INTELLIGIBILITY; QUANTIZATION; DIVERSITY; INDEX;
D O I
10.3390/e24020259
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
An important aspect of using entropy-based models and proposed "synthetic languages", is the seemingly simple task of knowing how to identify the probabilistic symbols. If the system has discrete features, then this task may be trivial; however, for observed analog behaviors described by continuous values, this raises the question of how we should determine such symbols. This task of symbolization extends the concept of scalar and vector quantization to consider explicit linguistic properties. Unlike previous quantization algorithms where the aim is primarily data compression and fidelity, the goal in this case is to produce a symbolic output sequence which incorporates some linguistic properties and hence is useful in forming language-based models. Hence, in this paper, we present methods for symbolization which take into account such properties in the form of probabilistic constraints. In particular, we propose new symbolization algorithms which constrain the symbols to have a Zipf-Mandelbrot-Li distribution which approximates the behavior of language elements. We introduce a novel constrained EM algorithm which is shown to effectively learn to produce symbols which approximate a Zipfian distribution. We demonstrate the efficacy of the proposed approaches on some examples using real world data in different tasks, including the translation of animal behavior into a possible human language understandable equivalent.
引用
收藏
页数:25
相关论文
共 50 条
  • [1] Information-theoretic approach to interactive learning
    Still, S.
    EPL, 2009, 85 (02)
  • [2] Fairness in Supervised Learning: An Information Theoretic Approach
    Ghassami, AmirEmad
    Khodadadian, Sajad
    Kiyavash, Negar
    2018 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2018, : 176 - 180
  • [3] A Graph Theoretic Approach for Preference Learning with Feature Information
    Saha, Aadirupa
    Rajkumar, Arun
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2024, 244 : 3138 - 3158
  • [5] An Information Theoretic Approach to Learning Generative Graph Prototypes
    Han, Lin
    Hancock, Edwin R.
    Wilson, Richard C.
    SIMILARITY-BASED PATTERN RECOGNITION, 2011, 7005 : 133 - 148
  • [6] An Information Theoretic Approach to Learning Generative Graph Prototypes
    Han, Lin
    Hancock, Edwin R.
    Wilson, Richard C.
    SIMILARITY-BASED PATTERN RECOGNITION: FIRST INTERNATIONAL WORKSHOP, SIMBAD 2011, 2011, 7005 : 133 - 148
  • [7] An Information-Theoretic Approach for Multi-task Learning
    Yang, Pei
    Tan, Qi
    Xu, Hao
    Ding, Yehua
    ADVANCED DATA MINING AND APPLICATIONS, PROCEEDINGS, 2009, 5678 : 386 - 396
  • [8] Robust active noise control: An information theoretic learning approach
    Kurian, Nikhil Cherian
    Patel, Kashyap
    George, Nithin V.
    APPLIED ACOUSTICS, 2017, 117 : 180 - 184
  • [9] AN INFORMATION-THEORETIC APPROACH TO TRANSFERABILITY IN TASK TRANSFER LEARNING
    Bao, Yajie
    Li, Yang
    Huang, Shao-Lun
    Zhang, Lin
    Zheng, Lizhong
    Zamir, Amir
    Guibas, Leonidas
    2019 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2019, : 2309 - 2313
  • [10] An information-theoretic approach to feature extraction in competitive learning
    Kamimura, Ryotaro
    NEUROCOMPUTING, 2009, 72 (10-12) : 2693 - 2704