Unsupervised Model-Free Representation Learning

被引:0
|
作者
Ryabko, Daniil [1 ]
机构
[1] INRIA Lille, Lille, France
来源
关键词
PATTERN;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Numerous control and learning problems face the situation where sequences of high-dimensional highly dependent data are available, but no or little feedback is provided to the learner. In such situations it may be useful to find a concise representation of the input signal, that would preserve as much as possible of the relevant information. In this work we are interested in the problems where the relevant information is in the time-series dependence. Thus, the problem can be formalized as follows. Given a series of observations X-0 ,..., X-n coming from a large (high-dimensional) space chi, find a representation function f mapping chi to a finite space Y such that the series f(X-0) ,..., f(X-n) preserve as much information as possible about the original time-series dependence in X-0 ,..., X-n. For stationary time series, the function f can be selected as the one maximizing the time-series information I-infinity(f) = h(0)(f(X)) - h(infinity)(f(X)) where h0(f(X)) is the Shannon entropy of f(X-0) and h(infinity)(f(X)) is the entropy rate of the time series f(X-0) ,..., f(X-n),.... In this paper we study the functional I-infinity(f) from the learning-theoretic point of view. Specifically, we provide some uniform approximation results, and study the behaviour of I-infinity(f) in the problem of optimal control.
引用
收藏
页码:354 / 366
页数:13
相关论文
共 50 条
  • [41] Model-free execution monitoring by learning from simulation
    Pettersson, O
    Karlsson, L
    Saffiotti, A
    2005 IEEE INTERNATIONAL SYMPOSIUM ON COMPUTATIONAL INTELLIGENCE IN ROBOTICS AND AUTOMATION, PROCEEDINGS, 2005, : 505 - 511
  • [42] Covariance matrix adaptation for model-free reinforcement learning
    Adaptation de la matrice de covariance pour l'apprentissage par renforcement direct
    2013, Lavoisier, 14 rue de Provigny, Cachan Cedex, F-94236, France (27)
  • [43] Driving in Dense Traffic with Model-Free Reinforcement Learning
    Saxena, Dhruv Mauria
    Bae, Sangjae
    Nakhaei, Alireza
    Fujimura, Kikuo
    Likhachev, Maxim
    2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 5385 - 5392
  • [44] Model-Free Reinforcement Learning with Continuous Action in Practice
    Degris, Thomas
    Pilarski, Patrick M.
    Sutton, Richard S.
    2012 AMERICAN CONTROL CONFERENCE (ACC), 2012, : 2177 - 2182
  • [45] MODEL-FREE ONLINE REINFORCEMENT LEARNING OF A ROBOTIC MANIPULATOR
    Sweafford, Jerry, Jr.
    Fahimi, Farbod
    MECHATRONIC SYSTEMS AND CONTROL, 2019, 47 (03): : 136 - 143
  • [46] Practical aspects of the model-free learning control initialization
    Stebel, Krzysztof
    2015 20TH INTERNATIONAL CONFERENCE ON METHODS AND MODELS IN AUTOMATION AND ROBOTICS (MMAR), 2015, : 453 - 458
  • [47] Model-Free Learning, Dopamine and Alcoholism: Review and Test
    Huys, Quentin
    NEUROPSYCHOPHARMACOLOGY, 2016, 41 : S22 - S22
  • [48] Robotic Table Tennis with Model-Free Reinforcement Learning
    Gao, Wenbo
    Graesser, Laura
    Choromanski, Krzysztof
    Song, Xingyou
    Lazic, Nevena
    Sanketi, Pannag
    Sindhwani, Vikas
    Jaitly, Navdeep
    2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 5556 - 5563
  • [49] Modular Neural Networks for Model-Free Behavioral Learning
    Takeuchi, Johane
    Shouno, Osamu
    Tsujino, Hiroshi
    ARTIFICIAL NEURAL NETWORKS - ICANN 2008, PT I, 2008, 5163 : 730 - 739
  • [50] DANSE: Data-Driven Non-Linear State Estimation of Model-Free Process in Unsupervised Learning Setup
    Ghosh, Anubhab
    Honore, Antoine
    Chatterjee, Saikat
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2024, 72 : 1824 - 1838