Complete Statistical Theory of Learning (Learning Using Statistical Invariants)

被引:0
|
作者
Vapnik, Vladimir [1 ]
Izmailov, Rauf [2 ]
机构
[1] Columbia Univ, New York, NY 10025 USA
[2] Perspecta Labs, Basking Ridge, NJ USA
关键词
Learning Theory; Weak convergence; Statistical Invariants; Complete solution of learning problem; Reproducing Kernel Hilbert Space; Kernel Machines; Statistical Invariants for Support Vector Classification; Statistical Invariants for Support Vector Regression; Statistical Invariants for Neural Nets; Predicates; Symmetries Invariants;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Statistical theory of learning considers methods of constructing approximations that converge to the desired function with increasing number of observations. This theory studies mechanisms that provide convergence in the space of functions in L-2 norm, i.e., it studies the so-called strong mode of convergence. However, in Hilbert space, along with the convergence in the space of functions, there also exists the so-called weak mode of convergence, i.e., convergence in the space of functionals. Under some conditions, this weak mode of convergence also implies the convergence of approximations to the desired function in L-2 norm, although such convergence is based on other mechanisms. The paper discusses new learning methods which use both modes of convergence (weak and strong) simultaneously. Such methods allow one to execute the following: (1) select an admissible subset of functions (i.e., the set of appropriate approximation functions), and (2) find the desired approximation in this admissible subset. Since only two modes of convergence exist in Hilbert space, we call the theory that uses both modes the complete statistical theory of learning. Along with general reasoning, we describe new learning algorithms referred to as Learning Using Statistical Invariants (LUSI). LUSI algorithms were developed for sets of functions belonging to Reproducing Kernel Hilbert Space (RKHS); they include the modified SVM method (LUSI-SVM method). Also, the paper presents a LUSI modification of Neural Networks (LUSI-NN). LUSI methods require fewer training examples that standard approaches for achieving the same performance. In conclusion, the paper discusses the general (philosophical) framework of a new learning paradigm that includes the concept of intelligence.
引用
收藏
页码:4 / 40
页数:37
相关论文
共 50 条
  • [31] Statistical learning theory of structured data
    Pastore, Mauro
    Rotondo, Pietro
    Erba, Vittorio
    Gherardi, Marco
    PHYSICAL REVIEW E, 2020, 102 (03)
  • [32] A short review of statistical learning theory
    Pontil, M
    NEURAL NETS, 2002, 2486 : 233 - 242
  • [33] Statistical learning theory and stochastic optimization
    Catoni, Olivier
    STATISTICAL LEARNING THEORY AND STOCHASTIC OPTIMIZATION: ECOLE D'ETE DE PROBABILITIES DE SAINT-FLOUR XXXI 2001, 2004, 1851 : 1 - +
  • [34] On Lower Bounds for Statistical Learning Theory
    Loh, Po-Ling
    ENTROPY, 2017, 19 (11)
  • [35] Statistical Asymptotic Theory of Active Learning
    Takafumi Kanamori
    Annals of the Institute of Statistical Mathematics, 2002, 54 : 459 - 475
  • [37] A few notes on statistical learning theory
    Mendelson, S
    ADVANCED LECTURES ON MACHINE LEARNING, 2002, 2600 : 1 - 40
  • [38] On Adaptive Estimators in Statistical Learning Theory
    Konyagin, S. V.
    Livshits, E. D.
    PROCEEDINGS OF THE STEKLOV INSTITUTE OF MATHEMATICS, 2008, 260 (01) : 185 - 193
  • [39] Consumer purchasing behavior extraction using statistical learning theory
    Zuo, Yi
    Ali, A. B. M. Shawkat
    Yada, Katsutoshi
    KNOWLEDGE-BASED AND INTELLIGENT INFORMATION & ENGINEERING SYSTEMS 18TH ANNUAL CONFERENCE, KES-2014, 2014, 35 : 1464 - 1473
  • [40] Extrapolation to the complete basis-set limit in density-functional theory using statistical learning
    Speckhard, Daniel T.
    Carbogno, Christian
    Ghiringhelli, Luca M.
    Lubeck, Sven
    Scheffler, Matthias
    Draxl, Claudia
    Physical Review Materials, 9 (01):