Statistical learning theory: a tutorial

被引:32
|
作者
Kulkarni, Sanjeev R. [1 ]
Harman, Gilbert [2 ]
机构
[1] Princeton Univ, Dept Elect Engn, Princeton, NJ 08544 USA
[2] Princeton Univ, Dept Philosophy, Princeton, NJ 08544 USA
关键词
statistical learning; pattern recognition; classification; supervised learning; kernel methods;
D O I
10.1002/wics.179
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
In this article, we provide a tutorial overview of some aspects of statistical learning theory, which also goes by other names such as statistical pattern recognition, nonparametric classification and estimation, and supervised learning. We focus on the problem of two-class pattern classification for various reasons. This problem is rich enough to capture many of the interesting aspects that are present in the cases of more than two classes and in the problem of estimation, and many of the results can be extended to these cases. Focusing on two-class pattern classification simplifies our discussion, and yet it is directly applicable to a wide range of practical settings. We begin with a description of the two-class pattern recognition problem. We then discuss various classical and state-of-the-art approaches to this problem, with a focus on fundamental formulations, algorithms, and theoretical results. In particular, we describe nearest neighbor methods, kernel methods, multilayer perceptrons, Vapnik-Chervonenkis theory, support vector machines, and boosting. (C) 2011 JohnWiley& Sons, Inc.
引用
收藏
页码:543 / 556
页数:14
相关论文
共 50 条