An overview of statistical learning theory

被引:4133
|
作者
Vapnik, VN [1 ]
机构
[1] AT&T Bell Labs, Res, Red Bank, NJ 07701 USA
来源
IEEE TRANSACTIONS ON NEURAL NETWORKS | 1999年 / 10卷 / 05期
关键词
D O I
10.1109/72.788640
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Statistical learning theory was introduced in the late 1960's, Until the 1990's it was a purely theoretical analysis of the problem of function estimation from a given collection of data. In the middle of the 1990's new types of learning algorithms (called support vector machines) based on the developed theory mere proposed, This made statistical learning theory not only a tool for the theoretical analysis hut also a tool for creating practical algorithms for estimating multidimensional functions. This article presents a very general overview of statistical learning theory including both theoretical and algorithmic aspects of the theory. The goal of this overview is to demonstrate how the abstract learning theory established conditions for generalization which are more general than those discussed in classical statistical paradigms and how the understanding of these conditions inspired new algorithmic approaches to function estimation problems. ih more detailed overview of the theory (without proofs) can be found in Vapnik (1995), In Vapnik (1998) one can find detailed description of the theory (including proofs).
引用
收藏
页码:988 / 999
页数:12
相关论文
共 50 条