Statistical asymptotic theory of active learning

被引:6
|
作者
Kanamori, T [1 ]
机构
[1] Tokyo Inst Technol, Dept Math & Comp Sci, Meguro Ku, Tokyo 1528552, Japan
关键词
active learning; Kullback-Leibler divergence; risk; optimal experimental design;
D O I
10.1023/A:1022446624428
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
We study a parametric estimation problem. Our aim is to estimate or to identify the conditional probability which is called the system. We suppose that we can select appropriate inputs to the system when we gather the training data. This kind of estimation is called active learning in the context of the artificial neural networks. In this paper we suggest new active learning algorithms and evaluate the risk of the algorithms by using statistical asymptotic theory. The algorithms are regarded as a version of the experimental design with two-stage sampling. We verify the efficiency of the active learning by simple computer simulations.
引用
收藏
页码:459 / 475
页数:17
相关论文
共 50 条