A decision-theoretic extension of stochastic complexity and its applications to learning

被引:52
|
作者
Yamanishi, K [1 ]
机构
[1] NEC Res Inst, Princeton, NJ 08540 USA
关键词
aggregating strategy; batch-learning; complexity regularization; extended stochastic complexity; MDL principle; on-line prediction; stochastic complexity;
D O I
10.1109/18.681319
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Rissanen has introduced stochastic complexity to define the amount of information in a given data sequence relative to a given hypothesis class of probability densities, where the information is measured in terms of the logarithmic loss associated with universal data compression. This paper introduces the notion of extended stochastic complexity (ESC) and demonstrates its effectiveness in design and analysis of learning algorithms in on-line prediction and batch-learning scenarios. ESC can be thought of as an extension of Rissanen's stochastic complexity to the decision-theoretic setting where a general real-valued function is used as a hypothesis and a general loss function is used as a distortion measure. As an application of ESC to online prediction, this paper shows that a sequential realization of ESC produces an on-line prediction algorithm called Vovk's aggregating strategy, which can be thought of as an extension of the Bayes algorithm. We derive upper bounds on the cumulative loss for the aggregating strategy both of an expected form and a worst case form in the case where the hypothesis class is continuous. As an application of ESC to batch-learning, this paper shows that a batch-approximation of ESC induces a batch-learning algorithm called the minimum L-complexity algorithm (MLC), which is an extension of the minimum description length (MDL) principle. We derive upper bounds on the statistical risk for MLC, which are least to date. Through ESC we give a unifying view of the most effective learning algorithms that have recently been explored in computational learning theory.
引用
收藏
页码:1424 / 1439
页数:16
相关论文
共 50 条
  • [31] Decision-theoretic perspective on racial mistrust
    Krueger, Joachim I.
    PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2011, 108 (32) : E410 - E410
  • [32] DTPROBLOG: A Decision-Theoretic Probabilistic Prolog
    Van den Broeck, Guy
    Thon, Ingo
    van Otterlo, Martijn
    De Raedt, Luc
    PROCEEDINGS OF THE TWENTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE (AAAI-10), 2010, : 1217 - 1222
  • [33] Decision-theoretic cooperative sensor planning
    Cook, DJ
    Gmytrasiewicz, P
    Holder, LB
    IMAGE UNDERSTANDING WORKSHOP, 1996 PROCEEDINGS, VOLS I AND II, 1996, : 1321 - 1332
  • [34] Multigranulation decision-theoretic rough sets
    Qian, Yuhua
    Zhang, Hu
    Sang, Yanli
    Liang, Jiye
    INTERNATIONAL JOURNAL OF APPROXIMATE REASONING, 2014, 55 (01) : 225 - 237
  • [35] A decision-theoretic approach to data mining
    Elovici, Y
    Braha, D
    IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART A-SYSTEMS AND HUMANS, 2003, 33 (01): : 42 - 51
  • [36] Decision-theoretic control of planetary rovers
    Zilberstein, S
    Washington, R
    Bernstein, DS
    Mouaddib, AI
    ADVANCES IN PLAN-BASED CONTROL OF ROBOTIC AGENTS, 2002, 2466 : 270 - 289
  • [37] The propensity theory: a decision-theoretic restatement
    Albert, M.
    SYNTHESE, 2007, 156 (03) : 587 - 603
  • [38] Decision-theoretic reflections on processing a fingermark
    Gittelson, S.
    Bozza, S.
    Biedermann, A.
    Taroni, F.
    FORENSIC SCIENCE INTERNATIONAL, 2013, 226 (1-3) : E42 - E47
  • [39] Decision-theoretic cooperative sensor planning
    Univ of Texas at Arlington, Arlington, United States
    IEEE Trans Pattern Anal Mach Intell, 10 (1013-1023):
  • [40] Modeling and Reasoning with Decision-Theoretic Goals
    Liaskos, Sotirios
    Khan, Shakil M.
    Soutchanski, Mikhail
    Mylopoulos, John
    CONCEPTUAL MODELING, ER 2013, 2013, 8217 : 19 - +