Replica theory for learning curves for Gaussian processes on random graphs

被引:3
|
作者
Urry, M. J. [1 ]
Sollich, P. [1 ]
机构
[1] Kings Coll London, Dept Math, London WC2R 2LS, England
关键词
PROCESS REGRESSION; BOUNDS; RANKING;
D O I
10.1088/1751-8113/45/42/425005
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
We use a statistical physics approach to derive accurate predictions for the challenging problem of predicting the performance of Gaussian process regression. Performance is quantified by the learning curve, defined as the average error versus number of training examples. We assume the Gaussian process prior is defined by a random walk kernel, inputs are vertices on a random graph and the outputs are noisy function values. We show that replica techniques can be used to obtain exact performance predictions in the limit of large graphs, after first rewriting the average error in terms of a graphical model. Conventionally, the Gaussian process kernel is only globally normalized, so that the prior variance can differ between vertices. As a more principled alternative we also consider local normalization, where the prior variance is uniform. The normalization constants for the prior then have to be defined as thermal averages in an unnormalized model and this requires the introduction of a second, auxiliary set of replicas. Our results for both types of kernel normalization apply generically to all random graph ensembles constrained by a fixed but arbitrary degree distribution. We compare with numerically simulated learning curves and find excellent agreement, a significant improvement over existing approximations.
引用
收藏
页数:28
相关论文
共 50 条
  • [1] Random Walk Kernels and Learning Curves for Gaussian Process Regression on Random Graphs
    Urry, Matthew J.
    Sollich, Peter
    JOURNAL OF MACHINE LEARNING RESEARCH, 2013, 14 : 1801 - 1835
  • [3] Learning curves for Gaussian processes
    Sollich, P
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 11, 1999, 11 : 344 - 350
  • [4] Approximate learning curves for Gaussian processes
    Sollich, P
    NINTH INTERNATIONAL CONFERENCE ON ARTIFICIAL NEURAL NETWORKS (ICANN99), VOLS 1 AND 2, 1999, (470): : 437 - 442
  • [5] Learning curves for LMS and regular Gaussian processes
    Hriljac, P
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2002, 47 (02) : 284 - 289
  • [6] Learning curves for Gaussian processes models: Fluctuations and universality
    Malzahn, D
    Opper, M
    ARTIFICIAL NEURAL NETWORKS-ICANN 2001, PROCEEDINGS, 2001, 2130 : 271 - 276
  • [7] Gaussian Processes on Graphs Via Spectral Kernel Learning
    Zhi, Yin-Cong
    Ng, Yin Cheng
    Dong, Xiaowen
    IEEE TRANSACTIONS ON SIGNAL AND INFORMATION PROCESSING OVER NETWORKS, 2023, 9 : 304 - 314
  • [8] GENERALIZED RANDOM PROCESSES - THEORY AND WHITE GAUSSIAN PROCESS
    INABA, H
    TAPLEY, BD
    SIAM JOURNAL ON CONTROL, 1975, 13 (03): : 719 - 735
  • [9] Replica methods for loopy sparse random graphs
    Coolen, A. C. C.
    INTERNATIONAL MEETING ON HIGH-DIMENSIONAL DATA-DRIVEN SCIENCE (HD3-2015), 2016, 699
  • [10] Learning curves for Gaussian processes regression: A framework for good approximations
    Malzahn, D
    Opper, M
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 13, 2001, 13 : 273 - 279