A two-layer aggregation model with effective consistency for large-scale Gaussian process regression

被引:1
|
作者
Wang, Wengsheng [1 ]
Zhou, Changkai [1 ]
机构
[1] Hangzhou Dianzi Univ, Sch Econ, Hangzhou, Peoples R China
基金
中国国家自然科学基金;
关键词
Gaussian process; Big data; Aggregation models; Multi-layer structure;
D O I
10.1016/j.engappai.2021.104449
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
To scale full Gaussian process (GP) to large-scale data sets, aggregation models divide the dataset into independent subsets for factorized training, and then aggregate predictions from distributed experts. Some aggregation models have been able to produce consistent predictions which converge to the latent function when data size approaches infinity. However, these consistent predictions will become ineffective due to the limited subset size of experts. Oriented by the transition from theory to practice, the key idea is using Generalized Robust Bayesian Committee Machine (GRBCM) with corrective function to replace experts of Generalized Product of Experts (GPoE) which focuses on global information, in order to get rid of the limitation of the experts' size. Such a nested two-layer structure enables the proposed Generalized Product of Generalized Robust Bayesian Committee Machine (GPoGRBCM) to provide effective predictions on largescale datasets and to inherit virtues of aggregations, e.g., a slightly flawed Bayesian inference framework, distributed/parallel computing. Furthermore, we perform comparisons of GPoGRBCM against the state-of-theart aggregation models on one toy example and six real-world datasets with up to more than 3 million training points, showing dramatic performance improvement on scalability, capability, controllability, and robustness.
引用
收藏
页数:8
相关论文
共 50 条
  • [1] Model for the dynamics of the large-scale circulations in two-layer turbulent convection
    Sun, Yu
    Xie, Yi-Chao
    Xie, Jin-Xiao
    Zhong, Jin-Qiang
    Zhang, Jianwei
    Xia, Ke-Qing
    [J]. PHYSICAL REVIEW FLUIDS, 2024, 9 (03)
  • [2] Generalized Local Aggregation for Large Scale Gaussian Process Regression
    Gao, Yinghua
    Li, Naiqi
    Ding, Ning
    Li, Yiming
    Dai, Tao
    Xia, Shu-Tao
    [J]. 2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [3] Large-Scale Heteroscedastic Regression via Gaussian Process
    Liu, Haitao
    Ong, Yew-Soon
    Cai, Jianfei
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (02) : 708 - 721
  • [4] Asynchronous Parallel Large-Scale Gaussian Process Regression
    Dang, Zhiyuan
    Gu, Bin
    Deng, Cheng
    Huang, Heng
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (06) : 8683 - 8694
  • [5] Patchwork Kriging for Large-scale Gaussian Process Regression
    Park, Chiwoo
    Apley, Daniel
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2018, 19
  • [6] Two-Layer Gaussian Process Regression With Example Selection for Image Dehazing
    Fan, Xin
    Wang, Yi
    Tang, Xianxuan
    Gao, Renjie
    Luo, Zhongxuan
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2017, 27 (12) : 2505 - 2517
  • [7] Fast transfer Gaussian process regression with large-scale sources
    Da, Bingshui
    Ong, Yew-Soon
    Gupta, Abhishek
    Feng, Liang
    Liu, Haitao
    [J]. KNOWLEDGE-BASED SYSTEMS, 2019, 165 : 208 - 218
  • [8] TLGProb: Two-Layer Gaussian Process Regression Model for Winning Probability Calculation in Two-Team Sports
    Lam, Max W. Y.
    [J]. ARTIFICIAL INTELLIGENCE AND SOFT COMPUTING, ICAISC 2017, PT II, 2017, 10246 : 280 - 291
  • [9] Multiresolution Representations for Large-Scale Terrain with Local Gaussian Process Regression
    Liu, Xu
    Li, Decai
    He, Yuqing
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 5497 - 5503
  • [10] Two-layer hierarchical control for large-scale urban traffic networks
    Kouvelas, Anastasios
    Triantafyllos, Dimitris
    Geroliminis, Nikolas
    [J]. 2018 EUROPEAN CONTROL CONFERENCE (ECC), 2018, : 1295 - 1300