Epoch-Evolving Gaussian Process Guided Learning for Classification

被引:0
|
作者
Cui, Jiabao [1 ]
Li, Xuewei [1 ]
Zhao, Hanbin [1 ]
Wang, Hui [1 ]
Li, Bin [2 ]
Li, Xi [3 ,4 ,5 ,6 ]
机构
[1] Zhejiang Univ, Coll Comp Sci & Technol, Hangzhou 310027, Peoples R China
[2] Zhejiang Univ, Coll Informat Sci & Elect Engn, Hangzhou 310027, Peoples R China
[3] Zhejiang Univ, Coll Comp Sci & Technol, Hangzhou 311121, Peoples R China
[4] Shanghai AI Lab, Shanghai 200000, Peoples R China
[5] Zhejiang Univ, Shanghai Inst Adv Study, Shanghai 200000, Peoples R China
[6] Zhejiang Singapore Innovat & AI Joint Res Lab, Hangzhou 311121, Peoples R China
基金
中国国家自然科学基金;
关键词
Computational modeling; Pipelines; Deep learning; Context modeling; Predictive models; Feature extraction; Data models; Gaussian process (GP); global distribution-aware learning; non-parametric modeling; top-down strategy; PROCESS REGRESSION; NEURAL-NETWORKS;
D O I
10.1109/TNNLS.2022.3174207
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The conventional mini-batch gradient descent algorithms are usually trapped in the local batch-level distribution information, resulting in the ``zig-zag'' effect in the learning process. To characterize the correlation information between the batch-level distribution and the global data distribution, we propose a novel learning scheme called epoch-evolving Gaussian process guided learning (GPGL) to encode the global data distribution information in a non-parametric way. Upon a set of class-aware anchor samples, our GP model is built to estimate the class distribution for each sample in mini-batch through label propagation from the anchor samples to the batch samples. The class distribution, also named the context label, is provided as a complement for the ground-truth one-hot label. Such a class distribution structure has a smooth property and usually carries a rich body of contextual information that is capable of speeding up the convergence process. With the guidance of the context label and ground-truth label, the GPGL scheme provides a more efficient optimization through updating the model parameters with a triangle consistency loss. Furthermore, our GPGL scheme can be generalized and naturally applied to the current deep models, outperforming the state-of-the-art optimization methods on six benchmark datasets.
引用
收藏
页码:326 / 337
页数:12
相关论文
共 50 条
  • [21] Approximations for Binary Gaussian Process Classification
    Nickisch, Hannes
    Rasmussen, Carl Edward
    JOURNAL OF MACHINE LEARNING RESEARCH, 2008, 9 : 2035 - 2078
  • [22] A new Gaussian process classification algorithm
    Zhang, Jun-Xing, 1600, Northeast University (29):
  • [23] Scalable Variational Gaussian Process Classification
    Hensman, James
    Matthews, Alex G. de G.
    Ghahramani, Zoubin
    ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 38, 2015, 38 : 351 - 360
  • [24] SCALABLE GAUSSIAN PROCESS FOR EXTREME CLASSIFICATION
    Dhaka, Akash Kumar
    Andersen, Michael Riis
    Moreno, Pablo Garcia
    Vehtari, Aki
    PROCEEDINGS OF THE 2020 IEEE 30TH INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING (MLSP), 2020,
  • [25] Efficient approaches to Gaussian Process classification
    Csató, L
    Fokoué, E
    Opper, M
    Schottky, B
    Winther, O
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 12, 2000, 12 : 251 - 257
  • [26] 21-cm signal from the Epoch of Reionization: a machine learning upgrade to foreground removal with Gaussian process regression
    Acharya, Anshuman
    Mertens, Florent
    Ciardi, Benedetta
    Ghara, Raghunath
    Koopmans, Leon V. E.
    Giri, Sambit K.
    Hothi, Ian
    Ma, Qing-Bo
    Mellema, Garrelt
    Munshi, Satyapan
    MONTHLY NOTICES OF THE ROYAL ASTRONOMICAL SOCIETY, 2024, 527 (03) : 7835 - 7846
  • [27] 21-cm signal from the Epoch of Reionization: a machine learning upgrade to foreground removal with Gaussian process regression
    Acharya, Anshuman
    Mertens, Florent
    Ciardi, Benedetta
    Ghara, Raghunath
    Koopmans, Leon V. E.
    Giri, Sambit K.
    Hothi, Ian
    Ma, Qing-Bo
    Mellema, Garrelt
    Munshi, Satyapan
    MONTHLY NOTICES OF THE ROYAL ASTRONOMICAL SOCIETY, 2023, 527 (03) : 7835 - 7846
  • [28] Stacked Gaussian Process Learning
    Neumann, Marion
    Kersting, Kristian
    Xu, Zhao
    Schulz, Daniel
    2009 9TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING, 2009, : 387 - +
  • [29] Learning Rates for Classification with Gaussian Kernels
    Lin, Shao-Bo
    Zeng, Jinshan
    Chang, Xiangyu
    NEURAL COMPUTATION, 2017, 29 (12) : 3353 - 3380
  • [30] Transfer learning-based Gaussian process classification for lattice structure damage detection
    Yang, Xin
    Farrokhabadi, Amin
    Rauf, Ali
    Liu, Yongcheng
    Talemi, Reza
    Kundu, Pradeep
    Chronopoulos, Dimitrios
    MEASUREMENT, 2024, 238