Mutual adaptation: Learning from prototype for time-series prediction

被引:0
|
作者
Chen J. [1 ]
Shi X. [1 ,2 ]
Zhang H. [3 ]
Li W. [1 ]
Li P. [1 ]
Yao Y. [1 ]
Song X. [4 ]
Shibasaki R. [1 ]
机构
[1] Center for Spatial Information Science, University of Tokyo, Chiba
[2] School of Business, Society and Technology, Mälardalens University, Västeras
[3] LocationMind Inc., Chiyoda-ku
[4] SUSTech-UTokyo Joint Research Center on Super SmartCity, Department OfComputer Science AndEngineering, Southern University of Science and Technology, Shenzhen
来源
关键词
Knowledge based systems; Machine learning; Swarm systems; Temporal learning;
D O I
10.1109/TAI.2023.3282201
中图分类号
学科分类号
摘要
Time-series prediction is a current research hotspot in deep learning. However, due to the complex nature of time-series data, the modeling in this task is often highly nonconvex, which canmake the final convergence unstable. To address this challenge, recentworks have proposed deep mutual learning frameworks that allow models to learn from both ground truth and knowledge of othermodels in order to locate a better convergence point.However, a key disadvantage of deep mutual learning is that models that converge to poor local optima may still share their knowledge, limiting the overall performance. To overcome this limitation, in this article, we propose a new learning framework called mutual adaptation, which selects a prototypemodel that has the least error among all the models in the framework as the common teacher model. In addition, we incorporate a strategy of learning from each individual model's best local optimum in the history of training. Our experimental results show that, on average across multiple datasets, our method improves the performance of both Informer and long short-Term memory (LSTM) models compared to deep mutual learning by 4.73% in mean absolute error (MAE) and 6.99% in mean squared error (MSE) for Informer, and 11.54% in MAE and 18.15% in MSE for LSTM. We also demonstrate the importance of memory of individual best local optima and provide sensitivity analysis and visualization of error and the loss descending process. Our method represents a new state-of-The-Art in group learning for time-series prediction. Impact Statement-Deep mutual learning locates a more robust and better modelling of deep learning by merging different models for interactive learning. Every model learns fromboth groundtruth and knowledge of other models. However, its limitation is that models, which find bad local optima, would also share knowledge and probably lead others to worse performance. In this paper, we propose a framework named mutual adaptation. It selects a prototype model that has the least error from groundtruth among the models as the common teacher model of others. In addition, we recommend each model learn from individual's best output in the history of learning. In the experiment, among all of datasets, the performance of different models is improved by mutual adaptation and shows better performance than deep mutual learning. We believe our method could represents the advances in group deep learning. © 2023 IEEE.
引用
收藏
页码:1247 / 1262
页数:15
相关论文
共 50 条
  • [11] Learning Time-Series Shapelets
    Grabocka, Josif
    Schilling, Nicolas
    Wistuba, Martin
    Schmidt-Thieme, Lars
    PROCEEDINGS OF THE 20TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING (KDD'14), 2014, : 392 - 401
  • [12] Time-series information and learning
    Ryabko, Daniil
    2013 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY PROCEEDINGS (ISIT), 2013, : 1392 - 1395
  • [13] Time-series failure prediction on small datasets using machine learning
    Maior, Caio B. S.
    Silva, Thaylon G.
    IEEE LATIN AMERICA TRANSACTIONS, 2024, 22 (05) : 362 - 371
  • [14] Deep Learning for Time-Series Prediction in IIoT: Progress, Challenges, and Prospects
    Ren, Lei
    Jia, Zidi
    Laili, Yuanjun
    Huang, Di
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (11) : 15072 - 15091
  • [15] Double-Layered Cortical Learning Algorithm for Time-Series Prediction
    Aoki, Takeru
    Takadama, Keiki
    Sato, Hiroyuki
    BIO-INSPIRED INFORMATION AND COMMUNICATIONS TECHNOLOGIES, BICT 2021, 2021, 403 : 33 - 44
  • [16] Efficient learning of nonlinear prediction models with time-series privileged information
    Jung, Bastian
    Johansson, Fredrik D.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [17] A Time-Series Approach for Shock Outcome Prediction Using Machine Learning
    Shandilya, Sharad
    Ward, Kevin R.
    Najarian, Kayvan
    2010 IEEE INTERNATIONAL CONFERENCE ON BIOINFORMATICS AND BIOMEDICINE WORKSHOPS (BIBMW), 2010, : 440 - 446
  • [18] Time-series prediction of organomineral fertilizer moisture using machine learning
    Korkmaz, Cem
    Kacar, Ilyas
    APPLIED SOFT COMPUTING, 2024, 165
  • [19] Comparison of Neural-Network Learning Algorithms for Time-Series Prediction
    George, Koshy
    Harish, Madhumita
    Rao, Sneha
    Murali, Kruthi
    2017 INTERNATIONAL CONFERENCE ON ADVANCES IN COMPUTING, COMMUNICATIONS AND INFORMATICS (ICACCI), 2017, : 7 - 13
  • [20] Robust Unsupervised Feature Learning from Time-Series
    Miao, Jianyu
    Shi, Yong
    Niu, Lingfeng
    2016 IEEE/WIC/ACM INTERNATIONAL CONFERENCE ON WEB INTELLIGENCE WORKSHOPS (WIW 2016), 2016, : 37 - 40