Hyperparameter self-tuning for data streams

被引:26
|
作者
Veloso, Bruno [1 ,2 ]
Gama, Joao [1 ,3 ]
Malheiro, Benedita [1 ,4 ]
Vinagre, Joao [1 ,5 ]
机构
[1] INESC TEC, Porto, Portugal
[2] Univ Portucalense, Porto, Portugal
[3] FEP Univ Porto, Porto, Portugal
[4] ISEP Polytech Inst Porto, Porto, Portugal
[5] FCUP Univ Porto, Porto, Portugal
关键词
Data Streams; Optimisation; Hyperparameters;
D O I
10.1016/j.inffus.2021.04.011
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The number of Internet of Things devices generating data streams is expected to grow exponentially with the support of emergent technologies such as 5G networks. Therefore, the online processing of these data streams requires the design and development of suitable machine learning algorithms, able to learn online, as data is generated. Like their batch-learning counterparts, stream-based learning algorithms require careful hyperparameter settings. However, this problem is exacerbated in online learning settings, especially with the occurrence of concept drifts, which frequently require the reconfiguration of hyperparameters. In this article, we present SSPT, an extension of the Self Parameter Tuning (SPT) optimisation algorithm for data streams. We apply the Nelder-Mead algorithm to dynamically-sized samples, converging to optimal settings in a single pass over data while using a relatively small number of hyperparameter configurations. In addition, our proposal automatically readjusts hyperparameters when concept drift occurs. To assess the effectiveness of SSPT, the algorithm is evaluated with three different machine learning problems: recommendation, regression, and classification. Experiments with well-known data sets show that the proposed algorithm can outperform previous hyperparameter tuning efforts by human experts. Results also show that SSPT converges significantly faster and presents at least similar accuracy when compared with the previous double-pass version of the SPT algorithm.
引用
收藏
页码:75 / 86
页数:12
相关论文
共 50 条
  • [1] Self-Tuning, Bandwidth-Aware Monitoring for Dynamic Data Streams
    Jain, Navendu
    Yalagandula, Praveen
    Dahlin, Mike
    Zhang, Yin
    ICDE: 2009 IEEE 25TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING, VOLS 1-3, 2009, : 114 - 125
  • [2] A Self-Tuning Fuzzy Rule-Based Classifier for Data Streams
    Shahparast, Homeira
    Hamzeloo, Sam
    Jahromi, Mansoor Zolghadri
    INTERNATIONAL JOURNAL OF UNCERTAINTY FUZZINESS AND KNOWLEDGE-BASED SYSTEMS, 2014, 22 (02) : 293 - 303
  • [3] Improving hyper-parameter self-tuning for data streams by adapting an evolutionary approach
    Moya, Antonio R.
    Veloso, Bruno
    Gama, Joao
    Ventura, Sebastian
    DATA MINING AND KNOWLEDGE DISCOVERY, 2024, 38 (03) : 1289 - 1315
  • [4] Improving hyper-parameter self-tuning for data streams by adapting an evolutionary approach
    Antonio R. Moya
    Bruno Veloso
    João Gama
    Sebastián Ventura
    Data Mining and Knowledge Discovery, 2024, 38 : 1289 - 1315
  • [5] A Meta-Learning Approach for Automated Hyperparameter Tuning in Evolving Data Streams
    Lacombe, Thomas
    Koh, Yun Sing
    Dobbie, Gillian
    Wu, Ocean
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [6] Self-tuning and conformality
    Kakushadze, Z
    MODERN PHYSICS LETTERS A, 2000, 15 (30) : 1879 - 1890
  • [7] Lightweight, self-tuning data dissemination for dense nanonetworks
    Tsioliaridou, A.
    Liaskos, C.
    Ioannidis, S.
    Pitsillides, A.
    NANO COMMUNICATION NETWORKS, 2016, 8 : 2 - 15
  • [8] Self-tuning Eventually-Consistent Data Stores
    Chatterjee, Shankha
    Golab, Wojciech
    STABILIZATION, SAFETY, AND SECURITY OF DISTRIBUTED SYSTEMS, SSS 2017, 2018, 10616 : 78 - 92
  • [9] Self-tuning clustering for high-dimensional data
    Wen, Guoqiu
    Zhu, Yonghua
    Cai, Zhiguo
    Zheng, Wei
    WORLD WIDE WEB-INTERNET AND WEB INFORMATION SYSTEMS, 2018, 21 (06): : 1563 - 1573
  • [10] Self-Tuning for Data-Efficient Deep Learning
    Wang, Ximei
    Gao, Jinghan
    Long, Mingsheng
    Wang, Jianmin
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139 : 7748 - 7759