Incremental k-Nearest Neighbors Using Reservoir Sampling for Data Streams

被引:1
|
作者
Bahri, Maroua [1 ]
Bifet, Albert [1 ,2 ]
机构
[1] IP Paris, LTCI, Telecom Paris, Paris, France
[2] Univ Waikato, Hamilton, New Zealand
来源
DISCOVERY SCIENCE (DS 2021) | 2021年 / 12986卷
关键词
Data stream classification; K-nearest neighbors; Reservoir sampling; Sliding window;
D O I
10.1007/978-3-030-88942-5_10
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The online and potentially infinite nature of data streams leads to the inability to store the flow in its entirety and thus restricts the storage to a part of - and/or synopsis information from - the stream. To process these evolving data, we need efficient and accurate methodologies and systems, such as window models (e.g., sliding windows) and summarization techniques (e.g., sampling, sketching, dimensionality reduction). In this paper, we propose, RW-kNN, a k-Nearest Neighbors (kNN) algorithm that employs a practical way to store information about past instances using the biased reservoir sampling to sample the input instances along with a sliding window to maintain the most recent instances from the stream. We evaluate our proposal on a diverse set of synthetic and real datasets and compare against state-of-the-art algorithms in a traditional test-then-train evaluation. Results show how our proposed RW-kNN approach produces high-predictive performance for both real and synthetic datasets while using a feasible amount of resources.
引用
收藏
页码:122 / 137
页数:16
相关论文
共 50 条
  • [1] Compressed k-Nearest Neighbors Ensembles for Evolving Data Streams
    Bahri, Maroua
    Bifet, Albert
    Maniu, Silviu
    de Mello, Rodrigo F.
    Tziortziotis, Nikolaos
    ECAI 2020: 24TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, 325 : 961 - 968
  • [2] Adaptive Global k-Nearest Neighbors for Hierarchical Classification of Data Streams
    Tieppo, Eduardo
    Barddal, Jean Paul
    Nievola, Julio Cesar
    2021 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2021, : 631 - 636
  • [3] Forecasting Earnings Using k-Nearest Neighbors
    Easton, Peter D.
    Kapons, Martin M.
    Monahan, Steven J.
    Schutt, Harm H.
    Weisbrod, Eric H.
    ACCOUNTING REVIEW, 2024, 99 (03): : 115 - 140
  • [4] K-Nearest Neighbors Hashing
    He, Xiangyu
    Wang, Peisong
    Cheng, Jian
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 2834 - 2843
  • [5] Modernizing k-nearest neighbors
    Elizabeth Yancey, Robin
    Xin, Bochao
    Matloff, Norm
    STAT, 2021, 10 (01):
  • [6] Compressed kNN: K-Nearest Neighbors with Data Compression
    Salvador-Meneses, Jaime
    Ruiz-Chavez, Zoila
    Garcia-Rodriguez, Jose
    ENTROPY, 2019, 21 (03)
  • [7] Consistency of the k-nearest neighbors rule for functional data
    Younso, Ahmad
    COMPTES RENDUS MATHEMATIQUE, 2023, 361 (01) : 237 - 242
  • [8] A new k-nearest neighbors classifier for functional data
    Zhu, Tianming
    Zhang, Jin-ting
    STATISTICS AND ITS INTERFACE, 2022, 15 (02) : 247 - 260
  • [9] Particles Contaminations Detection during Plasma Etching Process by using k-Nearest Neighbors and Fuzzy k-Nearest Neighbors
    Somari, Noratika Mohammad
    Abdullah, Mohd Firdaus
    Osman, Muhammad Khusairi
    Nazelan, Abdul Mu'iz
    Ahmad, Khairul Azman
    Appanan, Sooria Pragash Rao S.
    Hooi, Loh Kwang
    2016 6TH IEEE INTERNATIONAL CONFERENCE ON CONTROL SYSTEM, COMPUTING AND ENGINEERING (ICCSCE), 2016, : 512 - 516
  • [10] k-nearest neighbors prediction and classification for spatial data
    Mohamed-Salem Ahmed
    Mamadou N’diaye
    Mohammed Kadi Attouch
    Sophie Dabo-Niange
    Journal of Spatial Econometrics, 2023, 4 (1):