A Large-Scale Ensemble Learning Framework for Demand Forecasting

被引:1
|
作者
Park, Young-Jin [1 ]
Kim, Donghyun [2 ]
Odermatt, Frederic [3 ]
Lee, Juho [4 ]
Kim, Kyung-Min [5 ,6 ]
机构
[1] MIT, Cambridge, MA 02139 USA
[2] Seoul Natl Univ, Seoul, South Korea
[3] ETH Z urich, Zurich, Switzerland
[4] Superpetual Inc, Seoul, South Korea
[5] NAVER CLOVA, Seongnam, South Korea
[6] NAVER Lab, Seongnam, South Korea
关键词
D O I
10.1109/ICDM54844.2022.00048
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Demand forecasting is a crucial component of supply chain management for revenue optimization and inventory planning. Traditional time series forecasting methods, however, have resulted in small models with limited expressive power because they have difficulty in scaling their model size up while maintaining high accuracy. In this paper, we propose Forecasting orchestra (Forchestra), a simple but powerful ensemble framework capable of accurately predicting future demand for a diverse range of items. Forchestra consists of two parts: 1) base predictors and 2) a neural conductor. For a given time series, each base predictor outputs its respective forecast based on historical observations. On top of the base predictors, the neural conductor adaptively assigns the importance weight for each predictor by looking at the representation vector provided by a representation module. Finally, Forchestra aggregates the predictions by the weights and constructs a final prediction. In contrast to previous ensemble approaches, the neural conductor and all base predictors of Forchestra are trained in an endto-end manner; this allows each base predictor to modify its reaction to different inputs, while supporting other predictors and constructing a final prediction jointly. We empirically show that the model size is scalable to up to 0.8 billion parameters (approximate to 400-layer LSTM). The proposed method is evaluated on our proprietary E-Commerce (100K) and the public M5 (30K) datasets, and it outperforms existing forecasting models with a significant margin. In addition, we observe that our framework generalizes well to unseen data points when evaluated in a zeroshot fashion on downstream datasets. Last but not least, we present extensive qualitative and quantitative studies to analyze how the proposed model outperforms baseline models and differs from conventional ensemble approaches. The code is available at https://github.com/young-j-park/22-ICDM-Forchestra.
引用
收藏
页码:378 / 387
页数:10
相关论文
共 50 条
  • [31] A Large-Scale Graph Learning Framework of Technological Gatekeepers by MapReduce
    Liu Tong
    Guo Wensheng
    2012 IEEE 26TH INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM WORKSHOPS & PHD FORUM (IPDPSW), 2012, : 1997 - 2003
  • [32] A large-scale evaluation framework for EEG deep learning architectures
    Heilmeyer, Felix A.
    Schirrmeister, Robin T.
    Fiederer, Lukas D. J.
    Voelker, Martin
    Behncke, Joos
    Ball, Tonio
    2018 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2018, : 1039 - 1045
  • [33] Large-scale forecasting of information spreading
    Severiukhina, Oksana
    Kesarev, Sergey
    Bochenina, Klavdiya
    Boukhanovsky, Alexander
    Lees, Michael H.
    Sloot, Peter M. A.
    JOURNAL OF BIG DATA, 2020, 7 (01)
  • [34] Study on forecasting human resources demand of large-scale hydropower project development for owners
    Peng, Q. (pengqingfeng@ehdc.com.cn), 1600, Tsinghua University (31):
  • [35] Large-scale forecasting of information spreading
    Oksana Severiukhina
    Sergey Kesarev
    Klavdiya Bochenina
    Alexander Boukhanovsky
    Michael H. Lees
    Peter M. A. Sloot
    Journal of Big Data, 7
  • [36] Deep Learning Framework for Forecasting Tourism Demand
    Laaroussi, Houria
    Guerouate, Fatima
    Sbihi, Mohamed
    2020 IEEE INTERNATIONAL CONFERENCE ON TECHNOLOGY MANAGEMENT, OPERATIONS AND DECISIONS (ICTMOD), 2020,
  • [37] Agent-based demand-modeling framework for large-scale microsimulations
    Balmer, Michael
    Axhausen, Kay W.
    Nagel, Kai
    TRAVELER BEHAVIOR AND VALUES 2006, 2006, (1985): : 125 - 134
  • [38] Large-scale paper supercapacitors on demand
    Brooke, Robert
    Ahlin, Jessica
    Huebscher, Kathrin
    Hagel, Olle
    Strandberg, Jan
    Sawatdee, Anurak
    Edberg, Jesper
    JOURNAL OF ENERGY STORAGE, 2022, 50
  • [39] Hybrid Deep Learning Ensemble Model for Improved Large-Scale Car Recognition
    Verma, Abhishek
    Liu, Yu
    2017 IEEE SMARTWORLD, UBIQUITOUS INTELLIGENCE & COMPUTING, ADVANCED & TRUSTED COMPUTED, SCALABLE COMPUTING & COMMUNICATIONS, CLOUD & BIG DATA COMPUTING, INTERNET OF PEOPLE AND SMART CITY INNOVATION (SMARTWORLD/SCALCOM/UIC/ATC/CBDCOM/IOP/SCI), 2017,
  • [40] SVM ensemble based transfer learning for large-scale membrane proteins discrimination
    Mei, Suyu
    JOURNAL OF THEORETICAL BIOLOGY, 2014, 340 : 105 - 110