Evaluating time series forecasting models: an empirical study on performance estimation methods

被引:107
|
作者
Cerqueira, Vitor [1 ]
Torgo, Luis [1 ,2 ,3 ]
Mozetic, Igor [4 ]
机构
[1] LIAAD INESC TEC, Porto, Portugal
[2] Univ Porto, Porto, Portugal
[3] Dalhousie Univ, Halifax, NS, Canada
[4] Jozef Stefan Inst, Ljubljana, Slovenia
关键词
Performance estimation; Model selection; Cross validation; Time series; Forecasting; CROSS-VALIDATION; SELECTION;
D O I
10.1007/s10994-020-05910-7
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Performance estimation aims at estimating the loss that a predictive model will incur on unseen data. This process is a fundamental stage in any machine learning project. In this paper we study the application of these methods to time series forecasting tasks. For independent and identically distributed data the most common approach is cross-validation. However, the dependency among observations in time series raises some caveats about the most appropriate way to estimate performance in this type of data. Currently, there is no consensual approach. We contribute to the literature by presenting an extensive empirical study which compares different performance estimation methods for time series forecasting tasks. These methods include variants of cross-validation, out-of-sample (holdout), and prequential approaches. Two case studies are analysed: One with 174 real-world time series and another with three synthetic time series. Results show noticeable differences in the performance estimation methods in the two scenarios. In particular, empirical experiments suggest that blocked cross-validation can be applied to stationary time series. However, when the time series are non-stationary, the most accurate estimates are produced by out-of-sample methods, particularly the holdout approach repeated in multiple testing periods.
引用
收藏
页码:1997 / 2028
页数:32
相关论文
共 50 条