Optimizing dynamic time warping's window width for time series data mining applications

被引:52
|
作者
Hoang Anh Dau [1 ]
Silva, Diego Furtado [2 ]
Petitjean, Francois [3 ]
Forestier, Germain [4 ]
Bagnall, Anthony [5 ]
Mueen, Abdullah [6 ]
Keogh, Eamonn [1 ]
机构
[1] Univ Calif Riverside, Riverside, CA 92521 USA
[2] Univ Fed Sao Carlos, Sao Carlos, SP, Brazil
[3] Monash Univ, Melbourne, Vic, Australia
[4] Univ Haute Alsace, Mulhouse, France
[5] Univ East Anglia, Norwich, Norfolk, England
[6] Univ New Mexico, Albuquerque, NM 87131 USA
基金
澳大利亚研究理事会; 英国工程与自然科学研究理事会;
关键词
Time series; Clustering; Classification; Dynamic time warping; Semi-supervised learning; CLASSIFICATION;
D O I
10.1007/s10618-018-0565-y
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Dynamic Time Warping (DTW) is a highly competitive distance measure for most time series data mining problems. Obtaining the best performance from DTW requires setting its only parameter, the maximum amount of warping (w). In the supervised case with ample data, w is typically set by cross-validation in the training stage. However, this method is likely to yield suboptimal results for small training sets. For the unsupervised case, learning via cross-validation is not possible because we do not have access to labeled data. Many practitioners have thus resorted to assuming that "the larger the better", and they use the largest value of w permitted by the computational resources. However, as we will show, in most circumstances, this is a na < ve approach that produces inferior clusterings. Moreover, the best warping window width is generally non-transferable between the two tasks, i.e., for a single dataset, practitioners cannot simply apply the best w learned for classification on clustering or vice versa. In addition, we will demonstrate that the appropriate amount of warping not only depends on the data structure, but also on the dataset size. Thus, even if a practitioner knows the best setting for a given dataset, they will likely be at a lost if they apply that setting on a bigger size version of that data. All these issues seem largely unknown or at least unappreciated in the community. In this work, we demonstrate the importance of setting DTW's warping window width correctly, and we also propose novel methods to learn this parameter in both supervised and unsupervised settings. The algorithms we propose to learn w can produce significant improvements in classification accuracy and clustering quality. We demonstrate the correctness of our novel observations and the utility of our ideas by testing them with more than one hundred publicly available datasets. Our forceful results allow us to make a perhaps unexpected claim; an underappreciated "low hanging fruit" in optimizing DTW's performance can produce improvements that make it an even stronger baseline, closing most or all the improvement gap of the more sophisticated methods proposed in recent years.
引用
收藏
页码:1074 / 1120
页数:47
相关论文
共 50 条
  • [1] Optimizing dynamic time warping’s window width for time series data mining applications
    Hoang Anh Dau
    Diego Furtado Silva
    François Petitjean
    Germain Forestier
    Anthony Bagnall
    Abdullah Mueen
    Eamonn Keogh
    [J]. Data Mining and Knowledge Discovery, 2018, 32 : 1074 - 1120
  • [2] On-line and dynamic time warping for time series data mining
    Hailin Li
    [J]. International Journal of Machine Learning and Cybernetics, 2015, 6 : 145 - 153
  • [3] On-line and dynamic time warping for time series data mining
    Li, Hailin
    [J]. INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2015, 6 (01) : 145 - 153
  • [4] Judicious Setting of Dynamic Time Warping's Window Width Allows More Accurate Classification of Time Series
    Dau, Hoang Anh
    Silva, Diego Furtado
    Petitjean, Francois
    Forestier, Germain
    Bagnall, Anthony
    Keogh, Eamonn
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2017, : 917 - 922
  • [5] Parallelization of Searching and Mining Time Series Data using Dynamic Time Warping
    Shabib, Ahmed
    Narang, Anish
    Niddodi, Chaitra Prasad
    Das, Madhura
    Pradeep, Rachita
    Shenoy, Varun
    Auradkar, Prafullata
    Vignesh, T. S.
    Sitaram, Dinkar
    [J]. 2015 INTERNATIONAL CONFERENCE ON ADVANCES IN COMPUTING, COMMUNICATIONS AND INFORMATICS (ICACCI), 2015, : 343 - 348
  • [6] Similarity Measure Based on Incremental Warping Window for Time Series Data Mining
    Li, Hailin
    Wang, Cheng
    [J]. IEEE ACCESS, 2019, 7 : 3909 - 3917
  • [7] Time works well: Dynamic time warping based on time weighting for time series data mining
    Li, Hailin
    [J]. INFORMATION SCIENCES, 2021, 547 : 592 - 608
  • [8] Dynamic time warping based on cubic spline interpolation for time series data mining
    Li, Hailin
    Wan, Xiaoji
    Liang, Ye
    Gao, Shile
    [J]. 2014 IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOP (ICDMW), 2014, : 19 - 26
  • [9] Addressing Big Data Time Series: Mining Trillions of Time Series Subsequences Under Dynamic Time Warping
    Rakthanmanon, Thanawin
    Campana, Bilson
    Mueen, Abdullah
    Batista, Gustavo
    Westover, Brandon
    Zhu, Qiang
    Zakaria, Jesin
    Keogh, Eamonn
    [J]. ACM TRANSACTIONS ON KNOWLEDGE DISCOVERY FROM DATA, 2013, 7 (03)
  • [10] A local segmented dynamic time warping distance measure algorithm for time series data mining
    Dong, Xiao-Li
    Gu, Cheng-Kui
    Wang, Zheng-Ou
    [J]. PROCEEDINGS OF 2006 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS, VOLS 1-7, 2006, : 1247 - +