A Comparative Study of Performance Estimation Methods for Time Series Forecasting. Cerqueira, V., Torgo, L., Smailović, J., & Mozetic, I. IEEE International Conference on Data Science and Advanced Analytics (DSAA), 2017.
A Comparative Study of Performance Estimation Methods for Time Series Forecasting [link]Paper  doi  abstract   bibtex   
Performance estimation denotes a task of estimating the loss that a predictive model will incur on unseen data. These procedures are part of the pipeline in every machine learning task and are used for assessing the overall generalisation ability of models. In this paper we address the application of these methods to time series forecasting tasks. For independent and identically distributed data the most common approach is cross-validation. However, the dependency among observations in time series raises some caveats about the most appropriate way to estimate performance in these datasets and currently there is no settled way to do so. We compare different variants of cross-validation and different variants of out-of-sample approaches using two case studies: One with 53 real-world time series and another with three synthetic time series. Results show noticeable differences in the performance estimation methods in the two scenarios. In particular, empirical experiments suggest that cross-validation approaches can be applied to stationary synthetic time series. However, in real-world scenarios the most accurate estimates are produced by the out-of-sample methods, which preserve the temporal order of observations.
@article{cerqueira2017comparative,
abstract = {Performance estimation denotes a task of estimating the loss that a predictive model will incur on unseen data. These procedures are part of the pipeline in every machine learning task and are used for assessing the overall generalisation ability of models. In this paper we address the application of these methods to time series forecasting tasks. For independent and identically distributed data the most common approach is cross-validation. However, the dependency among observations in time series raises some caveats about the most appropriate way to estimate performance in these datasets and currently there is no settled way to do so. We compare different variants of cross-validation and different variants of out-of-sample approaches using two case studies: One with 53 real-world time series and another with three synthetic time series. Results show noticeable differences in the performance estimation methods in the two scenarios. In particular, empirical experiments suggest that cross-validation approaches can be applied to stationary synthetic time series. However, in real-world scenarios the most accurate estimates are produced by the out-of-sample methods, which preserve the temporal order of observations.},
author = {Cerqueira, Vitor and Torgo, Luis and Smailovi{\'{c}}, Jasmina and Mozetic, Igor},
doi = {10.1109/DSAA.2017.7},
file = {::},
isbn = {978-1-5090-5004-8},
journal = {IEEE International Conference on Data Science and Advanced Analytics (DSAA)},
keywords = {DOLFINS{\_}T3.1,DOLFINS{\_}WP3},
mendeley-tags = {DOLFINS{\_}T3.1,DOLFINS{\_}WP3},
pages = {529--538},
title = {{A Comparative Study of Performance Estimation Methods for Time Series Forecasting}},
url = {https://www.researchgate.net/publication/322586715{\_}A{\_}Comparative{\_}Study{\_}of{\_}Performance{\_}Estimation{\_}Methods{\_}for{\_}Time{\_}Series{\_}Forecasting},
year = {2017}
}

Downloads: 0