Solved – 10-fold cross validation for forecasting time series with explanatory data

cross-validationforecastingmachine learningtime series

I saw that the question was asked some years ago here, but I wasn't satisfied with the answers so I'm asking it again. Is there some theoretical foundations about not doing k-fold cross validation with time series ? I've read that it makes more sense to not use forward looking data to predict past data, but if you expect the dependencies to have evolved with time, you should have modelled it so this is not shoking to mix all the data for tuning the algorithm, and use the end of the window for testing (for aesthetic purposes and plots).

Plus most of approaches don't take into account the time dependency so it seems artificial to use the rolling windows (local methodes use distances between regressors in the sense of the metric, not in a timely way).
I was working with this paper that doesn't even evoke the question.

The only point I would underline is that if you use lags and regressors and lags of regressors, the observations of the training test would be correlated which may be problematic for some methods more than others, right ?

Best Answer

I am not an expert on the field, but I believe the paper "On the use of cross-validation for time series predictor evaluation" by Christoph Bergmeir can help answer your question. It basically addresses the question of whether the use of specialized evaluation techniques for time series provide an advantage to take care of dependencies and time-evolving effects that may occur. They do not provide a strong theoretical background but performed several experiments and basically concluded that no practical problems with standard cross-validation could be found. However, they suggest the use of blocked cross-validation, together with an adequate control for stationarity, since it makes full use of all available information both for training and testing, thus yielding a robust error estimate.

Related Question