So a group of statisticians (from Cyprus and Greece) have written an easy-to-read paper comparing statistical and machine learning methods in time series forecasting, and found that statistical methods do better, both in terms of accuracy and computational complexity.
To me, there’s no surprise in the conclusion, since in the statistical methods, there is some human intelligence involved, in terms of removing seasonality, making the time series stationary and then using statistical methods that have been built specifically for time series forecasting (including some incredibly simple stuff like exponential smoothing).
Machine learning methods, on the other hand, are more general purpose – the same neural networks used for forecasting these time series, with changed parameters, can be used for predicting something else.
In a way, using machine learning for time series forecasting is like using that little screwdriver from a Swiss army knife, rather than a proper screwdriver. Yes, it might do the job, but it’s in general inefficient and not an effective use of resources.
Yet, it is important that this paper has been written since the trend in industry nowadays has been that given cheap computing power, machine learning be used for pretty much any problem, irrespective of whether it is the most appropriate method for doing so. You also see the rise of “machine learning purists” who insist that no human intelligence should “contaminate” these models, and machines should do everything.
By pointing out that statistical techniques are superior at time series forecasting compared to general machine learning techniques, the authors bring to attention that using purpose-built techniques can actually do much better, and that we can build better systems by using a combination of human and machine intelligence.
They also helpfully include this nice picture that summarises what machine learning is good for, and I wholeheartedly agree:
The paper also has some other gems. A few samples here:
Knowing that a certain sophisticated method is not as accurate as a much simpler one is upsetting from a scientific point of view as the former requires a great deal of academic expertise and ample computer time to be applied.
[…] the post-sample predictions of simple statistical methods were found to be at least as accurate as the sophisticated ones. This finding was furiously objected to by theoretical statisticians , who claimed that a simple method being a special case of e.g. ARIMA models, could not be more accurate than the ARIMA one, refusing to accept the empirical evidence proving the opposite.
A problem with the academic ML forecasting literature is that the majority of published studies provide forecasts and claim satisfactory accuracies without comparing them with simple statistical methods or even naive benchmarks. Doing so raises expectations that ML methods provide accurate predictions, but without any empirical proof that this is the case.
At present, the issue of uncertainty has not been included in the research agenda of the ML field, leaving a huge vacuum that must be filled as estimating the uncertainty in future predictions is as important as the forecasts themselves.
2 thoughts on “Statistics and machine learning”
Read the paper. A few things:
0) let’s assume there’s a dichotomy of stat vs. ML (there really isn’t, in my opinion, especially when you’re faced with an application and want to get the job done.)
1) ML can be shittily applied, stat is less likely to be shittily applied because it is not all reduced to an API like sklearn or tensorflow and one has to be smarter and more rigorous to understand statistical inference. So by selection, stat inference is used by smarter people than the average sklearn user. To that extent, their point is argued well.
The part of stat that is shittily applied is the p-value stuff, by biomedical folks, who interact with stat using a dumb API.
2) With that out of the way, the problem with the paper is that they’re really creating the weakest possible straw man out of ML to destroy it. They apply ML methods in very amateurish and indisciplined ways. They’re comparing ML as practiced by an amateur in the tail end of ML IQ curve with stat as practiced by a professional.
Perhaps that itself is the lesson we can learn from this paper: that ML libraries and hype have made the barrier for entry into ML so low that people apply it without much rigor or thoughtfulness, as a hammer seeking a nail.
Moral may be: libraries and APIs make it easier for people to use tools and a consequence of this is that people will apply the tools shittily (like they do in their own paper 🙂 ).
*I read the paper.
So, “read” as in past tense, not “you read” 😀