New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[ENH] set random seed in TestAllForecasters
data generation - potential solution for sporadic failures
#6382
base: main
Are you sure you want to change the base?
Conversation
hm, no failures but long runtimes - |
Was this one run or multiples? After taking a look into the runtimes. Based on the one action run I observe:
|
Maybe it is latency? Perhaps this is some DDoS protection kicking in, not allowing too many downloads from the same IP address? This is perhaps related to a new problem I have been seeing: sometimes when I try to access the logs, my virus scanner says the IP has been blacklisted. Hypothesis:
Could it be this, @benHeid, @yarnabrina? If yes, it may indicate we need to think carefully about testing of hugging face based models - we had similar issues with downloads earlier, so we moved them out to a separate "downloads" CI element. Only that this time, these are downloads attached to models. |
We need to test if there are really that many downloads. As far as I know, hugging face is caching downloads. Thus, once the model is downloaded it shouldn't be downloaded again.. |
Regarding the long runtimes of the AutoModel from Statsforecast. I suppose that there is something strange with python 3.8 (and Mac). Furthermore, I think that this issue is not located in sktime: Local measurement of the execution time of the unit test with different initializations of the model:
Measurements of the direct fit time. (20 fits with a random time series with a length 1000). I executed it multiple times, since the numbers are fluctuating a lot... Random data 20 fits
|
@benHeid suggested that sporadic test failures and long test times in #6344 could be related to LU decomposition or similar issues in ARIMA - compare #6201.
This PR aims to help with diagnosis, by setting the random seed in data generation in
TestAllForecasters
. This should greatly reduce the number of different time series occurring, and hopefully make the failure behaviour - if impacted - deterministic.FYI @yarnabrina
The CI should run all forecaster tests because of the "test class has changed" criterion.