Replies: 3 comments 1 reply
-
Hi all, Would appreciate if someone could provide any leads on the above. In summary, best model suggested from cross validation does not visually suitable for predictions as per the attached image and earlier commentary. Thank you in advance. |
Beta Was this translation helpful? Give feedback.
-
@jmoralez , @federicogarza: Could you please look into this discrepancy. I am observing that the best models across unique_ids are always producing almost constant forecast which does not seem correct. Adding an image for your reference, where the best model as per cross validation is SES. Pls let me know if this is indeed an issue or possibly the way I am running the code. Following are the forecast and cv settings used:
Attached is the data for this unique_id:
And the forecasts from various models for next 6 months (56 to 61) of which SES has been selected as the best model as per RMSE criterion:
Thank you in advance. |
Beta Was this translation helpful? Give feedback.
-
Hi - visual interpretation may be misleading - what is the forecast error that all these models produce on the cross-validation set? (see the tutorial for computing the errors) |
Beta Was this translation helpful? Give feedback.
-
Hi all,
I am exploring statsforecast using the reference page: https://nixtlaverse.nixtla.io/statsforecast/docs/getting-started/getting_started_complete.html.
There are roughly 200 unique_ids which I'd like to forecast + do a back-test/cross validation as well. Using rmse as the evaluation metric to identify best model while cross validation.
During this course, I observed the following which seems a bit off for 10 sample unique_ids:
The best model suggested by statsforecast (using cross validation) does not seem to hold true when observing/visualizing the predictions from various models. Though I understand this may not be the like to like comparison, I've ensured that using correct settings for horizon, step_size and n_windows, I am able to expose increasing amount of 'training' data sequentially to all the possible models.
For example in the attached screenshot, the best model suggested is ARIMA whereas visually it seems that one of CES or Holts-Winter seem more appropriate. This happens for all the 10 samples in one way or another. "Current approach" is just a custom approach which is not available in statsforecast but has been incorporated as an additional algorithm.
Would request inputs/comments if this is the right way of looking at the data/results or is there anything more that might need to be included.
Statsforecast version: v.1.7.3. Thanks in advance.
Beta Was this translation helpful? Give feedback.
All reactions