-
Notifications
You must be signed in to change notification settings - Fork 2
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
auto ARIMA for influenza hospitalizations directly #37
Comments
interesting. can you add your code over at https://github.com/signaturescience/fiphde/tree/hosp-arima i was working on this at https://github.com/signaturescience/fiphde/blob/hosp-arima/scratch/hosp-arima.R holding the current four weeks (which includes the dramatic rise in hospitalizations) as the test set .... this doesnt look nearly as good: |
Alright, you're right. Pushed some relatively well-commented code in 710147b. Forecasting the next four weeks, ETS looks possibly too high, ARIMA too low, ensemble just right 🥣 🧸 📈 But backing up and removing four weeks of data and forecasting those weeks, it looks pretty bad OTOH, if you look at the 95% confidence interval on the ETS model, it's almost covering the true value. One of the things I want to look at is adding those ili ranks and hosp ranks as predictors here. We could treat this as an ex-post predictor as noted in https://otexts.com/fpp3/forecasting-regression.html. I have some ideas on how to do this. More here shortly. |
a972758 shows how to bring in historical severity into the ARIMA modelling and forecasting steps. Code shoudl be fairly well commented. Without historical hospitalization rank data, forecasting next four weeks: With hospitalization rank data: Previous four weeks, without historical predictor: With the historical predictor, maybe, slightly better? |
A few more commits in that script looking at how these models performed throughout the time series we have. @vpnagraj, Download this pdf and hold the pagedown/pageup key to scroll through the weeks. Keep in mind I'm only looking at the 10% interval to keep it tight. Note the ETS intervals are pretty wide at 95%. |
this is awesome. i added some code at 3cc73f2 might seem redundant. but i actually think what i did there could be really useful. basically recreated similar plots but using the heres what the updated plots look like (with 95% PI): bottom line. getting the forecasts into a common format / visualization paradigm sets us up to have a "fair" evaluation of forecasts from multiple methods (right now our glm-s and these several time series approaches). and we dont have to eyeball it. we can basically scoot back to any (every?) previous week where we have data and run the modeling => calculate WIS. from there we could justify any model selection by going with the framework that gives us the lowest (summed? averaged? median?) WIS over all forecast time points. |
About to blow up your tab with this monster image. Good/bad news. Good news. tsibble/fable "just works" as advertised. When the tsibble is keyed by a location, the modeling and forecasting steps work with nearly zero adjustments to the code 🔮 Bad news. I think we kind of knew this was going to happen based on some work we did at fluforce-init. Some states have extremely sparse data. I'm thinking we ought to implement some kind of filter on the incoming data, where if there aren't at least X number of observations with at least Y number of counts over some period Z, we throw out that location. |
initial work on state level forecasting #37
I removed a few states reporting extremely low numbers, and added a few transformations on the ets model in |
closed in #65 |
up to now we have discussed a two step forecasting workflow:
what if we tried auto ARIMA on hospitalizations directly?
Do this for both state and national level
The text was updated successfully, but these errors were encountered: