Sorry, maybe I wasn’t very clear. The case of predicting a time series with a pre-trained model on that same time series could be understood as a very specific case of transfer learning.
In other words, you could follow the Colab and use the model you trained on your data to predict your own and same data. Hopefully, you will achieve your practical goals. Namely: saving time on new training without compromising too much accuracy.
Regarding `statsforecast`: currently yo do have to retrain every time you want to update the model with new data. That doesn’t necessarily mean you have too retrain with every new observation.
Example: you want to predict the next 48 hours every 24 hours although you get new info every hour. In that case you would retrain from scratch every 24 hours, say at noon.
In an oversimplification, how people tend to deploy this models into production is by orchestrating workflows or pipelines that:
1. Extract data, do some wrangling and load to memory every so and so or given a specific event
2. Run the algorithms. Sometimes people run different models and then choose the best one bases on defined error metrics.
3. Output the prediction somewhere
For simple pipelines we have seen people use tools like cron on servers on the cloud.
For more complicated pipelines you can use things like Prefect, Airflow, Ray, Mlflow and even run the computation on clusters of different machines. If you use spot instances, this process is actually quite cheap and fast.
I’m sure you knew most of this stuff but wanted to be as clear as possible.
If you need help with your particular implementation we could schedule a call or something.