Introducing Modeltime Ensemble: Time Series Forecast Stacking
Want to share your content on Rbloggers? click here if you have a blog, or here if you don't.
I’m SUPER EXCITED to introduce modeltime.ensemble
, the time series ensemble extension to modeltime
. This tutorial (view original article) introduces our new R Package, Modeltime Ensemble, which makes it easy to perform stacked forecasts that improve forecast accuracy. If you like what you see, I have an Advanced Time Series Course where you will become the timeseries expert for your organization by learning modeltime
, modeltime.ensemble
, and timetk
.
Forecasting and Time Series Software Announcements
Articles on the modeltime
and timetk
forecasting and time series ecosystem.
 Introducing Modeltime Ensemble: Time Series Forecast Stacking
 Introducing Modeltime: Tidy Time Series Forecasting using Tidymodels
 Timetk 2.0.0: Visualize Time Series Data in 1Line of Code
Like these articles?
👉 Register to stay in the know
👈
on new cuttingedge R software like modeltime
.
Modeltime Ensemble
The time series ensemble extension to Modeltime
Three months ago I introduced modeltime
, a new R package that speeds up forecasting experimentation and model selection with Machine Learning (e.g. XGBoost, GLMNET, Prophet, Prophet Boost, ARIMA, and ARIMA Boost).
Fastforward to now. I’m thrilled to announce the first extension to Modeltime: modeltime.ensemble
.
Modeltime Ensemble is a cuttingedge package that integrates 3 competitionwinning time series ensembling strategies:

Super Learners (MetaLearners): Use
modeltime_fit_resamples()
andensemble_model_spec()
to create super learners (models that learn from the predictions of submodels) 
Weighted Ensembles: Use
ensemble_weighted()
to create weighted ensembles. 
Average Ensembles: Use
ensemble_average()
to build simple average and median ensembles.
HighPerformance Forecasting Stacks
Using these modeltime.ensemble
, you can build highperformance forecasting stacks. Here’s a MultiLevel Stack, which won the Kaggle Grupo Bimbo Inventory Demand Forecasting Competition (I teach this technique in my HighPerformance Time Series Forecasting Course).
The MultiLevel Stacked Ensemble that won the Kaggle Grupo Bimbo Inventory Demand Challenge
Ensemble Tutorial
Forecasting Product Sales with Average Ensembles
Today, I’ll cover forecasting Product Sales with Average and Weighted Ensembles, which are fast to implement and can have good performance (although superlearner’s tend to have better performance).
Weighted Stacking with Modeltime Ensemble
Ensemble Key Concepts:
The idea is that we have several submodels (Level 1) that make predictions. We can then take these predictions and combine them using a simple average (mean), median average, or a weighted average:
 Simple Average: Weights all models with the same proportion. Selects the average for each timestamp. Use
ensemble_average(type = "mean")
.  Median Average: No weighting. Selects prediction using the centered value for each time stamp. Use
ensemble_average(type = "median")
.  Weighted Average: User defines the weights (loadings). Applies a weighted average at each of the timestamps. Use
ensemble_weighted(loadings = c(1, 2, 3, 4))
.
More Advanced Ensembles:
The average and weighted ensembles are the simplest approaches to ensembling. One method that Modeltime Ensemble has integrated is Super Learners. We won’t cover these in this tutorial. But, I teach them in my HighPerformance Time Series Course. 💪
Getting Started
Let’s kick the tires on modeltime.ensemble
Install modeltime.ensemble
.
Load the following libraries.
Get Your Data
Forecasting Product Sales
Our Business objective is to forecast the next 12weeks of Product Sales given 2year sales history.
We’ll start with a walmart_sales_weekly
time series data set that includes Walmart Product Transactions from several stores, which is a small sample of the dataset from Kaggle Walmart Recruiting – Store Sales Forecasting. We’ll simplify the data set to a univariate time series with columns, “Date” and “Weekly_Sales” from Store 1 and Department 1.
Next, visualize the dataset with the plot_time_series()
function. Toggle .interactive = TRUE
to get a plotly interactive plot. FALSE
returns a ggplot2 static plot.
Seasonality Evaluation
Let’s do a quick seasonality evaluation to hone in on important features using plot_seasonal_diagnostics()
.
We can see that certain weeks and months of the year have higher sales. These anomalies are likely due to events. The Kaggle Competition informed competitors that Super Bowl, Labor Day, Thanksgiving, and Christmas were special holidays. To approximate the events, week number and month may be good features. Let’s come back to this when we preprocess our data.
Train / Test
Split your time series into training and testing sets
Give the objective to forecast 12 weeks of product sales, we use time_series_split()
to make a train/test set consisting of 12weeks of test data (hold out) and the rest for training.
 Setting
assess = "12 weeks"
tells the function to use the last 12weeks of data as the testing set.  Setting
cumulative = TRUE
tells the sampling to use all of the prior data as the training set.
Next, visualize the train/test split.
tk_time_series_cv_plan()
: Converts the splits object to a data frameplot_time_series_cv_plan()
: Plots the time series sampling data using the “date” and “value” columns.
Feature Engineering
We’ll make a number of calendar features using recipes
. Most of the heavy lifting is done by timetk::step_timeseries_signature()
, which generates a series of common time series features. We remove the ones that won’t help. After dummying we have 74 total columns, 72 of which are engineered calendar features.
Make SubModels
Let’s make some submodels with Modeltime
Now for the fun part! Let’s make some models using functions from modeltime
and parsnip
.
Auto ARIMA
Here’s the basic Auto ARIMA Model.
 Model Spec:
arima_reg()
<– This sets up your general model algorithm and key parameters  Set Engine:
set_engine("auto_arima")
<– This selects the specific packagefunction to use and you can add any functionlevel arguments here.  Fit Model:
fit(Weekly_Sales ~ Date, training(splits))
<– All Modeltime Models require a date column to be a regressor.
Elastic Net
Making an Elastic NET model is easy to do. Just set up your model spec using linear_reg()
and set_engine("glmnet")
. Note that we have not fitted the model yet (as we did in previous steps).
Next, make a fitted workflow:
 Start with a
workflow()
 Add a Model Spec:
add_model(model_spec_glmnet)
 Add Preprocessing:
add_recipe(recipe_spec %>% step_rm(date))
<– Note that I’m removing the “date” column since Machine Learning algorithms don’t typically know how to deal with date or datetime features  Fit the Workflow:
fit(training(splits))
XGBoost
We can fit a XGBoost Model using a similar process as the Elastic Net.
NNETAR
We can use a NNETAR model. Note that add_recipe()
uses the full recipe (with the Date column) because this is a Modeltime Model.
Prophet w/ Regressors
We’ll build a Prophet Model with Regressors. This uses the Facebook Prophet forecasting algorithm and supplies all of the 72 features as regressors to the model. Note – Because this is a Modeltime Model we need to have a Date Feature in the recipe.
SubModel Evaluation
Let’s take a look at our progress so far. We have 5 models. We’ll put them into a Modeltime Table to organize them using modeltime_table()
.
We can get the accuracy on the holdout set using modeltime_accuracy()
and table_modeltime_accuracy()
. The best model is the Prophet with Regressors with a MAE of 1031.
Accuracy Table  

.model_id  .model_desc  .type  mae  mape  mase  smape  rmse  rsq 
1  ARIMA(0,0,1)(0,1,0)[52]  Test  1359.99  6.77  1.02  6.93  1721.47  0.95 
2  GLMNET  Test  1222.38  6.47  0.91  6.73  1349.88  0.98 
3  XGBOOST  Test  1089.56  5.22  0.82  5.20  1266.62  0.96 
4  NNAR(4,1,10)[52]  Test  2529.92  11.68  1.89  10.73  3507.55  0.93 
5  PROPHET W/ REGRESSORS  Test  1031.53  5.13  0.77  5.22  1226.80  0.98 
And, we can visualize the forecasts with modeltime_forecast()
and plot_modeltime_forecast()
.
Build Modeltime Ensembles
This is exciting.
We’ll make Average, Median, and Weighted Ensembles. If you are interested in making Super Learners (MetaLearner Models that leverage submodel predictions), I teach this in my new HighPerformance Time Series course.
I’ve made it super simple to build an ensemble from a Modeltime Tables. Here’s how to use ensemble_average()
.
 Start with your Modeltime Table of SubModels
 Pipe into
ensemble_average(type = "mean")
You now have a fitted average ensemble.
We can make median and weighted ensembles just as easily. Note  For the weighted ensemble I’m loading the better performing models higher.
Ensemble Evaluation
Let’s see how we did
We need to have Modeltime Tables that organize our ensembles before we can assess performance. Just use modeltime_table()
to organize ensembles just like we did for the SubModels.
Let’s check out the Accuracy Table using modeltime_accuracy()
and table_modeltime_accuracy()
.
 From MAE, Ensemble Model ID 1 has 1000 MAE, a 3% improvement over our best submodel (MAE 1031).
 From RMSE, Ensemble Model ID 3 has 1228, which is on par with our best submodel.
Accuracy Table  

.model_id  .model_desc  .type  mae  mape  mase  smape  rmse  rsq 
1  ENSEMBLE (MEAN): 5 MODELS  Test  1000.01  4.63  0.75  4.58  1408.68  0.97 
2  ENSEMBLE (MEDIAN): 5 MODELS  Test  1146.60  5.68  0.86  5.77  1310.30  0.98 
3  ENSEMBLE (WEIGHTED): 5 MODELS  Test  1056.59  5.15  0.79  5.20  1228.45  0.98 
And finally we can visualize the performance of the ensembles.
It gets better
You’ve just scratched the surface, here’s what’s coming…
The modeltime.ensemble
package functionality is much more featurerich than what we’ve covered here (I couldn’t possibly cover everything in this post). 😀
Here’s what I didn’t cover:

SuperLearners: We can make use resample predictions from our submodels as inputs to a metalearner. This can result is significantly better accuracy (5% improvement is what we achieve in my Time Series Course).

MultiLevel Modeling: This is the strategy that won the Grupo Bimbo Inventory Demand Forecasting Challenge where multiple layers of esembles are used.

Refitting SubModels and MetaLearners: Refitting is special task that is needed prior to forecasting future data. Refitting requires careful attention to control the submodel and metalearner retraining process.
I teach each of these techniques and strategies so you become the time series expert for your organization. Here’s how. 👇
Advanced Time Series Course
Become the times series domain expert in your organization.
Make sure you’re notified when my new Advanced Time Series Forecasting in R course comes out. You’ll learn timetk
and modeltime
plus the most powerful time series forecasting techiniques available. Become the times series domain expert in your organization.
👉 Advanced Time Series Course.
You will learn:
 Time Series Preprocessing, Noise Reduction, & Anomaly Detection
 Feature Engineering using lagged variables & external regressors
 Hyperparameter Tuning
 Time Series CrossValidation
 Ensembling Multiple Machine Learning & Univariate Modeling Techniques (Competition Winner)
 NEW  Deep Learning with GluonTS (Competition Winner)
 and more.
Unlock the HighPerformance Time Series Course
Documentation
More information on the modeltime
ecosystem can be found in the software documentation Modeltime, Modeltime Ensemble, and
Timetk.
Have questions about Modeltime Ensemble?
Make a comment in the chat below. 👇
And, if you plan on using modeltime.ensemble
for your business, it’s a nobrainer  Take my Time Series Course.
Rbloggers.com offers daily email updates about R news and tutorials about learning R and many other topics. Click here if you're looking to post or find an R/datascience job.
Want to share your content on Rbloggers? click here if you have a blog, or here if you don't.