Training and predicting for multiple households
We have picked a few models (LassoCV
, XGBRFRegressor
, and LGBMRegressor
) that are doing better in terms of metrics, as well as runtime, to run on all the selected households in our validation dataset. The process is straightforward: loop over all the unique combinations, inner loop over the different models to run, and then train, predict, and evaluate. The code is available in the 01-Forecasting with ML.ipynb
notebook in chapter08
, under the Running an ML Forecast For All Consumers heading. You can run the code and take a break because this is going to take a little less than an hour. The notebook also calculates the metrics and contains a summary table that will be ready for you when you’re back. Let’s look at the summary now:
Figure 8.19 – Aggregate metrics on all the households in the validation dataset
Here, we can see that even at the aggregated level, the different models we used perform...