HierarchicalForecast offers a collection of reconciliation methods, including BottomUp
, TopDown
, MiddleOut
, MinTrace
and ERM
. And Probabilistic coherent predictions including Normality
, Bootstrap
, and PERMBU
.
A vast amount of time series datasets are organized into structures with different levels or hierarchies of aggregation. Examples include categories, brands, or geographical groupings. Coherent forecasts across levels are necessary for consistent decision-making and planning. Hierachical Forecast offers different reconciliation methods that render coherent forecasts across hierachies. Until recent, this methods were mainly avaiable in the R ecosystem. This Python-based framework aims to bridge the gap between statistical modeling and Machine Learning in the time series field.
- Classic reconciliation methods:
BottomUp
: Simple addition to the upper levels.TopDown
: Distributes the top levels forecasts trough the hierarchies.
- Alternative reconciliation methods:
MiddleOut
: It anchors the base predictions in a middle level. The levels above the base predictions use the bottom-up approach, while the levels below use a top-down.MinTrace
: Minimizes the total forecast variance of the space of coherent forecasts, with the Minimum Trace reconciliation.ERM
: Optimizes the reconciliation matrix minimizing an L1 regularized objective.
- Probabilistic coherent methods:
Normality
: Uses MinTrace variance-covariance closed form matrix under a normality assumption.Bootstrap
: Generates distribution of hierarchically reconciled predictions using Gamakumara's bootstrap approach.PERMBU
: Reconciles independent sample predictions by reinjecting multivariate dependence with estimated rank permutation copulas, and performing a Bottom-Up aggregation.
Missing something? Please open an issue here or write us in
Short: We want to contribute to the ML field by providing reliable baselines and benchmarks for hierarchical forecasting task in industry and academia. Here's the complete paper.
Verbose: HierarchicalForecast
integrates publicly available processed datasets, evaluation metrics, and a curated set of standard statistical baselines. In this library we provide usage examples and references to extensive experiments where we showcase the baseline's use and evaluate the accuracy of their predictions. With this work, we hope to contribute to Machine Learning forecasting by bridging the gap to statistical and econometric modeling, as well as providing tools for the development of novel hierarchical forecasting algorithms rooted in a thorough comparison of these well-established models. We intend to continue maintaining and increasing the repository, promoting collaboration across the forecasting community.
You can install HierarchicalForecast
's the Python package index pip with:
pip install hierarchicalforecast
You can also can install HierarchicalForecast
's from conda with:
conda install -c conda-forge hierarchicalforecast
The following example needs statsforecast
and datasetsforecast
as additional packages. If not installed, install it via your preferred method, e.g. pip install statsforecast datasetsforecast
.
The datasetsforecast
library allows us to download hierarhical datasets and we will use statsforecast
to compute the base forecasts to be reconciled.
You can open a complete example in Colab
Minimal Example:
# !pip install -U numba statsforecast datasetsforecast
import numpy as np
import pandas as pd
#obtain hierarchical dataset
from datasetsforecast.hierarchical import HierarchicalData
# compute base forecast no coherent
from statsforecast.core import StatsForecast
from statsforecast.models import AutoARIMA, Naive
#obtain hierarchical reconciliation methods and evaluation
from hierarchicalforecast.core import HierarchicalReconciliation
from hierarchicalforecast.evaluation import evaluate
from hierarchicalforecast.methods import BottomUp, TopDown, MiddleOut
from utilsforecast.losses import mse
# Load TourismSmall dataset
Y_df, S, tags = HierarchicalData.load('./data', 'TourismSmall')
Y_df['ds'] = pd.to_datetime(Y_df['ds'])
S = S.reset_index(names="unique_id")
#split train/test sets
Y_test_df = Y_df.groupby('unique_id').tail(4)
Y_train_df = Y_df.drop(Y_test_df.index)
# Compute base auto-ARIMA predictions
fcst = StatsForecast(df=Y_train_df,
models=[AutoARIMA(season_length=4), Naive()],
freq='Q', n_jobs=-1)
Y_hat_df = fcst.forecast(h=4)
# Reconcile the base predictions
reconcilers = [
BottomUp(),
TopDown(method='forecast_proportions'),
MiddleOut(middle_level='Country/Purpose/State',
top_down_method='forecast_proportions')
]
hrec = HierarchicalReconciliation(reconcilers=reconcilers)
Y_rec_df = hrec.reconcile(Y_hat_df=Y_hat_df, Y_df=Y_train_df,
S=S, tags=tags)
Assumes you have a test dataframe.
df = Y_rec_df.merge(Y_test_df, on=['unique_id', 'ds'])
evaluation = evaluate(df = df,
tags = tags,
train_df = Y_train_df,
metrics = [mse],
benchmark = "Naive")
Here is a link to the documentation.
This project is licensed under the MIT License - see the LICENSE file for details.
In the R ecosystem, we recommend checking out fable, and the now-retired hts. In Python we want to acknowledge the following libraries hiere2e, sktime, darts, pyhts, scikit-hts.
This work is highly influenced by the fantastic work of previous contributors and other scholars who previously proposed the reconciliation methods presented here. We want to highlight the work of Rob Hyndman, George Athanasopoulos, Shanika L. Wickramasuriya, Souhaib Ben Taieb, and Bonsoo Koo. For a full reference link, please visit the Reference section of this paper. We encourage users to explore this literature review.
If you enjoy or benefit from using these Python implementations, a citation to this hierarchical forecasting reference paper will be greatly appreciated.
@article{olivares2024hierarchicalforecastreferenceframeworkhierarchical,
title={HierarchicalForecast: A Reference Framework for Hierarchical Forecasting in Python},
author={Kin G. Olivares and Azul Garza and David Luo and Cristian ChallΓΊ and Max Mergenthaler and Souhaib Ben Taieb and Shanika L. Wickramasuriya and Artur Dubrawski},
year={2024},
eprint={2207.03517},
archivePrefix={arXiv},
primaryClass={stat.ML},
url={https://arxiv.org/abs/2207.03517},
}