My work focuses on addressing the growing need for reliable, day-ahead energy demand forecasts in smart grids. In particular, we have been developing structured ensemble models for probabilistic forecasting that are able to incorporate information from a number of sources. I have undertaken this EDF-sponsored project with the help of my supervisors Matteo Fasiolo (UoB) and Yannig Goude (EDF) and in collaboration Christian Capezza (University of Naples Federico II).
One of the largest challenges society faces is climate change. Decarbonisation will lead to both a considerable increase in demand for electricity and a change in the way it is produced. Reliable demand forecasts will play a key role in enabling this transition. Historically, electricity has been produced by large, centralised power plants. This allows production to be relatively easily tailored to demand with little need for large-scale storage infrastructure. However, renewable methods are typically decentralised, less flexible and supply is subject to weather conditions or other unpredictable factors. A consequence of this is that electricity production will less able to react to sudden changes in demand, instead it will need to be generated in advance and stored. To limit the need for large-scale and expensive electricity storage and transportation infrastructure, smart grid management systems can instead be employed. This will involve, for example, smaller, more localised energy storage options. This increases the reliance on accurate demand forecasts to inform storage management decisions, not only at the aggregate level, but possibly down at the individual household level. The recent impact of the Covid-19 pandemic also highlighted problems in current forecasting methods which struggled to cope with the sudden change in demand patterns. These issues call attention to the need to develop a framework for more flexible energy forecasting models that are accurate at the household level. At this level, demand is characterised by a low signal-to-noise ratio, with frequent abrupt changepoints in demand dynamics. This can be seen in Figure 1 below.
Additive stacking is a probabilistic ensemble method where we form a weighted mixture distribution of multiple predictive densities. We begin by dividing our data into 3 parts.
- Expert training: The data we use to train the component models of the ensemble
- Stacking: The data we use to fit the weights
- Validation set: The data we use to evaluate the performance of our ensemble.
We first fit our experts (individual models) using the expert training data. We then find the predictive densities of our fitted experts on the stacking data. Suppose in our stacking data we have data points for and fitted predictive densities for . For ease of notation we will let . Then our stacked model has the general likelihood:
Where are the weights. We want to find the weights that maximse this likelihood with respect to the constraints:
We want the weights to depend on some features from our data. This idea was explored by Capezza using multinomial weights of the form:
Where is determined by some set of features. We do not only include linear combinations of features but also, random or smooth effects based on spline basis expansions. These can be fitted using the methods in Wood et al (2016). A key limitation of the multinomial weights is that there is a huge amount of flexibility as each weight has it’s own which can lead to overfitting. Another problem is that the number of unknown parameters required to estimate grows linearly with the number of experts. This limits the number of experts and thus the amount of information that can be given to the model as it becomes too computationally complex to fit. This motivated the idea of more structured weights that, at the cost of flexibility, would be less prone to overfitting and more efficient to compute.
In order to reduce the number of unknown parameters, we first need to make an assumption about the experts. In this case we assume that the experts have some associated ordering. For example, we could use a set of experts that looked at the previous day, week and month and order them accordingly. We treat the experts as though they are responses in an ordinal regression and use the framework of Winship and Mare (1984) to parametrise the weights. This involves fitting a set of thresholds such that . We can ensure this property holds by parameterising the thresholds as and,
The weight of expert at point is then,
Where can be any cumulative distribution function, but as a standard we use the logit function,
The weights of the experts at each data point are now entirely determined by a single and the cut points. To see how the weights change with you can use the interactive plot below. The area under the curve for each of the colours corresponds to the weighting. The cutpoints are represented by the vertical lines.
A limitation of the ordered structure is that it relies on an assumption about the experts. That is, that they must have some associated order. This significantly restricts the number of available experts that you can use. The next step is to create a framework that allows an ordinal weighting structure to be used in conjunction with unordered experts as well. Thus we would retain the modelling advantages of the ordered framework with the flexibility of a multinomial structure.
 Christian Capezza, Biagio Palumbo, Yannig Goude, Simon N. Wood, and Matteo Fasiolo. Additive stacking for disaggregate electricity demand forecasting, 2020.
 David H. Wolpert. Stacked generalization. Neural Networks, 5(2):241–259, 1992a. ISSN 0893-6080. doi: https://doi.org/10.1016/S0893-6080(05)80023-1.
 Simon N. Wood, Natalya Pya, and Benjamin S ̈afken. Smoothing parameter and model selection for general smooth models. Journal of the American Statistical Association, 111(516):1548–1563, 2016. doi: 10.1080/01621459.2016.1180986.
 Trindade (2015), Electricity Load Diagrams, Retrieved from: https://archive.ics.uci.edu/ml/datasets/ElectricityLoadDiagrams20112014