This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Towards Accurate Predictions and Causal ‘What-if’ Analyses for Planning and Policy-making: A Case Study in Emergency Medical Services Demand

Kasun Bandara1, Christoph Bergmeir1, Sam Campbell2, Deborah Scott2, Dan Lubman2 1Faculty of Information Technology, Monash University, Melbourne, Australia.
[email protected], [email protected]
2 Turning Point, Eastern Health Clinical School, Monash University, Melbourne, Australia.
[email protected], [email protected], [email protected]
Abstract

Emergency Medical Services (EMS) demand load has become a considerable burden for many government authorities, and EMS demand is often an early indicator for stress in communities, a warning sign of emerging problems. In this paper, we introduce Deep Planning and Policy Making Net (DeepPPMNet), a Long Short-Term Memory network based, global forecasting and inference framework to forecast the EMS demand, analyse causal relationships, and perform ‘what-if’ analyses for policy-making across multiple local government areas. Unless traditional univariate forecasting techniques, the proposed method follows the global forecasting methodology, where a model is trained across all the available EMS demand time series to exploit the potential cross-series information available. DeepPPMNet also uses seasonal decomposition techniques, incorporated in two different training paradigms into the framework, to suit various characteristics of the EMS related time series data. We then explore causal relationships using the notion of Granger Causality, where the global forecasting framework enables us to perform ‘what-if’ analyses that could be used for the national policy-making process. We empirically evaluate our method, using a set of EMS datasets related to alcohol, drug use and self-harm in Australia. The proposed framework is able to outperform many state-of-the-art techniques and achieve competitive results in terms of forecasting accuracy. We finally illustrate its use for policy-making in an example regarding alcohol outlet licenses.

Index Terms:
Emergency Medical Services, Time Series, Demand Forecasting, LSTM
©2020 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. Accepted to be Published in: Proceedings of the IJCNN 2020: International Joint Conference on Neural Networks, 19 - 24 July, 2020, Glasgow (UK)

I Introduction

Emergency Medical Services (EMS), also known as ambulance services or paramedic services, provide out-of-hospital medical care and transportation to a medical facility for those in need of urgent medical attention. In general, minimising the response time to emergency calls is crucial for EMS to treat patients efficiently and effectively.

Refer to caption

Figure 1: The reported number of ambulance attendances related to alcohol, drug use, self-harm, and mental health, for one of the states in Australia, over the past eight years.

Generating accurate and reliable spatio-temporal estimations of the future EMS demand is essential for authorities to proactively allocate the available resources, i.e., number of paramedics and size of ambulance fleets. A successful resource deployment strategy assures EMS deliver high-quality medical services in a timely manner. Furthermore, EMS demand patterns are often an early warning sign for certain problems in a community, and can therefore be useful for national-level policy makers, when introducing intervention strategies to address those problems.

The staggering growth of incidents requiring EMS assistance, combined with the scarcity of the resources have made EMS planning a challenging task for many countries [1, 2, 3, 4]. As a result, accurate EMS demand forecasting is becoming ever more important in the emergency services industry, so that the authorities can anticipate EMS requirements in advance to optimally utilise the available resources. Over the past two decades, numerous studies have been developed to improve the accuracy of EMS demand forecasts [1, 5, 2, 6, 7, 8, 3, 9, 4]. However, many of the existing approaches hold strong prior distributional and structural assumptions about the EMS data. Furthermore, these methods are univariate forecasting models, which treat each time series separately, and forecast them in isolation. As a result, these models are unable to exploit the common EMS demand patterns present in collections of time series. Motivated by this need, we propose a forecasting framework, Deep Planning and Policy Making Net (DeepPPMNet), training a Long Short-Term Memory Network (LSTM) globally across different series, and therewith accounting for the cross-series information available in a set of EMS demand time series.

In this work, we use a case study of ambulance attendances related to alcohol, drug use, and self-harm in Australia. Similar to other developed nations, Australia is experiencing an increasing demand for EMS over the past few years. For example, Fig. 1 illustrates the growth of EMS demand related to alcohol, drug use, self-harm, and mental health, for one particular Australian state. In the EMS capacity planning, generating sub-region demand forecasts has proven more useful than aggregated forecasts for an entire region [1, 2, 3]. In Australia, states and territories are subdivided into statistical sub-regions known as Local Government Areas (LGAs). This breakdown enables authorities to deploy a fine-grained resource allocation strategy. Fig. 2 illustrates such an example of EMS demand for suicide attempts, over the past eight years. Here, the EMS demand patterns of various LGAs can be similar and may share salient features in common, i.e., similar ambulance demand peaks and drops. In order to exploit these similar EMS demand patterns, the proposed DeepPPMNet framework employs a global forecasting methodology (GFM) that simultaneously learns from a collection of time series [10].

We subsequently perform causal inference on the potential factors that influence the EMS demand, employing the concept of Granger Causality (GC) [11, 12], that we define according to Granger [11, 12] as follows. Given a target time series YtY_{t}, and (multivariate) external factors WtW_{t} and ZtZ_{t}. Let us assume that we have a predictive model that uses past values of YtY_{t} and WtW_{t} to predict Yt+1Y_{t+1}. We now define that ZtZ_{t} Granger causes Yt+1Y_{t+1}, if we are able to improve the accuracy of predicting Yt+1Y_{t+1} by including past values of ZtZ_{t} as exogenous values in our model. Granger [11] notes that causality is a complicated concept and no universally accepted definition exists. Granger causality assumes that a cause has to precede the effect in time, i.e., the future cannot cause the past. Then, in the hypothetical case in which WtW_{t} and ZtZ_{t} together cover all possible factors, Granger argues that Granger causality is true causality. In practice, the strength of Granger causality depends heavily on the careful analysis and inclusion of possible influencing factors WtW_{t}, which effectively determines if Granger causality measures rather a correlation or suggests true causal relationships.

The GC analysis subsequently allows us to perform ‘what-if’ analyses, in the sense that if we are confident that we have identified exogenous variables ZtZ_{t} as causal factors, we can deliberately change their values to analyse how the predicted outcomes change accordingly. We note that using a GFM based forecasting model like DeepPPMNet greatly improves this methodology, as the model is trained across all the variations that exist among a collection of time series, which makes the ‘what-if’ analysis feasible even for relatively constant features or feature values that have not been observed in a particular time series before.

Refer to caption


Figure 2: Monthly ambulance attendances for suicide attempts of three different LGAs for one of the states in Australia, over a time period of eight years.

The main contributions of our paper are as follows:

  • To the best of our knowledge, our work is the first study that employs a GFM in a collection of EMS demand time series for forecasting. We evaluate our model on a real-world EMS demand dataset, which consists of ambulance attendances related to alcohol overdose, suicide attempts, and other drug related harms reported in Australia, and show that our framework achieves competitive forecasting accuracy.

  • Building on these results, we then show how DeepPPMNet provides a platform to conduct GC-based ‘what-if’ analyses, and allows decision makers to assess the factors that could drive the EMS demand, for which they can implement timely intervention or prevention strategies.

The remainder of this paper is organized as follows. In Section II, we review the state of the art in this domain, and discuss the advent of global forecasting models. In Section III, we discuss the proposed DeepPPMNet framework in detail. Our experimental setup is presented in Section IV, where we apply the proposed DeepPPMNet framework to multiple EMS demand datasets extracted from an internationally unique dataset of coded EMS clinical records. Finally, Section V concludes the paper.

II Prior Work

Early studies to forecast EMS demand are mostly based on exponential smoothing models [13], autoregressive moving average models [4, 6], factor models [7], and simple averaging models [14]. Baker and Fitzpatrick [13] use a Holt-Winter’s exponential smoothing model along with goal programming to forecast EMS demand in South Carolina. Channouf et al. [4] investigate the use of ARIMA models with special-day effects to predict EMS calls in Calgary, Alberta. Furthermore, Wong et al. [6] use ARIMA models to measure the effect of weather factors for daily ambulance call out demand. Also, Brown et al. [14] use various forms of averaging techniques to forecast the hourly emergency call volumes. Matteson et al. [7] propose a combined inter-valued time series model with a dynamic latent factor structure using hourly emergency call demand data from Toronto, Canada.

Overcoming the limitations of traditional EMS prediction approaches, Setzler et al. [3] introduce a neural network (NN) based methodology to generate spatio-temporal forecasts for an EMS agency in North Carolina. Based on demand data provided by the Welsh Ambulance Service Trust, Vile et al. [8] evaluate the application of singular spectrum analysis for daily EMS demand forecasting. Zhou and Matteson [2] develop a spatio-temporal kernel density estimation approach to predict ambulance demand in Toronto, Canada. More recently, Chen et al. [5] use a combination of machine learning techniques to forecast the pre-hospital emergency demand in New Taipei city, while Steins et al. [1] use a zero-inflated poisson regression model to forecast ambulance call demand in Sweden.

In parallel to these developments, deep learning based applications have achieved state-of-the-art performance in many domains [15, 16, 17]. In the time series forecasting field, the modern advances are mainly around deep learning based GFMs. These models are becoming a strong alternative to the univariate time series forecasting models, in situations where a collection of related time series is available. This is particularly evident from the recent success of such models in forecasting competitions [15] and the use in industry applications [16, 17]. As highlighted in Section I, compared to univariate time series models, which rely only on the individual level time series data, GFM models benefit from training across multiple time series, allowing them to learn key patterns available in a collection of time series.

III DeepPPMNet Framework

DeepPPMNet is a GFM based forecasting framework, which is designed to train across multiple time series. The DeepPPMNet consist of three layers, namely: 1) the pre-processing layer, 2) the DeepPPMNet training layer, and 3) the post-processing layer. In the following, we discuss the functionality of these layers in detail.

III-A Pre-processing layer

The pre-processing layer of DeepPPMNet is a three-stepped process. First, the set of EMS demand time series are normalised using a meanscale transformation strategy to scale various ambulance call volume ranges of multiple LGAs [17, 18], as DeepPPMNet is trained across a group of time series.

In the second step of pre-processing layer, following the suggestions of [19, 20], we perform a log transformation strategy to stabilise the variance of the EMS demand time series. In addition to stabilising the variance, the log transformation also assures additive seasonality and trend in the series, which is necessary for our decomposition layer, which is the last step of our pre-processing layer.

We use a time series decomposition layer as the final pre-processing step of DeepPPMNet. The main objective of this layer is to seasonally adjust the time series, i.e., removing the seasonal component from a time series, and thereby supplementing the subsequent DeepPPMNet learning phase. This deseasonalisation strategy is advocated by many studies, when using neural networks for time series forecasting [21, 20]. Due to the promising performance of Prophet [22] as a multi-seasonal decomposition technique in the recent past [21], we use Prophet as an additive, single-seasonal decomposition technique to obtain the seasonal components present in a time series. This additive decomposition can be defined as follows:

Xt=S^t+T^t+R^t+H^tX_{t}=\hat{S}_{t}+\hat{T}_{t}+\hat{R}_{t}+\hat{H}_{t} (1)

Here, XtX_{t} denotes the EMS demand observation at time tt, and the decomposed seasonal, trend, remainder, and holiday effect components of the time series are represented by S^t\hat{S}_{t}, T^t\hat{T}_{t}, R^t\hat{R}_{t}, and H^t\hat{H}_{t} respectively. We perform the seasonal extraction, after fitting the Prophet model using the prophet package implemented in R [23]. Also, this deseasonalisation process can be replaced with other additive decomposition techniques such as seasonal and trend decomposition using loess (STL) [24].

III-B DeepPPMNet Training layer

As the primary prediction unit of DeepPPMNet, we use an LSTM [25], a special type of neural network which is naturally suited for modelling time series data, because of its ability to capture long-term dependencies in sequences. LSTMs have been heavily used in sequence modelling tasks, such as Natural Language Processing, speech recognition, and more recently have received significant attention in the time series forecasting domain [15, 20, 17]. It is also worth mentioning that the proposed DeepPPMNet framework can potentially be employed with other RNN variants such as the Elman RNN [26] or the Gated Recurrent Unit (GRU) cell [27]. For more discussions on the LSTM architecture, we refer to Bandara et al. [20].

III-B1 Moving Window Transformation

In order to train the DeepPPMNet, we use the past observations of time series (XiX_{i}) in the form of input and output frames. Following recommendations in the literature [15, 20, 19], we achieve this by applying the Moving Window (MW) tranformation strategy to each pre-processed time series. In this process, the size of the output frame is equivalent to the size of the intended forecasting horizon MM, i.e., a Multi-Input Multi-Output (MIMO) principle is used in multi-step forecasting. This enables the DeepPPMNet to directly predict all the future observations up to the intended forecasting horizon XiMX^{M}_{i}, while avoiding prediction error accumulation at each forecasting step [28]. For further discussions of the MW transformation strategy we refer to Hewamalage et al. [19].

III-B2 DeepPPMNet Training Paradigms

Following Bandara et al. [21], we apply two training paradigms to the output generated by the decomposition layer.

In the first training paradigm, Deseasonalised Approach (DS), the MW strategy is applied to a seasonally adjusted time series, and the transformed input and output windows are used to train the DeepPPMNet. As the seasonal components are excluded from the time series, an additional reseasonalisation process is introduced in the post-processing layer to predict the future seasonal values of the time series. These seasonal values are estimated by repeating the last seasonal components of the time series (extracted by Prophet) to the intended forecast horizon. The DS approach can be seen as a ‘boosting’ ensemble technique [29], where the Prophet method used for the deseasonalisation process acts as a weak base learner to supplement the LSTM training procedure.

In the second training paradigm, Seasonal Exogenous Approach (SE), instead of detaching the seasonal components from the time series, we use the extracted seasonal components as exogenous variables together with the original time series (after normalising and log-transforming). As the original distribution of the time series is used to generate the training windows (using the MW process), the DeepPPMNet is expected to forecast all components of a time series, including extrapolation of the seasonality. In this way, SE supplements the DeepPPMNet by assisting to model the seasonal trajectories of a time series.

Fig. 3 illustrates the overview of the proposed two training paradigms and the overall procedure of the DeepEMSFNet .

Refer to caption

(a) The proposed DS training paradigm used in the DeepPPMNet framework

Refer to caption

(b) The proposed SE training paradigm used in the DeepPPMNet framework
Figure 3: An overview of the proposed DeepPPMNet framework. A reseasonalisation phase is required for the post-processing layer of the DS approach, as the seasonal components are not included in the DeepPPMNet training phase. Whereas, the SE approach uses the pre-processed time series along with the seasonal components as external variables, and directly forecast all the components of the time series, thus a reseasonalisation phase is not required.

III-B3 DeepPPMNet Learning Scheme

Following the recommendations of Bandara et al. [20] and Hewamalage et al. [19], we implement the “LSTM Stacking Layer” design pattern in DeepPPMNet. Here, a stacking of LSTM layers is followed by an affine neural layer (excluding the bias component) to map each LSTM cell output to the dimension of the intended forecasting horizon. As we follow the MIMO approach, when generating the input and output windows, the dimension of this fully connected neural layer is equivalent to the size of the output window.

To avoid possible network saturation effects that occur in neural networks [20, 19], we perform a local normalisation process on the training windows, before using them for the DeepPPMNet training procedure. We implement two different normalisation strategies for each proposed training paradigm. The trend component of the last value of the input window is used in the DS approach, whereas in the SE approach, the mean value of each input window is used.

We implement the above training paradigms and learning schemes of the DeepPPMNet using TensorFlow, an open-source deep-learning toolkit [30]. The operating environment in EMS is often volatile and may present irregular demand patterns and anomalies. To be more robust under these circumstances, we use L1-norm as the primary learning objective function of DeepPPMNet. Moreover, we also include a L2-regularisation term to minimise possible overfitting of the network.

III-C Post-processing Layer

The post-processing layer of DeepPPMNet consists of two phases, namely the reseasonalisation phase and the renormalisation phase. Firstly, in the reseasonalisation phase, the seasonal components estimated by the Prophet method are added back to the forecasts generated by the LSTM. As discussed in Section III-B2, this phase is only required by the DS training paradigm, as the seasonal components of the time series are not included in the DeepPPMNet training procedure. Secondly, in the renormalisation phase, the corresponding forecasts are back-transformed into the original scale, by adding back the local normalisation factor used in each training window, taking the exponent of the values, and finally multiplying by the mean scaling factor of each time series that is used in the normalisation process. The renormalisation phase is required by both the DS and SE training paradigms, as the mean normalisation and log transformation are common processes for both training paradigms.

IV Experiments

In this section, we present the experimental setup used to evaluate the DeepPPMNet framework on real-world EMS datasets. This includes a summary of the datasets, error metrics, hyper-parameter selection, benchmarks and a discussion of the results obtained.

IV-A Datasets

We use an internationally unique, national dataset of coded ambulance clinical records held by Turning Point, an Australian addiction research and education centre. The dataset (The National Ambulance Surveillance System or NASS) [31] holds surveillance data on alcohol and other drug, self-harm and mental health-related ambulance attendances across 5 of the 6 Australian states and 2 territories. We subdivide the dataset into five categories of ambulance attendance datasets namely: 1) the alcohol overdose (AO) dataset, 2) the suicide-attempt (SA) dataset, 3) the heroin overdose (HO) dataset, 4) the methamphetamine overdose (MO) dataset, and 5) the prescription Opioids overdose (PO) dataset. For further definitions of these categories, we refer to [32].

We use 8 years worth of monthly EMS data available from September-2011 to May-2019. The last 12 months of EMS observations, i.e., May-2018 to May-2019, are reserved for testing. The size of the training output window is 12, as this is the intended forecasting horizon. Also, we choose the size of the training input window as 15 (12*1.25), following the heuristic proposed in [19, 16]. The number of LGAs, i.e., number of time series, present in the AO, SA, HO, MO and PO categories are 89, 88, 80, 87 and 87, respectively.

Fig. 4 shows violin plots generated using the seasonal components of each EMS category. This enables us to analyse how seasonal strength can affect the performance of the DeepPPMNet training paradigms, i.e., DS and SE approaches. We observe that the AO category carries a higher seasonal strength compared to other categories. This can be mainly attributed to the higher consumption of alcohol during the holiday periods.

Refer to caption

Figure 4: The yearly seasonal strength distributions of the AO, SA, HO, MO and PO categories. The seasonal components are extracted using Prophet as a decomposition technique.

IV-B Error Metrics

To assess the accuracy of the DeepPPMNet, we use two evaluation metrics commonly used in the time series forecasting literature, namely the symmetric Mean Absolute Percentage Error (sMAPE) and the Mean Absolute Scaled Error (MASE), which are scale independent error measures [33]. Using these measures to calculate the forecast errors independently of the scale of the time series leads to effectively every series contributing equally to the error, even though the LGA-wise EMS dataset may contain different scales of time series. This may be appropriate for certain applications but not for others. However, these measures are also commonly used in forecasting competitions nowadays, and can be considered as standard evaluation measures in forecasting. The sMAPE and MASE error measures are defined as follows:

sMAPE=2mt=1m(|FtYt||Ft|+|Yt|)\text{sMAPE}=\frac{2}{m}\sum_{t=1}^{m}\left(\frac{\left|F_{t}-Y_{t}\right|}{\left|F_{t}\right|+\left|Y_{t}\right|}\right) (2)
MASE=1mt=1m|FtYt|1nSt=S+1n|YtYtS|\text{MASE}=\frac{1}{m}\frac{\sum_{t=1}^{m}|F_{t}-Y_{t}|}{\frac{1}{n-S}\sum_{t=S+1}^{n}|Y_{t}-Y_{t-S}|} (3)

Here, FtF_{t} denotes the forecasts generated by DeepPPMNet, and YtY_{t} is the actual EMS observation at time tt. Furthermore, mm represents the number of data points in the test set and nn is the number of observations in the training set of a time series. SS refers to the frequency of the seasonality in a given time series, set to 12 in our experiments, as we use monthly EMS demand time series. The mean and median of these error measures are reported in our evaluation.

IV-C Hyper-parameter Tuning and Optimisation

A Python implementation of a Bayesian hyper-parameter optimisation process, the sequential model-based algorithm configuration (SMAC) [34] is used to autonomously determine the optimal hyper-parameters of DeepPPMNet. As we employ COntinuous COin Betting [35], tuning the learning rate is not necessary in DeepPPMNet. Table I summarises the bounds of the hyper-parameter values used in the DeepPPMNet learning process.

We also address the parameter uncertainty of our proposed framework by training DeepPPMNet on 10 different Tensorflow graph seeds, using the optimised hyperparameters from SMAC. We generate the final forecasts of DeepPPMNet by computing the median of the forecasts that are produced from different initialisation seeds of DeepPPMNet.

TABLE I: DeepPPMNet hyper-parameter grid
Model Parameter Minimum value Maximum value
LSTM-cell-dimension 20 50
Mini-batch-size 1 10
Epoch-size 2 10
Maximum-epochs 3 40
Hidden Layers 1 2
Gaussian-noise-injection 10410^{-4} 81048\cdot 10^{-4}
Random-Normal-Initialiser 10410^{-4} 81048\cdot 10^{-4}
L2-regularisation-weight 10410^{-4} 81048\cdot 10^{-4}

IV-D Benchmarks and DeepPPMNet variants

We use a collection of state-of-the-art univariate forecasting techniques to benchmark against. This includes well-established methods for monthly time series forecasting, such as ETS [36], ARIMA [37], and recently developed forecasting methods, such as BaggedETS [38], and Prophet [22]. Here, the ETS, ARIMA, and BaggedETS methods are used from the forecast package implemented in R [39], while the Prophet method is used from the prophet package in R.

When building GFMs, as the first approach, we train separate DeepPPMNet models on each EMS category (GFM-CAT). As the second approach, we build one GFM across all the available EMS category datasets (GFM-ALL). Based on the proposed training paradigms and the GFM modelling approaches, we define the following variants of DeepPPMNet.

  • DeepPPMNet-DS: The DeepPPMNet framework that uses DS as the training paradigm and a separate GFM is built on each EMS category.

  • DeepPPMNet-SE: The DeepPPMNet framework that uses SE as the training paradigm and a separate GFM is built on each EMS category.

  • DeepPPMNet-DS-ALL: The DeepPPMNet framework that uses DS as the training paradigm and one GFM is trained across all the EMS categories.

  • DeepPPMNet-SE-ALL: The DeepPPMNet framework that uses SE as the training paradigm and one GFM is trained across all the EMS categories.

IV-E Results and Discussion

Table II shows the evaluation summary of the DeepPPMNet variants and benchmarks for the 89 LGAs of the AO category, ordered by the Mean sMAPE. For each column, the results of the best performing methods are marked in boldface. According to Table II, we see that the proposed DeepPPMNet-DS-ALL variant achieves the best Mean sMAPE, Mean MASE, and Median sMAPE, while Prophet obtains the best Median MASE. Also, among the proposed variants, we observe that the DeepPPMNet variants with the DS training paradigm achieve better accuracies than the SE training paradigm variants. Furthermore, overall the GFM-ALL approach obtains better results than the GFM-CAT approach. Most importantly, we see that the DeepPPMNet-DS-ALL variant outperforms state-of-the-art methods, such as ETS, BaggedETS, and ARIMA consistently across all performance metrics.

TABLE II: AO Dataset Results
Method Mean sMAPE Median sMAPE Mean MASE Median MASE
DeepPPMNet-DS-ALL 0.1243 0.1251 0.9182 0.8959
DeepPPMNet-DS 0.1254 0.1318 0.9235 0.9119
Prophet 0.1281 0.1261 0.9585 0.8964
BaggedETS 0.1286 0.1299 0.9472 0.9466
ETS 0.1291 0.1284 0.9499 0.9568
ARIMA 0.1332 0.1351 0.9691 0.9762
DeepPPMNet-SE-ALL 0.1334 0.1392 0.9883 0.9755
DeepPPMNet-SE 0.1417 0.1435 1.0340 1.0289

Table III summarises the results for the evaluations on the 88 LGAs of the SA category. It can be seen that the proposed DeepPPMNet-SE variant outperforms all other methods consistently across all error measures. Most importantly, we observe that contrary to our previous findings from the AO dataset, the SE variants achieve better results than the DS variants. Also, among the SE variants, we see that training a separate GFM gives better results than building one GFM across all the available time series. However, the observations among the DS based variants are reciprocal to those of SE.

TABLE III: SA Dataset Results
Method Mean sMAPE Median sMAPE Mean MASE Median MASE
DeepPPMNet-SE 0.1128 0.1126 0.8817 0.8088
DeepPPMNet-SE-ALL 0.1131 0.1153 0.8835 0.8263
ETS 0.1146 0.1201 0.8893 0.8212
BaggedETS 0.1148 0.1193 0.8996 0.8299
ARIMA 0.1159 0.1247 0.8946 0.8281
Prophet 0.1206 0.1233 1.0174 0.8988
DeepPPMNet-DS-ALL 0.1231 0.1189 0.9547 0.8858
DeepPPMNet-DS 0.1266 0.1210 0.9652 0.9164

Table IV shows the evaluation summary for the 87 LGAs of the MO category. We observe that the BaggesETS method achieves the best accuracy in terms of Mean sMAPE, and the DeepPPMNet-SE-ALL variant obtains the best Median sMAPE. Meanwhile, the DeepPPMNet-SE variant achieves the best Mean MASE, while the DeepPPMNet-SE-ALL variant and the Prophet method are the best-performing methods with respect to Median MASE. Furthermore, similar to our previous findings from Table III, the DS variants show poor results compared to the SE variants, and overall the GFM-ALL approach obtains better results than its counterpart.

TABLE IV: MO Dataset Results
Method Mean sMAPE Median sMAPE Mean MASE Median MASE
BaggedETS 0.0917 0.0859 1.0751 0.9946
DeepPPMNet-SE-ALL 0.0919 0.0832 1.0739 0.9861
DeepPPMNet-SE 0.0927 0.0919 1.0703 0.9946
Prophet 0.0929 0.0951 1.1264 0.9861
ETS 0.0930 0.0912 1.0861 0.9946
ARIMA 0.0935 0.0845 1.0953 0.9946
DeepPPMNet-DS-ALL 0.0973 0.0983 1.1195 1.0703
DeepPPMNet-DS 0.1056 0.1043 1.1607 1.1221

Table V provides the results for the evaluations on the 80 LGAs of the HO category. Except for Median sMAPE, DeepPPMNet-SE-ALL outperforms all the benchmarks on each performance metric, whereas the proposed DeepPPMNet-DS achieves the best Median sMAPE. Similar to Table III and Table IV, variants based on the SE training paradigm achieve better accuracies compared to those of DS. Overall, the GFM-ALL approach obtains better results compared to GFM-CAT approach.

TABLE V: HO Dataset Results
Method Mean sMAPE Median sMAPE Mean MASE Median MASE
DeepPPMNet-SE-ALL 0.0780 0.0634 1.2118 0.9685
ETS 0.0786 0.0634 1.2254 0.9807
BaggedETS 0.0788 0.0598 1.2655 0.9989
DeepPPMNet-SE 0.0804 0.0634 1.2137 0.9925
DeepPPMNet-DS-ALL 0.0814 0.0610 1.2314 1.0964
ARIMA 0.0809 0.0654 1.2811 0.9989
Prophet 0.0813 0.0647 1.3476 1.0857
DeepPPMNet-DS 0.0825 0.0597 1.2433 1.0964

Table VI shows the results of the benchmarks for the 87 LGAs of the PO category. It can be seen that the proposed DeepPPMNet-SE variant achieves the best results on each performance metric. Once again, we observe that among the proposed training paradigms, the SE based variants perform better than the DS based variants. However, contrary to our previous findings, the DeepPPMNet variants that use the GFM-CAT approach, i.e., a separate DeepPPMNet is built only using the PO dataset, achieve better results compared to the GFM-ALL.

TABLE VI: PO Dataset Results
Method Mean sMAPE Median sMAPE Mean MASE Median MASE
DeepPPMNet-SE 0.0620 0.0528 0.8606 0.7759
ETS 0.0629 0.0639 0.8653 0.8068
BaggedETS 0.0634 0.0662 0.8803 0.8068
DeepPPMNet-DS 0.0669 0.0661 0.9366 0.8539
ARIMA 0.0645 0.0639 0.8821 0.8068
DeepPPMNet-SE-ALL 0.0664 0.0664 0.9071 0.8068
DeepPPMNet-DS-ALL 0.0679 0.0675 0.9640 0.8677
Prophet 0.0699 0.0699 1.0456 0.9297

In summary, we observe that the proposed DeepPPMNet variants achieve competitive results on the benchmark datasets, outperforming many state-of-the-art univariate forecasting techniques. This highlights the contribution made by employing a GFM strategy to exploit the similar EMS demand patterns. Moreover, we see that in the majority of cases, the GFM-ALL approach that uses all the available EMS category datasets performs better than the GFM-CAT approach that uses only the data specific to an EMS category. This can be mainly attributed to the low number of time series available in each EMS category, i.e., a maximum of 89 time series, for the AO category. On the other hand, grouping all the EMS categories together allows us to build a GFM across a total of 431 time series, and therewith allows our deep-learning based framework to learn better with more available data. It is also noteworthy to mention that the proposed DS training paradigm only outperforms the SE training paradigm in the AO dataset. This can be mainly attributed to the higher seasonal strength present in the AO dataset, compared to other EMS categories (see Fig. 4).

Refer to caption
(a) Type-1: 5% ALI sensitivity
Refer to caption
(b) Type-2: 5% ALI sensitivity
Refer to caption
(c) Type-1: 10% ALI sensitivity
Refer to caption
(d) Type-2: 10% ALI sensitivity
Figure 5: The application of ‘what-if’ scenario analysis, using the number of ALI (as a percentage of change) against the AO related EMS demand.

IV-F DeepPPMNet for Policy-making

The competitiveness of the DeepPPMNet forecasting framework allows us to work towards assessing causing factors of EMS demand. In the following, we demonstrate the use of the DeepPPMNet framework as a platform to perform GC analyses using external time series. As a use case, we investigate how the number of alcohol licenses issued (ALI) for a certain period of time can affect the AO related EMS demand patterns. This can enable government policy makers to understand effects of different numbers of ALI by conducting ‘what-if’ scenario analyses.

IV-F1 Experimental Setup

For this experiment, we use the monthly EMS data related to AO demand for 79 LGAs, and add the number of ALI as an exogenous variable to the current training input window. In particular, the DeepPPMNet uses a concatenation of the EMS demand and the ALI to learn the actual EMS demand of the output window. We consider the five year period from January-2015 to May-2019, as for this period we have both the EMS data and the corresponding ALI data available. The observations of the last 12 months are used to test the DeepPPMNet, the rest of the data are used for training. We use the DeepPPMNet in the GFM-CAT mode, i.e., the GFM is built only using the alcohol related data, assuming other datasets are not available. Also, in the decomposition layer, we replace the Prophet method with STL, as in this experiment the time series are shorter (42 monthly observiations), and STL is more robust and better suited for shorter series. In this way, two additional DeepPPMNet variants are introduced, DeepPPMNet-SE-License and DeepPPMNet-DS-License, to represent the models that use the number of ALI as an exogenous variable. Furthermore, analogously we use the ARIMA benchmark in the exogenous-regressive mode, using the number of ALI as an additional regressor (ARIMAX-License).

IV-F2 Accuracy Results

Table VII shows the results for the evaluations on the 79 LGAs of the AO category. Furthermore, we perform non-parametric Friedman rank-sum statistical significance testing to assess the statistical significance within the benchmarks, and the Hochberg’s post-hoc procedure is used to further examine these differences [40]. The statistical testing is carried out on the sMAPE error measure, with a significance level of α\alpha = 0.05. Table VIII shows the corresponding results of the statistical testing evaluation. A horizontal line is used to separate the methods that perform significantly worse than the best performing method. The overall result of the Friedman rank sum test gives a pp-value of 6.91×10116.91\times 10^{-11}, which means the differences are highly significant.

We see that DeepPPMNet-SE-License performs best, and achieves significantly better results than the DeepPPMNet-DS variants and the other benchmarks. However, we also observe that the DeepPPMNet-SE, BaggedETS, and ARIMAX-License variants do not perform significantly worse than the control method, DeepPPMNet-SE-License. The results confirm the earlier findings of the overall competitiveness of our methodology and the benefits of employing a GFM based forecasting model in this framework.

IV-F3 Effect of Number of ALI

Regarding the effect of the number of ALI, we see that after incorporating it as an exogenous variable (DeepPPMNet-SE-License), the accuracy of DeepPPMNet-SE has improved, outperforming the rest of the DeepPPMNet variants and the statistical benchmarks. We also observe that the ARIMAX-License variant has uplifted the results of the original ARIMA method, showing that inclusion of ALI as an exogenous variable is able to improve predictive accuracy over different methods. Furthermore, we perform a non-parametric paired Wilcoxon signed rank test to explore the statistical significance of the differences. The test indicates a pp-value of 0.687 for the DeepPPMNet-SE-License and DeepPPMNet-SE variants, and a pp-value of 0.704 for the ARIMAX-License and ARIMA variants, which indicates that the differences are not statistically significant.

As the results indicate that incorporating the number of ALI issued as an exogenous variable has improved the base accuracies of DeepPPMNet-SE and ARIMA, the number of ALI is a viable candidate for further exploration of a causal relationship. However, as the significance tests have not detected significant differences, further experiments with a larger dataset are needed in the future to clearly establish the relationship.

IV-F4 Discussion of Causality of the Observed Effect

As discussed in Section I, GC assesses an external regressor ZtZ_{t} with respect to other regressors WtW_{t}. We have established that the inclusion of ALI as ZtZ_{t} into our model is in fact able to improve forecasting accuracy across different methods. However, the strength of GC heavily depends on careful inclusion of additional factors WtW_{t}. In the current work, we suspect that ALI could work as a mere proxy to distinguish between rural and metropolitan areas, so, additional factors to be considered as WtW_{t} could be, e.g., the amount of supermarkets in a given area, the amount of cafes or other non-alcohol-serving outlets, the amount of GPs, and others. The proposed DeepPPMNet framework is a platform where conveniently all these factors can be included in the model, to carefully assess the relationships. However, we leave such a careful analysis for future work, and illustrate next how to perform a series of ‘what-if’ analyses on the AO related EMS demand once the GC analysis is done and the number of ALI is identified as a driving factor.

TABLE VII: AO Results with Alcohol Licenses
Method Mean sMAPE Median sMAPE Mean MASE Median MASE
DeepPPMNet-SE-License 0.1411 0.1334 0.9824 0.9031
DeepPPMNet-SE 0.1424 0.1414 1.0009 0.9964
BaggedETS 0.1476 0.1403 1.0501 1.0000
ARIMAX-License 0.1494 0.1485 1.0472 1.0116
ARIMA 0.1507 0.1462 1.0518 0.9879
ETS 0.1505 0.1447 1.0636 1.0121
DeepPPMNet-DS 0.1528 0.1467 1.0635 0.9667
DeepPPMNet-DS-License 0.1539 0.1478 1.0703 1.0175
Prophet 0.1870 0.1862 1.2522 1.2083
TABLE VIII: Significance testing for AO Results
Method pHochp_{Hoch}
DeepPPMNet-SE-License -
DeepPPMNet-SE 0.559
BaggedETS 0.217
ARIMAX-License 0.089
ETS 0.038
ARIMA 0.035
DeepPPMNet-DS 0.006
DeepPPMNet-DS-License 1.729 ×\times 10410^{-4}
Prophet 2.254 ×\times 101510^{-15}

IV-F5 ‘What-if’ analysis

Fig. 5 illustrates an example of performing a ‘what-if’ scenario analysis to determine the sensitivity of ALI towards the AO specific EMS demand. Here, a 5% increase/decrease (Fig. 5a and Fig. 5b) and a 10% increase/decrease (Fig. 5c and Fig. 5d) is applied to the current number of ALI. We present two different types of demand variations present in two different LGAs. Fig. 5a represents the 5% increase/decrease effect of the ALI on LGA-1, while Fig. 5c shows the 10% increase/decrease effect of the ALI on LGA-1. We see that the increase of ALI has increased the predicted EMS demand, while a decrease in ALI has decreased EMS demand. Likewise, Fig. 5b and Fig. 5d show the 5% increase/decrease and the 10% increase/decrease effect of ALI on LGA-2. In this case, though decreasing the ALI has reduced the predicted EMS demand related to AO, the increase of ALI has not substantially affected the predicted EMS demand. Such insights may help government policy makers, when taking strategic decisions about the number of alcohol licenses to be issued for a specific LGA. Likewise, the DeepPPMNet can be used to conduct GC and ‘what-if’ analyses on other EMS categories. For example, the number of pharmacy licenses issued could be explored as a causal factor that affects PO related EMS demand.

V Conclusion

Overcoming the limitations of the current univariate state-of-the-art models, we have introduced DeepPPMNet, a three-layered, unified global forecasting framework that is capable of exploiting the non-trivial demand relationships present in EMS demand time series. We have evaluated DeepPPMNet using a real-world EMS dataset. The results have shown that the proposed DeepPPMNet framework has outperformed state-of-the-art univariate forecasting techniques, indicating the benefit of exploiting cross-series information available in the EMS demand time series.

Furthermore, we have demonstrated how DeepPPMNet can be employed straightforwardly for causal analyses using the notion of Granger Causality, and how we subsequently can perform ‘what-if’ scenario analyses to assist government authorities in the policy-making process.

As future work, we want to include further external factors in the analysis around alcohol outlet licenses, and we want to use DeepPPMNet on related cases, to demonstrate and make use of the full potential of our framework. From a methodologic point of view, we will enable the framework to perform probabilistic forecasting, for an even more informed policy making process.

VI Acknowledgements

This research was supported by the Australian Research Council under grant DE190100045, by a Facebook Statistics for Improving Insights and Decisions research award, and by Monash Institute of Medical Engineering seed funding. We acknowledge Sharon Matthews and the research assistants who code the NASS data and the ambulance services and paramedics who create and provide that data.

References

  • [1] K. Steins, N. Matinrad, and T. Andersson Granberg, “Forecasting the demand for emergency medical services,” in 52nd Hawaii International Conference on System Sciences.   University of Hawai’i at Manoa, 2019, pp. 1855–1864.
  • [2] Z. Zhou and D. S. Matteson, “Predicting ambulance demand: A Spatio-Temporal kernel approach,” in Proceedings of the 21th ACM SIGKDD, ser. KDD ’15.   New York, NY, USA: ACM, 2015, pp. 2297–2303.
  • [3] H. Setzler, C. Saydam, and S. Park, “EMS call volume predictions: A comparative study,” Comput. Oper. Res., vol. 36, no. 6, pp. 1843–1851, Jun. 2009.
  • [4] N. Channouf, P. Ecuyer, A. Ingolfsson, and A. N. Avramidis, “The application of forecasting techniques to modeling emergency medical system calls in calgary, alberta,” Health Care Manag. Sci., vol. 10, no. 1, pp. 25–45, Feb. 2007.
  • [5] A. Y. Chen, T. Lu, M. H. Ma, and W. Sun, “Demand forecast using data analytics for the preallocation of ambulances,” IEEE Journal of Biomedical and Health Informatics, vol. 20, no. 4, pp. 1178–1187, 2016.
  • [6] H.-T. Wong and P.-C. Lai, “Weather factors in the short-term forecasting of daily ambulance calls,” Int. J. Biometeorol., vol. 58, no. 5, pp. 669–678, Jul. 2014.
  • [7] D. S. Matteson, M. W. McLean, D. B. Woodard, and S. G. Henderson, “Forecasting emergency medical service call arrival rates,” Jul. 2011.
  • [8] J. L. Vile, J. W. Gillard, P. R. Harper, and V. A. Knight, “Predicting ambulance demand using singular spectrum analysis,” J. Oper. Res. Soc., vol. 63, no. 11, pp. 1556–1565, Nov. 2012.
  • [9] S. S. Jones, R. S. Evans, T. L. Allen, A. Thomas, P. J. Haug, S. J. Welch, and G. L. Snow, “A multivariate time series approach to modeling and forecasting demand in the emergency department,” J. Biomed. Inform., vol. 42, no. 1, pp. 123–139, Feb. 2009.
  • [10] T. Januschowski, J. Gasthaus, Y. Wang, D. Salinas, V. Flunkert, M. Bohlke-Schneider, and L. Callot, “Criteria for classifying forecasting methods,” Int. J. Forecast., vol. 36, no. 1, pp. 167–177, Jan. 2020.
  • [11] C. W. J. Granger, “Testing for causality: A personal viewpoint,” J. Econ. Dyn. Control, vol. 2, pp. 329–352, Jan. 1980.
  • [12] A. Seth, “Granger causality,” Scholarpedia J., vol. 2, no. 7, p. 1667, 2007.
  • [13] J. R. Baker and K. E. Fitzpatrick, “Determination of an optimal forecast model for ambulance demand using goal programming,” J. Oper. Res. Soc., vol. 37, no. 11, pp. 1047–1059, Nov. 1986.
  • [14] L. H. Brown, E. B. Lerner, B. Larmon, T. LeGassick, and M. Taigman, “Are EMS call volume predictions based on demand pattern analysis accurate?” Prehosp. Emerg. Care, vol. 11, no. 2, pp. 199–203, Apr. 2007.
  • [15] S. Smyl, “A hybrid method of exponential smoothing and recurrent neural networks for time series forecasting,” Int. J. Forecast., Jul. 2019.
  • [16] K. Bandara, P. Shi, C. Bergmeir, H. Hewamalage, Q. Tran, and B. Seaman, “Sales demand forecast in e-commerce using a long Short-Term memory neural network methodology,” Jan. 2019.
  • [17] V. Flunkert, D. Salinas, and J. Gasthaus, “DeepAR: Probabilistic forecasting with autoregressive recurrent networks,” Apr. 2017.
  • [18] K. Bandara, P. Shi, C. Bergmeir, H. Hewamalage, Q. Tran, and B. Seaman, “Sales demand forecast in e-commerce using a long Short-Term memory neural network methodology,” in Neural Information Processing.   Springer International Publishing, 2019, pp. 462–474.
  • [19] H. Hewamalage, C. Bergmeir, and K. Bandara, “Recurrent neural networks for time series forecasting: Current status and future directions,” arXiv [cs.LG], 2019.
  • [20] K. Bandara, C. Bergmeir, and S. Smyl, “Forecasting across time series databases using recurrent neural networks on groups of similar series: A clustering approach,” Expert Syst. Appl., vol. (in press), 2019.
  • [21] K. Bandara, C. Bergmeir, and H. Hewamalage, “LSTM-MSNet: Leveraging forecasts on sets of related time series with multiple seasonal patterns,” arXiv preprint arXiv:1909.04293, 2019.
  • [22] S. J. Taylor and B. Letham, “Forecasting at scale,” PeerJ Preprints, Tech. Rep. e3190v2, Sep. 2017.
  • [23] S. Taylor and B. Letham, “prophet: Automatic forecasting procedure,” 2018.
  • [24] R. B. Cleveland, W. S. Cleveland, and I. Terpenning, “STL: A seasonal-trend decomposition procedure based on loess,” J. Off. Stat., vol. 6, no. 1, p. 3, 1990.
  • [25] S. Hochreiter, “Untersuchungen zu dynamischen Neuronalen Netzen”,” Diploma, Technische Universität München, vol. 91, 1991.
  • [26] J. L. Elman, “Finding structure in time,” Cogn. Sci., vol. 14, no. 2, pp. 179–211, Mar. 1990.
  • [27] K. Cho, B. van Merrienboer, C. Gulcehre, F. Bougares, H. Schwenk, and Y. Bengio, “Learning phrase representations using RNN encoder-decoder for statistical machine translation,” in Conference on Empirical Methods in Natural Language Processing (EMNLP 2014), 2014.
  • [28] S. Ben Taieb, G. Bontempi, A. F. Atiya, and A. Sorjamaa, “A review and comparison of strategies for multi-step ahead time series forecasting based on the NN5 forecasting competition,” Expert Syst. Appl., vol. 39, no. 8, pp. 7067–7083, Jun. 2012.
  • [29] R. E. Schapire, “The boosting approach to machine learning: An overview,” in Nonlinear Estimation and Classification, D. D. Denison, M. H. Hansen, C. C. Holmes, B. Mallick, and B. Yu, Eds.   New York, NY: Springer New York, 2003, pp. 149–171.
  • [30] M. Abadi, A. Agarwal, P. Barham, E. Brevdo, Chen, and et al., “TensorFlow: Large-Scale machine learning on heterogeneous distributed systems,” 2016.
  • [31] Dan I. Lubman, Sharon Matthews, Cherie Heilbronn, Jessica J. Killian Rowan P. Ogeil, and et al., “The national ambulance surveillance system: A novel method surveillance system for monitoring acute alcohol, illicit and pharmaceutical drug-related harms using coded australian ambulance clinical records,” PLoS One, 2020.
  • [32] Self Harm and Mental Health Related Ambulance Attendances in Australia.   Turning Point : Eastern Health, 2013.
  • [33] R. J. Hyndman and A. B. Koehler, “Another look at measures of forecast accuracy,” Int. J. Forecast., 2006.
  • [34] F. Hutter, H. H. Hoos, and K. Leyton-Brown, “Sequential model-based optimization for general algorithm configuration,” in Proceedings of the 5th International Conference on Learning and Intelligent Optimization, ser. LION’05.   Berlin, Heidelberg: Springer-Verlag, 2011, pp. 507–523.
  • [35] F. Orabona and T. Tommasi, “Training deep networks without learning rates through coin betting,” May 2017.
  • [36] R. Hyndman, A. B. Koehler, J. Keith Ord, and R. D. Snyder, Forecasting with Exponential Smoothing: The State Space Approach.   Springer Science & Business Media, 2008.
  • [37] G. E. P. Box, G. M. Jenkins, G. C. Reinsel, and G. M. Ljung, Time Series Analysis: Forecasting and Control.   John Wiley & Sons, 2015.
  • [38] C. Bergmeir, R. J. Hyndman, and J. M. Benítez, “Bagging exponential smoothing methods using STL decomposition and Box–Cox transformation,” Int. J. Forecast., vol. 32, no. 2, pp. 303–312, 2016.
  • [39] R. J. Hyndman, G. Athanasopoulos, S. Razbash, D. Schmidt, Z. Zhou, Y. Khan, C. Bergmeir, and E. Wang, “forecast: Forecasting functions for time series and linear models,” R package version, vol. 6, no. 6, p. 7, 2015.
  • [40] S. García, A. Fernández, J. Luengo, and F. Herrera, “Advanced nonparametric tests for multiple comparisons in the design of experiments in computational intelligence and data mining: Experimental analysis of power,” Inf. Sci., vol. 180, no. 10, pp. 2044–2064, May 2010.