RNN-BOF: A Multivariate Global Recurrent Neural Network for Binary Outcome Forecasting of Inpatient Aggression
††thanks: This research was supported by the Australian Research Council under grant DE190100045.
Abstract
Psychometric assessment instruments aid clinicians by providing methods of assessing the future risk of adverse events such as aggression. Existing machine learning approaches have treated this as a classification problem, predicting the probability of an adverse event in a fixed future time period from the scores produced by both psychometric instruments and clinical and demographic covariates. We instead propose modelling a patient’s future risk using a time series methodology that learns from longitudinal data and produces a probabilistic binary forecast that indicates the presence of the adverse event in the next time period. Based on the recent success of Deep Neural Nets for globally forecasting across many time series, we introduce a global multivariate Recurrent Neural Network for Binary Outcome Forecasting, that trains from and for a population of patient time series to produce individual probabilistic risk assessments. We use a moving window training scheme on a real world dataset of 83 patients, where the main binary time series represents the presence of aggressive events and covariate time series represent clinical or demographic features and psychometric measures. On this dataset our approach was capable of a significant performance increase against both benchmark psychometric instruments and previously used machine learning methodologies.
Index Terms:
deep learning, time series, clinical psychology, risk assessment.I Introduction
Psychometric risk assessment instruments offer a structured measure that can aid clinicians in management of their patients by indicating the future risk of an adverse event such as aggression or violence. Tests such as the Dynamic Appraisal of Situational Aggression (DASA) [1] and Short-Term Assessment of Risk and Treatability (START) [2] produce scores on an ordinal scale based on the presence of multiple risk domains, which in turn correspond to the likelihood of an adverse event occurring. Probabilistic machine learning classifiers have been built from these instruments and additional clinical and demographic covariates [3, 4, 5], although this approach has no explicit way of learning temporal relationships that may occur between recurrent assessments or in the history of adverse events. As such, we propose a methodology for modelling this type of risk assessment as a probabilistic global multivariable time series forecasting task.
A local multivariable forecasting approach would produce a per-patient model from a binary time series representing the adverse event and covariate series representing clinical and demographic features. Instead, we train a single global forecaster for a population of patients allowing for cross learning between patients, more complex modelling methods, and no requirement for homogeneity between patients’ time series [6] or an instance of the adverse event within the patient’s history. For this task, we propose a sliding window time series based modelling scheme and a Recurrent Neural Network architecture for Binary Outcome Forecasting, RNN-BOF. An implementation of our method is available online111Implementation available at https://github.com/aidanpcquinn/RNNBOF. Unlike other recent global multivariable forecasters [7, 8], our model is capable of producing probabilities from a main binary time series, allowing us to build a generalised and time aware risk assessment tool.
RNN-BOF uses stacked Long-Short Term Memory cells (LSTM) constructed as a many-to-one forecaster with a single neuron sigmoid activation output layer, an adaptation from other RNN forecasters [9] to allow for the probabilistic and binary day-ahead outcome prediction. The model assigns any given patient a probability associated with the presence of the adverse event occurring in the next time interval from fixed length input windows. We compare this approach on a real world dataset against both the psychometric instruments themselves, and the state-of-art machine learning classifiers as used for clinical risk assessment.
We evaluate RNN-BOF on the Thomas Embling Hospital risk assessment dataset (TEH dataset) which contains daily recordings indicating the presence of an aggressive event, daily psychometric measures using the START and DASA instruments, and other static and dynamic clinical and demographic features. We then evaluate forecasting performance using the Precision Recall Gain curve (PRG) [10], a scoring rule that we propose is best suited to assess a model’s performance.
The main contributions of our work are as follows: (A) We present RNN-BOF, a novel implementation of an RNN as a scored global multivariable binary time series forecaster using a pooled sliding window training scheme. (B) We show experimental evidence that our methodology is capable of outperforming benchmark psychometric instruments and traditional machine learning models on the task of predicting the presence of inpatient aggression on a real world dataset. (C) Furthermore, our evidence suggests that longitudinal data which represents the presence of an event, alongside covariate data and categorical features, could benefit from being modelled as a global time series problem, even when a probability is required.
II Background and related works
In this section we briefly describe related works in psychometric evaluation and time series forecasting.
II-A Psychometrics and machine learning
We have identified two studies that directly employed psychometric instruments and clinical covariates alongside machine learning techniques. Galatzer-Levy et al. [3] tested the efficacy of various models including Support Vector Machines (SVM), Random Forests (RF) and kernel ridge regressions trained to predict the presence of a traumatic stress disorder diagnosis within 15 months of a traumatic event. The models were trained on patient data collected at the time of the event, general demographic data, physiological responses, and an assessment on the Acute Stress Disorder scale psychometric instrument. Their approach was able to outperform the benchmark predictive power of the scale itself. A similar work [4] assessed high-risk psychiatric inpatients using the Suicide Crisis Inventory instrument which is composed of 49 five-point ordinal scales. They trained models including RFs and gradient boosted decision trees (GBDT) to predict the occurrence of suicide attempts over a one month period although they made no comparison to the baseline psychometric evaluation.
II-B Global multivariable time series forecasting
Statistical local forecasting methods have consistently out-performed neural networks and other machine learning techniques in forecasting competitions for decades [11]. This changed only relatively recently in the M4 forecasting competition, which was won by a hybrid model that employed both statistical and neural network forecasting, training models globally across all time series [12]. Following this success and a subsequent rise in popularity, Montero-Manso and Hyndman [6] established theoretical bounds of effectiveness for global cross-series approaches as compared to local (per series) methods, suggested that the multivariable equivalent would inherit the beneficial properties of global cross-series methods, and offered empirical evidence for the lack of necessity of homogeneity across series. Hewamalage et al. [13, 14] supported these conclusions and identified RNNs with LSTM cells (RNN-LSTM) and Light Gradient Boosted Machines (LGBM) as strong performers on both real world and simulated datasets.
However, to the best of our knowledge there is limited literature on multivariable time series forecasting where models are trained globally across many homogeneous multivariable datasets. The two multivariable global forecasters we have identified are DeepGLO [7] and TEDGE [8]. Both of these models learn from a set of time series, where each series has covariate time series and static categorical features. Both use initial matrix factorisation to extract global trends; DeepGLO then uses a temporal convolution network to learn local patterns, while TEDGE makes use of a multivariable RNN-LSTM architecture. Either of these forecasters could suit our needs if not for the probabilistic or scored binary forecast requirement.
II-C Binary time series forecasters
The majority of models for binary time series forecasting are from the auto-regressive family, with neural network approaches being scarce [15]. gbARMA [16], GLARMA [17] and PC-ARIMA [18] are capable forecasters given a multivariate forecasting problem with a binary target series. However, they are all locally trained models [7], meaning for a patient without any instances of the adverse event, the model would have no historical reference and therefore could not predict future instances. Bautu et al. [15] used a probabilistic hyper-network to predict the direction of a stock price. Although theoretically globally trainable, the model’s implementation is univariate, making it again unsuitable for our dataset.
III Methodology
In this section we formally outline the learning task and compare this to previous formulations of similar tasks. We then describe the proposed model architecture, reasoning for the structure and components, details of the window training scheme, and the learning algorithm.
III-A Learning task
We first define the learning tasks as used in supervised learning classification works with similar feature sets [3] and similar learning tasks [4] to the TEH dataset. For a population of patients, , we have the dataset with features , where each can be binary, , the value of some bound ordinal scale , , or a value of continuous measurement, . Examples of are general clinical data such as ‘history of drug use’ or an outcome on a specific psychometric measure, such as ‘irritability’ on the DASA assessment [1]. Each could represent the summation of the binary measure components of an assessment such as DASA, or a score on a Likert scale [19], while continuous measures, , could refer to age or blood pressure at time of assessment. Each patient would have associated label, , representing the presence of a future event within a fixed future time frame. From this data we would learn a model that predicts a score or probability for the positive outcome (presence of the event) within the fixed time frame for any given patient, .
The TEH dataset consists of daily longitudinal psychometric and clinical data with a feature set, , that exhibits the same feature types as those described in , alongside a daily recording representing the presence of an instance of aggression over the same daily periods. We propose modelling the future presence of aggression using a global multivariable binary time series forecaster, where the target series is a patient’s binary time series indicating presence of aggression and associated covariate time series are features that vary over time. The task is, for any given patient, to forecast a score or probability associated with the presence of aggression in their next time interval. We note that the TEH dataset does not contain equal history length for each of the patients, contains static features that do not vary over time and is relatively small with only 83 patients total and a mean series length of 140 with a standard deviation of 106.
We formally define the new learning task as follows: For a patient , we represent the time series of the history of a dynamic feature, as the vector where is the total series length of a patient, and is a dynamic feature . We then define the set of all of a given patient’s dynamic feature time series as . The static features for a patient, , , are represented as the set . The specific adverse event for a patient is represented by a binary vector , where indicates the presence of the event. The complete dataset then contains the sets: , and . The objective is to learn a model that takes a fixed length window of length from a given patient to produce a probability associated with the future positive patient outcome, namely , where . We note that when only a length of is considered by the model, this learning task matches the traditional future classification style approach with a bound of one day placed on the future time period of when the event can occur.
III-B RNN-BOF architecture
RNN-BOF uses a stacked Recurrent Neural Network architecture similar to that described by Bandara et al. [20], but with only the last LSTM cell’s hidden state vector from the last stacked layer being fully connected to a single sigmoid output neuron for the single probabilistic binary forecast point (Fig. 1). For input data, we concatenate the dynamic, static, and adverse event series elements per patient per time step, so an input from time is defined as , where denotes vector concatenation and where . For a fixed input window size of , a stacked RNN architecture passes hidden state outputs from the layer , , as inputs to the next layer , where . For our RNN-BOF with stacked layers, the output corresponds with the hidden state vector represented by , which is fully connected to a single neuron with sigmoid activation for learning the binary probability. This differs from other multivariate time series forecasters [20] where the last layer is instead fully connected to the output window for the learning of many continuous values. This also differs from stacked RNNs as implemented for binary classification tasks, where all hidden states from the final layer, , would be flattened and connected to a dense layer before again being connected to a single dense neuron with sigmoid activation. The output of the final neuron is defined by the equation: , where is the learned weight vector, and is the learned bias and is the learned probability that the next day in the adverse event series of a given patient belongs to the positive class.

We make use of two techniques to deal with potential overfitting on small datasets. Dropout layers [21] are added after each LSTM layer (Fig. 1) which set values in the hidden layer output vectors, , to 0 at a fixed rate, and L2 kernel regularisation which has been used in RNNs as a tool to reduce overfitting [22] calculated by , where is a specific hidden weight during back propagation and is a learned hyper-parameter.
III-C Sliding window global training
We train our model using a supervised learning sliding window training scheme, again similar to that described by Bandara et al. [20], but where the forecasting point is only the next value in the series, which we will refer to as the label associated with any given input window. We employ this training scheme as it completely maximises use of the relatively small dataset. For each patient, , we define their ordered set of -length dynamic feature windows as , similarly, their adverse event windows are defined as , and static covariates are still the set . Given windows, w, for patient p, and have associated label where . We can then define the label vector associated with , and as (Fig. 2).
III-D Training and evaluation sets
For training and evaluation purposes we split each individual patient’s windows sequentially into the train and evaluation sets instead of using a held-out patient group. This represents the real world case where a forecaster is implemented on a group of series to predict future values of those series and is common practice in global time series forecasting evaluation [9, 20]. We define this for a 80-20 train test split by taking the last windows per patient, where so that across different window lengths , the set of test windows will match exactly. The train and test window sets are then defined as , , with associated labels, and , , with associated labels, . Thus, we maintain temporal integrity as no window labels are shared across the sets.
III-E Loss function and learning algorithm
Unlike time series forecasting problems with continuous outputs where loss functions such as L2 or L1 are used to learn model weights, we use weighted binary cross entropy batch loss :
(1) |
where is the size of the batch, and weight, is the ratio of positive to negative days in the training set, and the opposite. For each batch, the model’s weights are updated via the Adam learning algorithm. This learning scheme closely represents that of a standard supervised learning classification problem.
IV Experiments
In this section we detail our experimental procedure for evaluating the performance of RNN-BOF as compared to benchmarks.
IV-A Dataset
We evaluate RNN-BOF on the Thomas Embling Hospital risk assessment dataset (TEH dataset) provided by Swinburne University of Technology333Ethics approval granted by Swinburne University of Technology (Project #1258). The dataset contains 40 static demographic and general clinical features, 73 features that vary daily or over longer periods, and daily recordings indicating the presence of an aggressive incident. The features include binary representations of general clinical data such as ‘history of drug use’, all components of START and DASA assessments, ordinal summations of both measures as well as general demographic data such as age.
IV-B Pre-Processing
We categorised the following three types of missing values found in the dataset: all patients had entries where the collection of dynamic features, psychometric assessments and presence of aggressive incidents spanned more than 24 hours; some patients had components of or full psychometric measures missing for some days; and a small subset were missing the label for presence of aggressive events. The frequency and length of the data entries lasting more than 24 hours were greater at the start and end of a given patient’s admission, with one or two long spans of daily assessments. We defined these long frequent assessment periods as a sequence of at least 30 days, starting and ending with 5 daily entries, and with an average of 1.1 days or less between entries over the entire period. As the contents of assessments lasting more than 24 hours were still accurate, no imputation or duplication was performed but a flag was added. For the dynamic missing values, nearest value linear interpolation was used and another flag was added, while patients that were missing values in the aggression series were dropped entirely. We performed z-score standardisation calculated per feature across the entire population’s training set. The cleaned frequent assessment periods were then windowed and split 80-20 into train and test window sets as described in Section III-C. Descriptive statistics for the dataset after pre-processing can be seen in Table I.
Description | Value |
---|---|
Windows in train and test set respectively | 9343 / 2296 |
Windows with aggression labels in train and test set | 545 / 122 |
Avg. of series length | 140.23 |
Std. Dev. of series length | 105.78 |
Avg. of positive labeled windows per patient | 8.03 |
Std. Dev. of positive labeled windows per patient | 20.20 |
Static / dynamic feature counts | 40 / 74 |
IV-C Evaluation metric
Previous works in the clinical psychology domain have commonly evaluated the predictive performance of their psychometric instruments using the ROC curve and its respective area [4, 5]. The ROC curve is not attenuated by imbalanced datasets such as ours and as such optimistically evaluates performance on the minority class, masking poorly performing models [23]. We instead use the Precision Recall Gain curve (PRG) [10], a variation on the Precision Recall (PR) curve which has been shown to be a stronger predictive evaluator than the ROC on imbalanced datasets [23]. While representing the same relationship between precision and recall, the Area Under the PRG Curve (AUC-PRG) is valid under cross-validation averaging and the curve does not require non-linear interpolation between points. The AUC-PRG is defined by the following equation:
(2) |
Here, FP, FN, TP are false positive, false negative, and true positive, respectively, and
(3) |
IV-D Benchmark models
We chose benchmark models that have seen use in clinical psychology machine learning papers as probabilistic or scored predictive classifiers. This list includes Random Forests (RF) [4, 5], Support Vector Machines (SVM) [3], Feed Forward Neural Networks (FFNN) [5] and logistic regression [4, 5]. We also included Light Gradient Boosted Machines (LGBM) due to their probabilistic output capability and for being strong global time series forecasters [13]. All benchmark models were trained using the same global sliding window training scheme across all patients as described in our modelling approach (Section III-C). The benchmarks were trained and tested using the same input window lengths of 10 and 20 days and therefore are also acting as binary times series forecasters. An additional one day length input window was also tested, representing the traditional classification approach described in Section III-A.
For psychometric instrument benchmarks we used the two measures collected in the TEH dataset; the DASA assessment [1] and the vulnerability component of START assessment [2]. The DASA total score is calculated by summing seven individual binary scored components, each representing the presence of a risk indicating behavior. The START vulnerability assessment contains 20 components, each scored 0, 1 or 2 on a scale representing vulnerability that is summed to give a score between 0 and 40.
IV-E Overall procedure
The complete set of windows were split sequentially 80-20 per patient, then pooled as the train and test set respectively (Section III-C). For each model’s hyper-parameter tuning we used 5-fold expanding sequential cross-validation on the pool of patients’ training windows [24], with folds instead of single instances and pooling across a population. The cross-validation procedure splits the training windows per patient into six sequential and equal size window pools. A validation model is then trained on the first pool and validated on the second, and a second validation model is then trained on the first two pools and validated on the third and so on. This still maintains the temporal integrity described by Hyndman and Athanasopoulos [24] as no labels are shared between train and validation folds. The AUC-PRG is averaged across validation folds to inform a Bayesian hyper-parameter tuning algorithm (namely the Hyperopt algorithm, for 100 iterations). After parameter selection, each model is trained on the full training set [9] with forecasts then performed on the held-out test windows for final performance evaluation.

IV-F Hyper-parameters
We chose the hyper-parameters and their ranges through existing literature [21, 22, 9] and manual exploration on the training set cross-validation folds. The tuned parameters and their respective ranges identified for RNN-BOF were: number of hidden nodes per layer, 10 - 100; number of stacked layers, 1-4; dropout percent, 10% - 50%; L2 regularisation, 0.000001 - 0.001; and number of epochs, 10 - 100. We used a fixed batch size of 512 and found that too high of a learning rate led to especially poor performance so we fixed it at 0.0001. We note that all benchmarks except the FFNN showed worse performance with the weighted loss despite the imbalanced nature of the output windows, so standard binary cross entropy loss was used.
V Results and discussion
In this section we provide a comparison of all tested models and psychometrics instruments, as well as interpretation and discussion of these results. The test set contains a total of windows with having a positive label (and in the training window set, Table I). As described in Section III-C, regardless of input window length, windows contain the same forecast labels which allows us to make like-to-like comparisons. We calculated confidence intervals for the AUC-PRG using stratified moving block bootstrapping (the requirement of normality was validated using the Kolmogorov-Smirnov Test) [25, 26] on the complete set of test window predictions. We compare the performance of models directly by calculating the bootstrapped distribution of differences in AUC-PRG across the same samples, using the null hypothesis that our model performs equal or worse we can get a -value indicating a significant increase of AUC-PRG. In Table II we provide all models’ AUC-PRG with associated 95% confidence intervals, along with the difference between AUC-PRG of RNN-BOF with 10 day input windows (RNN-BOF10) and all other models where -values indicate a bootstrapped significant difference in area (B=2000). RNN-BOF10 is the best performing model with an AUC of on the test set and an increase in AUC-PRG against all benchmarks at the significance level.
Modela | AUC-PRG | AUC-PRG CI 95% | AUC-PRG Dif. | -valueb |
RNN-BOF10 | 0.9761 | [0.9535,0.9852] | - | - |
RNN-BOF20 | 0.9664 | [0.9428,0.9815] | 0.0097 | |
FFNN20 | 0.9602 | [0.9336,0.9781] | 0.0159 | |
FFNN10 | 0.9599 | [0.9324,0.9777] | 0.0162 | |
FFNN1 | 0.9598 | [0.9265,0.9771] | 0.0164 | |
RF1 | 0.9566 | [0.9312,0.9745] | 0.0195 | |
RF10 | 0.9559 | [0.9241,0.9754] | 0.0202 | |
LGBM20 | 0.9518 | [0.9191,0.9734] | 0.0243 | |
RF20 | 0.9489 | [0.9150,0.9703] | 0.0272 | |
LGBM10 | 0.9477 | [0.9180,0.9694] | 0.0284 | |
LR1 | 0.9441 | [0.9088,0.9705] | 0.0320 | |
LGBM1 | 0.9394 | [0.6599,0.7872] | 0.0367 | |
SVM20 | 0.9072 | [0.8335,0.9457] | 0.0689 | |
SVM10 | 0.878 | [0.8016,0.9263] | 0.0981 | |
SVM1 | 0.8733 | [0.7781,0.9233] | 0.1028 | |
DASA | 0.8679 | [0.7733,0.9203] | 0.1082 | |
LR10 | 0.8542 | [0.7595,0.9187] | 0.1219 | |
LR20 | 0.7895 | [0.6828,0.8694] | 0.1866 | |
STARTVuln | 0.7264 | [0.6599,0.7872] | 0.2497 | |
aNumber after model name indicates input window length. | ||||
bCalculated on stratified moving block bootstrapped samples with B = 2000 (a horizontal line indicates the significance threshold of 0.05). |
Unlike for ROC curves where the main diagonal, , is the baseline random classifier (AUC-ROC = 0.5), for PRG the equivalent diagonal, , represents the always positive classifier [10]. Subsequently, for PRG analysis of a dataset with low amounts of positive labels the values of precG and recG will be higher due to a worse performing baseline. The drawback of this can be that differences at high precision and recall levels on an imbalanced dataset are harder to determine and indicate a bigger difference than they would for the equivalent ROC and PR curves.
We provide three comparison PRG curves to better understand where our model gains its performance increase. In Fig. 3A we compare RNN-BOF10 with the psychometric instruments START and DASA where we can see that the majority of the difference in area comes from the top-rightmost region, and high precision gain values. In practical terms the higher precision gain values indicate our model can determine a higher number of true positives from false positives for the same amount of false negatives. The points in the top-rightmost area represent the common threshold calibration use case where precision is roughly equal to recall and points on the line parallel to the always positive classifier and that are closest to precG = 1 and recG = 1 represent peak calibrated scores for the model [10]. The difference between RNNBOF10 and the best performing benchmark machine learning model, FFNN with 20 day input windows (FFNN20) (Fig. 3B), are much smaller, with only slight gaps at high precision gain levels and in the top-rightmost region, although these regions still result in a significant performance improvement.


VI Conclusion
Our research builds upon previous machine learning works in both the field of clinical risk assessment and global time series forecasting. We presented a novel approach to clinical risk assessment, specifically for predicting instances of inpatient aggression that leveraged daily longitudinal data collection. We achieved this using RNN-BOF, which unlike previous global multivariable approaches is capable of probabilistic forecasts on binary time series. Our model uses a modified multivariable Recurrent Neural Network with a single probabilistic output value for the estimation of next day patient risk, differing architecturally from RNNs as used for multivariable classification and forecasting tasks. The model was globally trained on a real world dataset across sequentially constructed windows of all patients consisting of: the target binary time series, representing aggressive incidents; covariate series, representing psychometric measurements and dynamic clinical features; and static clinical and demographic features. Unlike the benchmark models, RNN-BOF had an explicit understanding of the sequential nature of the data, allowing for the potential to learn temporal relations that may exist within a patient’s historical data. Our model outperformed both the currently used psychometric tools and accepted machine learning methods on a dataset comprised of many clinical variables recorded daily from many patients.
Our study is not without limitations: the dataset contained only 83 patients from a single population with some patients having as little as 30 days of usable sequential data. Further works would need to be performed before statements can be made about the general efficacy of our methodology as a clinical risk assessment tool. Additionally, we focused our analysis of predictive performance on the PRG curve and we believe it to be the most appropriate metric. However, for different use cases, different metrics and analysis may be necessary.
With the popularity and performance of global time series models, forecasters are no longer limited to homogeneous series or single series modelling [6]. We hope this work serves to indicate that tasks which traditionally used classification techniques, and where frequent data collection occurs, may benefit from this global time series approach which can explicitly learn temporal relationships.
References
- [1] J. Ogloff and M. Daffern, “The dynamic appraisal of situational aggression: An instrument to assess risk for imminent aggression in psychiatric inpatients,” Behavioral sciences & the law, vol. 24, pp. 799–813, 2006.
- [2] C. D. Webster, T. L. Nicholls, M.-L. Martin, S. L. Desmarais, and J. Brink, “Short-term assessment of risk and treatability (start): the case for a new structured professional judgment scheme,” Behavioral sciences & the law, vol. 24, no. 6, pp. 747–766, 2006.
- [3] I. R. Galatzer-Levy, K.-I. Karstoft, A. Statnikov, and A. Y. Shalev, “Quantitative forecasting of ptsd from early trauma responses: A machine learning application,” Journal of Psychiatric Research, vol. 59, pp. 68–76, 2014.
- [4] N. Parghi, L. Chennapragada, S. Barzilay, S. Newkirk, B. Ahmedani, B. Lok, and I. Galynker, “Assessing the predictive ability of the suicide crisis inventory for near-term suicidal behavior using machine learning approaches,” International Journal of Methods in Psychiatric Research, vol. 30, p. e1863, 2021.
- [5] G. Orrù, M. Monaro, C. Conversano, A. Gemignani, and G. Sartori, “Machine learning in psychometrics and psychological research,” Frontiers in Psychology, vol. 10, p. 2970, 2020.
- [6] P. Montero-Manso and R. J. Hyndman, “Principles and algorithms for forecasting groups of time series: Locality and globality,” International Journal of Forecasting, vol. 37, no. 4, pp. 1632–1653, 2021.
- [7] R. Sen, H.-F. Yu, and I. Dhillon, “Think globally, act locally: A deep neural network approach to high-dimensional time series forecasting,” in Advances in Neural Information Processing Systems, H. Wallach, H. Larochelle, A. Beygelzimer, F. d’Alché Buc, E. Fox, and R. Garnett, Eds., vol. 32. Curran Associates, Inc., 2019.
- [8] Z. Li, J. He, H. Liu, and X. Du, “Combining global and sequential patterns for multivariate time series forecasting,” in 2020 IEEE International Conference on Big Data, 2020, pp. 180–187.
- [9] K. Bandara, P. Shi, C. Bergmeir, H. Hewamalage, Q. Tran, and B. Seaman, “Sales demand forecast in e-commerce using a long short-term memory neural network methodology,” in Neural Information Processing, T. Gedeon, K. W. Wong, and M. Lee, Eds., vol. 1142. Cham: Springer, 2019, pp. 462–474.
- [10] P. Flach and M. Kull, “Precision-recall-gain curves: Pr analysis done right,” in Advances in Neural Information Processing Systems, C. Cortes, N. Lawrence, D. Lee, M. Sugiyama, and R. Garnett, Eds., vol. 28. Curran Associates, Inc., 2015.
- [11] S. Makridakis, E. Spiliotis, and V. Assimakopoulos, “The m4 competition: Results, findings, conclusion and way forward,” International Journal of Forecasting, vol. 34, no. 4, pp. 802–808, 2018.
- [12] S. Smyl, “A hybrid method of exponential smoothing and recurrent neural networks for time series forecasting,” International Journal of Forecasting, vol. 36, no. 1, pp. 75–85, 2020.
- [13] H. Hewamalage, C. Bergmeir, and K. Bandara, “Global Models for Time Series Forecasting: A Simulation Study,” arXiv, 2020.
- [14] H. Hewamalage, C. Bergmeir, and K. Bandara, “Recurrent neural networks for time series forecasting: Current status and future directions,” International Journal of Forecasting, vol. 37, pp. 388–427, 2021.
- [15] E. Bautu, S. Kim, A. Bautu, H. Luchian, and B.-T. Zhang, “Evolving hypernetwork models of binary time series for forecasting price movements on stock markets,” IEEE Congress on Evolutionary Computation, vol. 11, pp. 166–173, 2009.
- [16] C. Jentsch and L. Reichmann, “Generalized binary time series models,” Econometrics, vol. 7, p. 47, 2019.
- [17] W. T. M. Dunsmuir and D. J. Scott, “The glarma package for observation-driven time series regression of counts,” Journal of Statistical Software, vol. 67, no. 7, p. 1–36, 2015.
- [18] A. M. Aguilera, M. Escabias, and M. J. Valderrama, “Forecasting binary longitudinal data by a functional pc-arima model,” Computational Statistics & Data Analysis, vol. 52, no. 6, pp. 3187–3197, 2008.
- [19] R. Likert, “A technique for the measurement of attitudes,” Archives of psychology, vol. 22, no. 140, pp. 5–55, 1933.
- [20] K. Bandara, C. Bergmeir, and S. Smyl, “Forecasting across time series databases using recurrent neural networks on groups of similar series: A clustering approach,” Expert Systems with Applications, vol. 140, p. 112896, 2020.
- [21] N. Srivastava, G. Hinton, A. Krizhevsky, I. Sutskever, and R. Salakhutdinov, “Dropout: A simple way to prevent neural networks from overfitting,” J. Mach. Learn. Res., vol. 15, no. 1, p. 1929–1958, 2014.
- [22] B. Zoph and Q. V. Le, “Neural architecture search with reinforcement learning,” ArXiv: Machine Learning, vol. 1611.01578, 2017.
- [23] L. A. Jeni, J. F. Cohn, and F. De La Torre, “Facing imbalanced data–recommendations for the use of performance metrics,” in 2013 Humaine Association Conference on Affective Computing and Intelligent Interaction, 2013, pp. 245–251.
- [24] R. Hyndman and G. Athanasopoulos, Forecasting: Principles and Practice, 2nd ed. Melbourne, Australia: OTexts, 2018.
- [25] H. R. Kunsch, “The Jackknife and the Bootstrap for General Stationary Observations,” The Annals of Statistics, vol. 17, no. 3, pp. 1217–1241, 1989.
- [26] O. Pons, “Bootstrap of means under stratified sampling,” Electronic Journal of Statistics, vol. 1, pp. 381–391, 2007.