This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Stochastic loss reserving with mixture density neural networks

Muhammed Taher Al-Mudafer [email protected] Benjamin Avanzi [email protected] Greg Taylor [email protected] Bernard Wong [email protected] School of Risk and Actuarial Studies, UNSW Australia Business School, UNSW Sydney NSW 2052, Australia Centre for Actuarial Studies, Department of Economics, University of Melbourne VIC 3010, Australia
Abstract

In recent years, new techniques based on artificial intelligence and machine learning in particular have been making a revolution in the work of actuaries, including in loss reserving. A particularly promising technique is that of neural networks, which have been shown to offer a versatile, flexible and accurate approach to loss reserving. However, applications of neural networks in loss reserving to date have been primarily focused on the (important) problem of fitting accurate central estimates of the outstanding claims. In practice, properties regarding the variability of outstanding claims are equally important (e.g., quantiles for regulatory purposes).

In this paper we fill this gap by applying a Mixture Density Network (“MDN”) to loss reserving. The approach combines a neural network architecture with a mixture Gaussian distribution to achieve simultaneously an accurate central estimate along with flexible distributional choice. Model fitting is done using a rolling-origin approach. Our approach consistently outperforms the classical over-dispersed model both for central estimates and quantiles of interest, when applied to a wide range of simulated environments of various complexity and specifications.

We further extend the MDN approach by proposing two extensions. Firstly, we present a hybrid GLM-MDN approach called “ResMDN“. This hybrid approach balances the tractability and ease of understanding of a traditional GLM model on one hand, with the additional accuracy and distributional flexibility provided by the MDN on the other. We show that it can successfully improve the errors of the baseline ccODP, although there is generally a loss of performance when compared to the MDN in the examples we considered. Secondly, we allow for explicit projection constraints, so that actuarial judgement can be directly incorporated in the modelling process.

Throughout, we focus on aggregate loss triangles, and show that our methodologies are tractable, and that they out-perform traditional approaches even with relatively limited amounts of data. We use both simulated data—to validate properties, and real data—to illustrate and ascertain practicality of the approaches.

keywords:
Loss reserving, neural network, Mixture Density Network, Distributional Forecasting, machine learning JEL codes: C45 , C53 , G22 , MSC classes: 91G70 , 91G60 , 62P05

1 Introduction

1.1 Background

Forecasting outstanding claims (“OSC”) is essential in insurance for effective loss reserving, liability reporting and pricing. In addition to producing accurate central estimates, forecasting the distribution of future claims accurately is important in allocating capital which satisfies regulatory requirements. In Australia, the calculation of regulated risk margins involves determining quantiles such as the 75% percentile. In Europe, Solvency II regulations require calculations as far as a 99.5% percentile. Having an enhanced understanding of the distribution of OSC is beneficial beyond regulatory obligations, of course. With point forecasts of OSC being the predominant focus on the literature, this paper focuses on distributional forecasting of outstanding claims using neural networks.

Machine learning models, especially neural networks (“NN”), have gained momentum in the actuarial field in the past few years. Neural networks have stood out for providing ‘state of the art results’ (Rossouw and Richman, 2019). Their earliest use in reserving, to our knowledge, goes to Mulquiney (2006). Since 2018, this field has accelerated, with recent NN applications to reserving showcasing its accuracy and versatility. Kuo (2019); Gabrielli, Richman and Wüthrich (2020) use NNs to learn claim development trends from multiple lines of business simultaneously. Wüthrich and Merz (2019) develop a GLM-NN hybrid model, which is used successfully by Gabrielli, Richman and Wüthrich (2020); Poon (2019); Gabrielli (2021). The NNs ability to handle large granular datasets has allowed individual claims modelling to flourish, with recent success demonstrated by Kuo (2020); Delong, Lindholm and Wüthrich (2021); Gabrielli (2021).

1.2 Motivation and contributions

Despite the potential shown by NNs, several gaps in their current implementation can be observed, in particular with aggregate data. Firstly, most of the neural network loss reserving literature focuses on obtaining accurate central estimates. However, providing accurate distributional forecasts for outstanding claims is essential for optimising capital allocation, reporting liabilities with more accurate risk margins, and allowing profit margins that suit the company’s risk appetite to be set more accurately. Secondly, there is a lack of an explicit model selection framework. The performance of neural networks is heavily dependent on its design, therefore having no explicit selection procedure risks choosing a model with significantly reduced accuracy. Furthermore, model building becomes more reliant on expert input, hindering their applicability in practice. This paper aims to address those issues, as developed below.

1.2.1 Distributional forecasting with Neural Networks using the MDN

Probabilistic forecasts of outstanding claims are essential for capital allocation and optimising the risk margins set when reporting liabilities and pricing products. In this paper, we explain how such forecasts can be obtained with the Mixture Density Network (“MDN”) in a flexible way, working with aggregate loss triangles. The MDN is a design developed by Bishop (1994) (see also Bishop, 2006), with applications found in many fields, such as financial modelling (Ormoneit and Neuneier, 1996), acoustics (Zen and Senior, 2014) and electrical engineering (Vossen, Feron and Monti, 2018). The essential idea is that the parameters of the distribution will be estimated by the NN architecture, allowing for heterogeneity. Related works in the (non mixture) Gaussian setting can be found in Nix and Weigend (1994) and Lakshminarayanan et al. . The MDN assumes that the target output follows a mixture (typically Gaussian) distribution, allowing effectively a flexible distribution to be fit to the data. In this paper, we will assume that incremental claims follow a mixture Gaussian or mixture Log-Gaussian distribution.

A special case of an MDN was used by Kuo (2020) who mixed a shifted lognormal with a degenerate distribution to represent (positive) and zero cashflows for individual claims, respectively. In contrast, in our application we utilise a mixture of Gaussian distributions, whereby the number of components as well as all parameters are kept flexible.

In loss reserving, previous literature has fitted members of the exponential family using NNs. For instance, Gabrielli, Richman and Wüthrich (2020) considered aggregate claims, and Delong et al. (2021); Gabrielli (2021) considered individual claims. However, they typically focused on the central estimate rather than distributional properties.

A typical approach is to utilise the NN to replace the linear component in a GLM setup (see also Denuit and Trufin, 2020, for additional discussions), including the GLM-based assumption of a homogeneous mean-variance relationship. Such an assumption is not as flexible for the modelling of volatility, as compared to the mixture Gaussian fit by the MDN, which allows for heterogeneity for each random variable. In addition, the mixture Gaussian, given sufficient components, can approximate any distribution within a desired accuracy (see Nguyen and McLachlan, 2019, for details). As the MDN considers a wide range of distributions of varying volatility and shape, the training and fitting process of the network is akin to distribution selection. Another benefit of MDNs is that the central estimates can be derived directly from the fitted mixture Gaussian parameters, which means that location and shape are fitted simultaneously under the single model.

In addition to analysing central estimates, we use qualitative and quantitative measures to assess the distributional and quantile accuracy of the fitted MDN forecasts; see Section 4.4. Overall, we show that the MDN’s flexibility yields more accurate probabilistic forecasts when compared to the cross-classified over-dispersed Poisson (“ccODP”), in both simulated and real data; see also Section 1.2.4.

1.2.2 Model calibration and selection

Neural networks are highly flexible in their design. Since different designs produce varying results, it is vital to have a clear methodology for testing between these different model designs.

Commonly, the data is partitioned into training, validation and testing sets. The network is trained on the training set until the validation loss is minimised, then projected onto the testing set. The model producing the lowest test error is preferred.

To test between different models, a loss triangle must be split into training, validation and testing sets. Tashman (2000) and Bergmeir and Benítez (2012) explore two popular methods; fixed origin and rolling origin, used to partition sequential data. Balona and Richman (2020) recently apply the rolling origin method to loss triangles, in order to compare between different traditional reserving models.

This paper contributes by performing the rolling origin data partition exclusively within the aggregate loss triangle, using it to select neural network designs. Taking the latest calendar periods for testing has been done by Kuo (2019); Ramos-Pérez, Alonso-González and Núñez-Velázquez (2019), which was facilitated by simultaneously training the NN on multiple triangles. This paper performs model testing, selection and fitting using only one triangle at a time. Furthermore, the neural network model testing framework implemented in this paper outlines a validation set, which is essential in training NNs.

Furthermore, this paper contributes by implementing a model searching and selection algorithm to loss triangle reserving, which methodically searches and selects different network features such as the number of layers, nodes, and components. This algorithm extends the methodology implemented by Gabrielli, Richman and Wüthrich (2020) by also searching for the best-performing regularisation coefficients. This algorithm also makes NN design selection more methodical, requiring less expert input and assisting the popularisation of NNs in practice.

1.2.3 Acceptability of results in practice: ResMDN, and projection constraints

While the current neural network applications to loss reserving have shown their accuracy and flexibility in modelling, there are practical and technical obstacles which have hindered the acceptance and use of neural networks by the actuarial community (Rossouw and Richman, 2019; Wüthrich and Merz, 2019).

Producing interpretable forecasts is important for justifying business decisions to stakeholders. The neural network’s lack of interpretability has contributed to its lack of acceptance by actuaries. Wüthrich and Merz (2019) tackle this by adopting the resnet design from He et al. (2016) to form the GLM-NN hybrid CANN (Combined Actuarial Neural Network) architecture.

In this paper, we adapt the MDN to the above approach, to create the ResMDN. While Gabrielli, Richman and Wüthrich (2020), Poon (2019) and Richman and Wuthrich (2021) have adopted a similar methodology in loss reserving, the ResMDN provides the additional distributional flexibility of a (heterogeneous) mixture Gaussian distribution assumption. It boosts all parameters of the mixture Gaussian simultaneously and directly within one network, while maintaining a fixed GLM initialisation for interpretability.

We show that the ResMDN can successfully boost the embedded GLM, the ccODP, improving structural deficiencies in both mean and volatility estimates where visible, while maintaining the interpretable GLM backbone in its forecasts.

As a further consideration, the NN’s black box modelling can cause long range forecasts to become unstable (Rossouw and Richman, 2019). In other words, it may not be adequate to project the behaviour of the highly flexible function fit to the upper triangle to the lower triangle. In this paper, central estimate projections are able to be explicitly constrained by the actuary, forcing the neural network to only fit functions which produce reasonable mean forecasts. This framework thus also allows actuarial judgement to be explicitly incorporated into the modelling process.

1.2.4 Aggregate data in loss triangles

Recent applications of machine learning to loss reserving have been mainly focused on more granular claims datasets, such as individual claims data. This article, however, works exclusively with aggregate loss triangles. Developing a neural network model that is applied exclusively to loss triangles is highly relevant, as it enhances comparability and interpretability of results (as opposed to traditional reserving), and some insurers may still have limited data, preventing them from applying individual loss reserving methodologies. Furthermore, it is not obvious that machine learning methods can replicate their documented accuracy when presented with limited data, which is an interesting research question in itself.

In this paper, we demonstrate how neural networks can improve on traditional models, even with limited triangular data:

  • Data scarcity: The MDN was applied by Kuo (2020), but it was used in Individual Claims modelling. The current paper shows that the MDN also finds success when applied to loss triangles.

  • Model selection: Rossouw & Richman (2019) outline a fixed origin data partition (see Bergmeir & Benitez (2012) for more detail), but use more granular data, hence this methodology hasn’t been tested on a sparse loss triangle. While Gabrielli et al (2020) works with loss triangles, individual claims data are used to partition into training and validation sets. Balona & Richman (2020) recently apply the rolling origin model validation methodology (see Bergmeir & Benitez (2012) for more details) exclusively to loss triangles. However, as only non-machine learning models were tested, no validation set was constructed in the paper.

    With the rolling origin methodology and model searching algorithm, the neural network designs chosen in this paper produce robust, smooth, and accurate central and distributional forecasts, showcasing their combined effectiveness. Given that neural networks can perform unreliably with small datasets, these results show that the rolling origin partition is feasible for loss triangles of a certain size (40 ×\times 40 triangles were used in this paper), which should encourage further neural network modelling with loss triangles.

  • Data variety: Additionally, for NNs to prove their practicality and reliability, they must provide accurate results in a variety of different triangles (of instance, in a range of claim development situations which would invalidate the chain ladder assumptions). The current paper contributes to the literature by testing the MDN on a variety of environments of varying complexity and specifications, using both simulated and real data. Some of the environments tested are derived in concept from triangles used by Harej et al. (2017); Gabrielli et al. (2020). The MDN achieved superior probabilistic forecasts relative to the ccODP in all environments tested (see Section 3.3 for details).

1.3 Structure of the paper

Section 2 provides a detailed overview of the models used in this paper; the MDN, ResMDN and the benchmark ccODP models. Section 3 outlines the model development methodologies, including rolling origin validation and the hyper-parameter selection algorithm. The section concludes with an overview of the environments used. Section 4 provides details into the network training procedure and evaluation metrics. The standard MDN’s results are analysed in Section 5, with practical considerations, including the ResMDN, explored in Section 6. Section 7 concludes.

2 Description of the ccODP, MDN and ResMDN models

2.1 Notation

Let us first introduce some basic notation. Let

  • 1.

    Φ(x|μ,σ)=P(Zx)\Phi(x|\mu,\sigma)=P(Z\leq x) be the distribution function of a normal distribution with mean μ\mu and standard deviation σ\sigma, that is, ZN(μ,σ)Z\sim N(\mu,\sigma);

  • 2.

    dΦ(x|μ,σ)=ϕ(x|μ,σ)dxd\Phi(x|\mu,\sigma)=\phi(x|\mu,\sigma)dx, that is, ϕ(x|μ,σ)\phi(x|\mu,\sigma) is the probability density function of ZZ;

  • 3.

    Xi,jX_{i,j} be the incremental claims paid in accident period ii and Development period jj;

  • 4.

    X^i,j\hat{X}_{i,j} be a random variable which a model predicts to match the distribution of Xi,jX_{i,j};

  • 5.

    AQ and DQ be abbreviations for accident quarter and development quarter, respectively.

2.2 GLM: Cross-Classified Over-Dispersed Poisson (ccODP)

The benchmark model used in this paper is the Cross-Classified Over-Dispersed Poisson model (“ccODP”), in line with the existing NN loss reserving literature (Kuo, 2019; Gabrielli, Richman and Wüthrich, 2020; Gabrielli, 2020; Delong, Lindholm and Wüthrich, 2021; Kuo, 2020; Wüthrich, 2018). The ccODP model assumes that incremental claims, Xi,jX_{i,j}, follow a Cross-Classified Over-Dispersed Poisson distribution

X^i,jDPoi(AiBjD), where E[X^i,j]=AiBj and Var(X^i,j)=DAiBj.\frac{\hat{X}_{i,j}}{D}\sim\text{Poi}\bigg{(}\frac{A_{i}B_{j}}{D}\bigg{)},\text{ where }E[\hat{X}_{i,j}]=A_{i}B_{j}\text{ and }Var(\hat{X}_{i,j})=DA_{i}B_{j}. (2.1)

The Cross-Classified structure of the model, as well as the assumption of a constant DD, leads it to produce mean estimates that are identical to the Chain Ladder. Hence, the ccODP’s strengths and weaknesses in central estimate accuracy follow from the Chain Ladder’s characteristics. In practice, assuming the ccODP model is applied, some of these weaknesses can be easily mitigated. For example, a low value for X40,1X_{40,1} will lead the ccODP to estimate low losses for all of AQ40. Hence, in this paper, if the ccODP coefficient for AQ40, ln(A40)ln(A_{40}), is below the average of all AQ coefficients, it is taken as an average of the previous 3 quarters, as such:

If ln(A40)<i=140ln(Ai)40, then ln(A40)ln(A37)+ln(A38)+ln(A39)3.\text{If }ln(A_{40})<\frac{\sum_{i=1}^{40}ln(A_{i})}{40}\text{, then }ln(A_{40})\leftarrow\frac{ln(A_{37})+ln(A_{38})+ln(A_{39})}{3}. (2.2)

2.3 Mixture Density Network (MDN)

In this paper, we use Mixture Density Networks (MDNs) to perform probabilistic forecasting of outstanding claims.

2.3.1 Distribution of the incremental claims

Incremental claims Xi,jX_{i,j} are assumed to follow a mixture Gaussian distribution

fX^i,j(x)=k=1Kαi,j,kϕ(x|μi,j,k,σi,j,k)f_{\hat{X}_{i,j}}(x)=\sum_{k=1}^{K}\alpha_{i,j,k}\phi({x|\mu_{i,j,k},{\sigma_{i,j,k}}}) (2.3)

With that distributional assumption, the output layer of the MDN estimates the parameters of the mixture distribution, (𝜶,𝝁,𝝈(\boldsymbol{\alpha},\boldsymbol{\mu},\boldsymbol{\sigma}), which are used to form a mixture Gaussian density. A Negative Log Likelihood (NLL) loss function

NLLLoss(X,X^|w)=1|X|i,j:Xi,jTrainln(fX^i,j(Xi,j|w))NLLLoss(\textbf{X},\hat{\textbf{X}}|\textbf{w})=-\frac{1}{|\textbf{X}|}\sum_{i,j:X_{i,j}\in Train}ln(f_{\hat{X}_{i,j}}(X_{i,j}|\textbf{w})) (2.4)

is used to train the MDN, where X is the set of cells Xi,jX_{i,j} in the training set, |X||\textbf{X}| is the cardinal of X, X^\hat{\textbf{X}} is the set of predicted distributions of Xi,jX_{i,j} and w is the set of weights in the MDN.

The MDN isn’t structurally restricted to fitting mixture Gaussians to the response; it can estimate the parameters of any desired distribution so long as the loss function is specified accordingly. In this paper, only the mixture Gaussian framework was considered, with the output layer estimating the (𝜶,𝝁,𝝈)(\boldsymbol{\alpha},\boldsymbol{\mu},\boldsymbol{\sigma}) parameters.

As Bishop (1994) notes, the mixture Gaussian distribution, given a sufficient number of components and hidden layers, is capable of approximating any desired distribution within a desired accuracy. Mixture densities with more components will certainly be more flexible, however, practical obstacles such as over-parametrisation and data insufficiency will limit the range of distributions fit by the MDN.

Alternatively, while maintaining a mixture Gaussian output layer, a mixture Log-Gaussian can also be fit to Xi,jX_{i,j} by fitting a mixture Gaussian to ln(Xi,j)ln(X_{i,j})—This holds by the definition of a mixture random variable; see Appendix A for details). This distribution helps to address the practical limitations to the flexibility of the mixture Gaussian by providing a positive, heavier-tailed option. Furthermore, taking the log of incremental claims linearises the data, which can make training simpler and more efficient. Both the mixture Gaussian and mixture Log-Gaussian distributions achieved impressive results, analysed in Section 5.

Remark 2.1.

We modelled the log of the data where justified (as is often the case in actuarial applications), leading to a mixture Log-Gaussian, but alternative transforms could be readily used by the modeller if needed.

2.3.2 Structure of the MDN

Figure 1 provides a basic visualisation of the MDN’s design. Mixture Density Networks differ from other neural networks due to their output layer, which estimates a mixture distribution, commonly mixture Gaussian, to the response variable. For a detailed overview of neural networks, their design, mechanism and terminology, see, for instance, Richman (2020). In the case of aggregate loss triangles that we consider, the input variables of the MDN are ii and jj, the accident and development periods, respectively. These variables are passed through a fully connected hidden layer, which consists of hidden layers, each layer consisting of neurons. Each neuron in a hidden layer takes a weighted sum of the previous layer’s output, before passing it through an activation function. The final hidden layer’s output is then passed to the output layer, which produces the desired distribution parameters.

Refer to caption

Figure 1: The basic design of the Mixture Density Network (MDN). The inputs (i,j)(i,j) are the accident and development quarters respectively. The outputs are the parameters of the mixture Gaussian distribution, α,μ,σ\alpha,\mu,\sigma

Let wa,clw_{a,c}^{l} be the weight parameter connecting the atha^{th} neuron in the lthl^{th} layer to the cthc^{th} neuron in the (l+1)th(l+1)^{th} layer. Define bal1b_{a}^{l-1} as the bias term added to the atha^{th} neuron in the lthl^{th} layer. Weighted sums of the inputs, (i,j)(i,j), are passed into the first hidden layer, before an activation function gg yields such output for the pthp^{th} node in that layer:

zi,j,p1=g(i×w1,p0+j×w2,p0+bp0).\textbf{z}_{i,j,p}^{1}=g(i\times w_{1,p}^{0}+j\times w_{2,p}^{0}+b_{p}^{0}). (2.5)

We assume LL hidden layers in the MDN, with DD nodes in each layer. Each node in successive hidden layers take a weighted sum of the output from nodes in the previous layer, such that zi,j,pL\textbf{z}_{i,j,p}^{L}, the pthp^{th} node in the final hidden layer, is calculated as

zi,j,pL=g(d=1Dwd,pL1zi,j,dL1+bpL1).\textbf{z}_{i,j,p}^{L}=g(\sum_{d=1}^{D}w_{d,p}^{L-1}\textbf{z}_{i,j,d}^{L-1}+b_{p}^{L-1}). (2.6)

The output layer is split into three sections, each with KK nodes, KK being the number of components in the Mixture Density. Let’s call these sections the alpha, mu and sigma sections, respectively. Similarly to the hidden layers, each node in the output layer takes a weighted sum of the output of all nodes in the last hidden layer, Layer LL. The weighted sums are then passed through different activation functions for each section to yield the final output of the MDN, (𝜶,𝝁,𝝈)(\boldsymbol{\alpha},\boldsymbol{\mu},\boldsymbol{\sigma}). Specifically:

Alpha:

The output of the kthk^{th} nodes of the alpha section

zi,j,kα=d=1Dwd,kL,αzi,j,kL+bkL,α, for k=1,2,..K\textbf{z}_{i,j,k}^{\alpha}=\sum_{d=1}^{D}w_{d,k}^{L,\alpha}\textbf{z}_{i,j,k}^{L}+b_{k}^{L,\alpha}\text{, for }k=1,2,..K (2.7)

leads to

αi,j,k=ezi,j,kαk=1Kezi,j,kα.\alpha_{i,j,k}=\frac{e^{\textbf{z}_{i,j,k}^{\alpha}}}{\sum_{k=1}^{K}e^{\textbf{z}_{i,j,k}^{\alpha}}}. (2.8)

Note that the output zi,j,kα\textbf{z}_{i,j,k}^{\alpha} was passed through a Softmax activation function, which ensures that k=1Kαi,j,k=1\sum_{k=1}^{K}\alpha_{i,j,k}=1.

Mu:

Similarly,

zi,j,kμ=d=1Dwd,kL,μzi,j,kL+bkL,μ, for k=1,2,..K\textbf{z}_{i,j,k}^{\mu}=\sum_{d=1}^{D}w_{d,k}^{L,\mu}\textbf{z}_{i,j,k}^{L}+b_{k}^{L,\mu}\text{, for }k=1,2,..K (2.9)

leads to

μi,j,k=zi,j,kμ\mu_{i,j,k}=\textbf{z}_{i,j,k}^{\mu} (2.10)

as there are no constraints on the mu layer which would require an activation function. Bishop (1994) notes that such a design represents an ‘un-informative prior’ on μ\mu, which befits the lack of constraints on the mean.

Sigma:

Finally, the sigma output

zi,j,kσ=d=1Dwd,kL,σzi,j,kL+bkL,σ, for k=1,2,..K\textbf{z}_{i,j,k}^{\sigma}=\sum_{d=1}^{D}w_{d,k}^{L,\sigma}\textbf{z}_{i,j,k}^{L}+b_{k}^{L,\sigma}\text{, for }k=1,2,..K (2.11)

is passed through an exponential function,

σi,j,k=ezi,j,kσ,{\sigma_{i,j,k}}=e^{\textbf{z}_{i,j,k}^{\sigma}}, (2.12)

which ensures the standard deviation is always positive (Hjorth and Nabney, 2000).

Here, wd,kL,α,wd,kL,μ,wd,kL,σw_{d,k}^{L,\alpha},w_{d,k}^{L,\mu},w_{d,k}^{L,\sigma} are the weights connecting the output of node dd in layer LL to node kk in the alpha, mu and sigma layers, respectively. Thus, for each input cell (i,j)(i,j), a unique combination of parameters,

(αi,j,1,αi,j,2..αi,j,K,μi,j,1,μi,j,2..μi,j,K,σi,j,1,σi,j,2..σi,j,K),(\alpha_{i,j,1},\alpha_{i,j,2}.....\alpha_{i,j,K},\mu_{i,j,1},\mu_{i,j,2}.....\mu_{i,j,K},\sigma_{i,j,1},\sigma_{i,j,2}.....\sigma_{i,j,K}),

is produced in the output layer of the MDN, which then generates the probability density for X^i,j\hat{X}_{i,j}

fX^i,j(x)=k=1Kαi,j,kϕ(x|μi,j,k,σi,j,k).f_{\hat{X}_{i,j}}(x)=\sum_{k=1}^{K}\alpha_{i,j,k}\phi(x|\mu_{i,j,k},\sigma_{i,j,k}). (2.13)

2.4 Boosting GLM based models with the MDN: ResMDN

The Mixture Density Network (“MDN”) has greater computational complexity than the standard feedforward neural network, hence its interpretability is even lower. To implement a more interpretable structure, this paper adapts the residual neural network (“ResNet”) design implemented successfully by Gabrielli, Richman and Wüthrich (2020), Gabrielli (2020) and Poon (2019), which boosted a GLM model with a neural network - resulting in a more interpretable and stable model. In this paper, this boosting design was adapted to the MDN to create the“ResMDN”. Note that in our ResMDN approach the mean of the resulting model can be interpreted as a boosted version of the GLM backbone, but the other probabilistic distributional properties for the resulting model are inherited from the MDN. This makes the ResMDN approach in principle quite different from the ResNet approach by Gabrielli, Richman and Wüthrich (2020), which focused on the mean.

The ResMDN uses a skip connection, applied in the form of an Embedding Layer, to connect the input layer directly to the output layer. This skip connection allows the MDN to initialise with an approximate GLM fit, subsequently enabling the feedforward module to boost the GLM during training.

In the following, we chose to illustrate the ResMDN by boosting the well known (and used) ccODP model. It is worthwhile to note that—quite naturally—some of the benefits and drawbacks of the ccODP will flow on to the outcomes of the resulting ResMDN model to a certain extent. Indeed, this is illustrated in Section 6.

2.4.1 Distribution of the incremental claims

The ResMDN embeds an approximation of the Cross-Classified Over-Dispersed Poisson (ccODP) model (see Section 2.2 for more detail), which follows the distribution outlined in (2.1). In neural network terminology, an embedding is the mapping of a discrete or categorical input variable into a numerical vector, which is then fed into the network (Richman, 2020). Since the output of the ResMDN takes the form of parameters for a mixture Gaussian distribution, the GLM initialisation’s density is approximated by a mixture Gaussian, spread evenly over KK components. The parameters of this approximation are fed into the ResMDN as embeddings of the GLM. The distribution fX^i,jccODP(x)f_{\hat{X}_{i,j}^{ccODP}}(x) of Xi,jX_{i,j} as estimated by the ccODP is approximated by

fX^i,jccODP(x)k=1Kαi,j,kGLMϕi,j,k(x|μi,j,kGLM,σi,j,kGLM),f_{\hat{X}_{i,j}^{ccODP}}(x)\approx\sum_{k=1}^{K}\alpha^{GLM}_{i,j,k}\phi_{i,j,k}(x|{\mu^{GLM}_{i,j,k},\sigma^{GLM}_{i,j,k}}), (2.14)

where

αi,j,kGLM=1K,μi,j,kGLM=E[X^i,jccODP]=AiBj,σi,j,kGLM=Var[X^i,jccODP]=DAiBj.\alpha_{i,j,k}^{GLM}=\frac{1}{K},\quad\mu_{i,j,k}^{GLM}=E[\hat{X}_{i,j}^{ccODP}]=A_{i}B_{j},\quad\sigma_{i,j,k}^{GLM}=\sqrt{Var[\hat{X}_{i,j}^{ccODP}]}=\sqrt{DA_{i}B_{j}}. (2.15)

2.4.2 ResMDN structure

The ResMDN’s structure resembles the MDN very closely. The Input Layer of the ResMDN consists of the accident and development periods, (i,j)(i,j), as well as a unique categorical integer, ci,j=40(i1)+jc_{i,j}=40*(i-1)+j, which allows the ResMDN’s embedding layer to identify the specific cell (i,j)(i,j) and produce the corresponding GLM loss estimate for that cell as output. Hence, each cell (i,j)(i,j) is assigned a number from (0,1599)(0,1599). The variables (i,j)(i,j) are passed through an MDN which excludes the activations of the final output layer. An embedding layer takes the categorical input ci,jc_{i,j} and produces as output:

(ln(𝜶𝒊,𝒋𝑮𝑳𝑴),𝝁𝒊,𝒋𝑮𝑳𝑴,ln(𝝈𝒊,𝒋𝑮𝑳𝑴))\bigg{(}ln(\boldsymbol{\alpha^{GLM}_{i,j}}),\boldsymbol{\mu^{GLM}_{i,j}},ln(\boldsymbol{\sigma^{GLM}_{i,j}})\bigg{)} (2.16)

The outputs from both the fully connected component and embedding layer are added together, before the Softmax and exponential activations are applied to the alpha and sigma additions, respectively. The final output of the ResMDN consists of the mixture Gaussian parameter estimates, (𝜶𝑹𝒆𝒔𝑴𝑫𝑵,𝝁𝑹𝒆𝒔𝑴𝑫𝑵,𝝈𝑹𝒆𝒔𝑴𝑫𝑵)(\boldsymbol{\alpha^{ResMDN}},\boldsymbol{\mu^{ResMDN}},\boldsymbol{\sigma^{ResMDN}}). Figure 2 provides a visualisation of the ResMDN model. The key design features of the ResMDN are outlined below:

Refer to caption

Figure 2: The ResMDN design with a mixture Gaussian output. The embedding layer converts the input to mixture Gaussian parameters approximating the GLM backbone. The feedforward module boosts the GLM initialisation during training.
  • 1.

    Fully Connected Module: Let zi,j,kα,zi,j,kμ,zi,j,kσ\textbf{z}_{i,j,k}^{\alpha},\textbf{z}_{i,j,k}^{\mu},\textbf{z}_{i,j,k}^{\sigma} be as described in (2.7), (2.9) and (2.11), that is, the MDN’s output before the output layer activations are applied. The fully connected module of the ResMDN performs the function

    (i,j){(zi,j,kα,zi,j,kμ,zi,j,kσ),k=1,2,..K},(i,j)\mapsto\{(\textbf{z}_{i,j,k}^{\alpha},\textbf{z}_{i,j,k}^{\mu},\textbf{z}_{i,j,k}^{\sigma}),k=1,2,..K\}, (2.17)
  • 2.

    Embedding Layer: The embedding layer weights are pre-set to provide the mapping

    ci,j(ln(𝜶𝒊,𝒋𝑮𝑳𝑴),𝝁𝒊,𝒋𝑮𝑳𝑴,ln(𝝈𝒊,𝒋𝑮𝑳𝑴).c_{i,j}\mapsto\bigg{(}ln(\boldsymbol{\alpha^{GLM}_{i,j}}),\boldsymbol{\mu^{GLM}_{i,j}},ln(\boldsymbol{\sigma^{GLM}_{i,j}}\bigg{)}. (2.18)

    The log of the 𝜶\boldsymbol{\alpha} and 𝝈\boldsymbol{\sigma} parameters are produced in the embedding layer, since the Softmax and exponential Activation functions will take the exponent in the output layer nodes.

  • 3.

    Addition and Final Activation: The output from the embedding and fully connected modules are added together element-wise:

    Addition: (ln(αi,j,kGLM),μi,j,kGLM,ln(σi,j,kGLM),zi,j,kα,zi,j,kμ,zi,j,kσ)\displaystyle\bigg{(}ln(\alpha^{GLM}_{i,j,k}),\mu^{GLM}_{i,j,k},ln(\sigma^{GLM}_{i,j,k}),\textbf{z}_{i,j,k}^{\alpha},\textbf{z}_{i,j,k}^{\mu},\textbf{z}_{i,j,k}^{\sigma}\bigg{)}
    (ln(αi,j,kGLM)+zi,j,kα,μi,j,kGLM+zi,j,kμ,ln(σi,j,kGLM)+zi,j,kσ)\displaystyle\mapsto\bigg{(}ln(\alpha^{GLM}_{i,j,k})+\textbf{z}_{i,j,k}^{\alpha},\mu^{GLM}_{i,j,k}+\textbf{z}_{i,j,k}^{\mu},ln(\sigma^{GLM}_{i,j,k})+\textbf{z}_{i,j,k}^{\sigma}\bigg{)}
    =(ln(αi,j,kGLM)+zi,j,kα,μi,j,kResMDN,ln(σi,j,kResMDN)),\displaystyle\quad=\bigg{(}ln(\alpha^{GLM}_{i,j,k})+\textbf{z}_{i,j,k}^{\alpha},\mu^{ResMDN}_{i,j,k},ln(\sigma^{ResMDN}_{i,j,k})\bigg{)}, (2.19)

    before the Softmax and exponential activations are applied to the alpha and sigma layers, respectively:

    Final Activations: (ln(αi,j,kGLM)+zi,j,kα,μi,j,kResMDN,ln(σi,j,kResMDN))\displaystyle\bigg{(}ln(\alpha^{GLM}_{i,j,k})+\textbf{z}_{i,j,k}^{\alpha},\mu^{ResMDN}_{i,j,k},ln(\sigma^{ResMDN}_{i,j,k})\bigg{)}
    (αi,j,kGLMezi,j,kαk=1Kαi,j,kGLMezi,j,kα,μi,j,kResMDN,eln(σi,j,kResMDN))\displaystyle\mapsto\left(\frac{{\alpha^{GLM}_{i,j,k}e^{\textbf{z}_{i,j,k}^{\alpha}}}}{\sum_{k=1}^{K}\alpha^{GLM}_{i,j,k}e^{\textbf{z}_{i,j,k}^{\alpha}}},\mu^{ResMDN}_{i,j,k},e^{ln(\sigma^{ResMDN}_{i,j,k})}\right)
    =(αi,j,kResMDN,μi,j,kResMDN,σi,j,kResMDN).\displaystyle\quad=\left(\alpha^{ResMDN}_{i,j,k},\mu^{ResMDN}_{i,j,k},\sigma^{ResMDN}_{i,j,k}\right). (2.20)

    Hence, the boosted mixture Gaussian parameters are produced in the Output Layer.

  • 4.

    Initialisation: The activations, zi,j,kα,zi,j,kμ,zi,j,kσ\textbf{z}_{i,j,k}^{\alpha},\textbf{z}_{i,j,k}^{\mu},\textbf{z}_{i,j,k}^{\sigma}, are generated using the parameters, (wL,bL)(\textbf{w}_{L},\textbf{b}_{L}), defined in (2.7)–(2.12). These parameters, representing the weights in the final hidden layer, are initialised at 0, such that

    zi,j,kα,zi,j,kμ,zi,j,kσ=0,αi,j,kResMDN=αi,j,kGLM,μi,j,kResMDN=μi,j,kGLM,σi,j,kResMDN=σi,j,kGLM,\textbf{z}_{i,j,k}^{\alpha},\textbf{z}_{i,j,k}^{\mu},\textbf{z}_{i,j,k}^{\sigma}=0,\alpha_{i,j,k}^{ResMDN}=\alpha_{i,j,k}^{GLM},\mu_{i,j,k}^{ResMDN}=\mu_{i,j,k}^{GLM},\sigma_{i,j,k}^{ResMDN}=\sigma_{i,j,k}^{GLM},

    hence producing the GLM approximation in the Output Layer at the initialisation of the ResMDN. This initialisation follows the methodology of Gabrielli, Richman and Wüthrich (2020) closely.

During training, the embedding layer maintains constant output, while the fully connected module adjusts its weights to capture non-linearities which the GLM has missed. The ResMDN’s overall function at the termination of training is

(i,j,ci,j)\displaystyle(i,j,c_{i,j}) (αi,j,kGLMezi,j,kαk=1Kαi,j,kGLMezi,j,kα,μi,j,kGLM+zi,j,kμ,σi,j,kGLMezi,j,kσ)\displaystyle\mapsto\left(\frac{{\alpha^{GLM}_{i,j,k}e^{\textbf{z}_{i,j,k}^{\alpha}}}}{\sum_{k=1}^{K}\alpha^{GLM}_{i,j,k}e^{\textbf{z}_{i,j,k}^{\alpha}}},\mu^{GLM}_{i,j,k}+\textbf{z}_{i,j,k}^{\mu},{\sigma^{GLM}_{i,j,k}e^{\textbf{z}_{i,j,k}^{\sigma}}}\right)
=(αi,j,kResMDN,μi,j,kResMDN,σi,j,kResMDN), for k=1,2,,K.\displaystyle\quad=\left(\alpha^{ResMDN}_{i,j,k},\mu^{ResMDN}_{i,j,k},\sigma^{ResMDN}_{i,j,k}\right)\text{, for }k=1,2,...,K. (2.21)

The NN boosting terms are relatively easy to analyse in relation to the GLM fit, especially the mean and volatility terms. Furthermore, the black box neural network modelling is only applied to the residuals, meaning the lack of interpretability is restricted to that domain only. Hence the ResMDN improves the interpretability of the model compared to the MDN.

3 Model Development

3.1 Model selection using the rolling origin method for training, validating, and testing

The accuracy of the neural network depends heavily on its hyper-parameters, such as the number of hidden layers, number of neurons, weight regularisation penalty, etc. Therefore, to assume that one model design will work well in all environments will lead to sub-optimal performance. Hence, a training/testing split is required to assess different model designs and choose the best one found. This paper partitions the loss triangle using the rolling origin method, which performs the training and testing split in multiple stages, each one progressively shifting the testing set forward in time (see Tashman (2000); Bergmeir and Benítez (2012); Balona and Richman (2020) for details). The total test error of the model is a weighted average of the test error in each stage. This methodology allows for a systematic hyper-parameter fine-tuning algorithm to be implemented; see Section 3.2.

The loss triangle has the characteristics of a time series, with the incremental claims generally decaying over successive development periods. Where the objective of modelling is to improve interpolation accuracy, randomly splitting the data into training, validation and testing sets is common and sufficient. With loss triangles, however, the objective is extrapolation, hence the testing set needs to focus on assessing the model’s projection accuracy. This is done by assigning (a chosen number of) the latest calendar periods of the triangle to the testing set and the earliest to training. Similarly, the Validation set is chosen to be the latest calendar periods which aren’t assigned for testing. That way, when combined with Early Stopping (see Section 4), the MDN stops training when short term projection accuracy is maximised.

Hence, it is important to sequentially split the data into training, validation and testing sets to more effectively assess the model’s accuracy when extrapolating. In our illustrative example of a 40×\times40 triangle, the rolling origin validation method was used in two partitions:

  • 1.

    In the first partition, the data is assumed to comprise a 30×\times30 triangle, which leaves the latest 10 calendar periods for the testing set. This partition focuses on assessing the model’s long term forecasting accuracy.

  • 2.

    The second partition works with a 36×\times36 triangle, leaving 4 calendar periods for testing. Building on the first partition, later calendar periods are included in training, which helps to assess the model’s ability to capture the more recent and more holistic trends present in the triangle.

For all partitions, the validation set included the 4 latest non-testing calendar periods, excluding the first 3 accident and development periods. This exclusion was done to provide the MDN more training data for the latest accident and development periods. Instead, the DQ2 and DQ3 validation points are taken evenly from earlier AQs, an arbitrary but simplistic approach. Figure 3 visualises the data partitions.

A potential downside of the rolling origin method is that the training data does not include points from the latest calendar quarters. Therefore, in situations where we expect losses to possess substantially different characteristics in the later periods, this method might not be able to effectively capture the change in trends. In such situations, we implement an adjusted data partitioning methodology to allow more training data in those periods (see Section B).

Refer to caption
(a) Partition 1: Assesses projection accuracy
Refer to caption
(b) Partition 2: Assesses trend fitting
Figure 3: The 2-stage partition of the triangle into training, validation and testing sets. The first partition focuses on assessing projection accuracy, while the second assesses the model’s ability to fit more recent trends in the data.

3.2 Network hyper-parameter selection algorithm

There is a vast array of literature surrounding neural network model search and selection, and that no set algorithm succeeds the rest. The MDN’s architecture was selected using an algorithm that successively optimised one hyper-parameter at a time. The number of components in the density is increased so long as the test error decreases, which allows the algorithm to consider fitting densities of infinite flexibility. The loss triangle size restricts the parametrisation of the mixture density, hence this algorithm aims to find the optimal distributional flexibility allowed by the given data. Before the algorithm is run, certain aspects of the MDN’s design are set constant:

  • 1.

    The sigmoid activation function is used for all hidden layers;

  • 2.

    The number of neurons is equal for all hidden layers.

With this modelling framework, two additional considerations improved the MDN’s performance:

  • 1.

    Mixture Gaussian or mixture Log-Gaussian: Fitting a mixture Log-Gaussian linearises the data (in our methodology), which provided smoother results in some instances. When used, the mixture Log-Gaussian gave noticeably improved results over the mixture Gaussian in the upper triangle.

  • 2.

    Mean Squared Error term to the loss function: A Negative Log-Likelihood loss function sometimes failed to capture sharp points in the data, preferring to allocate volatility to these points. This issue was solved by adding a Mean Squared Error (“MSE”) term to the loss, which encouraged the MDN to provide more accurate central estimates.

The hyper-parameters chosen to fine-tune are:

  1. 1.

    λw\lambda_{w}, the L2 weight penalty. The values [0,0.0001,0.001,0.01,0.1][0,0.0001,0.001,0.01,0.1] were tested.

  2. 2.

    λσ\lambda_{\sigma}, the L2 sigma activity penalty. When central estimates are inaccurate, the MDN may increase volatility estimates unreasonably to reduce the Negative Log Likelihood Loss (2.4) ( Goodfellow et al. (2016, Ch6.2)), hence a penalty was applied to the sigma output. The values [0,0.0001,0.001,0.01,0.1][0,0.0001,0.001,0.01,0.1] were tested.

  3. 3.

    pp, the dropout rate (See Srivastava, Hinton, Krizhevsky, Sutskever and Salakhutdinov (2014) for details). The values [0,0.1,0.2][0,0.1,0.2] were tested.

  4. 4.

    nn, the number of neurons in each hidden layer. The values [20,40,60,80,100][20,40,60,80,100] were tested.

  5. 5.

    hh, the number of hidden layers. The values [1,2,3,4][1,2,3,4] were tested.

  6. 6.

    KK, the number of components in the mixture density.

The training loss becomes

Loss(X,X^|w,λw,λσ)\displaystyle Loss(\textbf{X},\hat{\textbf{X}}|\textbf{w},\lambda_{w},\lambda_{\sigma}) =1|X|i,j:Xi,jXln(fX^i,j(Xi,j|w))\displaystyle=-\frac{1}{|\textbf{X}|}\sum_{i,j:X_{i,j}\in\textbf{X}}ln(f_{\hat{X}_{i,j}}(X_{i,j}|\textbf{w}))
+λwww+λσi,j:Xi,jXk=1Kσi,j,k2\displaystyle+\lambda_{w}\textbf{w}\cdot\textbf{w}+\lambda_{\sigma}\sum_{i,j:X_{i,j}\in\textbf{X}}\sum_{k=1}^{K}\sigma_{i,j,k}^{2} (3.1)

Denote 𝜽={λw,λσ,p,n,h,K}\boldsymbol{\theta}=\{\lambda_{w},\lambda_{\sigma},p,n,h,K\} as the set of hyper-parameters to fine-tune. The hyper-parameter selection algorithm was conducted as such:

  1. 1.

    Start with 𝜽initial={0,0,0,ninitial,hinitial,Kinitial}\boldsymbol{\theta}^{initial}=\{0,0,0,n^{initial},h^{initial},K^{initial}\}, a set of initial hyper-parameters deemed suitable through judgement. Setting 𝜽initial={0,0,0,60,2,2}\boldsymbol{\theta}^{initial}=\{0,0,0,60,2,2\} worked well in this paper, as allowing the algorithm to explore unregularised models vastly improved the fit in some instances.

  2. 2.

    Using 𝜽initial\boldsymbol{\theta}^{initial} and keeping all other hyper-parameters fixed, use Grid Search to test all desired values of λw\lambda_{w}, the weight penalty coefficient. Select the coefficient with the lowest test error, λ^w\hat{\lambda}_{w}, and update 𝜽1={λ^w,0,0,ninitial,hinitial,Kinitial}\boldsymbol{\theta}^{1}=\{\hat{\lambda}_{w},0,0,n^{initial},h^{initial},K^{initial}\}

  3. 3.

    Using 𝜽1\boldsymbol{\theta}^{1} and keeping all other hyper-parameters fixed, use Grid Search to test all desired values of λσ\lambda_{\sigma}, the sigma activity penalty coefficient. Select the coefficient with the lowest test error, λ^σ\hat{\lambda}_{\sigma}, and update 𝜽2={λ^w,λ^σ,0,ninitial,hinitial,Kinitial}\boldsymbol{\theta}^{2}=\{\hat{\lambda}_{w},\hat{\lambda}_{\sigma},0,n^{initial},h^{initial},K^{initial}\}

  4. 4.

    Using 𝜽2\boldsymbol{\theta}^{2} and keeping all other hyper-parameters fixed, use Grid Search to test all desired values of pp, the dropout rate. Select the rate with the lowest test error, p^\hat{p}, and update 𝜽3={λ^w,λ^σ,p^,ninitial,hinitial,Kinitial}\boldsymbol{\theta}^{3}=\{\hat{\lambda}_{w},\hat{\lambda}_{\sigma},\hat{p},n^{initial},\linebreak[0]h^{initial},K^{initial}\}

  5. 5.

    Using 𝜽3\boldsymbol{\theta}^{3} and keeping all other hyper-parameters fixed, use Grid Search to test all desired values of hh, the number of hidden layers. Select the number with the lowest test error, h^\hat{h}, and update 𝜽4={λ^w,λ^σ,p^,ninitial,h^,Kinitial}\boldsymbol{\theta}^{4}=\{\hat{\lambda}_{w},\hat{\lambda}_{\sigma},\hat{p},n^{initial},\hat{h},K^{initial}\}

  6. 6.

    Using 𝜽4\boldsymbol{\theta}^{4} and keeping all other hyper-parameters fixed, fine-tune the number of neurons and components. This process tests an increasing number of components until the test error ceases to improve. Let nKn_{K} be the number of neurons which minimises the test error for a KK-component model (among the values tested). Let EnK,KE_{n_{K},K} be the test error of a model with nKn_{K} neurons and KK components (with the other hyper-parameters as in 𝜽4\boldsymbol{\theta}^{4}). Starting at K=1K=1, increment KK until EnK,K<EnK+1,K+1E_{n_{K},K}<E_{n_{K+1},K+1}. At this final increment, set K^=K\hat{K}=K and n^=nK\hat{n}=n_{K}. Update the hyper-parameters, 𝜽5={λ^w,λ^σ,p^,n^,h^,K^}\boldsymbol{\theta}^{5}=\{\hat{\lambda}_{w},\hat{\lambda}_{\sigma},\hat{p},\hat{n},\hat{h},\hat{K}\}

Following the algorithm, select 𝜽5\boldsymbol{\theta}^{5} as the final set of hyper-parameters and run the final model (see Section 4.

3.3 Data

This paper tests the MDN’s performance on both simulated and real data. Using simulated data allows the practitioner to simulate any desired trend, providing a controlled environment where the MDN can be directly assessed on its ability to capture these trends embedded in the data (Avanzi, Taylor, Wang and Wong, 2021). A downside of simulating claims, as noted by Murray, Ryan and Reisinger (2011) is that complex interactions in real data may not be captured by the simulator, hindering model development. Fitting a model on unrealistic data will reduce its validity. However, we mitigate this by applying the MDN on real data (AUSI environment; see Section 3.3.2) as well as the default SynthETIC dataset from Avanzi, Taylor, Wang and Wong (2021), which can mimic data in a wide range of realistic situations.

Note that while the AUSI Dataset was partitioned randomly into 10 triangles (via subdivision), we simulated 200 triangles (from four separate realistic scenarios) with SynthETIC, meaning that the MDN was fit on 210 triangles in total.

3.3.1 Simulated environments from SynthETIC

Thanks to the flexibility offered by SynthETIC (see Avanzi, Taylor, Wang and Wong, 2021, for a description of the R simulation package), four different claim environments were simulated, of various features and complexities. Simulating different environments was done to test the MDN’s versatility and ability to capture complex trends and produce accurate forecasts in a variety of controlled, challenging environments. As a loss triangle is a collection of random variables, it is important to run the MDN on a large sample of triangles to gain a better understanding of its accuracy and also test its ability to provide consistent results. Hence, for each simulated environment, 50 independent triangles were simulated (of size 40×4040\times 40), leading to 200 triangles. Namely, the scenarios, or environments, were:

  1. 1.

    Environment 1 - Simple, short tail claims: This environment simulates short tail claims which are homogeneous in composition for all accident quarters. The reporting and settlement delays have been approximately calibrated to show similar characteristics to the simulator developed by Gabrielli and Wüthrich (2018). Figure 4 plots the incremental claims for Environment 1. A spike in claim payment in DQ2 complicates this dataset, but such a feature is not uncommon in practice. This environment was a preliminary test to the feasibility of MDNs in modelling 40x40 triangles and producing reasonable results.

    Refer to caption

    Figure 4: A plot of the incremental claims of environment 1, for selected accident quarters. Solid lines represents data in the upper triangle, while dashed lines represent data in the lower triangle
  2. 2.

    Environment 2 - Increase in claim processing speed: A gradual shift from long tail to short tail claims along accident quarters is simulated, that is, an increase in claims processing speed. Initially, there are more long tail claims, however, the proportion of these claims decreases, while the proportion of short claims increases. From the incremental claim plot shown below, later AQs see higher losses early on, due to the increasing proportion of short tail claims. Figure 5 plot the incremental claims of environment 2. The main question to be answered in testing this dataset is, given the systematic volatility in the claims data, can the MDN accurately distinguish between systematic and unsystematic volatility and capture the distribution of data points accurately? That is, will the MDN learn that claims are getting shorter, or will it attribute the trend to noise?

    Refer to caption

    Figure 5: A plot of the incremental claims of environment 2, for selected accident quarters. Solid lines represents data in the upper triangle, while dashed lines represent data in the lower triangle
  3. 3.

    Environment 3 - Inflation shock: Superimposed inflation is changed instantly from 0% to 8% per annum, starting at AQ30. The 8% inflation remains constant in the lower triangle. This environment tests the ability of the MDN to recognise changes in calendar effects and adapt projections accordingly.

    Only the last 10 calendar quarters in the upper triangle contain information regarding the inflation shock, which increases the difficulty for the MDN. A further complication is that the rolling origin partition contains little training points featuring the change in inflation, hence this environment assesses its ability to capture recent trends. Figure 6 plots the incremental claims.

    Refer to caption

    Figure 6: A plot of the incremental claims of environment 3, for selected accident quarters. Solid lines represents data in the upper triangle, while dashed lines represent data in the lower triangle
  4. 4.

    Environment 4 - High Systematic Complexity: This environment is the default triangle generated by the SynthETIC simulator, which was designed to mimic features seen in real data (see Avanzi, Taylor, Wang and Wong, 2021, for details). Complex dependencies exist between claim size, reporting and settlement delay, and superimposed inflation. Settlement delay, which depends on claim size, declines over the first 20 AQs. Superimposed inflation is up to 30%, but declines for larger claims. A legislative change at AQ20 causes small claims to face a reduction in size and settlement speed until AQ30. The general trend can be summarised by slow development, high volatility and high superimposed inflation. The volatility is primarily caused by the low claim frequency and highly volatile severity. Hence, it is normal for claims in one AQ to follow a completely different pattern (reporting, settlement, volume, development pattern) than claims in the adjacent AQ. Figure 7 plots the incremental claims. This environment assesses the MDN’s ability to produce accurate forecasts in a volatile environment.

    Refer to caption

    Figure 7: A plot of the incremental claims of environment 4, for selected accident quarters. Solid lines represents data in the upper triangle, while dashed lines represent data in the lower triangle

3.3.2 Real Dataset - AUSI Auto Bodily Injury

We apply the MDN to a dataset obtained through a collaborative project between Allianz, University of New South Wales (UNSW), Suncorp and Insurance Australia Group (IAG). This forms the AUSI acronym, which we will use to refer to this dataset. The Auto Bodily Injury line of business is used, which features slow claim development and high volatility. The AUSI dataset consists of transactional data for individual claims, which we aggregated into quarterly triangles. We use quarterly data from January 2005 to December 2014, which provides a 36×\times36 upper triangle and a 4 quarter forecasting period, which will be used to compare the MDN’s results to the ccODP. Using the individual claims data, each claim was randomly allocated to one of ten triangles, meaning that ten aggregate triangles of roughly equal size were created using this dataset. As mentioned earlier, running the MDN on multiple triangles better assessed the model’s consistency. Figure 8 plots the incremental claims for this dataset. This environment aims to assess the MDN’s ability to provide accurate forecasts for real data.

Refer to caption

Figure 8: A plot of the incremental claims of the AUSI environment, for selected accident quarters. Solid lines represents data in the upper triangle, while dashed lines represent data in the lower triangle

4 Model Training and Evaluation

The SynthETIC Simulator produces individual claims, which are aggregated into a 40x40 triangle. It is a common procedure in neural network modelling to standardise the input variables, in order to stabilise the training. Through early experimentation, normalising the response as well (Xi,jX_{i,j}) was crucial in achieving convergence during training.

For each hyperparameter combination tested, 𝜽={λw,λσ,p,n,h,K}\boldsymbol{\theta}=\{\lambda_{w},\lambda_{\sigma},p,n,h,K\}, an MDN with such hyperparameters is trained on the training set of each partition, then projected on to the testing set. The loss function we selected is the Negative Log-Likelihood, with weight and sigma activation penalties applied during training:

Loss(X,X^|w,λw,λσ)=1|X|i,j:Xi,jXln(fXi,j^(Xi,j|w))+λwww+λσi,j:Xi,jXk=1Kσi,j,k2.Loss(\textbf{X},\hat{\textbf{X}}|\textbf{w},\lambda_{w},\lambda_{\sigma})=-\frac{1}{|\textbf{X}|}\sum_{i,j:X_{i,j}\in\textbf{X}}ln(f_{\hat{X_{i,j}}}(X_{i,j}|\textbf{w}))+\lambda_{w}\textbf{w}\cdot\textbf{w}+\lambda_{\sigma}\sum_{i,j:X_{i,j}\in\textbf{X}}\sum_{k=1}^{K}\sigma_{i,j,k}^{2}. (4.1)

Via experimentation, the Adam optimiser (Kingma and Ba, 2014), with a learning rate of 0.001 provided the most stable training compared with other optimisers like RMSProp and Stochastic Gradient Descent. To further minimise over-fitting, Early Stopping was used to stop training as soon as the validation loss was minimised (Goodfellow et al., 2016). The validation loss rarely decreased steadily, hence training was only stopped when it did not hit new lows in the last 1000 epochs. This is referred to as the patience measure in the Keras interface; a lower patience than 1000 would sometimes prematurely stop training. Training would usually last for several thousand epochs, with higher dropout rates and larger networks often requiring up to 10-15 thousand iterations. A 10000 epoch limit was set when running the hyper-parameter optimisation algorithm, in order to increase efficiency.

4.1 Test error

Denote 𝜽\boldsymbol{\theta} as the hyper-parameter values of the MDN being run. In addition, let fX^i,j(x|w,θ)f_{\hat{X}_{i,j}}(x|\textbf{w},\theta) be the density of X^i,j\hat{X}_{i,j} projected by an MDN with hyper-parameters 𝜽\boldsymbol{\theta} and weights w. Let T1T1 and T2T2 be the set of cells (i,j)(i,j) in the testing set of the first and second partitions, respectively. A separate MDN is trained TT times in each partition; let wp,t\textbf{w}_{p,t} be the weights of the ttht^{th} model trained on the pthp^{th} partition. The test error of the MDN with hyper-parameters 𝜽\boldsymbol{\theta} is calculated from (4.2) - (4.4).

TestError(𝜽,Partition 1)=1T|T1|t=1Ti,j:(i,j)T1ln(fX^i,j(Xi,j|w1,t,𝜽))\text{TestError}(\boldsymbol{\theta},\text{Partition }1)=-\frac{1}{T|T1|}\sum_{t=1}^{T}\sum_{i,j:(i,j)\in T1}ln(f_{\hat{X}_{i,j}}(X_{i,j}|\textbf{w}_{1,t},\boldsymbol{\theta})) (4.2)
TestError(𝜽,Partition 2)=1T|T2|t=1Ti,j:(i,j)T2ln(fX^i,j(Xi,j|w2,t,𝜽))\text{TestError}(\boldsymbol{\theta},\text{Partition }2)=-\frac{1}{T|T2|}\sum_{t=1}^{T}\sum_{i,j:(i,j)\in T2}ln(f_{\hat{X}_{i,j}}(X_{i,j}|\textbf{w}_{2,t},\boldsymbol{\theta})) (4.3)
TestError(𝜽)=|T1|TestError(𝜽,Partition 1)+|T2|TestError(𝜽,Partition 2)|T1|+|T2|\text{TestError}(\boldsymbol{\theta})=\frac{|T1|*\text{TestError}(\boldsymbol{\theta},\text{Partition }1)+|T2|*\text{TestError}(\boldsymbol{\theta},\text{Partition }2)}{|T1|+|T2|} (4.4)

Hence, the MDN is trained 2T2T times for each set of hyper-parameters θ\theta, as each run has a different weight initialisation and hence a different fit. Averaging the error of these runs reduces the impact of random weight initialisations on the performance of the hyper-parameter set 𝜽\boldsymbol{\theta}.

4.2 Projection constraints

This paper implements a mechanism for directly constraining central estimates of cells in the lower triangle, thereby directly controlling projections made by the MDN. This method allows for the practitioner’s judgement to be incorporated if required. The practitioner can place upper and lower bounds on the central estimates of any desired set of cells (i,j)(i,j) in the lower triangle and penalise the MDN if its central estimates fall outside those boundaries.

Let C be the set of cells (i,j)(i,j) in the lower triangle, which have had constraints placed on their projections. Let Ci,jLowerC_{i,j}^{Lower} and Ci,jUpperC_{i,j}^{Upper} be the lower and upper constraints of the central estimates for cell (i,j)C(i,j)\in\textbf{C}. Let μ^i,j=E[X^i,j]\hat{\mu}_{i,j}=E[\hat{X}_{i,j}]. The loss function during training follows (4.6):

NLLLoss(X,X^|w)\displaystyle NLLLoss(\textbf{X},\hat{\textbf{X}}|\textbf{w}) =1|X|i,j:Xi,jTrainln(fX^i,j(Xi,j|w))\displaystyle=-\frac{1}{|\textbf{X}|}\sum_{i,j:X_{i,j}\in Train}ln(f_{\hat{X}_{i,j}}(X_{i,j}|\textbf{w})) (4.5)
+Regularisation+λC|C|i,j:(i,j)C[max(0,μ^i,jCi,jUpper)]2+[max(0,Ci,jLowerμ^i,j)]2\displaystyle+Regularisation+\frac{\lambda_{C}}{|\textbf{C}|}\sum_{i,j:(i,j)\in\textbf{C}}[max(0,\hat{\mu}_{i,j}-C_{i,j}^{Upper})]^{2}+[max(0,C_{i,j}^{Lower}-\hat{\mu}_{i,j})]^{2} (4.6)

Where Regularisation=λwww+λσi,j:Xi,jXTraink=1Kσi,j,k2Regularisation=\lambda_{w}\textbf{w}\cdot\textbf{w}+\lambda_{\sigma}\sum_{i,j:X_{i,j}\in\textbf{X}_{Train}}\sum_{k=1}^{K}\sigma_{i,j,k}^{2} from (4.1), and λC\lambda_{C} is a constraint violation penalty coefficient. The constraints apply a square distance penalty to the loss function if the central estimate of constrained cells in the lower triangle violate the constraints. With a sufficiently high penalty coefficient, the MDN’s projection will satisfy the constraints specified, providing projections that are more reasonable. The cells in C are randomly split in half between the training and validation sets, as the validation loss should indicate how well the projection constraints have been met in order for Early Stopping to be used effectively.

4.3 Fitting the final model

Once all desired hyper-parameter combinations are tested, the combination with the lowest test error, 𝜽min\boldsymbol{\theta}^{min} (see Section 3.2 for details), is set as the model architecture of choice. To produce distributional forecasts of claims in the lower triangle, the chosen MDN is run on the entire upper triangle. Only a training/validation split is needed, since the testing set was only used to compare different hyper-parameters. The training/validation partition of the upper triangle was done sequentially, visualised in Figure 9.

Refer to caption

Figure 9: The training/validation partition of the upper triangle. The chosen MDN design is fit on the training data and used to project claims in the lower triangle

An MDN with hyper-parameters 𝜽min\boldsymbol{\theta}^{min} is fit 5 times on the training data of Partition 3, under different weight initialisations. The 5 fitted distributions are ensembled to produce the final forecast. This ensemble of models produces more robust results and reduces the error of bad runs that stop training at a poor local minimum of loss (Perrone and Cooper, 1993). Let wzw_{z} be the set of the MDN’s final weights in the zthz^{th} run. The 5 fits are ensembled to produce the distribution of incremental claims shown in (4.7).

fX^i,j(x)=15z=15k=1Kαi,j,kwzϕ(x|μi,j,kwz,σi,j,kwz)f_{\hat{X}_{i,j}}(x)=\frac{1}{5}\sum_{z=1}^{5}\sum_{k=1}^{K}\alpha_{i,j,k}^{w_{z}}\phi(x|\mu_{i,j,k}^{w_{z}},\sigma_{i,j,k}^{w_{z}}) (4.7)

4.4 Model evaluation

The final MDN model is fit on the lower triangle and compared to the ccODP model. The results of two main variables were analysed:

  1. 1.

    Individual cells, Xi,jX_{i,j}

  2. 2.

    Total reserves, R=i,j:i+j>41Xi,jR=\sum_{i,j:i+j>41}X_{i,j}

Capital standards set by APRA and Solvency II require reserve allocations to meet total reserves in the lower triangle with a 75% and 99.5% probability of sufficiency, respectively. Hence measuring the accuracy of total reserves is important. However, it is also desirable for a model to achieve accurate total reserves by correctly modelling the individual cells, Xi,jX_{i,j}.

Results were analysed qualitatively and quantitatively. Qualitative analysis allowed the MDN’s strengths and weaknesses to be located graphically, while quantitative analysis provided a more objective measure of the model’s accuracy. The qualitative analysis conducted included the following plots:

  1. 1.

    Central Estimates: Plots of the MDN and ccODP’s central estimates μ^i,j\hat{\mu}_{i,j} were compared to actual losses of the dataset Xi,jX_{i,j}, as well as the empirical mean calculated from hundreds of simulations of the same dataset.

  2. 2.

    Risk Margins: Plots of the MDN and ccODP’s mean-centred risk margins (at the 25%, 75% and 95% level) were compared to empirical risk margins.

  3. 3.

    Total reserves: The distributions of total reserves estimated by the MDN and ccODP, R^\hat{R}, were plotted alongside the empirical distribution of total reserves.

When analysing individual cells, Xi,jX_{i,j}, the RMSE, Log Score and Quantile Score statistics were calculated on each loss triangle involved in the modelling. For total reserves, RR, the MDN and ccODP was fit on 50 triangles for each of the four simulated data environments, and 10 independent triangles randomly partitioned from the AUSI dataset, to generate reserve estimates for each triangle, R^i\hat{R}_{i} for i=1,2,3..50i=1,2,3..50. Let X={Xi,j:i+j>41}\textbf{X}=\{X_{i,j}:i+j>41\}, R={Ri,i=1,2,3,..50}\textbf{R}=\{R_{i},i=1,2,3,..50\}, fX^={fX^i,j:Xi,jX}f_{\hat{\textbf{X}}}=\{f_{\hat{X}_{i,j}}:X_{i,j}\in\textbf{X}\} and XqX_{q} be the qthq^{th} quantile estimate of the variable XX. The quantitative metrics used are calculated as follows:

1. Distributional forecast accuracy, using the log score metric ((4.8))

:

LogScore(X,fX^)=(i,j):Xi,jXln(fX^i,j(Xi,j))|X|LogScore(\textbf{X},f_{\hat{\textbf{X}}})=\frac{\sum_{(i,j):X_{i,j}\in\textbf{X}}ln(f_{\hat{X}_{i,j}}(X_{i,j}))}{|\textbf{X}|} (4.8)

A higher log score is desirable as it indicates a more accurate distributional fit for the lower triangle. The log score wasn’t calculated when analysing total reserves, as the fitted distributions usually fell completely outside the simulated empirical distribution, setting the likelihood to 0.

2. Central estimate forecast accuracy, using the RMSE metric ((4.9), (4.10))

:

RMSE(X,X^)=(i,j):Xi,jX(Xi,jX^i,j)2|X|RMSE(\textbf{X},{\hat{\textbf{X}}})=\sqrt{\frac{\sum_{(i,j):X_{i,j}\in\textbf{X}}(X_{i,j}-\hat{X}_{i,j})^{2}}{|\textbf{X}|}} (4.9)
RMSE(R,R^)=i=1D(RiRi^)2DRMSE(\textbf{R},{\hat{\textbf{R}}})=\sqrt{\frac{\sum_{i=1}^{D}(R_{i}-\hat{R_{i}})^{2}}{D}} (4.10)

A lower RMSE indicates more accurate central estimates for the lower triangle and total reserves.

3. Quantile forecast accuracy (75% and 95%), using quantile scores ((4.11), (4.12) )

:

QS(𝑿^𝒒,X)=(i,j):Xi,jX(𝟏(Xi,j<X^i,j,q)q)(X^i,j,qXi,j)|X|QS(\boldsymbol{\hat{X}_{q}},\textbf{X})=\frac{\sum_{(i,j):X_{i,j}\in\textbf{X}}(\mathbf{1}(X_{i,j}<\hat{X}_{i,j,q})-q)(\hat{X}_{i,j,q}-X_{i,j})}{|\textbf{X}|} (4.11)
QS(𝑹^𝒒,R)=i=1D(𝟏(Xi,j<X^i,j,q)q)(X^i,j,qXi,j)DQS(\boldsymbol{\hat{R}_{q}},\textbf{R})=\frac{\sum_{i=1}^{D}(\mathbf{1}(X_{i,j}<\hat{X}_{i,j,q})-q)(\hat{X}_{i,j,q}-X_{i,j})}{D} (4.12)

A lower quantile score indicates more accurate quantile estimates, for individual cells and total reserves.

5 Results

In this section, we analyse the results of the MDN, with the ResMDN separately analysed in Section 6.

5.1 Stable forecasts - rolling origin model validation

Generally, the set of hyper-parameters selected by the rolling origin method produced very reasonable and accurate central and distributional forecasts. All models were successful in predicting a decrease in the mean and volatility of claims in the later development quarters (DQs), which is a significant achievement given the low quantity of data available to the MDN in those periods. Figure 10 plots the MDN’s mean and volatility estimates on environment 2, showing the accuracy of projections despite its systematic complexity.

Refer to caption

Figure 10: Environment 2 (speed up in claim processing): Plots of the mean (red) and standard deviation (black dotted) estimates of the MDN against actual losses (blue). The grey area represents the lower triangle, the forecasting region. These plots show the MDN producing reasonable and accurate forecasts, which were consistently observed.

The MDN also produced smooth, robust predictions. This can especially be seen in Figure 11, which plots the MDN’s and ccODP’s fits to the highly volatile environment 4. The ccODP’s coefficients are derived from few data points, especially in the later AQs and DQs, leading to a more volatile fit throughout. Meanwhile, the MDN produced a more holistic fit, resulting in a smoother more robust forecast.

Refer to caption

Figure 11: Environment 4: Plots of the MDN’s central estimate (red) and standard deviation (dotted black) fits against the ccODP model (green) fit and actual losses (blue). The grey area represents the lower triangle, the forecasting region. These plots demonstrate the smooth and robust forecasts produced by the MDN, especially relative to the ccODP, despite the volatile data given.

The rolling origin method, in the third partition, uses the latest calendar quarters for validation. A model that overfits the data will not project accurately, and hence the MDN is encouraged to produce a smooth fit. In addition to Figure 11, the smoothness can be visualised in Figure 12, where the MDN produces a significantly smooth fit despite the huge volatility present in the dataset.

The rolling origin model validation method proved successful at partitioning triangles of a size as small as 36×\times36. The scarcity of data relative to the large datasets to which neural networks are usually applied would normally discourage the use of this method. However, both the MDN and rolling origin partition performed well on less than 700 data points, showing their appropriateness in a practical loss triangle reserving setting.

Refer to caption

Figure 12: AUSI: Plotting the MDN’s central estimate (red) and standard deviation (black dotted) fits against actual losses (blue). The grey area represents the lower triangle, the forecasting region. The MDN provides smooth and accurate forecasts using real data.

However, as suspected, the rolling origin method was visibly unable to capture the inflation shock accurately for environment 3, due to a lack of training data in the later calendar periods. This shortcoming was detected through analysis of the residuals in the upper triangle, which were consistently negative in the high inflation periods. Hence, an adjusted data partition methodology was implemented for environment 3, visualised in Section B, which allocated more training data to the later calendar periods, allowing the MDN more exposure to the inflation shock. This adjustment enabled the MDN to capture the later trends more effectively, and is recommended for data sets where a significant change in the claim pattern is observed in later calendar periods.

Refer to caption

Figure 13: Environment 2: Plots comparing the mean estimates of the MDN (red) and ccODP (green) models to the empirical mean claims based on 250 simulations (black). The grey area represents the lower triangle, the forecasting region. The MDN captured the increase in claim settlement speed, while the ccODP did not.

Refer to caption

Figure 14: Environment 3: Plots comparing the mean estimates of the MDN (red) and ccODP (green) models to the empirical mean claims based on 500 simulations (black). The grey area represent the lower triangle, the forecasting region. The MDN captured the inflation shock more accurately than the ccODP.

5.2 Distributional forecasting with the Mixture Density Network

The Mixture Density Network (MDN), overall, outperformed the ccODP for all environments in all qualitative and quantitative metrics. When analysing both individual cells and total reserves, the MDN outperformed the ccODP, often in a decisive manner.

Generally, the MDN’s out-performance relative to the ccODP is due to its high flexibility, being able to fit a highly flexible function to the data and capture non-linearities. This flexibility can easily lead to over-fitting, but the rolling origin model validation method ensured that the MDN’s flexibility was enough to capture the relevant trends in the data while minimising over-fitting.

5.2.1 Central estimate analysis

The MDN produced excellent central estimate projections in all environments. Where the environment had more structural heterogeneity, i.e. where the ccODP assumptions are not satisfied, the MDN decisively outperformed the ccODP in all metrics. Several key observations can be made:

  • 1.

    In environment 1, the data (by design) satisfies the ccODP assumptions well, hence the ccODP was very competitive. Nevertheless, the MDN slightly outperformed in all quantitative metrics when measuring the accuracy of incremental claims, Xi,jX_{i,j}. This can be attributed to the smooth function fit by the MDN.

  • 2.

    In environment 2, the MDN successfully learned that claims processing speed is increasing, predicting a sharper spike in claim payments in the later AQs. Figure 13 plots the results for this environment. The ccODP, assuming homogeneity in claim development, approximated claims as medium-tailed, leading to a clear over-estimation of claims in later AQs.

  • 3.

    For environment 3, the MDN accurately captured the inflation shock at calendar quarter 30 (CQ30) onwards. Figure 14 plots the results. The ccODP did not keep pace with the increased inflation due to its limited ability to handle heterogeneity, leading to its under-estimation of claims from CQ30 onwards.

  • 4.

    In both environment 4 and the AUSI environment, the MDN handled volatile data well and provided accurate central estimates, outperforming the ccODP. This accuracy is shown in Figure 15, which plots the MDN’s central estimates against the empirical mean based on 10 simulations.

Refer to caption

Figure 15: AUSI: Plots comparing the mean estimates of the MDN (red) and ccODP (green) models to the empirical mean claims based on 10 triangles (black). The grey area represent the lower triangle, the forecasting region. These plots show the MDN producing fairly accurate mean forecasts in a real environment, outperforming the ccODP.

Figure 16 provides boxplots of the MDN’s % reduction of the RMSE relative to the ccODP for 50 triangles in each of the environments tested (10 for the AUSI environment). The boxplots show that the MDN achieved a positive reduction in the RMSE (lower RMSE) relative to the ccODP for the majority of triangles in each environment, further showing the MDN’s higher forecasting accuracy.

Refer to caption

Figure 16: Left: conventional boxplots displaying the MDN’s RMSE as a percentage of ccODP’s for each of the 50 triangles run for environments 1,2,3,4, and 10 triangles for AUSI. Right: conventional boxplots displaying the MDN’s increase in log score relative to the ccODP for each of the 50 triangles run for environments 1,2,3,4, and 10 triangles for AUSI.

Despite the MDN’s success, it showed weaknesses in several areas, some of which were mitigated:

  • 1.

    Using the NLL loss function alone can encourage the MDN to over-estimate the volatility when its central estimate is inaccurate. This was seen in environment 2, where a MDN with a NLL loss function under-estimates claims in the (40,2) cell, leading to an excessively high volatility estimate for that region. In addition to sigma activity regularisation (see Section 3.2, adding an MSE term to the loss function helped to resolve this issue, as it encouraged the MDN to achieve more accurate central estimates. Hence, an MSE term was added to the loss function for environments 1 and 2, and is recommended for loss triangles with sharp shifts in claims development.

  • 2.

    Taking the log of aggregate claims linearised the data, which often led to faster and more accurate modelling. In this paper, environments 1 and 3 were fit with a mixture Log-Gaussian, as they showed significantly more accurate results, especially capturing the claims decay in later DQs.

Refer to caption

Figure 17: Environment 2: Plots comparing the 25% (solid) and 75% (dashed) risk margin estimates of the MDN (red) and ccODP (green) models to the empirical margins based on 250 simulations (black). The grey area represents the lower triangle, the forecasting region. These plots demonstrate the MDN providing more accurate volatility forecasts than the ccODP benchmark.

Refer to caption

Figure 18: Environment 3 (Inflation Shock): Plots comparing the 25% (solid), 75% (dashed) and 95% (dotted) risk margin estimates of the MDN (red) and ccODP (green) models to the empirical margins based on 500 simulations (black). The grey area represents the lower triangle, the forecasting region. Similarly to Figure 17, this figure demonstrates the MDN’s ability to capture the increase in claim volatility due to the inflation shock, while the ccODP did not.

5.2.2 Volatility estimate analysis

The Mixture Density Network produced very smooth and accurate volatility estimates. Where noise in the data was low, the MDN projected low volatility, and vice versa. Overall, it outperformed the ccODP qualitatively and quantitatively in estimating the volatility of individual cells.

In relation to individual cells, the MDN’s risk margin estimates at the 25th and 75th percentile were more accurate overall than the ccODP’s margins in almost all environments tested. The ccODP’s variance is a function of its mean, and hence it failed where the central estimates failed. For example, in environment 2 (Figure 17), the ccODP over-estimates claims in later AQs, which led to it over-estimating margins in that same period. In environment 3, the ccODP under-estimated claims in later calendar Quarters (CQs), as it didn’t effectively capture the inflation shock. This led to volatility estimates also being too low in those periods, as Figure 18 illustrates. The MDN dealt with these structural issues more effectively, leading to more accurate dispersion estimates of claims. In the volatile AUSI environment, the MDN produced smooth and accurate margin estimates, as Figure 19 illustrates.

Refer to caption

Figure 19: AUSI: Plots comparing the 25% (solid) and 75% (dashed) risk margin estimates of the MDN (red) and ccODP (green) models to the empirical margins based on 10 triangles (black). The grey area represents the lower triangle, the forecasting region. This figure demonstrates the MDN producing accurate and smooth volatility forecasts on real data.

While some correlation was found in the results between central and volatility forecasts, the MDN allows a large degree of independence between the μ\mu and σ\sigma parameters estimated, allowing it to fit a much wider range of distributions than the ccODP.

Figure 16 provides boxplots of the MDN’s increase in the log score relative to the ccODP for 50 triangles in each of environments 1,2,3 and 4. The AUSI boxplot was based on 10 triangles. The boxplots show that the MDN achieved a higher log score relative to the ccODP for the majority of triangles in each environment, indicating a more accurate probabilistic forecast.

With the MDN’s success, there are some weaknesses which must be addressed:

  • 1.

    The MDN still showed signs of attributing noise to systematic trends. In environment 2, even though the DQ2 spike was fixed, the volatility was still too high for AQ30 and AQ40.

  • 1.

    Similar to central estimates, the MDN often over-estimates volatility in later DQs, also due to a lack of data in that region.

5.2.3 Quantile estimate analysis

The MDN provided more accurate 75% and 95% quantiles for all environments in the majority of triangles run for each environment. These results follow from the MDN’s ability to provide more accurate central estimates and volatility estimates. The quantile analysis was mainly quantitative, using the quantile scores. Table 1 confirms that in all environments, the MDN reduces the 75% and 95% quantile scores for the majority of triangles, indicating more accurate quantile estimates at the 75% and 95% levels.

The MDN and ccODP models were run on fifty triangles of each of environments 1,2,3,4, and the ten triangles partitioned from the AUSI data. The quantitative metrics are calculated for the 50 triangles and averaged, with results between the MDN and ccODP models compared in Table 1. As the table shows, the MDN, on average, had a lower RMSE and Quantiles Scores and had a higher log score for each environment, which is a significant out-performance by the MDN. Table 2 further reinforces these results by showing the percentage of triangles in which the MDN outperformed the ccODP for each quantitative metric. In each environment, the MDN outperforms the ccODP in each metric for the majority of triangles.

Environment Model Mean RMSE RMSE Mean LS Mean QS Mean QS
(% of ccODP) (75%) (95%)
1 ccODP 1,656,921.0 100 -14.99 380,968.5 144,423.7
1 MDN 1,527,799 92.2 -14.93 375,413.6 140,754.3
2 ccODP 591,505.4 100 -14.97 111,733.3 31,213.2
2 MDN 182,041.1 30.8 -13.31 50,628.9 16,326.1
3 ccODP 190,482.4 100 -13.46 57,562.2 32,164.1
3 MDN 162,621.8 85.4 -13.05 47,837.0 18,817.4
4 ccODP 1,053,008.0 100 -14.72 232,011.3 96,506.1
4 MDN 652,230.7 61.9 -13.88 210,694.1 95,235.0
AUSI ccODP - 100 -14.04 124,976.3 58,760.6
AUSI MDN - 84.2 -13.07 105,559.9 53,773.9
Table 1: The average score, over 50 triangles, of each quantitative metric; the RMSE, log score (LS) and quantile scores (QS) for the 75% and 95% levels. The MDN outperformed the ccODP in all environments and metrics when the average is taken.
Environment Model RMSE Log Score Quantile Score (75%) Quantile Score (95%)
1 MDN 76 80 60 58
2 MDN 100 100 100 100
3 MDN 88 94 90 100
4 MDN 84 98 66 50
AUSI MDN 100 100 100 90
Table 2: The percentage of triangles in which the MDN outperformed the ccODP for each environment and metric.

5.2.4 Total reserves

The MDN, in all environments except environment 1, showed more accurate central and dispersion estimates of total reserves compared to the ccODP estimate (the dispersion accuracy is qualitatively measured through visualising reserve density plots, Figure 20 plots these results). An empirical distribution of total reserves, based on many simulations, is used as an estimate of the actual distribution of RR. This out-performance follows from the MDN modelling the mean and volatility of individual cells more accurately than the ccODP. Because the claims in environment 1 are homogeneous in development, the ccODP provides highly competitive results and the MDN didn’t outperform. However, for the more complicated environments 2,3 and 4, the MDN had more accurate 75% and 95% quantiles of total reserves compared to the ccODP. Table 3 calculated the quantitative metrics for both models for total reserve estimates, R^\hat{R}. The qualitative analysis (Figure 20) also supports these results.

Refer to caption

Figure 20: A plot of the total reserve density estimates for all environments, R^\hat{R}, showing the MDN (red) and ccODP’s (green) estimated densities against the empirical density (black) based on hundreds of simulations. For each environment, only one triangle is analysed for each plot. The MDN consistently provides more accurate results, except for environment 1
Environment Model RMSE (×106\times 10^{6}) QS(75%) (×106\times 10^{6}) QS(95%) (×106\times 10^{6})
1 MDN 111.7 37.3 15.4
1 ccODP 92.9 28.9 13.2
2 MDN 80.4 20.1 4.25
2 ccODP 260.2 65.9 13.45
3 MDN 39.5 19.7 20.00
3 ccODP 53.0 37.0 44.37
4 MDN 99.4 24.1 9.00
4 ccODP 322.8 56.5 11.98
Table 3: The RMSE and quantile scores (QS) at the 75% and 95% levels, calculated for total reserve estimates, R^\hat{R}. The ccODP outperforms for Environment 1, but the MDN outperforms otherwise.

6 Practical Considerations

6.1 Projection constraints

For both the MDN and ResMDN, central estimate projections in the lower triangles can be explicitly constrained for any desired set of cells, Xi,jX_{i,j}. Without constraints these projections can, and sometimes do, produce negative results in individual cells, and even for total reserves for some accident years (as was seen for the ResMDN in Section 6.2). This is unrealistic in some circumstances, and is best prevented by constraining projections to non-negativity. Similarly, it may be desirable to force projected payments to converge toward zero with increasing DQ. There might be other “reasonableness constraints” that the actuary wishes to apply. In this paper, projection constraints were applied successfully to the ResMDN for environments 2 and 3. Mean forecasts were constrained in the later DQs to be non-negative. There are some points to note:

  • 1.

    The MDN’s forecast was virtually unchanged in the upper triangle, meaning the constraints set did not distract the model from fitting the in-sample data accurately.

  • 2.

    The MDN produced a natural, smooth curve while still meeting the constraints. There was no evidence of a sudden jolt in the fitted function, nor did the function simply rest on the closest constraint boundary. Consequently, only a small proportion of cells need to be constrained to achieve reasonable results. For example, the non-negativity constraint described above was only applied to approximately 10% of cells in the lower triangle.

To visually demonstrate the effect of constraining projections, we apply this methodology in environment 4, shown in Figure 21. The volatile data in this environment caused the MDN to occasionally over-estimate losses in the later DQs. Hence, the mean was constrained in that period to be approximately 0, which the model followed once set. In general, the actuary can set any desirable boundary to any cell in the lower triangle, to ensure the MDN strongly leans towards fitting functions with sensible projections.

Refer to caption

Figure 21: Environment 4: A plot of the central estimates forecasted by the MDN (red) and MDN with projection constraints (orange), against the actual losses (blue). The grey area represents the lower triangle, the forecasting region. This figure demonstrates the MDN producing more reasonable results when projections are constrained.

6.2 Interpretability: ResMDN

The ResMDN shows plenty of potential in boosting the residuals of its GLM backbone while providing more interpretable results than the MDN. In this paper, we analyse the ResMDN on environments 2 and 3, as in both environments, the ccODP provided a smooth backbone with clear and detectable flaws which could be analysed. In both environments, the ResMDN successfully detected the ccODP’s shortcomings and corrected them, to an extent. Both mean and volatility estimates were corrected; the ResMDN didn’t just boost the mean, but also the distribution. A visible shortcoming of the ResMDN is that it can produce unreasonable forecasts; for environment 2 the model occasionally under-estimates claims in later DQs, while for environment 3 there were some instances where it over-estimated the inflation shock. Hence we apply projection constraints in the later DQs to mitigate these deviations.

  • 1.

    The ResMDN demonstrated the ability to recognise errors in the ccODP’s central estimates and correct them. In environment 2, where the claim processing speed gradually increases, the ccODP models the claims as being middle-tailed, as it assumes homogeneity in claim development. Hence, the ccODP under-estimated claims in early AQs and over-estimated claims later on. Figures 22(a) and 22(b) show heatmaps of the ccODP and ResMDN’s residuals for environment 2, respectively. As can be seen, the ResMDN successfully understood the ccODP’s shortcomings just mentioned, and reduced the residuals. Figure 23 plots the central estimates of the ResMDN and ccODP models, also showing the ResMDN’s corrections producing a more accurate forecast.

  • 2.

    The ResMDN also demonstrated the ability to recognise errors in the ccODP’s volatility estimates and correct them. In environment 2, the ccODP models the speed up in claims processing with a higher dispersion parameter, leading to consistently over-estimating volatility. The ResMDN successfully learns this shortcoming and reduces volatility estimates accordingly. Figure 24 visualises the ResMDN’s boosting for environment 2; the ResMDN has corrected volatility to almost match the empirical trend, indicating its higher distributional forecast accuracy.

Refer to caption
(a) A heatmap of the ccODP’s residuals
Refer to caption
(b) A heatmap of the ResMDN’s residuals
Figure 22: Environment 2: Heatmaps showing the ccODP’s initial residuals in (a), calculated as μi,jccODPXi,j\mu_{i,j}^{ccODP}-X_{i,j}. The ResMDN’s residuals, calculated as μi,jResMDNXi,j\mu_{i,j}^{ResMDN}-X_{i,j}, are shown in (b). The lighter colours in (b) show that the ResMDN partially corrected the ccODP’s residuals, producing more accurate forecasts.

Refer to caption


Figure 23: Environment 2: Plots comparing the mean estimates of the ResMDN (orange) and ccODP (green) models to actual losses (blue). The grey area represents the lower triangle, the forecasting region. These plots demonstrate the ResMDN partially correcting the ccODP’s residuals, leading to more accurate forecasts.

Refer to caption

Figure 24: Environment 2: Plots comparing the 25% (solid) and 75% (dashed) risk margin estimates of the ResMDN (orange) and ccODP (green) models to the empirical margins based on 250 simulations (black). The grey area represents the lower triangle, the forecasting region. While Figure 23 demonstrates the ResMDN’s ability to correct the ccODP’s mean estimates, this figure shows the ResMDN correcting the ccODP’s volatility estimates, allowing distributional forecasts to be more in line with empirical data.

The ResMDN showed promise in detecting and partially correcting the embedded GLM’s mean and volatility estimates. Given these results, several considerations arose that are worth mentioning:

  • 1.

    The ResMDN projected the ccODP’s residuals unreasonably in some instances. For example, in environment 2, it learnt that the ccODP over-estimates claims in later AQs, so it adjusted this error by reducing the loss estimates for those periods, and continued that correction excessively in the later DQs. This issue was fixed by constraining projections for DQ 38-40. Central estimate bounds for these periods were set between 0-500,000 (for environment 2) and 0-200,000 (for environment 3). These bounds are judgemental, but are reasonably wide to accommodate the uncertainty in forecasting. Figure 25 plots the difference in forecasts between the constrained and unconstrained ResMDN. While the plotted triangle is an extreme scenario (out of the 50 triangle sample), it justifies using the constrained ResMDN to ensure incremental claims tend to 0.

  • 2.

    Despite the ResMDN correcting residuals, its log score is disadvantaged compared to the ccODP. This is mainly driven by the high coefficient of variation of incremental claims observed in the later DQs (of which only a small amount are in-sample), which naturally favoured the ODP distribution over the mixture Gaussian.

Refer to caption

Figure 25: Environment 2: Plots comparing the mean estimates of the constrained (orange) and unconstrained (purple) ResMDN models against the ccODP backbone (green) and actual losses (blue). The grey area represents the lower triangle, the forecasting region. These plots show that the unconstrained ResMDN can significantly under-estimate claims, hence justifying the use of constraints to stabilise forecasts.

Tables 4, 5 and 6 display the quantitative results of the unconstrained ResMDN (ResMDN) and constrained ResMDN (ResMDN-PC), in a similar fashion to Tables 1, 2 and 3. In accordance with conclusions from visual analysis in Figures 23 and 24, the ResMDN produced more accurate central estimates, 75th and 95th quantiles when looking at individual cells, demonstrating that it has improved the mean and distribution over its ccODP backbone. In the majority of triangles, the ResMDN achieved a more accurate RMSE and Quantile Score. Boosting residuals translated to more accurate total reserve estimates, for the mean, 75th and 95th quantiles. We note that the unconstrained ResMDN yielded more accurate central estimates for environment 2 than the constrained model. This is due to the constraints slightly discouraging boosting to avoid negative forecasts. Moreover, the apparent improvement in performance embodied in the reduced RMSE can come at the cost of negative projections.

Environment Model Mean RMSE RMSE Mean LS Mean QS Mean QS
(% of ccODP) (75%) (95%)
2 ResMDN 378,898.3 64.1 -15.23 81,641.2 30,300.2
2 ResMDN-PC 436,414.9 73.8 -14.95 78,584.0 21,644.8
3 ResMDN 190,471.4 100.0 -14.01 55,497.3 25,348.4
3 ResMDN-PC 182,944.4 96.0 -13.65 53,003.2 26,971.7
Table 4: The average score, over 50 triangles, of each quantitative metric; the RMSE, log score (LS) and quantile scores (QS) for the 75% and 95% levels for the ResMDN. The ResMDN-PC model features constrained projections
Environment Model RMSE Log Score Quantile Score (75%) Quantile Score (95%)
2 ResMDN 96 58 90 72
2 ResMDN-PC 96 60 96 100
3 ResMDN 78 38 80 76
3 ResMDN-PC 90 46 92 76
Table 5: The percentage of triangles in which the ResMDN outperformed the ccODP in each metric. The ResMDN-PC model features constrained projections
Environment Model RMSE (×106\times 10^{6}) QS(75%) (×106\times 10^{6}) QS(95%) (×106\times 10^{6})
2 ResMDN 101.1 (39%) 27.0 (41%) 13.69 (102%)
2 ResMDN-PC 180.2 (69%) 44.85 (68%) 9.13 (68%)
3 ResMDN 55.9 (105%) 19.9 (54%) 18.3 (41%)
3 ResMDN-PC 36.7 (69%) 22.3 (60%) 25.8 (58%)
Table 6: The RMSE and quantile scores (QS) at the 75% and 95% levels, calculated for total reserve estimates, R^\hat{R}. The number in brackets represents the corresponding metric as a %\% of the corresponding ccODP results. The ResMDN-PC model features constrained projections

6.2.1 Comparison to the MDN

While the ResMDN was able to detect the ccODP’s systematic errors, it generally failed to outperform the non-embedded MDN in the above examples as it only partially corrected the errors (see Figure 22 for an illustration). Nevertheless, the ResMDN showed great function and potential in correcting an embedded GLM’s mean and volatility estimates, while maintaining a fair level of interpretability. While the current implementation is not as accurate as the MDN, if this is criteria is essential then one possibility is to consider a more sophisticated GLM backbone.

Finally, it should also be noted that an additional benefit of the ResMDN is that the ResMDN’s training time was noticeably faster than the MDN. For environment 2, when a few triangles were analysed, the ResMDN finished training in 2218 epochs on average, compared to 3868 for the MDN. This 43% reduction in training time is due to the ResMDN’s GLM initialisation being more accurate than the MDN’s random starting fit, meaning that less parameter adjustment is required. This observation is in parallel to the findings of Gabrielli, Richman and Wüthrich (2020), who also noted faster training times under the CANN structured model.

7 Conclusion

In this paper, we identified, addressed, and mitigated a number of obstacles, which so far hindered the popularisation of neural networks in loss triangle reserving. A neural network design which specialises in distributional forecasting, the MDN, was applied successfully to a variety of environments. The MDN produced more accurate central and volatility estimates, for both the individual cell and total reserve measures. The rolling origin model validation method provided a framework for model testing and selection, suited to sequential data. This sequential data partition gave preference to smooth, robust models, while also producing accurate forecasts.

We considered additional extensions involving projection controls and hybrid GLM-MDN approaches. We demonstrated that the MDN was able to significantly outperform the ccODP model in a variety of environments and metrics. While the ccODP model is not representative of the full potential of GLMs in loss reserving, it is a gold standard, and as such the results are compelling.

Acknowledgements

Earlier versions of this paper were presented at the Actuaries Institute 2021 Virtual Summit, and at the ASTIN Online Colloquium. The authors are grateful for constructive comments received from colleagues who attended those events.

This research was supported under Australian Research Council’s Linkage (LP130100723, with funding partners Allianz Australia Insurance Ltd, Insurance Australia Group Ltd, and Suncorp Metway Ltd) and Discovery (DP200101859) Projects funding schemes. The views expressed herein are those of the authors and are not necessarily those of the supporting organisations.

Data and Code

We are unable to provide the dataset that was used in the empirical case study due to confidentiality. However, simulated data sets with similar features, as well as all relevant codes, can be found at https://github.com/agi-lab/reserving-MDN-ResMDN.

References

References

  • Avanzi et al. (2021) Avanzi, B., Taylor, G.C., Wang, M., Wong, B., 2021. SynthETIC: an individual insurance claim simulator with feature control. Upcoming, Insurance: Mathematics and Economics .
  • Balona and Richman (2020) Balona, C., Richman, R., 2020. The actuary and ibnr techniques: A machine learning approach. Available at SSRN 3697256 .
  • Bergmeir and Benítez (2012) Bergmeir, C., Benítez, J.M., 2012. On the use of cross-validation for time series predictor evaluation. Information Sciences 191, 192–213.
  • Bishop (1994) Bishop, C.M., 1994. Mixture density networks. Aston University.
  • Bishop (2006) Bishop, C.M., 2006. Pattern Recognition and Machine Learning (Information Science and Statistics). Springer-Verlag, Berlin, Heidelberg.
  • Delong et al. (2021) Delong, L., Lindholm, M., Wüthrich, M.V., 2021. Collective reserving using individual claims data. Scandinavian Actuarial Journal In press.
  • Denuit and Trufin (2020) Denuit, M., Trufin, J., 2020. Effective statistical learning methods for actuaries, Volume 3. Springer Nature, Switzerland.
  • Gabrielli (2020) Gabrielli, A., 2020. A neural network boosted double overdispersed poisson claims reserving model. ASTIN Bulletin 50:1, 25–60.
  • Gabrielli (2021) Gabrielli, A., 2021. An individual claims reserving model for reported claims. European Actuarial Journal .
  • Gabrielli et al. (2020) Gabrielli, A., Richman, R., Wüthrich, M.V., 2020. Neural network embedding of the over-dispersed poisson reserving model. Scandinavian Actuarial Journal 2020, 1–29.
  • Gabrielli and Wüthrich (2018) Gabrielli, A., Wüthrich, M., 2018. An individual claims history simulation machine. Risks 6.
  • Goodfellow et al. (2016) Goodfellow, I., Bengio, Y., Courville, A., Bengio, Y., 2016. Deep learning. volume 1. MIT press Cambridge.
  • Harej et al. (2017) Harej, B., Gächter, R., Jamal, S., 2017. Individual claim development with machine learning. Report of the ASTIN Working Party of the International Actuarial Association .
  • He et al. (2016) He, K., Zhang, X., Ren, S., Sun, J., 2016. Deep residual learning for image recognition, in: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770–778. doi:10.1109/CVPR.2016.90.
  • Hjorth and Nabney (2000) Hjorth, L.U., Nabney, I.T., 2000. Bayesian training of mixture density networks, in: Proceedings of the IEEE-INNS-ENNS International Joint Conference on Neural Networks. IJCNN 2000. Neural Computing: New Challenges and Perspectives for the New Millennium, IEEE. pp. 455–460.
  • Kingma and Ba (2014) Kingma, D.P., Ba, J., 2014. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 .
  • Kuo (2019) Kuo, K., 2019. Deeptriangle: A deep learning approach to loss reserving. Risks 7, 97.
  • Kuo (2020) Kuo, K., 2020. Individual claims forecasting with bayesian mixture density networks. arXiv preprint arXiv:2003.02453 .
  • (19) Lakshminarayanan, B., Pritzel, A., Blundell, C., . Simple and scalable predictive uncertainty estimation using deep ensembles, in: Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R. (Eds.), Advances in Neural Information Processing Systems, Curran Associates, Inc.
  • Mulquiney (2006) Mulquiney, P., 2006. Artificial neural networks in insurance loss reserving, in: 9th Joint International Conference on Information Sciences (JCIS-06), Atlantis Press.
  • Murray et al. (2011) Murray, R.E., Ryan, P.B., Reisinger, S.J., 2011. Design and validation of a data simulation model for longitudinal healthcare data, in: AMIA Annual Symposium Proceedings, American Medical Informatics Association. p. 1176.
  • Nguyen and McLachlan (2019) Nguyen, H.D., McLachlan, G., 2019. On approximations via convolution-defined mixture models. Communications in Statistics-Theory and Methods 48, 3945–3955.
  • Nix and Weigend (1994) Nix, D., Weigend, A., 1994. Estimating the mean and variance of the target probability distribution, in: Proceedings of 1994 IEEE International Conference on Neural Networks (ICNN’94), pp. 55–60 vol.1.
  • Ormoneit and Neuneier (1996) Ormoneit, D., Neuneier, R., 1996. Experiments in predicting the german stock index dax with density estimating neural networks, in: IEEE/IAFE 1996 Conference on Computational Intelligence for Financial Engineering (CIFEr), IEEE. pp. 66–71.
  • Perrone and Cooper (1993) Perrone, M.P., Cooper, L.N., 1993. When networks disagree: Ensemble methods for hybrid neural networks, in: (Ed.), R.M. (Ed.), Artificial Neural Networks for Speech and Vision. Chapman & Hall, New York, pp. 126–142.
  • Poon (2019) Poon, J.H., 2019. Penalising unexplainability in neural networks for predicting payments per claim incurred. Risks 7, 95.
  • Ramos-Pérez et al. (2019) Ramos-Pérez, E., Alonso-González, P.J., Núñez-Velázquez, J.J., 2019. Forecasting volatility with a stacked model based on a hybridized artificial neural network. Expert Systems with Applications 129, 1–9.
  • Richman (2020) Richman, R., 2020. AI in actuarial science– a review of recent advances – part 1. Annals of Actuarial Science .
  • Richman and Wuthrich (2021) Richman, R., Wuthrich, M.V., 2021. Localglmnet: interpretable deep learning for tabular data. Available at SSRN 3892015 .
  • Rossouw and Richman (2019) Rossouw, L., Richman, R., 2019. Using machine learning to model claims experience and reporting delays for pricing and reserving. Available at SSRN 3465424 .
  • Srivastava et al. (2014) Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., Salakhutdinov, R., 2014. Dropout: a simple way to prevent neural networks from overfitting. The journal of machine learning research 15, 1929–1958.
  • Tashman (2000) Tashman, L.J., 2000. Out-of-sample tests of forecasting accuracy: an analysis and review. International journal of forecasting 16, 437–450.
  • Vossen et al. (2018) Vossen, J., Feron, B., Monti, A., 2018. Probabilistic forecasting of household electrical load using artificial neural networks, in: 2018 IEEE International Conference on Probabilistic Methods Applied to Power Systems (PMAPS), IEEE. pp. 1–6.
  • Wüthrich (2018) Wüthrich, M.V., 2018. Neural networks applied to chain–ladder reserving. European Actuarial Journal 8, 407–436.
  • Wüthrich and Merz (2019) Wüthrich, M.V., Merz, M., 2019. Yes, we cann! ASTIN Bulletin: The Journal of the IAA 49, 1–3.
  • Zen and Senior (2014) Zen, H., Senior, A., 2014. Deep mixture density networks for acoustic modeling in statistical parametric speech synthesis, in: 2014 IEEE international conference on acoustics, speech and signal processing (ICASSP), IEEE. pp. 3844–3848.

A Fitting a mixture Log-Gaussian

Lemma A.1.

Let YY be a strictly positive random variable. Suppose that X=ln(Y)X=ln(Y) follows a mixture Gaussian distribution with parameters (𝛂,𝛍,𝛔)(\boldsymbol{\alpha},\boldsymbol{\mu},\boldsymbol{\sigma}), such that:

fX(x)=k=1Kαkϕ(x|μk,σk).f_{X}(x)=\sum_{k=1}^{K}\alpha_{k}\phi({x|\mu_{k},\sigma_{k}}).

Then YY follows a mixture Log-Gaussian distribution with parameters (𝛂,𝛍,𝛔)(\boldsymbol{\alpha},\boldsymbol{\mu},\boldsymbol{\sigma}), such that:

fY(y)=1yk=1Kαkϕ(ln(y)|μk,σk).f_{Y}(y)=\frac{1}{y}\sum_{k=1}^{K}\alpha_{k}\phi(ln(y)|\mu_{k},\sigma_{k}).

Proof: By definition XX is defined by

X={Xi, w.p αi,i=1,..,K},X=\{X_{i}\textrm{, w.p }\alpha_{i},i=1,..,K\},

where 0<αi<10<\alpha_{i}<1 and i=1Kαi=1\sum_{i=1}^{K}\alpha_{i}=1. Then for some function g()g(\cdot), if we consider Y=g(X)Y=g(X), we then have by definition

Y={g(Xi), w.p αi,i=1,..,K}.Y=\{g(X_{i})\textrm{, w.p }\alpha_{i},i=1,..,K\}.

i.e. in the log-gaussian case (where we have g()g(\cdot) being the exponential function) we have YY defined as a mixture of g(Xi)g(X_{i}), which are individually log-gaussian.

B Training Data in the Later Calendar Periods

For environments where a significant shift in claim development pattern is observed in the later calendar periods (e.g. in environment 3, and observable in the generated data), the rolling origin method is less likely to capture that shift. Hence, an adjusted data partition is implemented in such situations, as Figure A illustrates. The focus remains on ensuring the validation and testing sets are in the later calendar periods, but also includes some training data there to help the MDN model the shift. The partition is made as such:

  • 1.

    All partitions used the entire upper triangle

  • 2.

    10% of the data was assigned to each of the validation and testing sets

  • 3.

    For Partition 1, the testing set was randomly assigned in the latest 11 calendar quarters. Then, half the validation set was assigned to non-testing points in the latest 11 calendar quarters, while the rest of the validation set was randomly assigned to the whole triangle.

  • 4.

    For Partition 2, the process for Partition 1 was repeated, except the testing set excluded any data points that were in the test set for Partition 1. This allowed both testing sets to cover a wider variety of data points.

  • 5.

    For Partitions 3 and 4, no testing set was assigned as the final model was trained on them. The validation set was assigned similarly to Partitions 1 and 2, in that half the set was chosen randomly from the latest 11 calendar quarters, while the other half was chosen randomly from the entire triangle.

  • 6.

    The final model was trained five times; three times under Partition 3 and twice using Partition 4

Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Refer to caption
(d)
Figure A: Diagrams of the four data partitions used when the latest calendar quarters were included in the training set

C Final Model Designs

Table A lists the most accurate model design found by the algorithm in Section 3.2.

Environment Model Component Distribution MSE weight λw\lambda_{w} λσ\lambda_{\sigma} pp nn hh KK
1 MDN Log-Gaussian 3 0 0.0001 0 60 4 2
2 MDN Gaussian 3 0.001 0.1 0.1 100 2 4
2 ResMDN Gaussian 4 0 0 0 20 2 4
3 MDN Log-Gaussian 0 0 0.0001 0 80 3 1
3 ResMDN Gaussian 4 0* 0 0.1 60 2 2
4 MDN Gaussian 0 0 0 0 40 4 3
AUSI MDN Gaussian 0 0 0 0.1 20 3 3
Table A: The most accurate model designs determined by the algorithm listed in Section 3.2. λw\lambda_{w} represents the L2 weight regularisation coefficient, λσ\lambda_{\sigma} represents the L2 activity regularisation coefficient on the σ\sigma output, pp is the dropout rate, nn is the number of neurons in each hidden layer, hh is the number of hidden layers and KK is the number of components in the mixture density. *The weight regularisation coefficient was manually adjusted to 0 in order to encourage boosting.

D Additional Notes on Modelling

While effort was taken to keep the modelling as methodical and simple as possible, there were some additional details worth mentioning regarding the methodology in this paper:

  • 1.

    The ResMDN was trained with an MSE term (with a weight of 4) to encourage the model to boost mean estimates.

  • 2.

    The number of components for the ResMDN was pre-determined, set as the number of components deemed best for the mixture Gaussian MDN on the same environment.

  • 3.

    For environments 1 and 3, fitting mixture Log-Gaussians gave visibly more accurate results than fitting a mixture Gaussian, hence that distribution was chosen for these environments.

  • 4.

    An MSE term (with a weight of 3) was added to environments 1 and 2, to assist in capturing the peak in claims in the later development quarters.

  • 5.

    The LogScore for each individual cell was restricted to a minimum of 50-50. The statistic was often heavily influenced by low-volume data points in the later DQs, which tended to score lower due to their high coefficient of variation. As the log score in the lower triangle is usually between -8 and -20, this restriction predominantly applies to these low volume data points and reduces their influence on the score.

  • 6.

    When processing the fit of a mixture Log-Gaussian distribution, the σk\sigma_{k} for k=1,2,..Kk=1,2,..K was set to a maximum of 2Var(X)2\sqrt{Var(\textbf{X})}, where X={Xi,j,i+j41}\textbf{X}=\{X_{i,j},i+j\leq 41\}. This was done to avoid high σ\sigma values which, when transformed from mixture Gaussian to mixture Log-Gaussian, would lead to unreasonably high volatility estimates.

  • 7.

    The hyper-parameter selection algorithm was only run on a single triangle from each environment. The chosen model for that triangle was used to fit all 50 triangles of that environment. This was done to increase the efficiency of modelling.

  • 8.

    When running the ResMDN on environment 3, a positive L2 penalty on the weights led to almost no boosting. Hence, the penalty was manually adjusted to 0 to encourage network activity.

  • 9.

    Out of the 210 triangles the MDN and ResMDN were fit on, the results displayed in Section 5 generally aim to illustrate the average results obtained. Of course, some triangles yielded better results, while others yielded worse results.