1. Introduction
Aggregating is a statistical technique that helps to reduce noise and uncertainty in predictors. Typically, it is applied to an i.i.d. sequence of predictors and it is justified theoretically using the law of large numbers. In many practical applications one is not in the comfortable situation of having an i.i.d. sequence of predictors to which aggregating could be applied. For this reason, Breiman (1996) combined bootstrapping and aggregating, called bagging, bootstrapping being used to generate a sequence of randomized predictors and then aggregating them to receive an averaged bootstrap predictor. The title of this paper has been inspired by Breiman (1996), in other words, it is not meant in the sense of grumbling or the like, but we are going to combine networks and aggregating to receive the nagging predictor. Thereby, we benefit from the fact that typically neural network regression models have infinitely many equally good predictors which are determined by gradient descent algorithms. These equally good predictors depend on the (random) starting point of the gradient descent algorithm, thus, starting the algorithm in i.i.d. seeds for different runs results in an infinite sequence of (equally good) neural network predictors. Moreover, other common neural network training techniques may even lead to more randomness in neural network results, for example, dropout, which relies on randomly setting parts of the network to zero during training to improve performance at test time, see Srivastava et al. (2014) for more details, or the random selection of data to implement stochastic or mini-batch gradient descent methods. On the one hand, this creates difficulties for using neural network models within an insurance pricing context, since the predictive performance of the models measured at portfolio level will vary with each run and the predictions for individual policies will vary even more, leading to uncertainty about the prices that should ultimately be charged to individuals. On the other hand, having multiple network predictors puts us in the same situation as Breiman (1996) after having received the bootstrap samples, and we can aggregate them, leading to more stable results and enhanced predictive performance. In this paper, we explore statistical properties of nagging predictors at a portfolio and at a policy level, in the context of insurance pricing.
We give a short review of related literature. Aggregation of predictors is also known as ensembling. Dietterich (2000a, 2000b) discusses ensembling within various machine learning methods such as decision trees and neural networks. As stated in Dietterich (2000b), successful ensembling crucially relies on the fact that one has to be able to construct multiple suitable predictors. This is usually achieved by injecting randomness either into the data or into the fitting algorithm. In essence, this is what we do in this study, however, our main objective is to reduce randomness (in the results of the fitting algorithms) because, from a practical perspective, we need to have uniqueness in insurance pricing, in particular, we prove rates of convergence at which the random element can be diminished. Zhou (2012) and Zhou et al. (2002) study optimal ensembling of multiple predictors. Their proposal leads to a quadratic optimization problem that can be solved with the method of Lagrange. This is related to our task below of finding an optimal meta model using individual uncertainties which we discuss in Section 5.8, below. Related empirical studies on financial data are given in Di Persio and Honchar (2016), du Jardin (2016) and Wang et al. (2011). These studies conclude that ensembling is very beneficial to improve predictive performance which, indeed, is also one of our main findings.
Organization of manuscript. In the next section we introduce the framework of neural network regression models, and we discuss their calibration within Tweedie’s compound Poisson models. In Section 3 we give the theoretical foundation of aggregating predictors, in particular, we prove that aggregating leads to more stability in prediction in terms of a central limit theorem. In Section 4 we implement aggregation within the framework of neural network predictors providing the nagging predictor. In Section 5 we empirically prove the effectiveness of nagging predictors based on a French car insurance data set. Since nagging predictors involve simultaneously dealing with multiple networks, we provide a more practical meta network that approximates the nagging predictor at a minimal loss of accuracy.
2. Feed-Forward Neural Network Regression Models
These days neural networks are state-of-the-art for performing complex regression and classification tasks, particularly on unstructured data such as images or text. For a general introduction to neural networks we refer to LeCun et al. (2015), Goodfellow et al. (2016) and the references therein. In the present work, we follow the terminology and notation of Wüthrich (2019). We design feed-forward neural network regression models, referred to in short as networks, to predict insurance claims. In the spirit of Wüthrich (2019), we understand these networks as extensions of generalized linear models (GLMs), the latter being introduced by Nelder and Wedderburn (1972).
2.1. Generic Definition of Feed-Forward Neural Networks
Assume we have independent observations(Yi,xi,vi),i=1,…,n; the variablesYidescribe the responses (here: insurance claims),xi∈X⊂Rq0 describe the real-valued feature information (also known as covariates, explanatory variables, independent variables or predictors), andvi>0are known exposures. The main goal is to find a regression functionalμ(·)that appropriately describes the expected insurance claimsE[Yi]as a function of the feature informationxi, i.e.,
X→R,xi↦μ(xi)=E[Yi].
Since, typically, the true regression functionμis unknown we approximate it by a network regression function. Under a suitable link function choiceg(·), we assume thatE[Yi]can be described by the following network of depthd∈N
xi↦gE[Yi]=β(d+1),z(d)∘⋯∘z(1)(xi),
where〈·,·〉is the scalar product in Euclidean spaceRqd+1, the operation ○ gives a composition of hidden network layersz(m),1≤m≤d, of dimensionsqm+1∈N, andβ(d+1)∈Rqd+1is the readout parameter. For a given non-linear activation functionϕ:R→R, the m-th hidden network layer is a mapping
z(m):{1}×Rqm−1 →{1}×Rqm ,x↦z(m)(x)=1,z1(m)(x),…,zqm(m)(x)⊤,
with hidden neuronszj(m)(·),1≤j≤qm, being described by ridge functions
x↦zj(m)(z)=ϕβj(m),x,
for network parametersβj(m)∈Rqm−1+1. This network regression function has a network parameterβ=(β1(1),…,βqd(q),β(d+1))∈Rrof dimensionr=∑m=1d+1(qm−1+1)qm.
Based on so-called universality theorems, see, for example, Cybenko (1989) and Hornik et al. (1989), we know that networks are very flexible and can approximate any continuous and compactly supported regression function arbitrarily well if one allows for a sufficiently complex network architecture. In practical applications this means that one has to choose sufficiently many hidden layers with sufficiently many hidden neurons to be assured of adequate approximation capacity, and then one can work with this network architecture as a replacement of the unknown regression function. This complex network is calibrated to (learning) data using the gradient descent algorithm. To prevent the network from over-fitting to the learning data, one exercises early stopping of this calibration algorithm by applying a given stopping rule. This early stopping of the algorithm before convergence typically implies that one cannot expect to receive a “unique best” model, in fact, the early stopped calibration depends on the starting point of the gradient descent algorithm and, usually, since each run has a different starting point, one receives a different solution each time early-stopped gradient descent is run. As described in Section 3.3.1 of Wüthrich (2019), this exactly results in the problem of having infinitely many equally good models for a fixed stopping rule, and it is not clear which particular solution should be chosen, say, for insurance pricing. This is the main point on which we elaborate in this work, and we come back to this discussion after Equation (7), below.
2.2. Tweedie’s Compound Poisson Model
We assume that the responseYi in Equation (1) belongs to the family of Tweedie’s compound Poisson (CP) models, see Tweedie (1984), having a density of the form
Yi∼f(y;θi,vi/φ,p)=expyθi−κp(θi)φ/vi+ap(y;vi/φ),
withp∈[1,2]being the power variance (hyper-)parameter, and
vi>0isagivenexposure(weight,volume),φ>0isthedispersionparameter,θi∈ΘpisthecanonicalparameterintheeffectivedomainΘp,κp:Θp→RisacumulantfunctionoftheformEquation(3),below,ap(·;·)isthenormalization,notdependingonthecanonicalparameterθi.
Tweedie’s CP models belong to the exponential dispersion family (EDF), the latter being more general because it allows for more general cumulant functionsκ , we refer to Jørgensen (1986, 1987). We focus here on Tweedie’s CP family and not on the entire EDF because for certain calculations we need explicit properties of cumulant functions. However, similar results can be derived for other members of the EDF. We mention that the effective domainΘp⊂Ris a convex set, and that the cumulant functionκpis smooth and strictly convex in the interior of the effective domain, having the following explicit form
κp(θ)=exp(θ)forp=1,12−p((1−p)θ)2−p1−pforp∈(1,2),−log(−θ)forp=2.
p=1is the Poisson model (withφ=1),p=2is the gamma model, and forp∈(1,2)we receive compound Poisson models with i.i.d. gamma claim sizes having shape parameterγ=(2−p)/(p−1)∈(0,∞) , see Jørgensen and de Souza (1994) and Smyth and Jørgensen (2002). The first two moments are given by
μi=def.E[Yi]=κp′(θi)andVar(Yi)=φviκp″(θi)=φviVp(μi),
with power variance functionVp(μ)=μpforp∈[1,2].
Remark 1.
In this paper we work under Tweedie’s CP models characterized by cumulant functions of the form Equation (3) because we need certain (explicit) properties ofκp in the proofs of the statements below. For regression modeling one usually starts from a bigger class of models, namely, the EDF, which only requires that the cumulant function is smooth and strictly convex on the interior of the corresponding effective domainΘ, see Jørgensen (1986, 1987). A sub-family of the EDF is the so-called Tweedie’s distributions which have a cumulant function that allows for a power mean-variance relationship Equation (4) for anyp∈R∖(0,1) , see Table 1 in Jørgensen (1987), for instance,p=0is the Gaussian model,p=1is the Poisson model,p=2is the gamma model andp=3is the inverse Gaussian model. As mentioned, the models generated by cumulant function Equation (3) cover the interval(1,2) and correspond to compound Poisson models with i.i.d. gamma claim size, we also refer to Delong et al. (2020).
Relating Equation (3) to network regression function Equation (1) implies that we aim at modeling the canonical parameterθiof policy i by
xi↦θi=θ(xi)=κp′−1μ(xi)=κp′−1g−1β(d+1),z(d)∘⋯∘z(1)(xi),
and(κp′)−1is the canonical link, in contrast to the general link function g. If we choose the canonical link for g then(κp′)−1∘g−1provides the identity function in Equation (5). Network parameterβ∈Rris estimated with maximum likelihood estimation (MLE). This is achieved either by maximizing the log-likelihood function or by minimizing the corresponding deviance loss function. We use the framework of the deviance loss function here because it is more closely related to the philosophy of minimizing an objective function in gradient descent methods. The average deviance loss for independent random variablesYiis under Tweedie’s CP model assumption given by
L(D;β)=2n∑i=1nviφYi κp′−1(Yi)−κpκp′−1(Yi)−Yi θi+κp(θi)=2n∑i=1nviφYi κp′−1(Yi)−κpκp′−1(Yi)−Yi κp′−1μi+κpκp′−1μi,
for given dataD={(Yi,xi,vi);i=1,…,n}. The canonical parameter is given byθi=θ(xi,β)if we understand it as a function defined through Equation (5). Naturally we have for any mean parameterμiof any policy i
δ(Yi,μi)=def.2viφYi κp′−1(Yi)−κpκp′−1(Yi)−Yi κp′−1μi+κpκp′−1μi≥0,
because these terms subtract twice the log-likelihood of theμi-parametrized model from its saturated counterpart.δ(Yi,μi)is called the unit deviance ofYiw.r.t. mean parameterμi.
The gradient descent algorithm now tries to make the objective functionL(D;β)small through optimizing network parameterβ. This is done globally, i.e., simultaneously on the entire portfolioi=1,…,n. Naturally, two different parametersβ(1)≠β(2)withL(D;β(1))=L(D;β(2))may provide very different models on an individual policy level i. This is exactly the point raised in the previous section, namely, that early stopping in model calibration w.r.t. to a global objective functionLon observationsDmay provide equally good models on that portfolio level, but they may be very different on an individual policy level (if the gradient descent algorithm has not converged to the same extremal point of the loss function). The goal of this paper is to study such differences on individual policy level.
3. Aggregating Predictors
We introduce aggregating in this section. This can be described on one single policy i. Assume thatμ^iis a predictor for responseYi(and an estimator for mean parameterμi=E[Yi]). In general, we assume that predictorμ^iand responseYiare independent. This independence reflects that we perform an out-of-sample analysis, meaning that the mean parameter has been estimated on a learning data set that is disjoint fromYi, and we aim at performing a generalization analysis by considering the average loss described by the expected unit deviance (subject to existence)
Eδ(Yi,μ^i).
We typically assume that the predictorμ^iis chosen such that this expected generalization loss is finite.
Proposition 1.
Choose responseYi∼f(·;θi,vi/φ,p)with power variance parameterp∈[1,2]and canonical parameterθi∈Θp. Assumeμ^iis an unbiased estimator for the mean parameterμi=κp′(θi), being independent ofYi, and additionally satisfyingϵ<μ^i≤p/(p−1)μi, a.s., for someϵ∈(0,p/(p−1)μi). We have expected generalization loss
Eδ(Yi,μ^i)≥Eδ(Yi,μi).
We remark that the bounds onμ^iensure that Equation (8) is finite, moreover, the upper boundp/(p−1)μiis needed to ensure that the predictors lie in the domain such that the expected unit deviancesm↦E[δ(Yi,m)]are convex functions in m. This is needed in the following proofs.
Proof.
We calculate the expected generalization loss received by the mean of the unit deviance
Eδ(Yi,μ^i)=2viφEYi κp′−1(Yi)−κpκp′−1(Yi)−Yi κp′−1μ^i+κpκp′−1μ^i=Eδ(Yi,μi)+2viφEYi κp′−1μi−κpκp′−1μi−Yi κp′−1μ^i+κpκp′−1μ^i=Eδ(Yi,μi)+2viφhp(μi)−Ehpμ^i,
where in the last step we have used independence betweenYiandμ^iand where we use function
m>0↦hp(m)=μi κp′−1m−κpκp′−1m=μilog(m)−mforp=1,μim1−p1−p−m2−p2−pforp∈(1,2),−μi/m−log(m)forp=2.
We calculate the second derivative of this function. Forp∈[1,2], it is given by
∂2∂m2hp(m)=−pμi m−p−1−(1−p)m−p=m1+p−pμi−(1−p)m≤0,
where for the last inequality we have used that the square bracket is non-negative under our assumptions. This implies thathpis a concave function, and applying Jensen’s inequality we obtain
Eδ(Yi,μ^i)=Eδ(Yi,μi)+2viφhp(μi)−Ehpμ^i≥Eδ(Yi,μi)+2viφhp(μi)−hpEμ^i=Eδ(Yi,μi),
where in the last step we have used unbiasedness of estimatorμ^i. This finishes the proof. □
Proposition 1 tells us that the estimated modelμ^ihas an expected generalization loss Equation (8) which is bounded below by the one of the true model meanμiofYi . Using aggregating we now try to come as close as possible to this lower bound. Breiman (1996) has analyzed this question in terms of the square loss function, and further results under the square loss function are given in Bühlmann and Yu (2002). We prefer to work with the deviance loss function here because this is the objective function used for fitting in the gradient descent algorithm. For this reason we prove the subsequent results, however, in their deeper nature these results are equivalent to the ones in Breiman (1996) and Bühlmann and Yu (2002).
Assume thatμ^i(j)are i.i.d. copies of unbiased predictorμ^i. We define the aggregated predictor
μ¯i(M)=1M∑j=1Mμ^i(j).
Proposition 2.
Assume thatμ^i(j),j≥1, are i.i.d. copies ofμ^isatisfying the assumptions of Proposition 1, and being all independent fromYi. We have for allM≥1
EδYi,μ^i(1)≥EδYi,μ¯i(M)≥EδYi,μ¯i(M+1)≥Eδ(Yi,μi).
Proof.
The last bound is immediately clear because the aggregated predictors themselves fulfill the assumptions of Proposition 1, and henceforth the corresponding statement. Thus, we focus on the inequalities forM≥1. Consider decomposition of the aggregate predictor forM+1
μ¯i(M+1)=1M+1∑j=1M+1μ¯i(−j),whereμ¯i(−j)=1M∑k=1M+1μ^i(k) ??{k≠j}.
The predictorsμ¯i(−j),j≥1, are copies ofμ¯i(M), though not independent ones. We have, using functionhpdefined in Equation (9),
Eδ(Yi,μ¯i(M))=Eδ(Yi,μ¯i(M+1))+2viφEhpμ¯i(M+1)−Ehpμ¯i(M)=Eδ(Yi,μ¯i(M+1))+2viφEhp1M+1∑j=1M+1μ¯i(−j)−Ehpμ¯i(M)≥Eδ(Yi,μ¯i(M+1))+2viφE1M+1∑j=1M+1hpμ¯i(−j)−Ehpμ¯i(M)=Eδ(Yi,μ¯i(M+1)),
where the inequality follows from applying Jensen’s inequality to the concave functionhp, and the last identity follows from the fact thatμ¯i(−j),j≥1, are copies ofμ¯i(M). This finishes the proof. □
Proposition 2 says that aggregation works, i.e., aggregating i.i.d. predictors Equation (10) leads to a monotonically decreasing expected generalization loss. Moreover, notice that the i.i.d. assumption can be relaxed, indeed, it is sufficient that everyμ¯i(−j)in the above proof has the same distribution asμ¯i(M). This does not require independence between the predictorsμ^i(j),j≥1, but exchangeability is sufficient.
Proposition 3.
Assume thatμ^i(j),j≥1, are i.i.d. copies ofμ^isatisfying the assumptions of Proposition 1, and being all independent fromYi. In the Poisson casep=1we additionally assume that the sequence of aggregated predictors Equation (8) has a uniform integrable upper bound. We have
limM→∞EδYi,μ¯i(M)=ElimM→∞δYi,μ¯i(M)=Eδ(Yi,μi).
Proof.
We have the identity
Eδ(Yi,μ¯i(M))=2viφEYi κp′−1(Yi)−Eκpκp′−1(Yi)−Ehpμ¯i(M),
thus, it suffices to consider the last term. The law of large numbers implies a.s. convergencelimM→∞ hp(μ¯i(M))=hp(μi)because we have i.i.d. unbiased predictorsμ^i(j),j≥1(in particular we have consistency). Thus, it suffices to provide a uniform integrable bound and then the claim follows from Lebesgue’s dominated convergence theorem. Note that by assumption we have uniform boundsμ¯i(M)∈(ϵ,p/(p−1)μi), a.s., which proves the claim forp∈(1,2]. Thus, there remains the Poisson casep=1. In the Poisson case we haveh1(m)=μilog(m)−m. The leading term of this function is linear form→∞, henceforth, the uniform integrable bound assumption on the sequence Equation (8) provides the proof. □
The previous statement is based on the law of large numbers. Of course, we can also study a central limit theorem (CLT) that provides asymptotic normality and the rate of convergence. For the aggregated predictors we have convergence in distribution
M1/2μ¯i(M)−μiVar(μ^i)1/2⟹N(0,1),asM→∞,
noting that, under the Poisson casep=1we need to assume, in addition to the assumptions of Proposition 1, that the second moment ofμ^iis finite. Consider the expected generalization loss function on policy i for given mean estimatem>0being independent ofYi
m>0↦δ¯i(m)=Eδ(Yi,m)|m=2viφEYi κp′−1(Yi)−Eκpκp′−1(Yi)−hp(m),
where functionhpwas defined in Equation (9). Thus, for a CLT of the expected generalization loss it suffices to understand the asymptotic behavior ofhp(μ¯i(M)), because we have, using the tower property for conditional expectation and using independence betweenμ¯i(M)andYi,
EδYi,μ¯i(M)=EEδYi,μ¯i(M)μ¯i(M)=Eδ¯iμ¯i(M).
Proposition 4.
Assume thatμ^i(j),j≥1, are i.i.d. copies ofμ^isatisfying the assumptions of Proposition 1, and all being independent fromYi. In the Poisson casep=1we additionally assume thatμ^ihas finite second moment. We have
M1/2hp(μ¯i(M))−hp(μi)hp′(μi)Var(μ^i)1/2⟹N(0,1),asM→∞.
This shows how the speed of convergence of aggregated predictors translates to the speed of convergence of deviance loss functions, in particular, from Taylor’s expansion we receive a first order termhp′(μi) . Basically, this is Theorem 5.2 in Lehmann (1983); we provide a proof because it is instructive.
Proof.
We have Taylor expansion (using the Lagrange form for the remainder)
hp(μ¯i(M))=hp(μi)+hp′(μi)μ¯i(M)−μi+12!hp″(m)μ¯i(M)−μi2,
for some m betweenμ¯i(M)andμi. This provides
M1/2hp(μ¯i(M))−hp(μi)hp′(μi)Var(μ^i)1/2=M1/2μ¯i(M)−μiVar(μ^i)1/2+12!hp′(μi)Var(μ^i)1/2M1/2hp″(m)μ¯i(M)−μi2.
The first term on the right-hand side converges in distribution to the standard Gaussian distribution asM→∞, because of the CLT Equation (11). Therefore, the claim follows by proving that the last term converges in probability to zero. Consider the eventAM={|μ¯i(M)−μi|>M−3/8}={M1/2|μ¯i(M)−μi|>M1/8}. CLT Equation (11) implieslimM→∞P[AM]=0. Chooseε>0. For the last term on the right-hand side of the last identity we have bound
limM→∞PM1/2max−|μ¯i(M)−μi|≤ξ≤|μ¯i(M)−μi||hp″(μi+ξ)|μ¯i(M)−μi2>ε≤limM→∞PM1/2max−|μ¯i(M)−μi|≤ξ≤|μ¯i(M)−μi||hp″(μi+ξ)|μ¯i(M)−μi2>ε,AMc+limM→∞P[AM]=0.
This finishes the proof. □
We remark that Propositions 1–4 have been formulated for predictorsμ^i, and a crucial assumption in these propositions is that these predictors are unbiased for mean parameterμi. We could state similar aggregation results on the canonical parameter scale forθ^iandθi. This would have the advantage that we do not need any side constraints onθ^ibecause the cumulant functionκpis convex over the entire effective domainΘp. The drawback of this latter approach is that typicallyθ^iis not unbiased forθi, in particular, if we choose the canonical linkg=(κp′)−1in a non-Gaussian situation. In general, we can only expect asymptotic unbiasedness in this latter situation.
4. Networks and Aggregating
The results on aggregated predictors in Propositions 1–4 are essentially based on the assumption that we can generate a suitable i.i.d. sequence of unbiased predictorsμ^i(j),j≥1 . Of course, in any reasonable practical application this is not the case because the data generating mechanism is unknown. Therefore, in practice, we can only make statements that are based on empirical approximations to the true model. Breiman (1996) used bootstrapping to obtain multiple predictorsμ^i(j),j≥1 , and inserting these bootstrap estimates into Equation (10), he called the resulting predictor the bagging predictor. Dietterich (2000b) discusses other methods of receiving multiple predictors. We do not use bootstrapping here, but rather a method highlighted in Section 2.5 of Dietterich (2000b), and we will discuss similarities and differences between our approach and Breiman’s bagging predictor in this section.
4.1. Empirical Generalization Losses
We start from two disjoint sets of observationsD={(Yi,xi,vi);i=1,…,n}andT={(Yt†,xt†,vt†);t=1,…,m}. Assume that these two sets of observations have been generated independently and by exactly the same data generating mechanism providing independent observationsYiinDandYt†inT.Dwill be termed learning data set, and it will be used for model calibration by making objective functionβ↦L(D;β)small inβ; we typically use deviance loss function Equation (6) as the objective function. The resulting estimatorβ^=β^Dofβis then used to receive estimated meansμ^t†=μ(xt†,β^)=μ(xt†,β^D)for the featuresx1†,…,xm†from the test data setT. This allows us to study an empirical version of the expected generalization loss (8) onTgiven by considering
L(T;β^)=1m∑t=1mδ(Yt†,μ^t†)=1m∑t=1mδYt†,μ(xt†,β^)=2m∑t=1mvt†φYt† κp′−1(Yt†)−κpκp′−1(Yt†)−Yt† κp′−1μ^t†+κpκp′−1μ^t†.
Three remarks:
(1)
Parameter estimateβ^=β^Dis solely based on the learning data setDand lossL(T;β^)is purely evaluated on the observationsYt†of the test data setT. Independence between the two data sets implies independence between parameter estimator and observationsYt†inT. Thus, this provides a proper out-of-sample generalization analysis.
(2)
Expected generalization loss Equation (8) is based on the assumption that we can calculate the necessary moments ofYt†under its true distributional model. Since this is not possible, in practice, we replace moments by empirical moments onTvia Equation (12), however, the individual observations inTare not necessarily i.i.d., they are only independent and have the same structure for the mean parameterμt†=μ(xt†,β), but they typically have different variances, see Equation (4).
(3)
Evaluation Equation (12) is based on one single estimatorβ^, only. Bootstrapping may provide an empirical mean also in parameter estimatesβ^.
4.2. The Nagging Predictor
4.2.1. Definition of Nagging Predictors
In order to calculate aggregated predictors we need to have multiple parameter estimatesβ^(j),j≥1, for network parameterβ∈Rr . Assume that we work in the network regression model framework as introduced in Section 2. We choose a sufficiently complex network so that it has sufficient approximation capacity to model fairly well a suitable family of regression functions. The selection of this network architecture involves the choices of the depth d of the network, the numbers of hidden neuronsqmin each hidden layer1≤m≤d, as well as activation functionϕ and link function g, see Section 2.1. This network regression function is then calibrated to the learning data setD assumed to satisfy Tweedie’s CP model assumptions as specified in Section 2.2. The network parameterβ∈Rrof dimensionr=∑m=1d+1(qm−1+1)qmis now calibrated to this learning data by making deviance loss Equation (6) small. Since typically the dimension r is large, we do not aim at finding the MLE forβby minimizing deviance lossβ↦L(D;β), because this would over-fit the predictive model to the learning dataD, and it would poorly generalize to the (independent) test dataT. For this reason we early stop the fitting algorithm before it reaches a (local) minimum of the deviance loss functionβ↦L(D;β). This early stopping is done according to a sensible stopping rule that tracks over-fitting on validation dataVwhich typically is a subset of the learning dataL, and, in particular, disjoint from the test data setT.
The crucial point now is that the smoothness of this minimization problem in combination with an early stopping rule will provide for each different starting point of the fitting algorithm a different parameter estimateβ^(j),j≥1. Thus, this puts us in the situation of having an (infinite) sequence of parameter estimates that obey the same stopping rule and have a comparable deviance loss, i.e.,L(D;β^(j))≈L(D;β^(j′)), of course, supposed that the stopping rule is based on this objective function. This then allows us to define and explore the resulting nagging predictor, defined by
μ¯¯t(M)=1M∑j=1Mμ^t(j)=1M∑j=1Mμ(xt†,β^(j)),
where the estimatorsμ^t(j)are calculated according to Equation (5) using featurext†∈Xand network parameter estimatesβ^(j)of different runs (seeds)j≥1of the gradient descent algorithm.
In the remainder of this manuscript, we describe differences and commonalities of nagging predictorsμ¯¯t(M)with bootstrapping and bagging predictors. We describe properties of these nagging predictorsμ¯¯t(M) , and we empirically analyze rates of convergence both on the portfolio level and on the individual insurance policy level, this analysis being based on an explicit car insurance data set, which is shown in Section 5.1, below.
4.2.2. Interpretation and Comparison to Bagging Predictors
Dependence on ObservationsD
The nagging and the bagging predictors both have in common that they are fully based on the observed dataD. Thus, similar to bootstrapping, we can only extract information that is already contained in the dataD, and all deduced results and properties have to be understood conditionally, givenD. In particular, if for some reason dataDis atypical, then this is reflected in bagging and nagging predictors and, therefore, these predictors may not generalize well to more typical data.
Differences between Bagging and Nagging
The crucial difference between bagging and nagging predictors is that the former performs re-sampling on observations, thus, it tries to create new (bootstrap) observations from the dataDthat follow a similar law as this original data. This re-sampling is either done in a non-parametric way, using an appropriate definition of residuals, or in a parametric way, with parameter estimates based onD. Both re-sampling versions involve randomness and, therefore, bootstrapping is able to generate multiple random predictorsμ^i(j). Typically, these bootstrap predictors are i.i.d. by applying the same algorithm using i.i.d. seeds, but (again) this i.i.d. property has to be understood conditionally on the given observationsD, as mentioned in the previous paragraph.
The nagging predictor is not based on re-sampling data, but it always works on exactly the same data set, and multiple predictors are obtained by exploring multiple models, or rather multiple parametrizations of the same model using gradient descent methods combined with early stopping. Naturally, this involves less randomness compared to bootstrapping because the underlying data set for the different predictors is always the same.
Remark 2.Bootstrapping is often used to assess model and parameter estimation uncertainty. For such an analysis, the bootstrapped observation is thought of being a second independent observation from the same underlying model, and one aims at understanding how parameter estimates change under such a change in the observations. As mentioned above, bootstrapping (re-sampling) adds an element of randomness, for instance, a large observation may occur on a different insurance policy that has slightly different features compared to the original data that generated the large observation and, henceforth, parameter estimates slightly change. Our nagging predictor does not have this element of randomness because it always works on exactly the same observations. One could combine the nagging predictor with bootstrapping on the estimated model. However, we refrain from doing so here, because we aim at fully understanding different (early stopped) gradient descent solutions, and additional bootstrap noise is not helpful for getting this understanding. Nevertheless, we view combining bootstrapping with nagging as a next step to explore model uncertainty. This is beyond the present manuscript, also because convergence results will require more work.
Unbiasedness
The results of Propositions 1–4 are crucially based on the assumption that the predictorsμ^i(j)are unbiased forμi. In practice, this might be a critical point if there is a bias in the dataD: as explained in the previous paragraphs, all deduced results are conditional on this dataD , and therefore a potential bias will also be reflected in bagging and nagging predictors. In this sense, the empirical analysis is a bit different from the theoretical results in Section 3; the propositions in that section build on the assumption of being able to simulate infinitely many observations and, henceforth, by the law of large numbers a potential bias asymptotically vanishes. Typically, on finite samples there is a bias and the best that bagging and nagging can do is to exactly match this bias. Thus, these methods do not help to reduce bias, but they reduce volatility in predictors, as highlighted in Propositions 1–4 but always in an empirical sense, conditionally onD.
5. Example: French Motor Third-Party Liability Insurance 5.1. Data
Our goal is to explore bagging predictors on real data. We therefore use the French motor third-party liability (MTPL) claim counts data set of Dutang and Charpentier (2019). This data set has already been used in Noll et al. (2018) and Wüthrich (2019). The data are available through the R package CASdatasets, see Dutang and Charpentier (2019); we use version CASdatasets_1.0–8.
Listing 1 gives a short excerpt of the data. We have a unique policy ID (IDpol) for each policy i, the claim countsNi per policy (ClaimNb), and the exposurevi (Exposure) on lines 3–5 of Listing 1, and lines 6–14 provide the covariate informationxi . An extensive descriptive and exploratory data analysis of the French MTPL data are provided in Noll et al. (2018) and Lorentzen and Mayer (2020), therefore, we do not repeat such an analysis here, but refer to these references.
Risks 08 00083 i001 550
5.2. Regression Design for Predictive Modeling
Our goal is to build a predictive model for the claim countsNi (ClaimNb) in Listing 1. We choose a Poisson regression model for data(Ni,xi,vi)satisfying
Ni∼Poiμ(xi)vi,withexpectedfrequencyfunctionxi↦μ(xi).
The Poisson distribution belongs to the family of Tweedie’s CP models with power variance parameterp=1, effective domainΘ=R, cumulant functionκ1(θ)=exp(θ)and dispersion parameterϕ=1. In fact, if we defineYi=Ni/vi, we have the following density w.r.t. the counting measure onN0/vi
Yi∼f(y;θi,vi,p)=expviyθi−exp(θi)+a1(y;vi).
This provides the first two moments as
μi=E[Yi]=exp(θi)andVar(Yi)=1viexp(θi)=1viμi.
The canonical link is given by the log-link, and we choose links(κp′)−1(·)=g(·)=log(·). These choices provide the network regression function on the canonical scale, see Equation (5),
xi↦θi=θ(xi)=κ1′−1μ(xi)=β(d+1),z(d)∘⋯∘z(1)(xi),
thus, the network predictor on the right-hand side exactly gives the canonical parameter under the canonical link choice forg(·). We are now almost ready to fit the model to data, i.e., to find a good network parameterβ∈Rrusing the gradient descent algorithm. As objective function for parameter estimation we choose the Poisson deviance loss function which is given by inserting all Poisson distribution related terms into Equation (6),
L(D;β)=2n∑i=1nviYilog(Yi)−Yi−Yi θi+exp(θi)=2n∑i=1nμ(xi)vi−Ni−Nilogμ(xi)vi/Ni,
for regression functionxi↦μi=μ(xi)defined through Equation (14), and where the terms under the summations are set equal to2μ(xi)viin caseNi=0.
5.3. Selection of Learning and Test Data
Next we describe data selection and feature pre-processing so that they can be used in a network regression function of the form Equation (14). Features are pre-processed completely analogously to the example in Section 3.3.2 of Wüthrich (2019), i.e., we use the MinMaxScaler for continuous explanatory variables and we use two-dimensional embedding layers for categorical covariates.
Having this pre-processed data, we specify the choice of learning dataDon which the model is learned, and the test dataT on which we perform the out-of-sample analysis. To keep comparability with the results in Noll et al. (2018); Wüthrich (2019) we use exactly the same partition of all data of Listing 1. Namely, 90% of all policies in Listing 1 are allocated to learning dataDand the remaining 10% are allocated to test dataT . This allocation is done at random, and we use the same seed as in Noll et al. (2018); Wüthrich (2019). This provides us with the two data sets highlighted in Table 1.
5.4. Network Architecture and Gradient Descent Fitting
Finally, we need to specify the network architecture. We choose a network of depthd=3having(q1,q2,q3)=(20,15,10) hidden neurons in the three hidden layers. We have 7 continuous features components (we model Area as a continuous component, see Wüthrich (2019) and two categorical ones having 11 and 22 labels, respectively. Using embedding dimensions 2 for the two categorical variables provides us with a network architecture having a network parameter of dimensionr=792 ; this includes the 66 embedding weights of the two categorical feature components. As activation function we choose the hyperbolic tangent. We implement this in R using the keras library, the corresponding code is given in Listing A1 in the Appendix A. Line 25 of Listing A1 highlights that we choose the Poisson deviance loss function as objective function, and we use the nadam version of gradient descent; for different versions of the gradient descent algorithm we refer to Goodfellow et al. (2016).
We are now ready to fit this network regression model to the learning data setD. Running this algorithm involves some more choices. First of all, we need to ensure that the network does not over-fit to the learning dataD. To ensure this we partition at random 9:1 the learning data into a training data setD(−)and a validation data setV. The network parameter is learned on the training dataD(−)and over-fitting is tracked on the validation dataV . We run the nadam gradient descent algorithm over 1000 epochs on random mini-batches of size 5000 from the training dataD(−) . Using a callback we retrieve the network parameter that has the smallest (validation) loss onV , this is exactly the stopping rule that we set in place, here. This is illustrated in Figure 1 giving training losses onD(−)in blue color and validation losses onVin green color. We observe that the validation loss increases after roughly 150 training epochs (green color), and we retrieve the network parameter with the lowest validation loss.
The fact that the resulting network model has been received by an early stopping of the gradient descent algorithm implies that this network has a bias w.r.t. the learning dataD . For this reason we additionally apply the bias regularization step proposed in Section 3.4 of Wüthrich (2019), see Listing 5 in Wüthrich (2019). This gives us an estimated network parameterβ^(1)and corresponding mean estimatesμ^i(1)for all insurance policies inDandT . This procedure leads to the results on line (d) of Table 2.
We compare the network results to the ones received in Table 11 of Noll et al. (2018), and we conclude that our network approach is competitive with these other methods (being a classical GLM and a boosting regression model), see the out-of-sample losses on lines (a)–(d).
Remark 3.
Our analysis is based on claim counts observations, therefore, the Poisson model is the canonical choice. As mentioned in Remark 1, depending on the underlying modeling problem other distributional choices from the EDF may be more appropriate. In any of these other model choices the proposed methodology works similarly, only minor changes to theRcode in Listing A1 will be necessary:
- Firstly, the complexity of the network architecture should be adapted to the problem, both the sizes of the hidden layers and the depth of the network will vary with the complexity of the problem and the complexity of the regression function. In general, the bigger the network the better the approximation capacity (this follows from the universality theorems). This says that the chosen network should have a certain complexity otherwise it will be not sufficiently flexible to approximate the true (but unknown) regression function. On the other hand, for computational reasons, the chosen network should not be too large.
-
Secondly, different distributions will require different choices of loss functions on line 25 of Listing A1. Some loss functions are already implemented in thekeraslibrary, others will require custom loss functions. An example of a custom loss function implementation is given in Listing 2 of Delong et al. (2020). In general, the loss function of any density that allows for an EDF representation can be implemented inkeras.
-
Thirdly, we could add more specialized layers to the network architecture of Listing A1. For instance, we could add dropout layers after each hidden layer on lines 15–17 of Listing A1, for dropout layers see Srivastava et al. (2014). Dropout layers add an additional element of randomness during training, because certain network connections are switched off (at random) for certain training steps when using dropout. This switching off of connections acts as regularization during training because it prevents certain neurons from over-fitting to special tasks. In fact, under certain assumptions one can prove that dropout acts similarly to ridge regularization, see Section 18.6 in Efron and Hastie (2016). In our study we refrain from using dropout.
5.5. Comparison of Different Network Calibrations
The issue with the network result on line (d) of Table 2 now is that it involves quite some randomness:
(R1)
we randomly split learning dataDinto training dataD(−)and validation dataV;
(R2)
we randomly split training dataD(−)into mini-batches of size 5000 (to more efficiently calculate gradient descent steps); and
(R3)
we randomly choose the starting point of the gradient descent algorithm; the default initialization in keras is the glorot_uniform initialization which involves simulation from uniform distributions.
Changing seeds in these points (R1)–(R3) will change the network regression parameter estimateβ^(j)and, hence, the results. For this study we explore the randomness of changing the seeds in (R1)–(R3). However, one may also be interested in sensitivities of results when changing sizes of the mini-batches, hyper-parameters of the gradient descent methods, or when using stratified versions of validation data. We will not explore this here.
We run the above calibration procedure under identical choices of all hyper-parameters, but we choose different seeds for the random choices (R1)–(R3). The boxplot in Figure 2 shows the in-samples lossesL(D;β^(j))and out-of-samples lossesL(T;β^(j))over 400 network calibrationsβ^(j)by only randomly changing the seeds in the above mentioned points (R1)–(R3). We note that these losses have a rather large range which indicates that results of single network calibrations are not very robust. We can calculate empirical mean and standard deviation for the 400 seeds j given by
L¯(T;β^(1:400))=1400∑j=1400L(T;β^(j))and1399∑j=1400L¯(T;β^(1:400))−L(T;β^(j))2.
The first gives an estimate for the expected generalization loss Equation (8) averaged over the corresponding portfolios. We emphasize in notationβ^(1:400)that we do not average over network parameters, but over deviance losses on individual network parametersβ^(j) . The resulting numbers are given on line (e) of Table 2. This shows that the early stopped network calibrations have quite significant differences, which motivates the study of the nagging predictor.
Figure 3 gives a scatter plot of in-sample and out-of-sample losses over the 400 different runs of the gradient descent fitting (complemented with a natural cubic spline). We note in this example that both small and big in-sample losses do not lead to favorable generalization, small in-sample losses may indicate over-fitting and big out-of-sample losses may indicate the that early stopping rule has not found a good model based on the validation dataV.
5.6. Nagging Predictor
We are now in the situation where we are able to calculate the nagging predictorsμ¯t(M)over the test data setT. ForM→∞this provides us with empirical counterparts of Propositions 3 and 4. We therefore consider forM≥1the sequence of out-of-sample losses, see Equation (12),
L(T;μ¯¯t=1,…,m(M))=1m∑t=1mδ(Yt†,μ¯¯t(M)),
whereμ¯t(M)are the nagging predictors Equation (13) received from the i.i.d. sequenceμ^t(j)=μ(xt†,β^(j))and, again, where the i.i.d. property is conditional onDand refers to the i.i.d. starting points chosen to start the gradient descent fitting algorithm.
Figure 4 gives the out-of-sample losses of the nagging predictorsL(T;μ¯¯t=1,…,m(M))forM=1,…,100. Most noticeable is that nagging leads to a substantial improvement in out-of-sample losses, forM→∞ the out-of-sample loss converges to 31.272 which is much smaller then the figures reported on lines (b)–(e) of Table 2, in fact, the out-of-sample loss decreases by more than 3 standard deviations (as reported on line (e) of Table 2). From this we conclude that nagging helps to improve the predictive model substantially. From Figure 4 we also observe that this convergence mainly takes place over the first 20 aggregating steps in our example. That is, we need to aggregate over roughly 20 network calibrationsβ^(j) to get the maximal predictive power. The dotted orange lines in Figure 4 give corresponding 1 standard deviation confidence bounds (received by repeating the nagging procedure for different seeds). To get sufficiently small confidence bounds in our example we need to average over roughly 40 network calibrations.
Conclusion 1.Nagging helps to sufficiently improve out-of-sample performance of network regression models. In our example we need to average over 20 different calibrations to get optimal predictive power, and we need an average of 40 calibrations to ensure that we cannot further improve this predictive power.
5.7. Pricing of Individual Insurance Policies
In the previous sections we have proved that nagging successfully improves models. However, all these considerations have mainly been based on portfolio considerations, i.e., our focus has been on the question whether the insurance company has a model that gives good predictive power to predict the portfolio claim amount. For insurance pricing, this is not sufficient because we should also ensure that we have robustness of prices on an individual insurance policy level. In this section we analyze by how much individual insurance policy prices may differ if we select two different network calibrationsβ^(j)andβ^(j′). This will tell us whether aggregating over 20 or 40 network calibrations is sufficient as stated in Conclusion 1. Naturally, we expect that we need to average over more networks because the former statement includes an average overT, i.e., overm=67,801insurance policies (though there is dependence between these policies because they simultaneously use the same network parameter estimateβ^(j)).
To analyze this question on individual insurance policies we calculate for each policyt=1,…,mof the test dataTthe nagging predictorμ¯t(M)overM=400different network calibrationsβ^(j),j=1,…,M, and we calculate the empirical coefficients of variation in the individual network predictors given by
CoV^t=σ^t μ¯¯t(M)=1M−1∑j=1M μ^t(j)−μ¯¯t(M)2μ¯¯t(M),
these are the empirical standard deviations normalized by the corresponding mean estimates. We plot these coefficients of variationsCoV^t in Figure 5 (lhs) against the nagging predictorsμ¯t(M)for each single insurance policyt=1,…,m(out-of-sample onT ). Figure 5 (rhs) shows the resulting histogram. We observe that on most insurance policies (73%) we have a coefficient of variation of less than 0.2, however, on 11 of them=67,801insurance policies we have a coefficient of variation bigger then 1. Thus, for the latter, if we average over 400 different network calibrationsβ^(j)we still have an uncertainty of1/400=0.05, i.e., the prices have a precision of 5% to 10% in these latter cases (this is always conditional givenD). From this we conclude that on individual insurance policies we need to aggregate over a considerable number of networks to receive stable network regression prices.
We show these 11 insurance policies with a coefficient of variation bigger 1 in Listing 2. Striking is that all these insurance policies have vehicle ageVehAge=0 . In Figure 6 (top row) we show a scatter plot and histogram of these insurance policies, and Figure 6 (bottom row) gives the corresponding graphs forVehAge>0. We indeed confirm that mainly the policies withVehAge=0 are difficult to price. From the histogram in Figure 5 (rhs) we receive in total 1423 policies that have a coefficient of variationCoV^t bigger than 0.4, and in view of Figure 6 (rhs), 1163 of these insurance policies haveVehAge=0.
Risks 08 00083 i002 550
To better understand this uncertainty inVehAge=0 , we provide some empirical plots for those vehicles with age 0; these plots are taken from Noll et al. (2018). Figure 7 shows the total exposure per vehicle age (top-lhs), the empirical frequency per vehicle age (top-rhs), the vehicle ages per vehicle brands (bottom-lhs) and total exposures per vehicle brand (bottom-rhs). From these graphs we observe that vehicle age 0 has a completely different frequency than all other vehicle ages (Figure 7, top-rhs). Moreover, vehicle age 0 is dominated by vehicle brand B12 (Figure 7, bottom). It seems in view of Listing 2 that this configuration results in quite some uncertainty in frequency prediction, note that 1011 of the 1163 insurance policies with vehicle age 0 and a coefficient of variation bigger than 0.4 are cars of vehicle brand B12. Naturally, in a next step we would have to analyze vehicle age 0 and vehicle brand B12 in more depth, we suspect that these could likely be rental cars (or some other special cases). Unfortunately, there is no further information available for this data set that allows us to investigate this problem more thoroughly. We remark that König and Loser (2020) come to a similar conclusion for cars of vehicle brand B12 with vehicle age 0, see regression tree in Chapter 1.2 of König and Loser (2020) where furthermore the split w.r.t. vehicle gas is important.
5.8. Meta Network Regression Model
From the previous sections we conclude that the nagging predictor substantially improves the predictive model. Nevertheless, the nagging predictor may not be fully satisfactory in practice. The difficulty is that it involves aggregating overM=400predictorsμ^i(j)for each policy i. Such blended predictors (and models) are not easy to maintain nor is it simple to study model properties, updating models with new observations, etc., for instance, if we have a new insurance policy having covariatex, then we need to run this new insurance policy through allM=400networks to receive the price. For this reason we propose to build a meta model that fits a new network to the nagging predictorsμ¯i(M),i=1,…,n . In the wider machine learning literature, building such meta models is also referred to as “model distillation”, see Hinton et al. (2015). Since these nagging predictors are aggregated predictors over M network models, and since the network regression functions themselves are smooth functions in input variables (at least in the continuous features), the nagging predictors describe smooth surfaces. Therefore, it is comparably simple to fit a network to the smooth surface described by nagging predictorsμ¯i(M),i=1,…,n , and over-fitting will not be an issue. To build this meta model we use exactly the same network architecture as above, and as outlined in Listing A1 in the Appendix A. The only parts that we change are the loss function and the response variables. We replace the original claim count responsesYiby the nagging predictorsμ¯i(M), and for the loss function we choose the square loss function. We can either choose an unweighted square loss function or we can weight the individual observations with the inverse standard deviation estimates1/σ^i , see Equation (15), the latter reflects the idea of giving more weight to predictors that have less uncertainty. In fact, the latter is a little bit similar to Zhou et al. (2002) where the authors argue that equally weighting over network predictors is non-optimal. Our approach is slightly different here because we do not weight networks, but we weight nagging predictors of individual insurance policies to receive better rates of convergence for meta model training.
We present the gradient descent fitting performance in Figure 8 and the resulting in-sample and out-of-sample losses in Table 3. We conclude that the weighted version of the square loss has better convergence properties in gradient descent fitting, and the resulting model has a better loss performance, compare lines (g1) and (g2) of Table 3. The resulting meta model on line (g2) is slightly worse than the nagging predictor model on line (f), however, substantially better than the individual network models (d)–(e) and much more easy in handling than the nagging predictor. For this reason, we are quite satisfied by the meta model, and we propose to hold on to this model for further analysis and insurance pricing.
Remark 4.We have mentioned in Remark 3 that model implementation will slightly change if we move from our Poisson example to another distributional example, for instance, the loss function has to be adapted to the choice of the distribution function. The last step of fitting a meta model, however, does not require any changes. Note that the meta model fits a regression function to another regression function (which is the nagging predictor in our case). This fitting uses the square loss as the objective function, and the nature of the distribution function of the observations is not important because we fit a deterministic function to another deterministic function here.
In Figure 9 we present the two predictors in a scatter plot, we observe that they are reasonably equal, the biggest differences are highlighted in blue color, and they refer to policies with vehicle age 0, that is, the feature component within the data that is the most difficult to fit with the network model.
We provide a concluding analysis of this meta model. As emphasized in Section 7 of Richman et al. (2019), we analyze and try to understand the representations learned in the last hidden layer of the network over all insurance policies. In view of Equation (1) this means that we study the learned representations fort=1,…,mgiven by
xt†↦x˘t†=z(d)∘⋯∘z(1)(xt†)∈Rqd+1,
to which the GLM step with readout parameterβ(d+1)∈Rqd+1is applied to. In our example of Listing A1 we haved=3hidden layers withq3=10hidden neurons in this last hidden layer. In order to illustrate the learned representations we apply a principal component analysis (PCA) to these learned representations(x˘t)t=1,…,m. This provides us with the principal component (PC) scores for each policy t and the corresponding singular valuesλq3 ≥…≥λ1≥0. Note that these PC scores are ordered w.r.t. importance measured by the singular values. We scale the individual PC scoresk=1,…,q3withλk to reflect this importance. In Figure 10 (top row) we plot these scaled PC scores for the first 4 principal components and averaged per vehicle age (lhs), driver age (middle), and bonus malus level (rhs). We remark that these 3 feature components are the most predictive ones according to the descriptive analysis in Noll et al. (2018). From Figure 10 (top row) we observe that the first principal component (in red color) reflects the marginal empirical frequencies illustrated in Figure 10 (bottom row), thus, the first PC reflects the influence of each of these three explanatory variables on the regression function, and higher order PCs are used to refine this picture and model interactions between feature components.
6. Conclusions and Outlook This work has defined the nagging predictor which produces accurate and stable portfolio predictions on the basis of random network calibrations, and it provided convergence results in the context of Tweedie’s compound Poisson generalized linear models. Focusing on an example in motor third-party liability insurance pricing, we have shown that stable portfolio results are achieved after 20 network training runs, and by increasing the number of network training runs to 400, we have shown that quite stable results are produced at the level of individual policies, which is an important requirement for the use of networks for insurance pricing, and more general actuarial tasks. The coefficient of variation of the nagging predictor is shown to be a useful data-driven metric for measuring the relative difficulty with which a network is able to fit to individual training examples, and we have used it to calibrate an accurate meta network which approximates the nagging predictor. Whereas this work has examined the stability of network predictions in the case of a portfolio at a point in time, another important aspect of consistency within insurance is stable pricing over time, thus, future work could consider methods for stabilizing network predictions as new information becomes available.
Numbers of Observed Claims | Empirical | Size of | |||||
---|---|---|---|---|---|---|---|
0 | 1 | 2 | 3 | 4 | Frequency | Data Sets | |
empirical probability onD | 94.99% | 4.74% | 0.36% | 0.01% | 0.002% | 10.02% | n=610,212 |
empirical probability onT | 94.83% | 4.85% | 0.31% | 0.01% | 0.003% | 10.41% | m=67,801 |
In-Sample | Out-of-Sample | |||
---|---|---|---|---|
Loss onD | Loss onT | |||
(a) homogeneous model | 32.935 | 33.861 | ||
(b) generalized linear model | 31.267 | 32.171 | ||
(c) boosting regression model | 30.132 | 31.468 | ||
(d) network regression model (seedj=1) | 30.184 | 31.464 | ||
(e) average over 400 network calibrations | 30.230 | (0.089) | 31.480 | (0.061) |
(f) nagging predictor forM=400 | 30.060 | 31.272 |
In-Sample | Out-of-Sample | |||
---|---|---|---|---|
Loss onD | Loss onT | |||
(d) network regression model (seedj=1) | 30.184 | 31.464 | ||
(e) average over 400 network calibrations | 30.230 | (0.089) | 31.480 | (0.061) |
(f) nagging predictor forM=400 | 30.060 | 31.272 | ||
(g1) meta network model (un-weighted) | 30.260 | 31.342 | ||
(g2) meta network model (weighted) | 30.257 | 31.332 |
Author Contributions
Both authors (R.R., M.V.W.) have equally contributed to the this project, this concerns the concept, the methodology, the writing, and the numerical analysis of this project. All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Conflicts of Interest
The authors declare no conflict of interest.
Appendix A. R Code
Listing A1 provides the R code used to fit the data.
Risks 08 00083 i0a1 550
Listing A1.R code for a network of depth d = 3.
Listing A1.R code for a network of depth d = 3.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2020. This work is licensed under http://creativecommons.org/licenses/by/3.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
We define the nagging predictor, which, instead of using bootstrapping to produce a series of i.i.d. predictors, exploits the randomness of neural network calibrations to provide a more stable and accurate predictor than is available from a single neural network run. Convergence results for the family of Tweedie’s compound Poisson models, which are usually used for general insurance pricing, are provided. In the context of a French motor third-party liability insurance example, the nagging predictor achieves stability at portfolio level after about 20 runs. At an insurance policy level, we show that for some policies up to 400 neural network runs are required to achieve stability. Since working with 400 neural networks is impractical, we calibrate two meta models to the nagging predictor, one unweighted, and one using the coefficient of variation of the nagging predictor as a weight, finding that these latter meta networks can approximate the nagging predictor well, only with a small loss of accuracy.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer