Loss Reserving: Past, Present and Future
|
|
|
- Clifford Jacobs
- 10 years ago
- Views:
Transcription
1 Loss reserving: past, present and future 1 G. Taylor, G. McGuire and A. Greenfield ISBN Loss Reserving: Past, Present and Future by Greg Taylor Taylor Fry Consulting Actuaries and The University of Melbourne and Gráinne McGuire and Alan Greenfield Taylor Fry Consulting Actuaries RESEARCH PAPER NUMBER 19 SEPTEMBER 23 Centre for Actuarial Studies Department of Economics The University of Melbourne Victoria 31 Australia
2 Loss reserving: past, present and future 2 Loss reserving: past, present and future Greg Taylor Taylor Fry Consulting Actuaries & University of Melbourne Gráinne McGuire Taylor Fry Consulting Actuaries Alan Greenfield Taylor Fry Consulting Actuaries An invited lecture to the XXXIVth ASTIN Colloquium, Berlin, August 23
3 Loss reserving: past, present and future i Table of Contents Abstract Taxonomy of loss reserving models Adaptive reserving From the present to the future Conclusion References...29 c:\documents and settings\challis boland\eudora mail\attach\lecture1583.doc
4 Loss reserving: past, present and future 1 Abstract The paper reviews the development of loss reserving models over the past, classifying them according to an elementary taxonomy. The taxonomic components include (1) the algebraic structure of the model, (2) the form of its parameter estimation, (3) whether or not it is explicitly stochastic, and (4) whether or not its parameters evolve over time. Particular attention is given to one of the higher species of model, involving complex structure, optimal estimation, and evolutionary parameters. A generalisation of the Kalman filter is considered as a basis of adaptive loss reserving in this case. Real life numerical examples are provided. Some implications of this type of data analysis for loss reserving are discussed, with particular reference to the form of data set used. The use of triangular arrays is questioned, and alternatives examined. Again, real life numerical examples are provided. Keywords: loss reserving, dynamic models, Kalman filter, exponential dispersion family. c:\documents and settings\challis boland\eudora mail\attach\lecture1583.doc
5 Loss reserving: past, present and future 2 1. Taxonomy of loss reserving models 1.1 Taxonomic considerations The taxonomic classification of loss reserving models was considered by Taylor (1986). The main classification variables recognised there were: Whether or not the model was stochastic; Model structure (whether a macro- or micro-model ); The model s dependent variables: o Whether paid or incurred losses; o Whether or not claim counts were modelled; Whether or not specific explanatory variables were included. When Taylor came to research a new book on loss reserving (Taylor, 2), he found that the literature published in the 1½ decades since the previous book was roughly equal in volume to all the prior literature. A number of new techniques had been introduced. In short, the situation had changed sufficiently to require a re-consideration of taxonomic questions. These questions were not pursued in the book, but are examined here. Four dimensions are identified for a re-classification of loss reserving models, namely: Stochasticity (as before); Dynamism; Model structure; Parameter estimation. The space of models may be thought of as a 4-dimensional hyper-cube in these dimensions. Their meaning is now discussed briefly. For the sake of the discussion, it is assumed that the data being modelled follows the conventional format of a triangle of values of some observed claim statistic C(i,j) illustrated in Figure 1.1, where i,j denote accident period and development period respectively. c:\documents and settings\challis boland\eudora mail\attach\lecture1583.doc
6 Loss reserving: past, present and future 3 Figure 1.1 Claims triangle j i C(i,j) Stochasticity For a stochastic model, observations are assumed to have the form: C(i,j) = µ(i,j) + e(i,j) (1.1) where µ(i,j) is a parameter and e(i,j) an explicitly stochastic error term. Though it may seem surprising now, not all models have this form. The earlier models, of which the chain ladder was an example, were developed heuristically, and essentially deterministically. The earliest stochastic models in the literature appear to be those of Kramreiter and Straub (1973) and Hachemeister and Stanard (1975). Others followed, e.g Reid (1978), and by the early 198 s they were appearing rapidly (De Jong and Zehnwirth, 1983; Pollard, 1983; Taylor and Ashe, 1983). Dynamism A dynamic model is characterised by containing parameters that evolve over time. For example, E[C(i,j)] = µ(i,j) = f (β(i),j) (1.2) where β(i) is a parameter vector that evolves as follows: β(i) = β(i-1) + w(i) (1.3) with w(i) a stochastic perturbation. c:\documents and settings\challis boland\eudora mail\attach\lecture1583.doc
7 Loss reserving: past, present and future 4 Model structure Whereas stochasticity and dynamism are dichotomous classifications, model structure introduces a spectrum of possibilities, as illustrated in Figure 1.2. Figure 1.2 Spectrum of model structures Micro-structural models are those whose structure is not related to aspects of the claim process that have a direct physical meaning. The incurred loss chain ladder model would lie towards this end of the spectrum, with its log-linear structure in terms of somewhat abstract parameters, and its emphasis on age-toage factors. Micro-structural models, on the other hand, are concerned with the fine structure of the claims process. An example might be a statistical case estimation model (Taylor and Campbell, 22), i.e. a model of individual claim sizes according to claim-specific covariates such as age of claimant, nature of injury, etc). This spectrum cannot be regarded as any more than a broad concept. It is difficult to think of a case that lies absolutely at one end or the other. For example, even the chain ladder parameters have some form of physical meaning, such as proportion of an accident year s cost paid in a particular development year, but they are rather remote from the mechanics of the claim payment process itself. Parameter estimation There are two main possibilities as to the form of parameter estimation: Heuristic; Optimal. Under optimal estimation parameters are estimated by optimisation of some statistical criterion, e.g. likelihood. Heuristic estimation is typical of non-stochastic models; in the absence of a defined stochastic structure, formulation of a suitable objective function for optimisation may be difficult. There are, however, exceptions in which optimal estimation is applied to a non-stochastic structure, e.g. De Vylder (1978). Prior to Hachemeister and Stanard (1975), the chain ladder model was heuristic in its estimation. These authors placed it on an optimal basis in relation to Poisson claim counts, but it remained heuristic for claim amounts. England and Verrall (22) placed it on an optimal basis for claim amounts with overdispersed Poisson distributions. But it remains heuristic for many of its applications. c:\documents and settings\challis boland\eudora mail\attach\lecture1583.doc
8 Loss reserving: past, present and future Phylogeny The taxonomy of models defined by the four characteristics described above can be set out in a phylogenetic tree, as in Figure 1.3, where only the more realistic combinations of characteristics are represented. Figures 1.4 to 1.6 mark the main branches of the tree. Figure 1.3 Phylogenetic tree of loss reserving models Loss reserving models Static Dynamic Deterministic Stochastic Stochastic Heuristic Heuristic Optimal Optimal Microstructural Microstructural Figure 1.4 Figure 1.5 Deterministic branch Static stochastic branch Loss reserving models Loss reserving models Static Dynamic Static Dynamic Deterministic Stochastic Stochastic Deterministic Stochastic Stochastic Heuristic Heuristic Optimal Optimal Heuristic Heuristic Optimal Optimal Microstructural Microstructural Microstructural Microstructural c:\documents and settings\challis boland\eudora mail\attach\lecture1583.doc
9 Loss reserving: past, present and future 6 Figure 1.6 Dynamic stochastic branch Loss reserving models Static Dynamic Deterministic Stochastic Stochastic Heuristic Heuristic Optimal Optimal Microstructural Microstructural 1.3 The Darwinian representation The Darwinian view of evolution regarded it as a process of progressive improvement from the earliest species to homo sapiens. This was the concept of The Ascent of Man. Nowadays, value judgements about the superiority of one species over another are regarded as invalid. One need have no such reservations about the different species of loss reserving models, and Figure 1.3 may be organised into a diagrammatic representation of The Ascent of the Loss Reserving Model, as in Figure 1.7. The criterion for the ranking of species here is informal and intuitive. Broadly, however, it might equate to the volume of information encompassed by the model s output. c:\documents and settings\challis boland\eudora mail\attach\lecture1583.doc
10 Loss reserving: past, present and future 7 Figure 1.7 Darwinian representation of the Loss Reserving phylogeny 5 Dynamic Stochastic Optimal Dynamic Stochastic Micro-structural Optimal 6 3 Static Stochastic Optimal Static Stochastic Micro-structural Optimal 4 Static Stochastic Heuristic 2 Static Deterministic Heuristic 1 The earliest models, most of those up to the late 197 s, fell within Box 1. These included methods like: The chain ladder (as it was then viewed) The separation method (Taylor, 1977) The payment per claim finalised model (Fisher and Lange, 1973; Sawkins, 1979). Any deterministic model may be stochasticised by the addition of error terms. If these terms are left distribution-free, parameter estimation is likely to remain heuristic, as in Box 2. An example of such a model is Mack s (1993) version of the stochastic chain ladder. Alternatively, optimal parameter estimation may be applied to the case of distribution-free error terms, as in least squares chain ladder estimation (De Vylder, 1978), to produce a model in Box 3. c:\documents and settings\challis boland\eudora mail\attach\lecture1583.doc
11 Loss reserving: past, present and future 8 Optimal parameter estimation may also be employed if an error structure added. Examples are: Chain ladder for triangle of Poisson counts (Hachemeister & Stanard, 1975) Chain ladder with log normal age-to-age factors (Hertig, 1985) Chain ladder with triangle of over-dispersed Poisson cells (England & Verrall, 22). Box 4 is similar to Box 3 but with finer structure inserted into the model, as in, for example: The payments per claim finalised structure of Taylor and Ashe (1983); Reid s (1978) model of claim sizes in operational time. Evolution of parameters over time may be incorporated by means of the Kalman filter (Kalman, 196). This may be bolted onto many stochastic models, converting Boxes 3 and 4 to 5 and 6 respectively. Examples are: Model of distribution of an accident period s claim payments over development periods (De Jong and Zehnwirth, 1983); Chain ladder model (England and Verrall, 22). The blind application of the Kalman filter can, however, bring some shortcomings with it. This is a subject to be discussed further in Section 2. c:\documents and settings\challis boland\eudora mail\attach\lecture1583.doc
12 Loss reserving: past, present and future 9 2. Adaptive reserving 2.1 Kalman filter The term adaptive reserving will be used to refer to loss reserving based on dynamic models, as defined in Section 1. The Kalman filter has been noted as a means of converting a static model to a dynamic model. It was introduced into: the engineering literature by Kalman (196); the statistical literature by Harrison and Stevens (1976); the actuarial literature by De Jong and Zehnwirth (1983). The model underlying the Kalman filter consists of two equations, called the system equation and observation equation respectively. The former describes the model s parameter evolution, while the latter describes the model of observations conditional on the parameters. The two equations are as follows. System equation β j+1 = G j+1 β j + w j+1 (2.1) p 1 p p p 1 p 1 Observation equation Y j = X j β j + v j (2.2) p 1 p p p 1 p 1 These equations are written in vector and matrix form with dimensions written beneath, and Y j denotes the vector of observations made at time j (=1,2, ) β j denotes the vector of parameters at time j X j is the design matrix applying to time j G j+1 is a transition matrix governing the evolution of the expected parameter values from one epoch to the next w j+1 and v j are stochastic perturbations, each with zero mean, and with V[w j+1 ] = W j+1, V[v j ] = V j. (2.3) The Kalman filter is an algorithm for calculating estimates of the β j. It revises the parameter estimates iteratively over time. Each iteration introduces the additional data Y j. The formal statement of the filter requires a little extra notation. Let Y j k denote the filter s estimate of Y j on the basis of information up to and including epoch k; and similarly for other symbols with the subscript j k. Also, let Γ j k denote V[β j k ]. Figure 2.1 sets out a single iteration of the Kalman filter. c:\documents and settings\challis boland\eudora mail\attach\lecture1583.doc
13 Loss reserving: past, present and future 1 Figure 2.1 Single iteration of the Kalman filter Forecast new epoch s parameters and observations without new information β j+1 j =G j+1 β j j Γ j+1 j = G j+1 Γ j j G T j+1 + W j+1 Y j+1 j = X j+1 β j+1 j Update parameter estimates to incorporate new observation β j+1 j+1 = β j+1 j + K j+1 (Y j+1 -Y j+1 j ) Γ j+1 j+1 = (1 - K j+1 X j+1 ) Γ j+1 j Calculate gain matrix (credibility of new observation) L j+1 j = X j+1 Γ j+1 j X T j+1 + V j+1 K j+1 = Γ j+1 j X T j+1 [L j+1 j ]-1 To operate the filter, one commences at the top of the diagram with a prior estimate β and the associated dispersion Γ. The top box generates estimates β 1 and Γ 1 for the next epoch, still based only on the prior. Working through a complete iteration of the filter introduces the new data vector Y 1 and generates revised estimates β 1 1 and Γ 1 1. The procedure is repeated to generate β 2 2 and Γ 2 2, β 3 3 and Γ 3 3, etc. The key equation for revising parameter estimates appears under the bottom left box, and is as follows: βj+1 j+1 = βj+1 j + Kj+1 (Yj+1 - Yj+1 j). (2.4) The main characteristics of the formula of note are: It is linear in the data vector Y j+1. β j+1 j+1 is the Bayesian revision of the estimate of β j+1 in the case that w j+1 and v j+1 are normally distributed. Consequently, K j+1 is equivalent to the credibility of Y j Application of the Kalman filter to loss reserving Let Y j denote some vector of loss experience statistics. For example, suppose that Y j = (Y j1, Y j2, ) T where Y jm = log [paid losses in (j,m) cell] is and normally distributed. Suppose in addition that E[Y j ] = X j β j. (2.5)
14 Loss reserving: past, present and future 11 Here paid losses are log normal with a log-linear dependency of expectations on parameters. A specific example, as in De Jong and Zehnwith (1983) or Wright (199), involves Hoerl curves to represent the distribution of paid losses over development periods m, thus: Y jm = β j + β j1 log(m+1) + β j2 m, m=,1,2,etc. (2.6) where β j = (β j, β j1, β j2 ) T. The restricted nature of the Kalman filter can cause some difficulties in the formulation of loss reserving models. The linear nature of (2.4) means that the range of parameter estimates is doubly infinite. Hence, by (2.5), Y j j = X j β j j (2.7) also has a doubly infinite range. This is compatible with log normal Y j, but not with other distributions. If, for example, Y j denotes claim count data, negative estimates will be inappropriate. Distributional difficulties of this sort can sometimes be overcome by the use of transformed data, just as the log normal distribution relies on a log transformation. However, with the transformation comes the need for subsequent inversion, usually with an associated bias correction. The determination of a suitable bias correction, particularly when the distribution of Y j is only vaguely known, can be problematic. 2.3 Dynamic models with non-normal error terms Consider the following model as an alternative to that consisting of (2.1) and (2.2), underlying the Kalman filter: System equation: As in (2.1). Observation equation: Y j satisfies a generalised linear model (GLM) (see McCullagh and Nelder, 1989) with link function h and linear predictor X j β j, i.e. the distribution of Y j is chosen from the exponential dispersion family (EDF), and E[Y j ] = h -1 (X j β j ). (2.8) When observation equation (2.8) replaces (2.2), a question arises as to how the filter illustrated in Figure 2.1 should be modified. To answer this question, return to (2.4), and note that it is linear in both the prior estimate β j+1 j and the data Y j+1. This leads to the remark, made many times before, that (2.4) may be viewed as a regression of the vector [Y T j+1, β T j+1 j] T on β j+1.
15 Loss reserving: past, present and future 12 Indeed, (2.1) may be stacked with an equation setting out the dispersion of β j+1 j as follows: Y j+1 = X j+1 β j+1 + v j+1 β j+1 j 1 u j+1 (2.9) where V v j+1 = V j+1 u j+1 Γ j+1 j (2.1) and generally Γ j+1 j is not diagonal. Estimation of β j+1 by weighted least squares regression yields the estimate β j+1 j+1 given by (2.4). This regression is therefore equivalent to the Kalman filter estimation of β j+1. Moreover, this estimation will be maximum likelihood in the case in which v j+1 and w j+1 are both normal which, as remarked above, is also the case in which the Kalman filter produces Bayesian revision of the estimate of β j+1. It is now interesting to consider the case in which the system and observation equations of the present sub-section hold, with v j+1 and w j+1 from the EDF but not necessarily normal. Then (2.9) is replaced by: Y j+1 = h -1 X j+1 β j+1 + v j+1 β j+1 j 1 u j+1 (2.11) where the (n+p)-dimensional vector function h is the identity in its last p components, but not necessarily in its first n. Equation (2.1) continues to hold. By (2.9), the variate u j+1 is the centred version of β j+1 j, whose computation for the present case has not yet been specified. Therefore, it may or may not be from the EDF. Suppose, however, that it may be approximated by a member of the EDF. Then (2.9) holds with the stochastic error terms from the EDF. Apply the term EDF filter to an algorithm that parallels the Kalman filter but applies to a model conforming with (2.11). It is evident that the Kalman filter is a special case. The situation is illustrated by Figure 2.2. c:\documents and settings\challis boland\eudora mail\attach\lecture1583.doc
16 Loss reserving: past, present and future 13 Figure 2.2 Kalman filter and EDF filter Kalman filter Identity Normal Y j+1 = h -1 X j+1 β j+1 + v j+1 β j+1 j 1 u j+1 Non-identity Non-normal EDF filter Note that v j+1 and u j+1 may be from different sub-families of the EDF, e.g. v j+1 Poisson and u j+1 gamma. For simplicity, however, consider first the case in which both come from the same sub-family P 1, e.g. v j+1 and u j+1 both gamma. As previously remarked in connection with the Kalman filter, β j+1 j+1 is obtained by weighted least squares regression of the vector [Y T j+1, β T j+1 j] T on β j+1. In the case of the EDF filter, it may be obtained by GLM regression with link function h and error term from P 1. Figure 2.3 generalises Figure 2.1 to accommodate the EDF filter. To reduce this to the special case of the Kalman filter, replace the GLM regression mentioned at the foot of the diagram by least squares regression. c:\documents and settings\challis boland\eudora mail\attach\lecture1583.doc
17 Loss reserving: past, present and future 14 Figure 2.3 EDF filter Forecast new epoch s parameters and observations without new information EDF filter Update parameter estimates to incorporate new observation GLM regression of augmented data vector on parameter vector Calculate gain matrix (credibility of new observation) It is usual in GLM applications for all components of the error vector to be chosen from the same sub-family of the EDF, as was assumed above. This assumption is certainly made within GLM software packages such as SAS GENMOD, GLIM, EMBLEM, etc. It is not essential, however. The same packages assume that all components of the error vector are stochastically independent. If they are to be used to carry out the regression stage of Figure 2.3, then it will be necessary first to diagonalise the covariance matrix Γ j+1 j appearing in (2.1). The full details of the regression will be given in a forthcoming paper (Taylor and McGuire, 23). With this step included in the procedure, the EDF filter may be represented diagrammatically as in Figure 2.4. The bottom right box in this diagram is shown as stippled, and by-passed by the process. This reflects the fact that this step is implicit in the regression when it is carried out by a software package. c:\documents and settings\challis boland\eudora mail\attach\lecture1583.doc
18 Loss reserving: past, present and future 15 Figure 2.4 EDF filter implemented with GLM software Forecast new epoch s parameters and observations without new information Update parameter estimates to incorporate new observation GLM regression of augmented data vector on parameter vector EDF filter For use of GLM regression software Linear transformation of estimated parameter vector to diagonal covariance matrix Calculate gain matrix (credibility of new observation) Software performs this step There is a substantial literature dealing with the properties of estimates produced by the EDF filter for the simple special case of 1-dimensional observations (n=1). When the prior at j= is suitably conjugate to the EDF sub-family from which the distribution of data points is taken, the estimates are Bayes (Jewell, 1974; Nelder and Verrall, 1997; Landsman and Makov, 1998). For n>1, the parameter estimates produced by the EDF filter will not be Bayes in general. They can, however, be shown to approximately Bayes. A further conjecture is that they may be derivable as a form of stochastic approximation. Landsman and Makov (1999, 23) have demonstrated a stochastic approximation interpretation of credibility formulas, which, like Kalman filter parameter estimates, are also linear Bayes. The properties of EDF filter estimates are to be studied in the forthcoming paper (Taylor and McGuire, 23) referred to above. 2.4 Numerical examples Example 1: workers compensation payments This is an example of a workers compensation portfolio, in which benefits are dominated by payments of weekly compensation. Paid loss experience has been corrected for past inflation and recorded according to accident half-years and development half-years. The data are represented in a triangle of payments per claim incurred (PPCI), i.e. paid losses in each cell have been divided by the number of claims estimated to have been incurred in the relevant accident half-year. c:\documents and settings\challis boland\eudora mail\attach\lecture1583.doc
19 Loss reserving: past, present and future 16 Simple inspection of the data triangle reveals that there have been gradual but systematic shifts in PPCI from one past accident period to the next. Figure 2.5 illustrates this, displaying the PPCI of four accident half-years dispersed in time. The evident features in these data are: The steady upward movement of PPCI at the low development halfyears; The dramatic upward movement in later accident periods of PPCI in the mid-range of development half-years. Figure 2.5 PPCI by development half-year for various accident half-years PPCI by accident half-year PPCI development half-year 89H2 9H1 94H2 98H2 These changes are modelled using an EDF filter with log link and gamma error. Data vectors relate to accident half-years, so that the progress of the filter is as illustrated in Figure 2.6. Figure 2.6 Diagrammatic illustration of application of filter c:\documents and settings\challis boland\eudora mail\attach\lecture1583.doc
20 Loss reserving: past, present and future 17 The observation equation takes the generic form contained in (2.11), with Y j denoting the vector of PPCIs from accident half-year j, and the linear response term X j β j has m-th component β j + β j1 (m+1) + β j2 /(m+1) + β j3 /(m+1) 2 + β j4 δ m, m=,1,2,etc. (2.12) and with δ denoting the Kronecker delta. Figures 2.7 to 2.13 display data and filtered results for selected accident halfyears, showing how the models fitted to individual accident periods track the data. Figure 2.7 Figure 2.8 Initiation of the filter The next accident half-year s data PPCI 1 PPCI development half-year development half-year 89H2 data 89H2 fitted 89H2 data 89H2 fitted 9H1 data Figure 2.9 Figure 2.1 Filtered results More filtered results PPCI 1 PPCI development half-year development half-year 89H2 fitted 9H1 data 9H1 fitted 92H2 data 92H2 fitted 93H1 data Figure 2.11 Figure 2.12 More filtered results More filtered results PPCI 1 PPCI development half-year development half-year 92H2 fitted 93H1 data 93H1 fitted 97H2 data 97H2 fitted 98H1 data c:\documents and settings\challis boland\eudora mail\attach\lecture1583.doc
21 Loss reserving: past, present and future 18 Figure 2.13 More filtered results PPCI development half-year 97H2 fitted 98H1 data 98H1 fitted Example 2: motor vehicle bodily injury claim closure rates This example is taken directly from Chapter 1 of Taylor (2), and uses the data set that underlies the numerical examples from that book. The data are represented in a triangle of claim closure rates in which accident and development periods are both annual. As in the reference, the closure rate in any cell is calculated as: Number of claims closed in the cell Number open at the start + 1 / 3 x number newly reported in cell. Inspection of the data triangle reveals that closure rates are relatively flat over development years, but are subject to upward or downward shocks that tend to affect whole experience years (diagonals). Figure 2.14 illustrates, displaying closure rates for three separate experience years. Figure 2.14 Claim closures rates for different experience years 7% Claim closure rates 6% 5% Closure rate 4% 3% 2% 1% % >13 De velopment year Exp yrs Exp yrs Exp yrs c:\documents and settings\challis boland\eudora mail\attach\lecture1583.doc
22 Loss reserving: past, present and future 19 It is evident that numbers-based closure rates should be modelled with a discrete error structure, such as binomial. However, in the present case, most of the numbers of closures underlying the closure rates are sufficiently large that a normal approximation suffices. Hence, the EDF filter is applied to the rates with identity link, normal error, and weights equal to the denominators of the rates. This choice of link and error structure reduces the EDF filter to the Kalman filter. Since the shocks hit diagonals, the diagonals of the data triangle are taken as the data vectors, so that the progress of the filter is as illustrated in Figure Figure 2.15 Diagrammatic illustration of application of filter Let C jm denote the closure rate for accident year j and development year m. The model takes the form: C jm = β j + β j1 /(m+1) + β j2 /(m+1) 2 + γ k δ j+m,k, m=,1,2,etc. (2.13) where γ k denotes the shock to experience year k. When filtering takes place along diagonals in the manner illustrated, each data vector Y k (now subscripted by experience year k, instead of j) takes the form: Y k T = [C k, C 1,k-1, C 2,k-2,, C k ]. (2.14) The vector of parameters involved in this is the stacked vector [β T, β 1 T, β 2 T,,β k T, γ k ] T, (2.15) where β j T denotes [β j T, β j1 T, β j2 T ] T. That is, at the iteration of the filter that incorporates data diagonal k the parameter vector includes the parameters associated with all rows up to row k. This defines the observation equation for this example. This form of model was introduced by Verrall (1989). Full details of its application to the present example are given in Chapter 1 of Taylor (2). The system equation (2.1) takes the form: β j+1 = β j + w (β) j+1 (2.16) γ k+1 = γ k + w (γ) k+1. (2.17) c:\documents and settings\challis boland\eudora mail\attach\lecture1583.doc
23 Loss reserving: past, present and future 2 Figures 2.16 to 2.19 display data and filtered results for selected experience years. Each graph displays the closure rates of specific experience years, plotted against development year. Figure 2.16 Figure 2.17 Adding the 1991 data diagonal Fitting the 1991 diagonal Claim closure rate Claim closure rate Development year Development year 199 data 199 fitted 1991 data 199 fitted 1991 data 1991 fitted Figure 2.18 Figure 2.19 Adding the 1992 data diagonal Fitting the 1992 diagonal Clai m clo su re rate Development year Claim closure rate Development year 1991 fitted 1992 data 1992 fitted 1991 data 1991 fitted 1992 data The examples deal with the relatively high closure rates of 1991, and the much lower ones of The filtered results follow the data closely in their general level, i.e. there is little smoothing of the diagonal effects. There is, however, considerable smoothing across development years. Note that the fitted curves shown in these figures are not quite smooth. This is because each represents a diagonal, hence a cross-section of accident years. Consecutive points along one of these curves represent different accident years, and therefore, according to (2.16), are subject to different β parameters in (2.13).
24 Loss reserving: past, present and future From the present to the future 3.1 Loss experience triangles All discussion of loss reserving models to this point has been framed within the context of the triangle set out in Figure 1.1. This format of data is fundamental to the loss reserving literature. Indeed, the literature contains little else. The raw data underlying the triangle is, of course, much more extensive, relating as it does to individual claims. Typically, the raw loss data might comprise: A unit record claim header file, containing static information about the claim, such as claim number, name of claimant, date of occurrence, date of notification, etc. A unit record claim transaction file, containing a record for each payment made, each change of claim status (e.g. from open to closed), each change of estimate of incurred loss, etc. These are the real data. The triangle is a summary, whose origins are very much driven by the computational restrictions of a bygone era. The position of the triangle in the general schema of the construction of a loss reserve is as illustrated in Figure 3.1. Figure 3.1 General schema for construction of a loss reserve Loss Raw experience Loss Data triangle Model reserve statistical estimation The first stage of this procedure is essentially one of simplification, in which a data set of many dimensions is reduced to one of much lower dimensionality. The next stage, labelled statistical estimation, may be sophisticated or not, depending on where the loss reserving model fits into the Model structure and Parameter estimation dimensions of the phylogeny discussed in Section 1. If, however, the model is such that this stage involves a high degree of complexity, questions may well arise about the appropriateness of a pre-processing simplification stage. c:\documents and settings\challis boland\eudora mail\attach\lecture1583.doc
25 Loss reserving: past, present and future 22 The data summary stage of the loss reserving procedure, and even the loss experience triangle, is indeed integral to some of the more phenomenological model structures. Moreover, it may well be a useful data exploration tool at the stage at which one seeks to gain a basic understanding of the data in order to formulate a model. However, it is difficult to avoid the conclusion that one of its major motivations is the compression of data to meet the computational needs of a bygone era. Nowadays, when it is quite customary to use statistical software to process large data sets using sophisticated models, it seems natural to think of loss reserving models adapted to unit record claim data sets. Indeed, one can imagine future generations of students, educated on the basis of such models, finding the compression of data into a triangle quite artificial. This theme is pursued further by means of another numerical example. 3.2 Example 3: a non-triangular numerical example This example deals with another motor vehicle bodily injury portfolio. The data set comprises unit record data on all claims closed at non-zero cost, each record including: accident quarter closure quarter operational time at closure cost of claim. Here, all payments contributing to the cost of the claim are inflation corrected on the basis of some agreed index. The operational time at closure of claim X is, as usual, equal to the proportion of claims from X s accident quarter closed before it. Consistently with the previous example, let j = accident quarter m = development quarter k = j+m = experience quarter. In addition, let t = operational time at closure C(t,j,k) = claim cost. Assume that C(t,j,k) is subject to a quasi-likelihood from the EDF with V[C] = φ {E[C]} p (3.1) and p=2.3. This makes claim size substantially longer tailed than gamma (p=2). The model will take the general form:
26 Loss reserving: past, present and future 23 E[C(t,j,k)] = exp[f(t,j) + g(t,k)] (3.2) for functions f(t,j) and g(t,k) whose form is yet to be specified. First, however, note their purposes. The function f(t,j) is primarily a description of the pattern of claim size by operational time. Generally, the higher the value of t, the greater the expected claim size. The dependency on accident quarter recognises a discrete change in the form of the model at j=j on account of changes to legislation affecting claim sizes. That is, f(t,j) takes one form for j j, and another for j>j. The function f(t,j) depends on a sub-vector β (f) of the full parameter vector β which evolves according to (2.16), i.e. with changes from one accident period to the next. The function g(t,k) is primarily a description of superimposed inflation through its dependency on k, e.g. g(t,k) = γk would represent superimposed inflation at a constant rate of e γ -1 per quarter. The dependency on t allows the rate of superimposed inflation to vary with operational time, reflecting the fact that in this portfolio claims of small to medium injury severity tend to inflate rapidly whereas high severity claims do not. The function g(t,k) is given the more specific form: g(t,k) = k h(t) (3.3) where the appearance of k as a multiplier indicates the superimposed inflation effect, and h(t) = X (h) β (h) (3.4) the superscript (h) indicating that the design relates to just the function h(t) in (3.3) and (3.4) and β (h) is another sub-vector of the full parameter vector β. Let β (h) k denote the dynamic form of this sub-vector relating to experience period k. Its dynamics are as follows: β (h) k = β (h) k-1 + ε (h) k (3.5) with the ε (h) k stochastically independent error terms. Note the difference between (3.5) and (2.17). In the latter the diagonal effect is subject to upward and downward shocks, whereas in the former the rate of superimposed inflation h(t) receives the shocks. This means that the diagonal effect will evolve more smoothly than in the earlier example. The form (3.5) is also better adapted than (2.17) to a conventional understanding of superimposed inflation, whose rate is precisely the first difference of the diagonal effect in the linear response in a log-linear model (as here). Note also the difference in evolution of β (f) and β (h) in that one evolves over rows, the other over diagonals.
27 Loss reserving: past, present and future 24 This example illustrates a number of aspects of the power of the EDF filter, specifically: it models a large data set with a variety of structural complexities; it incorporates a long tailed distribution of individual claim size; it does so dynamically, with a subset of the parameters evolving over accident periods and another subset over experience periods; all of this can be expressed in concise and relatively simple mathematical form. As to the last point, the model is applied to data which, if put in triangular form, would form a 35x35 triangle, i.e. with 63 quarterly data cells. However, with some tens of thousands of claims and these labelled by operational time rather than development quarter, the data input has much greater dimension than this. Yet it is adequately modelled with 1 parameters: 7 describing the dependency on operational time of closure (f(t,j) in (3.2)); and 3 describing superimposed inflation (g(t,k) in (3.2)). In the application of the filter to this structure, each data vector relates to a new diagonal of claim closure, but consists of the vector of sizes of individual claims closed in the quarter, rather than a summary of these into development quarters or the like. Figures 3.1 to 3.4 illustrate the model fit of the filter as it stood at a couple of different closure quarters. In each of these plots: There is a comparison of actual and model claim sizes by development quarter. For purposes of comparison, the actual claim size (appearing in the plots as data ) is the average observed claim size in the development quarter. Model claim size is the fitted claim size at the average operational time in that development quarter. All claims in the plot are subject to the same superimposed inflation parameters β (h) k since they have the same value of k though, as seen in (3.3), superimposed inflation depends on operational time. Figure 3.1 Figure 3.2 Adding the December 1997 quarter diagonal Fitting the December 1997 quarter diagonal 6, 6, 5, 5, Average claim size 4, 3, 2, g 1, Average claim size 4, 3, 2, 1, Development quarter Sep-97 data Sep-97 fitted Dec-97 data Development quarter Sep-97 fitted Dec-97 data Dec-97 fitted
28 Loss reserving: past, present and future 25 Figure 3.3 Figure 3.4 Adding the December 2 quarter diagonal Fitting the December 2 quarter diagonal Average claim size 2, 18, 16, 14, 12, 1, 8, 6, 4, 2, Average claim size 2, 18, 16, 14, 12, 1, 8, 6, 4, 2, Development quarter Sep- data Sep- fitted Dec- data Development quarter Sep- fitted Dec- data Dec- fitted It is interesting to examine the trends in the filtered estimates of superimposed inflation parameters. The rate of superimposed inflation is modelled as a piecewise linear function of operational time with a peak at operational time 1 (%). Other analysis had suggested: Low superimposed inflation up to September 2; A much higher rate thereafter; and A further increase in the rate after March 22. Figures 3.5 to 3.1 track the shift in the estimated rate of superimposed inflation over the critical period from June 2 to March 23. Each plot shows the filtered estimates over three successive quarters. For the most part, successive plots overlap by one quarter. The general upward drift is visible over the course of the plots. For example, from the June 2 to the March 23 quarters, the estimated rate of superimposed inflation: Changes from about 6% p.a. to 7% p.a. at operational time 1% (its peak); Changes from about -4% p.a. to +3% p.a. at operational time 5%. Figure 3.5 Figure 3.6 Filtered estimates of superimposed inflation Filtered estimates of superimposed inflation June to December 2 quarters December 2 to June 21 quarters 1% 1% 8% 8% 6% 6% 4% 4% annualised SI g 2% % -2% -4% -6% annualised SI 2% % -2% -4% -6% % -8% -1% operational time -1% operational time Jun- Sep- Dec- Dec- Mar-1 Jun-1
29 Loss reserving: past, present and future 26 Figure 3.7 Figure 3.8 Filtered estimates of superimposed inflation Filtered estimates of superimposed inflation June to December 21 quarters December 21 to June 22 quarters annualised SI 1% 8% 6% 4% 2% % % -4% -6% -8% -1% operational time 9 1 annuali sed SI 1% 8% 6% 4% 2% % -2% -4% -6% -8% -1% operational time Jun-1 Sep-1 Dec-1 Dec-1 Mar-2 Jun-2 Figure 3.9 Figure 3.1 Filtered estimates of superimposed inflation Filtered estimates of superimposed inflation March to September 22 quarters September 22 to March 23 quarters 1% 1% 8% 8% 6% 6% 4% 4% annualised SI 2% % -2% annualised SI 2% % -2% % -4% -6% -6% -8% -8% -1% -1% operational time operational time Mar-2 Jun-2 Sep-2 Sep-2 Dec-2 Mar-3
30 Loss reserving: past, present and future Conclusion While triangular data formats may be useful tools for data exploration, there is no apparent reason why they should be retained as the structures for data analysis. It may often be preferable to model unit record claim data, or at least a considerably more comprehensive data structure than the conventional triangle. Section 3 gives an example to illustrate this point. The example is in many ways typical of loss reserving exercises involving large data sets. The main characteristics are that: The volume of data is sufficient that a number of subtleties in the algebraic structure to be estimated become manifest, e.g. o Dependency of the rate of superimposed inflation on some variables, rather than just a flat rate; o Changes in the relation between the dependent variable and the measure of development time (operational time in the example), possibly deriving from legislative change, or other specific change to the external claims environment; With the manifest nature of the fine structure of the data comes the likelihood that its underlying parameters will shift over time. The first characteristic motivates a detailed model structure such as a GLM. The second requires that time variation be incorporated in that GLM. One can, of course, include time-based interactions directly in a GLM. This is, however, rather different from converting it to a dynamic model by randomisation of parameters. On the one hand, the inclusion of time-based interactions in a GLM implies a model in which the time-dependent changes in model structure are assumed known with certainty, and only their parameters remain to be estimated. Shifts in underlying parameters will initially be ignored by such a model. Over time, the user will observe the failure of the model to fit more recent data, and will revise the model s algebraic structure. This will create a fundamental discontinuity in the sequence of estimates, such as loss reserves, flowing from the model. On the other hand, the dynamic models considered in the foregoing sections are of a Bayesian nature, and so result in more gradual recognition of underlying change. The degree of recognition will relate directly to the extent of the accumulated evidence of change. This will lead in turn to smooth changes in parameter estimates and in any derivatives of these estimates, such as loss reserves. Dynamic models may take various forms. The Kalman filter is an extremely powerful model, but is essentially adapted to the general linear model, in which there is a linear relation between explanatory variable and response, and all stochastic terms are normally distributed.
31 Loss reserving: past, present and future 28 This may not always be adequate for loss reserving problems, which tend to involve non-linearities and non-normal errors. For this reason the extension of the Kalman filter to the EDF filter in Section 2 may be useful in the loss reserving context.
32 Loss reserving: past, present and future 29 References De Jong, P. and Zehnwirth, B. (1983). Claims reserving state space models and the Kalman filter. Journal of the Institute of Actuaries, 11, De Vylder, F. (1978). Estimation of IBNR claims by least squares. Mitteilungen der Vereinigung Schweizerischer Versicherungsmathematiker, 78, England, P.D. and Verrall, R.J. (22). Stochastic claims reserving in general insurance. British Actuarial Journal. 8, Fisher, W.H. and Lange, J.T. (1973). Loss reserve testing: a report year approach. Proceedings of the Casualty Actuarial Society, 6, Hachemeister, C.A. and Stanard, J.N. (1975). IBNR claims count estimation with static lag functions. Paper presented to the 12 th Astin Colloquium, Portimao, Portugal, Harrison, P.J. and Stevens, C.F. (1976). Bayesian forecasting. Journal of the Royal Statistical Society, Series B, 38, Hertig, J. (1985). A statistical approach to the IBNR-reserves in marine reinsurance. ASTIN Bulletin, 15, Jewell, W.S. (1974). The credible distribution. Astin Bulletin,7, Kalman, R.E. (196). A new approach to linear filtering and prediction problems. Journal of Basic Engineering, 82, Kramreiter, H. and Straub, E. (1973). On the calculation of IBNR reserves II. Mitteilungen der Vereinigung Schweizerischer Versicherungsmathematiker, 73, Landsman, Z. and Makov, U. (1998). Exponential dispersion models and credibility. Scandinavian Actuarial Journal, Landsman, Z. and Makov, U. (1999). On stochastic approximation and credibility. Scandinavian Actuarial Journal, Landsman, Z. and Makov, U. (23). Sequential quasi credibility for scale dispersion models. Scandinavian Actuarial Journal, McCullagh P. and Nelder J.A. (1989). Generalised Linear Models (3 rd Edition). Chapman and Hall, New York. Mack, T. (1993). Distribution-free calculation of the standard error of chain ladder reserve estimates, Astin Bulletin, 23, Nelder, J.A. and Verrall, R.J. (1997). Credibility theory and generalised linear models. Astin Bulletin, 27,
33 Loss reserving: past, present and future 3 Pollard, J.H. (1983). Outstanding claims provisions: a distribution-free statistical approach. Journal of the Institute of Actuaries, 19, Also appears in Institute of Actuaries of Australia General Insurance Seminar, 3, Reid, D.H. (1978). Claim reserves in general insurance. Journal of the Institute of Actuaries, 15, Sawkins, R.W. (1979). Methods of analysing claim payments in general insurance. Transactions of the Institute of Actuaries of Australia, Taylor, G.C. (1977). Separation of inflation and other effects from the distribution of non-life insurance claim delays. Astin Bulletin, 9, Taylor, G.C. (1986). Claim Reserving in Non-life Insurance. North-Holland, Amsterdam. Taylor, G.C. (2). Loss reserving an actuarial perspective. Kluwer Academic Publishers, Boston. Taylor, G.C. and Ashe, F.R. (1983). Second moments of estimates of outstanding claims. Journal of Econometrics, 23, Taylor, G.C. and Campbell, M. (22). Statistical Case Estimation. Research Paper No. 14, Centre for Actuarial Studies, University of Melbourne. Verrall, R.J. (1989). A State Space Representation of the Chain Ladder Linear Model. Journal of the Institute of Actuaries, 116, Wright, T.S. (199). A stochastic method for claims reserving in general insurance. Journal of the Institute of Actuaries, 117,
34 Loss reserving: past, present and future 31 RECENT ADDITIONS TO THE RESEARCH PAPER SERIES No. Date Subject Author 84 FEBRUARY 21 DISCRETE TIME RISK MODELS UNDER STOCHASTIC FORCES OF INTEREST 85 FEBRUARY 21 MODERN LANDMARKS IN ACTUARIAL SCIENCE Inaugural Professorial Address 86 JUNE 21 LUNDBERG INEQUALITIES FOR RENEWAL EQUATIONS 87 SEPTEMBER 21 VOLATILITY, BETA AND RETURN WAS THERE EVER A MEANINGFUL RELATIONSHIP? 88 NOVEMBER 21 EXPLICIT, FINITE TIME RUIN PROBABILITIES FOR DISCRETE, DEPENDENT CLAIMS 89 NOVEMBER 21 ON THE DISTRIBUTION OF THE DEFICIT AT RUIN WHEN CLAIMS ARE PHASE-TYPE Jun Cai David C M Dickson Gordon E Willmot Jun Cai X Sheldon Lin Richard Fitzherbert Zvetan G Ignatov Vladimir K Kaishev Rossen S Krachunov Steve Drekic David C M Dickson David A Stanford Gordon E Willmot 9 NOVEMBER 21 THE INTEGRATED SQUARE-ROOT PROCESS Daniel Dufresne 91 NOVEMBER 21 ON THE EXPECTED DISCOUNTED PENALTY FUNCTION AT RUIN OF A SURPLUS PROCESS WITH INTEREST Jun Cai David C M Dickson 92 JANUARY 22 CHAIN LADDER BIAS Greg Taylor 93 JANUARY 22 FURTHER OBSERVATIONS ON CHAIN LADDER BIAS Greg Taylor 94 JANUARY 22 A GENERAL CLASS OF RISK MODELS Daniel Dufresne 95 JANUARY 22 THE DISTRIBUTION OF THE TIME TO RUIN IN THE CLASSICAL RISK MODEL 96 MAY 22 A NOTE ON THE MAXIMUM SEVERITY OF RUIN AND RELATED PROBLEMS 97 JUNE 22 UPPER BOUNDS FOR ULTIMATE RUIN PROBABILITIES IN THE SPARRE ANDERSEN MODEL WITH INTEREST 98 JUNE 22 CONTINUOUS COMPOUNDING, VOLATILITY AND THE EQUITY PREMIUM David C M Dickson Howard R Waters David C M Dickson Jun Cai David C M Dickson Richard Fitzherbert
35 Loss reserving: past, present and future 32 No. Date Subject Author 99 JUNE 22 THE DEFICIT AT RUIN IN THE STATIONARY RENEWAL RISK MODEL Gordon E Willmot David C M Dickson Steve Drekic David A Stanford 1 AUGUST 22 ASIAN AND BASKET ASYMPTOTICS Daniel Dufresne 11 AUGUST 22 RUIN PROBABILITIES WITH A MARKOV CHAIN INTEREST MODEL 12 AUGUST 22 THE GERBER-SHIU DISCOUNTED PENALTY FUNCTION IN THE STATIONARY RENEWAL RISK MODEL 13 NOVEMBER 22 INITIAL CAPITAL AND MARGINS REQUIRED TO SECURE A JAPANESE LIFE INSURANCE POLICY PORTFOLIO UNDER VARIABLE INTEREST RATES Jun Cai David C M Dickson Gordon E Willmot David C M Dickson Manabu Sato David C M Dickson Richard M Fitzherbert 14 NOVEMBER 22 STATISTICAL CASE ESTIMATION Greg Taylor Mireille Campbell 15 MARCH 23 ACTUARIAL PRACTICE AND CONTROL: OBJECTIVES AND CAPABILITIES 16 MARCH 23 THE IDENTIFICATION AND MEASUREMENT OF SPECULATIVE RISK Julian D Gribble Richard Fitzherbert 17 AUGUST 23 SOME OPTIMAL DIVIDENDS PROBLEMS David C M Dickson Howard R Waters 18 AUGUST 23 THE JOINT DISTRIBUTION OF THE SURPLUS PRIOR TO RUIN AND THE DEFICIT AT RUIN IN SOME SPARRE ANDERSEN MODELS David C M Dickson Steve Drekic 19 SEPTEMBER 23 LOSS RESERVING: PAST, PRESENT AND FUTURE Greg Taylor Gráinne McGuire Alan Greenfield A complete list of papers can be found at
SAS Software to Fit the Generalized Linear Model
SAS Software to Fit the Generalized Linear Model Gordon Johnston, SAS Institute Inc., Cary, NC Abstract In recent years, the class of generalized linear models has gained popularity as a statistical modeling
Combining GLM and datamining techniques for modelling accident compensation data. Peter Mulquiney
Combining GLM and datamining techniques for modelling accident compensation data Peter Mulquiney Introduction Accident compensation data exhibit features which complicate loss reserving and premium rate
A three dimensional stochastic Model for Claim Reserving
A three dimensional stochastic Model for Claim Reserving Magda Schiegl Haydnstr. 6, D - 84088 Neufahrn, [email protected] and Cologne University of Applied Sciences Claudiusstr. 1, D-50678 Köln
11. Time series and dynamic linear models
11. Time series and dynamic linear models Objective To introduce the Bayesian approach to the modeling and forecasting of time series. Recommended reading West, M. and Harrison, J. (1997). models, (2 nd
STOCHASTIC CLAIMS RESERVING IN GENERAL INSURANCE. By P. D. England and R. J. Verrall. abstract. keywords. contact address. ".
B.A.J. 8, III, 443-544 (2002) STOCHASTIC CLAIMS RESERVING IN GENERAL INSURANCE By P. D. England and R. J. Verrall [Presented to the Institute of Actuaries, 28 January 2002] abstract This paper considers
An Extension Model of Financially-balanced Bonus-Malus System
An Extension Model of Financially-balanced Bonus-Malus System Other : Ratemaking, Experience Rating XIAO, Yugu Center for Applied Statistics, Renmin University of China, Beijing, 00872, P.R. China Phone:
Development Period 1 2 3 4 5 6 7 8 9 Observed Payments
Pricing and reserving in the general insurance industry Solutions developed in The SAS System John Hansen & Christian Larsen, Larsen & Partners Ltd 1. Introduction The two business solutions presented
Runoff of the Claims Reserving Uncertainty in Non-Life Insurance: A Case Study
1 Runoff of the Claims Reserving Uncertainty in Non-Life Insurance: A Case Study Mario V. Wüthrich Abstract: The market-consistent value of insurance liabilities consists of the best-estimate prediction
Underwriting risk control in non-life insurance via generalized linear models and stochastic programming
Underwriting risk control in non-life insurance via generalized linear models and stochastic programming 1 Introduction Martin Branda 1 Abstract. We focus on rating of non-life insurance contracts. We
PREDICTIVE DISTRIBUTIONS OF OUTSTANDING LIABILITIES IN GENERAL INSURANCE
PREDICTIVE DISTRIBUTIONS OF OUTSTANDING LIABILITIES IN GENERAL INSURANCE BY P.D. ENGLAND AND R.J. VERRALL ABSTRACT This paper extends the methods introduced in England & Verrall (00), and shows how predictive
Motor and Household Insurance: Pricing to Maximise Profit in a Competitive Market
Motor and Household Insurance: Pricing to Maximise Profit in a Competitive Market by Tom Wright, Partner, English Wright & Brockman 1. Introduction This paper describes one way in which statistical modelling
STATISTICA Formula Guide: Logistic Regression. Table of Contents
: Table of Contents... 1 Overview of Model... 1 Dispersion... 2 Parameterization... 3 Sigma-Restricted Model... 3 Overparameterized Model... 4 Reference Coding... 4 Model Summary (Summary Tab)... 5 Summary
The zero-adjusted Inverse Gaussian distribution as a model for insurance claims
The zero-adjusted Inverse Gaussian distribution as a model for insurance claims Gillian Heller 1, Mikis Stasinopoulos 2 and Bob Rigby 2 1 Dept of Statistics, Macquarie University, Sydney, Australia. email:
Chapter 6: The Information Function 129. CHAPTER 7 Test Calibration
Chapter 6: The Information Function 129 CHAPTER 7 Test Calibration 130 Chapter 7: Test Calibration CHAPTER 7 Test Calibration For didactic purposes, all of the preceding chapters have assumed that the
MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.
MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0-534-40596-7. Systems of Linear Equations Definition. An n-dimensional vector is a row or a column
Master of Mathematical Finance: Course Descriptions
Master of Mathematical Finance: Course Descriptions CS 522 Data Mining Computer Science This course provides continued exploration of data mining algorithms. More sophisticated algorithms such as support
Stochastic programming approaches to pricing in non-life insurance
Stochastic programming approaches to pricing in non-life insurance Martin Branda Charles University in Prague Department of Probability and Mathematical Statistics 11th International Conference on COMPUTATIONAL
Machine Learning and Pattern Recognition Logistic Regression
Machine Learning and Pattern Recognition Logistic Regression Course Lecturer:Amos J Storkey Institute for Adaptive and Neural Computation School of Informatics University of Edinburgh Crichton Street,
One-year reserve risk including a tail factor : closed formula and bootstrap approaches
One-year reserve risk including a tail factor : closed formula and bootstrap approaches Alexandre Boumezoued R&D Consultant Milliman Paris [email protected] Yoboua Angoua Non-Life Consultant
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS 1. SYSTEMS OF EQUATIONS AND MATRICES 1.1. Representation of a linear system. The general system of m equations in n unknowns can be written a 11 x 1 + a 12 x 2 +
Cover. Optimal Retentions with Ruin Probability Target in The case of Fire. Insurance in Iran
Cover Optimal Retentions with Ruin Probability Target in The case of Fire Insurance in Iran Ghadir Mahdavi Ass. Prof., ECO College of Insurance, Allameh Tabataba i University, Iran Omid Ghavibazoo MS in
Stochastic Claims Reserving Methods in Non-Life Insurance
Stochastic Claims Reserving Methods in Non-Life Insurance Mario V. Wüthrich 1 Department of Mathematics ETH Zürich Michael Merz 2 Faculty of Economics University Tübingen Version 1.1 1 ETH Zürich, CH-8092
Matrices and Polynomials
APPENDIX 9 Matrices and Polynomials he Multiplication of Polynomials Let α(z) =α 0 +α 1 z+α 2 z 2 + α p z p and y(z) =y 0 +y 1 z+y 2 z 2 + y n z n be two polynomials of degrees p and n respectively. hen,
Bootstrap Modeling: Beyond the Basics
Mark R. Shapland, FCAS, ASA, MAAA Jessica (Weng Kah) Leong, FIAA, FCAS, MAAA Abstract Motivation. Bootstrapping is a very versatile model for estimating a distribution of possible outcomes for the unpaid
Statistics in Retail Finance. Chapter 6: Behavioural models
Statistics in Retail Finance 1 Overview > So far we have focussed mainly on application scorecards. In this chapter we shall look at behavioural models. We shall cover the following topics:- Behavioural
Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model
Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model 1 September 004 A. Introduction and assumptions The classical normal linear regression model can be written
A credibility method for profitable cross-selling of insurance products
Submitted to Annals of Actuarial Science manuscript 2 A credibility method for profitable cross-selling of insurance products Fredrik Thuring Faculty of Actuarial Science and Insurance, Cass Business School,
STRATEGIC CAPACITY PLANNING USING STOCK CONTROL MODEL
Session 6. Applications of Mathematical Methods to Logistics and Business Proceedings of the 9th International Conference Reliability and Statistics in Transportation and Communication (RelStat 09), 21
Principle Component Analysis and Partial Least Squares: Two Dimension Reduction Techniques for Regression
Principle Component Analysis and Partial Least Squares: Two Dimension Reduction Techniques for Regression Saikat Maitra and Jun Yan Abstract: Dimension reduction is one of the major tasks for multivariate
Analysis of a Production/Inventory System with Multiple Retailers
Analysis of a Production/Inventory System with Multiple Retailers Ann M. Noblesse 1, Robert N. Boute 1,2, Marc R. Lambrecht 1, Benny Van Houdt 3 1 Research Center for Operations Management, University
Least Squares Estimation
Least Squares Estimation SARA A VAN DE GEER Volume 2, pp 1041 1045 in Encyclopedia of Statistics in Behavioral Science ISBN-13: 978-0-470-86080-9 ISBN-10: 0-470-86080-4 Editors Brian S Everitt & David
Section A. Index. Section A. Planning, Budgeting and Forecasting Section A.2 Forecasting techniques... 1. Page 1 of 11. EduPristine CMA - Part I
Index Section A. Planning, Budgeting and Forecasting Section A.2 Forecasting techniques... 1 EduPristine CMA - Part I Page 1 of 11 Section A. Planning, Budgeting and Forecasting Section A.2 Forecasting
OPTIMAl PREMIUM CONTROl IN A NON-liFE INSURANCE BUSINESS
ONDERZOEKSRAPPORT NR 8904 OPTIMAl PREMIUM CONTROl IN A NON-liFE INSURANCE BUSINESS BY M. VANDEBROEK & J. DHAENE D/1989/2376/5 1 IN A OPTIMAl PREMIUM CONTROl NON-liFE INSURANCE BUSINESS By Martina Vandebroek
Non-Life Insurance Mathematics
Thomas Mikosch Non-Life Insurance Mathematics An Introduction with the Poisson Process Second Edition 4y Springer Contents Part I Collective Risk Models 1 The Basic Model 3 2 Models for the Claim Number
Time series Forecasting using Holt-Winters Exponential Smoothing
Time series Forecasting using Holt-Winters Exponential Smoothing Prajakta S. Kalekar(04329008) Kanwal Rekhi School of Information Technology Under the guidance of Prof. Bernard December 6, 2004 Abstract
Financial Simulation Models in General Insurance
Financial Simulation Models in General Insurance By - Peter D. England Abstract Increases in computer power and advances in statistical modelling have conspired to change the way financial modelling is
Review of Basic Options Concepts and Terminology
Review of Basic Options Concepts and Terminology March 24, 2005 1 Introduction The purchase of an options contract gives the buyer the right to buy call options contract or sell put options contract some
STA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! [email protected]! http://www.cs.toronto.edu/~rsalakhu/ Lecture 6 Three Approaches to Classification Construct
business statistics using Excel OXFORD UNIVERSITY PRESS Glyn Davis & Branko Pecar
business statistics using Excel Glyn Davis & Branko Pecar OXFORD UNIVERSITY PRESS Detailed contents Introduction to Microsoft Excel 2003 Overview Learning Objectives 1.1 Introduction to Microsoft Excel
Predictive Modeling in Actuarial Science
1 Predictive Modeling in Actuarial Science Edward W. Frees, Richard A. Derrig, and Glenn Meyers Chapter Preview. Predictive modeling involves the use of data to forecast future events. It relies on capturing
CHAPTER 2 Estimating Probabilities
CHAPTER 2 Estimating Probabilities Machine Learning Copyright c 2016. Tom M. Mitchell. All rights reserved. *DRAFT OF January 24, 2016* *PLEASE DO NOT DISTRIBUTE WITHOUT AUTHOR S PERMISSION* This is a
Poisson Models for Count Data
Chapter 4 Poisson Models for Count Data In this chapter we study log-linear models for count data under the assumption of a Poisson error structure. These models have many applications, not only to the
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION Introduction In the previous chapter, we explored a class of regression models having particularly simple analytical
The Study of Chinese P&C Insurance Risk for the Purpose of. Solvency Capital Requirement
The Study of Chinese P&C Insurance Risk for the Purpose of Solvency Capital Requirement Xie Zhigang, Wang Shangwen, Zhou Jinhan School of Finance, Shanghai University of Finance & Economics 777 Guoding
STATISTICAL METHODS IN GENERAL INSURANCE. Philip J. Boland National University of Ireland, Dublin [email protected]
STATISTICAL METHODS IN GENERAL INSURANCE Philip J. Boland National University of Ireland, Dublin [email protected] The Actuarial profession appeals to many with excellent quantitative skills who aspire
15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
Analysis of Bayesian Dynamic Linear Models
Analysis of Bayesian Dynamic Linear Models Emily M. Casleton December 17, 2010 1 Introduction The main purpose of this project is to explore the Bayesian analysis of Dynamic Linear Models (DLMs). The main
CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES. From Exploratory Factor Analysis Ledyard R Tucker and Robert C.
CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES From Exploratory Factor Analysis Ledyard R Tucker and Robert C MacCallum 1997 180 CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES In
S T O C H A S T I C C L A I M S R E S E R V I N G O F A S W E D I S H L I F E I N S U R A N C E P O R T F O L I O
S T O C H A S T I C C L A I M S R E S E R V I N G O F A S W E D I S H L I F E I N S U R A N C E P O R T F O L I O DIPLOMA THESIS PRESENTED FOR THE. FOR SWEDISH ACTUARIAL SO CIETY PETER NIMAN 2007 ABSTRACT
Logistic Regression (1/24/13)
STA63/CBB540: Statistical methods in computational biology Logistic Regression (/24/3) Lecturer: Barbara Engelhardt Scribe: Dinesh Manandhar Introduction Logistic regression is model for regression used
Chapter 10 Run-off triangles
0 INTRODUCTION 1 Chapter 10 Run-off triangles 0 Introduction 0.1 Run-off triangles Run-off triangles are used in general insurance to forecast future claim numbers and amounts. In this chapter we will
Penalized Logistic Regression and Classification of Microarray Data
Penalized Logistic Regression and Classification of Microarray Data Milan, May 2003 Anestis Antoniadis Laboratoire IMAG-LMC University Joseph Fourier Grenoble, France Penalized Logistic Regression andclassification
How To Understand The Theory Of Probability
Graduate Programs in Statistics Course Titles STAT 100 CALCULUS AND MATR IX ALGEBRA FOR STATISTICS. Differential and integral calculus; infinite series; matrix algebra STAT 195 INTRODUCTION TO MATHEMATICAL
Statistical Machine Learning
Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes
ECON20310 LECTURE SYNOPSIS REAL BUSINESS CYCLE
ECON20310 LECTURE SYNOPSIS REAL BUSINESS CYCLE YUAN TIAN This synopsis is designed merely for keep a record of the materials covered in lectures. Please refer to your own lecture notes for all proofs.
The Basics of Graphical Models
The Basics of Graphical Models David M. Blei Columbia University October 3, 2015 Introduction These notes follow Chapter 2 of An Introduction to Probabilistic Graphical Models by Michael Jordan. Many figures
Small Business Lending *
Reserve Small Business Bank of Lending Australia Bulletin Small Business Lending * These notes were prepared in response to a request from the House of Representatives Standing Committee on Financial Institutions
Nonparametric adaptive age replacement with a one-cycle criterion
Nonparametric adaptive age replacement with a one-cycle criterion P. Coolen-Schrijner, F.P.A. Coolen Department of Mathematical Sciences University of Durham, Durham, DH1 3LE, UK e-mail: [email protected]
Bayesian Statistics: Indian Buffet Process
Bayesian Statistics: Indian Buffet Process Ilker Yildirim Department of Brain and Cognitive Sciences University of Rochester Rochester, NY 14627 August 2012 Reference: Most of the material in this note
Offset Techniques for Predictive Modeling for Insurance
Offset Techniques for Predictive Modeling for Insurance Matthew Flynn, Ph.D, ISO Innovative Analytics, W. Hartford CT Jun Yan, Ph.D, Deloitte & Touche LLP, Hartford CT ABSTRACT This paper presents the
By choosing to view this document, you agree to all provisions of the copyright laws protecting it.
This material is posted here with permission of the IEEE Such permission of the IEEE does not in any way imply IEEE endorsement of any of Helsinki University of Technology's products or services Internal
Adaptive Demand-Forecasting Approach based on Principal Components Time-series an application of data-mining technique to detection of market movement
Adaptive Demand-Forecasting Approach based on Principal Components Time-series an application of data-mining technique to detection of market movement Toshio Sugihara Abstract In this study, an adaptive
FULL LIST OF REFEREED JOURNAL PUBLICATIONS Qihe Tang
FULL LIST OF REFEREED JOURNAL PUBLICATIONS Qihe Tang 87. Li, J.; Tang, Q. Interplay of insurance and financial risks in a discrete-time model with strongly regular variation. Bernoulli 21 (2015), no. 3,
GENERALIZED LINEAR MODELS IN VEHICLE INSURANCE
ACTA UNIVERSITATIS AGRICULTURAE ET SILVICULTURAE MENDELIANAE BRUNENSIS Volume 62 41 Number 2, 2014 http://dx.doi.org/10.11118/actaun201462020383 GENERALIZED LINEAR MODELS IN VEHICLE INSURANCE Silvie Kafková
Exam Introduction Mathematical Finance and Insurance
Exam Introduction Mathematical Finance and Insurance Date: January 8, 2013. Duration: 3 hours. This is a closed-book exam. The exam does not use scrap cards. Simple calculators are allowed. The questions
QUALITY ENGINEERING PROGRAM
QUALITY ENGINEERING PROGRAM Production engineering deals with the practical engineering problems that occur in manufacturing planning, manufacturing processes and in the integration of the facilities and
Application of Credibility Theory to Group Life Pricing
Prepared by: Manuel Tschupp, MSc ETH Application of Credibility Theory to Group Life Pricing Extended Techniques TABLE OF CONTENTS 1. Introduction 3 1.1 Motivation 3 1.2 Fundamentals 1.3 Structure 3 4
A Deeper Look Inside Generalized Linear Models
A Deeper Look Inside Generalized Linear Models University of Minnesota February 3 rd, 2012 Nathan Hubbell, FCAS Agenda Property & Casualty (P&C Insurance) in one slide The Actuarial Profession Travelers
1. (First passage/hitting times/gambler s ruin problem:) Suppose that X has a discrete state space and let i be a fixed state. Let
Copyright c 2009 by Karl Sigman 1 Stopping Times 1.1 Stopping Times: Definition Given a stochastic process X = {X n : n 0}, a random time τ is a discrete random variable on the same probability space as
GLM I An Introduction to Generalized Linear Models
GLM I An Introduction to Generalized Linear Models CAS Ratemaking and Product Management Seminar March 2009 Presented by: Tanya D. Havlicek, Actuarial Assistant 0 ANTITRUST Notice The Casualty Actuarial
The Scientific Data Mining Process
Chapter 4 The Scientific Data Mining Process When I use a word, Humpty Dumpty said, in rather a scornful tone, it means just what I choose it to mean neither more nor less. Lewis Carroll [87, p. 214] In
Modelling the Claims Development Result for Solvency Purposes
Modelling the Claims Development Result for Solvency Purposes Michael Merz, Mario V. Wüthrich Version: June 10, 008 Abstract We assume that the claims liability process satisfies the distribution-free
Statistical Models in R
Statistical Models in R Some Examples Steven Buechler Department of Mathematics 276B Hurley Hall; 1-6233 Fall, 2007 Outline Statistical Models Structure of models in R Model Assessment (Part IA) Anova
Chapter 9 Experience rating
0 INTRODUCTION 1 Chapter 9 Experience rating 0 Introduction The rating process is the process of deciding on an appropriate level of premium for a particular class of insurance business. The contents of
Industry Environment and Concepts for Forecasting 1
Table of Contents Industry Environment and Concepts for Forecasting 1 Forecasting Methods Overview...2 Multilevel Forecasting...3 Demand Forecasting...4 Integrating Information...5 Simplifying the Forecast...6
6 Scalar, Stochastic, Discrete Dynamic Systems
47 6 Scalar, Stochastic, Discrete Dynamic Systems Consider modeling a population of sand-hill cranes in year n by the first-order, deterministic recurrence equation y(n + 1) = Ry(n) where R = 1 + r = 1
pwc.com.au NT WorkSafe Actuarial review of Northern Territory workers compensation scheme as at June 2014
pwc.com.au NT WorkSafe Actuarial review of Northern Territory workers compensation scheme as at June 2014 30 June 2013 Key findings Funding ratio The following table shows the funding ratio for the scheme
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a
IASB Educational Session Non-Life Claims Liability
IASB Board Meeting Observer Note- Agenda Paper 10 January 2005 IASB Educational Session Non-Life Claims Liability Presented by the International Actuarial Association January 19, 2005 Sam Gutterman and
Predictive Modeling in Workers Compensation 2008 CAS Ratemaking Seminar
Predictive Modeling in Workers Compensation 2008 CAS Ratemaking Seminar Prepared by Louise Francis, FCAS, MAAA Francis Analytics and Actuarial Data Mining, Inc. www.data-mines.com [email protected]
Stocks paying discrete dividends: modelling and option pricing
Stocks paying discrete dividends: modelling and option pricing Ralf Korn 1 and L. C. G. Rogers 2 Abstract In the Black-Scholes model, any dividends on stocks are paid continuously, but in reality dividends
Marketing Mix Modelling and Big Data P. M Cain
1) Introduction Marketing Mix Modelling and Big Data P. M Cain Big data is generally defined in terms of the volume and variety of structured and unstructured information. Whereas structured data is stored
Handling attrition and non-response in longitudinal data
Longitudinal and Life Course Studies 2009 Volume 1 Issue 1 Pp 63-72 Handling attrition and non-response in longitudinal data Harvey Goldstein University of Bristol Correspondence. Professor H. Goldstein
MEMORANDUM. Re: Technical input to the ACCC submission relationship between the regulatory asset base and the renewals annuity
MEMORANDUM To: Peter McGahan From: Jeff Balchin Director Date: 4 July 2008 Re: Technical input to the ACCC submission relationship between the regulatory asset base and the renewals annuity A. Introduction
Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.
Statistical Learning: Chapter 4 Classification 4.1 Introduction Supervised learning with a categorical (Qualitative) response Notation: - Feature vector X, - qualitative response Y, taking values in C
Load Balancing and Switch Scheduling
EE384Y Project Final Report Load Balancing and Switch Scheduling Xiangheng Liu Department of Electrical Engineering Stanford University, Stanford CA 94305 Email: [email protected] Abstract Load
Automated Biosurveillance Data from England and Wales, 1991 2011
Article DOI: http://dx.doi.org/10.3201/eid1901.120493 Automated Biosurveillance Data from England and Wales, 1991 2011 Technical Appendix This online appendix provides technical details of statistical
TESTING THE ASSUMPTIONS OF AGE-TO-AGE FACTORS. Abstract
TESTING THE ASSUMPTIONS OF AGE-TO-AGE FACTORS GARY G. VENTER Abstract The use of age-to-age factors applied to cumulative losses has been shown to produce least-squares optimal reserve estimates when certain
Nonlinear Iterative Partial Least Squares Method
Numerical Methods for Determining Principal Component Analysis Abstract Factors Béchu, S., Richard-Plouet, M., Fernandez, V., Walton, J., and Fairley, N. (2016) Developments in numerical treatments for
Introduction to General and Generalized Linear Models
Introduction to General and Generalized Linear Models General Linear Models - part I Henrik Madsen Poul Thyregod Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby
GUIDANCE NOTE 253 - DETERMINATION OF LIFE INSURANCE POLICY LIABILITIES
THE INSTITUTE OF ACTUARIES OF AUSTRALIA A.C.N. 000 423 656 GUIDANCE NOTE 253 - DETERMINATION OF LIFE INSURANCE POLICY LIABILITIES APPLICATION Appointed Actuaries of Life Insurance Companies. LEGISLATION
