/
Business School Self-assembling insurance claim models Business School Self-assembling insurance claim models

Business School Self-assembling insurance claim models - PowerPoint Presentation

sherrill-nordquist
sherrill-nordquist . @sherrill-nordquist
Follow
383 views
Uploaded On 2018-03-16

Business School Self-assembling insurance claim models - PPT Presentation

Greg Taylor School of Risk and Actuarial Studies UNSW Australia Sydney Australia Hugh Miller Gráinne McGuire Taylor Fry Analytics amp Actuarial Consulting Sydney Australia 2nd International ID: 653049

set data model loss data set loss model lasso reserving tracking regression basis simulated framework function glm parameters ladder

Share:

Link:

Embed:

Download Presentation from below link

Download Presentation The PPT/PDF document "Business School Self-assembling insuranc..." is the property of its rightful owner. Permission is granted to download and print the materials on this web site for personal, non-commercial use only, and to display it on your personal computer provided you do not modify the materials and that you retain all copyright notices contained in the materials. By downloading content from our website, you accept the terms of this agreement.


Presentation Transcript

Slide1

Business School

Self-assembling insurance claim modelsGreg TaylorSchool of Risk and Actuarial Studies, UNSW AustraliaSydney, AustraliaHugh MillerGráinne McGuireTaylor Fry Analytics & Actuarial Consulting, Sydney, Australia2nd International Congress on Actuarial Science and Quantitative FinanceCartagena, Colombia, June 15-18, 2016.Slide2

OverviewMotivationRegularized regressionIn generalThe lassoLoss reserving framework and notationApplication of lasso to loss reservingTest on simulated dataData setResultsFurther testing and developmentConclusion Slide3

Motivation (1)Consider loss reserving on the basis of a conventional triangular data sete.g. paid losses, incurred losses, etc.A model often used is the chain ladderThis involves a very simple model structureIt will be inadequate for the capture of certain claim data characteristics from the real world (illustrated later)Slide4

Motivation (2)When such features are present, they may be modelled by means of a Generalized Linear Model (GLM) (McGuire, 2007; Taylor & McGuire, 2004, 2016)But construction of this type of model requires many hours (perhaps a week) of a highly skilled analystTime-consumingExpensiveObjective is to consider more automated modelling that produces a similar GLM but at much less time and expenseSlide5

OverviewMotivationRegularized regressionIn generalThe lassoLoss reserving framework and notationApplication of lasso to loss reservingTest on simulated dataData setResultsFurther testing and developmentConclusion

Slide6

Regularized regression: in generalConsider a standard (multivariate) linear regression problem, expressed in vector and matrix form:

Estimation of parameter vector by OLS loss function is

[

least squares

]

where

denotes the

-norm

:

regularized

regression

loss function

is

 

Penalty for poor fit

Tuning constant (𝜆)

 

Penalty for additional parametersSlide7

Regularized regression: the lassoRegularized regression loss function (previous slide)

Special cases: OLS regression (no penalty): Ridge regression

: Lasso (Least Absolute Shrinkage and Selection Operator)

Adaptation to Generalized Linear Models (GLM)GLM takes form

Regularized regression loss

function becomes

 

Link function

Stochastic error (EDF)

Log-likelihoodSlide8

Formal derivations of the GLM lassoConstrained parametersFit GLM by MLE subject to parameter constraint

Random effects GLMMAP (maximum a posteriori) estimation of when parameters subject to random effects with independent Laplace distributed priors:

 Slide9

Application of GLM lassoRegularized GLM regression loss function (earlier slide)

Lasso version ()

The second member of the loss function tends to force parameters to zero

: model approaches conventional GLM

:

all parameter estimates approach zero

Intermediate values

of

control the complexity of the model (number of non-zero parameters)

 Slide10

OverviewMotivationRegularized regressionIn generalThe lassoLoss reserving framework and notationApplication of lasso to loss reservingTest on simulated dataData setResultsFurther testing and developmentConclusion

Slide11

Loss reserving framework and notation (1)Experimental simulated data setsIncremental quarterly paid claim triangles (40x40)Notation

accident quarter development quarter payment quarter incremental paid losses in cell

Assumed that

(generalized chain ladder)

The

are selected throughout to be consistent with the Mack formulation of the chain ladder model

 Slide12

Loss reserving framework and notation (2)It is necessary to decide the set of regressors to be included in the model, i.e. the rows of the design matrix Each regressor

is some function of The present application includes all of the following regressors (“basis functions”)All unit step functions of or Various locations of the stepAll unit gradient ramp functions of or Various locations of the start and end of the rampCombinations of these (linear splines) can approximate smooth curves(later) interactions between the step basis functions Slide13

OverviewMotivationRegularized regressionIn generalThe lassoLoss reserving framework and notationApplication of lasso to loss reservingTest on simulated dataData setResultsFurther testing and developmentConclusion

Slide14

Application of lasso to loss reserving4 data sets with different underlying model structures (of ) consideredIn increasing order of stress to the model

Lasso applied to each datasetOnce the tuning constant selected, models are self-assemblingInterest in examination of the extent to which the model self-assembles the structures concealed in the dataAlso compare model forecasts with those from the chain ladder Slide15

OverviewMotivationRegularized regressionIn generalThe lassoLoss reserving framework and notationApplication of lasso to loss reservingTest on simulated dataData setResultsFurther testing and developmentConclusion

Slide16

Data set 1: set-upRecall

Observations (both past and future, whole square, not just triangle) simulated according to Assumed that the

are in constant dollar values (inflation corrected)Any calendar quarter trend represents superimposed inflation (“SI”)Upper triangle forms training data set

Lower triangle forms test data setAssume

follows

Hoerl

curve as function of

=0 (no payment year effect)

appears as in diagram

 Slide17

Data set 1: model selectionThe number of basis functions (regressors) was 2,380Model fitted to the training data set for a large number of values of tuning constant As

increases, number of non-zero parameters decreasesModel performance (for any given ) measured by:AICTraining error [sum of (actual-fitted)2/fitted values for training data set]Test error [sum of (actual-fitted)2/fitted values for test data set] (N.B. unobservable in practice)8-fold cross-validation error based on training data set Slide18

Data set 1: model selection (cont’d)Model selected to minimize CV errorOther approaches are possible to reduce model complexitySelected model contains 152 non-zero parametersSlide19

Data set 1: resultsAQ trackingTracking appears reasonable

DQ trackingSlide20

Data set 1: results (cont’d)Loss reserve by AQLasso forecast appears satisfactory

Distribution of total loss reserveLasso forecast tighter than chain ladder Slide21

Data set 2: set-up and model selectionRecall

Assume as for data set 1 as for data set 1

appears as in diagramModel includes:about 3,200 basis functionsExperimentation suggests inclusion of an

unpenalized constant SI term (

) in regression

84 non-zero parameters

 Slide22

Data set 2: resultsCQ tracking : at DQ 5Tracking again appears reasonable

CQ tracking : at DQ 15Slide23

Data set 2: results (cont’d)Loss reserve by AQ

Chain ladder now based on last 8 calendar quartersLasso CQ trends stopped at last diagonalHence lasso biased downward relative to CLTotal loss reserveChain ladder highly volatileSlide24

Data set 3: set-up and model selectionThis time

Assume as for data sets 1 & 2

as for data sets 1 & 2

as for data set 2

Interaction between AQ and DQ

For

,

increases by 0.3 for

Difficult to detect: affects only 6 cells in the triangle of 820 cells

Model includes:

about 3,200 basis functions

103 non-zero parameters

 

  Slide25

Data set 3: resultsDQ tracking at AQ 25DQ tracking surprisingly

accurateAQ tracking at DQ 25Though under-estimation of tail at higher AQsSlide26

Data set 3: results (cont’d)Loss reserve by AQChain ladder now based on last 8 calendar quartersCL and lasso both under-estimate

But CL under-estimation greaterTotal loss reserveChain ladder highly volatileSlide27

Data set 3: results (cont’d)

The AQxDQ interaction has been penalized like all other regressorsIn practice, one might be able to anticipate the changee.g. a legislated benefit change, taking effect in AQ 17In this case, one could apply no penalty to the interactionLoss reserve by AQInteraction unanticipated

Interaction anticipatedSlide28

Data set 4: set-up and model selectionThis time

Assume as for data sets 1-3 as for data

sets 1-3

as for data sets 2 & 3

(multiplier that varies SI with

)

Model includes:

about 3,200 basis functions

87 non-zero parameters

 

DQ

40Slide29

Data set 4: resultsDQ tracking at AQ 10

DQ tracking at AQ 25Slide30

Data set 4: results (cont’d)AQ tracking at DQ 10

AQ tracking at DQ 25Slide31

Data set 4: results (cont’d)CQ tracking at DQ 5

CQ tracking at DQ 15Slide32

Data set 4: results (cont’d)

Total loss reserveChain ladder comparable with lasso but with some outlying forecastsLoss reserve by AQ

Lasso more efficient predictor of individual accident quartersSlide33

OverviewMotivationRegularized regressionIn generalThe lassoLoss reserving framework and notationApplication of lasso to loss reservingTest on simulated dataData setResultsFurther testing and developmentConclusion

Slide34

Further testing and developmentExamination of additional scenariosParticularly those likely to stress the chain ladderDifferent basis functionse.g. Hoerl curve basis functions for DQ effectsConsideration of future SIHow well adapted to extrapolation is the lasso?Robustification Multi-line reserving (with dependencies)

Adaptive reservingHow might the lasso be adapted as a dynamic model?Slide35

OverviewMotivationRegularized regressionIn generalThe lassoLoss reserving framework and notationApplication of lasso to loss reservingTest on simulated dataData setResultsFurther testing and developmentConclusion

Slide36

ConclusionThe lasso appears promising as a platform for self-assembling modelsThe model calibration procedure follows a routine and is relatively quickPerhaps 30 minutes for routine calibration and examination of diagnosticsPerhaps an hour if one or two ad hoc changes require formulation and implementatione.g. superimposed inflation, legislative changeThe lasso appears to track eccentric features of the data reasonably wellIncluding in scenarios where the chain ladder has little hope of an accurate forecastSome further experimentation required before full confidence can be invested in it as an automated procedureSlide37

ReferencesMcGuire, G. 2007. “Individual Claim Modelling of CTP Data.” Institute of Actuaries of Australia XIth Accident Compensation Seminar, Melbourne, Australia. http://actuaries.asn.au/Library/6.a_ACS07_paper_McGuire_Individual%20claim%20modellingof%20CTP%20data.pdfTaylor, G., and G. McGuire. 2004. “Loss Reserving with GLMs: A Case Study.” Casualty Actuarial Society 2004 Discussion Paper Program, pp 327-392.Taylor, G., and G. McGuire. 2016. “

Stochastic Loss Reserving Using Generalized Linear Models”. CAS Monograph Series, Number 3. Casualty Actuarial Society, Arlington VA.Slide38

Questions?