Predictive Modeling - SoA Webinar - 2013 - Freakonometrics

Prediction using some (categorical) covariates. E.g. predicting someone's weight (Y) based on his/her sex (X1). Conditional parametric model assume that YX1 ...
7MB taille 2 téléchargements 196 vues
Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Perspectives of Predictive Modeling Arthur Charpentier [email protected] http ://freakonometrics.hypotheses.org/

(SOA Webcast, November 2013)

1

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Agenda • ◦ ◦ • ◦ ◦ • • ◦ ◦ ◦ •

Introduction to Predictive Modeling Prediction, best estimate, expected value and confidence interval Parametric versus nonparametric models Linear Models and (Ordinary) Least Squares From least squares to the Gaussian model Smoothing continuous covariates From Linear Models to G.L.M. Modeling a TRUE-FALSE variable The logistic regression R.O.C. curve Classification tree (and random forests) From individual to functional data

2

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

0.015

Prediction ? Best estimate ? E.g. predicting someone’s weight (Y )

ε is some unpredictable noise

Yb = y is our ‘best guess’...

●●

●●●



0.010

●●● ●●●●●● ●● ●● ●● ●● ●● ●●● ●● ●●●● ●●● ●● ●●●●●●● ●●●●●●●●●●●●● ●● ●●● ●

0.005

with E(ε) = 0 and Var(ε) = σ 2



0.000

Model : Yi = β0 + εi

−→ Density

Consider a sample {y1 , · · · , yn }

100

150

200

250

Weight (lbs.)

Predicting means estimating E(Y ).

ε

z }| {  Recall that E(Y ) = argmin{kY − ykL2 } = argmin{E [Y − y]2 } y∈R y∈R | {z } least squares

3

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Best estimate with some confidence 0.015

E.g. predicting someone’s weight (Y ) Give an interval [y− , y+ ] such that



●●

●●●



we can estimate the distribution of Y dF (x) F (y) = P(Y ≤ y) or f (y) = dx x=y

0.000

Confidence intervals can be derived if

0.005

Density

0.010

P(Y ∈ [y− , y+ ]) = 1 − α

●●● ●●●●●● ●● ●● ●● ●● ●● ●●● ●● ●●●● ●●● ●● ●●●●●●● ●●●●●●●●●●●●● ●● ●●● ●

100

150

200

250

Weight (lbs.)

(related to the idea of “quantifying uncertainty” in our prediction...) 4

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Parametric inference 0.015

E.g. predicting someone’s weight (Y )

−→ maximum likelihood techniques

0.005

90% 0.000

yb− = F −1 (α/2) b θ yb+ = F −1 (1 − α/2) b θ Standard estimation technique :

Density

2. Compute bound estimates

0.010

Assume that F ∈ F = {Fθ , θ ∈ Θ} b 1. Provide an estimate θ

100

150

200

250

Weight (lbs.)

n X b θ = argmax log fθ (yi ) θ∈Θ

i=1

|

{z

log likelihood

  explicit (analytical) expression for θ b  numerical optimization (Newton Raphson)

} 5

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Non-parametric inference

2. Compute bound estimates yb− = Fb−1 (α/2) yb+ = Fb−1 (1 − α/2)

●●●



0.010

●●

0.005

natural estimator for P(Y ≤ y)

●●● ●●●●●● ●● ●● ●● ●● ●● ●●● ●● ●●●● ●●● ●● ●●●●●●● ●●●●●●●●●●●●● ●● ●●● ●

90% 0.000

#{i such that yi ≤y}



Density

1. Empirical distribution function n X 1 1(yi ≤ y) Fb(y) = n i=1 | {z }

0.015

E.g. predicting someone’s weight (Y )

100

150

200

250

Weight (lbs.)

6

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Prediction using some covariates

Male

0.010

βF −βM

with E(ε) = 0 and Var(ε) = σ 2

0.000

βM

0.005

or Yi = β0 + β1 1(X1,i = F) + εi |{z} |{z}

Female

Density

based on his/her sex (X1 )   β + ε if X = F F i 1,i Model : Yi =  βH + εi if X1,i = M

0.015

E.g. predicting someone’s weight (Y )

100

150

200

250

Weight (lbs.)

7

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Prediction using some (categorical) covariates E.g. predicting someone’s weight (Y ) 0.015

based on his/her sex (X1 ) Conditional parametric model

0.005

Male

0.000

−→ our prediction will be

Female

Density

  F if X = F θF 1,i i.e. Yi ∼  Fθ if X1,i = M M

0.010

assume that Y |X1 = x1 ∼ Fθ(x1 )

100

conditional on the covariate

150

200

250

Weight (lbs.)

8

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Prediction using some (categorical) covariates

0.010 0.005

Density 0.005

Density

0.010

0.015

Prediction of Y when X1 = M

0.015

Prediction of Y when X1 = F

0.000

90%

0.000

90%

100

150

200 Weight (lbs.)

250

100

150

200

250

Weight (lbs.)

9

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Linear Models, and Ordinary Least Squares E.g. predicting someone’s weight (Y )

250



based on his/her height (X2 ) ● ●



Linear Model : Yi = β0 + β2 X2,i + εi 200

2

Conditional parametric model

● ● ●

150



100



● ● ●

● ● ● ●





● ●



● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●



● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●



● ●



● ● ●

● ●

● ●



● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●



● ● ● ● ●

● ● ●



● ●



● ●



● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ●

● ●

● ● ● ● ●



● ● ●

● ●

● ●



σ2

b = argmin −→ ordinary least squares, β





E.g. Gaussian Linear Model

β0 +β2 x2







Y |X2 = x2 ∼ N ( µ(x2 ) , σ 2 (x2 )) | {z } | {z }



● ●



assume that Y |X2 = x2 ∼ Fθ(x2 )



● ●



Weight (lbs.)

with E(ε) = 0 and Var(ε) = σ





5.0

( n X

5.5

)

6.0

6.5

Height (in.)

2 [Yi − X T β] i

i=1

b is also the M.L. estimator of β β 10

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Prediction using no covariates

250

● ●

● ● ● ●

● ●

● ●

● ●



● ●

● ● ● ● ● ●

Weight (lbs.)

● ● ● ● ● ● ● ●

● ●

● ● ● ● ● ● ● ●

● ●

150

● ● ● ●

● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ● ● ● ●

● ●

● ●

● ●

● ●

● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ●

● ● ● ●

● ●

● ● ● ● ● ●

● ● ● ●

● ● ● ●

● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ●

● ● ● ● ● ● ● ● ● ●

● ● ● ●

● ●

● ●

● ●

● ●

● ●

● ● ● ●

● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ● ● ● ●



● ● ● ●

● ● ● ● ● ●

● ● ● ●

● ●

● ●

● ● ● ●

● ●

● ●

● ●

● ●

● ●

● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ●● ● ● ● ● ●● ●●● ●●●● ● ● ● ●● ● ●● ● ● ●● ● ● ● ● ● ● ● ● ●● ● ●● ● ● ●● ●● ● ●● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ●● ●● ●●● ● ● ● ● ● ● ●● ● ●● ● ● ● ● ● ● ● ●● ● ●● ● ● ● ● ● ● ●● ● ● ● ●●●● ●●●● ● ● ● ●●●●●●●●●● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ●

● ●

90%

200

● ● ● ●

● ●

● ●

100

● ●











● ●

●●

● ●





● ●

5.0

5.5

6.0

6.5

Height (in.)

11

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Prediction using a categorical covariates E.g. predicting someone’s weight (Y ) based on his/her sex (X1 ) E.g. Gaussian linear model Y |X1 = M ∼ N (µM , σ 2 ) ●

b |X1 = M) = 1 E(Y nM

● ●

● ● ● ● ● ● ●●● ●●● ● ● ● ●● ● ● ● ●●● ●●● ●●●● ● ● ● ●● ● ●● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ●● ●● ●●●● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ●● ●● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●●● ●●●● ● ● ● ●●●●●●●●●● ● ● ● ● ● ●● ● ● ● ● ●●● ● ● ● ●







X

Yi = Yb (M)

i:X1,i =M









  b Y ∈ Y (M) ± u1−α/2 · σ b | {z }



●●

● ●





1.96

Remark In the linear model, Var(ε) = σ 2 does not depend on X1 . 12

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Prediction using a categorical covariates E.g. predicting someone’s weight (Y ) based on his/her sex (X1 ) E.g. Gaussian linear model Y |X1 = F ∼ N (µF , σ 2 ) ●

b |X1 = F) = 1 E(Y nF

● ●

● ● ● ● ● ● ●●● ●●● ● ● ● ●● ● ● ● ●●● ●●● ●●●● ● ● ● ●● ● ●● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ●● ●● ●●●● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ●● ●● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●●● ●●●● ● ● ● ●●●●●●●●●● ● ● ● ● ● ●● ● ● ● ● ●●● ● ● ● ●







X

Yi = Yb (F)

i:X1,i =F









  b Y ∈ Y (F) ± u1−α/2 · σ b | {z }



●●

● ●





1.96

Remark In the linear model, Var(ε) = σ 2 does not depend on X1 . 13

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Prediction using a continuous covariates

E.g. predicting someone’s weight (Y ) based on his/her height (X2 ) E.g. Gaussian linear model Y |X2 = x2 ∼ N (β0 + β1 x2 , σ 2 ) ●

b |X2 = x2 ) = βb0 + βb1 x2 = Yb (x2 ) E(Y   b b Y ∈ Y (x2 ) ± u1−α/2 · σ | {z }

● ●

● ● ● ● ● ● ● ●● ● ●●● ● ● ● ●● ● ● ● ● ●● ●●● ●●●● ● ● ● ●● ● ●● ● ● ●● ● ● ● ● ●● ● ● ●● ● ●● ● ● ●● ●● ● ●● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ●● ●● ●●● ● ● ● ● ● ● ●● ● ●● ● ● ● ● ● ● ● ●● ● ●● ● ● ● ● ● ● ●● ● ● ● ●●●● ●●●● ● ● ● ●●●●●●●●●● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ●











● ●

●●

● ●





1.96

14

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Improving our prediction ?

sex as

b (Empirical) residuals, εbi = Yi − X T iβ | {z } bi Y R2 or log-likelihood

covariate

height as covariate no covariates

−50

0

50

100

−50

0

50

100

parsimony principle ? −→ penalizing the likelihood with the number of covariates Akaike (AIC) or Schwarz (BIC) criteria

● ● ●



●●●



●●



15

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Relaxing the linear assumption in predictions Use of b-spline function basis to estimate µ(·) where µ(x) = E(Y |X = x)

250



250



● ●







● ● ●



● ● ● ●





● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ●

● ●







● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●



● ●



● ● ●

● ●



● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●



● ● ● ● ●

● ● ●



● ●



● ● ●

● ●



● ● ● ● ● ● ● ● ● ● ● ●

● ● ● ● ●





● ● ● ● ● ● ●



● ● ●







● ● ●

● ● ● ●

● ●



● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●



5.0





100

150

● ●

● ●





● ● ● ●

150

Weight (lbs.)







● ●





200

● ● ● ●





Weight (lbs.)

200

● ● ●

100



● ●

● ●







● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●



● ●



● ● ●

● ●

● ●



● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●



● ● ● ● ●

● ● ●



● ●



● ●



● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ●

● ●

● ● ● ● ●



● ● ●

● ●

● ●



5.5

6.0 Height (in.)

6.5

5.0

5.5

6.0

6.5

Height (in.)

16

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Relaxing the linear assumption in predictions

E.g. predicting someone’s weight (Y ) based on his/her height (X2 ) E.g. Gaussian linear model Y |X2 = x2 ∼ N (µ(x2 ), σ 2 ) b |X2 = x2 ) = µ E(Y b(x2 ) = Yb (x2 )   b Y ∈ Y (x2 ) ± u1−α/2 · σ b | {z }

● ● ●

● ● ● ● ● ● ● ●● ● ●●● ● ● ● ●● ● ● ● ● ●● ●●● ●●●● ● ● ● ●● ● ●● ● ● ● ● ● ●● ● ●● ● ● ●● ● ● ● ● ●● ●● ●●●● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ●● ●● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●● ● ●● ● ● ● ● ● ● ● ● ●● ● ● ● ●●●● ●●●● ● ● ● ●●●●●●●●●● ● ● ● ● ● ●● ● ● ● ● ●●● ● ● ● ●











● ●

●●

● ●





1.96

Gaussian model : E(Y |X = x) = µ(x) (e.g. xT β) and Var(Y |X = x) = σ 2 . 17

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Nonlinearities and missing covariates 250



E.g. predicting someone’s weight (Y ) ● ●



based on his/her height and sex





200

model mispecification





● ● ● ●

● ● ● ●



● ● ● ●



150



E.g. Gaussian linear model   β 0,F + β2,F X2,i + εi if X1,i = F Yi =  β0,M + β2,M X2,i + εi if X1,i = M





Weight (lbs.)

−→ nonlinearities can be related to



● ●

● ● ●

● ●



● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

100







● ● ●



● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ●



● ● ●

● ●

● ●



● ●





● ● ● ● ●

● ● ●



● ●



● ●



● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ●

● ●

● ● ● ● ●



● ● ●

● ●





● ●

● ●







5.0

5.5

6.0

6.

Height (in.)

b = argmin −→ local linear regression, β x

( n X

) 2 ωi (x) · [Yi − X T i β]

i=1 Tb

and set Yb (x) = x β x 18

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Local regression and smoothing techniques ● ●

250

250



● ●



● ●









● ●



● ●

● ● ● ● ●

● ● ● ● ● ●

● ●



100

● ●



● ●

● ●

● ● ● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ●

● ● ● ● ●

● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●



● ●

● ● ● ●



● ●

● ●







● ● ● ● ● ● ● ● ● ●

● ● ● ●





● ● ●



● ●

● ●

● ● ● ● ●



● ●



● ● ●



● ● ● ●

● ●

● ●

● ●

5.0

● ●

● ●

● ● ●

● ● ● ● ● ● ● ●

● ●

● ●

● ●

● ● ●

● ● ●

● ● ● ● ● ●

● ●



● ● ●

● ●

● ● ● ● ● ●





● ●

100

150

● ●

● ●



● ● ● ●





● ●

● ●

● ● ● ● ● ● ● ●

150

● ●

● ● ●



● ●

● ●



● ●

● ●

● ●

Weight (lbs.)

Weight (lbs.)

● ● ●

● ●

● ●



200

200





● ● ● ●

● ●



● ●

● ● ● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ●

● ● ● ●

● ●

● ● ● ● ● ●

● ●

● ● ● ●

● ●

● ●

● ● ● ● ● ● ● ● ● ●

● ● ● ●

● ●

● ●

● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ●

● ● ● ●

● ●

● ● ● ●

● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ● ● ● ●

● ● ● ●

● ● ● ● ● ●

● ● ● ●

● ●

● ●

● ● ● ●

● ●

● ●

● ●

● ●



● ●

● ●

● ●

5.5

6.0 Height (in.)

6.5

5.0

5.5

6.0

6.5

Height (in.)

19

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

k-nearest neighbours and smoothing techniques ● ●

250

250



● ●



● ●







● ●



● ●



● ●

● ● ●

● ● ●



● ●

● ●

● ●

● ● ● ●

● ● ● ●



150

● ●

● ● ●

● ● ●

● ●

● ● ●

● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●







● ● ● ● ●



● ●



● ● ●



● ●

● ●

● ●



● ● ●







● ●

● ● ● ● ●



● ●

● ●

● ●

● ● ● ● ● ●

● ●

● ●

● ● ● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ●

● ● ● ●

● ●

● ● ● ● ● ●

● ●

● ● ● ●

● ●

● ●

● ●

● ● ● ● ● ● ● ● ● ●

● ● ● ●

● ●

● ●

● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ●

● ● ● ●

● ● ● ●

● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ● ● ● ●

● ● ● ●

● ● ● ● ● ●

● ● ● ●

● ●

● ●

● ● ● ●

● ●

● ●

● ●

● ●



● ●

● ●

● ●



5.0

● ●

● ●

● ●

● ● ● ● ● ● ● ●

● ●

● ●

● ● ● ● ● ● ● ● ● ● ● ●

● ● ●

● ● ● ● ● ●

● ●

100

● ●



● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ●



● ● ●







● ●

100

● ● ● ●







● ●

● ● ● ●

150



● ●

● ●





● ●

● ●

Weight (lbs.)

Weight (lbs.)

● ● ●

● ●

● ●



200

200





● ● ● ●

● ●



5.5

6.0 Height (in.)

6.5

5.0

5.5

6.0

6.5

Height (in.)

20

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Kernel regression and smoothing techniques ● ●

250

250





● ● ● ●





● ●

● ●





● ●



● ●



● ●





● ●



● ●

● ●





● ●

● ●



● ● ● ● ● ●







● ●





● ●









● ●









● ● ●





● ●●















● ●









● ●

● ●



● ●



150













●●



















● ●

●● ●

















● ●

● ●











● ● ●

● ●







● ●





● ●





● ● ●



● ●





● ●

● ●

● ●

● ●

● ●

● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ●

● ● ● ●

● ●

● ● ● ● ● ●

● ●

● ● ● ●

● ●

● ●

● ● ● ● ● ● ● ● ● ●

● ● ● ●

● ●

● ●

● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ●

● ● ● ●

● ●

● ● ● ●

● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ● ● ● ●

● ● ● ●

● ● ● ● ● ●

● ● ● ●

● ●

● ●

● ● ● ●

● ●

● ●

● ●

● ●



● ●

● ●

● ●

● ●

5.0

● ●



● ●





















● ●













● ●

● ● ●













● ●

100













● ●







● ●

● ●





● ● ● ●



















● ●











● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ●













● ●

● ●







● ● ●

























● ● ● ● ● ● ● ●

● ●

● ●

● ● ● ● ● ● ● ● ● ●





● ● ● ●











● ● ● ● ● ●

● ●





















● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ●



● ● ●



100

● ●





150

● ●









●●●

● ●





● ●



● ●



● ●









● ●















● ● ● ●





● ●

Weight (lbs.)

Weight (lbs.)





● ● ●





● ●

● ●

200

200

● ●







5.5

6.0 Height (in.)

6.5

5.0

5.5

6.0

6.5

Height (in.)

21

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Multiple linear regression

10

E.g. predicting someone’s misperception

5

of his/her weight (Y )

0

based on his/her height and weight | {z } | {z }

−5

X2

−10 250

−→ linear model

200

ht

ig We

6.0

150

s.) (lb

E(Y |X2 , X3 ) = β0 + β2 X2 + β3 X3 Var(Y |X2 , X3 ) = σ

X3

5.5

2

g Hei

100

in.)

ht (

5.0

22

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Multiple non-linear regression

10

E.g. predicting someone’s misperception

5

of his/her weight (Y )

0

based on his/her height and weight | {z } | {z }

−5

X2

−→ non-linear model

250 200

ht

6.0

150

s.) (lb

Var(Y |X2 , X3 ) = σ

−10

ig We

E(Y |X2 , X3 ) = h(X2 , X3 )

X3

5.5

2

ght

100

Hei

)

(in.

5.0

23

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Away from the Gaussian model

Y is not necessarily Gaussian Y can be a counting variable, E.g. Poisson Y ∼ P(λ(x)) Y can be a FALSE-TRUE variable, ●

E.g. Binomial Y ∼ B(p(x)) −→ Generalized Linear Model









● ●

E.g. Y |X2 = x2 ∼ P e





(see next section) β0 +β2 x2

● ●

● ● ● ● ● ● ●●● ●● ● ● ● ● ●● ● ● ● ●●● ● ● ● ●●●● ● ● ● ●● ● ●● ● ● ●● ● ●● ● ●● ● ● ●● ● ● ● ● ●● ●● ●●●● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ●● ●● ●●● ● ● ● ● ● ● ●● ● ● ● ● ● ● ●● ●● ● ●● ● ● ● ● ● ● ● ●● ●●● ● ● ● ●● ● ●● ● ● ● ● ●●●●●●●●●● ● ● ● ● ●● ● ●● ● ● ●●● ● ● ● ●



●●

● ●





Remark With a Poisson model, E(Y |X2 = x2 ) = Var(Y |X2 = x2 ). 24

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Logistic regression

●●●

● ●



0.4

of his/her weight (Y )   1 if prediction > observed weight Yi =  0 if prediction ≤ observed weight

0.0

E.g. predicting someone’s misperception

●● ● ● ● ● ● ● ●● ●●● ●●●●● ●● ●●●

0.8

● ● ●

●● ● ● ● ● ●● ● ●● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ●● ● ●●

5.0

5.5

6.0

6.5

Height (in.)

1.0

Bernoulli variable, 0.8

●●

0.6 0.4

P(Y = y|X = x) = p(x)y (1 − p(x))1−y , where y ∈ {0, 1}



0.2

−→ logistic regression

0.0

P(Y = y) = py (1 − p)1−y , where y ∈ {0, 1}

●● ●●●●●●●●●●●●●●●●● ●●●● ●●● ●●● ● ● ●●



●● ● ●●●●●●●●●● ●●●●●●●●●●● ● ●●●●●●●●●● ●● ● ●

100

150

200







250

Weight (lbs.)

25

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Logistic regression

1.0

−→ logistic regression 0.8

y

P(Y = y|X = x) = p(x) (1 − p(x))

1−y

, 0.6

where y ∈ {0, 1}  P(Y = 1|X = x) T Odds ratio = exp x β P(Y = 0|X = x)

0.4

0.2

0.0 250



200

ht

ig We

6.0

150

s.) (lb

exp xT β E(Y |X = x) = 1 + exp (xT β) Estimation of β ? b (Newton - Raphson) −→ maximum likelihood β

5.5

g Hei

100

in.) ht (

5.0

26

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Smoothed logistic regression

●●●

● ●



0.4 0.0

GLMs are linear since  P(Y = 1|X = x) T = exp x β P(Y = 0|X = x)

●● ● ● ● ● ● ● ●● ●●● ●●●●● ●● ●●●

0.8

● ● ●

●● ● ● ● ● ●● ● ●● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ●● ● ●●

5.0

5.5

6.0

6.5



●●

0.8

●● ●●●●●●●●●●●●●●●●● ●●●● ●●● ●●● ● ● ●●

0.2 0.0

exp (h(x)) E(Y |X = x) = 1 + exp (h(x))

0.4

0.6

−→ smooth nonlinear function instead P(Y = 1|X = x) = exp (h(x)) P(Y = 0|X = x)

1.0

Height (in.)



●● ● ●●●●●●●●●● ●●●●●●●●●●● ● ●●●●●●●●●● ●● ● ●

100

150

200







250

Weight (lbs.)

27

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Smoothed logistic regression

1.0

−→ non linear logistic regression P(Y = 1|X = x) = exp (h(x)) P(Y = 0|X = x)

0.8

0.6

0.4

0.2

E(Y|X = x) =

exp (h(x)) 1 + exp (h(x))

0.0 250 200

ht

ig We s.) (lb

Remark we do not predict Y here,

6.0

150 5.5

ght Hei

100

but E(Y |X = x).

) (in.

5.0

28

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Predictive modeling for a {0, 1} variable 1.0

What is a good {0, 1}-model ? Observed

Observed

−→ decision theory

0.2

Y=1



0.6



Y=0

^ Y=1

● ● ● ● ● ● ●● ● ● ●● ● ● ● ● ● ● ● ●● ● ● ● ● ●● ● ●● ● ● ● ● ● ●● ● ●●●●● ●● ● ● ●●●●

0.4

^ Y=0



0.8

Predicted

0.0

  if P(Y |X = x) ≤ s, then Yb = 0  if P(Y |X = x) > s, then Yb = 1

● ●●●● ●● ● ●●● ●● ● ● ●● ● ● ●● ● ● ● ● ● ● ●● ● ●● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ●● ● ● ●●●● ●● ● ● ●●

0.0

0.2

0.4



0.6



0.8

1.0

0.8

1.0

1.0

Predicted

error

Y =1

error

fine

0.6

fine

0.4

Y =0



0.2

Yb = 1 ●

0.0

Yb = 0

True Positive Rate

0.8



0.0

0.2

0.4

0.6

False Positive Rate

29

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Observed

Observed

Y=0



0.0

0.2

Y=1



● ●●●● ●● ● ●● ● ● ●● ● ● ● ● ● ● ● ● ●● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ●● ● ● ●●●● ●● ● ● ●●

0.0

0.2

0.4



0.6



0.8

1.0

1.0

Predicted





0.6 0.4 0.0

0.2

True Positive Rate

0.8



R.O.C. curve is {F P (s), T P (s)), s ∈ (0, 1)}

^ Y=1

● ● ● ● ● ● ●● ● ● ●● ● ● ● ● ● ● ● ●● ● ● ● ● ● ●● ● ● ● ● ● ●● ● ●●●●● ● ● ●●●●

0.6

^ Y=0



0.8

Predicted

0.4

True positive rate T P (s) = P(Yb (s) = 1|Y = 1) nYb =1,Y =1 = nY =1 False positive rate F P (s) = P(Yb (s) = 1|Y = 0) nYb =1,Y =1 = nY =1

1.0

R.O.C. curve

0.0

0.2

0.4

0.6

0.8

1.0

False Positive Rate

(see also model gain curve) 30

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Classification tree (CART)

250



● ●



If Y is a TRUE-FALSE variable 200

E(Y |X = x) = pj if x ∈ Aj









● ● ● ●



● ● ● ●

150

● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

regions of the X-space.



100

● ●











● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ● ●



● ● ●

● ●





● ●

● ● ● ● ●



● ● ●



● ● ● ● ● ● ● ●

● ●





● ●



● ● ● ● ● ●

● ● ●

● ●

where A1 , · · · , Ak are disjoint





Weight (lbs.)

prediction is a classification problem.





● ●

● ●





● ● ●



● ●

● ●

● ●





5.0

5.5

6.0

6.5

Height (in.)

31

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Classification tree (CART) −→ iterative process



250

P(Y=1) = 38.2% P(Y=0) = 61.8%

Step 1. Find two subset of indices either A1 = {i, X1,i < s} or A1 = {i, X2,i < s}





● ● ●

200

● ●

● ● ●



150

● ●



100



& maximize heterogeneity between subsets



● ●





● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

maximize homogeneity within subsets





● ● ● ●





and A2 = {i, X2,i > s}





Weight (lbs.)

and A2 = {i, X1,i > s}

● ●









● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ● ●



● ● ●

● ●







● ●



● ● ●



● ● ● ● ● ● ● ●

● ●

● ● ● ●





● ●







● ● ●

● ●

● ●





● ● ●



● ●

● ●

● ●



P(Y=1) = 25% P(Y=0) = 75%



5.0

5.5

6.0

6.5

Height (in.)

32

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Classification tree (CART)

250



● ●







0.317



● ● ● ●



● ● ● ●



150

● ● ●

y∈{0,1}

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ●









● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ● ●



● ● ●

● ●





● ● ●

● ● ● ● ●



● ● ●



● ● ● ● ● ● ● ●

● ●





● ●



● ● ● ● ● ●

● ● ●



100

x∈{A1 ,A2 }





0.523

Weight (lbs.)

E.g. Gini index   X nx X nx,y nx,y − · 1− n nx nx

200

Need an impurity criteria





● ●

● ●





● ● ●



● ●

● ●

● ●



0.273



5.0

5.5

6.0

6.5

Height (in.)

33

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Classification tree (CART)

250



Step k. Given partition A1 , · · · , Ak

● ●



0.472





● ● ●



● ● ● ●



● ● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●



100

● ●









● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ● ●



● ● ●

● ●





● ●

● ● ● ● ●



● ● ●



● ● ● ● ● ● ● ●

● ●





● ●



● ● ● ● ● ●

● ● ●



maximize homogeneity within subsets & maximize heterogeneity between subsets



● ●

150

or according to X2



0.621 ●

0.444

Weight (lbs.)

either according to X1

200

find which subset Aj will be divided,





● ●

● ●



● ● 0.111

● ● ● ●



● ●

● ●



0.273



5.0

5.5

6.0

6.5

Height (in.)

34

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

Visualizing classification trees (CART) Y < 124.561 |

Y < 124.561 |

X < 5.39698

0.3429

X < 5.69226

0.1500

X < 5.69226 0.2727

X < 5.52822

0.4444

0.5231

0.3175

Y < 162.04

0.6207

0.1111

0.4722

35

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

From trees to forests ● ●

250

Problem CART tree are not robust −→ boosting and bagging

● ● ●

● ●

● ●

● ● ●

● ● ● ●

● ●

● ●

● ● ● ● ● ●





● ● ● ●

● ●

0.808

Then aggregate all the trees

● ●

● 0.520 ● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ●

100

● ●



● ●

● ●

● ●



● ●





● ● ● ● ● ●

150

repeat this resampling strategy



● ● ● ●

● ●

Weight (lbs.)

and generate a classification tree

0.467

200

use bootstrap : resample in the data

0.576





● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ●

● ●

● ● ●



● ● ● ● ●

● ●

● ●

● ●

● ●

● ●

● ● ●

● ●

● ● ● ●

● ● ● ● ● ● ● ● ●



● ● ● ● ● ●



● ● ●

● ●

● ●

● ●





● ● ●

● ● ● ● ● ● ● ●



● ●

● ●

● ● ● ●

● ●



● ●

● 0.040

● ● ● ● ● ● ● ●

● ●

● ●

● ●

● ● ● ●

● ●



0.217

● ●

5.0

5.5

6.0

6.5

Height (in.)

36

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

10

A short word on functional data ●

● ●



● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

−15

● ● ● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

Individual data, {Yi , (X1,i , · · · , Xk,i , · · · )}

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●











● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●



● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ●



5 0 −5 −10



● ● ● ● ● ● ● ● ●



−20

Temperature (in °Celsius)

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●







● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●

● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●







● ●



5

10

15

Week of winter (from Dec. 1st till March 31st)

Functional data, {Y i = (Yi,1 , · · · , Yi,t , · · · )} ● ●● ●

0 −10 −20 −30

Temperature (in °Celsius)

10

E.g. Winter temperature, in Montréal, QC

● ● ●● ● ●● ● ● ●● ● ● ● ●● ● ● ● ● ●● ●● ● ●● ● ● ● ● ● ● ● ●● ●● ● ● ●●● ●● ●● ●● ●● ●● ●● ●● ●● ● ● ● ●● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●●●●● ● ●● ● ●●● ● ● ● ● ● ● ●● ●● ● ● ● ●● ●●● ● ●● ● ●●●● ●●● ● ● ● ● ●● ●●● ● ● ● ● ●● ● ●●●●● ● ● ●●● ●● ●● ●●●●● ● ●●●● ●●● ● ●● ● ●● ● ●● ●●● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ●●●● ●● ● ● ● ●● ● ●● ●●●●●● ● ●● ●● ● ●●●●● ● ● ● ● ●● ●●●●●●●● ● ● ● ●●● ●● ● ●● ●●● ●● ● ● ● ● ●● ●●● ●● ●● ● ● ●● ●● ● ● ● ● ● ● ●● ● ● ● ●● ● ●● ● ● ● ● ●●● ● ●●● ● ● ●●●● ● ●●●●● ● ●● ●● ● ●●● ●●● ● ●●● ●● ● ●● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ●● ●●● ●● ●●● ●● ●●●● ● ●● ● ●● ● ● ●●●●● ●● ●● ● ●● ●● ● ● ●● ●● ●● ●● ● ●● ●●● ● ●●● ● ● ● ●●●●● ●● ●● ●● ●●●●● ●●● ●● ●● ●● ●●●● ●● ● ●●● ● ●●● ● ●● ● ● ●●●●●● ●● ●● ●● ●● ●● ● ●● ●● ●● ●● ● ●●● ●● ●●●● ●●●●● ●● ●●● ●● ●● ●● ● ●● ● ●● ●● ●●● ● ● ● ● ● ●●● ● ●● ●● ●● ● ●●●● ● ●●● ● ●● ●●● ● ●● ●●● ● ● ●● ●●● ● ●●● ●● ●●●●● ● ● ●● ● ●● ● ●●● ●● ● ●● ●● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ●●● ●●●● ●●●●●● ● ● ● ●●● ●●● ●●● ●●● ●●●●●● ●●● ●●●●●● ●●●●●● ●● ●● ●●●●● ●●●● ●●●● ●●●●● ●● ●● ●●●● ●●●●●●● ●● ●● ●● ●● ●● ●● ● ● ●●● ●●●●●● ● ●● ● ●● ●● ● ● ● ●●● ● ●● ●● ●● ● ● ●● ●● ●● ●● ●● ●●● ● ● ● ●● ●● ●● ●● ● ● ● ●●● ●● ● ●●● ●●●●● ● ●● ●●● ●● ●● ●● ●● ●● ●●● ● ●● ● ●●● ●● ●● ● ● ● ●● ●● ● ●● ●● ● ●● ● ● ●● ●●●● ●● ● ●● ●●●●●● ●● ●●● ●● ● ● ● ●● ●● ●●●● ●● ●● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●●●●●●● ● ●●● ● ●● ● ●● ●● ● ●●●● ●● ●● ●●● ●●● ●●● ●● ●● ●● ● ● ●●●●● ●●●●●●●●● ●●●● ●● ● ● ●●● ●●●● ●●● ●●● ●●● ●●● ● ●● ●●● ● ●● ●●●●● ●● ●●●● ●● ●● ● ●●● ●● ●● ●● ● ● ●● ●● ● ●● ●● ● ●● ●●●●●● ●●●●● ●● ●●● ● ●● ● ●●● ●●●●● ●● ●● ● ● ●● ●● ●● ●●● ●● ●● ● ●●●● ●● ●● ● ●● ●●●●● ●●● ●●●● ●● ●●● ● ●●● ● ●● ●● ●● ● ● ● ● ● ● ●● ●● ●● ●●● ● ●● ●● ●● ●●●● ●● ● ● ● ●● ●● ● ●● ●●● ●● ●● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●●● ●●●●●●●●●●●●●● ●●●●●●● ●●● ●●● ●●●● ●●●● ●●●● ●●●●●●● ● ●●● ●●●●●● ●●●●● ●● ●● ●● ●●● ● ●● ● ●● ●● ●● ●● ● ●● ●●●● ●●● ●●● ●● ●●●● ●● ● ●● ●●●● ●● ● ● ● ●● ●●● ● ●● ●● ●● ●● ● ●● ● ● ● ● ● ●●● ●● ●● ●●● ●● ● ●●●● ●● ●●● ●●● ●●● ●● ●● ● ●● ●● ●●● ●● ●● ●●● ●● ●● ● ●● ●●● ●● ● ● ● ●●●●● ●● ● ●● ● ●● ● ●●● ●● ● ● ●●●●●● ●●● ●● ● ●● ●● ● ● ●● ●● ● ● ●● ●● ● ● ●●●● ●● ●● ●● ●● ● ●●● ●● ●● ●●● ●● ●●● ●● ● ●● ●● ●● ● ●●● ● ●● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●● ●●●● ● ●● ●●●●● ● ●● ● ●●●●●●●● ●●● ●● ● ●●● ●● ●●●● ●● ●●●● ●● ●● ●● ●● ● ● ●●● ●●●● ●●● ●● ●●●● ●●● ● ●●●● ●● ● ●● ●●●● ● ● ● ●● ●●● ●● ●● ●● ●●● ●●● ●● ●●● ●● ●●● ●● ●● ●● ●● ●● ●● ●● ● ●● ● ● ● ●● ●● ●● ●● ●● ●● ●●● ● ●● ●● ● ● ● ● ●●● ● ●●●●●● ●● ● ●●●● ●●●● ● ● ● ● ● ● ●● ●● ● ● ●●●●●● ● ● ●●●● ●● ● ●● ●●●● ● ●● ●● ● ●● ● ● ●● ● ●● ●●●● ● ● ●● ● ●● ● ●● ●●● ● ●●● ●● ●● ●● ● ●● ● ●● ●● ●● ●● ●● ●● ●● ●● ● ● ● ●● ●● ● ●● ●● ● ●● ● ●● ●● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●● ●●● ●●● ●●● ●● ●● ●●●● ● ●● ● ●●● ●●●●●●●●● ● ●●●● ●●●●● ●●● ●● ●●● ●● ●●● ●● ●● ●● ●● ● ●●● ●●● ●● ●● ●●● ●● ●● ●●●● ●● ● ●● ● ● ● ● ● ● ●●●●●● ●● ●● ●● ●●● ●● ●● ●● ●●●● ● ● ● ●●●●●●●●● ● ●●● ●● ●● ●● ●● ●●● ●● ●● ● ●● ●● ●●● ●● ●● ● ● ● ● ●● ● ●● ●● ●●●●● ● ●● ●● ● ●● ● ● ●● ●● ● ● ● ●●● ● ● ●●●● ●● ●● ● ●● ● ● ●● ● ●● ●● ●●● ●●● ● ● ●● ●● ●● ● ●● ●●●● ●● ● ● ●●●● ●● ● ● ●● ●● ● ● ● ●● ●●● ● ●●● ●● ●● ●●●● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●●●● ●● ●● ●●●● ●● ●● ●●●●● ● ●● ● ●●●● ● ● ●●●● ● ●● ●● ● ●●●● ●●●● ●● ● ●●●● ●●●●● ● ●●● ●●● ●● ●●● ●● ●●● ●● ●● ●● ● ●● ●● ●●● ● ●● ●●● ● ●● ● ●●● ● ●● ●●● ● ● ●● ●● ●● ● ● ●●● ● ● ●●● ●●●●●●●● ●● ● ●●●● ● ●● ● ●● ●● ● ● ●● ●●● ● ● ● ● ●● ●●● ●●● ●●● ● ●● ● ●●● ●● ●● ●●●●● ● ● ● ● ●● ●● ●● ● ● ● ●●● ●● ●● ●● ●● ● ● ●● ● ● ●● ●●● ●● ●● ●● ●● ● ●● ● ●●●● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ●● ●● ● ●●●●●●●● ●●● ●●● ●●●●●● ● ●● ● ● ● ● ●●● ●● ●●● ●●● ●●● ● ●●●●● ●●●●● ● ●● ● ●● ●● ●●●●●●●● ●●● ●●●● ● ● ● ●● ●●● ●● ●●●●●●●●● ● ●●● ●●●●●●●●● ● ● ●● ● ● ●●● ●● ●● ●● ●●● ● ● ● ●● ●●● ● ●● ●●●● ●●● ● ●● ●● ● ●●●● ●●●● ● ● ●● ● ●●●●●● ●●●● ● ● ● ● ● ●● ● ● ● ● ●●●● ● ●● ● ●● ●● ● ● ●●● ●● ● ●●●● ● ●●● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●●●● ●● ●●●● ● ●● ●● ● ●● ●●●● ●● ●●●● ●● ●● ● ●● ● ●●● ●● ● ●●● ●●● ●●● ● ● ●● ●● ● ● ●●●●● ● ●● ●● ● ●●● ● ● ● ● ●●● ●●● ● ●● ● ●● ●●●● ●● ● ● ●● ●●● ●●● ●●● ●● ● ●● ●● ●●● ●●● ●● ●●●●● ● ●● ● ● ● ● ● ● ●●● ● ● ●● ●● ●●●●●● ●●●●● ● ● ● ● ● ●● ● ●● ●● ● ●● ●● ●● ● ●● ● ● ● ●●●● ●● ●●●●● ● ●●●●●●●●● ● ●●●●● ●● ●●● ● ●● ● ● ●●●●● ●●●● ●● ● ● ●● ●●● ● ●● ● ●● ● ●●● ●●● ●● ●● ●● ●● ●●●● ●● ● ●● ●● ●● ●● ●●●● ●● ●●●● ●● ● ●●● ●● ●● ●● ●●● ●● ● ●●● ●● ● ● ●●● ●●●● ● ●● ●● ●●●●●●● ●● ● ●●●● ● ● ●● ●●● ●●● ●● ●● ●●● ●● ● ●●●● ●● ●● ● ● ● ●● ● ●● ● ● ● ● ● ● ●●● ● ● ● ●●● ● ● ● ●● ●●●● ●●● ● ● ●● ● ● ●● ●●●●● ●● ● ● ●●●●● ● ●●● ●●● ●●●●● ●● ● ●●●● ●● ● ● ●● ● ●● ●●●●●● ●● ●●● ● ●● ● ●●● ●●●●● ●●● ● ●● ●●●●● ●● ●● ●●● ●● ●●● ●● ● ● ● ●● ● ●●●● ●● ●● ● ● ●● ●●●● ● ● ● ● ●● ● ● ●● ●● ● ●●● ● ●●●● ●●● ● ● ●● ●● ●● ● ●● ● ●● ● ●● ●●●● ●●●● ●● ● ● ●●●● ●● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ●● ●●● ● ●●● ●●● ●●●● ● ●●●● ●●● ●●●●● ● ●● ● ●●● ●● ●●● ●●●●●●● ●●●● ●●●●● ●● ● ● ● ● ●● ● ●● ● ● ●●● ●● ●●● ● ●●● ●●● ●● ● ● ● ●● ●● ●●●●● ● ●● ● ● ●● ●● ●●●●● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ●● ● ●● ● ●●● ●●●● ● ● ● ● ●●● ●●●● ● ● ●● ●●●●● ● ● ● ●● ●●● ● ●● ●● ● ●● ●● ●●●● ●● ● ●● ● ● ●● ● ● ● ● ●● ● ● ●● ● ● ●●●● ●● ● ● ● ● ●● ● ● ● ●● ●●● ●● ● ● ● ● ●● ●● ● ● ● ● ● ● ● ● ●● ● ● ● ●●● ● ● ● ●●● ● ● ●●● ● ●● ● ● ● ● ● ● ● ● ● ● ● ●●

0

20

40

60

80

100

120

Day of winter (from Dec. 1st till March 31st)

37

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

A short word on functional data Daily dataset

2001

1982 ●

1974 1999 2007

0

PCA Comp. 1

40

50



● ●



2006

2009





1957 ●

2011

1997

20





−50



1986

2005 ●

1961

●● ●● ●● ●● ● ●● ●● ● ●●● ●● ● ● ●● ●● ●● ● ●● ●●● ●● ●● ● ● ●● ●● ● ●●●●●● ●●●●● ● ●● ●● ●●● ●●●●●●●●●● ● ●● ● ● ● ● ● ● ● ●●● ● ●●●●● ● ● ●●● ● ●●●●●●●●●●●● ●● ●● ● ●● ●● ● ● ●● ● ● ● ●● ● ● ●● ●● ●●●●●●●●●●● ● ●● ●● ●● ● ● ●● ●●●● ●●●●● ● ● ●● ● ●● ●● ●● ● ●● ●● ● ●●● ● ●● ● ● ●● ● ● ● ● ●●● ●● ● ● ● ● ● ● ●● ● ●●● ● ●●● ● ● ●●●● ● ●●●●● ●●● ● ● ● ●● ● ● ● ● ● ●● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●●● ● ● ●● ● ●● ● ● ● ● ●● ● ●● ● ● ● ●● ●● ●● ● ●●● ●● ● ●● ● ●●●●● ●● ● ●●● ●● ● ●● ●● ● ●● ●● ● ●● ●● ● ●●●● ● ●●● ● ● ●● ● ● ● ●●●●●●● ●● ●● ●● ●● ●●●●● ●●● ●● ● ● ● ● ●● ● ●● ●●● ● ● ● ●●● ● ● ●● ● ●● ● ● ● ●●● ●●● ●● ● ● ●● ●●●● ●●●● ● ●● ● ●● ● ●● ●● ●●● ●●● ●● ● ●● ●●● ● ●● ● ●●● ● ● ●●●● ● ●● ● ● ●● ● ●●● ●●●● ●● ●● ● ●● ●● ●● ●●● ●●● ●● ● ● ● ● ●● ● ● ●● ● ●● ●● ● ● ●● ● ●●● ●● ● ● ●●● ● ● ● ● ●● ●● ●● ●●●●●● ● ● ● ●● ● ● ●● ●● ●● ●●● ●●● ●●●● ●●● ● ● ● ●● ● ●●● ●● ●● ● ● ●● ● ●●●● ●● ● ●● ●● ● ● ●● ●● ● ● ●●● ●● ● ●● ● ●● ●● ●● ●● ●● ●●● ●● ● ● ● ● ● ●●● ● ●●● ● ● ● ● ● ● ●● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ●●●●● ●●● ● ●●● ●●●●● ● ● ●●●●●● ● ● ● ●● ●●● ● ●● ●● ●● ●● ●●● ● ● ● ● ● ● ●● ● ●●● ●● ● ●●● ●● ● ● ● ●●● ● ●● ● ● ●●● ● ● ● ●● ●●● ● ●● ● ●● ● ● ● ●●● ● ●●● ●●● ● ●● ● ●● ●● ●● ● ● ● ● ● ● ● ●● ●● ● ● ● ● ●● ● ●● ●● ● ● ● ●●● ●● ● ●●●●●●● ● ●● ● ● ●●● ● ● ● ●● ●●●● ●● ●● ●●● ● ● ●●● ●●●●● ●●● ● ●● ●● ● ●●● ●●●● ● ● ●● ● ● ● ● ●●● ●● ● ● ● ●● ●● ● ● ● ● ● ●● ●●● ●● ●● ●●● ● ● ● ●● ●●●● ●● ●● ●●● ●● ● ● ● ●●● ●● ● ●●● ●● ●● ● ●●●● ● ●●● ● ●●● ● ●● ● ● ● ● ●● ● ● ●● ● ● ●● ● ●● ●● ● ● ● ● ●● ●● ● ●●●●● ●●● ● ●●●●● ● ● ● ● ●● ●● ●●● ● ● ● ● ●● ●● ●●●● ●● ● ● ● ● ● ● ● ●● ● ● ● ● ● ●● ● ●● ● ●● ●● ● ●● ● ● ● ●● ● ● ●● ●●● ●● ●● ●●● ●● ● ●●● ●●● ●● ●● ● ●● ● ● ● ● ●●● ● ● ●● ● ● ●● ● ● ● ● ● ● ● ● ●●● ●●● ● ●● ●● ● ● ●● ● ● ●● ● ● ● ●● ● ●● ● ● ●●● ● ●● ● ● ● ● ● ● ● ● ● ●● ● ●● ●● ●● ● ● ●●● ● ●● ●●● ●● ●● ● ● ● ●● ●●● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ●● ● ● ● ●●● ● ● ●●● ● ● ● ● ● ●● ●●● ● ●● ● ● ●● ● ●● ● ● ● ●● ● ● ● ● ● ●●● ●● ●● ●● ● ●● ●● ●● ●●● ● ● ● ● ● ●● ● ● ● ●●● ●● ● ● ● ● ● ● ● ● ●● ● ● ●● ● ●●●● ● ●● ● ● ●● ● ●● ●●● ●● ●● ●● ●●● ● ● ● ●● ● ●● ●● ●● ● ●● ● ● ●● ●●● ●● ●● ●● ● ●● ● ● ●● ●● ●● ● ● ● ● ●●● ● ●●●●●●● ●● ● ●● ●● ● ● ● ●●● ●● ●● ●●●● ●● ● ●● ● ●● ● ● ●● ● ● ●● ● ● ●● ●● ● ● ●● ● ●● ● ● ● ●● ● ●●● ●●●● ● ●● ●● ●● ● ● ● ● ● ●● ●● ●● ●● ●● ● ●● ● ● ●● ●● ● ● ●● ●●●● ● ● ●● ● ● ● ● ●● ● ●●● ●● ● ●●● ●● ●● ● ● ● ● ● ● ●● ●● ●● ● ●● ● ●● ● ●● ● ●● ● ●● ●● ●● ● ●●● ●●●● ●●●● ● ●● ● ●● ● ●● ● ● ● ●●● ●●● ● ●●●● ●●● ●●● ● ●● ● ● ● ● ●● ● ●● ●● ● ●● ● ● ●● ●●●●● ●● ● ● ●●●● ●● ●● ● ●● ●● ● ● ● ● ● ● ●● ●● ●● ● ● ● ● ●● ● ● ● ● ● ●●● ●● ● ●● ● ● ●●● ● ● ●● ● ● ● ● ● ● ● ●● ● ●● ●● ● ● ●● ● ●● ●● ● ● ● ●● ● ● ●● ● ●● ● ●●● ●●● ●●●● ● ● ● ● ●● ●● ●● ● ● ● ●● ● ● ●● ●●● ● ●●●●● ● ●● ●● ● ● ● ● ●● ●●●●● ●● ●●● ●●● ● ● ● ● ● ● ● ●● ●● ● ●●● ●● ●● ●●● ● ● ● ●●● ●● ● ●●● ●● ●● ● ●● ●● ●● ●●● ● ● ●● ●● ●● ●●●●● ● ●● ●●●●●● ● ●●●●● ●●● ●●● ● ● ● ●● ● ● ● ● ● ● ●● ● ●● ● ● ● ● ●● ●● ● ●● ● ● ● ● ●●● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●● ● ●● ● ●●● ● ●●● ●●● ●● ● ●● ● ● ● ● ● ●● ●●●●●●●●●● ●● ● ● ● ●● ● ● ● ● ●●● ● ● ● ●● ●●● ● ●● ● ● ●●●● ●● ●● ● ● ● ● ●●● ● ● ● ● ● ●● ● ● ● ●● ●● ● ●● ● ●●●● ● ●●●● ●● ●● ●● ●● ●● ● ●● ●● ● ●● ● ● ● ●● ● ● ● ● ●● ●● ● ● ●● ●●● ●● ●● ● ●● ●● ● ●● ● ● ●● ●● ● ●● ●● ●● ●● ●● ●● ● ●●● ●● ● ● ●●● ● ● ●●● ● ● ●● ●● ● ●● ● ●● ● ●● ●● ●● ●● ●● ●●●● ●● ● ● ● ● ●●●● ●● ● ● ●● ● ●● ●●● ●●● ● ● ● ● ● ● ●● ● ● ●● ● ●●● ● ● ● ●● ● ●● ●● ● ● ● ●● ●● ●● ●● ●● ● ●● ● ● ●● ●● ● ● ● ●● ● ● ● ● ● ●● ● ● ● ● ●●●●●● ● ●● ● ●● ●● ● ● ●● ●● ● ●● ● ● ● ●● ● ● ●● ● ● ●● ●● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ●● ●● ● ● ● ●● ● ● ● ● ● ● ● ●● ● ● ● ●● ● ● ●● ●

JANUARY

DECEMBER

FEBRUARY

MARCH



1979 ●

Comp.2

0









1994

1996

1966 1964 1973 2010 1955

1978



20

40

60

80

100

120

Day of winter (from Dec. 1st till March 31st)

1987 ●

●●

1953 ●



1962 1954

1991







1989 ●

1960

0



2000 ●

1988 1963 1965 1959 1990





30





1992 1971

2004









2002

1998 ●

1977

1972 1985

20







1984





2008 ●

1968 ●

10

1956 1995 ●



2003 ●

1969 1975

1983





1976





1993 ●

1980 1967

−20





−20

−10

0

10 Comp.1

20

30

40

50

−30

−20

1981

0





−10

1958 1970

PCA Comp. 2



● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●● ● ●● ● ●● ● ● ● ●● ● ● ● ● ● ●● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ●● ●●● ● ●● ● ● ● ●● ● ● ● ● ●● ●● ●● ●● ● ● ● ●● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ●● ●● ● ● ●● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ●● ● ● ● ●● ● ● ● ● ● ●● ● ● ●● ● ● ● ●● ●● ●● ●● ● ●● ● ●●● ● ● ● ●●● ● ● ● ●●● ● ● ●● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●● ● ● ●● ● ●●● ● ● ● ● ●● ● ● ● ● ●● ●●● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ●● ● ●●● ●●● ●● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ●● ● ●● ● ●● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ●● ● ● ● ● ●● ● ● ● ● ●●● ● ● ● ● ● ● ● ●● ● ● ● ● ●● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●● ● ● ● ● ●●● ● ● ●● ● ●●● ●● ● ● ● ● ●● ● ● ● ● ●● ● ●● ● ● ● ● ● ● ● ● ●● ● ● ●● ● ●● ● ● ● ● ● ● ● ●● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●●● ● ● ● ●● ● ●● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ●● ● ● ● ● ● ●● ● ●●●● ● ● ● ●● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●●● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ●● ● ● ● ● ● ● ● ● ●● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●●● ● ● ● ● ●● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ●● ●● ●● ● ● ●● ●● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ●●● ●● ● ● ● ● ●● ● ● ● ● ●● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ●● ● ● ●● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●● ● ● ● ●● ●● ● ● ● ● ●● ● ● ● ● ● ● ●●● ● ●● ● ● ●● ● ● ● ● ●● ● ● ● ● ● ● ● ● ●● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●● ● ● ● ● ● ● ● ●● ● ● ● ● ● ●● ● ● ●● ● ● ● ●● ● ● ● ● ●● ● ● ● ●● ●● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●● ●● ● ● ● ● ● ● ● ● ● ●● ● ● ●● ● ●● ● ● ●● ● ●● ● ● ● ● ● ● ● ● ● ● ●●● ●● ●● ●● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ●●● ● ● ● ● ●● ●● ●● ● ● ●● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ●● ● ●● ●● ● ●● ● ● ● ● ● ● ● ●● ●●● ● ● ● ● ● ● ●● ●● ● ●● ●● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ●● ● ● ●●● ● ● ● ● ●● ● ●● ● ● ●● ● ● ●● ● ●●● ● ●● ● ●● ● ● ● ● ● ● ● ● ● ●● ● ● ● ●● ● ● ● ●● ● ● ●● ● ● ● ● ●● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●● ● ● ● ● ● ● ● ● ●● ●● ● ● ● ● ● ●● ● ● ●● ● ●● ● ● ● ● ●● ● ●● ● ●● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●●● ●● ●● ●● ● ● ● ● ●● ●● ● ● ● ●●● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ●● ● ● ●● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●● ● ●● ● ● ● ● ● ●● ● ● ●● ● ● ●● ●● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ●● ● ●● ● ● ● ● ●● ● ● ● ● ● ●● ● ●● ● ● ● ●● ● ●● ●● ● ●● ● ● ● ● ● ● ●● ● ● ●● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●●●● ● ●● ● ● ● ●● ● ● ● ● ●● ● ● ● ● ●● ● ●● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ●● ● ● ●● ● ● ● ● ● ● ●● ● ● ● ● ● ●

0

20

40

60

80

100

120

Day of winter (from Dec. 1st till March 31st)

38

Arthur CHARPENTIER - Predictive Modeling - SoA Webinar, 2013

To go further...

forthcoming book entitled Computational Actuarial Science with R

39