2 Simple, parametric distributions for frequency and severity data
2.1 The exponential distribution
In this tutorial you will simulate data from an exponential distribution with density
\[ f(x) = \lambda \cdot e^{-\lambda \cdot x}.\]
You will then explore and visualize these data. Finally, you will fit an exponential distribution to the data using Maximum Likelihood Estimation (MLE) (as discussed in Chapter 13 of the Loss Models book).
2.1.1 Simulating data
Use the R function rexp
to simulate 10 000
observations from an exponential distribution with mean \(5\).
- Create a variable
nsim
for the number of simulations; - Create a variable
lambda
for the \(\lambda\) value of the exponential distribution. Hint: the mean of the exponential distribution is given by \(\frac{1}{\lambda}\) when using the parametrization given above; - Check the documentation of
rexp
to see which parametrizationR
uses for the exponential distribution; - Simulate
nsim
observations from the exponential distribution. Store the result in the variablesim
; - Calculate
mean(sim)
and verify that the simulated mean is close to \(5\).
# @1.
nsim <- 10000;
# @2.
lambda <- 1/5;
# @3.
?rexp
# @4.
sim <- rexp(nsim, rate = lambda);
# @5.
mean(sim)
[1] 4.98
2.1.2 Exploratory analysis
- Calculate the (empirical) variance of the simulated sample;
- Calculate the (emprical) skewness of the simulated sample. The skewness is defined as \[ \frac{E((X-\mu)^3)}{\sigma^3}; \]
- Calculate (empirically) \(VaR_{0.95}\) and \(TVaR_{0.95}\) for the simulated sample.
# @1.
variance <- var(sim)
variance
[1] 25.05
# @2.
mu <- mean(sim)
sigma <- sqrt(var(sim))
numerator <- mean((sim - mu)^3)
skewness <- numerator / sigma^3
skewness
[1] 2.032
# @3.
var95 <- quantile(sim, 0.95);
var95
95%
14.83
tvar95 <- mean(sim[sim > var95])
tvar95
[1] 19.97
2.1.3 Data visualization with ggplot
- Load the package
ggplot2
; - You will construct step-by-step the following graph of the empirical CDF
Let \(x_{(i)}\) be the \(i\)-th simulated value when sorted ascending. The empirical CDF is given by
\[ \hat{F}(x_{(i)}) = \frac{\# \{\text{observations } \leq x_{(i)}\}}{n} = \frac{i}{n}.\]
2.1. Create a new ggplot. Add stat_ecdf
using the simulated data;
ggplot() +
stat_ecdf(aes(???))
2.2. Change the color of the line to blue by adding the option col = #99CCFF
to stat_ecdf
;
2.3. Add the black and white ggplot theme, theme_bw()
;
2.4. Add \(x\) and \(y\) labels to the graph. Hint: use xlab
, ylab
;
2.5. Add a vertical line to indicate the \(VaR_{0.95}\). Check the documentation for geom_vline
;
2.6. Add a title to the plot using ggtitle
;
2.7. Change the number of simulations nsim
to 50 and observe the effect on the empirical CDF.
- Use
geom_density
to create a density plot of the data. Improve the look of the graph using what you learned when creating the plot of the empirical CDF.
# @1
library(ggplot2)
# @2.1
p <- ggplot() +
stat_ecdf(aes(sim))
# @2.2
p <- ggplot() +
stat_ecdf(aes(sim), col = "#99CCFF")
# @2.3
p <- p + theme_bw()
# @2.4
p <- p +
xlab('x') +
ylab('F(x)')
# @2.5
p <- p +
geom_vline(xintercept = var95, linetype = 'dashed')
# @2.6
p <- p +
ggtitle('Empirical CDF')
print(p)
# @3
ggplot() +
geom_density(aes(sim), fill = "#99CCFF") +
theme_bw() +
ggtitle('Density') +
xlab('x')
2.1.4 Maximum Likelihood Estimation (MLE)
The density of the exponential distribution is given by
\[f(x) = \lambda \cdot e^{-\lambda \cdot x}.\]
You have nsim
simulated observations \(x_{i}, \ldots, x_{nsim}\) from this distribution. In this exercise you look for the MLE of the parameter \(\lambda\) using the simulated data.
- Write down (on paper) the formula for the likelihood of the observed data as a function of \(\lambda\);
The likelihood for the observed data is \[L(\lambda) = \prod_{i=1}^{\text{nsim}} \lambda \cdot e^{-\lambda \cdot x_i}.\]
- Derive (on paper) the loglikelihood;
The loglikelihood for the observed data is \[l(\lambda) = \sum_{i=1}^{\text{nsim}} \log(\lambda) -\lambda \cdot x_i.\]
- Compute (on paper) the MLE for \(\lambda\). Hint: put \(\frac{d \ell}{ d \lambda}(\hat{\lambda}) = 0\) and solve for the unknown \(\lambda\);
\[\frac{d l}{ d \lambda} = \frac{n}{\lambda} - \sum_{i=1}^{\text{nsim}} x_i\] This derivative is zero when
\[ \hat{\lambda} = \frac{\text{nsim}}{\sum_{i=1}^{\text{nsim}} x_i}.\] This is the MLE for \(\lambda\).
- You will now find the MLE numerically in R by optimizing the likelihood using the
nlm
procedure;
4.1. Define an R-function loglikelihood
which takes as input \(\lambda\) and returns the loglikelihood for the simulated sample;
loglikelihood <- function(lambda)
{
loglikelihood <- ???
return(loglikelihood)
}
4.2. The nlm
procedure minimizes a function. You will minimize the negative loglikelihood \(-l(\lambda)\) to find the maximum likelihood estimator \(\hat{\lambda}\). Start from the result of 4.1 and create a function negLoglikelihood
which returns the negative loglikelihood;
4.3. The nlm
procedure searches for the optimal parameter in the domain \((-\infty, \infty)\). You will use a transformation \(\lambda = \exp(\beta)\) and optimize the likelihood for this parameter \(\beta \in (-\infty, \infty)\). Update the function negloglikelihood
to take \(\beta\) as its input;
4.4. Minimize the function negLoglikelihood
you defined in 4.3. using the nlm
procedure. Add the option hessian = TRUE
;
4.5. Interpret the output of nlm
. What is the maximum likelihood estimate for \(\beta\) and what about \(\lambda\) (see the discussion in Section 13.3 on Variable and interval estimation). Do you find the same result as in 3.?
4.6. You will now construct a \(95\%\) confidence interval for the unknown parameter \(\beta\) and afterwards for \(\lambda\). Under MLE the actual parameter \(\beta\) is asymptotically distributed as (see Chapter 13 on Variance and interval estimation)
\[ \beta \sim \mathcal{N}(\hat{\beta}, \mathcal{I}^{-1}(\hat{\beta})),\] where \(\mathcal{I}\) denotes the Fisher information matrix. You calculate this matrix as the negative of the Hessian, the matrix with the second order derivatives of the log-likelihood, evaluated in \(\hat{\beta}\). Of course, since the Exponential distribution only has one paremeter, the matrix reduces to a scalar.
4.6.1. You added the option hessian = TRUE
in nlm
to obtain the Hessian (numerically) in the nlm
procedure. Use the Hessian to calculate the standard error of the MLE \(\hat{\beta}\). Because you calculated the Hessian of the negative log-likelihood, it suffices to take its inverse to obtain the (asymptotic) variance of the MLE.
4.6.2. A \(95\%\) confidence interval for the actual parameter \(\beta\) is then given by
\[ [\hat{\beta} - \Phi^{-1}(0.975) \cdot \text{se}_{\hat{\beta}}, \hat{\beta} + \Phi^{-1}(0.975) \cdot \text{se}_{\hat{\beta}}], \] where \(\Phi\) is the CDF of the standard normal distributon. Calculate the \(95\%\) confidence interval for the intensity \(\beta\) based on the simulated sample. Is the orignal \(\beta = \log \lambda\) (used for simulating the data) contained in this interval?
4.6.3 You will now use the delta method (see Section 13.3 in the book) to construct a confidence interval for the unknown \(\lambda\). The MLE for \(\lambda\) is obtained from the transformation \(\hat{\lambda}=\exp \hat{\beta}\). The corresponding se is calculated as \(se_{\hat{\lambda}} = (\exp \hat{\beta})^2 \cdot \text{se}_{\hat{\beta}}\). Using these ingredients you are now ready to construct the confidence interval for the unknown parameter \(\lambda\).
# @1
loglikelihood <- function(lambda)
{
loglikelihood <- nsim * log(lambda) - sum(lambda * sim)
return(loglikelihood)
}
# @2
negLoglikelihood <- function(lambda)
{
loglikelihood <- nsim * log(lambda) - sum(lambda * sim)
return(-loglikelihood)
}
# @3
negLoglikelihood <- function(beta)
{
lambda <- exp(beta)
loglikelihood <- nsim * log(lambda) - sum(lambda * sim)
return(-loglikelihood)
}
# @4
fit <- nlm(negLoglikelihood, p = 0, hessian = TRUE)
Warning in nlm(negLoglikelihood, p = 0, hessian = TRUE): NA/Inf replaced by
maximum positive value
fit
$minimum
[1] 26055
$estimate
[1] -1.606
$gradient
[1] -7.251e-05
$hessian
[,1]
[1,] 10001
$code
[1] 1
$iterations
[1] 8
# @5
lambdaMLE <- exp(fit$estimate)
lambdaMLE
[1] 0.2008
nsim / sum(sim)
[1] 0.2008
# @6
sigma.beta <- sqrt(solve(fit$hessian))
sigma.lambda <- sigma.beta * lambdaMLE^2
c(lambda - qnorm(0.975) * sigma.lambda, lambda + qnorm(0.975) * sigma.lambda)
[1] 0.1992 0.2008
2.2 Discrete distributions
In this computer lab you will simulate discrete data (e.g. claim counts). You will then explore and fit a statistical model to the simulated data set.
2.2.1 Simulating the data
You simulate 10 000
observations from a lognormal distribution (a continuous distribution!). You will then discretize the simulated data by rounding down.
- Simulate
10 000
observations from a lognormal distribution with density
\[f(x) = \frac{1}{x \cdot \sqrt{2 \pi}} \cdot \exp \left( -(\ln(x) + 1.5 \right)^2).\]
Hint: Check the specification of the lognormal distribution in R, ?rlnorm
.
- Discretize the data by rounding down using the
floor
function.
# Example of the floor function
x <- runif(6)*3
rbind(x = x, floor = floor(x))
[,1] [,2] [,3] [,4] [,5] [,6]
x 2.576 2.158 0.347 0.4791 2.429 0.6439
floor 2.000 2.000 0.000 0.0000 2.000 0.0000
nsim <- 10000;
sim <- exp(rnorm(nsim, mean = -1.5, sd = 1))
# we only observe the data discrete
sim <- floor(sim)
2.2.2 Exploratory analysis
You just obtained simulated discrete data. You now want to investigate which discrete distributions could be good candidates for modelling the simulated data.
- Start by calculating the mean and variance of the simulated data. Is the data underdispersed or overdispersed?
The variance is larger than the mean of the data. The data is overdispersed.
- Which of the following three distributions will most likely describe the data in a good way?
- Binomial
- Poisson
- Negative binomial
The Negative binomial distribution is the best candidate, since this distribution is overdispersed.
- Test visually whether the data belongs to the \((a, b, 0)\) class, i.e. see whether the relation \[ k \cdot \frac{p_k}{p_{k-1}} = a \cdot k + b, \quad k = 1,2, \ldots\] holds for the simulated data.
3.1 Compute the left hand side (lhs) of this relation.
- Use
prop.table(table(???))
to get the empirical probability distribution \(p_k\); - The data is heavy tailed and the lhs can become very large when \(p_{k-1}\) is small. You check the relation for \(k = 1, \dots, 5\);
- Create a vector \(k\), \(p_k\) and \(p_{k-1}\) for \(k = 1,\ldots,5\);
- Combine these results to obtain the lhs of the equation.
3.2 Use ggplot to construct a graph containg the points \((k, k \cdot \frac{p_k}{p_{k-1}})\). Your graph should look similar to
- Load the package
ggplot2
; - Create a new ggplot figure. Add a
geom_point
graphic using the data points \((k, lhs)\);
ggplot() +
geom_point(aes(???, ???))
- Change the color of the points to blue by adding the option
col = #99CCFF
togeom_point
; - You can further customize the graph with
theme_bw()
,xlab
,ylab
,ggtitle
, \(\ldots\).
3.3. Discuss. Is a distribution from the \((a, b, 0)\) class a good candidate for this data?
mean(sim)
[1] 0.0897
var(sim)
[1] 0.1555
prob <- prop.table(table(sim))
prob
sim
0 1 2 3 4 5 6 8 10
0.9319 0.0538 0.0104 0.0023 0.0010 0.0002 0.0001 0.0002 0.0001
k <- 1:5
pk <- prob[k+1]
pkmin1 <- prob[k]
lhs <- k * pk / pkmin1
ggplot() +
geom_point(aes(k[1:5], lhs[1:5]), color = "#99CCFF") +
theme_bw() +
xlab("k") +
ggtitle('(a, b, 0)-relation test plot')
Don't know how to automatically pick scale for object of type table. Defaulting to continuous.
2.2.3 Maximum Likelihood Estimation (MLE)
You will now fit two count distributions to the simulated data:
- Geometric
- Negative binomial (NB)
2.2.3.1 Geometric
For a Geometric distribution with parameter \(p \in [0, 1]\) the probability of observing \(k\) events is given by \[ P(N = k) = (1-p)^{k} \cdot p.\] In the Appendix ‘An inventory of discrete distributions’ of the Loss Models book you will find a different parameterization. That is \[ P(N=k) = \left(\frac{\theta}{1+\theta}\right)^k \cdot \frac{1}{(1+\theta)}.\] If you put \(p = \frac{1}{1+\theta}\) you can work from the second to the first parametrization. Verify this.
- Derive an expression for the loglikelihood;
The likelihood is given by \[ L(p) = \prod_{i=1}^{\text{nsim}} P(N = x_i) = \prod_{i=1}^{\text{nsim}} (1-p)^{x_i} \cdot p.\] The loglikelihood is
\[ l(p) = \sum_{i=1}^{\text{nsim}} \left( \log(1-p) \cdot x_i + \log(p) \right).\]
- Implement the negative loglikelihood as a function in R;
geom.negLoglikelihood <- function(p)
{
loglikelihood <- ???
return(-loglikelihood)
}
- The probability \(p\) can only take values in \([0, 1]\). Change the function
geom.negLoglikelihood
to take a parameter \(\beta \in (-\infty, \infty)\). Then transform the interval \((-\infty, \infty)\) to \([0, 1]\) using the logit transform
\[ \text{logit}(p) = \log\left( \frac{p}{1-p} \right) = \beta. \]
Inverting this expression, you find (verify this!)
\[ p = \frac{\exp(\beta)}{ 1 + \exp(\beta) }.\]
- Maximize the likelihood using the
nlm
procedure in R and interpret the results.
geom.negLoglikelihood <- function(beta)
{
p <- exp(beta) / (1+exp(beta))
loglikelihood <- sum(log(1-p) * sim) + nsim * log(p)
return(-loglikelihood)
}
fit <- nlm(geom.negLoglikelihood, 1)
Warning in nlm(geom.negLoglikelihood, 1): NA/Inf replaced by maximum positive
value
Warning in nlm(geom.negLoglikelihood, 1): NA/Inf replaced by maximum positive
value
fit
$minimum
[1] 3099
$estimate
[1] 2.411
$gradient
[1] -0.000222
$code
[1] 1
$iterations
[1] 6
geom.p <- exp(fit$estimate) / (1+exp(fit$estimate))
geom.p
[1] 0.9177
geom.loglik <- -fit$minimum
2.2.3.2 Negative binomial
You will now go from the one parameter geometric distribution to a two parameter discrete distribution, the Negative Binomial. Its pf is specified as follows:
\[Pr(N=k) = \frac{\Gamma(a+k)}{\Gamma(a) k!}\left(\frac{\mu}{\mu+a}\right)^{k}\left(\frac{a}{\mu+a}\right)^{a}.\]
- Follow the same steps as with the geometric distribution to fit the NB distribution to the simulated data.
- The parameters \(\mu\) and \(a\) can take values on the positive real line \([0, \infty)\). Choose an appropriate transormation to convert this interval to the whole real line, \((-\infty, \infty)\).
NB.negativeLoglikelihood <- function(beta)
{
mu <- exp(beta[1])
a <- exp(beta[2])
loglikelihood <- sum(lgamma(a + sim) - lgamma(a) - lfactorial(sim) + sim * log(mu/(mu + a)) + a * log(a / (mu + a)))
return(-loglikelihood)
}
fit <- nlm(NB.negativeLoglikelihood, c(0, 0), hessian=TRUE)
Warning in nlm(NB.negativeLoglikelihood, c(0, 0), hessian = TRUE): NA/Inf
replaced by maximum positive value
Warning in nlm(NB.negativeLoglikelihood, c(0, 0), hessian = TRUE): NA/Inf
replaced by maximum positive value
Warning in nlm(NB.negativeLoglikelihood, c(0, 0), hessian = TRUE): NA/Inf
replaced by maximum positive value
fit
$minimum
[1] 2975
$estimate
[1] -2.411 -1.893
$gradient
[1] 8.958e-04 2.859e-05
$hessian
[,1] [,2]
[1,] 562.22093 0.01064
[2,] 0.01064 121.47852
$code
[1] 1
$iterations
[1] 19
# Store the fitted values
nb.mu <- exp(fit$estimate[1])
nb.a <- exp(fit$estimate[2])
c(mu = nb.mu, a = nb.a)
mu a
0.0897 0.1506
nb.loglik <- -fit$minimum
2.2.4 Comparing fitted models
You will now compare which model best fits the data using AIC as well as some visual inspection tools.
2.2.4.1 AIC
Suppose that you have a statistical model calibrated on some data. Let \(k\) be the number of estimated parameters in the model. Let \({\displaystyle {\hat {L}}}\) be the maximum value of the likelihood function for the model. Then the AIC of the investigated model is the following
\[ \text{AIC} = 2 k - 2 \ln ( \hat{L}). \]
Given a set of candidate models for the data, the preferred model is the one with the minimum AIC value. Thus, AIC rewards goodness of fit (as assessed by the likelihood function), but it also includes a penalty that is an increasing function of the number of estimated parameters. The penalty discourages overfitting, because increasing the number of parameters in the model almost always improves the goodness of the fit. For more information see wikipedia.
Calculate the AIC for both fitted models. Hint: \(-\ln ( \hat{L})\) is the minimum reached by the
nlm
procedure.Which of the two models does AIC prefer?
aic <- c(geom = 2 * 1 - 2 * geom.loglik,
nb = 2 * 2 - 2 * nb.loglik)
# the nb distribution has the lowest AIC
print(aic)
geom nb
6200 5955
2.2.4.2 Visual inspection
Using the fitted parameters you will now simulate new datasets of nsim
observations from the Geometric and Negative Binomial distribution. You compare the shapes of the fitted and the original data.
- Simulate a dataset from the Geometric distribution using the fitted parameters;
- Simulate a dataset from the Negative binomial distribution using the fitted parameters;
- You will now create a barplot using
geom_barplot
. First the data has to be merged into a single data frame containing two columns:
- x: the simulated values;
- method: a string, referring to the method used to simulate the data (i.e. observed, geom or nb).
x method
0 observed
0 observed
0 geom
1 geom
0 nb
3.1 Create datasets df.observed
, df.geom
, df.nb
with the simulated data in one column and a string referring to the method used in the other column.
df.observed <- data.frame(x = ???, method = 'observed')
3.2 Combine these three datasets into a single dataset using rbind
.
df <- rbind(df.observed, df.geom, df.nb);
- Create a barplot using
geom_bar
.
ggplot() +
geom_bar(aes(???, fill = ???)) +
theme_bw()
By default
geom_bar
stacks the bars for the different methods. To show the bars sideways add the optionposition = position_dodge()
.Discuss. Which distribution best mimicks the original data?
sim.geom <- rgeom(nsim, geom.p)
sim.nb <- rnbinom(nsim, mu = nb.mu, size = nb.a)
df.observed <- data.frame(x = sim, method = 'observed')
df.geom <- data.frame(x = sim.geom, method = 'geom')
df.nb <- data.frame(x = sim.nb, method = 'nb')
df <- rbind(df.observed, df.geom, df.nb);
ggplot() +
geom_bar(aes(df$x, fill=df$method), position = position_dodge()) +
theme_bw()