Loss distributions for modeling claim severity
In this part, we will focus on the following aspects to discuss the distributions:
- Characteristics of loss distributions: we will explore the Moment Generating Function (MGF), probability density function (PDF), cumulative distribution function (CDF), survival function and etc.
- Propositions of loss distributions: we will discuss the properties of the distributions, such as the mean, variance, skewness, and kurtosis.
- Estimation of parameters: we will introduce the methods to estimate the parameters of the distributions, such as the method of moments (MM), maximum likelihood estimation (MLE), and etc.
- Applications in insurance: we will discuss the applications of the distributions in insurance contexts, such as pricing insurance policies, setting reserves, and etc.
Note
More details will be added into Probability and Statistics section.
Preliminaries
Before we delve into the specific distributions, let's first introduce some basic concepts that are essential for understanding the subsequent discussions.
When you study the loss distributions, you should have already known that the modeling is based on the random variables and the probability. We use to denote the probability of the random variable .
Then we have the following definitions:
Probability density function (PDF): The probability density function (PDF) of a continuous random variable is a function that describes the likelihood of the variable taking on a particular value . The PDF is non-negative and integrates to 1 over the entire range of possible values of .
Cumulative distribution function (CDF): The cumulative distribution function (CDF) of a random variable is a function that describes the probability that the variable takes on a value less than or equal to .
Survival function: The survival function of a random variable is a function that describes the probability that the variable takes on a value greater than .
Exponential distribution
The exponential distribution is a continuous probability distribution that describes the time between events in a Poisson process, where events occur continuously and independently at a constant average rate. The exponential distribution is characterised by a single parameter , which represents the rate at which events occur.
Characteristics of the exponential distribution
If a random variable is exponential distributed with rate parameter , we write .
PDF:
CDF:
Survival function:
Propositions of the exponential distribution
The exponential distribution has the following properties:
MGF(Moment Generating Function):
CGF (Cumulant Generating Function):
Mean: The mean can be generated from the MGF:
Variance:
or we can use CGF to calculate the variance:
Skewness: According to the definition of CDF, we can calculate the skewness:
Estimation of parameters
Method of moments (MM)
From the above propositions, we can see that the first moment of the exponential distribution is . We only have one parameter to estimate, so we can use the method of moments to estimate the parameter . It can be solved by the following criteria, the first origin moment is equal to the first sample moment:
Then we can get the estimator of :
Maximum likelihood estimation (MLE)
The likelihood function of the exponential distribution is:
Then the log-likelihood function is:
To find the MLE of , we need to solve the following equation:
Then we can get the MLE of :
Gamma distribution
The gamma distribution is a continuous probability distribution that generalises the exponential distribution. It is characterized by two parameters: a shape parameter and a rate parameter (or ). The exponential distribution is a special case of the gamma distribution when .
Characteristics of the gamma distribution
If a random variable is gamma distributed with shape parameter and rate parameter , we write or .
PDF:
where is the gamma function defined as .
CDF:
Survival function:
where is the lower incomplete gamma function defined as .
Propositions of the gamma distribution
The gamma distribution has the following properties:
MGF:
CGF:
Mean: The mean can be generated from the MGF:
Variance:
Skewness:
Estimation of parameters
Method of moments (MM)
In the gamma distribution, we have two parameters to estimate, so we need to use the method of moments to estimate the parameters and . The first moment of the gamma distribution is , and the second moment is . Both of them should be equal to the sample moments.
We can solve the following criteria:
Solve the first moment,
Then, replace with in the second moment, we can get the estimator of :
and the estimator of :
The method of moments estimators for and are:
Maximum likelihood estimation (MLE)
The likelihood function of the gamma distribution is:
Then the log-likelihood function is given by:
Taking the derivative of the log-likelihood function with respect to and ,
we have the following equations:
Solve the equation, we have the MLE of :
Then, taking the derivative of the log-likelihood function with respect to ,
where is the digamma function.
Replace with , we can get the MLE of :
This equatio can be solved by R or Python.
x <- rgamma(1000, shape = 5, rate = 0.5) # nolint R
aux <- log(mean(x)) - mean(log(x)) # nolint: assignment_linter.
f <- function(z) {
log(z) - digamma(z) - aux
}
alpha <- uniroot(f, c(1e-8, 1e8))$root
lambda <- alpha / mean(x)Lognormal distribution
The lognormal distribution is a continuous probability distribution of a random variable whose logarithm is normally distributed. It is characterised by two parameters: the mean and the standard deviation of the logarithm of the variable.
Characteristics of the lognormal distribution
If a random variable is lognormally distributed with parameters and , we write or .
PDF: Letting , we can derive the PDF of the lognormal distribution:
CDF:
Note
CDF of the lognormal distribution will not be used normally in the actuarial practice.
- Survival function:
Propositions of the lognormal distribution
The lognormal distribution has the following properties:
MGF: The lognormal distribution does not have a closed-form MGF.
Mean:
According to the relationship between and , we can derive the mean for lognormal distribution by:
Variance: Similar to the mean, we can derive the variance for lognormal distribution by:
where is the MGF of the normal distribution .
Skewness: The skewness of lognormal distribution should be derived by the standard formula:
Estimation of parameters
The lognormal distributions are similar to the normal distribution that they have two parameters and .
Method of moments (MM)
like the previous estimation, we can have the following first tww origin moments from the above propositions:
Solve the first equation with taking logarithem:
Then, replace with in the second equation, we can get the estimator of :
Then we use the sample moments to estimate the parameters and .
Replace the above estimators into the equations, we can get the estimators of and :
Maximum likelihood estimation (MLE)
The likelihood function of the lognormal distribution is:
Then the log-likelihood function is:
Taking the derivative of the log-likelihood function with respect to and , we have:
Then, solve for , we have
The MLE of and are:
Conclusion
In this part, we have discussed the loss distributions that are commonly used in actuarial practice, including the exponential distribution, gamma distribution and lognormal distribution. We have explored the characteristics of these distributions, their key properties, and the methods for estimating their parameters. These distributions are essential for modeling claim severity in insurance contexts and are widely used for pricing insurance policies, setting reserves, and assessing risk.