Modeling the Frequency and Severity of Auto Insurance Claims Using Statistical Distributions

Claims experience in non-life insurance is contingent on random eventualities of claim frequency and claim severity. By design, a single policy may possibly incur more than one claim such that the total number of claims as well as the total size of claims due on any given portfolio is unpredictable. For insurers to be able to settle claims that may occur from existing portfolios of policies at some future time periods, it is imperative that they adequately model historical and current data on claims experience; this can be used to project the expected future claims experience and setting sufficient reserves. Non-life insurance companies are often faced with two challenges when modeling claims data; selecting appropriate statistical distributions for claims data and establishing how well the selected statistical distributions fit the claims data. Accurate evaluation of claim frequency and claim severity plays a critical role in determining: An adequate premium loading factor, required reserve levels, product profitability and the impact of policy modifications. Whilst the assessment of insurers’ actuarial risks in respect of their solvency status is a complex process, the first step toward the solution is the modeling of individual claims frequency and severity. This paper presents a methodical framework for choosing a suitable probability model that best describes automobile claim frequency and loss severity as well as their application in risk management. Selected statistical distributions are fitted to historical automobile claims data and parameters estimated using the maximum likelihood method. The Chi-square test is used to check the goodness-of-fit for claim frequency distributions whereas the Kolmogorov-Smirnov and Anderson-Darling tests are applied to claim severity distributions. The Akaike information criterion (AIC) is used to choose between competing distributions. Empirical results indicate that claim severity data is better modeled using heavy-tailed and skewed distributions. The lognormal distribution is selected as the best distribution to model the claim size while negative binomial and geometric disHow to cite this paper: Omari, C.O., Nyambura, S.G. and Mwangi, J.M.W. (2018) Modeling the Frequency and Severity of Auto Insurance Claims Using Statistical Distributions. Journal of Mathematical Finance, 8, 137-160. https://doi.org/10.4236/jmf.2018.81012 Received: November 27, 2017 Accepted: February 23, 2018 Published: February 26, 2018 Copyright © 2018 by authors and Scientific Research Publishing Inc. This work is licensed under the Creative Commons Attribution International License (CC BY 4.0). http://creativecommons.org/licenses/by/4.0/ Open Access


Introduction
The development of insurance business is driven by general demands of the society for protection against various types of risks of undesirable random events with a significant economic impact.Insurance is a process that entails the provision of an equitable method of offsetting the risk of a likely future loss with a payment of a premium.The underlying concept is to create a fund to which the insured members contribute predetermined amounts of the premium for given levels of loss.When the random events that policyholders are protected against occur giving rise to claims then claims are settled from the fund.The characteristic feature of such an arrangement is that the insured members are faced with a homogeneous set of risks.The positive aspect of forming such communities is the pooling together of risks which enables members to benefit from the weak law of large numbers.
In the non-life insurance industry, there is increased interest in the automobile insurance because it requires the management of a large number of risk events.These include cases of theft and damage to vehicles due to accidents or other causes as well as the extent of damage and parties involved [1].General insurance companies deal with large volumes of data and need to handle this data in the most efficient way possible.According to Klugman et al. [2], actuarial models assist insurance companies to handle these vast amounts of data.The models are intended to represent the real world of insurance problems and to depict the uncertain behavior of future claims payments.This uncertainty necessitates the use of probability distributions to model the occurrence of claims, the timing of the settlement and the severity of the payment, that is, the amount paid.
One of the main challenges that non-life insurance companies face is to accurately forecast the expected future claims experience and consequently determining an appropriate reserve level and premium loading.The erratic results often reported by non-life insurance companies lead to the pre-conclusion that it is virtually impossible to conduct the business on a sound basis.Most non-life insurance companies base their estimations of claim frequency and severity on their own historical claims data.This is sometimes complemented with data from external sources and is used as a base for managerial decisions [3].The management, with the guidance of actuaries, ensures that the insurance compa-ny is run on a sound financial basis by charging the appropriate premium corresponding to the specific risks and also maintaining a suitable reserve level.
A general approach to claims data modeling is to consider separately claim count experience from the claim size experience.Both claim frequency and severity random variables; hence there is a risk that future claims experience will deviate from past eventualities.It is therefore imperative that appropriate statistical distributions are applied in modeling of claims data variables.Although the empirical distribution functions are useful tools in analyzing claims processes, it is always convenient to model claims data by fitting a probability distribution with mathematically tractable features.
Statistical modeling of claims data has gained popularity in the past recent years, particularly in the actuarial literature all in an attempt to address the problem of premium setting and reserves calculations.Hogg and Klugman [4] provide an extensive description of the subject of insurance loss models.Loss models are useful in revealing claim-sensitive information to insurance companies and giving insight into decisions on required premium levels, premium loading reserves and assessing the profitability of insurance products.The main components of aggregate claims of a non-life insurance company are frequency/count and size/severity.Claim frequency refers to the total count of claims that occur on a block of policies, while claim severity or claim size is the monetary amount of loss on each policy or on the portfolio as a whole.The claims experience random variables are often assumed to follow certain probability distributions, referred to as loss distributions [5].Loss distributions are typically skewed to the right with relatively high right-tail probabilities.Various studies available literature describes these probability distributions as being long-tailed or heavy-tailed.
When assessing portfolio claim frequency, it is often found that certain policies did not incur any claims since the insured loss event did not occur to the insured.Such cases result in many zero claim counts such that the claim frequency random variable takes the value zero with high probability.Antonio et al. [6] present the Poisson distribution as the modeling archetype of claim frequency.In available economic literature, considerations have been made for models of count data on claims frequency that allow for excess zeros.For instance, Yip and Yau [7] fit a zero-inflated count model to their insurance data.
Boucher et al. [8] considered zero-inflated and hurdles models with application to real data from a Spanish insurance company.Mouatassim and Ezzahid [9] analyzed the zero-inflated models with an application to private health insurance data.
Modeling claim frequency and claim severity of an insurance company is an essential part of insurance pricing and forecasting future claims.Being able to forecast future claim experience enables the insurance company to make appropriate prior arrangements to reduce the chances of making a loss.Such arrangements include setting a suitable premium for the policies and setting aside The remainder of the paper is organized as follows: Section 2 discusses the statistical modeling procedure for claims data.Section 3 presents the empirical results of the study.Finally, Section 4 concludes the study.

Statistical Modeling of Claims Data
The statistical modeling of claims data involves the fitting of standard probability distributions to the observed claims data.Kaishev and Krachunov [11] suggested the following four-step statistical procedure for fitting an appropriate statistical distribution to claims data: 1) Selection of the claims distribution family 2) Estimation of parameters of the chosen fitted distributions 3) Specification of the criteria to select the appropriate distribution from the family of distributions.
4) Testing the goodness of fit of the approximate distributions.

Selection of Claims Distributions
The initial selection of the models is based on prior knowledge on the nature and form of claim's data.Claim frequency is usually modeled using non-negative discrete probability distributions since the number of claims is discrete and nonzero.Claim severity is known to be best modeled using non-zero continuous distributions which are skewed to the right and have heavy tails.Kaas, et al. [12] attributes this to the fact that extremely large claim values often occur in the upper-right tails of the distribution.Prior knowledge of claims experiences in nonlife insurance coupled with descriptive analysis of prominent features of the claims data and graphical techniques a used to guide the selection of the initial approximate probability distributions of claim size and frequency.Merz and Wüthrich [13] noted that majority of claims data arising from the general insurance industry are positively skewed and have heavy tails.They argued that statistical distributions which exhibit these characteristics may be appropriate for modeling such claims.This study proposes a number of standard probability distributions that could be used to approximate the distributions for claim amount and claim count random variables.The binomial, geometric, negative-binomial and Poisson distributions are considered for modeling claim frequency as they are discrete.On the other hand, five standard continuous probability distributions are proposed for modeling claim severity.These are the exponential, gamma, Weibull, Pareto and the lognormal distributions.The probability distribution functions along with their parameters estimation and their respective properties are discussed subsequently.

Estimation of Parameters
In this paper, the parameters of the selected claim distributions are estimated using the maximum likelihood estimation (MLE) technique.The MLE is a commonly applied method of estimation in a variety of problems.It often yields better estimates compared to other methods like; least-squares estimation (LSE), the method of quantile and method of moments especially when the sample size is large.Boucher et al. [14] argued that the MLE approach fully utilizes all the information about the parameters contained in the data and yields highly flexible estimator with better asymptotic properties.
is a random sample of independent and identically distributed (iid) observations drawn from an unknown population.Let X x = denote a realization of a random variable or vector X with probability mass or density function where θ is a scalar or a vector of unknown para- meters to be estimated.The objective of statistical inference is to infer θ from the observed data.The MLE involves obtaining the likelihood function of a random variable.The likelihood function ( ) L θ is the probability mass or density function of the observed data x, expressed as a function of the unknown parameter θ .Given that The probability distribution functions of the selected claim distributions along with their maximum likelihood estimates for the parameters are given as follows:

Binomial Distribution
The binomial distribution is a popular discrete distribution for modeling count data.Given a portfolio of n independent insurance policies, let X denote a binomially distributed random variable that represents the number of policies in the portfolio that result in a claim.The claim count variable X can be said to follow a binomial distribution with parameters n and p, where n is a known positive integer representing the number of policies on the portfolio and p is the probabili-ty that there is at least one claim on an individual policy.The probability distribution function of X is defined as: ( ) ( ) The expected value of the binomial distribution is np and variance ( ) Hence, the variance of the binomial distribution is smaller than the mean.
The corresponding binomial likelihood function is ( ) ( ) Therefore the log-likelihood function is where k is a constant that does not involve the parameter p.Therefore, to determine the parameter estimate p s we obtain the derivative of the log-likelihood function with respect to the parameter and equate to zero.
( ) Solving the Equation ( 6) gives the MLE.Thus, the MLE is p x n = .

Poisson Distribution
The Poisson distribution is a discrete distribution for modeling the count of randomly occurring events in a given time interval.Let X be the number of claim events in a given interval of time and λ is the parameter of the Poisson distribution representing the mean number of claim events per interval.The probability of recording x claim events in a given interval is given by A Poisson random variable can take on any positive integer value.Unlike the Binomial distribution which always has a finite upper limit.In general, the expected value and variance functions of the Poisson distribution are both equal λ.
The Poisson likelihood function is ( ) Therefore, the log-likelihood function is ( ) Differentiating the log-likelihood function with respect to λ, ignoring the constant term that does not depend on λ and equating to zero Solving the Equation ( 9) gives the MLE.Thus, the MLE is .

Geometric Distribution
The geometric distribution is a discrete distribution that generates the probability that the first occurrence of an event of interest requires x independent trials, where each trial results in either success or failure, and the probability of success in any individual trial is constant.The probability distribution function of the geometric distribution is ( ) ( ) where p is the probability of success, and x is the number of failures before the first success.The geometric distribution is the only discrete memoryless random distribution analogous to the exponential distribution.
The likelihood function for n independent and identically distributed trials is The log-likelihood function is: Thus, to determine the parameter estimate, equate the derivative of the loglikelihood function to zero.
( ) Therefore, the MLE is

Negative Binomial Distribution
The negative binomial distribution is a generality of the geometric distribution.
Consider a sequence of independent Bernoulli trials, with common probability p, until we get r successes.If X denotes the number of failures which occur before the r-th success, then X has a negative binomial distribution given by The negative binomial distribution has an advantage over the Poisson distribution in modeling because it has two positive parameters 0 p > and 0 r > .
The most important feature of this distribution is that the variance is bigger than the expected value.A further significant feature in comparing these three discrete distributions is that the binomial distribution has a fixed range but the negative binomial and Poisson distributions both have infinite ranges.
The likelihood function is The log-likelihood function is obtained by taking logarithms ( ) ( ) Taking the derivative with respect to p and equating to zero.
The resulting MLE estimate is ˆnr p nr x = + .Therefore, the negative binomial random variable can be expressed as a sum of r independent, identically distributed (geometric) random variables.

Exponential Distribution
The exponential distribution is a continuous distribution that is usually used to model the time until the occurrence of an event of interest in the process.A continuous random variable X is said to have an exponential distribution if its probability density function (pdf) is given by: where 0 λ > is the rate parameter of the distribution.The exponential distribu- tion is a special case for both the gamma and Weibull distribution.
The likelihood function is ( ) ( ) The log-likelihood functions is therefore ( ) The parameter estimator λ is be obtained by setting the derivative of the log-likelihood function to zero and solving for λ ( ) The resulting MLE is given by ˆ1 x λ = where = ∑ is the mean ob- served time.

Gamma Distribution
The gamma distribution and Weibull distributions are extensions of the exponential distribution.Both of them are expressed in terms of the gamma function which is defined by ( ) The two-parameter gamma distribution function is given by where 0 α > is the shape parameter and 0 β > is the scale parameter.
The likelihood function for the gamma distribution function is given by ( ) ( ) The log-likelihood functions is , ln ln 1 ln Thus, to determine the parameter estimates, we equate the derivatives of the log-likelihood function to zero and solve the following equations Substituting ˆˆx β α = in Equation ( 21) results in the following relationship for α , ( ) This result is a non-linear equation in α that cannot be solved in a closed form.This can be solved numerically using the root-finding methods.

Weibull Distribution
The Weibull distribution is a continuous distribution that is commonly used to model the lifetimes of components.The Weibull probability density function has two parameters, both positive constants that determine its location and shape.
The probability density function of the Weibull distribution is where γ is the shape parameter and β the scale parameter.When 1 γ = , the Weibull distribution is reduced to the exponential distribution with parameter λ β = .
The likelihood function for the Weibull distribution is given by: ( ) The log-likelihood function is therefore given by , ln ln 1 ln Thus, to determine the parameter estimates, we equate the derivatives of the log-likelihood function to zero and solve the following equations , 1 ln ln ln 0 , 0 By eliminating β and simplifying, we obtained the following non-linear eq- uation This can be solved numerically to obtain the estimate of γ by using the Newton-Raphson method.The MLE estimate for β is given by

Log-Normal Distribution
A positive random variable X is log-normally distributed if the logarithm of the random variable is normally distributed.Hence X follows a lognormal distribution if its probability density function is given by ( ) ( ) with parameters: location µ , scale 0 σ > .
The likelihood function for the lognormal distribution is Therefore the log-likelihood function is given by ( ) ( ) The parameter estimators μ and σ for the parameters µ and σ can be determined by equating the derivatives of the log-likelihood function to zero and solve the following two equations ( ) The resulting estimates are

Pareto Distribution
The Pareto distribution with parameter 0 α > and 0 γ > , is given by where 0 α > represent the shape parameter and 0 γ > the scale parameter.
The likelihood function is It is important to note that the best way to maximize the log-likelihood function is by adjusting as follows,

{ }
ˆmin i x γ = such that γ cannot be larger than the smallest value of x in the data.Thus, to determine the parameter estimate α equates the derivative of the log-likelihood function to zero we solve the following equations Thus, we obtain After estimating the parameters of the selected distributions, the next step is typically the determination of the best fitting distribution, i.e., the distribution that provides the best fit to the data set at hand.Establishing the underlying distribution of a data set is fundamental for the correct implementation of claims modeling procedures.This is an important step that must be implemented before employing the selected distributions for forecasting insurance claims or pricing insurance contacts.The consequences of misspecification of the underlying distribution may prove very costly.One way of dealing with this problem is to assess the goodness of fit of the selected distributions.First, they are among the most common statistical test for small samples (and they can as well be used for large samples).Secondly, the tests are implemented in most statistical packages and are therefore widely used in practice.The Chi-Square test is based on the PDF while both, the K-S and A-D GoF tests use the CDF approach and therefore they are generally considered to be more powerful than the Chi-Square of fit test.

Goodness of Fit Tests
For all the GoF tests, the hypotheses of interest are: H 0 : The claims data sample follows a particular distribution, H 1 : The claims data samples do not follow the particular distribution.

The Chi-Square Goodness of Fit Test
The Chi-Square goodness of fit test is used to test the hypothesis that the distribution of a set of observed data follows a particular distribution.The Chi-square statistic measures how well the expected frequency of the fitted distribution compares with the observed frequency of a histogram of the observed data.The Chi-square test statistic is: where j O is the observed number of cases in interval j, j E is the expected number of cases in interval j of the specified distribution and k is the number of intervals the sample data is divided into.The test statistic approximately follows a Chi-Squared distribution with 1 k p − − degrees of freedom where p is the number of parameters estimated from the (sample) data used to generate the hypothesized distribution.

The Kolmogorov-Smirnov Test
The Kolmogorov-Smirnov (K-S) test compares a hypothetical or fitted cumulative distribution function (CDF) ( ) x CDF in order to assess the goodness-of-fit of a given data set to a theoretical distribution.
The CDF uniquely characterizes a probability distribution.The empirical ( ) n F x CDF is expressed as the proportion of the observed values that are less than or equal to x and is defined as where n is the size of the random sample, ( ) ( ) ( ) F x is the theoretical cumulative distribution of the distribution being tested which must be a continuous distribution, and it must be fully specified.

Data Description
The data set used in modeling the claims frequency and claims severity distributions consists of three data sets: AutoCollision, dataCar and dataOhlsson obtained from the R package insuranceData.Autocollision data set is a sample of 32 observations due to [16] who considered 8942 collision losses from private passenger United Kingdom (UK) automobile insurance policies.The average severity is in Sterling Pounds adjusted for inflation.The dataCar dataset consists of 67,856 one-year motor vehicle insurance policies issued in 2004 and 2005.Finally, dataOhlsson dataset contains 64,548 aggregated claims data for motorcycle insurance policyholders on all insurance policies and claims during 1994-1998.
Table 1 presents the summary statistics for the three claims frequency and claims severity datasets.The summary statistics include the number of observations, minimum, maximum, mean, median, standard deviation, skewness, and kurtosis.The Auto Collision data set has the highest mean for both the claims frequency and claims severity given that the number observations were only 32.
For both data Car and data Ohlsson, the mean and the standard deviations are close to zero for the claim frequency.This is because the majority of the observations in the datasets are zeros.More specifically, throughout the analyzed period the descriptive statistics show that the Data Car and Data Ohlsson losses are more extreme with respect to skewness and kurtosis than the Auto Collision losses.All the datasets are thus significantly skewed to the right and exhibit excess kurtosis, especially the claims severity have high kurtosis values compared to claims frequency.This, in addition, confirms that the data sets do not follow the normal distribution.The descriptive statistics confirm the skewed nature of the claims data.As a result of these findings, the skewed-heavy tailed distributions are the most appropriate to fit this type of data since they account for skewness, excess kurtosis as well as heavy tails.

Parameters Estimation
The parameters of the claims frequency and claims severity fitted distributions are estimated using the maximum likelihood estimation method and is implemented via packages in R statistical software.Table 4  Table 5 reports the estimated parameters their (asymptotic) standard errors,

Goodness-of-Fit Tests
The purpose of goodness-of-fit tests is typically to measure the distance between the fitted parametric distribution and the empirical distribution: e.g., the DOI: 10.4236/jmf.2018.81012144 Journal of Mathematical Finance

A
goodness-of-fit (GoF) test is a statistical procedure that describes how well a distribution fits a set of observations by measuring the quantifiable compatibility between the estimated theoretical distributions against the empirical distributions of the sample data.The GoF tests are effectively based on either of the two distribution functions: the probability density function (PDF) or cumulative distribution function (CDF) in order to test the null hypothesis that the unknown distribution function is, in fact, a known specified function.The tests considered for testing the suitability of the fitted distributions to claims data include; the Chi-Square goodness of fit test, Kolmogorov-Smirnov (K-S) test, and the Anderson-Darling (A-D) test.The three GoF tests were selected for two reasons.C. O. Omari et al.DOI: 10.4236/jmf.2018.81012149 Journal of Mathematical Finance x is the number of xi's less than or equal to x.To test the null hypothesis, the maximum absolute distance be-

Figure 1 Figure 1 .
Figure 1 displays the histograms of the claims frequency and claims severity data.Most of the claims frequency and claims severity for data Car and data Ohlsson data are zero and it is clear that we have zero-inflation problem.For the Auto Collison dataset for claims frequency, most of the values are less than 200 and the data is slightly positively skewed.The claims severity data values mostly lie between 200 and 300 and also we have very few extremely large values that lie between 700 and 800.Also, the histogram in Figure 1 illustrates that claims data in our case are skewed and exhibits heavy-tailed distributions.Table 2 reports the summary number of claims occurrences distribution recorded and their respective percentages for the data Car and data Ohlsson datasets.The maximum numbers of reported claims occurrences are four in data Car and two in data Ohlsson respectively.The zero claim account for the significant percentage of the claim occurrences recorded.For data Car, the percentage of total zero claims occurrences are 93.18% and 98.96% for data Ohlsson respectively.This kind of data is defined as zero-inflated because of the zero response of the unaffected policyholders.Therefore it is obvious that the distribution of claim occurrences should be one with a very high probability for a claim occurrence of zero.Such distributions would be discrete distributions with a large parameter (probability of success) since they have the mode at zero and the probability of the other values would become progressively smaller.When data displays over-dispersion, the most likely to be used distribution is the negative

Figure 2
Figure 2 displays the histograms of logarithms of claims frequency and claims severity sets.The histograms look much promising for purposes of fitting the distributions compared to the raw data histograms in Figure 1 especially for the distribution of the claim severity datasets.The Auto Collison data has most of the observations concentrated at the center.The data Car and data Ohisson data sets for claims severity are skewed to the right and left respectively.
represents the parameter estimates and their corresponding (asymptotic) standard errors, the maximum log likelihood function (LLF), AIC and BIC values for the fitted claims frequency distributions to the Auto Collision, data Car and data Ohlsson datasets.The parameter estimates for all the fitted distributions are obtained except for the LLF, AIC, and BIC for the binomial distribution for both data Car and data Ohlsson data.In order to have better comparisons among the fitted distributions, we us the LLF, AIC and BIC values of the fitted distributions.For the Auto Collision the geometric distribution has the lowest AIC followed by the negative binomial distribution.Both data Car and data Ohlsson have the Poisson distribution with the lowest AIC value followed by the negative binomial distribution.The over-dispersion observed in the data would suggest that Negative binomial or Geometric distributions might be good candidates for the modeling claims frequency data.This will be verified later by the Chi-square goodness-of-fit test.

Figure 2 .
Figure 2. The Histogram of logarithms of claims frequency and claims severity data.
The Chi-square goodness-of-fit test is used to select the most appropriate claims frequency distribution among the fitted discrete probability distributions.Table6summarizes the Chi-square goodness-of-fit test-statistic values and the corresponding p-value to further compare fitted distributions.The p-values are used to determine whether or not to reject the null hypotheses.The Negative-Binomial and geometric distributions have p-values greater than 0.01 α = in almost all cases except for the geometric distribution for data Ohlsson.The p-values for the binomial and Poisson distributions for all the three data sets are zero; hence the null hypothesis that claims data follows a particular distribution is rejected for the binomial and Poisson distributions with a 99% confidence level.Thus, the Chi-square goodness-of-fit test confirmed that both negative binomial and geometric distributions are appropriate distributions for modeling claims frequency while binomial and Poisson distributions provide the inappropriate fit.To select the most appropriate continuous distributions for the claims severity, two goodness-of-fit statistics are typically considered: K-S and A-D statistics [17] are used to test the suitability of fitted claims severity distributions to the data sets.Table 7 reports the K-S and A-D test statistic values for the claim severity distributions that are used to model the claims severity data.The K-S and A-D test statistic values are used to compare the fit of competing distributions as opposed to an absolute measure of how a particular distribution fits the data.Smaller K-S and A-D test statistic values indicate that the distribution fits the data better.However, these statistics should be used cautiously especially when comparing fits of various distributions.The Kolmogorov-Smirnov and Anderson-Darling statistics computed for several claims severity distributions fitted on the same data set like in our case are hypothetically complicated to compare.Moreover, such a statistic, as K-S one, doesn't take into account the complexity of the distribution (i.e. the number of parameters).It is less problematic where the compared distributions are characterized by the same number of parameters, but again it could methodically encourage the selection of the more complex distributions.Both the K-S and A-D goodness-of-fit statistics based on the CDF distance choose the lognormal distribution, which is also the preference for the AIC and BIC values for all the datasets as the most appropriate claims severity distribution.
settle future claims (reserves).A suitable premium level for an individual policy, with proper investment, should be enough to at least cover the cost of paying a claim on the said policy.Adequate reserves set aside should enable the insurance company to remain solvent, such that it can adequately set- DOI: 10.4236/jmf.2018.81012140Journal of Mathematical Finance money required to eling claim severity.Other distributions for claim size are the exponential, Weibull, and Pareto distributions.Achieng [2] modeled the claim amounts from First Assurance Company limited, Kenya for motor comprehensive policy.The log-normal distribution was chosen as the appropriate model that would provide a good fit to the claims data.The Pareto distribution has been shown to adequately mimic the tail-behavior of claim amounts and thus provides a good fit.The objective of this paper is to present a hypothetical procedure for selecting probability models that approximately describe auto-insurance frequency and severity losses.The method employs fitting of standard statistical distributions

Table 1 .
Descriptive statistics of the claims frequency and claims severity data.

Table 2 .
The number of claims in the datasets.

Table 3
presents the descriptive statistics of the claims data without zero claims.The claims frequency in both data Car and data Ohlsson without zero claims still has a mean and standard deviation close to zero.All the datasets are still significantly skewed to the right and exhibit excess kurtosis.This still confirms the data sets do not follow the normal distribution.

Table 3 .
Descriptive statistics of the claims data without zero claims.
Auto Collision data.The LLF, AIC and BIC criteria are employed for purposes of selecting the appropriate distribution among the fitted distributions.The distribution function with the maximum LLF and lowest AIC or BIC values is

Table 4 .
Estimated parameters for claims frequency distributions.

Table 5 .
Estimated parameters for claims severity distributions.

Table 6 .
Chi-squared test for claim frequency distributions.