Next Article in Journal
The Dynamic Endogeneity Issue between Corporate Ownership Structure and Real-Based Earnings Manipulation in an Emerging Market: Advanced Dynamic Panel Model
Previous Article in Journal
Rank-Based Multivariate Sarmanov for Modeling Dependence between Loss Reserves
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Model Uncertainty and Selection of Risk Models for Left-Truncated and Right-Censored Loss Data

1
Department of Mathematics, Robert Morris University, Moon Township, PA 15108, USA
2
Department of Mathematics, Central Washington University, Ellensburg, WA 98926, USA
3
Department of Mathematical Sciences, University of Wisconsin-Milwaukee, Milwaukee, WI 53211, USA
*
Author to whom correspondence should be addressed.
Risks 2023, 11(11), 188; https://doi.org/10.3390/risks11110188
Submission received: 6 October 2023 / Revised: 24 October 2023 / Accepted: 26 October 2023 / Published: 30 October 2023

Abstract

:
Insurance loss data are usually in the form of left-truncation and right-censoring due to deductibles and policy limits, respectively. This paper investigates the model uncertainty and selection procedure when various parametric models are constructed to accommodate such left-truncated and right-censored data. The joint asymptotic properties of the estimators have been established using the Delta method along with Maximum Likelihood Estimation when the model is specified. We conduct the simulation studies using Fisk, Lognormal, Lomax, Paralogistic, and Weibull distributions with various proportions of loss data below deductibles and above policy limits. A variety of graphic tools, hypothesis tests, and penalized likelihood criteria are employed to validate the models, and their performances on the model selection are evaluated through the probability of each parent distribution being correctly selected. The effectiveness of each tool on model selection is also illustrated using well-studied data that represent Wisconsin property losses in the United States from 2007 to 2010.

1. Introduction

In actuarial science, calculating financial and insurance risk measures plays an important role in the modeling of financial and insurance loss data. In many cases, loss data exhibit complex features such as left-truncation and right-censoring. These features pose challenges on model selection and the analysis of model uncertainty.
The selection of appropriate risk models for left-truncated and right-censored (LTRC) loss data is an essential task in actuarial practice. Selection of inaccurate models can lead to incorrect estimation of financial risk, resulting in sub-optimal decision making. There is a need for a thorough analysis of model uncertainty and model selection for LTRC loss data. This article aims to address the model uncertainty and the corresponding effect on model selection.
In the actuarial literature, Brazauskas and Upretee (2019) studied various aspects of LTRC insurance loss data such as the probability density function (PDF), cumulative distribution function (CDF), and quantile function (QF), building up a framework to derive asymptotic distributions of parametric and empirical estimators. Moreover, Poudyal et al. (2023) designed some robust parametric estimation procedures for the insurance payment data affected by deductibles, policy limits, and coinsurance factors. These authors clearly presented examples and showed how model uncertainty arises in the model fitting process of the actuarial loss data and how to deal with model mis-specification.
Left-truncation and right-censoring are both types of data incompleteness commonly encountered in survival analysis. Mitra (2012) studied the left-truncation and right-censoring in lifetime data. Left-truncation is the situation where the individuals in the study are only observed from a certain point in time forward, and those who have experienced the event of interest prior to this time point are excluded from the analysis. Right-censoring refers to the situation where the event of interest has not occurred for some individuals by the end of the study period. In our study, left-truncation means that losses below a certain deductible value have not been observed, and right-censoring means that frequencies are only observed for losses above a certain policy limit value without further information about severity. This context of insurance loss data differs from the context of lifetime data for survival analysis.
In the model selection procedure, the selection criteria play a critical role in determining the best candidate model. The performance of typically penalized likelihood measures, the Akaike Information Criterion (AIC) and the Bayesian Information Criterion (BIC), have been reviewed (see Emura and Michimae (2022), for example) and discussed using various incomplete data sets with the data truncation and censoring facilitated according to their industrial applications. Balakrishnan and Mitra (2014) demonstrated that AIC is superior to BIC when the true model comes from the Gamma, Lognormal, and Weibull distributions. Mitra et al. (2021) suggested that the likelihood-based selection criterion is preferred to the distance-based selection criterion (such as Kolmogorov–Smirnov distance) within the Lehmann family distributions. We will also employ the Information Complexity (ICOMP) criterion, a measure that penalizes the interdependencies among parameter estimators of Maximum Likelihood Estimation, as a supplemental tool, in addition to AIC and BIC, to identify the structural differences among widely used medium and heavy tailed distributions. In this research, we compare all of the above-mentioned selection indicators and consider the Anderson–Darling test, another goodness-of-fit test that puts more weight on the tails, to evaluate the model performance on a wider range of loss models.
The remainder of the paper is organized as follows. In Section 2, we present key probabilistic and log-likelihood functions of the Fisk, Frechet, Lognormal, Lomax, Paralogistic, and Weibull distributions. In Section 3, we discuss the Quantitile–Quantitile plot, Kolmogorov–Smirnov test, and Anderson–Darling test for model validation. Further, AIC, BIC, and ICOMP criteria will be discussed to select the model in Section 4. The simulation study has been conducted in Section 5. A real data analysis has been conducted in Section 6. Furthermore, the main results of the paper are summarized in Section 7, which concludes the paper.

2. Theoretical Derivations for Candidate Distributions

In this Section, we study six candidate loss distributions for medium and heavy tailed risks. They are Fisk, Frechet, Lognormal, Lomax, Paralogistic, and Weibull distributions. Let X represent the conditional random variable X | d X u .

2.1. Model Distribution with Left-Truncation and Right-Censoring

Candidate 1: Suppose random variable X follows a Fisk distribution (see Fisk (1961), for example) with scale parameter θ and shape parameter α . The pdf and cdf for a ground-up Fisk loss distribution are:
PDF : f ( x ) = α θ α x α 1 ( x α + θ α ) 2 , x > 0 , α > 0 , θ > 0 CDF : F ( x ) = x α x α + θ α , x > 0 , α > 0 , θ > 0
This distribution is also called a Loglogistic distribution. It resembles the Lognormal near zero when the shape parameter is greater than one and behaves more like Lomax in the tail. The cdf of X representing the loss data with deductible (d) and policy limit (u) is:
F ( x ) = 0 ; if x d , F ( x ) F ( d ) 1 F ( d ) = ( x ) α d α ( x ) α + θ α ; if d < x < u , 1 if x u .
Further, the qf of X is:
F 1 ( p ) = θ d α + p θ α ( 1 p ) θ α 1 / α for 0 p p u , u for p u p 1 .
Note that p u = F ( u ) F ( d ) 1 F ( d ) = u θ α d α + θ α u α + θ α d θ α . Let H ( x ) = ( x ) α + θ α , then the log-likelihood function becomes
log L = i = 1 n log f ( x i ) 1 { d < x i < u } + log [ 1 F ( u ) ] 1 ( x i = u ) n log 1 F ( d ) = i = 1 n log α θ α x i ( α 1 ) H 2 ( x i ) 1 { d < x i < u } + log θ α H ( u ) 1 { x i = u } n log θ α H ( d ) = i = 1 n log α + ( α 1 ) log x 2 log H ( x i ) 1 { d < x i < u } log H ( u ) 1 { x i = u } + n log H ( d ) .
Candidate 2: The Frechet, known as the inverse Weibull distribution, is a special case of the generalized extreme value distribution with much fatter tails on the right. It has a ground-up pdf and cdf as follows:
PDF : f ( x ) = α ( θ / x ) α e ( θ / x ) α x , x > 0 , α > 0 , θ > 0 , CDF : F ( x ) = e ( θ / x ) α , x > 0 , θ > 0 .
Then, the cdf of X representing the loss data with deductible (d) and policy limit (u) is:
F ( x ) = 0 ; if x d , F ( x ) F ( d ) 1 F ( d ) = e ( θ / x ) α e ( θ / d ) α 1 e ( θ / d ) α , if d < x < u , 1 , if x u .
Further, the qf of X is:
F 1 ( p ) = θ log p + ( 1 p ) e ( θ / d ) α 1 / α , for 0 p p u , u , for p u p 1 .
Note that p u = F ( u ) F ( d ) 1 F ( d ) = e ( θ / u ) α e ( θ / d ) α 1 e ( θ / d ) α . Let S ( x ) = 1 F ( x ) = 1 e ( θ / x ) α , the log-likelihood function becomes
log L = i = 1 n log f ( x i ) 1 { d < x < u } + log [ 1 F ( u ) ] 1 { x = u } n log 1 F ( d ) = i = 1 n log [ α θ α x ( α 1 ) e ( θ / x ) θ ] 1 { d < x < u } + log S ( u ) 1 { x = u } n log S ( d ) = i = 1 n log α + α log θ ( α + 1 ) log x ( θ / x α 1 { d < x < u } + log S ( u ) 1 { x = u } n log S ( d ) .
Candidate 3: Suppose random variable X is distributed according to a Lognormal distribution with log-location parameter < μ < , and log-scale parameter σ > 0 . We will denote this fact as X LN μ , σ . Its pdf and cdf are:
PDF : f ( x ) = ( σ ( x ) ) 1 φ log ( x ) μ σ , x > 0 , CDF : F ( x ) = Φ log ( x ) μ σ , x > 0 .
Here, Φ , φ , Φ 1 denote the cdf, pdf, and qf of the standard normal distribution, respectively. Next, let us introduce the following abbreviations:
c d : = log ( d ) μ σ , c x : = log ( x ) μ σ , c u : = log ( u ) μ σ .
Then, the cdf of X representing the loss data with deductible and policy limit is
F ( x | d ; u ) = 0 , x d , Φ c x Φ c d 1 Φ c d , d < x < u , 1 , x u .
Further, the qf of X is:
F 1 ( p | d ; u ) = exp μ + σ Φ 1 p + ( 1 p ) Φ c d , 0 p < p u , u , p u p 1 .
Note that p u = [ Φ ( c u ) Φ ( c d ) ] / [ 1 Φ ( c d ) ] . Finally, the log-likelihood function is
log L = i = 1 n log ( σ ) log ( x i ) + log ( φ ( c x i ) ) 1 { d < x i < u } + log 1 Φ ( c u ) i = 1 n 1 { x i = u } n log 1 Φ ( c d ) .
Candidate 4: Suppose random variable X is distributed according to the Lomax distribution (see Lomax (1954), for example) (also known as Pareto Type II, see Arnold (2015), for example). The pdf and cdf for a ground-up Lomax loss distribution are:
PDF : f ( x ) = α θ α ( x + θ ) α + 1 , x > 0 , CDF : F ( x ) = 1 θ x + θ α , x > 0 .
Then, the cdf of X representing the loss data with deductible (d) and policy (u) limit is:
F ( x | d ; u ) = 0 ; if x d , 1 d + θ x + θ α ; if d < x < u , 1 , x u .
Further, the qf of X is:
F 1 ( p | d ; u ) = ( 1 p ) 1 / α ( d + θ ) θ , if 0 p < p u , u , if p u p 1 .
Note that p u = 1 d + θ u + θ α . The log-likelihood function becomes
log L = i = 1 n log α ( α + 1 ) log ( x i + θ ) 1 { d < x i < u } α log ( u + θ ) 1 { x i = u } + α log ( d + θ ) .
Candidate 5: Suppose random variable X is distributed according to Paralogistic distribution. The pdf and cdf for a ground-up Paralogistic loss distribution are:
PDF : f ( x ) = α 2 x α 1 θ α 2 θ α + x α α + 1 , x > 0 , CDF : F ( x ) = 1 θ α θ α + x α α , x > 0 .
Then, the cdf of X representing the loss data with deductible (d) and policy (u) limit is:
F ( x | d ; u ) = 0 , i f x d , F ( x ) F ( d ) 1 F ( d ) = 1 θ α + d α θ α + x α α , if d < x < u , 1 , if x u .
Further, the qf of X is:
F 1 ( p | d ; u ) = F 1 p + ( 1 p ) F ( d ) = θ α + d α θ α ( 1 p ) 1 / α ( 1 p ) 1 / α 1 / α , for 0 p < p u , u , for p u p 1 .
Note that p u = F ( u ) F ( d ) 1 F ( d ) = 1 θ α + d α θ α + u α α . Then, the log-likelihood function becomes
log L = i = 1 n log f ( x i ) 1 { ( d < x i < u ) } + α 2 log θ α log θ α + u α 1 { x i = u } n α 2 log θ α log θ α + d α = i = 1 n 2 log α + ( α 1 ) log x i ( α + 1 ) log θ α + x i α 1 { ( d < x i < u ) } α log θ α + u α 1 { x i = u } + n α log θ α + d α .
Candidate 6: Suppose random variable X follows a Weibull distribution. The pdf and cdf for a ground-up Weibull loss distribution are:
PDF : f ( x ) = α x x θ α e x / θ α , x > 0 , CDF : F ( x ) = 1 e x / θ α , x > 0 .
Then, the cdf of X representing the loss data with deductible (d) and policy (u) limit is:
F ( x | d ; u ) = 0 , if x d , F ( x ) F ( d ) 1 F ( d ) = 1 e d / θ α x / θ α , if d < x < u , 1 , if x u .
Further, the qf of X is:
F 1 ( p | d ; u ) = θ log ( 1 p ) + d θ α 1 / α , for 0 p < p u , u , for p u p 1 .
Note that p u = F ( u ) F ( d ) 1 F ( d ) = 1 e d / θ α u / θ α . Then, the log-likelihood function is
log L = i = 1 n log f ( x i ) 1 { d < x i < u ) } u θ α 1 { x i = u } + n d θ α = i = 1 n log α log x i + α log x i α log θ x i θ α 1 { d < x i < u ) } u θ α 1 { x i = u } + n d θ α .

2.2. MLE Approach

Parametric methods use the observed data x 1 , , x n and fully recognize their distributional properties. The Maximum Likelihood Estimation (MLE) approach is one of the most common estimation techniques. Parameter estimates are found by maximizing the following log-likelihood function:
log L θ | x 1 , , x n = log i = 1 n f ( x i | d ; u ) = log i = 1 n f ( x i ) 1 F ( d ) 1 { d < x i < u } 1 F ( u ) 1 F ( d ) 1 { x i = u }
= i = 1 n log f ( x i ) 1 { d < x i < u } n log 1 F ( d ) + log 1 F ( u ) i = 1 n 1 { x i = u } ,
where 1 { · } denotes the indicator function and u represents a point closing to u from the left.
Once parameter MLEs, θ ^ 1 , , θ ^ k , are available, the pth quantile estimate is found by plugging those MLE values into the parametric expression of F 1 ( p ) = h ( θ 1 , , θ k ) , where h ( θ 1 , , θ k ) is a function of parameters.

2.3. Fisher Information

Let θ ^ = θ ^ 1 , , θ ^ k denote the MLE of parameter θ = ( θ 1 , , θ k ) . Then, we have the Fisher information matrix I, with the entries given by
I i j = E log g ( X ) θ i · log g ( X ) θ j ,
where E [ · ] denotes the expectation operator, and g is the true underlying pdf.
The observed Fisher information represents the amount of information that the observed data provides about the parameter θ . Intuitively, a high Fisher information indicates that the data contain valuable information about the parameter, and it implies that the estimate of the parameter based on the data is likely to be more precise. On the other hand, a low Fisher information suggests that the data contain little information about the parameter, leading to higher uncertainty in the estimate. The variance–covariance matrix is related to the inverse of the Fisher information matrix.

3. Model Validation

A variety of graphic tools, hypothesis tests, and penalized likelihood criteria are employed to select the model and validate the performance.

3.1. Quantile–Quantile Plot

A Quantile–Quantile (QQ) plot is a graphical representation used to compare two probability distributions. To construct a QQ plot, we plot the sample distribution on the Y-axis and the theoretical distribution on the X-axis. When the points in the QQ plot align closely to a diagonal line at 45 degrees, it indicates a high level of agreement between the sample distribution and the theoretical distribution.
In order to avoid visual distortions due to large spacings between the most extreme observations, both axes can be measured on the logarithmic scale. The QQ plot is generated using the following formula.
log F ^ 1 u i + F ^ ( d ) ( 1 u i ) , log x ( i ) , i = 1 , , n ,
where F ^ ( d ) is the estimated parametric cdf evaluated at d, F ^ 1 is the estimated parametric qf, x ( 1 ) < < x ( n ) denote the ordered claim severities, and u i = ( i 0.5 ) / n is the quantile level.

3.2. Kolmogorov–Smirnov Test and Anderson–Darling Test

Besides the graphical visual check, multiple quantitative hypothesis tests can help compare the model performance and specify an appropriate fitted parametric distribution. Here, the Kolmogorov–Smirnov (KS) test and Anderson–Darling (AD) test are considered in the following model validation procedure.
For a KS test, the test statistic is
K S = max d x u | F n ( x ) F ^ ( x ) | ,
where d is the left-truncation point ( d = 0 if there is no truncation) and u is the right-censoring point ( u = if there is no censoring).
The AD test is similar to the KS test but uses a different measure of the difference between the two distribution functions. The test statistic is
A D = n d u [ F n ( x ) F ^ ( x ) ] 2 F ^ ( x ) [ 1 F ^ ( x ) ] f ( x ) d x .
The AD goodness-of-fit statistic measures the cumulative weighted quadratic distance (with more weight on the tails) between the empirical cdf and the parametrically estimated cdf. The computational formula for LTRC data is given by
AD n = n F ^ ( u ) + n i = 1 k i / n 2 log F ^ ( X ( i + 1 ) ) F ^ ( X ( i ) ) n i = 0 k 1 1 i / n 2 log 1 F ^ ( X ( i + 1 ) ) 1 F ^ ( X ( i ) ) ( n k ) 2 n log 1 F ^ ( u ) 1 F ^ ( X ( k ) ) ,
where the unique noncensored data points are d = x 0 < x 1 < x k < x k + 1 = u and F ^ ( u ) = F ^ ( u ) F ^ ( d ) / 1 F ^ ( d ) (see more details in Klugman et al. (2019, sct. 15.4)).

4. Model Selection Criteria and Validation Methods

4.1. Selection Criteria: AIC, BIC, and ICOMP

Akaike Information Criteria ( A I C ) is a statistical measure that can be used to assess and compare the goodness-of-fit and complexity of different statistical models. A I C is a useful tool to compare different models in the context of model selection and is relevant when dealing with multiple candidate models. Akaike (1973) has provided the formula as
A I C = 2 log f x n | θ ^ p + 2 p ,
where p is the number of parameters. Bayesian Information Criteria ( B I C ) is another statistical measure which is used to assess the quality of the goodness-of-fit and selection of models. Schwarz (1978) has provided the formula to compute B I C as
B I C = 2 log f x n | θ ^ p + p · log ( n ) .
Similar to AIC, the model with the lowest BIC value is considered the most appropriate choice. It balances the trade-off between the model accurately capturing the data and employing a concise number of model parameters. Moreover, similar to A I C and B I C , Bozdogan (1987, 1990) has introduced another metric, Information Complexity (ICOMP), and its corresponding value is given by
I C O M P = 2 log f x n | θ ^ p + 2 C Σ ( θ ^ p )
with
C Σ ( θ ^ p ) = s 2 log tr Σ ( θ ^ p ) s 1 2 log det Σ ( θ ^ p ) ,
with s, tr, and det denoting the rank, trace, and determinant of Σ (the variance–covariance matrix), respectively.

4.2. Validation Methods

As we know, the smaller that the value of a likelihood-based measure is, the better the performance of a parametric loss model. However, how much difference in a likelihood-based measure really matters and separates the models? Suppose we set up a null hypothesis H 0 : Model i is the parent candidate that fits a data set and uses BIC to select the candidate. Fabozzi et al. (2014, p. 403) defined Δ i = B I C i B I C min , the difference in BIC between the model i and the model with the smallest BIC, and then used the range of Δ i (see Table 1) to indicate the strength of evidence to take against the model i as the parent distribution.
From the definition, the value of BIC highly depends on the sample size n; the larger the sample size, the higher the scale of a BIC could be. Therefore, for the same difference Δ i = 10 , it is relatively easier to separate fitted models for a data set with sample n = 1000 rather than a data set with n = 10,000. To avoid such an impact of sample size on model selection and validation, we report two types of probabilities instead of the pure BIC values (or the differences) in the following simulation study.
  • Type 1—the probability of each parent distribution being correctly selected. Inspired by Mitra et al. (2021), we can fit all of the candidate models including the parent one to LTRC data, and then choose the model with the lowest BIC number as the best fit. We repeat this process N times and calculate the proportion of times each candidate is chosen as the best, resulting in the probability for each of the candidate models to be selected. The same process is followed for other likelihood-based criteria: AIC and ICOMP, and the distance-based measures, the KS statistic and AD statistic.
  • Type 2—the posterior model probabilities. Suppose we only have k models at hand and assume that at the beginning, each model has the same chance ( 1 / k ) being the underlying distribution. Then, the difference in likelihood can update the possibility of a model being selected and provides an insight about the best candidate model. Defined in Raftery (1995) and Burnham and Anderson (2004), the posterior probability of model i is
    p i = exp ( 1 2 Δ i ) j = 1 k exp ( 1 2 Δ j ) .
    Here, the constant 1 / 2 in the exponent will balance the multiplier 2 in the BIC definition, and the ratio structure will reduce the impact of sample size on BIC comparison. The posterior probability of AIC and ICOMP can be calculated using the same formula, and we just need to change the term Δ as the difference for those two criteria.

5. Simulation Study

In this section, we conduct an analysis of model uncertainty due to both left-truncation and right-censoring structure, and investigate the effects of various combinations of proportions on the model selection for the above-mentioned loss models. We fix the deductible d and the policy limit u for all models and tune the parameter values of each model so that F ( d ) and F ( u ) are both pre-specified constants. The parameter values can be determined using the percentile matching at two percentile levels, namely F ( d ) and F ( u ) . The data are then generated according to those parameter values of an underlying distribution and the MLE estimations are conducted to fit the data to each of the candidate models. Multiple model selection rules are implemented to specify the parent distribution of data and their performances are compared in terms of selection accuracy and likelihood probabilities, ultimately.
The simulation study is designed with the following setups:
(i)
Five selected loss models: Fisk, Lognormal, Lomax, Paralogistic, Weibull
(ii)
Sample size: n = 1000 and 100,000
(iii)
Truncation and censoring thresholds: d = 500 , u = 10,000
(iv)
Truncation and non-censoring probabilities ( F ( d ) , F ( u ) ) : (0.10, 0.85) and (0.50, 0.80)
(v)
Model selection indicators: KS statistic, AD statistic, AIC, BIC, ICOMP
NOTE: Frechet is not included in the simulation study due to the instability of parameter estimation.
The choice of sample size, truncation and censoring thresholds, truncation probability and censoring probability depends on practical considerations and research objectives. A sample size of 1000 can be considered moderate for many statistical studies. It is often chosen due to constraints on resources. It can provide a reasonable level of statistical power to detect meaningful effects. A sample size of 100,000 is quite large and may be chosen for studies where researchers need to achieve a very high level of statistical power, detect small effect sizes, and adequately capture rare events. A deductible of 500 and a policy limit of 10,000 are very common in the insurance industry, which provides a practical foundation for our choice of truncation threshold being 500 and censoring threshold being 10,000. A truncation probability of 10% is relatively low in insurance, which means there is not too much information loss due to truncation but it is useful for comparison with another scenario of more information loss due to a higher truncation probability. A truncation probability of 50% is high in theory but quite ordinary in insurance practice, due to the fact of many loss events with magnitudes below deductibles. Both a censoring probability of 15% and that of 20% imply that a substantial portion of the subjects experience censoring during the study. This could be chosen based on prior knowledge or to simulate realistic scenarios. The choice of these parameters depends on the balance between theoretical rigor, practical considerations, and the objectives of the study.
Table 2, Table 3, Table 4 and Table 5 demonstrate how often LTRC data can correctly identify its underlying distribution with various combinations of left-truncation and right-censoring proportions and how the sample size affects correct identification. The selecting probabilities of a model within all indicator measurements (KS, AD, AIC, BIC, and ICOMP) are recorded as the first value in each cell and the likelihood-based approach also reports posterior probability in additional parenthesis, where the former number is the mean of 100 simulations while the latter one is the median. The computation is run through Matlab R2023b.
The records of a best-case scenario of F ( d ) = 0.1 and F ( u ) = 0.85 are displayed in Table 2 and Table 3, in which 75% of the data are used to estimate the parameters and validate the model performance. Overall, each model is distinguishable from the others. When the sample size is n = 1000 in Table 2, the likelihood-based indicators AIC and BIC are significantly better than the distance-based indicators KS and AD on identifying correct distribution for Fisk and Lomax data, whereas the opposite performance is observed for Weibull data. In particular, when all five candidate distributions are used to fit Weibull data, KS and AD have 59% and 50% chances, respectively, to correctly select the underlying Weibull distribution, while AIC and BIC only have around 20% to 23% correction rate. Meanwhile, ICOMP always specifies Lognormal model no matter what the underlying distribution, thus it is not a reliable selection criterion under this situation. Moreover, selection probabilities and posterior probabilities are generally consistent to indicate the best fit, which means that when a selection probability is large enough to specify a candidate model, the corresponding posterior probability will move to a significantly higher value than the benchmark of 20%.
As the sample size increases to n = 100,000, all measures work very well (with at least 88% selection accuracy) in Table 3. Although there is no significant difference in selection performance, we observe that AD and ICOMP are slightly better for identifying Weibull data. Posterior probability and selection probability are almost identical within each likelihood approach comparison and the median posterior even could specify a correct parent distribution with 100% probability.
In a worst-case scenario of F ( d ) = 0.5 and F ( u ) = 0.8 in Table 4, we only use 30% of the middle values to estimate the parameters and identify the parent distributions. As a result, the true underlying model is generally not distinguishable from other candidates when n = 1000 . It is clear that AIC and BIC always tend to pick up the Fisk model as the best fit no matter what the parent model is, except for Weibull as the parent distribution. In particular, if the data are generated from the Paralogistic distribution, there is no chance for AIC and BIC to identify it, since the probabilities that the Paralogistic model is chosen as the best candidate are 0% and 2%, respectively. Meanwhile, KS and AD prefer to select Weibull, but this tendency overall is not as strong as the likelihood-based measures, AIC and BIC, to select Fisk. These outcomes give rise to a scenario of model uncertainty, wherein posterior probabilities appear to fluctuate equally around 20%, leading to no signal for model preference. As the sample size is increased to 100,000 in Table 5, all measures except ICOMP work reasonably well for separating models. The correct selection rate, especially for Paralogistic data, has moved up from 0% to 66% in likelihood-based measures, although there is still an 8% gap to the distance-based indicators.

6. Real Data Illustration

Next, a real data analysis is conducted to illustrate the procedure of parametric model selection when the loss deductible and policy limit are under consideration. The target data are the Local Government Property Insurance Fund (LGPIF), an insurance pool administered by the Wisconsin Office of the Insurance Commissioner, which has been studied by Frees (2018). This insurance fund covers local government properties that include counties, cities, towns, villages, school districts, and library boards. Our goal is to investigate what effect initial assumptions have on the MLE estimates with various typical loss models and how to pick up an appropriate candidate to predict the claims for potential risk management.
With a 500 deductible ( d = 500 ), 1890 LGPIF loss payments were recorded in the portfolio and the histogram of losses is displayed in Figure 1. It is clear that the loss values resemble a heavy-tailed distribution even after the logarithmic transformation. The above-mentioned Fisk, Frechet, Lognormal, Lomax, Paralogistic, and Weibull could be good candidates to investigate the uncertainty of claim payment for this insurance fund.
This fund covers seven different types of property. Due to the various ranges of loss magnitude, we set a flexible policy limit for each individual type, stopping at the 80th percentile of the loss (that is F ( u ) = 0.8 ), respectively. The estimated parameters and the model selection indicators are recorded in Table 6. All values of AD, AIC, and BIC are the lowest for Frechet model and the differences in measure indicators are tremendously significant. This indicates a signal to prepare for risk management, that is, LGPIF may have an extremely heavy tail to their potential claim payment.
We also present plots of the fitted-versus-observed quantiles for the Fisk, Frechet, Lognormal, Lomax, Paralogistic, and Weibull distributions in Figure 2. Clearly, Frechet-estimated quantiles fall almost perfectly on the 45 line against the empirical quantiles. On the other hand, Fisk, Lognormal, Lomax, Paralogistic, and Weibull QQ plots do not look as good, and having bottom observations above the 45 line indicates that these models underestimate the right tail of the data.

7. Discussion and Conclusions

We have presented our research on model uncertainty and selection in the modeling of insurance loss data using left-truncated and right-censored (LTRC) distributions. We have considered six candidate distributions: Fisk, Frechet, Lognormal, Lomax, Paralogistic, and Weibull. Maximum Likelihood Estimation (MLE)-based approaches are used to estimate the parameters of the distributions under consideration.
A variety of model validation methods, including Quantile–Quantile (QQ) plots, Kolmogorov–Smirnov (KS) and Anderson–Darling (AD) statistics are employed. These tools are critical for assessing how well each distribution fits the data. Model selection adopts information criteria such as Akaike Information Criterion (AIC), Bayesian Information Criterion (BIC), and Information Complexity (ICOMP). These criteria help us choose the most suitable model while penalizing the complexity of the model.
A simulation study has been conducted to evaluate the performance of each model validation method and model selection criterion under different scenarios of left-truncation and right-censoring. This is achieved through calculating both selection probability and posterior probability for each candidate model. Uncertainty has been observed in the patterns regarding which model validation method or model selection criterion is outperforming others, depending on a lot of factors, such as sample size, the underlying true distribution, left-truncation proportion, and right-censoring proportion, etc.
In the simulation study, we have noticed that multiple models may be difficult to distinguish, and various tests and/or criteria may lead to the selection of different models being the best. It seems that distance-based criteria tend to select the Weibull model and likelihood-based criteria prefer to select the Fisk model among the considered candidate models. Then, model averaging can be considered at this time. Model averaging can provide more robust estimates, especially when there is uncertainty in identifying the true underlying distribution.
Analyzing actual local government property loss data in Wisconsin serves as a practical application of our modeling methodology. Based on the outputs of the model selection indicators and the QQ plots, we identify Frechet distribution as the best fit to LGPIF, which is a significant finding, especially given its relevance for modeling extreme events.
In summary, our study combines theoretical and empirical approaches to analyze LTRC insurance loss data. The use of various statistical tools and criteria helps ensure the reliability of our modeling results, which exhibit a certain degree of uncertainty in the model selection procedure illustrated through both simulation and case studies. Through our research in this paper, we would like to give the following recommendations in regard to uncertainty in the model selection procedure, using our selection probability framework. If KS and AD can provide strong evidence to select a best candidate model, then it is very likely that AIC and BIC will come up with the same decision. When KS and AD can only provide fairly weak evidence for the selection of a best model, then we may consider further use of AIC and BIC to reinforce the decision if a common best model has been selected by AIC/BIC as KS/AD. Otherwise, when the use of AIC and BIC leads to the selection of a different best model from the use of KS and AD, we may consider model averaging of two or more competing best models.
Our comprehensive study of modeling LTRC insurance loss data addresses various aspects of model uncertainty and selection, but there are also areas where some limitations and potential avenues for future research can be discussed.
There are some limitations to the assumptions about candidate distributions. Our study considers six candidate distributions for modeling insurance loss data. However, it is worth noting that these distributions may not capture all possible underlying distributions. There are other limitations to the sensitivity analysis. We have already mentioned that the performance of model validation methods and selection criteria will vary with factors such as sample size, and left-truncation and right-censoring ratios. It would be useful to further quantify specific sensitivity levels. This information can help practitioners better understand the applicability of our methodology in various real-world scenarios.
In addition to exploring other distributions and conducting more quantitative sensitivity analysis, future research avenues could include, but are not limited to, robust model averaging, out-of-sample testing, incorporation of expert opinions, quantification of model uncertainty, and applications in other areas. First, we mention that model averaging can provide more robust estimates when there is uncertainty in identifying the true underlying distribution. Future research could delve deeper into methods and techniques for robust model averaging, exploring different weighting schemes or Bayesian model averaging. Second, while we have conducted simulation studies and applied our methods to real data, future research could focus on out-of-sample testing to evaluate how well the selected models generalize to new, unseen data. Third, incorporating expert opinions or domain knowledge into the model selection process can improve the relevance and accuracy of the selected distributions in practical applications. Fourth, it seems interesting to develop techniques that more explicitly quantify and visualize model uncertainty which can help decision makers understand the range of potential outcomes based on different candidate distributions. Finally, we can explore how our modeling approach can be applied to areas beyond insurance loss data, such as finance, healthcare, or environmental modeling.
Incorporating these limitations and future research considerations will help to enrich our work and provide a more complete understanding of the challenges and opportunities of modeling insurance loss data using LTRC distributions. It will also enhance the practical applicability of our findings.

Author Contributions

Conceptualization, Q.Z., S.U. and D.Y.; Methodology, Q.Z., S.U. and D.Y.; Software, Q.Z., S.U. and D.Y.; Validation, Q.Z., S.U. and D.Y.; Formal analysis, Q.Z., S.U. and D.Y.; Data curation, Q.Z., S.U. and D.Y.; Writing—original draft, Q.Z., S.U. and D.Y.; Writing—review and editing, Q.Z., S.U. and D.Y.; Visualization, Q.Z., S.U. and D.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Not applicable.

Acknowledgments

The authors are very appreciative of valuable insights and useful comments provided by Vytaras Brazauskas, which helped to improve the paper significantly.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Arnold, Barry C. 2015. Pareto Distributions, 2nd ed. Boca Raton: Chapman and Hall/CRC. [Google Scholar]
  2. Balakrishnan, N., and Debanjan Mitra. 2014. EM-based likelihood inference for some lifetime distributions based on left truncated and right censored data and associated model discrimination. South African Statistical Journal 48: 125–71. [Google Scholar]
  3. Bozdogan, Hamparsum. 1987. Model Selection and Akaike’s Information Criterion (AIC): The General Theory and Its Analytical Extensions. Psychometrika 52: 345–70. [Google Scholar] [CrossRef]
  4. Bozdogan, Hamparsum. 1990. On the information based measure of covariance complexity and its application to the evaluation of multivariate linear models. Communications in Statistics: Theory and Methods 19: 221–78. [Google Scholar] [CrossRef]
  5. Brazauskas, Vytaras, and Sahadeb Upretee. 2019. Model Efficiency and Uncertainty in Quantile Estimation of Loss Severity Distributions. Risks 7: 55. [Google Scholar] [CrossRef]
  6. Burnham, Kenneth P., and David R. Anderson. 2004. Multimodel Inference: Understanding AIC and BIC in Model Selection. Sociological Methods & Research 33: 261–304. [Google Scholar]
  7. Emura, Takeshi, and Hirofumi Michimae. 2022. Left-truncated and right-censored field failure data: Review of parametric analysis for reliability. Quality and Reliability Engineering International 38: 3919–34. [Google Scholar] [CrossRef]
  8. Fabozzi, Frank, Sergio Focardi, Teodosii Rachev, and Bala Arshanapalli. 2014. The Basics of Financial Econometrics: Tools, Concepts, and Asset Management Applications. Hoboken: Wiley. [Google Scholar]
  9. Fisk, Peter R. 1961. The graduation of income distributions. Econometrica 29: 171–85. [Google Scholar] [CrossRef]
  10. Frees, Edward. 2018. Loss Data Analytics. An Open Text Authored by the Actuarial Community. Available online: https://openacttexts.github.io/Loss-Data-Analytics/ (accessed on 1 October 2023).
  11. Klugman, Stuart A., Harry H. Panjer, and Gordon E. Willmot. 2019. Loss Models: From Data to Decisions. Wiley Series in Probability and Statistics. Hoboken: Wiley. [Google Scholar]
  12. Lomax, K. S. 1954. Business failures: Another example of the analysis of failure data. Journal of the American Statistical Association 49: 847–52. [Google Scholar] [CrossRef]
  13. Mitra, Debanjan. 2012. Likelihood Inference for Left Truncated and Right Censored Lifetime Data. Ph.D. thesis, McMaster University, Hamilton, ON, Canada. [Google Scholar]
  14. Mitra, Debanjan, Debasis Kundu, and Narayanaswamy Balakrishnan. 2021. Likelihood analysis and stochastic EM algorithm for left truncated right censored data and associated model selection from the Lehmann family of life distributions. Japanese Journal of Statistics and Data Science 4: 1019–48. [Google Scholar] [CrossRef]
  15. Poudyal, Chudamani, Qian Zhao, and Vytaras Brazauskas. 2023. Method of Winsorized Moments for Robust Fitting of Truncated and Censored Lognormal Distributions. North American Actuarial Journal, 1–25. [Google Scholar] [CrossRef]
  16. Raftery, Adrian E. 1995. Bayesian Model Selection in Social Research. Sociological Methodology 25: 111. [Google Scholar] [CrossRef]
Figure 1. Loss data of LGPIF when the deductible is 500.
Figure 1. Loss data of LGPIF when the deductible is 500.
Risks 11 00188 g001
Figure 2. QQ plot of the fitted models to the real data example.
Figure 2. QQ plot of the fitted models to the real data example.
Risks 11 00188 g002aRisks 11 00188 g002b
Table 1. Strength of evidence to take against the model i.
Table 1. Strength of evidence to take against the model i.
Range of Δ i Evidence against a Candidate
[ 0 , 2 ] little
( 2 , 6 ] positive
( 6 , 10 ] strong
( 10 , ) very strong
Table 2. Shown are 100 simulations of Fisk data ( α = 1.31 , θ = 2667 ); Logn data ( μ = 7.87 , θ = 1.29 ); Lomax data ( α = 8.69 , θ = 41,007); Paralog data ( α = 1.24 , θ = 3533 ); Weibull data ( α = 0.96 , θ = 5150 ); d = 500 , u = 10,000, n = 1000 . F ( d ) = 0.10 , F ( u ) = 0.85 .
Table 2. Shown are 100 simulations of Fisk data ( α = 1.31 , θ = 2667 ); Logn data ( μ = 7.87 , θ = 1.29 ); Lomax data ( α = 8.69 , θ = 41,007); Paralog data ( α = 1.24 , θ = 3533 ); Weibull data ( α = 0.96 , θ = 5150 ); d = 500 , u = 10,000, n = 1000 . F ( d ) = 0.10 , F ( u ) = 0.85 .
FiskLognLomaxParalogWeibull
KS 0.54 0.150.100.210.00
AD0.560.150.060.230.00
FiskAIC0.72 ( 0.38 , 0.36 ) 0.19 (0.21, 0.20)0.03 (0.12, 0.10)0.06 (0.29, 0.30)0.00 (0.01, 0.00)
DataBIC0.72 (0.37, 0.36)0.18 (0.21, 0.20)0.05 (0.12, 0.10)0.06 (0.29, 0.30)0.00 (0.01, 0.00)
ICOMP0.00 (0.00, 0.00)1.00 (1.00, 1.00)0.00 (0.00, 0.00)0.00 (0.00, 0.00)0.00 (0.00, 0.00)
KS0.260.160.270.180.13
AD0.270.300.170.150.11
LognAIC0.33 (0.20, 0.16)0.43 (0.26, 0.27)0.12 (0.23, 0.23)0.03 (0.21, 0.23)0.09 (0.11, 0.02)
DataBIC0.33 (0.19, 0.16)0.44 (0.26, 0.26)0.10 (0.23, 0.23)0.02 (0.20, 0.23)0.11 (0.11, 0.02)
ICOMP0.00 (0.00, 0.00)1.00 (1.00, 1.00)0.00 (0.00, 0.00)0.00 (0.00, 0.00)0.00 (0.00, 0.00)
KS0.000.020.420.140.42
AD0.000.020.390.140.45
LomaxAIC0.00 (0.02, 0.00)0.06 (0.05, 0.01)0.73 (0.45, 0.48)0.06 (0.09, 0.02)0.15 (0.40, 0.48)
DataBIC0.00 (0.02, 0.00)0.06 (0.05, 0.01)0.68 (0.45, 0.48)0.05 (0.09, 0.02)0.21 (0.40, 0.47)
ICOMP0.00 (0.00, 0.00)0.93 (0.86, 0.98)0.03 (0.04, 0.00)0.04 (0.00, 0.01)0.05 (0.10, 0.01)
KS0.390.220.170.210.01
AD0.370.290.120.220.00
ParalogAIC0.53 (0.29, 0.29)0.27 (0.26, 0.26)0.11 (0.16, 0.12)0.09 (0.28, 0.28)0.00 (0.02, 0.00)
DataBIC0.51 (0.29, 0.27)0.30 (0.26, 0.27)0.07 (0.15, 0.12)0.12 (0.28, 0.29)0.00 (0.02, 0.00)
ICOMP0.00 (0.00, 0.00)1.00 (1.00, 1.00)0.00 (0.00, 0.00)0.00 (0.00, 0.00)0.00 (0.00, 0.00)
KS0.000.010.320.080.59
AD0.000.000.420.080.50
WeibullAIC0.00 (0.01, 0.00)0.00 (0.02, 0.00)0.75 (0.46, 0.50)0.05 (0.06, 0.01)0.20 (0.46, 0.50)
DataBIC0.00 (0.01, 0.00)0.00 (0.02, 0.00)0.72 (0.45, 0.49)0.05 (0.06, 0.01)0.23 (0.46, 0.50)
ICOMP0.00 (0.00, 0.00)0.45 (0.61, 0.82)0.21 (0.21, 0.00)0.00 (0.00, 0.00)0.14 (0.18, 0.02)
NOTE: Probability of selecting a model in various parent data distributions. The likelihood-based approach has additional posterior probability in parenthesis ( a , b ) , where a represents the mean, and b is the median.
Table 3. Shown are 100 simulations of Fisk data ( α = 1.31 , θ = 2667 ); Logn data ( μ = 7.87 , θ = 1.29 ); Lomax data ( α = 8.69 , θ = 41,007); Paralog data ( α = 1.24 , θ = 3533); Weibull data ( α = 0.96 , θ = 5150); d = 500 , u = 10,000, n = 100,000. F ( d ) = 0.10 , F ( u ) = 0.85 .
Table 3. Shown are 100 simulations of Fisk data ( α = 1.31 , θ = 2667 ); Logn data ( μ = 7.87 , θ = 1.29 ); Lomax data ( α = 8.69 , θ = 41,007); Paralog data ( α = 1.24 , θ = 3533); Weibull data ( α = 0.96 , θ = 5150); d = 500 , u = 10,000, n = 100,000. F ( d ) = 0.10 , F ( u ) = 0.85 .
FiskLognLomaxParalogWeibull
KS0.960.000.000.040.00
AD0.960.000.000.040.00
FiskAIC0.96 (0.96, 1.00)0.00 (0.00, 0.00)0.00 (0.00, 0.00)0.04 (0.04, 0.04)0.00 (0.00, 0.00)
DataBIC0.96 (0.96, 1.00)0.00 (0.00, 0.00)0.00 (0.00, 0.00)0.04 (0.04, 0.04)0.00 (0.00, 0.00)
ICOMP0.97 (0.96, 1.00)0.00 (0.00, 0.00)0.00 (0.00, 0.00)0.03 (0.04, 0.00)0.00 (0.00, 0.00)
KS0.000.940.050.010.00
AD0.000.990.010.000.00
LognAIC0.00 (0.00, 0.00)0.99 (0.98, 1.00)0.01 (0.02, 0.00)0.00 (0.00, 0.00)0.00 (0.00, 0.00)
DataBIC0.00 (0.00, 0.00)0.99 (0.98, 1.00)0.01 (0.02, 0.00)0.00 (0.00, 0.00)0.00 (0.00, 0.00)
ICOMP0.00 (0.00, 0.00)1.00 (1.00, 1.00)0.00 (0.00, 0.00)0.00 (0.00, 0.00)0.00 (0.00, 0.00)
KS0.000.000.980.000.02
AD0.000.000.980.000.02
LomaxAIC0.00 (0.00, 0.00)0.00 (0.00, 0.00)0.99 (0.99, 1.00)0.00 (0.00, 0.00)0.01 (0.01, 0.00)
DataBIC0.00 (0.00, 0.00)0.00 (0.00, 0.00)0.99 (0.99, 1.00)0.00 (0.00, 0.00)0.01 (0.01, 0.00)
ICOMP0.00 (0.00, 0.00)0.00 (0.00, 0.00)0.95 (0.96, 1.00)0.00 (0.00, 0.00)0.05 (0.04, 0.00)
KS0.020.010.000.970.00
AD0.010.010.000.980.00
ParalogAIC0.02 (0.02, 0.00)0.01 (0.01, 0.00)0.00 (0.00, 0.00)0.97 (0.98, 1.00)0.00 (0.00, 0.00)
DataBIC0.02 (0.02, 0.00)0.01 (0.01, 0.00)0.00 (0.00, 0.00)0.97 (0.98, 1.00)0.00 (0.00, 0.00)
ICOMP0.02 (0.02, 0.00)0.07 (0.07, 0.00)0.00 (0.00, 0.00)0.91 (0.91, 1.00)0.00 (0.00, 0.00)
KS0.000.000.120.000.88
AD0.000.000.080.000.92
WeibullAIC0.00 (0.00, 0.00)0.00 (0.00, 0.00)0.11 (0.14, 0.04)0.00 (0.00, 0.00)0.89 (0.86, 0.96)
DataBIC0.00 (0.00, 0.00)0.00 (0.00, 0.00)0.11 (0.14, 0.04)0.00 (0.00, 0.00)0.89 (0.86, 0.96)
ICOMP0.00 (0.00, 0.00)0.00 (0.00, 0.00)0.01 (0.01, 0.00)0.00 (0.00, 1.00)0.99 (0.99, 1.00)
Table 4. Shown are 100 simulations of Fisk data ( α = 0.46 , θ = 500 ); Lognormal data ( μ = 6.21 , θ = 3.56 ); Lomax data ( α = 0.32 , θ = 63.91 ); Paralogistic data ( α = 0.61 , θ = 144.04 ); Weibull data ( α = 0.28 , θ = 184.09 ); d = 500 , u = 100,000, n = 1000 . F ( d ) = 0.50 , F ( u ) = 0.80 .
Table 4. Shown are 100 simulations of Fisk data ( α = 0.46 , θ = 500 ); Lognormal data ( μ = 6.21 , θ = 3.56 ); Lomax data ( α = 0.32 , θ = 63.91 ); Paralogistic data ( α = 0.61 , θ = 144.04 ); Weibull data ( α = 0.28 , θ = 184.09 ); d = 500 , u = 100,000, n = 1000 . F ( d ) = 0.50 , F ( u ) = 0.80 .
FiskLognLomaxParalogWeibull
KS0.110.110.220.100.46
AD0.130.110.260.090.41
FiskAIC0.56 (0.20, 0.20)0.16 (0.22, 0.22)0.16 (0.17, 0.16)0.00 (0.19, 0.20)0.12 (0.22, 0.22)
DataBIC0.55 (0.20, 0.20)0.16 (0.21, 0.22)0.17 (0.18, 0.16)0.01 (0.19, 0.19)0.11 (0.23, 0.22)
ICOMP0.00 (0.00, 0.00)1.00 (0.99, 1.00)0.00 (0.00, 0.00)0.00 (0.00, 0.00)0.00 (0.00, 0.00)
KS0.050.120.280.170.48
AD0.060.080.310.080.47
LognAIC0.46 (0.20, 0.21)0.14 (0.22, 0.22)0.24 (0.17, 0.14)0.00 (0.18, 0.20)0.16 (0.23, 0.22)
DataBIC0.40 (0.20, 0.20)0.22 (0.22, 0.22)0.23 (0.18, 0.14)0.00 (0.18, 0.20)0.15 (0.00, 0.00)
ICOMP0.00 (0.00, 0.00)1.00 (1.00, 1.00)0.00 (0.00, 0.00)0.00 (0.00, 0.00)0.00 (0.00, 0.00)
KS0.350.070.230.160.19
AD0.270.050.430.110.14
LomaxAIC0.80 (0.23, 0.22)0.05 (0.14, 0.13)0.15 (0.24, 0.25)0.00 (0.23, 0.22)0.10 (0.17, 0.16)
DataBIC0.76 (0.23, 0.22)0.07 (0.14, 0.13)0.12 (0.23, 0.24)0.00 (0.22, 0.22)0.05 (0.18, 0.18)
ICOMP0.07 (0.07, 0.00)0.70 (0.71, 0.96)0.00 (0.00, 0.00)0.23 (0.21, 0.00)0.00 (0.01, 0.00)
KS0.110.150.280.120.34
AD0.170.090.370.080.29
ParalogAIC0.71 (0.20, 0.20)0.11 (0.20, 0.20)0.15 (0.20, 0.20)0.00 (0.20, 0.20)0.03 (0.20, 0.20)
DataBIC0.64 (0.20, 0.20)0.10 (0.19, 0.20)0.18 (0.20, 0.20)0.02 (0.20, 0.20)0.06 (0.20, 0.20)
ICOMP0.02 (0.02, 0.00)0.96 (0.95, 1.00)0.00 (0.00, 0.00)0.02 (0.02, 0.00)0.00 (0.02, 0.00)
KS0.080.090.050.140.64
AD0.040.130.070.090.67
WeibullAIC0.39 (0.19, 0.20)0.16 (0.24, 0.23)0.04 (0.10, 0.08)0.01 (0.16, 0.15)0.40 (0.31, 0.29)
DataBIC0.35 (0.19, 0.20)0.16 (0.24, 0.24)0.006(0.10, 0.08)0.01 (0.16, 0.15)0.42 (0.31, 0.29)
ICOMP0.00 (0.00, 0.00)1.00 (1.00, 1.00)0.00 (0.00, 0.00)0.00 (0.00, 0.00)0.00 (0.00, 0.00)
Table 5. Shown are 100 simulations of Fisk data ( α = 0.46 , θ = 500 ); Lognormal data ( μ = 6.21 , θ = 3.56 ); Lomax data ( α = 0.32 , θ = 63.91 ); Paralogistic data ( α = 0.61 , θ = 144.04 ); Weibull data ( α = 0.28 , θ = 184.09 ); d = 500 , u = 100,000, n = 100,000. F ( d ) = 0.50 , F ( u ) = 0.80 .
Table 5. Shown are 100 simulations of Fisk data ( α = 0.46 , θ = 500 ); Lognormal data ( μ = 6.21 , θ = 3.56 ); Lomax data ( α = 0.32 , θ = 63.91 ); Paralogistic data ( α = 0.61 , θ = 144.04 ); Weibull data ( α = 0.28 , θ = 184.09 ); d = 500 , u = 100,000, n = 100,000. F ( d ) = 0.50 , F ( u ) = 0.80 .
FiskLognLomaxParalogWeibull
KS0.670.100.000.230.00
AD0.700.080.000.220.00
FiskAIC0.75 (0.62, 0.72)0.09 (0.13, 0.02)0.00 (0.00, 0.00)0.16 (0.24, 0.18)0.00 (0.01, 0.00)
DataBIC0.77 (0.62, 0.72)0.09 (0.13, 0.02)0.00 (0.00, 0.00)0.14 (0.24, 0.18)0.00 (0.01, 0.00)
ICOMP0.00 (0.02, 0.01)0.99 (0.93, 0.99)0.00 (0.00, 0.00)0.11 (0.05, 0.00)0.00 (0.00, 0.00)
KS0.080.780.000.000.14
AD0.080.780.000.000.14
LognAIC0.09 (0.12, 0.03)0.83 (0.67, 0.73)0.00 (0.00, 0.00)0.00 (0.001, 0.000)0.08 (0.21, 0.15)
DataBIC0.09 (0.12, 0.03)0.83 (0.68, 0.73)0.00 (0.00, 0.00)0.00 (0.00, 0.00)0.08 (0.21, 0.12)
ICOMP0.00 (0.00, 0.00)1.00 (1.00, 1.00)0.00 (0.00, 0.00)0.00 (0.00, 0.00)0.00 (0.00, 0.00)
KS0.060.000.760.180.00
AD0.040.000.760.200.00
LomaxAIC0.14 (0.14, 0.09)0.00 (0.00, 0.00)0.81 (0.59, 0.67)0.05 (0.27, 0.25)0.00 (0.00, 0.00)
DataBIC0.14 (0.14, 0.09)0.00 (0.00, 0.00)0.81 (0.59, 0.67)0.05 (0.27, 0.25)0.00 (0.00, 0.00)
ICOMP0.42 (0.29, 0.25)0.00 (0.00, 0.00)0.58 (0.44, 0.45)0.00 (0.27, 0.27)0.00 (0.00, 0.00)
KS0.190.000.070.740.00
AD0.180.000.080.740.00
ParalogAIC0.23 (0.23, 0.18)0.00 (0.02, 0.00)0.11 (0.12, 0.02)0.66 (0.63, 0.71)0.00 (0.00, 0.00)
DataBIC0.23 (0.24, 0.18)0.00 (0.02, 0.00)0.11 (0.12, 0.02)0.66 (0.63, 0.69)0.00 (0.00, 0.00)
ICOMP0.00 (0.03, 0.03)0.65 (0.57, 0.59)0.14 (0.13, 0.01)0.21 (0.28, 0.26)0.00 (0.00, 0.00)
KS0.000.020.000.000.98
AD0.000.030.000.000.97
WeibullAIC0.00 (0.00, 0.00)0.03 (0.06, 0.00)0.00 (0.00, 0.00)0.00 (0.00, 0.00)0.97 (0.94, 1.00)
DataBIC0.00 (0.00, 0.00)0.03 (0.06, 0.00)0.00 (0.00, 0.00)0.00 (0.00, 0.00)0.97 (0.94, 1.00)
ICOMP0.00 (0.00, 0.00)0.90 (0.86, 0.99)0.00 (0.00, 0.00)0.00 (0.00, 0.00)0.10 (0.14, 0.01)
Table 6. Real data illustration with d = 500 , and empirical F ( u ) = 0.8 .
Table 6. Real data illustration with d = 500 , and empirical F ( u ) = 0.8 .
FiskFrechetLognLomaxParalogWeibull
θ ^ 2744.42125.47.983221,9993912.43731.9
α ^ 1.86611.42960.84536.74641.53031.051
KS0.07410.06780.06650.11610.08650.0999
AD15.47026.658317.651330.628819.292427.5664
AIC27,66927,47227,66427,82127,72027,828
BIC27,68027,48327,67827,83327,73127,839
ICOMP27,678 27,66327,83827,73227,839
F ( d ) ^ , F ( u ) ^ 0.04, 0.84400.0004, 0.82670.0182, 0.83950.1407, 0.83680.0623, 0.84070.1139, 0.8465
Note: θ ^ and α ^ are estimates based on the formulas in Section 2. For Lognormal distribution, θ ^ represents μ ^ and α ^ represents σ ^ .
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, Q.; Upretee, S.; Yu, D. Model Uncertainty and Selection of Risk Models for Left-Truncated and Right-Censored Loss Data. Risks 2023, 11, 188. https://doi.org/10.3390/risks11110188

AMA Style

Zhao Q, Upretee S, Yu D. Model Uncertainty and Selection of Risk Models for Left-Truncated and Right-Censored Loss Data. Risks. 2023; 11(11):188. https://doi.org/10.3390/risks11110188

Chicago/Turabian Style

Zhao, Qian, Sahadeb Upretee, and Daoping Yu. 2023. "Model Uncertainty and Selection of Risk Models for Left-Truncated and Right-Censored Loss Data" Risks 11, no. 11: 188. https://doi.org/10.3390/risks11110188

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop