Next Article in Journal
Spatio-Functional Local Linear Asymmetric Least Square Regression Estimation: Application for Spatial Prediction of COVID-19 Propagation
Previous Article in Journal
A Novel Concept of Level Graph in Interval-Valued Fuzzy Graphs with Application
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A New Effective Jackknifing Estimator in the Negative Binomial Regression Model

Department of Statistics, Faculty of Science, Cankiri Karatekin University, Cankiri 18100, Turkey
*
Author to whom correspondence should be addressed.
Symmetry 2023, 15(12), 2107; https://doi.org/10.3390/sym15122107
Submission received: 27 September 2023 / Revised: 15 November 2023 / Accepted: 21 November 2023 / Published: 23 November 2023
(This article belongs to the Section Mathematics)

Abstract

:
The negative binomial regression model is a widely adopted approach when dealing with dependent variables that consist of non-negative integers or counts. This model serves as an alternative regression technique for addressing issues related to overdispersion in count data. Typically, the maximum likelihood estimator is employed to estimate the parameters of the negative binomial regression model. However, the maximum likelihood estimator can be highly sensitive to multicollinearity, leading to unreliable results. To eliminate the adverse effects of multicollinearity in the negative binomial regression model, we propose the use of a jackknife version of the Kibria–Lukman estimator. In this study, we conducted a theoretical comparison between the proposed jackknife Kibria–Lukman negative binomial regression estimator and several existing estimators documented in the literature. To assess the performance of the proposed estimator, we conducted two simulation studies and performed a real data application. The results from both the simulation studies and the real data application consistently demonstrated that the proposed jackknife Kibria–Lukman negative binomial regression estimator outperforms other estimators.

1. Introduction

Count data refers to data that represent the occurrence of events within a given period, taking on positive values indicating the frequency of those events. In the case of count data, the distribution typically exhibits right skewness. Due to the violation of normality assumptions, classical regression methods are unsuitable for count data analysis [1]. The use of classical methods often leads to biased parameter estimates. Consequently, researchers have focused on developing regression models specifically tailored for count data analysis, aiming to enhance data interpretation. Several regression models are commonly employed to analyze count data, including the Poisson regression, negative binomial regression (NBR), zero-inflated regression, zero-truncated regression, and hurdle regression models [2]. Count data often exhibit overdispersion, which violates the assumptions of the Poisson regression model. Using the Poisson regression model for overdispersed data results in biased parameter estimates and unreliable outcomes. To address this, the negative binomial regression model is frequently utilized as it assumes a negative binomial distribution for the response variable. The logarithm of the expected value of the response variable can be modeled using a linear combination of unknown parameters [3]. The negative binomial regression model finds application across various fields, such as health, actuarial science, education, economics, biostatistics, and healthcare. The issue of multicollinearity arises when there is a relationship between some of the variables under study, leading to errors induced by the researcher [4]. Maximum likelihood estimation (MLE) applied to a model with multicollinearity can result in the incorrect estimation of regression coefficients, sign mismatches between correlation coefficients and regression coefficients, inflated standard error values, and decreased model power. Various approaches can be employed to mitigate multicollinearity, such as adding new variables or removing multicollinear variables from the model. However, removing variables from the model risks discarding variables that truly contribute to the model, resulting in a loss of information. Alternative methods, such as the ridge and Liu estimators, are commonly employed to estimate regression coefficients without removing variables, albeit at the expense of introducing bias to the estimates. The ridge and Liu estimators have been developed as an alternative to MLE. Both estimators have a smaller mean square error (MSE) than MLE [5]. Månsson [6] suggested the negative binomial ridge regression estimator (RNBR) as an alternative to the MLE method in the NBR. Månsson [7] proposed a Liu estimator to solve an inflated mean square error problem of the maximum likelihood method in the presence of multicollinearity in the NBR. Asar [8] proposed a new biased estimator for the negative binomial regression model, which is a generalization of the Liu-type estimator. Türkan and Özel [9] proposed the modified jackknife estimator in the NBR. Alobaidi et al. [10] proposed a new ridge estimator in the NBR to address the multicollinearity problem. Duran and Akdeniz [11] proposed a new estimator named the modified jackknife generalized Liu-type estimator. This estimator is based on a combination of both the generalized Liu estimator and the jackknife generalized Liu estimator. Arum et al. [12] combined the modified jackknife ridge estimator with the transformed M-estimator to solve the problem of both multicollinearity and outliers in the Poisson regression model. As a result, they showed that the method they proposed was dominated by the existing methods. Akram et al. [13] proposed a new estimator, the zero-inflated negative binomial Stein estimator, to alleviate the effects of the multicollinearity problem in zero-inflated negative binomial regression models. According to the simulation and application results, they showed that the proposed new estimator is superior to the existing estimators. Dawoud et al. [14] proposed new robust estimators to handle multicollinearity and outliers in the Poisson Model. Akram et al. [15] proposed a new Conway–Maxwell–Poisson (COMP) Liu estimator for the COMP regression model to address the problem of multicollinearity. This estimator covers situations of overdispersion, equal dispersion, and underdispersion. Amin et al. [16] proposed a new adjusted Poisson Liu estimator as a robust solution for dealing with multicollinear explanatory variables in the Poisson regression. The adjusted Poisson Liu estimator addresses the problem of negative shrinkage parameter values encountered by the traditional Liu estimator. Through simulation studies and empirical applications, the adjusted Poisson Liu estimator is shown to outperform the MLE and other competing estimators in terms of robustness and consistency. Sami et al. [17] proposed several ridge regression estimators in the context of the COMP regression model to overcome the effects of multicollinearity, considering overdispersion, equidispersion, and underdispersion. Amin et al. [18] proposed a Poisson James–Stein estimator as a solution to the problems of increased variance and standard error associated with the MLE when dealing with multicollinear explanatory variables in the Poisson regression. Amin et al. [19] introduced the Bell Ridge regression estimator as a solution to mitigate the multicollinearity issue in the Bell regression model. Batool et al. [20] suggested a ridge estimator as a method to address the issue of overdispersion in the Poisson inverse Gaussian regression model. Abonazel [21] proposed a new modified Liu estimator for the COMP regression model based on two shrinkage parameters to reduce the multicollinearity effect in the COMP and used the mean square error criterion to evaluate the performance of the proposed estimator. Algamal et al. [22] proposed a jackknife version of the ridge estimator for the Bell regression model by jackknifing the ridge estimator to reduce bias. As a result, they showed that their proposed estimator improved the performance of the ridge estimator. Algamal et al. [23] proposed a modified jackknife ridge estimator for the COMP model. As a result, they found that their proposed estimator had minimum bias and minimum mean square error.
In recent years, Kibria and Lukman [24] developed the Kibria–Lukman (KL) estimator, which was found to outperform the ridge and Liu estimators in the linear regression model. Lukman et al. [25] proposed a new biased estimator called the KL estimator for linear mixed models. Aladeitan et al. [26] proposed a modified KL estimator in the Poisson regression model. Ugwuowo et al. [27] proposed the jackknife version of the KL estimator in the linear regression model. Rasheed et al. [28] proposed the jackknife approach and its modified version to reduce the effects of multicollinearity and the bias of using the Liu-type estimator simultaneously in the COMP regression model. Jabur et al. [29] proposed a jackknife Liu-type negative binomial estimator. Dawoud et al. [30] proposed a generalized version of the KL estimator along with the optimal biasing parameter of their proposed estimator derived by minimizing the scalar mean square error. Abonazel et al. [31] proposed the Kibria–Lukman estimator to overcome the multicollinearity problem in the COMP regression model.
In this study, we introduce a modification to the KL estimator proposed by Kibria and Lukman [24] for the NBR model. We employ the jackknife method to address the issue of multicollinearity and reduce the MSE in the NBR model. To evaluate the effectiveness of our proposed estimator, we conduct a comprehensive comparison with existing estimators through theoretical analysis, simulation studies, and real data application.
The article is organized as follows: In Section 2, we present the existing estimators. Section 3 provides a detailed discussion of the proposed estimator. A theoretical comparison of efficiency is presented in Section 4. In Section 5, we conduct two simulation studies and analyze a real dataset example. Finally, in Section 6, we provide concluding remarks.

2. Existing Estimators

2.1. Negative Binomial Regression Model MLE Estimator

The NBR model is widely used to model the response variable as count or non-negative integers. The negative binomial probability density function is given as
f y i ; μ i , θ = Γ ( θ + y i ) Γ ( θ ) Γ ( 1 + y i ) θ θ + μ i θ μ i θ + μ i y i
where Γ . is the gamma function and θ is the overdispersion parameter. The mean and variance of the negative binomial probability distribution are, respectively, E y i = μ i and V a r y i = μ i + μ i 2 θ [32].
The NBR model parameter can be estimated using the MLE method or it can be estimated as a member of the generalized linear models [32]. The log-likelihood function of the negative binomial regression model is given by:
L θ , γ = i = 1 n t = 0 y i 1 l o g ( t + θ l o g y i ! y i + θ l o g 1 + 1 θ μ i + y i l o g 1 θ + y i μ i ,
where μ i = e x p ( x i γ ) , x i is a p × 1 vector of the covariates, and γ is a p × 1 vector of the parameters. Differentiating Equation (2) with respect to γ using the iteratively weighted least squared algorithm or Fisher’s scoring algorithm, we have
γ ^ M L E = X T W ^ X 1 X T W ^ u ^ ,
where u ^ is a vector where the i th elements equals log μ ^ i + y i μ ^ i / μ ^ i and
W ^ = d i a g ( μ ^ i / ( 1 + 1 θ μ ^ i ) .
The MLE estimator of γ is normally distributed with the asymptotic mean vector E γ ^ M L E = 0 and the asymptotic covariance matrix
C o v γ ^ M L = X T W ^ X 1 .
The MSE based on the asymptotic covariance matrix is as follows:
M S E γ ^ M L E = t r X T W ^ X 1 = j = 1 p 1 ω j ,
where ω j is the eigenvalue of the X T W ^ X matrix.

2.2. Liu-Type Negative Binomial Regression Estimator (LNBR)

Asar [8] extended the Liu-type estimator to the NBR model, as follows:
γ ^ L N B R = X T W ^ X + I 1 X T W ^ X + d I γ ^ M L , 0 < d < 1
where d is the Liu parameter. The mean square error matrix (MSEM) of the LNBR estimator is
M S E M γ ^ L N B R = Q Ω k 1 Ω d Ω 1 Ω d Ω k 1 Q T + d 1 2 Q Ω k 1 γ γ T Ω k 1 Q T ,
where Ω k = d i a g ω 1 + k , , ω p + k and Ω d = d i a g ω 1 + d , , ω p + d . Then, the MSE of the LNBR estimator is
M S E γ ^ L N B R = j = 1 p 1 ω j ω j + d ω j + 1 2 + d 1 2 j = 1 p γ j 2 ω j + 1 2 .

2.3. Ridge Negative Binomial Regression Estimator (RNBR)

The ridge negative binomial regression (RNBR) estimator was presented by Manson and Shukur [33] and Jabur et al. [29], as shown below:
γ ^ R N B R = X T W ^ X + k I 1 X T W ^ X γ ^ M L ,
where 0 < k < 1 .
The MSEM for the RNBR estimator is as follows:
M S E M γ ^ R N B R = C o v ( γ ^ R N B R ) + B i a s ( γ ^ R N B R ) B i a s ( γ ^ R N B R ) T , = Q Ω k 1 Ω Ω k 1 Q T + k 2 Q Ω k 1 γ γ T Ω k 1 Q T ,
where k > 0, Ω = d i a g ω 1 , , ω p = Q T X T W ^ X Q , Ω k = d i a g ω 1 + k , ω 2 + k , , ω p + k and Q is the matrix whose columns are eigenvectors of the   X T W ^ X . The MSE for the RNBR estimator is as follows:
M S E ( γ ^ R N B R ) = j = 1 P ω j ω j + k 2 + k 2 j = 1 P γ j 2 ω j + k 2 .

3. Proposed Estimator

Considering the Kibria and Lukman [24] and Algamal and Abonazel [34], we defined the KL negative binomial regression (KLNBR) estimator as follows:
γ ^ K L N B R = X W ^ X + k I 1 X W ^ X k I γ ^ M L .
The canonical form of the KL negative binomial regression estimator is given as:
γ ^ K L N B R = Ω + k I 1 Ω k I γ ^ M L ,
where Ω = d i a g ω 1 , ω 2 , , ω p = T W ^ T is a diagonal matrix, whose columns are the eigenvectors of   X W ^ X , T = X Q such that Q is an orthogonal matrix whose rows are normalized eigenvectors of the X′WX matrix. We rewrite the negative binomial regression ML in Equation (3)
γ ^ M L E = Ω 1 T W ^ u ^ ,
γ ^ K L N B R = Ω + k I 1 Ω k I Ω 1 T W ^ u ^ ,
with C 1 = Ω + k I 1 Ω k I Ω 1 , γ ^ K L N B R = C 1 T W ^ u ^ .
Singh et al. [35] used the jackknife procedure to reduce the bias of the estimator. We applied the jackknife procedure to the KL negative binomial regression estimator, and we derived the jackknife KL negative binomial regression estimator, as follows:
γ ^ K L N B R _ i = Ω i + k I 1 Ω i k I γ ^ M L E .
Considering Ugwuowo et al. [27], we defined the jackknife KL negative binomial regression (JKLNBR) estimator, as follows:
γ ^ K L N B R _ i = C t i w ^ i t i 1 · T W ^ u ^ t i w ^ i u ^ i ,
γ ^ K L N B R _ i = C 1 + C 1 t i w ^ i t i C 1 1 t i w ^ i C 1 t i T W ^ u ^ t i w ^ i u ^ i ,
γ ^ K L N B R _ i = C 1 T W ^ u ^ C 1 t i w ^ i u ^ i + C 1 t i w ^ i t i C 1 t i w ^ i u ^ i 1 t i w ^ i C 1 t i C 1 t i w ^ i t i C 1 1 t i w ^ i C 1 t i t i w ^ i u ^ i , = γ ^ K L N B R C 1 w ^ i u ^ i 1 + t i w ^ i C 1 t i 1 t i w ^ i C 1 t i + C 1 t i w ^ i t i C 1 t i w ^ i u ^ i 1 t i w ^ i C 1 t i , = γ ^ K L N B R C 1 t i w ^ i u ^ i 1 1 t i w ^ i C 1 t i + C 1 t i w ^ i t i γ ^ K L N B 1 t i w ^ i C 1 t i , = γ ^ K L N B R C 1 t i w ^ i u ^ i t i w ^ i γ ^ K L N B 1 t i w ^ i C 1 t i , = γ ^ K L N B R C 1 t i w ^ i e i 1 h i ,
where h i = 1 t i w ^ i C 1 t i , t i is the ith row of T , and e i is the ith KLNBR residual.
Using the weighted pseudo-values defined by Khurana et al. [36], the jackknife KL negative binomial regression estimator is obtained as
γ ^ J K L N B R = γ ^ K L N B R + C 1 1 Ω + k I 1 Ω k I T W ^ u ^ ,
γ ^ J K L N B R = γ ^ K L N B R + C 1 T W ^ u ^ C 1 1 Ω + k I 1 Ω k I T W ^ u ^ ,
γ ^ J K L N B R = γ ^ K L N B R + γ ^ K L N B R Ω + k I 1 Ω k I γ ^ K L N B R ,
γ ^ J K L N B R = γ ^ K L N B R ( 2 I Ω + k I 1 Ω k I ,
γ ^ J K L N B R = Ω + k I 1 [ 2 Ω + k I Ω k I ] γ ^ K L N B R .
Consequently, the jackknife KL negative binomial regression estimator is
γ ^ J K L N B R = γ ^ M L E [ I 2 k Ω + k I 1 2 ] .
Following Duran and Akdeniz [11], subsequently, we modified Equation (25)
γ ^ J K L N B R = γ ^ K L N B R [ I 2 k Ω + k I 1 2 ] ,
where γ ^ K L N B R = γ ^ M L [ I 2 k Ω + k I 1 ] .
Therefore,
    γ ^ J K L N B R = γ ^ M L E [ I + 2 k Ω + k I 1 ] [ I 2 k Ω + k I 1 2 ] ,
The expectation of γ ^ J K L N B R is as follows:
E γ ^ J K L N B R = I 2 k Ω + k I 1 2 [ I + 2 k Ω + k I 1 ] γ .
The bias of γ ^ J K L N B R is given by
B i a s γ ^ J K L N B R = I 2 k Ω + k I 1 2 I + 2 k Ω + k I 1 I γ .
The MSEM for the JKLNBR estimator is
M S E M γ ^ J K L N B R = C o v γ ^ J K L N B R + B i a s γ ^ J K L N B R · B i a s γ ^ J K L N B R T ,
M S E M γ ^ J K L N B R = I 2 k Ω + k I 1 2 2 Ω I 2 k Ω + k I 1 2 B i a s γ ^ J K L N B R · B i a s γ ^ J K L N B R T
The MSE of the JKLNBR estimator is
M S E γ ^ J K L N B R = t r M S E M γ ^ J K L N B R , = j = 1 p ω j + k 2 4 k 2 2 ω j k 2 ω j ω j + k 6 + j = 1 p ω j k 2 ω j + 3 k ω j + k 3 2 γ ^ 2 ω j + k 6 .

4. Comparison of Efficiency

Following Oranye and Ugwuowo [37], Algamal and Abonazel [34], and Ugwuowo et al. [27], we provide the efficiency of the proposed jackknife KL negative binomial regression estimator over the existing estimators. The Farebrother [38] and Trenkler and Toutenburg [39] lemmas are adopted for theoretical comparisons.
  • The suggested estimator γ ^ J K L N B R is more efficient than to estimate γ ^ M L if and only if γ I 2 k Ω + k I 1 2 I + 2 k Ω + k I 1 I [ Ω 1 [ I 2 k Ω + k I 1 2 ] 2 Ω [ I 2 k Ω + k I 1 2 ] 1 ] [ I 2 k Ω + k I 1 ] 2 I + 2 k Ω + k I 1 I γ < 1 for the NBR model.
Proof. 
The difference between the MSEM functions of the MLE and JKLNBR estimators are obtained as follows:
M S E M γ ^ M L E M S E M γ ^ J K L N B R = d i a g 1 ω j ω j + k 2 4 k 2 2 ω j k 2 ω j ω j + k 6 j = 1 p .
Therefore, Ω 1 [ I 2 k Ω + k I 1 2 ] 2 Ω [ I 2 k Ω + k I 1 ] 2 is positive definite provided ω j ω j + k 6 > ω j ω j + k 2 4 k 2 2 ω j k 2 . Thus, the proof is finished. □
ii.
The suggested estimator γ ^ J K L N B R is more efficient than to estimate γ ^ L N B R if and only if γ I 2 k Ω + k I 1 2 I + 2 k Ω + k I 1 I [ Ω + d I 2 Ω 1 Ω + I 2 [ I 2 k Ω + k I 1 2 ] 2 Ω I 2 k Ω + k I 1 2 + d 1 2 Ω + I 2 γ γ ] 1 I 2 k Ω + k I 1 2 I + 2 k Ω + k I 1 I γ < 1 under the NRB model.
Proof. 
The MSEM of the LNBR estimator
M S E M γ ^ L N B R = Q Ω k 1 Ω d Ω 1 Ω d Ω k 1 Q T + d 1 2 Q Ω k 1 γ γ T Ω k 1 Q T ,
where d is the Liu parameter, Ω k = d i a g ω 1 + k , , ω p + k and Ω d = d i a g ω 1 + d , , ω p + d .
The difference between the MSEM functions of the LNBR and JKLNBR estimators are obtained as follows:
M S E M γ ^ L N B R M S E M γ ^ J K L N B R = d i a g 1 ω j ω j + d ω j + 1 2 ω j + k 2 4 k 2 2 ω j k 2 ω j ω j + k 6 j = 1 p .
Thus, Ω + d I 2 Ω 1 Ω + I 2 I 2 k Ω + k I 1 2 2 Ω I 2 k Ω + k I 1 2 is positive definite provided   ω j + d 2 ω j + k 6 > ω j + k 2 4 k 2 2 ω j k 2 ω j + 1 2 . □
iii.
The suggested estimator γ ^ J K L N B R is more efficient than to estimate γ ^ R N B R if and only if γ I 2 k Ω + k I 1 2 I + 2 k Ω + k I 1 I [ Ω 2 Ω + k I 2 [ I 2 k Ω + k I 1 2 ] 2 Ω I 2 k Ω + k I 1 2 + k 2 Ω + k I 2 γ γ ] 1 I 2 k Ω + k I 1 2 I + 2 k Ω + k I 1 I γ < 1 under the NBR model.
Proof. 
The MSEM of the RNBR estimator is
M S E M γ ^ R N B R = Q Ω k 1 Ω Ω k 1 Q T + k 2 Q Ω k 1 γ γ T Ω k 1 Q T ,
where k > 0 and Ω k = d i a g ω 1 + k , ω 2 + k , , ω p + k
The difference between the MSEM functions of the RNBR and JKLNBR estimators are obtained as follows:
M S E M γ ^ R N B R M S E M γ ^ J K L N B R = d i a g ω j 2 ω j + k 2 ω j + k 2 4 k 2 2 ω j k 2 ω j ω j + k 6 j = 1 p .
Therefore, [ Ω 2 Ω + k I 2 [ I 2 k Ω + k I 1 2 ] 2 Ω [ I 2 k Ω + k I 1 2 ] ] is positive definite provided ω j 3 ω j + k 4 > ω j + k 2 4 k 2 2 ω j k 2 .
iv.
The suggested estimator γ ^ J K L N B R is more efficient than to estimate γ ^ K L N B R if and only if γ I 2 k Ω + k I 1 2 I + 2 k Ω + k I 1 I [ Ω k I 2 Ω 1 Ω + k I 2 [ I 2 k Ω + k I 1 2 ] 2 Ω I 2 k Ω + k I 1 2 + 4 k 2 Ω + k I 2 γ γ ] 1 I 2 k Ω + k I 1 2 I + 2 k Ω + k I 1 I γ < 1 under the NBR model.
Proof. 
The MSE of the KLNBR estimator is
M S E γ ^ K L N B R = j = 1 p ω j k 2 ω j ω j + k 2 + 4 k 2 j = 1 p γ j 2 ω j + k 2 .
The difference between the MSEM functions of the KLNBR and JKLNBR estimators is obtained as follows:
M S E M γ ^ K L N B R M S E M γ ^ J K L N B R = d i a g 1 ω j ω j k ω j + k 2 ω j + k 2 4 k 2 2 ω j k 2 ω j ω j + k 6 j = 1 p .
Consequently, Ω k I 2 Ω 1 Ω + k I 2 I 2 k Ω + k I 1 2 2 Ω I 2 k Ω + k I 1 2 is positive definite provided ω j k 2 ω j + k 4 > ω j + k 2 4 k 2 2 ω j k 2 . □
There is no definite rule for estimating the k and d parameters. In this paper, we use the parameter d derived by [27]. We consider one method to estimate the ridge parameter k. However, in practice, there are countless ways available to estimate the ridge parameter k. For different methods of estimating the ridge parameter k for both linear and non-linear regression models, we refer to Kibria [40], Kibria and Banik [41], and, very recently, Kibria [42], among others. The biasing parameter for the ridge and KL estimators is as follows:
k ^ = p j = 1 p γ ^ j 2 j = 1,2 , , p ,
where p is the number of estimated coefficients.
d ^ = m i n γ ^ j 2 1 ω j + γ ^ j 2 .

5. Application

In this section, we conducted three numerical studies to evaluate the performance of the proposed jackknife KL negative binomial regression estimator in comparison to traditional estimators. These numerical studies consisted of two Monte Carlo simulations and an analysis of a real dataset. For a good-quality MC simulation, the use of a random number generator alone is not good enough [43]. So, in the simulation designs, we generated multiple datasets for negative binomial regression models according to the following procedure:
μ = e x p ( β 1 X 1 + + β p X p ) .
The response variable is simulated as Y ~ N B ( μ , μ + θ 1 μ ) , and θ = 1 is a suitable parameter value [44].
We calculate the mean square error of the negative binomial regression coefficient estimates in the simulation parts as follows:
M S E β = 1 100 i = 1 100 β i β r T β i β r ,
where β r are the true regression coefficients and β i are the estimated negative binomial regression coefficients. The simulation settings aim to denote the precision of the estimates β values. The number of runs is 100, and the sample size is n = 20, 25, 30, 50,100, and 300 in the simulation settings.
In the real data analysis part, we divided the data as test–train and evaluated the predictive performance of the negative binomial estimators. The prediction performances of the estimators are computed using the following MSE formula:
M S E Y = 1 n i = 1 n y p r e d ( i ) y r e a l ( i ) 2 ,
where y p r e d shows the predictions and y r e a l shows the real response values. This formula is used to demonstrate the predictive performance of the negative binomial regression estimators for both the test and train sets in the real data analysis part.
Both the simulation parts and real dataset analysis were chosen to have multicollinearity among the variables. We obtained the entire results using R software 4.3.2 [45].

5.1. Simulation Studies

In this section, we present two distinct simulation designs that incorporate different structures for the independent variables. Each simulation design aims to assess the performance of the proposed jackknife KL negative binomial regression estimator and traditional estimators under these specific structures.

5.1.1. Simulation Study 1

In this setting, we generate the independent variables (X) as follows:
X i j = 1 r 2 Z i j + r Z i ( p + 1 ) i = 1 , 2 , n j = 1 , 2 , , p
where r controls the correlation value among the predictors, and Z~N(0,1) shows the standard normally distributed variables [46,47]. We set the correlation as r = 0.90, 0.95, 0.99 with p = 3, 6.
Table 1 and Table 2 show the results of the first simulation for p = 3, 6, respectively. The MSE values decrease as the sample size increases for all the estimators. When looking at the results, we see that our proposed estimators are superior to the maximum likelihood, Liu-type, ridge, and KL negative binomial estimators. In all cases, the JKLNBR has the lowest MSE values. Even in a very high correlation (r = 0.99), the JKLNBR achieves the best for all the sample sizes.

5.1.2. Simulation Study 2

In this subsection, we generated highly correlated predictors by following the method of Kartal Koc and Bozdogan [48] and Dünder [49,50]. The multicollinearity level is controlled using the c parameter. In total, we simulated five different predictors, as follows:
X 1 = 10 + ε 1
X 2 = 10 + 0.3 ε 1 + c ε 2
X 3 = 10 + 0.3 ε 1 + 0.5604 c ε 2 + 0.8282 c ε 2
X 4 = 8 + X 1 + 0.5 X 2 0.3 X 3 + 0.5 ε 4
X 5 = 5 + 0.5 X 1 + X 2 + 0.5 ε 5
where ε 1 , ε 2 , ε 3 , ε 4 , and ε 5 ~ N(0,1) are the standard normally distributed variables. To introduce strong increasing multicollinearity, we set the parameter ‘c’ to three different values: c = 0.3, 0.6, and 0.9. These values ensure a progressively stronger correlation among the independent variables in the simulation designs.
We report the results of the second simulation design in Table 3. Based on the Mean Square Error (MSE) values, the JKLNBR estimator consistently demonstrates the lowest prediction error across all the cases in the negative binomial regression models. On average, the JKLNBR estimator outperforms the other estimators in terms of the MSE values, indicating superior performance.

5.2. Real Dataset Example

In this section, we evaluate the performance of the negative binomial estimators using a real dataset called the “docvisit” data. The “docvisit” dataset is suitable for count data analysis as it contains the number of doctor visits as the response variable. The description of the regressors of the docvisits dataset is given in Table 4. The dataset consists of 14 different regressors and a sample size of n = 1812. The “docvisit” dataset can be accessed from the “zic” package in R [51].
We divided the “docvisit” data into two parts: a training set and a test set. The training set comprised the first 1500 observations, while the remaining 318 observations were allocated to the test set. Using the training set, we estimated negative binomial regression models employing five different estimators. Subsequently, we evaluated the predictive performance of these models by examining their ability to generalize to unseen data, specifically the test set. This evaluation allows us to assess the models’ effectiveness in predicting the number of doctor visits based on data not used for the model estimation.
To investigate the existence of multicollinearity, we computed the condition index via the following formulation:
κ = λ m a x λ m i n
where λ values represent the eigenvalues of X T X [50]. The regressors have the multicollinearity problem due to the high condition index = 362.881, which is greater than 100 [52]. Besides this fact, there are two correlations greater than 0.90 (correlation = 0.993, 0.913) between the regressors. These indicators demonstrate the existence of the multicollinearity problem in our dataset.
We also investigate the overdispersion problem using the chi-squared statistic (CHISQ) and the overdispersion parameter (OP), as follows:
C H I S Q = i = 1 n y i μ i 2 μ i
O P = i = 1 n C H I S Q n p
where μ i denotes the predictions and y i shows the actual vales. In the formulation, df = n − p shows the degrees of freedom of the residuals. Based on the overdispersion test, we conclude that the model is overdispersed (CHISQ = 2850.934, p < 0.001). The overdispersion parameter is equal to 1.918, which is also greater than 1, and it indicates the existence of the overdispersion problem as well [53]. The correlation matrix of the predictor variables is given in Table 5.
Table 6 shows the regression coefficients of the negative binomial regression estimators. The regression coefficients have different signs, and the intercept terms slightly differ in the KL estimators. The estimator with the highest mean square error is the MLE due to the presence of multicollinearity. The proposed estimator, JKLNBR, has outperformed the other estimators because it has the smallest MSE.
Table 7 reports the average MSE(Y) values (given in Equation (46)) for the test and train datasets for five type negative binomial regression estimators. We divided the dataset for training and testing with the proportions 20–80%, respectively. In both the train and test data, the proposed JKLNBR estimators apparently have the lowest MSE values. Even the classical KL estimator suffers; our proposed one is better than the other competitors.

6. Conclusions

The NBR model, which is a mixed distribution, is a solution to the overdispersion problem in modeling overdispersed count data. However, when faced with the problem of multicollinearity, the variance of the MLE of the NBR model tends to increase, leading to unreliable results. To address this problem, various biased estimators have been proposed in the literature. In this study, we introduce a new estimator for NBR models using the jackknife approach. Our proposed estimator effectively reduces the negative impact of multicollinearity on the standard error of the estimates, resulting in a smaller MSE with minimal bias. The theoretical results show that our proposed estimators outperform several existing estimators, including the MLE, LNBR, RNBR, and KLNBR. Furthermore, we evaluate the performance of our proposed JKLNBR estimator through two simulation studies and a real data application. The empirical results strongly support the effectiveness of the proposed JKLNBR estimator on several fronts. In our simulation results, the JKLNBR estimator exhibits the highest estimation accuracy and consistently provides the smallest MSE values when assessing the correctness of the regression coefficient estimates. In addition, the JKLNBR estimator outperforms other estimators in the real data example, showing the lowest prediction error on the test data. Even in the presence of high multicollinearity, our new estimator is the best when compared to the classical MLE, Liu-type, ridge, and classical KL estimators in terms of the NBRM. In future studies, we aim to investigate the performance of the Kibria–Lukman estimator in other types of regression models where the response variable is in the form of count data. In addition, we will investigate the behavior of the shrinkage parameters under different selection procedures and develop estimators that can deal not only with the multicollinearity problem but also with the outlier problem.

Author Contributions

T.K. and H.K. contributed equally to this work. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pittman, B.; Buta, E.; Krishnan-Sarin, S.; O’Malley, S.S.; Liss, T.; Gueorguieva, R. Models for analyzing zero-inflated and overdispersed count data: An application to cigarette and marijuana use. Nicotine Tob. Res. 2020, 22, 1390–1398. [Google Scholar] [CrossRef] [PubMed]
  2. Alrumayh, A.; Khogeer, H.A. A New Two-Parameter Discrete Distribution for overdispersed and Asymmetric Data: Its Properties, Estimation, Regression Model, and Applications. Symmetry 2023, 15, 1289. [Google Scholar] [CrossRef]
  3. Cameron, A.C.; Trivedi, P.K. Regression Analysis of Count Data; Cambridge University Press: Cambridge, UK, 2013. [Google Scholar]
  4. Mamun, A.; Paul, S. Model Selection in Generalized Linear Models. Symmetry 2023, 15, 1905. [Google Scholar] [CrossRef]
  5. Liu, K. A new class of biased estimate in linear regression. Commun. Stat.-Theory Methods 1993, 22, 393–402. [Google Scholar]
  6. Månsson, K. On ridge estimators for the negative binomial regression model. Econ. Model. 2012, 29, 178–184. [Google Scholar] [CrossRef]
  7. Månsson, K. Developing a Liu estimator for the negative binomial regression model: Method and application. J. Stat. Comput. Simul. 2013, 83, 1773–1780. [Google Scholar] [CrossRef]
  8. Asar, Y. Liu-type negative binomial regression: A comparison of recent estimators and applications. In Trends and Perspectives in Linear Statistical Inference; Springer: Cham, Switzerland, 2018; pp. 23–39. [Google Scholar]
  9. Türkan, S.; Özel, G. A Jackknifed estimator for the negative binomial regression model. Commun. Stat.-Simul. Comput. 2018, 47, 1845–1865. [Google Scholar] [CrossRef]
  10. Alobaidi, N.N.; Shamany, R.E.; Algamal, Z.Y. A new ridge estimator for the negative binomial regression model. Thail. Stat. 2021, 19, 116–125. [Google Scholar]
  11. Akdeniz Duran, E.; Akdeniz, F. Efficiency of the modified jackknifed Liu-type estimator. Stat. Pap. 2012, 53, 265–280. [Google Scholar] [CrossRef]
  12. Arum, K.C.; Ugwuowo, F.I.; Oranye, H.E. Robust Modified Jackknife Ridge Estimator for the Poisson Regression Model with Multicollinearity and outliers. Sci. Afr. 2022, 17, e01386. [Google Scholar] [CrossRef]
  13. Akram, M.N.; Abonazel, M.R.; Amin, M.; Kibria, B.M.G.; Afzal, N. A new Stein estimator for the zero-inflated negative binomial regression model. Concurr. Comput. Pract. Exp. 2022, 34, e7045. [Google Scholar] [CrossRef]
  14. Dawoud, I.; Awwad, F.A.; Tag Eldin, E.; Abonazel, M.R. New Robust Estimators for Handling Multicollinearity and Outliers in the Poisson Model: Methods, Simulation and Applications. Axioms 2022, 11, 612. [Google Scholar] [CrossRef]
  15. Akram, M.N.; Amin, M.; Sami, F.; Mastor, A.B.; Egeh, O.M.; Muse, A.H. A new Conway Maxwell–Poisson Liu regression estimator method and application. J. Math. 2022, 2022, 3323955. [Google Scholar] [CrossRef]
  16. Amin, M.; Akram, M.N.; Kibria, B.G. A new adjusted Liu estimator for the Poisson regression model. Concurr. Comput. Pract. Exp. 2021, 33, e6340. [Google Scholar] [CrossRef]
  17. Sami, F.; Amin, M.; Butt, M.M. On the ridge estimation of the Conway-Maxwell Poisson regression model with multicollinearity: Methods and applications. Concurr. Comput. Pract. Exp. 2022, 34, e6477. [Google Scholar] [CrossRef]
  18. Amin, M.; Akram, M.N.; Amanullah, M. On the James-Stein estimator for the Poisson regression model. Commun. Stat.-Simul. Comput. 2022, 51, 5596–5608. [Google Scholar] [CrossRef]
  19. Amin, M.; Akram, M.N.; Majid, A. On the estimation of Bell regression model using ridge estimator. Commun. Stat.-Simul. Comput. 2023, 52, 854–867. [Google Scholar] [CrossRef]
  20. Batool, A.; Amin, M.; Elhassanein, A. On the performance of some new ridge parameter estimators in the Poisson-inverse Gaussian ridge regression. Alex. Eng. J. 2023, 70, 231–245. [Google Scholar] [CrossRef]
  21. Abonazel, M.R. New modified two-parameter Liu estimator for the Conway–Maxwell Poisson regression model. J. Stat. Comput. Simul. 2023, 93, 1976–1996. [Google Scholar] [CrossRef]
  22. Algamal, Z.; Lukman, A.; Golam, B.K.; Taofik, A. Modified Jackknifed Ridge Estimator in Bell Regression Model: Theory, Simulation and Applications. Iraqi J. Comput. Sci. Math. 2023, 4, 146–154. [Google Scholar]
  23. Algamal, Z.Y.; Abonazel, M.R.; Awwad, F.A.; Eldin, E.T. Modified Jackknife Ridge Estimator for the Conway-Maxwell-Poisson Model. Sci. Afr. 2023, 19, e01543. [Google Scholar] [CrossRef]
  24. Kibria, B.M.G.; Lukman, A.F. A new ridge-type estimator for the linear regression model: Simulations and applications. Scientifica 2020, 2020, 9758378. [Google Scholar] [CrossRef] [PubMed]
  25. Lukman, A.F.; Amin, M.; Kibria, B.G. K-L estimator for the linear mixed models: Computation and simulation. Concurr. Comput. Pract. Exp. 2022, 34, e6780. [Google Scholar] [CrossRef]
  26. Aladeitan, B.B.; Adebimpe, O.; Lukman, A.F.; Oludoun, O.; Abiodun, O.E. Modified Kibria-Lukman (MKL) estimator for the Poisson Regression Model: Application and simulation. F1000Research 2021, 10, 548. [Google Scholar] [CrossRef]
  27. Ugwuowo, F.I.; Oranye, H.E.; Arum, K.C. On the jackknife Kibria-Lukman estimator for the linear regression model. Commun. Stat.-Simul. Comput. 2021, 1–13. [Google Scholar] [CrossRef]
  28. Rasheed, H.A.; Sadik, N.J.; Algamal, Z.Y. Jackknifed Liu-type estimator in the Conway-Maxwell Poisson regression model. Int. J. Nonlinear Anal. Appl. 2022, 13, 3153–3168. [Google Scholar]
  29. Jabur, D.M.; Rashad, N.K.; Algamal, Z.Y. Jackknifed Liu-type estimator in the negative binomial regression model. Int. J. Nonlinear Anal. Appl. 2022, 13, 2675–2684. [Google Scholar]
  30. Dawoud, I.; Abonazel, M.R.; Awwad, F.A. Generalized Kibria-Lukman estimator: Method, simulation, and application. Front. Appl. Math. Stat. 2022, 8, 31. [Google Scholar] [CrossRef]
  31. Abonazel, M.R.; Saber, A.A.; Awwad, F.A. Kibria–Lukman estimator for the Conway–Maxwell Poisson regression model: Simulation and applications. Sci. Afr. 2023, 19, e01553. [Google Scholar] [CrossRef]
  32. Hilbe, J.M. Negative Binomial Regression; Cambridge University Press: Cambridge, UK, 2011. [Google Scholar]
  33. Månsson, K.; Shukur, G. A Poisson ridge regression estimator. Econ. Model. 2011, 28, 1475–1481. [Google Scholar] [CrossRef]
  34. Algamal, Z.Y.; Abonazel, M.R. Developing a Liu-type estimator in beta regression model. Concurr. Comput. Pract. Exp. 2022, 34, e6685. [Google Scholar] [CrossRef]
  35. Singh, B.; Chaubey, Y.P.; Dwivedi, T.D. An almost unbiased ridge estimator. Sankhyā Indian J. Stat. Ser. B 1986, 48, 342–346. [Google Scholar]
  36. Khurana, M.; Chaubey, Y.P.; Chandra, S. Jackknifing the ridge regression estimator: A revisit. Commun. Stat.-Theory Methods 2014, 43, 5249–5262. [Google Scholar] [CrossRef]
  37. Oranye, H.E.; Ugwuowo, F.I. Modified jackknife Kibria–Lukman estimator for the Poisson regression model. Concurr. Comput. Pract. Exp. 2022, 34, e6757. [Google Scholar] [CrossRef]
  38. Farebrother, R.W. Further results on the mean square error of ridge regression. J. R. Stat. Soc. Ser. B 1976, 38, 248–250. [Google Scholar] [CrossRef]
  39. Trenkler, G.; Toutenburg, H. Mean squared error matrix comparisons between biased estimators—An overview of recent results. Stat. Pap. 1990, 31, 165–179. [Google Scholar] [CrossRef]
  40. Kibria, B.M.G. Performance of some new ridge regression estimators. Commun. Stat.-Simul. Comput. 2003, 32, 419–435. [Google Scholar] [CrossRef]
  41. Kibria, B.M.G.; Banik, S. Some Ridge Regression Estimators and Their Performances. J. Mod. Appl. Stat. Methods 2016, 15, 206–238. [Google Scholar] [CrossRef]
  42. Kibria, B.M.G. More than hundred (100) estimators for estimating the shrinkage parameter in a linear and generalized linear ridge regression models. J. Econom. Stat. 2022, 2, 233–252. [Google Scholar]
  43. Jäntschi, L. Detecting extreme values with order statistics in samples from continuous distributions. Mathematics 2020, 8, 216. [Google Scholar] [CrossRef]
  44. Huang, J.; Yang, H. A two-parameter estimator in the negative binomial regression model. J. Stat. Comput. Simul. 2014, 84, 124–134. [Google Scholar] [CrossRef]
  45. R Core Team. R: A Language and Environment for Statistical Computing. R Foundation for Statistical Computing: Vienna, Austria, 2022. Available online: https://www.R-project.org/ (accessed on 25 January 2022).
  46. Lukman, A.F.; Kibria, B.M.G.; Nziku, C.K.; Amin, M.; Adewuyi, E.T.; Farghali, R. KL estimator: Dealing with multicollinearity in the logistic regression model. Mathematics 2023, 11, 340. [Google Scholar] [CrossRef]
  47. Koç, T.; Dünder, E. Jackknife Kibria-Lukman estimator for the beta regression model. Commun. Stat.-Theory Methods 2023, 1–17. [Google Scholar] [CrossRef]
  48. Kartal Koc, E.; Bozdogan, H. Model selection in multivariate adaptive regression splines (MARS) using information complexity as the fitness function. Mach. Learn. 2015, 101, 35–58. [Google Scholar] [CrossRef]
  49. Dünder, E. A hybridized consistent Akaike type information criterion for regression models in the presence of multicollinearity. Commun. Stat.-Simul. Comput. 2023, 1–10. [Google Scholar] [CrossRef]
  50. Dünder, E.; Gümüştekin, S.; Murat, N.; Cengiz, M.A. Variable selection in linear regression analysis with alternative Bayesian information criteria using differential evaluation algorithm. Commun. Stat.-Simul. Comput. 2018, 47, 605–614. [Google Scholar] [CrossRef]
  51. Jochmann, M. zic: Bayesian Inference for Zero-Inflated Count Models. R Package Version 0.9, 1. England. 2017. Available online: https://cran.r-project.org/web/packages/zic/zic.pdf (accessed on 25 January 2022).
  52. Alin, A. Multicollinearity. Wiley Interdiscip. Rev. Comput. Stat. 2010, 2, 370–374. [Google Scholar] [CrossRef]
  53. Payne, E.H.; Gebregziabher, M.; Hardin, J.W.; Ramakrishnan, V.; Egede, L.E. An empirical approach to determine a threshold for assessing overdispersion in Poisson and negative binomial models for count data. Commun. Stat.-Simul. Comput. 2018, 47, 1722–1738. [Google Scholar] [CrossRef]
Table 1. Simulated MSE values for the first simulation design when p = 3.
Table 1. Simulated MSE values for the first simulation design when p = 3.
θ = 1 θ = 2
rnMLELNBRRNBRKLNBRJKLNBRMLELNBRRNBRKLNBRJKLNBR
r = 0.90200.822810.809170.731240.618290.594650.529610.519950.487500.412200.39651
250.578730.542480.510450.437760.408740.369490.346340.325890.279480.26096
300.358490.332460.322990.286930.258460.226670.210210.204220.181470.16342
500.190540.182370.180160.166440.148050.120970.115780.114380.105670.09399
1000.083540.080360.080840.075860.069270.053420.051390.051700.048510.04430
3000.024080.023680.023820.023240.022490.015410.015160.015250.014870.01439
r = 0.95200.984880.897120.735300.627150.615190.626850.573810.470310.401130.39343
250.769010.699270.608340.529460.496440.474590.431550.375430.326750.30638
300.556370.509260.465670.407480.356690.354790.324750.296950.259840.22746
500.347710.324620.307290.271000.226170.216010.214090.201670.168350.14050
1000.142160.134610.133200.120020.104880.091710.086840.085930.077430.06766
3000.039640.038720.038990.037610.035830.025180.024590.024760.023890.02276
r = 0.99206.783095.531152.251372.251111.723584.522963.688171.501211.501041.14928
254.852534.078131.971891.925841.432082.876132.417141.168751.141460.84880
303.175432.758771.575961.498061.086581.920941.668880.953350.906230.65731
501.751791.573071.063580.973170.690031.089990.978780.661770.605520.42935
1000.581610.521030.434750.351170.241480.372360.333650.278330.224820.15460
3000.135370.127230.124400.109770.091050.086320.081330.079320.070000.05809
Average1.232091.064630.642230.595020.477870.776300.671250.405370.374930.30162
Table 2. Simulated MSE values for the first simulation design when p = 6.
Table 2. Simulated MSE values for the first simulation design when p = 6.
θ = 1 θ = 2
rnMLELNBRRNBRKLNBRJKLNBRMLELNBRRNBRKLNBRJKLNBR
r = 0.90201.761501.743621.712811.608921.464441.065601.054781.036150.973300.88590
251.155641.127311.125891.087790.909200.710650.693230.692350.668920.55910
300.712590.676700.645110.591640.494230.440480.418300.398770.365720.30551
500.297870.286840.283290.260570.219530.192960.185820.183520.168800.14221
1000.102620.099470.100430.094530.085100.067510.065440.066070.062190.05598
3000.034460.034080.034280.033690.032570.021440.021210.021330.020960.02027
r = 0.95203.766353.465022.999982.820942.287792.232352.053751.778111.672001.35599
252.278312.102011.850311.716501.379871.581311.458951.284251.191380.95773
301.186081.104481.000590.909140.722490.782170.728360.659850.599540.47645
500.499670.472450.450840.398860.315650.325820.308070.293970.260080.20582
1000.209080.201750.201060.185660.159350.134290.129580.129140.119250.10235
3000.060940.060100.060430.058900.056050.036090.035590.035790.034880.03320
r = 0.992019.0837117.5500411.2423611.1092110.0192011.5147310.589356.783426.703086.04539
2511.1766810.273466.677846.261685.552106.872976.317544.106463.850553.41420
305.505935.067163.410393.491662.472343.403473.132252.108122.158361.52827
502.082471.931141.440011.299150.779941.295721.201560.895980.808340.48528
1000.906300.865400.766700.684150.474900.587100.560610.496670.443190.30764
3000.248330.242470.240010.225190.193910.163350.159500.157880.148130.12756
Average2.837142.627971.902351.824341.534371.746001.617441.173771.124930.94494
Table 3. Simulated MSE values for the second simulation design when p = 5.
Table 3. Simulated MSE values for the second simulation design when p = 5.
θ = 1 θ = 2
rnMLELNBRRNBRKLNBRJKLNBRMLELNBRRNBRKLNBRJKLNBR
c = 0.3201.254890.387920.296440.280270.166220.759130.234670.179330.169550.10055
250.891410.362930.305540.293360.166240.548160.223180.187890.180400.10223
300.638440.335030.318640.266680.176220.394650.207100.196970.164850.10893
500.475490.306990.335800.238740.183650.308030.198870.217530.154660.11897
1000.377250.274410.335590.252140.201060.248170.180520.220770.165870.13227
3000.337040.271740.328880.279190.236170.209710.169080.204630.173720.14695
c = 0.6201.221760.531340.386740.275310.165820.724150.314930.229220.163180.09828
250.830480.423470.359600.259510.170690.576410.293920.249590.180120.11847
300.573220.335570.338070.243910.176060.378010.221290.222940.160850.11610
500.426850.272690.322540.228200.181420.278330.177810.210320.148800.11830
1000.368790.270470.332220.246790.197950.236870.173720.213380.158510.12714
3000.338520.270970.330740.279060.235010.200490.160490.195880.165280.13919
c = 0.9200.864080.352490.390520.251980.168020.521370.212690.235630.152040.10138
250.709370.345650.370790.248760.172140.436220.212550.228010.152970.10586
300.575780.332350.350010.246820.178450.355920.205440.216360.152570.11031
500.462430.309050.353380.238020.184420.287730.192290.219870.148100.11475
1000.366480.261860.335260.243170.195460.237410.169630.217180.157530.12662
3000.332850.261790.326220.272940.228910.218950.172210.214590.179540.15058
Average0.613620.328150.339830.258050.188000.384430.206690.214450.162700.11871
Table 4. Description of the regressors of the docvisits dataset.
Table 4. Description of the regressors of the docvisits dataset.
RegressorDescription
ageAge of the people
agesqSquare of the age
healthHealth satisfaction point between [0, 10]
handicapA dummy variable (0–1) the handicapping situation
hdegreeThe percentage of the handicap degree
marriedA dummy variable (0–1) for the marital status
schoolingThe number of years for schooling
hhincomeHousehold income per month
selfA dummy variable (0–1) for the self-employed status
civilA dummy variable (0–1) for the civil servant
bluecA dummy variable (0–1) for the blue-collar employee
employedA dummy variable (0–1) for the employed status
publicA dummy variable (0–1) for the health insurance
addonA dummy variable (0–1) for the add-on insurance
Table 5. Correlation matrix.
Table 5. Correlation matrix.
AgeAgesqHealthHandicapHdegreeMarriedSchoolingHhincomeSelfCivilBluecEmployedPublicAddon
age1
agesq0.9931
health−0.242−0.2431
handicap0.2900.303−0.3131
hdegree0.2740.290−0.3250.9131
married0.4470.414−0.0670.0840.0561
schooling−0.078−0.0920.097−0.154−0.162−0.0551
hhincome0.0390.0200.121−0.113−0.1230.1320.3101
self0.005−0.0070.048−0.083−0.0820.0220.0750.1061
civil−0.004−0.0160.057−0.043−0.0490.0140.2630.146−0.1061
bluec−0.114−0.1220.021−0.069−0.0940.040−0.351−0.124−0.210−0.2291
employed−0.207−0.2540.200−0.311−0.3640.0820.1120.2290.1440.1570.3111
public0.0010.017−0.0770.0780.0840.010−0.323−0.247−0.169−0.6390.275−0.1171
addon0.0060.0030.002−0.006−0.0170.0240.0510.0220.028−0.013−0.0370.0490.0561
Table 6. Regression coefficients of the negative binomial estimators and MSEs of the docvisit data.
Table 6. Regression coefficients of the negative binomial estimators and MSEs of the docvisit data.
CoefficientMLELNBRRNBRKLNBRJKLNBR
(Intercept)3.375792.085902.25278−0.71307−0.21224
age−0.06361−0.00910−0.016010.105310.10237
agesq0.892630.279950.35745−0.99982−0.50932
health−0.26084−0.24793−0.24971−0.21675−0.21561
handicap0.154410.124730.129580.036610.00472
hdegree0.001910.003050.002890.006060.00606
married−0.11917−0.16474−0.15907−0.25701−0.23221
schooling0.001660.006700.005970.019760.01963
hhincome0.006650.007430.007350.008700.00867
self−0.23726−0.22652−0.22812−0.19742−0.16125
civil−0.05374−0.02211−0.026170.044920.03463
bluec0.191220.208490.206290.244960.21506
employed0.03120−0.01845−0.01260−0.11041−0.09170
public0.156950.217420.209440.352220.28912
addon0.425150.361280.372080.160400.03457
MSE6.713254.139463.419330.764370.25113
K = 1.255031
d = 7.794738e-10
Table 7. Predictive performance of the negative binomial regression estimators.
Table 7. Predictive performance of the negative binomial regression estimators.
DataMLLNBRRNBRKLNBRJKLNBR
Test30.604630.780630.756031.235826.5837
Train22.127422.376822.342623.000719.4763
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Koç, T.; Koç, H. A New Effective Jackknifing Estimator in the Negative Binomial Regression Model. Symmetry 2023, 15, 2107. https://doi.org/10.3390/sym15122107

AMA Style

Koç T, Koç H. A New Effective Jackknifing Estimator in the Negative Binomial Regression Model. Symmetry. 2023; 15(12):2107. https://doi.org/10.3390/sym15122107

Chicago/Turabian Style

Koç, Tuba, and Haydar Koç. 2023. "A New Effective Jackknifing Estimator in the Negative Binomial Regression Model" Symmetry 15, no. 12: 2107. https://doi.org/10.3390/sym15122107

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop