Next Article in Journal
Information Theoretic Metagenome Assembly Allows the Discovery of Disease Biomarkers in Human Microbiome
Previous Article in Journal
Minimum Divergence Estimators, Maximum Likelihood and the Generalized Bootstrap
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Statistical Inference of Truncated Normal Distribution Based on the Generalized Progressive Hybrid Censoring

Department of Mathematics, Beijing Jiaotong University, Beijing 100044, China
*
Author to whom correspondence should be addressed.
Entropy 2021, 23(2), 186; https://doi.org/10.3390/e23020186
Submission received: 27 December 2020 / Revised: 25 January 2021 / Accepted: 28 January 2021 / Published: 2 February 2021

Abstract

:
In this paper, the parameter estimation problem of a truncated normal distribution is discussed based on the generalized progressive hybrid censored data. The desired maximum likelihood estimates of unknown quantities are firstly derived through the Newton–Raphson algorithm and the expectation maximization algorithm. Based on the asymptotic normality of the maximum likelihood estimators, we develop the asymptotic confidence intervals. The percentile bootstrap method is also employed in the case of the small sample size. Further, the Bayes estimates are evaluated under various loss functions like squared error, general entropy, and linex loss functions. Tierney and Kadane approximation, as well as the importance sampling approach, is applied to obtain the Bayesian estimates under proper prior distributions. The associated Bayesian credible intervals are constructed in the meantime. Extensive numerical simulations are implemented to compare the performance of different estimation methods. Finally, an authentic example is analyzed to illustrate the inference approaches.

1. Introduction

1.1. Truncated Normal Distribution

Normal distribution has played a crucial role in a diversity of research fields like reliability analysis and economics, as well as many other scientific developments. However, in many practical situations, experimental data are available from a certain range, so the truncated form of normal distribution is more applicable in actual life.
The truncated normal distribution, with its characteristic of practical, has gained some attention among researchers and interesting results have been obtained. Ref. [1] studied the maximum likelihood estimations for singly and doubly truncated normal distributions. Ref. [2] applied the approach of moments to estimate unknown parameters of singly truncated normal distributions from the first three sample moments. Ref. [3] investigated the estimators of unknown parameters from the normal distribution under the singly censored data. Ref. [4] utilized an iterative procedure to transform singly right censored samples into pseudo-complete samples and then estimated the parameters of interest through the transformed data. One can refer to [5] to find more details about the truncated normal distribution. Ref. [6] adopted the maximum likelihood estimation and Bayesian methods to estimate the unknown parameters for the truncated normal distribution under the progressive type-II censoring scheme. Meanwhile, optimal censoring plans under different optimality criteria had been discussed. While the above-mentioned works are based on the known truncation point. Ref. [7] developed a standard truncated normal distribution. The truncated mean and variance of this distribution are zero and one, respectively, regardless of the location of the truncation points. Ref. [8] considered the maximum likelihood estimators of unknown parameters with the known truncation point and the unknown truncation point respectively.
Generally, lifetime data are non-negative and under this circumstance, the left-truncated normal distribution (the truncation point is zero) can be applied to investigate statistical inference of unknown parameters.
Provided that a variable X is subject to a left-truncated normal distribution TN( μ , τ ), which has a probability density function (pdf) when truncation ranges from zero to infinity.
f ( x ; μ , τ ) = e 1 2 τ ( x μ ) 2 2 π τ Φ ( μ τ ) , x > 0 ,
where μ > 0 and τ > 0, μ is the mean of the corresponding normal distribution and τ is the variance accordingly. Φ ( . ) is the cumulative distribution function of the standard normal distribution.
The corresponding cumulative distribution function (cdf) takes the form as
F ( x ; μ , τ ) = 1 1 Φ ( x μ τ ) Φ ( μ τ ) , x > 0 .
The hazard rate function (hrf) of the left-truncated normal distribution at zero is expressed as below. In addition, we obtain the associated numeral characteristics such as the expectation and variance respectively.
h ( x ; μ , τ ) = e 1 2 τ ( x μ ) 2 2 π τ ( 1 Φ ( x μ τ ) ) , x > 0 .
E ( X ) = μ τ ϕ ( μ τ ) Φ ( μ τ )
V a r ( X ) = τ 1 + μ τ ϕ ( μ τ ) Φ ( μ τ ) ϕ ( μ τ ) Φ ( μ τ ) 2
Figure 1 presents the pdfs and hrfs of the left-truncated normal distributions when the truncation point is zero with different μ and τ . It is observed that the pdf is unimodal and the hrf is monotone increasing.

1.2. Generalized Progressive Hybrid Censoring Scheme

As technology has developed by leaps and bounds in the past few years, products are more reliable so that we can not get enough lifetime data to estimate the unknown parameters under the constraints of time and cost. As it is unavoidable to lose some experimental units during an experiment, an increasing number of researchers turned their attention to censored data. The earliest censoring schemes proposed are type-I and type-II censoring. Furthermore, combining type-I with type-II censoring schemes, researchers developed the hybrid censoring scheme. The features of all these schemes do not enable experimenters to withdraw any experimental unit at any stage before the experiment is over. Ref. [9] introduced the progressive censoring scheme afterwards. However, this scheme takes a huge amount of hours spent in conducting the experiment. In face of such defect, Ref. [10] developed an effective method-the progressive hybrid censoring scheme.
A progressive hybrid censoring sample can be generated as follows. First of all, n identical units are put into a test. We denote X 1 , X 2 , , X m as the ordered failure times and R = ( R 1 , R 2 , , R m ) as the censoring scheme which satisfies that i = 1 m R i + m = n . Once the m-th failed unit happens or the threshold time T is reached, the test is stopped. That is, the stop time T * = min X m , T . When the first failure happens, we randomly remove R 1 units out of the test. On the occasion of the second failure, we withdraw R 2 survival units randomly from the test. Analogously, R i units are removed at random when the i-th failure occurs. Finally, either when the m-th failure has happened or the threshold time T has been reached, the rest of the survival units are taken out of the test together.
But in the progressive hybrid censoring scheme, we cannot get accurate estimates of unknown parameters since there might be only few failures before the threshold time T which is pre-fixed. For this reason, Ref. [11] introduced a generalized progressive hybrid censoring scheme (GPHCS). This censoring scheme can realize a compromise between time restriction and the number of failed observations by terminating the test at the random time T e n d = max X k , T * . It guarantees that at least k failures can be observed before the terminal time T e n d . Assume that a test starts with n identical units and the acceptable minimum number k of failures and the expected number m of failures observed are pre-fixed between zero and n, we choose the threshold time T and the censoring scheme that satisfies i = 1 m R i + m = n ahead of the test. Similarly, we remove R 1 units randomly on the occasion of the first failure. On the arrival of the second failure, we randomly withdraw R 2 units from the rest of the experimental units. The procedure keeps repeating until the end time T e n d = max X k , min X m , T is reached.
Schematically, Figure 2 gives the process that how to generate the generalized progressive hybrid censored data under different conditions of the pre-fixed time.
The generalized progressive hybrid censoring scheme modifies the terminal time T e n d to attain sufficient failed observations within a reasonable experimental period and brings about more accurate estimations of unknown parameters.
Since its higher efficiency of statistical inference, attention to generalized progressive hybrid censoring scheme has mounted. Ref. [11] employed the classical and Bayesian estimation techniques to estimate the entropy of Weibull distribution. Base on the method he introduced previously, Ref. [12] further derived the exact confidence intervals. Ref. [13] chose a competing risks model when data were sampled in the generalized progressive hybrid censoring scheme from an exponential distribution and derived the estimates through the maximum likelihood estimation approach and importance sampling method. On the basis of GPHCS, Ref. [14] investigated the two-parameter Rayleigh competing risk data adopting the maximum likelihood estimation and Gibbs sampling technique was employed to approximate the associated Bayes estimates.
The aim of our work is to obtain the classical and Bayesian estimations of the left-truncated normal distribution at zero when data are derived from the generalized progressive hybrid censoring. To begin with, the Newton–Raphson (N-R) algorithm is proposed to compute the maximum likelihood estimates (MLEs) of unknown parameters of TN( μ , τ ). Another iterative approach—expectation maximization (EM) algorithm is also introduced to calculate the estimates. Subsequently, the observed Fisher information matrix and percentile bootstrap (Boot-p) method are considered to obtain the confidence interval estimations. In the Bayesian framework, we employ Tierney and Kadane (T-K) approximation and importance sampling (IS) technique to estimate Bayesian estimators. The associated higher posterior density (HPD) intervals are developed as well. Statistical inference of the left-truncated normal distribution at zero with the generalized progressive hybrid censored samples has not yet been carried out previously in terms of what we know about it.
The rest of this paper is made up of the following sections. The maximum likelihood estimators of μ and τ are theoretically derived respectively via the N-R approach and EM method in Section 2. In Section 3, we obtain the asymptotic confidence intervals using the asymptotic distributions of the MLEs, the asymptotic distributions of log-transformed MLEs and the Boot-p method. In Section 4, Bayes estimates of all unknown quantities are achieved applying T-K approximation under different loss functions. Besides, we figure out the Bayesian estimates of parameters using the importance sampling procedure. Based on this approach, the corresponding HPD intervals are developed. Numerical simulations and an analysis of one authentic example are carried out in Section 5. Finally, we arrive at some conclusive remarks in Section 6.

2. Maximum Likelihood Estimation

Our interest in this section is to obtain the maximum likelihood estimates of μ and τ with generalized progressive hybrid censored data. Based on the pdf and cdf of the truncated normal distribution when the left truncation point is zero, the likelihood and log-likelihood functions of three cases are expressed as
  • Case I:
    L 1 ( 2 π τ ) k 2 [ Φ ( μ τ ) ] n e 1 2 τ i = 1 k ( x i μ ) 2 i = 1 k [ 1 Φ ( x i μ τ ) ] R i l 1 k 2 log ( τ ) n log Φ ( μ τ ) 1 2 τ i = 1 k ( x i μ ) 2 i = 1 k R i log ( 1 Φ ( x i μ τ ) )
  • Case II:
    L 2 ( 2 π τ ) D 2 [ Φ ( μ τ ) ] n e 1 2 τ i = 1 D ( x i μ ) 2 i = 1 D [ 1 Φ ( x i μ τ ) ] R i [ 1 Φ ( T μ τ ) ] R 2 * l 2 D 2 log ( τ ) n log Φ ( μ τ ) 1 2 τ i = 1 D ( x i μ ) 2 i = 1 D R i log ( 1 Φ ( x i μ τ ) ) + R 2 * log ( 1 Φ ( T μ τ ) )
  • Case III:
    L 3 ( 2 π τ ) m 2 [ Φ ( μ τ ) ] n e 1 2 τ i = 1 m ( x i μ ) 2 i = 1 m [ 1 Φ ( x i μ τ ) ] R i l 3 m 2 log ( τ ) n log Φ ( μ τ ) 1 2 τ i = 1 m ( x i μ ) 2 i = 1 m R i log ( 1 Φ ( x i μ τ ) )
Then the likelihood and log-likelihood functions for three kinds of cases can be combined in a general expression as
L ( μ , τ | x ˜ ) τ J 2 e 1 2 τ i = 1 J ( x i μ ) 2 Φ μ τ n i = 1 J 1 Φ x i μ τ R i H ( μ , τ )
l ( μ , τ | x ˜ ) = l J 2 log ( τ ) n log Φ μ τ 1 2 τ i = 1 J ( x i μ ) 2 + i = 1 J R i log 1 Φ x i μ τ + h ( μ , τ ) .
where x ˜ = x 1 , x 2 , , x J denotes the observed data and J = k , R k = R 1 * , H ( μ , τ ) = 1 , h ( μ , τ ) = 0 for Case I; J = D , H ( μ , τ ) = 1 Φ T μ τ R 2 * , h ( μ , τ ) = R 2 * log 1 Φ T μ τ for Case II; J = m , R m = R 3 * , H ( μ , τ ) = 1 , h ( μ , τ ) = 0 for Case III.
Next, take the first derivatives of (1) for μ and τ respectively and make them equal to zero. A set of score equations can be attained as follows.
l μ = 1 τ i = 1 J η i + 1 τ i = 1 J R i ϕ ( η i ) 1 Φ ( η i ) n τ ϕ ( η ) Φ ( η ) + h 1 ( μ , τ ) = 0
l τ = J 2 τ + 1 2 τ i = 1 J η i 2 + 1 2 τ i = 1 J R i η i ϕ ( η i ) 1 Φ ( η i ) + n η 2 τ ϕ ( η ) Φ ( η ) + h 2 ( μ , τ ) = 0
where η = μ τ , η i = x i μ τ , η T = T μ τ and H 1 ( μ , τ ) = 0 , H 2 ( μ , τ ) = 0 for Case I and Case II; h 1 ( μ , τ ) = R 2 * τ ϕ ( η T ) 1 Φ ( η T ) , h 2 ( μ , τ ) = R 2 * η T 2 τ 2 ϕ ( η T ) 1 Φ ( η T ) for Case III.
The maximum likelihood estimates of unknown parameters are the solutions to the above equations. Apparently, the expressions of μ and τ are involved in a nonlinear problem and the analytic solutions are not available. Therefore, we have to depend on some numerical methods such as the N-R method and EM algorithm to approximate the values of unknown parameters.

2.1. Newton–Raphson Algorithm

Since the first and second-order derivations of the log-likelihood function are available, the Newton–Raphson algorithm is appropriate to maximize the log-likelihood function.
The second-order derivations of (1) for the parameters are given by
l 20 = 2 l μ 2 = J τ n τ Q Q 2 i = 1 J R i τ Q i Q i 2 + h 20 ( μ , τ )
l 11 = 2 l μ τ = 1 τ 3 / 2 i = 1 J η i + n 2 τ 3 / 2 Q + n μ 2 τ 2 Q Q 2 i = 1 J R i 2 τ 3 / 2 Q i i = 1 J R i η i 2 τ 3 / 2 Q i Q i 2 + h 11 ( μ , τ )
l 02 = 2 l τ 2 = J 2 τ 2 1 τ 2 i = 1 J η i 2 3 n μ 4 τ 5 / 2 Q n μ 2 4 τ 3 Q Q 2 i = 1 J 3 R i η i 4 τ 2 Q i i = 1 J R i η i 2 4 τ 2 Q i Q i 2 + h 02 ( μ , τ )
with
h 20 ( μ , τ ) = R 2 * τ Q T + Q T 2 h 11 ( μ , τ ) = R 2 * 2 τ 3 / 2 Q T + η T Q T + η T Q T 2 h 22 ( μ , τ ) = R 2 * 4 τ 2 3 η T Q T + η T 2 Q T + η T 2 Q T 2
where η = μ / τ , Q = ϕ ( η ) / Φ ( η ) , Q = ϕ ( η ) / Φ ( η ) , η i = ( x i μ ) / τ , Q i = ϕ ( η i ) / ( 1 Φ ( η i ) ) and Q i = ϕ ( η i ) / ( 1 Φ ( η i ) ) . Similarly, η T = ( T μ ) / τ , Q T = ϕ ( η T ) / ( 1 Φ ( η T ) ) and Q T = ϕ ( η T ) / ( 1 Φ ( η T ) ) .
The estimates can be updated by
μ ( l + 1 ) τ ( l + 1 ) = μ ( l ) τ ( l ) J ( μ ( l ) , τ ( l ) ) 1 μ ( l ) τ ( l )
where
J ( μ ( l ) , τ ( l ) ) = l 20 l 11 l 11 l 02 ( μ , τ ) = ( μ ( l ) , τ ( l ) )
The process repeats until | μ l + 1 μ l | < ε and | τ l + 1 τ l | < ε , where ε is pre-fixed as the tolerance limit.

2.2. Expectation Maximization Algorithm

Here EM algorithm discussed in [15], is employed to obtain the maximum likelihood estimates of μ and τ for the left-truncated normal distribution. It is an effective procedure to calculate the MLEs in the case of censored data. The expectation step (E-step) and maximization step (M-step) are two steps in the progress of the EM algorithm to calculate estimates for the given model. The former is to compute adequate information of censored data based on observed data, whereas the latter one is to re-estimate current parameters.
As mentioned earlier, we can only observe J failure units for three cases under consideration. Assume that X = ( X 1 , X 2 , , X J ) stands for the observed data which is subject to the left-truncated normal distribution at zero and Z = ( Z 1 , Z 2 , , Z J ) ( Z T j , j = 1 , 2 , , R T ) denotes censored data, where Z i j , i = 1 , 2,⋯, J, j = 1 , 2, ⋯, R J is the j-th unit that was withdrawn at the failure time X i and Z T j , j = 1 , 2, ⋯, R T , is the j-th unit that was withdrawn at the end time T e n d in Case II. Thus, we denote the complete data as C = ( X , Z ) , then the likelihood function of complete data under the GPHCS takes the form as
L c ( μ , τ ) = i = 1 J f ( x i ) j = 1 R i f ( z i j ) j = 1 R T f ( z T j ) = ( 2 π τ ) n 2 Φ μ τ n e 1 2 τ i = 1 J ( x i μ ) 2 + i = 1 J j = 1 R i ( z i j μ ) 2 + j = 1 R T ( z T j μ ) 2 .
Leaving out the constant term, the corresponding log-likelihood function is transformed into
l c ( μ , τ ) = n 2 log ( τ ) n log Φ ( η ) n μ 2 2 τ 1 2 τ i = 1 J x i 2 + i = 1 J j = 1 R i z i j 2 + j = 1 R T z T j 2 + μ τ i = 1 J x i + i = 1 J j = 1 R i z i j + j = 1 R T z T j .
Following two steps of the EM algorithm implemented in [16], we can get the ‘pseudo-log-likelihood’ function by utilizing the associated expected values to take place of the censored data.
E-step: Under the complete data, the ‘pseudo-log-likelihood’ function is
l s ( μ , τ ) = n 2 log ( τ ) n log Φ ( η ) n μ 2 2 τ + μ τ i = 1 J x i + i = 1 J j = 1 R i E ( z i j | z i j > x i ) + j = 1 R T * E ( z T j | z T j > T ) 1 2 τ i = 1 J x i 2 + i = 1 J j = 1 R i E ( z i j 2 | z i j > x i ) + j = 1 R T E ( z T j 2 | z T j > T ) .
For i = 1 , 2,⋯, J; j = 1 , 2,⋯, R J , the conditional expectations mentioned above are deduced as
E ( z i j | z i j > x i ) = μ + τ Q i = A ( x i , μ ( l ) , τ ( l ) ) E ( z i j 2 | z i j > x i ) = μ 2 + τ + 2 μ τ Q i + τ η i Q i = B ( x i , μ ( l ) , τ ( l ) ) ,
By analogy, for j = 1 , 2, ⋯, R T ,
E ( z T j | z T j > T ) = μ + τ Q T = C ( T , μ ( l ) , τ ( l ) ) E ( z T j 2 | z T j > T ) = μ 2 + τ + 2 μ τ Q T + τ η T Q T = D ( T , μ ( l ) , τ ( l ) ) .
where η = μ / τ , η i = ( x i μ ) / τ , η T = ( T μ ) / τ , Q i = ϕ ( η i ) / ( 1 Φ ( η i ) ) and Q T = ( T μ ) / τ .
By substituting (3) and (4) into (2), the ‘pseudo log-likelihood’ function becomes
l s ( μ , τ ) = n 2 log τ n log Φ ( μ τ ) n μ 2 2 τ + μ τ i = 1 J x i + i = 1 J j = 1 R i A ( x i , μ ( l ) , τ ( l ) ) + j = 1 R T C ( T , μ ( l ) , τ ( l ) ) 1 2 τ i = 1 J x i 2 + i = 1 J j = 1 R i B ( x i , μ ( l ) , τ ( l ) ) + j = 1 R T D ( T , μ ( l ) , τ ( l ) ) .
M-step: The major purpose of this step is to maximize the ‘pseudo log-likelihood’ function to calculate the next iterate. Let (5) take derivatives with regard to μ and τ respectively and equal them to zero, the corresponding score equations can be obtained as below. Suppose that ( μ ( l ) , τ ( l ) ) is the l-th iteration estimate of ( μ , τ ) , ( μ ( l + 1 ) , τ ( l + 1 ) ) is derived afterwards.
n μ ( l + 1 ) + n τ ( l ) ϕ ( μ ( l + 1 ) τ ( l ) ) Φ ( μ ( l + 1 ) τ ( l ) ) i = 1 J x i + i = 1 J j = 1 R i A ( x i , μ ( l ) , τ ( l ) ) + j = 1 R T C ( x i , μ ( l ) , τ ( l ) ) = 0
n τ ( l + 1 ) n μ ( l + 1 ) 2 n μ ( l + 1 ) τ ( l + 1 ) ϕ ( μ ( l + 1 ) τ ( l + 1 ) ) Φ ( μ ( l + 1 ) τ ( l + 1 ) ) + 2 μ ( l + 1 ) i = 1 J x i + i = 1 J j = 1 R i A ( x i , μ ( l ) , τ ( l ) ) + j = 1 R T C ( x i , μ ( l ) , τ ( l ) ) i = 1 J x i 2 + i = 1 J j = 1 R i B ( x i , μ ( l ) , τ ( l ) ) + j = 1 R T D ( x i , μ ( l ) , τ ( l ) ) = 0
Next upgrade μ ( l ) and τ ( l ) into μ ( l + 1 ) and τ ( l + 1 ) by soving the equations shown above. The E-step and M-step continue until the precision satisfies the tolerance limit that is fixed ahead of time μ ( l + 1 ) μ ( l ) + τ ( l + 1 ) τ ( l ) < ε . By that time, the convergence values are seen as the estimated values of μ and τ based on EM algorithm.

3. Confidence Interval Estimation

In this section, confidence interval estimates are provided for unknown parameters μ and τ using the MLE-based asymptotic confidence intervals (ACIs), the log-transformed MLE-based asymptotic confidence intervals (Log-CIs) and bootstrap confidence intervals (Boot-p CIs).

3.1. Asymptotic Confidence Intervals for Mles

Based on the asymptotic normal property of the MLEs, the asymptotic distribution of ( μ ^ M , τ ^ M ) is ( μ ^ M , τ ^ M ) N 2 ( μ , τ ) , I 1 ( μ , τ ) , where μ ^ M , τ ^ M are the MLEs of μ and τ , I 1 ( μ , τ ) stands for the inverse of the Fisher information matrix. Then the variance-covariance Fisher information matrix of ( μ , τ ) is able to obtain theoretically from the following of Fisher information matrix.
I 1 ( μ , τ ) = E 2 l μ i τ j 1 i + j = 2 , i , j = 0 , 1 , 2
At times it’s troublesome to figure out the exact Fisher information matrix. Therefore, I o b s 1 ( μ ^ M , τ ^ M ) the inverse of the observed Fisher information matrix is employed to make an estimation of I 1 ( μ , τ ) . Here we express the observed Fisher information matrix of unknown variables μ and τ as the following form.
I o b s ( μ , τ ) = 2 l μ 2 2 l μ τ 2 l τ μ 2 l τ 2 ( μ , τ ) = ( μ ^ M , τ ^ M ) .
The elements of the matrix (7) are calculated in Section 2.
Subsequently, the asymptotic variance-covariance matrix is derived from
I o b s 1 ( μ , τ ) = V a r ( μ ^ M ) C o v ( μ ^ M , τ ^ M ) C o v ( τ ^ M , μ ^ M ) V a r ( τ ^ M ) = v 11 v 12 v 21 v 22 .
Based on the matrix (8), the approximate variance of two parameters μ and τ can be derived, then we can obtain the 100 1 ζ % ACIs for these two parameters
μ ^ M Z 1 ζ / 2 v 11 , μ ^ M + Z 1 ζ / 2 v 11 , τ ^ M Z 1 ζ / 2 v 22 , τ ^ M + Z 1 ζ / 2 v 22 .
where Z 1 ζ / 2 satisfies P ( X Z 1 ζ / 2 ) = 1 ζ 2 when X follows the standard normal distribution.

3.2. Asymptotic Confidence Intervals for Log-Transformed Mles

Occasionally, the lower bound of ACI is less than zero. To conquer this setback, the logarithmic transformation and delta method is suggested to make sure that the lower bound of ACI is nonnegative.
Let α = ( μ , τ ) denotes the unknown parameter vector, in accordance to [17], the distribution of
log ( α ^ ) = log ( α ^ ) log ( α ) V a r ( log ( α ^ ) )
is approximately subject to the standard normal distribution, where α 1 = μ and α 2 = τ . Hence, a 100 1 ζ % logarithmic transformation confidence interval for α i is further constructed as
α i ^ exp Z 1 ζ / 2 V a r ( log α i ^ ) , α i ^ exp ( Z 1 ζ / 2 V a r ( log α i ^ ) ) , i = 1 , 2
where α i ^ is the MLE of α i and V a r ( log α i ^ ) = V a r ( α i ^ ) α i 2 = I 1 ( α ^ ) i i α i 2 , i = 1 , 2 .

3.3. Percentile Bootstrap Approach

The methods of asymptotic confidence intervals introduced above are both originated from the laws of large numbers. They do not perform effectively when faced with a small sample size. We suggest the percentile bootstrap approach to overcome the drawback and construct Boot-p CIs for μ and τ as well. According to [18], the following steps can be implemented to generate the bootstrap samples to develop Boot-p CIs.
Step 1:
Calculate the MLEs of two parameters μ and τ from the original generalized progressive hybrid censored sample.
Step 2:
Utilize the same censoring scheme ( T , n , m , k , R ) and μ ^ M and τ ^ M to generate a generalized progressive hybrid censored bootstrap sample.
Step 3:
Calculate the bootstrap estimators of μ and τ , denote as μ * and τ * , from the bootstrap sample of truncated normal distribution.
Step 4:
Perform Step 2 and Step 3, N times to obtain a sequence of bootstrap estimators.
Step 5:
Sort μ 1 * , μ 2 * , , μ N * and τ 1 * , τ 2 * , , τ N * in ascending order respectively. Then we get μ ( 1 ) * , μ ( 2 ) * , , μ ( N ) * and τ ( 1 ) * , τ ( 2 ) * , , τ ( N ) * .
Step 6:
The 100 1 ζ % Boot-p CIs of μ and τ are ( μ ( N ( ζ / 2 ) ) * , μ ( N ( 1 ζ / 2 ) ) * ) and ( τ ( N ( ζ / 2 ) ) * , τ ( N ( 1 ζ / 2 ) ) * ) respectively.

4. Bayes Estimation

Bayes point and interval estimations can be evaluated for unknown variables μ and τ belong to truncated normal distribution under the GPHCS in this section. All the Bayes estimations can be deduced theoretically under symmetric and asymmetric loss functions using Tierney and Kadane approximation and the importance sampling technique.

4.1. Prior and Posterior Distribution

Since a conjugate prior distribution for μ and τ does not exist, we make the same assumption as in [6] that μ and τ have a conditional bivariate prior distribution as the following form.
π ( μ , τ ) = π 1 ( τ ) π 2 ( μ | τ )
where
π 1 ( τ ) = d 2 c Γ ( c ) 1 τ c + 1 e d 2 τ
π 2 ( μ | τ ) = e b 2 τ ( μ a ) 2 2 π τ b Φ ( a b τ )
The prior distribution of τ is the inverse Gamma distribution IG(c, d / 2 ). While under the circumstance that τ is known, μ follows the truncated normal distribution TN(a, τ / b ) when the truncation point is zero. Here a, b, c, and d are treated as the hyper-parameters, whose domain range from zero to positive infinity.
Therefore the joint prior distribution is written as
π 0 ( μ , τ ) 1 Φ a b τ 1 τ c + 3 2 e 1 2 τ [ b ( μ a ) 2 + d ] .
Using (1) and (9), the posterior distribution of the left-truncated normal distribution TN( μ , τ ) at zero is
π ( μ , τ | x ˜ ) = P Φ μ τ n Φ a b τ 1 τ ( c + 1 + J 2 ) + 1 e 1 2 τ d + i = 1 J ( x i μ ) 2 + b ( μ a ) 2 × i = 1 J 1 Φ x i μ τ R i H ( μ , τ ) .
where P is the normalizing constant satisfying that
P 1 = 0 0 Φ μ τ n Φ a b τ 1 τ c + 3 + J 2 e 1 2 τ d + i = 1 J ( x i μ ) 2 + b ( μ a ) 2 × i = 1 J 1 Φ x i μ τ R i H ( μ , τ ) d μ d τ .
To start with, ( μ , τ ) denotes a function of μ and τ . So the expectation of ( μ , τ ) given x ˜ is shown as
E ( μ , τ ) | x ˜ = 0 0 ( μ , τ ) L ( μ , τ | x ˜ ) π 0 ( μ , τ ) d μ d τ 0 0 L ( μ , τ | x ˜ ) π 0 ( μ , τ ) d μ d τ .

4.2. Loss Functions

In the Bayesian framework, the Bayes estimate of a function ( μ , τ ) can be derived based on a prescribed loss function. We discuss three kinds of loss functions, namely squared error, general entropy, and linex loss functions.
  • Squared error loss function
Squared error loss (SEL) function is the most universally applicable loss function to obtain Bayes estimators of unknown parameters. We can express its definition as
L s ( ϖ , ϖ ^ ) = ( ϖ ^ ϖ ) 2 .
In the following cases, we denote ϖ ^ as an estimator of ϖ .
Under this circumstance, the corresponding Bayes estimator ϖ ^ s of ϖ can be derived from
ϖ s = E ( ϖ | x ˜ ) .
Then we can obtain the Bayes estimate ^ ( μ , τ ) s of ( μ , τ ) under SEL function
^ ( μ , τ ) s = 0 0 L ( μ , τ | x ˜ ) ( μ , τ ) π 0 ( μ , τ ) d μ d τ 0 0 L ( μ , τ | x ˜ ) π 0 ( μ , τ ) d μ d τ .
  • General entropy loss function
The expression of general entropy loss (GEL) function is
L e ( ϖ , ϖ ^ ) = ϖ ^ ϖ q q l o g ϖ ^ ϖ 1 , q 0 .
Under this circumstance, the Bayes estimator ϖ ^ e of ϖ can be derived from
ϖ e = E ( ϖ q | X ) 1 q .
Finally, the Bayes estimate ^ ( μ , τ ) e of ( μ , τ ) results to be the following form based on GEL function.
^ ( μ , τ ) e = 0 0 ( μ , τ ) q L ( μ , τ | x ˜ ) π 0 ( μ , τ ) d μ d τ 0 0 L ( μ , τ | x ˜ ) π 0 ( μ , τ ) d μ d τ 1 q .
  • Linex loss function
The linex loss (LL) function is defined in the form as
L l ( ϖ , ϖ ^ ) = e ( ϖ ^ ϖ ) ( ϖ ^ ϖ ) 1 , 0 .
Under this circumstance, the corresponding Bayes estimator ϖ ^ l of ϖ can be derived from
ϖ l = 1 E ( e ϖ | x ˜ ) .
Next, the Bayes estimate ^ ( μ , τ ) l of ( μ , τ ) under LL function is computed by
^ ( μ , τ ) l = 1 log 0 0 e ( μ , τ ) π 0 ( μ , τ ) L ( μ , τ | x ˜ ) d μ d τ 0 0 L ( μ , τ | x ˜ ) π 0 ( μ , τ ) d μ d τ .
Obviously, the Bayes estimatons of unknown parameters μ and τ in three kinds of loss functions cannot be expressed in closed forms. For this reason, we derive the Bayes estimations by the means of Tierney and Kadane method, as well as the importance sampling procedure.

4.3. Tierney and Kadane Method

It is hard to reduce the Bayes estimations into closed forms in the shape of the ratio of two integrals. Ref. [19] introduced an alternative method to approximate such ratio of integrals to derive the Bayes estimates of unknown parameters. It is just a simple Taylor approximation around the maximum a posteriori estimate up to second order (also known as saddle-point approximation, see [20]). We regard ( μ , τ ) as a function of μ and τ . Then the approximation of Tierney and Kadane approach is summarized as follows.
λ ( μ , τ ) = l ( μ , τ | X ) + log ( π 0 ( μ , τ ) ) n , λ * ( μ , τ ) = λ ( μ , τ ) + log ( ( μ , τ ) ) n
Based on this method, (11) is re-expressed as
E ( μ , τ ) | x ˜ = 0 0 e n λ * ( μ , τ ) d μ d τ 0 0 e n λ ( μ , τ ) d μ d τ = Λ * Λ e n [ λ * ( μ ^ λ * , τ ^ λ * ) λ ( μ ^ λ , τ ^ λ ) ]
where ( μ ^ λ , τ ^ λ ) and ( μ ^ λ * , τ ^ λ * ) are the maximum points of λ ( μ , τ ) and λ * ( μ , τ ) , respectively. The inverse of the negative Hessian matrix of λ ( μ , τ ) and λ * ( μ , τ ) at ( μ ^ λ , τ ^ λ ) and ( μ ^ λ * , τ ^ λ * ) are denoted as Λ and Λ * repectively.
For the truncated normal distribution, we have
λ ( μ , τ ) = 1 n c + 3 2 + J 2 log ( τ ) 1 2 τ b ( μ a ) 2 + d + i = 1 J ( x i μ ) 2 + i = 1 J R i log 1 Φ ( x i μ n ) n log Φ ( μ τ ) log Φ ( a b τ ) + h ( μ , τ )
Differentiating (16) with regard to μ and τ so as to obtain μ ^ λ , τ ^ λ , we can obtain the following equations
λ μ = 1 n 1 τ ( b + J ) μ a b i = 1 J x i + i = 1 J R i τ ϕ ( η i ) 1 Φ ( η i ) n τ ϕ ( η ) Φ ( η ) + h 1 ( μ , τ ) = 0
λ τ = 1 n c + 3 2 + J 2 τ + 1 2 τ 2 i = 1 J ( x i μ ) 2 + d + b ( μ a ) 2 + i = 1 J R i η i 2 τ ϕ ( η i ) 1 Φ ( η i ) n μ 2 τ 3 / 2 ϕ ( η ) Φ ( η ) + a b 2 τ 3 / 2 ϕ ( a b τ ) Φ ( a b τ ) + h 2 ( μ , τ ) = 0
Thus, we can calculate Λ , which is derived from
Λ = 2 λ μ 2 2 λ τ 2 2 λ μ τ 2 λ τ μ ( μ ^ λ , τ ^ λ ) 1
where
2 λ μ 2 = 1 n b + J τ i = 1 J R i τ ϕ ( η i ) 1 Φ ( η i ) + ϕ ( η i ) 1 Φ ( η i ) 2 n τ ϕ ( η ) Φ ( η ) ϕ ( η ) Φ ( η ) 2 + h 20 ( μ , τ )
2 λ μ τ = 1 n ( b + J ) μ a b i = 1 J x i τ 2 i = 1 J R i η i 2 τ 3 / 2 ϕ ( η i ) 1 Φ ( η i ) + ϕ ( η i ) 1 Φ ( η i ) 2 i = 1 J R i 2 τ 3 / 2 ϕ ( η i ) 1 Φ ( η i ) + n μ 4 τ 2 ϕ ( η ) Φ ( η ) ϕ ( η ) Φ ( η ) 2 + n 2 τ 3 / 2 ϕ ( η ) Φ ( η ) + h 11 ( μ , τ )
2 λ τ 2 = 1 n c + 3 2 + J 2 τ 2 1 τ 3 b ( μ a ) 2 + d + i = 1 J ( x i μ ) 2 i = 1 J R i η i 2 4 τ 2 ϕ ( η i ) 1 Φ ( η i ) + ϕ ( η i ) 1 Φ ( η i ) 2 i = 1 J 3 R i η i 4 τ 2 ϕ ( η i ) 1 Φ ( η i ) n μ 2 4 τ 3 ϕ ( η ) Φ ( η ) ϕ ( η ) Φ ( η ) 2 3 n μ 4 τ 5 / 2 ϕ ( η ) Φ ( η ) 3 a b 4 τ 5 / 2 ϕ ( a b τ ) Φ ( a b τ ) a 2 b 4 τ 3 ϕ ( a b τ ) Φ ( a b τ ) ϕ ( a b τ ) Φ ( a b τ ) 2 + h 02 ( μ , τ )
Note that Λ * and λ * depend on ( μ , τ ) .
For the SEL function ( μ , τ ) = μ , then the corresponding function λ μ s * ( μ , τ ) is obtained as
λ μ s * ( μ , τ ) = λ ( μ , τ ) + log μ n
In the following, μ ^ λ * , τ ^ λ * are obtained by solving the following two equations.
λ μ s * μ = λ μ + 1 n μ = 0 , λ μ s * τ = λ τ = 0
Then we have Λ μ s * , which is given by
Λ μ s * = 2 λ μ s * μ 2 2 λ μ s * τ 2 2 λ μ s * μ τ 2 λ μ s * τ μ 1
where
2 λ μ s * μ 2 = 2 λ μ 2 1 n μ 2 , 2 λ μ s * τ μ = 2 λ μ s * μ τ = 2 λ μ τ , 2 λ μ s * τ 2 = 2 λ τ 2 .
Using the equations above, the Bayes estimator of μ comes to be
μ ^ s = Λ μ s * Λ e n [ λ μ s * ( μ ^ λ * , τ ^ λ * ) λ ( μ ^ λ , τ ^ λ ) ]
Going through a similar process, under the same loss function the Bayesian estimator of τ is achieved.
As for the GEL function, we consider ( μ , τ ) = μ q for μ and then the corresponding function λ μ e * ( μ , τ ) is expressed as
λ μ e * ( μ , τ ) = λ ( μ , τ ) q log μ n
Subsequently, the following equations are solved to obtain ( μ ^ λ * , τ ^ λ * )
λ μ e * μ = λ μ q n μ = 0 , λ μ e * τ = λ τ = 0
Then, we have Λ μ e * , which is given by
Λ μ e * = 2 λ μ e * μ 2 2 λ μ e * τ 2 2 λ μ e * μ τ 2 λ μ e * τ μ 1
where
2 λ μ e * μ 2 = 2 λ μ 2 + q n μ 2 , 2 λ μ e * τ μ = 2 λ μ e * μ τ = 2 λ μ τ , 2 λ μ e * τ 2 = 2 λ τ 2 .
After that, the Bayes estimator of μ is
μ ^ e = Λ μ e * Λ e n q [ λ μ e * ( μ ^ λ * , τ ^ λ * ) λ ( μ ^ λ , τ ^ λ ) ]
Likewise, the Bayesian estimator of τ is realized based on this loss function.
When it comes to the LL function, ( μ , τ ) = e μ for μ is under consideration and then the corresponding function λ μ l * ( μ , τ ) is given by
λ μ l * ( μ , τ ) = λ ( μ , τ ) μ n
Later, μ ^ λ * and τ ^ λ * are derived by solving the following equations
λ μ l * μ = λ μ n = 0 , λ μ l * τ = λ τ = 0
Then, we have Λ μ l * , which is given by
Λ μ l * = 2 λ μ l * μ 2 2 λ μ l * τ 2 2 λ μ l * μ τ 2 λ μ l * τ μ 1
where
2 λ μ l * μ 2 = 2 λ μ 2 , 2 λ μ l * τ μ = 2 λ μ l * μ τ = 2 λ μ τ , 2 λ μ l * τ 2 = 2 λ τ 2 .
The Bayes estimator of μ turns into
μ ^ l = 1 log Λ μ l * Λ e n [ λ μ l * ( μ ^ λ * , τ ^ λ * ) λ ( μ ^ λ , τ ^ λ ) ]
Similarly, under the LL function, the Bayesian estimator of τ can be attained.

4.4. Importance Sampling Procedure

Since the T-K method fails to develop the interval estimators, an importance sampling procedure is proposed to construct Bayesian credible intervals in this part. The importance sampling procedure is an effective approach to attain Bayes estimates for TN( μ , τ ). In the meanwhile, the HPD intervals can be constructed through this method under generalized progressive hybrid censored data. Recall that the posterior distribution of μ and τ for μ > 0, τ > 0 is the following form
π ( μ , τ | x ˜ ) Φ μ τ n Φ a b τ 1 τ ( c + 1 + J 2 ) + 1 e 1 2 τ d + i = 1 J ( x i μ ) 2 + b ( μ a ) 2 × i = 1 J 1 Φ x i μ τ R i H ( μ , τ ) .
After some calculations, (18) is reduced to
π ( μ , τ | x ˜ ) I G τ c + J 2 , 1 2 i = 1 J x i + a 2 b + d ( a b + i = 1 J x i ) 2 b + J T N μ | τ a b + i = 1 J x i b + J , τ b + J W ( μ , τ )
where
W ( μ , τ ) = H ( μ , τ ) i = 1 J 1 Φ x i μ τ R i Φ a b + i = 1 J x i τ ( b + J ) Φ μ τ n Φ a b τ
Through the importance sampling procedure, we attain the Bayes estimates of μ and τ . The importance sampling procedure method is briefly described as follows:
Step 1:
Generate τ 1 from I G τ c + J 2 , 1 2 i = 1 J x i + a 2 b + d ( a b + i = 1 J x i ) 2 b + J ;
Step 2:
Sample μ 1 randomly from T N μ | τ a b + i = 1 J x i b + J , τ b + J ;
Step 3:
Perform Step 1 and Step 2, k times to obtain ( μ 1 , τ 1 ) , ( μ 2 , τ 2 ) , ⋯, ( μ k , τ k )
Step 4:
Now the Bayes estimation of ( μ , τ ) can be derived as follows.
^ I S = i = 1 k ( μ i , τ i ) W ( μ i , τ i ) i = 1 k W ( μ i , τ i )
Furthermore, the method introduced in [21] is applied to derive the 100 ( 1 ζ ) % Bayesian credible intervals for the given truncated normal distribution. Assume that 0 < ζ < 1 and ζ satisfies P ( ( μ , τ ) ζ ) = ζ . For a prefixed ζ , we attain the estimation of ζ and use it to establish the HPD intervals for ( μ , τ ) .
First, we substitute ( μ i , τ i ) with i for simplicity and suppose
ω i = W ( μ i , τ i ) i = 1 k W ( μ i , τ i ) , i = 1 , , k .
Sort ( 1 , ω 1 ) , , ( k , ω k ) into ( ( 1 ) , ω ( 1 ) ) , , ( ( k ) , ω ( k ) ) , where ( 1 ) < ⋯ < ( k ) and ( i ) is connected with ω i for i = 1 , , k . Then the estimator of ζ is ζ ^ = ( z ζ ) , where z ζ is an integer satisfying
i = 1 z ζ ω ( i ) ζ i = 1 z ζ + 1 ω ( i )
Hence, a 100 ( 1 ζ ) % Bayesian credible interval of ( μ , τ ) can be obtained by ( υ ^ , υ + 1 ζ ^ ) for ν = ω ( 1 ) , ω ( 1 ) + ω ( 2 ) , ⋯, i = 1 z ( 1 ζ ) ω ( i ) . Eventually, the 100 ( 1 ζ ) % HPD interval of ( μ , τ ) is obtained by ( ^ ν * , ^ ν * + 1 ζ ) where ν * satisafying
ν * + 1 ζ ^ ν * ^ ν + 1 ζ ^ ν ^
for all ν .

5. Simulation Study

5.1. Simulation

In attempts to analyze the performance of different methods introduced in previous sections, we utilize R software to conduct the simulation experiments. In light of the algorithm proposed in [22], a progressive type-II censored sample under any continuous distribution can be generated. By adopting this method, we can generate a generalized progressive hybrid censoring sample that is subject to the truncated normal distribution. Please see Algorithm 1.
Algorithm 1 Generate a generalized progressive hybrid censoring sample from truncated normal distribution.
1:
Pre-assign n,m and the censoring scheme ( R 1 , R 2 , , R m ) .
2:
Generate m independent observations W 1 , W 2 , , W m , all of which are subject to the standard uniform distribution U ( 0 , 1 ) .
3:
For pre-fixed censoring scheme, we compute V i = W i 1 / ( i + j = ( m i + 1 ) m R j ) , i = 1 , 2 , , m .
4:
Set U i = 1 j = m i + 1 m V j , for i = 1 , 2 , , m . Then U 1 , U 2 , , U m is a progressive Type-II censored sample with size m from a uniform distribution from zero to one.
5:
For known values of parameters μ and τ , the desired progressive Type-II censored sample out of the truncated normal distribution TN ( μ , τ ) , is X i = μ + τ Φ 1 ( 1 ( 1 U i ) Φ ( μ τ ) ) , for i = 1 , 2 , , m , where Φ 1 is the inverse cumulative distribution function of truncated normal distribution.
6:
If T < X k < X m , the generalized progressive hybrid censored sample is X = ( X 1 , X 2 , , X k ) .
7:
If X k < T < X m , J can be obtained which satisfies X J < T < X J + 1 , the generalized progressive hybrid censored sample is ( X 1 , X 2 , , X J ) .
8:
If X k < X m < T , the generalized progressive hybrid censoring sample is ( X 1 , X 2 , , X m ) .
Without loss of generality, we take μ = 0.5, τ = 1 for diverse values of T, n, m and k to generate a generalized progressive hybrid censored sample from the left-truncated normal distribution when truncation point is zero. Meanwhile, two kinds of censoring schemes are considered, which are:
Scheme I: R m = n m , R i = 0 , i = 1 , , m 1 .
Scheme II: R 1 = n m , R i = 0 , i = 2 , , m .
In point estimation, we compute the MLEs and Bayes estimates. For MLEs, the EM estimation method is employed to calculate the MLEs of μ and τ , where the nleqslv package with Broyden method in R software is applied to solve the nonlinear equations in the maximization step. Besides, the MLEs are derived by the N-R approach for comparative purposes. This method can be achieved by the function ‘optim’ in R software. Here the true values of parameters are considered as the initial values for the N-R method and the EM algorithm. The value of the tolerance limit ε is 0.0001 for all simulations. Table A1 and Table A2, shown in the Appendix A, compare the results of the EM and N-R methods in terms of average absolute biases (ABs), the associated mean squared errors (MSEs) and average number of iterations (AIs) until convergence. They are
A B s = 1 N i = 1 N | ϖ ^ i ϖ |
M S E s = 1 N i = 1 N ( ϖ ^ i ϖ ) 2 .
where N stands for the simulation times, ϖ is the true value, ϖ ^ i denotes the i-th estimate of ϖ .
Besides, the Bayesian estimates against diverse loss functions including SEL, LL, and GEL functions are obtained by the means of the T-K method and importance sampling procedure. In order to appropriate true values better, the values of hyper-parameters a = 0.05 , b = 1 , c = 8 , d = 0.3 in prior distributions might be a wise choice based on numerous experimental simulations. The desired estimates are obtained with = 0.35, 0.45 for the linex loss function. The general entropy loss function is considered with q = 0.8 and q = 1.1 to calculate the relative estimates. The ABs and MSEs are derived to evaluate the accuracy of the estimations. All the estimates are derived by replicating 1000 times in each case. The simulation results, Table A3 and Table A4, are shown in the Appendix A.
From these tables, some conclusions can be drawn:
(1)
There is no significant difference between the EM algorithm and N-R algorithm in terms of ABs and MSEs.
(2)
The N-R method takes fewer steps until convergence than the EM.
(3)
The results of = 0.35 show a bit more precise than those of = 0.45 for the linex loss function.
(4)
q = 0.8 is a wiser choice than q = 1.1 when the general entropy loss function is under consideration.
(5)
Set T, n, m and k invariant, Scheme II in which the censored units happen when the first failure is observed shows a more precise estimate than Scheme I for most cases.
(6)
Between Bayes estimation methods, neither the TK method nor the importance sampling technique performs consistently better since in some cases T-K estimates behave better and in some cases important sampling estimates perform better.
(7)
When the sample size n increases, the ABs of all estimates show downward trends.
(8)
When T, n and m keep invariable, the behaviors of both MLEs and Bayes estimates become better concerning the values of MSEs and ABs with the larger values of k.
(9)
When k, n and m keep invariable, the ABs and MSEs of all estimates fluctuate slightly, and the tendency is not significant with the growth of T.
(10)
When T, n and k keep invariable, both MLEs and Bayes estimates tend to have smaller MSEs with the larger values of m.
(11)
Overall, the Bayes estimates seem to be marginally better compared to the MLEs.
In interval estimation, we derive 95% confidence intervals (CIs) of parameters using the MLEs, log-transformed MLEs, Boot-p approach and 95% HPD intervals for Scheme I and Scheme II. We compare the coverage probabilities (CPs) and average lengths (ALs) of these interval estimates. The simulation results are reported in Table A5 and Table A6, shown in the Appendix A.
From these table, some conclusions are summarized as follows.
(1)
For confidence intervals, the Log-CIs perform much better than the ACIs in the sense of having higher coverage probabilities.
(2)
When the sample size n gets larger, the CPs of all interval estimates tend to decrease.
(3)
Boot-p CIs show higher coverage probabilities and narrower interval lengths than the ACIs and Log-CIs when the sample size is small.
(4)
With n, m, and k keeping invariant, the CPs and ALs of all estimates fluctuate slightly and the tendency is not significant with an increase of T.
(5)
The HPD intervals are slightly better than other interval estimates based on the CPs.
(6)
Scheme II usually performs better than Scheme I with regard to the CPs.

5.2. Real Data Analysis

One authentic example has been considered to demonstrate the performance of the proposed estimation approaches in this section. The data set is obtained from [23] (Lawless 1982, page 288), which shows the number of million revolutions before 23 ball bearings had failed. The data are given in Table 1 and shown as a histogram in Figure 3.
Prior to analyzing the example, one question that may arise is whether the data set comes from the truncated normal distribution. In order to validate the hypothesis, we fit the truncated normal distribution to the data set, competing with folded normal distribution (FN) and half-normal distribution (HN). The associated probability density functions of FN and HN for x > 0 are respectively written as follows.
f ( x ; μ , τ ) = e ( x μ ) 2 2 τ + e ( x + μ ) 2 2 τ 2 π τ , x > 0 ; μ > 0 , τ > 0 .
and
f ( x ; τ ) = 2 π τ e x 2 2 τ , x > 0 ; τ > 0 .
In order to assess the goodness of fit of these given models, we take advantage of log ( L ) , and Kolmogorov-Smirnov (K-S) statistic, defined by D n = sup x | F n ( x ) F ( x ) | , where L is the maximum of the likelihood function, n is the number of observations, F n ( x ) is the cumulative distribution function of the sample set, and F ( x ) is the assumed cumulative distribution function. These estimated values are shown in Table 2. Under the complete data, the classical estimates of these given distributions are obtained additionally. In view of the fact that the truncated normal distribution has the lowest values for K-S and log ( L ) statistics, it is reasonable to say we have no proof leading to rejection of the null hypothesis.
Then the following generalized progressive hybrid censored samples can be generated by setting m = 18 , and R 1 = 5 , R 2 = = R 18 = 0 . In this example, we take T = 1 , k = 16 for Case I, T = 1 , k = 12 for Case II, as well as T = 1.8 , k = 12 for Case III.
Table 3 presents the exact values of point estimations of μ and τ under the generalized progressive hybrid censored sample. The N-R algorithm and the EM algorithm are employed to derive the MLEs of parameters of the truncated normal distribution. We take the estimates under the complete data for initializing these two methods. For Bayes estimations, non-informative prior distributions, a = b = c = d = 0 , are applied to compute the exact values under symmetric and asymmetric loss functions based on T-K approximation and importance sampling procedure. We choose q = 0.8 and q = 1.1 for general entropy loss function, and fix = 0.35 and = 0.45 for the linex loss function. Table 4 shows the results of the 95% CIs and HPD intervals of unknown parameters μ and τ . As can be seen from the tables, one can notice that the results of MLEs, Bayes estimates under SEL function and importance sampling procedure are quite close to each other.

6. Conclusive Remarks

In the paper, we address the problem of making statistical inference for the truncated normal distribution when the truncation point is zero with the generalized progressive hybrid censored data. The N-R and EM algorithms are applied to calculate the MLEs of unknown parameters. The associated estimates are computed through numerous simulations by taking the true unknown parameters as an initial guess. Further, making use of the asymptotic normality property of the maximum likelihood estimator, we construct the 95% ACIs as well. As for the small sample size, Boot-p method is suggested to develop the Boot-p intervals. Considering different kinds of loss functions, we derive Bayes estimates through T-K approximation and the importance sampling procedure. The latter one is also employed to construct the Bayesian credible intervals. Extensive simulations are conducted to examine how the proposed approaches work. It is notable that if proper prior information is available on unknown parameters then corresponding Bayes estimates are superior to respective MLEs. Among intervals, Bayesian credible intervals are slightly better than other invertals. One authentic example is studied for illustrative purposes. The considered model is found to be suitable for this case and the proposed approaches perform well.
While we consider the statistical inference of the truncated normal distribution at zero in this paper, the proposed approaches can be extended to the doubly truncated normal distribution and the singly truncated normal distribution at any fixed truncation point. The statistical inference on the doubly truncated distribution with unknown truncation points in [24] provides a good starting point for discussion and further research. Extensive work needs to be carried out in this direction.

Author Contributions

Investigation, X.Z.; Supervision, W.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Project 202110004003 which was supported by National Training Program of Innovation and Entrepreneurship for Undergraduates.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Simulation Results

Table A1. Comparisons between EM and N-R algorithms with μ = 0.5, τ = 1 when T = 0.4.
Table A1. Comparisons between EM and N-R algorithms with μ = 0.5, τ = 1 when T = 0.4.
nmkSchMethodABs( μ )ABs( τ )MSEs( μ )MSEs( τ )AIs
302016IEM0.17660.31790.05030.156321.35
N-R0.17080.30030.04550.13857.39
IIEM0.17370.30530.04730.145921.03
N-R0.18140.27230.05300.11686.73
18IEM0.16410.28020.04230.125420.30
N-R0.16440.29570.04210.13497.11
IIEM0.16420.29570.04340.129820.57
N-R0.17860.26540.05080.10996.43
302516IEM0.17170.29920.04580.131221.35
N-R0.17320.30730.04770.14597.43
IIEM0.16960.31600.04530.152021.37
N-R0.17180.27900.04660.12206.96
18IEM0.16720.28850.04380.125320.68
N-R0.16270.29570.04200.13497.13
IIEM0.16140.29290.04170.132420.81
N-R0.17130.27790.04530.11686.79
605038IEM0.11510.20350.02050.064919.08
N-R0.11520.19910.02030.06146.58
IIEM0.11590.20040.02130.061618.85
N-R0.12200.18560.02300.05526.27
42IEM0.11290.18540.01960.053818.58
N-R0.10920.18090.01890.05296.30
IIEM0.10900.18620.01840.053418.73
N-R0.12280.18470.02340.05316.12
605538IEM0.11380.19450.02070.059518.99
N-R0.11680.19640.02180.05976.54
IIEM0.11820.20810.02200.066619.14
N-R0.10800.17790.01850.04976.17
42IEM0.10890.18670.01850.053918.63
N-R0.10950.18640.01920.05526.35
IIEM0.10430.18200.01730.050218.62
N-R0.10990.18020.01950.05026.17
1008062IEM0.08930.15750.01230.038418.21
N-R0.08930.15490.01260.03746.25
IIEM0.08890.15360.01250.036718.14
N-R0.09780.15050.01480.03495.99
70IEM0.08370.14630.01090.033817.87
N-R0.08520.14520.01150.03326.11
IIEM0.08510.14560.01140.032918.05
N-R0.09110.13990.01280.03135.84
1009062IEM0.09110.15470.01320.366618.29
N-R0.08830.15470.01240.03676.27
IIEM0.08510.15630.01120.038618.17
N-R0.09020.15040.01260.0366.07
70IEM0.08720.14350.01180.031517.92
N-R0.08340.14240.01080.03166.1
IIEM0.08180.14620.01090.033417.83
N-R0.08690.13730.01190.03035.95
Table A2. Comparisons between EM and N-R algorithms with μ = 0.5, τ = 1 when T = 0.6.
Table A2. Comparisons between EM and N-R algorithms with μ = 0.5, τ = 1 when T = 0.6.
nmkSchMethodABs( μ )ABs( τ )MSEs( μ )MSEs( τ )AIs
302016IEM0.17440.30390.04780.136021.60
N-R0.17190.30430.04590.14357.37
IIEM0.17390.30640.04760.144020.76
N-R0.18570.27370.05520.11746.75
18IEM0.16480.26850.04320.110220.25
N-R0.16710.30060.04360.14027.18
IIEM0.16980.29520.04450.136720.79
N-R0.18280.25530.05220.10016.44
302516IEM0.17480.31590.04870.151621.17
N-R0.16390.30880.04270.14427.41
IIEM0.17960.31190.05130.153821.24
N-R0.16780.27680.04380.11756.76
18IEM0.16780.28540.04320.128820.42
N-R0.16870.28650.04360.12946.78
IIEM0.16140.29290.04170.132420.81
N-R0.17130.27790.04530.11686.79
605038IEM0.11190.19790.01990.060118.95
N-R0.11250.20350.01990.06336.61
IIEM0.11270.19250.02010.057018.72
N-R0.12190.18640.02310.05536.30
42IEM0.10880.18660.01870.054818.46
N-R0.11010.18330.01860.05216.32
IIEM0.10980.18460.01840.054518.53
N-R0.12270.17560.02290.04936.06
605538IEM0.11270.19220.01960.056218.76
N-R0.11730.20170.02170.06286.59
IIEM0.11520.19990.02050.062418.99
N-R0.11870.18690.02200.05446.44
42IEM0.10940.18570.01830.054118.50
N-R0.11150.18320.01990.05166.30
IIEM0.11270.18720.01970.056219.01
N-R0.11290.18420.01980.05426.21
1008062IEM0.09030.15570.01310.037918.12
N-R0.08730.15820.01220.03826.28
IIEM0.08760.16030.01210.040118.36
N-R0.09500.15400.01410.03746.06
70IEM0.08640.14540.01160.032517.88
N-R0.08570.13960.01180.03096.10
IIEM0.08980.14160.01250.031717.80
N-R0.09000.13310.01280.02805.81
1009062IEM0.09340.15130.01350.036217.99
N-R0.08850.15810.01240.03986.26
IIEM0.08390.15270.01100.037018.09
N-R0.09300.14960.01360.03476.14
70IEM0.08710.14210.01180.031317.67
N-R0.08610.14720.01630.03356.12
IIEM0.08260.13930.01080.030817.81
N-R0.08360.14040.01110.03125.95
Table A3. ABs and MSEs (within bracket) of different estimates with μ = 0.5, τ = 1 when T = 0.4.
Table A3. ABs and MSEs (within bracket) of different estimates with μ = 0.5, τ = 1 when T = 0.4.
nmkSchParSELLLGELIS
= 0.35 = 0.45q = 0.8q = 1.1
302016I μ 0.78930.16610.39530.16740.17820.0951
(0.6679)(0.0398)(0.2963)(0.0646)(0.0775)(0.0028)
τ 0.19530.15300.19610.08110.36200.1523
(0.0496)(0.0530)(0.0403)(0.0127)(0.1515)(0.0157)
II μ 0.77300.16470.33240.15040.18730.0947
(0.6410)(0.0365)(0.0915)(0.0533)(0.0802)(0.0072)
τ 0.16770.15290.19610.09840.24040.1505
(0.0553)(0.0328)(0.0402)(0.0191)(0.0718)(0.0153)
18I μ 0.76620.16270.38000.14230.17090.0942
(0.6308)(0.0856)(0.2318)(0.0275)(0.0350)0.0029
τ 0.17100.17190.18480.07300.09060.1570
(0.0398)(0.0237)(0.0357)(0.0144)(0.0158)(0.0126)
II μ 0.65880.14760.23540.14220.17050.0891
(0.4745)(0.0304)(0.1213)(0.0271)(0.0339)(0.0022)
τ 0.15630.14490.18380.06980.04120.1525
(0.0384)(0.0283)(0.0322)(0.0138)(0.0112)(0.0163)
302516I μ 0.74700.11700.33260.14320.16740.0812
(0.5952)(0.0175)(0.2722)(0.0441)(0.0646)(0.0037)
τ 0.19800.15520.19450.07300.09060.1661
(0.0538)(0.0483)(0.0396)(0.0144)(0.0158)(0.0155)
II μ 0.62770.10080.27100.14150.16430.0638
(0.4396)(0.0167)(0.2059)(0.0413)(0.0636)(0.0031)
τ 0.16750.15200.17610.07310.11670.1545
(0.0383)(0.0472)(0.0240)(0.0119)(0.0246)(0.0049)
18I μ 0.58220.05380.22580.14080.16640.0576
(0.3814)(0.0042)(0.1235)(0.0256)(0.0337)(0.0028)
τ 0.18150.14050.18370.07120.04820.1583
(0.0445)(0.0284)(0.0352)(0.0117)(0.0131)(0.0130)
II μ 0.50820.05570.19920.14020.16580.0529
(0.3019)(0.0040)(0.2391)(0.0250)(0.0333)(0.0082)
τ 0.15040.14030.14730.07090.04340.1437
(0.0372)(0.0306)(0.0309)(0.0111)(0.0124)(0.0099)
605038I μ 0.72150.12080.35170.09580.16890.0911
(0.5395)(0.0184)(0.1457)(0.0232)(0.0457)(0.0419)
τ 0.09870.09400.24460.02120.03880.0989
(0.0146)(0.0121)(0.0643)(0.0070)(0.0139)(0.0752)
II μ 0.71190.11920.35030.09030.15670.0853
(0.5248)0.0204(0.1447)(0.0145)(0.0443)(0.0420)
τ 0.09010.08600.16200.02100.03820.1423
(0.0221)(0.0111)(0.0312)(0.0072)(0.0133)(0.0704)
42I μ 0.60140.12070.28420.08910.10950.0650
(0.3794)(0.0196)(0.1041)(0.0226)(0.0244)(0.0371)
τ 0.09790.09940.19920.01140.03610.0600
(0.0147)(0.0025)(0.0444)(0.0067)(0.0149)(0.0599)
II μ 0.56980.11210.27930.08570.09010.0623
(0.3432)(0.0278)(0.1003)(0.0257)(0.0294)(0.0291)
τ 0.09060.08860.10720.01110.03590.0344
(0.0233)(0.0050)(0.0176)(0.0065)(0.0147)(0.0658)
605538I μ 0.71910.10480.34740.07500.09580.0835
(0.5353)(0.0361)(0.1427)(0.0190)(0.0230)(0.0219)
τ 0.09820.08920.24180.01950.03210.0929
(0.0151)(0.0104)(0.0628)(0.0067)(0.0143)(0.0731)
II μ 0.61930.09360.33910.06700.06780.0761
(0.4036)(0.0347)(0.1357)(0.0196)(0.0196)(0.0216)
τ 0.09780.08540.16270.02050.03590.0284
(0.0178)(0.0074)(0.0317)(0.0068)(0.0211)(0.0616)
42I μ 0.60070.09610.19760.03570.08390.0726
(0.3786)(0.0124)(0.0639)(0.0185)(0.0217)(0.0458)
τ 0.09330.09290.19700.01070.03140.0569
(0.0135)(0.0013)(0.0433)(0.0062)(0.0138)(0.0594)
II μ 0.46170.11220.18790.03900.08200.0582
(0.2329)(0.0287)(0.0597)(0.0145)(0.0215)(0.0280)
τ 0.09220.08850.10720.01950.03220.0649
(0.0130)(0.0019)(0.0176)(0.0066)(0.0211)(0.0142)
1008062I μ 0.65730.09140.25920.09420.13740.0889
(0.4440)(0.0158)(0.0635)(0.0237)(0.0370)(0.0128)
τ 0.09800.07190.36200.12240.17620.0783
(0.0129)(0.0068)(0.1515)(0.0673)(0.0394)(0.0232)
II μ 0.54490.09030.18540.09330.11310.0790
(0.3092)(0.0158)(0.0371)(0.0227)(0.0328)(0.0198)
τ 0.31530.05860.24040.12230.14000.0693
(0.1125)(0.0030)(0.0718)(0.0626)(0.0398)(0.0102)
70I μ 0.52700.06060.23370.08660.12740.0680
(0.2887)(0.0113)(0.0771)(0.0172)(0.0399)(0.0135)
τ 0.07660.04520.09060.12190.16240.0780
(0.0089)(0.0029)(0.0158)(0.0673)(0.0396)(0.0235)
II μ 0.41080.06040.15920.07410.12440.0675
(0.1809)(0.0117)(0.0763)(0.0160)(0.0350)(0.0159)
τ 0.21020.04380.04120.09720.13770.0686
(0.0540)(0.0008)(0.0112)(0.0169)(0.0369)(0.0235)
1009062I μ 0.65310.08730.23930.09800.13030.0897
(0.4416)(0.0151)(0.0797)(0.0169)(0.0359)(0.0118)
τ 0.09490.06100.11670.65630.17210.0789
(0.0125)(0.0029)(0.0246)(0.4456)(0.0395)(0.0227)
II μ 0.41230.07580.20120.21250.12550.0716
(0.1837)(0.0122)(0.0714)(0.1698)(0.0360)(0.0240)
τ 0.11820.05100.08960.14020.16690.0685
(0.0212)(0.0030)(0.0154)(0.0851)(0.0390)(0.0212)
70I μ 0.52700.05390.15120.09550.11330.0676
(0.2800)(0.0053)(0.0880)(0.0169)(0.0898)(0.0232)
τ 0.07450.04880.04820.13110.17270.0728
(0.0084)(0.0008)(0.0131)(0.0766)(0.0947)(0.0236)
II μ 0.26390.02980.09210.09090.12660.0545
(0.0827)(0.0241)(0.0299)(0.0170)(0.0929)(0.0211)
τ 0.07210.04210.03980.12660.15480.0669
(0.0087)(0.0008)(0.0102)(0.0668)(0.0935)(0.0237)
Table A4. ABs and MSEs (within bracket) of different estimates with μ = 0.5, τ = 1 when T = 0.6.
Table A4. ABs and MSEs (within bracket) of different estimates with μ = 0.5, τ = 1 when T = 0.6.
nmkSchParSELLLGELIS
= 0.35 = 0.45q = 0.8q = 1.1
302016I μ 0.78110.11610.38330.15370.17380.0949
(0.6541)(0.0172)(0.2961)(0.0531)(0.0731)(0.0029)
τ 0.19560.15540.19850.07400.08960.1697
(0.0500)(0.0323)(0.0433)(0.0117)(0.0154)(0.0182)
II μ 0.62700.11470.38020.13880.08960.0905
(0.4396)(0.0167)(0.2438)(0.0372)(0.0154)(0.0028)
τ 0.16750.15180.19570.06530.03980.1687
(0.0549)(0.0321)(0.0400)(0.0107)(0.0102)(0.0164)
18I μ 0.67150.05510.23160.14760.17190.0889
(0.4910)(0.0039)(0.1054)(0.0488)(0.0728)(0.0025)
τ 0.18310.14030.19730.03810.05260.1573
(0.0455)(0.0306)(0.0420)(0.0091)(0.0192)(0.0127)
II μ 0.50670.05560.19590.14660.17000.0805
(0.3001)(0.0041)(0.1296)(0.0412)(0.0711)(0.0050)
τ 0.18150.14020.18670.16610.04200.1547
(0.0422)(0.0299)(0.0365)(0.0381)(0.0112)(0.0123)
302516I μ 0.77930.18950.22970.13880.16190.0629
(0.6508)(0.0456)(0.1262)(0.0372)(0.0638)(0.0074)
τ 0.19750.16250.18460.07180.11940.1557
(0.0508)(0.0394)(0.0356)(0.0119)(0.0263)(0.0122)
II μ 0.50420.16080.17090.13850.16100.0605
(0.2968)(0.0840)(0.1049)(0.0372)(0.0638)(0.0077)
τ 0.14420.53280.14660.00260.09110.1241
(0.0365)(0.3006)(0.0227)(0.0075)(0.0158)(0.0046)
18I μ 0.67170.17490.22490.11670.15920.0575
(0.4918)(0.0328)(0.1260)(0.0246)(0.0576)(0.0159)
τ 0.17590.14190.19370.07590.05340.1592
(0.0422)(0.0291)(0.0392)(0.0121)(0.0124)(0.0113)
II μ 0.41760.15830.39920.10970.15640.0317
(0.2249)(0.0426)(0.2387)(0.0326)(0.0573)(0.0373)
τ 0.14050.13580.14190.04580.03790.1433
(0.0363)(0.0238)(0.0262)(0.0099)(0.0110)(0.0096)
605038I μ 0.72020.11930.35160.09450.19180.0862
(0.5371)0.0203(0.1445)(0.0231)(0.0979)(0.0416)
τ 0.09990.09410.24230.02380.03840.0451
(0.0151)(0.0124)(0.0631)(0.0071)(0.0149)(0.0689)
II μ 0.71030.11820.34910.09470.16050.8068
(0.5229)0.0190(0.1432)(0.0239)(0.0283)(0.0415)
τ 0.09750.08510.06020.02360.03790.0612
(0.0149)(0.0110)(0.0102)(0.0070)(0.0105)(0.0677)
42I μ 0.60290.12750.27320.09600.17000.0726
(0.3818)0.0104(0.0975)(0.0292)(0.0901)(0.1899)
τ 0.09120.09870.20110.01060.03700.0588
(0.0134)(0.0103)(0.0451)(0.0065)(0.0140)(0.0700)
II μ 0.57080.11290.27000.08200.16980.0562
(0.3435)0.0274(0.0945)(0.0215)(0.0900)(0.0149)
τ 0.09100.08840.03240.02500.03960.0612
(0.0112)(0.0267)(0.0091)(0.0059)(0.0230)(0.0834)
605538I μ 0.71900.10460.34710.07730.15340.0719
(0.5364)0.0361(0.1421)(0.0249)(0.0398)(0.0210)
τ 0.10020.08920.24310.01970.03200.0808
(0.0149)(0.0259)(0.0634)(0.0066)(0.0143)(0.0777)
II μ 0.61980.09350.34140.09030.13150.6755
(0.4035)0.0448(0.1385)(0.0245)(0.0326)(0.0203)
τ 0.09760.08460.16180.02070.03200.0899
(0.0142)(0.0047)(0.0383)(0.0066)(0.0143)(0.0594)
42I μ 0.60000.12600.19410.07930.16980.0713
(0.3784)0.0124(0.0604)(0.0251)(0.0913)(0.0478)
τ 0.09120.09270.19420.01010.03150.0856
(0.0134)(0.0271)(0.0430)(0.0064)(0.0137)(0.0872)
II μ 0.46040.11240.19330.05290.06590.0376
(0.2308)0.0280(0.0595)(0.0222)(0.0187)(0.0146)
τ 0.09110.08840.10800.01870.05290.1554
(0.0129)(0.0201)(0.0180)(0.0066)(0.0215)(0.0799)
1008062I μ 0.65590.08850.25900.09870.12690.0893
(0.4429)(0.0153)(0.1089)(0.0169)(0.0993)(0.0120)
τ 0.09200.09460.08960.13540.16480.0856
(0.0137)(0.0062)(0.0154)(0.0657)(0.0938)(0.0209)
II μ 0.54760.07100.18930.08820.12270.0813
(0.3123)(0.0111)(0.0394)(0.0160)(0.0915)(0.0199)
τ 0.31320.09320.06760.12320.15500.0725
(0.1115)(0.0042)(0.0172)(0.0672)(0.0970)(0.0215)
70I μ 0.52630.06980.23990.09970.12330.0732
(0.2883)(0.0100)(0.1005)(0.0157)(0.0925)(0.0224)
τ 0.07820.05900.05260.13320.17450.0780
(0.0091)(0.0053)(0.0192)(0.0660)(0.0944)(0.0234)
II μ 0.40960.06950.12280.08380.10210.0679
(0.1797)(0.0109)(0.0373)(0.0157)(0.0895)(0.0161)
τ 0.20910.03710.04200.11140.15070.0690
(0.0545)(0.0008)(0.0112)(0.0683)(0.0943)(0.0235)
1009062I μ 0.65490.05430.22900.08000.13150.0892
(0.4404)(0.0054)(0.0737)(0.0150)(0.0920)(0.0121)
τ 0.09770.06680.11940.12030.16900.0771
(0.0130)(0.0043)(0.0263)(0.0675)(0.0952)(0.0240)
II μ 0.42170.04230.20230.09450.12390.0720
(0.1917)(0.0078)(0.0812)(0.0164)(0.0994)(0.0236)
τ 0.07320.05820.09110.14510.15670.0726
(0.0087)(0.0010)(0.0158)(0.0644)(0.0968)(0.0211)
70I μ 0.52580.04110.15420.08450.13070.0683
(0.2854)(0.0075)(0.0968)(0.0579)(0.0906)(0.0237)
τ 0.07380.03630.05340.12850.16950.0634
(0.0080)(0.0008)(0.0124)(0.0666)(0.0952)(0.0239)
II μ 0.25370.03180.08670.07940.12480.0664
(0.0763)(0.0004)(0.0285)(0.0099)(0.0938)(0.0239)
τ 0.11680.02100.03790.11250.16160.0689
(0.0208)(0.0016)(0.0110)(0.0682)(0.0390)(0.0239)
Table A5. CPs and ALs (within bracket) of 95% CIs for the parameters with μ = 0.5, τ = 1, when T = 0.4.
Table A5. CPs and ALs (within bracket) of 95% CIs for the parameters with μ = 0.5, τ = 1, when T = 0.4.
nmkSchACIsLog-CIsBoot-p CIsHPD Intervals
μ τ μ τ μ τ μ τ
302016I91.00%85.32%91.14%91.23%91.36%91.40%91.18%91.12%
(0.8139)(1.4571)(0.8044)(1.5610)(0.8172)(1.3819)(0.8034)(1.5568)
II91.60%83.21%92.26%92.63%92.48%92.18%91.46%91.38%
(0.8134)(1.3467)(0.8634(1.4140)(0.8653)(1.2618)(0.8556)(1.3877)
18I92.60%85.73%90.71%91.03%91.82%92.18%91.14%91.92%
(0.7731)(1.3532)(0.7691)(1.4545)(0.7771)(1.3162)(0.7691)(1.4527)
II93.40%89.62%94.43%92.50%92.64%92.84%91.52%92.16%
(0.8485)(1.2287)(0.8429)(1.2999)(0.8384)(1.1713)(0.8407)(1.2888)
302516I92.70%85.10%91.82%91.92%92.03%91.70%91.72%91.60%
(0.8212)(1.3716)(0.8065)(1.5668)(0.8231)(1.4005)(0.7712)(1.4605)
II92.20%86.23%92.93%93.32%92.48%92.02%91.84%91.88%
(0.8172)(1.3527)(0.8198)(1.4811)(0.8272)(1.3351)(0.7948)(1.3854)
18I92.30%84.55%93.56%92.30%91.78%92.35%91.68%92.22%
(0.7684)(1.3399)(0.7656)(1.4376)(0.7724)(1.3021)(0.7681)(1.4486)
II0.92985.76%92.18%91.70%92.32%92.38%91.68%92.52%
(0.7910)(1.2766)(0.7928)(1.3786)(0.7897)(1.2489)(0.7920)(1.3751)
605038I92.57%88.64%93.07%93.62%93.20%91.54%93.50%93.80%
0.5479)(0.9431)(0.5459)(0.9824)(0.5368)(0.9170)(0.5158)(0.7909)
II93.21%89.90%94.16%93.74%93.70%92.80%93.62%93.82%
(0.5183)(0.8916)(0.5656)(0.9396)(0.5316)(0.8707)(0.5085)(0.7485)
42I93.53%90.21%94.08%94.03%96.24%92.50%93.78%94.08%
(0.5436)(0.9493)(0.5270)(0.9147)(0.5412)(0.8654)(0.5289)(0.7429)
II93.92%91.37%92.19%93.62%93.81%92.67%93.91%94.16%
(0.5090)(0.9156)(0.5517)(0.8698)(0.5194)(0.8479)(0.5153)(0.7259)
605538I93.27%90.23%94.12%93.33%93.50%92.20%93.67%94.48%
(0.5312)(0.8481)(0.5439)(0.9763)(0.5431)(0.9157)(0.4972)(0.7352)
II93.64%91.68%93.71%94.31%93.60%93.30%93.82%94.36%
(0.5292)(0.8462)(0.5559)(0.9686)(0.5345)(0.8917)(0.5019)(0.7353)
42I93.66%90.51%94.13%93.86%93.91%92.91%94.16%94.50%
(0.5150)(0.9174)(0.5323)(0.9334)(0.5273)(0.8644)(0.5066)(0.7076)
II93.91%91.03%94.35%94.51%94.15%93.73%94.40%94.93%
(0.5452)(0.7920)(0.5421)(0.9092)(0.5175)(0.8249)(0.4912)(0.7031)
1008062I93.25%90.28%94.80%93.90%93.52%90.50%94.86%95.08%
(0.4313)(0.8709)(0.4294)(0.7688)(0.4361)(0.7311)(0.4299)(0.6750)
II94.10%91.56%93.21%94.03%94.23%91.57%94.38%95.60%
(0.4386)(0.7537)(0.4476)(0.7270)(0.4172)(0.6800)(0.4148)(0.6714)
70I93.71%91.74%95.26%93.24%93.73%92.03%95.72%95.24%
(0.4129)(0.6561)(0.4131)(0.7126)(0.4219)(0.6807)(0.4138)(0.6603)
II94.82%92.36%93.33%92.53%94.96%92.40%94.34%95.52%
(0.4228)(0.6480)(0.4354)(0.6686)(0.3985)(0.6311)(0.4180)(0.6478)
1009062I94.5%91.93%95.28%95.30%94.56%92.59%95.74%95.64%
(0.4314)(0.8907)(0.4327)(0.7778)(0.4234)(0.7243)(0.4017)(0.6513)
II94.64%92.03%94.62%94.06%94.50%92.50%94.88%96.02%
(0.4311)(0.9071)(0.4387)(0.7569)(0.4209)(0.7078)(0.3929)(0.6138)
70I94.71%91.92%95.17%95.10%94.88%92.03%95.03%96.04%
(0.4380)(0.7111)(0.4152)(0.7195)(0.4198)(0.6256)(0.3987)(0.6131)
II94.83%92.65%95.63%93.70%95.01%92.80%95.97%96.78%
(0.4134)(0.6985)(0.4242)(0.6977)(0.3867)(0.5862)(0.3716)(0.5252)
Table A6. CPs and ALs (within bracket) of 95% CIs for the parameters with μ = 0.5, τ = 1, when T = 0.6.
Table A6. CPs and ALs (within bracket) of 95% CIs for the parameters with μ = 0.5, τ = 1, when T = 0.6.
nmkSchACIsLog-CIsBoot-p CIsHPD Intervals
μ τ μ τ μ τ μ τ
302016I92.21%85.12%92.80%91.37%91.86%91.54%91.28%91.36%
(0.8082)(1.4333)(0.8086)(1.5806)(0.8245)(1.3886)(0.8094)(1.3806)
II92.63%85.12%93.54%91.92%93.20%92.68%91.10%92.63%
(0.8552)(1.2903)(0.8567)(1.3925)(0.8575)(1.2579)(0.8610)(1.2051)
18I92.42%83.91%92.73%91.11%90.90%92.36%91.33%91.64%
(0.7682)(1.3373)(0.7644)(1.4355)(0.7752)(1.311)(0.7709)(1.2593)
II91.70%87.90%92.31%92.00%92.52%92.94%91.42%92.78%
(0.8416)(1.2125)(0.8379)(1.2806)(0.8335)(1.1767)(0.8422)(1.1927)
302516I91.66%85.00%91.93%92.73%91.42%91.58%91.34%91.85%
(0.8136)(1.4535)(0.8081)(1.5726)(0.8180)(1.3935)(0.8094)(1.3579)
II92.81%86.08%91.00%92.48%91.98%92.26%92.52%91.36%
(0.856)(1.2924)(0.8239)(1.4979)(0.8218)(1.3367)(0.8240)(1.2990)
18I91.80%85.53%92.71%91.40%91.62%91.48%92.14%92.14%
(0.7705)(1.3431)(0.7663)(1.4428)(0.7846)(1.3042)(0.7673)(1.2944)
II92.93%86.32%92.80%91.73%92.06%92.00%92.46%92.38%
(0.7971)(1.2958)(0.7905)(1.3683)(0.7961)(1.2453)(0.7911)(1.2373)
605038I93.33%89.20%95.40%94.60%93.40%94.53%93.46%94.76%
(0.5389)(0.9433)(0.5480)(0.9901)(0.5554)(0.9074)(0.5162)(0.7907)
II93.58%88.87%93.52%93.33%93.80%94.10%94.00%94.20%
(0.5491)(0.9340)(0.5672)(0.9454)(0.5378)(0.8759)(0.5066)(0.7491)
42I93.74%90.10%93.21%94.61%93.85%94.33%94.10%94.84%
(0.4133)(0.8990)(0.5286)(0.9198)(0.5426)(0.8560)(0.5290)(0.7463)
II94.00%91.34%95.73%93.82%94.00%94.50%94.20%94.62%
(0.5338)(0.9572)(0.5551)(0.8818)(0.5124)(0.8221)(0.5143)(0.7439)
605538I94.31%90.62%93.20%92.79%94.50%94.87%94.40%95.10%
(0.4308)(0.8645)(0.5422)(0.9686)(0.5301)(0.8556)(0.5077)(0.7355)
II94.22%90.62%93.10%93.64%94.31%94.80%94.55%95.50%
(0.4474)(0.8358)(0.5524)(0.9559)(0.5084)(0.8406)(0.4970)(0.7103)
42I93.50%90.71%93.66%94.00%93.58%94.93%94.42%95.78%
(0.5529)(0.8817)0.5572)(0.9157)(0.5263)(0.8129)(0.5031)(0.7025)
II94.92%91.63%93.00%94.44%94.81%94.90%94.70%96.08%
(0.4365)(0.8042)(0.5417)(0.9079)(0.5210)(0.7729)(0.4963)(0.6911)
1008062I93.70%90.91%93.91%93.85%93.70%91.00%94.00%95.20%
(0.4319)(0.7260)(0.4278)(0.7610)(0.4374)(0.6958)(0.4206)(0.6753)
II93.81%92.00%93.90%95.43%94.00%92.50%94.06%95.66%
(0.4390)(0.7574)(0.4509)(0.7375)(0.4125)(0.6858)(0.4112)(0.6711)
70I93.76%91.63%94.70%95.20%94.12%92.80%94.54%95.22%
(0.4125)(0.6491)(0.4128)(0.0717)(0.4191)(0.6545)(0.4038)(0.6472)
II94.06%92.22%94.50%94.36%94.31%92.28%94.68%95.10%
(0.4240)(0.6150)(0.4379)(0.6721)(0.4120)(0.6355)(0.4022)(0.6312)
1009062I94.45%92.10%94.00%93.57%94.20%93.07%94.66%96.40%
(0.4011)(0.7081)(0.4312)(0.7731)(0.4243)(0.7031)(0.4018)(0.6516)
II94.70%92.70%94.03%94.51%94.85%93.07%94.90%96.28%
(0.4380)(0.6986)(0.4377)(0.7531)(0.4325)(0.6733)(0.3921)(0.6402)
70I95.31%92.16%93.76%94.11%95.55%92.90%95.81%96.38%
(0.4008)(0.7392)(0.4142)(0.7164)(0.3951)(0.6256)(0.3925)(0.6113)
II95.50%93.95%94.81%93.90%95.89%94.10%96.01%96.14%
(0.4313)(0.6955)(0.4236)(0.6954)(0.3893)(0.6090)(0.3756)(0.5907)

References

  1. Cohen, A. Estimating the mean and variance of normal populations from singly truncated and doubly truncated samples. Ann. Math. Stat. 1950, 21, 557–569. [Google Scholar] [CrossRef]
  2. Cohen, A. On estimating the mean and variance of singly truncated normal frequency distributions from the first three sample moments. Ann. Inst. Stat. Math. 1951, 3, 37–44. [Google Scholar] [CrossRef]
  3. Cohen, A. Tables for maximum likelihood estimates: Singly truncated and singly censored samples. Technometrics 1961, 3, 535–541. [Google Scholar] [CrossRef]
  4. Whitten, B.J.; Sundaraiyer, V. A pseudo-complete sample technique for estimation from censored samples. Commun. Stat. Theory Methods 1988, 17, 2239–2258. [Google Scholar] [CrossRef]
  5. Cohen, A. Truncated and Censored Samples: Theory and Applications; CRC Press: New York, NY, USA, 1991. [Google Scholar]
  6. Lodhi, C.; Tripathi, Y.; Rastogi, M. Estimating the parameters of a truncated normal distribution under progressive type-II censoring. Commun. Stat. Simul. Comput. 2019. [Google Scholar] [CrossRef]
  7. Cha, J.; Cho, B.; Sharp, J. Rethinking the truncated normal distribution. Int. J. Exp. Des. Process Optim. 2013, 3, 27–63. [Google Scholar] [CrossRef]
  8. Akahira, M. Maximum likelihood estimation for a one-sided truncated family of distributions. Jpn. Stat. Data Sci. 2020. [Google Scholar] [CrossRef]
  9. Kundu, D.; Joarder, A. Analysis of Type-II progressively hybrid censored data. Comput. Stat. Data Anal. 2006, 50, 2509–2528. [Google Scholar] [CrossRef]
  10. Balakrishnan, N.; Cramer, E. The Art of Progressive Censoring: Applications to Reliability and Quality; Springer: New York, NY, USA, 2014. [Google Scholar] [CrossRef]
  11. Cho, Y.; Sun, H.; Lee, K. Estimating the Entropy of a Weibull Distribution under Generalized Progressive Hybrid Censoring. Entropy 2015, 17, 102–122. [Google Scholar] [CrossRef] [Green Version]
  12. Cho, Y.; Sun, H.; Lee, K. Exact likelihood inference of the exponential parameter under generalized Type-II progressive hybrid censoring. Stat. Methodol. 2015, 23, 18–34. [Google Scholar] [CrossRef]
  13. Wang, L.; Li, H. Inference for exponential competing risks data under generalized progressive hybrid censoring. Commun. Stat. Simul. Comput. 2019. [Google Scholar] [CrossRef]
  14. Singh, D.P.; Lodhi, C.; Tripathi, Y.; Wang, L. Inference for two-parameter Rayleigh competing risks data under generalized progressive hybrid censoring. Qual. Reliab. Eng. Int. 2020. [Google Scholar] [CrossRef]
  15. McLachlan, G.; Krishnan, T. The EM Algorithm and Extensions. J. Classif. 1998, 15, 154–156. [Google Scholar] [CrossRef]
  16. Belaghi, R.; Noori Asl, M.; Alma, O.; Singh, S.; Vasfi, M. Estimation and Prediction for the Poisson-Exponential Distribution Based on Type-II Censored Data. Am. J. Math. Manag. Sci. 2018, 38, 1–20. [Google Scholar] [CrossRef] [Green Version]
  17. Ren, J.; Gui, W. Inference and optimal censoring scheme for progressively Type-II censored competing risks model for generalized Rayleigh distribution. Comput. Stat. 2021, 36, 479–513. [Google Scholar] [CrossRef]
  18. Hall, P. The Bootstrap and Edgeworth Expansion; Springer: New York, NY, USA, 1993. [Google Scholar]
  19. Tierney, L.; Kadane, J. Accurate Approximations for Posterior Moments and Marginal Densities. J. Am. Stat. Assoc. 1986, 81, 82–86. [Google Scholar] [CrossRef]
  20. Von der Linden, W.; Dose, V.; von Toussaint, U. Bayesian Probability Theory: Application to the Physical Sciences; Cambridge University Press: Cambridge, UK, 2014. [Google Scholar]
  21. Akgül, F.G.; Yu, K.; Senoglu, B. Classical and Bayesian Inferences in Step-Stress Partially Accelerated Life Tests for Inverse Weibull Distribution Under Type-I Censoring. Strength Mater. 2020, 52, 480–496. [Google Scholar] [CrossRef]
  22. Balakrishnan, N.; Sandhu, R. A Simple Simulational Algorithm for Generating Progressive Type-II Censored Samples. Am. Stat. 1995, 49, 229–230. [Google Scholar] [CrossRef]
  23. Lawless, J.F. Statistical Models and Methods for Lifetime Data; Wiley: New York, NY, USA, 1982. [Google Scholar]
  24. Dörre, A.; Emura, T. Analysis of Doubly Truncated Data, an Introduction; Springer: New York, NY, USA, 2019. [Google Scholar]
Figure 1. The pdfs and hrfs of the left-truncated normal distributions at zero.
Figure 1. The pdfs and hrfs of the left-truncated normal distributions at zero.
Entropy 23 00186 g001
Figure 2. Three cases of the generalized progressive hybrid censoring scheme.
Figure 2. Three cases of the generalized progressive hybrid censoring scheme.
Entropy 23 00186 g002
Figure 3. Histogram of the real data set.
Figure 3. Histogram of the real data set.
Entropy 23 00186 g003
Table 1. The number of revolutions (in millions) before failure of 23 ball bearings.
Table 1. The number of revolutions (in millions) before failure of 23 ball bearings.
0.17880.28920.33000.41520.42120.45600.48400.5184
0.51960.54120.55560.67800.68640.68640.68880.8412
0.93120.98641.05121.05841.27921.28041.7340
Table 2. Estimated values of various criteria for different distributions.
Table 2. Estimated values of various criteria for different distributions.
Distribution μ τ log ( L ) K-S
TN0.680790.164368.800690.16832
FN0.714010.146239.265710.17858
HN-0.6560111.845420.26135
Table 3. Point estimations of the real dataset.
Table 3. Point estimations of the real dataset.
CaseParameterMLEMLESELGELLLIS
(N-R)(EM)q = 0.8q = 1.1 = 0.35 = 0.45
I μ 0.804620.726340.713070.647640.559001.889661.889660.78965
τ 0.121250.145010.094440.094620.042571.116281.162780.14589
II μ 0.789570.719850.690740.814900.756481.311321.417160.75745
τ 0.099590.136720.111790.221410.138911.045351.058780.10030
III μ 0.808950.730290.750610.792980.730611.298321.399460.79024
τ 0.137800.163310.242710.306960.219091.069501.090430.17267
Table 4. Interval estimations of the real dataset.
Table 4. Interval estimations of the real dataset.
CaseParameterACIsLog-CIsHPD IntervalsBoot-p CIs
I μ (0.63866, 0.97059)(0.63854, 0.98013)(0.64351, 0.95027)(0.63901, 0.96593)
τ (0.02390, 0.21861)(0.02407, 0.21534)(0.02519, 0.20918)(0.02432, 0.21568)
II μ (0.63542, 0.94373)(0.62853, 0.93851)(0.64864, 0.93061)(0.63854, 0.94837)
τ (0.02336, 0.18983)(0.02161, 0.17734)(0.02053, 0.17918)(0.02241, 0.17782)
III μ (0.63058, 0.98731)(0.63509, 0.98602)(0.63837, 0.98069)(0.64013, 0.98165)
τ (0.03544, 0.24016)(0.03641, 0.24505)(0.04068, 0.24033)(0.03871, 0.24186)
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zeng, X.; Gui, W. Statistical Inference of Truncated Normal Distribution Based on the Generalized Progressive Hybrid Censoring. Entropy 2021, 23, 186. https://doi.org/10.3390/e23020186

AMA Style

Zeng X, Gui W. Statistical Inference of Truncated Normal Distribution Based on the Generalized Progressive Hybrid Censoring. Entropy. 2021; 23(2):186. https://doi.org/10.3390/e23020186

Chicago/Turabian Style

Zeng, Xinyi, and Wenhao Gui. 2021. "Statistical Inference of Truncated Normal Distribution Based on the Generalized Progressive Hybrid Censoring" Entropy 23, no. 2: 186. https://doi.org/10.3390/e23020186

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop