Next Article in Journal
A New Class of Alternative Bivariate Kumaraswamy-Type Models: Properties and Applications
Previous Article in Journal
Comparing Robust Linking and Regularized Estimation for Linking Two Groups in the 1PL and 2PL Models in the Presence of Sparse Uniform Differential Item Functioning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bayesian Logistic Regression Model for Sub-Areas

1
National Institute of Statistical Sciences, 1750 K Street NW Suite 1100, Washington, DC 20006, USA
2
Worcester Polytechnic Institute, Stratton Hall 103, Worcester, MA 01609, USA
*
Author to whom correspondence should be addressed.
Stats 2023, 6(1), 209-231; https://doi.org/10.3390/stats6010013
Submission received: 27 October 2022 / Revised: 4 January 2023 / Accepted: 17 January 2023 / Published: 29 January 2023

Abstract

:
Many population-based surveys have binary responses from a large number of individuals in each household within small areas. One example is the Nepal Living Standards Survey (NLSS II), in which health status binary data (good versus poor) for each individual from sampled households (sub-areas) are available in the sampled wards (small areas). To make an inference for the finite population proportion of individuals in each household, we use the sub-area logistic regression model with reliable auxiliary information. The contribution of this model is twofold. First, we extend an area-level model to a sub-area level model. Second, because there are numerous sub-areas, standard Markov chain Monte Carlo (MCMC) methods to find the joint posterior density are very time-consuming. Therefore, we provide a sampling-based method, the integrated nested normal approximation (INNA), which permits fast computation. Our main goal is to describe this hierarchical Bayesian logistic regression model and to show that the computation is much faster than the exact MCMC method and also reasonably accurate. The performance of our method is studied by using NLSS II data. Our model can borrow strength from both areas and sub-areas to obtain more efficient and precise estimates. The hierarchical structure of our model captures the variation in the binary data reasonably well.

1. Introduction

The Nepal Living Standard Survey (NLSS) II (see [1]) is a two-stage stratified sampling. A random sample of wards (areas) were selected from six strata and 12 households (sub-areas) were selected from each sampled ward. All individuals in each sampled household were interviewed. One interest is on health status, a binary variable. To make smooth estimates of the finite population proportion of the individuals with good health in each household, we focus on hierarchical Bayesian (HB) models with sub-area random effects to obtain reliable “indirect” estimates for numerous small areas or sub-areas. Most of the sample surveys are designed to provide reliable “direct” estimates of interests for large areas or domains (e.g., state level, national level). However, direct estimates are not reliable for areas or domains for which only small samples or no samples are available—see [2].
In many applications, some areas, e.g., states and wards, are sampled; in each sampled area, a sample of sub-areas, e.g., counties and households, is further selected. Ref. [3] proposed a one-fold hierarchical Bayesian logistic regression model and applied the model to NLSS II data. The main objective is to make an inference for the finite population proportion of individuals with a specific character for each area. However, the one-fold model ignores the sub-area level structure in the data. As an extension of [3], we are particularly interested in small area models that can capture the hierarchical structure of the NLSS II data in this paper. Although the one-fold basic models are very popular and in common use in producing reliable estimates, the hierarchical structure of the data and the consistency between the estimates for different levels may not hold. In particular, the sampling designs of many population-based surveys were two-stage stratified sampling as NLSS II. But if we use a one-fold unit level model to fit the data, the sub-area level effects will have been ignored. Ref. [4] studied the case that the data follow a normal model with a two-stage (three-stage) hierarchical structure, while the fitted model has a one-stage (two-stage) hierarchical structure using posterior predictive p-values. Ref. [5] discussed the ability to detect a three-stage model when a two-stage model is actually fitted.
Two-fold models are an important extension of basic small area models. Many authors have considered the problems and proposed these kinds of models. Much of the literature focuses on continuous data. Ref. [6] proposed a sub-area level model which provides model-based estimates that account for the hierarchical structure of data. Two-fold sub-area level models were studied by [2,7,8,9], and many others. This type of model is an area-level model which extends the Fay-Harriot model (see [10]) to the sub-area level. Two-fold nested error regression models were considered by [11,12]. On the other hand, some literature focus on the categorical data. Ref. [13] described a HB model to make an inference about the finite population proportion under two-stage clustering sampling. Ref. [14] extended the Beta-Binomial model to the two-fold model and used Gibbs sampling to obtain the posterior estimates. Ref. [15] showed that the two-fold Beta-Binomial model is preferable over the one-fold one if the data have a hierarchical structure. Ref. [16] extended [15] to accommodate heterogeneous correlations. They used a HB model to make a posterior inference about the finite population proportion of each area, accounting for intracluster correlations. Ref. [17] discussed the sub-area Beta-Binomial model and applied the model to estimate the finite population proportion of healthy individuals in each household covered by the NLSS II, assuming no covariate was available.
Bayesian logistic regression models with random effects are suitable for handling binary data with covariates. Ref. [18] discussed discrimination between the logit and the complementary log-log link functions by using the logistic regression model. Roberts, Ref. [19], discussed logistic regression for the sample survey data (not small area estimation). Ref. [20] showed how to accelerate the Gibbs sampler for a model with latent variables introduced earlier by [21] for Bayesian probit analysis. Ref. [22], discussed the logistic regression model by using the empirical Bayesian approach. Ref. [23] showed how to analyze binary data with covariates to maintain conjugacy for both the logistic and Poisson regression models. The analysis of binary data with covariates under nonignorable nonresponse was discussed by [24]. Ref. [3] proposed a hierarchical Bayesian logistic regression model for binary data in a small area estimation. This model is a unit level model without a sub-area effect. Our two-fold sub-area model is an extension of this logistic regression model. We add the sub-area level random effect into the model which can capture the hierarchical structure of the sampled data. At the same time, we add more hyper-parameters into the model, which make the inference more complicated. However, we propose an approximation method called the integrated nested normal approximation (INNA), which solves the difficulties.
The other side of our application is that there are numerous small areas (households and individuals) and MCMC methods, which involve complicated integrals, and cannot handle them efficiently. “Big data” are defined as data that are too big to comfortably process on a single machine [25]. The researchers considered consensus Monte Carlo methods that split the data across several machines. They proposed algorithms that perform distributed approximate Bayesian analyses in order to minimize the communication between computers. The parallel MCMC methods for non-Gaussian posterior distributions were discussed by [26]. Fortunately, in survey sampling, the design generally uses a stratification which is not artificial, and, in this case, consensus Monte Carlo may not be needed; it will be a good idea for a large stratum.
The integration involved in Bayesian inference is usually intractable, which is true for our logistic regression model. The approximation techniques are desired. The procedure we used to approximate the posterior density of the parameters of the logistic regression sub-area model, INNA, is similar to the integrated nested Laplace approximation (INLA) originally proposed by [27], but they are actually different. INLA is a quite popular algorithm and an alternative to MCMC for big data analysis if the joint posterior density is very complicated. It requires posterior modes, and, for numerous small areas, the computation of modes becomes time-consuming and challenging for the logistic regression model or any generalized linear mixed models. Yet, INLA has found many useful applications, such as in Poisson regression by [28], and in spatial point pattern data by [29]. We note that INLA can be problematic, especially for logistic and Poisson hierarchical regression models, even if the modes can be computed. Ref. [30], attempted to improve INLA using a copula-based correction, which adds complexity to INLA. Our approximation method, INNA, which does not require finding posterior modes, uses a sampling-based procedure accommodated by the multiplication rule of probability. Instead of finding the posterior modes, INNA finds the approximate modes in closed form, facilitated by the empirical logistic transform ([31]) and the second-order Taylor series approximation.
On the other hand, two-fold models can capture the heterogeneity between samples within not only areas but also sub-areas. Many model-based estimation techniques for the sampling variances have been considered in the literature, but most of them for the area-level model: see [32,33,34].
In Section 2, a full description of a sub-area HB logistic regression model is given. In Section 3, we describe the integrated nested normal approximation (INNA) computation method and some theoretical results are provided. The exact MCMC method is presented in Appendix A. The exact method refers to MCMC methods without further approximation. In Section 4, we apply the model to the NLSS II data to provide smoothed estimates of the household proportions of members in good health for both sampled and nonsampled households. Some comparisons between INNA and the exact method are presented. Finally, in Section 5, we make concluding remarks and discuss the future work.

2. Sub-Area Logistic Regression Model

In this section, we discuss the sub-area HB logistic regression model at the unit level. In the NLSS II data, we have binary data (good health versus poor health) for each individual within a household, and these households are within wards. The observations are available at the unit level and so is the reliable auxiliary information. However, the model and method we proposed for small areas and sub-areas is not only for this application on NLSS II data. It can be also applied to other population-based surveys with binary responses which contain small areas or/and sub-areas.
Suppose that there are L small areas (wards) in the finite population and that, within the i th area, there are N i sub-areas (households). Within the j th sub-area, there are M i j individuals. We assume that ( < L ) areas are sampled and a simple random sample of n i ( < N i ) households is taken from the i th area. All individuals in the sampled households are sampled. Here, we assume the survey weights are the same within all households in each area. Actually, the design is almost self-weighting.
Let y i j k , k = 1 , , m i j , j = 1 , , n i , i = 1 , , denote the binary responses. Let y ˜ = ( y i j k , k = 1 , , m i j , j = 1 , , n i , i = 1 , , ) . Let y i j = k = 1 m i j y i j k be the number with response 1 and m i j be the total number of people who responded. Let x ˜ i j k = ( 1 , x i j k 1 , , x i j k p ) be the ( p + 1 ) vector with p covariates for individuals and an intercept.
We use P to represent the population proportion and p as the sample proportion. Let p i j be the corresponding sample probability of y i j , j = 1 , , n i , i = 1 , , .
The primary interests are the finite population proportions of the households, which are P i j = 1 M i j k = 1 M i j y i j k , j = 1 , , N i , i = 1 , . . , and the finite population proportions of the areas, which are P i = 1 N i j = 1 N i k = 1 M i j y i j k , i = 1 , , .
In the content of the logistic regression model, the two-fold hierarchical Bayesian logistic regression model for the sub-area means, μ i j , is
y i j k | β ˜ , ν i , μ i j i n d Bernoulli e x ˜ i j k β ˜ + ν i + μ i j 1 + e x ˜ i j k β ˜ + ν i + μ i j , k = 1 , , m i j ,
μ i j | σ 2 i i d Normal ( 0 , σ 2 ) , j = 1 , , n i ,
ν i | δ 2 i i d Normal ( 0 , δ 2 ) , i = 1 , , ,
π ( β ˜ , δ 2 , σ 2 ) 1 ( 1 + δ 2 ) 2 1 ( 1 + σ 2 ) 2 , δ 2 > 0 , σ 2 > 0 .
Here, μ i j , i = 1 , , ; j = 1 , , n i are the sub-area level random effects, which are not in the area-level model in [3]. ν i , i = 1 , , are the area random effects and β ˜ = ( β 0 , β 1 , , β p ) are the regression coefficients, with σ 2 , δ 2 as the variance of the random effects, respectively.
In order to apply our approximation method and make an inference for posterior distribution, we use an equivalent model.
First, we separate β ˜ into β 0 and β ˜ ( 0 ) , where β ˜ ( 0 ) = ( β 1 , β 2 , , β p ) T . We set β 0 as the mean of ν ˜ , and then we can omit the intercept term from the covariate x ˜ i j k . Second, we introduce a new parameter, w i j = ν i + μ i j , in order to set ν i and μ i j independently and make it easy to make an inference for both of them. We have
y i j k | w i j , β ˜ ( 0 ) i n d Bernoulli e x ˜ i j k β ˜ ( 0 ) + w i j 1 + e x ˜ i j k β ˜ ( 0 ) + w i j , k = 1 , , m i j ,
w i j | ν i , σ 2 i n d Normal ( ν i , σ 2 ) , j = 1 , , n i ,
ν i | β 0 , δ 2 i i d Normal ( β 0 , δ 2 ) , i = 1 , , ,
π ( β ˜ , δ 2 , σ 2 ) 1 ( 1 + δ 2 ) 2 1 ( 1 + σ 2 ) 2 , δ 2 > 0 , σ 2 > 0 .
The joint posterior density for the parameters is
π ( v ˜ , w ˜ , β ˜ , σ 2 , δ 2 | y ˜ ) i = 1 j = 1 n i k = 1 m i j e ( x ˜ i j k β ˜ ( 0 ) + w i j ) y i j k 1 + e x ˜ i j k β ˜ ( 0 ) + w i j × 1 2 π σ 2 n exp i = 1 l j = 1 n i ( w i j ν i ) 2 2 σ 2 × 1 2 π δ 2 l exp i = 1 l ( ν i β 0 ) 2 2 δ 2 1 ( 1 + σ 2 ) 2 1 ( 1 + δ 2 ) 2 .
The posterior density is a non-standard multivariate density, and there are difficulties in fitting it using MCMC methods, more so when n i , m i j are large. This motivates our approximate methods.

3. Integrated Nested Normal Approximation Method

In this section, we discuss the INNA method for the sub-area HB logistic regression model. It is an extension of the INNA method in [3]. INNA method is not required to find the posterior modes. Due to the large amount of sub-areas, it would be time-consuming to find all posterior modes, which is why we did not choose the popular INLA method. In detail, we discuss the approximation of the joint posterior density (3).
Notice that the joint posterior density (3) is very complicated and it is the expit part, i = 1 j = 1 n i k = 1 m i j e ( x ˜ i j k β ˜ ( 0 ) + w i j ) y i j k 1 + e x ˜ i j k β ˜ ( 0 ) + w i j , that causes the difficulties. In the following, we discuss how to approximate this term to normal density functions by using Laplace approximation, the second-order multivariate Taylor-series approximation and the empirical logistic transform (ELT). This is the key contribution in the paper. Then we use the multiplication rule to approximate the joint posterior density,
π a ( w ˜ , ν ˜ , β ˜ , σ 2 , δ 2 y ˜ ) π a ( w ˜ ν ˜ , β ˜ ( 0 ) , σ 2 , y ˜ ) π a ( ν ˜ β 0 , δ 2 , y ˜ ) π a ( β ˜ ( 0 ) y ˜ ) π a ( β ˜ , σ 2 , δ 2 y ˜ ) ,
where the first three densities on the right-hand side are all multivariate normal densities. Therefore, we can draw samples and make inference through the approximate joint posterior density.
Let f ( τ ˜ ) = e h ( τ ˜ ) denote the density of a vector of parameters τ ˜ . Let g ˜ denote the gradient vector and H the Hessian matrix at some point τ ˜ * .
Lemma 1. 
Let h ( τ ˜ ) be a logconcave density function with the parameter τ ˜ . Then, τ ˜ approximately has a multivariate normal distribution,
τ ˜ N o r m a l ( τ ˜ * H 1 g ˜ , H 1 ) .
Proof. 
Simply applying the second-order multivariate Taylor series of h ( τ ˜ ) at τ ˜ * gives
f ( τ ˜ ) f ( τ ˜ * ) + ( τ ˜ τ ˜ * ) g ˜ + 1 2 ( τ ˜ τ ˜ * ) H ( τ ˜ τ ˜ * ) .
Due to the logconcavity of h ( τ ˜ ) , its Hessian Matrix H is posit-definite, which can be the covariance matrix. Notice that we are not required to use the mode of h ( τ ˜ ) . We do not need to find the solution of the gradient vector g ˜ = 0 . Therefore, τ ˜ * does not have to be the solution but some other point. It is worth noticing that the term, H 1 g ˜ , is a correction to τ ˜ * . □
To illustrate the approximation steps, we start with a simpler model with flat priors for β ˜ ( 0 ) and the w ˜ , according to model (2). That is,
y i j k | w i j , β ˜ ( 0 ) i n d Bernoulli e x ˜ β ˜ ( 0 ) + w i j 1 + e x ˜ i j k β ˜ ( 0 ) + w i j , k = 1 , , m i j , j = 1 , , n i , i = 1 , , ,
p ( w ˜ , β ˜ ( 0 ) ) = 1 .
The joint posterior density is
π ( w ˜ , β ˜ ( 0 ) | y ˜ ) i = 1 j = 1 n i k = 1 m i j e ( x ˜ i j k β ˜ ( 0 ) + w i j ) y i j k 1 + e x ˜ i j k β ˜ ( 0 ) + w i j .
The logarithm of the joint posterior density (or log likelihood) is
Δ = h ( τ ˜ ) = i = 1 j = 1 n i k = 1 m i j ( x ˜ i j k β ˜ ( 0 ) + w i j ) y i j k log ( 1 + e x ˜ i j k β ˜ ( 0 ) + w i j ) .
Let τ ˜ = ( μ ˜ , β ˜ ( 0 ) ) . In our method, we find a convenient point to expand the log-likelihood in a second-order multivariate Taylor-series expansion.
To begin with, let y ¯ i j = 1 m i j k = 1 m i j y i j k . We use the empirical logistic transform z i j to get an estimate of w i j , where
w ^ i j * = z i j = log y ¯ i j + 1 2 m i j 1 y ¯ i j + 1 2 m i j , i = 1 , , ; j = 1 , , n i .
First, we discuss how to find the quasi mode of β ˜ ( 0 ) . We plug w ^ i j * into the log likelihood function Δ and consider it as a function of β ˜ ( 0 ) only as q ( β ˜ ( 0 ) ) , and we get
q ( β ˜ ( 0 ) ) = i = 1 j = 1 n i k = 1 m i j ( x ˜ i j k β ˜ ( 0 ) + w ^ i j * ) y i j k log ( 1 + e x ˜ i j k β ˜ ( 0 ) + w ^ i j * ) .
The first derivative of q ( β ˜ ( 0 ) ) is
q ( β ˜ ( 0 ) ) = i = 1 j = 1 n i k = 1 m i j x ˜ i j k y i j k x ˜ i j k e ( x ˜ i j k β ˜ ( 0 ) + w ^ i j * ) 1 + e x ˜ i j k β ˜ ( 0 ) + w ^ i j * = i = 1 j = 1 n i k = 1 m i j x ˜ i j k y i j k x ˜ i j k 1 + e ( x ˜ i j k β ˜ ( 0 ) + w ^ i j * ) 1 .
Usually we should set q ( β ˜ ( 0 ) ) equal to zero and find the modes as the maximum likelihood estimator (MLE) of β ˜ ( 0 ) . But here, it is not easy to solve the equation due to the complexity of q ( β ˜ ( 0 ) ) . We use the first-order Taylor series to approximate it and then simplify q ( β ˜ ( 0 ) ) so that we can get quasi-modes of β ˜ ( 0 ) .
The first-order Taylor expansion of ( 1 + e x ˜ i j k β ˜ ( 0 ) + w ^ i j * ) 1 equals ( 1 e ( x ˜ i j k β ˜ ( 0 ) + w ^ i j * ) ) . Notice that by Taylor series, e ( x ˜ i j k β ˜ ( 0 ) + w ^ i j * ) 1 ( x ˜ i j k β ˜ ( 0 ) + w ^ i j * ) . Then we can get
q ( β ˜ ( 0 ) ) i = 1 j = 1 n i k = 1 m i j x ˜ i j k y i j k x ˜ i j k ( 1 e x ˜ i j k β ˜ ( 0 ) + w ^ i j * ) i = 1 j = 1 n i k = 1 m i j x ˜ i j k y i j k x ˜ i j k ( 1 1 + ( x ˜ i j k β ˜ ( 0 ) + w ^ i j * ) ) = i = 1 j = 1 n i k = 1 m i j x ˜ i j k ( y i j k w ^ i j * ) x ˜ i j k x ˜ i j k β ˜ ( 0 ) .
We can get the quasi-modes of β ˜ ( 0 ) by solving the equation q ( β ˜ ( 0 ) ) = 0 . That is,
β ˜ ( 0 ) * = [ i = 1 j = 1 n i k = 1 m i j x ˜ i j k x ˜ i j k ] 1 [ i = 1 j = 1 n i k = 1 m i j x ˜ i j k ( y i j k w ^ i j * ) ] .
Second, we obtain quasi-modes for the w i j , a refinement of the z i . Plug β ˜ ( 0 ) * into the likelihood function Δ and consider it as function w i j only:
g ( w i j ) = k = 1 m i j ( x ˜ i j k β ˜ ( 0 ) * + w i j ) y i j k log ( 1 + e x ˜ i j k β ˜ ( 0 ) * + w i j ) .
Similarly, after applying Taylor expansion, we get the approximate first derivative of g ( w i j )
g ( w i j ) = k = 1 m i j y i j k 1 + e ( x ˜ i j k β ˜ ( 0 ) * + w i j ) 1 k = 1 m i j y i j k 1 e w i j e x ˜ i j k β ˜ ( 0 ) * .
We can obtain the approximate posterior mode of w i j by solving the equation g ( w i j ) = 0 .
w i j * = log k = 1 m i j e x ˜ i j k β ˜ ( 0 ) * m i j ( 1 y ¯ i j ) .
Notice that the term 1 y ¯ i j in denominator may cause trouble if y ¯ i j = 1 for some is and js. Here, we borrow the idea from ELT and make a small adjustment in order to avoid a zero denominator. That is,
w i j * log k = 1 m i j e x ˜ i j k β ˜ ( 0 ) * m i j ( 1 y ¯ i j + 1 2 m i j ) i = 1 , , , j = 1 , , n i .
Let τ ˜ * = ( μ ˜ * , β ˜ ( 0 ) * ) . Next, we evaluate g ˜ and H at the quasi-modes τ ˜ = τ ˜ * can also be obtained as
g ˜ = Δ w 11 Δ w n Δ β ˜ ( 0 ) w ˜ = w ˜ * , β ˜ ( 0 ) = β ˜ ( 0 ) * T ,
H = 2 Δ w 11 2 2 Δ w 11 w n 2 Δ w 11 β ˜ ( 0 ) 0 2 Δ w n 2 2 Δ w n β ˜ ( 0 ) 2 Δ w 11 β ˜ ( 0 ) 2 Δ w n β ˜ ( 0 ) 2 Δ β ˜ ( 0 ) 2 w ˜ = w ˜ * , β ˜ ( 0 ) = β ˜ ( 0 ) * .
The partial derivatives can be expressed in terms of response y i j k and covariates x ˜ i j k as
Δ β ˜ ( 0 ) = i = 1 j = 1 n i k = 1 m i j x ˜ i j k y i j k x ˜ i j k e x ˜ i j k β ˜ ( 0 ) * + w i j * 1 + e x ˜ i j k β ˜ ( 0 ) * + w i j * ,
Δ w i j = k = 1 m i j ( y i j k e x ˜ i j k β ˜ ( 0 ) * + w i j * 1 + e x ˜ i j k β ˜ ( 0 ) * + w i j * ) ,
2 Δ β ˜ ( 0 ) 2 = i = 1 j = 1 n i k = 1 m i j x ˜ i j k x ˜ i j k e x ˜ i j k β ˜ ( 0 ) * + w i j * ( 1 + e x ˜ i j β ˜ ( 0 ) * + w i j * ) 2 ,
2 Δ w i j 2 = k = 1 m i j e x ˜ i j k β ˜ ( 0 ) * + w i j * ( 1 + e x ˜ i j k β ˜ ( 0 ) * + w i j * ) 2 ,
2 Δ μ i β ˜ ( 0 ) = k = 1 m i j x ˜ i j k e x ˜ i j k β ˜ ( 0 ) * + w i j * ( 1 + e x ˜ i j k β ˜ ( 0 ) * + w i j * ) 2 ,
where i = 1 , , , j = 1 , , n i .
For the convenience of computation, denote g ˜ = g ˜ 1 g ˜ 2 and H = D C C B , where
g ˜ 1 = Δ w 11 Δ w n T , g ˜ 2 = Δ β ˜ ( 0 ) ,
B = 2 Δ β ˜ ( 0 ) 2 , C = 2 Δ w 11 β ˜ ( 0 ) 2 Δ w n β ˜ ( 0 ) , D = 2 Δ w 11 2 0 : : 0 2 Δ w n 2 .
Let H 1 = D C C B 1 = E F F G , where
E = D 1 + D 1 C ( B C D 1 C ) 1 C D 1 , F = ( B C D 1 C ) 1 C D 1 , G = ( B C D 1 C ) 1 .
Lemma 2. 
Assuming that the design matrix is full-rank and 0 < k = 1 m i j y i j k < m i j , j = 1 , , n i ; i = 1 , , , the posterior density, τ ˜ | y ˜ in (5), is logconcave.
Proof. 
If 0 < k = 1 m i j y i j k < m i j , i = 1 , , , j = 1 , , n i , there are solutions to the gradient vector set to zero.
Let p i j k = e x ˜ i j k β ˜ ( 0 ) + w i j 1 + e x ˜ i j k β ˜ ( 0 ) + w i j , k = 1 , , m i j , j = 1 , , n i , i = 1 , , . Then, A, B and C of the negative Hessian matrix can be written as,
B = 2 Δ β ˜ ( 0 ) 2 = i = 1 j = 1 n i k = 1 m i j p i j k ( 1 p i j k ) x ˜ i j k x ˜ i j k ,
D = d i a g o n a l ( d i j ) , d i j = 2 Δ w i j 2 = k = 1 m i j p i j k ( 1 p i j k ) ,
C = ( c ˜ i j ) , c ˜ i j = 2 Δ w i j β ˜ ( 0 ) = k = 1 m i j p i j k ( 1 p i j k ) x ˜ i j k ,
where j = 1 , , n i , i = 1 , , .
It is obvious that D is positive-definite. Thus, to show that H is positive-definite, we need to show that its Schur complement of D, S = B C D 1 C , is positive-definite (e.g., see [35]). Let ω i j k = p i j k ( 1 p i j k ) / k = 1 m i j p i j k ( 1 p i j k ) , k = 1 , , m i j , j = 1 , , n i , i = 1 , , . The Schur complement is
S = i = 1 j = 1 n i k = 1 m i j p i j k ( 1 p i j k ) k = 1 m i j ω i j k x ˜ i j k x ˜ i j k i = 1 j = 1 n i k = 1 m i j p i j k ( 1 p i j k ) k = 1 m i j ω i j k x ˜ i j k k = 1 m i j ω i j k x ˜ i j k .
It is now easy to show that
S = i = 1 j = 1 n i k = 1 m i j ω i j k ( x ˜ i j k k = 1 m i j ω i j k x ˜ i j k ) ( x ˜ i j k k = 1 m i j ω i j k x ˜ i j k ) .
Therefore, H is positive-definite, and τ ˜ | y ˜ is logconcave. □
Finally, according to the Lemmas 1 and 2, we can establish the approximation Theorem.
Theorem 1. 
Assuming that the design matrix is full-rank and 0 < k = 1 m i j y i j k < m i j , j = 1 , , n i , i = 1 , , , the posterior density, τ ˜ | y ˜ in (5) is approximately a multivariate normal density, and the conditional posterior density of w ˜ | β ˜ ( 0 ) , y ˜ and β ˜ ( 0 ) | y ˜ can also be approximated by multivariate normal distributions.
Proof. 
The proof is given in Appendix B. □
Therefore, we can approximate that logit expit term i = 1 j = 1 n i k = 1 m i j e ( x ˜ i j k β ˜ ( 0 ) + w i j ) y i j k 1 + e x ˜ i j k β ˜ ( 0 ) + w i j into two multivariate densities by Theorem 1. And then we can get our approximate two-fold Bayesian logistic regression model.
Recall the posterior density of our two-fold logistic model is
π ( w ˜ , ν ˜ , β ˜ , σ 2 , δ 2 y ˜ ) π ( y ˜ | w ˜ , β ˜ ( 0 ) ) π ( w ˜ ν ˜ , σ 2 ) π ( ν ˜ β 0 , δ 2 ) π ( β ˜ ( 0 ) , β 0 , σ 2 , δ 2 )
The likelihood function π ( y ˜ | w ˜ , β ˜ ( 0 ) ) can be approximated by the multivariate normal distribution by Theorem 1. Combining the prior values of w ˜ and ν ˜ given by our Bayesian Logistic model and the results in Theroem 1, we can obtain our INNA model
w ˜ | β ˜ ( 0 ) , y ˜ Normal { μ ˜ w D 1 C ( β ˜ ( 0 ) μ ˜ β ) , D 1 }
β ˜ ( 0 ) | y ˜ Normal { μ ˜ β , G }
w ˜ | ν ˜ , σ 2 i n d Normal ( μ ˜ ν , σ 2 I ) ,
ν ˜ | β 0 , δ 2 i i d Normal ( β 0 j ˜ , δ 2 I ) ,
π ( β ˜ ( 0 ) , β 0 , δ 2 , σ 2 ) 1 ( 1 + δ 2 ) 2 1 ( 1 + σ 2 ) 2 , δ 2 > 0 , σ 2 > 0 ,
where μ ˜ ν = ν 1 , , ν 1 n 1 ν , , ν n and j ˜ is a vector of ones.
Using Bayes’ Theorem and the multiplication rule, the posterior density π ( w ˜ , ν ˜ , β ˜ , σ 2 , δ 2 y ˜ ) can be approximated as
π a ( w ˜ , ν ˜ , β ˜ , σ 2 , δ 2 y ˜ ) π a ( w ˜ ν ˜ , β ˜ ( 0 ) , σ 2 , y ˜ ) π a ( ν ˜ β 0 , δ 2 , y ˜ ) π a ( β ˜ ( 0 ) y ˜ ) π a ( β ˜ , σ 2 , δ 2 y ˜ ) = e 1 2 w ˜ μ ˜ w D 1 C ( β ˜ ( 0 ) μ ˜ β ) D w ˜ μ ˜ w D 1 C ( β ˜ ( 0 ) μ ˜ β ) × e 1 2 w ˜ μ ˜ ν ( σ 2 I ) 1 w ˜ μ ˜ ν + ν ˜ β 0 j ˜ ( δ 2 I ) 1 ν ˜ β 0 j ˜ + β ˜ ( 0 ) μ ˜ β G 1 β ˜ ( 0 ) μ ˜ β × | D | 1 / 2 | δ 2 I | 1 / 2 | σ 2 I | 1 / 2 | G | 1 / 2 1 ( 1 + σ 2 ) 2 1 ( 1 + δ 2 ) 2
Therefore, we can get the following key result.
Theorem 2. 
Using the multiplication rule, the joint posterior density, π ( w ˜ , ν ˜ , β ˜ , σ 2 , δ 2 y ˜ ) in (6), can be approximated by
π a ( w ˜ , ν ˜ , β ˜ , σ 2 , δ 2 y ˜ ) π a ( w ˜ ν ˜ , β ˜ ( 0 ) , σ 2 , y ˜ ) π a ( ν ˜ β 0 , δ 2 , y ˜ ) π a ( β ˜ ( 0 ) y ˜ ) π a ( β ˜ , σ 2 , δ 2 y ˜ ) ,
where the first three densities on the right-hand side are all multivariate normal densities.
Proof. 
The proof is given in Appendix C. □
The INNA is actually a random sampler. First, we draw samples for σ 2 , δ 2 from π ( σ 2 , δ 2 | y ˜ ) . The posterior distribution of σ 2 , δ 2 | y ˜ does not have standardized form. Here, we use the grid method and numerical integration to sample σ 2 and δ 2 . Since 0 < σ 2 < and 0 < δ 2 < , we make a transformation to ϕ 1 = 1 1 + σ 2 and ϕ 2 = 1 1 + δ 2 so that we get 0 < ϕ 1 < 1 and 0 < ϕ 2 < 1 . Then, the posterior density of ϕ 1 , ϕ 2 | y ˜ is
π a ( ϕ 1 , ϕ 2 | y ˜ ) δ 0 2 γ ˜ γ ˜ Δ ( 0 ) 1 2 × i = 1 l 1 j = 1 n i σ i j 2 + 1 δ 2 1 2 1 | δ 2 D + δ 2 σ 2 I | 1 / 2 ϕ 1 = 1 1 + σ 2 , ϕ 2 = 1 1 + δ 2 × exp 1 2 ( μ ˜ w + D 1 C μ ˜ β ) ( D 1 + σ 2 I + δ 2 I ) 1 ( μ ˜ w + D 1 C μ ˜ β ) + μ ˜ β G 1 μ ˜ β × exp 1 2 β 0 ω 0 β ˜ ( 0 ) ω ˜ ( 0 ) δ 0 2 γ ˜ γ ˜ Δ ( 0 ) β 0 ω 0 β ˜ ( 0 ) ω ˜ ( 0 ) ϕ 1 = 1 1 + σ 2 , ϕ 2 = 1 1 + δ 2 .
We need to draw ϕ 1 , ϕ 2 together. The joint density can be rewritten as
π ( ϕ 1 , ϕ 2 | y ˜ ) = π ( ϕ 2 | ϕ 1 ) π ( ϕ 1 | y ˜ ) = π ( ϕ 2 | ϕ 1 ) 0 1 π ( ϕ 1 , ϕ 2 | y ˜ ) d ϕ 2 .
We plug each grid of ϕ 1 ( 0 , 1 ) into 0 1 π ( ϕ 1 , ϕ 2 | y ˜ ) d ϕ 2 and then use numerical integration to get the density of ( ϕ 1 | y ˜ ) . After we plug all the 100 grids, we can get 100 value of π ( ϕ 1 | y ˜ ) and then draw ϕ 1 from them, i.e. ϕ 1 ( h ) . Next, we plug ϕ 1 ( h ) into π ( ϕ 2 | ϕ 1 ) and use grid method to draw ϕ 2 ( h ) . We repeat those steps 10,000 times to get the sample of ( ϕ 1 ( h ) , ϕ 2 ( h ) ) , h = 1 , , 10,000 . Once we get samples for ϕ 1 , ϕ 2 , we transform them back to σ 2 and δ 2 respectively. Second, given σ 2 , δ 2 , we can simply draw samples of β ˜ from the approximate multivariate normal distribution π a ( β ˜ σ 2 , δ 2 , y ˜ ) . Third, we can draw samples of ν i independently given β ˜ , δ 2 and data from the approximate normal distribution π a ( ν ˜ | β 0 , δ 2 , y ˜ ) . Finally, samples of w i j independently given ν ˜ , β ˜ , σ 2 can be obtained from the approximate normal distribution π a ( w ˜ | ν ˜ , β ˜ ( 0 ) , σ 2 , y ˜ ) . Notice that the last three steps are very simple, just drawing samples from normal densities. In addition, w i j and ν i are all independent so that we can draw them simultaneously. Therefore, those latter steps permit fast computing.
In order to check if INNA method can provide resonal results, we apply the MCMC logistic regression exact method to the sub-area model. The idea of exact method is to get full conditional posterior distributions for all of the parameters in the model, and then get a large number of independent samples of each parameter with its full conditional posterior density. Details are given in Appendix A.
There are two differences between these two methods. First, both methods are sampling-based. The approximate method implements random samples and the exact method uses numerical integration method and Markov chains. Second, π a ( β ˜ , σ 2 , δ 2 y ˜ ) is used for the INNA method. In the exact method, a Metroplis step is used for the π ( β ˜ , σ 2 , δ 2 y ˜ ) . This is very time-consuming in the exact method. On the other hand, the exact method actually uses the INNA method. We use M-H sampler draw samples for ν ˜ and w ˜ , respectively. Proposal functions are π a ( ν ˜ β ˜ ( 0 ) , β 0 , σ 2 , y ˜ ) and π a ( w ˜ ν ˜ , β ˜ ( 0 ) , σ 2 , y ˜ ) , respectively, from the INNA method.

4. Numerical Example

4.1. Nepal Living Standards Survey II

The performance of our method is studied using the Nepal Living Standard Survey (NLSS II), conducted in the years 2003–2004. The main objective of the NLSS II is to track changes in and the progress of national living standards and social indicators of the Nepalese population. It is an integrated survey which covers samples from the whole country and runs throughout the year.
The NLSS II gathers information on a variety of aspects. It has collected data on demographics, housing, education, health, fertility, employment, income, agricultural activity, consumption, and various other areas. The sampling design of NLSS II is two-stage stratified sampling. Nepal is stratified into Primary Sampling Units (PSUs) and, within each PSU, there are a number of households (sub-area) selected. All household members in the sample were interviewed.
In detail, the NLSS II has records for 20,263 individuals from 3912 households (sub-areas) from 326 PSUs (areas) from a population of 60,262 households and about two million Nepalese. A sample of PSUs was selected from the strata using probability proportional to size (PPS) sampling and 12 households were systematically selected from each PSU. The survey is self-weighted and some adjustments were made after conducting the survey for non-responses or missing data. For simplicity, in this paper, we assume all samples have the same weight. Table 1 shows the distribution of all samples by stratum.
We chose four relevant covariates which can influence health status from the same NLSS II survey for our two-fold logistic regression model. They are age, nativity, sex and religion. We created binary variables for nativity (Indigenous = 1, Non-indigenous = 0) and religion ((Hindu = 1, Non-Hindu = 0), sex (Male = 1, Female = 0). Table 2 shows the details of these four covariates. In the model fitting, we standardized the age covariate. Older age and a child’s age are more vulnerable times than younger age. Indigenous people can have different health statuses from migrated people.
According to the 2001 census data, only about 0.091% of households and only 0.904% of PSU were sampled. The NLSS II was designed to provide reliable estimates only at the stratum level, or even larger areas than the stratum. It cannot give estimates in small areas (PSU or household level) since the sample sizes are too small. Therefore, we need to use statistical models to fit the available data and find reliable estimates in small areas. In our study, we chose the binary variable, health status, from the health section of the questionnaire.

4.2. Numerical Comparison

We used data from NLSS II to illustrate our sub-area logistic regression model. We predicted the household proportions of members in good health for 18,924 households (sampled and non-sampled). Bayesian bootstrap by [36] was applied to get non-sampled auxiliary information. This analysis was based on 1224 sample households from 102 wards (PSUs) in strata 6. Our primary purposes were to show that our model can provide good estimates and to compare the approximate method with the exact method when there are random effects at the household level.
We used Rcpp [37] and RcppArmadillo [38] packages in R [39] to fit the model based on both the approximation, INNA, method and the exact method to this NLSS II dataset. For the INNA method, we began with 10,000 iterations and a burn-in of 1000 and we kept only every ninth sample. Finally, 1000 samples were obtained for constructing the posterior distributions of all the parameters. The exact method was very time-consuming, taking about 30 h to finish. However, the INNA approximation method can get samples in 8 min. When we have a large number of areas or sub-areas, the approximation method will make enormous savings.
Convergence diagnostics were conducted. The convergences of the hyperparameters ( β ˜ , σ 2 , δ 2 ) were monitored by the Geweke test of stationarity [40] and the effective sample sizes. The p-values and effective sample sizes are shown in Table 3, resulting in good convergence for both methods. Table 3 also shows the posterior means (PMs) and associated posterior standard deviations (PMs) of the hyperparameters. The PMs are very close between these two methods. The PSDs are slightly larger for the exact method than for the INNA method, but they are reasonably close.
In Figure 1, Figure 2 and Figure 3, we compare the PMs, PSDs and posterior coefficient of variations (PCVs) in the household level as our primary purpose. We can see that the PMs are very close, nearly lying on the 45-degree line through the origin. The PSDs are slightly spread out and thicker, but all points still lie on the 45-degree line and so do the PCVs. Overall, these approximations are acceptable in the data analysis. Figure 4, Figure 5 and Figure 6 we compare respectively to the PMs, PSDs and PCVs at the ward level. The plots of the PMs are still very good. Notice that other two plots of PSDs and PCVs are more spread out than those in the household level. Again, though, the approximate method and the exact method are reasonably close.
We also compare the approximate method with the exact method using the five number summaries (the minimum values, the first quartiles, the median, the third quartiles, and the maximum values) with respect to the PMs, PSDs and PCVs of the finite population proportions at the household level and ward level in Table 4 and Table 5. The PMs from both methods at the household level have larger variations than those at the ward level. The PCVs at the ward level are generally much smaller than at the household level. The summaries of the PMs, PSDs and PCVs within households and wards between the approximate and exact methods are very close.
We conclude that the approximation method at the household level is reasonable. The approximation is desirable because one can perform the computations in real time.

5. Conclusions and Future Work

The sub-area HB logistic regression model can be applied to analyze the binary response variable. This model is an extension of the HB logistic regression area-level model, which ignores the actual hierarchical structure of the data. We propose an approximation method, INNA, to fit the model. For large datasets, it is very unrealistic to use the MCMC method to fit the model. We propose the approximation method, INNA, which saves time significantly because there is no need to compute numerous modes. In the numerical example, we can show that INNA can provide reliable estimates as well. An illustrative example of the NLSS II is presented in order to compare the approximation method and the exact method. It shows that, when there are a large number of areas and sub-areas, the approximation method can be efficient and it can also provide reasonable estimates.
INNA is a method for approximate Bayesian inference based on Laplace’s method, the second-order multivariate Taylor-series approximation and the empirical logistic transform (ELT). It can be applied to all HB logistic regression models, for which it can be a fast and accurate alternative to the Markov chain Monte Carlo methods. The comparison and model results illustrate the performance of the INNA methods based on the sub-area model.
There will be many future works on the two-fold small areas model. First, in this paper, we assume equal survey weights since the NLSS II is a self-weighted sampling. However, after the data are collected, the sampling weights are usually adjusted for various characteristics or based on nonresponse as well. Incorporating those survey weights into the model is also very important. Generally, we need to consider these weights in the model. The NLSS II is a national population-based survey. We should rescale the sample weights to sum to an equivalent sample size. That is, we consider the adjusted weight as w i j k * = n ^ ( w i j k i = 1 j = 1 n i k = 1 m i j w i j k ) , where n ^ = ( i = 1 j = 1 n i k = 1 m i j w i j k ) 2 i = 1 j = 1 n i k = 1 m i j w i j k 2 as an equivalent sample. Introducing the sampling weights, we can obtain an updated normalized likelihood function. Based on the updated likelihood function and the same prior in the two-fold model, we can have a full Bayesian analysis on the updated model and then project the finite population proportion of the family members with good health in each household.
Second, we focus on the binary data. Actually, there are four options in the health status questionnaire. The Multinomial-Dirichlet model can be an extension of the polychotomous data. Third, the two-fold sub-area level models can also be extended to three-fold models if the data have an additional hierarchical structure; actually, the NLSS II has this structure (households within wards, wards within districts). Fourth, in our models, we consider parametric priors. Introducing the Dirichlet process as a prior might make our method more robust to its specifications.

Author Contributions

Conceptualization, B.N. and L.C.; methodology, B.N. and L.C.; software, L.C.; validation, B.N. and L.C.; formal analysis, L.C.; investigation, L.C.; resources, B.N.; writing—original draft preparation, L.C.; writing—review and editing, B.N. and L.C.; visualization, L.C.; supervision, B.N. All authors have read and agreed to the published version of the manuscript.

Funding

Balgobin Nandram was supported by a grant from the Simons Foundation (#353953, Balgobin Nandram).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Exact Method for Sub-Area Logistic Regression Model

Recall that the joint posterior distribution of our two-fold logistic regression model is the joint posterior density for the parameters is
π ( v ˜ , w ˜ , β ˜ , σ 2 , δ 2 | y ˜ ) i = 1 j = 1 n i k = 1 m i j e ( x ˜ i j k β ˜ ( 0 ) + w i j ) y i j k 1 + e x ˜ i j k β ˜ ( 0 ) + w i j × 1 2 π σ 2 n exp i = 1 l j = 1 n i ( w i j ν i ) 2 2 σ 2 × 1 2 π δ 2 l exp i = 1 l ( ν i β 0 ) 2 2 δ 2 1 ( 1 + σ 2 ) 2 1 ( 1 + δ 2 ) 2 .
We can see that the form of the joint posterior density is very complicated. It is very time consuming to draw all the posterior samples if applying the exact MCMC method. But the exact method will provide reliable estimates of all parameters, so in order to test the performance of our approximation method, we need to apply MCMC method on our model and then compare the performance of two methods. We use Metropolis-Hastings sampler to draw samples for β ˜ , σ 2 , δ 2 together and then draw ν ˜ given β ˜ , σ 2 , δ 2 samples. At last, we use MH method to draw w ˜ given ν ˜ , β ˜ , σ 2 , δ 2 samples.
In order to draw samples for β ˜ , σ 2 , δ 2 together, we need to integrate out w ˜ and v ˜ . First, we integrate out w ˜ from the joint posterior density π ( v ˜ , w ˜ , β ˜ , σ 2 , δ 2 | y ˜ ) to get
π ( ν ˜ , β ˜ , σ 2 , δ 2 | y ˜ ) Ω i = 1 j = 1 n i k = 1 m i j e ( x ˜ i j k β ˜ ( 0 ) + μ i + w i j ) y i j 1 + e x ˜ i j k β ˜ ( 0 ) + μ i + w i j 1 2 π σ 2 e ( w i j ν i ) 2 2 σ 2 d w ˜ × 1 2 π δ 2 l exp i = 1 l ( ν i β 0 ) 2 2 δ 2 1 ( 1 + σ 2 ) 2 1 ( 1 + δ 2 ) 2 .
Notice that the integrant is not any simple distribution function, so we use Monte Carlo numberical integration to approximate the integrals. Let z i j w = w i j ν i σ . Notice that z i j w follows standard normal distribution. For standard normal density, 99.7% of data will fall within 3 standard deviations of the mean, which corresponds to the interval [ 3 , 3 ] . Therefore, we bounded the integration domain to [ 3 , 3 ] and divide the interval to M equal subintervals [ p a 1 , p a ] , a = 1 , , M . Then we can get an approximate but very accurate joint density
π ( ν ˜ , β ˜ , σ 2 , δ 2 | y ˜ ) i = 1 j = 1 n i a = 1 M p a 1 p a e k = 1 m i j ( x ˜ i j k β ˜ ( 0 ) + w i j ) y i j k k = 1 m i j 1 + e x ˜ i j k β ˜ ( 0 ) + w i j 1 2 π σ 2 e ( w i j ν i ) 2 2 σ 2 d w i j × 1 2 π δ 2 l exp i = 1 l ( ν i β 0 ) 2 2 δ 2 1 ( 1 + σ 2 ) 2 1 ( 1 + δ 2 ) 2 i = 1 j = 1 n i a = 1 M p a 1 p a e k = 1 m i j ( x ˜ i j k β ˜ ( 0 ) + σ z i j w + ν i ) y i j k k = 1 m i j 1 + e x ˜ i j k β ˜ ( 0 ) + σ z i j w + ν i 1 2 π e ( z i j w ) 2 2 d z i j w × 1 2 π δ 2 l exp i = 1 l ( ν i β 0 ) 2 2 δ 2 1 ( 1 + σ 2 ) 2 1 ( 1 + δ 2 ) 2
Let z ¯ a w = p a p a 1 2 , which is the midpoint of each interval [ p a 1 , p a ] , a = 1 , , M . We use midpoint rule to approximate the definite integrals. We divide the interval [ 3 , 3 ] into 100 subintervals, and so we use 100 midpoints to get the approximate joint posterior distribution
π ( ν ˜ , β ˜ , σ 2 , δ 2 | y ˜ ) i = 1 j = 1 n i a = 1 100 e k = 1 m i j ( x ˜ i j k β ˜ ( 0 ) + σ z ¯ a w + ν i ) y i j k k = 1 m i j 1 + e x ˜ i j k β ˜ ( 0 ) + σ z ¯ a w + ν i Φ ( a ) Φ ( a 1 ) × 1 2 π δ 2 l exp i = 1 l ( ν i β 0 ) 2 2 δ 2 1 ( 1 + σ 2 ) 2 1 ( 1 + δ 2 ) 2
Similarly, let z i ν = ν i β 0 δ and z ¯ b ν = p b p b 1 2 , b = 1 , , 100 . We use the midpoint rule to approximate the definite integral with respect to ν ˜ and then get the posterior density of β ˜ , σ 2 , δ 2 | y ˜
π ( β ˜ , σ 2 , δ 2 | y ˜ ) i = 1 b = 1 100 j = 1 n i a = 1 100 e k = 1 m i j ( x ˜ i j k β ˜ ( 0 ) + β 0 + z ¯ a w δ + z ¯ b ν σ ) y i j k k = 1 m i j 1 + e x ˜ i j k β ˜ ( 0 ) + β 0 + z ¯ a w δ + z ¯ b ν σ Δ Φ ( p a ) Δ Φ ( p b ) × 1 ( 1 + σ 2 ) 2 1 ( 1 + δ 2 ) 2 .
We propose to draw samples from β ˜ , σ 2 , δ 2 jointly by applying M-H sampler. Target function is π ( β ˜ , σ 2 , δ 2 | y ˜ ) . We set the proposal function as
β ˜ log σ 2 log δ 2 | y ˜ Normal β ˜ ¯ a log σ a 2 ¯ log δ a 2 ¯ , σ t 2 Σ a ,
where t σ t 2 χ t 2 , Chi-square on t degree of freedom, i.e. β ˜ log σ 2 log δ 2 | y ˜ Student’s t. Here t is tuning constant.
We also use M-H sampler draw samples for ν ˜ and w ˜ respectively. Proposal functions are π a ( ν ˜ β ˜ ( 0 ) , β 0 , σ 2 , y ˜ ) and π a ( w ˜ ν ˜ , β ˜ ( 0 ) , σ 2 , y ˜ ) respectively from the INNA method.
The target function to draw ν ˜ is
π ( ν ˜ | β ˜ , σ 2 , δ 2 , y ˜ ) i = 1 j = 1 n i a = 1 100 e k = 1 m i j ( x ˜ i j k β ˜ ( 0 ) + σ z ¯ a w + ν i ) y i j k k = 1 m i j 1 + e x ˜ i j k β ˜ ( 0 ) + σ z ¯ a w + ν i Φ ( a ) Φ ( a 1 ) × 1 2 π δ 2 l exp i = 1 l ( ν i β 0 ) 2 2 δ 2 .
After we get samples for ν ˜ , we can use M-H sampler to draw
π ( w ˜ | ν ˜ , β ˜ ( 0 ) , σ 2 , y ˜ ) i = 1 j = 1 n i k = 1 m i j e ( x ˜ i j k β ˜ ( 0 ) + w i j ) y i j k 1 + e x ˜ i j k β ˜ ( 0 ) + w i j × 1 2 π σ 2 n exp i = 1 l j = 1 n i ( w i j ν i ) 2 2 σ 2 .

Appendix B. Proof of Theorem 1

Proof. 
By Lemma 2, the posterior density is logconcave. Then according to Lemma 1, the posterior distribution τ ˜ | y ˜ is approximately a multivariate normal distribution.
By Lemma 1, evaluating all quantities at τ ˜ * , the mean is
μ ˜ w μ ˜ β = τ ˜ * H 1 g ˜ = w ˜ * β ˜ ( 0 ) * + E F F G g ˜ 1 g ˜ 2 = w ˜ * + E g ˜ 1 + F g ˜ 2 β ˜ ( 0 ) * + F g ˜ 1 + G g ˜ 2 .
Also, the covariance matrix is
H 1 = D C C B 1 = E F F G .
Therefore, by Lemma 1, the approximate joint posterior density of w ˜ , β ˜ ( 0 ) | y ˜ is
w ˜ β ˜ ( 0 ) | y ˜ Normal μ ˜ w μ ˜ β , E F F G .
Finally, using the property of the multivariate normal density, the conditional posterior density of w ˜ | β ˜ ( 0 ) , y ˜ and β ˜ ( 0 ) | y ˜ can also be approximated by multivariate normal distributions,
w ˜ | β ˜ ( 0 ) , y ˜ Normal { μ ˜ w D 1 C ( β ˜ ( 0 ) μ ˜ β ) , D 1 } a n d β ˜ ( 0 ) | y ˜ Normal { μ ˜ β , G } ,
where
μ ˜ w = w ˜ * + E g ˜ 1 + F g ˜ 2 a n d μ ˜ β = β ˜ ( 0 ) * + F g ˜ 1 + G g ˜ 2 .

Appendix C. Proof of Theorem 2

Proof. 
First, look at the exponent terms containing w ˜ in the above approximate posterior density function
w ˜ μ ˜ w D 1 C ( β ˜ ( 0 ) μ ˜ β ) D w ˜ μ ˜ w D 1 C ( β ˜ ( 0 ) μ ˜ β ) + w ˜ μ ˜ ν ( σ 2 I ) 1 w ˜ μ ˜ ν = Σ w ˜ ( D + 1 σ 2 I ) Σ w ˜ + μ ˜ w D 1 C β ˜ ( 0 ) μ ˜ β μ ˜ ν ( D 1 + σ 2 I ) 1 μ ˜ w D 1 C β ˜ ( 0 ) μ ˜ β μ ˜ ν ,
where Σ w ˜ = w ˜ ( D + 1 σ 2 I ) 1 D μ ˜ w C ( β ˜ ( 0 ) μ ˜ β ) + 1 σ 2 μ ˜ ν .
Then it can show that the π a ( w ˜ ν ˜ , β ˜ ( 0 ) , σ 2 , y ˜ ) is
w ˜ | ν ˜ , β ˜ ( 0 ) , σ 2 , y ˜ app Normal ( D + 1 σ 2 I ) 1 D μ ˜ w C ( β ˜ ( 0 ) μ ˜ β ) + 1 σ 2 μ ˜ ν , ( D + 1 σ 2 I ) 1 .
Notice that ( D + 1 σ 2 I ) is diagonal matrix. Then given ν ˜ , β ˜ ( 0 ) , σ 2 , y ˜ , all w i j s are independent. This is an important result because parallel computation can be done for w i j , which accommodates time-consuming and massive storage challenges in big data analysis. This result holds for the exact conditional posterior density of the μ i j . Since w ˜ has a multivariate normal distribution, we can integrate out w ˜ from the joint approximate posterior density π a ( w ˜ , ν ˜ , β ˜ , σ 2 , δ 2 y ˜ ) , and obtain the joint posterior density of ν ˜ , β ˜ , σ 2 and δ 2
π a ( ν ˜ , β ˜ , σ 2 , δ 2 y ˜ ) e 1 2 μ ˜ ν D 1 C β ˜ ( 0 ) μ ˜ β μ ˜ w ( D 1 + σ 2 I ) 1 μ ˜ ν D 1 C β ˜ ( 0 ) μ ˜ β μ ˜ w × e 1 2 ν ˜ β 0 j ˜ ( δ 2 I ) 1 ν ˜ β 0 j ˜ + β ˜ ( 0 ) μ ˜ β G 1 β ˜ ( 0 ) μ ˜ β × | D | 1 / 2 | δ 2 I | 1 / 2 | D + 1 σ 2 I | 1 / 2 | G | 1 / 2 1 ( 1 + σ 2 ) 2 1 ( 1 + δ 2 ) 2 .
Next, we will show that the approximate conditional posterior density of ν i is also normal distribution and all ν i s are independent as well. Here we consider each ν i . Let i = 1 j = 1 n i = n , ( σ i j 2 ) n × n = D 1 + σ 2 I , ( t ˜ i j ) n × 1 = D 1 C and ( μ w i j ) n × 1 = μ ˜ w .
Look at the exponent only containing ν i , i = 1 , , . in the π a ( ν ˜ , β ˜ , σ 2 , δ 2 y ˜ )
i = 1 j = 1 n i 1 σ i j 2 ν i μ w i j + t ˜ i j ( β ˜ ( 0 ) μ ˜ β ) 2 + 1 δ 2 i = 1 ( ν i β 0 ) 2 = i = 1 ( 1 j = 1 n i σ i j 2 + 1 δ 2 ) 1 ν i 1 j = 1 n i σ i j 2 μ ¯ w i t ˜ ¯ i ( β ˜ ( 0 ) μ ˜ β ) + 1 δ 2 β 0 1 j = 1 n i σ i j 2 + 1 δ 2 2 + i = 1 1 1 / j = 1 n i σ i j 2 + δ 2 1 μ ¯ w i t ˜ ¯ i ( β ˜ ( 0 ) μ ˜ β ) β 0 2 + i = 1 j = 1 n i 1 σ i j 2 ( t ˜ ¯ i t ˜ i j ) ( β ˜ ( 0 ) μ ˜ β ) ( μ ¯ w i μ w i j ) 2 ,
where μ ¯ w i = 1 n i j = 1 n i μ w i j and t ˜ ¯ i = 1 n i j = 1 n i t ˜ i j .
Then it is easy to see that
ν i | β 0 , σ 2 , δ 2 , y ˜ app Normal 1 j = 1 n i σ i j 2 μ ¯ w i t ˜ ¯ i ( β ˜ ( 0 ) μ ˜ β ) + 1 δ 2 β 0 1 j = 1 n i σ i j 2 + 1 δ 2 , 1 j = 1 n i σ i j 2 + 1 δ 2 .
Similarly, we can use parallel computing to draw ν i , i = 1 , , as well since all of them are independent given β ˜ ( 0 ) , β 0 , σ 2 , δ 2 . Then we can integrate out ν ˜ from the joint approximate posterior density π a ( ν ˜ , β ˜ , σ 2 , δ 2 y ˜ ) and obtain the joint posterior density of β ˜ , σ 2 and δ 2
π a ( β ˜ , σ 2 , δ 2 | y ˜ ) exp 1 2 i = 1 1 1 / j = 1 n i σ i j 2 + δ 2 1 μ ¯ w i t ˜ ¯ i ( β ˜ ( 0 ) μ ˜ β ) β 0 2 × exp 1 2 i = 1 j = 1 n i 1 σ i j 2 { ( t ˜ ¯ i t ˜ i j ) ( β ˜ ( 0 ) μ ˜ β ) ( μ ¯ w i μ w i j ) 2 × i = 1 l 1 j = 1 n i σ i j 2 + 1 δ 2 1 2 1 | δ 2 I | 1 / 2 | D + 1 σ 2 I | 1 / 2 1 ( 1 + σ 2 ) 2 1 ( 1 + δ 2 ) 2 = e 1 2 μ ˜ w D 1 C ( β ˜ ( 0 ) μ ˜ β ) β 0 j ˜ D 1 + σ 2 I + δ 2 I 1 μ ˜ w D 1 C ( β ˜ ( 0 ) μ ˜ β ) β 0 j ˜ × e 1 2 β ˜ ( 0 ) μ ˜ β G 1 β ˜ ( 0 ) μ ˜ β × i = 1 l 1 j = 1 n i σ i j 2 + 1 δ 2 1 2 1 | δ 2 D + δ 2 σ 2 I | 1 / 2 1 ( 1 + σ 2 ) 2 1 ( 1 + δ 2 ) 2 .
Next we assume that the conditional posterior density of β ˜ | σ 2 , δ 2 , y ˜ has an approximate multivariate normal density,
β 0 β ˜ ( 0 ) | σ 2 , δ 2 , y ˜ Normal ω 0 ω ˜ ( 0 ) , δ 0 2 γ ˜ γ ˜ Δ ( 0 ) 1 ,
which is denoted by π a ( β ˜ σ 2 , δ 2 , y ˜ ) . The density function is
π a ( β ˜ σ 2 , δ 2 , y ˜ ) δ 0 2 γ ˜ γ ˜ Δ ( 0 ) 1 2 × e 1 2 β 0 ω 0 β ˜ ( 0 ) ω ˜ ( 0 ) δ 0 2 γ ˜ γ ˜ Δ ( 0 ) β 0 ω 0 β ˜ ( 0 ) ω ˜ ( 0 )
So the exponent terms are
β 0 ω 0 β ˜ ( 0 ) ω ˜ ( 0 ) δ 0 2 γ ˜ γ ˜ Δ ( 0 ) β 0 ω 0 β ˜ ( 0 ) ω ˜ ( 0 ) .
Consider the exponent terms containing β ˜ ( 0 ) and β 0
μ ˜ w D 1 C ( β ˜ ( 0 ) μ ˜ β ) β 0 j ˜ D 1 + σ 2 I + δ 2 I 1 μ ˜ w D 1 C ( β ˜ ( 0 ) μ ˜ β ) β 0 j ˜ + β ˜ ( 0 ) μ ˜ β G 1 β ˜ ( 0 ) μ ˜ β = β ˜ ( 0 ) C D 1 ( D 1 + σ 2 I + δ 2 I ) 1 D 1 C + G 1 β ˜ ( 0 ) + j ˜ ( D 1 + σ 2 I + δ 2 I ) 1 j ˜ β 0 2 2 ( μ ˜ w + D 1 C μ ˜ β ) ( D 1 + σ 2 I + δ 2 I ) D 1 C + μ ˜ β G 1 β ˜ ( 0 ) 2 ( μ ˜ w + D 1 C μ ˜ β ) ( D 1 + σ 2 I + δ 2 I ) 1 j ˜ β 0 + 2 C D 1 ( D 1 + σ 2 I + δ 2 I ) 1 j ˜ β 0 β ˜ ( 0 ) + ( D 1 C μ ˜ β + μ ˜ w ) ( D 1 + σ 2 I + δ 2 I ) 1 j ˜ β 0 β ˜ ( 0 ) ( μ ˜ w + D 1 C μ ˜ β ) ( D 1 + σ 2 I + δ 2 I ) 1 ( μ ˜ w + D 1 C μ ˜ β ) + μ ˜ β G 1 μ ˜ β .
We know those two exponent parts are equal, so we have
Δ ( 0 ) = C D 1 ( D 1 + σ 2 I + δ 2 I ) 1 D 1 C + G 1 ,
δ 0 2 = j ˜ ( D 1 + σ 2 I + δ 2 I ) 1 j ˜ ,
γ ˜ = C D 1 ( D 1 + σ 2 I + δ 2 I ) 1 j ˜ ,
ω 0 ω ˜ ( 0 ) = δ 0 2 γ ˜ γ ˜ Δ ( 0 ) 1 ( μ ˜ w + D 1 C μ ˜ β ) ( D 1 + σ 2 I + δ 2 I ) 1 j ˜ ( μ ˜ w + D 1 C μ ˜ β ) ( D 1 + σ 2 I + δ 2 I ) D 1 C + μ ˜ β G 1 .
That is, β ˜ | σ 2 , δ 2 , y ˜ approximately follows multivariate normal distribution,
β 0 β ˜ ( 0 ) | σ 2 , δ 2 , y ˜ Normal ω 0 ω ˜ ( 0 ) , δ 0 2 γ ˜ γ ˜ Δ ( 0 ) 1 ,
Then we can easily integrate out β ˜ from the joint density of β ˜ , σ 2 , δ 2 | y ˜ , and get the posterior density of σ 2 , δ 2 | y ˜
π a ( σ 2 , δ 2 | y ˜ ) δ 0 2 γ ˜ γ ˜ Δ ( 0 ) 1 2 × i = 1 l 1 j = 1 n i σ i j 2 + 1 δ 2 1 2 1 | δ 2 D + δ 2 σ 2 I | 1 / 2 1 ( 1 + σ 2 ) 2 1 ( 1 + δ 2 ) 2 × exp 1 2 ( μ ˜ w + D 1 C μ ˜ β ) ( D 1 + σ 2 I + δ 2 I ) 1 ( μ ˜ w + D 1 C μ ˜ β ) + μ ˜ β G 1 μ ˜ β × exp 1 2 β 0 ω 0 β ˜ ( 0 ) ω ˜ ( 0 ) δ 0 2 γ ˜ γ ˜ Δ ( 0 ) β 0 ω 0 β ˜ ( 0 ) ω ˜ ( 0 ) .

References

  1. Central Bureau of Statistics Thapathali. Nepal living Standards Survey 2003/04; Statistical Report; Central Bureau of Statistics Thapathali: Kathmandu, Nepal, 2004; Volume 1.
  2. Rao, J.N.; Molina, I. Small Area Estimation; John Wiley & Sons: Hoboken, NJ, USA, 2015. [Google Scholar]
  3. Nandram, B.; Chen, L.; Fu, S.-T.; Manandhar, B. Bayesian Logistic Regression for Small Areas with Numerous Households. Stat. Appl. 2018, 1, 171–205. [Google Scholar]
  4. Yan, G.; Sedransk, J. Bayesian Diagnostic Techniques for Detecting Hierarchical Structure. Bayesian Anal. 2007, 2, 735–760. [Google Scholar]
  5. Yan, G.; Sedransk, J. A Note on Bayesian Residuals as A Hierarchical Model Diagnostic Technique. In Stat. Pap. 2010, 51, 1. [Google Scholar] [CrossRef]
  6. Fuller, W.A.; Goyeneche, J.J. Estimation of the State Variance Component. Unpublished Manuscript. 1988. [Google Scholar]
  7. Torabi, M.; Rao, J.N.K. On Small Area Estimation under a Sub-Area Level Model. In J. Multivar. Anal. 2014, 127, 36–55. [Google Scholar] [CrossRef]
  8. Chen, L.; Nandram, B.; Cruze, N.B. Hierarchical Bayesian Model with Inequality Constraints for US County Estimates. J. Off. Stat. 2022, 38, 709–732. [Google Scholar] [CrossRef]
  9. Erciulescu, A.L.; Cruze, N.B.; Nandram, B. Model-Based County Level Crop Estimates Incorporating Auxiliary Sources of Information. J. R. Stat. Soc. Ser. (Stat. Soc.) 2019, 182, 283–303. [Google Scholar] [CrossRef]
  10. Fay, R.E.; Herriot, R.A. Estimates of Income for Small Places: An Application of James-Stein Procedures to Census Data. J. Am. Stat. Assoc. 1979, 74, 269–277. [Google Scholar] [CrossRef]
  11. Stukel, D.M.; Rao, J.N.K. Estimation of Regression Models with Nested Error Structure and Unequal Error Variances Under Two and Three Stage Cluster Sampling. Stat. Probab. Lett. 1997, 35, 401–407. [Google Scholar] [CrossRef]
  12. Stukel, D.M.; Rao, J.N.K. On Small-Area Estimation under Two-Fold Nested Error Regression Models. J. Stat. Plan. Inference 1999, 78, 131–147. [Google Scholar] [CrossRef]
  13. Nandram, B.; Sedransk, J. Bayesian Predictive Inference for a Finite Population Proportion: Two-Stage Cluster Sampling. J. R. Stat. Soc. Ser. B (Methodol.) 1993, 55, 399–408. [Google Scholar] [CrossRef]
  14. You, Y.; Reiss, P. Hierarchical Bayes Small Area Estimation of Response Rates for an Expenditure Survey. In Proceedings of the Survey Methods Section; Statistical Society of Canada: Ottawa, ON, Canada, 2000; pp. 123–128. [Google Scholar]
  15. Nandram, B. Bayesian Predictive Inference of a Proportion Under a Twofold Small-Area Model. J. Off. Stat. 2016, 32, 187–208. [Google Scholar] [CrossRef]
  16. Lee, D.; Nandram, B.; Kim, D. Bayesian Predictive Inference of a Proportion under a Two-Fold Small Area Model with Heterogeneous Correlations. Surv. Methodol. 2017, 17, 69–92. [Google Scholar]
  17. Chen, L.; Nandram, B. A Hierarchical Bayesian Beta-Binomial Model for Sub-areas. In Applied Statistical Methods. ISGES 2020, Springer Proceedings in Mathematics & Statistics; Hanagal, D.D., Latpate, R.V., Chandra, G., Eds.; Springer: Singapore, 2022; pp. 23–40. [Google Scholar]
  18. Nandram, B. Discrimination between Complementary Log-log and Logistic Model for Ordinal Data. Commun. Stat. Theory Methods 1989, 18, 2155–2164. [Google Scholar] [CrossRef]
  19. Roberts, G.; Rao, J.N.K.; Kumar, S. Logistic Regression Analysis of Sample Survey Data. Biometrika 1987, 74, 1–12. [Google Scholar] [CrossRef]
  20. Nandram, B.; Chen, M.-H. Reparameterizing the Generalized Linear Model to Accelerate Gibbs Sampler Convergence. J. Stat. Comput. Simul. 1996, 54, 129–144. [Google Scholar] [CrossRef]
  21. Albert, J.H.; Chib, S. Bayesian Analysis of Binary and Polychotomous Response Data. J. Am. Stat. Assoc. 1993, 88, 669–679. [Google Scholar] [CrossRef]
  22. Farrell, P.J.; MacGibbon, B.; Tomberlin, T.J. Empirical Bayes Small Area Estimation Using Logistic Regression Models and Summary Statistics. J. Bus. Econ. Stat. 1997, 15, 101–108. [Google Scholar]
  23. Nandram, B.; Erhardt, E. Fitting Bayesian Two-Stage Generalized Linear Models Using Random Samples via the SIR Algorithm. Sankhya 2010, 66, 733–755. [Google Scholar]
  24. Nandram, B.; Choi, J.W. A Bayesian Analysis of Body Mass Index Data from Small Domains Under Nonignorable Nonresponse and Selection. J. Am. Stat. Assoc. 2010, 105, 120–135. [Google Scholar] [CrossRef]
  25. Scott, S.L.; Blocker, A.W.; Bonassi, F.V.; Chipman, H.A.; George, E.I.; McCulloch, R.E. Bayes and Big Data: The Consensus Monte Carlo Algorithm; Technical Report; Google, Inc.: Mountain View, CA, USA, 2013; pp. 1–22. [Google Scholar]
  26. Miroshnikov, A.; Wei, Z.; Conlon, E.M. Parallel Markov chain Monte Carlo for non-Gaussian posterior distributions. Stat 2015, 4, 304–319. [Google Scholar] [CrossRef] [Green Version]
  27. Rue, H.; Martino, S.; Chopin, N. Approximate Bayesian Inference for Latent Gaussian Models Using Integrated Nested Laplace Approximations. J. R. Stat. Soc. (Ser. B) 2009, 71, 319–392. [Google Scholar] [CrossRef]
  28. Fong, Y.; Rue, H.; Wakefield, J. Bayesian inference for generalized linear mixed models. Biostatistics 2010, 11, 397–412. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Illian, J.B.; Sørbye, S.H.; Rue, H. A toolbox for fitting complex spatial point process models using integrated nested Laplace approximation (INLA). Ann. Appl. Stat. 2012, 6, 1499–1530. [Google Scholar] [CrossRef] [Green Version]
  30. Ferkingstad, E.; Rue, H. Improving the INLA approach for approximate Bayesian inference for latent Gaussian models. Electron. J. Stat. 2015, 9, 2706–2731. [Google Scholar] [CrossRef]
  31. Cox, D.R.; Snell, E.J. Analysis of Binary Data, 2nd ed.; Chapman and Hall/CRC: London, UK, 1989. [Google Scholar]
  32. Wang, J.; Fuller, W.A. The Mean Squared Error of Small Area Predictors Constructed With Estimated Area Variances. J. Am. Stat. Assoc. 2003, 98, 716–723. [Google Scholar] [CrossRef]
  33. Yan, G.; Sedransk, J. Small Area Estimation Using Area Level Models and Estimated Sampling Variances. Surv. Methodol. 2006, 32, 97–103. [Google Scholar]
  34. Erciulescu, A.L.; Berg, E. Small area estimates for the conservation effects assessment project. In In Frontiers of Hierarchical Modeling in Observational Studies, Complex Surveys and Big Data: A Conference Honoring Professor Malay Ghosh, College Park, MD, USA; Women in Statistics Conference: Cary, NC, USA, 2014. [Google Scholar]
  35. Boyd, S.; Vandenberghe, L. Convex Optimization; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar]
  36. Rubin, D.B. The Bayesian Bootstrap. Ann. Stat. 1981, 9, 130–134. [Google Scholar] [CrossRef]
  37. Eddelbuettel, D.; Francois, R. Rcpp: Seamless R and C++ Integration. J. Stat. Softw. 2011, 40, 1–18. [Google Scholar] [CrossRef] [Green Version]
  38. Eddelbuettel, D.; Sanderson, C. Rcpparmadillo: Accelerating R with High-performance C++ Linear Algebra. Comput. Stat. Data Anal. 2014, 71, 1054–1063. [Google Scholar] [CrossRef]
  39. R Core Team. R: A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2019. [Google Scholar]
  40. Geweke, J. Evaluating the Accuracy of Sampling-Based Approaches to the Calculation of Posterior Moments. Bayesian Stat. 1992, 169–193. [Google Scholar]
Figure 1. Comparison of the INNA method and the exact method using the PSDs of the household proportions.
Figure 1. Comparison of the INNA method and the exact method using the PSDs of the household proportions.
Stats 06 00013 g001
Figure 2. Comparison of the INNA method and the exact method using the PSDs of the household proportions.
Figure 2. Comparison of the INNA method and the exact method using the PSDs of the household proportions.
Stats 06 00013 g002
Figure 3. Comparison of the INNA method and the exact method using the CVs of the household proportions.
Figure 3. Comparison of the INNA method and the exact method using the CVs of the household proportions.
Stats 06 00013 g003
Figure 4. Comparison of the INNA method and the exact method using the PMs of the ward proportions.
Figure 4. Comparison of the INNA method and the exact method using the PMs of the ward proportions.
Stats 06 00013 g004
Figure 5. Comparison of the INNA method and the exact method using the PSDs of the ward proportions.
Figure 5. Comparison of the INNA method and the exact method using the PSDs of the ward proportions.
Stats 06 00013 g005
Figure 6. Comparison of the INNA method and the exact method using the CVs of the ward proportions.
Figure 6. Comparison of the INNA method and the exact method using the CVs of the ward proportions.
Stats 06 00013 g006
Table 1. Distribution of wards and households in the sample.
Table 1. Distribution of wards and households in the sample.
StrataMountainsKathemanduUrban HillRural HillsUrban TaraiRural TaraiTotal
PSU3234289634102326
Households384408336115240812243912
Individuals19491954146757552104703420,263
Table 2. The descriptives of 4 covariates.
Table 2. The descriptives of 4 covariates.
CovariatesFrequencyPercentage
Age0–14776538.32
15–5910,95154.04
60+15477.64
GenderMale976348.18
Female10,50051.82
NativityIndigeous11,90341.25
Non-Indigous8,36058.75
ReligionHingdu16,37880.83
Non-Hingdu338519.17
Table 3. Posterior means (PM), associated posterior standard deviations (PSD), Geweke test p-values and effective sample sizes (ESS) of hyperparameters based on the INNA and exact method.
Table 3. Posterior means (PM), associated posterior standard deviations (PSD), Geweke test p-values and effective sample sizes (ESS) of hyperparameters based on the INNA and exact method.
MethodINNAExact
EstimatorPMPSDp-ValueESSPMPSDp-ValueESS
β 0 0.8020.3230.6887800.8400.4050.2641000
β 1 0.6750.0150.7598570.6540.0160.156858
β 2 0.3020.0170.4538940.3130.0190.6551000
β 3 −0.8620.0170.605915−0.8240.0170.418863
β 4 −0.3380.0130.718937−0.3600.0110.8391000
σ 2 115.24835.6010.4081000108.95938.7190.615911
δ 2 30.0021.1460.44873129.0341.6970.490780
Table 4. Comparison of posterior inference about the finite population proportions using the five-number summaries at the household level.
Table 4. Comparison of posterior inference about the finite population proportions using the five-number summaries at the household level.
Households MinQ1MeanQ3Max
PMINNA0.0490.5410.5650.5840.981
Exact0.0500.5420.5660.5840.984
PSDINNA0.0400.2460.4410.4650.500
Exact0.0380.2480.4420.4660.501
PCVINNA0.0410.3790.7620.8521.549
Exact0.0390.3840.7680.8521.577
Table 5. Comparison of posterior inference about the finite population proportions using five-number summaries at the ward level.
Table 5. Comparison of posterior inference about the finite population proportions using five-number summaries at the ward level.
Wards MinQ1MeanQ3Max
PMINNA0.4490.5370.5630.5890.684
Exact0.4500.5370.5640.5900.683
PSDINNA0.0560.0590.0630.0660.077
Exact0.0580.0610.0640.0660.077
PCVINNA0.0950.1030.1130.1210.163
Exact0.0970.1030.1130.1220.165
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, L.; Nandram, B. Bayesian Logistic Regression Model for Sub-Areas. Stats 2023, 6, 209-231. https://doi.org/10.3390/stats6010013

AMA Style

Chen L, Nandram B. Bayesian Logistic Regression Model for Sub-Areas. Stats. 2023; 6(1):209-231. https://doi.org/10.3390/stats6010013

Chicago/Turabian Style

Chen, Lu, and Balgobin Nandram. 2023. "Bayesian Logistic Regression Model for Sub-Areas" Stats 6, no. 1: 209-231. https://doi.org/10.3390/stats6010013

Article Metrics

Back to TopTop