Next Article in Journal
Optimal Multi-Step-Ahead Prediction of ARCH/GARCH Models and NoVaS Transformation
Previous Article in Journal
Misclassification in Binary Choice Models with Sample Selection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Comparison of Some Bayesian and Classical Procedures for Simultaneous Equation Models with Weak Instruments

1
Fannie Mae, 1100 15th St NW, Washington, DC 20005, USA
2
Department of Economics, University at Albany, SUNY, Albany, NY 12222, USA
*
Author to whom correspondence should be addressed.
Econometrics 2019, 7(3), 33; https://doi.org/10.3390/econometrics7030033
Submission received: 18 March 2019 / Revised: 3 July 2019 / Accepted: 23 July 2019 / Published: 29 July 2019

Abstract

:
We compare the finite sample performance of a number of Bayesian and classical procedures for limited information simultaneous equations models with weak instruments by a Monte Carlo study. We consider Bayesian approaches developed by Chao and Phillips, Geweke, Kleibergen and van Dijk, and Zellner. Amongst the sampling theory methods, OLS, 2SLS, LIML, Fuller’s modified LIML, and the jackknife instrumental variable estimator (JIVE) due to Angrist et al. and Blomquist and Dahlberg are also considered. Since the posterior densities and their conditionals in Chao and Phillips and Kleibergen and van Dijk are nonstandard, we use a novel “Gibbs within Metropolis–Hastings” algorithm, which only requires the availability of the conditional densities from the candidate generating density. Our results show that with very weak instruments, there is no single estimator that is superior to others in all cases. When endogeneity is weak, Zellner’s MELO does the best. When the endogeneity is not weak and ρ ω 12 > 0 , where ρ is the correlation coefficient between the structural and reduced form errors, and ω 12 is the covariance between the unrestricted reduced form errors, the Bayesian method of moments (BMOM) outperforms all other estimators by a wide margin. When the endogeneity is not weak and β ρ < 0 ( β being the structural parameter), the Kleibergen and van Dijk approach seems to work very well. Surprisingly, the performance of JIVE was disappointing in all our experiments.

1. Introduction

Research on Bayesian analysis of the simultaneous equations models addresses a problem, raised initially by Maddala (1976), and now recognized as related to the problem of local nonidentification when diffuse/flat priors are used in traditional Bayesian analysis, e.g., Drèze (1976); Drèze and Morales (1976), and Drèze and Richard (1983).1 In this paper, we examine the approaches developed by Chao and Phillips (1998, hereafter CP), Geweke (1996), Kleibergen and van Dijk (1998, hereafter KVD), and Zellner (1998). The idea in KVD was to treat an overidentified simultaneous equations model (SEM) as a linear model with nonlinear parameter restrictions, and has been extended further in Kleibergen and Zivot (2003). While KVD focused mainly on resolving the problem of local nonidentification, CP explored further the consequences of using a Jeffreys prior. By deriving the exact and (asymptotically) approximate representations for the posterior density of the structural parameter, CP showed that the use of a Jeffreys prior brings Bayesian inference closer to classical inference in the sense that this prior choice leads to posterior distributions which exhibit Cauchy-like tail behavior akin to the LIML estimator. Geweke (1996), being aware of the potential problem of local nonidentification, suggests a shrinkage prior such that the posterior density is properly defined for each parameter. In another approach, Zellner (1998) suggested a finite sample Bayesian method of moments (BMOM) procedure based on given data without specifying a likelihood function or introducing any sampling assumptions.
For the Bayesian approaches considered, while Geweke (1996) proposed Gibbs sampling (GS) to evaluate the posterior density with a shrinkage prior, the posterior densities as well as their conditional densities resulting from CP and KVD are nonstandard and cannot be readily simulated. In the category of “block-at-a-time” approach, we suggest a novel MCMC procedure, which we call a “Gibbs within M–H” algorithm. The advantage of this algorithm is that it only requires the availability of the conditional densities from the candidate generating density. These conditional densities are used in a Gibbs sampler to simulate the candidate generating density, whose drawings on convergence are then weighted to generate drawings from the target density in a Metropolis–Hastings (M–H) algorithm. In this study, we will focus on weak instruments, where the classical approach has been particularly problematic.2 Ni and Sun (2003) have studied similar issues in the context of vector autoregressive models, see also Ni et al. (2007). Radchenko and Tsurumi (2006) used many of the procedures analyzed in this paper to estimate a gasoline demand model using an MCMC algorithm.
The main objective of the present paper is to compare the small sample performance of some Bayesian and classical approaches using Monte Carlo simulations. For the purpose of comparison, a number of classical methods including OLS, 2SLS, LIML, Fuller’s modified LIML, and a jackknife instrumental variables estimator (JIVE) originally due to Angrist et al. (1999) and Blomquist and Dahlberg (1999) are also computed from the generated data. Our simulation results from repeated sampling experiments provide some unambiguous guidelines for empirical practitioners.
The plan of the paper is as follows. In Section 2, we set up the model. Section 3 reviews in limited details the recent Bayesian approaches and JIVE. Section 4 suggests a new MCMC procedure for evaluating the posterior distributions for CP and KVD, and discusses the convergence diagnostics implemented. Section 5 presents simulation results and some discussions. Section 6 contains the main conclusions.

2. The Model

Consider the following limited information formulation of the m-equation simultaneous equations model (LISEM):
y 1 = Y 2 β + Z 1 γ + u
Y 2 = Z 1 Π 1 + Z 2 Π 2 + V 2
where y1: (T × 1) and Y2: (T × (m − 1)) are the m included endogenous variables; Z1: (T × k1) is an observation matrix of exogenous variables included in the structural Equation (1); Z2: (T × k2) is an observation matrix of exogenous variables excluded from (1); and u and V2 are, respectively, a T × 1 vector and a T × (m − 1) matrix of random disturbances to the system. We assume that (u, V2) ∼ N (0, Σ ⊗ IT), where the m × m covariance matrix is positive definite symmetric (pds) and is partitioned conformably with the rows of (u, V2) as follows
Σ = ( σ 11 σ 21 σ 21 Σ 22 )
The likelihood function for the model described by (1) and (2) can be written as
L ( β , γ , Π 1 , Π 2 , Σ | Y , Z ) = ( 2 π ) T m / 2 | Σ | T / 2 e x p { 1 2 t r [ Σ 1 ( u , V 2 ) ( u , V 2 ) ] } ,
where Y = ( y 1 , Y 2 ) and Z = ( Z 1 , Z 2 ) .
The structural model described by (1) and (2) can alternatively be written in its reduced form
( y 1 Y 2 ) = ( Z 1 Z 2 ) ( π 1 Π 1 Π 2 β Π 2 ) + ( ξ 1 V 2 )
where π 1 = γ + Π 1 β , ξ 1 = u + V 2 β , ( ξ 1 ,   V 2 ) ~ N ( 0 , Ω I T ) , and Σ = C Ω C , C = ( 1 0 β I m 1 ) . The likelihood function corresponding to this alternative representation is:
L ( β , γ , Π 1 , Π 2 , Ω | Y , Z ) = ( 2 π ) T m / 2 | Ω | T / 2 e x p { 1 2 t r [ Ω 1 ( ξ 1 , V 2 ) ( ξ 1 , V 2 ) ] }
The likelihood functions (3) and (5) are equivalent since the Jacobian between Ω and Σ is unity.
Geweke (1996) considers the following reduced rank regression specification3
Y = Z 1 A + Z 2 Θ + E ,
where A = ( Π 1 , π 1 ) , Θ = Π 2 Φ _ and Φ _ = ( I m 1 ,   β ) , E = ( V 2 , ξ 1 ) ~ N ( 0 , Σ _ I T with Σ _ 1 = ( Σ _ 11   Σ _ 12   Σ _ 21   Σ _ 22   ) partitioned conformably with the rows of ( V 2 , ξ 1 ) . Obviously, (6) is equivalent to (4) and the corresponding likelihood function is similar to (5).
Note that in the absence of restrictions on the covariance structure, (1) is fully identified if and only if r a n k ( Π 1 ) = ( m 1 ) k 2 .

3. Review of Some Bayesian Formulations

Among the Bayesian approaches, Geweke (1996) used a shrinkage prior such that all parameters are identified (in the sense that a proper posterior distribution exists) even when Π 2 has reduced rank. KVD treated overidentified SEMs as linear models with nonlinear parameter restrictions using the singular value decomposition. A diffuse or natural conjugate prior for the parameters of the embedding linear model results in the posterior for the parameters of the SEM having zero weight in the region of parameter space where Π 2 has reduced rank. This is a feature of the Jacobian of transformation from the multivariate linear model to the SEM. CP used a prior by applying Jeffreys principle on the model described by (1) and (2) and the assumptions regarding the disturbances. An important advantage of the Jeffreys prior in the present context is that it places no weight in the region of the parameter space where r a n k ( Π 2 ) < ( m 1 ) and relatively low weight in close neighborhoods of this region where the model is nearly unidentified.

3.1. Zellner’s Bayesian Method of Moments Approach (BMOM)

Among the various Bayesian treatments of SEM proposed by Zellner (1971, 1978, 1986, 1994, 1998), the Bayesian method of moments approach applies the principle of maximum entropy and generates optimal estimates which can be evaluated by double K-class estimators. Given the unrestricted reduced form equation y 1 = Z π ¯ 1 + ξ 1 , Zellner (1998) considered a balanced loss function,
L b = ω L g + ( 1 ω ) L p = ω ( y 1 X δ ^ ) ( y 1 X δ ^ ) + ( 1 ω ) ( Z π ¯ 1 X δ ^ ) ( Z π ¯ 1 X δ ^ ) ,   for   0 ω 1
where X = ( Y 2 , Z 1 ) , δ = ( β , γ ) , and δ ^ is an estimate of δ . The BMOM estimate that minimizes E L b , where the expectation is taken with respect to a probability density function of the π matrices of the unrestricted reduced form equations, is given by
( β ^ γ ^ ) = ( Y 2 Y 2 K 1 V ^ 2 V 2 ^ Y 2 Z 1 Z 1 Y 2 Z 1 Z 1 ) 1 ( ( Y 2 K 2 V 2 ^ ) y 1 Z 1 y 1 ) ,
where K 1 = 1     k   /   ( T k ) ,   K 2 =   1   ( 1 ω ) k / ( T k ) with 0   ω 1 and V 2 ^ = ( I Z ( Z Z ) 1 Z ) Y 2 .
BMOM estimate will vary depending on the value of ω . When ω = 1 , it is the optimal estimate resulting from a “goodness of fit” loss function Lg. When ω = 0, it is the optimal estimate given by a precision of estimation loss function Lp. Meanwhile, the well-known minimum expected loss (MELO) estimator is derived using a precision of estimation loss function and may be evaluated as a K-class estimator with
K 1 = K 2 = 1 k / ( T k m 1 ) .
Similar to the BMOM method, Conley et al. (2008) developed a Bayesian semiparametric approach to the instrumental variable problem assuming linear structural and reduced form equations, but with unspecified error distributions.

3.2. The Geweke Approach

Geweke (1996) assumes the following reference prior
| Σ _ | ( m + v + 1 ) / 2 exp [ 1 2 t r S Σ _ 1 ] exp [ τ 2 2 ( β β + t r Π 2 Π 2 + t r   A   A ) ] ,  
which is the product of an independent inverted Wishart distribution for Σ with v degrees of freedom and scale matrix S, and an independent N(0, τ 2 ) shrinkage priors for each element of β and Π 2 . Geweke derived the respective conditional posterior distributions, which may be used to generate drawings through Gibbs sampling from the joint posterior distribution. Regarding the vector of parameters ( Σ _ 1 , A , Π 2 , β ) , we obtain the full conditional densities as follows:
(1)
Conditional density of Σ _ 1
Σ _ 1 | ( Π 2 , β , A , Z , Y ) ~ Wishart   ( T + v ,   G 1 ) ,
where G   =   S   +   ( Y     Z 1 A     Z 2 Θ ) ( Y     Z 1 A     Z 2 Θ ) .
(2)
Conditional density of A
vec ( A ) | ( Π 2 , β , Σ _ 1 ,   Z ,   Y ) ~ N ( [ Σ _ 1 Z 1 Z 1 + τ 2 I m k 1 ] 1 [ Σ _ 1 Z 1 Z 1 ] vec ( A ^ ) , [ Σ _ 1 Z 1 Z 1   + τ 2 I m k 1 ] 1 ) ,  
where A ^   = ( Z 1 Z 1 ) 1   Z 1 ( Y     Z 2 Θ ) .
(3)
Conditional density of Π 2 4
vec ( Π 2 ) | ( β , Σ _ 1 ,   A ,   Z ,   Y ) ~ N ( [ Σ ˜ 11 Z 2 Z 2 + τ 2 I k 2 ( m 1 ) ] 1 [ Σ ˜ 11 Z 2 Z 2 ] vec ( Π ^ 2 ) , [ Σ ˜ 11 Z 2 Z 2 + τ 2 I k 2 ( m 1 ) ] 1 ) ,  
where Π ^ 2 = Θ ^ [ Φ _ + + Φ _ 0 Σ ˜ 21 ( Σ ˜ 11 ) 1 ] , Θ ^ = ( Z 2 Z 2 ) 1 Z 2 ( Y Z 1 A ) . Φ _ + is the Moore–Penrose generalized inverse of Φ _ and the columns of Φ _ + and Φ _ 0 are orthogonal, and C ≡ ( Φ _ + , Φ _ 0 ) is and m × m nonsingular matrix. Finally, Σ ˜ i j denotes the partitioning of Σ ˜ 1 = ( C Σ _ C ) 1 conformably with Y = ( Y 2 , y 1 ) .
(4)
Conditional density of β
β | ( Π 2 , Σ _ 1 , A ,   Z ,   Y )  
~   N ( [ Σ _ 22 Π 2 Z 2 Z 2 Π 2 + τ 2 I m 1 ] 1 [ Σ _ 22 Π 2 Z 2 Z 2 Π 2 ] β ^ , [ Σ _ 22 Π 2 Z 2 Z 2 Π 2 + τ 2 I m 1 ] 1 ) ,
where
β ^ = ( Π 2 Z 2 Z 2 Π 2 ) 1 Π 2 Z 2 ( Y 2 Z 1 Π 1 ) Σ _ 12 ( Σ _ 22 ) 1   Σ _ 12 ( Σ _ 22 ) 1 + ( Π 2 Z 2 Z 2 Π 2 ) 1 Π 2 Z 2 ( y 1 Z 1 π 1 )

3.3. The Chao and Phillips Approach

Using Jeffreys prior, CP obtains exact and approximate analytic expressions for the posterior density of the structural coefficient β in the LISEM (1) and (2). Their formulas are found to exhibit Cauchy-like tails analogous to comparable results in the classical literature on LIML estimation. For the model (1) and (2) under normality assumption for the disturbances, a Jeffreys prior on the parameters, θ = ( β , γ , Π 1 , Π 2 , Σ ) , is of the form
p ( β , γ , Π 1 , Π 2 , Σ ) | E { 2 θ   θ   ln L ( θ | Y , Z ) } | 1 / 2 | σ 11 | ( k 2 m + 1 ) / 2 | Σ | ( k + m + 1 ) / 2 | Π 2 Z 2 Q Z 1 Z 2 Π 2 | 1 / 2
where l n   L ( θ | Y ,   Z ) is the log-likelihood function as specified in (3), and Q X = I T     P X ,   P X = X ( X X ) 1 X . As first noted by Poirier (1996), the prior in (13) places no weight where r a n k ( Π 2 ) < ( m 1 ) through the factor | Π 2 Z 2 Q Z 1 Z 2 Π 2 | 1 / 2 .
The joint posterior of the parameters of LISEM (1) and (2) is constructed as proportional to the product of the prior (13) and the likelihood function (3),
p ( β , γ , Π 1 , Π 2 , Σ | Y , Z ) p ( β , γ , Π 1 , Π 2 , Σ ) L ( β , γ , Π 1 , Π 2 , Σ | Y , Z )
| σ 11 | ( k 2 m + 1 ) / 2 | Σ | ( T + k + m + 1 ) / 2 | Π 2 Z 2 Q Z 1 Z 2 Π 2 | 1 / 2 × exp { 1 2 t r [ Σ 1 ( u , V 2 ) ( u , V 2 ) ] }
where (u, V2) is defined in (1) and (2). Note that (14) or its conditionals do not belong to any standard class of probability density functions.

3.4. The Kleibergen and van Dijk Approach

To solve the problem of local nonidentification and to avoid the so-called Borel–Kolmogorov paradox, see Billingsley (1986) and Poirier (1995), KVD considered (4) as a multivariate linear model with nonlinear parameter restrictions:
( y 1 Y 2 ) = ( Z 1 Z 2 ) ( π 1 Π 1 ϕ 2 Φ 2 ) + ( ξ 1 V 2 ) ,
where ϕ 1 is a k 2 × 1 vector, Φ 2 is a k 2 × ( m 1 ) matrix. Denote Φ = ( ϕ 1 , Φ 2 ) . The reduced form model (4) is obtained if a reduced rank restriction is imposed on the linear model (15) such that r a n k ( Φ ) = ( m 1 ) instead of m.
Using a singular value decomposition (SVD) of Φ , they show that (15) is identical to the so-called unrestricted reduced form (URF) model,5
( y 1 Y 2 ) = Z 1 ( π 1 Π 1 ) + Z 2 Π 2 B + Z 2 Π 2 λ B + ( ξ 1 V 2 ) ,
where B = ( β I m 1 ) , λ is a ( k 2 m 1 ) × 1 vector. Π 2 and B are the orthogonal complements of Π 2 and B, respectively, such that Π 2 Π 2 0 , BB 0 , and Π 2 Π 2 I k 2   m 1 , B B   1 (i.e., Π 2 = ( Π 22 Π 21 1   I k 2 m 1 ) ( I k 2 m 1 + Π 22 Π 21 1 Π 21 1 Π 22 )   1 / 2 , where Π 2 =   ( Π 21   Π 22 ) , Π 21 :   ( m 1 ) × ( m 1 ) , Π 22 :   ( k 2 m 1 ) × ( m 1 ) , and B = ( 1 + β β ) 1 / 2 ( 1 β ) ).
There is one-to-one correspondence between the parameters in (15) and (16). The SVD of Φ is,
Φ   =   U S V ,
where U :   k 2 ×   k 2 , U U = I k 2 ; V :   m × m ; V V = I m ; and S :   k 2 ×   m is a rectangular matrix containing the (nonnegative) singular values (in decreasing order) on its main diagonal (i.e., (s11, s11, …, smm)). Rewrite
U = ( U 11 U 12 U 21 U 22 ) , = ( S 1 0 0 S 2 ) ,   and   V = ( v 11 v 12 v 21 v 22 ) ,  
where U 11 , S 1 , v 21 : ( m 1 ) × ( m 1 ) ; v 12 : 1 × 1 ; v 11 , v 22 : ( m 1 ) × 1 ; U 12 :   ( m 1 ) × ( k 2 m + 1 ) × ( m 1 ) ; U 21 : ( k 2 m 1 ) × ( m 1 ) ; U 22 :   ( k 2     m 1 ) × ( k 2 m + 1 ) ; S 2 :   ( k 2     m 1 ) × 1 , then the following relationship between ( Π 2 , β ,   λ ) and ( U ,   S ,   V )   results, Π 2 = ( U 11 U 12 ) S 1 V 21 , β = V 21 1 v 11   , and
λ = ( U 22 U 22 ) 1 / 2 U 22 S 2 v 12 ( v 12 v 12 ) 1 / 2 .
Note that λ is obtained through pre- and postmultiplication of s2 by orthogonal matrices while s2 contains the smallest singular values of Φ and is invariant with respect to the ordering of variables contained in Y and Z2.
According to KVD, the above shows that the model described by (1) and (2) can be considered as equivalent to the linear model (16) with a nonlinear (reduced rank) restriction λ = 0 on the parameters. Therefore, the priors and posteriors of the parameters of the LISEM (1) and (2) may be constructed as proportional to the priors and posteriors of the parameters of the linear model (16) evaluated at λ = 0 .
A diffuse (Jeffreys) prior for the parameters ( π 1 , Π 1 , Φ , ) of the linear model6
p ( π 1 , Π 1 , Φ , Ω ) | Ω | ( k + m + 1 ) / 2 | Ω | ( m + 1 ) / 2 | Ω 1 Z Z | 1 / 2
where k = k 1 + k 2 , implies the prior for the parameters ( β , π 1 , Π 1 , Π 2 , ) of the LISEM (4) as
p ( β ,   π 1 , Π 1 , Π 2 , Ω ) p ( π 1 , Π 1 Φ ( Π 2 ,   β , λ ) , Ω ) λ = 0 J ( Φ , ( Π 2 , β , λ ) ) | λ = 0 | | Ω | ( m + 1 ) / 2 | Ω 1 Z Z | 1 / 2   |   J ( Φ , ( Π 2 , β , λ ) ) | λ = 0 | | Ω | ( m + 1 ) / 2 | Ω 1 Z Z | 1 / 2 × | ( B     I k 2   e 1     Π 2   B   Π 2 ) | ,  
where e 1 = ( 1 , 0 , 0 , , 0 ) . Note that the prior (21) is the Jeffreys prior of the unrestricted reduced form (16) evaluated at λ = 0 . Most importantly, |J(Φ, (Π2, β, λ))|λ=0 = 0 when Π 2 has reduced rank. This feature in KVD approach eliminates the potential impact of local nonidentification.
The joint posterior of the parameters of the LISEM (4) is readily constructed as proportional to the product of the prior (21) and the likelihood function (5),
p ( β ,   π 1 , Π 1 , Π 2 , Ω   |   Y ,   Z ) p ( β , π 1 , Π 1 , Π 2 ,   Ω ) L * ( β ,   γ , Π 1 , Π 2 , Ω   |   Y ,   Z )   | Ω | ( T + m + 1 ) / 2 | Ω 1 Z Z | 1 / 2   × | ( B I k 2   e 1 Π 2   B Π 2 )   | × exp { 1 2 t r [ Ω 1 ( ( y 1 Y 2 ) ( Z 1 Z 2 ) ( π 1 Π 1 Π 2 β Π 2 ) ) ( ( y 1 Y 2 ) ( Z 1 Z 2 ) ( π 1 Π 1 Π 2 β Π 2 ) ) ] }  
Unfortunately, the above posterior or its conditional densities do not belong to a known class of probability density functions.

3.5. The Jackknife Instrumental Variable Estimator (JIVE)

Motivated by split sample instrumental variables estimators, Angrist et al. (1999) and Blomquist and Dahlberg (1999) independently suggested a jackknife instrumental variable estimator (JIVE). For model (1) and (2), JIVE is given by
δ ^ jive = ( X ^ jive X ) 1 ( X ^ jive y 1 )
where X ^ jive is the T × ( m 1 + k 1 ) matrix with t-th row defined by
Z t Π ^ t =   Z t ( Z t Z t ) 1 ( Z t X t ) = Z t Π ^ h t X t 1 h t ,
Z t and X t are ( T 1 )   ×   k and ( T 1 ) × ( m 1 + k 1 ) matrices obtained after eliminating the t-th rows of Z and X matrices respectively, Π ^ = ( Z Z ) 1 ( Z X ) and h t = Z t ( Z Z ) 1 Z t . In JIVE, the instrument is independent of the disturbances even in finite samples, which is achieved by using a ‘leave-one-out’ jackknife-type fitted value in place of the usual unrestricted reduced form predictions.
Angrist et al. (1999) also proposed a second jackknife estimator that is a slight modification of (23). Similar to their study, we found that its performance is very similar to JIVE, and is not reported here.

4. Posterior Simulator: “Gibbs within M–H” Algorithm

Given the full conditional densities in (9) through (12) for the four blocks of parameters, evaluating the joint posterior densities by Gibbs sampling is straightforward, see Geweke (1996) for a detailed description. Although Geweke’s (1996) shrinkage prior does not meet the argument in KVD that the implied prior/posterior on the parameters of an embedding linear model should be well-behaved, we found that the use of Geweke’s shrinkage prior does not lead to a reducible Markov Chain. With the specification of a shrinkage prior, when Π 2 has reduced rank, the joint posterior density still depends on β and will not exhibit any asymptotic cusp. In the following, we only discuss the posterior simulation for CP and KVD.
KVD suggested two simulation algorithms for the posterior (22): an Importance sampler and a Metropolis–Hastings algorithm. We found that their M–H algorithm performs unsatisfactorily with low acceptance rate even for reasonable parameter specifications.7 As mentioned earlier, since the posteriors (14) and (22) as well as their conditional posteriors do not belong to any standard class of probability density functions, Gibbs sampling cannot be used. In this section, we suggest an alternative simulation algorithm which combines Gibbs sampling (see Casella and George (1992) and Chib and Greenberg (1996)) and the Metropolis–Hastings algorithm (see Metropolis et al. 1953; Hastings 1970; Smith and Roberts 1993; Tierney 1994; Chib and Greenberg 1995). Our algorithm is different from the “M–H within Gibbs” algorithm and can find its usefulness in other applications as well.
To generate drawings from the target density p(x), we use a candidate-generating density r(x). An Independence sampler, which is a special case of the M–H sampler, in algorithmic form is as follows:
0.
Choose starting values x0
1.
Draw xi from r(x)
2.
Accept xi with probability
α ( x i 1 , x i ) = {   min   ( p ( x i ) r ( x i 1 ) p ( x i 1 ) r ( x i ) , 1 ) ,   if   p ( x i 1 ) r ( x i ) > 0 1 ,   if   p ( x i 1 ) r ( x i )   =   0 ,
otherwise x i = x i 1
3.
i = i   + 1 . Go to 1.
It is generally not feasible to draw all elements of the vector x simultaneously. A block-at-a-time possibility was first discussed in (Hastings 1970, sct. 2.4) and then in Chib and Greenberg (1995) along with an example.
Chib and Greenberg (1995) considered applying the M–H algorithm in turn to sub-blocks of the vector x , which presumes that the target density p(x) may be manipulated to generate full conditional densities for each of the sub-blocks of x , conditioning on other elements of x. However, the full conditionals are sometimes not readily available from the target density for empirical investigators. The posteriors (14) and (22) happen to fall in this category. In this latter case, problems come up at step 1 while trying to generate drawings from the joint marginal density r ( x ) . Note that these drawings, whether accepted or rejected at step 2, satisfy the necessary reversibility condition if step 1 is performed successfully.
To simplify the notation, we consider a vector x which contains two blocks, x = ( x 1 , x 2 ) . KVD used the fact that
r ( x 1 , x 2 ) = r ( x 1 ) r ( x 2 | x 1 )
and suggested to draw x 1 i from r ( x 1 ) and then draw x 2 i from r ( x 2 | x 1 i ) . The pair ( x 1 i , x 2 i ) is then taken as a drawing from r ( x ) . It turns out that this strategy gives very low acceptance rate at step 2 in simulation studies for various reasonable parameter values. Sometimes the move never takes place and the posterior has all its mass at the parameter values of the first drawing. The reason for the failure is that information is not updated at subsequent drawings and the transition kernel of (25) is static.
If the full conditionals r ( x 1 | x 2 ) and r ( x 2 | x 1 ) are available, which is usually true for many standard densities, we propose to use them in a Gibbs sampler to make independent drawings from the invariant density r ( x ) after the Markov chain has converged.
The combined algorithm is thus as follows, which we call “Gibbs within M–H”:
0.
Choose starting values x 0 = ( x 1 0 + x 2 0 )
1.
Draw x 1 i from r ( x 1 | x 2 i 1 ) , draw x 2 i from r ( x 2 | x 1 i ) .
2.
Accept x i = ( x 1 i   x 2 i ) with probability α ( x 1 i 1 ,   x 2 i ) as defined in (24), otherwise x i = x i 1 .
3.
i = i + 1 . Go to 1.
As explained, step 2 is the Gibbs step and step 3 is the M–H step in our combined algorithm. In the following subsections, we describe the steps for implementing the above procedure to generate drawings from the posteriors (14) and (22).8

4.1. Implementing the CP Approach

Note that the posterior in the CP approach is proportional to the product of the prior, which is uniformly bounded, and the likelihood function, which can be sampled by a Gibbs sampler. Therefore, we choose the candidate-generating density the way suggested by Chib and Greenberg (1995): we use the likelihood function, L ( β , γ , Π 1 , Π 2 , Σ | Y , Z ) , as the candidate generating density for the posterior (14). Using precision matrix Σ−1, the simulation steps are as follows,
0.
Choose starting values ( β 0 , γ 0 , Π 1 0 , Π 2 0 , Σ 1 , 0 )
1.
Draw Σ−1,i from p ( Σ 1 | β i 1 , γ i 1 , Π 1 i 1 , Π 2 i 1 , Y , Z )
Draw ( β i , γ i , Π 1 i , Π 2 i ) from p ( β , γ , Π 1 , Π 2 | Σ 1 , i , Y , Z )
2.
Accept β i , γ i , Π 1 i , Π 2 i , Σ 1 , i as a drawing from the posterior (14) with probability,
min ( | σ 11 i | ( k 2 m + 1 ) / 2 | Σ 1 , i | ( k m + 1 ) / 2 | Π 2 i Z 2 Q Z 2 Z 2 Π 2 i | 1 / 2 | σ 11 i 1 | ( k 2 m + 1 ) / 2 | Σ 1 , ( i 1 ) | ( k m + 1 ) / 2 | Π 2 i 1 Z 2 Q Z 1 Z 2 Π 2 i 1 | 1 / 2 , 1 ) ,
Otherwise,   β i , γ i , Π 1 i , Π 2 i , Σ 1 , i   = ( β i 1 , γ i 1 , Π 1 i 1 , Π 2 i 1 , Σ 1 , ( i 1 ) ) .
3.
i = i + 1 . Go to 1.
The conditional densities used in the first step are constructed as follows (see Percy (1992) and Chib and Greenberg (1996)): Rewrite the model (1) and (2) as a SUR model,
y t = W t δ + ( u t V 2 , t ) ,
where
y t = ( y 1 , t Y 2 , t )
W t = ( ( Y 2 , t   Z 1 , t ) 0 0 ( I m 1 Z t ) ) ,
δ = ( β , γ , ( vec ( Π 1 Π 2 ) ) ) . Then
p ( Σ 1   |   δ ,   Y ,   Z ) | Σ 1 | ( T 2 ( m + 1 ) ) / 2 exp [ 1 2 t r ( Σ 1 H ) ]
which follows a Wishart distribution with ( T m 1 ) degrees of freedom, where H = t = 1 T ( y t W t δ ) ( y t W t δ ) , and
p ( δ   |   Σ 1 ,   Y ,   Z ) = N ( ( t = 1 T W t Σ 1 W t ) 1 ( t = 1 T W t Σ 1 y t ) , ( t = 1 T W t Σ 1 W t ) 1 )

4.2. Implementing the KVD Approach

KVD proposed to use the posterior of the unrestricted linear model (16), p ( β , λ , Π 2 , | Y ,   Z ) , as the candidate generating density of the posterior (22), p ( β , Π 2 , | Y , Z ) , where the parameters ( π 1 , Π 1 ) have been concentrated out. First ( Φ ,   ) is generated from p ( Φ , | Y , Z ) , and then ( β , λ , Π 2 ) is obtained from Φ using (19). However, λ is also sampled which is not present in the posterior p ( β , Π 2 , | Y , Z ) . Therefore, KVD assumes that λ is generated by a conditional density of the form,
g ( λ | β , Π 2 , Ω ) = ( 2 π ) ( k 2 m + 1 ) / 2 | B Ω 1 B | ( k 2 m + 1 ) / 2 | Π 2 Z 2 M Z 1 Z 2 Π 2 | 1 / 2     × exp [ 1 2 t r ( B Ω 1 B ( λ λ ^ ) Π 2 Z 2 M Z 1 Z 2 Π 2 ) ( λ λ ^ ) ) ] ,  
where λ ^ = ( Π 2 Z 2 M Z 1 Z 2 Π 2 ) 1 Π 2 Z 2 M Z 1 ( Y Z 2 Π 2 B ) Ω 1 B ( B Ω 1 B ) 1 .
Therefore, the density p ( β , λ , Π ,   |   Y , Z ) is used to approximate the posterior g ( λ | β , Π 2 , ) p ( β , Π 2 , | Y , Z ) . The weight function, defined as the ratio of the posterior and the candidate generating density, becomes
ω ( β , λ , Π 2 , ) =   g ( λ | β , Π 2 , ) p ( β , Π 2 , | Y , Z ) p ( β , λ , Π 2 , | Y , Z ) = |   J ( Φ , ( Π 2 , β , λ ) ) | λ = 0 |   | J ( Φ , ( Π 2 , β , λ ) ) | g ( λ | β , Π 2 , ) | λ = 0 ,  
where the Jacobian matrix J ( Φ , ( Π 2 , β , λ ) ) as well as J ( Φ , ( Π 2 , β , λ ) ) | λ = 0 have been carefully derived in KVD9. Note that ω ( · ) = p ( · ) / r ( · ) , so (30) may be used in the “GS within M–H” algorithm to simplify (24).
Similar to the way we implemented the CP approach, it is more convenient to work with the precision matrix Ω−1 in the conditional densities. Applying the procedure outlined above, the steps involved in constructing the Markov chain for the posterior (22) are summarized as follows,
0.
Choose starting values ( Φ 0 , 1 , 0 )
1.
Draw 1 , i from p ( Ω 1 | Φ i 1 , Y , Z )
Draw Φ i from p ( Φ | Ω i 1 , Y , Z )
2.
Perform a singular value decomposition of Φ i = U i S i V i
3.
Compute β i , λ i , Π 2 i according to (18) and (19)
4.
Compute ω ( β i , λ i , π 1 i , Π 1 i , Π 2 i , Ω 1 , i ) according to (29) and (30)
5.
Draw ( π 1 i , Π 1 i ) from   p ( π 1 , Π 1 | Ω 1 , i , Φ i ( Π 2 i , β i , λ ) Y , Z ) | λ = 0
6.
Accept ( β i , π 1 i , Π 1 i , Π 2 i , Ω 1 , i ) as a drawing from the posterior with probability,
min ( ω ( β i , λ i , Π 2 i , Ω 1 , i ) ω ( β i 1 , λ i 1 , Π 2 i 1 , Ω 1 , ( i 1 ) )   ,   1 ) ,
otherwise, ( β i , λ i , Π 2 i , Ω 1 , i )   = ( β i 1 , λ i 1 , Π 2 i 1 , Ω 1 , ( i 1 ) ) .
7.
i = i + 1 . Go to 1.
Note that the conditional densities used in the first step are as follows:
p ( Σ 1   |   Φ ,   Y ,   Z ) | Ω 1 | ( T + k 2 m 1 ) ) / 2 exp [ 1 2 t r ( Ω 1 G ) ] ,
which follows a Wishart distribution W m ( T + k 2 , G 1 ) with ( T + k 2 ) degree of freedom, where G = Y Q z Y + ( Φ Φ ^ ) Z 2 M Z 1 Z 2 ( Φ Φ ^ ) , and Φ ^ = ( Z 2 M Z 1 Z 2 ) 1 Z 2 M Z 1 Y . In addition,
p ( Φ   |   Ω 1 ,   Y ,   Z ) | Ω 1 | k 2 / 2 exp [ 1 2 t r [ Ω 1 ( Φ Φ ^ ) Z 2 M Z 1 Z 2 ( Φ Φ ^ ) ] ] ,  
which is matric-variate normal density.
The conditional density used in step 5 is
p ( π 1 , Π 1 | Ω 1 , Φ ( Π 2 , β , λ ) , Y , Z ) | Ω 1 | k 1 / 2 exp [ 1 2 t r [ Ω 1 ( Λ Λ ^ ) Z 1 Z 1 ( Λ Λ ^ ) ] ] ,  
Evaluated at λ =   0 , where Λ   =   ( π 1   Π 1 ) , Λ ^ = ( Z 1 Z 1 ) 1 Z 1 ( Y     Z 2 Φ ) .

4.3. Convergence Diagnosis

One important implementation issue associated with MCMC methods is that of determining the number of iterations required. There are various informal or formal methods for the diagnosis of convergence, see Cowles and Carlin (1996) and Brooks and Roberts (1998) for comprehensive reviews and recommendations. Since the posterior densities in (14) and (22) resulting from CP and KVD do not have moments of any positive integer order, most of the methods proposed in the MCMC literature which require the existence of at least the first moment (posterior mean) are ruled out. We are left with a very few alternatives that can be used in our context.
First, the popular Raftery and Lewis (1992) method has been recognized as the best for estimating the convergence rate of the Markov chain if quantiles of the posterior density are of major interest, although the method does not provide any information as to the convergence rate of the chain as a whole. Because we are interested in the posterior modes and medians for β associated with the Bayesian approaches, we will largely rely on Raftery and Lewis’ method to determine the number of burn-ins and the subsequent number of iterations required to attain specified accuracy (e.g., estimating the 0.50 quantile in any posterior within ±0.05 with probability 0.95). However, we do not adopt their suggested skip-interval. MacEachern and Berliner (1994) showed that estimation quality is always degraded by discarding samples. We also experimented with using the skip-intervals and found that the results are basically the same if a sufficient number of iterations are run. This seems to be inefficient and sometimes infeasible in terms of computation time.
For each specification in our Monte Carlo study with repeated experiments, we determined the number of burn-ins and subsequent number of iterations by running the publicly available FORTRAN code gibbsit on MCMC output of 10,000 iterations from three or more testing replications. For KVD and CP approaches, the number of burn-ins for both the GS step and the M–H algorithm were estimated. It was found that the number of burn-ins in the GS step is negligible for most cases. However, we discarded more iterations as the transient phase than the estimated number of burn-ins.10 The estimated number of subsequent iterations across testing replications was stable for the Gibbs sampler (in both Geweke approach and the GS step for KVD and CP approaches), but it varied a lot for the M–H procedures, which is also demonstrated by the variation in acceptance rates over repeated experiments. We used a generous value for the number of subsequent iterations when feasible.
Second, for MCMC output from each testing replication, we also applied other convergence diagnostic methods, including percentiles derived from every quarter of the long chain, Yu and Mykland (1998)’s CUSUM plot, and Brooks (1996)’s D-sequence statistic. While the CUSUM partial sums actually involve averaging over sampling drawings, the computation of Brooks’ statistic is justified on the basis that it is designed to measure the frequency of back and forth movement in the MCMC algorithm. However, these diagnostics may sometimes provide contradictory outcomes so that one has to be extra careful in interpreting them before making a judgment on convergence.

5. Simulation Results and Discussions

In this section, we present results of Monte Carlo experiments and discuss some of the findings. As mentioned before, for the purpose of comparison, we also computed a number of single K-class estimators including OLS, 2SLS, LIML, and Fuller’s modified LIML. In summary, the set of K-class estimators for the structural coefficients in model (1) and (2) is given by:
( β ^ γ ^ ) = ( Y 2 Y 2 K 1 V ^ 2 V 2 ^ Y 2 Z 1 Z 1 Y 2 Z 1 Z 1 ) 1 ( ( Y 2 K 2 V 2 ^ ) y 1 Z 1 y 1 )
where V 2 ^ = Q Z Y 2 —see Equation (7) above.
The following LISEM estimators have been considered:
(1)
Ordinary least squares (OLS)
K 1 = K 2 = 0 .
(2)
Two stage least squares (2SLS)
K 1 = K 2 = 1 .
(3)
Zellner’s (1978) Bayesian minimum expected loss estimator (MELO)
K 1 = K 2 = 1 k / ( T k m 1 ) .
(4)
Zellner’s Bayesian method of moments relative to balanced loss function (BMOM)11
K 1 = 1 k / ( T k )   ,   K 2 = 1 ( 1 ω ) k / ( T k )   with   ω = 0.75
(5)
Classical LIML. We compute classical LIML as an iterated Aitken estimator (see Pagan (1979) and Gao and Lahiri (2000a)).
(6)
Fuller (1977) modified LIML estimators (Fuller1 and Fuller4)
K 1 = K 2 = λ * α / ( T k )   for   α = 1 ,   4
where
λ * = min β ( y 1 Y 2 β ) Q Z 1 ( y 1 Y 2 β ) ( y 1 Y 2 β ) Q Z ( y 1 Y 2 β )
and it is computed using the LIML estimate.
(7)
JIVE.
(8)
Posterior mode and median from the Geweke (1996) approach using Gibbs Sampling. The values of the hyperparameters are chosen to be τ 2 = 0.01 , v = m ( m + 1 ) / 2 , S = 0.01 I m .12
(9)
Mode and median of the marginal density of β based on classical LIML from Gibbs sampling (LIML-GS). LIML-GS is a byproduct of the “Gibbs within M–H” algorithm for the CP approach since the likelihood function is used as the candidate-generating density to explore the CP posterior.
(10)
Posterior mode and median from CP approach using “Gibbs within M–H” algorithm.
(11)
Posterior mode and median from KVD approach using “Gibbs within M–H” algorithm.
For the Bayesian approaches and LIML-GS, we report both (posterior) mode and median to show possible asymmetry in the marginal densities of β . Any preference for one over the other will depend on the researcher’s loss function. We obtain 16 estimates for each generated data set. The data are generated from the model,
y 1 = Y 2 β + u Y 2 = Z 2 π + V 2 ,  
where y1, Y2 are T × 1 such that m = 2, and Z 2 : T × k 2 . We further specify β =   1 and
Σ =   ( 1 ρ ρ 1 )
For | ρ | we used 0.20, 0.60, and 0.95.13 Z2 is simulated from a N ( 0 , I k 2 I T ) distribution and ( u , V 2 ) from a N ( 0 , Σ I T ) distribution. A constant term is added in each equation, i.e., Z1 is a T × 1 vector of 1 s.
The simulation results are reported in Table 1, Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11, Table 12 and Table 13. Table 1, Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11 and Table 12 are for cases with   ρ   >   0 , each table reporting results for one specification.
Table 13 summarizes the results for cases with ρ < 0 for BMOM and KVD for whom negative ρ made a surprising difference. As mentioned before, we focus on the estimates of the structural parameter β . Specifically, we analyze the sensitivity of the various estimates of β with respect to the strength of the instrumental variables Z, the degree of overidentification ( k 2 m + 1 ) , the degree of endogeneity ( ρ ) , and the sample size (T). Also, we will examine whether the performance of an estimator is symmetric with respect to the sign of parameter ρ , an issue generally overlooked in the literature.14
Note that the strength of the instrumental variables for the included endogenous variable Y2 is measured in terms of the adjusted R 2 by regressing Y2 on Z = (Z1, Z2). In the data generating process, we controlled R ¯ 2 to be within ±2.5% of the specified value to reduce unnecessary variation. We did not experiment with extremely small R ¯ 2 (say, 0.01 or less). In these cases, the mean values of all estimators approached the point of concentration ω 12 / ω 22 , which is equal to ( β + ρ ) for our data generating process (DGP).
For each specification, the number of replications is 400. The number of burn-ins (nburn_GS and nburn_MH), and subsequent number of iterations (n) determined at the convergence diagnosis step are reported in the footnotes to each table.
The average acceptance rate and its standard deviation (in parentheses) across replications for each M–H routine are reported as well. To evaluate alternative estimators, we computed mean, standard deviation (Std), root of mean squared errors (RMSE), and mean absolute deviation (MAD) over repeated experiments for all the estimators considered.15 Since LIML, posterior densities for CP and KVD, as well as 2SLS in the just-identified case do not have finite moments of positive order in finite samples, one should interpret the computed mean, standard deviation and RMSE across replications for these estimators with caution. In this sense, the MAD across replications is a preferred measure to consider.
We will first look at cases reported in Table 1, Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11 and Table 12 with ρ   >   0 . In Table 1, we consider a case (T = 50, ρ = 0.60, k2 = 4) with moderately strong instruments ( R ¯ 2 = 0.40 ). It is found that with reasonably strong instruments all estimators designed for simultaneous equations perform reasonably well. As expected, OLS is seriously biased. BMOM has a slight edge over others in terms of RMSE and MAD. For all Bayesian approaches and LIML-GS, the medians perform a little better than modes, and CP over KVD, in terms of bias, RMSE, and MAD. Notice that the classical LIML estimates are different from LIML-GS (mode or median). As noted by Drèze (1976), from a Bayesian viewpoint, LIML produces an estimate of β conditionally on the overidentifying restrictions, the modal values of all the remaining parameters, and a uniform prior. In other words, the concentrated likelihood function of β after concentrating out (i.e., maximizing with respect to) other reduced-form and nuisance parameters is a conditional density.
However, LIML-GS is a marginal density with all other parameters being integrated out. Due to possible asymmetry in the distribution of the nuisance parameters, the modal/median values of LIML-GS may not coincide with classical LIML estimates. In all our experiments, we find that the median-unbiasedness property of (conditional) LIML does not carry over to the marginal LIML (i.e., LIML-GS); however, the former generally has a much larger standard deviation than the latter. In a way, LIML-GS brings the classical LIML estimator close to its Bayesian counterpart for the purpose of comparison.
It is interesting to note that across all our tables, the difference between LIML-GS and CP can only be attributed to the importance of Jeffreys prior. Compared to LIML-GS, typically CP has a smaller bias, but slightly larger standard deviation, even though the differences are very small. In some cases, however, the use of Jeffreys prior reduces the bias in CP quite substantially. For example, in Table 4 with T = 50 and a high degree of overidentification, the bias is reduced from 0.36 to 0.25.
A simple case when the structural model is just identified (k2 = 1) is reported in Table 2. For this case it is well known that classical LIML coincides with 2SLS. The KVD approach does not accommodate the case of just-identification since (15) requires k2 > (m − 1).16 In this case, we find that CP-Mode produces results closer to LIML-GS-Mode than to LIML. CP (1998) showed that for a two-equation just-identified SEM in orthonormal canonical form, the posterior density of β with Jeffreys prior has precisely the same functional form as the density of the finite sample distribution of the corresponding LIML estimator as obtained by Mariano and McDonald (1979). Our simulation results show that the assumption of orthonormal canonical form is crucial for their exact correspondence, which cannot be extended to a general SEM.17 In general, the Bayesian marginal density is not the same as the classical conditional density. Interestingly, JIVE is considerably more biased and has larger standard deviation than 2SLS. Also, CP-Median and LIML-GS-Median perform significantly worse than their modes. This is because in an exactly identified model with weak instruments, the probability of local nonidentification is substantial, and the resulting nonstandard marginal density exhibits a very high variance. The same result holds true for Geweke-Median, but to a lesser extent. Thus, for exactly identified SEMs with very weak instruments, mode of the marginal density is a more dependable measure of β . We should point out that in all other cases in this study, the medians generally turned out to be more preferable than the modes in terms of bias, RMSE, and MAD (see Table 11 and Table 12, for instance).
Results reported in Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11 and Table 12 consider cases with general overidentification and weak instruments. As noted in the literature, OLS and 2SLS are median-biased in the direction of the correlation coefficient ρ , and the bias in 2SLS grows with the degree of over identification, and decreases as sample size increases. Results in Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9 and Table 10 confirm these results. Since MELO is a single K-class estimator with 0 < K < 1, its performance is always between OLS and 2SLS estimates. The bias in MELO shows the same pattern as that of 2SLS. With moderate simultaneity, the median-bias in 2SLS can be as large as about 40% of the true value (see Table 8). We note that MELO, LIML-GS-Mode, and KVD-Mode or KVD-Median are also median-biased in the direction of ρ. However, the bias in JIVE is consistently in the opposite direction of ρ . Classical LIML is remarkably median-unbiased when the instrumental variables are not very weak, which is well documented in the literature. We find that LIML is median-biased in the direction of ρ when the instruments are very weak (Table 8), which is consistent with the finding in Staiger and Stock (1997) using local-to-zero asymptotic theory. Even in this situation, the bias of LIML is much smaller than that of any other estimator, except BMOM.
The MAD of OLS is very close to its bias (i.e., relatively small Std) across all cases and it implies that OLS method is robust in the sense that it does not suffer from heavy tails or outlying estimates, see Zellner (1998). In this sense, MELO and BMOM are all robust with relatively small standard deviations across replications. However, OLS exhibits large bias in the presence of simultaneity and is not so appealing. It is known that for a degree of overidentification strictly less than seven, 2SLS would have a smaller asymptotic mean squared error (AMSE) than LIML, cf. Mariano and McDonald (1979) and references therein. In cases with weak instruments the situation gets more complicated in finite samples. In our experiments, LIML has larger RMSE and MAD than 2SLS except in Table 11 and Table 12 where ρ was 0.95. Note that the degree of overidentification is 8.0 in Table 4, Table 6, Table 8 and Table 10.
Among classical estimators, JIVE turns out to be least appealing. Monte Carlo simulations in Angrist et al. (1999) showed that JIVE has slight median bias in the opposite direction of ρ (but less than 2SLS) and have heavier tails than LIML. Our Table 6 is comparable to panel 2 of their Table 1, and the results are similar. Our other experiments show that JIVE may also have large absolute bias (larger than LIML) in the case with weak instruments, sometimes even greater than 2SLS (see Table 2). Generally, JIVE has slightly less bias than 2SLS, but this gain is overshadowed by enlarged standard deviation such that in finite samples it has no advantage over 2SLS in terms of MAD and RMSE. We also find that JIVE has greater RMSE and MAD than LIML. Blomquist and Dahlberg (1999) experimented with much larger sample sizes than ours. Comparing our Table 4 with Table 6 and with an unreported simulation with a sample size of 500, we found that the relative gain in JIVE is more than other estimators as sample size increases, even though its relative low standing remains valid. Examined from different angles, these results are very similar to those reported by Davidson and MacKinnon (2006a, 2006b).18
Fuller’s modified LIML estimators are included because Fuller1 is designed to minimize the median-bias, and Fuller4 to minimize the mean-squared error. It seems that this conclusion is also problematic in the presence of weak instruments. Between the two, Fuller1 has smaller median-bias, and Fuller4 has smaller standard deviations across replications. However, in terms of RMSE or MAD, Fuller4 shows no advantage over Fuller1 in most of the cases.
Because all the estimators except OLS are consistent and their asymptotic distributions are also the same, results in Table 3, Table 4, Table 5 and Table 6 confirm that their bias and dispersion decrease as sample size increases. But if the instruments are very weak (see Table 7 and Table 8), their bias and dispersion may remain significant, a point emphasized forcefully by Zellner (1998). However, when the endogeneity is not strong (see Table 9 and Table 10), their bias and dispersion may not be a big concern for some of the estimators.
Across all cases, we find that the bias in BMOM is small if ρ is not too small and the structural Equation (1) is overidentified. As sample size increases or degree of over-identification rises, the observed bias in BMOM decreases. The most striking feature of BMOM is that it exhibits the smallest MAD and Std when ρ is not too small. MELO shows slightly smaller MAD and Std than BMOM if ρ is small (see Table 9 and Table 10). In cases with very weak instruments and a high degree of overidentification, the MAD of BMOM is only one-fourth of that of other estimators (see Table 8). These are in accordance with Tsurumi (1990)’s finding that in many cases, ZEM has the least relative mean absolute deviation. Meanwhile, if ρ is very small and the structural equation is overidentified, the bias in BMOM can be large; 2SLS, LIML-GS, Geweke, and CP perform remarkably well in these situations.
Next, we examine in more detail the performance of the Bayesian approaches. Overall, the median bias resulting from these approaches exhibits the same pattern as the bias of 2SLS, it increases with the degree of overidentification, and decreases as sample size rises. The Geweke (1996) approach used a shrinkage prior but its performance is comparable with LIML-GS and CP. The median-bias from PMOD-Geweke is the same or slightly less than that of LIML-GS-Mode, and the bias from Geweke-Median is always slightly less than that of LIML-GS-Median. Similar relationships are observed for MADs. These reflect the impact of the (informative) shrinkage prior on the posterior density.
For each specification, the acceptance rate in the M–H algorithm using CP approach is stable while that using KVD approach shows huge variation across replications. The acceptance rate for CP is generally above 40%, except when sample size is small and the degree of overidentification is high. This shows that the posterior of CP is largely dominated by the likelihood function (3) and the Jeffreys prior generally carries little information. Second, in terms of the computed standard deviations (Stds) of the estimates across replications, CP-Mode has larger dispersion than LIML-GS-Mode, and CP-Median has larger dispersion than LIML-GS-Median. These also shed light on the notion that Jeffreys prior is less informative than a uniform prior. However, between the Jeffreys prior (13) used by CP and the implied prior (21) resulting from diffuse/Jeffreys prior on a linear model used by KVD, it is not clear which one is less informative.
As for the KVD (1998) approach, we observe that it performs as well as any other estimator if the instruments are not weak (see Table 1). But when the instruments are weak, and ρ is positive, KVD shows more bias and higher MAD than those from CP. In Table 4 with T = 50 and high degree of overidentification, KVD performs as bad as OLS.
Next, we consider cases with negative ρ , and the results are summarized in Table 13. We replicate each case in Table 1, Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11 and Table 12 with the same specification except ρ being negative. Since the performance of all estimators except BMOM and KVD were basically the same with respect to the sign of ρ , we only report results on these two in Table 13. We find that when ρ changes sign, the bias of BMOM does not change sign and even increases in magnitude. Also note that the computed Stds for BMOM when ρ   <   0 are close to the respective ones when ρ   >   0 . Therefore, for cases with ρ   <   0 , BMOM has large RMSEs/MADs and loses its attraction. Note that BMOM is the same as the double K-class estimator (DKC) with K values fixed. This asymmetry in the performance in DKC is not well recognized in the literature, and has been discussed in Gao and Lahiri (2001). The observed asymmetry in its bias with respect to ρ in our experiments is readily explained by examining an expression for the mean of double K-class estimator (DKC) in (Dwivedi and Srivastava (1984, Theorem (1)). We can express δ ^ D K C as:
δ ^ D K C = δ ^ K 1 +   ( Y 2 Y 2 K 1 V ^ 2 V 2 ^ Y 2 Z 1 Z 1 Y 2 Z 1 Z 1 ) 1 ( ( K 1 K 2 ) V ^ 2 y 1 0 ) ,
where δ ^ K 1 is a single K-class estimator with characterizing scalar K 1 . When Z 1 Z 2 = 0 , which is satisfied in our experimental specifications, a double K-class estimator of β may be written as
β ^ D K C = β ^ K 1 + ( K 1 K 2 ) Y 2 Q Z y 1 Y 2 Δ Y 2
where Δ = ( 1 K 1 ) Q z 1 + K 1 P Z 2 . Observe that for 0 < K 1 < 1 ,   β   ^ K 1 is biased in the direction of ρ , as noted in Mariano and McDonald (1979). Note also that Y 2 Δ Y 2 > 0 , and Y 2 Q Z y 1 provides an estimate of ω 12 . Although Dwivedi and Srivastava (1984) explored the dominance of double K-class over K-class using the exact MSE criterion, their guidelines for the selection of K2 for a given K1 are not entirely valid, because the conditions were derived from a small Monte Carlo simulation with positive ω 12 and negative ρ only. Since K1 < K2 for BMOM, when ρ   and ω 12 have the opposite sign, the second term in β ^ D K C will be of the same sign as the bias of β ^ K 1 , therefore β ^ D K C (hence BMOM) will exhibit large bias. Otherwise, when ρ . ω 12 >   0 , the bias is mitigated. Based on our simulation results, we found that the sign of ρ has no effect on the standard deviation of BMOM. This finding shows that the greater RMSE of BMOM when ρ ω 12 < 0 is due to the aggravated bias. For the specification corresponding to Table 4 in Table 13 (i.e., T = 50, ρ = −0.60, K2 = 4, R ¯ 2 = 0.10), we find that for given K1 = 0.947, RMSE is minimized if K2 is chosen to be 0.829, which is much less than K1, and less than K2 = 0.987 used in BMOM. See Gao and Lahiri (2001) for further details.
In Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11 and Table 12 we found that KVD with ρ > 0 performs very poorly, often with substantial bias and relatively high RMSE and MAD. CP uniformly dominates KVD in these cases. However, with ρ < 0 the picture turns around remarkably well in favor of KVD. As we see in Table 13, across all cases the bias tends to be negative and relatively small. With other parameter values being the same, KVD with ρ < 0 has significantly less RMSE and MAD than cases when ρ > 0, and performs unequivocally the best among all estimators when endogeneity is strong. However, since this observed asymmetry is essentially a finite sample problem with KVD, the improved performance when ρ < 0 becomes less significant when the sample size increases from 50 to 100. With ρ < 0 the overall performance of KVD is very comparable to that of CP, if not slightly better in some cases.
After experimenting with widely different negative and positive values of β and ρ , we found out that the performance of KVD is dependent on the sign of β ρ , rather than on the sign of ρ . When β ρ > 0, it performs very unsatisfactorily as documented in Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11 and Table 12. Kleibergen and Zivot (2003) have derived exact analytical expressions for the conditional densities of β given Ω for both the KVD and CP posteriors. They show that the difference between the two is in the Jacobian relating the unrestricted linear multivariate model to the restricted reduced form model. We expect that this additional term may account for the asymmetry in KVD with respect to β ρ . In our experiments, we found that in finite samples, when β ρ > 0, the reduced rank restriction using singular value decomposition shifts the marginal posterior for KVD away from the marginal posterior of the linear multivariate model. However, when the sample size gets large, the problem seems to go away.

6. Conclusions

This paper examines the relative merits of some contemporary developments in the Bayesian and classical analysis of limited information simultaneous equations models in situations where the instruments are very weak. Since the posterior densities and their conditionals in the Bayesian approaches developed by Chao and Phillips (1998) and Kleibergen and van Dijk (1998) are nonstandard, we proposed and implemented a “Gibbs within Metropolis–Hastings” algorithm, which only requires the availability of the conditional densities from the candidate-generating density. These conditional densities are used in a Gibbs sampler (GS) to simulate the candidate generating density, whose drawings, after convergence, are then weighted to generate drawings from the target density in a Metropolis–Hastings (M–H) algorithm. We rely on Raftery and Lewis (1992) to determine the number of burn-ins, and the subsequent number of required iterations in order to ensure convergence. Through a MCMC simulation study, our results provide useful guidelines for empirical practitioners.
The first comforting result is that with reasonably strong instruments (marginal R ¯ 2 in excess of 0.40), all estimators perform equally well in finite samples. In cases with very weak instruments (marginal R ¯ 2 less than 0.10), there is no single estimator that is superior to others in all cases—a conclusion also reached by Andrews and Stock (2005). When endogeneity is weak ( ρ less than 0.20), Zellner’s MELO does the best. When the endogeneity is relatively strong (ρ in excess of 0.60) and ρ ω 12 > 0, BMOM outperforms all other estimators by wide margins. When the endogeneity is strong but β ρ < 0, the KVD approach seems to get very appealing; but, otherwise, its performance is surprisingly poor. With β ρ > 0, as the sample size gets larger, the performance of KVD improves rapidly. Fortunately, the Geweke and CP approaches exhibit no such asymmetry and their performances based on bias, RMSE, and MAD are very similar. Based on the medians of marginal posteriors, their performance ranking is consistently a distant second. The record of JIVE is quite disappointing across all our experiments and is not recommended in practice. Even though JIVE is slightly less biased than 2SLS in most cases, its standard deviation is considerably higher, particularly in small samples. The most remarkable result in this study is that poor instruments can affect the performance of different estimators differently, depending on the signs and magnitudes of certain key parameters of the model. Given the finding that even in finite samples with very weak instruments BMOM and KVD perform remarkably well on certain parts of the parameter space, more research is needed to understand the reasons for the asymmetry and find ways to fix the problem. Another important caveat of our comparative study is that it was done in an iid setting. Heteroskedastic and autocorrelated errors, particularly in highly leveraged regressions, can affect inferences based on alternative instrument variable regressions differentially relative to ordinary least squares, see Young (2019). These issues remain unresolved.

Author Contributions

The two authors contributed equally to this work.

Funding

This research received no external funding.

Acknowledgments

An earlier version of this paper was presented at the 1999 Joint Statistical Meetings and the 2000 Winter Meetings of the Econometric Society. We are grateful to Late Arnold Zellner for his early encouragement with this work, and to Ingolf Dittmann, Yuichi Kitamura, Roberto Mariano, Eric Zivot, Herman van Dijk and two anonymous referees for many helpful comments and suggestions. Peter Dang re-typed the whole manuscript. However, the responsibility for any remaining errors and shortcomings is solely ours.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Proof that |J(Φ, (Π2, β, λ))| ≥ |J(Φ, (Π2, β, λ))|λ=0 is invalid, cf. footnote 9.
In current notation, we need to show that |J(Φ, (Π21, θ2, β, λ))| ≥ |J(Φ, (Π21, θ2, β, λ))|λ=0| fails, where θ 2 = Π 22 Π 21 1 . The fact is that J(Φ, (Π21, θ2, β, λ))|λ=0 and W are not orthogonal, where W = J(Φ, (Π21, θ2, β, λ))−J(Φ, (Π21, θ2, β, λ))|λ=0.
Consider a simple case with m = k2 = 2. In this case,
Φ = ( 1 θ 2 ) Π 21 ( β 1 ) + ( θ 2 1 ) ( 1 + θ 2 2 ) 1 / 2 λ ( 1 + β 2 ) 1 / 2 ( 1 β ) .
Denote K = ( 1 + θ 2 2 ) 1 / 2 ( 1 + β 2 ) 1 / 2 . Therefore,
J ( Φ ,   ( Π 21 ,   θ 2 ,   β ,   λ ) ) | λ = 0 = ( β 0 Π 21 K ( θ 2 ) β θ 2 β Π 21 θ 2 Π 21 K ( 1 ) 1 0 0 K ( β θ 2 ) θ 2 Π 21 0 K ( β ) ) ,
W = ( 0 λ K ( 1 + θ 2 ) 1 ( 1 ) λ K ( 1 + β 2 ) 1 ( β θ 2 ) 0 0 λ K ( 1 + θ 2 ) 1 ( θ 2 ) λ K ( 1 + β 2 ) 1 ( β ) 0 0 λ K ( 1 + θ 2 ) 1 ( β ) λ K ( 1 + β 2 ) 1 ( θ 2 ) 0 0 λ K ( 1 + θ 2 ) 1 ( β θ 2 ) λ K ( 1 + β 2 ) 1 ( 1 ) 0 ) .
It is easy to check that |J(Φ, (Π21, θ2, β, λ))|λ=0|W′ is not a zero matrix but with its third row being 0 s. Interestingly,
( J ( Φ ,   ( Π 21 ,   θ 2 ,   β ,   λ ) ) | λ = 0 ) W = ( 0 0 0 0 0 0 λ K ( Π 21 ) 0 0 λ K ( Π 21 ) 0 0 0 0 0 0 ) .

References

  1. Ackerberg, Daniel A., and Paul J. Devereux. 2006. Comment on case against JIVE. Journal of Applied Econometrics 21: 835–38. [Google Scholar] [CrossRef]
  2. Andrews, Donald W. K., and James H. Stock. 2005. Inference with weak instruments. In Advances in Economics and Econometrics, Theory and Applications. Edited by Richard Blundell, Whitney K. Newey and Torsen Persson. Ninth World Congress of the Econometric Society. Cambridge: Cambridge University Press, vol. III. [Google Scholar]
  3. Andrews, Donald W. K., James H. Stock, and Liang Sun. 2019. Weak Instruments in IV Regression: Theory and Practice. Annual Review of Economics. forthcoming. [Google Scholar]
  4. Angrist, Joshua D., Guido W. Imbens, and Alan Krueger. 1999. Jackknife instrumental variables estimation. Journal of Applied Econometrics 14: 57–67. [Google Scholar] [CrossRef]
  5. Billingsley, Patrik. 1986. Probability and Measure. New York: Wiley. [Google Scholar]
  6. Blomquist, Sören, and Matz Dahlberg. 1999. Small sample properties of LIML and jackknife IV estimators: Experiments with weak instruments. Journal of Applied Econometrics 14: 69–88. [Google Scholar] [CrossRef]
  7. Blomquist, Sören, and Matz Dahlberg. 2006. The case against Jive: A comment. Journal of Applied Econometrics 21: 839–41. [Google Scholar] [CrossRef]
  8. Bound, John, David A. Jaeger, and Regina M. Baker. 1995. Problems with instrumental variables estimation when the correlation between the instruments and the endogenous explanatory variable is weak. Journal of the American Statistical Association 90: 443–50. [Google Scholar] [CrossRef]
  9. Brooks, Stephen P. 1996. Quantitative Convergence Diagnosis for MCMC via CUSUMS. Technical Report. Bristol: University of Bristol. [Google Scholar]
  10. Brooks, Stephen P., and Gareth O. Roberts. 1998. Assessing convergence of Markov chain Monte Carlo algorithms. Statistics and Computing 8: 319–35. [Google Scholar] [CrossRef]
  11. Buse, Adolf. 1992. The bias of instrumental variable estimators. Econometrica 60: 173–80. [Google Scholar] [CrossRef]
  12. Casella, George, and Edward I. George. 1992. Explaining the Gibbs sampler. The American Statistician 46: 167–74. [Google Scholar]
  13. Chao, John C., and Peter C. B. Phillips. 1998. Posterior distributions in limited information analysis of the simultaneous equations model using the Jeffreys prior. Journal of Econometrics 87: 49–86. [Google Scholar] [CrossRef]
  14. Chib, Siddhartha, and Edward Greenberg. 1995. Understanding the Metropolis–Hastings algorithm. The American Statistician 49: 327–35. [Google Scholar]
  15. Chib, Siddhartha, and Edward Greenberg. 1996. Markov chain Monte Carlo simulation methods in econometrics. Econometric Theory 12: 409–31. [Google Scholar] [CrossRef]
  16. Conley, Timothy G., Christian B. Hansen, Robert McCulloch, and Peter E. Rossi. 2008. A semi-parametric Bayesian approach to the instrumental variable problem. Journal of Econometrics 144: 276–305. [Google Scholar] [CrossRef]
  17. Cowles, Mary K., and Bradley P. Carlin. 1996. Markov chain Monte Carlo convergence diagnosis: A comparative review. Journal of the American Statistical Association 91: 883–904. [Google Scholar] [CrossRef]
  18. Davidson, Russell, and James G. MacKinnon. 2006a. The case against JIVE. Journal of Applied Econometrics 21: 827–33. [Google Scholar] [CrossRef]
  19. Davidson, Russell, and James G. MacKinnon. 2006b. The case against JIVE: Reply. Journal of Applied Econometrics 21: 843–44. [Google Scholar] [CrossRef]
  20. Drèze, Jacques H. 1976. Bayesian limited information analysis of the simultaneous equations model. Econometrica 44: 1045–75. [Google Scholar] [CrossRef]
  21. Drèze, Jacques H., and Juan-Antonio A. Morales. 1976. Bayesian full information analysis of simultaneous equations. Journal of American Statistical Association 71: 329–54. [Google Scholar]
  22. Drèze, Jacques H., and Jean François Richard. 1983. Bayesian analysis of simultaneous equation systems. In Handbook of Econometrics. Edited by Zvi Griliches and Michael Intriligator. Amsterdam: North Holland. [Google Scholar]
  23. Dwivedi, Tryambakeshwar D., and Virendra K. Srivastava. 1984. Exact finite sample properties of double k-class estimators in simultaneous equations. Journal of Econometrics 25: 263–83. [Google Scholar] [CrossRef]
  24. Fuller, Wayne. A. 1977. Some properties of a modification of the limited information estimator. Econometrica 45: 939–53. [Google Scholar] [CrossRef]
  25. Gao, Chuanming, and Kajal Lahiri. 2000a. Further consequences of viewing LIML as an iterated Aitken estimator. Journal of Econometrics 98: 187–202. [Google Scholar] [CrossRef]
  26. Gao, Chuanming, and Kajal Lahiri. 2000b. MCMC algorithms for two recent Bayesian limited information estimators. Economics Letters 66: 121–26. [Google Scholar] [CrossRef]
  27. Gao, Chuanming, and Kajal Lahiri. 2001. A Note on the double k-class estimator in simultaneous equations. Journal of Econometrics 108: 101–11. [Google Scholar] [CrossRef]
  28. Geweke, John. 1996. Bayesian reduced rank regression in econometrics. Journal of Econometrics 75: 121–46. [Google Scholar] [CrossRef]
  29. Hastings, Willeen K. 1970. Monte Carlo sampling methods using Markov chains and their applications. Biometrica 57: 97–109. [Google Scholar] [CrossRef]
  30. Kleibergen, Frank. 1997. Equality Restricted Random Variables: Densities and Sampling Algorithms. Econometric Institute Report 9662/A. Rotterdam: Erasmus University Rotterdam. [Google Scholar]
  31. Kleibergen, Frank. 1998. Conditional Densities in Econometrics. Econometric Institute Research Papers EI 9853, Erasmus School of Economics (ESE), Discussion Paper. Rotterdam: Erasmus University Rotterdam. [Google Scholar]
  32. Kleibergen, Frank, and Herman K. van Dijk. 1998. Bayesian simultaneous equation analysis using reduced rank structures. Econometric Theory 14: 701–43. [Google Scholar] [CrossRef]
  33. Kleibergen, Frank, and Eric Zivot. 2003. Bayesian and classical approaches to in-strumental variable regression. Journal of Econometrics 114: 29–72. [Google Scholar] [CrossRef]
  34. MacEachern, Steven N., and L. Mark Berliner. 1994. Subsampling the Gibbs sampler. The American Statistician 48: 188–90. [Google Scholar]
  35. Maddala, Gangadharrao S. 1976. Weak priors and sharp posteriors in simultaneous equation models. Econometrica 44: 345–51. [Google Scholar] [CrossRef]
  36. Maddala, Gangadharrao S., and Jinook Jeong. 1992. On the exact small sample distribution of the instrumental variable estimator. Econometrica 60: 181–83. [Google Scholar] [CrossRef]
  37. Mariano, Roberto S., and James B. McDonald. 1979. A note on the distribution functions of LIML and 2SLS structural coefficient in exactly identified case. Journal of the American Statistical Association 74: 847–48. [Google Scholar] [CrossRef]
  38. Metropolis, Nicholas, Arianna. W. Rosenbluth, Marshall N. Rosenbluth, Augusta H. Teller, and Edward Teller. 1953. Equations of state calculations by fast computing machines. Journal of Chemical Physics 21: 1087–92. [Google Scholar] [CrossRef]
  39. Ni, Shawn, and Dongchu Sun. 2003. Noninformative priors and frequentist risks of Bayesian estimators of vector-autoregressive models. Journal of Econometrics 115: 159–97. [Google Scholar] [CrossRef]
  40. Ni, Shawn, Dongchu Sun, and Xiaogian Sun. 2007. Intrinsic Bayesian estimation of vector autoregression impulse responses. Journal of Business & Economic Statistics 25: 163–76. [Google Scholar]
  41. Pagan, Adrian R. 1979. Some consequences of viewing LIML as an iterated Aitken estimator. Economics Letters 3: 269–372. [Google Scholar] [CrossRef]
  42. Percy, David F. 1992. Prediction for seemingly unrelated regressions. Journal of the Royal Statistical Society B 54: 243–52. [Google Scholar] [CrossRef]
  43. Poirier, Dale J. 1995. Intermediate Statistics and Econometrics. Cambridge: MIT Press. [Google Scholar]
  44. Poirier, Dale J. 1996. Prior beliefs about fit. In Bayesian Statistics 5: Proceedings of the Fifth Valencia International Meeting. Edited by Jose M. Bernardo, James O. Berger, A. Philip Dawid and Adrian F. M. Smith. Oxford: Clarendon Press. [Google Scholar]
  45. Radchenko, Stanislav, and Hiroki Tsurumi. 2006. Limited information Bayesian analysis of a simultaneous equation with an autocorrelated error term and its application to the U.S. gasoline market. Journal of Econometrics 133: 31–49. [Google Scholar] [CrossRef]
  46. Raftery, Adrian E., and Stephen M. Lewis. 1992. How many iterations in the Gibbs sampler? In Bayesian Statistics 4. Proceedings of the Fourth Valencia International Meeting. Edited by Jose M. Bernardo, Adrian F. M. Smith, A. Philip Dawid and James O. Berger. Oxford: Oxford University Press. [Google Scholar]
  47. Smith, Adrian F. M., and Gareth O. Roberts. 1993. Bayesian computation via the Gibbs sampler and related Markov chain Monte Carlo methods. Journal of the Royal Statistical Society B 55: 3–23. [Google Scholar] [CrossRef]
  48. Staiger, Douglas, and James H. Stock. 1997. Instrumental variables regression with weak instruments. Econometrica 65: 557–86. [Google Scholar] [CrossRef]
  49. Tierney, Luke. 1994. Markov chains for exploring posterior distributions. Annals of Statistics 22: 1701–67. [Google Scholar] [CrossRef]
  50. Tsurumi, Hiroki. 1990. Comparing Bayesian and non-Bayesian limited information estimators. In Bayesian and Likelihood Methods in Statistics and Econometrics. Edited by Seymour Geisser, James S. Hodges, S. James Press and Arnold Zellner. Amsterdam: North-Holland. [Google Scholar]
  51. Young, Alwyn. 2019. Consistency without Inference: Instrumental Variables in Practical Application. London: London School of Economics. [Google Scholar]
  52. Yu, Bin, and Per Mykland. 1998. Looking at Markov samplers through Cusum path plots: A simple diagnostic idea. Statistics and Computing 8: 275–86. [Google Scholar] [CrossRef]
  53. Zellner, Arnold. 1971. An Introduction to Bayesian Inference in Econometrics. New York: Wiley. [Google Scholar]
  54. Zellner, Arnold. 1978. Estimation of functions of population means and regression coefficients: A minimum expected loss (MELO) approach. Journal of Econometrics 8: 127–58. [Google Scholar] [CrossRef]
  55. Zellner, Arnold. 1986. Further results on Bayesian minimum expected loss (MELO) estimates and posterior distributions for structural coefficients. In Advances in Econometrics. Edited by Daniel L. Slottje. Amsterdam: Elsevier, vol. 5, pp. 171–82. [Google Scholar]
  56. Zellner, Arnold. 1994. Bayesian and Non-Bayesian estimation using balanced loss functions. In Statistical Decision Theory and Related Topics. Edited by Shanti S. Gupta and James O. Berger. New York: Springer, vol. V, chp. 28. pp. 377–90. [Google Scholar]
  57. Zellner, Arnold. 1998. The finite sample properties of simultaneous equations’ estimates and estimators: Bayesian and non-Bayesian approaches. Journal of Econometrics 83: 185–212. [Google Scholar] [CrossRef]
  58. Zellner, Arnold, Luc Bauwens, and Harman K. van Dijk. 1988. Bayesian specification analysis and estimation of simultaneous equation models using Monte Carlo methods. Journal of Econometrics 38: 39–72. [Google Scholar] [CrossRef] [Green Version]
  59. Zellner, Arnold, Tomohiro Ando, Nalan Baştürk, Lennart Hoogerheide, and Herman K. van Dijk. 2014. Bayesian analysis of instrumental variable models: Acceptance-rejection within direct Monte Carlo. Econometric Reviews 33: 3–35. [Google Scholar] [CrossRef]
1
Zellner (1998) and Zellner et al. (2014) contain a comprehensive review of the finite sample properties of SEM estimators, and emphasize the need for finite sample optimal estimation procedure for such models. Andrews and Stock (2005) reviews recent developments in methods that deal with weak in IV regression models, and presents new testing results under “many weak-IV asymptotics”.
2
There has been a lot of interest in the estimation of LISEM with weak instruments. See Buse (1992); Bound et al. (1995); Staiger and Stock (1997); Angrist et al. (1999); Blomquist and Dahlberg (1999), among others. More recently, Andrews et al. (2019) review the literature on weak instruments in linear IV regression, and suggest that weak instruments remain an important issue in empirical practice.
3
Geweke (1996) considered a more general specification. To facilitate comparison, for Geweke approach only, we have denoted Y = (Y2, y1).
4
The expressions for the conditional densities of Π2 and β given in (Geweke 1996, expressions (11) and (13)) contain some typographical errors and are corrected here in (11) and (12).
5
Note that this formulation or the singular value decomposition does not change the identification status of the LISEM specified by (1) and (2). If r a n k ( Π 2 ) < ( m 1 ) ,   β is locally nonidentified.
6
This is the prior suggested in Drèze (1976). Zellner (1971) and Zellner et al. (1988) used a similar prior with −(m + 1)/2 in the exponent.
7
Zellner et al. (2014) suggested a variant of this approach called Acceptance-Rejection within Direct Monte Carlo (ARDMC) to evaluate the posterior density, and report substantial gain in computational efficiency, particularly with weak instruments. They also studied the existence conditions for posterior moments of the parameters of interest in terms of the number of available instruments being greater than the number of endogenous variables plus the order of the moment.
8
Gao and Lahiri (2000b) illustrated the algorithm empirically with a simple labor supply model.
9
See also Kleibergen (1997, 1998). Note that their claimed relationship that |J(Φ, (Π2, β, λ))| ≥ |J(Φ, (Π2, β, λ))|λ=0 is analytically incorrect; see the Appendix A for proof.
10
In practice, there is often a concern about possible underestimation of true length of the burn-in period using the Raftery and Lewis method if the quantile of interest is not properly pre-prescribed, see Brooks and Roberts (1998).
11
Tsurumi (1990) used ω = 0.75 for Zellner’s extended MELO (ZEM) in his experiments. BMOM and ZEF are almost identical in our context.
12
We found that the median-bias and dispersion of the posterior density of β from the Geweke (1996) approach increase as τ 2 gets larger. Although one might suspect that the convergence the Gibbs sampler could be slow with smaller values of τ 2 , our convergence diagnostics did confirm this concern.
13
We do not report cases with |ρ| = 0.99 or 1. As pointed out by Maddala and Jeong (1992), when the instruments are weak and |ρ| is very close to one, the exact finite sample distribution of IV estimator is bimodal. Our experiments show that the marginal posterior density of β from the Bayesian approaches exhibits a similar pattern.
14
Denote Ω = [ ω 11 ω 12 ω 12 ω 22 ] . Using Σ = C Ω C , we have σ 12 = ω 11 2 β ω 12 + β 2 ω 22 , σ 12 = ω 12 β ω 22 , and σ 22 = ω 22 . Letting ρ = σ 12 / σ 11 σ 22 , the second relationship may be rewritten as:
β ω 12 ω 22 = ρ σ 11 ω 22
If Σ is normalized as in (35) with σ 11 = ω 22 = 1 , then ω12 = β + ρ. Therefore, in our context, given β = 1, the sign and magnitude of ρ (or ω12) has a special significance.
15
Medians were also calculated. Since they were very close to the corresponding means in all our experiments, we did not report them in this paper.
16
When k2 = (m − 1), a diffuse prior in (20) for the linear model implies that the prior for the parameters of the LISEM (4) is
p ( β , π 1 , Π 1 , Π 2 , ) | | ( k + m + 1 ) / 2 | Π 2 | ,
and the prior for the parameters of the LISEM (1) and (2) is
p ( β , γ , Π 1 , Π 2 , Σ ) | Σ | ( k + m + 1 ) / 2 | Π 2 |
which is identical to the Jeffreys prior; see also expressions (22) and (42) in CP.
17
Note that the relationship between the standardized parameter vector and the original parameter vector involves the nuisance parameters, cf. Chao and Phillips (1998). However, when a SEM is in orthonormal canonical form (i.e., the exogenous regressors are orthonormal and the disturbance covariance matrix Ω is an identity matrix), both the density of random parameter β from the CP approach and the probability density of the classical LIML estimator for β are conditional on these information.
18
Ackerberg and Devereux (2006) and Blomquist and Dahlberg (2006) have suggested some ad hoc adjustments to the original JIVE formula to improve its performance.
Table 1. T = 50, ρ = 0.60, k2 = 4, R2 = 0.40.
Table 1. T = 50, ρ = 0.60, k2 = 4, R2 = 0.40.
MeanStdRMSEMAD
OLS1.3480.0890.3590.348
2SLS1.0450.1440.1510.121
MELO1.1150.1260.1710.144
BMOM0.9670.1270.1310.102
LIML0.9980.1520.1520.118
Fuller11.0150.1470.1480.116
Fuller41.0610.1360.1490.120
JIVE0.9570.1780.1830.141
Geweke_Mode1.0560.1400.1510.122
Geweke_Median1.0310.1430.1460.116
LIML_GS_Mode1.0610.1390.1520.123
LIML_GS_Median1.0360.1420.1460.116
CP_Mode1.0460.1440.1510.121
CP_Median1.0210.1450.1470.115
KVD_Mode1.0900.1480.1730.143
KVD_Median1.0790.1370.1580.130
Notes: Number of replications: 400. Geweke: nburn = 100, n = 2000. CP: nburn_GS = 100, nburn_MH = 100, n = 5000, acceptance rate = 0.482 (0.015). KVD: nburn_GS = 100, nburn_MH = 100, n = 4000, acceptance rate = 0.215 (0.136).
Table 2. T = 50, ρ = 0.60, k2 = 1, R2 = 0.10.
Table 2. T = 50, ρ = 0.60, k2 = 1, R2 = 0.10.
MeanStdRMSEMAD
OLS1.5370.1110.5480.537
2SLS1.0300.3450.3460.267
MELO1.1730.2620.3140.248
BMOM0.8810.2640.2900.229
LIML1.0300.3450.3460.267
Fuller11.1070.3000.3190.245
Fuller41.2500.2190.3320.277
JIVE0.8030.4910.5290.409
Geweke_Mode1.0890.3310.3430.265
Geweke_Median0.9070.5180.5260.358
LIML_GS_Mode1.0910.3130.3260.255
LIML_GS_Median0.7781.3861.4040.592
CP_Mode1.1080.3090.3270.256
CP_Median0.7971.3831.3980.580
KVD_Moden.a.n.a.n.a.n.a.
KVD_Mediann.a.n.a.n.a.n.a.
Notes: Number of replications: 400. Geweke: nburn = 100, n = 3000. CP: nburn_GS = 200, nburn_MH = 200, n = 10,000, acceptance rate = 0.551 (0.023).
Table 3. T = 50, ρ = 0.60, k2 = 4, R2 = 0.10.
Table 3. T = 50, ρ = 0.60, k2 = 4, R2 = 0.10.
MeanStdRMSEMAD
OLS1.5390.1110.5500.539
2SLS1.2310.2790.3620.296
MELO1.3660.1860.4110.368
BMOM0.9430.1840.1930.154
LIML1.0430.5790.5810.386
Fuller11.1430.3670.3940.307
Fuller41.2810.2440.3720.307
JIVE0.8160.5680.5970.474
Geweke_Mode1.2440.2870.3770.309
Geweke_Median1.2040.3090.3700.300
LIML_GS_Mode1.2600.2680.3730.308
LIML_GS_Median1.2200.2980.3700.300
CP_Mode1.2300.2930.3720.301
CP_Median1.1940.3150.3700.298
KVD_Mode1.3510.3840.5200.389
KVD_Median1.3810.3670.5290.405
Notes: Number of replications: 400. Geweke: nburn = 100, n = 2000. CP: nburn_GS = 100, nburn_MH = 100, n = 10,000, acceptance rate = 0.475 (0.010). KVD: nburn_GS = 100, nburn_MH = 100, n = 3000, acceptance rate = 0.400 (0.217).
Table 4. T = 50, ρ = 0.60, k2 = 9, R2 = 0.10.
Table 4. T = 50, ρ = 0.60, k2 = 9, R2 = 0.10.
MeanStdRMSEMAD
OLS1.5350.1110.5460.535
2SLS1.3630.2210.4250.371
MELO1.4630.1390.4830.463
BMOM0.9690.1320.1360.106
LIML1.0900.8640.8690.534
Fuller11.1820.4790.5120.366
Fuller41.3020.2910.4190.333
JIVE0.7060.9330.9780.728
Geweke_Mode1.3570.2390.4300.367
Geweke_Median1.3500.2450.4270.361
LIML_GS_Mode1.3750.2180.3280.380
LIML_GS_Median1.3670.2280.4320.374
CP_Mode1.2150.6290.6650.466
CP_Median1.2550.3880.4640.346
KVD_Mode1.5500.3760.6660.556
KVD_Median1.5730.3220.6570.576
Notes: Number of replications: 400. Geweke: nburn = 100, n = 1000. CP: nburn_GS = 200, nburn_MH = 200, n = 10,000, acceptance rate = 0.242 (0.040). KVD: nburn_GS = 200, nburn_MH = 100, n = 10,000, acceptance rate = 0.267 (0.188).
Table 5. T = 100, ρ = 0.60, k2 = 4, R2 = 0.10.
Table 5. T = 100, ρ = 0.60, k2 = 4, R2 = 0.10.
MeanStdRMSEMAD
OLS1.5380.0770.5430.538
2SLS1.1380.2080.2500.200
MELO1.2570.1560.3010.264
BMOM0.9540.1560.1630.127
LIML1.0230.2800.2810.210
Fuller11.0690.2500.2590.197
Fuller41.1710.1950.2590.209
JIVE0.9140.3200.3310.262
Geweke_Mode1.1490.2150.2620.208
Geweke_Median1.1110.2280.2540.198
LIML_GS_Mode1.1620.2050.2610.209
LIML_GS_Median1.1170.2250.2540.199
CP_Mode1.1550.2070.2590.206
CP_Median1.1070.2280.2520.196
KVD_Mode1.2330.2050.3100.258
KVD_Median1.2150.2100.3010.243
Notes: Number of replications: 400. Geweke: nburn = 100, n = 2000. CP: nburn_GS = 200, nburn_MH = 200, n = 10,000, acceptance rate = 0.616 (0.008). KVD: nburn_GS = 200, nburn_MH = 100, n = 10,000, acceptance rate = 0.312 (0.175).
Table 6. T = 100, ρ = 0.60, k2 = 9, R2 = 0.10.
Table 6. T = 100, ρ = 0.60, k2 = 9, R2 = 0.10.
MeanStdRMSEMAD
OLS1.5420.0780.5480.542
2SLS1.2580.1970.3250.274
MELO1.3760.1340.3990.376
BMOM0.9720.1320.1350.110
LIML1.0030.4370.4370.291
Fuller11.0710.3110.3190.243
Fuller41.1800.2330.2940.232
JIVE0.9270.4080.4140.333
Geweke_Mode1.2530.2010.3230.269
Geweke_Median1.2380.2060.3150.261
LIML_GS_Mode1.2650.1960.3300.278
LIML_GS_Median1.2470.2020.3190.266
CP_Mode1.1960.2640.3290.266
CP_Median1.1920.2320.3010.240
KVD_Mode1.3710.2780.4640.382
KVD_Median1.3950.2690.4780.397
Notes: Number of replications: 400. Geweke: nburn = 100, n = 1000. CP: nburn_GS = 200, nburn_MH = 200, n = 6000, acceptance rate = 0.434 (0.029). KVD: nburn_GS = 200, nburn_MH = 200, n = 10,000, acceptance rate = 0.210 (0.179).
Table 7. T = 100, ρ = 0.60, k2 = 4, R2 = 0.05.
Table 7. T = 100, ρ = 0.60, k2 = 4, R2 = 0.05.
MeanStdRMSEMAD
OLS1.5650.0800.5710.565
2SLS1.2540.2820.3800.309
MELO1.3760.1840.4190.379
BMOM0.9530.1830.1890.150
LIML1.0520.5840.5860.392
Fuller11.1580.3770.4090.307
Fuller41.2960.2440.3840.317
JIVE0.8330.6380.6590.527
Geweke_Mode1.2640.2850.3880.314
Geweke_Median1.2240.3160.3870.305
LIML_GS_Mode1.2740.2830.3940.320
LIML_GS_Median1.2320.3100.3870.306
CP_Mode1.2630.2950.3950.318
CP_Median1.2230.3160.3870.304
KVD_Mode1.3880.3890.5490.418
KVD_Median1.3940.3150.5040.414
Notes: Number of replications: 400. Geweke: nburn = 100, n = 2000. CP: nburn_GS = 100, nburn_MH = 100, n = 4000, acceptance rate = 0.611 (0.009). KVD: nburn_GS = 200, nburn_MH = 200, n = 8000, acceptance rate = 0.442 (0.224).
Table 8. T = 100, ρ = 0.60, k2 = 9, R2 = 0.05.
Table 8. T = 100, ρ = 0.60, k2 = 9, R2 = 0.05.
MeanStdRMSEMAD
OLS1.5740.0760.5790.574
2SLS1.3860.2190.4440.394
MELO1.4780.1310.4960.478
BMOM0.9790.1290.1310.105
LIML1.1390.8820.8930.545
Fuller11.2240.4770.5270.389
Fuller41.3350.2800.4370.358
JIVE0.8440.8230.8380.663
Geweke_Mode1.3850.2430.4550.395
Geweke_Median1.3800.2460.4530.390
LIML_GS_Mode1.3970.2300.4590.404
LIML_GS_Median1.3870.2360.4530.396
CP_Mode1.3380.4650.5750.433
CP_Median1.3370.3110.4590.376
KVD_Mode1.5840.4620.7450.592
KVD_Median1.6080.3680.7110.610
Notes: Number of replications: 400. Geweke: nburn = 100, n = 2000. CP: nburn_GS = 200, nburn_MH = 200, n = 10,000, acceptance rate = 0.433 (0.035). KVD: nburn_GS = 200, nburn_MH = 200, n = 10,000, acceptance rate = 0.371 (0.221).
Table 9. T = 100, ρ = 0.20, k2 = 4, R2 = 0.10.
Table 9. T = 100, ρ = 0.20, k2 = 4, R2 = 0.10.
MeanStdRMSEMAD
OLS1.1720.0900.1940.174
2SLS1.0460.2530.2570.206
MELO1.0830.1890.2060.164
BMOM0.8590.1900.2370.195
LIML1.0170.3330.3330.260
Fuller11.0290.2980.2990.236
Fuller41.0590.2350.2420.192
JIVE0.9570.4170.4190.340
Geweke_Mode1.0530.2510.2570.200
Geweke_Median1.0410.2670.2700.214
LIML_GS_Mode1.0580.2440.2510.197
LIML_GS_Median1.0440.2650.2690.212
CP_Mode1.0540.2550.2610.205
CP_Median1.0400.2710.2740.218
KVD_Mode1.1310.3680.3910.237
KVD_Median1.1610.3280.3650.245
Notes: Number of replications: 400. Geweke: nburn = 100, n = 1000. CP: nburn_GS = 100, nburn_MH = 100, n = 5000, acceptance rate = 0.615 (0.011). KVD: nburn_GS = 100, nburn_MH = 100, n = 1000, acceptance rate = 0.548 (0.200).
Table 10. T = 100, ρ = 0.20, k2 = 9, R2 = 0.10.
Table 10. T = 100, ρ = 0.20, k2 = 9, R2 = 0.10.
MeanStdRMSEMAD
OLS1.1790.0960.2030.181
2SLS1.0850.2140.2300.182
MELO1.1240.1460.1920.154
BMOM0.8230.1430.2280.193
LIML0.9920.3970.3970.301
Fuller11.0150.3470.3470.270
Fuller41.0550.2670.2730.216
JIVE0.9910.4810.4810.390
Geweke_Mode1.0840.2180.2340.184
Geweke_Median1.0790.2230.2370.187
LIML_GS_Mode1.0870.2120.2290.181
LIML_GS_Median1.0820.2180.2330.185
CP_Mode1.0540.3080.3130.223
CP_Median1.0630.2540.2620.207
KVD_Mode1.2490.2340.3420.283
KVD_Median1.2860.2350.3700.308
Notes: Number of replications: 400. Geweke: nburn = 100, n = 1000. CP: nburn_GS = 100, nburn_MH = 200, n = 5000, acceptance rate = 0.456 (0.023). KVD: nburn_GS = 100, nburn_MH = 100, n = 5000, acceptance rate = 0.413 (0.202).
Table 11. T = 50, ρ = 0.95, k2 = 4, R2 = 0.10.
Table 11. T = 50, ρ = 0.95, k2 = 4, R2 = 0.10.
MeanStdRMSEMAD
OLS1.8460.0520.8480.846
2SLS1.3590.1800.4020.363
MELO1.5720.1180.5840.572
BMOM1.0570.1180.1310.102
LIML0.9880.4040.4040.255
Fuller11.1690.1960.2590.221
Fuller41.4170.1200.4340.417
JIVE0.6370.6110.7110.478
Geweke_Mode1.3470.3020.4600.358
Geweke_Median1.2770.3770.4680.305
LIML_GS_Mode1.3380.1550.3720.345
LIML_GS_Median1.2520.1940.3180.281
CP_Mode1.3140.1620.3530.325
CP_Median1.2340.1940.3040.266
KVD_Mode1.4110.3790.5590.428
KVD_Median1.4620.4630.6540.514
Notes: Number of replications: 400. Geweke: nburn = 100, n = 3000. CP: nburn_GS = 200, nburn_MH = 200, n = 10,000, acceptance rate = 0.476 (0.010). KVD: nburn_GS = 200, nburn_MH = 200, n = 10,000, acceptance rate = 0.036 (0.038).
Table 12. T = 100, ρ = 0.95, k2 = 4, R2 = 0.10.
Table 12. T = 100, ρ = 0.95, k2 = 4, R2 = 0.10.
MeanStdRMSEMAD
OLS1.8500.0330.8510.850
2SLS1.2300.1260.2620.234
MELO1.4140.0940.4250.414
BMOM1.0440.0950.1050.082
LIML1.0250.1700.1720.132
Fuller11.0950.1420.1710.143
Fuller41.2640.0990.2820.265
JIVE0.8730.1990.2360.191
Geweke_Mode1.2160.1170.2460.223
Geweke_Median1.1500.1270.1970.172
LIML_GS_Mode1.2270.1180.2560.235
LIML_GS_Median1.1580.1280.2030.180
CP_Mode1.2210.1160.2500.228
CP_Median1.1540.1270.2000.176
KVD_Mode1.2580.2070.3310.280
KVD_Median1.2520.2940.3870.260
Notes: Number of replications: 400. Geweke: nburn = 100, n = 3000. CP: nburn_GS = 200, nburn_MH = 200, n = 10,000, acceptance rate = 0.626 (0.007). KVD: nburn_GS = 200, nburn_MH = 200, n = 10,000, acceptance rate = 0.022 (0.022).
Table 13. Performance of BMOM and KVD when ρ < 0.
Table 13. Performance of BMOM and KVD when ρ < 0.
MeanStdRMSEMADRemarks
T = 50, ρ = −0.60, k2 = 4, R2 = 0.40
BMOM0.8520.1290.1960.165Compare Table 1.
KVD_Mode0.9710.1500.1520.119Acceptance rate for
KVD_Median0.9990.1530.1520.119KVD: 0.713 (0.130)
T = 50, ρ = −0.60, k2 = 4, R2 = 0.10
BMOM0.5510.1910.4880.453Compare Table 3.
KVD_Mode0.8510.3270.3590.271Acceptance rate for
KVD_Median0.9340.3410.3470.267KVD: 0.680 (0.133)
T = 50, ρ = −0.60, k2 = 9, R2 = 0.10
BMOM0.4200.1360.6000.580Compare Table 4.
KVD_Mode0.8570.3670.3930.296Acceptance rate for
KVD_Median0.9270.3990.4060.291KVD: 0.482 (0.155)
T = 100, ρ = −0.60, k2 = 4, R2 = 0.10
BMOM0.6760.1600.3620.326Compare Table 5.
KVD_Mode0.9010.2130.2350.186Acceptance rate for
KVD_Median0.9640.2370.2390.190KVD: 0.772 (0.110)
T = 100, ρ = −0.60, k2 = 9, R2 = 0.10
BMOM0.5310.1290.4860.469Compare Table 6.
KVD_Mode0.9030.2400.2580.200Acceptance rate for
KVD_Median0.9520.2470.2520.198KVD: 0.614 (0.138)
T = 100, ρ = −0.60, k2 = 4, R2 = 0.05
BMOM0.5140.1810.5190.486Compare Table 7.
KVD_Mode0.8130.3060.3580.285Acceptance rate for
KVD_Median0.9080.3620.3730.287KVD: 0.720 (0.128)
T = 100, ρ = −0.60, k2 = 9, R2 = 0.05
BMOM0.4070.1310.6080.593Compare Table 8.
KVD_Mode0.8480.4240.4500.312Acceptance rate for
KVD_Median0.9070.3490.3610.275KVD: 0.585 (0.144)
T = 100, ρ = −0.20, k2 = 4, R2 = 0.10
BMOM0.7530.1950.3140.266Compare Table 9.
KVD_Mode1.0020.2670.2670.208Acceptance rate for
KVD_Median1.0370.2910.2930.218KVD: 0.699 (0.162)
T = 100, ρ = −0.20, k2 = 9, R2 = 0.10
BMOM0.6730.1590.3640.328Compare Table 10.
KVD_Mode1.0930.3180.3310.233Acceptance rate for
KVD_Median1.1290.2790.3070.241KVD: 0.553 (0.181)
T = 50, ρ = −0.95, k2 = 4, R2 = 0.10
BMOM0.4270.1200.5850.573Compare Table 11.
KVD_Mode0.7370.2440.3590.312Acceptance rate for
KVD_Median0.8360.2460.2950.239KVD: 0.173 (0.112)
T = 100, ρ = −0.95, k2 = 4, R2 = 0.10
BMOM0.5890.0970.4220.411Compare Table 12.
KVD_Mode0.8150.1550.2410.209Acceptance rate for
KVD_Median0.8890.1530.1890.156KVD: 0.179 (0.103)
Notes: Number of replications: 500.

Share and Cite

MDPI and ACS Style

Gao, C.; Lahiri, K. A Comparison of Some Bayesian and Classical Procedures for Simultaneous Equation Models with Weak Instruments. Econometrics 2019, 7, 33. https://doi.org/10.3390/econometrics7030033

AMA Style

Gao C, Lahiri K. A Comparison of Some Bayesian and Classical Procedures for Simultaneous Equation Models with Weak Instruments. Econometrics. 2019; 7(3):33. https://doi.org/10.3390/econometrics7030033

Chicago/Turabian Style

Gao, Chuanming, and Kajal Lahiri. 2019. "A Comparison of Some Bayesian and Classical Procedures for Simultaneous Equation Models with Weak Instruments" Econometrics 7, no. 3: 33. https://doi.org/10.3390/econometrics7030033

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop