Next Article in Journal
When It Counts—Econometric Identification of the Basic Factor Model Based on GLT Structures
Previous Article in Journal
Dirichlet Process Log Skew-Normal Mixture with a Missing-at-Random-Covariate in Insurance Claim Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On the Proper Computation of the Hausman Test Statistic in Standard Linear Panel Data Models: Some Clarifications and New Results

by
Julie Le Gallo
1,*,† and
Marc-Alexandre Sénégas
2,†
1
Center of Economics and Sociology Applied to Rural Areas, UMR1041, l’Institut Agro Dijon, INRAE, University Bourgogne Franche-Comté, 21000 Dijon, France
2
Bordeaux School of Economics, UMR CNRS6060, University of Bordeaux, 33000 Bordeaux, France
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Econometrics 2023, 11(4), 25; https://doi.org/10.3390/econometrics11040025
Submission received: 7 June 2023 / Revised: 31 October 2023 / Accepted: 6 November 2023 / Published: 8 November 2023

Abstract

:
We provide new analytical results for the implementation of the Hausman specification test statistic in a standard panel data model, comparing the version based on the estimators computed from the untransformed random effects model specification under Feasible Generalized Least Squares and the one computed from the quasi-demeaned model estimated by Ordinary Least Squares. We show that the quasi-demeaned model cannot provide a reliable magnitude when implementing the Hausman test in a finite sample setting, although it is the most common approach used to produce the test statistic in econometric software. The difference between the Hausman statistics computed under the two methods can be substantial and even lead to opposite conclusions for the test of orthogonality between the regressors and the individual-specific effects. Furthermore, this difference remains important even with large cross-sectional dimensions as it mainly depends on the within-between structure of the regressors and on the presence of a significant correlation between the individual effects and the covariates in the data. We propose to supplement the test outcomes that are provided in the main econometric software packages with some metrics to address the issue at hand.

1. Introduction

As is well known, the implementation of the Hausman specification test (Hausman 1978) might be affected, in practice and in a finite sample setting, by a non-positive definiteness or (in-) definiteness problem for the variance-covariance matrix corresponding to the difference between the efficient estimator and the consistent estimator1. This, in turn, can potentially lead to a negative value of the test statistic, which makes it unreliable for interpreting the test outcome. This problem is usually mentioned in the context of models using instrumental variables (IV) (Baum et al. 2003, pp. 19–22; Staiger and Stock 1997, pp. 567–68) where the Hausman test is performed to assess the endogeneity of the regressors, given a set of instruments. In this case, one solution to ensure a symmetric positive definite (hereafter, SPD) covariance matrix is to use a common and identical estimator for the variance of the (idiosyncratic) error term when confronting the Ordinary Least Squares (OLS) and the IV estimators (Hayashi 2000, pp. 220–33; Baum et al. 2003, pp. 19–22).
Yet, although the issue has been pointed out as a warning by Hausman himself when he addressed the case of a static balanced, panel data model in his seminal presentation of the test (Hausman 1978, footnote 25, p. 1267), it has not been, to the best of our knowledge, further formally examined in this specific framework2. This is surprising, as one of the most widespread applications of the Hausman test is for assessing the relevance of the random (RE) versus fixed effects (FE) specification in a panel data model. From our point of view, this application, however, deserves attention in its own right, and this article aims at filling this gap. Specifically, our contributions are twofold.
We first provide new and detailed analytical results for the implementation of the Hausman test in the case of a static and balanced panel data model with individual effects. In particular, we show that the test statistic is unreliable in a finite sample if the variance of the RE estimator is computed on the basis of the estimation of the quasi-demeaned model (we denote it in what follows QDM)3 rather than the conventional and direct implementation of the Feasible Generalized Least Squares (FGLS) on the RE panel data model. This result directly follows from the way in which standard errors are computed under the QDM approach and which can lead to a positive definiteness problem for the covariance matrix in the Hausman test statistic formula. To establish the unreliability, we perform a systematic analysis of the difference between the Hausman statistics computed under the two approaches as well as of the behavior of the statistic that uses the estimates based on the QDM regression framework. In particular, we show that the latter mainly depends on the within-between structure of the regressors and on the presence of a significant correlation between the individual effects and the covariates in the data4.
Second, based on a review of the main existing econometric software programs that deal with panel data models, we show that the vast majority of the related packages in those programs implement, by default, the Hausman test using the unreliable statistic. This leads us to assess different ways to supplement the test outcomes provided by those programs and/or to circumvent the reliability problem potentially raised by the use of the statistic involved.
The outline of the paper is as follows. First, in Section 2, we show how the two versions of the Hausman test statistics can produce diverging results for some well-known textbook examples in the context of panel data models. Then, in Section 3, we formalize the implications of the two approaches for the Hausman statistic and derive new analytical results regarding the comparison of the two versions of the statistic that follows from each of the two approaches. We also revisit textbook examples and provide some simulation results in the one-regressor case. Finally, in Section 4, we detail how the Hausman test is implemented in a variety of econometric software programs dealing with panel data models and discuss, on the basis of this review, some possible ways to implement this test in a reliable and robust manner with this software. The last section concludes.

2. Motivation

In this section, we illustrate the extent to which significant differences in the values of the Hausman test statistic may arise depending on the approach adopted to estimate the panel data model parameters and, more particularly, those pertaining to the variance components under the random effects (RE) specification.

2.1. Notation

All the case studies considered below are for a standard, linear, static, and balanced panel data model with individual effects (also called the one-way error component-panel data-model). Accordingly, we consider the following relationship:
y i t = α + X i t · β + u i t
where i = 1 , , N denotes the cross-section dimension; t = 1 , , T the time-series dimension; y i t the i t -th observation of the dependent variable; X i t a column vector of the i t -th observation on K explanatory variables; α an unknown scalar and β a K × 1 vector of unknown parameters, both to be estimated5. The error term u i t is assumed to take the following composite form:
u i t = α i * + ε i t
where α i * denotes the (unobservable) individual-specific (or individual) effect-also called the individual component of u i t and ε i t denotes the idiosyncratic component of u i t . Furthermore, we assume that α i * and ε i t are independent of each other and that α i * I I D ( 0 , σ α 2 ) and ε i t I I D ( 0 , σ ε 2 ) . It is also assumed that E [ X i s · ε i t ] = 0 for all s = 1 , , T .
As is well known, two alternative specifications are usually considered regarding the correlation between the individual effect, α i * and the regressors contained in X i t . On the one hand, the random-effects (RE) model assumes that this correlation is zero, ensuring that X i t is strictly exogenous for u i t in (1). On the other hand, the ‘fixed effects’ model allows for a non-zero correlation between the individual effect and the regressors. The Within transformation is then used to obtain an unbiased estimate for β .
The Hausman specification test (Hausman 1978) is widely used for testing the no-correlation assumption underlying the RE specification. In our setting, the test is based on the asymptotic properties of the RE and Within (or fixed-effects) estimators of β . Both estimators are consistent under the null hypothesis (of no correlation) while, under the alternative, only the Within estimator is consistent as the RE estimator is (asymptotically) biased. Accordingly, the test statistic is built on a distance measure between the Within and RE estimators. In its implementable version, the statistic writes as:
H M O j ( β ^ W β ^ R E ) · v a ^ r β ^ W v a ^ r j β ^ R E 1 · ( β ^ W β ^ R E )
where β ^ W and v a ^ r [ β ^ W ] denote the Within (or fixed effects) estimator of β and a consistent estimator of its asymptotic covariance matrix, respectively; β ^ R E and v a ^ r j [ β ^ R E ] denote the RE estimator of β and a consistent estimator of its asymptotic covariance matrix, respectively. The index j = 1 , 2 indicates that two approaches can be considered to obtain a consistent estimator of the asymptotic covariance matrix, leading to two versions of the test statistic (see Section 3.3).

2.2. Motivating Examples

We reproduce some outcomes of panel model estimations for well-known case study applications taken from main textbooks in the field. In the following tables, Std Err._1 and Std. Err._2 denote the two sets of standard errors for the parameters implied by the use of two different estimators for the covariance matrix of the RE estimator (detailed below). The values of the two related versions of Hausman test statistics are denoted H M O 1 and H M O 2 . Other variables and parameters are shown in the tables, the definitions and interpretations of which are left for discussion in Section 3 where we further comment on those results.

2.2.1. Motivating Example 1: Gasoline

Baltagi (2005) provides an interesting example of an important difference between the two versions of the Hausman test statistic in a study on the determinants of gasoline demand over the period 1960–1978 across 18 OECD countries6. The following specification is adopted:
log G a s / C a r = α + β 1 · log Y / N + β 2 · log P M G / P G D P + β 3 · log C a r / N + u
where G a s / C a r is motor gasoline consumption per auto, Y / N is real income per capita, P M G / P G D P is real motor gasoline price and C a r / N denotes the stock of cars per capita.
Table 1 clearly shows that the two values of the Hausman statistic deviate strongly from each other, even if the null hypothesis is rejected in both cases. It is also interesting to observe that the two sets of standard errors remain close to each other7.

2.2.2. Motivating Example 2: Airline

Greene (2000) relies on the following specification concerning the cost function in the airline industry8:
log c o s t = α + β 1 · log Q + β 2 · log f u e l p r i c e + β 3 · l o a d f a c t o r + u
where c o s t is the total cost, in $1000; Q is output, measured in “revenue passenger miles” (index number); f u e l p r i c e is fuel price and l o a d f a c t o r is a rate of capacity utilization: it is the average rate at which seats on the airline’s planes are filled. The dataset consists of six firms observed yearly for 15 years (1970 to 1984).
In this case (Table 2), H M O 1 > H M O 2 while the two values are much closer and clearly lower than in the previous case. They both lead to the rejection of the null. Also, the two sets of standard errors are quite close.
The airline case provides further interesting outcomes when the specification is estimated with two covariates. We single out the regression with log f u e l p r i c e and l o a d f a c t o r as the two covariates. The corresponding estimation results are provided in Table 3.
This new set of results is interesting for the negative sign of the computed H M O 2 and furthermore in that H M O 1 > H M O 2 . Using the absolute value of the H M O 2 statistic to perform the Hausman test (as some software packages do), the null hypothesis would not be rejected contrary to the outcome implied by H M O 1 . Note at this point that these results are obtained for a sample distribution of the covariate’s observations featuring a structure of the variance largely skewed in its within dimension.

2.2.3. Motivating Example 3: Wage determination

In another application, Greene (2012) relies on Cornwell and Rupert (2008)’s study about the determinants of the returns to schooling. The dataset is a balanced panel of 595 observations on heads of households that runs over the period (1976–1982). Among the specifications examined by Greene, we consider the following one:
log w a g e = α + β 1 · E X P + β 2 · E X P 2 + β 3 · W K S + β 4 · O C C + β 5 · I N D β 6 · S O U T H + β 7 · S M S A + β 8 · M S + β 9 · U N I O N + u
where E X P denotes the number of years of full-time work experience; W K S , the number of weeks worked; O C C = 1 if the status of the occupation is blue-collar occupation, 0 if not; I N D = 1 if the individual works in a manufacturing industry, 0 if not; S O U T H = 1 if the individual resides in the south, 0 if not; S M S A = 1 if the individual resides in a city, 0 if not; M S = 1 if the individual is married, 0 if not; U N I O N = 1 if the individual wage is set by a union contract, 0 if not; lastly log [ w a g e ] denotes the log of the (yearly) wage9.
The estimation results are presented in Table 4. Here, even with a large sample (4765 observations), we observe significant differences between the two Hausman statistics (which are very large) as well as the two sets of the standard errors for the random effects model (the ratio of the latter that is provided by h (see later) is close to 1.3 ).

3. The Two Versions of the Hausman Test Statistic

The two versions of the Hausman statistic refer to two possible approaches for estimating the covariance matrix of the estimator of the parameters of the model given in its RE specification (1) and (2). We first start with a formal and explicit presentation of these approaches since their implications for the computation of the Hausman test statistic have remained largely unnoticed.

3.1. The Original Hausman Test Specification in a Balanced Panel Data Model

We rewrite model (1) and (2), stacking the observations over the time and cross-sectional dimensions:
y = ι N T · α + X · β + u
where y is the ( N T × 1 ) vector for the observations of the dependent variable; ι N T a vector of ones of dimension N T and X is the ( N T × K ) matrix including the observations of the K explanatory variables; u ( u 11 u 1 T u N 1 u N T ) is the ( N T × 1 ) vector for the composite error terms. The covariance matrix of u is denoted by Ω . Given the properties of α i * and ε i t , it takes the following form:
Ω = σ ε 2 · Ω *
with Ω * = [ W + ( 1 / ψ 2 ) · B ] , ψ 2 σ ε 2 / σ α ε 2 , σ α ε 2 σ ε 2 + T · σ α 2 , B I N ( J T / T ) (the Between operator), W I N T B (the Within operator), where ⊗ is the Kronecker product, I N (resp. I T ) the ( N × N ) (resp. ( T × T ) ) identity matrix and J T = ι T · ι T a matrix of ones of dimension T.
It is useful to define Ω C * such that Ω C * = Ω * · ( I N T B N T ) = W + ( 1 / ψ 2 ) · B C with B C denoting the centered Between operator defined as B C B B N T and with B N T ι N T · ι N T · ι N T 1 · ι N T . We also use Ω * 1 2 defined as Ω * 1 2 · Ω * 1 2 = Ω * 1 . As B and W are symmetric and idempotent, Ω * 1 = [ W + ψ 2 · B ] and, in turn: Ω * 1 2 = [ W + ψ · B ] . The same applies to Ω C * 1 2 and Ω C * 1 with B C replacing B in the previous formulas.
Assume first, that the variance components (and thus Ω , Ω * and Ω C * ) are known. Then, it is well established that:
(1) The RE estimator of β ( β ^ R E ) corresponds to the Generalized Least Squares (GLS) estimator of β in (3), noted β ^ G L S , and is given by (4) with its covariance matrix by (5):
β ^ R E = β ^ G L S = X · Ω C * 1 · X 1 · X · Ω C * 1 · y
var β ^ R E = var β ^ G L S = σ ε 2 · X · Ω C * 1 · X 1
(2) The Within (or fixed effects) estimator of β , β ^ W , is given by (6) and its covariance matrix by (7):
β ^ W = X W · X W 1 · X W · y W
var β ^ W = σ ε 2 · X W · X W 1
with y W W · y and X W W · X .
The GLS estimator β ^ G L S can alternatively be obtained, using the quasi-demeaned model (also called the partial Within transformation model) built from (3) with the premultiplying factor10  Ω * 1 2 :
y * * = ω * * · α + X * * · β + u * *
with ω * * ψ · ι N T ;   X * * Ω * 1 2 · X ; y * * Ω * 1 2 · y ; u * * Ω * 1 2 · u . It is easy to check that the OLS estimator of β in (8)-denoted as β ^ O L S -corresponds to β ^ G L S 11. Also:
var β ^ O L S = v a r β ^ G L S = σ ε 2 · X * · X * 1
where X * Ω C * 1 2 · X which corresponds to (5) as X * · X * = X · Ω C * 1 · X .
Based on the former estimators, the Hausman specification test statistic initially proposed by Hausman (1978) is:
H M O * q ^ · var β ^ W var β ^ R E 1 · q ^
with q ^ β ^ W β ^ R E and var z ^ denoting the (finite sample) exact covariance matrix of z ^ . Under the null hypothesis of no-correlation, H M O * is asymptotically distributed as a χ 2 K .

3.2. Two Estimation Procedures

In practice, Ω , Ω * and Ω C * are unknown and replaced by consistent estimators (noted Ω ^ , Ω * ^ and Ω C * ^ , respectively12). In this case, the Hausman test statistic is written as:
H M O β ^ Δ W R E · v a ^ r β ^ W v a ^ r j β ^ R E 1 · β ^ Δ W R E
where:
(1) β ^ Δ W R E β ^ W β ^ R E with, now, β ^ R E = β ^ F G L S indicating that the RE estimator corresponds to the Feasible Generalized Least Squares (FGLS) estimator for β ( β ^ F G L S ), accounting for the use of Ω C * ^ instead of Ω C * .
(2) v a ^ r z ^ is a consistent estimator of the asymptotic covariance matrix of z ^ built upon the finite-sample estimator of the (exact) variance of z ^ and with j = 1 , 2 indicating that, for the RE estimator, two approaches are available to compute this matrix.
Under suitable conditions assumed to hold in what follows13, β ^ F G L S and β ^ G L S are asymptotically equivalent and H M O is, as H M O * , asymptotically distributed as a χ 2 K .
We now discuss the choice of variance component estimators that are required to compute H M O .
First, a consistent estimator for var [ β ^ W ] is: v a ^ r [ β ^ W ] = σ ^ ε w 2 · ( X W · X W ) 1 , where σ ^ ε w 2 ( u ^ W · u ^ W ) / ( N T 1 K ) denotes the consistent estimator for σ ε 2 built from the OLS estimation associated with the Within (transformed) regression model, u ^ W denoting the ( N T × 1 ) vector of the related residuals.
Second, to obtain v a ^ r [ β ^ F G L S ] , two approaches are possible.

3.2.1. Approach 1: The (Direct) FGLS Approach

Relying on the asymptotic equivalence between β ^ F G L S and β ^ G L S , the computation of a consistent, asymptotic covariance matrix estimator for var [ β ^ R E ] can be considered directly from the expression (5) where Ω C * ^ is substituted for Ω C * and σ ^ ε 2 for σ ε 2 , which yields:
v a ^ r 1 β ^ R E v a ^ r 1 β ^ F G L S = σ ^ ε 2 · X · Ω C * ^ 1 · X 1
with Ω C * ^ 1 = [ W + ψ ^ 2 · B C ] and ψ ^ 2 ( σ ^ ε 2 / σ ^ α ε 2 ) . σ ^ ε 2 in (11) should logically be chosen as the same estimator as the one entering into14  ψ ^ 2 for computing Ω C * ^ also appearing in (11). One usually relies on the estimator based on the ‘fixed effects model’ for this purpose, so that we set σ ^ ε 2 = σ ^ ε w 2 (Swamy-Arora approach). We use that correspondence in what follows.

3.2.2. Approach 2: The Quasi-Demeaning Approach

A consistent (asymptotic) covariance matrix estimator for β ^ R E can alternatively be obtained from the quasi-demeaned regression model (8) considered in its feasible version with Ω * ^ 1 2 substituting for Ω * 1 2  where Ω * ^ 1 2 = W + ψ ^ · B and ψ ^ = ψ ^ 2 . We note this feasible version of the QDM model as the FQDM regression model. In this case, and relying again on the asymptotic equivalence between β ^ G L S and β ^ F G L S , the resulting “plug-in” estimator can then be computed from the formula giving the covariance matrix for the OLS estimator of β in (8), i.e., (9), where X ˜ * substitutes for X * and σ ^ ε * 2 for σ ε 2 :
v a ^ r 2 β ^ R E v a ^ r 2 β ^ F G L S = σ ^ ε * 2 · X ˜ * · X ˜ * 1
In line with the quasi-demeaning approach, the computation of σ ^ ε * 2 is, here, usually considered as a byproduct of the OLS estimation process at work for the FQDM regression model. Accordingly, σ ^ ε * 2 is computed as σ ^ ε * 2 ( u ˜ ^ * * · u ˜ ^ * * ) / ( N T ( K + 1 ) ) where u ˜ ^ * * denotes the N T vector of the OLS residuals in the FQDM regression model.
From (11) and (12), we observe that v a ^ r 1 [ β ^ R E ] v a ^ r 2 [ β ^ R E ] as σ ^ ε w 2 σ ^ ε * 2 . Thus, the two approaches differ in providing two distinct estimators for the variance of the (RE) estimator insofar as they rely on two different estimators of the variance component, σ ε 2 . This leads to what we call, in the following, a disturbance variance disconnect problem.15

3.3. Comparing the Two Versions

From the previous results, it follows that two possible expressions are available for an implementable version of the Hausman test statistic, depending on which estimator of the asymptotic covariance matrix for β ^ R E is chosen.

3.3.1. Two Statistics

Using v a ^ r 1 [ β ^ R E ] , we have:
H M O 1 β ^ Δ W R E · σ ^ ε w 2 · X W · X W 1 σ ^ ε w 2 · X ˜ * · X ˜ * 1 1 · β ^ Δ W R E
or, using v a ^ r 2 [ β ^ R E ] , we have:
H M O 2 ( β ^ Δ W R E ) · σ ^ ε w 2 · X W · X W 1 σ ^ ε * 2 · X ˜ * · X ˜ * 1 1 · β ^ Δ W R E
Hausman (1978) originally proposed using H M O 1 , which he considered as the legitimate computational version of H M O * (Hausman 1978, footnote 25, p. 1267):16
“Note that the elements of q ^ and its standard errors are simply calculated given the estimates β ^ F E and of β ^ F G L S and their standarderrors, making sure to adjust to use the fixed effects estimate of  σ ε 2 ”.
On the other hand, as we will see in Section 4, most software programs compute the Hausman test statistic based on the second measure, H M O 2 . Consequently, it is important to analyze how H M O 2 behaves in relation to H M O 1 in finite sample settings. For that purpose, define the ratio h as h σ ^ ε * 2 / σ ^ ε w 2 . H M O 2 can then be rewritten as:
H M O 2 = β ^ Δ W R E · σ ^ ε w 2 1 · X W · X W 1 h · X ˜ * · X ˜ * 1 1 · β ^ Δ W R E
Comparing with (13), we note that H M O 2 diverges from H M O 1 as h 1 .

3.3.2. Main Results

To go further into the comparison of the two versions of the Hausman test statistic, we rewrite their expressions as:
H M O 1 = β ^ Δ W R E · σ ^ ε w 2 · Γ 1 · β ^ Δ W R E
H M O 2 = β ^ Δ W R E · σ ^ ε w 2 · Γ ˜ 1 · β ^ Δ W R E
where we define Γ = ( X W · X W ) 1 ( X ˜ * · X ˜ * ) 1 and Γ ˜ = ( X W · X W ) 1 h · ( X ˜ * · X ˜ * ) 1 .
From (16) and (17), the comparison between H M O 1 and H M O 2 is based on the one between Γ ˜ and Γ . Note that ( X ˜ * · X ˜ * ) = ( X W · X W ) + ψ ^ 2 · ( X B C · X B C ) with X B C B C · X and can be rewritten as ( X ˜ * · X ˜ * ) = H * · ( X W · X W ) where H * I K + ψ ^ 2 · ( X B C · X B C ) · ( X W · X W ) 1 . It follows that Γ writes as Γ = ( X W · X W ) 1 · ( H * I K ) · H * 1 and Γ ˜ as Γ ˜ = ( X W · X W ) 1 · ( H * h · I K ) · H * 1 .
Finally, define h min * min σ H * and h max * max σ H * with σ H * denoting the spectrum of H * and with: 1 < h min * < h max * .
We establish the following results (see Appendix A for details and related proofs):
  • Γ is a symmetric positive definite (SPD) matrix. It follows that H M O 1 is a positive-definite quadratic form.
  • Γ ˜ can be either a symmetric positive or a negative definite matrix or even an indefinite matrix depending on specific conditions holding for h. As a consequence, H M O 2 can be of either sign (and even of indeterminate sign a priori) depending on the values taken by h. Specifically, we have:
    (a)
    Γ ˜ is a symmetric positive definite (SPD) matrix iff h < h min * . In this case, H M O 2 is a positive-definite quadratic form
    (b)
    Γ ˜ is a symmetric negative definite (SND) matrix iff h > h max * . In this case, H M O 2 is a negative-definite quadratic form
    (c)
    Γ ˜ is indefinite iff h min * h h max * . In this case, H M O 2 can be of either sign, which is indeterminate a priori.
Based on those results, comparing H M O 2 relative to H M O 1 depends on whether Γ ˜ is SPD or SND. We then have:
  • If Γ ˜ is SPD, the relevant comparison relies upon the magnitude ( H M O 2 H M O 1 ) . We have H M O 2 H M O 1 whenever h 1
  • If Γ ˜ is SND, it necessarily follows that H M O 2 < 0 < H M O 1 .
In establishing those results, we directly echo the discussions, mentioned in the introduction, about the positive definiteness of the variance-covariance matrix estimator in the expression of the Hausman test statistic. As we observe, whether this matrix is SPD or not (which translates into whether Γ or Γ ˜ is SPD or not) depends on the choice of the estimator for σ ε 2 , which, itself, hinges on the approach that is adopted to estimate the variance components associated with the RE model. In other terms, Γ is, by construction, an SPD matrix, and this has to be related to the use of the same estimator for σ ε 2 , i.e., σ ^ ε w 2 , in the computation of the covariance matrix estimator. On the other hand, this is not necessarily the case for Γ ˜ and this has to do with the fact that two different estimators have been considered for σ ε 2 , σ ^ ε w 2 and σ ^ ε * 2 ( h 1 ).
Table 5 provides an overview of all possible cases. A wide spectrum of outcomes can be obtained for the value of H M O 2 depending on the value of h. Unreliable results for the test may arise notably when H M O 2 is negative.

3.3.3. Back to the Case Studies

Based on h and the proposed metrics for H * , we can now analyze the mechanisms driving the various outcomes observed for the case studies selected in Section 2.2.
From Table 1, Table 2, Table 3 and Table 4, in 3 cases out of 4, h min * < h < h max * . It follows that in those cases, Γ ˜ is an indefinite matrix. As a consequence, the sign of H M O 2 is a priori indeterminate as well the relative magnitudes of H M O 1 and H M O 2 . The observed outcome depends on the specific value taken by β ^ Δ W R E for the sample considered.
Conversely, in the two covariates’ regression case drawn from Greene, where H M O 2 is computed as a negative scalar, we logically have h > h max * and even h > 2 · h max * 1 which is consistent with H M O 1 > H M O 2 , what we, by the way, also observe.

3.4. What about h?

As we have shown, the value of h is key for determining the outcome of the Hausman test if it is measured through H M O 2 . In this section, we analyze the main determinants for this ratio and provide an illustration through simulations in the single regressor case.

3.4.1. Determinants

The value of h essentially derives from the comparison between σ ^ ε w 2 and σ ^ ε * 2 and therefore, in turn, from the two residual sums of squares that are associated with, respectively, the Within model ( u ^ W · u ^ W ) and the feasible, quasi demeaning model ( u ˜ ^ * * · u ˜ ^ * * ) . We show (see Appendix B) that the following relationship holds between the two expressions:
u ˜ ^ * * · u ˜ ^ * * = u ^ W · u ^ W + ψ ^ 2 · u ^ B · u ^ B + Δ *
where u ^ B denotes the N T vector of the OLS regression residuals for the Between (transformed) regression model y B = ι N T · α + X B · β + u B with y B B · y ; X B B · X , u B B · u and where Δ * is defined as:
Δ * β ^ W β ^ R E · Γ 1 · β ^ W β ^ R E
where Γ 1 is defined as in (16) and can be written as: Γ 1 = ( ψ ^ 2 ) 1 · ( X W · X W ) · ( X B C · X B C ) 1 · ( X ˜ * · X ˜ * ) . From the definition of σ ^ α ε 2 and σ ^ ε w 2 , we obtain u ^ B · u ^ B = ( N ( K + 1 ) ) · σ ^ α ε 2 and u ^ W · u ^ W = ( N T 1 K ) · σ ^ ε w 2 , so that (18) can be rewritten as:
σ ^ ε * 2 = 1 K N T K + 1 · σ ^ ε w 2 + Δ * N T K + 1
from which it follows that:
h = 1 + η · Δ * K · σ ^ ε w 2 1
with η K / ( N T K + 1 ) . Furthermore, considering the definition of H M O 1 given in (16), we have: Δ * = σ ^ ε w 2 · H M O 1 and therefore:
h = 1 + H M O 1 K N T K + 1
so that h 1 whenever H M O 1 K .
Taking advantage of the relationship between h and H M O 1 , we identify two categories of determinants for H M O 1 and h from (16) and (21):
  • The first category is related to the structure of the data at hand, i.e., the Between and Within components of the (empirical) covariance matrix of the explanatory variables. They are captured by the matrices ( X W · X W ) and  ( X B C · X B C ) , which influence the structure of H * (and therefore the magnitude of its eigenvalues).
  • The second category is linked to the correlation between the individual effects α i and the regressors contained in X i t , which determines the extent of the (asymptotic as well as finite sample) bias for β ^ R E (with respect to β ). This affects the gap between β ^ W (that is unbiased) and β ^ R E and therefore the value of β ^ Δ W R E .
These same factors in turn influence the determination of H M O 2 , the value of which is mostly linked to the comparison between h and the eigenvalues of H * .
With respect to the behavior of H M O 2 compared to H M O 1 , two mechanisms are at work. Assume that H M O 1 is large because of a significant distance between the RE and Within estimator. On the one hand, we could expect that H M O 2 will also be large and therefore that both statistics will correctly lead to reject the null hypothesis. This is because when H M O 1 is large, the further h will be from 1 from the upside, and in turn, the more likely it will be for the condition 1 < h to prevail, which leads to H M O 2 > H M O 1 as long as h remains below h m i n . On the other hand, the larger H M O 1 , the more likely it will actually be that h > h m i n and this could be all the more the case as the structure of the covariance matrix would be such that h m i n (or even h max ) is relatively small. In this case, the sign of H M O 2 becomes indeterminate which does not allow for a clear conclusion about the relative magnitude of the two test statistics and creates a possible divergence for the interpretation of the test.

3.4.2. Illustrations in the Single Regressor Case

To illustrate the role of the previous factors as well as to clarify their interpretation, we perform Monte-Carlo simulations on the behavior of the main magnitudes involved in the comparison between ( H M O 1 ) and ( H M O 2 ) in a single-regressor model ( K = 1 ) .

Preliminary Results

In such a setting, X = x and the various expressions for the main estimators and statistics simplify accordingly. We measure the total sample variance in the observations for x with s x 2 where s x 2 = ( N · T ) 1 · x T 2 and x T 2 x · I C · x , with I C I N T B N T . We then define x W 2 x · W · x and x B C 2 x · B C · x . By construction, we have x T 2 = x B C 2 + x W 2 . Hence, x W 2 and x B C 2 can be used as a measure of, respectively, the Within and the Between components (up to a N T - factor) of the total (empirical) variance in the NT observations contained in x . Finally, define θ W = x W 2 / x T 2 the share of the Within variance in the total variance. Then, substituting, we obtain:
h m i n = h m a x h * = 1 + ψ ^ 2 · x B C 2 x W 2 = 1 + ψ ^ 2 · 1 θ W θ W H M O 1 = ν · x W 2 · h * h * 1 h = 1 + η · ν · x W 2 · h * h * 1 1 H M O 2 = ν · x W 2 · h * h * h
with ν = ( σ ^ ε 2 ) 1 · ( β ^ Δ W R E ) 2 and η = 1 / ( N T 2 ) .

Design of Simulations

We perform Monte-Carlo simulations on the behavior of the previous four quantities, h * , h, H M O 1 and H M O 2 . The details of the simulation design are presented in Appendix C. We generate several series of y i t a based on a model y i t a = α + β · x i t a + u i t a , where we fix α = β = 1 and we let the other parameters of the simulation vary: N = ( 20 , 40 , 80 ) , T = ( 20 , 40 , 80 ) , s x 2 = ( 0.01 , 1 , 100 ) , θ W = ( 0.1 , 0.5 , 0.9 ) , σ u 2 = ( 0.01 , 1 , 100 ) , ρ x u = ( 0.99 , 0.9 , 0.5 , 0 , 0.5 , 0.9 , 0.99 ) and ρ u = ( 0.1 , 0.5 , 0.9 ) with the same notations as before, ρ x u is the correlation between x and u (on the cross-sectional dimension) and ρ u is the intra-class coefficient of the error term (share of the within variance in the total variance in u)17. Then, for each combination of the parameters, we perform 199 replications and compute the median of H M O 1 and H M O 2 . To explore the main dimensions of variability of H M O 1 and H M O 2 , we perform an ANCOVA analysis (Table A2 in Appendix C) where we regress, respectively, each of the means and medians on the levels of the different parameters and the interactions that we found significant.

Results of the Simulations

The results are similar whether we consider the mean or the medians over the replications of the levels of H M O 1 and H M O 2 for each combination of parameters. The value of H M O 1 and H M O 2 significantly depend on T, θ W , high absolute values of ρ x u and ρ u . Conversely, the scale parameters x T 2 and σ u 2 do not have a significant impact on H M O 1 and  H M O 2 .
The visual representation of the behavior of (the median) H M O 1 and H M O 2 for N = 80 , T = 80 , s x 2 = σ u 2 = 1 and ρ u = 0.5 is in Figure 1. This illustrates the strong dependence of the values of the Hausman statistics on θ W on the one hand and on the correlation between x and u on the other hand. In particular, for high absolute values of ρ x u and high values of θ W , H M O 2 even yields negative values.
This latter result is, for example, consistent with the regression outcomes obtained with log f u e l p r i c e as a single covariate in the framework of motivating Example 2 [Airline]. In this case (see Table A3), H M O 2 is computed as a negative scalar with H M O 1 > H M O 2 . Note also the extremely large value of θ W for the covariate which partly drives the values taken by h * and h.
Finally, looking more closely at the distribution of negative values for H M O 2 (Figure A1 in Appendix C), we clearly see that they are an increasing function of ρ u , θ W and ρ x u (correlation between x and u) in absolute value. While the empirical size of both tests remains around 5 % with little variations for the various values of the parameters, the power of H M O 2 consequently drops to 0 for the highest values of ρ u , θ W and ρ x u (Figure A2 in Appendix C).

4. The Implementation of the Hausman Test in Standard Econometric Software Packages for Panel Data: A Brief Review and Discussion

In this section, we review how six well-known econometric software packages deal with the implementation of the Hausman test in a standard panel data model and provide some discussion.

4.1. Review

  • STATA programming commands for the estimation of the random-effects panel data models (xtreg with the re option) rely on the specification of the quasi-demeaned model in (8). As a consequence, the random effects parameter estimates as well as its “conventional” covariance matrix estimate are provided as standard outputs of the OLS regression performed on that model. In particular, the vce(conventional) - default - command yields the (asymptotic) covariance matrix estimate based on the standard variance estimator for OLS regression. This corresponds to v a ^ r 2 [ β ^ F G L S ] with, accordingly, σ ^ ε * 2 used for the residual variance estimate. The default version of the command for implementing the standard Hausman test (the hausman command) corresponds accordingly to H M O 2 .
  • The R PLM package developed by Croissant and Millo (2008) allows estimating a wide range of panel data models with R software. Regarding the random-effects specification, the estimation process can be implemented via the plm function, whose model argument takes the random option. Croissant and Millo (2008) point out that it could have been possible to program the computation of the covariance matrix estimator for β ^ F G L S directly from the formula (11), “once the variance components have been estimated and hence the covariance matrix of errors”. However, to limit the computational costs associated with the inversion of the ( N T × N T ) matrix ( Ω ^ ) or ( Ω C * ^ ) and the related memory limits to store it, plm resorts to the specification and estimation of the quasi-demeaning estimator (8). Then, the coefficients’ covariance matrix estimator is readily calculated by applying the standard OLS formulas which, in the R language, go through the vcov() command.
    The phtest command computes the Hausman test in plm. Its main arguments are the two-panel model objects that underlie the comparison (ex. model = within and model = random). The corresponding estimates of the asymptotic covariances matrices provided under both models are thus used to compute the Hausman statistic corresponding to H M O 2 .
  • EViews estimates the random effects models using feasible GLS. The first step refers to the estimation of the covariance matrix for the composite error formed by the effects and the idiosyncratic disturbance. The EViews 9 User’s Guide II notes “Once the component variances have been estimated, we form an estimator of the composite residual covariance, and then GLS transform the dependent and regressor data”. As for the computation of the FGLS estimate, Eviews uses the quasi-demeaned model specification and proceeds on this basis. However, the calculation of the related coefficients’ covariance matrix is based on the direct application of the formula (11) and thus corresponds to v a ^ r 1 [ β ^ F G L S ] . The procedure for the Hausman test then corresponds to H M O 1 .
  • MATLAB provides estimation methods for the standard fixed (Within), between, and random effects models with the the panel data toolbox. Panel data models are estimated using the panel(·) function with the options argument set to re for the random effects model specification. The random effects FGLS estimates are based on the quasi-demeaned model and the asymptotic variance-covariance matrix for statistical inference is accordingly provided by v a ^ r 2 [ β ^ F G L S ] (see Equation (18) in Alvarez et al. (2017)). Then, hausmantest computes the Hausman test where the input of the hausmantest function requires the output structures of the two estimations to be compared. Accordingly, the statistics that are computed correspond to H M O 2 .
  • GAUSS: The GAUSS Times Serie∞s MT 3.0 TSMT provides a fixed effects and random effects models (TSCS) package that can be implemented through the tsmt library and the one-in-all tscsFit procedure. Another possibility is to use the pdlib GAUSS library and the randomEffects procedure in it. Both procedures implement the quasi-demeaning transformation on the original dataset and apply the standard OLS estimator on the transformed data so as to form the FGLS estimate. The covariance matrix estimate comes as a direct by-product of the OLS outcome so that v a ^ r 2 [ β ^ F G L S ] is used. The Hausman test provided in the tscsFit procedure is implemented accordingly and corresponds to H M O 2 .
  • SAS (SAS ETS 13.2) provides estimation methods for the standard fixed (Within), between, and random effects models in the balanced and unbalanced cases with the PANEL procedure toolbox. Standard panel data models are estimated using the PROC PANEL command with the MODEL statement specifying the regression model and the assumptions for the error structure. Specifically, FIXONE and RANONE must be used to specify the fixed-effect and the random-effect models, respectively (in the cross-sectional one-way case). In the latter case, various methods (but not the Swamy-Arora approach) are proposed to estimate, in the first stage, the variance components (through the VCOMP = option). It is explicitly indicated that the random effects FGLS estimates are then based, in the balanced case, on these variance components estimates through the quasi-demeaning approach, where ‘the random effects  β  is then the result of simple OLS on the transformed data’ (see SAS ETS 13.2 User Manual (2014), p. 1417). The estimator for the asymptotic variance-covariance matrix is thus provided by v a ^ r 2 [ β ^ F G L S ] . The Hausman statistic is automatically generated and reported as a conventional F statistic, with the statistic computed as H M O 2 .

4.2. Discussion

As the previous review indicates, in all but one of the packages discussed above, the Hausman test is, by default, implemented through the computation of H M O 2 with the quasi-demeaning estimator. The rationale for such a choice is computational. Indeed, the quasi-demeaning approach allows avoiding the inversion of the ( N T × N T ) matrix Ω ^ or Ω C * , which can be computationally costly (in terms of time and rounding errors). Conversely, the quasi-demeaned model only requires partially demeaning the variables with θ 1 ψ ^ as the partial demeaning factor. Yet, as a counterpart of this standard OLS regression, σ ^ ε * 2 is naturally chosen to compute the residual variance estimate and, in turn, yields to H M O 2 , which might, as we have seen, be an unreliable statistic for the Hausman test. In what follows, we explore some ways to circumvent the problems posed by the use of this statistic.
(1) First, it is possible to compute H M O 1 and still rely on the quasi-demeaning approach to estimate the parameters of the RE model (which, as mentioned before, is the default case in the vast majority of available econometric software). This can been easily seen from the relationship (21) that we established between h and H M O 1 . Once h has been determined, which only requires the OLS residual sums of squares from the estimation of the Within and the quasi-demeaning estimator, we can derive H M O 1 . Hence, the following procedure can be suggested, if required, to supplement the existing programs.
  • Use the quasi-demeaning estimator to compute the RE estimator for β and σ ^ ε * 2 .
  • Use σ ^ ε * 2 and σ ^ ε w 2 (within regression) to compute h.
  • Rearranging (21), obtain H M O 1 from h as: H M O 1 = N T K + 1 · h 1 + K .
  • Implement the Hausman test on the basis of H M O 1 .
(2) Second, depending on the software packages considered, some programming options can be used to fix the potential ‘variance disconnect’ problem associated with the use of the statistic H M O 2 .
For example, in STATA, it is possible to use the sigmamore and/or sigmaless option commands when implementing the Hausman test. As indicated in STATA instructions: “sigmamore and sigmaless specify that the two covariance matrices used in the test be based on a common estimate of disturbance variance. sigmamore specifies that the covariance matrices be based on the estimated disturbance variance from the efficient estimator. sigmaless specifies that the covariance matrices be based on the estimated disturbance variance from the consistent estimator”. Following the lines of Hausman’s seminal approach would lead to the choice of sigmaless option, whereby the variance estimator, is based on the Within model, would be used18. This would ensure the test to be performed upon the H M 0 1 statistics. The choice of sigmamore19 would imply to consider a third test statistics, H M 0 3 , where the common disturbance variance estimator would be based on the quasi-demeaned model, so that we would have: H M O 3 = β ^ Δ W R E · σ ^ ε * 2 · Γ 1 · β ^ Δ W R E . Comparing H M O 3 and H M O 1 , we observe that H M 0 1 > H M 0 3 whenever h < 1 . Thus, the more likely it would be to favor (even unduly) the rejection of the null hypothesis on the basis of H M 0 3 when h > 1 .
Some packages also offer the possibility to rely on an alternative expression for the Hausman statistic that does not involve the use of the RE estimator, so that it is immune to the variance disconnect problem. This expression was initially proposed by Hausman and Taylor (1981) and is based on the difference between the Between and Within estimators q ^ * * β ^ W β ^ B . Hausman and Taylor (1981, pp. 1382–83), establish that the resulting version of the Hausman statistic is numerically exactly identical to the one that is built upon q ^ and used above, that is: q ^ · var q ^ 1 · q ^ = q ^ * * · var q ^ * * 1 · q ^ * * . Such a solution can be notably implemented in the R PLM package using the phtest command and specifying as arguments model = within and model = between.
(3) Finally, two other approaches that depart from the separate estimation of the FE and RE model parameters - which underlies the standard implementation of the Hausman test - can be emphasized. They have the advantage of solving the disturbance variance estimator disconnect problem, while allowing, more generally, for a robust implementation of the Hausman test20.
(3.1) The first of these approaches relies on implementing an auxiliary regression, that was initially proposed by Hausman himself together with the presentation of the standard specification test (see also Mundlak 1978). This regression takes the following form:
y ˜ * * = Z ˜ * * · δ + X W · γ + η
with Z ˜ * * ω ˜ * * X ˜ * * where ω ˜ * * ψ ^ · ι N T and X ˜ * * Ω * ^ 1 2 · X ; y ˜ * * Ω * ^ 1 2 · y and η a vector of standard random disturbances.
It can be shown that the formula of the standard Wald test statistic for testing whether γ = 0 in the previous regression framework is equivalent to the one of the standard Hausman test statistic as expressed in terms of the difference between the Between and Within estimators (see Hausman and Taylor 1981 and above)21. Resorting to this auxiliary regression framework has two advantages. First, it involves only one estimator for the covariance matrix in the Wald test statistic formula, that one for γ , which is immune to the positive definiteness problem that can be encountered with the standard Hausman test statistic. Second, and as underlined by Baltagi and Liu (2007), it can be made robust to heteroskedasticity of unknown form (see, also, Arellano 1993).22 Once the variables have been transformed to be included as regressors in the auxiliary regression framework, the latter can be implemented in a rather standard way in any of the software econometric packages we have reviewed supra.
(3.2) The second approach goes through implementing White (1982)’s reformulated Hausman specification test that is based on the Maximum Likelihood (ML) estimation of the FE and RE model parameters. The related test statistic takes the form:
H W = n · β ˜ M L E W β ˜ M L E R E · S ˜ · β ˜ M L E W β ˜ M L E R E
with n N · T and β ˜ M L E W (resp. β ˜ M L E R E ) denoting the ML estimator related to the Within (resp RE) regression framework; S ˜ would serve as the covariance matrix estimator and involves the information matrices for both estimators of β , as well as outer products of scores within and between the two models under concern. White (1982) shows that S ˜ remains positive definite even under misspecification (including heteroskedasticity).
The last two procedures we have presented could even be suggested to be used in the first place when assessing the relevance of the RE-model specification as they allow for globally robust implementation of the Hausman test in the context of panel data models (if only, insofar as they do not require to be directly based on the use of the RE (FGLS) disturbance variance estimator).

5. Conclusions

In this paper, we provide new analytical results of the behavior of the Hausman statistic for the test of orthogonality between the individual effects and the error term in a static and balanced panel data model. We compare the Hausman statistic computed with direct FGLS implementation and the Hausman statistic computed on the quasi-demeaned model. We show that this difference depends upon several parameters; in particular, the between-within structure of the regressors. We show by means of a Monte Carlo simulation in the single regressor case and of a set of well-known textbook examples that the difference can be substantial and that in some cases, the Hausman statistic computed on the basis of the quasi-demeaned model can yield strong negative values. Therefore, despite its computational advantage, the quasi-demeaned model should not be used prima facie as the basis of the computation of the Hausman statistic. We suggest, if needed, to supplement the existing software instructions so as to be able to compute in any case the relevant statistic. Extensions can include deriving these analytical results for unbalanced panel data models, two-way component models and dynamic panel models.

Author Contributions

Conceptualization, M.-A.S.; methodology, J.L.G. and M.-A.S.; validation, J.L.G.; formal analysis, M.-A.S.; investigation, M.-A.S.; resources, J.L.G.; data curation, J.L.G. and M.-A.S.; writing—original draft preparation, M.-A.S.; writing—review and editing, J.L.G. and M.-A.S.; visualization, J.L.G.; supervision, J.L.G.; project administration, M.-A.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data are available upon request from the authors.

Acknowledgments

The authors thank the referees for their very useful comments and their careful reading of the first version submitted, which allowed to improve the paper. They also gratefully acknowledge the precious help of Alain Bachelot about the analysis of the definiteness of some matrices in the paper.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Hausman Test Statistics

Appendix A.1. Matrix Conditions

We first prove the following lemma (denoted below as Lemma A1):
Lemma A1.
Let R and M be two (square) matrices of size K. We assume that R and ( R · M ) are symmetric positive definite (SPD). Then, M can be diagonalized and there exists a non singular matrix P such that:
P · R · P = I
and
P · R · M · P = d i a g [ λ i ( M ) ]
with λ i ( M ) denoting the i-th eigenvalue of matrix M and d i a g [ λ i ( M ) ] the diagonal matrix whose elements are the eigenvalues of matrix M  ( i = 1 , K ) .
Proof. 
According to the spectral theorem (see Axler 2014), there exists a non-singular matrix O R such that
O R 1 · R · O R = d i a g [ λ i ( R ) ] , O R = O R 1 , 0 < λ i ( R ) , i
with λ i ( R ) denoting the i-th eigenvalue of matrix R and d i a g [ λ i ( R ) ] the diagonal matrix whose elements are the eigenvalues of matrix R  ( i = 1 , K ) .
It follows that:
d i a g 1 λ i ( R ) · O R · R · O R · d i a g 1 λ i ( R ) = I
so that:
Q · R · Q = I , with Q O R · d i a g 1 λ i ( R )
We then consider the symmetric (positive definite) matrix C, defined as:
C R · M
Consider in turn the matrix Q · C · Q that is symmetric. The spectral theorem ensures that there is an non-singular matrix O C that satisfies:
O C 1 · Q · C · Q · O C = d i a g [ λ i ( Q · C · Q ) ] with O C 1 = O C
Define the following matrix:
P Q · O C
(A4) ensures that
P · C · P = d i a g [ λ i ( Q · C · Q ) ] .
We have:
P · R · P = O C · Q · R · Q · O C = O C · O C = I
which leads to (A1).
Taking into account this result and refering on (A6), we see that:
d i a g [ λ i ( Q · C · Q ) ] = P · R · M · P = P · R · ( P · P 1 ) · M · P = ( P · R · P ) · P 1 · M · P = P 1 · M · P
This shows that M can be diagonalized and that
d i a g [ λ i ( M ) ] = d i a g [ λ i ( Q · C · Q ) ] = P · R · M · P
which is (A2). □

Appendix A.2. Applications

We obtain the following properties and results:
  • We first note that Γ ( X W · X W ) 1 ( X ˜ * · X ˜ * ) 1 and Γ ˜ ( X W · X W ) 1 h · ( X ˜ * · X ˜ * ) 1 are two symmetric (real valued) matrices. This comes from the definition of ( X ˜ * · X ˜ * ) which is computed as the sum of two symmetric matrices, ( X ˜ * · X ˜ * ) = ( X W · X W ) + ψ ^ 2 · ( X B C · X B C ) .
  • We have23  ( X ˜ * · X ˜ * ) ( X W · X W ) and, in turn24, ( X W · X W ) 1 ( X ˜ * · X ˜ * ) 1 . It follows that Γ is a symmetric positive definite (SPD) matrix.
As defined in the main text, H * I K + ψ ^ 2 · ( X B C · X B C ) · ( X W · X W ) 1 while X ˜ * · X ˜ * = H * · ( X W · X W ) . Then:
  • Note R = ( X W · X W ) 1 · H * 1 and M = H * . By construction R and ( R · M ) are two real positive symmetric definite matrices (this is because R · M = ( X W · X W ) 1 and that R can be written as R = [ ( X W · X W ) + ψ ^ 2 · ( X B C · X B C ) ] 1  which is the inverse of the sum of two symmetric positive definite matrices).
    We then deduce from Lemma1 that H * can be diagonalised and that there exists a non singular matrix P such that:
    d i a g [ λ i ( H * ) ] = P · X W · X W 1 · H * 1 · H * · P = P · X W · X W 1 · P
    As [ P · X W · X W 1 · P ] is SPD (given that X W · X W 1 is itself SPD), this ensures that the spectrum of H * is only composed of strictly positive elements.
  • Let R = ( X W · X W ) 1 and M = H * 1 . By construction, R and R · M are again two real symmetric positive matrices. We deduce from Lemma1 that H * 1 can be diagonalised and that there exists a non singular matrix P such that:
    P · X W · X W 1 · P = I P · X W · X W 1 · H * 1 · P = d i a g [ λ i ( H * 1 ) ] = d i a g [ λ i 1 ( H * ) ]
  • Noticing that Γ = X W · X W 1 · H * I K · H * 1 = X W · X W 1 X W · X W 1 · H * 1 , then, on the basis of the results above, we can write P · Γ · P as:
    P · Γ · P = d i a g 1 1 λ i ( H * )
    Then, let x denote a non-null vector, and setting y P 1 · x , we obtain the spectral decomposition of Γ as:
    x · Γ · x = y · P · Γ · P · y = i 1 1 λ i ( H * ) · y i 2
    Since we know that Γ is SPD, we obtain from the latter decomposition that i , 1 1 / ( λ i ( H * ) ) > 0 . which implies that h min * > 1 .
  • Further, noticing that Γ ˜ = ( X W · X W ) 1 · ( H * h · I K ) · H * 1 = ( X W · X W ) 1 h · ( X W · X W ) 1 · H * 1 and proceeding as for above, we can write ( P · Γ ˜ · P ) as:
    P · Γ ˜ · P = d i a g 1 h λ i ( H * )
    Let x denote a non-null vector, and setting y P 1 · x , we obtain the spectral decomposition of Γ ˜ as:
    x · Γ ˜ · x = y · P · Γ ˜ · P · y = i 1 h λ i ( H * ) · y i 2
    From this spectral decomposition, we conclude that
    Γ ˜ will be SPD if and only if i , 1 h λ i ( H * ) > 0 , that is, if and only if h < h min * is fulfilled.
    Γ ˜ will be SND if and only if i , 1 h λ i ( H * ) < 0 , that is, if and only if h > h max * is fulfilled.
We now use the previous results to compare H M O 1 and H M O 2 . For that purpose, we must distinguish according to whether H M O 2 is (for sure) a positive or (for sure) a non-positive quadratic form. This does in turn depend on whether Γ ˜ is SPD or SND.
  • If Γ ˜ is SPD, the relevant comparison can be built on the magnitude H M O 1 H M O 2 . The former can be written as:
    H M O 1 H M O 2 = β ^ Δ W R E · σ ^ ε w 2 1 · Γ 1 Γ ˜ 1 · β ^ Δ W R E = σ ^ ε w 2 1 · β ^ Δ W R E · Ξ · β ^ Δ W R E
    with Ξ Γ 1 Γ ˜ 1 .
    Thus, H M O 1 H M O 2 whenever Ξ is SPD or SND. Given the definition of Ξ , and since Γ ˜ is SPD, this, in turn, depends on whether Γ ˜ Γ is SPD or SND.
    Then, observe that Γ ˜ Γ = ( 1 h ) · ( X W · X W ) 1 · H * 1 .
    As ( X W · X W ) 1 · H * 1 is SPD, it follows that whether Γ ˜ Γ is SPD or SND depends on whether h 1 .
  • If Γ ˜ is SND, the relevant comparison is between H M O 1 and H M O 2 , and can be built on H M O 1 H M O 2 . The former magnitude can be written as:
    H M O 1 H M O 2 = β ^ Δ W R E · σ ^ ε w 2 1 · Γ 1 Γ ˜ * 1 · β ^ Δ W R E = σ ^ ε w 2 1 · β ^ Δ W R E · Ξ * · β ^ Δ W R E
    where Γ ˜ * = Γ ˜ and Ξ * Γ 1 Γ ˜ * 1 .
    Thus, H M O 1 H M O 2 whenever Ξ * is SPD or SND. Given the definition of Ξ * , and as Γ ˜ * is SPD (since Γ ˜ is SND), this, in turn, depends on whether Γ Γ ˜ * is SND or SPD.
    Then, observe that Γ Γ ˜ * = Γ + Γ ˜ = 2 · ( X W · X W ) 1 1 + h · ( X W · X W ) 1 · H * 1 .
    Proceeding as for Γ ˜ above, we can write ( P · Γ + Γ ˜ · P ) as:
    P · Γ + Γ ˜ · P = d i a g 2 1 + h λ i ( H * )
    Note by x a non-null vector, and setting y P 1 · x , we obtain the spectral decomposition of ( Γ + Γ ˜ ) as:
    x · ( Γ + Γ ˜ ) · x = y · ( P · ( Γ + Γ ˜ ) · P ) · y = i ( 2 1 + h λ i ( H * ) ) · y i 2
    From this spectral decomposition, we conclude that:
    ( Γ + Γ ˜ ) will be SPD if and only if i , 2 1 + h λ i ( H * ) > 0 , that is, if and only if h < 2 · h min * 1 .
    Conversely, ( Γ + Γ ˜ ) will be SND if and only if i , 2 1 + h λ i ( H * ) < 0 , that is, if and only if h > 2 · h max * 1 .
Hence, we obtain the following table covering all cases:
Table A1. Review of all cases.
Table A1. Review of all cases.
h < h min *
( Γ ˜ is SPD and H M O 2 > 0 )
0 < h < 1 < h min * Γ ˜ Γ H M O 1 > H M O 2
1 < h < h min * Γ Γ ˜ H M O 2 > H M O 1
h min * < h < h max *
( Γ ˜ and sign of H M O 2 a priori indefinite)
1 < h min * < h < h max * Γ ˜ indefinite H M O 1 H M O 2
h max * < h
( Γ ˜ is SND and H M O 2 < 0 )
1 < h max * < h < 2 · h min * 1 Γ Γ ˜ H M O 2 > H M O 1
2 · h min * 1 < h < 2 · h max * 1 Γ + Γ ˜ indefinite H M O 2 H M O 1
2 · h max * 1 < h Γ ˜ Γ H M O 1 > H M O 2

Appendix B. Relation between Within Residuals and Quasi-Demeaned Residuals

Preliminary Result: We first show that:
u ˜ ^ * * = u ^ W + ψ ^ · u ^ B + G · β ^ W β ^ B
Proof. 
We start with the definition of u ˜ ^ * * , the N T vector of the OLS residuals in the FQDM regression model. We have:
u ˜ ^ * * = y ˜ * * ω ˜ * * · α ^ F G L S X ˜ * * · β ^ F G L S
with ω ˜ * * ψ ^ · ι N T ; X ˜ * *   Ω * ^ 1 2 · X ; y ˜ * * Ω * ^ 1 2 · y . Using the definition of Ω * ^ 1 2 W + ψ ^ · B , we can rewrite (A8) as:
u ˜ ^ * * = u ^ W + ψ ^ · u ^ B + X W · β ^ W β ^ F G L S + ψ ^ · ι N T · α ^ B α ^ F G L S + ψ ^ · X B · β ^ B β ^ F G L S
where u ^ B denotes the N T vector of the OLS regression residuals for the Between (transformed) regression model; X B B · X and β ^ B is the OLS estimator of β in the Between transformed model which can be expressed as β ^ B = ( X B C · X B C ) 1 · ( X B C · y )  with X B C B C · X and (as a reminder) B C denoting the centered Between operator defined as B C B B N T with B N T ι N T · ( ι N T · ι N T ) 1 · ι N T .
Then, noting that α ^ J = ( ι N T · ι N T ) 1 · ι N T · ( y X · β ^ J ) for J = B ,   F G L S , which leads to: ι N T · ( α ^ B α ^ F G L S ) = B N T · X · ( β ^ F G L S β ^ B ) and using the following relationship between β ^ W , β ^ B and β ^ F G L S :
β ^ F G L S = Π · β ^ B + I K Π · β ^ W
with Π ψ ^ 2 · [ X · Ω C * ^ 1 · X ] 1 · ( X B C · X B C ) = ψ ^ 2 · [ X ˜ * · X ˜ * ] 1 · ( X B C · X B C )  and, as a reminder, Ω C * ^ 1 ( W + ψ ^ 2 · B C ) , we can rewrite equation (A9) as:
u ˜ ^ * * = u ^ W + ψ ^ · u ^ B + G · β ^ W β ^ B
with:
G X W · Π ψ ^ · B B N T · X · I K Π = X W · Π ψ ^ · X B C · I K Π
Other results: Define Π * and G * such that Π = ψ ^ · Π * and G = ψ ^ · G * . We have:
G * = X W · Π * X B C · I K ψ ^ · Π *
and replacing in (A11):
u ˜ ^ * * = u ^ W + ψ ^ · u ^ B + G * · β ^ W β ^ B
Using the property according to which B C · W = W · B C = 0 , we deduce, after manipulations, from (A12), the following relationship between the two residual sums of squares u ^ W · u ^ W and u ˜ ^ * * · u ˜ ^ * * :
u ˜ ^ * * · u ˜ ^ * * = u ^ W · u ^ W + ψ ^ 2 · u ^ B · u ^ B + Δ *
which corresponds to the equation provided in the main text in Section 3.4.1 and where:
Δ * ψ ^ 2 · β ^ W β ^ B · Σ * · β ^ W β ^ B
and25:
Σ * G * · G * = X B C · X B C · X ˜ * · X ˜ * 1 · X W · X W
Finally, noting from (A10) that: β ^ W β ^ B = Π 1 · [ β ^ W β ^ F G L S ] , we can rewrite (A13) as:
Δ * = ψ ^ 2 · β ^ W β ^ F G L S · Π 1 · Σ * · Π 1 · β ^ W β ^ F G L S
Define Σ ˜ * ( Π 1 ) · Σ * · Π 1 . Using the definition of Σ *  and Π , this can be written as:
Σ ˜ * = ψ ^ 4 1 · X W · X W · X B C · X B C 1 · X ˜ * · X ˜ * = ψ ^ 2 1 · Γ 1
with Γ 1 being written as: Γ 1 = ( ψ ^ 2 ) 1 · ( X W · X W ) · ( X B C · X B C ) 1 · ( X ˜ * · X ˜ * ) .
Replacing in (A14), we have:
Δ * = β ^ W β ^ F G L S · Γ 1 · β ^ W β ^ F G L S
which is the expression provided in the main text for Δ * .

Appendix C. Simulations in the Single Regressor Case ( K = 1 )

In the simulations for K = 1 , we focus on the role played by the structure in X and in u on the values taken by h and h * as well as by the Hausman statistics H M O 1 and H M O 2 . Those structural features are captured, respectively, by s x 2 and θ W on the one hand and by σ u 2 and ρ u on the other hand where σ u 2 σ α 2 + σ ε 2 and ρ u σ α 2 σ u 2 ( ρ u is called the intra-class correlation coefficient in the variance component literature). We let also the degree of correlation between the regressor and the (composite) error term on the cross-sectional dimension vary across the experiments (we call it ρ x u ).
Accordingly, for a given value of s x 2 ; θ W ; σ u 2 ; ρ u ; ρ x u as well as for a given size of the sample (N and T given), we generate (with replications) one series for u = u i t i = 1 , , N t = 1 , , T and one for x = x i t i = 1 , , N t = 1 , , T and consequently one series for y = y i t i = 1 , , N t = 1 , , T ), on the basis of which the different estimation procedures and the computation of the statistics of interest can be implemented.

Appendix C.1. Generating u it and x it

Each series for x and u is built such that the (sample) values of their total variances and Within variance shares in the total variance correspond to the ones that we set beforehand. To compute those series we adopt Nerlove’s approach (Nerlove 1971) and proceed in two steps.
  • First, we draw N pairs for the random vector Z b ( Z b x Z b u ) in the bivariate normal distribution N ( 0 , Σ Z b ) with: Σ Z b σ Z b x 2 σ Z b x u 2 σ Z b x u 2 σ Z b u 2 , σ Z b x u 2 = ρ x u · σ Z b x 2 · σ Z b u 2 ; σ Z b x 2 = 1 θ W · σ X 2 and σ Z b u 2 =   ρ u · σ u 2 .
  • Second, for each i = 1 , , N , we draw T pairs for the random vector Z w ( Z w x Z w u ) T in the bivariate normal distribution N ( 0 , Σ Z w ) with Σ Z w σ Z w x 2 0 0 σ Z b u 2 with σ Z w x 2 = θ W · σ X 2 and σ Z w u 2 = 1 ρ u · σ u 2 .
Then, the following variables are built:
x B i = σ Z b x 2 · z b i x z ¯ b x v a r z b x ; u B i = σ Z b u 2 · z b i u z ¯ b u v a r z b u x W i t = σ Z w x 2 · z w i t x z ¯ w i x v a r z w i x ; u W i t = σ Z w u 2 · z w i t u z ¯ w i u v a r z w i u i = 1 , , N
with for  j = x , u : z ¯ b j = 1 N · i = 1 i = N z b i j z ¯ w i j = 1 T · t = 1 t = T z w i t j ; v a r z b j = 1 N · i = 1 i = N z b i j z ¯ b j 2 ; v a r z w i j = 1 T · i = 1 i = N z w i t j z ¯ w i j 2 .
Finally we compute x i t as x i t = x B i + x W i t and u i t as u i t = u B i + u W i t .
This design ensures that s x 2 = σ X 2 , s u 2 = σ u 2 , θ W = x W 2 x T 2 and ρ u = 1 u T · W · u u T · u with s x 2 (resp. s u 2 ) denoting the sample variance for x (resp. u ).

Appendix C.2. Shaping the Experiment

We generate the different series for y i t i = 1 , , N t = 1 , , T according to the following relationship:
y i t = α + β · x i t + u i t
In the simulations, we set α = β = 1 . We let the other parameters of the simulation vary: N = ( 20 , 40 , 80 ) , T = ( 20 , 40 , 80 ) , x T 2 = ( 0.01 , 1 , 100 ) , θ W = ( 0.1 , 0.5 , 0.9 ) , σ u 2 = ( 0.01 , 1 , 100 ) , ρ x u = ( 0.1 , 0.5 , 0.9 ) and ρ u = ( 0.1 , 0.5 , 0.9 ) .

Appendix C.3. ANCOVA Results

Table A2. ANCOVA on simulation results.
Table A2. ANCOVA on simulation results.
Dependent Variable:
Mean HM 0 1
(1)
Mean HM 0 2
(2)
Median HM 0 1
(3)
Median HM 0 2
(4)
N = 40 (ref = 20)−24.158134.826 **−24.132136.294 **
(37.475)(56.029)(36.677)(55.435)
N = 80 (ref = 20)−72.354 *405.306 ***−72.281 **409.057 ***
(37.475)(56.029)(36.677)(55.435)
T = 40 (ref = 20)63.206 ***111.042 ***60.953 ***110.473 ***
(16.759)(25.057)(16.402)(24.791)
T = 80 (ref = 20)126.892 ***306.754 ***122.040 ***303.316 ***
(16.759)(25.057)(16.402)(24.791)
x T 2 = 1 (ref = 0.01)0.039−1.1050.322−0.300
(16.759)(25.057)(16.402)(24.791)
x T 2 = 100 (ref = 0.01)0.2772.7840.803−0.171
(16.759)(25.057)(16.402)(24.791)
θ W = 0.5 (ref = 0.1)90.452 ***−18.13483.656 ***−16.453
(29.028)(43.400)(28.410)(42.940)
θ W = 0.9 (ref = 0.1)184.362 ***−35.323167.161 ***−30.374
(29.028)(43.400)(28.410)(42.940)
ρ x u = 0.99 (ref = 0)1083.156 ***640.633 ***1057.507 ***643.296 ***
(25.600)(38.275)(25.055)(37.869)
ρ x u = 0.9 (ref = 0)147.213 ***214.549 ***141.792 ***202.157 ***
(25.600)(38.275)(25.055)(37.869)
ρ x u = 0.5 (ref = 0)13.48114.88212.5145.921
(25.600)(38.275)(25.055)(37.869)
ρ x u = 0.5 (ref = 0)13.4363.66412.4695.878
(25.600)(38.275)(25.055)(37.869)
ρ x u = 0.9 (ref = 0)147.237 ***215.027 ***141.974 ***201.985 ***
(25.600)(38.275)(25.055)(37.869)
ρ x u = 0.99 (ref = 0)1083.935 ***641.078 ***1058.341 ***643.762 ***
(25.600)(38.275)(25.055)(37.869)
σ u 2 = 1 (ref = 0.01)−0.2653.7230.3430.056
(16.759)(25.057)(16.402)(24.791)
σ u 2 = 100 (ref = 0.01)0.4000.7801.234−0.119
(16.759)(25.057)(16.402)(24.791)
ρ u = 0.5 (ref = 0.1)90.544 ***−20.71083.885 ***−15.796
(29.028)(43.400)(28.410)(42.940)
ρ u = 0.9 (ref = 0.1)184.478 ***−31.878167.171 ***−29.092
(29.028)(43.400)(28.410)(42.940)
N = 40 : θ W = 0.5 86.977 **−16.08087.205 **−15.864
(41.052)(61.376)(40.177)(60.726)
N = 80 : θ W = 0.5 261.702 ***−41.165262.250 ***−48.630
(41.052)(61.376)(40.177)(60.726)
N = 40 : θ W = 0.9 175.791 ***−27.656176.433 ***−30.434
(41.052)(61.376)(40.177)(60.726)
N = 80 : θ W = 0.9 525.795 ***−89.808526.255 ***−92.123
(41.052)(61.376)(40.177)(60.726)
N = 40 : ρ u = 0.5 86.826 **−13.05986.837 **−16.015
(41.052)(61.376)(40.177)(60.726)
N = 80 : ρ u = 0.5 261.064 ***−45.820261.054 ***−48.512
(41.052)(61.376)(40.177)(60.726)
N = 40 : ρ u = 0.9 175.736 ***−30.548176.043 ***−31.087
(41.052)(61.376)(40.177)(60.726)
N = 80 : ρ u = 0.9 526.388 ***−84.972526.704 ***−92.531
(41.052)(61.376)(40.177)(60.726)
Constant−446.938 ***−246.347 ***−430.229 ***−246.886 ***
(35.552)(53.154)(34.795)(52.590)
Observations5103510351035103
R20.5800.1650.5790.169
Adjusted R20.5780.1610.5770.165
Residual Std. Error (df = 5076)488.758730.740478.346722.994
F Statistic (df = 26; 5076)269.238 ***38.691 ***268.675 ***39.743 ***
Note: * p < 0.1; ** p < 0.05; *** p < 0.01.
Figure A1. Distribution of the proportion of negative values of H M O 2 , N = 80 , T = 80 , s x 2 = σ u 2 = 1 .
Figure A1. Distribution of the proportion of negative values of H M O 2 , N = 80 , T = 80 , s x 2 = σ u 2 = 1 .
Econometrics 11 00025 g0a1
Figure A2. Power of H M O 2 , T = 80 , s x 2 = σ u 2 = 1 .
Figure A2. Power of H M O 2 , T = 80 , s x 2 = σ u 2 = 1 .
Econometrics 11 00025 g0a2

Appendix D. Regression Outcomes in the Single Regressor Case for the Motivating Example 2 [Airline]

Table A3. Estimation results for Motivating example 2, Airline, single covariate.
Table A3. Estimation results for Motivating example 2, Airline, single covariate.
Specification Intercept log fuelprice
Within specificationCoef. 0.7785
Std Err. 0.0279
Between specificationCoef. 476.260 36.248
Std Err. 136.439 10.684
Random effect specificationCoef. 3.4264 0.7783
Std Err. 1 0.4245 0.0279
Std Err. 2 0.4502 0.0296
H M O 1 12.0100
H M O 2 0.0006
σ ^ ε 2 0.0456
σ ^ ε * 2 0.0513
ψ ^ 2 0.0095
h 1.1251
h * 1.0000
2 · h * 1 1.0000
Within variance 0.6521
Between variance 0.0005
Within variance/Total variance (in %) θ W 99.9282
Between variance/Total variance (in %) θ B 0.0718

Notes

1
Schreiber (2008), based on Holly (1982), examines cases in which this problem can arise even asymptotically, when the alternative hypothesis is true.
2
In particular, this issue is generally not addressed in the leading textbooks in panel data econometrics. An exception is Wooldridge (2010), (chp. 10, pp. 289–90), but he merely mentions the possibility of obtaining a non-positive definite covariance matrix if different estimates of the error term variance are used, suggesting a way out, which we further discuss below.
3
See Nerlove (1971) and Fuller and Battese (1973), for an original exposition of the transformed, quasi-demeaned model and the related approach.
4
Schreiber (2008) also considers a panel data framework as an illustration for the asymptotic results and highlights cases where the matrix is not SPD in a context where the error term variance estimates differ. He does not, however, focus on the comparison of the different estimation approaches in the random effects model and their implications for the computation of the Hausman test statistic, as we do.
5
We assume in what follows that there is no time-invariant regressor in X i t , so that it is possible to compute the β -estimator with the Within transformation of X i t (see below).
6
See the initial study by Baltagi and Griffin (1973). The dataset is available at: https://www.wiley.com/legacy/wileychi/baltagi/datasets.html, accessed on 2 May 2023.
7
Interestingly, in an updated version of his textbook, Baltagi (2021) modified the presentation of the Gasoline case study compared to the one provided in 2005 and presented here. In this update, there is no more disconnection between the two versions of the statistic, and only one version is considered, H M O 1 . While, in both presentations, the estimations are drawn from the STATA software package, the second presentation benefits from the use of the sigmaless option command that fixes the computation of the estimator for the idiosyncratic component of the error term. See infra in Section 4.2.
8
The original study is from Greene (1999). The dataset is available at: http://pages.stern.nyu.edu/~wgreene/Text/tables/tablelist5.htm, accessed on 2 May 2023.
9
The dataset is available at: http://pages.stern.nyu.edu/~wgreene/Text/Edition7/tablelist8new.htm, accessed on 2 May 2023.
10
A typical i t -observation for y * is given by y i t * = y i t θ ^ · y ¯ i . where θ 1 ψ ^ and y ¯ i . 1 T t = 1 t = T y i t . A similar transformation is applied for each of the components of X , hence the quasi-demeaning expression for the transformed model that is obtained in that way.
11
Indeed, β ^ O L S = ( X * * · ( I N T B N T ) · X * * ) 1 · ( X * * · I N T B N T · y * * ) , which is equivalent to (4) given the definition of X * * , y * * and Ω C * 1 .
12
Given the definition of those matrices, this replacement relies on the use of consistent estimators for the variance components, i.e., σ ε 2 , σ α 2 and/or σ α ε 2 . For a discussion about these variance component estimators, see, among others, Amemiya (1971); Fuller and Battese (1974); Maddala (1971); Nerlove (1971); Swamy and Arora (1972); Wallace and Hussain (1969).
13
In particular, we assume that Ω * ^ (resp. Ω C * ^ ) is a consistent estimator for Ω * (resp. Ω C * ). See Wooldridge (2010) for a discussion on these conditions.
14
The estimator for σ α ε 2 is usually computed from the sum of the squares of the OLS regression residuals, ( u ^ B · u ^ B ) , for the Between (transformed) regression model with σ ^ α ε 2 = u ^ B · u ^ B N K + 1 . (Swamy-Arora approach), see infra.
15
It can be easily checked that the two approaches give rise to the same (RE) estimator for the parameters, α and β .
16
The emphasis is added by us. The notation used by Hausman for the fixed-effects estimate of β ( β ^ F E ) corresponds to our β ^ W .
17
The intra-class coefficient drives the value of ψ 2 . It can be shown, indeed, that: ψ 2 = 1 ρ u 1 ρ u · 1 T .
18
Note that this estimator is also used to compute ψ ^ 2 and in turn Ω C * ^ , which makes it fully, logically consistent with respect to the FGLS regression model framework.
19
This is, e.g., recommended by Cameron and Trivedi (2009), p. 360.
20
We thank both referees for having highlighted those approaches and suggested to account for them in this discussion subsection.
21
In this case, indeed, γ ^ = q ^ * * .
22
Baltagi and Liu show in (Baltagi and Liu 2007) that the Hausman test can be obtained equivalently from other artificial regressions, involving the use of the set of Between-transformed regressor variables, X B , or even the set of the initial regressor variables, X . They also discuss the case where the auxiliary regression can accommodate the presence of potentially endogenous regressors. With respect to the issue of weak instruments in this context, see also (Staiger and Stock 1997).
23
Take two symmetric matrices A and B. We denote by A B the property according to which A B is a positive definite matrix (what we can also write as A B 0 ).
24
If A and B are two symmetric positive definite matrices and are non-singular, then A B B 1 A 1 .
25
As a reminder, note that ( X ˜ * · X ˜ * ) = ( X W · X W ) + ψ ^ 2 · ( X B C · X B C ) .

References

  1. Alvarez, Inmaculada C., Javier Barbero, and José L. Zofío. 2017. A Panel Data Toolbox for MATLAB. Journal of Statistical Software 76: 1–17. [Google Scholar] [CrossRef]
  2. Amemiya, Takeshi. 1971. The Estimation of the Variances in a Variance-Components Model. International Economic Review 12: 1–13. [Google Scholar] [CrossRef]
  3. Arellano, Manuel. 1993. On the testing of correlated effects with panel data. Journal of Econometrics 59: 87–97. [Google Scholar] [CrossRef]
  4. Axler, Sheldon. 2014. Linear Algebra Done Right. New York: Springer. [Google Scholar]
  5. Baltagi, Badi H. 2005. Econometric Analysis of Panel Data, 3rd ed. Chichester and Hoboken: J. Wiley & Sons. [Google Scholar]
  6. Baltagi, Badi H. 2021. Econometric Analysis of Panel Data, 6th ed. Springer texts in Business and Economics. Cham: Springer. [Google Scholar] [CrossRef]
  7. Baltagi, Badi H., and James M. Griffin. 1973. Gasoline demand in the oecd: An application of pooling and testing procedures. European Economic Review 22: 626–32. [Google Scholar] [CrossRef]
  8. Baltagi, Badi H., and Long Liu. 2007. Alternative ways of obtaining Hausman’s test using artificial regressions. Statistics & Probability Letters 77: 1413–17. [Google Scholar] [CrossRef]
  9. Baum, Christopher F., Mark E. Schaffer, and Steven Stillman. 2003. Instrumental Variables and GMM: Estimation and Testing. The Stata Journal: Promoting Communications on Statistics and Stata 3: 1–31. [Google Scholar] [CrossRef]
  10. Cameron, Adrian Colin, and Pravin K. Trivedi. 2009. Microeconometrics Using Stata. College Station: Stata Press. [Google Scholar]
  11. Cornwell, Christopher, and Peter Rupert. 2008. Efficient estimation with panel data: An empirical comparison of instrumental variable estimators. Journal of Applied Econometrics 3: 149–55. [Google Scholar]
  12. Croissant, Yves, and Giovanni Millo. 2008. Panel Data Econometrics in R: The plm Package. Journal of Statistical Software 27: 1–43. [Google Scholar] [CrossRef]
  13. Fuller, Wayne A., and George E. Battese. 1973. Transformations for Estimation of Linear Models with Nested-Error Structure. Journal of the American Statistical Association 68: 626–32. [Google Scholar] [CrossRef]
  14. Fuller, Wayne A., and George E. Battese. 1974. Estimation of linear models with crossed-error structure. Journal of Econometrics 2: 67–78. [Google Scholar] [CrossRef]
  15. Greene, William. 1999. Frontier Production Functions. In Handbook of Applied Econometrics Volume II: Microeconomics. Edited by Pesaran M. Hashem and Schmidt Peter. Oxford: Blackwell Publishing Ltd., pp. 75–153. [Google Scholar] [CrossRef]
  16. Greene, William. 2000. Econometric Analysis, 4th ed. Upper Saddle River: Prentice Hall Internat. [Google Scholar]
  17. Greene, William. 2012. Econometric Analysis, 7th ed. Pearson Series in Economics; Boston and Munich: Pearson. [Google Scholar]
  18. Hausman, Jerry A. 1978. Specification Tests in Econometrics. Econometrica 46: 1251–71. [Google Scholar] [CrossRef]
  19. Hausman, Jerry A., and William E. Taylor. 1981. Panel data and unobservable individual effects. Econometrica 49: 1377–98. [Google Scholar]
  20. Hayashi, Fumio. 2000. Econometrics. Princeton: Princeton University Press. [Google Scholar]
  21. Holly, Alberto. 1982. A Remark on Hausman’s Specification Test. Econometrica 50: 749–60. [Google Scholar] [CrossRef]
  22. Maddala, Gangadharrao Soundalyara. 1971. The Use of Variance Components Models in Pooling Cross-Section and Time Series Data. Econometrica 39: 341–57. [Google Scholar]
  23. Mundlak, Yair. 1978. On the pooling of time series and cross section data. Econometrics 46: 69–85. [Google Scholar]
  24. Nerlove, Marc. 1971. A Note on Error Components Models. Econometrica 39: 383–96. [Google Scholar] [CrossRef]
  25. Schreiber, Sven. 2008. The Hausman Test Statistic Can Be Negative even Asymptotically. Jahrbücher für Nationalökonomie und Statistik 228: 394–405. [Google Scholar] [CrossRef]
  26. Staiger, Douglas, and James H. Stock. 1997. Instrumental Variables Regression with Weak Instruments. Econometrica 65: 557–86. [Google Scholar] [CrossRef]
  27. Swamy, Paravastu Aananta Venkata Bhattandha, and Swarnjit S. Arora. 1972. The Exact Finite Sample Properties of the Estimators of Coefficients in the Error Components Regression Models. Econometrica 40: 261–75. [Google Scholar] [CrossRef]
  28. Wallace, T. Dudley, and Ashiq Hussain. 1969. The Use of Error Components Models in Combining Cross Section with Time Series Data. Econometrica 37: 55. [Google Scholar] [CrossRef]
  29. White, Halbert. 1982. Maximum Likelihood Estimation of Misspecified Models. Econometrica 50: 1–25. [Google Scholar] [CrossRef]
  30. Wooldridge, Jeffrey M. 2010. Econometric Analysis of Cross Section and Panel Data, 2nd ed. Cambridge, MA: MIT Press. [Google Scholar]
Figure 1. Distribution of the median of H M O 1 and H M O 2 , N = 80 , T = 80 , s x 2 = σ u 2 = 1 and ρ u = 0.5 .
Figure 1. Distribution of the median of H M O 1 and H M O 2 , N = 80 , T = 80 , s x 2 = σ u 2 = 1 and ρ u = 0.5 .
Econometrics 11 00025 g001
Table 1. Estimation results for motivating example 1, Gasoline.
Table 1. Estimation results for motivating example 1, Gasoline.
Specification Intercept log Y / N log P MG / P GDP log Car / N
Within/fixed effectsCoef. 0.6622 0.3217 0.6405
Std Err. 0.0734 0.0441 0.0297
BetweenCoef. 2.5416 0.9676 0.9635 0.7953
Std Err. 0.5268 0.1557 0.1329 0.0825
Random effectsCoef. 1.997 0.5550 0.4204 0.6068
Std Err._1 0.1782 0.0572 0.0387 0.0247
Std Err._2 0.1843 0.0591 0.0400 0.0255
H M O 1 26.49505
H M O 2 302.8037
σ ^ ε w 2 0.0085
σ ^ ε * 2 0.0091
ψ ^ 2 0.0116
h 1.069
h min * 1.0409
h max * 2.0837
Within variance 0.0507 0.0162 0.3089
Between variance 0.3508 0.4424 1.1725
With. var. share (in %) θ W 12.6255 3.5325 20.8518
Betw. var. share (in %) θ B 87.3745 96.4675 79.1482
Table 2. Estimation results for motivating example 2, Airline.
Table 2. Estimation results for motivating example 2, Airline.
Specification Intercept log Q log fuelprice loadfactor
WithinCoef. 0.9193 0.4175 1.0704
Std Err. 0.0299 0.0152 0.2017
BetweenCoef. 85.8087 0.7825 5.5239 1.7510
Std Err. 56.4830 0.1088 4.4788 2.7432
Random effectsCoef. 9.6279 0.9067 0.4228 1.0645
Std Err._1 0.2098 0.0256 0.0140 0.1998
Std Err._2 0.2102 0.0256 0.0140 0.2000
H M O 1 3.249
H M O 2 2.1247
σ ^ ε w 2 0.0036
σ ^ ε * 2 0.0036
ψ ^ 2 0.0152
h 1.0029
h min * 1.000
h max * 1.3690
Within variance 0.1751 0.6521 0.0021
Between variance 1.1340 0.0005 0.0007
With. var. share (in %) θ W 13.3777 99.9282 76.0391
Betw. var. share (in %) θ B 86.6223 0.0718 23.9609
Table 3. Estimation results for motivating example 2, Airline, two covariates.
Table 3. Estimation results for motivating example 2, Airline, two covariates.
Specification Intercept log fuelprice loadfactor
WithinCoef. 0.7445 0.8684
Std Err. 0.0384 0.6780
BetweenCoef. 419.760 32.304 10.964
Std Err. 136.203 10.541 8.880
Random effectsCoef. 3.3673 0.7393 0.9931
Std Err._1 0.4179 0.0384 0.6758
Std Err._2 0.4471 0.0411 0.7230
H M O 1 14.5905
H M O 2 0.2470
σ ^ ε w 2 0.0452
σ ^ ε * 2 0.0518
ψ ^ 2 0.0106
h 1.1447
h min * 1.0000
h max * 1.0066
Within variance (1) 0.6522 0.0021
Between variance (2) 0.00047 0.0007
With. var. share (in %) θ W 99.9282 76.0391
Betw. var. share (in %) θ B 0.0718 23.9609
Table 4. Estimation results for Motivating example 3, Wage.
Table 4. Estimation results for Motivating example 3, Wage.
Specification Intercept Exp Exp 2 WKS OCC IND SOUTH SMSA MS UNION
WithinCoef. 0.1132 0.0004 0.0008 0.0215 0.0192 0.0019 0.0425 0.0297 0.0328
Std Err. 0.0025 0.00005 0.0006 0.0138 0.0154 0.0343 0.0194 0.0190 0.014923
BetweenCoef. 5.7222 0.0275 0.0005 0.0089 0.3536 0.0460 0.1082 0.1815 0.3837 0.0891
Std Err. 0.1918 0.0053 0.0001 0.0040 0.0309 0.0282 0.0284 0.0283 0.0352 0.0324
Random effectsCoef. 5.4668 0.0838 0.0008 0.0012 0.1270 0.0194 0.0822 0.0030 0.0092 0.0374
Std Err._1 0.0417 0.0022 0.0001 0.006 0.0123 0.0134 0.0214 0.0157 0.0165 0.0133
Std Err._2 0.0554 0.0029 0.0001 0.0008 0.0164 0.0178 0.0284 0.0208 0.0219 0.01763
σ ^ ε w 2 0.0231 H M O 1 3177.583
σ ^ ε * 2 0.0407 H M O 2 7569.713
h 1.7626 ψ ^ 2 0.0368
h min * 1.0221
h max * 2.6757
h ˜ 2.0566
Within variance (1) 4 0.0087 15.5347 0.0300 0.0233 0.0048 0.0149 0.0155 0.0254
Between variance (2) 116.2324 0.0001 10.7666 0.2199 0.2157 0.2012 0.2114 0.1356 0.2061
With. var. share (in %) θ W 3.3269 3.3118 59.0643 11.9970 9.7561 2.3308 6.60679 10.2570 10.9640
Betw. var. share (in %) θ B 96.6731 96.6881 40.9357 88.0029 90.2439 97.6691 93.3932 89.7430 89.03604
Table 5. Review of all cases.
Table 5. Review of all cases.
h < h min *
( Γ ˜ is SPD and H M O 2 > 0 )
0 < h < 1 < h min * Γ ˜ Γ H M O 1 > H M O 2 > 0
1 < h < h min * Γ Γ ˜ H M O 2 > H M O 1 > 0
h min * < h < h max *
( Γ ˜ and sign of H M O 2  a priori indefinite)
1 < h min * < h < h max * Γ ˜ indefinite H M O 1 H M O 2
h max * < h
( Γ ˜ is SND and H M O 2 < 0 )
1 < h max * < h Γ ˜ negative definite H M O 1 > 0 > H M O 2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Le Gallo, J.; Sénégas, M.-A. On the Proper Computation of the Hausman Test Statistic in Standard Linear Panel Data Models: Some Clarifications and New Results. Econometrics 2023, 11, 25. https://doi.org/10.3390/econometrics11040025

AMA Style

Le Gallo J, Sénégas M-A. On the Proper Computation of the Hausman Test Statistic in Standard Linear Panel Data Models: Some Clarifications and New Results. Econometrics. 2023; 11(4):25. https://doi.org/10.3390/econometrics11040025

Chicago/Turabian Style

Le Gallo, Julie, and Marc-Alexandre Sénégas. 2023. "On the Proper Computation of the Hausman Test Statistic in Standard Linear Panel Data Models: Some Clarifications and New Results" Econometrics 11, no. 4: 25. https://doi.org/10.3390/econometrics11040025

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop