Next Article in Journal
Estimating Smoothness and Optimal Bandwidth for Probability Density Functions
Previous Article in Journal
A Semiparametric Tilt Optimality Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

Data Cloning Estimation and Identification of a Medium-Scale DSGE Model

1
Departament of Economics, Federal University of Santa Catarina, Florianópolis 88040-900, SC, Brazil
2
Department of Economics, School of Economics, Business Administration and Accounting at Ribeirão Preto (FEA-RP/USP), Av. dos Bandeirantes 3900, FEARP—University of São Paulo, Ribeirão Preto 14040-905, SP, Brazil
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
These authors acknowledge funding from Capes, CNPq (310646/2021-9) and FAPESP (2018/04654-9).
Stats 2023, 6(1), 17-29; https://doi.org/10.3390/stats6010002
Submission received: 4 December 2022 / Revised: 19 December 2022 / Accepted: 20 December 2022 / Published: 24 December 2022
(This article belongs to the Section Econometric Modelling)

Abstract

:
We apply the data cloning method to estimate a medium-scale dynamic stochastic general equilibrium model. The data cloning algorithm is a numerical method that employs replicas of the original sample to approximate the maximum likelihood estimator as the limit of Bayesian simulation-based estimators. We also analyze the identification properties of the model. We measure the individual identification strength of each parameter by observing the posterior volatility of data cloning estimates and access the identification problem globally through the maximum eigenvalue of the posterior data cloning covariance matrix. Our results corroborate existing evidence suggesting that the DSGE model of Smeets and Wouters is only poorly identified. The model displays weak global identification properties, and many of its parameters seem locally ill-identified.

1. Introduction

The data cloning (DC) methodology proposed by Lele et al. [1] is a numerical procedure for approximating maximum-likelihood estimates as the limit of Bayesian posterior simulation estimators when we repeat the original observed sample many times.
Because DC is applicable whenever we can employ Bayesian simulation techniques, such as Markov Chain Monte Carlo (MCMC), for estimating a model, the method has been adapted to many settings and applications, for example, Ponciano et al. [2], Baghishani and Mohammadzadeh [3], Torabi [4], Ponciano et al. [5], Torabi et al. [6], Picchini and Anderson [7], and Duan et al. [8]. We further discuss the DC method in Section 3.
Among econometricians, however, data cloning is not as popular. We only know of a few existing applications in the estimation of stochastic volatility models (Laurini [9], Marín et al. [10], de Zea Bermudez et al. [11]) A natural application of DC is in estimating Dynamic Stochastic General Equilibrium (DSGE) models. DSGE models take the form of nonlinear systems of expectational equations and play a central role in modern macroeconomic practice and research. Bayesian posterior simulation methods are often the preferred choice when estimating this wide class of models (Fernández-Villaverde [12]. Furlani et al. [13] use the DC method to estimate a small-scale DSGE model and report good properties with respect to sensitivity to initial values.
More complex DSGE models often display diverse problems regarding parameter identification (Canova and Sala [14]). Data cloning estimation has the convenient byproduct of uncovering parameter identification problems, which makes the method especially interesting to apply to DSGE models, since it becomes possible to, at the same time, estimate parameters and expose identification problems.
In this paper, we employ the DC method to estimate the DSGE model of Smets and Wouters [15] (SW) and then to assess parameter identification issues. SW is a medium-scale DSGE model with 36 free parameters, which are estimated from seven observed macroeconomic time series. Its general equilibrium structure is augmented with nominal rigidities and additional exogenous shocks. SW is a natural choice of subject for this analysis because it has been extensively studied and embodies all characteristic features of New Keynesian DSGE models. Indeed, much of the research on the identification of DSGE models, and even general critiques to modern macroeconomics, at times points to specific issues with the model of Smets and Wouters [15] (e.g., Canova and Sala [14], Chari et al. [16], Iskrev [17], Komunjer and Ng [18], Romer [19], and Chadha and Shibayama [20]).
The remainder of this paper is structured as follows. Section 2 discusses relevant estimation and identification issues with DSGE models, Section 3 overviews the data cloning idea and comments on some applications of the method, Section 4 details the setup of our exercise, Section 5 presents and discusses our results in relation to the literature, and Section 6 concludes.

2. Estimation and Identification of DSGE Models

A central focus of macroeconomic research effort, be it in academia or policymaker circles, is assessing the net effect of forces operating on different parts of the economy. DSGE models are the main tool employed for such task.
DSGE models are built up from microeconomic primitives, which makes them specially transparent with respect to underlying assumptions. Christiano et al. [21] argue that this openness is the main appeal of the DSGE approach, and what allowed diverse groups of researchers to make contributions to the field. While transparency is a great strength, it also makes DSGE models easy to criticize. Dubious and problematic assumptions can be singled out and scrutinized. Indeed, critics are plentiful, and we discuss some of them in relation to our results.
In this section, we follow some key steps when taking a DSGE model to data and extracting information about structural parameters; with each step, we highlight potential issues and discuss important results in the literature.

2.1. Estimation

As DeJong and Dave [22] put it, DSGE models start with a characterization of the environment in which decision makers reside, a set of decision rules that dictate their behavior, and a characterization of the uncertainty they face when making decisions. This structure takes the form of a nonlinear system of expectational equations, which are often linearized before solving.
Let z ¯ t be an m-dimensional vector of stationary variables, and z 🟉 be the steady state values of those variables such that z t = z ¯ t z 🟉 is the present deviation from the steady-state value. Following Iskrev [17], most log-linearized DSGE models can be represented as
Γ 0 ( θ ) z t = Γ 1 ( θ ) E t z t + 1 + Γ 2 ( θ ) z t 1 + Γ 3 ϵ t .
Since (1) depends on expectational components, it must be solved in terms of rational expectations. There are several algorithms for solving linear rational expectation models (e.g., Blanchard and Kahn [23], Anderson and Moore [24], King and Watson [25], Sims [26]). Assuming that a unique solution exists, the linearized DSGE model can be written as
z t = A ( θ ) z t 1 + B ( θ ) ϵ t .
Due to unobserved variables, the transition Equation (2) is augmented with the measurement equation
x t = C z t + D u t + ν t .
Equations (2) and (3) together form the state space representation of our DSGE model. Assuming that shocks ϵ t are jointly Gaussian, we can recover the likelihood function L ( · | θ ) of the data X through Kalman filter recursions. Then, the maximum likelihood estimate (MLE) of θ is
θ ^ M L = max θ Θ L ( X | θ ) .
The precision of the maximum-likelihood estimator θ ^ M L is given by the inverse of the Fisher information matrix
I ( X | θ ) = E L ( X | θ ) θ L ( X | θ ) θ .
In practice, obtaining θ ^ M L from the numerical optimization of L ( X | θ ) , as suggested in Equation (4), is not straightforward. The likelihood function of a typical DSGE model is an ill behaved object, with multiple local maxima/minima and flat surfaces. Due to those difficulties, Bayesian methods are often preferred when estimating DSGE models (Fernández-Villaverde [12]).
Let π ( θ ) denote the joint prior distribution of parameters θ . Bayes’ theorem implies the posterior distribution of the parameters given data is
π ( θ | X ) = L ( X | θ ) π ( θ ) Θ L ( X | θ ) π ( θ ) d θ .
The marginal likelihood of the denominator is constant over all parameter values, and in many applications it suffices to compute the nominator
π ( θ | X ) L ( X | θ ) π ( θ ) .
The common solution here is then to sample from π ( θ | X ) using a posterior-simulation Markov Chain Monte Carlo (MCMC) method [27]. Multiplying the likelihood by the prior distribution has the effect of attenuating the computational issues with optimizing the likelihood function that we alluded to before [28].

2.2. Identification

The classical result of identification in parametric models dates back to Rothenberg [29]. Local identifiability of a parameter vector is ensured by a non-singular Fisher information matrix. Komunjer and Ng [18] show this result does not apply to DSGE models. Because the reduced-form parameters of the state space representation are not generally identifiable, the non-singularity of the information matrix does not suffice to ensure local identification of the parameter vector.
Canova and Sala [14] frame the identification problem as having “to do with the ability to draw inference about the parameters of a theoretical model from an observed sample”. This broad definition of the identification problem is well suited for the DSGE context, as issues with parameter identification may arise in all stages from model specification to estimation from data, or, as the authors put it, from different “mappings” between the estimated parameter vector θ ^ and data X.
The “solution mapping” regards the relationship between structural parameters θ and the coefficients of matrices A ( θ ) and B ( θ ) . Common issues arise here when some individual parameter θ 1 disappears from the solution (under identification), or when two parameters θ 1 and θ 2 enter the solution only in relation to one another (partial identification). It is not uncommon for approximation methods to induce under-identification or partial identification problems. The SW model has two pairs of parameters that, due to log linear approximation, are only partially identified. Calvo [30] price reoptimization probabilities ξ π , ξ w feature in the solution only as functions of the Kimball curvature parameters ϵ π , ϵ w , respectively. Indeed, Smets and Wouters [15] fix ϵ π = ϵ w = 10 and estimate ξ π and ξ w accordingly.
Partial identification and under-identification are problems of the model, the objective function, and/or the approximation methods employed—therefore, a model that displays this sort of problem will display them no matter the data set at hand. Alternatively, we could say they involve only the “solution” and “objective function” mappings of Canova and Sala [14].
One fundamental identification problem arises when the likelihood function does not have a unique maximum. Thus, under the point of view of the objective function, two models, potentially having different economic interpretations, are indistinguishable, and information external to the model needs to be introduced in order to choose between structural forms. Observational equivalence is a well-known phenomenon in macroeconomics and has played part in many debates over the years. For example, Sargent [31] showed that it is difficult to distinguish between old Keynesian and neoclassical models using only parameters estimated from a single policy regime. Chari et al. [16] uses observationally equivalent, but structurally different, models to question to what extent shocks in the SW model are interpretable.
A fourth identification issue characteristic of the DSGE model is related to the accuracy with which some particular parameter θ can be estimated from finite samples. If the curvature of the likelihood function L ( X | θ ) around θ ^ is small, then large changes in θ ^ will have little impact on the value of the likelihood function. Canova and Sala [14] dub this phenomenon “weak identification” in allusion to the literature on weak instruments.
Iskrev [17] tackle the weak identification problem as a model feature, in both a mathematical and economic sense. He mentions that although the non-singularity of I ( X | θ ) ensures the model’s expected likelihood is not flat and achieves a locally unique maximum, the precision with which the true parameter θ may be estimated in finite samples depends on the degree of curvature of the likelihood around an open neighborhood of θ ^ , to which the rank condition provides no information. If the curvature of the log likelihood is nearly flat around θ ^ , then large changes in the estimate θ ^ will induce small relative changes in L ( X | θ ^ ) . The economic argument the author presents is that a weakly identified parameter is one that is either irrelevant, because it has only a negligible effect on the likelihood, or nearly redundant, because its effect on the likelihood may be closely approximated by other parameters, and thus the value of such parameters is hard to pin down on the basis of the information contained in the likelihood function alone.
Qu and Tkachenko [32] provide a characterization of necessary and sufficient conditions for the local identification of structural parameters in linearized DSGE models and propose a frequency domain quasi-maximum likelihood estimator. They show how in a quasi-Bayesian estimation scheme, the procedure can be used to access parameter identification properties. This method was employed by Tkachenko and Qu [33] to analyze the model of Smets and Wouters [15].
In a Bayesian estimation, we are interested in the posterior distribution π ( θ | X ) , which is proportional to the likelihood function L ( X | θ ) times the prior distribution π ( θ ) . As long as there is a proper prior, the posterior distribution is well defined. The extreme case is inference without data, where the posterior is the prior itself. In fact, as An and Schorfheide [28] discuss, priors “might add curvature to a likelihood function that is (nearly) flat in some dimensions of the parameter space and therefore strongly influence the shape of the posterior distribution”. In this sense, the presence of the prior distribution makes weak identification problems even harder to spot in real-world applications.
A naive Bayesian identification test used is to visually contrast prior and posterior shapes. Intuitively, if the posterior is very similar to the prior, then data have little to say about the parameter of interest. Canova and Sala [14] point out that this is only true if the parameter space is variation-free; i.e., there are no model-implied restrictions on the parameter space. DSGE models are seldom variation-free; they usually impose stability restrictions or economically motivated non-negativity constraints on some parameters. If the parameter space is not variation-free, then beliefs updates about one parameter might affect the parameter space of another, ultimately affecting the shape of the posterior distribution even if data are not informative.
Koop et al. [34] introduce two Bayesian identification indicators. The first indicator looks to the expected posterior shape relative to the prior and provides a yes-or-no answer to the identification problem, much in line with Iskrev [17] and Komunjer and Ng [18]. The second indicator is especially relevant to us because it is similar to the data cloning method in that both use a scheme to create artificial data and tackle weak identification problems computationally. The Bayesian Learning Rate indicator of Koop et al. [34] consists of augmenting the observed sample with model simulated path and then observing the standard error of MCMC estimates as the simulated sample size increases.
Morris [35] discusses the sampling distribution of estimators for DSGE parameters tends to be non-normal and/or pile up on the boundary of the theoretical support. This phenomenon is showcased in our exercise below.
More recently, Ivashchenko and Mutschler [36] combine the “a priori” rank identification criteria of Iskrev [17], Komunjer and Ng [18], and Qu and Tkachenko [32], with the simulation-based Bayesian Learning Rate indicator of Koop et al. [34], to study the identification properties of two traditional DSGE models. Their focus is on how different investment adjustment cost specifications and output-gap definition affect individual parameter identification. Qu and Tkachenko [37] discuss global identification properties of DSGE models while taking into account indeterminacy of monetary policy rules that can be (nearly) observationally equivalent to each other. An alternative approach to verify identification conditions in DSGE models was proposed by Meenagh et al. [38], in the context of the estimation of DSGE models using indirect inference. They verified the identification through Monte Carlo procedures, where artificial restrictions were progressively imposed on the [15] model parameters, and checked the rejection rate over many samples from the true model. With this method, identification problems are verified through the power of the tests to reject invalid parameters in the model; in the absence of identification, the power of these tests is low. A procedure to verify global identification was proposed in Kocięcki and Kolasa [39], where the condition for identification was formulated through the restrictions linking the observationally equivalent state space representations of the DSGE model and on the constraints imposed on the deep parameters by the model solution. Kocięcki and Kolasa [40] refine this method by using the concept of a Gröbner basis and new algorithms to analytically compute the complete set of observationally equivalent parameter vectors.

3. The Data Cloning Method

The data cloning (DC)methodology was developed by Lele et al. [1] and Lele et al. [41] as a computational scheme to obtain maximum-likelihood (ML) parameter estimates as the limit of Bayesian estimates when the original data are pooled many times. It was first designed to estimate complex hierarchical ecological systems and has been applied to a wide range of Generalized Linear Mixed models. Furlani et al. [13] demonstrate the possibility of estimating DSGE models through the DC method.
Original data X are repeated K times to obtain an artificially augmented, cloned dataset X ( K ) = ( X , , X ) . The new likelihood function based on the cloned data is given by [ L ( X | θ ) ] K . Then, the new posterior distribution is
π K ( θ | X ) = [ L ( X | θ ) ] K π ( θ ) Θ [ L ( X | θ ) ] K π ( θ ) d θ .
Lele et al. [1] and Lele [42] show that as the number of sample clones K grows larger, the DC posterior distribution converges to an approximately normal with mean θ ^ M L and variance 1 K I 1 ( X | θ ^ M L ) ,
π K ( θ | X ) K N θ ^ M L , 1 K I 1 ( X | θ ^ M L ) .
In other words, as the number of clones K increases, marginal distributions of individual parameters θ 0 should be nearly degenerated around the ML point estimate, and the variance of this distribution is K times the corresponding asymptotic variance of the ML estimator of θ .
Thus, the main result of the data cloning method is to guarantee that when we estimate the posterior distributions of model parameters using K clones of the original sample (Equation (8)), as K grows to infinity, the resulting estimator eliminates the influence of prior specification on parameter estimation by pinpointing maximum likelihood estimates as limiting estimates from the procedure. Intuitively, the influence of priors on posterior estimation is progressively eliminated by the sample cloning procedure because, since the contribution of the prior is fixed for any sample size, its influence is dominated by the information contained in an evaluated likelihood function from a K times replicated original sample. The factor 1 / K multiplying the inverse of the information matrix in Equation (9) is necessary in order to correct the variance of the estimators. Since they are obtained using K-clones of the original sample, their posterior variance is artificially reduced by the increased sample size. The asymptotic normality of DC posterior estimates is demonstrated by [1]. Thus, the DC estimator can be taken as a computational method employed in order to exploit the standard asymptotic properties of Bayesian estimators given by the Bernstein–von Mises theorem, which states that under some regularity conditions, discussed, for example, in van der Vaart [43], Bayesian estimators converge to the maximum-likelihood estimator as sample size grows.
Data cloning estimators belong to the class of estimators denoted as quasi-Bayesian/ Laplacian by Chernozhukov and Hong [44] and thus inherit the good computational properties that such methods exhibit when compared to directly maximizing the likelihood function. As the estimators are constructed through MCMC sampling, they are not obtained through numerical optimization and thus are robust to problems such as nonsmooth and discontinuous objective functions, do not suffer from the computational curse of dimensionality, and are as efficient as the extremum estimates. Chernozhukov and Hong [44] also provide evidence that these estimators are robust to multiple local maxima and initial values. However, these estimators can be affected by problems that affect the speed of convergence of the MCMC chains, for example, initial values in the tails of the posterior distribution, a problem that can be avoided through a choice of initial values based on economic considerations.
The implementation of the data cloning method is very straightforward. One has only to repeat the original sample K times to obtain an augmented sample with K × T observations and then sample from π K ( θ | X ) using a traditional Metropolis–Hastings algorithm. As discussed by Lele et al. [1], one does not assume that the artificial samples are independent and the sample repetition is only a scheme to numerically obtain the maximum-likelihood estimator of parameter θ . As the number of clones increases, the algorithm better approximates the true location of the likelihood function and the true inverse of the Fisher information for the observed data. The statistical accuracy of the estimator is a function of sample size and is not enhanced by the data cloning procedure; increasing the number of clones will only contribute to enhancing the numerical approximation of the ML estimates.
An alluring theoretical advantage of the data cloning methodology is that since ML estimates are invariant to the choice of prior distributions, DC helps attenuate the impact of subjectivity that prior specification can have on posterior estimates. There is no guarantee, however, that the DC method will be successful in pinpointing ML estimates. As Lele [42] discuss, Bayesian hierarchical models are easy to construct and straightforward to analyze due to MCMC, and, as a general principle, the complexity of the model should not exceed the information content in the data. Data cloning can be helpful by alerting the researcher to potential issues with the model such as nonestimability. Data cloning is used by Ponciano et al. [5] to assess parameter identifiability in Bayesian phylogenetic models. Ponciano et al. [2] show how the data cloning can be used to construct hypothesis tests, confidence intervals, and perform model selection exercises.
Within the DC framework, an individual parameter is considered identifiable if its marginal posterior variance converges to 0 at the rate of 1 / K as the number of K clones increases (Lele et al. [41], Ponciano et al. [5]). If, however, the likelihood function has insufficient curvature with respect to some individual parameter θ 0 , then as the number of clones increases and the prior distribution has a diminished impact on the posterior, marginal variance will not converge to zero at the predicted rate because the flat likelihood dominates the DC marginal posterior distribution.
The computational properties of data cloning estimators in the context of estimating DSGE models are analyzed in Furlani et al. [13], which through Monte Carlo studies, show that data cloning estimators are robust to the choice of initial values and local maximums, and it has superior performance in finite samples terms of bias and root mean squared error in relation to several other numerical methods of maximizing the likelihood function. As discussed in Furlani et al. [13], data cloning estimators do not rely on the evaluation of derivatives and Hessians of the objective function, and in this respect, they are more robust to problems related to the evaluation of these functions, such as near-flat-likelihood functions, discontinuities, and problems in the numerical approximations of these functions through the use of methods such as finite differences.
Given its both practical and theoretical appeal, the idea of DC has been adapted to many settings and applications. We enumerate a few for the sake of context. Baghishani and Mohammadzadeh [3] employ the DC method to estimate generalized linear mixed models in a spatial setting. Picchini and Anderson [7] combine the DC idea with an approximate Bayesian computation (ABC) approach. More recently, Duan et al. [8] incorporated data cloning into a sequential Monte Carlo algorithm. They employed DC to help with convergence and argue that their method has good computational properties that take advantage of parallel computing.
The data cloning method allows the construction of information criteria such as the AIC and the BIC. These criteria can be constructed by evaluating the likelihood function with the parameters estimated through the limit of the Bayesian estimator, using the penalties associated with each information criterion. In the case of our study, we are only analyzing a single model, and therefore, there is no need to compare different specifications using these criteria.

4. Model and Estimation Details

The model of Smets and Wouters [15] is a medium-scale general equilibrium model augmented with nominal rigidities and additional exogenous shocks that embodies many key features of New Keynesian DSGE models.
There are 41 parameters in the SW model: 17 dictate the dynamics of exogenous AR and ARMA processes, the remainder characterize structural features of the economy. Five parameters are calibrated, leaving a total of 36 parameters to be estimated from seven observable series. The model is estimated from seven macroeconomic variables observed at a quarterly frequency: real output, worked hours, inflation rate, consumption, investment, and capital stock. We omit further details on the model structure due to space constraints, and point to the original paper for more information. Smets and Wouters [15] employ data from the third quarter of 1947 through the fourth quarter of 2004, totaling 230 observations. Estimation was carried out in Matlab and Dynare (version 4.3.3, Adjemian et al. [45]), using two MCMC chains composed of 250.000 iterations each.
First we replicate the baseline estimation exercise of Smets and Wouters [15], employing the same data and prior specification. Then, using gradually more sample repetitions (5, 10, and 25), we compute data cloning posterior estimates. As discussed in Section 3, parameter identification can be accessed by observing the behavior of data cloning posterior variance as the number of sample repetitions increases, and a global measure of identification comes from the largest eigenvalue of the DC posterior covariance matrix.

5. Results

The present application of the data cloning method consists in first replicating the baseline exercise of Smets and Wouters [15], then estimating the model using gradually more sample replications—we consider 5, 10, and 25. Local identification can be accessed by observing the posterior volatility of the data cloning estimated parameters. We also measured the global identification identification properties of the DSGE system obtained through the maximum eigenvalue of the data cloning posterior covariance matrix. In this section, we report our findings and place them in relation to established results in the literature.
Table 1 summarizes our estimation results. The first two columns display parameters and their economic definition. Prior specification, which mimics Smets and Wouters [15], is reported in columns three to five. Following standard practice in the DSGE literature, prior specification is reported in terms of distribution family (N for Normal, B for Beta, G for Gamma, and I G for Inverse Gamma), mean, and standard deviation. Columns three through eight display our single-sample MCMC estimation results: the mode obtained through numerical optimization (initial values for posterior sampling), the posterior mean, and standard deviation. Then we report posterior mean and standard deviations for data cloning estimations using 5, 10, and 25 sample replications. The last column reproduces posterior means reported by Smets and Wouters [15] in their Table 1A,B for easier comparison.
Single sample-point estimates are mostly in line with benchmark SW values and remain, with notable exceptions, stable as the number of sample replication increases. As we increase the number of clones, estimates of the autorregressive coefficient of the price-markup exogenous process ρ p push the upper limit of unity. SW’s estimate of ρ p is 0.89, while our estimate with 20 sample clones is 0.99, implying qualitatively different dynamics for the exogenous price-markup process. The price-markup process is ARMA(1,1), and a similar issue seems to occur with the moving average parameter μ p , whose benchmark estimate of 0.69 changes to 0.927 with 20 sample clones. This phenomenon seems to be specific to the price-markup process. For some other exogenous processes, we do have autorregressive coefficients estimated with relatively high magnitude, namely ρ g and ρ w , but those parameters have stable estimates not far from benchmark values. Furthermore, this issue seems to be of a different sort than the one discussed by Morris [35], as in his Monte Carlo experiment of SW’s model, the sampling distribution of the estimator of ρ p is indistinguishable from its Beta prior.
Following Furlani et al. [13], we compute normalized standard errors (proportional to the standard error in the single-sample estimation) to evaluate whether the data-cloning algorithm is successful in reducing posterior volatility. Table 2 displays the normalized standard errors as a function of the number of clones K, denoted by s K * . To enhance visualization, we list the parameters in decreasing order relative to the 25-clone standard deviation estimate. Parameters in Table 2 are listed from worse to best identified with respect to the normalized standard error metric.
Parameters related to the dynamics of exogenous stochastic processes, in general, seem to be better identified relative to structural parameters. In Table 1 shock parameters are presented below the calibrated parameters, while structural parameters are placed above. This is a recurring finding in the identification literature ([14,17,20]). Indeed, we find the best-identified parameters to be ρ b and σ b , which describe the risk-premium exogenous process. Even though ρ p and σ p are considered well identified according to the normalized standard deviation metric, from the previous discussion on the autorregressive coefficient, we know that there must be estimation problems related to those parameters.
Among shock parameters, the one whose posterior volatility is least diminished by the data cloning method is the standard deviation of shocks to wage markups σ w . This is not unexpected, since this specific parameter is embroiled in the critique Chari et al. [16] made of the SW model, which is based on the implausible magnitude and interchangeable economic interpretation of exogenous wage markup shocks ε t w .
Several structural parameters that characterize the economic environment seem not to be identifiable. After 20 clones, the posterior standard deviation of Ψ , which determines costs to capital utilization readjustment, is almost twice the original standard deviation; in the case of wage elasticity of labor supply σ l , s 25 * is almost one and a half times s 1 * , indicating the data cloning algorithm is unable to reduce posterior volatility and thus unable to accurately determine maximum likelihood point estimates.
Calvo reoptimization probabilities ξ p and ξ w are badly identified, while partial indexation degrees ι p and ι w display somewhat better identifiability. The fact that we seem to better identify partial indexation parameters could be, as suggested by Iskrev [17] and Canova and Sala [14], because ξ and ι are highly pairwise collinear. We believe it is important to remember that price rigidities are not primitive model features, and there is nothing that justifies the introduction of these rigidities besides a higher level of realism in the model. However, the empirical validity of the indexed-Calvo model is questionable (Dixon and Kara [46], Bils et al. [47], Bils and Klenow [48]), which adds to the catalogue of critiques directed towards the class of NK-DSGE models, of which SW’s model is an important example.
Parameters that appear explicitly in the measurement equations are sometimes called “trend parameters”. In SW’s model, these are the growth rate trend γ , steady-state inflation π ¯ , steady-state worked hours l ¯ , and discount factor β . Most trend parameters seem to be badly identified, especially steady-state worked hours l ¯ and inflation π ¯ . The parameter γ is featured explicitly as linear growth trend of output, consumption, investment, and worked hours. According to the the data cloning method, γ seems to be only weakly identified. Canova and Sala [14] and Chadha and Shibayama [20] show that these trend parameters mainly affect first moments, and having constant terms changes the identification in general. Interestingly, we find the intertemporal discount factor β to be relatively well identified—the best performing among trend parameters. This result is unexpected since Iskrev [17] argues β has little expected impact over the likelihood function and should be especially hard to identify.
Figure 1 plots the standardized maximum eigenvalue of the data cloning covariance matrix λ K S against its expected value of 1 / K . We find λ K S above its reverence level for all samples considered. From this evidence, we cannot say the SW model is globally identified, which is to be expected given the widespread poor individual parameter identification performance.

6. Conclusions

In this paper, we applied the data cloning methodology of Lele et al. [1] to the flagship NK-DSGE model of Smets and Wouters [15]. The data-cloning method is a numerical scheme to obtain maximum-likelihood parameter estimates by artificially replicating the original sample. A byproduct of this method is an identification metric related to the posterior variability of the parameter as the number of sample clones increases.
The data cloning method is a simple way of obtaining maximum likelihood estimators, and in this respect, it allows verifying the impact of the prior structure on the estimation results of DSGE models with some simple modifications in the Bayesian methods traditionally used in the estimation of these models.
In addition, this method produces measures of local and global identification in this class of models as direct by-products. This result is particularly useful since identification diagnoses are notably more complex due to the non-linear structure and the use of the approximations and linearizations required for estimation. We believe that the data cloning method is an effective contribution to the estimation and analysis of DSGE models and can be built from the Bayesian estimation methods that are the standard in this literature.
Regarding the identification diagnostics obtained by the data cloning method for the Smets and Wouters [15] model analyzed in this work, shock-related parameters seem better identified relative to structural parameters, which has become a stylized fact. The former group displays identification problems nonetheless. Point estimates for the autoregressive coefficient of the exogenous price markup process ρ p become larger as the number of sample clones increases, pushing the upper limit of the unit. The standard deviation of shocks to the wage-markup process σ w does not seem well identified, which relates to the critique that Chari et al. [16] make of the NK-DSGE model of Smets and Wouters [15].

Author Contributions

The two authors participated equally in all stages of the work. All authors have read and agreed to the published version of the manuscript.

Funding

The authors acknowledge funding from CNPq (310646/2021-9), FAPESP (2018/04654-9) and Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)—Finance Code 001.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are from [15].

Conflicts of Interest

The authors report the absence of any type of conflict of interest.

References

  1. Lele, S.R.; Dennis, B.; Lutscher, F. Data cloning: Easy maximum likelihood estimation for complex ecological models using Bayesian Markov chain Monte Carlo methods. Ecol. Lett. 2007, 10, 551–563. [Google Scholar] [CrossRef] [PubMed]
  2. Ponciano, J.M.; Taper, M.L.; Dennis, B.; Lele, S.R. Hierarchical models in ecology: Confidence intervals, hypothesis testing, and model selection using data cloning. Ecology 2009, 90, 356–362. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Baghishani, H.; Mohammadzadeh, M. A data cloning algorithm for computing maximum likelihood estimates in spatial generalized linear mixed models. Comput. Stat. Data Anal. 2011, 55, 1748–1759. [Google Scholar] [CrossRef]
  4. Torabi, M. Likelihood inference in generalized linear mixed models with two components of dispersion using data cloning. Comput. Stat. Data Anal. 2012, 56, 4259–4265. [Google Scholar] [CrossRef]
  5. Ponciano, J.M.; Burleigh, J.G.; Braun, E.L.; Taper, M.L. Assessing parameter identifiability in phylogenetic models using data cloning. Syst. Biol. 2012, 61, 955–972. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Torabi, M.; Lele, S.R.; Prasad, N.G. Likelihood inference for small area estimation using data cloning. Comput. Stat. Data Anal. 2015, 89, 158–171. [Google Scholar] [CrossRef]
  7. Picchini, U.; Anderson, R. Approximate maximum likelihood estimation using data-cloning ABC. Comput. Stat. Data Anal. 2017, 105, 166–183. [Google Scholar] [CrossRef] [Green Version]
  8. Duan, J.C.; Fulop, A.; Hsieh, Y.W. Data-cloning SMC2: A global optimizer for maximum likelihood estimation of latent variable models. Comput. Stat. Data Anal. 2020, 143, 106841. [Google Scholar] [CrossRef]
  9. Laurini, M.P. A Hybrid Data Cloning Maximum Likelihood Estimator for Stochastic Volatility Models. J. Time Ser. Econom. 2013, 5, 193–229. [Google Scholar] [CrossRef]
  10. Marín, J.M.; Rodríguez-Bernal, M.T.; Romero, E. Data cloning estimation of GARCH and COGARCH models. J. Stat. Comput. Simul. 2015, 85, 1818–1831. [Google Scholar] [CrossRef]
  11. de Zea Bermudez, P.; Marín, J.M.; Veiga, H. Data cloning estimation for asymmetric stochastic volatility models. Econom. Rev. 2020, 39, 1057–1074. [Google Scholar] [CrossRef]
  12. Fernández-Villaverde, J. The econometrics of DSGE models. SERIEs 2010, 1, 3–49. [Google Scholar] [CrossRef] [Green Version]
  13. Furlani, L.G.C.; Laurini, M.P.; Portugal, M.S. Data cloning: Maximum likelihood estimation of DSGE models. Results Appl. Math. 2020, 7, 100121. [Google Scholar] [CrossRef]
  14. Canova, F.; Sala, L. Back to square one: Identification issues in DSGE models. J. Monet. Econ. 2009, 56, 431–449. [Google Scholar] [CrossRef] [Green Version]
  15. Smets, F.; Wouters, R. Shocks and frictions in US business cycles: A Bayesian DSGE approach. Am. Econ. Rev. 2007, 97, 586–606. [Google Scholar] [CrossRef] [Green Version]
  16. Chari, V.V.; Kehoe, P.J.; McGrattan, E.R. New Keynesian models: Not yet useful for policy analysis. Am. Econ. J. Macroecon. 2009, 1, 242–266. [Google Scholar] [CrossRef] [Green Version]
  17. Iskrev, N. Local identification in DSGE models. J. Monet. Econ. 2010, 57, 189–202. [Google Scholar] [CrossRef] [Green Version]
  18. Komunjer, I.; Ng, S. Dynamic identification of dynamic stochastic general equilibrium models. Econometrica 2011, 79, 1995–2032. [Google Scholar]
  19. Romer, P. The Trouble with Macroeconomics. Available online: https://paulromer.net/trouble-with-macroeconomics-update/WP-Trouble.pdf (accessed on 21 September 2016).
  20. Chadha, J.S.; Shibayama, K. Bayesian estimation of DSGE models: Identification using a diagnostic indicator. J. Econ. Dyn. Control. 2018, 95, 172–186. [Google Scholar] [CrossRef] [Green Version]
  21. Christiano, L.J.; Eichenbaum, M.S.; Trabandt, M. On DSGE models. J. Econ. Perspect. 2018, 32, 113–140. [Google Scholar] [CrossRef] [Green Version]
  22. DeJong, D.N.; Dave, C. Structural Macroeconometrics; Princeton University Press: Princeton, NJ, USA, 2011. [Google Scholar]
  23. Blanchard, O.J.; Kahn, C.M. The solution of linear difference models under rational expectations. Econometrica 1980, 48, 1305–1311. [Google Scholar] [CrossRef]
  24. Anderson, G.; Moore, G. A linear algebraic procedure for solving linear perfect foresight models. Econ. Lett. 1985, 17, 247–252. [Google Scholar] [CrossRef]
  25. King, R.G.; Watson, M.W. The solution of singular linear difference systems under rational expectations. Int. Econ. Rev. 1998, 39, 1015–1026. [Google Scholar] [CrossRef] [Green Version]
  26. Sims, C.A. Solving linear rational expectations models. Comput. Econ. 2002, 20, 1–20. [Google Scholar] [CrossRef]
  27. Gamerman, D.; Lopes, H. MCMC—Stochastic Simulation for Bayesian Inference; Chapman & Hall/CRC: London, UK, 2006. [Google Scholar]
  28. An, S.; Schorfheide, F. Bayesian analysis of DSGE models. Econom. Rev. 2007, 26, 113–172. [Google Scholar] [CrossRef] [Green Version]
  29. Rothenberg, T.J. Identification in parametric models. Econom. J. Econom. Soc. 1971, 39, 577–591. [Google Scholar] [CrossRef]
  30. Calvo, G.A. Staggered prices in a utility-maximizing framework. J. Monet. Econ. 1983, 12, 383–398. [Google Scholar] [CrossRef]
  31. Sargent, T.J. The observational equivalence of natural and unnatural rate theories of macroeconomics. J. Political Econ. 1976, 84, 631–640. [Google Scholar] [CrossRef]
  32. Qu, Z.; Tkachenko, D. Identification and frequency domain quasi-maximum likelihood estimation of linearized dynamic stochastic general equilibrium models. Quant. Econ. 2012, 3, 95–132. [Google Scholar] [CrossRef] [Green Version]
  33. Tkachenko, D.; Qu, Z. Frequency domain analysis of medium scale DSGE models with application to Smets and Wouters (2007). Adv. Econom. 2012, 28, 319. [Google Scholar]
  34. Koop, G.; Pesaran, M.H.; Smith, R.P. On identification of Bayesian DSGE models. J. Bus. Econ. Stat. 2013, 31, 300–314. [Google Scholar] [CrossRef] [Green Version]
  35. Morris, S.D. DSGE pileups. J. Econ. Dyn. Control. 2017, 74, 56–86. [Google Scholar] [CrossRef]
  36. Ivashchenko, S.; Mutschler, W. The effect of observables, functional specifications, model features and shocks on identification in linearized DSGE models. Econ. Model. 2019, 88, 280–292. [Google Scholar] [CrossRef] [Green Version]
  37. Qu, Z.; Tkachenko, D. Global Identification in DSGE Models Allowing for Indeterminacy. Rev. Econ. Stud. 2018, 84, 1306–1345. [Google Scholar] [CrossRef] [Green Version]
  38. Meenagh, D.; Minford, P.; Wickens, M.; Xu, Y. Testing DSGE Models by Indirect Inference: A Survey of Recent Findings. Open Econ. Rev. 2019, 30, 593–620. [Google Scholar] [CrossRef] [Green Version]
  39. Kocięcki, A.; Kolasa, M. Global identification of linearized DSGE models. Quant. Econ. 2018, 9, 1243–1263. [Google Scholar] [CrossRef]
  40. Kocięcki, A.; Kolasa, M. A Solution to the Global Identification Problem in DSGE Models; Working Papers; Faculty of Economic Sciences, University of Warsaw: Warsaw, Poland, 2022. [Google Scholar]
  41. Lele, S.R.; Nadeem, K.; Schmuland, B. Estimability and likelihood inference for generalized linear mixed models using data cloning. J. Am. Stat. Assoc. 2010, 105, 1617–1625. [Google Scholar] [CrossRef]
  42. Lele, S.R. Model complexity and information in the data: Could it be a house built on sand? Ecology 2010, 91, 3493–3496. [Google Scholar] [CrossRef] [Green Version]
  43. van der Vaart, A. Asymptotic Statistics; Cambridge University Press: Cambridge, UK, 1998. [Google Scholar]
  44. Chernozhukov, V.; Hong, H. An MCMC approach to classical estimation. J. Econom. 2003, 115, 293–346. [Google Scholar] [CrossRef] [Green Version]
  45. Adjemian, S.; Bastani, H.; Juillard, M.; Mihoubi, F.; Perendia, G.; Ratto, M.; Villemot, S. Dynare: Reference Manual, Version 4; Technical Report; CEPREMAP: Paris, France, 2011. [Google Scholar]
  46. Dixon, H.; Kara, E. Can we explain inflation persistence in a way that is consistent with the microevidence on nominal rigidity? J. Money Credit. Bank. 2010, 42, 151–170. [Google Scholar] [CrossRef] [Green Version]
  47. Bils, M.; Klenow, P.J.; Malin, B.A. Reset price inflation and the impact of monetary policy shocks. Am. Econ. Rev. 2012, 102, 2798–2825. [Google Scholar] [CrossRef] [Green Version]
  48. Bils, M.; Klenow, P.J. Some Evidence on the Importance of Sticky Prices. J. Political Econ. 2004, 112, 947–985. [Google Scholar] [CrossRef]
Figure 1. Standardized maximum eigenvalue of the posterior covariance matrix. Note: figure plots the standardized maximum eigenvalue of data cloning posterior the covariance matrix estimates against the 1 / K expected value for a well-identified model.
Figure 1. Standardized maximum eigenvalue of the posterior covariance matrix. Note: figure plots the standardized maximum eigenvalue of data cloning posterior the covariance matrix estimates against the 1 / K expected value for a well-identified model.
Stats 06 00002 g001
Table 1. Data cloning posterior estimates.
Table 1. Data cloning posterior estimates.
ParameterDefinitionPriorOptimizationSingle-Sample (MCMC)5 Clones10 Clones25 ClonesBenchmark
DensityMeanStdModeMeanStdMeanStdMeanStdMeanStdMean
ψ Investment adj. cost.N4.001.505.495.5882.2416.8511.0916.0040.6945.9741.0445.74
σ c Inv. elats. intert. subst.N1.500.381.471.4350.3281.4010.3271.4060.2431.4680.0611.38
hConsump. habitB0.700.100.700.7050.1020.7940.0710.7710.0320.7590.0540.71
ξ w Calvo wageB0.500.100.730.6920.1630.8960.0240.8550.0670.8380.1210.71
σ l Elast. labour supplyN2.000.751.671.6551.2593.7910.4333.2670.7313.3981.8521.83
ξ p Calvo priceB0.500.100.680.6760.1250.7840.0480.7370.0390.6950.0760.66
ι w Index. of wagesB0.500.150.560.5430.2770.4760.1450.4950.1410.5820.0370.58
ι p Index. of pricesB0.500.150.240.260.2020.2730.0920.2760.0730.2650.0390.24
Ψ Capital utilizationB0.500.150.400.4110.2110.1710.1220.1990.1020.2260.3880.54
Φ Fixed costN1.250.121.651.6430.1711.5810.0951.5910.0651.5750.0781.6
r π Response to inflationN1.500.251.982.0150.3752.1310.3161.9360.0721.9490.1932.04
ρ Interest rate smooth.N0.750.100.820.820.0520.8720.0160.8540.0180.8510.0210.81
r y Response to outputN0.130.050.090.0950.0520.1380.0690.1230.0290.1180.0070.08
r Δ y Response to output gapN0.130.050.220.2220.0640.1910.0130.1850.0270.1920.0470.22
π ¯ Steady state inflationG0.630.100.670.6790.1570.5340.1210.6040.1070.6040.0650.78
100 ( β 1 1 ) Discount factorG0.250.100.210.2410.2030.2320.1670.1860.0410.1950.0380.16
l ¯ Steady state hours workedN0.002.000.400.2962.2492.3511.9210.8761.1351.0381.2220.53
100 ( γ 1 ) Trend growthN0.400.100.440.4350.0350.4510.0110.4560.0250.4550.0110.43
α Share of capitalN0.300.050.320.3140.0920.3540.1060.3710.0310.3560.0670.19
δ Depreciation raten.a.0.025n.a.n.a.n.a.n.a.n.a.n.a.n.a.n.a.n.a.n.a.n.a.
g y Government/Output ration.a.0.18n.a.n.a.n.a.n.a.n.a.n.a.n.a.n.a.n.a.n.a.n.a.
ϕ w Wage mark-upn.a.1.5n.a.n.a.n.a.n.a.n.a.n.a.n.a.n.a.n.a.n.a.n.a.
ϵ w Kimball (wage)n.a.10n.a.n.a.n.a.n.a.n.a.n.a.n.a.n.a.n.a.n.a.n.a.
ϵ p Kimball (price)n.a.10n.a.n.a.n.a.n.a.n.a.n.a.n.a.n.a.n.a.n.a.n.a.
ρ g a Tech. shock to gov. spendingN0.500.250.600.5870.2040.5530.0970.6440.1160.6250.0360.52
ρ a AR nonspecific technologyB0.500.200.950.9480.03540.9740.0270.9680.0140.9670.0050.95
ρ b AR risk premiumB0.500.200.170.2120.19570.2190.1830.2170.0340.2090.0140.22
ρ g AR government spendingB0.500.200.970.9730.01960.9970.0010.9960.0160.9950.0190.97
ρ i AR investmentB0.500.200.740.7470.1340.6470.0830.6710.1010.6810.0450.71
ρ m AR monetary policyB0.500.200.120.1420.12940.0420.0490.1210.0950.1210.0310.15
ρ p AR price mark-upB0.500.200.900.8790.13260.9990.0010.9990.0750.9990.0160.89
ρ w AR wage mark-upB0.500.200.970.9590.03660.9720.0290.9690.0040.9670.0110.96
μ p MA price mark-upB0.500.200.770.7230.22170.9660.0130.9520.1150.9270.0810.69
μ w MA wage mark-upB0.500.200.870.8090.16940.9450.0480.9140.0260.9040.0290.84
σ a Std. technology shockIG0.102.000.430.4340.05060.4720.0290.4450.0110.4460.0260.45
σ b Std. risk premium shockIG0.102.000.240.2370.06060.2430.0520.2410.0380.2430.0030.23
σ g Std. government spendingIG0.102.000.510.5160.05190.520.0160.5040.0610.5120.0110.53
σ i Std. investment shockIG0.102.000.430.4350.06880.4690.0790.4580.0850.4610.0110.56
σ m Std. monetary policyIG0.102.000.240.2440.10140.2310.0110.2330.0060.2280.0090.24
σ π Std. price mark-up shockIG0.102.000.140.1410.03280.1450.0160.1390.0140.1330.0040.14
σ w Std. wage mark-up shockIG0.102.000.240.2350.03590.2310.0240.2170.0080.2240.0280.24
Note: table displays posterior estimates of the SW model. In the priors, specification N corresponds to a Normal distribution, B a Beta distribution, G a Gamma distribution, and IG an Inverse-Gamma distribution. The Single-Sample columns are the results of the usual estimation by MCMC. Calibrated parameters are represented by “n.a.” values.
Table 2. Normalized data cloning posterior standard deviation.
Table 2. Normalized data cloning posterior standard deviation.
Parameters s 1 * s 2 * s 3 * s 5 * s 10 * s 25 *
Ψ 1.0000.3890.3730.5820.4451.841
σ l 1.0000.4410.3480.3440.331.471
ρ g 1.0000.1780.1420.0810.1321.004
σ w 1.0000.4060.4340.6850.3280.799
ξ w 1.0000.1760.1740.1510.3180.738
α 1.0000.4280.5451.1490.4950.733
r Δ y 1.0000.3380.3610.2070.5150.733
ξ p 1.0000.3390.3430.3820.4090.61
l ¯ 1.0000.6210.5750.8531.0460.543
h1.0000.3830.2940.70.1270.534
σ a 1.0000.2840.5330.5880.7920.519
r π 1.0000.3520.5710.8430.2250.516
ψ 1.0000.4290.4230.4860.2060.466
Φ 1.0000.4880.6110.5560.4410.456
π ¯ 1.0000.5290.3750.7631.0590.413
ρ 1.0000.2990.3290.3160.1930.397
μ p 1.0000.0970.0860.0590.0870.362
ρ i 1.0000.4260.4520.6210.5550.341
100 ( γ 1 ) 1.0000.3140.2570.3110.2260.311
ρ w 1.0000.2650.2950.8030.5320.306
ρ m 1.0000.3260.3330.3830.9120.238
σ g 1.0000.4250.4830.3120.6760.204
ι p 1.0000.4850.2880.4550.4330.194
σ c 1.0000.4250.4650.9990.5360.187
100 ( β 1 1 ) 1.0000.7050.5250.8250.2260.186
ρ g a 1.0000.3570.4970.4750.3280.178
μ w 1.0000.1080.1090.2850.3240.171
σ i 1.0000.3020.7731.1561.3150.175
ρ a 1.0000.2090.3330.7790.5810.163
r y 1.0000.3160.7281.3240.2770.148
ι w 1.0000.5030.3840.5230.6280.135
σ p 1.0000.3070.3710.4960.3680.128
ρ p 1.0000.0270.0170.0070.0020.122
σ m 1.0000.3590.1160.1160.1460.088
ρ b 1.0000.3920.3860.9370.3250.072
σ b 1.0000.4020.4190.8710.7120.061
Note: Table reports the posterior standard deviation s K * of parameters of the SW model obtained from DC estimates as a proportion of the single sample standard deviation s 1 * .
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chaim, P.; Laurini, M.P. Data Cloning Estimation and Identification of a Medium-Scale DSGE Model. Stats 2023, 6, 17-29. https://doi.org/10.3390/stats6010002

AMA Style

Chaim P, Laurini MP. Data Cloning Estimation and Identification of a Medium-Scale DSGE Model. Stats. 2023; 6(1):17-29. https://doi.org/10.3390/stats6010002

Chicago/Turabian Style

Chaim, Pedro, and Márcio Poletti Laurini. 2023. "Data Cloning Estimation and Identification of a Medium-Scale DSGE Model" Stats 6, no. 1: 17-29. https://doi.org/10.3390/stats6010002

Article Metrics

Back to TopTop