Next Article in Journal
Adaptive Square-Root Unscented Kalman Filter-Based State-of-Charge Estimation for Lithium-Ion Batteries with Model Parameter Online Identification
Next Article in Special Issue
The Trend and Status of Energy Resources and Greenhouse Gas Emissions in the Malaysia Power Generation Mix
Previous Article in Journal
Effect of Silica, Activated Carbon, and Alumina Supports on NiMo Catalysts for Residue Upgrading
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Uncertainty Analysis of Greenhouse Gas (GHG) Emissions Simulated by the Parametric Monte Carlo Simulation and Nonparametric Bootstrap Method

1
Department of Environmental and Safety Engineering, Ajou University, 206 Worldcup-ro, Yeongtong-gu, Suwon 16499, Korea
2
Environmental Regulation Compliance Office, Korea Institute of Industrial Technology, Hanshin Intervalley 24 East B/D 18F 322, Teheran-ro, Gangnam-gu, Seoul 06211, Korea
3
Division of Policy Research, Green Technology Center, 173, Toegye-re, Jung-gu, Seoul 04554, Korea
4
Office of Carbon Upcycling R&D strategy, Environment & Sustainable Resources Research Center, Korea Research Institute of Chemical Technology, 141 Gajeong-ro, Yuseong-gu, Daejeon 34114, Korea
*
Author to whom correspondence should be addressed.
Energies 2020, 13(18), 4965; https://doi.org/10.3390/en13184965
Submission received: 30 July 2020 / Revised: 7 September 2020 / Accepted: 18 September 2020 / Published: 22 September 2020
(This article belongs to the Special Issue Life Cycle Assessment of Environmental System)

Abstract

:
Uncertainty of greenhouse gas (GHG) emissions was analyzed using the parametric Monte Carlo simulation (MCS) method and the non-parametric bootstrap method. There was a certain number of observations required of a dataset before GHG emissions reached an asymptotic value. Treating a coefficient (i.e., GHG emission factor) as a random variable did not alter the mean; however, it yielded higher uncertainty of GHG emissions compared to the case when treating a coefficient constant. The non-parametric bootstrap method reduces the variance of GHG. A mathematical model for estimating GHG emissions should treat the GHG emission factor as a random variable. When the estimated probability density function (PDF) of the original dataset is incorrect, the nonparametric bootstrap method, not the parametric MCS method, should be the method of choice for the uncertainty analysis of GHG emissions.

Graphical Abstract

1. Introduction

In accordance with the Paris Accord, many countries around the world set up voluntary goals for mitigating greenhouse gas (GHG) emissions and implement various action plans, including related research to achieve the goals. Many industrial sectors set up the goals and action plans to comply with the Paris Accord. They include, among others, goals related to energy, industrial processes and product use, agriculture, forest and other land use, and waste. The energy sector, including the transportation sector, formulated three different policy scenarios for the carbon emission goal in 2050 and computed GHG emissions for each scenario [1]. Vázquez-Rowe et al. [2] estimated GHG emissions of cement production in three relevant national cement plants to identify the main GHG mitigation strategies throughout the whole supply chain in industrial processes and the product use sector. Baek et al. [3] developed a GHG emission quantification procedure for dairy cow systems based on a life cycle assessment (LCA) approach incorporating the Intergovernmental Panel on Climate Change (IPCC)’s GHG emission computation equations in the agriculture sector. The waste sector computed GHG emissions from the composting and solid waste treatment processes and suggested recommendations for improving the current waste management policy [4].
The above literature [1,2,3,4] adopted the same methodological approach for estimating GHG emissions in each sector [5,6]. This approach involves integrating information on the level of human activity (i.e., activity data) with the quantified emission coefficients per unit activity (i.e., emission factor) [7]. Therefore, a model used extensively for the computation of GHG emissions from different sources in a given sector is a linear model, where the activity data are multiplied by the emission factor.
Recently, uncertainty analysis of a mathematical model such as the greenhouse gas (GHG) emission estimation model is receiving increasing attention. GHG has an environmental impact that is categorized as global warming impact, one of the many life cycle impact categories in life cycle assessment (LCA). When estimating GHG emissions, it is always necessary to evaluate and quantify the uncertainties of estimates. Uncertainty analyses help analysts identify how accurate the estimations are and the likely range in which the true value of the emissions fall [8].
Uncertainty refers to the difference between the true value and the measured value. All input random variables to the mathematical model have uncertainty, be it systematic or random [9,10]. Uncertainty of the input random variables often refers to error. As such, this paper defines the uncertainty of the model output as an uncertainty and the uncertainty of the input random variables as an error.
A model is a mathematical description of the real-world phenomenon. A model can have any number of parameters, and they are interrelated through various mathematical operations. Parameters (e.g., activities, processes, and emission factors) refer to any input random variables comprising a mathematical model. Error in the parameter originates from the lack of knowledge about the true value of the parameter [11]. These errors can vary widely from a few percent to orders of magnitude [12]. Errors associated with the parameters are bound to increase due to the compounding effect of the errors in the parameters. The compounding effect is often termed error propagation [8,13,14].
The variance of the mathematical model output is a measure of uncertainty of the model output. Statistics such as the confidence interval (CI) and relative uncertainty (U) are used widely for assessing the uncertainty [10]. The relative uncertainty is defined as a ratio between the half-width of the CI to the sample mean of the model output [10]. McMurray et al. [8] recommended that analysts apply the CI method for calculating uncertainty, considering that other methods such as the coefficient of variation (CV) underestimated uncertainty.
Uncertainty analysis methods often used for the mathematical model output are the error propagation, the Monte Carlo simulation (MCS) method and the bootstrap method. However, the error propagation method is intended for computing the variance of the model output, not for the CI. Many prior studies used the error propagation method for computing the uncertainty not based on the CI [9,10,14].
The MCS is run using algorithms that generate stochastic (i.e., random) values based on the probability distribution function (PDF) of the original dataset [8]. The normal distribution of the original dataset is often assumed [8]. However, non-normal assumptions such as log-normal assumptions were also made frequently [9,12,15].
Two widely used uncertainty analysis methods, the parametric MCS method and the non-parametric bootstrap method, were chosen to assess the uncertainty of GHG emissions. The parametric MCS method requires estimation of the probability distribution of the original dataset of a parameter, while the nonparametric bootstrap method does not. Lee et al. [16] showed that the nonparametric bootstrap method gives smaller relative uncertainty over the parametric MCS method. However, this study did not estimate the probability distribution of the parameter data but assumed normal and lognormal distributions. In addition, they relied on commercial software for performing the MCS.
The error propagation equation ignoring the covariance among the parameters in a model was recommended for use to calculate the uncertainties of GHG emissions from individual input uncertainty estimates [17]. However, the covariance among the parameters can be significant in most cases and cannot be ignored. Besides, there is a matrix operation that allows calculation of the variance–covariance matrix of the input parameter data [13,16]. A newer version of the IPCC guidelines [10] recommends MCS as the more detailed uncertainty analysis method.
Estimation of the probability distribution is faced with difficulties because of the limited number of observations of the original dataset. Inaccurate PDF will lead to inaccurate MCS output such that accurate estimation of PDF is a prerequisite for proper use of the parametric MCS method.
There are several studies on the uncertainty analysis of GHG emissions based on the MCS method with an assumed probability distribution. Hong et al. [18] stated that the statistical features of the total carbon emissions can be determined through its probability distribution when performing MCS. They assumed the beta distribution of the observed data in the construction industry. Others rely on expert judgment for choosing probability distribution [3,9].
However, it is rare to find studies that estimate probability distribution explicitly for MCS. Lee, et al. [16,19] used the Anderson–Darling (AD) test to estimate the probability distribution of the input variable. The p-value and AD statistics were used to test the null hypothesis for the assumed probability distributions. McMurray et al. [8] described a step-by-step approach for estimating PDF for GHG emissions using the goodness-to-fit test of the parameter data. Thus, there is a need to assess the effect of estimating probability distribution on the uncertainty of the model output.
The emission factor is often expressed as a single point estimate, not an interval estimate. However, the value of the emission factor can vary. Thus, there is a need to assess the effect of the varying emission factor (i.e., treating it as a random variable) on the uncertainty of the model output.
As the number of observations (n) of an original dataset increases, its variance decreases. When increasing n of the dataset, there would be a certain n where the variance of the model output, and subsequently the value of U, no longer decreases with the increase in n. Thus, there is a need to assess the effect of n of the original dataset on the value of U and mean of the model output.
Therefore, the objective of this study was to assess: (i) the effect of the number of observations of an original dataset; (ii) the effect of treating the emission factor as random variable; (iii) the effect of a probability distribution; and, (iv) the effect of different uncertainty analysis methods—all on the uncertainty of the model output or GHG emissions.

2. Materials and Methods

2.1. Mathematical Model

GHG emissions are estimated by multiplying the input variables by their corresponding GHG emission factors. Input variables are the input and output of materials and energies from a product system such as resources consumed and emissions discharged as well as activities such as transportation. Of these, those related to GHG emissions are chosen together with their corresponding GHG emission factor for use in Equation (1).
A mathematical model in Equation (1) for computing GHG emissions from a system is a linear combination of the input variables.
Z = AX
where Z = the model output (e.g., GHG emission), A = the coefficient vector, and X = the input variable matrix.
The model output has uncertainty originating from the errors of the input variables. The errors of the input variables to the mathematical model propagate such that the uncertainty of the model output increases [14,20]. The variance of Z measures the uncertainty of Z. Since Z is a linear combination of A and X, Equation (2) computes the variance of Z [21]. Most of the time, the input variables are not independent such that variance computation should include the covariance of X. Equation (2) is also known as the error propagation equation [20,21].
var ( Z ) = var ( AX ) = Avar ( X ) A T
where var ( X ) = the variance–covariance matrix of X , and A T = transpose of A.
In order to reduce errors of the input variables X in Equation (1), many observations of a dataset should be collected for each input variable. However, this is infeasible and impractical. As such, stochastic modeling is necessary where many observations of the input variables are generated artificially, and these observations are used to compute the value of Z by Equation (1).
Application of the uncertainty analysis method, the MCS method and the bootstrap method to Equation (1) yields many Z values, expressed as z (numerical value of the random variable Z). These z values are used to compute statistics of interest such as the mean, variance, confidence interval, relative uncertainty, bias, and standard error, among others.

2.2. Uncertainty Analysis by the Parametric MCS Method

MCS is a random experimentation based on random sampling from the uniform distribution that generates uniform random variates. These variates are then transformed into the random variates of the cumulative distribution function (CDF) of the original dataset [14]. The random variates generated are used to solve Equation (1).
MCS can be performed in four steps: (i) generate uniform random variates vector, u, from the uniform distribution of (0,1); (ii) the u vector is transformed into the corresponding values of CDF of the original dataset, the x vector, which are random variates of the original dataset distribution; (iii) this process repeats many times (e.g., 10,000 times) and generates many sets of x vectors; (iv) these x vectors are used to solve Equation (1) to obtain z and z vector. Statistics of interest such as the bias, standard error, CI, and U are computed from the z vector.
The use of the MCS method requires estimation of the probability distribution of the original dataset. Fitting distribution consists of choosing a probability distribution model that describes the dataset, finding parameter estimates for that distribution, and the goodness-of-fit test [14,22]. This requires judgment and expertise, and needs an iterative process of distribution choice, parameter estimation, and quality-of-fit assessment [22].
Before choosing a probability model for a given dataset, graphical techniques such as histograms, density estimate, empirical CDF, quantile–quantile (Q–Q) plot, and probability–probability (P–P) plot can be used to suggest the type of probability distribution [16]. Goodness-of-fit tests indicate whether a dataset comes from a specific probability distribution based on the hypothesis test [23].
The fitdistrplus package of R [22,24] was used to choose the probability distribution. The graphical comparison of multiple fitted distributions (cdfcomp) was used by plotting the empirical cumulative distribution and fitted distribution functions (y-axis) against the values of X (x-axis). Other types of plots such as the Q–Q plot (qqcomp) representing the empirical quantiles (y-axis) against the theoretical quantiles (x-axis) were also made, but are not presented here.
Akaike’s information criterion (AIC) and Schwarz Bayesian information criterion (BIC) are widely used indices for the probability distribution selection [25]. Kolmogorov–Smirnov, Cramer-von Mises and Anderson–Darling statistics [26] were also considered to assess the goodness of fit. Gofstat function was used to calculate these statistics [24].
All these statistics were computed, and a probability distribution having the smallest indices and statistic (i.e., distance) was identified. In addition, visual examination of the CDF plots focused not only on both tails of the plot but also on the center of the distribution plot. The probability distribution showing the smallest AIC/BIC criterion and smallest goodness-of-fit statistics (distance) and that which was closest to the empirical CDF was chosen as the probability distribution of the original dataset.

2.3. Unertainty Analysis by the Nonparametric Bootstrap Method

Relatively small samples generally provide very reliable information about the shape of the population. The bootstrap method treats this sample as a population and takes repeated samples with replacements from it; these are bootstrap samples and they give reliable insight into various sample estimates, and reliable confidence intervals can be constructed from these estimates [27]. The bootstrap is a widely applicable and extremely powerful statistical tool that can be used to quantify the uncertainty associated with a given estimator [27,28].
The standard error of an estimator is difficult to estimate in reality. Most of the time, there is no simple formula for the standard error such as the standard error of the sample mean ( σ / n ) [19,29].
The estimated variance of the R-bootstrapped estimates gives standard deviation, and this was used to compute the standard error of the estimator [28]. The diagonal elements of the square root of the variance–covariance matrix are the standard deviation of the bootstrapped datasets [30]. The bias of an estimator is the difference between the mean of the bootstrapped datasets and that of the original dataset [28,31]. With these, the confidence interval of Z is computed [32].
Bootstrapping consists of four steps: (i) collect a sample dataset of size n from the population (original dataset of size n); (ii) draw a bootstrapped sample of size n from the original dataset with a replacement and repeat bootstrap sampling R times; (iii) compute the statistics of interest for each bootstrapped dataset, and there will be a total of R estimates of the estimator; (iv) compute statistics of interest such as the standard error and CI with the R bootstrapped datasets and the original dataset.
In practice, a bootstrap analysis in R can be implemented in three steps [28]: (i) create a function that computes the statistic of interest; (ii) use the boot() function [33] in R to perform bootstrapping by repeatedly sampling from the original dataset with the replacement; and (iii) compute the variance, standard error, bias, CI and U of the model output Z using the bootstrapped datasets and original dataset.

3. Results and Discussion

The feed data for dairy cows were used as the input data to the mathematical model for computing GHG emissions [16]. Six different types of feed to cow and corresponding emission factors were used. These are dry cow feed, lactating cow feed, straw, soybean meal, electricity, and diesel. The corresponding emission factors are 0.38, 0.64, 0.95, 0.71, 0.50, 3.3 kg CO2-eq/kg (kwh for electricity, liter for diesel), respectively. There are six different batches of the dataset, and all are subsets of the dataset of n = 72 (Table A1). Each subset has a differing length of n, where n is 12, 24, 36, 48, 60, and 72.

3.1. The Effect of the Size of the Dataset (n) on the Uncertainty of the Model Output

In order to assess the effect of the size of the dataset (n) on the uncertainty of the model output, Z, three different scenarios for computing the mean and variance of Z (S1, S2, and S3) were made. The bootstrap method was used for the assessment. Scenario S1 computes the mean and variance of Z using Equation (1) and the error propagation equation in Equation (2), respectively, and no bootstrapping was used. Here, X is the input variable matrix consisting of six columns of the dataset, each column data with the size of n. Scenario S2 computes the mean and variance of Z by bootstrapping X while keeping A constant. The bootstrapped datasets together with the original dataset were used for the computation of the statistics.
Scenario S3 computes the mean and variance of Z similar to S2. Unlike S2, however, coefficient vector A was treated as a random variable. Since A was constant initially, a linear transformation is needed to generate random variates from A. Uniform distribution of [EF × fab, EF × (fab + 2 × (1 − fab))] was assumed, where EF is an emission factor and “fab” is a multiplication factor to generate uniform variates for A. The value of fab varies from 0.0 to 1.0. Scenario S3 computes the mean and variance of Z by bootstrapping both X and A. Table 1 shows the summary of the three different scenarios described above.
Figure 1 and Figure 2 show the effect of the size of a dataset (n) of the three different scenarios on the U and mean of Z (i.e., the mean GHG emission), respectively. As n increases, the U and mean GHG emission decrease in all three scenarios and reach an asymptotic value. The plot of U vs. n in Figure 1 shows that the U decreased rapidly initially and then reached a minimum value. This indicates that the increase in n decreases the variance of Z and remains unchanged beyond a certain number of n (here, n ≥ 48). The plot of the mean GHG emission vs. n in Figure 2 shows the same trend.
Scenario S3 exhibited the lowest U and the mean GHG emission followed by Scenario S2 and then Scenario S1. This was quite obvious for the U, while less obvious for the mean GHG emission. The variance of Z was the smallest in Scenario S3 (bootstrapping both A and X), followed by Scenario S2 (bootstrapping X only), and was the largest in Scenario S1 (no bootstrapping). This indicates that bootstrapping reduces the variance of Z.
In the case of the mean GHG emission, Scenario S1 and S2 exhibited essentially the same trend. However, Scenario S3 reached an asymptotic value when n was 36, while scenarios S1 and S2 reached the same asymptotic value when n was equal to and/or greater than 60. A significant decrease in the mean GHG emission when the number of data is small indicates that the mean GHG emission is not stable. As such, there is a certain number of observations of a dataset required before a stable mean GHG emission can be reached. In this case, the number of observations was 36 for Scenario S3 and 60 for scenarios S1 and S2.
Tong et al. [34] investigated the effect of sample size on the bias of the confidence interval of the original dataset and the bootstrapped dataset. Using the original dataset to construct a confidence interval for quantifying the uncertainty of GHG emissions may lead to a significant bias when the sample size is small. Compared with this, the bootstrapped confidence intervals have smaller interval mean and smaller interval standard deviation for small sample size (n < 30) under non-normal distributions [34]. This suggests that there will be a certain sample size above which the uncertainty of the model output reached a stable value.

3.2. The Effect of Treating the Emission Factor (Coefficient Vector) as a Random Variable on the Uncertainty of the Model Output

Figure 3 shows the effect of fab on the U and mean GHG emission for the bootstrap method. The value of U decreases by approximately 90% when fab varies from 0.0 to 1. This indicates that the variability of the coefficient affects the U. When fab was 0.0, the U was 8.8%. When fab reached 0.9, no further changes occurred to the U. Thus, treating the coefficient as a random variable increases the uncertainty of Z or U. This is because the variance of the dataset increases due to increased variance of the coefficient.
Meanwhile, the mean GHG emission remains flat in all values of fab, i.e., 440.2 kg CO2-eq/head-year at fab of 0.0, and 438.5 kg CO2-eq/head-year at fab of 1.0. This indicates that treating the coefficient in the mathematical model as a random variable does not alter the mean of Z. It affects adversely, however, the uncertainty of Z. The varying coefficient case yielded higher uncertainty of Z over the constant coefficient case.

3.3. The Effect of Probability Destribution on the Uncertainty of the Model

The very first step in performing MCS is the estimation of the probability distribution of the original dataset. Figure 4 shows the empirical and theoretical CDF (y-axis) against the input values of x (x-axis) for the lactating cow feed dataset.
Akaike’s information criterion (AIC) provides a measure of the model fit, with smaller values of AIC indicating a better-fitting model. The same is true for the Bayesian information criterion (BIC). Table 2 shows that loglogistic distribution fits best for the lactating cow feed dataset, judging from the smallest AIC and BIC values among the five distributions tested.
From the five different distributions shown in Figure 4, loglogistic distribution also exhibits close proximity to the empirical CDF in all ranges of the input values of X (x-axis). This is in agreement with the AIC and BIC values in Table 2. Thus, loglogistic distribution was chosen for the distribution of the lactating cow feed dataset. This process heavily depends on the goodness-of-fit criterion as well as the CDF plots. Inevitably, this process involves judgment call and thus implies difficulties associated with the estimation of the probability distribution of the dataset.
Table 3 shows the chosen distribution of the original dataset for the six columns of the dataset. It lists statistics of each distribution.

3.4. Comparison of the Confidence Interval Computation Method

The confidence interval is the base for assessing the uncertainty of Z. The U is based on the CI width. Three different CI computation methods including the percentile method, normal method, and basic method were tested to examine any difference among them. The bootstrap method was used for the test. Table 4 shows that the three CI computation methods produce essentially the same results.
Orloff and Bloom [35] emphasized that the percentile method is preferred to the other methods. This is because the X ¯ μ distribution is approximated by the X * ¯ X ¯ distribution, where μ , X ¯ and X * ¯ are the true mean, mean of the original dataset, and mean of the bootstrapped data, respectively. Tong et al. [34] compared different methods for computing the confidence interval; they were the normal method, percentile method, bias-corrected percentile method, and bias-corrected accelerated percentile (BCa) method. Essentially, there were no significant differences among the four methods; however, the BCa method gave shorter CI compared to the others. McMurray et al. [8] concluded that the benefit of using the percentile method to other methods is that it can be applied to any type of bootstrapped distribution. As such, this study adopted the percentile method for computing the CI of Z at the 95% confidence level.

3.5. The Effect of Different Uncertainty Analysis Methods on the Uncertainty of the Model Output

The effect of three uncertainty analysis methods on the uncertainty of the model output was assessed by comparing the magnitude of statistics of the model output. The statistics considered are the mean, bias, standard error, confidence interval, confidence interval width, and relative uncertainty. Three uncertainty methods chosen are no simulation method, the nonparametric bootstrap method, and the parametric MCS method. No simulation method refers to the case where the original dataset and emission factor were applied directly to Equation (1) (no simulation case is identical to Scenario S1). No simulation method was served as a control to both the nonparametric bootstrap method and the parametric MCS method. The nonparametric bootstrap method refers to the method described in Section 2.3 while treating coefficient vector A constant (this is identical to Scenario S2). The parametric MCS method refers to the method described in Section 2.2.
Table 5 shows that the mean GHG emission was the largest for the control, followed by the nonparametric bootstrap method and parametric MCS. The difference between the largest and smallest was less than 0.7%. This indicates that three mean GHG emissions were essentially identical.
In the case where is no simulated dataset for the control, we use the sample mean to estimate the true mean of the population, and, as such, the sample mean is the unbiased estimator [28]. Thus, the bias of the control is 0. The bias of the nonparametric bootstrap was smaller than that of the parametric MCS, as shown in Table 5; however, both are not 0. This indicates that both the bootstrap method and MCS method do not generate a sample space closely resembling the population, although the bootstrapped datasets are much closer to the population compared to the random variate datasets.
The standard error of the nonparametric bootstrap method was the smallest for the nonparametric bootstrap method, followed by the control and the largest for the parametric MCS method. The CI width and U of the nonparametric bootstrap method give a much narrower CI width and a smaller U when compared to those of the parametric MCS method. The control is situated in between the two. The CI width and U of the MCS method were approximately 10.27 and 10.28 times of those of the bootstrap method, respectively. The CI width and U of the control were approximately 5.97 and 5.94 times of those of the bootstrap method, respectively. Significantly large CI and U for the parametric MCS method and the control compared to the nonparametric bootstrap method indicate that the nonparametric bootstrap method is much more reliable than the parametric MCS method and the control (no simulation) in estimating the uncertainty of Z.
When using a normal distribution for a parameter with large uncertainty, there is a risk of having extremely large values, that is, values orders of magnitude larger than the mean value. Extreme values are an often occurring quality for the distribution of activity data and emission factors [12]. Krezo et al. [36] investigated the effect of bootstrapping resample sizes on the statistics such as the mean, standard deviation, bias, skewness, and confidence levels of the GHG emission intensity due to rail maintenance. The analysis showed that there is a very small bias when compared with the field data. The standard deviation and standard error were less than those of the field study. This supports the findings in this study that the bootstrapping method gives the smallest CI compared to the MCS method and the field data (no simulation) case.
The reliability of a parametric statistical method depends on the validity of an underlying probability distribution; however, the nonparametric bootstrap method does not [27]. It is generally more reliable (often much more so) than the parametric MCS method when the distribution that the parametric method relies upon for its validity does not exist, and it is often almost as reliable as a parametric approach when the distribution that the parametric method relies upon for its validity does exist [27].
Although estimated distribution was considered the most plausible from the goodness-of-fit tests and CDF plots, this does not mean that the estimated distribution fits the original dataset accurately. After all, estimation was approximate at best. Since the distribution that the parametric MCS method relies upon for its validity does not exist, the parametric MCS method did not give reliable uncertainty estimates compared to the nonparametric bootstrap method.
Since the nonparametric bootstrap method does not require estimation of the probability distribution of the dataset, it has an innate advantage over the parametric MCS method in the uncertainty analysis [19,29]. In addition, bootstrapping generates many bootstrapped datasets having the smallest standard error of the three methods and smaller bias compared to the parametric MCS method. As such, the nonparametric bootstrap method gives the lowest U (smallest CI width) among the three methods. This indicates that a collection of the bootstrapped datasets resembles the population of the original dataset such that a more accurate estimation of the statistics of Z can be possible. Therefore, it is reasonable to conclude that the estimation of the probability distribution of the original dataset could be the cause of the uncertainty of GHG emissions for the parametric MCS method.
There are a considerable number of studies that produced similar conclusions to those of this study, although they are not identical. Park et al. [37] collected data of the key variables repeatedly to reduce uncertainty of the model output and reduce the coefficient of variation (CV) value from 10.76% to 5.64%. Szczesny et al. [38] estimated uncertainty of the model output by applying various probability distributions including normal, uniform, and triangular etc. to the input data. From this study they found that discrepancy of the model output can be as high as 40%, and emphasized the importance of choosing accurate probability distribution of the input variable. Lee et al. [29] and Lee et al. [19] compared uncertainties of the model output by applying simulation methods including Monte Carlo simulation (MCS) and bootstrap, and found that the bootstrap method yielded much lower uncertainty compared to that of the MCS method.
Lee et al. [29] and Chen and Corson [39] defined activity data as well as emission factor as random variables as in this study and observed that uncertainty of the model output tends to decrease by 12 to 15% in the MCS method and by 13 to 20% in the bootstrap method.
From the above discussions, it is clear that uncertainty analysis of GHG emissions should be based on the nonparametric bootstrap method rather than the parametric MCS method. This applies, however, only when the estimated PDF of the original dataset is incorrect; nonetheless, quite often this is the case.

4. Conclusions

In recent years, the need for the uncertainty analysis of greenhouse gas (GHG) emissions has increased. It is a general practice to quantify uncertainty using a deterministic method such as the error propagation method, where statistics including standard error of the mean and variance, among others, were used. However, there are stochastic simulation methods available in the statistical inference field. In general, the stochastic simulation method gives more accurate results than the deterministic method. Besides, the GHG emission model relies on the IPCC tier GHG emission factor recommended by the IPCC. The tier I emission factor is a default value, lacks specificity and is intended to be applied to the broader area, and, as such, it has inherently high uncertainty. Thus, GHG emissions calculated from the model is envisaged to have high uncertainty.
This study aims at exploring the possibility of minimizing the uncertainty of GHG emissions computed from the model by applying two stochastic simulation methods: Monte Carlo simulation and the bootstrap method. The cause of the difference between the two methods in the GHG emission uncertainty was investigated and identified such that we can recommend one method over the other for the uncertainty analysis. Statistics used for expressing the uncertainty in the stochastic method, unlike the statistics used in the error propagation methods, were confidence interval and relative uncertainty (U). Furthermore, we tested the effect of constant and random GHG emission factors on the uncertainty of the GHG emission results. Finally, the number of data for the input variable in the GHG emission model was determined based on the asymptotic value of the GHG emission model results.
Many ongoing activities around the world attempt to quantify GHG emissions; however, most of the GHG emission results lack uncertainty information. This is mainly because the uncertainty analysis method was difficult to use, and no credible uncertainty analysis method was available. This study identified an easy-to-use and accurate uncertainty analysis method, that is, the bootstrap method, as well as a general procedure to apply it to the uncertainty analysis of GHG emissions. This will expedite the application of the uncertainty analysis to the quantification of GHG emissions worldwide.
The major conclusions of this study are:
  • There is a certain number of observations of a dataset required before an asymptotic value of the model output can be reached. In this study, the number of observations was 36 for Scenario S3 and 60 for scenarios S1 and S2.
  • Bootstrapping reduces the variance and standard error of Z. This is because bootstrapping generates a bootstrapped dataset resembling the population of the original dataset. The variance of Z was the smallest in Scenario S3 (bootstrapping both A and X), followed by in Scenario S2 (bootstrapping X only), and the largest was observed in Scenario S1 (no bootstrapping).
  • Uncertainty analysis of GHG emissions should be based on the nonparametric bootstrap method, not the parametric MCS method, when the estimated PDF of the original dataset is incorrect.
  • A mathematical model for estimating GHG emissions of a system should consider treating the GHG emission factor as a random variable.

Author Contributions

Conceptualization, K.M.L.; methodology, K.M.L. and M.H.L.; software, J.S.L.; validation, J.Y.L., J.S.L. and K.M.L.; formal analysis, investigation, resources, data curation, M.H.L.; writing—original draft preparation, K.M.L.; writing—review and editing, K.M.L.; visualization, J.S.L.; supervision, K.M.L.; project administration, M.H.L.; funding acquisition, K.M.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the “Cooperative Research Program for Agriculture Science and Technology Development” of the Korea Rural Development Administration (Project No. PJ011762022018).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. The feed and utility input variables and their values of dairy cow.
Table A1. The feed and utility input variables and their values of dairy cow.
MonthFeed for Dry Cows 1Feed for Lactating Cows 1Straw 1Soybean Meal 1Electricity 2Diesel 3
Mon.X1X2X3X4X5X6
111429954131615
226041113171135
327832034181213
4236676234616022
524874138142255
618338433162165
735844335131096
8172786271912510
98329433141345
10732211816274
1182411619355
12572395413813
1311434326171034
1426036525141093
1527836733191123
1623642231121285
1724838032191204
1818340530201334
1935840529181315
2017238931221194
218338730151114
227339029121125
23841834211265
245740729231215
2511437724201213
262603482216994
2727837128211164
2823638929181195
2924838329171184
3018339630191234
3135840034211234
3217240333231266
338340634281264
347342534161294
35841536171334
365740129141164
3711439026191154
3826035822191065
3927841718231455
4023643233211415
4124845532151444
4218343632181366
4335834926171084
4417240127181225
458337629181154
467341029221225
47836129101125
485737829241215
4911440831171225
502603553191124
5127836626151074
5223640534131235
5324834032231093
5418338431181162
5535845434211363
5617240632211195
578341633151234
587342329201194
59838027191184
605738329171234
6111435530141074
6226035125161091
6327838128221177
6423638131151236
6524844434251385
6618340327191224
6735840633161203
6817245438201425
698344837261264
707338023221153
71838635221144
725737522141134
EF 10.380.640.950.710.503.3
Notes: 1 kg CO2-eq/kg; 2 kWh electricity/head-month; 3 kg diesel/head-month.

References

  1. Chang, C.-C.; Chung, C.-L. Greenhouse gas mitigation policies in Taiwan’s road transportation sectors. Energy Policy 2018, 123, 299–307. [Google Scholar] [CrossRef]
  2. Vázquez-Rowe, I.; Ziegler-Rodriguez, K.; Laso, J.; Quispe, I.; Aldaco, R.; Kahhat, R. Production of cement in Peru: Understanding carbon-related environmental impacts and their policy implications. Resour. Conserv. Recycl. 2019, 142, 283–292. [Google Scholar] [CrossRef]
  3. Baek, C.-Y.; Lee, K.-M.; Park, K.-H. Quantification and control of the greenhouse gas emissions from a dairy cow system. J. Clean. Prod. 2014, 70, 50–60. [Google Scholar] [CrossRef]
  4. Shams, S.; Sahu, J.N.; Shamimur Rahman, S.M.; Ahsan, A. Sustainable waste management policy in Bangladesh for reduction of greenhouse gases. Sustain. Cities Soc. 2017, 33, 18–26. [Google Scholar] [CrossRef]
  5. Intergovernmental Panel on Climate Change (IPCC). Revised 1996 IPCC Guidelines for National Greenhouse Gas Inventories; Cambridge University Press: Cambridge, UK, 1996. [Google Scholar]
  6. Intergovernmental Panel on Climate Change (IPCC). 2006 IPCC Guidelines for National Greenhouse Gas Inventories; Cambridge University Press: Cambridge, UK, 2006. [Google Scholar]
  7. Mir, K.A.; Park, C.; Purohit, P.; Kim, S. Comparative analysis of greenhouse gas emission inventory for Pakistan: Part I energy and industrial processes and product use. Adv. Clim. Chang. Res. 2020, 11, 40–51. [Google Scholar] [CrossRef]
  8. McMurray, A.; Pearson, T.; Casarim, F. Guidance on Applying the Monte Carlo Approach to Uncertainty Analysis in Forestry and Greenhouse Gas Accounting. 2017, pp. 4–26. Available online: https://www.winrock.org/wp-content/uploads/2018/03/UncertaintyReport-12.26.17.pdf (accessed on 5 December 2019).
  9. Greenhous Gas Protocol. GHG Protocol Guidance on Uncertainty Assessment in GHG Inventories and Calculating Statistical Parameter. 2003, pp. 2–15. Available online: https://ghgprotocol.org/sites/default/files/ghg-uncertainty.pdf (accessed on 1 January 2020).
  10. Intergovernmental Panel on Climate Change (IPCC). Guidelines for National Greenhouse Gas Inventories; IPCC: Hayama, Japan, 2003; Volume 1, pp. 6–63. Available online: https://www.ipcc-nggip.iges.or.jp/public/2006gl/pdf/1_Volume1/V1_3_Ch3_Uncertainties.pdf (accessed on 4 February 2020).
  11. Huijbregts, M.A.; Gilijamse, W.; Ragas, A.M.; Reijnders, L. Evaluating uncertainty in environmental life cycle assessment. A case study comparing two insulation options for a Dutch one-family dwelling. Environ. Sci. Technol. 2003, 37, 2600–2608. [Google Scholar] [CrossRef] [PubMed]
  12. Fauser, P.; SØrensen, P.B.; Nielsen, M.; Winther, M.; Plejdrup, M.S.; Hoffmann, L.; GyldenkÆrne, S.; Mikkelsen, H.M.; Albrektsen, R.; Lyck, E.; et al. Monte Carlo (Tier 2) uncertainty analysis of Danish Greenhouse gas emission inventory. Greenh. Gas Meas. Manag. 2011, 1, 145–160. [Google Scholar] [CrossRef]
  13. Tellinghuisen, J. Statistical error propagation. J. Phys. Chem. A. 2001, 105, 3917–3921. [Google Scholar] [CrossRef]
  14. Bevington, P.R.; Robinson, D.K. Data Reduction and Error Analysis; McGraw-Hill: New York, NY, USA, 2003; pp. 36–50. [Google Scholar]
  15. Super, I.; Dellaert, S.; Visschedijk, A.; van der Gon, H. Uncertainty analysis of a European high-resolution emission inventory of CO2 and CO to support inverse modelling and network design. Atmos. Chem. Phys. 2020, 20, 1795–1816. [Google Scholar] [CrossRef] [Green Version]
  16. LEE, M.H.; LEE, J.S.; LEE, J.Y.; KIM, Y.H.; PARK, Y.S.; LEE, K.M. Uncertainty Analysis of a GHG Emission Model Output Using the Block Bootstrap and Monte Carlo Simulation. Sustainability 2017, 9, 1522. [Google Scholar] [CrossRef] [Green Version]
  17. Intergovernmental Panel on Climate Change (IPCC). Good Practice Guidance and Uncertainty Management in National Greenhouse Gas Inventories. 2000. Available online: https://www.ipcc-nggip.iges.or.jp/public/gp/english/6_Uncertainty.pdf (accessed on 5 February 2020).
  18. Hong, J.; Shen, G.Q.; Peng, Y.; Feng, Y.; Mao, C. Uncertainty analysis for measuring greenhouse gas emissions in the building construction phase: A case study in China. J. Clean. Prod. 2016, 129, 183–195. [Google Scholar] [CrossRef] [Green Version]
  19. Lee, J.S.; Lee, M.H.; Chun, Y.-Y.; Lee, K.M. Uncertainty Analysis of the Water Scarcity Footprint Based on the AWARE Model Considering Temporal Variations. Water 2018, 10, 341. [Google Scholar] [CrossRef] [Green Version]
  20. Heijungs, R.; Lenzen, M. Error propagation methods for lca—A comparison. Int. J. LCA 2014, 19, 1445–1461. [Google Scholar] [CrossRef]
  21. Kutner, M.; Nachtscheim, C.; Neter, J.; Li, W. Applied Linear Statistical Models, 5th ed.; McGraw Hill: New York, NY, USA, 2005; pp. 458–464. [Google Scholar]
  22. Delignette-Muller, M.L.; Dutang, C. fitdistrplus: An R Package for Fitting Distributions. J. Stat. Softw. 2015, 64, 1–34. [Google Scholar] [CrossRef] [Green Version]
  23. Vito, R. Fitting Distributions with R. 2005, pp. 3–23. Available online: https://cran.r-project.org/doc/contrib/Ricci-distributions-en.pdf (accessed on 2 February 2020).
  24. Delignette-Muller, M.L.; Dutang, C.; Pouillot, R.; Denis, J.-B.; Siberchicot, A. Fitdistrplus-Package. Help to Fit of a Parametric Distribution to Non-Censored or Censored Data. 2019, pp. 39–42. Available online: https://cran.r-project.org/web/packages/fitdistrplus/fitdistrplus.pdf (accessed on 1 January 2020).
  25. Maydeu-Olivares, A.; García-Forero, C. Goodness-of-Fit Testing. Int. Encycl. Educ. 2010, 7, 190–196. [Google Scholar]
  26. Wikipedia. Goodness of Fit. Available online: https://en.wikipedia.org/wiki/Goodness_of_fit (accessed on 5 December 2019).
  27. Significance. What is the bootstrap? Available online: https://www.significancemagazine.com/science/608-what-is-the-bootstrap (accessed on 7 December 2019).
  28. James, G.; Witten, D.; Hastie, T.; Tibshirani, R. An Introduction to Statistical Learning with Application in R; Springer: Berlin/Heidelberg, Germany, 2013; pp. 187–197. [Google Scholar]
  29. Lee, J.Y.; Lee, M.H.; Lee, J.S.; Chun, Y.-Y.; Kim, K.H.; Kim, M.S.; Lee, K.M. Developing emission factors for dairy cow enteric fermentation in Korea. J. Clean. Prod. 2018, 198, 754–762. [Google Scholar] [CrossRef]
  30. Freedman, D.A. Statistical Models Theory and Practice Revised Edition; Cambridge University Press: Cambridge, UK, 2009; pp. 45–46. [Google Scholar]
  31. Robert, K. R in Action: Data Analysis and Graphics with R; Manning: Shelter Island, NY, USA, 2011; pp. 303–309. [Google Scholar]
  32. DiCiccio, T.J.; Efron, B. Bootstrap Confidence Intervals. Stat. Sci. 1996, 11, 189–212. [Google Scholar]
  33. UCLA Institute for Digital Research and Education. How Can I Generate Bootstrap Statistics in R. Available online: https://stats.idre.ucla.edu/r/faq/how-can-i-generate-bootstrap-statistics-in-r/ (accessed on 10 December 2019).
  34. Tong, L.I.; Saminathan, R.; Chang, W.C. Uncertainty Assessment of Non-normal Emission Estimates Using Non-Parametric Bootstrap Confidence Intervals. J. Environ. Inform. 2016, 28, 61–70. [Google Scholar] [CrossRef] [Green Version]
  35. Orloff, J.; Bloom, J. Bootstrap Confidence Intervals Class 24, 18.05, Spring 2014; MIT: Boston, MA, USA, 2014; pp. 1–10. Available online: https://ocw.mit.edu/courses/mathematics/18-05-introduction-to-probability-and-statistics-spring-2014/ (accessed on 1 November 2019).
  36. Krezo, S.; Mirza, O.; He, Y.; Kaewunruen, S. Bootstrap statistical analysis of GHG emission from railway maintenance and renewal projects. In Proceedings of the Third International Conference on Railway Technology: Research, Development and Maintenance, Cagliari, Italy, 5–8 April 2016. [Google Scholar] [CrossRef]
  37. Park, Y.-S.; Yeon, S.-M.; Lee, G.-Y.; Park, G.-H. Proposed Consecutive Uncertainty Analysis Procedure of the Greenhouse Gas Emission Model Output for Products. Sustainability 2019, 11, 2712. [Google Scholar] [CrossRef] [Green Version]
  38. Szczesny, S.; Golijanek-Jędrzejczyk, A.; Świsulski, D. Impact of probability distribution on the uncertainty of resistance measurement. EDP Sci. 2019, 28, 1038. [Google Scholar] [CrossRef] [Green Version]
  39. Chen, X.; Corson, M.S. Influence of emission-factor uncertainty and farm-characteristic variability in LCA estimates of environmental impacts of French dairy farms. J. Clean. Prod. 2014, 81, 150–157. [Google Scholar] [CrossRef]
Figure 1. The effect of the size of the dataset (n) on the values of U.
Figure 1. The effect of the size of the dataset (n) on the values of U.
Energies 13 04965 g001
Figure 2. The effect of the size of the dataset (n) on the mean greenhouse gas (GHG) emission.
Figure 2. The effect of the size of the dataset (n) on the mean greenhouse gas (GHG) emission.
Energies 13 04965 g002
Figure 3. The effect of fab on the values of U for the bootstrap method.
Figure 3. The effect of fab on the values of U for the bootstrap method.
Energies 13 04965 g003
Figure 4. Cumulative distribution function (CDF) plots of the empirical distribution and fitted (theoretical) distributions against the input values of x for the lactating cow feed data.
Figure 4. Cumulative distribution function (CDF) plots of the empirical distribution and fitted (theoretical) distributions against the input values of x for the lactating cow feed data.
Energies 13 04965 g004
Table 1. Summary of the three different scenarios.
Table 1. Summary of the three different scenarios.
ScenarioVar(Z) CalculationTreatment of A Vector
S1Error propagation method (no bootstrapping)Keeping A vector constant
S2Bootstrapping X matrix onlyKeeping A vector constant
S3Bootstrapping both X matrix and A vectorTreating A vector as a
random variable
Table 2. Goodness-of-fit criterion for the lactating cow feed data case.
Table 2. Goodness-of-fit criterion for the lactating cow feed data case.
Goodness-of-Fit CriterionProbability Distribution
normallognormalWeibullgammaloglogistic
AIC845.9826.5864.9830.9803.5
BIC850.5831.1869.5835.4808.1
Table 3. Chosen distribution of the data with relevant statistics.
Table 3. Chosen distribution of the data with relevant statistics.
Column (Data) NameDistributionMeanSd 1ShapeScale
Dry cow feedWeibull 1.57189.82
Lactating cow feedloglogistic 12.28390.54
Strawnormal29.366.36
Soy bean mealloglogistic 6.9717.81
Electricityloglogistic 10.92120.97
Dieselloglogistic 6.204.36
1 Standard deviation.
Table 4. Comparison of the confidence interval computation methods.
Table 4. Comparison of the confidence interval computation methods.
StatisticPercentileNormalBasic
95% CI(418.13–458.57)(418.55–458.00)(416.88–457.32)
CI width40.4439.4540.44
U (%)4.644.514.64
Table 5. Uncertainty of GHG emissions of the three uncertainty analysis methods: no simulation, nonparametric bootstrap and parametric Monte Carlo simulation (MCS) (unit: kg CO2-eq/head-year).
Table 5. Uncertainty of GHG emissions of the three uncertainty analysis methods: no simulation, nonparametric bootstrap and parametric Monte Carlo simulation (MCS) (unit: kg CO2-eq/head-year).
StatisticUncertainty Analysis Method
No SimulationNonparametric BootstrapParametric MCS
Mean438.27437.72435.36
Bias0−0.55−2.91
Standard error61.6610.07105.47
95% CI(317.41–559.14)(418.13–458.57)(263.64–679.29)
CI width241.7340.44415.65
U (%)27.604.6447.74

Share and Cite

MDPI and ACS Style

Lee, K.M.; Lee, M.H.; Lee, J.S.; Lee, J.Y. Uncertainty Analysis of Greenhouse Gas (GHG) Emissions Simulated by the Parametric Monte Carlo Simulation and Nonparametric Bootstrap Method. Energies 2020, 13, 4965. https://doi.org/10.3390/en13184965

AMA Style

Lee KM, Lee MH, Lee JS, Lee JY. Uncertainty Analysis of Greenhouse Gas (GHG) Emissions Simulated by the Parametric Monte Carlo Simulation and Nonparametric Bootstrap Method. Energies. 2020; 13(18):4965. https://doi.org/10.3390/en13184965

Chicago/Turabian Style

Lee, Kun Mo, Min Hyeok Lee, Jong Seok Lee, and Joo Young Lee. 2020. "Uncertainty Analysis of Greenhouse Gas (GHG) Emissions Simulated by the Parametric Monte Carlo Simulation and Nonparametric Bootstrap Method" Energies 13, no. 18: 4965. https://doi.org/10.3390/en13184965

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop