Next Article in Journal
Poleward Migration of Spatial Asymmetry in Tropical Cyclone Precipitation over the Western North Pacific
Previous Article in Journal
Time-Frequency Domain Variation Analysis and LSTM Forecasting of Regional Visibility in the China Region Based on GSOD Station Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Reconstructing PM2.5 Data Record for the Kathmandu Valley Using a Machine Learning Model

1
Climate and Radiation Laboratory, NASA Goddard Space Flight Centre, Greenbelt, MD 20771, USA
2
Goddard Earth Sciences Technology and Research II, Morgan State University, Baltimore, MD 21251, USA
*
Author to whom correspondence should be addressed.
Atmosphere 2023, 14(7), 1073; https://doi.org/10.3390/atmos14071073
Submission received: 7 May 2023 / Revised: 21 June 2023 / Accepted: 22 June 2023 / Published: 25 June 2023
(This article belongs to the Section Air Quality)

Abstract

:
This paper presents a method for reconstructing the historical hourly concentrations of Particulate Matter 2.5 (PM2.5) over the Kathmandu Valley from 1980 to the present. The method uses a machine learning model that is trained using PM2.5 readings from the US Embassy (Phora Durbar) as a ground truth, and the meteorological data from the Modern-Era Retrospective Analysis for Research and Applications v2 (MERRA2) as input. The Extreme Gradient Boosting (XGBoost) model acquires a credible 10-fold cross-validation (CV) score of ~83.4%, an r2-score of ~84%, a Root Mean Square Error (RMSE) of ~15.82 µg/m3, and a Mean Absolute Error (MAE) of ~10.27 µg/m3. Further demonstrating the model’s applicability to years other than those for which truth values are unavailable, the multiple cross-tests with an unseen data set offered r2-scores for 2018, 2019, and 2020 ranging from 56% to 67%. The model-predicted data agree with true values and indicate that MERRA2 underestimates PM2.5 over the region. It strongly agrees with ground-based evidence showing substantially higher mass concentrations in the dry pre- and post-monsoon seasons than in the monsoon months. It also shows a strong anti-correlation between PM2.5 concentration and humidity. The results also demonstrate that none of the years fulfilled the annual mean air quality index (AQI) standards set by the World Health Organization (WHO).

1. Introduction

Air pollution poses a severe hazard to both the biotic (living organisms) and abiotic (hydrosphere, lithosphere, and atmosphere) components of our environment, as they are interconnected [1]. The rapid modernization of civilization, which encompasses increased traffic, construction, and industrialization, has had a significant impact on our air quality. The Particulate Matter 2.5 (PM2.5) particles, with an aerodynamic diameter of less than 2.5 µm, make a substantial contribution to air pollution. Exposure to both smaller and larger airborne particles is detrimental, but PM2.5 particulate matter directly contributes to cardiovascular and respiratory illnesses as well as mortality [2,3]. These particles can occasionally be toxic due to their chemical composition, which includes organic molecules, biological components, sulfate, nitrate, acid, and more [4].
The leading causes of high PM2.5 concentrations in urban areas include rapid urbanization (construction), road dust, increased energy (fossil) consumption, and inefficient combustion [5]. On the other hand, major wildfires produce a significant amount of aerosols globally and some of them cause the formation of pyro-cumulonimbus (Pyrocbs) clouds [6], with some even reaching the stratosphere [7]. Pyrocbs are fire-triggered clouds of smoke. However, most of these wildfires contribute to surface aerosols and serve as sources of (PM2.5) particles.
Nepal was among the ten nations with the worst air quality in the world in 2019, according to a report from the Health Effects Institute [8]. Nepal’s capital city, Kathmandu, is surrounded by other megacities with large population densities. Kathmandu has been identified as the most polluted city in Asia, according to Parajuly (2016) [9]. Temperature inversions caused by the unique mountainous environment of Kathmandu trap the polluted air. Pollution in Kathmandu includes the rising number of new and used diesel-engine vehicles, big trucks hauling sand and building supplies, deteriorating and unpaved roads, and hazardous metal operations on the streets near construction sites. Additionally, the extended disorganized rubbish management in open areas is a factor. Numerous brick-and-block production factories are dotted throughout the Kathmandu Valley’s three districts: Kathmandu, Bhaktapur, and Lalitpur, as well as on its outskirts [10]. The number of brick kilns in the Kathmandu Valley has increased by 200% and about 500 are in operation during the dry seasons [11]. These kilns contaminate the atmosphere by spewing smoke and dust [12].
The majority of people in Nepal follow the Hindu faith. A deceased person’s body is burned during cremation in the Hindu faith. In Nepal, open-air cremations are commonly performed, although electric indoor cremation has recently been practiced. The Pashupatinath temple, located on the bank of the Bagmati River in Kathmandu, is regarded as the holiest cremation site in Nepal. Kathmandu is a sacred city with several rivers of religious significance, including the Bagmati, Bishnumati, and others. Particulate matter (benzene, mercury, and polychlorinated dibenzodioxins and furans) is created during this process [13,14]. In light of this, Kathmandu’s PM2.5 compositions are distinct from those of the rest of the world (except for some Indian cities). Overall, the PM2.5 in Kathmandu is a unique and complex mixture of organic and inorganic materials.
Meteorological parameters along with precipitation are the major factors for the PM2.5 concentration distribution and washouts [15,16,17]. The movement of PM2.5 from emission sources to other places can also be impacted by the direction and speed of the wind. The mixing and vertical dispersion of pollutants, especially PM2.5, can be influenced by temperature and air stability [18]. During steady atmospheric conditions (such as temperature inversions), pollutants frequently become trapped near the surface, leading to higher PM2.5 concentrations. In contrast, unstable conditions promote vertical mixing and dispersion, which reduce PM2.5 concentrations. PM2.5 concentrations are not directly impacted by atmospheric pressure. However, variations in atmospheric pressure can have an impact on wind patterns, which in turn can affect how PM2.5 is transported and dispersed. Therefore, existing sources, weather patterns, and geological characteristics all directly influence the occurrence of PM2.5 [19]. Their dispersal is mainly influenced by wind and atmospheric stability [1,20]. Temperature, pressure, water vapor concentration, and other factors all affect the removal process, chemical production, and conversion of them. Thus, by identifying PM2.5 precursors, we may increase our comprehension of the effects of PM2.5 on several facets of life. A solid track record of air quality measurement is essential for pollution management plans. There have not been many long-term studies utilizing PM2.5 to track the decadal trends in Nepal’s air quality. There are very few in situ measurement data available. Becker et al. (2021) and Mahapatra et al. (2019) list some earlier initiatives to detect air pollution in Kathmandu [21,22].
The Kathmandu Valley’s air quality was initially measured in the 1980s, however, seldom and only during specific seasons. The limited air quality measurements at the time included carbon monoxide, nitrous oxide, sulfur dioxide, and nitrous oxide. These investigations disclosed data on the initial pollutant concentrations and the seasonality of the valley’s air pollution [22]. The Nepalese government-built air quality monitoring stations at several locations throughout the valley between 2002 and 2007 to gauge particulate matter (PM10 and PM2.5) concentrations. Because the campaign was so short, no long-term trends could be drawn. Nonetheless, these campaigns showed that the air in urban (Kathmandu Valley) areas is two to four times more polluted than in rural areas [19,23].
Several other air pollution monitoring campaigns have been conducted over the years. From 2003 to 2005, the Ministry of Population and Environment (MOPE) conducted the first extensive monitoring campaigns on particulate air pollutants with the Danish International Development Agency’s (DANIDA) assistance. Nepal Health Research Council (NHRC) in the spring of 2014 [24] and a 2-week campaign in April 2015 [12] were among the critical PM2.5-focused measurements. In the Kathmandu Valley, Black Carbon aerosol mass was measured for the first time in an urban environment between May 2009 and April 2010 [25]. At Paknajole, The International Centre for Integrated Mountain Development (ICIMOD) conducted measurements between February 2013 and January 2014. Since then, other large-scale, multi-country collaboration-based programs (including “Sus-Kat” and “NAMaSTE”) have been implemented to understand the various air quality-related concerns in Nepal [26,27,28].
Nonetheless, a reliable long-term record of air pollution over Kathmandu is still lacking, yet it is crucial for establishing patterns and conducting a social and health impact analysis. To address the issue, we offer a machine learning approach in this research to reconstruct the hourly PM2.5 concentration in the Kathmandu Valley using available meteorological data. This work attempts to bridge the PM2.5 data gap over Kathmandu. A long-term dataset with comprehensive meteorological parameters is needed as input to achieve this goal. Such datasets can be obtained from the Modern-Era Retrospective Analysis for Research and Applications-2 (MERRA-2) reanalysis data from the National Aeronautics and Space Administration (NASA) [29]. This study presents the reconstructed data record of the PM2.5 mass concentrations in the Kathmandu Valley and examines its long-term climatology.

2. Data Sources and Pre-Processing

The US Embassy has set up an ambient air quality monitoring station in Phora Durbar (P.D.), Kathmandu (Latitude: 27.71° N, Longitude: 85.32° E), providing ground-based PM2.5 data on an hourly basis since March 2017. In this study, we utilize these hourly PM2.5 observations as the ground truth values for training the model. The data used for the analysis cover the period from March 2017 to March 2021.
MERRA2 is the atmospheric reanalysis data of the NASA Global Modeling and Assimilation Office [29]. It assimilated aerosol optical depth (AOD) data over the ocean from the Advanced Very High-Resolution Radiometer (AVHRR) from 1979 to 2002 [30]. Similarly, it assimilated the bias-corrected Moderate Resolution Imaging Spectroradiometer (MODIS) AOD from 2002 to the present [31], the Multiangle Imaging Spectro Radiometer (MISR) AOD from 2000 to 2014 (over the bright surface and desert only), and the AOD from ground-based Aerosol Robotic Network (AERONET) (1994 to 2014) [32,33,34,35]. The GOCART model is connected with the GOES atmospheric model to simulate mixed aerosols, including dust, sea salt, black carbon, organic carbon, and sulfate [29,36].
This study utilizes MERRA2 time-averaged hourly data for the Kathmandu Valley (Latitude: 27.5, Longitude: 85.625) from 1980 to 2021. The data used as input include surface pressure, total ozone column, wind speeds, temperature, and total precipitable water vapor. Additionally, the study incorporates information on the extinction, scattering, and mass concentrations of various aerosols, such as black carbon, dust, organic carbon, sulfate, sea salt, and aerosols. Specifically, the Dust Surface Mass Concentration (PM2.5) is considered. In total, there are 28 variables as presented in Table 1 that are included in the model for both training and prediction purposes. These variables collectively contribute to the model’s ability to analyze and make predictions.
The calculation of MERRA2 PM2.5 is typically performed using Equation (1), as described in (https://gmao.gsfc.nasa.gov/reanalysis/MERRA-2/FAQ/#Q4, accessed on 10 December 2022) [32]. However, it should be noted that Equation (1) does not encompass the entire input list of PM2.5. In order to address this limitation, a comprehensive comparison was conducted, taking into account both the PM2.5 values predicted by the machine learning (ML) model and the values calculated using Equation (1). This comparison allowed for overcoming the constraints posed by the incomplete coverage of variables in Equation (1) and provided a more comprehensive analysis.
PM 2.5 = Dust 2.5 + SS 2.5 + BC + 1.6 × OC + 1.375 × Sulfate
In Equation (1), Dust2.5 is the Dust Surface Mass Concentration of dust (with radii < 2.5 µm). The above equation considers the surface mass concentration of dust PM2.5, sea salt2.5, black carbon, and sulfate. Still, it does not account for nitrate emissions (primarily produced by industrial processes and vehicle exhaust), ammonium, silicon, sodium ions, elemental carbons, etc. [29]. A study in China shows that the absence of those elements results in a significant underestimation of PM2.5 retrievals. [15]. Such discrepancies can be verified with trustworthy ground-based measurements, and the long-term bias-corrected data records would be strengthened and supported by simulating the rectified PM2.5 data.
The distribution of PM2.5 is significantly affected by many factors, such as meteorological parameters, surface conditions, pollutant emissions, and population distributions [37]. In our ML model, population distribution is not included as a factor. This is primarily due to the unavailability of a reliable, long-term source of population data that can be consistently incorporated into the model. Estimating the missing PM2.5 components is made possible by comparing the hourly proportion/variation of meteorological variables from MERRA2 with the variation in the PM2.5 precursors that are already accessible. The MERRA2 variables can be related to the missing components in the total mass of PM2.5 and used to offset it partially. For example, the entire ozone column, coupled with the temperature and pressure, all imply a certain concentration of nitrous. Many studies have shown that the NOx concentration directly correlates with the morning formation and evening–night breakdown rates of ozone as well as the variance of volatile organic compounds (VOCs) [38,39]. The shift in temperature, longwave (terrestrial), and shortwave (solar) radiation intensities throughout the day cause such variations. Among VOCs, organic carbon represents a portion of it. Hence, the missing proportion, primarily caused by the nitrate concentration in the local PM2.5, can be reduced by integrating ozone data with metrological factors. The application of machine learning can significantly improve the situation, as demonstrated by the metrics analysis of the model in the test data in the Results section. Table 1 gives a list of factors from MERRA2 that are used for the PM2.5 data record reconstruction.
Although it is conceivable that not every variable we picked directly causes or contributes to PM2.5, their oscillation and fluctuation with other variables suggest and aid in estimating PM2.5. By building a suitable model, machine learning excels at assessing and characterizing the distinctions between them.

3. Machine Learning

Machine learning is potentially a helpful method for capturing the complex inter/intra play of selected variables with the target values. The US Embassy at Phora Durbar has continuous ground-based PM2.5 data going back almost five years, which can serve as the truth for training the machine learning model. A machine learning algorithm makes predictions by mapping input features to a single output based on the relationship between input and truth values (regression or classification). The missing components of Equation (1) have a combined influence on the surface PM2.5 mass concentration, as was mentioned in the previous section. Potentially relevant MERRA2 meteorological variables include local temperature, pressure, relative humidity, wind speed and direction [40,41,42], total ozone columns, and various aerosol extinction and optical depths. All of these elements interact, and that interaction can significantly impact how much PM2.5 is present and distributed in the area [17,43]. With MERRA2 meteorological and environmental data as input features (Table 1) and Phora Durbar ground-based PM2.5 mass concentration as a truth value, we are better equipped to apply and evaluate various Machine Learning models.
There are numerous types of machine learning regression models. To select the best model for this study, we chose to compare the following models: Linear Regression (L.R.), Decision Tree (D.T.), Random Forest (R.F.), and Extreme Gradient Boosting (XGBoosting (X.G.)). The common theme of these methods is to train the model to find the best prediction by minimizing the errors between the output and the input “truth.” The L.R. approach looks for the best fit using multi-linear regression; the D.T. method establishes regression using a tree structure. The model reaches the final results using decision nodes and leaf nodes. The R.F. technique uses multiple independent decision trees to predict a response given a set of predictors. The algorithm then merges the outcomes by averaging the final outcomes. The X.G. method is also a decision tree ensemble algorithm, similar to the R.F. methods in this regard. The difference is that the X.G. method improves a single weak model by combining it with other weak models. It does that through iteratively training a new model using the error residual of the previous model. In other words, the R.F. method generates decision trees in parallel, while the X.G. is a sequential model, where each subsequent tree is dependent on the last one. More details about X.G. will be given later in this section.
The statistical metric R squared (r2-score) value is used to assess the fitness of a regression model. While using 28 input features in Table 1, the performances of each of these models are contrasted in terms of their cross-validation score, r2-score, Mean Absolute Error (MAE), and Root Mean Squared Error (RMSE). Such metrics are calculated as follows:
r 2 s c o r e = 1 R e s i d u a l   S u m   o f   S q u a r e d   E r r o r T o t a l   S u m   o f   S q u a r e d   E r r o r s = 1 i = 1 N   y i y p r e d 2 i = 1 N   y i y m e a n 2
R o o t   M e a n   S q u a r e   E r r o r   R M S E = i = 1 N y i y p r e d 2 N
M e a n   A b s o l u t e   E r r o r M A E = i = 1 N y i y p r e d N
where yi, ypred, and ymean are the ith measurement; the model predicted value, and the mean of truth value, respectively, where N is the total number of samples. Figure 1a demonstrates that the X.G. algorithm has the highest r2-score and Cross-validation score. The lowest MAE and RMSE figures are also produced by X.G. (Figure 1b,c). The most notable regression overall is the X.G. one. In contrast to R.F., which is vulnerable to overfitting in noisy data, it has also been demonstrated that the X.G. efficiently prevents overfitting and minimizes computing complexity [44]. We decided to use the X.G. regression method in light of this.
The collection of serialized decision trees that cooperate to finish a task is at the core of the XGBoost ML. It is superior because it considers the contribution of each tree before building a serial model that incorporates every variable.
The XGBoost algorithm, a gradient-boosted (trees are serialized to reduce the loss function in subsequent trees) decision tree technique, was first introduced by Chen and Guestrin (2016) [45]. Since then, there has been constant growth and progress. Recently, the XGBoost, a scalable machine learning technique that outperforms several widely used existing classifiers and uses tree boosting to prevent overfitting, has caught the attention of researchers. Its multiprocessing algorithm can process massive amounts of data.
Boosting is a group learning method in which weak learners are united to produce strong learners to work together to reduce training errors and boost the model’s performance. Even when random forests fit trees that depend on one another to lessen the bias of the strong learner, boosting is often more effective than bagging. Gradient boosting (G.B.), a sequential training method, devalues models that have previously been incorrectly classified. By parallelizing the training procedure, Extreme Gradient Boosting (XGBoost) improves computational speed while utilizing multiple cores [45].
The base model is first created, which yields outputs for each instance, and then the residue (prediction error) is obtained. A further model is trained to fix the previous error, obtaining a new, improved gain value and an improved model. It continues in sequence until all true values are correctly trained, or a defined number of trees are reached, with the largest gain for each tree. In this way, several models are created with multiple gain values.
The major steps of the XGBoost model are expressed in the summary equations, as explained by Chen and Guestrin (2016) [45]:
First, the model assumes the base model, as shown in Equation (5). For a given dataset having n examples and m features, a tree ensemble model uses K additive functions to predict the output:
y i ^ = φ x i = k = 1 K f k x i  
where the symbols are y i ^ = Model predicted PM2.5, x i = the vector representation of 28 input variables and f k is an independent tree and contains a continuous score on each leaf. Based on the predicted value from Equation (5), a model is trained in an additive manner. As the equation describes, a new tree is created to minimize the error (objective) observed during the previous base model, as expressed by (5). Scores of corresponding leaves sum up the final model prediction. The objective function used to train the model is a sum of a differentiable loss function and a regularization term. So, to learn the set of models, the regularized objective function term is minimized as:
L t = i = 1 n l y i , y ^ i t 1 + f t x i + Regularization   term
.
In Equation (6), l is a training loss function that expresses the relation between the truth ( y i ) and the predicted value y ^ i t . The common way to evaluate it is using the Mean Squared Error (MSE). Similarly, y ^ i t 1 + f t x i   is predicted at iteration t−1. The addition of each new f t balances the error in each iteration. The ‘Regularization term’ in XGBoost is influenced by both the learning rate and the minimum child weight. Additionally, it is dependent on the number of leaf nodes in a tree and the weights assigned to those leaves. Using a gradient boosting approach, the XGBoost algorithm minimizes this objective function (shown in Equation (6)). Each decision tree in the ensemble is trained to minimize the negative gradient of the objective function with respect to the predicted values. The final prediction is the weighted sum of the predictions of all the trees in the ensemble [45]. So, instead of using the average of each tree to forecast the final model output values, one can learn from past mistakes and develop a robust model.

4. Methodology

To build the desired machine learning model, we start with coupling the P.D. PM2.5 data (truth value) to the MERRA2 data (input variables) for the time within a 15-min window. With the 28 features in Table 1 as input, the algorithm was tested and trained using data from March 2017 to March 2021. Note that this period covers a wide range of PM2.5 situations. For example, in 2020, several primary human sources of PM2.5 were substantially lower compared to other years due to the COVID-19 lockdowns.
As the next step, all the data are randomly divided into two groups: 20% is set aside for testing and 80% is used to train the model (80/20% split). The X.G. Model’s numerous hyperparameters are tweaked and tested to choose the ideal combination of input variables and hyperparameters using the ‘Randomized Search’ 10-fold cross-validation scores. Those parameters and their contributions [45,46] are listed below in Table 2:
The results of such pairings are noted in testing data using the r2-score. In testing data, the r2-score ranged from ~64% to ~84% for various combinations of hyperparameters. So, a good variety of properly hyperparameter-tuned models significantly improves the model’s performance. The base model to be deployed is 84%. At first glance, some of the 28 listed criteria appear to be only distantly related to PM2.5. However, combining them will yield the optimal metrics for the model. So, even if their bulk has no direct impact on the PM2.5, their interaction with other variables does. Examples include temperature, pressure, etc., which are not PM2.5 components but suggest the PM2.5 mass variation scenario. In this case, ML is helpful in capturing the variances.
Figure 2 displays the feature ratings of the input variables for the XGBoost’s default importance for the top 10 variables. The meteorological environment affects air quality and PM2.5 levels. The specific humidity seems to play the most prominent role in forecasting the PM2.5 concentration in the Kathmandu Valley. Various sources of PM2.5, such as incomplete combustion, forest fires, dust, etc., rank lower compared to other input components regarding their relative significance. In Section 4, we further explore its relationship to the PM2.5 levels. These rankings demonstrate that when estimating PM2.5 levels, meteorological parameters are more crucial in identifying the actual contributors. As a result, we can rely on the model to take into account some of the missing PM2.5 contributors that are not easily accessible to us in explicit forms, such as the various nitrate compounds, but are well suggested by meteorological data in its implicit form and will be handled by the model in the proper proportions. However, the most accurate way to assess a model’s performance is to test it against data that it has never seen before.

5. Results

The trained model, which was developed using 80% of the data, is then applied to the remaining 20% of the randomly selected unused data from the period spanning 2017 to 2021. The relation between the actual PM2.5 mass concentration and the model-predicted PM2.5 mass concentration is depicted in Figure 3a. The mean absolute error is 10.27 µg/m3, the root mean square error is 15.82 µg/m3, the mean difference is 0.4 µg/m3, and the coefficient of determination between the projected value and the actual value is 84%. It is also revealed that nearly one-third of PM2.5 concentrations are greater than 100 µg/m3, which is a health-frightening amount.
The monthly average time series for the same 20% tested, truth values, and MERRA2 PM2.5 are displayed in Figure 3b. With the 84% coefficient of determination, we anticipate a higher degree of agreement between the truth and the predicted values. Here, averaging them over a month reveals a nearly perfect match, indicating its applicability to climatology and long-term trend monitoring. MERRA2 PM2.5 displays a similar seasonal pattern in months with low-mass concentrations and a better correlation with actual data and predictions. When high PM2.5 concentrations are present, MERRA2 significantly biases low. This comparison demonstrates that the MERRA2 does not sufficiently account for the emission of particles in its computations, which was previously discussed in this paper and in the work by Jin et al. (2022) [15]. Long-term climatology investigations and data creation are sparked by the frequent low-bias MERRA2 representations in various studies and the literature mentioned above, which motivates more research and model use. Figure 3b also illustrates the seasonality in the readings of PM2.5 over Kathmandu. Higher concentrations sustain the periodicity near the end and beginning of the year, while dropping concentrations support it in the middle.
It has been shown that ML development works quite well with split samplings of 80%/20% [47]. Yet, because aerosol concentrations involve complex factors that are constantly changing, we should proceed with caution before pronouncing such a model to be the best. Even while we think that a 20% random sample of the data correctly captures the distribution of the data, there are times when testing data is entirely new, such as when selecting a different year, the ML metrics start to decline. We must, thus, conduct several tests in various scenarios based on earlier predictions that have been proven valid scientifically to construct a usable model. Truth values from 2017 to 2021 also consider that multiple COVID-19 lockdowns in 2020 resulted in a very different PM2.5 period, as mentioned before. In terms of pollution brought on by human activity, such as the significantly decreased usage of vehicles, almost closed kilns, the restricted amount of construction, etc., it is akin to the early 1980s and the 1990s. As a result, it provides the leverage needed to determine whether the model can accurately predict occurrences, such as those in the 1980s and 1990s. To test the validity of this machine learning approach, we purposely left out the data from the year to be assessed when training the model. For example, for the 2018 testing, we used data from 2017 and from 2019 to 2021 to train the model and left out the entire 2018 data as unobserved data to test the model. We used the same strategy in 2019 and 2020. Not having data for all 12 months, 2017 and 2021 are not assessed with this method. This alternative testing methodology does not use an 80/20 split. In this instance, the testing data distribution is entirely new and unseen for the model, which may not be the case for splits of 80/20 %.
The full metrics analysis of all these tests is shown in Figure 4. As seen from the figure, compared to the metrics derived using the 80/20% split, the r2-score decreased, and the MAE and RMSE increased. For example, the r2-score is 67% for the 2018 testing, the MAE is 16.35 µg/m3, and RMSE is 24.23 µg/m3, while the corresponding values when using the 80/20% split are 84%, 10.27 µg/m3, and 15.82 µg/m3, respectively. This is expected because, for the 80/20% split, data from all years are sampled; the test data likely share the same distribution as the training data since they are from the same years. When the entire year is used for testing, the data distribution patterns can differ from the training dataset. In this regard, the test result for the year 2020 is especially suited for evaluating the model’s performance because the aerosol data patterns can be very different due to the pandemic lockdowns. However, the model still performed well, with an r2-score of 56%, an MAE of 18.52, and an RMSE of 25.3 µg/m3.
The monthly averages from the various cross-tests are shown in Figure 5, along with the truth values. Each hue represents the PM2.5 concentration anticipated by the model for each particular year with a shaded standard deviation. It also contains the projected 80/20% split test results, which match a truth value. The remaining test data for each year using the model trained for the remaining year demonstrate good agreement in comparing true monthly averages. If we perform an hour-by-hour examination, the disparity between all projected concentrations and the actual values is more prominent. However, for a trend and seasonal comparison, the repeated cross-tests of the model yield a solid estimate.
We anticipated a wealth of information to be available for the analysis and research into the climatology and history of PM2.5 by looking at the long-term data record. We used the trained model to reconstruct the hourly PM2.5 data from 1980 to 2021. Figure 6a represents the monthly distribution of PM2.5 during all years. It further confirms the seasonality of PM2.5, established in testing the data, over a lengthy period. Based on the distribution of PM2.5 mass, there are primarily two seasonal patterns: rainy and non-rainy. Along with the dynamics and chemistry of the various elements, human-made effects also contribute to seasonal variance, but local weather phenomena play a vital role.
The rainy (monsoon) season typically lasts roughly from June to September in Kathmandu, Nepal [48]. We included May also in this season, which is full of rainy days, some of which might linger for several days and can occasionally result in life-threatening flooding. We noticed that the PM2.5 concentration dropped low starting in May. The PM2.5 concentration was at its lowest point of the year when the rainy season peaked in July and August, with a mean value of 25 µg/m3 and a reasonably small interquartile range, as shown in Figure 6a,b. Figure 7a demonstrates that the distribution of PM2.5 during monsoon months is more concentrated around mean and median values, with a condensed interquartile range. It aids in calculating the Inter Quartile Range (IQR), or the difference between Q3 (upper) and Q1 (lower) quartiles. Analyzing how compact the data distribution is can be helpful too. Two whisker caps up and down, commonly called the outlier threshold, similarly represent the specified maximum and minimum values gathered by the Q3 + 1.5 × IQR and Q1 − 1.5 × IQR (in this case). Since these occurrences do not always match the distribution of most data from the same groups, they are called outliers. However, here, they represent some high-concentration events of PM2.5 rather than outliers. Consequently, taking these into account in subsequent calculations depends on the purpose, result, and impact of these data.
Building and brick manufacturing stoppages and frequent rain-related washouts are among the factors that cause PM2.5 levels to drop during monsoon seasons. On the other hand, pre- and post-monsoon months greatly vary from the median and mean with considerable expansions of many points above 75% and frequently exceeding 175 µg/m3 to show a significant enhancement of PM2.5 in these seasons.
Figure 6b compares the monthly mean PM2.5 between the dry and rainy seasons. The air quality (PM2.5) in Kathmandu is at its healthiest during the rainy months, albeit not meeting World Health Organization (WHO) standards of the Air Quality Index (AQI), as shown in Figure 6b. According to the WHO, the yearly mean PM2.5 AQI for healthy air is 10 µg/m3 and the 24-h daily average is 25 µg/m3 [49]. It clearly distinguishes between the monthly average of PM2.5 for rainy and dry months. For the dry months, the construction (big buildings, houses, roads, etc.) process picks up significantly, the brick factories begin to reopen, there is less rain to wash the dust away, and the lack of moisture causes the muddy roadways to turn dusty. Due to the landscape of the Kathmandu Valley, a temperature inversion can easily develop during the cold season, retaining pollution until there is strong wind assistance or rain washout. As a result, high concentrations are shown to endure. As per these models’ predicted data, the average number of healthy days in Kathmandu from 1980 to 2021 was just 160. It shows how seriously polluted air is in the Kathmandu Valley.
The decadal rolling average of that statistic shows an increase in the PM2.5 concentration from 1980 to 2005. As shown in Figure 7, the annual average has some ups and downs, but the concentration does increase consistently until 2002. The years 2001 and 2002 appear to have had the highest PM2.5 levels. This is also accurately reflected in the five-year rolling average. By combining yearly, half-decadal, and decadal averages, it was possible to summarize the distribution trend throughout the Kathmandu Valley, which had been rising until 2002 before beginning to modestly decline and demonstrating a steady concentration of decadal average PM2.5 concentrations after that. Figure 7 displays the mean of all individuals as a horizontal black baseline, with a concentration of 51.19 µg/m3. The rolling decadal average continuously climbs from 1980 to 2005, with a slight decline. However, specific years exhibit a modest deviation from the mean on occasion. Due to COVID-19 lockdowns, as discussed earlier, 2020 brings a concentration level much lower than typical, which is well demonstrated by the model.
The two most important input factors that influence the model’s predictions are specific humidity and the total amount of water vapor that can precipitate. Figure 8a,b illustrate the strong anti-correlation of PM2.5 with specific humidity. The big blue band in Figure 8b indicates the middle of the year monsoon months, generally from May to September, with a lower PM2.5 mass concentration than the other months. Compared to the monsoon months, other months have greater PM2.5 concentrations. Figure 8a shows a distribution of lower specific humidity during the months with higher PM2.5 concentrations shown in Figure 8b, which is precisely the opposite distribution in Figure 8a,b. It suggests that PM2.5 levels in the Kathmandu Valley are lower during extremely humid seasons and vice versa. This conclusion aligns with Liu et al. (2020) [50]. High humidity is a manifestation of rainy weather, which washes out aerosols. In addition, hygroscopic growth at times of high humidity makes the aerosol particles (PM2.5) heavier and causes them to fall by dry deposition [51]. As a result, PM2.5 concentrations are reduced. In addition, Wang et al. (2013) conducted a thorough analysis of Beijing, China, to look at the contribution of meteorological factors to PM2.5. Beijing’s typically dry, chilly winter has greater PM2.5 levels than the muggy, hot, rainy summer [52]. As expected, similar results were achieved in the projected model and actual PM2.5 concentrations over the Kathmandu Valley confirming the rationality of the model’s performance.

6. Summary

The concentration of PM2.5 is predicted using various machine learning regression models. These models were created and assessed for approximately five years, from March 2017 to March 2021, using surface level MERRA2 time average meteorological and aerosol parameters as input features and Phora Durbar PM2.5 mass concentration data for training. This work generated long-term historical PM2.5 concentrations in the Kathmandu Valley, a metropolis with a high population density. With a high r2-score of almost 84% and a low root mean square of 15.82 µg/m3, the XGBoost machine learning model stands out among the various regression models. In multi-cross tests, it succeeds, demonstrating that the statistics produced are satisfactory for the specified study. The trained model is used to reconstruct the long-term PM2.5 mass concentrations in the Kathmandu Valley beginning in 1980. The primary contribution of this research is a viable approach to close the data gap for PM2.5 in the climatological study of the Kathmandu Valley.
Analysis of the reconstructed PM2.5 time series shows that only around 160 days from 1980 to 2021 fulfill the WHO’s AQI of PM2.5 requirements. It highlights the significant health risks that the Valley’s PM2.5 presents. Both ground-based data and machine learning forecasts support the seasonal trend of the Kathmandu Valley’s PM2.5. This study demonstrates increased trends in PM2.5 from the 1980s to 2002. After 2002, it begins to decrease slightly. The result also shows clearly the COVID-19 lockdowns’ influence on lowering PM2.5 in 2020 in terms of annual mass concentration.
This approach appears to be effective at reconstructing long-term PM2.5 levels. The applicability of the model can be expanded by considering new locations and truth values. By including additional years’ worth of future true value, the accuracy of this model can be further improved. This study’s practical application is to offer a framework for rebuilding PM2.5 data records in areas without comprehensive historical data. Based on the connections with meteorological factors, we may fill in missing PM2.5 data using machine learning models. This reconstructed dataset can be used for a number of tasks, including analyzing trends in air quality, carrying out epidemiological research, and assisting with the formulation of air pollution mitigation legislation.

Author Contributions

Concept initiation, machine learning model design, plan implementation, calculation, and paper drafting are conducted by S.B. Y.Y. provided a concept review and conducted a paper revision. All authors have read and agreed to the published version of the manuscript.

Funding

Funding support for this research is from NASA’s Modeling, Analysis, and Prediction (MAP) managed by David Considine.

Institutional Review Board Statement

Not Applicable.

Informed Consent Statement

Not Applicable.

Data Availability Statement

This analysis’s training and testing data are publicly available. (1) MERRA2 data are available at https://disc.gsfc.nasa.gov/datasets?project=MERRA-2, managed by the NASA Goddard Earth Sciences (GES) Data and Information Services Center (DISC) (accessed on 10 December 2022). (2) US embassy Kathmandu (Phora Durbar) data (ground-based data) available at https://opendatanepal.com/dataset/air-quality-data-in-kathmandu/resource/0d91e90d-db6a-4c23-9274-96c757aaedae (accessed on 10 December 2022). (3) All calculation is carried out by using Python open sources and various libraries such as Sklearn, XGBoost, Pandas, SciPy, NumPy, etc. (4) The model predicted hourly PM2.5 from 1980 to 2021 for Kathmandu Valley will be available at https://github.com/surenbhatta2/Kathmandu_PM2.5 (accessed on 10 December 2022).

Acknowledgments

GESTAR II MSU, Phora Durbar US embassy and for ground-based data, NASA Goddard for MERRA2 data. We extend our sincere appreciation to the five anonymous reviewers for providing valuable feedback and constructive reviews.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Manisalidis, I.; Stavropoulou, E.; Stavropoulos, A.; Bezirtzoglou, E. Environmental and Health Impacts of Air Pollution: A Review. Front. Public Health 2020, 8, 14. [Google Scholar] [CrossRef] [Green Version]
  2. Ma, Z.; Hu, X.; Sayer, A.M.; Levy, R.; Zhang, Q.; Xue, Y.; Tong, S.; Bi, J.; Huang, L.; Liu, Y. Satellite-Based Spatiotemporal Trends in PM2.5 Concentrations: China, 2004–2013. Environ. Health Perspect. 2016, 124, 184–192. [Google Scholar] [CrossRef] [Green Version]
  3. Pope, C.A.; Dockery, D.W. Health Effects of Fine Particulate Air Pollution: Lines That Connect. J. Air Waste Manag. Assoc. 2006, 56, 709–742. [Google Scholar] [CrossRef]
  4. Majumder, A.; Murthy, V.; Bajracharya, R.; Khanal, S.; Islam, K.; Giri, D. Spatial and Temporal Variation of Ambient PM2.5: A Case Study of Banepa Valley, Nepal. Kathmandu Univ. J. Sci. Eng. Technol. 2012, 8, 23–32. [Google Scholar] [CrossRef] [Green Version]
  5. Xu, B.; Lin, B. What Cause Large Regional Differences in PM2.5 Pollutions in China? Evidence from Quantile Regression Model. J. Clean. Prod. 2018, 174, 447–461. [Google Scholar] [CrossRef]
  6. Bhatta, S.; Pandit, A.K.; Loughman, R.P.; Vernier, J.-P. Three-Wavelength Approach for Aerosol-Cloud Discrimination in the SAGE III/ISS Aerosol Extinction Dataset. Appl. Opt. 2023, 62, 3454–3466. [Google Scholar] [CrossRef] [PubMed]
  7. Bhatta, S. High-Altitude Cloud/Aerosol Detection from SAGE III-ISS and Comparison with OMPS/CALIPSO; Hampton University: Hampton, VA, USA, 2021; ISBN 9798759968900. [Google Scholar]
  8. Health Effects Institute. Global Burden of Disease—State of Global Air 2020: A Special Report on Global Exposure to Air Pollution and Its Health Impacts—2020; Health Effects Institude: Boston, MA, USA, 2020. [Google Scholar]
  9. Parajuly, K. Clean up the Air in Kathmandu. Nature 2016, 533, 321. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Thygerson, S.M.; Sanjel, S.; Johnson, S. Occupational and Environmental Health Hazards in the Brick Manufacturing Industry in Kathmandu Valley, Nepal. Occup. Med. Health Aff. 2016, 4, 2–7. [Google Scholar] [CrossRef] [Green Version]
  11. Pariyar, S.K.; Das, T.; Ferdous, T. Environment and Health Impact for Brick Kilns in Kathmandu Valley. Int. J. Sci. Technol. Res. 2013, 2, 184–187. [Google Scholar]
  12. Islam, M.; Jayarathne, T.; Simpson, I.J.; Werden, B.; Maben, J.; Gilbert, A.; Praveen, P.S.; Adhikari, S.; Panday, A.K.; Rupakheti, M. Ambient Air Quality in the Kathmandu Valley, Nepal, during the Pre-Monsoon: Concentrations and Sources of Particulate Matter and Trace Gases. Atmos. Chem. Phys. 2020, 20, 2927–2951. [Google Scholar] [CrossRef] [Green Version]
  13. Green, L.C.; Crouch, E.A.; Zemba, S.G. Cremation, Air Pollution, and Special Use Permitting: A Case Study. Hum. Ecol. Risk Assess. Int. J. 2014, 20, 559–565. [Google Scholar] [CrossRef]
  14. Xue, Y.; Cheng, L.; Chen, X.; Zhai, X.; Wang, W.; Zhang, W.; Bai, Y.; Tian, H.; Nie, L.; Zhang, S. Emission Characteristics of Harmful Air Pollutants from Cremators in Beijing, China. PLoS ONE 2018, 13, e0194226. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Jin, C.; Wang, Y.; Li, T.; Yuan, Q. Global Validation and Hybrid Calibration of CAMS and MERRA-2 PM2.5 Reanalysis Products Based on OpenAQ Platform. Atmos. Environ. 2022, 274, 118972. [Google Scholar] [CrossRef]
  16. Seinfeld, J.H.; Pandis, S.N. Atmospheric Chemistry and Physics: From Air Pollution to Climate Change; John Wiley & Sons: Hoboken, NJ, USA, 2016; ISBN 1-118-94740-1. [Google Scholar]
  17. Gupta, P.; Zhan, S.; Mishra, V.; Aekakkararungroj, A.; Markert, A.; Paibong, S.; Chishtie, F. Machine Learning Algorithm for Estimating Surface PM2.5 in Thailand. Aerosol Air Qual. Res. 2021, 21, 210105. [Google Scholar] [CrossRef]
  18. DeGaetano, A.T.; Doherty, O.M. Temporal, Spatial and Meteorological Variations in Hourly PM2.5 Concentration Extremes in New York City. Atmos. Environ. 2004, 38, 1547–1558. [Google Scholar] [CrossRef]
  19. Aryal, R.K.; Lee, B.-K.; Karki, R.; Gurung, A.; Baral, B.; Byeon, S.-H. Dynamics of PM2.5 Concentrations in Kathmandu Valley, Nepal. J. Hazard. Mater. 2009, 168, 732. [Google Scholar] [CrossRef]
  20. Kelishadi, R.; Poursafa, P. Air Pollution and Non-Respiratory Health Hazards for Children. Arch. Med. Sci. 2010, 6, 483–495. [Google Scholar] [CrossRef]
  21. Becker, S.; Sapkota, R.P.; Pokharel, B.; Adhikari, L.; Pokhrel, R.P.; Khanal, S.; Giri, B. Particulate Matter Variability in Kathmandu Based on In-Situ Measurements, Remote Sensing, and Reanalysis Data. Atmos. Res. 2021, 258, 105623. [Google Scholar] [CrossRef]
  22. Mahapatra, P.S.; Puppala, S.P.; Adhikary, B.; Shrestha, K.L.; Dawadi, D.P.; Paudel, S.P.; Panday, A.K. Air Quality Trends of the Kathmandu Valley: A Satellite, Observation and Modeling Perspective. Atmos. Environ. 2019, 201, 334–347. [Google Scholar] [CrossRef]
  23. Gurung, A.; Bell, M.L. Exposure to Airborne Particulate Matter in Kathmandu Valley, Nepal. J. Expo. Sci. Environ. Epidemiol. 2012, 22, 235–242. [Google Scholar] [CrossRef]
  24. Shrestha, S. Assessment of Ambient Particulate Air Pollution and Its Attribution to Environmental Burden of Disease in Kathmandu Valley, Nepal: A Review. Environ. Anal. Ecol. Stud. 2018, 4, 1–3. [Google Scholar]
  25. Sharma, R.; Bhattarai, B.; Sapkota, B.; Gewali, M.; Kjeldstad, B. Black Carbon Aerosols Variation in Kathmandu Valley, Nepal. Atmos. Environ. 2012, 63, 282–288. [Google Scholar] [CrossRef]
  26. Bhardwaj, P.; Naja, M.; Rupakheti, M.; Lupascu, A.; Mues, A.; Panday, A.K.; Kumar, R.; Mahata, K.S.; Lal, S.; Chandola, H.C. Variations in Surface Ozone and Carbon Monoxide in the Kathmandu Valley and Surrounding Broader Regions during SusKat-ABC Field Campaign: Role of Local and Regional Sources. Atmos. Chem. Phys. 2018, 18, 11949–11971. [Google Scholar] [CrossRef] [Green Version]
  27. Jayarathne, T.; Stockwell, C.E.; Bhave, P.V.; Praveen, P.S.; Rathnayake, C.M.; Islam, M.; Panday, A.K.; Adhikari, S.; Maharjan, R.; Goetz, J.D. Nepal Ambient Monitoring and Source Testing Experiment (NAMaSTE): Emissions of Particulate Matter from Wood-and Dung-Fueled Cooking Fires, Garbage and Crop Residue Burning, Brick Kilns, and Other Sources. Atmos. Chem. Phys. 2018, 18, 2259–2286. [Google Scholar] [CrossRef] [Green Version]
  28. Rupakheti, D.; Adhikary, B.; Praveen, P.S.; Rupakheti, M.; Kang, S.; Mahata, K.S.; Naja, M.; Zhang, Q.; Panday, A.K.; Lawrence, M.G. Pre-Monsoon Air Quality over Lumbini, a World Heritage Site along the Himalayan Foothills. Atmos. Chem. Phys. 2017, 17, 11041–11063. [Google Scholar] [CrossRef] [Green Version]
  29. Gelaro, R.; McCarty, W.; Suárez, M.J.; Todling, R.; Molod, A.; Takacs, L.; Randles, C.A.; Darmenov, A.; Bosilovich, M.G.; Reichle, R.; et al. The Modern-Era Retrospective Analysis for Research and Applications, Version 2 (MERRA-2). J. Clim. 2017, 30, 5419–5454. [Google Scholar] [CrossRef]
  30. Heidinger, A.K.; Cao, C.; Sullivan, J.T. Using Moderate Resolution Imaging Spectrometer (MODIS) to Calibrate Advanced Very High Resolution Radiometer Reflectance Channels. J. Geophys. Res. Atmos. 2002, 107, AAC 11-1–AAC 11-10. [Google Scholar] [CrossRef] [Green Version]
  31. Remer, L.A.; Kaufman, Y.J.; Tanré, D.; Mattoo, S.; Chu, D.A.; Martins, J.V.; Li, R.-R.; Ichoku, C.; Levy, R.C.; Kleidman, R.G.; et al. The MODIS Aerosol Algorithm, Products, and Validation. J. Atmos. Sci. 2005, 62, 947–973. [Google Scholar] [CrossRef] [Green Version]
  32. Buchard, V.; Da Silva, A.; Randles, C.; Colarco, P.; Ferrare, R.; Hair, J.; Hostetler, C.; Tackett, J.; Winker, D. Evaluation of the Surface PM2.5 in Version 1 of the NASA MERRA Aerosol Reanalysis over the United States. Atmos. Environ. 2016, 125, 100–111. [Google Scholar] [CrossRef]
  33. Heidinger, A.K.; Foster, M.J.; Walther, A.; Zhao, X.T. The Pathfinder Atmospheres–Extended AVHRR Climate Dataset. Bull. Am. Meteorol. Soc. 2014, 95, 909–922. [Google Scholar] [CrossRef]
  34. Holben, B.N.; Eck, T.F.; al Slutsker, I.; Tanre, D.; Buis, J.; Setzer, A.; Vermote, E.; Reagan, J.A.; Kaufman, Y.; Nakajima, T. AERONET—A Federated Instrument Network and Data Archive for Aerosol Characterization. Remote Sens. Environ. 1998, 66, 1–16. [Google Scholar] [CrossRef]
  35. Kahn, R.A.; Gaitley, B.J.; Martonchik, J.V.; Diner, D.J.; Crean, K.A.; Holben, B. Multiangle Imaging Spectroradiometer (MISR) Global Aerosol Optical Depth Validation Based on 2 Years of Coincident Aerosol Robotic Network (AERONET) Observations. J. Geophys. Res. Atmos. 2005, 110. [Google Scholar] [CrossRef] [Green Version]
  36. Chin, M.; Ginoux, P.; Kinne, S.; Torres, O.; Holben, B.N.; Duncan, B.N.; Martin, R.V.; Logan, J.A.; Higurashi, A.; Nakajima, T. Tropospheric Aerosol Optical Thickness from the GOCART Model and Comparisons with Satellite and Sun Photometer Measurements. J. Atmos. Sci. 2002, 59, 461–483. [Google Scholar] [CrossRef]
  37. Wei, J.; Li, Z.; Lyapustin, A.; Sun, L.; Peng, Y.; Xue, W.; Su, T.; Cribb, M. Reconstructing 1-Km-Resolution High-Quality PM2.5 Data Records from 2000 to 2018 in China: Spatiotemporal Variations and Policy Implications. Remote Sens. Environ. 2021, 252, 112136. [Google Scholar] [CrossRef]
  38. Chameides, W.; Walker, J.C.G. A Photochemical Theory of Tropospheric Ozone. J. Geophys. Res. 1973, 78, 8751–8760. [Google Scholar] [CrossRef]
  39. David, L.M.; Nair, P.R. Diurnal and Seasonal Variability of Surface Ozone and NOx at a Tropical Coastal Site: Association with Mesoscale and Synoptic Meteorological Conditions. J. Geophys. Res. Atmos. 2011, 116. [Google Scholar] [CrossRef]
  40. Shi, C.; Nduka, I.C.; Yang, Y.; Huang, Y.; Yao, R.; Zhang, H.; He, B.; Xie, C.; Wang, Z.; Yim, S.H.L. Characteristics and Meteorological Mechanisms of Transboundary Air Pollution in a Persistent Heavy PM2.5 Pollution Episode in Central-East China. Atmos. Environ. 2020, 223, 117239. [Google Scholar] [CrossRef]
  41. Wen, H.; Dang, Y.; Li, L. Short-Term PM2.5 Concentration Prediction by Combining GNSS and Meteorological Factors. IEEE Access 2020, 8, 115202–115216. [Google Scholar] [CrossRef]
  42. Xu, Y.; Xue, W.; Lei, Y.; Huang, Q.; Zhao, Y.; Cheng, S.; Ren, Z.; Wang, J. Spatiotemporal Variation in the Impact of Meteorological Conditions on PM2.5 Pollution in China from 2000 to 2017. Atmos. Environ. 2020, 223, 117215. [Google Scholar] [CrossRef]
  43. Marsha, A.; Larkin, N.K. A Statistical Model for Predicting PM2.5 for the Western United States. J. Air Waste Manag. Assoc. 2019, 69, 1215–1229. [Google Scholar] [CrossRef]
  44. Liu, Y.; Luo, H.; Zhao, B.; Zhao, X.; Han, Z. Short-Term Power Load Forecasting Based on Clustering and XGBoost Method. In Proceedings of the 2018 IEEE 9th International Conference on Software Engineering and Service Science (ICSESS), Beijing, China, 23–25 November 2018; pp. 536–539. [Google Scholar]
  45. Chen, T.; Guestrin, C. Xgboost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
  46. XGBoost Parameters—Xgboost 1.7.5 Documentation. Available online: https://xgboost.readthedocs.io/en/stable/parameter.html (accessed on 29 April 2023).
  47. Yang, Y.; Kiv, D.; Bhatta, S.; Ganeshan, M.; Lu, X.; Palm, S. Diagnosis of Antarctic Blowing Snow Properties Using MERRA-2 Reanalysis with a Machine Learning Model. J. Appl. Meteorol. Climatol. 2023. [Google Scholar] [CrossRef]
  48. Shrestha, M. Interannual Variation of Summer Monsoon Rainfall over Nepal and Its Relation to Southern Oscillation Index. Meteorol. Atmos. Phys. 2000, 75, 21–28. [Google Scholar] [CrossRef]
  49. Ambient (Outdoor) Air Pollution. Available online: https://www.who.int/news-room/fact-sheets/detail/ambient-(outdoor)-air-quality-and-health (accessed on 14 June 2023).
  50. Liu, Y.; Zhou, Y.; Lu, J. Exploring the Relationship between Air Pollution and Meteorological Conditions in China under Environmental Governance. Sci. Rep. 2020, 10, 14518. [Google Scholar] [CrossRef] [PubMed]
  51. Wang, J.; Ogawa, S. Effects of Meteorological Conditions on PM2.5 Concentrations in Nagasaki, Japan. Int. J. Environ. Res. Public Health 2015, 12, 9089–9101. [Google Scholar] [CrossRef]
  52. Wang, J.; Wang, Y.; Liu, H.; Yang, Y.; Zhang, X.; Li, Y.; Zhang, Y.; Deng, G. Diagnostic Identification of the Impact of Meteorological Conditions on PM2.5 Concentrations in Beijing. Atmos. Environ. 2013, 81, 158–165. [Google Scholar] [CrossRef]
Figure 1. Performance evaluations of different models (a) The r2-score for training data is shown in the left panel, for testing data is shown in the middle panel, and the overall cross-validation score is shown in the right panel. (b) the testing data’s Root Mean Square Error (RMSE), and (c) the mean absolute error (MAE) in µg/m3.
Figure 1. Performance evaluations of different models (a) The r2-score for training data is shown in the left panel, for testing data is shown in the middle panel, and the overall cross-validation score is shown in the right panel. (b) the testing data’s Root Mean Square Error (RMSE), and (c) the mean absolute error (MAE) in µg/m3.
Atmosphere 14 01073 g001
Figure 2. The first top ten crucial variables from the feature importance list contributed to the XGBoost regressor model training in terms of their percentage contribution normalized to 1.
Figure 2. The first top ten crucial variables from the feature importance list contributed to the XGBoost regressor model training in terms of their percentage contribution normalized to 1.
Atmosphere 14 01073 g002
Figure 3. (a) The scatter plot represents the correlation between model-predicted PM2.5 and Phora Durbar ground-based observations for the testing dataset. The red line represents the best-fit line with the equation along with r2-score, RMSE, and MAE. (b) Monthly average time series of truth (Phora Durbar), predicted randomly selected 20% testing data, and MERRA2 PM2.5 mass concentration. The blue line with a closed circle represents the predicted PM2.5 mass concentration from the model; the red line with a square is the true ground-based Phora Durbar station data; and the green line with a closed circle represents the MERRA2 data.
Figure 3. (a) The scatter plot represents the correlation between model-predicted PM2.5 and Phora Durbar ground-based observations for the testing dataset. The red line represents the best-fit line with the equation along with r2-score, RMSE, and MAE. (b) Monthly average time series of truth (Phora Durbar), predicted randomly selected 20% testing data, and MERRA2 PM2.5 mass concentration. The blue line with a closed circle represents the predicted PM2.5 mass concentration from the model; the red line with a square is the true ground-based Phora Durbar station data; and the green line with a closed circle represents the MERRA2 data.
Atmosphere 14 01073 g003
Figure 4. Comparisons of model performance in multi-tests. It represents the model train-test for an 80/20% split of all the years, individual years 2018, 2019, and 2020 being tested while the remaining years are utilized for training the model in each case.
Figure 4. Comparisons of model performance in multi-tests. It represents the model train-test for an 80/20% split of all the years, individual years 2018, 2019, and 2020 being tested while the remaining years are utilized for training the model in each case.
Atmosphere 14 01073 g004
Figure 5. PM2.5 monthly averages from March 2017 through March 2021 in various scenarios, including actual and model predicted. A different hue represents the various outputs specified in the legends. It comprises a truth Phora Durbar output for all years from the model trained with an 80% and 20% split, model output for 2018 using the remaining data to train, and similarly for 2019 and 2020.
Figure 5. PM2.5 monthly averages from March 2017 through March 2021 in various scenarios, including actual and model predicted. A different hue represents the various outputs specified in the legends. It comprises a truth Phora Durbar output for all years from the model trained with an 80% and 20% split, model output for 2018 using the remaining data to train, and similarly for 2019 and 2020.
Atmosphere 14 01073 g005
Figure 6. (a) Monthly grouping of the entire year’s statistics of the model predicted hourly PM2.5 mass concentrations from 1980 to 2021, showing lower PM2.5 concentrations from May to September with the mean of each month and showing interquartile range and overall distribution among the interquartile range. The center lines of each box represent the median, while the top and bottom margins represent the higher (Q3) and lower (Q1) quartiles, respectively. Two whisker caps are the maximum and minimum for each month. (b) Monthly mean PM2.5 dry and rainy month’s groups show the seasonality for years.
Figure 6. (a) Monthly grouping of the entire year’s statistics of the model predicted hourly PM2.5 mass concentrations from 1980 to 2021, showing lower PM2.5 concentrations from May to September with the mean of each month and showing interquartile range and overall distribution among the interquartile range. The center lines of each box represent the median, while the top and bottom margins represent the higher (Q3) and lower (Q1) quartiles, respectively. Two whisker caps are the maximum and minimum for each month. (b) Monthly mean PM2.5 dry and rainy month’s groups show the seasonality for years.
Atmosphere 14 01073 g006
Figure 7. The various mean yearly, half-decadal, and decadal rolling mean for entire years predicted PM2.5 representing the concentration distribution over the past four decades over Kathmandu Valley. The black horizontal line represents the mean of all hourly data from 1980 to 2021.
Figure 7. The various mean yearly, half-decadal, and decadal rolling mean for entire years predicted PM2.5 representing the concentration distribution over the past four decades over Kathmandu Valley. The black horizontal line represents the mean of all hourly data from 1980 to 2021.
Atmosphere 14 01073 g007
Figure 8. (a) The monthly distribution of the MERRA2 hourly specific humidity; (b) Monthly PM2.5 mass concentration distribution.
Figure 8. (a) The monthly distribution of the MERRA2 hourly specific humidity; (b) Monthly PM2.5 mass concentration distribution.
Atmosphere 14 01073 g008
Table 1. MERRA2 time-average hourly input data for training the model include a list of meteorological data and a list of MERRA2 aerosol-related variables.
Table 1. MERRA2 time-average hourly input data for training the model include a list of meteorological data and a list of MERRA2 aerosol-related variables.
Meteorological data:
Total precipitable water vapor Total column
ozone
Specific humidity
(2M)
Surface Pressure 10M Wind
(U and V)
Temperature (2M)
Aerosols and dust species:
Surface Mass concentration Extinction
(O.D.)
Scattering
(O. D.)
PM2.5
(Extinction)
PM2.5
(Mass)
PM2.5
(Scattering)
Dust Dust Dust Dust Dust Dust
Organic Carbon Organic Carbon Organic Carbon -- -- --
Sea salt -- -- Sea salt Sea salt --
SO4 SO4 SO4 -- -- --
SO2 -- -- -- -- --
-- Total aerosol -- -- -- --
DMS -- -- -- -- --
Black carbon Black carbon Black carbon -- -- --
Table 2. XGBoost Regression Model’s selected tuned hyperparameters and their short description.
Table 2. XGBoost Regression Model’s selected tuned hyperparameters and their short description.
HyperparametersDescription
Eta (learning rate)Controls the step size at each iteration while updating the weights of a decision tree by preventing overfitting and generalization.
n_estimatorNumber of decision trees to be built. The larger number of trees makes the model more complex and better but may lead to overfitting.
max_depthMaximum depth of each decision tree. Higher values allow more splits and capture the complex interaction between each feature.
min_child_weightDetermines the thresholds on the sample to split.
subsampleIt is the fraction of samples to be randomly sampled for each tree in the ensemble.
colsample_bytreeFraction of each tree’s features also helps prevent overfitting.
objectiveSpecifies the loss function to be optimized during training.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bhatta, S.; Yang, Y. Reconstructing PM2.5 Data Record for the Kathmandu Valley Using a Machine Learning Model. Atmosphere 2023, 14, 1073. https://doi.org/10.3390/atmos14071073

AMA Style

Bhatta S, Yang Y. Reconstructing PM2.5 Data Record for the Kathmandu Valley Using a Machine Learning Model. Atmosphere. 2023; 14(7):1073. https://doi.org/10.3390/atmos14071073

Chicago/Turabian Style

Bhatta, Surendra, and Yuekui Yang. 2023. "Reconstructing PM2.5 Data Record for the Kathmandu Valley Using a Machine Learning Model" Atmosphere 14, no. 7: 1073. https://doi.org/10.3390/atmos14071073

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop