Next Article in Journal
Heterogenous Canopy in a Lagrangian-Stochastic Dispersion Model for Particulate Matter from Multiple Sources over the Haifa Bay Area
Next Article in Special Issue
Performance of Bayesian Model Averaging (BMA) for Short-Term Prediction of PM10 Concentration in the Peninsular Malaysia
Previous Article in Journal
Estimating Site-Specific Wind Speeds Using Gridded Data: A Comparison of Multiple Machine Learning Models
Previous Article in Special Issue
A PM2.5 Concentration Prediction Model Based on CART–BLS
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Co-Training Semi-Supervised Learning for Fine-Grained Air Quality Analysis

1
School of Electronics and Information Engineering, Hebei University of Technology, Tianjin 300401, China
2
Convergence Media Center, Hebei University of Technology, Tianjin 300401, China
3
Glasgow College, University of Electronic Science and Technology of China, Chengdu 611731, China
*
Author to whom correspondence should be addressed.
Atmosphere 2023, 14(1), 143; https://doi.org/10.3390/atmos14010143
Submission received: 13 December 2022 / Revised: 29 December 2022 / Accepted: 4 January 2023 / Published: 9 January 2023
(This article belongs to the Special Issue Air Quality Prediction and Modeling)

Abstract

:
Due to the limited number of air quality monitoring stations, the data collected are limited. Using supervised learning for air quality fine-grained analysis, that is used to predict the air quality index (AQI) of the locations without air quality monitoring stations, may lead to overfitting in that the models have superior performance on the training set but perform poorly on the validation and testing set. In order to avoid this problem in supervised learning, the most effective solution is to increase the amount of data, but in this study, this is not realistic. Fortunately, semi-supervised learning can obtain knowledge from unlabeled samples, thus solving the problem caused by insufficient training samples. Therefore, a co-training semi-supervised learning method combining the K-nearest neighbors (KNN) algorithm and deep neural network (DNN) is proposed, named KNN-DNN, which makes full use of unlabeled samples to improve the model performance for fine-grained air quality analysis. Temperature, humidity, the concentrations of pollutants and source type are used as input variables, and the KNN algorithm and DNN model are used as learners. For each learner, the labeled data are used as the initial training set to model the relationship between the input variables and the AQI. In the iterative process, by labeling the unlabeled samples, a pseudo-sample with the highest confidence is selected to expand the training set. The proposed model is evaluated on a real dataset collected by monitoring stations from 1 February to 30 April 2018 over a region between 118° E–118°53′ E and 39°45′ N–39°89′ N. Practical application shows that the proposed model has a significant effect on the fine-grained analysis of air quality. The coefficient of determination between the predicted value and the true value is 0.97, which is better than other models.

1. Introduction

Air quality is one of the main environmental issues in the world. Over the past few decades, rapid urbanization has brought economic maturity, but there is a decoupling trend between economic growth and pollution emission strategies [1]. Severe air pollution poses a persistent threat to humans. According to statistics, about 1 million people die each year from air pollution in China [2], and about 4.24 million people die prematurely each year globally due to air pollution [3]. In addition, air pollution can cause serious ecological damage, such as acid rain, climate change and global warming. The frequency of extreme weather events is increasing [4,5].
In such a situation, air quality is receiving more and more attention. In order to know the air quality at a specific time, the environmental protection department has set up monitoring stations with equipment that can automatically monitor air quality [6]. However, each station requires expensive construction costs and a large number of human resources to maintain them regularly, and the construction may be limited by urban land-use strategies [7,8]. As a result, the number of air quality monitoring stations is often limited, and the spatial distribution is highly uneven. Compared with the occupied area, the effective area covered by the monitoring station is limited. For example, Beijing covers an area of 16,410 km2 but has only 36 monitoring stations; Taiwan covers an area of 36,193 km2 but has only 78 monitoring stations. Therefore, it is necessary to design an adequate air quality fine-grained analysis model to predict the air quality at the location without monitoring stations.
At present, the methods applied to fine-grained analysis of air quality include numerical models and data-driven models. As the atmospheric environment has been increasingly explored, researchers have found ways to simulate the complex dispersion and transport processes of air pollutants using empirical assumptions, and various numerical models that can estimate the pollutant concentrations at any specified location have emerged [9]. Numerical models include physical models and chemistry transport models (CTMs). Physical models include the Gaussian plume model (GPM) [10], the convection–diffusion model, the urban street canyon model [11], computational fluid dynamics (CFD) [12,13], etc. Ghenai et al. [14] used the GPM to simulate the dispersion of PM10 from point sources and area sources in the air around a 35-meter-high factory to obtain the average concentrations over the next 48 h. Rangel et al. [15] used the GPM to estimate the emissions and diffusion of CO, PM2.5 and NOX from burning sugarcane biomass in rural areas of Northeast Brazil. Yang et al. [16] modified the GPM to estimate the transport of ammonia and particulate matter from ventilation tunnel fans of poultry houses. Karim and Nolan [17] developed a CFD model to simulate the dispersion of motor vehicle pollution on motorways, mainly for NO2 concentrations. Tonbon et al. [18] used a CFD model for the 3D simulation of CO in a real city. Kwak et al. [12] used CFD to simulate the concentration distribution of NO2 and O3 based on measured data from roadside air quality monitoring stations. Nie et al. [19] used CFD to simulate the dispersion and distribution of vehicle emissions as a guide for adjusting the location of pollutant monitoring devices and thereby improving air quality.
With the study of turbulence characteristics in the atmospheric boundary layer, researchers have carried out a lot of indoor numerical experiments and field observations, adding more complex meteorological models and nonlinear reaction mechanisms to the physical models. CTMs perform air quality prediction by simulating the chemical processes of emission, transport and mixing of pollutants in the atmosphere. Qiao et al. [20] used a source-oriented version of the Community Multiscale Air Quality (CMAQ) model to determine the source contributions to PM2.5 during 2013 for 25 Chinese province capitals and municipalities. Koo et al. [21] used CAM_X to develop a PM10 emission inventory, and a posteriori emissions to minimize the gap between the predicted and observed PM10 concentrations were proposed. Manders et al. [22] studied the ability of CTM LOTOS-EUROS to predict PM10 concentrations in the Netherlands. However, these numerical models are usually one or several mathematical formulas, and the setting of their parameters may not apply to all atmospheric environments, resulting in limited accuracy of the results.
Data-driven models such as random forest [23,24], combining a feedforward neural network with a recurrent neural network [25,26], K-nearest neighbor (KNN) [27], XGBoost [28], LightGBM [29,30], support vector machine (SVM) [31], deep neural network (DNN) [32], artificial neural network (ANN) [33], graph convolution network (GCN) [34,35], long and short-term memory (LSTM) [36,37], etc., are widely used. Compared with numerical models, data-driven models can automatically extract complex knowledge from data, model the changing trend of air quality and obtain unknown air quality information. Xu et al. [27] proposed a two-stage method for predicting the air quality index (AQI). In the first stage, ANN is used to predict the AQI at some locations to alleviate the problem of data sparseness. In the second stage, the tensor decomposition method is used to predict the complete AQI distribution for all locations. Based on causal analysis, Zhu et al. [38] selected the most relevant data, using back propagation neural networks to obtain the AQI in unobserved grids. Zhong et al. [26] considered the environmental similarity between regions with and without monitoring stations to select the stations most relevant to the target location. The attributes of candidate stations and target locations are input into the regressor composed of a feedforward neural network and LSTM for air quality fine-grained analysis. Hu et al. [32] estimated PM2.5 distribution maps in real-time by spatial KNN and DNN based on data collected by air–ground wireless sensors and urban dynamic information captured online. Liu et al. [35] used GCN to construct the topology of the monitoring stations, introduced LSTM to dynamically capture the correlation between historical data and proposed a GC-LSTM to achieve real-time and future AQI prediction. Han et al. [36] proposed a multichannel attention model that models static and dynamic spatial correlations as independent channels to learn the effect of a station on the target location. Song et al. [39] proposed an FCN-based multipollutant spatio-temporal network to estimate grid-level concentrations of multiple air pollutant species based on fixed station measurements and multisource urban features. Qin et al. [24] proposed an environment-aware locally adaptive deep forest model to infer the fine-grained distribution of NO2 at a resolution of 100 m and 1 h. Hsieh et al. [40] designed a hybrid predictor to predict the long-term AQI of cities without monitoring stations by using dynamic weighting and combining time-related and time-independent predictors. Chen et al. [41] proposed a cascade forest algorithm used to fit the relationship between multisource spatio-temporal characteristics and AQI. Dai et al. [28] built a model based on the GARCH family-XGBoost-MLP to capture the nonlinearity of PM2.5 concentrations due to different meteorological changes and characteristics to describe the spatio-temporal volatility of PM2.5.
There are also several hybrid methods for air quality fine-grained analysis. Ma et al. [42,43] quantitatively combined a convective diffusion model and a neural network for multitask learning, where the neural network learned the mapping of attributes to PM2.5 concentrations and then fitted it to a convective diffusion model. Later, the ConvLSTM was chosen to model the pollution diffusion process, and an autoencoder-based algorithm for air pollution map recovery [44] was proposed, which directly simulates the emergence and diffusion of air pollution instead of modeling the mapping between observed data and unobserved samples. Xu et al. [45] proposed a model coupling Gaussian diffusion and GRU. Based on the diffusion law of pollutants, a multivariate Gaussian diffusion model was developed. Chen et al. [46] proposed a physically guided adaptive model to cope with the sparse coverage of mobile sensors, which adds to the physically guided model when the coverage is sparse and gives more importance to the data-driven model when the sensing coverage is dense. Hong et al. [47] proposed a hybrid model combining the CMAQ model and LSTM to predict PM2.5 concentrations, improving the performance of large-area spatial distribution prediction at a relatively low cost.
However, the above models need a large number of labeled samples in order to be built. In contrast, the limited number of monitoring stations leads to a limited number of labeled samples, significantly reducing the performance of supervised learning methods. To avoid the problems, Lv et al. [48] proposed a multi-view transfer semi-supervised learning method for air quality estimation, which trains the initial models by leveraging the labeled data and refines the models using a semi-supervised regression algorithm by exploiting the unlabeled data from the target location. Zheng et al. [49] proposed a semi-supervised learning method named U-Air based on co-training [50] to infer the grade of fine-grained urban air quality. The enlightenment for us is that co-training semi-supervised learning can improve the prediction performance of the model by using unlabeled samples to enhance labeled data [51].
In this paper, the most important content focuses on predicting the accurate AQI for locations where monitoring stations are not deployed, which is a regression problem. Therefore, we proposed a new co-training semi-supervised learning method named KNN-DNN that trains two different regressors to model the relationship between the AQI and meteorological and pollutants factors, and constantly complements and improves each regressor. It can effectively enhance the real-time prediction of the AQI at the locations without monitoring stations, provide fine-grained distribution of the AQI and provide sufficient information for air quality prediction and pollutant prevention.
The contributions of this study are as follows:
1.
A co-training semi-supervised learning method combining KNN and DNN is proposed, which can use unlabeled data to improve the prediction accuracy. It avoids the overfitting of supervised learning methods caused by too few labeled samples in the fine-grained analysis of air quality;
2.
The method proposed in this paper combines two different types of learners, which can overcome the limitation of the co-training method combining two similar types of learners, and further improve the fine-grained analysis effect of air quality;
3.
The method proposed in this paper is evaluated on a real dataset to verify that the method is effective and superior to other models.

2. Materials and Methods

2.1. Study Area

Beijing–Tianjin–Hebei region is known as the gathering place of rapid development. Both sides of the region are surrounded by mountains, so air pollutants are difficult to drive away, and this region has experienced severe pollution events. The study area of this paper is between 118°–118°53′ east longitude and 39°45′–39°89′ north latitude, with a total area of 13,472 km2. Here, heavy industry has developed, energy consumption is significant and air pollution is also severe.
Figure 1 shows all the air quality monitoring micro stations in the region, a total of 498, of which 6 are state-controlled stations and the rest are provincial-controlled micro-stations. It can be seen the number is limited, and the distribution is highly concentrated. This paper collected the hourly records of these monitoring stations from 1 February to 30 April 2018, with a total of 2136 moments.

2.2. Input Data

The data include the temperature, humidity, pollutant concentrations, source type and AQI of each monitoring station. Among them, AQI is used to describe the degree of air pollution quantitatively. The higher the AQI, indicating that the more serious the air pollution, the greater the harm. AQI is used as a label in the experiment, and the rest are related to the attributes that affect its changing trend. The impact of these attributes on AQI are discussed below.
1.
The changes in AQI are affected by meteorological factors.
Figure 2 shows the correlation between AQI and the two meteorological attributes of temperature and humidity. The horizontal axis represents the humidity, and the vertical axis represents the temperature. Each point in the graph reflects the AQI under specific temperature and humidity conditions. It can be seen that high humidity usually leads to poor air quality, and the effect of temperature on AQI is not apparent, but when the temperature is low and the humidity is high, the air quality is poor.
2.
Air pollutants’ concentrations are the most significant cause of AQI change.
In the correlation matrix shown in Figure 3, subgraphs per row/column represent the same specific air pollutant, including PM2.5, PM10, SO2, NO2, CO and O3. The pollutant concentration unit is ug/m3, omitted in the graphs. Each subgraph represents the AQI under different pollutant concentrations. It can be seen that no matter which pollutant concentration increases, the air quality becomes worse, and the influence of PM2.5 and PM10 on AQI is especially obvious.
3.
AQI is directly related to the geographical environment.
On environments such as building sites, sand and gravel and other building materials easily produce dust, leading to poor air quality. We call the geographical environment of the air quality monitoring station the source type. Figure 4 shows the changing trend of the 24 h average AQI of different types of monitoring stations within the research time range of the study area. It can be seen that compared with other source types of monitoring stations, the AQI, which belongs to the locations of main pollution source and key enterprise, maintains a high level as a whole. This attribute is classified, so the single hot code method encodes it.

2.3. Method and Principles

2.3.1. DNN

DNN is a network model suitable for dealing with complex nonlinear relationships and has excellent regression performance. As shown in Figure 5, the DNN model is a hierarchical architecture. Each arrow in the network has its linear coefficient w and bias b. In fact, the local model, such as the perceptron, is composed of a linear relationship and an activation function σ(·). To represent the operation with matrices, the linear coefficient from the k-th neuron of layer l to the j-th neuron of layer l + 1 is defined as w j k l , and the bias of j-th neuron of layer l is defined as b j l , the activation value is expressed as a j l . All the linear coefficients of the l layer constitute the matrix Wl and all the biases constitute the vector bl, then the output of layer l is:
a l = σ ( z l ) = σ ( W l a l 1 + b l )
By analogy, we can obtain the final outputs. This process is called forward propagation of DNN, and the black solid line arrow in Figure 5 is the direction.
To obtain a DNN model with high accuracy through training, an appropriate loss function is needed to measure the output loss. By optimizing this function to minimize the loss, a series of corresponding linear coefficient matrices and bias vectors can be obtained to construct an excellent DNN. This process is called the backpropagation of DNN, and the dotted arrow in Figure 5 shows the direction.

2.3.2. KNN

The core idea of the KNN regression algorithm is: the label value of a sample to be predicted is the arithmetic mean of the label values of K samples with the smallest distance from this sample in the feature space. The algorithm flow is shown in Figure 6.
In the process of determining regression decision, KNN only determines the label value of the sample to be predicted according to the label values of the nearest sample or samples. The measure of proximity degree is the distance between the sample to be predicted and each sample, usually using Euclidean distance. The calculation formula is as follows:
d = ( x 1 x 2 ) 2 + ( y 1 y 2 ) 2
where (x1, y1) and (x2, y2) are the coordinates of two samples in 2D space.

2.3.3. Co-Training

Semi-supervised learning can integrate a small number of labeled samples and a large number of unlabeled samples to improve the learning performance of learners. The labeled sample refers to the sample with both attributes and label, while the unlabeled sample refers to the data with only attributes but no label. Co-training is one of the most well-known semi-supervised learning methods. It was first proposed by Blum et al. [50], and trains two classifiers on two different views, using the prediction of unlabeled examples of each classifier to enhance the training set of the other classifier. In other words, the standard co-training algorithm requires that the attributes are naturally divided into two views. Each view is sufficient for training and is conditionally independent of another view [52].
In practical applications, the requirement of multi-view is challenging to meet, so co-training is improved. The COREG algorithm proposed by Zhou and Li [53] is a single-view co-training method that uses two KNNs as regressors to distinguish between different K and p values. Liang et al. [54] implemented a single-view co-training algorithm using two ANN with different structures. Therefore, the single-view co-training method can make up for the problem that the data provide a challenge to meeting the requirements of multi-views by meeting the diversity of the learners.
The specific steps of single-view co-training are as follows:
  • Train two different learners using training sets with labels to build a diverse learning framework that learns the relationship between sample labels and related attributes.
  • Label the unlabeled samples and calculate the confidence of each sample.
  • Select the unlabeled samples with the highest confidence for each learner, together with the predicted results, add them to the labeled sample set to achieve the purpose of expanding the labeled sample set.
  • Train each learner with the expanded labeled sample set.
  • Repeat steps 2, 3 and 4 until the preset termination goal is reached or the unlabeled sample set is empty, and two learners with excellent performance are obtained.
The validation set or sample to be predicted is input into two learners, and the average value of the prediction result are taken as the final result.

2.4. Co-Training Semi-Supervised Learning Method Combining KNN and DNN

2.4.1. Construction of a Framework

In the single-view co-training algorithm, if learners with the same type are combined, even if the initial parameters are different, the difference between the learners will continue to shrink with continuous training. Therefore, the prediction effect of each learning is difficult to improve significantly.
For this reason, this paper adopts a strategy of combining different types of learners and conducting co-training from a single view. A new co-training regression algorithm, a co-training semi-supervised learning method combining KNN and DNN named KNN-DNN, is proposed. This method can thoroughly incorporate the advantages of KNN and DNN in semi-supervised learning and overcome the ability limitation of the co-training semi-supervised learning method, which combines two learners with the same type.
Therefore, for the fine-grained analysis of air quality, this improved method can use unlabeled data to expand the training set. It can effectively avoid the overfitting problem caused by the limited number of labeled samples due to the limited number of air quality monitoring stations, thereby improving the model effect. The overall framework is shown in Figure 7. The original data are divided into three parts. The first part is the labeled sample set for the initialization of the learners in KNN-DNN; the second part is the unlabeled sample set, examples in it are used as candidates, which are labeled and added to the labeled sample set during the iterative process; the third part is the validation set: when the model is trained, it is input into the trained model to obtain the AQI of the target locations.

2.4.2. KNN-DNN

The implementation process of the KNN-DNN algorithm is as follows:
1.
The input dataset is divided into a labeled sample set L = { ( x 1 , y 1 ) , ( x 2 , y 2 ) , , ( x L , y L ) } and an unlabeled sample set U = { x 1 , x 2 , , x U } . The attributes and labels in L are used to train the initial learners KNN and DNN.
2.
Each learner receives a batch of randomly selected samples u = { x 1 , x 2 , , x u | x i U } from U and predicts them. The confidence of each sample is calculated, and the sample with the highest confidence is taken as a pseudo-label sample. The predicted value yi’, together with the attributes xi’, are added to the training set of another learner.
3.
Through continuous iterative training, a large number of unlabeled samples are labeled and added to the labeled sample set, so they are fully applied.
It is worth noting that in specific training, in order to select the appropriate pseudo-label samples to join the training set of another learner, it is necessary to estimate the confidence of each sample in u. For example, in the classification problem, it can be estimated by calculating the probability that unlabeled samples are labeled to different categories. However, the air quality fine-grained analysis in this paper used to estimate AQI value is a regression problem, and the predicted value is infinite. No estimated probability can be used as with the one in classification.
Therefore, a critical problem in the process of constructing the KNN-DNN algorithm is how to scientifically calculate the confidence of the pseudo-label samples and select the sample with the highest confidence to enhance the performance of each learner. On regression problems such as fine-grained analysis of air quality, the samples that improve the performance of the learner most significantly should be the samples that reduce the error of the learner the most. Thus, calculate the error of the learners h1 and h2 trained by the original labeled sample first. Then calculate the error of the learners h1′ and h2′ trained by the original labeled set plus each pseudo-label sample. Finally, calculate the change in the error, the sample with the most error reduction is the most confident pseudo-label sample. For each learner, the confidence of each pseudo-labeled sample can be designed as follows:
Δ = x i u ( ( y i h k ( x i ) ) 2 ( y i h k ( x i ) ) 2 )
where k = 1, 2.
Algorithm 1 shows the KNN-DNN algorithm steps.
Algorithm 1. KNN-DNN.
Input: labeled sample set L and unlabeled sampled set U
Process:
h1 = KNN(L1), h2 = DNN(L2)
for k (number of iterations)
  for each x’ in u
y1 = h1 (x’), y2 = h2(x’)
h1’ = KNN(L1∪(x’, y1)), h2’ = DNN(L2∪(x’, y2))
    Calculate Δx1, Δx2 according to Formula (3)
end for
  if there exist Δx1 > 0
  then select x’, which makes Δx1 maximum, as x 1 , y 1 = h1( x 1 ), u1 = ( x 1 , y 1 ), u = uu1
  else u1 = ∅
  if there exist Δx1 > 0
  then select x’, which makes Δx2 maximum, as x 2 , y 2 = h2( x 2 ), u2 = ( x 2 , y 2 ), u = uu2
  else u2 = ∅
  L1 = L1 + u2, L2 = L2 + u1
  if u not changed
  then exit
  else h1 = KNN(L1), h2 = DNN(L2)
       Reselect u in U
end for
output: 1/2(h1(x) + h2(x))

3. Results

3.1. Evaluation Indicators

This paper predicts the AQI of the target location without monitoring equipment, which belongs to the regression problem. The root mean square error (RMSE), mean absolute error (MAE) and determination coefficient (R2) are used to evaluate the fine-grained prediction effect of each model. The smaller the RMSE and MAE are, the closer R2 is to 1, and the better the effect of the model is.

3.2. Baseline Models and Parameter Settings

In the course of the experiment, the supervised learning methods and co-training semi-supervised learning methods were compared, respectively, which mainly included:
  • Kriging: The AQI of the target location is estimated by the weighted values of all existing monitoring stations. The coefficient is not the reciprocal of the distance but a set of optimal coefficients that can satisfy the minimum difference between the estimated value and the real value at the target location;
  • Liner Regression (LR): Use the curve to fit the relationship between the AQI of the existing monitoring stations and the relevant attributes. When the attributes of the target position are inputted, the AQI can be obtained according to the curve;
  • Support Vector Regression (SVR): Use the hyperplane to fit the relationship between the AQI of the existing monitoring stations and the relevant attributes;
  • XGBoost: A deep learning algorithm based on gradient boosting decision tree, which uses the second-order Taylor expansion to calculate the loss to overcome overfitting;
  • CNN: A classical deep learning algorithm, which is based on local connection and parameter sharing, that automatically extracts features from the data;
  • GCN: Construct the collected data into graphs according to the distribution of the monitoring stations, and then perform convolution operations;
  • KNN: A single learner for supervised learning;
  • DNN: A single learner for supervised learning;
  • KNN-KNN: A co-training semi-supervised learning method combining the same type of learners, KNN1 and KNN2.
Because the real AQI data cannot be collected at the location without monitoring equipment, it can be used as an unlabeled sample in practical application. However, to verify the effectiveness of the models, in the experiment, for the co-training semi-supervised learning methods, the data of 100 monitoring stations among 498 monitoring stations were randomly selected as the training set, and 20 were randomly selected as the testing set. Data from other monitoring stations were stripped of labels and used as unlabeled samples.
The number of iterative rounds was set to 100. That is, in the training process, the data of up to 100 monitoring stations were selected for pseudo-labeling. After training, the unselected samples were used as the validation set to verify the effect of the trained model. To make a fair comparison, for the supervised learning model in the baseline models, the settings of the training set, testing set and validation set were the same as the semi-supervised learning methods.
Among the models, KNN has two key parameters, the nearest neighbor number K and the distance metric p of the vector space. In this paper, the cross-validation method was used to optimize these two values on the training set. Figure 8 shows that the fitting effect of KNN is the best when K = 3 and p = 2. That is, the Euclidean distance is used as the metric, and the number of neighbors is 3. The KNN learner in KNN-DNN also uses this parameter combination.

3.3. Analysis of Experimental Results

The proposed method KNN-DNN and the KNN-KNN were verified on the dataset. The relevant indicators of the prediction results of the two models and each learner were compared, as shown in Table 1.
As seen from Table 1, compared with the prediction results of its own two KNN learners, the KNN-KNN model decreased by 26% and 6% in RMSE, 16% and 8% in MAE and increased by 0.03 and 0 on R2. Compared with the prediction results of its own KNN learner and the DNN learner, the DNN-KNN model decreased by 26% and 6% in RMSE, 16% and 8% in MAE index, and increased by 0.03 and 0 on R2. It can be seen that in the process of co-training, by labeling the unlabeled samples and enriching the training set of each learner, the performance of each learner can be effectively improved. Therefore, a smaller error and higher fitting accuracy can be obtained.
In addition, the prediction result of the KNN-DNN model has a better performance on all three indicators than that of the KNN-KNN model, which decreases by 52% and 43% on RMSE and MAE, respectively, and increases by 0.15 on R2. It can be seen that the co-training semi-supervised learning method combining two different types of learners can better fuse the respective advantages of both learners and break through the limitation of combining single-type learners.
The KNN-DNN model and all the baseline models mentioned in Section 3.2 were verified on the dataset. The relevant indicators of the prediction results of each model are shown in Table 2.
As can be seen from Table 2, compared with the prediction results of the Kriging, LR, SVR, XGBoost, CNN, GCN, KNN and DNN models, the prediction results of KNN-DNN decreased by 90%, 80%, 74%, 78%, 77%, 69%, 73% and 79%, respectively, on RMSE, 90%, 80%, 72%, 78, 73%, 68%, 73% and 80%, respectively, on MAE, and increased 0.85, 0.07, 0.23, 0.08, 0.08, 0.04, 0.10 and 0.07, respectively, on R2. The prediction effect of the KNN-DNN model is better than that of each baseline model.
In summary, the co-training semi-supervised learning method continuously strengthens the view used to train the learners by using unlabeled samples, so that KNN-DNN achieves more competitive air quality fine-grained analysis performance than the existing methods when the number of samples is limited.
The trained KNN-DNN model was applied to the verification set, and the predicted value was compared with the true value, as shown in Figure 9. In theory, the predicted value should be equal to the real value; that is, the trend line is y = x, as shown by the solid line in Figure 9. The trend line obtained by the model is shown by the dotted line in the diagram, with an angle of about 4.4 inches, which is very close to the solid line.

4. Discussion

In this study, we adopt a single-view strategy and propose a co-training semi-supervised learning method that combines multi-type learners to achieve fine-grained analysis of air quality. It provides air quality information in areas that cannot be covered by current air quality monitoring systems. Through the evaluation, the proposed model performs well in all aspects and has better performance than other methods. Compared with supervised learning models such as KNN, DNN, CNN, GCN, XGBoost, etc., our advantage is to make full use of unlabeled samples to improve model performance by expanding the training set. When the number of air quality monitoring stations is limited, the superiority of our model is especially obvious. Compared with U-Air, our advantage is that we do not need to divide a dataset into two fully redundant views, reducing the requirements for the dataset. Compared with KNN-KNN, our advantage is that we increase the difference between the two learners and avoid the ability limitation of using the same type of learner in the co-training algorithm.
There are still some issues to be solved in this study. For example, the spread of air pollution has spatial and temporal correlation, so the distribution of the AQI also has spatial and temporal correlation. The proposed model uses data collected by the air quality monitoring stations to predict the AQI of locations without the deployment of the monitoring station at the current time. It is in real time, but it cannot predict the AQI in the future. In addition, the data we collected are not deep enough, which may lead to the lack of persuasion and generalization of the model.
In the future, we will collect more abundant and comprehensive data, such as topographical, human flow and traffic-related factors and so on so as to achieve a comprehensive understanding of the atmospheric environment and obtain a more rigorous and accurate air quality fine-grained analysis model. At the same time, we will improve the model structure and establish a model that can consider both the temporal and spatial propagation trends of the AQI. It will provide reasonable and effective prediction for the current and future air quality information in the area not covered by the monitoring system. It also facilitates residents to arrange outdoor activities reasonably, and policy makers to take pollution prevention measures in advance, so as to serve the sustainable development of the city.

5. Conclusions

In this paper, a co-training semi-supervised learning method combining KNN and DNN is proposed for fine-grained analysis of air quality. As two different types of learners, KNNs and DNN adopt a single-view co-training strategy. In each iteration, each learner selects a sample with the highest confidence for another learner among the unlabeled samples and constantly expands the training set to achieve the purpose of using unlabeled data to improve prediction performance.
By comparing the results with the co-training semi-supervised learning method of combining learners with the same type, the KNN-DNN proposed in this paper obtains a higher prediction accuracy and smaller error. The superiority and effectiveness of the combination of different types of learners in co-training is proven. More importantly, by comparing the prediction results of the interpolation method and supervised learning method with a single learner, KNN-DNN has the highest regression accuracy and the minimum error in the fine-grained analysis of air quality. It is proved that the co-training semi-supervised learning method effectively overcomes the problem of overfitting due to the limited number of air quality monitoring stations. It has a good application prospect in the fine-grained analysis of air quality.

Author Contributions

Conceptualization, Y.Z. and L.W.; methodology, Y.Z.; software, Y.Z. and L.W. validation, Y.Z., N.Z. and L.Y.; formal analysis, N.Z. and X.H.; investigation, N.Z. and X.H.; resources, W.Y.; data curation, Y.Z.; writing—original draft preparation, Y.Z.; writing—review and editing, L.W.; visualization, Y.Z.; supervision, L.Y. and W.Y.; project administration, L.W.; funding acquisition, L.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number 42075129 and the Key Research and Development Program of Hebei, grant number 21351803D.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Balsa-Barreiro, J.; Li, Y.; Morales, A.J. Globalization and the shifting centers of gravity of world’s human dynamics: Implications for sustainability. J. Clean. Prod. 2019, 239, 117923. [Google Scholar] [CrossRef]
  2. Hong, C.; Zhang, Q.; Zhang, Y.; Davis, S.J.; Tong, D.; Zheng, Y.; Liu, Z.; Guan, D.; He, K.; Schellnhuber, H.J. Impacts of climate change on future air quality and human health in China. Proc. Natl. Acad. Sci. USA 2019, 116, 17193–17200. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Apte, J.S.; Brauer, M.; Cohen, A.J.; Ezzati, M.; Pope, C.A. Ambient PM2.5 Reduces Global and Regional Life Expectancy. Environ. Sci. Technol. Lett. 2018, 5, 546–551. [Google Scholar] [CrossRef] [Green Version]
  4. Zou, Y.; Wang, Y.; Zhang, Y.; Koo, J.-H. Arctic sea ice, Eurasia snow, and extreme winter haze in China. Sci. Adv. 2017, 3, e1602751. [Google Scholar] [CrossRef] [Green Version]
  5. Han, Z.; Zhou, B.; Xu, Y.; Wu, J.; Shi, Y. Projected changes in haze pollution potential in China: An ensemble of regional climate model simulations. Atmos. Chem. Phys. 2017, 17, 10109–10123. [Google Scholar] [CrossRef] [Green Version]
  6. Jiang, P.; Li, C.; Li, R.; Yang, H. An innovative hybrid air pollution early-warning system based on pollutants forecasting and Extenics evaluation. Knowl.-Based Syst. 2019, 164, 174–192. [Google Scholar] [CrossRef]
  7. Hermosilla, T.; Palomar-Vázquez, J.; Balaguer-Beser, Á.; Balsa-Barreiro, J.; Ruiz, L.A. Using street based metrics to characterize urban typologies. Comput. Environ. Urban Syst. 2014, 44, 68–79. [Google Scholar] [CrossRef] [Green Version]
  8. Hermosilla, T.; Ruiz, L.A.; Recio, J.A.; Balsa-Barreiro, J. Land-use mapping of Valencia city area from aerial images and LiDAR data. In Proceedings of the GEOProcessing 2012: The Fourth International Conference in Advanced Geographic Information Systems, Applications and Services, Valencia, Spain, 30 January–4 February 2012; pp. 232–237. [Google Scholar]
  9. Voordeckers, D.; Lauriks, T.; Denys, S.; Billen, P.; Tytgat, T.; Van Acker, M. Guidelines for passive control of traffic-related air pollution in street canyons: An overview for urban planning. Landsc. Urban Plan. 2021, 207, 103980. [Google Scholar] [CrossRef]
  10. Yang, Y.; Zheng, Z.; Bian, K.; Song, L.; Han, Z. Real-Time Profiling of Fine-Grained Air Quality Index Distribution Using UAV Sensing. IEEE Internet Things J. 2018, 5, 186–198. [Google Scholar] [CrossRef]
  11. Venegas, L.E.; Mazzeo, N.A.; Dezzutti, M.C. A simple model for calculating air pollution within street canyons. Atmos. Environ. 2014, 87, 77–86. [Google Scholar] [CrossRef]
  12. Kwak, K.-H.; Baik, J.-J.; Ryu, Y.-H.; Lee, S.-H. Urban air quality simulation in a high-rise building area using a CFD model coupled with mesoscale meteorological and chemistry-transport models. Atmos. Environ. 2015, 100, 167–177. [Google Scholar] [CrossRef]
  13. Hong, D.; Yan, Z.; Jie, X.; Zheng, L.; Cun, J.; Wei, F. Numerical simulation of pollutant propagation characteristics in a three-dimensional urban traffic system (in Chinese). China Environ. Sci. 2018, 38, 51–58. [Google Scholar] [CrossRef]
  14. Ghenai, C.; Lin, C. Dispersion modeling of PM10 released during decontamination activities. J. Hazard. Mater. 2006, 132, 58–67. [Google Scholar] [CrossRef] [PubMed]
  15. Rangel, M.G.L.; Henríquez, J.R.; Costa, J.A.P.; de Lira Junior, J.C. An assessment of dispersing pollutants from the pre-harvest burning of sugarcane in rural areas in the northeast of Brazil. Atmos. Environ. 2018, 178, 265–281. [Google Scholar] [CrossRef]
  16. Yang, Z.; Yao, Q.; Buser, M.D.; Alfieri, J.G.; Li, H.; Torrents, A.; McConnell, L.L.; Downey, P.M.; Hapeman, C.J. Modification and validation of the Gaussian plume model (GPM) to predict ammonia and particulate matter dispersion. Atmos. Pollut. Res. 2020, 11, 1063–1072. [Google Scholar] [CrossRef]
  17. Karim, A.A.; Nolan, P.F. Modelling reacting localized air pollution using Computational Fluid Dynamics (CFD). Atmos. Environ. 2011, 45, 889–895. [Google Scholar] [CrossRef]
  18. Tobon, A.M.; Moncho-Esteve, I.J.; Martinez-Corral, J.; Palau-Salvador, G. Dispersion of CO Using Computational Fluid Dynamics in a Real Urban Canyon in the City Center of Valencia (Spain). Atmosphere 2020, 11, 693. [Google Scholar] [CrossRef]
  19. Nie, W.; Liu, X.; Liu, C.; Guo, L.; Hua, Y. Prediction of dispersion behavior of typical exhaust pollutants from hydraulic support transporters based on numerical simulation. Environ. Sci. Pollut. Res. 2022, 29, 38110–38125. [Google Scholar] [CrossRef]
  20. Qiao, X.; Ying, Q.; Li, X.H.; Zhang, H.L.; Hu, J.L.; Tang, Y.; Chen, X. Source apportionment of PM2.5 for 25 Chinese provincial capitals and municipalities using a source-oriented Community Multiscale Air Quality model. Sci. Total Environ. 2018, 612, 462–471. [Google Scholar] [CrossRef]
  21. Koo, Y.S.; Choi, D.R.; Kwon, H.Y.; Jang, Y.K.; Han, J.S. Improvement of PM10 prediction in East Asia using inverse modeling. Atmos. Environ. 2015, 106, 318–328. [Google Scholar] [CrossRef]
  22. Manders, A.M.M.; Schaap, M.; Hoogerbrugge, R. Testing the capability of the chemistry transport model LOTOS-EUROS to forecast PM10 levels in the Netherlands. Atmos. Environ. 2009, 43, 4050–4059. [Google Scholar] [CrossRef]
  23. Araki, S.; Shima, M.; Yamamoto, K. Spatiotemporal land use random forest model for estimating metropolitan NO2 exposure in Japan. Sci. Total Environ. 2018, 634, 1269–1277. [Google Scholar] [CrossRef] [PubMed]
  24. Qin, X.; Do, T.H.; Hofman, J.; Bonet, E.R.; La Manna, V.P.; Deligiannis, N.; Philips, W. Fine-Grained Urban Air Quality Mapping from Sparse Mobile Air Pollution Measurements and Dense Traffic Density. Remote Sens. 2022, 14, 2613. [Google Scholar] [CrossRef]
  25. Cheng, W.; Shen, Y.; Zhu, Y.; Huang, L. A Neural Attention Model for Urban Air Quality Inference: Learning the Weights of Monitoring Stations. Proc. AAAI Conf. Artif. Intell. 2018, 32, 2151–2158. [Google Scholar] [CrossRef]
  26. Zhong, H.; Yin, C.; Wu, X.; Luo, J.; He, J. AirRL: A Reinforcement Learning Approach to Urban Air Quality Inference. arXiv 2020, arXiv:12205. [Google Scholar] [CrossRef]
  27. Xu, Y.; Zhu, Y.; Shen, Y.; Yu, J. Fine-Grained Air Quality Inference with Remote Sensing Data and Ubiquitous Urban Data. ACM Trans. Knowl. Discov. Data 2019, 13, 1–47. [Google Scholar] [CrossRef]
  28. Dai, H.; Huang, G.; Zeng, H.; Zhou, F. PM2.5 volatility prediction by XGBoost-MLP based on GARCH models. J. Clean. Prod. 2022, 356, 131898. [Google Scholar] [CrossRef]
  29. Ke, G.L.; Meng, Q.; Finley, T.; Wang, T.F.; Chen, W.; Ma, W.D.; Ye, Q.W.; Liu, T.Y. LightGBM: A Highly Efficient Gradient Boosting Decision Tree. In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; pp. 3147–3155. [Google Scholar]
  30. Dai, H.; Huang, G.; Zeng, H.; Yu, R. Haze Risk Assessment Based on Improved PCA-MEE and ISPO-LightGBM Model. Systems 2022, 10, 263. [Google Scholar] [CrossRef]
  31. Liu, B.C.; Binaykia, A.; Chang, P.C.; Tiwari, M.K.; Tsao, C.C. Urban air quality forecasting based on multidimensional collaborative Support Vector Regression (SVR): A case study of BeijingTianjin-Shijiazhuang. PLoS ONE 2017, 12, e0179763. [Google Scholar] [CrossRef] [Green Version]
  32. Hu, Z.; Bai, Z.; Yang, Y.; Zheng, Z.; Bian, K.; Song, L. UAV Aided Aerial-Ground IoT for Air Quality Sensing in Smart City: Architecture, Technologies, and Implementation. IEEE Netw. 2019, 33, 14–22. [Google Scholar] [CrossRef]
  33. Park, S.; Kim, M.; Kim, M.; Namgung, H.G.; Kim, K.T.; Cho, K.H.; Kwon, S.B. Predicting PM10 concentration in Seoul metropolitan subway stations using artificial neural network (ANN). J. Hazard. Mater. 2018, 341, 75–82. [Google Scholar] [CrossRef] [PubMed]
  34. Yi, P.; Wei, Y. Attention based PM2.5 multi-order spatio-temporal graph convolutional network inference model (in chinese). Appl. Res. Comput. 2022, 39, 1–6. [Google Scholar] [CrossRef]
  35. Liu, Y.; Nie, J.; Li, X.; Ahmed, S.H.; Lim, W.Y.B.; Miao, C. Federated learning in the sky: Aerial-ground air quality sensing framework with UAV swarms. IEEE Internet Things J. 2021, 8, 9827–9837. [Google Scholar] [CrossRef]
  36. Han, Q.; Lu, D.; Chen, R. Fine-Grained Air Quality Inference via Multi-Channel Attention Model. In Proceedings of the International Joint Conference on Artificial Intelligence, Montreal, QC, Canada, 19–26 August 2021; pp. 2512–2518. [Google Scholar] [CrossRef]
  37. Chen, L.; Ding, Y.; Lyu, D.; Liu, X.; Long, H. Deep Multi-Task Learning Based Urban Air Quality Index Modelling. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2019, 3, 1–17. [Google Scholar] [CrossRef]
  38. Zhu, J.Y.; Sun, C.; Li, V.O.K. An Extended Spatio-Temporal Granger Causality Model for Air Quality Estimation with Heterogeneous Urban Big Data. IEEE Trans. Big Data 2017, 3, 307–319. [Google Scholar] [CrossRef]
  39. Song, J.; Stettler, M.E.J. A novel multi-pollutant space-time learning network for air pollution inference. Sci. Total Environ. 2022, 811, 152254. [Google Scholar] [CrossRef]
  40. Hsieh, H.-P.; Wu, S.; Ko, C.-C.; Shei, C.; Yao, Z.-T.; Chen, Y.-W. Forecasting Fine-Grained Air Quality for Locations without Monitoring Stations Based on a Hybrid Predictor with Spatial-Temporal Attention Based Network. Appl. Sci. 2022, 12, 4268. [Google Scholar] [CrossRef]
  41. Chen, L.; Wang, J.; Wang, H.; Jin, T. Urban Air Quality Assessment by Fusing Spatial and Temporal Data from Multiple Study Sources Using Refined Estimation Methods. ISPRS Int. J. Geo-Inf. 2022, 11, 330. [Google Scholar] [CrossRef]
  42. Ma, R.; Xu, X.; Wang, Y.; Noh, H.Y.; Zhang, P.; Zhang, L. Guiding the Data Learning Process with Physical Model in Air Pollution Inference. In Proceedings of the 2018 IEEE International Conference on Big Data (Big Data), Seattle, WA, USA, 10–13 December 2018; pp. 4475–4483. [Google Scholar] [CrossRef]
  43. Ma, R.; Liu, N.; Xu, X.; Wang, Y.; Noh, H.Y.; Zhang, P.; Zhang, L. Fine-Grained Air Pollution Inference with Mobile Sensing Systems: A Weather-Related Deep Autoencoder Model. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2020, 4, 1–21. [Google Scholar] [CrossRef]
  44. Ma, R.; Liu, N.; Xu, X.; Wang, Y.; Noh, H.Y.; Zhang, P.; Zhang, L. A deep autoencoder model for pollution map recovery with mobile sensing networks. In Proceedings of the Adjunct Proceedings of the 2019 ACM International Joint Conference on Pervasive and Ubiquitous Computing and Proceedings of the 2019 ACM International Symposium on Wearable Computers, London, UK, 9–13 September 2019; pp. 577–583. [Google Scholar] [CrossRef]
  45. Xu, R.; Deng, X.; Wan, H.; Cai, Y.; Pan, X. A deep learning method to repair atmospheric environmental quality data based on Gaussian diffusion. J. Clean. Prod. 2021, 308, 127446. [Google Scholar] [CrossRef]
  46. Chen, X.; Xu, X.; Liu, X.; Pan, S.; He, J.; Noh, H.Y.; Zhang, L.; Zhang, P. PGA: Physics Guided and Adaptive Approach for Mobile Fine-Grained Air Pollution Estimation. In Proceedings of the 2018 ACM International Joint Conference and 2018 International Symposium on Pervasive and Ubiquitous Computing and Wearable Computers, Singapore, 8–12 October 2018; pp. 1321–1330. [Google Scholar] [CrossRef]
  47. Hong, H.; Choi, I.; Jeon, H.; Kim, Y.; Lee, J.-B.; Park, C.H.; Kim, H.S.J.A. An Air Pollutants Prediction Method Integrating Numerical Models and Artificial Intelligence Models Targeting the Area around Busan Port in Korea. Atmosphere 2022, 13, 1462. [Google Scholar] [CrossRef]
  48. Lv, M.Q.; Li, Y.F.; Chen, L.; Chen, T.M. Air quality estimation by exploiting terrain features and multi-view transfer semi-supervised regression. Inf. Sci. 2019, 483, 82–95. [Google Scholar] [CrossRef]
  49. Zheng, Y.; Liu, F.; Hsieh, H.-P. U-Air: When urban air quality inference meets big data. In Proceedings of the 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Chicago, IL, USA, 11–14 August 2013; pp. 1436–1444. [Google Scholar] [CrossRef]
  50. Blum, A.; Mitchell, T. Combining labeled and unlabeled data with co-training. In Proceedings of the Eleventh Annual Conference on Computational Learning Theory, Madison, WI, USA, 24–26 July 1998; pp. 92–100. [Google Scholar]
  51. Kostopoulos, G.; Karlos, S.; Kotsiantis, S.; Ragos, O. Semi-supervised regression: A recent review. J. Intell. Fuzzy Syst. 2018, 35, 1483–1500. [Google Scholar] [CrossRef]
  52. Páez, A.; Fei, L.; Farber, S. Moving Window Approaches for Hedonic Price Estimation: An Empirical Comparison of Modelling Techniques. Urban Stud. 2008, 45, 1565–1581. [Google Scholar] [CrossRef]
  53. Zhou, Z.; Li, M. Semisupervised Regression with Cotraining-Style Algorithms. IEEE Trans. Knowl. Data Eng. 2007, 19, 1479–1493. [Google Scholar] [CrossRef]
  54. Liang, Y.; Liu, Z.; Liu, W. A co-training style semi-supervised artificial neural network modeling and its application in thermal conductivity prediction of polymeric composites filled with BN sheets. Energy AI 2021, 4, 10052. [Google Scholar] [CrossRef]
Figure 1. Distribution of air quality monitoring stations in the study area.
Figure 1. Distribution of air quality monitoring stations in the study area.
Atmosphere 14 00143 g001
Figure 2. Correlation between AQI with temperature and humidity.
Figure 2. Correlation between AQI with temperature and humidity.
Atmosphere 14 00143 g002
Figure 3. Correlation matrix between AQI with concentrations of six pollutants.
Figure 3. Correlation matrix between AQI with concentrations of six pollutants.
Atmosphere 14 00143 g003
Figure 4. Daily variation in AQI at different source type monitoring stations.
Figure 4. Daily variation in AQI at different source type monitoring stations.
Atmosphere 14 00143 g004
Figure 5. Structure of DNN.
Figure 5. Structure of DNN.
Atmosphere 14 00143 g005
Figure 6. KNN algorithm flow.
Figure 6. KNN algorithm flow.
Atmosphere 14 00143 g006
Figure 7. The overall framework of KNN-DNN for fine-grained analysis of air quality.
Figure 7. The overall framework of KNN-DNN for fine-grained analysis of air quality.
Atmosphere 14 00143 g007
Figure 8. Selection of KNN hyperparameters.
Figure 8. Selection of KNN hyperparameters.
Atmosphere 14 00143 g008
Figure 9. Fitting effect of predicted value and true value.
Figure 9. Fitting effect of predicted value and true value.
Atmosphere 14 00143 g009
Table 1. Comparison of each learner in co-training semi-supervised learning models.
Table 1. Comparison of each learner in co-training semi-supervised learning models.
Title 1RMSEMAER2
KNN-KNN3.122.300.84
KNN14.002.960.72
KNN23.252.470.82
KNN-DNN1.511.300.97
KNN2.051.540.94
DNN1.611.410.97
Table 2. Comparison of different models.
Table 2. Comparison of different models.
ModelsRMSEMAER2
Kriging15.5613.230.12
LR7.406.660.90
SVR5.824.670.74
XGBoost6.986.010.89
CNN6.614.840.89
GCN4.874.010.93
KNN5.634.830.87
DNN7.356.620.90
KNN-KNN3.122.300.84
KNN-DNN1.511.300.97
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, Y.; Wang, L.; Zhang, N.; Huang, X.; Yang, L.; Yang, W. Co-Training Semi-Supervised Learning for Fine-Grained Air Quality Analysis. Atmosphere 2023, 14, 143. https://doi.org/10.3390/atmos14010143

AMA Style

Zhao Y, Wang L, Zhang N, Huang X, Yang L, Yang W. Co-Training Semi-Supervised Learning for Fine-Grained Air Quality Analysis. Atmosphere. 2023; 14(1):143. https://doi.org/10.3390/atmos14010143

Chicago/Turabian Style

Zhao, Yaning, Li Wang, Nannan Zhang, Xiangwei Huang, Lunke Yang, and Wenbiao Yang. 2023. "Co-Training Semi-Supervised Learning for Fine-Grained Air Quality Analysis" Atmosphere 14, no. 1: 143. https://doi.org/10.3390/atmos14010143

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop