Next Article in Journal
Seasonal Cambial Activity and Formation of Secondary Phloem and Xylem in White Oaks (Quercus alba L.)
Previous Article in Journal
Temporal and Spatial Analysis of the Changes in the Supply of Ecological Products in State-Owned Forest Farms after the Reform
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Integrating Machine Learning and a Spatial Contextual Algorithm to Detect Wildfire from Himawari-8 Data in Southwest China

School of Resources and Environment, University of Electronic Science and Technology of China, Chengdu 611731, China
*
Author to whom correspondence should be addressed.
Forests 2023, 14(5), 919; https://doi.org/10.3390/f14050919
Submission received: 31 March 2023 / Revised: 23 April 2023 / Accepted: 28 April 2023 / Published: 28 April 2023
(This article belongs to the Section Forest Inventory, Modeling and Remote Sensing)

Abstract

:
Timely wildfire detection is helpful for fire monitoring and fighting. However, the available wildfire products with high temporal resolutions face problems, such as high omission error and commission error (false alarm) rates. This study proposed a wildfire detection algorithm combined with an improved spatial contextual algorithm and machine learning method in southwest China. First, a dataset consisting of a formation of high-confidence fire pixels combining the WLF (Himawari Wild Fire product) and VIIRS wildfire products was constructed. Then, a model to extract potential fire pixels was built using the random forest method. Additionally, an improved spatial contextual algorithm was used to identify actual fire pixels from potential fire pixels. Finally, strategies such as sun glint rejection were used to remove false alarms. As a result, the proposed algorithm performed better, with both a lower omission error rate and a lower commission error rate than the WLF product. It had a higher F1 score (0.47) than WLF (0.43) with VIIRS for reference, which means it is more suitable for wildfire detection.

1. Introduction

With the acceleration of industrialization, the amounts of wildfires worldwide have been increasing since the 1950s [1]. China is also a country that suffers from wildfire. Wildfire, which causes significant damage to the ecological environment and socio-economic development, threatens the safety of people’s lives and property and destroys surface vegetation and topsoil [2,3,4,5]. Therefore, timely and accurate wildfire detection is necessary for protecting the ecological environment and stabilizing the social economy.
Commonly used methods of wildfire detection include manual patrol monitoring [6], video monitoring [7,8,9], UAV (Unmanned Aerial Vehicle) monitoring [10,11], and remote sensing monitoring [12,13]. Manual patrol monitoring requires many human resources. Meanwhile, video and UAV monitoring have problems, such as their small monitoring scope and high cost. These problems make them difficult to apply to large-scale wildfire detection. On the other hand, remote sensing technology has an extensive monitoring range, high monitoring timeliness, and low monitoring cost, making it available for large-scale wildfire detection. Today, high spatial resolution sensors borne by polar-orbiting satellites such as MODIS, AVHRR, and VIIRS are widely used to detect wildfire. However, these satellite data have problems, such as low temporal resolution, making them unable to meet the requirements of near-real-time wildfire detection. Compared to the above data, the geostationary satellite Himawari-8 has a high temporal resolution (full-disk observation for every 10 min), stable data quality, and a large amount of data accumulation. These advantages make it more suitable for near-real-time wildfire detection.
Wildfire detection methods based on remote sensing mainly include the fixed threshold method [14,15], spatial contextual method [16], multi-temporal method [12,17,18] and machine learning method [18,19]. The main principle of the detection methods is to identify anomalies in the surface temperature [20,21].
Wildfire detection based on a fixed threshold value aims to find pixels with abnormal brightness temperatures at infrared bands [22]. Though this method can detect some wildfire pixels, its poor robustness may result in a high omission error rate [23,24]. The multi-temporal method extracts fire pixels by detecting the abnormal changes in the brightness temperature in terms of time dimension. However, the previous data required are easily contaminated by clouds, making it difficult to obtain an accurate estimation of the brightness temperature [12]. The spatial contextual wildfire detection algorithm, which is widely used in wildfire detection of satellite data, has advantages such as fewer omission errors [25,26,27]. This method needs to set a loose threshold to extract potential fire pixels in the image. Then, by comparing the brightness temperature difference between each potential fire spot and its surrounding background pixels, the non-fire spots are removed and accurate fire spot pixels are obtained [17,23,28]. However, this algorithm also has a high commission error rate. Moreover, using a fixed threshold based on experience to screen potential fire spots can easily miss out on smaller, lower-temperature fire pixels. Machine learning methods are also widely used for wildfire detection [29,30,31]. They have an efficient information-mining ability, strong learning ability, and good adaptability [18,32]. The method first analyzes and extracts the hidden features in the dataset for training. Then, a model that best fits all the samples in the training dataset is obtained through iteration [33]. However, the accuracy of machine learning models largely depends on the quality of the dataset.
Taking Yunnan, Sichuan, and Guizhou provinces as the research areas and using Himawair-8 satellite data, a high-precision, near-real-time wildfire detection algorithm combining machine learning and spatial contextual methods was proposed in this study to solve the above problems. This algorithm alleviates the problem of missing potential fire pixels extracted by fixed thresholds in traditional spatial contextual algorithms by adding a random forest model to the screening stage of potential fire pixels. On this basis, the spectral and spatial contextual information of the pixels is integrated to detect the real fire pixels in the potential fire pixels and remove false alarms. False alarm rejection methods, including sun glint rejection, desert boundary rejection, and forest-clearing rejection, are added to further remove false alarms due to the study area’s complex physical and geographical environment. Finally, the accuracy of the proposed algorithm is evaluated. The results show that the proposed algorithm has advantages in terms of the overall detection accuracy (F1 score = 0.47). Furthermore, its omission error rate (0.66) and commission error rate (0.23) are lower than the available near-real-time wildfire detection product WLF. Therefore, the results indicate that this algorithm is more suitable for near-real-time wildfire detection.

2. Materials and Methods

2.1. Study Area

The study area of this study is southwest China, including Sichuan Province, Yunnan Province, and Guizhou Province, which includes the Sichuan Basin, its surrounding mountains, and the high mountains and hills in the Yunnan–Guizhou Plateau. The terrain there is complex and varied, with an average altitude of more than 500 m (as shown in Figure 1b). The study area is a subtropical monsoon climate zone, with overall low precipitation and a relatively dry spring climate, making these areas prone to wildfire. Due to these reasons, southern Sichuan, northern and eastern Yunnan, and western Guizhou are high-risk areas for fire incidents (as shown in Figure 1a). Regarding forest resources, the depth and breadth of vegetation cover in southwest China are relatively high, and the types of forest resources are diverse [34]. The rich forest resources, complex terrain, and relatively dry climate in the study area mean that destructive fires can easily occur in these areas. Therefore, there are high requirements for wildfire monitoring and control.

2.2. Materials

2.2.1. Himawari-8 Image Data

Himawari-8 is a geostationary satellite operated by Japan Meteorological Agency (JMA). It was launched from Tanegashima Space Center using an H-IIA rocket on October 2014 and has been generating data since July 2015. The AHI (Advanced Himawari Imagers) imager borne by Himawari-8, which is greatly improved compared with those of the MTSAT series, contains 16 bands of data, from visible to infrared [35]. Table 1 shows the specifications of the AHI imager, whose spatial resolution varies from 500 m to 2 km. The temporal resolution of AHI full-disk data is 10 min, and its scope includes regions such as East Asia and Australia. Additionally, the image data provided by Himawari-8 are in NetCDF (network Common Data Form) format. The data used in this study include albedo in visible and near-infrared bands (bands 1–6), brightness temperature in infrared bands (bands 7–16), and geometric observation information of the sun and satellite, such as solar zenith angle.

2.2.2. Wildfire Data

This study used WLF and VIIRS wildfire products as the wildfire dataset for model training and testing. See Section 2.3.1 for relevant content on dataset construction.
WLF, a wildfire position and fire radiation power product produced based on the data of the Himawari-8 satellite, is published by Japan Aerospace Exploration Agency (JAXA). It provides data between 60° N to 60° S and 80° E to 160° W, with a spatial resolution of 2 km and a temporal resolution of 10 min. Compared to WLF, the VIIRS wildfire product published by NASA, with a spatial resolution of 375 m, can provide a more accurate location of wildfire pixels and cause fewer omission errors and false alarms due to its higher spatial resolution. However, VIIRS has a low temporal resolution of four times a day. Taking both of them into account, WLF and VIIRS data at the corresponding fire time in the study area are collected to evaluate the detection effect of the algorithm quantitatively. Before calculating accuracy, samples marked as low confidence in the products are eliminated.
Some examples of fires that occurred in the study area from 2020 to 2022 are collected. On the one hand, they serve as time reference for image data collection. On the other hand, they help verify whether the algorithm can accurately detect the occurrence of wildfires in practical applications.

2.3. Methods

This study proposes a near-real-time wildfire detection algorithm based on Himawari-8 satellite data, combining random forest model with spatial contextual algorithm. The process flowchart is shown in Figure 2. For each scene of image to be detected, cloud masking and water masking are first performed to reduce the interference to subsequent process. After that, a random forest model is used to detect potential fire pixels with relatively strict fixed thresholds to reduce omission errors. Then, an improved spatial contextual algorithm is used to filter these potential fire pixels, and various false alarm rejection methods are applied to minimize commission error. Finally, two kinds of methods are used to evaluate detection accuracy.

2.3.1. Cloud Masking and Water Masking

Clouds and water can interfere with detecting wildfire in the mid-infrared band, so conducting cloud masking and water masking before wildfire detection is required. The algorithm for cloud masking varies between day and night. Pixels satisfying condition (1) are marked as night.
s o z > 85 °
where s o z is the solar zenith angle. Cloud has high albedo in the visible and near-infrared channels, while its brightness temperature in the far infrared band is low. A pixel is considered to be cloud-obscured if the following conditions are satisfied:
b t 15 < 265 K
a 3 + a 4 > 1.2
a 3 + a 4 > 0.7   a n d   b t 15 < 285 K
where a 3 and a 4 are the albedo of band 3 and band 4, respectively, b t 15 is the brightness temperature of band 15. During the daytime, pixels that satisfy any of the conditions in 2, 3, and 4 are marked as clouds. In addition, at night, all pixels that satisfy condition 2 will be marked as clouds.
As most of the study area is inland, the majority of the water in the study area is from inland rivers and lakes, which may vary with seasons. Because of the possible differences in water distribution between seasons, using existing water products for water masking is not feasible. The proposed algorithm uses the strong absorption of water at band 6 of Himawari-8 data to extract water.
N D V I = a 4 a 3 a 4 + a 3
a 6 < 0.05   a n d   N D V I < 0
where a 6 is the albedo of band 6, N D V I is the normalized difference vegetation index calculated based on Himawari-8 image data. Pixels that satisfy condition 6 will be marked as water.

2.3.2. Construction of Wildfire Dataset

This study selected remote sensing image data and existing wildfire products at the corresponding times of fire occurrence in the study area from 2017 to 2018 to construct a wildfire dataset. The wildfire data in the dataset comes from WLF and VIIRS wildfire products. When a pixel in a Himawari-8 image is marked as having medium or high confidence of fire by both wildfire products, it is added to the dataset as a fire pixel sample. A total of 576 fire point pixels were selected as positive samples in the dataset. After that, 10 times the number of non-cloud, non-water, and non-fire pixels are randomly chosen as negative samples. Meanwhile, to minimize the impact of erroneous samples on the detection results, all fire point data in WLF and VIIRS are not involved in the construction process of negative samples. After that, all the samples are divided into test sets and training sets in a 1:4 ratio. The extracted features for each sample pixel include albedo information, brightness temperature information, angle information, and spatial contextual information. The detailed information of feature is shown in Table 2, where dev means the mean deviation and w means the value calculated in the window. The determination of window information is detailed in Section 2.3.4.
To make the model detect more fires with lower temperatures, an enhanced dataset is constructed as follows: 314 selected pixels in this dataset are those marked as having medium or high confidence by VIIRS products with relatively high brightness temperatures but not recognized as fires by WLF products. These pixels may be missed due to the low spatial resolution of Himawari-8. This portion will not be divided into test sets and will all be input to the model as the training set. The distribution of all samples in the dataset is shown in Figure 3.

2.3.3. Potential Fire Pixel Detection

Detecting potential fire pixels based on fixed thresholds will miss some smaller and lower temperature fire spots. Therefore, random forest model is required to screen the potential fire spots further. Random forest model was proposed by Breiman in 2001 [36], and has been applied to remote sensing wildfire detection [37].
The essence of random forest is a combination classifier based on decision trees. Each decision tree is composed of a root node and multiple internal leaf nodes [38]. The final output of the model is obtained by voting based on the results of each independent tree. Compared with other machine learning models, random forest model has better robustness to overfitting. Moreover, random forest model can derive the importance of each feature during the training process.
Firstly, the constructed fire point dataset is divided into test set and training set at a ratio of 1:4, and the enhanced dataset and training set are merged. Then, the integrated dataset is input into the model for training. Due to the relatively small number of positive samples in the dataset, attention mechanism is introduced in the training process to give higher weights to fire samples. In the algorithm process, the random forest model will test each non-water pixel that is not cloud-obscured and output its probability of belonging to a fire spot. Pixels classified as fire by random forest model are marked as potential fire pixels.
In addition, to avoid possible omission errors in random forests, this algorithm also uses relatively strict fixed thresholds to screen potential fire spots. Pixels that satisfy conditions 7 or 8 will also be marked as potential fire spots.
b t 7 > 307 K   a n d   b t 7 b t 14 > 7 K   a n d   a 4 < 0.4
b t 7 > 305 K   a n d   b t 7 b t 14 > 7 K   a n d   n i g h t
where b t 7 is the brightness temperature of Himawari-8 at band 7 and b t 14 is the brightness temperature of Himawari-8 at band 14, and n i g h t is judged by condition (1). The output of the fixed threshold algorithm is combined with the potential fire pixels obtained from the random forest model to obtain the final set of potential fire pixels.

2.3.4. Fire Detection Based on Spatial Contextual Algorithm

The aim of spatial contextual algorithm is to identify the real fire pixels in potential fire pixels by selecting appropriate background pixels around the potential fire pixels and determining whether the brightness temperature of the critical bands of the pixel is significantly different from the background ones [24,39]. It mainly includes global contextual method [40] and window contextual method. Global contextual method regards all the valid non-fire pixels in the study area as background pixels, which may lead to an inaccurate estimation of the background features of fire pixels. Window contextual method solves this problem by choosing background pixels from the ambient pixels around the potential fire pixel.
In this study, a window contextual algorithm is used to filter all the extracted potential fire pixels, which can not only reduce the false alarms generated by the random forest algorithm but also extract real fire pixels. Firstly, all pixels that satisfied condition 9 are directly marked as fire.
b t 7 > 345 K   o r   ( b t 7 > 320 K   a n d   n i g h t )
Condition 9 indicates that the fire has reached a certain scale, resulting in a sharp increase in the brightness temperature of pixels compared to normal conditions. Therefore, it is needless to use background pixels for screening. Pixels that do not satisfy condition 9 need to be delimited around a window to select background pixels. At least eight eligible background pixels are required in a window, and at least 25% of the pixels in the window are required to be eligible background pixels. Eligible background pixels refer to pixels with non-cloud, non-water, and non-fire properties, where the background fire pixel is determined by condition 10. The study of Giglio et al. [41] shows that selecting a window within a range of 20 km around the pixel to be detected can accurately represent the background characteristics of it. However, Himawari-8 has a low spatial resolution in the infrared band, which means that only 10 pixels correspond to 20 km. So, the problem of insufficient background pixels in the window is likely to occur. To solve this problem, this study sets the maximum window size to 15 × 15, which is the range of 30 km around the potential fire pixel.
b t 7 > 304 K   a n d   b t 7 b t 14 > 7 K
The initial size of the window is 5 × 5. If the number of background pixels does not meet the requirements, the scope will be expanded, and the window will be retested. When the window size reaches the preset upper limit, the potential fire pixel in the center will be discarded. After determining the window size, the spatial information of background pixels is calculated.
d t > d t a + 3 d t d
d t > d t a + 4.5 K
b t 7 > b t 7 a + 3 b t 7 d
b t 14 > b t 14 a + b t 14 d 4.5 K
f p d 7 > 3 K
where dt is the brightness temperature difference between band 7 and band 14 of Himawari-8 satellite data. Subscripts a and d are the mean and mean deviation of corresponding spectral band in the background pixels within the window, respectively. f p d 7 is the mean deviation of brightness temperature of band 7 for all background fire pixels in the window. Pixels that satisfy conditions 16 or 17 are marked as fire pixels.
11   a n d   12   a n d   13   a n d   ( 14   o r   ( 15 ) )
11   a n d   12   a n d   13   a n d   n i g h t
The purpose of the above criteria is to screen abnormal pixels that have significant differences in brightness temperature from surrounding background pixels. It is particularly important to point out that the aim of condition 15 is to compensate for possible misjudgment caused by condition 14 when a large fire is included in the background pixel.

2.3.5. False Alarm Rejection

This step aims to remove as many non-fire pixels as possible that have spectral similarities to fire pixels. Considering the complex natural geographical environment and climatic conditions of the study area, methods including sun glint rejection, desert boundary rejection, and forest-clearing rejection are taken to remove false alarms and reduce the commission error rate.
1.
Sun glint rejection
In some areas, such as water bodies and humid soil, the light spots reflected by the sun can be captured by sensors, resulting in the saturation of specific channel values. When the infrared band is saturated, the algorithm may recognize it as a brightness temperature anomaly, which results in a false alarm. Such false alarms can be removed effectively by calculating the angle between vectors pointing in the surface-to-satellite and specular reflection directions. The angle θ g is calculated by Equation (18).
c o s θ g = c o s   ( s o z ) c o s   ( s a z ) s i n   ( s o z ) s i n   ( s a z ) c o s φ
where s o z and s a z are solar zenith angle and satellite zenith angle, respectively, and φ is the relative azimuth between satellite and sun. The criteria for sun glint rejection are as follows:
θ g < 2 °
θ g < 10 °   a n d   a 3 > 0.1   a n d   a 4 > 0.2   a n d   a 6 > 0.12
θ g < 15 °   a n d   N W > 0  
where N W is the total number of pixels recognized as water in the window. The addition of this indicator can reduce false alarms caused by pixels with high reflectivity around water. Pixels that satisfy any one of the conditions 19, 20, and 21 are marked as false alarms.
2.
Desert boundary rejection
When filtering fire pixels through spatial contextual algorithm, some ground pixels with higher temperatures may be incorrectly identified as background fire pixels. This may lead to a generally low brightness temperature of the background pixel in the window and result in false alarms in the detection results. By analyzing the number and spectral characteristics of fire pixels in the window, it is found that they have a higher mean and mean deviation in band 7 compared to ordinary, high-temperature background pixels. This feature can be used to remove such false alarms. The criteria are as below, where N b k and N f are the number of background pixels and fire pixels in the window, respectively, and f p a 7 is the mean of band 7 brightness temperature of all fire pixels in the window. Pixels that satisfy any one of the conditions 22, 23, and 24 are marked as false alarms.
N f > 0.1 N b k   a n d   N f > 3
a 4 > 0.18   a n d   f p a 7 < 320 K   a n d   f p d 7 < 2.25 K
b t 7 < f p a 7 + 6 f p d 7  
3.
Forest-clearing rejection
The study area has a high vegetation coverage and relatively dense forests, but there is some vacant space between the trees. Due to significant spectral differences between these pixels and their surrounding ground objects, they may be mistakenly identified as fire pixels sometimes. Compared to remote sensing image data with high spatial resolution, the lower spatial resolution of Himawari-8 somewhat reduces this false alarm, but it still needs to be removed. Pixels satisfying both conditions 25 and 26 are marked as false alarms, where a 4 a is the mean of albedo of band 4 in the window. These conditions mainly aim to identify forest-clearing pixels with significantly higher brightness temperatures than surrounding vegetation pixels.
b t 14 > b t 14 a + 3.7 b t 14 d
a 4 a > 0.28   a n d   b t 7 < 325 K
After removing false alarms from the wildfire pixels obtained by the spatial contextual algorithm in three ways, the resulting fire pixels are the output of the algorithm.

2.4. Accuracy Evaluation Methods and Indicators

This study will evaluate the accuracy of the algorithm from two aspects. On the one hand, the detection effect of the algorithm is evaluated with examples by testing images corresponding to the time of previous fires. On the other hand, the omission error rate and commission error rate of the algorithm are quantitatively calculated with WLF and VIIRS wildfire products for reference to evaluate the detection accuracy of the algorithm.
The omission error rate and commission error rate represent the degree of omission and false alarm of the wildfire pixels detected by the algorithm compared to the reference wildfire product, respectively. The omission error rate is defined by Equation (27).
o m i = 1 r i g h t f p r e s t o t a l f p ( r e f )  
where r i g h t f p r e s is the number of pixels in the reference data that coincides with the pixel position of the algorithm detection result, and t o t a l f p ( r e f ) is the total number of pixels in the reference fire point data. The commission error rate is defined by Equation (28).
c o m = 1 r i g h t f p ( r e f ) t o t a l f p ( r e s )  
where r i g h t f p r e f is the number of pixels in the algorithm detection result that coincides with the reference data position, and t o t a l f p ( r e s ) is the total number of fire point pixels in the algorithm detection result. The omission error rate and commission error rate can intuitively characterize the accuracy of the algorithm through a large number of sample analyses. From the machine learning perspective, the omission error rate and commission error rate are defined by Equations (29) and (30), respectively.
o m i = 1 T P T P + F N  
c o m = 1 T P T P + F P  
where TP, FP, and FN represent the number of true positive samples, false positive samples, and false negative samples, respectively, which can be obtained by constructing a confusion matrix. The omission error rate is 1-Recall, and the commission error rate is 1-Precision. In addition, to comprehensively evaluate the accuracy of the model, F1 score is introduced as an evaluation index. F1 score is defined by Equation (31). It can be regarded as a harmonic average of model precision and recall, and it comprehensively represents the performance of fire detection algorithms in both omissions and false alarms.
F 1 = 2 × ( 1 c o m ) × ( 1 o m i ) 2 c o m o m i  
To enhance the persuasiveness of the verification results, previous fire cases are selected to verify the algorithm. The primary purpose of this process is to test whether the algorithm can achieve forest fire monitoring in practical applications.

3. Results

3.1. Wildfire Dataset and Machine Learning Model Analysis

The feature selection in the wildfire dataset is based on the feature’s important parameters, which are derived from the random forest model and represent the degree to which each feature affects the model. The initial dataset contains more information, but some of its features are of low importance. Therefore, some almost unimportant features are removed based on computational efficiency considerations. After filtering, the importance of each feature in the dataset is shown in Figure 4. The brightness temperature of band 7 and the difference in the brightness temperature between band 7 and band 14 accounts for more than half of the importance of the model, which is the same as the basic principle.
To further evaluate the performance of the random forest model, a comparison to CNN (convolutional neural network) is used. The CNN is simplified from the model proposed by Hong et al. [18]. It consists of two convolutional layers, which have a combined output and input to an ANN consisting of nine layers of decreasing size. This study uses the image of 20 April 2021 at 9 am (UTC) in the test. As there are no VIIRS data at the corresponding time, WLF is used as reference data to evaluate whether the model can extract potential fire pixels effectively. As shown in Table 3, the random forest model extracts 45 potential fire pixels, of which 23 are also detected by WLF. However, CNN only detects 14 pixels in the WLF product, and the total amount of fire pixels detected is also the lowest, which means that it has a worse ability to extract potential fire pixels. CNN requires much more time and better equipment to run. Therefore, the random forest model is used in the proposed algorithm to identify potential fire pixels.

3.2. Algorithm Application Effect Demonstration

Most pixels provided by wildfire products are not caused by fire but by thermal anomalies caused by various reasons. Therefore, to further illustrate the reliability of the algorithm, actual fire cases are selected (Table 4) for verification. The algorithm proposed in this study detected ten fire cases between 2020 and 2021. The results are shown in Figure 5. However, the algorithm also missed some fire cases. All the missed fire cases have a burning area of less than 6 hectares, which is just 1.5 percent of the area of a pixel. Therefore, the abnormal brightness temperature they caused is insufficient to cause an increase in the brightness temperature of the entire pixel, which demonstrates its difficulty in being detected. Moreover, to verify the detection performance of the algorithm, a time dimension verification was conducted for the fire that occurred in Mianning on 20 April 2021 (as shown in Figure 6). The algorithm detected many fire spots when the fire happened at 16:30 (UTC + 8). Over time, the fire gradually became under control and the number of fire spots gradually decreased.

3.3. Algorithm Accuracy Evaluation

By comparing the algorithm results with the WLF data provided by JAXA, the ability of the algorithm to extract fire points is evaluated, and thereby a preliminary evaluation of the accuracy of the algorithm is achieved, as shown in Table 5.
Using the WLF product for reference, the omission error and commission error rates of the proposed algorithm are 0.3 and 0.26, respectively. Table 5 illustrates that the detection result of the algorithm is the same as that of the WLF wildfire product for most pixels. However, due to the limitation of satellite spatial resolution, the wildfire product provided by WLF itself needs to be corrected. Therefore, comparing the two can only roughly prove the reliability of the algorithm, and it is not possible to prove whether there is an improvement in the detection results of the algorithm in the omission error rate or commission error rate. The high spatial resolution VIIRS wildfire product is chosen as a reference to solve this problem. The spatial resolution of VIIRS data is 375 m, which has high accuracy and fire sensitivity. Comparative experiments using VIIRS data as a reference can objectively evaluate the detection effect of the algorithm. The result of the accuracy evaluation is shown in Table 6.
Compared to the VIIRS data, the omission error and commission error rates of the WLF data are relatively high. This is mainly affected by the low spatial resolution of Himawari-8 satellite data. The omission error rate of the random forest model is reduced (0.64), whereas the commission error rate is high. After filtering, the omission error rate of the proposed algorithm (0.66) is not significantly increased compared to the random forest model, but it still decreased compared to WLF products. The commission error rate (0.23) is lower than the random forest model and the WLF product. The F1 score is used as an indicator to evaluate the comprehensive performance of the algorithm. The F1 score of the proposed algorithm (0.47) is higher than WLF, indicating that the overall accuracy of the algorithm is improved compared to the WLF product.
To further illustrate the improvements in the proposed algorithm, a comparison to other algorithms is used. Algorithms B and C are the spatial contextual algorithms proposed by Giglio et al., in 2003 [41] and 2016 [26], respectively, without any further process. As shown in Table 7, algorithms B and C have lower commission error rates in the table, but they missed more fire pixels. This is because these methods are used for a high spatial resolution MODIS imager, and may not be so suitable for the AHI of Himawari-8 satellite without any improvement. Compared with them, the proposed algorithm has a lower omission error rate and higher F1 score, and can achieve fire detection relatively accurately.

4. Discussion

Based on the Himawari-8 image data, this study improves the traditional spatial contextual algorithm by adding a random forest model to detect potential fire pixels. The algorithm performs better than the existing high-temporal resolution wildfire product WLF with a high accuracy of detection, including both the lower omission error rate and lower commission error rate [28]. Unlike these algorithms that detect wildfire with temporal information [12,17], once the image data at the corresponding time is published by JAXA, the wildfire can be detected. This means that it can achieve near-real-time active fire detection with a low publish latency of Himawari-8 data.
Due to the low spatial resolution, wildfire detection based on Himawari-8 satellite usually has a higher omission error rate and commission error rate than those found on high spatial resolution sensors. This is mainly because a larger pixel area means that the fire temperature needs to be higher to cause temperature anomalies at the pixel [42]. A wider range of background pixels may not correctly represents the environment around the fire pixel or there may not be enough background pixels in the best range proposed [41].
This study adopts various methods to reduce omission errors and false alarms for wildfire detection. To reduce omission errors, this study introduces a random forest model. The random forest model was used to detect wildfires with few fixed thresholds [37], which may lead to weak robustness and strong dependence on the dataset. Because filtering wildfire pixels just by a deep learning model needs to take both precision and recall of the model into consideration [30,31], the omission of some wildfire pixels is inevitable [18]. The proposed algorithm uses strategies to improve the recall of the machine learning model as much as possible regardless of precision, with the result that fewer wildfire pixels will be omitted. An enhanced training set is added and an attention mechanism is introduced during the training process. The fire samples in the enhanced dataset are all relatively low-temperature fire points, which can make the model more sensitive to low-temperature fires. The introduction of an attention mechanism can help improve the accuracy of the model by making it pay more attention to fire samples [9].
To reduce false alarms, the results from the random forest model are first filtered through spatial contextual information. Then, three false alarm rejection algorithms are used to eliminate false alarms further. Most spatial contextual algorithms applied to a geostationary satellite use the variance of ambient background pixels to assess whether the potential fire pixel has a significant difference in the brightness temperature compared to the background [23,27,39]. However, the variance may not be enough to describe the dispersion degree of background pixels. Therefore, the mean deviation, which is used in wildfire detection algorithms for polar-orbit satellites [26,41], is introduced to better describe the distribution of the brightness temperature around potential fire pixels. The results show that this introduction makes the algorithm perform better than those that only use the variance. The false alarm rejection algorithms were first proposed to remove false alarms in MODIS Collection 6 active fire product [26,41]. Considering the environment of the study area, this study makes some improvements to it for better performance. Compared to those without false alarm rejection, the results show that these algorithms, especially the desert boundary rejection algorithm, play an important role in the algorithm [40,43].
However, the ability of the random forest model to detect small fires can be improved, and other kinds of machine learning or deep learning models [8,18] may perform better. The false alarm rejection algorithms remove some fire pixels by mistake, which shows that these methods can be improved. Based on the above discussion, there are some areas of research that should be finished in the future. The first one is to exploit the potential of the random forest models to detect fire spots further while improving the robustness of the model by adding more samples to the dataset. At the same time, the performance of deep learning models, such as artificial neural networks and convolutional neural networks, in fire detection based on remote sensing data will be explored. The second one is to optimize the false alarm rejection algorithm. The method can effectively remove false alarms while mistakenly identifying a few real fire points as false alarms. With further improvement, the false alarm rejection algorithm can minimize the false alarm pixels that were erroneously deleted.

5. Conclusions

This study proposes a near-real-time fire detection algorithm that combines a random forest model with a spatial contextual algorithm. The proposed algorithm has a lower omission error rate and commission error rate compared to existing high-temporal wildfire detection products, making it more suitable for near-real-time wildfire detection.
First, we built a high-quality wildfire dataset by integrating the VIIRS product and the WLF product. An additional portion of positive samples that were not detected by WLF but detected by VIIRS were added to enhance the sensitivity of the model to small fires. Then, we used the dataset to train a random forest model, which will be used to screen potential fire spots. Finally, we adopted the improved spatial contextual algorithm to filter potential fire spots, and introduced false alarm rejection methods to remove false alarms. The proposed algorithm has better accuracy than the WLF product. Therefore, it has practical application value in southwest China.

Author Contributions

Conceptualization, C.L. and B.H.; methodology, C.L.; validation, C.L.; formal analysis, C.L.; investigation, C.L.; resources, B.H.; data curation, C.L.; writing—original draft preparation, C.L.; writing—review and editing, C.L., R.C. and B.H.; visualization, C.L., R.C. and B.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research is founded by the National Natural Science Foundation of China (Contract No. U20A2090) and the Sichuan Science and Technology Program (Contract No. 2023YFS0432).

Data Availability Statement

Not applicable.

Acknowledgments

We remain indebted to Jiarun Huang, Qiming Zhang, Juncen Zhou, Xuping Zheng, Hongguo Zhang, Chunquan fan, and Jianpeng Yin for their invaluable comments in the early stages of this study.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Dupuy, J.L.; Fargeon, H.; Martin-Stpaul, N.K.; Pimont, F.; Ruffault, J.; Guijarro, M.; Hernando, C.; Madrigal, J.; Fernandes, P. Climate change impact on future wildfire danger and activity in southern Europe: A review. Ann. For. Sci. 2020, 77, 35. [Google Scholar] [CrossRef]
  2. Rajora, O.P.; Fageria, M.S.; Fitzsimmons, M. Effects of Wild Forest Fires on Genetic Diversity and Population Structure of a Boreal Conifer, White Spruce (Picea glauca (Moench) Voss): Implications for Genetic Resource Management and Adaptive Potential under Climate Change. Forests 2023, 14, 157. [Google Scholar] [CrossRef]
  3. Moya, D.; Fonturbel, T.; Peña, E.; Alfaro-Sanchez, R.; Plaza-Álvarez, P.A.; González-Romero, J.; Lucas-Borja, M.E.; de Las Heras, J. Fire Damage to the Soil Bacterial Structure and Function Depends on Burn Severity: Experimental Burnings at a Lysimetric Facility (MedForECOtron). Forests 2022, 13, 1118. [Google Scholar] [CrossRef]
  4. Papagiannaki, K.; Giannaros, T.M.; Lykoudis, S.; Kotroni, V.; Lagouvardos, K. Weather-related thresholds for wildfire danger in a Mediterranean region: The case of Greece. Agric. For. Meteorol. 2020, 291, 108076. [Google Scholar] [CrossRef]
  5. Quan, X.W.; Xie, Q.; He, B.B.; Luo, K.W.; Liu, X.Z. Integrating remotely sensed fuel variables into wildfire danger assessment for China. Int. J. Wildland Fire 2021, 30, 807–821. [Google Scholar] [CrossRef]
  6. Kułakowski, P.; Calle, E.; Marzo, J.L. Performance study of wireless sensor and actuator networks in forest fire scenarios. Int. J. Commun. Syst. 2013, 26, 515–529. [Google Scholar] [CrossRef]
  7. Li, A.; Zhao, Y.; Zheng, Z. Novel Recursive BiFPN Combining with Swin Transformer for Wildland Fire Smoke Detection. Forests 2022, 13, 2032. [Google Scholar] [CrossRef]
  8. Guo, Y.-Q.; Chen, G.; Wang, Y.-N.; Zha, X.-M.; Xu, Z.-D. Wildfire Identification Based on an Improved Two-Channel Convolutional Neural Network. Forests 2022, 13, 1302. [Google Scholar] [CrossRef]
  9. Huang, J.; He, Z.; Guan, Y.; Zhang, H. Real-Time Forest Fire Detection by Ensemble Lightweight YOLOX-L and Defogging Method. Sensors 2023, 23, 1894. [Google Scholar] [CrossRef]
  10. Zhang, L.; Wang, M.; Fu, Y.; Ding, Y. A Forest Fire Recognition Method Using UAV Images Based on Transfer Learning. Forests 2022, 13, 975. [Google Scholar] [CrossRef]
  11. Yuan, C.; Liu, Z.X.; Zhang, Y.M.; IEEE. UAV-based Forest Fire Detection and Tracking Using Image Processing Techniques. In Proceedings of the International Conference on Unmanned Aircraft Systems (ICUAS), Denver, CO, USA, 9–12 June 2015; pp. 639–643. [Google Scholar]
  12. Roberts, G.; Wooster, M.J. Development of a multi-temporal Kalman filter approach to geostationary active fire detection & fire radiative power (FRP) estimation. Remote Sens. Environ. 2014, 152, 392–412. [Google Scholar]
  13. Liew, S.C.; Lim, A.; Kwoh, L.K. A Stochastic model for active fire detection using the thermal bands of MODIS data. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Anchorage, AK, USA, 20–24 September 2004; pp. 4465–4467. [Google Scholar]
  14. Ding, Y.; Wang, M.; Fu, Y.; Zhang, L.; Wang, X. A Wildfire Detection Algorithm Based on the Dynamic Brightness Temperature Threshold. Forests 2023, 14, 477. [Google Scholar] [CrossRef]
  15. Barducci, A.; Guzzi, D.; Marcoionni, P.; Pippi, I. Comparison of fire temperature retrieved from SWIR and TIR hyperspectral data. Infrared Phys. Technol. 2004, 46, 1–9. [Google Scholar] [CrossRef]
  16. Di Biase, V.; Laneve, G. Geostationary Sensor Based Forest Fire Detection and Monitoring: An Improved Version of the SFIDE Algorithm. Remote Sens. 2018, 10, 741. [Google Scholar] [CrossRef]
  17. Xie, Z.X.; Song, W.G.; Ba, R.; Li, X.L.; Xia, L. A Spatiotemporal Contextual Model for Forest Fire Detection Using Himawari-8 Satellite Data. Remote Sens. 2018, 10, 1992. [Google Scholar] [CrossRef]
  18. Hong, Z.H.; Tang, Z.Z.; Pan, H.Y.; Zhang, Y.W.; Zheng, Z.S.; Zhou, R.Y.; Ma, Z.L.; Zhang, Y.; Han, Y.L.; Wang, J.; et al. Active Fire Detection Using a Novel Convolutional Neural Network Based on Himawari-8 Satellite Images. Front. Environ. Sci. 2022, 10, 794028. [Google Scholar] [CrossRef]
  19. Li, X.L.; Song, W.G.; Lian, L.P.; Wei, X.G. Forest Fire Smoke Detection Using Back-Propagation Neural Network Based on MODIS Data. Remote Sens. 2015, 7, 4473–4498. [Google Scholar] [CrossRef]
  20. Kaufman, Y.J.; Justice, C.O.; Flynn, L.P.; Kendall, J.D.; Prins, E.M.; Giglio, L.; Ward, D.E.; Menzel, W.P.; Setzer, A.W. Potential global fire monitoring from EOS-MODIS. J. Geophys. Res.-Atmos. 1998, 103, 32215–32238. [Google Scholar] [CrossRef]
  21. Robinson, J.M. Fire from space-global fire evaluation using infrared remote-sensing. Int. J. Remote Sens. 1991, 12, 3–24. [Google Scholar] [CrossRef]
  22. Plank, S.; Fuchs, E.M.; Frey, C. A Fully Automatic Instantaneous Fire Hotspot Detection Processor Based on AVHRR ImageryA TIMELINE Thematic Processor. Remote Sens. 2017, 9, 30. [Google Scholar] [CrossRef]
  23. Liu, X.Z.; He, B.B.; Quan, X.W.; Yebra, M.; Qiu, S.; Yin, C.M.; Liao, Z.M.; Zhang, H.G. Near Real-Time Extracting Wildfire Spread Rate from Himawari-8 Satellite Data. Remote Sens. 2018, 10, 1654. [Google Scholar] [CrossRef]
  24. Roberts, G.J.; Wooster, M.J. Fire detection and fire characterization over Africa using Meteosat SEVIRI. IEEE Trans. Geosci. Remote Sens. 2008, 46, 1200–1218. [Google Scholar] [CrossRef]
  25. Prata, A.J. Observations of volcanic ash clouds in the 10–12 μm window using AVHRR/2 data. Int. J. Remote Sens. 1989, 10, 751–761. [Google Scholar] [CrossRef]
  26. Giglio, L.; Schroeder, W.; Justice, C.O. The collection 6 MODIS active fire detection algorithm and fire products. Remote Sens. Environ. 2016, 178, 31–41. [Google Scholar] [CrossRef]
  27. Schroeder, W.; Oliva, P.; Giglio, L.; Csiszar, I.A. The New VIIRS 375 m active fire detection data product: Algorithm description and initial assessment. Remote Sens. Environ. 2014, 143, 85–96. [Google Scholar] [CrossRef]
  28. Na, L.; Zhang, J.Q.; Bao, Y.L.; Bao, Y.B.; Na, R.S.; Tong, S.Q.; Si, A. Himawari-8 Satellite Based Dynamic Monitoring of Grassland Fire in China-Mongolia Border Regions. Sensors 2018, 18, 276. [Google Scholar] [CrossRef]
  29. Jain, P.; Coogan, S.C.P.; Subramanian, S.G.; Crowley, M.; Taylor, S.; Flannigan, M.D. A review of machine learning applications in wildfire science and management. Environ. Rev. 2020, 28, 478–505. [Google Scholar] [CrossRef]
  30. Miller, J.; Borne, K.; Thomas, B.; Huang, Z.; Chi, Y. Automated Wildfire Detection Through Artificial Neural Networks. In Proceedings of the East FIRE Conference, Fairfax, VA, USA, 11–13 May 2005; pp. 293–304. [Google Scholar]
  31. Pinto, M.M.; Libonati, R.; Trigo, R.M.; Trigo, I.F.; DaCamara, C.C. A deep learning approach for mapping and dating burned areas using temporal sequences of satellite images. Isprs J. Photogramm. Remote Sens. 2020, 160, 260–274. [Google Scholar] [CrossRef]
  32. Priya, R.S.; Vani, K. Deep Learning Based Forest Fire Classification and Detection in Satellite Images. In Proceedings of the 2019 11th International Conference on Advanced Computing (ICoAC), Chennai, India, 18–20 December 2019; pp. 61–65. [Google Scholar]
  33. Udousoro, I.C. Machine Learning: A Review. Semicond. Sci. Inf. Devices 2020, 2, 5–14. [Google Scholar]
  34. Chen, W.; Bai, S.; Zhao, H.M.; Han, X.R.; Li, L.H. Spatiotemporal analysis and potential impact factors of vegetation variation in the karst region of Southwest China. Environ. Sci. Pollut. Res. 2021, 28, 61258–61273. [Google Scholar] [CrossRef]
  35. Bessho, K.; Date, K.; Hayashi, M.; Ikeda, A.; Imai, T.; Inoue, H.; Kumagai, Y.; Miyakawa, T.; Murata, H.; Ohno, T.; et al. An Introduction to Himawari-8/9-Japan’s New-Generation Geostationary Meteorological Satellites. J. Meteorol. Soc. Jpn. 2016, 94, 151–183. [Google Scholar] [CrossRef]
  36. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  37. Jang, E.; Kang, Y.; Im, J.; Lee, D.W.; Yoon, J.; Kim, S.K. Detection and Monitoring of Forest Fires Using Himawari-8 Geostationary Satellite Data in South Korea. Remote Sens. 2019, 11, 271. [Google Scholar] [CrossRef]
  38. Liaw, A.; Wiener, M. Classification and Regression by randomForest. R News 2002, 23, 18–22. [Google Scholar]
  39. Schroeder, W.; Oliva, P.; Giglio, L.; Quayle, B.; Lorenz, E.; Morelli, F. Active fire detection using Landsat-8/OLI data. Remote Sens. Environ. 2016, 185, 210–220. [Google Scholar] [CrossRef]
  40. Xu, G.; Zhong, X. Real-time wildfire detection and tracking in Australia using geostationary satellite: Himawari-8. Remote Sens. Lett. 2017, 8, 1052–1061. [Google Scholar] [CrossRef]
  41. Giglio, L.; Descloitres, J.; Justice, C.O.; Kaufman, Y.J. An enhanced contextual fire detection algorithm for MODIS. Remote Sens. Environ. 2003, 87, 273–282. [Google Scholar] [CrossRef]
  42. Atkinson, P.M. Sub-pixel target mapping from soft-classified, remotely sensed imagery. Photogramm. Eng. Remote Sens. 2005, 71, 839–846. [Google Scholar] [CrossRef]
  43. Flasse, S.P.; Ceccato, P. A contextual algorithm for AVHRR fire detection. Int. J. Remote Sens. 1996, 17, 419–424. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of the research area. (a) depicts some of the previous fire point distributions. (b) depicts the elevation data of the study area.
Figure 1. Schematic diagram of the research area. (a) depicts some of the previous fire point distributions. (b) depicts the elevation data of the study area.
Forests 14 00919 g001
Figure 2. The process flow of the algorithm.
Figure 2. The process flow of the algorithm.
Forests 14 00919 g002
Figure 3. The distribution of samples in the dataset.
Figure 3. The distribution of samples in the dataset.
Forests 14 00919 g003
Figure 4. Feature importance of the dataset. The horizontal axis is the abbreviation of features, the meaning of which is shown in Table 2.
Figure 4. Feature importance of the dataset. The horizontal axis is the abbreviation of features, the meaning of which is shown in Table 2.
Forests 14 00919 g004
Figure 5. Presentation of fire event verification results. The red point represents the fire pixels that the algorithm detected (detected fire cases in Table 4).
Figure 5. Presentation of fire event verification results. The red point represents the fire pixels that the algorithm detected (detected fire cases in Table 4).
Forests 14 00919 g005
Figure 6. Detection of fire that occurred in Mianning on 20 April 2021; the red point represents the fire pixels that the algorithm detected.
Figure 6. Detection of fire that occurred in Mianning on 20 April 2021; the red point represents the fire pixels that the algorithm detected.
Forests 14 00919 g006
Table 1. Himawari-8 AHI Imager specifications.
Table 1. Himawari-8 AHI Imager specifications.
TypeBand Wavelength   ( μ m ) Spatial   Resolution   ( k m )
Visible10.471
20.57
30.610.5
Near-infrared40.861
51.62
62.3
Infrared73.9
86.2
96.9
107.3
118.6
129.6
1310.4
1411.2
1512.4
1613.3
Table 2. Features included in the dataset.
Table 2. Features included in the dataset.
Information CategoryContent and Abbreviation
AlbedoAlbedo of bands 1 (a1), 2 (a2), 3 (a3), 4 (a4), and 6 (a6)
Brightness temperatureBrightness temperature of bands 7 (bt7), 11 (bt11), 14 (bt14), and 15 (bt15) and the difference between the brightness temperature of band 7 and band 14 (dt)
AngleSolar zenith angle (soz) and satellite zenith angle (saz)
Spatial contextual informationThe mean, variance, and mean deviation of the brightness temperature of the background pixels at bands 7 (meanw7, devw7, and varw7) and 14 (meanw14, and devw14, varw14) and the difference between them in the window (dmeanw, ddev, and dvarw)
Table 3. Machine learning model performance evaluation with WLF for reference.
Table 3. Machine learning model performance evaluation with WLF for reference.
Model/ProductTotal NumberNumber of Same Pixels to Reference
WLF2525
Random forest model4523
CNN2414
Table 4. Fire cases detected.
Table 4. Fire cases detected.
NumberDateLatitude (N)Longitude (E)SiteDetected or Not
124 April 202130.4100.24Ganzidetected
229 November 202232.92102.94Hongyuandetected
316 February 202229.04101.46Jiulongdetected
415 March 202229.04103.92Muchuandetected
528 March 202027.96101.26Mulidetected
620 April 202128.5102.2Mianningdetected
723 April 202126.7299.72Lijiangdetected
827 March 202127.32100.82Lijinagdetected
914 March 202125.22102.72Kunmingdetected
1030 March 202027.84102.18Xichangdetected
113 January 202130.24101.68Ganzimissed
1213 March 202128.6101.94Mianningmissed
1319 March 202128.68102.24Mianningmissed
142 May 202128.46104.46Yibinmissed
Table 5. Accuracy evaluation with WLF for reference.
Table 5. Accuracy evaluation with WLF for reference.
Data to Be EvaluatedReference DataOmission Error RateCommission Error RateF1 Score
Result of the proposed algorithmWLF0.300.260.72
Table 6. Accuracy evaluation with VIIRS for reference.
Table 6. Accuracy evaluation with VIIRS for reference.
Data to Be EvaluatedReference DataOmission Error RateCommission Error RateF1 Score
WLFVIIRS0.70.240.43
Random forest model in this studyVIIRS0.640.470.42
Result of the proposed algorithmVIIRS0.660.230.47
Table 7. Accuracy comparison with VIIRS for reference.
Table 7. Accuracy comparison with VIIRS for reference.
Data to Be EvaluatedReference DataOmission Error RateCommission Error RateF1 Score
WLFVIIRS0.70.240.43
Result of the proposed algorithmVIIRS0.660.230.47
Results of algorithm BVIIRS0.820.130.29
Results of algorithm CVIIRS0.730.140.41
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, C.; Chen, R.; He, B. Integrating Machine Learning and a Spatial Contextual Algorithm to Detect Wildfire from Himawari-8 Data in Southwest China. Forests 2023, 14, 919. https://doi.org/10.3390/f14050919

AMA Style

Liu C, Chen R, He B. Integrating Machine Learning and a Spatial Contextual Algorithm to Detect Wildfire from Himawari-8 Data in Southwest China. Forests. 2023; 14(5):919. https://doi.org/10.3390/f14050919

Chicago/Turabian Style

Liu, Chuanfeng, Rui Chen, and Binbin He. 2023. "Integrating Machine Learning and a Spatial Contextual Algorithm to Detect Wildfire from Himawari-8 Data in Southwest China" Forests 14, no. 5: 919. https://doi.org/10.3390/f14050919

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop