Next Article in Journal
Anchor-free Convolutional Network with Dense Attention Feature Aggregation for Ship Detection in SAR Images
Next Article in Special Issue
Automatic Grapevine Trunk Detection on UAV-Based Point Cloud
Previous Article in Journal
A New Fast Factorized Back-Projection Algorithm with Reduced Topography Sensibility for Missile-Borne SAR Focusing with Diving Movement
Previous Article in Special Issue
Crop Row Detection through UAV Surveys to Optimize On-Farm Irrigation Management
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Evaluation of Rapeseed Winter Crop Damage Using UAV-Based Multispectral Imagery

by
Łukasz Jełowicki
1,
Konrad Sosnowicz
2,
Wojciech Ostrowski
3,
Katarzyna Osińska-Skotak
3,* and
Krzysztof Bakuła
3
1
OPEGIEKA Sp. z o.o., 82-300 Elbląg, Poland
2
Skysnap Sp. z o.o., 02-001 Warsaw, Poland
3
Department of Photogrammetry, Remote Sensing and Spatial Information Systems, Faculty of Geodesy and Cartography, Warsaw University of Technology, 00-661 Warsaw, Poland
*
Author to whom correspondence should be addressed.
Remote Sens. 2020, 12(16), 2618; https://doi.org/10.3390/rs12162618
Submission received: 30 June 2020 / Revised: 31 July 2020 / Accepted: 12 August 2020 / Published: 13 August 2020
(This article belongs to the Special Issue UAS-Remote Sensing Methods for Mapping, Monitoring and Modeling Crops)

Abstract

:
This research is related to the exploitation of multispectral imagery from an unmanned aerial vehicle (UAV) in the assessment of damage to rapeseed after winter. Such damage is one of a few cases for which reimbursement may be claimed in agricultural insurance. Since direct measurements are difficult in such a case, mainly because of large, unreachable areas, it is therefore important to be able to use remote sensing in the assessment of the plant surface affected by frost damage. In this experiment, UAV images were taken using a Sequoia multispectral camera that collected data in four spectral bands: green, red, red-edge, and near-infrared. Data were acquired from three altitudes above the ground, which resulted in different ground sampling distances. Within several tests, various vegetation indices, calculated based on four spectral bands, were used in the experiment (normalized difference vegetation index (NDVI), normalized difference vegetation index—red edge (NDVI_RE), optimized soil adjusted vegetation index (OSAVI), optimized soil adjusted vegetation index—red edge (OSAVI_RE), soil adjusted vegetation index (SAVI), soil adjusted vegetation index—red edge (SAVI_RE)). As a result, selected vegetation indices were provided to classify the areas which qualified for reimbursement due to frost damage. The negative influence of visible technical roads was proved and eliminated using OBIA (object-based image analysis) to select and remove roads from classified images selected for classification. Detection of damaged areas was performed using three different approaches, one object-based and two pixel-based. Different ground sampling distances and different vegetation indices were tested within the experiment, which demonstrated the possibility of using the modern low-altitude photogrammetry of a UAV platform with a multispectral sensor in applications related to agriculture. Within the tests performed, it was shown that detection using UAV-based multispectral data can be a successful alternative for direct measurements in a field to estimate the area of winterkill damage. The best results were achieved in the study of damage detection using OSAVI and NDVI and images with ground sampling distance (GSD) = 10 cm, with an overall classification accuracy of 95% and a F1-score value of 0.87. Other results of approaches with different flight settings and vegetation indices were also promising.

Graphical Abstract

1. Introduction

The European Union is the global leader in rapeseed production, and Poland is one of the largest rapeseed producers and processors in Europe. In 2019, it was second only to France [1]. In 2010–2019, the area of rapeseed sown in Poland ranged from 720,000 to 950,000 hectares, representing about 8–10% of the total crops, and about 95–96% of the oilseed acreage [2].
Due to the climate changes observed in recent years (i.e., snow-free winters), yield losses caused by poor wintering of plants are becoming more frequent. Low temperatures are the most dangerous for crops, combined with no snow or a thin layer of snow cover. Rapid warming causing thaws, followed by frosts, are also not conducive to wintering plants. The result of such winter weather is freezing winter crops, i.e., damage to the crops. If there is no snow cover, winter rapeseed may freeze at −10 °C [3]. Annual losses in rapeseed yields in 2015–2019 were estimated to range from 10% to 20% [2]. To assess damages for agricultural insurance, an assessment of its occurrence in the field is performed. These are imprecise measurements made based on a sample selected in the field whereby an expert determines the number of plants per square meter and takes this as a representation of one hectare of the affected field. The results of these measurements depend on the expert’s experience. In the case of rapeseed crops, compensation is awarded when there are less than 15 healthy plants per square meter and the damage covers at least 10% of the field area. The damage threshold values (min. 10% and max. 50% of the parcel area) are particularly important because, according to the scope of insurance, they affect the final decision on the qualification of loss in crops (partial or total compensation). Since the methods of assessing damages used by insurance companies are not very accurate and are time consuming, it appears that the use of remote sensing techniques may be an excellent alternative.
Currently, research conducted in the area of remote sensing applications with the use of unmanned aerial vehicles (UAV) is often related to aspects of vegetation monitoring, particularly relating to forest and agricultural areas. UAV-based remote sensing in agriculture is particularly related to pest and disease detection, development of crops during their growth cycle, assessment of biomass, and water stress in plants [4]. To date, these phenomena have been assessed through satellite and airborne photogrammetry [5,6,7,8]. The breakthrough solutions in terms of image acquisition techniques are those using UAVs [9,10,11]. There is a need to receive continuous spatial information with a high level of accuracy and actuality, while retaining low operating costs. Drones can provide an alternative, providing high resolution images with a short revisit time, even every couple of hours, contrary to optical satellite systems or aerial images which have a much lower time resolution.
To date, analyses connected with object geometry, such as an assessment of parcel areas, landslide monitoring, and cropland and forest inventory, have been widely applied in many countries [12,13,14]). However, applications of sensors with more than three spectral bands (e.g., multispectral or hyperspectral cameras) mounted on unmanned aerial vehicles are not as commonly used. Research on implementation of these cameras and integration of various remote sensing techniques is a key area of interest for many groups of scientists [15,16,17,18].
Remote sensing techniques for vegetation mapping play an important role in precise agriculture [9,19] and phenotyping research of crops [20,21,22]. Remote monitoring of arable lands using spectral libraries (obtained from field measurements, satellite imagery, or aerial and close-range images) could be helpful in the analysis of plant growth [19,22], assessment of the size of harvest [23] and crop fertilizing needs [24,25], pest control [24], and the extraction of dead plants. For many years, the Remote Sensing Centre at the Institute of Geodesy and Cartography in Poland has carried out monitoring of farmland with the use of satellite imagery to forecast harvests [26]. The accuracy of the prepared models for the prediction of crop volumes using NOAA AVHRR data is estimated at 90–95%. A high correlation between a harvest’s size and its spectral characteristics based on field measurements [27] is a clear indicator of the potential of remote sensing data for such research. This translates into numerous areas of scientific research, involving spectral vegetation indices in the assessment of the physical condition of flora [5,19,27,28].
The use of spectral indices aims to extract essential information about vegetation condition, and the indices should indicate a proper correlation between their values and biophysical parameters of plant cover, such as biomass or leaf area index (LAI) [5,19,27,29]. The most popular spectral indices in this area are those that use big reflectance differences between the near infrared region and the red band. This kind of index includes the normalized difference vegetation index (NDVI) and simple ratio (SR). Based on the research carried out to date, it is observed that there is a link between the above-mentioned parameters and the size of the biomass and LAI [5]. The outcome of scientific studies has confirmed that a high accuracy of the performed estimations (85%) [18,19] was found, and that there are strong and statistically significant relationships between NDVI (obtained from UAV) and crop biophysical variables (for LAI, R2 was 0.95; for ground truth percent canopy cover, R2 was 0.93). The obtained strong relationships suggest that UAV multispectral data can be used for estimation of LAI and percent canopy cover. Wei et al. (2017) [30] concluded in their research that derivative spectral indices formulated using optimized narrow wavebands were most effective in quantifying the changes in pigment and water content of leaves subjected to freezing injury.
Frost is one of the environmental stresses of plants. The plant response to frost can be a decrease in chlorophyll, inhibition of photosynthesis, altered leaf angle, and plant freezing [31]. The influence of frost on biochemical and biophysical changes, as well as spectral properties of various plant species, have been investigated [30,32,33,34,35,36]. It was found that freezing causes changes in the spatial differentiation of chlorophyll content on the lamina surface [32] and the structural changes of mesophyll cells [30]. In turn, these changes affect the spectral reflection of various plant parts, especially in blue-green, green, red, and near infrared radiation [30,33,34,35,36]. However, as the analyzed studies show, these changes depend on the species studied and the degree of freezing.
In order to analyze the effects of crop freezing in winter, it is essential that the influence of the spectral reflectance of soil should be included. For practical implementation, it is common to use the soil-adjusted vegetation index (SAVI). Because of the difficulty in choosing the optimal value of the coefficient describing soil brightness and the low sensitivity to a small amount of chlorophyll, the modified chlorophyll absorption ratio index (MCARI) was introduced. This parameter highlights an amount of chlorophyll absorption in the range of about 670 nm in relation to spectral characteristics centered at 550 and 700 nm [6]. With the use of band normalization in the spectral range of red edge and red (R700, R670), it is possible to minimize the influence of soil and extract information about vegetation without green pigments. Another method of implementation is to use the transformed chlorophyll absorption reflectance index (TCARI)/optimized soil adjusted vegetation index (OSAVI). The use of normalization enables the detection of biomass variability even at low LAI values [6]. Furthermore, Hunt’s research (2008) [28] confirms the effectiveness of the green normalized difference vegetation index (GNDVI, [37]), which differentiates biomass content within the area of winter wheat crops. Moreover, studies carried out by many scientists provides evidence that GNDVI correlates significantly with the nutrient ingredients (mainly nitrogen compounds) [28,38]. Among the methods used, it is worth mentioning the fusion of chosen spectral indices. As an example, the use of NDVI, the modified soil-adjusted vegetation index (MSAVI)/the transformed soil-adjusted vegetation index (TSAVI) (in the case where spectral bands in the shortwave infrared (SWIR) range are available) and GNDVI enables the assessment of the spatial variability of cropland, including its impact on agricultural yields. The opportunities from using remote sensing data outlined above appear to be a reasonable solution to crop insurance, including estimating the winterkill losses. Multispectral sensors have been significantly miniaturized in recent years, and their current price allows for the commercialization of the presented method without significant financial outlay. In contrast to satellite data, the quality of which is determined by cloud cover range, the use of UAVs in time-series analysis is not as strongly related to weather conditions. This fact is essential for regions of moderate climate zones because winter crop inventories are conducted in the late autumn which is usually very foggy. Furthermore, an application of the appropriate methodology for data processing and analysis enables farmers and insurance providers to accurately estimate winterkill losses and the amounts of compensation paid by insurers. One of the most important steps in this process is damage detection at the beginning of the vegetation period.
The purpose of this study was to: (1) evaluate the possibility of using UAVs for estimating losses in rapeseed crops caused by poor wintering, (2) determine the impact of data acquisition parameters (mainly the altitude influencing the ground sampling distance (GSD)), and (3) present the methodology of data processing (proper vegetation index) to obtain the best results compared to the traditional method of field inventory.

2. Materials and Methods

The experiment was carried out to analyze the possibility of using UAV-based multispectral imagery in the detection of damaged rapeseed after winter and the estimation of the damage area. The experiment was carried out in cooperation with an insurance company. Figure 1 shows the scheme of the experiment performed. Data were registered during three test flights on different flight heights in order to evaluate the influence of image spatial resolution. Each time, remote sensing data were acquired in four spectral bands, namely green, red, red edge, and near infrared, which were used to calculate tested vegetation indices. RGB images were acquired and combined with high-resolution spectral data, which were later used for manual preparation of reference data. Training samples were obtained during the field campaign.
In the methodology, there were a few crucial elements which are described in this section. Firstly, selected indices were tested to evaluate their usefulness in further analysis. The important part of the experiment was to delineate technical roads and replace those areas with a mean value from neighborhoods to exclude the effect of machinery influence (technical roads) on estimating damage. Another issue in the experiment was to define the influence of spatial resolution of the multispectral data on the obtained results.

2.1. Test Area and Data Used

For the planned tests, a 34 ha parcel with a rapeseed field was selected. It is situated in Olecko County in Northern Poland (Figure 2). This region is well known for its low temperatures and harsh climate throughout the year, hence, it was chosen for this experiment on the detection of winter crop damage. The winter of 2016/2017 in this region was characterized by high variability of weather conditions, with alternate periods of frost and thaw. During the period of severe frosts (from −12 to −27 °C) there was a thin layer of snow of several centimeters. This caused winter damage, which was confirmed during a field inspection.
The testing data were acquired with the use of an eBee UAV equipped with a Sequoia multispectral camera. This allowed for the acquisition of data in four different spectral bands: green (530–570 nm), red (640–680 nm), red edge (730–740 nm), and near infrared (770–810 nm). Three flights were performed on 20 April 2017, at three different altitudes of 50, 100, and 200 m above the ground. This resulted in images with three different spatial resolutions—ground sampling distances (GSD) of 5, 10, and 20 cm, respectively. This allowed us to assess the impact of spatial resolution on the imagery in the detection of a crop’s health when undamaged by frost. The overlapping of successive images along a flight strip (endlap) was 80%, and the overlap between flight strips (sidelap) was 70%. Flights were carried out in midday hours (12:00 p.m.–2:00 p.m.), with a cloudless or slightly cloudy sky. Wind speed was 1–2 m/s, air temperature was 5–6 °C, and humidity was 32–37%. The visibility was 35.5–43.8 km.
The orientation of images was performed in Pix4DMapper Pro software (Pix4D SA, Switzerland). This software allows the user to select the appropriate calculation template. For multispectral data calculation, the AG (agriculture) multispectral template is preferred. The program automatically recognizes the camera, retrieves its parameters from the database, groups the images into the appropriate spectral bands, and enables them to be calibrated. A useful feature of the Sequoia camera is that it can measure the amount of irradiance coming from the sun for each image and for each band with the sunshine sensor. This enables the user to take irradiance into account and normalize the value of the images amongst themselves. In the orientation process, self-calibration with 15 ground control points (control and check points) measured with a GNSS receiver was also performed to give a final root mean square error (RMS) of aerial triangulation of 10 cm. The result of the images processed in Pix4DMapper Pro were reflectance orthomosaics.
Based on the orthomosaics generated from the collected images in four spectral bands, different vegetation indices were calculated, based on which further experiments were conducted. To perform the correct classification of healthy and damaged crops, 23 training fields were created using data from field inspection and verified on RGB images. The training fields contained information about the number of healthy plants over a surface of 1 square meter (Figure 3). This helped to describe the relationship between the vegetation indices values and the number of plants in a surface unit, which resulted in thresholds for image classification.

2.2. Selection of Vegetation Indices

The most important process in determining post-winter damage is the correct detection of areas where the crop has not grown properly, or where the plants have died. Remote sensing as a multidisciplinary technique is a tool with many indices, which can help test the health of plants and solve the problem of the value of compensation for farmers for damaged crops. The Sequoia camera was the equipment used on the UAV in this study and allowed recording in four spectral bands: green, red, red edge, and near infrared. As a result of a literature review, a set of vegetation indicators was identified for further analysis. Consequently, after preliminary tests, the final selection was based on three basic spectral indices of vegetation: NDVI, SAVI, and OSAVI. In addition, an interesting solution in the calculation of these indices was to replace the red channel with the red edge, which was supported by the multispectral camera used in the experiment. Such modification was carried out and three further indices were examined: NDVI_RE, SAVI_RE, OSAVI_RE.
The NDVI (normalized difference vegetation index) is one of the most basic and commonly used indices [39]. It is a measure of photosynthetic activity and determines the condition of the plant and its development stage. Plants absorb most of the red light that hits it while reflecting much of the near infrared light. When vegetation is dead or stressed it reflects more red light and less near infrared light. Based on these properties, NDVI is calculated as the normalized difference between the red and near infrared bands from an image. The basic formula for calculating the index value is as follows:
NDVI = NIR     RED NIR   +   RED
where NIR is the near infrared band value and RED is the red band value.
The SAVI (soil adjusted vegetation index) is a modification of the NDVI index [40]. Due to changes in the numerator and denominator of the typical NDVI equation (soil brightness correction factor), the SAVI index further improves the final result for the influence of the soil brightness. For this reason, it is used in situations where a large part of the crop is not covered with lush vegetation. Therefore, the beginning of the growing season appears to be an excellent time for examining the condition of plants using this index. The formula for calculating SAVI is as follows:
SAVI = ( 1 + L ) ( NIR     RED ) NIR   +   RED + L
where L denotes the coverage of the vegetation area. In most cases, and particularly for the case of intermediate vegetation canopy levels, optimal results are achieved using L = 0.5. For this experiment, a value 0.5 was used during the tests.
The OSAVI (optimized soil adjusted vegetation index) was developed by Rondeaux et al. (1996) [41] and was first presented in the work entitled “Optimized Soil Adjusted Vegetation Index”. This index is a modification of the SAVI that has been optimized for agricultural monitoring. It is more sensitive to changes in plant condition than the original SAVI and does not need a priori knowledge of the soil type. A value of 0.16 as the soil adjustment coefficient was selected as the optimal value to minimize variation with soil background. The formula for calculating OSAVI is as follows:
OSAVI = NIR     RED NIR   +   RED + 0.16
In addition to the above three spectral indices, their modifications were tested in which the red channel was replaced by red edge. By analyzing spectral reflectance properties for plants in red edge wavelength, we can observe quick changes in reflection and can thus expect to see more discreet differences between healthy and damaged plants. To test this premise, NDVI_RE, SAVI_RE, and OSAVI_RE indices were calculated. The application of such indices with a modified channel has been discussed previously [42,43].
The correctness of damage recognition by the vegetation index was examined using training fields measured during a field inventory with the participation of insurance specialists. Twenty-tree sites were surveyed with an area of approximately 1 square meter and the condition of the plants was determined in their range. These fields were then characterized by experts as to whether they would qualify for compensation or if the plants in the field area were healthy. A training field was also specified that could be treated as two classes on the border: damaged and healthy crops. The training fields were ranked according to the condition of the plants. After calculating the mean values of the tested indices for each training field, the values obtained were compared with the reference from the field inventory. The threshold value was established based on the training fields which, according to experts, were found not to meet the conditions for granting compensation. Based on the healthy training field, threshold values for each index were used to classify (or not) training fields for compensation. All training fields were verified with an expert in the field. Their correctness was ensured so the outlier fields were eliminated in defining the thresholds. The threshold value was calculated as the mean value from the lowest mean value of the index for an accepted healthy crop training field and the highest mean value of the index for an accepted damaged crop training field. A comparison of results for training fields with defined thresholds is shown in Figure 4. As can be seen, the training fields were characterized by differentiation of the values of the vegetation indices. The highest standard deviation was noted in the fields with healthy plants, and the lowest in damaged areas. The greatest variability of the index values was related to the occurrence of plants of different sizes and condition. Training field no. 9 recorded higher NDVI and SAVI values due to the presence of weeds; furthermore, in training field no. 11, there were 10 healthy, well-developed plants, which also resulted in higher NDVI values.
Based on the selected thresholds, referring to each index, Figure 5 illustrates the errors in training fields, verified by inspection of the field. The red areas represent the test fields that, according to specialists, should be classified as damaged areas. The green areas represent fields without visible frozen influence. The threshold shown by a horizontal line in Figure 4 determined the correctness of interpretation for each training field: red areas should be below the threshold and green above. If the situation is reversed, it means that tested index does not work properly. After summarizing the errors, the top 3 results were finally selected, i.e., NDVI, OSAVI, and OSAVI_RE, to carry out further experiments.
In addition, considering the training fields, it is clear that the worst results were obtained for field no. 21. This was a unique area with healthy but very small plants (significant influence of the soil background). This suggests that the accepted threshold method may be sensitive to the stage in which the plants are grown, and results in slight over detection of damaged areas.

2.3. Elimination of Road Influence

Regardless of the selected vegetation index, the influence of technical roads is clearly visible on the classification results. The classification of technical roads, which are visible as traces of tires on a part of the field in poor condition, could have a big impact on the final results of detection, because this area should not be included in the calculation of the area of rapeseed damaged by frost.
The main insurer’s condition applied to an area for reimbursement is that the area of a detected polygon must cover some minimum area (i.e., larger than 100 m2). Therefore, many smaller objects are removed from the analysis. However, small isolated damaged areas are sometimes related to technical roads, which can result in detection of one bigger area, and are thus mistakenly detected as an area eligible for reimbursement. To prevent such situations, it was decided to work on rasters without the influence of roads on the reimbursement area.
For this purpose, road detection was conducted, based on compositions consisting of NDVI and green, red, and red edge channels. Object-based classification was performed in eCognition software for each dataset (data with spatial resolution equal to 5, 10, and 20 cm). In such a classification, information about spectral values, object shape, and extent were used. This resulted in three different road images, one for each spatial resolution of 5, 10, and 20 cm. Detected road images differed significantly from one another. The most precise and complete images came from images of 5 cm, and the worst from the lowest resolution (see Figure 6).
The biggest problems with areas of roads are presented in Figure 7, where there is no visible border between road and bare soil, which is why some tires traces were not detected. However, these places do not need to be detected because they do not cause the described problem of connecting polygons. Subsequently, a road mask was subtracted from the index rasters and the resulting holes were replaced with a mean value from corresponding cells (see Figure 7). Thus, the negative influence of technical roads on the final results of classification and detection of regions for reimbursement was reduced.

2.4. Approaches for Classification

The detection of damaged areas was performed using three different vegetation indices (NDVI, OSAVI, OSAVI_RE) in three different classification approaches, which were used to avoid the influence of technical roads on the result of winter damage crop estimation. The first method was a simple object-based classification to delineate the areas eligible for reimbursement. The main assumption behind this method was that it should appropriately suit borders of reference polygons which represent real situations and were vectorized without the generalization of polygons.
The second and third approach exploited pixel-based classification with thresholds. Index images were intersected with grids of cell sizes of 1 × 1 m and 5 × 5 m. For cells, a mean value of the index was assigned, and each cell was classified and merged with corresponding cells with the same class, creating homogenous regions. Thus, the resulting polygon geometry was simplified. By using those two approaches, we could also simulate a decrease in the resolution of input data.
In every approach, objects larger than 100 m2 were finally recognized as detected damages. However, smaller regions lying within 5 m of other suspected damaged objects were also used, if the sum of their area was bigger than 100 m2.
The possibility of proper detection of crops damaged by frost was examined using many variants. The influence of road elimination was examined, in addition to using a vegetation index for detection and spatial resolution of collected data. This resulted in 42 variants.

2.5. Accuracy Assessment

To estimate the accuracy of the final indications of which areas in each variant were eligible for reimbursement, a reference image was created by vectorization of damaged crops on an orthomosaic (see Figure 8).
The results were compared with reference areas by creating an error matrix which allowed for calculating such parameters as producer’s accuracy (PA, completeness), user’s accuracy (UA, correctness), F1-score, and overall accuracy (OA) [44,45]. The F1-score is the harmonic mean of precision and sensitivity and is usually used as an accuracy measure of a dichotomous model [45], so it is suitable for one-class delineation. In compensation assessment, the accurate area size of detected damage is much more important for the insurance company than its position, so another parameter was also used. This was the basic ratio of detected damaged areas to the reference area, and for the purpose of this study we named it “area index” (AI).

3. Results

The results of all variants for damaged rapeseed detection are presented in Table A1 in Appendix A. Figure 9 illustrates the comparison of the overall accuracy, the F1-score, and AI for all variants of the experiment. The visual effect of the classification for selected variants is presented in Figure 10.
Most variants of classification allow detection of winterkill with OA above 90% (Figure 9a). UA and PA ranged between 70–80% and 90–100%, respectively (Table A1 in Appendix A). This means that the algorithm detects almost every polygon from our reference set but also mistakenly indicates other areas (overestimation). The poorer result in UA may indicate some mistakes in the reference set, which was difficult to vectorize objectively in a field, and because the algorithm may overestimate the detection of crop losses, which is confirmed by the area of damage often exceeding 110% to 130% (Figure 9c). However, it should be noted that the overestimation relates to transition areas, which are inherently difficult to clearly separate (see Figure 11). Generally, the best results were achieved using OSAVI and NDVI indices (the highest UA, OA, and F1-score values). The highest accuracy (measured by F1-score and OA) with the smallest overestimation of the area of losses in crops (AI) was obtained for a spatial resolution of 10 cm. The worst results (low OA and F1, and high AI) were obtained for OSAVI_RE, using images with GSD = 5 cm. The area for compensation was overestimated by 29–31%, which may be caused, for example, by inaccuracies in the georeference of an image with a spatial resolution of 5 cm.
In the next section, a deeper analysis of the influence of road removal, the spatial resolution of collected images, and the vegetation indices used in this experiment is provided.

4. Discussion

4.1. Influence of Road Removal and Approaches in Damage Detection

As expected, road removal was a crucial step for obtaining better results. It allowed us to avoid many errors in the classification of smaller objects connected with roads (see Figure 7). It also prevented the incorrect enlargement of detected losses by incorporating road areas in detected polygons.
The negative influence of roads is highlighted in all three approaches on 10 cm input rasters. In both the second (grid 1 × 1 m) and the third (grid 5 × 5 m) approaches, each index increases UA by 3–5 percentage points, with a slight decrease of PA (around three percentage points) with only one exception of a higher outlier (10 percentage points for OSAVI in grid 5 × 5 m) (see Table A1 in Appendix A). The direction of changes in the first approach (object-based) are similar. The decrease of PA is on the same level as in other approaches, however, the increase in UA equals 17 percentage points on OSAVI and NDVI indices, and only two percentage points for OSAVI _RE, which is due to a much bigger UA value for this index before road removal (73–63% and 64% for NDVI and OSAVI, respectively). Although the differences in the range of increase of UA are high, the discrepancies in the final results between approaches 1 and 2 do not exceed three percentage points. This is also reflected by the area index (see Figure 11). The area index’s values decrease in all three approaches after using the road removal technique. For OSAVI and NDVI indices, the difference is over two times higher in the object-based approach, but for OSAVI_RE, the decrease is on the same level as in the pixel-based approaches, which may be due to the fact that in the OSAVI_RE variant, technical roads are slightly less visible and are harder to distinguish in the images. This results in a lower number of roads that qualified as winter damage polygons.
The conducted analysis indicates that without removing road influence, object-based classification is a significantly worse method for proper winter damage detection. The generalization that is performed during the process of averaging the cell value in approaches 2 and 3 (resolution of grids 1 and 5 m, respectively), causes the rejection of some fraction of roads, which is not the case in the object-based approach because it provides a more precise classification. Because of this characteristic, a higher resistance to road influence in the OSAVI_RE index is most observable in the object-based scenario.

4.2. Influence of Spatial Resolution

The comparison of PA and UA for a 1 × 1 m grid with respect to vegetation indices is shown in Figure 12. Comparing NDVI results in this approach, we can assume that the smaller the GSD (ground sampling distance of obtained images), the better the results (an increase in the F1-score is observed). However, UA and PA values are almost equal for resolutions of 5 and 10 cm. The visible difference comes only at the 20 cm raster (a decrease of three percentage points in UA and 0.02 in F1-score value). In the case of the OSAVI index, we can confirm that results are slightly better on the 10 cm raster than 5 cm, but 20 cm still results in the worst outcome. The OSAVI_RE index trend is contrary to NDVI, with most of the indicator’s values rising with decreasing pixel size. In the second approach, a 20 cm input OSAVI_RE raster results in a F1-score value of 0.84 and an OA equal to 93% (Table A1 in Appendix A), which is only slightly worse than the results obtained from other indices. The detected damage area is also more similar to the reference data than in comparable variants. This information allows us to suppose that, with an increasing size of GSD, OSAVI_RE could be the best index to use in the whole process. Generally, it can be assumed that any tested spatial resolution of UAV-based imagery is enough for winter damage assessment for insurance purposes and there are few differences in terms of accuracy in the tested approaches in the detection of these areas without plants.
For the results obtained in the described experiments with NDVI, decreasing the size of GSD provided better results, while the research of Candiago et al. (2015) [9] showed that the use of such a high resolution can be a problem when applying algorithms originally developed for aerial and satellite data. In this paper, three approaches were tested with OBIA, or resampling to grid sizes of 1 × 1 m and 5 × 5 m, which limited this problem.
The highest overall accuracy of detection wais achieved using the NDVI index on two different input data sets with 5 and 20 cm GSD (see Table A1 in Appendix A). For rasters with 10 cm resolution, the better choice was the OSAVI index, which was clearly confirmed in the majority of variants. However, the difference between the overall accuracy of damage detection using NDVI and OSAVI is not significant (1–2%). The usefulness of the NDVI vegetation index to determine the vegetation fraction (the percentage of green vegetation per unit of ground surface) and the number of plants per unit of ground surface is confirmed by the results of other authors (among others [19,22]). In our experiment, the results related to OSAVI_RE index are slightly better.

4.3. Influence of Vegetation Index Used

Figure 13 presents the results of damage detection for different vegetation indices in the case of the three approaches. Analysis of this figure allows us to draw a few crucial conclusions. Firstly, we can state that the producer’s accuracy is stable in all three approaches regardless of chosen vegetation index as input data for classification. Its values vary from 0 to 3 percentage points in the same approach. However, approach 3 (grid 5 × 5 m) provides smaller absolute PA (around 85%) and UA (77%) values than the other approaches (mean values of 91% and 80%, respectively). This is mostly the result of much stronger generalization in this approach. Secondly, it is observed that there are significant differences in the user’s accuracy between the chosen indices in all three approaches.
In each variant, OSAVI_RE has the lowest value. OSAVI provides slightly better UA than NDVI (difference of 1–2 percentage points), but NDVI compensates this disparity with slightly higher PA values. Analysis of the overall accuracy (see Table A1 in Appendix A) indicates that those two indices are the same, however, the area of damage in all three approaches is 5 percentage points better for OSAVI than NDVI. The difference in the F1-score is only 0.01. Nonetheless, for data with a resolution of 5 cm (for both approaches) and 20 cm (for the third approach), the results indicate that NDVI is slightly better than OSAVI (see Table A1 in Appendix A).

5. Conclusions

The experiments performed demonstrate that low-altitude remote sensing can be used for the assessment of damaged areas in the case of insurance compensation. The case study used multispectral UAV images for the damage detection of rapeseed caused during winter. Data were acquired in April, when plant vegetation starts. This is the most important period in estimating damage since it directly exposes the effect of the winter period on vegetation. The presented methodology includes selecting possible vegetation indices, ground sampling resolutions, and approaches for technical road influence elimination, which can be used as an effective alternative for direct measurements in a field that currently are often used during the refurbishment process.
The analysis of GSD images (in the range of 5–20 cm) demonstrates a relatively small (a few percentage points) influence on the obtained results of winterkill detection. The best approximation of the area of the winter damage to crops was obtained using images with GSD of 10 cm. Results for most variants indicate that there is no need to acquire data with a higher resolution than 10 cm. This is a crucial conclusion for UAV application when the size of the area and the GSD of the images are considered within mission planning
The conclusion of the presented experiment is of great practical importance because cameras with four spectral channels (R, G, B, NIR) are becoming increasingly popular and relatively inexpensive. This provides the possibility to choose various vegetation indices appropriately for the analyzed crop and the purpose of the study related to the monitoring of the crop condition. The best results were obtained using NDVI and OSAVI vegetation indices; however, results from other indices were not significantly worse.
Referring to the utilization of the red edge spectral band, an optimized Soil-Adjusted Vegetation Index with the use of the red edge channel instead of the red channel is the most resilient index to the negative influence of technical roads, which is crucial when road detection cannot be conducted. The optimized soil-adjusted vegetation index with red edge channel (OSAVI_RE) may be the best for crop condition monitoring using lower spatial resolution data.
During tests, it was also necessary to develop and implement elimination of technical roads in calculations of rapeseed area damaged with frost. Including the detected area in the classification helps in the reliable calculation of the area for compensation. Without removing road influence, object-based classification is a significantly worse method, but elimination of road influence can be processed using both object- and pixel-based approaches. In tests with NDVI and OSAVI, overestimation of 10–30% was noticed for calculations without the elimination of the impact of roads, and all of the remote sensing-based approaches in all experiments were characterized by overestimation of at least a few percent considering the Area Index.
The implementation of the presented solutions for delimiting areas for refurbishment in the comprehensive damage estimation method, which also includes periodic monitoring and time analysis, can significantly improve the compensation process, and provide it with scientific certainty for obtained results.

Author Contributions

Conceptualization, K.S., W.O., and K.B.; methodology, W.O.; software, K.S.; validation, W.O.; formal analysis, K.O.-S.; investigation, Ł.J.; resources, K.S. and W.O.; writing—original draft preparation, Ł.J., W.O., K.O.-S., and K.B.; writing—review and editing, K.B. and K.O.-S.; visualization, K.O.-S. and Ł.J.; supervision, project administration and funding acquisition, K.O.-S. and K.B. All authors have read and agreed to the published version of the manuscript.

Funding

The presented research was financed by statutory subsidies of the Polish Ministry of Science and Higher Education for Warsaw University of Technology.

Acknowledgments

We would like to thank Skysnap Sp. z o.o. for providing the data from the Sequoia camera and participation in the calculations and analyzes, and Paulina Bartkowiak for the help in reference data preparation.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. The results of all variants within the experiment, assessed by producer’s accuracy (PA) and user’s accuracy (UA) for class damaged crops, and area index (AI), F1-score, and overall accuracy (OA) of classification for damaged crops and healthy crops.
Table A1. The results of all variants within the experiment, assessed by producer’s accuracy (PA) and user’s accuracy (UA) for class damaged crops, and area index (AI), F1-score, and overall accuracy (OA) of classification for damaged crops and healthy crops.
Resolution5 cm5 cm
ApproachGrid 1 × 1Grid 5 × 5
IndexUAPAAIF1OAUAPAAIF1OA
without influence of roadsNDVI83%93%113%0.8894%78%87%111%0.8292%
OSAVI80%93%116%0.8694%76%87%114%0.8192%
OSAVI_RE71%93%131%0.8191%68%88%129%0.7789%
with influence of roadsNDVI73%96%131%0.8392%73%90%123%0.8191%
OSAVI69%96%138%0.8091%70%90%129%0.7990%
OSAVI_RE65%95%146%0.7789%65%90%139%0.7588%
Resolution10 cm10 cm10 cm
ApproachGrid 1 × 1Grid 5 × 5Object-Based
IndexUAPAAIF1OAUAPAAIF1OAUAPAAIF1OA
without influence of roadsNDVI82%93%113%0.8795%78%86%111%0.8292%80%92%115%0.8694%
OSAVI84%91%108%0.8795%79%84%106%0.8192%81%90%110%0.8594%
OSAVI_RE77%92%120%0.8493%73%86%117%0.7991%76%89%117%0.8292%
with influence of roadsNDVI74%96%130%0.8492%73%90%123%0.8191%63%96%153%0.7688%
OSAVI75%94%126%0.8392%75%94%126%0.8392%64%94%146%0.7688%
OSAVI_RE72%94%130%0.8291%70%88%126%0.7890%73%91%126%0.8191%
Resolution20 cm20 cm
ApproachGrid 1 × 1Grid 5 × 5
IndexUAPAAIF1OAUAPAAIF1OA
without influence of roadsNDVI79%94%118%0.8694%78%86%111%0.8292%
OSAVI80%92%115%0.8694%76%86%113%0.8192%
OSAVI_RE79%90%114%0.8493%76%85%112%0.8092%
with influence of roadsNDVI76%100%131%0.8694%73%89%121%0.8091%
OSAVI77%93%121%0.8493%74%87%118%0.8091%
OSAVI_RE78%91%117%0.8493%75%86%115%0.8091%

References

  1. EUROSTAT DATA. 2019. Available online: https://ec.europa.eu/eurostat/databrowser/view/tag00100/default/table?lang=en (accessed on 17 July 2020).
  2. GUS. Statistical Yearbook of Agriculture. Warsaw. 2020. Available online: https://stat.gov.pl (accessed on 17 July 2020).
  3. Szulc, K. Wymarzanie Rzepaku Ozimego, Farmer.pl. 2018. Available online: https://www.farmer.pl/produkcja-roslinna/rosliny-oleiste/wymarzanie-rzepaku-ozimego,76896.html (accessed on 17 July 2020).
  4. Wójtowicz, M.; Wójtowicz, A.; Piekarczyk, J. Application of remote sensing methods in agriculture. Commun. Biometry Crop Sci. 2016, 11, 31–50. [Google Scholar]
  5. Brown, R.J.; Staenz, K.; McNairn, H.; Hopp, B.; Van Acker, R. Application of high-resolution optical imagery to precision agriculture. In Proceedings of the International Symposium Geomatics in the Era of RADARSAT (GER’97), Ottawa, ON, Canada, 25–30 May 1997; p. 9. [Google Scholar]
  6. Haboudane, D.; Miller, J.R.; Tremblay, N.; Zarco-Tejada, P.J.; Dextraze, L. Integrated narrow-band vegetation indices for prediction of crop chlorophyll content for application to precision agriculture. Remote Sens. Environ. 2002, 81, 416–426. [Google Scholar] [CrossRef]
  7. Doraiswamy, P.C.; Moulin, S.; Cook, P.W.; Stern, A. Crop yield assessment from remote sensing. Photogramm. Eng. Remote Sens. 2003, 69, 665–674. [Google Scholar] [CrossRef]
  8. Rembold, F.; Atzberger, C.; Savin, I.; Rojas, O. Using low resolution satellite imagery for yield prediction and yield anomaly detection. Remote Sens. 2013, 5, 1704–1733. [Google Scholar] [CrossRef] [Green Version]
  9. Candiago, S.; Remondino, F.; De Giglio, M.; Dubbini, M.; Gattelli, M. Evaluating multispectral images and vegetation indices for precision farming applications from UAV images. Remote Sens. 2015, 7, 4026–4047. [Google Scholar] [CrossRef] [Green Version]
  10. Gago, J.; Douthe, C.; Coopman, R.; Gallego, P.; Ribas-Carbo, M.; Flexas, J.; Medrano, H. UAVs challenge to assess water stress for sustainable agriculture. Agric. Water Manag. 2015, 153, 9–19. [Google Scholar] [CrossRef]
  11. Holman, F.H.; Riche, A.B.; Michalski, A.; Castle, M.; Wooster, M.J.; Hawkesford, M.J. High throughput field phenotyping of wheat plant height and growth rate in field plot trials using UAV based remote sensing. Remote Sens. 2016, 8, 1031. [Google Scholar] [CrossRef]
  12. Puliti, S.; Ørka, H.O.; Gobakken, T.; Næsset, E. Inventory of Small Forest Areas Using an Unmanned Aerial System. Remote Sens. 2015, 7, 9632–9654. [Google Scholar] [CrossRef] [Green Version]
  13. Lucieer, A.; Jong, S.M.D.; Turner, D. Mapping landslide displacements using Structure from Motion (SfM) and image correlation of multi-temporal UAV photography. Prog. Phys. Geogr. 2014, 38, 97–116. [Google Scholar] [CrossRef]
  14. Wei, Z.; Han, Y.; Li, M.; Yang, K.; Yang, Y.; Luo, Y.; Ong, S.-H. A Small UAV Based Multi-Temporal Image Registration for Dynamic Agricultural Terrace Monitoring. Remote Sens. 2017, 9, 904. [Google Scholar] [CrossRef] [Green Version]
  15. Khot, L.R.; Sankaran, S.; Carter, A.H.; Johnson, D.A.; Cummings, T.F. UAS imaging-based decision tools for arid winter wheat and irrigated potato production management. Int. J. Remote Sens. 2016, 37, 125–137. [Google Scholar] [CrossRef]
  16. Kipp, S.; Mistele, B.; Baresel, P.; Schmidhalter, U. High-throughput phenotyping early plant vigour of winter wheat. Eur. J. Agron. 2014, 52, 271–278. [Google Scholar] [CrossRef]
  17. Zhou, J.; Pavek, M.J.; Shelton, S.C.; Holden, Z.J.; Sankaran, S. Aerial multispectral imaging for crop hail damage assessment in potato. Comput. Electron. Agric. 2016, 127, 406–412. [Google Scholar] [CrossRef] [Green Version]
  18. Salamí, E.; Barrado, C.; Pastor, E. UAV flight experiments applied to the remote sensing of vegetated areas. Remote Sens. 2014, 6, 11051–11081. [Google Scholar] [CrossRef] [Green Version]
  19. Shi, Y.; Thomasson, J.A.; Murray, S.C.; Pugh, N.A.; Rooney, W.L.; Shafian, S.; Rajan, N.; Rouze, G.; Morgan, C.L.S.; Neely, H.L.; et al. Unmanned Aerial Vehicles for High-Throughput Phenotyping and Agronomic Research. PLoS ONE 2016, 11, e0159781. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Zaman-Allah, M.; Vergara, O.; Araus, J.L.; Tarekegne, A.; Magorokosho, C.; Zarco-Tejada, P.J.; Hornero, A.; Hernández Albà, A.; Das, B.; Craufurd, P.; et al. Unmanned aerial platform-based multi-spectral imaging for field phenotyping of maize. Plant Methods 2015, 11, 1–10. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Yang, G.; Liu, J.; Zhao, C.; Li, Z.; Huang, Y.; Yu, H.; Xu, B.; Yang, X.; Zhu, D.; Zhang, X.; et al. Unmanned Aerial Vehicle Remote Sensing for Field-Based Crop Phenotyping: Current Status and Perspectives. Front. Plant Sci. 2017, 8, 1–26. [Google Scholar] [CrossRef]
  22. Ni, J.; Yao, L.; Zhang, J.; Cao, W.; Zhu, Y.; Tai, X. Development of an Unmanned Aerial Vehicle-Borne Crop-Growth Monitoring System. Sensors 2017, 17, 502. [Google Scholar] [CrossRef] [Green Version]
  23. Shanahan, J.F.; Schepers, J.S.; Francis, D.D.; Varvel, G.E.; Wilhelm, W.W.; Tringe, J.M.; Schlemmer, M.R.; Major, D.J. Use of Remote-Sensing Imagery to Estimate Corn Grain Yield. Agron. J. 2001, 93, 583–589. [Google Scholar] [CrossRef] [Green Version]
  24. Zhang, C.; Walters, D.; Kovacs, J.M. Applications of Low Altitude Remote Sensing in Agriculture upon Farmers’ Requests—A Case Study in Northeastern Ontario, Canada. PLoS ONE 2014, 9, e112894. [Google Scholar] [CrossRef]
  25. Hunt, E.R.; Hively, W.D.; Daughtry, C.S.; McCarty, G.W.; Fujikawa, S.J.; Ng, T.L.; Yoel, D.W. Remote sensing of crop leaf area index using unmanned airborne vehicles. In Proceedings of the Pecora 17 Symposium, Denver, CO, USA, 18–20 November 2008. [Google Scholar]
  26. Bochenek, Z.; Dąbrowska-Zielińska, K.; Ciołkosz, A.; Drupka, S.; Boken, V.K. Monitoring Agricultural Drought in Poland. Monitoring and Predicting Agricultural Drought; Oxford University Press: Oxford, UK, 2005; pp. 171–180. [Google Scholar]
  27. Piekarczyk, J.; Sulewska, H.; Szymańska, G. Winter oilseed-rape yield estimates from hyperspectral radiometer measurements. Quaest. Geogr. 2011, 30, 77–84. [Google Scholar] [CrossRef] [Green Version]
  28. Hunt, E.R.; Hively, W.D.; Fujikawa, S.J.; Linden, D.S.; Daughtry, C.S.T.; McCarty, G.W. Acquisition of NIR-Green-Blue Digital Photographs from Unmanned Aircraft for Crop Monitoring. Remote Sens. 2010, 2, 290–305. [Google Scholar] [CrossRef] [Green Version]
  29. Mathews, A.J.; Jensen, J.L.R. Visualizing and quantifying vineyard canopy LAI using an unmanned aerial vehicle (UAV) collected high density structure from motion point cloud. Remote Sens. 2013, 5, 2164–2183. [Google Scholar] [CrossRef] [Green Version]
  30. Wei, C.; Huang, J.; Wang, X.; Blackburn, G.A.; Zhang, Y.; Wang, S.; Mansaray, L.R. Hyperspectral characterization of freezing injury and its biochemical impacts in oilseed rape leaves. Remote Sens. Environ. 2017, 195, 56–66. [Google Scholar] [CrossRef] [Green Version]
  31. Jones, H.G.; Vaughan, R.A. Remote Sensing of Vegetation—Principles, Techniques, and Applications; Oxford University Press: New York, NY, USA, 2010. [Google Scholar]
  32. Nicotra, A.B.; Hofmann, M.; Siebke, K.; Ball, M.C. Spatial patterning of pigmentation in evergreen leaves in response to freezing stress. Plant Cell Environ. 2003, 26, 1893–1904. [Google Scholar] [CrossRef]
  33. Wu, Q.; Zhu, D.; Wang, C.; Ma, Z.; Wang, J. Diagnosis of freezing stress in wheat seedlings using hyperspectral imaging. Biosyst. Eng. 2012, 112, 253–260. [Google Scholar] [CrossRef]
  34. Flower, K.; Boru, B.; Nansen, C.; Jones, H.; Thompson, S.; Lacoste, C.; Murphy, M. Proof of Concept: Remote Sensing Frost-Induced Stress in Wheat Paddocks; Grains Research and Development Corporation: Canberra, Australia, 2014. [Google Scholar]
  35. Murphy, M.E.; Boruff, B.; Callow, J.N.; Flower, K.C. Detecting Frost Stress in Wheat: A Controlled Environment Hyperspectral Study on Wheat Plant Components and Implications for Multispectral Field Sensing. Remote Sens. 2020, 12, 477. [Google Scholar] [CrossRef] [Green Version]
  36. Xie, Y.; Wang, C.; Yang, W.; Feng, M.; Qiao, X.; Song, J. Canopy hyperspectral characteristics and yield estimation of winter wheat (Triticum aestivum) under low temperature injury. Sci. Rep. 2020, 10, 244. [Google Scholar] [CrossRef] [Green Version]
  37. Gitelson, A.A.; Kaufman, Y.J.; Merzlyak, M.N. Use of a green channel in remote sensing of global vegetation from EOS-MODIS. Remote Sens. Environ. 1996, 58, 289–298. [Google Scholar] [CrossRef]
  38. Daughtry, C.S.T.; Walthall, C.L.; Kim, M.S.; Brown de Colstoun, E.; McMurtrey, J.E., III. Estimating corn leaf chlorophyll concentration from leaf and canopy reflectance. Remote Sens. Environ. 2000, 74, 229–239. [Google Scholar] [CrossRef]
  39. Rouse, J.W. Monitoring the Vernal Advancement and Retrogradation (Green Wave Effect) of Natural Vegetation; Technical Report; NASA: Washington, DC, USA, 1972. [Google Scholar]
  40. Huete, A.R. A soil-adjusted vegetation index (SAVI). Remote Sens. Environ. 1988, 25, 295–309. [Google Scholar] [CrossRef]
  41. Rondeaux, C.; Steven, M.D.; Baret, F. Optimized of Soil-adjusted vegetation indices. Remote Sens Environ. 1996, 55, 95–107. [Google Scholar] [CrossRef]
  42. Schuster, C.; Förster, M.; Kleinschmit, B. Testing the red edge channel for improving land-use classifications based on high-resolution multi-spectral satellite data. Int. J. Remote Sens. 2012, 33, 5583–5599. [Google Scholar] [CrossRef]
  43. Li, F.; Miao, Y.; Feng, G.; Yuan, F.; Yue, S.; Gao, X.; Liu, Y.; Liu, B.; Ustin, S.L.; Chen, X. Improving estimation of summer maize nitrogen status with red edge-based spectral vegetation indices. Field Crop. Res. 2014, 157, 111–123. [Google Scholar] [CrossRef]
  44. Congalton, R. A Review of Assessing the Accuracy of Classifications of Remotely Sensed Data. Remote Sens. Environ. 1991, 37, 35–46. [Google Scholar] [CrossRef]
  45. Koga, Y.; Miyazaki, H.; Shibasaki, R. A CNN-based Method of vehicle detection from aerial images using hard example mining. Remote Sens. 2018, 10, 124. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The scheme of the experiment performed.
Figure 1. The scheme of the experiment performed.
Remotesensing 12 02618 g001
Figure 2. Test area, rapeseed field in Olecko, Poland.
Figure 2. Test area, rapeseed field in Olecko, Poland.
Remotesensing 12 02618 g002
Figure 3. Examples of the training fields (1 m × 1 m).
Figure 3. Examples of the training fields (1 m × 1 m).
Remotesensing 12 02618 g003
Figure 4. Interpretation of training fields with selected thresholds according to indices: Normalized Difference Vegetation Index (NDVI) (a), Optimized Soil Adjusted Vegetation Index (OSAVI) (b), Soil Adjusted Vegetation Index (SAVI) (c), Normalized Difference Vegetation Index—Red Edge (NDVI_RE) (d), Optimized Soil Adjusted Vegetation Index—Red Edge (OSAVI_RE) (e), and Soil Adjusted Vegetation Index—Red Edge (SAVI_RE) (f). The training fields are ranked according to the condition of plants. The chart shows the average value of index for each of the training fields with the standard deviation.
Figure 4. Interpretation of training fields with selected thresholds according to indices: Normalized Difference Vegetation Index (NDVI) (a), Optimized Soil Adjusted Vegetation Index (OSAVI) (b), Soil Adjusted Vegetation Index (SAVI) (c), Normalized Difference Vegetation Index—Red Edge (NDVI_RE) (d), Optimized Soil Adjusted Vegetation Index—Red Edge (OSAVI_RE) (e), and Soil Adjusted Vegetation Index—Red Edge (SAVI_RE) (f). The training fields are ranked according to the condition of plants. The chart shows the average value of index for each of the training fields with the standard deviation.
Remotesensing 12 02618 g004
Figure 5. Errors of the classification for training fields intended for compensation process.
Figure 5. Errors of the classification for training fields intended for compensation process.
Remotesensing 12 02618 g005
Figure 6. Detected roads on orthomosaics with ground sampling distance (GSD) of 5 cm (a), 10 cm (b), and 20 cm (c).
Figure 6. Detected roads on orthomosaics with ground sampling distance (GSD) of 5 cm (a), 10 cm (b), and 20 cm (c).
Remotesensing 12 02618 g006
Figure 7. Presentation of impact of tires traces shown on NDVI images and the results of classification of winter damage for crops: (a) NDVI image, (b) NDVI image after elimination of influence of roads, (c,d) the results of classification of winter damages for crop based on the NDVI image, before and after elimination of influence of roads.
Figure 7. Presentation of impact of tires traces shown on NDVI images and the results of classification of winter damage for crops: (a) NDVI image, (b) NDVI image after elimination of influence of roads, (c,d) the results of classification of winter damages for crop based on the NDVI image, before and after elimination of influence of roads.
Remotesensing 12 02618 g007
Figure 8. Reference data: crop field map manually vectorized.
Figure 8. Reference data: crop field map manually vectorized.
Remotesensing 12 02618 g008
Figure 9. The comparison of the overall accuracy (OA) (a), the F1-score (b), and area index (AI) (c) of all variants of classifications.
Figure 9. The comparison of the overall accuracy (OA) (a), the F1-score (b), and area index (AI) (c) of all variants of classifications.
Remotesensing 12 02618 g009
Figure 10. Small fragment of true color orthomosaic with reference mask and detected areas of damaged rapeseed without the influence of roads: (a) in three different approaches, (b) with three different indices, (c) with three different resolutions (d) true color orthomosaic.
Figure 10. Small fragment of true color orthomosaic with reference mask and detected areas of damaged rapeseed without the influence of roads: (a) in three different approaches, (b) with three different indices, (c) with three different resolutions (d) true color orthomosaic.
Remotesensing 12 02618 g010
Figure 11. Area index (AI) in three approaches with 10 cm resolution of input data.
Figure 11. Area index (AI) in three approaches with 10 cm resolution of input data.
Remotesensing 12 02618 g011
Figure 12. Producer’s (PA) and user’s (UA) accuracy of the detected damaged crops, without the influence of roads, with three different resolution of input data in the second approach (1 × 1 m).
Figure 12. Producer’s (PA) and user’s (UA) accuracy of the detected damaged crops, without the influence of roads, with three different resolution of input data in the second approach (1 × 1 m).
Remotesensing 12 02618 g012
Figure 13. Producer’s (PA) and user’s (UA) accuracy of results without the influence of roads, with three different approaches performed on 10 cm input data.
Figure 13. Producer’s (PA) and user’s (UA) accuracy of results without the influence of roads, with three different approaches performed on 10 cm input data.
Remotesensing 12 02618 g013

Share and Cite

MDPI and ACS Style

Jełowicki, Ł.; Sosnowicz, K.; Ostrowski, W.; Osińska-Skotak, K.; Bakuła, K. Evaluation of Rapeseed Winter Crop Damage Using UAV-Based Multispectral Imagery. Remote Sens. 2020, 12, 2618. https://doi.org/10.3390/rs12162618

AMA Style

Jełowicki Ł, Sosnowicz K, Ostrowski W, Osińska-Skotak K, Bakuła K. Evaluation of Rapeseed Winter Crop Damage Using UAV-Based Multispectral Imagery. Remote Sensing. 2020; 12(16):2618. https://doi.org/10.3390/rs12162618

Chicago/Turabian Style

Jełowicki, Łukasz, Konrad Sosnowicz, Wojciech Ostrowski, Katarzyna Osińska-Skotak, and Krzysztof Bakuła. 2020. "Evaluation of Rapeseed Winter Crop Damage Using UAV-Based Multispectral Imagery" Remote Sensing 12, no. 16: 2618. https://doi.org/10.3390/rs12162618

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop