Next Article in Journal
Fault Estimation Method for Nonlinear Time-Delay System Based on Intermediate Observer-Application on Quadrotor Unmanned Aerial Vehicle
Previous Article in Journal
Federated Learning Attacks Revisited: A Critical Discussion of Gaps, Assumptions, and Evaluation Setups
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automated Detection of Rice Bakanae Disease via Drone Imagery

1
IT Application Research Center, Korea Electronics Technology Institute, Jeonju 54853, Republic of Korea
2
Horticultural Research Institute, Jeonnam Agricultural Research & Extension Services, Iksan 58213, Republic of Korea
3
Ministry of Agriculture, Food and Rural Affairs, 94 Dasom 2-ro, Government Complex-Sejong, Sejong-si 30110, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(1), 32; https://doi.org/10.3390/s23010032
Submission received: 26 October 2022 / Revised: 9 December 2022 / Accepted: 14 December 2022 / Published: 20 December 2022
(This article belongs to the Section Smart Agriculture)

Abstract

:
This paper proposes a system for the forecasting and automated inspection of rice Bakanae disease (RBD) infection rates via drone imagery. The proposed system synthesizes camera calibrations and area calculations in the optimal data domain to detect infected bunches and classify infected rice culm numbers. Optimal heights and angles for identification were examined via linear discriminant analysis and gradient magnitude by targeting the morphological features of RBD in drone imagery. Camera calibration and area calculation enabled distortion correction and simultaneous calculation of image area using a perspective transform matrix. For infection detection, a two-step configuration was used to recognize the infected culms through deep learning classifiers. The YOLOv3 and RestNETV2 101 models were used for detection of infected bunches and classification of the infected culm numbers, respectively. Accordingly, 3 m drone height and 0° angle to the ground were found to be optimal, yielding an infected bunches detection rate with a mean average precision of 90.49. The classification of number of infected culms in the infected bunch matched with an 80.36% accuracy. The RBD detection system that we propose can be used to minimize confusion and inefficiency during rice field inspection.

1. Introduction

Rice is an essential food resource in Asia, and global concerns over rice self-sufficiency have increased as the topic of food sovereignty has been emphasized due to instabilities in trade associated with COVID 19 and the Ukraine–Russia war [1]. In addition to its role as a major resource for stabilizing food supply and demand, rice occupies an important position in terms arable land utilization and employment as well.
With recent dietary changes due to globalization, an increase in the consumption of high-quality rice has been documented [2,3]. Accordingly, seed quality control is increasing in importance. To this end, rice seed and field inspections are conducted in Korea. For the field inspections, cultivar purity, weed germination, and disease occurrence tests are performed on paddy fields that have been sufficiently isolated to prevent natural hybridization between different seeds. Since it is essentially an inspection of seed dissemination, field inspections are largely carried out during seeding seasons. However, the frequency and timing of inspection can be adjusted for early detection and removal of specific diseases [4]. There are various diseases that can occur in rice, namely blast disease, sheath blight disease, bacterial blight disease, rice stripe virus, and rice Bakanae disease (RBD) [5].
RBD is caused by a fungal pathogen. The disease continues to spread from the infected rice to the surrounding area. As the infected seedlings die or produce chaff, rice production is directly affected, and the fungus attaches to the normal rice seeds, causing the disease to reoccur when sowing in the following year [5]. Therefore, prevention of repeated infection cycles through early detection of RBD is essential for its effective control. Specifically, RBD-infected rice bunches are easier to distinguish from normal rice bunches before seeds are formed (Figure 1). Therefore, field inspections are conducted at that period. RBD field inspections generally calculate the morbidity rate based on the number of infected culms compared to the total rice field area, as determined by experts [6]. When inspecting an entire rice paddy field, inspection objectivity is often inconsistent due to expert bias, and is exacerbated by differences in weather, sunlight, location, or imaging angle [7]. Since multiple cross-validation by numerous experts is time-intensive and costly, automated detection of RBD is required to continuously achieve consistent test results and reduce resource costs [5,8].
Preprocessing for disease classification and detection is conventionally performed using plant imagery. Morphological information, such as the length of a plant or the width of a leaf, can also be directly collected by an observer. Image preprocessing into an advantageous form to achieve accurate identification is generally performed to enhance the differences between diseased and normal regions. Accordingly, classical preprocessing methods, such as support vector machine (SVM) or machine vision models have been employed [9,10]. However, since the distribution of preprocessed data is distinct from that of existing data, it may result in the over-fitting of more slight environmental changes [11].
RBD detection studies have commonly employed quantifying characteristics such as seedling height and width, the angle of the forked culm, S a t u r a t i o n from RBD images taken according to a certain standard, and SVM classification using the quantified data [5,12,13]. Importantly, it remains difficult to collect data in hard-to-access places, such as the interior of a rice paddy field. To this end, accessibility restrictions have been overcome by performing data collection via drones [8]. Accordingly, this prior study exploited the distinct color and S a t u r a t i o n corresponding to disease-affected parts of rice plants. Therefore, preprocessing was performed by applying a threshold value to the H u e and I n t e n s i t y values of the image for increasing its contrast, followed by segmentation. Elsewhere, a deep learning object detection model was trained to detect the location of various plant disease parts such as apple black rot or tomato leaf spot within an image, and the resulting validity of the detection results was verified based on the learning results and visualization to produce a heat map of infection [14,15,16]. A further study on sequential systems performed disease detection over a wide range to select suspected infected areas, and detecting and classifying the diseased areas within these ranges [17,18]. The results showed that false detection over a wide range could be reduced, and the infection site could be more accurately detected, which inspired the two-step method of our study.
The present study aimed to devise a system for identifying the RBD-infected culms during field inspections for early detection and control. In order to analyze the infection rate via field inspections, precise areal calculations and detailed detections of culm numbers are required. Since the camera of the drone used for field inspection is a mono-vision camera, the area cannot be calculated alone. Therefore, to determine the exact area, image height and angle were specified through camera calibration, and data collection was limited to a controlled environment. For RBD infection detection, a two-step configuration was used to recognize the infected culms through deep learning classifiers. The YOLOv3 and RestNETV2 101 models were used to detect infected bunches and determine the infected culm numbers, respectively. In the process of classifying the number of infected culms in a bunch, focal loss function and transfer learning were adopted to resolve the data imbalance between each class.

2. Proposed System

For RBD-infected rice, aging progresses faster than normal. As a result, the infected rice grows thinner and longer than normal rice, while its color is yellower and less saturated than that of normal rice leaves, as shown in Figure 2. Assuming that such patterns are found in paddy fields, observations from above are advantageous, as the infected portions are covered by normal rice when viewed from the side.
In general, deep learning models exhibit higher levels of performance with an increase in training data amount. Therefore, certain limitations can be overcome by increasing the training data amount. Since the collection and labeling of drone images is both time and labor intensive, it remains a challenge to collect sufficient data for adequate generalization of the deep learning model across all height and angle imaging conditions. Furthermore, since the automation of rice field inspection to be performed encompasses the total rice-planted area, maintaining the imaging height and angle constant is advantageous. When detecting RBD from images collected via drone, symptoms may or may not appear clearly, depending on the specific imaging height or angle, in conjunction with the environmental conditions. Therefore, in the present study, the optimal shooting height and angle for RBD detection were selected. Subsequently, through camera calibration and the construction of a viewpoint transformation matrix, it was possible to calculate a specific area at the selected height and angle [19,20].
Rice bunches were positioned at equal intervals (Figure 3). A single bunch of rice contains several stems that can bear seeds (Figure 4), which are defined as individual “culms”. During field inspection of the RBD-infected culms, calculations were performed by summing the number of infected culms, whereas when the detection model was trained through the RBD data collected at the selected shooting height and angle, the output value refers to the infected bunch, so additional analyses are required. Therefore, in this study, the RBD detection model was configured in two stages: to find the candidate regions for RBD infected bunches via the deep learning model, and to classify the number of culms infected with the disease.
Figure 5 describes the system proposed for the analysis of the RBD-infected rate based on the aforementioned related research and prior information. The proposed system configuration preceded RBD image collection using drones. The data for deciding optimal shooting height and angle were collected for each condition across specific vertical heights and angles. Subsequently, the optimal vertical height and angle values were inferred by analyzing the data for each condition. Afterward, camera calibration was performed to minimize the effects of distortion, and automatically calculate the ground area under inferred vertical height and angle. While collecting data inferring the number of infected culms, RBD-containing images were collected under optimal shooting conditions, and bounding boxes of bunches were labeled and the number of culms counted, using the images. The labeled data were used to train a detection model for extracting the total area of RBD bunches, and to train a classification model to identify the number of infected culms contained within the infected bunches. When the configuration and learning of each module was completed, the entire system detected the RBD bunch area with respect to the area of interest from the drone image. By performing classification based on the detected bunch areas, it is possible to quantitatively calculate the number of RBD-infected culms relative to the total area.

2.1. Selection of Drone Shooting Height and Angle Using Linear Discriminant Analysis and Gradient Intensity

If the cost of data collection and labeling is low, the data collected from all shooting heights and angles can be used as input data for training RBD detection and classification models. However, since deep learning models generally require a lot of data, it is practically impossible to collect data in all cases. Therefore, the optimal shooting height and angle were inferred using a simpler model in the present study, focusing on the morphological aspects of RBD (Figure 6). The collected data were cropped. Specifically, a single bunch of RBD infections observed at the same location in 16 separate imaging conditions was cropped to fit. Additionally, a corresponding normal rice bunch that was the same size as the infected bunch was cropped, which yielded 16 RBD bunches and 16 samples of normal rice bunches corresponding to each height and angle. In total, 91 sets of data were collected, consisting of a total of 2912 (16 × 91 × 2) image data points. Figure 7 displays a set of example images in which the area infected with RBD, and normal rice areas were cropped out.
The cropped images of the RBD and normal rice bunch areas have different resolutions, respectively. However, to analyze this with linear discriminant analysis (LDA) and gradient magnitude, all images must have the same resolution. Therefore, all images were resized to 29(W) × 29(H), which was the smallest resolution among all images. LDA and gradient magnitude were used to quantify RBD. Specifically, LDA was used to compare the color information of RBD. LDA is a linear dimension reduction supervised learning method. Converting an image from the RGB color space to the HSI color space gives requires the calculation of H u e and S a t u r a t i o n (Equations (1) and (2)):
H u e = cos 1 0.5 × ( ( R G ) + ( R B ) ) ( R G ) 2 + ( R B ) ( G B )
S a t u r a t i o n = 1 3 R + G + B min ( R , G , B )
As stated, it is possible to use the distinct yellowish color and lower S a t u r a t i o n compared to normal rice to distinguish RBD. Accordingly, H u e can be advantageous in the extraction of RBD information, whereas S a t u r a t i o n can help identify areas of low S a t u r a t i o n . Of the 2912 normal and RBD bunches image data, each image became a 1682 (29 × 29 × 2) dimensional vector, which was divided into two classes—RBD and normal—via LDA to create a one-dimensional projection space. Scalar values can be obtained by projecting the data for each of the 16 shooting environments into a one-dimensional projection space. Scalar values computed from 2912 image data points belong to one of 16 classes. The calculated scalar values were normalized using the Mahalanobis distance or Batarachaya distance, and then the average value was used to obtain the recognition degree (Figure 8a).
Gradient magnitude was used to calculate the average texture feature values when images contained RBD, as it was expected that edges with a stronger I n t e n s i t y than that of normal rice would be included in terms of edge detection. Accordingly, the I n t e n s i t y was obtained using Equation (3):
I n t e n s i t y = R + G + B 3
The gradient values in horizontal and vertical directions are obtained by applying an edge filter to the I n t e n s i t y . Accordingly, the magnitude values can be acquired using the Euclidean distance from the gradient values in the two directions (Figure 8). The gradient magnitude value is the average value of the gradient magnitudes of each pixel in the 29 × 29 image, and it corresponds to the right portion of Figure 8. The gradient magnitude values of the RBD and normal rice classes were 21.94 and 20.67, respectively, indicating that the RBD class had a high gradient magnitude value.
Since the one-dimensional LDA and gradient magnitude values obtained using H u e , S a t u r a t i o n , and I n t e n s i t y maintain different scales, the values were normalized by converting them to a standard normal distribution, with a mean of 0 and a variance of 1 using the whitening formula shown in Equation (4):
x ˜ = x μ σ
where x is the data before whitening, μ is the average of the data, σ is the standard deviation of the data, and x ˜ is the whitened data. The average value of each whitening-processed one-dimensional LDA and gradient magnitude value became the final recognition degree.

2.2. Camera Calibration

Due to resolution limitations, the drone-attached camera could not accurately calculate the number of rice plants planted in the paddy fields, no matter which algorithm was applied. Therefore, it was impossible to calculate the infected rate of RBD by bunches or by culms based on imagery information alone. Due to the mechanization of agriculture, however, the distances between rows and columns of transplanted rice were standardized to 20 cm and 30 cm, respectively (Figure 3), and accordingly, it became possible to calculate the amount of rice grown per area. Specifically, 16.67 bunches were cultivated per 1 m2, and an average of 23 rice culms were grown per bunch, resulting in 383.33 culms cultivated per 1 m2. Based on this information, accurate camera calibrations could calculate the RBD-infected.

2.3. RBD Bunch Area Detection and Culm Classification

Characteristic symptoms of include plants growing 1.5 times taller than normal during the growing season, as well as the death of the culm as the leaves turn light green [21]. Detecting RBD from drone imagery based on these features is across various environments is thus complicated due to its relative similarity to normal, non-infected rice. Furthermore, the inconspicuous nature and relatively low areas of RBD occurrence also inhibit the efficient extraction of information. Accordingly, to reduce the misdetection of RBD, it was necessary to sacrifice the processing speed instead of increasing the accuracy of the algorithm. In addition, considering that the field inspection of RBD-infected was carried out in the middle of a rice field, it was difficult to operate equipment requiring additional electricity supply. However, it remains necessary to present the results onsite wherever possible for efficient exchange of information. Therefore, a deep learning object detection model with high accuracy and fast detection speed, even across various and complex environments, was selected. In addition, to identify the number of RBD-infected culms across the total area of field inspection, classifications of culms with RBD-detected bunches were run as well. Here, a series of interwoven processes for the proposed RBD detection are shown in Figure 9, as well as the number of RBD culms in the detected bunch region.

2.3.1. Model Structure

First, five candidate deep learning object detection models were selected for pre-evaluation of RBD-infected bunch region detection. Based on a comparison of mean average precision (mAP) and frames per second (FPS) in the open Pascal VOC 2007 dataset [22], the YOLOv3 model was selected.
Table 1 shows the performance comparison information of the deep learning object detection models for the dataset.
The YOLOv3 model was configured to detect the area bunches of RBD (Figure 9). Since a large amount of high-quality data is required when training deep learning object detection models, transfer learning [30,31,32] was performed using the weights of the DarkNet 53 model pre-trained with ImageNet [33], an open dataset used for image recognition. Since the pre-trained DarkNet 53 model used an input with an aspect ratio of 1:1, a cropped image of size 3208 × 555 was divided into size 555 × 555 . To minimize the loss occurring in the boundary region when dividing the image, a patch image was generated by overlapping a certain region in the form of a sliding window (Figure 10).

2.3.2. Classification of Culm Number in the Detected RBD-Infected Bunch Area

For the bunch areas detected through the YOLOv3 model, the images of the bunch areas were extracted from the bounding box coordinates predicted from the input image, and then the number of culms was classified by going through an additional convolutional neural network (Figure 9). When using this method for RBD identification, the ambiguity of culm classification, as well as locations where RBD bunch regions do not show clear features due to changes in weather or light sources, can affect the classification performance. As shown in Figure 7a, an RBD-infected culm is composed of one or more leaves. It should yield only one infested culm, even if two or more leaves appear to be infected. However, inaccurate labeling can lead to a significant drop in classification performance with a single infected culm misclassified as two or more.
To prevent performance degradation due to such ambiguity, consistent expert labeling is required. Here, the width and height of the RBD bounding box detected was not uniform. Accordingly, bounding boxes of various sizes were resized so that the width and height were 1:1, and they could then be input into the classifier. In the present study, a classification experiment was conducted using the deep residual network model with an input size of 224 × 224 pixels. Deep residual networks can be used to resolve deep learning model problems, such as vanishing gradients, exploding gradients, as well as increasing parameter numbers for increasing the number of layers of a neural network using a residual block structure, and 1 × 1 convolution structures [34]. With such characteristics, the model showed high performance feature extraction and classification. Accordingly, since the deep residual network has the potential to show high performance while readily controlling the speed and performance according to the increase in parameter number and layers by changing the number of repeated structures, this model was used for culm number classification.

3. Experiments and Results

3.1. Experiment Environment

In this paper, the DJI ‘Mavic2 Pro’ drones were operated to collect data for the detection of RBD under different environmental conditions, such as different illumination, diverse seasons, or weather conditions. The camera installed in the device has a resolution of 5472 pixels horizontally, 3648 pixels vertically, and a field of view of 77°. When the shooting view was calculated according to the technique shown in Figure 11, where the drone is operated with a 3 m vertical height, and a downward shooting angle of 0°, an image that detects 9 m left and right at a minimum detection distance of 6.8 m can be obtained.
Experiments and evaluations were performed on a workstation in Windows 10 environment, equipped with Intel Xeon(R) Gold 6230 R CPU 2.10 GHz and 128 GB RAM. Each module for the experiment was implemented in Python 3.6. Tensorflow 1.15 library was used for training and evaluation of deep learning algorithms, and 2 GeForce RTX 2080TI 11 GB GPUs were used for learning and evaluating each deep learning module.

3.2. Dataset Details

3.2.1. Drone Shooting Height and Angle Selection

Table 2 shows the composition of the dataset collected for selecting the imaging height and angle.
In order to increase accuracy, various environmental conditions were included in the data. For this purpose, images were collected from July to August 2020 in Namwon-si, Gimje-si, and Gunsan-si, Jeollabuk-do, Republic of Korea, with all areas corresponding to ordinary paddy fields in the plain area. Drone imaging vertical heights from the ground were set at 1, 2, 3, and 4 m. When the observation height exceeded 4 m, RBD was not visually identified due to the drone camera resolution. Therefore, only heights up to 4 m were used in the experiment. Furthermore, if the shooting down angle exceeded 15°, diffuse reflection would be induced, making it challenging to distinguish between RBD-infected and normal culm using the naked eye. Therefore, shooting angles were set to 0, 5, 10, and 15°.
The imaging was performed on various days so that various weather conditions were included under a range of back light and front light conditions. To observe the disease of rice in the same bunch, the drone remained focused in the same position and direction to take pictures in the order of 0°, 5°, 10°, and 15° at 1 m, and in the same order at 2, 3, and 4 m (Figure 12). When shooting, the shutter was pressed twice at a specific height and angle to take a picture, to avoid instances where infected plants cannot be identified because rice is bent by the wind, and shooting failure due to user error or data transmission error between drone and remote control. The normal number of shots was 126. However, for the reasons mentioned above, 124 to 128 images were acquired per condition, as shown in Table 2.
In the experiment, the same location RBD-infected bunch sample was labeled across all conditions. Since there few samples existed in the same location, 91 matched data were obtained for each height and angle condition. Similarly, normal bunches with the same location were also collected around 91 infected bunches under 16 conditions. Accordingly, 1456 normal bunches of data for 16 conditions, and 1456 RBD-infected bunch data were used in the analysis.

3.2.2. Detecting RBD Bunch Areas and Culm Classification

Data collection was performed from July to August 2021 for training the deep learning model for detecting RBD-infected bunch area and culm classification. Here, images collected from Jeollabuk-do, Gyeongsang-do, and Chungcheong-do Republic of Korea were used. All areas correspond to ordinary paddy fields in the plain area. The total number of images collected for training the deep learning model is shown in Table 3.
The collected 20,514 images contained data making it difficult to identify RBD, such as images taken under strong winds, or blurred focus due to photographer error. In addition, data in which the identification of RBD bunches was ambiguous or infeasible due to various factors, or data that did not include actual RBD bunches were included. As a result of removal such unrecognized images, 10,003 usable images were obtained.
Random cropping was performed from the 10,003 datasets (Figure 13). This cropped image was labeled by both experts and non-experts, and the bounding box labeling was performed via expert inspection for the amount of work of non-experts. Thereafter, dividing the cropped data of 10,003 sheets through a sliding window (Figure 10), 28,365 pieces of RBD-containing cropped patch image data were generated. Cropped patch image data were then used for training and evaluation of the detection model, where 26,812, 769, and 784 pieces of data were used for model training, validation, and evaluation, respectively.
In order to obtain the culm number classification model for RBD, it was necessary to also classify the number of culms in each bounding box (Figure 14). As mentioned in Section 2.3.2, this required expertise, and thus is an intensive task for an expert to process all data. Therefore, labeling was performed by non-agricultural experts, and then reviewed by experts. Table 4 shows the total amount of data collected using this method. In July and August, when disease identification is easier, it is common that the disease does not spread to the entire bunches. Therefore, it can be seen that the data were biased in the range of 1 to 5 culms of infection.

3.3. Quantitative Evaluation Indicators

The evaluations were based on two metrics: mAP and a confusion matrix. mAP is an index used to evaluate the performance of a deep learning object detection model, and is based on average precision. Here, only the evaluation based on I o U 0.5 was used for assessing PASCAL VOC (Equation (5)), where the average precision corresponds to the area under the precision-recall graph, while the recall maintains a regular interval, such as [ 0 , 0.1 , 0.2 , , 1 ] . In Equations (6) and (7), P i n t e r p ( r ) is the maximum precision for recalls > r , P ( r ˜ ) is the precision evaluated at recall r ˜ , and A P is calculated as the average of P i n t e r p ( r ) over all recall ranges:
I o U = a r e a ( g t b o x   t r b o x ) a r e a ( g t b o x   t r b o x ) { I o U T h r e s h o l d ( 0.5 ) : P o s I o U < T h r e s h o l d ( 0.5 ) : N e g
P i n t e r p ( r ) = max r ^ r P ( r ˜ )
A P = 1 11 r { 0 , 0.1 , , 1 } P i n t e r p ( r )

3.4. Drone Shooting Height and Angle Selection

Figure 15 shows the recognition degree obtained by calculating sample data for a total of 16 shooting conditions according to the algorithm shown in Figure 6. It can be seen that the value of one-dimensional LDA reached its maximum of 1.255 at 3 m and 10°, whereas the value of gradient magnitude peaked (1.312) at 3 m and 15°. Lastly, for the one-dimensional LDA and the average of the gradient magnitude values, the recognition degree of RBD reached its maximum (1.23) at 3 m and 0°. Figure 16 shows the results of visualizing the images by projecting the vectors for the image onto a one-dimensional straight line via LDA. It is readily apparent that few data exist that are likely to be confused at 3 m and 0° values near the class boundary.

3.5. Camera Calibration

A large calibration banner measuring 8 × 6 m (width × height) was produced considering that 9 m left and right was imaged at a minimum sensing distance of 6.8 m (Figure 11). To identify the optimal detection position, the drone was fixed on a ladder, and images were taken at 1 m intervals from 6 to 20 m (Figure 17). As shown in Figure 11, the front part of the calibration banner is missing from the field of view at 6 m, whereas the banner is completely within the camera field of view at ≥7 m.
Figure 18A shows the image of the calibration banner at a distance of 7 m from the camera attached to the drone. The red gridded lines and dots in Figure 18A indicate intersections in the calibration banner, and the red line extending vertically is a straight line that crosses the center of the calibration banner in the longitudinal direction, whereas the blue vertical line crosses the center of the image in the longitudinal direction, revealing that the calibration banner was not placed exactly in the center of the image. To compensate for this, perspective transformation was used. In Figure 18A, with horizontal and vertical dimensions of 5472 and 3648, respectively, the upper-left coordinate of the intersection point of the calibration banner is located at (1329, 2874), while those of the upper-right, lower-left, and lower-right coordinates are (3489, 2872), (759, 3416), and (3990, 3386), respectively.
Perspective transformation shifted the ground to the coordinate plane, the horizontal and vertical sizes to 20,000 and 22,000, respectively, the camera coordinates to (7000, 14,000), and the upper-left, upper-right, lower-left, and lower-right coordinates to (4000, 2000), (10,000, 2000), (4000, 6000), and (10,000, 6000) respectively (Figure 18B).
As mentioned, since the center of the calibration banner does not coincide with the center of the image, the calibration area was slightly biased to the left (Figure 18B). To compensate for this, the distorted angle was calculated using the resulting angle from the straight line between the center of the camera and the calibration banner, as well as the camera central axis in the perspective-transformed world coordinate system. The resulting distortion angle calculated in the analysis was 3.77°.
A region calibrated with respect to the central axis of the camera was obtained via rotation transforming the calibration area by a distorted angle with respect to the central axis of the camera, and converting it into a pixel coordinate system using inverse perspective transformation. Figure 18C illustrates the result of correcting the calibration area using rotation transformation in the world coordinate system.
Figure 18D shows the results of image correction based on the information of the calibration banner, and the camera central axis included in the perspective transformation of the pixel coordinate system data belonging to the drone image in the world coordinate system. The resulting coordinates of the calibration areas were upper left (1660, 2858), upper right (3808, 2891), lower left (1127, 3377), and lower right (4338, 3425), while the green rectangle in Figure 18D contains these areas, and can be used as the input for object detection. Here, the blue rectangular areas depict that of one piece of checkerboard used for calibration (equivalent to 1 m2). Therefore, it was possible to precisely calculate the area for 24 m2.

3.6. Detection and Classification of RBD

In order to verify the performance of the proposed system, the details of performance evaluation for each module of RBD bunch area and infected culm number detection were assessed. Culm classification using deep learning model structure included in the system consisted of two steps (Figure 9). Since end-to-end learning could not be performed because it included the step of projecting and extracting information of the detected region to the input image (as described above), different datasets were used for training and evaluating each module.

3.6.1. RBD Bunch Area Detection

The training and evaluation data for the YOLOv3 model for RBD bunch area detection contained a total of 28,365 images, consisting of 26,812 training images, 769 verification images, and 784 evaluation images. During analysis, an augmentation method was applied to avoid over-fitting when training the YOLOv3 model. The weighting and batch size of the DarkNet53 network pre-trained with 9 anchors and ImageNet datasets determined in the experiment were set to 5.
Here, when analyzing the dataset for detecting the area of RBD bunches, there was an imbalance [35,36] between the foreground and background regarding the number of RBD-infected culms and normal rice. To address this problem, a focal loss function [10,37,38] was introduced, and parameter values of α = 0.25 and γ = 2 were applied. In addition, a gradual learning rate reduction was applied using a cosine learning rate decay method [39]. The training parameters of the YOLOv3 model for detecting RBD bunch areas were set as shown in Table 5 by adjusting the batch size, epochs, learning rate, etc., and adding cosine attenuation, in accordance with previous research related to apple detection [40].
The experiments were conducted on the application warm-up techniques and label smoothing [41,42] during the training process using the learning parameters presented in Table 5. For the warmup, issues were addressed regarding the problem of excessive gradient values occurring at the beginning of learning due to the data being biased toward normal data, as well as the issue of excessively increasing confidence during the detection of RBD ambiguous features using label smoothing. As label smoothing was applied, the data were augmented [43,44], and the deleterious effects of data imbalance problem can partially alleviated. The performance comparison evaluation was conducted using the mAP value based on an I o U threshold value of 0.5. As a result of applying warm up and label smoothing in the corresponding experiment, it was confirmed that the performance increased by 9.16% compared to when they were not applied (Table 6). Accordingly, it was presumed that the regularization effect of the data was achieved by suppressing the over-fitting of inaccurate training data.
Figure 19 shows the comparison results of the same YOLOv3 model with and without the warmup and label smoothing techniques applied. Specifically, Figure 19A shows that when detecting RBD bunch area through the learning results without applying warm up and label smoothing techniques, missed detection was more frequent. In contrast, Figure 19B shows relatively more fitting detection results and avoidance of missed detections, as a result of training using warm-up and label-smoothing techniques.
Figure 20 presents the non-maximum suppression result image with a threshold value of 0.45 for the bounding box, pertaining to the bunch area of RBD detected through the YOLOv3 model for cropped patch data. The yellow bounding box indicated the actual RBD bunch areas, whereas the red bounding box indicates the predicted results of the RBD bunch areas. In spite of the trainings applying the data reinforcement mix-up method, the warm-up method, and the label smoothing method, misdetection could be confirmed.

3.6.2. Classification of Culm Number in the Detected RBD Bunch Area

Based on the YOLOv3 model producing 90.49% detection performance for the evaluation data (Figure 9), the RBD bunch areas were detected, and extracted through the bounding box coordinates. After reconstructing the RBD bunch area to a size of 224 × 224 , infected culm number classification was performed through the deep residual network. The number of culm classification learning and evaluation data in the RBD bunch area total 31,826 images. To train and evaluate the deep residual network, labeling was performed according to the number of RBD culms, and consisted of 78% training, 20% validation, and 2% testing. Table 7 shows the dataset for deep residual network training and evaluation.
The classification performance of RBD was evaluated using the relationship between the actual and predicted classes according to the data in Table 4. Table 8 shows the deep residual network performance by network depth for the evaluation data. In general, as the depth of the convolutional neural network increased, the expression level of the filters used varied, providing high performance feature extraction. However, the result of the deep residual network with 50 layers was 80.36%, approximately 9.91% and 4.41% higher than that of the deep residual network with 101 and 152 layers, respectively (Table 8). Accordingly, performance verification was performed on the detection of RBD in the deep residual network with 50 trained layers.
The data in Table 9 confirms that the classification performance with ≥6 RBD-infected culms was ≤70%. Although the focal loss was applied to the difficulty of data augmentation due to the lack of data, a data imbalance between classes was confirmed. Figure 21 shows the results of predicting RBD-infected culm numbers using the learned classification results for 50 layers of the deep residual network.
In the previous experiment, 1–8 RBD-infected culms were included in the bunch area, and produced an accuracy of 80.36%. However, there is notable room for improvement regarding classes with limited data. To confirm this, a classification analysis was conducted using only all classes except 6–8 culms, which maintained the smallest number of classification data, and the results when including only 1–5 infected culms are shown in Table 10.
In order to perform a comparative experiment based on the depth of the deep residual network on the reconstructed RBD culm classification learning and evaluation data, the learning parameters were set and trained according to Table 5. Table 11 shows the performance comparison results for the RBD classification data according to the reconstructed network depth. The performance of the deep residual network with 101 layers was 85.36%, representing increases of approximately 3.64% and 0.92%, when compared with the deep residual networks with 50 and 152 layers, respectively. Therefore, performance verification was conducted for each number of RBD on the learned deep residual network with 101 layers.
The data in Table 12 confirm that the classification performance for all number of culms was ≤80%. Compared with the experiment in which the data were divided into 1-8-culm classes, it can be confirmed that the performance degradation caused by data imbalance between classes is improved. Figure 22 shows the predicted number of RBD-infected culms for 101 layers of the deep residual network of the reconstructed classification data.
Figure 23 compares the classification performance for the number of RBD culms with respect to Table 8 and Table 11. The performances following reorganization of the number of RBD culms from 1 to 5 were improved by 1.91%, 11.29%, and 13.91%. Therefore, it was concluded that it was derived by resolving the issue of imbalance in the amount of data.

4. Conclusions

This study proposed an automatic detection system using drone imagery and deep learning models for the early detection of RBD, and automation of rice field inspections. Prior to training the deep learning model, the LDA and gradient intensity values of the data were collected according to certain rules were analyzed, paying particular attention to the color and morphological characteristics of RBD. Subsequently, at the selected optimal 3-m height and 0° angle, a viewpoint transformation matrix was derived for the correction of camera distortion and area calculation. The YOLOv3 deep learning model was selected as the module for detecting the RBD-infected bunch area, and the training result had a mAP of 90.49. RBD-infected culm number classification was performed using the ResNet 50 layer model, with an accuracy of 80.36% achieved when labeling data for culms 1 to 8 were used. As a result of the experiment by reducing the class range from 1 to 5, the class imbalance problem was somewhat resolved and the accuracy increased to 85.36 in the ResNet 101 layers model. Therefore, it is expected that maintaining only 5 classes will have a stable effect on classification until sufficient data on rice bunches with ≥6 culms are obtained.

Author Contributions

Conceptualization, D.K. and K.P.; methodology, D.K., S.J. (Sunghwan Jeong) and K.P.; software, B.K. and K.P.; validation, B.K., S.-j.K. and K.P.; formal analysis, B.K. and K.P.; investigation, S.-j.K., H.K., S.J. (Sunghwan Jeong), S.J. (Sooho Jeong), K.-Y.K. and G.-y.Y.; resources, H.K., S.J. (Sunghwan Jeong), S.J. (Sooho Jeong), K.-Y.K. and G.-y.Y.; data curation, B.K., S.-j.K. and K.P.; writing—original draft preparation, D.K. and K.P.; writing—review and editing, D.K. and K.P.; visualization, D.K., B.K. and K.P.; supervision, S.J. (Sunghwan Jeong) and K.P.; project administration, S.J. (Sunghwan Jeong) and K.P.; funding acquisition, S.J. (Sunghwan Jeong) and H.K. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Korea Institute of Planning and Evaluation for Technology in Food, Agriculture and Forestry (IPET) and Korea Smart Farm R&D Foundation (KosFarm) through Smart Farm Innovation Technology Development Program, funded by Ministry of Agriculture, Food and Rural Affairs (MAFRA) and Ministry of Science and ICT(MSIT), Rural Development Administration (RDA) (421003-04).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets generated and/or analyzed during the current study are not publicly available due lack of obtaining an online database but are available from the corresponding author on reasonable request.

Conflicts of Interest

The authors declare no conflict of interest. The sponsors had no role in the design, execution, interpretation, or writing of the study.

References

  1. Center for Strategic and International Studies. The Russia-Ukraine War and Global Food Security: A Seven-Week Assessment, and the Way Forward for Policymakers. Available online: https://www.csis.org/analysis/russia-ukraine-war-and-global-food-security-seven-week-assessment-and-way-forward (accessed on 12 September 2022).
  2. Oh, S. Development of rice varieties for processing and trend of food industry. Food Ind. Nut. 2016, 21, 8–14. [Google Scholar]
  3. The JoongAng. ‘Eat Less, but Eat Better’. Corona 19, Rice Selection Criteria have Changed. Available online: https://www.joongang.co.kr/article/24018279#home (accessed on 2 August 2022).
  4. National Legislation Information Center, Ministry of Legislation. Seed Inspection Instructions. Available online: https://www.law.go.kr/%ED%96%89%EC%A0%95%EA%B7%9C%EC%B9%99/%EC%A2%85%EC%9E%90%EA%B2%80%EC%82%AC%EC%9A%94%EB%A0%B9/(2020-1,20200123) (accessed on 2 August 2022).
  5. Liu, Z.; Zhu, Y.; Shi, H.; Qiu, J.; Ding, X.; Kou, Y. Recent progress in rice broad spectrum disease resistance. Int. J. Mol. Sci. 2021, 22, 11658. [Google Scholar] [CrossRef]
  6. Seed Inspection Tips. Korea Law Information Center. Available online: https://www.law.go.kr/%ED%96%89%EC%A0%95%EA%B7%9C%EC%B9%99/%EC%A2%85%EC%9E%90%EA%B2%80%EC%82%AC%EC%9A%94%EB%A0%B9 (accessed on 4 December 2022).
  7. Pacific Pests & Pathogens—Full Size Fact Sheets, Rice Bakanae Disease (429). Available online: https://apps.lucidcentral.org/ppp/text/web_full/entities/rice_bakanae_disease_429.htm (accessed on 12 September 2022).
  8. Chawathe, S.S. Rice disease detection by image analysis. In Proceedings of the 2020 10th Annual Computing and Communication Workshop and Conference (CCWC), Las Vegas, NV, USA, 6–8 January 2022; pp. 524–530. [Google Scholar]
  9. Sulistyaningrum, D.R.; Rasyida, A.; Setiyono, B. Rice disease classification based on leaf image using multilevel support vector machine (SVM). J. Phys. Conf. Ser. 2020, 1490, 012053. [Google Scholar] [CrossRef]
  10. Chung, C.L.; Huang, K.J.; Chen, S.Y.; Lai, M.H.; Chen, Y.C.; Kuo, Y.F. Detecting Bakanae disease in rice seedlings by machine vision. Comput. Electron. Agric. 2016, 121, 404–411. [Google Scholar] [CrossRef]
  11. Kim, S.H.; Oh, S.J.; Yoon, G.Y.; Jung, Y.G.; Kang, M.S. Influence on overfitting and reliability due to change in training data. Int. J. Adv. Cult. Technol. 2017, 5, 82–89. [Google Scholar]
  12. Huang, K.J.; Chen, S.Y.; Chen, Y.C.; Lai, M.H.; Chung, C.L.; Kuo, Y.F. Image-based approach to detect Bakanae disease on rice seedlings. In Proceedings of the 2015 ASABE Annual International Meeting, New Orleans, LA, USA, 26–29 July 2015; American Society of Agricultural and Biological Engineers (ASABE): St. Joseph, MI, USA, 2015. [Google Scholar]
  13. Patrício, D.I.; Rafael., R. Computer Vision and Artificial Intelligence in Precision Agriculture for Grain Crops: A Systematic Review. Comput. Electron. Agric. 2018, 153, 69–81. [Google Scholar] [CrossRef] [Green Version]
  14. Saleem, M.H.; Potgieter, J.; Arif, K.M. Plant disease detection and classification by deep learning. Plants 2019, 8, 468. [Google Scholar] [CrossRef] [Green Version]
  15. Brahimi, M.; Arsenovic, M.; Laraba, S.; Sladojevic, S.; Boukhalfa, K.; Moussaoui, A. Deep learning for plant diseases: Detection and saliency map visualization. In Human and Machine Learning, 1st ed.; Zhou, J., Chen, F., Eds.; Springer: Cham, Switzerland, 2018; pp. 93–117. [Google Scholar]
  16. DeChant, C.; Wiesner-Hanks, T.; Chen, S.; Stewart, E.L.; Yosinski, J.; Gore, M.A.; Nelson, R.J.; Lipson, H. Automated identification of northern leaf blight-infected maize plants from field imagery using deep learning. Phytopathology 2017, 107, 1426–1432. [Google Scholar] [CrossRef] [Green Version]
  17. Kim, B.; Han, Y.K.; Park, J.H.; Lee, J. Improved vision based detection of strawberry diseases using a deep neural network. Front. Plant Sci. 2021, 11, 559172. [Google Scholar] [CrossRef]
  18. Fuentes, A.; Yoon, S.; Kim, S.C.; Park, D.S. A robust deep-learning-based detector for real-time tomato plant diseases and pests recognition. Sensors 2017, 17, 2022. [Google Scholar] [CrossRef] [Green Version]
  19. Park, K.; Choi, K.; Kim, S.; Ahn, H.G.; Jeong, S. Selecting the shooting height and angle method for detection of Bakanae disease using a drone. In Proceedings of the 12th International Conference on Computer Science and its Applications, Jeju, Republic of Korea, 14–16 December 2020. [Google Scholar]
  20. Park, K.; Kim, D.; Choi, K.; Kim, S.; Jeong, S. Method of calculating ground area of drone image using perspective transformation. In Proceedings of the KICS Winter Conference, Pyeongchang, Republic of Korea, 9–11 February 2022. [Google Scholar]
  21. Nongsaro. Pests by Crop—Bakanae Disease. Available online: https://url.kr/jp7n8b (accessed on 2 August 2022).
  22. The PASCAL Visual Object Classes. The PASCAL Visual Object Classes Challenge 2007. Available online: http://host.robots.ox.ac.uk/pascal/VOC/voc2007/ (accessed on 2 August 2022).
  23. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. Nips 2015, 1, 91–99. [Google Scholar] [CrossRef] [Green Version]
  24. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition(CVPR) 2017, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  25. Seferbekov, S.; Iglovikov, V.; Buslaev, A.; Shvets, A. Feature pyramid network for multi-class land segmentation. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–20 June 2018; pp. 272–275. [Google Scholar]
  26. Zhu, L.; Deng, Z.; Hu, X.W.; Fu, C.W.; Xu, X.; Qin, J.; Heng, P.A. Bidirectional feature pyramid network with recurrent attention residual modules for shadow detection. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 121–136. [Google Scholar]
  27. Wu, X.; Zhang, D.; Zhu, J.; Hoi, C.H.S. Single-shot bidirectional pyramid networks for high-quality object detection. Neurocomputing 2020, 401, 1–9. [Google Scholar] [CrossRef]
  28. Redmon, J.; Ali, F. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  29. Redmon, J.; Ali, F. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  30. Tan, C.; Sun, F.; Kong, T.; Zhang, W.; Yang, C.; Liu, C. A survey on deep transfer learning. In Proceedings of the International Conference on Artificial Neural Networks, Rhodes, Greece, 4–7 October 2018; pp. 270–279. [Google Scholar]
  31. Talukdar, J.; Gupta, S.; Rajpura, P.S.; Hegde, R.S. Transfer learning for object detection using state-of-the-art deep neural networks. In Proceedings of the 2018 5th International Conference on Signal Processing and Integrated Networks (SPIN), Noida, India, 22–23 February 2018; pp. 78–83. [Google Scholar]
  32. Singla, A.; Bertino, E.; Verma, D. Overcoming the lack of labeled data: Training intrusion detection models using transfer learning. In Proceedings of the 2019 IEEE International Conference on Smart Computing (SMARTCOMP), Washington, DC, USA, 12–15 June 2019; pp. 69–74. [Google Scholar]
  33. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
  34. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  35. Longadge, R.; Snehalata, D. Class imbalance problem in data mining review. arXiv 2013, arXiv:1305.1707. [Google Scholar]
  36. Thabtah, F.; Hammoud, S.; Kamalov, F.; Gonsalves, A. Data imbalance in classification: Experimental evaluation. Inf. Sci. 2022, 513, 429–441. [Google Scholar] [CrossRef]
  37. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision(ICCV), Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  38. Kiratiratanapruk, K.; Temniranrat, P.; Kitvimonrat, A.; Sinthupinyo, W.; Patarapuwadol, S. Using deep learning techniques to detect rice diseases from images of rice fields. In Proceedings of the International Conference on Industrial, Engineering and Other Applications of Applied Intelligent Systems, Kitakyushu, Japan, 22–25 September 2020; pp. 225–237. [Google Scholar]
  39. He, T.; Zhang, Z.; Zhang, H.; Zhang, Z.; Xie, J.; Li, M. Bag of tricks for image classification with convolutional neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition(CVPR) 2019, Long Beach, CA, USA, 16–20 June 2019; pp. 558–567. [Google Scholar]
  40. Tian, Y.; Yang, G.; Wang, Z.; Wang, H.; Li, E.; Liang, Z. Apple detection during different growth stages in orchards using the improved YOLO-V3 model. Comp. Electron. Agric. 2019, 157, 417–426. [Google Scholar] [CrossRef]
  41. Zhang, Z.; He, T.; Zhang, H.; Zhang, Z.; Xie, J.; Li, M. Bag of freebies for training object detection neural networks. arXiv 2019, arXiv:1902.04103. [Google Scholar]
  42. Michal, L.; Srinadh, B.; Aditya, K.M.; Sanjiv, K. Does label smoothing mitigate label noise? arXiv 2020, arXiv:2003.02819. [Google Scholar]
  43. Zhong, Z.; Zheng, L.; Kang, G.; Li, S.; Yang, Y. Random erasing data augmentation. arXiv 2020, arXiv:1708.04896. [Google Scholar] [CrossRef]
  44. Shorten, C.; Taghi, M.K. A survey on image data augmentation for deep learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
Figure 1. Rice infected with Bakanae disease (left) and normal rice (right).
Figure 1. Rice infected with Bakanae disease (left) and normal rice (right).
Sensors 23 00032 g001
Figure 2. The rice infected with Bakanae disease photographed from above (Red box corresponds to RBD-infected bunch).
Figure 2. The rice infected with Bakanae disease photographed from above (Red box corresponds to RBD-infected bunch).
Sensors 23 00032 g002
Figure 3. The standardized rice transfer interval due to agricultural mechanization.
Figure 3. The standardized rice transfer interval due to agricultural mechanization.
Sensors 23 00032 g003
Figure 4. Example bunch and culm of rice plant.
Figure 4. Example bunch and culm of rice plant.
Sensors 23 00032 g004
Figure 5. The system structure for reading the RBD-infected.
Figure 5. The system structure for reading the RBD-infected.
Sensors 23 00032 g005
Figure 6. The algorithm flowchart for calculating RBD recognition degree.
Figure 6. The algorithm flowchart for calculating RBD recognition degree.
Sensors 23 00032 g006
Figure 7. An example of image cropping RBD and normal rice plants: (a) Images in which the area containing rice infected with Bakanae disease is cropped, and (b) images in which an area containing normal rice is cropped.
Figure 7. An example of image cropping RBD and normal rice plants: (a) Images in which the area containing rice infected with Bakanae disease is cropped, and (b) images in which an area containing normal rice is cropped.
Sensors 23 00032 g007
Figure 8. The edge detection of images containing RBD: (a) Original image, and (b) edge image containing gradient magnitude values.
Figure 8. The edge detection of images containing RBD: (a) Original image, and (b) edge image containing gradient magnitude values.
Sensors 23 00032 g008
Figure 9. The process of determining the number of RBD-infected culms.
Figure 9. The process of determining the number of RBD-infected culms.
Sensors 23 00032 g009
Figure 10. The method of creating patch image data using sliding window.
Figure 10. The method of creating patch image data using sliding window.
Sensors 23 00032 g010
Figure 11. Calculation of the camera field of view for the operating drone.
Figure 11. Calculation of the camera field of view for the operating drone.
Sensors 23 00032 g011
Figure 12. Drone images collected from various height and angles conditions (AD) heights of 1, 2, 3, and 4 m, respectively; (14) 0°, 10°, 20°, and 30° down angles, respectively.
Figure 12. Drone images collected from various height and angles conditions (AD) heights of 1, 2, 3, and 4 m, respectively; (14) 0°, 10°, 20°, and 30° down angles, respectively.
Sensors 23 00032 g012
Figure 13. Random cropping procedure from original image.
Figure 13. Random cropping procedure from original image.
Sensors 23 00032 g013
Figure 14. Example of data by number of culms of RBD (The numbers in the upper left mean the actual number of culm).
Figure 14. Example of data by number of culms of RBD (The numbers in the upper left mean the actual number of culm).
Sensors 23 00032 g014
Figure 15. The rest results of recognition degree for RBD based on the selection of shooting height and angle (The red box indicates the height and angle at which the mean of the gradient magnitude, hue, and saturation LDA is the highest).
Figure 15. The rest results of recognition degree for RBD based on the selection of shooting height and angle (The red box indicates the height and angle at which the mean of the gradient magnitude, hue, and saturation LDA is the highest).
Sensors 23 00032 g015
Figure 16. Visualization of LDA results by shooting conditions (The red box indicates the height and angle at which the mean of the gradient magnitude, hue, and saturation LDA is the highest).
Figure 16. Visualization of LDA results by shooting conditions (The red box indicates the height and angle at which the mean of the gradient magnitude, hue, and saturation LDA is the highest).
Sensors 23 00032 g016
Figure 17. The images taken for camera calibration.
Figure 17. The images taken for camera calibration.
Sensors 23 00032 g017
Figure 18. The perspective transformation for camera calibration: (A) reference lines and points for performing perspective transformations, (B) image projection from pixel coordinate to world coordinate, (C) center compensation based on the camera center point, (D) calculated recognition range using camera calibration.
Figure 18. The perspective transformation for camera calibration: (A) reference lines and points for performing perspective transformations, (B) image projection from pixel coordinate to world coordinate, (C) center compensation based on the camera center point, (D) calculated recognition range using camera calibration.
Sensors 23 00032 g018
Figure 19. The detection results of RBD bunch area for patch segmentation data (Red boxes represent detected RBD-infected bunches): (A) warm up and label smoothing technique not applied, and (B) warmup and label smoothing techniques applied.
Figure 19. The detection results of RBD bunch area for patch segmentation data (Red boxes represent detected RBD-infected bunches): (A) warm up and label smoothing technique not applied, and (B) warmup and label smoothing techniques applied.
Sensors 23 00032 g019
Figure 20. YOLOv3 model-based RBD bunch area detection results (Yellow boxes represent the ground truths, and red boxes represent the detection results): (AC) Each detection task, (1) Ground truth, and (2) predicted result.
Figure 20. YOLOv3 model-based RBD bunch area detection results (Yellow boxes represent the ground truths, and red boxes represent the detection results): (AC) Each detection task, (1) Ground truth, and (2) predicted result.
Sensors 23 00032 g020
Figure 21. The prediction result of RBD for 50 layers of deep residual network.
Figure 21. The prediction result of RBD for 50 layers of deep residual network.
Sensors 23 00032 g021
Figure 22. The prediction result of RBD based on reconstructed number of culms data.
Figure 22. The prediction result of RBD based on reconstructed number of culms data.
Sensors 23 00032 g022
Figure 23. The comparison of classification performance for RBD infected culm data.
Figure 23. The comparison of classification performance for RBD infected culm data.
Sensors 23 00032 g023
Table 1. A comparison of object detection models based on Pascal VOC 2007 dataset.
Table 1. A comparison of object detection models based on Pascal VOC 2007 dataset.
ModelmAPFPS
Faster R-CNN (ResNetV2 101) [23]74.637
FPN (ResNetV2 101) [24,25]75.896
Bi-FPN (ResNetV2 101) [26,27]76.0423
YOLOv2 (DarkNet 53) [28]78.6041
YOLOv3 (DarkNet 53) [29]80.1536
Table 2. The data collected to select the drone’s shooting height and angle.
Table 2. The data collected to select the drone’s shooting height and angle.
Taking Height (m)Taking Downward Angle (°)Total
051015
1126126128126506
2126126128126506
3126124126127503
4126124125124499
Total5045005075032014
Table 3. The number of data collected for the training detection and classification model.
Table 3. The number of data collected for the training detection and classification model.
Collection PeriodAreaNumber of Images
7/2020~8/2020Jeolla-do100
7/2021~8/2021Jeolla-do7548
Gyeongsang-do11,095
Chungcheong-do1771
Total20,514
Table 4. The data for classification of RBD infected culms.
Table 4. The data for classification of RBD infected culms.
Number of RBD Infected CulmsTrainValidationTestTotal
13258838944190
2708718391809106
3765518701769701
4525212651126629
51282326701678
62817120372
7921817127
8162523
Total24,923622967431,826
Table 5. YOLOv3 model training parameters.
Table 5. YOLOv3 model training parameters.
Parameter NameParameter Value
Batch size5
Total epochs120
Batch normalizationTrue
Batch normalization decay0.99
Weight decay5 × 10 4
Learning rate1 × 10 4
Cosine learning rate decayTrue
Non-maximum suppression threshold0.45
Pre-train weightImageNet
Table 6. YOLOv3 model performance results for culm area detection of RBD.
Table 6. YOLOv3 model performance results for culm area detection of RBD.
NetworkWarm UpLabel SmoothingmAP
TrainValidationTest
DarkNet53FalseFalse95.09%80.75%81.33%
FalseTrue95.28%91.36%89.22%
TrueFalse91.64%89.95%88.13%
TrueTrue97.84%92.68%90.49%
Table 7. The deep residual network training parameters.
Table 7. The deep residual network training parameters.
ParameterMethod
Image size224 × 224 × 3
Batch size16
Total epochs300
Learning rate0.001
Weight decay0.00004
Batch normalizationTrue
OptimizerAdam
Loss functionFocal loss ( α   = 4.0, γ = 2.0)
Weight initializationXavier initialization
Pre-train weightImageNet
NetworkResNetV2(50/101/152)
Table 8. The comparison of culm classification performance by deep residual network depth.
Table 8. The comparison of culm classification performance by deep residual network depth.
Depth50 Layers101 Layers152 Layers
Accuracy80.36%70.45%75.95%
Table 9. The number of culm classification performance results for 50 layers of deep residual network.
Table 9. The number of culm classification performance results for 50 layers of deep residual network.
Number of RBD Infected CulmsTestTrue (Pred)False (Pred)Accuracy
19489594.68%
21801681293.33%
31761542287.50%
4112991388.39%
570591184.28%
62014670.00%
71711664.70%
853260.00%
Total6745977780.36%
Table 10. The reconstructed number of RBD infected culm classification training and evaluation datasets.
Table 10. The reconstructed number of RBD infected culm classification training and evaluation datasets.
Number of RBD Infected CulmsOriginalTrainValidationTest
14190325883894
2910670871839180
3970176551870176
4662952521265112
51678128232670
Total31,30424,5346139632
Table 11. The comparison of culm classification performance by deep residual network depth.
Table 11. The comparison of culm classification performance by deep residual network depth.
Depth50 Layers101 Layers152 Layers
Accuracy81.72%85.36%84.44%
Table 12. Culm number classification performance results for deep residual network 101.
Table 12. Culm number classification performance results for deep residual network 101.
Number of RBD Infected CulmsTestTrue (Pred)False (Pred)Accuracy
19485990.42%
21801522884.44%
31761532386.93%
4112922082.14%
570581282.85%
Total6325409285.36%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kim, D.; Jeong, S.; Kim, B.; Kim, S.-j.; Kim, H.; Jeong, S.; Yun, G.-y.; Kim, K.-Y.; Park, K. Automated Detection of Rice Bakanae Disease via Drone Imagery. Sensors 2023, 23, 32. https://doi.org/10.3390/s23010032

AMA Style

Kim D, Jeong S, Kim B, Kim S-j, Kim H, Jeong S, Yun G-y, Kim K-Y, Park K. Automated Detection of Rice Bakanae Disease via Drone Imagery. Sensors. 2023; 23(1):32. https://doi.org/10.3390/s23010032

Chicago/Turabian Style

Kim, Donghoon, Sunghwan Jeong, Byoungjun Kim, Seo-jeong Kim, Heegon Kim, Sooho Jeong, Ga-yun Yun, Kee-Yeun Kim, and Keunho Park. 2023. "Automated Detection of Rice Bakanae Disease via Drone Imagery" Sensors 23, no. 1: 32. https://doi.org/10.3390/s23010032

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop