Next Article in Journal
A Moderate Wetting and Drying Regime Combined with Appropriate Nitrogen Application Increases Grain Yield and Nitrogen Use Efficiency in Rice
Previous Article in Journal
Melatonin Mitigated Salinity Stress on Alfalfa by Improving Antioxidant Defense and Osmoregulation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Plot-Level Maize Early Stage Stand Counting and Spacing Detection Using Advanced Deep Learning Algorithms Based on UAV Imagery

1
Biological System Engineering, University of Wisconsin-Madison, Madison, WI 53706, USA
2
Department of Agronomy, University of Wisconsin-Madison, Madison, WI 53706, USA
*
Author to whom correspondence should be addressed.
Agronomy 2023, 13(7), 1728; https://doi.org/10.3390/agronomy13071728
Submission received: 25 May 2023 / Revised: 22 June 2023 / Accepted: 25 June 2023 / Published: 27 June 2023
(This article belongs to the Section Precision and Digital Agriculture)

Abstract

:
Phenotyping is one of the most important processes in modern breeding, especially for maize, which is an important crop for food, feeds, and industrial uses. Breeders invest considerable time in identifying genotypes with high productivity and stress tolerance. Plant spacing plays a critical role in determining the yield of crops in production settings to provide useful management information. In this study, we propose an automated solution using unmanned aerial vehicle (UAV) imagery and deep learning algorithms to provide accurate stand counting and plant-level spacing variabilities (PSV) in order to facilitate the breeders’ decision making. A high-resolution UAV was used to train three deep learning models, namely, YOLOv5, YOLOX, and YOLOR, for both maize stand counting and PSV detection. The results indicate that after optimizing the non-maximum suppression (NMS) intersection of union (IoU) threshold, YOLOv5 obtained the best stand counting accuracy, with a coefficient of determination (R2) of 0.936 and mean absolute error (MAE) of 1.958. Furthermore, the YOLOX model subsequently achieved an F1-score value of 0.896 for PSV detection. This study shows the promising accuracy and reliability of processed UAV imagery for automating stand counting and spacing evaluation and its potential to be implemented further into real-time breeding decision making.

1. Introduction

Crop breeding is a critical process underlying global food security and raw material production [1]. Data of high quality and accuracy produced in breeding field trials are a fundamental component of breeding success. Variation is regularly observed on the field, region, and country levels due to genetic variation, environmental effects, and the differential responses of genotypes grown in multiple environments on the expression of a given phenotype (e.g., grain yield), known as genotype by environment (G × E) interaction [2]. Non-genetic error variation and G × E impact almost every aspect of the decision-making process involved in plant-breeding programs [3]. Therefore, accounting for and modeling variation are important for improving the breeding process [4].
Maize (Zea mays ssp. Mays) is a major crop grown in the United States and one of the most widely consumed grains in the world. Since the 1930s, the combination of improved genetics and production practices has resulted in consistent increases in yield per acre, with increments in grain yields of around 2.5 bu/ac per year [5]. The annual average maize grain yield fluctuates due to yearly weather patterns, resulting in variable impacts on the economy and global food security. It has been reported that these fluctuations are mainly due to variability in growing conditions from year to year [6,7].
Crops respond differently to environmental factors, including agronomic management practices. For maize, both planting density and plant-level spacing variabilities (PSV) have significant effects on grain yield [8,9], and different hybrids have different optimal densities required to achieve the maximum yield. A reduction in the number of plants or an increase in the variability of plant spacing will result in the decline in the grain yield achieved per unit of planted land [10,11]. Unlike the grain yield, which is determined by harvesting an entire plot with a plot combine, the number of plants per plot and the variability in space in each breeding row can only be obtained through visual inspections at the plot level. This activity is time-consuming and labor-intensive. Furthermore, while the stand count is recorded more regularly, PSV is a much more difficult trait to quantify, as it requires measuring the distance between every pair of plants in a row [9] and is thus very rarely measured or reported. PSV is also an important source of error persisting in the analysis of maize trials, as statistical models that account for patterns of variation across fields cannot account for plot-specific factors such as PSV.
Unmanned aerial vehicles (UAVs) have vast potential for applications in agriculture and provide an opportunity to support a high-throughput method with which to detect plant stands [12]. UAV-based sensing platforms cover large areas in a short amount of time compared with ground vehicles or robots. They can provide data with higher spatial and temporal resolutions than satellite imagery [13]. With a superior data collection speed and resolution, UAVs have become increasingly important tools used to acquire high-resolution RGB (visible light spectrum) [14], multispectral (specific wavelength ranges across the spectrum) [15], and hyperspectral (light emitted outside of the visible spectrum) [16,17] imagery for plant phenotyping in the digital agriculture field.
To date, there have been successful applications combining UAV imagery and deep learning in precision agriculture, such as estimations of crop counts and estimations of plant density in field conditions. Kuznetsova et al. [18] compared the performance of You Only Look Once (YOLO) v3 and v5 in orchard apple detection. In the experiment, YOLOv5 outperformed YOLOv3 with a false negative rate (FNR) of 2.8% and false positive Rate of 3.5% (FPR). Ong et al. [19] used RGB UAV imagery and a convolutional neural network (CNN) to detect weeds in Chinese cabbage, and the overall accuracy was 92.41%. Fend et al. [20] developed a nearly real-time UAV image-processing algorithm based on a convolutional neural network (CNN) and obtained encouraging results for cotton stand counting (R2 = 0.95). Lin et al. [21] compared the MobileNet and CenterNet algorithms and concluded that CenterNet has a better performance in cotton plant counting. Ranđelović et al. [22] applied RGB UAV imagery and a machine learning model to predict soybean density and achieved promising results, with a correlation coefficient equal to 0.87. Mhango et al. applied a faster region-based convolutional neural network (FRCNN) to estimate potato densities using UAV imagery. The estimated results had high correlation with the manually determined ground truth (R2 = 0.80) [23]. To the best of our knowledge, previous studies mostly focused on applying UAV imagery and deep learning techniques in calculating either the stand count or plant density but not PSV.
To build an object detection model that is efficient and reliable in detecting both maize stands and PSV, we compared the performance of two YOLO series models, namely, YOLOv5 and YOLOX, and one unified network for multiple tasks, called YOLOR (You Only Learn One Representation) [24,25]. The YOLO model is the best-known and most popular one-stage object detection algorithm due to its small model size and fast calculation speed [26]. It was first introduced by Redmon et al. in 2015 [27]. Enhanced through its generations, YOLOv5 has more flexible control of the model size without a loss of accuracy. The YOLOX model is built on the YOLOv3 baseline with a DarkNet53 model [28] as the backbone and has been shown to have a better detection accuracy [24]. The YOLOR was designed to mimic human learning; it can learn from both features and implicit knowledge and has demonstrated promising object detection results. Therefore, this study aimed to evaluate the performance of the three different deep learning object detection models and optimize their thresholds in estimating stand counting for maize-breeding plots and quantifying the PSV.

2. Materials and Methods

2.1. Experimental Sites and Design

The experiments were conducted at the West Madison Agricultural Research Station (WMARS) in Verona, Wisconsin, USA (43°03’37” N 89°31’54” W), and the Arlington Agricultural Research Station (AARS) in Arlington, Wisconsin, USA (43°18’13” N 89°23’04” W). Five maize hybrids with good germination rates were selected as the experimental materials, among which three were from the Field Corn Breeding Program and two from the private sector (Dekalb: DKC46-60 and DKC48-12RIB). The five hybrids are in the 96–102 relative maturity group. The experimental materials were planted following a split-plot design on 11 May 2021 at the WMARS and 18 May 2021 at the AARS, respectively. Each plot was 6.7 m in length and 0.97 m in row spacing. At each of the sites, the plots received one of the five plot-level spacing treatments (Figure 1), and the subplots accommodated the five maize hybrids with five replicates, leading to 125 plots in total.
Each plot was planted as a two-row plot with 40 seeds evenly distributed every 0.15 m at each row, resulting in approximately 80 seeds in each plot. Thirteen days after planting (DAP), the maize seedlings at the West Madison site were manually removed to create varied spacing distances between pairs so as to form the four plot-level treatment conditions shown in Figure 1. The plot-level treatment conditions 1 to 4, shown in Figure 1, were Perfect, Not Enough Seeds (NES), Skips, and Big Gaps. The plot-level treatment condition 5 (Misplacements) was manual planting with two plants next to each other, which increased the planting density by 10%. The average seedling numbers of all the plots at this site were 65, 63, 64, 80, and 94 for Skips, Not Enough Seeds, Big Gaps, Perfect Spacings, and Misplacements, respectively. In Arlington, the plots were planted in the same way as above but did not receive treatments for NES, Skips, or Big Gaps. Only the PSV for Perfect and Misplacements is counted here, resulting in 100 and 25 plots for the conditions of Perfect Spacing and Misplacements, respectively.

2.2. UAV Image Acquisition

Aerial image data were collected using a Phantom 4 Pro V2.0 (DJI, Shenzhen, Guangdong, China) with an onboard RGB camera. The flight camera had a resolution of 4864 × 3648 pixels and was set to take timelapse photos at 2 frames per second. The drone was set to fly at 12 m above the ground with 80% image overlapping, and the ground sampling distance (GSD) was 0.37 cm. The flight plan was pre-defined in a flight control app called FieldAgent (Sentera, Saint Paul, MN, USA).
The aerial images were taken on 3 June 2021, at both sites, when the maize plants developed the third leaf collar (roughly in the V3 stage). The leaves could be clearly seen in the images. Plots with a plant density not meeting our experiment design were discarded, leading to a total of 249 images for validating the deep learning methods.

2.3. Image Processing and Labeling

Each maize plot was visually identified and cropped from the center of the raw RGB image that was taken directly above the plot. Thus, no orientated adjustment was needed for the plot image. The image dimensions for individual plots were, on average, 1900 × 400 pixels. A free and open-source image-labeling package, LabelImg [29], was used to manually label the maize stands and PSV for individual plots. All targets (either maize stands or PSV) were visually detected and labeled by drawing rectangular bounding boxes around the edges of the targets. In stand counting, each bounding box corresponds to an individual maize plant. If the plants overlap with each other, the bounding boxes will also overlap. For PSV, each individual box represents a specific spacing variation. A single bounding box for the Misplacements contains all the maize plants in a particular location, while Skips and Big Gaps include plants both before and after the spacing, along with the gap itself. In the case of Big Gaps, the spacing length typically accommodates more than two plants, whereas the Skips involve only a single plant. Two lists each containing the pixel positions of the four vertices of the labeled bounding boxes were saved as separate Extensible Markup Language (.xml) files for the maize stands and PSV in each plot. Figure 2a shows a single labeled maize stand in an example maize plot, and Figure 2b shows its labeled PSV spots.

2.4. Model Development

2.4.1. Detecting and Counting Maize Stands

Three models were developed for detecting maize stands, namely, YOLOv5, YOLOX, and YOLOR. The YOLOv5 model has a similar structure compared to the previous YOLO–YOLOv3 that uses Mosaic data augmentation for the input data. However, the addition of an auto-anchor check in YOLOv5 enables YOLOv5 to converge faster than the previous YOLO series and maintain high accuracy.
Based on the modified version of YOLOv3 with Darkent-53 as its backbone, YOLOX is an anchor-free object detection algorithm with a decoupled head to improve object detection accuracy. This mechanism ensures that YOLOX has a better performance compared with the other YOLO series, which use coupled detection heads. Similarly, the YOLOX model was built on the version with minimum parameters.
Though it has a similar name, YOLOR does not adopt any of the architectures in the YOLO series. YOLOR is a unified network used to encode implicit knowledge and explicit knowledge together. Given the small sample size of our dataset, all three models were configured with the version of the smallest model size, including YOLOv5s, YOLOX-s, and YOLOR-P6, to prevent potential overfitting.
The models were developed and trained using Google Colab. The images of 249 maize plots at the two sites and their bounding boxes for the maize stands were randomly split into training, testing, and validation sets with proportions of 7:2:1. In order to enhance the training process, all three models were trained using transfer learning, leveraging the pre-trained weights from the MS COCO dataset [24,25,30]. During training, the dataset was divided into mini-batches with a batch size of 16. An epoch was counted when all the minibatches were fed to the model, and three epoch numbers (500, 1000, 1500) were used for comparing the training and testing performances. As increasing the epoch may yield a better accuracy but can also lead to overfitting, it is important to choose the best weights during the whole training process. Thus, the best weights will be saved and used to evaluate the accuracy of each model and further compare their performances.
All other parameters, such as the learning rate, were kept with the original settings that each algorithm was designed with, which are shown in Table 1.

2.4.2. Detecting the PSV

The three model architectures introduced in Section 2.4 were also used for detecting the PSV. Instead of taking the positions of the maize stands as training responses, the models for different PSV were trained to output the locations of the spacing variation spots (plant pairs with abnormal distances) and of the categories (Big Gaps, Skips, and Misplacements) for each spot. Compared with the stand counting, the data only focused on PSV, while Perfect and NES were not counted in the dataset, so that the size of the training and testing data was smaller. Thus, the training epochs needed to be decreased, and each of the models was trained with 200, 500, 800, and 1000 epochs.
The workflow of both detecting maize stands and the PSV is shown in Figure 3.

2.5. Evaluation of Model Performance

The performance of the developed models in detecting the maize stands was evaluated based on the classification confusion matrix between the model outputs and the manual labels of the testing dataset [31]. Precision, recall, and F1-scores are the indicators most widely used to summarize object detection performance, using Equations (1)–(3). Equation (4) computes the average precision by taking the area under the precision–recall curve p(r). The precision and recall values of the model are plotted as a function of the model to obtain the p(r) curve.
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
A P = r = 0 1 p r d r
The mean absolute error (MAE) and coefficient of determination (R2) between the model and manual counts were used as metrics for evaluating the performance in counting the maize stands from images, computed as shown in Equations (5) and (6), where TP, FP, and FN stand for true positive, false positive, and false negative, respectively. They were obtained from the confusion matrices between true classes and model outputs.
M A E = 1 n 1 n y i y ^ i
R 2 = 1 1 n y i y ^ i 2 1 n y i y ¯ i 2
where n is the number of images, and yi and y ^ i are the manual and model counts (i.e., number of bounding boxes) of the ith image, respectively. y ¯ i is the mean manual count of the images. Pearson’s correlation coefficient was used to further investigate the linear relationship between the maize stand counts and ground truth under the Misplacement treatment using Equation (7).
ρ x y = C o v x , y σ x σ y
Similarly, the performance in detecting the PSV was evaluated using the precision, recall, and F1-scores of each algorithm. F1-score is sufficient for performance evaluation due to the lack of overlapping bounding boxes for each variability in PSV detection.

3. Results

3.1. Detecting and Counting Maize Stands from UAV-Based Imagery

The performance of the three models in detecting the maize stands from the UAV canopy images is shown in Table 2. The best AP were 0.931, 0.898, and 0.920 for YOLOv5, YOLOX, and YOLOR. These findings align with a previous study that achieved a mean average precision (mAP) of 86% by employing YOLOv3 and training it with 200 annotated images for estimating cotton stand count using UAS (unmanned aerial system) images [32].
Regarding the performance in counting the maize stands, the YOLOX model under 500 training epochs achieved the smallest MAE = 3.542, resulting from the strong agreement between the image-based and manual counts, especially for the groups of Skips, NES, Big Gaps, and Perfect Spacings. The errors of YOLOX were mainly caused by the underestimated stand counts of the Misplacement group. The underestimation of overlapping plants is highlighted with orange dots in Figure 4 for YOLOX, deviating from the red line. This issue is due to the large proportion of overlapping areas between the maize stands in the Misplacement group when the data were collected. It is also notable that the further training of the model (under 1000 and 1500 epochs) could not aid in reducing the underestimation of overlapping plants caused by Misplacement.
Table 2 reveals inconsistencies among the three evaluation criteria: AP, R2, and MAE. YOLOv5 has the highest AP, while YOLOR performs best in terms of R2. It is important to note that MAE has the nature of measuring the average error, in contrast to the others [33]. R2 may negatively impact YOLOX due to the underestimated Misplacement group. AP only emphasizes the accuracy in the detection of the bounding boxes and fails to reflect the counting accuracy. Hence, MAE serves as the most appropriate evaluation metric here.
Compared with YOLOX, both the YOLOv5 and YOLOR outputs mostly agreed with those in the Misplacement group’s detection. However, clear offsets were observed between the image-based and manual counts for the other four groups. These overestimations were caused by the fact that the threshold setting of the intersection over union (IoU) of non-maximum suppression (NMS) for determining the detection bounding boxes as positive or negative outputs for each stand (Section 3.2) was not optimized. Therefore, it was observed that the YOLOX model was more suitable for counting maize stands under the Perfect, Skips, NES or Big Gaps conditions, where the plants were not crowded and did not severely overlap with each other.
Through visual observation of the testing dataset for all 48 images, it was found that both YOLOv5 and YOLOR produced redundant bounding boxes under the Perfect, Skips, NES, and Big Gaps spacing treatments compared with YOLOX. However, YOLOX detected a value less than the true value under the Misplacement treatment. Figure 5 shows the original plot-level image and the maize stand counting demo for the three deep learning models under the Misplacement spacing treatment. From the demo, we can see that both YOLOv5 and YOLOR have overlapping bounding boxes that constitute multiple entries at the same location, whereas YOLOX tends to detect the plant without overlapping. Compared with Figure 6, which presents another spacing treatment, namely, Skips, both YOLOv5 and YOLOR count 84 maize stands, as compared with ground truth 76. Both YOLOv5 and YOLOR have some redundant counts because both models output more than one bounding box for a single maize stand, and the defined IoU threshold was not able to differentiate between false and true positive boxes. Consequently, though the average precision (AP) of YOLOv5 and YOLOR were higher than that of YOLOX, the false positive boxes caused a higher MAE, as shown in Table 1.

3.2. Determining the Optimal NMS IoU Threshold

In the previous sections, we demonstrated that both the YOLOv5 and YOLOR models overestimated the samples from the Perfect, Skips, NES, and Big Gaps groups, in which the maize plants were not crowded or severely overlapping with each other. The YOLOX performed well for these four groups but showed underestimation for the Misplacement group. This was due to varied ability in eliminating redundant detection boxes among the three models in different PSV scenarios, which was controlled using the NMS IoU threshold for the three models. The NMS IoU is a method commonly used to select single entity (i.e., the bounding boxes in this case) out of many overlapping entities by discarding those that are below a given thresholding parameter. The thresholding parameter is calculated as the IoU between the pairs of detected bounding boxes.
The PSV treatments with varied overlapping conditions might favor different IoU values. The Misplacement group underestimated by YOLOX might benefit from a looser threshold (a higher IoU) to retain the bounding boxes that were previously removed, while the groups overestimated by YOLOv5 and YOLOR might favor a more restrictive threshold (lower IoU values) for eliminating redundancies. Therefore, the determination of the optimal NMS IoU value in each scenario for the models plays a key role in improving the overall counting accuracy.
The results in the previous section were based on the default NMS IoU thresholds of 0.5, 0.5, and 0.3 for YOLOv5, YOLOR, and YOLOX, respectively. Nonetheless, as a result of the unique treatment of spacing variabilities in the dataset. Misplacement led to the presence of some bounding boxes that overlap with one another more than 90% in the training dataset. Both YOLOv5 and YOLOR learned this special feature and tended to detect more stand counts under the other four PSV. Thus, maize stand detection under these special spacing variabilities required further evaluations of the NMS IoU thresholds. Table 3 shows the initial evaluation of the performance when changing the NMS thresholds from 0.1 to 0.5.
In general, the average precision values for all three models did not differ significantly across the different NMS IoU thresholds. However, there is still an improvement in the AP results when comparing the numbers in Table 3 with those in Table 2. Since the main purpose of optimizing the NMS IoU threshold was to help the models to have the best performance under both the special spacing treatment, Misplacement, and the other four groups (Perfect, Skips, NES, and Big Gaps), instead of demonstrating the models’ performance based on all the spacing variabilities, Misplacement was separated from the other four groups. Table 4 shows the performance for the Misplacement group and the other four groups under changing NMS IoU thresholds. R2 is not used here to evaluate the models' performance, since the selection of the sample size can significantly impact the anticipated R2. In this case, Misplacement only has 10 samples. With small numbers of samples, the mean of the coefficient of determination may exhibit a high value, even if there is no correlation between the variables [34]. Here, the MAE is a more meaningful metric with which to assess accuracy, as it solely quantifies the differences in quality between the predicted stand count value and the ground truth. Pearson’s correlation is used here to demonstrate the linear correlation between the updated stand counts and the ground truth.
Table 4 reveals that the models’ accuracy is influenced by the NMS IoU, as determined by the MAE. For the Misplacement PSV, both YOLOv5 and YOLOR display an optimal performance when the threshold is set to 0.5, yielding MAE values of 4.9 and 3.9, respectively. As described in Section 3.1, YOLOX tends to underestimate the maize stand counts but can achieve better results when the NMS IoU is set to 0.6, with an MAE equal to 4.9. These outcomes indicate that the NMS IoU threshold must be optimized and can have different values under different PSV. For example, while YOLOv5 achieves optimal results for Misplacement when the threshold is set to 0.5, it shows diminished accuracy for the other four groups. Similarly, with different optimized NMS IoU thresholds under various PSV, the detection results can be enhanced, as shown for YOLOX and YOLOR. The correlation coefficient for all three models across various thresholds also demonstrates a linear correlation between the detected maize stand counts and the ground truth.
Table 5 shows the re-evaluation results after choosing the optimized NMS IoU threshold for the PSV of Misplacement and the other four groups. The R2 and MAE are improved dramatically compared with the results shown in Section 3.1, with all three models having an R2 above 0.9 and an MAE of approximately 2.0. In Figure 7, the Misplacement (orange dots) for YOLOX shift towards the ground truth. Additionally, for both YOLOv5 and YOLOR, Figure 7 shows that the problem of overestimating for the other four groups can be resolved by choosing an optimized NMS IoU threshold, as compared with the results in Figure 4. The experiment results indicate that after carefully choosing the optimized NMS IoU threshold, all three models could yield promising results under different PSVs.

3.3. Detecting and Visualizing the PSV

The performance in the detection of PSV from UAV images on the plant level is shown in Table 6. Overall, all three models displayed a promising performance, with all showing an F1 above 0.80. YOLOX under 200 training epochs achieved the highest accuracy, with a precision of 0.941 and F1-score of 0.896, followed by YOLOv5 under 200 training epochs. For both YOLOv5 and YOLOX, increasing the number of training epochs did not increase the F1-score, which was caused by the fact that the size of the training dataset was even smaller than that for maize stand counting. However, YOLOR under 800 training epochs reached the best performance, with an F1-score equal to 0.878, which may mean that YOLOR needs more training epochs to learn the features from the training data.
It is difficult to determine the best model by only evaluating the F1-scores in the table, since all the F1-scores are close to each other. This is mainly caused by the fact that when detecting the PSV, one will only locate the variabilities in a single bounding box, instead of multiple bounding boxes as in the previous maize stand counting. The examples in Figure 8 demonstrate the performance of the three models in detecting PSV. In general, all three models could successfully detect multiple spacing variabilities (indicated by the color and label of each bounding box) in the maize plots. The detected positions and corresponding categories closely matched the manual labels. However, it can be noted from Figure 8 that YOLOv5 and YOLOR tended to detect empty spaces at the end of each plot as Big Gaps, while YOLOX could avoid this issue.
From the visualization results, we can consider the developed method to be promising in detecting PSV in maize plots for practical use.

4. Discussion

It was proven in previous studies that crop stand counting could be achieved using deep learning object detection algorithms [21,35,36]. In the case of our study, YOLOX initially achieved the lowest MAE among all the advanced deep learning models that were used, the others being YOLOv5 and YOLOR, for plot-level maize stand counting. After optimizing the NMS IoU threshold, all three models demonstrated promising results, with an R2 above 0.9 and MAE of around 2. YOLOX outperformed YOLOv5 and YOLOR in the PSV detection applications.
In the field of object detection, newly developed algorithms such as the ones used in our study have been shown to have much more accurate results than the traditionally adopted methods, such as Faster-RCNN or YOLOv3. For traditional object detection, YOLOX-s [24] outperforms YOLOv5-s with a 2.9% AP, and YOLOR [25] proves that a single model architecture is still effective for multi-task learning. Despite their seemingly better performance in traditional object detection, there are only a few studies showing that these methods have the potential to be used in the field of precision agriculture or related areas. Zhang et al. [37] adopted YOLOX for the fruit-counting area and achieved a value of 99.519% with respect to the best detection rate. Additionally, Song et al. [38] used the improved YOLOX-tiny for tree height estimation based on fisheye images, and the highest relative error of the tree measurements was 4.06%. Zhong et al. [39] successfully adopted YOLOR to detect real-time marine animals in a coral reef ecosystem, and it achieved an AP above 0.79 in both the fish and turtle categories.
Plant spacing is an important topic in maize production, and it has been reported that with well-tuned planters and uniform plant spacing, the yields could be 20% greater than those of fields with PSV [9]. Wang et al. [40] presented a vision-based method to measure maize plant spacing and population in the early growth stage. Tang et al. [41] proposed different methods to detect and measure the spacing between corn plant stem centers through mosaic images. Nevertheless, these works focused on normal spacing situations and the distance between each plant, which requires the information regarding the center of the maize to be known before the measurements. This can be problematic, since this information can sometimes be very vague in the case of special spacings and can introduce additional uncertainties. Additionally, the plant distance for certain types of plant spacings can vary considerably. Thus, our experiment not only proved that YOLOv5, YOLOX, and YOLOR can be applied in small crop object detection and have a better accuracy than the previous advanced models but also demonstrated that the different successfully detected PSV can be used for further analysis.
However, some limitations should be noted. First, the flying height is an important factor that will influence the detection accuracy, since the object of interest is relatively small [42]. This presents a challenging problem, as small objects may lack the information required to distinguish them from background objects [43,44] such as weeds and soil. The ideal strategy for maximizing the information collected requires the UAVs to fly at a relatively low altitude. The tradeoff between the flying height and image quality is that drones can cover a larger area by flying at higher altitudes or a smaller area at a lower altitude. Additionally, most deep learning algorithms require a large training dataset, and the current traditional object detection datasets, such as COCO [45], contain 328 k images. Additionally, researchers [46] showed that a small training dataset may cause an unstable deep learning algorithm performance. Based on our experiment results, when the training epoch increased, the results were not guaranteed to be improved.
In future studies, platforms other than UAVs could be applied to stand counting in order to acquire higher-resolution imagery, as in the study of Wang et al. [36], who mounted a camera on a cart and collected video sequences. This study showed a promising stand counting result with an accuracy over 98%. Flying height derivation may influence the results of deep learning because the imagery may contain different information. Thus, different flying heights should be accounted for in further analyses. Since this experiment only focused on the V3 stage, it was possible to acquire the earlier stage of maize growth in order to avoid leaf overlapping. Additionally, data augmentation is another tool that could be considered as a solution to enlarge the number of training datasets in order to prevent model overfitting due to limited data, as stated in [47]. Data augmentation could improve deep learning accuracy in traditional object detection tasks. Therefore, more studies need to be conducted to examine the height influence in order to find the optimal flying altitude as well as the optimal size of the dataset and thus obtain a stable and, therefore, more reliable result.

5. Conclusions

In this study, three deep learning models, YOLOv5, YOLOX, and YOLOR, were applied to detect and count maize and spacing variabilities in the V3 stage using UAV imagery. Overall, all three deep learning models were able to conduct plot-level stand counting and determine PSV at a satisfactory level. Under different spacing variabilities, the NMS IoU threshold is an important parameter that needs to be optimized in order to obtain a more robust model. For the Misplacement, the NMS IoU should be set to around 0.5 to 0.6, compared with the other PSV, which can have a lower threshold. After adjusting the optimal NMS IoU threshold, the YOLOv5 model had a better overall performance for the different spacing treatments, indicated by its better value of R2 = 0.936 and lowest MAE value = 1.958. All three algorithms achieved closely satisfied results in detecting PSV, with the best F1-score above 0.85. When detecting PSV through visual examination, YOLOX outperformed the other models, with an F1-score = 0.896. In our experiment, we proved that deep learning could potentially be used to accurately and efficiently detect maize stands and PSV from UAVs imagery. The detection outcomes could be used in breeding experiments as phenotypic information in future analyses for genotype–phenotype mapping, understanding genotype by environment interaction, or breeding selection. The proposed method could also be adopted in crop production settings, providing a reference to maize growers for precision management activities.

Author Contributions

Conceptualization, B.W., J.Z. and Z.Z.; data collection B.W. and M.C.; methodology, B.W., J.Z. and Z.Z.; validation, B.W., J.Z. and Z.Z.; data curation, B.W. and M.C.; writing—original draft, B.W.; writing—review and editing, B.W., J.Z., Z.Z., M.C. and S.M.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the USDA National Institute of Food and Agriculture Hatch project, accession No. 7002632; and the USDA National Institute of Food and Agriculture, AFRI project accession No. 1028196.

Data Availability Statement

The data and programming codes are freely available upon request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lemke, S.; Tao, X.; Kushner, G.J. Assuring the Food Safety of Crops Developed through Breeding. ACS Agric. Sci. Technol. 2022, 2, 1151–1165. [Google Scholar] [CrossRef]
  2. Hudson, A.I.; Odell, S.G.; Dubreuil, P.; Tixier, M.H.; Praud, S.; Runcie, D.E.; Ross-Ibarra, J. Analysis of genotype-by-environment interactions in a maize mapping population. G3 Genes Genomes Genet. 2022, 12, jkac013. [Google Scholar] [CrossRef] [PubMed]
  3. NLeon, D.; Jannink, J.L.; Edwards, J.W.; Kaeppler, S.M. Introduction to a special issue on genotype by environment interaction. Crop Sci. 2016, 56, 2081–2089. [Google Scholar] [CrossRef] [Green Version]
  4. Elias, A.A.; Robbins, K.R.; Doerge, R.W.; Tuinstra, M.R. Half a century of studying genotype × Environment interactions in plant breeding experiments. Crop Sci. 2016, 56, 2090–2105. [Google Scholar] [CrossRef]
  5. USDA, N.A.S.S. National Statistics for Corn Grain Yield Yearly. Available online: https://quickstats.nass.usda.gov/results/FFD23882-EACC-32EB-BDB6-7F4F4748BE6B?pivot=short_desc (accessed on 12 January 2021).
  6. Ckucharik, J.; Ramankutty, N. Trends and Variability in U.S. Corn Yields over the Twentieth Century. 2005. Available online: http://EarthInteractions.org (accessed on 12 January 2021).
  7. Teasdale, J.R.; Cavigelli, M.A. Meteorological fluctuations define long-Term crop yield patterns in conventional and organic production systems. Sci. Rep. 2017, 7, 688. [Google Scholar] [CrossRef] [Green Version]
  8. Assefa, Y.; Vara Prasad, P.V.; Carter, P.; Hinds, M.; Bhalla, G.; Schon, R.; Jeschke, M.; Paszkiewicz, S.; Ciampitti, I.A. Yield responses to planting density for US modern corn hybrids: A synthesis-analysis. Crop Sci. 2016, 56, 2802–2817. [Google Scholar] [CrossRef] [Green Version]
  9. Lauer, J.G.; Rankin, M. Corn response to within row plant spacing variation. Agron J. 2004, 96, 1464–1468. [Google Scholar] [CrossRef] [Green Version]
  10. Edwards, J.W. Genotype × environment interaction for plant density response in maize (Zea mays L.). Crop Sci. 2016, 56, 1493–1505. [Google Scholar] [CrossRef] [Green Version]
  11. Nielsen, R. Stand Establishment Variability In Corn. Agronomy Department, Purdue University. Available online: https://www.agry.purdue.edu/cca/2006/PDF/Nielsen.pdf (accessed on 12 January 2021).
  12. Varela, S.; Dhodda, P.R.; Hsu, W.H.; Prasad, P.V.; Assefa, Y.; Peralta, N.R.; Griffin, T.; Sharda, A.; Ferguson, A.; Ciampitti, I.A. Early-season stand count determination in Corn via integration of imagery from unmanned aerial systems (UAS) and supervised learning techniques. Remote Sens. 2018, 10, 343. [Google Scholar] [CrossRef] [Green Version]
  13. Zhang, C.; Kovacs, J.M. The application of small unmanned aerial systems for precision agriculture: A review. Precis. Agric. 2012, 13, 693–712. [Google Scholar] [CrossRef]
  14. Niu, Y.; Zhang, L.; Zhang, H.; Han, W.; Peng, X. Estimating above-ground biomass of maize using features derived from UAV-based RGB imagery. Remote Sens. 2019, 11, 1261. [Google Scholar] [CrossRef] [Green Version]
  15. Zhang, L.; Zhang, H.; Niu, Y.; Han, W. Mapping maize water stress based on UAV multispectral remote sensing. Remote Sens. 2019, 11, 605. [Google Scholar] [CrossRef] [Green Version]
  16. Sarić, R.; Nguyen, V.D.; Burge, T.; Berkowitz, O.; Trtílek, M.; Whelan, J.; Lewsey, M.G.; Čustović, E. Applications of hyperspectral imaging in plant phenotyping. Trends Plant Sci. 2022, 27, 301–315. [Google Scholar] [CrossRef]
  17. Feng, A.; Zhou, J.; Vories, E.; Sudduth, K.A. Evaluation of cotton emergence using uav-based narrow-band spectral imagery with customized image alignment and stitching algorithms. Remote Sens. 2020, 12, 1764. [Google Scholar] [CrossRef]
  18. Kuznetsova, A.; Maleva, T.; Soloviev, V. Detecting Apples in Orchards Using YOLOv3 and YOLOv5 in General and Close-Up Images. In Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Berlin/Heidelberg, Germany, 2020; pp. 233–243. [Google Scholar] [CrossRef]
  19. Ong, P.; Teo, K.S.; Sia, C.K. UAV-based weed detection in Chinese cabbage using deep learning. Smart Agric. Technol. 2023, 4, 100181. [Google Scholar] [CrossRef]
  20. Feng, A.; Zhou, J.; Vories, E.; Sudduth, K.A. Evaluation of cotton emergence using UAV-based imagery and deep learning. Comput. Electron. Agric. 2020, 177, 105711. [Google Scholar] [CrossRef]
  21. Lin, Z.; Guo, W. Cotton Stand Counting from Unmanned Aerial System Imagery Using MobileNet and CenterNet Deep Learning Models. Remote Sens. 2021, 13, 142822. [Google Scholar] [CrossRef]
  22. Ranđelović, P.; Đorđević, V.; Milić, S.; Balešević-Tubić, S.; Petrović, K.; Miladinović, J.; Đukić, V. Prediction of soybean plant density using a machine learning model and vegetation indices extracted from RGB images taken with a UAV. Agronomy 2020, 10, 1108. [Google Scholar] [CrossRef]
  23. Mhango, J.K.; Harris, E.W.; Green, R.; Monaghan, J.M. Mapping potato plant density variation using aerial imagery and deep learning techniques for precision agriculture. Remote Sens. 2021, 13, 2705. [Google Scholar] [CrossRef]
  24. Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. Yolox: Exceeding Yolo Series in 2021. arXiv 2021, arXiv:2107.08430. Available online: https://arxiv.org/abs/2107.08430 (accessed on 6 August 2021).
  25. Wang, C.-Y.; Yeh, I.-H.; Liao, H.-Y.M. You Only Learn One Representation: Unified Network for Multiple Tasks. arXiv 2021, arXiv:2105.04206. Available online: http://arxiv.org/abs/2105.04206 (accessed on 10 May 2021).
  26. Gallo, I.; Rehman, A.U.; Dehkordi, R.H.; Landro, N.; La Grassa, R.; Boschetti, M. Deep Object Detection of Crop Weeds: Performance of YOLOv7 on a Real Case Dataset from UAV Images. Remote Sens. 2023, 15, 539. [Google Scholar] [CrossRef]
  27. Jiang, P.; Ergu, D.; Liu, F.; Cai, Y.; Ma, B. A Review of Yolo Algorithm Developments. In Procedia Computer Science; Elsevier: Amsterdam, The Netherlands, 2022; pp. 1066–1073. [Google Scholar] [CrossRef]
  28. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. Available online: http://arxiv.org/abs/1804.02767 (accessed on 8 April 2018).
  29. Tzutalin, “LabelImg”. Tzutalin, D. LabelImg. Git Code. Available online: https://github.com/tzutalin/labelImg (accessed on 3 March 2019).
  30. Jocher, G. YOLOv5 by Ultralytics. Available online: https://github.com/ultralytics/yolov5 (accessed on 25 June 2020).
  31. Singh, P.; Singh, N.; Singh, K.K.; Singh, A. Diagnosing of disease using machine learning. In Machine Learning and the Internet of Medical Things in Healthcare; Elsevier: Amsterdam, The Netherlands, 2021; pp. 89–111. [Google Scholar] [CrossRef]
  32. Oh, S.; Chang, A.; Ashapure, A.; Jung, J.; Dube, N.; Maeda, M.; Gonzalez, D.; Landivar, J. Plant counting of cotton from UAS imagery using deep learning-based object detection framework. Remote Sens. 2020, 12, 2981. [Google Scholar] [CrossRef]
  33. Willmott, C.J.; Matsuura, K. Advantages of the mean absolute error (MAE) over the root mean square error (RMSE) in assessing average model performance. Clim. Res. 2005, 30, 79–82. [Google Scholar] [CrossRef]
  34. Ali, M.A. Effect of Sample Size on the Size of the Coefficient of Determination in Simple Linear Regression. J. Inf. Optim. Sci. 1987, 8, 209–219. [Google Scholar] [CrossRef]
  35. Vong, C.N.; Conway, L.S.; Zhou, J.; Kitchen, N.R.; Sudduth, K.A. Early corn stand count of different cropping systems using UAV-imagery and deep learning. Comput. Electron. Agric. 2021, 186, 106214. [Google Scholar] [CrossRef]
  36. Wang, L.; Xiang, L.; Tang, L.; Jiang, H. A convolutional neural network-based method for corn stand counting in the field. Sensors 2021, 21, 507. [Google Scholar] [CrossRef]
  37. Zhang, Y.; Zhang, W.; Yu, J.; He, L.; Chen, J.; He, Y. Complete and accurate holly fruits counting using YOLOX object detection. Comput. Electron. Agric. 2022, 198, 107062. [Google Scholar] [CrossRef]
  38. Song, J.; Zhao, Y.; Song, W.; Zhou, H.; Zhu, D.; Huang, Q.; Fan, Y.; Lu, C. Fisheye Image Detection of Trees Using Improved YOLOX for Tree Height Estimation. Sensors 2022, 22, 3636. [Google Scholar] [CrossRef]
  39. Zhong, J.; Li, M.; Qin, J.; Cui, Y.; Yang, K.; Zhang, H. Real-time marine animal detection using yolo-based deep learning networks in the coral reef ecosystem. In International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences—ISPRS Archives; International Society for Photogrammetry and Remote Sensing: Bethesda, MD, USA, 2022; pp. 301–306. [Google Scholar] [CrossRef]
  40. Wang, C.; Guo, X.; Zhao, C. Detection of corn plant population and row spacing using computer vision. In Proceedings of the 2011 2nd International Conference on Digital Manufacturing and Automation, ICDMA 2011, Zhangjiajie, China, 5–7 August 2011; pp. 405–408. [Google Scholar] [CrossRef]
  41. Tang, L.; Tian, L.F. Plant Identification In Mosaicked Crop Row Images For Automatic Emerged Corn Plant Spacing Measurement. Trans. ASABE 2008, 51, 2181–2191. [Google Scholar] [CrossRef]
  42. Seifert, E.; Seifert, S.; Vogt, H.; Drew, D.; Van Aardt, J.; Kunneke, A.; Seifert, T. Influence of drone altitude, image overlap, and optical sensor resolution on multi-view reconstruction of forest images. Remote Sens. 2019, 11, 1252. [Google Scholar] [CrossRef] [Green Version]
  43. Tong, K.; Wu, Y.; Zhou, F. Recent advances in small object detection based on deep learning: A review. Image Vis. Comput. 2020, 97, 103910. [Google Scholar] [CrossRef]
  44. Zhao, B.; Zhang, J.; Yang, C.; Zhou, G.; Ding, Y.; Shi, Y.; Zhang, D.; Xie, J.; Liao, Q. Rapeseed seedling stand counting and seeding performance evaluation at two early growth stages based on unmanned aerial vehicle imagery. Front Plant Sci. 2018, 9, 1362. [Google Scholar] [CrossRef]
  45. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common objects in context. In Lecture Notes in Computer Science (including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Elsevier: Amsterdam, The Netherlands, 2014; Volume 8693, pp. 740–755. [Google Scholar] [CrossRef] [Green Version]
  46. Lin, Z.; Guo, W. Sorghum Panicle Detection and Counting Using Unmanned Aerial System Images and Deep Learning. Front Plant Sci. 2020, 11, 534853. [Google Scholar] [CrossRef]
  47. Shorten, C.; Khoshgoftaar, T.M. A survey on Image Data Augmentation for Deep Learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Five treatment conditions of plot-level planting space variability.
Figure 1. Five treatment conditions of plot-level planting space variability.
Agronomy 13 01728 g001
Figure 2. Examples of labeled bounding boxes for maize stands and PSV spots for a maize plot. (a) The labeled maize stands, with each bounding box representing a maize plant. (b) The labeled PSV locations within each plot. Each maize plot had one to three different plant spacings, namely, big_gap (Big Gaps), skips (Skips), and/or misplacements (Misplacements).
Figure 2. Examples of labeled bounding boxes for maize stands and PSV spots for a maize plot. (a) The labeled maize stands, with each bounding box representing a maize plant. (b) The labeled PSV locations within each plot. Each maize plot had one to three different plant spacings, namely, big_gap (Big Gaps), skips (Skips), and/or misplacements (Misplacements).
Agronomy 13 01728 g002
Figure 3. The workflow of the whole process throughout the experiment, including data preprocessing, the training of the models, and the result output expectations. The whole experiment was divided into two parallel sub-experiment branches. The left branch applied deep learning algorithms to focus on maize stand counting. The right branch was designed solely based on deep learning algorithms to detect PSV.
Figure 3. The workflow of the whole process throughout the experiment, including data preprocessing, the training of the models, and the result output expectations. The whole experiment was divided into two parallel sub-experiment branches. The left branch applied deep learning algorithms to focus on maize stand counting. The right branch was designed solely based on deep learning algorithms to detect PSV.
Agronomy 13 01728 g003
Figure 4. Scatter plot of predicted count vs. ground count from YOLOv5, YOLOX, and YOLOR under different training epochs (IoU = 0.5).
Figure 4. Scatter plot of predicted count vs. ground count from YOLOv5, YOLOX, and YOLOR under different training epochs (IoU = 0.5).
Agronomy 13 01728 g004
Figure 5. Original plot-level image and a segment stand counting demo for YOLOv5, YOLOX, and YOLOR under the Misplacement treatment. Both YOLOv5 and YOLOR have overlapping bounding boxes, as compared with YOLOX. Each color bounding box represents a single maize stand.
Figure 5. Original plot-level image and a segment stand counting demo for YOLOv5, YOLOX, and YOLOR under the Misplacement treatment. Both YOLOv5 and YOLOR have overlapping bounding boxes, as compared with YOLOX. Each color bounding box represents a single maize stand.
Agronomy 13 01728 g005
Figure 6. Original plot-level image and a segment stand counting demo for YOLOv5, YOLOX, and YOLOR under the Skips treatment. YOLOv5 and YOLOR have worse stand counting results compared with YOLOX due to redundant counts. Each color bounding box represents a single maize stand.
Figure 6. Original plot-level image and a segment stand counting demo for YOLOv5, YOLOX, and YOLOR under the Skips treatment. YOLOv5 and YOLOR have worse stand counting results compared with YOLOX due to redundant counts. Each color bounding box represents a single maize stand.
Agronomy 13 01728 g006
Figure 7. Scatter plot of predicted value vs. ground truth for YOLOv5, YOLOX, and YOLOR under the optimized NMS IoU.
Figure 7. Scatter plot of predicted value vs. ground truth for YOLOv5, YOLOX, and YOLOR under the optimized NMS IoU.
Agronomy 13 01728 g007
Figure 8. Visualization of different PSVs for the three models: YOLOv5, YOLOX, and YOLOR. (a) Misplacement, (b) Skips, and (c) Big Gaps. Both YOLOv5 and YOLOR mis-detected the boundaries as Big Gap conditions.
Figure 8. Visualization of different PSVs for the three models: YOLOv5, YOLOX, and YOLOR. (a) Misplacement, (b) Skips, and (c) Big Gaps. Both YOLOv5 and YOLOR mis-detected the boundaries as Big Gap conditions.
Agronomy 13 01728 g008aAgronomy 13 01728 g008bAgronomy 13 01728 g008c
Table 1. Parameters used for the three models for training. All the parameters are the original settings.
Table 1. Parameters used for the three models for training. All the parameters are the original settings.
ModelParameterValues
YOLOv5Learning rate1 × 10−3
Momentum0.98
Learning decay1 × 10−2
YOLOXLearning rate0.01 × BatchSize/64
Momentum0.9
Learning decay5 × 10−4
YOLORLearning rate2.61 × 10−3
Momentum0.949
Learning decay5 × 10−4
Table 2. Performance in maize stand detection and counting.
Table 2. Performance in maize stand detection and counting.
ModelTraining EpochAverage PrecisionCoefficient of Determination (R2)Mean Absolute Error (MAE)
YOLOv55000.9170.6216.208
10000.9310.7085.688
15000.9210.7245.333
YOLOX5000.8980.8053.542
10000.8890.7733.792
15000.8820.7104.354
YOLOR5000.9200.7904.583
10000.9040.7674.958
15000.9020.8384.104
Table 3. Average precision under different NMS IoU thresholds.
Table 3. Average precision under different NMS IoU thresholds.
ModelTraining EpochNMS IoU ThresholdAP
YOLOv55000.10.921
0.20.924
0.30.923
0.40.920
0.50.917
10000.10.934
0.20.934
0.30.933
0.40.931
0.50.931
15000.10.922
0.20.929
0.30.929
0.40.926
0.50.921
YOLOX5000.10.897
0.20.898
0.30.898
0.40.898
0.50.898
10000.10.806
0.20.892
0.30.891
0.40.891
0.50.891
15000.10.800
0.20.799
0.30.799
0.40.878
0.50.882
YOLOR5000.10.927
0.20.931
0.30.932
0.40.930
0.50.920
10000.10.931
0.20.935
0.30.936
0.40.937
0.50.904
15000.10.927
0.20.931
0.30.933
0.40.932
0.50.902
Table 4. AP, Pearson’s correlation, and MAE under various NMS IoU thresholds.
Table 4. AP, Pearson’s correlation, and MAE under various NMS IoU thresholds.
ModelNMS IoUMisplacementPerfect, Skips, NES, or Big Gaps
APPearson’s CorrelationMAEAPPearson’s
Correlation
MAE
YOLOv5-10000.10.8510.16215.50.9550.9782.711
0.20.8510.50312.40.9580.9861.605
0.30.850.35810.60.9570.9901.316
0.40.8460.4708.10.9570.9861.184
0.50.840.3194.90.9560.9645.895
0.60.840.37912.20.9570.93212.421
0.70.8360.44123.40.9570.88819.447
0.80.8170.44042.60.9530.78134.316
0.90.7080.57095.00.8680.62488.105
YOLOX-5000.10.7160.56913.10.9020.9692.868
0.20.7990.92010.90.9020.9672.053
0.30.7990.85310.20.9020.9751.789
0.40.7980.7918.70.9020.9761.553
0.50.7980.8337.00.9020.9731.579
0.60.7980.6154.90.9020.9681.763
0.70.7980.6105.00.9020.9651.947
0.80.7960.61512.20.9020.9524.000
0.90.7760.72343.80.8930.82425.053
YOLOR-10000.10.8560.18214.50.9550.9832.632
0.20.8570.39512.40.9580.9792.000
0.30.8610.68810.20.9590.9791.763
0.40.8580.6928.90.9610.9791.579
0.50.8530.7073.90.9590.9595.289
0.60.8480.5297.70.9590.9459.974
0.70.8420.59614.20.9590.93213.158
0.80.8340.66324.50.9570.92619.974
0.90.7740.32368.50.9240.84859.184
Table 5. Stand counting results after choosing the best average precision NMS IoU threshold.
Table 5. Stand counting results after choosing the best average precision NMS IoU threshold.
ModelTraining EpochNMS IoU Threshold (Misplacement/Other Four)Coefficient of Determination (R2)Mean Absolute Error (MAE)
YOLOv510000.5/0.40.9361.958
YOLOX5000.6/0.40.9182.417
YOLOR10000.5/0.40.9462.063
Table 6. Precision, Recall, and F1-score for maize spacing treatment detection with three different deep learning models under different epochs.
Table 6. Precision, Recall, and F1-score for maize spacing treatment detection with three different deep learning models under different epochs.
ModelTraining EpochPrecisionRecallF1
YOLOv52000.8610.9210.890
5000.8620.890.876
8000.8860.8620.876
10000.8450.8340.840
YOLOX2000.9410.8540.896
5000.9030.8250.862
8000.8310.8640.847
10000.8990.8560.877
YOLOR2000.7850.8770.829
5000.8760.8670.872
8000.8790.8760.878
10000.8750.8520.863
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, B.; Zhou, J.; Costa, M.; Kaeppler, S.M.; Zhang, Z. Plot-Level Maize Early Stage Stand Counting and Spacing Detection Using Advanced Deep Learning Algorithms Based on UAV Imagery. Agronomy 2023, 13, 1728. https://doi.org/10.3390/agronomy13071728

AMA Style

Wang B, Zhou J, Costa M, Kaeppler SM, Zhang Z. Plot-Level Maize Early Stage Stand Counting and Spacing Detection Using Advanced Deep Learning Algorithms Based on UAV Imagery. Agronomy. 2023; 13(7):1728. https://doi.org/10.3390/agronomy13071728

Chicago/Turabian Style

Wang, Biwen, Jing Zhou, Martin Costa, Shawn M. Kaeppler, and Zhou Zhang. 2023. "Plot-Level Maize Early Stage Stand Counting and Spacing Detection Using Advanced Deep Learning Algorithms Based on UAV Imagery" Agronomy 13, no. 7: 1728. https://doi.org/10.3390/agronomy13071728

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop