Next Article in Journal
Untargeted UHPLC-MS Metabolomics Reveals the Metabolic Perturbations of Helicoverpa armigera under the Stress of Novel Insect Growth Regulator ZQ-8
Next Article in Special Issue
Counting Crowded Soybean Pods Based on Deformable Attention Recursive Feature Pyramid
Previous Article in Journal
Relationships between Rhizosphere Environments and Growth of 10-Year-Old Wild-Simulated Ginseng
Previous Article in Special Issue
Detection of Male and Female Litchi Flowers Using YOLO-HPFD Multi-Teacher Feature Distillation and FPGA-Embedded Platform
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detection of Litchi Leaf Diseases and Insect Pests Based on Improved FCOS

1
College of Electronic Engineering (College of Artificial Intelligence), South China Agricultural University, Guangzhou 510642, China
2
Guangdong Laboratory for Lingnan Modern Agriculture, South China Agricultural University, Guangzhou 510642, China
3
Engineering Research Center for Monit Oring Agricultural Information of Guangdong Province, Guangzhou 510642, China
4
Zhujiang College, South China Agricultural University, Guangzhou 510900, China
5
College of Engineering, South China Agricultural University, Guangzhou 510642, China
*
Author to whom correspondence should be addressed.
Agronomy 2023, 13(5), 1314; https://doi.org/10.3390/agronomy13051314
Submission received: 4 April 2023 / Revised: 5 May 2023 / Accepted: 5 May 2023 / Published: 7 May 2023
(This article belongs to the Special Issue Precision Operation Technology and Intelligent Equipment in Farmland)

Abstract

:
Litchi leaf diseases and pests can lead to issues such as a decreased Litchi yield, reduced fruit quality, and decreased farmer income. In this study, we aimed to explore a real-time and accurate method for identifying Litchi leaf diseases and pests. We selected three different orchards for field investigation and identified five common Litchi leaf diseases and pests (Litchi leaf mite, Litchi sooty mold, Litchi anthracnose, Mayetiola sp., and Litchi algal spot) as our research objects. Finally, we proposed an improved fully convolutional one-stage object detection (FCOS) network for Litchi leaf disease and pest detection, called FCOS for Litch (FCOS-FL). The proposed method employs G-GhostNet-3.2 as the backbone network to achieve a model that is lightweight. The central moment pooling attention (CMPA) mechanism is introduced to enhance the features of Litchi leaf diseases and pests. In addition, the center sampling and center loss of the model are improved by utilizing the width and height information of the real target, which effectively improves the model’s generalization performance. We propose an improved localization loss function to enhance the localization accuracy of the model in object detection. According to the characteristics of Litchi small target diseases and pests, the network structure was redesigned to improve the detection effect of small targets. FCOS-FL has a detection accuracy of 91.3% (intersection over union (IoU) = 0.5) in the images of five types of Litchi leaf diseases and pests, a detection rate of 62.0/ms, and a model parameter size of 17.65 M. Among them, the detection accuracy of Mayetiola sp. and Litchi algal spot, which are difficult to detect, reached 93.2% and 92%, respectively. The FCOS-FL model can rapidly and accurately detect five common diseases and pests in Litchi leaf. The research outcome is suitable for deployment on embedded devices with limited resources such as mobile terminals, and can contribute to achieving real-time and precise identification of Litchi leaf diseases and pests, providing technical support for Litchi leaf diseases’ and pests’ prevention and control.

1. Introduction

Litchi (Litchi chinensis Sonn) is an evergreen tree belonging to the Sapindaceae family. China is the country with the widest cultivation area and the largest yield of Litchi in the world. Litchi is rich in nutrients and has a delicious taste. As a cash crop tree species, it is widely planted in South China. In China’s mainland, the total output value of Litchi was 33.389 billion yuan in 2022 according to the acquisition price [1]. Therefore, promoting the development of the Litchi planting industry is beneficial to boosting the economic benefits of farmers and meeting the market’s demands for scaled and high-quality Litchis [2,3]. The climate in southern China is usually hot and humid, making Litchi susceptible to various pest and disease problems. Litchi leaf disease and pest problems can result in a decrease in Litchi production and a decline in fruit quality. These issues significantly obstruct the steady development of the Litchi industry [4]. The key to solving this problem is to detect, identify, and provide feedback on the disease and insect pests at the early stages of their occurrence. Therefore, timely and efficient detection of Litchi leaf diseases and insect pests is an important measure to ensure the yield of Litchi. Traditional artificial recognition is greatly influenced by subjective factors and may not be able to successfully diagnose specific diseases, leading to erroneous conclusions and treatment [5,6]. With the rapid popularization of computer technology, it is possible to quickly identify and detect pests and diseases, greatly promoting the development of intelligent agriculture [7].
In early agricultural pest detection research, traditional image processing methods using support vector machine (SVM) [8], K-nearest neighbor (KNN) [9], and other algorithms to detect target pests [10,11,12] were used for agricultural pest detection. Liu et al. [13] used the histogram of oriented gradients (HOG) feature extraction method to extract features from wheat aphid images in the wheat field, and then used SVM classifiers to identify and locate wheat aphids. Sabro et al. [14] used the Otsu threshold method to segment tomato leaf diseases and insect pests, manually extracting the shape, color, texture, and other characteristics of the disease, and then using a decision tree to classify tomato leaf diseases and insect pests. Pydipati et al. [15] used the color co-occurrence method (CCM) to determine whether texture-based hue, saturation, and intensity (HSI) color features in conjunction with statistical classification algorithms could be used to identify diseased and normal citrus leaves under laboratory conditions. Traditional pest detection methods have achieved some results, but there are still problems such as low efficiency, poor segmentation results, and low recognition rates in complex backgrounds. Therefore, they cannot meet the need for the real-time and accurate detection of Litchi leaf diseases and pests in complex field environments [16].
In recent years, significant progress has been made in image recognition technology based on convolutional neural networks [17,18]. Target detection is a branch of image recognition based on convolutional neural network algorithms. A large number of researchers have applied target detection technology to research crop disease and pest detection, making significant breakthroughs [19,20,21]. Zhang et al. [22] proposed an improved YOLOv5 network that combines DenseNet, attention mechanism, and Bi-FPN to accurately detect unopened cotton bolls in the field at a lower cost. Liu et al. [23] proposed an improved convolutional neural network and a PestNet algorithm with a modular channel attention mechanism to identify 16 pests. Experiments showed that the average accuracy rate reached 75.46%. Dai et al. [24] proposed an improved YOLOv5m method, introducing the Swin Transformer mechanism into the YOLOv5m network to capture more global features and increase receptive fields, enabling more accurate detection of different pests from the dataset. Zhang et al. [25] used the attention method of dynamic mechanism fusion to improve DenseNet and proposed the Dense Channel and Position Self-Attention Fusion Network model, with a high recognition accuracy of 96.90% for six types of navel orange diseases and pests. Liu et al. [26] developed a tomato pest recognition algorithm based on the improved YOLOv4 fusion triple attention mechanism (YOLOv4-TAM), which solves the problem of uneven positive and negative sample numbers in the recognition system and improves the accuracy in identifying tomato pests and diseases.
Research on the intelligent recognition of disease and insect images based on convolutional neural networks has made steady progress in recent years. However, due to the wide variety of diseases and insect pests in Litchi and the complex image background, two types of diseases and insect pests, Mayetiola sp. and Litchi algal spot, are characterized by small and dense pest features, which are likely to cause false detection and missed detection of disease and insect targets, making detection difficult [27,28]. The existing neural network algorithms have high computational complexity, large parameter quantities, and low detection accuracy for small target diseases and insect pests, such as Mayetiola sp. and Litchi algal spot. There is still room for improvement in real-time detection and small target disease and insect pest identification. To address the issues of low accuracy and large model parameters in existing methods for detecting Litchi leaf diseases and pests, this paper proposes an FCOS-FL model for detecting Litchi leaf diseases and pests based on the images of five types of Litchi leaf diseases and pests collected from orchards. The proposed model achieves efficient and accurate recognition of the five common types of Litchi leaf diseases and pests, with a focus on solving the low detection accuracy of Mayetiola sp. and Litchi algal spot. This work provides a reference for the real-time and accurate detection of Litchi leaf diseases and pests and offers an effective technical support for the prevention and control of Litchi leaf diseases and pests. The innovation points of this study include:
(1)
Achieving a lightweight model by replacing the backbone network.
(2)
Enhancing the features of lychee pests and diseases by adding attention mechanisms.
(3)
Improving the generalization of the model by improving the central sampling and central measurement of the model.
(4)
Improving the positioning accuracy of the model by improving the loss function.

2. Materials and Methods

2.1. Dataset Production

2.1.1. Data Collection

The experimental data in this study were collected from three different orchards: the Guangdong Agricultural Technology Extension Station, Guangzhou City, Guangdong Province, China, where the variety collected was glutinous rice dumpling Litchi; the State Key Laboratory of South China Agricultural University, Guangzhou City, Guangdong Province, China, where the varieties collected were mainly Gui Wei Litchi and glutinous rice dumpling Litchi; and the Rijin Orchard in Matouling Farm, Gaozhou City, Guangdong Province and Xiaoliang Town, Maoming City, Guangdong Province, China, where the main variety collected was Gui Wei Litchi. The data collection period was from 11 July 2021 to 19 November 2021. Smartphones were chosen as the data collection devices because they meet the convenience requirements of data collection in actual orchard scenarios and can meet the demand for image pixel quality for disease and pest detection. The devices used were iPhone 12 and Xiaomi 6 smartphones. In order to obtain clear images of pests and diseases, the distance between the device and the diseased leaf was set to between 0.2 m and 0.5 m during shooting, and the resolution included 4032 × 3016 and 3016 × 3016, among others. To more realistically recreate the complex environment of the field and improve the model’s generalization, environmental factors were taken into consideration during data collection, including sunny, cloudy, and rainy days; the data collection time period was from 8:30 a.m. to 12:00 p.m. and from 2:00 p.m. to 6:00 p.m. to consider the effect of lighting conditions. To simulate the actual fruit collection process, the shooting angle was set as vertical 90°, upward 45°, and downward 45°. During the collection process, five common characteristics of Litchi leaf diseases and pests were identified by visual inspection and by consulting local agricultural experts and farmers [29,30], including Litchi sooty mold, Litchi anthracnose, Litchi algal spot, Litchi leaf mite, and Mayetiola sp. The characteristics of the five pests and diseases are shown in Table 1. In total, 3900 original images were collected.
Among them, Litchi leaf mite has 795 images, and Litchi sooty mold has 770 images. Litchi anthracnose has 783 images, Mayetiola sp. has 738 images, and Litchi algal spot has 814 images.

2.1.2. Data Cleaning and Labeling

In order to reduce the number of repetitive and blurred images, the method of manual selection was used to clean the collected images, i.e., images that were blurry due to handheld shooting or images with unclear features of Litchi leaf diseases and pests were deleted. The final dataset consisted of 3725 images, including 1061 images taken in direct sunlight, 960 images taken against backlight, 720 images taken after rain, and 984 images taken on cloudy days.
This experiment focuses on visible light images of Litchi leaf diseases and pests and aims to design a recognition model for Litchi leaf diseases and pests. The collected Litchi images require manual annotation to mark the regions with Litchi leaf diseases and pests, and to write the coordinates of the marked regions into XML files. Then, these annotated images can be fed into the neural network model to learn the features of Litchi leaf diseases and pests. In this study, the open-source annotation software “labelImg” was used for manual annotation of the Litchi leaf disease and pest dataset. LabelImg was developed by Chinese computer scientist Yude Wang, and the version used in this article is V1.8.5. With this software, the target region of Litchi leaf diseases and pests can be selected and labeled with its type of disease or pest, and an XML file containing the location coordinates and disease or pest categories of the target box can be obtained for neural network model training. The annotation process is illustrated in Figure 1.

2.1.3. Dataset Partitioning

After annotating the Litchi diseases and pests dataset using the labeling software LabelImg, the images and label files were divided into training and testing sets according to the requirements of different object detection models. The Litchi disease and pest image data were randomly divided into training sets, validation sets, and test sets at a ratio of 8:1:1. The basic information of the dataset is shown in Table 2.

2.2. Experimental Protocol

This paper proposes a Litchi leaf diseases and pests detection method that consists of three main steps: data collection and processing, model construction and optimization, and detection result output, as illustrated in Figure 2. Firstly, a Litchi leaf diseases and pests dataset is constructed by collecting and cleaning image data from field orchards, followed by manual annotation. Then, a Litchi leaf diseases and pests recognition model based on the FCOS model is designed to address the challenges of recognizing small targets with multiple categories and improving the model’s lightweight performance for practical application. By improving the FCOS model, a Litchi leaf diseases and pests detection model based on FCOS is constructed to achieve accurate and efficient identification of diseases and pests on Litchi leaves, providing technical support for the prevention and control of Litchi diseases and pests and improving the yield and quality of Litchi.

2.3. FCOS-FL Construction

2.3.1. FCOS Model

The FCOS algorithm is a fully convolutional, anchor-free, single-stage target detection network proposed by Tian et al. [31]. It is one of the representative methods for target detection based on deep learning, and it has achieved good results in conventional target detection tasks. The basic idea of the FCOS algorithm is to achieve the goal of target detection by classifying and regressing each position in the image. The FCOS network structure is shown in Figure 3. The FCOS network mainly includes three parts: the feature extraction backbone of the residual neural network (ResNet) [32]; the feature pyramid network (FPN) for object detection [33]; FCOS for header detection.
Specifically, the Litchi leaf diseases and pests dataset serves as the input, and a series of convolutional and pooling operations are performed to extract the features of the manually annotated target areas of the diseases and pests. These feature maps are then passed into a residual network for further feature extraction, in order to accurately identify the diseases and pests. Next, the feature maps are passed into a feature pyramid network, which upsamples the maps to different resolutions to adapt to the uneven distribution of sizes in the Litchi leaf diseases and pests. Finally, each feature map is classified, regressed, and center point detected to determine whether a target is present at that location. The regressor predicts the target box and detects the center point of each disease or pest target to determine the precise location. The model assigns a confidence score to each location and category to indicate whether it belongs to a certain type of disease or pest. Based on the score, the most likely target is determined and information such as its category, location, and confidence score is output.

2.3.2. G-GhostNet

HAN [34] proposed the improved G-GhostNet in 2022. This article uses G-GhostNet-3.2 as the backbone network extraction feature for improving FCOS. G-GhostNet is a new lightweight network architecture optimized for server GPUs. G-GhostNet utilizes neural networks to extract effective redundant feature maps in each stage, with the feature map size unchanged between different structural blocks. The structure is shown in Figure 4. The redundant feature map obtained directly from shallow structural blocks in Figure 3a lacks high semantic information; thus, as shown in Figure 4b, a mix operation is added for information compensation. The mix operation with information compensation is depicted in Figure 4c.
The mix operation of information compensation is shown in Figure 4c, where Y i c represents the output characteristic diagram of the i-th structural block, and the stacked characteristic diagram is represented as Z R   c × h × w . This is followed by τ Z average pooling and convolution, which can be expressed as follows:
τ Z = W × Pooling Z + b .
From the perspective of practical application of the model, this paper considers the difficulty of deploying the model on mobile and embedded devices. To address this issue, the G-GhostNet module, which is characterized by its versatility and scalability, is introduced into the FCOS model to reduce the computational and parameter complexity of the FCOS-FL model and optimize its lightweight design.

2.3.3. Attention Mechanism

The attention mechanism is a method that utilizes global pooling operations to extract global information from a feature, and then applies the global information to the feature graph. This method utilizes global maximum pooling and average pooling to extract global information, enhancing the ability of network feature extraction. Wu Tong et al. [35] proposed a central moment pooling spatial attention mechanism based on statistical analysis.
This paper addresses the challenge of recognizing Litchi leaf diseases and pests due to their small size and high difficulty. To enhance the recognition accuracy of disease and pest targets, this paper combines spatial attention mechanism with central moment pooling, which provides more statistical information on the features of Litchi leaf diseases and pests. Specifically, the central moment pooling attention (CMPA) mechanism is introduced into the FCOS model to enhance the Litchi leaf disease and pest features using a superior spatial attention mechanism. The structure of CMPA is shown in Figure 5. By adopting this approach, the model achieves better recognition accuracy for Litchi leaf diseases and pests.

2.3.4. Improved Center Sampling and Center Metrics

In response to the uneven size and significant differences in the morphology of Litchi leaf diseases and pests, using the original center sampling formula can lead to inaccurate recognition of the target. As shown in Figure 6b, if the original center sampling method is used, sampling is conducted using a fixed square, which may miss some positive sample points [36] for targets with significant differences in width and height, resulting in a positive sample imbalance. To address this issue, an improved center sampling method that can adaptively adjust based on the aspect ratio of the target is proposed, as shown in Figure 6c. This method increases the number of positive samples during the training process and improves the problem of positive sample imbalance.
This paper proposes a new center measurement method to address the problem of predicting bounding boxes for Litchi leaf disease and pest targets with large differences in width and height. The new centerness formula is designed as Equation (2). In the Equation (2), w and h represent the width and height of the true bounding box of the object, respectively. l*, t*, r*, and b* respectively denote the distances between a certain positive sample point and the left, top, right, and bottom edges of the true bounding box.
C e n t e r n e s s * = min ( l * , r * ) max ( l * , r * ) × min ( t * , b * ) max ( t * , b * ) 1 2 × min ( w , h ) max ( w , h )
This paper addresses the problem of box regression for Litchi leaf disease and pest targets with large aspect ratios. The original centerness metric is not sensitive enough to the aspect ratio of the targets, and when a sample point is located far from the center of the longer side but still within the center of the shorter side (as shown in the green grid of Figure 7a, its centerness value is lower, and its predicted box score is forced to decrease, leading to excessive suppression of the sample point [37]. As shown in (b) of Figure 7, based on Equation (2), when there are Litchi leaf diseases and pests targets with large aspect ratios in the image, the centerness metric of the above sample point will not be too low, and its predicted box will not be excessively suppressed. The improved centerness can still suppress the effect of sample points located at the edge, making FCOS pay more attention to the center of the Litchi leaf diseases and pests targets, thereby improving the accuracy of box regression.

2.3.5. Improved Loss Function

The loss function is a function used to measure the difference between the model’s predicted results and the true results, and can guide the model’s optimization and evaluate its performance. The original model used the IoU loss to perform regression training, as shown in Equation (3).
L I o U = ln I o U .
L r e g = L I o U * = α I o U γ L I o U .
The original IoU loss did not reflect the true differences between the width, height, and confidence of the predicted bounding boxes, leading to errors in the predicted boxes. To address the problem of Litchi leaf diseases and pests being densely connected and difficult to distinguish, resulting in significant differences between positive and negative samples in small target images, this paper introduces the concept of hard example mining to focus the training process on beneficial samples, thereby improving the accuracy of the model in recognizing and locating difficult to identify disease and pest targets. The improved loss function is expressed as Equation (4).
Similarly to focal loss, α I o U γ , it is mainly used for difficultly divided sample mining, which is herein referred to as difficultly divided sample weight. A larger IoU value of a sample and a real target indicates more accurate positioning. At this point, the sample is defined as a separable sample. When the IoU value is smaller, it indicates that its positioning is more inaccurate. At this time, the sample is defined as hard to distinguish. In order for the model to focus on the location and regression of difficultly divided samples during the training process, an adaptive weight is added to the original loss, and the model detection accuracy is improved by amplifying the location loss of difficultly divided samples.

2.3.6. Detection Model of Litchi Leaf Diseases and Pests Based on FCOS-FL

The FCOS-FL Litchi leaf disease and pest detection model proposed in this article is shown in Figure 8. The original ResNet-50 is replaced with G-GhostNet-3.2 as the backbone network to extract features, reducing the weight of FCOS and accelerating detection speed. A central moment pooling attention mechanism (CMPA) is added to the last three stages of G-GhostNet-3.2 to enhance the feature maps. Using the width and height information of real targets to improve the center sampling and center measurement of FCOS, it is possible to obtain more positive samples based on the width and height of real targets during model sampling, thus improving the center measurement, avoiding excessive suppression of positive samples of targets with large width and height differences, and improving the generalization of FCOS. An improved location loss function is proposed, which adds difficultly divided sample weights to the FCOS training process to improve the location accuracy of the model. Aiming at the small size of Litchi leaf diseases and insect pests, the FCOS detection layer was adjusted, and the outputs from C2 to C5 in the last four stages of the backbone network G-GhostNet-3.2 were extracted as the input characteristic map of the FPN. The two downsamples in the FPN were changed to one downsample to obtain P6, and P3 was upsampled and fused with C2 to obtain P2. The input of the detection head was adjusted to P2 to P6, maintaining five layers.

2.4. Training Environment and Parameter Settings

The server hardware configuration and software environment configuration used in the experiment are shown in the Table 3.
This article used an enhanced training set of 3725 Litchi images for model training. To enhance test reliability, the image resolution was uniformly scaled to 512 × 512 before inputting network model training. A single graphics card was used for training. The training involved 200 rounds, the batch size was set to 8, the number of iterations was set to 48, and the initial learning rate was set to 0.01. Warmup [38] was used to warm up the learning rate during training to improve the stability of the model. The cosine annealing strategy was used to update the learning rate. The optimizer used SGD [39] to adjust the error. When evaluating the detection performance of the trained model, the detection results were postprocessed using NMS with non-maximum suppression, and the IoU threshold of NMS was set to 0.5.

2.5. Model Evaluators

In this study, precision, recall, average precision(AP), mean average precision (mAP), parameters (Params), and average time per single image in the test set (Time) were used as model evaluation metrics. Recall and precision are two of the important indexes for evaluating the model. The larger the area of the precision–recall curve, the better the comprehensive performance of the model. mAP is a measure of detection accuracy in target detection, and the higher the mAP, the better the detection effect of the model. Model parameter quantity (Params) is the quantity of parameters contained in the model structure, and the smaller the parameter quantity, the smaller the memory space required for model operations.

3. Experiments and Results

3.1. Ablation Contrast Experiment

In order to verify the effectiveness of the various improvement strategies proposed, a series of ablation experiments were conducted in this paper. The network model was trained using the same training parameters on the same platform and under the same experimental framework. This mainly included a comparison of different backbone networks, the exploration of the optimal location of the central moment pooling attention mechanism (CMPA) in the backbone network, the exploration of parameter settings for difficult sample weights in the improved loss function, the comparison of various improvement schemes in FCOS and FCOS-FL, and the comparison of different models.

3.1.1. Comparison of Different Backbone Networks

In order to verify the effectiveness of using lightweight convolutional neural network G-GhostNet-3.2 as the backbone network of FCOS, different backbone networks were introduced in FCOS models, and then trained on the constructed Litchi leaf disease and pest dataset. The comparison results of the detection performance of FCOS under different backbone networks are shown in Table 4.
As can be seen from Table 4, when the IoU threshold was 0.5 and 0.75, and the average of other IoU thresholds was AP, G-GhostNet-3.2 had the highest detection accuracy compared to other backbone networks, reaching 88.4%, 66.8%, and 60.2%, respectively. Among them, the accuracy of the original FCOS using ResNet-50 as the backbone network was second only to G-GhostNet-3.2, but its disadvantage was that the number of parameters, amount of computation, and detection time were higher than those of other backbone networks. In addition, the parameter number and calculation amount of GhostNet were not the lowest in this comparison, but the detection time was the lowest, while the detection accuracy was not inferior to EfficientNet-b2, proving its efficiency. Although G-GhostNet-3.2 was not superior to MobileNet-v2 in terms of its light weight, it achieved a greater degree of light weight based on improved accuracy compared to ResNet-50 of the initial FCOS. Figure 9 shows the performance of the G-GhostNet backbone network in terms of loss and AP50. It can be seen from the figure that the G-GhostNet network backbone performed best. The above comparative experiments fully prove that using G-GhostNet-3.2 as the backbone network to extract features could meet the real-time requirements of Litchi leaf disease and pest detection tasks, with a greater advantage in detection accuracy, which is significant for improving FCOS.

3.1.2. Comparison of CMPA at Different Locations

To verify the effectiveness of the central moment pooling attention mechanism (CMPA) and to explore its effectiveness in feature enhancement in FCOS models, comparative experiments were conducted with CMPA placed in different locations. The results are shown in Table 5.
As can be seen from Table 5, when the CMPA was placed in the last three stages, the average AP50 of various pests and diseases reached the highest level at 89.3%. Moreover, the detection effect of difficult-to-detect Mayetiola sp. and Litchi algal spot pest and disease samples was excellent. Among them, Litchi leaf mite and Litchi algal spot had the highest detection effect on AP50, with 86.3% and 85.8%, respectively. Litchi anthracnose and Mayetiola sp. were second only to the highest detection rates of 92.1% and 88.4%, with detection results of 91.9% and 87.6%. Although the detection accuracy of Litchi sooty mold differed by 1% from the optimal accuracy, this type of leaf disease and insect pest is relatively easy to separate from the other four types, thus having little impact on the overall detection effect. Therefore, this article proposes adding the CMPA to the last three stages of the backbone network G-GhostNet-3.2 to provide a better input characteristic graph for FPN.

3.1.3. Comparison of Improved Loss Functions with Different Hyperparameters

In order to verify the effectiveness of the improved loss function and explore the optimal combination of different hyperparameters that are difficult to distinguish between sample weights in the improved loss function, the comparison results are shown in Table 4. First, the weight of difficultly divided samples α IoU γ was fixed. A γ value of 1 indicates that the weight of difficultly divided samples only varied linearly, without influencing the size of the IoU. Thus, a smaller value of 1.5 was considered to compare the differences in α. Then, the value of α was fixed to compare differences in γ. During the test, when the detection performance of the target detection model FCOS improved to a certain extent and started to decline, the test was stopped. A higher value of γ was achieved when α increased to 4, after which there was a downward trend in detection performance.
From Table 6, it can be seen that, when α = 3, AP and AP75 reached the highest level, whereas AP50 and recall began to show a downward trend. When α = 4, all indicators began to decline. When α = 2, although AP50 and recall reached the highest level, overall superior results were obtained when α = 3, at which point the maximum improvement in FCOS could be achieved with AP50 reaching a maximum of 89.4%. Therefore, in this paper, the hard-to-distinguish sample weight hyperparameters of the loss function α and γ were set to 3.

3.1.4. Comparison of Different Improvement Methods

The effectiveness of the improved algorithm’s introduction of the CMPA attention mechanism module and the loss function optimization method was verified by conducting ablation experiments on the improved FCOS algorithm. The test results are shown in Table 7.
As can be seen from Table 7, after adding the four improvements, the model AP50 increased to 91.3%. Specifically, the accuracy rate of the FCOS model that only replaced the backbone network with G-GhostNet-3.2 increased by 0.9 percentage points. After adding the four improvements, the accuracy rate of the model reached 91.3%, i.e., 2.9 percentage points higher than the original model, verifying the effectiveness of the improved method.

3.2. Comparison of Detection Effects between FCOS and FCOS-FL

The experimental data were collected in four weather conditions: direct sunlight, backlight, after rain, and overcast days. The collected image data were experimentally compared using the original FCOS and FCOS-FL, and the results are shown in Figure 10. In Figure 10, the yellow box is marked as missed inspection. The red box in Figure 10a shows Litchi leaf mite. The purple block box in Figure 10b shows Litchi algal spot, the light blue box in Figure 10c shows Mayetiola sp, and the dark blue box in Figure 10d shows Litchi anthracnose. As can be seen from Figure 10, the original FCOS model had different degrees of missed detection when detecting various pests and diseases. When exposed to direct sunlight, Litchi leaf mites could cause FCOS to miss detection due to interference from light spots, whereas FCOS-FS could eliminate the interference from light spots for accurate detection (Figure 10a). For Mayetiola sp. and Litchi algal spot (Figure 10b,c), FCOS experienced a large number of missed detections due to their extremely small and dense disease volume, whereas FCOS-FL performed excellently in detecting these two small target pests and diseases. In addition, FCOS-FL could accurately identify multiple pests and diseases on a single leaf (Figure 10d).

3.3. Performance Comparison of Different Models

To verify the effectiveness of the FCOS-FL model proposed in this article, comparative tests were conducted with faster R-CNN, RetinaNet, VFNet, and the original FCOS model under the same training conditions. The comparison of detection performance is shown in Table 8.
As can be seen from Table 8, compared to faster R-CNN, RetinaNet, and VFNet, the accuracy of FCOS-FL in terms of AP50 increased by 16.1, 14.2, and 19.6 percentage points, respectively, with respect to model detection accuracy, while it increased by 4.2 percentage points compared to the original FCOS. In terms of detection speed and model parameter size, FCOS-FL improved the detection rate by 19.1% and reduced the model parameter size by 45% compared to the original FCOS. In FCOS-FL, the detection accuracy of four pests and diseases (Litchi sooty mold, Litchi anthracnose, Mayetiola sp., and Litchi algal spot) was the best. For Mayetiola sp. and Litchi algal spot, with high detection difficulty, FCOS-FL improved the detection accuracy by 9.4 and 11 percentage points compared to the original FCOS, respectively, demonstrating the superiority of FCOS-FL in the detection of small target pests and diseases. The above comparison verifies that FCOS-FL achieved a lightweight model while improving detection accuracy.

4. Discussion

Litchi diseases and pests can reduce yield, affect the quality of Litchi, and reduce the benefits of farmers. However, there are many types of Litchi diseases and pests, and their characteristics are small, making detection difficult. Currently, the main method for detecting Litchi diseases and pests in orchards is still visual identification, but deep learning target detection methods have been widely applied to Litchi detection [40,41,42,43]. However, research on the application of deep learning target detection methods to Litchi diseases’ and pests’ detection is relatively scarce. In view of these problems, this paper further optimized the model structure based on the advantages of existing research at the network design level, and proposed a strong targeted model optimization method for the characteristics of the five Litchi leaf diseases and pests. The FCOS-FL model was experimentally proven to be superior in terms of accuracy and lightweight performance compared to the pre-improved model on the self-built Litchi leaf diseases and pests dataset. This method can meet the real-time detection needs of Litchi leaf diseases and pests, and the model can be deployed on embedded resource-limited devices such as mobile terminals to achieve real-time recognition of Litchi leaf diseases and pests, providing a new solution for the prevention and control of Litchi diseases and pests. The overall experimental conclusions are as follows:
(1)
Regarding the lack of research on deep learning in Litchi diseases and pests and the low detection accuracy of the existing FCOS model in detecting Litchi leaf diseases and pests, the FCOS-FL model proposed in this paper adds an attention mechanism and redesigns the network structure to achieve an average detection accuracy of 91.3% on the detection of five types of Litchi leaf diseases and pests under different natural environments on the AP50, which is 4.2% higher than the detection accuracy of the original FCOS model. Among them, the detection accuracy of the small target pests with high detection difficulty, Mayetiola sp. and Litchi algal spot, reached 93.2% and 92.0%, respectively. It proves that the detection accuracy of the proposed FCOS-FL model is sufficient for detecting five types of Litchi leaf diseases and pests.
(2)
To solve the problem of the large parameter size of the original FCOS model, G-GhostNet-3.2 was used as the backbone network to achieve a model that was lightweight. The parameter size of the FCOS-FL model proposed in this paper is only 17.65 M, which is 45% smaller than the parameter size of the original FCOS model. The single-image detection speed is 62.0 ms, which proves that the proposed model is suitable for deployment on embedded resource-limited devices such as mobile terminals for real-time and fast recognition and detection of Litchi diseases and pests.
(3)
The proposed FCOS-FL model outperforms other commonly used models, including Faster R-CNN, RetinaNet, VFNet, and the original FCOS model, in terms of detection accuracy and model parameter size. These results show that the FCOS-FL model has higher accuracy and smaller model parameters, providing more reliable support for real-time and accurate detection of the five types of Litchi diseases and pests.
(4)
In summary, the Litchi diseases and pests detection model FCOS-FL proposed in this paper can meet the real-time recognition of Litchi diseases and pests in practical applications. The recognition effect and model lightweight degree perform well on the self-built Litchi diseases and pests dataset, which can provide a reference for the prevention and control of Litchi diseases and pests. Compared with other Litchi diseases and pests detection studies, the model and optimization methods used in this study are more in line with the disease characteristics, and the constructed dataset has more complete types of diseases. However, there are still shortcomings in this study. The dataset collected in this study still has room for improvement, and future research should collect more varieties and full-cycle Litchi images to supplement the dataset. There are still a small number of missed detections during the detection process, and further optimization of the detection model should be performed to improve its accuracy. Future work should consider deploying the model to edge computing devices for field applications.

Author Contributions

Conceptualization, J.X., X.Z. and Z.L.; methodology, J.X., X.Z. and Z.L.; software, X.Z. and Z.L.; validation, X.Z., Z.L. and F.L.; formal analysis, F.L. and Z.L.; investigation, J.X., X.Z. and Z.L.; resources, J.X. and W.W.; data curation, F.L.; writing—original draft preparation, J.X. and X.Z.; writing—review and editing, J.X.; visualization, X.Z. and Z.L.; supervision, W.W. and J.L.; project administration, W.W. and J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Co-constructing Cooperative Project on Agricultural Sci-tech of New Rural Development Research Institute of South China Agricultural University (No. 2021XNYNYKJHZGJ032). It was also partly supported by the China Agriculture Research System of MOF and MARA, China (No. CARS-32-11); the Guangdong Provincial Special Fund for Modern Agriculture Industry Technology Innovation Teams, China (No. 2023KJ108); the Laboratory of Lingnan Modern Agriculture Project, China (No. NT2021009); the Guangdong Province Rural Revitalization Strategy Projects (No. TS-1-4); and the Guangdong Science and Technology Innovation Cultivation Special Fund Project for College Students (“Climbing Program” Special Fund), China (No. pdjh2023a0074 and No. pdjh2021b0077).

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to the privacy policy of the organization.

Acknowledgments

The authors would like to thank the anonymous reviewers for their critical comments and suggestions for improving the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Qi, W.N.; Chen, H.B.; Li, J.X.; Kumar, M. Status, Trend and Countermeasures of Development of Litchi lndustry in theMainland of China in 2022. Guangdong Agric. Sci. 2023, 50, 147–155. [Google Scholar]
  2. Punia, S.; Kumar, M. Litchi (Litchi Chinenis) Seed: Nutritional Profile, Bioactivities, and Its Industrial Applications. Trends Food Sci. Technol. 2021, 108, 58–70. [Google Scholar] [CrossRef]
  3. Zhuang, L.; Qiu, Z. Development Characteristics and Policy Suggestions of Chinese Litchi Industry in 2019. South China Fruits 2021, 50, 184–188. [Google Scholar]
  4. Ling, J.F.; Song, X.B.; Xi, P.G.; Cheng, B.P.; Cui, Y.P.; Chen, X.; Peng, A.T.; Jiang, Z.D.; Zhang, L.H. Identification of Colletotrichum Siamense Causing Litchi Pepper Spot Disease in Mainland China. Plant Pathol. 2019, 68, 1533–1542. [Google Scholar] [CrossRef]
  5. Ferentinos, K.P. Deep Learning Models for Plant Disease Detection and Diagnosis. Comput. Electron. Agric. 2018, 145, 311–318. [Google Scholar] [CrossRef]
  6. Barman, U.; Choudhury, R.D.; Sahu, D.; Barman, G.G. Comparison of Convolution Neural Networks for Smartphone Image Based Real Time Classification of Citrus Leaf Disease. Comput. Electron. Agric. 2020, 177, 105661. [Google Scholar] [CrossRef]
  7. Nagar, H.; Sharma, R.S. A Comprehensive Survey on Pest Detection Techniques Using Image Processing. In Proceedings of the 2020 4th International Conference on Intelligent Computing and Control Systems (ICICCS), Madurai, India, 13–15 May 2020; pp. 43–48. [Google Scholar]
  8. Ebrahimi, M.A.; Khoshtaghaza, M.H.; Minaei, S.; Jamshidi, B. Vision-Based Pest Detection Based on SVM Classification Method. Comput. Electron. Agric. 2017, 137, 52–58. [Google Scholar] [CrossRef]
  9. Kasinathan, T.; Singaraju, D.; Uyyala, S.R. Insect Classification and Detection in Field Crops Using Modern Machine Learning Techniques. Inf. Process. Agric. 2021, 8, 446–457. [Google Scholar] [CrossRef]
  10. Wen, C.; Guyer, D.E.; Li, W. Local Feature-Based Identification and Classification for Orchard Insects. Biosyst. Eng. 2009, 104, 299–307. [Google Scholar] [CrossRef]
  11. Wen, C.; Guyer, D. Image-Based Orchard Insect Automated Identification and Classification Method. Comput. Electron. Agric. 2012, 89, 110–115. [Google Scholar] [CrossRef]
  12. Singh, C.B.; Jayas, D.S.; Paliwal, J.; White, N.D.G. Detection of Insect-Damaged Wheat Kernels Using near-Infrared Hyperspectral Imaging. J. Stored Prod. Res. 2009, 45, 151–158. [Google Scholar] [CrossRef]
  13. Liu, T.; Chen, W.; Wu, W.; Sun, C.; Guo, W.; Zhu, X. Detection of Aphids in Wheat Fields Using a Computer Vision Technique. Biosyst. Eng. 2016, 141, 82–93. [Google Scholar] [CrossRef]
  14. Sabrol, H.; Satish, K. Tomato Plant Disease Classification in Digital Images Using Classification Tree. In Proceedings of the 2016 International Conference on Communication and Signal Processing (ICCSP), Melmaruvathur, Tamilnadu, India, 6–8 April 2016; pp. 1242–1246. [Google Scholar]
  15. Pydipati, R.; Burks, T.F.; Lee, W.S. Identification of Citrus Disease Using Color Texture Features and Discriminant Analysis. Comput. Electron. Agric. 2006, 52, 49–59. [Google Scholar] [CrossRef]
  16. Mao, R.; Wang, Z.; Li, F.; Zhou, J.; Chen, Y.; Hu, X. GSEYOLOX-s: An Improved Lightweight Network for Identifying the Severity of Wheat Fusarium Head Blight. Agronomy 2023, 13, 242. [Google Scholar] [CrossRef]
  17. LeCun, Y.; Bengio, Y.; Hinton, G. Deep Learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  18. LeCun, Y.; Kavukcuoglu, K.; Farabet, C. Convolutional Networks and Applications in Vision. In Proceedings of the 2010 IEEE International Symposium on Circuits and Systems, Paris, France, 30 May–2 June 2010; pp. 253–256. [Google Scholar]
  19. Zhang, W.; Xia, X.; Zhou, G.; Du, J.; Chen, T.; Zhang, Z.; Ma, X. Research on the Identification and Detection of Field Pests in the Complex Background Based on the Rotation Detection Algorithm. Front. Plant Sci. 2022, 13. [Google Scholar] [CrossRef] [PubMed]
  20. Li, L.; Zhang, S.; Wang, B. Plant Disease Detection and Classification by Deep Learning—A Review. IEEE Access 2021, 9, 56683–56698. [Google Scholar] [CrossRef]
  21. Lyu, S.; Ke, Z.; Li, Z.; Xie, J.; Zhou, X.; Liu, Y. Accurate Detection Algorithm of Citrus Psyllid Using the YOLOv5s-BC Model. Agronomy 2023, 13, 896. [Google Scholar] [CrossRef]
  22. Zhang, Y.; Yang, G.; Liu, Y.; Wang, C.; Yin, Y. An Improved YOLO Network for Unopened Cotton Boll Detection in the Field. J. Intell. Fuzzy Syst. 2022, 42, 2193–2206. [Google Scholar] [CrossRef]
  23. Liu, L.; Wang, R.; Xie, C.; Yang, P.; Wang, F.; Sudirman, S.; Liu, W. PestNet: An End-to-End Deep Learning Approach for Large-Scale Multi-Class Pest Detection and Classification. IEEE Access 2019, 7, 45301–45312. [Google Scholar] [CrossRef]
  24. Dai, M.; Dorjoy, M.M.H.; Miao, H.; Zhang, S. A New Pest Detection Method Based on Improved YOLOv5m. Insects 2023, 14, 54. [Google Scholar] [CrossRef] [PubMed]
  25. Zhang, Y.; Liu, Y.P. Identification of Navel Orange Diseases and Pests Based on the Fusion of Densenet and Self-Attention Mechanism. Comput. Intell. Neurosci. 2021, 2021, 5436729. [Google Scholar] [CrossRef]
  26. Liu, J.; Wang, X.; Liu, G. Tomato Pests Recognition Algorithm Based on Improved YOLOv4. Front. Plant Sci. 2022, 1894. [Google Scholar] [CrossRef] [PubMed]
  27. Anal, A.K.D.; Kumar, V.; Tripathi, M.; Nath, V. Important Diseases of Litchi and Their Management. Lychee Dis. Manag. 2017, 89–97. [Google Scholar] [CrossRef]
  28. Kumar, V. Diseases of Litchi and Their Management. In Diseases of Fruits and Vegetable Crops; Apple Academic Press: Waretown, New Jersey, 2020; pp. 107–124. [Google Scholar]
  29. Xiao, Z. Green control technology of Litchi diseases and insect pests. Agric. Dev. Equip. 2021, 233–234. [Google Scholar]
  30. Zhou, Z. Green control technology of Litchi diseases and insect pests. South China Agric. 2021, 15, 26–27. [Google Scholar]
  31. Tian, Z.; Shen, C.; Chen, H.; He, T. Fcos: Fully Convolutional One-Stage Object Detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9627–9636. [Google Scholar]
  32. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  33. Lin, T.-Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  34. Han, K.; Wang, Y.; Xu, C.; Guo, J.; Xu, C.; Wu, E.; Tian, Q. GhostNets on Heterogeneous Devices via Cheap Operations. Int. J. Comput. Vis. 2022, 130, 1050–1069. [Google Scholar] [CrossRef]
  35. Wu, T.; Li, B.; FEI, S.; LIAN, D. Anti-Vibration Hammer Detection of overhead transmission lines based on improved FCOS algorithm. J. Electr. Eng. 2022, 1–10. [Google Scholar]
  36. Jia, S. Research on Multi-scale Object Detection Method Based on Anchor Free. Master’s Thesis, Xidian University, Xi’an, China, 2021. [Google Scholar]
  37. Lai, R. Detection of Extreme Size Defects on MetalSurface Based on Advanced-FCOS. Master’s Thesis, Huazhong University of Science and Technology, Wuhan, China, 2021. [Google Scholar]
  38. Xiong, R.; Yang, Y.; He, D.; Zheng, K.; Zheng, S.; Xing, C.; Zhang, H.; Lan, Y.; Wang, L.; Liu, T. On Layer Normalization in the Transformer Architecture. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual Event, 13–18 July 2020; pp. 10524–10533. [Google Scholar]
  39. Ruder, S. An Overview of Gradient Descent Optimization Algorithms. arXiv 2016, arXiv:1609.04747. [Google Scholar]
  40. Liang, C.; Xiong, J.; Zheng, Z.; Zhong, Z.; Li, Z.; Chen, S.; Yang, Z. A Visual Detection Method for Nighttime Litchi Fruits and Fruiting Stems. Comput. Electron. Agric. 2020, 169, 105192. [Google Scholar] [CrossRef]
  41. Jiao, Z.; Huang, K.; Jia, G.; Lei, H.; Cai, Y.; Zhong, Z. An Effective Litchi Detection Method Based on Edge Devices in a Complex Scene. Biosyst. Eng. 2022, 222, 15–28. [Google Scholar] [CrossRef]
  42. Peng, H.; Xue, C.; Shao, Y.; Chen, K.; Liu, H.; Xiong, J.; Chen, H.; Gao, Z.; Yang, Z. Litchi Detection in the Field Using an Improved YOLOv3 Model. Int. J. Agric. Biol. Eng. 2022, 15, 211–220. [Google Scholar] [CrossRef]
  43. Yu, L.; Xiong, J.; Fang, X.; Yang, Z.; Chen, Y.; Lin, X.; Chen, S. A Litchi Fruit Recognition Method in a Natural Environment Using RGB-D Images. Biosyst. Eng. 2021, 204, 50–63. [Google Scholar] [CrossRef]
Figure 1. LabelImg annotation software interface diagram.
Figure 1. LabelImg annotation software interface diagram.
Agronomy 13 01314 g001
Figure 2. The experimental protocol workflow diagram.
Figure 2. The experimental protocol workflow diagram.
Agronomy 13 01314 g002
Figure 3. The FCOS model.
Figure 3. The FCOS model.
Agronomy 13 01314 g003
Figure 4. G-GhostNet low-cost operation.
Figure 4. G-GhostNet low-cost operation.
Agronomy 13 01314 g004
Figure 5. The CMPA model.
Figure 5. The CMPA model.
Agronomy 13 01314 g005
Figure 6. The sampling methods of FCOS.
Figure 6. The sampling methods of FCOS.
Agronomy 13 01314 g006
Figure 7. Improved centerness thermal diagram before and after.
Figure 7. Improved centerness thermal diagram before and after.
Agronomy 13 01314 g007
Figure 8. The FCOS-FL model.
Figure 8. The FCOS-FL model.
Agronomy 13 01314 g008
Figure 9. Loss (left) and AP50 (right) variation in different backbone networks.
Figure 9. Loss (left) and AP50 (right) variation in different backbone networks.
Agronomy 13 01314 g009
Figure 10. Detection results of FCOS and FCOS-Fl under different weather conditions.
Figure 10. Detection results of FCOS and FCOS-Fl under different weather conditions.
Agronomy 13 01314 g010
Table 1. Characteristics of 5 pests and diseases.
Table 1. Characteristics of 5 pests and diseases.
NameCharacteristicsPicture
Litchi leaf miteAt the initial stage of damage to the leaves, small chartreuse spots appear, and then gradually expand to form irregular large patches. The back of the diseased leaves is fluffy, like felt, uneven, yellow at first, then turns to dark brown to reddish brown, and the back of the leaf margin curls.Agronomy 13 01314 i001
Litchi sooty moldA dark brown to dark brown mold layer forms on the surface of the leaves, which thickens into a coal-smoke-like layer. Scattered black small particles on the mold layerAgronomy 13 01314 i002
Litchi anthracnoseBrown spots are produced at the leaf tips and edges, and the brown spots on the leaves are circular in shape, followed by the formation of small black particles.Agronomy 13 01314 i003
Mayetiola sp.By using larvae to grow hoops, the affected spots gradually cause protrusions on the surface of the leaves, forming nodules. In severe cases, there are more than ten or even hundreds of galls on a single leaf, causing the leaf to twist. As the leaves age and dry up, some of them become perforated.Agronomy 13 01314 i004
Litchi algal spotWhen the disease occurs, the leaves have black brown or brick red spots, and there are gray–green or yellow–brown plush-like substances on the disease spots. Spindle-shaped to short-strip-shaped black spots often appear on both sides of the midvein of the leaf, with a gray white center Agronomy 13 01314 i005
Table 2. Dataset of Litchi diseases and pests.
Table 2. Dataset of Litchi diseases and pests.
CategoryTrainingValidationTestTotal
Litchi leaf mite6307979788
Litchi sooty mold5777272721
Litchi anthracnose5677171709
Mayetiola sp.6368080796
Litchi algal spot5697171711
Total29793733733725
Table 3. The configuration of server hardware and software environment.
Table 3. The configuration of server hardware and software environment.
Hardware ConfigurationSoftware Environment
CPU: Inter(R) Xeon(R) Gold 6240 CPU@2.59 GHzOS: Windows10
Memory: 192 GBPycharm2021.3.3 + python 3.8.10
Disk: 4 TBCUDA 11.1 + CuDNN 8.0.5
GPU*2: NVIDIA GeForce GTX 3090TiPyTorch 1.8.0
Table 4. The detection results of different backbone networks.
Table 4. The detection results of different backbone networks.
Backbone NetworkPrecision (%)Time (ms)Params (M)
APAP50AP75
ResNet-5059.487.164.176.632.12
EfficientNet-b25381.554.162.115.03
GhostNet52.182.95549.710.3
MobileNet-v258.787.163.651.510.04
G-GhostNet-3.260.288.466.853.717.62
Note: Precision means the proportion of correct predictions in the positive samples predicted by the classifier. Rate means average time per single image in the test set. Params means the number of model parameters.
Table 5. The detection results of CMPA at different locations.
Table 5. The detection results of CMPA at different locations.
StageRecall (%)AP50 (%)AP50 (%)
Litchi Leaf MiteLitchi Sooty MoldLitchi AnthracnoseMayetiola sp.Litchi Algal Spot
C593.988.383.995.292.187.183.2
C3 + C494.188.084.395.391.986.282.3
C3 + C593.888.884.795.391.688.483.8
C4 + C593.388.684.095.991.787.184.2
C3 + C4 + C593.989.386.394.991.987.685.8
C2 + C3 + C4 + C593.288.784.795.591.587.484.2
Note: Recall means evaluate the percentage of coverage predicted for all actual positive examples. AP50 means the detection accuracy of the model at an IoU threshold of 0.5.
Table 6. The parameter results for the weight of difficult samples.
Table 6. The parameter results for the weight of difficult samples.
AlphaGammaAP (%)AP50 (%)AP75 (%)Recall (%)
1.51.559.588.666.093.1
260.688.967.493.8
362.288.469.593.5
462.288.469.494.0
3261.288.367.593.3
361.289.468.194.1
460.988.667.394.1
Note: AP (average precision) means the area under the P–R curve for a single category. This index can comprehensively indicate the precision and recall of a model. AP75 means the detection accuracy of the model at an IoU threshold of 0.75.
Table 7. The results of different improvement methods.
Table 7. The results of different improvement methods.
Backbone NetworkABCDAP (%)AP50 (%)AP75 (%)Recall (%)
G-GhostNet-3.2 60.288.466.892.4
G-GhostNet-3.2 60.989.366.993.9
G-GhostNet-3.2 60.289.067.694.1
G-GhostNet-3.2 61.289.468.194.1
G-GhostNet-3.2 63.090.772.795.0
G-GhostNet-3.265.191.373.296.2
A: CMPA was added to the last three stages of G-GhostNet-3.2; B: improved center sampling and center measurement with width and height adaptability; C: modified loss function with the weight of difficult samples; D: adjusted detection layer of FCOS.
Table 8. The detection results of different networks.
Table 8. The detection results of different networks.
Network ModelTime (ms)Params (M)AP50 (%)AP50 (%)
Litchi Leaf MiteLitchi Sooty MoldLitchi AnthracnoseMayetiola sp.Litchi Algal Spot
Faster R-CNN86.141.1475.282.594.290.547.856.5
RetinaNet80.136.1977.183.193.688.651.868.2
VFNet84.332.4971.772.989.286.451.059.0
FCOS76.632.1287.185.495.390.483.881.0
FCOS-FL62.017.6591.384.895.790.993.292.0
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xie, J.; Zhang, X.; Liu, Z.; Liao, F.; Wang, W.; Li, J. Detection of Litchi Leaf Diseases and Insect Pests Based on Improved FCOS. Agronomy 2023, 13, 1314. https://doi.org/10.3390/agronomy13051314

AMA Style

Xie J, Zhang X, Liu Z, Liao F, Wang W, Li J. Detection of Litchi Leaf Diseases and Insect Pests Based on Improved FCOS. Agronomy. 2023; 13(5):1314. https://doi.org/10.3390/agronomy13051314

Chicago/Turabian Style

Xie, Jiaxing, Xiaowei Zhang, Zeqian Liu, Fei Liao, Weixing Wang, and Jun Li. 2023. "Detection of Litchi Leaf Diseases and Insect Pests Based on Improved FCOS" Agronomy 13, no. 5: 1314. https://doi.org/10.3390/agronomy13051314

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop