Next Article in Journal
Effect of Nitrogen Nutrition and Planting Date on the Yield and Physicochemical Parameters of Flowering Chinese Cabbage
Next Article in Special Issue
Development of Weed Detection Method in Soybean Fields Utilizing Improved DeepLabv3+ Platform
Previous Article in Journal
Smallholders’ Water Management Decisions in the Face of Water Scarcity from a Socio-Cognitive Perspective, Case Study of Viticulture in Mendoza
Previous Article in Special Issue
Determining Strawberries’ Varying Maturity Levels by Utilizing Image Segmentation Methods of Improved DeepLabV3+
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Faster R-CNN-Based Model for the Identification of Weed Seedling

1
College of Information Technology, Jilin Agricultural University, Changchun 130118, China
2
Jilin Province Agricultural Internet of Things Technology Collaborative Innovation Center, Changchun 130118, China
3
Jilin Province Intelligent Environmental Engineering Research Center, Changchun 130118, China
4
Jilin Province Colleges and Universities and the 13th Five-Year Engineering Research Center, Changchun 130118, China
5
School of Management, Changchun University, Changchun 130022, China
6
College of Information Technology, Wuzhou University, Wuzhou 543003, China
7
Guangxi Key Laboratory of Machine Vision and Inteligent Control, Wuzhou 543003, China
8
College of Agriculture, Jilin Agricultural University, Changchun 130118, China
*
Author to whom correspondence should be addressed.
Agronomy 2022, 12(11), 2867; https://doi.org/10.3390/agronomy12112867
Submission received: 6 October 2022 / Revised: 6 November 2022 / Accepted: 11 November 2022 / Published: 16 November 2022

Abstract

:
The accurate and rapid acquisition of crop and weed information is an important prerequisite for automated weeding operations. This paper proposes the application of a network model based on Faster R-CNN for weed identification in images of cropping areas. The feature pyramid network (FPN) algorithm is integrated into the Faster R-CNN network to improve recognition accuracy. The Faster R-CNN deep learning network model is used to share convolution features, and the ResNeXt network is fused with FPN for feature extractions. Tests using >3000 images for training and >1000 images for testing demonstrate a recognition accuracy of >95%. The proposed method can effectively detect weeds in images with complex backgrounds taken in the field, thereby facilitating accurate automated weed control systems.

1. Introduction

Weeds include all kinds of herbaceous plants that grow where they are not wanted [1,2]. Weeds cause much harm to the agricultural economy [3]. Not only do they compete with crops for sunlight, water and fertilizer, but they also compete for living space. If not dealt with in time, weeds will reduce crop growth, yield and quality, and can even cause crop failure. With the development of computer technology, rapid and accurate machine vision-based recognition technology has been increasingly used for weed identification [4,5,6,7]. Deep learning has achieved good results in the recognition of human behaviour, crop fruits and weeds. The image recognition method of kiwifruit based on a convolutional neural network (CNN) shows that CNNs have good application prospects in field fruit recognition. Recognition of corn weeds based on CNNs, hash codes and multi-scale hierarchical features has proven the effectiveness of CNNs in recognising weeds in field images. In a study on broccoli seedlings, a crop detection method based on the Faster R-CNN model was proposed. Using a dropout value of 0.6, the ResNet101 network was used as the feature extraction network, and an average precision of 91.73% was achieved. At present, the Faster R-CNN model is also widely applied in the field of vehicle detection [8], ground object recognition in remote sensing images [9,10], appearance defect detection [11], pedestrian detection and recognition [12], and field image detection [13], and has excellent recognition accuracy.
Since most current models have the shortcomings of high complexity and difficulty in modification, an improved Faster R-CNN field weed detection model is proposed in this study. This network model was fused with the Feature Pyramid Network (FPN) to improve the detection accuracy of the model in weed identification. On the basis of obtaining a large amount of weed and plant seedling data, a deep learning network is established, and the ResNeXt feature extraction network is used to generate target recognition models for different types of weeds. After training, it can effectively identify weeds and plant seedlings in images. In this way, a Faster R-CNN-based rapid weed identification system for use in the field is obtained, which can identify weeds in multiple crops.

2. Materials and Methods

2.1. Image Data Acquisition

Image data were obtained from the V2 Plant Seedlings Dataset [14]. This dataset consists of images of nine field weeds—scentless mayweed, common chickweed, shepherd’s purse, cleavers, charlock, fat hen, small-flowered cranesbill, black-grass, and loose silky-bent—and three crop seedlings—maize, common wheat, and sugar beet (Figure 1). To ensure experimental accuracy, the images were collected under different conditions, including sunny, cloudy, and rainy days. There were 5539 images in total, including 598 of scentless mayweed, 713 of common chickweed, 274 of shepherd’s purse, 335 of cleavers, 452 of charlock, 538 of fat hen, 576 of small-flowered cranesbill, 309 of black-grass, 762 of loose silky-bent, 257 of maize, 253 of common wheat, and 463 of sugar beet. The resolution of the images after processing was 227 × 227 pixels. All processed images were divided as follows: 3329 in a training set, 1107 in a validation set, and 1103 in a testing set. The pictures in these sets did not overlap each other. The training set was used to train the model parameters, the test set was used to evaluate the generalization error of the model applied to the samples after training, and the validation set was used to tune the hyperparameters of the model during the training process. After the experiment was completed, deep learning was used to apply evaluation indicators typically used in the field of object recognition to measure the performance of the detector model.
This dataset is due to the use of public datasets, so there are some limitations. Because the pictures in the dataset exist as a single crop, they cannot be used for large-area crop identification between fields. Moreover, because the dataset collection site is under laboratory conditions, a large amount of sand and gravel is selected for plant cultivation, so the detection accuracy of soil-growing plants cannot be guaranteed.

2.2. Image Pre-Processing

Since the images were affected by the environmental conditions at the time of collection, extreme highlights and shadows could decrease the model’s recognition success rate. Therefore, the images were processed and segmented; the green part of the plant was extracted and non-green areas were suppressed. The application of normalized colour components can effectively improve the effects of lighting and shadows on image quality [15,16]. After the greyscale images were obtained, the Otsu method [17,18] was used to convert them into binary images and separate the plants. After processing, clear plant images were obtained. They were then filled with noise to obtain single-leaf binary images after segmentation image processing, as shown in Figure 2.

3. Model Refinement

Currently, there are two types of object detection models. The first is one complete object recognition and object positioning in two steps. Typical representatives of this approach are the R-CNN, Fast R-CNN, and Faster R-CNN families. This type of framework has a low recognition error rate and a low omission rate and can be used in real-time detection scenarios. The typical representative of the second category are YOLO (you only look once), SSD (Single Shot Multibox Detector), YOLOv2, and YOLOv3. To complete object classification and object localization in one step [19,20,21], SSD directly regresses the position and category of the target at the output layer. Although these methods have fast recognition speed, their accuracy rate is lower than that of Faster R-CNN. Therefore, this paper selected the Faster R-CNN model framework to identify weed images.
The network structure of Faster R-CNN is shown in Figure 3. The network can be roughly divided into four parts: (1) a feature extraction layer, (2) a Region Proposal Network (RPN), (3) a Region of Interest pooling (ROI pooling) layer, and (4) classification and regression.
The deep network in the convolutional network is used to respond to the semantics, and the shallow layer responds to the image. However, in object detection, because the feature map size is too small, the high-level network can respond to the semantic features but is not conducive to object detection [22]. Therefore, a Feature Pyramid Network (FPN) is introduced to improve the weed detection accuracy of the algorithm in the field identification process. The structure of the FPN network is shown in Figure 4.
Bottom-to-top process: It can be seen that the bottom-to-top extraction process is a conventional feature extraction process. The output of the last layer of each stage is selected as the input of the feature map. The first layer is not used here due to its large memory footprint.
Top-down process: Upsampling starting from the highest layer yields better feature maps, and nearest-neighbour 2× upsampling is used for simplicity.
Horizontal connection process: Each horizontal connection fuses feature maps of the same spatial size, specifically, the upsampling result and the bottom-up C2, C3, C4, and C5 layers are fused with the same-sized feature maps generated by 1 × 1 convolution. The output channels are all set to the same 256 channels.
The ResNeXt target extraction network and FPN are selected for fusion in the Faster R-CNN network. This is used to extract target features in target detection so that the learning of the target features is more complete.
The building blocks of ResNet and ResNeXt are shown in Figure 5. While the ResNeXt network structure retains the basic stacking method of ResNet, it splits its single path into 32 independent paths, which perform convolution operations on the input image at the same time. Finally, the cumulative summation of outputs from different paths is used as the final result. This operation makes the division of labour of the network clearer and the local adaptability stronger. Since each path shares the same topology and convolution parameters, and the design method is the same, the network parameters will not increase, which is convenient for model transplantation.
The improved model is shown in Figure 6. At present, there are mainly two training methods for Faster R-CNN: alternating training and approximate joint training. The alternate training method trains two networks—RPN and Fast-RCNN—in a total of two stages, with each stage training the RPN and Fast-RCNN once. In the approximate joint training process, only one weighted network is trained, which requires slightly less memory. This training method saves 25–50% of the training time compared with the alternate training method, and the two training methods have similar accuracy, so the approximate joint training method was selected.
First, the weed image is input into the model and the improved ResNeXt network is used to fuse the FPN network to extract features from the image. The overall design of FPN is presented in a multi-scale pyramid structure, and each layer of FPN corresponds to {P2, P3, P4, P5, P6} in ResNeXt pyramid through anchors, as shown in Figure 7. Using 3 proportions {1:2, 1:1, 2:1}, 15 types of anchors were used to predict the target objects in weed images in the field. After the feature map is obtained, it is input into the RPN network and propagated forward to obtain a higher-dimensional feature map. When the feature map is passed to the RPN network, the proposed box is obtained and a non-maximum suppression value operation is performed on the proposed box. The top N highest-scoring proposal boxes are used as ROIs. The feature map and ROIs are passed to the ROI pooling layer for a pooling operation. Then, the ROI pooling layer is output to the full link layer and linear regression is performed on each area of interest to obtain accurate detection of weeds. Then, the Softmax regression model is used to perform multi-classification target detection and, finally, the categories of weeds and crop seedlings in different fields can be identified.

4. Experimental Results and Analysis

4.1. Test Platform

The operating environment for the test was a desktop computer with a Windows 10 64-bit operating system, and the computer memory was a 32 GB, NVIDIA GeForce RTX 3070 graphics card, equipped with an AMD Ryzen 7 5800H processor.
The software environments used are Anaconda 3.9.12 (Developed by Anaconda, Inc. in Austin, Texas, USA), Python 3.7.9, CUDA 11.2 (Developed by NVIDIA Corporation of California, USA), and cuDNN 8.2.1 (Developed by NVIDIA Corporation of California, USA). The open-source deep learning framework Tensorflow 2.0 (Developed by Google, Inc. in California, USA) was used as the development environment.

4.2. Parameter Design

To improve the performance of the model and reduce overfitting, the pre-trained model was used to initialize the parameters and the stochastic gradient descent (SGD) method was used to improve the model. The learning rate was set to 0.001, the momentum factor was set to 0.9, the epoch was set to 1500, the maximum number of iterations was 200,000, and the learning rate was adjusted to 0.0001 after 80,000 iterations and to 0.00001 after 160,000 iterations. After the model calculation was completed, the model with the highest accuracy was selected for weed identification. The non-maximum suppression (NMS) value was set to 0.3 to obtain the best candidate box. Finally, to obtain the trained network model, the test set was used to further verify the modelling effect, and the recognition result was the output.

4.3. Evaluating Indicator

In this paper, two evaluation indexes, accuracy (P) and recall (R), were used to verify whether the model can be used for image recognition of weed seedlings. The range of both is [0, 1]. In addition, the F1 value was used to harmonize the average evaluation of the calculation results [23], and the Mean Intersection over Union (MIoU) was used to evaluate the image segmentation results. The assessment is calculated as follows:
P = T P T P + F P   ×   100 % ,
R = T P T P + F N   ×   100 % ,
F 1 = 2 P R P + R   ×   100 % ,
M I o U = 1 k + 1 i = 0 k T P F N + F P + T P
where P stands for accuracy for weed image recognition; R is a recall for weed image recognition; F1 is the harmonic mean of P and R; T P is the true positive rate (number of correctly identified crop seedlings and weed targets); F P is the number of incorrectly identified crop seedlings and weed targets; F N is the number of unidentified crop seedlings and weed targets; k is how many categories there are in the dataset.

5. Results and Analysis

5.1. Model Training Results

According to the above experimental methods, a weed recognition network based on Faster R-CNN fused with FPN was trained and the improved ResNeXt-101 feature extraction network was adopted. The loss degree and accuracy of this model in recognition are shown in Figure 8 and Figure 9.

5.2. Experimental Results

It can be seen from Figure 8 and Figure 9 that the weed recognition network based on Faster R-CNN integrated with FPN has high recognition accuracy when using the ResNeXt-101 feature extraction network. To select a better deep network model and a suitable feature extraction network, this experiment compares the Faster R-CNN integrated with FPN and the ordinary Faster R-CNN network under the condition of the ResNeXt-101 feature extraction network. The results are shown in Table 1.
It can be seen from Table 1 that Faster R-CNN has a higher accuracy rate when it does not integrate the FPN and only uses the ResNeXt-101 feature extraction network. However, with integrated FPN, the accuracy is higher than that of the ordinary Faster R-CNN network, and the F1-value is higher; in addition, the value of MIOU is about 4% higher than the unimproved model. It can be seen that the FPN can optimize the network after being integrated into the Faster R-CNN network, which improves the recognition accuracy.
With the Faster R-CNN deep network model integrated with FPN, compared with the ResNet-50 feature extraction network in the literature [22] that proposed FPN, ResNet-50 and ResNeXt-101 were used for training. With increases in the number of model training iterations, the real-time losses of the overall loss functions of the two feature extraction networks used in the Faster R-CNN deep network model were compared (Figure 10 and Table 2).
Figure 10 shows the losses of the overall loss functions of the two feature extraction networks with the Faster R-CNN deep network model. The Faster R-CNN deep network model based on the ResNeXt-101 feature extraction network can converge faster and is obviously better than the ResNet-50 feature extraction network model. The Faster R-CNN deep network model of the ResNet-50 feature extraction network oscillated to varying degrees after 120,000 iterations. The comparison in Table 2 shows that although the ResNeXt-101 feature extraction network uses 101 layers, there is basically no difference between the operation time, but the F1-value and MIoU value of ResNeXt-101 as the result of the feature extraction network are higher than those of the ResNet-50 feature extraction network; therefore, the ResNeXt-101 network has a higher operation accuracy. Since the ResNeXt network replaces the three-layer convolutional blocks of the original ResNet with blocks of the same topology stacked in parallel, it can improve the accuracy of the model without significantly increasing the number of parameters. Compared with ResNet-50, it can be seen that ResNeXt-101 has certain advantages.
To determine the superiority of the deep network model comprised of the Faster R-CNN integrated with the FPN in weed recognition, the results of the three deep network models and the corresponding feature extraction networks were compared. The recognition performance of the ResNeXt-101 feature extraction network was the best. Therefore, on the premise of satisfying the recognition time, ResNeXt-101 was selected as the feature extraction network to compare the overall loss of the Faster R-CNN model, YOLOv3 model, and SSD model, and the accuracy of model verification. The results are shown in Figure 11 and Figure 12 and Table 3.
The Faster R-CNN, SSD, and YOLOv3 deep network models are compared under the same image processing method and RESNEXT-101 feature extraction network. It can be seen from Figure 11 and Figure 12 that when using the ResNeXt-101 feature extraction network, the SSD network has a slight oscillation at 150,000 steps, while the Faster R-CNN network tends to stabilize after 120,000 steps. The network loss curve of Faster R-CNN with FPN is lower than that of the SSD deep network model after 4000 iterations, and the network model of Faster R-CNN integrated with FPN is better than the SSD model in terms of loss degree. By comparing the accuracy, recall rate and F-value, it can be seen that, compared with the SSD network, the accuracy rate of Faster R-CNN with the integrated FPN network is about 0.7 higher. Although the detection time is longer than that of the SSD network, the overall efficiency is still higher than that of the SSD network.
According to the model evaluation, this experiment makes a detailed comparison of the two indexes of accuracy and recall, F1 value, as well as MIoU and detection time (single image). The comparative results of the deep network model are shown in Table 3. It can be concluded from Table 3 that the Faster R-CNN deep network model based on the ResNext-101 network shows better detection performance, while the accuracy and recall rate of the YOLOv3 deep network model are both somewhat lower than those of the Faster R-CNN network. The accuracy of the deep network model can only reach about 84% and the recall rate can only reach about 80%, the F1 value is about 8% lower than that of Faster R-CNN and the difference in the value of MIoU can reach about 9%, although it has a shorter detection time. Hence, the advantages of the Faster R-CNN network are greater.
The Faster R-CNN network model integrated with FPN was analysed via a confusion matrix (Figure 13). The Faster R-CNN network model integrated with FPN was used for the detection of weeds and plant seedlings. The light-coloured areas show very low FN and PN values, most of which are 0 and a few are 1–4. The dark areas show very high values of TP and TN, which shows that the deep learning network predicts a high proportion of correct results; hence, it has high accuracy.
These model training results show the advantages and disadvantages of the two feature extraction networks (ResNet-50 and ResNeXt-101) in weed recognition. It is found that, compared with the Faster R-CNN deep network model, the YOLOv3 and SSD deep network models have slightly inferior weed recognition accuracy. The Faster R-CNN network model obtains much better accuracy by integration with the FPN network, while the improved ResNeXt feature extraction network has higher computational efficiency. The Faster R-CNN network model integrates the FPN network into the feature extraction network and adopts the process of bottom-to-top, bottom-up, and horizontal connection to realize the simultaneous use of high-resolution low-level features and high-level features. Integrating the features of these different layers to achieve the prediction effect can greatly improve the detection accuracy. The selected ResNeXt-101 feature extraction network achieves more accurate target recognition by using a convolution operation with multiple paths. The experimental results show that the Faster R-CNN deep network model based on the ResNeXt-101 feature extraction network has obvious advantages for weed recognition in field images after being integrated with the FPN network.

6. Conclusions

This paper used corn beet and other crop seedling and weed image data to build a Faster R-CNN deep network model based on the ResNeXt-101 feature extraction network and the Tensorflow 2.0(Developed by Google, Inc. in California, USA) deep learning framework. The FPN network and improved ResNeXt network were applied to the identification of weeds. The actual effects of the three feature extraction networks were compared and analysed, and the key parameters of the Faster R-CNN were optimized to make it more suitable for weed identification in field images with complex backgrounds.
The experimental results show that the Faster R-CNN deep network model obtains improved recognition accuracy by using the ResNeXt feature extraction network and incorporating the FPN network. It has obvious advantages compared with the ResNet feature extraction network in achieving rapid and effective target recognition, and demonstrates the great efficiency of deep learning methods in this field.

Author Contributions

Conceptualization, Y.M. and R.F.; methodology, Y.M.; software, R.F.; validation, R.N., J.L. and T.L. (Tianye Luo); formal analysis, Y.S.; investigation, T.L. (Tonghe Liu); resources, H.G.; data curation, X.L.; writing—original draft preparation, Y.M.; writing—review and editing, S.L. and Y.W.; visualization, Y.B.; supervision, Y.G.; project administration, T.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Jilin Province Science and Technology Development Plan (focuses on research and development projects), funding number 20200402006NC (http://kjt.jl.gov.cn) accessed on 1 January 2020, Key technology R&D project of Changchun Science and Technology Bureau of Jilin Province, funding number 21ZGN29 (http://kjj.changchun.gov.cn) accessed on 1 November 2021 and Science and Technology Research Project of Jilin Provincial Department of Education, funding number JJKH20220337KJ (http://jyt.jl.gov.cn/) accessed on 1 January 2022.

Data Availability Statement

All new research data were presented in this contribution.

Conflicts of Interest

The authors declare that they have no conflict of interest.

References

  1. Petrich, L.; Lohrmann, G.; Neumann, M.; Martin, F.; Frey, A.; Stoll, A.; Schmidt, V. Detection of Colchicum Autumnale in Drone Images, Using a Machine-Learning Approach. Precis. Agric. 2020, 21, 1291–1303. [Google Scholar] [CrossRef]
  2. Osorio, K.; Puerto, A.; Pedraza, C.; Jamaica, D.; Rodríguez, L. A Deep Learning Approach for Weed Detection in Lettuce Crops Using Multispectral Images. AgriEngineering 2020, 2, 471–488. [Google Scholar] [CrossRef]
  3. Sabzi, S.; Abbaspour-Gilandeh, Y. Using Video Processing to Classify Potato Plant and Three Types of Weed Using Hybrid of Artificial Neural Network and Particle Swarm Algorithm. Measurement 2018, 126, 22–36. [Google Scholar] [CrossRef]
  4. Bakhshipour, A.; Jafari, A. Evaluation of Support Vector Machine and Artificial Neural Networks in Weed Detection Using Shape Features. Comput. Electron. Agric. 2018, 145, 153–160. [Google Scholar] [CrossRef]
  5. Wu, G.; Zhu, M.; Jiang, Q.; Sun, X. Fast-Axis Collimating Lens Recognition Algorithm Based on Machine Vision. J. Phys. Conf. Ser. 2021, 1820, 012157. [Google Scholar] [CrossRef]
  6. Wu, Z.; Chen, Y.; Zhao, B.; Kang, X.; Ding, Y. Review of Weed Detection Methods Based on Computer Vision. Sensors 2021, 21, 3647. [Google Scholar] [CrossRef] [PubMed]
  7. Anthoniraj, S.; Karthikeyan, P.; Vivek, V. Weed Detection Model Using the Generative Adversarial Network and Deep Convolutional Neural Network. J. Mob. Multimed. 2021, 18, 275–292. [Google Scholar] [CrossRef]
  8. Cui, F.; Ning, M.; Shen, J.; Shu, X. Automatic Recognition and Tracking of Highway Layer-Interface Using Faster R-CNN. J. Appl. Geophy. 2022, 196, 104477. [Google Scholar] [CrossRef]
  9. Yan, D.; Li, G.; Li, X.; Zhang, H.; Lei, H.; Lu, K.; Cheng, M.; Zhu, F. An Improved Faster R-CNN Method to Detect Tailings Ponds from High-Resolution Remote Sensing Images. Remote Sens. 2021, 13, 2052. [Google Scholar] [CrossRef]
  10. Yujian, C.; Guoming, X.; Guochuan, S. Low Altitude Armored Target Detection Based on Rotation Invariant Faster R-CNN. Laser Optoelectron. Prog. 2018, 55, 101501. [Google Scholar] [CrossRef]
  11. Ma, X.; Zhu, D.; Jin, C.; Tong, X. Bullet Appearance Defect Detection Based on Improved Faster Region-Convolutional Neural Network. Laser Optoelectron. Prog. 2019, 56, 151202. [Google Scholar]
  12. Chen, E.; Tang, X.; Fu, B. Pedestrian Search Method Based on Faster R-CNN with the Integration of Pedestrian Detection and Re-Identification. J. Comput. Aided Des. Comput. Graph. 2019, 31, 332–339. [Google Scholar]
  13. Prasanna Kumar, A.; Mishra, D. Harnessing Feedback Region Proposals for Multi-object Tracking. IET Comput. Vis. 2020, 14, 434–442. [Google Scholar] [CrossRef]
  14. Giselsson, T.M.; Jørgensen, R.N.; Jensen, P.K.; Dyrmann, M.; Midtiby, H.S. A Public Image Database for Benchmark of Plant Seedling Classification Algorithms. arXiv 2017, arXiv:1711.05458. [Google Scholar]
  15. Wang, S.; Yang, L. Feature dimension reduction and category identification of weeds in cotton field based on GA-ANN complex algorithm. J. Henan Agric. Sci. 2018, 47, 148–154, 160. [Google Scholar]
  16. Wang, H.; Yang, J.; Xue, B.; Yan, X.; Tao, J. A Novel Color Calibration Method of Multi-Spectral Camera Based on Normalized RGB Color Model. Results Phys. 2020, 19, 103498. [Google Scholar] [CrossRef]
  17. Liu, S.; Xue, Q.; Li, Z.; Li, C.; Gong, Z.; Li, N. An air temperature predict model based on BP neural networks for solar greenhouse in North China. J. China Agric. Univ. 2015, 20, 176–184. [Google Scholar]
  18. Ranjitha, P.; Shreelakshmi, T.D. A Hybrid Ostu Based Niblack Binarization for Degraded Image Documents. In Proceedings of the 2021 2nd International Conference for Emerging Technology (INCET), Belagavi, India, 21–23 May 2021; IEEE: New York, NY, USA, 2021; pp. 914–920. [Google Scholar]
  19. Xu, Z.-F.; Jia, R.-S.; Sun, H.-M.; Liu, Q.-M.; Cui, Z. Light-YOLOv3: Fast Method for Detecting Green Mangoes in Complex Scenes Using Picking Robots. Appl. Intell. 2020, 50, 4670–4687. [Google Scholar] [CrossRef]
  20. Huang, Z.; Wang, J.; Fu, X.; Yu, T.; Guo, Y.; Wang, R. DC-SPP-YOLO: Dense Connection and Spatial Pyramid Pooling Based YOLO for Object Detection. Inf. Sci. 2020, 522, 241–258. [Google Scholar] [CrossRef] [Green Version]
  21. Kuznetsova, A.; Maleva, T.; Soloviev, V. Detecting Apples in Orchards Using YOLOv3. In Proceedings of the Computational Science and Its Applications—ICCSA 2020, Cagliari, Italy, 1–4 July 2020; Springer International Publishing: Cham, Switzerland, 2020; pp. 923–934. [Google Scholar]
  22. Lin, T.; Dollar, P.; Girshick, R. Feature Pyramid Networks for Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 936–944. [Google Scholar]
  23. Wang, D.D.; He, D.J. Apple Target Recognition of Robot before Fruit Thinning Based on R-FCN Deep Convolution Neural Network. Trans. Chin. Soc. Agric. Eng. 2019, 35, 156–163. [Google Scholar]
Figure 1. Examples of partial samples of (a) black-grass; (b) charlock; (c) maize; (d) scentless mayweed; (e) fat hen; (f) common wheat; (g) sugar beet; and (h) shepherd’s purse.
Figure 1. Examples of partial samples of (a) black-grass; (b) charlock; (c) maize; (d) scentless mayweed; (e) fat hen; (f) common wheat; (g) sugar beet; and (h) shepherd’s purse.
Agronomy 12 02867 g001
Figure 2. Example of single-leaf binary images after partial processing.
Figure 2. Example of single-leaf binary images after partial processing.
Agronomy 12 02867 g002
Figure 3. The Faster R-CNN network structure.
Figure 3. The Faster R-CNN network structure.
Agronomy 12 02867 g003
Figure 4. The FPN network structure diagram.
Figure 4. The FPN network structure diagram.
Agronomy 12 02867 g004
Figure 5. Building blocks of (a) ResNet and (b) ResNeXt.
Figure 5. Building blocks of (a) ResNet and (b) ResNeXt.
Agronomy 12 02867 g005
Figure 6. Improved model plots.
Figure 6. Improved model plots.
Agronomy 12 02867 g006
Figure 7. Schematic representation of the ResNeXt + FPN structure.
Figure 7. Schematic representation of the ResNeXt + FPN structure.
Agronomy 12 02867 g007
Figure 8. Loss of the Faster R-CNN model incorporating FPN.
Figure 8. Loss of the Faster R-CNN model incorporating FPN.
Agronomy 12 02867 g008
Figure 9. Precision of the Faster R-CNN model incorporating FPN.
Figure 9. Precision of the Faster R-CNN model incorporating FPN.
Agronomy 12 02867 g009
Figure 10. Overall loss comparison with the Faster R-CNN model.
Figure 10. Overall loss comparison with the Faster R-CNN model.
Agronomy 12 02867 g010
Figure 11. Comparison of the overall losses of the Faster R-CNN and SSD models.
Figure 11. Comparison of the overall losses of the Faster R-CNN and SSD models.
Agronomy 12 02867 g011
Figure 12. Comparison of the precision of the Faster R-CNN and SSD models.
Figure 12. Comparison of the precision of the Faster R-CNN and SSD models.
Agronomy 12 02867 g012
Figure 13. Confusion matrix diagram of identification results.
Figure 13. Confusion matrix diagram of identification results.
Agronomy 12 02867 g013
Table 1. Performance comparison of the faster R-CNN-FPN network and ordinary Faster R-CNN.
Table 1. Performance comparison of the faster R-CNN-FPN network and ordinary Faster R-CNN.
Model (With
ResNeXt-101)
Accuracy (%)Recall (%) F 1 -Value
(%)
MIoU
(%)
Detection Time (ms)
Faster R-CNN-FPN95.6187.2691.2493.7330
Faster R-CNN92.485.288.6589.6319
Table 2. Feature extraction network comparison.
Table 2. Feature extraction network comparison.
ModelExtraction
Network
Accuracy (%)Recall (%) F 1 -Value (%) MIoU
(%)
Detection Time (ms)
Faster R-CNN-FPNResNeXt-10195.6187.2691.2493.7330
Faster R-CNN-FPNResNet-5091.6785.1988.3189.5319
Table 3. Deep network model comparison.
Table 3. Deep network model comparison.
ModelExtraction
Network
Accuracy (%)Recall (%) F 1 -Value (%) MIoU
(%)
Detection Time (ms)
Faster R-CNN-FPNResNeXt-10195.6187.2691.2493.7330
YOLOv3 ResNeXt-10184.3479.8582.0384.5215
SSDResNeXt-10188.1279.6983.6985.3203
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Mu, Y.; Feng, R.; Ni, R.; Li, J.; Luo, T.; Liu, T.; Li, X.; Gong, H.; Guo, Y.; Sun, Y.; et al. A Faster R-CNN-Based Model for the Identification of Weed Seedling. Agronomy 2022, 12, 2867. https://doi.org/10.3390/agronomy12112867

AMA Style

Mu Y, Feng R, Ni R, Li J, Luo T, Liu T, Li X, Gong H, Guo Y, Sun Y, et al. A Faster R-CNN-Based Model for the Identification of Weed Seedling. Agronomy. 2022; 12(11):2867. https://doi.org/10.3390/agronomy12112867

Chicago/Turabian Style

Mu, Ye, Ruilong Feng, Ruiwen Ni, Ji Li, Tianye Luo, Tonghe Liu, Xue Li, He Gong, Ying Guo, Yu Sun, and et al. 2022. "A Faster R-CNN-Based Model for the Identification of Weed Seedling" Agronomy 12, no. 11: 2867. https://doi.org/10.3390/agronomy12112867

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop