Next Article in Journal
Regulation of bta-miRNA29d-3p on Lipid Accumulation via GPAM in Bovine Mammary Epithelial Cells
Next Article in Special Issue
Yield Prediction for Winter Wheat with Machine Learning Models Using Sentinel-1, Topography, and Weather Data
Previous Article in Journal
Modelling Vineyard Spraying by Precisely Assessing the Duty Cycles of a Blast Sprayer Controlled by Pulse-Width-Modulated Nozzles
Previous Article in Special Issue
BAG: A Linear-Nonlinear Hybrid Time Series Prediction Model for Soil Moisture
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On Precision Agriculture: Enhanced Automated Fruit Disease Identification and Classification Using a New Ensemble Classification Method

1
Department of Management Information Systems, College of Business Administration, King Faisal University, Al Ahsa 31982, Saudi Arabia
2
Department of Human and Digital Interface, Woosong University, Daejeon 34606, Republic of Korea
3
Department of Information Systems, College of Computer Sciences and Information Technology, King Faisal University, Al Ahsa 31982, Saudi Arabia
*
Authors to whom correspondence should be addressed.
Agriculture 2023, 13(2), 500; https://doi.org/10.3390/agriculture13020500
Submission received: 15 January 2023 / Revised: 6 February 2023 / Accepted: 16 February 2023 / Published: 20 February 2023
(This article belongs to the Special Issue Applications of Data Analysis in Agriculture)

Abstract

:
Fruits are considered among the most nutrient-dense cash crops around the globe. Since fruits come in different types, sizes, shapes, colors, and textures, the manual classification and disease identification of a large quantity of fruit is time-consuming and sluggish, requiring massive human intervention. We propose a multilevel fusion method for fruit disease identification and fruit classification that includes intensive fruit image pre-processing, customized image kernels for feature extraction with state-of-the-art (SOTA) deep methods, Gini-index-based controlled feature selection, and a hybrid ensemble method for identification and classification. We noticed certain limitations in the existing literature of adopting a single data source, in terms of limited data sizes, variability in fruit types, variability in quality, and variability in disease type. Therefore, we extensively aggregated and pre-processed multi-fruit data to simulate our proposed ensemble model on comprehensive datasets to cover both fruit classification and disease identification aspects. The multi-fruit imagery data contained regular and augmented images of fruits including apple, apricot, avocado, banana, cherry, fig, grape, guava, kiwi, mango, orange, peach, pear, pineapple, and strawberry. Similarly, we considered normal and augmented images of rotten fruits including beans (two categories), strawberries (seven categories), and tomatoes (three categories). For consistency, we normalized the images and designed an auto-labeling mechanism based on the existing image clusters to label inconsistent data to appropriate classes. Finally, we verified the auto-labeled data with a complete inspection to correctly assign it to the relevant classes. The proposed ensemble classifier outperforms all other classification methods, achieving 100% and 99% accuracy for fruit classification and disease identification. Further, we performed the analysis of variance (ANOVA) test to validate the statistical significance of the classifiers’ outcomes at α = 0.05. We achieved F-values of 32.41 and 11.42 against F-critical values of 2.62 and 2.86, resulting in p-values of 0.00 (<0.05) for fruit classification and disease identification.

1. Introduction

Food is one of the most fundamental requirements for human life and survival. Agriculture plays a vital role in fulfilling food requirements globally. The rate of growth in agricultural production has declined recently, according to the Food and Agricultural Organization (FAO) [1]. Global food security is seriously threatened by this trend and other risks like climate change, population growth, rural-to-urban migration, and biofuel demand [2]. As a result, to increase plant yield for food production, a higher yielding and more sustainable environment is required [3,4].
Given that there will hardly be an increase in the size of land available for agricultural use in the future, the best way to deal with the problem of food insecurity is by increasing the productivity of existing farmland. For this purpose, innovative methods to increase agricultural yield will be continuously required. Furthermore, since 30% of the global agricultural workforce is predicted to move to work in other sectors of the economy between 2017 and 2030 [5], the adoption of agricultural technology becomes even more critical. Therefore, technology has been adopted, modified, and optimized increasingly over the past century to increase agricultural yields [6]. Precision Agriculture (PA) is gaining popularity in this setting of sustainably increasing yields. PA is a broad term that encompasses all approaches used to improve the accuracy and control of farming management [7]. These techniques include GPS navigation of tractors, robotics, remote sensing, data analytics, and unmanned aerial and terrestrial vehicles [8]. PA relies heavily on early and precise plant disease detection.
Fruit production and cultivation comprise a sizeable portion of all agricultural activity and are essential to the human diet. On the other hand, fruit diseases (e.g., gray mold, apple scab, canker, black knot, brown rot) are among the key factors that can negatively influence the quality of fruits and reduce their production. Such diseases essentially affect the quality and quantity of fruits by interfering with several processes, such as plant growth, the development of flowers and fruits, and absorbent capacity [9,10]. As a result, early detection and precise classification of fruit diseases is an intriguing problem related to increasing the potential economic value of fruits [11]. The manual inspection of plants and fruits to detect any signs of disease involves various issues. Firstly, it demands ongoing monitoring, disease identification expertise, and other related efforts and time [12]. Furthermore, this method might not detect all infections.
As far as automated techniques for the detection and classification of fruit diseases are concerned, we can categorize them as laboratory methods (e.g., serological methods and molecular techniques [13]) and image processing methods. As image processing methods are inexpensive and non-destructive while yielding superior performance, they have been widely used [14,15]. Fruit plants exhibit various symptoms from their early to late stages, which are, in most cases, distinctly observable to the naked eye on infected fruits and plants [16]. Thus, image processing algorithms can be created to diagnose these conditions quickly, accurately, and affordably using typical digital photographs [17,18].
The use of image processing techniques and machine learning algorithms in disease detection and recognition is a growing research field with great potential to address the challenges involved in the early detection of diseases [13,18,19]. These methods have been developed to achieve accuracies exceeding human-level perception.
As far as the workflow of disease identification and classification using image processing is concerned, it comprises almost identical phases in all approaches. Specifically, the captured image passes through four steps: image pre-processing, image segmentation, feature extraction, and classification. To this end, early works employed classical image processing procedures and ‘hand-crafted’ feature extraction from leaf and fruit images. The extracted features were then used to train shallow classifier algorithms such as support vector machines, principal component analysis, maximum likelihood classification, k-nearest neighbors, naïve Bayes, decision trees, random forest, and artificial neural networks (e.g., [20,21,22,23]). Another recent line of work has mainly focused on deep learning architectures comprising convolutional neural networks to automate feature extraction and image classification (e.g., [24,25,26,27]).
Despite the exciting results reported by the aforementioned methods, the reliance on a single classifier is deemed to affect the performance of disease detection and classification. In this way, the classification methods based on a single method underperform when applied to large heaps of fruit because of the diversity in texture, shape, color, size, disease, and maturity levels. Here, we propose a hybrid ensemble method for fruit disease identification and fruit classification that employs a multilevel fusion method, intensive fruit image pre-processing, customized image kernels for feature extraction using state-of-the-art deep methods, and Gini-index-based controlled feature selection. The major contributions of this study are as follows:
  • The hybrid ensemble classification method is inspired by multilevel fusion methodology that includes intensive fruit image pre-processing, adoption of customized image kernels for feature extraction with state-of-the-art (SOTA) deep methods, and Gini-index-based controlled feature selection.
  • Considering different use cases of convolutional neural networks (CNNs) simulated on multiple imagery data, we evaluated the performance of Squeeze Net with defined/customized image kernels.
  • Although the performance of state-of-the-art (SOTA) methods is architecture-dependent, the performance of Squeeze Net was noted to be better with relatively faster training time. It ultimately helped to improve classification accuracy by empowering the weaker classifiers.
  • The ensemble model outperforms individual classifiers with the help of the best feature extractor, Squeeze Net, among other convolution networks, including Inception-V3, VGG-16, VGG-19, and DeepLoc, simulated on larger sets of imagery data.
  • The proposed ensemble method lays the foundation for efficient solutions to similar problems on embedded devices by adopting Squeeze Net (in general) and MobileNetV2 (in particular) for remote imagery data.
The rest of the manuscript is organized as follows. A brief description and the limitations of related work are presented in Section 2. Section 3 presents the proposed method with a description of the essential components of the architecture. Section 4 discusses the significant results, and Section 5 concludes the research work.

2. Related Work

There are two approaches for fruit disease classification. The first approach, applicable at an early stage of plant and fruit development, attempts to identify diseases in plants and leaves [24,28,29]. The second approach involves analyzing the skin and appearance of the fruit for disease identification and classification. The scope of this study is limited to the latter approach. In the following, we provide a brief overview of methods and techniques proposed by researchers for disease identification and classification in several fruits, such as citrus, apple, mango, papaya, and strawberry.
Citrus is one of the most popular fruit in the world. Citrus fruit may suffer from various diseases, such as canker, greening, anthracnose, scab, black spot, and melanose. Cubero et al. provide an excellent review of the application of machine learning techniques for various issues related to citrus fruit, including disease detection [30]. Sharif et al. used the support vector machine (SVM) algorithm to identify and classify these diseases in citrus fruit [31]. The proposed method identifies the lesion spots on the fruit’s skin using a segmentation method followed by the classification of diseases. The proposed method is tested on three datasets and achieved 97%, 90.4%, and 89% accuracy for these datasets. Abdulridha et al. used unmanned aerial vehicles to capture hyperspectral images of a citrus orchard for real-time identification of the citrus canker disease [32]. The radial basis function achieved a classification accuracy of 94%, 96%, and 100% for asymptomatic, early, and late symptoms of citrus canker, respectively. The corresponding results for the k-nearest neighbors method were 94%, 95%, and 96%. Rauf et al. prepared a dataset comprising 759 images of healthy and unhealthy citrus leaves and fruit to help other researchers develop methods and techniques for citrus disease detection [33].
Apple is another fruit consumed throughout the world for the whole year. The most common diseases in apple fruit include black rot canker, apple scab, powdery mildew, core rot, white/brown/root rot, collar rot, and seedling blight [34]. Turkoglu et al. proposed an ensemble classification method for apple disease detection [35]. The study applied deep feature extraction on pre-trained models and used the SVM and long short-term memory (LSTM)-based convolutional neural network for disease classification. The authors used a real-time dataset comprising leaves and fruit images. The proposed method achieved 96.4% accuracy with SVM and 99.2% using LSTM. Tian et al. proposed a deep learning model for detecting anthracnose lesions in apple fruit [36]. The authors highlight the difficulty in collecting sufficient data for fruit diseases and use Cycle-Consistent Adversarial Network to augment the available data. The authors used the YOLO-V3 model for disease identification. Fan et al. also used a deep learning model for segregating defective apples from healthy fruit [37]. The study also employed conventional machine learning algorithms, but reported better results using deep learning models. Al-Shawwa and Abu-Naser developed a traditional expert system for detecting all apple fruit diseases mentioned above [34]. A limitation of the study is that the detailed design is not discussed. Moreover, the results are not reported in the study to compare with other solutions.
Strawberries provide great economic value to farmers. However, diseases and pests pose a significant threat to the strawberry fruit. Dong et al. used a deep learning model to identify nine diseases and pests in strawberry fruit, including root rot, slug, thrips, powdery mildew, sunburn, botrytis cinerea, vegetable green insect, virus disease, and aphid [38]. The authors used an enhanced AlexNet by transfer learning that reduced the learning time for the proposed model. The classification accuracy was improved using the inner product and vector max norm. Further details of the proposed model are provided by [39]. Siedliska et al. propose using hyperspectral imaging to detect fungal infections in strawberries [40]. The authors used visible, near-infrared, and shortwave-infrared (VNIR/SWIR) spectroscopy to capture images of two groups of strawberries; a control group without any treatment and a group inoculated with fungi. The study used four classification methods: backpropagation neural network, random forest, naïve Bayes, and support vector machine. The backpropagation neural network achieved the best results.
Habib et al. developed a near real-time computer vision model to detect and identify five diseases in papaya fruit: brown spot, black spot, phytophthora blight, anthracnose, and powdery mildew [41]. A user-captured image from the field is pre-processed by: (a) resizing the image to a fixed size, (b) enhancing contrast through histogram equalization, and (c) converting the image from color to L*a*b* color space. The k-means clustering algorithm is used for image segmentation into healthy and unhealthy fruit regions. The unhealthy regions are then subjected to a feature extraction phase comprising gray level co-occurrence matrix and statistical features. Finally, a support vector machine classifier is used to identify the disease. Andrushia and Patricia use ant colony optimization for feature identification and support vector machine to identify anthracnose and stem end rot diseases of the mango fruit [42].
Several review articles have also addressed the problem of automatic fruit disease identification and classification. Naranjo-Torres et al. have reviewed the application of the convolutional neural network in the problem domain [43]. Bhargava and Bansal have addressed the fruit disease identification problem while reviewing fruit and vegetable grading and quality inspection [15]. Wani et al. provide an excellent review of conventional and deep learning techniques for plant and fruit disease detection [44].

3. Methodology

We propose a multilevel fusion methodology for fruit disease identification and fruit classification that includes intensive fruit image pre-processing, customized image kernels for feature extraction with state-of-the-art (SOTA) deep methods, Gini-index-based controlled feature selection, and hybrid ensemble method for identification and classification.
Figure 1 presents the salient components of the proposed methodology in two segments. The first segment (represented with pink objects) shows a systemic flow of multi-fruit classification. The multi-fruit imagery data contains regular and augmented images of various fruits, including apple, apricot, avocado, banana, cherry, fig, grape, guava, kiwi, mango, orange, peach, pear, pineapple, and strawberry. We identified images of fruits with different shapes, sizes, quality, saturation, and labels that required intensive pre-processing to normalize the fruit for an enhanced classification. In addition, we designed an auto-labeling mechanism based on the existing image clusters to label inconsistent data to belong to appropriate classes. Lastly, we verified the auto-labeled data with a complete inspection to correctly assign it to the relevant classes. Our proposed methodology can be generalized to any larger set of diverse fruit types to meet the agricultural industry requirements for bulk fruit classification.
The state-of-the-art (SOTA) convolutional neural networks generate enormous feature sets of image data that are computationally expensive for conventional classification methods. We employed the custom image kernels to extract the relevant and most significant features. Further, we pruned the extracted features using the Gini index, information gain, and entropy of feature values. Eventually, the pruned features were used as inputs for the ensemble classifier. We used decision trees, AdaBoost, support vector machine, random forest, and gradient boosting methods. We employed individual classifiers on diverse imagery data, although classification outcomes fluctuated. The classification accuracy of an individual classifier on certain imagery data was very different on another data set and vice versa. The variations in the outcomes of single classification methods require devising a new ensemble classification method based on the majority voting policy of several distinct classifiers.
The second segment (represented by green objects) depicts a systemic flow of multi-fruit disease classification. The existing multi-fruit disease imagery data also contains regular and augmented images of different fruits including beans (two categories), strawberries (seven categories), and tomatoes (three categories). Here we also noticed that fruit disease data comes in different shapes, sizes, quality, saturation, and labels. Therefore, as with the fruit imagery data, we performed intensive pre-processing to bring the disease imagery data to a consistent format acceptable for an enhanced classification. In the second segment, we devised an auto-labeling method for correctly labeling unlabeled augmented data (ensuring complete inspection of annotations). At this stage, we also encountered challenges of achieving stable and consistent classification outcomes from single classification methods for various fruit diseases. Here, the proposed ensemble method equally bestowed the significant classification outcomes, tested on various disease data. We illustrate the process in the subsequent section.
Let us consider a set IINPUT = {I1, I2, I3, I4…IK, ∀ K ∈ N} of input fruit images. Let IOUT = {O1, O2, O3, O4…OM, ∀ M ∈ N} be a scaled set of IINPUT images. Let IK, ∀ K ∈ N carry a scale w × h, where w, h ∈ [N, N]. If W, H are new dimensions of IK, W × H is a new scale of OM, ∀ M ∈ N, mapped with OM (x, y) = OM (W, H) = IK (w/W, h/H), where x ∈ [0, N] and y ∈ [0, N]. Further, we consider any fruit image I (a, b) ∈ IINPUT, 0 ≤ a, b ≤ N − 1 of size N × N. Let I (a’, b’) ∈ IOUT, where 0 ≤ a’, b’ ≤ M − 1 of size M × M ∀ M ∈ N. We construct a mapping function F: IINPUTIOUT defines a zoomed image Z (I (a’, b’)) = I (ka, kb), where k ∈ N.
Now, we consider a fruit image J (p, q) ∈ IINPUT, 0 ≤ p, q ≤ N − 1 of size N × N. Let J (p’, q’) ∈ IOUT, where 0 ≤ p’, q’ ≤ M − 1 of size M × M ∀ M ∈ N. Here, we consider a new mapping function S: IINPUTIOUT defines a sheared image Z in the x-direction by J (p’, q’)) = J (p, kq), where k ∈ R. Similarly, we can also construct another mapping function S’: IINPUTIOUT that defines a sheared image Z’ in the y-direction by J (p’, q’)) = J (kp, q), where k ∈ R. Let us consider a fruit image F (r, s) ∈ IINPUT, 0 ≤ r, s ≤ N − 1 of size N × N. Let F (r’, s’) ∈ IOUT, where 0 ≤ r’, s’ ≤ M − 1 of size M × M ∀ M ∈ N. Now, we construct a flip function T: IINPUTIOUT that defines a flip image in the x-direction by F (r’, s’)) = F (r, s-k), where k ∈ R. Further, let function T’: IINPUTIOUT for a flipped image in the y-direction by F (r’, s’) = F (r-k, s), where k ∈ R. Algorithms 1–4 describe the pre-processing steps.
Algorithm 1: Preprocessing (Scaling) of Fruit Images
Inputs: IINPUT = {I1, I2, I3, I4…IK,∀ K∈ N} of input fruit images.
Outputs: IOUT = {O1, O2, O3, O4…OM,∀ M∈ N}
  Start PROC_SCALE
  Let w, h = IKIINPUT, KN, where w, h are width and height of current image
  Let w’, h’ = 100, 100IKIINPUT, KN, where w’ and h’ are new dimensions of image
  Let s_x = w’/w, s_y = h’/h, where s_x and s_y are revised scales of image IK, K ≤ N
    ForA in range (h’):
      ForB in range (w’):
        scaled_IK→populate(x, y), where populate is a procedure that scales IK, K ≤ N
    save_IK→ OM,∀ M∈ N, OMIOUT
      End For
    End For
  End PROC_SCALE
Algorithm 2: Preprocessing (Zooming) of Fruit Images
Inputs: IINPUT = {I1, I2, I3, I4…IK,∀ K∈ N} of input fruit images.
Outputs: IOUT = {O1, O2, O3, O4…OM,∀ M∈ N}
  Start PROC_ZOOM
  Let a, b = IKIINPUT, KN, where a, b are old dimensions of current image
  Let a’, b’ = IMIOUT, MN, where a’ and b’ are new zoomed dimensions of image
    ForX in range (a’):
      ForY in range (b’):
        Zoomed_IK→generate(a, b), where generate will map new dimensions to old dimensionsIK, K ≤ N
      save_IK→ OM,∀ M∈ N, OMIOUT
      End For
    End For
  End PROC_ZOOM
Algorithm 3: Preprocessing (Shearing) of Fruit Images
Inputs: IINPUT = {I1, I2, I3, I4…IK,∀ K∈ N} of input fruit images.
Outputs: IOUT = {O1, O2, O3, O4…OM,∀ M∈ N}
  Start PROC_SHEAR
  Let p, q = J (p, q)IINPUT, p, q ≤ N, where p, q are old coordinates of any image
  Let p’, q’ = J’(p’, q’)IOUT, p’, q’ ≤ N, where p’ and q’ are new sheared dimensions
    ForS in range (p’):
      ForT in range (q’):
        Define S: IINPUTIOUT for sheared image Z in x-direction by J (p’, q’)) = J (p, kq), where kR
        Define S’: IINPUTIOUT for sheared image Z’ in y-direction by J (p’, q’)) = J (kp, q), where kR
      Save Z, Z’→ OM,∀ M∈ N, OMIOUT
      End For
    End For
  End PROC_SHEAR
Algorithm 4: Preprocessing (Flipping) of Fruit Images
Inputs: IINPUT = {I1, I2, I3, I4…IK,∀ K∈ N} of input fruit images.
Outputs: IOUT = {O1, O2, O3, O4…OM,∀ M∈ N}
  Start PROC_FLIP
  Let r, s = F (r, s)IINPUT, r, s ≤ N, where r, s are old coordinates of any image
  Let r’, s’ = F’(r’, s’)IOUT, r’, s’ ≤ N, where r’ and s’ are new flipped dimensions
    ForA in range (r’):
      ForB in range (s’):
        Define T: IINPUTIOUT for Flipped image F in x-direction by F (r’, s’)) = F (r, s-k), where kR
        Define T’: IINPUTIOUT for Flipped image F in y-direction by F (r’, s’)) = F (r-k, s), where kR
      Save F, F’→ OM,∀ M∈ N, OM∈ IOUT
      End For
    End For
  End PROC_FLIP
Let us now consider IOUT = {O1, O2, O3, O4…OM, ∀ M ∈ N} and I’OUT = {O1, O2, O3, O4…OM, ∀ M ∈ N} as pre-processed images of normal and rotten fruit, respectively. We extract feature vectors F, and F’ of said datasets, and analyze with classification methods described in Algorithms 5 and 6 below.
Algorithm 5: Ensemble Classification of pre-processed Images
Start: PROC_Fruits-Classification
Inputs: Feature Vector F of IOUT = {O1, O2, O3, O4…OM,∀ M∈ N} images of different fruits
Outputs: Classification Outcomes vector of different fruits C achieved for P instances
  Let us consider a collection IOUT = {O1, O2, O3, O4…OM,∀ M∈ N} of processed images
    Let us extract the features by applying image embedding to extract F = {F1, F2, F3, Fi} features of images∀ i ≤ N
      Analyze Pi instances with features F using AdaBoost classifier where each Pi in P
      Analyze Pi instances with features F using the Decision tree classifier, each Pi in P
      Analyze Pi instances having features F using Support Vector Machine (SVM) classifier, each Pi in P
      Analyze Pi instances having features F using Random Forest classifier, each Pi in P
      Analyze Pi instances with features F using GradientBoost where each Pi in P
      Analyze the individual performance of all classifiers on Pi attributes of P for i ≤ N
      Output the classification as a result C (Y ≤ 5) of Y classifiers
    End
End: PROC_Fruits-Classification
Algorithm 6: Ensemble Fruit Disease Classification of pre-processed Images
Start: PROC_Fruits-Disease_Classification
  Inputs: Feature Vector F’ of I’OUT = {O1, O2, O3, O4…OM,∀ M∈ N} images of rotten fruits
  Outputs: Classification Outcomes of fruit disease vector C’ achieved for P instances
  Let us consider a collection I’OUT = {O1, O2, O3, O4…OM,∀ M∈ N} of processed fruit disease images
    Let us extract the features by applying image embedding to extract F = {F’1, F’2, F’3, F’i} features of fruit diseases images∀ i ≤ N
      Analyze Pi instances with features F’ using AdaBoost classifier where each Pi in P
      Analyze Pi instances with features F using the Decision tree classifier, each Pi in P
      Analyze Pi instances having features F using k-Nearest Neighbor (kNN) classifier, each Pi in P
      Analyze Pi instances having features F using Random Forest classifier, each Pi in P
      Analyze the individual performance of all classifiers on Pi attributes of P for i ≤ N
      Output the classification as a result C’ (Y ≤ 4) of Y classifiers
    End
  End: PROC_Fruits-Disease_Classification
Further, we implemented the bagging strategy to seek a classification model with the lowest variance compared to single classification methods. The proposed ensemble model achieved the best approximation of bootstrap samples. Let us consider M bootstrap samples of size L as described,
S 1 1 ,   S 2 1 ,     S 3 1 ,   S L 1 ,   S 1 2 ,   S 2 2 ,     S 3 2 ,   S L 2 ,   S 1 3 ,   S 2 3 ,     S 3 3 ,   S L 3 , , S 1 M ,   S 2 M ,     S 3 M ,   S L M
In the above expression given by (*), S L M represents L-th training examples of M-th bootstrap samples. Based on M samples, we can fit about M weak learners Learner1 (.), Learner2 (.), Learner3 (.), …….…, LearnerM(.) to seek lower variance by averaging their outcomes. The ensemble model thus becomes:
S M . = arg T m a x   c a r d m | L e a r n e r m . = T
Using equation 1, we apply soft voting policy, aggregate, and average the probabilities of models presented in Figure 1, and retain the class with the highest mean probability. In addition, the chosen models for voting policy adopted parallelism and provided the flexibility to include as many models as required.

4. Results and Discussion

In the following Section 4.1 we present the numerical results as the performance analysis of different classification methods, while Section 4.2 further illustrates the findings.

4.1. Results

We reviewed several publicly available fruit imagery datasets based on the reported literature on fruit classification and disease identification. We noticed certain limitations in adopting a single data source regarding limited data sizes, variability in fruit types, variability in quality, and variability in disease type. Eventually, we extensively aggregated and pre-processed multi-fruit data to simulate our proposed ensemble model on comprehensive datasets to cover fruit classification and disease identification. The multi-fruit imagery data contained regular and augmented images of fruit including apple, apricot, avocado, banana, cherry, fig, grape, guava, kiwi, mango, orange, peach, pear, pineapple, and strawberry. Similarly, the multi-fruit disease imagery data also contains regular and augmented images of different fruits including beans (two categories), strawberries (seven categories), and tomatoes (three categories). We also identified images of fruits with different shapes, sizes, quality, saturation, and labels that required intensive pre-processing to normalize the fruit for an enhanced classification. In addition, we designed an auto-labeling mechanism based on existing image clusters to label inconsistent data to the appropriate classes. Lastly, we verified the auto-labeled data with a complete inspection to assign the relevant classes correctly.
We employed the following evaluation measures to investigate the performance of the classification methods:
True Positive (TP): The outcome of the model when the model correctly predicts the positive class
False Positive (FP): The outcome of the model when the model incorrectly predicts the positive class
True Negative (TN): The outcome of the model when the model correctly predicts the negative class
False Negative (FN): The outcome of the model when the model incorrectly predicts the positive class
(a) Accuracy: The proportion of true results to the total number of cases examined,
Classification   Accuracy   CA = TP + TN TP + FP + FN + TN   ×   100 %
(b) Precision: The proportion of predicted positives to be truly positive,
Precision = TP TP + FP   ×   100 %
(c) Recall: The proportion of actual positives correctly classified,
Recall = TP TP + FN   ×   100 %
(d) AUC: How well the probabilities from the positive classes are separated from the negative classes.
(e) F1-Score: The overall performance of the model is measured
F 1 Score = 2 TP 2 TP + FP + FN   ×   100 %
Figure 2 shows the performance of different fruit classification methods. The decision tree classifiers performed the worst, with an average of 0.85 for recall, F1 score, and classification accuracy. The SVM and random forest classifiers depict similar outcomes on almost all evaluation metrics, with evaluation measures of 0.93 to 0.95. The proposed ensemble classifier outperforms all other classification methods, achieving 100% for AUC, CA, F1 score, precision, and recall.
Table 1 presents the misclassification (partial confusion matrix) of different classification methods (only cells colored red refer to misclassifications). The decision tree classifier depicts a more significant misclassification for some fruits, i.e., fig, guava, kiwi, mango, orange, and pear. Gradient boosting and random forest misclassified guava, mango, and pear. The ensemble classification had no misclassification except for guava (5%) and outperformed all single classification methods. Almost all classification methods carried a more considerable misclassification for the guava fruit. Similarly, mango was also misclassified to some extent by decision trees, AdaBoost, SVM, gradient boost, and random forest algorithms.
Figure 3 presents the ROC analysis of the classification methods. The AdaBoost and decision tree classifier methods performed the worst. Gradient boosting and random forest classification methods show similar true positive/false positive rates. In contrast, the ensemble classification method stands out with a significant curve leaning towards the upper left corner of the ROC graph.
Further, we calculated the performance metrics of different classification methods for fruit disease identification. We recorded the following observations.
Figure 4 demonstrates the performance analysis of different classification methods for fruit disease identification. Here, we can notice that the decision tree classifier underperformed with an average of 0.67 for recall, precision, F1 score, and CA, while it achieved 0.86 for the area under the curve. Random forest showed an intermediate performance compared to other classification methods, with an average score of 0.78 for recall, precision, F1, and CA, while it scored 0.96 for the AUC metric. Similarly, the kNN performed well for all evaluation metrics. We can notice the outperformance of the ensemble classification method that scored 0.99 for the AUC and 0.93 for other evaluation metrics.
Table 2 shows the misclassification of classification methods for fruit disease identification (only cells colored red refer to misclassifications). We can see that the decision tree classifier had a more significant misclassification of 51%, 26%, 9%, 3%, and 6%, for Beans-1, Beans-2, Strawberry-1, Strawberry-2, and Strawberry-4, respectively. Similarly, AdaBoost missed fruits Beans-1, Beans-2, and Strawberry-5 with a more significant 77%, 19%, and 82%, respectively. The kNN and random forest misclassified Beans-1 and Beans-2 by 82% and 76%, respectively. The ensemble classifier stood out with misclassifications of 3% and 7% for Beans-1 and Beans-2, respectively. Almost all classification methods misclassified Beans-1 and Beans-2 to varying extents.
Figure 5 shows the receiver operating characteristic (ROC) curve analysis of the classification methods. The AdaBoost and Decision Tree classifiers performed the worst, and kNN and Random Forest classification methods show similar true positive/false positive rates. In contrast, the ensemble classification method stands out with a significant curve leaning towards the upper left corner of the ROC graph.

4.2. Discussion

To analyze the performance of our proposed ensemble classification method, we compiled a larger dataset of over 20,000 images. A careful review of the literature reported the employment of various datasets. The reported datasets varied to a larger extent regarding the size, shape, quality, and variety of fruits for classification and disease identification.
We segregated our datasets into four major segments. We applied bootstrapping on each segment to identify an optimal method for feature extraction. The study considered Inception V3, Squeeze Net, VGG-16, VGG-19, and DeepLoc as the best candidate networks for feature extraction.
Table 3 highlights the comparative analysis of different feature extraction methods. We employed the convolution layers of these models. Based on the images’ embedding, we noticed the following cumulative statistics:
Figure 6 describes the performance metrics of different feature extraction models. For our particular domain of fruit classification and disease identification, we recorded that Squeeze Net outperformed the other models with an extent of 0.99 for AUC, 0.95 for CA, F1 score, and recall. In addition, it showed a precision of 0.97 compared to the other models.
Figure 7 shows the selection of features based on information gain, gain ratio, and Gini index. The principal components are represented with letters "PC", for instance "PC3" represents the third principal component. The Gini index is considered an excellent split criterion for most machine learning algorithms. This is the reason that we restricted our findings to the Gini index. In addition, the other measures, i.e., information gain and gain ratio, depicted similar characteristics. We can see the features sorted in descending order with respect to the Gini index referring to the most robust features (candidates) for performance analysis of ensemble classification models and other classifiers.
Furthermore, we validated the feature selection using central data tendency measures as shown in Figure 8. Figure 8 is a high-resolution figure prepared at 600 pixels per inch. It may be zoomed in to view the contents.
Figure 8 presents an exponential number format which can be read as follows, the subsequent signed number (either plus or minus) succeeding “e” is raised to the power of 10, for instance number 8 × 10−3 will be read as 8 × 10−3. The central data tendency measure gives the data distribution with respect to mean, mode, and median. It also glimpses the data dispersion, minimum, maximum, and missing values (if any) in certain features. We can validate the chosen strongest candidates concerning variance in the data and their distance from the mean. This gives us a second opinion regarding the selection of better candidate features.
In addition, we also measured the relevance of features with each other as follows:
Figure 9 shows the impact between features. For instance, we can identify different fruit clusters among principal components (1) and (2). Apple and cherry fruits are well segregated, while mango, peach, pear, and strawberry overlap at certain data points of two features. This is another dimension that gives us the co-relevance among principal components to understand the dispersion of data into clusters shared by two or more principal components.
Figure 10 presents the classification accuracy and error of different classification methods. It should be noted that hyphen sign with vertical/horizontal scale represent minus sign, for instance -8 can be read as (−, 8) for clarity. The actual accuracy is represented with a green bar and the residual error with red bars. We can notice that the ensemble method has negligible (or no) errors in the correct classification of fruits. Further, the random forest and SVM classifiers show a similar residual error of 0.45. The decision tree classifier carries the highest residual error in the correct classification of fruits.
Figure 11 shows the classification accuracy and error of methods for fruit disease identification. The decision tree classifier carries the most significant residual error, while kNN and AdaBoost contain an error in an intermediate range of 0.10. The random forest algorithm shows a 0.21 error. The ensemble classifier produced the slightest error in fruit disease identification.
Further, we performed the analysis of variance (ANOVA) test on our simulated results (including both fruit classification and disease identification). ANOVA employs the F-test to investigate the variability between group means within and between the groups. We achieved the following outcomes:
In Table 4, chosen on a 95% confidence interval (α = 0.05), the F value is 32.41, which is significantly larger than the F-critical value of 2.62, achieving a p-value of 0.00, which is significantly smaller than 0.05.
Similarly, we performed the ANOVA test on the outcomes of SOTA methods used for feature extraction.
In Table 5, chosen on a 95% confidence interval (α = 0.05), the F value is 11.42, which is significantly larger than the F-critical value of 2.866, achieving a p-value of 0.00, significantly smaller than 0.05.

5. Conclusions

We proposed a hybrid ensemble classification method for fruit classification and disease identification, inspired by multilevel fusion methodology that includes intensive fruit image pre-processing, adoption of customized image kernels for feature extraction with state-of-the-art (SOTA) deep learning methods, and Gini-index-based controlled feature selection. Realizing the limitations of publicly available fruit imagery datasets regarding data sizes, variability in fruit types, quality, and diseases, we extensively aggregated and pre-processed multi-fruit data to achieve augmented images of various fruits, including apple, apricot, avocado, banana, cherry, fig, grape, guava, kiwi, mango, orange, peach, pear, pineapple, beans (two categories), strawberries (seven categories), and tomatoes (three categories). The ensemble classifier outperformed all classification methods, achieving 100% and 99% classification accuracies for fruit classification and disease identification. We validated the statistical significance of the classifiers’ outcomes at α = 0.05 and achieved F-values of 32. 41 and 11.42 against F-critical values of 2.62 and 2.86, resulting in p-values of 0.00 (<0.05) for fruit classification and disease identification. As a future work direction, we consider implementing the hybrid ensemble algorithm to embedded devices for remote classification of image sensory data. Another limitation of the study regards multiple simultaneous diseases in a fruit. New data sets need to be developed to test the performance of our approach for such instances.

Author Contributions

Conceptualization, A.M., M.A. and Q.M.I.; methodology, Q.M.I., A.M. and M.A.; formal analysis, A.M., M.A. and Q.M.I.; investigation, A.M., M.A. and Q.M.I.; resources, A.M.; data curation, A.M., M.A. and Q.M.I.; writing—original draft preparation, A.M., M.A. and Q.M.I.; project administration, A.M.; funding acquisition, A.M. All authors have read and agreed to the published version of the manuscript.

Funding

The authors extend their appreciation to the Deputyship for Research & Innovation, Ministry of Education in Saudi Arabia for funding this research work through the project number INST030.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Food and Agricultural Organization (FAO). Crop Production and Natural Resource Use. Available online: http://www.fao.org/3/y4252e/y4252e06.htm (accessed on 7 December 2022).
  2. Calicioglu, O.; Flammini, A.; Bracco, S.; Bellù, L.; Sims, R. The Future Challenges of Food and Agriculture: An Integrated Analysis of Trends and Solutions. Sustainability 2019, 11, 222. [Google Scholar] [CrossRef] [Green Version]
  3. Rathore, S.S.; Shekhawat, K.; Singh, R.K.; Babu, S.; Singh, V.K. Diversification for Restoration of Ecosystems and Sustainable Livelihood. Sustain. Agric. Syst. Technol. 2022, 21–36. [Google Scholar] [CrossRef]
  4. Hunter, M.C.; Smith, R.G.; Schipanski, M.E.; Atwood, L.W.; Mortensen, D.A. Agriculture in 2050: Recalibrating Targets for Sustainable Intensification. Bioscience 2017, 67, 386–391. [Google Scholar] [CrossRef] [Green Version]
  5. Research for AGRI Committee - The EU Farming Employment: Current Challenges and Future Prospects. Available online: https://www.europarl.europa.eu/thinktank/en/document/IPOL_STU(2019)629209 (accessed on 7 December 2022).
  6. Horton, P.; Long, S.P.; Smith, P.; Banwart, S.A.; Beerling, D.J. Technologies to deliver food and climate security through agriculture. Nat. Plants 2021, 7, 250–255. [Google Scholar] [CrossRef]
  7. Cisternas, I.; Velásquez, I.; Caro, A.; Rodríguez, A. Systematic literature review of implementations of precision agriculture. Comput. Electron. Agric. 2020, 176, 105626. [Google Scholar] [CrossRef]
  8. Lassoued, R.; Macall, D.M.; Smyth, S.J.; Phillips, P.W.B.; Hesseln, H. Expert Insights on the Impacts of, and Potential for, Agricultural Big Data. Sustainability 2021, 13, 2521. [Google Scholar] [CrossRef]
  9. Ali, M.M.; Yousef, A.F.; Li, B.; Chen, F. Effect of Environmental Factors on Growth and Development of Fruits. Trop. Plant Biol. 2021, 14, 226–238. [Google Scholar] [CrossRef]
  10. Paradiso, R.; Proietti, S. Light-Quality Manipulation to Control Plant Growth and Photomorphogenesis in Greenhouse Horticulture: The State of the Art and the Opportunities of Modern LED Systems. J. Plant Growth Regul. 2022, 41, 742–780. [Google Scholar] [CrossRef]
  11. Wieme, J.; Mollazade, K.; Malounas, I.; Zude-Sasse, M.; Zhao, M.; Gowen, A.; Argyropoulos, D.; Fountas, S.; Van Beek, J. Application of hyperspectral imaging systems and artificial intelligence for quality assessment of fruit, vegetables and mushrooms: A review. Biosyst. Eng. 2022, 222, 156–176. [Google Scholar] [CrossRef]
  12. Khan, M.A.; Akram, T.; Sharif, M.; Alhaisoni, M.; Saba, T.; Nawaz, N. A probabilistic segmentation and entropy-rank correlation-based feature selection approach for the recognition of fruit diseases. Eurasip J. Image Video Process. 2021, 2021, 1–28. [Google Scholar] [CrossRef]
  13. Manavalan, R. Automatic identification of diseases in grains crops through computational approaches: A review. Comput. Electron. Agric. 2020, 178, 105802. [Google Scholar] [CrossRef]
  14. Ouhami, M.; Hafiane, A.; Es-Saady, Y.; El Hajji, M.; Canals, R. Computer Vision, IoT and Data Fusion for Crop Disease Detection Using Machine Learning: A Survey and Ongoing Research. Remote Sens. 2021, 13, 2486. [Google Scholar] [CrossRef]
  15. Bhargava, A.; Bansal, A. Fruits and vegetables quality evaluation using computer vision: A review. J. King Saud Univ.-Comput. Inf. Sci. 2021, 33, 243–257. [Google Scholar] [CrossRef]
  16. Zhang, X.; Qiao, Y.; Meng, F.; Fan, C.; Zhang, M. Identification of Maize Leaf Diseases Using Improved Deep Convolutional Neural Networks. IEEE Access 2018, 6, 30370–30377. [Google Scholar] [CrossRef]
  17. Thakur, P.S.; Khanna, P.; Sheorey, T.; Ojha, A. Trends in vision-based machine learning techniques for plant disease identification: A systematic review. Expert Syst. Appl. 2022, 208, 118117. [Google Scholar] [CrossRef]
  18. Ngugi, L.C.; Abelwahab, M.; Abo-Zahhad, M. Recent advances in image processing techniques for automated leaf pest and disease recognition–A review. Inf. Process. Agric. 2020, 8, 27–51. [Google Scholar] [CrossRef]
  19. Shashi; Singh, J. Automated Disease Detection and Classification of Plants Using Image Processing Approaches: A Review. Lect. Notes Netw. Syst. 2021, 203, 641–651. [Google Scholar] [CrossRef]
  20. Johannes, A.; Picon, A.; Alvarez-Gila, A.; Echazarra, J.; Rodriguez-Vaamonde, S.; Navajas, A.D.; Ortiz-Barredo, A. Automatic plant disease diagnosis using mobile capture devices, applied on a wheat use case. Comput. Electron. Agric. 2017, 138, 200–209. [Google Scholar] [CrossRef]
  21. Krithika, N.; Selvarani, A.G. An individual grape leaf disease identification using leaf skeletons and KNN classification. In Proceedings of the 2017 International Conference on Innovations in Information, Embedded and Communication Systems (ICIIECS), Coimbatore, India, 17–18 March 2017; Volume 138, pp. 1–5. [Google Scholar] [CrossRef]
  22. Islam, M.; Anh, D.; Wahid, K.; Bhowmik, P. Detection of potato diseases using image segmentation and multiclass support vector machine. In Proceedings of the 2017 IEEE 30th Canadian Conference on Electrical and Computer Engineering (CCECE), Windsor, ON, Canada, 30 April–3 May 2017; pp. 1–4. [Google Scholar] [CrossRef] [Green Version]
  23. Qin, F.; Liu, D.; Sun, B.; Ruan, L.; Ma, Z.; Wang, H. Identification of Alfalfa Leaf Diseases Using Image Recognition Technology. PLoS ONE 2016, 11, e0168274. [Google Scholar] [CrossRef] [Green Version]
  24. Li, L.; Zhang, S.; Wang, B. Plant Disease Detection and Classification by Deep Learning—A Review. IEEE Access 2021, 9, 56683–56698. [Google Scholar] [CrossRef]
  25. Bierman, A.; LaPlumm, T.; Cadle-Davidson, L.; Gadoury, D.; Martinez, D.; Sapkota, S.; Rea, M. A High-Throughput Phenotyping System Using Machine Vision to Quantify Severity of Grapevine Powdery Mildew. Plant Phenomics 2019, 2019, 1–13. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. JBarbedo, J.G. Factors influencing the use of deep learning for plant disease recognition. Biosyst. Eng. 2018, 172, 84–91. [Google Scholar] [CrossRef]
  27. Fuentes, A.; Yoon, S.; Kim, C.S.; Park, S.D. A Robust Deep-Learning-Based Detector For Real-Time Tomato Plant Diseases and Pests Recognition. Precis. Agric. 2017, 17, 2022. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Vishnoi, V.K.; Kumar, K.; Kumar, B. Plant disease detection using computational intelligence and image processing. J. Plant Dis. Prot. 2020, 128, 19–53. [Google Scholar] [CrossRef]
  29. Buja, I.; Sabella, E.; Monteduro, A.; Chiriacò, M.; De Bellis, L.; Luvisi, A.; Maruccio, G. Advances in Plant Disease Detection and Monitoring: From Traditional Assays to In-Field Diagnostics. Sensors 2021, 21, 2129. [Google Scholar] [CrossRef]
  30. Cubero, S.; Lee, W.S.; Aleixos, N.; Albert, F.; Blasco, J. Automated Systems Based on Machine Vision for Inspecting Citrus Fruits from the Field to Postharvest—A Review. Food Bioprocess Technol. 2016, 9, 1623–1639. [Google Scholar] [CrossRef] [Green Version]
  31. Sharif, M.; Khan, M.A.; Iqbal, Z.; Azam, M.F.; Lali, M.I.U.; Javed, M.Y. Detection and classification of citrus diseases in agriculture based on optimized weighted segmentation and feature selection. Comput. Electron. Agric. 2018, 150, 220–234. [Google Scholar] [CrossRef]
  32. Abdulridha, J.; Batuman, O.; Ampatzidis, Y. UAV-Based Remote Sensing Technique to Detect Citrus Canker Disease Utilizing Hyperspectral Imaging and Machine Learning. Remote Sens. 2019, 11, 1373. [Google Scholar] [CrossRef] [Green Version]
  33. Rauf, H.T.; Saleem, B.A.; Lali, M.I.U.; Khan, M.A.; Sharif, M.; Bukhari, S.A.C. A citrus fruits and leaves dataset for detection and classification of citrus diseases through machine learning. Data Brief 2019, 26, 104340. [Google Scholar] [CrossRef]
  34. Al-Shawwa, M.; Abu-Naser, S.S. Knowledge Based System for Apple Problems Using CLIPS. Int. J. Acad. Eng. Res. 2019, 3, 1–11. Available online: www.ijeais.org/ijaer (accessed on 22 November 2022).
  35. Turkoglu, M.; Hanbay, D.; Sengur, A. Multi-model LSTM-based convolutional neural networks for detection of apple diseases and pests. J. Ambient Intell. Humaniz. Comput. 2019, 13, 3335–3345. [Google Scholar] [CrossRef]
  36. Tian, Y.; Yang, G.; Wang, Z.; Li, E.; Liang, Z. Detection of Apple Lesions in Orchards Based on Deep Learning Methods of CycleGAN and YOLOV3-Dense. J. Sens. 2019, 2019, 1–13. [Google Scholar] [CrossRef]
  37. Fan, S.; Li, J.; Zhang, Y.; Tian, X.; Wang, Q.; He, X.; Zhang, C.; Huang, W. On line detection of defective apples using computer vision system combined with deep learning methods. J. Food Eng. 2020, 286, 110102. [Google Scholar] [CrossRef]
  38. Dong, C.; Zhang, Z.; Yue, J.; Zhou, L. Classification of strawberry diseases and pests by improved AlexNet deep learning networks. In Proceedings of the 2021 13th International Conference on Advanced Computational Intelligence (ICACI), Wanzhou, China, 14–16 May 2021; pp. 359–364. [Google Scholar] [CrossRef]
  39. Dong, C.; Zhang, Z.; Yue, J.; Zhou, L. Automatic recognition of strawberry diseases and pests using convolutional neural network. Smart Agric. Technol. 2021, 1, 100009. [Google Scholar] [CrossRef]
  40. Siedliska, A.; Baranowski, P.; Zubik, M.; Mazurek, W.; Sosnowska, B. Detection of fungal infections in strawberry fruit by VNIR/SWIR hyperspectral imaging. Postharvest Biol. Technol. 2018, 139, 115–126. [Google Scholar] [CrossRef]
  41. Habib, T.; Majumder, A.; Jakaria, A.; Akter, M.; Uddin, M.S.; Ahmed, F. Machine vision based papaya disease recognition. J. King Saud Univ.-Comput. Inf. Sci. 2020, 32, 300–309. [Google Scholar] [CrossRef]
  42. Andrushia, A.D.; Patricia, A.T. Artificial bee colony based feature selection for automatic skin disease identification of mango fruit. In Nature Inspired Optimization Techniques for Image Processing Applications; Springer: Berlin/Heidelberg, Germany, 2019; Volume 150. [Google Scholar] [CrossRef]
  43. Naranjo-Torres, J.; Mora, M.; Hernández-García, R.; Barrientos, R.J.; Fredes, C.; Valenzuela, A. A Review of Convolutional Neural Network Applied to Fruit Image Processing. Appl. Sci. 2020, 10, 3443. [Google Scholar] [CrossRef]
  44. Wani, J.A.; Sharma, S.; Muzamil, M.; Ahmed, S.; Sharma, S.; Singh, S. Machine Learning and Deep Learning Based Computational Techniques in Automatic Agricultural Diseases Detection: Methodologies, Applications, and Challenges. Arch. Comput. Methods Eng. 2021, 29, 641–677. [Google Scholar] [CrossRef]
Figure 1. Block diagram of the proposed methodology.
Figure 1. Block diagram of the proposed methodology.
Agriculture 13 00500 g001
Figure 2. Performance analysis of fruit classification methods.
Figure 2. Performance analysis of fruit classification methods.
Agriculture 13 00500 g002
Figure 3. Receiver operating characteristic curve analysis.
Figure 3. Receiver operating characteristic curve analysis.
Agriculture 13 00500 g003
Figure 4. Performance metrics for fruit disease identification.
Figure 4. Performance metrics for fruit disease identification.
Agriculture 13 00500 g004
Figure 5. ROC analysis of classification methods for fruit disease identification.
Figure 5. ROC analysis of classification methods for fruit disease identification.
Agriculture 13 00500 g005
Figure 6. Cumulative performance analysis of models.
Figure 6. Cumulative performance analysis of models.
Agriculture 13 00500 g006
Figure 7. Feature selection based on characteristics.
Figure 7. Feature selection based on characteristics.
Agriculture 13 00500 g007
Figure 8. Basic consideration of measures for feature selection.
Figure 8. Basic consideration of measures for feature selection.
Agriculture 13 00500 g008
Figure 9. Scatter plot analysis of selected features.
Figure 9. Scatter plot analysis of selected features.
Agriculture 13 00500 g009
Figure 10. Classification accuracy and error in fruit classification.
Figure 10. Classification accuracy and error in fruit classification.
Agriculture 13 00500 g010
Figure 11. Classification accuracy and error in fruit disease identification.
Figure 11. Classification accuracy and error in fruit disease identification.
Agriculture 13 00500 g011
Table 1. Misclassification of classification methods.
Table 1. Misclassification of classification methods.
FruitModel
Decision TreeAdaBoostSVMRandom ForestGradient BoostingEnsemble
Apple100%100%100%100%100%100%
Apricot100%20%100%100%100%100%
Avocado100%100%100%100%100%100%
Banana100%100%100%100%100%100%
Cherry100%100%100%100%100%100%
Fig14%17%100%100%100%100%
Grape100%100%100%100%100%100%
Guava20%50%17%13%33%5%
Kiwi20%25%100%100%100%100%
Mango14%17%17%13%33%100%
Orange50%100%100%100%100%100%
Peach100%100%100%100%100%100%
Pear40%17%100%13%17%100%
Pineapple100%100%100%100%100%100%
Strawberry100%100%100%100%100%100%
Table 2. Misclassification of classification methods for fruit disease identification.
Table 2. Misclassification of classification methods for fruit disease identification.
FruitModel
Decision TreeAdaBoostkNNRandom ForestEnsemble
Beans-151%77%82%76%3%
Beans-226%19%11%24%7%
Strawberry-19%4%7%100%100%
Strawberry-23%100%100%100%100%
Strawberry-3100%100%100%100%100%
Strawberry-46%100%30%50%100%
Strawberry-5100%82%70%100%100%
Strawberry-6100%100%100%100%100%
Strawberry-76%12%100%60%100%
Tomato-1100%100%100%100%100%
Tomato-2100%100%100%100%100%
Tomato-3100%100%100%100%100%
Table 3. Comparative analysis of feature extraction methods.
Table 3. Comparative analysis of feature extraction methods.
Segment 1
ModelAUCCAF1PrecisionRecall
Inception V30.9300.8680.8660.8870.868
Squeeze Net0.9800.9620.9630.9720.962
VGG-160.9500.9060.9070.9280.906
VGG-190.9300.8680.8400.8320.868
DeepLoc0.9710.9430.9450.9600.943
Segment 2
ModelAUCCAF1PrecisionRecall
Inception V30.9210.8490.8560.9160.849
Squeeze Net0.9690.9430.9420.9560.943
VGG-160.9500.9060.9160.9580.906
VGG-190.9500.9060.9080.9310.906
DeepLoc0.9380.8680.8690.9000.868
Segment 3
ModelAUCCAF1PrecisionRecall
Inception V30.9980.9620.9640.9750.962
Squeeze Net0.9980.9620.9620.9700.962
VGG-160.9920.9620.9640.9750.962
VGG-190.9920.9620.9640.9750.962
DeepLoc0.9890.9620.9640.9750.962
Segment 4
ModelAUCCAF1PrecisionRecall
Inception V30.9980.9350.9390.9550.935
Squeeze Net0.9890.9460.9510.9640.946
VGG-160.9870.9350.9400.9570.935
VGG-190.9870.9270.9340.9560.927
DeepLoc0.9840.9040.9080.9270.904
Table 4. ANOVA: Statistical significance of outcomes of fruit classification methods (α = 0.05).
Table 4. ANOVA: Statistical significance of outcomes of fruit classification methods (α = 0.05).
SUMMARY
GroupsCountSumAverageVariance
Decision Tree54.390.8780.0013647
AdaBoost54.410.8830.0007402
SVM54.830.9663.28 × 10−5
Random Forest54.810.9630.0000222
Gradient Boosting54.740.9484.6 × 10−5
Ensemble5510
Source of VariationSSdfMSFp-valueF crit
Between Groups0.0595750.0119132.41030.000002.620
Within Groups0.00882240.00036
Total0.06840129
Table 5. ANOVA: Statistical significance of outcomes of disease identification methods (α = 0.05).
Table 5. ANOVA: Statistical significance of outcomes of disease identification methods (α = 0.05).
SUMMARY
GroupsCountSumAverageVariance
Decision Tree53.590.7180.0058
AdaBoost54.410.8830.0013
kNN54.520.9040.0008
Random Forest54.130.8260.0053
Ensemble54.630.9260.0012
Source of VariationSSdfMSFp-valueF crit
Between Groups0.13812040.034511.4240.0000542.866
Within Groups0.060448200.0030
Total0.1985624
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mehmood, A.; Ahmad, M.; Ilyas, Q.M. On Precision Agriculture: Enhanced Automated Fruit Disease Identification and Classification Using a New Ensemble Classification Method. Agriculture 2023, 13, 500. https://doi.org/10.3390/agriculture13020500

AMA Style

Mehmood A, Ahmad M, Ilyas QM. On Precision Agriculture: Enhanced Automated Fruit Disease Identification and Classification Using a New Ensemble Classification Method. Agriculture. 2023; 13(2):500. https://doi.org/10.3390/agriculture13020500

Chicago/Turabian Style

Mehmood, Abid, Muneer Ahmad, and Qazi Mudassar Ilyas. 2023. "On Precision Agriculture: Enhanced Automated Fruit Disease Identification and Classification Using a New Ensemble Classification Method" Agriculture 13, no. 2: 500. https://doi.org/10.3390/agriculture13020500

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop