Next Article in Journal
Farmer Perception of Links between Grassland Diversity and Animal Health in Relation to Farm Structure
Previous Article in Journal
Study on Synergistic Emission Reduction in Greenhouse Gases and Air Pollutants in Hebei Province
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Sustainable Coffee Leaf Diagnosis: A Deep Knowledgeable Meta-Learning Approach

by
Abdullah Ali Salamai
1,* and
Waleed Tawfiq Al-Nami
2
1
Department of Management, Applied College, Jazan University, Jazan 82822, Saudi Arabia
2
Department of Computer, Applied College, Jazan University, Jazan 82822, Saudi Arabia
*
Author to whom correspondence should be addressed.
Sustainability 2023, 15(24), 16791; https://doi.org/10.3390/su152416791
Submission received: 21 August 2023 / Revised: 28 September 2023 / Accepted: 28 September 2023 / Published: 13 December 2023

Abstract

:
Multi-task visual recognition plays a pivotal role in addressing the composite challenges encountered during the monitoring of crop health, pest infestations, and disease outbreaks in precision agriculture. Machine learning approaches have been revolutionizing the diagnosis of plant disease in recent years; however, they require a large amount of training data and suffer from limited generalizability for unseen data. This work introduces a novel knowledgeable meta-learning framework for the few-shot multi-task diagnosis of biotic stress in coffee leaves. A mixed vision transformer (MVT) learner is presented to generate mixed contextual attention maps from discriminatory latent representations between support and query images to give more emphasis to the biotic stress lesions in coffee leaves. Then, a knowledge distillation strategy is introduced to avoid disastrous forgetting phenomena during inner-loop training. An adaptive meta-training rule is designed to automatically update the parameters of the meta-learner according to the current task. The competitive results from exhaustive experimentations on public datasets demonstrate the superior performance of our approach over the traditional methods. This is not only restricted to enhancing the accuracy and efficiency of coffee leaf disease diagnosis but also contributes to reducing the environmental footprint through optimizing resource utilization and minimizing the need for chemical treatments, hence aligning with broader sustainability goals in agriculture.

1. Introduction

Multi-task visual recognition is a revolutionary computer vision paradigm that allows training single artificial intelligence algorithms to perform many visual recognition tasks instantaneously. Different from single-task learning, in which independent models are trained for each separate task, resulting in redundant computation and enlarged complexity of the recognition algorithm, the multi-task paradigm makes use of shared representations across tasks, enabling the model to learn common designs and representations that are valuable for multiple tasks [1]. This method encourages the model to learn from its experiences with different tasks, thus enhancing its performance across the board. The model’s efficiency and efficacy are both improved by the shared learning across tasks, and more comprehensive comprehension of the visual input is provided, allowing for the modeling of complex interdependencies and linkages. Better generalization is another benefit of multi-task visual recognition since the model is trained to extract high-level features that are robust and meaningful across different recognition tasks [2].
Deep learning (DL) has been revolutionizing the visual recognition tasks (i.e., classification, detection, segmentation, colorization, etc.) in many application domains owing to the powerful learning capabilities that enable the DL models to automatically process and extract patterns of disease from plant images without the need for any handcrafted features or engineered features [3,4]. Among the DL methods, convolutional neural networks (CNNs) are the dominating family of deep networks that have achieved many breakthroughs in different visual recognition tasks due to their capabilities for feature extraction stacked convolutional kernels. Recently, the vision transformers (ViTs) have evolved as state-of-the-art visual learning models that can model long-range dependencies, which enables outperforming the performance of CNNs on different vision tasks. However, their potential has still not been investigated for visual recognition tasks [5].
Knowledge distillation is a powerful technique in deep learning that enables model compression and facilitates the transfer of knowledge from a larger, complex teacher network to a smaller, more compact student network. The process involves training the student network to mimic the soft probabilities (logits) produced by the teacher network rather than learning from ground-truth hard labels [6]. By doing so, the student network learns not only from the teacher’s final predictions but also from the rich intermediate information that the teacher encodes during training [7]. This distilled knowledge acts as a form of regularization, guiding the student network to focus on important patterns and generalize better to unseen data. Knowledge distillation has shown remarkable success in model compression, allowing the creation of more lightweight networks that retain the performance of their larger counterparts. In the context of multi-task visual recognition in agriculture, knowledge distillation can be particularly advantageous, as it allows the creation of efficient and accurate models capable of handling diverse tasks with reduced computational overhead [8].
Meta-learning, popularly recognized as “learning to learn,” is a front-line learning method in machines that empowers the models to be able to adapt rapidly and successfully to new tasks with minimum training data [6]. Different from conventional learning methods, meta-learning algorithms can obtain more generalizable knowledge that could be straightforwardly applied to new and unseen tasks as they learn from a distribution of tasks rather than just one. This capacity is particularly useful in fields such as agriculture, where fluctuating weather patterns and new pests and diseases always present fresh recognition obstacles [7,8]. The integration of knowledge learned from several activities can empower meta-learners to learn a representation that is autonomous of the specific tasks being acted on. It can quickly adapt and learn from just a few samples or shots from the new activity since this meta-learner efficiently encodes knowledge about how to approach the assignment. Since the model becomes skilled at performing a variety of jobs and effectively learning from little data, meta-learning can contribute greatly to greater efficiency, speedier deployment, and better agility in the setting of multi-task visual recognition in agriculture [9,10].
The complexity of coffee farming and the breadth of tasks involved in leaf analysis make this an intriguing and challenging case study for multi-task visual recognition. The widespread crop loss caused by coffee leaf diseases is a major threat to agricultural production on multiple fronts, including yield and quality [11]. Pathogens, like fungi, bacteria, and viruses, cause these agricultural diseases by invading and damaging coffee plant tissues and organs. Hence, correct recognition of the types of these diseases, their causes, and their severity is all-important to implement effective management strategies to control the distribution of these diseases [12]. The four most important diseases that can impact coffee leaves are Cercospora leaf spot, brown leaf spot, rust, and leaf miner. The use of computer vision in precision agriculture has enabled the development of data-driven approaches with a continued emphasis on the early identification of crop losses from leaf diseases. Multi-task recognition of coffee leaf diseases in visual images is one of the many promising directions to improve disease diagnosis on smart farms [13]. Combining intelligent strategies for detecting coffee leaf disease can guarantee to safeguard long-term viability losses due to production declines resulting from coffee health reduction, thus ensuring continued superior quality satisfaction found in robust global demand for coffee products [6,7,8,9,10,11,12]. Moreover, it is vital to decide the severity of the stress impacting the coffee leaves, which is estimated by the proportion of the leaf surface that is wounded [8]. However, this multi-tasking activity is challenging, even for professionals, for many reasons, such as the absence of exact delineation of the lesion; the variability of attributes between nodules of the same category (including color, outline, and magnitude); the occurrence of numerous lesions within the same leaf; the intra- and inter-observer variability, and the reality that various stresses could have similar effects [14]. Therefore, meta-learning and knowledge distillation are particularly applicable in this setting due to the complexity of coffee leaf analysis and the accompanying necessity for effective and adaptive learning strategies. The availability or practicality of the requisite labeled data and computer resources for training in traditional deep-learning approaches may be limited in agricultural contexts. Knowledge distillation allows us to efficiently transfer know-how, reducing the volume of data needed while simultaneously improving the model’s replicability. In addition, meta-learning allows the model to learn from a variety of tasks and to adjust to novel tasks in the domain of coffee leaf cultivation [10,11,12,13,14,15].

1.1. Research Gaps

Despite the prevailing use of DL in a broad range of precision agriculture, however, these solutions usually require big datasets to be trained in an efficient manner, which is difficult to collect and annotate in real-world settings [11,12]. This, in turn, highlights the need for powerful hardware like graphical processing units (GPUs), which implies extra costs and expenses. This convergence of the above factors signifies the misalignment between the current DL solutions and the sustainability principles intrinsic to precision agriculture. Thus, it is clear that these solutions may not completely fulfill the sustainability requirements of precision agriculture, which is a great motive to further explore more resource-efficient and reasonably feasible alternatives [13]. Humans, conversely, can pick up new information and relocate fast based on a small number of examples, leading us to question whether intelligence gained from processing large datasets is the goal. Because of the low cost of a few samples of data, few-shot learning (FSL) has evolved as a paradigm to enable DL to learn from a small amount of data, and hence it is an interesting and potentially fruitful field of study for the detection of biotic stresses in coffee leaves [2]. Presently, transfer learning (TL), data augmentation, and meta-learning are the main three approaches to FSL. The purpose of TL is to acquire and apply information from one domain to another. Assuming there is adequate data in the source domain for learning, the learning algorithm would be fine-tuned by a few examples in the target domain to keep a satisfactory performance [3]. While TL solves the difficulties of model fragility and slow convergence by fine-tuning parameters, it still has some basic issues to deal with, such as the need for comparatively large datasets. For instance, the labeling and training of the classification model must be reperformed if multiple additional classification tasks are introduced. Intuitively, we can generate additional new instances or features with the help of data augmentation by rotating and scaling images, mixing them, oversampling, and other similar techniques. However, the performance improvement gained by data augmentation is still limited [4].

1.2. Novelty and Contribution

In response to the above-mentioned challenges, this study presents a new meta-learning framework for the sustainable diagnosis of coffee leaves through a few-shot estimation of biotic stresses and severity, with a powerful ability to generalize on unseen coffee data. The main contributions of this framework can be pointed out as follows:
  • Motivated by the ability of ViTs to learn long-term dependencies, a mixed vision transformer (MVT) is introduced to simultaneously extract rich contextual representations from both support and query images rather than extracting visual leaf features independently from the query and support inputs. In the MVT layer, we introduced contextual attention to capturing the significant information regarding the biotic stress and its severity degree in a way that enables perfect alignment of visual representations from both the query and support set.
  • A new meta-training strategy is presented to empower the meta-learner to adapt its learnable according to the underlying task in a way that improves the generalizability of unseen data.
  • A novel knowledge distillation is introduced in our framework to keep a balance between learning new knowledge and preservation of previously acquired knowledge in a way that removes the appalling forgetting problem.
  • Exhaustive proof-of-concept experiments are conducted to evaluate and compare the performance of the proposed framework against the state-of-the-art methods on a case study of coffee leaf datasets. The findings demonstrate the proposed is qualified to be adopted as an efficient tool for delivering insightful diagnoses for coffee diseases in smart farming environments.

1.3. Outline

The overall structure of this work is composed of six primary sections. Section 2 reviews the related literature. Section 3 debates the methodology of the proposed framework. The experimental settings of this work are given in Section 4. Section 5 presents the experimental results and the corresponding discussion. Finally, the concluding remarks of this work are provided in Section 6.

2. Literature Review

This section provides an in-depth analysis of the related literature with a primary emphasis on visual recognition studies for the detection of leaf diseases in coffee plants. Two branches of studies are discovered in this spectrum, as described in the following subsections.

2.1. DL-Based Visual Recognition

Research on multi-task visual recognition has been gaining increasing interest in recent years. However, it is still in its infancy stage. For example, Esgario et al. [5] explored the potential of CNNs for multi-task diagnosis of biotic stress from images of coffee leaves by classifying biotic stress and estimating their severity. Similarly, Putra et al. [6] applied a CNN and compared its performance against common models, such as LeNet, AlexNet, ResNet-50, and GoogleNet, for classifying between Arabica and Robusta coffee plants. In [7], a segmentation of the symptomatology was carried out using a threshold-based method to create a system that can classify the individual symptoms of coffee leaves. Manual features were used to categorize the symptoms into leaf miner and rust groups. The severity of the symptoms could be measured, and specific symptoms could be identified, thanks to the lesion segmentation findings. Despite the impressive outcomes, this method of segmentation is highly susceptible to environmental factors, like lighting changes and specular reflection, because it ignores the positional relation between the pixels. Novtahaning et al. [8] introduced an ensemble DL system that exploits the notion of TL to integrate multiple pre-trained CNNs to extract the disease patterns from the in-field images of coffee leaves. Then, the most powerful feature extractor is nominated to dominate the ensemble architecture, and its features are passed to the final classification head. Additionally, Esgario et al. [9] introduced a multi-step approach that includes a semantic segmentation step followed by a severity estimating step for detecting diseases and pests of coffee leaves. U-net architecture is used in the earlier step, while a CNN was used to implement the later step. These two steps were evaluated disjointedly to shed light on the optimistic and undesirable arguments of each one. To aid professionals and farmers in identifying and quantifying biotic stresses utilizing pictures of coffee leaves captured by mobile phones, a custom app was developed and deployed on the Android platform, including segmentation, severity computation, and symptom categorization. Similarly, Tassis et al. [10] introduced a three-stage DL system that combined different CNNs to systematize the detection of lesions presented in mobile-captured in-field imageries encompassing fragments of the coffee tree. The first stage applied Mask R-CNN to perform instance segmentation for coffee leaves. Then, the UNet and PSPNet models were applied to perform semantic segmentation at the second stage, and lastly, residual convolution networks were adopted to perform classification. In addition, Yamashita and Leite [11] emphasized developing a lightweight CNN, based on the MobileNet model, to detect coffee leaf disease on edge computing devices namely low-cost microcontroller board. It demonstrated the practicability of using CNNs for detection of diseases in coffee leaf images in resource-constrained devices with no internet connection.
A prominent and common downside for the above methods is the data-hungry nature that makes them unable to learn well from small-size data, which is a common case for coffee biotic stress. This, in turn, implies the weak generalizability of these methods for an unseen dataset.

2.2. Meta-Learning in Visual Recognition

In response to the challenges of learning from small-size data, the research community moved toward meta-learning approaches. For instance, Li and Yang [12] introduced a few-shot learning framework to identify cotton pests using only a few images, in which two datasets were used to validate the efficiency and viability of the model. This framework trained the parameters of CNN as a feature extractor according to a triplet loss function to help differentiate various pest species to guarantee the system toughness. Not only that, but the compiled CNN circuit in the FPGA and the control program in the ARM were both successfully used to obtain the model functioning in an embedded gate. In addition, Hu et al. [13], introduced a low-shot learning approach to disorder recognition in tea leaves to facilitate timely prevention and control measures. Disease regions on tea leaf images were segmented utilizing the support vector machine (SVM) technique, with color and texture features extracted. By feeding segmented disease spot images into an enhanced conditional deep convolutional generative adversarial network (C-DCGAN) for data augmentation, a VGG16 model is taught to recognize diseases in tea leaves. In [15], Li and Chao proposed a semi-supervised few-shot learning strategy for identifying plant leaf diseases, in which a subset of the publicly available PlantVillage dataset is used, and it is further subdivided into a source domain and an end-user domain. The experiments were conducted taking into account the domain separation and few-shot parameters (N-way, k-shot) to verify the rightness and generalization of developed semi-supervised few-shot methodologies. To flexibly choose the amount of unlabeled data for pseudo-labeling in the semi-supervised process, a confidence interval was used in this work. Moreover, Tseng et al. [14] looked into how model-agnostic meta-learning (MAML) can be used to gain insight from a wide variety of global datasets and boost results in areas with limited available information. The findings of this work stated that MAML outperforms pre-trained and random initial weights in a wide range of tasks and countries (including Togo, Kenya, and Brazil). It also investigated MAML’s potential benefits across a variety of target data size commands. Across a broad range of training set sizes and positive-to-negative label ratios, the findings supported that MAML performs better than alternative meta-learners, suggesting its general applicability for territory usage multiple crop mapping.
Despite the research efforts devoted to the few-shot computer vision applications, this family of solutions is still in its infancy and has not been deeply investigated in visual recognition tasks. This implies the promise of this research direction for the detection of biotic stresses in coffee leaves. Thus, in this work, we seek to fill this gap by studying meta-learning for improving the multi-task diagnosis of coffee leaf disease under a data scarcity scenario. To this end, this work presents a multi-task meta-learning framework as an important step towards flexible and efficient solutions to problems in precision agriculture that require a wide range of resources. The design principles of the proposed model validate remarkable flexibility in handling emerging tasks and obstacles, radically cutting down on the time and effort spent retraining and freeing up precious computational resources in the process. With the adaptability and multitasking abilities of our model, the process of coffee disease management may be directed with more efficiency, which makes it an asset in the context of changing agricultural dynamics, where a quick response to new issues is of utmost importance.

3. Methodological Framework

3.1. Problem Formulation

The multi-task classification of coffee leaves is designated as a few-shot learning dilemma, whereby a set of labeled coffee leaf images is given as a training set, and a set of labeled coffee leaf images is given as a testing set. Both sets share non-overlapping label space. Therefore, given a few samples from the labeled set; our objective is to learn a model that could generalize unseen coffee leaf images from the unlabeled set. To achieve this objective, meta-learning with episodic training could be applied to fine-tune the generalizability by imitating the limited data scenario met during the inference through generating stable episodes. An episode is shaped with two subsets: one contains a few labeled coffee leaf images for learning, and the other one contains labeled images, in which the labels are utilized to compute the error rate of the model’s prediction in each episode. In addition, the “N-way, k-shot” task is defined by every training episode, with N pointing to the number of biotic/severity classes in each episode, while k denotes the number of images in each class. The problem of multi-task diagnosis of coffee leaves is formulated as a generic meta-setup known as N-way, k-shot classification.
This formulation is built up with three phases described below:
  • Meta-training phase: the multi-task learner is initiated to learn from a set of meta-training data.
  • Meta-validation phase: the meta-learner leverages meta-validation data to assess its classification performance on unobserved tasks that are not encountered during the meta-training phase. This assessment estimates the meta-generalizability of the trained learner and provides feedback to update and fine-tune the parameters of the meta-learner.
  • Meta-testing phase: different from the previous phase, the meta-testing data are exploited here to evaluate the final performance of the meta-learner for both biotic stress classification and severity estimation.
In the first phase, a big enough coffee dataset is used to train the multi-task learner, A , over a finite number of episodes. For each episode, the data T j = ( D T j t r , D T j v a l ) are attained through sampling coffee leaf images ( x x i , y i ) from the complete dataset D . The tuple ( D T j t r , D T j v a l ) describes a training and a validation set of an episode, in which each set comprises a small number of coffee leaf images. According to the N-way, k-shot definition, the process of the data sampling process poses that the training set D T j t r comprise precisely N classes with k images belonging to each class, suggesting that | D T j t r | = N × k . Additionally, the factual labels of images in the validation set D T j v a l need to exist in the train set D T j t r of a certain episode  T j . The training set D T j t r = { ( x s , y s ) } s = 1 S is also a support set, because it figuratively supports the learner’s decisions on the validation set D T j v a l = { ( x q , y q ) } q = 1 Q known as query set. The multi-task learner, A , can be defined as y * = f θ x * , with * signifying s or q, and is prone to high variance because of the large dimensions of x * . Thus, coffee leaf images in both support and query sets are encoded into latent space via an embedding function Φ * = f ϕ ( x * ) . Supposing that the embedding function is static throughout the training of A on each episode, then the multi-task learner seeks to optimize the following objective:
θ = A D T j t r ; ϕ = arg m i n θ L base D T j t r ; θ , ϕ + R θ ,
The terms L base and R denote cost function and regularization factor, respectively. Here, ϕ stipulates expectations regarding ‘how to learn’, e.g., the optimizer. The goal of our meta-learning coffee leaf classifier is to train a decent embedding function, with the aim of reducing the average loss of the multi-task learner on the query set. Officially,
ϕ = arg m i n ϕ E T L meta D val ; θ , ϕ
By the completion of meta-training, the learner’s performance is evaluated on a meta-testing set S = { ( D j t r , D j t s ) } j = 1 J . The assessment is performed using the distribution of the test coffee leaf images, as follows:
E S [ L m e t a ( D t s ; θ , ϕ ) ,   with θ = A ( D t r ; ϕ ) ] .

3.2. Mixed Vision Transformer (MVT)

This section argues for the design of our multi-task learner backbone that seeks to handle the challenge of generalizability on unseen types of biotic stresses from the interactive standpoint. Figure 1 displays the construction of our multi-task learner, comprising three chief learnable modules and dual classification heads. The design of our MVT extractor consists of a stack of three MVT blocks for multiparty contextual representation learning of the support and query images. At the early block of the backbone, there is a batch of support images, X s R B s × H x s × w x s × 3 , and one query image, X q R B q × H x q × w x q × 3 . Inspired by ViT [16], the proposed MVT block is composed of the contextual multi-head contextual attention, two fully connected layers (FCLs) with residual connectivity, LayerNorm (LN), as well as Gaussian error linear unit (GELU) activation. Also, the design of the backbone feature extractor follows the non-hierarchical design, in which the MVT block shares the same learnable parameters, embedding and manipulating feature maps with shared dimensions. In this case, our model handles the adjustable extent of the token sequence.
The next MVT blocks share a comparable structure and calculate representational maps through the gradual reduction in sequence lengths and expansion of the number of channels. In standard ViT, a patching operation was applied to split the raw images into a group of non-overlapping patches. Next, a linear embedding layer is applied to flatten the generated patches into C 1 dimensions. Then, positional embedding is generated for those patches before being passed into the ViT layer. However, traditional patch embedding suffers from discontinuity that limits the ability to model local features. To overcome this limitation, an interactive patch embedding (IPE) block is built up by stacking multiple strided convolution layers of k = 3 × 3 convolution, switch normalization (SN) [17], and GELU. The patch size is computed as P = 2 k . This design enables conserving significant low-level features and local spatial information in coffee leaf images because of the injection of inductive biases by convolutions. Moreover, the IPE block empowers the MVT to have more elasticity over original ViTs, by alleviating the constraints such that the input dimensions (whether in query or support set) are firmly dividable by the predefined patch size. The IPE uses two affine transformations, ℵ, prior to and following the stacked convolutions for rescaling and shifting the image feature, hence making the learning performance steadier on small datasets, which is common for the coffee leaf dataset. The pipeline of IPE can be expressed as follows (Algorithm 1):
Algorithm 1: Interactive Patch Embedding (IPE)
1IPE(x):
2 x = Diag α x + β
3For i = 1 , , k do:
4   F i x = GELU SN Conv 3 × 3 x
5 x = F k F 2 F 1 x
6 x = Reshape ( x )
7Return  x
In our MVT block, the IPE is applied to generate embedding for both support and query images as follows.
X s = I P E ( X s ) X q = I P E ( X q )
For each attention head, h = 1 , , H , the embedding of support X s and X q are mapped to Q s i , K s i , V s i and Q q i , K q i , V q i , respectively. To maintain lightweight computation for our attention, a strided depth-wise separable convolution (dwconv) is introduced to promote the subsampling of visual representation maps for matrices K and V.
Q q i = X q W Q i , & Q s i = X s W Q i
K q i = d w c o n v s t r = 2 X q W K i ,   & K s i = d w c o n v s t r = 2 X s W K i
V q i = d w c o n v s t r = 2 X q W V i ,   & V s i = d w c o n v s t r = 2 X s W V i
In the above formula, W Q i R C 1 × d h ,   W K i R C 1 × d h ,   and W V i R C 1 × d h denote the trainable parameters shared for the projection layer for both input sets. d h = C 1 H denotes the dimension of the projected features.
In our case, the size of the support batch varies from the size of the query batch. Also, biotic detection or/and severity assessment is performed for each query instance disjointedly, since query imageries are unrelated, and thereby the recognition of each is independent of others. The simple way to address the variation in sizes is to direct a pair of images (one from query and one from support) at each time; however, replicating this procedure for each image in the support set leads to high time complexity. Thus, we propose enabling contextual attention to consider the attention between the query image and all images belonging to the same class in the support set. The contextual attention mechanism is introduced into the MVT layer to fuse the key-value information related to query and support input attention. To fuse the key-value information from the support-related feature maps to the query-related feature maps and fuse the key-value information from the query-related feature maps to the support-related feature maps, a pointwise convolution is applied to update the channel dimensions of maps to be matched and then the key-value information of both inputs is concatenated.
Q ˜ q i = K q i , 1 B s B s   Q s i , V ˜ q i = V q i , 1 B s B s   V s i
Q ˜ s i = K s i , 1 B q B q   Q q i ,    V ˜ s i = V s i , 1 B q B q   V q i
The term · represents the concatenation layer. To sum up, the computation of multi-head contextual attention can be summarized as follows:
X q = Concat h q 1 , , h q H W O
h q i = Attention Q ˜ q i , K q i , V ˜ q i
X s = Concat h s 1 , , h s h W O
h s i = Attention Q ˜ q i , K q i , V ˜ q i
Next, the linear layers are applied to each patch with robust feature representations, while the residual connection is used to improve the gradient flow.
X q = L i n e a r LN X q + X q X s = L i n e a r LN X s + X s

3.3. Adaptive Meta-Training

The primary objective of meta-training is to allow the multi-task learner to perfectly train an embedding function, f ϕ , to be able to generalize well on new, unseen coffee data. Instead of rebuilding and/or retraining new learners to learn the embedding function, our framework suggests that the pre-training of vision learners on single- or multi-task data is favorable for empowering the embeddings of the underlying multi-task learner. Thus, the meta-training data were combined from all episodes according to the following formula:
D c o m b i n e = { ( x i , y i ) } k = 1 K = D 1 t r , , D i t r , , D I t r ,
Next, the embedding function can be defined as follows:
ϕ = arg m i n ϕ E T L cf D combine ; θ , ϕ
where the L cf represent the categorical focal loss [18] between the learner’s predictions and coffee leaves labels.
Different from early meta-learning mechanisms, the proposed framework seeks to optimize an adaptive cost function to enforce the inner-loop optimization process to achieve high generalizability for new episodes during the training. To this end, an adaptive cost function L ϕ · is presented and updated with a small learner with trainable parameters, ϕ . This way, the inner-loop update can be formulated as follows:
θ i , j + 1 = θ i , j α θ i , j L ϕ τ i , j
The term τ i , j represents the state of the episode T i at time-step j, which is always defined by the training set in the conventional meta-learning paradigm. Various episodes might require including some regularization factors or secondary cost functions throughout training to generalize well. Unlike gradient descent adaptation, which exhibits high computing complexity, we proposed to take advantage of deformable transformation for the creation of adaptive feature interactions, thereby making the updates of the inner loop more adaptive. This can be expressed as follows:
ϕ = γ ϕ + π
where ϕ symbolizes the parameters of the multi-task learner and γ denotes π the deformation parameters made by the learner g τ j   ; ψ with parameter ψ . The meta-training process seeks to enable the multi-task learner to generalize across diverse episodes through optimizing the parameters θ ,   ϕ ,   a n d   ψ , and then we perform outer-loop optimization for episode T i using their own learner and query samples as formulated below:
( θ , ϕ , ψ ) ( θ , ϕ , ψ ) η ( θ , ϕ , ψ ) τ   L D i t s ; θ i
Knowledge distillation [19] is a popular technique for knowledge transmission from a robust teacher–learner to a minor student learner. At each step of the model’s learning process, a distillation loss function is required to make sure that knowledge retaining and acquisition are balanced in the best way possible. Rather than directly applying the embedding function on the meta-testing set, the learned knowledge in the embedding function is distilled into a new embedding function sharing the same building structure. Then, we train the new model with a set of parameters ϕ to optimize dual-task focal loss and the Jensen–Shannon (JS) divergence between outputs and soft targets:
ϕ = arg m i n ϕ α L c f D c o m b i n e ; ϕ + β   J S f D c o m b i n e ; ϕ , f D c o m b i n e ; ϕ
where usually β = 1 − α.
To evade the problem of disastrous forgetting in our framework, the above distillation loss is integrated as part of our loss function to empower the learner of the current episode to keep achieving good performance similar to the learners’ performance in the previous episodes. The JS distillation is attributed to its ability to deal with probability distributions containing extreme values which is often the case in multi-task learning scenarios, in which there is a high need to balance the contributions of different tasks, mitigating the impact of outliers or noisy data. In addition, JS divergence possesses the desirable properties of symmetry and smoothness, which guarantees that the order of the input distributions does not affect the result. JS divergence can be interpreted as a softened form of discrimination between probability distributions. In our design, we follow the approach from Born again [20] to put on JS successively to engender manifold generations. For each episode, the knowledge of the embedding function at the k t h generation is transmitted to the embedding function of the next generation:
ϕ k = arg m i n ϕ α L c f D c o m b i n e ; ϕ + β J S f D c o m b i n e ; ϕ   , f D c o m b i n e ;   ϕ k
This operation is repeated for K times, and the ϕ K is designated as an embedding function to learn discriminative features.
For each meta-testing tuple ( D j t r , D j t s ) , the multi-task learner is instantiated as Bayesian multivariate logistic regression with weight W and bias parameters, b , as defined below:
θ = arg m i n { W , b } t = 1 T   L t c f W f ϕ x t + b , y t + R ( W , b )

4. Experimental Materials and Setups

This section provides an in-depth discussion of the experimentation of this work in terms of the coffee data adopted to evaluate the proposed model, the competing baselines with which our model is compared, the evaluation indicators used in experiments, and the implementation setups.

4.1. Materials

In this work, we train and evaluate the proposed model on the multi-task biotic stresses dataset from [5]. It contains images of Arabica coffee leaves exaggerated by the core biotic stresses that influence the coffee tree, which was captured using different smartphones at diverse times of the year in the state of Espírito Santo, Brazil. The photos were taken from the abaxial (lower) side of the leaves under partially controlled conditions and placed on a white background. In addition to that, in-field images were not used due to the occlusion of leaf parts and their inclination concerning the camera, making the disease severity estimation infeasible in several cases. The images were captured from the underside, or abaxial side, of the leaves and then superimposed onto a white background for viewing. In addition, in-field imagery was not leveraged because of the obstruction of leaf slices and their predisposition relative to the camera, which makes the estimate of disease severity impracticable in some cases. There were no criteria used during image collection, which resulted in a more diverse dataset. Leaf miner, brown leaf spot, rust, and Cercospora leaf spot were just some of the biotic challenges experienced by the Arabica coffee plants that led to the collection of 1747 photos of healthy and diseased leaves. An expert used the acquired photos to perform the procedure of biotic stress identification for dataset labeling. Two datasets were created from the collected pictures: symptom-only pictures and the original, full-leaf images from which they were extracted. The following paragraphs provide descriptions of each dataset.
  • Leaf dataset: There are complete-leaf photos from the original source, and they have been annotated with information about the most common biotic stress and its intensity. There were 372 photos showing leaves experiencing multiple stressors, of which 62 indicated stresses of similar severity. Given that this study can only identify a single stress per leaf, multi-stress categorization is outside the purview of this work. Since it is challenging to visually determine which stress is predominant, these 62 photos with equal severity were not used on this dataset. The severity of stress was determined by applying automatic image processing techniques, such as a mask, to isolate symptoms and identify which leaves are stressed. All picture segmentation findings were manually checked for accuracy. Images with inadequate segmentation were removed from the severity calculation and were examined manually by a trained professional using visual estimate techniques. This led to five degrees of severity, namely healthy, very low, low, high, and very high.
  • Symptom dataset: This dataset was produced by selectively cropping the original photos so that just a single stress was visible in each. The cropping process involved 2147 photos of symptoms.
To make things simpler, for both tasks, we provide a detailed description of the class distribution of the data, presented in Table 1. Moreover, samples for different types of coffee leaves from the above datasets are provided in Figure 2.

4.2. Baselines

The evaluation of the proposed method should involve a comparison with the state-of-the-art meta-learning algorithms. To this end, this experimental part of this work used six meta-learning methods as our baselines, namely SiameseNet [21], TripletNet [22], ProtoNet [23], MatchingNet [24], MAML [25], and RelationNet [26]. For each baseline, three vision transformers are applied as a backbone, namely MobileViTv2 [27], LeViT [28], and Mobile-Former [29].

4.3. Evaluation Metrics

The evaluation of the performance of deep learning models for both classification and severity estimations is designated using a different set of metrics. Four of them (i.e., accuracy, precision, recall, F1-measure) are calculated according to the confusion matrix comprising true positive (TP), true negative (TN), false positive (FP), and false negatives (FN). These metrics are computed as follows:
A c c u r a c y = T P + T N T P + T N + F P + F N
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 m e a s u r e = 2 R e c a l l × P r e c i s i o n R e c a l l + P r e c i s i o n
The area under the receiver operating characteristic (AUROC) curve [30] is also reported as a numerical metric for model evaluation in both tasks.

4.4. Environmental Setups

The implementation of deep learning models is coded with the Pytorch library running on Python 3.8.0 environment. All experimental tools are fixed on a Dell workstation equipped with 128 GB RAM, Intel® Xeon® Silver 4316 Processor (30 M Cache, 2.30 GHz), and operated with Windows 10 64-bit OS. The learning of the models is accelerated with Nvidia GeForce RTX 3090. To guarantee consistent comparative results, experiments are run with standardized hyper-parameter settings. The specifics of model hyper-parameters are presented in Table 2.

5. Results and Discussion

To determine how well the proposed models fit the challenges of biotic stress identification and severity estimate, a different set of experiments was carried out in this section. The findings are reported in the following parts, for both datasets (coffee leaves and symptoms).

5.1. Comparative Analysis

In this section, we provide a detailed discussion of the numerical analysis of the experimental comparisons between the proposed method and competing baselines. For each dataset, we report the performance of the models under five-way one-shot and under five-way five-shot scenarios. The experimental conditions are kept constant for all experiments to maintain the fairness of the comparisons. The results of comparative experiments are reported in terms of the average and standard deviation of the results from training data folds. In Table 3, we present the quantitative results obtained from the different methods with five-way five-shot training on the leaf dataset. The tabulated results represent the performance of the models on coffee biotic stress classification as well as the severity estimation tasks. Notably, the SiameseNet achieves the lowest stress classification accuracy (ranging from 91.73% to 92.49%) for all backbones, and it also achieves the lowest severity estimation accuracy (ranging from 88.38% to 90.55%) for all backbones. Comparatively, the TripletNet can achieve improved stress classification accuracy but achieve similar severity estimation accuracy except for Mobile-Former (with 92.38% accuracy). This reflects that the appropriate selection of backbone network is important to severity estimation. In addition, ProtoNet can notably achieve higher results than TripletNet across all performance indicators for both tasks. Similar performance can be observed for MatchingNet with biotic stress classification accuracy in the range from/to and severity estimation accuracy in the range from/to. The performance of RelationNet is similar to TripletNet in the biotic stress classification task; however, this is not the case for the severity estimation task. Further, it is worth noting that the NAML achieved the highest performance (with accuracy between 96.15 and 96.69), overcoming the other baselines for both tasks. This reflects the advantage of combining optimization-based meta-learning into FSL for improving the generalizability of the underlying model. More importantly, the proposed model is achieving remarkable improvement over all baselines reflecting the ability of our model to learn the discriminatory features necessary for biotic stress classification in coffee leaves and simultaneously to learn the severity attributes from small data scenarios.
To further interpret and analyze the performance of the proposed model (five-way five-shot), Figure 3 provides a visualization of the class-level performance by plotting the confusion matrices for both classification and severity estimation tasks. For coffee biotic stress classification, it is worth noting that the class “rust” attains the lowest detection performance with 97.2% precision, while the classes “healthy and Brown leaf spot and Cercospora leaf spot” attain the highest detection performance with 100% precision. For severity estimation, it is worth noting that the class “very high” attains the lowest detection performance with 90% precision, while the class “very low” attains the highest detection performance with 98.2% precision. These findings coincide with the findings from previous studies [2,5], where the lowest performance was attained in this same class. This can be attributed to the high resemblance between lesions in these classes. However, different from the previous studies [2,5], the proposed model can recognize the different types of stresses with high confidence, which further demonstrates the discriminatory power of our model.
Using the leaf dataset for five-way one-shot training, the quantitative results of biotic stress categorization using various approaches are shown in Table 4. The results demonstrated the effectiveness of the models in classifying and estimating the severity of biotic stresses on coffee. SiameseNet uses Mobile-Former as its backbone to achieve the lowest stress classification accuracy (89.48%) and the lowest severity estimation accuracy (87.8%). While TripletNet outperforms the other backbones and SiameseNet by a small margin (93.5%) when trained using LeViT for stress classification, the other backbones perform about the same. Based on these findings, it appears that selecting the right backbone network is crucial for the multi-task diagnosis of coffee leaves. Furthermore, the ProtoNet performs better than the TripletNet on both tasks across the board. It is also noteworthy that the NAML, MatchingNet, and RelationNet all outperform the other baselines by around the same margins on both tasks. Our model’s remarkable ability to learn multi-task features allows it to significantly beat all baselines. Although it was expected that model performance would suffer when going from five shots to one, the findings show that the suggested model still does well in a five-way one-shot scenario.
To further understand and scrutinize the behavior of our model (five-way one-shot), Figure 4 provides a visualization of the class-level performance by plotting the confusion matrices for both classification and severity estimation tasks. For coffee biotic stress classification, it is worth noting that the class “Brown leaf spot” attains the lowest detection performance with 95.7% precision, while the class “healthy” attains the highest detection performance with 98.1% precision. For severity estimation, it is worth noting that the class “very high” attains the lowest detection performance with 91% precision, while the class “high” attains the highest detection performance with 95% precision. This can be attributed to the class imbalance or the in-between severity levels that cause the model to be confused.
In Table 5, we present the quantitative results obtained from the different methods under both five-way five-shot and five-way one-shot training on the symptom dataset. The tabulated results only represent the performance of the models on coffee biotic stress classification. It is worth noting that the NAML achieves the highest performance among other baselines, while SiameseNet gets the lowest performance. This observation applies to both five-way one-shot as well as five-way five-shot settings. Conforming to our findings on the leaf dataset, the proposed model shows competing for performance outperforming overall baselines on the symptom dataset. Notably, the competing baselines don’t show significant performance gain when increasing the number of shots. The same behavior applies to the proposed model, in which the classification performance only reduces by 1% when shifting from five shots to one shot. This can be attributed to the fact that the leaf dataset encompasses only a few categories, which makes the learner not require many support instances per category to achieve good generalization.
In Figure 5, we present the confusion matrix of the proposed model on the symptom dataset. For five-way five-shot settings, it is observable that the proposed model recognizes different types of biotic stresses with an average precision above 98%. For the five-way one-shot setting, it is worth noting that the class “Brown leaf spot” attains the lowest detection performance with an average precision of 94.8%, while the class “Leaf miner” attains the highest detection performance with an average precision of 99.1%. This is explained by the fact that the variation in the representativeness of support symptom samples has a significant impact on the class-level performance and overall model performance.
In Table 6, we display the results of two-tailed t-test experiments applied to assess and statistically analyze the role significance of achieved improvements under both five-way five-shot and five-way one-shot settings. In these experiments, we use a threshold value of 0.05. As observed, the statistical results reveal the remarkable classification accuracy in categorizing coffee leaf diseases and determining their severity level. The statistical significance is observed for both types of settings validating the robustness of our model and its potential to address the complications of multi-task leaf diagnosis efficiently. Based on these outcomes, we can conclude that the performance improvements achieved by our approach are realistic and not achieved by chance, which proves the competitive advantage of our mode over the competing models.

5.2. Ablation Analysis

In this section, the results from ablation experiments are discussed to explain the contribution of different building blocks to the final performance of the proposed model. In this context, a set of learner ablation experiments is implemented to compare the performance of the MVT to other multi-path learners from the literature. In Figure 6, the results of these experiments on the leaf dataset are reported for both five-way five-shot and five-way one-shot settings. It is worth noting that the proposed MVT enables our model to achieve higher performance on biotic stress classification and severity estimation tasks. The findings further demonstrate the efficiency of MVT. This can be attributed to the ability of MVT to extract the fine-grained features from both support and query images while considering the feature interaction among them.
Moreover, to interpret the performance of our MVT, a comprehensive ablation analysis is conducted to evaluate the efficacy of our IPE against the conventional PE, as shown in Figure 7. This experimental analysis is expected to elucidate the distinct contributions of IPE to the overall model performance. As shown, the detection performance improvements achieved by our IPE module, demonstrated its effectiveness in modeling complex spatial relations and context representations within leaf images. Amazingly, the results on the severity estimation task conform with the findings on the detection task, which further supports our claims regarding the role of the IPE module. These findings collectively lay the groundwork for future research intended to enhance embedding strategies in the context of multi-task leaf diagnosis, promoting an in-depth interpretation of the complicated associations between spatial information and accurate disease classification.
According to the above findings, it is worth noting that our work has important ramifications for the social and economic fabric of rural areas in addition to their experimental merits. The proposed framework has the potential to completely alter the coffee farming environment, especially in areas where the industry is vital to the economy. With the quick and accurate multi-task diagnosis of coffee leaf diseases, the proposed model can provide farmers with the data they need to take corrective and preventative steps. The consequence is improved food security for populations that rely heavily on coffee for both economic and nutritional support. The financial security of farmers is bolstered by higher yields, and their ability to weather agricultural crises is strengthened [30,31,32]. Our findings support sustainable agriculture’s broader mission to improve the lives of farmers and rural communities while also bolstering regional economies and ensuring a reliable supply of staple food. Though the origins of our study are in the fields of machine learning and multi-task leaf diagnostics, its implications are substantially more general. The potential of our model to solve real-world difficulties faced by coffee farmers increases the program’s potential social and economic impact. The capacity to correctly identify biotic stress affecting coffee leaves is vital not just for effective disease management but also for the continued existence of agricultural communities. Since correct diagnosis improves harvests, it helps ensure that people in coffee-growing areas have access to a safe and sufficient food supply. As a result of increasing yields and farmer incomes, the economy stands a better chance of remaining stable thanks to this productivity boost. Our novel strategy has contributed to a robust and sustainable agriculture sector [33,34,35,36].

6. Conclusions

This work presents a novel meta-learning approach for a few-shot multi-task diagnosis of biotic stresses from coffee leaf images. A new MVT network is proposed to perform mixed contextual attention on both query and support images of the meta-training dataset. An adaptive meta-training strategy is introduced based on an objective function that acclimates to each episode according to its state throughout the inner-loop optimization, hence empowering the model’s generalizability. The experimental results demonstrated that the proposed targeted approach not only enhances the diagnosis and management of coffee leaves but also empowers coffee farmers with efficient and accurate tools to combat diseases, safeguard crop productivity, and promote sustainable practices. The proposed holistic approach enables site-specific disease management policies tailored to the exceptional conditions of each coffee plantation. However, the centralized scenario of our model remains a significant limit to the use and deployment of our model in real-world smart farms which are geographically distributed across different locations.
In line with continuous advancements in the field of multi-task coffee leaf diagnosis, many promising avenues are envisioned for future exploration. Firstly, our knowledgeable meta-learning approach can be fine-tuned to optimize its adaptability to a variety of crop diseases and varied environmental settings. In addition, the continuous evolution of sustainable agriculture highlights the importance of integrating real-time monitoring and IoT technologies into the proposed approach to allow timely and data-driven decision-making for farmers. The potential for our investigation to grow to incorporate remote sensing data and satellite pictures for disease forecasting and early diagnosis is quite encouraging. We also plan to work with agronomists and ecologists to better understand the intricate interplay between crop health and environmental factors through the development of interdisciplinary collaborations. Furthermore, Figure 8 displays the comprehensive ablation analysis that thoroughly assesses the performance of our model when implemented with the proposed adaptive meta-training, and when implemented with the standard meta-training approach. As shown, the experimental results show that adaptive meta-training reveals a noteworthy advantage and demonstrates superior adaptability to developing task distributions and complexities. With the dynamic adjustment ability of our meta-learner, our approach excels in capturing task-specific nuances, thus showcasing higher accuracy and robustness in multi-task leaf diagnosis. On the other hand, with the conventional meta-training approach, our model suffers from limited adaptability in the face of changing circumstances.

Author Contributions

Methodology, A.A.S. and W.T.A.-N.; Software, A.A.S. and W.T.A.-N.; Validation, A.A.S. and W.T.A.-N.; Investigation, A.A.S.; Data curation, A.A.S.; Writing—original draft, A.A.S.; Writing—review & editing, W.T.A.-N.; Supervision, A.A.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets generated during and/or analyzed during the current study are not publicly available due to the privacy-preserving nature of the data but are available from the corresponding author upon reasonable request.

Acknowledgments

The authors extend their appreciation to the Deputyship for Research & Innovation, Ministry of Education, in Saudi Arabia for funding this research work through the project number ISP22-35.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Han, K.; Wang, Y.; Chen, H.; Chen, X.; Guo, J.; Liu, Z.; Tang, Y.; Xiao, A.; Xu, C.; Xu, Y.; et al. A Survey on Vision Transformer. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 87–110. [Google Scholar] [CrossRef]
  2. Tassis, L.M.; Krohling, R.A. Few-shot learning for biotic stress classification of coffee leaves. Artif. Intell. Agric. 2022, 6, 55–67. [Google Scholar] [CrossRef]
  3. Zhao, W.; Yamada, W.; Li, T.; Digman, M.; Runge, T. Augmenting crop detection for precision agriculture with deep visual transfer learning—A case study of bale detection. Remote Sens. 2021, 13, 23. [Google Scholar] [CrossRef]
  4. Nitin, S.B.; Gupta, R.K.; Yadav, F.; Bovand; Tyagi, P.K. Developing precision agriculture using data augmentation framework for automatic identification of castor insect pests. Front. Plant Sci. 2023, 14, 1101943. [Google Scholar] [CrossRef] [PubMed]
  5. Esgario, J.G.M.; Krohling, R.A.; Ventura, J.A. Deep learning for classification and severity estimation of coffee leaf biotic stress. Comput. Electron. Agric. 2020, 169, 105162. [Google Scholar] [CrossRef]
  6. Putra, B.T.W.; Amirudin, R.; Marhaenanto, B. The Evaluation of Deep Learning Using Convolutional Neural Network (CNN) Approach for Identifying Arabica and Robusta Coffee Plants. J. Biosyst. Eng. 2022, 47, 118–129. [Google Scholar] [CrossRef]
  7. Mansoa, G.L.; Knidela, H.; Krohlinga, R.A.; Venturad, J.A. A smartphone application to detection and classification of coffee leaf miner and coffee leaf rust. arXiv 2019, arXiv:1904.00742. [Google Scholar]
  8. Novtahaning, D.; Shah, H.A.; Kang, J.-M. Deep Learning Ensemble-Based Automated and High-Performing Recognition of Coffee Leaf Disease. Agriculture 2022, 12, 1909. [Google Scholar] [CrossRef]
  9. Esgario, J.G.M.; de Castro, P.B.C.; Tassis, L.M.; Krohling, R.A. An app to assist farmers in the identification of diseases and pests of coffee leaves using deep learning. Inf. Process. Agric. 2022, 9, 38–47. [Google Scholar] [CrossRef]
  10. Tassis, L.M.; de Souza, J.E.T.; Krohling, R.A. A deep learning approach combining instance and semantic segmentation to identify diseases and pests of coffee leaves from in-field images. Comput. Electron. Agric. 2021, 186, 106191. [Google Scholar] [CrossRef]
  11. Bordin Yamashita, J.V.Y.; Leite, J.P.R.R. Coffee disease classification at the edge using deep learning. Smart Agric. Technol. 2023, 4, 100183. [Google Scholar] [CrossRef]
  12. Li, Y.; Yang, J. Few-shot cotton pest recognition and terminal realization. Comput. Electron. Agric. 2020, 169, 105240. [Google Scholar] [CrossRef]
  13. Hu, G.; Wu, H.; Zhang, Y.; Wan, M. A low shot learning method for tea leaf’s disease identification. Comput. Electron. Agric. 2019, 163, 104852. [Google Scholar] [CrossRef]
  14. Tseng, G.; Kerner, H.; Nakalembe, C.; Becker-Reshef, I. Learning to predict crop type from heterogeneous sparse labels using meta-learning. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Nashville, TN, USA, 19–25 June 2021; pp. 1111–1120. [Google Scholar] [CrossRef]
  15. Li, Y.; Chao, X. Semi-supervised few-shot learning approach for plant diseases recognition. Plant Methods 2021, 17, 1–10. [Google Scholar] [CrossRef] [PubMed]
  16. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Antiga, L. Attention Is All You Need. Adv. Neural Inf. Process. Syst. 2019, 13. [Google Scholar]
  17. Luo, P.; Zhang, R.; Ren, J.; Peng, Z.; Li, J. Switchable Normalization for Learning-to-Normalize Deep Representation. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 712–728. [Google Scholar] [CrossRef] [PubMed]
  18. Charoenphakdee, N.; Vongkulbhisal, J.; Chairatanakul, N.; Sugiyama, M. On Focal Loss for Class-Posterior Probability Estimation: A Theoretical Perspective. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 21 June 2021. [Google Scholar] [CrossRef]
  19. Wang, L.; Yoon, K.J. Knowledge Distillation and Student-Teacher Learning for Visual Intelligence: A Review and New Outlooks. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 3048–3068. [Google Scholar] [CrossRef]
  20. Furlanello, T.; Lipton, Z.C.; Tschannen, M.; Itti, L.; Anandkumar, A. Born-Again Neural Networks. In Proceedings of the 35th International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018. [Google Scholar]
  21. Zhu, Z.; Wang, Q.; Li, B.; Wu, W.; Yan, J.; Hu, W. Siamese Neural Networks for One-shot Image Recognition Gregory. ICML Deep Learn Work. 2018, 2, 2015. [Google Scholar]
  22. Hoffer, E.; Ailon, N. Deep metric learning using triplet network. In Similarity-Based Pattern Recognition: Third International Workshop; Springer International Publishing: Berlin/Heidelberg, Germany, 2015. [Google Scholar] [CrossRef]
  23. Snell, J.; Swersky, K.; Zemel, R. Prototypical networks for few-shot learning. Adv. Neural Inf. Process. Syst. 2017, 30. Available online: https://proceedings.neurips.cc/paper_files/paper/2017/file/cb8da6767461f2812ae4290eac7cbc42-Paper.pdf (accessed on 20 August 2023).
  24. Vinyals, O.; Blundell, C.; Lillicrap, T.; Kavukcuoglu, K.; Wierstra, D. Matching networks for one shot learning. Adv. Neural Inf. Process. Syst. 2016, 29. Available online: https://proceedings.neurips.cc/paper/2016/file/90e1357833654983612fb05e3ec9148c-Paper.pdf (accessed on 20 August 2023).
  25. Antoniou, A.; Storkey, A.; Edwards, H. How to train your MAML. arXiv 2019, arXiv:1810.09502. [Google Scholar]
  26. Sung, F.; Yang, Y.; Zhang, L.; Xiang, T.; Torr, P.H.S.; Hospedales, T.M. Learning to Compare: Relation Network for Few-Shot Learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar] [CrossRef]
  27. Mehta, S.; Rastegari, M. Separable Self-attention for Mobile Vision Transformers. arXiv 2022, arXiv:2206.02680. [Google Scholar]
  28. Graham, B.; El-Nouby, A.; Touvron, H.; Stock, P.; Joulin, A.; Jégou, H.; Douze, M. LeViT: A Vision Transformer in ConvNet’s Clothing for Faster Inference. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 17 April 2021. [Google Scholar]
  29. Chen, Y.; Dai, X.; Chen, D.; Liu, M.; Dong, X.; Yuan, L.; Liu, Z. Mobile-Former: Bridging MobileNet and Transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 August 2021. [Google Scholar]
  30. Hanley, J.A.; McNeil, B.J. The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiology 1982, 143, 29–36. [Google Scholar] [CrossRef] [PubMed]
  31. Abdelhafeez, A.; Aziz, A.; Khalil, N. Building a Sustainable Social Feedback Loop: A Machine Intelligence Approach for Twitter Opinion Mining. Sustain. Mach. Intell. J. 2022, 1. [Google Scholar] [CrossRef]
  32. Abdel-Monem, A.; Abouhawwash, M. A Machine Learning Solution for Securing the Internet of Things Infrastructures. Sustain. Mach. Intell. J. 2022, 1. [Google Scholar] [CrossRef]
  33. Ali, A.M.; Abdelhafeez, A. DeepHAR-Net: A Novel Machine Intelligence Approach for Human Activity Recognition from Inertial Sensors. Sustain. Mach. Intell. J. 2022, 1. [Google Scholar] [CrossRef]
  34. Hassan, M.M.; Mollick, S.; Yasmin, F. An unsupervised cluster-based feature grouping model for early diabetes detection. Healthc. Anal. 2022, 2, 100112. [Google Scholar] [CrossRef]
  35. Yasmin, F.; Hassan, M.M.; Zaman, S.; Aung, S.T.; Karim, A.; Azam, S. A Forecasting Prognosis of the Monkeypox Outbreak Based on a Comprehensive Statistical and Regression Analysis. Computation 2022, 10, 177. [Google Scholar] [CrossRef]
  36. Yasmin, F.; Hassan, M.M.; Hasan, M.; Zaman, S.; Angon, J.H.; Bairagi, A.K.; Changchun, Y. AQIPred: A Hybrid Model for High Precision Time Specific Forecasting of Air Quality Index with Cluster Analysis. Hum.-Centric Intell. Syst. 2023, 3, 275–295. [Google Scholar] [CrossRef]
Figure 1. Systematic diagram of the architecture of the proposed meta-learning approach for multi-task few-shot classification of coffee leaves.
Figure 1. Systematic diagram of the architecture of the proposed meta-learning approach for multi-task few-shot classification of coffee leaves.
Sustainability 15 16791 g001
Figure 2. Visualization of samples from different classes in coffee leaf dataset and coffee symptoms dataset.
Figure 2. Visualization of samples from different classes in coffee leaf dataset and coffee symptoms dataset.
Sustainability 15 16791 g002
Figure 3. Confusion matrix of the proposed model on leaf dataset for coffee biotic stress classification (Left) and severity estimation (Right) in 5-way 1-shot setting.
Figure 3. Confusion matrix of the proposed model on leaf dataset for coffee biotic stress classification (Left) and severity estimation (Right) in 5-way 1-shot setting.
Sustainability 15 16791 g003
Figure 4. Confusion matrix of the proposed model on leaf dataset for coffee biotic stress classification (Left) and severity estimation (Right) in 5-way 1-shot setting.
Figure 4. Confusion matrix of the proposed model on leaf dataset for coffee biotic stress classification (Left) and severity estimation (Right) in 5-way 1-shot setting.
Sustainability 15 16791 g004
Figure 5. Confusion matrix of the proposed model on symptom dataset in 5-way 5-shot setting (Left) and in 5-way 1-shot setting (Right).
Figure 5. Confusion matrix of the proposed model on symptom dataset in 5-way 5-shot setting (Left) and in 5-way 1-shot setting (Right).
Sustainability 15 16791 g005
Figure 6. Results of learner ablation experiments on leaf dataset for coffee biotic stress classification (Left) and severity estimation (Right).
Figure 6. Results of learner ablation experiments on leaf dataset for coffee biotic stress classification (Left) and severity estimation (Right).
Sustainability 15 16791 g006
Figure 7. Ablation Analysis of Model Performance with IPE vs. Standard PE.
Figure 7. Ablation Analysis of Model Performance with IPE vs. Standard PE.
Sustainability 15 16791 g007aSustainability 15 16791 g007b
Figure 8. Ablation Analysis of Model Performance with Adaptive Meta-training vs. Standard Meta-training.
Figure 8. Ablation Analysis of Model Performance with Adaptive Meta-training vs. Standard Meta-training.
Sustainability 15 16791 g008
Table 1. Summary of class distribution of the dataset.
Table 1. Summary of class distribution of the dataset.
Biotic StressHealthyLeaf MinerRustBrown Leaf SpotCercospora Leaf SpotTotal
Leaf dataset2723875313481471685
Symptom dataset2565939915043782722
SeverityHealthyVery lowLowHighVery high/
Leaf dataset272924332101561685
Table 2. Summary of training hyper-parameters in our benchmark.
Table 2. Summary of training hyper-parameters in our benchmark.
Hyper-ParametersAssigned Value
Batch Size32
# Epochs80
Input size244 × 244 × 3
Patch size16
OptimizersAdamW
Loss functionFocal loss
# Attention heads8
Learning rate0.001
Decay0.0003
Kernel3 × 3
Linear128
Dropout rate0.3
Table 3. Comparison of numerical results of the proposed model against baselines on leaf dataset in 5-way 5-shot setting.
Table 3. Comparison of numerical results of the proposed model against baselines on leaf dataset in 5-way 5-shot setting.
Biotic Stress ClassificationSeverity Estimation
FSL MethodBackboneAccuracyPrecisionRecallF1-ScoreAUCAccuracyPrecisionRecallF1-ScoreAUC
SiameseNetMobileViTv292.34 ± 5.2292.88 ± 3.1092.48 ± 3.7792.68 ± 3.8995.61 ± 2.7988.38 ± 2.7289.06 5.9690.68 3.4889.86 3.7493.09 ± 6.00
LeViT92.49 ± 5.2291.49 ± 2.5992.3 ± 2.2391.9 ± 3.4696.55 ± 2.2890.55 ± 2.2787.79 ± 5.1391.08 ± 3.8789.41 ± 3.1496.38 ± 2.12
Mobile-Former91.73 ± 2.0991.88 ± 4.2792.24 ± 3.0392.06 ± 2.8195.5 ± 2.1490.37 ± 2.0589.99 ± 3.6388.71 ± 4.0489.35 ± 2.7994.31 ± 2.10
TripletNetMobileViTv294.31 ± 3.4394.81 ± 4.0194.18 ± 3.2894.5 ± 3.6997.31 ± 2.0690.55 ± 2.3491.39 ± 2.8091.13 ± 4.3391.26 ± 2.5596.67 ± 2.19
LeViT94.34 ± 5.2993.80 ± 5.3793.95 ± 2.3193.87 ± 5.3396.67 ± 2.7990.68 ± 4.0989.99 ± 2.8590.71 ± 2.5890.35 ± 3.3695.71 ± 5.60
Mobile-Former94.98 ± 3.1894.65 ± 5.8095.48 ± 2.3795.06 ± 4.1197.43 ± 2.5792.38 ± 3.1592.66 ± 2.8693.08 ± 4.1592.87 ± 3.0096.94 ± 2.04
ProtoNetMobileViTv295.31 ± 4.5495.15 ± 4.9596.06 ± 2.7795.6 ± 4.7397.81 ± 3.1891.5 ± 2.6792.34 ± 2.8094.88 ± 4.0793.59 ± 2.7396.2 ± 2.06
LeViT95.62 ± 2.0895.61 ± 3.7994.96 ± 3.0895.29 ± 2.6898.21 ± 5.5893.91 ± 3.0993.85 ± 5.9692.6 ± 4.4393.22 ± 4.0798.13 ± 4.05
Mobile-Former95.60 ± 4.3394.61 ± 3.1994.49 ± 2.4394.55 ± 3.6896.98 ± 2.7394.42 ± 2.4592.42 ± 2.7592.19 ± 5.5992.3 ± 2.5995.02 ± 2.47
MatchingNetMobileViTv295.91 ± 4.7696.25 ± 4.4296.19 ± 4.4996.22 ± 4.5897.88 ± 3.9694 ± 5.5295.2 ± 2.2794.87 ± 4.8695.03 ± 3.2296.98 ± 3.56
LeViT95.86 ± 4.6396.26 ± 4.9496.68 ± 4.4296.47 ± 4.7898.84 ± 3.5094.49 ± 5.1694.4 ± 5.7893.06 ± 5.9193.72 ± 5.4698.27 ± 2.62
Mobile-Former96.88 ± 2.8796.27 ± 3.4595.93 ± 2.1196.1 ± 3.1397.4 ± 3.9095.22 ± 4.1794.07 ± 2.9092.28 ± 3.5593.17 ± 3.4296.15 ± 2.55
MAMLMobileViTv296.15 ± 4.1696.48 ± 3.8695.5 ± 2.2795.99 ± 4.0098.16 ± 5.3393.14 ± 5.1992.61 ± 2.7093.71 ± 3.1593.16 ± 3.5595.51 ± 2.46
LeViT96.69 ± 5.6397.46 ± 5.0897.27 ± 4.1697.37 ± 5.3498.36 ± 2.1395.1 ± 3.8695.05 ± 3.3695.32 ± 3.2895.19 ± 3.5997.26 ± 3.63
Mobile-Former96.55 ± 2.1595.96 ± 5.5796.33 ± 4.7996.15 ± 3.1098.5 ± 2.3894.58 ± 4.5293.74 ± 2.8693.38 ± 2.6893.56 ± 3.5197.04 ± 3.70
RelationNetMobileViTv294.93 ± 5.1394.69 ± 5.1694.11 ± 5.0994.4 ± 5.1497.65 ± 5.6992.46 ± 3.6291.51 ± 2.2792.71 ± 5.7492.1 ± 2.7994.94 ± 3.75
LeViT94.00 ± 2.1993.18 ± 2.4692.59 ± 2.1992.89 ± 2.3198.54 ± 2.5091.06 ± 2.5491.02 ± 3.9991.12 ± 5.4291.07 ± 3.1095.69 ± 4.95
Mobile-Former95.32 ± 5.9595.36 ± 2.6995.39 ± 5.8195.37 ± 3.7096.82 ± 2.1491.91 ± 5.4591.96 ± 4.2193.1 ± 4.4892.52 ± 4.7596.46 ± 2.49
Proposed98.51 ± 2.0998.92 ± 2.6598.31 ± 3.3198.61 ± 3.4999.63 ± 5.6796.13 ± 2.1794.22 ± 2.3594.55 ± 2.3994.35 ± 2.2699.31 ± 2.47
Table 4. Comparison of numerical results of the proposed model against baselines on leaf dataset in the 5-way 1-shot setting.
Table 4. Comparison of numerical results of the proposed model against baselines on leaf dataset in the 5-way 1-shot setting.
Biotic Stress ClassificationSeverity Estimation
FSL MethodBackboneAccuracyPrecisionRecallF1-ScoreAUCAccuracyPrecisionRecallF1-ScoreAUC
SiameseNetMobileViTv292.32 ± 2.9891.18 ± 4.8491.86 ± 2.8591.52 ± 3.6994.95 ± 1.5887.8 ± 2.4689.57 ± 3.5988.81 ± 1.8889.19 ± 2.9294.22 ± 3.87
LeViT90.1 ± 1.7691.46 ± 4.2490.79 ± 1.1091.12 ± 2.4895.12 ± 3.1990.28 ± 3.9689.93 ± 4.0490.97 ± 1.8290.44 ± 4.0092.9 ± 1.58
Mobile-Former89.48 ± 4.6588.09 ± 4.8088.8 ± 3.4188.44 ± 4.7294.06 ± 1.9288.96 ± 2.6487.95 ± 2.5387.88 ± 4.1387.91 ± 2.5993.23 ± 2.59
TripletNetMobileViTv294.73 ± 1.0793.18 ± 4.4690.51 ± 4.9691.83 ± 1.7397.95 ± 3.4290.85 ± 1.1793.06 ± 1.5891.2 ± 2.3292.12 ± 1.3495.21 ± 2.95
LeViT93.56 ± 4.2892.91 ± 4.9793.48 ± 1.7393.2 ± 4.6095.51 ± 2.5892.25 ± 1.6390.19 ± 2.9789.34 ± 4.4989.76 ± 2.1194.53 ± 1.19
Mobile-Former91.15 ± 1.5492.56 ± 4.1991.77 ± 2.0492.16 ± 2.2595.45 ± 4.3388.5 ± 1.9389.14 ± 3.5188.32 ± 3.7988.73 ± 2.4993.45 ± 2.95
ProtoNetMobileViTv293.1 ± 2.5893.02 ± 1.7694.04 ± 1.3593.53 ± 2.0996.45 ± 4.0891.89 ± 3.9690.44 ± 2.5390.43 ± 3.6190.43 ± 3.0895.04 ± 3.02
LeViT92.01 ± 2.0391.44 ± 4.8792.59 ± 1.2692.01 ± 2.8797.04 ± 2.2293.15 ± 3.5090.51 ± 2.1592.73 ± 1.3591.61 ± 2.6694.4 ± 3.96
Mobile-Former92.47 ± 3.8394.63 ± 2.7593.08 ± 3.6493.85 ± 3.2097.94 ± 4.9591.26 ± 1.0792.59 ± 2.1593.71 ± 2.7293.15 ± 1.4296.28 ± 3.44
MatchingNetMobileViTv294.27 ± 3.4192.86 ± 3.8994.93 ± 3.2493.89 ± 3.6397.69 ± 1.7091.01 ± 1.0491.36 ± 3.1593.32 ± 4.0492.33 ± 1.5697.26 ± 2.94
LeViT94.68 ± 2.8493.11 ± 1.5894.62 ± 3.6793.86 ± 2.0397.08 ± 1.4291.19 ± 1.7692.13 ± 4.6091.96 ± 2.7392.05 ± 2.5595 ± 2.64
Mobile-Former94.25 ± 2.4293.11 ± 2.0395.48 ± 2.8394.28 ± 2.2097.93 ± 4.8791.08 ± 4.5492.58 ± 2.0091.74 ± 4.6192.16 ± 2.7896.76 ± 2.14
MAMLMobileViTv294.31 ± 2.5392.98 ± 3.5195.04 ± 4.6694 ± 2.9497.47 ± 4.8992.14 ± 1.3091.36 ± 4.3393.19 ± 2.1092.27 ± 2.0096.99 ± 2.07
LeViT94.55 ± 3.6493.29 ± 1.7396.49 ± 3.9194.87 ± 2.3598.81 ± 1.3391.8 ± 3.8892.73 ± 4.9892.4 ± 2.9592.57 ± 4.3697.14 ± 4.05
Mobile-Former93.3 ± 1.0192.86 ± 3.9192.78 ± 3.5892.82 ± 1.6195.88 ± 1.7191.21 ± 1.2191.27 ± 4.5194.19 ± 2.2992.7 ± 1.9096.31 ± 2.09
RelationNetMobileViTv294.11 ± 4.4792.05 ± 2.8493.88 ± 3.1192.96 ± 3.4897.75 ± 3.9092.36 ± 4.5190.74 ± 4.8591.22 ± 4.8890.98 ± 4.6797.28 ± 4.59
LeViT94.69 ± 2.5794.32 ± 1.8394.18 ± 4.4894.25 ± 2.1498.01 ± 1.3790.75 ± 3.0589.79 ± 2.6587.27 ± 4.0688.51 ± 2.8396.96 ± 3.88
Mobile-Former92.69 ± 3.4394.03 ± 3.1194.61 ± 2.1094.32 ± 3.2697.12 ± 4.2790.06 ± 1.6193.11 ± 4.7088.56 ± 2.1990.78 ± 2.4097.65 ± 1.78
Proposed 96.72 ± 2.6796.72 ± 5.7696.22 ± 3.9996.46 ± 2.6598.89 ± 2.6094.34 ± 2.0294.26 ± 2.0194.37 ± 2.4994.24 ± 2.0298.84 ± 4.74
Table 5. Comparison of numerical results of the proposed model against baselines on symptom dataset.
Table 5. Comparison of numerical results of the proposed model against baselines on symptom dataset.
Biotic Stress Classification (5-Way 5-Shot)Biotic Stress Classification (5-Way 1-Shot)
FSL MethodBackboneAccuracyPrecisionRecallF1-ScoreAUCAccuracyPrecisionRecallF1-ScoreAUC
SiameseNetMobileViTv292.34 ± 2.6492.88 ± 3.6292.48 ± 5.6292.68 ± 3.0595.11 ± 4.3192.24 ± 3.8793.24 ± 2.3391.81 ± 3.4192.52 ± 2.9195.46 ± 2.07
LeViT91.99 ± 2.7890.99 ± 2.6191.3 ± 2.5691.14 ± 2.6996.55 ± 2.2892.36 ± 3.6490.47 ± 3.0790.38 ± 4.2590.42 ± 3.3397.08 ± 4.85
Mobile-Former91.23 ± 5.5191.38 ± 3.7892.24 ± 5.0991.81 ± 4.4994.5 ± 2.8291.61 ± 5.4491.04 ± 4.3692.46 ± 3.6391.74 ± 4.8494.47 ± 3.70
TripletNetMobileViTv293.31 ± 2.3993.81 ± 2.3894.18 ± 2.2493.99 ± 2.3997.31 ± 5.5793.33 ± 5.6094.32 ± 4.7893.99 ± 5.8894.15 ± 5.1696.81 ± 5.11
LeViT93.84 ± 3.6293.8 ± 3.8392.95 ± 3.9593.37 ± 3.7296.17 ± 3.5993.83 ± 2.4994.52 ± 2.5393.31 ± 3.2993.92 ± 2.5195.83 ± 5.85
Mobile-Former94.98 ± 5.2293.65 ± 3.7495.48 ± 4.4494.56 ± 4.3696.93 ± 3.2394.91 ± 4.3494.45 ± 4.3094.5 ± 5.2894.48 ± 4.3297.07 ± 3.05
ProtoNetMobileViTv295.31 ± 4.1094.15 ± 3.9495.56 ± 3.8294.85 ± 4.0296.81 ± 4.3495.69 ± 3.3493.5 ± 2.5194.83 ± 4.6694.16 ± 2.8796.25 ± 2.65
LeViT95.31 ± 5.0695.61 ± 3.9194.46 ± 2.0895.03 ± 4.4198.21 ± 4.8095.09 ± 4.4795.84 ± 5.1895.4 ± 2.3095.62 ± 4.8097.27 ± 3.40
Mobile-Former94.6 ± 5.4494.61 ± 3.2894.49 ± 3.4594.55 ± 4.0996.48 ± 2.0394.75 ± 2.9894.54 ± 4.9393.69 ± 2.3094.11 ± 3.7196.91 ± 4.15
MatchingNetMobileViTv295.91 ± 4.7095.25 ± 5.4695.19 ± 2.1995.22 ± 5.0597.88 ± 3.7195.57 ± 3.3695.31 ± 4.7494.41 ± 4.3194.86 ± 3.9398.12 ± 4.55
LeViT95.86 ± 3.5595.76 ± 2.3995.68 ± 2.3995.72 ± 2.8697.84 ± 2.0495.99 ± 4.3895.83 ± 4.1494.92 ± 5.5495.37 ± 4.2697.24 ± 5.43
Mobile-Former95.88 ± 3.5095.27 ± 5.3294.93 ± 5.3995.1 ± 4.2296.4 ± 4.7296.81 ± 2.8795.18 ± 2.4195.05 ± 3.2495.12 ± 2.6296.99 ± 4.74
MAMLMobileViTv296.15 ± 5.1995.98 ± 3.9895.5 ± 4.3595.74 ± 4.5197.16 ± 5.7796.83 ± 4.5895.89 ± 2.8395.03 ± 5.6495.46 ± 3.5096.87 ± 4.22
LeViT95.89 ± 4.0796.46 ± 2.7796.77 ± 3.7396.61 ± 3.2997.36 ± 2.6295.02 ± 3.5895.82 ± 3.1697.57 ± 4.0796.69 ± 3.3697.83 ± 4.01
Mobile-Former96.05 ± 3.8194.96 ± 2.1996.33 ± 4.4395.64 ± 2.7898.5 ± 3.3995.96 ± 5.3094.66 ± 2.8896.56 ± 4.6195.6 ± 3.7399.3 ± 2.18
RelationNetMobileViTv294.93 ± 5.8693.69 ± 2.0493.11 ± 3.4193.4 ± 3.0397.65 ± 4.5195.92 ± 4.2994.37 ± 3.3492.24 ± 4.5693.29 ± 3.7697.02 ± 3.84
LeViT93.5 ± 4.7592.18 ± 5.1391.59 ± 3.7591.88 ± 4.9397.54 ± 3.4693.58 ± 2.7991.75 ± 5.2090.75 ± 4.4691.25 ± 3.6397.97 ± 3.41
Mobile-Former94.32 ± 2.5095.36 ± 4.8494.39 ± 5.9294.87 ± 3.2996.32 ± 3.8993.33 ± 2.4895.7 ± 3.5393.4 ± 3.2494.54 ± 2.9297.19 ± 5.40
Proposed 98.34 ± 1.2998.4 ± 3.1498.14 ± 2.2398.27 ± 1.8399.68 ± 2.3597.79 ± 1.1897.22 ± 1.4797.38 ± 1.6797.30 ± 1.3199.12 ± 3.93
Table 6. Statistical Analysis Results for 5-Way 5-Shot and 5-Way 1-Shot Settings.
Table 6. Statistical Analysis Results for 5-Way 5-Shot and 5-Way 1-Shot Settings.
5-Way 5-Shot5-Way 1-Shot
Biotic Stress ClassificationSeverity EstimationBiotic Stress ClassificationSeverity Estimation
Proposed vs. SiameseNet3.983 × 10−118.405 × 10−118.014 × 10−152.377 × 10−70
Proposed vs. TripletNet3.399 × 10−404.268 × 10−147.375 × 10−801.662 × 10−90
Proposed vs. ProtoNet8.185 × 10−201.363 × 10−218.313 × 10−906.254 × 10−40
Proposed vs. MatchingNet4.482 × 10−116.694 × 10−318.539 × 10−122.501 × 10−50
Proposed vs. MAML6.641 × 10−508.805 × 10−124.539 × 10−133.357 × 10−80
Proposed vs. RelationNet4.369 × 10−36.220 × 10−85.713 × 10−72.819 × 10−11
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Salamai, A.A.; Al-Nami, W.T. Sustainable Coffee Leaf Diagnosis: A Deep Knowledgeable Meta-Learning Approach. Sustainability 2023, 15, 16791. https://doi.org/10.3390/su152416791

AMA Style

Salamai AA, Al-Nami WT. Sustainable Coffee Leaf Diagnosis: A Deep Knowledgeable Meta-Learning Approach. Sustainability. 2023; 15(24):16791. https://doi.org/10.3390/su152416791

Chicago/Turabian Style

Salamai, Abdullah Ali, and Waleed Tawfiq Al-Nami. 2023. "Sustainable Coffee Leaf Diagnosis: A Deep Knowledgeable Meta-Learning Approach" Sustainability 15, no. 24: 16791. https://doi.org/10.3390/su152416791

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop