Next Article in Journal
Slope with Predetermined Shear Plane Stability Predictions under Cyclic Loading with Innovative Time Series Analysis by Mechanical Learning Approach
Next Article in Special Issue
A Neural Network-Based Model for Predicting Saybolt Color of Petroleum Products
Previous Article in Journal
Development of a Smartphone-Integrated Reflective Scatterometer for Bacterial Identification
Previous Article in Special Issue
Few-Shot Object Detection Using Multimodal Sensor Systems of Unmanned Surface Vehicles
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Word Embedding Distribution Propagation Graph Network for Few-Shot Learning

College of Computer Science and Technology, Changchun University of Science and Technology, Changchun 130022, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(7), 2648; https://doi.org/10.3390/s22072648
Submission received: 14 January 2022 / Revised: 4 March 2022 / Accepted: 15 March 2022 / Published: 30 March 2022
(This article belongs to the Special Issue Intelligent Sensors and Machine Learning)

Abstract

:
Few-shot learning (FSL) is of great significance to the field of machine learning. The ability to learn and generalize using a small number of samples is an obvious distinction between artificial intelligence and humans. In the FSL domain, most graph neural networks (GNNs) focus on transferring labeled sample information to an unlabeled query sample, ignoring the important role of semantic information during the classification process. Our proposed method embeds semantic information of classes into a GNN, creating a word embedding distribution propagation graph network (WPGN) for FSL. We merge the attention mechanism with our backbone network, use the Mahalanobis distance to calculate the similarity of classes, select the Funnel ReLU (FReLU) function as the activation function of the Transform layer, and update the point graph and word embedding distribution graph. In extensive experiments on FSL benchmarks, compared with the baseline model, the accuracy of the WPGN on the 5-way-1/2/5 shot tasks increased by 9.03, 4.56, and 4.15%, respectively.

1. Introduction

Most successful deep-learning architectures are based on rich labeled datasets. However, in special practical application scenarios, only small numbers of labeled data may be available, owing to certain limitations. Therefore, there is a need to acquire information on new classes based on limited labeled data. This is known as few-shot learning, in which tasks use a small number of labeled samples to predict unlabeled ones. There are a variety of approaches that have been proposed for FSL to address the deficiency of labeled data.
Meta-learning is one of the main methods used in FSL. Model-Agnostic Meta-Learning (MAML) [1] uses an initialization parameter that requires only a few samples to perform gradient descent and to achieve good results when encountering new problems. Because the MAML method only adjusts parameters according to different tasks, the trained model is prone to overfitting. Task Agnostic Meta-learning for few-shot Learning (TAML) [2] is an improvement on the MAML algorithm. The original design explicitly requires that the parameters of the model have no preference for different tasks during regularization. Meta-learning with memory Augmented Neural Networks (MANN) [3] uses a recurrent neural network (RNN) to memorize the representation of the previous task. Although this method is helpful for learning new tasks, the weight of the RNN updating process is still slow, which makes the training process difficult. Meta-learning with differentiable closed-form solvers [4] uses simpler differentiable regression methods that have closed-form solutions to replace the original learning algorithms (e.g., the k-nearest-neighbor (KNN) algorithm and the convolutional neural network) and has inspired our idea of combining a traditional learning algorithm with a neural network. Belonging Network (BeNet) [5] uses basic statistical information of the target class to find the simple mean and variance information to improve performance of the training image set. Regularization using knowledge distillation when learning small datasets [6] leverages the knowledge distillation method, which found that increasing the distillation parameter “temperature” can improve model accuracy, especially for small numbers of training data. However, if the testing and training sets show a large difference with regard to their distribution, the output of the model will be highly unsatisfactory. Task-Aware Feature Embeddings for low-shot learning (TAFE-Net) [7] innovatively uses the meta-learning method to dynamically select weight parameters. According to various tasks, different weight parameters are selected, and the weight decomposition method is used to render this computationally possible. Because the few-shot dataset has no corresponding class description information, the meta-learner’s ability to represent the embedded features of the task is affected. Hence, the experimental effect is only slightly improved compared with other algorithms. Therefore, how to effectively use the limited dataset information has become the focus of our research.
Metric learning maps images into an embedding space where images of the same class are located close to each other, and the images of different classes are farther apart. The Siamese Neural Network [8] limits the structure of the input images and can automatically discover the generalizable features of the new samples. However, because it is sensitive to the differences between the two images, it may easily result in misclassification. Matching networks [9] construct an end-to-end nearest-neighbor classifier. Through meta-learning training, the classifier can quickly adapt to new tasks with few samples. However, when the label distribution has obvious deviations (e.g., being fine-grained), the model becomes unusable. Few-shot image classification with Differentiable Earth Mover’s Distance and structured classifiers (DeepEMD) [10] splits an image into multiple tiles and introduces a new distance measurement method (i.e., Earth Mover’s Distance (EMD)) and calculates the best matching cost between each tile of the query set and the support-set image, which indicates the degree of similarity between the two. Boosting few-shot learning with adaptive margin loss [11] improves the classification effect of the original algorithm by introducing adaptive margin loss of class-relevant or task-relevant information and uses the semantic similarity between different categories to generate adaptive edges. However, there is no correlation between the two proposed edge generation methods. Improved few-shot visual classification [12] uses the Mahalanobis distance to calculate the distance between samples; however, it focuses largely on dividing the most accurate inter-class interval for the existing samples and neglects learning the image features. The metric learning above makes us realize the importance of a distance measurement in FSL.
In this study, we propose a new model for FSL called the word embedding distribution propagation graph network. As illustrated in Figure 1, the word embedding distribution graph is merged into the GNN, and the fine-grained few-shot classification task is solved by the cyclic calculation method. The contributions of this work are summarized below.
First, the WPGN uses the GloVe model to extract the class label information as a word vector. The WordNet model is used to weigh the class distribution similarity. It embeds class semantic information into the GNN. Using the word embedding distribution graph, the WPGN solves the problem of low classification accuracy caused by the similarity of fine-grained image features.
Second, we replace the ReLU activation function of the GNN with the FReLU [13] function. Compared with ReLU, FReLU is more suitable for processing vision tasks and can further improve classification accuracy. At the same time, according to our experiments, the Mahalanobis distance exhibits a better classification performance than Euclidean distance for FSL. Furthermore, the covariance matrix is used to eliminate the variance and dimensionality between each component. Therefore, we used Mahalanobis distance instead of Euclidean distance to calculate the distance between samples.
Finally, we combine the Efficient Channel Attention (ECA) [14] and our backbone network (ResNet-12), called ECAResNet-12, which can introduce few extra parameters, almost ignore calculations, and bring performance gain. ECAResNet-12 can better extract image feature information through one-dimensional convolution to efficiently achieve local cross-channel interaction, which extracts the dependencies between channels and avoids dimensionality reduction, and further improves the classification performance of GNN.
The remainder of this paper is organized as follows. Section 2 describes related work. Section 3 focuses on the few-shot task and introduces the framework of WPGN in detail. Section 4 presents a discussion of the results of the WPGN comparison experiments, ablation studies and practical application example, demonstrating the effectiveness of WPGN on FSL.

2. Related Work

Recently, there have been many high-quality works in Few-Shot Learning with novel ideas, which are different from meta-learning and metric learning. We will introduce them from four parts: graph neural network, transfer learning, semantic information, attention mechanism and application.

2.1. Graph Neural Network

The GNN, which is a new method for performing FSL, is a multi-layer network with a weight-sharing property. The GNN is a graph model composed of nodes and edges. Each node represents an input image, and the weight on each edge represents the relationship between different images. This relationship can consist of the distance or the similarity between the images. FSL with graph neural networks [15] transfers the distance metric from a Euclidean space to a non-Euclidean one, and its core idea is the same as that of most GNNs. It transfers the label information of labeled images to unlabeled query images. On the basis of the abovementioned GNN method, the Edge-labeling Graph Neural Network (EGNN) [16] uses edge label prediction instead of node label prediction as well as a two-dimensional edge features to explicitly express the similarities and differences between classes. Although it shows the great potential of GNN in FSL, the distribution information of the sample is ignored. Most GNN methods based on meta-learning focus on the instance-level relationship between samples, and the Distribution Propagation Graph Network (DPGN) [17] further extends this idea by explicitly simulating the distribution relationship from one instance to all other instances in a 1-vs.-N manner. DPGN proposes a dual-graph neural network model, which establishes the cyclic propagation process between the two graphs. The instance- and distribution-level relationships are then combined to create a better environment for few-shot classification. Unfortunately, DPGN completely ignores semantic information, which is important for fine-grained classification. Therefore, DPGN shows poor classification performance. The looking-back [18] method proposed absorbs the lower-level information in the hidden layers of the neural network using a GNN. Denoising AutoEncoders (DAE) [19] proposed a structure comprising denoising autoencoders using a graph neural network, exploited the connection between nodes to perform reconstruction, and updated the classification weight vector. However, the DAE completely ignores the semantic information of the image.

2.2. Transfer Learning

The transfer-learning scheme for semi-supervised few-shot learning (TransMatch) [20] uses transfer learning and weight imprinting to generate classifier weights, and it uses the MixUp method to perform semi-supervised training, demonstrating that transfer learning can achieve better results in small sample scenarios. Although the accuracy of transfer learning is lower than that of an FSL model because of its flexibility, we use this idea to extract semantic information.

2.3. Semantic Information

The classification method based on semantic information uses target label information as prior knowledge to assist target classification. Multiple-Semantics [21] uses a variety of semantic information to enrich the information source of small sample learning, which is closer to the situation of humans learning new things and reminds us that human beings will refer to label information when classifying. Variational AutoEncoders [22] proposed a cross and distribution aligned-variational autoencoder that combines image feature information and semantic information to construct latent features containing important multi-model information to infer the classification of unseen samples. However, methods for obtaining accurate semantic information are an important factor that limits the application of these methods. Learning Compositional Representations [23] decomposed the image representation into multiple attribute representations and improved the representation ability of the feature extraction network by adding loss-function constraints. The characteristics of manual labeling limit the applicability of the model. A classification hierarchical structure [24] uses the semantic relationship between classes to perform additional supervision on the feature extraction network and guide it to learn additional transferable feature information. This helps the KNN algorithm obtain more accurate classification results. Semantic feature augmentation in few-shot learning [25] also uses semantic information to expand the data. It encodes the feature data to map the semantic space, and it then performs classification by decoding the enhanced information. The result is better than that of data augmentation at the image level. A new hierarchical semantic embedding [26] framework effectively uses the hierarchical classification structure to guide network feature learning, encodes the correlation between different hierarchical classes, and achieves better performance on fine-grained image classification. However, this framework requires the manual annotation of datasets, which affects its practical applicability because manual annotation is tedious and time consuming. To sum up, our research focuses on acquiring accurate semantic information without manual annotation.

2.4. Attention and Application

Few-shot classification via adaptive attention [27] introduces a channel attention mechanism and a spatial attention mechanism to optimize feature maps. Although simpler and more effective than other models, it lacks the ability to adapt to new tasks. As there are fewer samples for learning in FSL, it means that richer features should be extracted from each image, and an attention mechanism can meet this demand. Voronoi Decomposition-based Random Region Erasing [28] generates new artificial images with randomly covered regions of irregular shape and augments the original dataset to train the neural network more efficiently. Multimodal few-shot learning for gait recognition [29] combines CNNs and RNNs using multi-modal time-series learning to map the latent embedding vector space and to address the open-set gait recognition problem. The wide application of FSL makes us realize the great research potential in this field.
Based on this research and our own work, we now introduce the WPGN and follow the dual-graph cyclic calculation method of the DPGN to construct a point graph and word embedding distribution graph that extract semantic information using the GloVe [30] model, and we eliminate the need for manual labeling, thereby achieving better classification results than extant methods.

3. Method

In this section, we first provide the background of our FSL task, and we then discuss our proposed WPGN in detail.

3.1. Problem Definition

The goal of an FSL task is to train a model to perform well in cases when only few samples are given. Our few-shot task provides a support set S, a query set Q, and training image dataset Dtrain. Each task contains N classes with K samples per class, which implies an N-way-K shot setting. We compare the few-shot task to the traditional image classification task in Figure 2, where Figure 2a represents the traditional image classification. We train the model on the training set in 10 classes, and then use the trained model to test the accuracy on the testing set. Conversely, Figure 2b represents a few-shot classification. There are five classes in the training, and five images of S in each class, indicating the 5-way-5 shot task category. After obtaining the trained model, the accuracy of the model is tested on the S and Q of the testing set. The testing task process is the same as the training task: a 5-way-5 shot task.

3.2. Feature Extraction with ECA-Net

The image contains a foreground and background, and the quality of feature extraction will directly affect the classification effect of the GNN. In ResNet-12, different regional features of the image are treated equally. However, in the classification task, we hope that the network can pay more attention to the foreground and ignore the background. Therefore, as shown in Figure 3, we have added a channel attention mechanism, ECA-Net, to the ResNet-12 network, to enable our backbone to ignore the background, highlight the foreground, and further improve the quality of feature extraction. In the backbone, with the increase in the number of channels, the resolution of the feature map will decrease. In the course of the channel attention learning, if either the resolution or the number of channels is too low, it will result in a decline in the image extraction quality. Because the channel attention mechanism has certain requirements on the resolution and channel number of feature map, we choose the feature map with medium channel number and resolution, that is, an ECA mechanism is added to the feature map with 128 channels.

3.3. Word Embedding Distribution Propagation Graph Network

As shown in Figure 1, the WPGN consists of L layers, and each layer contains a point graph and a word embedding distribution graph, which are designed based on a GNN. The ECAResnet-12 backbone network is used for feature extraction. First, an image feature is extracted in the feature extraction layer as the initialization information of the point graph. Second, according to the corresponding class label of the images, the word vectors of each class are embedded using the GloVe model to provide the initial information of the word embedding distribution graph. Third, we merge the point graph and word embedded distribution graph, update the position of the node in the point graph, and cyclically generate the point graph and word-embedded distribution graph for each layer. Finally, the distance between nodes in the point graph judges the similarity between the query set and the support set, and the class of the query set is predicted.

3.3.1. Point Graph

The point graph was designed based on the GNN and generated according to the extracted image feature information. The point graph represents the position of each instance in the sample space. The nodes in the point graph are initialized as follows:
V 0 , i p = f e x t r a c t ( g i )
where g i represents each image sample instance, f e x t r a c t ( ) represents the backbone network for image extraction. Edge E l , i j p in the point graph represents the image feature similarity, which can be calculated using the following formula:
E l , i j p = M ( V l , i p , V l , j p ) × E l 1 , i j p
where V l , i p and V l , j p respectively represent nodes i and j in the point graph, and when l equals zero, E l 1 , i j p equals one. M represents the Mahalanobis distance. The calculation formula is as follows:
M ( i , j ) = 1 2 ( i j ) T ( Q τ ) - 1 ( i j )
Because FSL uses tasks as a unit of learning, we use τ for specific task and c for a class in task τ . Q τ represents the estimated value of the covariance matrix between images in task τ . Q c τ is the estimated value of the covariance matrix between images of task τ and class c .
Q τ = c = 1 N Q c τ N = c = 1 N 1 N ( λ c τ Q c τ + ( 1 λ c τ ) Q τ )
where N represents the number of classes in task τ . Q c τ is the true value of the covariance matrix between images of class c in task τ .
Q c τ = 1 | K | g i K ( f e x t r a c t ( g i ) μ c ) ( f e x t r a c t ( g i ) μ c ) T
where μ c represents the mean value of the feature embedding matrix, f e x t r a c t ( ) , in the support set image in task τ . K represents the number of images in the support set of class c in task τ . Q τ is the true value of the covariance matrix between all classes of images in task τ .
Q τ = c = 1 N Q c τ N
The weight λ c τ is calculated by the following method:
λ c τ = K K + 1
In existing GNNs, the focus is on the use of methods for embedding information. The choice of metric is also important. Previous selection of the metric involves two unrealistic assumptions: feature dimensions are not correlated, and there is consistent covariance. Mahalanobis believes that different types of images can have different covariances, and the distribution of those images is closer to the real situation and must remain in focus. The Mahalanobis distance can deal with the problem of non-independent and identical distributions among various dimensions in high-dimensional linearly distributed data. Because the amount of data in the FSL task is small, it is important to consider the difference in image covariance of different categories. The Mahalanobis distance is the normalized distance of non-uniform distribution in Euclidean space, without considering the influence of data dimension. Furthermore, the correlation between variables is considered according to the distribution of features in the whole space. Therefore, it can better describe the similarity between data.
The Mahalanobis distance is an excellent way to solve the differences between classes; hence, the WPGN uses it to calculate the distance between them. To verify the effectiveness of Mahalanobis distance, we compare the classification results of three datasets by changing the measurement method of the similarity calculation. The experimental results are shown in Table 1.
As seen in Table 1, the Manhattan distance appears to have the lowest accuracy. The Mahalanobis distance has certain advantages compared with the Manhattan and Euclidean distances; therefore, we choose it as the best method to calculate the similarity of class in the WPGN.

3.3.2. Word Embedding Distribution Graph

Similar to the point graph, the word embedding distribution graph is designed based on the GNN, which is generated by semantic information. The GloVe word embedding model is adopted to vectorize object labels in the training. WordNet [31] is used to calculate the similarity between nodes.
The GloVe model performs a vector representation of words, which makes the vectors contain as much semantic and grammatical information as possible. The GloVe model first constructs a word co-occurrence matrix based on a large corpus, and then trains word vectors with fixed dimensions, such that the distance between word vectors of different words is close to the distance between different words in the co-occurrence matrix. While reducing the dimension of data, the loss of accuracy is reduced as much as possible, and the trained word vectors are finally obtained. The word vectors of the GloVe model can be added and subtracted. For example,
f g ( K i n g ) f g ( M a n ) + f g ( W o m a n ) = f g ( Q u e e n )
where f g ( ) represents the word vectors trained by the GloVe model. The fact that the word vectors can be added and subtracted allows us to create and generate it in a full sample space. This method is conducive to expanding the distance between classes in the full sample space, thus improving the classification performance. We extract the label of each class in the dataset as word vectors through the pre-trained GloVe model to use it as the initial node of the word embedding distribution graph.
The node V l , i w in the word embedding distribution graph represents the instance of each image g i embedded by semantic information, and the V 0 , i w initial value is as follows:
V 0 , i w = f g ( l a b e l c )
where l a b e l c represents the label of class c .
Each edge in the word distribution graph E l , i j w stands for the similarity between the semantic distribution features of different samples, and the calculation formula is as follows:
E l , i j w = M ( V l , i w , V l , j w ) × E l 1 , i j w × f w ( V l , i w , V l , j w )
where l equals zero, and E l - 1 , i j w equals one. To prevent E 0 , i j w from disappearing after many iterations of E l , i j w , the number of iteration layers is discussed in the experiment. The number of layers is set to five to avoid the disappearance of E 0 , i j w .
Here, f w represents the similarity calculated by the WordNet model, and the calculation method is as follows:
f w ( x , y ) = 1 a b s ( m min x min ) + a b s ( m min y min )
where x min , y min , and m min represent the minimum depths of the word set. These three parameters are calculated as follows:
x m i n = m i n d e p t h ( x ) y m i n = m i n d e p t h ( y ) m m i n = m i n d e p t h ( m )
Parameter m represents the lowest common upper word set, and the calculation formula is as follows:
m = L o w h y ( x , y )

3.3.3. Loop Computation

The positions between instances in the word embedding distribution graph represent the distribution of different instances in the sample space. After the WPGN is initialized, the model will perform cyclic calculations combined with word embedding to learn image features and predict the classification of the images. The cycle calculation process is shown in Figure 4.
At first, the image through feature extraction network ECAResNet-12 is used as the initial node V 0 , i p of point graph 0 layer. The GloVe model is used to extract semantic information as the initial node V 0 , i w of word embedding distribution graph 0 layer. Then, Mahalanobis distance is used as a measure of the distance between nodes and the edges of the point graph E 0 , i j p , and the word embedding distribution graph E 0 , i j w is calculated. Second, E 0 , i j w and V 0 , i p will update the point graph through W2P and obtain point graph node V 1 , i p . After the calculation of E 1 , i j p , E 1 , i j p and V 0 , i w , it will update the word embedding distribution graph through P2W and obtain node V 1 , i w . Finally, the E 1 , i j w is calculated according to its nodes, where the dual-graph of layer 1 is calculated, and the calculation is repeated as described above until layer L is reached.

Updating the Point Graph

The process of updating the point graph is shown in Figure 5a. The point graph is adjusted in the Transform layer, which consists of a Conv layer, a BatchNorm layer, and a FReLU activation function, which transmits information in reverse. The adjustment strategy W2P is shown as follows:
V l , i p = f F R f B N f c o n v 2 d j = 1 N ( E l - 1 , i j w V l - 1 , j p ) , V l - 1 , i p
where f c o n v 2 d ( ) represents the convolution operation, and f B N ( ) represents the Batch Normalization operation. Then, node V l , i p of the next layer is obtained.

Updating the Word Embedding Distribution Graph

The process of updating the word embedding distribution is shown in Figure 5b. The word embedding distribution graph is adjusted by the Transform layer, which includes the full connection layer and the FReLU activation function, which provides fusion transfer adjustments. We propagate the query set images without label information from the point graph to the word embedding distribution graph. The node adjustment method P2W is as follows:
V l , i w = f F R ( f F C ( |   | j = 1 N ( E l , i j p ) , V l - 1 , i w ) )
where |   | is the concatenation operator, which aggregates scalar E l , i j p into vectors.

FReLU Activation Function

FReLU is a simple and effective activation function suitable for processing visual tasks. It improves the ReLU by adding negligible spatial conditions overhead. FReLU is more suitable for the GNN than ReLU. We use FReLU to combine features in the Transform layer of W2P and P2W in the same way that WPGN obtains the point graph and word embedding distribution graph of the next layer. The FReLU is calculated as follows:
f F R ( x ) = M a x ( x , T ( x ) )
where T ( x ) represents a simple and efficient spatial context feature extractor. T ( x ) is defined as follows:
T ( x ) = f B N ( f c o n v 2 d ( x ) )
To verify the effectiveness of FReLU in the WPGN, we compare the classification results using different activation functions on the CUB-200-2011 dataset. The experimental results are shown in Table 2.
Here, the activation function has a non-negligible impact on the WPGN model. Compared with the LeakyReLU, FReLU used in this study has a certain degree of accuracy improvement. From ReLU to the latest FReLU, the accuracy of the WPGN has increased by 1.83%.

3.3.4. Loss Function

We use the Softmax function as the classification function, combining point graph loss and word embedding distribution graph loss as the loss value of the WPGN. The prediction process of each node is as follows:
P y i | x i = S o f t m a x j = 1 K E l , i j p one-hot ( y j )
where P ( y i | x i ) is the most probable class, given sample x i that belongs to the point graph, and y i is the label of the j th sample in the support set.
The calculation steps of the loss function are as follows.
1.
Calculate the point graph loss:
L l p = L C E ( P ( y i | x i ) , y i )
where L l p represents the loss of the L-layer point graph, and L C E is the cross-entropy loss function.
2.
Calculate the word embedding distribution graph loss:
P w y i | x i = Softmax j = 1 K E l , i j w one-hot ( y j )
where P w ( y i | x i ) is the most probable class, given sample x i that belongs to the word embedding distribution graph.
L l w = L C E ( P w ( y i | x i ) , y i )
3.
Calculate the model loss.
To balance the two losses, we introduce weight λ and calculate the total loss as follows:
L l o s s = λ L l p + ( 1 λ ) L l w
The classification accuracy is shown in the Table 3 as λ takes different values.
It can be seen from Table 3, with the increase in λ value, the classification accuracy gradually improves. When λ is 0.9, the highest accuracy rate can be obtained, and when it is greater than 0.9, the classification accuracy begins to decrease. The WPGN obtains the minimum loss value when λ is 0.9; thus, we set λ to 0.9.

4. Experiment

4.1. Experimental Environment and Datasets

The experimental environment of this paper is shown in Table 4.
We selected three types of standard datasets in FSL: MiniImageNet [9], CUB-200-2011 [32] and CIFAR-FS [4]. The details of the images, classes, training/validation/test set divisions and the image resolutions of each dataset are shown in Table 5.
As illustrated in Figure 6, the image features of the four different classes of birds in the CUB-200-2011 dataset are similar and more difficult to distinguish.

4.2. Experimental Settings

The WPGN uses cyclic computation to construct the network structure, including the point graph and word embedding distribution graph. Mutual updating between the dual-graph is the biggest feature of the WPGN. Therefore, the total number of layers of the WPGN affects the final classification results. To find the layer number that best fits the network structure, we trained the WPGN on the CUB-200-2011 dataset by changing the layer number to obtain the classification accuracy of each training model. The experimental results are shown in Figure 7.
Here, we can see that the abscissa represents the number of layers, zero represents no cyclic calculation, and one represents one cycle calculation. When the layer number increases from zero to five, the classification accuracy increases by nearly 17%. However, the growth of classification accuracy tends to be flat and slightly oscillates when the layer number is greater than five. Therefore, we chose five as the final layer number of the WPGN.
To more intuitively show the impact of different layer numbers on the WPGN’s classification accuracy, the labeled class [1,2,3,4,5] was selected for the experiment, and a heat map was used to show the change of classification accuracy with the increase in layer numbers.
The brighter parts indicate high confidence. Figure 8a did not use a cycle for calculation; hence, the classification accuracy was low, resulting in fuzzy predictions and a greater possibility of predicting the wrong label. Figure 8e has five layers, and with the exception of the ground-truth location, the other parts have darker colors, meaning that the probability of an accurate prediction is much higher than that of the prediction error.
The parameter settings obtained in the WPGN are shown in Table 6 and listed by experiment.

4.3. Evaluation

In this paper, classification accuracy was used to evaluate the performance of the model. The higher the accuracy, the better the performance of the model. We randomly selected n = 10,000 tasks, and we published the mean accuracy and the 95% confidence interval. The calculation formula of accuracy is as follows:
i = 1 n A c c i n

4.4. Experimental Results

For this study, we used ConvNet, RestNet-12 and ECAResNet-12 as the backbone network of features traction with three tasks: 5-way-1 shot/2 shot/5 shot. The experimental results are shown in Table 7 on the CUB-200-2011 dataset.
As can be seen from Table 7, the classification accuracy of the WPGN under three backbone networks and three tasks is higher than the other methods. When the feature extraction backbone network is ECAResNet-12 and the tasks are 5-way-1 shot, 5-way-2 shot, and 5-way-5 shot, the accuracy of the WPGN is improved by nearly 9.0%, 4.5%, and 4.1%, respectively, compared with the DPGN. The accuracy of the WPGN under the 5-way-2 shot task is approximately 2% higher than that of the DPGN under 5-way-5 shot. The experimental results prove and demonstrate that our WPGN is robust in fine-grained classification.
The experimental results are shown in Figure 9 on the MiniImagenet and CIFAR-FS datasets.
Here, the DPGN Conv represents the feature extraction backbone network as ConvNet on the DPGN, the WPGN ResNet represents the feature extraction backbone network as ResNet-12 on the WPGN, whereas the WPGN ECARes represents the feature extraction backbone network as ECAResNet-12 on the WPGN. From Figure 9, we can see that on the MiniImagenet dataset and the CIFAR-FS dataset, the classification accuracy of the WPGN is higher than that of the DPGN on the three tasks. Moreover, when the feature extraction backbone network adopted ECAResNet-12, its classification effect was better than that of ConvNet and ResNet-12. The experiment demonstrated that the WPGN performs better in a dataset having fewer obfuscating features. The accuracy of the CIFAR-FS dataset was lower than that of the MiniImagenet dataset because its background had a much smaller impact on the classification accuracy.
Moreover, compared with the DPGN, our model had less computational overhead while improving accuracy. This is because FSL is trained in task units. For each task, the first layer of the DPGN distribution graph requires a large number of calculations to initialize, whereas the first layer of the word distribution graph associated with the WPGN only needs to obtain the word vectors of the corresponding category. Thus, initialization is completed much faster. The time used for the same number of steps of WPGN and DPGN training is shown in Table 8. For the same number of rounds of training, WPGN requires far fewer calculations than DPGN.
In addition, compared with the number of training rounds, as shown in the left side of Figure 10, the loss convergence speed of WPGN is significantly faster than that of DPGN, which shows that the WPGN is better in total training time. We found that WPGN converged in 12,000 rounds. Hence, we reduced the learning rate for further optimization. DPGN requires at least 15,000 rounds to converge such that the learning rate can be reduced. We tried to reduce the learning rate for DPGN at 12,000 rounds, but experimental results show that the accuracy of DPGN was reduced by ~2%. The right side of Figure 10 shows that, compared with DPGN, WPGN converges faster and significantly improves test accuracy. At the same time, after several rounds of training, the performance of WPGN will not obviously decline, which proves that the model is more robust and less prone to overfitting.
Because the WPGN model is better than the DPGN in terms of calculation overhead and accuracy, it demonstrates that our research content has a good prospect for real-world applications.

4.5. Ablation Studies

In order to verify the validity of each component in WPGN, we will add word embedding, Mahalanobis distance, FReLU and ECA to the baseline model one by one. In the baseline model, the distance measurement method is Euclidean distance, and the activation function is the Leakey ReLU function. The ablation experiment can fully prove the effectiveness of the components. The results of the ablation experiment on the CUB-200-2011 and CIFAR-FS datasets under 5-way-1 shot tasks are shown in Table 9.
As observed from the results in the table, after the word embedding distribution graph was added to the WPGN, the classification accuracy of the two datasets increased by 7.23% and 2.1%. Using Mahalanobis distance in the similarity calculation method, the classification accuracy increased by ~0.4%. The activation FReLU function also improved the classification accuracy of the model, and although it is not as great as the first two innovations, it contributes to the improvement of model accuracy. Finally, by integrating the ECA attention module into ResNet-12, the accuracy of our model has increased by 1.2%. From the experimental results, it can be seen that for these two datasets, the four innovations described in this paper improved the classification accuracy of the model.

4.6. Practical Application Example

In order to prove the great potential of WPGN in practical application, we added an example to apply the trained WPGN to the classification of specific rare birds. In this case, seven species of rare birds in bird habitats were selected, as shown in Figure 11, of which two belonged to storks as shown in the upper part and five belong to cranes at the bottom. We can see that the similarity between these birds is high, although these birds belong to different categories. It is difficult for ordinary people to distinguish these seven species of birds if you are not a professional ornithologist. Generally speaking, compared with common image classification problems, fine-grained classification faces the images with more similar appearance characteristics. In addition, there are interference factors such as posture, illumination, viewing angle, occlusion and background in the collection, which lead to the characteristics of small differences between classes and large differences with classes. By using category labels, WPGN can first increase the distance between storks and cranes, and the distance between storks and cranes from the word vectors will be greater than the distance between subcategories. Second, in the subcategories of cranes or storks, word vectors can also divide well according to the category labels. Finally, image information is embedded into GNN to classify birds with the aid of semantic information.
The example contains 350 images of seven species of birds, and we use WPGN trained on the CUB-200-2011 dataset to test this example with 7-way-1 shot task. In this example, the accuracy of WPGN on the 7-way-1 shot task is 82.45%, while our baseline model is 72.14%. Importantly, semantic information can be obtained without manual tagging. This example illustrates the huge potential of WPGN in practical application.

5. Conclusions

In this paper, we proposed the WPGN. To the best of our knowledge, this is the first attempt at combining semantic information with the GNN in FSL. Several experiments showed that our method achieves state-of-the-art results in fine-grained FSL. Compared with the baseline, the 5-way-1 shot task was improved by nearly 9%. At the same time, we can see that WPGN has a greater accuracy improvement on the CUB-200-2011 dataset than the other two datasets. The CUB-200-2011 dataset is a fine-grained dataset containing 200 categories of birds. Because of the great similarity between the classes, the fusion of semantic information and visual information can play a better role in this application scenario. In the use of semantic information, we used the GloVe model to extract word vectors, which greatly enhanced the practicability of this method. The effect of WPGN in the classification of rare birds shows that the method is flexible and feasible in practical applications. However, there are some limitations to our model. The WPGN fails to integrate semantic information with the GNN to the maximum extent and must therefore be improved in this regard. In future works, we will explore further improved methods to embed semantic information in fine-grained FSL to improve the influence of semantic information on classification accuracy. The uncertainty of a graph neural network in terms of layers will also be the focus of our future work.

Author Contributions

C.Z., data curation, conceptualization, software, validation, writing—original draft. L.W., data curation, resources, methodology, supervision, writing—review and editing. C.H., funding acquisition, project administration, visualization, investigation, writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Science Foundation for Young Scientists of China (grant no. 61702051, grant no. 61602058).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: CUB-200-2011: https://resolver.caltech.edu/CaltechAUTHORS:20111026-120541847, MiniImagenet: https://www.image-net.org/, CIFAR-FS: DOI: 10.1109/IROS45743.2020.9341282.

Acknowledgments

Special thanks to Zhang Yuqiang for her contribution to the methodology and polishing of our paper.

Conflicts of Interest

The authors declared that they have no conflict of interest for this work. We declare that we do not have any commercial or associative interest that represents a conflict of interest in connection with the work submitted.

References

  1. Finn, C.; Abbeel, P.; Levine, S. Model-agnostic meta-learning for fast adaptation of deep networks. arXiv, 2017; arXiv:1703.03400. [Google Scholar]
  2. Jamal, M.A.; Qi, G.J. Task agnostic meta-learning for few-shot learning. In Proceedings of the IEEE/CVF Conference on CVPR, Long Beach, CA, USA, 15–20 June 2019; pp. 11719–11727. [Google Scholar]
  3. Santoro, A.; Bartunov, S.; Botvinick, M.; Wierstra, D.; Lillicrap, T. Meta-learning with memory augmented neural networks. In Proceedings of the International Conference on Machine Learning (ICML), New York, NY, USA, 19–24 June 2016; pp. 1842–1850. [Google Scholar]
  4. Bertinetto, L.; Henriques, J.F.; Torr, P.H.; Vedaldi, A. Meta-learning with differentiable closed-form solvers. arXiv 2019, arXiv:1805.08136. [Google Scholar]
  5. Ohkuma, T.; Nakayama, H. Belonging network. In Proceedings of the 2020 IEICE-120 IEICE-PRMU-36-IEICE-PRMU-41, Online, 2–4 December 2020. [Google Scholar]
  6. Higashi, R.; Wada, T. Regularization using knowledge distillation in learning small datasets. IEICE Tech. Rep. 2020, 120, 133–138. [Google Scholar]
  7. Wang, X.; Yu, F.; Wang, R.; Darrell, T.; Gonzalez, J.E. TAFE-Net: Task-aware feature embeddings for low shot learning. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 1831–1840. [Google Scholar]
  8. Koch, G.; Zemel, R.; Salakhutdinov, R. Siamese neural for one-shot image recognition. In Proceedings of the ICML Deep Learning Workshop, Lille, France, 6–11 July 2015; Volume 2. [Google Scholar]
  9. Vinyals, O.; Blundell, C.; Lillicrap, T.; Wierstra, D. Matching networks for one shot learning. Adv. Neural Inf. Process. Syst. 2016, 29, 3630–3638. [Google Scholar]
  10. Zhang, C.; Cai, Y.; Lin, G.; Shen, C. DeepEMD: Few-Shot Image Classification with Differentiable Earth Mover’s Distance and Structured Classifiers. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 12203–12213. [Google Scholar]
  11. Li, A.; Huang, W.; Lan, X.; Feng, J.; Li, Z.; Wang, L. Boosting few-shot learning with adaptive margin loss. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 12576–12584. [Google Scholar]
  12. Bateni, P.; Goyal, R.; Masrani, V.; Wood, F.; Sigal, L. Improved few-shot visual classification. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 14493–14502. [Google Scholar]
  13. Ma, N.; Zhang, X.; Sun, J. Funnel activation for visual recognition. arXiv 2020, arXiv:2007.11824. [Google Scholar]
  14. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient channel attention for deep convolutional neural networks. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  15. Garcia, V.; Bruna, J. Few-shot learning with graph neural networks. arXiv 2017, arXiv:1711.04043. [Google Scholar]
  16. Kim, J.; Kim, T.; Kim, S.; Yoo, C.D. Edge-labeling graph neural network for few-shot learning. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 11–20. [Google Scholar]
  17. Yang, L.; Li, L.; Zhang, Z.; Zhou, X.; Zhou, E.; Liu, Y. DPGN: Distribution propagation graph network for few-shot learning. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 13390–13399. [Google Scholar]
  18. Yu, Z.; Raschka, S. Looking back to lower-level information in few-shot learning. Information 2020, 11, 345. [Google Scholar] [CrossRef]
  19. Gidaris, S.; Komodakis, N. Generating classification weights with gnn denoising autoencoders for few-shot learning. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 21–30. [Google Scholar]
  20. Yu, Z.; Chen, L.; Cheng, Z.; Luo, J. Transmatch: A transfer-learning scheme for semi-supervised few-shot learning. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 12856–12864. [Google Scholar]
  21. Schwartz, E.; Karlinsky, L.; Feris, R.; Giryes, R.; Bronstein, A.M. Baby steps towards few-shot learning with multiple semantics. arXiv 2020, arXiv:1906.01905. [Google Scholar]
  22. Schonfeld, E.; Ebrahimi, S.; Sinha, S.; Darrell, T.; Akata, Z. Generalized zero and few-shot learning via aligned variational autoencoders. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 8247–8255. [Google Scholar]
  23. Tokmakov, P.; Wang, Y.-X.; Hebert, M. Learning compositional representations for few-shot recognition. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27–28 October 2019; pp. 6372–6381. [Google Scholar]
  24. Li, A.; Luo, T.; Lu, Z.; Xiang, T.; Wang, L. Large-Scale few-shot learning: Knowledge transfer with class hierarchy. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 7212–7220. [Google Scholar]
  25. Cheny, Z.; Fuy, Y.; Zhang, Y.; Jiang, Y.-G.; Xue, X.; Sigal, L. Multi-level semantic feature augmentation for one-shot learning. IEEE Trans. Image Process. 2019, 28, 4594–4605. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Chen, T.; Wu, W.; Gao, Y.; Dong, L.; Luo, X.; Lin, L. Fine-Grained Representation Learning and Recognition by Exploiting Semantic Embedding. In Proceedings of the 26th ACM International Conference on Multimedia, Seoul, Korea, 22–26 October 2018; pp. 2023–2031. [Google Scholar]
  27. Jiang, Z.; Kang, B.; Zhou, K.; Feng, J. Few-shot classification via adaptive attention. arXiv 2020, arXiv:2008.02465. [Google Scholar]
  28. Abayomi-Alli, O.; Damaševičius, R.; Maskeliūnas, R.; Misra, S. Few-shot learning with a novel Voronoi tessellation-based image augmentation method for facial palsy detection. Electronics 2021, 10, 978. [Google Scholar] [CrossRef]
  29. Moon, J.; Le, N.; Minaya, N.; Choi, S.-I. Multimodal few-shot learning for gait recognition. Appl. Sci. 2020, 10, 7619. [Google Scholar] [CrossRef]
  30. Pennington, J.; Socher, R.; Manning, C.D. GloVe: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014. [Google Scholar]
  31. Miller, G.A. WordNet: A lexical database for English. Commun. ACM 1995, 11, 39–41. [Google Scholar] [CrossRef]
  32. Wah, C.; Branson, S.; Welinder, P.; Perona, P.; Belongie, S. The Caltechucsd Birds-200-2011 Dataset; California Institute of Technology: Pasadena, CA, USA, 2011. [Google Scholar]
  33. Chen, W.Y.; Liu, Y.C.; Kira, Z.; Wang, Y.C.F.; Huang, J.B. A closer look at few-shot classification. arXiv 2019, arXiv:1904.04232. [Google Scholar]
Figure 1. The figure shows a 3-way-1 shot task, where 1, 2, 3 represent the support set, 4 represents the query set, and the red line segment represents the mini-mum similarity between 4 and other support set samples. After cyclic calculation of the L-layer Point Graph and Word Distribution Graph, the class output with the closest distance between 4 and the other three types of samples in the Word Distribution Graph is finally selected as the prediction result.
Figure 1. The figure shows a 3-way-1 shot task, where 1, 2, 3 represent the support set, 4 represents the query set, and the red line segment represents the mini-mum similarity between 4 and other support set samples. After cyclic calculation of the L-layer Point Graph and Word Distribution Graph, the class output with the closest distance between 4 and the other three types of samples in the Word Distribution Graph is finally selected as the prediction result.
Sensors 22 02648 g001
Figure 2. Few-shot tasks and traditional image classification tasks in CUB-200-2011 dataset: (a) traditional classification; (b) few-shot classification.
Figure 2. Few-shot tasks and traditional image classification tasks in CUB-200-2011 dataset: (a) traditional classification; (b) few-shot classification.
Sensors 22 02648 g002
Figure 3. ECAResNet-12 network architecture.
Figure 3. ECAResNet-12 network architecture.
Sensors 22 02648 g003
Figure 4. Cycle calculation process of the WPGN.
Figure 4. Cycle calculation process of the WPGN.
Sensors 22 02648 g004
Figure 5. Details about W2P strategy and P2W strategy in the WPGN. (a) W2P; (b) P2W.
Figure 5. Details about W2P strategy and P2W strategy in the WPGN. (a) W2P; (b) P2W.
Sensors 22 02648 g005
Figure 6. Four different classes of birds in CUB-200-2011.
Figure 6. Four different classes of birds in CUB-200-2011.
Sensors 22 02648 g006
Figure 7. Effects of different layer numbers on classification accuracy.
Figure 7. Effects of different layer numbers on classification accuracy.
Sensors 22 02648 g007
Figure 8. Impact of different layer numbers in the WPGN on classification accuracy: (a) layer 0; (b) layer 1; (c) layer 2; (d) layer 3; (e) layer 5; (f) ground truth.
Figure 8. Impact of different layer numbers in the WPGN on classification accuracy: (a) layer 0; (b) layer 1; (c) layer 2; (d) layer 3; (e) layer 5; (f) ground truth.
Sensors 22 02648 g008aSensors 22 02648 g008b
Figure 9. Experimental results on the MiniImagenet and CIFAR-FS. (a) MiniImagenet; (b) CIFAR-FS.
Figure 9. Experimental results on the MiniImagenet and CIFAR-FS. (a) MiniImagenet; (b) CIFAR-FS.
Sensors 22 02648 g009
Figure 10. Training loss and test accuracy comparison.
Figure 10. Training loss and test accuracy comparison.
Sensors 22 02648 g010
Figure 11. Practical application example of rare bird classification.
Figure 11. Practical application example of rare bird classification.
Sensors 22 02648 g011
Table 1. Different measurement methods.
Table 1. Different measurement methods.
DistanceManhattanEuclideanMahalanobis
Dataset
CUB-200-201183.5183.8184.34
MiniImageNet69.9270.3470.69
CIFAR-FS79.679.980.4
Table 2. Effectiveness of introducing FReLU into the WPGN.
Table 2. Effectiveness of introducing FReLU into the WPGN.
FunctionReLUPReLUSwishLeakyReLUFReLU
A c c 82.5182.9283.5583.9584.34
Table 3. λ’s influence on classification accuracy.
Table 3. λ’s influence on classification accuracy.
λ0.10.20.30.40.5
A c c 81.8682.3282.5982.8583.15
λ0.60.70.80.91.0
A c c 83.3283.5183.8284.3483.88
Table 4. Hardware and software environments.
Table 4. Hardware and software environments.
GPUPythonTorchCUDATorchvisionTorchnet
1080Ti3.5.21.110.10.3.00.0.4
Table 5. Details for few-shot dataset benchmarks.
Table 5. Details for few-shot dataset benchmarks.
DatasetImagesClassesTrain/Val/TestResolution
MiniImageNet60 k10064/16/20 [9]84 × 84
CUB-20011.7 k200100/50/50 [33]84 × 84
CIFAR-FS60 k10064/16/20 [4]32 × 32
Table 6. Hyperparameter settings.
Table 6. Hyperparameter settings.
ParameterValue
Adam learning rate10−3
Decay learning rate10−1
Decay iterations12,000
Weight decay10−5
Layer number5
Table 7. Few-shot classification accuracies on CUB-200-2011.
Table 7. Few-shot classification accuracies on CUB-200-2011.
ModelBackbone1 Shot2 Shot5 Shot
MAMLConvNet55.92 ± 0.87/72.09 ± 0.76
MatchingNetConvNet61.16 ± 0.95/72.86 ± 0.69
RelationNetConvNet62.45 ± 0.89/76.11 ± 0.66
CloserLookConvNet60.53 ± 0.87/79.34 ± 0.69
DPGN (Batch size: 30)ConvNet75.52 ± 0.5985.65 ± 0.5289.31 ± 0.51
DPGN (Batch size: 40)ConvNet76.05 ± 0.51/89.08 ± 0.38
WPGN (Batch size: 30)ConvNet81.25 ± 0.4688.62 ± 0.3892.65 ± 0.42
FEATResNet-1268.87 ± 0.22/82.90 ± 0.15
DPGN (Batch size: 30)ResNet-1275.31 ± 0.3187.72 ± 0.4190.26 ± 0.30
DPGN (Batch size: 40)ResNet-1275.71 ± 0.47/91.48 ± 0.33
WPGN (Batch size: 30)ResNet-1283.05 ± 0.4591.31 ± 0.3493.91 ± 0.33
WPGN (Batch size: 30)ECARes-1284.34 ± 0.6692.28 ± 0.4194.41 ± 0.28
Table 8. Training time comparison.
Table 8. Training time comparison.
WPGNDPGN
Steps40,00040,000
Time (minutes)416724
Table 9. Ablation experiment on CUB-200-2011 and CIFAR-FS.
Table 9. Ablation experiment on CUB-200-2011 and CIFAR-FS.
DatasetsWord
Embedding
Mahalanobis DistanceFReLUECA
Attention
Accuracy
××××75.31
×××82.54
CUB-200-2011××82.91
×83.10
84.34
××××76.8
×××78.9
CIFAR-FS××79.2
×79.5
80.4
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhu, C.; Wang, L.; Han, C. Word Embedding Distribution Propagation Graph Network for Few-Shot Learning. Sensors 2022, 22, 2648. https://doi.org/10.3390/s22072648

AMA Style

Zhu C, Wang L, Han C. Word Embedding Distribution Propagation Graph Network for Few-Shot Learning. Sensors. 2022; 22(7):2648. https://doi.org/10.3390/s22072648

Chicago/Turabian Style

Zhu, Chaoran, Ling Wang, and Cheng Han. 2022. "Word Embedding Distribution Propagation Graph Network for Few-Shot Learning" Sensors 22, no. 7: 2648. https://doi.org/10.3390/s22072648

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop