Next Article in Journal
Using a Hybrid Neural Network and a Regularized Extreme Learning Machine for Human Activity Recognition with Smartphone and Smartwatch
Previous Article in Journal
A Scheduling Method of Using Multiple SAR Satellites to Observe a Large Area
Previous Article in Special Issue
ADE-CycleGAN: A Detail Enhanced Image Dehazing CycleGAN Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Recognition of Occluded Goods under Prior Inference Based on Generative Adversarial Network

1
School of Electronic and Information, Yangtze University, Jingzhou 434023, China
2
National Electrical and Electronic Experimental Teaching Demonstration Center, Yangtze University, Jingzhou 434023, China
3
Western Research Institute, Yangtze University, Karamay 834000, China
4
School of Computer Science, Yangtze University, Jingzhou 434023, China
5
School of Computer Science and Engineering, Central South University, Changsha 410083, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2023, 23(6), 3355; https://doi.org/10.3390/s23063355
Submission received: 23 December 2022 / Revised: 11 February 2023 / Accepted: 21 March 2023 / Published: 22 March 2023
(This article belongs to the Special Issue Image Processing and Pattern Recognition Based on Deep Learning)

Abstract

:
Aiming at the recognition of intelligent retail dynamic visual container goods, two problems that lead to low recognition accuracy must be addressed; one is the lack of goods features caused by the occlusion of the hand, and the other is the high similarity of goods. Therefore, this study proposes an approach for occluding goods recognition based on a generative adversarial network combined with prior inference to address the two abovementioned problems. With DarkNet53 as the backbone network, semantic segmentation is used to locate the occluded part in the feature extraction network, and simultaneously, the YOLOX decoupling head is used to obtain the detection frame. Subsequently, a generative adversarial network under prior inference is used to restore and expand the features of the occluded parts, and a multi-scale spatial attention and effective channel attention weighted attention mechanism module is proposed to select fine-grained features of goods. Finally, a metric learning method based on von Mises–Fisher distribution is proposed to increase the class spacing of features to achieve the effect of feature distinction, whilst the distinguished features are utilized to recognize goods at a fine-grained level. The experimental data used in this study were all obtained from the self-made smart retail container dataset, which contains a total of 12 types of goods used for recognition and includes four couples of similar goods. Experimental results reveal that the peak signal-to-noise ratio and structural similarity under improved prior inference are 0.7743 and 0.0183 higher than those of the other models, respectively. Compared with other optimal models, mAP improves the recognition accuracy by 1.2% and the recognition accuracy by 2.82%. This study solves two problems: one is the occlusion caused by hands, and the other is the high similarity of goods, thus meeting the requirements of commodity recognition accuracy in the field of intelligent retail and exhibiting good application prospects.

1. Introduction

With the expansion of the internet in recent years, purchasing and consumption have become more convenient. For example, vending machines may be employed in a variety of complicated applications owing to their low labor costs, compact footprint, low running expenses, and high income per unit. The rapid development of consumer poverty alleviation vending machines is a concrete practice for consolidating and expanding the achievements of poverty alleviation and rural revitalization; additionally, it is consistent with the new tendency of “non-touch commerce” in the face of the epidemic [1,2,3]. Traditional vending cabinets are based primarily on the radio frequency identification (RFID) [4] and static object detection [5] technologies; however, they tend to have significant manufacturing and maintenance expenses. Dynamic and smart vision-based cabinets, as opposed to traditional cabinets, integrate computer vision and cloud computing to complete the purchasing process of customers, with precise recognition of commodities being one of its most significant and key functions. The entire purchasing process is recorded by a camera to recognize the classes of commodities in the users’ hands [6]; however, because of the varying degrees of occlusion brought by the actions of manual picking up of commodities, the ability of convolutional neural network to extract features for commodities is limited, which results in inaccurate recognition results.
Two main issues need to be resolved to recognize commodities with occlusion caused by hand and occasion when the features of goods are highly similar: 1. the lack of commodity features owing to occlusion caused by human hands, thus necessitating restoration and expansion; 2. the low recognition accuracy due to the high degree of similarity of some products, which dictates the need to distinguish between similar features.
To address these issues, this study proposes an algorithm for goods recognition under occlusion based on prior inference and spherical clustering. First, generative adversarial network (GAN) is combined with semantic inference whilst appropriate noise priors are matched with pre-trained generators and noise predictors. Next, a Hausdorff distance-based contour structure loss function is used during the training process to render the process of feature expansion more effective and accurate. Then, an attention mechanism is utilized to select the most discriminative fine-grained features for the expanded features. Subsequently, the von Mises–Fisher (vMF) distribution [7] is utilized to map the expanded and selected fine-grained features onto a unit hypersphere for clustering to increase the spacing of features. During the process, an angle loss is designed to improve the effect of clustering, and finally, the softmax function is utilized to obtain the ultimate recognition results.

2. Related Work

2.1. Conventional Recognizing Method

Several algorithms have been used to recognize smart retail commodities. The YOLO series [8] and SSD [9] are typical examples of single-stage algorithms that combine detection with classification tasks into a single stage. Faster-region-based convolutional neural network (Faster-RCNN) [10] and cascade R-CNN [11] are examples of multi-stage algorithms that divide feature extraction and regression tasks into two stages. Recognition of objects and faces under occlusion has advanced rapidly in recent years, thus yielding novel ideas and approaches for recognizing occluded commodities [12,13,14], which are broadly classified into two classes, one of which is adding weight to the unoccluded part and another is restoring the occluded part. Kortylewski et al. [15] proposed a compositional convolutional neural network (CNN) model to recognize products based on unoccluded parts. Wang et al. [16] proposed an object shape feature extraction approach called slope difference distribution (SDD), which extracts features of shape as a sparse representation and utilizes the detected SDD features of all shape models and the minimum distance between SDD features for object recognition. Ma et al. [17] proposed a robust face recognition approach based on a sparse network with limited probability, built a sparse image network with limited probability, and acquired the overall training images from a global perspective for recognition. Heo et al. [18] proposed an occlusion-aware spatial attention transformer (OSAT) architecture based on a visual transformer (ViT), CutMix strengthening, and occlusion mask predictor (OMP) to solve the occlusion problem. Xu et al. [19] proposed a double-active-layer-based CNN to recognize facial expressions with high accuracy by learning robust and discriminative features from data. However, because of the occlusion caused by the hand and the great resemblance of the commodities, the accuracy of goods recognition using conventional methods is low.

2.2. Recognizing Based on Feature Expansion Method

Although the network topologies and approaches described above can successfully improve the recognition of items and faces under occlusion, the recognition accuracy remains low because of the lack of goods features. A GAN [20] is an adversarial game-based neural network that can be utilized to generate missing features of commodities. Arjovsky et al. [21] proposed Wasserstein GAN (WGAN), which addresses the difficulties in gradient disappearance and collapse during the training of GAN networks. Gulrajani et al. [22] proposed the WGAN-gradient penalty (WGAN-GP) to boost the convergence speed of networks. Liao et al. [23,24] proposed a semantic guidance and evaluation network (SGE-Net) to iteratively update the structural prior and restore images in an interactive framework of semantic extraction and image restoration; additionally, they designed a semantic wise attention propagation (SWAP) module to restore the integral details of texture. Li et al. [25] proposed a recursive feature reasoning (RFR) network to restore the largely missing textures of images. However, feature expansion and restoration can not address the problem of high similarity of goods.

2.3. Recognizing Based on Fine-Grained Method

Feature restoration and expansion are excellent solutions to the problem of the small number of product features; however, their similarity to other commodities is a crucial challenge for the recognition of comparable commodities. Geng et al. [26] proposed an approach for the recognition of fine-grained commodities based on feature matching and one-time deep learning to handle the issue of high feature similarity. Lee et al. [27] proposed a network that combines a linear model with a deep learning model, considering both discrete features and the content of continuous images. Rao et al. [28] proposed a counterfactual attention learning approach based on causal inference to concentrate on fine-grained features through counterfactual intervention. Wang et al. [29] proposed an improved fine-grained classification approach based on self-attention destruction and constructive learning (SADCL). Liu et al. [30] proposed a scale-consistent attention part network (SCAPNet) for fine-grained image recognition. Although the fine-grained recognition accuracy for unoccluded goods is high, it cannot achieve its best efficiency for the recognition of goods under occlusion.

3. Methods

Figure 1 shows the flowchart of the overall algorithm in this study. The algorithm includes three elements, namely GAN pretraining, feature expansion and selection of commodities, and features classification of commodities. (1) The GAN model and parameters are trained with the addition of a semantic inference module and noise prior to increase the accuracy and efficiency of the generated features. (2) The pre-trained GAN is utilized for the restoration and expansion of features of the occluded parts and then combines them with LBP features [31]. Simultaneously, the attention mechanism is utilized to select features with a greater degree of discrimination after expansion. (3) Metric learning based on vMF distribution is utilized to distinguish and classify among selected features.

3.1. GAN Pretraining

Regarding the similarity of object recognition, every time a new object is under recognition when the network is trained, the recognition efficiency of the processing is low. To increase the accuracy and efficiency of the entire recognition process, all the images of commodities in the library are sent to the GAN for pre-training, during which each input image is matched with an appropriate latent space noise distribution [32], which in turn is utilized in the actual recognition process of commodities under occlusion. The predicted noise prior is used for the following process of feature restoration and expansion to guarantee the accuracy of the generated features.

3.1.1. Generator Pretraining

In the network training process, the generator input is set as random latent spatial noise [33] and a commodity feature map is generated during the overall process. After completing the iterative training, the generator parameters are saved and await feature expansion of the commodity in the next process. The generator is a neural network with an encoder–decoder architecture, where the encoder is associated with five dense blocks in sequence, whereas the decoder is associated with four dense blocks and three semantic inference modules (SIM) connected crosswise in sequence. The encoder and the decoder are linked using a context inference module (CIM). Each dense block of the encoder comprises three layers of batch normalization, an activating function layer (LeakyReLU), and convolution layer (Conv2D), wherein the corresponding size of the convolution kernel in the convolution layer is 3 × 3. Padding is not utilized in the convolution layer, and the alpha value of LeakyReLU is 0.01. Each dense block is followed by a downsample layer made up of a 1 × 1 convolution layer and a 2 × 2 pooling layer to halve the size of the feature map. The encoded contextual features are sent to the CIM for the inference of contextual features, after which the encoded feature map is sent for decoding. The structure of the four dense blocks in the decoder is identical to those of the encoder. Each dense block is followed by an oversampling layer composed of a 1 × 1 convolution layer and transposed convolution layer, which doubles the size of the feature map. The SIM process is illustrated in Figure 2.
The current encoding feature  ϕ l and decoding features  φ l 1 are simultaneously sent to SIM wherein they are first fused with a skipping connection, as follows:
f l 1 = i = 1 C ( φ l 1 + ϕ l K i T ) K i ,
where  K i and  K i T are the convolution kernels on the i-th channel and its transpose, respectively; and  represents convolution,  l ∈ 1,2,3,4.
Next, semantic segmentation is implemented on the image to acquire a semantic segmentation map  S l . This allows for the implement spatial adaptive normalization on its parameters to acquire the updated inferred features:
F l 1 = γ f l 1 max ( f l 1 ) max ( f l 1 ) min ( f l 1 ) + β ,
where  γ and  β are parameters in the semantic segmentation map  S l , with  γ representing the normalized scaling coefficient and  β is the term of bias; and ⊙ representing the corresponding multiplication of matrix elements.
The generator and discriminator constantly learn from each other, and the generator parameters are fixed when reaching the Nash equilibrium. The network structure of the generator is illustrated in Figure 3.

3.1.2. Discriminator Pretraining

The discriminator constantly learns from the generator to increase the recognition accuracy of the sample. Herein, the CNN was utilized in the discriminator training process, with the real and feature maps generated by the generator as the inputs, to determine whether the generated feature map was suitable for a real distribution. Additionally, dimensionality reduction was accomplished using a convolutional layer with a step size of 2, convolutional kernel size of 3 × 3, activation function of LeakyReLU, dropout operation to avoid overfitting, parameter p = 0.2, and output layer activated by the Tanh activation function. The fully linked layer outputted the final discriminant result.

3.1.3. Noise Prior Pretraining

In this study, the precision of the feature expansion was improved by training the predictor to match the precise prior noise during the feature-generating process. First, the images of the occluded commodities were input, and the predictor was utilized to match the appropriate noise prior to guiding the generator for the generation of correct features. These are highly similar to the known features of commodities and thus ensure the accuracy of the generated features. The parameters were fixed and utilized for the expansion of commodity features after numerous rounds of iterative training.

3.1.4. Loss Function

Through an adversarial game involving generator G and discriminator D, the GAN optimizes the value function V(G,D) as follows:
min G max D V ( D , G ) = E x p r ( x ) [ log D ( x ) ] + E z p g ( z ) [ log ( 1 D ( G ( z ) ) ) ] ,
where z is the random noise input in the latent space; G(z) denotes the generated feature matrix; D(x) denotes the probability that the discriminator’s judgment of the generated features of the generator is true;  p r ( x ) denotes the distribution of the real features; and  p g ( z ) denotes the distribution of the generated features.
The JS divergence between  p r ( x ) and  p g ( z ) was optimized to maximize the training effect of G and D in the game between the generator and discriminator to reach  p r ( x ) p g ( z ) . However, the gradient vanishing phenomenon often occurs during the training process of the GAN, thereby resulting in the divergence of JS being a constant and producing a gradient vanishing issue. The similarity index of the two distributions is calculated using the Wasserstein distance combined with the Sobolev constraint [34], which is more versatile compared with the Lipschitz constraint.
The Wasserstein distance of the Sobolev constraint is
min G max D L s ( D , G ) = E x p r D ( x ) E z p g D ( G ( z ) ) .
The constraint condition is
E x f x i , x j x D ( x ) 2 1 0 , x i p r , x j p g ,
where  f x i , x j ( x ) is the probability density function of random variable x between  x i and  x j . Assuming that t is a random variable obeying a uniform distribution between [0, 1], the linear interpolation of the real sample  x i and generated sample  x j is
x = t x i + ( 1 t ) x j .
Then,  f x i , x j ( x ) can be represented as
f x i , x j ( x ) = { x i x j 1 2 π σ e ( x μ ) 2 2 σ 2 d x 0 ,   otherwise   , x = t x i + ( 1 t ) x j ,
where x is the linear interpolation between the real sample  x i and the generated sample  x j ; and  μ and  σ ( σ > 0) are the mean and standard deviation of the Gaussian distribution, respectively.
For D(x) to satisfy the Sobolev constraint, the gradient penalty GP was used, which is defined as
G P = E x p x [ ( x D ( x ) 2 1 ) 2 ] .
When discriminating the contours of generated features and real features, a contour structure loss function called Hausdorff Distance loss was designed using the Hausdorff Distance model, which is more sensitive to the contour structure. The pixel points on the shallow contour feature map of the network were viewed as a set of k pixel points, where the contour of the feature map extracted from the real image was  C r = { A 1 , A 2 , , A k } , and the goods’ contour of the generated feature map was  C g = { B 1 , B 2 , , B k } . The directional Hausdorff Distance of the contours of the goods in the two images can be expressed as
h ( C r , C g ) = max A k C r min B k C g A k B k ,
h ( C g , C r ) = max B k C g min A k C r B k A k .
The coordinates of the pixel points in the real and generated feature maps correspond to each other.
Thus, the Hausdorff Distance of the coordinates is
H ( C r , C g ) = m a x { h ( C g , C g ) , h ( C g , C r ) } .
Therefore, the contour structure loss function  L H a u s d o r f f   d i s t a n c e is
L H a u s d o r f f d i s t a n c e = H ( C r , C g ) .
Consequently, the loss function of the discriminator in this study is
min D L s ( G , D ) = E x ~ p r [ D ( x ) ] E x ~ p g [ D ( x ) ] , + λ E x ~ f x i , x j [ ( x D ( x ) 2 1 ) 2 ] + α H ( C r , C g ) ,
where  P g is the distribution of the generated feature map;  P r is the distribution of the real feature map;  x represents the linear interpolation between the real sample  x i and the generated sample  x j ; and  λ is the weight coefficient, which is updated during the iteration process and taken as  λ = 8,  α = 4.
To guarantee the consistency between the generated and real features during the training process of the generator, the generator was trained iteratively using the combined loss of adversarial loss  L 1 and gradient loss  L 2 .
If only the contour structure loss function is available in the discriminator, the generator can guarantee only the generation of accurate coarse contours but not clear texture features. Therefore, an adversarial loss function was added, and its expression is
L 1 = max D [ log D ( ( x r ) ) + log ( 1 D ( x g ) ) ] ,
where  x r denotes the real feature distribution; and  x g denotes the generated feature distribution.
The  L 1 loss function ensures that the generated texture features are accurate; however, in the back-propagation process of the generated features, ensuring the feature coherence of the generated feature maps is challenging. Therefore, this study introduced the gradient loss function  L 2 to ensure coherence between the generated features and the real features, as follows:
L 2 = F r F g | | F g ( x , y ) | | d x d y ,
where  F r denotes the real feature map;  F g denotes the generated feature map; and  | | F g ( x , y ) | | denotes the pixel gradient modulus of the point (x, y) in the generated feature map.
Thus, the loss function of the generator is as follows:
L g = β L 1 + γ L 2 ,
where  β and  γ are weight coefficients.
The overall objective function is as follows:
min G max D L ( G , D ) = E x ~ p r [ D ( x ) ] E x ~ p g [ D ( x ) ] , + λ E x ~ f x i , x j [ ( x D ( x ) 2 1 ) 2 ] , + α max { h ( C r , C g ) , h ( C g , C r ) } , + β max D [ log D ( ( x r ) ) + log ( 1 D ( x g ) ) ] , + γ F r F g F g ( x , y ) d x d y ,
where  λ α β , and  γ are the weight coefficients used to guarantee that the loss functions are balanced.

3.2. Expansion and Selection of Features

When recognizing similar occluded commodities, the number of features of the commodities typically determines the recognition accuracy. When a human hand picks up a product and creates different degrees of occlusion, the effective features available for recognition are reduced, and the commodity’s effective features need to be restored and expanded to increase the accuracy of commodity recognition. This study utilized a pre-trained GAN, semantic inference module, and predictor P to expand the number of commodity features.

3.2.1. Three-Channel Feature Extraction

When a commodity is occluded manually, using only a single gray channel to extract the features results in poor commodity recognition accuracy. Consequently, in this study, the recognition accuracy was improved by increasing the number of effective features of the commodity by extracting the features of goods using red–green–blue (RGB) three-channel feature extractor.
First, the OpenCV library was used to intercept the frame of the product image from a video stream photographed; subsequently, DarkNet53 was utilized as the backbone network to extract and concatenate the RGB three-channel feature map. That is,
F * = [ i = 1 c F i ] K ,
where  F * is the feature map after feature fusion; c = 3 is the number of channels;  F i is the feature map on the i-th channel; and K is a 1 × 1 convolution kernel.
The flow of feature extraction of RGB three-channel is shown in Figure 4.

3.2.2. Feature Restoration and Expansion

Herein, targeting the situation where the goods have few valid features owing to being occluded by hands, a combination of generative adversarial networks and semantic inference was used to extract relevant detailed features such as contours and textures.
First, the LBP operator was used to extract the texture features of the image of a good as it excels at describing the local texture features of the goods. Second, the previously extracted image features were sent to a pre-trained predictor (P) that predicts the latent spatial noise prior to adjusting its feature distribution. Finally, feature-matched noise was utilized to guide the expansion of features using a pre-trained generator.

3.2.3. Features Selection

Given that a large variety and number of features are provided, which is not beneficial for recognizing similar goods, herein, a self-designed lightweight attention mechanism, multi-scale spatial attention (MSSA), combined with effective channel attention (ECA), which focuses on the most distinctive features of the feature map and selects them in a short period, was used.
In MSSA, a 1 × 1, 2 × 2, and 3 × 3 convolution kernels were used to extract the multi-scale features to gain a feature pyramid, and the discrimination scores were calculated. The higher the discrimination score, the greater the distinction between features. The calculation process is as follows:
score = W 2 · SiLU ( W 1 G ( x ) ) ,
where G denotes global average pooling;  W 1 and  W 2 are different full connection layers; and the SiLU activation function was used to enhance the ability of non-linear activation.
Thus, the attention feature vectors were obtained through the discrimination score pyramid under spatial pyramid pooling (SPPF).
In ECA, first, pooled feature maps were acquired utilizing global average pooling; subsequently, a 3 × 1 fast convolution was utilized and the attention feature map was acquired using a sigmoid activation function. Next, the expanded features were fused with the attentional feature map, and the feature vectors under MSSA and ECA were weightedly fused to acquire the final selected features of goods. A flowchart of the feature selection is presented in Figure 5.

3.3. Features Distinction

The high similarity among features in the process of similar goods recognition under occlusion is the most significant problem that must be resolved. Herein, first, the similarities were classified at a finer scale to increase the spacing of the selected fine-grained features, and then the most identifiable features were selected to further improve the recognition accuracy. Further, metric learning based on the vMF distribution was used and a continuous probability distribution that is a Gaussian distribution on a sphere which is a distribution function for modeling high-dimensional spaces was used to increase the spacing of features. The accuracy and effectiveness of the classification can be improved by mapping the feature vectors to the unit hypersphere, treating them as feature points on the sphere, and then clustering them. This increases the separation among various classes of features, which provides each class with more distinct classification boundaries.

3.3.1. Feature Mapping and Clustering

Herein, the feature vectors were mapped onto the hypersphere using vMF distribution. The feature vector x must conform to the vMF distribution on the sphere if it conforms to the Gaussian distribution and the probability density function, which can be expressed as
f ( x | μ , k ) = c p ( k ) e k μ T x ,
where  μ , k are the mean vector and aggregation parameters, both of which parameterize the vMF distribution, respectively; and  c p ( k ) is the normalization constant, denoted as
c p ( k ) = k p / 2 1 ( 2 π ) p / 2 I p / 2 1 ( k ) ,
where  p is the number of variables and  I p / 2 1 ( k ) denotes the  P / 2 1 first- class modified Bessel function.
The vMF clustering centers were determined by maximizing the vMF likelihood of the training feature vectors. The vMF likelihood is defined as
p ( f j | λ ) = e σ μ T f i Z ( σ ) , | | f j | | = 1 , | | μ | | = 1 ,
where  μ is the mean value of the vMF distribution;  σ is the standard deviation;  λ = { μ , σ } Z ( σ ) is the normalization constant;  f j is the feature vector;  i is the number of cluster centers; and  f i is the cluster centroid feature vector.
Note that the loss function must be minimized to maximize the vMF probability; moreover, the feature vector  f j is assumed to be assigned to the vMF clustering center  μ i during training. In the clustering process, first, a random point was chosen from the hyperspherical feature points as the centroid of the first class of features; subsequently, the corresponding loss function was set to classify the nearest k points into one class. Here, k is the aggregation parameter, which represents the number of feature points clustered close to the centroid. Next, the second centroid point was determined, and so on, until the spherical feature points were clustered into n classes to complete the overall clustering process.
Given that the spacing between the features is in the form of arcs as they are situated on a sphere and move around the sphere as they cluster, their distances cannot be calculated using only the Euclidean distance. Therefore, the distances between the feature vectors and center vectors were quantified according to the angle  θ between them, and an angle-based loss function was constructed to constrain the entire clustering process. Figure 6 shows a cross-section of the unit hypersphere feature.
The following is the process of derivation. According to the cosine theorem,
cos θ = i N ( x i 2 + y i 2 ) d 2 2 i N x i 2 · i N y i 2 ,
where  ( x 1 , x 2 , , x N ) , ( y 1 , , y N ) denotes two random feature vectors with  N dimensions on the hypersphere; and  d denotes the Euclidean distance between them.
Given that the features lie in the unit hypersphere, the radius is 1, which yields
θ = arccos 2 d 2 2 .
The angle-based loss function  L θ can be defined as
L θ = arccos 2 n = 1 N ( x i _ n x j _ n ) 2 2 ,
where  ( x i _ 1 , x i _ 2 , , x i _ N ) are the coordinates of the i-th cluster center on the unit hypersphere of dimension  N , and the size of  N depends on the dimension of the feature vectors; and  ( x j _ 1 , x j _ 2 , , x j _ N ) are the coordinates of the jth feature vector on the unit hypersphere of dimension  N . Minimizing  L θ enables better clustering of the k points closest to the clustering center.
In the process of finding the (i + 1)-th centroid, a probability formula was used, defined as follows:
p = d ( f i , f j ) 2 f j F d ( f i , f j ) 2 ,
where  f i is the i-th cluster center feature point on the hypersphere; and  f j is the jth feature point on the hypersphere.
Under the constraint of the loss function, feature points are constantly clustered to form n classes of features on the hypersphere, with each class consisting of k feature points.

3.3.2. Feature Classification

The softmax loss function is normally used for the classification of features. After inputting the final feature vectors of the fully connected layer, the result is normalized, the input features are converted into the form of probability, and the one with the highest probability score is the final result of classification. Unlike the general softmax loss function, this study used spherical softmax loss which can better describe the distribution of feature points on the sphere and classify them.
This loss function can be defined as
L s = 1 N i = 1 N log e s cos θ y i j = 1 C e s cos θ j ,
where  s is the scale factor and  s = 5; N is the number of feature points;  C is the number of categories;  θ y i is the angle between the weight matrix of the sample labels and the feature vector  x i ; and  θ j is the angle between the weight matrix of the jth class of sample labels and the feature vector  x j .
For the final classification, the fine-grained characteristics that were clustered and mapped onto the hypersphere were fed into the softmax function, and the results of goods detection under occlusion were output.

4. Experimental Results and Analysis

To evaluate the performance of the proposed algorithm in feature restoration and product recognition accuracy, an ablation experiment and a comparative experiment are set up, and the performance of the algorithm is evaluated according to the experimental data.

4.1. Dataset

Owing to the complex situation of smart retail containers in practical applications, at present, only commodities appear in the open-source dataset, with no character interaction or alternative complex environments; therefore, the recognition accuracy of the algorithm in the actual application process cannot be guaranteed. To ensure that commercial standards can be reached in the actual operation process, the experimental data used in this study were all obtained from the self-made smart retail container dataset, which contains a total of 12 types of goods, including canned, bagged, and bottled goods. By filming the entire process from opening the door to picking up the product and closing the door, the video frame of the product picture from each angle was captured for training. Among them, the lighting conditions were normal indoor lighting and the background was dark owing to the light strip. Figure 7 shows the number of images for each product in the dataset. Figure 8 shows some of the pictures in our dataset.

4.2. Experiment Platform

The experimental platform included an operating system Windows 10, graphics card NVIDIA GeForce RTX3060, processor AMD Ryzen 7 5800H with Radeon Graphics. The Pytorch1.11 deep neural network framework and Python3.7 were used to build the network models. Figure 9 shows the UI Interface of proposed method.

4.3. Feature Restoration and Expansion Result

Note that owing to the occlusion caused by the hand when purchasing goods, some of the features of goods are missing, thus causing a decrease in recognition accuracy. Therefore, feature restoration and expansion were utilized to increase recognition accuracy. The following experiments were conducted to evaluate the effects of feature restoration and expansion. Further, a self-made dataset was utilized for feature restoration and expansion, in which a mask was used to simulate the occlusion caused by the hand.

4.3.1. Ablation Experiment

To justify the effectiveness of the Hausdorff Distance applied in the loss function of the discriminator in the proposed feature restoration and expansion algorithm in this study, a video frame during the user picking up the goods was shot and the ablation experiment was performed with the loss function in WGAN and WGAN-GP, which are the peak signal-to-noise ratio (PSNR) and structural similarity index measure (SSIM). After several debugging and optimization cycles, the parameters in the code were set with 300 epochs, a batch size of 32, Adam optimizer, learning rates of the generator and discriminator as 0.0001 and 0.0004, respectively, and the hyperparameters  β 1 and  β 2 as 0.5 and 0.9, respectively. The results of the experiments are listed in Table 1.

4.3.2. Comparison Experiment

To evaluate the capability of the Semantic inference module applied in generative adversarial network for feature restoration and expansion, the commonly utilized image inpainting algorithms partial convolution (PC) [35] and edge connect [36] were compared with the proposed method. Convolution visual evaluation was used to evaluate the effect of feature restoration and expansion, with experimental results indicating that the proposed method is superior to the abovementioned methods. The experimental results are shown in Figure 10.

4.4. Recognition Results of Occluded Goods

4.4.1. Ablation Experiment

To evaluate the effectiveness of prior GAN and spherical clustering in the occluded goods recognition algorithm in this study, a neural network ablation experiment based on DarkNet53 was performed, and its capability was evaluated using the F1 score and mean average precision (mAP).
F 1 Score = 2 × precision × recall precision + recall ,
mAP = n = 1 N AveP ( n ) N ,
where  N denotes the number of queries; and  AveP ( n ) denotes the average precision.
precision = TP TP + FP ,
recall = TP TP + FN ,
where TP denotes the positive samples predicted to be positive by the model; FN denotes the positive samples predicted to be negative by the model; and FP denotes the negative samples predicted to be positive by the model.
After debugging and optimization, the parameters in the code were set as 100 epochs, batch size 8, Adam optimizer, a learning rate of 0.0001, and the hyperparameters  β 1 β 2 were both equal to 0.999. The following experiments are based on the parameters used to train the neural networks and the experimental results are as Table 2. Although both our F1 score and mAP are superior to other combinations, the optimal parameters remain to be explored.
To evaluate the recognition performance of various neural network combinations on the dataset, the P–R curve, denoting the precision–recall curve, was utilized. In theory, the larger the area enclosed by the P–R curve and the horizontal and vertical axes, the stronger the capability of the algorithm. Results demonstrate that the P–R curve of the proposed method is superior to the following network combinations. The experimental results are shown in Figure 11.

4.4.2. Comparison Experiment

To evaluate the effectiveness of feature selection under the proposed attention mechanism, that is, ECA + MSSA, a comparison experiment based on squeeze excitation (SE) [37], convolutional block attention module (CBAM) [38], and ECA [39] was conducted to compare the attention heatmap with the proposed method. The experimental results are shown in Figure 12, with the proposed method achieving the best performance.
To justify the improvement of proposed attention mechanism on increasing the recognition accuracy, the effects of the proposed attention mechanism were compared with other attention mechanisms through recognition accuracy; that is, the actual videos of the purchasing process were recognized and the recognition results of each frame of the videos were averaged such that the class corresponding to the highest confidence was the ultimate recognition result of this purchasing process. Recognition accuracy denotes the number of correctly recognized videos divided by the total number of videos. The recognition accuracy in the following experimental results was based on the standard. The experimental results, including the F1 score, mAP, and recognition accuracy, are listed in Table 3.
To evaluate the proposed algorithm for occluded goods recognition, several mainstream object detection algorithms were compared with the proposed algorithm, and the F1 score and mAP of each algorithm during the training process were recorded for comparison. The experimental results are shown in Figure 13.
During the process of good recognition, a detection frame was drawn on the detected goods, and on this detection frame, the class and corresponding confidence were marked. A schematic is shown in Figure 14, which shows different scenarios and the purchasing process based on different persons.
To evaluate the generalization capability of various object recognition algorithms, the smart retail goods were divided into three shapes and the videos of each shape of goods were recognized in the actual purchasing process. The training effect of each algorithm reached an optimal state on the same dataset. The recognition accuracy results are presented in Table 4. The approach in this paper achieved good results, but there is little progress in the recognition accuracy of canned goods, which needs improvement.
To reflect the improvement of the proposed algorithm on the recognition accuracy of occluded similar goods, a comparison experiment comprising four groups of similar goods on various algorithms was designed. The recognition accuracy of each good was recorded, as summarized in Table 5. For similar goods recognition, Table 5 shows that the proposed approach performs well in most cases, but still has shortcomings in a small number of cases such as in recognizing “Sprite_Bottled” and “Little Nongfu Spring”, which needs improvement.

5. Conclusions

This study proposed an algorithm for the recognition of occluded goods under prior inference and spherical clustering, which resolves not only the problem of few features of goods caused by hand occlusion but also the occasion when the features of goods are highly similar.
  • To solve the problem of the small number of good features due to occlusion, the generator and predictor of the network were pretrained, and noise was used prior to guide the generation of features. By jumping and connecting context features for semantic inference, these features could be restored and expanded. The contour structure loss function based on the Hausdorff Distance designed in this study was used in the process of feature restoration and expansion to ensure the generation of accurate contour features.
  • To resolve the situation in which the product features are highly similar, MSSA+ECA was used to select the features in a fine-grained manner and the most discriminative features were screened out. The spherical model under the vMF distribution was used to map the feature vectors to the unit hypersphere and cluster it, thereby increasing the distance between features. In the spherical feature clustering process, the angle loss function designed in this study was used to effectively improve the clustering effect.
The experimental results in this study achieved good subjective and objective evaluation results, thus proving the feasibility and effectiveness of the proposed method. However, the proposed algorithm must be optimized and improved. Given that feature restoration and clustering require a certain amount of time and the consumption scenarios are complex, the real-time performance and generalization ability of the algorithm need to be improved in future research. Additionally, the algorithm process can be further explored to improve real-time and generalization capabilities to deal with more complex situations. In the future, the algorithm can be applied to the unmanned smart vending method of goods in shopping malls, supermarkets, and other places to provide a reliable and good recognition method for the smart retail field.

Author Contributions

Conceptualization, M.C. and K.X.; methodology, M.C.; software, M.C. and F.L.; validation, M.C., F.L. and B.L.; formal analysis, C.W.; investigation, J.H.; resources, M.C.; data curation, K.X.; writing—original draft preparation, M.C.; writing—review and editing, F.L.; visualization, K.X.; supervision, K.X.; project administration, W.Z.; funding acquisition, K.X. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by National Natural Science Foundation of China (No. 62272485) and Natural Science Foundation of Xinjiang Uygur Autonomous Region (Grant No.2020DO1A131), under the cooperation this work was successfully established and conducted. This work was mainly supported by the project sponsored by the Undergraduate Training Programs for Innovation and Entrepreneurship of Yangtze University under Grant Yz2021040 and Teaching Research Fund of Yangtze University (Grant No. JY2020101). The project enabled us to purchase and establish a smart retail container, and conduct the experiments in this article to complete the recognition of smart retail goods under occlusion.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data sharing not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Ethical Statement

The mention of different brands of products in the article is only for illustrative and explanatory purposes, and does not imply any intention of advertising or marketing.

References

  1. Han, P.; Zhang, J. Comparison of Investment Direction and Efficiency Analysis of China’s Rural Poverty Alleviation Fund. Int. Conf. Econ. Manag. Model Eng. (ICEMME) 2020, 2, 715–719. [Google Scholar]
  2. Zhang, L. The Construction of Rural Poverty Alleviation Audit using Big Data Anomaly Detection. In Proceedings of the 2021 Fifth International Conference on I-SMAC (IoT in Social, Mobile, Analytics and Cloud) (I-SMAC), Palladam, India, 11–13 November 2021. [Google Scholar]
  3. Cao, X.; Shi, D. Research on Poverty Alleviation by Informatization Education under the Background of Rural Revitalization. In Proceedings of the 2021 International Conference on Education, Information Management and Service Science (EIMSS), Xi’an, China, 16–18 July 2021. [Google Scholar]
  4. Lodhi, E.; Zhu, F.; Lodhi, Z.; Saleem, Q.; Xiong, G.; Wang, F. Design and Implementation of RFID Based Smart Shopping Booth; IEEE: Barcelona, Spain, 2019. [Google Scholar] [CrossRef]
  5. Liu, L.; Cui, J.; Huan, Y.; Zou, Z.; Hu, X.; Zheng, L. A Design of Smart Unmanned Vending Machine for New Retail Based on Binocular Camera and Machine Vision. IEEE Consum. Electron. Mag. 2022, 11, 21–31. [Google Scholar] [CrossRef]
  6. Zhang, H.; Li, D.; Ji, Y.; Zhou, H.; Wu, W.; Liu, K. Toward New Retail: A Benchmark Dataset for Smart Unmanned Vending Machines. IEEE Trans. Ind. Inform. 2020, 16, 7722–7731. [Google Scholar] [CrossRef]
  7. Kobayashi, T.; Otsu, N. Von Mises-Fisher Mean Shift for Clustering on a Hypersphere. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 2130–2133. [Google Scholar] [CrossRef]
  8. Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. YOLOX: Exceeding YOLO Series in 2021. In Proceedings of the Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 1–6. [Google Scholar]
  9. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single Shot Multibox Detector. In Proceedings of the European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2016; pp. 21–37. [Google Scholar]
  10. Meivel, S.; Sindhwani, N.; Anand, R.; Pandey, D.; Alnuaim, A.A.; Altheneyan, A.S.; Jabarulla, M.Y.; Lelisho, M.E. Mask Detection and Social Distance Identification Using Internet of Things and Faster R-CNN Algorithm. Comput. Intell. Neurosci. 2022, 2022, 2103975. [Google Scholar] [CrossRef] [PubMed]
  11. Li, Y.; Wang, Z. Research on Textile Defect Detection Based on Improved Cascade R-CNN. In Proceedings of the 2021 International Conference on Artificial Intelligence and Electromechanical Automation (AIEA), Guangzhou, China, 14–16 May 2021. [Google Scholar]
  12. Dapogny, A.; Bailly, K.; Dubuisson, S. Confidence-Weighted Local Expression Predictions for Occlusion Handling in Expression Recognition and Action Unit Detection. Int. J. Comput. Vis. 2018, 126, 255–271. [Google Scholar] [CrossRef] [Green Version]
  13. Chaudhury, A.; Barron, J.L. Plant Species Identification from Occluded Leaf Images. IEEE/ACM Trans. Comput. Biol. Bioinform. 2020, 17, 1042–1055. [Google Scholar] [CrossRef] [PubMed]
  14. Putri, D.U.K.; Musdholifah, A.; Makhrus, F.; Duong, V.H.; Le, P.T.; Chen, B.W.; Wang, J.C. Occluded Face Recognition Using Sparse Complex Matrix Factorization with Ridge Regularization. In Proceedings of the 2021 International Symposium on Intelligent Signal Processing and Communication Systems (ISPACS), Hualien City, Taiwan, 16–19 November 2021. [Google Scholar]
  15. Kortylewski, A.; Liu, Q.; Wang, A.; Sun, Y.; Yuille, A. Compositional Convolutional Neural Networks: A Robust and Interpretable Model for Object Recognition Under Occlusion. Int. J. Comput. Vis. 2020, 129, 736–760. [Google Scholar] [CrossRef]
  16. Wang, Z. Recognition of occluded objects by slope difference distribution features. Appl. Soft Comput. 2022, 120, 108622. [Google Scholar] [CrossRef]
  17. Ma, X.; Ma, Q.; Ma, Q.; Han, X. Robust face recognition for occluded real-world images using constrained probabilistic sparse network. IET Image Process 2022, 16, 1359–1375. [Google Scholar] [CrossRef]
  18. Heo, J.; Wang, Y.; Park, J. Occlusion-aware spatial attention transformer for occluded object recognition. Pattern Recognit. Lett. 2022, 159, 70–76. [Google Scholar] [CrossRef]
  19. Xu, Y.; Liu, J.; Zhai, Y.; Gan, J.; Zeng, J.; Cao, H.; Scotti, F.; Piuri, V.; Labati, R.D. Weakly supervised facial expression recognition via transferred DAL-CNN and active incremental learning. Soft Comput. 2020, 24, 5971–5985. [Google Scholar] [CrossRef]
  20. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the 2014 Advances in Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; Curran Associates: New York, NY, USA, 2014; pp. 2672–2680. [Google Scholar]
  21. Arjovsky, M.; Chintala, S.; Bottou, L. Wasserstein Gan. [EB/OL]. Available online: https://arxiv.org/abs/1701.07875 (accessed on 20 April 2021).
  22. Gulrajani, I.; Ahmed, F.; Arjovsky, M.; Dumoulin, V.; Courville, A.C. Improved training of Wasserstein gans. In Proceedings of the 2017 Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Curran Associates: New York, NY, USA, 2017; pp. 5767–5777. [Google Scholar]
  23. Liao, L.; Xiao, J.; Wang, Z.; Lin, C.-W.; Satoh, S. Guidance and Evaluation: Semantic-Aware Image Inpainting for Mixed Scenes. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2020; pp. 683–700. [Google Scholar]
  24. Liao, L.; Xiao, J.; Wang, Z.; Lin, C.-W.; Satoh, S. Image Inpainting Guided by Coherence Priors of Semantics and Textures. In Proceedings of the Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  25. Li, J.; Wang, N.; Zhang, L.; Du, B.; Tao, D. Recurrent Feature Reasoning for Image Inpainting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 7760–7768. [Google Scholar]
  26. Geng, W.; Han, F.; Lin, J.; Zhu, L.; Bai, J.; Wang, S.; He, L.; Xiao, Q.; Lai, Z. Fine-Grained Grocery Product Recognition by One-Shot Learning; ACM: New York, NY, USA, 2018; pp. 1706–1714. [Google Scholar]
  27. Lee, J.W.; Yoon, Y.C. Fine-Grained Plant Identification Using Wide and Deep Learning Model; IEEE: Barcelona, Spain, 2019. [Google Scholar]
  28. Rao, Y.; Chen, G.; Lu, J.; Zhou, J. Counterfactual Attention Learning for Fine-Grained Visual Categorization and Re-identification. Computer Vision and Pattern Recognition (CVPR). In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 1025–1034. [Google Scholar]
  29. Wang, W.; Cui, Y.; Li, G.; Jiang, C.; Deng, S. A self-attention-based destruction and construction learning fine-grained image classification method for retail product recognition. Neural Comput. Appl. 2020, 32, 14613–14622. [Google Scholar] [CrossRef]
  30. Liu, H.; Li, J.; Li, D.; See, J.; Lin, W. Learning Scale-Consistent Attention Part Network for Fine-Grained Image Recognition. IEEE Trans. Multimed. 2022, 24, 2902–2913. [Google Scholar] [CrossRef]
  31. Kato, K.; Takano, H.; Saiko, M.; Hiraike, R.; Imaoka, H. Effectiveness of Feature-Selected LBP-TOP for Emotional Estimation by Micro-Expression; IEEE: Barcelona, Spain, 2020. [Google Scholar]
  32. Lahiri, A.; Jain, A.K.; Agrawal, S.; Mitra, P.; Biswas, P.K. Prior Guided GAN Based Semantic Inpainting. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 13693–13702. [Google Scholar]
  33. Shen, Y.; Gu, J.; Tang, X.; Zhou, B. Interpreting the Latent Space of GANs for Semantic Face Editing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  34. Xu, M.; Zhou, Z.; Lu, G.; Tang, J.; Zhang, W.; Yu, Y. Sobolev Wasserstein GAN. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), Montreal, QC, Canada, 21–26 August 2020. [Google Scholar]
  35. Liu, G.; Shih, K.; Wang, T.C.; Tao, A.; Catanzaro, B. Image Inpainting for Irregular Holes Using Partial Convolutions. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  36. Nazeri, K.; Ng, E.; Joseph, T.; Qureshi, F.Z.; Ebrahimi, M. EdgeConnect: Generative Image Inpainting with Adversarial Edge Learning; IEEE: Barcelona, Spain, 2019; pp. 1–17. [Google Scholar]
  37. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Transactions on Pattern Analysis and Machine Intelligence, Salt Lake City, UT, USA, 18–23 June 2018; Volume 42, pp. 2011–2023. [Google Scholar] [CrossRef] [Green Version]
  38. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Computer Vision—ECCV 2018; Lecture Notes in Computer, Science; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Springer: Cham, Switzerland, 2018; Volume 11211. [Google Scholar]
  39. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020; pp. 11531–11539. [Google Scholar] [CrossRef]
Figure 1. Algorithm flow of the proposed architecture for goods recognition. The numbers 1 and 2 in the upper left and upper right corners represent the first and second parts of GAN pretraining.
Figure 1. Algorithm flow of the proposed architecture for goods recognition. The numbers 1 and 2 in the upper left and upper right corners represent the first and second parts of GAN pretraining.
Sensors 23 03355 g001
Figure 2. Semantic Inference Module. The current encoding feature  ϕ l and decoding features  φ l 1 are sent to SIM to be fused with a skipping connection.
Figure 2. Semantic Inference Module. The current encoding feature  ϕ l and decoding features  φ l 1 are sent to SIM to be fused with a skipping connection.
Sensors 23 03355 g002
Figure 3. Structure of generator.
Figure 3. Structure of generator.
Sensors 23 03355 g003
Figure 4. Feature extraction of RGB three-channel.
Figure 4. Feature extraction of RGB three-channel.
Sensors 23 03355 g004
Figure 5. Flow of features selection.
Figure 5. Flow of features selection.
Sensors 23 03355 g005
Figure 6. Cross-section of features on unit hypersphere. Different colored points represent feature vectors of different classes.
Figure 6. Cross-section of features on unit hypersphere. Different colored points represent feature vectors of different classes.
Sensors 23 03355 g006
Figure 7. The number of each kind of good in our dataset.
Figure 7. The number of each kind of good in our dataset.
Sensors 23 03355 g007
Figure 8. Some of the pictures in our dataset. Below these images are the names of these products.
Figure 8. Some of the pictures in our dataset. Below these images are the names of these products.
Sensors 23 03355 g008
Figure 9. The UI interface of proposed method.
Figure 9. The UI interface of proposed method.
Sensors 23 03355 g009
Figure 10. Comparison experiment of feature restoration and expansion.
Figure 10. Comparison experiment of feature restoration and expansion.
Sensors 23 03355 g010
Figure 11. Ablation experiment of P–R curve under different combinations of networks.
Figure 11. Ablation experiment of P–R curve under different combinations of networks.
Sensors 23 03355 g011
Figure 12. Heatmaps of various attention mechanism algorithms.
Figure 12. Heatmaps of various attention mechanism algorithms.
Sensors 23 03355 g012
Figure 13. Comparison experiment of training parameters on various algorithms. Note: Bold is the best result.
Figure 13. Comparison experiment of training parameters on various algorithms. Note: Bold is the best result.
Sensors 23 03355 g013
Figure 14. Diagrams of recognition results on various scenarios and persons.
Figure 14. Diagrams of recognition results on various scenarios and persons.
Sensors 23 03355 g014
Table 1. Ablation experiment of loss function.
Table 1. Ablation experiment of loss function.
Loss FunctionPSNRSSIM
  L o s s G A N + l o s s W a s s e r s t e i n 19.07420.7367
  L o s s G A N + l o s s W a s s e r s t e i n G P 19.20390.7481
Ours 119.97820.7664
1 Ours ( l o s s G A N + l o s s W a s s e r s t e i n G P + l o s s H a u s d o r f f ). Note: Bold is the best result.
Table 2. Parameters of ablation experiment of neural network.
Table 2. Parameters of ablation experiment of neural network.
BackboneF1 ScoremAP
DarkNet53 (baseline)0.940.934
DarkNet53 + Attention0.950.980
DarkNet53 + Pr-GAN + Attention0.950.985
Ours0.980.996
Note: Bold is the best result.
Table 3. Parameters of comparison experiment of attention mechanism.
Table 3. Parameters of comparison experiment of attention mechanism.
Network CombinationsF1 ScoremAPAccuracy
DarkNet53 (baseline)0.940.9340.894
DarkNet53 + SE0.950.9800.907
DarkNet53 + CBAM0.960.9830.899
DarkNet53 + ECA0.960.9840.916
Ours (DarkNet53 + MSSA + ECA)0.980.9950.937
Note: Bold is the best result.
Table 4. Comparison experiment of recognition accuracy of different kinds of goods on various algorithms.
Table 4. Comparison experiment of recognition accuracy of different kinds of goods on various algorithms.
Algorithm ModelsBottle ShapedCan ShapedBag Shaped
EfficientDet84.7488.2376.94
CenterNet81.3162.3961.87
SSD82.1484.2374.18
Faster-RCNN82.5665.6071.49
RetinaNet88.4781.0474.17
YOLOv591.1383.4380.57
YOLOX91.7796.4384.62
Ours94.6496.8091.42
Note: Bold is the best result.
Table 5. Comparison experiment of recognition accuracy of occluded similar goods on various algorithms.
Table 5. Comparison experiment of recognition accuracy of occluded similar goods on various algorithms.
Classes of Similar GoodsFaster-RCNNSSD RetinaNetYOLOv5YOLOXOurs
Large C’estbon95.9688.6795.9499.8495.8799.91
C’estbon75.1476.2175.2680.0996.7497.62
Sprite_Canned59.7683.5478.9181.4896.4496.53
Sprite_Bottled90.9293.3690.3799.8996.8797.46
Large Pepsi Cola86.7686.0498.8197.6297.6898.94
Pepsi Cola83.1177.6488.6097.2482.7997.49
Nongfu Spring77.1463.0894.4691.8292.1695.27
Little Nongfu Spring62.7484.0974.1890.1689.6287.46
Average Accuracy78.9481.5887.0792.2793.5296.34
Note: Bold is the best result.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cao, M.; Xie, K.; Liu, F.; Li, B.; Wen, C.; He, J.; Zhang, W. Recognition of Occluded Goods under Prior Inference Based on Generative Adversarial Network. Sensors 2023, 23, 3355. https://doi.org/10.3390/s23063355

AMA Style

Cao M, Xie K, Liu F, Li B, Wen C, He J, Zhang W. Recognition of Occluded Goods under Prior Inference Based on Generative Adversarial Network. Sensors. 2023; 23(6):3355. https://doi.org/10.3390/s23063355

Chicago/Turabian Style

Cao, Mingxuan, Kai Xie, Feng Liu, Bohao Li, Chang Wen, Jianbiao He, and Wei Zhang. 2023. "Recognition of Occluded Goods under Prior Inference Based on Generative Adversarial Network" Sensors 23, no. 6: 3355. https://doi.org/10.3390/s23063355

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop