Next Article in Journal
End-to-End Detection of a Landing Platform for Offshore UAVs Based on a Multimodal Early Fusion Approach
Next Article in Special Issue
Underwater Object Detection Using TC-YOLO with Attention Mechanisms
Previous Article in Journal
Crack Monitoring in Rotating Shaft Using Rotational Speed Sensor-Based Torsional Stiffness Estimation with Adaptive Extended Kalman Filters
Previous Article in Special Issue
Lightweight Helmet Detection Algorithm Using an Improved YOLOv4
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detection of Occluded Small Commodities Based on Feature Enhancement under Super-Resolution

1
School of Electronic Information, Yangtze University, Jingzhou 434023, China
2
Western Research Institute, Yangtze University, Karamay 834000, China
3
School of Computer Science, Central South University, Changsha 410083, China
4
National Super-Computer Center in Changsha, Hunan University, Changsha 410082, China
5
School of Information Science and Engineering, Hunan University, Changsha 410082, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2023, 23(5), 2439; https://doi.org/10.3390/s23052439
Submission received: 4 January 2023 / Revised: 19 February 2023 / Accepted: 21 February 2023 / Published: 22 February 2023
(This article belongs to the Special Issue Image Processing and Analysis for Object Detection)

Abstract

:
As small commodity features are often few in number and easily occluded by hands, the overall detection accuracy is low, and small commodity detection is still a great challenge. Therefore, in this study, a new algorithm for occlusion detection is proposed. Firstly, a super-resolution algorithm with an outline feature extraction module is used to process the input video frames to restore high-frequency details, such as the contours and textures of the commodities. Next, residual dense networks are used for feature extraction, and the network is guided to extract commodity feature information under the effects of an attention mechanism. As small commodity features are easily ignored by the network, a new local adaptive feature enhancement module is designed to enhance the regional commodity features in the shallow feature map to enhance the expression of the small commodity feature information. Finally, a small commodity detection box is generated through the regional regression network to complete the small commodity detection task. Compared to RetinaNet, the F1-score improved by 2.6%, and the mean average precision improved by 2.45%. The experimental results reveal that the proposed method can effectively enhance the expressions of the salient features of small commodities and further improve the detection accuracy for small commodities.

1. Introduction

In recent years, owing to the continuous development of technologies such as big data and the Internet of Things, artificial intelligence has gradually matured. The government has issued relevant policies to support the transformation of the retail industry to digital platforms and further promote the development of an intelligent retail industry; consequently, the offline retail market has witnessed continuous expansion. The global retail industry market size reached 27 trillion USD in 2021, and as per estimates, artificial intelligence will contribute additional growth of 2 trillion USD to retail by 2035, thereby providing massive business value.
Currently, two main solutions exist for retail containers: non-visual and visual methods. Non-visual methods primarily include gravity sensing and radio frequency identification technologies. However, these methods exhibit poor flexibility and increase the cost of commodities. At present, numerous target detection methods are based on convolutional neural networks (CNNs), such as faster-region-based CNNs (Faster-RCNN) [1], single-shot detection (SSD) [2], YOLO [3], and RetinaNet [4]. Retail containers in the market are increasingly using visual container technology [5] based on deep learning for commodity detection [6,7] and identification to realize the deduction of commodities purchased by customers and corresponding settlements.
Owing to the influence of light, transmission equipment, and the surrounding environment, the details of a video image can be substantially lost. Some researchers have conducted studies related to image super-resolution (SR) [8] to solve the problem of image blurring. For example, Noh divided an input low-resolution image into textured and non-textured regions [9] and then interpolated the image according to the features of local structures, thereby retaining the texture and structure information of the image while ignoring the contour information. To reduce the number of parameters and ensure good performance of the network, an ultra-lightweight SR network [10] was proposed to retain high-frequency details. However, the restored images exhibit structural distortions. Therefore, Ma [11,12] proposed a structure-preserving SR method with gradient guidance to alleviate the geometric distortion prevalent in the SR results from perception-driven methods. Additionally, a new module called feature texture transfer (FTT) [13] was used to extract trusted regional details. A texture- and detail-preserving network [14] was proposed, which can not only learn local and regional features but also pay attention to texture and detail features and restore high-resolution ratio images with better perceptual effects. In addition, some experts [15] decoupled the reference-based super-resolution from a new perspective, eliminating the interference between the LR image and the reference image. However, the generated image is easily lacking constraints with the original image. In this article, we combine texture, content, and contour features to obtain rich SR image information.
Recently, in terms of feature extraction, CNNs [16], residual networks [17,18], and other networks have been used to extract the target features. As small commodities are occluded [19], the effective features of such commodities are often missing, so image inpainting [20] algorithms are usually used to repair the incomplete image. However, existing studies can only display excellent results in accomplishing simple image structures and generating image content with a complex overall structure, and high fidelity of detail remains a huge challenge. Therefore, optimized residual mapping [21] was used to improve the learning ability of the residual network. Zhang [22] used densely connected convolution layers in residual dense blocks to extract rich local features. They reported that stacking additional residual blocks enhances the normalization preservation of a network [23]. Although these networks perform well at extracting features, they are extremely complex, resulting in a significant loss of efficiency. Some scholars [24,25] have adopted residual learning to gradually improve by learning the residual in each output, which can be achieved with only a few convolution parameters, thereby achieving high compactness and efficiency. A novel squeeze-and-excitation module (SENet) [26] was proposed. This attention mechanism focuses on each input channel, and the network focuses on the important channel after obtaining the weight of the corresponding channel, thus significantly improving the performance of the CNN. Wang [27] designed an efficient channel attention module to significantly improve model performance while using fewer parameters. Liu [28] proposed a pixel-level context attention network for selectively focusing on the context location information of pixels and generating an attention force to generate the context features of salient targets. Compared to SENet [26], which focuses only on the attention mechanism of the channel, Woo [29] conceived a lightweight convolutional attention module (CBAM). This module infers an attention map from the channel and spatial dimensions in turn and outputs refined features. Instead of simply using the residual network to perform feature extraction, we add the attention module based on this, which makes the network pay more attention to the detailed features of commodity regions, fully extracts the spatial information of multi-scale feature maps, and realizes the interaction between important features of cross-dimensional channels and spatial attention.
In the process of target prediction, small targets are easily ignored by the network because of their relatively few features [30]. The detection effect for small targets can be significantly improved by enhancing their features. However, the accuracy of target detection is unstable owing to uncertainty in the features of multiscale fusion. To efficiently express small target features, a new enhanced feature pyramid network (FPN) [31] was proposed, which can suppress redundant semantic information, ensure the enhancement of target features, and significantly improve the detection performance of objects. To improve the detection performance caused by weak features, neighborhood erasing and neighborhood transmission modules [32] were introduced to erase the salient features of large targets and emphasize the features of small and medium targets in shallow layers, respectively. Additionally, recognizing that boundary and texture features help to detect targets, researchers use boundary and texture enhancement networks [33] to embed feature information into object features to predict targets. Wang [34] proposed an “Attentive WaveBlock” module that can be embedded in dual networks to enhance the complementarity between the two parts and further suppress noise.
At present, object detection networks with deep learning as the mainstream are widely used in intelligent retail containers [35], but there is still a lot of room for improvement in the accuracy of commodity detection, especially in the detection of small commodities occluded by hands. There are still problems such as a low detection rate, false detection, and missed detection. Based on this, it is necessary to conduct in-depth and detailed research on the detection of the occlusion of small commodities. In this article, aiming at the detection of small commodities in a smart retail container, especially in the situation where customers’ hands occlude commodities during the purchase process, this paper proposes a feature enhancement occlusion detection algorithm for small commodities with SR. Since the video needs to be compressed and uploaded to the cloud server for corresponding small commodity detection to obtain high-definition images, it is processed with SR, and the corresponding feature expression ability is enhanced to effectively improve the detection performance of small commodities when the features of small commodities are occluded during the detection process.
In summary, this study makes the following three contributions:
(1)
During the experiments, we found that the image clarity of the video frames was low; therefore, we processed the images with SR and image super-reconstruction to recover clear images containing more detailed features of commodities;
(2)
To obtain more information about commodity features, a convolutional attention mechanism was used to guide the network to extract important features of the commodity while suppressing irrelevant features to fully extract the effective features of the commodity;
(3)
As small commodities have fewer features, extracting discriminative features is challenging. Therefore, we enhanced the contour and texture features of the commodity regions to ensure that the features of small commodities could be efficiently expressed and the detection accuracy could be improved.
The structure of this article is as follows: Section 2 describes the overall algorithm and related theories. Section 3 presents the experimental details, including the experimental platform, comparison experiment, ablation experiment, experimental results, and analysis. Section 4 summarizes the proposed algorithm.

2. General Architecture

Figure 1 shows the small commodity detection method employed in this study. The flowchart of the algorithm is divided into three primary steps: (1) preprocessing the input video frame to obtain the SR image; (2) extracting features from SR images and extracting feature information of different dimensions of commodities through the attention module; and (3) enhancing the small commodity area of the shallow feature map in the feature pyramid and classifying the commodities by adaptive regression through the fusion of the multi-scale feature maps.

2.1. Video Frame Preprocessing

SR Processing

Owing to the low number of pixels in an input video frame, the high-frequency detail information in the image may be substantially missing. This is not conducive to the small commodity detection task. Inspired by the FTT [13], the approach proposed in this study extracts the corresponding semantic features of images through content, texture, and contour extractors and thus improves the resolution of content and texture features to four times those of the original images. Simultaneously, it extracts the contour features of the features extracted from the content. Subsequently, it stitches the content, texture, and contour features together to obtain high-resolution features. Thus, it achieves the goal of using low-resolution images to output high-resolution images. The network structure is illustrated in Figure 2.
The structure consists of three parts corresponding to three functions, namely content extraction, texture extraction, and outline extraction. F 1 is the content input, F 2 is the texture input, and F 3 is the outline input. Herein, sub-pixel convolution was used to perform advanced spatial resolution processing on the input features. Subpixel convolution is a transformation for processing pixels in the channel dimension, whereby F 0 R H × W × C · l 2 is transformed into F 0 R H · l × W · l × C . F 1 extracts the image’s semantic features through the content extractor and converts the output multi-channel feature map into a single-channel feature map. F 2 initially maintains its resolution and then fuses it with F 1 for the feature map. The semantic features play an important role in SR image restoration; however, the generated video frames lack contour features. Therefore, in this study, an outline extractor was added to represent the contour information. F 3 was input into the outline extractor to obtain contour features, and the output multi-channel feature map was converted into a single-channel feature map. Then, we input the content feature map. Subsequently, the content feature map, F 2 , and contour feature map were input into the texture extractor for texture feature extraction. The generated feature map contains rich semantic information. Finally, it was stitched with the feature map output using the content and contour extractors to obtain the feature map F s containing the texture, content, and contour information. The expression is as follows:
F s = R T ( F 2 ( R C ( F 1 ) 2 × ) R O ( F 3 ) 2 × + R C ( F 1 ) 2 × + R O ( F 3 ) 2 × )
where R C ( · ) is the content extraction module; R O ( · ) is the outline extraction module; and R T ( · ) is the texture extraction module.
The original image and detailed feature map were fused, and F l o w represents the source image features, which were merged with F s through the fusion layer and described as follows:
F f u s i o n = H C o n c a t ( F l o w + F s )
where H C o n c a t denotes the fusion operation. The fusion layer is essentially a bottleneck layer for providing feature fusion and increasing the nonlinear relationship between the high- and low-resolution features. Subsequently, the merged image features are input into the image reconstruction network for high-quality SR image reconstruction [8]. The formula is expressed as follows:
I S R = F I R N ( F f u s i o n )
where F I R N stands for image reshaping operation and I S R represents the reconstructed image.

2.2. Feature Extraction

In the feature extraction process, a residual network is used for feature extraction. To fully extract the features of commodity regions, an attention mechanism is added to the network to focus on commodity regions and extract fine commodity feature information.

2.2.1. Residual Network

In this study, a residual network comprising multiple residual units stacked together was used. Owing to the general lack of small commodity features, the proposed approach uses residual dense blocks to connect and supplement local features when extracting features and reduces the number of network parameters through parameter sharing. The network adds a skip connection between each residual unit and the next one and fuses the output features of the different residual units. The skip connection in the residual block helps maintain the norm of the gradient and ensures stable backpropagation.
Essentially, the feature map F W × H × C · l 2 is sent to the three channels of the residual network for effective feature extraction. The proposed network was divided into two parts: the residual backbone network and the attention mechanism module [29]. In the backbone network, a 3 × 3 convolution kernel was used for feature extraction, and a F 1 , , F i , , F n feature map was obtained. The map can be expressed as follows:
F i = H C o n v ( F i 1 )
where H C o n v ( · ) includes the convolution layer, batch normalization (BN) layer, and rectified linear unit (ReLU) function.
The residual dense blocks are fused in each branch to obtain a dense feature map. The corresponding equation is as follows:
F k + 1 , C t = H C o n v , 1 × 1 ( H C o n c a t ( F k + 1 , C 1 , , F k + 1 , C t 1 ) )
where H C o n c a t ( · ) represents the feature fusion and H C o n v , 1 × 1 ( · ) is the convolutional layer, BN layer, and a Relu nonlinear layer.
The output feature map F k + 2 was obtained by adding the input feature map F k and dense feature map F k + 1 [36]. The formula is as follows:
F k + 2 = F k + F k + 1
The structure of the residual network is shown in Figure 3.

2.2.2. Attention Mechanism Module

To address the problem of insufficient utilization of features in the middle of the network, CBAM was introduced in the middle of the residual network [29] to enhance its representational ability. To avoid the loss of salient features of small items in the extraction process, an attention mechanism based on both channel and spatial attention was used. A convolution operation was employed to mix the cross-channel and spatial information and extract the important feature information of the small commodities. The structure of the attention mechanism is shown in Figure 4.
The input feature map F R C × H × W , one-dimensional channel attention map F R C × 1 × 1 , and two-dimensional spatial attention map M s R 1 × H × W are described as follows:
F = M c ( F ) F
F = M s ( F ) F
where is the pixel-by-pixel multiplication and F is the refined feature of the output.

2.3. Feature Fusion and Detection Frame Regression

2.3.1. Feature Pyramid Network

Existing object detectors have achieved good results for large objects; however, their performance for small objects is unsatisfactory. In this study, to detect smaller commodities, an image FPN was constructed to realize detection across the scale range. In particular, a lightweight architecture that efficiently generates image feature pyramids in the detection framework was used. The structure of the FPN is shown in Figure 5.
The extracted features were sampled to obtain multi-scale feature maps, namely, G 1 , G 2 , G 3 , G 4 . The feature maps of different scales were upsampled and fused to obtain S 1 , S 2 , S 3 , S 4 . This approach can fully utilize different context regions to obtain global information, including high-level semantic and shallow location information. The region proposal network adaptively generates proposal regions and sends them to the subsequent network.

2.3.2. Feature Enhancement

Small commodities contain less information in the feature maps, and such information can easily be ignored. To efficiently detect small commodities, the salient features of commodities are emphasized and expressed, which is helpful in achieving the rapid detection of commodities. By improving the neighborhood transmission module [32], a feature enhancement network was designed herein. Compared to the deep feature map, the shallow feature map contains richer information regarding the locations, textures, and outlines of commodities. Therefore, the shallow feature map from the FPN was input into the feature enhancement module to enhance the location, contour, and texture information along with other features of small items to improve the detection accuracy and speed performance. The feature enhancement network structure is shown in Figure 6.
The feature enhancement module was used to enhance the features of small commodities, that is, to enhance the features of the shallow feature maps S 1 and S 2 . First, S 1 was upsampled. Subsequently, spatial channels were generated using a gate function, and a feature map S 1 was obtained based on the activation function. The input S 2 underwent convolution by a 1 × 1 convolution kernel, and S 2 was obtained by a gate function operation. By multiplying the features of S 1 and S 2 , calculations were obtained as follows:
S 1 = σ ( G ( U ( S 1 ) ) ) = 1 1 + e G ( U ( S 1 ) )
S 2 = G ( H C o n v , 1 × 1 ( S 2 ) )
where σ ( · ) is the activation function; U ( · ) is the upsampling operation on the feature map; and G ( · ) is the self-attention gate function, which generates a spatial channel to enhance commodity features. The formula is as follows:
G ( S i ) = H C o n v ( S i )
The combination of the two results in S p , which can be expressed as follows:
S p = S 1 S 2
where denotes the Hadamard product. These features were summed element-by-element to obtain the details S k ¯ , as follows:
S k ¯ = S 1 H C o n v , 1 × 1 ( D ( S p ) )
where stands for a pixel-by-pixel summation; D ( · ) is the downsampling operation on the feature map; and the detailed features of the commodity area are enhanced to facilitate subsequent classification and detection box regression. Thus, enhanced feature maps S k ¯ and high-level feature maps S 1 , S 3 , and S 4 , containing the location information, contour information and center point of the commodity, were obtained. These can be used to effectively predict different scales and to generate subsequent product detection frames.

2.3.3. Commodity Detection Frame

As our task was to generate a commodity detection box, a region proposal network was introduced for commodity region regression. In the training phase, 10,000 regression boxes with the highest scores were obtained through a non-maximum suppression operation, and 1500 of them were selected as small-item proposals. In the test phase, 400 proposals were selected from 10,000 regression frames. Owing to the occlusion of small commodities and relatively few features, detection in regression detection frames can be easily missed. Therefore, inspired by a previous study [37], a new loss function was proposed to train the network. The loss function in this article consists of three parts. The first part is the regression loss function, which has a great influence on the regression of the detection box owing to the variety of shapes of the commodities. To solve this problem, the intersection-over-union (IOU) factor log ( I O U ) L 1 ( v k j , v k j ) was introduced to optimize positioning accuracy and accurately return the detection box for commodities. The formula is as follows:
L 1 ( x ) = 0.5 x 2 , x < 1 x 0.5 , o t h e r w i s e
L r e g = 1 N i = 1 N j x , y , w , h L 1 ( v k j , v k j ) L 1 ( v k j , v k j ) log ( I O U )
where L 1 is the smoothing loss; N is the number of regression frames; and I O U represents the overlap between prediction frames and real frames.
By regressing the size of the target commodity c, the regressed commodity width and height were obtained as S c = ( x 2 ( c ) x 1 ( c ) , y 2 ( c ) y 1 ( c ) ) . The second part is a loss function for the commodity width and height. The loss function was used to measure the losses of commodity width and height. The formula is described as follows:
L h , w = 1 N 2 i = 1 N S r e a l S c 2
S r e a l is the true width and height of the commodity.
The third part is the classification loss function, which comprises the cross-entropy function, and is expressed as follows:
L c l s = 1 N i = 1 N log e W C i T x i + b C i i = 1 N e W j T x i + b j
where W C i T is the learned weight; and b j is the bias term.
The formula for the total loss function is defined as follows:
L t o t a l = λ 1 L r e g + λ 2 L h , w + λ 3 L c l s
Among them, the distribution of the super-parameters, λ 1 = 1 3 , λ 2 = 1 3 , λ 3 = 1 3 controls the weight of each loss function. Through the constraint of the loss function, an accurate detection frame was derived, thereby completing the detection task for small commodities.

3. Experiments

The specific structure of the experimental content in this study is shown in Figure 7 and is mainly divided into five parts: experimental setting, algorithm evaluation, comparison study, ablation study, and experimental analysis.

3.1. Experiment Setting

3.1.1. Dataset

In this study, self-made retail containers were used to collect 16 commodity datasets, including training, validation, and test sets. Herein, the definition of the small commodity is the size of the small commodity, which is related to the size of a human hand. In particular, under extreme conditions, whether the entire hand of a consumer can fully cover the features of the effective area of the commodity to the greatest extent such that the commodity cannot be detected was determined. Commodities meeting this condition were considered small commodities.
To facilitate the subsequent commodity inspection, the names of the commodities were simplified, as shown in Table 1.
To illustrate the feasibility of the experimental data, commodity datasets were collected under appropriate lighting conditions. The datasets for each commodity, including both large and small commodities, are shown in Figure 8 and Figure 9.

3.1.2. Experimental Platform

In this work, the system platform was Windows 10, the GPU model was an NVIDIA GeForce RTX 3060, the CPU was an I5-12400F, the memory was 16 GB, and the software environment was Python3.7 and Pytorch2.3.

3.2. Algorithm Evaluation

3.2.1. Index of SR Algorithm

To illustrate the processing results from the SR algorithm, two quantitative indicators, namely the peak signal-to-noise ratio (PSNR) and structural similarity measure (SSIM), were introduced. The PSNR formula is expressed as follows:
P S N R = 20 × log 10 ( M A X I M S E )
where M A X I represents the maximum pixel value in the image pixels; and M S E represents the mean square error of the corresponding pixels between the generated image f i j and original image f i j . M S E is calculated as follows:
M S E = 1 M N i = 1 N j = 1 M ( f i j f i j ) 2
SSIM is a measure of the similarity between two images, and it is calculated as follows:
S S I M ( x , y ) = 2 μ X μ Y + c 1 μ X 2 + μ Y 2 + c 1 2 σ X σ Y + c 2 σ X 2 + σ Y 2 + c 2
where, μ X and μ Y are the pixel mean of image X and image Y, respectively; σ X and σ Y are the pixel variances of image X and image Y, respectively; c 1 = ( 0.01 l ) 2 ; and c 2 = ( 0.03 l ) 2 .
Note that the higher the PSNR, the less distorted the processed image is. A higher SSIM indicates higher image similarity and better image quality.

3.2.2. Index of Detection Algorithm

To evaluate the proposed algorithm, the average precision (AP) and mean AP (mAP) were selected as evaluation indicators.
A c c u r a c y = T P + T N T P + T N + F P + F N
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 s c o r e = 2 P r e c i s e · R e c a l l P r e c i s e + R e c a l l
A P i = 1 i P · d r
m A P = i N A P i N
where TP denotes a positive sample and a positive prediction result; TN denotes a positive sample and a negative prediction result; FP denotes a negative sample and a positive prediction result; and FN denotes a negative sample and negative prediction result; Accuracy represents the proportion of all correct predictions. Precision represents the percentage of true positive predictions; Recall denotes the proportion of true positives; AP is the area of the Precision-Recall curve; and mAP is the mean average accuracy across all classes.

3.3. Comparative Study

The algorithm proposed herein has good completeness. The relevant parameters were set to achieve high-performance small commodity detection. The total training batch was 100, the epoch was 80, and the learning rate was set to 0.0001. The loss function curve is shown in Figure 10.
As can be seen from the figure, when the training reached approximately 80 iterations, the loss function converged.

3.3.1. Effect of SR

To demonstrate the effectiveness of the SR method, it was compared to the SRGAN [38], EDSR [39], and CARN [40] algorithms in the context of image blurring during the detection process. The abovementioned experiments revealed that the image details processed by the proposed method were richer, and the generated image had a high degree of similarity to the original image. Table 2 lists the experimental results obtained by the different methods.
Compared to the other methods, the method proposed herein was superior in terms of index performance. The PSNR and SSIM values were the highest. The SR images significantly restored the details of the original image, and the original image features were retained to the greatest extent, thus demonstrating the feasibility and superiority of this algorithm.
The experimental results are shown in Figure 11, with original images (a) and (c) and SR images (b) and (d) obtained by the proposed method.
Evidently, from the above figure, the SR processing retained the texture and structure information of the original image and enhanced the high-frequency information, such as the commodity contents and contours in the image. The semantic information of commodities was restored, thereby avoiding the problem of image distortion (which led to a decline in detection accuracy).
The SR processing not only retains the original image but also restores the commodity contour and other features to a certain extent, thereby playing a positive role in the subsequent commodity contour feature extraction.

3.3.2. Small Commodity Detection Performance

When consumers buy commodities, the camera captures different degrees of occlusion of the commodities from its perspective, which can considerably affect the effects of small-object detection. To verify the efficiency of this algorithm, the model was compared with four other algorithms, and four types of commodities with different degrees of occlusion were selected. These included slight occlusions (occlusion degrees of 0–10%), partial occlusions (occlusion degrees of 10–20%), moderate occlusions (occlusion degrees of 20–30%), and severe occlusions (occlusion degrees of 30–60%). The experimental results are listed in Table 3.
According to the above table, the proposed method was superior to the SSD, Faster-RCNN, Yolov5, and RetinaNet algorithms in terms of detection accuracy on commodity datasets with different degrees of occlusion. The horizontal comparison indicates that the proposed method performed well in terms of detection accuracy for different commodities. Under severe occlusion, the detection accuracy of the proposed algorithm was improved by more than 3% relative to the mainstream methods Yolov5 and RetinaNet. The longitudinal comparison indicated that with an increase in the degree of occlusion, detection accuracy exhibited a gradual downward trend. Compared with other algorithms, the detection accuracy of the proposed method generally remained above 81%, and the proposed model was relatively stable compared with the other models. In addition, the results demonstrated the superiority of the proposed algorithm. In terms of detection speed, the proposed model reached an average of 15.13 frames/s, thus meeting the requirements for real-time performance.

3.3.3. Comparison of Model Complexity and Computation Time

To evaluate the computational complexity of each model, relevant comparison experiments were performed in terms of the number of model parameters and training time, and the results are shown in Table 4. The time indicates the training time. As can be seen from the table, the Faster R-CNN model has the largest number of parameters, while the RetinaNet model has a relatively small number of parameters. Compared with other models, the number of parameters and training time of this paper need to be further reduced, and the complexity of the model needs to be optimized to meet the commercialization requirements.

3.4. Ablation Study

3.4.1. Effect of SR

The comparative experiments revealed that the quality of the image generated by the SR algorithm was high and that the contours of the commodities and other information were significantly restored. To explore whether SR commodity detection was efficient, ablation experiments were performed under two conditions: (1) lack of SR commodity detection and (2) commodity detection under SR. The experimental results are listed in Table 5.
The image detection effect after SR processing was significantly higher than that of the original image. In terms of detection accuracy, the performance with SR processing was better than that without SR processing, with an increase of more than 9%. The SR-processed image contour feature information was more abundant, the network could further extract the semantic details of small commodities, and the small commodity detection accuracy was significantly improved, further verifying the effectiveness of the SR algorithm.

3.4.2. Effect of Attention Mechanism Module

Insufficient information extraction from small commodities can easily occur in the feature extraction process. This study focuses on commodity feature information using an attention mechanism. To verify the effectiveness of the method, experiments on feature extraction with an attention mechanism were conducted. The results of the commodity detection are shown in Figure 12.
According to the above experimental results, the detection accuracy of the right figure was significantly improved compared to that of the left figure, which indicates that the small commodity detection effect was significantly improved under the effect of the attention mechanism.
To understand which parts of the network were focused on based on the attention mechanism, the feature map of the feature extraction part could be visualized through heatmaps. The size of the output feature maps was set to 600 × 600 in this experiment, as shown in Figure 13.
The left and right panels, respectively, show heatmaps without and with the attention mechanism. Evidently, under the effect of the attention mechanism, the network focus area is significantly reduced, and the network extracts more refined commodity regional features from the channel and spatial dimensions, improving the efficiency of commodity feature extraction to a certain extent. In addition, it further verifies the feasibility and scientificity of using the attention mechanism in the network.

3.4.3. Effect of Feature Enhancement

To evaluate the effect of the feature enhancement module on the detection of occluded small commodities, experiments were conducted using both an ordinary FPN and an FPN with a feature enhancement module (FPN + FEM). The experimental results are shown in Figure 14.
The commodity detection accuracy after adding the feature enhancement could reach approximately 96%, and the highest accuracy reached was 98%. Compared with direct prediction, it had a higher accuracy. Simultaneously, for the small bottles of Sprite with fewer effective features, the accuracy was increased by 32.4% relative to the original, and the detection effect was more significant.
Figure 15 shows the results for the commodity feature region after feature enhancement. According to the figure, the important features of the commodity were enhanced, and the detection performance of the network was further improved, thus verifying the effectiveness of the feature enhancement module proposed in this study.

3.5. Experiment Analysis

3.5.1. Qualitative Analysis

In this study, different types of small commodities with different degrees of occlusion were selected to analyze the proposed algorithm, and the results are shown in Figure 16.
The results revealed excellent performance in the detection of small commodities of different types and occlusion levels. As shown in Figure 17, the detection performance of the network was stable under different occlusion levels, and the detection accuracy of the commodities still reached as high as 80% in the case of severe occlusion. This method increased the high-frequency information of the image through SR; simultaneously, the feature enhancement module in the FPN could effectively improve the feature expression of small commodities, and therefore, the detection effect was significantly improved. The detection results qualitatively illustrate the feasibility and efficiency of the algorithm used in this study.

3.5.2. Quantitative Analysis

To illustrate the commodity detection performance of the proposed algorithm, a quantitative comparison was performed with five mainstream networks, and Table 6 lists the commodity detection results of the different algorithms.
According to the above table, the different algorithms achieved good detection performance. The algorithm in this study achieved excellent performance in terms of the F1-score and mAP indicators, with an F1-score of 0.983 and an mAP of 0.9847, which are superior to the results for other algorithms. The algorithm network in this study comprised Resent50 + CBAM + FPN (FAM). Compared to RetinaNet, the F1-score improved by 2.6 % and the mAP improved by 2.45%.
To illustrate the detection accuracy of the algorithm, different lightly occluded commodity data were selected for comparative experiments. The results are presented in Table 7.
The detection accuracies for different commodity categories varied considerably. Because the features of Red Bull were more significant, each network had high detection accuracy. Compared with YOLOv5, the overall accuracy of this method was higher. The proposed method had a detection effect on different small commodities. The detection performance of the model was more stable than that of the other methods while maintaining accuracy.

4. Conclusions

In this article, a local adaptive feature enhancement detection algorithm for occluded small commodities under super-resolution is proposed. Based on the low image clarity, a new SR algorithm is designed that effectively improves the image clarity by adding contour features to the feature texture transmission module, fusing them with texture and content features, obtaining rich fine features, and obtaining high-frequency image information through reconstruction. To effectively express small commodities occluded in complex environments, a self-attention gate function is used to generate commodity space channels, enhance commodity texture features, and other characteristics, and further improve the detection accuracy of small commodities. Experimental results show that the proposed algorithm has good detection accuracy and can effectively reduce the false or missed detections caused by complex occlusion. However, the method in this article pursues commodity detection accuracy and ignores the light weight of the model, which considerably limits the detection speed of the model. In the future, the network model will be further explored and optimized to reduce the number of model parameters and achieve real-time detection of small commodities.

Author Contributions

Conceptualization, H.D. and K.X.; Methodology, H.D.; Software, A.X.; Writing—original draft preparation, H.D.; Writing—review and editing, H.D.; Visualization, C.W. and J.H.; Investigation, W.Z., D.Y. and S.Y.; Funding acquisition, J.H.; Project administration, H.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported in part by the National Natural Science Foundation of China (62272485). The project conducted work on Intelligent Retail Containers, and the research was conducted in cooperation with it, in part by the Natural Science Foundation of Xinjiang Uygur Autonomous Region (Grant No. 2020DO1A131). The research was sponsored by the project and in part by the Teaching and Research Fund of Yangtze University (Grant No. JY2020101). The research was sponsored and conducted under the project, and was mainly conducted under the Undergraduate Training Programs for Innovation and Entrepreneurship of Yangtze University under Grant Yz2022056. The project was about commodity detection under Intelligent Retail Containers, and the research work in the article was carried out under the project.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data sharing is not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. He, Z.; Zhang, L. Multi-adversarial faster-rcnn for unrestricted object detection. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October 2019–2 November 2019; pp. 6668–6677. Available online: https://ieeexplore.ieee.org/document/9010003 (accessed on 3 October 2022).
  2. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the 2016 European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 11–14 October 2016; Springer: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar] [CrossRef] [Green Version]
  3. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. Available online: https://ieeexplore.ieee.org/document/7780460 (accessed on 5 October 2022).
  4. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollar, P. Focal Loss for dense object detection. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 318–327. [Google Scholar] [CrossRef] [Green Version]
  5. Zhang, H.; Li, D.; Ji, Y.; Zhou, H.; Wu, W.; Liu, K. Toward new retail: A benchmark dataset for smart unmanned vending machines. IEEE Trans. Ind. Inform. 2020, 16, 7722–7731. [Google Scholar] [CrossRef]
  6. Zou, Z.; Shi, Z.; Guo, Y.; Ye, J. Object detection in 20 years: A survey. arXiv 2019, arXiv:1905.05055. [Google Scholar] [CrossRef]
  7. Zou, X.F.; Zhou, L.Q.; Li, K.L.; Ouyang, A.J.; Chen, C. Multi-task cascade deep convolutional neural network for large-scale commodity recognition. Neural Comput. Appl. 2020, 32, 5633–5647. [Google Scholar] [CrossRef]
  8. Fang, F.; Li, J.; Zeng, T. Soft-edge assisted network for single image super-resolution. IEEE Trans. Image Process. 2020, 29, 4656–4668. [Google Scholar] [CrossRef]
  9. Zhang, Y.; Fan, Q.; Bao, F.; Liu, Y.; Zhang, C. Single-image super-resolution based on rational fractal interpolation. IEEE Trans. Image Process. 2018, 27, 3782–3797. [Google Scholar]
  10. Li, B.; Wang, B.; Liu, J.; Qi, Z.; Shi, Y. s-lwsr: Super lightweight super-resolution network. IEEE Trans. Image Process. 2020, 29, 8368–8380. [Google Scholar] [CrossRef]
  11. Ma, C.; Rao, Y.; Cheng, Y.; Chen, C.; Lu, J.; Zhou, J. Structure-preserving super resolution with gradient guidance. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 7769–7778. Available online: https://ieeexplore.ieee.org/document/9156994 (accessed on 7 September 2022).
  12. Ma, C.; Rao, Y.; Lu, J.; Zhou, J. Structure-preserving image super-resolution. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 7898–7911. [Google Scholar] [CrossRef]
  13. Deng, C.; Wang, M.; Liu, L.; Liu, Y.; Jiang, Y. Extended feature pyramid network for small object Detection. IEEE Trans. Multimed. 2021, 24, 1968–1979. [Google Scholar] [CrossRef]
  14. Cai, Q.; Li, J.; Li, H.; Yang, Y.; Wu, F.; Zhang, D. TDPN: Texture and detail-preserving network for single image super-resolution. IEEE Trans. Image Process. 2022, 31, 2375–2389. [Google Scholar] [CrossRef]
  15. Huang, Y.X.; Zhang, X.Y.; Fu, Y.; Chen, S.H.; Zhang, Y.; Wang, Y.F.; He, D.Z. Task decoupled framework for reference-based super-resolution. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 5921–5930. Available online: https://ieeexplore.ieee.org/document/9879135 (accessed on 17 September 2022).
  16. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [Green Version]
  17. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. Available online: https://ieeexplore.ieee.org/document/7780459 (accessed on 8 November 2022).
  18. Jian, M.; Jin, H.; Liu, X.; Zhang, L. Multiscale Cascaded Attention Network for Saliency Detection Based on ResNet. Sensors 2022, 22, 9950. [Google Scholar] [CrossRef]
  19. Xia, R.L.; Chen, Y.T.; Ren, B.B. Improved anti-occlusion object tracking algorithm using unscented rauch-tung-striebel smoother and kernel correlation filter. J. King Saud Univ.-Comput. Inf. Sci. 2022, 34, 6008–6018. [Google Scholar] [CrossRef]
  20. Chen, Y.T.; Xia, R.L.; Zou, K.; Yang, K. FFTI: Image inpainting algorithm via features fusion and two-steps inpainting. J. Vis. Commun. Image Represent. 2023, 91, 103776. [Google Scholar] [CrossRef]
  21. Zhang, K.; Sun, M.; Han, T.X.; Yuan, X.; Guo, L.; Liu, T. Residual networks of residual networks: Multilevel residual networks. IEEE Trans. Circuits Syst. Video Technol. 2017, 28, 1303–1314. [Google Scholar] [CrossRef] [Green Version]
  22. Zhang, Y.; Tian, Y.; Kong, Y.; Zhong, B.; Fu, Y. Residual dense network for image restoration. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 2480–2495. [Google Scholar] [CrossRef] [Green Version]
  23. Zaeemzadeh, A.; Rahnavard, N.; Shah, M. Norm-preservation: Why residual networks can become extremely deep? IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 3980–3990. [Google Scholar] [CrossRef]
  24. Chen, S.; Tan, X.; Wang, B.; Lu, H.; Hu, X.; Fu, Y. Reverse attention-based residual network for salient object detection. IEEE Trans. Image Process. 2020, 29, 3763–3776. [Google Scholar] [CrossRef]
  25. Feng, M.; Lu, H.; Yu, Y. Residual learning for salient object detection. IEEE Trans. Image Process. 2020, 29, 4696–4708. [Google Scholar] [CrossRef]
  26. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-excitation networks. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 2011–2023. [Google Scholar] [CrossRef] [Green Version]
  27. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient channel attention for deep convolutional neural networks. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 11531–11539. Available online: https://ieeexplore.ieee.org/document/9156697 (accessed on 26 November 2022).
  28. Liu, N.; Han, J.; Yang, M.H. PiCANet: Pixel-wise contextual attention learning for accurate saliency detection. IEEE Trans. Image Process. 2020, 29, 6438–6451. [Google Scholar] [CrossRef] [Green Version]
  29. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the 2018 European Conference on Computer Vision (ECCV); Springer: Cham, Switzerland, 2018; Volume 11211, pp. 3–19. Available online: https://link.springer.com/chapter/10.1007/978-3-030-01234-2_1 (accessed on 24 October 2022).
  30. Zhong, Y.; Wang, Y.; Zhang, S. Progressive feature enhancement for person re-identification. IEEE Trans. Image Process. 2021, 30, 8384–8395. [Google Scholar] [CrossRef]
  31. Jin, Z.; Liu, B.; Chu, Q.; Yu, N. SAFNet: A semi-anchor-free network with enhanced feature pyramid for object detection. IEEE Trans. Image Process. 2020, 29, 9445–9457. [Google Scholar] [CrossRef]
  32. Li, Y.; Pang, Y.; Cao, J.; Shen, J.; Shao, L. Improving single shot object detection with feature scale unmixing. IEEE Trans. Image Process. 2021, 30, 2708–2721. [Google Scholar] [CrossRef]
  33. Li, P.; Yan, X.F.; Zhu, H.W.; Wei, M.Q.; Zhang, X.P.; Qin, J. Findnet: Can you find me? boundary-and-texture enhancement network for camouflaged object detection. IEEE Trans. Image Process. 2022, 31, 6396–6411. [Google Scholar] [CrossRef]
  34. Wang, W.; Zhao, F.; Liao, S.; Shao, L. Attentive waveblock: Complementarity-enhanced mutual networks for unsupervised domain adaptation in person re-identification and beyond. IEEE Trans. Image Process. 2022, 31, 1532–1544. [Google Scholar] [CrossRef]
  35. Zhu, P.F.; Sun, Y.M.; Cao, B.; Liu, X.Y.; Liu, X.; Hu, Q.H. Self-supervised fully automatic learning machine for intelligent retail container. IEEE Trans. Instrum. Meas. 2022, 71, 1–15. [Google Scholar] [CrossRef]
  36. Zhang, J.; Hu, H.; Feng, S. Robust facial landmark detection via heatmap-offset regression. IEEE Trans. Image Process. 2020, 29, 5050–5064. [Google Scholar] [CrossRef]
  37. Yang, X.; Yang, J.; Yan, J.; Zhang, T.; Guo, Z.; Sun, X.; Fu, K. Scrdet: Towards more robust detection for small, cluttered and rotated objects. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October 2019–2 November 2019; pp. 8232–8241. Available online: https://ieeexplore.ieee.org/document/9008772 (accessed on 8 December 2022).
  38. Ledig, C.; Theis, L.; Huszar, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-realistic single image super-resolution using a generative adversarial network. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4681–4690. Available online: https://ieeexplore.ieee.org/document/8099502 (accessed on 1 January 2023).
  39. Lim, B.; Son, S.; Kim, H.; Nah, S.; Mu Lee, K. Enhanced deep residual networks for single image super-resolution. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; pp. 136–144. Available online: https://ieeexplore.ieee.org/document/8014885 (accessed on 28 November 2022).
  40. Ahn, N.; Kang, B.; Sohn, K.A. Fast, accurate, and lightweight super-resolution with cascading residual network. In Proceedings of the 2018 European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; Springer: Cham, Switzerland, 2018; Volume 11214, pp. 252–268. [Google Scholar] [CrossRef] [Green Version]
  41. Ge, Z.; Liu, S.T.; Wang, F.; Li, Z.M.; Sun, J. Yolox: Exceeding yolo series in 2021. arXiv 2021, arXiv:210.08430. [Google Scholar]
  42. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End to end object detection with transformers. In Proceedings of the 2020 European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; Springer: Cham, Switzerland, 2020; Volume 12346, pp. 213–229. [Google Scholar] [CrossRef]
Figure 1. Algorithm flow. (A) The input video frames are first preprocessed to obtain the SR images. (B) Residual dense blocks are used to extract features from SR images and extract feature information for different dimensions of commodities through the CBAM. (C) The feature pyramid network enhances the small commodity region of the shallow feature map, and the fusion of multi-scale feature maps is used to classify the commodities by adaptive regression. Downsample denotes the downsampling operation on the feature map. Upsample denotes the upsampling operation on the feature map.
Figure 1. Algorithm flow. (A) The input video frames are first preprocessed to obtain the SR images. (B) Residual dense blocks are used to extract features from SR images and extract feature information for different dimensions of commodities through the CBAM. (C) The feature pyramid network enhances the small commodity region of the shallow feature map, and the fusion of multi-scale feature maps is used to classify the commodities by adaptive regression. Downsample denotes the downsampling operation on the feature map. Upsample denotes the upsampling operation on the feature map.
Sensors 23 02439 g001
Figure 2. The architecture of image super-resolution (SR) network. F 1 denotes the content input, F 2 denotes the texture input, and F 3 denotes the outline input.
Figure 2. The architecture of image super-resolution (SR) network. F 1 denotes the content input, F 2 denotes the texture input, and F 3 denotes the outline input.
Sensors 23 02439 g002
Figure 3. The architecture of the residual network.
Figure 3. The architecture of the residual network.
Sensors 23 02439 g003
Figure 4. The architecture of the attention mechanism module.
Figure 4. The architecture of the attention mechanism module.
Sensors 23 02439 g004
Figure 5. The architecture of the feature pyramid network. The FPN is used to produce multi-scale feature representations.
Figure 5. The architecture of the feature pyramid network. The FPN is used to produce multi-scale feature representations.
Sensors 23 02439 g005
Figure 6. The architecture of the feature enhancement network. C 1 is the 1 × 1 convolution kernel.
Figure 6. The architecture of the feature enhancement network. C 1 is the 1 × 1 convolution kernel.
Sensors 23 02439 g006
Figure 7. Overview of experiment arrangement.
Figure 7. Overview of experiment arrangement.
Sensors 23 02439 g007
Figure 8. Sample datasets of each commodity.
Figure 8. Sample datasets of each commodity.
Sensors 23 02439 g008aSensors 23 02439 g008b
Figure 9. Quantity distribution of each commodity.
Figure 9. Quantity distribution of each commodity.
Sensors 23 02439 g009
Figure 10. Loss function curve.
Figure 10. Loss function curve.
Sensors 23 02439 g010
Figure 11. Comparison of original images and SR-processed images.
Figure 11. Comparison of original images and SR-processed images.
Sensors 23 02439 g011
Figure 12. Detection results with and without the attention mechanism.
Figure 12. Detection results with and without the attention mechanism.
Sensors 23 02439 g012
Figure 13. Heatmaps with and without the attentional mechanism.
Figure 13. Heatmaps with and without the attentional mechanism.
Sensors 23 02439 g013
Figure 14. Ablation experiment of detection results with and without the feature enhancement.
Figure 14. Ablation experiment of detection results with and without the feature enhancement.
Sensors 23 02439 g014
Figure 15. Feature-enhanced visualization.
Figure 15. Feature-enhanced visualization.
Sensors 23 02439 g015
Figure 16. Detection results for different types of small commodities.
Figure 16. Detection results for different types of small commodities.
Sensors 23 02439 g016
Figure 17. Detection results for small commodities with different degrees of occlusion.
Figure 17. Detection results for small commodities with different degrees of occlusion.
Sensors 23 02439 g017
Table 1. The simplified table of commodity names.
Table 1. The simplified table of commodity names.
Name of CommoditySimplification of Commodity Name
Canned Pepsibskl_gz
Pepsi Colabskl
Rainbow Popcornchbmh
Red Bullhn
Fiber Drinkjj
Kumquat Lemonjjnm
Small bottle of Coca Colakkkl_s
Green Tealc
Mirindamnd
Canned Mirindamnd_gz
Mai Xiang Chicken Flavor Blockmxjwk
Nongfu Springnfsq
Wang Zai Milkwz
Small bottle of Spritexb_s
C’estbonyb_m
Small bottle of C’estbonyb_s
Table 2. Super-resolution results of different algorithms.
Table 2. Super-resolution results of different algorithms.
MethodSRGANEDSRCARNOurs
PSNR/dB28.1630.4632.1232.72
SSIM0.8890.8870.8580.894
Note: Bold is the best result.
Table 3. Comparison of detection performance of different algorithm models.
Table 3. Comparison of detection performance of different algorithm models.
Occlusion Degree of Different CommoditySSDFaster-RCNNYOLOv5RetinaNetOurs
Red Bull (slight occlusion)0.96670.96930.96820.98830.9891
Red Bull (partial occlusion)0.94300.94510.96710.95540.9557
Red Bull (moderate occlusion)0.87470.83660.89640.91890.9273
Red Bull (heavy occlusion)0.67460.75290.81530.84820.8828
C’estbon (slight occlusion)0.67580.98330.96340.98390.9857
C’estbon (partial occlusion)0.56930.94080.95670.96730.9724
C’estbon (moderate occlusion)0.53510.80210.84670.80170.9152
C’estbon (heavy occlusion)0.43320.74750.76760.76310.8281
Sprite (slight occlusion)0.88070.88630.97570.96030.9787
Sprite (partial occlusion)0.83560.85250.93630.85920.9576
Sprite (moderate occlusion)0.72210.78890.84850.79750.8972
Sprite (heavy occlusion)0.45330.72180.72270.79190.8483
Wang Zai Milk (slight occlusion)0.75070.91630.95740.97520.9793
Wang Zai Milk (partial occlusion)0.74360.80110.93690.92390.9679
Wang Zai Milk (moderate occlusion)0.64920.61250.74950.87020.9362
Wang Zai Milk (heavy occlusion)0.47500.53820.67560.76470.81251
Note: Bold is the best result. 1 All results are the AP.
Table 4. Comparison of model complexity.
Table 4. Comparison of model complexity.
ModelBackboneParameters (M)Time (min)
YOLOv4CSPDarknet5342.3977
YOLOv5CSPDarknet5338.4854
SSDVGG16139.73063
Faster R-CNNVGG16148.43368
RetinaNetResnet5027.5617
OursResnet5041.1918
Table 5. Detection accuracy of original image and super-resolution.
Table 5. Detection accuracy of original image and super-resolution.
MethodThe Lack of SR Commodity DetectionCommodity Detection under SR
Mai Xiang Chicken Flavor Block0.79270.8938
Small bottle of C’estbon0.80830.9673
Wang Zai Milk0.86920.9651
Rainbow Popcorn0.77430.9081
Small bottle of Coca Cola0.73570.8635 1
1 All results are the AP.
Table 6. Comparison of the accuracy of commodity detection with different algorithms.
Table 6. Comparison of the accuracy of commodity detection with different algorithms.
MethodBackboneF1-ScoremAP
YOLOv4CSPDarknet530.9320.9544
YOLOv5CSPDarknet530.9720.9740
SSDVGG0.9790.9784
Faster R-CNNVGG0.9520.9768
RetinaNetResnet500.9570.9602
YOLOX [41]CSPDarknet530.9740.9831
DETR [42]Resnet500.9800.9753
OursResnet500.9830.9847
Note: Bold is the best result.
Table 7. The detection results of different commodity types.
Table 7. The detection results of different commodity types.
CategorySSDFaster R-CNNRetinaNetYOLOv5Ours
Coca Cola0.44090.55730.55310.88650.8537
Wang Zai milk0.68070.91560.90150.93320.9671
Small bottle of Sprite0.87840.94870.94950.88560.9635
Small bottle of C’estbon0..69830.94350.96470.95600.9676
Red Bull0.95340.96430.97790.97830.9883
Canned Mirinda0.76060.93110.93530.92400.9512 1
Note: Bold is the best result. 1 All results are the AP.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Dong, H.; Xie, K.; Xie, A.; Wen, C.; He, J.; Zhang, W.; Yi, D.; Yang, S. Detection of Occluded Small Commodities Based on Feature Enhancement under Super-Resolution. Sensors 2023, 23, 2439. https://doi.org/10.3390/s23052439

AMA Style

Dong H, Xie K, Xie A, Wen C, He J, Zhang W, Yi D, Yang S. Detection of Occluded Small Commodities Based on Feature Enhancement under Super-Resolution. Sensors. 2023; 23(5):2439. https://doi.org/10.3390/s23052439

Chicago/Turabian Style

Dong, Haonan, Kai Xie, An Xie, Chang Wen, Jianbiao He, Wei Zhang, Dajiang Yi, and Sheng Yang. 2023. "Detection of Occluded Small Commodities Based on Feature Enhancement under Super-Resolution" Sensors 23, no. 5: 2439. https://doi.org/10.3390/s23052439

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop