Next Article in Journal
A Study on the Geometric and Kinematic Descriptors of Trajectories in the Classification of Ship Types
Next Article in Special Issue
Novel Spatio-Temporal Continuous Sign Language Recognition Using an Attentive Multi-Feature Network
Previous Article in Journal
Low-Light Image Enhancement via Retinex-Style Decomposition of Denoised Deep Image Prior
Previous Article in Special Issue
A Hierarchical Approach for Traffic Sign Recognition Based on Shape Detection and Image Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimisation of Deep Learning Small-Object Detectors with Novel Explainable Verification

by
Elhassan Mohamed
1,*,
Konstantinos Sirlantzis
1,*,
Gareth Howells
2 and
Sanaul Hoque
1
1
School of Engineering, University of Kent, Canterbury CT2 7NT, UK
2
School of Computing, University of Kent, Canterbury CT2 7NZ, UK
*
Authors to whom correspondence should be addressed.
Sensors 2022, 22(15), 5596; https://doi.org/10.3390/s22155596
Submission received: 5 June 2022 / Revised: 13 July 2022 / Accepted: 18 July 2022 / Published: 26 July 2022
(This article belongs to the Special Issue Vision Sensors for Object Detection and Recognition)

Abstract

:
In this paper, we present a novel methodology based on machine learning for identifying the most appropriate from a set of available state-of-the-art object detectors for a given application. Our particular interest is to develop a road map for identifying verifiably optimal selections, especially for challenging applications such as detecting small objects in a mixed-size object dataset. State-of-the-art object detection systems often find the localisation of small-size objects challenging since most are usually trained on large-size objects. These contain abundant information as they occupy a large number of pixels relative to the total image size. This fact is normally exploited by the model during training and inference processes. To dissect and understand this process, our approach systematically examines detectors’ performances using two very distinct deep convolutional networks. The first is the single-stage YOLO V3 and the second is the double-stage Faster R-CNN. Specifically, our proposed method explores and visually illustrates the impact of feature extraction layers, number of anchor boxes, data augmentation, etc., utilising ideas from the field of explainable Artificial Intelligence (XAI). Our results, for example, show that multi-head YOLO V3 detectors trained using augmented data produce better performance even with a fewer number of anchor boxes. Moreover, robustness regarding the detector’s ability to explain how a specific decision was reached is investigated using different explanation techniques. Finally, two new visualisation techniques are proposed, WS-Grad and Concat-Grad, for identifying explanation cues of different detectors. These are applied to specific object detection tasks to illustrate their reliability and transparency with respect to the decision process. It is shown that the proposed techniques can result in high resolution and comprehensive heatmaps of the image areas, significantly affecting detector decisions as compared to the state-of-the-art techniques tested.

1. Introduction

Machine learning in general and deep learning in specific has demonstrated efficiency in a wide range of applications, such as chemical process analysis [1], defect detections [2] and medical image analysis [3]. Moreover, state-of-the-art object detector systems based on deep learning and convolutional neural networks have shown significant performances in terms of accuracy and speed on standard datasets [4,5]. Such datasets mainly contain large objects that occupy a large area of an image. This helps the detector to exploit more pixels during the training and inference stages. Consequently, more information can be used in the training and prediction steps. The case becomes more challenging for small size objects as the information available to the system for training or inference is limited. Furthermore, small size objects may appear in groups which further complicate their detection. When it comes to small-size objects, the performance of state-of-the-art systems has not been systematically investigated.
Anchor boxes that are estimated from the training data are used as initial priors to enhance the predicted bounding boxes. Small anchor boxes are efficient with small objects but fail to capture large ones and vice versa. This limits the capabilities of state-of-the-art object detectors to detect small and large size objects simultaneously. Thus, redesigning the model’s architecture and anchor boxes may help to tackle this problem. However, increasing the number of anchor boxes to fit both object sizes can negatively impact the detector’s speed and accuracy. Speed-accuracy trade-off is another major challenge for object detectors [6]. Therefore, object detectors are considered application-oriented.
This paper investigates the impact of different object detection architectures on detector performance. State-of-the-art object detectors are tested on an application-specific dataset. The proposed dataset is challenging as its major components are small size objects. Nevertheless, it contains large and medium size objects as well, which introduces more challenges. Optimisation parameters, such as the minimum number of anchor boxes that can efficiently represent the proposed datasets, which base network to use, which feature extraction layer to use and the impact of different detector architectures on the accuracy, are investigated. Besides, the training strategies and data augmentation implications are sought to be discussed.
The contribution of this paper is twofold:
  • A deep investigation of state-of-the-art object detectors is presented on a challenging dataset that has multi-size objects but the majority of these objects are small size. The investigation presents a road map for researchers and developers to choose the most appropriate detector for a given application.
  • Novel techniques to visualise the contributing features to the detection decision to ensure the reliability of the produced detectors are proposed. The proposed techniques, besides state-of-the-art ones, are applied for the object detection task. Explainable Artificial Intelligence (XAI) techniques are usually used for classification tasks. However, applying these techniques for a different task (object detection) is a novel contribution.
This paper is organised as follows: state-of-the-art systems are presented in Section 2. The methodology is introduced in Section 3, where the system architecture, the dataset, the training parameters and the evaluation metrics are discussed. In Section 4, results are presented and discussed. Section 5 proposes novel visualisation techniques for convolutional neural networks predictions with application to object detection tasks. Finally, the paper is concluded in Section 6.

2. State-of-the-Art

Object detection architectures can be split into two main categories: one-stage and two-stage detectors. Unlike one-stage detectors, two-stage detectors have an extra step to generate region proposals. Every generation of the region proposal-based methods offer an improvement over the previous generation. For instance, Spatial Pyramid Pooling (SPP) network [7] modified region-based Convolutional Neural Network (R-CNN) [8] with an SPP layer. The SPP layer uses the convolutional layer’s feature map to generate fixed-size vectors for object proposals. Fast R-CNN [9] is proposed to overcome R-CNN and SPP-Net problems. Without caching features, it can be trained end-to-end on a multi-task loss function for classification and bounding box regression. Fast R-CNN uses a special case of SPP layer called the Region of Interest (RoI) pooling layer, which has one pyramid level. In addition, softmax classifier that is used in Fast R-CNN outperforms linear Support Vector Machines (SVMs) classifiers used in R-CNN and SPP-Net. Further, it does not require disk storage. Besides, it improves both accuracy and efficiency.
Object detection algorithms such as SSP-Net and Fast R-CNN suffer from bottleneck computations due to region proposals. Ren et al. [5] introduced Region Proposal Network (RPN), a fully convolutional network that shares full-image convolutional features with the detection network. It can be trained end-to-end to generate region proposals at nearly cost-free computations.
Unlike previous methods [7,9,10,11], which use pyramids of images or filters, Faster R-CNN [5] introduces the concept of anchor boxes. RPN adapts anchors (reference boxes) with three different scales and three different aspect ratios. The regression towards the output bounding boxes is achieved by comparing the proposed boxes with the anchors. RPN guides the Fast R-CNN network to the places where it is most likely to detect objects.
On the other hand, Lenc et al. [12] introduce one of the early attempts to accelerate the two-stage object detection networks. The study suggests dropping the region proposal section from the R-CNN [8] as it represents the bottleneck of the architecture. Instead of using a selective search for region proposals, the proposed system uses an image-independent list of candidate regions sampled from the distribution of the bounding boxes in the dataset. The investigation found that the CNN architecture by itself, without the fully connected layers, contains sufficient geometric information (spatial information) for accurate object detection. However, the accuracy of the proposed system without the region proposals network is negatively impacted.
Unlike detection approaches that modify classifier networks to perform detection, YOLO [13] (You Only Look Once) dealt with the object detection task as a regression problem in which bounding boxes are spatially separated and associated with class probabilities. YOLO approach uses a CNN to predict both the class probabilities and the bounding boxes from an image. It is a unified real-time object detection system with a design resembling GoogleNet [14] that needs one evaluation (forward pass) for predictions. Besides, the network can be trained end-to-end.
A disadvantage of the YOLO [13] approach is that it can only predict two bounding boxes for each grid cell. In addition, each grid cell can only have one class. As this spatial constraint, the detection of close objects is limited. Consequently, the network struggle with objects that appear in groups.
YOLO V2 [15], V3 [4] and V4 [16] are introduced to solve some of the challenges of YOLO V1 [13] and to enhance the detector performance. For instance, YOLO V2 uses Batch Normalisation (BN) [17], anchor boxes and multi-scale training. Whereas YOLO V3 uses residual connections [18] and Feature Pyramid Network (FPN) [19] with three predictions at different layers to process the image at different spatial resolutions. YOLO V4 [16] uses a different backbone network called CSPDarknet53. The introduced backbone network uses Cross Stage Partial Network (CSPNet) strategy to partition the feature map of the base layer into two parts and then merge them through a cross-stage hierarchy. The split and merge strategy allows for more gradient flow through the network. Comparisons of YOLO detector performances on different datasets are presented in [20,21,22]. RetinaNet [23] introduces two improvements over previous single-stage detectors. It uses Feature Pyramid Network (FPN) [19] and a novel focal loss instead of cross-entropy loss.
Small object detection represents a challenge for state-of-the-art detectors. These detectors are fine-tuned on datasets that contain large size objects. Besides, the base networks of these detectors are trained on general datasets such as ImageNet [24]. Studies show that state-of-the-art models [25,26,27,28,29], as well as standard datasets, such as PASCAL [30] and Microsoft COCO [31], do not give much consideration to small object detection. The performance of these models on small size objects is not deeply investigated as the evaluation of these models with the focus on the detection of small size objects is limited [27].
Moreover, the definition of small object size is not unified, which presents another challenge for researchers. Chen et al. [25] classify objects from the PASCAL dataset to be small if the ratio between the bounding box area to the image area, averaged over all the instances of that class, is in the range of 0.08% to 0.58%. This corresponds to 16 × 16 to 42 × 42 pixels. The small object can vary in size according to the image size, which is not constant for the PASCAL dataset. To compare, the median relative areas of the PASCAL dataset are between 1.38% to 46.0%.
Torralba et al. [32] introduce a dataset for tiny images with 32 × 32 pixels. Zhu et al. [26] follow the definition of the Microsoft COCO dataset for small-size objects to be equal to or less than 32 × 32 pixels. Microsoft COCO contains small objects, but they occupy large parts of the images. The variation in small-size objects definition is attributed to the dataset image size. For the PASCAL dataset, the image size varies. Whereas for the Microsoft COCO dataset, the image size is fixed and is equal to 640 × 480 pixels.
In the light of the previously mentioned definitions, this paper follows a new definition for small size objects. An object is categorised as small if its size equals to or is less than 42 × 42 pixels. This definition is adopted as the image size in the proposed object detection dataset is 512 × 512.
Chen et al. [25] is one of the first works that try to enhance the performance of R-CNN on small-size objects. The study introduces ContextNet, at which the region proposals and the context of the regions are forward propagated through two CNNs. Then the results of the two networks are concatenated. A limitation of the proposed system is that the two CNNs do not share any weights. Consequently, the system requires more training time and resources.
Several strategies have been introduced to enhance the detector performance on small-size objects, such as feature learning, context-based detection, data augmentation and training strategies. In addition, Generative-Adversarial Networks (GAN) [33] achieved good results on the task of small object detection. Tong et al. [34] review deep learning methods for small object detection. The review highlights the following remarks: multi-scale feature learning, context modelling and data augmentation can enhance the performance of state-of-the-art detection methods in the detection of small-size objects. Input image resolution and base networks have a great impact on detection performance. The combination of multiple techniques to enhance object detectors can further improve the performance [35,36]. Lastly, large datasets and the combination of multiple datasets can boost the detector to learn the better representation of small-size objects.
This paper investigates the performance of state-of-the-art object detection systems on the proposed dataset. The detailed investigation of different detector architectures and different training strategies gives a road map to choosing the most optimal system for a given application.

3. Methodology

3.1. System Architecture

Two of the most widely adopted object detection systems are used for investigation and performance comparison. The first detector is the one-stage YOLO V3 [4], while the second is the two-stage Faster R-CNN [5]. Both detectors use ResNet-18 as the base network.
The pipeline of the Faster R-CNN network consists of a feature extraction network, an RPN, and two sub-networks for class prediction and bounding box regression. The feature extraction network is a pretrained network that extracts the features from the input image. The RPN is trained to extract region proposals from the feature maps produced by the feature extraction network. Lastly, the classification and regression networks predict the class category and the bounding box of each region proposal. On the other hand, YOLO V3 uses a CNN to predict both the class probabilities and the bounding boxes from an image without the need for a region proposals step.
The choice of the feature extraction network is based on the application requirements. A deep network results in high accuracy but low processing speed and vice versa. Thus, the choice of the base network is a trade-off between accuracy and speed.
The choice of the feature extraction layer that feeds into the RPN is also a trade-off between the strength of the extracted features and the spatial resolution. High feature extraction layers (deep layers) down the network result in strong encoded features, but the object’s spatial information is lost. However, feature extraction layers up the network (early layers) have a better spatial resolution but weak encoded features. Empirical analysis can identify the optimal feature extraction layer for a specific application.
The detection of small size objects is a delicate task. The spatial information and features of small size objects are limited. Consequently, these objects can get lost as the feature maps are down-sampled through the network layers.
ResNet-18 [18] has been used as the feature extraction network in the experiments. ResNet-18 is the smallest version of the ResNet family. Nevertheless, it is a powerful network that uses residual blocks. It can achieve adequate processing speed with high accuracy. Residual blocks help to overcome deep network problems of vanishing and exploding gradients [37,38]. Residual blocks reuse the activations from previous layers until the adjacent layer learns its weights [18]. Four different feature extraction layers are used in the experiments to investigate the trade-off between spatial resolution and discriminative features: ‘res4a_relu’, ‘res4b_relu’, ‘res5a_relu’ and ‘res5b_relu’. These are the ReLU layers after the last four residual blocks of the ResNet-18 network.

3.2. Object Detection Dataset

The proposed object detection dataset contains 3292 images that are collected using a handheld camera. The object detection dataset is annotated at the bounding box level. Objects are categorised into eight classes. The number of object instances per class and the number of images which contain that object are shown in Table 1. The highest number of instances is for the ‘Door’ class. Whereas the lowest is for the ‘Push button’. Images of the dataset are shuffled and split randomly into 60% for training (1975), 10% for validation (330 images) and 30% for testing (987 images).
The proposed object detection dataset mainly contains small size objects. Objects with sizes less than or equal to 42 × 42 pixels are categorised as small-size objects. Objects bigger than 42 × 42 pixels and less than 96 × 96 pixels are categorised as medium size objects. Objects greater than 96 × 96 pixels are categorised as large-size objects. The definition of object sizes follows that of the Traffic [26] and COCO [31] datasets except for small size objects because the proposed images are larger than that of the traffic dataset. Consequently, the definition of small-size objects is 42 × 42 pixels instead of 32 × 32 pixels.
Figure 1 shows the sizes and the aspect ratios of the object detection dataset. For ease of understanding and to better distinguish between objects, a set of randomly selected 100 objects from each category are displayed. The majority of the object sizes can be categorised as small and medium size objects. Consequently, the performance of state-of-the-art detection systems trained on large-size objects can differ due to the different nature of the proposed dataset. These detectors are designed with anchor boxes to accommodate general dataset objects (mainly large objects). Therefore, task-specific datasets need different designs for the anchor boxes and different techniques to capture small-size objects along with medium and large-size ones.

3.3. Training

3.3.1. Training Parameters

Several training parameters are tried to find the optimal ones that can achieve the highest performance. The chosen training parameters for both detectors (Faster R-CNN and YOLO V3) are as follows: Stochastic Gradient Descent with Momentum (SGDM) is used as the training optimiser with 0.9 momentum. The Learning rate starts at 0.001 and then drops by a factor of 0.1 every six epochs. L2 regularisation of 0.005 is utilised to avoid overfitting. Training examples are shuffled every epoch to limit sequence memorising and avoid computing the gradients for the same batch of images.

3.3.2. Data Augmentation

Data augmentation refers to increasing the number of images or instances of small size objects by image transformation that includes flipping, cropping, scaling, etc. The main idea is to extend the dataset with a large amount of data by increasing the representation of small size objects, which can help to boost the performance of detectors on small size objects [39].
Data augmentation techniques, such as image flipping can be employed to increase the variations and the number of training samples. Augmentation techniques can result in improved accuracy and enhanced model generalisation. Data augmentation techniques are applied to the training data only to produce a robust model and avoid evaluation bias. In the experiments, data augmentation techniques are employed on the training dataset by horizontal flipping of the images and associated boxes.

3.4. Evaluation

Average Precision ( A P ) that can be computed from the Precision ( P ) Recall ( R ) curves is the standard metric of evaluating object detectors. Precision can be calculated using Equation (1) as the ratio between True Positive ( T P ) instances to all positive instances. Whereas Recall can be calculated using Equation (2) as the ratio between T P instances to the sum of T P and False Negative ( F N ) instances (ground truth positives).
P = T P T P + F P
R = T P T P + F N
Intersection over Union is used to determine which detection is T P , False Positive ( F P ) , or F N . If there is an overlap between the detected object bounding box and the ground truth bounding box above a certain threshold (in the experiments, the threshold is set to 0.5), the detection is considered a T P . If the IoU is less than the threshold, the detection is F P . Lastly, if there is an object, but it has not been detected, or the object is detected with a wrong category, then it is a F N .
A P is then calculated as the area under the Precision/Recall curve for a specific class of objects using Equation (3). A high A P value indicates the ability of the model to detect a specific class of objects efficiently and vice versa.
A P = 0 1 P ( R ) d R
Mean Average Precision ( m A P ) is used to assess the detector’s abilities over all the dataset objects. m A P can be calculated using Equation (4), where A P k is the A P for class K and N is the total number of classes. The metric reflects the detector’s performance over the whole dataset objects. In the experiments, A P for each class and m A P for all classes are reported.
m A P = 1 N K = 1 N A P K

4. Experimental Analysis and Discussion

Anchor boxes are a set of predefined boxes with different sizes and aspect ratios that represent the objects of the dataset. They are estimated from the training data and used as initial priors to enhance the predicted bounding boxes.
Anchor boxes are used to eliminate the need to scan the whole image using different sizes and aspect ratios sliding windows. Consequently, the whole image can be processed in a single propagation through the network, which enhances the overall prediction speed. Different sizes of anchor boxes enable the detection of multi-scale objects. The model predicts the offsets of the anchor boxes to refine the boxes’ locations and sizes.
The final detector output is produced by removing the anchor boxes that belong to the background. Moreover, other anchor boxes with confidence scores below a specific threshold are ignored. Lastly, the multiple detections of the same object are refined using the Non-Maximum Suppression (NMS) technique. Anchor boxes enable the prediction of multiple objects with different sizes and scales, besides overlapping objects.
Manually selecting the anchor boxes for the dataset is challenging as object groups are scattered with varying sizes and aspect ratios (Figure 1). A clustering algorithm, such as k-means [40], can group boxes of similar aspect ratios and sizes based on a specific metric. The Intersection over Union (IoU) distance metric is used to estimate the anchor boxes that better represent the dataset objects. IoU distance metric-based clustering algorithm can produce anchor boxes that fit the dataset objects efficiently as it is invariant to the boxes’ sizes [15]. Whereas other metrics such as Euclidean distance can lead to large errors when the boxes’ sizes increase [15].
The number of anchor boxes is a hyper-parameter that can be selected empirically. However, the mean IoU (mIoU) between the training data boxes and the estimated anchor boxes can be used to assess the number and validity of the estimated boxes. Figure 2 shows the estimated number of anchor boxes w.r.t the training data bounding boxes and the corresponding mIoU. The maximum number of anchor boxes is set to 30 as the mIoU plateaus or degrades after this point. Arbitrary increasing the number of anchor boxes can negatively affect the detector performance. Many anchor boxes can result in overfitting of the training data. Besides, the computation cost is directly proportional to the number of anchor boxes. Consequently, it is a trade-off process where the lowest number of anchor boxes that can achieve the highest mIoU is the objective.
A large number of anchor boxes results in low-performance detectors. Thus, a mIoU greater than 0.5 indicates adequate overlap between the training boxes and the estimated anchor boxes. Usually, marginal improvement can be achieved with many anchor boxes (mIoU starts to oscillate between 0.6 and 0.75 after 15 anchor boxes).
Three data points are selected to understand the impact of the anchor boxes on the detector performance (Figure 2). First, the point at which the mIoU is more than or equal to 0.5 with the lowest number of anchor boxes (number of anchor boxes = 3, mIoU = 0.518). Second, the point at which the highest mIoU can be achieved (number of anchor boxes = 23, mIoU = 0.757). Third, the point with the highest number of anchor boxes (number of anchor boxes = 30, mIoU = 0.756).
The adequate number of anchor boxes to achieve high accuracy, fast processing speed, or a trade-off between both metrics can be attained by analysing the dataset objects. Nevertheless, the application requirements are the main motive for choosing the number of anchor boxes.
Detectors are trained on a personal computer with an NVIDIA GeForce RTX 2080. Training time varies as the training process can be stopped early when the loss of the validation dataset plateaus or when the training process reaches the maximum number of epochs (30 epochs). The largest mini-batch size that can accommodate the available memory is sought. The largest mini-batch sizes are 2 and 16 in the case of the Faster R-CNN and YOLO V3 detectors, respectively. Table 2 and Table 3 show the training time of each model, the used mini-batch size, the stopping epoch, the trained model size and the number of layers.
Generally, Faster R-CNN detectors take significant training time compared to YOLO V3 detectors. The long training time is attributed to the detector architecture, which comprises an RPN attached to a Fast R-CNN [9]. This is translated into many layers and large footprints (Table 2). In contrast, the footprints and number of layers of YOLO V3 detectors vary depending on the feature extraction layer and the number of prediction heads. The smallest YOLO V3 detector has 48 layers and occupies a memory of 10 MB (Table 3).
The loss functions that have been used in the training process of Faster R-CNN and YOLO V3 are different, which can explain the difference in the results of Table 2 and Table 3. The objective function of Faster R-CNN follows the multi-task loss function of Fast R-CNN. However, it is minimised by a combination of the object classification loss and the bounding box regression loss (Equation (5)). The classification loss is a l o g loss over two classes, while the regression loss is the smooth L 1 loss [9]. Smooth L 1 loss is less sensitive to outliers compared to L 2 loss, especially when regression targets are unbounded, which may cause exploding gradients when L 2 loss is used.
L Faster R - CNN = L cls + L box
On the other hand, the YOLO V3 loss function optimises the training process over three different losses. Like Faster R-CNN, the classification loss is the binary cross-entropy loss. Unlike Faster R-CNN, Mean Square Error (MSE) is used for the bounding box loss. Besides, YOLO V3 introduces the bounding boxes objectness loss [4], which is an additional binary cross-entropy loss for the overlapping between the predicted and the ground truth boxes. Ideally, the objectness score should equal one when the best overlapping anchor box among all anchor boxes overlaps with the ground truth box. The predictions are ignored when other anchor boxes (not the best overlapping anchor box) overlap with the object box. This means that there is one anchor box assigned for each ground truth object [4].
L YOLO V 3 = L cls + L box + L obj
Table 2 shows that the lowest validation loss achieved using Faster R-CNN is 0.192 using res5a_relu as the feature extraction layer with three anchor boxes. The lowest validation losses achieved using YOLO V3 with single, double and triple heads are 2.52, 2.35 and 2.38, respectively (Table 3). Like Faster R-CNN, the best-achieved validation loss YOLO V3 detectors use only three anchor boxes. Unlike Faster R-CNN, these YOLO detectors are trained using augmented data and with different feature extraction layers.
Earlier feature extraction layers in the network have higher spatial resolutions but may extract less semantic information compared to layers further down the network. High spatial resolution features are better for small and medium size objects but not for large size ones. In contrast, strong semantic information is important for large size objects. However, due to the successive down-sampling of the feature maps as the network goes deep, this information is lost for small objects. This makes the choice of the feature extraction layer a challenging task. As an example from Table 4 and Table 5, the A P of the smallest object in the proposed dataset (key slot) using earlier feature extraction layers such as res4a_relu or res4b_relu is significantly better than the A P when later layers such as res5a_relu or res5b_relu are used.
On the other hand, using res5a_relu or res5b_relu as the feature extraction layers on the largest size object in the proposed dataset (door) produces better A P than using res4a_relu or res4b_relu. This can be clearly seen from the YOLO V3 results (Table 5). In contrast, Faster R-CNN results do not reflect this fact (Table 4).
YOLO V3 detector can make predictions using multiple prediction heads over different scale feature maps in a similar approach to Feature Pyramid Network (FPN) [19]. The first head makes predictions over the first feature map. The second head makes predictions over a concatenation of the current feature map, after up-sampling and the previous feature map. The same approach is followed for the other heads. Thus, semantic information and fine-grained details can be obtained from the up-sampled and high-resolution feature maps. This approach allows the prediction of different scale objects, where small size objects can be detected from the high-resolution maps and large size objects can be extracted from strong semantic feature maps.
On the other hand, using a single feature map for prediction is less efficient than predictions over multiple feature maps, even with multiple scale anchor boxes (pyramid of anchors) that are used in Faster R-CNN [5]. Overall, the performance of YOLO V3 using single or multiple prediction heads are significantly better than that of Faster R-CNN for all object sizes.
Table 4 shows that the best Faster R-CNN detector achieved a m A P of 0.434. Whereas the best single-head YOLO V3 detector achieved a m A P of 0.765 (Table 5). A faster R-CNN detector uses res4b_relu as the feature extraction layer with 23 anchor boxes. Similarly, the YOLO V3 detector uses res4b_relu as the feature extraction layer but with only three anchor boxes. The best double and triple heads YOLO V3 detectors achieved a m A P of 0.786 and 0.818, respectively. Both of them use only three anchor boxes. Lastly, the overall best performance detector is YOLO V3 with triple heads and trained on augmented data.
Unlike the presented investigation, Liu et al. [41] survey shows that Faster R-CNN produces slightly better results than YOLO V3 on small object datasets [42,43,44]. Reflecting on the survey results [41] that use off-the-shelf Faster R-CNN and YOLO V3, the presented implementation unified the detectors set up in terms of using the same base networks, the same number of anchor boxes and the same feature extraction layers to facilitate the comparisons. Consequently, the detectors’ comparisons are robust and reliable. Moreover, the proposed dataset contains mainly small size objects but also medium and large size objects which can justify the differences.
On the other hand, the presented results align with the findings of Zeng et al. [2] on detecting tiny surface defects of the printed circuit board where the YOLO V3 detector can perform better on small-size objects. However, Faster R-CNN can better detect large-size objects [2]. Similarly, the proposed YOLO V3 detector can better detect small-size objects such as push buttons and key slots Table 4 and Table 5.
Figure 3 shows detection examples of the best performing Faster R-CNN (with 23 anchor boxes and res4b_relu feature extraction layer) and YOLO V3 (with three anchor boxes, res4b_relu feature extraction layer and trained using augmented training data) single head detectors on two test images. The two networks can predict the class categories and the bounding boxes with high confidence. However, Faster R-CNN predicts two bounding boxes for the same object (the moveable door handle in Figure 3f), one with a high confidence score of 0.99, while the second confidence score is relatively low (0.58). The confidence score threshold value used in the experiments is 0.5. A higher threshold value can discard the second box. Other Faster R-CNN detectors find it challenging to detect all the objects in the test images.
In contrast, other YOLO V3 detectors can localise all the objects in the test images with minor differences in the confidence scores. Another observation is that the predicted bounding box using Faster R-CNN for the fire extinguisher in Figure 3c covers the whole object, unlike the produced bounding box from the YOLO V3 detector (Figure 3b). In comparison, the bounding box for the door object in the same images is fully covered by the YOLO V3 detector and partially covered by the Faster R-CNN detector.
Detection examples of the best performing three heads YOLO V3 detector are shown in Figure 4. The detector can localise small size objects, such as ID readers and pull door handles, along with large and medium size objects.
Detailed investigation of different detectors using different parameters can give insights into the suitable detector for a given application. YOLO V3 with three detection heads and three anchor boxes has achieved the best performance using data augmentation techniques during training on the proposed dataset. Moreover, the system needs less time for training compared to Faster R-CNN detectors.
It can be concluded from the results that small-size objects can be accurately detected using YOLO V3 detectors with an earlier feature extraction layer. On the other hand, increasing the number of anchor boxes is not enhancing the detector’s accuracy. However, feature extraction layers significantly impact the detector’s performance. Feature extraction layers greatly affect the detector’s ability to capture different size objects. An earlier layer can better localise small size objects, while a deeper layer can better encode large size objects. Consequently, multi-head detectors can capture different size objects efficiently because the predictions are made over several feature maps.

5. Visualisation of Detector Predictions

Visualisation techniques are essential tools to understand CNN behaviours. Reliable systems based on deep learning techniques need to reason about their predictions. For this reason, the transparency of the proposed systems is tested to ensure their robustness and accelerate their approval for real-life applications.
As an input pattern causes a given activation in the feature maps, Zeiler et al. [45] map this activation back to the input pixel space using deconvolutional networks [46]. The steps can be explained as follows: an input image is presented to the CNN and the features are computed through the networks’ layers. To analyse a given activation, all other activations in that layer are set to zero. Then the feature maps are passed to the attached deconvolutional layer. Finally, the input pixel space is reached through successive un-pooling, rectifying and filtering operations to reconstruct the layer’s activity.
Gradients approach [47], also known as backpropagation or saliency method, visualises the derivatives of the target object score with respect to the input image (Equation (7)). Saliency maps are generated for the trained network and not during the training process (i.e., the networks’ weights are constant). Backpropagation is the process of increasing or decreasing networks’ weights to minimise the loss of function during the training process [48]. Saliency maps return the spatial discriminative pixels locations of a particular class in an image.
Although Gradient heatmaps are computationally faster than Occlusion [46] as it only needs one backward propagation through the network, they do not fully explain the image prediction. The calculated map measures pixels change that would make an image belong to a specific category. However, it does not explain the classifier decision as argued by [49] or the direct relation to the variation of the output [50,51].
DeconvNet approach [46], which zeros negative values of the top gradients and backpropagation [47], which zeros negative values from the bottom inputs, are then combined to produce Guided Backpropagation (GBP) [52], which zeros both negative values (Equation (8)). The signal from higher layers guides the backpropagation; hence the name is derived. It works as the switches of the DeconvNet approach [46]. Doing so prevents negative gradients from flowing back, which can undesirably impact the activation’s visualisation.
Many approaches based on Gradients are proposed, such as Integrated Gradients (IG) [53]. IG [53] approach accumulates gradients over scaled versions of the input that follow a baseline defined by the user, i.e., they integrate the gradients of all points that fall on the straight-line path from the baseline to the input (Equation (9)).
Z i j G r a d : = Y c ( X ) X i j
Z i j G B P : = Y c ( X ) X i j X i j > 0 & Y c X i j > 0 0 otherwise
Z i j I G : = ( X i j X ¯ i j ) α = 0 1 Y c ( X ¯ i j + α ( X i j X ¯ i j ) ) X i j d α
where,
  • X: Input image
  • X i j : Image pixel
  • X ¯ i j : Baseline image pixel
  • Y c : Output prediction for class c
  • α : A real value parameter that defines the path steps between the baseline image and the input image
  • Z i j : Generated heatmap
  • Gradient calculations following the chain-rule.
Gradient-based methods, such as saliency [47], GBP [52] and IG [53], can produce high-resolution heatmaps. However, each approach captures specific features that contribute to the overall output. Two techniques are proposed to attain the benefits of these methods: weighted sum of gradient approach (WS-Grad) and concatenation of gradient approach (Concat-Grad).

5.1. Weighted Sum of Gradients Approach (WS-Grad)

Figure 5b shows the weighted sum gradients approach. First, gradients-based heatmaps are generated (Figure 5a). Then each map is scaled by a weight that the user can determine to highlight specific features because different maps can highlight various elements (Equation (10)). For example, saliency maps (Gradient approach) highlight all the features that contribute equally to the prediction. However, GBP focuses on the most discriminative features and ignores supplementary ones. IG approach accumulates gradients over scaled versions of the input that follow a baseline defined by the user. The flexibility of choosing the weights is a powerful tool that can be utilised differently according to the application. The produced heatmap is more expressive than the individual ones, where the most important features are strongly highlighted with high resolution.
Z i j W S G r a d = W 1 Z i j Grad + W 2 Z i j GBP + W 3 Z i j I G

5.2. Concatenation of Gradients Approach (Concat-Grad)

Figure 5c shows the concatenated gradients approach. Similar to WS-Grad, Gradient-based heatmaps are generated. Then, the generated maps are weighted and concatenated as a single image with three channels (Equation (11)).
The Concat-Grad approach produces high-resolution heatmaps where all the important features using different visualisation approaches can be seen and identified in one image. The generated map has three channels (similar to an RGB image). Consequently, the produced heatmap reflects each map using a different colour. This means Gradient, GBP and IG features are depicted in red, green and blue colours, respectively. The novelty of this method can be seen in the ability to distinguish various features of different approaches using distinctive colours in one map, which is very informative as it gives better insights into the important features and their corresponding approach.
Z i j C o n c a t G r a d = W 1 Z i j G r a d W 2 Z i j G B P W 3 Z i j I G

5.3. Visualisation Results

To validate the reliability of the best single-head YOLO V3 detector, novel techniques for visualising the network decisions are applied. The proposed techniques can be applied to classification tasks like other visualisation techniques [54]. However, applying them to different tasks, such as object detection, is a novel contribution.
The output of a YOLO V3 single head detector is N × N × [ 3 × ( 4 + 1 + 8 ) ] where N represents the convolution filter size, 3 represents the number of anchor boxes, 4 represents the bounding box offsets, 1 represents the objectness score and 8 represents the scores of classes. The convolutional filter size is 16, resulting in an output tensor of dimension 16 × 16 × 39 . The target object score needs to be tracked back through the network layers to find the contributing input pixels to the target object score. However, the final output scores are obtained by multiplying the confidence score and the objectness score.
Tracking the final output scores through the output tensor is challenging because of the post-processing step that extracts the confidence and objectness score from the output tensor and multiplies them. This post-processing step needs to be reversed. Besides, the two values that produce the final output score need to be tracked back (using backpropagation) through the network to find the corresponding contributing features. In contrast, the final output score can be directly tracked in classification tasks that use the softmax layer in the output, whereas YOLO V3 does not have any softmax layers. The locations of the confidence and objectness scores are identified by analysing the output tensor. Consequently, the output target score has become feasible to be tracked through the detector network. In other words, by identifying the location of the confidence and objectness score in the output tensor, the output score of the target object has become trackable.
Figure 6 and Figure 7 show the contributing pixels to the detector predictions of the door and the fire extinguisher, respectively. The weighting parameter of the proposed methods is set to one for a fair comparison with state-of-the-art methods. Gradient [47], GBP [52], IG [53], WS-Grad and Concat-Grad attribution maps are compared. Gradient, GBP and IG show specific individual features. However, the proposed methods show comprehensive maps that contain all of the contributing pixels. Furthermore, the Concat-Grad method shows the features of each individual method in different colours on a single map which enriches the output and makes it more descriptive and understandable.
As an example, the WS-Grad heatmap of the door in Figure 6e highlights the important common features among all other individual visualisation techniques. On the other hand, the Concat-Grad technique visualises the individual features in different colours. This allows the identification of correct features that contributes to the network decision. For instant, the Gradient method that is depicted in red shows that the network-based its decision on the context information. Whereas the GBP features that are depicted in green show that the network-based its decision on the door handle. Lastly, the IG, which is depicted in blue, shows that the network-based its decision on the door panel. While some features should not contribute to the network decision, such as the door panel, visualising different contributing features using the proposed method can greatly help to understand and enhance the detector performance. In this case, the detector performance and ability to generalise can be enhanced further by training on door images without panels.
Generally, the research area of visualising the network decision for object detection tasks has not been adequately explored. The proposed techniques can help researchers in the field of object detection to understand and trust the decisions of their systems. Besides, visualising the detector prediction can help to debug the system in the case of bias or error.
One limitation of this study is in the visualisation of Faster R-CNN predictions. Faster R-CNN detectors contain some layers that are challenging to reverse (backpropagate the output through them) such as region proposal and region of interest pooling layers. Further, the 4-step alternating training method [5] that is used by Faster R-CNN to train the region proposal and region classification sub-networks separately makes it challenging to track the activations through the network layers.

6. Conclusions

This paper presents a comprehensive investigation of object detector performance. Detector performance using different feature extraction layers and a different number of anchor boxes is investigated. Besides, the impact of training the detectors using data augmentation techniques is highlighted.
Data augmentation positively impacts the generated detectors and results in lower validation loss compared to detectors trained on data without augmentation techniques. However, increasing the number of anchor boxes does not enhance the detector’s performance. In contrast, it can negatively impact performance. YOLO V3 detectors with multi-prediction heads achieved the best performance. Furthermore, YOLO V3 detectors have fewer layers, less footprint and train faster than Faster R-CNN detectors.
Feature extraction layer can significantly impact the ability of the detector to localise different size objects. An earlier feature extraction layer can better detect small size objects, as it can preserve spatial information. Whereas later layers are better with large size objects because it better encodes the object’s features. Successive down-sampling of features as the object propagates through the network layers strengthen the encoded features, but spatial information of small size object can be lost. Consequently, earlier feature extraction layer is advisable with small-size object detection applications.
The paper greatly contributes to the visualisation and explanation techniques applied to the object detection task as research in this area is very limited. It is important to attain not only an accurate system but also a system that can explain its predictions. Black box systems, like deep convolutional networks, must provide adequate insights into the system’s predictions. Developers, policymakers and legislators often require a certain level of system transparency to approve/appreciate such technologies and can self-assuredly conclude that the underlying system is robust and reliable. Consequently, these kinds of transparent systems can be approved and used for critical real-life applications. The proposed explanation techniques help achieve this by providing high-resolution and sharp heatmaps for the contributing features to the network decision compared to the state-of-the-art ones. This can greatly help to understand and explain the detector’s behaviour.

Author Contributions

Conceptualization, E.M., K.S., G.H. and S.H.; methodology, E.M.; software, E.M., K.S., G.H. and S.H.; validation, E.M., K.S., G.H. and S.H.; formal analysis, E.M., K.S., G.H. and S.H.; investigation, E.M., K.S., G.H. and S.H.; resources, E.M.; data curation, E.M.; writing—original draft preparation, E.M., K.S., G.H. and S.H.; writing—review and editing, E.M.; visualisation, E.M.; supervision, K.S. and G.H.; project administration, K.S.; funding acquisition, K.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Engineering and Physical Sciences Research Council (EPSRC), U.K., under Grant NPIF2018-2112938. European Regional Development Fund (ERDF), INTERREG VA France (Channel) England Programme, ADAPT project.

Acknowledgments

This work is supported by the Assistive Devices for empowering dis-Abled People through robotic Technologies (ADAPT) project and the Engineering and Physical Sciences Research Council (EPSRC), U.K., under Grant NPIF2018-2112938. ADAPT is selected for funding by the INTERREG VA France (Channel) England Programme which is co-financed by the European Regional Development Fund (ERDF). The European Regional Development Fund (ERDF) is one of the main financial instruments of the European Union (EU) cohesion policy.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jiang, Y.; Zhang, G.; Wang, J.; Vaferi, B. Hydrogen solubility in aromatic/cyclic compounds: Prediction by different machine learning techniques. Int. J. Hydrogen Energy 2021, 46, 23591–23602. [Google Scholar] [CrossRef]
  2. Zeng, N.; Wu, P.; Wang, Z.; Li, H.; Liu, W.; Liu, X. A Small-Sized Object Detection Oriented Multi-Scale Feature Fusion Approach With Application to Defect Detection. IEEE Trans. Instrum. Meas. 2022, 71, 3507014. [Google Scholar] [CrossRef]
  3. Davies, R.H.; Augusto, J.B.; Bhuva, A.; Xue, H.; Treibel, T.A.; Ye, Y.; Hughes, R.K.; Bai, W.; Lau, C.; Shiwani, H.; et al. Precision measurement of cardiac structure and function in cardiovascular magnetic resonance using machine learning. J. Cardiovasc. Magn. Reson. 2022, 24, 1–11. [Google Scholar] [CrossRef]
  4. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  5. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28, 91–99. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Huang, J.; Rathod, V.; Sun, C.; Zhu, M.; Korattikara, A.; Fathi, A.; Fischer, I.; Wojna, Z.; Song, Y.; Guadarrama, S.; et al. Speed/accuracy trade-offs for modern convolutional object detectors. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7310–7311. [Google Scholar]
  7. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  9. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Washington, DC, USA, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  10. Felzenszwalb, P.F.; Girshick, R.B.; McAllester, D.; Ramanan, D. Object detection with discriminatively trained part-based models. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 32, 1627–1645. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Sermanet, P.; Eigen, D.; Zhang, X.; Mathieu, M.; Fergus, R.; LeCun, Y. Overfeat: Integrated recognition, localisation and detection using convolutional networks. arXiv 2013, arXiv:1312.6229. [Google Scholar]
  12. Lenc, K.; Vedaldi, A. R-cnn minus r. arXiv 2015, arXiv:1506.06981. [Google Scholar]
  13. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  14. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. arXiv 2015, arXiv:1409.4842. [Google Scholar]
  15. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  16. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  17. Ioffe, S.; Szegedy, C. Batch normalisation: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning, PMLR, Lille, France, 6–11 July 2015; pp. 448–456. [Google Scholar]
  18. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. arXiv 2016, arXiv:1512.03385. [Google Scholar]
  19. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. arXiv 2017, arXiv:1612.03144. [Google Scholar]
  20. Nepal, U.; Eslamiat, H. Comparing YOLOv3, YOLOv4 and YOLOv5 for Autonomous Landing Spot Detection in Faulty UAVs. Sensors 2022, 22, 464. [Google Scholar] [CrossRef]
  21. Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. Yolox: Exceeding yolo series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
  22. Long, X.; Deng, K.; Wang, G.; Zhang, Y.; Dang, Q.; Gao, Y.; Shen, H.; Ren, J.; Han, S.; Ding, E.; et al. PP-YOLO: An effective and efficient implementation of object detector. arXiv 2020, arXiv:2007.12099. [Google Scholar]
  23. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. arXiv 2017, arXiv:1708.02002. [Google Scholar]
  24. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  25. Chen, C.; Liu, M.Y.; Tuzel, O.; Xiao, J. R-CNN for small object detection. In Proceedings of the Asian Conference on Computer Vision, Taipei, Taiwan, 20–24 November 2016; pp. 214–230. [Google Scholar]
  26. Zhu, Z.; Liang, D.; Zhang, S.; Huang, X.; Li, B.; Hu, S. Traffic-sign detection and classification in the wild. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2110–2118. [Google Scholar]
  27. Pham, P.; Nguyen, D.; Do, T.; Ngo, T.D.; Le, D.D. Evaluation of deep models for real-time small object detection. In Proceedings of the International Conference on Neural Information Processing, Long Beach, CA, USA, 4–9 December 2017; pp. 516–526. [Google Scholar]
  28. Zhao, Z.Q.; Zheng, P.; Xu, S.t.; Wu, X. Object detection with deep learning: A review. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 3212–3232. [Google Scholar] [CrossRef] [Green Version]
  29. Nguyen, N.D.; Do, T.; Ngo, T.D.; Le, D.D. An evaluation of deep learning methods for small object detection. J. Electr. Comput. Eng. 2020, 2020. [Google Scholar] [CrossRef]
  30. Everingham, M.; Eslami, S.A.; Van Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The pascal visual object classes challenge: A retrospective. Int. J. Comput. Vis. 2015, 111, 98–136. [Google Scholar] [CrossRef]
  31. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 740–755. [Google Scholar]
  32. Torralba, A.; Fergus, R.; Freeman, W.T. 80 million tiny images: A large data set for nonparametric object and scene recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2008, 30, 1958–1970. [Google Scholar] [CrossRef]
  33. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. Commun. ACM 2020, 63, 139–144. [Google Scholar] [CrossRef]
  34. Tong, K.; Wu, Y.; Zhou, F. Recent advances in small object detection based on deep learning: A review. Image Vis. Comput. 2020, 97, 103910. [Google Scholar] [CrossRef]
  35. Singh, B.; Davis, L.S. An Analysis of Scale Invariance in Object Detection—SNIP. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3578–3587. [Google Scholar]
  36. Bell, S.; Zitnick, C.L.; Bala, K.; Girshick, R. Inside-outside net: Detecting objects in context with skip pooling and recurrent neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2874–2883. [Google Scholar]
  37. Bengio, Y.; Simard, P.; Frasconi, P. Learning long-term dependencies with gradient descent is difficult. IEEE Trans. Neural Netw. 1994, 5, 157–166. [Google Scholar] [CrossRef]
  38. Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. Proc. Mach. Learn. Res. 2010, 9, 249–256. [Google Scholar]
  39. Kisantal, M.; Wojna, Z.; Murawski, J.; Naruniec, J.; Cho, K. Augmentation for small object detection. arXiv 2019, arXiv:1902.07296. [Google Scholar]
  40. Arthur, D.; Vassilvitskii, S. k-Means++: The Advantages of Careful Seeding. In Proceedings of the Eighteenth Annual ACM-SIAM Symposium on Discrete Algorithms, New Orleans, LA, USA, 7–9 January 2007. [Google Scholar]
  41. Liu, Y.; Sun, P.; Wergeles, N.; Shang, Y. A survey and performance evaluation of deep learning methods for small object detection. Expert Syst. Appl. 2021, 172, 114602. [Google Scholar] [CrossRef]
  42. Xia, G.S.; Bai, X.; Ding, J.; Zhu, Z.; Belongie, S.; Luo, J.; Datcu, M.; Pelillo, M.; Zhang, L. DOTA: A large-scale dataset for object detection in aerial images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3974–3983. [Google Scholar]
  43. Xiao, J.; Hays, J.; Ehinger, K.A.; Oliva, A.; Torralba, A. Sun database: Large-scale scene recognition from abbey to zoo. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and PATTERN Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 3485–3492. [Google Scholar]
  44. Yang, S.; Luo, P.; Loy, C.C.; Tang, X. Wider face: A face detection benchmark. arXiv 2016, arXiv:1511.06523. [Google Scholar]
  45. Zeiler, M.D.; Taylor, G.W.; Fergus, R. Adaptive deconvolutional networks for mid and high level feature learning. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 2018–2025. [Google Scholar]
  46. Zeiler, M.D.; Fergus, R. Visualizing and understanding convolutional networks. arXiv 2014, arXiv:1311.2901. [Google Scholar]
  47. Simonyan, K.; Vedaldi, A.; Zisserman, A. Deep inside convolutional networks: Visualising image classification models and saliency maps. In Proceedings of the in Workshop at International Conference on Learning Representations, Banff, Canada, 14–16 April 2014. [Google Scholar]
  48. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning representations by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
  49. Samek, W.; Binder, A.; Montavon, G.; Lapuschkin, S.; Müller, K.R. Evaluating the visualisation of what a deep neural network has learned. IEEE Trans. Neural Netw. Learn. Syst. 2016, 28, 2660–2673. [Google Scholar] [CrossRef] [Green Version]
  50. Ancona, M.; Ceolini, E.; Öztireli, C.; Gross, M. Towards better understanding of gradient-based attribution methods for deep neural networks. arXiv 2017, arXiv:1711.06104. [Google Scholar]
  51. Kindermans, P.J.; Schütt, K.T.; Alber, M.; Müller, K.R.; Erhan, D.; Kim, B.; Dähne, S. Learning how to explain neural networks: Patternnet and patternattribution. arXiv 2017, arXiv:1705.05598. [Google Scholar]
  52. Springenberg, J.T.; Dosovitskiy, A.; Brox, T.; Riedmiller, M. Striving for simplicity: The all convolutional net. arXiv 2014, arXiv:1412.6806. [Google Scholar]
  53. Sundararajan, M.; Taly, A.; Yan, Q. Axiomatic attribution for deep networks. In Proceedings of the International Conference on Machine Learning, PMLR, Sydney, NSW, Australia, 6–11 August 2017; pp. 3319–3328. [Google Scholar]
  54. Mohamed, E.; Sirlantzis, K.; Howells, G. A review of visualisation-as-explanation techniques for convolutional neural networks and their evaluation. Displays 2022, 73, 102239. [Google Scholar] [CrossRef]
Figure 1. Sizes and aspect ratios of objects in the detection dataset.
Figure 1. Sizes and aspect ratios of objects in the detection dataset.
Sensors 22 05596 g001
Figure 2. Estimated number of the anchor boxes with the corresponding mIoU.
Figure 2. Estimated number of the anchor boxes with the corresponding mIoU.
Sensors 22 05596 g002
Figure 3. Prediction results of single head YOLO V3 and Faster R-CNN detectors on two test images. (a,d) Test images. (b,e) Output of YOLO V3. (c,f) Output of Faster R-CNN.
Figure 3. Prediction results of single head YOLO V3 and Faster R-CNN detectors on two test images. (a,d) Test images. (b,e) Output of YOLO V3. (c,f) Output of Faster R-CNN.
Sensors 22 05596 g003
Figure 4. Prediction results of the best performing YOLO V3 detector with three heads.
Figure 4. Prediction results of the best performing YOLO V3 detector with three heads.
Sensors 22 05596 g004
Figure 5. Proposed visualisation approaches using an image from a classification dataset. (a) activation maps generation, (b) Weighted sum of gradients approach (WS-Grad) and (c) Concatenation of gradients approach (Concat-Grad).
Figure 5. Proposed visualisation approaches using an image from a classification dataset. (a) activation maps generation, (b) Weighted sum of gradients approach (WS-Grad) and (c) Concatenation of gradients approach (Concat-Grad).
Sensors 22 05596 g005
Figure 6. Gradients based attribution methods and the proposed methods to visualise the door prediction using YOLO V3. (a) YOLO prediction. (b) Gradient. (c) GBP. (d) IG. (e) WS-Grad (proposed). (f) Concat-Grad (proposed).
Figure 6. Gradients based attribution methods and the proposed methods to visualise the door prediction using YOLO V3. (a) YOLO prediction. (b) Gradient. (c) GBP. (d) IG. (e) WS-Grad (proposed). (f) Concat-Grad (proposed).
Sensors 22 05596 g006
Figure 7. Gradients based attribution methods and the proposed methods to visualise the fire extinguisher prediction using YOLO V3. (a) YOLO prediction. (b) Gradient. (c) GBP. (d) IG. (e) WS-Grad (proposed). (f) Concat-Grad (proposed).
Figure 7. Gradients based attribution methods and the proposed methods to visualise the fire extinguisher prediction using YOLO V3. (a) YOLO prediction. (b) Gradient. (c) GBP. (d) IG. (e) WS-Grad (proposed). (f) Concat-Grad (proposed).
Sensors 22 05596 g007
Table 1. Number of instances per class and the image count.
Table 1. Number of instances per class and the image count.
ClassNo# of InstanceNo# of Images
Door34432548
Pull door handle362227
Push button9292
Moveable door handle20351797
Push door handle437367
Fire extinguisher536536
Key slot826763
ID reader500485
Table 2. Training details (Faster R-CNN with ResNet-18 using mini-batch size of 2).
Table 2. Training details (Faster R-CNN with ResNet-18 using mini-batch size of 2).
ModelFeature
Layer
No# of
Anchors
Training
Time (≈Hours)
Stopping
Epoch
Size (MB)/
No# of
Layers
Validation
Loss
Faster R-CNNres4a_relu3016.31542/820.681
Faster R-CNN *res4a_relu3018.61542/820.719
Faster R-CNNres4b_relu30141742/820.575
Faster R-CNN *res4b_relu3017.51842/820.627
Faster R-CNNres5a_relu3051548.5/820.369
Faster R-CNN *res5a_relu3041148.5/820.399
Faster R-CNNres5b_relu302.51448.5/820.396
Faster R-CNN *res5b_relu3031748.5/820.479
Faster R-CNNres4a_relu2334.7942/820.639
Faster R-CNN *res4a_relu23561442/820.639
Faster R-CNNres4b_relu2341.51942/820.521
Faster R-CNN *res4b_relu23321542/820.529
Faster R-CNNres5a_relu234.51348.4/820.303
Faster R-CNN *res5a_relu2351348.4/820.311
Faster R-CNNres5b_relu232.71548.4/820.318
Faster R-CNN *res5b_relu232.31448.4/820.342
Faster R-CNNres4a_relu317.61041.9/820.366
Faster R-CNN *res4a_relu323.31441.9/820.380
Faster R-CNNres4b_relu3121041.9/820.310
Faster R-CNN *res4b_relu311.11141.9/820.328
Faster R-CNNres5a_relu31948.2/820.192
Faster R-CNN *res5a_relu31.11248.2/820.201
Faster R-CNNres5b_relu30.5648.2/820.336
Faster R-CNN *res5b_relu30.4648.2/820.264
* System trained on augmented data. MB = Megabyte.
Table 3. Training details (YOLO V3 with ResNet-18 using mini-batch size of 16).
Table 3. Training details (YOLO V3 with ResNet-18 using mini-batch size of 16).
ModelFeature
Layer
No# of
Anchors
Training
Time (≈Hours)
Stopping
Epoch
Size (MB)/
No# of
Layers
Validation
Loss
YOLO V3res4a_relu300.92510.6/487.91
YOLO V3 *res4a_relu301.12910.6/486.40
YOLO V3res4b_relu3012414.8/556.89
YOLO V3 *res4b_relu301.12814.8/555.23
YOLO V3res5a_relu301.12541.1/644.68
YOLO V3 *res5a_relu301.22841.1/643.93
YOLO V3res5b_relu301.32757.9/714.20
YOLO V3 *res5b_relu301.53057.9/713.57
YOLO V3res4a_relu230.82410.4/487.64
YOLO V3 *res4a_relu231.12810.4/486.25
YOLO V3res4b_relu230.92314.6/556.31
YOLO V3 *res4b_relu231.12714.6/555.17
YOLO V3res5a_relu231.22740.8/644.37
YOLO V3 *res5a_relu231.33040.8/643.90
YOLO V3res5b_relu231.42957.6/713.96
YOLO V3 *res5b_relu231.53057.6/713.40
YOLO V3res4a_relu30.82410/486.22
YOLO V3 *res4a_relu30.92710/484.73
YOLO V3res4b_relu30.92614.2/554.67
YOLO V3 *res4b_relu30.92514.2/553.75
YOLO V3res5a_relu31.33039.8/643.27
YOLO V3 *res5a_relu31.22839.8/642.85
YOLO V3res5b_relu31.22556.6/712.74
YOLO V3 *res5b_relu31.53056.6/712.52
YOLO V3res4a&5a_relu31.22443.2/733.56
YOLO V3 *res4a&5a_relu31.22443.2/732.95
YOLO V3res4a&5b_relu31.52760/802.95
YOLO V3 *res4a&5b_relu31.52860/802.35
YOLO V3res4b&5a_relu31.32743.2/733.52
YOLO V3 *res4b&5a_relu31.52843.2/732.97
YOLO V3res4b&5b_relu31.52760/802.93
YOLO V3 *res4b&5b_relu31.62760/802.51
YOLO V3res4a&4b&5a_relu31.41646.4/823.79
YOLO V3 *res4a&4b&5a_relu31.83046.4/822.38
YOLO V3res4a&4b&5b_relu31.42363.2/893.28
YOLO V3 *res4a&4b&5b_relu31.72863.2/892.94
YOLO V3res4a&5a&5b_relu31.52472.9/893.27
YOLO V3 *res4a&5a&5b_relu31.72772.9/892.46
YOLO V3res4b&5a&5b_relu31.93072.9/893.32
YOLO V3 *res4b&5a&5b_relu31.93072.9/892.57
* System trained on augmented data. MB = Megabyte.
Table 4. Detectors detailed results on the test set (Faster R-CNN with ResNet-18).
Table 4. Detectors detailed results on the test set (Faster R-CNN with ResNet-18).
ModelFeature
Layer
No# of
Anchors
AP for Each ClassmAP
DoorKey SlotFire
Extinguisher
ID ReaderMoveable
Door Handle
Pull
Door Handle
Push ButtonPush
Door Handle
Faster R-CNNres4a_relu300.9110.0610.5870.3150.3300.3540.0000.2050.345
Faster R-CNN *res4a_relu300.9160.0660.6390.2850.3790.3400.0000.2500.359
Faster R-CNNres4b_relu300.9110.0890.6660.2740.4040.4150.0000.2890.381
Faster R-CNN *res4b_relu300.9050.0880.7090.2150.3940.3790.0000.2380.366
Faster R-CNNres5a_relu300.8100.0000.4070.0710.0810.0000.0000.0290.174
Faster R-CNN *res5a_relu300.7990.0000.3960.0880.0560.0000.0000.0000.167
Faster R-CNNres5b_relu300.5040.0000.1540.0000.0110.0000.0000.0000.083
Faster R-CNN *res5b_relu300.4320.0000.0970.0000.0020.0000.0000.0000.066
Faster R-CNNres4a_relu230.8830.0920.6730.1900.2980.3170.0000.2800.341
Faster R-CNN *res4a_relu230.8900.0860.6330.2030.4110.3510.0000.2440.352
Faster R-CNNres4b_relu230.9240.0680.7420.4260.4270.3840.2790.2220.434
Faster R-CNN *res4b_relu230.9100.1090.6420.2850.4090.3440.1970.2420.392
Faster R-CNNres5a_relu230.7800.0000.3880.1320.0690.0000.0000.0710.180
Faster R-CNN *res5a_relu230.7770.0000.3610.0000.0410.0000.0000.0000.147
Faster R-CNNres5b_relu230.5050.0000.2420.0680.0350.0000.0000.0000.106
Faster R-CNN *res5b_relu230.4310.0000.0880.0000.0090.0000.0000.0000.066
Faster R-CNNres4a_relu30.8730.0180.6920.2110.3900.2020.0000.0820.308
Faster R-CNN *res4a_relu30.8690.0660.7290.1770.3610.3560.0000.1080.333
Faster R-CNNres4b_relu30.9090.1050.6020.4810.3960.3380.0000.1900.378
Faster R-CNN *res4b_relu30.8890.0340.7540.1800.4180.3540.0000.2330.357
Faster R-CNNres5a_relu30.7450.0000.4150.0000.0280.0000.0000.0000.148
Faster R-CNN *res5a_relu30.7490.0000.2570.0000.0330.0000.0000.0000.130
Faster R-CNNres5b_relu30.1230.0000.1160.0000.0040.0000.0000.0000.030
Faster R-CNN *res5b_relu30.5020.0000.1270.0000.0020.0000.0000.0000.079
* System trained on augmented data.
Table 5. Detectors detailed results on the test set (YOLO V3 with ResNet-18).
Table 5. Detectors detailed results on the test set (YOLO V3 with ResNet-18).
ModelFeature
Layer
No# of
Anchors
AP for Each ClassmAP
DoorKey SlotFire
Extinguisher
ID ReaderMoveable
Door Handle
Pull
Door Handle
Push ButtonPush
Door Handle
YOLO V3res4a_relu300.5160.4620.7140.6570.6320.5360.5000.4580.559
YOLO V3 *res4a_relu300.5610.5560.7870.7390.7440.5430.5660.4700.620
YOLO V3res4b_relu300.5730.5100.7880.7510.7160.5660.6660.5380.638
YOLO V3 *res4b_relu300.6700.5460.7970.8070.7790.5750.6890.6480.668
YOLO V3res5a_relu300.7210.1560.8430.6180.5930.5190.8330.4370.590
YOLO V3 *res5a_relu300.7860.2300.8380.6700.6160.5100.7390.4120.600
YOLO V3res5b_relu300.7600.1920.8560.6040.6210.4820.7970.4370.593
YOLO V3 *res5b_relu300.8090.1960.8260.6340.6170.4620.7740.4000.589
YOLO V3res4a_relu230.5630.4240.7200.6870.6330.5780.5660.4070.572
YOLO V3 *res4a_relu230.6090.4740.7540.6800.7120.5390.4930.5050.596
YOLO V3res4b_relu230.6320.4450.7660.7310.6750.6580.6000.4550.620
YOLO V3 *res4b_relu230.6990.4960.8250.7600.7430.5670.6220.5150.653
YOLO V3res5a_relu230.7710.1940.8060.6230.6220.5170.7400.4050.584
YOLO V3 *res5a_relu230.7910.1760.7530.6450.5800.4540.8000.4050.575
YOLO V3res5b_relu230.8090.1920.8260.6170.5920.4770.8330.5270.609
YOLO V3 *res5b_relu230.8210.1660.7770.6540.5910.4710.8660.3620.588
YOLO V3res4a_relu30.6520.5020.8050.7620.7210.6430.7580.5550.674
YOLO V3 *res4a_relu30.7260.5230.8820.8170.7670.5720.8490.6730.726
YOLO V3res4b_relu30.7450.5580.8460.8040.7500.6410.7550.6860.723
YOLO V3 *res4b_relu30.8150.5820.9220.8680.7770.6650.7550.7360.765
YOLO V3res5a_relu30.8340.1950.8590.7120.6920.5620.8610.5110.653
YOLO V3 *res5a_relu30.8550.1880.8770.7250.7280.5010.6910.4830.631
YOLO V3res5b_relu30.8730.2120.8740.7660.7360.5230.6950.5810.657
YOLO V3 *res5b_relu30.8870.1890.8810.7680.7320.5280.8330.5220.667
YOLO V3res4a&5a_relu30.8340.4990.9140.7920.7540.5990.8000.5120.713
YOLO V3 *res4a&5a_relu30.8560.5950.9110.8430.8050.5170.8660.6820.759
YOLO V3res4a&5b_relu30.8790.6360.9150.8560.8210.6170.7690.6880.772
YOLO V3 *res4a&5b_relu30.9020.6140.9080.8650.8430.6300.7770.7430.785
YOLO V3res4b&5a_relu30.8350.5370.9040.8250.7670.6130.7980.5740.731
YOLO V3 *res4b&5a_relu30.8620.5820.9140.9040.8120.6060.8500.6820.776
YOLO V3res4b&5b_relu30.8880.5730.9220.8550.8070.6840.8810.6790.786
YOLO V3*res4b&5b_relu30.8950.6040.9150.8920.8440.6210.7530.7070.778
YOLO V3res4a&4b &5a_relu30.8470.5100.8960.8030.7660.6950.7930.6670.747
YOLO V3 *res4a&4b &5a_relu30.9070.6730.9370.8620.8690.6380.8930.7660.818
YOLO V3res4a&4b &5b_relu30.8920.4540.8960.7730.7590.7240.7570.6070.732
YOLO V3 *res4a&4b &5b_relu30.9170.5710.9410.8870.8310.6550.7750.7300.788
YOLO V3res4a&5a &5b_relu30.8780.5930.9210.7920.7750.6160.7750.6510.750
YOLO V3 *res4a&5a &5b_relu30.8920.5610.9540.8770.8330.6120.7490.6410.764
YOLO V3res4b&5a &5b_relu30.8770.5390.9020.7870.7700.5520.7820.5350.718
YOLO V3 *res4b&5a &5b_relu30.9000.5160.9320.8470.8370.6010.8700.7090.776
* System trained on augmented data.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Mohamed, E.; Sirlantzis, K.; Howells, G.; Hoque, S. Optimisation of Deep Learning Small-Object Detectors with Novel Explainable Verification. Sensors 2022, 22, 5596. https://doi.org/10.3390/s22155596

AMA Style

Mohamed E, Sirlantzis K, Howells G, Hoque S. Optimisation of Deep Learning Small-Object Detectors with Novel Explainable Verification. Sensors. 2022; 22(15):5596. https://doi.org/10.3390/s22155596

Chicago/Turabian Style

Mohamed, Elhassan, Konstantinos Sirlantzis, Gareth Howells, and Sanaul Hoque. 2022. "Optimisation of Deep Learning Small-Object Detectors with Novel Explainable Verification" Sensors 22, no. 15: 5596. https://doi.org/10.3390/s22155596

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop