Next Article in Journal
Optimization Method for Wide Beam Sonar Transmit Beamforming
Next Article in Special Issue
Deep Non-Line-of-Sight Imaging Using Echolocation
Previous Article in Journal
Verifiable Delay Function and Its Blockchain-Related Application: A Survey
Previous Article in Special Issue
Low-Light Image Enhancement Using Hybrid Deep-Learning and Mixed-Norm Loss Functions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Instance-Level Contrastive Learning for Weakly Supervised Object Detection

School of Information and Communication Engineering, University of Electronic Science and Technology of China, No. 2006, Xiyuan Avenue, West Hi-Tech Zone, Chengdu 611731, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(19), 7525; https://doi.org/10.3390/s22197525
Submission received: 2 September 2022 / Revised: 30 September 2022 / Accepted: 3 October 2022 / Published: 4 October 2022
(This article belongs to the Special Issue Deep Learning Technology and Image Sensing)

Abstract

:
Weakly supervised object detection (WSOD) has received increasing attention in object detection field, because it only requires image-level annotations to indicate the presence or absence of target objects, which greatly reduces the labeling costs. Existing methods usually focus on the current individual image to learn object instance representations, while ignoring instance correlations between different images. To address this problem, we propose an instance-level contrastive learning (ICL) framework to mine reliable instance representations from all learned images, and use the contrastive loss to guide instance representation learning for the current image. Due to the diversity of instances, with different appearances, sizes or shapes, we propose an instance-diverse memory updating (IMU) algorithm to mine different instance representations and store them in a memory bank with multiple representation vectors per class, which also considers background information to enhance foreground representations. With the help of memory bank, we further propose a memory-aware instance mining (MIM) algorithm that combines proposal confidence and instance similarity across images to mine more reliable object instances. In addition, we also propose a memory-aware proposal sampling (MPS) algorithm to sample more positive proposals and remove some negative proposals to balance the learning of positive-negative samples. We conduct extensive experiments on the PASCAL VOC2007 and VOC2012 datasets, which are widely used in WSOD, to demonstrate the effectiveness of our method. Compared to our baseline, our method brings 14.2 % mAP and 13.4 % CorLoc gains on PASCAL VOC2007 dataset, and 12.2 % mAP and 8.3 % CorLoc gains on PASCAL VOC2012 dataset.

1. Introduction

Object detection is a fundamental task in computer vision, which requires to identify object categories and use bounding boxes to locate their complete region positions. With the development of convolutional neural network (CNN) [1,2,3], some object detection methods [4,5,6,7,8,9,10,11,12,13], such as Fast R-CNN [4], Faster R-CNN [5], SSD [6] and YOLO [7], have made significant progress. However, these methods require fully supervised information, i.e., instance-level annotations, which are time-consuming and labor-intensive to label. To reduce the burden of annotations, weakly supervised object detection (WSOD) removes bounding boxes and only requires image-level annotations, i.e., image tags, to indicate whether object categories are present in an image.
Due to lack of object bounding box position supervision, most current WSOD methods [14,15,16,17,18,19,20,21,22,23,24,25,26,27] use multiple instance learning (MIL) [28] to mine object instances from pre-generated proposals, and treat them as pseudo instance-level annotations to train weakly supervised detectors. However, these methods only focus on a single image to learn object representations without considering the internal relevance of various object instances across images. When there are object appearance variations in the complex diverse image scenes, it is easy to cause false detection. For example, when a horse is occluded in Figure 1, it focuses more on local feature representation, which is not enough to represent the whole object, resulting in the learned object instance only covering the head of the horse.
To deal with the problem, we propose an instance-level contrastive learning (ICL) framework to store reliable instance representations from all learned images, and utilize contrastive learning [29] mechanism to explicitly establish semantic correlations with other image instances. It attempts to enhances the discriminative and robustness of object instance representations in the current input image, pulling it close to the instance representations of same class from all training images and pushing it away instance representations of different classes. As shown in Figure 1b, owning to instance correlation with other images, our method can effectively learn the instance representation for the whole horse.
To sufficiently represent diversity instances in all training data, we next propose an instance-diverse memory updating (IMU) algorithm. It mines reliable instance representations from proposal features and builds a memory bank with multiple representation vectors for each class to store them based on similarity, where background information is also consider to enhance foreground representations. Based on the memory bank, we further propose a memory-aware instance mining (MIM) algorithm. Unlike most methods [14,15,20,21,22,26,27] that mine object instances only based on proposal confidence, we also compute the similarity with stored diverse instances to evaluate the completeness of proposals to mine more reliable object instances. Instead of selecting the top-scoring proposal as an instance, we also consider the multi-instance case to mine more instances in an image. During the training process of weakly supervised object detectors, we propose a memory-aware proposal sampling (MPS) algorithm to alleviate the imbalance problem between positive and negative samples. According to the similarity with instance representations, we select more positive proposals to increase the number of positive samples. Based on the similarity with the background information, we remove some negative proposals with low similarity to reduce the number of negative samples.
To verify the effectiveness of our method, we conduct extensive experiments on the PASCAL VOC2007 and VOC2012 datasets, which are widely used for weakly supervised object detection. In this paper, we adopt typical WSOD method OICR [15] as our baseline, which can be easily embedded into our ICL framework to further improve the performance. On PASCAL VOC2007 dataset, our method improves detection performance and localization accuracy by 14.2 % and 13.4 % in terms of mAP and CorLoc, respectively. On PASCAL VOC2012 dataset, our method improves performance by 12.2 % mAP and 8.3 % CorLoc.
The contributions of this paper are summarized as follows:
  • We propose an instance-level contrastive learning (ICL) framework to guide the weakly supervised detector to learn instance representations. To the best of our knowledge, we are the first to explore contrastive learning in weakly supervised object detection.
  • We propose an instance-diverse memory update (IMU) algorithm to store reliable instance representations into a memory bank, where multiple representation vectors are used in each class to maintain the diversity of instance representations.
  • With the help of memory, we further propose a memory-aware instance mining (MIM) algorithm to efficiently mine object instances by combining proposal confidence and instance similarity.
  • With the help of memory, we also propose a memory-aware proposal sampling (MPS) algorithm to alleviate the imbalance between positive and negative samples by finding more positive proposals and removing some unreliable negative proposals.

2. Related Work

In this section, we present the two most relevant to this paper, weakly supervised object detection and contrastive learning.

2.1. Weakly Supervised Object Detection

Since Hakan Bilen and Andrea Vedaldi proposed a weakly supervised deep detection network (WSDDN) [20] to combine MIL and CNN into an end-to-end network, most WSOD methods follow [14,15,16,17,18,19,21,22,26,27] this pipeline to train the weakly supervised detector. In MIL, an image is treated as a bag of proposals. If the image contains an object class, this bag is labelled as positive bag, i.e., at least containing one object instance of this class, otherwise labelled as negative bag. Due to lack of instance-level annotations, MIL is tend to get stuck in a local optimum to locate the most representative part of target objects. Subsequently, most researchers have proposed promising approaches to alleviate this problem. For instance, Kantorov et al. [14] proposed additional and contrastive context-aware guidance models to improve localization by using the surrounding contextual region of proposals. Tang et al. [15] proposed an online instance classifier refinement (OICR) method that uses spatial correlations between proposals to refine mined instances. Wan et al. [21] proposed continuation multiple instance learning (C-MIL) to alleviate the problem that MIL is prone to falling into local optima, which uses some smooth loss function to approximate the original non-convex loss function. Lin et al. [22] proposed object instance mining (OIM) framework to build spatial and appearance graphs of proposals to mine all possible object instances. Furthermore, some methods [16,17,18,19] introduce segmentation information to assist instance mining. Shen et al. [16] proposed a recurrent guidance strategy for weakly supervised detection and segmentation, where the detection module generates seeds for semantic segmentation and the segmentation module provides prior information for object detection. Yang et al. [17] proposed an objectness consistent representation method to exploit segmentation map to mine more high-quality proposals. Wei et al. [18] used segmentation context information around proposals to discover tight object bounding boxes. Li et al. [19] leveraged the segmentation map to reweight proposals scores. However, these methods only consider information from a single image, which are difficult to deal with diverse object instances. In this paper, our method explores the semantic correlation beyond the input image to assist object instance mining.

2.2. Contrastive Learning

Contrastive learning [29] has been widely used in unsupervised representation learning (e.g., SimCLR [30] and MoCo [31]), which compares positive and negative pairs to compress together different view representations of the same image and separate view representations of different images. In addition, contrastive learning-based methods [32,33,34,35] have also achieved promising performance in other vision tasks. For instance, Yan et al. [33] proposed a semantics-guided contrastive network that introduces contrastive learning into zero-shot object detection to transfer available semantic information for unseen classes. Wu et al. [34] proposed a contrastive learning-based robust object detection algorithm to detect objects under smoky conditions, which applies contrastive learning to maximize the consistency between different augmented views of the same smoke image. Li et al. [35] introduced contrastive learning into remote sensing image semantic segmentation to learn global and local image representations. However, these methods are difficult to directly apply to WSOD that learns the detector based on image-level annotations. In this paper, we introduce contrastive learning into weakly supervised object detection and propose an instance-level contrastive learning framework. To our best knowledge, we are the first to explore contrastive learning for weakly supervised object detection.

3. Method

In this section, we first describe our instance-level contrastive learning (ICL) framework in detail. Then, we present the instance-diverse memory updating (IMU) algorithm, memory-aware instance mining (MIM) algorithm and memory-aware proposal sampling (MPS) algorithm.

3.1. Instance-Level Contrastive Learning

In Figure 2, we present the pipeline of the instance-level contrastive learning (ICL) framework. Given an input image I and the corresponding proposals R generated by the proposal generation methods [36,37,38], we first extract image features F I using convolutional neural network. Based on the pre-generated proposals R, we convert image features F I to proposal features F R through a RoI-pooling layer, and use two fully connected (FC) Layers to obtain proposal vector representations F V . By mining reliable instance representations from F V , we then perform contrastive learning (CL) and store them in the memory bank M. In addition, F V is fed into several parallel detection heads, where a base head is supervised by the image label Y = [ y 1 , y 2 , , y C ] T R C × 1 and K refined heads are supervised by the output results of previous heads, where C is the number of classes. In this paper, we set K = 3 to be the same as our baseline method [15].
Unsupervised representation learning [30,31] performs contrastive learning by augmenting image to different views, where views of the same image are pulled closer and views of different images are pulled apart. In this paper, we introduce the contrastive learning of object instance representations to guide the detector to learn the entire representation of the instance. Specifically, we first denote all outputs of refined heads as ( { φ 1 , φ 2 , , φ K } , { t 1 , t 2 , , t K } ) . To mine more reliable instances, we average these outputs to obtain the proposal scores φ = 1 K k = 1 K φ k and the bounding box coordinate offsets t = 1 K k = 1 K t k . Applying the coordinate offset t to transform R, we obtain the transformed proposal P. Then, we exploit non-maximum suppression (NMS) to mine as many object instances as possible. For a positive class c, we use NMS to gradually select object instances from the transpose proposal P c according to the proposal score φ c from high to low, and remove redundant proposals. Then, we set a score threshold T 1 to obtain more reliable object instances D, and extract the corresponding instance feature representations F D from F V . The detailed procedure can be seen in Algorithm 1. For each mined instance representation q F D , we utilize the memory instance representation M including all training data to assist each instance learning in current image. Assume that a positive representation from memory bank k + M represent the same class as q, and a negative representation from memory bank k M represent different classes. Then, we use the contrastive loss [39] to pull q close to k + of the same class while pushing it away from negative keys k of other classes, and thus enhance the discrimination and generalization of current instance representation:
L C L = 1 | D | q F D φ q l o g e x p ( q · k + / τ ) e x p ( q · k + / τ ) + k e x p ( q · k / τ ) ,
where we take φ q as the loss weight and τ means the temperature hyperparameter.
Algorithm 1 Instance representation mining algorithm.
Input:  The pre-generated proposals R, the pre-defined score threshold T 1 , the image label Y, the memory bank M, the outputs of instance refined heads ( { φ 1 , φ 2 , . . . , φ K } , { t 1 , t 2 , , t K } ) and the proposal feature vectors F V .
   (I) average proposal scores φ = 1 K k = 1 K φ k
   (II) average coordinate offsets t = 1 K k = 1 K t k
   (III) obtain transformed proposals P by adding t to R
   (IV) instance representations F D = and the corresponding confidences φ F D =
   For  c = 0 to C + 1
      If  y c = = 1 or c = = C + 1
         (1) k e e p = N M S ( P c , φ c )
         (2) P k e e p = P c [ k e e p ]
         (3) φ k e e p = φ [ k e e p ]
         (4) D c = P k e e p [ φ k e e p > T 1 ]
         (5) φ D c = φ k e e p [ φ k e e p > T 1 ]
         (6) F D c = F V [ D c ]
         (7) F D = F D F D c , φ F D = φ F D φ D c
Output: F D , φ F D .
Subsequently, we describe the training of heads. The base head has two parallel branches. One branch uses an FC layer to generate a matrix x c R C × | R | ( | R | is the number of proposals), which is then input to a class-wise softmax layer: [ σ ( x c ) ] i j = e x i j c q = 1 C e x q j c . In another branch, there is an FC layer and a proposal-wise softmax layer to generate another normalized matrix σ ( x r ) , where x r R C × | R | and [ σ ( x r ) ] i j = e x i j r q = 1 | R | e x i q r . Then, element-wise matrix multiplication is performed on these two matrices to generate proposal scores x R = σ ( x c ) σ ( x r ) . Finally, the image class score is calculated by summing all proposal scores: ϕ = r = 1 | R | x R . According to the image label Y = [ y 1 , y 2 , . . . , y C ] T , the loss of base head is computed by Equation (2).
L b = c = 1 C { y c log ϕ c + ( 1 y c ) log ( 1 ϕ c ) } .
For K refined heads, their training process is consistent. Specifically, for the k t h head, there is a classifier and a regressor. In the classifier, an FC layer and a class-wise softmax are used to generate proposals scores φ k R ( C + 1 ) × | R | , where C + 1 means background is included. In the regressor, an FC layer is used to produce the coordinate offsets of proposals t k R 4 C × | R | , where 4 means the dimension of coordinate offsets ( x 1 , y 1 , x 2 , y 2 ) . In order to generate their supervision, we first use the memory-aware instance mining (MIM) algorithm to mine multiple representative object instances B k based on the score outputs and offset outputs of previous head ( φ k 1 , t k 1 ) and memory bank M. The details can be seen in Section 3.3. Then, we use the memory-aware proposal sampling (MPS) algorithm of Section 3.4 to sample negative and positive proposals ( R p o s , R n e g ) from proposals R and assign labels for these proposals. For a positive class c, if a proposal p is selected as positive sample, p is labeled as class c, i.e., y c , p k = 1 . All negative proposals R n e g are labeled as background class C + 1 . In this way, the classifier can be trained by a cross entropy loss:
L c l s k = 1 | R p o s | + | R n e g | p R p o s R n e g c = 1 C + 1 φ p k y c , p k log φ c , p k ,
where we also use the confidence φ p k as the loss weight. For the regressor, only positive proposals R p o s are used to calculate loss by the smooth L1 loss [4]:
L r e g k = smoothL 1 ( t k , T k ) ,
where T k is the supervision of coordinate offsets.
In summary, our ICL framework can be end-to-end trained in Equation (5).
L = L C L + L b + k = 1 K ( L c l s k + L r e g k ) .

3.2. Instance-Diverse Memory Updating Algorithm

In order to enable the network to memory the instance representations from previous training images, we first initialize M R ( C + 1 ) × N × L , where N means the number of stored instance feature representations in each class and L represents the length of the feature vector F V . Since instances of the same class differ in size, shape, and appearance, we use multiple feature vectors to store richer instance representations instead of a single vector. We first use the Algorithm 1 to obtain some reliable instance representations F D from F V . For each instance representation f d , c F D , we calculate the similarity between f d , c and with M c = { f c , 1 , f c , 2 , , f c , N } in Equation (6).
S d , c = | | f d , c | | × | | M c T | | ,
where | | · | | , × and T mean L 2 normalization, matrix multiplication and transpose, respectively. Then, we select the most similarity feature f c , j from M c in Equation (7) to maximize the assistance of the current instance.
j = a r g m a x { S d , c } ,
Finally, we update the feature vector f c , j according Equation (8) for the instance contrastive learning of subsequent images.
f c , j = r f c , j + ( 1 r ) φ f d , c f d , c ,
where r is the momentum coefficient [31] and φ f d , c is the confidence of instance representation, which aim to control the weight balance between previous instance representation and current representation. The whole process can be seen in the Algorithm 2.
Algorithm 2 Instance-diverse memory updating algorithm.
Input: The pre-generated proposals R, the pre-defined score threshold T 1 , the image label Y, the memory bank M, the outputs of instance refined heads ( { φ 1 , φ 2 , , φ K } , { t 1 , t 2 , , t K } ) and the proposal feature vectors F V .
   (I) obtain reliable instance representations F D using Algorithm 1
   For each representation f d , c in F D
      (a) compute the similarity between f d , c and with M c in Equation (6)
      (b) choose the most similarity feature f c , j from M c in Equation (7)
      (c) update f c , j in Equation (8)
Output: Updated memory M.

3.3. Memory-Aware Instance Mining Algorithm

With the help of memory bank M, we propose a memory-aware instance mining (MIM) algorithm to effectively mine some reliable object instances. Different from our baseline [15], which only selects the top-scoring proposal as pseudo instance annotations, we comprehensively consider the confidence of proposals and the similarity between proposal features and memory bank covering previous training data to effectively mine object instances. Specifically, we first calculate the similarity S between F V and M according to Equation (9).
S = | | F V | | × | | M T | | .
Then, we select the highest similarity along the N feature vectors and apply the class-wise softmax to generate memory-base confidence φ M through Equation 10.
φ M = s o f t m a x ( m a x N { S } ) .
For the k t h branch in instance refinement heads, we further calculate the combination confidence ψ k in Equation (11).
ψ k = φ k + μ φ M ,
where μ is the combination coefficient. Next, we use the NMS algorithm to remove redundant proposals and set a score threshold T 2 to remove unreliable proposals. In this way, we can obtain some reliable instances B k . More details can be found in Algorithm 3.
Algorithm 3 Memory-aware instance mining algorithm.
Input: The pre-generated proposals R, the pre-defined score threshold T 2 , the image label Y, the memory bank M, the outputs of k t h instance refined head ( φ k , t k ) and the proposal feature vectors F V .
   (I) obtain transformed proposals R t k by adding t k to R
   (II) calculate the memory-based confidence φ M by Equation (10)
   (III) compute the combination confidence ψ k with Equation (11)
   For  c = 0 to C
      If  y c = = 1
         (1) k e e p = N M S ( R t k , ψ c k )
         (2) R 1 = R t k [ k e e p ]
         (3) ψ 1 = ψ c k [ k e e p ]
         (4) B k = R 1 [ φ 1 > T 2 ]
         (5) ψ B k k = φ 1 [ φ 1 > T 2 ]
Output: B k , ψ B k k .

3.4. Memory-Aware Proposal Sampling Algorithm

After mining object instances, we further propose a memory-aware proposal sampling (MPS) algorithm to effectively sample positive and negative proposals from R. Some methods [15,16,17,18,19,26] simply divide R into two parts by computing the IoU with B k : highly overlapped proposals are taken as positive samples, and the rest are taken as negative samples, while ignoring the imbalance of positive and negative samples with overwhelmingly negative proposals. To alleviate this problem, we leverage the memory bank to select more positive proposals and remove some unreliable negative proposals. We first calculate the IoU between B k and R to separate R into two parts R 1 k and R 2 k in Equation (12).
R 1 k = p R | b B k , I o U ( p , b ) 0.5 R 2 k = p R | b B k , 0.1 < I o U ( p , b ) < 0.5
Then, we extract the feature representations F R 2 k of R 2 k from F V . For each positive class c, we compute the similarity between F R 2 k and M c , and use Equation (13) to choose the most similar proposal p c into R 1 .
j = a r g m a x ( | F R 2 k | × | M c T | ) .
For the remaining proposals in R 2 k , we compute the similarity S R 2 k = m a x N { | F R 2 k | × | M c T | } between R 2 k and background information M C + 1 and sort S R 2 k according to similarity from high to low. Finally, We removed the last low-similarity 1 / λ proposals from R 2 k to obtain the negative samples. More details can be found in Algorithm 4.
Algorithm 4 Memory-aware proposal sampling algorithm.
Input: The pre-generated proposals R, the image label Y, the memory bank M, the mined object instance ( B k , ψ B k k ) and the proposal feature vectors F V .
   (I) positive samples R p o s = , negative samples R n e g =
   (II) calculate I o U ( B k , R )
   (III) separate R into two parts R 1 k and R 2 k using Equation (12).
   (IV) R p o s = R p o s R 1 k
   (V) extract feature representations F R 2 k from F V
   For  c = 0 to C
      If  y c = = 1
         (1) calculate the similarity between F R 2 k and M c
         (2) select the most similar proposal p c from R 2 k by Equation (13)
         (3) R p o s = R p o s p c , R 2 k = R 2 k / p c , F R 2 k = F R 2 k / F p c
   (VI) S R 2 k = m a x N { | F R 2 k | × | M c T | }
   (VII) sort S R 2 k from from high to low
   (VIII) obtain R n e g by removing the last low-similarity 1 / λ proposals from R 2 k
Output:  R p o s , R n e g .

3.5. Test

After training, only the instance refined heads are used for testing. We perform the same operations as our baseline method [15]. We average the outputs of all refined heads to generate the final detection results.

4. Experiments

In this section, we first introduce experimental data and evaluation criteria, and elaborate on experimental details. Then we validate the advantages of our method by comparing with some recent methods. Finally, we conduct extensive ablation experiments to demonstrate the effectiveness of our method.

4.1. Datasets and Evaluation Measures

We conduct experiments on PASCAL VOC2007 [40] and VOC2012 [41] datasets, which are widely used in weakly supervised object detection setting [14,15,16,17,18,19,20,21,22,25,26,27]. In the PASCAL VOC2007 dataset, there are 9962 images belonging to 20 categories. These images are divided into three sets: train, val, test. According to the widely used WSOD setting, the trainval set (5011 images) is used for training. The PASCAL VOC2012 dataset has 22531 images split into train, val and test sets. The trainval set has 11540 images for training. It is important to note that all experiments have only image-level labels for training. For evaluation, there are two evaluation measures mean average precision (mAP [40]) and correct localization (CorLoc [42]). mAP is the standard PASCAL VOC protocol, which first computes the average precision (AP) for each class and then averages over all classes. AP for each class is obtained by calculating the area under the precision-recall curve. The mAP is used to evaluate performance on the test set. The second metric CorLoc is used to measure the localization accuracy of the trainval set. For each class, CorLoc is calculated as the ratio of images where at least one object is correctly localized. Both mAP and CorLoc are based on the PASCAL criterion. The object is considered to be successfully detected, when the intersection over union (IoU) between the ground-truth and predicted boxes is greater than 0.5 .

4.2. Experimental Details

All experiments are performed on the Detectron2 (https://github.com/facebookresearch/detectron2) deep learning framework and 4 NVIDIA GTX 1080ti GPUs. Following our baseline method OICR [15], we use the the VGG16 [2] model as our backbone, pre-trained on the ImageNet dataset [43]. For pre-generated proposals, we use multiscale combinatorial grouping (MCG) method [38] to generate approximately 2000 proposals per image. During the training phase, we set the learning rate to 0.001 for the first 28 epochs and divide it by 10 for the next 12 epochs. In addition, we set the momentum and weight decay to 0.9 and 0.0005, respectively. The mini-batch size is set to 4, i.e., an image is run by one GPU. Regarding the data augmentation, we use 5 scales {480, 576, 688, 864, 1200} to randomly resize the shortest side of the image and make the longest side no more than 2000, where the random horizontal flips are also used. During the test stage, We average the output of all augmented data to generate final detection results. For Hyperparameters in our method, we set K = 3 , N = 5 , μ = 0.1 , 1 / λ = 1 / 4 , and T 1 = T 2 = 0.5 . All settings in the PASCAL VOC2007 and VOC2012 datasets are the same.

4.3. Comparison with Other Methods

On PASCAL VOC2007 and VOC2012 datasets, we compare our method with some recent methods [14,15,16,17,18,19,20,21,22,25,26,27] to present our advantages.
For the PASCAL VOC2007 dataset, we present the detection performance (mAP) and localization accuracy (CorLoc) in Table 1 and Table 2, respectively. In terms of mAP, our method ICL achieves a detection performance of 55.4 % , which brings a significant improvement (about 14.2 % ) compared to our baseline OICR [15] ( 41.2 % ). Our method also outperforms methods [16,17,18,19] that exploit segmentation information to learn instance representation. For example, compare with [17], our method has an advantage of about 4.8 % . Furthermore, our method also has some improvements (about 1.9 % ) compared to recent methods SLV[27] and D-MIL [25]. In terms of CorLoc, our method ICL achieves 74.0 % localization accuracy. Compared with our baseline ( 60.6 % ), our method improves the performance by about 13.4 % . Compared to the segmentation-assisted method WS-JDS [16] or SDCN [19], our method improves the performance by more than 7.2 % . In addition, our method also shows significant advantages (more than 3 % ) compared to recent methods SLV [27] and D-MIL [25].
For the PASCAL VOC2012 dataset, we show both detection performance (mAP) and localization accuracy (CorLoc) in the Table 3. Our method achieves 50.1 % mAP and 70.4 % CorLoc, which are 12.2 % and 8.3 % improvement over the baseline, respectively. Compare to some recent methods [25,27], our method also bring some gains. In terms of mAP, our method outperforms the methods [27] and [25] by 0.9 % and 0.5 % , respectively. In terms of CorLoc, our method brings gains of 1.2 % and 0.3 % , respectively. These results further demonstrate the effectiveness of our method.

4.4. Ablation Study

In this part, we conduct extensive experiments to further discuss the effects of main components of our method. Without loss of generality, all experiments are performed on the PASCAL VOC2007 dataset.
The effect of IMU algorithm. We first analyze the effect of IMU algorithm on our method ILC. In Table 4, after removing IMU, our method achieves 53.6 % mAP and 72.9 % CorLoc, respectively. There are 1.8 % performance reduction and 1.1 % accuracy reduction in terms of mAP and CorLoc, respectively, which proves the effectiveness of the instance-diverse memory updating algorithm. Furthermore, we analyze the effect of the number of feature vector N on the IMU algorithm in Figure 3. We can see that both mAP and CorLoc first increase and then decrease as N increases. When N is too small, the memory bank is difficult to store the diversity of instance representations well, and when N is too large, it is easy to cause there are internal differences during the learning of instance representations. In this paper, we recommend setting N = 5 to balance the number of stored instance vectors.
The effect of MIM algorithm. As shown in the Table 4, MIM brings 5.7 % and 4.4 % gains to ICL in mAP and CorLoc, which shows the effectiveness of memory-aware instance mining algorithm (MIM). In addition, we also analyze the effect of memory on MIM by setting different combination coefficients μ in Figure 4. During the change of μ from 0 to 1, we can see that both the detection performance and the localization accuracy are the highest at μ = 0.1 , which demonstrates that it is useful to introducing the similarity between memory features of previous training data and proposal features for mining effective instances. When μ becomes larger, the memory from previous images may hinder the learning of new instances from the current image, resulting in performance degradation. Therefore, we set μ = 0.1 in this paper.
The effect of MPS algorithm. Removing MPS from ICL, our method achieves 52.9 % detection performance (mAP) and 71.4 % localization accuracy (CorLoc). There are 2.5 % and 2.6 % reductions in mAP and CorLoc, respectively, which shows the effectiveness of the memory-aware proposal sampling algorithm. In addition, we analyze the effect of the removal coefficient 1 / λ on MPS in Table 5. We achieve the best performance when 1 / λ = 1 / 4 . Continuing to increase 1 / λ may remove too many negative samples and affect the training of the detector. When 1 / λ is too small, it cannot achieve the purpose of balancing positive and negative samples.
The performance of COCO metrics. In Table 4, we also analyze the contribution of each component under the COCO metrics [44]. The performance of each component on AP, AP 50 , and AP 75 is similar to that under the PASCAL metrics. For AP S , AP M and AP L , the objects are divided into three sizes of small, medium and large for evaluation. Our method ICL can achieve the best performance on large objects, while removing MPS algorithm can achieve better performance on small and medium objects. Compared with small and medium-sized objects that are difficult to perceive, MPS algorithm is more conducive to sampling region proposals of large objects.
The analysis of training process. In Figure 5, we further provide training loss curves to verify the rationality of our method. We can see that the loss curves of refined heads rise first and then decrease to convergence. The rising phase of the loss is due to the weight, which is the confidence of mined object instances. At the beginning of training, the low discrimination of the model makes the confidence of object instances very low (almost close to 0). As model capabilities increase, the confidence starts to increase and so does the loss. Since confidence range is [ 0 , 1 ] , the loss will reach the maximum value, and finally start to decrease due to the enhanced model generalization until it converges. In Figure 6, we also provide the performance of the model during training. Both mAP and CorLoc continue to increase, which further demonstrates the effectiveness of our method.
Qualitative results. In Figure 7, we provide qualitative results to more intuitively compare the proposed ICL with our baseline. On the trainval set of PASCAL VOC2007 dataset, we compare the learned object instances in Figure 7a. For the simple image in the first column, the baseline method can learn effective information about the car well. For the horse in the second column and the cow in the third column, when the foreground and background are relatively similar or the objects are occluded, the instances learned by the baseline method may contain more background. Our method ICL can learn more reliable object instances guided by instance correlations. On the test set, we compare the detection results in Figure 7b. Since the baseline method is more easily disturbed by background information during the training process, its detection results also contain more background information, such as the first two columns. Our method can better locate the boundary of the object. When there is interaction between objects, such as the third column, our method can also provide better detection results. For a more comprehensive analysis of our method, we present failure cases in the last column. For example, for the smaller aeroplane in Figure 7a, it is difficult for our method to learn its instance representation. Our method also fails to detect the highly overlapping sheep and distant little sheep in Figure 7b.

5. Conclusions

In this paper, we propose an instance-level contrastive learning (ICL) framework to guide the weakly supervised detector to learning entire instance representations by constructing instance correlations with other images. To store diverse object instance representations in a memory bank, we propose an instance-diverse memory updating (IMU) algorithm. With the help of memory, we further propose a memory-aware instance mining (MIM) algorithm to effectively mine object instances. To alleviate the imbalance of positive and negative proposals, we propose a memory-aware proposal sampling (MPS) algorithm. We conduct extensive experiments on PASCAL VOC2007 and VOC2012 datasets to verify the effectiveness of our method.
Our proposed method mines object instance representations from other images and stores them in a memory bank to guide instance learning on the current image. If the memory contains noisy representations, it will make the learned object instances inaccurate. The performance of weakly supervised detectors is also limited by the quality of the stored representations. In order to mine more reliable instance representations, our future studies will explore contextual information of region proposals or segmentation information of images to perceive object boundaries and locate object instances accurately.

Author Contributions

Conceptualization, M.Z.; methodology, M.Z.; software, M.Z.; validation, M.Z.; formal analysis, M.Z.; investigation, M.Z.; resources, M.Z.; data curation, M.Z.; writing—original draft preparation, M.Z.; writing—review and editing, B.Z.; visualization, M.Z.; supervision, B.Z.; project administration, M.Z.; funding acquisition, B.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported in part by the National Natural Science Foundation of China under Grant 61720106004, and in part by the 111 Projects under Grant B17008.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The PASCAL VOC2007 and VOC2012 datasets can be available in the following link: http://host.robots.ox.ac.uk/pascal/VOC/.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, Spain, 3–8 December 2012; pp. 1097–1105. [Google Scholar]
  2. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  3. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  4. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 13–16 December 2015; pp. 1440–1448. [Google Scholar]
  5. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; pp. 91–99. [Google Scholar]
  6. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single shot multibox detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 21–37. [Google Scholar]
  7. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  8. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  9. Tian, Z.; Shen, C.; Chen, H.; He, T. Fcos: Fully convolutional one-stage object detection. In Proceedings of the IEEE International Conference on Computer Vision, Long Beach, CA, USA, 16–20 June 2019; pp. 9627–9636. [Google Scholar]
  10. Qiu, H.; Li, H.; Wu, Q.; Shi, H. Offset bin classification network for accurate object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–18 June 2020; pp. 13188–13197. [Google Scholar]
  11. Zhao, H.; Zhou, Y.; Zhang, L.; Peng, Y.; Hu, X.; Peng, H.; Cai, X. Mixed YOLOv3-LITE: A lightweight real-time object detection method. Sensors 2020, 20, 1861. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Lian, J.; Yin, Y.; Li, L.; Wang, Z.; Zhou, Y. Small object detection in traffic scenes based on attention feature fusion. Sensors 2021, 21, 3031. [Google Scholar] [CrossRef]
  13. Xiang, Y.; Zhao, B.; Zhao, K.; Wu, L.; Wang, X. Improved Dual Attention for Anchor-Free Object Detection. Sensors 2022, 22, 4971. [Google Scholar] [CrossRef]
  14. Kantorov, V.; Oquab, M.; Cho, M.; Laptev, I. Contextlocnet: Context-aware deep network models for weakly supervised localization. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 350–365. [Google Scholar]
  15. Tang, P.; Wang, X.; Bai, X.; Liu, W. Multiple instance detection network with online instance classifier refinement. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Venice, Italy, 22–29 October 2017; pp. 2843–2851. [Google Scholar]
  16. Shen, Y.; Ji, R.; Wang, Y.; Wu, Y.; Cao, L. Cyclic guidance for weakly supervised joint detection and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 697–707. [Google Scholar]
  17. Yang, K.; Zhang, P.; Qiao, P.; Wang, Z.; Dai, H.; Shen, T.; Dou, Y. Rethinking segmentation guidance for weakly supervised object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 946–947. [Google Scholar]
  18. Wei, Y.; Shen, Z.; Cheng, B.; Shi, H.; Xiong, J.; Feng, J.; Huang, T. Ts2c: Tight box mining with surrounding segmentation context for weakly supervised object detection. In Proceedings of the European Conference on ComputerbVision, Salt Lake City, UT, USA, 18–23 June 2018; pp. 434–450. [Google Scholar]
  19. Li, X.; Kan, M.; Shan, S.; Chen, X. Weakly supervised object detection with segmentation collaboration. In Proceedings of the IEEE International Conference on Computer Vision, Long Beach, CA, USA, 16–20 June 2019; pp. 9735–9744. [Google Scholar]
  20. Bilen, H.; Vedaldi, A. Weakly supervised deep detection networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2846–2854. [Google Scholar]
  21. Wan, F.; Liu, C.; Ke, W.; Ji, X.; Jiao, J.; Ye, Q. C-mil: Continuation multiple instance learning for weakly supervised object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 2199–2208. [Google Scholar]
  22. Lin, C.; Wang, S.; Xu, D.; Lu, Y.; Zhang, W. Object instance mining for weakly supervised object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 11482–11489. [Google Scholar]
  23. Xu, Y.; Zhou, C.; Yu, X.; Xiao, B.; Yang, Y. Pyramidal multiple instance detection network with mask guided self-correction for weakly supervised object detection. IEEE Trans. Image Process. 2021, 30, 3029–3040. [Google Scholar] [CrossRef]
  24. Wu, Z.; Wen, J.; Xu, Y.; Yang, J.; Li, X.; Zhang, D. Enhanced Spatial Feature Learning for Weakly Supervised Object Detection. IEEE Trans. Neural Netw. Learn. Syst. 2022. [Google Scholar] [CrossRef] [PubMed]
  25. Gao, W.; Wan, F.; Yue, J.; Xu, S.; Ye, Q. Discrepant multiple instance learning for weakly supervised object detection. Pattern Recognit. 2022, 122, 108233. [Google Scholar] [CrossRef]
  26. Tang, P.; Wang, X.; Bai, S.; Shen, W.; Bai, X.; Liu, W.; Yuille, A. Pcl: Proposal cluster learning for weakly supervised object detection. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 42, 176–191. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Chen, Z.; Fu, Z.; Jiang, R.; Chen, Y.; Hua, X.S. Slv: Spatial likelihood voting for weakly supervised object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–18 June 2020; pp. 12995–13004. [Google Scholar]
  28. Dietterich, T.G.; Lathrop, R.H.; Lozano-Pérez, T. Solving the multiple instance problem with axis-parallel rectangles. Artif. Intell. 1997, 89, 31–71. [Google Scholar] [CrossRef] [Green Version]
  29. Hadsell, R.; Chopra, S.; LeCun, Y. Dimensionality reduction by learning an invariant mapping. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, New York, NY, USA, 17–22 June 2006; pp. 1735–1742. [Google Scholar]
  30. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the International Conference on Machine Learning, Virtual Event, 13–18 July 2020; pp. 1597–1607. [Google Scholar]
  31. He, K.; Fan, H.; Wu, Y.; Xie, S.; Girshick, R. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–18 June 2020; pp. 9729–9738. [Google Scholar]
  32. Sun, B.; Li, B.; Cai, S.; Yuan, Y.; Zhang, C. Fsce: Few-shot object detection via contrastive proposal encoding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Virtual Event, 19–25 June 2021; pp. 7352–7362. [Google Scholar]
  33. Yan, C.; Chang, X.; Luo, M.; Liu, H.; Zhang, X.; Zheng, Q. Semantics-guided contrastive network for zero-shot object detection. IEEE Trans. Pattern Anal. Mach. Intell. 2022. [Google Scholar] [CrossRef] [PubMed]
  34. Wu, W.; Chang, H.; Zheng, Y.; Li, Z.; Chen, Z.; Zhang, Z. Contrastive Learning-Based Robust Object Detection Under Smoky Conditions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, New Orleans, LO, USA, 21–24 June 2022; pp. 4295–4302. [Google Scholar]
  35. Li, H.; Li, Y.; Zhang, G.; Liu, R.; Huang, H.; Zhu, Q.; Tao, C. Global and local contrastive self-supervised learning for semantic segmentation of HR remote sensing images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5618014. [Google Scholar] [CrossRef]
  36. Uijlings, J.R.; Van De Sande, K.E.; Gevers, T.; Smeulders, A.W. Selective search for object recognition. Int. J. Comput. Vis. 2013, 104, 154–171. [Google Scholar] [CrossRef] [Green Version]
  37. Zitnick, C.L.; Dollár, P. Edge boxes: Locating object proposals from edges. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 391–405. [Google Scholar]
  38. Arbeláez, P.; Pont-Tuset, J.; Barron, J.T.; Marques, F.; Malik, J. Multiscale combinatorial grouping. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 328–335. [Google Scholar]
  39. Oord, A.V.D.; Li, Y.; Vinyals, O. Representation learning with contrastive predictive coding. arXiv 2018, arXiv:1807.03748. [Google Scholar]
  40. Everingham, M.; Van Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The pascal visual object classes (voc) challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef] [Green Version]
  41. Everingham, M.; Eslami, S.M.; Van Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The pascal visual object classes challenge: A retrospective. Int. J. Comput. Vis. 2015, 111, 98–136. [Google Scholar] [CrossRef]
  42. Deselaers, T.; Alexe, B.; Ferrari, V. Weakly supervised localization and learning with generic knowledge. Int. J. Comput. Vis. 2012, 100, 275–293. [Google Scholar] [CrossRef]
  43. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Miami Beach, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  44. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 740–755. [Google Scholar]
Figure 1. Illustration of our motivation through data flow. (a) Most WSOD methods usually learn object instances from region proposals of current input image. (b) Our method establishes instance correlations with other images to guide instance representation learning for the current input image.
Figure 1. Illustration of our motivation through data flow. (a) Most WSOD methods usually learn object instances from region proposals of current input image. (b) Our method establishes instance correlations with other images to guide instance representation learning for the current input image.
Sensors 22 07525 g001
Figure 2. The pipeline of instance-level contrastive learning (ICL) framework. The upper part shows the overall network structure, where only the blue arrows backpropagate the gradients. There are one base head (Base-H) and three refined heads (R-H1, R-H2, R-H3). The base head is supervised by image labels, while each refined head is supervised by the previous parallel head. Dashed arrows indicate the supervision information. The detailed network structure of these heads can be found in the lower boxes, where each box corresponds to a submodule in the pipeline. Three refined heads have the same network structure but do not share parameters. The processes of memory-aware instance mining (MIM) algorithm, memory-aware proposal sampling (MPS) algorithm and contrastive learning (CL) are also shown in boxes.
Figure 2. The pipeline of instance-level contrastive learning (ICL) framework. The upper part shows the overall network structure, where only the blue arrows backpropagate the gradients. There are one base head (Base-H) and three refined heads (R-H1, R-H2, R-H3). The base head is supervised by image labels, while each refined head is supervised by the previous parallel head. Dashed arrows indicate the supervision information. The detailed network structure of these heads can be found in the lower boxes, where each box corresponds to a submodule in the pipeline. Three refined heads have the same network structure but do not share parameters. The processes of memory-aware instance mining (MIM) algorithm, memory-aware proposal sampling (MPS) algorithm and contrastive learning (CL) are also shown in boxes.
Sensors 22 07525 g002
Figure 3. The effect of number of vectors N.
Figure 3. The effect of number of vectors N.
Sensors 22 07525 g003
Figure 4. The effect of combination coefficient μ .
Figure 4. The effect of combination coefficient μ .
Sensors 22 07525 g004
Figure 5. Training loss.
Figure 5. Training loss.
Sensors 22 07525 g005
Figure 6. Model performance during the training process. Evaluations are performed every 10K without data augmentation.
Figure 6. Model performance during the training process. Evaluations are performed every 10K without data augmentation.
Sensors 22 07525 g006
Figure 7. Qualitative results on PASCAL VOC2007 dataset. Yellow boxes mean ground truths. Green and red boxes represent correct and failed cases, respectively.
Figure 7. Qualitative results on PASCAL VOC2007 dataset. Yellow boxes mean ground truths. Green and red boxes represent correct and failed cases, respectively.
Sensors 22 07525 g007
Table 1. Comparison with other methods on Pascal VOC2007 test set. means our baseline.
Table 1. Comparison with other methods on Pascal VOC2007 test set. means our baseline.
MethodaerbikbirboabotbuscarcatchacowtabdoghormotperplashesoftratvmAP
WSDDN [20]39.450.131.516.312.664.542.842.610.135.724.938.234.455.69.414.730.240.754.746.934.8
Kantorov et al. [14]57.152.031.57.611.555.053.134.11.733.149.242.047.356.615.312.824.848.944.447.836.3
OICR [15] 58.062.431.119.413.065.162.228.424.844.730.625.337.865.515.724.141.746.964.362.641.2
PCL [26]54.469.039.319.215.762.964.430.025.152.544.419.639.367.717.822.946.657.558.663.043.5
C-MIL [21]62.558.449.532.119.870.566.163.420.060.552.953.557.468.98.424.651.858.766.763.550.5
Lin et al. [22]55.667.045.827.921.169.068.370.521.360.240.354.556.570.112.525.052.955.265.063.750.1
TS 2 C [18]59.357.543.727.313.563.961.759.924.146.936.745.639.962.610.323.641.752.458.756.644.3
WS-JDS [16]52.064.545.526.727.960.547.859.713.050.446.456.349.660.725.428.250.051.466.529.745.6
SDCN [19]59.867.132.034.722.867.163.867.922.548.947.860.551.765.211.820.642.154.760.864.348.3
Yang et al. [17]--------------------50.6
SLV [27]65.671.449.037.124.669.670.370.630.863.136.061.465.368.412.429.952.460.067.664.553.5
D-MIL [25]60.471.351.125.423.870.470.371.925.263.442.667.157.770.115.526.658.763.366.967.653.5
ICL61.973.044.033.332.975.374.773.82.670.662.060.872.271.326.025.457.357.772.760.955.4
Table 2. Comparison with other methods on Pascal VOC2007 trainval set. means our baseline.
Table 2. Comparison with other methods on Pascal VOC2007 trainval set. means our baseline.
Methodaerbikbirboabotbuscarcatchacowtabdoghormotperplashesoftratvmean
WSDDN [20]65.158.858.533.139.868.360.259.634.864.530.543.056.882.425.541.661.555.965.963.753.5
Kantorov et al. [14]83.368.654.723.418.373.674.154.18.665.147.159.567.083.535.339.967.049.763.565.255.1
OICR [15] 81.780.448.749.532.881.785.440.140.679.535.733.760.588.821.857.976.359.975.381.460.6
PCL [26]79.685.562.247.937.083.883.443.038.380.150.630.957.890.827.058.275.368.575.778.962.7
C-MIL [21]--------------------65.0
Lin et al. [22]--------------------67.2
TS 2 C [18]84.274.161.352.132.176.782.966.642.370.639.557.061.288.49.354.672.260.065.070.361.0
WS-JDS [16]82.974.073.447.160.980.477.578.818.670.056.767.064.584.047.050.171.957.683.343.564.5
SDCN [19]85.883.156.258.544.780.285.077.929.678.853.674.273.188.418.257.574.260.876.179.266.8
SLV [27]84.684.373.358.549.280.287.079.446.883.641.879.388.890.419.559.779.467.782.983.271.0
D-MIL [25]81.382.072.748.942.080.286.178.543.980.242.276.568.791.232.756.081.469.678.779.968.7
ICL85.388.965.557.557.486.090.785.815.188.778.074.489.293.539.257.688.571.686.280.174.0
Table 3. Comparison with other methods on Pascal VOC 2012 dataset. means our baseline.
Table 3. Comparison with other methods on Pascal VOC 2012 dataset. means our baseline.
MethodmAPCorLoc
Kantorov et al. [14]35.354.8
OICR [15] 37.962.1
PCL [26]40.663.2
C-MIL [21]46.767.4
Lin et al. [22]45.367.1
TS 2 C [18]40.064.4
WS-JDS [16]39.163.5
SDCN [19]43.567.9
SLV [27]49.269.2
D-MIL [25]49.670.1
ICL50.170.4
Table 4. The contribution of each component of our method. Both PASCAL metrics and COCO metrics are applied.
Table 4. The contribution of each component of our method. Both PASCAL metrics and COCO metrics are applied.
MethodPASCAL MetricsCOCO Metrics
mAPCorLocAPAP 50 AP 75 AP S AP M AP L
Baseline41.260.614.937.010.71.88.619.4
ICL55.474.020.848.814.42.310.327.1
ICL w/o IMU53.672.919.846.814.22.810.125.9
ICL w/o MIM49.769.618.043.112.52.18.923.8
ICL w/o MPS52.971.419.546.713.53.310.525.0
Table 5. The effect of the removal coefficient 1 / λ .
Table 5. The effect of the removal coefficient 1 / λ .
1 / λ mAPCorLoc
1/1052.770.8
1/852.070.1
1/653.672.6
1/455.474.0
1/252.470.8
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, M.; Zeng, B. Instance-Level Contrastive Learning for Weakly Supervised Object Detection. Sensors 2022, 22, 7525. https://doi.org/10.3390/s22197525

AMA Style

Zhang M, Zeng B. Instance-Level Contrastive Learning for Weakly Supervised Object Detection. Sensors. 2022; 22(19):7525. https://doi.org/10.3390/s22197525

Chicago/Turabian Style

Zhang, Ming, and Bing Zeng. 2022. "Instance-Level Contrastive Learning for Weakly Supervised Object Detection" Sensors 22, no. 19: 7525. https://doi.org/10.3390/s22197525

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop