Next Article in Journal
π-FBG Fiber Optic Acoustic Emission Sensor for the Crack Detection of Wind Turbine Blades
Next Article in Special Issue
Thermal Image Super-Resolution Based on Lightweight Dynamic Attention Network for Infrared Sensors
Previous Article in Journal
Understanding the Role of Sensor Optimisation in Complex Systems
Previous Article in Special Issue
Deep Sensing for Compressive Video Acquisition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Scale-Hybrid Group Distillation with Knowledge Disentangling for Continual Semantic Segmentation

School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(18), 7820; https://doi.org/10.3390/s23187820
Submission received: 10 July 2023 / Revised: 6 September 2023 / Accepted: 8 September 2023 / Published: 12 September 2023

Abstract

:
Continual semantic segmentation (CSS) aims to learn new tasks sequentially and extract object(s) and stuff represented by pixel-level maps of new categories while preserving the original segmentation capabilities even when the old class data is absent. Current CSS methods typically preserve the capacities of segmenting old classes via knowledge distillation, which encounters the limitations of insufficient utilization of the semantic knowledge, i.e., only distilling the last layer of the feature encoder, and the semantic shift of background caused by directly distilling the entire feature map of the decoder. In this paper, we propose a novel CCS method based on scale-hybrid distillation and knowledge disentangling to address these limitations. Firstly, we propose a scale-hybrid group semantic distillation (SGD) method for encoding, which transfers the multi-scale knowledge from the old model’s feature encoder with group pooling refinement to improve the stability of new models. Then, the knowledge disentangling distillation (KDD) method for decoding is proposed to distillate feature maps with the guidance of the old class regions and reduce incorrect guides from old models towards better plasticity. Extensive experiments are conducted on the Pascal VOC and ADE20K datasets. Competitive performance compared with other state-of-the-art methods demonstrates the effectiveness of our proposed method.

1. Introduction

Semantic segmentation [1,2] is a fundamental task in the field of computer vision that aims to assign a category to each pixel in an image. With the help of Convolutional Neural Networks (CNNs), semantic segmentation methods have achieved significant progress under the condition of all of the fixed classes that have been given. However, the main challenge in the real-world lies in the constantly changing environment (i.e., the new classes are generated progressively), which means a semantic segmentation model needs to continually learn from newly emerged classes while preserving the learned knowledge from old classes without retraining from scratch. Such a learning process is called Continual Semantic Segmentation (CSS), which has gained widespread interest from researchers and plays an important role in wide potential applications in dynamically changing environments, such as automatic driving, medical imaging, robotics, augmented reality, and so on.
A straightforward solution is to fine tune [3] the trained old model on data from the new class to adjust the parameters to fit the distribution of the new class without using the old class of data. Nevertheless, the model suffers from a catastrophic forgetting problem [4,5] when it is updated incrementally by a gradient-based update method, which will directly lead to rapid degradation of the model performance, i.e., the model will forget how to solve the old classes after learning the new classes due to the interference caused by the parameter updation. To alleviate the problem, some researchers have attempted to adopt the knowledge distillation strategy [6,7,8,9,10,11], which transfers knowledge from the old class to the new class of models to preserve the model’s original capacity. Despite the success of the knowledge distillation-based method for CSS, two common limitations remain. The first limitation is inadequate distillation. Specifically, only the knowledge of the last layer of the encoder is transferred by simply pooling the feature map, which fails to leverage the richer knowledge of the old model such as multi-scale object information, spatial position, and channel semantics. The second limitation is the semantic shift [12]: the old CSS models consider all data that is not the current class as background, which may contain non-background information in the new task. However, existing CSS methods force distill the entire feature map from the old task decoder to the new task, which may disturb the learning of new classes since some new classes are mistakenly considered as background by the old model.
To address the aforementioned problems, we propose a novel scale-hybrid group distillation with the knowledge disentangling continual semantic segmentation method. In particular, a scale-hybrid group distillation for encoding is proposed for transferring richer semantic knowledge from the feature encoder of the old model to avoid inadequate distillation. On the one hand, we perform multi-scale distillation of the encoder to preserve comprehensive semantic information. On the other hand, we design a group pooling strategy to retain the spatial position and channel semantic knowledge. In addition, we propose the knowledge disentangling distillation method composed of old class distillation and new class learning for decoding to reduce the semantic shift of the background class. Unlike previous indiscriminate distillation, we utilize the old model to generate pseudo labels for the current task, and then only focus on the knowledge transfer from the old non-background classes to avoid the semantic shift where the class is considered as background in the old task and foreground in the new task. Additionally, the new class learning is supervised by the cross entropy loss function in order to enhance the plasticity of the model.
The main contributions can be summarized as follows:
  • We propose a scale-hybrid group distillation (SGD) for encoding to transfer richer semantic knowledge from the old model’s feature encoder in different scales in a novel group pooling manner to preserve comprehensive knowledge without the catastrophic forgetting problem.
  • We propose a knowledge disentangling distillation (KDD) for decoding to decompose the learning of old and new knowledge based on the corresponding model. This approach can reduce the interference of incorrect guides from old models for the new knowledge.
  • Extensive experiments on Pascal VOC and ADE20k datasets are conducted on the typical continual semantic segmentation settings, and the results demonstrate the effectiveness of our proposed method.

2. Related Work

In this section, we firstly make a overview of two important research fields: semantic segmentation and continual learning. Then, we further make an in-depth exploration of recent advancements in continual semantic segmentation field.

2.1. Semantic Segmentation

In recent years, significant progress has been made in the field of semantic segmentation benefit from the availability of large datasets [13,14,15,16] and advancements [1,2] in deep convolutional neural networks. Long et al. [1] proposed an end-to-end architecture for semantic segmentation called fully convolutional networks (FCN), which can output pixel-wise prediction of the object class. However, FCN suffers from the loss of spatial information and insufficient contextual information. Chen et al. [17,18,19] proposed the Deeplab series to capture more spatial information through atrous convolutions. Some works [2,20,21,22] have adopted an encoder-decoder structure for retaining spatial information. In order to capture contextual information in images, some works [23,24,25,26,27] have adopted an attention mechanism to build up connections between image contexts. Recently, semantic segmentation approaches based on the transformer architecture have had a great deal of success [28,29,30,31,32], mainly in their ability to capture long-range dependencies in images. Despite the remarkable achievement of the semantic segmentation field, these methods are not capable of dealing with the new emerging classes of continual learning, which are more applicable to real-world scenarios.

2.2. Continual Learning

Continual learning aims to adapt and learn from new data while retaining knowledge acquired from previous data. However, catastrophic forgetting of old knowledge is a major challenge when learning new knowledge. To address this issue, some approaches have been proposed for continual learning, including replay-based methods [33,34,35,36,37,38], regularization-based methods, and architecture-based methods. Replay-based methods solve the catastrophic forgetting problems by retaining the old knowledge in the learning of new data. The stored old knowledge can be divided into many types, including partial raw data [33,36], synthetic data [34,35], and prototype information [37,38]. Without the need of storing old data, regularization-based methods adopt regularization techniques to encourage the model to retain previous knowledge, including the knowledge distillation [39,40,41,42,43], adversarial learning [44,45], and vanilla regularization methods [46,47,48]. Besides, the architecture-based approaches [49,50,51] dynamically adjust the network architecture to preserve the learned knowledge from the old task while acquiring new information from the current task. Recently, continual learning techniques have been applied to a series of computer vision tasks, including object detection [52,53], semantic segmentation [6,7,12], and instance segmentation [54,55].

2.3. Continual Semantic Segmentation

Continual semantic segmentation (CSS) [6,7,11,12,56,57] is a very challenging task in computer vision, which aims to solve the catastrophic forgetting [46] in semantic segmentation. ILT [6] first proposed the continual semantic segmentation task and a Deeplab-based [18] CSS framework. Cermelli et al. [12] further built unbiased knowledge distillation to avoid catastrophic forgetting.
In recent years, more CSS methods [7,8,9,10,11,56,57,58] have been proposed. SDR [8] made full use of the prototype matching to reduce forgetting and improve the representation ability of new classes. PLOP [7] extracted multi-scale features in intermediate layers for knowledge distillation. Replay-based RECALL [9] hoped to add extra data of old classes and the background class from the online website or by using the GAN network. Taking into account the influence of class similarities, REMINDER [57] built different weights during the processing of distillation. Architectural-based RCN [10] creatively designed two parallel network branches for storing old knowledge and learning new categories to avoid forgetting problems. RBC [11] believed that the background of old classes in the new image is more similar to the old image, which can significantly exacerbate the old class forgetting and the new level of learning. So a biased-context-rectified CSS framework was proposed, which decoupled different classes by using context-rectified image-duplet learning.
However, the above existing methods directly utilize the knowledge of old models by merely distilling the last layer of the feature encoder and simply pooling the feature map for distillation without paying attention to the different impacts of backgrounds and old classes.

3. Method

The overall framework of our method is shown in Figure 1. Given an image, we first input the image into the old model in the last step to predict the segmentation masks of the old classes learned in the previous step. Then, the predicted masks are used as pseudo-labels of the old classes, and they are combined with the ground truth of the new classes to train the current model. Next, to prevent catastrophic forgetting, we propose scale-hybrid group distillation (SGD) for the encoder and knowledge disentangling distillation (KDD) for the decoder. In the following sections, we first introduce the basic settings and preliminaries of continual semantic segmentation in Section 3.1. The framework of our method is then illustrated in Section 3.2. Moreover, we present the proposed scale-hybrid group distillation in Section 3.3 and knowledge disentangling distillation in Section 3.4.

3.1. Preliminaries

Compared with traditional semantic segmentation with all classes of labels available, the continual semantic segmentation (CSS) task divides the dataset into multiple subsets based on classes, each of which only contains a part of the class labels. The model learns one subset at each learning step without labels from previous subsets, and there is no class intersection between different subsets.
Given a multi-step dataset D = { D 1 , D 2 , , D T } , and the corresponding classes C = { C 1 , C 2 , , C T } contained in each subset of D , where T is the total learning steps. The subset D t contains a series of sample pairs { I i t , G i t } , where I i t R 3 × H × W is the input images and G i t is the corresponding ground truth, H and W are the height and width of the input images, and the subscript i means the i- th sample pair. We denote | D t | as the number of subset D t , and | C t | as the number of classes contained in subset D t . Only the subset D t is available and the labels of class C t are contained in G t during the t- th learning step, while the previously learned classes C 1 : t 1 is regarded as the background class C 0 . Due to the disjointness of classes contained in distinct subsets, then C 1 : t 1 C t = . After the learning step t, the labels of C t are no longer available, and the classes C t in the t-th step are regarded as the background in the current subset D t + 1 during the learning step t + 1 . Therefore, learning new classes and maintaining the segmentation capacity of the old classes is a challenging research topic in recent years.
There are two main issues with the CSS task. On the one hand, if the old model is directly fine-tuned on the new data with the unavailable labels of the old class, the model will quickly forget the learned knowledge, which will lead to a sharp performance decline for the old tasks, i.e., catastrophic forgetting.
On the other hand, in the CSS task, the forthcoming classes are considered as the background in the old task, while when learning new classes, the old classes are considered as the background because the corresponding labels are not available. Therefore, the semantics of the background changes with the different learning steps—the so-called background semantic shift [12] problem. Once the model is supervised directly by the labels of data from the new step, the background semantic shift will constrain the model to reclassify the old classes as the background to aggravate catastrophic forgetting. The details of our method for alleviating the two problems are illustrated as follows:

3.2. Basic Framework

We first introduce the basic framework of continual semantic segmentation. We build our method on a fully convolutional neural network M consisting of a convolutional encoder E n c and a decoder, where the decoder consists of a decoding network D e c and an output layer O. In the learning step t, given an image I t , we first input it into the encoder E n c t to get the deep feature map of the image F t R C × H 16 × W 16 , then feed it into the decoding network D e c t to get the decoded feature map F o t . Finally, the feature map F o t R C o × H 16 × W 16 is fed into the output layer O t to get the final segmentation result S t R | C 0 : t | × H × W . Our framework is formulated as:
F t = E n c t ( I t ) ; F o t = D e c t ( F t ) S t = O t ( F o t )
In the new learning step t + 1 , we first initialize the current model M t + 1 using the old model M t obtained from the last training step t. To prevent the model from catastrophic forgetting during the learning process of new classes, we use the old model M t in the last step to perform knowledge distillation on the current model M t + 1 to maintain the knowledge of the old classes learned by the model and we perform novel knowledge distillation methods for the encoder and decoder of the model respectively to improve the stability of the model for old classes and the plasticity for new classes.
For the encoder, we focus on the comprehensive distillation of semantic information from multiple feature layers and propose the scale-hybrid group encoder semantic distillation (SGD). For the decoder, we propose a knowledge disentangling distillation (KDD) method that disentangles the old and new classes in the image and distills the old class features. In the following subsections, we will introduce our proposed distillation method in detail.

3.3. Scale-Hybrid Group Distillation

To mitigate catastrophic forgetting in continual semantic segmentation, most studies adopt the technique of knowledge distillation [7,8,10], which minimizes dissimilarities between features of the teacher and features of the student [59]. We denote the intermediate features of l- th layer in the encoder as F l , the distillation loss between the old model and the current model defined by L 2 norm is as follows:
L l d i s = 1 H W i = 1 H W f l , i t f l , i t 1 2
where f l , i t 1 are the features of F l at position i in the old model and f l , i t are the features of the current model. However, simply applying vanilla L 2 norm to measure dissimilarities between features commonly can not achieve ideal performance, due to the lack of utilization of the richer semantic knowledge of the old model, including multi-scale object information, spatial position, and channel semantics. If we regularize L l d i s too much, it will constrain the features of the current model close to the old features, and damage the process of learning the new tasks. If we loosen L l d i s too much, it will make the model ignore the semantic knowledge of old tasks and lead to catastrophic forgetting. Therefore, adaptively distilling the important semantic features of the previous tasks and leaving more plasticity for the new tasks is vital for improving the performance.
POD [43] distills pooled features at different layers between the old model and the new model to alleviate catastrophic forgetting. However, semantic segmentation requires more fine-grained information than the task of classification. We devise a novel Scale-hybrid Group Distillation method to distill features within each small cuboid of features, as shown in Figure 2. Specifically, we first divide features F R C × H × W into G 2 K groups. For each group, the features can be represented as:
F ˜ i , j , k = F [ k C K : ( k + 1 ) C K , i H G : ( i + 1 ) H G j W G : ( j + 1 ) W G ]
where i , j = 1 , , G and k = 1 , , K . Then, we apply the average pooling on each f i , j , k :
f ¯ i , j , k = A v g P o o l ( F ˜ i , j , k )
We concatenate the pooled results of these cubes and obtain the final features for distillation:
f ¯ = [ f ¯ 1 , 1 , 1 , , f ¯ G , G , K ]
The process of the proposed group-wise distillation measuring the dissimilarities by the pooled features is formulated as follows:
L l d i s = f ¯ l t f ¯ l t 1 2
To further keep the consistency of different scales between the old and current models, an average loss over multi-scale features of different layers is constructed:
L d i s = 1 L l = 1 L L l d i s
where L is the total amount of layers we distill. Through this scale-hybrid group distillation, we can effectively and comprehensively reserve the semantic knowledge of previous tasks and continually learn new tasks.

3.4. Knowledge Disentangling Distillation

In the process of distilling the features generated by the encoder, the feature map is considered as a whole, without distinguishing between different classes. However, an image often contains both new and old classes simultaneously and the background semantic shift problem makes the definition of background dynamically change over time, i.e., the foreground classes in the current task are regarded as the background in the old task. If the entire feature map is indiscriminately distilled using the old model, the semantic shift of the old model to the background will mislead the current model and limit the plasticity of the model. At the same time, the old model has a richer knowledge of identifying old classes and a stronger ability to represent the features of the old classes, which is also the most desired knowledge to be preserved during the distillation process. Therefore, in order to further explore the distillation of old classes by old models, we propose a knowledge disentangling distillation method to disentangle old and new classes and use the old model to distill the old class features more intently while reducing the interference in learning the new classes.
Specifically, our method disentangles the old and new classes and consists of three parts. First, we use the segmentation masks predicted by the old model in the last step as the pseudo-labels of the old classes. These pseudo-labels are used for capturing regions of the old classes and supervising the current segmentation results. Second, based on the spatial regions corresponding to the old classes in the pseudo-labels, the class-specific features are captured by the current model, and they are distilled by the old model. Finally, the ground truths of the current new classes are combined with the pseudo-labels of the old classes for the supervision of the output results of the current model.

3.4.1. Pseudo-Label Generation

During different learning steps, old classes are annotated as background since their labels are not available. Therefore, the model cannot perceive the region corresponding to the old classes, and it is difficult to use the old model to distill the features of the old class regions generated by the current model. Moreover, the background semantic shift caused by the background semantics is constantly changing in different learning steps, which aggravates the catastrophic forgetting of the model.
To address this problem, since the old model has a stronger segmentation ability for the learned old classes, following Ref. [7], we leverage the old model to generate pseudo-labels for the old classes. In learning step t, given an image I t in the current dataset D t , we first feed I t into the old model M t 1 in the last step to generate segmentation predictions S t 1 for the old classes. However, the segmentation mask S t 1 of the old classes generated by the old model is not completely accurate. If the regions of old classes are captured completely according to S t 1 , many features of the non-old classes will also be included. To reduce these noises and interference, we use the median of the entropy of the old model’s prediction probabilities for each class in the current dataset as the threshold for determining the confidence of the model’s prediction. For position i, when the entropy of the probability of class c predicted by the old model is less than the threshold τ c , it means that the prediction confidence of position i is higher, and it is set as the foreground, otherwise the position is ignored. The updated pseudo-label S ˜ t 1 is formulated as:
S ˜ i t 1 = S i t 1 , i f e n t r o p y ( S i t 1 ) < τ c μ , o t h e r w i s e
where μ denotes the position that is ignored and not calculated. Based on the obtained pseudo-labels S ˜ t 1 , we further perform the distillation of old class features and supervision of the segmentation results.

3.4.2. Distillation of Old Class

Since the features of different classes are not distinguished, the incorrect recognition of the new classes by the old model will mislead the learning processing, and it also causes difficulties in distilling the knowledge of the old classes. To enhance the old classes distillation, after obtaining accurate old class pseudo-label S ˜ t 1 , we generate class region masks M corresponding to different old classes according to S ˜ t 1 , and distill the old class-specific features. In learning step t, given an image I t , we first generate an output feature map F o t 1 and F o t through the old and new models. For the old class c, we first obtain the area corresponding to c in the pseudo-label S ˜ t 1 , and get the area mask M c of class c.
Then we conduct element-wise multiplication of M c with F o t 1 and F o t and use average pooling to fuse the features in the c- th class region to obtain the embedding features of class c generated by the old and new models. The equations are formulated as:
f c t 1 = A v g P o o l ( M c · F o t 1 ) f c t = A v g P o o l ( M c · F o t )
Following the above approach, we disentangle the old and new classes and obtain the embedding features of the old classes generated by the old and the current model. Next, we perform distillation exclusively on old class features. Specifically, we constrain the embedding features of each old class generated by the current model to be similar to those generated by the old model, to more intently distill the old knowledge learned and reduce the interference of the current model to learn new classes. The equation for distillation loss is formulated as:
L d i s t _ o l d = 1 | C 1 : t 1 | c = 1 | C 1 : t 1 | | | f c t f c t 1 | | 2

3.4.3. Learning of New Class

While maintaining the knowledge of old classes, the current models need to further learn new classes. To prevent catastrophic forgetting due to background semantic shift and supervise the predictions of new classes, we combine the pseudo-label of the old classes S ˜ t 1 with the ground truth G t of the current new classes as the supervision of the current model. The combined label G ^ t is formulated as:
G ^ i t = G i t , G i t C t S ˜ i t 1 , S ˜ i t 1 C 1 : t 1 a n d G i t = 0 0 , o t h e r w i s e
where we replace the background of G t with the pseudo-labels of the old classes of S ˜ t 1 .
We utilize the obtained combined labels for training to get the segmentation predictions generated by the current model. The segmentation loss is:
L s e g = 1 H W i = 1 H W c C 0 : t G ^ c , i t l o g S c , i t

4. Experiments

In this section, we evaluate the performance of our method on the Pascal VOC dataset [13] and ADE20K dataset [15].

4.1. Datasets, Protocols, and Metrics

4.1.1. Datasets

The Pascal VOC dataset [13,60] is a mainstream benchmark dataset for object detection and semantic segmentation tasks in computer vision, which contains 20 different semantic classes with 10,582 images for training and 1449 images for testing.
The ADE20K dataset [15] is a large-scale scene understanding dataset, which contains 150 semantic classes, with 20,210 fully annotated images for the training process and 2000 fully annotated images for validation.

4.1.2. Protocols

There are two different settings for continual learning for semantic segmentation: Overlapped and Disjoint. For overlapped setting, the data of each step further contains future classes, which is more consistent with realistic scenes. For disjoint setting, the data in each step only contains old classes C 0 : t 1 learned in the previous steps and the current classes C t , without the future classes, while old classes are labeled as the background.
Following existing methods [7,11], for Pascal VOC dataset [13], we evaluate our method on traditional continual semantic segmentation protocols, including VOC-19-1 (2 steps, first training on 19 classes, and then on 1 new class), VOC-15-5 (2 steps, first training on 15 classes, and then on 5 new classes), and VOC-15-1 (6 steps, first training on 15 classes, and then on 1 new class in each of the next 5 steps).
For ADE20K dataset [15], we evaluate on our method with similarly continual semantic segmentation protocols, including 100-50 (2 steps, first training on 100 classes, and then on 50 new classes), 50-50 (3 steps, first training on 50 classes, and then on 50 new classes in each of the next 2 steps) and 100-10 (6 steps, first training on 100 classes, and then on 10 new classes in each of the next 5 steps).

4.1.3. Metrics

We use mean Intersection over Union (mIoU) as metrics to evaluate the performance of our method. This is a widely used evaluation metric in the field of semantic segmentation [1], which measures the accuracy of a segmentation algorithm by comparing the overlap between predicted and ground truth regions in an image. Specifically, we compute mIoU after the last step T for the initial classes C 1 , for the incremented classes C 2 , and for all classes C T . These metrics represent the stability (the robustness to catastrophic forgetting), the plasticity (capacity to learn new classes), and the overall performance of the proposed model, respectively.

4.1.4. Implementation Details

We utilized the Deeplab-V3 [18] architecture with ResNet-101 [61] as the backbone following current popular continual semantic segmentation methods [7,8,12]. The output stride of Deeplab-V3 is set to 16, and the in-place activated batch normalization [62] is applied in the backbone, which is pre-trained on the ImageNet [63].
We train the model on the Pascal VOC dataset for 30 epochs for every continual semantic segmentation step and the ADE20K dataset for 60 epochs for every continual semantic segmentation step, where the initial learning rate is 2 × 10 2 for the first step and 10 3 for all of the remaining steps and the batch size is set to 24 for both datasets. The learning rate is reduced exponentially with a decay rate of 0.9. We use the SGD optimizer with a momentum of 0.9 and weight decay rate of 10 4 during training. The feature maps used in our method are applied before ReLU with squared pixel values. During the process of continuous learning, we use the loss function proposed by MiB [12] following RCN [10]. Adhering to the definition of incremental learning, the task ID during inference is not accessible by the model, which needs to predict the target class in the set of all the seen classes.

4.2. Main Results

4.2.1. Pascal VOC

As we can see in Table 1, we first compare our method with the current state-of-the-art continual semantic segmentation methods on the Pascal VOC dataset [13] under the overlapped settings. Our method is evaluated on different continual learning tasks, namely 19-1, 15-5, and 15-1, and shows consistent improvements over current popular methods. Compared with the most recent method GSC [64], our method achieves competitive performance under setting of 19-1, and outperforms it by 0.83% and 2.28% under settings of 15-5 and 15-1 for all classes respectively.
Under the disjoint settings, as shown in Table 2, our method still achieves significant performance improvements. Especially under the setting of long learning steps 15–1, our method outperforms GSC [64] by 1.58%, which has the best performance previously except for introducing additional data of RECALL [9]. Our method reduces the forgetting of old classes while improving the learning ability of new classes, and the model’s stability and plasticity are enhanced.

4.2.2. ADE20K

In this part, we conduct experiments on the most challenging ADE20K dataset [15] for semantic segmentation to verify the effectiveness of the proposed method. The quantitative results are shown in Table 3. Our method is evaluated on multiple continual learning tasks, i.e., 100-50, 100-10, and 50-50, and it outperforms the current popular continual semantic segmentation methods.
ILT [6] builds a bridge between continual learning and semantic segmentation with weak effectiveness. Our method outperforms it by a large margin of 18.72%, 23.52%, and 34.05% under the 100-50, 100-10, and 50-50 settings, respectively. MiB [12] aims to address the background class shift problem, while our method proposes knowledge disentangling distillation to achieve higher performance, i.e., our method outperforms it by 2.93%, 3.91%, and 5.9% under settings of 100-50, 100-10, and 50-50, respectively. PLOP [7] extracts features in intermediate layers to make knowledge distillation in order to alleviate the issue of catastrophic forgetting, while our method proposes scale-hybrid group distillation to obtain greater improvement, e.g., our method outperforms it by 2.78%, 2.82%, and 3.55% under settings of 100-50, 100-10, and 50-50, respectively. Architectural-based RCN [10] designs two parallel network branches for storing previous information and learning new categories to avoid catastrophic forgetting, while our method obtains greater improvement, e.g., our method outperforms it by 1.22%, 0.72%, and 3.04% under settings of 100-50, 100-10, and 50-50, respectively. In addition, because the proposed method can retain most of the information from previous tasks, it has achieved greater improvement in more challenging experimental settings (6 steps) and obtains a 1.87% improvement compared with the second place overall.

4.3. Ablation Study

In this section, we conduct a set of ablation experiments to analyze the effectiveness of the different components of our proposed method.

4.3.1. Distillation Mechanism

We conduct ablation experiments under the overlapped setting of 15-1 on the Pascal VOC dataset. As shown in Table 4, the different components of our method achieve considerable improvements. PLOP [7] extracts features in intermediate layers, and RCN [10] designs two parallel network branches for storing previous information and learning new categories to alleviate the issue of catastrophic forgetting caused by knowledge distillation, respectively. The proposed scale-hybrid group distillation (SGD) brings performance improvement for both PLOP and RCN. At the same time, knowledge disentangling distillation (KDD) obtains consistent improvements both in combination with PLOP and RCN. In our method, both SGD and KDD have their own improvements. When combining SGD and KDD, it achieves the state-of-the-art performance of 63.08%, which demonstrates the effectiveness of the proposed method.

4.3.2. Different Group Pooling Kernel Sizes

In this part, we conduct ablation experiments under the overlapped setting of 15-1 on the Pascal VOC dataset. As shown in Table 5, the different group pooling kernel sizes play an important role in the proposed scale-hybrid group distillation (SGD) method. In practice, we set the group pooling kernel as a cube and divide the intermediate feature maps of ResNet [61] into three stages (C2C3, C4, C5). The performances of five different kernel sizes combinations of the three stages are compared. From the Table, the kernel size (8,8,8) achieves the worst performance, with one of the reasons being that the huge kernel size leads the network to ignore some detailed information. With the kernel size decreasing, e.g., (8,8,8) to (4,4,4), (4,4,4) to (2,2,2), the performance becomes better, where the kernel size (2,2,2) achieves a performance of 63.02% and outperforms the kernel size (8,8,8) by 13.12% mIoU. In addition, we conduct the experiments with mixed kernel size combinations, e.g., (2,4,8) and (8,4,2). Finally, the kernel size (8,4,2) achieves the best performance of 63.08%.

4.3.3. Different Pooling Methods

In this part, we conduct ablation experiments under the overlapped setting of 15-1 on the Pascal VOC dataset. As shown in Table 6, we compared five pooling methods of knowledge distillation. It can be observed that the proposed group pooling method preserves more semantic information compared with other pooling methods and achieves the best performance.

4.4. Visualization

In this section, we provide several visualization examples of the typical PLOP [7] and our method for qualitative comparison, as shown in Figure 3 and Figure 4. The visualization results demonstrate that the proposed method is significantly improved compared with PLOP. Some old categories are forgotten and not segmented well by PLOP, but our method makes an accurate segmentation. Besides, regarding the results shown in Figure 5 and Figure 6, while mIoU for PLOP deteriorates after only a handful of steps, our method’s mIoU remains very high throughout, indicating improved resilience to catastrophic forgetting and background semantic shift. In addition, we present two typical failure examples in Figure 7. The first row represents the problem of category confusion caused by catastrophic forgetting of the old categories, while the second row represents the challenge of segmentation for hard samples of new categories due to background semantic shift. In future work, we will employ more powerful feature extraction networks such as a transformer to further address these challenges.

5. Conclusions

In this paper, we propose a novel continual semantic segmentation method that utilizes two key components: the scale-hybrid group distillation for the encoder and the knowledge disentangling distillation for the decoder. Our scale-hybrid group distillation facilitates knowledge transfer from the feature encoder of the old to the new model at different scales through group pooling refinement so that the new model can preserve the old model’s abundant semantic information. Moreover, the knowledge disentangling distillation prevents the semantic shift of the background by focusing on the regions of old classes and reducing incorrect guidance from the old model. We evaluated our method on two challenging continual semantic segmentation datasets: the Pascal VOC and ADE20K datasets. The results demonstrate the effectiveness of our method and its superiority over other state-of-the-art methods.

Author Contributions

Conceptualization, Z.S. (Zichen Song); data curation, Z.S. (Zichen Song); formal analysis, Z.S. (Zichen Song); investigation, Z.S. (Zichen Song); methodology, Z.S. (Zichen Song); project administration, Z.S. (Zhaofeng Shi); resources, Z.S. (Zichen Song); software, Z.S. (Zichen Song); supervision, X.Z.; validation, X.Z. and Z.S. (Zhaofeng Shi); visualization, Z.S. (Zichen Song) and X.Z.; writing—original draft, Z.S. (Zichen Song) and X.Z.; writing—review and editing, X.Z., and Z.S. (Zhaofeng Shi); funding acquisition, Z.S. (Zichen Song) and Z.S. (Zhaofeng Shi). All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under Grant 62071086 and the Fundamental Research Funds for the Central Universities under Grant ZYGX2021YGLH210.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. The Pascal VOC dataset can be found here (accessed on 1 October 2022): http://host.robots.ox.ac.uk/pascal/VOC/. The ADE20K dataset can be found here (accessed on 1 October 2022): https://groups.csail.mit.edu/vision/datasets/ADE20K/.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  2. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  3. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  4. McCloskey, M.; Cohen, N.J. Catastrophic interference in connectionist networks: The sequential learning problem. In Psychology of Learning and Motivation; Academic Press: New York, NY, USA, 1989; Volume 24, pp. 109–165. [Google Scholar]
  5. Goodfellow, I.J.; Mirza, M.; Xiao, D.; Courville, A.; Bengio, Y. An empirical investigation of catastrophic forgetting in gradient-based neural networks. arXiv 2013, arXiv:1312.6211. [Google Scholar]
  6. Michieli, U.; Zanuttigh, P. Incremental Learning Techniques for Semantic Segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Long Beach, CA, USA, 15–20 June 2019; pp. 3205–3212. [Google Scholar]
  7. Douillard, A.; Chen, Y.; Dapogny, A.; Cord, M. PLOP: Learning Without Forgetting for Continual Semantic Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 4040–4050. [Google Scholar]
  8. Michieli, U.; Zanuttigh, P. Continual Semantic Segmentation via Repulsion-Attraction of Sparse and Disentangled Latent Representations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 1114–1124. [Google Scholar]
  9. Maracani, A.; Michieli, U.; Toldo, M.; Zanuttigh, P. RECALL: Replay-Based Continual Learning in Semantic Segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 7026–7035. [Google Scholar]
  10. Zhang, C.B.; Xiao, J.W.; Liu, X.; Chen, Y.C.; Cheng, M.M. Representation Compensation Networks for Continual Semantic Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 7053–7064. [Google Scholar]
  11. Zhao, H.; Yang, F.; Fu, X.; Li, X. RBC: Rectifying the Biased Context in Continual Semantic Segmentation. In Proceedings of the European conference on computer vision (ECCV), Tel Aviv, Israel, 23–27 October 2022; pp. 55–72. [Google Scholar]
  12. Cermelli, F.; Mancini, M.; Bulo, S.R.; Ricci, E.; Caputo, B. Modeling the Background for Incremental Learning in Semantic Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 9230–9239. [Google Scholar]
  13. Everingham, M.; Van Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The Pascal Visual Object Classes (VOC) Challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef]
  14. Cordts, M.; Omran, M.; Ramos, S.; Rehfeld, T.; Enzweiler, M.; Benenson, R.; Franke, U.; Roth, S.; Schiele, B. The cityscapes dataset for semantic urban scene understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 3213–3223. [Google Scholar]
  15. Zhou, B.; Zhao, H.; Puig, X.; Fidler, S.; Barriuso, A.; Torralba, A. Scene Parsing Through ADE20K Dataset. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 June 2017. [Google Scholar]
  16. Caesar, H.; Uijlings, J.; Ferrari, V. Coco-stuff: Thing and stuff classes in context. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 1209–1218. [Google Scholar]
  17. Liang-Chieh, C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Semantic Image Segmentation with Deep Convolutional Nets and Fully Connected CRFs. In Proceedings of the International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  18. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef] [PubMed]
  19. Chen, L.C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
  20. Noh, H.; Hong, S.; Han, B. Learning deconvolution network for semantic segmentation. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1520–1528. [Google Scholar]
  21. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef] [PubMed]
  22. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  23. Chen, L.C.; Yang, Y.; Wang, J.; Xu, W.; Yuille, A.L. Attention to scale: Scale-aware semantic image segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3640–3649. [Google Scholar]
  24. Ding, H.; Jiang, X.; Shuai, B.; Liu, A.Q.; Wang, G. Context contrasted feature and gated multi-scale aggregation for scene segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2393–2402. [Google Scholar]
  25. Li, X.; Zhong, Z.; Wu, J.; Yang, Y.; Lin, Z.; Liu, H. Expectation-maximization attention networks for semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9167–9176. [Google Scholar]
  26. Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual attention network for scene segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3146–3154. [Google Scholar]
  27. Hou, Q.; Zhang, L.; Cheng, M.M.; Feng, J. Strip pooling: Rethinking spatial pooling for scene parsing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 4003–4012. [Google Scholar]
  28. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. In Proceedings of the International Conference on Learning Representations, Vienna, Austria, 3–7 May 2021. [Google Scholar]
  29. Zheng, S.; Lu, J.; Zhao, H.; Zhu, X.; Luo, Z.; Wang, Y.; Fu, Y.; Feng, J.; Xiang, T.; Torr, P.H.; et al. Rethinking Semantic Segmentation From a Sequence-to-Sequence Perspective With Transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 6881–6890. [Google Scholar]
  30. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 10012–10022. [Google Scholar]
  31. Strudel, R.; Garcia, R.; Laptev, I.; Schmid, C. Segmenter: Transformer for Semantic Segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 7262–7272. [Google Scholar]
  32. Cheng, B.; Misra, I.; Schwing, A.G.; Kirillov, A.; Girdhar, R. Masked-Attention Mask Transformer for Universal Image Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 1290–1299. [Google Scholar]
  33. Rebuffi, S.A.; Kolesnikov, A.; Sperl, G.; Lampert, C.H. icarl: Incremental classifier and representation learning. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 June 2017; pp. 2001–2010. [Google Scholar]
  34. Shin, H.; Lee, J.K.; Kim, J.; Kim, J. Continual learning with deep generative replay. In Proceedings of the 30th Conference on Neural Information Processing Systems (NeurIPS), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  35. Kemker, R.; Kanan, C. FearNet: Brain-Inspired Model for Incremental Learning. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  36. Castro, F.M.; Marín-Jiménez, M.J.; Guil, N.; Schmid, C.; Alahari, K. End-to-end incremental learning. In Proceedings of the European conference on computer vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 233–248. [Google Scholar]
  37. Hayes, T.L.; Kafle, K.; Shrestha, R.; Acharya, M.; Kanan, C. Remind your neural network to prevent catastrophic forgetting. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Springer: Cham, Switzerland, 2020; pp. 466–483. [Google Scholar]
  38. Zhu, F.; Zhang, X.Y.; Wang, C.; Yin, F.; Liu, C.L. Prototype augmentation and self-supervision for incremental learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 5871–5880. [Google Scholar]
  39. Zenke, F.; Poole, B.; Ganguli, S. Continual learning through synaptic intelligence. In Proceedings of the International Conference on Machine Learning (ICML), Sydney, Australia, 6–11 August 2017; pp. 3987–3995. [Google Scholar]
  40. Aljundi, R.; Babiloni, F.; Elhoseiny, M.; Rohrbach, M.; Tuytelaars, T. Memory aware synapses: Learning what (not) to forget. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 139–154. [Google Scholar]
  41. Hou, S.; Pan, X.; Loy, C.C.; Wang, Z.; Lin, D. Learning a unified classifier incrementally via rebalancing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 831–839. [Google Scholar]
  42. Dhar, P.; Singh, R.V.; Peng, K.C.; Wu, Z.; Chellappa, R. Learning without memorizing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5138–5146. [Google Scholar]
  43. Douillard, A.; Cord, M.; Ollion, C.; Robert, T.; Valle, E. Podnet: Pooled outputs distillation for small-tasks incremental learning. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Springer: Cham, Switzerland, 2020; pp. 86–102. [Google Scholar]
  44. Xiang, Y.; Fu, Y.; Ji, P.; Huang, H. Incremental learning using conditional adversarial networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6619–6628. [Google Scholar]
  45. Ebrahimi, S.; Meier, F.; Calandra, R.; Darrell, T.; Rohrbach, M. Adversarial continual learning. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Springer: Cham, Switzerland, 2020; pp. 386–402. [Google Scholar]
  46. Kirkpatrick, J.; Pascanu, R.; Rabinowitz, N.; Veness, J.; Desjardins, G.; Rusu, A.A.; Milan, K.; Quan, J.; Ramalho, T.; Grabska-Barwinska, A.; et al. Overcoming catastrophic forgetting in neural networks. Proc. Natl. Acad. Sci. USA 2017, 114, 3521–3526. [Google Scholar] [CrossRef] [PubMed]
  47. Pan, P.; Swaroop, S.; Immer, A.; Eschenhagen, R.; Turner, R.; Khan, M.E.E. Continual deep learning by functional regularisation of memorable past. Adv. Neural Inf. Process. Syst. 2020, 33, 4453–4464. [Google Scholar]
  48. Zhao, B.; Xiao, X.; Gan, G.; Zhang, B.; Xia, S.T. Maintaining discrimination and fairness in class incremental learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 13208–13217. [Google Scholar]
  49. Frankle, J.; Carbin, M. The lottery ticket hypothesis: Finding sparse, trainable neural networks. arXiv 2018, arXiv:1803.03635. [Google Scholar]
  50. Golkar, S.; Kagan, M.; Cho, K. Continual learning via neural pruning. arXiv 2019, arXiv:1903.04476. [Google Scholar]
  51. Hung, C.Y.; Tu, C.H.; Wu, C.E.; Chen, C.H.; Chan, Y.M.; Chen, C.S. Compacting, picking and growing for unforgetting continual learning. In Proceedings of the 32nd Conference on Neural Information Processing Systems (NeurIPS), Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
  52. Feng, T.; Wang, M.; Yuan, H. Overcoming catastrophic forgetting in incremental object detection via elastic response distillation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 9427–9436. [Google Scholar]
  53. Yang, B.; Deng, X.; Shi, H.; Li, C.; Zhang, G.; Xu, H.; Zhao, S.; Lin, L.; Liang, X. Continual object detection via prototypical task correlation guided gating mechanism. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 9255–9264. [Google Scholar]
  54. Ganea, D.A.; Boom, B.; Poppe, R. Incremental few-shot instance segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 1185–1194. [Google Scholar]
  55. Nguyen, K.; Todorovic, S. ifs-rcnn: An incremental few-shot instance segmenter. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 7010–7019. [Google Scholar]
  56. Yan, S.; Zhou, J.; Xie, J.; Zhang, S.; He, X. An em framework for online incremental learning of semantic segmentation. In Proceedings of the 29th ACM International Conference on Multimedia, Chengdu, China, 20–24 October 2021; pp. 3052–3060. [Google Scholar]
  57. Phan, M.H.; Phung, S.L.; Tran-Thanh, L.; Bouzerdoum, A. Class Similarity Weighted Knowledge Distillation for Continual Semantic Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 16866–16875. [Google Scholar]
  58. Stan, S.; Rostami, M. Unsupervised model adaptation for continual semantic segmentation. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 2–9 February 2021; Volume 35, pp. 2593–2601. [Google Scholar]
  59. Hinton, G.; Vinyals, O.; Dean, J. Distilling the knowledge in a neural network. arXiv 2015, arXiv:1503.02531. [Google Scholar]
  60. Hariharan, B.; Arbelaez, P.; Bourdev, L.; Maji, S.; Malik, J. Semantic Contours from Inverse Detectors. In Proceedings of the International Conference on Computer Vision (ICCV), Barcelona, Spain, 6–13 November 2011. [Google Scholar]
  61. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  62. Bulo, S.R.; Porzi, L.; Kontschieder, P. In-place activated batchnorm for memory-optimized training of dnns. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5639–5647. [Google Scholar]
  63. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  64. Cong, W.; Cong, Y.; Dong, J.; Sun, G.; Ding, H. Gradient-Semantic Compensation for Incremental Semantic Segmentation. arXiv 2023, arXiv:2307.10822. [Google Scholar]
  65. Yang, G.; Fini, E.; Xu, D.; Rota, P.; Ding, M.; Nabi, M.; Alameda-Pineda, X.; Ricci, E. Uncertainty-Aware Contrastive Distillation for Incremental Semantic Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 2567–2581. [Google Scholar] [CrossRef] [PubMed]
  66. Yang, G.; Fini, E.; Xu, D.; Rota, P.; Ding, M.; Hao, T.; Alameda-Pineda, X.; Ricci, E. Continual Attentive Fusion for Incremental Learning in Semantic Segmentation. IEEE Trans. Multimed. 2022, 25, 3841–3854. [Google Scholar] [CrossRef]
  67. Lin, Z.; Wang, Z.; Zhang, Y. Continual Semantic Segmentation via Structure Preserving and Projected Feature Alignment. In Proceedings of the European conference on computer vision (ECCV), Tel Aviv, Israel, 23–27 October 2022; pp. 345–361. [Google Scholar]
  68. Goswami, D.; Schuster, R.; van de Weijer, J.; Stricker, D. Attribution-Aware Weight Transfer: A Warm-Start Initialization for Class-Incremental Semantic Segmentation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–7 January 2023; pp. 3195–3204. [Google Scholar]
Figure 1. An overview of our method. Our method consists of the Scale-hybrid Group Distillation in the encoder and the Knowledge Disentangling Distillation in the decoder. For the encoder, we propose group distillation for feature layers with different scales to preserve the old knowledge more adequately. For the decoder, we distill the output feature map of Atrous Spatial Pyramid Pooling (ASPP) module. Specifically, we disentangle the old and new classes and only distill the old class features, so as to alleviate the catastrophic forgetting background caused by the semantic shift problem.
Figure 1. An overview of our method. Our method consists of the Scale-hybrid Group Distillation in the encoder and the Knowledge Disentangling Distillation in the decoder. For the encoder, we propose group distillation for feature layers with different scales to preserve the old knowledge more adequately. For the decoder, we distill the output feature map of Atrous Spatial Pyramid Pooling (ASPP) module. Specifically, we disentangle the old and new classes and only distill the old class features, so as to alleviate the catastrophic forgetting background caused by the semantic shift problem.
Sensors 23 07820 g001
Figure 2. Illustration of scale-hybrid group distillation.
Figure 2. Illustration of scale-hybrid group distillation.
Sensors 23 07820 g002
Figure 3. Visualization of PLOP [7] and our predictions on Pascal VOC dataset [13]. ‘Image’ denotes the original image, and ‘GT’ denotes the ground truth.
Figure 3. Visualization of PLOP [7] and our predictions on Pascal VOC dataset [13]. ‘Image’ denotes the original image, and ‘GT’ denotes the ground truth.
Sensors 23 07820 g003
Figure 4. Visualization of PLOP [7] and our predictions on the ADE20K dataset [15]. ‘Image’ denotes the original image, and ‘GT’ denotes the ground truth.
Figure 4. Visualization of PLOP [7] and our predictions on the ADE20K dataset [15]. ‘Image’ denotes the original image, and ‘GT’ denotes the ground truth.
Sensors 23 07820 g004
Figure 5. mIoU evaluation under the overlapped setting of 15-1 on the Pascal VOC dataset.
Figure 5. mIoU evaluation under the overlapped setting of 15-1 on the Pascal VOC dataset.
Sensors 23 07820 g005
Figure 6. mIoU evaluation under the overlapped setting of 100-10 on ADE20K dataset.
Figure 6. mIoU evaluation under the overlapped setting of 100-10 on ADE20K dataset.
Sensors 23 07820 g006
Figure 7. Visualization of two typical failure examples. ‘Image’ denotes the original image, and ‘GT’ denotes the ground truth.
Figure 7. Visualization of two typical failure examples. ‘Image’ denotes the original image, and ‘GT’ denotes the ground truth.
Sensors 23 07820 g007
Table 1. Continual semantic segmentation results under the Overlapped settings of VOC-19-1, VOC-15-5, and VOC-15-1 benchmarks. Best in bold.
Table 1. Continual semantic segmentation results under the Overlapped settings of VOC-19-1, VOC-15-5, and VOC-15-1 benchmarks. Best in bold.
Method19-1 (2 Steps)15-5 (2 Steps)15-1 (6 Steps)
0–1920All0–1516–20All0–1516–20All
ILT [6]67.7510.8865.0567.0839.2360.458.757.998.56
MiB [12]71.4323.5969.1576.3749.9770.0834.2213.5029.29
SDR [8]69.1032.6067.4075.4052.6069.9044.7021.8039.20
PLOP [7]75.3537.3573.5475.7351.7170.0965.1221.1154.64
RECALL [9]67.9053.5068.4066.6050.9064.0065.7047.8062.70
UCD [65]71.4047.3070.0077.5053.1071.3049.0019.5041.90
CAF [66]75.5034.8073.4077.2049.9070.4055.7014.1045.30
RCN [10]---78.8052.0072.4070.6023.7059.40
RBC [11]77.2655.6076.2376.5952.7870.9269.5438.4462.14
SPPFA [67]76.5036.2074.6078.1052.9072.1066.2023.3056.00
AWT [68]---77.3052.9071.5059.1017.2049.10
GSC [64]76.9042.7075.3078.3054.2072.6072.1024.4060.80
Ours77.0139.9775.2578.8256.1673.4373.9228.3763.08
Joint77.4078.0077.4079.1072.5677.3979.1072.5677.39
Table 2. Continual semantic segmentation results under the Disjoint settings of VOC-19-1, VOC-15-5, and VOC-15-1 benchmarks. Best in bold.
Table 2. Continual semantic segmentation results under the Disjoint settings of VOC-19-1, VOC-15-5, and VOC-15-1 benchmarks. Best in bold.
Method19-1 (2 Steps)15-5 (2 Steps)15-1 (6 Steps)
0–1920All0–1516–20All0–1516–20All
ILT [6]69.1016.4066.4063.2039.5057.303.705.704.20
MiB [12]69.6025.6067.4071.8043.3064.7046.2012.9037.90
SDR [8]69.9037.3068.4073.5047.3067.2059.2012.9048.10
PLOP [7]75.3738.8973.6471.0042.8264.2957.8613.6746.48
RECALL [9]65.2050.1065.8066.3049.8063.5066.0044.9062.10
UCD [65]73.4033.7071.5071.9049.5066.2053.1013.0042.90
CAF [66]75.5030.8073.3072.9042.1065.2057.2015.5046.70
RCN [10]---75.0042.8067.3066.1018.2054.70
RBC [11]76.4345.7975.0175.1249.7169.8961.6819.5251.60
SPPFA [67]75.5038.0073.7075.3048.7069.0059.6015.6049.10
GSC [64]75.9031.0074.0074.4045.8067.6067.2019.2055.80
Ours77.1039.9175.3375.4244.8468.1469.3819.0057.38
Joint77.4078.0077.4079.1072.5677.3979.1072.5677.39
Table 3. Continual semantic segmentation results under the Overlapped settings of ADE-100-50, ADE50-50, and ADE-100-10 benchmarks. Best in bold.
Table 3. Continual semantic segmentation results under the Overlapped settings of ADE-100-50, ADE50-50, and ADE-100-10 benchmarks. Best in bold.
Method100-50 (2 Steps)50-50 (3 Steps)100-10 (6 Steps)
0–100101–150All0–5051–150All0–100101–150All
ILT [6]18.2914.4017.003.5312.859.700.113.061.09
MiB [12]40.5217.1732.7945.5721.0129.3138.2111.1229.24
PLOP [7]41.8714.8932.9448.8320.9930.4040.4813.6131.59
UCD [65]42.1215.8433.3147.1224.1231.7940.8015.2332.29
RCN [10]42.3018.8034.5048.3025.0032.5039.3017.6032.10
RBC [11]42.9021.4935.8149.5926.3234.1839.0121.6733.27
SPPFA [67]42.9019.9035.2049.8023.9032.5041.0012.5031.50
AWT [68]40.9024.7035.6046.6026.8533.5039.1021.2833.20
GSC [64]42.4019.2034.8046.2026.2033.0040.8017.6032.60
Ours42.3222.3835.7248.7125.1833.2243.2320.8335.14
Joint43.9027.2038.3050.9032.1038.3043.9027.2038.30
Table 4. The final mIoU (%) of the ablation study of our method. SGD denotes Scale-hybrid Group Distillation and KDD denotes Knowledge Disentangling Distillation. All experiments are conducted under the overlapped setting of 15-1 on Pascal VOC dataset.
Table 4. The final mIoU (%) of the ablation study of our method. SGD denotes Scale-hybrid Group Distillation and KDD denotes Knowledge Disentangling Distillation. All experiments are conducted under the overlapped setting of 15-1 on Pascal VOC dataset.
PLOP [7]RCN [10]SGDKDD15-1 (6 Steps)
58.32
59.64
62.03
62.17
62.81
63.08
Table 5. The final mIoU (%) of the ablation study about different group pooling kernel sizes. The numbers in brackets denote the pooling kernel sizes of three stages in sequence. All experiments are conducted under the overlapped setting of 15-1 on Pascal VOC dataset.
Table 5. The final mIoU (%) of the ablation study about different group pooling kernel sizes. The numbers in brackets denote the pooling kernel sizes of three stages in sequence. All experiments are conducted under the overlapped setting of 15-1 on Pascal VOC dataset.
Kernel Sizes15-1 (6 Steps)
(8,8,8)49.90
(4,4,4)58.96
(2,2,2)63.02
(2,4,8)49.97
(8,4,2)63.08
Table 6. The final mIoU (%) of the ablation study about different pooling methods. All experiments are conducted under the overlapped setting of 15-1 on Pascal VOC dataset.
Table 6. The final mIoU (%) of the ablation study about different pooling methods. All experiments are conducted under the overlapped setting of 15-1 on Pascal VOC dataset.
Method15-1 (6 Steps)
Strip Pooling62.17
Spatial Pooling48.55
Channel Pooling44.12
Spatial and Channel Pooling62.81
Max Pooling62.42
Group Pooling63.08
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Song, Z.; Zhang, X.; Shi, Z. Scale-Hybrid Group Distillation with Knowledge Disentangling for Continual Semantic Segmentation. Sensors 2023, 23, 7820. https://doi.org/10.3390/s23187820

AMA Style

Song Z, Zhang X, Shi Z. Scale-Hybrid Group Distillation with Knowledge Disentangling for Continual Semantic Segmentation. Sensors. 2023; 23(18):7820. https://doi.org/10.3390/s23187820

Chicago/Turabian Style

Song, Zichen, Xiaoliang Zhang, and Zhaofeng Shi. 2023. "Scale-Hybrid Group Distillation with Knowledge Disentangling for Continual Semantic Segmentation" Sensors 23, no. 18: 7820. https://doi.org/10.3390/s23187820

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop