Next Article in Journal
Intelligent Soft Sensors
Next Article in Special Issue
EnViTSA: Ensemble of Vision Transformer with SpecAugment for Acoustic Event Classification
Previous Article in Journal
Continuous Glucose Monitoring: A Possible Aid for Detecting Hypoglycemic Events during Insulin Tolerance Tests
Previous Article in Special Issue
Affective Neural Responses Sonified through Labeled Correlation Alignment
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Online Continual Learning in Acoustic Scene Classification: An Empirical Study

1
Artificial Intelligence Research Laboratory, Electronics and Telecommunications Research Institute, 218 Gajeong-ro, Daejeon 34129, Republic of Korea
2
Artificial Intelligence, University of Science and Technology, 217 Gajeong-ro, Daejeon 34113, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(15), 6893; https://doi.org/10.3390/s23156893
Submission received: 30 June 2023 / Revised: 27 July 2023 / Accepted: 1 August 2023 / Published: 3 August 2023
(This article belongs to the Special Issue Advances in Acoustic Sensors and Deep Audio Pattern Recognition)

Abstract

:
Numerous deep learning methods for acoustic scene classification (ASC) have been proposed to improve the classification accuracy of sound events. However, only a few studies have focused on continual learning (CL) wherein a model continually learns to solve issues with task changes. Therefore, in this study, we systematically analyzed the performance of ten recent CL methods to provide guidelines regarding their performances. The CL methods included two regularization-based methods and eight replay-based methods. First, we defined realistic and difficult scenarios such as online class-incremental (OCI) and online domain-incremental (ODI) cases for three public sound datasets. Then, we systematically analyzed the performance of each CL method in terms of average accuracy, average forgetting, and training time. In OCI scenarios, iCaRL and SCR showed the best performance for small buffer sizes, and GDumb showed the best performance for large buffer sizes. In ODI scenarios, SCR adopting supervised contrastive learning consistently outperformed the other methods, regardless of the memory buffer size. Most replay-based methods have an almost constant training time, regardless of the memory buffer size, and their performance increases with an increase in the memory buffer size. Based on these results, we must first consider GDumb/SCR for the continual learning methods for ASC.

1. Introduction

Acoustic scene classification (ASC) involves classifying a sound event into a predefined acoustic scene class [1,2,3,4]. Because deep-learning-based methods significantly improve performance in various tasks, numerous deep learning methods have been applied to ASC tasks to improve the classification accuracy of sound events [5,6,7,8,9,10,11,12,13,14,15]. In the detection and classification of acoustic scenes and events (DCASE) challenge, which is a representative event in the field of ASC, deep-learning-based methods improved the performance of sound event classification [10,12]. However, if a trained model with the DCASE dataset is deployed directly in the real world, unexpected results with severe performance degradation are often obtained. This phenomenon occurs because of device heterogeneities, that is, differences between the device used to build the training dataset and the device using the trained model. In addition, depending on the user’s environment, predefined classes may not be able to identify the right class assignment when there is no exact class matching. Therefore, to deploy a network model in the real world, re-training or fine-tuning the network model with new data obtained by the user’s sensors is inevitable [7,16]. However, re-training or fine-tuning the neural network incurs considerable computational costs and a large amount of memory space for training data.
Continual learning (CL) methods attempt to resolve the problems of task changes, including those associated with device heterogeneity, and class-incremental and task-incremental scenarios by transferring the knowledge obtained from previous tasks to the new classification task. Several studies on CL have been conducted, and various mechanisms have been proposed [17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38]. Thus, CL methods can effectively reduce the training time of the network for new data adaptation, and they are used to reduce the cost of re-training or fine-tuning models in various fields in the real world. When a neural network is trained for the current task based on knowledge obtained from a previous task, it suffers from catastrophic forgetting and intransigence. Catastrophic forgetting implies the loss of important information from preceding tasks caused by incorporating new information relevant to the current task [17,22], and intransigence refers to the inhibition of learning information from the new task [18,22]. Thus, the goal of CL methods is to improve the performance of neural networks by alleviating catastrophic forgetting and avoiding intransigence.
Previous methods on overcoming catastrophic forgetting and intransigence can be classified into three categories: regularization-based methods, replay-based methods, and dynamic-architecture-based methods [17,18,19,20,22,39,40,41,42,43]. Regularization-based methods [23,24,25] alleviate catastrophic forgetting by preserving previous capabilities or imposing constraints when the weights of the neural network are updated. Replay-based methods [26,27,28,30,31,32,33,35,36,38,44] store examples from previous tasks in a memory buffer or reproduce old examples using a generative model, and, then, they train the network using both the old examples and the examples of the current task. Dynamic-architecture-based methods [34,45,46,47] improve performance by dynamically changing neural network resources by analyzing old examples and examples of the current task. Because the performance of CL methods changes according to the data of the current task and the data of the previous tasks, the scenario that defines the relationship between the current task and previous tasks is an important factor in CL.
Based on the assumption of the training data and given information, such as a task identifier (ID), various CL scenarios can be classified into three categories [17,18,20,22,41,42]: task incremental, class incremental, and domain incremental. The task-incremental scenario assumes that each task has a task ID, which indicates information on the class distribution and input distribution. The class-incremental scenario has the same setting as the task-incremental scenario; however, the task ID remains unknown. The domain-incremental scenario assumes that each task is set with different nonstationarity, different distributions of input data, and same class distribution. In general, class- and domain-incremental scenarios are more realistic and difficult because of the absence of a task ID.
From another viewpoint, CL scenarios are classified into offline and online approaches based on data accessibility of the current and previous tasks [22]. The offline approach of CL methods assumes that all data for the current task and previous tasks are accessible; however, the online approach of CL methods assumes that only one small batch is available, and that the previous batch data as well as the data of preceding tasks are not accessible. Considering that ASC tasks are suitable for mobile devices, which have limited computing power and memory resources, and because privacy of sound data is important for users, online approaches to CL methods are more practical.
Numerous studies have applied CL in various fields such as natural language processing (NLP) [19], human activity recognition (HAR) [39,40], and computer vision [22]; however, few studies have been conducted on ASC tasks. In addition, most CL studies on ASC tasks use task-incremental settings and offline approaches, which are relatively simple and unrealistic. In ASC tasks, class-incremental and domain-incremental scenarios are necessary to reflect real-world scenarios where sound classes are varied and device heterogeneity exists. In addition, for reasons discussed earlier, online approaches to CL methods are more suitable for ASC tasks. Therefore, in this study, we developed more realistic CL scenarios, such as online class-incremental (OCI) and online domain-incremental (ODI) scenarios, for various ASC datasets. Additionally, we systematically investigated the performance of recent CL methods using two proposed scenarios on three public datasets. Finally, we comprehensively analyzed the features of each CL method by comparing the performance of CL methods on various datasets.
This study makes the following contributions:
  • We proposed OCI and ODI scenarios on three public datasets for evaluating the performance of CL methods in ASC tasks. For ODI scenarios, we used noise addition and SpecAugment to mimic the domain change. The public dataset contains the DCASE-2021 dataset [8], UrbanSound8K dataset [48], and ESC-50 dataset [49].
  • Using the two scenarios on three public datasets, we systematically analyzed the performance of the recent CL methods. In the OCI scenario, we experimented with two regularization-based methods and eight replay-based methods. For the ODI scenario, we analyzed two regularization-based methods and seven replay-based methods, excluding iCaRL.
  • We analyzed the experimental results using performance metrics such as accuracy and forgetting for both scenarios. Moreover, we evaluated the training time of the CL methods in accordance with the changes in buffer size for model training to validate the effectiveness of the CL methods.
The remainder of this paper is organized as follows. Section 2 reviews related works, and Section 3 describes the problem definition for online CL in ASC tasks. The CL methods are compared in Section 4. The experimental setup is presented in Section 5, and the experimental results are described in Section 6. Finally, conclusions and discussion are presented in Section 7.

2. Related Works

2.1. Acoustic Scene Classification

ASC involves the task of classifying a sound event into predefined classes such as acoustic scenes or acoustic environments [1,2,3,4]. Traditional methods for ASC tasks comprise feature extraction and classification models [2,3,4]. These methods first extract hand-crafted features such as the Mel-frequency cepstral coefficient (MFCC) and then classify the category of sound events using the extracted features and machine learning algorithms, such as support vector machines (SVMs) and K-nearest neighbors (KNN). However, these methods are limited in their ability to increase the classification performance.
To use the superior discriminative power of deep neural networks, deep-learning-based methods have been proposed for ASC tasks and have shown better classification performance than traditional methods. Most deep-learning-based methods use the convolutional neural network (CNN) architecture, and ResNet [9] is widely used as a backbone network in CNNs because of its high classification accuracy [10,11,12,13,50].
Although deep-learning-based methods show superior performance, they do not flexibly reflect the changes in sensor devices that gather sound events and the environments that determine the class of sound events. To deal with these changes that affect the performance of the trained network, re-training or fine-tuning of the networks, which incurs a high computational cost, is required.
CL involves a model that continually learns new tasks without forgetting the knowledge from preceding tasks [17,18,19,20,22,39,40,41,42,43]. CL can flexibly handle changes in tasks and incurs a lower computational cost than re-training or fine-tuning the network model. Although CL has gained considerable interest in various fields in recent years, few studies have focused [17,21] on ASC tasks to date.

2.2. Benchmark of Continual Learning Methods

Extensive research on CL is actively being conducted in the field of computer vision, and other fields such as NLP and HAR have received substantial attention. Thus, numerous methods have been proposed for CL; however, all these methods have reported different performances using different scenarios, different datasets for training the model, and different experimental environments. Therefore, there is a growing interest in benchmark studies to compare the performances of CL methods in a fair manner.
In a previous study by [17], biological research and a complete learning system were compared based on the plasticity and stability of the human brain. In addition, the performances of the CL methods, including traditional methods and deep-learning-based methods, were summarized for image and audio classification tasks. Another study benchmarked CL methods for image classification tasks in the field of computer vision [22]. In addition, refs. [39,40] applied regularization-based and replay-based methods to HAR tasks and proposed a framework to evaluate the performance of such techniques on various public HAR datasets. Ref. [19] compared the performance of CL methods for various NLP tasks, including word and sentence representations, language modeling, and question answering. Although studies comparing the performance of CL methods have been conducted in various fields, a comprehensive benchmark study on ASC tasks is yet to be conducted.
CL scenarios constitute one of the most important factors in benchmark studies on CL methods because they determine the accessibility to the current and previous tasks and prior knowledge, such as the task ID. Widely used scenarios for CL are task-, domain-, and class-incremental scenarios [41].
In addition, a previous study [20] used an offline approach that can access all data for the current and previous tasks multiple times. However, these settings are unrealistic; therefore, a recent benchmark study [22] was conducted using more realistic scenarios. It used an online approach that can access only one small batch and cannot access the previous batch data and data of the preceding tasks. Another ASC study [21] using a task-incremental setting and an offline approach has also been reported. However, considering the features of sound data, such as privacy issues and device heterogeneity, a benchmark study using incremental class and domain settings and an online approach is necessary. Further, it is necessary to explore the performance of CL methods in ASC tasks because the characteristics of sound data are different from those of language, sensor, and image data. Thus, in this paper, we propose more realistic and difficult scenarios, such as OCI and ODI, for various publicly available sound datasets.

3. Problem Definition for Online Continual Learning in Acoustic Scene Classification Tasks

The performance of CL methods is significantly influenced by the scenarios that define the problem. Therefore, in this section, we describe the scenarios for the ASC tasks.
CL scenarios have a sequential set of tasks,  T = T 1 , , T t . At time t, CL methods receive the data and information on task t. According to the input data and information, typical CL scenarios can be categorized into three types: task-, class-, and domain-incremental scenarios. A conceptual illustration of the three CL scenarios is depicted in Figure 1. As shown in Figure 1, task-incremental scenarios have a task ID as prior information, unlike in the other scenarios. At time t in task-incremental scenarios, data distribution,  D t = ( x 1 , y 1 ) , , ( x n , y n ) , t a s k _ I D , consists of the task ID and n samples with input data x and label y. Because the neural network cannot recognize the task ID in the real world, the task-incremental scenario is unrealistic. In the case of the class-incremental scenario, each task takes different input data and class distribution without a task ID. Data distributions of the class-incremental scenario are the same as those of task-incremental scenarios without the task ID,  D t = ( x 1 , y 1 ) , , ( x n , y n ) . Without the task ID, class-incremental scenarios are more difficult and realistic compared with the task-incremental scenarios. In the domain-incremental scenario, each task takes different input data but may have the same class distribution. Data distributions of domain-incremental scenarios are represented by  D t = ( x 1 , y 1 ) , , ( x n , y n ) , where  x  represents a different input augmented from a previous task and y represents the same label with a previous task. To mimic a domain change that has different input data for the same class, data-augmentation techniques such as noise addition and SpecAugment [51,52] are used for domain-incremental scenarios.
CL scenarios are also classified into offline and online approaches based on data accessibility for current and previous tasks during model training. A conceptual illustration of the offline and online approaches for class-incremental scenarios is shown in Figure 2. As shown in Figure 2, the offline approach for CL can access all data from the first task to the current task, T, when training the network model at the time of task T. However, the online approach for CL can only use the current batch data of task T when training the network model for task T.
Recently, applications of ASC tasks have received substantial attention for real-time sound recognition [5,7] and sound-based monitoring systems for smart factories [6]. Realistic CL scenarios that require limited computing resources and no prior information are necessary for such applications. Therefore, in this paper, we propose OCI and ODI scenarios for various public datasets to consider real scenarios.

4. Comparison of Continual Learning Methods

Existing CL methods can be classified into three categories: regularization-based, replay-based, and dynamic architecture-based. Only a few studies using dynamic architecture-based methods have been conducted for online scenarios. In addition, dynamic architecture-based methods require considerable time to adapt the network structure to the current task dynamically. Therefore, in this study, we focused on regularization-based and replay-based methods.
The network architecture can vary depending on the CL scenario [25]. For task-incremental scenarios, a neural network with a multihead architecture has been used [36,39,40]. In contrast, a neural network with a single-head architecture has been used to evaluate the performance of CL methods for class- and domain-incremental scenarios [22,38,44]. Therefore, we used a single-head architecture for all CL methods and modified the network to apply online scenarios.

4.1. Regularization-Based Methods

Regularization-based methods alleviate catastrophic forgetting by preserving weights important for previous tasks while updating the weights of the neural network. Representative methods include elastic weight consolidation (EWC) [24] and learning without forgetting (LwF) [23].
EWC mitigates catastrophic forgetting by selectively decelerating the update of weights according to the impact of previous works. The Fisher information matrix is used to define the importance of the network parameters, and the regularization loss with the Fisher information matrix is used to update the parameters for the new task while maintaining the parameters learned in the previous tasks. Given network parameters  θ , A, and B, the loss function of EWC is defined as [24]
L ( θ ) = L B ( θ ) + i λ 2 F i ( θ i θ A , i * ) 2
where F, i λ , and  L B ( θ )  represent the Fisher information matrix, each parameter, weight when considering the previous task, and loss for task B, respectively. However, EWC cannot be applied directly to online scenarios [22], and, therefore, we used EWC++ [25], which can set online scenarios for the experiments.
For efficient computation and memory usage, EWC++ maintains a single diagonal Fisher matrix, F, and uses the moving average to update the Fisher matrix.
F θ t = α F θ t + ( 1 α ) F θ t 1
where EWC++ calculates  F θ t  with the current batch and  α  is a hyperparameter between 0 and 1. The Fisher matrix computed by EWC++ eliminates an additional pass over the dataset, and thus it can efficiently use memory by only storing two sets of the Fisher matrix.
LwF [23] trains a neural network with only new task data and uses knowledge distillation to preserve previous capabilities. The output of the old network is computed using new task data, which are used as soft labels for knowledge distillation. Given that the network parameter,  θ , consists of shared parameters, task-specific parameters for each task, and parameters for the current tasks, the loss function of LwF is defined as [23]
L ( θ ) = λ L o l d ( Y o , Y ^ o ) + L n e w ( Y n , Y ^ n ) + R ( θ )
where  Y o Y n Y ^ o , and  Y ^ n  represent the output of the old network for the new task data, ground truth for the new task data, predictions of the current network for old tasks, and predictions of the current network for current tasks, respectively.  λ  represents the weight that sets the degree to which a previous task affects a new task.

4.2. Replay-Based Methods

Replay-based methods store the information of the examples of preceding tasks in a memory buffer [29,30,31,32,33,35,36,38,44] or generative model [26,27]. Then, examples of the preceding tasks and new data for the current task are used together to train the network model for the current task. Therefore, replay-based methods consist of three steps: memory retrieval, model update, and memory update. In the memory-retrieval step, old examples from the stored information are retrieved using memory-retrieval algorithms. Then, in the model-update step, the parameters of the network model are trained using old examples and new data for the current task. Finally, in the memory-update step, some new task data that have representative features for the current task are stored in the memory buffer or reflect the generative model by training the network model.
Training the network model using all older examples and new data can mitigate catastrophic forgetting; however, storing all examples is inefficient because of the limited memory space and data redundancy. Thus, experience replay (ER) [28,35] uses a sampling method to select a limited number of examples. The reservoir sampling method was used to replace the old samples in the memory-update step, and the random sampling method was used to select the old samples in the memory buffer. In addition, ER uses cross-entropy loss to train the neural network with new data and old examples stored in the memory buffer. Recently, Ref. [35] highlighted that, although ER is simple, it exhibits superior performance compared with conventional CL methods.
Greedy sampler and dumb learner (GDumb) [31] was originally designed to solve the open-set recognition problem; however, it also solves the CL problem with superior performance. GDumb comprises a greedy sampler and a dumb learner. When new task data arrive, the greedy sampler manages the memory buffer to maintain the balance of the sample by category. The dumb learner trains the neural network from scratch using examples provided by the greedy sampler.
Gradient episodic memory (GEM) [32] uses the information of the loss gradient vector calculated using examples in the memory buffer to decide whether to update the parameters. Because GEM calculates the loss gradient vector for each previous task, it incurs considerable computational cost. In contrast, averaged gradient episodic memory (A-GEM) [33] uses the average loss gradient vector calculated for all previous tasks to reduce the computational costs. From the viewpoint of network training, GEM tries to decrease the loss of examples per task, whereas A-GEM tries to decrease the average loss of examples for all tasks. We used A-GEM, which is an efficient version of GEM, to compare the performance of various CL methods.
Gradient-based sample selection (GSS) [29] also formulates the CL problem as a constrained optimization problem, similar to GEM and A-GEM. However, GEM and A-GEM focus on the model-update step, whereas GSS focuses on the memory-update step by retrieving examples with a sample selection algorithm formulated as a constraint reduction problem. GSS selects examples that maximize the diversity of samples in the memory buffer by analyzing the parameter gradient. Two online sample selection algorithms were proposed for the GSS: GSS-integer quadratic programming (IQP) and GSS-greedy. In this study, we adopted the GSS-greedy method used in [22] for our experiments.
Incremental classifier and representation learning (iCaRL) [36] introduces the nearest-class-mean (NCM) classifier [53] for class-incremental learning and has a superior classification accuracy because it uses the mean of exemplar images robust to small changes in data. iCaRL manages memories for replay so that exemplars have priority, and the number of exemplars in each class is kept the same. To train the network model, the loss, consisting of classification and distillation terms [37], is calculated using exemplar sets and new task data. The labels of the new task data are classified by comparing the distance between the average feature vectors of all exemplars for each class and the input feature vector.
Maximally interfered retrieval (MIR) [30] was proposed to increase the performance of the CL method by improving the memory sampling method for replay. Unlike random sampling, which is a widely used sampling method for memory retrieval, MIR reduces catastrophic forgetting by selecting samples that interfere the most with new tasks from among the samples of old tasks used for training the network model. MIR can be applied to both the experience replay method and generative replay method, and can be used with other replay-based methods such as ER, GEM, and iCaRL. In this study, we used ER with MIR as the MIR method for the experiments.
Adversarial Shapley value experience replay (ASER) [38] introduced a method that can evaluate the effect of samples in terms of the performance of the training model and used them in the memory-retrieval step. ASER uses the Shapley value (SV) to measure the contribution of each data point; the SV is calculated by measuring the marginal improvement in the classification accuracy as each sample is added. Because the SV requires computational complexity,  O ( N 2 ), to calculate the training dataset, ASER uses the KNN-SV method [54] that can be computed in  O ( N l o g N )  to speed up the computation of SV. ASER uses the adversarial SV (ASV), containing the cooperative and adversarial terms, as [38]
A S V ( i ) = max j S s u b s j ( i ) min k B n s k ( i )
where  S s u b  represents the subset of memories,  B n  represents the current batch data, and  s j ( i )  denotes the marginal contribution of each sample, i, with regard to j. In Equation (4), the first term is related to the cooperative notion that has a high positive SV for data in the memory buffer; the second term is related to the adversarial notion that has a negative SV with a large magnitude for new task data.
Replay-based methods suffer from class imbalance between the preceding and current tasks. SoftMax classifiers cannot effectively deal with the problem of class imbalance. Thus, supervised contrastive replay (SCR) [44] substitutes SoftMax classifiers with an NCM classifier [53] and introduces supervised contrastive loss to create better clusters in the feature space of the NCM classifier. The experimental results of SCR show that it outperforms the existing methods in online CL.

5. Experiments

5.1. Acoustic Scene Classification Datasets

To evaluate the performance of the CL method on ASC tasks, we used three public datasets of ASC tasks: DCASE-2021 [8], UrbanSound8K [48], and ESC-50 [49]. Table 1 lists the detailed information of the three datasets.
DCASE-2021 is an audio dataset used for classifying acoustic scenes in Task 1 A of the DCASE 2021 challenge. It has 10 classes, and the numbers of samples for training and validation are 13,962 and 2970, respectively. We pre-processed the audio data in a manner similar to that in [55]. We calculated 128 log-Mel energies under the original sampling rate of 44.1 kHz for each time slice by taking 2048 FFT points with a 50% overlap. After pre-processing, the input size of the audio data was 423 × 128 × 3. In DCASE-2021, we compared the methods using various memory buffer sizes including 200 (0.2k), 500 (0.5k), 1000 (1k), 2000 (2k), and 5000 (5k). In addition, we used the OCI/ODI scenarios to evaluate the performance of all the aforementioned CL methods on the DCASE dataset.
UrbanSound8K is a widely used dataset for evaluating the performance of the environmental sound classification methods, and contains 8732 labeled environmental sound clips of up to 4 s for 10 classes. To pre-process audio data, we ensured that all audio clips were 4 s long through zero padding, as in [56,57]. We extracted features with a 128-dimensional log-Mel filterbank at a sampling rate of 11,025 Hz for each time slice by taking 2048 FFT points with a 50% overlap. For UrbanSound8K, we used the same settings as for DCASE-2021 for the size of the memory buffer and online scenarios.
ESC-50 is a dataset used for classifying environmental sounds, and consists of 2000 labeled environmental recordings equally balanced between 50 classes of 40 clips per class. We divided the ESC-50 into 1400 audio clips for training and 600 audio clips for evaluation. Because the total number of samples for ESC-50 was 2000, a memory buffer larger than 2k was not adequate for an online scenario setting. Therefore, we experimented with the compared methods to obtain memory buffer sizes of 0.2, 0.5, and 1k. We pre-processed the dataset similar to [56,57], sampled the audio at 32k Hz, and converted the audio waveform to a sequence of 128-dimensional log-Mel filterbank features computed with a 25 ms Hamming window every 10 ms. We used the ODI scenario only for ESC-50.

5.2. Scenarios of Continual Learning for Acoustic Scene Classification Tasks

Considering the features of sound data, such as privacy issues, the online approach is appropriate for ASC tasks because it accesses only one small batch and cannot access the previous batch data and data of preceding tasks. In addition, class- and domain-incremental scenarios are useful for considering the characteristics of sound data, which can vary depending on the microphone of the devices and data-gathering environments. Thus, in this study, we used more realistic and difficult scenarios, such as OCI and ODI, to evaluate the performance of the CL methods.

5.2.1. OCI Scenario

DCASE-2021 and UrbanSound8K with 10 classes were divided into five tasks, with two classes for each task. ESC-50 with 50 classes was split into 10 tasks, and each task contained five classes.

5.2.2. ODI Scenario

The datasets for domain-incremental scenarios used in computer vision were permutated MNIST and Cifar datasets with a nonstationary distribution caused by noise, blur, and occlusion [22]. To emulate the domain change in the sound data, we applied data-augmentation techniques such as noise addition and SpecAugment [51,52] for the ODI scenarios. For each ODI scenario, we divided the five tasks into DCASE-2021 and UrbanSound8K.

5.3. Evaluation Metrics

To evaluate the performance of CL methods in the two CL scenarios, we measured the accuracy and extent of forgetting per task and calculated the average accuracy and forgetting for all tasks, given that the accuracy of task i a T , i  is trained on task T, and the average accuracy,  A T , is defined as [22] A T = 1 T i = 1 T a T , i .
Average forgetting measures the degree of forgetfulness of each preceding task after training the neural network for the last task, T. The average forgetting,  F T , was calculated using [22] F T = 1 T 1 i = 1 T 1 f T , i , where  f T , i = max k 1 , , t 1 a k , i a T , i .
Because the average accuracy is related to the intransigence and the average forgetting is related to catastrophic forgetting, we can evaluate the performance of the CL methods using the two metrics.

5.4. Implementation Details

ResNet is widely used as a backbone network for classifying acoustic scenes and has demonstrated superior performance [10,11,12,13]. Hence, we adopted the reduced ResNet-18 model, considering the lack of memory and long training time [22,38,44]. We set the batch size to 10 and used the stochastic gradient descent optimizer with a tuned learning rate to train the reduced ResNet-18 model. We use single-head architecture with output units for all classes to experiment with the CL method in the OCI and ODI scenarios. As the first task, we tuned the learning rate with various learning rates set from 0.0001 to 0.1. Then, we trained the reduced ResNet-18 model with a tuned learning rate from the first task.
To compare the performance of CL methods on ASC tasks, we modified the benchmark framework of CL for computer vision tasks [22,44]. To tune the hyperparameters without violating the online CL setting, previous studies [22,33] separated the data streams for cross-validation and for training and evaluation. They used the data stream of the first and second tasks for tuning the hyperparameters, and the data stream of the other tasks for evaluating the performance. Similar to previous studies, we used the data stream of the first and second tasks for tuning the hyperparameters. However, we evaluated the performance from the first task to the last task using optimal hyperparameters because this is more suitable for an online setting and yields results similar to realistic scenarios.

6. Experimental Results

We evaluated the performance of the three baseline methods to compare the CL methods with conventional methods: offline, online, and fine-tuning. The offline method trains the network model using all data of the dataset for 50 epochs with a mini-batch size of 128, and this provides the upper bound of the performance for the CL methods. The online method trains the network model using all data of the dataset for one epoch with a mini-batch size of 10; all data of the dataset can be seen once. The fine-tuning method always updates the network model for the current task and provides a lower-bound performance for the CL methods. For the OCI scenario, we analyzed ten recent CL methods comprising two regularization-based methods and eight replay-based methods. In the ODI scenario, we analyzed nine CL methods excluding iCaRL, because iCaRL is not suitable for the ODI scenarios. iCaRL updates the model with distillation loss calculated from the data of old classes; however, the ODI scenarios have different input data for the same class from all tasks.

6.1. Performance Comparison in the Online Class-Incremental Scenario

6.1.1. Overall Analysis

From our studies, we showed that the regularization-based methods demonstrated a performance in terms of accuracy similar to that of the fine-tuning methods, providing a lower-bound performance; however, LwF showed the best performance in terms of forgetting compared with other methods, including the replay-based methods. This result indicates that LwF is good for catastrophic forgetting, but not for intransigence.
The replay-based methods showed superior performance compared with the baseline and regularization-based methods. GDumb, iCaRL, and SCR showed superior performance in terms of accuracy compared with other methods. This tendency is similar to the benchmark results of the CL methods for computer vision tasks [22]. These experimental results show that we must consider both accuracy and forgetting metrics to devise a CL method for ASC tasks.

6.1.2. Average Accuracy

Table 2 summarizes the experimental results of the average accuracy for the OCI scenario in ASC tasks. For the three datasets, the replay-based methods showed higher accuracies compared with the regularization-based methods, and the regularization-based methods had an accuracy similar to that of the fine-tuning methods, which provide lower-bound performance. Among the replay-based methods, iCaRL, GDumb, and SCR had higher accuracies compared with the other replay-based methods. If the size of the memory buffer was large in the three datasets, GDumb had the highest accuracy. The experimental result of GDumb for a buffer size of over 0.5k in the ESC-50 dataset shows that GDumb had the best performance by a large margin. For small buffer sizes (0.2k and 0.5k), iCaRL had the best accuracy in DCASE-2021, and SCR had the best accuracy in UrbanSound8K. In addition, in ESC-50, the performance of SCR for a 0.2k buffer size was the best. Among the replay-based methods, A-GEM showed the lowest performance for all sizes of memory buffers in all datasets and it was similar to the baseline methods and regularization-based methods. This tendency is similar to the benchmark results of the CL methods for computer vision tasks [22]. Although MIR has the best performance in terms of accuracy in CORe-50-NC for image classification tasks [22], it did not show impressive results in the ASC tasks. Experimental results indicate that the performance of the numerous methods was not affected by the size of the memory buffer; however, the performance of GDumb increased in proportion to the size of the memory buffer. Of all CL methods, GDumb had the best performance, which was similar to that of offline methods with the largest buffer size in all datasets. In addition, the size of the memory buffer with the highest accuracy differs depending on the method. Therefore, the size of the memory buffer is another important parameter to consider when deploying CL methods for the ASC tasks.

6.1.3. Average Forgetting

Table 3 presents the experimental results of the average forgetting for the OCI scenario in the ASC tasks. Average forgetting is related to average accuracy because average forgetting indicates the extent to which the network model forgets each preceding task after being trained on the last task, T. A previous study [22] on computer vision tasks showed that a method with high accuracy shows lower forgetting, and this is similar to our experimental results. Among the replay-based methods for ASC tasks, SCR, iCaRL, and GDumb showed superior performance in terms of average accuracy and forgetting. In most cases, SCR exhibited the best performance in terms of average forgetting among the replay-based methods. iCaRL had the highest accuracy and lowest forgetting in the case of a buffer size of 0.2k on DCASE-2021. GDumb had the highest accuracy and lowest forgetting when the memory buffer sizes were 2k and 5k for UrbanSound8K. In DCASE-2021 and ESC-50, LwF had the lowest average forgetting among the compared methods; however, it had lower performance in terms of average accuracy. This result indicates that LwF is good for catastrophic forgetting, not intransigence. EWC showed a performance in terms of average forgetting similar to that of the fine-tuning method. The experimental results of the average accuracy showed that ASER, ER, GSS, A-GEM, and MIR did not show good performance in terms of average forgetting. The experimental results showed that LwF could alleviate only intransigence, and GDumb, iCaRL, and SCR could significantly reduce both catastrophic forgetting and intransigence. Therefore, we must consider both the accuracy and forgetting metrics to devise a CL method for ASC tasks.

6.2. Performance Comparison in Online Domain-Incremental Scenario

6.2.1. Overall Analysis

In the two ODI scenarios, the performance gap between the CL methods and fine-tuning methods was lower than that of the OCI scenario, which indicates that the ODI scenario is more difficult than the OCI scenario. In ODI scenarios, regularization-based methods have similar performances in terms of the average accuracy of the fine-tuning method. However, both LwF and EWC perform better than the fine-tuning method. This implies that regularization-based methods such as LwF and EWC are suitable for catastrophic forgetting in the ODI scenario. Most replay-based methods, except for A-GEM, have performances in terms of average accuracy and average forgetting that are superior compared with the fine-tuning method. As in the experimental results for the OCI scenario, GDumb had the highest accuracy when used with a large-size memory buffer in DCASE-2021; however, A-GEM had a lower accuracy compared with the regularization-based methods and fine-tuning method in DCASE-2021 and UrbanSound8K. SCR consistently outperformed the compared methods in terms of accuracy. These experimental results show that we can consider SCR first in ODI scenarios.

6.2.2. Average Accuracy

Table 4 and Table 5 list the experimental results of the average accuracy for ODI scenarios in DCASE-2021 and UrbanSound8K, respectively. In DCASE-2021 and UrbanSound8K, the average accuracy of regularization-based methods, including LwF and EWC, was similar to that of the fine-tuning method. This result showed a tendency similar to that of previous work [22] for image classification tasks. However, replay-based methods have different tendencies. ER and MIR have shown impressive performance, which outperforms the fine-tuning method in image classification tasks; however, they do not show the best performance in ASC tasks. In contrast, GDumb and SCR have shown superior performance compared with other methods, unlike the results of image classification tasks. In DCASE-2021 with a 5k buffer size, GDumb had the best performance among all CL methods. The SCR consistently outperformed the compared methods in terms of accuracy in most cases, except for DCASE-2021 with a 5k buffer size. SCR with a 5k memory buffer had higher accuracy compared with the offline method on UrbanSound8K. SCR adopts supervised contrastive learning, and, therefore, augmenting the training data is effective for ODI scenarios with a different distribution of input data for the same class. In addition, GDumb had the highest accuracy with a 5k memory buffer in the two types of scenarios on the DCASE-2021. In most cases, the A-GEM had a lower average accuracy compared with the fine-tuning method. Similar to the results of the OCI scenario, the A-GEM had a lower performance compared with the fine-tuning method. ASER and GSS did not perform well in the ODI scenarios. From the experimental results on average accuracy, we must first consider SCR and GDumb.

6.2.3. Average Forgetting

Table 6 and Table 7 list the experimental results of the average forgetting for ODI scenarios in DCASE-2021 and UrbanSound8K, respectively. As shown in Table 3, LwF exhibited lower average forgetting compared with the fine-tuning method. Although the average forgetting of EWC in the OCI scenario showed lower performance, EWC in the ODI scenario showed good performance, which was similar to that of LwF. Regularization-based methods such as LwF and EWC had relatively lower performance, similar to the results [44] of image classification tasks. Thus, the regularization-based method in the ODI scenario is suitable for catastrophic forgetting. Replay-based methods exhibit various performances, from bad to good. Our results indicate that A-GEM, ASER, and ER showed lower performances compared with replay-based methods, and this tendency is similar to the results of the OCI scenario. For small sizes (0.2k), MIR exhibited good performance in the noise of DCASE-2021. This result is similar to that of the OCI scenario. Unlike the results of the OCI scenario among the replay-based methods, GSS for 0.2k, 0.5k, and 1k in SpecAugment of DCASE-2021 showed a higher performance in terms of average forgetting. The average forgetting of GSS for 0.2k shows that its performance is the best among all CL methods in DCASE-2021. The SCR outperformed the compared methods in terms of the average forgetting in most cases of ODI scenarios, whereas LwF outperformed the compared methods in terms of the average forgetting in the OCI scenario. Considering both the average accuracy and average forgetting, we found that SCR is the most suitable method for ODI scenarios.

6.3. Training Time

We measured the training time of each CL method because training time is a key factor for assessing the performance of CL methods. For a fair comparison, we excluded the time for grid search because each method has a different search space, which significantly affects the search time. We experimented with CL methods with all memory buffers from 0.2k to 5k and applied the OCI scenario on the DCASE-2021 dataset. The average training time was calculated by taking the average over the sum of the training times; this was repeated three times on an NVIDIA RTX 8000 GPU.
The average training time, set to a memory buffer size of 1k, for CL methods on DCASE-2021 is shown in Figure 3. In Figure 3, the training time is plotted on a logarithmic scale, and the offline method using 50 epochs had the longest training time compared with the other methods. Replay-based methods have a longer training time compared with regularization-based methods because replay-based methods need to manage the memory buffer to retrieve old samples or update the samples. Methods with superior performances such as GDumb and SCR have a longer training time among the replay-based methods. iCaRL showed good performance in terms of both accuracy and training time, whereas GSS showed lower performance in both accuracy and training time. The experimental results showed that the average accuracy of the CL methods was highly related to training time, except in the case of GSS.
A previous study [22] showed only the training time for the compared methods on a specific memory buffer size. However, the size of the memory buffer is a very important factor for determining the performance of replay-based methods, as shown in Section 6.1 and Section 6.2. Therefore, we assessed the training time of the replay-based methods for various sizes of memory buffers. We applied the OCI scenario to DCASE-2021 to measure the training time.
The average training time and accuracy of the replay-based methods with all memory buffer sizes from 0.2k to 5k are shown in Figure 4. In Figure 4, the experimental results show that ASER, ER, and MIR had a constant training time, regardless of the size of the memory buffer; their performance increased with an increase in the size of the memory buffer. The training time of ASER was slightly reduced in proportion to the size of the memory because the large buffer size reduced the number of memory updates.
In contrast, A-GEM had a constant training time for the size of the memory buffer; however, the performance of A-GEM decreased with an increase in the size of the memory buffer. The training times of GSS, iCaRL, and SCR slightly increased with an increase in the size of the memory buffer; however, their performance irregularly increased or decreased by a large margin with an increase in the size of the memory buffer.
Although the average accuracy of GDumb increased in proportion to the size of the memory buffer, the training time of GDumb increased significantly according to the memory buffer size. The training time of GDumb was 242 s when the memory buffer size was 0.2k; however, the training time of GDumb increased to 4753 s when the memory buffer size increased to 5k.

7. Conclusions

We used CL for ASC tasks for realistic scenarios considering the features of sound data such as privacy issues and device heterogeneity. The online approach that accesses only one small batch and cannot access the previous batch data and data of preceding tasks is appropriate for ASC tasks because of privacy issues; hence, the proposed scenarios are based on an online approach. The proposed scenarios contain OCI and ODI scenarios that reflect the device heterogeneity and change in the data-gathering environments. We defined OCI and ODI scenarios using three public datasets: DCASE-2021, UrbanSound8K, and ESC-50. With the two proposed scenarios on three public datasets, we systematically analyzed the performance of ten recent CL methods (two regularization-based methods and eight replay-based methods) for the OCI scenario, and nine CL methods (except for iCaRL). iCaRL, which requires both old and new classes to use knowledge distillation, is not suitable for ODI scenarios that have the same class for all tasks.
In the OCI scenarios, regularization-based methods showed a performance similar in terms of accuracy to the fine-tuning method, which provides a lower-bound performance; however, LwF showed the best performance in terms of forgetting compared with the other methods, which includes replay-based methods. Replay-based methods showed superior performance compared with the baseline and regularization-based methods. GDumb, iCaRL, and SCR showed superior performance in terms of accuracy compared with the other methods; this tendency is similar to the benchmark results of the CL methods for computer vision tasks. In the ODI scenarios, regularization-based methods showed a performance similar in terms of average accuracy and average forgetting to that of the fine-tuning method. Most replay-based methods had superior performance in terms of the average accuracy and average forgetting compared with the fine-tuning method. In the two ODI scenarios on the two datasets, SCR consistently outperformed the compared methods in terms of accuracy. In addition, we analyzed the training time of each replay-based method for various memory buffer sizes. Most replay-based methods containing ASER, ER, and MIR had almost constant training times, regardless of the size of the memory buffer; their performance increased with an increase in the size of the memory buffer. In addition, the training times of GSS, iCaRL, and SCR slightly increased with an increase in the memory buffer size. Although the average accuracy of GDumb increased in proportion to the size of the memory buffer, the training time of GDumb significantly increased with an increase in the memory buffer size.
We focused on analyzing the performance of regularization-based and replay-based methods for ASC tasks. Because dynamic-architecture-based methods are representative methods for CL, we plan to analyze the performance of dynamic-architecture-based methods for ASC tasks. The experimental results of this study indicate that most replay-based methods with superior accuracy suffer from a long training time. Therefore, in future work, we plan to devise an efficient replay-based method that has a short training time for various memory buffer sizes and superior performance in terms of accuracy and forgetting.

Author Contributions

Conceptualization, D.H., M.K. and C.Y.J.; methodology, D.H., M.K. and C.Y.J.; software, D.H.; validation, D.H., M.K. and C.Y.J.; formal analysis, D.H. and C.Y.J.; resources, M.K. and C.Y.J.; writing—original draft preparation, D.H.; writing—review and editing, D.H., M.K. and C.Y.J.; supervision, M.K.; project administration, M.K. and C.Y.J.; All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Electronics and Telecommunications Research Institute (ETRI) grant by the Korean government (23ZS1200, Fundamental Technology Research for Human-Centric Autonomous Intelligent Systems).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ASCAcoustic scene classification
ASERAdversarial Shapley value experience replay
ASVAdversarial Shapley value
A-GEMAveraged gradient episodic memory
CLContinual learning
DCASEThe detection and classification of acoustic scenes and events
ERExperience replay
EWCElastic weight consolidation
GDumbGreedy sampler and dumb learner
GEMGradient episodic memory
GSSGradient-based sample selection
HARHuman activity recognition
iCaRLIncremental classifier and representation learning
IDIdentifier
LwFLearning without forgetting
MIRMaximally interfered retrieval
NCMNearest-class-mean
NLPNatural language processing
OCIOnline class incremental
ODIOnline domain incremental
SCRSupervised contrastive replay
SVShapley value

References

  1. Mu, W.; Yin, B.; Huang, X.; Xu, J.; Du, Z. Environmental sound classification using temporal-frequency attention based convolutional neural network. Sci. Rep. 2021, 11, 21552. [Google Scholar] [CrossRef]
  2. Chachada, S.; Kuo, C.C.J. Environmental sound recognition: A survey. APSIPA Trans. Signal Inf. Process. 2014, 3, e14. [Google Scholar] [CrossRef] [Green Version]
  3. Sophiya, E.; Jothilakshmi, S. Deep Learning Based Audio Scene Classification. In Proceedings of the International Conference on Computational Intelligence, Cyber Security, and Computational Models; Springer: Berlin/Heidelberg, Germany, 2017; pp. 98–109. [Google Scholar] [CrossRef]
  4. Abeßer, J. A Review of Deep Learning Based Methods for Acoustic Scene Classification. Appl. Sci. 2020, 10, 2020. [Google Scholar] [CrossRef] [Green Version]
  5. Kücüktopcu, O.; Masazade, E.; Ünsalan, C.; Varshney, P.K. A real-time bird sound recognition system using a low-cost microcontroller. Appl. Acoust. 2019, 148, 194–201. [Google Scholar] [CrossRef]
  6. Kim, J.; Lee, H.; Jeong, S.; Ahn, S.H. Sound-based remote real-time multi-device operational monitoring system using a convolutional neural network (CNN). J. Manuf. Syst. 2021, 58, 431–441. [Google Scholar] [CrossRef]
  7. Ha, D.; Kim, M.; Moon, K.; Jeong, C.Y. Accelerating On-Device Learning with Layer-Wise Processor Selection Method on Unified Memory. Sensors 2021, 21, 2364. [Google Scholar] [CrossRef]
  8. Mesaros, A.; Heittola, T.; Virtanen, T. A multi-device dataset for urban acoustic scene classification. In Proceedings of the Detection and Classification of Acoustic Scenes and Events 2018 Workshop (DCASE2018), Surrey, UK, 19–20 November 2018; pp. 9–13. [Google Scholar]
  9. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  10. Suh, S.; Park, S.; Jeong, Y.; Lee, T. Designing Acoustic Scene Classification Models with CNN Variants; Technical report, DCASE2020 Challenge; IEEE: Manhattan, NY, USA, 2020. [Google Scholar]
  11. Koutini, K.; Eghbal-zadeh, H.; Widmer, G.; Kepler, J. CP-JKU Submissions to DCASE’19: Acoustic Scene Classification and Audio Tagging with REceptive-Field-Regularized CNNs. In Proceedings of the Detection and Classification of Acoustic Scenes and Events Workshop (DCASE), New York, NY, USA, 25–26 October 2019; pp. 25–26. [Google Scholar]
  12. McDonnell, M.D.; Gao, W. Acoustic Scene Classification Using Deep Residual Networks with Late Fusion of Separated High and Low Frequency Paths. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 141–145. [Google Scholar] [CrossRef]
  13. Liu, M.; Wang, W.; Li, Y. The System for Acoustic Scene Classification Using Resnet; Technical report, DCASE2019 Challenge; IEEE: Manhattan, NY, USA, 2019. [Google Scholar]
  14. Hershey, S.; Chaudhuri, S.; Ellis, D.P.W.; Gemmeke, J.F.; Jansen, A.; Moore, R.C.; Plakal, M.; Platt, D.; Saurous, R.A.; Seybold, B.; et al. CNN architectures for large-scale audio classification. In Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 131–135. [Google Scholar] [CrossRef] [Green Version]
  15. Piczak, K.J. Environmental sound classification with convolutional neural networks. In Proceedings of the 2015 IEEE 25th International Workshop on Machine Learning for Signal Processing (MLSP), Boston, MA, USA, 17–20 September 2015; pp. 1–6. [Google Scholar] [CrossRef]
  16. Chung, S.; Jeong, C.Y.; Lim, J.M.; Lim, J.; Noh, K.J.; Kim, G.; Jeong, H. Real-world multimodal lifelog dataset for human behavior study. ETRI J. 2022, 44, 426–437. [Google Scholar] [CrossRef]
  17. Parisi, G.I.; Kemker, R.; Part, J.L.; Kanan, C.; Wermter, S. Continual lifelong learning with neural networks: A review. Neural Netw. 2019, 113, 54–71. [Google Scholar]
  18. Masana, M.; Liu, X.; Twardowski, B.; Menta, M.; Bagdanov, A.D.; van de Weijer, J. Class-incremental learning: Survey and performance evaluation on image classification. arXiv 2020, arXiv:2010.15277. [Google Scholar]
  19. Biesialska, M.; Biesialska, K.; Costa-jussà, M.R. Continual Lifelong Learning in Natural Language Processing: A Survey. In Proceedings of the 28th International Conference on Computational Linguistics, Barcelona, Spain (Online), 8–13 December 2020; International Committee on Computational Linguistics: Barcelona, Spain, 2020; pp. 6523–6541. [Google Scholar] [CrossRef]
  20. Delange, M.; Aljundi, R.; Masana, M.; Parisot, S.; Jia, X.; Leonardis, A.; Slabaugh, G.; Tuytelaars, T. A continual learning survey: Defying forgetting in classification tasks. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 3366–3385. [Google Scholar]
  21. Karam, S.; Ruan, S.J.; ul Haq, Q.M. Task Incremental Learning with Static Memory for Audio Classification without Catastrophic Interference. IEEE Consum. Electron. Mag. 2022, 11, 101–108. [Google Scholar] [CrossRef]
  22. Mai, Z.; Li, R.; Jeong, J.; Quispe, D.; Kim, H.; Sanner, S. Online continual learning in image classification: An empirical survey. Neurocomputing 2022, 469, 28–51. [Google Scholar] [CrossRef]
  23. Li, Z.; Hoiem, D. Learning Without Forgetting. In Proceedings of the Computer Vision—ECCV 2016, Amsterdam, The Netherlands, 11–14 October 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 614–629. [Google Scholar]
  24. Kirkpatrick, J.; Pascanu, R.; Rabinowitz, N.; Veness, J.; Desjardins, G.; Rusu, A.A.; Milan, K.; Quan, J.; Ramalho, T.; Grabska-Barwinska, A.; et al. Overcoming catastrophic forgetting in neural networks. Proc. Natl. Acad. Sci. USA 2017, 114, 3521–3526. [Google Scholar] [CrossRef]
  25. Chaudhry, A.; Dokania, P.K.; Ajanthan, T.; Torr, P.H. Riemannian Walk for Incremental Learning: Understanding Forgetting and Intransigence. In Proceedings of the ECCV, Munich, Germany, 8–14 September 2018. [Google Scholar]
  26. Shin, H.; Lee, J.K.; Kim, J.; Kim, J. Continual learning with deep generative replay. Adv. Neural Inf. Process. Syst. 2017, 30, 2994–3003. [Google Scholar]
  27. Wu, Y.; Chen, Y.; Wang, L.; Ye, Y.; Liu, Z.; Guo, Y.; Zhang, Z.; Fu, Y. Incremental classifier learning with generative adversarial networks. arXiv 2018, arXiv:1802.00853. [Google Scholar]
  28. Hayes, T.L.; Cahill, N.D.; Kanan, C. Memory efficient experience replay for streaming learRning. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 9769–9776. [Google Scholar]
  29. Aljundi, R.; Lin, M.; Goujaud, B.; Bengio, Y. Gradient based sample selection for online continual learning. Adv. Neural Inf. Process. Syst. 2019, 32, 11816–11825. [Google Scholar]
  30. Aljundi, R.; Belilovsky, E.; Tuytelaars, T.; Charlin, L.; Caccia, M.; Lin, M.; Page-Caccia, L. Online Continual Learning with Maximal Interfered Retrieval. In Proceedings of the Advances in Neural Information Processing Systems; Wallach, H., Larochelle, H., Beygelzimer, A., d’ Alché-Buc, F., Fox, E., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2019; Volume 32. [Google Scholar]
  31. Prabhu, A.; Torr, P.; Dokania, P. GDumb: A Simple Approach that Questions Our Progress in Continual Learning. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020. [Google Scholar]
  32. Lopez-Paz, D.; Ranzato, M. Gradient episodic memory for continual learning. Adv. Neural Inf. Process. Syst. 2017, 30, 6467–6476. [Google Scholar]
  33. Chaudhry, A.; Ranzato, M.; Rohrbach, M.; Elhoseiny, M. Efficient Lifelong Learning with A-GEM. In Proceedings of the ICLR, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  34. Lee, S.; Ha, J.; Zhang, D.; Kim, G. A neural dirichlet process mixture model for task-free continual learning. arXiv 2020, arXiv:2001.00689. [Google Scholar]
  35. Chaudhry, A.; Rohrbach, M.; Elhoseiny, M.; Ajanthan, T.; Dokania, P.K.; Torr, P.H.; Ranzato, M. On tiny episodic memories in continual learning. arXiv 2019, arXiv:1902.10486. [Google Scholar]
  36. Rebuffi, S.A.; Kolesnikov, A.; Sperl, G.; Lampert, C.H. icarl: Incremental classifier and representation learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2001–2010. [Google Scholar]
  37. Hinton, G.; Vinyals, O.; Dean, J. Distilling the knowledge in a neural network. arXiv 2015, arXiv:1503.02531. [Google Scholar]
  38. Shim, D.; Mai, Z.; Jeong, J.; Sanner, S.; Kim, H.; Jang, J. Online class-incremental continual learning with adversarial shapley value. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtually, 2–9 February 2021; Volume 35, pp. 9630–9638. [Google Scholar]
  39. Jha, S.; Schiemer, M.; Zambonelli, F.; Ye, J. Continual Learning in Sensor-based Human Activity Recognition: An Empirical Benchmark Analysis. Inf. Sci. 2021, 575, 1–21. [Google Scholar]
  40. Jha, S.; Schiemer, M.; Ye, J. Continual learning in human activity recognition: An empirical analysis of regularization. In Proceedings of the Thirty-Seventh International Conference on Machine Learning(ICML), Vienna, Austria, 12–18 July 2020. [Google Scholar]
  41. Van de Ven, G.M.; Tolias, A.S. Three scenarios for continual learning. arXiv 2019, arXiv:1904.07734. [Google Scholar]
  42. Belouadah, E.; Popescu, A.; Kanellos, I. A comprehensive study of class incremental learning algorithms for visual tasks. Neural Netw. 2021, 135, 38–54. [Google Scholar] [PubMed]
  43. Hayes, T.L.; Krishnan, G.P.; Bazhenov, M.; Siegelmann, H.T.; Sejnowski, T.J.; Kanan, C. Replay in Deep Learning: Current Approaches and Missing Biological Elements. Neural Comput. 2021, 33, 2908–2950. [Google Scholar] [CrossRef]
  44. Mai, Z.; Li, R.; Kim, H.; Sanner, S. Supervised Contrastive Replay: Revisiting the Nearest Class Mean Classifier in Online Class-Incremental Continual Learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 21–25 June 2021; pp. 3589–3599. [Google Scholar]
  45. Yoon, J.; Yang, E.; Lee, J.; Hwang, S.J. Lifelong Learning with Dynamically Expandable Networks. In Proceedings of the International Conference on Learning Representations (ICLR), Vancouver, BC, Canada, 30 April–3 May 2018; pp. 1–11. [Google Scholar]
  46. Pham, Q.; Liu, C.; Hoi, S. DualNet: Continual Learning, Fast and Slow. Adv. Neural Inf. Process. Syst. 2021, 34, 16131–16144. [Google Scholar]
  47. Rusu, A.A.; Rabinowitz, N.C.; Desjardins, G.; Soyer, H.; Kirkpatrick, J.; Kavukcuoglu, K.; Pascanu, R.; Hadsell, R. Progressive neural networks. arXiv 2016, arXiv:1606.04671. [Google Scholar]
  48. Salamon, J.; Jacoby, C.; Bello, J.P. A dataset and taxonomy for urban sound research. In Proceedings of the 22nd ACM International Conference on Multimedia, Taipei, Taiwan, 11–14 April 2014; pp. 1041–1044. [Google Scholar]
  49. Piczak, K.J. ESC: Dataset for Environmental Sound Classification. In Proceedings of the 23rd Annual ACM Conference on Multimedia, Brisbane, Australia, 26–30 October 2015; ACM Press: New York, NY, USA, 2015; pp. 1015–1018. [Google Scholar] [CrossRef]
  50. Jo, S.H.; Jeong, C.Y.; Moon, K.; Kim, C.K. Global-Local Three-Stream Network for Acoustic Scene Classification. In Proceedings of the 2021 International Conference on Information and Communication Technology Convergence (ICTC), Jeju Island, Republic of Korea, 20–22 October 2021; pp. 1567–1571. [Google Scholar] [CrossRef]
  51. Park, D.S.; Chan, W.; Zhang, Y.; Chiu, C.C.; Zoph, B.; Cubuk, E.D.; Le, Q.V. Specaugment: A simple data augmentation method for automatic speech recognition. arXiv 2019, arXiv:1904.08779. [Google Scholar]
  52. Wang, H.; Zou, Y.; Wang, W. Specaugment++: A hidden space data augmentation method for acoustic scene classification. arXiv 2021, arXiv:2103.16858. [Google Scholar]
  53. Mensink, T.; Verbeek, J.; Perronnin, F.; Csurka, G. Distance-based image classification: Generalizing to new classes at near-zero cost. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 2624–2637. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  54. Jia, R.; Dao, D.; Wang, B.; Hubis, F.A.; Gurel, N.M.; Li, B.; Zhang, C.; Spanos, C.; Song, D. Efficient Task-Specific Data Valuation for Nearest Neighbor Algorithms. Proc. VLDB Endow. 2019, 12, 1610–1623. [Google Scholar] [CrossRef] [Green Version]
  55. Gao, W.; McDonnell, M. Acoustic Scene Classification Using Deep Residual Networks with Focal Loss and Mild Domain Adaptation. Technical Report, Detection and Classification of Acoustic Scenes and Events 2020 Challenge, 1 March–1 July 2020. Available online: https://dcase.community/documents/challenge2020/technical_reports/DCASE2020_Gao_132.pdf (accessed on 29 June 2023).
  56. Gong, Y.; Chung, Y.A.; Glass, J. PSLA: Improving Audio Tagging with Pretraining, Sampling, Labeling, and Aggregation. IEEE/ACM Trans. Audio Speech Lang. Process. 2021, 29, 3292–3306. [Google Scholar] [CrossRef]
  57. Gong, Y.; Chung, Y.A.; Glass, J. AST: Audio Spectrogram Transformer. In Proceedings of the Interspeech 2021, Brno, Czechia, 30 August–3 September 2021; pp. 571–575. [Google Scholar] [CrossRef]
Figure 1. Conceptual illustration of the three types of CL scenarios. The task-incremental scenarios only have a task ID per data. The class-incremental scenarios have different input data and class distributions without the task ID. The domain-incremental scenario assumes that each task is set with different nonstationarity, different distributions of input data, and the same class distribution.
Figure 1. Conceptual illustration of the three types of CL scenarios. The task-incremental scenarios only have a task ID per data. The class-incremental scenarios have different input data and class distributions without the task ID. The domain-incremental scenario assumes that each task is set with different nonstationarity, different distributions of input data, and the same class distribution.
Sensors 23 06893 g001
Figure 2. Conceptual illustration of the offline and online approaches for class-incremental scenarios. The offline approach uses all data from previous tasks while training a neural network with a new task. In contrast, the online approach uses only new tasks.
Figure 2. Conceptual illustration of the offline and online approaches for class-incremental scenarios. The offline approach uses all data from previous tasks while training a neural network with a new task. In contrast, the online approach uses only new tasks.
Sensors 23 06893 g002
Figure 3. Comparison of training times for different models in the case of a memory buffer size with 1k. The training time is plotted on a logarithmic scale. Offline is the upper bound of the training time. Baseline and regularization-based methods have lower training times compared with replay-based methods. With a 1k memory buffer size, GSS, GDumb, and SCR have higher training times.
Figure 3. Comparison of training times for different models in the case of a memory buffer size with 1k. The training time is plotted on a logarithmic scale. Offline is the upper bound of the training time. Baseline and regularization-based methods have lower training times compared with replay-based methods. With a 1k memory buffer size, GSS, GDumb, and SCR have higher training times.
Sensors 23 06893 g003
Figure 4. Averaged training time and accuracy of different memory-based methods according to changes in the size of the memory buffer. The vertical bar indicates the averaged training time, and dots represent the accuracy of each model according to the buffer size. For easy identification, we used different colors according to the size of the memory buffer.
Figure 4. Averaged training time and accuracy of different memory-based methods according to changes in the size of the memory buffer. The vertical bar indicates the averaged training time, and dots represent the accuracy of each model according to the buffer size. For easy identification, we used different colors according to the size of the memory buffer.
Sensors 23 06893 g004
Table 1. Detailed information of the ASC datasets. The attributes include the number of classes, tasks, and total data. The number of classes per task is 2 in DCASE-2021 and UrbanSound8K, and 5 in ESC-50. The information contains the input size and online scenarios for experiments. DCASE-2021 and UrbanSound8K experiments with CL methods are under OCI and ODI scenarios, but ESC-50 experiments with CL methods are under only the OCI scenario.
Table 1. Detailed information of the ASC datasets. The attributes include the number of classes, tasks, and total data. The number of classes per task is 2 in DCASE-2021 and UrbanSound8K, and 5 in ESC-50. The information contains the input size and online scenarios for experiments. DCASE-2021 and UrbanSound8K experiments with CL methods are under OCI and ODI scenarios, but ESC-50 experiments with CL methods are under only the OCI scenario.
DatasetDCASE-2021UrbanSound8KESC-50
#Classes101050
#Task5510
#Total data23,04087322000
Input size423 × 128 × 344 × 60 × 1512 × 128 × 1
Online scenariosOCI/ODIOCI/ODIOCI
Table 2. Average accuracy (%) of the CL methods in the OCI scenario. A higher score is better.
Table 2. Average accuracy (%) of the CL methods in the OCI scenario. A higher score is better.
DatasetDCASE-2021UrbanSound8KESC-50
Offline55.3389.7859.96
Online38.4660.5710.53
Fine-tuning16.3917.283.17
LwF15.4517.674.53
EWC16.1015.914.61
Buffer Size0.2k0.5k1k2k5k0.2k0.5k1k2k5k0.2k0.5k1k
A-GEM15.0115.7615.0114.5013.7218.0218.1517.2517.3516.915.275.613.99
ASER20.0725.4022.6529.3729.3440.2748.5351.7258.2756.0810.1410.9011.35
GDumb27.5933.8638.0343.5847.9447.3861.0969.2678.6187.4225.6141.1654.42
GSS24.1829.7826.7731.9426.8838.9346.5052.0049.5253.1111.299.7812.02
iCaRL36.2836.1835.2836.0236.0550.7752.2155.5652.3352.7316.1321.8023.29
ER24.5031.3426.6129.0832.7439.9148.0053.3145.3148.6310.718.9511.48
MIR26.5928.3130.3326.7028.9633.6250.9849.6653.6654.6210.5711.8310.22
SCR31.8535.4835.6338.2439.6156.7361.4262.4964.5765.6225.9034.7237.70
Table 3. Average forgetting (%) of CL methods in the OCI scenario. A lower score is better.
Table 3. Average forgetting (%) of CL methods in the OCI scenario. A lower score is better.
DatasetDCASE-2021UrbanSound8KESC-50
Fine-tuning60.1471.6227.95
LwF4.357.412.01
EWC62.4465.5435.48
Buffer Size0.2k0.5k1k2k5k0.2k0.5k1k2k5k0.2k0.5k1k
A-GEM60.1962.8660.1963.7764.3269.0770.6963.3570.8567.5040.0340.2339.73
ASER52.8253.5049.9744.2243.1246.6339.0632.3223.8630.0433.3333.3238.06
GDumb22.5823.2820.8718.5419.4819.4615.6213.317.934.0224.2719.2312.32
GSS48.2146.3643.9431.3240.8748.3037.4925.9932.7931.1532.0633.4931.74
iCaRL16.8818.4116.0013.6113.2619.3514.9614.3614.3016.0914.4910.7910.05
ER49.5039.3039.4244.2135.1644.9935.9931.2637.0037.0036.2835.4932.70
MIR40.5635.0029.5933.8833.8651.5727.8327.5422.9020.6132.8431.5632.07
SCR17.6112.0213.9313.6710.8415.6312.1111.499.087.6214.289.149.08
Table 4. Average accuracy (%) of the CL methods in DCASE-2021 in the ODI scenario. A higher score is better.
Table 4. Average accuracy (%) of the CL methods in DCASE-2021 in the ODI scenario. A higher score is better.
DatasetDCASE-2021
ODI scenarioNoiseSpecAugment
Offline46.3550.43
Online27.0828.36
Fine-tuning21.2926.90
LwF22.9026.21
EWC21.1926.63
Buffer Size0.2k0.5k1k2k5k0.2k0.5k1k2k5k
A-GEM12.6217.2815.3615.2213.2815.3015.4514.8615.8017.76
ASER22.3423.5423.4527.1430.2224.4825.8625.6523.4325.72
GDumb14.6815.5215.8816.9936.0014.0915.1715.7917.4140.21
GSS23.0022.5525.4220.6425.3424.6321.9020.1022.2227.12
ER25.0128.5924.6427.2125.0325.4924.7924.7425.8129.99
MIR27.6329.7427.6529.3730.5027.9926.6629.8628.9826.84
SCR27.7730.9933.2632.3934.9728.0031.6630.8534.5734.99
Table 5. Average accuracy (%) of the CL methods in UrbanSound8K in the ODI scenario. A higher score is better.
Table 5. Average accuracy (%) of the CL methods in UrbanSound8K in the ODI scenario. A higher score is better.
DatasetUrbanSound8K
ODI ScenarioNoiseSpecAugment
Offline33.2434.15
Online27.8925.40
Fine-tuning24.5426.95
LwF26.2128.53
EWC27.0524.36
Buffer Size0.2k0.5k1k2k5k0.2k0.5k1k2k5k
A-GEM16.1515.5617.2317.5915.3414.0816.5216.0014.6816.44
ASER25.3427.0027.8228.5427.6725.5624.7127.2429.1623.02
GDumb17.5316.2814.3730.0833.6315.8616.8314.7929.5634.19
GSS24.0522.8221.8927.1124.5623.3423.6424.3728.6129.87
ER22.7127.4727.0826.8228.0725.4226.2127.1027.1030.30
MIR25.6826.9826.1830.6128.5624.8428.5227.2925.7324.97
SCR33.8236.5135.3735.7335.6034.2535.3635.3035.1036.60
Table 6. Average forgetting (%) of CL methods in DCASE-2021 in the ODI scenario. A lower score is better.
Table 6. Average forgetting (%) of CL methods in DCASE-2021 in the ODI scenario. A lower score is better.
DatasetDCASE-2021
Type of ODINoiseSpecAugment
Fine-tuning7.435.35
LwF5.442.46
EWC6.344.00
Buffer Size0.2k0.5k1k2k5k0.2k0.5k1k2k5k
A-GEM3.063.812.964.444.785.174.303.085.683.75
ASER9.285.704.054.793.335.384.395.577.685.95
GDumb1.212.371.503.531.062.701.892.612.511.86
GSS1.282.371.773.033.230.740.991.280.972.27
ER2.281.352.893.434.132.372.945.292.901.65
MIR1.141.161.921.741.942.091.921.372.144.43
SCR1.171.080.530.520.551.451.301.610.340.32
Table 7. Average forgetting (%) of CL methods in UrbanSound8K in the ODI scenario. A lower score is better.
Table 7. Average forgetting (%) of CL methods in UrbanSound8K in the ODI scenario. A lower score is better.
DatasetUrbanSound8K
Type of ODINoiseSpecAugment
Fine-tuning7.605.06
LwF3.793.14
EWC4.445.82
Buffer Size0.2k0.5k1k2k5k0.2k0.5k1k2k5k
A-GEM5.734.552.903.094.215.276.013.693.713.59
ASER5.906.026.645.044.696.875.637.634.478.12
GDumb1.894.315.841.732.034.313.475.172.212.08
GSS2.801.474.224.275.132.201.371.532.842.41
ER5.272.764.564.182.902.665.075.435.432.92
MIR4.033.443.950.883.665.061.253.983.667.05
SCR1.310.550.630.880.802.060.731.670.680.47
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ha, D.; Kim, M.; Jeong, C.Y. Online Continual Learning in Acoustic Scene Classification: An Empirical Study. Sensors 2023, 23, 6893. https://doi.org/10.3390/s23156893

AMA Style

Ha D, Kim M, Jeong CY. Online Continual Learning in Acoustic Scene Classification: An Empirical Study. Sensors. 2023; 23(15):6893. https://doi.org/10.3390/s23156893

Chicago/Turabian Style

Ha, Donghee, Mooseop Kim, and Chi Yoon Jeong. 2023. "Online Continual Learning in Acoustic Scene Classification: An Empirical Study" Sensors 23, no. 15: 6893. https://doi.org/10.3390/s23156893

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop