Next Article in Journal
Canopy-Height and Stand-Age Estimation in Northeast China at Sub-Compartment Level Using Multi-Resource Remote Sensing Data
Next Article in Special Issue
Adaptive Adjacent Layer Feature Fusion for Object Detection in Remote Sensing Images
Previous Article in Journal
DPSDA-Net: Dual-Path Convolutional Neural Network with Strip Dilated Attention Module for Road Extraction from High-Resolution Remote Sensing Images
Previous Article in Special Issue
Multi-Oriented Enhancement Branch and Context-Aware Module for Few-Shot Oriented Object Detection in Remote Sensing Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MFSFNet: Multi-Scale Feature Subtraction Fusion Network for Remote Sensing Image Change Detection

1
Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100094, China
2
Key Laboratory of Technology in Geo-Spatial Information Processing and Application System, Chinese Academy of Sciences, Beijing 100190, China
3
School of Electronic, Electrical and Communication Engineering, University of Chinese Academy of Sciences, Beijing 101408, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(15), 3740; https://doi.org/10.3390/rs15153740
Submission received: 13 June 2023 / Revised: 19 July 2023 / Accepted: 26 July 2023 / Published: 27 July 2023

Abstract

:
Change detection plays a crucial role in remote sensing by identifying surface modifications between two sets of temporal remote sensing images. Recent advancements in deep learning techniques have yielded significant achievements in this field. However, there are still some challenges: (1) Existing change feature fusion methods often introduce redundant information. (2) The complexity of network structures leads to a large number of parameters and difficulties in model training. To overcome these challenges, this paper proposes a Multi-Scale Feature Subtraction Fusion Network (MFSF-Net). It comprises two primary modules: the Multi-scale Feature Subtraction Fusion (MFSF) module and the Feature Deep Supervision (FDS) module. MFSF enhances change features and reduces redundant pseudo-change features. FDS provides additional supervision on different scales of change features in the decoder, improving the training efficiency performance of the network. Additionally, to address the problem of imbalanced samples, the Dice loss strategy is introduced as a means to mitigate this issue. Through comprehensive experiments, MFSF-Net achieves an F1 score of 91.15% and 95.64% on LEVIR-CD and CDD benchmark datasets, respectively, outperforming six state-of-the-art algorithms. Moreover, it attains an improved balance between model complexity and performance, showcasing the efficacy of the proposed approach.

Graphical Abstract

1. Introduction

The process of change detection in remote sensing images involves the detection of modifications that occur on the Earth’s surface between two separate sets of temporal remote sensing images. This task holds significant importance in remote sensing [1]. By enabling effective monitoring of surface changes, the application of remote sensing image change detection spans various fields and domains. such as urban built-up area expansion monitoring [2], natural disaster assessment [3], and environmental monitoring [4].
Based on the analysis unit, conventional approaches to remote sensing image change detection can be classified into two categories: pixel-based and object-based methods. Pixel-based methods operate on a pixel-by-pixel basis, extracting spectral and texture features from the input images pixel by pixel. They subsequently utilize predefined thresholds to identify change areas for each pixel. Pixel-based methods include arithmetic image differencing [5], change vector analysis based on transformations [6,7], principal component analysis [8,9], and independent component analysis [10]. Object-based change detection methods first segment the images into super-pixel objects using image segmentation techniques based on spectral and texture features (segmentation methods such as quadtree-based segmentation [11], multi-resolution segmentation [12], etc.). Then, in order to derive the outcomes of change detection, the segmented results from different time periods are compared in object-based methods. In contrast to the pixel-based approach, object-based methods consider contextual information but are more affected by the results of image segmentation. Additionally, both pixel-based and object-based approaches need significant manual intervention and are prone to pseudo-changes caused by sensor and illumination conditions.
Over the past few years, deep learning methods have received considerable recognition within the remote sensing domain, prompting researchers to integrate these techniques into various tasks related to remote sensing, including scene classification [13,14], semantic segmentation [15,16,17], object detection [18,19], and change detection [20,21,22,23,24], achieving remarkable performance. Deep learning-based approaches for change detection tasks have the ability to incorporate spatial contextual information during pixel-level change identification in the images. CNN (Convolutional neural network) models, with their excellent feature representation capability and end-to-end simplicity, not only reduce manual intervention but also enhance the accuracy and generalization. CNN-based change detection methods transform the two temporal remote sensing images into high-level features, extract semantic context of change regions by fusing the features of the two time-phased images, and mitigate artificial errors stemming from preprocessing. There are two types of CNN-based change detection methods, categorized based on the fusion strategy utilized: image-level fusion [25,26,27,28,29] and feature-level fusion [30,31,32,33,34,35,36,37]. In image-level fusion networks, the two remote sensing images from different temporal input as a whole into CNN to obtain a representation of image differences. However, the absence of deep feature extraction from individual temporal images can result in boundary disturbances within the predicted outcomes, thereby constraining the accuracy of change detection. In contrast to image-level fusion networks, feature-level fusion networks employ two networks with shared parameters. These networks independently learn features from individual temporal images and then combine these features as inputs to the classifier, overcoming the limitations mentioned above.
With the development of satellites and airborne sensors, more detailed and objective representations of the surface can be observed, thus offering finer-grained data for detecting surface changes. However, the diversity of surface features, especially the variability in object shapes, the complexity of background objects, as well as differences in weather conditions, imaging angles, and sensors, can easily lead to false detections or missed detections of actual change areas. Additionally, objects in remote sensing images exhibit different sizes, and a robust and generalizable model should be capable of handling various object scales. Moreover, change detection tasks often encounter a substantial class imbalance problem, where the count of unchanged pixels significantly outweighs the count of changed pixels. This presents a significant challenge in deep learning-based change detection methods for high-resolution remote sensing, as it becomes crucial to effectively extract and utilize the abundant feature information from high-resolution imagery to mitigate the influence of pseudo-changes and enhance the accuracy of detection.
To address this challenge and achieve better representation capabilities of deep features, designing deeper and more complex feature extraction networks has gotten significant attention as a primary research focus. Many researchers have put forward several enhanced models to achieve more discriminative feature representations, such as combining Generative Adversarial Networks (GAN) [38,39,40] or Recurrent Neural Networks (RNN) [41,42], or using feature extraction models based on the Transformer architecture [43,44,45] to expand the receptive field. Some studies focus on the effective utilization of features, such as using spatial or channel attention mechanisms [30,31,32,36,46] or employing multi-scale feature fusion for feature enhancement [25,29,47,48,49]. However, along with the increased complexity of the models, there is a proliferation of parameters and redundant feature information. This not only imposes a heavy burden on model training but also increases the risk of pseudo-changes detections due to the presence of excessive redundant information.
In light of the aforementioned challenges, we present a novel approach called the Multi-Scale Feature Subtraction Fusion Network (MFSFNet). Our network is constructed based on the following four requirements. Firstly, the model should maximize the utilization of feature information derived from the dual-temporal imagery, emphasizing real change areas while reducing the generation of redundant information to minimize false detections caused by complex backgrounds or imaging differences. Secondly, the network must be able to effectively represent features of diverse structures and different sizes of objects. Thirdly, the model should be easy to train, avoiding the issue of gradient vanishing. Lastly, it is essential for the model to effectively tackle the issue of sample imbalance and enhance the accuracy of detection. To address the first two requirements, we extract the dual-temporal imagery multi-scale features and design the Multi-scale Feature Subtraction Fusion (MFSF) module to fuse these features. Unlike existing fusion strategies, our subtraction fusion strategy enhances change features while reducing the interference of redundant features. To meet the third requirement, we use ConvnNext v2-Atto, as the lightweight multi-scale feature encoder with less parameters. We also introduce the Feature Deep Supervision (FDS) module in the decoder to provide additional supervision for deep change features, which improves the model’s feature extraction capability while accelerating convergence. To address the fourth requirement, we incorporate Dice loss into the loss function to mitigate the imbalance between change and non-change pixels. With these efforts, our network can effectively capture change features in high-resolution imagery.
The primary contributions of this work can be summarized as follows:
  • We propose the MFSFNet for high-resolution remote sensing image change detection. This network enhances change features and reduces redundant pseudo-change features through a multi-scale subtraction fusion strategy.
  • We utilize a lightweight feature extraction network and introduce a novel deep supervision strategy in the change decoder, which enhances the training performance of the network.
The paper is structured as follows: Section 1 provides an introduction to the background and problem addressed in this study. Section 2 discusses relevant literature and related works. Section 3 presents the comprehensive details of MFSFNet. Section 4 describes the experimental design, parameter settings, and analysis of the obtained results. Section 5 offers a discussion of our method. Finally, Section 6 concludes the paper, summarizing the key findings and contributions.

2. Related Works

2.1. Encoder in Change Detection Task

To enhance the learning and representing change features, recent studies have concentrated on improving the architecture and capabilities of change detection encoders. One approach is to increase the depth of the encoder by incorporating more convolutional layers. Regarding network structures, Peng et al. introduced the Unet++_MSOF [25], which builds upon the Unet++ architecture [50,51]. Unet++_MSOF employs a technique of combining two individual temporal images, each having C channels, to create a 2C data representation. This combined data representation is then inputted into the network for the purpose of identifying the changed regions. This method allows for the incorporation of temporal information and enhances ability to capture changes between the two images. Zhan et al. [52] introduced the use of Siamese convolutional networks for change detection tasks, which are composed of twin networks that share weights and are trained to identify similarities or differences between two input samples. The concept of end-to-end training for change detection was introduced by Tao et al. [37]. In their work, FC-EF employed a fully connected layer to establish a connection between the two temporal images, serving as input to the network. On the other hand, FC-Siam-conc and FC-Siam-diff employed a Siamese structure, enabling direct processing of the temporal images. This approach facilitated the integration of temporal information and resulted in improved change detection performance. Chen and Shi [53], as well as Chen et al. [31], introduced STANet and DASNet, respectively, using ResNet [54] as the backbone. These architectures took advantage of ResNets powerful feature extraction capabilities to boost the effectiveness of change detection models. To address the inherent localness of convolutional operations, Zhang et al. [55] adopted dilated convolutions instead of traditional convolutions and achieved promising results. Furthermore, dense connections between features have been demonstrated to improve the performance of networks. Thus, Fang et al. [29] and Peng et al. [46] employed dense connections between features at layers, enhancing the network’s capabilities. Additionally, for the task of change detection in multispectral time-series images, 3D CNN models are developed to handle spatial-spectral-temporal features [56]. However, it is important to acknowledge that these methods introduce additional complexity and computational costs to change detection algorithms. Therefore, it becomes crucial to carefully consider and strike a balance between achieving high detection accuracy and managing the computational complexity involved.

2.2. Multi-Scale Feature Fusion in Change Detection Task

The incorporation of multi-scale feature fusion modules has gained significant attention to effectively represent the multi-scale characteristics of geo-objects. One notable approach is PSPNet-CONC [49], which introduced the PSP module [17] for multi-scale feature extraction. The PSP module allows the network to capture contextual information at multiple scales, enabling a more comprehensive representation of the objects in the image. Another approach is utilized by Unet++_MSOF [25] and NestNet [48]. This module allows for the integration of features from various levels of the network, combining both detailed and high-level information to enhance the precision of change detection. To further enhance the detection accuracy, these methods also employ a multi-output fusion strategy. For instance, ADS-Net [47] integrates features from different branches in the decoder and calculates the F1 score for each scale output. Subsequently, the change maps are weighted according to the F1 score to emphasize more reliable detections. It is important to mention that the fusion strategies mentioned above are based on addition, although commonly used, may introduce noise in the fused results and increase the complexity of the model. This can make convergence challenging and potentially lead to false change detections. Therefore, researchers need to carefully balance the complexity and fusion strategies to ensure accurate and reliable results.
Table 1 demonstrates our summary and comparison analysis of encoder and feature fusion strategies for above change detection methods.

3. Methods

This section presents a comprehensive overview of the MFSFNet (Multi-Scale Feature Subtraction Fusion Network) architecture, highlighting its key components and functionality. Firstly, we present the MFSFNets flowchart and overall architecture (refer to Figure 1 and Figure 2). Then, we describe in detail the MSSF module and the Feature Deep Supervision (FDS) module that we have designed. Lastly, we define the loss function.

3.1. Flowchart and Overall Architecture of MFSFNet

As the flowchart of MFSFNet shown in Figure 1, firstly, the inputs of MFSFNet are two temporal remote sensing images. Secondly, the images are fed into a weight-shared Siamese CNN to extract multi-scale features based on ConvNext V2. Thirdly, multi-scale features are fused by multi-scale subtraction fusion module. Fourthly, the fused features are decoded by a multi-scale feature decoder. The multiscale feature decoder has two branches, one that outputs the change detection results of the model, and a feature deep supervision branch that generates an additional change detection result through feature deep supervision. Lastly, both results are used with ground truth to compute the loss and perform loss back propagation. The module of feature deep supervision subjects the model to additional supervision. The black solid arrows in the figure show the sinusoidal propagation path of the data and the red dashed arrows show the backward propagation path of the gradient.
Figure 2 illustrates the overall architecture of MFSFNet. MFSFNet is composed of three components: the Siamese encoder, the Multi-scale Subtraction Feature Fusion (MSSF), and the decoder. The decoder includes the Feature Deep Supervision (FDS) module for multi-scale supervision. Firstly, we employ a ConvNext architecture [57,58] as the feature encoder to obtain multi-level features while enhancing the generalization of feature representations, enabling the model to tackle the challenges posed by diverse land cover structures. Secondly, the MSSF module performs feature fusion on different scales. It reduces feature redundancy while emphasizing different-sized objects and minimizing interference from complex background information. This helps to generate more accurate boundaries. Next, the MSSF module’s output features are bottom-up integrated by the decoder as part of the feature decoding process to produce change features. Subsequently, FDS is applied to predict change detection results on different scales of the change features. Multiple predictions are compared with the label, and losses are computed accordingly. Additionally, we incorporate the Dice Loss to increase the model’s attention to object boundaries and alleviate issues related to irregular or misaligned boundaries. Finally, the model achieves convergence through loss backpropagation, resulting in improved change detection results. Overall, the MFSFNet network leverages the Siamese feature encoder, MSSF module, and decoder module to effectively fuse multi-scale features and obtain accurate results. The incorporation of the FDS module enhances the training process by providing multi-scale supervision. The addition of the Dice Loss promotes better handling of object boundaries. Through loss propagation, the model achieves convergence and produces superior change detection outcomes.

3.2. The Siamese Feature Encoder

The Siamese feature encoder is responsible for extracting multi-scale change features and semantic features that are essential for change detection. To achieve a generalized representation of both semantic and change features, the improved version of the ConvNext network, called ConvNext V2-Atto [58], is employed in MFSFNet. ConvNext V2-Atto is a lightweight variant of ConvNextV2, with a small number of network parameters while maintaining good feature generalization. Compared to ConvNextV1 [57], ConvNext V2 incorporates a fully convolutional masked autoencoder (FCMAE) framework similar to MAE [59] for pre-training, which enhances the feature extraction capability. The feature encoder based on ConvNext V2-Atto is illustrated in Figure 3a.
The basic unit of the ConvNext V2-Atto encoder is the ConvNext V2 Block. As shown in Figure 3b, within each ConvNext V2 Block, the input feature map F 0 of size h × w × c undergoes a 7 × 7 Depthwise convolution [60] and a Layer Norm layer [61], resulting in the feature map F 1 of size h × w × c . Then, a 1 × 1 convolution layer, GeLU activation layer [62], and Global Response Normalization (GRN) layer [58] are applied to get the feature map F 2 of size h × w × 4 c . The channel count is then restored to c using a 1 × 1 convolution layer and Drop Path layers are employed to prevent overfitting to get F 3 . Finally, an element-wise addition operation is performed between F 3 and the input F 0 , resulting in a feature map F of size h × w × c . In the ConvNext V2-Atto encoder, four groups of ConvNext V2 Blocks with a ratio of {2:2:6:2} are combined with down-sampling layers to generate four feature maps at different scales: size of H 4 × W 4 × C , H 8 × W 8 × 2 C , H 16 × W 16 × 4 C , H 32 × W 32 × 8 C , respectively.
For MFSFNet, two shared-weight ConvNext V2-Atto Encoders are used to get multi-scale features for the two images. The images from two time periods, denoted as H × W × C image I1 and I2, are inputted into the twin ConvNext V2-Atto Encoders. This allows for obtaining semantic features for each image at four different scales. Specifically, for I1, the multi-scale semantic features are   F 1 1 of size H 4 × W 4 × C , F 2 1 of size H 8 × W 8 × 2 C , F 3 1 of size H 16   × W 16 × 4 C , and F 4 1 of size H 32 × W 32 × 8 C . Similarly, I2 yields the multi-scale semantic features F 1 2 , F 2 2 , F 3 2 , and F 4 2 , which have the same size as the corresponding features of I1.
To obtain multi-scale change features for the objects, the semantic features from the two images are concatenated along the channel dimension at each scale. This concatenation ensures that the change features have the same width and height as the semantic features, but with the channel count being twice that of the semantic features of a single image. Finally, we obtain change features F 1 C h a n g e ,   F 2 C h a n g e ,   F 3 C h a n g e and   F 4 C h a n g e of sizes H 4 × W 4 × 2 C , H 8 × W 8 × 4 C , H 16 × W 16 × 8 C , and   H 32 × W 32 × 16 C , respectively. These change features serve as the input to the multi-scale subtraction feature fusion module.

3.3. Multi-Scale Feature Subtraction Fusion (MSSF) Module

To enhance the model’s generalization ability for objects at various scales and prioritize features relevant to changes while suppressing irrelevant ones, we introduce the Multi-scale Feature Subtraction Fusion (MFSF) module. This module combines the multi-scale features from the encoder and employs a subtraction fusion method. By subtracting corresponding features from different scales, redundant information during feature fusion is reduced. This subtraction fusion approach enhances the representation of relevant features associated with changes and helps mitigate the interference caused by complex backgrounds. The MFSF module emphasizes discriminative information and improving the accuracy.
To facilitate the fusion of different-scale features, we first employ a 3 × 3 convolutional layer to lessen the dimension of the multi-scale change features F 1 C h a n g e ,   F 2 C h a n g e ,   F 3 C h a n g e , and F 4 C h a n g e to 64, resulting in M S 1 0 , M S 2 0 , M S 3 0 , and M S 4 0 . Then, we apply subtraction units to fuse the features not only within the same scale but also between adjacent scales. The subtraction units enhance the change features and emphasize the change regions at different scales.
The subtraction unit (Figure 4) takes inputs F A and F B . F A is derived from the features M S j i 1 at the same scale, while F B is obtained by up-sampling the features M S j + 1 i 1 from the adjacent scale to match the size of F A . The feature fusion is achieved by Equation (1):
M S j i = S U F A , F B = C o n v a b s M S j i 1 U p s a m p l e M S j + 1 i 1
where j = 1 ,   2 ,   3 ,   4 indices the scale level, and i = 0 , 1 , , 4 j indices the number of feature fusion operations at each scale level. C o n v · denotes a 3 × 3 convolution, a b s · represents the absolute value operation, indicates element-wise subtraction, and U p s a m p l e · denotes the up-sampling layer. M S j i represents feature fusion result, which has the same size as M S j i 1 .
The absolute value operation in the subtraction unit serves a similar purpose as an activation function. When performing fusion operations between features of different scales, the features represented by operation results with the same absolute values indicate the same feature differences. Therefore, the absolute value operation is used for activation. If the commonly used ReLU activation function is employed, the negative values would be discarded directly, overlooking a portion of the feature differences, which would affect the effectiveness of the subtraction fusion.
After the multi-scale feature fusion, at each scale j, the features M S j i are aggregated using element-wise addition to further enhance the focus on the changed features. The computation is performed according to Equation (2):
S F j   = i = 0 4 j M S j i
where j = 1 ,   2 ,   3 ,   4 indices the scale level, M S j i represents the feature fusion result after i iterations of subtraction units at scale j. S F j   denotes the result of adding features at scale j, and its size is the same as M S j i . S F j will serve as the input to the decoder module.

3.4. Decoder and Feature Deep Supervision Module

Due to the varying expressive capabilities of different feature levels in change detection tasks, we employ a FDS module in the multi-scale decoder. By introducing additional supervisory signals at different levels of the network, the network is able to better utilize multi-level feature representations during the learning process.
The decoder of MFSFNet is composed of four stages and a FDS module (Figure 5). Each stage includes a 3 × 3 convolutional, a batch normalization (BN), a ReLU activation, and an up-sampling. Stages 1 to 3 progressively up-sample the channel dimensions of the input feature maps by a factor of 2, while maintaining a consistent channel dimension of 64. This is done to enable element-wise addition with the multi-scale features from MSSF. In Stage 4, the fused feature map of size H 4 × W 4 × 64 is converted to a change detection result P r e 1   of size H × W × 1 , which serves as one of the decoder’s outputs. It is then compared with the ground truth (GT) labels to calculate L o s s 1   .
Additionally, when implementing deep supervision, we took into consideration that the feature map sizes of Stage 3 and Stage 4 are closer to the size of the original image compared to Stage 1 and Stage 2. The larger feature maps indicate the presence of more feature information. Therefore, in addition to supervising the output of Stage 4, supervising Stage 3 may yield better results than supervising Stage 1 and Stage 2. This idea is supported by the results of the ablation experiments in Section 5.1. As a result, we apply deep supervision to the input feature map of Stage 3. In FDS module, Stage 2′s output is fused with S F 2   and passed through a series of a 3 × 3 convolutional, a BN, a ReLU activation, and an up-sampling. This generates a change detection result   P r e 2   of size H × W × 1, which is compared with the labels to calculate L o s s 2   . This process represents the multi-scale feature deep supervision. Finally, after training, the decoder of MFSFNet uses the   P r e 1   result as the model’s change detection output.

3.5. Loss Function of MFSFNet

The loss function of MFSFNet consists of two parts: the difference between the decoder’s final output   P r e 1   and the GT, denoted as   L o s s 1   , and the difference between the output   P r e 2   of the multi-scale deep supervision module and GT, denoted as   L o s s 2   . The calculation of both losses follows the same approach.
In MFSFNet, the proposed approach incorporates the Dice loss [63] along with the binary cross-entropy (BCE) loss to mitigate the issue of class imbalance and alleviate the boundary fuzziness in the results. Therefore, the calculation of   L o s s 1   and   L o s s 2   is described by Equation (3).
L o s s i α , β = α L B C E P r e i , G T + β L d i c e P r e i , G T ,
where   P r e i   ( i = 1 ,   2 ) , represents the binary maps of the two change detection predictions. L o s s i   ( i = 1 ,   2 ) , represents the losses of the two predictions. L B C E denotes the binary cross-entropy loss. L d i c e refers to the Dice loss, calculated as described in Equation (5). The coefficients α and β are used, with the common setting in this paper being α = 0.6 and β = 0.4.
L B C E can be calculated as follows:
L B C E = 1 N i , j = 0 N y i , j l o g   x i , j + 1 y i , j l o g 1   x i , j ,
where y i , j denotes the ground truth label and x i , j denotes the predicted probability of change at a specific point (𝑖,𝑗).
The Dice loss [63] is a metric that can calculated the similarity between two sets from a global viewpoint. It is particularly useful in scenarios where there are only a few positive samples in the image, as it remains effective in such cases. By utilizing the Dice loss, the issue of blurred boundaries caused by the class imbalance in change detection results can be addressed. The calculation of the Dice loss is as follows:
L d i c e = 1 2 T P 2 T P + F P + F N ,
where TP represents the count of correctly predicted positive pixels (true positives), FP represents the count of incorrectly predicted positive pixels (false positives), and FN represents the count of incorrectly predicted negative pixels (false negatives). By considering these values, the Dice loss can effectively evaluate the model in capturing the boundaries and overall similarity of the predicted and true regions in the context of binary classification tasks such as change detection.
The overall loss of MFSFNet is determined by combining L o s s 1 and L o s s 2 through summation, and it can be expressed as follows:
L t o t a l = L o s s 1 0.6 , 0.4 + L o s s 2 0.6 , 0.4 = 0.6 ( L B C E P r e 1 , G T + L B C E P r e 2 , G T ) + 0.4 ( L d i c e P r e 1 , G T + L d i c e P r e 2 , G T ) ,
where, P r e 1 and P r e 2 represent the change detection prediction results.

4. Experiments

4.1. Datasets

To thoroughly evaluate the effectiveness of our method, we conducted experiments on LEVIR-CD and CDD benchmark datasets.
(1)
LEVIR-CD
The LEVIR-CD dataset was introduced by Beihang University. It comprises a large-scale collection of 637 pairs of high-resolution images. The images are of size 1024 × 1024 pixels (0.5 m/pixel) and cover diverse urban and rural areas. This diversity in building types and land cover makes the LEVIR-CD dataset suitable for evaluating the performance of change detection models in complex scenarios. To accommodate memory constraints and sample size, the dataset was divided into non-overlapping patches of size 256 × 256, following the settings in [53]. According to the default dataset partition, the paper used 7120/1024/2048 pairs for training, validation, and testing.
(2)
CDD
The CDD dataset [40] is a popular dataset widely employed for change detection tasks, specifically focusing on seasonal variations in urban landscapes. The dataset comprises real remote sensing images (256 × 256 pixels). One notable characteristic of the CDD dataset is the diverse range of spatial resolutions it offers, spanning from 0.03 to 1 m. Ten thousand pairs were designated for training the model, while 3000 pairs each were allocated for model validation and testing, ensuring an adequate evaluation of the model’s performance on unseen data.

4.2. Evaluation Metrics

Four commonly used accuracy metrics were employed to assess the experimental results. These metrics are precision, recall, F1 score, and IoU (Intersection over Union). For the binary classification problem of change detection, each pixel can belong to one of two classes. The formulas for calculating the four accuracy metrics are as follows.
P r e c i s i o n = T P T P + F P ,
R e c a l l = T P T P + F N ,
F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l ,
I o U = T P F P + T P + F N .
where TP represents the count of correctly predicted positive pixels (true positives), FP represents the count of incorrectly predicted positive pixels (false positives), and FN represents the count of incorrectly predicted negative pixels (false negatives).

4.3. Implementation Details

The PyTorch framework was utilized for implementing the proposed model, and all experiments were performed on a GPU, specifically an RTX 3090. The Adam optimizer was utilized with a weight decay of 1 × 10−4. To update the learning rate, a cosine annealing strategy was employed, with a maximum learning rate of 1 × 10−4 and a minimum learning rate of 1 × 10−6. The learning rate was updated every 20 epochs. During the model training process, several data augmentation techniques were applied to enhance the model’s robustness across different scenes. These techniques included random flipping, color augmentation, and multi-scale training. For the experiments conducted on the LEVIR-CD dataset, the model was trained for a total of 60 epochs. Due to the larger sample size of the CDD dataset, the model was trained for a total of 100 epochs. To initialize the feature extraction backbone of the model, pre-trained parameters from ConvNeXt V2-Atto were used, which were obtained through self-supervised training. This initialization helped in leveraging the learned representations from a large-scale dataset and boosted the model’s performance.

4.4. Comparative Methods

We performed comprehensive experiments comparing it with several SOTA change detection methods. These methods include FC-EF [37], FC-Siam-Diff [37], FC-Siam-Conc [37], STANet [53], BiT [35], and ChangeFormer [64].
  • FC-EF [37], FC-Siam-conc [37], and FC-Siam-diff [37] were the first methods to introduce Siamese networks into change detection. EF refers to a fusion technique in which the dual-temporal images are combined or merged at the input stage. Siam-conc represents the concatenation fusion model based on Siamese networks, which combines the dual-temporal features. Siam-diff represents the difference fusion model based on Siamese networks.
  • STANet [53] incorporates a self-attention feature fusion module, enabling the model to capture the spatiotemporal dependencies present in various sub-regions of the input images. The self-attention mechanism allows the network to concentrate on important regions and relationships within the images, enhancing its ability to detect changes effectively.
  • BiT [35] leverages the Transformer architecture as a change feature fusion network, integrating it with a convolutional neural network (CNN) backbone. This combination enables BiT to capture and model the global semantic information from dual-temporal features. By incorporating Transformers, which excel at capturing long-range dependencies and contextual information, BiT enhances the representation learning process and performance. The CNN backbone complements the Transformer by extracting spatial features from the input images. Together, they form a powerful framework for effectively detecting changes in remote-sensing data.
  • ChangeFormer [64] is a change detection method that utilizes a pure Transformer architecture. Unlike traditional methods that combine CNNs and Transformers, ChangeFormer solely relies on Transformers for the entire change detection process. By leveraging the self-attention mechanism, ChangeFormer efficiently captures multi-scale long-range details. The Transformer architecture allows for the modeling of global dependencies and contextual information across the input images, enhancing the overall performance of change detection tasks.
The implemented CD networks above were developed using their publicly available codes.

4.5. Results Evaluation

4.5.1. Experimental Results on LEVIR-CD

Table 2 presents the performance of different methods on the LEVIR-CD test set based on the four accuracy metrics. It is worth mentioning that for BiT and ChangeFormer, we directly used the pre-trained model weights provided in their respective papers. In Table 1, the first three fully convolutional neural network methods, FC-EF, FC-Siam-Di, and FC-Siam-Conc, perform poorly. This is mainly because their feature extractors struggle to effectively capture features from complex remote-sensing images. The latter three comparative methods, STANet, BiT, and ChangeFormer, all employ different attention mechanisms and achieve good performance on the LEVIR-CD dataset. Overall, MFSFNet demonstrated superior performance compared to all six comparative methods across various evaluation metrics. Specifically, it achieved impressive scores of 92.16% for precision, 90.17% for recall, 91.15% for F1 score, and 83.75% for IoU. Notably, our method showcased the most significant improvement in recall. These results highlight the effectiveness and robustness of our approach in accurately detecting changes, particularly in terms of identifying true positive instances.
Figure 6 visualizes the prediction results of our method and other comparative models on the test set images. As indicated by the accuracy results in Table 2, FC-EF, FC-Siam-Diff, and FC-Siam-Conc performed poorly due to their insufficient feature extraction capabilities. For relatively simple scenes (Figure 6a,e), the three methods STANet, BiT, and ChangeFormer, accurately located the changed buildings, achieving comparable overall performance to our method. However, for relatively complex scenes (Figure 6b–d), our method significantly outperformed all the comparative methods. Specifically, MFSFNet achieved the best results in localizing the changed regions and extracting the boundaries of multi-scale changed regions in complex scenes. The former is attributed to the powerful feature extraction capability of our feature extractor, while the latter relies on the effectiveness of the multi-scale subtraction fusion method.

4.5.2. Experimental Results on CDD

Table 3 shows that MFSFNet surpasses all the comparative models in terms of recall, F1 score, and IoU, with only a slightly lower precision compared to BiT. Particularly noteworthy is the high recall achieved by our method, reaching 95.7%, which demonstrates a substantial superiority over other methods and surpasses the ChangeFormer by 2.18%. From an application perspective, recall is more important than precision in change detection tasks, and our effective feature fusion manner helps to lessen the count of missed change regions.
Additionally, the F1 score of the FC-EF, FC-Siam-Di, and FC-Siam-Conc models is only around 60%, much lower than the other four methods, and this performance gap is even more pronounced on the LEVIR-CD dataset. Analyzing the feature extractors of these three methods, we found that their feature extraction capability is weak, making it challenging to accurately locate change regions. We believe that the limited feature extraction capability is the main reason for the poor performance. Furthermore, the images in the CDD dataset exhibit more diverse scenes, such as seasonal variations, which pose additional challenges for feature extraction.
The prediction results of various methods in Figure 7 show our proposed method produced predictions that are closest to the true change regions in the test images. Examining the predicted result images, it is evident that the first three methods struggle to identify change regions in seasonal variations, while the latter four models are able to adapt to some extent to the effects of seasonal changes. In the case of scenes in Figure 7b,c, our model demonstrates more accurate localization of change regions compared to the other models, and the detected change regions are more complete. In the case of the widened road in scene Figure 7e, our proposed method exhibits the best connectivity of change regions, with smoother edges and the closest resemblance to the ground truth.

5. Discussion

5.1. Ablation Experiments

5.1.1. The Ablation of MFSF and FDS

To conduct a detailed analysis of MFSFNet, ablation experiments were performed on the LEVIR-CD dataset, as summarized in Table 4.
Firstly, we used the lightweight ConvNeXt V2-Atto to get image features. Without using the multi-scale feature subtraction fusion and deep supervision strategy, the F1 score achieved 88.52% for MFSFNet. Next, we successively added the subtraction fusion module and deep supervision at different stages, which resulted in varying degrees of performance improvement. We also conducted a comparative experiment on deep supervision at different stages of the outputs, and the results showed that the model achieved the highest performance of 89.51% when the supervision constraints were applied only in the last two stages. The possible reason for this result is that the feature map sizes of Stage 3 and Stage 4 are closer to the size of the original image compared to Stage 1 and Stage 2. The larger feature maps indicate the presence of more feature information. On the other hand, the feature maps of Stage 1 and Stage 2 contain limited effective information, and directly up-sampling them to the size of the original image leads to poorer performance in change detection results. This, in turn, results in erroneous supervision for the model when all four stages are deep supervised. As a result, the F1 score is lower when all four stages are deep supervised compared to deep supervising only Stage 3 and Stage 4. Finally, we replaced the backbone network with the ConvNeXt V2-Tiny [58], which has a similar scale to ResNet50, and the F1 score further improved to 91.15%, reaching the state-of-the-art performance.

5.1.2. The Effectiveness of MFSF

In this section, a comparison is made between the subtraction fusion strategy and other feature fusion strategies, including product operation, concatenate operation, maximum operation, average operation, and addition operation on the LEVIR-CD dataset. Additionally, we compared the experimental results of using absolute value activation in the subtraction unit with those using ReLU activation to demonstrate the effectiveness of the subtraction fusion strategy in feature fusion.
Table 5 presents the experimental results of change detection on the LEVIR-CD dataset using different feature fusion strategies based on ConvNeXt V2-Atto. The results indicate that our subtraction fusion strategy achieves the highest scores in Precision, F1, and IoU metrics. In terms of the F1, the Product fusion strategy, concatenate fusion strategy, maximum fusion strategy, average fusion strategy, and addition fusion strategy yield similar results, all lower than our subtraction fusion strategy. This is because the change detection task emphasizes the differentiated features between different temporal images, and the subtraction-based feature fusion strategy is more suitable for expressing these differentiated features.
Table 6 presents the experimental results of change detection on the LEVIR-CD dataset using ReLU and absolute value as activation functions in the Subtraction Unit based on ConvNeXt V2-Atto. The results indicate that using absolute value achieves higher scores than ReLU in Precision, Recall, F1, and IoU metrics. This suggests that using absolute value as the activation function provides a better expression of feature differences, while ReLU discards the negative values in the feature differences, thereby affecting the effectiveness of the subtraction fusion.

5.2. Parameter Analysis

To further analyze the complexity of MFSFNet, we calculated the parameter size and computational complexity of different models in Figure 8. The F1 score of different methods on the LEVIR-CD dataset was represented. Although FC-EF, FC-Siam-Di, and FC-Siam-Conc models have the fewest parameters, their accuracy, and performance are the poorest, making them less suitable for practical scenarios. Therefore, we exclude these three models from the following analysis.
For the other models, our proposed model using ConvNeXt V2-Atto has the lowest FLOPs (floating-point operations per second) while still achieving higher accuracy than STANet and BiT. Regarding ChangeFormer, we assume the use of ChangeFormer V6 [64], which has the maximum parameter size and FLOPs. When we employ ConvNeXt V2-Tiny as the feature extractor in our model, the parameter size of MFSFNet is slightly smaller compared to ChangeFormer V6, and the number of floating-point operations (FLOPs) is also lower. Despite this, our proposed method achieves the highest F1 score of 91.15% on the LEVIR-CD dataset in this configuration. This demonstrates the efficiency and effectiveness of our approach in achieving superior performance with smaller model size and computational cost compared to the comparative method.
Overall, our proposed model not only achieves superior accuracy compared to other comparative models but also demonstrates better performance in terms of complexity. Our model achieves SOTA results while also having a smaller model size and lower computational cost. This highlights the efficiency and effectiveness of our approach, making it a promising solution for change detection tasks.

6. Conclusions

In this paper, we propose MFSFNet for change detection in remote sensing images. The key contributions of our approach lie in its twin lightweight encoder-decoder structure and the incorporation of the multi-scale feature subtraction fusion and feature deep supervision modules. The MFSFNet architecture is designed to accurately extract change features. The multi-scale feature subtraction fusion module enhances the representation of change features by emphasizing the differences at various scales while reducing the influence of irrelevant pseudo-change features. This module plays a crucial role in capturing meaningful change regions. To further improve the training performance and enable effective feature learning, we introduce the feature deep supervision module. This module provides additional supervision to the change features at different scales in the decoder, promoting the learning of discriminative representations. By leveraging multi-scale deep supervision, our network effectively captures hierarchical information inherent in the change detection task. With the Dice loss along, our method mitigates the class imbalance problem and reduces boundary fuzziness in the change detection results. The experimental results demonstrate the superior performance of our MFSFNet compared to existing state-of-the-art methods. Our approach achieves a better balance between complexity and performance, indicating its robustness and effectiveness in change detection tasks. In future work, we will focus on making network models and change detection solutions more lightweight, applicable to small sample sizes, and adaptable to weak supervision. This will contribute to the rapid and accurate extraction of surface change information, reducing excessive reliance on change samples and thus improving the applicability of this method in regular surface monitoring.

Author Contributions

Conceptualization, Z.H. and H.Y.; methodology, Z.H.; software, Z.H.; validation, Z.H.; formal analysis, Z.H.; investigation, Z.H.; resources, Z.H.; data curation, Z.H.; writing—original draft preparation, Z.H.; writing—review and editing, Z.H.; visualization, Z.H.; supervision, H.Y.; project administration, H.Y.; funding acquisition, H.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Not applicable.

Conflicts of Interest

All authors declare no conflict of interest.

References

  1. Singh, A. Review article digital change detection techniques using remotely-sensed data. Int. J. Remote Sens. 1989, 10, 989–1003. [Google Scholar] [CrossRef] [Green Version]
  2. Zhong, Q.; Ma, J.; Zhao, B.; Wang, X.; Zong, J.; Xiao, X. Assessing spatial-temporal dynamics of urban expansion, vegetation greenness and photosynthesis in megacity Shanghai, China during 2000–2016. Remote Sens. Environ. 2019, 233, 111374. [Google Scholar] [CrossRef]
  3. De Moel, H.; Jongman, B.; Kreibich, H.; Merz, B.; Penning-Rowsell, E.; Ward, P.J. Flood risk assessments at different spatial scales. Mitig. Adapt. Strat. Glob. Chang. 2015, 20, 865–890. [Google Scholar] [CrossRef] [Green Version]
  4. Demir, B.; Bovolo, F.; Bruzzone, L. Updating Land-Cover Maps by Classification of Image Time Series: A Novel Change-Detection-Driven Transfer Learning Approach. IEEE Trans. Geosci. Remote Sens. 2013, 51, 300–312. [Google Scholar] [CrossRef]
  5. Mondini, A.C.; Guzzetti, F.; Reichenbach, P.; Rossi, M.; Cardinali, M.; Ardizzone, F. Semi-automatic recognition and mapping of rainfall induced shallow landslides using optical satellite images. Remote Sens. Environ. 2011, 115, 1743–1757. [Google Scholar] [CrossRef]
  6. Du, P.; Wang, X.; Chen, D.; Liu, S.; Lin, C.; Meng, Y. An improved change detection approach using tri-temporal logic-verified change vector analysis. ISPRS J. Photogramm. Remote Sens. 2020, 161, 278–293. [Google Scholar] [CrossRef]
  7. Lambin, E.F.; Strahlers, A.H. Change-vector analysis in multitemporal space: A tool to detect and categorize land-cover change processes using high temporal-resolution satellite data. Remote Sens. Environ. 1994, 48, 231–244. [Google Scholar] [CrossRef]
  8. Kesikoglu, M.H.; Atasever, .H.; Özkan, C. Unsupervised change detection in satellite images using fuzzy c-means clustering and principal component analysis. ISPRS Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2013, XL7, 129–132. [Google Scholar] [CrossRef] [Green Version]
  9. Celik, T. Unsupervised Change Detection in Satellite Images Using Principal Component Analysis and k-Means Clustering. IEEE Geosci. Remote Sens. Lett. 2009, 6, 772–776. [Google Scholar] [CrossRef]
  10. Zhong, J.; Wang, R. Multi-temporal remote sensing change detection based on independent component analysis. Int. J. Remote Sens. 2006, 27, 2055–2061. [Google Scholar] [CrossRef]
  11. Gong, M.; Yang, Y.-H. Quadtree-based genetic algorithm and its applications to computer vision. Pattern Recognit. 2004, 37, 1723–1733. [Google Scholar] [CrossRef]
  12. Baraldi, A.; Boschetti, L. Operational Automatic Remote Sensing Image Understanding Systems: Beyond Geographic Object-Based and Object-Oriented Image Analysis (GEOBIA/GEOOIA). Part 1: Introduction. Remote Sens. 2012, 4, 2694–2735. [Google Scholar] [CrossRef] [Green Version]
  13. Roy, S.K.; Krishna, G.; Dubey, S.R.; Chaudhuri, B.B. HybridSN: Exploring 3-D–2-D CNN Feature Hierarchy for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2020, 17, 277–281. [Google Scholar] [CrossRef] [Green Version]
  14. Mou, L.; Ghamisi, P.; Zhu, X.X. Deep Recurrent Neural Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3639–3655. [Google Scholar] [CrossRef] [Green Version]
  15. Kampffmeyer, M.; Salberg, A.-B.; Jenssen, R. Semantic Segmentation of Small Objects and Modeling of Uncertainty in Urban Remote Sensing Images Using Deep Convolutional Neural Networks. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 1–9. [Google Scholar] [CrossRef]
  16. Kemker, R.; Salvaggio, C.; Kanan, C. Algorithms for semantic segmentation of multispectral remote sensing imagery using deep learning. ISPRS J. Photogramm. Remote Sens. 2018, 145, 60–77. [Google Scholar] [CrossRef] [Green Version]
  17. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid scene parsing network. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2881–2890. [Google Scholar]
  18. Shivappriya, S.N.; Priyadarsini, M.J.P.; Stateczny, A.; Puttamadappa, C.; Parameshachari, B.D. Cascade Object Detection and Remote Sensing Object Detection Method Based on Trainable Activation Function. Remote Sens. 2021, 13, 200. [Google Scholar] [CrossRef]
  19. Avola, D.; Cinque, L.; Diko, A.; Fagioli, A.; Foresti, G.L.; Mecca, A.; Pannone, D.; Piciarelli, C. MS-Faster R-CNN: Multi-Stream Backbone for Improved Faster R-CNN Object Detection and Aerial Tracking from UAV Images. Remote Sens. 2021, 13, 1670. [Google Scholar] [CrossRef]
  20. Yang, M.; Jiao, L.; Liu, F.; Hou, B.; Yang, S. Transferred Deep Learning-Based Change Detection in Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6960–6973. [Google Scholar] [CrossRef]
  21. Li, X.; He, M.; Li, H.; Shen, H. A Combined Loss-Based Multiscale Fully Convolutional Network for High-Resolution Remote Sensing Image Change Detection. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  22. Zhang, H.; Gong, M.; Zhang, P.; Su, L.; Shi, J. Feature-Level Change Detection Using Deep Representation and Feature Change Analysis for Multispectral Imagery. IEEE Geosci. Remote Sens. Lett. 2016, 13, 1666–1670. [Google Scholar] [CrossRef]
  23. Lv, Z.; Wang, F.; Cui, G.; Benediktsson, J.A.; Lei, T.; Sun, W. Spatial–Spectral Attention Network Guided with Change Magnitude Image for Land Cover Change Detection Using Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–12. [Google Scholar] [CrossRef]
  24. Chen, T.; Lu, Z.; Yang, Y.; Zhang, Y.; Du, B.; Plaza, A.J. A Siamese Network Based U-Net for Change Detection in High Resolution Remote Sensing Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 2357–2369. [Google Scholar] [CrossRef]
  25. Peng, D.; Zhang, Y.; Guan, H. End-to-End Change Detection for High Resolution Satellite Images Using Improved UNet++. Remote Sens. 2019, 11, 1382. [Google Scholar] [CrossRef] [Green Version]
  26. Du, P.; Liu, S.; Xia, J.; Zhao, Y. Information Fusion Techniques for Change Detection from Multi-Temporal Remote Sensing Images. Inf. Fusion 2013, 14, 19–27. [Google Scholar] [CrossRef]
  27. Alimjan, G.; Jiaermuhamaiti, Y.; Jumahong, H.; Zhu, S.; Nurmamat, P. An Image Change Detection Algorithm Based on Multi-Feature Self-Attention Fusion Mechanism UNet Network. Int. J. Pattern Recogn. Artif. Intell. 2021, 35, 2159049. [Google Scholar] [CrossRef]
  28. Jiang, Y.; Hu, L.; Zhang, Y.; Yang, X. WRICNet: A Weighted Rich-Scale Inception Coder Network for Multi-Resolution Remote Sensing Image Change Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–13. [Google Scholar] [CrossRef]
  29. Fang, S.; Li, K.; Shao, J.; Li, Z. SNUNet-CD: A Densely Connected Siamese Network for Change Detection of VHR Images. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  30. Shi, Q.; Liu, M.; Li, S.; Liu, X.; Wang, F.; Zhang, L. A Deeply Supervised Attention Metric-Based Network and an Open Aerial Image Dataset for Remote Sensing Change Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–16. [Google Scholar] [CrossRef]
  31. Chen, J.; Yuan, Z.; Peng, J.; Chen, L.; Huang, H.; Zhu, J.; Liu, Y.; Li, H. DASNet: Dual Attentive Fully Convolutional Siamese Networks for Change Detection in High-Resolution Satellite Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 1194–1206. [Google Scholar] [CrossRef]
  32. Ding, Q.; Shao, Z.; Huang, X.; Altan, O. DSA-Net: A novel deeply supervised attention-guided network for building change detection in high-resolution remote sensing images. Int. J. Appl. Earth Obs. Geoinf. 2021, 105, 102591. [Google Scholar] [CrossRef]
  33. Yang, L.; Chen, Y.; Song, S.; Li, F.; Huang, G. Deep Siamese Networks Based Change Detection with Remote Sensing Images. Remote Sens. 2021, 13, 3394. [Google Scholar] [CrossRef]
  34. Chen, H.; Wu, C.; Du, B. Towards Deep and Efficient: A Deep Siamese Self-Attention Fully Efficient Convolutional Network for Change Detection in VHR Images. arXiv 2021, arXiv:2108.08157. [Google Scholar]
  35. Chen, H.; Qi, Z.; Shi, Z. Remote Sensing Image Change Detection with Transformers. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–14. [Google Scholar] [CrossRef]
  36. Zhang, C.; Yue, P.; Tapete, D.; Jiang, L.; Shangguan, B.; Huang, L.; Liu, G. A deeply supervised image fusion network for change detection in high resolution bi-temporal remote sensing images. ISPRS J. Photogramm. Remote Sens. 2020, 166, 183–200. [Google Scholar] [CrossRef]
  37. Caye Daudt, R.; Le Saux, B.; Boulch, A. Fully Convolutional Siamese Networks for Change Detection. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; pp. 4063–4067. [Google Scholar]
  38. Zhao, W.; Mou, L.; Chen, J.; Bo, Y.; Emery, W.J. Incorporating Metric Learning and Adversarial Network for Seasonal Invariant Change Detection. IEEE Trans. Geosci. Remote Sens. 2020, 58, 2720–2731. [Google Scholar] [CrossRef]
  39. Hou, B.; Liu, Q.; Wang, H.; Wang, Y. From W-Net to CDGAN: Bitemporal Change Detection via Deep Learning Techniques. IEEE Trans. Geosci. Remote Sens. 2020, 58, 1790–1802. [Google Scholar] [CrossRef] [Green Version]
  40. Lebedev, M.A.; Vizilter, Y.V.; Vygolov, O.V.; Knyaz, V.A.; Rubis, A.Y. Change detection in remote sensing images using conditional adversarial networks. ISPRS Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2018, 422, 565–571. [Google Scholar] [CrossRef] [Green Version]
  41. Khusni, U.; Dewangkoro, H.I.; Arymurthy, A.M. Urban Area Change Detection with Combining CNN and RNN from Sentinel-2 Multispectral Remote Sensing Data. In Proceedings of the 2020 3rd International Conference on Computer and Informatics Engineering (IC2IE), Yogyakarta, Indonesia, 15–16 September 2020; pp. 171–175. [Google Scholar]
  42. Papadomanolaki, M.; Verma, S.; Vakalopoulou, M.; Gupta, S.; Karantzalos, K. Detecting Urban Changes with Recurrent Neural Networks from Multitemporal Sentinel-2 Data. In Proceedings of the IGARSS 2019—2019 IEEE International Geoscience and Remote Sensing Symposium, Yokohama, Japan, 28 July–2 August 2019; pp. 214–217. [Google Scholar]
  43. Mohammadian, A.; Ghaderi, F. SiamixFormer: A fully-transformer Siamese network with temporal Fusion for accurate building detection and change detection in bi-temporal remote sensing images. Int. J. Remote Sens. 2023, 44, 3660–3678. [Google Scholar] [CrossRef]
  44. Shafique, A.; Seydi, S.T.; Alipour-Fard, T.; Cao, G.; Yang, D. SSViT-HCD: A Spatial Spectral Convolutional Vision Transformer for Hyperspectral Change Detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 6487–6504. [Google Scholar] [CrossRef]
  45. Marsocci, V.; Coletta, V.; Ravanelli, R.; Scardapane, S.; Crespi, M. Inferring 3D change detection from bitemporal optical images. ISPRS J. Photogramm. Remote Sens. 2023, 196, 325–339. [Google Scholar] [CrossRef]
  46. Peng, X.; Zhong, R.; Li, Z.; Li, Q. Optical Remote Sensing Image Change Detection Based on Attention Mechanism and Image Difference. IEEE Trans. Geosci. Remote Sens. 2021, 59, 7296–7307. [Google Scholar] [CrossRef]
  47. Wang, D.; Chen, X.; Jiang, M.; Du, S.; Xu, B.; Wang, J. ADS-Net:An Attention-Based deeply supervised network for remote sensing image change detection. Int. J. Appl. Earth Obs. Geoinf. 2021, 101, 102348. [Google Scholar] [CrossRef]
  48. Yu, X.; Fan, J.; Chen, J.; Zhang, P.; Zhou, Y.; Han, L. NestNet: A multiscale convolutional neural network for remote sensing image change detection. Int. J. Remote Sens. 2021, 42, 4898–4921. [Google Scholar] [CrossRef]
  49. Yu, X.; Fan, J.; Zhang, P.; Han, L.; Zhang, D.; Sun, G. Multi-Scale Convolutional Neural Network for Remote Sensing Image Change Detection. In Proceedings of the Geoinformatics in Sustainable Ecosystem and Society; Xie, Y., Li, Y., Yang, J., Xu, J., Deng, Y., Eds.; Springer: Singapore, 2020; pp. 234–242. [Google Scholar]
  50. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. UNet++: A Nested U-Net Architecture for Medical Image Segmentation. In Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Stoyanov, D., Taylor, Z., Carneiro, G., Syeda-Mahmood, T., Martel, A., Maier-Hein, L., Tavares, J.M.R.S., Bradley, A., Papa, J.P., Belagiannis, V., et al., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 3–11. [Google Scholar]
  51. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015; Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F., Eds.; Springer International Publishing: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  52. Zhan, Y.; Fu, K.; Yan, M.; Sun, X.; Wang, H.; Qiu, X. Change Detection Based on Deep Siamese Convolutional Network for Optical Aerial Images. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1845–1849. [Google Scholar] [CrossRef]
  53. Chen, H.; Shi, Z. A Spatial-Temporal Attention-Based Method and a New Dataset for Remote Sensing Image Change Detection. Remote Sens. 2020, 12, 1662. [Google Scholar] [CrossRef]
  54. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  55. Zhang, M.; Xu, G.; Chen, K.; Yan, M.; Sun, X. Triplet-Based Semantic Relation Learning for Aerial Remote Sensing Image Change Detection. IEEE Geosci. Remote Sens. Lett. 2019, 16, 266–270. [Google Scholar] [CrossRef]
  56. Meshkini, K.; Bovolo, F.; Bruzzone, L. A 3D CNN approach for change detection in HR satellite image time series based on a pretrained 2D CNN. ISPRS—Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2022, 43, 143–150. [Google Scholar] [CrossRef]
  57. Liu, Z.; Mao, H.; Wu, C.-Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A ConvNet for the 2020s. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 11966–11976. [Google Scholar] [CrossRef]
  58. Woo, S.; Debnath, S.; Hu, R.; Chen, X.; Liu, Z.; Kweon, I.S.; Xie, S. ConvNeXt V2: Co-Designing and Scaling ConvNets with Masked Autoencoders. arXiv 2023, arXiv:2301.00808. [Google Scholar]
  59. He, K.; Chen, X.; Xie, S.; Li, Y.; Dollar, P.; Girshick, R. Masked Autoencoders Are Scalable Vision Learners. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 16000–16009. [Google Scholar] [CrossRef]
  60. Chollet, F. Xception: Deep Learning with Depthwise Separable Convolutions. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1800–1807. [Google Scholar]
  61. Ba, J.L.; Kiros, J.R.; Hinton, G.E. Layer Normalization. arXiv 2016, arXiv:1607.06450, 2016. [Google Scholar]
  62. Hendrycks, D.; Gimpel, K. Gaussian Error Linear Units (GELUs). arXiv 2020, arXiv:1606.08415. [Google Scholar]
  63. Milletari, F.; Navab, N.; Ahmadi, S.-A. V-net: Fully convolutional neural networks for volumetric medical image segmentation. arXiv 2016, arXiv:1606.04797. [Google Scholar]
  64. Bandara, W.G.C.; Patel, V.M. A Transformer-Based Siamese Network for Change Detection. In Proceedings of the IGARSS 2022–2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 207–210. [Google Scholar]
Figure 1. The Flowchart of MFSFNet.
Figure 1. The Flowchart of MFSFNet.
Remotesensing 15 03740 g001
Figure 2. The overall architecture of MFSFNet. I1 and I2 are two input remote sensing images of different time periods. H is the height of the input image, W is the width of the image, C is the number of channels. M S j i represents feature fusion result, where j = 1 ,   2 ,   3 ,   4 indices the scale level, and   i = 0 , 1 , , 4 j indices the number of feature fusion operations at each scale level. Pre1 denotes the final change detection result. Pre2 denotes the change detection result of the feature deep supervision. GT denotes the ground truth of change detection result. Loss1 and Loss2 denote the loss between the final change detection result and the true value of the model, and the loss between the change detection result and the true value of the deep supervised change detection result, respectively.
Figure 2. The overall architecture of MFSFNet. I1 and I2 are two input remote sensing images of different time periods. H is the height of the input image, W is the width of the image, C is the number of channels. M S j i represents feature fusion result, where j = 1 ,   2 ,   3 ,   4 indices the scale level, and   i = 0 , 1 , , 4 j indices the number of feature fusion operations at each scale level. Pre1 denotes the final change detection result. Pre2 denotes the change detection result of the feature deep supervision. GT denotes the ground truth of change detection result. Loss1 and Loss2 denote the loss between the final change detection result and the true value of the model, and the loss between the change detection result and the true value of the deep supervised change detection result, respectively.
Remotesensing 15 03740 g002
Figure 3. The Details of the Encoder.
Figure 3. The Details of the Encoder.
Remotesensing 15 03740 g003
Figure 4. Subtraction Unit.
Figure 4. Subtraction Unit.
Remotesensing 15 03740 g004
Figure 5. Decoder and Feature Deep Supervision Module. S F j denotes the result of adding features at scale j, where j = 1 ,   2 ,   3 ,   4 indices the scale level. H and W denote the length and width of the feature map. Pre1 denotes the final change detection result. Pre2 denotes the change detection result of the feature deep supervision. GT denotes the ground truth of change detection result. Loss1 and Loss2 denote the loss between the final change detection result and the true value of the model, and the loss between the change detection result and the true value of the deep supervised change detection result, respectively.
Figure 5. Decoder and Feature Deep Supervision Module. S F j denotes the result of adding features at scale j, where j = 1 ,   2 ,   3 ,   4 indices the scale level. H and W denote the length and width of the feature map. Pre1 denotes the final change detection result. Pre2 denotes the change detection result of the feature deep supervision. GT denotes the ground truth of change detection result. Loss1 and Loss2 denote the loss between the final change detection result and the true value of the model, and the loss between the change detection result and the true value of the deep supervised change detection result, respectively.
Remotesensing 15 03740 g005
Figure 6. The visualizations of the prediction results on the LEVIR-CD dataset. The red pixels in the figure represent false positives, the blue pixels represent false negatives, and the white pixels represent true positives. (ae) denote different samples and results in the test set of LEVIR dataset.
Figure 6. The visualizations of the prediction results on the LEVIR-CD dataset. The red pixels in the figure represent false positives, the blue pixels represent false negatives, and the white pixels represent true positives. (ae) denote different samples and results in the test set of LEVIR dataset.
Remotesensing 15 03740 g006
Figure 7. The visualizations of the prediction results on the CDD dataset. The red pixels in the figure represent false positives, the blue pixels represent false negatives, and the white pixels represent true positives. (ae) denote different samples and results in the test set of CDD dataset.
Figure 7. The visualizations of the prediction results on the CDD dataset. The red pixels in the figure represent false positives, the blue pixels represent false negatives, and the white pixels represent true positives. (ae) denote different samples and results in the test set of CDD dataset.
Remotesensing 15 03740 g007
Figure 8. Comparison of parameters and FLOPs of different models.
Figure 8. Comparison of parameters and FLOPs of different models.
Remotesensing 15 03740 g008
Table 1. The literature review summary and comparison analysis.
Table 1. The literature review summary and comparison analysis.
MethodEncoder ArchitectureFeature Fusion StrategyProposed Year
Siamese convolutional networks [52]Twin networks with shared weights.N/A2017
FC-EF, FC-Siam-conc, FC-Siam-diff [37]Fully connected layer skip connecting. Concatenation2018
Unet++_MSOF [25]Multi-scale feature for combining low-level and high-level information.Addition2019
Dilated convolutions [55]Twin networks with shared weights and dilated convolutions used instead of traditional convolutions.Concatenation2019
STANet [53]Twin networks with shared weights and ResNet used as the backbone.Addition2020
PSPNet-CONC [49]Introduces PSP module for multi-scale feature extraction and contextual information capture.Addition2020
DASNet [31]Twin networks with shared weights and ResNet used as the backbone.Addition2021
NestNet [48]Multi-scale feature for combining low-level and high-level information.Addition2021
ADS-Net [47]Multi-scale feature for combining low-level and high-level information.Addition2021
SNUNet-CD [29]Employs dense connections between features at layers.Addition2022
3D CNN [56]A 3D CNN based on a pretrained 2D CNNConcatenation2022
Table 2. The performance of different methods on the LEVIR-CD.
Table 2. The performance of different methods on the LEVIR-CD.
Precision (%)Recall (%)F1 (%)IoU (%)
FC-EF80.2470.3174.9559.93
FC-Siam-Di85.3274.8279.7266.28
FC-Siam-conc83.8281.9882.8970.77
STANet91.9085.0088.1079.12
BIT89.2489.3789.3180.68
ChangeFormer92.0588.8090.4082.84
Ours92.1690.1791.1583.75
Table 3. The performance of different methods on the CDD.
Table 3. The performance of different methods on the CDD.
Precision (%)Recall (%)F1 (%)IoU (%)
FC-EF66.7354.0859.7442.59
FC-Siam-Di81.5151.6863.2546.25
FC-Siam-conc72.6046.5856.7539.62
STANet92.2885.4488.6180.12
BIT96.0293.2694.6189.78
ChangeFormer94.5093.5294.2389.09
Ours95.5995.7095.6491.65
Table 4. Ablation experiments on the LEVIR-CD dataset for the MFSF module and FDS module. (√ denotes the module is employed).
Table 4. Ablation experiments on the LEVIR-CD dataset for the MFSF module and FDS module. (√ denotes the module is employed).
EncoderMFSF ModuleThe Stage of Deep Supervision for FDS Module F1 (%)
AttoTiny1234
88.52
89.25
89.51
89.20
89.31
91.15
Table 5. Comparison of Different Feature Fusion Strategies.
Table 5. Comparison of Different Feature Fusion Strategies.
Fusion StrategyPrecision (%)Recall (%)F1 (%)IoU (%)
Product90.6587.6689.1380.39
Concatenate90.3088.3389.3180.68
Maximum90.5987.4989.0280.21
Average90.4287.6289.0080.18
Addition90.8887.1788.9880.15
Ours91.2688.1789.5181.30
Table 6. Comparison of Different Activation Function in Subtraction Unit.
Table 6. Comparison of Different Activation Function in Subtraction Unit.
Activation FunctionPrecision (%)Recall (%)F1 (%)IoU (%)
ReLU90.8088.1289.4081.16
Absolute Value91.2688.1789.5181.30
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huang, Z.; You, H. MFSFNet: Multi-Scale Feature Subtraction Fusion Network for Remote Sensing Image Change Detection. Remote Sens. 2023, 15, 3740. https://doi.org/10.3390/rs15153740

AMA Style

Huang Z, You H. MFSFNet: Multi-Scale Feature Subtraction Fusion Network for Remote Sensing Image Change Detection. Remote Sensing. 2023; 15(15):3740. https://doi.org/10.3390/rs15153740

Chicago/Turabian Style

Huang, Zhiqi, and Hongjian You. 2023. "MFSFNet: Multi-Scale Feature Subtraction Fusion Network for Remote Sensing Image Change Detection" Remote Sensing 15, no. 15: 3740. https://doi.org/10.3390/rs15153740

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop