Next Article in Journal
Detection of White Leaf Disease in Sugarcane Crops Using UAV-Derived RGB Imagery with Existing Deep Learning Models
Next Article in Special Issue
Two-Branch Convolutional Neural Network with Polarized Full Attention for Hyperspectral Image Classification
Previous Article in Journal
Remote Radio-Physical Harbingers of Drought in Steppes of the South of Western Siberia
Previous Article in Special Issue
Improved Central Attention Network-Based Tensor RX for Hyperspectral Anomaly Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Image Inpainting with Bilateral Convolution

The Institute of Artificial Intelligence and Robotic, Xi’an Jiaotong University, Xian Ning West Road No.28, Xi’an 710049, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(23), 6140; https://doi.org/10.3390/rs14236140
Submission received: 14 October 2022 / Revised: 17 November 2022 / Accepted: 25 November 2022 / Published: 3 December 2022

Abstract

:
Due to sensor malfunctions and poor atmospheric conditions, remote sensing images often miss important information/pixels, which affects downstream tasks, therefore requiring reconstruction. Current image reconstruction methods use deep convolutional neural networks to improve inpainting performances as they have a powerful modeling capability. However, deep convolutional networks learn different features with the same group of convolutional kernels, which restricts their ability to handle diverse image corruptions and often results in color discrepancy and blurriness in the recovered images. To mitigate this problem, in this paper, we propose an operator called Bilateral Convolution (BC) to adaptively preserve and propagate information from known regions to missing data regions. On the basis of vanilla convolution, the BC dynamically propagates more confident features, which weights the input features of a patch according to their spatial location and feature value. Furthermore, to capture different range dependencies, we designed a Multi-range Window Attention (MWA) module, in which the input feature is divided into multiple sizes of non-overlapped patches for several heads, and then these feature patches are processed by the window self-attention. With BC and MWA, we designed a bilateral convolution network for image inpainting. We conducted experiments on remote sensing datasets and several typical image inpainting datasets to verify the effectiveness and generalization of our network. The results show that our network adaptively captures features between known and unknown regions, generates appropriate content for various corrupted images, and has a competitive performance compared with state-of-the-art methods.

1. Introduction

Remote sensing (RS) images often suffer the problem of missing information/pixels caused by sensor malfunctions and poor atmospheric environment conditions [1], such as dead-pixel recovery and cloud obscuration [2,3,4], as shown in Figure 1. Image inpainting is a task that aims to fill in the missing information/pixels with semantically coherent and visually plausible contents [5], which can improve the performance of downstream tasks, such as classification, detection, and segmentation [1,6].
Benefiting from the excellent capacities of Convolutional Neural Networks (CNNs), learning-based works [7] have made significant progress in image inpainting by formulating the problem as a conditional image generation task, viewing pixels of the unbroken region as a condition and predicting pixels of missing regions by maximizing the posterior probability. The static convolutional kernels are not able to deal with diverse corruption regions. Because they process a feature map through the same convolution kernel coefficient on valid and invalid positions, this limits the ability of networks to capture rich contextual cues for the restored image [8,9]. This problem leads to some learning-based inpainting methods suffering from color discrepancy and blurriness in recovered regions.
Figure 1. Example results from different inpainting methods, including PC [8], GC [9], DSN [10], and ours. (a) Results of PatternNet [11] with simulated dead pixels, (b) results of PatternNet with simulated cloud obscuration, (c) results of PatternNet with irregular corrupted regions, (d) results of Paris StreetView [12] with irregular corrupted regions, and (e) results of Places2 [13] with irregular corrupted regions.
Figure 1. Example results from different inpainting methods, including PC [8], GC [9], DSN [10], and ours. (a) Results of PatternNet [11] with simulated dead pixels, (b) results of PatternNet with simulated cloud obscuration, (c) results of PatternNet with irregular corrupted regions, (d) results of Paris StreetView [12] with irregular corrupted regions, and (e) results of Places2 [13] with irregular corrupted regions.
Remotesensing 14 06140 g001
Researchers have reported some attempts to overcome this problem. Partial Convolution (PC) [8] and Gated Convolution (GC) [9] proposed rule-based updating or the learnable feature selection mechanism to treat features differently in corrupted and uncorrupted regions. Furthermore, the recently proposed Dynamic Selection Network (DSN) [10] method selects valid information by deformable convolution and regional normalization for invalid regions. Although these methods can gradually distinguish the invalid regions and use different characterization strategies for valid and invalid regions to help recover images, they are vulnerable to the effect of distance between features located at different locations in the image. This is crucial for the task of inpainting, which relies on known content to infer missing content, since features at a certain point or region in the image tend to be more similar to closer features, as described in [14]. Hence, these methods can suffer from color discrepancy, blurry textures, and distorted/discontinuous structures, especially when the missing regions are large, as shown in Figure 1.
Inspired by the information propagation strategy in exemplar-based methods [5,15], in this paper, we propose a Bilateral Convolution (BC) operator to learn the preservation and propagation of features with a confidence-based estimation for image inpainting. Instead of simply averaging the surrounding features, in BC, patches of features are summed with weights and output the corresponding feature. Specifically, for a feature in a local patch, it is reweighted by the feature distance between it and the feature at the patch center. Moreover, the feature distance is measured by a Gaussian filter [16]. BC expands receptive fields compared with vanilla convolutional neural networks, and dynamically adjusts the results of regular convolution.
Vanilla convolution has a limited capacity for long-range dependency modeling because of the local inductive priors and narrow size of the convolution kernel [17,18,19]. To address this problem, some attention operators have been used in inpainting networks, such as the contextual attention [17] used in GC [9] and DSN [10]. Nevertheless, they often produce artifacts or inconsistent structures in the inpainting results. Furthermore, a further series of studies [20,21,22] found that global attention can be over-concentrated and dispersed; hence, local attention, represented by window attention, may work better in comparison. However, the window attention limits the range of action of the attention, which may create a hindrance for the inpainting model in searching for valid information for broken regions.
We introduce an attention module called Multi-range Window Attention (MWA) to capture different range dependencies and achieve competitive results for inpainting tasks. This approach first divides input features into multiple sizes of non-overlapped patches for several heads. Then, these feature patches are processed by the window self-attention. Our MWA allows each point on the feature to be associated with different features at different ranges, thus enabling the model to synthesize more appropriate contents for broken regions.
With the proposed BC and MWA, we designed a one-stage U-Net [23] network for image inpainting tasks. We conducted extensive experiments on several remote sensing and typical image inpainting datasets. The results demonstrate that our model can adaptively process images with irregular corrupted regions, effectively preserve and propagate information from known regions into unknown regions, and improve the performance of inpainting and downstream tasks.
The main contributions of this paper are summarized as follows:
  • We designed a novel BC operator to adaptively preserve and propagate informative features from known regions to damaged regions with feature distance and surrounding elements. This expands the convolution receptive fields and dynamically adjusts the results of regular convolution.
  • We proposed a MWA module to model multi-range dependencies from different size patches of features, which can effectively associate features at different ranges to synthesize more appropriate contents for damaged regions.
  • The presented bilateral convolution network is based on the proposed BC and MWA with a one-stage image inpainting model, which can reconstruct various irregular corrupted images and improve qualitative and quantitative inpainting performance.
The remainder of this paper is organized as follows: Section 2 presents related works about the missing information reconstruction of remote sensing images and learning-based image inpainting. Section 3 introduces the proposed BC operator and MWA module and designs a one-stage U-shape network for image inpainting. Section 4 presents experimental datasets, settings, and comparison models. In Section 5, we present the qualitative and quantitative results in simulated and real-data experiments and discuss the outcome of ablation studies to validate the effectiveness of our proposed module. Finally, Section 6 summarizes the conclusions and expectations of our approach. Our network architecture is detailed in Appendix A.

2. Related Work

2.1. Missing Information Reconstruction of Remote Sensing Image

Most of the current reconstruction methods for missing information in remote sensing (RS) images require complementary data, such as spectral-based and temporal-based methods, to restore the corrupted images [1]. Spectral-based methods [24,25] use information from intact spectral bands to reconstruct the missing information of the target band in multispectral and hyperspectral images. These methods strongly depend on the corrupted spectral band possessing intact data in other bands, and cannot process missing information in all spectral bands, e.g., missing information caused by cloud obscuration and sensor faults [1]. Temporal-based methods [26,27] utilize the complementary data of multiple periods to reconstruct missing information. However, the time interval is difficult to adjust, and these methods cannot process abrupt variations in the scene. The various complementary data limit these two methods.
Some researchers have proposed non-complementary spatial-based methods [1], also called image inpainting methods in computer vision, which utilize the uncorrupted pixels of images to restore the corrupted pixels without auxiliary spectral or temporal information. Some classical methods include interpolation [28], exemplar-based [29,30], propagated diffusion [31], and variation-based methods [32,33,34]. Nevertheless, these methods are unable to reconstruct complicated or extensively corrupted regions.
Extensive research on neural networks has resulted in some learning-based methods [2,35,36] to reconstruct corrupted RS images. For example, Gao et al. [35] used the deep self-regression network on the reference image to reconstruct the RS image. Wang et al. [36] designed a content and sequence-texture network to restore the missing information of time series images. Zhang et al. [37] adopted a convolutional network for spatial-temporal spectral data to address missing information reconstruction. Lin et al. [38] utilized residual connections and dense blocks with the Charbonnier loss to directly recover corrupted RS images. Shao et al. [2] utilized the coarse-to-fine network with gated convolution and a multiscale discriminator to reconstruct RS images. Singh et al. used CycleGAN [39] and Spatial Attention GAN (SpA-GAN) [40] to restore cloud regions in images. Shao et al. [41] adopted pyramidal GAN to reconstruct missing data in RS images. Czerkawski et al. [42] utilized the inpainting model to recover cloud-free images. Zheng et al. [43] utilized the fully connected network decomposition for RS image inpainting. He et al. [44] used inpainting to assist with the semantic segmentation of remote sensing images. Du et al. [45] proposed a two-stage inpainting model with spatial semantic attention for the recovery of RS images. These learning-based inpainting methods on RS images mainly depend on the CNNs to train the model using a massive dataset, which can achieve sufficient information to predict convincing results in complicated scenes. The details of these methods are introduced in the following subsection.

2.2. Learning-Based Image Inpainting

Learning-based image inpainting methods formulate the inpainting problem as a conditional image generation task [7]. Convolutional neural networks are commonly used to learn weights from massive datasets and predict the missing contents. Pathak et al. [7] first used the Generative Adversarial Network (GAN) to improve the semantic information of inpainting results. Some methods utilized auxiliary information as inpainting guidance and refined details, such as edges [46], structural information [47,48], semantic segmentation maps [49], and foreground object contours [50]; then, they refined the details. However, these methods often suffer from inconsistency in the generated structures and texture details due to the lack of the effective alignment of features in the two stages. Some researchers have utilized parallel networks [51] and recursive networks [52] to achieve powerful results. These methods have more parameters, making the training of models more time-consuming. Some pluralistic inpainting methods, such as VQ-VAE [48], PD-GAN [53], and PUT [54], were proposed to generate multiple completion images. In addition, standard convolution treats all features with the same kernel weights, which is unsuitable for the image completion of various corrupted areas. Some methods use modified convolution in different manners to address the inadaptability of corrupted and uncorrupted regions. For example, Partial Convolution (PC) [8] inferred missing pixels depending on valid pixels and automatically updated the valid region for the next layer; Gated Convolution (GC) [9] learned the feature selection mechanism for each channel and spatial location; and Dynamic Selection Network (DSN) [10] modified the deformable convolution to select valid information. Furthermore, some researchers [55,56] modified the instance normalization operation to adjust valuable features.
Furthermore, fully convolutional networks still lack the capacity to capture long-term relationships in distant contexts. The long-range correlation is vital because it is familiar to the human perception, which is sensitive to semantic relevance [57,58]. Yu et al. [17] proposed contextual attention to borrow features by matching patches from distant locations. Furthermore, other inpainting approaches [9,10,18,19,57,59] also integrated the attention operations into their networks to capture the long-term correlation, improving the inpainting performance.

3. Our Approach

This section demonstrates a bilateral convolution network for image inpainting, the architecture of which is shown in Figure 2a. Our inpainting network adopts a simple yet efficient one-stage U-Net architecture [23], and stacks a series of bilateral convolution blocks and MWA modules into the encoder and decoder. We first design a BC operator to propagate features using the surrounding information, considering distance. Then, we design a residual bilateral convolution block based on our proposed BC operator. Moreover, we introduce an MWA module to build multi-range dependencies for various scales of objects. Finally, we jointly optimize several loss functions to train the image completion network. The details of our network can be found in Appendix A.

3.1. Bilateral Convolution (BC)

3.1.1. Preliminaries

First, we explain why standard convolutions are inadaptable to corrupted and uncorrupted regions for image inpainting tasks. Theoretically, a convolution operation on the input features is defined as follows:
F x , y = W · I x , y ,
where F x , y R C o u t represents the output map of the convolution operation located at ( x , y ) in C o u t channels, I x , y R C i n is the input feature at ( x , y ) , W is the convolution filter weights with the dimensions k h × k w × C i n × C o u t , and k h and k w represent the kernel size. Here, the convolution bias is ignored. The equation shows that the same filter coefficients are applied at any spatial position in the standard convolution. This feature is suitable for tasks in which the input image is not treated differently, such as classification, detection, and segmentation. However, it is not suitable for inpainting tasks, for which the input consists of damaged and undamaged contexts.
Partial Convolution (PC) [8] conditions the completed results only in the valid regions, using the mask-update and re-normalization steps. This can be expressed as
F x , y = β W · I · M , if sum ( M ) > 0 0 , otherwise ,
where M is the corresponding binary mask in the convolution window; 1 represents a valid pixel, and 0 invalid. β is the scaling factor, which equals the size of M divided by the sum of M elements. After the partial convolution operation, the mask is updated with the following rule:
M x , y = 1 , if sum ( M ) > 0 0 , otherwise .
Although PC improved the inpainting results, it still has some problems. The mask-update operation is too rough, being a binary hard update with a sliding window as the calculation unit, which cannot be smoothly propagated as a new mask layer by layer. If one valid pixel in the sliding window exists, this location mask value is updated as valid, as shown in the red edge windows of M o u t in Figure 3a. M has only one channel, and all F channels share the same non-smooth mask. These designs cause some propagation features to be invalid in the hole, as shown in the red box of F in Figure 3a, where some values are close to zero. Moreover, the rule-based update strategy is non-differential and also increases the difficulty of end-to-end learning.
Gated Convolution (GC) [9] improved PC by providing a learnable soft mask for each channel and each spatial location, defined as follows:
M x , y = θ W m · I , F ˜ x , y = ϕ W f · I , F x , y = F ˜ x , y M x , y ,
where W m and W f are two convolutional filters, θ is a sigmoid function that produces soft gating values M x , y R C o u t , ϕ is an activation function, and ⊙ is element-wise multiplication. Yu et al. [9] showed that gating values could learn feature selection values to help generate higher-quality results. However, there is no strategy to obtain better features while the gating value is low. For example, some mask values are close to zero in the corrupted hole region, and due to the corresponding feature values also being invalid, as shown in the red box of F in Figure 3b.

3.1.2. Bilateral Convolution (BC)

We propose a BC operator for image inpainting networks; the architecture is shown in Figure 3c. The BC operator comprises the feature preservation term and propagation term. Both are based on the feature confidence level M. Specifically, the BC operator is similar to GC for feature selection. It uses a point-wise convolution and a 3 × 3 depth-wise convolution with a Sigmoid activation function to compute its corresponding confidence map M R H × W × C , and then uses the confidence map to determine the feature preservation rate. This can be formulated as
M = θ W m · I , F ˜ p = ϕ W f · I , F p = F ˜ p M ,
where I R H × W × C is the input feature of the BC operator, and θ is the sigmoid activation function to ensure that the output confidence values are between 0 and 1. We use this soft gating value M to represent our confidence level, and a higher value in the corresponding position should contain a more helpful feature. ϕ is an Instance Norm (IN) operator. F p is the preservation feature term, which means that a higher confidence level for the values of corresponding location features is better preserved.
The process of BC operator can be given as follows:
F g = 1 M F ˜ , F = F p + F g ,
where F g is the propagation feature term; low confidence values in M represent unfavorable conditions for image inpainting and the requirement to propagate sufficient guidance.
As for the intermediate propagation feature F ˜ , inspired by diffusion-based approaches, we assume that every point in the features can be approximated from spatially close points. We execute convolution on the preservation feature F p with a Gaussian filter. The Gaussian filter measures the feature distance between it and the feature at the patch center. The Gaussian spatial function is defined as follows:
G ( x , y ) = 1 2 π exp x 2 + y 2 2 σ 2 .
With the help of Equation (7), we can obtain the intermediate propagation result F ˜ , reweighted by the feature distance. In our implementation, σ = 1.
These designs of BC expand the receptive field and can dynamically adapt diverse corrupted regions of input. Furthermore, more valuable features can be preserved and propagated. For example, the output feature F shown in Figure 3c has higher, smoother values in the orange box and high-confidence values in the red box.
We adopt the residual block idea to design our bilateral convolution block. Its architecture is shown in Figure 2b. It can be expressed as follows:
x l + 1 = x l + Conv BC ψ Conv ( x l ) ,
where x l and x l + 1 are the input and output of the l-th block, respectively, and ψ represents the Instance Norm (IN) and the GELU activation function. Then, bilateral convolution BC dynamically adjusts the results of regular convolution to preserve features in valid positions and propagate new features with distance information in invalid positions.

3.2. Multi-Range Window Attention (MWA)

Many inpainting models [17,18,19,57,59] have integrated attention modules in order to mitigate the inefficiency of modeling the long-range dependencies of convolution neural networks. Although these models consider the similarity of deep features on the global spatial scale, they fail to build the dependencies between different scales of the feature [60], and the attention usually suffers from being over-concentrated and dispersed [20,61]. We note that some attention mechanisms for dividing windows in transformer architecture [21,60] provide a more efficient and favorable supervisory signal for modeling dependencies. In fact, an image often contains objects of various scales during inpainting tasks. Constructing the multi-range dependencies between different object features is beneficial when recovering the image. Motivated by this phenomenon, we propose the Multi-range Window Attention (MWA) to build dependencies among various feature ranges, as shown in Figure 4.
In the MWA module, we first split features into multiple heads along the channel. Then, we divide the split features of all heads into non-overlapped patches with different window sizes by window and shifted-window strategies. Specifically, when the window size is equal to the head’s feature size, it can obtain global attention. Self-attention only happens within each patch or window. Finally, we can obtain multi-range window attention via joint multi-branch attention.
The first step is splitting feature maps into multiple heads. We use a 1 × 1 convolution to convert the feature F i n R H × W × C to a new feature F R H × W × C , then split the new feature F along the channel into several heads. These operations can be defined as follows:
F = W b · F i n , { F 1 , F 2 , , F B } = split F ,
where F b R H × W × d is the feature on each head, d = C / B is the channel of each head feature, and B is the total number of heads.
We introduce windows with different sizes to divide the feature maps into non-overlapping patches, which can cause each head of the various feature maps to have different range dependencies. For each window size, we deploy it on two head features. On one head feature, we execute normal window-based self-attention. On the other head feature, we execute shifted window-based self-attention, similar to Swin-Transformer [21]. After dividing the windows, each patch feature can be expressed as F b , w R k 1 × k 2 × d , where k 1 × k 2 is the window size, w { 1 , 2 , , m × n } is the sequence number of the patches, and m = H / k 1 , n = W / k 2 . When k 1 = H and k 2 = W , the patch is a global window.
Then, we compute self-attention within each patch, which includes three steps: embedding, matching, and attending [61]. We encode each patch feature map F b , w into a latent embedding space, which can be expressed as
Q b , w = W b , w q · F b , w , K b , w = W b , w k · F b , w , V b , w = W b , w v · F b , w ,
where W b , w q , W b , w k , and W b , w v are three convolutions that map feature maps into the embedding space. Q b , w , K b , w , and V b , w are the query, key, and value features of patches in regular self-attention, respectively. The output of a window’s self-attention is calculated by
Attention ( Q b , w , K b , w , V b , w ) = Softmax Q b , w K b , w T d V b , w ,
where d is a scaling factor based on the dimension of feature maps. The matching operator is calculated by Q b , w K b , w T to obtain the attention map, which represents the correlation between all the tokens within the window. The attending operator performs a weighted summation of value features V b , w to obtain the output.
After determining self-attention in windows, we can obtain scale attention results for each joint window attention on each head. Specifically, we concatenate the attention results of the number of B branches. This jointly requires information from different window sizes and considers various range dependencies:
F a = Concatenate ( WSA 1 , , WSA B ) ,
where F a R H × W × C , C = B d , WSA b R H × W × d is the attention result representing jointed window attention on the b-th head. Next, we pass the concatenated result F a to a 3 × 3 convolution to obtain the final result of multi-range window attention F o u t .
In summary, with the help of multi-range local and global strategies on multiple heads, MWA can effectively borrow relevant features from different ranges, which better models the multi-range dependencies inside multi-range viewpoints.
We present the complexity analysis of our MWA module comparing multi-head self-attention [61]. In multi-head self-attention and our MWA, the same embedding operation is used; only the calculation cost in the matching operation is different. For simplicity, we only discuss the computation cost of the matching operation.
The shape of the input feature maps is H × W × C , where H and W are the height and width of feature maps, and C is the number of channels. The computation complexity of Global Self-Attention (GSA) is below:
O G S A = 2 H W 2 C .
The b-th head contains m × n windows, with a window size of k 1 = H / m and k 2 = W / n . Then, the window self-attention cost of one head can be computed as follows:
O W S A = 2 H W 2 m n C = 2 k 1 k 2 H W C .
From the above computation equations, when k 1 H and k 2 W , the window self-attention of one head is more efficient, and the computation cost declines with m n repetitions.
Our MWA includes one global self-attention head and several window self-attention heads with different window sizes. The computation cost is
O G S A + b = 1 B 1 O W S A b .
The computation cost of multi-head self-attention with B heads is B × O G S A . MWA has 2 H W C × b = 1 B 1 ( H W k b , 1 k b , 2 ) lower complexity than multi-head self-attention [61]. Theoretically, the feature splits into many windows, H W k b , 1 k b , 2 , so our MWA is more efficient than multi-head self-attention.

3.3. Loss Functions

Following some previous works [8,9,46,52], we train the networks with common loss functions to obtain high-quality and semantic levels of completed images, including adversarial loss [62], perceptual loss [63], style loss [62], and reconstruction loss [8]. We define the overall objective of our network as follows:
L o b j I g t , I o u t = α a L a I g t , I o u t + α p L p I g t , I o u t + α s L s I g t , I o u t + α r L r I g t , I o u t ,
where L a is adversarial loss, L p is perceptual loss, L s is style loss, and L r is reconstruction loss. I g t denotes the ground truth image. I o u t denotes the output of our network. α a , α p , α s , and α r are the weights of each loss function.
We use the adversarial loss along with the Patch-GAN [62] to classify the recovery result I o u t and ground truth image I g t . The adversarial loss can guide the model to produce more plausible and local details, which can be expressed as
L a I g t , I o u t = E I g t log D I g t + E I o u t log 1 D I o u t ,
where D is the discriminator network.
We use perceptual loss L p to simulate the human perception of image quality. This method computes the L 1 -norm distance on the high-level semantics features [63] and is expressed as follows:
L p I g t , I o u t = i ϕ r e l u i I o u t ϕ r e l u i I g t 1 ,
where ϕ is the VGG-19 classification model pre-trained on the ImageNet dataset [64], which can obtain the high-level semantics features. ϕ r e l u i is the activation map of the ith specified relu layer of VGG-19. In our training stage, ϕ i corresponds to layers relu1_1, relu2_1, relu3_1, relu4_1, and relu5_1 [63].
Since our network includes window-based attention and up-sampling operations, the checkerboard phenomenon often appears. Therefore, we use style loss to suppress this situation [65], computed by
L s I g t , I o u t = i ϕ r e l u i I o u t ϕ r e l u i T I o u t ϕ r e l u i I g t ϕ r e l u i T I g t 1 ,
where ϕ is the same as the VGG-19 model in perceptual loss, ϕ r e l u i ϕ r e l u i T is a Gram matrix constructed by activation maps ϕ r e l u i , and ϕ r e l u i are activation maps from relu2_2, relu3_4, relu4_4, and relu5_2 in the VGG-19 model [62].
The reconstruction loss is calculated by the L 1 -norm distance between the output I o u t and the ground truth I g t at the pixel level [8], defined as
L r I g t , I o u t = I o u t I g t 1 .

4. Experiments

This section reports the details of our inpainting network’s experimental datasets, settings, and model training.

4.1. Datasets

In image inpainting tasks, three remote sensing image datasets, three typical image inpainting datasets, and one cloud removal dataset were used.
The three RS image datasets were AID [66], NWPU-RESISC45 [67], and PatternNet [11], and their descriptions are as follows:
  • The AID dataset was released by Wuhan University, containing data extracted from different Google Earth sensors [66]. It contains 10,000 RS images with 30 scene categories. We used 500 images for testing, with approximately 17 images in each category.
  • The NWPU-RESISC45 dataset was released by Northwestern Polytechnical University containing data extracted from Google Earth [67]. It includes 31,500 images with 45 scene categories. We selected 500 images for testing, with approximately 11 images in each category.
  • PatternNet is a remote sensing image dataset with 256 × 256 resolution. It mainly obtains 38 scenes with 30,400 images from Google Earth imagery or the Google Map API [11]. We used 400 images for testing, including 10 or 11 images of each scene.
Three typical image inpainting datasets include Paris StreetView [12], CelebA-HQ [68], and Places2 [13], and their introductions are as follows:
  • The Paris StreetView dataset contains images of Paris from Google StreetView. It mainly contains the structural information of windows, doors, and buildings, and has 14,900 training images and 100 test images [12].
  • The CelebA-HQ dataset mainly consists of celebrity faces and includes 30,000 images [68]. We used the first 2000 images as the test set. This dataset is a high-quality version of the CelebA dataset [69].
  • The Places2 dataset was released by MIT. It contains over 1.8 million training images from over 365 scenes [13]. We randomly selected 4000 images from the original validation set for testing.
The cloud removal dataset is the Remote sensing Image Cloud rEmoving dataset (RICE) [70], which consists of RICE1 and RICE2. The RICE2 dataset contains 736 images from the Landsat 8 OLI/TIRS dataset. We randomly selected 630 images for training and 106 images for testing.
Furthermore, in the image completion task, when verifying the algorithm, we need to determine the location of damaged regions. Therefore, we utilized the irregular mask dataset from PC [8] as the test mask dataset. This dataset contains six groups of masks with different ratios of mask regions, and there are 2000 images in each group.

4.2. Experimental Setting

We trained our network on a single RTX3090 GPU with a batch size of eight images. We used the AdamW optimizer with betas ( 0.5 , 0.9 ) . We set an initial learning rate of 10 4 , then decreased it to 10 5 for the fine-tuning of the model. For the weights of each loss function, we set α a = 0.1 , α p = 0.1 , α s = 250 , α r = 1 to train the joint optimization of the loss functions, with reference to previous methods [46,51]. We employed B = 5 attention heads, and the dimension feature map of each branch d was 32. We set the window sizes as 8 and 16 for the local window and shift window attention heads, and set the window size on the global attention head as 64. For AID, NWPU-RESISC45, PatternNet, and Paris StreetView, we trained and fine-tuned our model for 70 epochs. For CelebA-HQ, we trained and fine-tuned our model for 100 and 50 epochs, respectively. As for Places2, we trained our model for four epochs and fine-tuned it for two epochs. The input images of our model were 256 × 256 . Following previous inpainting works [8], we also employ PSNR (Peak Signal-to-Noise Ratio), SSIM (Structural Similarity Index Measure), and FID (Fréchet Inception Distance) to quantitatively compare our results with SOTA methods.

4.3. Comparison Models

We compared our network with several state-of-the-art inpainting methods, including Partial Convolution (PC) [8], Gated Convolution (GC) [9], Recurrent Feature Reasoning Network (RFR) [52], Dynamic Selection Network (DSN) [10], PUT [54], and SpA-GAN [40]. Some of these methods are similar to solving the inadaptability of random corrupted regions of convolution, and some match the good performance of the inpainting model:
  • PC [8]: An encoder–decoder inpainting model that adopts a rule-based updating mask, renormalized only on valid pixels in the U-Net architecture;
  • GC [9]: A two-stage inpainting model representing an improvement on the PC model, having learned a dynamic gating mask. This model can achieve better inpainting results;
  • DSN [10]: A two-parallel U-Net inpainting model used to validate migratable convolution and regional composite normalization in order to select valid information dynamically;
  • RFR [52]: A recurrent encoder–decoder inpainting model adopting the partial convolution of U-Net architecture, which recurrently gathers the hole boundary features and recovers structures to strengthen the results;
  • PUT [54]: A pluralistic inpainting model adopting the patch-based VQVAE without downsampling, using a transformer without quantization to reduce information loss;
  • SpA-GAN [40]: The SpA-GAN model uses the spatial attention generative adversarial network to recover the corrupted or cloud obscuration regions and generate realistic images of remotely sensed scenes.

5. Results

In this section, we compare our results with state-of-the-art methods, both qualitatively and quantitatively, on five datasets. Then, we conduct experiments on actual clouded scenario images to remove the clouds. Next, we analyze the training loss of our network. We also present ablation studies to validate the effectiveness of our proposed BC block and MWA module.

5.1. Quantitative Comparison

We quantitatively compare our network with comparison models in terms of SSIM, PSNR, and FID. Table 1, Table 2 and Table 3 show the results with different corruption ratios on the AID, NWPU-RESISC45, PatternNet, Paris StreetView, CelebA-HQ, and Places2 datasets, respectively. The best results in each group are highlighted in bold. The ↓ indicates that a lower value is better, while ↑ indicates that higher is better. The quantitative results show that our network has comparable performance against other comparison models and produces the best evaluation results on FID and SSIM. Compared with DSN, our method increases the average value of FID (27.0%, 29.9%, 37.4%, 20.2%, 31.8%, and 37.2%), the average value of SSIM (0.9%, 1.4%, 1.8%, 1.65%, 0.4%, and 0.6%), and the average value of PSNR (0.6%, 1.3%, 2.2%, 3.5%, 1.1%, and 0.7%) on AID, NWPU-RESISC45, PatternNet, Paris, CelebA-HQ, and Places2, respectively. Compared with PUT, our method increases the average value of FID (19.7%, 9.7%, 7.2%, 11.5%, 27.4%, 68.4%), the average value of SSIM (1.77%, 1.9%, 0.74%, 1.62%, 1.29%, 3.77%), and the average value of PSNR (2%, 2.2%, 1.5%, 2.5%, 3.8%, 6.0%) on AID, NWPU-RESISC45, PatternNet, Paris, CelebA-HQ, and Places2, respectively. These results show that our method has greatly improved FID metrics and produces better completion results for the surrounding features with distance and multi-range dependencies.

5.2. Qualitative Comparison

We also provide some qualitative comparisons to intuitively analyze the results of different methods. Figure 5 and Figure 6 show the results of the visual comparison of our method and six state-of-the-art methods on remote sensing and natural datasets. For each dataset, the results in the first two rows are the overall completion quality, and the results in the last two rows are the local details displayed in the red box. PC [8], based on the rule-based mask-update method, is unable to obtain valid information to fill the middle part of the hole. Especially for images with large holes, regions are filled with over-saturated and structurally distorted patches. GC [9], based on the learnable feature selection module, can obtain color consistency results but easily generate blurry content. The completed images produced by RFR [52] are relatively good. However, this method still suffers from inconsistent structures, e.g., the blurry bridges of AID and PatternNet, some breaking of the window railings in the third row of Paris StreetView, and a missing human ear in the fourth row of CelebA-HQ. DSN [10] can complete some good images, but it misses some content at the object edge. For example, the car and the human ear lose some content in Paris StreetView and CelebA-HQ, respectively. PUT [54] produces inconsistent contents, incontinuous edges, and abnormal colors, e.g., the windows, railings, and stairs of the buildings in Paris StreetView and Places2. SpA-GAN [40] can restore small damaged regions but cannot process large damaged regions, and some completion results have inconsistent content.
Compared with these methods, our network results in rich texture, consistent structure, and intact objects on both remote sensing and natural images. For example, rich texture as a result of our method is shown in the excellent brick texture of the buildings and the continual window fence of Paris StreetView, the natural hair covering the eye of CelebA-HQ, and the continuous cabinet door strip of Places2. Consistent structure is shown in the buildings of the RS results, the two consistent eyes of CelebA-HQ, and the steps of Places2. Intact objects are shown in the intact bridge of AID and PatternNet, the intact car of Paris StreetView, the intact ear of CelebA-HQ, and the intact red ball of Places2.

5.3. Experiments of Cloud Removal on Real Images

We conducted cloud removal experiments on RICE2 [70] to verify the performance of our network on real data. We trained our bilateral convolution network for 200 epochs and fine-tuned 70 epochs, and compared our cloud removal results with SpA-GAN [40], as shown in Table 4 and Figure 7. The quantitative results show that our network performs better than SpA-GAN on the cloud removal RICE2 dataset. The visual results in Figure 7 show that our model achieves rich texture and consistent structure in bareland and grassland scenes. However, the SpA-GAN results suffer from blurring, ghosting, and artifact problems.

5.4. Loss Analysis

In order to analyze the training process of our network, the training loss curves on four datasets are shown in Figure 8. It can be seen that reconstruction loss, perceptual loss, and style loss decrease with the increase in the number of training epochs. The Patch-GAN generator and discriminator network adversarially learn from each other, which leads the adversarial loss to fluctuate. This situation is in line with the training process of the GAN network. These four losses constitute our objective loss, and the value of the objective loss is shown to be declining in Figure 8, which shows that the network can continue to learn until the convergence state. Furthermore, it can be seen from the objective loss curves that the objective loss stabilizes in the range of the 60th epoch, while it decreases after the 70th epoch. This is because the fine-tuning stage starts in the 70th epoch, which shows that adjusting the learning rate during this stage is helpful to the training process.

5.5. Ablation Studies

To better analyze the impact of the BC block and MWA module in our network, we perform ablation studies on the AID and Paris StreetView datasets. The designs of the experiments are as follows:
  • Two experiments are conducted to analyze the effectiveness of the BC block. The first experiment is denoted as “CONV”, which replaces each layer with standard convolution operators in our designed U-Net style network. The second experiment is denoted as “BC”, which uses the proposed BC block of each layer in our network.
  • Two experiments are conducted to validate the effectiveness of the MWA. The “w/ CA” experiment consists of replacing the MWA in our network with the contextual attention (CA) [17]. The “w/o g” experiment consists of replacing the header of the global attention with a window attention of size 8 in MWA.
All networks mentioned above have the same number of layers and loss functions. The details of our network are shown in Appendix A. The quantitative results of the ablation experiments are displayed in Table 5 and Table 6, and the corresponding qualitative results are shown in Figure 9.

5.5.1. Effectiveness of the BC Block

The proposed BC block is a residual block that includes bilateral convolution. This convolution is composed of the feature preservation term and propagation term. To evaluate the effects of the proposed BC block on inpainting, we replace all standard convolution operators with the BC blocks in our designed U-Net style network. We denote the standard convolution as CONV.
Comparing the quantitative metrics of the CONV and BC block networks in Table 5, the BC block network generates better results than the CONV network. Furthermore, the visual results of the BC block network are better than those of the CONV network, as shown in Figure 9. These results demonstrate that our BC block can preserve better features, and propagate new features with distance information. Therefore, these characteristics are indeed suitable and beneficial for image inpainting.

5.5.2. Effectiveness of the MWA

We compare our MWA with CA and without the global header (w/o g). The experimental results are shown in Table 6. Compared with CA, we achieved higher scores when using the MWA module. This shows that our multi-range window strategy can capture more features that are suitable for completion than the global contextual attention. Compared with the “w/o g” model, this shows that global dependency is still needed in the image restoration task.
Comparing the BC column of Table 5 with the column representing our method in Table 6, we find that our method performs better than using only the BC block. Furthermore, our model can generate more real textures than using only the BC block, as exemplified by our results in Figure 9 for the AID dataset (the dead-line recovery result (row 1), the circle stadium (row 2), the cars in the parking (row 3), the continuity river (row 4)) and in Figure 9 for the Paris StreetView dataset (the bridge hole behind the tree (row 1), the continuous window fence and edge (rows 2 and 3), and the building’s edge (row 4)). These visual results show that our proposed MWA attention can capture multiple ranges of dependencies among pixels and improve the integrity and consistency of the inpainted contents.

6. Conclusions

In this paper, we customize a bilateral convolution network to reconstruct corrupted images, which stacks a series of BC blocks and MWA modules. With the BC operator, the BC block could simultaneously consider the reliability of features based on the spatial location and the feature value, which could be used to adaptively handle diverse corrupted regions and efficiently preserve and propagate information from known regions to unknown regions. Furthermore, the MWA with multi-range window self-attention can model different ranges of spatial dependencies among pixels. The results of quantitative and qualitative experiments on typical remote sensing datasets and image inpainting datasets demonstrate that our network is robust to various scene images and generates more appropriate and consistent content than several state-of-the-art methods. However, our BC operator has the limitation of using a Gaussian filter with a static kernel, which may not work well on large damaged regions. In future studies, we will consider using deformable convolution to expand the receptive field of the Gaussian kernel and convolution kernel, which may help to generate satisfactory completion results on various complex damaged images.

Author Contributions

Conceptualization, W.H. and Y.D.; Data curation, W.H.; Formal analysis, S.H.; Funding acquisition, J.W.; Investigation, W.H.; Methodology, W.H. and Y.D.; Project administration, W.H.; Resources, W.H.; Software, W.H.; Supervision, W.H.; Validation, Y.D. and S.H.; Visualization, W.H.; Writing—original draft, W.H.; Writing—review and editing, J.W., Y.D. and S.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key Research and Development Program of China under Grant No. 2017YFA0700800.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to acknowledge the reviewers and editors for their careful work in improving the quality and presentation of this paper.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
RSRemote Sensing
BCBilateral Convolution
MWAMulti-range Window Attention
CNNsConvolutional Neural Networks
PCPartial Convolution
GCGated Convolution
DSNDynamic Selection Network
RFRRecurrent Feature Reasoning Network
CEContext Encoders
PGNProgressive Generative Networks
INInstance Normalization
GANGenerative Adversarial Network
PSNR Peak Signal-to-Noise Ratio
SSIMStructural Similarity Index Measure
FIDFréchet Inception Distance

Appendix A. Network Architecture

This appendix presents details of our network architectures.

Appendix A.1. Bilateral Convolution Network

We design an encoder–decoder network that adopts the proposed Bilateral Convolution (BC) blocks and Multi-range Window Attention (MWA). Our network includes 23 BC blocks, 7 MWA modules, and 2 vanilla convolution layers. The encoder has four stages, including 16 BC blocks and 7 MWA modules. At the end of each stage, the encoder uses a convolution with a stride of 2 to downsample the feature twice. The input image I i n is R H × W × 3 . The encoder first uses a 7 × 7 convolution layer to raise the channel of the I i n to C. Then, the feature passes through four stages, and the coding feature of each stage E i R H / 2 i 1 × W / 2 i 1 × C 2 i 1 , i = 1 , 2 , 3 , 4 is obtained.
The decoder has seven BC blocks and one convolution layer. It upsamples the features of the last encoder stage feature E 4 , then concatenates E 3 and the upsampled feature that favors U-Net [23]. Then, a 1 × 1 convolution operator is used to fuse features, and features are put into three BC blocks to obtain this stage decoder feature D 3 . Then, the output feature D 1 R H × W × C of the decoder is obtained via two stages. Finally, a 7 × 7 convolution and a tanh activation function are used to obtain the completed RGB image I o u t . The details of the network are shown in Table A1.
Table A1. Details of the proposed inpainting network. “Conv-k-sy” denotes convolution with kernel size k and stride y. “BCB”, "MWA", and "Tanh" represent the BC block, the MWA module, and the tanh activation function, respectively. "×" denotes multiplication. "* N" denotes stacking N modules.
Table A1. Details of the proposed inpainting network. “Conv-k-sy” denotes convolution with kernel size k and stride y. “BCB”, "MWA", and "Tanh" represent the BC block, the MWA module, and the tanh activation function, respectively. "×" denotes multiplication. "* N" denotes stacking N modules.
Stage NameInput Feature SizeOperators
Input256 × 256 × 3-
Stem256 × 256 × 48Conv-7-s1
Encoder1256 × 256 × 48BCB * 1
Down1256 × 256 × 48Conv-3-s2
Encoder2128 × 128 × 96BCB * 3
Down2128 × 128 × 96Conv-3-s2
Encoder364 × 64 × 192BCB * 4
Down364 × 64 × 192Conv-3-s2
Encoder432 × 32 × 384[BCB, MWA] * 7, BCB × 1
Up332 × 32 × 384Upsample 2 times, Conv-3-s1
Fuse364 × 64 × 192Concat(Up3, Encoder3), Conv-1-s1
Decoder364 × 64 × 192BCB * 3
Up264 × 64 × 192Upsample 2 times, Conv-3-s1
Fuse2128 × 128 × 96Concat(Up2, Encoder 2), Conv-1-s1
Decoder2128 × 128 × 96BCB * 2
Up1128 × 128 × 96Upsample 2 times, Conv-3-s1
Fuse1256 × 256 × 48Concat(Up1, Encoder 1), Conv-1-s1
Decoder1256 × 256 × 48BCB * 1
Refine256 × 256 × 48Conv-7-s1
Output256 × 256 × 3Tanh

Appendix A.2. Discriminator

Our network is built upon the framework of a generative adversarial network. We adopt a PatchGAN network as our discriminator [71]. Specifically, the PatchGAN learns to classify each output as real or fake, while the inpainting network tries to fool the PatchGAN. This adversarial training allows our model to generate better content for large corrupted regions. Conv2D uses Spectral Normalization (SN) [72] and the LeakyReLU(0.2) activation function to stabilize GAN training in the PatchGAN. The input of the discriminator is 256 × 256 × 3 , and the final convolution layer produces scores to predict whether image patches are real or fake. We provide the details of the architectures of the PatchGAN in Table A2.
Table A2. Details of the PatchGAN discriminator. “SN-Conv2d” denotes a 2D convolution operator with spectral normalization. "×" denotes multiplication.
Table A2. Details of the PatchGAN discriminator. “SN-Conv2d” denotes a 2D convolution operator with spectral normalization. "×" denotes multiplication.
Module NameFilter SizeOutput ChannelsStrideOutput Feature Size
SN-Conv2d4 × 4642128 × 128
SN-Conv2d4 × 4128264 × 64
SN-Conv2d4 × 4256232 × 32
SN-Conv2d4 × 4512132 × 32
SN-Conv2d4 × 41132 × 32

References

  1. Shen, H.; Li, X.; Cheng, Q.; Zeng, C.; Yang, G.; Li, H.; Zhang, L. Missing information reconstruction of remote sensing data: A technical review. IEEE Geosci. Remote Sens. Mag. 2015, 3, 61–85. [Google Scholar] [CrossRef]
  2. Shao, M.; Wang, C.; Wu, T.; Meng, D.; Luo, J. Context-based multiscale unified network for missing data reconstruction in remote sensing images. IEEE Geosci. Remote Sens. Lett. 2020, 19, 1–5. [Google Scholar] [CrossRef]
  3. Ng, M.K.P.; Yuan, Q.; Yan, L.; Sun, J. An adaptive weighted tensor completion method for the recovery of remote sensing images with missing data. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3367–3381. [Google Scholar] [CrossRef]
  4. Zhao, H.; Duan, S.; Liu, J.; Sun, L.; Reymondin, L. Evaluation of five deep learning models for crop type mapping using sentinel-2 time series images with missing information. Remote Sens. 2021, 13, 2790. [Google Scholar] [CrossRef]
  5. Bertalmio, M.; Sapiro, G.; Caselles, V.; Ballester, C. Image inpainting. In Proceedings of the 27th Annual Conference on Computer Graphics and Interactive Techniques, New Orleans, LA, USA, 23–28 July 2000; pp. 417–424. [Google Scholar]
  6. Gao, Y.; Sun, X.; Liu, C. A General Self-Supervised Framework for Remote Sensing Image Classification. Remote Sens. 2022, 14, 4824. [Google Scholar] [CrossRef]
  7. Pathak, D.; Krahenbuhl, P.; Donahue, J.; Darrell, T.; Efros, A.A. Context Encoders: Feature Learning by Inpainting. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  8. Liu, G.; Reda, F.A.; Shih, K.J.; Wang, T.C.; Tao, A.; Catanzaro, B. Image Inpainting for Irregular Holes Using Partial Convolutions. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 85–100. [Google Scholar]
  9. Yu, J.; Lin, Z.; Yang, J.; Shen, X.; Lu, X.; Huang, T.S. Free-form image inpainting with gated convolution. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 4471–4480. [Google Scholar]
  10. Wang, N.; Zhang, Y.; Zhang, L. Dynamic selection network for image inpainting. IEEE Trans. Image Process. 2021, 30, 1784–1798. [Google Scholar] [CrossRef] [PubMed]
  11. Zhou, W.; Newsam, S.; Li, C.; Shao, Z. PatternNet: A benchmark dataset for performance evaluation of remote sensing image retrieval. ISPRS J. Photogramm. Remote Sens. 2018, 145, 197–209. [Google Scholar] [CrossRef] [Green Version]
  12. Doersch, C.; Singh, S.; Gupta, A.; Sivic, J.; Efros, A. What makes paris look like paris? ACM Trans. Graph. 2012, 31, hal-01053876. [Google Scholar] [CrossRef]
  13. Zhou, B.; Lapedriza, A.; Khosla, A.; Oliva, A.; Torralba, A. Places: A 10 Million Image Database for Scene Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 1452–1464. [Google Scholar] [CrossRef] [Green Version]
  14. Liu, H.; Jiang, B.; Song, Y.; Huang, W.; Yang, C. Rethinking image inpainting via a mutual encoder-decoder with feature equalizations. In Computer Vision–ECCV 2020, Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 2–28 August 2020; Springer: Cham, Switzerland, 2020; pp. 725–741. [Google Scholar]
  15. Ballester, C.; Bertalmio, M.; Caselles, V.; Sapiro, G.; Verdera, J. Filling-in by joint interpolation of vector fields and gray levels. IEEE Trans. Image Process. 2001, 10, 1200–1211. [Google Scholar] [CrossRef] [PubMed]
  16. Tomasi, C.; Manduchi, R. Bilateral filtering for gray and color images. In Proceedings of the Sixth International Conference on Computer Vision, Bombay, India, 7 January 1998; pp. 839–846. [Google Scholar]
  17. Yu, J.; Lin, Z.; Yang, J.; Shen, X.; Lu, X.; Huang, T.S. Generative Image Inpainting with Contextual Attention. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  18. Zeng, Y.; Fu, J.; Chao, H.; Guo, B. Learning Pyramid-Context Encoder Network for High-Quality Image Inpainting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  19. Liu, H.; Jiang, B.; Xiao, Y.; Yang, C. Coherent Semantic Attention for Image Inpainting. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  20. Yang, R.; Ma, H.; Wu, J.; Tang, Y.; Xiao, X.; Zheng, M.; Li, X. ScalableViT: Rethinking the Context-oriented Generalization of Vision Transformer. arXiv 2022, arXiv:2203.10790. [Google Scholar]
  21. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 10012–10022. [Google Scholar]
  22. Vaswani, A.; Ramachandran, P.; Srinivas, A.; Parmar, N.; Hechtman, B.; Shlens, J. Scaling local self-attention for parameter efficient visual backbones. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 12894–12904. [Google Scholar]
  23. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  24. Li, X.; Shen, H.; Zhang, L.; Zhang, H.; Yuan, Q. Dead pixel completion of aqua MODIS band 6 using a robust M-estimator multiregression. IEEE Geosci. Remote Sens. Lett. 2013, 11, 768–772. [Google Scholar]
  25. Wang, Q.; Wang, L.; Li, Z.; Tong, X.; Atkinson, P.M. Spatial–spectral radial basis function-based interpolation for Landsat ETM+ SLC-off image gap filling. IEEE Trans. Geosci. Remote Sens. 2020, 59, 7901–7917. [Google Scholar] [CrossRef]
  26. Zeng, C.; Shen, H.; Zhang, L. Recovering missing pixels for Landsat ETM+ SLC-off imagery using multi-temporal regression analysis and a regularization method. Remote Sens. Environ. 2013, 131, 182–194. [Google Scholar] [CrossRef]
  27. Li, X.; Shen, H.; Zhang, L.; Zhang, H.; Yuan, Q.; Yang, G. Recovering quantitative remote sensing products contaminated by thick clouds and shadows using multitemporal dictionary learning. IEEE Trans. Geosci. Remote Sens. 2014, 52, 7086–7098. [Google Scholar]
  28. Siu, W.C.; Hung, K.W. Review of image interpolation and super-resolution. In Proceedings of the Asia Pacific Signal and Information Processing Association Annual Summit and Conference, Hollywood, CA, USA, 3–6 December 2012; pp. 1–10. [Google Scholar]
  29. Criminisi, A.; Perez, P.; Toyama, K. Region Filling and Object Removal by Exemplar-Based Image Inpainting. IEEE Trans. Image Process. 2004, 13, 1200–1212. [Google Scholar] [CrossRef]
  30. Barnes, C.; Shechtman, E.; Finkelstein, A.; Goldman, D.B. PatchMatch: A randomized correspondence algorithm for structural image editing. ACM Trans. Graph. 2009, 28, 24–35. [Google Scholar] [CrossRef]
  31. Chan, T.F.; Shen, J. Nontexture inpainting by curvature-driven diffusions. J. Vis. Commun. Image Represent. 2001, 12, 436–449. [Google Scholar] [CrossRef]
  32. Shen, H.; Zhang, L. A MAP-based algorithm for destriping and inpainting of remotely sensed images. IEEE Trans. Geosci. Remote Sens. 2008, 47, 1492–1502. [Google Scholar] [CrossRef]
  33. Bugeau, A.; Bertalmío, M.; Caselles, V.; Sapiro, G. A comprehensive framework for image inpainting. IEEE Trans. Image Process. 2010, 19, 2634–2645. [Google Scholar] [CrossRef] [PubMed]
  34. Cheng, Q.; Shen, H.; Zhang, L.; Li, P. Inpainting for remotely sensed images with a multichannel nonlocal total variation model. IEEE Trans. Geosci. Remote Sens. 2013, 52, 175–187. [Google Scholar] [CrossRef]
  35. Gao, J.; Yuan, Q.; Li, J.; Su, X. Unsupervised missing information reconstruction for single remote sensing image with Deep Code Regression. Int. J. Appl. Earth Obs. Geoinf. 2021, 105, 102599. [Google Scholar] [CrossRef]
  36. Wang, Y.; Zhou, X.; Ao, Z.; Xiao, K.; Yan, C.; Xin, Q. Gap-Filling and Missing Information Recovery for Time Series of MODIS Data Using Deep Learning-Based Methods. Remote Sens. 2022, 14, 4692. [Google Scholar] [CrossRef]
  37. Zhang, Q.; Yuan, Q.; Zeng, C.; Li, X.; Wei, Y. Missing data reconstruction in remote sensing image with a unified spatial–temporal–spectral deep convolutional neural network. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4274–4288. [Google Scholar] [CrossRef] [Green Version]
  38. Lin, D.; Xu, G.; Wang, Y.; Sun, X.; Fu, K. Dense-Add Net: An novel convolutional neural network for remote sensing image inpainting. In Proceedings of the IGARSS 2018—2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 4985–4988. [Google Scholar]
  39. Singh, P.; Komodakis, N. Cloud-gan: Cloud removal for sentinel-2 imagery using a cyclic consistent generative adversarial networks. In Proceedings of the IGARSS 2018—2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 1772–1775. [Google Scholar]
  40. Pan, H. Cloud removal for remote sensing imagery via spatial attention generative adversarial network. arXiv 2020, arXiv:2009.13015. [Google Scholar]
  41. Shao, M.; Wang, C.; Zuo, W.; Meng, D. Efficient Pyramidal GAN for Versatile Missing Data Reconstruction in Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–14. [Google Scholar] [CrossRef]
  42. Czerkawski, M.; Upadhyay, P.; Davison, C.; Werkmeister, A.; Cardona, J.; Atkinson, R.; Michie, C.; Andonovic, I.; Macdonald, M.; Tachtatzis, C. Deep internal learning for inpainting of cloud-affected regions in satellite imagery. Remote Sens. 2022, 14, 1342. [Google Scholar] [CrossRef]
  43. Zheng, W.J.; Zhao, X.L.; Zheng, Y.B.; Pang, Z.F. Nonlocal Patch-Based Fully Connected Tensor Network Decomposition for Multispectral Image Inpainting. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  44. He, S.; Li, Q.; Liu, Y.; Wang, W. Semantic Segmentation of Remote Sensing Images With Self-Supervised Semantic-Aware Inpainting. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  45. Du, Y.; He, J.; Huang, Q.; Sheng, Q.; Tian, G. A Coarse-to-Fine Deep Generative Model With Spatial Semantic Attention for High-Resolution Remote Sensing Image Inpainting. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–13. [Google Scholar] [CrossRef]
  46. Nazeri, K.; Ng, E.; Joseph, T.; Qureshi, F.; Ebrahimi, M. Edgeconnect: Structure guided image inpainting using edge prediction. In Proceedings of the Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Seoul, Republic of Korea, 27–28 October 2019. [Google Scholar]
  47. Ren, Y.; Yu, X.; Zhang, R.; Li, T.H.; Li, G. StructureFlow: Image Inpainting via Structure-aware Appearance Flow. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  48. Peng, J.; Liu, D.; Xu, S.; Li, H. Generating Diverse Structure for Image Inpainting With Hierarchical VQ-VAE. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  49. Song, Y.; Chao, Y.; Shen, Y.; Peng, W.; Kuo, C. Spg-net: Segmentation prediction and guidance network for image inpainting. arXiv 2018, arXiv:1805.03356. [Google Scholar]
  50. Xiong, W.; Yu, J.; Lin, Z.; Yang, J.; Lu, X.; Barnes, C.; Luo, J. Foreground-Aware Image Inpainting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  51. Guo, X.; Yang, H.; Huang, D. Image Inpainting via Conditional Texture and Structure Dual Generation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 14134–14143. [Google Scholar]
  52. Li, J.; Wang, N.; Zhang, L.; Du, B.; Tao, D. Recurrent Feature Reasoning for Image Inpainting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  53. Liu, H.; Wan, Z.; Huang, W.; Song, Y.; Han, X.; Liao, J. PD-GAN: Probabilistic Diverse GAN for Image Inpainting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 9371–9381. [Google Scholar]
  54. Liu, Q.; Tan, Z.; Chen, D.; Chu, Q.; Dai, X.; Chen, Y.; Liu, M.; Yuan, L.; Yu, N. Reduce Information Loss in Transformers for Pluralistic Image Inpainting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 11347–11357. [Google Scholar]
  55. Yu, T.; Guo, Z.; Jin, X.; Wu, S.; Chen, Z.; Li, W.; Zhang, Z.; Liu, S. Region normalization for image inpainting. In Proceedings of the the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12733–12740. [Google Scholar]
  56. Wang, Y.; Chen, Y.C.; Tao, X.; Jia, J. Vcnet: A robust approach to blind image inpainting. In Computer Vision—ECCV 2020, Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Cham, Switzerland, 2020; pp. 752–768. [Google Scholar]
  57. Wang, Y.; Chen, Y.C.; Zhang, X.; Sun, J.; Jia, J. Attentive normalization for conditional image generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 5094–5103. [Google Scholar]
  58. Wang, N.; Ma, S.; Li, J.; Zhang, Y.; Zhang, L. Multistage attention network for image inpainting. Pattern Recognit. 2020, 106, 107448. [Google Scholar] [CrossRef]
  59. Xie, C.; Liu, S.; Li, C.; Cheng, M.M.; Ding, E. Image Inpainting with Learnable Bidirectional Attention Maps. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  60. Wang, W.; Yao, L.; Chen, L.; Cai, D.; He, X.; Liu, W. Crossformer: A versatile vision transformer based on cross-scale attention. arXiv 2021, arXiv:2108.00154. [Google Scholar]
  61. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  62. Isola, P.; Zhu, J.Y.; Zhou, T.; Efros, A.A. Image-to-Image Translation with Conditional Adversarial Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  63. Johnson, J.; Alahi, A.; Fei-Fei, L. Perceptual Losses for Real-Time Style Transfer and Super-Resolution. In Computer Vision—ECCV 2016, Proceedings of the European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 11–14 October 2016; Springer: Cham, Switzerland, 2016; pp. 694–711. [Google Scholar]
  64. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  65. Sajjadi, M.S.; Scholkopf, B.; Hirsch, M. Enhancenet: Single image super-resolution through automated texture synthesis. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 4491–4500. [Google Scholar]
  66. Xia, G.S.; Hu, J.; Hu, F.; Shi, B.; Bai, X.; Zhong, Y.; Zhang, L.; Lu, X. AID: A benchmark data set for performance evaluation of aerial scene classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3965–3981. [Google Scholar] [CrossRef] [Green Version]
  67. Cheng, G.; Han, J.; Lu, X. Remote sensing image scene classification: Benchmark and state of the art. Proc. IEEE 2017, 105, 1865–1883. [Google Scholar] [CrossRef]
  68. Karras, T.; Aila, T.; Laine, S.; Lehtinen, J. Progressive Growing of GANs for Improved Quality, Stability, and Variation. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  69. Liu, Z.; Luo, P.; Wang, X.; Tang, X. Deep learning face attributes in the wild. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 3730–3738. [Google Scholar]
  70. Lin, D.; Xu, G.; Wang, X.; Wang, Y.; Sun, X.; Fu, K. A remote sensing image dataset for cloud removal. arXiv 2019, arXiv:1901.00600. [Google Scholar]
  71. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
  72. Miyato, T.; Kataoka, T.; Koyama, M.; Yoshida, Y. Spectral Normalization for Generative Adversarial Networks. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
Figure 2. Overview of our inpainting network.
Figure 2. Overview of our inpainting network.
Remotesensing 14 06140 g002
Figure 3. Illustration of Partial Convolution, Gated Convolution, and our proposed Bilateral Convolution method. The value of feature F is close to 1, representing a more confident feature. The value of the mask map M is closer to 1, representing greater validity. The middle of the feature map is the damaged hole region, marked with a white contour.
Figure 3. Illustration of Partial Convolution, Gated Convolution, and our proposed Bilateral Convolution method. The value of feature F is close to 1, representing a more confident feature. The value of the mask map M is closer to 1, representing greater validity. The middle of the feature map is the damaged hole region, marked with a white contour.
Remotesensing 14 06140 g003
Figure 4. The architecture of our proposed MWA.
Figure 4. The architecture of our proposed MWA.
Remotesensing 14 06140 g004
Figure 5. Qualitative results of remote sensing datasets (AID, NWPU-RESISC45, and PatternNet). Local details are displayed in the red box.
Figure 5. Qualitative results of remote sensing datasets (AID, NWPU-RESISC45, and PatternNet). Local details are displayed in the red box.
Remotesensing 14 06140 g005
Figure 6. Qualitative results of typical image inpainting datasets (Paris StreetView, CelebA-HQ, and Places2). Local details are displayed in the red box.
Figure 6. Qualitative results of typical image inpainting datasets (Paris StreetView, CelebA-HQ, and Places2). Local details are displayed in the red box.
Remotesensing 14 06140 g006
Figure 7. Our cloud removal results compared with SpA-GAN on the RICE2 dataset.
Figure 7. Our cloud removal results compared with SpA-GAN on the RICE2 dataset.
Remotesensing 14 06140 g007
Figure 8. Loss curves of the AID, NWPU-RESISC45, PatternNet, and Paris StreetView datasets.
Figure 8. Loss curves of the AID, NWPU-RESISC45, PatternNet, and Paris StreetView datasets.
Remotesensing 14 06140 g008
Figure 9. Ablation qualitative results of the AID and Paris StreetView datasets. Local details are displayed in the red box.
Figure 9. Ablation qualitative results of the AID and Paris StreetView datasets. Local details are displayed in the red box.
Remotesensing 14 06140 g009
Table 1. Numerical comparison of the AID and NWPU-RESISC45 datasets.
Table 1. Numerical comparison of the AID and NWPU-RESISC45 datasets.
DatasetAIDNWPU-RESISC45
MetricsMask RatioPCGCRFRDSNPUTSpA-GANOursPCGCRFRDSNPUTSpA-GANOurs
FID ↓10–20%55.5639.0531.0420.7720.2952.2916.8646.3931.7325.5317.1314.9741.3413.09
20–30%83.3065.2845.9535.7534.0479.6228.0470.6054.5633.3529.0224.8664.0322.32
30–40%106.2487.6158.1448.5645.65106.7538.0390.1075.6141.3940.4233.7983.9330.76
40–50%122.87114.6272.2861.3556.85130.7148.14107.52100.1151.2351.7043.19108.6440.29
AVG91.9976.6451.8541.6139.2192.3432.7778.6565.5037.8834.5729.2074.4926.62
SSIM ↑10–20%0.8090.9040.8870.9170.9150.8350.9200.8190.9160.9050.9310.9310.8620.935
20–30%0.6720.8260.8220.8450.8410.7170.8510.6950.8470.8580.8680.8650.7690.877
30–40%0.5500.7420.7420.7630.7540.5930.7720.5790.7660.7880.7910.7860.6590.806
40–50%0.4360.6440.6460.6700.6570.4610.6810.4570.6680.6990.6990.6910.5280.717
AVG0.6170.7790.7740.7990.7920.6520.8060.6380.7990.8130.8220.8180.7050.834
PSNR ↑10–20%23.6728.3127.3928.9628.6524.9029.0524.3229.3628.5630.1430.0226.2630.43
20–30%20.6125.6525.1826.0525.7722.1326.2721.4426.6826.6227.1826.9723.7727.52
30–40%18.6523.9123.5924.1623.7919.9024.3519.3924.8324.9025.0824.8121.4825.47
40–50%17.2322.4822.2122.6822.2117.8822.8117.7523.3123.3923.4323.0619.3123.75
AVG20.0425.0924.5925.4625.1121.2025.6220.7326.0525.8726.4626.2222.7126.79
The ↓ indicates lower is better, while ↑ indicates higher is better. The bold number indicates that it has the best metric value compared with other methods on the same dataset.
Table 2. Numerical comparison of the PatternNet and Paris StreetView datasets.
Table 2. Numerical comparison of the PatternNet and Paris StreetView datasets.
DatasetPatternNetParis StreetView
MetricsMask RatioPCGCRFRDSNPUTSpA-GANOursPCGCRFRDSNPUTSpA-GANOurs
FID ↓10–20%53.6327.4833.7917.3314.1539.7812.7263.7820.6920.3316.2815.1437.7013.29
20–30%82.5549.4648.7728.8523.2864.9521.60102.1139.4828.9329.3927.5763.4923.32
30–40%105.7872.8462.8142.5832.7694.6029.98131.9158.6639.8442.0238.8585.0834.14
40–50%128.5197.7276.6154.1641.31124.0839.74158.8282.5149.9653.6649.56114.1846.83
AVG92.6261.8855.5035.7327.8880.8526.01114.1650.3434.7735.3432.7875.1129.40
SSIM ↑10–20%0.8220.9300.8980.9330.9370.8770.9390.8430.9520.9430.9520.9560.9070.960
20–30%0.6960.8720.8410.8750.8810.7830.8870.7300.9150.9080.9140.9130.8370.926
30–40%0.5800.7970.7660.8030.8130.6720.8220.6220.8650.8610.8590.8570.7420.877
40–50%0.4560.7000.6730.7180.7340.5410.7420.4980.7920.7990.7910.7910.6080.812
AVG0.6390.8250.7950.8320.8410.7180.8480.6730.8810.8780.8790.8790.7740.894
PSNR ↑10–20%24.0029.6627.8429.7629.9826.3930.3224.0631.5230.1831.0631.7127.5932.13
20–30%21.1227.0525.7926.9127.0523.6827.5321.0228.5027.7628.0528.2524.8229.18
30–40%19.1324.9024.0224.6624.8621.3325.2819.0226.0325.9925.9226.0622.3726.87
40–50%17.4723.1722.5523.0523.2019.1623.5917.1424.8024.2524.0524.1519.7124.87
AVG20.4326.2025.0526.1026.2722.6426.6820.3127.7127.0527.2727.5423.6228.26
The ↓ indicates lower is better, while ↑ indicates higher is better. The bold number indicates that it has the best metric value compared with other methods on the same dataset.
Table 3. Numerical comparison of the CelebA-HQ and Places2 datasets.
Table 3. Numerical comparison of the CelebA-HQ and Places2 datasets.
DatasetCelebA-HQPlaces2
MetricsMask RatioPCGCRFRDSNPUTOursPCGCRFRDSNPUTOurs
FID ↓10–20%9.202.545.171.911.871.4211.275.575.233.634.192.90
20–30%18.534.494.063.183.212.4322.0210.676.586.287.654.92
30–40%34.636.844.894.704.573.5835.7317.608.649.5912.006.96
40–50%56.459.836.116.205.804.7053.0127.8012.0013.7116.939.43
AVG29.705.935.064.003.863.0330.5115.418.118.3010.196.05
SSIM ↑10–20%0.9250.9790.9690.9810.9780.9830.8800.9510.9370.9540.9450.956
20–30%0.8600.9590.9580.9630.9580.9660.7910.9060.9030.9100.8890.914
30–40%0.7850.9310.9390.9390.9290.9440.7000.8520.8580.8560.8220.863
40–50%0.6990.8960.9130.9100.8950.9160.6000.7840.7970.7910.7420.798
AVG0.8170.9410.9450.9480.9400.9520.7430.8730.8740.8780.8500.883
PSNR ↑10–20%25.5032.2630.9332.7232.1033.1323.1628.5727.3428.5127.4228.72
20–30%22.2129.1028.9429.5328.8129.9320.2325.4624.9925.3224.0625.53
30–40%19.7826.7127.1127.1526.3427.4818.2423.3323.1523.1221.7423.32
40–50%17.8724.7825.4725.3424.4325.5516.6221.5321.4721.3319.8421.45
AVG21.3428.2128.1128.6927.9229.0219.5624.7224.2424.5723.2724.76
The ↓ indicates lower is better, while ↑ indicates higher is better. The bold number indicates that it has the best mertic value compared with other methods on the same dataset.
Table 4. Numerical comparison of cloud removal on the RICE2 dataset.
Table 4. Numerical comparison of cloud removal on the RICE2 dataset.
ModelsFID ↓SSIM ↑PSNR ↑
SpA-GAN45.440.73129.74
Ours39.680.78930.97
The ↓ indicates lower is better, while ↑ indicates higher is better. The bold number indicates that it has the best metric value compared with other methods.
Table 5. Analysis of the BC block network on the AID and Paris StreetView datasets.
Table 5. Analysis of the BC block network on the AID and Paris StreetView datasets.
DatasetAIDParis StreetView
MetricsFID ↓SSIM ↑PSNR ↑FID ↓SSIM ↑PSNR ↑
Mask RatioCONVBCCONVBCCONVBCCONVBCCONVBCCONVBC
10–20%22.5918.810.9050.91428.1228.6916.7613.440.9510.95831.0531.95
20–30%37.4731.980.8300.84225.4825.9431.2824.090.9110.92328.0928.98
30–40%51.0943.360.7440.75923.6924.0545.3834.960.8470.87225.6926.72
40–50%64.8955.820.6440.66222.1622.5061.8846.070.7630.80723.4824.71
AVG44.0137.490.7810.79424.8625.3038.8329.640.8680.89027.0828.09
The ↓ indicates lower is better, while ↑ indicates higher is better. The bold number indicates that it has the best metric value compared with other methods on the same dataset.
Table 6. Analysis of the MWA module on the AID and Paris StreetView datasets.
Table 6. Analysis of the MWA module on the AID and Paris StreetView datasets.
DatasetAIDParis StreetView
MetricsFID ↓SSIM ↑PSNR ↑FID ↓SSIM ↑PSNR ↑
Mask Ratiow/ CAw/o gOursw/ CAw/o gOursw/ CAw/o gOursw/ CAw/o gOursw/ CAw/o gOursw/ CAw/o gOurs
10–20%20.6217.0916.860.9120.9170.92028.5128.8429.0514.4513.5513.290.9570.9590.96031.7432.0632.13
20–30%35.1228.2428.040.8370.8470.85125.7726.0926.2726.5724.6023.320.9210.9250.92628.8129.1329.18
30–40%47.6138.7338.030.7530.7670.77223.9124.1924.3537.5834.8034.140.8660.8750.87726.4626.7826.87
40–50%60.9148.8448.140.6550.6730.68122.3722.6422.8150.0448.1846.830.7960.8070.81224.3824.7524.87
AVG41.0733.2332.770.7890.8010.80625.1425.4425.6232.1630.2929.390.8850.8920.89427.8528.1828.26
The ↓ indicates lower is better, while ↑ indicates higher is better. The bold number indicates that it has the best metric value compared with other methods on the same dataset.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Huang, W.; Deng, Y.; Hui, S.; Wang, J. Image Inpainting with Bilateral Convolution. Remote Sens. 2022, 14, 6140. https://doi.org/10.3390/rs14236140

AMA Style

Huang W, Deng Y, Hui S, Wang J. Image Inpainting with Bilateral Convolution. Remote Sensing. 2022; 14(23):6140. https://doi.org/10.3390/rs14236140

Chicago/Turabian Style

Huang, Wenli, Ye Deng, Siqi Hui, and Jinjun Wang. 2022. "Image Inpainting with Bilateral Convolution" Remote Sensing 14, no. 23: 6140. https://doi.org/10.3390/rs14236140

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop