Next Article in Journal
Evacuation Simulation Implemented by ABM-BIM of Unity in Students’ Dormitory Based on Delay Time
Previous Article in Journal
A Tale of Two Cities: COVID-19 Vaccine Hesitancy as a Result of Racial, Socioeconomic, Digital, and Partisan Divides
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SAM-GAN: Supervised Learning-Based Aerial Image-to-Map Translation via Generative Adversarial Networks

School of Geoscience and Technology, Zhengzhou University, Zhengzhou 450001, China
*
Author to whom correspondence should be addressed.
ISPRS Int. J. Geo-Inf. 2023, 12(4), 159; https://doi.org/10.3390/ijgi12040159
Submission received: 18 February 2023 / Revised: 5 April 2023 / Accepted: 6 April 2023 / Published: 7 April 2023

Abstract

:
Accurate translation of aerial imagery to maps is a direction of great value and challenge in mapping, a method of generating maps that does not require using vector data as traditional mapping methods do. The tremendous progress made in recent years in image translation based on generative adversarial networks has led to rapid progress in aerial image-to-map translation. Still, the generated results could be better regarding quality, accuracy, and visual impact. This paper proposes a supervised model (SAM-GAN) based on generative adversarial networks (GAN) to improve the performance of aerial image-to-map translation. In the model, we introduce a new generator and multi-scale discriminator. The generator is a conditional GAN model that extracts the content and style space from aerial images and maps and learns to generalize the patterns of aerial image-to-map style transformation. We introduce image style loss and topological consistency loss to improve the model’s pixel-level accuracy and topological performance. Furthermore, using the Maps dataset, a comprehensive qualitative and quantitative comparison is made between the SAM-GAN model and previous methods used for aerial image-to-map translation in combination with excellent evaluation metrics. Experiments showed that SAM-GAN outperformed existing methods in both quantitative and qualitative results.

1. Introduction

In modern life, maps are relevant to us. From general geo-information location search to in-car navigation, maps have become indispensable to human society. Traditionally, in the geographic information industry, maps are first measured in the field and then vectorized using professional mapping tools, which involves much time and repetitive human resources. Getting maps corresponding to the current situation in real-time is still challenging. Although various map vendors (Google Maps (Google Inc. in Mountain View, CA, USA), Baidu Maps (Baidu Inc. in Beijing, China)) already provide massive amounts of map information, there is a considerable iteration time gap as map vendors rely on manual collection to update map data. Therefore, this does not meet the needs of individual users. However, with the development of deep learning in the direction of image migration, a completely different means of mapping has emerged. In a review of the development of cartography in China, Liao Ke [1] suggests that future map mapping will be more intelligent and personalized. To realize the above needs, many scholars have used aerial or remote sensing imagery combined with image migration techniques to generate maps in an automated manner. This work makes it possible to generate maps without needing vector data. The success of such models will not only introduce new cartographic ideas to the field of map generation but also promote further development of the field of map mapping towards personalization and intelligence.
Recent image-to-image translation algorithms for map generation based on aerial imagery have achieved great success. However, there is still massive room for image accuracy and quality improvement.
In recent years, adversarial generative models have flourished and have had good results in various fields. Previous map generation models for aerial imagery did not focus on learning the target domain style from the model input level. The popular approach in the past was to guide the model to generate the target domain style results through pixel-level loss. Still, in this paper, these were generated by encoding the content and style of the aerial image and the map separately, i.e., the generator of the generative model includes not only the content encoder but also the style encoder. This allows the generated map to learn the map style better and faster while retaining the aerial image content. In addition, current generative models proposed in this field do not handle the topological relationships in the generated maps well. Our model adds topological consistency loss to bring about a significant visual improvement in the model results. In this paper, a deep learning model based on adversarial mechanisms aims to find a framework for aerial image-to-map transformation based on generative adversarial networks. We propose a novel supervised learning-based generative adversarial network for aerial image-to-map conversion. The model improves the applicability of map generation models by solving the problems of inadequate learning of map styles and topological integrity of map generation models. The model achieves generalization from aerial images of unknown regions to target map classes by learning the potential encoding space of content and styles extracted from aerial images and maps. Not only does our model include topological consistency constraints and style encoders, but we also find that adding an attention mechanism to the adversarial generation model can motivate the encoders to acquire more critical information when learning the potential encoding space corresponding to the images.
In this paper, we learn the content and style of images in different domain spaces and introduce attention mechanisms into the GAN architecture, proposing a novel supervised learning-based generative adversarial network for aerial image-to-map translation (in the paper, we will use SAM-GAN as a shorthand for the model) that aims to improve the quality and accuracy of the aerial image-to-map translation. Unlike traditional models, the generator of SAM-GAN has three modules: style encoder, content encoder, and decoder, while using multi-scale discriminators to integrate image information of different sizes to increase the guidance capability of the generator. Our network uses paired datasets, and this model performs better in terms of accuracy and visual perception by comparing it with the previous SOTA model.

2. Related Work

We divide the current work into three aspects: Firstly, the current state of research on image-to-image translation algorithms based on generative adversarial networks; secondly, the role of attention mechanisms for image conversion models; finally, the current state of development of aerial image-to-map translation work.
Image-to-image translation: In recent years, GAN [2] and image translation techniques have been used to generate images for a variety of purposes, in directions such as data enhancement [3,4], road extraction [5], and super-resolution images [6,7].
As a model with “infinite” generative power, a direct application of GAN is modeling, generating data samples consistent with real data distribution. Thus, GAN-based image-to-image translation has continued to achieve compelling results [8]. The pix2pix proposed by Phillip Isola [9] learns the mapping relationship between two domains using paired images and achieves better results. One of the most classical algorithms in unpaired inter-image translation, the CycleGAN model [10], learns the mapping relationship between the original and target domains well using cyclic consistency loss. The contemporaneous DualGAN [11] and DiscoGAN also illuminate the non-pairing domain. DualGAN uses the idea of coupling to extend the native GAN into two mutually coupled GANs, which provide substantial performance gains compared to a single GAN. DiscoGAN, proposed by Taeksoo Kim et al. [12], learns the relationship between different domains and then uses the learned domain relationships to successfully transfer styles from one domain to another while retaining the key attributes of the original domain. However, the generated results of these models applied to aerial imagery and map transformation have not been able to achieve better visual results. The subsequent StarGANv2 proposed by Yunjey Choi et al. [13] aims to generate a single framework for image diversity and multi-domain scalability.
Recently, the UNIT [14] model based on the assumption that two images with the same semantic content but different style domain spaces have the same latent space vector has achieved good results in unsupervised image transformation. Later, the MUNIT model [15] argues that the shared space can encode both the content space and the space in which they should differ, i.e., the style space, thus proposing a multimodal unsupervised image-to-image migration framework. It generates polymorphic images by encoding different combinations of styles. The FUNIT [16] model, based on UNIT, achieves better results and improves the generalization ability of the model when trained on small- sample datasets.
Attention mechanisms in deep learning: The attention model comes from the study of human vision, where humans selectively focus on parts of the comprehensive information, which is the attention mechanism. Similarly, attention can be added to deep learning methods to motivate the model to assign weights to different regions to learn critical information and thus improve the model’s performance. The attention mechanism has demonstrated its performance in various tasks, such as machine translation [17,18], image generation [19,20], and image classification [21,22,23].
The plug-and-play performance of the attention mechanism has made it widely used in deep learning applications and has minimal impact on model overhead. Volodymyr Mnih [24] combined recurrent neural networks with visual attention to improve image classification accuracy, and attention mechanisms are gaining popularity. Ashish Vaswani [25] used an attention mechanism to achieve excellent results in machine translation. The SeNet [26] model proposed by the autonomous driving company Momenta proposes the general-purpose channel attention module SeBlock, which makes better use of inter-channel correlation by both compression and excitation, and achieves outstanding performance in image recognition. CBAM [27] proposes a lightweight, general-purpose module combining spatial attention and channel attention, which can be used with convolutional neural networks for end-to-end training, improving model classification and detection performance with minimal overhead. ECA-Net [28] implements a local cross-channel interaction strategy without dimensionality reduction and adaptive selection of one-dimensional convolutional kernel size on top of SeNet, thus achieving improved performance.
Recent studies have found that incorporating attention mechanisms into generative adversarial network models can also achieve good performance. Self-AttentionGAN [19] uses spectral normalization and self-attentive mechanisms to learn the interdependencies between global features better, combining self-attentive mechanisms and convolutional neural networks to model long-range multi-level image regions, providing a method to combine global information. It is a better solution to the problem that GANs do not quickly learn specific structural and geometric features. Attention-GAN [29] decomposes the generator into two separate networks. The model will acquire the attention graph and then fuse the attention graph with the original image to obtain the final generated result. Hajar Emami’s proposed SPA-GAN model [20] improves the model effect by using a discriminator to get the attention graph back to the generator, thus achieving better results on unpaired datasets.
Style translation of aerial images-to-maps: Inspired by generative adversarial networks, deep learning adversarial mechanisms to learn mapping patterns between aerial images and maps started to attract attention in map mapping. The data used in this paper is the Maps dataset (http://efrosgans.eecs.berkeley.edu/pix2pix/datasets/maps.tar.gz (accessed on 10 February 2023)) [9]. In that related work, when the models involved use that dataset, we provide a simple and clear comparative analysis of their performance on that dataset. Jun Gu et al. [30] better extracted feature edge information using a generator architecture with cyclic consistency loss and pix2pixHD [31]. However, combined with the model’s performance on the Maps dataset, we argue that the model only combines the generator of the Pix2pixHD model and the discriminator along the lines of the Pix2pix model [9], and does not improve the quality of the generated maps any better. GeoGAN [32] proposed a RealNVP bijection model based on CGAN [33] that can better learn the style space of the target domain, although the RealNVP flow model idea used in this model presents a new perspective on the research problem of this paper. However, in combination with the model’s results on the dataset Maps, the feature boundaries of the generated maps are very blurred. Furthermore, the generated maps are poorly rendered in color due to insufficient learning of the map domain. Users are unable to obtain useful information from the generated maps. The SG-GAN [34] model later rendered the crowdsourced GPS data into images using layers and integrated them into GAN while combining semantic rules to estimate the high-level information of the images, significantly improving the road in image migration accuracy. However, it requires a large amount of GPS data, which may affect the model’s usefulness. On the other hand, the MapGAN model [35] uses a rendering matrix to focus on geographic entity attributes and aesthetic color rendering to give a better visual effect to the generated map. Although MapGAN has improved the performance of the generated maps in terms of color rendering, it can lead to overfitting problems with the model. As shown by its performance on the Maps dataset, it does not solve the problem of generating maps with clear boundaries between feature boundaries nor does it guarantee the accuracy of the color rendering locations.

3. Materials and Methods

In this section, we first present the proposed SAM-GAN’s general architecture and then detail the model’s components. Finally, we explain the role of each loss function used in the learning of the model.

3.1. The Core of SAM-GAN

We propose SAM-GAN, a supervised learning-based architecture for transforming aerial images and maps. Figure 1 shows the core idea of SAM-GAN, which is inherited from the UNIT [14] model. It indicates that the potential space of an image can be decomposed into a content space and a style space where the content space represents the content of the source domain and the style space represents the style of the target domain. By decomposing and selectively combining the content and style spaces from different domains, the model is motivated to consider both the content space of the source domain and the style space of the target domain. In this paper, the source domain refers to aerial images, and the target domain refers to maps.

3.2. The Overall Architecture of SAM-GAN

Our model aims to train generators to learn mapping relationships between source and target domains. The SAM-GAN model is divided into two main parts: a generator and a discriminator, where the generator consists of a content encoder, a style encoder, and a decoder. Similar to other GAN problems, SAM-GAN needs to have excellent capabilities in the training and testing process to transform unlearned source domain images well into images. In a data form, we use x to represent aerial images and y to represent map images, where y denotes a collection of K maps in the target domain, i.e., y = {y1, y2, … yK}.
Figure 2 shows the detailed architecture of the model. The input to the model is divided into two parts: the first part is an aerial image, and the second is K maps, where the aerial image is used in the content encoder to extract the content space. In contrast, the maps are fed into the style encoder to learn the style space of the target domain, resulting in a model architecture suitable for two-domain transformation. One of the things to note is that the order of the images input to the model during training is random. The discriminator guides the generator during training by learning the distribution of images in the source and target domains, thus allowing the generator to generate a more realistic map. In model testing, the model is fed with aerial images that have never been seen in training. The model must generate the corresponding maps when we provide the completed model with aerial images from the source domain.

3.2.1. Generators

The generator of SAM-GAN, an aerial image-to-map translation model, consists of three parts: a content encoder, a style encoder, and a decoder.
Unlike previous aerial image-to-map translation methods, SAM-GAN introduces a style encoder to learn the style space of the target domain. The input of the style encoder contains K map images from the target domain; each map is passed through the style encoder to obtain a one-dimensional vector, and the final output of the style encoder is obtained by averaging the K potential style encoding vectors. The result represents the potential coding vectors in the style space of the map domain, which in turn guides the decoder to work better.
In addition to the fully connected layer, the content encoder and the decoder are based on an improved Pix2pix [9] model architecture. The content encoder consists of four parts: the input convolution block, the attention block, the downsampling block, and the residual block. The decoder also has four components: the residual block, the upsampling block, the output convolution block, and the multilayer perceptron. As shown in Figure 2, PCIR denotes the input convolutional layer, CIR denotes the downsampling layer, SeBlk [26] denotes the attentional layer, ResBlk denotes the residual block, CTIR denotes the upsampling layer, and PCTanh denotes the output convolutional layer. Three consecutive FC blocks denote a multilayer perceptron. The main improvements include three aspects:
  • Adding the channel attention mechanism SeBlock (abbreviated as SeBlk) at the beginning of the encoder, which can obtain the importance level of each channel in the feature map, then use this importance level to assign the corresponding weight value to each feature, thus allowing the model to focus on specific feature channels and enhance the channel importance of certain feature maps, which ultimately allows the model to better filter out important information in a large amount of primary.
  • We use nine residual blocks, of which seven are used in the content encoder and two in the decoder using adaptive instance normalization, which we call adaptive residual blocks, where the initial weights of the adaptive residual blocks come from the multilayer perceptron. Adaptive residual blocks enable the decoder to combine the content space obtained from the content encoder and the style space obtained from the content encoder and the style encoder to generate high-quality translation results.
  • We establish connections at the channel level between the content encoder and the decoder via jump connections which allows the model to better combine low-level features from earlier stages with high-level semantic features from later stages to achieve cross-layer information propagation and serves to prevent the model from gradient disappearance during the training process.
  • In the decoder part of SAM-GAN, the adaptive instance normalization (AdaIN) [36] method that we use is a fundamental reason for the decoder to make efficient use of the style space learned by the style encoder. AdaIN works in detail as follows: each channel of the latent content encoding feature map ZC obtained from the content encoder is first normalized, during which a multi-layer perceptron consisting of three fully connected layers of the decoder computes the style latent encoding vector ZS from the style encoder to obtain the mean and variance vectors (μ,σ2); then, μ is applied to the adaptive residual block of the decoder as a deviation and σ2 as a scaling factor to jointly parameterize the affine transformation.

3.2.2. Discriminator

The discriminator idea of the SAM-GAN model comes from the PatchGAN [9] architecture, which uses a receptive field of size 70 × 70 to judge the resulting feature map. The specific discriminator architecture is shown in the right-hand area of Figure 2. During the model’s training, the discriminator not only learns aerial and map images but also feeds the results back to the generator to guide its generation process.

3.3. Loss Function

For SAM-GAN to efficiently understand the coding space learned by content and style encoders, our model combines the following loss terms: adversarial loss, content loss, and style loss. They ultimately constitute the total loss function of the model.

3.3.1. Adversarial Loss

The model SAM-GAN aims to convert aerial images into maps, where the generator’s purpose is to generate realistic maps through the learned style space, while the purpose of the discriminator is to distinguish between the actual and false maps of the input. Thus, the generator minimizes the adversarial loss, while the discriminator maximizes the objective function. The two form a game relationship in which the model moves towards equilibrium in a constant adversarial process. The equation can therefore be expressed as
min G   max D   L GAN G , D = E y ~ p y log D y + E x ~ P x , y ~ P y log 1 D G x , y

3.3.2. Content Loss

The content loss of the SAM-GAN model consists of three aspects: VGG loss, L1 pixel-level loss, and topological consistency loss.
The VGG loss [37] compares the gap between the generated image and its corresponding real map. The loss is achieved using a pre-trained VGG-19 model [38] to extract features from the style-transformed image and its corresponding real map, obtaining multiple feature maps corresponding to both. Then we use the L1 loss function and set different weights to calculate the gap between the corresponding feature maps, which is expressed as
L VGG = E x ~ P x , y ~ P y i , j F i j l P i j l 1
where Nl is used to denote the number of channels contained in the lth feature map, Ml denotes the Nl feature map contained in the lth feature map, and Ml is the height of the feature map multiplied by its width so that the lth feature map can be expressed as F l R N l × M l , where F i j l denotes the feature value of the jth position of the ith channel in the lth feature layer.
The L1 pixel-level loss [39] is a direct calculation of the L1 loss between the style translation result and the real map, as follows:
L p i x e l = E x ~ P x , y ~ P y G x y 1
In geography, topological relations reflect the logical relationships between spatial entities and describe the connections between lines and planes in space. So, a map that can express better topological information is one of the prerequisites for its practical application. Topological consistency loss is added to the optimization objective in SAM-GAN to have better structural characteristics of the elements in the generated map. The topological consistency loss refers to the content loss function of the SMAPGAN model [40]. In maps, features are mainly composed of points and lines, and the topological relationships between entities can be represented using their edges, whereas in digital image processing, gradients can be used for edge detection. Thus, we can use image gradients to represent the edges of image features and thus express the topological relationships between map elements.
Let the pixel value of a pixel point (i, j) of the image be H(i, j). The horizontal gradient of this pixel point is
v i i , j = H i + 1 , j H i 1 , j
The vertical gradients are
v j i , j = H i , j + 1 H i , j 1
Therefore, the gradient of the image at the pixel point is known to be
L t o p = E x ~ P x , y ~ P y v G x v y 1
The loss of the elements comprising the above three areas is therefore
L c o n t e n t =   α 1 L v g g + α 2 L p i x e l +   α 3 L t o p
where α1, α2 and α3 represent the weights of the corresponding loss terms.

3.3.3. Style Loss

Gatys [41] proposed that the representation of content and style in a convolutional neural network is separable and that the image has a good target domain style after translation by using style loss to measure the style before and after the image style translation. We input the style-transformed image with its corresponding real map into the VGG-19 network, extract each feature map layer, and calculate the corresponding Gram-style matrix [42]. The Gram-style matrix can describe the correlation between features and also possesses the feature of positional insensitivity. Hence, we use the Gram-style matrix as a quantified description of image style. Since one feature map corresponds to one Gram- style matrix, for multiple feature maps, we use the set {GM1, GM2, …, GMl} to denote the set of Gram-style matrices corresponding to the feature map F. We use the L1 loss function, which is represented by the following equation:
G M i j l = k F i k l F j k l
L s t y l e = E x ~ P x , y ~ P y ( | 1 N l M l G M l G x G M l y | 1 )

4. Results

In this section, we first describe the sources of the dataset. The model is then compared quantitatively using four objective evaluation metrics. Then, the qualitative comparison phase illustrates the performance of the SAM-GAN-generated results in terms of visual perception using three subjective evaluation metrics. Finally, we perform ablation experiments to analyze the functionality of the model components.

4.1. Dataset and Experimental Setups

The Maps dataset [9] was created from satellite and electronic map tiles from Google Maps, with an image size of 600 × 600 pixels. It consists of 2194 data pairs, each consisting of an aerial image and its corresponding map tile image extracted from Google Maps. The sample area included in the dataset is New York City and its surrounding area, and the aerial imagery has a spatial resolution of 2.15 m/pixel in the visible band.
We used the PyTorch (version 1.9.0) framework to implement our model. In terms of hardware, an NVIDIA GeForce RTX 3090 graphics card with 24 GB of video memory was used for training. We choose the Adam optimizer in the model training details and set β1 = 0.9, β2 = 0.999. We also set the initial learning rate to 0.0002 and set it to decrease by half every 50 cycles. The training batch size of the model is 4, the size of the model input and output images are both 256 × 256, and the weights in the content loss function are α1 = 10, α2 = 10, and α3 = 1.
To fairly compare the performance of each method, all models in the experiment use the same image pre-processing process in the training process.

4.2. Baselines

We chose four SOTA methods to compare with the model proposed in this paper: AttentionGAN [29], CycleGAN [14], Pix2pix [9], and UNIT [14]. Among them, pix2pix requires paired datasets for training, while AttentionGAN, CycleGAN, and UNIT are unsupervised training models and do not require paired datasets.
Pix2pix, an image transformation model based on the CGAN model, was used as one of the baseline models for this experiment due to the generality of its architecture and its better performance in supervised learning of image-to-image transformations.
CycleGAN, as one of the classical unsupervised learning architectures for image transformation models, proposes a cyclic consistency loss that can still perform image transformation between two domains in the absence of paired data, which first attempts to map images from the source domain to the target domain and then back to the original domain, reducing the difference between the reconstructed image and the original domain while motivating the model to learn better the mapping relationship between the source and target domains. The relationship between the source and target domains is reduced.
AttentionGAN minimizes the difference between the source and target domains in the data generation distribution by adding an attention mechanism to CycleGAN.
The UNIT model is a novel unsupervised image-to-image translation model that learns the spatial distribution of the original domain image by separating its content encoding from the style encoding of the target domain image, which in turn fuses to obtain the converted target domain image eventually.

4.3. Evaluation Metrics

In our experiments, we combined objective and subjective evaluations to assess the maps generated by the model. We have paired data so that the generated images have their counterparts in real maps. We considered the images’ quality, diversity, and accuracy in choosing the evaluation metrics. For the quantitative evaluation, we chose four evaluation metrics: inception score [43], Fréchet inception distance [44], structural similarity [45], and pixel accuracy [46]. For the qualitative evaluation, we use three metrics: content retention, style similarity, and map availability [47].

4.3.1. Inception Score (IS)

IS is a class of metrics used to assess the quality of the generated images. It evaluates the quality of images in two main aspects: sharpness and diversity. In the comparison experiments, we calculated the IS scores of the images generated by each model on the test set and then compared them. The higher the IS score, the better the quality of the images generated by the model and the better the model’s generalization ability.

4.3.2. Fréchet Inception Distance (FID)

As the IS metric only evaluates image quality, we introduced the FID to reflect the distance between the real and generated images. The lower the FID score, the better, which indicates that the model generates image results that are more similar to the real image.

4.3.3. Structural Similarity (SSIM)

The purpose of SSIM is to measure the similarity index between two images. It considers three main characteristics of an image: brightness, contrast, and structure. As a perception-based evaluation metric, SSIM is visually more aligned with human intuition. The formula is as follows:
SSIM x , y = 2 u x μ y + c 1 2 σ x y + c 2 u x 2 + u y 2 + c 1 σ x 2 + σ y 2 + c 2
where x is the image obtained by the generator and y is the real map image corresponding to x. μx is the mean of x, μy is the mean of y, σ x 2 is the variance of x, σ y 2 is the variance of y, and σ x y is the covariance of x and y. c1 = (k1L)2 and c2 = (k2L)2 are constants used to maintain stability (k1 = 0.01, k2 = 0.03, and L is the dynamic range of pixel values). The range of SSIM is [−1, 1], and the value of SSIM is equal to 1 when the two images are identical.

4.3.4. Pixel Accuracy

The abbreviation for pixel accuracy is PixelACC (%). It is implemented as follows: first, the RGB values (ri, gi, bi) of the original map and the RGB values r i ,   g i ,   b i of the generated map are obtained. Then the difference between each channel is calculated, and the maximum difference is compared with the thresholdδ. If the result is less than δ, the generated image is considered similar to the real map and generated accurately.
r i r i + g i g i + b i b i < δ
In this paper, we take δ to be 1 and 5.

4.4. Comparisons with Baselines

We used objective and subjective evaluation metrics to compare the model proposed in this paper with the four baseline models mentioned above, thus demonstrating the effectiveness of SAM-GAN in the aerial image and map translation.

4.4.1. Objective Evaluation

This section compares the SAM-GAN model with each baseline model to illustrate the SAM-GAN model’s powerful capabilities. In the comparison experiments, we set the generator’s style encoder parameter K equal to 1 and tested the translation results of SAM-GAN and the baseline model on the same test set using four different image quality evaluation metrics to illustrate the generalization capabilities of each model. Table 1 shows the quantitative comparison between the baseline model and SAM-GAN. Table 2 gives the training time required for each model.
The evaluation metric scores show that the SAM-GAN framework significantly outperforms the baseline model in the four metrics of IS, FID, SSIM, and PixelACC(%) in the aerial image-to-map translation task. Regarding IS, by comparing the IS values of the models, we found that SAM-GAN has the largest IS value, which indicates that SAM-GAN generates the highest-quality images. In terms of FID, SAM-GAN still has the best performance, with a two-fold improvement relative to the Pix2pix model and a three-fold improvement compared to CycleGAN. In terms of SSIM, SAM-GAN doubles its performance compared to the baseline model AttentionGAN. Regarding pixel accuracy, SAM-GAN achieves an accuracy of approximately 33.364% with parameter δ equal to 5, which is over 9% better than CycleGAN and nearly 10% better than UNIT. FID, SSIM, and pixel accuracy results show that SAM-GAN generates the closest results to the real map. Table 1 illustrates that previous image-to-image translation methods do not make good use of a priori knowledge to learn target classes better when applied to aerial image-to-map translations because these models cannot learn the style space and topological feature details of the target domain well. SAM-GAN uses content and style encoders to combine the content of the original domain with the style of the target and the attention module to focus on the essential parts of the different domain spaces while preserving the topology of the image to achieve a higher-quality image transformation.
The training time of the models given in Table 2 shows that the training time of Pix2pix using supervised learning strategies is in the range of 3–4 h, while the three models using unsupervised learning strategies, CycleGAN, AttentionGAN, and UNIT, take longer, with the first two of them having cyclic consistency loss taking around 7–10 h to train, while UNIT took the longest time of approximately 30 h. In contrast, the SAM-GAN model in this paper required the shortest training time, at 1–2 h. So SAM-GAN has higher translation results in terms of image quality and better model convergence and learning ability in the aerial image and map translation applications.

4.4.2. Subjective Evaluation

The results of the objective evaluation metrics show that the accuracy scores of the baseline models and SAM-GAN appear to be low. However, from a visual perspective, slight differences in accuracy do not have a decisive impact on using the maps. Therefore, to determine the visual effectiveness of the models, 25 people from the GIS profession with knowledge of map mapping were invited to qualitatively evaluate the output of the baseline and SAM-GAN models in terms of content retention, style similarity, and map usability. The scores for the three metrics take a range of 1–10, with higher scores representing the better performance of the image on that metric.
Content retention: indicates whether the generated map matches the content information of the corresponding aerial image; a higher score means the complete content retention of the generated map.
Style similarity: indicates the degree of similarity between the generated map and the corresponding real map; a higher score means that the generated map is more similar to the corresponding real map.
Map availability: indicates whether the generated map can be used in reality; a higher score represents its ability. Higher scores represent better information on features and the ability to express map elements in the generated map.
Table 3 shows a qualitative comparison between the baseline model and the SAM-GAN model, and Figure 3 shows the details of the images generated by the model. The comparison of the images in Figure 3 shows that the SAM-GAN model performs better in terms of the quality and structure of the generated results compared to the Pix2pix supervised learning method, e.g., the degree of correct distribution of colors corresponding to different features and the detail of buildings. Compared to the other three unsupervised learning methods, SAM-GAN performs better in terms of topological integrity of features, clarity of boundary lines between different features, and color distribution.
The data in Table 3 show that SAM-GAN is the best compared to the baseline models regarding content retention, style similarity, and map availability, indicating that the model is the best regarding subjective perception. The results for map availability show that although the SAM-GAN model does not score well in some quantitative evaluation metrics, this does not negate its potential use in some practical scenarios that require personalization scenarios, e.g., rescue and disaster relief.

4.5. Exploration of the Number of Style Encoder Input Maps K in Training

In this section, the main focus is to explore whether there is a relationship between the ability of the model and the K-value. We plan to observe the model’s performance by setting different K values with the help of four types of objective evaluation metrics. As shown in the table below, we input an aerial image to the content encoder in training, and K randomly selected real maps for the style encoder. We set the values of K to be 1, 3, 5, 8, and 10 and evaluate the scores of the model under different K values on the test set using four evaluation metrics.
Table 4 shows the model scores on each evaluation metric for different K values. Figure 4 and Figure 5 show the trends of the evaluation metrics for different K values taken by the model, while Table 5 shows the time taken to train the model for different K values.
The scores in Table 4 show that the model achieves first place in IS, FID, and PixelACC (%) when K is taken as 1. Although the score of SSIM (K = 1) is smaller than that of SSIM (K = 5), the difference is only about 0.001. The trend of each evaluation metric with K shown in Figure 4 and Figure 5 shows that the performance of SAM-GAN does not increase with K. There is a more obvious fluctuation in the number of domain classes. The training time of the model varies for different values of K. The K value is proportional to the training time of the model, and the model has the shortest training time and the best generalization ability on the test set when K is set to 1. Unlike traditional pairwise training models, the above findings demonstrate that the model in this paper can achieve fast style migration with small samples.
By combining SAM-GAN’s results in quantitative evaluation and training time, we concluded that the number of maps input to the training process affects the performance of SAM-GAN, where the model performs best with a K range of 1–5. This also indicates no mutual influence between the content encoder learning image content and the style encoder learning image style in SAM-GAN training. The style encoder uses the learned target domain style encoding vector to guide the decoder in generating content images with the target domain style.

4.6. Ablation Study

To determine the plausibility of SAM-GAN for aerial image translation applications, we performed ablation tests where K was taken as 1. We validated the importance of each SAM-GAN component on the same test set. First, we kept all parts of the model. Second, we removed only the style encoder. Third, we removed only the VGG loss in content loss. Fourth, we removed only the topological consistency loss in content loss. Finally, we removed only the SeBlock module, the attention mechanism used. These ablation methods are described as follows:
  • SAM-GAN-no-StyleEncoder: SAM-GAN has the style encoder removed from the generator.
  • SAM-GAN-no-VGGLoss: SAM-GAN does not use VGG content loss.
  • SAM-GAN-no-TOPLoss: SAM-GAN does not use topological consistency loss.
  • SAM-GAN-no-SeBlock: the attention module is removed from the generator of SAM-GAN.
Table 6 shows the objective evaluation scores for each ablation experiment. Table 7 shows the corresponding training time for the model under each experiment. Figure 6 shows the generated results for each experiment on the dataset Maps.
Comparing the results in Table 6, the SAM-GAN-no-StyleEncoder with the missing style encoder has the worst results. At the same time, in Figure 6, it can be seen that this experiment corresponds to a substantial decrease in the quality of the generated images and the pixel-level accuracy of the images. This is followed by the performance of SAM-GAN-no-VGGLoss and SAM-GAN-no-TOPLoss, which differ from SAM-GAN in objective evaluation, mainly regarding the FID and SSIM metrics of image quality. In Figure 6, it can be seen that SAM-GAN-no-VGGLoss generates results with a larger gap in terms of the accuracy of the color and distribution of features in the images compared to SAM-GAN. The difference between SAM-GAN-no-TOPLoss and SAM-GAN is mainly in the topological integrity of the roads and the clarity of the boundary lines between features. For the channel attention mechanism used in the model, we find that SAM-GAN-no-SeBlock performance is the closest to SAM-GAN. By, combining the time required for training each ablation experiment provided in Table 7, we know that SAM-GAN-no-SeBlock is second only to SAM-GAN in terms of time needed; despite this, the time cost of the former is close to twice the time cost of the latter, which indicates that SeBlock acts as an accelerator for model learning.
In summary, the primary role of the style encoder is to improve the image quality by learning the underlying coding space, while the VGG and topological losses are mainly used to improve the accuracy of feature boundary lines and the accuracy of the model for rendering feature colors in the generated maps, thus increasing the structural correlation between the generated maps and the real maps and also improving the visual performance of the generated results. The attention mechanism used in SAM-GAN enhances the learning ability of the model during the training process and accelerates the convergence of the model. Combining the qualitative and quantitative results shows that SAM-GAN generates high-quality maps while also enhancing the visual aspects of the images.
SAM-GAN provides better visual translation results from aerial images than SAM-GAN-no-StyleEncoder, SAM-GAN-no-VGGLoss, SAM-GAN-no-TOPLoss, and SAM-GAN-no-SeBlock, especially in terms of the edge detail of buildings in the image, completeness of roads, features, the rendering of corresponding colors, etc. This suggests that using the style encoder, the content loss, and the attention mechanism proposed in this paper can improve the model’s ability to capture image details.

5. Discussion

SAM-GAN presents an end-to-end aerial image-to-map-based translation model, which performed well compared to some SOAT models. Although SAM-GAN can achieve better results than other image translation models in never-before-seen test sets, the results are not always positive. For example, when there are blurred areas in the aerial image itself and when there are colored areas in the image that makeup very little of the image as a whole, the model results may show inaccurate color rendering and blurred details. The above illustrates the model’s shortcomings but simultaneously will prompt us to explore more options.
  • The model’s performance depends not only on the number of datasets but also on the data quality. We note that some of the aerial images in the training set appear to be in the same area but correspond to different features on the map. Moreover, the maps sometimes do not contain complete road information, so we speculate that improving the image quality of the dataset may improve the model’s learning of feature classes and edge information.
  • The SAM-GAN in this paper is a two-domain translation model. Given that the content and style encoders in the model generator do not interact with each other in learning information in the corresponding domains, the dataset can be expanded to use aerial images of different resolutions and map tiles of different styles in combination with each other to achieve multi-domain translation, which may enhance the generality of the model. However, as the number and variety of datasets increase, higher demands on the model’s training time and hardware equipment are expected.
  • More target domain styles can be used, not limited to map styles, e.g., art maps and game maps.
The current SAM-GAN model demonstrates its notable performance in generating maps using a novel mapping approach.
We also present a partial mechanism analysis of the methods and results of this paper from a cartographic perspective.
The traditional process of map mapping involves the acquisition of vectorized data and later processing by a large number of professional cartographers. This not only has strict requirements on the data source but also requires a lot of human resources and time, making the whole mapping process long and difficult to adapt to the needs of time-sensitive mapping in today’s big data era. Since the 21st century, with the development and advancement of Internet information technology and visualization, the huge amount of data generated has become the driving force of the times, bringing new opportunities and challenges to various fields. The combination of digital information technology and cartography has led several innovative studies. Compared with the data used in traditional mapping, aerial remote sensing image data are richer in data type and volume and have significant advantages in terms of access, fineness, coverage, and real-time. Therefore, in this paper, we choose to use aerial remote-sensing images as the data source to improve the timeliness and convenience of map making.
As batch map generation is the current goal in aerial image and map translation, a deep learning model based on generative adversarial networks was chosen for this study, which is ideal for batch data generation. Based on the experimental results of this paper, the SAM-GAN method proposed in this paper can achieve the effect of processing a large amount of map data efficiently and accomplish the task of batch map generation excellently. Therefore, the research of the model based on aerial remote sensing image big data is beneficial to improve the efficiency of mapping and reduce the difficulty of mapping.
The model in this paper realizes the fitting of remote sensing images to map tile data, which can improve the clarity of map information expression and increase the information expression ability based on improving map mapping efficiency.
Firstly, compared to aerial remote sensing image data, map tiles can represent the information and categories of features more clearly. Users do not need professional knowledge of image interpretation to recognize the information expressed in the map. The resulting maps greatly extend the range of map users with an unambiguous representation. Secondly, the spatial relationships between features in the map (orientation, inclusion, separation, and intersection, etc.), which can be expressed in the map utilizing different style symbols, colors, etc., make the spatial relationships between features clearer, the content more hierarchical, and the elements such as points, lines, and surfaces more readable. Finally, maps generated using aerial imagery can inherit the high timeliness. The high timeliness enables fast task execution and meets emergency response needs so that users can access the generated map products as soon as possible and understand the changes in spatial information, thus meeting personalized needs. Examples include law enforcement agencies’ supervision of illegal structures and visualization tools for emergency response departments in the face of natural disasters. This will promote more personalized and intelligent map mapping applications in the future, prompting the development of the map mapping field in the direction of artificial intelligence.

6. Conclusions

This paper proposes SAM-GAN, a model for aerial image-to-map translation. We show that SAM-GAN produces better quality maps in terms of IS, FID, SSIM, and other evaluation metrics and that SAM-GAN, with the help of a style encoder, learns the style of the target domain well enough to guide the decoder to produce high-quality converted images. In addition, we use the attention mechanism, VGG loss, and topological consistency loss to improve the model’s results at the visual level and speed up the model’s convergence, helping it pay better attention to details.
In conclusion, SAM-GAN achieves better results in both objective and subjective evaluations compared to previous baseline models used for aerial image translation, which illustrates the effectiveness of this model in aerial image-to-map translation work, and also provides new thinking for future map mapping for areas such as personalized and intelligent map use.
We plan to expand the dataset in the next step by setting up more styles and resolutions to test the model. We will also improve the model’s accuracy in the translation process and explore image translation in multiple domains.

Author Contributions

Conceptualization, methodology, formal analysis, investigation, and writing–original draft preparation, Jian Xu; supervision, project management, funding acquisition and resources, Hongwei Li; data curation, Xiaowen Zhou; validation and visualization, Chaolin Han and Bing Dong; writing—review and editing, Jian Xu and Hongwei Li. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported by the key project of National Natural Science Foundation of China—“Machine Map Theory and Modeling Method”, whose support number is 42130112, and “Theory and Method of Map and Spatial Cognition under Human-Machine-Environment Collaboration” of High-level Talents Research Project of Zhengzhou University, whose support number is 135-32310276.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset used in this experiment is from the open-source dataset of the pix2pix model.

Acknowledgments

The author would like to thank the model and dataset contributed by the open source. At the same time, we thank the editors and reviewers for their valuable comments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liao, K. Retrospect and Prospect of the Development of Chinese Cartography. J. Geod. Geoinf. Sci. 2017, 46, 1517–1525. [Google Scholar] [CrossRef]
  2. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Nets. In Proceedings of the 27th International Conference on Neural Information Processing Systems, Palais des Congrès de Montréal, Montreal, QC, Canada, 8–13 December 2014. [Google Scholar]
  3. Zhang, X.F.; Wu, G. Data Augmentation Method Based on Generative Adversarial Network. Comput. Syst. Appl. 2019, 28, 201–206. [Google Scholar]
  4. Chen, W.; Guan, Z.; Chen, Y. Data augmentation method based on conditional generative adversarial net model. J. Comput. Appl. 2018, 38, 3305–3311. [Google Scholar]
  5. Lu, C.W.; Sun, Q.; Zhao, Y.P.; Sun, S.J.; Ma, J.Z.; Cheng, M.M.; Li, Y.F. A Road Extraction Method Based on Conditional Generative Adversarial Nets. Geomat. Inf. Sci. Wuhan Univ. 2021, 46, 807–815. [Google Scholar] [CrossRef]
  6. Li, C.; Zhang, Y.; Huang, C.H. Improved super-resolution reconstruction of image based on generative adversarial networks. Comput. Eng. Appl. 2020, 56, 191–196. [Google Scholar]
  7. Min, R.; Yang, X.Z.; Dong, Z.Y.; Chen, J. SAR Image Super-Resolution Reconstruction Based on Enhanced Structural Generative Adversarial Network. Geogr. Geo-Inf. Sci. 2021, 37, 47–53. [Google Scholar]
  8. Wang, K.F.; Gou, C.; Duan, Y.J.; Lin, Y.L.; Zheng, X.H.; Wang, F.Y. Generative Adversarial Networks: The State of the Art and Beyond. Acta Autom. Sin. 2017, 43, 321–332. [Google Scholar] [CrossRef]
  9. Isola, P.; Zhu, J.Y.; Zhou, T.H.; Efros, A.A. Image-to-Image Translation with Conditional Adversarial Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–27 July 2017; pp. 1125–1134. [Google Scholar] [CrossRef] [Green Version]
  10. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar] [CrossRef] [Green Version]
  11. Yi, Z.L.; Zhang, H.; Tan, P.; Gong, M.L. DualGAN: Unsupervised Dual Learning for Image-to-Image Translation. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2849–2857. [Google Scholar] [CrossRef] [Green Version]
  12. Kim, T.; Cha, M.; Kim, H.; Lee, J.K.; Kim, J. Learning to Discover Cross-Domain Relations with Generative Adversarial Networks. arXiv 2017, arXiv:1703.05192. [Google Scholar]
  13. Choi, Y.; Uh, Y.; Yoo, J.; Ha, J.W. Stargan v2: Diverse image synthesis for multiple domains. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 8188–8197. [Google Scholar]
  14. Liu, M.Y.; Breuel, T.; Kautz, J. Unsupervised Image-to-Image Translation Networks. arXiv 2017, arXiv:1703.00848. [Google Scholar]
  15. Huang, X.; Liu, M.Y.; Belongie, S.; Kautz, J. Multimodal Unsupervised Image-to-Image Translation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 179–196. [Google Scholar] [CrossRef] [Green Version]
  16. Liu, M.Y.; Huang, X.; Mallya, A.; Karras, T.; Aila, T.; Lehtinen, J.; Kautz, J. Few-Shot Unsupervised Image-to-Image Translation. In Proceedings of the IEEE/CVF international conference on computer vision Seoul, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 10550–10559. [Google Scholar] [CrossRef] [Green Version]
  17. Bahdanau, D.; Cho, K.; Bengio, Y. Neural machine translation by jointly learning to align and translate. arXiv 2014, arXiv:1409.0473. [Google Scholar]
  18. Luong, M.-T.; Pham, H.; Manning, C.D. Effective approaches to attention-based neural machine translation. arXiv 2015, arXiv:1508.04025. [Google Scholar]
  19. Zhang, H.; Goodfellow, I.; Metaxas, D.; Odena, A. Self-Attention Generative Adversarial Networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 10–15 June 2019; pp. 7354–7363. [Google Scholar]
  20. Emami, H.; Aliabadi, M.M.; Dong, M.; Chinnam, R.B. SPA-GAN: Spatial Attention GAN for Image-to-Image Translation. IEEE T Multimed. 2021, 23, 391–401. [Google Scholar] [CrossRef] [Green Version]
  21. Chen, C.-F.R.; Fan, Q.; Panda, R. Crossvit: Cross-attention multi-scale vision transformer for image classification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 357–366. [Google Scholar]
  22. Wang, F.; Jiang, M.; Qian, C.; Yang, S.; Li, C.; Zhang, H.; Wang, X.; Tang, X. Residual attention network for image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3156–3164. [Google Scholar]
  23. Xiao, T.; Xu, Y.; Yang, K.; Zhang, J.; Peng, Y.; Zhang, Z. The application of two-level attention models in deep convolutional neural network for fine-grained image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 842–850. [Google Scholar]
  24. Mnih, V.; Heess, N.; Graves, A.; Kavukcuoglu, K. Recurrent Models of Visual Attention. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; p. 27. [Google Scholar]
  25. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; p. 30. [Google Scholar]
  26. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar] [CrossRef] [Green Version]
  27. Woo, S.H.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar] [CrossRef] [Green Version]
  28. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 11534–11542. [Google Scholar] [CrossRef]
  29. Tang, H.; Liu, H.; Xu, D.; Torr, P.H.S.; Sebe, N. AttentionGAN: Unpaired Image-to-Image Translation Using Attention-Guided Generative Adversarial Networks. IEEE Trans. Neural. Netw. Learn Syst. 2021; Online ahead of print. [Google Scholar] [CrossRef]
  30. Gu, J.; Zhang, Y.; Zhang, W.K.; Yu, H.F.; Wang, S.Y.; Wang, Y.L.; Wang, L. Aerial Image and Map Synthesis Using Generative Adversarial Networks. In Proceedings of the IGARSS 2019-2019 IEEE International Geoscience and Remote Sensing Symposium, Yokohama, Japan, 28 July–2 August 2019; pp. 9803–9806. [Google Scholar]
  31. Wang, T.-C.; Liu, M.-Y.; Zhu, J.-Y.; Tao, A.; Kautz, J.; Catanzaro, B. High-resolution image synthesis and semantic manipulation with conditional gans. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 8798–8807. [Google Scholar]
  32. Ganguli, S.; Garzon, P.; Glaser, N. GeoGAN: A conditional GAN with reconstruction and style loss to generate standard layer of maps from satellite images. arXiv 2019, arXiv:1902.05611. [Google Scholar]
  33. Mirza, M.; Osindero, S. Conditional generative adversarial nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
  34. Zhang, Y.; Yin, Y.F.; Zimmermann, R.; Wang, G.F.; Varadarajan, J.; Ng, S.K. An Enhanced GAN Model for Automatic Satellite-to-Map Image Conversion. IEEE Access 2020, 8, 176704–176716. [Google Scholar] [CrossRef]
  35. Li, J.T.; Chen, Z.L.; Zhao, X.Z.; Shao, L.J. MapGAN: An Intelligent Generation Model for Network Tile Maps. Sensors 2020, 20, 3119. [Google Scholar] [CrossRef]
  36. Huang, X.; Belongie, S. Arbitrary Style Transfer in Real-Time with Adaptive Instance Normalization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1510–1519. [Google Scholar] [CrossRef] [Green Version]
  37. Johnson, J.; Alahi, A.; Li, F.F. Perceptual Losses for Real-Time Style Transfer and Super-Resolution. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 8–16 October 2016; pp. 694–711. [Google Scholar] [CrossRef] [Green Version]
  38. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  39. Zhao, H.; Gallo, O.; Frosio, I.; Kautz, J. Loss Functions for Image Restoration with Neural Networks. IEEE Trans. Comput. Imaging 2016, 3, 47–57. [Google Scholar] [CrossRef]
  40. Chen, X.; Chen, S.Q.; Xu, T.; Yin, B.G.; Peng, J.; Mei, X.M.; Li, H.F. SMAPGAN: Generative Adversarial Network-Based Semisupervised Styled Map Tile Generation Method. IEEE Trans. Geosci. Remote. Sens. 2021, 59, 4388–4406. [Google Scholar] [CrossRef]
  41. Gatys, L.A.; Ecker, A.S.; Bethge, M. A neural algorithm of artistic style. arXiv 2015, arXiv:1508.06576. [Google Scholar] [CrossRef]
  42. Gatys, L.A.; Ecker, A.S.; Bethge, M. Image Style Transfer Using Convolutional Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2414–2423. [Google Scholar] [CrossRef]
  43. Xu, Q.; Huang, G.; Yuan, Y.; Guo, C.; Sun, Y.; Wu, F.; Weinberger, K. An empirical study on evaluation metrics of generative adversarial networks. arXiv 2018, arXiv:1806.07755. [Google Scholar]
  44. Hensel, M.; Ramsauer, H.; Unterthiner, T.; Nessler, B.; Hochreiter, S. GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium. arXiv 2017, arXiv:1706.08500. [Google Scholar]
  45. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
  46. Fu, H.; Gong, M.M.; Wang, C.H.; Batmanghelich, K.; Zhang, K.; Tao, D.C. Geometry-Consistent Generative Adversarial Networks for One-Sided Unsupervised Domain Mapping. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2427–2436. [Google Scholar] [CrossRef]
  47. Song, J.Q.; Li, J.; Chen, H.; Wu, J. RSMT: A Remote Sensing Image-to-Map Translation Model via Adversarial Deep Transfer Learning. Remote Sens. 2022, 14, 919. [Google Scholar] [CrossRef]
Figure 1. The core of the SAM-GAN model. x1 represents the source domain (aerial image), and x2 represents the target domain (map). C1 represents the content space of the aerial image, and S2 represents the style space of the map. Different types of encoders can separate the content space or style space from the domain, while decoders must combine the content space and style space.
Figure 1. The core of the SAM-GAN model. x1 represents the source domain (aerial image), and x2 represents the target domain (map). C1 represents the content space of the aerial image, and S2 represents the style space of the map. Different types of encoders can separate the content space or style space from the domain, while decoders must combine the content space and style space.
Ijgi 12 00159 g001
Figure 2. Detailed architecture of the SAM-GAN model, where PCR represents the combination of Padding, Convolution, and ReLU activation functions; SeBlk represents the attention module, PCIR represents the combination of Padding, Convolution, Instance Normalization, and ReLU activation functions; FC represents the fully connected layer; CTIR represents the Transpose Convolution, Instance Normalization, and ReLU combination; ZS denotes the style potential coding vector, and ZC denotes the content potential coding feature map.
Figure 2. Detailed architecture of the SAM-GAN model, where PCR represents the combination of Padding, Convolution, and ReLU activation functions; SeBlk represents the attention module, PCIR represents the combination of Padding, Convolution, Instance Normalization, and ReLU activation functions; FC represents the fully connected layer; CTIR represents the Transpose Convolution, Instance Normalization, and ReLU combination; ZS denotes the style potential coding vector, and ZC denotes the content potential coding feature map.
Ijgi 12 00159 g002
Figure 3. Results of SAM-GAN and each baseline model in terms of aerial image-to-map translation performance, from left to right, aerial images, real maps, results of AttentionGAN, results of CycleGAN, results of Pix2pix, results of SAM-GAN, and results of UNIT.
Figure 3. Results of SAM-GAN and each baseline model in terms of aerial image-to-map translation performance, from left to right, aerial images, real maps, results of AttentionGAN, results of CycleGAN, results of Pix2pix, results of SAM-GAN, and results of UNIT.
Ijgi 12 00159 g003
Figure 4. Trend plots of IS and SSIM scores for the SAM-GAN-K (K = 1, 3, 5, 8, 10) model on the test set.
Figure 4. Trend plots of IS and SSIM scores for the SAM-GAN-K (K = 1, 3, 5, 8, 10) model on the test set.
Ijgi 12 00159 g004
Figure 5. Trend of FID and PixelACC(%) scores for the SAM-GAN-K (K = 1, 3, 5, 8, 10) model on the test set.
Figure 5. Trend of FID and PixelACC(%) scores for the SAM-GAN-K (K = 1, 3, 5, 8, 10) model on the test set.
Ijgi 12 00159 g005
Figure 6. Qualitative comparison between aerial imagery and the map transformation model SAM-GAN in terms of ablation experiments. From left to right: aerial images, real maps, results of SAM-GAN, results of SAM-GAN-no-StyleEncoder, results of SAM-GAN-no-TOPLoss, results of SAM-GAN-no-VGGLoss, and results of SAM-GAN-no-SeBlock.
Figure 6. Qualitative comparison between aerial imagery and the map transformation model SAM-GAN in terms of ablation experiments. From left to right: aerial images, real maps, results of SAM-GAN, results of SAM-GAN-no-StyleEncoder, results of SAM-GAN-no-TOPLoss, results of SAM-GAN-no-VGGLoss, and results of SAM-GAN-no-SeBlock.
Ijgi 12 00159 g006
Table 1. Quantitative results of the baseline model and SAM-GAN (K = 1) on the performance of aerial image-to-map translation for different evaluation metrics, where δ is taken as 5. (Bold: performance first; underlined: performance second).
Table 1. Quantitative results of the baseline model and SAM-GAN (K = 1) on the performance of aerial image-to-map translation for different evaluation metrics, where δ is taken as 5. (Bold: performance first; underlined: performance second).
ModelsISFIDSSIMPixelACC (%)
AttentionGAN2.945482.00070.195830.8973
CycleGAN2.6103211.18720.299324.2836
Pix2pix3.0706136.03180.399027.7821
SAM-GAN(K = 1)3.755765.73820.434333.3638
UNIT3.0576149.95810.325223.9213
Table 2. Training time required for the baseline and SAM-GAN models (K = 1). (Bold: performance first; underlined: performance second).
Table 2. Training time required for the baseline and SAM-GAN models (K = 1). (Bold: performance first; underlined: performance second).
AttentionGANCycleGANPix2pixSAM-GANUNIT
Time (h)9.567.733.331.2130.27
Table 3. Qualitative comparison results between the baseline model and SAM-GAN on the performance of aerial image-to-map translation. (Bold: performance first; underlined: performance second).
Table 3. Qualitative comparison results between the baseline model and SAM-GAN on the performance of aerial image-to-map translation. (Bold: performance first; underlined: performance second).
ModelsContent RetentionStyle SimilarityMap Availability
AttentionGAN6.656.686.45
CycleGAN5.635.605.46
Pix2pix6.246.456.32
SAM-GAN(K=1)7.948.057.77
UNIT4.834.894.73
Table 4. Quantitative assessment scores of the SAM-GAN model for different K values where δ is taken as 5. (Bold: first in performance; underlined: second in performance).
Table 4. Quantitative assessment scores of the SAM-GAN model for different K values where δ is taken as 5. (Bold: first in performance; underlined: second in performance).
ModelsISFIDSSIMPixelACC(%)
SAM-GAN (K = 1)3.755765.73820.434332.5677
SAM-GAN (K = 3)3.673771.83030.434032.5300
SAM-GAN (K = 5)3.702171.15990.435332.5526
SAM-GAN (K = 7)3.539471.39840.431132.4577
SAM-GAN (K = 10)3.621771.44940.429932.5048
Table 5. Time required to train the SAM-GAN model for different K values.
Table 5. Time required to train the SAM-GAN model for different K values.
K = 1K = 3K = 5K = 8K = 10
Time(h)1.213.033.384.084.65
Table 6. Quantitative assessment scores for SAM-GAN ablation experiments, where δ is taken as 1. (Bold: performance first; underlined: performance second).
Table 6. Quantitative assessment scores for SAM-GAN ablation experiments, where δ is taken as 1. (Bold: performance first; underlined: performance second).
MetricsISFIDSSIMPixelACC (%)
SAM-GAN-no-StyleEncoder3.191497.21840.39270.2862
SAM-GAN-no-VGGLoss3.495681.72140.41580.3237
SAM-GAN-no-TOPLoss3.601370.12930.42880.3236
SAM-GAN-no-SeBlock3.701971.45340.42980.3256
SAM-GAN3.755765.73820.43430.3257
Table 7. Quantitative assessment scores for SAM-GAN ablation experiments, where δ is taken as 1. (Bold: performance first; underlined: performance second).
Table 7. Quantitative assessment scores for SAM-GAN ablation experiments, where δ is taken as 1. (Bold: performance first; underlined: performance second).
SAM-GAN-No-StyleEncoderSAM-GAN-
No-VGGLoss
SAM-GAN-
No-TOPLoss
SAM-GAN-
No-SeBlock
SAM-GAN
Time(h)3.523.683.312.371.21
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, J.; Zhou, X.; Han, C.; Dong, B.; Li, H. SAM-GAN: Supervised Learning-Based Aerial Image-to-Map Translation via Generative Adversarial Networks. ISPRS Int. J. Geo-Inf. 2023, 12, 159. https://doi.org/10.3390/ijgi12040159

AMA Style

Xu J, Zhou X, Han C, Dong B, Li H. SAM-GAN: Supervised Learning-Based Aerial Image-to-Map Translation via Generative Adversarial Networks. ISPRS International Journal of Geo-Information. 2023; 12(4):159. https://doi.org/10.3390/ijgi12040159

Chicago/Turabian Style

Xu, Jian, Xiaowen Zhou, Chaolin Han, Bing Dong, and Hongwei Li. 2023. "SAM-GAN: Supervised Learning-Based Aerial Image-to-Map Translation via Generative Adversarial Networks" ISPRS International Journal of Geo-Information 12, no. 4: 159. https://doi.org/10.3390/ijgi12040159

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop