Next Article in Journal
Fabrication of Low-Twist and High-Strength Metallic Fibre Hybrid Spun Yarns
Next Article in Special Issue
A Comprehensive Review of Computer Vision in Sports: Open Issues, Future Trends and Research Directions
Previous Article in Journal
Air Flow Study around Isolated Cubical Building in the City of Athens under Various Climate Conditions
Previous Article in Special Issue
Automatic Understanding and Mapping of Regions in Cities Using Google Street View Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bootstrapped SSL CycleGAN for Asymmetric Domain Transfer

1
Faculty of Technical Sciences, University of Novi Sad, Trg D. Obradovića 6, 21000 Novi Sad, Serbia
2
Institute of Mathematics, Serbian Academy of Sciences and Arts, Kneza Mihaila 36, 11000 Belgrade, Serbia
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(7), 3411; https://doi.org/10.3390/app12073411
Submission received: 18 February 2022 / Revised: 19 March 2022 / Accepted: 22 March 2022 / Published: 27 March 2022
(This article belongs to the Special Issue Computer Vision-Based Intelligent Systems: Challenges and Approaches)

Abstract

:
Most CycleGAN domain transfer architectures require a large amount of data belonging to domains on which the domain transfer task is to be applied. Nevertheless, in many real-world applications one of the domains is reduced, i.e., scarce. This means that it has much less training data available in comparison to the other domain, which is fully observable. In order to tackle the problem of using CycleGAN framework in such unfavorable application scenarios, we propose and invoke a novel Bootstrapped SSL CycleGAN architecture (BTS-SSL), where the mentioned problem is overcome using two strategies. Firstly, by using a relatively small percentage of available labelled training data from the reduced or scarce domain and a Semi-Supervised Learning (SSL) approach, we prevent overfitting of the discriminator belonging to the reduced domain, which would otherwise occur during initial training iterations due to the small amount of available training data in the scarce domain. Secondly, after initial learning guided by the described SSL strategy, additional bootstrapping (BTS) of the reduced data domain is performed by inserting artifically generated training examples into the training poll of the data discriminator belonging to the scarce domain. Bootstrapped samples are generated by the already trained neural network that performs transferring from the fully observable to the scarce domain. The described procedure is periodically repeated during the training process several times and results in significantly improved performance of the final model in comparison to the original unsupervised CycleGAN approach. The same also holds in comparison to the solutions that are exclusively based either on the described SSL, or on the bootstrapping strategy, i.e., when these are applied separately. Moreover, in the considered scarce scenarios it also shows competitive results in comparison to the fully supervised solution based on the pix2pix method. In that sense, it is directly applicable to many domain transfer tasks that are relying on the CycleGAN architecture.

1. Introduction

The task of image-to-image domain translation (or much broader domain transfer task) is to transfer or translate images (i.e., to learn the mapping) from one domain into another, while preserving the content, where the term domain denotes a particular style (if images are for example drawings), season (like winter or summer, if images are photographs), etc. This can be achieved by using sets of paired training samples (images), which is referred as the supervised domain adaption, or by using unpaired training samples, which is also known as the unsupervised domain adaption. Domain adaption refers to a similar task.
Transfer of images from one source domain into another target domain found its application in various image processing, computer graphics, computer vision and related problems. For example, in semantic image synthesis [1,2,3,4], image-to-image translation [5,6,7,8], image inpainting [9,10], image super-resolution [11,12], etc.
In the case of supervised approaches that are mostly based on the conditional Generative Adversarial Networks (GANs, or cGANs), and which were developed first, all images are assumed to be available in pairs corresponding to the samples denoting the same thing in the source and the target domain (i.e., in the “first” and the “second” domain). Isola et al. [13] have first applied cGAN in the domain translation task, by proposing supervised pix2pix learning strategy. This approach was further developed by [14,15,16], but those solutions still failed to capture the complex structural relationships of the scenes in the cases when, e.g., two domains have drastically different views, i.e., while trying to achieve mapping through a single translation network.
Nevertheless, despite the efforts to improve the supervised methods, for many domain transfer tasks paired training images are not fully available. Therefore, besides the described approaches, almost at the same time an unsupervised domain transfer method known as CycleGAN was proposed in [17]. It utilizes two GANs oriented in opposite directions, i.e., from one domain into another, and vice versa. As introduced mappings that are learned are highly under-constrained, cycle consistency loss was introduced in [17] to reduce the mentioned anomaly.
Cycle consistency constraint proved to be very effective as a technique for preserving semantic information of the data with respect to a task of domain transfer, and is therefore applied in various tasks, e.g., in image-to-image translation [17], emotion style transfer [18], speech enhancement [19], etc. However, many of these problems, such as emotion style transfer, speech enhancement, speaker domain transfer, medical image domain transfer, have inherent domain asymmetry, in the sense that one of the domains has significantly less training data available. Therefore, such domain is usually named as reduced or scarce in comparison to another one, which is considered as fully observable. According to the authors’ knowledge, only several approaches tackled the mentioned problem in the literature, i.e., analyzed domain transfer in cases when one of the two domains is scarce. In [20], an augmented cyclic adversarial learning model that enforces the cycle-consistency constraint via an external task-specific model is proposed. Additionally, in [21], the authors add semi-supervised task-specific attention modules to generate images that are used for improving the performance in a medical image classification task. However, in all those papers, a task-specific approach is used in forming additional cost functions, which means that the analysis of each specific task is required for problem formulation, as well during the training procedure.
In this work, we propose a different approach. We combine the original unsupervised CycleGAN architecture with a semi-supervised learning (SSL) strategy, as well as with a bootstrapping (BTS) procedure. SSL is used to prevent overfitting of the CycleGAN discriminator that is belonging (or being related) to the assumed scarce domain in the considered domain transfer mapping. Discriminator overfitting could occur during the initial learning iterations of the unsupervised CycleGAN due to insufficient number of training samples in the scarce domain, and consequently lead to an underperforming GAN. On the other hand, the second part of the proposed method, i.e., the bootstrapping strategy (BTS), has the aim to bootstrapp the statistics of the problematic discriminator belonging to the scarce domain (i.e., its training pool statistics). This is accomplished by inserting (bootstrapping) artificially generated example data into the discriminator’s training pool from time to time during the training procedure. Namely, the corresponding generator residing in the second, fully observable domain of the CycleGAN is periodically called during the training to randomly produce the necessary samples in the scarce domain, and thus compensate the fact that the original training pool of the discriminator in the scarce domain was inadequate. This sampling procedure is applied periodically during the training, but for the first time only after the necessary generator is already sufficiently well trained, i.e., after the initial SSL based on a small amount of paired samples from both domains is performed. Elements of the described BTS-SSL method, i.e., combined SSL and BTS procedure (“SSL+BTS”) that is applied during the proposed SSL of CycleGAN in the considered asymmetric domain size scenarios, are illustrated in Figure 1.
The initial SSL strategy that is based on a small set of paired observations allows discriminator as well the generator that are residing in the scarce domain to avoid overfitting and learn necessary parameters to some extent. In the latter stages of the training procedure the learning process is further guided by the combined SSL as well as by the usual adversarial training strategy, but with an additional augmentation of the discriminator’s training pool in the scarce domain, from time to time. This is illustrated in the right part of Figure 1, where it is highlighted that the training involves three types of data, as compared to only one in the baseline unsupervised case. Actually, during the training, after a number of initial iterations, the pool corresponding to the discriminator of the scarce domain is constantly (periodically) filed by training examples generated by the network F in Figure 1, which is transferring the training examples from the full to the scarce domain.
Thus, contrary to the approach proposed in [20], where the domain attention mechanism is utilized, the proposed BTS-SSL model is independent of the specific domain transfer task, and requires only a relatively small amount of paired examples in the very beginning of the training procedure. This property makes it directly applicable to different tasks.
The paper is organized as follows. Firstly, in Section 2 we give the preliminaries on GANs as well as CycleGAN training strategies, on which we further build the proposed method. Secondly, in Section 3 we introduce the combined SSL and a novel bootstrapping strategy (BTS) in the context of CycleGAN architecture, in order to deal with imbalanced domain transfer tasks that are characterized by one of the two domains being scarce. In Section 4 we describe the elements of the adopted network architectures, training details and experimental setup. Finally, in Section 5 results of extensive experiments involving different training strategies, domain transfer tasks, and varying level of scarcity are presented and discussed. At the end, based on gathered insights, final conclusions are drawn in Section 6.

2. Preliminaries

In this section, we give an overview of methods and algorithms that we relate to and further adapt in the paper.

2.1. Generative Adversarial Networks

In [22], the ground-breaking method of learning the true data distribution in a nonparametric way was proposed in the form of Generative Adversarial Networks (GANs). In that constellation, the discriminator network D is trying to discriminate between the synthetic samples artificially generated by the generator network G and the ground truth observations available in the training data. The generator G models the true data distribution by learning to confuse, i.e., deceive the discriminator D by providing D with synthetic examples that are hardly distinguishable from the real ones. Thus, the two structures, the discriminator D and the generator G, are mutually competing in order to reach the Nash equilibrium expressed by the minimax loss of the training procedure, where the optimization problem is given by:
min G max D E x p ( x ) ln D ( x ) + E z p ( z ) ln 1 D G ( z ) ,
where p ( x ) represents the true data distribution over domain X , while the latent variable z is sampled by the distribution p ( z ) .
The first term in Equation (1), ln D ( x ) , corresponds to the cross-entropy between true distribution [ 1 0 ] T and distribution [ D ( x ) 1 D ( x ) ] T corresponding to discriminator’s probabilistic output. Similarly, the second term: ln 1 D G ( z ) , is the cross-entropy between [ 1 0 ] T and [ D G ( z ) 1 D G ( z ) ] T .
In Figure 1, generators in both directions are assumed to be in form of GANs, and are denoted by mappings F and G, while the corresponding discriminators over domains X and Y are denoted by D X and D Y .

2.2. CycleGAN Networks

The CycleGAN Network is designed to capture special characteristics of one image collection and to figure out how these could be translated into another image collection in the absence of any supervisor, i.e., paired training examples, as finding those is difficult, as well as expensive in the sense of labelling effort that has to be employed simultaneously over two different domains.
In [17], Zhu et al. have proposed invoking the “cycle consistency” loss in the overall loss function, by designing two domain translators or mappers G and F, in the mutually opposite domain-wise directions. Namely, by learning:
G : X Y ,
F : Y X ,
in the form of GAN networks, and by encouraging both mappings given by Equations (2) and () and illustrated in Figure 1, to be as “close” as possible to bijection, i.e., by making:
F G ( x ) x ,
G F ( y ) y ,
it was shown to be possible to compensate the lack of paired data samples, and effectively learn the complex nonlinear transformations in a fully unsupervised manner.
Therefore, if we denote the generator networks implementing mappings G and F by G X Y and G Y X , respectively, and denote the discriminator networks corresponding to domains X and Y , i.e., true distributions p X ( x ) and p Y ( y ) , by  D X and D Y , respectively, from Equation (1) we get the following adversarial optimization objectives:
L a d v ( G X Y , D Y ) = E y p Y ( y ) ln D Y ( y ) + E x p X ( x ) ln ( 1 D Y ( G X Y ( x ) ) )
L a d v ( G Y X , D X ) = E x p X ( x ) ln D X ( x ) + E y p Y ( y ) ln ( 1 D X ( G Y X ( y ) ) )
as well as the cycle-consistency objective:
L c y c ( G X Y , G Y X ) = E x p X ( x ) G Y X ( G X Y ( x ) ) x
+ E y p Y ( y ) G X Y ( G Y X ( y ) ) y
making the full optimization objective given by:
L ( G X Y , G Y X , D X , D Y ) = L a d v ( G X Y , D Y ) + L a d v ( G Y X , D X ) + λ c y c L c y c ( G X Y , G Y X )
the one being optimized during the unsupervised training of the model shown in the left part of Figure 1. Effect of the enforced cycle-consistency during the training (learning) is regulated by the penalty parameter λ c y c .

2.3. Unsupervised Training Procedure

In the following lines of Algorithm 1 are formally summarized the main steps of the unsupervised training procedure that was discussed in more details in Section 2.2. The algorithm implements an iterative optimization of the objective function given by Equation (10).
The presented steps are the basis for the proposed semi-supervised extension of the original CycleGAN algorithm, presented in Section 3.
Algorithm 1 CycleGAN training procedure
  • procedureCycleGAN(process in the left part of Figure 1)
  •     N, number of iterations; m, minibatch size; η > 0 , learning rate; X X , Y Y , unpaired or unlabeld training sets, such that | X | | Y | ;
  •     Randomly initialize parameters of discriminators D X , D Y , and generators G X Y , G Y X : θ D X , θ D Y , θ G X Y , θ G Y X
  •     for  k = 1 to N do
  •         Sample minibatch of unpaired training data { x 1 , , x m } X , { y 1 , , y m } Y
  •          L ^ a d v ( G X Y , D Y ) = 1 m i = 1 m ln D Y ( y i ) + 1 m i = 1 m ln 1 D Y ( G X Y ( x i ) )
  •          L ^ a d v ( G Y X , D X ) = 1 m i = 1 m ln D X ( x i ) + 1 m i = 1 m ln 1 D X ( G Y X ( y i ) )
  •          θ D X ( k + 1 ) θ D X ( k ) η θ D X L ^ a d v ( G Y X , D X )
  •          θ D Y ( k + 1 ) θ D Y ( k ) η θ D Y L ^ a d v ( G X Y , D Y )
  •          θ G Y X ( k + 1 ) θ G Y X ( k ) η θ G Y X L ^ a d v ( G Y X , D X ) + λ c y c L c y c ( G X Y , G Y X )
  •          θ G X Y ( k + 1 ) θ G X Y ( k ) η θ G X Y L ^ a d v ( G X Y , D Y ) + λ c y c L c y c ( G X Y , G Y X )
  •     end for
  • end procedure

3. Proposed Approach

In order to deal with the mutually imbalanced domains, i.e., a problem where one of the domains involved in the translation task is scarce, we utilize an approach with two additional concepts, which are added to the standard CycleGAN model. Nevertheless, as in the original CycleGAN architecture, the proposed solution remains task-independent and avoids introduction of any task-dependent optimization constraints. The general idea of the method is depicted by the functional diagram in the right part of Figure 1.

3.1. Avoiding the Overfitting by SSL Strategy

Firstly, we employ the semi supervised learning (SSL) strategy based on the simple · 1 norm error term, which is computed over all pairs or elements in the provided small set of labeled or paired training samples ( x i p , y i p ) that represent the same thing in the involved scarce X and a fully observable domain Y , respectively.
Thus, despite an unfavourable situation in which the number of unlabeled training samples in domain X is considered as significantly lower than the number of available unlabeled observations in the domain Y , i.e.,  X Y , it is assumed that there still exists a small amount of labeled training pairs ( x i p , y i p ) that are possible to be reliably matched into pairs based on their ground-truth labels, and thus provide the basis for the SSL strategy.
Based on the valuable information contained in this small set of M S S L paired observations, i.e.,  P d a t a = { ( x i p , y i p ) | i = 1 , , M S S L } , it is possible to define an additional SSL optimization term L S S L , which will enforce that the learned CycleGAN mappings in Equations (2) and (3) respect the fact that x i p X and y i p Y are paired, i.e., have the same label. This is illustrated in Figure 2, where the expected closeness of the samples x p and y p after transformation of x p by the learned mapping G is depicted.
As the result of previous considerations, the closeness of the “transformed” G ( x i p ) and the “true” labeled samples y i p during the SSL phase of the proposed training procedure can be described by the following optimization objective:
L S S L ( G X Y , G Y X ) = 1 m i = 1 m G X Y ( x i p ) y i p 1 + G Y X ( y i p ) x i p 1 ,
which evaluates and imposes the corresponding closeness over a predefined amount of M S S L available paired examples in both directions, from  X into Y , as well to the opposite side, from  Y into X , Figure 1 and Figure 2.
Thus, by applying the described SSL procedure and the introduced L S S L objective, we prevent discriminator D X in the scarce domain X to overfit during the initial iterations of the learning process. In the standard unsupervised learning setting, characteristic for any CycleGAN model, overfitting of discriminator D X would mostly be inevitable, due to assumed limited number of available unlabeled training samples in the scarce domain, i.e., adopted assumption that X Y .
By combining Equation (11) with the standard CycleGAN objective presented in Equation (10), the full SSL objective is made of:
L ( G X Y , G Y X , D X , D Y ) = L a d v ( G X Y , D Y ) + L a d v ( G Y X , D X ) + λ c y c L c y c ( G X Y , G Y X ) + λ S S L L S S L ( G X Y , G Y X ) ,
where λ S S L is the corresponding regularization parameter.

3.2. Improving the Learning Process by BTS Strategy

As the second part of the proposed learning procedure, we employ the bootstrapping (BTS) strategy based on the internal structures already available in the CycleGAN model.
Since the domain X is assumed to be scarce, during the training discriminator D X at its disposal has a significantly smaller pool of unlabeled training samples in comparison to the discriminator D Y in the domain Y . Therefore, the aim of the proposed BTS strategy is to overcome this imbalance, by artificially expanding the size of the unlabeled training pool of the discriminator D X . This is illustrated in the left part of Figure 3.
This is achieved by exploiting the already learned mapping F, shown in the right part of Figure 1, i.e., the GAN implementing the generator function G Y X . Under the term “already learned”, it is assumed that in the initial phase of the training procedure, parameters of the CycleGAN model are first initially optimized for some time using the previously described SSL strategy from Section 3.1. Only after this “initial training” phase, the  G Y X structure from the CycleGAN model should be considered as reliable enough to be used as the bootstrapping data sampler for the training pool of the discriminator D X residing in the scarce domain X . Of course, the decribed BTS strategy could also be applied independently from the SSL one, but in the assumed scarce X scenarios such approach would not result in the same level of performance, as in the “SSL+BTS” case, when these two strategies are synergistically combined together. The only requirement is that semi-supervision is feasible, i.e., that there exists a small amount of paired data in set P d a t a with M S S L matched samples from both domains, as described in Section 3.1 and illustrated in Figure 2.
BTS of the D X ’s training pool, i.e., BTS of its statistics, is periodically repeated during the whole learning procedure in order to exploit the improved versions of the utilized generator G Y X each time. Furthermore, available set of paired, i.e., labeled samples P d a t a is used alongside the larger unlabeled training sets X and Y, from both X and Y domains, during the whole training procedure. This means that even after the initial SSL is already accomplished, i.e., after the proposed BTS strategy is invoked for the first time during the training, P d a t a samples are continued to be used in the SSL objective, Equation (12). In such a way, the initial constraints that are expressing the closeness between the “transformed” and the “true” samples in Figure 2, i.e., expressing the quality of generators G Y X and G X Y in some training iteration k, are also preserved in the future – after the BTS of D X statistics is periodically performed, and the training pool of D X is re-expanded again.
In Figure 1, the proposed periodical bootstrapping of the D X ’s statistics, which is achieved by G Y X GAN that is transferring examples from the full into the scarce domain, is illustrated by an additional third arrow, oriented from Y towards X and centrally positioned in the functional diagram of the SSL CycleGAN architecture, in the right part of Figure 1. The same process is also illustrated in Figure 3, where it can be seen that the X B T S denotes the set of novel unlabeled training samples that are added to D X ’s training pool. The samples are randomly generated based on the parameters of the generator G Y X that have been learned up to that moment.
After performed BTS, the training pool { X X B T S } of D X can be considered as being closer in size to the training pool Y of the discriminator D Y , Figure 3. Thus, despite the unfavourable training conditions, an improved CycleGAN performance is expected to be achieved in comparison to the standard unsupervised training, as well as in comparison to the basic SSL strategy without BTS. It should be mentioned that X B T S is replaced each time with a new set of samples when G Y X is periodically invoked for the BTS—randomly generated samples from previous BTS iterations are not accumulated in the training pool of D X .
In summary, after a number of initial iterations, when the discriminator D X , as well as the generator G Y X for the scarce domain X , are sufficiently pre-trained by using the combination of adversarial and SSL loss, Equation (12), the described BTS strategy for improving the learning process comes into play. In essence, it results in improving the statistics of p ^ X ( x ) that is describing the unlabeled data in the training poll { X X B T S } of D X . Improved statistics result in better discrimination of x, and consequently in the better performance of whole CycleGAN SSL.
Although it has already been discussed, we would like to point out once again that the “unlabeled” refers to training samples of { X X B T S } and { Y } for which the exact match in the opposite domain is not known exactly and in advance, in contrast to the paired or “labeled” training pools { X p } and { Y p } , i.e., set of paired samples P d a t a , which is used in the SSL objective L S S L in Equations (11) and (12), Figure 2. In that sense, the BTS strategy directly affects only the discriminator term D X in the cost functions in Equation (12),, i.e.,  L a d v ( G Y X , D X ) . However, the indirect effect of the increased sample size in the training pool of D X is successfully propagated to all model elements in Figure 1 during the subsequent training iterations, until the next BTS with a likely better training pool is generated again.
In practice, the proposed BTS strategy consists of a function call to G Y X in every k + K -th training iteration, for some fixed period K. Each BTS results in a fixed amount of novel samples added to X, where samples represent, e.g., images in the image translation task. Images in X B T S are generated by the GAN network G Y X ( k ) , where ( k ) denotes the state of parameters in k-th training iteration. This is done by using a uniform distribution and randomly choosing a fixed percentage of images in the training set Y, which are than transferred to scarce domain by G Y X ( k ) . Thus, in every k + K -th iteration, the discriminator D X is trained on a pool { X X B T S } , augmented in the previously described manner.

3.3. Proposed SSL+BTS Training Procedure

In the following lines are formally summarized the main steps of the introduced SSL and BTS strategies, described in Section 3.1 and Section 3.2.
Algorithm 2 begins by the order of steps that in the very beginning perform exclusively SSL strategy, during the first K 0 out of N iterations, K 0 N . Only after the SSL secures that discriminator D X is not overfitting, the BTS strategy is invoked for the first time in iteration: K 0 + 1 . Afterwards, the BTS is periodically invoked every K-th iterations, and jointly employed with the SSL strategy till the end of the procedure. 
Algorithm 2 BTS-SSL CycleGAN training
  • procedureCycleGAN SSL+BTS(process in the right part of Figure 1)
  •     N, number of iterations; K 0 , number of initial SSL iterations without BTS, K 0 N ; K, period of BTS repetition; m, minibatch size; M S S L = | P d a t a | , number of paired examples in P d a t a = { ( x i p , y i p ) | i = 1 , , M S S L } = X p × Y p ; m S S L , SSL minibatch size, m S S L < M S S L ; q ( 0 , 1 ) , fixed percentage of training data Y that are randomly chosen to produce the bootstrapping samples for the training pool of D X ; η > 0 , learning rate; X X , Y Y , unpaired or unlabeled training sets, while X p X , Y p Y are paired or labeled training subsets, such that | X | | Y | , | X | | Y | , and  | X p | = | Y p | , but  | X p | | X | ;
  •     Randomly initialize parameters of D X , D Y , and  G X Y , G Y X : θ D X , θ D Y , θ G X Y , θ G Y X
  •     for  k = 1 to N do
  •         Sample minibatch of unpaired training data { x 1 , , x m } X , { y 1 , , y m } Y
  •         Sample minibatch of paired training data { ( x 1 p , y 1 p ) , , ( x m S S L p , y m S S L p ) } X p × Y p
  •          L ^ a d v ( G X Y , D Y ) = 1 m i = 1 m ln D Y ( y i ) + 1 m i = 1 m ln 1 D Y ( G X Y ( x i ) )
  •          L ^ a d v ( G Y X , D X ) = 1 m i = 1 m ln D X ( x i ) + 1 m i = 1 m ln 1 D X ( G Y X ( y i ) )
  •          L ^ S S L = 1 m i = 1 m G X Y ( x i ) y i 1 + G Y X ( y i ) x i 1
  •          L 1 = L ^ a d v ( G Y X , D X ) + λ c y c L c y c ( G X Y , G Y X ) + λ S S L L ^ S S L ( G X Y , G Y X )
  •          L 2 = L ^ a d v ( G X Y , D Y ) + λ c y c L c y c ( G X Y , G Y X ) + λ S S L L ^ S S L ( G X Y , G Y X )
  •          θ D X ( k + 1 ) θ D X ( k ) η θ D X L ^ a d v ( G Y X , D X )
  •          θ D Y ( k + 1 ) θ D Y ( k ) η θ D Y L ^ a d v ( G X Y , D Y )
  •          θ G Y X ( k + 1 ) θ G Y X ( k ) η θ G Y X L 1
  •          θ G X Y ( k + 1 ) θ G X Y ( k ) η θ G X Y L 2
  •         if  k = = K 0 + 1 ( k > K 0 ) ( ( ( k K 0 ) mod K ) = = 0 )  then
  •            Bootstrapping the training pool X D of discriminator D X .
  •             X D { X X B T S } , X B T S = r a n d q G Y X ( k ) ( Y ) , where r a n d q ( · ) represents the operator of choosing q | Y | samples from the training pool Y Y , and transforming them by G Y X ( k ) to scarce domain X , i.e., generating X B T S .
  •         end if
  •     end for
  • end procedure
In that sense, the adopted abbreviation “SSL+BTS” in Algorithm 2 better reflects the order of appearance of SSL and BTS strategies during the learning process. However, since BTS is periodically repeated, and also represents an extension of the SSL CycleGAN setup, the method is named as BTS-SSL CycleGAN.
Algorithm 2 has all elements of standard CycleGAN training procedure given by Algorithm 1 in Section 2.3. In that sense it can be used in any task where the unsupervised learning approach based on the cycle-consistency loss has already been applied before.
As already explained in Section 3.2, the main characteristic of the presented SSL+BTS procedure is that every K iterations after k = K 0 the training pool X D of discriminator D X is updated with q | Y | synthetic samples generated by G Y X ( k ) , i.e., X D X X B T S . However, X B T S is used only for D X , and is not added to unlabeled training set X.
Improved performance of D X indirectly boosts the entire training procedure and leads to better results of the domain transfer in comparison to the baseline learning Algorithm 1, as reported by the results of the conducted experiments in Section 5.

4. Elements of the Adopted Experimental Setup

In this section, we provide a description of the elements that were used in order to obtain experimental results presented in Section 5. The experimental setup was designed in such a way to provide extensive comparison between the proposed BST-SSL CycleGAN method and the baseline CycleGAN, as well comparison against one supervised domain transfer method. In order to get better insights into the performance of the proposed SSL+BTS training procedure, experiments included different types of assessments, both quantitative and visual. Therefore, in the following we point out the main characteristics of the individual elements of the adopted experimental setup and provide details of the corresponding training procedures.

4.1. Network Architecture and Training Details

For the purpose of experimental comparisons we have implemented the corresponding neural networks based on the best practices from the literature. Since the proposed semi-supervised learning is generally applicable to any variant of the original CycleGAN architecture, we have decided to rely on the referent implementations form the papers were the corresponding methods were originally introduced.
Thus, for the main neural network architecture of the corresponding GANs we have adopted the architecture proposed in [17], where the CycyleGAN model was originally proposed), and which was also utilized in [19]. This implementation of unsupervised domain translation was also the basis for implementation of the SSL+BTS method proposed in this work. In addition, since the original CycyleGAN was introduced in the context of image-to-image translation tasks, the same type of domain transfer tasks was also chosen for the experimental comparisons in this work. One of the advantages of this type of experiments is also that they provide visual insights into the quality of obtained domain transfer results.
The main characteristics of the adopted architecture are that it contains two stride-2 convolutions, as well as several residual blocks and 2 fractionally strided convolutions with stride 1 2 . It uses 6 blocks for 128 × 128 and 9 blocks for 256 × 256 and higher-resolution type images. It also uses instance normalization as in [17,23,24].
For the discriminator network, we have relied on 70 × 70 PatchGAN in order to classify whether the image patches are real or fake, as it was more efficient than the usual full-image network.
When it comes to training details, besides the main algorithmic steps of the performed training procedures, which have already been described in detail in Section 2.2 and Section 3, some of the specific settings and values of control parameters that were used in the experiments include the following. In the training procedure, as in [17], instead of negative log likelihood loss figuring in (1), we used more stable L 2 loss approach consisting of training G to minimize E z p ( z ) [ ( D ( G ( z ) ) 1 ) 2 ] as well as training discriminator D to minimize E x p ( x ) [ ( D ( x ) 1 ) 2 ] + E z p ( z ) [ D ( G ( z ) ) ] . Furthermore, similarly as in [17], we use the history of generated images (50 of them) in order to calculate the average score. We also keep λ c y c = λ S S L = 10 , for all of our experiments. Sizes of minibatches m and m S S L are kept equal and set to m = m S S L = 50 . Random initialization of the network parameters was performed by using random samples drawn from normal distribution N ( 0 , 0.02 ) . All networks were trained using the learning rate of η = 0.0002 which is kept constant during the first 100 epochs and then linearly decays to zero during the next 100 epochs.
In experiments with all used datasets we have controlled the size of the scarce domain, by changing the percentage S X of the domain used: 25%, 50% and 100%, and performed both the unsupervised, as well as the proposed SSL approach (loss function (12)), with and without the proposed bootstrapping of the statistic of the discriminator D X from the scarce domain (variants SSL+BTS, and SSL). In all experiments, as all datasets used actually contain paired images, we use a fixed 20% of the scarce domain training paired examples for evaluation in the SSL manner. In each experiment, after initial K 0 iterations, during the next k + K -th training iteration, where K = 50 , 20% of randomly chosen examples from the target domain is transformed by G Y X and added to the pool of the discriminator D X .

4.2. Considered Domain Translation Tasks

The particular tasks, as well as datasets that are used in experiments, are the following: Semantic label↔photo task on CityScapes dataset [13,25]. The dataset consists of 2975 training images of the size 128 × 128 , as well as evaluation set for testing; Architectural labels↔photo task [13,26] on CMP Facade dataset, containing 400 training images; as well as Map↔aerial photo task on Google Maps dataset [13], containing 1096 training images of the size 256 × 256 .
All experiments were performed in an imbalanced domain scenario, where we keep the “left” domain (original images domain in all experiments) scarce, i.e., we simulate that particular situation by using only a certain percentage of the original left domain.

4.3. Considered Baseline Domain Translation Methods

We evaluate the proposed approach in comparison to well-established baseline image-to-image translation methods, where we choose the original CycleGAN method proposed in [17], as well as pix2pix image translation method proposed in [13].

4.4. Utilized Measures for Results Comparison

Evaluating the quality of synthesized images is an open and difficult problem. We use some classical measures such as Peak Signal-to-Noise Ratio (PSNR) as well as more advanced Structural Similarity Index Measure (SSIM), which is a perception-based model that considers image degradation as perceived change in structural information. It also incorporates important perceptual phenomena, including both luminance masking and contrast masking terms. The SSIM measure is much more appropriate for measuring image degradation than PSNR.
The PSNR is evaluated as P S N R = 20 l o g M A X I / M S E , where M A X I is the maximum possible pixel value of the ground truth images, while M S E is the squared Euclidean norm between the generated and ground truth images. The SSIM measure between images generated by the considered GAN algorithms and ground truth images is calculated on various windows x, y of an image, by using the following formula:
S S I M ( x , y ) = ( 2 μ x μ y + c 1 ) ( 2 σ x y + c 2 ) ( μ x 2 + μ y 2 + c 1 ) ( σ x 2 + σ y 2 + c 2 )
where μ x and μ y are average values of x and y, while σ x 2 and σ y 2 are variances and c 1 and c 2 are constants, set as reported in [27].
In addition to the mentioned classical image quality measures, we have also used some more specific measures, such as measures based on human judgments, which we call Perceptual Realistic Measure (PRM), as well as the FCN score measure proposed in [28], which is based on pre-trained semantic classifiers and measures how discriminative the generated output is. We then score the synthesized photos using the classification accuracy against the labels these photos were synthesized from.
There are various measures of FCN accuracy used in [28]: pixel accuracy evaluated as i n i i / i t i , mean accuracy evaluated as ( 1 / n c l ) i i n i i / t i and mean region intersection over union accuracy evaluated as ( 1 / n c l ) i n i i / t i + j n j i n i i . Term n i j is the number of pixels of class i predicted to belong to class j, n c l is the number of classes and t i = j n i j is the total number of pixels belonging to the class i.
Concerning the PRM measure, we have asked student participants (10 of them) at the Faculty of Technical Sciences, University of Novi Sad, to evaluate the perceptual quality of images generated by the proposed method in comparison to baseline methods. We note that for us it was much more accessible to organize our own pool of human participants, than to use some internet service such as AMT perceptual studies. In the Map↔ aerial photo experiments conducted on Google Maps, we follow the procedure reported in [17].
The participants were shown pairs of images, each pair containing one real image (i.e., the ground truth) and one synthesized, generated by the proposed as well as baseline algorithms. Next, they were asked to choose an image they thought it was real. The first 5 trials of each session were conducted for practice and a feedback was given whether the participant’s response was correct or incorrect. The remaining 15 trials were used to assess the rate at which each algorithm fooled the participants.
In addition, the relative improvement of the proposed method against the CycleGAN baseline, denoted by δ , was computed for each of the described measures separately.

5. Experimental Results

In this section, we present experiments on several real image datasets in the task of image-to-image translation, showing the efficiency of the proposed domain transformation method that we call BTS-SSL CycleGAN.

5.1. Quantitative Evaluation of the Conducted Experiments

Results of the conducted experiments are summarized in Table 1 and Table 2. The relative improvement brought by the proposed method in Table 1 is expressed by average δ PSNR and δ SSIM , while SSL and BTS denote results of the methods corresponding to individual elements of the proposed SSL-BTS learning strategy.
BTS training includes only bootstrapping of the discriminator D X , without paired samples, while SSL involves using paired samples, but without bootstrapping. As an example of the performance of supervised methods on the same tasks, results of pix2pix method are also presented.
In Table 1 the results of the PSNR, as well as SSIM measure between images generated by the considered GAN algorithms (baseline pix2pix, CycleGAN as well as the proposed BTS-SSL CycleGAN, in an image-to-image translation task, from scarce to target domain) and ground truth images, are presented, for all the dataset used.
From Table 1 it can be seen that the proposed BTS-SSL CycleGAN algorithm obtains better results on average, by both PSNR and SSIM measures, in all of the experiments, in comparison to the CycleGAN method, especially on CityScapes and CMP Facade datasets. It can also be seen that both SSL and BTS components of the proposed algorithm contribute significantly to the obtained results.
In Table 2, the PRM results for the Google Maps dataset and the Photo→ Map task, for the proposed method in comparison to the baseline methods, are given in form of the accuracy of detecting generated, i.e., false images by the participants, and expressed in %.
It can be seen that, similarly to other experiments, the proposed BTS-SSL CycleGAN outperforms the baseline CycleGAN, but as expected, does not outperform the fully supervised pix2pix method. Furthermore, both BTS as well as SSL components of the method contribute to the final result.
Results of FCN accuracy measure for the proposed BTS-SSL CycleGAN in comparison to baseline methods are provided in Table 2 for the CityScapes dataset and the Semantic label↔photo task, are presented. It could be seen that the proposed BTS-SSL CycleGAN obtains better results in comparison to the CycleGAN, but it is outperformed by the fully supervised pix2pix, as expected. It could also be seen that both BTS as well as SSL components of the algorithm contribute to the final result.

5.2. Visual Comparison between Analyzed Methods

As an example of the conducted visual comparisons, in the following are presented several figures that illustrate visual performance of the proposed BTS-SSL CycleGAN method against considered baselines.
In Figure 4, Figure 5 and Figure 6, visual examples are given for the proposed BTS-SSL CycleGAN vs. baseline algorithm comparisons, for 100 % of the scarce domain data used: Real A (scarce domain) and Real B (target domain) correspond to image pair examples. Examples are shown for Google Maps, CityScapes and Facade datasets. It can be seen that in those examples, the proposed BTS-SSL CycleGAN obtains visually more accurate results than the baseline methods.

6. Conclusions

In this work, in order to deal with imbalanced domains problem in the context of domain translation or domain transfer tasks, we have proposed a novel solution for overcoming the difficulties of the unsupervised learning process in such case. The proposed method combines the unsupervised CycleGAN architecture with SSL learning strategy that is additionally improved through an internal bootstrapping procedure (BTS). Thus, the BTS-SSL CycleGAN semi-supervised domain translation model was formed. Based on experimental results, it has exhibited potential to be a method of choice for the improvement of any CycleGAN-based domain translation task, especially in the considered imbalanced domains scenario. The proposed SSL strategy was used during the initial training stages to prevent discriminator related to the scarce domain from overfitting, while the employed BTS strategy for bootstrapping the statistics of the training pool of the discriminator was used to improve the learning performance in the unfavourable scenario of imbalanced domains. BTS was performed by inserting the example data generated by the corresponding internal generator, i.e., mapping from the fully observable into the scarce domain. We have manage to obtain significantly better results in comparison to the original CycleGAN method, and comparable with the fully supervised pix2pix method, on several image domain transfer datasets. In further work, we will focus on other more intelligent ways to bootstrap the statistics of the discriminator of the scarce domain, by inserting into its pool only a certain amount of transferred image examples that “fit” the initial statistics of the mentioned pool sufficiently well.

Author Contributions

Conceptualization, M.J., L.K., B.P. and B.B.; Data curation, B.P.; Formal analysis, L.K.; Funding acquisition, L.K. and B.P.; Investigation, M.J. and B.B.; Methodology, M.J., L.K. and B.P.; Project administration, L.K.; Software, B.P.; Supervision, L.K. and B.P.; Visualization L.K,; Writing—original draft, M.J. and L.K.; Writing—review & editing, B.B., L.K. and B.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the “Science Fund of the Republic of Serbia“, through the project grant agreement No. 6524560: “Speaker/Style Adaptation for Digital Voice Assistants Based on Image Processing Methods (AI-S-ADAPT); and by the “Serbian Ministry of Education, Science and Technological Development“ through the research project No. 451-03-68/2020-14/200156: “Innovative Scientific and Artistic Research from the Faculty of Technical Sciences Activity Domain”.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All data sources are publicly available and mentioned in the references.

Acknowledgments

This research was supported by the “Science Fund of the Republic of Serbia”, through the project grant agreement No. 6524560: “Speaker/Style Adaptation for Digital Voice Assistants Based on Image Processing Methods (AI-S-ADAPT), and by the “Serbian Ministry of Education, Science and Technological Development” through the research project No. 451-03-68/2020-14/200156: “Innovative Scientific and Artistic Research from the Faculty of Technical Sciences Activity Domain”. The authors would also like to acknowledge support by the EU’s H2020 research and innovation programme under the project grant agreement No. 872614: “Selfsustained Cross Border Customized Cyberphysical System Experiments for Capacity Building Among European Stakeholders” (Smart4All).

Conflicts of Interest

The authors declare that there is no conflict of interest.

References

  1. Park, T.; Liu, M.Y.; Wang, T.C.; Zhu, J.Y. Semantic image synthesis with spatially-adaptive normalization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2337–2346. [Google Scholar]
  2. Zhu, P.; Abdal, R.; Qin, Y.; Wonka, P. SEAN: Image synthesis with semantic region-adaptive normalization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 5104–5113. [Google Scholar]
  3. Lee, C.H.; Liu, Z.; Wu, L.; Luo, P. MaskGAN: Towards diverse and interactive facial image manipulation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 5549–5558. [Google Scholar]
  4. Tang, H.; Xu, D.; Yan, Y.; Torr, P.H.; Sebe, N. Local class-specific and global image-level generative adversarial networks for semantic-guided scene generation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 7870–7879. [Google Scholar]
  5. Yi, Z.; Zhang, H.; Tan, P.; Gong, M. DualGAN: Unsupervised dual learning for image-to-image translation. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2849–2857. [Google Scholar]
  6. Mejjati, Y.A.; Richardt, C.; Tompkin, J.; Cosker, D.; Kim, K.I. Unsupervised attention-guided image-to-image translation. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018; pp. 3697–3707. [Google Scholar]
  7. Tomei, M.; Cornia, M.; Baraldi, L.; Cucchiara, R. Art2Real: Unfolding the reality of artworks via semantically-aware image-to-image translation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5849–5859. [Google Scholar]
  8. Liu, M.Y.; Huang, X.; Mallya, A.; Karras, T.; Aila, T.; Lehtinen, J.; Kautz, J. Few-shot unsupervised image-to-image translation. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 10551–10560. [Google Scholar]
  9. Song, Y.; Yang, C.; Lin, Z.; Liu, X.; Huang, Q.; Li, H.; Kuo, C.C.J. Contextual-based image inpainting: Infer, match, and translate. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  10. Zhao, L.; Mo, Q.; Lin, S.; Wang, Z.; Zuo, Z.; Chen, H.; Xing, W.; Lu, D. UCTGAN: Diverse image inpainting based on unsupervised cross-space translation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 5741–5750. [Google Scholar]
  11. Yuan, Y.; Liu, S.; Zhang, J.; Zhang, Y.; Dong, C.; Lin, L. Unsupervised image super-resolution using cycle-in-cycle generative adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–23 June 2018; pp. 701–710. [Google Scholar]
  12. Zhang, Y.; Liu, S.; Dong, C.; Zhang, X.; Yuan, Y. Multiple cycle-in-cycle generative adversarial networks for unsupervised image super-resolution. IEEE Trans. Image Process. 2019, 29, 1101–1112. [Google Scholar]
  13. Isola, P.; Zhu, J.Y.; Zhou, T.; Efros, A.A. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1125–1134. [Google Scholar]
  14. Wang, C.; Zheng, H.; Yu, Z.; Zheng, Z.; Gu, Z.; Zheng, B. Discriminative region proposal adversarial networks for high-quality image-to-image translation. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 770–785. [Google Scholar]
  15. Wang, T.C.; Liu, M.Y.; Zhu, J.Y.; Tao, A.; Kautz, J.; Catanzaro, B. High-resolution image synthesis and semantic manipulation with conditional GANs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8798–8807. [Google Scholar]
  16. AlBahar, B.; Huang, J.B. Guided image-to-image translation with bi-directional feature transformation. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 9016–9025. [Google Scholar]
  17. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
  18. Bao, F.; Neumann, M.; Vu, N.T. CycleGAN-based emotion style transfer as data augmentation for speech emotion recognition. In Proceedings of the Conference of the International Speech Communication Association (Interspeech), Graz, Austria, 15–19 September 2019; pp. 2828–2832. [Google Scholar]
  19. Meng, Z.; Li, J.; Gong, Y.; Juang, B.H. Cycle-consistent speech enhancement. In Proceedings of the Conference of the International Speech Communication Association (Interspeech), Hyderabad, India, 2–6 September 2018; pp. 1165–1169. [Google Scholar]
  20. Hosseini-Asl, E.; Zhou, Y.; Xiong, C.; Socher, R. Robust domain adaptation by augmented cyclic adversarial learning. In Proceedings of the 31st International Conference on Neural Information Processing Systems—Interpretability and Robustness for Audio, Speech and Language Workshop, Montreal, QC, Canada, 3–8 December 2018; pp. 1–11. [Google Scholar]
  21. Qi, C.; Chen, J.; Xu, G.; Xu, Z.; Lukasiewicz, T.; Liu, Y. SAG-GAN: Semi-supervised attention-guided GANs for data augmentation on medical images. arXiv 2020, arXiv:2011.07534. [Google Scholar]
  22. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the 27th International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; pp. 2672–2680. [Google Scholar]
  23. Ulyanov, D.; Vedaldi, A.; Lempitsky, V. Instance normalization: The missing ingredient for fast stylization. arXiv 2017, arXiv:1607.08022v3. [Google Scholar]
  24. Johnson, J.; Alahi, A.; Fei-Fei, L. Perceptual losses for real-time style transfer and super-resolution. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2016; pp. 694–711. [Google Scholar]
  25. Cordts, M.; Omran, M.; Ramos, S.; Rehfeld, T.; Enzweiler, M.; Benenson, R.; Franke, U.; Roth, S.; Schiele, B. The cityscapes dataset for semantic urban scene understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3213–3223. [Google Scholar]
  26. Reed, S.; Akata, Z.; Yan, X.; Logeswaran, L.; Schiele, B.; Lee, H. Generative adversarial text to image synthesis. In Proceedings of the International Conference on Machine Learning, PMLR, New York, NY, USA, 20–22 June 2016; pp. 1060–1069. [Google Scholar]
  27. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [PubMed] [Green Version]
  28. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
Figure 1. Elements of the proposed SSL method, as compared to the original CycleGAN architecture. It is assumed that one of the domains, e.g., X is scarce, i.e., X Y , which makes the unsupervised CycleGAN training much harder. Therefore, the proposed BTS-SSL CycleGAN is invoked.
Figure 1. Elements of the proposed SSL method, as compared to the original CycleGAN architecture. It is assumed that one of the domains, e.g., X is scarce, i.e., X Y , which makes the unsupervised CycleGAN training much harder. Therefore, the proposed BTS-SSL CycleGAN is invoked.
Applsci 12 03411 g001
Figure 2. SSL loss L S S L , Equation (11), which enforces closeness of the “transformed” y ^ i p = G ( x i p ) and the “true” y i p over a small set of paired, i.e., labeled samples P d a t a = { ( x i p , y i p ) | i = 1 , , M S S L } from a scarce domain X and a fully observable Y . The same also holds in the opposite direction, for F ( y i p ) .
Figure 2. SSL loss L S S L , Equation (11), which enforces closeness of the “transformed” y ^ i p = G ( x i p ) and the “true” y i p over a small set of paired, i.e., labeled samples P d a t a = { ( x i p , y i p ) | i = 1 , , M S S L } from a scarce domain X and a fully observable Y . The same also holds in the opposite direction, for F ( y i p ) .
Applsci 12 03411 g002
Figure 3. Bootstrapping of the statistics of the discriminator D X in the scarce domain X , by adding a set of novel examples X B T S to the original pool of unlabeled training data X. Bootstrapped samples are randomly generated by G Y X .
Figure 3. Bootstrapping of the statistics of the discriminator D X in the scarce domain X , by adding a set of novel examples X B T S to the original pool of unlabeled training data X. Bootstrapped samples are randomly generated by G Y X .
Applsci 12 03411 g003
Figure 4. Visual examples of the proposed vs. baseline algorithm comparisons for 25 % of the scarce domain data used: Real A (scarce domain) and Real B (target domain) correspond to image pair examples. Examples are given for Google Maps, CityScapes and Facade datasets. SSL denotes that only SSL on the control part of pared images is included in the training of CycleGAN, BTS denotes that only bootstrapping of the statistic of the A discriminator pool is included in the training, while SSL+BTS denotes that both SSL and BTS are included.
Figure 4. Visual examples of the proposed vs. baseline algorithm comparisons for 25 % of the scarce domain data used: Real A (scarce domain) and Real B (target domain) correspond to image pair examples. Examples are given for Google Maps, CityScapes and Facade datasets. SSL denotes that only SSL on the control part of pared images is included in the training of CycleGAN, BTS denotes that only bootstrapping of the statistic of the A discriminator pool is included in the training, while SSL+BTS denotes that both SSL and BTS are included.
Applsci 12 03411 g004
Figure 5. Visual examples of the proposed vs. baseline algorithm comparisons for 50 % of the scarce domain data used: Real A (scarce domain) and Real B (target domain) correspond to image pair examples. Examples are given for Google Maps, CityScapes and Facade datasets. SSL denotes that only SSL on the control part of pared images is included in the training of CycleGAN, BTS denotes that only bootstrapping of the statistic of the A discriminator pool is included in the training, while SSL+BTS denotes that both SSL and BTS are included.
Figure 5. Visual examples of the proposed vs. baseline algorithm comparisons for 50 % of the scarce domain data used: Real A (scarce domain) and Real B (target domain) correspond to image pair examples. Examples are given for Google Maps, CityScapes and Facade datasets. SSL denotes that only SSL on the control part of pared images is included in the training of CycleGAN, BTS denotes that only bootstrapping of the statistic of the A discriminator pool is included in the training, while SSL+BTS denotes that both SSL and BTS are included.
Applsci 12 03411 g005
Figure 6. Visual examples of the proposed vs. baseline algorithm comparisons for 100 % of the scarce domain data used: Real A (scarce domain) and Real B (target domain) correspond to image pair examples. Examples are given for Google Maps, CityScapes and Facade datasets. SSL denotes that only SSL on the control part of pared images is included in the training of CycleGAN, BTS denotes that only bootstrapping of the statistic of the A discriminator pool is included in the training, while SSL+BTS denotes that both SSL and BTS are included.
Figure 6. Visual examples of the proposed vs. baseline algorithm comparisons for 100 % of the scarce domain data used: Real A (scarce domain) and Real B (target domain) correspond to image pair examples. Examples are given for Google Maps, CityScapes and Facade datasets. SSL denotes that only SSL on the control part of pared images is included in the training of CycleGAN, BTS denotes that only bootstrapping of the statistic of the A discriminator pool is included in the training, while SSL+BTS denotes that both SSL and BTS are included.
Applsci 12 03411 g006
Table 1. Experimental comparison between proposed semi-supervised (SSL+BTS) and the unsupervised CycleGAN, under different scenarios: varying sample size S X of scarce domain X , as well as when applied to different tasks/datasets—CityScapes, Facade dataset, Google Maps.
Table 1. Experimental comparison between proposed semi-supervised (SSL+BTS) and the unsupervised CycleGAN, under different scenarios: varying sample size S X of scarce domain X , as well as when applied to different tasks/datasets—CityScapes, Facade dataset, Google Maps.
S X pix2pixCycleGANSSLBTSSSL+BTS δ PSNR δ SSIM
[%]PSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIM[%][%]
CityScapes2519.980.6017.200.5618.300.5917.300.5818.770.619.18.9
5020.450.6417.000.5518.950.6117.180.5919.040.6412.0016.4
10019.510.5917.120.5420.030.5817.750.6320.470.6519.620.4
Facade dataset2513.780.3510.930.2511.780.3110.810.2711.830.328.2328.0
5014.240.4011.000.2513.220.3711.920.2813.750.4025.060.0
10014.250.4210.980.2712.880.3311.520.3513.210.4120.351.8
Google Maps2530.350.6730.470.7130.620.7330.550.7531.200.772.48.4
5030.550.6829.780.7230.680.7529.810.7630.880.793.79.7
10030.010.6930.240.7330.920.7530.270.7731.230.813.311.0
Note: maximum for each S X is typeset in bold (PSNR) and blue bold with blue highlight (SSIM), while the largest δ over all S X is shaded in green.
Table 2. Results of the second type of experiments. Performance of the proposed SSL+BTS was compared against unsupervised CycleGAN, and supervised pix2pix by using Perceptual Realistic Measure (PRM) for the ‘Photo Map’ domain translation task, and by assessing image segmentation quality in the case of ‘Photo Semantic label’ task. Relative improvement of SSL+BTS over CycleGAN is expressed by δ , while the respective datasets are listed in the table.
Table 2. Results of the second type of experiments. Performance of the proposed SSL+BTS was compared against unsupervised CycleGAN, and supervised pix2pix by using Perceptual Realistic Measure (PRM) for the ‘Photo Map’ domain translation task, and by assessing image segmentation quality in the case of ‘Photo Semantic label’ task. Relative improvement of SSL+BTS over CycleGAN is expressed by δ , while the respective datasets are listed in the table.
Task/DatasetMeasure S X pix2pixCycleGANSSLBTSSSL+BTS δ
Photo Map
Google Maps
PRM2518.817.518.018.118.45.1
5019.818.719.319.219.54.3
10021.720.420.820.721.13.4
Photo Semantic label
CityScapes
FCN
per-pixel acc.
250.570.430.450.460.5016.3
500.630.490.520.510.5410.2
1000.710.520.560.550.5811.5
FCN
mean acc.
250.160.130.150.160.1838.5
500.200.150.170.180.1926.7
1000.250.170.190.200.2229.4
FCN
mean IoU
250.130.090.100.100.1233.3
500.160.100.110.120.1330.0
1000.180.110.130.130.1536.4
Note: all values are expressed in [%]; higher PRM is better (% of gen. images misinterpreted as real); maximum for each S X is typeset in bold and the largest δ over all S X is shaded in green.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Krstanović, L.; Popović, B.; Janev, M.; Brkljač, B. Bootstrapped SSL CycleGAN for Asymmetric Domain Transfer. Appl. Sci. 2022, 12, 3411. https://doi.org/10.3390/app12073411

AMA Style

Krstanović L, Popović B, Janev M, Brkljač B. Bootstrapped SSL CycleGAN for Asymmetric Domain Transfer. Applied Sciences. 2022; 12(7):3411. https://doi.org/10.3390/app12073411

Chicago/Turabian Style

Krstanović, Lidija, Branislav Popović, Marko Janev, and Branko Brkljač. 2022. "Bootstrapped SSL CycleGAN for Asymmetric Domain Transfer" Applied Sciences 12, no. 7: 3411. https://doi.org/10.3390/app12073411

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop