Next Article in Journal
Efficient IoT-Assisted Waste Collection for Urban Smart Cities
Previous Article in Journal
A Discussion of Building a Smart SHM Platform for Long-Span Bridge Monitoring
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Human-Unrecognizable Differential Private Noised Image Generation Method

School of Computer Science and Engineering, Pusan National University, Busan 46241, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(10), 3166; https://doi.org/10.3390/s24103166
Submission received: 11 March 2024 / Revised: 29 April 2024 / Accepted: 7 May 2024 / Published: 16 May 2024
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Differential privacy has emerged as a practical technique for privacy-preserving deep learning. However, recent studies on privacy attacks have demonstrated vulnerabilities in the existing differential privacy implementations for deep models. While encryption-based methods offer robust security, their computational overheads are often prohibitive. To address these challenges, we propose a novel differential privacy-based image generation method. Our approach employs two distinct noise types: one makes the image unrecognizable to humans, preserving privacy during transmission, while the other maintains features essential for machine learning analysis. This allows the deep learning service to provide accurate results, without compromising data privacy. We demonstrate the feasibility of our method on the CIFAR100 dataset, which offers a realistic complexity for evaluation.

1. Introduction

The rapid advancement in deep neural networks (DNNs) has enabled their widespread application in personalized services across diverse fields, including advertising [1], finance [2,3], and medicine [4,5,6]. To train these DNN models for such personalized services, institutions often collect and utilize extensive datasets. This data frequently contain sensitive information, raising concerns about user privacy and data protection.
The data memorization effect of DNNs, where models retain information beyond what is strictly necessary for their intended task [7], presents a significant privacy risk. This vulnerability enables malicious actors to target and extract sensitive data ‘memorized’ by the DNN. A prominent example is a model inversion attack [8,9,10], in which adversaries reconstruct representative input data from the DNN model itself, potentially exposing confidential information.
To mitigate sensitive information leakages, researchers have actively explored privacy-preserving deep learning (PPDL) techniques designed to maintain performance while protecting data. Two main categories have emerged: (1) encryption-based techniques, and (2) perturbation-based techniques.
Prominent encryption-based approaches include homomorphic encryption (HE) [11,12,13,14,15] and secure multi-party computation (SMPC) [16,17,18]. These methods encrypt DNN computations and values, ensuring the data remain unintelligible to unauthorized users without the necessary decryption keys. However, the substantial computational overheads introduced by these techniques often make them impractical. Additionally, the complex service architectures, frequently involving trusted third parties, can introduce vulnerabilities to privacy attacks [19].
In contrast to encryption-based techniques, perturbation-based methods modify DNN models or input data to prevent the reconstruction of the original information. Differential privacy (DP) [20] has emerged as a widely adopted model modification technique, due to its low computational overheads. A prominent example is differential private stochastic gradient descent (DP-SGD) [21], which introduces DP noise during the DNN training process. However, An et al. demonstrated that DP-SGD remains vulnerable to model inversion attacks, especially with clean, raw input data and a known input domain [22].
Input data modification strategies, such as marginal distribution (MD)-based techniques [23,24], offer an alternative. While effective for structured tabular data, their application is limited in this domain. Generative adversarial networks (GANs) [25,26,27] can address both structured and unstructured (e.g., image) data, but their success has been primarily confined to simple datasets like MNIST. Instance-hiding based methods [28] have shown good performance for accuracy on various datasets. However, they can be applied in the training phase only. Thus, such methods cannot be applied to protect a client’s data privacy.
Inspired by a recent study demonstrating that machines can achieve higher recognition rates on strategically noised images [29], we propose a differential privacy-based image generation method for image datasets. Our approach aims to make images unrecognizable to humans, preserving privacy, while enhancing machine-recognizability for specific tasks. We address two core questions: (1) how to augment machine-interpretable characteristics within the noised images, and (2) how to disrupt the original image’s explanatory cues, mitigating privacy vulnerabilities. To enhance machine-readability, we leverage explainable artificial intelligence (XAI) techniques. By extracting pretrained features from machine models and strategically embedding them into the noised images, we ensure that they remain interpretable by subsequent machine learning systems.
Main contributions of this paper can be summarized as follows:
  • We propose an image de-identification method that strategically combines two types of noise. The first type makes the image unrecognizable to humans, protecting privacy. The second type preserves essential features for machine learning tasks. This dual-noise approach effectively removes private information from input images, while maintaining the data’s utility for downstream analysis.
  • Our proposed method enables privacy-preserving deep learning-based services, by de-identifying client input images, protecting their data during service interactions.
  • Our experimental results demonstrate the potential of XAI techniques for generating features or clues that facilitate DNN classification. We anticipate that these findings will stimulate future research directions focused on enhancing privacy across diverse data modalities.
The rest of this paper is organized as follows. In Section 2, we describe related works. We propose our differential private image de-identification method and give a theoretical analysis in Section 3 and Section 4, respectively. In Section 5, we show the feasibility of the proposed method with experimental results. Finally, we summarize this paper in Section 6.

2. Related Works

2.1. Perturbation-Based Privacy-Preserving Deep Learning

Differential private stochastic gradient descent (DP-SGD) [21] stands as a prominent perturbation-based PPDL method. It protects privacy by injecting calibrated noise into the model’s gradients during the training process. This allows the model to learn essential features, while adhering to the rigorous mathematical guarantees of differential privacy, which quantify privacy protection based on the chosen parameters. However, recent advancements in model inversion attacks have demonstrated that DP-SGD remains vulnerable, even with carefully tuned privacy settings [22].
In contrast to model modification methods, input modification approaches probabilistically alter sensitive information in the original data, hindering malicious exploitation of DNN models. Two notable examples include DataSynthesizer by Ping et al. and the Bayesian network method proposed by Zhang et al. [23,24]. DataSynthesizer offers differential privacy by analyzing dataset distributions and feature correlations, while Zhang et al. employed a differential private Bayesian network for data synthesis. However, a key limitation of both methods is their reliance on Bayesian theory, making them primarily suitable for structured tabular data, where data distributions are well-defined.
GAN-based approaches offer an alternative for handling unstructured data, where data distributions are less well defined. Notable examples include DP-GAN by Xie et al. [25], which incorporates DP-SGD into the GAN’s generator, and the teacher–student GAN model proposed by Jordan et al. [26] for differential private generation. Torkzadehmahani et al. [27] further extended DP-GAN with a conditional framework for label generation. However, a significant limitation remains: these methods have primarily demonstrated feasibility on simple datasets, either structured or like MNIST.
Huang et al. introduced instance-hiding as an alternative input modification approach [28]. Their method strategically blends pixels from private data with those from a large, diverse public dataset. This obfuscates the original content, while selectively preserving key features, enabling machine learning on the modified data and subsequent classification of clean inputs. However, a critical vulnerability remains: during service, the instance-hiding method receives clean, raw data, exposing its distribution and leaving it susceptible to state-of-the-art model inversion attacks. Furthermore, as illustrated in Figure 1a, this lack of service-phase protection prevents clients from securely submitting their private data for analysis by the service model.
To address the service-phase vulnerability of instance-hiding, Gao et al. proposed an image obfuscation method specifically for medical images [18]. Their method randomizes both pixel values and positions within a defined distribution, achieving a balance between practicality and client-side applicability. However, this randomization approach presents a significant limitation: it disrupts the spatial relationships between pixels, making the method unsuitable for services where object location is crucial, such as object-detection tasks.

2.2. Model Inversion Attack

Since the seminal work on model inversion attacks in [8], the field has witnessed the rapid development of increasingly sophisticated techniques. In this section, we survey recent advancements in state-of-the-art model inversion attack research.
Balle et al. introduced a black-box attack capable of reconstructing training data [30]. Assuming knowledge of the data distribution, they leveraged shadow models and confidence scores from the target model. Wang et al. proposed a method combining variational autoencoders with StyleGAN [31,32]. By training a prior distribution for the latent space that reflects the training data distribution, they could generate representative latent vectors suitable for StyleGAN. An et al. also employed StyleGAN, but with a different approach [22]. They trained a latent space mapping network using the target classifier’s confidence scores. This allowed them to extract representative data of the target class, bypassing the defenses of DP-SGD-based PPDL models, even with strong privacy settings.
The current landscape of model inversion attacks is dominated by black-box approaches. These attacks require minimal information—only the target model and the data distribution—posing a significant challenge to privacy-preserving deep learning. Since DP-based PPDL methods inherently reveal the data distribution, they are particularly susceptible to such attacks. This highlights the critical need for further research into developing robust privacy-preserving techniques that can withstand these increasingly sophisticated black-box attacks.

2.3. Explainable Artificial Intelligent

The widespread adoption of machine learning (ML) has spurred questions about whether models arrive at their classifications using features that align with human understanding. To address this, explainable artificial intelligence (XAI) has emerged as a field dedicated to developing techniques that reveal the features influencing ML model outputs. XAI aims to provide transparency and insights into the decision-making processes of these complex models.
Several XAI techniques exist for explaining DNN behavior. Gradient-based methods, such as those proposed by Simonyan et al. [33], utilize backpropagation gradients to generate heatmap-style explanation maps. Guided backpropagation [34] builds upon this concept but drops negative gradients, aiming to highlight only the positive contributions of input features to the final output. Layer-wise relevance propagation (LRP) [35,36] takes a different approach, analyzing the activated weights within the DNN. By calculating the contribution of each weight, LRP generates an explanation map that reveals how individual features influenced the model’s decision.

3. Human-Unrecognizable Differential Private Noised Image Generation Method

Existing input modification-based PPDL schemes often utilize unmodified original images during the service time, as shown in Figure 1a. This exposes client data to potential privacy risks, including network hijacking. While encryption offers a degree of protection, it has known vulnerabilities to certain privacy attacks [19]. Alternatively, adding substantial noise to the data can provide robust privacy by completely obscuring sensitive information. However, this extreme obfuscation makes the image unusable for both authorized humans and machine learning models, negating its utility.
To address the limitations of traditional privacy-preserving methods, we propose a noise generation method that makes images unrecognizable to humans, while preserving machine-readability. Our approach combines two core elements:
  • Human Obfuscation: We introduce strong Gaussian noise into the image, disrupting the visual coherence and making it unrecognizable to humans.
  • Machine-Readable Enhancement: Leveraging XAI techniques, we extract crucial image features, such as structural information, and re-embed them into the noised image. This maintains machine-recognizability, despite the obfuscation.
This hybrid method effectively safeguards privacy, while enabling machine learning analysis.
Recent research on DNN security, particularly in adversarial examples and XAI, provides a compelling foundation for our approach. These studies highlight a critical disparity: features readily identifiable by humans may not be the same features crucial for machine recognition [37,38]. For instance, an image and its adversarial counterpart can appear identical to humans, yet be classified differently by a machine learning model. Similarly, Dombrowski et al. demonstrated that explanation maps, tools used for understanding DNN decisions, can diverge for visually indistinguishable images [38]. By leveraging this inherent disparity, our method injects machine-recognizable features into images, achieving human-imperceptible obfuscation, while preserving machine readability.

3.1. Overview

Figure 1b shows a simplified architecture of the proposed differential private image generation method. The proposed method consists of two components; i.e., Conversion  g ( · ) and Service Model  f ( · ) . The Conversion component transforms the input image into a noised format, preserving machine-readability, while obscuring the image from human recognition, and the Service Model component conducts training and analysis tasks on the converted, noised image.
Our two-component system, consisting of a data de-identification module and an analysis service, offers a promising approach for securing client privacy during data transfer. This architecture leverages DP and is divided into three phases:
Phase 1: Training with Privacy Protection
  • Distribution Learning: The de-identification module, denoted as g ( · ) , learns the distribution of Gaussian noise and the machine-readable features essential for accurate analysis. XAI techniques are employed to extract these features. However, it is crucial to acknowledge that such features might introduce some level of information leakage.
  • Privacy-Preserving Noise Injection: To mitigate information leakage risks arising from a potential malicious analysis of the service model ( f ( · ) ), DP noise is incorporated into the training process of g ( · ) . This added noise mathematically guarantees a level of privacy protection against leakage through the service model.
  • Service Model Training: Once g ( · ) has been trained, the noisy images generated by g ( · ) are used to train the service model f ( · ) .
Phase 2: Secure Distribution of De-identification Module
The trained g ( · ) component is then securely delivered to authorized clients through a secure channel.
Phase 3: Client-Side Data Conversion and Secure Analysis
  • Client-Side Conversion: When a client has private data containing sensitive information to be analyzed, the client utilizes the locally deployed g ( · ) component. This component transforms the client’s data into a machine-readable but human-imperceptible noisy image.
  • Secure Transmission and Analysis: The anonymized noisy image is then transmitted to the service provider. The service model f ( · ) , previously trained on similar noisy images, can perform the required analysis, without compromising the client’s raw data privacy, due to the DP guarantees and human-imperceptibility of the noise.

3.2. Conversion Component

This section delves into our core component, termed the conversion component ( g ( · ) ), which addresses two crucial questions in privacy-preserving deep learning: (1) How can images be obfuscated for privacy, while retaining features essential for machine learning?; and (2) How can we eliminate potential information leakages through explanation maps that arise from obfuscated images?
To move beyond the limitations of purely random Gaussian noise, we replace it with a generator that learns to produce noise with a controlled distribution. This balances obfuscation with structure preservation. Furthermore, we employ XAI techniques to extract machine-readable features. Unlike traditional CNN feature layers or pretrained teacher models, which are strongly tied to specific details within the original image, XAI results have a weaker connection to the image itself. XAI results instead highlight how the model reached its output, similar to the backpropagation process. This characteristic makes them ideal for extracting machine-readable features without excessively compromising privacy.
A significant challenge lies in the potential for an image’s partial shape or features to persist within its XAI-generated explanation map, even after obfuscation. To combat this, we add a controlled degree of randomness to the noisy image’s explanation map during the training process of our conversion component. However, indiscriminate random perturbations risk damaging the very features we aim to preserve for analysis. Instead, we carefully overlay the explanation map with that of a different, randomly chosen class. This process is meticulously managed using DP, ensuring mathematical guarantees regarding the privacy level offered.
To train our conversion component ( g ( · ) ), we employ a specialized generative adversarial network (GAN). This GAN incorporates a discriminator that guides the generator towards producing structured, target-magnitude Gaussian noise, and an explainer ( f ( · ) ) that facilitates the generation of machine-readable features using XAI-derived explanation maps (see Figure 2). The explainer’s architecture mirrors that of the service model ( f ( · ) ), and it is pretrained with the original, unmodified data. For the given explainer XAI method h ( · ) , Gaussian noise generation method n ( · ) , differential private noise generation method N ( · ) , and specified input image x, the objective function of our GAN algorithm can be expressed as
min g max D E n ( x ) p data ( n ( x ) ) [ log D ( n ( x ) ) ) ] + E x p data ( x ) [ 1 log D ( g ( x ) ) + | h     f ( x ) h N f ( g ( x ) ) | ]
where ∘ is the composition of functions, g ( · ) is a generator, which is the Conversion component, and D is a discriminator. In Equation (1), the term | h     f ( x ) h N f ( g ( x ) ) | means the distance between the explain map of the original data and the explain map of the converted data with differential private noise. Here, the loss function of the discriminator is the same as that of the original GAN’s discriminator, except that the distribution of the discriminator’s input is changed from the original images to noised images. Contrarily, the loss function of the generator is defined as a joint loss consisting of noise loss and explanation loss.
The noise loss L n o i s e measures the errors of the discriminator for modified image data G ( x ) and is calculated as follows:
L n o i s e = log ( 1 D ( g ( x ) ) )
which encourages the Conversion component to learn the distribution of Gaussian noise.
The explanation loss L e x p measures the difference between the explain maps of the original image and the modified image and is calculated as follows:
L e x p = d ( h f ( x ) , h N f ( g ( x ) ) )
where d ( · ) is a distance function that measures the difference between the explain maps of the original image h f ( x ) and the converted data with differential private noise h N f ( g ( x ) ) . The explanation loss makes features of the original image and the converted image more similar when analyzed by machine; that is, it encourages the Conversion component to inject machine-recognizable features into the converted image.
Consequently, the loss function of the generator L G is defined as a weighted summation of two such loss functions:
L g = α L n o i s e + β L e x p
where α and β are hyperparameters to control the characteristics of the generator.
As mentioned above, the role of the discriminator is to encourage the Conversion component to learn the distribution of Gaussian noise. In other words, the distribution of the input images is changed to the distribution of noised images. Thus, the loss function of the discriminator L D is defined as follows:
L D = log ( 1 D ( n ( x ) ) ) + log ( D ( g ( x ) ) )

4. Theoretical Analysis

To show that the security of the proposed method can be supported by DP, we describe our theoretical basis in this section. According to the definition of DP, the proposed method has to satisfy following equation:
P r [ f ( D 1 ) = S ] P r [ f ( D 2 ) = S ] e ϵ + δ
where f ( · ) is a deterministic function that hides a single data point. D 1 and D 2 are neighboring data points in a specific data distribution. ϵ and δ are privacy parameters. The distinguishability of two processed data points, f ( D 1 ) and f ( D 2 ) , increases when the value of ϵ and δ increases.
In Equation (6), to address neighboring data D 1 and D 2 , we need to define the field of data. Since the proposed method is focusing on image data, the field of data can be width · height · channels · 255 in RGB expression. However, such a field only represents the field of each pixel, not the contents in the image. Since the content in the image is expressed by a group of pixels, we define a set of classification results as a field of data.
Therefore, the conversion component of the proposed method can be expressed as follows:
h N f ( g ( x ) )
where N is a deterministic function, which generates noise satisfying DP, which is equivalent to f in Equation (6). Here, we apply a theorem that defines the properties of DP as follow [39]:
Theorem 1.
If F ( x ) satisfies DP, then for any deterministic or randomized operation g on F ( x ) , g ( F ( x ) ) satisfies DP.
Therefore, following the Theorem 1, the conversion component satisfies DP.

5. Experiment

To show the feasibility of our proposed scheme, we show experimental results as follows: (1) the classification performance according to the values of noise magnitude σ and privacy parameter ϵ and δ ; (2) a de-identification performance comparison of the image with the original Gaussian noise and the image with the proposed scheme; (3) an analysis on machine-recognizable features from XAI; and (4) the effect of DP on the explanation map.

5.1. Experimental Configuration

Since most reference implementations of XAI do not support gradient calculation from explanation maps, we implemented an explainer network that calculates gradients from explanation maps referring to Dombrowski et al.’s implementation [38]. The Conversion component was implemented with a generator, as shown in Table 1, and a Service Model component that performed classification on the generated images, the VGG16 network shown in Table 2 with input size of 224 × 224 pixels, was used.
In addition, we performed all experiments using the CIFAR100 [40] dataset. Existing DP-based data-modification-based PPDL approaches, such as [25,27], were used MNIST as a benchmark dataset. However, since the MNIST dataset has a small number of classes and too simple a structure, the MNIST dataset was not suitable to show the possibility of general applications to various fields. On the other hand, the CIFAR100 dataset consists of very diverse types of images divided into 100 classes. Due to the diversity of the composed images, it is used to show the general applicability of new techniques in state-of-the-art research related to image processing [41,42,43,44]. Therefore, to show the feasibility of the proposed human-unrecognizable differential private noised image generation method, we used the CIFAR100 dataset in the experiments.
The differential private noise generation mechanism and the XAI method for the explainer were set up as a Gaussian mechanism [45] and guided backpropagation [34], respectively.

5.2. Classification Performance

To show the feasibility of the proposed human-unrecognizable differential private noised image generation method, we evaluated the classification accuracy on the CIFAR100 dataset in two ways: (1) measurement of the classification accuracy under various parameters; and an (2) ablation test without L e x p . Here, the accuracy of the explainer model used to train the Conversion component that satisfied the parameters was 0.720. The overall results are shown in Table 3.
First, we measured the accuracy of the Service Model component trained with the generated noised training images using the Conversion component. Each Conversion component was trained with noise parameter σ values of 3.0, 4.0, and 5.0 and privacy parameters ( ϵ , δ ) value (0.1, 10 5 ), (0.1, 0.9), (0.9, 10 5 ) and (0.9, 0.9), respectively. When the σ value was 3.0 and ϵ value was 1, the accuracy was observed to be about 0.62 with both of δ values. Similarly, with a σ value of 4.0, the accuracy was about 0.61 and 0.63 with ϵ values of 0.1 and 0.9, respectively. For a σ value of 5.0, accuracy was about 0.60 and 0.62 with ϵ values of 0.1 and 0.9, respectively. From such results, we observed a tendency that matched with the theoretical characteristic of DP, where the value of ϵ affected the noise more than that of δ . In the ablation test, we observed accuracies of 0.603, 0.586, and 0.577 with respect to σ values of 3.0, 4.0, and 5.0, respectively. Such results imply that the explanation loss L e x p is a key feature for injecting machine-recognizable features into a converted image. Considering the simple network architecture of the Service Model component and the complexity of the CIFAR100 dataset, such results were considered acceptable [46].
The Conversion component of the proposed scheme mimics the Gaussian noise and differential private explainer’s explanation maps; that is, it is infeasible to show the privacy of modified images with mathematical theory. Instead, we show that the noise generated from the Conversion component is similar enough to the noised image with real Gaussian noise.

5.3. Image De-Identification

Figure 3 shows graphical examples of images with the original Gaussian noise and converted images using the proposed scheme. According to the characteristic of Gaussian noise, the magnitude of noise increases when the σ value increases. In other words, details of the original image fade out and the probability of a sensitive information leakage through human analysis decreases, Figure 3a. Specifically, it is very difficult to recognize the presence of any object in the image with a value of σ around 3.0 and more. Images converted by the proposed Conversion component under each target σ are shown in Figure 3b. When comparing the images with Gaussian noise and converted images, it is almost impossible to see a distinctive difference.
To show the de-identification performance concretely, we measured a quantitative indicator of Structural Similarity (SSIM) [47], which is widely used in the image processing field to measure a generated image’s quality. The SSIM quantifies the perceived similarity between two images, often referred to as the reference image (x) and the test image (y). It goes beyond a simple pixel intensity comparison by incorporating luminance ( l ( x , y ) ), contrast ( c ( x , y ) ), and structure ( s ( x , y ) ) comparisons. The SSIM value ranges from −1 to 1, with 1 indicating perfect structural similarity and values closer to −1 signifying significant structural dissimilarities. The specific equation for SSIM involves a combination of these three components:
S S I M ( x , y ) = [ l ( x , y ) × c ( x , y ) × s ( x , y ) ]
where each component is calculated based on the means ( μ x , μ y ), standard deviations ( σ x , σ y ), and covariance ( σ x y ) of x and y within a local window, along with two parameters (C1 and C2) to stabilize the division by small denominators. This metric provides a valuable tool for assessing image quality and compression effectiveness in image processing applications. Therefore, we calculated the SSIM between the original image and the modified image, to show the average difference from the original image. In other words, a smaller SSIM value means a lower recognition probability by humans. Table 4 shows the average SSIM of the test datasets of CIFAR10 and CIFAR100, which were generated with the original Gaussian noise and the proposed scheme. In both datasets, the average SSIM value decreased as the target σ increased. In addition, we observed that there was no difference in the average SSIM of noised images and converted images according to each target σ and privacy parameter ϵ . That is, the proposed conversion component showed a very stable de-identification performance.

5.4. Machine-Recognizable Features from XAI

To analyze the feasibility of the proposed machine-recognizable feature injection method, we visualized some features analyzed by machine using XAI, i.e., guided backpropagation, and Figure 4 shows the results. The first row shows the original image of samples, and the corresponding converted image by Conversion component g ( x ) is shown in the second row. In the third row, an explanation map of the classification results of the explainer model of the original image h f ( x ) is shown. Explanation maps of the classification results of the explainer model h f g ( x ) and the Service Model component h f g ( x ) are shown in the fourth and fifth row, respectively. The last row shows explanation maps of the classification results of the model where DP was not applied. All samples were selected from the test set of the CIFAR100 dataset, which was classified correctly by the Service Model trained with σ = 3.0 , ϵ = 0.1 and δ = 10 5 .
Considering Equation (3), the explanation map of h f ( x ) and h f g ( x ) should have become similar as the model shows better performance. However, the explanation maps of h f g ( x ) in the fourth row only shows very strange images that do not match with the corresponding explanation map in the third row. In addition, the explanation map of h f ( x ) and h f g ( x ) show different shapes with proper privacy parameter settings. However, the experimental results in Section 5.2 show a clear performance difference using L e x p . Additionally, compared to the last row, which shows almost all details, the explanation maps in the fifth row hide details of the object very well. As a consequence of such observations, we conjecture that the explanation loss L e x p preserves machine-recognizable features that are not human-unrecognizable, even in the presence of noise from DP.

5.5. Effect of DP on the Explanation Map

To observe the effect of DP on the converted image, we extracted an explanation map from arbitrary selected samples under various privacy parameters. All samples were selected from the test set of the CIFAR100 dataset that was classified correctly by all Service Models with each privacy parameter. Figure 5 shows the partial results of the selected samples. In each sub-figure, the first row shows the original image of each sample; the second row shows the corresponding explanation map with a stronger privacy parameter; and the third row shows another corresponding explanation map with weaker privacy parameters.
As shown in Figure 5a, most explanation maps show the outlines of the object with a relatively low target σ value. However, when considering privacy parameters, we can observe that explanation maps with weaker parameters show colors and shapes similar to those of the original object. In particular, in the first sample from left, the shape of the fish has almost disappeared in the second row. Conversely, the third row shows a much clearer shape and colors of the yellow fish. Meanwhile, with a relatively high target σ value, the explanation maps show very noisy images, regardless of the privacy parameters. However, we can intermittently observe relatively clear explanation maps, such as the last image of the third row from the left.

6. Conclusions

The proposed method presents a new privacy-preserving deep learning method that tackles the challenge of protecting client data during service interactions. Our contribution lies in differential privacy-based image de-identification. This method strategically injects noise to obfuscate visual content, while strategically embedding machine-readable, XAI-derived features. We achieve this balance using a customized GAN architecture that explicitly incorporates explanation maps during the training process.
In addition, our approach addresses the limitations of the existing perturbation-based methods, which can be vulnerable to state-of-the-art model inversion attacks. The integration of differential privacy (DP) provides theoretical guarantees of privacy, with the level of protection controlled by the DP parameters.
The optimization of the accuracy and privacy trade-offs caused by image de-identification, as well as differential privacy and extensions to other types of tasks, such as object detection, will be our future work.

Author Contributions

Conceptualization, H.-G.K. and J.S.; methodology, H.-G.K. and J.S.; software, J.S.; validation, H.-G.K., J.S. and Y.-H.C.; formal analysis, J.S.; investigation, H.-G.K. and J.S.; resources, Y.-H.C.; data curation, J.S.; writing—original draft preparation, H.-G.K.; writing—review and editing, J.S.; visualization, J.S.; supervision, Y.-H.C.; project administration, Y.-H.C.; funding acquisition, Y.-H.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the MSIT (Ministry of Science and ICT), Korea, under the ITRC (Information Technology Research Center) support program (IITP-2023-RS-2023-00259967) supervised by the IITP (Institute for Information & Communications Technology Planning & Evaluation); Institute for Information & communications Technology Planning & Evaluation (IITP) grant funded by the Korea government (MSIT) (No. 2023-0-01201, Convergence security core talent training business (Pusan National University)); and Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (2022R1I1A3055233).

Institutional Review Board Statement

Ethical review and approval were waived for this study due to the use of a publicly available dataset.

Informed Consent Statement

Patient consent was waived due to the use of a publicly available dataset for researches.

Data Availability Statement

The all data including source codes for implementation is available on https://github.com/sinryang/You-Know-Nothing, accessed on 10 March 2024.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, C.; Kong, Y.; Zhou, X.; Zhang, H.; Zhang, X.; Geng, C.; Chu, D.; Wu, X. An Effective Deep Learning Approach for Personalized Advertisement Service Recommend. In Proceedings of the 2021 International Conference on Service Science (ICSS), Xi’an, China, 14–16 May 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 96–101. [Google Scholar]
  2. Ozbayoglu, A.M.; Gudelek, M.U.; Sezer, O.B. Deep learning for financial applications: A survey. Appl. Soft Comput. 2020, 93, 106384. [Google Scholar] [CrossRef]
  3. Wang, W.; Li, W.; Zhang, N.; Liu, K. Portfolio formation with preselection using deep learning from long-term financial data. Expert Syst. Appl. 2020, 143, 113042. [Google Scholar] [CrossRef]
  4. Zhou, X.; Li, Y.; Liang, W. CNN-RNN based intelligent recommendation for online medical pre-diagnosis support. IEEE/ACM Trans. Comput. Biol. Bioinform. 2020, 18, 912–921. [Google Scholar] [CrossRef] [PubMed]
  5. Bullock, J.; Cuesta-Lázaro, C.; Quera-Bofarull, A. XNet: A convolutional neural network (CNN) implementation for medical X-ray image segmentation suitable for small datasets. In Proceedings of the Medical Imaging 2019: Biomedical Applications in Molecular, Structural, and Functional Imaging, San Diego, CA, USA, 16–21 February 2019; SPIE: Bellingham, WA, USA, 2019; Volume 10953, pp. 453–463. [Google Scholar]
  6. Xie, Y.; Zhang, J.; Shen, C.; Xia, Y. Cotr: Efficiently bridging cnn and transformer for 3d medical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Strasbourg, France, 27 September–1 October 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 171–180. [Google Scholar]
  7. Carlini, N.; Liu, C.; Erlingsson, Ú.; Kos, J.; Song, D. The secret sharer: Evaluating and testing unintended memorization in neural networks. In Proceedings of the 28th USENIX Security Symposium (USENIX Security 19), Santa Clara, CA, USA, 14–16 August 2019; pp. 267–284. [Google Scholar]
  8. Fredrikson, M.; Jha, S.; Ristenpart, T. Model inversion attacks that exploit confidence information and basic countermeasures. In Proceedings of the 22nd ACM SIGSAC Conference on Computer and Communications Security, Denver, CO, USA, 12–16 October 2015; pp. 1322–1333. [Google Scholar]
  9. Zhang, Y.; Jia, R.; Pei, H.; Wang, W.; Li, B.; Song, D. The secret revealer: Generative model-inversion attacks against deep neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 253–261. [Google Scholar]
  10. Kumar, A.; Levine, S. Model inversion networks for model-based optimization. Adv. Neural Inf. Process. Syst. 2020, 33, 5126–5137. [Google Scholar]
  11. Gilad-Bachrach, R.; Dowlin, N.; Laine, K.; Lauter, K.; Naehrig, M.; Wernsing, J. Cryptonets: Applying neural networks to encrypted data with high throughput and accuracy. In Proceedings of the International Conference on Machine Learning, New York, NY, USA, 19–24 June 2016; PMLR: Westminster, UK, 2016; pp. 201–210. [Google Scholar]
  12. Hesamifard, E.; Takabi, H.; Ghasemi, M. Cryptodl: Towards deep learning over encrypted data. In Proceedings of the Annual Computer Security Applications Conference (ACSAC 2016), Los Angeles, CA, USA, 5–8 December 2016; Volume 11. [Google Scholar]
  13. Boemer, F.; Costache, A.; Cammarota, R.; Wierzynski, C. ngraph-he2: A high-throughput framework for neural network inference on encrypted data. In Proceedings of the 7th ACM Workshop on Encrypted Computing & Applied Homomorphic Cryptography, London, UK, 11 November 2019; pp. 45–56. [Google Scholar]
  14. Falcetta, A.; Roveri, M. Privacy-preserving deep learning with homomorphic encryption: An introduction. IEEE Comput. Intell. Mag. 2022, 17, 14–25. [Google Scholar] [CrossRef]
  15. Podschwadt, R.; Takabi, D.; Hu, P.; Rafiei, M.H.; Cai, Z. A survey of deep learning architectures for privacy-preserving machine learning with fully homomorphic encryption. IEEE Access 2022, 10, 117477–117500. [Google Scholar] [CrossRef]
  16. Ma, X.; Zhang, F.; Chen, X.; Shen, J. Privacy preserving multi-party computation delegation for deep learning in cloud computing. Inf. Sci. 2018, 459, 103–116. [Google Scholar] [CrossRef]
  17. Sayyad, S. Privacy preserving deep learning using secure multiparty computation. In Proceedings of the 2020 Second International Conference on Inventive Research in Computing Applications (ICIRCA), Coimbatore, India, 15–17 July 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 139–142. [Google Scholar]
  18. Gao, C.; Yu, J. SecureRC: A system for privacy-preserving relation classification using secure multi-party computation. Comput. Secur. 2023, 128, 103142. [Google Scholar] [CrossRef]
  19. Shin, J.; Choi, S.H.; Choi, Y.H. Is Homomorphic Encryption-Based Deep Learning Secure Enough? Sensors 2021, 21, 7806. [Google Scholar] [CrossRef]
  20. Dwork, C. Differential privacy: A survey of results. In Proceedings of the International Conference on Theory and Applications of Models of Computation, Xi’an, China, 25–29 April 2008; Springer: Berlin/Heidelberg, Germany, 2008; pp. 1–19. [Google Scholar]
  21. Abadi, M.; Chu, A.; Goodfellow, I.; McMahan, H.B.; Mironov, I.; Talwar, K.; Zhang, L. Deep learning with differential privacy. In Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security, Vienna, Austria, 24–28 October 2016; pp. 308–318. [Google Scholar]
  22. An, S.; Tao, G.; Xu, Q.; Liu, Y.; Shen, G.; Yao, Y.; Xu, J.; Zhang, X. MIRROR: Model Inversion for Deep Learning Network with High Fidelity. In Proceedings of the Network and Distributed Systems Security Symposium (NDSS 2022), San Diego, CA, USA, 24–28 April 2022. [Google Scholar]
  23. Ping, H.; Stoyanovich, J.; Howe, B. Datasynthesizer: Privacy-preserving synthetic datasets. In Proceedings of the 29th International Conference on Scientific and Statistical Database Management, Chicago, IL, USA, 27–29 June 2017; pp. 1–5. [Google Scholar]
  24. Zhang, J.; Cormode, G.; Procopiuc, C.M.; Srivastava, D.; Xiao, X. Privbayes: Private data release via bayesian networks. ACM Trans. Database Syst. (TODS) 2017, 42, 1–41. [Google Scholar] [CrossRef]
  25. Xie, L.; Lin, K.; Wang, S.; Wang, F.; Zhou, J. Differentially Private Generative Adversarial Network. arXiv 2018, arXiv:1802.06739. [Google Scholar]
  26. Jordon, J.; Yoon, J.; Van Der Schaar, M. PATE-GAN: Generating synthetic data with differential privacy guarantees. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  27. Torkzadehmahani, R.; Kairouz, P.; Paten, B. DP-CGAN: Differentially Private Synthetic Data and Label Generation. arXiv 2020, arXiv:2001.09700. [Google Scholar]
  28. Huang, Y.; Song, Z.; Li, K.; Arora, S. Instahide: Instance-hiding schemes for private distributed learning. In Proceedings of the International Conference on Machine Learning, Virtual, 13–18 July 2020; PMLR: Westminster, UK, 2020; pp. 4507–4518. [Google Scholar]
  29. Jang, H.; McCormack, D.; Tong, F. Noise-trained deep neural networks effectively predict human vision and its neural responses to challenging images. PLoS Biol. 2021, 19, e3001418. [Google Scholar] [CrossRef] [PubMed]
  30. Balle, B.; Cherubin, G.; Hayes, J. Reconstructing training data with informed adversaries. In Proceedings of the 2022 IEEE Symposium on Security and Privacy (SP), Francisco, CA, USA, 22–26 May 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1138–1156. [Google Scholar]
  31. Karras, T.; Laine, S.; Aila, T. A style-based generator architecture for generative adversarial networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 4401–4410. [Google Scholar]
  32. Wang, K.C.; Fu, Y.; Li, K.; Khisti, A.; Zemel, R.; Makhzani, A. Variational model inversion attacks. Adv. Neural Inf. Process. Syst. 2021, 34, 9706–9719. [Google Scholar]
  33. Simonyan, K.; Vedaldi, A.; Zisserman, A. Deep inside convolutional networks: Visualising image classification models and saliency maps. arXiv 2013, arXiv:1312.6034. [Google Scholar]
  34. Springenberg, J.T.; Dosovitskiy, A.; Brox, T.; Riedmiller, M. Striving for simplicity: The all convolutional net. arXiv 2014, arXiv:1412.6806. [Google Scholar]
  35. Binder, A.; Bach, S.; Montavon, G.; Müller, K.R.; Samek, W. Layer-wise relevance propagation for deep neural network architectures. In Information Science and Applications (ICISA) 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 913–922. [Google Scholar]
  36. Montavon, G.; Lapuschkin, S.; Binder, A.; Samek, W.; Müller, K.R. Explaining nonlinear classification decisions with deep taylor decomposition. Pattern Recognit. 2017, 65, 211–222. [Google Scholar] [CrossRef]
  37. Choi, S.H.; Shin, J.; Liu, P.; Choi, Y.H. EEJE: Two-step input transformation for robust DNN against adversarial examples. IEEE Trans. Netw. Sci. Eng. 2020, 8, 908–920. [Google Scholar] [CrossRef]
  38. Dombrowski, A.K.; Alber, M.; Anders, C.; Ackermann, M.; Müller, K.R.; Kessel, P. Explanations can be manipulated and geometry is to blame. Adv. Neural Inf. Process. Syst. 2019, 32. [Google Scholar]
  39. Dwork, C.; McSherry, F.; Nissim, K.; Smith, A. Calibrating noise to sensitivity in private data analysis. In Proceedings of the Theory of Cryptography: Third Theory of Cryptography Conference, TCC 2006, New York, NY, USA, 4–7 March 2006; Springer: Berlin/Heidelberg, Germany, 2006; pp. 265–284. [Google Scholar]
  40. Krizhevsky, A.; Hinton, G. Learning Multiple Layers of Features from Tiny Images. 2009. Available online: https://www.cs.toronto.edu/~kriz/learning-features-2009-TR.pdf (accessed on 10 March 2024).
  41. Cazenavette, G.; Wang, T.; Torralba, A.; Efros, A.A.; Zhu, J.Y. Dataset distillation by matching training trajectories. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 4750–4759. [Google Scholar]
  42. Wang, Q.; Fink, O.; Van Gool, L.; Dai, D. Continual test-time domain adaptation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 7201–7211. [Google Scholar]
  43. Wang, X.; Fan, H.; Tian, Y.; Kihara, D.; Chen, X. On the importance of asymmetry for siamese representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 16570–16579. [Google Scholar]
  44. Zhang, H.; Zu, K.; Lu, J.; Zou, Y.; Meng, D. EPSANet: An efficient pyramid squeeze attention block on convolutional neural network. In Proceedings of the Asian Conference on Computer Vision, Macao, China, 4–8 December 2022; pp. 1161–1177. [Google Scholar]
  45. Dwork, C.; Roth, A. The algorithmic foundations of differential privacy. Found. Trends Theor. Comput. Sci. 2014, 9, 211–407. [Google Scholar] [CrossRef]
  46. Jaiswal, A.K.; Ma, H.; Chen, T.; Ding, Y.; Wang, Z. Training your sparse neural network better with any mask. In Proceedings of the International Conference on Machine Learning, Baltimore, MD, USA, 17–23 July 2022; PMLR: Westminster, UK, 2022; pp. 9833–9844. [Google Scholar]
  47. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Architecture comparison of the existing and the proposed input modification schemes.
Figure 1. Architecture comparison of the existing and the proposed input modification schemes.
Sensors 24 03166 g001
Figure 2. GAN architecture for the proposed image conversion component.
Figure 2. GAN architecture for the proposed image conversion component.
Sensors 24 03166 g002
Figure 3. Comparison of noised images with Gaussian noise and converted images using the proposed scheme. (a) Noised images with Gaussian noise with a target σ value; (b) Converted images using the proposed scheme with a target σ value ( ϵ = 0.1 , δ = 0.9 ).
Figure 3. Comparison of noised images with Gaussian noise and converted images using the proposed scheme. (a) Noised images with Gaussian noise with a target σ value; (b) Converted images using the proposed scheme with a target σ value ( ϵ = 0.1 , δ = 0.9 ).
Sensors 24 03166 g003
Figure 4. Comparison of the explanation map extracted from the explainer and the service model ( σ = 3.0 , ϵ = 0.1 , δ = 10 5 ).
Figure 4. Comparison of the explanation map extracted from the explainer and the service model ( σ = 3.0 , ϵ = 0.1 , δ = 10 5 ).
Sensors 24 03166 g004
Figure 5. Comparison of noised images with Gaussian noise and converted images using the proposed method.
Figure 5. Comparison of noised images with Gaussian noise and converted images using the proposed method.
Sensors 24 03166 g005
Table 1. Details of the implemented generator network for the conversion component.
Table 1. Details of the implemented generator network for the conversion component.
LayerDescriptionNumber of Parameters
Convolution (7 × 7, stride 1, padding 3)64 filters1792
Instance Norm--
ReLU activation--
(6 residual blocks)
Deconvolution (3 × 3, stride 1, padding 1)64 filters4352
Instance Norm--
ReLU activation--
Convolution (7 × 7, stride 1, padding 3)Target image channels1792
tanh activation--
Table 2. Details of the implemented VGG16 network.
Table 2. Details of the implemented VGG16 network.
LayerDescriptionNumber of Parameters
Convolution (3 × 3 , same padding)64 filters, ReLU activation1792
Convolution (3 × 3, same padding)64 filters, ReLU activation36,864
Max Pooling (2 × 2, stride 2)-
Convolution (3 × 3, same padding)128 filters, ReLU activation73,792
Convolution (3 × 3, same padding)128 filters, ReLU activation147,520
Max Pooling (2 × 2, stride 2)-
Convolution (3 × 3, same padding)256 filters, ReLU activation295,136
Convolution (3 × 3, same padding)256 filters, ReLU activation590,080
Convolution (1 × 1)256 filters, ReLU activation65,536
Convolution (3 × 3, same padding)256 filters, ReLU activation590,080
Max Pooling (2 × 2, stride 2)-
Fully-connected4096 units, ReLU activation1,048,576
Fully-connected4096 units, ReLU activation16,777,216
Output (Softmax)100 units40,960
Table 3. Classification accuracy of the proposed method under various parameter settings ( w . base accuracy 0.720).
Table 3. Classification accuracy of the proposed method under various parameter settings ( w . base accuracy 0.720).
Privacy ParameterNoise Parameter
σ = 3.0 σ = 4.0 σ = 5.0
without L e x p 0.6030.5860.577
ϵ = 0.1 δ = 10 5 0.6200.6110.601
δ = 0.9 0.6230.6120.598
ϵ = 0.9 δ = 10 5 0.6520.6290.624
δ = 0.9 0.6510.6320.620
Table 4. Average SSIM of images where the original Gaussian noise and the proposed image conversion method were applied ( δ = 0.9 , scaled by × 10 2 ).
Table 4. Average SSIM of images where the original Gaussian noise and the proposed image conversion method were applied ( δ = 0.9 , scaled by × 10 2 ).
SchemeGaussianProposed
ϵ = 0.9 ϵ = 0.5 ϵ = 0.1
Target  σ 3.04.05.03.04.05.03.04.05.03.04.05.0
SSIM1.541.291.161.521.211.121.541.301.151.551.281.22
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kim, H.-G.; Shin, J.; Choi, Y.-H. Human-Unrecognizable Differential Private Noised Image Generation Method. Sensors 2024, 24, 3166. https://doi.org/10.3390/s24103166

AMA Style

Kim H-G, Shin J, Choi Y-H. Human-Unrecognizable Differential Private Noised Image Generation Method. Sensors. 2024; 24(10):3166. https://doi.org/10.3390/s24103166

Chicago/Turabian Style

Kim, Hyeong-Geon, Jinmyeong Shin, and Yoon-Ho Choi. 2024. "Human-Unrecognizable Differential Private Noised Image Generation Method" Sensors 24, no. 10: 3166. https://doi.org/10.3390/s24103166

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop