Next Article in Journal
A LiDAR–Inertial SLAM Method Based on Virtual Inertial Navigation System
Next Article in Special Issue
A Deep-Learning-Based Approach to Keystroke-Injection Payload Generation
Previous Article in Journal
Analysis and Suppression of High-Frequency Resonance for Offshore Wind Power Grid-Connected Converter Considering Cable Capacitance Effect
Previous Article in Special Issue
Applications and Technologies of Big Data in the Aerospace Domain
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Low-Pass Filtering Detection on Small Digital Images Using Hybrid Deep Learning

1
Department of Software Convergence, Andong National University, Andong 36729, Gyeongbuk, Republic of Korea
2
Amity School of Engineering & Technology, Amity University Uttar Pradesh, Noida 201313, India
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(12), 2637; https://doi.org/10.3390/electronics12122637
Submission received: 25 April 2023 / Revised: 25 May 2023 / Accepted: 7 June 2023 / Published: 12 June 2023
(This article belongs to the Special Issue Data Driven Security)

Abstract

:
Detecting image manipulation is essential for investigating the processing history of digital images. In this paper, a novel scheme is proposed to detect the use of low-pass filters in image processing. A new convolutional neural network with a reasonable size was designed to identify three types of low-pass filters. The learning experiences of the three solvers were combined to enhance the detection ability of the proposed approach. Global pooling layers were employed to protect the information loss between the convolutional layers, and a new global variance pooling layer was introduced to improve detection accuracy. The extracted features from the convolutional neural network were mapped to the frequency domain to enrich the feature set. A leaky Rectified Linear Unit (ReLU) layer was discovered to perform better than the traditional ReLU layer. A tri-layered neural network classifier was employed to classify low-pass filters with various parameters into two, four, and ten classes. As detecting low-pass filtering is relatively easy on large-dimension images, the experimental environment was restricted to small images of 30 × 30 and 60 × 60 pixels. The proposed scheme achieved 80.12% and 90.65% detection accuracy on ten categories of images compressed with JPEG and a quality factor 75 on 30 × 30 and 60 × 60 images, respectively.

1. Introduction

Digital images are a commonly used medium, especially for informal information sharing. However, image contents can be changed without difficulty by using specific software. Usually, resizing [1], gamma correction [2], sharpening [3], and smoothing [4] operations are used to produce a convincing fake image. Image forensics provides the details of past operations applied to the image. This paper’s primary objective was to detect the low-pass filter operation usually used for smoothing. Three commonly used low-pass filters—Gaussian, median, and averaging—are addressed in this paper. The images displayed in Figure 1 show the visual effect on the image after low-pass filtering and JPEG compression. Uncompressed images are displayed in the first row, and their compressed version of 75 quality factor in the second row. The structural similarity index measure (SSIM) [5] relates to visual quality. The first image in the first row is the reference image, and its assumed SSIM quality value (SQ) is 1.0. In the first row, the second, third, and fourth images are averaging- (SQ = 0.48), Gaussian- (SQ = 0.92), and median-filtered (SQ = 0.52) images, respectively. The Gaussian-filtered image is created using a Gaussian filter with a standard deviation of 0.5 and a filter window size of 3 × 3. The SQ is highest for the Gaussian-filtered image after the original image. The SQ is related to visual quality. In the second row, compressed images are displayed with their corresponding SQs. The SQ has dropped greatly for a compressed Gaussian-filtered image compared to other filtered images. However, the SQ value of the compressed Gaussian-filtered image is highest in the compressed filtered images.
The means of SSIM quality values (SQ) are displayed in Figure 2 to observe the general behavior of filtered images. The low-pass filters were applied to one thousand images with dimensions 30 × 30 and 60 × 60. ORI denotes the non-filtered images. GA1, GA2, and GA3 indicate the Gaussian-filtered images of filter sizes 3 × 3, 5 × 5, and 7 × 7. Likewise, AG represents averaging-filtered images, and MD denotes median-filtered images. JPEG-compressed images are also considered with quality factor (JQF) 75. It can be observed from the figure that as the filter size is enlarged, the SQ value declines. The SQ values of GA are followed by those for AG and MD.
One more image quality indicator—perception-based image quality evaluator (PIQUE) [6]—is considered for analysis. PIQUE is a no-reference quality metric for blind image quality assessment. The covariance plot of the PIQUE score is displayed in Figure 3 to reveal the difference in nature of non-filtered, low-pass-filtered images for all ten classes, including three low-pass filters with different parameters. In Figure 3a, uncompressed images are considered; in Figure 3b, compressed images with JQF = 75 are considered to evaluate the PIQUE score.
As discussed above, the low-pass-filtered images with low-potency filters, such as filter size 3 × 3, have less difference in SQ compared to non-filtered images. Researchers suggest several schemes to address the issue of low-pass filtering. Kirchner and Fridrich [7] detected median filtering by measuring the co-occurrence between pixels. Similarly, Chen et al. [8] and Agarwal & Jung [4] measured the co-occurrence of pixels using some additional modifications. Chen et al. [8] also added local features by calculating local correlation. Agarwal and Jung [4] utilized the higher-significance bit-plane to improve the detection capability on low-pass-filtered images. Gupta and Singhal [9] exploited the discrete cosine domain to extract four features of first-order moments. The autoregressive model detects filtering in schemes [10,11,12]. Kang et al. [10] considered the model of order ten to detect median filtering. Yang et al. [11] applied the 2D model on several types of difference arrays for better results. Rhee [12] considered the bit planes to boost performance using the autoregressive model. Chen et al. [13] introduced the deep neural network to recognize median-filtered images. The network had five convolutional layers, and median-filter residual images were considered for training. The deep neural network of Liu et al. [14] utilized the discrete Fourier transform domain images to detect low-pass filtering. The network had merely two convolutional layers and pooling layers. The Zhang et al. [15] deep neural network contained squeeze and excitation blocks to detect three types of low-pass filtering. The computational cost of the Zhang et al. network is expensive. Yu et al. [16] preprocessed the images with several high-pass filters before sending them to a deep neural network with four convolutional layers. The co-occurrence and autoregressive model extracted the prevailing manual scheme features. In the prevailing convolutional neural network (CNN)-based schemes, altered arrangements of layers are performed to improve the performance. However, deep neural network schemes provide better results than the prevailing manual schemes. In this paper, a hybrid scheme is proposed that has a deep neural network at its core. The proposed scheme has superior recognition competency and reasonable computational expense. The key highlights of the proposed scheme are as follows:
  • The proposed scheme is suitable for detecting various types and strengths of low-pass filters. The scheme is suitable for large- and small-dimension uncompressed and JPEG-compressed images.
  • The convolutional neural network of the proposed scheme comprises merely nine convolutional layers with a lesser number of kernels.
  • The global variance pooling layer is introduced to improve the detection performance of the deep neural network of the proposed scheme.
  • Feature arrays from trained networks are combined to boost detection accuracy. The network is trained using three different solvers.
  • The spatial domain and discrete Fourier transform domain features are concatenated to enhance the potential of the proposed scheme.
  • A leaky ReLU layer is employed instead of a conventional ReLU layer to optimize the results.
  • A tri-layered neural network classifier is employed to classify a different set of images that are processed by low-pass filters.
The remainder of the paper is structured as follows: Section 2 describes the suggested technique. Multiple features are fetched in the proposed scheme using two global pooling layers and three optimizers. Spatial domain features are also mapped in the frequency domain. Both domain features are concatenated for better results. The experiment’s findings for the proposed scheme and the paper’s conclusions are discussed in Section 3 and Section 4, respectively.

2. Proposed Scheme

Low-pass filtering is applied for smoothing and blurring the image. In this paper, a robust scheme is suggested to detect low-pass filtering. The proposed scheme incorporates the features of spatial and frequency domains by using three different solvers in a convolutional neural network. First, the proposed deep neural network is explained. It is found from the regressive experimental analysis that for low-pass filtering detection, a light neural network provides better results than a CNN that has residual connections and a high number of parameters. The ResNet [17] and other computationally expensive CNNs do not provide satisfactory detection performance. Figure 4 presents a proposed deep neural network containing nine two-dimensional (2D) convolutional (Conv) layers. A batch normalization (Batch Norm) layer and a leaky ReLU layer trail each 2D convolutional layer. A leaky ReLU layer is used instead of the conventional ReLU layer in the proposed convolutional neural network to improve detection capability. In a leaky ReLU layer, every negative element is multiplied by a constant real positive number (c). The elements greater than or equivalent to zero stay the same. The leaky ReLU layer (LR) can be defined as follows:
L R n = n , n 0 c × n , n < 0
However, every negative element is assigned a zero value, and positive elements remain unchanged in the conventional ReLU layer. ReLU layer (R) can be defined as follows:
R n = n , n 0 0 , n < 0
There are 96, 48, 24, 96, 48, 24, 96, 48, and 24 kernels in the head 2D convolutional layer to the last convolutional layer. Several combinations of layers and the number of kernels were tried. However, the suggested convolutional neural network was found more suitable. No other pooling layers were used in the network to protect against the loss of information between the convolutional layers. The global pooling layers were used to amplify the performance. Global average pooling (GAP) and global variance pooling (GVP) layers were considered.
Subsequently, in the last leaky ReLU layer, global pooling layers were used to boost the detection capability. In some previous schemes [18,19], the global average pooling layer enhanced the stego image detection accuracy. The proposed scheme also introduces a global variance pooling layer to make the robust feature array. Figure 5 displays layouts of the global average and variance pooling layers. Instead of calculating the average of the feature map, variance is calculated in the global variance pooling layer.
The CNN was trained twice, once using the GAP layer and once using the GVP layer. Further, an activation function was applied, as displayed in Figure 6, to extract the feature arrays. The feature arrays i and j were concatenated. The variance provides the additional details of low-pass-filtering traces that assist in distinguishing non-filtered and filtered images.
Three different solvers were used in network training to strengthen the feature array. These solvers are adaptive moment (Adam), root mean square propagation (RMSprop), and stochastic gradient descent with moment (Sgdm). The activation function was applied to the trained network, as displayed in Figure 7. Feature arrays were fetched from the trained networks to acquire the benefit of three solvers. Each feature array was also converted into the DFT domain. The conversion of features in the DFT domain is beneficial in steganalysis by Wang et al. [20]. This fact is also found relevant in the experimental analysis of our work. This process created six feature arrays. In the proposed CNN, two types of global pooling layers were used, one layer at a time. Therefore, a total of twelve feature arrays were formed. These twelve feature arrays were concatenated to generate a single feature array.
Further, a tri-layered neural network classifier was applied to feature arrays to classify non-filtered and low-pass-filtered images. A tri-layered neural network classifier provided better results than softmax, support vector machine, and linear discriminant analysis classifiers in our experimental analysis. Fifty epochs were engaged to train the network. Mini batch sizes of 300 and 150 were deliberated separately for images with dimensions 60 × 60 and 30 × 30. The proposed work was performed on a system with the following hardware details: NVIDIA GeForce RTX3070, Intel(R) Core (TM) i7-10700K CPU @ 3.80 GHz, and 32 GB RAM. The network learning duration is displayed in Table 1 with Adam, RMSprop, and Sgdm solvers for ten class categorizations. In the case of uncompressed images, learning time is compared to compressed images due to more information content.
In Figure 8, feature arrays are displayed for four classes of non-filtered images and median-filtered images that were processed with filter sizes 3 × 3, 5 × 5, and 7 × 7. The first plot shows the feature arrays of the uncompressed images with dimensions 30 × 30, and the second plot displays the feature arrays of the compressed images with JQF = 75 and size 30 × 30. It can be perceived from Figure 8 that feature arrays of uncompressed images are more distinguishable than compressed images. This distinguishability of feature arrays is also reflected in the experimental analysis, indicating that the detection performance on uncompressed images is superior to compressed images.

3. Experimental Analysis and Discussion

This paper proposes a scheme to detect low-pass filtering in images. An image set was created from three different image datasets, i.e., UCID [21], BOSSBase 1.01 [22], and NRCS [23], to assess the detection scheme. UCID dataset comprises 1338 uncompressed color images with dimensions 512 × 384. The BOSSBase 1.01 dataset has ten thousand uncompressed grayscale images with dimensions 512 × 512. Two thousand high-resolution images were taken from an NRCS photo gallery containing landscapes, soil, etc. Thirty-five thousand images of size 60 × 60 were generated by cropping images in random order. The images with dimensions 30 × 30 were generated by selecting an interior chunk of images with dimensions 60 × 60. Respective low-pass filtered images were generated by applying a particularly low-pass filter. In the case of training the network, twenty thousand images in the required category were taken. Fifteen thousand images were employed for training. Uncompressed and JPEG-compressed images of quality factor 75 were considered for rigorous experimental analysis. JPEG compression was applied after filtering.
The proposed scheme is denoted as Z1, Z2, and Z3 while using Sgdm, Adam, and RMSprop, respectively, as solvers. ZALL represents the final scheme by concatenating the feature arrays of Z1, Z2, and Z3. ORI stands for unfiltered images. Both spatial and frequency domain features were concatenated for better results. The Gaussian-filtered images of sizes 3 × 3, 5 × 5, and 7 × 7 are designated as GA1, GA2, and GA3, respectively. In the same way, AG stands for averaging-filtered images, and MD for median-filtered images. Two, four, and ten categories were classified using a tri-layered neural network classifier for in-depth analysis.
Table 2 and Table 3 display the detection accuracy of two-class classification. The detection accuracy of the four-class classification is shown in Table 4 and Table 5. The detection accuracy of the ten-class classification is given in Table 6 and Table 7.
In Table 2, the classification accuracy is shown between non-filtered (ORI) and filtered images, i.e., between two classes. The classification accuracy is more than 99%, except for the ORI-GA1 pair (84.72%) for 30 × 30 uncompressed images. Images with dimensions 60 × 60 follow a similar pattern; except for the ORI-GA1 pair (91.87%), the classification accuracy is greater than 99%. Therefore, the results for the 60 × 60 uncompressed images are not displayed in the table.
In Table 3, outcomes for compressed images with JQF = 75 and dimensions of 30 × 30 or 60 × 60 are exhibited. Gaussian-filtered images GA1, GA2, and GA3 are created using the 0.5, 1.0, and 1.5 standard deviations; due to this, ORI-GA1 has the least detection accuracy. ZALL provides 93.06% classification accuracy on ORI-MD1 pairs on the 30 × 30 images. In the case of the 3 × 3 filter, the ORI-AG1 pair has the highest detection accuracy, i.e., 96.49%, compared to pairs ORI-GA1 and ORI-MD1. For low-pass filters of size 5 × 5, the detection accuracies are 96.26%, 99.39%, and 98.87% for ORI-GA2, ORI-AG2, and ORI-MD2, respectively. The detection accuracies for low-pass filters of size 7 × 7 are nearly 99%.
In Table 4, detection accuracies are demonstrated for four-class classification in the case of the Gaussian filter using the proposed scheme on uncompressed images. ORI, GA1, GA2, and GA3 are the four classes. Detection accuracies are displayed for individuals and four classes (overall). Most of the misclassification is in the ORI and GA1 classes. However, the ZALL delivers 95.71% and 98.85% overall classification accuracy for the 30 × 30 and 60 × 60 images, respectively. The median- and averaging-filtered results are not included in Table 4. Regarding median and averaging filtering, ZALL delivers 98.89% and 99.93% overall classification accuracy, respectively, for 30 × 30 images. The general classification accuracy of median and averaging filtering is more than 99.61% and 99.97%, respectively, for the 60 × 60 images.
In Table 5, four-class classification was executed using the proposed scheme on compressed images of JQF = 75. Overall detection accuracy on test images is 81.62% and 89.25% in the case of Gaussian-filtered images for the 30 × 30 and 60 × 60 images, respectively. The overall detection accuracy for averaging-filtered images is 96.85% for the 30 × 30 images and 99.02% for the 60 × 60 images. Overall classification accuracy for median-filtered images is 92.09% for the 30 × 30 images and 97.81% for the 60 × 60 images.
ZALL provides the highest detection accuracy, as shown in Table 6, on uncompressed images of dimensions 30 × 30 or 60 × 60 for ten-class classification. In ten classes, one is ORI class; three are Gaussian-filtered, three are averaging-filtered, and three are median-filtered class. Images were classified simultaneously using a tri-layered neural network classifier. The overall detection accuracies are 97.83% and 99.26% on the 30 × 30 and 60 × 60 images, respectively. On observing the misclassified images, the number of misclassified images is highest between ORI and GA1, as the GA1 variance is only 0.5.
Table 7 displays results for ten-class classification on compressed images (JQF = 75) of dimensions 30 × 30 or 60 × 60. ZALL also delivers the maximum number of correct predictions. The average detection accuracies are 79.25% and 89.05% on the 30 × 30 and 60 × 60 images, respectively. There is approximately a ten percent difference between detection accuracies for the 30 × 30 and 60 × 60 images. The environment is challenging, as the same filter with a different configuration is also included with other filters.
The proposed scheme was compared with some other prevalent schemes [8,10,11,13,14,15] for ten-class classification overall detection accuracy. The schemes [8,10,11] rely on manual feature fetching. The scheme in [8] is well known as GLF. The features are fetched by co-occurrence and correlation within the pixels. The schemes [10,11] exploit the autoregressive model. The results of deep neural network schemes are far better than manual schemes, as apparent from Table 8. The techniques [13,14,15] originated using a deep learning network. Chen et al. [13] utilized the five convolutional layers and considered the residual images and pooling layers. Liu et al. [14] considered the frequency domain, max-pooling layers, and only two convolutional layers. Zhang et al. [15] used squeeze and excitation blocks. In the proposed scheme, multiple measures are applied to improve the detection accuracy. The suggested design adds the global variance pooling layer to enhance the deep neural network’s detection capabilities. The spatial and discrete Fourier transform domain features are concatenated to enrich the feature array. Feature arrays from three trained networks are combined, and a leaky ReLU layer is used instead of a traditional ReLU layer to improve the outcomes. Therefore, the results of the proposed scheme are better. The proposed scheme results are superior to Zhang et al. [15] by at least three percent. The presented scheme results are also much higher than results from other schemes [8,10,11,13,14].

4. Conclusions

Image manipulation has become increasingly prevalent on web platforms, with various operations used to generate fake images, including low-pass filtering. In this paper, a hybrid scheme has been proposed to detect low-pass filtering in small digital images. The proposed approach used a deep neural network with three different solvers and a leaky ReLU layer to improve classification accuracy. We also introduced a global variance pooling layer to boost performance. In addition, frequency domain features were utilized to generate a more robust feature array. A tri-layered neural network classifier was used to classify Gaussian, median, and averaging filters with diverse parameters. The experimental results of the proposed scheme are superior to several other existing techniques. The proposed scheme acquired 81.62%, 96.85%, and 92.05% detection accuracies for Gaussian-, averaging-, and median-filtered JPEG-compressed 30 × 30 images on four-class classification.

Author Contributions

Conceptualization, S.A. and K.-H.J.; software, S.A.; validation, K.-H.J.; formal analysis, S.A.; investigation, S.A.; resources, S.A.; data curation, K.-H.J.; writing—original draft preparation, S.A.; writing—review and editing, K.-H.J.; visualization, S.A.; supervision, K.-H.J.; project administration, K.-H.J.; funding acquisition, K.-H.J. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by a Research Grant from Andong National University.

Data Availability Statement

The datasets used in this paper are publicly available, and their links are provided in the reference section.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Danon, D.; Arar, M.; Cohen-Or, D.; Shamir, A. Image resizing by reconstruction from deep features. Comput. Vis. Media 2021, 7, 453–466. [Google Scholar] [CrossRef]
  2. Sengupta, D.; Biswas, A.; Gupta, P. Non-linear weight adjustment in adaptive gamma correction for image contrast enhancement. Multimed. Tools Appl. 2021, 80, 3835–3862. [Google Scholar] [CrossRef]
  3. Hu, D.; Tan, J.; Zhang, L.; Ge, X. Image deblurring based on enhanced salient edge selection. Vis. Comput. 2023, 39, 281–296. [Google Scholar] [CrossRef]
  4. Agarwal, S.; Jung, K.-H. HSB-SPAM: An Efficient Image Filtering Detection Technique. Appl. Sci. 2021, 11, 3749. [Google Scholar] [CrossRef]
  5. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image Quality Assessment: From Error Visibility to Structural Similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Venkatanath, N.; Praneeth, D.; Bh, M.C.; Channappayya, S.S.; Medasani, S.S. Blind image quality evaluation using perception based features. In Proceedings of the 2015 Twenty First National Conference on Communications (NCC), Mumbai, India, 27 Feburary–1 March 2015; pp. 1–6. [Google Scholar] [CrossRef] [Green Version]
  7. Kirchner, M.; Fridrich, J. On detection of median filtering in digital images. In Media Forensics and Security II; SPIE: Bellingham, DC, USA, 2010; Volume 7541, p. 754110. [Google Scholar] [CrossRef] [Green Version]
  8. Chen, C.; Ni, J.; Huang, J. Blind Detection of Median Filtering in Digital Images: A Difference Domain Based Approach. IEEE Trans. Image Process. 2013, 22, 4699–4710. [Google Scholar] [CrossRef] [PubMed]
  9. Gupta, A.; Singhal, D. A simplistic global median filtering forensics based on frequency domain analysis of image residuals. ACM Trans. Multimed. Comput. Commun. Appl. 2019, 15, 1–23. [Google Scholar] [CrossRef]
  10. Kang, X.; Stamm, M.C.; Peng, A.; Liu, K.J.R. Robust median filtering forensics using an autoregressive model. IEEE Trans. Inf. Forensics Secur. 2013, 8, 1456–1468. [Google Scholar] [CrossRef] [Green Version]
  11. Yang, J.; Ren, H.; Zhu, G.; Huang, J.; Shi, Y.-Q. Detecting median filtering via two-dimensional AR models of multiple filtered residuals. Multimed. Tools Appl. 2018, 77, 7931–7953. [Google Scholar] [CrossRef]
  12. Rhee, K.H. Forensic Detection Using Bit-Planes Slicing of Median Filtering Image. IEEE Access 2019, 7, 92586–92597. [Google Scholar] [CrossRef]
  13. Chen, J.; Kang, X.; Liu, Y.; Wang, Z.J. Median Filtering Forensics Based on Convolutional Neural Networks. IEEE Signal Process. Lett. 2015, 22, 1849–1853. [Google Scholar] [CrossRef]
  14. Liu, A.; Zhao, Z.; Zhang, C.; Su, Y. Smooth filtering identification based on convolutional neural networks. Multimed. Tools Appl. 2019, 78, 26851–26865. [Google Scholar] [CrossRef]
  15. Zhang, Y.; Yu, L.; Fang, Z.; Xiong, N.N.; Zhang, L.; Tian, H. An end-to-end deep learning model for robust smooth filtering identification. Futur. Gener. Comput. Syst. 2022, 127, 263–275. [Google Scholar] [CrossRef]
  16. Yu, L.; Zhang, Y.; Han, H.; Zhang, L.; Wu, F. Robust Median Filtering Forensics by CNN-Based Multiple Residuals Learning. IEEE Access 2019, 7, 120594–120602. [Google Scholar] [CrossRef]
  17. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; Volume 2016, pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  18. Yedroudj, M.; Comby, F.; Chaumont, M. Yedroudj-Net: An Efficient CNN for Spatial Steganalysis. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 2092–2096. [Google Scholar] [CrossRef] [Green Version]
  19. Xu, G. Deep convolutional neural network to detect J-UNIWARD. In Proceedings of the 5th ACM Workshop Inf. Hiding Multimedia Security (IH&MMSec), Philadelphia, PA, USA, 20–22 June 2017; pp. 67–73. [Google Scholar] [CrossRef] [Green Version]
  20. Wang, P.; Liu, F.; Yang, C. Towards feature representation for steganalysis of spatial steganography. Signal Process. 2020, 169, 107422. [Google Scholar] [CrossRef]
  21. Schaefer, G.; Stich, M. UCID: An uncompressed color image database. In Storage and Retrieval Methods and Applications for Multimedia 2004; SPIE: Bellingham, DC, USA, 2003; Volume 5307, pp. 472–480. [Google Scholar] [CrossRef]
  22. Bas, P.; Filler, T.; Pevný, T. Break Our Steganographic System: The Ins and Outs of Organizing BOSS. In Lecture Notes in Computer Science (including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Berlin/Heidelberg, Germany, 2011; Volume 6958, pp. 59–70. [Google Scholar]
  23. NRCS Photo Gallery Database. Available online: https://photogallery.sc.egov.usda.gov/photogallery/#/ (accessed on 21 February 2022).
Figure 1. Uncompressed and compressed low-pass-filtered images.
Figure 1. Uncompressed and compressed low-pass-filtered images.
Electronics 12 02637 g001
Figure 2. SQ values for uncompressed and compressed images.
Figure 2. SQ values for uncompressed and compressed images.
Electronics 12 02637 g002
Figure 3. Covariance plot of PIQUE score on ten categories of images. (a) Uncompressed images; (b) compressed images.
Figure 3. Covariance plot of PIQUE score on ten categories of images. (a) Uncompressed images; (b) compressed images.
Electronics 12 02637 g003
Figure 4. Design of the proposed convolutional neural network.
Figure 4. Design of the proposed convolutional neural network.
Electronics 12 02637 g004
Figure 5. GAP and GVP layers.
Figure 5. GAP and GVP layers.
Electronics 12 02637 g005
Figure 6. Features fetched while using GAP and GVP layer.
Figure 6. Features fetched while using GAP and GVP layer.
Electronics 12 02637 g006
Figure 7. Features fetched from three different solvers and converted into the DFT domain.
Figure 7. Features fetched from three different solvers and converted into the DFT domain.
Electronics 12 02637 g007
Figure 8. Feature arrays of uncompressed and compressed images for the median filter.
Figure 8. Feature arrays of uncompressed and compressed images for the median filter.
Electronics 12 02637 g008
Table 1. CNN learning duration in minutes.
Table 1. CNN learning duration in minutes.
Image Dimensions/SolverAdamRMSpropSgdm
30 × 30121.12104.67125.92
60 × 60198.29213.81219.35
30 × 30 (JQF = 75)52.6153.6264.88
60 × 60 (JQF = 75)163.86170.92154.16
Table 2. Two-class classification for 30 × 30 uncompressed images.
Table 2. Two-class classification for 30 × 30 uncompressed images.
Image Dimensions 30 × 30
Image PairORI-GA1ORI-GA2ORI-GA3ORI-AG1ORI-AG2ORI-AG3ORI-MD1ORI-MD2ORI-MD3
Z183.7199.7499.9499.9499.6999.8698.9799.3499.76
Z283.6899.7199.9199.9099.6799.8398.9499.3299.73
Z383.7099.7399.9499.9299.7099.3298.5499.3599.65
ZALL84.7299.7499.9399.9299.7099.8799.8799.3499.75
Table 3. Two-class classification of compressed images.
Table 3. Two-class classification of compressed images.
Image Dimensions 30 × 30 JQF = 75
Image PairORI-GA1ORI-GA2ORI-GA3ORI-AG1ORI-AG2ORI-AG3ORI-MD1ORI-MD2ORI-MD3
Z172.4095.2598.9695.7099.0799.2891.7697.6598.17
Z273.0095.2398.7896.1299.2699.1191.6797.9297.65
Z372.0795.7598.3895.7098.8399.3091.9598.0398.52
ZALL74.0796.2699.3296.4999.3999.4593.0698.8798.70
Image dimensions 60 × 60 JQF = 75
Image pairORI-GA1ORI-GA2ORI-GA3ORI-AG1ORI-AG2ORI-AG3ORI-MD1ORI-MD2ORI-MD3
Z180.5398.1599.6097.6399.8199.6397.1099.3299.56
Z280.4397.9199.4697.4899.5799.5296.9599.1799.85
Z380.5898.1199.7497.6499.7399.7497.2499.3299.68
ZALL81.6398.8999.8398.4599.7499.7998.2399.4599.98
Table 4. Four-class classification on uncompressed images for Gaussian filtering.
Table 4. Four-class classification on uncompressed images for Gaussian filtering.
ClassImage Dimensions 30 × 30Image Dimensions 60 × 60
ORIGA1GA2GA3OverallORIGA1GA2GA3Overall
Z187.2192.0397.2999.1493.9294.8696.6799.1999.4997.55
Z288.2593.2995.6999.6394.2196.0997.2398.6899.8397.96
Z390.7589.6797.2899.4994.3095.3196.5599.3399.5097.67
ZALL92.2092.0698.9999.6195.7197.5098.2299.9699.7398.85
Table 5. Four-class classification of compressed images.
Table 5. Four-class classification of compressed images.
Image Dimensions 30 × 30 JQF = 75Image Dimensions 60 × 60 JQF = 75
ClassORIGA1GA2GA3OverallClassORIGA1GA2GA3Overall
Z162.0571.9185.6597.5079.28Z177.4083.2589.9199.5087.51
Z268.6368.9285.6094.8679.50Z269.6688.2890.3599.3986.92
Z364.7573.9886.4794.8380.01Z377.5983.1791.9799.2687.99
ZALL70.4771.4786.8797.6981.62ZALL81.2883.5393.5498.6589.25
ClassORIAG1AG2AG3OverallClassORIAG1AG2AG3Overall
Z193.2794.8697.4097.3195.71Z197.6397.8198.4099.5598.34
Z292.1194.4996.0997.6595.09Z297.3397.8598.6999.4298.32
Z393.6395.0696.2398.1195.76Z397.0398.3597.9999.6998.26
ZALL94.9995.7997.8398.8096.85ZALL98.6998.8698.9199.6199.02
ClassORIMD1MD2MD3OverallClassORIMD1MD2MD3Overall
Z190.0689.3992.6689.9190.51Z194.2597.0196.7197.1796.29
Z286.6491.1090.8594.2390.71Z295.7396.3196.3598.2296.65
Z391.9985.6691.2694.3190.80Z396.7995.5696.5398.1396.75
ZALL90.7790.2091.5195.8992.09ZALL96.8397.6797.7598.9997.81
Table 6. Ten-class classification of uncompressed images.
Table 6. Ten-class classification of uncompressed images.
Image Dimensions 30 × 30
ClassORIMD1MD2MD3GA1GA2GA3AG1AG2AG3Overall
Z191.8499.0397.1198.5385.6197.8298.0794.6799.2799.4396.14
Z291.5299.0998.4897.1593.8597.0398.9696.4098.8397.8696.92
Z390.7198.0597.8398.1692.1996.5398.2196.1799.5399.0596.64
ZALL92.1799.4198.4598.8794.7398.0199.1298.0899.6999.7297.83
Image dimensions 60 × 60
Z193.6999.5499.2599.5297.8198.7499.6699.0799.5699.3798.62
Z295.7299.5499.4699.2896.3199.1199.7398.9999.1299.6398.69
Z397.4399.2098.8499.4181.6998.5998.7898.8599.3799.8497.20
ZALL97.5999.7799.5699.6097.9399.2699.7599.4699.8499.8499.26
Table 7. Ten-class classification of compressed images.
Table 7. Ten-class classification of compressed images.
Image dimensions 30 × 30 JQF = 75
ClassORIMD1MD2MD3GA1GA2GA3AG1AG2AG3Overall
Z169.5278.9986.1188.1157.2256.0685.1863.4692.6197.3477.46
Z266.4879.7081.4390.8563.9362.4784.9159.9991.7296.4977.80
Z373.7480.0984.9789.4153.8360.5786.3560.5089.8395.9177.52
ZALL70.1180.2884.4190.7963.1461.3187.2564.9392.8397.4979.25
Image dimensions 60 × 60 JQF = 75
Z178.1988.9594.5494.8576.7487.6196.9262.5992.3998.7187.15
Z280.5791.5493.4397.3972.0977.8795.0173.5497.4599.4687.83
Z380.6693.3994.2597.3672.9967.5685.9782.7097.6798.8987.14
ZALL80.9792.4294.3597.6978.5378.4094.9677.0296.8599.3589.05
Table 8. Proposed scheme compared with other schemes.
Table 8. Proposed scheme compared with other schemes.
Image
Dimensions
Kang et al. [10]Chen et al. [8]Yang et al. [11]Chen et al. [13]Liu et al. [14]Zhang et al. [15]Proposed Scheme
Uncompressed images
30 × 3058.2867.4869.3976.8581.7994.1697.83
60 × 6068.8883.9078.3483.9791.8796.3399.26
Compressed images with JQF = 75
30 × 3051.3155.7253.3358.3968.4475.3779.25
60 × 6062.2265.6074.1070.3968.7283.2189.05
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Agarwal, S.; Jung, K.-H. Enhancing Low-Pass Filtering Detection on Small Digital Images Using Hybrid Deep Learning. Electronics 2023, 12, 2637. https://doi.org/10.3390/electronics12122637

AMA Style

Agarwal S, Jung K-H. Enhancing Low-Pass Filtering Detection on Small Digital Images Using Hybrid Deep Learning. Electronics. 2023; 12(12):2637. https://doi.org/10.3390/electronics12122637

Chicago/Turabian Style

Agarwal, Saurabh, and Ki-Hyun Jung. 2023. "Enhancing Low-Pass Filtering Detection on Small Digital Images Using Hybrid Deep Learning" Electronics 12, no. 12: 2637. https://doi.org/10.3390/electronics12122637

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop