Next Article in Journal
Multi-Objective Green Closed-Loop Supply Chain Management with Bundling Strategy, Perishable Products, and Quality Deterioration
Previous Article in Journal
Distributed Bootstrap Simultaneous Inference for High-Dimensional Quantile Regression
Previous Article in Special Issue
Pooled Steganalysis via Model Discrepancy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bio-Inspired Watermarking Method for Authentication of Fundus Images in Computer-Aided Diagnosis of Retinopathy

by
Ernesto Moya-Albor
1,*,†,
Sandra L. Gomez-Coronel
2,†,
Jorge Brieva
1,*,† and
Alberto Lopez-Figueroa
1,†
1
Facultad de Ingeniería, Universidad Panamericana, Augusto Rodin 498, Ciudad de México 03920, Mexico
2
Instituto Politécnico Nacional, UPIITA, Departamento de Ingeniería, Av. IPN No. 2580, Col. La Laguna Ticomán, Ciudad de México 07340, Mexico
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Mathematics 2024, 12(5), 734; https://doi.org/10.3390/math12050734
Submission received: 1 February 2024 / Revised: 24 February 2024 / Accepted: 25 February 2024 / Published: 29 February 2024
(This article belongs to the Special Issue Data Hiding, Steganography and Its Application)

Abstract

:
Nowadays, medical imaging has become an indispensable tool for the diagnosis of some pathologies and as a health prevention instrument. In addition, medical images are transmitted over all types of computer networks, many of them insecure or susceptible to intervention, making sensitive patient information vulnerable. Thus, image watermarking is a popular approach to embed copyright protection, Electronic Patient Information (EPR), institution information, or other digital image into medical images. However, in the medical field, the watermark must preserve the quality of the image for diagnosis purposes. In addition, the inserted watermark must be robust both to intentional and unintentional attacks, which try to delete or weaken it. This work presents a bio-inspired watermarking algorithm applied to retinal fundus images used in computer-aided retinopathy diagnosis. The proposed system uses the Steered Hermite Transform (SHT), an image model inspired by the Human Vision System (HVS), as a spread spectrum watermarking technique, by leveraging its bio-inspired nature to give imperceptibility to the watermark. In addition, the Singular Value Decomposition (SVD) is used to incorporate the robustness of the watermark against attacks. Moreover, the watermark is embedded into the RGB fundus images through the blood vessel patterns extracted by the SHT and using the luma band of Y’CbCr color model. Also, the watermark was encrypted using the Jigsaw Transform (JST) to incorporate an extra level of security. The proposed approach was tested using the image public dataset MESSIDOR-2, which contains 1748 8-bit color images of different sizes and presenting different Diabetic Retinopathy (DR). Thus, on the one hand, in the experiments we evaluate the proposed bio-inspired watermarking method over the entire MESSIDOR-2 dataset, showing that the embedding process does not affect the quality of the fundus images and the extracted watermark, by obtaining average Peak Signal-to-Noise Ratio (PSNR) values higher to 53 dB for the watermarked images and average PSNR values higher to 32 dB to the extracted watermark for the entire dataset. Also, we tested the method against image processing and geometric attacks successfully extracting the watermarking. A comparison of the proposed method against state-of-the-art was performed, obtaining competitive results. On the other hand, we classified the DR grade of the fundus image dataset using four trained deep learning models (VGG16, ResNet50, InceptionV3, and YOLOv8) to evaluate the inference results using the originals and marked images. Thus, the results show that DR grading remains both in the non-marked and marked images.

1. Introduction

The World Health Organization (WHO) classifies visual impairment into two groups (International Classification of Diseases 11 (2018)): visual impairment of distance and near presentation. For the distance vision disability group, blindness is considered when the person has a visual acuity of less than 3/60, considering a normal visual acuity of 20/20 [1].
Globally, at least 2.2 billion people have near or far vision impairment, whereas in at least 1 billion people the vision impairment could have been prevented or has not yet been treated. In addition, of these 1 billion people, about 3.9 million people have Diabetic Retinopathy (DR) as the main cause of their visual impairment [1].
In general, the term retinopathy is a generic term used to refer to any non-inflammatory disease affecting the retina, being used to group a set of different conditions, each with its characteristics. The most common retinopathies are retinopathy of prematurity, diabetic retinopathy, hypertensive retinopathy, and central serous retinopathy [2].
In Latin America and the Caribbean, there is a high prevalence of diabetes, in 2015 it was estimated that 29.6 million people were living with diabetes. In addition, it is estimated that more than 75% of patients who have had diabetes mellitus for more than 20 years will have some form of diabetic retinopathy, which is responsible for 2.6% of blindness worldwide. On the other hand, it is estimated that after 15 years of diabetes, approximately 2% of patients will become blind and 10% will develop severe visual impairment [3,4]. Particularly, it is estimated that in Mexico there is a 71% of DR incidence in diabetic patients [5].
Currently, some treatments can significantly reduce the risks of blindness and moderate vision loss by more than 90% [3]. In this sense, fluor angiography is one of the techniques that has contributed to the understanding, diagnosis, and treatment of many chorioretinal diseases [6], which consists of the intravenous administration of a dye substance called fluorescein and photographs are taken, using a special camera, of the back of the eye to assess how the dye flows in the arteries, capillaries, and veins of the inner part of the eye, known as fundus images. However, there is the possibility of developing adverse reactions such as urticaria, fever, and chills that correspond to individual susceptibility [6]. On the other hand, retinography is a lower cost non-invasive diagnostic technique that does not use contrast agents in the acquisition process avoiding possible reactions. This technique produces RGB fundus images that can be used to detect some of the most distinctive structures that characterize retinopathies, such as neovascularization, hemorrhages, exudates, and microaneurysms [7].
Fundus imaging is a process in which the 3D structure of the retina is projected onto the 2D plane. The intensity of the image represents the amount of reflected light. The fundus camera consists of a low-power microscope and a camera attached to the top of the microscope. The camera can capture the retinal area at an angle of 30° to 50° with a magnification of 2.5× (5× using auxiliary lenses). Color filters, fluorescein, and indocyanine green dyes are used to obtain the fundus image. There are mainly three modalities for fundus photography of the retina [8]:
  • Fundus photography (red-free): A wavelength band is used to capture the amount of reflected light.
  • RGB fundus photography: The red (R), green (G), and blue (B) wavelength bands are used to capture the amount of reflected light.
  • Fluorescein angiography (fluorangiography) and indocyanine: The image is generated from the amount of photons emitted by the fluorescein and indocyanine dyes injected into the patient.
Another area of study in recent years is the vulnerability of patient digital information, hence, copyright protection. For example, steganography and image watermarking have been widely used to hide digital information in a cover image. In the case of watermarking, several methods have been developed with the objective of these methods to insert information into the cover image. Depending on the application, each method uses different tools to achieve its goal and they must consider the requirements to design the algorithm: imperceptibility, robustness, legibility, ambiguity, and security. However, discerning when a watermarking method is focused on medical images is also very important to preserve the diagnosis after, including the watermark and protecting the watermark. In this case, generally, the watermark includes is information about the patient, i.e., the Electronic Patients Record (EPR), but in some cases, authors use QR codes, logos, etc., as a watermark. So, to achieve these elements, designing an imperceptible watermarking algorithm is usually required. The state-of-the-art indicated that the proposals can be designed in the spatial domain, the transform domain, and in recent years with hybrid techniques. The spatial domain techniques have the advantage of low computational cost, but because pixels are directly modified, the images suffer visible modifications, altering the diagnosis. So it is preferable to design a method in the transform domain or combine different transforms (hybrid method). The most popular transforms employed are the Discrete Fourier Transform (DFT), Discrete Cosine Transform (DCT), and Discrete Wavelet Transform (DWT). In addition, when algorithms are focused on medical images, the state-of-the-art indicates classified methods. According to [9], medical image watermarking methods can be classified as Region Of Interest (ROI)-based watermarking, reversible watermarking, and imperceptible watermarking. In [10], the authors indicated the next classification: ROI, reversible watermarking, and zero watermarking. Finally, the authors of [11] indicated that the watermark algorithms for medical images could be classified as Region Of No Interest (RONI), reversible watermarking, and conventional digital watermarking. As we can see, there are algorithms focused on determining the region of interest to insert the watermark. So, different proposals to watermark digital images consider the classification to define their algorithm as a hybrid because they combine it, i.e., the proposals that combine RONI with a reversible schema sometimes are not successful because trying to recover the watermark fails in the manipulated areas by protecting the integrity of the image.
Regarding the watermarking of medical images, a very important point is not only to evaluate a watermarking method in terms of robustness, security, and imperceptibility but also to evaluate the proposal with real medical images and determine if the real diagnosis changes or not. Different watermark algorithms use retinal fundus images and evaluate if the method modified the diagnosis or not. For example, ref. [9] presents an imperceptible watermark method for medical images and the authors include tests demonstrating that it does not affect computer vision-based automated diagnosis of retinal diseases. On the other hand, Dey et al. [12] reported a watermarking algorithm for fundus images by inserting patient information, and evaluated that the embedding watermark process slightly modifies the blood vessel extraction.
In this work, we propose a hybrid watermark method based on the Steered Hermite Transform (SHT), the Singular Value Decomposition (SVD), and the Jigsaw Transform (JST) applied to a public dataset of RGB fundus images presenting a grade of diabetic retinopathy. Thus, the SHT, a bio-inspired image model, is used as a spatial frequency decomposition tool to embed the watermark, providing its imperceptibility; the JST, a popular image scrambling technique, is used to increase the security; and the SVD generates robust watermarks against attacks. Additionally, we evaluate, through a deep learning classification strategy, if the watermarking algorithm applied to these medical images modifies the diagnosis.
The rest of the document is structured as follows: Section 2 presents a revision of the state-of-the-art works, both those for watermarking applied to medical images, in particular to fundus images and those deep learning-based works used as support to the diagnosis of diabetic retinopathy. Section 3 reports the fundus image dataset and the watermark used to test the watermarking approach and used to classify the DR grade through deep learning models. In addition, this section presents an overview of the proposed method and introduces the Steered Hermite transform, the Jigsaw transform, the SVD, and the basis of the Convolutional Neural Networks (CNN). On the other hand, Section 4 describes the proposed bio-inspired watermarking method, reporting both the insertion and extraction processes, as well as the trained deep learning models used to classify the DR grade presented in the images. Later, Section 5 shows first the metrics used to evaluate the performance of the proposal and a sensitivity analysis of the scaling factor. Moreover, this section reports the watermarking performance over the whole fundus image dataset, the robustness against image processing and geometric attacks, a comparison of the proposed watermarking method versus other state-of-the-art works, and an evaluation of the DR diagnosis of both marked and non-marked images. Finally, Section 6 discusses the results of the present work, followed by the conclusions and future work given in Section 7.

2. Related Work

Storing patient information is becoming easier with digital medical records. These records can hold text data about the patient and their medical images, all in one secure digital file. One way to achieve this is by using digital watermarking, where patient information is hidden within the medical images themselves. However, this process must not alter the original medical images, as any changes could affect diagnosis. Therefore, a key measure of success for this technology is to guarantee the original diagnosis accuracy of the images after watermarking them.
In this section, first, we report the recent watermarking algorithms in the transform domain and those applied to fundus images. Secondly, we present the deep learning approaches used for DR classification in this kind of image.

2.1. Watermarking Algorithms

In [13], the authors describe a watermarking method to store and transmit digital fundus images with patient information. The authors used as a watermark the patient data. They calculated the histogram from the medical image, to determine a zero point and then a peak point because the number of pixels that are associated with the peak point is the number of bits that it is possible to insert. Klington et al. [14] developed a watermarking algorithm to authenticate the digital fundus images using SVD and DWT. As a watermark, the authors used textual information about the fundus images and the original image. According to the results, the maximum capacity to insert is 329,960 bits, and with the jittering attack, the algorithm modifies 43% of the total number of embedded pixels. The authors indicate that they used the green channel of the original image to generate the watermark, and the other channels (red and blue) were used to embed the watermark. In [12], Dey et al. presented a watermarking method applied to fundus images to insert an EPR into the blood vessels extracted using K-means segmentation. Additionally, the authors evaluated and found a percentage difference in accuracy (0.25%) in blood vessel extraction before and after watermarking. On the other hand, Singh et al. [15] reported a watermarking approach to insert a unique identification code into the blood vessels. The blood vessels are detected through the matched filter and derivative of Gaussian and the pattern found formed the personal identification code.
In other papers, the authors use all channels. For example, in the paper An Imperceptible Semi-blind Color Image Watermarking Using RDWT and SVD [16], the authors use the RGB color space and employ the three channels to insert the watermark. Also, they used Redundant Discrete Wavelet Transform (RDWT) and SVD. One of the advantages of this method is the amount of information to use as a watermark because the original image and the watermark are the same size. In [17], the authors used a hybrid method to design a robust and imperceptible watermark method for digital images. The authors combine Lifting Wavelet Transform (LWT), Schur Decomposition, and SVD. The watermark employed is a QR code of size 256 × 256 . According to the results, the method has better robustness in comparison with other techniques. A novel hybrid watermarking algorithm is described in [10]. The authors present a brief review of the state-of-the-art and describe their proposal. It consists of a hybrid reversible-zero watermarking to verify the copyright and authenticity of medical images. This kind of scheme has the advantage of no distortion in the medical image because the copyright information is not directly embedded in it. The proposal is focused on a distortion-free method because in this way the accuracy of medical diagnosis is better. The results show that the watermark is recovered efficiently and the authors compare their proposal with zero watermarking schemes and reversible watermarking algorithms.
In [18], the authors describe their method as the first algorithm in which ROI and RONI are divided only for watermark generation, not for watermark embedding. To embed the watermark they use Slantlet Transform (SLT)-SVD and Recursive Dither Modulation (RDM). Their results demonstrate a robust method for Average Filter, Gaussian Blurring, Gaussian Noise, JPEG Compression, Median Filter, Crop, Salt and Noise, Resizing, and Wiener Filter. Another hybrid method is [19]. In this paper, the authors present a method that uses Fast Discrete Curvelet Transform (FDCuT), DCT, SVD, and Arnold cat map. To increase security they employed the Arnold cat map because the original image is distorted to encrypt it and does not change its intensity. The hybrid method is focused on applications in medical images because of their imperceptibility and high resistance. They use different datasets to probe their algorithm, including [20,21]. The problem is the authors do not compare their method with other similar ones, even though their results are good in the insertion process and extraction process. The Peak Signal-to-Noise Ratio (PSNR) is close to 50 dB and the correlation is 1.00. A method to protect biometric images is presented in [22]. This method uses DWT, and SVD with Chaotic encryption. As a watermark, they use fingerprints and gait biometrics (20,000) and 1000 fundus images of different categories as original images, but also the authors probe their algorithm with common images. Their results are good in terms of imperceptibility because when using medical images PSNR = 50.43 dB, and when they used common images PSNR = 52.97 dB. The method was tested with different attacks: Gamma Correction, Resize, Rotate, Crop, Sharp, Speckle Noise, Adjust, and Salt and Pepper Noise, but they do not indicate the parameter values of each attack. Finally, they test their algorithm with different techniques to demonstrate their high imperceptibility.

2.2. Deep Learning Classification Algorithms

On the other hand, there is a lot of research focused on DR detection using Artificial Neural Networks (ANN) or some variants, with some using different techniques. The state-of-the-art is very extensive because the authors try to develop the best technique for DR detection or some similar conditions, for example in [23], Kapoor and Arora presented a set of steps for DR detection and degree classification using deep CNNs. The authors applied the Enhance Local Contrast (CLAHE) method to obtain noise-free images from the Kaggle DR dataset [24], allowing the lesions to be visible, they classified the images using the Deep CNN. On the other hand, Radha et al. [25] presented a study of fundus eyes images, including the normalization of shape and size, segmentation, and automatic retinal lesion classification using the Atrous Convolutional Neural Network (ACNN) to extract the relevant features. Dutta et al. [26] proposed a feature extraction method from retinal images to perform binary and multi-class classification through various machine learning models. The authors used a variant of CNN using a transfer learning approach and hyper-parameter tuning of the VGG-19 model. Gayathri et al. [27] developed an automated DR grading method from the fundus images. A Multipath Convolutional Neural Network (M-CNN) was used for global and local feature extraction from images, and the machine learning classifiers Support Vector Machine (SVM), Random Forest, and J48 were used to categorize the input according to the severity. The model was evaluated across the publicly available databases IDRiD, Kaggle, and MESSIDOR. In [28], the authors Chetoui and Akhloufi presented a study to develop a deep learning algorithm capable of detecting DR on retinal fundus images. Also, the proposed deep learning algorithm fine-tunes a pre-trained deep CNN for DR detection. A new variant of fully convolutional networks, with its expansive path redesigned, is Lesion-Net [29]. A dual loss that leverages both semantic segmentation and image classification losses is introduced. Lastly, the authors have built a multi-task network that employs Lesion-Net as a side-attention branch for both DR grading and result interpretation. In [30], Ni et al. presented a deep convolutional neural network for DR stage classification, trained and evaluated. In addition, the model uses high-resolution retinal fundus images as inputs to take advantage of more detailed retinal lesion information in images and a strong correlation between both eyes. Randive et al. [31], made a model that includes preprocessing, feature extraction using Spherical Directional Local Ternary Pattern (SDLTP), and classification using traditional distance measure and learning-based distance measure using ANN. Also, the SDLTP was used for extracting the directional feature in the 3D plane and to reduce the feature vector length. Loheswaran [32] developed a classification system using Fuzzy C-Means and a Recurrent Neural Network (RNN). Other proposals use different techniques or methods. In [33], Shorfuzzaman et al. presented an explainable deep learning ensemble model by fusing the weights from different models into a single model. It extracts salient features from various retinal lesions found on retinal fundus images. Then, the extracted features were fed to a custom classifier to obtain a DR severity level. The model was trained on the APTOS dataset and was tested using the APTOS, MESSIDOR, and IDRiD datasets. Suresh et al. [34] presented a screening technique that relies on the texture analysis of the retinal background using Local Ternary Patterns (LTP). Also, it compared the results obtained using the proposed approach with Local Binary Patterns (LBP) instead of LTP. They performed three experiments separating, DR from normal, Age-related Macular Degeneration (AMD) from normal, and DR from AMD. Sharif and Shah [35] presented an automatic design for the retinal lesions screening to grade the DR system. Also, the system is comprised of a preprocessing determination of biomarkers and formulation of a profile set for classification. In [36], Wang et al. presented a method requiring only a series of normal and abnormal retinal images without the need to specifically annotate their locations and types. Additionally, the proposed method encodes both the background knowledge of fundus images and the background noise into one unique model. On the other hand, Kaur and Mittal [37] presented a reliable segmentation of lesions that have been performed using iterative clustering irrespective of associated heterogeneity, and bright and faint edges. Afterwards, a computer-aided severity level detection method that they proposed to diagnose non-proliferative diabetic retinopathy. DelaPava et al. [38] designed a model for automatic DR classification on eye fundus images. The approach that has identifies the main ocular lesions related to DR and subsequently diagnoses the illness. Additionally, the Kaggle EyePACS subset is used as a training set, and the MESSIDOR-2 as a test set for lesions and DR classification models. A comprehensive machine learning computer-aided diagnosis (CAD) system based on deep learning techniques [39], eliminates noise, enhances quality, and standardizes the sizes of the retinal images. Also, it distinguishes between healthy and DR cases. Finally, the proposal automatically extracts the four changes: exudates, microaneurysms, hemorrhages, and blood vessels. Biswas et al. [40] developed a model called intelligent system for diabetic retinopathy for early detection of DR using a SVM.

3. Materials and Methods

In this section, the database of retinal fundus images used in the proposed watermarking method is introduced. Moreover, the fundamentals of the SHT, JT, and SVD are given. In addition, the theory of the CNN and the transfer learning strategy are explained in detail.

3.1. Fundus Image Dataset and Watermark Image

The proposed image watermarking and classification method of DR uses the image public dataset MESSIDOR-2 [20,21]. MESSIDOR stands for “Methods to Evaluate Segmentation and Indexing Techniques in the Field of Retinal Ophthalmology” (from french: Méthodes d’Évaluation de Systèmes de Segmentation et d’Indexation Dédiées à l’Ophtalmologie Rétinienne). The main objective of MESSIDOR project is to compare and evaluate segmentation algorithms for detecting lesions in retinal RGB images and to facilitate computer-assisted diagnoses of DR.
The MESSIDOR dataset is a collection of DR examinations, each one consisting of two macula-centered eye fundus RGB images. The MESSIDOR-original dataset was provided by the MESSIDOR program partners, containing 1058 images in PNG format (529 examinations). The MESSIDOR-Extension dataset includes examinations from Brest University Hospital. It contains 690 images in JPEG format (345 examinations).
The MESSIDOR-2 dataset contains 1748 8-bit color images (874 examinations) of different sizes ( 1440 × 960 and 2240 × 1488 pixels). It includes the MESSIDOR-original and the MESSIDOR-Extension. This dataset comes with a list containing image pairing and it does not contain annotations of DR. However, there are third parties that include these annotations. Currently, the MESSIDOR-2 dataset has errors from its site, and the dataset used in this work was downloaded from the Kaggle public dataset MESSIDOR-2: https://www.kaggle.com/datasets/geracollante/messidor2/, (accessed on 10 August 2023).
In addition, in [41], Krause et al. reported the DR grades (0, 1, 2, 3, 4) and Diabetic Macular Edema (DME) presence (0, 1) for the MESSIDOR-2 fundus image dataset, which is publicly available at: https://www.kaggle.com/datasets/google-brain/messidor2-dr-grades, (accessed on 10 August 2023). The grades of the images were adjudicated by a panel of specialists. This is because the authors of the original MESSIDOR-2 dataset did not include any diabetic retinopathy ground truth. Figure 1 shows 25 examples of the MESSIDOR-2 fundus image dataset.
As a watermark, we use the medicine symbol Caduceus, a Greek symbol representing the staff carried by the deity Hermes. The image of the symbol Caduceus used in this work is a gray-scale image of 256 × 256 pixels as is shown in Figure 2:

3.2. Overview of the Proposed Bio-Inspired Watermarking Method

The proposed bio-inspired watermarking algorithm is shown in Figure 3. It is based on the Steered Hermite transform, Singular Value Decomposition, and Jigsaw transform. Our proposal consists of an insertion process to embed a watermark image into the MESSIDOR-2 fundus image dataset and an extraction process to recover the watermark. In addition, a CNN model is used to estimate the DR grading. This parameter allows us to define if the original diagnosis is modified because of the watermark. We detail each method in the following sections.
The watermarking insertion process has as input the color MESSIDOR-2 fundus image dataset I R G B ( x , y , z ) i and the watermark W ( x , y ) , where i = 1 , , 1748 is the image index, x , y represent the spatial coordinates, and z = 1 , 2 , 3 the color band (Red, Green, Blue). Thus, the insertion process generates a set of watermarked fundus images I R G B ( x , y , z ) i . In addition, the watermarking process produces some elements, which are stored in the Key Area (yellow-shaded rectangle) and that will be used later in the extraction process.
On the other hand, the watermarking extraction process recovers the set of extracted watermarks I ^ R G B ( x , y , z ) i . Later, invisibility of the watermark, and robustness against attacks is performed over the watermarked dataset.
Finally, a CNN model is independently trained either with the original image dataset I R G B ( x , y , z ) i (dashed red line) or the watermarked image dataset I ^ R G B ( x , y , z ) i (dash-dotted blue line). Thus, both trained CNN models estimate the DR grading and are compared to evaluate whether the watermarking process modifies the DR diagnosis.

3.3. Steered Hermite Transform

The Steered Hermite transform is obtained by rotating the image decomposition of the Hermite Transform (HT) [42], a bio-inspired image model used for a spatial-frequency decomposition of the digital image. Figure 4 shows the steps to calculate the SHT:
First, the Cartesian Hermite coefficients ( I m , n m ( x , y ) ) are obtained by convoluting a gray-scale image (2D image) ( I ( x , y ) ) with the Hermite analysis filters ( D m , n m ), as is shown in Equation (1), followed by sub-sampling with factor T.
I m , n m ( x 0 , y 0 ) = ( x 0 , y 0 ) S I ( x , y ) D m , n m ( x 0 x , y 0 y ) ,
where m and n m represents the decomposition order in the spatial directions x and y, respectively, with m = 0 n and n = 0 D , D is the maximum order of the expansion, and ( x 0 , y 0 ) is the spatial position in the discrete sampling lattice S.
The Hermite analysis filters ( D m , n m ) are obtained by the Equation (2):
D m , n m ( x , y ) = K m , n m ( x , y ) ω 2 ( x , y ) ,
where ω 2 ( x , y ) represents the 2D version of a binomial window function:
ω ( x ) = 1 2 N C N x , x = 0 , 1 , , N 1 ,
C N x is the binomial function:
C N x = N ! ( N x ) ! x ! , x = 0 , 1 , , N 1 ,
K m , n m are the orthogonal polynomials associated with the binomial window ω 2 ( x , y ) :
K n [ x ] = 1 ( C N n k = 0 n ( 1 ) n k C N x n k C x k ,
and N + 1 is the size of the binomial window, and it is related to the size of the Gaussian window spread ( σ ) , so that defines the discrete sampling lattice S. Moreover, the maximum order of the expansion must fulfill the relationship D 2 N .
Then, the steered Hermite coefficients ( I n , θ ( x , y ) ) are calculated by rotating the Cartesian Hermite coefficients towards an angle θ (Equation (4)):
I n , θ ( x 0 , y 0 ) = k = 0 n I k , n k ( x 0 , y 0 ) ϕ k , n k ( θ ) ,
where ϕ m , n m ( θ ) are the angular functions, of order n, which are defined by Equation (5):
ϕ m , n m ( θ ) = n m cos m θ sin n m θ ,
where θ is an angle of maximum energy, e.g., the gradient angle.
The gradient angle could be approximated using the I 0 , 1 ( x , y ) and I 1 , 0 ( x , y ) Cartesian Hermite coefficients, which represent the edges in the horizontal and vertical directions of the image I ( x , y ) , respectively. Equation (6) shows the approximation of the angle θ :
θ ( x , y ) = t g 1 I 0 , 1 ( x , y ) I 1 , 0 ( x , y ) .
On the other hand, to reconstruct the original image the Inverse Steered Hermite Transform (ISHT) is applied as is shown in Figure 5:
The Cartesian Hermite coefficients ( I m , n m ( x , y ) ) are recovered by inversely rotating the steered Hermite coefficients ( I n , θ ( x , y ) ), by applying Equations (4) and (6) with the stored angles ( θ ( x , y ) ).
Next, the original image is reconstructed by calculating the Inverse Hermite Transform (IHT) through Equation (7). Before performing the reconstruction, an over-sampling with factor T is applied to the Cartesian Hermite coefficients.
I ( x , y ) = n N m = 0 n ( x 0 , y 0 ) S I m , n m ( x 0 , y 0 ) P m , n n ( x x 0 , y y 0 ) ,
where P m , n n are the Hermite synthesis filters given by Equation (8):
P m , n m ( x , y ) = D m , n m ( x , y ) W ( x , y ) ,
and W ( x , y ) is a weight function:
V ( x , y ) = ( x 0 , y 0 ) S ω 2 ( x x 0 , y y 0 ) 0 .
For multi-dimensional images (e.g., RGB images), the SHT is obtained by applying the HT to each band and then rotating the coefficients, generating a set of coefficients per band. Moreover, it is possible to decompose the whole color image using 3-D Hermite filters and then steering the Cartesian Hermite coefficients towards two local orientation angles ( θ and ϕ ). See Mira et al. [43] for more details.

3.4. Jigsaw Transform

The Jigsaw transform is a popular scrambling technique used to hide visual information on digital images. It is calculated by relocating blocks of pixels of fixed size [44]. Thus, a digital image ( I ( x , y ) ) of X × Y pixels could be divided into k × l blocks of s 1 × s 2 pixels each one, so that k = X / s 1 and l = Y / s 2 . A random number generator defines the new location of the j-th block, with j = 1 , , k × l , and the original positions of each block are stored to recover the original image. Thus, the encryption/decryption process is symmetric and can be applied to both RGB and gray-scale images.
Figure 6 shows a fundus image of 960 × 11,140 pixels, and the JST results, applied to each color band of the RGB image, varying the number of blocks ( k × l ) and the size of each one ( s 1 × s 2 ).

3.5. Singular Value Decomposition

Singular Value Decomposition is a mathematical tool used to determine the intrinsic algebra present in a matrix. For example, let I a gray-scale image of X × Y , the SVD decomposed it into three matrices as is shown in Equation (9):
I = i = 1 r σ i u i v i T = U S V T ,
where U R Y × Y , V R X × X are orthogonal matrices and S R X × Y is diagonal:
S = σ 1 0 0 0 σ 2 0 0 0 0 0 0 σ r ,
and the values σ 1 σ 2 σ r are unique, and are the singular values of I. The number r m i n ( X , Y ) is equal to the rank of I.
In image processing, SVD has been applied to several problems, such as compression, noise reduction, and watermarking [45]. One interesting property of SVD in images is related to the singular values, which do not change when the image is altered, for example, by image processing operations.

3.6. Convolutional Neural Networks

Nowadays, the convolutional neural network is a novelty approach that has obtained efficient results in different fields of science and technology, such as computer vision, illness detection, image segmentation, self-driving, biometric authentication, and the entertainment industry.
The basis of CNNs is convolutional layers that extract local features (e.g., edges and textures) from a set of input images through several convolution kernels [46]. The 2-D kernels or filters are represented by matrices, and layers output is known as a feature map. Thus, each convolutional layer could be trained to detect determinate local features, generating a output feature map. In addition, pooling layers (or down-sampling layer) are used to reduce the dimension of the output of each convolutional layer, decreasing the number of neurons and reducing in consequence overfitting and computational complexity. computational complexity. Max-pooling layer, min-pooling layer, and average-pooling layer are three classes of pooling layers, being the max-pooling layer the most commonly used. Figure 7 shows a classical architecture of a CNN, consisting of various convolutional-pooling layer pairs for feature extraction. Moreover, fully connected layers are used for classification tasks, and in the final process an activation function generated the output classification labels.
The unknown parameters of a CNN model are the weights and biases of the connections. Thus, an iterative training process adjusts these weights for a later classification process [46]. In a typical training process, a high number of training samples are used. For example, ImageNet [47] is a dataset containing more than 14 million images in 1000 classes. However, in those cases where a large set of samples is not available, the transfer learning approach is an alternative. This method transfers the weights of a pre-trained network, with many samples, to the desired network to be trained by performing a fine-tuning [48]. Thus, the transfer learning replaces the last fully connected layer of the pre-trained CNN with a specific fully connected layer of the problem to be resolved [46].
For many classification problems, a fine-tuning of the last layers is performed, whereas the first layers remain without changes. However, if there is a significant difference between the data source of the pre-trained network and the data source of the current network, a fine-tuning of the first layers could contribute to extracting primitive features (gray levels, edges, colors, etc.) of the desired data samples [49].
Regarding pre-trained networks, VGG16, Inception-v3, and ResNet50 are three relevant networks used on medical image problems by applied transfer learning. These networks obtained the highest performing CNNs on the ImageNet Large Scale Visual Recognition Challenge (ILSVRC2014) [50]. On the other hand, YOLO (You Only Look Once) is a real-time object detection algorithm that has gained popularity since 2016, when YOLOv2 was released [51].
On the one hand, VGG-16 is a network with 16 convolutional layers, two fully-connected layers, a softmax classifier, 3 × 3 convolutional filters, and 2 × 2 max-pooling operations [52]. Inception-v3 CNN contains 48 convolutional layers and incorporates the so-called “Inception modules”, which reduce the number of parameters while holding the network efficiency [53]. On the other hand, ResNet50 contains 50 convolutional layers into five blocks and it uses residual blocks by the so-called “identity shortcut connection” that skips one or more layers. The first block is composed of a convolutional and a pooling layer. The following four blocks use a layer stack of 1 × 1 , 3 × 3 , and 1 × 1 convolutional filter size, respectively, with three, four, six, and three layers for each block.
In addition, YOLOv8 from Ultralytics [54] incorporates anchor boxes from Faster R-CNN, which are used to size and shape real-time prediction of the objects [51]. It has two main parts: the backbone and the head. The backbone of YOLOv8 is a modified version of the CSPDarknet53 architecture, consisting of 53 convolutional layers that use cross-stage partial connections to improve information flow between the layers. The head of YOLOv8 consists of multiple convolutional layers followed by a series of fully connected layers. These layers are responsible for predicting bounding boxes, objectness scores, and class probabilities for the objects detected in an image. Finally, YOLOv8 can perform multi-scaled object detection using a feature pyramid network to detect objects of different sizes and scales within an image.

4. Proposed Method for Image Watermarking

The present article is an extension of a previous work published in [55]. In that work, we proposed a hybrid watermark method based on the SHT, SVD, and JST. The original algorithm uses the SHT because it is possible to insert the watermark in oriented structures. This has the advantage that the watermark is more imperceptible to the Human Vision System (HVS). On the other hand, we employ SVD because we ensure more robustness to different attacks using the second level of SVD. In addition, the algorithm was tested with chest X-ray gray-scale images extracted from the Kaggle public dataset COVID-19 Radiography Dataset (https://www.kaggle.com/datasets/tawsifurrahman/covid19-radiography-database/, (accessed on 14 February 2022)). The results showed that medical images do not suffer visual alterations reporting PSNR values up to 30 dB, Mean Structure Similarity Index (MSIM) values are within 0.9800, and Normalized Crossed-Correlation (NCC) values equal to 0.9900.
In this paper, we propose to use a variation of the algorithm presented in [55], applied to RGB fundus images, and evaluate whether the watermark insertion process modifies the DR diagnosis.
Thus, the main differences with the previous work are listed as follows: (i) The proposed watermarking method was adapted and tested on RGB fundus images instead of gray-scale images, as was reported in [55]. (ii) The fundus images were first transformed into the Y’CbCr color model, then the luma channel was selected to insert the watermark, and the SHT for 2D images (Equations (1) and (4)) was applied. (iii) The watermark was embedded into the blood vessel patterns extracted by the SHT in the luma channel, obtaining in consequence a natural and automatic ROI-based watermarking method, and differing from [55] where the watermark is embedded in the whole image. (iv) The encryption security of the watermark was increased by using blocks of 1 × 1 of the JST. (v) The watermarking method was tested over 1748 8-bit color images of different sizes ( 1440 × 960 and 2240 × 1488 pixels). (vi) A complete robust analysis was performed by applying 11 different attacks instead only of six attacks from the previous work. (vii) A classification task using different deep learning architectures was carried out to identify diagnosis changes in the watermarked images. Finally, (viii), with the proposed modifications we overcome the evaluation metrics of the previous work.

4.1. Watermarking Insertion Process

Figure 8 shows the details of the proposed watermarking insertion schema using the steered Hermite transform and the SVD technique. The solid blue line corresponds to steps for the color model conversion of the cover fundus image, its spatial-frequency decomposition using the SHT and the SVD, and the watermarking embedding step. On the other hand, the dashed red line represents the following steps for pre-processing and encrypting the watermark using the JST. In addition, the green dotted line shows steps to reconstruct the watermarked fundus image using the inverse SVD, the ISHT, and the color model conversion. Finally, the yellow-shaded ellipses correspond to those elements stored in the Key Area used in the extraction process, and the symbols * and + represent the multiplication and addition operations, respectively.
Thus, the steps of embedding the watermark ( W ( x , y ) ) are described below:
  • Input the RGB fundus image I R G B ( x , y , z ) .
  • Convert the fundus image I R G B ( x , y , z ) from the RGB model to the Y’CbCr (Y represents the luma component or the brightness in an image, Cb and Cr are the blue-difference and red-difference chrome components, respectively). In the Equation (11) we present this conversion as is described in [56]:
    I Y ( x , y ) I C b ( x , y ) I C r ( x , y ) = 16 128 128 + 65.481 128.553 24.966 37.797 74.203 112 112 93.786 18.214 · I R ( x , y ) I G ( x , y ) I B ( x , y ) ,
    where I R ( x , y ) , I G ( x , y ) and I B ( x , y ) correspond to the color bands of the image I R G B ( x , y ) , and I Y ( x , y ) , I C b ( x , y ) and I C r ( x , y ) correspond to the components of the image I Y C b C r ( x , y ) .
  • Select the luma band I Y ( x , y ) of I Y C b C r ( x , y ) . We found that the Y’ component is the band most suitable to embed the watermark.
  • Calculate the SHT for 2D images (Equations (1) and (4)) to the luma band I Y ( x , y ) , obtaining the steered Hermite coefficients I n , θ ( x , y ) . The parameters used are N = 2 , D = 2 N , and T = 2 .
  • Select the steered Hermite coefficient I j , θ ( x , y ) for embedding the watermark.
  • Apply the SVD to I j , θ ( x , y ) obtaining the matrices U 1 ( x , y ) , S 1 ( x , y ) and V 1 ( x , y ) . Store the diagonal matrix of singular values S 1 ( x , y ) into the Key Area.
  • Input the watermark image W ( x , y ) .
  • Resize W ( x , y ) to have the same size of I j , θ ( x , y ) .
  • Encrypt the resized matrix W ( x , y ) by applying the JST, by dividing the resized watermark into k × l = X × Y blocks of s 1 × s 2 = 1 × 1 pixels each one, and obtaining the matrix W J S T ( x , y ) . The original indexes I d x of the blocks are stored in the Key Area.
  • Embed the encrypted watermark W J S T ( x , y ) in the matrix S 1 ( x , y ) applying the Equation (12):
    S W ( x , y ) = S 1 ( x , y ) + α W J S T ( x , y ) ,
    where α is a scaling factor, which defines the imperceptibility of the watermark image.
  • Apply a second SVD to S W ( x , y ) abstaining the matrices U 2 ( x , y ) , S 2 ( x , y ) and V 2 ( x , y ) . Store the left and right singular matrices U 2 ( x , y ) and V 2 ( x , y ) into the Key Area.
  • Apply the inverse SVD using the matrices S 2 ( x , y ) , U 1 ( x , y ) and V 1 ( x , y ) obtaining the marked steered Hermite coefficient I ^ j , θ ( x , y ) .
  • Perform the assembly of the steered Hermite coefficients using the coefficients I n , θ ( x , y ) (step 4) and including the marked coefficient I ^ j , θ ( x , y ) (step 12).
  • Calculate the ISHT (Equation (4) with θ ( x , y ) = θ ( x , y ) and Equation (7)), obtaining the marked luma band I ^ Y ( x , y ) .
  • Finally, using the bands I C b ( x , y ) and I C r ( x , y ) of I Y C b C r ( x , y , z ) (step 2) and the marked luma band I ^ Y ( x , y ) , convert from the Y’CbCr color model to the RGB model obtaining the marked fundus image I ^ R G B ( x , y , z ) . Equation (13) show this conversion [56]:
    I R ( x , y ) I G ( x , y ) I B ( x , y ) = 0.00456621 0 0.00625893 0.00456621 0.00153632 0.00318811 0.00456621 0.00791071 0 · I Y ( x , y ) I C b ( x , y ) I C r ( x , y ) 16 128 128 .
An important factor in the proposed watermarking method corresponds to the selection of the steered Hermite coefficient I j , θ ( x , y ) for embedding the watermark because is important to have robustness and imperceptibility. So, we performed several tests and we selected the coefficient I 2 , θ ( x , y ) . Figure 9 shows the luma component I Y ( x , y ) (Figure 9a) of a fundus image and its corresponding coefficient I 2 , θ ( x , y ) (Figure 9b).
The selected coefficient allows for inserting the watermark into the blood vessel patterns, giving a natural and automatic ROI-based watermarking method. Thus, the proposed method is an alternative to Dey et al. [12], where the blood vessels are first extracted using K-means segmentation and then the EPR is hidden into them using interpolation and trigonometric functions. Also, our approach differs from the work of Singh et al. [15], where a unique identification code is inserted into the blood vessels detected in fundus images. In that work, the blood vessels are detected through the matched filter and derivative of Gaussian, and these patterns, unique per patient, represent the identification code.

4.2. Watermarking Extraction Process

In Figure 10, we can see the proposed watermarking extraction schema. It is composed of two stages. First, the solid blue line corresponds to the decomposition of the marked fundus image using the SHT and SVD, followed by the operations to extract the watermark. Second, the dashed red line represents the decryption steps, using the inverse JST and the post-processing operations to recover the watermark. On the other hand, the yellow-shaded ellipses are those elements stored in the Key Area in the insertion phase and used to recover the watermark.
The extraction process is described below:
  • Input the marked fundus image I ^ R G B ( x , y ) .
  • Convert the marked fundus image I ^ R G B ( x , y ) to the Y’CbCr color model (Equation (11)), obtaining the image I ^ Y C b C r ( x , y ) .
  • Select the luma band of I ^ Y C b C r ( x , y ) .
  • Calculate the SHT for 2D images (Equations (1) and (4)) to the luma band I ^ Y ( x , y ) , with the same parameters employed in the insertion process, obtaining the steered Hermite coefficients I ^ n , θ ( x , y ) .
  • Select the steered Hermite coefficient I ^ j , θ ( x , y ) to extract the watermark.
  • Calculate the SVD to I ^ j , θ ( x , y ) obtaining the matrices U ^ 1 ( x , y ) , S ^ 1 ( x , y ) and V ^ 1 ( x , y ) .
  • Calculate a second SVD to the matrix S ^ 1 ( x , y ) obtaining the matrices U ^ 2 ( x , y ) , S ^ 2 ( x , y ) and V ^ 2 ( x , y ) .
  • Read the left and right singular matrices U 2 ( x , y ) and V 2 ( x , y ) from the Key Area (yellow-shaded rectangle).
  • Apply an inverse SVD decomposition using matrices S ^ 2 ( x , y ) , U 2 ( x , y ) , and U 2 ( x , y ) to obtain the diagonal matrix of singular values S ^ W ( x , y ) .
  • Read the diagonal matrix of singular values S 1 ( x , y ) from the Key Area (yellow-shaded rectangle).
  • To extract the first version of the watermark we apply the Equation (14):
    W ^ J S T ( x , y ) = S ^ W ( x , y ) S 1 ( x , y ) α .
  • Read the indexes I d x from the Key Area (yellow-shaded rectangle) used in the JST of the insertion process.
  • Apply the inverse Jigsaw transform to W ^ J S T ( x , y ) , using the indexes I d x , to obtain the original position of the pixels, and recover the watermark image.
  • Finally, the extracted watermark is resized to its original dimensions to obtain the final watermark image W ^ ( x , y ) .

4.3. Network Architectures for DR Classification

For the DR classification of the images present in MESSIDOR-2 dataset we used the CNN architectures VGG16, ResNet50, and InceptionV3 through Tesorflow-Keras [57,58]. In addition, we used the YOLOv8 network due to its simplicity, speed, and accuracy through the implementation of Ultralytics [54].
As to VGG16, ResNet50, and InceptionV3, we used their standard Keras Implementation as defined in the “tf.keras.applications” module [58] and as it is shown in Table 1:
For the YOLOv8 architecture, we selected the YOLOv8n variant, which is the smallest and simplest one.
The MESSIDOR-2 dataset was divided into two parts: 80% for training and 20% for testing. Tools like Keras and YOLO offer straightforward methods to create classes without manual annotation. However, data should be organized in a specific file structure.
The root directory contains sub-folders named with class identifiers (0, 1, 2, …). Each sub-folder corresponds to a class and contains images (img1.jpg, img2.jpg, …) belonging to that class. This structure is effective as each image features a retinal image and is associated with only one class:
  • root/
  • |– 0/
  • | |– img1.jpg
  • | |– img2.jpg
  • | |– …
  • |
  • |– 1/
  • | |– img3.jpg
  • | |– img4.jpg
  • | |– …
  • |
  • |– 2/
  • | |– img5.jpg
  • | |– img6.jpg
  • | |– …
  • |
  • |– …
Regarding the classification categories, we utilized the “adjudicated DR Grade” from the MESSIDOR-2 DR Grades dataset [41] as the basis for our classes. This classification is crucial for evaluating how watermarking affects the neural network’s ability to correctly identify classes.
As per described previously, the categories used for classification (adjudicated_dr_grade) are as follows, according to the five-point International Clinical Diabetic Retinopathy (ICDR) grade:
  • 0 = None
  • 1 = Mild DR
  • 2 = Moderate DR
  • 3 = Severe DR
  • 4 = PDR

5. Experiments and Results

5.1. Experimental Setup

We implemented the proposed watermarking algorithm in a Dell laptop Inspiron 7380 (Dell, Inc., Round Rock, TX, USA) on an Intel Core i7 @ 1.6 GHz (Intel Corporation, Santa Clara, CA, USA) with 16 GB RAM. The method was programmed in a non-optimized script in MATLAB R2018b (MathWorks, Natick, MA, USA) without a parallel configuration. The parameters used were: (i) For the SHT an analysis window size N = 2 , a maximum expansion order D = 2 N = 4 , and a sub-sampling factor T = 2 . (ii) Blocks of size s 1 × s 2 = 1 × 1 pixels for the JST. (iii) A scaling factor α = 1 × 10 5 (see Section 5.3).
Alternatively, the CNN architectures were developed on a personal custom built computer operating with Manjaro Linux KDE Edition (Manjaro GmbH & Co. KG, Grafing b. München, Bavaria, Germany). This system was equipped with a Ryzen 5 3600 processor (Advanced Micro Devices, Inc., Santa Clara, CA, USA), an NVIDIA GeForce RTX 3070 8 GB graphics card (Nvidia Corporation, Santa Clara, CA, USA), and 32 GB of RAM.
The watermarking algorithm has a time-consumption of 0.67061 s for the insertion stage and 0.78226 s for the extraction using images of 2240 × 1488 pixels. These times were obtained without the JST step, which is the stage that consumes more time when small size blocks are used ( 1 × 1 , 2 × 2 , 3 × 3 , etc.).

5.2. Evaluation Metrics

In this section, we describe two sets of evaluation metrics. First, we reported the watermarking metrics to measure the quality of the watermarked images and the extracted watermark. Secondly, we presented the metric used to test the influence of the watermark on the diagnosis of DR.

5.2.1. Watermarking Metrics

To evaluate the algorithm we use different metrics that are commonly used in watermarking. Some of them are focused on evaluating image quality and others on determining if the image has suffered modifications compared with the original. Considering the images I 1 (original image), I 2 (watermarked image), with X × Y size, and x and y represent the spatial coordinates we define the metrics as follow:
  • Mean Square Error (MSE) is a quality criterion in image processing. It evaluates the similarity between two images ( I 1 , I 2 ). See Equation (15):
    M S E = 1 X Y x = 1 X y = 1 Y I 1 ( x , y ) I 2 ( x , y ) 2 .
  • PSNR is an objective metric to compare two images, using numerical criteria [59]. In Equation (16) it is defined:
    P S N R = 10 log 10 255 2 M S E .
    Close values to zero of MSE and higher values of PSNR correspond to a better quality of the watermarked image.
  • NCC has been used as a metric to evaluate the degree of similarity between two images. Compared with ordinary cross-correlation, it is less sensitive to linear changes in the amplitude of illumination in two images [60]. We calculated it using Equation (17).
    N C C = x X y Y I 1 ( x , y ) I ¯ 1 I 2 ( x , y ) I ¯ 2 x X y Y I 1 ( x , y ) I ¯ 1 2 x X y Y I 2 ( x , y ) I ¯ 2 2 1 / 2 ,
    where I ¯ * represents the average value of I * .
    Close to one value of NCC corresponds to a watermarked image of better quality.
  • Structural Similarity Index (SSIM) is a quality metric to measure the similarity between two images. It is considered to be correlated with the quality perception of the HVS [59]. It takes into account three factors: loss of correlation, luminance distortion, and contrast distortion. It is defined by Equation (18).
    S S I M ( I , I ^ ) = ( 2 μ I μ I ^ + C 1 ) ( 2 σ I I ^ + C 2 ) ( μ I 2 + μ I ^ 2 + C 1 ) ( σ I 2 + σ I ^ 2 + C 2 ) ,
    where I and I ^ correspond to the original and the distorted image, respectively. The averages of I and I ^ are given by μ I and μ I ^ , respectively, and their standard deviations are given by σ I and σ I ^ , the covariance between both images is represented by σ I I ^ , and the constants C 1 and C 2 are used to prevent instability if the denominator happens to have a value close to zero.
  • MSSIM represents the SSIM mean. Is calculated as we show in Equation (19).
    M S S I M ( I , I ^ ) = 1 M j = 1 M S S I M ( I j , I ^ j ) ,
    where I j and I ^ j represent their j-th local window, and M stands for the number of local windows of the image. In the case that I j and I ^ j have no negative values.
    A better quality of the watermarked image is achieved for close to one value of SSIM and MSSIM metrics.

5.2.2. Classification Metrics

To evaluate how the watermark embedding process affects the DR diagnosis in the fundus images, the accuracy classification metric was used as a support metric for the tests in all architectures (YOLOv8, VGG16, InceptionV3, and ResNet50), as it is shown in Equation (20):
A c c u r a c y = T P + T N T P + T N + F P + F N ,
where T P , T N , F P , and F N correspond to the True Positive, True Negative, False Positive, and False Negative values, respectively.
Accuracy is measured within the range [ 0 , 1 ] , with 0 as the worst and 1 as the best possible score.

5.3. Sensitivity Analysis of the Scaling Factor

The scaling factor α , used in the insertion and extraction process, defines the imperceptibility of the watermark and its robustness against attacks. In this sense, low values give imperceptibility but decrease robustness. On the other hand, high values of the scaling factor provide robustness but affect the quality of the watermarked image. Therefore, a sensitivity analysis of the scaling factor was performed by marking with different values of α a set of ten images from the image dataset and randomly selected. The values of the scaling factor varied from 1 × 10 6 to 1 × 10 4 with steps of 1 × 10 6 , generating 100 different results per image. Thus, we marked a total of 1000 images and we calculated the watermarking metrics reported in Section 5.2. Table 2 shows the average metrics obtained over the selected image set by varying the scale factor.
The metrics shown in Table 2 mean that the scale factor α does not affect the quality of the watermarked image in the insertion process.
On the other hand, Figure 11 shows the PSNR and the SSIM metrics of each selected image (A–J images) by varying the scale factor α . In addition, we graph the threshold scale factor α t h = 1 × 10−5 (vertical black dotted line) which shows that a scale factor below α t h , the PSNR value (Figure 11a) of some images decrease. In the case of the SSIM (Figure 11b), the same behavior occurs for values below α t h .
In Figure 12, we show the images used for sensitivity analysis of Figure 11 (A–J images). We can see that for images B, C, D, G, I, and J, the PSNR and SSIM values remain constants for the values range of α . In addition, for images A and E metrics increase and then hold constant for values of α α t h . On the other hand, for images F and H, The metrics show higher values for α α t h and lower performance for values α t h α 5 × 10 5 , which corresponds to those images where blood vessel are not visible or are very subtle. Thus, we fixed the scaling factor to α = 1 × 10 5 for all images of the dataset, which ensures both a high quality of the marked images and a successful extraction of the watermark.

5.4. Watermarking Performance Analysis

The proposed watermarking method was tested using the complete image public dataset MESSIDOR-2 [20,21]. The Caduceus symbol (Figure 2) was embedded following the steps presented in Section 4.1. Below we present the performance measures of the proposed method.
Table 3 presents the average performance metrics for the watermarked images (insertion stage) and the recovered watermarks (extraction stage).
On the other hand, Figure 13 shows some of the best examples of the insertion and extraction process. The left column corresponds to the original fundus images. The middle columns show the watermarked images and the right column shows the extracted watermarks, with their PSNR value, respectively. As we can see, these results show a good balance between the invisibility of the watermarks and the quality of the watermarked images, with high PSNR values of the watermarked images, and the quality of the extracted watermarks with high PSNR values.
In addition, Figure 14 shows some results with worse performance in the extracted watermark process, but with high metrics in the watermarked image. Thus, the original fundus images are shown in the left column. The middle column shows the watermark images and the right column shows the extracted watermarks with their PSNR values, respectively. In these results, the PSNR values of the extracted watermarks decrease a bit. However, the high PSNR values of the watermarked images demonstrate that the invisibility of the watermarks and the quality of the watermarked images remain unchanging compared with the best results.
Comparing Figure 13 and Figure 14, we can see that visually original and watermark images do not suffer alterations, and even though recovery watermarks in Figure 14 have PSNR values from 20 to 30 dB the images are clear. Therefore, a critical point studied and reported later is to determine if the DR grade diagnosis is modified in the watermarked images respecting the originals.

5.5. Watermarking Robustness against Attacks

To measure the robustness of the present watermarking approach, we applied different attacks to the watermarked images of the complete dataset. To this aim, we vary the parameter that defines the operation in each attack. Following, in Table 4 we list the type of attack, the operation name, and the corresponding parameter that was applied:
In Table 5, we show the average results of the completed dataset by applying the following image processing operations to the watermarked image: Gaussian filter, Median Filter, Gaussian Noise, and Salt and Pepper Noise.
In Table 6, we show the average results of the completed dataset by applying the following image processing operations to the watermarked image: Contrast Enhancement, Histogram Equalization, JPEG Compression, and Image Scaling.
In Table 7, we present the average results of the completed dataset by applying the following geometric attacks: Rotation, Cropping, and Translation.
On the other hand, Figure 15, Figure 16 and Figure 17 show some results of the watermarked and attacked images and the corresponding extracted watermark. The first two columns in each figure correspond to some satisfying examples of extracting the watermark, whereas the last two columns represent those cases where the extracted watermark presented worse performance. Moreover, each row corresponds to a different attack.
In addition, Table 8, Table 9 and Table 10, show the metrics obtained of the extracted watermarks for each attack of Figure 15, Figure 16 and Figure 17, respectively, both the most satisfying and worst performance result. It is essential to mention that the extracted watermarks are visually recognizable in most cases in the worse-performance results (the last two columns).

5.6. Comparison with Other Watermarking Methods

An important point to define the effectiveness of our proposal is to compare it with other similar algorithms, focusing on medical images. As we present in Section 5, our proposal was tested using fundus images, so this is an element to consider to selecting the approaches to compare. In Table 11, we present the comparison values of different watermarking schemes [9,14,16,17,18,19,22]. We present the results indicated in each paper using fundus images. In Table 12 we indicated the type of watermarking and the dataset repository that each algorithm uses.
It is important to remark on some points of values reported by each method. In the method [14], the authors evaluated the three channels RGB, because the green channel is used for generating the watermark, and the other channels (blue and red) are used to insert the watermark. So in Table 11, we included the average value of red and blue channels. The paper [16] presents its results using different types of images, but the authors report their values metrics separately, so we only include the values referring to fundus images. The authors of the algorithm [17] used the LWT and tested their proposal using sub-bands LL and sub-bands HH, which is the reason why in Table 11 we present both results. The first row corresponds to sub-band LL and the second row is the value of sub-band HH. The methods [10,18] were evaluated with CT images, MRI images, ultrasound images, X-ray images, and fundus images and their results are the average of all of them, in these cases, we present this average. The authors of [10] indicated that used a total of 200 medical images, but we reported only the fundus images (40) that they used, in Table 12.
According to Table 11 and Table 12, we can see, on the one hand, the authors do not calculate all metrics to evaluate their algorithm, so if we consider only PSNR values, even though the algorithms [9,14,17] have better PSNR values, our technique has a good result too. Also, considering the NCC value our technique is competitive because these values demonstrate that the image does not suffer visual modifications. In addition, the MSE is very low and SSIM is very close to one, so with these values, we reinforce the imperceptibility of the watermark. On the other hand, to probe the effectiveness of an algorithm, it is necessary to use a considerable amount of data, in this case, our dataset is bigger than the others, and the capacity to hide information as a watermark is competitive compared to other state-of-the-art works. Finally, is important to compare taking into account the robustness of each one.
Particularly, the method described in [9] was not evaluated to probe its robustness, so this method has the highest value of PSNR (Table 11) but this does not facilitate guaranteed success against attacks. This paper only evaluates if the watermark affects the original diagnosis. According to its results, all tests are satisfactory indicating that the diagnosis does not change with the watermark.
Regarding the method reported by A. George Klington et al. [14], they only applied the jittering attack. They concluded that the proposed algorithm changes 43 % of the original watermark bits against this attack. In the case of the method presented in [10], the metric they used to compare with other algorithms is mean BER (Bit Error Rate).
With the rest of the methods it is not easy to make a comparison because the authors do not use the same attacks or in some cases the parameters of each attack are different. In Table 13, Table 14 and Table 15 we include the results after applying the Median Filter attack, JPEG Compression, and Cropping attack, respectively.
As we can see from Table 13, in all cases, our algorithm has the best results after applying the Median Filter. Concerning JPEGC, our results are low compared with other proposals ([16,18]), however, the metric values demonstrate that it is possible to recover the watermark and it is clear, as we demonstrate in Section 5. In the case of the Cropping attack, the method presented by Xiyao Liu et al. [18] has a better performance than our algorithm, the values indicated that the extracted watermark is very similar to the original watermark. So with this experiment, we demonstrated that our method is competitive with other similar methods. Regarding the attacks such as Gaussian Noise, and Salt and Pepper Noise, the parameters that we use to test our method are more strict as we can see in Table 5. While methods [16,17] use Salt and Pepper densities of 0.05, 0.01, and 0.02. For Gaussian Noise, they use values of 0.01 and 0.02 or 0.03. So, according to our results, our method is robust against attacks.

5.7. Analysis of the Influence of the Watermark on the Diagnosis of Diabetic Retinopathy

In this section, we evaluate the influence of the watermarking process on DR diagnosis using a deep learning approach. Thus, accuracy is reported as a support metric, as the main focus remains on the effects of the watermarking on the inference produced by the different models.

5.7.1. Yolov8 Model Results

In the evaluation of YoloV8, it was observed that the model’s top 1 prediction accuracy averaged 0.736. This is visually represented in the accompanying figure (Figure 18).
Notably, the inference performance of this model was exemplary, accurately classifying the majority of data into their respective classes as delineated description files of the dataset, which specifies the class to which each image belongs. This high level of accuracy is further evidenced by the model’s ability to generate consistent inference matrices for both marked and non-marked images, as illustrated in Figure 19.
The inference results for an original image (Figure 19a) and its watermarked counterpart (Figure 19b) are particularly revealing. The original and the watermarked images were classified as class 3 (severe). Both classifications align accurately with the DR grades assigned in the dataset labels (see Figure 19c). Crucially, the similarity in the inference matrices for both the marked and non-marked images demonstrates that watermarking does not adversely affect the model’s inference capabilities on this architecture.
Therefore, the YoloV8 model exhibits high accuracy in its top 1 predictions and maintains consistent performance in classifying both watermarked and non-watermarked images, thereby proving that watermarking does not affect the predicted class.

5.7.2. VGG16 Model Results

The VGG16 model achieved a peak accuracy of 58.31% as we demonstrated in the results. However, this accuracy level did not show any improvement after the initial two epochs. This stagnation in performance is clearly illustrated in Figure 20, where the graphical representations of the model’s accuracy and loss are presented.
Further analysis of the VGG16 model was conducted to assess the impact of image watermarking on the classification performance. The comparison involved using both watermarked and non-watermarked images in the inference process. The results, as depicted in Figure 21, indicate that watermarking does not adversely affect the model’s ability to correctly classify images for computer-aided DR diagnosis. Both watermarked and non-watermarked images yielded identical confidence levels in the predicted class, in the example class 0 (NO DR).

5.7.3. InceptionV3 Model Results

In Figure 22, we present the performance metrics for the InceptionV3 model. This figure is divided into two parts: Figure 22a illustrates the accuracy of the InceptionV3 model, and Figure 22b displays its loss metrics. Both sub-figures provide a comprehensive view of the model’s performance characteristics.
The InceptionV3 inference results indicate a slight but notable variation in the confidence level of the predicted class for certain images, with a deviation of Δ E = ± 1.9941 % . Despite this fluctuation, the classification accuracy remains high, as evidenced in Figure 23. It is important to note that while there is a minor difference in confidence values for certain images, the model consistently produces high-confidence results. This variation does not significantly impact the overall accuracy, as demonstrated in the comparative analysis of marked and non-marked images.
Further analysis, as shown in Figure 23, confirms the aforementioned slight discrepancy in the confidence levels between marked and non-marked images. While this variance is minimal and does not affect the correct classification of images, it is an interesting observation that non-marked images tend to exhibit marginally higher confidence levels. This pattern is not consistent across all images, as some show identical inference results for both their marked and non-marked versions as the one presented in Figure 24.

5.7.4. ResNet50 Model Results

The ResNet50 model, trained over 100 epochs, achieved an outstanding accuracy of 1.0. An analysis of the inference matrix for both marked and non-marked images revealed a minuscule average difference ( Δ E = ± 0.002 % ), indicating a negligible impact of watermarking on the model’s performance. Notably, ResNet50 demonstrated exceptional confidence in its predictions, averaging a 99.2% confidence level for the top predicted class. Remarkably, some images were classified with a perfect confidence score of 100%. The performance metrics for the ResNet50 model, including accuracy and loss, are comprehensively detailed in Figure 25.
Figure 26 and Figure 27 further elucidate the effect of watermarking on the inference process. In these figures, we observe the minimal impact of watermarking on the inference matrix, as well as instances where watermarking has no perceivable effect on the model’s inference accuracy.
In the evaluation of the ResNet50 model, the model achieved an unmatched accuracy rate of 100% over 100 epochs. A critical aspect of our analysis was examining the impact of watermarking on image classification. Remarkably, the ResNet50 model demonstrated remarkable resilience to watermarking, with an average deviation of just 0.002% in the inference matrix between marked and non-marked images. This negligible impact underscores the model’s ability to maintain accuracy and reliability in classification, despite minor alterations in the input data. Furthermore, the model consistently displayed high confidence in its predictions, with an average confidence level of 99.2% for the top predicted class, and in some cases, reaching a perfect confidence score of 100%. The minimal variation in performance, irrespective of watermarking, affirms that watermarking medical images using the proposed method does not present an effect in computer-aided and vision-based diagnostics.

6. Discussion

The method described in this paper focuses on a method to watermark digital medical images (fundus images) and four trained deep-learning models to evaluate if the original diagnosis is modified by the watermark. To probe the performance of the proposed method, typical metrics employed to evaluate quantitatively (MSE, PSNR and NCC) and visually (SSIM and MSSIM) any alteration of the watermarked image and the extracted watermark were used. Thus, the results reported in Table 3 show that the average performance metrics, using the 1748 fundus images, present high quality of the watermarked images with values of the MSE = 4.6976 × 10 6 PSNR = 53.8638 dB, NCC = 0.9993, SSIM = 0.9937, and MSSIM = 0.9938, demonstrating that the invisibility of the watermarks on the whole image dataset. On the other hand, the recovered watermarks in all marked images had excellent performance reporting values of MSE = 7 × 10 4 , PSNR = 32.0690 dB, NCC = 0.9975, SSIM = 0.9937, and MSSIM = 0.9943. This can be verified, in Figure 13 and Figure 14, where are given some examples of the best and worst performing, respectively. Thus, the reporting PSNR values, for these watermarked images, were between 53 and 54 dB for both the best and worst results. Regarding the extracted watermarks, the PSNR values were between 33 and 34 dB for the best results, and they decreased between 21 and 30 dB for the worst ones, but with extracted watermarks visually clear and discernible.
To evaluate the robustness of the proposed watermarking algorithm, the more common attacks in this application were applied. Thus, both image processing and geometric attacks were employed by varying their parameter from low and high effect over the watermarked images. Among the image processing attacks were applied: Gaussian and Median Filter, Gaussian and Salt and Pepper Noise, Contrast Enhancement, Histogram Equalization, JPEG Compression, and Image Scaling. Additionally, Rotation, Cropping, and Translation operations were applied as geometric attacks. Thus, from Table 5, Table 6 and Table 7, the average metrics show that the watermark is successfully extracted for the Median Filter, Gaussian and Salt and Pepper Noise, Histogram Equalization, Rotation, and Translation attacks, obtaining PSNR average values between 31.40636 and 32.92558 dB as a quantitative metric and MSSIM average values between 0.98996 and 0.99557 as a qualitative metric, allowing to extract the watermark with high visual quality as it is shown in Figure 15, Figure 16 and Figure 17. For the Contrast Enhancement attack, average values of PSNR = 29.29871 dB and MSSIM = 0.91870 (edge values) were obtained (see Table 6 and Figure 16). In addition, for the Image Cropping attack (Table 7 and Figure 17), average values of PSNR = 26.84495 dB and MSSIM = 0.89405 are consistent because up to 60% of the image is removed, destroying the watermark as a consequence.
On the other hand, worse performance was obtained for the following attacks: Gaussian Filter (PSNR = 28.24112 dB, MSSIM = 0.90795) and JPEG Compression (PSNR = 22.30630 dB, MSSIM = 0.78557). However, in both cases it is possible to recover the watermark and distinguish the visual information in it for parameters with high effect, as shown in Figure 15d and Figure 16l, respectively. In the Gaussian Filter case, the low values are caused by a 9 × 9 window when the edges of the images are considerably affected. For a JPEG Compression attack, the low values are present for quality percentages less than 70% due to information loss.
Finally, for Image Scaling, Table 6 shows the lowest average values of PSNR = 19.56431 dB and MSSIM = 0.66882, which is because the watermark extraction is unsuccessful for scaling factor values less than 1 or image downsampling (see Figure 16p), affecting the general performance for this attack. Succinctly, in image downsampling, it is not possible to recover the watermark. This has a similar behavior to the Cropping attack where the information of the image is also removed.
To determine if our algorithm is competitive with other similar approaches, we compare it with other methods taking into consideration their use of fundus images, because an important key of our investigation is to determine if the watermark modifies the original diagnosis. So as a first step, we compare the different methods [9,10,14,16,17,18,19,22] with our proposal to evaluate the invisibility of the watermarks. The results in Table 11 and Table 12 demonstrate that our technique is competitive with the state-of-the-art, including the contribution that our algorithm was tested to determine if the watermark modifies the original diagnosis. Regarding robustness, we presented the outcomes of the algorithms employing identical parameters to our proposal, see Table 10, Table 14 and Table 15. The values in these attacks indicated that our algorithm is robust and it is possible to extract the watermark without visible modifications, having the worst value when we applied a JPEG Compression of 70%. In the case of the Median Filter, we have the best results. With the rest of the attacks that we select to test our algorithm, the parameters employed are more strict than the parameters used by other proposals, so it demonstrates the robustness of our method. Referring to the capacity of the watermark, we use a logo and its dimensions are similar to the watermarks employed in the state-of-the-art.
Regarding the effect of watermarking in diagnosing diabetic retinopathy, we used various deep-learning models by evaluating their inference capabilities rather than solely concentrating on accuracy metrics.
Thus, the YoloV8 model displayed noteworthy performance, with a top 1 prediction accuracy averaging 0.736. This high accuracy level was maintained consistently across both watermarked and non-watermarked images. The model adeptly classified the majority of the data into their respective classes, as defined in the dataset description file. An in-depth examination of the inference results revealed that watermarking did not detrimentally affect the model’s ability to accurately classify images, a critical aspect in medical image diagnostics.
On the other hand, the VGG16 model showed a peak accuracy of 58.31%, a figure that notably stagnated after the initial two epochs. This stagnation is visually depicted in the corresponding figures. Despite this, the VGG16 model effectively classified both watermarked and non-watermarked images, with no significant difference in the confidence levels of the predicted classes. This finding was vital in understanding the impact of watermarking in medical image analysis, especially in the context of diabetic retinopathy.
The performance of the InceptionV3 model was also evaluated. Although there was a slight variation in confidence levels between watermarked and non-watermarked images, this deviation was minor and did not significantly influence the model’s overall classification accuracy. The InceptionV3 model successfully managed to maintain high-confidence results across different image types, thus underscoring its robustness in handling watermarked medical images.
Lastly, the ResNet50 model demonstrated exceptional performance, achieving a 100% accuracy rate over 100 epochs. The analysis of the inference matrix for both marked and non-watermarked images revealed a minuscule average difference, indicating the minimal impact of watermarking on the model’s performance. The model displayed extraordinary confidence in its predictions, further affirming its reliability in accurately classifying medical images irrespective of watermarking.
Additionally, this finding is encouraging for the future of automated medical diagnosis systems. It suggests that these systems can handle watermarked images without losing accuracy, which is important for maintaining patient confidence and data protection.

7. Conclusions and Future Work

This study introduces a novel, bio-inspired watermarking algorithm for fundus images leveraging the transform domain. Our approach integrates a combination of mathematical tools to achieve robust and imperceptible watermarking without compromising clinical diagnosis.
Firstly, the Steered Hermite transform acts as a natural detector for blood vessels, enabling automatic insertion of the watermark within this region of interest. This ensures imperceptibility while maintaining robustness. We further enhance security by incorporating the Jigsaw transform, an encryption process that scrambles the watermark information.
Secondly, Singular Value Decomposition bolsters the watermark’s resilience against various image processing and geometric attacks. Our algorithm’s effectiveness was evaluated on the MESSIDOR-2 dataset, encompassing 1748 RGB fundus images. Results confirmed superior invisibility and robustness compared to existing state-of-the-art methods.
Critically, we investigated the impact of watermarking on DR diagnosis. Four deep learning models (YoloV8, VGG16, InceptionV3, and ResNet50) were trained and tested on watermarked and original images. Remarkably, the analysis revealed a negligible influence on both disease classification accuracy and confidence scores. This finding underscores the compatibility of digital watermarking with computer-aided diagnostic systems, paving the way for enhanced data security without compromising diagnostic quality.
Looking forward, we aim to explore methods for strengthening the watermark’s resistance against attacks that manipulate or remove image information, potentially impacting the watermark itself. Additionally, we will delve deeper into blood vessel-based watermarking to optimize both imperceptibility and robustness.
By integrating these advancements, we can contribute to a future where robust and secure medical image management coexists seamlessly with accurate clinical diagnosis, ultimately benefiting patient care and overall healthcare security.

Author Contributions

Conceptualization, E.M.-A., S.L.G.-C. and J.B.; methodology, E.M.-A., S.L.G.-C., J.B. and A.L.-F.; software, E.M.-A., S.L.G.-C., J.B. and A.L.-F.; validation, E.M.-A., S.L.G.-C., J.B. and A.L.-F.; formal analysis, E.M.-A., S.L.G.-C. and J.B.; investigation, E.M.-A., S.L.G.-C., J.B. and A.L.-F.; resources, E.M.-A., S.L.G.-C. and A.L.-F.; data curation, E.M.-A., S.L.G.-C. and A.L.-F.; writing—original draft preparation, E.M.-A., S.L.G.-C., J.B. and A.L.-F.; writing—review and editing, E.M.-A., S.L.G.-C., J.B. and A.L.-F.; visualization, E.M.-A., S.L.G.-C. and A.L.-F.; supervision, E.M.-A. and S.L.G.-C.; project administration, E.M.-A. All authors have read and agreed to the published version of the manuscript.

Funding

This research has been funded by Universidad Panamericana under the Program “Fomento a la Investigación UP 2021” grant UP-CI-2021-MEX-24-ING.

Data Availability Statement

The image dataset used in this work is publicly available from the Kaggle public dataset MESSIDOR-2: https://www.kaggle.com/datasets/geracollante/messidor2/ (accessed on 10 August 2023).

Acknowledgments

Ernesto Moya-Albor, Jorge Brieva, and Alberto Lopez-Figueroa would like to thank Facultad de Ingeniería of Universidad Panamericana for all support in this work. Sandra L. Gomez-Coronel thanks Instituto Politécnico Nacional (UPIITA) for the support in this work. The image dataset was kindly provided by the MESSIDOR program partners (see https://www.adcis.net/en/third-party/messidor/, accessed on 10 August 2023).

Conflicts of Interest

The authors declare there is no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ACNNAtrous Convolutional Neural Network
AMDAge-Related Macular Degeneration
ANNArtificial Neural Network
BERBit Error Rate
BRBBlood-Retina Barrier
CADComputer-Aided Diagnosis
CEContrast Enhancement
CLAHEEnhance Local Contrast
CNNConvolutional Neural Network
CROPCropping
DCTDiscrete Cosine Transform
DFTDiscrete Fourier Transform
DMEDiabetic Macular Edema
DRDiabetic Retinopathy
DWTDiscrete Wavelet Transform
EPRElectronic Patients Record
EQHistogram Equalization
FDCuTFast Discrete Curvelet Transforms
GFGaussian Filter
GNGaussian Noise
HTHermite Transform
HVSHuman Vision System
ICDRInternational Clinical Diabetic Retinopathy
IHTInverse Hermite Transform
ISHTInverse Steered Hermite Transform
JPEGCJPEG Compression
JSTJigsaw Transform
LBPLocal Binary Patterns
LTPLocal Ternary Patterns
LVEqualization Levels number
LWTLifting Wavelet Transform
M-CNNMultipath Convolutional Neural Network
MFMedian Filter
MSEMean Square Error
MSSIMMean Structural Similarity Index
NCCNormalized Cross-Correlation
PCEPercent Saturation
PCRCropping Percentage
PSNRPeak Signal to Noise Ratio
QPQuality Percentage
RDMRecursive Dither Modulation
RDWTRedundant Discrete Wavelet Transform
RNNRecurrent Neural Network
ROIRegion Of Interest
RONIRegion Of No Interest
ROTRotation
SCImage Scaling
SDLTPSpherical Directional Local Ternary Pattern
SFScaling Factor
SHTSteered Hermite Transform
SLTSlantlet Transform
SPSalt and Pepper Noise
SSIMStructural Similarity Index
SVDSingular Value Decomposition
SVMSupport Vector Machine
TRANTranslation
WHOWorld Health Organization
YOLOYou Only Look Once

References

  1. World Health Organization. Blindness and Vision Impairment; Technical report; World Health Organization: Geneva, Switzerland, 2021. Available online: https://www.who.int/news-room/fact-sheets/detail/blindness-and-visual-impairment (accessed on 10 August 2023).
  2. Harvard Health Publishing. Retinopathy; Technical report; Harvard Medical School: Boston, MA, USA, 2020; Available online: https://www.health.harvard.edu/a_to_z/retinopathy-a-to-z (accessed on 10 August 2023).
  3. The Pan American Health Organization. Prevention of Blindness and Eye Care—Blindness; Technical report; The Pan American Health Organization: Washington, DC, USA, 2019; Available online: https://www.paho.org/hq/index.php?option=com_content&view=article&id=13693:prevention-blindness-eye-care-blindness&Itemid=39604&lang=en (accessed on 10 August 2023).
  4. Salamanca, O.; Geary, A.; Suárez, N.; Benavent, S.; Gonzalez, M. Implementation of a diabetic retinopathy referral network, Peru. Bull. World Health Organ. 2018, 96, 674–681. [Google Scholar] [CrossRef]
  5. Carrillo-Alarcón, L.; López-López, E.; Hernández-Aguilar, C.; Martínez-Cervantes, J. Prevalencia de retinopatía diabética en pacientes con diabetes mellitus tipo 2 en Hidalgo, México. Rev. Mex. Oftalmol. 2011, 3, 125–178. [Google Scholar]
  6. Prado-Serrano, A.; Guido-Jiménez, M.; Camas-Benítez, J. Prevalencia de retinopatía diabética en población mexicana. Rev. Mex. Oftalmol. 2009, 83, 261–266. [Google Scholar]
  7. Hervella, A.S.; Rouco, J.; Novo, J.; Ortega, M. Learning the retinal anatomy from scarce annotated data using self-supervised multimodal reconstruction. Appl. Soft Comput. 2020, 91, 106210. [Google Scholar] [CrossRef]
  8. Mookiah, M.R.K.; Acharya, U.R.; Chua, C.K.; Lim, C.M.; Ng, E.; Laude, A. Computer-aided diagnosis of diabetic retinopathy: A review. Comput. Biol. Med. 2013, 43, 2136–2155. [Google Scholar] [CrossRef]
  9. Singh, A.; Dutta, M.K. Imperceptible watermarking for security of fundus images in teleophthalmology applications and computer-aided diagnosis of retina diseases. Int. J. Med. Inform. 2017, 108, 110–124. [Google Scholar] [CrossRef]
  10. Dai, Z.; Lian, C.; He, Z.; Jiang, H.; Wang, Y. A Novel Hybrid Reversible-Zero Watermarking Scheme to Protect Medical Image. IEEE Access 2022, 10, 58005–58016. [Google Scholar] [CrossRef]
  11. Liu, J.; Li, J.; Ma, J.; Sadiq, N.; Bhatti, U.A.; Ai, Y. A Robust Multi-Watermarking Algorithm for Medical Images Based on DTCWT-DCT and Henon Map. Appl. Sci. 2019, 9, 700. [Google Scholar] [CrossRef]
  12. Dey, N.; Ahmed, S.S.; Chakraborty, S.; Maji, P.; Das, A.; Chaudhuri, S.S. Effect of trigonometric functions-based watermarking on blood vessel extraction: An application in ophthalmology imaging. Int. J. Embed. Syst. 2017, 9, 90–100. [Google Scholar] [CrossRef]
  13. Nayak, J.; Subbanna Bhat, P.; Acharya, U.R.; Sathish Kumar, M. Efficient storage and transmission of digital fundus images with patient information using reversible watermarking technique and error control codes. J. Med. Syst. 2009, 33, 163–171. [Google Scholar] [CrossRef] [PubMed]
  14. Klington, G.; Ramesh, K.; Kadry, S. Cost-Effective watermarking scheme for authentication of digital fundus images in healthcare data management. Inf. Technol. Control 2021, 50, 645–655. [Google Scholar] [CrossRef]
  15. Singh, A.; Dutta, M.K.; Sharma, D.K. Unique identification code for medical fundus images using blood vessel pattern for tele-ophthalmology applications. Comput. Methods Programs Biomed. 2016, 135, 61–75. [Google Scholar] [CrossRef] [PubMed]
  16. Dwivedi, R.; Srivastava, V.K. An Imperceptible Semi-blind Color Image Watermarking Using RDWT and SVD. In Proceedings of the International Conference on VLSI, Communication and Signal Processing, Prayagraj, India, 14–16 October 2022; Springer: Prayagraj, India, 2022; pp. 283–293. [Google Scholar]
  17. Awasthi, D.; Srivastava, V.K. Robust and Imperceptible Color Image Watermarking Using LWT, Schur Decomposition, and SVD in YCbCr Color Space. In Proceedings of the International Conference on VLSI, Communication and Signal Processing, Prayagraj, India, 14–16 October 2022; Springer: Prayagraj, India, 2022; pp. 259–271. [Google Scholar]
  18. Liu, X.; Lou, J.; Fang, H.; Chen, Y.; Ouyang, P.; Wang, Y.; Zou, B.; Wang, L. A Novel Robust Reversible Watermarking Scheme for Protecting Authenticity and Integrity of Medical Images. IEEE Access 2019, 7, 76580–76598. [Google Scholar] [CrossRef]
  19. Mahyudin, M.F.; Novamizanti, L.; Sa’idah, S. Robust Watermarking using Arnold and Hybrid Transform in Medical Images. In Proceedings of the 2021 IEEE International Conference on Industry 4.0, Artificial Intelligence, and Communications Technology (IAICT), Bandung, Indonesia, 27–28 July 2021; pp. 180–185. [Google Scholar] [CrossRef]
  20. Decencière, E.; Zhang, X.; Cazuguel, G.; Laÿ, B.; Cochener, B.; Trone, C.; Gain, P.; Ordóñez Varela, J.R.; Massin, P.; Erginay, A.; et al. Feedback on a publicly distributed image database: The Messidor database. Image Anal. Stereol. 2014, 33, 231–234. [Google Scholar] [CrossRef]
  21. Abràmoff, M.D.; Folk, J.C.; Han, D.P.; Walker, J.D.; Williams, D.F.; Russell, S.R.; Massin, P.; Cochener, B.; Gain, P.; Tang, L.; et al. Automated analysis of retinal images for detection of referable diabetic retinopathy. JAMA Ophthalmol. 2013, 131, 351–357. [Google Scholar] [CrossRef]
  22. Garg, P.; Jain, A. A robust technique for biometric image authentication using invisible watermarking. Multimed. Tools Appl. 2023, 82, 2237–2253. [Google Scholar] [CrossRef]
  23. Kapoor, P.; Arora, S. Applications of Deep Learning in Diabetic Retinopathy Detection and Classification: A Critical Review. Lect. Notes Data Eng. Commun. Technol. 2022, 91, 505–535. [Google Scholar] [CrossRef]
  24. Kaggle. Diabetic Retinopathy Detection Competition. 2015. Available online: https://www.kaggle.com/c/diabetic-retinopathy-detection (accessed on 15 August 2023).
  25. Radha; Suchetha; Raman, R.; Madhumitha; Meena, S.; Sruthi; Philip, N. Classification of Retinal Lesions in Fundus Images Using Atrous Convolutional Neural Network. In Futuristic Communication and Network Technologies. VICFCNT 2020; Lecture Notes in Electrical Engineering; Springer: Singapore, 2022; Volume 792, pp. 551–564. [Google Scholar] [CrossRef]
  26. Dutta, A.; Agarwal, P.; Mittal, A.; Khandelwal, S. Detecting grades of diabetic retinopathy by extraction of retinal lesions using digital fundus images. Res. Biomed. Eng. 2021, 37, 641–656. [Google Scholar] [CrossRef]
  27. Gayathri, S.; Gopi, V.; Palanisamy, P. Diabetic retinopathy classification based on multipath CNN and machine learning classifiers. Phys. Eng. Sci. Med. 2021, 44, 639–653. [Google Scholar] [CrossRef]
  28. Chetoui, M.; Akhloufi, M. Explainable end-to-end deep learning for diabetic retinopathy detection across multiple datasets. J. Med. Imaging 2020, 7. [Google Scholar] [CrossRef]
  29. Wei, Q.; Li, X.; Yu, W.; Zhang, X.; Zhang, Y.; Hu, B.; Mo, B.; Gong, D.; Chen, N.; Ding, D.; et al. Learn to Segment Retinal Lesions and Beyond. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 7403–7410. [Google Scholar] [CrossRef]
  30. Ni, J.; Chen, Q.; Liu, C.; Wang, H.; Cao, Y.; Liu, B. An Effective CNN Approach for Diabetic Retinopathy Stage Classification with Dual Inputs and Selective Data Sampling. In Proceedings of the 2019 18th IEEE International Conference On Machine Learning And Applications (ICMLA), Boca Raton, FL, USA, 16–19 December 2019; pp. 1578–1584. [Google Scholar] [CrossRef]
  31. Randive, S.; Senapati, R.; Bhosle, N. Spherical Directional Feature Extraction with Artificial Neural Networkfor Diabetic Retinopathy Classiftcation. In Proceedings of the 2018 IEEE 13th International Conference on Industrial and Information Systems (ICIIS), Rupnagar, India, 1–2 December 2018; pp. 152–157. [Google Scholar] [CrossRef]
  32. Loheswaran, K. Optimized KFCM Segmentation and RNN Based Classification System for Diabetic Retinopathy Detection. In ICCCE 2020; Lecture Notes in Electrical Engineering; Springer: Singapore, 2021; Volume 698, pp. 1309–1322. [Google Scholar] [CrossRef]
  33. Shorfuzzaman, M.; Hossain, M.; El Saddik, A. An explainable deep learning ensemble model for robust diagnosis of diabetic retinopathy grading. ACM Trans. Multimed. Comput. Commun. Appl. 2021, 17, 113. [Google Scholar] [CrossRef]
  34. Suresh, M.; Indira, S.; Ramachandran, S. Classification of Fundus Images Based on Non-binary Patterns for the Automated Screening of Retinal Lesions. Lect. Notes Netw. Syst. 2021, 204, 773–787. [Google Scholar] [CrossRef]
  35. Sharif, M.; Shah, J. Automatic screening of retinal lesions for grading diabetic retinopathy. Int. Arab. J. Inf. Technol. 2019, 16, 766–774. [Google Scholar]
  36. Wang, R.; Chen, B.; Meng, D.; Wang, L. Weakly Supervised Lesion Detection From Fundus Images. IEEE Trans. Med. Imaging 2019, 38, 1501–1512. [Google Scholar] [CrossRef] [PubMed]
  37. Kaur, J.; Mittal, D. Estimation of severity level of non-proliferative diabetic retinopathy for clinical aid. Biocybern. Biomed. Eng. 2018, 38, 708–732. [Google Scholar] [CrossRef]
  38. DelaPava, M.; Ríos, H.; Rodríguez, F.; Perdomo, O.; González, F. A deep learning model for classification of diabetic retinopathy in eye fundus images based on retinal lesion detection. In Proceedings of the 17th International Symposium on Medical Information Processing and Analysis, Campinas, Brazil, 17–19 November 2021; SPIE: Bellingham, WA, USA, 2021; Volume 12088. [Google Scholar] [CrossRef]
  39. Abdelmaksoud, E.; El-Sappagh, S.; Barakat, S.; Abuhmed, T.; Elmogy, M. Automatic Diabetic Retinopathy Grading System Based on Detecting Multiple Retinal Lesions. IEEE Access 2021, 9, 15939–15960. [Google Scholar] [CrossRef]
  40. Biswas, S.; Upadhya, R.; Das, N.; Das, D.; Chakraborty, M.; Purkayastha, B. An Intelligent System for Diagnosis of Diabetic Retinopathy. Adv. Intell. Syst. Comput. 2020, 1139, 97–110. [Google Scholar] [CrossRef]
  41. Krause, J.; Gulshan, V.; Rahimy, E.; Karth, P.; Widner, K.; Corrado, G.S.; Peng, L.; Webster, D.R. Grader Variability and the Importance of Reference Standards for Evaluating Machine Learning Models for Diabetic Retinopathy. Ophthalmology 2018, 125, 1264–1272. [Google Scholar] [CrossRef] [PubMed]
  42. Martens, J.B. The Hermite Transform-Theory. IEEE Trans. Acoust. Speech Signal Process. 1990, 38, 1595–1606. [Google Scholar] [CrossRef]
  43. Mira, C.; Moya-Albor, E.; Escalante-Ramírez, B.; Olveres, J.; Brieva, J.; Vallejo, E. 3D Hermite transform optical flow estimation in left ventricle CT sequences. Sensors 2020, 20, 595. [Google Scholar] [CrossRef]
  44. Hennelly, B.M.; Sheridan, J.T. Image encryption techniques based on the fractional Fourier transform. In Optical Information Systems; Javidi, B., Psaltis, D., Eds.; International Society for Optics and Photonics, SPIE: Bellingham, WA, USA, 2003; Volume 5202, pp. 76–87. [Google Scholar] [CrossRef]
  45. Chang, C.C.; Hu, Y.S.; Lin, C.C. A digital watermarking scheme based on singular value decomposition. In Combinatorics, Algorithms, Probabilistic and Experimental Methodologies. ESCAPE 2007; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Berlin/Heidelberg, Germany, 2007; Volume 4614 LNCS, pp. 82–93. [Google Scholar] [CrossRef]
  46. Ovalle-Magallanes, E.; Avina-Cervantes, J.G.; Cruz-Aceves, I.; Ruiz-Pinales, J. Transfer learning for stenosis detection in X-ray Coronary Angiography. Mathematics 2020, 8, 1510. [Google Scholar] [CrossRef]
  47. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. ImageNet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar] [CrossRef]
  48. Yosinski, J.; Clune, J.; Bengio, Y.; Lipson, H. How transferable are features in deep neural networks? In Advances in Neural Information Processing Systems 27 (NIPS ’14); Ghahramani, Z., Welling, M., Cortes, C., Lawrence, N., Weinberger, K., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2014; pp. 3320–3328. [Google Scholar]
  49. Tajbakhsh, N.; Shin, J.Y.; Gurudu, S.R.; Hurst, R.T.; Kendall, C.B.; Gotway, M.B.; Liang, J. Convolutional Neural Networks for Medical Image Analysis: Full Training or Fine Tuning? IEEE Trans. Med. Imaging 2016, 35, 1299–1312. [Google Scholar] [CrossRef] [PubMed]
  50. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. (IJCV) 2015, 115, 211–252. [Google Scholar] [CrossRef]
  51. Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. arXiv 2016, arXiv:1612.08242. [Google Scholar]
  52. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015—Conference Track Proceedings, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  53. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar] [CrossRef]
  54. Jocher, G.; Chaurasia, A.; Qiu, J. Ultralytics YOLOv8. 2023. Available online: https://github.com/ultralytics/ultralytics (accessed on 31 August 2023).
  55. Gomez-Coronel, S.L.; Moya-Albor, E.; Pérez-Daniel, K.R.; Brieva, J.; Cruz-Aceves, I.; Hernandez-Aguirre, A.; Soto-Alvarez, J.A. Authentication of medical images through a hybrid watermarking method based on Hermite-Jigsaw-SVD. In Proceedings of the 18th International Symposium on Medical Information Processing and Analysis, Valparaíso, Chile, 9–11 November 2022; Brieva, J., Guevara, P., Lepore, N., Linguraru, M.G., Rittner, L., Castro, E.R., Eds.; International Society for Optics and Photonics, SPIE: St Bellingham, WA, USA, 2023; Volume 12567, p. 125671G. [Google Scholar] [CrossRef]
  56. Poynton, C.A. A Technical Introduction to Digital Video; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 1996. [Google Scholar]
  57. Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.S.; Davis, A.; Dean, J.; Devin, M.; et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems. 2015. Available online: www.tensorflow.org (accessed on 31 August 2023).
  58. Chollet, F. Keras. 2015. Available online: https://keras.io (accessed on 1 August 2023).
  59. Horé, A.; Ziou, D. Image Quality Metrics: PSNR vs. SSIM. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 2366–2369. [Google Scholar] [CrossRef]
  60. Rao, Y.R.; Prathapani, N.; Nagabhooshanam, E. Application of normalized cross correlation to image registration. Int. J. Res. Eng. Technol. 2014, 3, 12–16. [Google Scholar]
Figure 1. The 25 examples of the MESSIDOR-2 fundus image dataset.
Figure 1. The 25 examples of the MESSIDOR-2 fundus image dataset.
Mathematics 12 00734 g001
Figure 2. Medicine symbol Caduceus used as watermark.
Figure 2. Medicine symbol Caduceus used as watermark.
Mathematics 12 00734 g002
Figure 3. Proposed hybrid watermarking algorithm.
Figure 3. Proposed hybrid watermarking algorithm.
Mathematics 12 00734 g003
Figure 4. Steered Hermite transform schema.
Figure 4. Steered Hermite transform schema.
Mathematics 12 00734 g004
Figure 5. Inverse Steered Hermite transform schema.
Figure 5. Inverse Steered Hermite transform schema.
Mathematics 12 00734 g005
Figure 6. Examples of the JST a fundus image of 960 × 1140 pixels, the number of blocks (N) and the size of each one ( s 1 × s 2 ) were varied. (a) Fundus image. (b) k × l = 96 and s 1 × s 2 = 120 × 120 . (c) k × l = 384 and s 1 × s 2 = 60 × 60 . (d) k × l = 1536 and s 1 × s 2 = 30 × 30 . (e) k × l = 6144 and s 1 × s 2 = 15 × 15 . (f) k × l = 1536 and s 1 × s 2 = 30 × 30 . k × l = 1,382,400 and s 1 × s 2 = 1 × 1 .
Figure 6. Examples of the JST a fundus image of 960 × 1140 pixels, the number of blocks (N) and the size of each one ( s 1 × s 2 ) were varied. (a) Fundus image. (b) k × l = 96 and s 1 × s 2 = 120 × 120 . (c) k × l = 384 and s 1 × s 2 = 60 × 60 . (d) k × l = 1536 and s 1 × s 2 = 30 × 30 . (e) k × l = 6144 and s 1 × s 2 = 15 × 15 . (f) k × l = 1536 and s 1 × s 2 = 30 × 30 . k × l = 1,382,400 and s 1 × s 2 = 1 × 1 .
Mathematics 12 00734 g006
Figure 7. Classic architecture of a convolutional neural network.
Figure 7. Classic architecture of a convolutional neural network.
Mathematics 12 00734 g007
Figure 8. Watermarking insertion schema.
Figure 8. Watermarking insertion schema.
Mathematics 12 00734 g008
Figure 9. (a) Luma component ( I Y ( x , y ) ). (b) Selected Steered Hermite coefficient ( I 2 , θ ( x , y ) ).
Figure 9. (a) Luma component ( I Y ( x , y ) ). (b) Selected Steered Hermite coefficient ( I 2 , θ ( x , y ) ).
Mathematics 12 00734 g009
Figure 10. Watermarking extraction schema.
Figure 10. Watermarking extraction schema.
Mathematics 12 00734 g010
Figure 11. Sensitivity analysis of the scaling factor α , for the ten selected images (A–J images), by varying the scale factor. (a) PSNR values. (b) SSIM values.
Figure 11. Sensitivity analysis of the scaling factor α , for the ten selected images (A–J images), by varying the scale factor. (a) PSNR values. (b) SSIM values.
Mathematics 12 00734 g011
Figure 12. Images used for sensitivity analysis of Figure 11: images (AJ).
Figure 12. Images used for sensitivity analysis of Figure 11: images (AJ).
Mathematics 12 00734 g012
Figure 13. Best results examples. Original images: (a,d,g). Watermarked images: (b,e,h). Recovery watermark: (c,f,i).
Figure 13. Best results examples. Original images: (a,d,g). Watermarked images: (b,e,h). Recovery watermark: (c,f,i).
Mathematics 12 00734 g013
Figure 14. Worst results examples. Original images: (a,d,g). Watermarked images: (b,e,h). Recovery watermark: (c,f,i).
Figure 14. Worst results examples. Original images: (a,d,g). Watermarked images: (b,e,h). Recovery watermark: (c,f,i).
Mathematics 12 00734 g014
Figure 15. Results examples of the watermarked and attacked images (GF, MF, GN, SP) and the corresponding extracted watermark, showing satisfying results (b,f,j,n) and those with worse performance (d,h,l,p).
Figure 15. Results examples of the watermarked and attacked images (GF, MF, GN, SP) and the corresponding extracted watermark, showing satisfying results (b,f,j,n) and those with worse performance (d,h,l,p).
Mathematics 12 00734 g015
Figure 16. Results examples of the watermarked and attacked images (CE, EQ, JPEGC, SC) and the corresponding extracted watermark, showing satisfying results (b,f,j,n) and those with worse performance (d,h,l,p).
Figure 16. Results examples of the watermarked and attacked images (CE, EQ, JPEGC, SC) and the corresponding extracted watermark, showing satisfying results (b,f,j,n) and those with worse performance (d,h,l,p).
Mathematics 12 00734 g016
Figure 17. Results examples of the watermarked and attacked images (CE, EQ, JPEGC, SC) and the corresponding extracted watermark, showing satisfying results (b,f,j) and those with worse performance (d,h,l).
Figure 17. Results examples of the watermarked and attacked images (CE, EQ, JPEGC, SC) and the corresponding extracted watermark, showing satisfying results (b,f,j) and those with worse performance (d,h,l).
Mathematics 12 00734 g017
Figure 18. YOLOv8 top 1 accuracy.
Figure 18. YOLOv8 top 1 accuracy.
Mathematics 12 00734 g018
Figure 19. YoloV8 (a) Original image. (b) Watermarked image. (c) Inference match.
Figure 19. YoloV8 (a) Original image. (b) Watermarked image. (c) Inference match.
Mathematics 12 00734 g019
Figure 20. VGGC16 inference results. (a) VGG16 Accuracy. (b) VGG16 Loss.
Figure 20. VGGC16 inference results. (a) VGG16 Accuracy. (b) VGG16 Loss.
Mathematics 12 00734 g020
Figure 21. VGG16 inference results. (a) Original image. (b) Watermarked image. (c) Inference match.
Figure 21. VGG16 inference results. (a) Original image. (b) Watermarked image. (c) Inference match.
Mathematics 12 00734 g021
Figure 22. InceptionV3 inference results. (a) InceptionV3 Accuracy. (b) InceptionV3 Loss.
Figure 22. InceptionV3 inference results. (a) InceptionV3 Accuracy. (b) InceptionV3 Loss.
Mathematics 12 00734 g022
Figure 23. InceptionV3 inference results. (a) Original image. (b) Watermarked image. (c) Inference mismatch.
Figure 23. InceptionV3 inference results. (a) Original image. (b) Watermarked image. (c) Inference mismatch.
Mathematics 12 00734 g023
Figure 24. InceptionV3 inference results. (a) Original image. (b) Watermarked image. (c) Inference match.
Figure 24. InceptionV3 inference results. (a) Original image. (b) Watermarked image. (c) Inference match.
Mathematics 12 00734 g024
Figure 25. ResNet50 inference results. (a) ResNet50 Accuracy. (b) ResNet50 Loss.
Figure 25. ResNet50 inference results. (a) ResNet50 Accuracy. (b) ResNet50 Loss.
Mathematics 12 00734 g025
Figure 26. ResNet50 slight inference results. (a) Original image. (b) Watermarked image. (c) Inference match.
Figure 26. ResNet50 slight inference results. (a) Original image. (b) Watermarked image. (c) Inference match.
Mathematics 12 00734 g026
Figure 27. ResNet50 complete inference results. (a) Original image. (b) Watermarked image. (c) Inference match.
Figure 27. ResNet50 complete inference results. (a) Original image. (b) Watermarked image. (c) Inference match.
Mathematics 12 00734 g027
Table 1. VGG16, ResNet50, and InceptionV3 CNN implementation.
Table 1. VGG16, ResNet50, and InceptionV3 CNN implementation.
NetworkNumber of LayersType of FiltersResidual Connections
VGG1616 3 × 3 No
InceptionV348 1 × 1 , 3 × 3 , 5 × 5 No
ResNet5050 3 × 3 Yes
Table 2. Average performance metrics for the extracted watermark by varying the scale factor.
Table 2. Average performance metrics for the extracted watermark by varying the scale factor.
MSEPSNR (dB)NCCSSIMMSSIM
4.87209 × 10 6 53.503730.999360.993300.99354
Table 3. Average performance metrics using the 1748 fundus images of the MESSIDOR-2 dataset. The insertion section corresponds to the watermarked images and the extraction section to the recovered watermarks.
Table 3. Average performance metrics using the 1748 fundus images of the MESSIDOR-2 dataset. The insertion section corresponds to the watermarked images and the extraction section to the recovered watermarks.
Insertion Extraction
MSEPSNR (dB)NCCSSIMMSSIMMSEPSNR (dB)NCCSSIMMSSIM
4.6976 × 10 6 53.86380.99930.99370.99380.000732.06900.99750.99370.9943
Table 4. Definition of the attacks applied and their corresponding parameters.
Table 4. Definition of the attacks applied and their corresponding parameters.
Attack TypeOperation NameParameter NameParameter Value
Image ProcessingGaussian Filter (GF)Filter size N × N
Median Filter (MF)Window size N × N
Gaussian Noise (GN)Variance σ
Salt and Pepper Noise (SP)Noise densityd
Contrast Enhancement (CE)Percent saturation P C E (%)
Histogram Equalization (EQ)Equalization Levels number L V
JPEG Compression (JPEGC)Quality Percentage Q P (%)
Image Scaling (SC)Scaling factor S F
GeometricRotation (ROT)Rotation angle ϕ (°)
Cropping (CROP)Cropping percentage P C R (%)
Translation (TRAN)Displaced pixels number Δ x , Δ y
Table 5. Average performance metrics for the extracted watermark using the 1748 fundus images of the MESSIDOR-2 dataset and applying the attacks: GF, MF, GN, and SP.
Table 5. Average performance metrics for the extracted watermark using the 1748 fundus images of the MESSIDOR-2 dataset and applying the attacks: GF, MF, GN, and SP.
Attack/ParameterParameter ValueMSEPSNR (dB)NCCSSIMMSSIM
GF/( N × N ) 3 × 3 6.9476 × 10 4 32.141170.997660.994090.99457
5 × 5 7.2134 × 10 4 32.069830.997570.993830.99442
7 × 7 3.4721 × 10 3 30.063350.989590.966750.97031
9 × 9 2.2502 × 10 1 18.690120.714250.666720.67248
Average 5.7477 × 10 2 28.241120.924770.905350.90795
MF/( N × N ) 2 × 2 5.7127 × 10 4 32.699450.998080.995330.99534
3 × 3 5.5048 × 10 4 32.831690.998150.995510.99550
4 × 4 5.5852 × 10 4 32.788040.998120.995450.99544
5 × 5 5.9831 × 10 4 32.775500.998000.995010.99502
Average 5.6965 × 10 4 32.773670.998090.995320.99532
GN/( σ )0.10 2.3761 × 10 3 31.780010.995460.991080.99191
0.20 3.1808 × 10 3 31.592090.994370.989490.99052
0.40 3.2338 × 10 3 31.441540.994170.988850.99010
0.50 3.2449 × 10 3 31.409800.994130.988690.98998
Average 3.0089 × 10 3 31.555860.994530.989520.99063
SP/(d)0.20 3.2188 × 10 3 31.490940.994220.988980.99016
0.40 3.2418 × 10 3 31.430300.994160.988770.99002
0.60 3.2713 × 10 3 31.370930.994070.988520.98987
0.75 3.2793 × 10 3 31.333290.994040.988400.98979
Average 3.2528 × 10 3 31.406360.994120.988670.98996
Table 6. Average performance metrics for the extracted watermark using the 1748 fundus images of the MESSIDOR-2 dataset and applying the attacks: CE, EQ, JPEG, and SC.
Table 6. Average performance metrics for the extracted watermark using the 1748 fundus images of the MESSIDOR-2 dataset and applying the attacks: CE, EQ, JPEG, and SC.
Attack/ParameterParameter ValueMSEPSNR (dB)NCCSSIMMSSIM
CE/( P C E ) 0.5 % 3.5275 × 10 2 28.507710.946520.911860.91404
1.0 % 2.9811 × 10 2 29.821470.957420.935470.93700
2.0 % 4.4006 × 10 2 29.354760.938530.911390.91290
3.0 % 4.0111 × 10 2 29.323480.941570.909690.91140
4.0 % 3.6314 × 10 2 29.486120.946820.916290.91815
Average 3.7104 × 10 2 29.298710.946170.916940.91870
EQ/( L V )8 6.6205 × 10 4 32.065280.997770.994600.99466
32 7.5019 × 10 4 31.513110.997470.993920.99397
64 7.6718 × 10 4 31.417610.997420.993780.99384
128 7.7192 × 10 4 31.390450.997400.993740.99380
256 7.7103 × 10 4 31.392860.997400.993750.99381
Average 7.4448 × 10 4 31.555860.997490.993960.99402
JPEGC/( Q P ) 90 % 8.5412 × 10 4 31.392710.997130.992520.99359
80 % 2.4153 × 10 2 24.049040.945050.882270.89217
70 % 9.7825 × 10 2 19.767810.833120.717750.72688
60 % 1.5889 × 10 1 18.573340.762770.653390.66094
50 % 1.5515 × 10 1 17.748610.763660.646560.65428
Average 8.7375 × 10 2 22.306300.860340.778500.78557
SC/( S F )0.25× 8.1853 × 10 1 0.869670.003460.130620.13400
0.50× 5.9573 × 10 1 5.241810.394170.268750.27398
1.50× 1.8756 × 10 3 30.877950.994130.982530.98540
1.75× 2.1085 × 10 3 30.739330.993470.980170.98318
2.00× 3.9879 × 10 3 30.092810.988530.963740.96752
Average 2.8445 × 10 1 19.564310.674750.665160.66882
Table 7. Average performance metrics for the extracted watermark using the 1748 fundus images of the MESSIDOR-2 dataset and applying the attacks: ROT, CROP and TRAN.
Table 7. Average performance metrics for the extracted watermark using the 1748 fundus images of the MESSIDOR-2 dataset and applying the attacks: ROT, CROP and TRAN.
Attack/ParameterParameter ValueMSEPSNR (dB)NCCSSIMMSSIM
ROT/( ϕ ) 5 9.9763 × 10 4 31.918580.996690.990970.99239
15 1.0727 × 10 3 31.876680.996490.990330.99181
45 1.1402 × 10 3 31.524960.996280.989540.99103
65 7.9339 × 10 4 32.116630.997340.993010.99383
90 1.4098 × 10 3 28.801350.995270.986680.99021
190 7.0773 × 10 4 32.114810.997630.993910.99438
Average 1.0202 × 10 3 31.392170.996620.990740.99227
CROP/( P C R ) 10 % 5.3050 × 10 4 33.015040.998210.995690.99565
20 % 5.3289 × 10 4 33.004320.998210.995670.99563
30 % 5.5777 × 10 4 32.785660.998120.995450.99545
40 % 1.4770 × 10 3 28.497380.995050.985990.98966
50 % 9.0257 × 10 3 21.707040.971440.892020.90520
60 % 1.0349 × 10 1 12.060250.784170.471570.48271
Average 1.9270 × 10 2 26.844950.957530.889400.89405
TRAN/ ( Δ x , Δ y ) (100, 100) px 5.4622 × 10 4 32.882750.998160.995550.99553
(250, 250) px 5.4341 × 10 4 32.896090.998170.995580.99555
(400, 400) px 5.4083 × 10 4 32.930440.998180.995600.99557
(550, 550) px 5.3478 × 10 4 32.993040.998200.995660.99562
Average 5.4131 × 10 4 32.925580.998180.995600.99557
Table 8. Metrics of the extracted watermarks for attacked images from Figure 15.
Table 8. Metrics of the extracted watermarks for attacked images from Figure 15.
Attack/ParameterRow (Figure 15)Parameter ValuePSNR (dB)NCCMSSIM
GF/( N × N )1 (left) 3 × 3 34.443870.998790.99697
1 (right) 9 × 9 16.770390.934240.77541
MF/( N × N )2 (left) 3 × 3 34.403090.998780.99694
2 (right) 5 × 5 27.857570.994490.98804
GN/( σ )3 (left)0.134.376110.998770.99692
3 (right)0.521.467120.976590.92703
SP/(d)4 (left)0.634.269860.998740.99685
4 (right)0.7516.623600.932400.72361
Table 9. Metrics of the extracted watermarks for attacked images from Figure 16.
Table 9. Metrics of the extracted watermarks for attacked images from Figure 16.
Attack/ParameterRow (Figure 16)Parameter ValuePSNR (dB)NCCMSSIM
CE/( P C E )1 (left) 4 % 34.498840.998810.99701
1 (right) 4 % 12.996990.857360.51012
EQ/( L V )2 (left)12834.403090.998780.99694
2 (right)828.037740.994720.98875
JPEGC/( Q P )3 (left) 90 % 34.443870.998790.99697
3 (right) 50 % 13.457550.869720.69389
SC/( S F )4 (left)1.50x34.403090.998780.99694
4 (right)0.25x0.869100.037470.14016
Table 10. Metrics of the extracted watermarks for attacked images from Figure 17.
Table 10. Metrics of the extracted watermarks for attacked images from Figure 17.
Attack/ParameterRow (Figure 17)Parameter ValuePSNR (dB)NCCMSSIM
ROT/( ϕ )1 (left) 45 34.457540.998800.99698
1 (right) 15 9.705830.739190.30531
CROP/( P C R )2 (left) 10 % 34.554520.998820.99705
2 (right) 50 % 24.330870.987730.97136
TRAN/( Δ x , Δ y )3 (left) ( 550 , 550 ) px34.526590.998820.99703
3 (right) ( 100 , 100 ) px29.463970.996190.99115
Table 11. Comparison values of different algorithms.
Table 11. Comparison values of different algorithms.
Watermarking TechniquePSNR (dB)NCCSSIMMSE
Anushikha Singh et al. [9]158.41831.0000--
Zhen Dai et al. [10]46.9631---
A. George Klington et al. [14]54.1572---
Ranjana Dwivedi et al. [16]46.8600-0.9914-
Divyanshu Awasthi et al. [17]39.45810.99570.99865.0534 × 10−5
67.14751.00001.00008.2033 × 10−8
Xiyao Liu et al. [18]41.29950.9607--
Muhammad Fachri et al. [19]50.52280.9607--
Payal Garg et al. [22]51.7040---
Proposed scheme53.86380.99930.99374.6976 × 10−6
Table 12. Comparison of different algorithms about dataset and type of watermarking.
Table 12. Comparison of different algorithms about dataset and type of watermarking.
TechniqueImage DatasetWatermark TypeCapacity (Size)
Anushikha Singh et al. [9]42Digital patient ID-
Zhen Dai et al. [10]40Binary image32 × 32
A. George Klington et al. [14]1000Fundus image & textual information329,960 bits
Ranjana Dwivedi et al. [16]10Color image512 × 512
Divyanshu Awasthi et al. [17]1QR code256 × 256
Xiyao Liu et al. [18]40Hospital logo32 × 32
Muhammad Fachri et al. [19]-Binary image64 × 64
Payal Garg et al. [22]1000Fingerprints and gait images-
Proposed scheme1748Binary image256 × 256
Table 13. Comparison of different algorithms with our algorithm using Median Filter.
Table 13. Comparison of different algorithms with our algorithm using Median Filter.
TechniqueParameter ValueNCC
Ranjana Dwivedi et al. [16] 2 × 2 0.9937
3 × 3 0.8945
Xiyao Liu et al. [18] 3 × 3 0.9759
5 × 5 0.9215
Proposed scheme 2 × 2 0.9980
3 × 3 0.9981
5 × 5 0.9980
Table 14. Comparison of different algorithms with our algorithm using JPEG Compression.
Table 14. Comparison of different algorithms with our algorithm using JPEG Compression.
TechniqueParameter ValueNCC
Ranjana Dwivedi et al. [16] 90 % 0.9988
Xiyao Liu et al. [18] 80 % 0.9982
70 % 0.9939
Proposed scheme 90 % 0.9971
80 % 0.9450
70 % 0.8331
Table 15. Comparison between Xiyao Liu et al. [18] and our algorithm after Cropping attack.
Table 15. Comparison between Xiyao Liu et al. [18] and our algorithm after Cropping attack.
TechniqueParameter ValueNCC
Xiyao Liu et al. [18] 10 % 1.0000
20 % 0.9803
Proposed scheme 10 % 0.9982
20 % 0.9982
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Moya-Albor, E.; Gomez-Coronel, S.L.; Brieva, J.; Lopez-Figueroa, A. Bio-Inspired Watermarking Method for Authentication of Fundus Images in Computer-Aided Diagnosis of Retinopathy. Mathematics 2024, 12, 734. https://doi.org/10.3390/math12050734

AMA Style

Moya-Albor E, Gomez-Coronel SL, Brieva J, Lopez-Figueroa A. Bio-Inspired Watermarking Method for Authentication of Fundus Images in Computer-Aided Diagnosis of Retinopathy. Mathematics. 2024; 12(5):734. https://doi.org/10.3390/math12050734

Chicago/Turabian Style

Moya-Albor, Ernesto, Sandra L. Gomez-Coronel, Jorge Brieva, and Alberto Lopez-Figueroa. 2024. "Bio-Inspired Watermarking Method for Authentication of Fundus Images in Computer-Aided Diagnosis of Retinopathy" Mathematics 12, no. 5: 734. https://doi.org/10.3390/math12050734

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop