1 Introduction

Transmission of Electronic Patient Record (EPR) between medical research centers, schools, and hospitals is necessary. Sending of EPR is essential for educational purposes and e-healthcare applications. Many applications, such as teleconsulting, telediagnosis, and remote surgery, require EPR. EPR encompasses confidential medical information about the patient, medical situation, diagnosis, and treatment [8]. Since EPR is transmitted through a public network such as the Internet, the confidentiality protection of EPR is a real concern in the Health Information System (HIS). Watermarking is one solution to the confidentiality protection problem in HIS. Besides watermarking, various other methods have been proposed for improving the confidentiality protection of EPR. One approach is to embed encrypted EPR into a cover image with encryption algorithms such as chaotic encryption [25] and compressive sensing [34]. Another scheme is to insert the EPR into selected areas of the cover image rather than the whole cover image and use the binary location map as a key for embedding [9, 36]. In the extraction process, the binary key used for the embedding process is required to extract the embedded EPR.

In the watermarking systems, capacity and imperceptibility are two vital concerns [10]. One concern in the watermarking schemes is to maximize embedding capacity. Another concern is to avoid the watermarked image to suffer subtle distortions after the embedding process. Imperceptibility means that the watermarked image should be perceptually indistinguishable from the original cover image. Another concern in watermarking systems is the concept of blindness [10]. Blind systems have been thoroughly investigated [6, 12, 23, 30], and they generally lead to higher complexity. However, blind systems are practically superior to non-blind systems, which require various side information for the extraction or recovery on the receiver side [3, 17, 31].

Due to the rapid growth of machine learning and artificial intelligence in the last decades, several research works have used machine-learning tools for watermarking [2, 13, 18, 20, 22, 33]. Some researches propose prediction based watermarking methods for natural images [22, 33]. In natural images, pixels have a high correlation with their neighbors and can be predicted based on neighbor values. The prediction process may utilize various machine-learning tools such as the Extreme Learning Machine (ELM) [33] and Lagrangian Support Vector Regression (LSVR) [22]. Heidari et al. [13] embed the watermark data redundantly in multiple spectral zones. For extraction, they use SVM to detect a zone with minimum distortion, from which watermark is extracted. Abdelhakim et al. [2], use K-Nearest Neighborhood (K-NN) regression to find the optimum value of the embedding parameter. Among machine learning tools, deep Convolutional Neural Networks (CNN) have become popular in many computer vision applications such as object detection [27], pattern recognition [39], image classification [15], and watermarking [18, 20]. For instance, Kandi et al. [18] use auto-encoders for feature extraction. In their method, watermark data is embedded in the feature space, generated by the auto-encoder, rather than embedding in a transform domain such as DCT.

Medical images are divided into two regions, ROI and NROI. Since the ROI region includes critical information for diagnostic purposes, small distortions of ROI may cause problems in medical diagnosis. On the other hand, NROI contains nonessential medical information. Hence, small deformations of NROI are tolerable and do not cause serious issues. In watermarking systems introduced by [2, 3, 13], the whole cover image is considered for embedding, and this distortion caused during embedding is irreversible. Two approaches for medical watermarking are utilized to assure that diagnostic information is not distorted during watermarking or undesired distortion can be restored. The first approach is ROI-based watermarking, in which the cover image is divided into two regions, ROI and NROI. Then the watermark is embedded only in NROI. Thus, ROI remains intact, and only NROI is distorted during embedding. The second approach is the lossless watermarking. In this approach, the whole cover image is considered for embedding with lossless methods, so that in the recovery process, the original cover image is recovered without any loss of information.

One limitation of ROI based watermarking approaches is that they cannot detect the ROI map on the receiver side. Consequently, the ROI map needs to be sent as side information to the receiver side for the extraction of the watermark [11, 32]. Hence, regardless of their segmentation method, they are non-blind algorithms, which require side information for the extraction of watermark data or the recovery of the original cover image. On the other hand, some other researches [7, 26, 37] do not send an ROI map as side information. Therefore, they are blind algorithms, and the watermark data can be extracted solely from the watermarked image. Yang et al. [37] and Chaitanya et al. [7] embed ROI map inside the watermarked image to make it blind. However, the watermarking capacity is reduced as the embedded ROI map occupies part of data capacity. The most relevant research to our work is [19]. Their algorithm recognizes the ROI map from the watermarked image on the receiver side and does not need to send an ROI map as additional information. They segment the image by Otsu and then utilize a histogram-shifting method for the embedding of the watermark. This embedding method processes pixel values that are close to the peak bin. Hence, the ROI map is not changed during the embedding process and can be recognized before the extraction process. Despite the elegant idea suggested in this work, it is only applicable to the specific watermarking method utilized in this work. Hence, it cannot be generalized for other watermarking approaches. Furthermore, the method of [19] lacks a wide variety of experimental evaluations, as it is only tested on three medical image samples.

In lossless watermarking approaches, the whole cover image is distorted during embedding since all regions of the image are considered for embedding. However, due to utilizing lossless methods, the original cover image is entirely recovered after watermark extraction [36], [4, 5, 14, 16, 21, 35, 38]. Despite complete recovery and extraction in the receiver side, some papers are non-blind, implying that some side information from the embedding module is required by the extraction or recovery modules on the receiver side [36], [4, 5, 38], [21].

In this paper, we propose a recursive method for blind diagnostically-lossless watermarking in medical applications, which is named BlessMark. We use a deep network for segmentation and generation of ROI map. Watermark is only embedded in NROI blocks within a novel iterative scheme. Therefore, sensitive medical information remains intact, which leads to a diagnostically-lossless watermarking system, where only NROI is distorted during the embedding process. The proposed framework is blind, as neither of the ROI map or the original cover image and information about watermark is transmitted to the receiver side. The utilized segmentation procedure helps further improve the confidentiality of the embedded information, due to the proprietary network weights, which are not publicly known by others. In other words, the ROI map may be considered as a key for confidentiality since watermark data is extracted solely from NROI blocks.

Furthermore, we train a classifier to detect distorted NROI regions. Hence, distorted areas of NROI are mostly recovered to the original form.

To test our proposed framework, we use a simple embedding method in the DCT domain. The choice of the DCT domain for embedding is just a convenient option for the proof of concept. Hence, different embedding methods in other transform domains, such as DWT and Hadamard, may be applied. Also, to evaluate the framework, we use a CNN for ROI segmentation and a three-layer fully connected neural network for the detection of distorted NROI blocks.

The main contributions of this work are as follows:

  1. 1)

    Introducing a framework that can be used as a platform for applying a desirable watermarking method to any medical image by utilizing various network structures in different transform domains.

  2. 2)

    In the extraction process, the same ROI map, which is used for the embedding process, is generated without any additional information. Therefore, the watermark is blindly extracted from NROI blocks.

  3. 3)

    NROI blocks, which are modified due to the embedding process, are recovered to their original state wherever possible.

The remainder of the paper is organized as follows: Details of the proposed framework is explained in Section 2. The experimental results are explained in Section 3. Finally, we conclude our proposed framework in Section 4.

2 General Watermarking Framework

In this section, we go through the technical details of the proposed watermarking framework, BlessMark. As shown in the block diagram of Fig. 1, the framework is composed of three separate modules for the embedding of the watermark, extraction of the watermark, and the recovery of the original cover image. There is a standard core block, ROI segmentation network, which is responsible for segmentation the ROI regions and generating the ROI map for the embedding, extraction and recovery modules. The problem is that embedding of the watermark data into the image may cause distortions, which could affect the segmentation results. We embed the watermark data into NROI. Hence, the ROI map is key for the extraction process. It is crucial to obtain the same segmentation on the watermarked image; otherwise, accurate extraction of the watermark data is not possible. To this end, we propose a novel iterative scheme for embedding based on the ROI segmentation function. This is done such that the ROI segmentation network can accurately detect the same ROI map in both transmitter and receiver sides. ROI map is automatically recognized on the receiver side without requiring any additional information. Therefore, the ROI map works as a key for confidentiality, as any third party cannot recognize it without having access to our proprietary segmentation tool.

Fig. 1
figure 1

Block diagram of the proposed blind ROI-based watermarking framework

On the other hand, due to embedding watermark data in NROI, crucial medical information in ROI remains intact. Another core block of the framework is a distortion detection network used in the recovery module. This network is responsible for the detection of distorted blocks due to the embedding process for the recovery operation in the next step.

In the following sections, we elaborate on the details of each block separately. We first introduce the proposed ROI segmentation network. In Sections 2.2 and 2.3, embedding and extraction modules are described. Details of the distortion detection network and the recovery module are presented in Section 2.4.

2.1 ROI segmentation network

The ROI Segmentation unit is responsible for the detection of ROI pixels from NROI pixels, and the generation of ROI block map based on the ROI detected pixels. A block is considered as NROI if all of its pixels are detected as NROI. Otherwise, the block is considered as ROI. In this work, we use a CNN structure inspired by U-Net [29], without max-pooling, up-sampling, and concatenation layers, for the segmentation of the ROI pixels. Figure 2 demonstrates the structure of the utilized network. As shown in Fig. 2, the ROI segmentation network is composed of 11 convolutional layers. Input to the network is a small m × m image block, while block size remains constant across all the convolution layers. Hidden layers consist of 3 × 3 convolutions followed by ReLU (Rectified Linear Unit) activation function. At the final layer, a 1 × 1 convolution is applied, followed by the same ReLU activation function. The number of channels in each layer is shown on the top of each layer in Fig. 2. The network output represents a segmentation probability map for pixels of the input block. This probability map is binarized by using a 0.5 threshold.

Fig. 2
figure 2

ROI segmentation network structure. Each blue box shows a multichannel feature map. Numbers above the boxes represent the number of channels

2.2 Embedding module

We propose a novel iterative approach for the embedding process. The block diagram of the embedding module is shown in Fig. 3. The ROI block map of the cover image is generated based on the strategy discussed in Section 2.1. Watermark is embedded into NROI blocks of the cover image. Then NROI and ROI blocks are merged to construct the tentative watermarked image. The ROI block map of the tentative watermarked image is then generated based on the strategy discussed in Section 2.1. One image block might be detected as NROI before embedding and the same block may be detected as ROI after the embedding. If the watermarking process causes a change in the ROI block map, then the modified cover image is constructed, and the watermark is embedded into the NROI blocks of the modified cover image. The embedding process may cause some of the NROI blocks of the original cover image to be identified as ROI. The changed versions of such NROI blocks are placed in the cover image and from there on, they will be considered as ROI. We repeat the embedding process into the modified cover image until the ROI block map remains unchanged, and the final watermarked image is produced. Thus, we do not need to send the ROI block map to the receiver side with the watermarked image, and the proposed framework is blind.

Fig. 3
figure 3

Block diagram of the BlessMark embedding module

figure f

We use a simple embedding method in the DCT domain for proof of concept. However, the proposed framework is not limited to this specific domain, and watermarking can be performed in other known transform domains. A pseudo-code of the embedding module is presented in Algorithm 1. In the first step, m × m NROI blocks are detected. Then one bit of watermark is embedded in every NROI block. The embedding process is continued until the whole watermark is embedded. When two DCT coefficients are swapped, we add a constant threshold to secure a minimum distance between the two coefficients. Changing DCT coefficients may lead to under/overflow in the spatial domain, i.e., when they are transformed back to the spatial domain, the pixel values may exceed the valid range [0, 255]. In this situation, the under/overflowed pixels are assigned 0 and 255, respectively.

figure g
figure h

2.3 Extraction module

In the extraction module, we blindly extract the watermark data from the watermarked image. The block diagram of the extraction module is shown in Fig. 4. The ROI map of the watermarked image is generated based on the strategy discussed in Section 2.1. We accurately attain the same ROI map from the final embedding loop. The watermark is extracted from the NROI blocks of the watermarked image. The pseudo-code of the extraction module is presented in Algorithm 2. In the first step, the m × m NROI blocks are detected. Then one bit of watermark is extracted from each NROI block. The extraction process is continued until the whole watermark is extracted.

Fig. 4
figure 4

Block diagram of the extraction module

2.4 Recovery module

The proposed recovery module performs recovery of distorted NROI blocks of the cover image to the extent feasible. We introduce a distortion detection classifier which is responsible for the detection of distorted NROI blocks that are altered during the embedding process from original NROI blocks. Then distorted blocks are recovered to the closest estimate of the original block. The block diagram of the recovery module is shown in Fig. 5. This module gets the NROI blocks of the watermarked image and the extracted watermark as its inputs and recovers the NROI blocks of the original cover image to the extent feasible. Recovered NROI blocks are merged with ROI blocks to construct the mostly recovered cover image. The pseudocode for the recovery module is presented in Algorithm 3. In the first step, m × m NROI blocks, which have been distorted through the embedding process, are detected. The changed blocks are recovered to their original state by reversing the embedding operation. Underflow/Overflow pixels are assigned 0 and 255 respectively.

Fig. 5
figure 5

Block diagram of the BlessMark recovery module

3 Experimental Results

The proposed framework is implemented in Tensor-flow [1] and executed on NVidia GeForce® GTX 1080 Ti. The framework is evaluated on two datasets of Retina [28] and X-ray Angiography [24], independently. The datasets are composed of 40 color images of size 584 × 565 and 44 grayscale images of size 512 × 512, respectively. We divided datasets into equal sets for training and test purposes. The CNN network and three-layer fully connected neural network performing ROI segmentation and distortion detection are separately trained on each dataset. We generate binary random sequences to be used as watermark data. For color images, watermark data is embedded in all three color channels. Thus, the distortion detection classifier is separately applied on single channels to detect the distorted blocks.

To evaluate the embedding method, we use the two standard measures, PSNR (Peak Signal to Noise Ratio) and SSIM (Structural Similarity Index). We utilize the Dice score to evaluate the segmentation network and the accuracy metric for evaluating the distortion detection classifier. Definitions of all the metrics are presented in Table 1. In the PSNR formula, w and h are the image dimensions. In this formula, I and IW are the cover image and the watermarked image, respectively. In the SSIM formula, I and IW are the cover and the watermarked images, respectively. In this formula, μ and σ2 represent mean and variance values, respectively. Also, \( {\sigma}_{I,{I}_w} \) is the covariance between I and IW and c1 and c2 are constants. In Dice score and Accuracy measures, variables TP and FP are cases that are predicted as positive, while actual outputs are positive and negative, respectively. Also, TN and FN are cases that are predicted as negative while the actual outputs are negative and positive, respectively.

Table 1 Evaluation metrics employed for testing the framework

All the networks are evaluated in Section 3.1. Performance of the whole framework is analyzed in Section 3.2. Finally, we compare BlessMark with state-of-the-art watermarking systems in Section 3.3.

3.1 Independent evaluation of networks

3.1.1 Segmentation network

As discussed in previous sections, the segmentation network is responsible for attaining ROI pixel map. This network takes a block as the input and produces an ROI pixel map for that. block. The network input is a gray-scale image block. Therefore, color images are converted to grayscale, and one ROI map is generated for all channels. A softmax activation function is used for creating a probability map in the last layer. We use a cross-entropy loss function with stochastic gradient descent (SGD) optimizer for training. Table 2 presents the Dice scores of the trained networks on two datasets with different block sizes. Two training sets constructed from 20 Retina images and 22 Angiography images are used separately for training each network within 150 epochs. The other half of the datasets is used as a test set. As shown in Table 2, the trained networks recognize 73% of ROI pixels for Retina and 61% of the ROI pixels for Angiography test sets, when the block size is 6 × 6. Dice scores increase by using larger block sizes, usually. Also, it is shown that a trained network for retina reaches a higher Dice score than the trained network for angiography test sets for the same block size. As an example, the produced ROI pixel maps for retina and angiography are shown in Fig. 6.

Table 2 Performance of segmentation network for two datasets with different block sizes
Fig. 6
figure 6

ROI segmentation for Retina and Angiography image. (a) Image. (b) Produced ROI pixel map

3.1.2 Distortion detection network

As discussed in previous sections, the distortion detection network is responsible for detecting the distorted blocks, i.e., the blocks that have distorted during the embedding process. In this work, we train a three-layer fully connected neural network for distortion detection. The structure of this network is shown in Fig. 7. The network input is a small m × m NROI block. In the first step, the block is flattened by a flatten layer. The two hidden layers are dense layers with m2 node and each layer is followed by a ReLU activation function. At the final layer, a dense layer with one neuron is used, which is followed by a sigmoid activation function. We use Adam with default parameters as an optimizer and cross-entropy as a loss function. The network output represents a classification probability, in which this probability is converted to a binary by thresholding the output layer on 0.5.

Fig. 7
figure 7

Distortion detection network structure

Every block has an intrinsic embedded value. Hence, for arranging a training set for the network, we pick all the NROI blocks of all channels and invert the intrinsic embedded value inside them. The training set is composed of all the NROI blocks and their inverted versions. Thus, the three-layer fully connected neural network is trained to classify these two sets of blocks for all channels. It worth mentioning that we can embed different watermark data in various channels. Consequently, the classifier conducts independent analyses on separate channels.

The accuracy of the classifier for the two datasets is demonstrated in Table 3. Training set in the Retina and the Angiography datasets are used for the training of the two networks with 100 epochs. Test sets are constructed similar to the training sets by using a test set in the Retina and the Angiography datasets. Index i in Table 3 represents the two DCT coefficients c(i, i + 1) and c(i + 1, i), which are swapped for embedding. As shown in Table 3, DCT coefficients swapped for embedding in various block sizes (‘6 × 6’, ‘8 × 8’, ‘10 × 10’) are different. The parameter th = 0.01 represents the threshold as used in our embedding algorithm (Algorithm 1). As shown in Table 3, the trained networks detect the distorted blocks with an accuracy of 94% for Retina and 97% for Angiography, when the block size is 6 × 6.

3.2 Evaluation of the whole framework

In Table 4, we evaluate our framework for the 20 Retina and 22 Angiography test images in terms of imperceptibility and variations in the ROI map. Since the image is segmented by CNN, false classifications are probable, i.e., it is probable that one pixel is segmented as NROI by CNN while it is labeled as ROI in the ground truth or vice versa. Therefore, we use the ground truth ROI map for calculating PSNR of images.

Table 4 Performance of the proposed framework on Retina and Angiography test sets
Table 3 Performance of classifier for two datasets with different block sizes. th = 0.01
Table 5 Training times of two networks on Retina training set
Table 6 Average times of the proposed framework on Retina test set
Table 8 Comparison of average PSNR and side information with other method for different block sizes
Table 7 Qualitative comparison of the proposed framework with other methods
Table 9 Details of the segmentation and detection networks

In Table 4, PSNR of the whole image, as well as PSNR of ROI and NROI regions, are presented. For Retina image with block size 6 × 6, PSNR of the watermarked image NROI is 54.96 (dB) on the maximum capacity of 0.021 (BPP). The image NROI imperceptibility is enhanced to 59.81 (dB) after recovery, i.e., imperceptibility of NROI may be improved by 4.85 (dB). The recovery process is not perfect due to the distortion detection classifier errors, irreversible distortions caused by under/overflow clipping of embedding operation, and switching of segmentation results after embedding. We also demonstrate the average percentage of NROI blocks, which are switched to ROI blocks during the embedding process. It is shown that distortions caused by the iterative embedding process are minimal and decrease by using larger block sizes.

Boxplot of Fig. 8 shows the distribution of PSNR and SSIM for watermarked images across the Retina and Angiography test sets. Each box corresponds to block size. For block sizes 6 × 6, 8 × 8, 10 × 10, maximum experimental capacities are 0.021, 0.010, and 0.006 (BPP) for Retina and 0.022, 0.012 and 0.007 (BPP) for Angiography test sets.

Fig. 8
figure 8

Box-plot of watermarked image PSNR and SSIM. Each box represents the distribution of PSNR and SSIM versus block size. (a) Retina. (b) Angiography

In Table 5, the training durations, for the segmentation and classifier networks, are shown for a retina training set with a 6 × 6 block size. We used 2 M blocks to train the segmentation network in 150 epochs. Also, the classifier network is trained by using 820,950 blocks in 100 epochs. The segmentation and the classification networks are trained in 23 h and 4.5 h, respectively.

In Table 6, embedding, extraction, and recovery times with GPU and without GPU are evaluated for the Retina test set. Average experimental results for 20 Retina test images with block size 6 × 6 and capacity 0.021 (BPP) is demonstrated in Table 6. Embedding time with GPU is 7 s. Extraction and recovery time with GPU is 3 s. These times are increased to 78 and 18 s without GPU.

In Fig. 9, the visual qualities of the embedding result with block size 6 × 6 for two test images are shown. The left column shows the original cover image, the middle column is the embedded watermark, and the right column demonstrates the watermarked image with its corresponding PSNR and SSIM. We have not seen any distinguishable difference between the watermarked image and the cover image.

Fig. 9
figure 9

(a), (c) Original cover images, (b), (d) watermarked images

3.3 Comparison with State-of-the-art Methods

Properties of various watermarking algorithms [36], [11, 32], [4, 5, 7, 14, 16, 19, 21, 35, 37, 38] are compared with our framework in Table 7. The eight lossless watermarking methods [36], [4, 5, 14, 16, 21, 35, 38] use the whole cover image for embedding, causing distortions across all image regions including ROI. Therefore, they are not applicable to medical watermarking, since diagnostic information in ROI may be corrupted. Also methods of [36], [4, 5, 38], [21] are non-blind.

The two non-blind ROI based watermarking methods [11, 32] send an ROI map with the watermarked image to the receiver side. The two blind ROI based watermarking methods [7, 37] embed ROI map inside the watermarked image. Hence, the extraction of an ROI map from the watermarked image is a prior step for the extraction of the watermark data. The ROI based watermarking method [19] utilize the same segmentation method for recognizing the ROI map in embedding and extraction modules of their proposed system. In our framework, the watermark is embedded in NROI blocks to keep the sensitive ROI information intact during the embedding process. ROI map can be recognized from the watermarked image in the extraction module by the trained segmentation network.

In Table 8, imperceptibility and side-information of one lossless method [38] are compared with our system. Similar to Table 4, we use the ground truth ROI map for calculating PSNR of ROI and NROI. Note that the false classification of CNN is probable, i.e., a pixel may be segmented as NROI by CNN while it is labeled as ROI in the ground truth or vice versa.

The lossless method [38] embeds the watermark into 1-level. of IWT of the whole image in two iterations so that iteration-2 compensate produced distortion in iteration-1.

In Table 8, the PSNR and side-information results are calculated over the 20 Retina test images. For a fair comparison, a single watermark is embedded in all methods. The infinity values of PSNR in the table demonstrate that the original cover image has been completely recovered. The proposed framework does not produce any side information, while the method of [38] produce 16,516.5 bits of side information on the capacity of 0.020 (BPP). However, the lossless method [38] can completely recover the original cover image. Since NROI does not contain valuable information for medical diagnosis, accurate recovery of NROI is not critical in medical applications.

Figures 2 and 7 show the general structure of the ROI segmentation network and the distortion detection network. Details of layers of these two networks are listed in Table 9.

4 Conclusion

In this paper, we presented BlessMark, a framework for blind diagnostically-lossless watermarking. The proposed watermarking scheme is used as a means for the simultaneous improvement of confidentiality and preservation of diagnostic medical information. BlessMark consists of a deep neural network for image segmentation and one fully connected neural network for classification.

BlessMark ROI segmentation network generates an ROI map for the embedding, extraction, and recovery modules. The ROI segmentation network is applied using an iterative scheme to accurately produce the same ROI map, both in the transmitter and receiver sides. Hence, in the proposed blind watermarking framework, an ROI map is automatically generated on the receiver side without requiring any additional information. The. proposed ROI map is critical for improving confidentiality protection in our system. Third parties cannot create the ROI map without having access to our proprietary segmentation tool. Since the watermark is embedded only in NROI blocks, ROI remains intact and leading to a diagnostically-lossless watermarking system. The distortion detection classifier used in the recovery module helps the detection of blocks that have distorted during the embedding process. Distorted blocks are mostly recovered to their original form.

The choice of a simple embedding method in the DCT domain is just a convenient option to prove the concept of the proposed framework. Hence, different embedding methods in other transform domains such as DWT and Hadamard may be applied.

Furthermore, we used a CNN and a three-layer fully connected neural network for the ROI segmentation and detection of distorted NROI blocks. However, other structures can be investigated for these purposes. Our watermarking method is non-robust since the ROI map of the watermarked image can be changed as a result of attacks. Since the ROI map is vital for embedding and extraction, it is necessary to detect the same ROI map on both sides.

We evaluated our framework on two Retina and Angiography datasets. The proposed framework demonstrates producing any side information in comparison with one lossless watermarking method. Our proposed method is advantageous to recent methods such as the work of [19]. We are offering a framework while other research works implement a specific embedding routine.

One future research is to explore the capabilities of the proposed framework by using it as a platform for testing other networks and embedding domains. We are also planning to work on the robustness of the proposed framework such that the watermark can be extracted accurately in the presence of attacks. Furthermore, the new robust framework needs to be compared against more recent watermarking schemes, some of which are reviewed in the introduction section.