Open Access
29 December 2021 Survey on deep learning applications in digital image security
Zhenjie Bao, Ru Xue
Author Affiliations +
Abstract

In the digital era, sharing pictures on social media has become a common privacy issue. To prevent private images from being eavesdropped on and destroyed, developing secure and efficient image steganography, image cryptography, and image authentication has been difficult. Deep learning provides a solution for digital image security. First, we make an overall conclusion on deep learning applications in image steganography to generate five aspects: the cover image, stego-image, embedding change probabilities, coverless steganography, and steganalysis. Second, we also combine and compare deep learning methods used in six aspects: image cryptography from image compression, image resolution improvement, image object detection and classification, key generation, end-to-end image encryption, and image cryptoanalysis. Third, we collect deep learning methods in image authentication from five perspectives: image forgery detection, watermarked image generation, image watermark extraction and detection, image watermarking attack, and image watermark removal. Finally, we summarize future research directions of deep learning utilization in image steganography, image cryptography, and image authentication.

1.

Introduction

In such a highly information-oriented era, digital image transmission and delivery on the Internet have been increasingly frequent. the deliverer hopes that this only occurs in a secure channel while on the Internet; however; there are many eavesdroppers and destroyers. To protect individual privacy on public network platforms, researchers need to find an approach that satisfies both private image security and robustness. Image encryption, image steganography, and image authentication are three efficient methods that balance the characteristics of images and the requirement of security. Deep learning is a powerful tool in image processing that has reached impressive successes in image object detection,13 image classification,69 image segmentation,1013 image style transfer,1417 image denoising,1820 and image compression.2123 Applying deep learning methods to the field of image security to solve the traditional problems has also received extensive attention and achieved breakthrough progress recently. But how to better use the advantages of deep learning in image steganography, image cryptography, and image authentication always attracts many scholars’ attention. To help relevant researchers understand the field of deep learning applications in digital image security and its future development more quickly, in this paper, we order the origin and development process of deep learning methods in image steganography, cryptography, and authentication from multiple aspects, as can be seen in Fig. 1; we then compare these methods, analyze the advantages and disadvantages of each, and finally suggest future research directions of this field.

Fig. 1

The overall framework of the survey on digital image security.

OE_60_12_120901_f001.png

This survey covers around 90 papers about deep learning for image steganography, cryptography, and authentication. The main contributions of this paper can be summarized as follows.

  • 1. This survey collects and analyzes the deep learning techniques in the field of digital image security from image steganography, image cryptography, and image authentication.

  • 2. This survey estimates and compares the steganography, encryption, and watermarking performance of these approaches from quantitative indicators to reach the existing challenges.

  • 3. This survey suggests research trends for deep learning in the use of image steganography, cryptography, and authentication to collision sparks in a wider research field.

The remainder of this paper is formed as follows. Section 2 collects the related works for traditional algorithms of image steganography, image encryption, and image authentication and compares them with the deep learning methods. Section 3 illustrates the deep learning applications for image steganography as well as a comparison and future research directions of these methods. Section 4 discusses the deep learning mechanisms in image cryptography, presents a performance comparison of the image encryption methods, and then points out the existing challenges. Section 5 represents the deep learning techniques in image authentication and explains the desire for future research. Section 6 suggests the future scope of deep learning in image security. Section 7 elaborates the survey’s conclusions.

2.

Related Works

2.1.

Image Steganography

The purpose of image steganography is to send the stego-images like innocent normal images to the receiver and avoid the secret data being noticed by the attacker. As can be seen in Fig. 2, the process of image steganography is to embed the secret data into the cover image and then arrive at the stego-image that steganalysis such as spatial rich model (SRM)24 and maxSRMd2,25 which are two feature-based classifiers, and Xu-Net,26 which is a convolutional neural network classifier, tries to distinguish the stego-image from the cover image.

Fig. 2

The process of image steganography and steganalysis.

OE_60_12_120901_f002.png

The least significant bit (LSB) replacement is a classical steganography algorithm that replaces the LSBs of the cover image with the secret data bits.27 However, the LSBs of the pixel of the cover image take up a small part of the cover image, so the capacity is limited. Meanwhile, if we modify more bits of the pixel of the cover image to get a larger capacity, the possibilities to be detected by the attacker are higher. For higher capacity and higher undetectability, Pevný et al.28 introduced HUGO, which provided a larger capacity while having equal safety compared with LSB matching.

Considering the relationship between the content of the image itself and the size of the secret image, adaptive steganography selects the edge29 or texture30 areas of the cover image as the embedding location as much as possible to strengthen the invisibility of stego-images. Li et al.31 proposed a cost function for spatial image steganography that used a high-pass filter to locate the less predictable parts in an image, and then utilized two low-pass filters to make the low-cost values more clustered; it achieved better performance on resisting SRM steganalysis over HUGO28 and S-UNIWARD.30

Deep learning image steganography methods can reach a higher capacity and have higher undetectability by not only the traditional feature-based steganalysis model but also deep learning steganalysis models over the traditional steganography algorithms. The capacity is measured by bits-per-pixel (bpp), which is the average number of bits concealed into each pixel of the cover image.32 An ideal image steganography method could embed as much secret data as possible into the cover image without being detected by steganalysis methods. For example, Ref. 33 described that 0.2, 0.3, 0.4, and 0.5 bpp of secret data were embedded into the image from the BOSSbase34 dataset using three algorithms. As illustrated in Table 1, the detection error rates of three steganalysis methods SRM,24 maxSRMd2,25 and Xu-Net26 on the deep learning model in Ref. 33 are all higher than on traditional image steganography methods such as HILL31 and S-UNIWARD,30 which shows that deep learning-based image steganography has superiority in escape steganalysis.

Table 1

Different steganalysis models detection error rate of different image steganography methods.

Steganalysis modelsSteganography methods0.2 bpp (%)0.3 bpp (%)0.4 bpp (%)0.5 bpp (%)
SRMRef. 3338.5233.6329.1124.89
HILL38.4032.4827.8222.88
S-UNIWARD33.8427.4121.9717.72
maxSRMd2Ref. 3334.7030.2625.9722.98
HILL32.9727.5323.8620.63
S-UNIWARD30.4225.2320.5417.16
Xu-NetRef. 3342.6438.5333.5629.71
HILL36.8031.0625.7621.91
S-UNIWARD37.5530.4324.3419.25

2.2.

Image Encryption

Image encryption keeps image content invisible until someone has the correct key. As we can see in Fig. 3, image cryptography involves encrypting the plaintext image to the ciphertext image by the encryption key and decrypting the ciphertext image to the decrypted image by the decryption key. Cryptanalysis involves cracking the cryptography to get the encryption/decryption keys or the secret image; these includes ciphertext only attacks,35 known-plaintext attacks,36 chosen plaintext attacks,37 and chosen ciphertext attacks.38

Fig. 3

The process of image cryptography and cryptanalysis.

OE_60_12_120901_f003.png

Chaos is sensitive to initial states and possesses a complex and unpredictable long-term behavior,39 so it is widely used in image encryption. Bentoutou et al.40 introduced an efficient image encryption method based on chaotic maps and the Advanced Encryption Standard. Naim et al.41 described a satellite image encryption algorithm based on the linear-feedback shift register generator, SHA 512 hash function, hyperchaotic systems, and Josephus problem.

Based on the characteristic of optical systems,4244 cellular automata,4547 quantum,4850 and DNA computing,5153 there are many applications in image encryption.

In most deep learning image encryption methods, deep learning plays a supporting role, which gives full play to their special advantages. End-to-end deep learning image encryption is a new research direction and has achieved security that is the same as or even better than the traditional encryption methods. Table 2 compares three different image encryption methods in ciphertext image entropy; correlation coefficients of two horizontal, vertical, diagonal adjacent pixels; the number of pixels change rate (NPCR), and the unified averaged changed intensity (UACI) on satellite images, and the best values are shown in bold. As we can see in Table 2, the end-to-end deep learning image algorithm in Ref. 54 has better values in correlation coefficients and NPCR than the other two traditional image encryption methods. Furthermore, in other indexes, the deep learning method also has a similar value to the other two methods.

Table 2

Ciphertext image security performance comparisons of different image encryption schemes.

MethodsImage entropyHorizontal correlation coefficientsVertical correlation coefficientsDiagonal correlation coefficientsNPCRUACI
Ref. 407.99930.00080.00130.00170.99630.3527
Ref. 417.99770.0018−0.0020−0.00120.99620.3345
Ref. 547.99720.00040.00050.00110.99640.3349

2.3.

Image Authentication

Checking the image identity or image integrity is the target of image authentication. Image watermarking is a significant technique for authenticating images. As can be seen in Fig. 4, the process of image watermarking is like image steganography. The watermark is embedded into the container and arrives at the watermarked image. Watermark extraction or detection involves extracting or detecting and further authenticating the watermark. From the perspective of watermark visibility, an image watermark can be divided into the visible watermark55 and the invisible watermark,56 whereas from the perspective of watermark robustness, an image watermark can be classified by fragile watermark,57 semifragile watermark,58 and robust watermark.59 Furthermore, from the mode of watermark extraction, an image watermark can be itemized into blind,60 semiblind,61 and nonblind extraction.62

Fig. 4

The scheme of image watermarking and watermark extraction or detection.

OE_60_12_120901_f004.png

Compared with the traditional image watermark methods, the deep learning methods have better imperceptibility and robustness. Table 3 describes the peak signal-to-noise ratio (PSNR) and normalized correlation (NC) between the extracted watermark from the watermarked image under multiattacks and the original watermark. Suppose that x and y are two images with the size of M×N and (i,j) is the pixel location in an image. The PSNR is calculated as Eq. (1), and NC is computed as Eq. (2), where if a=b, then f(a,b)=1, and otherwise, f(a,b)=0:63

Eq. (1)

PSNR=10log10{25521MNi=0M1j=0N1[x(i,j)y(i,j)]2},

Eq. (2)

NC=1MNi=0M1j=0N1f(a,b).

Table 3

Comparisons of different methods facing multiattacks.

Attack typesMetricsRef. 59Ref. 62Ref. 63
No attackPSNR (dB)48.4751.4558.91
NC1.001.001.00
Median filtering (3×3)PSNR (dB)32.1236.7136.42
NC0.440.780.82
Salt and pepper (δ: 0.005)PSNR (dB)32.7431.6654.12
NC0.810.921.00
JPEG compression (QF: 20)PSNR25.3816.3430.21
NC0.440.320.56

As can be seen in Table 3, when watermarked images face the attacks like median filtering of 3×3, salt and pepper noise with noise level δ of 0.005, and JPEG compression with quality factor 20, the deep learning method in Ref. 63 reaches almost all of the highest watermark quality assessment values, which are shown in bold, while it also obtains the most complete watermark with respect to no attack.

3.

Deep Learning in Image Steganography

With regards to the traditional problem of image steganography for embedding capacity and security, the deep learning method in image steganography helps a lot (Table 4).

Table 4

Some deep learning models for image steganography tasks.

SeriesReferencesCharacteristics
Learning to generate cover imagesVolkhonskiy et al. (2017)65Use deep convolutional GANs to generate the cover image
Shi et al. (2017)67Use WGAN to generate cover image and GNCNN to analyze steganography
Learning to generate stego-imageHayes et al. (2017)70Define a game between three parties represented by the neural network
Baluja (2017)71Embed a color image into another image of the same size using deep neural networks
Chattopadhyay et al. (2018)72Use multistage feed-forward artificial neural network
Zhu et al. (2018)73Jointly train encoder and decoder networks
Hu et al. (2018)74Map the secret data into a noise vector used by the trained generator to produce the carrier image
Zhang et al. (2019)76Hide binary data in images using GANs
Wang et al. (2019)77Utilize a new secret information preprocessing method, Inception-ResNet block, GAN, and perceptual loss
Zhang et al. (2019)32Use an inception-module-based neural network to embed a secret gray image into an image with the same size, which was the Y channel of the cover image, and propose a mixed loss function
Duan et al. (2020)78Use residual block in the hidden network to generate stego-image
Learning embedding change probabilities for every pixel in the cover imageTang et al. (2017)84A generator to learn the probability map, then the secret message embedding simulated by the TES that was presented by a neural network
Yang et al. (2019)33Has three modules: a generator with a U-Net architecture, a no-pretraining-required double-tanh function, and an enhanced steganography analyzer
Tang et al. (2021)86Employ reinforcement learning to learn the embedding policy
Coverless image steganography based on deep learningDuan et al. (2018)87A coverless image steganography based on a generative model
Luo et al. (2020)88A coverless image steganography method based on multiobject recognition
Liu et al. (2019)89A coverless image steganography algorithm based on image retrieval of DenseNet features and DWT sequence mapping
Zhang et al. (2019)91A diversity image style transfer network using multilevel noise encoding
Zhou et al. (2019)92Use a faster region-based CNN and a dictionary that defined the objects and corresponding codes
Duan et al. (2020)94Coverless information hiding method that constructs improved Wasserstein GAN model
Steganalysis methods based on deep learningQian et al. (2015)69A single neural network called GNCNN employing Gaussian function as the activation function
Xu et al. (2016)26Xu-Net, a CNN architecture in consideration of the knowledge of steganalysis
Yang et al. (2017)95Incorporate selection-channel awareness into modified Xu-Net
Zeng et al. (2017)96A hybrid CNN using the domain knowledge behind rich models for JPEG steganalysis
Ye et al. (2017)97An alternative approach to steganalysis of digital images based on a convolutional neural network named Ye-Net
Yedroudj et al. (2018)98Yedroudj-Net, which improves the architecture of the convolutional neural network
Zhang et al. (2018)99Zhu-Net, which adopts small-sized convolutional kernel for preprocessing, separable convolution to enhance the stego-signal, spatial pyramid pooling, and data augmentation
Boroumand et al. (2019)100SR-Net, which adopts an expanded front part in a deep residual neural network
Reinel et al. (2021)101GBRAS-Net using filter banks for preprocessing, and depth-wise, separable convolution, skip connections for feature extraction
Liu et al. (2021)102DFSE-Net involving diverse filter modules and squeeze-and-excitation modules
Iskanderani et al. (2021)103An efficient θ-nondominated sorting genetic algorithm-III based DCNN model
Singhal et al. (2021)104Blind steganalysis for multiple categories in spatial and JPEG images by the deep residual network

3.1.

Learning to Generate Cover Images

As can be seen in Fig. 5, using deep convolutional generative adversarial networks (DCGAN),64 Volkhonskiy et al.65 proposed steganographic generative adversarial networks (SGAN), which was trained on the Celebrities dataset. Liu et al.66 generated cover images and embedded secret messages using the LSB algorithm to deceive the steganography analyzer. Similar to SGAN, secure steganography based on generative adversarial networks (SSGAN)67 used wasserstein generative adversarial networks (WGAN)68 as the cover images generator and Gaussian-neuron convolutional neural networks (GNCNN)69 as the steganography analyzer trained on the CelebA66 database to improve the training performance and image quality. They opened the field of container generation, although they lacked considerations in steganography properties.

Fig. 5

The structure of SGAN and SSGAN.

OE_60_12_120901_f005.png

3.2.

Learning to Generate Stego-Image

As can be seen in Fig. 6, the deep learning model in stego-image generation consists of three parts: the generator translates the cover image and secret data to the stego-image, the restorer recovers the secret data from the stego-image, and the steganography analyzer determines if the input image is either the stego-image or the cover image.

Fig. 6

The framework of the deep learning model in stego-image generation.

OE_60_12_120901_f006.png

Hayes and Danezis70 simulated a communication scenario in which three neural networks trained on the CelebA dataset engaged: Alice and Bob communicated by concealing a secret message in the carrier image, and Eve eavesdropped on the image and distinguished the embedded image from the innocent image. Baluja71 designed three neural networks: a preparation network gained the features of the secret image, a hiding network concealed the extracted features into the cover image across all available bits, and a revealing network reconstructed the secret image from the stego-image. Chattopadhyay et al.72 adopted a multistage feed-forward artificial neural network to complete image steganography. Zhu et al.73 transferred the input message and cover image to a discriminator-indetectable image by a neural network and recovered the secret message from the encoded image by the other neural network simultaneously. Hu et al.74 trained a generator to translate a carrier-like image deceiving the discriminator using the vector calculated by the secret information and an extractor to reconstruct the vector and map it to the secret message reversely. The neural networks were trained on the CelebA and Food-10175 datasets. However, the disadvantages of DCGAN result in some drawbacks. For example, some generated stego-images were not sufficiently natural, the size of the stego-image was small, and the steganography capacity was not big adequate. Zhang et al.76 hid arbitrary binary data in images using generative adversarial networks (GAN). Wang et al.77 utilized a secret information preprocessing method, Inception-ResNet block, GAN, and perceptual loss to improve the undetectability, imperceptibility, and capacity, although the model had certain limitations on the length of secret information. Zhang et al.32 used an inception-module-based neural network to embed a secret gray image into an image with the same size, which was the Y channel of the cover image; a neural network to recover the secret image from the Y channel of the stego-image, which tried to minimize a mean square error (MSE), structural similarity (SSIM), and multiscale SSIM mixed loss function; and another neural network to judge whether the input image was the stego-image or not. Duan et al.78 employed the residual learning block in the hiding network to directly generate a stego-image that looked like the cover image and designed the reveal network to recover the secret image.

Table 5 estimates PSNR and SSIM indices between the cover image and stego-image obtained by SteGAN,70 HiDDeN,73 SteganoGAN,76 and HidingGAN77 with embedding capacities of 0.4, 0.4, 4.4, and 4 bpp, respectively, in the COCO79 dataset; ISGAN32 with embedding capacity of 8 bpp; Ref. 71 with embedding capacity of close to 24 bpp in the ImageNet80 dataset; and Ref. 78 with embedding capacity of 23.8 bpp in the ImageNet, LFW,81 or Pascal-VOC82 datasets. The invisibility requires the stego-image to be similar to the cover image. And the higher the values of PSNR and SSIM are, the more similar the two images are. PSNR estimates the error between corresponding pixels of two images, but it does not considered the human visual characteristics; the SSIM83 measures the image similarity from comparisons of luminance, contrast, and structure.

Table 5

Comparison of different deep learning methods in PSNR and SSIM between stego-images and cover images on different image datasets.

Methods (capacity, datasets)PSNR (dB)SSIM
SteGAN70 (0.4 bpp, COCO)21.430.69
HiDDeN73 (0.4 bpp, COCO)33.400.96
SteganoGAN76 (4.4 bpp, COCO)36.330.88
HidingGAN77 (4 bpp, COCO)33.160.96
ISGAN32 (8 bpp, ImageNet)34.890.97
Ref. 78 (23.8 bpp, ImageNet, LFW, or Pascal-VOC)40.620.98
Ref. 71 (close to 24 bpp, ImageNet)41.20.98

In practice, these two indices should both be considered. Table 6 compares the steganography capacity and cover image size of different deep learning methods. As can be seen in Fig. 7, the method proposed by Ref. 71 has the highest PSNR, SSIM, and capacity values, and its SSIM value is close to the theoretical maximum value 1, which is the best stego-image imperceptibility and the highest embedding capacity among all compared methods.

Table 6

Comparison of different deep learning methods in capacity and cover image size.

MethodsCover image sizeCapacity (bits/pixel)
SteGAN32×320.4
HiDDeN512×5120.203
Ref. 7464×640.009
SteganoGAN4.4
HidingGAN256×2564
ISGAN256×2568
Ref. 78256×25623.8
Ref. 71200×200Close to 24

Fig. 7

Histogram comparison of different deep learning methods in PSNR and SSIM between the stego-image and the cover image and embedding capacity.

OE_60_12_120901_f007.png

The deep learning methods generate the stego-image and recover the secret data automatically while having a large embedding capacity and a good performance in undetectability. But they do not recover the secret data 100% when the stego-image does not suffer any attacks. Deep learning methods are not completely accurate secret data extraction algorithms in steganography although they perform well in real complex communication situations. The secret data extraction accuracy is limited by the characteristics of the neural network. Furthermore, the fixed input and output size of deep learning model limits the size of the image that it could process.

3.3.

Learning Embedding Change Probabilities for Every Pixel in the Cover Image

As illustrated in Fig. 8, Tang et al.84 trained a generator to learn the probability map and then embedded the secret message simulated by the ternary embedding simulator (TES), which was presented by a neural network; finally the obtained modification map was added to the cover image to achieve the stego-image that attempts to deceive the steganography analyzer. On the basis of Ref. 84, Yang et al.33 presented a U-Net-based85 generator and replaced the TES with the double-tanh function, which does not need to be trained. In addition, six high-pass filters were integrated into the steganography analyzer. Similar to Refs. 33 and 84, Tang et al.86 employed reinforcement learning to learn the embedding policy, which performed pixel-level actions and rewards.

Fig. 8

The structure of deep learning model in probability map generation.

OE_60_12_120901_f008.png

The above three methods were tested in embedding 0.1, 0.2, 0.3, 0.4, and 0.5 bpp of secret data into the image from the BOSSbase dataset, and SRM was used to detect them. As illustrated in Table 7, the detection error rates of Ref. 86 are all higher than the other compared algorithms and have the best undetectability.

Table 7

Detection error rate of different image steganography methods using SRM.

Methods0.1 bpp (%)0.2 bpp (%)0.3 bpp (%)0.4 bpp (%)0.5 bpp (%)
ASDL-GAN8436.1931.2425.4121.9518.04
UT-GAN3343.5336.8732.2627.3022.52
SPAR-RL8645.1538.4332.6828.3023.80

3.4.

Coverless Image Steganography

The coverless image steganography does not change the cover image but transmits it directly, so it does not easily raise the suspicion of the attacker, which can effectively resist steganalysis from a new angle. As shown in Fig. 9, Duan et al.87 generated the stego-image directly using two generators and two discriminators trained on the CelebA dataset. Generator 1 translates the secret image to the cover image, and discriminator 1 distinguishes the stego-image from the cover image. Furthermore, generator 2 translates the stego-image to the secret image, and discriminator 2 determines whether the input image is generated or not. Luo et al.88 showed a coverless image steganography scheme based on multiobject recognition. Liu et al.89 retrieved images according to the features DenseNet90 trained on the ImageNet80 dataset and generated feature sequences using discrete wavelet transform (DWT) coefficients, which presented the secret data and had good robust and security performance in resisting most image attacks. Zhang et al.91 discussed a diverse image style transfer network trained on the Microsoft COCO dataset79 using multilevel noise encoding with the image style transfer results presenting different codewords. However, the images with particularly dense textures were not suitable for datasets of steganography. Zhou et al.92 created a dictionary that defined the objects and corresponding codes and used a faster region-based convolutional neural network trained on the PASCAL VOC 200782 and VOC 201293 datasets to detect objects in stego-images that would extract the secret message. Duan et al.94 trained an improved Wasserstein GAN model on the LFW dataset81 and transmitted and input the disguised images with the generator outputting the images similar to the secret images in visual mode.

Fig. 9

The structure of deep learning model in stego-image direct generation in Ref. 87.

OE_60_12_120901_f009.png

3.5.

Steganalysis Methods Based on Deep Learning

A deep learning model can automatically learn image features, which contributes to better classification precision. Utilizing the advantages of the deep learning model, Qian et al.69 applied a single neural network called GNCNN, which employed six convolutional layers for extracting the image features, three fully connected layers for classifying, and a Gaussian function as the activation function for separating the cover and stego-signals. Xu et al.26 proposed Xu-Net, which took absolute values of elements in the feature maps generated from the first convolutional layer to force the model to take into account the sign symmetry24 that existed in noise residuals and constrained the range of data values with the saturation regions of tanh in the first two convolutional layers, while constraining the strength of the model using 1 × 1 convolutions in the last three convolutional layers to avoid overfitting. Yang et al.95 incorporated selection-channel awareness into a modified Xu-Net architecture trained on the BOSSbase v1.01 dataset,34 which applied large weights to features learned from complex texture regions and small weights to features learned from smooth regions. Zeng et al.96 convolved the images with a set of kernels, then calculated the different quantized and truncated features, and finally used the CNN model trained on the ImageNet dataset to process the extracted features for JPEG steganalysis. Jian et al.97 presented an image steganalysis CNN-based model called Ye-Net, which was initialized with the filters used in calculating residual maps in SRM and integrated them with the truncated linear unit to suit the distribution of embedding signals (with low signal-to-noise ratio) and selection channel awareness. Yedroudj-Net98 involved a predefined convolutional layer for extracting the noise component residuals and adopted scale operations in the last three convolutional layers. Zhu-Net99 applied 3×3 convolutional kernels to reduce the number of parameters and model the features in a small local region, used separable convolution to enhance the stego-signal-to-noise ratio, and employed spatial pyramid pooling to aggregate the local features for strengthening the representation ability of features. SR-Net100 adopted an expanded front part in a deep residual neural network without max-pooling operations to minimize the use of heuristics and externally enforced elements. GBRAS-Net101 used filter banks in the preprocessing phase and depth-wise, separable convolution, and skip connections in the feature extraction phase. Liu et al.102 constructed DFSE-Net with diverse filter parts that combined three different scale convolution filters that could process information diversely and squeeze-and-excitation parts that could enhance the effective channels out from diverse filter parts. Based on an efficient θ-nondominated sorting genetic algorithm-III, Iskanderani et al.103 described a densely connected CNN (DCNN) for image steganalysis. θ NSGA-III was utilized to tune the initial parameters of the DCNN model. It could control the accuracy and f-measure of the DCNN model by utilizing them as the multiobjective fitness function. Singhal and Bedi104 proposed multiclass blind steganalysis that included 60 layers to record demographic features and residual mapping to retain weak stego-signals generated by embedding payload and thus making classification easier and utilized a high-pass filter to preprocess images.

Table 8 and Fig. 10 illustrate the detection accuracy of different deep learning steganalysis algorithms in multiple steganography conditions, which consist of using the WOW105 steganography algorithm with 0.2 and 0.4 bpp embedding capacity and using the S-UNIWARD30 steganography algorithm with 0.2 and 0.4 bpp embedding capacity on the BOSSbase dataset. Compared with other algorithms, GBRAS-Net101 achieves the highest steganalysis accuracy.

Table 8

Comparison of detection accuracy of deep learning steganalysis methods for WOW and S-UIWARD at embedding capacity of 0.2 and 0.4 bpp.

AlgorithmsWOW 0.2 bpp (%)WOW 0.4 bpp (%)S-UIWARD 0.2 bpp (%)S-UIWARD 0.4 bpp (%)
GNCNN6961.470.753.769.1
Xu-Net2667.579.360.972.7
Ye-Net9766.976.760.168.7
Yedroudj-Net9872.385.163.577.4
Zhu-Net9976.988.171.484.5
SR-Net10075.586.467.781.3
GBRAS-Net10180.389.873.687.1
DFSE-Net10275.385.165.978.5

Fig. 10

Steganalysis accuracy comparisons of the deep learning steganalysis techniques against WOW and S-UNIWARD algorithms with the embedding capacity of 0.2 and 0.4 bpp.

OE_60_12_120901_f010.png

Despite the image steganography method based on deep learning achieving an impressive result in embedding capacity and stego-image quality, these methods have a lack of robustness. The robustness of image steganography algorithms based on deep learning needs to be further discussed and improved in future studies. Furthermore, the input and output of the deep learning model are fixed, so the deep learning models only process the images of a fixed size. Developing a model that could process images of various sizes in the field of image steganography is a meaningful task.

4.

Deep Learning in Image Cryptography

4.1.

Image Compression in Image Encryption Algorithms

Image compression can increase the efficiency of image encryption by reducing the size of data. Chen et al.106 utilized a deep learning model trained on a color image dataset107 to compress and reconstruct the plaintext image and compound the chaotic system to encryption. Hu et al.108 employed stacked autoencoder for compression and chaotic logistic map to encrypt the compressed vector. Suhail and Sankar109 presented an application of image compression and encryption using an autoencoder and chaotic logistic map. Selvi et al.110 suggested a competent adaptive sigma filterized synorr certificateless signcryptive Levenshtein entropy coding-based deep neural learning technique trained on a dataset of chest x-ray images to develop the image encryption and compression (Table 9).

Table 9

Some deep learning models for image cryptography tasks.

SeriesReferencesCharacteristics
Deep learning for image compression in image encryption algorithmsChen et al. (2020)106Utilize a deep learning model to compress and reconstruct the plaintext image and compound chaotic system to encryption
Hu et al. (2016)108Stacked autoencoder for compression and chaotic logistic map for encryption
Suhail et al. (2020)109Autoencoder for compression and chaotic logistic map for encryption
Selvi et al. (2021)110A competent adaptive sigma filterized synorr certificateless signcryptive Levenshtein entropy coding-based deep learning technique
Deep learning for image resolution improvement or denoising in image encryption algorithmsZhang et al. (2019)111Employ ghost imaging as a transmission and encryption mode and a CNN to improve the reconstructed image resolution
Chen et al. (2019)113Utilize a dilated deep CNN denoiser that improves the resolution of the fractional Fourier transform-based decrypted images
Deep learning for image object detection and classification in image encryption algorithmsZhao et al. (2020)115Utilize the MTCNN to seek key feature points of human faces, then adopt a combination of chaotic logic diagrams and RC4 stream ciphers to encrypt features
Alqaralleh et al. (2021)116Apply elliptic curve cryptography, employ the neighborhood indexing sequence with burrow wheeler transform to encrypt the hash values, and utilize a deep belief network for the classification process to diagnose the existence of the disease
Asgari-Chenaghlu et al. (2021)117A method based on YoloV3 object detection and chaotic image encryption
Deep learning for image private key generation in image encryption algorithmsLi et al. (2018)118Train deep learning model to gain the features of iris image, then use the RS error correcting code to calculate the encryption key, finally encrypt the image using XOR operation
Ding et al. (2021)120Use the GAN to generate the private key
Jin et al. (2020)122The method based on deep neural network learning to induce the symmetric key creation
Maniyath et al. (2020)123Adopt a robust deep neural network that generates secret key resistive of different forms of attack and chaotic map to encrypt
Erkan et al. (2020)124Use sensitive key generation by deep convolution neural network to produce a diverse chaotic sequence for encrypting operations
Fratalocchi et al. (2021)125Train a neural architecture to learn the mapping algorithm between the key and the physical unclonable function
Deep learning for end-to-end image encryptionLi et al. (2020)127An optical image encryption learning scheme based on Cycle-GANs
Ding et al. (2021)128Employ Cycle-GAN to encrypt and decrypt the medical images like a style transfer task
Bao et al. (2021)129Adversarial autoencoder for image scrambling based on asymmetric encryption
Bao et al. (2021)54Employ the traditional diffusion technique to enhance the avalanche effect of Cycle-GAN-based image encryption methods
Image cryptanalysis method based on deep learningXu et al. (2021)132A deep learning method to attack the phase truncated Fourier transform encryption system
Hai et al. (2019)134Train a deep neural network model to learn and crack the Random Phase Encoding based optical cryptosystems
Wu et al. (2020)136A model trained with large numbers of ciphertext-plaintext pairs to crack the modified diffractive-imaging-based image cryptosystem
Chen et al. (2020)138A CNN directly converting the ciphertext image encrypted by the joint transform correlation structure to the original plaintext image
He et al. (2019)139A deep learning-based decrypted image generation approach to unravel the image encryption method in Ref. 140

4.2.

Image Resolution Improvement or Denoising in Image Encryption Algorithms

Zhang et al.111 employed ghost imaging as a transmission and encryption mode and a CNN that was trained on the IAPR TC-12 benchmark112 to improve the recovered image resolution. Chen et al.113 utilized a dilated deep CNN denoiser trained on the Waterloo exploration database,114 which improved the resolution of the fractional Fourier transform-based decrypted images and resistance against multiclass attacks.

4.3.

Image Object Detection and Classification in Image Encryption Algorithms

Zhao et al.115 applied the multitask cascaded convolution network (MTCNN) to seek five key feature points of human faces and then adopted a combination of chaotic logic diagrams and Rivest Cipher 4 stream ciphers to encrypt eigenvalues. At the same time, the face coordinates generated by MTCNN and user passwords were hash-converted and double-encrypted by a hash table, which reduced the size of the images to be encrypted. Alqaralleh et al.116 employed elliptic curve cryptography with an optimal key generated by hybridization of grasshopper with the fruit fly optimization algorithm, then used the neighborhood indexing sequence with burrow wheeler transform to encrypt the hash values, and finally adopted a deep belief network to classify the existence of the disease. Asgari-Chenaghlu et al.117 described a technique based on YoloV3 object detection and chaotic image encryption that had the ability of automatic image encryption on both full or user-selected regions.

4.4.

Image Private Key Generation in Image Encryption Algorithms

Li et al.118 trained a CNN on the CASIA iris database version 4.0119 to extract the feature of the iris image, then employed the RS error correcting code to encode the feature vector, and calculated the encryption key that was adopted to encrypt the plaintext image by the XOR operation. As can be seen in Fig. 11, Ding et al.120 trained a GAN trained on Montgomery County’s chest x-ray121 dataset to generate the private key and input a seed image that had a large key space, pseudorandomness, a one-time pad, high sensitivity to change, and resistance to different kinds of attacks; then the bit-wise XOR algorithm was adopted as an encryption and decryption algorithm. Jin and Kim122 illustrated a method based on deep neural network learning without sharing the preshared key between systems and created and used the key that was variably used through the symmetric key encryption system of the 3D cube algorithm, which provided good security. Maniyath and Thanikaiselvan123 proposed a robust deep neural network trained on an SIPI image database that generated a secret key resistive of multiattacks and applied a chaotic map to encrypt the image without any negative effect on image quality. Erkan et al.124 mechanized a CNN trained on the ImageNet database to generate sensitive keys and then produced initial values and controlled parameters for the hyperchaotic log-map; thus they obtained a diverse chaotic sequence for image encryption. Fratalocchi et al.125 decoupled the design of the physical unclonable functions from the key generation and trained a neural structure to learn the mapping between the key and the physical unclonable function, which could address the shortcomings of unreliability and weak unpredictability of cryptographic keys.

Fig. 11

The structure of the deep learning model in image private key generation and image encryption in Ref. 120.

OE_60_12_120901_f011.png

4.5.

End-to-End Image Encryption

As can be seen in Fig. 12, the main end-to-end image encryption deep learning scheme consists of the encryption network that generates a random-like ciphertext image, the decryption network that reconstructs the plaintext image, and the discriminator that distinguishes the ciphertext images from the pixel-random images.

Fig. 12

The main scheme of the deep learning model in end-to-end image encryption.

OE_60_12_120901_f012.png

Furthermore, cycle-consistent generative adversarial network (Cycle-GAN)126 has a good performance in image style transfer, in which the process of image encryption is regarded as translating the usual images to images with randomly distributed pixels. Thus the neural network structure of Cycle-GAN was widely used as the encryption or decryption network in end-to-end image encryption methods based on deep learning. The neural network structure, described in Fig. 13, down-samples the input, then extracts the feature map through nine residual blocks,9 and finally up-samples the feature map to output the image with the objective style. One style transfer process of Cycle-GAN can be seen in Fig. 14, where the generator translates the original image to the generated image with objective style and the discriminator distinguishes whether the image is a real image.

Fig. 13

The generator neural network architecture of Cycle-GAN.

OE_60_12_120901_f013.png

Fig. 14

One style transfer process of Cycle-GAN.

OE_60_12_120901_f014.png

Li et al.127 demonstrated an optical image encryption learning scheme based on Cycle-GAN that was trained by the plaintext-ciphertext training set of satellite images in which the ciphertext images were encrypted by double random phase encoding. Ding et al.128 employed Cycle-GAN trained on a dataset of chest x-rays121 to encrypt and decrypt medical images as a style transfer task. In addition, a neural network is projected to gain the interested object from the ciphertext image. Bao et al.129 constructed an encoder–decoder and discriminator framework trained on the Corel-1000 dataset130 to imitate the process of image scrambling and reconstruction in which the parameters of the encoder and decoder are different. However, the cipher pixels were not uniformly distributed, the decrypted images quality and the generalization ability of model were not good, and the plaintext and ciphertext image sensitivities were weak. Bao and Xue54 analyzed the causes of the weak avalanche effect in the neural network of Cycle-GAN and integrated the traditional diffusion algorithm into the Cycle-GAN-based image encryption methods trained on a dataset of satellite images scraped from Google Maps to enhance the avalanche effect, although its decryption performance was not well.

Table 10 compares the image encryption methods illustrated in Refs. 106, 109, 117, 120, 127128.129, and 54 in PSNR and SSIM values between recovered decrypted images and plaintext images, the key space, and the encryption speed per plaintext image with 256×256 resolution. Table 11 compares the image entropy, correlation coefficients of two horizontally, vertically, and diagonally adjacent pixels of the ciphertext image obtained by different methods, NPCR values, and UACI values when these methods face a chosen plaintext attack, especially, changing 1% of pixels of the plaintext image in Refs. 128 and 129. As can be seen in Tables 10 and 11, the end-to-end key generation and image encryption methods based on the deep learning model have a large key space and a quick encryption efficiency. However, correlation coefficients of two horizontally, vertically, and diagonally adjacent pixels of the ciphertext image and the resistance to the chosen plaintext attack need to be further improved generally.

Table 10

Comparison of PSNR and SSIM values between the decrypted image and the plaintext image, key space, and encryption efficiency in different deep learning methods.

MethodPSNR (dB)SSIMKey spaceEfficiency (s)
Ref. 106 (Sec. 4.1)32.55160.9456101350.85
Ref. 109 (Sec. 4.1)1.00
Ref. 117 (Sec. 4.3)0.097
Ref. 120 (Sec. 4.4)(28)196608
Ref. 127 (Sec. 4.5)30.16640.90810.044
Ref. 128 (Sec. 4.5)37.430.93(1010)27579360.07
Ref. 129 (Sec. 4.5)27.50870.9115(232)60674590.6423
Ref. 54 (Sec. 4.5)33.18000.9360(232)16698307+(28)196608

Table 11

Comparison of different deep learning methods in image entropy, correlation coefficients of two horizontally, vertically, and diagonally adjacent pixels of the ciphertext image, NPCR values, and UACI values facing a chosen plaintext attack.

MethodImage entropyHorizontal correlation coefficientsVertical correlation coefficientsDiagonal correlation coefficientsNPCRUACI
Ref. 106 (Sec. 4.1)7.9944−0.00240.00120.00350.99610.3357
Ref. 109 (Sec. 4.1)0.960.33
Ref. 117 (Sec. 4.3)−0.00210.00140.0031
Ref. 120 (Sec. 4.4)7.99860.03830.22590.11580.99590.2319
Ref. 127 (Sec. 4.5)0.08770.13790.0349
Ref. 128 (Sec. 4.5)7.960.9421 (change 1% pixels)
Ref. 129 (section 4.5)7.97720.02910.0363−0.02330.9045 (change 1% pixels)0.1237 (change 1% pixels)
Ref. 54 (Sec. 4.5)7.99720.00040.0005−0.00110.99640.3349

Some problems exist with end-to-end key generation and image encryption based on the deep learning model. For example, the histogram of the generated key and ciphertext image is not seriously randomly distributed, and as can be seen in Table 11, the UACI values facing chosen ciphertext or plaintext attacks is not high in general. Although Ref. 54 proved that the employment of traditional diffusion techniques to Cycle-GAN-based image encryption methods could enhance the ability against chosen plaintext attack, the complexity of the image encryption methods also increased. Using a deep learning method to take the place of the diffusion algorithm is a possible solution to reduce the time costs. Then the decrypted image quality could be further enhanced. In addition, the deep learning model needs a lot of training time, and the huge amount of model calculations results in a poor encryption/decryption speed. Thus the training and encryption/decryption efficiency of the end-to-end image encryption techniques based on Cycle-GAN could be further improved. Furthermore, because the model is trained according to a specific dataset, the generalization ability of the encryption and decryption model should be further discussed, analyzed, and improved. However, the automatic generation of keys and ciphertext images using deep learning methods has the advantages of convience, large key space and reduced reliance on complex cryptography design knowledge. How to realize the image cryptography on an even more profound level using deep learning models and thus making a breakthrough is still a potential research direction in the future.

4.6.

Image Cryptanalysis Method Based on Deep Learning

Some progress of image cryptanalysis using deep learning has been made, especially in optical image encryption. The deep learning model is trained with large ciphertext and corresponding plaintext images to learn the ability to crack the optical image encryption method.

Because of the nonlinear operation of phase truncation, the cryptography based on phase truncated Fourier transforms (PTFT)131 has high robustness against existing attacks.

Figure 15 illustrates the encryption processes of PTFT, assuming that the original image is f(x) FT(·) and IFT(·) are the Fourier transform and inverse Fourier transform, respectively. The Fourier transform is given in Eq. (3). Equation (4) gives the phase truncation operation PT(·). Suppose that R1(x) and R2(u) are a pair of independent random phase masks; the encryption procedure of ciphertext image g(x) is seen in Eqs. (5) and (6):131

Eq. (3)

F(u)=FT[f(x)]=|F(u)|exp(i2πφ(u)),

Eq. (4)

PT[F(u)]=|F(u)|,

Eq. (5)

g1(u)=PT[FT(f(x)·R1(x))],

Eq. (6)

g(x)=PT[IFT(g1(u)·R2(u))].

Fig. 15

Process diagram of phase truncation Fourier transform encryption.

OE_60_12_120901_f015.png

Xu et al.132 proposed a deep learning method to attack the PTFT encryption system131 and used a dataset with pairs of plaintext images on the MNIST handwritten dataset133 and corresponding ciphertext images constructed through the PTFT encryption system to train residual network,9 which automatically learned the decryption characteristics of the encryption system by reducing the MSE between the decrypted images obtained by the deep learning model with the secret image as shown in Fig. 16. However, the quality of recovered images obtained by this method was not good.

Fig. 16

The training process of deep learning decryption network for phase truncation Fourier transform encryption.

OE_60_12_120901_f016.png

Hai et al.134 trained a neural network to crack the random phase encoding-based optical cryptosystems.135 Wu et al.136 trained a model that involved a module for obtaining the features of the ciphertext image and a module for recovering the plaintext image according to the obtained features with a large numbers of ciphertext-plaintext image pairs to attack the modified diffractive-imaging-based image encryption137 cryptosystem. Chen et al.138 demonstrated a CNN trained with a large amount of ciphertext image data encrypted by the joint transform correlation structure and its corresponding plaintext image, directly converting the ciphertext image to the original plaintext image.

Deep learning also shows a certain ability to detect or crack other encryption methods. He et al.139 mapped the ciphertext images encrypted by the chaos-based image encryption algorithm demonstrated in Ref. 140 into the low-dimensional space and then regenerated visually consistent decrypted images utilizing a deconvolutional generator.

5.

Image Authentication

5.1.

Image Forgery Detection

Bondi et al.141 employed a CNN trained on the Dresden image database142 to extract characteristic camera model features, which were analyzed through iterative clustering techniques for image tampering detection and localization using characteristic footprints left on images by different camera models. Elaskily et al.143 exploited a CNN trained on the MICC-F220,144 MICC-F2000,144 MICC-F600,145 and SATs-130146 datasets to extract features from images to detect the copy-move forgery. Diallo et al.147 presented a camera identification CNN model trained with a mixture of different qualities of compressed and uncompressed images on the Dresden dataset142 for image forgery detection. Patil and Jariwala63 carried out the intensive and incremental learning phase and then implemented a hybrid CNN to detect the image and video forgery. Bappy et al.148 introduced a manipulation localization method using resampling features, long-short-term memory cells, and an encoder–decoder network to segment out manipulated regions from nonmanipulated ones. Xiao et al.149 suggested a splicing forgery detection algorithm with diluted adaptive clustering and a coarse-to-refined CNN trained on the CASIA,150 COLUMB,151 and FORENSICS152 datasets, which cascaded a coarse CNN and a refined CNN and extracted the differences in the image properties between untampered and tampered regions from image patches with different scales. However, the detection only focused on a single tampered region in an image owing to a restriction of the postprocessing approach. Zhang and Ni153 employed a cross-layer intersection mechanism to dense U-Net85 for image forgery detection and localization. Biach et al.154 described an encoder using an architecture that was topologically the same as that of Resnet-50;9 it analyzed the discriminating characteristics between the manipulated and nonmanipulated regions, and a decoder localized the manipulated regions. However, there were a few poorly detected images, especially on the NIST’16155 dataset. Moulin and Goel156 derived locally optimal statistical tests for identifying forgeries and showed a procedure for learning a forgery detector trained on the CIFAR-10157 and MNIST handwritten datasets. To combat image recapture attacks such as recapturing high-quality images from high-fidelity liquid crystal display screens, Zhu et al.158 described a recaptured image detection method based on CNN in which the local binary patterns coding coded maps were extracted as the input (Table 12).

Table 12

Some deep learning models for image authentication tasks.

SeriesReferencesCharacteristics
Deep learning in image forgery detectionBondi et al. (2017)141Employ a CNN to extract characteristic camera model features that are analyzed through iterative clustering techniques for image tampering detection
Elaskily et al. (2020)143A CNN is used for detecting the copy-move forgery and original images
Diallo et al. (2020)147A camera identification CNN model trained with a mixture of different qualities of compressed and uncompressed images
Pramod et al. (2021)63Ameliorate the image and video forgery detection’s efficiency utilizing hybrid CNN
Bappy et al. (2019)148A manipulation localization method that utilizes resampling features, long-short-term memory cells, an encoder–decoder
Xiao et al. (2020)149A splicing forgery detection method with diluted adaptive clustering and a coarse-to-refined CNN
Zhang et al. (2020)153Apply cross-layer intersection mechanism to dense U-Net for image forgery detection and localization
Biach et al. (2021)154A CNN method based on an encoder/decoder to locate the manipulated regions
Moulin et al. (2017)156Derive locally optimal statistical tests for identifying forgeries
Zhu et al. (2019)158A recaptured image detection method based on convolutional neural networks
Deep learning in image watermark generationVukotić et al. (2018)159A new family of transformations based on deep learning networks that were useful in image watermarking
Kandi et al. (2017)160An autoencoder CNN for watermark embedding and extraction
Fierro-Radilla et al. (2019)161A robust zero-watermarking algorithm
Ahmadi et al. (2019)162An end-to-end diffusion blind watermarking framework
Mun et al. (2019)163A reinforcement learning for robust and blind watermarking
Zhong et al. (2020)164An encoder encodes the watermark and then input to an embedder with the cover image to reach the watermarked image
Zhang et al. (2021)165A watermarking framework for protecting deep networks
Deep learning in image watermark extraction and detectionLi et al. (2021)167A single-exposure optical image watermarking framework
Huynh-The et al. (2019)169A blind image watermarking framework based on an encoder–decoder network watermark extraction model
Li et al. (2018)170A cooperative neural network to recognize the suspected watermark signal
Hayes et al. (2020)171Resilient signal watermarking via adversarial training
Chen et al. (2021)172A model based on deep learning technology that accurately identifies the watermark copyright
Deep learning in image watermarking attackWang et al. (2021)173Digital image watermark fakers using generative adversarial learning
Hatoum et al. (2021)175A fully convolutional neural network as a denoising attack on watermarked images
Sharma et al. (2020)176An adversarial watermarking attack based on a CNN-based autoencoder scheme
Deep learning in image watermark removalCheng et al. (2018)177A deep learning model for visible watermark removal task that consists of two parts: watermark detection and removal
Gandelsman et al. (2019)178A coupled “Deep-Image-Prior” network to remove image watermark
Hertz et al. (2019)179Estimate the visual motif matte and reconstruct the latent image without opaque and semitransparent visual motifs
Li et al. (2019)180A watermark processing framework using the conditional GAN
Pei et al. (2021)181A watermark removal structure including watermark extraction and image inpainting networks
Cun et al. (2020)183A multitask feature extractor and a watermarked region smoother
Shafieinejad et al. (2019)184Focus on backdoor-based watermarking
Chen et al. (2019)185A unified watermark removal framework based on fine-tuning and incorporated with an adaption of the elastic weight consolidation algorithm and unlabeled data augmentation
William et al. (2021)186A neural network “laundering” algorithm to remove black-box backdoor watermarks from neural networks

F1-score takes both false negatives and false positives into account. Table 13 shows that Ref. 154 achieves the highest F1-score and has the best performance in image forgery detection among all compared methods.

Table 13

F1-score comparisons of different image forgery detection methods on the CASIA151 and NIST’16156 datasets.

DatasetsRef. 153Ref. 149Ref. 154
CASIA v1.00.57220.67580.7362
NIST’160.51400.6389

Although deep learning has achieved good results in image forgery detection for several types of forgery, there is shortage of large and perfect datasets that include images tampered by methods for more types of forgery and research studies on deep learning forgery detection methods for more complete forgery types.

5.2.

Watermarked Image Generation

To dynamically adapt image watermarking algorithms, deep learning-based image watermarking schemes have attracted increased attention, and experiments and estimation results have confirmed the advantages of the deep learning mechanisms in image watermarking. Vukotić159 investigated a new family of transformations based on deep learning networks that were useful in image watermarking. As can be seen in Fig. 17, Kandi et al.160 proposed an autoencoder CNN for watermark embedding and extraction; the attack layer simulated different attacks, and the strength factor controlled the level of watermarked images robustness versus imperceptibility. Fierro-Radilla et al.161 demonstrated a zero-watermarking algorithm in which features of the image were gained by the CNN and combined with the watermark sequence using the XOR operation. Ahmadi et al.162 described two fully CNNs with the residual structure trained on the CIFAR-10 and Pascal VOC201293 datasets for watermarks embedding and extraction. Mun et al.163 exploited a reinforcement learning trained on the BOSSbase dataset for robust and blind watermarking. Zhong et al.164 introduced an encoder to encode the watermark and input the result into an embedder with the cover image to reach the watermarked image, with the encoder and embedder being two CNNs trained on the ImageNet and CIFAR157 datasets. Zhang et al.165 exploited an embedding network and an extractor network to embed and gain the watermark, respectively, and a surrogate network to boost the watermark, revealing ability of an extractor network; these were trained on the PASCAL VOC82 and Chestx-ray8166 datasets. However, the method was not robust enough to some preprocessing techniques such as random cropping and resizing.

Fig. 17

The watermarking framework of Ref. 160.

OE_60_12_120901_f017.png

Table 14 compares the bit error rate (BER) for the method recovering watermark information against normal digital image manipulations (including 20% cropping, 5% pepper and salt noise, and lossy JPEG compression with factor 10), capacity, and the PSNR between the container image and the watermarked image gained by Refs. 161162.163.164. The higher the BER value is, the more robust the watermark recovery ability of the methods is. The higher the PSNR values between the container image and the watermarked image is, the more imperceptible the method is.

Table 14

Comparison of different deep learning techniques in PSNR between watermarked images and container images, capacity, and BER for the methods recovering watermark information against normal digital image processing operations.

MethodRobustness to multiattacksPSNR (dB)Capacity (bpp)
JPEG compression with factor 1020% cropping5% salt and pepper noise
Ref. 161233.153.7×104
Ref. 16211.344.140.0156
Ref. 1636.617.9838.010.0052
Ref. 1648.1600.9739.720.0208

As can be seen in Table 14, different methods have their distinctive advantages. For example, Ref. 161 has better resistance in JPEG compression with factor 10, whereas Ref. 162 has a higher PSNR index between watermarked images and original images, and Ref. 164 has a better resistance in 5% salt and pepper noise and 20% cropping and larger capacity. As a result, a deep learning method in image watermarking that has a high resistance to multiattacks and a good imperceptibility of the watermarked image while having a large capacity needs to be further developed in future research.

5.3.

Image Watermark Extraction and Detection

Li et al.167 extracted a watermark image from a single-frame watermarked hologram by a conditional GAN trained on the fashion-MNIST168 and MNIST handwritten datasets. Huynh-The et al.169 exploited a blind image watermarking framework based on a deep convolutional encoder–decoder network watermark extraction model trained with various attacked watermarked images on the BOSSbase v1.01 database. Li et al.170 embedded a processed watermark image into the block discrete cosine transform component and used a cooperative neural network to recognize the suspected watermark signal with a single hologram, which improved the transmission efficiency. Hayes et al.171 learned a transformation-resilient watermark detector trained on the CIFAR-10 and ImageNet datasets to detect watermarks that could be employed in various carriers such as in the image, audio, and video domains. Chen et al.172 performed a simulated process to generate a large number of distorted watermarks and then collected them to form a training dataset to train a CNN model that could accurately identify the watermark copyright.

5.4.

Image Watermarking Attack

Wang et al.173 trained a watermark faker based on U-Net trained on the Caltech256174 dataset with the input being an original image and the output being a fake watermarked image after preprocessing; a set of paired images of original and watermarked images was obtained by the targeted image watermarking algorithms. However, this method did not perform well at generating the watermark in the frequency domain. Hatoum et al.175 employed a fully CNN trained on th BOSSbase dataset to denoise watermarked images and destroy the watermarks while preserving a satisfied quality of the denoised images. Sharma and Chandrasekaran176 implemented an enhanced hybrid watermarking scheme using DWT and singular value decomposition methods and proposed an adversarial attack based on a CNN-based autoencoder scheme trained on the CIFAR-10 database that could produce a perceptually close image.

5.5.

Image Watermark Removal

As can be seen in Fig. 18, for the visible watermark removal task, deep learning models often consist of two parts:177 watermark detection and removal. First, the deep learning model detects the watermark object in the watermarked images and then removes the watermark object from the watermarked image to obtain the watermark-less image. Gandelsman et al.178 proposed a coupled “Deep-Image-Prior” network to remove the image watermark that needed no training examples other than the input image/video. Hertz et al.179 trained on the Microsoft COCO val2014 dataset79 and learned to separate the visual motif from the image by estimating the visual motif matte and reconstructing the latent image for blind removal of both opaque and semitransparent motifs. Li et al.180 suggested a watermark processing framework using the conditional GAN trained on a large-scale visible watermark dataset177 and the PASCAL VOC2012 dataset for visible watermark removal in a real-world application. The generated watermark-less image had photorealistic quality but not good performance in standard quantitative evaluation metrics such as PSNR. Jiang et al.181 presented a watermark removal structure consisting of a watermark extraction network that removed the watermark in the watermarked image and an image inpainting network that inpainted the image for a watermark-less image. The two networks were trained on the PASCAL VOC2012 and places2182 datasets. Cun and Pun183 introduced a multitask feature extractor and a watermarked region smoother incorporated with multiple perceptual losses trained on the VAL2014 subset of the MSCOCO79 dataset and a dataset of logos to simulate the procedure of image watermark detection, removal, and refinement. However, when the detection failed or the textures in the watermark and background were similar, the network could not remove the watermark perfectly.

Fig. 18

The general process of deep learning methods for watermark removal.

OE_60_12_120901_f018.png

To remove the image watermark generated by the deep learning model, many scholars have proposed their methods from different perspectives. Shafieinejad et al.184 focused on backdoor-based watermarking, removingd the watermark fully by just relying on public data and proposing an attack that detected whether a model contained a watermark trained on th MNIST and CIFAR-10 datasets. Chen et al.185 exploited a unified watermark removal framework based on fine-tuning and incorporated it with an adaption of the elastic weight consolidation algorithm and unlabeled data augmentation. William et al.186 described a neural network “laundering” algorithm to remove black-box backdoor watermarks from neural networks trained on the MNIST and CIFAR-10 datasets.

6.

Future Scope

Bringing deep learning methods into the field of image security have solved many problems that cannot be solved by traditional methods. Deep learning image methods need a lot of pretraining time and depend too much on the training datasets, which are its characteristics. Furthermore, being good at using these characteristics or not decides the performance of deep learning models. In the future, the development directions could be summarized into five points.

  • 1. Deep learning models should be designed to take more consideration of the property of image security tasks and balance the model performance of all aspects. For example, to enhance the robustness or antiattack ability of deep learning methods, possible restrictions can be considered and set in advance in the design of the model architecture and training methods. Because the stego-signal representing only a small part in stego-image is not strong, finding a suitable method to enhance the stego-signal-to-noise ratio properly for better performance in image steganalysis is important. Meanwhile, the difference between the tampered image and the original image is very small, and understanding how to better use deep learning according to this property for forgery detection needs to be continually explored. Because the low-avalanche effect of neural network in end-to-end image encryption, considering the diffusion part is essential for the security.

  • 2. The internal working principles of deep learning should be better understood, and the techniques of deep learning models should be develeop for better use. The input and output of deep learning model are usually fixed, and designing a more flexible input and output size contributes to deep learning methods being more widely used in practice. A stronger image features extraction ability will make the action of deep learning more accurate, which is an area that many authors have been studying. Promoting the interpretability of deep learning helps people better design models according to the characteristics of deep learning. Improving the generalization ability of deep learning model will make the deep learning image security methods adapt to images in more scenarios. The large amount of computation of neural networks has been criticized. It is imperative to design a lightweight neural network using knowledge distillation, neural network pruning, and other technologies to reduce the computational complexity, which is conducive to applications especially in industry.

  • 3. There is a shortage of theories for deep learning image security, so establishing and continuously improving the theoretical system of deep learning image security are urgent. Some deep learning methods in image security need to be explained from the view of mathematics, while the targeted tests should be expanded from other angles. A better theoretical basis will guide faster and better development of the field of deep learning image security.

  • 4. Deep learning is a dataset driven technology, but the datasets established for some special tasks are not perfect. It is necessary to establish larger and richer datasets for special tasks. For example, the dataset used in the field of image forgery detection needs to include images tampered by various tampering methods. There is an urgent need to establish a more comprehensive dataset to promote the rapid development and application of diversified tests of deep learning on special tasks.

  • 5. Deep learning should be taken into other areas of image security to solve more traditional image security problems. For example, using neural network to exchange keys in image cryptography and neural network image homomorphic encryption are interesting research directions.

Above all, there are still many challenges and development opportunities in the field of image security for deep learning. It is significant to develop deep learning in image security.

7.

Conclusion

This paper describes deep learning with respect to image steganography to generate the cover image, the stego-image, embedding change probabilities, coverless steganography, and steganalysis. As a result, we know that the image steganography method based on deep learning has reached a good performance in embedding capacity and stego-image imperceptibility quality. However, the robustness of deep learning-based image steganography algorithms needs further detailed testing, analysis, and improvements; the embedded secret data extraction should be more accurate; and the input and output size should be more flexible in future studies. Furthermore, this paper combines and compares deep learning techniques used in image cryptography as concerns in image compression, image resolution improvement, image object detection and classification, key generation, end-to-end image encryption, and image cryptoanalysis. We find that end-to-end key generation and image encryption based on the deep learning model have advantages in large key space and automatic generation, with a reduced reliance on complex cryptography design knowledge. Furthermore, the improvement in the randomness of generated ciphertext image and keys, quality of decrypted image, generalization ability and efficiency of the encryption and decryption model, and resistance of facing chosen ciphertext or plaintext attack are still significant research directions for the future. In addition, this paper relates deep learning methods in image authentication from image forgery detection, watermarked image generation, image watermark extraction and detection, image watermarking attack, and image watermark removal and predicts the development of an image watermarking method based on deep learning that has a high resistance to multiattacks, good imperceptibility of the watermarked image, and a large capacity, which are future research directions for this topic. Finally, we summarize three future development directions through the whole analysis that have enlightening significance for relevant researchers.

Acknowledgements

This project was supported by the Natural Science Foundation of Xizang Autonomous Region of China (Grant No. XZ202001ZR0048G). The authors declare no conflicts of interest.

References

1. 

W. Liu et al., “SSD: single shot multibox detector,” Lect. Notes Comput. Sci., 9905 21 –37 (2016). https://doi.org/10.1007/978-3-319-46448-0_2 LNCSD9 0302-9743 Google Scholar

2. 

J. Redmon et al., “You only look once: unified, real-time object detection,” in IEEE Conf. Comput. Vision and Pattern Recognit. (CVPR), 779 –788 (2016). https://doi.org/10.1109/CVPR.2016.91 Google Scholar

3. 

J. Redmon and A. Farhadi, “YOLO9000: better, faster, stronger,” in IEEE Conf. Comput. Vision and Pattern Recognit. (CVPR), 6517 –6525 (2017). https://doi.org/10.1109/CVPR.2017.690 Google Scholar

4. 

J. Redmon and A. Farhadi, “YOLOv3: an incremental improvement,” (2018). https://arxiv.org/abs/1804.02767 Google Scholar

5. 

A. Bochkovskiy, C. Y. Wang and H. Y. Liao, “YOLOv4: optimal speed and accuracy of object detection,” (2020). https://arxiv.org/abs/2004.10934 Google Scholar

6. 

K. M. Hosny, M. A. Kassem and M. M. Fouad, “Classification of skin lesions into seven classes using transfer learning with AlexNet,” J. Digital Imaging, 33 1325 –1334 (2020). https://doi.org/10.1007/s10278-020-00371-9 Google Scholar

7. 

A. Krizhevsky, I. Sutskever and G. E. Hinton, “ImageNet classification with deep convolutional neural networks,” in Proc. NIPS, 1097 –1105 (2012). Google Scholar

8. 

N. Dey et al., “Customized VGG19 architecture for pneumonia detection in chest X-rays,” Pattern Recognit. Lett., 143 67 –74 (2021). https://doi.org/10.1016/j.patrec.2020.12.010 PRLEDG 0167-8655 Google Scholar

9. 

K. He et al., “Deep residual learning for image recognition,” in IEEE Conf. Comput. Vision and Pattern Recognit. (CVPR), 770 –778 (2016). https://doi.org/10.1109/CVPR.2016.90 Google Scholar

10. 

R. Ke et al., “Multi-task deep learning for image segmentation using recursive approximation tasks,” IEEE Trans. Image Process., 30 3555 –3567 (2021). https://doi.org/10.1109/TIP.2021.3062726 IIPRE4 1057-7149 Google Scholar

11. 

J. Zhang et al., “LCU-Net: a novel low-cost U-Net for environmental microorganism image segmentation,” Pattern Recognit., 115 107885 (2021). https://doi.org/10.1016/j.patcog.2021.107885 Google Scholar

12. 

B. Olimov et al., “FU-Net: fast biomedical image segmentation model based on bottleneck convolution layers,” Multimedia Syst., 27 (4), 637 –650 (2021). https://doi.org/10.1007/s00530-020-00726-w MUSYEW 1432-1882 Google Scholar

13. 

Y. Lei et al., “Echocardiographic image multi‐structure segmentation using Cardiac‐SegNet,” Med. Phys., 48 (5), 2426 –2473 (2021). Google Scholar

14. 

L. A. Gatys, A. S. Ecker and M. Bethge, “Image style transfer using convolutional neural networks,” in IEEE Conf. Comput. Vision and Pattern Recognit. (CVPR), 2414 –2423 (2016). https://doi.org/10.1109/CVPR.2016.265 Google Scholar

15. 

Z. Ma et al., “Image style transfer with collection representation space and semantic-guided reconstruction,” Neural Networks, 129 123 –137 (2020). https://doi.org/10.1016/j.neunet.2020.05.028 NNETEB 0893-6080 Google Scholar

16. 

C. T. Lin et al., “GAN-based day-to-night image style transfer for nighttime vehicle detection,” IEEE Trans. Intell. Transp. Syst., 22 (2), 951 –963 (2021). https://doi.org/10.1109/TITS.2019.2961679 Google Scholar

17. 

Z. Wang et al., “GLStyleNet: exquisite style transfer combining global and local pyramid features,” IET Comput. Vision, 14 575 –586 (2020). https://doi.org/10.1049/iet-cvi.2019.0844 Google Scholar

18. 

C. Tian et al., “Designing and training of a dual CNN for image denoising,” Knowl.-Based Syst., 226 106949 (2021). https://doi.org/10.1016/j.knosys.2021.106949 KNSYET 0950-7051 Google Scholar

19. 

H. Xia et al., “Combination of multi‐scale and residual learning in deep CNN for image denoising,” IET Image Process., 14 2013 –2019 (2020). https://doi.org/10.1049/iet-ipr.2019.1386 Google Scholar

20. 

P. Jia et al., “Point spread function modelling for wide-field small-aperture telescopes with a denoising autoencoder,” Mon. Not. R. Astron. Soc., 493 (1), 651 –660 (2020). https://doi.org/10.1093/mnras/staa319 MNRAA4 0035-8711 Google Scholar

21. 

D. Liu et al., “View synthesis-based light field image compression using a generative adversarial network,” Inf. Sci., 545 118 –131 (2021). https://doi.org/10.1016/j.ins.2020.07.073 Google Scholar

22. 

H. Liu et al., “Deep learning-based picture-wise just noticeable distortion prediction model for image compression,” IEEE Trans. Image Process., 29 641 –656 (2020). https://doi.org/10.1109/TIP.2019.2933743 IIPRE4 1057-7149 Google Scholar

23. 

I. Schiopu and A. Munteanu, “Residual‐error prediction based on deep learning for lossless image compression,” Electron. Lett., 54 1032 –1034 (2018). https://doi.org/10.1049/el.2018.0889 ELLEAK 0013-5194 Google Scholar

24. 

J. Fridrich and J. Kodovsky, “Rich models for steganalysis of digital images,” IEEE Trans. Inf. Forensics Secur., 7 (3), 868 –882 (2012). https://doi.org/10.1109/TIFS.2012.2190402 Google Scholar

25. 

T. Denemark et al., “Selection-channel-aware rich model for steganalysis of digital images,” in Proc. IEEE Int. Workshop Inf. Forensics Secur. (WIFS), 48 –53 (2014). https://doi.org/10.1109/WIFS.2014.7084302 Google Scholar

26. 

G. Xu, H. Wu and Y. Shi, “Structural design of convolutional neural networks for steganalysis,” IEEE Signal Process. Lett., 23 (5), 708 –712 (2016). https://doi.org/10.1109/LSP.2016.2548421 IESPEJ 1070-9908 Google Scholar

27. 

M. Jarno, “LSB matching revisited,” IEEE Signal Process. Lett., 13 (5), 285 –287 (2006). https://doi.org/10.1109/LSP.2006.870357 IESPEJ 1070-9908 Google Scholar

28. 

T. Pevný, T. Filler and P. Bas, “Using high-dimensional image models to perform highly undetectable steganography,” Lect. Notes Comput. Sci., 6387 161 –177 (2010). https://doi.org/10.1007/978-3-642-16435-4_13 LNCSD9 0302-9743 Google Scholar

29. 

W. Luo, F. Huang and J. Huang, “Edge adaptive image steganography based on LSB matching revisited,” IEEE Trans. Inf. Forensics Secur., 5 (2), 201 –214 (2010). https://doi.org/10.1109/TIFS.2010.2041812 Google Scholar

30. 

V. Holub, J. Fridrich and T. Denemark, “Universal distortion function for steganography in an arbitrary domain,” EURASIP J. Inf. Secur., 2014 (1), 1 –13 (2014). https://doi.org/10.1186/1687-417X-2014-1 Google Scholar

31. 

B. Li et al., “A new cost function for spatial image steganography,” in Proc. IEEE ICIP, 4206 –4210 (2014). https://doi.org/10.1109/ICIP.2014.7025854 Google Scholar

32. 

R. Zhang, S. Dong and J. Liu, “Invisible steganography via generative adversarial networks,” Multimedia Tools Appl., 78 (7), 8559 –8575 (2019). https://doi.org/10.1007/s11042-018-6951-z Google Scholar

33. 

J. Yang et al., “An embedding cost learning framework using GAN,” IEEE Trans. Inf. Forensics Secur., 15 839 –851 (2019). https://doi.org/10.1109/TIFS.2019.2922229 Google Scholar

34. 

P. Bas, T. Filler and T. Pevný, “Break our steganographic system: the ins and outs of organizing BOSS,” in Proc. 13th Int. Conf. Inf. Hiding, 59 –70 (2011). Google Scholar

35. 

S. Jiao et al., “Known-plaintext attack and ciphertext-only attack for encrypted single-pixel imaging,” IEEE Access, 7 119557 –119565 (2019). https://doi.org/10.1109/ACCESS.2019.2936119 Google Scholar

36. 

S. K. Rajput and N. K. Nishchal, “Known-plaintext attack on encryption domain independent optical asymmetric cryptosystem,” Opt. Commun., 309 (Complete), 231 –235 (2013). https://doi.org/10.1016/j.optcom.2013.06.036 OPCOB8 0030-4018 Google Scholar

37. 

I. E. Hanouti, H. E. Fadili and K. Zenkouar, “Cryptanalysis of an embedded systems’ image encryption,” Multimedia Tools Appl., 80 (9), 13801 –13820 (2021). https://doi.org/10.1007/s11042-020-10289-7 Google Scholar

38. 

E. J. Yoon and K. Y. Yoo, “Cryptanalysis of a modulo image encryption scheme with fractal keys,” Opt. Lasers Eng., 48 (7–8), 821 –826 (2010). https://doi.org/10.1016/j.optlaseng.2010.02.004 Google Scholar

39. 

J. J. Chen et al., “Memristor-based hyper-chaotic circuit for image encryption,” Chin. Phys. B, 29 (11), 299 –310 (2020). 1674-1056 Google Scholar

40. 

Y. Bentoutou et al., “An improved image encryption algorithm for satellite applications,” Adv. Space Res., 66 176 –192 (2020). https://doi.org/10.1016/j.asr.2019.09.027 ASRSDW 0273-1177 Google Scholar

41. 

M. Naim, A. A. Pacha and C. Serief, “A novel satellite image encryption algorithm based on hyperchaotic systems and Josephus problem,” Adv. Space Res., 67 (7), 2077 –2103 (2021). https://doi.org/10.1016/j.asr.2021.01.018 ASRSDW 0273-1177 Google Scholar

42. 

M. D. Zhao et al., “Image encryption based on fractal-structured phase mask in fractional Fourier transform domain,” J. Opt., 20 (4), 045703 (2018). https://doi.org/10.1088/2040-8986/aab247 Google Scholar

43. 

F. A. Yatish and N. K. Nishchal, “Optical image encryption using triplet of functions,” Opt. Eng., 57 (3), 033103 (2018). https://doi.org/10.1117/1.OE.57.3.033103 Google Scholar

44. 

Y. Shi et al., “Multiple-image double-encryption via 2D rotations of a random phase mask with spatially incoherent illumination,” Opt. Express, 27 (18), 26050 –26059 (2019). https://doi.org/10.1364/OE.27.026050 OPEXFF 1094-4087 Google Scholar

45. 

P. Ping, F. Xu and Z. J. Wang, “Color image encryption based on two-dimensional cellular automata,” Int. J. Mod. Phys. C, 24 (10), 1350071 (2013). https://doi.org/10.1142/S012918311350071X Google Scholar

46. 

A. Souyah and K. M. Faraoun, “An image encryption scheme combining chaos-memory cellular automata and weighted histogram,” Nonlinear Dyn., 86 639 –653 (2016). https://doi.org/10.1007/s11071-016-2912-0 NODYES 0924-090X Google Scholar

47. 

P. Naskar et al., “A robust image encryption scheme using chaotic tent map and cellular automata,” Nonlinear Dyn., 100 2877 –2898 (2020). https://doi.org/10.1007/s11071-020-05625-3 NODYES 0924-090X Google Scholar

48. 

H. S. Li et al., “Quantum image encryption based on phase-shift transform and quantum Haar wavelet packet transform,” Mod. Phys. Lett. A, 34 (26), 1950214 (2019). https://doi.org/10.1142/S0217732319502146 MPLAEQ 0217-7323 Google Scholar

49. 

C. Hou, X. Liu and S. Feng, “Quantum image scrambling algorithm based on discrete Baker map,” Mod. Phys. Lett. A, 35 (17), 2050145 (2020). https://doi.org/10.1142/S021773232050145X MPLAEQ 0217-7323 Google Scholar

50. 

G. Ye, K. Jiao and X. Huang, “Quantum logistic image encryption algorithm based on SHA-3 and RSA,” Nonlinear Dyn., 104 (3), 2807 –2827 (2021). https://doi.org/10.1007/s11071-021-06422-2 NODYES 0924-090X Google Scholar

51. 

M. Guan, X. Yang and W. Hu, “Chaotic image encryption algorithm using frequency‐domain DNA encoding,” IET Image Process., 13 1535 –1539 (2019). https://doi.org/10.1049/iet-ipr.2019.0051 Google Scholar

52. 

X. Chai et al., “A novel image encryption algorithm based on the chaotic system and DNA computing,” Int. J. Mod. Phys. C, 28 (5), 1750069 (2017). https://doi.org/10.1142/S0129183117500693 Google Scholar

53. 

D. Ravichandran et al., “An efficient medical image encryption using hybrid DNA computing and chaos in transform domain,” Med. Biol. Eng. Comput., 59 589 –605 (2021). https://doi.org/10.1007/s11517-021-02328-8 MBECDY 0140-0118 Google Scholar

54. 

Z. Bao and R. Xue, “Research on the avalanche effect of image encryption based on the Cycle-GAN,” Appl. Opt., 60 (18), 5320 –5334 (2021). https://doi.org/10.1364/AO.428203 APOPAI 0003-6935 Google Scholar

55. 

F. H. Hsu et al., “Visible watermarking with reversibility of multimedia images for ownership declarations,” J. Supercomput., 70 (1), 247 –268 (2014). https://doi.org/10.1007/s11227-014-1258-y JOSUED 0920-8542 Google Scholar

56. 

J. C. Patra, J. E. Phua and C. Bornand, “A novel DCT domain CRT-based watermarking scheme for image authentication surviving jpeg compression,” Digital Signal Process., 20 (6), 1597 –1611 (2010). https://doi.org/10.1016/j.dsp.2010.03.010 DSPREJ 1051-2004 Google Scholar

57. 

T. S. Nguyen, “Fragile watermarking for image authentication based on DWT-SVD-DCT techniques,” Multimedia Tools Appl., 80 25107 –25119 (2021). Google Scholar

58. 

W. Zhang and F. Y. Shih, “Semi-fragile spatial watermarking based on local binary pattern operators,” Opt. Commun., 284 (16–17), 3904 –3912 (2011). https://doi.org/10.1016/j.optcom.2011.04.004 OPCOB8 0030-4018 Google Scholar

59. 

Y. Pathak, Y. K. Jain and S. Dehariya, “A secure transmission of medical images by single label SWT and SVD based non-blind watermarking technique,” Infocomp J. Comput. Sci., 14 (1), 50 –59 (2015). https://doi.org/10.18760/IC.14120155 Google Scholar

60. 

B. Z. Li and Y. Guo, “Blind image watermarking method based on linear canonical wavelet transform and QR decomposition,” IET Image Process., 10 (10), 773 –786 (2016). https://doi.org/10.1049/iet-ipr.2015.0818 Google Scholar

61. 

A. Sleit et al., “An enhanced semi-blind. DWT-SVD-based watermarking technique for digital images,” Imaging Sci. J., 60 (1), 29 –38 (2012). https://doi.org/10.1179/1743131X11Y.0000000010 Google Scholar

62. 

H. Agarwal, P. Atrey and B. Raman, “Image watermarking in real oriented wavelet transform domain,” Multimedia Tools Appl., 74 (23), 10883 –10921 (2015). https://doi.org/10.1007/s11042-014-2212-y Google Scholar

63. 

S. P. Patil and K. N. Jariwala, “Improving the efficiency of image and video forgery detection using hybrid convolutional neural networks,” Int. J. Uncertain. Fuzz. Knowl.-Based Syst., 29 (Supp1), 101 –117 (2021). https://doi.org/10.1142/S0218488521400067 Google Scholar

64. 

A. Radford, L. Metz and S. Chintala, “Unsupervised representation learning with deep convolutional generative adversarial networks,” (2015). https://arxiv.org/abs/1511.06434 Google Scholar

65. 

D. Volkhonskiy, B. Borisenko and E. Burnaev, ““Generative adversarial networks for image steganography,” in ICLR 2017 Open Rev., (2017). Google Scholar

66. 

Z. Liu et al., “Deep learning face attributes in the wild,” in IEEE Int. Conf. Comput. Vision (ICCV), 3730 –3738 (2015). https://doi.org/10.1109/ICCV.2015.425 Google Scholar

67. 

H. Shi et al., “SSGAN: secure steganography based on generative adversarial networks,” in Pac. Rim Conf. Multimedia, 534 –544 (2017). Google Scholar

68. 

M. Arjovsky, S. Chintala and L. Bottou, “Wasserstein GAN,” (2017). https://arxiv.org/abs/1701.07875 Google Scholar

69. 

Y. Qian et al., “Deep learning for steganalysis via convolutional neural networks,” Proc. SPIE, 9409 94090J (2015). https://doi.org/10.1117/12.2083479 PSISDG 0277-786X Google Scholar

70. 

J. Hayes and G. Danezis, “Generating steganographic images via adversarial training,” in NIPS 2017: Neural Inf. Process. Syst., 4 –9 (2017). Google Scholar

71. 

S. Baluja, “Hiding images within images,” IEEE Trans. Pattern Anal. Mach. Intell., 42 (7), 1685 –1697 (2020). https://doi.org/10.1109/TPAMI.2019.2901877 ITPIDJ 0162-8828 Google Scholar

72. 

S. Chattopadhyay, P. P. Acharjya and C. Koner, “A deep learning approach to implement slab based image steganography algorithm of RGB images,” in 3rd Int. Conf. Invent. Comput. Technol. (ICICT), (2018). Google Scholar

73. 

J. Zhu et al., “Hidden: hiding data with deep networks,” in Proc. Eur. Conf. Comput. Vision (ECCV), 657 –672 (2018). Google Scholar

74. 

D. Hu et al., “A novel image steganography method via deep convolutional generative adversarial networks,” IEEE Access, 6 38303 –38314 (2018). https://doi.org/10.1109/ACCESS.2018.2852771 Google Scholar

75. 

L. Bossard, M. Guillaumin and L. V. Gool, “Food-101—mining discriminative components with random forests,” Lect. Notes Comput. Sci., 8694 446 –461 (2014). https://doi.org/10.1007/978-3-319-10599-4_29 LNCSD9 0302-9743 Google Scholar

76. 

K. A. Zhang et al., “SteganoGAN: high capacity image steganography with GANs,” (2019). https://arxiv.org/abs/1901.03892v2 Google Scholar

77. 

Z. Wang et al., “HidingGAN: high capacity information hiding with generative adversarial network,” Comput. Graphics Forum, 38 (7), 393 –401 (2019). https://doi.org/10.1111/cgf.13846 CGFODY 0167-7055 Google Scholar

78. 

X. Duan et al., “High-capacity information hiding based on residual network,” IETE Tech. Rev., 38 (1), 172 –183 (2021). https://doi.org/10.1080/02564602.2020.1808097 ITREEI Google Scholar

79. 

T. Y. Lin et al., “Microsoft COCO: common objects in context,” in Proc. Eur. Conf. Comput. Vision, (2014). Google Scholar

80. 

J. Deng et al., “ImageNet: a large-scale hierarchical image database,” in IEEE Conf. Comput. Vision and Pattern Recognit., 248 –255 (2009). https://doi.org/10.1109/CVPR.2009.5206848 Google Scholar

81. 

G. B. Huang et al., Labeled Faces in the Wild: A Database for Studying Face Recognition in Unconstrained Environments, University of Massachusetts, Amherst, Massachusetts (2007). Google Scholar

82. 

M. Everingham et al., “The Pascal visual object classes (VOC) challenge,” Int. J. Comput. Vision, 88 (2), 303 –338 (2010). https://doi.org/10.1007/s11263-009-0275-4 IJCVEQ 0920-5691 Google Scholar

83. 

Z. Wang et al., “Image quality assessment: from error measurement to structural similarity,” IEEE Trans. Image Process., 13 600 –612 (2004). https://doi.org/10.1109/TIP.2003.819861 IIPRE4 1057-7149 Google Scholar

84. 

W. Tang et al., “Automatic steganographic distortion learning using a generative adversarial network,” IEEE Signal Process. Lett., 24 (10), 1547 –1551 (2017). https://doi.org/10.1109/LSP.2017.2745572 IESPEJ 1070-9908 Google Scholar

85. 

O. Ronneberger, P. Fischer and T. Brox, “U-Net: convolutional networks for biomedical image segmentation,” Lect. Notes Comput. Sci., 9351 234 –241 (2015). https://doi.org/10.1007/978-3-319-24574-4_28 LNCSD9 0302-9743 Google Scholar

86. 

W. Tang et al., “An automatic cost learning framework for image steganography using deep reinforcement learning,” IEEE Trans. Inf. Forensics Secur., 16 952 –967 (2021). https://doi.org/10.1109/TIFS.2020.3025438 Google Scholar

87. 

X. Duan et al., “Coverless steganography for digital images based on a generative model,” Comput. Mater. Contin., 55 (3), 483 –493 (2018). https://doi.org/10.3970/cmc.2018.01798 Google Scholar

88. 

Y. Luo et al., “Coverless image steganography based on multi-object recognition,” IEEE Trans. Circuits Syst. Video Technol., 31 2779 –2791 (2020). https://doi.org/10.1109/TCSVT.2020.3033945 Google Scholar

89. 

Q. Liu et al., “Coverless steganography based on image retrieval of DenseNet features and DWT sequence mapping,” Knowl.-Based Syst., 192 105375 (2019). https://doi.org/10.1016/j.knosys.2019.105375 Google Scholar

90. 

H. Gao, L. Zhuang and M. D. V. Laurens, “Densely connected convolutional networks,” in IEEE Conf. Comput. Vision and Pattern Recognit. (CVPR), 2261 –2269 (2017). Google Scholar

91. 

S. Zhang et al., “An image style transfer network using multilevel noise encoding and its application in coverless steganography,” Symmetry, 11 (9), 1152 (2019). https://doi.org/10.3390/sym11091152 SYMMAM 2073-8994 Google Scholar

92. 

Z. Zhou et al., “Faster-RCNN based robust coverless information hiding system in cloud environment,” IEEE Access, 7 179891 –179897 (2019). https://doi.org/10.1109/ACCESS.2019.2955990 Google Scholar

93. 

M. Everingham et al., “The PASCAL visual object classes challenge 2012 (VOC2012) results,” (2012). http://www.pascal-network.org/challenges/VOC/voc2012/workshop/index.html Google Scholar

94. 

X. Duan et al., “A coverless steganography method based on generative adversarial network,” J. Image Video Proc., 2020 18 (2020). https://doi.org/10.1186/s13640-020-00506-6 Google Scholar

95. 

J. Yang et al., “Steganalysis based on awareness of selection-channel and deep learning,” Lect. Notes Comput. Sci., 10431 263 –272 (2017). LNCSD9 0302-9743 Google Scholar

96. 

J. Zeng et al., “Large-scale jpeg steganalysis using hybrid deep-learning framework,” IEEE Trans. Inf. Forensics Secur., 13 (5), 1200 –1214 (2017). https://doi.org/10.1109/TIFS.2017.2779446 Google Scholar

97. 

Y. Jian, J. Ni and Y. Yang, “Deep learning hierarchical representations for image steganalysis,” IEEE Trans. Inf. Forensics Secur., 12 (11), 2545 –2557 (2017). https://doi.org/10.1109/TIFS.2017.2710946 Google Scholar

98. 

M. Yedroudj, F. Comby and M. Chaumont, “Yedroudj-Net: an efficient CNN for spatial steganalysis,” in IEEE Int. Conf. Acoust. Speech Signal Process. (ICASSP), 2092 –2096 (2018). https://doi.org/10.1109/ICASSP.2018.8461438 Google Scholar

99. 

R. Zhang et al., “Efficient feature learning and multi-size image steganalysis based on CNN,” (2018). https://arxiv.org/abs/1807.11428 Google Scholar

100. 

M. Boroumand, M. Chen and J. Fridrich, “Deep residual network for steganalysis of digital images,” IEEE Trans. Inf. Forensics Secur., 14 (5), 1181 –1193 (2019). https://doi.org/10.1109/TIFS.2018.2871749 Google Scholar

101. 

T. S. Reinel et al., “GBRAS-Net: a convolutional neural network architecture for spatial image steganalysis,” IEEE Access, 9 14340 –14350 (2021). https://doi.org/10.1109/ACCESS.2021.3052494 Google Scholar

102. 

F. Liu et al., “Image steganalysis via diverse filters and squeeze-and-excitation convolutional neural network,” Mathematics, 9 (2), 189 (2021). https://doi.org/10.3390/math9020189 Google Scholar

103. 

A. I. Iskanderani et al., “Artificial intelligence-based digital image steganalysis,” Secur. Commun. Networks, 2021 (11), 1 –9 (2021). https://doi.org/10.1155/2021/9923389 Google Scholar

104. 

A. Singhal and P. Bedi, “Multi-class blind steganalysis using deep residual networks,” Multimedia Tools Appl., 80 13931 –13956 (2021). https://doi.org/10.1007/s11042-020-10353-2 Google Scholar

105. 

V. Holub and J. Fridrich, “Designing steganographic distortion using directional filters,” in IEEE Int. Workshop Inf. Forensics Secur., WIFS’2012, 234 –239 (2012). https://doi.org/10.1109/WIFS.2012.6412655 Google Scholar

106. 

W. Chen, Y. Guo and S. W. Jing, “General image encryption algorithm based on deep learning compressed sensing and compound chaotic system,” Acta Phys. Sin., 69 (24), 240502 (2020). https://doi.org/10.7498/aps.69.20201019 WLHPAR 1000-3290 Google Scholar

107. 

Q. S. Lian et al., “A compressed sensing algorithm based on multi-scale residual reconstruction network,” Acta Autom. Sin., 45 (11), 2082 –2091 (2019). https://doi.org/10.11834/jig.200168 THHPAY 0254-4156 Google Scholar

108. 

F. Hu et al., “An image compression and encryption scheme based on deep learning,” (2016). https://arxiv.org/abs/1608.05001 Google Scholar

109. 

K. M. A. Suhail and S. Sankar, “Image compression and encryption combining autoencoder and chaotic logistic map,” Iran J. Sci. Technol. Trans. Sci., 44 1091 –1100 (2020). https://doi.org/10.1007/s40995-020-00905-4 Google Scholar

110. 

C. T. Selvi, J. Amudha and R. Sudhakar, “Medical image encryption and compression by adaptive sigma filterized synorr certificateless signcryptive Levenshtein entropy-coding-based deep neural learning,” Multimedia Syst., 27 1059 –1074 (2021). MUSYEW 1432-1882 Google Scholar

111. 

L. Zhang et al., “Optical image compression and encryption transmission-based on deep learning and ghost imaging,” Appl. Phys. B, 126 (1), 061802 –993 (2019). https://doi.org/10.1007/s00340-019-7362-1 Google Scholar

112. 

M. Grubinger et al., “The IAPR TC-12 benchmark: a new evaluation resource for visual information systems,” in Int. Workshop OntoImage, 13 –23 (2006). Google Scholar

113. 

J. Chen, X. W. Li and Q. H. Wang, “Deep learning for improving the robustness of image encryption,” IEEE Access, 7 181083 –181091 (2019). https://doi.org/10.1109/ACCESS.2019.2959031 Google Scholar

114. 

K. Ma et al., “Waterloo exploration database: new challenges for image quality assessment models,” IEEE Trans. Image Process., 26 (2), 1004 –1016 (2017). https://doi.org/10.1109/TIP.2016.2631888 IIPRE4 1057-7149 Google Scholar

115. 

X. Zhao et al., “Application of face image detection based on deep learning in privacy security of intelligent cloud platform,” Multimedia Tools Appl., 79 16707 –16718 (2020). https://doi.org/10.1007/s11042-019-08014-0 Google Scholar

116. 

B. A. Y. Alqaralleh et al., “Blockchain-assisted secure image transmission and diagnosis model on Internet of Medical Things Environment,” Pers Ubiquit Comput., (2021). Google Scholar

117. 

M. Asgari-Chenaghlu et al., “Cy: Chaotic yolo for user intended image encryption and sharing in social media,” Inf. Sci., 542 212 –227 (2021). https://doi.org/10.1016/j.ins.2020.07.007 Google Scholar

118. 

X. Li et al., “Research on iris image encryption based on deep learning,” J. Image Video Proc., 2018 (1), 1 –10 (2018). https://doi.org/10.1186/s13640-018-0358-7 Google Scholar

119. 

L. Debiasi and A. Uhl, “Techniques for a forensic analysis of the CASIA-IRIS V4 database,” in 3rd Int. Workshop Biometrics and Forensics (IWBF 2015), (2015). Google Scholar

120. 

Y. Ding et al., “DeepKeyGen: a deep learning-based stream cipher generator for medical image encryption and decryption,” IEEE Trans. Neural Networks Learn. Syst., (2021). Google Scholar

121. 

S. Jaeger et al., “Two public chest X-ray datasets for computer-aided screening of pulmonary diseases,” Quant. Imaging Med. Surg., 4 475 –477 (2014). https://doi.org/10.3978/j.issn.2223-4292.2014.11.20 Google Scholar

122. 

J. Jin and K. Kim, “3D CUBE algorithm for the key generation method: applying deep neural network learning-based,” IEEE Access, 8 33689 –33702 (2020). https://doi.org/10.1109/ACCESS.2020.2973695 Google Scholar

123. 

S. R. Maniyath and V. Thanikaiselvan, “An efficient image encryption using deep neural network and chaotic map,” Microprocess. Microsyst., 77 103134 (2020). https://doi.org/10.1016/j.micpro.2020.103134 MIMID5 0141-9331 Google Scholar

124. 

U. Erkan et al., “An image encryption scheme based on chaotic logarithmic map and key generation using deep CNN,” (2020). https://arxiv.org/abs/2012.14156v1 Google Scholar

125. 

A. Fratalocchi et al., “NIST-certified secure key generation via deep learning of physical unclonable functions in silica aerogels,” Nanophotonics, 10 (1), 457 –464 (2021). https://doi.org/10.1515/nanoph-2020-0368 Google Scholar

126. 

J. Y. Zhu et al., “Unpaired image-to-image translation using cycle-consistent adversarial networks,” in IEEE Int. Conf. Comput. Vision, 2223 –2232 (2017). https://doi.org/10.1109/ICCV.2017.244 Google Scholar

127. 

J. Li, J. Zhou and X. Di, “A learning optical image encryption scheme based on CycleGAN,” J. Jilin Univ. (Eng. Technol. Ed.), 51 (3), 1060 –1066 (2021). https://doi.org/10.13229/j.cnki.jdxbgxb20200521 Google Scholar

128. 

Y. Ding et al., “DeepEDN: a deep-learning-based image encryption and decryption network for internet of medical things,” IEEE Internet Things J., 8 (3), 1504 –1518 (2021). https://doi.org/10.1109/JIOT.2020.3012452 Google Scholar

129. 

Z. Bao, R. Xue and Y. D. Jin, “Image scrambling adversarial autoencoder based on the asymmetric encryption,” Multimedia Tools Appl., 80 (18), 28265 –28301 (2021). https://doi.org/10.1007/s11042-021-11043-3 Google Scholar

130. 

J. Z. Wang, J. Li and G. Wiederhold, “SIMPLIcity: semantics-sensitive integrated matching for picture librarie,” IEEE Trans. Pattern Anal. Mach. Intell., 23 (9), 947 –963 (2001). https://doi.org/10.1109/34.955109 ITPIDJ 0162-8828 Google Scholar

131. 

W. Qin and X. Peng, “Asymmetric cryptosystem based on phase-truncated Fourier transforms,” Opt. Lett., 35 (2), 118 –120 (2010). https://doi.org/10.1364/OL.35.000118 OPLEDP 0146-9592 Google Scholar

132. 

Z. Xu et al., “Attacking the asymmetric cryptosystem based on phase truncated Fourier transforms by deep learning,” Acta Phys. Sin., 70 (14), 226 –232 (2021). https://doi.org/10.7498/aps.70.20202075 WLHPAR 1000-3290 Google Scholar

133. 

D. Ciresan, U. Meier and J. Schmidhuber, “Multi-column deep neural networks for image classification,” in IEEE Int. Conf. Comput. Vision and Pattern Recognit. (CVPR), 3642 –3649 (2012). https://doi.org/10.1109/CVPR.2012.6248110 Google Scholar

134. 

H. Hai et al., “Cryptanalysis of random-phase-encoding-based optical cryptosystem via deep learning,” Opt. Express, 27 (15), 21204 –21213 (2019). https://doi.org/10.1364/OE.27.021204 OPEXFF 1094-4087 Google Scholar

135. 

P. Refregier and B. Javidi, “Optical image encryption based on input plane and Fourier plane random encoding,” Opt. Lett., 20 (7), 767 –769 (1995). https://doi.org/10.1364/OL.20.000767 OPLEDP 0146-9592 Google Scholar

136. 

C. Wu et al., “Cryptoanalysis of the modified diffractive-imaging-based image encryption by deep learning attack,” J. Mod. Opt., 67 (17), 1398 –1409 (2020). https://doi.org/10.1080/09500340.2020.1862329 JMOPEW 0950-0340 Google Scholar

137. 

Q. A. Gong et al., “Modified diffractive-imaging-based image encryption,” Opt. Lasers Eng., 121 66 –73 (2019). https://doi.org/10.1016/j.optlaseng.2019.03.013 Google Scholar

138. 

L. Chen et al., “Plaintext attack on joint transform correlation encryption system by convolutional neural network,” Opt. Express, 28 (19), 28154 –28163 (2020). https://doi.org/10.1364/OE.402958 OPEXFF 1094-4087 Google Scholar

139. 

C. He et al., “A deep learning based attack for the chaos-based image encryption,” (2019). https://arxiv.org/abs/1907.12245 Google Scholar

140. 

Z. H. Guan, F. Huang and W. Guan, “Chaos-based image encryption algorithm,” Phys. Lett. A, 346 (1–3), 153 –157 (2005). https://doi.org/10.1016/j.physleta.2005.08.006 PYLAAG 0375-9601 Google Scholar

141. 

L. Bondi et al., “Tampering detection and localization through clustering of camera-based CNN features,” in IEEE Conf. Comput. Vision and Pattern Recognit. Workshops., 1855 –1864 (2017). https://doi.org/10.1109/CVPRW.2017.232 Google Scholar

142. 

T. Gloe and R. B¨ohme, “The Dresden image database for benchmarking digital image forensics,” J. Digital Forensic Pract., 3 150 –159 (2010). https://doi.org/10.1080/15567281.2010.531500 Google Scholar

143. 

M. A. Elaskily et al., “A novel deep learning framework for copy-move forgery detection in images,” Multimedia Tools Appl., 79 19167 –19192 (2020). https://doi.org/10.1007/s11042-020-08751-7 Google Scholar

144. 

I. Amerini et al., “A SIFT-based forensic method for copy–move attack detection and transformation recovery,” IEEE Trans. Inf. Forensics. Secur., 6 (3), 1099 –1110 (2011). https://doi.org/10.1109/TIFS.2011.2129512 Google Scholar

145. 

I. Amerini et al., “Copy-move forgery detection and localization by means of robust clustering with J-linkage,” Signal Process.: Image Commun., 28 (6), 659 –669 (2013). https://doi.org/10.1016/j.image.2013.03.006 SPICEF 0923-5965 Google Scholar

146. 

V. Christlein et al., “An evaluation of popular copy-move forgery detection approaches,” IEEE Trans. Inf. Forensics Secur., 7 (6), 1841 –1854 (2012). https://doi.org/10.1109/TIFS.2012.2218597 Google Scholar

147. 

B. Diallo et al., “Robust forgery detection for compressed images using CNN supervision,” Forensic Sci. Int.: Rep., 2 100112 (2020). https://doi.org/10.1016/j.fsir.2020.100112 Google Scholar

148. 

J. H. Bappy et al., “Hybrid LSTM and encoder–decoder architecture for detection of image forgeries,” IEEE Trans. Image Process., 28 (7), 3286 –3300 (2019). https://doi.org/10.1109/TIP.2019.2895466 IIPRE4 1057-7149 Google Scholar

149. 

B. Xiao et al., “Image splicing forgery detection combining coarse to refined convolutional neural network and adaptive clustering,” Inf. Sci., 511 172 –191 (2020). https://doi.org/10.1016/j.ins.2019.09.038 Google Scholar

150. 

J. Dong and W. Wang, “CASIA tampered image detection evaluation (tide) database v1.0 and v2.0,” (2011) http://forensics.idealtest.org/ Google Scholar

151. 

Y. Hsu and S. Chang, “Detecting image splicing using geometry invariants and camera characteristics consistency,” in IEEE Int. Conf. Multimedia and Expo, 549 –552 (2006). https://doi.org/10.1109/ICME.2006.262447 Google Scholar

153. 

R. Zhang and J. Ni, “A dense U-Net with cross-layer intersection for detection and localization of image forgery,” in ICASSP 2020-2020 IEEE Int. Conf. Acoust Speech and Signal Process. (ICASSP), 2982 –2986 (2020). https://doi.org/10.1109/ICASSP40776.2020.9054068 Google Scholar

154. 

F. Z. E. Biach et al., “Encoder–decoder based convolutional neural networks for image forgery detection,” Multimedia Tools Appl., (2021). Google Scholar

156. 

P. Moulin and A. Goel, “Locally optimal detection of adversarial inputs to image classifiers,” in IEEE Int. Conf. Multimedia & Expo Workshops (ICMEW), 459 –464 (2017). https://doi.org/10.1109/ICMEW.2017.8026257 Google Scholar

157. 

A. Krizhevsky and G. Hinton, “Learning multiple layers of features from tiny images,” (2009) http://citeseerx.ist.psu.edu/viewdoc/download;jsessionid=0D60E5DD558A91470E0EA1725FF36E0A?doi=10.1.1.222.9220&rep=rep1&type=pdf Google Scholar

158. 

N. Zhu, M. Qin and Y. Yin, “Recaptured image detection based on convolutional neural networks with local binary patterns coding,” Proc. SPIE, 11198 1119804 (2019). https://doi.org/10.1117/12.2540496 PSISDG 0277-786X Google Scholar

159. 

V. Vukotić, V. Chappelier and T. Furon, “Are deep neural networks good for blind image watermarking?,” in IEEE Int. Workshop Inf. Forensics and Secur. (WIFS), 1 –7 (2018). https://doi.org/10.1109/WIFS.2018.8630768 Google Scholar

160. 

H. Kandi, D. Mishra and S. Gorthi, “Exploring the learning capabilities of convolutional neural networks for robust image watermarking,” Comput. Secur., 65 (March), 247 –268 (2017). https://doi.org/10.1016/j.cose.2016.11.016 Google Scholar

161. 

A. Fierro-Radilla et al., “A robust image zero-watermarking using convolutional neural networks,” in 7th Int. Workshop Biometrics and Forensics (IWBF), 1 –5 (2019). https://doi.org/10.1109/IWBF.2019.8739245 Google Scholar

162. 

M. Ahmadi et al., “Redmark: framework for residual diffusion watermarking based on deep networks,” Expert Syst. Appl., 146 113157 (2019). https://doi.org/10.1016/j.eswa.2019.113157 Google Scholar

163. 

S. M. Mun et al., “Finding robust domain from attacks: a learning framework for blind watermarking,” Neurocomputing, 337 (April 14), 191 –202 (2019). https://doi.org/10.1016/j.neucom.2019.01.067 NRCGEO 0925-2312 Google Scholar

164. 

X. Zhong et al., “An automated and robust image watermarking scheme based on deep neural networks,” IEEE Trans. Multimedia, 23 1951 –1961 (2021). Google Scholar

165. 

J. Zhang et al., “Deep model intellectual property protection via deep watermarking,” IEEE Trans. Pattern Anal. Mach. Intell., (2021). ITPIDJ 0162-8828 Google Scholar

166. 

X. Wang et al., “ChestX-ray8: hospital-scale chest x-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases,” in CVPR, (2017). Google Scholar

167. 

J. Li et al., “Single exposure optical image watermarking using a cGAN network,” IEEE Photonics J., 13 (2), 6900111 (2021). https://doi.org/10.1109/JPHOT.2021.3068299 Google Scholar

168. 

H. Xiao, K. Rasul and R. Vollgraf, “Fashion-MNIST: a novel image dataset for benchmarking machine learning algorithms,” (2017). https://arxiv.org/abs/1708.07747v2 Google Scholar

169. 

T. Huynh-The et al., “Robust image watermarking framework powered by convolutional encoder-decoder network,” in Digital Image Comput.: Tech. and Appl. (DICTA), 1 –7 (2019). https://doi.org/10.1109/DICTA47822.2019.8945866 Google Scholar

170. 

D. Li et al., “A novel CNN based security guaranteed image watermarking generation scenario for smart city applications,” Inf. Sci., 479 432 –447 (2018). https://doi.org/10.1016/j.ins.2018.02.060 Google Scholar

171. 

J. Hayes et al., “Towards transformation-resilient provenance detection of digital media,” (2020). https://arxiv.org/abs/2011.07355v1 Google Scholar

172. 

Y. Chen, T. Fan and H. Chao, “WMNet: a lossless watermarking technique using deep learning for medical image authentication,” Electronics, 10 (8), 932 –932 (2021). https://doi.org/10.3390/electronics10080932 ELECAD 0013-5070 Google Scholar

173. 

R. Wang et al., “Watermark faker: towards forgery of digital image watermarking,” (2021). https://arxiv.org/abs/2103.12489 Google Scholar

174. 

G. Griffin, A. Holub and P. Perona, “Caltech-256 object category dataset,” (2007) https://authors.library.caltech.edu/7694/ Google Scholar

175. 

M. W. Hatoum et al., “Using deep learning for image watermarking attack,” Signal Process. Image Commun., 90 116019 (2021). https://doi.org/10.1016/j.image.2020.116019 SPICEF 0923-5965 Google Scholar

176. 

S. S. Sharma and V. Chandrasekaran, “A robust hybrid digital watermarking technique against a powerful CNN-based adversarial attack,” Multimedia Tools Appl., 79 (43), 32769 –32790 (2020). https://doi.org/10.1007/s11042-020-09555-5 Google Scholar

177. 

D. Cheng et al., “Large-scale visible watermark detection and removal with deep convolutional networks,” in Chin. Conf. Pattern Recognit. And Comput. Vision (PRCV), 27 –40 (2018). Google Scholar

178. 

Y. Gandelsman, A. Shocher and M. Irani, “‘Double-DIP’: unsupervised image decomposition via coupled Deep-Image-Priors,” in IEEE/CVF Conf. Comput. Vision and Pattern Recognit. (CVPR), 11018 –11027 (2019). Google Scholar

179. 

A. Hertz et al., “Blind visual motif removal from a single image,” in Proc. IEEE Conf. Comput. Vision and Pattern Recognit. (CVPR), 6858 –6867 (2019). Google Scholar

180. 

X. Li et al., “Towards photo-realistic visible watermark removal with conditional generative adversarial networks,” in Int. Conf. Image and Graphics, 345 –356 (2019). Google Scholar

181. 

P. Jiang et al., “Two‐stage visible watermark removal architecture based on deep learning,” IET Image Process., 14 (15), 3819 –3828 (2021). https://doi.org/10.1049/iet-ipr.2020.0444 Google Scholar

182. 

B. Zhou et al., “Places: a 10 million image database for scene recognition,” IEEE Trans. Pattern Anal. Mach. Intell., 40 (6), 1452 –1464 (2017). https://doi.org/10.1109/TPAMI.2017.2723009 ITPIDJ 0162-8828 Google Scholar

183. 

X. Cun and C. M. Pun, “Split then refine: stacked attention-guided ResUNets for blind single image visible watermark removal,” (2020). https://arxiv.org/abs/2012.07007v1 Google Scholar

184. 

M. Shafieinejad et al., “On the robustness of the backdoor-based watermarking in deep neural networks,” (2019). https://arxiv.org/abs/1906.07745v2 Google Scholar

185. 

X. Chen et al., “REFIT: a unified watermark removal framework for deep learning systems with limited data,” in ASIA CCS ‘21: ACM Asia Conf. Comput. And Commun. Secur., (2021). Google Scholar

186. 

A. William et al., “Neural network laundering: removing black-box backdoor watermarks from deep neural networks,” Comput. Secur., 106 102277 (2021). Google Scholar

Biographies of the authors are not available.

© 2021 Society of Photo-Optical Instrumentation Engineers (SPIE)
Zhenjie Bao and Ru Xue "Survey on deep learning applications in digital image security," Optical Engineering 60(12), 120901 (29 December 2021). https://doi.org/10.1117/1.OE.60.12.120901
Received: 29 August 2021; Accepted: 9 December 2021; Published: 29 December 2021
Lens.org Logo
CITATIONS
Cited by 5 scholarly publications.
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Image encryption

Digital watermarking

Digital imaging

Steganography

Image compression

Image processing

Neural networks

Back to Top