This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Image Steganography based on Style Transfer

Abstract

Image steganography is the art and science of using images as cover for covert communications. With the development of neural networks, traditional image steganography is more likely to be detected by deep learning-based steganalysis. To improve upon this, we propose image steganography network based on style transfer, and the embedding of secret messages can be disguised as image stylization. We embed secret information while transforming the content image style. In latent space, the secret information is integrated into the latent representation of the cover image to generate the stego images, which are indistinguishable from normal stylized images. It is an end-to-end unsupervised model without pre-training. Extensive experiments on the benchmark dataset demonstrate the reliability, quality and security of stego images generated by our steganographic network.

Index Terms—  Image steganography, style transfer, steganalysis

1 Introduction

Image steganography aims to hide secret information in images without being detected on social networks. The cover image is the one that will be modified to carry the secret message, while the stego image is the output of the steganography algorithm. Traditional image steganography [1] mainly adopts the Least Significant Bits (LSB) distribution method to replace the less important pixels in the cover image to carry secret information, however, this method cannot resist steganalysis based on high-dimensional features [2, 3, 4, 5, 6]. Then adaptive steganography algorithm [7, 8, 9] proposed based on the minimum distortion embedding framework, which tends to embed secret messages into textured and noisy regions, but still faces the threat of detection by deep learning-based steganalysis [10].

Therefore, many researchers have attempted to combine steganography with deep learning. To date, Generative Adversarial Networks (GANs) [11] have been widely used for image generation. It has achieved state-of-the-art performance in many areas, including super-resolution imaging, style transfer, etc. Based on GAN, Hu et al. [12] and Wang et al. [13] proposed the embedding-free steganography (SwE) method respectively, which learns the steganographic algorithm in an unsupervised manner and directly generates the stego image from the secret message. However, the secret information inevitably interferes with the generated image features and the generated image looks unnatural and thus susceptible to suspicion. Recently, style transfer has received a lot of attention, Wang et al. [14] proposed steganography network STNet based on style transfer, but it relies on pre-trained CNNs on ImageNet and introduces additional bias because it is assembled without artistic considerations.

To resolve the above-mentioned limitation, in this paper, we propose an image steganography model based on style transfer. We adopted an encoder-decoder architecture, which includes information preprocessing module, information hiding module and information recovery module as shown in Fig.1. The secret information is input into the information hiding module after passing through the information preprocessing module. The information hiding module includes encoder, decoder and discriminator. The encoder maps the content image into latent representations and the processed secret information is embedded in this process. The decoder generates the stylized stego image. The discriminator is used to distinguish the generated image from the real artistic image to optimize the generated image. The extractor in the information recovery module is used to extract the secret information in the stego image. The steganographic network combined with the style transfer makes it difficult for the steganalyzer to distinguish between the stego image and the stylized cover image.

In summary, the main contributions of this work are as follows:

  • We designed a steganography model based on style transfer, which embeds secret information while transforming the content image style, thus could reduce the interference of secret information on the generated image.

  • It is an end-to-end unsupervised model without pre-training, greatly reducing the computational overhead.

  • We added channel attention to the extractor and obtained a high message extraction accuracy.

Refer to caption
Fig. 1: The overall framework of the image steganography model based on style transfer.

2 Related work

With the rapid development of social networking and multimedia technologies, image stylization has been widely used in various social media. Style transfer is a method to modify the image style while retaining the original image content. Gatys et al. [15, 16] proposed an iterative optimization method based on a VGG19 network [17]. This method produces high-quality results and is suitable for any inputs, but each optimization step requires forward propagation and backpropagation through the VGG19 network. Sanakoyeu et al. [18] proposed a method can be trained on arbitrary unpaired content and style images without back propagation.

Based on style transfer, Yi et al. [19] proposed a scheme that uses an edge marker embedding scheme to realize information hiding and convert content images into comic images. The comic style generated by this traditional steganography method lacks diversity. Wang et al. [14] proposed steganography network STNet based on style transfer, wherein the encoder uses the VGG19 [17] network. However, this model is limited to the combination of a single style image and a content image to show a similar combined image and does not consider the details of the artistic images.

3 Approach

Our model is shown in Fig.1. The steganography framework is mainly divided into three modules, information preprocessing module, information hiding module and information recovery module. The information preprocessing module maps the secret information mm to the secret feature MM, binds MM to the content feature FF through the dot product operation, and outputs the preprocessed secret feature MM^{{}^{\prime}}. Among them, FF is the feature extracted from the content image II by a set of high-pass filters. The information hiding module has an encoder-decoder structure. II is input to the encoder network to generate content features AA, and then, AA is connected with MM^{{}^{\prime}} to feed to the decoder. The style stego image stegostego is the output of the decoder. The discriminator is used to measure the gap between stegostego and the real style image II^{{}^{\prime}}. We trained an extractor network with channel attention mechanism and used the optimized filters in the first layer of the extractor. The secret information mm is recovered from stegostego by the extractor network. Next, we will introduce each part in detail.

3.1 Information preprocessing module

Refer to caption
Fig. 2: Secret message mapping.

To better obtain the content feature FF used for secret information mapping, we used a set of high-pass filters (a total of 32 filters, each with a size of 5×5 and three channels), including the high-pass filter of SRM [20] and its rotating counterparts. The high-pass filter with learnable parameters is used to sample the high-frequency signal from the image. The texture features and spatial position information in the image are extracted as a whole and then passed through a 4-conv layer network and activated by the Leaky ReLU function and the output is the content feature FF. In this paper, rather than directly concatenate the secret information with the content image that might not fully exploit the semantic information represented by the latent space, as well as corrupting the original features of the input image. Inspired by image enhancement, we propose a new secret information embedding scheme. As shown in Fig. 2, the combined features:

M\displaystyle M^{\prime} =\displaystyle= MF,\displaystyle M\odot F, (1)

where MM is the secret information matrix, FF is the image feature map, MRH×W×CM^{{}^{\prime}}\in R^{H\times W\times C} is the processed feature by dot product operation. HH and WW represent the height and width of the matrix, CC is the number of channels. ”\odot” represents the dot product operation of the corresponding position elements of the matrix. The secret information matrix is multiplied by the image feature map, which is feature selection. Therefore, the actual embedded secret information is closely related to the texture. Due to the above reasons, when the secret information vector is mapped to the matrix space, it will be forced to learn to arrange the secret information into the matrix according to the image texture characteristics.

3.2 Information hiding module

Information hiding module is designed on the basis of the style transfer model [18]. The model mainly includes encoder EE, decoder GG and discriminator DD. As shown in Fig. 3, the encoder network consists of 5 convolutional layers: 1×\timesconv-stride-1 and 4×\timesconv-stride-2. The decoder network contains 9 residual blocks [21], 4 up-sampling blocks and 1×\timesconv-stride-1. The discriminator is a fully convolutional network with 7×\timesconv-stride-2 layers.

Refer to caption
Fig. 3: Network structure of information hiding module.

The style of the generated stego image is controlled by DD. Therefore, the style loss function is also the discriminator loss function to distinguish the output stylized image G(E(xi))G(E(x_{i})) from the real style sample siIs_{i}\in I^{{}^{\prime}} , xiIx_{i}\in I is a content image. The style loss function of our model is given by

Lstyle=𝔼spstyle[logD(s)]+𝔼xpx(x)[log(1D(G(E(x))))].\displaystyle\begin{aligned} L_{style}=&\mathop{\mathbb{E}}_{s\sim p_{style}}[\log D(s)]+\\ &\mathop{\mathbb{E}}_{x\sim p_{x}(x)}[\log(1-D(G(E(x))))].\end{aligned} (2)

We map the secret information into a feature map and hide the secret while generating the stylized image. The full objective function of our model is

L(E,G,D)=λLstyle+Lc(E,G),\displaystyle\begin{aligned} L(E,G,D)&=&\lambda L_{style}+L_{c}(E,G),\end{aligned} (3)

EE and GG are trained on the content image set and the style image set to generate stylized stego images. λ\lambda controls the relative importance of style loss.

3.3 Information recovery module

When recovering secret information from the received stego image, a positive number represents 11 and a negative number represents 0. The secret loss function is used to measure the difference between the output of the extractor network and the original secret information and is given by

Lm=min(max(0,(mm^)2δ))\displaystyle L_{m}\ =\ min(max(0,(m-\widehat{m})^{2}-\delta))
mpsecret(m),\displaystyle m\sim p_{secret}(m), (4)

where mm is the secret information, m^\widehat{m} is the extracted secret information by the extractor network, δ\delta is the upper threshold of the error accuracy of the secret information. We use a set of high-pass filter in the first layer of the extractor to better obtain image features with secret information and improve the secret information recovery capability of the extractor network. In the extractor of our model, we also added a channel attention mechanism [22] so that the extractor can pay attention to image features with rich textures because the secret information is embedded in this area. We first extract the feature map IfcI_{f_{c}} from the stego image. Then the IfcI_{f_{c}} will be fed into the attention model and the model is formulated as

sc\displaystyle s_{c} =\displaystyle= δ(W2(δ(W1(ϕ(Ifc))))),\displaystyle\delta(W_{2}(\delta(W_{1}(\phi(I_{f_{c}}))))), (5)

where cc represents the number of feature channels, δ\delta is the ReLU [23] function, ϕ\phi is the pooling layer. W1Rcr×cW_{1}\in R^{\frac{c}{r}\times c}, W2Rc×crW_{2}\in R^{c\times\frac{c}{r}} the hyperparameter rr is reduced to a coefficient. Finally, multiply the learned activation value scs_{c} by the original feature IfcI_{f_{c}} to obtain a feature map with different channel attention,

xc~\displaystyle\widetilde{x_{c}} =\displaystyle= scIfc.\displaystyle s_{c}\cdot I_{f_{c}}. (6)

Channel attention mechanism make the extractor focus on regions with rich texture and improve the extraction of secret information.

4 Experiment

In the training process, we use the Places365 [24] dataset as the content image, and the WikiArt [25] dataset style image block is sampled. The dataset Places365365 contains approximately 1.81.8 million images from 365365 scene categories, each with a maximum of 50005000 images. Therefore, it can provide rich image content for online learning. The WikiArt art image data set contains a large number of art style images of various genres and styles, which are used to provide a variety of training styles, such as Cézanne, Monet, Gauguin and other styles of artistic images. During the training process, we randomly crop an area with a size of 256×256256\times 256. We use batch size 11, learning rate 0.00020.0002 and Adam [26] optimizer training for 300,000300,000 iterations. A string of 0 or 11 bits is randomly generated as the secret information vector. In information recovery module, the recovered information is compared with the input secret information in bit unit, which is used as an evaluation index for the accuracy of information extraction.

4.1 Reliability

Here we evaluate the accuracy of secret information recovery. To improve the recovery rate of secret information, we optimized the filters for extracting the texture features of the stego image and added a channel attention mechanism to the extractor. To verify whether the above two modules are useful for message extraction, we designed three ablation experiments at a message embedding size of 1000 bits. The experimental results are shown in Table 1. As we can see that with the optimized filters and channel attention can improve the recovery accuracy of secret information.

Then we compare the accuracy with STNet [14], which is also a steganography network based on style transfer, at different embedding size. The experimental results are shown in Table 2. We can see that the message recovery performance of these two methods is comparable, although our method is slightly weaker, but in practice we can also achieve 100%100\% after using the error correction code. In addition, STNet is limited to a blunt combination of images and styles, and the quality of the generated images is weaker than our method as shown below.

Table 1: Message recovery accuracy under different experimental conditions.
Channel attention w/ w/o w/
Optimized filters w/o w/ w/
Recovery accuracy 0.96 0.98 0.99
Table 2: Comparison of message recovery accuracy of different methods at different embedding size.
Embedding size (bits) 1000 2000
STNet [14] 0.99 0.99
Ours 0.99 0.98

4.2 Quality

In Fig.4, we show the stylized stego image generated by our model and STNet. Stego images generated by our steganography model are like oil paintings created by artists, which are highly artistic. In order to quantitatively evaluate the quality of the stylized stego images generated by our model and STNet model, we evaluate these images through SSIM. The results are displayed in Table 3. In all three art styles, the SSIM evaluation index of images generated by our model is comparable to that of STNet. In addition,STNet can only be trained to combine style images and content images and cannot learn the general features of one class of style images.In contrast, the image generated by our method is more like an artistic oil painting, and our method is more artistic.

Refer to caption
Fig. 4: Comparison of the images generated by our steganography model with the STNet model.
Table 3: The quality of the stego images.
SSIM Cézanne Gauguin Monet
STNet 0.93 0.92 0.92
Ours 0.93 0.93 0.91

4.3 Security

Unlike the security pursued by traditional steganography, the goal is to make stego and cover indistinguishable. Our model aims to make the generated stylized steganographic images indistinguishable from the normal stylized images. Moreover, our model generates subtly different images for different training under the same message embedding, making the steganalysis network more difficult to train. Fig. 5 shows the stylized stego images generated by the same model in different iterations. The images from (a) to (e) have slight color differences in the same style. This is consistent with a realistic scenario, as the images created are not the same each time, even for the same work created by the same artist. From a security point of view, generating similar but different steganographic images from the same model can further resist the training of steganalysis networks.

Refer to caption
Fig. 5: The training step size has a subtle effect on the generated image. The color difference can be seen at the indicated position.

Our model embeds secret information during the stylization of content images, and current steganalysis cannot determine whether the modified part of the generated images come from style transfer or information embedding. And it is meaningless for steganalyzer to distinguish between the generated image and the content image, because the generated stego image is based on the stylized image. In our security experiment design, after receiving the generated image, the steganalyzer needs to determine whether the stylized image contains secret information. Therefore the cover image is a stylized image without secret information, stego image is a stylized image with secret information. We use SRNet [10], the typical steganalysis network, to evaluate the security of our model. The experimental data settings are: the artistic style of the generated image is Cézanne, the style image generated by the style transfer model [18] is cover, and the style stego image generated by our model is stego.

To evaluate the security of our stylized stego images, we distinguish three possible kinds of steganalyzers: 1) An Ignorant Steganalyzer who has obtained our model, but does not know which iterative model and only has one iterative model. Therefore, we set the training dataset as the cover and stego image pairs generated by corresponding models saved in a certain iteration, the testing dataset as the image pair generated by corresponding models saved in another iteration; 2) A Knowledgeable Steganalyzer who knows different iterative models, but the number of iterations of the generated stego is not known; and 3) An Omniscient Steganalyzer who knows iterations of generated stego. The experiment results are shown in Table 4. Under the embedding size of 1000 bits, for the ignorant steganalyzer, the detection accuracy of ours is 0.55, 20%20\% less than STNet, which is equivalent to blind guessing. The knowledgeable steganalyzer achieves a detection accuracy of 0.60 and omniscient steganalyzer achieves a detection accuracy of 0.71, 27%27\% and 22%22\% less than STNet, respectively. These results show that our security is stronger than STNet in different situations and difficult to be detected by ignorant steganalyzer in practical applications.

Table 4: Classification accuracy of SRNet under different steganalyzers.
Steganalyzer\Embedding size (bits) 1000 2000
Ignorant STNet 0.83 0.98
Ours 0.55 0.59
Knowledgeable STNet 0.87 0.99
Ours 0.60 0.73
Omniscient STNet 0.93 0.99
Ours 0.71 0.84

5 Conclusions

In this paper, we propose a steganography model based on style transfer, which hides secret information during the process of style transfer. The stylized stego images generated by this model are artistic in nature, making it less susceptible to suspicion when transmitted over the Internet. In the experiments, we test the performance of the proposed method compared with the STNet. Our method has a high message recovery accuracy and is more practical and is more artistic, because it can capture the subtle details of style images through training. In terms of security, in resisting deep-learning based steganalysis network detection, our method has high security performance in all three scenarios, satisfying the pursuit of steganography. Our future work is to improve the robustness and embedding rate of the method.

References

  • [1] Xiaolong Li, Tieyong Zeng, and Bin Yang, “Improvement of the embedding efficiency of lsb matching by sum and difference covering set,” in 2008 IEEE International Conference on Multimedia and Expo, 2008, pp. 209–212.
  • [2] S. Dumitrescu and Xiaolin Wu, “Steganalysis of lsb embedding in multimedia signals,” in Proceedings. IEEE International Conference on Multimedia and Expo, 2002, vol. 1, pp. 581–584 vol.1.
  • [3] Xiaoyi Yu and Noboru Babaguchi, “Weighted stego-image based steganalysis in multiple least significant bits,” in 2008 IEEE International Conference on Multimedia and Expo, 2008, pp. 265–268.
  • [4] Xiaoyi Yu and Noboru Babaguchi, “Run length based steganalysis for lsb matching steganography,” in 2008 IEEE International Conference on Multimedia and Expo, 2008, pp. 353–356.
  • [5] Seongho Cho, Jingwei Wang, C.-C. Jay Kuo, and Byung-Ho Cha, “Block-based image steganalysis for a multi-classifier,” in 2010 IEEE International Conference on Multimedia and Expo, 2010, pp. 1457–1462.
  • [6] J. Fridrich and M. Long, “Steganalysis of lsb encoding in color images,” in 2000 IEEE International Conference on Multimedia and Expo. ICME2000. Proceedings. Latest Advances in the Fast Changing World of Multimedia (Cat. No.00TH8532), 2000, vol. 3, pp. 1279–1282 vol.3.
  • [7] Tomáš Pevnỳ, Tomáš Filler, and Patrick Bas, “Using high-dimensional image models to perform highly undetectable steganography,” in International Workshop on Information Hiding. Springer, 2010, pp. 161–177.
  • [8] Vojtěch Holub, Jessica Fridrich, and Tomáš Denemark, “Universal distortion function for steganography in an arbitrary domain,” EURASIP Journal on Information Security, vol. 2014, no. 1, pp. 1–13, 2014.
  • [9] Vojtěch Holub and Jessica Fridrich, “Designing steganographic distortion using directional filters,” in 2012 IEEE International workshop on information forensics and security (WIFS). IEEE, 2012, pp. 234–239.
  • [10] Mehdi Boroumand, Mo Chen, and Jessica Fridrich, “Deep residual network for steganalysis of digital images,” IEEE Transactions on Information Forensics and Security, vol. 14, no. 5, pp. 1181–1193, 2018.
  • [11] Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio, “Generative adversarial nets,” Advances in neural information processing systems, vol. 27, 2014.
  • [12] Donghui Hu, Liang Wang, Wenjie Jiang, Shuli Zheng, and Bin Li, “A novel image steganography method via deep convolutional generative adversarial networks,” IEEE Access, vol. 6, pp. 38303–38314, 2018.
  • [13] Zihan Wang, Neng Gao, Xin Wang, Xuexin Qu, and Linghui Li, “Sstegan: self-learning steganography based on generative adversarial networks,” in International Conference on Neural Information Processing. Springer, 2018, pp. 253–264.
  • [14] Zihan Wang, Neng Gao, Xin Wang, Ji Xiang, and Guanqun Liu, “Stnet: A style transformation network for deep image steganography,” in International Conference on Neural Information Processing. Springer, 2019, pp. 3–14.
  • [15] Leon A Gatys, Alexander S Ecker, and Matthias Bethge, “Texture synthesis and the controlled generation of natural stimuli using convolutional neural networks,” in Bernstein Conference 2015, 2015, pp. 219–219.
  • [16] Leon A Gatys, Alexander S Ecker, and Matthias Bethge, “Image style transfer using convolutional neural networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 2414–2423.
  • [17] Karen Simonyan and Andrew Zisserman, “Very deep convolutional networks for large-scale image recognition,” arXiv preprint arXiv:1409.1556, 2014.
  • [18] Artsiom Sanakoyeu, Dmytro Kotovenko, Sabine Lang, and Bjorn Ommer, “A style-aware content loss for real-time hd style transfer,” in proceedings of the European conference on computer vision (ECCV), 2018, pp. 698–714.
  • [19] Yi Puyang, Zhenxing Qian, Zhaoxia Yin, and Xinpeng Zhang, “Style transferring based data hiding for color images,” in International Conference on Cloud Computing and Security. Springer, 2018, pp. 440–449.
  • [20] Jessica Fridrich and Jan Kodovsky, “Rich models for steganalysis of digital images,” IEEE Transactions on Information Forensics and Security, vol. 7, no. 3, pp. 868–882, 2012.
  • [21] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.
  • [22] Jie Hu, Li Shen, and Gang Sun, “Squeeze-and-excitation networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 7132–7141.
  • [23] Vinod Nair and Geoffrey E Hinton, “Rectified linear units improve restricted boltzmann machines,” in Icml, 2010.
  • [24] Bolei Zhou, Agata Lapedriza, Jianxiong Xiao, Antonio Torralba, and Aude Oliva, “Learning deep features for scene recognition using places database,” 2014.
  • [25] Sergey Karayev, Matthew Trentacoste, Helen Han, Aseem Agarwala, Trevor Darrell, Aaron Hertzmann, and Holger Winnemoeller, “Recognizing image style,” arXiv preprint arXiv:1311.3715, 2013.
  • [26] Diederik P Kingma and Jimmy Ba, “Adam: A method for stochastic optimization,” arXiv preprint arXiv:1412.6980, 2014.