FaceSigns: Semi-Fragile Neural Watermarks for Media Authentication and Countering Deepfakes
Abstract
Deepfakes and manipulated media are becoming a prominent threat due to the recent advances in realistic image and video synthesis techniques. There have been several attempts at combating Deepfakes using machine learning classifiers. However, such classifiers do not generalize well to black-box image synthesis techniques and have been shown to be vulnerable to adversarial examples. To address these challenges, we introduce a deep learning based semi-fragile watermarking technique that allows media authentication by verifying an invisible secret message embedded in the image pixels. Instead of identifying and detecting fake media using visual artifacts, we propose to proactively embed a semi-fragile watermark into a real image so that we can prove its authenticity when needed. Our watermarking framework is designed to be fragile to facial manipulations or tampering while being robust to benign image-processing operations such as image compression, scaling, saturation, contrast adjustments etc. This allows images shared over the internet to retain the verifiable watermark as long as face-swapping or any other Deepfake modification technique is not applied. We demonstrate that FaceSigns can embed a 128 bit secret as an imperceptible image watermark that can be recovered with a high bit recovery accuracy at several compression levels, while being non-recoverable when unseen Deepfake manipulations are applied. For a set of unseen benign and Deepfake manipulations studied in our work, FaceSigns can reliably detect manipulated content with an AUC score of which is significantly higher than prior image watermarking and steganography techniques111Inference Demo: https://github.com/paarthneekhara/FaceSignsDemo.
Keywords:
Deepfakes, watermarking, media forensics1 Introduction
Media authentication, despite having been a long-term challenge, has become even more difficult with the advent of deep learning based generative models. Deep Neural Network (DNN) based media synthesis techniques [45, 10, 23, 50, 31, 22], have enabled the creation of high-quality convincing fake photos and videos. One such example of synthetic videos is Deepfakes, in which a subject’s face is swapped with a target face to create convincingly realistic footage of events that never occurred [37, 19]. Such manipulated videos can fuel misinformation, defame individuals and reduce trust in media [29]. Recent Deepfake detection methods rely on DNN based classifiers to distinguish synthetic videos from real videos [37, 12]. However, with advances in Deepfake synthesis methods, DNN based classification methods are failing to stay a step ahead in the race to reliably detect Deepfakes. This is largely because classifiers trained in a supervised manner on existing Deepfake generation methods cannot be reliably secure against black-box Deepfake generation methods. Moreover, the current best performing Deepfake detectors can be easily bypassed by attackers using adversarial examples [19, 15, 30].

We believe that proactively embedding a secret verifiable message into the images when they are captured from a device, can establish the provenance of real images and videos and address the limitations of Deepfake classifiers. Several prior works have explored digital image watermarking and deep learning based steganography techniques [9, 14, 44, 54, 27] to hide secret messages in image pixels. However, these works are either fragile to basic image processing operations such as compression and color adjustments or overly robust to the point that the secret can be recovered even after occluding major portions of the embedded image [44]. In fact, we experimentally demonstrate that past works on robust neural watermarks [44, 54] can recover messages even from images that have undergone Deepfake manipulations.
To address the above challenges of Deepfake classifiers and neural watermarking frameworks, we introduce FaceSigns — a deep learning based semi-fragile watermarking system that embeds a recoverable message as an imperceptible perturbation in the image pixels. The watermark can contain a secret message or device-specific codes which can be used for authenticating images and videos. The desirable property of the watermark is that it should break if a malicious facial manipulation such as a Deepfake transformation is applied to the image, but it should be robust against harmless image transformations such as image compression, color and lighting adjustments which are commonly applied on pictures before uploading them to online sharing platforms. To achieve this goal, we develop an encoder-decoder based training framework that encourages message recovery under benign transformations and discourages message recovery if the watermark is tampered in the facial regions of the image. In contrast to hand-designed pipelines used in previous work for semi-fragile watermarking [25, 17, 48, 35, 4], our framework is end-to-end and learns to be robust to a wide range of real-world digital image processing operations such as social media filters and compression techniques, while being fragile to various Deepfake tampering techniques. The technical contributions of our work are as follows:
-
1.
We develop a neural semi-fragile image watermarking framework as a proactive defense against Deepfakes. To the best of our knowledge, our work is the first deep learning based semi-fragile watermarking system that can certify the authenticity of digital media.
-
2.
We design a differentiable procedure to simulate facial watermark tampering during training such that our framework can achieve selective fragility against unseen malicious transformations (Section 3.2.2).
-
3.
For a set of previously unseen benign and malicious image transformations, FaceSigns achieves the goal of selective fragility and reliably detects Deepfake manipulations with an AUC score of which is significantly higher than alternate robust and semi-fragile watermarking frameworks.
2 Background
2.1 Facial Forgery
Until recently, the ease of generating manipulated photos and videos has been limited by manual editing tools. However, since the advent of deep learning there has been significant work in developing new techniques for automatic digital forgery. DNN based facial manipulation methods [23, 6, 10, 7] operate end-to-end on a source video and target face, and require minimal human expertise to generate fake videos in real-time. In our work we show effectiveness against popular GAN based Deeepfake generation methods - SimSwap [6] and FSFT [39], and a classical computer graphics based face replacement approach - FaceSwap [23].
The best performing Deepfake detectors [1, 13, 37, 38, 46] rely on convolutional neural network (CNN) based architectures. Such Deepfake detectors model Deepfake detection as a per-frame binary classification problem, using a face-tracking method prior to CNN classification to effectively detect facial forgeries in both uncompressed and compressed videos. While CNN based classifiers achieve promising detection accuracy on a fixed in-domain test set of real and fake videos, they suffer from two main drawbacks: 1) Lack of generalizability to unseen Deepfake synthesis techniques. 2) Vulberability to adversarial examples in both black-box and white-box attack settings. We refer the reader to past works [15, 18, 19, 28] that explore such limitations of CNN Deepfake detectors.
2.2 Digital Watermarking
Digital watermarking [9] similar to steganography [14], is the task of embedding information into an image in a visually imperceptible manner. These techniques broadly seek to generate three different types of watermarks: fragile [11, 4], robust [8, 34, 5, 41, 54, 33] and semi-fragile [25, 43, 49] watermarks. Fragile and semi-fragile watermarks are primarily used to certify the integrity and authenticity of image data. Fragile watermarks are used to achieve accurate authentication of digital media, where even a one-bit change to an image will lead it to fail the certification system. In contrast, robust watermarks aim to be recoverable under several image manipulations, in order to allow media producers to assert ownership over their content even if the video is redistributed and modified. Semi-fragile watermarks combine the advantages of both robust and fragile watermarks, and are mainly used for fuzzy authentication of digital images and identification of image tampering [49]. The use of semi-fragile watermarks is justified by the fact that images and videos are generally transmitted and stored in a compressed form, which should not break the watermark. However when the image gets tampered, the watermark should also get damaged, indicating image tampering.
Several past works have proposed hand-engineered pipelines to embed semi-fragile watermark information in the spatial and frequency (transform) domain of images and videos. In the spatial domain, the pixels of digital images are processed directly using block-based embedding [4] and least significant bits modification [47, 48] to embed watermarks. In the frequency domain, the watermark can be embedded by modifying the coefficients produced with transformations such as the Discrete Cosine Transform (DCT) [35, 17] and Discrete Wavelet Transform [24, 3, 40]. However, we demonstrate in our experiments the major limitations of traditional approaches lies in higher visibility of the embedded watermarks, increased distortions in generated images and low robustness to compression techniques like JPEG transforms. Moreover, these works have not been designed to be fragile against Deepfake manipulations.
More recently, CNNs have been used to provide an end-to-end solution to the watermarking problem. They replace hand-crafted hiding procedures with neural network encoding [2, 16, 54, 52, 27, 44]. Notably, both StegaStamp [44] and HiDDeN [54] propose frameworks to embed robust watermarks that can hide and transmit data in a way that is robust to various real-world transformations. All of these works focus on generating robust watermarks, with the goal of ensuring robustness and recovery of the embedded secret information under various physical and digital image distortions. We empirically demonstrate that these techniques are unable to generate semi-fragile watermarks and are therefore not suitable for identifying tampered media such as Deepfakes.
3 Methodology

We aim to develop an image watermarking framework that is robust to a set of benign image transformations (eg. brightness and contrast adjustments, resizing, JPEG compression (etc.)) while being fragile to facial manipulations like GAN-based face swapping or conventional face editing. Additionally, it is desirable to have an imperceptible watermark so that devices can store only the watermarked images without revealing the original image to the end user. To this end, our system consists of three main components: an encoder network , a decoder network and an adversarial discriminator network where are learnable parameters.
The encoder network takes as input an image and a bit string of length , and produces an encoded (watermarked) image . That is, . The watermarked image then goes through two image transformation functions — one sampled from a set of benign transformations () and the other sampled from a set of malicious transformations () to produce a benign image and a malicious image . The benign and malicious watermarked images are then fed to the decoder network which predicts the messages and respectively.
For optimizing secret retrieval during training, we use the distortion between the predicted and ground-truth bit strings. The decoder is encouraged to be robust to benign transformations by minimizing the message distortion ; and fragile for malicious manipulations by maximizing the error . Therefore the secret retrieval error for an image is obtained as follows:
(1) |
The watermarked image is encouraged to look visually similar to the original image by optimizing three image distortion metrics: , and [51] distortions. Additionally, we use an adversarial loss from the discriminator which is trained simultaneously to distinguish original images from watermarked images. That is, our image reconstruction loss is obtained as follows:
(2) |
Therefore, the parameters of the encoder and decoder network are trained using mini-batch gradient descent to optimize the following loss over a distribution of input messages and images:
(3) |
The discriminator parameters are trained to distinguish original images from watermarked images as follows:
(4) |
In the above equations, , , are scalar coefficients for the respective loss terms. We refer the readers to our supplementary material for the values we use for these coefficients and other implementation details.
3.0.1 Message encoding
The encoder network accepts watermarking data as a bit string of length . This watermarking data can contain information about the device that captured the image or a secret message that can be used to authenticate the image. To prevent adversaries (who have gained white-box access to the encoder network) from encoding a target message, we can encrypt the message using symmetric or asymmetric encryption algorithms or hashing. In our experiments, we embed encrypted messages of size 128 bits which allows the network to encode unique messages. We discuss the possible threats and defenses to our watermarking framework in Section 5.
3.1 Network Architectures
Our encoder and decoder networks are based on the U-Net CNN architecture [36, 20, 44] and operate on images. The encrypted message , which is an length bit string, is first projected to a tensor of size using a trainable fully connected layer; then resized to using bilinear interpolation and finally added as the fourth channel to the original RGB image to be fed as an input to the encoder network. The encoder U-Net contains 8 downsampling and 8 unsampling layers. We modify the original U-Net architecture and replace the transposed convolution in the upsampling layers with convolutions followed by nearest-neighbour upsampling as per the recommendations given by [32]. In our preliminary experiments, we found this change to significantly improve the image quality and training speed of our framework. The decoder network also follows the U-Net architecture similar to our encoder network. The decoder U-Net first outputs a intermediate output, which is downsized to using bilinear down-sampling to produce and then projected to a vector of size using a fully connected layer followed by a sigmoid layer to scale values between and .
For the discriminator network, we use the patch discriminator from [20]. The discriminator is trained to classify each image patch as real or fake. We average discriminator responses across all patches to obtain the discriminator output. Our discriminator network consists of three convolutional blocks of stride 2 thereby classifying patches of size .
3.2 Transformation functions
The choice of benign and malicious transformation functions is critical to achieve selective fragility and robustness of the watermark. While we can only use a limited set of image transformations during training, the list of possible benign and malicious transforms in real-world setting is non exhaustive. In our experiments (Section 4.3), we demonstrate that by incorporating the below described transformation functions, we are able to generalize to unseen benign and malicious transformations.
3.2.1 Benign Transforms
Our goal is to authenticate real images shared over online platforms that generally undergo diverse color or lighting adjustments (e.g. Instagram filters). Therefore, to approximate standard image processing distortions, we apply a diverse set of differentiable benign image transformations () to our watermarked images during training:
1. Gaussian Blur: We convolve the original image with a Gaussian kernel . This transform is given by where is the convolution operator. We use kernel sizes ranging from to
2. JPEG compression: Digital images are usually stored in a lossy format such as JPEG. We approximate JPEG compression with the differentiable JPEG function proposed in [42]. During training, we apply JPEG compression with quality , and .
3. Saturation adjustments: To account for various color adjustments from social media filters, we randomly linearly interpolate between the original (full RGB) image and its grayscale equivalent.
4. Contrast adjustments: We linearly rescale the image histogram using a contrast factor
5. Downsizing and Upsizing: The image is first downsized by a factor scale and then up-sampled by the same factor using bilinear upsampling. We use
6. Translation and rotation: The image is shifted horizontally and vertically by and pixels where and rotated by degrees where .
For each mini-batch iteration, we sample one transformation function from the above list (and an Identity transform) and apply it to all the images in the batch.
3.2.2 Malicious Transforms

Our semi-fragile watermarks have to be unrecoverable when facial tampering such as face swapping or GAN based manipulations are applied. Explicitly including Deepfake manipulation systems in our training pipeline can create same drawbacks as supervised Deepfake classifiers and lead to lack of generalization to unseen Deepfake synthesis techniques.
The common operation across Deepfake techniques is modifying facial features to achieve the appearance of the target identity. Therefore, we model malicious tampering as a transformation function that selectively modifies the watermark in certain regions of the face. For each image, we initialize a mask of all ones. Next, we extract the facial feature polygons for eyes, nose and lips and set the values for all pixels inside the polygons to a small watermark retention percentage . That is, for all in the facial feature polygons. Finally, the maliciously transformed image is obtained as follows:
Figure 3 illustrates the malicious transform procedure. The underlying assumption is that Deepfake manipulations modify facial features thereby tampering the watermark in those regions.
4 Experiments
4.1 Datasets and Experimental Setup
We conduct our experiments on the CelebA dataset [26] which is a large scale database of over face images of unique celebrities. We set aside images for testing the watermarking models and split the remaining data into training and validation. We train our models for mini-batch iterations with a batch-size of and use an Adam optimizer with a fixed learning rate of . All our models are trained using images of size which are obtained after center-cropping and resizing the CelebA images. We conduct experiments with message length . To evaluate the effectiveness of using transformation functions during training, we conduct an ablation study by training a FaceSigns (No Transform) model that does not incorporate any input transformations and a FaceSigns (Robust) model that uses only benign transformations during training. We evaluate watermarking techniques primarily on the following aspects:
1. Imperceptibility: We compare the original and watermarked images to compute: peak signal to noise ratio (PSNR) and structural similarity index (SSIM). Higher values for both PSNR and SSIM are desirable for a more imperceptible watermark.
2. Robustness and Fragility: To measure the robustness and fragility of the watermarking system we measure the bit recovery accuracy (BRA) of the bit string when unseen (not used in training) benign and malicious image transformations are applied. For robustness, it is desirable to have a high BRA against benign transformations like social media filters and image compression. For fragility against Deepfake synthesis techniques, it is desirable to have a low BRA when a Deepfake facial manipulation is applied. To make a fair comparison with past works, we do not apply any bit error correcting codes while calculating the BRA and compare the input string with the raw decoder output. A detector can classify an input as manipulated if the BRA of the decoded message is below a set threshold and benign if the BRA is more than the threshold. We measure the performance of such a detector using the AUC score - Area under the ROC curve.
3. Capacity: Capacity measures the amount of information that can be embedded in the image. We measure the capacity as the bits per pixel (BPP) that is the number of bits of the encrypted message embedded per pixel of the image which is simply .
It is important to note the trade-off between the above metrics—e.g. models with higher capacity, sacrifice on the imperceptibility or bit recovery accuracy. Similarly, more robust models sacrifice capacity or imperceptibility. We compare our watermarking framework against three prior works on image watermarking — a DCT based semi-fragile watermarking system [17] and two neural image watermarking systems HiDDeN [54] and StegaStamp [44]. Both HiDDeN and StegaStamp embed a bit string message into a square RGB image while ensuring robustness to a set of image transformations.
Capacity | Imperceptibility | ||||
---|---|---|---|---|---|
Method | H,W | BPP | PSNR | SSIM | |
SemiFragile DCT [17] | |||||
HiDDeN [54] | |||||
StegaStamp [44] | |||||
FaceSigns (No Transform) | |||||
FaceSigns (Robust) | |||||
FaceSigns (Semi-Fragile) |

4.2 Imperceptibility and Capacity
We report the image similarity and capacity metrics of different watermarking techniques in Table 1. We find that even at a higher message capacity, FaceSigns can encode messages with better imperceptibility as compared to StegaStamp and HiDDeN. As noted by the authors of StegaStamp and visible in our Figure 4, the residual added by their model is perceptible in large low frequency regions of the image. We believe that this is primarily due to the difference in our network architecture choices. In our initial experiments, we found that using a UNet architecture for the decoder with an intermediate message reconstruction loss described in Section 3.1, performed significantly better than a down-sampling CNN architecture used in prior work. Additionally, we use nearest neighbour upsampling instead of transposed convolutions in our U-Net architectures which helps reduce the perceptibility of the watermark by removing upsampling artifacts.
4.3 Robustness and Fragility
To study the robustness and fragility of different DNN based watermarking techniques, we transform the watermarked images using unseen benign and malicious transformations and then attempt to decode the message from the transformed message. In this evaluation we consider out-of-domain and practical transformations which are generally applied before images are uploaded to the internet.

For benign transformations, we compress the image using different levels of JPEG compression (separate from training) and also apply Instagram filters namely Aden, Brooklyn and Clarendon which we use from an open-source python library - Pilgram [21]. Some example images from these transformations are shown in Figure 5. We report the BRA of different watermarking frameworks after undergoing benign transformations in Table 2. We find that both StegaStamp and our robust and semi-fragile models can decode secrets with a high BRA for these image transformations. We find that FaceSigns (Robust), which does not use malicious transforms during training, is slightly more robust to benign transformations as compared to FaceSigns (Semi-Fragile). However, this improved robustness comes at the cost of being non-fragile to malicious transformations and being able to decode messages with high BRA even for Deepfake manipulations. The model FaceSigns (No-Transform) which does not incorporate any benign or malicious transformations during training is fragile to both JPEG compression and malicious transforms as indicated by the low BRA for both methods.

To evaluate fragility of the watermark against unseen facial manipulations, we apply three face-swapping techniques on the watermarked images: FaceSwap [23], SimSwap [6] and Few-Shot Face Translation (FSFT) [39]. FaceSwap [23] is a computer graphics based technique that swaps the face by aligning the facial landmarks of the two images. SimSwap [6] and FSFT [39] are deep learning based techniques that use CNN encoder-decoder networks trained using adversarial loss to generate Deepfakes. As reported in Table 2, we find that StegaStamp and FaceSigns (Robust) can decode signatures from Deepfake images with a high BRA thereby making them unsuitable for authenticating the integrity of digital media. This is understandable since these methods prioritize robustness over fragility. StegaStamp has been shown to be robust to occlusions even though occlusions were not explicitly a part of their set of training transformations. As evident by the ROC plots and AUC scores shown in Figure 7, in contrast to prior works, our semi-fragile model demonstrates robustness to benign transformations while being fragile toward out-of-domain malicious Deepfake transformations, thereby achieving our goal of selective fragility and an AUC score of for manipulation detection.
BRA (%) - Benign Transforms | BRA (%) - Malicious Transforms | ||||||||
---|---|---|---|---|---|---|---|---|---|
Method | None | JPG-75 | JPG-50 | Aden | Brooklyn | Clarendon | SimSwap [6] | FSFT [39] | FaceSwap [23] |
SemiFragile DCT [17] | |||||||||
HiDDeN [54] | |||||||||
StegaStamp [44] | |||||||||
FaceSigns (No Transform) | |||||||||
FaceSigns (Robust) | |||||||||
FaceSigns (Semi-Fragile) |

4.4 Watermarking images with multiple faces
While our networks are trained on the CelebA dataset that has only one face per image, our watermarking framework can also be applied to images containing multiple faces. To achieve this, we use a face detection model to extract a square bounding box of each face in the image containing multiple faces. The faces from the bounding boxes are cropped out and resized to and passed as input to our encoder model to embed individual semi-fragile watermarks. The watermarked faces are then resized back to their original size and placed back into the original images. During decoding, a similar process is repeated where the faces are cropped and resized to before being fed into the decoder. Since the benign transforms during training tolerate small image translations, it makes our watermarking robust to small shifts in the face detection network. We conduct experiments on 400 test images containing faces each from the Celebrity Together dataset [53]. Our FaceSigns (Semi-Fragile) model achieves a BRA of 99.50% demonstrating that we can effectively encode and retrieve watermarks embedded in images with multiple faces. Figure 8 shows watermarked images with multiple identities.

5 Discussion - Threat Models
Both watermark embedding techniques and Deepfake detection systems face adversarial threats from attackers who attempt to bypass the detectors by authenticating manipulated media. In this section, we discuss some of the threat models faced by our system and how these challenges can be addressed:
Attack 1. Querying the decoder network for performing adversarial attacks: The attacker may query the decoder network with an image to get the decoded message and adversarially perturb the query image until the decoded message matches the target message.
Defense:
The attacker does not know what target messages can prove media authenticity since these messages can be kept as a secret and updated frequently. If the attacker gains access to the secret message by querying the decoder with a watermarked image, the encryption key secrecy can prevent the attacker from knowing the target encrypted message for the decoder. Lastly, the decoder network can be hosted securely and can only output a binary label indicating whether the image is authentic or manipulated by matching the decoded secret with the list of trusted secrets.
This would make the decoder’s signal unusable for performing adversarial attacks to match a target message out of the total possible messages.
Attack 2. Copying the watermark perturbation from one image to another: The adversary may attempt to extract the added perturbation of the watermark and add it onto a Deepfake image to authenticate the manipulated media.
Defense:
Since FaceSigns generates an image and message specific perturbation, we hypothesize that the same perturbation when applied on alternate images should not be recoverable by the decoder. We verify this hypothesis by conducting an experiment in which we extract added perturbations from 100 watermarked images, and apply extracted perturbation to 100 alternate images. The bit recovery accuracy of such an attack is just 17.6% which is worse than random prediction.
Attack 3. Training a proxy encoder: The adversary can collect a dataset of original and watermarked images and train a neural network based encoder-decoder image-to-image translation network to map any new image to a watermarked image.
Defense:
One defense strategy is to only store watermarked images on devices so that an attacker never gains access to pairs of original and watermarked images.
Also, the above attack can only work if the encoded images all contain the same secret message, so that the adversary can learn a generator for watermarking a new image with the same secret message. To prevent the creation of such a dataset, some bits of the message can be kept dynamic and contain a unique time-stamp and device specific codes so that each embedded bit-string is different.
Regularly updating the trusted message or encryption key is another preventative strategy against such attacks.
6 Conclusion
We introduce a deep learning based semi-fragile watermarking system that can certify the integrity of digital images and reliably detect facial tampering. Through our experiments and evaluations, we demonstrate that FaceSigns generates more imperceptible watermarks than previous state-of-the-art methods while upholding the desired semi-fragile characteristics. By carefully designing a fixed set of differentiable benign and malicious transformations during training, our framework achieves generalizability to real-world image transformations and can reliably detect Deepfake facial manipulations unlike prior image watermarking techniques. FaceSigns can be vital to media authenticators in social media platforms, news agencies and legal offices and help create more trustworthy platforms and establish consumer trust in digital media.
References
- [1] Afchar, D., Nozick, V., Yamagishi, J., Echizen, I.: Mesonet: a compact facial video forgery detection network. In: 2018 IEEE International Workshop on Information Forensics and Security (WIFS). IEEE (2018)
- [2] Baluja, S.: Hiding images in plain sight: Deep steganography. NIPS (2017)
- [3] Benrhouma, O., Hermassi, H., Belghith, S.: Tamper detection and self-recovery scheme by dwt watermarking. Nonlinear Dynamics (2015)
- [4] Bhalerao, S., Ansari, I.A., Kumar, A.: A secure image watermarking for tamper detection and localization. Journal of Ambient Intelligence and Humanized Computing (2021)
- [5] Bi, N., Sun, Q., Huang, D., Yang, Z., Huang, J.: Robust image watermarking based on multiband wavelets and empirical mode decomposition. IEEE Transactions on Image Processing (2007)
- [6] Chen, R., Chen, X., Ni, B., Ge, Y.: Simswap: An efficient framework for high fidelity face swapping. In: MM ’20: The 28th ACM International Conference on Multimedia. ACM (2020)
- [7] Choi, Y., Uh, Y., Yoo, J., Ha, J.W.: Stargan v2: Diverse image synthesis for multiple domains. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2020)
- [8] Cox, I.J., Kilian, J., Leighton, F.T., Shamoon, T.: Secure spread spectrum watermarking for multimedia. IEEE transactions on image processing (1997)
- [9] Cox, I.J., Miller, M.L., Bloom, J.A., Honsinger, C.: Digital watermarking, vol. 53. Springer (2002)
- [10] DeepFakes: https://github.com/deepfakes/faceswap (2017)
- [11] Di Martino, F., Sessa, S.: Fragile watermarking tamper detection via bilinear fuzzy relation equations. Journal of Ambient Intelligence and Humanized Computing (2019)
- [12] Dolhansky, B., Bitton, J., Pflaum, B., Lu, J., Howes, R., Wang, M., Ferrer, C.C.: The DeepFake Detection Challenge (DFDC) dataset. arXiv preprint arXiv:2006.07397 (2020)
- [13] Dolhansky, B., Howes, R., Pflaum, B., Baram, N., Ferrer, C.C.: The deepfake detection challenge (dfdc) dataset. arXiv preprint arXiv:2006.07397 (2020)
- [14] Fridrich, J.: Steganography in digital media: principles, algorithms, and applications. Cambridge University Press (2009)
- [15] Gandhi, A., Jain, S.: Adversarial perturbations fool deepfake detectors. In: 2020 International Joint Conference on Neural Networks (IJCNN). IEEE (2020)
- [16] Hayes, J., Danezis, G.: Generating steganographic images via adversarial training. In: International Conference on Neural Information Processing Systems (2017)
- [17] Ho, C.K., Li, C.T.: Semi-fragile watermarking scheme for authentication of jpeg images. In: International Conference on Information Technology: Coding and Computing, 2004. Proceedings. ITCC 2004. IEEE (2004)
- [18] Hussain, S., Neekhara, P., Dolhansky, B., Bitton, J., Canton Ferrer, C., McAuley, J., Koushanfar, F.: Exposing vulnerabilities of deepfake detection systems with robust attacks. ACM Journal of Digital Threats: Research and Practice . https://doi.org/10.1145/3464307
- [19] Hussain, S., Neekhara, P., Jere, M., Koushanfar, F., McAuley, J.: Adversarial deepfakes: Evaluating vulnerability of deepfake detectors to adversarial examples. In: WACV (2021)
- [20] Isola, P., Zhu, J.Y., Zhou, T., Efros, A.A.: Image-to-image translation with conditional adversarial networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition (2017)
- [21] Kamakura, A.: pilgram https://github.com/akiomik/pilgram
- [22] Karras, T., Laine, S., Aila, T.: A style-based generator architecture for generative adversarial networks. In: CVPR (2019)
- [23] Kowalski, M.: Faceswap https://github.com/marekkowalski/faceswap/ (2018)
- [24] Li, C., Zhang, A., Liu, Z., Liao, L., Huang, D.: Semi-fragile self-recoverable watermarking algorithm based on wavelet group quantization and double authentication. Multimedia tools and applications (2015)
- [25] Lin, E.T., Podilchuk, C.I., Delp III, E.J.: Detection of image alterations using semifragile watermarks. In: Security and Watermarking of Multimedia Contents II. International Society for Optics and Photonics (2000)
- [26] Liu, Z., Luo, P., Wang, X., Tang, X.: Deep learning face attributes in the wild. In: ICCV (December 2015)
- [27] Luo, X., Zhan, R., Chang, H., Yang, F., Milanfar, P.: Distortion agnostic deep watermarking. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (2020)
- [28] Lyu, S.: Deepfake detection: Current challenges and next steps. In: 2020 IEEE International Conference on Multimedia Expo Workshops (ICMEW) (2020)
- [29] Mirsky, Y., Lee, W.: The creation and detection of deepfakes: A survey. ACM Comput. Surv. 54(1) (jan 2021). https://doi.org/10.1145/3425780, https://doi.org/10.1145/3425780
- [30] Neekhara, P., Dolhansky, B., Bitton, J., Ferrer, C.C.: Adversarial threats to deepfake detection: A practical perspective. In: 2021 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW). IEEE (2021)
- [31] Nirkin, Y., Keller, Y., Hassner, T.: FSGAN: Subject agnostic face swapping and reenactment. In: CVPR (2019)
- [32] Odena, A., Dumoulin, V., Olah, C.: Deconvolution and checkerboard artifacts. Distill (2016). https://doi.org/10.23915/distill.00003, http://distill.pub/2016/deconv-checkerboard
- [33] Pereira, S., Pun, T.: Robust template matching for affine resistant image watermarks. IEEE transactions on image Processing (2000)
- [34] Pereira, S., Ruanaidh, J.J.O., Deguillaume, F., Csurka, G., Pun, T.: Template based recovery of fourier-based watermarks using log-polar and log-log maps. In: Proceedings IEEE International Conference on Multimedia Computing and Systems. IEEE (1999)
- [35] Preda, R., Vizireanu, D.: Watermarking-based image authentication robust to jpeg compression. Electronics Letters (2015)
- [36] Ronneberger, O., Fischer, P., Brox, T.: U-net: Convolutional networks for biomedical image segmentation. In: International Conference on Medical image computing and computer-assisted intervention. Springer (2015)
- [37] Rossler, A., Cozzolino, D., Verdoliva, L., Riess, C., Thies, J., Niessner, M.: Faceforensics++: Learning to detect manipulated facial images. In: ICCV (October 2019)
- [38] Seferbekov, S.: https://github.com/selimsef/dfdc_deepfake-_challenge (2020)
- [39] Shaoanlu: Few-shot face translation https://github.com/shaoanlu/fewshot-face-translation-gan (2019)
- [40] Shefali, S., Deshpande, S.: Information security through semi-fragile watermarking. In: International Conference on Computational Intelligence and Multimedia Applications (ICCIMA 2007). IEEE (2007)
- [41] Shehab, A., Elhoseny, M., Muhammad, K., Sangaiah, A.K., Yang, P., Huang, H., Hou, G.: Secure and robust fragile watermarking scheme for medical images. IEEE Access (2018)
- [42] Shin, R., Song, D.: Jpeg-resistant adversarial images. In: NIPS 2017 Workshop on Machine Learning and Computer Security (2017)
- [43] Sun, R., Sun, H., Yao, T.: A svd-and quantization based semi-fragile watermarking technique for image authentication. In: 6th International Conference on Signal Processing, 2002. IEEE (2002)
- [44] Tancik, M., Mildenhall, B., Ng, R.: Stegastamp: Invisible hyperlinks in physical photographs. In: CVPR (2020)
- [45] Thies, J., Zollhofer, M., Stamminger, M., Theobalt, C., Nießner, M.: Face2face: Real-time face capture and reenactment of rgb videos. In: CVPR (2016)
- [46] Wang, S.Y., Wang, O., Zhang, R., Owens, A., Efros, A.A.: Cnn-generated images are surprisingly easy to spot… for now. In: CVPR (2020)
- [47] Xiao, J., Wang, Y.: A semi-fragile watermarking tolerant of laplacian sharpening. In: 2008 International Conference on Computer Science and Software Engineering. IEEE (2008)
- [48] Yang, H., Sun, X., Sun, G.: A semi-fragile watermarking algorithm using adaptive least significant bit substitution. Information Technology Journal (2009)
- [49] Yu, X., Wang, C., Zhou, X.: Review on semi-fragile watermarking algorithms for content authentication of digital images. Future Internet (2017)
- [50] Zakharov, E., Shysheya, A., Burkov, E., Lempitsky, V.: Few-shot adversarial learning of realistic neural talking head models. In: ICCV (2019)
- [51] Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: CVPR (2018)
- [52] Zhang, R., Dong, S., Liu, J.: Invisible steganography via generative adversarial networks. Multimedia tools and applications (2019)
- [53] Zhong, Y., Arandjelović, R., Zisserman, A.: Compact deep aggregation for set retrieval. In: Workshop on Compact and Efficient Feature Representation and Learning in Computer Vision, ECCV (2018)
- [54] Zhu, J., Kaplan, R., Johnson, J., Fei-Fei, L.: Hidden: Hiding data with deep networks. In: ECCV (2018)