Professional Documents
Culture Documents
In this paper we investigate on this interesting issue, III. E XPOSING GAN FINGERPRINTS
and provide a first answer to the above questions. Our In this Section we show evidence on the existence of GAN
experiments with several popular GAN architectures and fingerprints. This goal is pursued in a minimal experimental
datasets, show that GAN do leave specific fingerprints on setting, considering only two GANs, a Cycle-GAN trained to
the image they generate, which can be used to carry out convert orange images into apple images and a Progressive-
reliable forensic analyses. GAN (Pro-GAN) trained to generate kitchen images, call
them GAN A and B, from now on. Lacking any statistical
model, we consider an extraction pipeline similar to that
II. R ELATED W ORK
of the PRNU pattern. For the generic image Xi generated
by a given GAN, the fingerprint represents a disturbance,
Recently there has been a growing interest in distinguish- unrelated with the image semantics. Therefore, we first
ing GAN-generated images from real ones. As shown in estimate the high-level image content, Xi = f (Xi ), through
Fig.1, the current state of the art in GANs is far from a suitable denoising filter f (·), then subtract it from the
perfection, and often generated images exhibit strong visual original image to extract the noise residual
artifacts that can be exploited for forensic use. For example,
to detect fake faces, [14] exploits visual features regarding Ri = Xi − f (Xi ) (1)
eyes, teeth and facial contours. Tellingly, the authors observe
that in GAN-generated images the color of left and right eye Then, we assume the residual to be the sum of a non-zero
are often inconsistent. Color information is also used in [15], deterministic component, the fingerprint F , and a random
[16]. In particular, [15] proposes to use some features shared noise component Wi
by different GAN architectures, related to the synthesis of R i = F + Wi (2)
RGB color channels.
Other methods rely on deep learning. Several architectures Accordingly, the fingerprint is estimated by a simple average
have been tested so far [17]–[19] showing a good accuracy over the available residuals
N
in detecting GAN-generated images, even on compressed 1
images. Unfortunately, if a network is trained on a specific F = Ri (3)
N i=1
architecture, its performance degrades sharply when used
to detect image generated by another architecture [20]. Fig.2 shows (suitably amplified) the fingerprints of the
This observation suggests the presence of different artifacts two GANs, estimated over a growing number of residuals,
peculiar of each specific GAN model. N = 2, 8, 32, 128, 512. Of course, for low values on N , the
Recently, it has also been shown [21] that a deep network estimates are dominated by image-related noise. However, as
can reliably discriminate images generated with different N grows, the additive noise component tends to vanish and
architectures. However, the network requires intensive train- both estimates converge to stable quasi-periodical patterns,
ing on an aligned dataset, and there is no hint, let alone which we regard as accurate approximations of the true GAN
exploitation, of the presence of GAN-induced fingerprints. fingerprints. In Fig.3 we show the energy E(N ) of these
507
True energy and fitting curve for the Cycle-GAN and
Figure 3. Figure 5. Correlation of Cycle-GAN (left) and Pro-GAN (right)
Pro-GAN fingerprints of Figure 2. residuals with same/cross-GAN fingerprints.
508
Table I
C AMERAS AND GAN S USED IN THE EXPERIMENTS
509
Figure 8. Source identification confusion matrix. Entries below 1% are canceled.
510
VI. ACKNOWLEDGEMENT [11] J. Lukàš, J. Fridrich, and M. Goljan, “Digital camera iden-
tification from sensor pattern noise,” IEEE Transactions on
This material is based on research sponsored by the Air Information Forensics and Security, vol. 1, no. 2, pp. 205–
Force Research Laboratory and the Defense Advanced Re- 214, 2006.
search Projects Agency under agreement number FA8750-
16-2-0204. The U.S. Government is authorized to reproduce [12] D. Cozzolino and L. Verdoliva, “Noiseprint: a CNN-based
and distribute reprints for Governmental purposes notwith- camera model fingerprint,” arXiv preprint arXiv:1808.08396,
2018.
standing any copyright notation thereon. The views and con-
clusions contained herein are those of the authors and should [13] M. Chen, J. Fridrich, M. Goljan, and J. Lukàš, “Determining
not be interpreted as necessarily representing the official image origin and integrity using sensor noise,” IEEE Trans-
policies or endorsements, either expressed or implied, of the actions on Information Forensics and Security, vol. 3, pp.
Air Force Research Laboratory and the Defense Advanced 74–90, 2008.
Research Projects Agency or the U.S. Government. [14] F. Matern, C. Riess, and M. Stamminger, “Exploiting visual
R EFERENCES artifacts to expose deepfakes and face manipulations,” in
IEEE WACV Workshop on Image and Video Forensics, 2019.
[1] T. Salimans, I. Goodfellow, W. Zaremba, V. Cheung, A. Rad-
ford, and X. Chen, “Improved Techniques for Training [15] S. McCloskey and M. Albright, “Detecting GAN-
GANs,” in Advances in Neural Information Processing Sys- generated Imagery using Color Cues,” arXiv preprint
tems, 2016, pp. 2234–2242. arXiv:1812.08247v1, 2018.
[2] I. Gulrajani, F. Ahmed, M. Arjovsky, V. Dumoulin, and [16] H. Li, B. Li, S. Tan, and J. Huang, “Detection of deep network
A. Courville, “Improved Training of Wasserstein GANs,” in generated images using disparities in color components,”
Advances in Neural Information Processing Systems, 2017, arXiv preprint arXiv:1808.07276v1, 2018.
pp. 5767–5777.
[17] F. Marra, D. Gragnaniello, D. Cozzolino, and L. Verdo-
[3] D. Berthelot, T. Schumm, and L. Metz, “BEGAN: Bound- liva, “Detection of GAN-Generated Fake Images over Social
ary Equilibrium Generative Adversarial Networks,” arXiv Networks,” in IEEE Conference on Multimedia Information
preprint: abs/1703.10717, 2017. Processing and Retrieval, April 2018, pp. 384–389.
[4] J. Thies, M. Zollhöfer, M. Stamminger, C. Theobalt, and [18] H. Mo, B. Chen, and W. Luo, “Fake Faces Identification via
M. Nießner, “Face2Face: Real-Time Face Capture and Reen- Convolutional Neural Network,” in Proc. of the 6th ACM
actment of RGB Videos,” in IEEE Conference on Computer Workshop on Information Hiding and Multimedia Security,
Vision and Pattern Recognition, 2016, pp. 2387–2395. June 2018.
[5] P. Isola, J.-Y. Zhu, T. Zhou, and A. A. Efros, “Image-to-image [19] S. Tariq, S. Lee, H. Kim, Y. Shin, and S. Woo, “Detecting both
translation with conditional adversarial networks,” in IEEE machine and human created fake face images in the wild,”
Conference on Computer Vision and Pattern Recognition, in Proc. of the 2nd International Workshop on Multimedia
2017. Privacy and Security, 2018, pp. 81–87.
[6] J. Y. Zhu, T. Park, P. Isola, and A. A. Efros, “Unpaired [20] D. Cozzolino, J. Thies, A. Rössler, C. Riess, M. Nießner,
image-to-image translation using cycle-consistent adversarial and L. Verdoliva, “ForensicTransfer: Weakly-supervised Do-
networks,” in IEEE International Conference on Computer main Adaptation for Forgery Detection,” arXiv preprint
Vision, 2017. arXiv:1812.02510, 2018.
[7] M.-Y. Liu, T. Breuel, and J. Kautz, “Unsupervised image-to-
[21] N. Yu, L. Davis, and M. Fritz, “Attributing Fake Images to
image translation networks,” in Neural Information Process-
GANs: Analyzing Fingerprints in Generated Images,” arXiv
ing Systems, 2017.
preprint arXiv:1811.08180v1, 2018.
[8] Y. Choi, M. Choi, M. Kim, J.-W. Ha, S. Kim, and J. Choo,
“StarGAN: Unified Generative Adversarial Networks for [22] D. Dang-Nguyen, C. Pasquini, V. Conotter, and G. Boato,
Multi-Domain Image-to-Image Translation,” in IEEE Confer- “RAISE: a raw images dataset for digital image forensics,” in
ence on Computer Vision and Pattern Recognition, 2018. Proc. of the 6th ACM Multimedia Systems Conference, 2015,
pp. 219–1224.
[9] T. Karras, T. Aila, S. Laine, and J. Lehtinen, “Progressive
Growing of GANs for Improved Quality, Stability, and Vari- [23] National Institute of Standards and Technology, “Media
ation,” in International Conference on Learning Representa- Forensics Challenge,” https://www.nist.gov/itl/iad/mig/media-
tions, 2018. forensics-challenge-2018, 2018.
511