Attributing Fake Images to GANs:
Learning and Analyzing GAN Fingerprints

ICCV 2019


Ning Yu1,2      Larry Davis1      Mario Fritz3
1. University of Maryland      2. Max Planck Institute for Informatics      3. CISPA Helmholtz Center for Information Security

Abstract


Recent advances in Generative Adversarial Networks (GANs) have shown increasing success in generating photorealistic images. But they also raise challenges to visual forensics and model attribution. We present the first study of learning GAN fingerprints towards image attribution and using them to classify an image as real or GAN-generated. For GAN-generated images, we further identify their sources. Our experiments show that (1) GANs carry distinct model fingerprints and leave stable fingerprints in their generated images, which support image attribution; (2) even minor differences in GAN training can result in different fingerprints, which enables fine-grained model authentication; (3) fingerprints persist across different image frequencies and patches and are not biased by GAN artifacts; (4) fingerprint finetuning is effective in immunizing against five types of adversarial image perturbations; and (5) comparisons also show our learned fingerprints consistently outperform several baselines in a variety of setups.

Demos

Materials




Paper



Poster

Code

Press coverage


thejiangmen Academia News

Citation

@inproceedings{yu2019attributing,
  author = {Yu, Ning and Davis, Larry and Fritz, Mario},
  title = {Attributing Fake Images to GANs: Learning and Analyzing GAN Fingerprints},
  booktitle = {IEEE International Conference on Computer Vision (ICCV)},
  year = {2019}
}

Acknowledgement


We thank Hao Zhou for helping with the relighting experiments. We also thank Yaser Yacoob and Abhinav Shrivastava for constructive advice. This research is partially funded by DARPA MediFor program under cooperative agreement FA87501620191. Any opinions, findings, conclusions, or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the DARPA.

Related Work


F. Marra, D. Gragnaniello, L. Verdoliva, G. Poggi. Do GANs leave artificial fingerprints? MIPR 2019.
Comment: A GAN detection baseline method that extracts shallow fingerprints to detect GAN-generated images.
T. Karras, T. Aila, S. Laine, J. Lehtinen. Progressive Growing of Gans for Improved Quality, Stability, and Variation. ICLR 2018.
Comment: A state-of-the-art GAN source for our generated image detection and attribution.
T. Miyato, T. Kataoka, M. Koyama, Y. Yoshida. Spectral Normalization for Generative Adversarial Networks. ICLR 2018.
Comment: A state-of-the-art GAN source for our generated image detection and attribution.
M. Bellemare, I. Danihelka, W. Dabney, S. Mohamed, B. Lakshminarayanan, S. Hoyer, R. Munos. The Cramer Distance as a Solution to Biased Wasserstein Gradients. arXiv 2017.
Comment: A state-of-the-art GAN source for our generated image detection and attribution.
C.L. Li, W.C. Chang, Y. Cheng, Y. Yang, B. Póczos. MMD GAN: Towards Deeper Understanding of Moment Matching Network. NeurIPS 2017.
Comment: A state-of-the-art GAN source for our generated image detection and attribution.