TY - JOUR
T1 - Deepfake attribution
T2 - On the source identification of artificially generated images
AU - Khoo, Brandon
AU - Phan, Raphaël C.W.
AU - Lim, Chern-Hong
N1 - Publisher Copyright:
© 2021 Wiley Periodicals LLC.
PY - 2022/5
Y1 - 2022/5
N2 - Synthetic media or "deepfakes" are making great advances in visual quality, diversity, and verisimilitude, empowered by large-scale publicly accessible datasets and rapid technical progress in deep generative modeling. Heralding a paradigm shift in how online content is trusted, researchers in digital image forensics have responded with different proposals to reliably detect AI-generated images in the wild. However, binary classification of image authenticity is insufficient to regulate the ethical usage of deepfake technology as new applications are developed. This article provides an overview of the major innovations in synthetic forgery detection as of 2020, while highlighting the recent shift in research towards ways to attribute AI-generated images to their generative sources with evidence. We define the various categories of deepfakes in existence, the subtle processing traces and fingerprints that distinguish AI-generated images from reality and each other, and the different degrees of attribution possible with current understanding of generative algorithms. Additionally, we describe the limitations of synthetic image recognition methods in practice, the counter-forensic attacks devised to exploit these limitations, and directions for new research to assure the long-term relevance of deepfake forensics. Reliable, explainable, and generalizable attribution methods would hold malicious users accountable for AI-enabled disinformation, grant plausible deniability to appropriate users, and facilitate intellectual property protection of deepfake technology.
AB - Synthetic media or "deepfakes" are making great advances in visual quality, diversity, and verisimilitude, empowered by large-scale publicly accessible datasets and rapid technical progress in deep generative modeling. Heralding a paradigm shift in how online content is trusted, researchers in digital image forensics have responded with different proposals to reliably detect AI-generated images in the wild. However, binary classification of image authenticity is insufficient to regulate the ethical usage of deepfake technology as new applications are developed. This article provides an overview of the major innovations in synthetic forgery detection as of 2020, while highlighting the recent shift in research towards ways to attribute AI-generated images to their generative sources with evidence. We define the various categories of deepfakes in existence, the subtle processing traces and fingerprints that distinguish AI-generated images from reality and each other, and the different degrees of attribution possible with current understanding of generative algorithms. Additionally, we describe the limitations of synthetic image recognition methods in practice, the counter-forensic attacks devised to exploit these limitations, and directions for new research to assure the long-term relevance of deepfake forensics. Reliable, explainable, and generalizable attribution methods would hold malicious users accountable for AI-enabled disinformation, grant plausible deniability to appropriate users, and facilitate intellectual property protection of deepfake technology.
KW - computer vision
KW - deepfakes
KW - machine learning
KW - source identification
KW - synthetic media
UR - http://www.scopus.com/inward/record.url?scp=85120417867&partnerID=8YFLogxK
U2 - 10.1002/widm.1438
DO - 10.1002/widm.1438
M3 - Review Article
AN - SCOPUS:85120417867
SN - 1942-4795
VL - 12
JO - WIREs Data Mining and Knowledge Discovery
JF - WIREs Data Mining and Knowledge Discovery
IS - 3
M1 - e1438
ER -