With today's technologies, the possibility to generate highly realistic visual fakes is within everyone's reach, leading to major threats in terms of misinformation and data trustworthiness. This holds in particular for synthetically generated faces, which are able to deceive even the most experienced observers, and can be exploited to create fake digital identities with synthetic facial attributes, to be used on social networks and online services. In response to this threat, researchers have employed artificial intelligence to detect synthetic images by analysing patterns and artifacts introduced by the generative models. However, most online images are subject to repeated sharing operations by social media platforms. Said platforms process uploaded images by applying operations (like compression) that progressively degrade those useful forensic traces, compromising the effectiveness of the developed detectors. To solve the synthetic-vs-real problem “in the wild”, more realistic image databases are needed to train specialised detectors. In this work, we present TrueFace, a first dataset of social-media-processed real and synthetic faces, obtained by the successful StyleGAN generative models, and shared on Facebook, Twitter and Telegram. The dataset is used to validate a ResNet-based image classification model addressing the discrimination of synthetic-vs-real faces in both presocial and post-social scenarios. The results demonstrate that even detectors with extremely high performance on non-shared images struggle to retain their accuracy on images from social media, while fine-tuning with shared images strongly mitigates such performance issues.
TrueFace: a Dataset for the Detection of Synthetic Face Images from Social Networks
Pasquini, C.;
2022-01-01
Abstract
With today's technologies, the possibility to generate highly realistic visual fakes is within everyone's reach, leading to major threats in terms of misinformation and data trustworthiness. This holds in particular for synthetically generated faces, which are able to deceive even the most experienced observers, and can be exploited to create fake digital identities with synthetic facial attributes, to be used on social networks and online services. In response to this threat, researchers have employed artificial intelligence to detect synthetic images by analysing patterns and artifacts introduced by the generative models. However, most online images are subject to repeated sharing operations by social media platforms. Said platforms process uploaded images by applying operations (like compression) that progressively degrade those useful forensic traces, compromising the effectiveness of the developed detectors. To solve the synthetic-vs-real problem “in the wild”, more realistic image databases are needed to train specialised detectors. In this work, we present TrueFace, a first dataset of social-media-processed real and synthetic faces, obtained by the successful StyleGAN generative models, and shared on Facebook, Twitter and Telegram. The dataset is used to validate a ResNet-based image classification model addressing the discrimination of synthetic-vs-real faces in both presocial and post-social scenarios. The results demonstrate that even detectors with extremely high performance on non-shared images struggle to retain their accuracy on images from social media, while fine-tuning with shared images strongly mitigates such performance issues.File | Dimensione | Formato | |
---|---|---|---|
IJCB2022.pdf
solo utenti autorizzati
Tipologia:
Documento in Post-print
Licenza:
NON PUBBLICO - Accesso privato/ristretto
Dimensione
402.1 kB
Formato
Adobe PDF
|
402.1 kB | Adobe PDF | Visualizza/Apri Richiedi una copia |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.