other one with synthetic images). The clearly best
classification rates were achieved using only real-
world data (91%), followed by using only synthetic
data (81-83%), and the worst results were achieved
using mixed data (50-76%). A comparison of the fre-
quency spectra of the synthetic images showed that
each generator exhibits unique model-specific char-
acteristics (a model-specific fingerprint). It’s likely
that this fingerprint is one of the reasons for the de-
graded performance of the classifiers trained on syn-
thetic data and, even more so, on mixed data. In any
case, the experiments clearly demonstrate that, de-
spite substantial developments in the field of synthe-
sis, synthetic training data should always be used with
caution, especially when synthetic data is used to re-
place only data of a single class. There is always the
risk that a deep learning based classifier learns to dif-
ferentiate based on model-specific characteristics and
not, as intended, based on class-specific features.
The employed FP removal techniques did not
work as intended and were unable to bridge the gap
between real and synthetic data. In future work,
we therefore plan to develop better FP removal tech-
niques.
ACKNOWLEDGEMENTS
This work has been partially supported by the
Salzburg State Government within the Science and In-
novation Strategy Salzburg 2025 (WISS 2025) under
the project AIIV Salzburg (Artificial Intelligence in
Industrial Vision), project no 20102-F2100737- FPR.
REFERENCES
Carlini, N., Hayes, J., Nasr, M., Jagielski, M., Sehwag, V.,
Tramèr, F., Balle, B., Ippolito, D., and Wallace, E.
(2023). Extracting training data from diffusion mod-
els.
Chen, Y., Yang, X.-H., Wei, Z., Heidari, A. A., Zheng, N.,
Li, Z., Chen, H., Hu, H., Zhou, Q., and Guan, Q.
(2022). Generative adversarial networks in medical
image augmentation: A review. Computers in Biology
and Medicine, 144:105382.
Corvi, R., Cozzolino, D., Poggi, G., Nagano, K., and Ver-
doliva, L. (2023). Intriguing properties of synthetic
images: from generative adversarial networks to dif-
fusion models.
Dhariwal, P. and Nichol, A. (2021). Diffusion models beat
gans on image synthesis. Advances in neural informa-
tion processing systems, 34:8780–8794.
Elreedy, D. and Atiya, A. F. (2019). A comprehensive
analysis of synthetic minority oversampling technique
(smote) for handling class imbalance. Information
Sciences, 505:32–64.
Frank, J., Eisenhofer, T., Schönherr, L., Fischer, A.,
Kolossa, D., and Holz, T. (2020). Leveraging fre-
quency analysis for deep fake image recognition. In
International conference on machine learning, pages
3247–3258. PMLR.
Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., and
Hochreiter, S. (2017). Gans trained by a two time-
scale update rule converge to a local nash equilibrium.
Advances in neural information processing systems,
30.
Ho, J., Jain, A., and Abbeel, P. (2020). Denoising diffusion
probabilistic models. Advances in neural information
processing systems, 33:6840–6851.
Li, G., Ji, Z., Qu, X., Zhou, R., and Cao, D. (2022). Cross-
domain object detection for autonomous driving: A
stepwise domain adaptative yolo approach. IEEE
Transactions on Intelligent Vehicles, 7(3):603–615.
Man, K. and Chahl, J. (2022). A review of synthetic im-
age data and its use in computer vision. Journal of
Imaging, 8(11).
Qiu, H., Yu, B., Gong, D., Li, Z., Liu, W., and Tao, D.
(2021). Synface: Face recognition with synthetic data.
In Proceedings of the IEEE/CVF International Con-
ference on Computer Vision, pages 10880–10890.
Saharia, C., Chan, W., Chang, H., Lee, C., Ho, J., Sali-
mans, T., Fleet, D., and Norouzi, M. (2022a). Palette:
Image-to-image diffusion models. In ACM SIG-
GRAPH 2022 Conference Proceedings, pages 1–10.
Saharia, C., Ho, J., Chan, W., Salimans, T., Fleet, D. J., and
Norouzi, M. (2022b). Image super-resolution via iter-
ative refinement. IEEE Transactions on Pattern Anal-
ysis and Machine Intelligence, 45(4):4713–4726.
Song, Y., Sohl-Dickstein, J., Kingma, D. P., Kumar, A.,
Ermon, S., and Poole, B. (2020). Score-based gen-
erative modeling through stochastic differential equa-
tions. arXiv preprint arXiv:2011.13456.
Tan, M. and Le, Q. (2019). Efficientnet: Rethinking model
scaling for convolutional neural networks. In Interna-
tional conference on machine learning, pages 6105–
6114. PMLR.
Torfi, A., Fox, E. A., and Reddy, C. K. (2022). Differen-
tially private synthetic medical data generation using
convolutional gans. Information Sciences, 586:485–
500.
Wesselkamp, V., Rieck, K., Arp, D., and Quiring, E. (2022).
Misleading deep-fake detection with gan fingerprints.
In 2022 IEEE Security and Privacy Workshops (SPW),
pages 59–65, Los Alamitos, CA, USA. IEEE Com-
puter Society.
Zhang, H., Grimmer, M., Ramachandra, R., Raja, K., and
Busch, C. (2021). On the applicability of synthetic
data for face recognition. In 2021 IEEE International
Workshop on Biometrics and Forensics (IWBF), pages
1–6. IEEE.
Zhu, J., Zhang, R., Pathak, D., Darrell, T., Efros,
A. A., Wang, O., and Shechtman, E. (2017). To-
ward multimodal image-to-image translation. CoRR,
abs/1711.11586.
VISAPP 2024 - 19th International Conference on Computer Vision Theory and Applications
206