
tuning of the target model did not lead to significantly
worse performance of the MIA. It should however be
noted that this was only tested by reducing the share
of target members to 1:2.
It is found that by introducing visible watermarks to
the target dataset, our MIA sees a significant boost
in performance. Using hidden watermarks was not
found to have a positive impact on the performance of
the MIA. No significant effect was found when inves-
tigating the influence of the relationship between the
labels used for fine-tuning the target model and the
prompt used for inference, i.e. whether they match or
not. Upon investigating the importance of the guid-
ance scale used by the target model, it is found to
have a significant influence on the performance of our
MIA, with best performance at s ∼ 8.
Overall the proposed MIA is a realistic and feasi-
ble attack in a real-life application. However, it is
computationally expensive to fine-tune a generative
”shadow model” for the task of producing an auxil-
iary dataset related to the domain of interest as well
as training the Resnet-18 attack model. The nature
of the tests performed restricts our conclusion to the
case of LDMs fine-tuned on face images. The small-
est amount of fine-tuning that was still found to be
effective was 50 epochs on the member images (how-
ever it could be lower - as it was not tested). The
only LDM used for testing was Stable-Diffusion-v1.5
(Rombach et al., 2022), which limits the generaliz-
ability of the conclusions drawn. The approach using
a Resnet-18 as an attack model is found to be gener-
ally stable on several different hyperparameters in the
target LDM. In conclusion, the method for Member-
ship Inference Attack shown in this paper is realistic
and could be used as a tool to infer if one’s face im-
ages have been used to fine-tune a Latent Diffusion
Model in a black-box setup.
REFERENCES
Carlini, N., Hayes, J., Nasr, M., Jagielski, M., Sehwag, V.,
Tram
`
er, F., Balle, B., Ippolito, D., and Wallace, E.
(2023). Extracting training data from diffusion mod-
els. ArXiv, abs/2301.13188.
Chen, D., Yu, N., Zhang, Y., and Fritz, M. (2020). Gan-
leaks: A taxonomy of membership inference attacks
against generative models. In Proceedings of the 2020
ACM SIGSAC Conference on Computer and Commu-
nications Security, CCS ’20. ACM.
Dubi’nski, J., Kowalczuk, A. D., Pawlak, S., Rokita, P.,
Trzci’nski, T., and Morawiecki, P. (2023). Towards
more realistic membership inference attacks on large
diffusion models. 2024 IEEE/CVF Winter Conference
on Applications of Computer Vision (WACV), pages
4848–4857.
He, K., Zhang, X., Ren, S., and Sun, J. (2015). Deep
residual learning for image recognition. CoRR,
abs/1512.03385.
Huang, G. B., Ramesh, M., Berg, T., and Learned-Miller,
E. (2007). Labeled faces in the wild: A database for
studying face recognition in unconstrained environ-
ments. Technical Report 07-49, University of Mas-
sachusetts, Amherst.
Li, J., Li, D., Xiong, C., and Hoi, S. (2022). Blip:
Bootstrapping language-image pre-training for unified
vision-language understanding and generation.
Radford, A., Kim, J. W., Hallacy, C., Ramesh, A., Goh, G.,
Agarwal, S., Sastry, G., Askell, A., Mishkin, P., Clark,
J., Krueger, G., and Sutskever, I. (2021). Learning
transferable visual models from natural language su-
pervision. CoRR, abs/2103.00020.
Rombach, R., Blattmann, A., Lorenz, D., Esser, P., and Om-
mer, B. (2022). High-resolution image synthesis with
latent diffusion models.
Shokri, R., Stronati, M., Song, C., and Shmatikov, V.
(2017). Membership inference attacks against ma-
chine learning models.
Zhang, M., Yu, N., Wen, R., Backes, M., and Zhang, Y.
(2023). Generated distributions are all you need for
membership inference attacks against generative mod-
els.
VISAPP 2025 - 20th International Conference on Computer Vision Theory and Applications
446