
Zhang, 2021), where a “boundary attack” for mem-
bership inference is presented, which bypasses the
creation and usage of shadow models altogether.
Also, our defense mechanism could be transferred
to a hybrid scenario. Instead of using a static adver-
sary during the adversarial training of a target model,
the adversary is retrained on the modified output of
the target model in every epoch of the target model
training, as in (Grari et al., 2020). Such a hybrid
adversarial training could potentially further reduce
the leaked property information of a target model, al-
though further investigation is necessary.
Last but not least, the adversarial training for fair-
ness in (Grari et al., 2020) could have side effects
on the success rate of PIAs. Since Grari et al. train
their models to yield outputs independent from a sen-
sitive property p, it would be interesting to investi-
gate whether their approach could also defend a target
model against a PIA focussing on property p.
7 CONCLUSION
In this work, we have expanded upon existing black-
box PIAs by using an arbitrary attack dataset, which
can be based on other datasets than the training
dataset. As the natural fit for many ratio-based prop-
erties, we have modeled the PIAs in this work as
regression problems. We have explored a defense
mechanism based on adversarial training which hard-
ens a target model against black-box PIAs during its
training process. We have evaluated our approach on
three datasets, comparing the attack against white-box
benchmarks and related work. In our experiments, we
have shown our defense scheme to be both effective
(by decreasing the adversary’s performance from an
R² of 0.63–0.64 to 0.07) and practical, decreasing the
mean accuracy of target models by less than 0.2 per-
centage points.
ACKNOWLEDGEMENTS
The authors would like to thank Jona Zantz for his
helpful comments and insights.
REFERENCES
Al-Rubaie, M. and Chang, J. M. (2019). Privacy-preserving
machine learning: Threats and solutions. IEEE Secu-
rity & Privacy, 17(2):49–58.
Alex, K. (2009). Learning multiple layers of fea-
tures from tiny images. https://www. cs. toronto.
edu/kriz/learning-features-2009-TR. pdf.
Ateniese, G., Mancini, L. V., Spognardi, A., Villani, A.,
Vitali, D., and Felici, G. (2015). Hacking smart ma-
chines with smarter ones: How to extract meaningful
data from machine learning classifiers. International
Journal of Security and Networks, 10(3):137–150.
Ganju, K., Wang, Q., Yang, W., Gunter, C. A., and Borisov,
N. (2018). Property inference attacks on fully con-
nected neural networks using permutation invariant
representations. In Proceedings of the 2018 ACM
SIGSAC conference on computer and communications
security, pages 619–633.
Grari, V., Ruf, B., Lamprier, S., and Detyniecki, M. (2020).
Achieving fairness with decision trees: An adversarial
approach. Data Science and Engineering, 5(2):99–
110.
Hartmann, V., Meynent, L., Peyrard, M., Dimitriadis, D.,
Tople, S., and West, R. (2023). Distribution inference
risks: Identifying and mitigating sources of leakage.
In 2023 IEEE Conference on Secure and Trustworthy
Machine Learning (SaTML), pages 136–149. IEEE.
Huang, G. B., Mattar, M., Berg, T., and Learned-Miller,
E. (2008). Labeled faces in the wild: A database
forstudying face recognition in unconstrained envi-
ronments. In Workshop on faces in’Real-Life’Images:
detection, alignment, and recognition.
Kohavi, R. et al. (1996). Scaling up the accuracy of naive-
bayes classifiers: A decision-tree hybrid. In Kdd, vol-
ume 96, pages 202–207.
Li, Z. and Zhang, Y. (2021). Membership leakage in label-
only exposures. In Proceedings of the 2021 ACM
SIGSAC Conference on Computer and Communica-
tions Security, pages 880–895.
Liu, Y., Wen, R., He, X., Salem, A., Zhang, Z., Backes,
M., Cristofaro, E. D., Fritz, M., and Zhang, Y. (2022).
ML-Doctor: Holistic risk assessment of inference at-
tacks against machine learning models. In USENIX
Security.
Nasr, M., Shokri, R., and Houmansadr, A. (2018). Machine
Learning with Membership Privacy using Adversarial
Regularization. In CCS.
Shokri, R., Stronati, M., Song, C., and Shmatikov, V.
(2017). Membership inference attacks against ma-
chine learning models. In 2017 IEEE symposium on
security and privacy (SP), pages 3–18. IEEE.
Song, L. and Mittal, P. (2021). Systematic evaluation of
privacy risks of machine learning models. In USENIX
Security.
Stock, J., Wettlaufer, J., Demmler, D., and Federrath, H.
(2023). Lessons learned: Defending against prop-
erty inference attacks. In di Vimercati, S. D. C. and
Samarati, P., editors, Proceedings of the 20th Interna-
tional Conference on Security and Cryptography (SE-
CRYPT), pages 312–323. SCITEPRESS.
Suri, A. and Evans, D. (2022). Formalizing and estimating
distribution inference risks. Proceedings on Privacy
Enhancing Technologies.
Tang, X., Mahloujifar, S., Song, L., Shejwalkar, V., Nasr,
M., Houmansadr, A., and Mittal, P. (2021). Mitigat-
SECRYPT 2024 - 21st International Conference on Security and Cryptography
884