ability. We validate our approach on the Adult dataset
using three classification models, demonstrating the
potential for trustworthy data analysis while control-
ling privacy levels. In future work, we aim to ex-
tend our approach to other types of datasets (i.e. non-
tabular) and analysis techniques using various privacy
preservation techniques.
ACKNOWLEDGEMENTS
This work has been partially funded by the EU-funded
project H2020 SIFIS-Home GA ID:952652.
REFERENCES
Abadi, M., Chu, A., Goodfellow, I., McMahan, H. B.,
Mironov, I., Talwar, K., and Zhang, L. (2016). Deep
learning with differential privacy. In Proceedings of
the 2016 ACM SIGSAC conference on computer and
communications security, pages 308–318.
Arjovsky, M., Chintala, S., and Bottou, L. (2017). Wasser-
stein generative adversarial networks. In Interna-
tional conference on machine learning, pages 214–
223. PMLR.
Biecek, P. and Burzykowski, T. (2021). Explanatory model
analysis: explore, explain, and examine predictive
models. CRC Press.
Blake, C. (1998). Cj merz uci repository of machine learn-
ing databases. University of California at Irvine.
Breiman, L. (2001). Random forests. Machine learning,
45(1):5–32.
Budig, T., Herrmann, S., and Dietz, A. (2020). Trade-offs
between privacy-preserving and explainable machine
learning in healthcare. In Seminar Paper, Inst. Appl.
Informat. Formal Description Methods (AIFB), KIT
Dept. Econom. Manage., Karlsruhe, Germany.
Dwork, C. (2008). Differential privacy: A survey of results.
In International conference on theory and applica-
tions of models of computation, pages 1–19. Springer.
Dwork, C., Kenthapadi, K., McSherry, F., Mironov, I., and
Naor, M. (2006). Our data, ourselves: Privacy via dis-
tributed noise generation. In Annual int. conference
on the theory and applications of cryptographic tech-
niques, pages 486–503. Springer.
Fung, B. C., Wang, K., Chen, R., and Yu, P. S. (2010).
Privacy-preserving data publishing: A survey of re-
cent developments. ACM Computing Surveys (Csur),
42(4):1–53.
Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B.,
Warde-Farley, D., Ozair, S., Courville, A., and Ben-
gio, Y. (2014). Generative adversarial nets. Advances
in neural information processing systems, 27.
Harder, F., Bauer, M., and Park, M. (2020). Interpretable
and differentially private predictions. In Proceedings
of the AAAI Conference on Artificial Intelligence, vol-
ume 34, pages 4083–4090.
Hleg, A. (2019). Ethics guidelines for trustworthy ai. B-
1049 Brussels.
Jakku, E., Taylor, B., Fleming, A., Mason, C., Fielke,
S., Sounness, C., and Thorburn, P. (2019). “if they
don’t tell us what they do with it, why would we
trust them?” trust, transparency and benefit-sharing in
smart farming. NJAS-Wageningen Journal of Life Sci-
ences, 90:100285.
Jaseena, K., David, J. M., et al. (2014). Issues, challenges,
and solutions: big data mining. CS & IT-CSCP,
4(13):131–140.
Jayaraman, B. and Evans, D. (2019). Evaluating differen-
tially private machine learning in practice. In 28th
{USENIX} Security Symposium ({USENIX} Security
19), pages 1895–1912.
Li, N., Li, T., and Venkatasubramanian, S. (2007).
t-closeness: Privacy beyond k-anonymity and l-
diversity. In 2007 IEEE 23rd International Confer-
ence on Data Engineering, pages 106–115. IEEE.
Linardatos, P., Papastefanopoulos, V., and Kotsiantis, S.
(2021). Explainable ai: A review of machine learn-
ing interpretability methods. Entropy, 23(1):18.
Machanavajjhala, A., Kifer, D., Gehrke, J., and Venkita-
subramaniam, M. (2007). l-diversity: Privacy beyond
k-anonymity. ACM Transactions on Knowledge Dis-
covery from Data (TKDD), 1(1):3–es.
Patel, N., Shokri, R., and Zick, Y. (2020). Model ex-
planations with differential privacy. arXiv preprint
arXiv:2006.09129.
Rai, A. (2020). Explainable ai: From black box to glass
box. Journal of the Academy of Marketing Science,
48(1):137–141.
Ramon, J. and Basu, M. (2020). Interpretable privacy with
optimizable utility. In ECML/PKDD workshop on eX-
plainable Knowledge Discovery in Data mining.
Rasheed, K., Qayyum, A., Ghaly, M., Al-Fuqaha, A., Razi,
A., and Qadir, J. (2021). Explainable, trustworthy, and
ethical machine learning for healthcare: A survey.
Raskhodnikova, S., Smith, A., Lee, H. K., Nissim, K., and
Kasiviswanathan, S. P. (2008). What can we learn pri-
vately. In Proceedings of the 54th Annual Symposium
on Foundations of Computer Science, pages 531–540.
Samarati, P. (2001). Protecting respondents identities in mi-
crodata release. IEEE transactions on Knowledge and
Data Engineering, 13(6):1010–1027.
Sheikhalishahi, M., Saracino, A., Martinelli, F., and Marra,
A. L. (2021). Privacy preserving data sharing and
analysis for edge-based architectures. International
Journal of Information Security.
Shokri, R., Strobel, M., and Zick, Y. (2019). Privacy risks
of explaining machine learning models.
Sweeney, L. (2002). k-anonymity: A model for protecting
privacy. International Journal of Uncertainty, Fuzzi-
ness and Knowledge-Based Systems, 10(05):557–570.
Weisberg, S. (2005). Applied linear regression, volume 528.
John Wiley & Sons.
Xie, L., Lin, K., Wang, S., Wang, F., and Zhou, J. (2018).
Differentially private generative adversarial network.
arXiv preprint arXiv:1802.06739.
The Explainability-Privacy-Utility Trade-Off for Machine Learning-Based Tabular Data Analysis
519