8 CONCLUSION
We introduced the first implementation of a federated
Naive Bayes classifier under differential privacy. The
relative performance of our approach to its centralized
counterpart varies depending on the characteristics of
the dataset. According to our Monte Carlo analysis,
in most cases, a small privacy penalty must be paid
to achieve the same accuracy level. However, when
the local sensitivity in each federated data partition is
much lower than the sensitivity of the whole dataset,
we show that the federated Laplace mechanism better
estimates the true parameters of the model.
Moreover, we demostrated that our algorithm can
be easily extended to incorporate additional features
such as online updates and full decentralization, while
more research is needed to achieve byzantine re-
silience.
ACKNOWLEDGEMENTS
This project has received funding from the Euro-
pean Union’s Horizon 2020 research and innovation
programme under the Marie Skłodowska-Curie grant
agreement No 813162: RAIS – Real-time Analytics
for the Internet of Sports. The content of this paper
reflects the views only of their author (s). The Eu-
ropean Commission/ Research Executive Agency are
not responsible for any use that may be made of the
information it contains.
REFERENCES
Abadi, M., Chu, A., Goodfellow, I., McMahan, H. B.,
Mironov, I., Talwar, K., and Zhang, L. (2016). Deep
learning with differential privacy. In Proceedings of
the 2016 ACM SIGSAC conference on computer and
communications security, pages 308–318.
Awan, S., Luo, B., and Li, F. (2021). Contra: Defending
against poisoning attacks in federated learning. In Eu-
ropean Symposium on Research in Computer Security,
pages 455–475. Springer.
Bernal, D. G., Giaretta, L., Girdzijauskas, S., and Sahlgren,
M. (2021). Federated word2vec: Leveraging feder-
ated learning to encourage collaborative representa-
tion learning. arXiv preprint arXiv:2105.00831.
Blanchard, P., El Mhamdi, E. M., Guerraoui, R., and
Stainer, J. (2017). Machine learning with adversaries:
Byzantine tolerant gradient descent. Advances in Neu-
ral Information Processing Systems, 30.
Dwork, C. (2008). Differential privacy: A survey of results.
In International conference on theory and applica-
tions of models of computation, pages 1–19. Springer.
Geyer, R. C., Klein, T., and Nabi, M. (2017). Differentially
private federated learning: A client level perspective.
arXiv preprint arXiv:1712.07557.
Giaretta, L. and Girdzijauskas,
ˇ
S. (2019). Gossip learning:
Off the beaten path. In 2019 IEEE International Con-
ference on Big Data (Big Data), pages 1117–1124.
IEEE.
Guerraoui, R., Rouault, S., et al. (2018). The hidden vul-
nerability of distributed learning in byzantium. In In-
ternational Conference on Machine Learning, pages
3521–3530. PMLR.
Heged
˝
us, I., Danner, G., and Jelasity, M. (2021). Decentral-
ized learning works: An empirical comparison of gos-
sip learning and federated learning. Journal of Paral-
lel and Distributed Computing, 148:109–124.
Islam, T. U., Ghasemi, R., and Mohammed, N. (2022).
Privacy-preserving federated learning model for
healthcare data. In 2022 IEEE 12th Annual Com-
puting and Communication Workshop and Conference
(CCWC), pages 0281–0287. IEEE.
Jelasity, M., Montresor, A., and Babaoglu, O. (2005).
Gossip-based aggregation in large dynamic networks.
ACM Trans. Comput. Syst., 23(3):219–252.
Ji, Z., Lipton, Z. C., and Elkan, C. (2014). Differential pri-
vacy and machine learning: a survey and review. arXiv
preprint arXiv:1412.7584.
Kantarcıoglu, M., Vaidya, J., and Clifton, C. (2003). Pri-
vacy preserving naive bayes classifier for horizontally
partitioned data. In IEEE ICDM workshop on privacy
preserving data mining, pages 3–9.
Li, T., Li, J., Liu, Z., Li, P., and Jia, C. (2018). Differen-
tially private naive bayes learning over multiple data
sources. Information Sciences, 444:89–104.
Lopuha
¨
a-Zwakenberg, M., Alishahi, M., Kivits, J., Klaren-
beek, J., van der Velde, G.-J., and Zannone, N. (2021).
Comparing classifiers’ performance under differential
privacy. In International Conference on Security and
Cryptography (SECRYPT).
Lyu, L., Yu, H., Ma, X., Sun, L., Zhao, J., Yang, Q., and
Yu, P. S. (2020). Privacy and robustness in feder-
ated learning: Attacks and defenses. arXiv preprint
arXiv:2012.06337.
Marchioro., T., Kazlouski., A., and Markatos., E. (2021).
User identification from time series of fitness data. In
Proceedings of the 18th International Conference on
Security and Cryptography - SECRYPT,, pages 806–
811. INSTICC, SciTePress.
McMahan, B., Moore, E., Ramage, D., Hampson, S., and
Arcas, B. A. y. (2017). Communication-Efficient
Learning of Deep Networks from Decentralized Data.
In Singh, A. and Zhu, J., editors, Proceedings of
the 20th International Conference on Artificial Intelli-
gence and Statistics, volume 54 of Proceedings of Ma-
chine Learning Research, pages 1273–1282. PMLR.
McSherry, F. D. (2009). Privacy integrated queries: an ex-
tensible platform for privacy-preserving data analysis.
In Proceedings of the 2009 ACM SIGMOD Interna-
tional Conference on Management of data, pages 19–
30.
Federated Naive Bayes under Differential Privacy
179