
consistently lower RMSE values across all models,
particularly when Brightness bias was selectively ap-
plied. This suggests that biases can be exploited to
infer client data contributions, though the effective-
ness varies between datasets. Although the strategy
employed in this work is from a more theoretical na-
ture, we empirically proved that the Leukocyte dataset
is highly vulnerable to such threats. Only a few
collected data points were sufficient for a successful
knowledge retrieval.
In conclusion, the results highlight the importance
of understanding how bias type and dataset character-
istics interact to affect FL model performance. These
insights can help designing more robust and secure
FL systems, particularly in settings where data hetero-
geneity and malicious clients may pose risks. Overall,
one cannot draw general conclusions across different
datasets. Experiments must be carefully planned and
executed when it comes to data manipulation, such as
the injection of biases. Given the highly sensitive na-
ture of human health data, we recommend conducting
even more nuanced research regarding these datasets.
Especially in FL, where each client constitutes a vul-
nerability, one compromised client can cause serious
trouble, making it essential to pursue state-of-the-art
data security mechanisms.
For future work, it would be interesting to exam-
ine additional bias types to strategically extract dif-
ferent information from honest clients. Additionally,
none of the models presented in this work were op-
timized, and we used the same architectures to en-
sure a fair comparison. However, given that different
datasets can yield completely different conclusions
even with the same architecture and circumstances,
optimizing models for specific datasets and rerunning
the same attacks could be beneficial. Considering the
promising results, we believe this approach could lead
to a significant performance boost and would be worth
further investigation.
ACKNOWLEDGEMENTS
The authors would like to especially thank their colleagues
from the Heinz-Nixdorf Chair of Biomedical Electronics -
D. Heim and C. Klenk - for performing sample preparation
and measurements.
REFERENCES
Alhajjar, E., Maxwell, P., and Bastian, N. (2021). Ad-
versarial machine learning in Network Intrusion De-
tection Systems. Expert Systems with Applications,
186:115782.
Barcia, J. J. (2007). The Giemsa Stain: Its History and Ap-
plications. International Journal of Surgical Pathol-
ogy, 15(3):292–296.
Bishop, C. M. (2006). Pattern Recognition and Ma-
chine Learning. Information Science and Statistics.
Springer, New York.
Breiman, L. (2001). Random Forests. Machine Learning,
45(1):5–32.
Dubois, F. and Yourassowsky, C. (2008). Digital holo-
graphic microscope for 3D imaging and process using
it.
Dubois, F. and Yourassowsky, C. (2011). Off-axis interfer-
ometer.
Fresacher, D., R
¨
ohrl, S., Klenk, C., Erber, J., Irl, H.,
Heim, D., Lengl, M., Schumann, S., Knopp, M.,
Schlegel, M., Rasch, S., Hayden, O., and Diepold,
K. (2023). Composition counts: A machine learning
view on immunothrombosis using quantitative phase
imaging. Proceedings of Machine Learning Research,
219:208–229.
Gianfrancesco, M. A., Tamang, S., Yazdany, J., and Schma-
juk, G. (2018). Potential Biases in Machine Learn-
ing Algorithms Using Electronic Health Record Data.
JAMA Internal Medicine, 178(11):1544.
Gonzalez, R. C. and Woods, R. E. (2002). Digital Image
Processing. Prentice Hall, Upper Saddle River, N.J,
2nd ed edition.
Goodfellow, I., Bengio, Y., and Courville, A. (2016). Deep
Learning. Adaptive Computation and Machine Learn-
ing. The MIT Press, Cambridge.
Goodfellow, I. J., Shlens, J., and Szegedy, C. (2014). Ex-
plaining and harnessing adversarial examples. CoRR,
abs/1412.6572.
Hughes, J. F. (2014). Computer Graphics: Principles and
Practice. Addison-Wesley, Upper Saddle River, New
Jersey, third edition edition.
James, G., Witten, D., Hastie, T., and Tibshirani, R. (2013).
An Introduction to Statistical Learning, volume 103
of Springer Texts in Statistics. Springer, New York.
Jegorova, M., Kaul, C., Mayor, C., O’Neil, A. Q., Weir, A.,
Murray-Smith, R., and Tsaftaris, S. A. (2022). Sur-
vey: Leakage and Privacy at Inference Time. IEEE
Transactions on Pattern Analysis and Machine Intel-
ligence, pages 1–20.
Jo, Y., Cho, H., Lee, S. Y., Choi, G., Kim, G., Min, H.-s.,
and Park, Y. (2019). Quantitative Phase Imaging and
Artificial Intelligence: A Review. IEEE Journal of
Selected Topics in Quantum Electronics, 25(1):1–14.
Kholod, I., Yanaki, E., Fomichev, D., Shalugin, E.,
Novikova, E., Filippov, E., and Nordlund, M. (2020).
Open-Source Federated Learning Frameworks for
IoT: A Comparative Review and Analysis. Sensors,
21(1):167.
Klenk, C., Erber, J., Fresacher, D., R
¨
ohrl, S., Lengl, M.,
Heim, D., Irl, H., Schlegel, M., Haller, B., Lahmer,
T., Diepold, K., Rasch, S., and Hayden, O. (2023).
Platelet aggregates detected using quantitative phase
imaging associate with COVID-19 severity. Commu-
nications Medicine, 3(1):161.
BIOINFORMATICS 2025 - 16th International Conference on Bioinformatics Models, Methods and Algorithms
442