visualizations work for error analysis in the
production domain. Our insights are based on
interviews with students as well as practitioners from
production quality management. We created
synthetic data with predefined feature value ranges
for errors. Based on the synthetic data we created 15
different visualizations. We discussed requirements
and wishes for visualization to identify corrupted
products based on the provided data. We summarized
the results from the interviews and discuss them, to
show the best and worst visualizations. One of the
favored visualizations was the Surrogate Decision
Tree Model, because it reflects the requirements for a
plot that is easily understandable and interpretable.
The Scatter Plot is also useful as an easy-to-
understand visualization and ties with the Surrogate
Decision Tree Model on the first place. Furthermore,
we contribute with eight possible combinations to use
the visualizations. These should help to analyze the
data more precise and identify the error cause. We
also identified a desire to use interactive
visualizations by the participants. Therefore, future
investigation should address this aspect. Further, it
has to be tested how useful the presented
visualizations are in practice.
ACKNOWLEDGEMENTS
This project was funded by the German Federal
Ministry of Education and Research, funding line
“Forschung an Fachhochschulen mit Unternehmen
(FHProfUnt)“, contract number 13FH249PX6. The
responsibility for the content of this publication lies
with the authors. Also, we want to thank the company
SICK AG for the cooperation and partial funding.
Further, we thank Mr. McDouall for proofreading.
REFERENCES
Abu-Rmileh, A. (2019, February 20). Be careful when
interpreting your features importance in XGBoost!
Medium. https://towardsdatascience.com/be-careful-
when-interpreting-your-features-importance-in-
xgboost-6e16132588e7.
Apley, D. W., & Zhu, J. (2016). Visualizing the effects of
predictor variables in black box supervised learning
models.
Arrieta, A. B., Díaz-Rodríguez, N., Del Ser, J., Bennetot,
A., Tabik, S., Barbado, A., ... & Chatila, R. (2020).
Explainable Artificial Intelligence (XAI): Concepts,
taxonomies, opportunities and challenges toward
responsible AI. Information Fusion, 58, 82-115.
Bhatt, U., Xiang, A., Sharma, S., Weller, A., Taly, A., Jia,
Y., ... & Eckersley, P. (2020, January). Explainable
machine learning in deployment. In Proceedings of the
2020 Conference on Fairness, Accountability, and
Transparency (pp. 648-657).
Elshawi, R., Al-Mallah, M. H., & Sakr, S. (2019). On the
interpretability of machine learning-based model for
predicting hypertension. BMC medical informatics and
decision making, 19(1), 146.
Fisher, A., Rudin, C., & Dominici, F. (2018). Model class
reliance: Variable importance measures for any
machine learning model class, from the “Rashomon”
perspective. arXiv preprint arXiv:1801.01489,68
Hirsch, Vitali, Peter Reimann, and Bernhard Mitschang.
“Data-Driven Fault Diagnosis in End-of-Line Testing
of Complex Products.” 2019 IEEE International
Conference on Data Science and Advanced Analytics
(DSAA), 2019.
Hunter, J. (2007). Matplotlib: A 2D graphics environment.
Computing in Science & Engineering, 9(3), 90–95.
Jiangchun, L. (2018). SauceCat/PDPbox. GitHub.
https://github.com/SauceCat/PDPbox.
Li, Zhixiong, Ziyang Zhang, Junchuan Shi, and Dazhong
Wu. “Prediction of Surface Roughness in Extrusion-
Based Additive Manufacturing with Machine
Learning.” Robotics and Computer-Integrated
Manufacturing 57 (2019): 488–95.
Lundberg, S., & Lee, S. I. (2017). A unified approach to
interpreting model predictions. arXiv preprint
arXiv:1705.07874
Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V.,
Thirion, B., Grisel, O., Blondel, M., Prettenhofer, P.,
Weiss, R., Dubourg, V., Vanderplas, J., Passos, D.,
Brucher, M., Perrot, M., & Duchesnay, E. (2011).
Scikit-learn: Machine Learning in Python. Journal of
Machine Learning Research, 12, 2825–2830.Ribeiro,
M. T., Singh, S., & Guestrin, C. (2016). "Why
Should I Trust You?". Proceedings of the 22nd ACM
SIGKDD International Conference on Knowledge
Discovery and Data Mining.
https://doi.org/10.1145/2939672.2939778
Ribeiro, M. T., Singh, S., & Guestrin, C. (2018, April).
Anchors: High-precision model-agnostic explanations.
In Thirty-Second AAAI Conference on Artificial
Intelligence.
Roscher, R., Bohn, B., Duarte, M. F., & Garcke, J. (2020).
Explainable machine learning for scientific insights and
discoveries. IEEE Access, 8, 42200-42216.
Waskom, M. L. (2021). seaborn: statistical data
visualization. Journal of Open Source Software, 6(60),
3021.
Zhao, Q., & Hastie, T. (2019). Causal interpretations of
black-box models. Journal of Business & Economic
Statistics, 1-10.
Ziekow, H., Schreier, U., Saleh, A., Rudolph, C., Ketterer,
K., Grozinger, D., & Gerling, A. (2019). Proactive
Error Prevention in Manufacturing Based on an
Adaptable Machine Learning Environment. From
Research to Application, 113.