how the model performs. This section was the most
interesting section for one of the reviewers in terms
of understanding. However, for “Prediction
Distribution and Classification Reports,” one of the
comments suggests that they are unnecessary.
Features’ Importance Section. Several reviews
mentioned that this section is important to give an
idea about the data. The correlation plot got the most
attention; however, the size of the plots was too small
to read.
Sensitivity Analysis Section. Most of the comments
agreed that selecting a variable is very helpful to
understand the performance. However, one of the
comments found it hard to understand the categorical
attributes plots.
Finally, most of the comments were positive.
Comments related to the size of plots, typos, and
rewording were reflected on the dashboard. The other
suggestions would be considered as future work due
to time limitations.
5 CONCLUSION AND
DISCUSSION
The present work was designed to demonstrate an
approach to visualizing classification model
performance in a dashboard with three sections:
statistical measures, which provide an overview of
the model performance; feature importance which
gives an overview of the data; and sensitivity analysis
which identifies the relationship between the attribute
and the prediction. The dashboard adds to a growing
body of literature on understanding and evaluating
classification learning. The advantages of the
dashboard are that it visualizes any classification
model, uses visuals that are simple and easy to
understand, and summarizes all the results in one
place. Yet, unlike interactive dashboards, this
dashboard does not react to user changes.
5.1 Limitation and Future Work
The survey results cannot be generalized due to
sample size limitations. However, the purpose of the
survey was to understand how people interact with
the dashboard, and the most interesting part was the
reviewers’ comments. Second, some design-related
changes like the colors and sizes of the plots are
recommended. For example, when the names of the
columns are long, the size of the figures in the feature
importance section becomes small, which requires
zooming in to read. Third, visualizing the regression
model results and comparing models is considered a
future work.
REFERENCES
Adams, N. M., & Hand, D. J. (1999). Comparing classifiers
when the misallocation costs are uncertain. Pattern
Recognition, 32(7), 1139–1147. https://doi.org/
10.1016/S0031-3203(98)00154-X
Alsallakh, B., Hanbury, A., Hauser, H., Miksch, S., &
Rauber, A. (2014). Visual Methods for Analyzing
Probabilistic Classification Data. IEEE Transactions on
Visualization and Computer Graphics, 20(12), 1703–
1712. https://doi.org/10.1109/TVCG.2014.2346660
Amershi, S., Cakmak, M., Knox, W. B., & Kulesza, T.
(2014). Power to the People: The Role of Humans in
Interactive Machine Learning. AI Magazine, 35(4),
105–120. https://doi.org/10.1609/aimag.v35i4.2513
Apley, D. W., & Zhu, J. (2020). Visualizing the effects of
predictor variables in black box supervised learning
models. Journal of the Royal Statistical Society: Series
B (Statistical Methodology), 82(4), 1059–1086. https://
doi.org/10.1111/rssb.12377
Bagais, W. (2021). Dashboard for Machine Learning
Models in Health Care. George Mason University.
Brooks, M., Amershi, S., Lee, B., Drucker, S. M., Kapoor,
A., & Simard, P. (2015). FeatureInsight: Visual support
for error-driven feature ideation in text classification.
2015 IEEE Conference on Visual Analytics Science and
Technology (VAST), 105–112. https://doi.org/10.1109/
VAST.2015.7347637
Fekete, J.-D. (2013). Visual Analytics Infrastructures: From
Data Management to Exploration. Computer, 46(7),
22–29. https://doi.org/10.1109/MC.2013.120
Jerome H. Friedman. (2001). Greedy function
approximation: A gradient boosting machine. The
Annals of Statistics, 29(5), 1189–1232. https://doi.org/
10.1214/aos/1013203451
Krause, J., Perer, A., & Ng, K. (2016). Interacting with
Predictions: Visual Inspection of Black-Box Machine
Learning Models. In Proceedings of the 2016 CHI
Conference on Human Factors in Computing Systems
(pp. 5686–5697). Association for Computing
Machinery. https://doi.org/10.1145/2858036.2858529
Liu, S., Wang, X., Liu, M., & Zhu, J. (2017). Towards better
analysis of machine learning models: A visual analytics
perspective. Visual Informatics, 1(1), 48–56. https://
doi.org/10.1016/j.visinf.2017.01.006
Lundberg, S. M., & Lee, S.-I. (2017). A Unified Approach
to Interpreting Model Predictions. In I. Guyon, U. V.
Luxburg, S. Bengio, H. Wallach, R. Fergus, S.
Vishwanathan, & R. Garnett (Eds.), Advances in Neural
Information Processing Systems 30 (pp. 4765–4774).
Curran Associates, Inc. http://papers.nips.cc/paper/
7062-a-unified-approach-to-interpreting-model-
predictions.pdf