connection (RFC) time between SAP CRM and BW
systems is excluded). Since mean arrival time
between two consecutive fault incidents is
approximately 36.5 second, it is technically feasible
to perform a real-time spare part prediction.
Figure 12: View of spare part prediction result list. While
odd numbered lines at PRED_RESULT prediction result
list inform spare part predictions, even lines indicate
concomitant spare part consumptions. HZMBSLK is the
unique identifier for the corresponding fault incident.
5 CONCLUSIONS
This paper proposes a hybrid classification algorithm
for the underlying spare part prediction scenario such
that, while Apriori is adapted to handle data anomalies
and redundancies emerged at data exploration,
significance weights obtained at C4.5 incorporates the
inter-dimension similarity at interpreting the
neighborhood among fault instances. Finally, two
adaptations of weighted kNN are applied:
IkNNwC
gives more importance to major instances that are more
likely to represent a dominant class in neighborhood
region of feature space,
IkNNwAS aims to balance the
discriminative power of minor class.
According to experimental results, proposed
hybrid classification algorithm doubles the human-
level performance at spare part prediction, which is
approximately 40% accuracy. This performance
implies a 50% decrease at the average number of
customer visits per fault incident. Hence a significant
cutting at especially sales and distribution costs is
expected by the effect of spare part prediction model.
As future work, we plan to extend the corresponding
modeling to other product groups.
REFERENCES
Aguilera, J., Gonzalez, L. C., Montes-y-Gomez, M. &
Rosso, P. (2019). A New Weighted k-Nearest Neighbor
Algorithm based on Newton's Gravitational Force.
CIARP 2018, 305-313.
Boriah, S., Chandola, V., & Kumar, V. (2008). Similarity
Measures for Categorical Data: A Comparative
Evaluation, SDM, 243-254.
Chapman, P., Clinton, J., Kerber, R., Khabaza, T., Reinartz,
T., Shearer, C., Wirth, R. (1999). CRISP-DM 1.0 Step-
by-Step Data Mining Guide.
Domingos, P. (2015). The Master Algorithm: How the
Quest for the Ultimate Learning Machine Will Remake
Our World. Basic Books.
Dudani, S.A. (1976). The Distance-Weighted k-Nearest-
Neighbor Rule. IEEE Trans. Syst. Man Cybern. SMC-
6(4), 325-327.
Kantardzic, M. 2011). Data Mining: Concepts, Models,
Methods, and Algorithms. Wiley-IEEE Press.
Liu, C., Cao, L. & Yu, P. S. (2014). A Hybrid Coupled k-
Nearest Neighbor Algorithm on Imbalance Data.
International Joint Conference on Neural Networks
(IJCNN).
Liu, W. & Chawla, S. (2011). Class Confidence Weighted
kNN Algorithms for Imbalanced Data Sets, Advances
in Knowledge Discovery and Data Mining, 345-356.
Mateos-Garcia, D., Garcia-Gutierrez, J. & Riquelme-
Santos, J.C. (2016). An Evolutionary Voting For k-
Nearest Neighbors. Expert Syst. Appl. 43, 9-14.
Pang-Ning, T., Steinbach, M. & Kumar, V. (2006).
Introduction to Data Mining, Library of Congress, 74.
Parvinnia, E., Sabeti, M., Zolghadri Jahromi, M. & Boostani,
R. (2014). Classification of EEG Signals Using Adaptive
Weighted Distance Nearest Neighbor Algorithm. J. King
Saud Univ. Comput. Inf. Sci. 26(1), 1-6.
Song, Y., Huang, J., Zhou, D., Zha, H. & Giles, C. L.
(2007). IkNN: Informative K-Nearest Neighbor Pattern
Classification, Knowledge Discovery in Databases:
PKDD 2007, 248-264.
Tan, S. (2005). Neighbor-Weighted K-Nearest Neighbor for
Unbalanced Text Corpus. Expert Syst. Appl. 28(4),
667-671.
Wang, C., Cao, L., Wang, M. Li, J., Wei, W. & Ou, Y.
(2011). Coupled Nominal Similarity in Unsupervised
Learning, CIKM 2011, 973-978.
Yuxuan, L. & Zhang, X. (2011). Improving K Nearest
Neighbor with Exemplar Generalization for
Imbalanced Classification, 15th Pacific-Asia
Conference, PAKDD 2011, 1-12.
Zhang, S., Cheng, D., Deng, Z., Zong, M. & Deng, X.
(2017). A Novel kNN Algorithm with Data-Driven k
Parameter Computation. Pattern Recognit. Lett. 1-11.
Zhu, Q., Feng, J. & Huang, J. (2016). Natural Neighbor: A
Self-Adaptive Neighborhood Method without
Parameter k. Pattern Recognit. Lett. 80, 30-36.