Figure 9: ST-Predictionvs Residuals for Metric 2 in unusual
segment.
Figure 10: ERC-Prediction vs Residuals for Metric 2 in un-
usual segment.
zero mean value and show a constant variation be-
haviour, which is again can be seen as an indication
that both predictive models are adequate in modeling
the variation in the response variables in ”unusual”
test data. Although we observe some outliers in the
residuals. we could say that this is normal consider-
ing that the ”unusual” dataset is statistically different
than the dataset we train our machine learning algo-
rithms.
4 CONCLUSION
In this paper we proposedan ensemble machine learn-
ing algorithm in order to predict the finished yarn
quality. The data is first segmented into ten clusters
nine of which is denoted as ”normal” and the one
with the highest distance from the general mean as
”unusual” via local outlier factor method. The former
cluster refers to production data one may expect due
to the nature of the process and latter is the dataset
showing an usual pattern compared to expected pro-
cess data. Then a set of classical machine learning
algorithms are applied and performances of the algo-
rithms is compared. It is seen that for the unusual
segment, performance of the classical algorithms gets
worse especially for one of the quality metrics. As
a remedy, an ensemble algorithm based on regressor
chains is recommended and yielding higher predic-
tion performance in two thirds of the dataset.
As the next step, implemented algorithm will be
fully tested at the facility. If the prediction perfor-
mance remains satisfactory, we’re going to move on
the next phase and start using the predictive tool as a
recommendation engine for the machine operator. At
this stage, operator will be informed about the sug-
gested production settings for the machine and the
recommendation system will perform as a decision
support tool, meaning that the recommendations of
the tool are push forward to the machine only if the
operator gives consent. Once the second phase is suc-
cessful, the recommendation engine will be plugged
into PLC and start changing set parameters of the ma-
chine as a part of the automation system. After the
third phase, the manufacturing plant will have its first
full-scale industry 4.0 application.
REFERENCES
Borchani, H., Varando, G., Bielza, C., and Larra˜naga, P.
(2015). A survey on multi-output regression. Wiley
Interdisciplinary Reviews: Data Mining and Knowl-
edge Discovery, 5(5):216–233.
Breiman, L. (2001). Random forests. Machine learning,
45(1):5–32.
Breunig, M. M., Kriegel, H.-P., Ng, R. T., and Sander, J.
(2000). Lof: identifying density-based local outliers.
In ACM sigmod record, volume 29, pages 93–104.
ACM.
Chandola, V., Banerjee, A., and Kumar, V. (2009).
Anomaly detection: A survey. ACM computing sur-
veys (CSUR), 41(3):15.
De’Ath, G. (2002). Multivariate regression trees: a new
technique for modeling species–environment relation-
ships. Ecology, 83(4):1105–1117.
Friedman, J. H. et al. (1991). Multivariate adaptive regres-
sion splines. The annals of statistics, 19(1):1–67.
Han, J., Pei, J., and Kamber, M. (2011). Data mining: con-
cepts and techniques. Elsevier.
Quinlan, J. R. (1986). Induction of decision trees. Machine
learning, 1(1):81–106.
Simil¨a, T. and Tikka, J. (2007). Input selection and shrink-
age in multiresponse linear regression. Computational
Statistics & Data Analysis, 52(1):406–422.
Spyromitros-Xioufis, E., Tsoumakas, G., Groves, W.,
and Vlahavas, I. (2012). Multi-label classification
methods for multi-target regression. arXiv preprint
arXiv:1211.6581, pages 1159–1168.
Struyf, J. and Dˇzeroski, S. (2005). Constraint based induc-
tion of multi-objective regression trees. In Interna-
tional Workshop on Knowledge Discovery in Inductive
Databases, pages 222–233. Springer.
Vazquez, E. and Walter, E. (2003). Multi-output supp-
port vector regression. IFAC Proceedings Volumes,
36(16):1783–1788.
Zhang, W., Liu, X., Ding, Y., and Shi, D. (2012). Multi-
output ls-svr machine in extended feature space.
In 2012 IEEE International Conference on Compu-
tational Intelligence for Measurement Systems and
Applications (CIMSA) Proceedings, pages 130–134.
IEEE.