5 CONCLUSION AND FURTHER
WORK
Early results seem promising, as the system could
predict the tendencies of the workload behaviour as
it changed over time. However, we will shorten the
prediction horizon to gain higher prediction
accuracy in future. A prediction horizon of 2-4
hours, rather than the 7 hours used so far, is
sufficient for the future applications that are planned
to exploit that prediction system. We also plan to
investigate the inclusion of calendaring data into the
prediction method, to include prior knowledge about
load peaks on special days of the month.
The results of this project show that it is possible
to predict the PI, as a relevant performance
indicator, of a complex mainframe system cluster,
like a z/OS Sysplex. This enables us to develop
several functionalities that do resource assignment
or resource provisioning to workloads right in time.
This can avoid resource contention, like CPU or
memory usage, significantly. Especially in big
mainframe environments with very large numbers of
competing workloads this can improve the
throughput and optimal resource usage significantly
and thus optimise data processing costs.
Further work in this area is to analyse and
develop an automatic relearning environment for the
prediction of non-stationary processes. This way, the
prediction system will be able to permanently
improve its prediction quality for each customer by
learning more and more about its particular
environment and workload behaviour. As an
additional benefit, relearning adjusts the prediction
to changes in typical customer workloads, e.g. when
business related changes take effect.
We expect further improvements, when
additional explicit knowledge is incorporated, e.g.
changes or higher workload at special days of the
month or upcoming events.
No other results in the field of operating system
workload management prediction are known to us,
hindering the comparison of our early results with
others.
ACKNOWLEDGEMENTS
The authors thank Clemens Gebhard and Sarah
Kleeberg for their participation in this project.
REFERENCES
Aman, J., Eilert, C. K., Emmes, D., Yocom, P.,
Dillenberger, D., 1997. Adaptive algorithms for
managing a distributed data processing workload. In
IBM Systems Journal, 36(2): 242-283.
Bensch, M., Schröder, M., Bogdan, M., Rosenstiel, W.,
2005. Feature Selection for High-Dimensional Indus-
trial Data. In Proceedings of European Symposium on
Artificial Neural Networks (ESANN), Bruges.
Bigus, J. P., Hellerstein, J. L., Jayram, T. S., Squillante,
M. S., 2000. AutoTune: A Generic Agent for Auto-
mated Performance Tuning. In Practical Application
of Intelligent Agents and Multi Agent Technology.
Guyon, I., Weston, J., Barnhill, S., Vapnik, V., 2002. Gene
Selection for Cancer Classification using Support
Vector Machines. Machine Learning, 46(1): 389-422.
Kearns, M., Mansour, Y., Ng, A. Y., Ron, D., 1997. An
Experimental and Theoretical Comparison of Model
Selection Methods. Machine Learning, 27(1): 7–50.
Mohraz, K., Protzel, P., 1996. FlexNet: A Flexible Neural
Network Construction Algorithm. In Proceedings of
European Symposium on Artificial Neural Networks
(ESANN), Bruges.
Ortiz-Arroyo, D., Skov, M. K., Huynh, Q., 2005. Accurate
Electricity Load Forecasting with Artificial Neural
Networks. In Computational Intelligence for
Modelling, Control and Automation, 1: 94-99.
Polak, E., Ribiere, G., 1969. Note sur la Convergence de
Methodes de Directions Conjugees. Revue Francaise
d'Informatique et de Recherche Operationnelle, 3: 35-
43.
Rumelhart, D., Hinton, G., Williams, R., 1986. Learning
internal representations by error propagation. In D.
Rumelhart and J. McClelland, editors, Parallel
Distributed Processing, pages 318-362. MIT Press.
Vaupel, R., Teuffel, M., 2004. Das Betriebssystem z/OS
und die zSeries – Die Darstellung eines modernen
Großrechnersystems. ISBN 3-486-27528-3, Olden-
bourg Wissenschaftsverlag, Germany.
ICEIS 2007 - International Conference on Enterprise Information Systems
218