Table 5: Incremental learning performance for neural networks on the “varying classes” Optical Digits data set of Polikar et
al. (2002). Mean percentage rates of correct test set classification, with standard errors in brackets.
T
1
T
2
T
3
T
4
Traditional NN
48.47 (0.01) 75.49 (0.02) 79.87 (0.07) 77.07 (0.07)
Learn++
46.6 68.9 82.0 87.0
Evolved NN
48.59 (0.09) 76.39 (0.18) 84.78 (0.47) 87.41 (0.39)
Evolved NN with Dual-weights
47.34 (0.32) 80.13 (0.50) 87.32 (0.15) 93.84 (0.04)
number of epochs allowed. Both of these factors
form part of the natural evolved regularization
process, and further performance improvements may
be possible by simply increasing those maximum
values. However, the improvements achievable by
doing this prove to be rather limited in relation to the
enormous increase in the associated computational
costs. Indeed, the computational cost of the general
evolutionary process proposed in this paper is also
high, but even small improvements are often
extremely valuable, so it will generally remain a
complex problem dependent decision whether the
potential improvements are worth the extra effort.
6 CONCLUSIONS
This paper has provided a more general and more
statistically rigorous confirmation of earlier results
(Seipone & Bullinaria, 2005b) indicating that the
application of evolutionary computation techniques
can massively improve the incremental learning
abilities of standard neural networks on real-world
generalization tasks compared to existing systems
such as Learn++ (Polikar et al., 2001, 2002). It has
also demonstrated that the same approach can be
used to evolve more sophisticated dual-weight
architectures that have further improvements in
performance, particularly when the representation of
classes varies between training data batches. Thus
effective evolutionary neural network techniques
have been established that can straightforwardly be
tested on and applied to any future incremental
learning problems requiring good generalization.
REFERENCES
Ans, B., Rousset, S, French, R.M., Musca, S., 2002.
Preventing Catastrophic Interference in Multiple-
Sequence Learning Using Coupled Reverberating
Elman Networks. Proceedings of the Twenty-fourth
Annual Conference of the Cognitive Science Society,
71-76. Mahwah, NJ: LEA
Bishop, C.M., 1995. Neural Networks for Pattern
Recognition. Oxford, UK: Oxford University Press.
Blake, C.L., Merz, C.J., 1998. UCI Repository of Machine
Learning Databases. University of California, Irvine.
http://www.ics.uci.edu/~mlearn/MLRepository.html
Bullinaria, J.A., 2007. Using Evolution to Improve Neural
Network Learning: Pitfalls and Solutions. Neural
Computing & Applications, 16, 209-226.
Frean, M., Robins, A., 1999. Catastrophic Forgetting in
Simple Neural Networks: An Analysis of the
Pseudorehearsal Solution. Network: Computation in
Neural Systems, 10, 227-236.
French, R.M., 1999. Catastrophic Forgetting in
Connectionist Networks. Trends in Cognitive
Sciences, 4, 128-135.
Giraud-Carrier, C., 2000. A Note on the Utility of
Incremental Learning. AI Communications, 13, 215-
223.
Hinton, G.E., Plaut, D.C., 1987. Using Fast Weights to
Deblur Old Memories. Proceedings of the Ninth
Annual Conference of the Cognitive Science Society,
177-186. Hillsdale, NJ: LEA.
McClelland, J.L., McNaughton, B.L., O’Reilly, R.C.,
1995. Why There Are Complementary Learning
Systems in the Hippocampus and Neocortex: Insights
From the Successes and Failures of Connectionist
Models of Learning and Memory. Psychological
Review, 102, 419-457.
Polikar, R., Byorick, J., Krause, S., Marino, A., Moreton,
M., 2002. Learn++: A Classifier Independent
Incremental Learning Algorithm for Supervised
Neural Networks. Proceedings of the 2002
International Joint Conference on Neural Networks, 2,
1742-1747.
Polikar, R., Udpa, L., Udpa, S.S., Honavar, V., 2001.
Learn++, 2001. An Incremental Learning Algorithm
for Multi-Layer Perceptron Networks. IEEE
Transactions on Systems, Man, and Cybernetics-Part
C: Applications and Reviews, 31, 497-508.
Seipone, T., Bullinaria, J.A., 2005a. The Evolution of
Minimal Catastrophic Forgetting in Neural Systems.
Proceedings of the Twenty-Seventh Annual
Conference of the Cognitive Science Society, 1991-
1996. Mahwah, NJ: LEA.
Seipone, T., Bullinaria, J.A., 2005b. Evolving Improved
Incremental Learning Schemes for Neural Network
Systems. Proceedings of the 2005 IEEE Congress on
Evolutionary Computing (CEC 2005), 273-280.
Piscataway, NJ: IEEE.
Yao, X., 1999. Evolving Artificial Neural Networks.
Proceedings of the IEEE, 87, 1423-1447.
IJCCI 2009 - International Joint Conference on Computational Intelligence
434