units, and thus is likely not to play a useful role in
the task.
4.2 Pure Error-driven Learning
Figure 4 presents the activation-based receptive field
and projective field analysis for the lower left unit
group in layer V1 when pure error-driven learning
was used for training (no Hebbian injected).
Compared to the previous figure, there are many
units here that have not developed any useful feature
representation, and project equally strongly to a
large number of output units. In addition, quite a few
units are never activated during processing. This is
probably why generalization performance suffered
when Hebbian learning was excluded from the
learning mix (Equation 3; Figures 4).
5 CONCLUSIONS
In this article, we have studied the effect of mixing
error driven and Hebbian learning in bidirectional
hierarchical networks for object recognition. Error
driven learning alone is a powerful learning
mechanism which could solve the task at hand by
learning to relate individual pixels in the input
patterns to desired perceptual categories. However,
handwritten letters are intrinsically noisy as they
contain small variations due to different
handwritings, and this increases the risk for
overfitting—especially so in large networks. Hence,
there is a risk that error-driven learning might not
give optimal generalization performance for these
networks.
We run systematic training and generalization
tests on a handwritten letter recognition task using
pure error-driven learning as compared to using a
mixture of error-driven and Hebbian learning. The
simulations indicate that mixing Hebbian and error-
driven learning can be quite successful in terms of
improving the generalization performance of
bidirectional hierarchical networks in cases when
there is much noise in the input, and an increased
risk for overfitting.
Additionally, we also believe that Hebbian
learning can be a good candidate for generic, local
feature extraction for image processing and pattern
recognition tasks. In contrast to pre-wired feature
detectors, for example, Gabor-filters, Hebbian
leaning provides a more flexible means for detecting
the underlying statistical structure of the input
patterns as it has no a priori constraints on the size or
shape of these local features.
REFERENCES
Callaway, E. M., 2004. Feedforward, feedback and
inhibitory connections in primate visual cortex. Neural
Network, 17, 625-632.
Fukushima, K., 1993. Improved generalization ability
using constrained neural network architectures,
Proceedings of the International Joint Conference on
Neural Networks, 2049-2054.
Fukushima, K., 2008. Recent advances in the
neocognitron. Neural Information Processing, Lecture
Notes In Computer Science, 1041-1050,
Berlin/Heidelberg: Springer Verlag.
Hebb, D. O., 1949. The Organization of Behavior: A
Neuropsychological Theory. New York: Wiley.
O'Reilly, R. C. and Munakata, Y., 2000. Computational
Explorations in Cognitive Neuroscience:
Understanding the Mind by Simulating the Brain.
Cambridge, MA: MIT Press.
McClelland, J., 2005. How far can you go with Hebbian
learning, and when does it lead you astray? In
Munakata, Y. and Johnson, M.H. (eds) Attention and
Performance XXI: Processes of Change in Brain and
Cognitive Development. Oxford: Oxford University
Press.
O'Reilly, R., 2001. Generalization in interactive networks:
The benefits of inhibitory competition and Hebbian
learning. Neural Computation, 13, 1199-1241.
Wallis, G. and Rolls, E. T., 1997. Invariant face and object
recognition in the visual system. Progress in
Neurobiology. 51(2), 167-194.
Aisa, B., Mingus, B. and O'Reilly, R., 2008. The emergent
neural modeling system. Neural Networks, 21, 1146-
1152.
BIDIRECTIONAL HIERARCHICAL NEURAL NETWORKS - Hebbian Learning Improves Generalization
111