from experience in a continuous way when running,
this methodology has the potential to be an adaptable
system where we can add or remove any sensors or
actuators, and the controller can adapt autonomously
and online to the new situation.
8 FURTHER WORK
The different parameters that define the speed of
adapting connection weights and the way of creating
new neurons and connections have to be investigated
further to evaluate our novel methodology for creat-
ing controllers for concurrent tasks. These investiga-
tions will lead us to find an elaborate but still very
basic “artificial brain” model that enables a system to
achieve a sophisticated level compared to other artifi-
cial intelligence models by learning from experience
efficiently.
When the basic methods are investigated in detail,
some extensions can be added like Spike Time Depen-
dent Plasticity or a feedback prediction mechanism.
Initial ideas for both enhancements were discussed in
this paper. Those improvements would help the con-
trolled systems to deal with more complex situations,
especially when timing considerations are important.
As mentioned in section 3 assigning delayed feed-
back more efficiently or even predicting feedback will
be an interesting research issue for future work. The
idea is that a neuron that receives positive or nega-
tive reward very often when it is active will probably
receive the same reward also in the future. Predict-
ing reward could actually be one reason for producing
reward. This earlier reward may now be correlated
to the activity of another neuron. That neuron could
again produce reward when predicting it. By the re-
cursive process reward could potentially be predicted
progressively earlier.
REFERENCES
Daucé, E. and Henry, F. (2006). Hebbian learning in large
recurrent neural networks. Technical report, Move-
ment and Perception Lab, Marseille.
Elizondo, D., Fiesler, E., and Korczak, J. (1995). Non-
ontogenetic sparse neural networks. In International
Conference on Neural Networks 1995, IEEE, vol-
ume 26, pages 290–295.
Florian, R. V. (2005). A reinforcement learning algorithm
for spiking neural networks. In Proceedings of the
Seventh International Symposium on Symbolic and
Numeric Algorithms for Scientific Computing, pages
299–306.
Gers, F. A., Schraudolph, N. N., and Schmidhuber, J.
(2002). Learning precise timing with LSTM recurrent
networks. Journal of Machine Learning Research,
3:115–143.
Gómez, G., Lungarella, M., Hotz, P. E., Matsushita, K.,
and Pfeifer, R. (2004). Simulating development in
a real robot: On the concurrent increase of sen-
sory, motor, and neural complexity. In Proceedings
of the Fourth International Workshop on Epigenetic
Robotics, pages 119–122.
Hebb, D. O. (1949). The Organization of Behaviour: A
Neuropsychological Approach. John Wiley & Sons,
New York.
Izhikevich, E. M. (2006). Polychronization: Computation
with spikes. Neural Computation, 18:245–282.
Izhikevich, E. M. (2007). Solving the distal reward prob-
lem through linkage of STDP and dopamine signaling.
Cerebral Cortex, 10:1093–1102.
Katic, D. (2006). Leaky-Integrate-and-Fire und Spike Re-
sponse Modell. Technical report, Institut fï¿
1
2
r Tech-
nische Informatik, Universitï¿
1
2
t Karlsruhe.
Liu, J. and Buller, A. (2005). Self-development of motor
abilities resulting from the growth of a neural network
reinforced by pleasure and tension. In Proceedings of
the 4th International Conference on Development and
Learning 2005, pages 121–125.
Liu, J., Buller, A., and Joachimczak, M. (2006). Self-
motivated learning agent: Skill-development in a
growing network mediated by pleasure and tensions.
Transactions of the Institute of Systems, Control and
Information Engineers, 19(5):169–176.
van Leeuwen, M. (2004). Spike timing dependent structural
plasticity in a single model neuron. Master’s thesis,
Intelligent Systems Group, Institute for Information
and Computing Sciences, Utrecht University.
Vreeken, J. (2003). Spiking neural networks, an introduc-
tion. Technical report, Intelligent Systems Group,
Institute for Information and Computing Sciences,
Utrecht University.
SELF CONSTRUCTING NEURAL NETWORK ROBOT CONTROLLER BASED ON ON-LINE TASK
PERFORMANCE FEEDBACK
333