Figure 3: The class diagram of the simulator.
// Longitudinal Automaton
public double reward(int action){
int combine;
combine=Math.max(speedModule(action),
frontModule(action));
if (combine = = 2) combine = 0;
return combine;
}
// Lateral Automaton
public double reward(int action){
int combine;
combine=Math.max(
leftRightModule(action),
frontModule(action));
return combine;
}
Figure 4: The physical environment response.
8 CONCLUSIONS
Reinforcement learning has attracted rapidly
increasing interest in the machine learning and
artificial intelligence communities. Its promise is
beguiling - a way of programming agents by reward
and punishment without needing to specify how the
task (i.e., behavior) is to be achieved. Reinforcement
learning allows, at least in principle, to bypass the
problems of building an explicit model of the
behavior to be synthesized and its counterpart, a
meaningful learning base (supervised learning).
The reinforcement scheme presented in this
paper satisfies all necessary and sufficient conditions
for absolute expediency in a stationary environment
and the nonlinear algorithm based on this scheme is
found to converge to the ”optimal” action faster than
nonlinear schemes previously defined in (Ünsal,
1999).
Using this new reinforcement scheme was
developed a simulator for an Intelligent Vehicle
Control System, in a multi-agent approach. The
entire system was implemented in Java, and is based
on JADE platform
REFERENCES
Baba, N., 1984. New Topics in Learning Automata:
Theory and Applications, Lecture Notes in Control
and Information Sciences Berlin, Germany: Springer-
Verlag.
Barto, A., Mahadevan, S., 2003. Recent advances in
hierarchical reinforcement learning, Discrete-Event
Systems journal, Special issue on Reinforcement
Learning.
Bigus, J. P., Bigus, J., 2001. Constructing Intelligent
Agents using Java, 2nd ed., John Wiley & Sons, Inc.
Buffet, O., Dutech, A., Charpillet, F., 2001. Incremental
reinforcement learning for designing multi-agent
systems, In J. P. Müller, E. Andre, S. Sen, and C.
Frasson, editors, Proceedings of the Fifth
International Conference onAutonomous Agents, pp.
31–32, Montreal, Canada, ACM Press.
Lakshmivarahan, S., Thathachar, M.A.L., 1973.
Absolutely Expedient Learning Algorithms for
Stochastic Automata, IEEE Transactions on Systems,
Man and Cybernetics, vol. SMC-6, pp. 281-286.
Moody, J., Liu, Y., Saffell, M., Youn, K., 2004. Stochastic
direct reinforcement: Application to simple games
with recurrence, In Proceedings of Artificial
Multiagent Learning. Papers from the 2004 AAAI Fall
Symposium,Technical Report FS-04-02.
Narendra, K. S., Thathachar, M. A. L., 1989. Learning
Automata: an introduction, Prentice-Hall.
Rivero, C., 2003. Characterization of the absolutely
expedient learning algorithms for stochastic automata
in a non-discrete space of actions, ESANN'2003
proceedings - European Symposium on Artificial
Neural Networks Bruges (Belgium), ISBN 2-930307-
03-X, pp. 307-312
Stoica, F., Popa, E. M., 2007. An Absolutely Expedient
Learning Algorithm for Stochastic Automata, WSEAS
Transactions on Computers, Issue 2, Volume 6, ISSN
1109-2750, pp. 229-235.
Sutton, R., Barto, A., 1998. Reinforcement learning: An
introduction, MIT-press, Cambridge, MA.
Ünsal, C., Kachroo, P., Bay, J. S., 1999. Multiple
Stochastic Learning Automata for Vehicle Path
Control in an Automated Highway System, IEEE
Transactions on Systems, Man, and Cybernetics -part
A: systems and humans, vol. 29, no. 1, january 1999.
ICE-B 2008 - International Conference on e-Business
50