BEHAVIOUR NAVIGATION LEARNINIG USING FACL
ALGORITHM
Abdelkarim Souissi and Hacene Rezine
EMP , Bordj Elbahri, Alger, Algerie
Keywords: Mobile Robot Navigation, Reactive Navigation, Fuzzy Control, Reinforcement learning, Fuzzy Actor Critic
Learning.
Abstract: In this article, we are interested in the reactive behaviours navigation training of a mobile robot in an
unknown environment. The method we will suggest ensures navigation in unknown environments with
presence off different obstacles shape and consists in bringing the robot in a goal position, avoiding
obstacles and releasing it from the tight corners and deadlock obstacles shape. In this framework, we use
the reinforcement learning algorithm called Fuzzy Actor-Critic learning, based on temporal difference
prediction method. The application was tested in our experimental PIONEER II platform.
1 INTRODUCTION
In this article, we propose a reinforcement training
method where the apprentice explores actively its
environment. It applies various actions in order to
discover the states causing the emission of rewards
and punishments. The agent must find the action
which it must carry out when it is in a given
situation It must learn how to choose the optimal
actions to achieve the fixed goal. The environment
can punish or reward the system according to the
applied actions. Each time that an agent applies an
action, a critic gives him a reward or a penalty to
indicate if the resulting state is desirable or not
(Sutton, 1998), (Glorennec, 2000). The task of the
agent is to learn using these rewards the continuation
of actions which gets the greatest cumulative
reward.
Mobile robotics is a privileged application field
of t
he training by reinforcement (Fujii, 1998),
(Smart, 2002), (Babvey, 2003). This established fact
is related to the growing place which takes, since a
few years, an autonomous robotics without
knowledge of the environment. The goal is then to
regard behaviour as a function of mapping sensor-
effector. The training in robotics consists of the
automatic modification of the behaviour of the robot
to improve its behaviour in its environment. The
behaviours are synthesized starting from the simple
definition of objectives through a reinforcement
function.
The considered approach of the robot navigation
using fuzzy i
nference as apprentice is ready to
integrate certain errors in the information about the
system. For example, with fuzzy logic we can
process vague data. The perception of the
environment by ultrasounds sensors and the
reinforcement training thus prove to be particularly
well adapted one to the other (Beom, 1995), (Fukuda
1995), (Jouffe, 1997), (Faria, 2000).
It is very difficult to determinate correct
co
nclusions manually in a large base rule FIS to
ensure the releasing from tight corner and deadlock
obstacles, even when we use a gradient descent
method or a potential-field technique due to the
local-minimum problem. In such situations the robot
will be blocked.
Behaviours made up of a fusion of a «goal
seeki
ng» and of an "obstacle avoidance» issues are
presented. The method we will suggest ensures
navigation in unknown environments with presence
off different obstacles shape, the behaviour will be
realised with SIF whose conclusions are determined
by reinforcement training methods. The algorithms
are written using the Matlab software after having
integrated, in a Simulink block, the functions of
perception, localization and motricity of the robot.
339