BEHAVIOUR NAVIGATION LEARNINIG USING FACL 
ALGORITHM 
Abdelkarim Souissi and Hacene Rezine 
EMP , Bordj Elbahri, Alger, Algerie 
Keywords:  Mobile Robot Navigation, Reactive Navigation, Fuzzy Control, Reinforcement learning, Fuzzy Actor Critic 
Learning. 
Abstract:  In this article, we are interested in the reactive behaviours navigation training of a mobile robot in an 
unknown environment. The method we will suggest ensures navigation in unknown environments with 
presence off different obstacles shape and consists in bringing the robot in a goal position, avoiding 
obstacles and releasing it from the tight corners and deadlock obstacles shape.  In this framework, we use 
the reinforcement learning algorithm called Fuzzy Actor-Critic learning, based on temporal difference 
prediction method.  The application was tested in our experimental PIONEER II platform. 
1 INTRODUCTION 
In this article, we propose a reinforcement training 
method where the apprentice explores actively its 
environment. It applies various actions in order to 
discover the states causing the emission of rewards 
and punishments.  The agent must find the action 
which it must carry out when it is in a given 
situation It must learn how to choose the optimal 
actions to achieve the fixed goal. The environment 
can punish or reward the system according to the 
applied actions. Each time that an agent applies an 
action, a critic gives him a reward or a penalty to 
indicate if the resulting state is desirable or not 
(Sutton, 1998), (Glorennec, 2000).  The task of the 
agent is to learn using these rewards the continuation 
of actions which gets the greatest cumulative 
reward.  
Mobile robotics is a privileged application field 
of t
he training by reinforcement (Fujii, 1998), 
(Smart, 2002), (Babvey, 2003). This established fact 
is related to the growing place which takes, since a 
few years, an autonomous robotics without 
knowledge of the environment. The goal is then to 
regard behaviour as a function of mapping sensor-
effector. The training in robotics consists of the 
automatic modification of the behaviour of the robot 
to improve its behaviour in its environment. The 
behaviours are synthesized starting from the simple 
definition of objectives through a reinforcement 
function.   
The considered approach of the robot navigation 
using fuzzy i
nference as apprentice is ready to 
integrate certain errors in the information about the 
system. For example, with fuzzy logic we can 
process vague data. The perception of the 
environment by ultrasounds sensors and the 
reinforcement training thus prove to be particularly 
well adapted one to the other (Beom, 1995), (Fukuda 
1995), (Jouffe, 1997), (Faria, 2000).  
It is very difficult to determinate correct 
co
nclusions manually in a large base rule FIS to  
ensure  the releasing from tight corner and deadlock 
obstacles, even when we use a gradient descent 
method  or a potential-field technique due to the 
local-minimum problem. In such situations the robot 
will be blocked.   
Behaviours made up of a fusion of a «goal 
seeki
ng» and of an "obstacle avoidance» issues are 
presented.  The method we will suggest ensures 
navigation in unknown environments with presence 
off different obstacles shape, the behaviour will be 
realised with SIF whose conclusions are determined 
by reinforcement training methods. The algorithms 
are written using the Matlab software after having 
integrated, in a Simulink block, the functions of 
perception, localization and motricity of the robot. 
339