Figure 11: Robot’s path carrying out the visual wall fol-
lowing behaviour on a real environment (right wall). The
map is a certainty grid created with a LMS200 laser (white
pixels are free space, black obstacles and grey uncertainty).
out on the 3-D Gazebo simulator. A wall following
test in a real environment is also shown.
Using a fisheye camera the environment is per-
ceived in 3D, and it is possible to avoid obstacles that
are invisible to other sensors which are more common
in mobile robotics (laser or ultrasounds).
In the proposed methodology the states are de-
fined directly from the image after a simple prepro-
cessing (Sobel or colour filtering) with no calibration
process. With a 3x3 grid, we can define a state space
of only 64 states. It has a certain degree of “percep-
tual aliasing”, but RL algorithm converges. We have
also tested grids of different dimensions with similar
results but greater convergence time. A delicate bal-
ance need be struck between reducing the number of
states and avoiding “perceptual aliasing”.
The proposed codification and methodology is
general, not specific for the task, and has proved to
be efficient and valid, and easy to adapt to distint be-
haviours. The system works with different types of
reinforcement and filtering.
Various tests were carried out to verify the robust-
ness of the learned behaviours. We used obstacles that
were not detected by the laser device, and walls with
gaps. In both cases the system generalized perfectly
and the results were optimal. If the gaps in the walls
were large (over 40 cm) a large number of new states
appeared with respect to the training process, and the
final result deteriorated.
Future lines of work include on-line real robot
learning, the integration of several behaviours (e.g.
follow objects and avoid obstacles) and establishing
a mechanism for automatically defining the states of
RL (e.g. neural networks).
ACKNOWLEDGEMENTS
This paper was supported in part by the Xunta
de Galicia and Spanish Government under Grants
PGIDIT04-TIC206011PR, TIN2005-03844, and de-
partment colaboration grant (B.O.E. 16-06-2006).
REFERENCES
Boada, M., Barber, R., and Salichs, M. (2002). Visual ap-
proach skill for a mobile robot using learning and fu-
sion of simple skills. Robotics and Autonomous Sys-
tems, 38:157–170.
Gaskett, C., Fletcher, L., and Zelinsky, A. (2000). Rein-
forcement learning for a vision based mobile robot. In
Proc. IEEE/RSJ Int. Conf. on Intelligent Robots and
Systems (IROS), volume 1, pages 403–409.
Mart
´
ınez-Mar
´
ın, T. and Duckett, T. (2005). Fast reinforce-
ment learning for vision-guided mobile robots. In
IEEE Int. Conf. on Robotics and Automation (ICRA),
pages 4170–4175.
Michels, J., Saxena, A., and Ng, A. (2005). High speed
obstacle avoidance using monocular vision and rein-
forcement learning. In Proc. Int. Conf. on Machine
Learning.
Mill
´
an, J. R., Posenato, D., and Dedieu, E. (2002).
Continuous-action Q-Learning. Machine Learning,
49:247, 265.
Moreno, D., Regueiro, C., Iglesias, R., and Barro, S.
(2004). Using prior knowledge to improve reinforce-
ment learning in mobile robotics. In Towards Au-
tonomous Robotic Systems (TAROS).
Nakamura, T. and Asada, M. (1995). Motion sketch: Ac-
quisition of visual motion guided behaviors. In IJCAI,
pages 126–132.
Nehmzow, U. (1999). Vision processing for robot learning.
Industrial Robot, 26(2):121–130.
Regueiro, C., Domenech, J., Iglesias, R., and Correa, J.
(2006). Acquiring contour following behaviour in
robotics through q-learning and image-based states. In
Proc. XV Int. Conf. on Computer and Information Sci-
ences Engineering (CISE), pages 403–409.
Regueiro, C., Rodr
´
ıguez, M., Correa, J., Moreno, D., Igle-
sias, R., and Barro, S. (2002). A control architecture
for mobile robotics based on specialists. In Leondes,
C., editor, Intelligent Systems: Technology and Appli-
cations, volume 6, pages 337–360. CRC Press.
Ruiz, J., Montero, P., Mart
´
ın, F., and Matell
´
an, V. (2005).
Vision based behaviors for a legged robot. In Proc.
Workshop en Agentes F
´
ısicos (WAF), pages 59–66.
Wyatt, J. (1995). Issues in putting reinforcement learning
onto robots. In 10th Biennal Conference of the AISB,
Sheffield, UK.
ICINCO 2007 - International Conference on Informatics in Control, Automation and Robotics
156