Table 1: Planning Accuracies using the two planning
schemes. Global planning is better.
Config Plan (%)
Learn (global) 100
Learn (heuristic) 89.3± 6.7
Table 2: Localization Accuracies using the two planning
schemes. Global planning is better.
Config Localization error
X (cm) Y (cm) θ (deg)
Learn (global) 7.6± 3.7 11.1± 4.8 9± 6.3
Learn (heuristic) 11.6± 5.1 15.1± 7.8 11± 9.7
Hand-labeled 6.9± 4.1 9.2± 5.3 7.1± 5.9
global planning scheme to generate robust plans, and
the replanning feature of the heuristic approach can
be used when the plan fails due to unforeseen reasons.
The online color learning process takes a simi-
lar amount of time with either planning scheme (≈ 6
minutes of robot effort) instead of more than two
hours of human effort. The initial training of the mod-
els (in global planning) takes 1-2 hours, but it pro-
ceeds autonomously and needs to be done only once
for each environment. The heuristic planning scheme,
on the other hand, requires manual parameter tuning
over a few days, which is sensitive to minor environ-
mental changes.
6 CONCLUSIONS
The potential of mobile robots can be exploited
in real-world applications only if they function au-
tonomously. For mobile robots equipped with color
cameras, two major challenges are the manual cali-
bration and the sensitivity to illumination. Prior work
has managed to learn a few distinct colors (Thrun,
2006), model known illuminations (Rosenberg et al.,
2001), and use heuristic action sequences to facilitate
learning (Sridharan and Stone, 2007).
We present an algorithm that enables a mobile
robot to autonomously model its motion errors and
the feasibility of learning different colors at different
poses, thereby maximizing color learning opportuni-
ties while minimizing localization errors. The global
action selection provides robust performance that is
significantly better than that obtained with manually
tuned heuristics.
Both planning schemes require the environmen-
tal structure as input, which is easier to provide than
hand-labeling several images. One challenge is to
combine this work with autonomous vision-based
map building (SLAM) (Jensfelt et al., 2006). We also
aim to extend our learning approach to smoothly de-
tect and adapt to illumination changes, thereby mak-
ing the robot operate with minimal human supervision
under natural conditions.
REFERENCES
Anzani, F., Bosisio, D., Matteucci, M., and Sorrenti, D.
(2005). On-line color calibration in non-stationary en-
vironments. In RoboCup Symposium.
Bishop, C. M. (1995). Neural Networks for Pattern Recog-
nition. Oxford University Press.
Boutillier, C., Dean, T., and S.Hanks (1999). Decision the-
oretic planning: structural assumptions and computa-
tional leverage. Journal of AI Research, 11:1–94.
Cameron, D. and Barnes, N. (2003). Knowledge-based au-
tonomous dynamic color calibration. In The Interna-
tional RoboCup Symposium.
Cohen, D., Ooi, Y. H., Vernaza, P., and Lee, D. D. (2004).
UPenn TDP, RoboCup-2003: RoboCup Competitions
and Conferences.
Comaniciu, D. and Meer, P. (2002). Mean shift: A robust
approach toward feature space analysis. PAMI.
Ghallab, M., Nau, D., and Traverso, P. (2004). Automated
Planning: Theory and Practice. Morgan Kaufmann,
San Francisco, CA 94111.
Jensfelt, P., Folkesson, J., Kragic, D., and Christensen, H. I.
(2006). Exploiting distinguishable image features in
robotic mapping and localization. In The European
Robotics Symposium (EUROS).
Jungel, M. (2004). Using layered color precision for a
self-calibrating vision system. In The International
RoboCup Symposium.
Maloney, L. T. and Wandell, B. A. (1986). Color Con-
stancy: A Method for Recovering Surface Spectral
Reflectance. Opt. Soc. of Am. A, 3(1):29–33.
Minten, B. W., Murphy, R. R., Hyams, J., and Micire,
M. (2001). Low-order-complexity vision-based dock-
ing. IEEE Transactions on Robotics and Automation,
17(6):922–930.
Pineau, J., Montemerlo, M., Pollack, M., Roy, N., and
Thrun, S. (2003). Towards robotic assistants in nurs-
ing homes: Challenges and results. RAS Special Issue
on Socially Interactive Robots.
Rosenberg, C., Hebert, M., and Thrun, S. (2001). Color
constancy using kl-divergence. In The IEEE Interna-
tional Conference on Computer Vision (ICCV).
Schulz, D. and Fox, D. (2004). Bayesian color estimation
for adaptive vision-based robot localization. In IROS.
Shi, J. and Malik, J. (2000). Normalized cuts and image
segmentation. In IEEE Transactions on PAMI.
Sridharan, M. and Stone, P. (2007). Color learning on a
mobile robot: Towards full autonomy under changing
illumination. In IJCAI.
Thrun, S. (2006). Stanley: The Robot that Won the
DARPA Grand Challenge. Journal of Field Robotics,
23(9):661–692.
LONG-TERM VS. GREEDY ACTION PLANNING FOR COLOR LEARNING ON A MOBILE ROBOT
685