appropriate actions of agents as well as SLAPSOM.
They apply their method to a task of mobile robot
navigation in addition to the pursuit problem that we
take up as evaluation experiment. Their method dif-
fers from SLAPSOM in that each hunter agent ex-
cludes other hunter agents in the learning phase. They
have done a simulation of mobile robot navigation
with an only agent. And, in the simulation of the pur-
suit problem each hunter agent reinforces their actions
that they approach a prey agent, so that they exclude
other hunter agents. On the other hand, each hunter
agent learns by the SOM with including the position
of other hunter agents in SLAPSOM, so that it is pos-
sible to acquire advanced cooperative actions as am-
bush in addition to the action that is approach to a prey
agent.
The latter’s method is similar except the reinforce-
ment learning to the pursuit problem. They also pro-
pose a combination method between analytic hierar-
chy process (AHP) and the profit sharing. They show
that their method based on AHP that is superior in the
result of the early learning stage and the profit sharing
that is superior in the result of the later learning stage
help each other. Their method differs from SLAP-
SOM in excluding other hunter agents as well as the
former’s method. SLAPSOM can give operator’s in-
tuitive teaching that the operator overlooks the field
by using the coordinate data, and help acquiring co-
operative actions between hunter agents. As another
different point, their method can get only the direc-
tion that hunter agent will move, on the other hand
SLAPSOM can get the coordinate that hunter agent
will move at the next step. It is possible for SLAP-
SOM to cope with the real number environment and
acquire detailed cooperative actions potentially.
6 CONCLUSIONS
In our study, we proposed a multi-agent cooperative
method where each agent can cope with partial obser-
vation, interpolate between teaching data, and reduce
the number of them by using the Self-OrganizingMap
as supervised learning.
For evaluating our proposed method, we did two
experiments using the pursuit problem. As the results,
our proposed method helped reduction of the number
of teaching data significantly as compared with the
neural network and acquiring cooperative actions be-
tween hunter agents in the partially observable envi-
ronment.
In our future work, we have to do more complex
experiments, because the settings of this paper’s ex-
periment were relatively simple as the field size is
7 × 7 and the number of hunter agents is 2. We aim
to implement our proposed method for more compli-
cated tasks of multi-agent system such as RoboCup
Soccer Simulation. We consider making the GUI
tools that the operator can make teaching data sets
more conveniently, because we cite increase of the
operator’s work as a current problem of our proposed
method.
REFERENCES
Akiyama, H. and Noda, I. (2008). Triangulation based
approximation model for agent positioning problem.
Transactions of the Japanese Society for Artificial In-
telligence, 23(4):255–267.
Ito, A. and Kanabuchi, M. (2001). Speeding up multi-agent
reinforcement learning by coarse-graining of percep-
tion –hunter game as an example–. The transactions
of the Institute of Electronics, Information and Com-
munication Engineers, J84-D-1(3):285–293.
Katayama, K., Koshiishi, T., and Narihisa, H. (2004). Re-
inforcement learning agents with analytic hierarchy
process: a case study of pursuit problem. Trans-
actions of the Japanese Society for Artificial Intelli-
gence, 19(4):279–291.
Kohonen, T. (2001). Self-Organizing Maps(3rd ed.).
Springer.
Ono, N. and Fukumoto, K. (1997). A modular approach to
multi-agent reinforcement learning. Springer-Verlag.
Rumelhart, D. and McClelland, J. (1986). Learning inter-
nal representation by error propagation. Parallel Dis-
tributed Processing: Explorations in the Microstruc-
ture of Cognition, 1:318–362.
Tateyama, T., Kawate, S., and Oguchi, T. (2004). A teach-
ing method by using self-organizing map for rein-
forcement learning. Transactions of the Japan Society
of Mechanical Engineers, 70:1722–1729.
ICEIS 2010 - 12th International Conference on Enterprise Information Systems
372