presented in contiguity with an alarmed
demonstrator, the unconditioned stimulus. More
importantly, there are properties of socially acquired
predator avoidance (e.g., the intensity of the
unconditioned response increases with that of the
unconditioned stimulus, and the fact that there is
preferential learning about particular types of
stimuli) that provide support of the idea that socially
acquired behaviours are mediated by individual
learning processes and not by independent social
learning mechanisms.
This line of research is complementary to the
work done in imitation in the Artificial Intelligence
community. Such approach has used social learning
theories from psychology to develop adaptive agents
that learn from others by observing their behaviour.
In particular, (Mataric, 1994) has used vicarious
reinforcement to deal with the Credit Assignment
Problem.
5 CONCLUSIONS
It is our contention that the proposal outlined in this
position paper will strengthen the connection
between the study of computational and biological
systems. In particular, the approach we advocate will
contribute to answering the question of how
psychological concepts such as motivation, attention
and intention can be modelled in artificial organisms
to affect adaptive behavioural modifications and
control.
Reinforcement learning algorithms have
successfully been applied to simple domains in areas
such as navigation robotics, manufacturing, and
process control. More powerful algorithms will, no
doubt, benefit larger scenarios in industrial
applications such as telecommunications systems,
air traffic control, traffic and transportation
management, information filtering and gathering,
electronic commerce, business process management,
entertainment, and medical care.
REFERENCES
Alonso, E., (2002). AI and Agents: State of the Art. AI
Magazine, 23, 25-29.
Bowling, M. and Veloso, M., (2001), Rational and
convergent learning in stochastic games. In
Proceedings of the Seventeenth International Joint
Conference on Artificial Inelligence (IJCAI-2001),
pages 1021-1026, Seattle, WA.
Claus, C. and Boutilier, C., (1998), The dynamics of
reinforcement learning in cooperative multiagent
systems. In Proceedings of the Fifteenth National
Conference on Artificial Intelligence, pages 746-752.
Greenwald, A., Hall, K., and Serrano, R., (2002),
Correlated-Q learning. In NIPS Workshop on
Multiagent Learning.
Griffin, A. S., (2004), Social learning about predators: A
review and prospectus. Learning & Behavior 32(1),
131-140.
Hu, J. and Wellman, M., (2001), Learning about other
agents in a dynamic multiagent system. Journal of
Cognitive Systems Research, 2:67-79.
Littman, M. L., (2001), Friend-or-foe Q-learning in
general-sum games. In Proceedings of the Eighteen
International Conference on Machine Learning.
Littman, M. L., (1994), Markov games as a framework for
multi-agent reinforcement learning. In Proceedings of
the 11
th
International Conference on Machine
Learning, 157-163.
Mataric, M., (1994), Learning to behave socially. In
Proceedings of the Third International Conference on
Simulation and Adaptive Behavior.
Sen , S., Sekaran, M., and Hale, J., (1994), Learning to co-
ordinate without sharing information. In Proceedings
of the Twelfth National Conference on Artificial
Intelligence, 426-431, Seattle, WA.
Shoham, Y., Powers, R., and Grenager, T., (2003), Multi-
Agent Reinforcement Learning: a critical survey.
Technical Report.
Stone, P. and Veloso, M., (2000), Multiagent Systems: A
Survey from a Machine Learning Perspective,
Autonomous Robots, Volume 8( 3), 345-383.
Suematsu, N. and Hayashi, A., (2002), A multiagent
reinforcement learning algorithm using extended
optimal response. In Proceedings of the Fifteenth
International Joint Conference on Artificial
Intelligence (IJCAI-97), 370-377.
Sutton, R. S. and Barto, A. G., (1998), Reinforcement
Learning: An Introduction, Cambridge, MA: The MIT
Press.
Wang, X. and Sandholm, T., (2002), Reinforcement
learning to play an optimal Nash equilibrium in team
Markov games. In Advances in Neural Information
Processing Systems (NIPS-2002).
ICAART2013-InternationalConferenceonAgentsandArtificialIntelligence
146