rare or if models are only moderately well defended.
However, the impact of taste sampling is non-linear
especially in systems with highly defended models.
In such situations taste-sampling lowers
r *
.
Consequently, under the assumption of a fixed value
for
r *
and stability, a predator evolves a taste-
sampling strategy because mimics are less common
or models are better defended than in a comparable
stable environment where predators do not utilize
taste sampling.
Another interesting aspect is the effect of
different age distributions: in general longevity in
predators increases
r *
. The effects are linear with
regards to prey abundance but non-linear with
regards to prey toxicity where behavioural
expenditure gains increasing impact in the case of
defended prey and older predators, whereas
metabolic costs have an increased impact in the case
of non-defended prey. The main conclusions of this
paper are as follows:
On the predator’s side
r *
is related to the
nutritional value of prey and on the prey’s side
it relates to an energy inventory which can be
allocated, amongst other things, towards the
cost of defences or reproduction;
Behavioural expenditure has a greater impact
than metabolic costs when prey is rare and
undefended;
Metabolic costs have a greater impact when
prey is abundant or highly defended;
Longevity of the predator increases the
importance of behavioural expenditure in the
case of highly defended prey and the impact of
metabolic costs if prey is undefended;
Mimics generally lower
r *
which leads to less
nutritional prey or better defended models if
r *
is meant to be unchanged;
Predators utilize taste sampling if mimics are
rare or models are highly toxic.
REFERENCES
Alonso, E., Fairbank, M., and Mondragón, E. (2015).
Back to optimality: A formal framework to express the
dynamics of learning optimal behavior. Adaptive
Behavior, 23(4), 206-215.
Barto, A.G., Sutton, R.S., and Watkins, C.J.C.H. (1990).
Learning and sequential decision making. In Learning
and Computational Neuroscience: Foundations of
Adaptive Networks, M. Gabriel and J.W. Moore, Eds.,
pp. 539-602, Cambridge, Mass: MIT Press.
Dayan, P., and Daw, N.D. (2008). Decision theory,
reinforcement learning, and the brain, Cognitive,
Affective, & Behavioral Neuroscience 8, 429–453.
Dingemanse, N. J., and Réale, D. (2005). Natural selection
and animal personality, Behavior 142, 1159–1184.
Orr, H. A., (2009). Fitness and its role in evolutionary
genetics. Nature Review Genetics 10, 531-539.
Rangel, A., Camerer, C., and Montague, P.R. (2008). A
framework for studying the neurobiology of value-
based decision making, Nature Reviews Neuroscience
9, 545–556.
Schultz, W. (2008). Neuroeconomics: the promise and the
profit, Philosophical Transactions of the Royal Society
B: Biological Sciences 363, 3767–3769.
Staddon, J.E. (2007). Is animal behavior optimal? In A.
Bejan & G.W. Merkx (eds.) Constructal Theory of
Social Dynamics, NY: Springer.
Sutton, R.S., and Barto, A.G. (1998). Reinforcement
learning: An introduction, Boston, MA: Cambridge
University Press.
Teichmann, J. (2014). Models of aposematism and the role
of aversive learning. PhD dissertation, City University
London, London, UK.
Teichmann, J., Broom, M., and Alonso, E. (2014). The
application of temporal difference learning in optimal
diet models, Journal of Theoretical Biology 340, 11–
16.