left for future work.
It should be mentioned that our implementation of
behaviour characteristic was naive, yet even with this
basic BC our technique yielded improvement over
both the standard GA and NS.
EyAL and Quality-diversity. Since we did not op-
timize EyAL, we left direct comparison to state-of-
the-art QD methods for future work.
It should be mentioned that the principles of QD
and of EyAL are not mutually exclusive. While QD
methods use a fixed number of niches, the adaptive-
ness of EyAL can be introduced to increase and de-
crease the number of cells, or to allocate additional
computational resources to more promising niches
at the expense of less promising niches. Likewise,
the local-competition principles of QD can be intro-
duced to EyAL by various methods of fitness sharing
(McKay, 2000).
While the global competition of EyAL has been
shown to be inferior to local competition in (Colas
et al., 2020), the adaptiveness of EyAL is yet to be
explored in this context. An algorithm that exploits
both of these traits would be interesting to see.
REFERENCES
Beyer, H.-G. and Schwefel, H.-P. (2002). Evolution strate-
gies – a comprehensive introduction. Natural comput-
ing, 1(1):3–52.
Coello Coello, C. (2006). Evolutionary multi-objective op-
timization: a historical view of the field. IEEE Com-
putational Intelligence Magazine, 1(1):28–36.
Colas, C., Madhavan, V., Huizinga, J., and Clune, J. (2020).
Scaling map-elites to deep neuroevolution. In Pro-
ceedings of the 2020 Genetic and Evolutionary Com-
putation Conference, pages 67–75.
Holland, J. H. (1992). Genetic algorithms. Scientific Amer-
ican, 267(1):66–73.
Jackson, E. C. and Daley, M. (2019). Novelty search for
deep reinforcement learning policy network weights
by action sequence edit metric distance. In Proceed-
ings of the Genetic and Evolutionary Computation
Conference Companion, pages 173–174.
Koza, J. R. et al. (1992). Evolution of subsumption using
genetic programming. In Proceedings of the first Eu-
ropean conference on artificial life, pages 110–119.
MIT Press Cambridge, MA, USA.
Lehman, J. and Stanley, K. O. (2008). Exploiting open-
endedness to solve problems through the search for
novelty. In Proceedings of the Eleventh International
Conference on Artificial Life (ALIFE), Cambridge,
MA. MIT Press.
Lehman, J. and Stanley, K. O. (2011). Evolving a diver-
sity of virtual creatures through novelty search and lo-
cal competition. In Proceedings of the 13th annual
conference on Genetic and evolutionary computation,
pages 211–218.
McKay, R. I. (2000). Fitness sharing in genetic program-
ming. In GECCO, pages 435–442.
Mouret, J.-B. and Clune, J. (2015). Illuminating
search spaces by mapping elites. arXiv preprint
arXiv:1504.04909.
Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J.,
Chanan, G., Killeen, T., Lin, Z., Gimelshein, N.,
Antiga, L., et al. (2019). PyTorch: An imperative
style, high-performance deep learning library. arXiv
preprint arXiv:1912.01703.
Rechenberg, I. (1989). Evolution strategy: Natures way of
optimization. In Optimization: Methods and appli-
cations, possibilities and limitations, pages 106–126.
Springer.
Salimans, T., Ho, J., Chen, X., Sidor, S., and Sutskever, I.
(2017). Evolution strategies as a scalable alternative
to reinforcement learning. arXiv:1703.03864.
Schwefel, H.-P. (1981). Numerical optimization of com-
puter models. John Wiley & Sons, Inc.
Sipper, M., Moore, J. H., and Urbanowicz, R. J. (2019a).
Solution and fitness evolution (SAFE): A study of
multiobjective problems. In 2019 IEEE Congress on
Evolutionary Computation (CEC), pages 1868–1874.
IEEE.
Sipper, M., Moore, J. H., and Urbanowicz, R. J. (2019b).
Solution and fitness evolution (SAFE): Coevolving
solutions and their objective functions. In European
Conference on Genetic Programming, pages 146–161.
Springer.
Stanley, K. O. and Miikkulainen, R. (2002). Evolving neu-
ral networks through augmenting topologies. Evolu-
tionary computation, 10(2):99–127.
Such, F. P., Madhavan, V., Conti, E., Lehman, J., Stanley,
K. O., and Clune, J. (2017). Deep neuroevolution: Ge-
netic algorithms are a competitive alternative for train-
ing deep neural networks for reinforcement learning.
arxiv.1712.06567.
Sutton, R. S. and Barto, A. G. (2018). Reinforcement Learn-
ing: An Introduction. MIT press, 2nd edition.
Todorov, E., Erez, T., and Tassa, Y. (2012). Mujoco:
A physics engine for model-based control. In 2012
IEEE/RSJ International Conference on Intelligent
Robots and Systems, pages 5026–5033. IEEE.
ECTA 2022 - 14th International Conference on Evolutionary Computation Theory and Applications
150