contained an unusual stack of three convolutional
filter layers that use ELU, linear and linear activation
functions, and only one fully connected classification
layer. This example suggests that a NAS working on
a wider search space can find interesting designs that
the experimenters had not considered before.
There is a large potential for future improvement
by exploring variations of step size control and
mutation operators. This is a step towards more
general and more efficient NAS methods. If NAS can
be improved further and applied to even wider search
spaces, perhaps in the future, it can find surprising
new architectural improvements.
REFERENCES
Assunção, F., Lourenço, N., Machado, P., & Ribeiro, B.
(2017). Towards the evolution of multi-layered neural
networks: A dynamic structured grammatical evolution
approach. 393–400. https://doi.org/10.1145/
3071178.3071286
Assunção, F., Lourenço, N., Machado, P., & Ribeiro, B.
(2018). Evolving the Topology of Large Scale Deep
Neural Networks (pp. 19–34).
https://link.springer.com/chapter/10.1007/978-3-319-
77553-1_2
Assunção, F., Lourenço, N., Machado, P., & Ribeiro, B.
(2019). Fast DENSER: Efficient Deep NeuroEvolution.
In Lecture Notes in Computer Science (pp. 197–212).
Springer International Publishing.
https://dx.doi.org/10.1007/978-3-030-16670-0_13
Cramer, N. L. (1985). A Representation for the Adaptive
Generation of Simple Sequential Programs. In
Proceedings of the First International Conference on
Genetic Algorithms and their Applications. Psychology
Press.
Droste, S., & Wiesmann, D. (2000). Metric Based
Evolutionary Algorithms. 29–43.
Hansen, N., & Ostermeier, A. (2001). Completely
Derandomized Self-Adaptation in Evolution Strategies.
Evolutionary Computation, 9(2), 159–195.
https://doi.org/10.1162/106365601750190398
Holland, J. H. (1992). Adaptation in Natural and Artificial
Systems: An Introductory Analysis with Applications to
Biology, Control, and Artificial Intelligence. The MIT
Press. https://direct.mit.edu/books/book/2574/
Adaptation-in-Natural-and-Artificial-SystemsAn
Ioffe, S., & Szegedy, C. (2015). Batch Normalization:
Accelerating Deep Network Training by Reducing
Internal Covariate Shift. arXiv Pre-Print Server.
https://arxiv.org/abs/1502.03167
Koza, J. R. (1989). Hierarchical genetic algorithms
operating on populations of computer programs.
Proceedings of the 11th International Joint Conference
on Artificial Intelligence - Volume 1, 768–774.
Li, R., Emmerich, M., Eggermont, J., Bäck, T., Schütz, M.,
Dijkstra, J., & Reiber, J. (2013). Mixed Integer
Evolution Strategies for Parameter Optimization.
Evolutionary Computation.
https://doi.org/10.1162/EVCO_a_00059
Loshchilov, I., & Hutter, F. (2016). CMA-ES for
Hyperparameter Optimization of Deep Neural
Networks.
Loshchilov, I., Schoenauer, M., & Sèbag, M. (2013). Bi-
population CMA-ES agorithms with surrogate models
and line searches. Proceedings of the 15th Annual
Conference Companion on Genetic and Evolutionary
Computation, 1177–1184.
https://doi.org/10.1145/2464576.2482696
Ostermeier, A., Gawelczyk, A., & Hansen, N. (1994). A
derandomized approach to self-adaptation of evolution
strategies. Evolutionary Computation, 2(4), 369–380.
Rechenberg, I. (1965). Cybernetic solution path of an
experimental problem. Roy. Aircr. Establ., Libr.
Transl., 1122. https://cir.nii.ac.jp/crid/
1570291225013966592
Rechenberg, I. (1994). Evolutionsstrategie. Frommann-
Holzboog.
Ryan, C., Collins, J., & Neill, M. O. (1998). Grammatical
evolution: Evolving programs for an arbitrary language.
In W. Banzhaf, R. Poli, M. Schoenauer, & T. C. Fogarty
(Eds.), Genetic Programming (pp. 83–96). Springer.
https://doi.org/10.1007/BFb0055930
Schwefel, H.-P. (1995). Evolution and optimum seeking.
Sixth-generation computer technology series.
Storn, R., & Price, K. (1997). Differential evolution–a
simple and efficient heuristic for global optimization
over continuous spaces. Journal of Global
Optimization, 11(4), 341–359.
https://doi.org/10.1023/A:1008202821328
Sultan, A. (2022). A Modern Variant of LeNet by TF.keras
on MNIST. https://www.kaggle.com/code/abedsultan/
a-modern-variant-of-lenet-by-tf-keras-on-mnist
White, C., Safari, M., Sukthanker, R., Ru, B., Elsken, T.,
Zela, A., Dey, D., & Hutter, F. (2023). Neural
Architecture Search: Insights from 1000 Papers
(arXiv:2301.08727). arXiv. https://doi.org/10.48550/
arXiv.2301.08727
Yang, A., Esperança, P., & Carlucci, F. (2019). NAS
evaluation is frustratingly hard.
Yu, K., Suito, C., Jaggi, M., Musat, C.-C., & Salzmann, M.
(Eds.). (2020). Evaluating the search phase of neural
architecture search. ICRL 2020 Eighth International
Conference on Learning Representations.
Step Size Control in Evolutionary Algorithms for Neural Architecture Search
295