
6 OUTLOOK
Further studies will focus on scaling of the observed
phenomena with the size of the dataset - a very impor-
tant question for real applications and related to en-
ergy saving and sustainability questions. It may well
be that for certain data it is useless to increase the size
of the data, because model performance is already in
saturation. Further, we will study the shape of the
transition in detail for a statistically representative set
of models, e.g. CNN vs. RNN vs. perceptrons. Lastly
we will extend the study to more diverse datasets.
Information theory provides definitions and inter-
pretations of quantities also known in statistical me-
chanics, such as entropy, in a way that is readily us-
able in the context of our experiments. A future goal
concerns theoretical work on universality of phase
transitions and scaling laws for (model, data) tuples
and their representation with respect to given tasks,
like classification and regression. The goal is to uni-
versally compare models and datasets to one another
with respect to robustness, to infer both dataset com-
plexity and model robustness, ultimately defining uni-
versality classes, as are well known from the theory of
phase transitions.
Eventually, we expect that studies on the finite size
effect, i.e. little data and/or small networks play a cru-
cial role. Our final aim is to provide a tool to deter-
mine the optimum network size and data size for a
problem given, in particular when data quality is un-
known.
ACKNOWLEDGEMENTS
We acknowledge fruitful discussions with F. Em-
merich, M. Quade, and M. Schultz. This work is sup-
ported by the KI:STE project, grant no. 67KI2043 by
the German ministry for ecology. This research was
supported in part by the National Science Foundation
under Grant No. NSF PHY-1748958.
REFERENCES
Advani, M., Lahiri, S., and Ganguli, S. (2013). Statisti-
cal mechanics of complex neural systems and high
dimensional data. Journal of Statistical Mechanics:
Theory and Experiment, 2013(03):P03014.
Al-Jarrah, O. Y., Yoo, P. D., Muhaidat, S., Karagiannidis,
G. K., and Taha, K. (2015). Efficient machine learning
for big data: A review. Big Data Research, 2(3):87–
93. Big Data, Analytics, and High-Performance Com-
puting.
Bahri, Y., Kadmon, J., Pennington, J., Schoenholz, S. S.,
Sohl-Dickstein, J., and Ganguli, S. (2020). Statistical
mechanics of deep learning. Annual Review of Con-
densed Matter Physics, 11:501–528.
Baldominos, A., Saez, Y., and Isasi, P. (2019). A survey
of handwritten character recognition with MNIST and
EMNIST. Applied Sciences (Switzerland), 9.
Biehl, M., Ahr, M., and Schl
¨
osser, E. (2007). Statistical
physics of learning: Phase transitions in multilayered
neural networks. Advances in Solid State Physics 40,
pages 819–826.
Boncelet, C. (2009). Chapter 7 - Image Noise Models. In
Bovik, A., editor, The Essential Guide to Image Pro-
cessing, pages 143–167. Academic Press, Boston.
Canabarro, A., Fanchini, F. F., Malvezzi, A. L., Pereira, R.,
and Chaves, R. (2019). Unveiling phase transitions
with machine learning. Phys. Rev. B, 100:045129.
Carleo, G., Cirac, I., Cranmer, K., Daudet, L., Schuld,
M., Tishby, N., Vogt-Maranto, L., and Zdeborov
´
a, L.
(2019). Machine learning and the physical sciences.
Reviews of Modern Physics, 91:045002.
Carrasquilla, J. and Melko, R. G. (2017). Machine learning
phases of matter. Nature Physics, 13(5):431–434.
Cohen, G., Afshar, S., Tapson, J., and Schaik, A. V. (2017).
EMNIST: Extending MNIST to handwritten letters.
Proceedings of the International Joint Conference on
Neural Networks, 2017-May.
Cortes, C., Jackel, L. D., and Chiang, W.-P. (1994). Limits
on learning machine accuracy imposed by data qual-
ity. Advances in Neural Information Processing Sys-
tems, 7.
Dask Development Team (2016). Dask: Library for dy-
namic task scheduling.
Deng, L. (2012). The mnist database of handwritten digit
images for machine learning research. IEEE Signal
Processing Magazine, 29:141–142.
Drenkow, N., Sani, N., Shpitser, I., and Unberath, M.
(2021). A systematic review of robustness in deep
learning for computer vision: Mind the gap? arXiv
preprint arXiv:2112.00639.
Gardner, E. (1988). The space of interactions in neural net-
work models. Journal of Physics A: Mathematical and
General, 21(1):257.
Giannetti, C., Lucini, B., and Vadacchino, D. (2019). Ma-
chine learning as a universal tool for quantitative in-
vestigations of phase transitions. Nuclear Physics B,
944:114639.
Goodfellow, I., Bengio, Y., and Courville, A. (2016). Deep
Learning. MIT Press. http://www.deeplearningbook.
org.
Gy
¨
orgyi, G. (1990). First-order transition to perfect gen-
eralization in a neural network with binary synapses.
Physical Review A, 41(12):7097–7100.
Heffern, E., Huelskamp, H., Bahar, S., and Inglis, R. F.
(2021). Phase transitions in biology: From bird flocks
to population dynamics. Proceedings of the Royal So-
ciety B: Biological Sciences, 288.
Ho, J., Jain, A., and Abbeel, P. (2020). Denoising diffusion
probabilistic models. Advances in Neural Information
Processing Systems.
ICAART 2024 - 16th International Conference on Agents and Artificial Intelligence
1090