in understanding what is the effect of the optimisa-
tion metric used by the iLoT tool on the overall output
quality.
Thanks to the quantization we obtain a 1.73x
speed-up on inference time compared to the initial
float32 model. This brings the total speed-up with re-
spect to the standard Monte Carlo approach to several
orders of magnitude (67 000x). These results make
the proposed reduced precision strategies on GAN
models an attractive approach for future researches on
the detector simulations and could help to save large
computing resources in view of the future High Lu-
minosity LHC runs.
With this paper we hope to inspire colleagues to
follow us and to apply quantization to deep learning
models with more complex outputs than classification
tasks. More use cases, in turn, will lead to an im-
provement of quantization methods and tools leading
to possible further decrease of inference times in the
future.
ACKNOWLEDGEMENTS
This work has been sponsored by the Wolfgang Gen-
tner Programme of the German Federal Ministry of
Education and Research.
REFERENCES
(2020). TensorFlow Lite. TensorFlow For Mobile & IoT.
Agostinelli, S. et al. (2003). GEANT4–a simulation toolkit.
Nucl. Instrum. Meth. A, 506:250–303.
Albrecht, Alvesand, A. e. a. (2019). A roadmap for hep
software and computing r&d for the 2020s. Comput-
ing and Software for Big Science, 3.
Apollinari, G., B
´
ejar Alonso, I., Br
¨
uning, O., Fessia, P., La-
mont, M., Rossi, L., and Tavian, L. (2017). High-
Luminosity Large Hadron Collider (HL-LHC): Tech-
nical Design Report V. 0.1. 4/2017.
Arjovsky, M., Chintala, S., and Bottou, L. (2017). Wasser-
stein gan.
Borji, A. (2018). Pros and cons of gan evaluation measures.
Brown, R. and Cockerill, D. (2012). Electromagnetic
calorimetry. Nuclear Instruments and Methods in
Physics Research Section A, 666:47 – 79. Advanced
Instrumentation.
Brownlee, J. (2020). 18 Impressive Applications of Gener-
ative Adversarial Networks (GANs).
de Oliveira, L. and Paganini, Michela, B. (2017). Learning
particle physics by example: Location-aware genera-
tive adversarial networks for physics synthesis. Com-
puting and Software for Big Science, 1(1).
Erdmann M., G. J. . Q. T. (2019). Precise simulation of
electromagnetic calorimeter showers using a wasser-
stein generative adversarial network. Comput Softw
Big Sci 3.
Feng Tian, Chuanqi Wang, G. Z. P. C. P. Y. H. S. and Gong,
J. (2020). Intel
R
low precision optimization tool.
Ghosh, A. (2019). Deep generative models for fast shower
simulation in ATLAS. Technical Report ATL-SOFT-
PROC-2019-007, CERN, Geneva.
Goodfellow, I. J., Pouget-Abadie, J., Mirza, M., Xu, B.,
Warde-Farley, D., Ozair, S., Courville, A., and Ben-
gio, Y. (2014). Generative adversarial networks.
He, K., Zhang, X., Ren, S., and Sun, J. (2015). Deep resid-
ual learning for image recognition.
IEEE (2008). Ieee standard for floating-point arithmetic.
IEEE Std 754-2008, pages 1–70.
Intel. oneapi deep neural network library (onednn).
Jain, A., Bhattacharya, S., Masuda, M., Sharma, V., and
Wang, Y. (2020). Efficient execution of quantized
deep learning models: A compiler approach.
Khattak, G., Vallecorsa, S., and Carminati, F. (2018). Three
dimensional energy parametrized generative adversar-
ial networks for electromagnetic shower simulation.
In 2018 25th IEEE International Conference on Im-
age Processing (ICIP), pages 3913–3917.
Micikevicius, P., Narang, S., Alben, J., Diamos, G., Elsen,
E., Garcia, D., Ginsburg, B., Houston, M., Kuchaiev,
O., Venkatesh, G., and Wu, H. (2017). Mixed preci-
sion training.
Nandakumar, S. R., Le Gallo, M., Piveteau, C., Joshi, V.,
Mariani, G., Boybat, I., et al. (2020). Mixed-precision
deep learning based on computational memory. Fron-
tiers in Neuroscience, 14.
Osorio, J. Evaluating mixed-precision arithmetic for 3d
generative adversarial networks to simulate high en-
ergy physics detectors.
Salamani, D., Gadatsch, S., Golling, T., Stewart, G., Ghosh,
A., Rousseau, D., Hasib, A., and Schaarschmidt, J.
(2018). Deep generative models for fast shower sim-
ulation in atlas. pages 348–348.
Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., and Wojna,
Z. (2015). Rethinking the inception architecture for
computer vision.
Vallecorsa, S., Carminati, F., and Khattak, G. (2019). 3d
convolutional gan for fast simulation. EPJ Web of
Conferences, 214:02010.
Wu, H. (2019). Inference at reduced precision on gpus.
Wu, H., Judd, P., Zhang, X., Isaev, M., and Micikevicius, P.
(2020). Integer quantization for deep learning infer-
ence: Principles and empirical evaluation.
ICPRAM 2021 - 10th International Conference on Pattern Recognition Applications and Methods
258