Currently, our approach is dependent on the ca-
pability of the technique of artistic text style trans-
fer. In the future, we may incorporate differential
post-processing schemes (Zhan et al., 2019) into our
framework, to generate rich and varied adversarial
examples with real-world scenes. We may also ex-
plore to combine the techniques of manipulating la-
tent codes with style transfer, to further enhance the
generation process and the smoothness of the adver-
sarial style texture.
REFERENCES
Azadi, S., Fisher, M., Kim, V. G., Wang, Z., Shechtman,
E., and Darrell, T. (2018). Multi-content GAN for
few-shot font style transfer. In IEEE Conference on
Computer Vision and Pattern Recognition (CVPR).
Baek, J., Kim, G., Lee, J., Park, S., Han, D., Yun, S.,
Oh, S. J., and Lee, H. (2019). What is wrong with
scene text recognition model comparisons? dataset
and model analysis. In IEEE/CVF International Con-
ference on Computer Vision (ICCV).
Bhattad, A., Chong, M. J., Liang, K., Li, B., and Forsyth,
D. A. (2020). Unrestricted adversarial examples via
semantic manipulation. In International Conference
on Learning Representations (ICLR).
Carlini, N. and Wagner, D. (2017). Towards evaluating the
robustness of neural networks. In IEEE Symposium
on Security and Privacy (SP).
Duan, R., Ma, X., Wang, Y., Bailey, J., Qin, A. K.,
and Yang, Y. (2020). Adversarial camouflage: Hid-
ing physical-world attacks with natural styles. In
IEEE/CVF Conference on Computer Vision and Pat-
tern Recognition (CVPR).
Goodfellow, I. J., Shlens, J., and Szegedy, C. (2015). Ex-
plaining and harnessing adversarial examples. In In-
ternational Conference on Learning Representations.
Jaderberg, M., Simonyan, K., Vedaldi, A., and Zisser-
man, A. (2014). Synthetic data and artificial neural
networks for natural scene text recognition. CoRR,
abs/1406.2227.
Liu, H. D., Tao, M., Li, C., Nowrouzezahrai, D., and Jacob-
son, A. (2019). Beyond pixel norm-balls: Paramet-
ric adversaries using an analytically differentiable ren-
derer. In International Conference on Learning Rep-
resentations (ICLR).
Madry, A., Makelov, A., Schmidt, L., Tsipras, D., and
Vladu, A. (2018). Towards deep learning models re-
sistant to adversarial attacks. In International Confer-
ence on Learning Representations.
Papernot, N., Mcdaniel, P., Jha, S., Fredrikson, M., Celik,
Z. B., and Swami, A. (2016). The limitations of deep
learning in adversarial settings. In IEEE Symposium
on Security and Privacy.
Qiu, H., Xiao, C., Yang, L., Yan, X., Lee, H., and Li, B.
(2020). Semanticadv: Generating adversarial exam-
ples via attribute-conditional image editing. In Euro-
pean Conference on Computer Vision (ECCV).
Shetty, R., Fritz, M., and Schiele, B. (2020). Towards au-
tomated testing and robustification by semantic adver-
sarial data generation. In Vedaldi, A., Bischof, H.,
Brox, T., and Frahm, J., editors, European Conference
on Computer Vision (ECCV).
Song, C. and Shmatikov, V. (2018). Fooling OCR systems
with adversarial text images. CoRR, abs/1802.05385.
Song, Y., Shu, R., Kushman, N., and Ermon, S. (2018).
Constructing unrestricted adversarial examples with
generative models. In Annual Conference on Neural
Information Processing Systems 2018 (NeurIPS).
Szegedy, H., Zaremba, W., Sutskever, I., Bruna, J., Erhan,
D., Goodfellow, I., and Fergus, R. (2014). Intriguing
properties of neural networks. In International Con-
ference on Learning Representations.
Wang, S., Chen, S., Chen, T., Nepal, S., Rudolph, C.,
and Grobler, M. (2020). Generating semantic ad-
versarial examples via feature manipulation. ArXiv,
abs/2001.02297.
Xiao, C., Li, B., Zhu, J., He, W., Liu, M., and Song, D.
(2018a). Generating adversarial examples with adver-
sarial networks. In International Joint Conference on
Artificial Intelligence (IJCAI).
Xiao, C., Zhu, J.-Y., Li, B., He, W., Liu, M., and Song, D.
(2018b). Spatially transformed adversarial examples.
In International Conference on Learning Representa-
tions.
Xu, X., Chen, J., Xiao, J., Gao, L., Shen, F., and Shen,
H. T. (2020a). What machines see is not what they
get: Fooling scene text recognition models with ad-
versarial text images. In IEEE/CVF Conference on
Computer Vision and Pattern Recognition (CVPR).
Xu, X., Chen, J., Xiao, J., Wang, Z., Yang, Y., and Shen,
H. T. (2020b). Learning optimization-based adversar-
ial perturbations for attacking sequential recognition
models. In International Conference on Multimedia.
Yang, S., Liu, J., Wang, W., and Guo, Z. (2019a). TET-
GAN: text effects transfer via stylization and destyl-
ization. In AAAI Conference on Artificial Intelligence,
pages 1238–1245.
Yang, S., Wang, Z., Wang, Z., Xu, N., Liu, J., and Guo,
Z. (2019b). Controllable artistic text style transfer
via shape-matching GAN. In IEEE/CVF International
Conference on Computer Vision (ICCV).
Zhan, F., Zhu, H., and Lu, S. (2019). Spatial fusion GAN
for image synthesis. In IEEE Conference on Computer
Vision and Pattern Recognition (CVPR).
Zhao, S., Liu, Z., Lin, J., Zhu, J., and Han, S. (2020). Dif-
ferentiable augmentation for data-efficient GAN train-
ing. In Annual Conference on Neural Information
Processing Systems (NeurIPS).
Zhao, Z., Dua, D., and Singh, S. (2018). Generating natural
adversarial examples. In International Conference on
Learning Representations (ICLR).