tools for designers. Specifically, we are considering
using our predictive model as an objective function
to optimize the color scheme of buttons and text in
mobile UI using a genetic algorithm. Optimization
allows developers to easily create a UI with the in-
creased importance of the UI components they want
to make stand out. For optimization, we would like
to perform predictions for novel and unique UIs that
are not included in existing datasets and conduct user
experiments to see if the predicted visual importance
maps are appropriate.
For a better user experience, it is also necessary to
analyze where users direct their attention when they
see a UI and whether they can understand that UI cor-
rectly. There is already related research on how users
understand UI, such as icon annotation in mobile UI
(Zang et al., 2021) and predicting mobile UI tappabil-
ity (Swearngin and Li, 2019). We believe that com-
bining these related work with our visual importance
predictions will provide more useful feedback to de-
signers.
ACKNOWLEDGEMENTS
This work was supported by JSPS KAKENHI Grant
Numbers JP21H03496, JP22K12157.
REFERENCES
Bylinskii, Z., Kim, N. W., O’Donovan, P., Alsheikh, S.,
Madan, S., Pfister, H., Durand, F., Russell, B., and
Hertzmann, A. (2017). Learning visual importance for
graphic designs and data visualizations. Proceedings
of the 30th Annual ACM Symposium on User Interface
Software and Technology, pages 57–69.
Chen, L.-C., Papandreou, G., Kokkinos, I., Murphy, K., and
Yuille, A. L. (2018). Deeplab: Semantic image seg-
mentation with deep convolutional nets, atrous convo-
lution, and fully connected crfs. IEEE Transactions
on Pattern Analysis and Machine Intelligence 40(4),
pages 834–848.
Deka, B., Huang, Z., Franzen, C., Hibschman, J., Afergan,
D., Li, Y., Nichols, J., and Kumar, R. (2017). Rico:
A mobile app dataset for building data-driven design
applications. Proceedings of the 30th Annual ACM
Symposium on User Interface Software and Technol-
ogy, pages 845–854.
Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., and Fei-
Fei, L. (2009). Imagenet: A large-scale hierarchical
image database. IEEE conference on computer vision
and pattern recognition, pages 248–255.
Fosco, C., Casser, V., Bedi, A. K., O’Donovan, P., Hertz-
mann, A., and Bylinskii, Z. (2020). Predicting visual
importance across graphic design types. Proceedings
of the 33rd Annual ACM Symposium on User Interface
Software and Technology, pages 249–260.
Gupta, P., Gupta, S., Jayagopal, A., Pal, S., and Sinha,
R. (2018). Saliency prediction for mobile user inter-
faces. 2018 IEEE Winter Conference on Applications
of Computer Vision, pages 1529–1538.
Jiang, M., Huang, S., Duan, J., and Zhao, Q. (2015). Sali-
con: Saliency in context. IEEE conference on com-
puter vision and pattern recognition, pages 1072–
1080.
Kroner, A., Senden, M., Driessens, K., and Goebel, R.
(2020). Contextual encoder–decoder network for vi-
sual saliency prediction. Neural Networks 129, pages
261–270.
Leiva, L. A., Xue, Y., Bansal, A., Tavakoli, H. R., K
¨
oro
˘
glu,
T., Du, J., Dayama, N. R., and Oulasvirta, A. (2020).
Understanding visual saliency in mobile user inter-
faces. 22nd International Conference on Human-
Computer Interaction with Mobile Devices and Ser-
vices, pages 1–12.
Swearngin, A. and Li, Y. (2019). Modeling mobile inter-
face tappability using crowdsourcing and deep learn-
ing. Proceedings of the 2019 CHI Conference on Hu-
man Factors in Computing Systems, pages 1–11.
Zang, X., Xu, Y., and Chen, J. (2021). Multimodal icon
annotation for mobile applications. Proceedings of
the 23rd International Conference on Mobile Human-
Computer Interaction, pages 1–11.
ICAART 2023 - 15th International Conference on Agents and Artificial Intelligence
266