
5 CONCLUSION
This study aims at efficient learning of models that
balance accuracy and network scale as its goal. We
are also targeting the construction of SEN, which
evaluates the potential of a model utilizing the inter-
nal states of the network. In this paper, we focused on
the LSTM in the traffic accident prediction network
and defined the activity level of LSTM cells based on
the variation in the values of the forget gate. More-
over, we assumed that a model has more potential if
the proportion of high-activity LSTM cells is higher,
and hence defined LAL. By checking the transition of
LAL and AP, we demonstrated that the value of LAL
is effective for controlling SEN. Additionally, since a
positive correlation between LAL and AP was con-
firmed, we believed that it’s possible to predict future
accuracy by extrapolating LAL in the early stages of
training.
In our future research, firstly, we plan to create a
small validation data set for the calculation of LAL
and experiment on the effectiveness of LAL calcu-
lated at a low cost. Then, we will proceed to discuss
the specific methods of extrapolating LAL.
REFERENCES
Baker, B., Gupta, O., Raskar, R., and Naik, N. (2018).
Accelerating neural architecture search using perfor-
mance prediction. In International Conference on
Learning Representations.
Chan, F. H., Chen, Y. T., Xiang, Y., and Sun, M. (2016).
Anticipating accidents in dashcam videos. In Asian
Conference on Computer Vision, pages 136–153.
Chen, Y., Yang, T., Zhang, X., Meng, G., Xiao, X., and
Sun, J. (2019). Detnas: Backbone search for object
detection. In International Conference on Neural In-
formation Processing Systems.
Domhan, T., Springenberg, J. T., and Hutter, F. (2015).
Speeding up automatic hyperparameter optimization
of deep neural networks by extrapolation of learning
curves. In International Joint Conference on Artificial
Intelligence, pages 3460–3468.
Greff, K., Srivastava, R. K., Koutnık, J., Steunebrink, B. R.,
and Schmidhuber, J. (2017). Lstm: A search space
odyssey. IEEE Transactions on Neural Networks and
Learning Systems, pages 2222–2232.
Hutter, F., Kotthoff, L., and Vanschoren, J., editors (2019).
Automatic Machine Learning: Methods, Systems,
Challenges. Springer. In press, available at http:
//automl.org/book.
Karpathy, A., Johnson, J., and Fei-Fei, L. (2016). Visualiz-
ing and understanding recurrent networks. In Interna-
tional Conference on Learning Representations.
Klein, A., Falkner, S., Springenberg, J. T., and Hutter, F.
(2017). Learning curve prediction with bayesian neu-
ral networks. In International Conference on Learn-
ing Representations.
Liu, C., Chen, L.-C., Schroff, F., Adam, H., Hua, W., Yuille,
A., and Fei-Fei, L. (2019). Auto-deeplab: Hierarchi-
cal neural architecture search for semantic image seg-
mentation. In IEEE / CVF Computer Vision and Pat-
tern Recognition Conference.
Pascanu, R., Mikolov, T., and Bengio, Y. (2013). On the
difficulty of training recurrent neural networks. In In-
ternational Conference on Machine Learning, pages
1310–1318.
Real, E., Aggarwal, A., Huang, Y., and Le, Q. V. (2019).
Regularized evolution for image classifier architecture
search. In AAAI Conference on Artificial Intelligence.
Ren, S., He, K., Girshick, R., and Sun, J. (2015). Faster
r-cnn: Towards real-time object detection with region
proposal networks. In IEEE / CVF Computer Vision
and Pattern Recognition Conference.
Swersky, K., Snoek, J., and Adams, R. P. (2014). Freeze-
thaw bayesian optimization. arXiv:1406.3896.
Wang, N., Gao, Y., Chen, H., Wang, P., Tian, Z., Shen, C.,
and Zhang, Y. (2020). Nas-fcos: Fast neural architec-
ture search for object detection. In IEEE / CVF Com-
puter Vision and Pattern Recognition Conference.
Wistuba, M. and Pedapati, T. (2019). Inductive transfer for
neural architecture optimization. arXiv:1903.03536.
Wu, B., Dai, X., Zhang, P., Wang, Y., Sun, F., Wu, Y., Tian,
Y., Vajda, P., Jia, Y., and Keutzer, K. (2019). Fbnet:
Hardware-aware efficient convnet design via differen-
tiable neural architecture search. In IEEE / CVF Com-
puter Vision and Pattern Recognition Conference.
Yan, S., White, C., Savani, Y., and Hutter, F. (2021).
Nas-bench-x11 and the power of learning curves.
arXiv:2111.03602.
Zhang, Y., Qiu, Z., Liu, J., Yao, T., Liu, D., and Mei, T.
(2019). Customizable architecture search for semantic
segmentation. In IEEE / CVF Computer Vision and
Pattern Recognition Conference.
Zoph, B., Vasudevan, V., Shlens, J., and Le, Q. V. (2018).
Learning transferable architectures for scalable image
recognition. In IEEE / CVF Computer Vision and Pat-
tern Recognition Conference.
VISAPP 2024 - 19th International Conference on Computer Vision Theory and Applications
324