even without complex communication mechanisms,
the traffic can be controlled and achieve high perfor-
mance. Because the number of agents is not the func-
tion of vehicles but the function of intersections, it
can be said that the proposed method is scalable. Fu-
ture work may include applying our approach to more
complex large-scale maps including more vehicles.
ACKNOWLEDGEMENTS
This research was supported by Basic Science Re-
search Program through the National Research Foun-
dation of Korea (NRF) funded by the Ministry of Ed-
ucation (NRF-2016R1D1A1B04933156)
REFERENCES
Bengio, Y., Louradour, J., Collobert, R., and Weston, J.
(2009). Curriculum learning. In Proceedings of
the 26th annual international conference on machine
learning, pages 41–48.
Bernstein, D. S., Givan, R., Immerman, N., and Zilberstein,
S. (2002). The complexity of decentralized control
of markov decision processes. Mathematics of opera-
tions research, 27(4):819–840.
Bus¸oniu, L., Babu
ˇ
ska, R., and De Schutter, B. (2010).
Multi-agent reinforcement learning: An overview. In-
novations in multi-agent systems and applications-1,
pages 183–221.
Chouhan, A. P. and Banda, G. (2018). Autonomous inter-
section management: A heuristic approach. IEEE Ac-
cess, 6:53287–53295.
Das, A., Gervet, T., Romoff, J., Batra, D., Parikh, D., Rab-
bat, M., and Pineau, J. (2019). Tarmac: Targeted
multi-agent communication. In International Confer-
ence on Machine Learning, pages 1538–1546. PMLR.
Dresner, K. and Stone, P. (2008). A multiagent approach
to autonomous intersection management. Journal of
artificial intelligence research, 31:591–656.
Fiori, C., Arcidiacono, V., Fontaras, G., Makridis, M., Mat-
tas, K., Marzano, V., Thiel, C., and Ciuffo, B. (2019).
The effect of electrified mobility on the relationship
between traffic conditions and energy consumption.
Transportation Research Part D: Transport and En-
vironment, 67:275–290.
Foerster, J. N., Assael, Y. M., De Freitas, N., and White-
son, S. (2016). Learning to communicate with deep
multi-agent reinforcement learning. arXiv preprint
arXiv:1605.06676.
Gupta, J. K., Egorov, M., and Kochenderfer, M. (2017).
Cooperative multi-agent control using deep reinforce-
ment learning. In International Conference on Au-
tonomous Agents and Multiagent Systems, pages 66–
83. Springer.
Hartanti, D., Aziza, R. N., and Siswipraptini, P. C. (2019).
Optimization of smart traffic lights to prevent traf-
fic congestion using fuzzy logic. TELKOMNIKA
Telecommunication Computing Electronics and Con-
trol, 17(1):320–327.
Hausknecht, M., Au, T.-C., and Stone, P. (2011).
Autonomous intersection management: Multi-
intersection optimization. In 2011 IEEE/RSJ
International Conference on Intelligent Robots and
Systems, pages 4581–4586. IEEE.
Khoza, E., Tu, C., and Owolawi, P. A. (2020). Decreas-
ing traffic congestion in vanets using an improved hy-
brid ant colony optimization algorithm. J. Commun.,
15(9):676–686.
Lasley, P. (2019). 2019 urban mobility report.
Li, S., Gupta, J. K., Morales, P., Allen, R., and Kochender-
fer, M. J. (2020). Deep implicit coordination graphs
for multi-agent reinforcement learning. arXiv preprint
arXiv:2006.11438.
Liu, Y., Wang, W., Hu, Y., Hao, J., Chen, X., and Gao, Y.
(2020). Multi-agent game abstraction via graph atten-
tion neural network. In Proceedings of the AAAI Con-
ference on Artificial Intelligence, volume 34, pages
7211–7218.
Mnih, V., Kavukcuoglu, K., Silver, D., Graves, A.,
Antonoglou, I., Wierstra, D., and Riedmiller, M.
(2013). Playing atari with deep reinforcement learn-
ing. arXiv preprint arXiv:1312.5602.
Pal, S. and Mitra, S. (1992). Multilayer perceptron, fuzzy
sets, and classification. IEEE Transactions on Neural
Networks, 3:683–697.
Parker, A. and Nitschke, G. (2017). How to best automate
intersection management. In 2017 IEEE Congress on
Evolutionary Computation (CEC), pages 1247–1254.
IEEE.
Singh, A., Jain, T., and Sukhbaatar, S. (2018). Learn-
ing when to communicate at scale in multiagent
cooperative and competitive tasks. arXiv preprint
arXiv:1812.09755.
Su, J., Adams, S., and Beling, P. A. (2020). Coun-
terfactual multi-agent reinforcement learning with
graph convolution communication. arXiv preprint
arXiv:2004.00470.
Sukhbaatar, S., Fergus, R., et al. (2016). Learning multia-
gent communication with backpropagation. Advances
in neural information processing systems, 29:2244–
2252.
Tan, M. (1993). Multi-agent reinforcement learning: Inde-
pendent vs. cooperative agents. In Proceedings of the
tenth international conference on machine learning,
pages 330–337.
Walraven, E., Spaan, M. T., and Bakker, B. (2016). Traf-
fic flow optimization: A reinforcement learning ap-
proach. Engineering Applications of Artificial Intelli-
gence, 52:203–212.
Wang, Z., Schaul, T., Hessel, M., Hasselt, H., Lanctot, M.,
and Freitas, N. (2016). Dueling network architec-
tures for deep reinforcement learning. In International
conference on machine learning, pages 1995–2003.
PMLR.
Wei, H., Zheng, G., Gayah, V., and Li, Z. (2019). A sur-
vey on traffic signal control methods. arXiv preprint
arXiv:1904.08117.
Micro Junction Agent: A Scalable Multi-agent Reinforcement Learning Method for Traffic Control
515