loading
Papers Papers/2022 Papers Papers/2022

Research.Publish.Connect.

Paper

Authors: Keisuke Fujii 1 ; 2 ; 3 ; Kazushi Tsutsui 1 ; Atom Scott 1 ; Hiroshi Nakahara 1 ; Naoya Takeishi 2 ; 4 and Yoshinobu Kawahara 2 ; 5

Affiliations: 1 Graduate School of Informatics, Nagoya University, Nagoya, Aichi, Japan ; 2 Center for Advanced Intelligence Project, RIKEN, Osaka, Osaka, Japan ; 3 PRESTO, Japan Science and Technology Agency, Tokyo, Japan ; 4 The Graduate School of Engineering, The University of Tokyo, Tokyo, Japan ; 5 Graduate School of Information Science and Technology, Osaka University, Osaka, Japan

Keyword(s): Neural Networks, Trajectory, Simulation, Multi-Agent.

Abstract: Modeling of real-world biological multi-agents is a fundamental problem in various scientific and engineering fields. Reinforcement learning (RL) is a powerful framework to generate flexible and diverse behaviors in cyberspace; however, when modeling real-world biological multi-agents, there is a domain gap between behaviors in the source (i.e., real-world data) and the target (i.e., cyberspace for RL), and the source environment parameters are usually unknown. In this paper, we propose a method for adaptive action supervision in RL from real-world demonstrations in multi-agent scenarios. We adopt an approach that combines RL and supervised learning by selecting actions of demonstrations in RL based on the minimum distance of dynamic time warping for utilizing the information of the unknown source dynamics. This approach can be easily applied to many existing neural network architectures and provide us with an RL model balanced between reproducibility as imitation and generalization ability to obtain rewards in cyberspace. In the experiments, using chase-and-escape and football tasks with the different dynamics between the unknown source and target environments, we show that our approach achieved a balance between the reproducibility and the generalization ability compared with the baselines. In particular, we used the tracking data of professional football players as expert demonstrations in football and show successful performances despite the larger gap between behaviors in the source and target environments than the chase-and-escape task. (More)

CC BY-NC-ND 4.0

Sign In Guest: Register as new SciTePress user now for free.

Sign In SciTePress user: please login.

PDF ImageMy Papers

You are not signed in, therefore limits apply to your IP address 18.118.27.72

In the current month:
Recent papers: 100 available of 100 total
2+ years older papers: 200 available of 200 total

Paper citation in several formats:
Fujii, K.; Tsutsui, K.; Scott, A.; Nakahara, H.; Takeishi, N. and Kawahara, Y. (2024). Adaptive Action Supervision in Reinforcement Learning from Real-World Multi-Agent Demonstrations. In Proceedings of the 16th International Conference on Agents and Artificial Intelligence - Volume 2: ICAART; ISBN 978-989-758-680-4; ISSN 2184-433X, SciTePress, pages 27-39. DOI: 10.5220/0012261100003636

@conference{icaart24,
author={Keisuke Fujii. and Kazushi Tsutsui. and Atom Scott. and Hiroshi Nakahara. and Naoya Takeishi. and Yoshinobu Kawahara.},
title={Adaptive Action Supervision in Reinforcement Learning from Real-World Multi-Agent Demonstrations},
booktitle={Proceedings of the 16th International Conference on Agents and Artificial Intelligence - Volume 2: ICAART},
year={2024},
pages={27-39},
publisher={SciTePress},
organization={INSTICC},
doi={10.5220/0012261100003636},
isbn={978-989-758-680-4},
issn={2184-433X},
}

TY - CONF

JO - Proceedings of the 16th International Conference on Agents and Artificial Intelligence - Volume 2: ICAART
TI - Adaptive Action Supervision in Reinforcement Learning from Real-World Multi-Agent Demonstrations
SN - 978-989-758-680-4
IS - 2184-433X
AU - Fujii, K.
AU - Tsutsui, K.
AU - Scott, A.
AU - Nakahara, H.
AU - Takeishi, N.
AU - Kawahara, Y.
PY - 2024
SP - 27
EP - 39
DO - 10.5220/0012261100003636
PB - SciTePress