Journal
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS
Volume 34, Issue 10, Pages 7775-7783Publisher
IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TNNLS.2022.3146201
Keywords
Space exploration; Stochastic processes; Reinforcement learning; Training; Games; Task analysis; Markov processes; Deep reinforcement learning; exploration; multi-agent coordination; multi-agent reinforcement learning (MARL)
Ask authors/readers for more resources
This article investigates how multiple agents can learn to coordinate in order to achieve efficient exploration in reinforcement learning. The proposed feudal latent-space exploration (FLE) tackles the problem of exponentially increasing difficulty in independent exploration of the joint action space as the number of agents increases. Experimental results demonstrate that FLE outperforms baseline MARL approaches that use independent exploration strategy in terms of mean rewards, efficiency, and the expressiveness of coordination policies.
In this article, we investigate how multiple agents learn to coordinate to form efficient exploration in reinforcement learning. Though straightforward, independent exploration of the joint action space of multiple agents will become exponentially more difficult as the number of agents increases. To tackle this problem, we propose feudal latent-space exploration (FLE) for multi-agent reinforcement learning (MARL). FLE introduces a feudal commander to learn a low-dimensional global latent structure that instructs multiple agents to explore coordinately. Under this framework, the multi-agent policy gradient (PG) is adopted to optimize both the agent policy and latent structure end-to-end. We demonstrate the effectiveness of this method in two multi-agent environments that need explicit coordination. Experimental results validate that FLE outperforms baseline MARL approaches that use independent exploration strategy in terms of mean rewards, efficiency, and the expressiveness of coordination policies.
Authors
I am an author on this paper
Click your name to claim this paper and add it to your profile.
Reviews
Recommended
No Data Available