4.7 Article

Feudal Latent Space Exploration for Coordinated Multi-Agent Reinforcement Learning

Journal

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TNNLS.2022.3146201

Keywords

Space exploration; Stochastic processes; Reinforcement learning; Training; Games; Task analysis; Markov processes; Deep reinforcement learning; exploration; multi-agent coordination; multi-agent reinforcement learning (MARL)

Ask authors/readers for more resources

This article investigates how multiple agents can learn to coordinate in order to achieve efficient exploration in reinforcement learning. The proposed feudal latent-space exploration (FLE) tackles the problem of exponentially increasing difficulty in independent exploration of the joint action space as the number of agents increases. Experimental results demonstrate that FLE outperforms baseline MARL approaches that use independent exploration strategy in terms of mean rewards, efficiency, and the expressiveness of coordination policies.
In this article, we investigate how multiple agents learn to coordinate to form efficient exploration in reinforcement learning. Though straightforward, independent exploration of the joint action space of multiple agents will become exponentially more difficult as the number of agents increases. To tackle this problem, we propose feudal latent-space exploration (FLE) for multi-agent reinforcement learning (MARL). FLE introduces a feudal commander to learn a low-dimensional global latent structure that instructs multiple agents to explore coordinately. Under this framework, the multi-agent policy gradient (PG) is adopted to optimize both the agent policy and latent structure end-to-end. We demonstrate the effectiveness of this method in two multi-agent environments that need explicit coordination. Experimental results validate that FLE outperforms baseline MARL approaches that use independent exploration strategy in terms of mean rewards, efficiency, and the expressiveness of coordination policies.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available