4.7 Article

SMIX(λ): Enhancing Centralized Value Functions for Cooperative Multiagent Reinforcement Learning

Journal

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TNNLS.2021.3089493

Keywords

Training; Optimization; Reinforcement learning; Nash equilibrium; Task analysis; History; Learning systems; Deep reinforcement learning (DRL); multiagent reinforcement learning (MARL); multiagent systems; StarCraft Multiagent Challenge (SMAC)

Ask authors/readers for more resources

This article proposes a method named SMIX(lambda) to learn a stable and generalizable centralized value function (CVF) through off-policy training. By using the lambda-return as a proxy for computing the temporal difference error, the modified QMIX network structure is adopted to train the model. Experiments demonstrate the significant advantages of the proposed SMIX(lambda) method in multiagent reinforcement learning.
Learning a stable and generalizable centralized value function (CVF) is a crucial but challenging task in multiagent reinforcement learning (MARL), as it has to deal with the issue that the joint action space increases exponentially with the number of agents in such scenarios. This article proposes an approach, named SMIX(lambda), that uses an off-policy training to achieve this by avoiding the greedy assumption commonly made in CVF learning. As importance sampling for such off-policy training is both computationally costly and numerically unstable, we proposed to use the lambda-return as a proxy to compute the temporal difference (TD) error. With this new loss function objective, we adopt a modified QMIX network structure as the base to train our model. By further connecting it with the Q(lambda) approach from a unified expectation correction viewpoint, we show that the proposed SMIX(lambda) is equivalent to Q(lambda) and hence shares its convergence properties, while without being suffered from the aforementioned curse of dimensionality problem inherent in MARL. Experiments on the StarCraft Multiagent Challenge (SMAC) benchmark demonstrate that our approach not only outperforms several state-of-the-art MARL methods by a large margin but also can be used as a general tool to improve the overall performance of other centralized training with decentralized execution (CTDE)-type algorithms by enhancing their CVFs.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available