Our code is modified from https://github.com/starry-sky6688/StarCraft.
This repository is our Pytorch implementations for Decomposed Soft Actor-Critic Method for Cooperative Multi-Agent Reinforcement Learning(https://arxiv.org/abs/2104.06655).
In the paper, we incorporates the idea of the multi-agent value function decomposition and soft actor-critic framework effectively and proposes a new method mSAC. Experimental results demonstrate that mSAC significantly outperforms policy-based approach–COMA, and achieves competitive performance with SOTA value-based approach–Qmix on most tasks in terms of asymptotic performance metric. In addition, mSAC has achieved significantly better results than Qmix in some tasks with large action spaces (such as 2c_vs_64zg, MMM2).
We trained these algorithms on SMAC, the decentralised micromanagement scenario of StarCraft II.
- QMIX: Monotonic Value Function Factorisation for Deep Multi-Agent Reinforcement Learning
- Value-Decomposition Networks For Cooperative Multi-Agent Learning
- Counterfactual Multi-Agent Policy Gradients
$ python src/main.py --map=3m
To run different variant algorithms, you can change the first line in src/main.py [from runner_msac import Runner] to different runner_[alg. name].