Skip to content

Implementations of QMIX and VDN on SMAC,corresponding to paper 《QMIX:Monotonic Value Function Factorisation for Deep Multi-Agent Reinforcement Learning》 and 《Value-Decomposition Networks For Cooperative Multi-Agent Learning》

Notifications You must be signed in to change notification settings

canyon/StarCraft

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

38 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

StarCraft

This is a pytorch implementation of the multi-agent reinforcement learning algrithms, QMIX and VDN, both of which are the state of art MARL algrithms. We trained these algrithms on SMAC, which is the decentralised micromanagement scenario of StarCraft II.

Requirements

Acknowledgement

Quick Start

$ python main.py --evaluate_epoch=100 --map=3m

Directly run the main.py, then the two algrithms will be respectively tested on map '3m' for 100 episodes, using the pretrained model.

Result

Although qmix and vdn are the state of art multi-agent algrithms, they are unstable sometimes, you need to independently run several times to get better performence.

1. Win Rate of QMIX in Two Independent Runs on '3m'

2. Win Rate of VDN in Two Independent Runs on '3m'

About

Implementations of QMIX and VDN on SMAC,corresponding to paper 《QMIX:Monotonic Value Function Factorisation for Deep Multi-Agent Reinforcement Learning》 and 《Value-Decomposition Networks For Cooperative Multi-Agent Learning》

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%