Skip to content

PPSantos/hybrid-marl

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Centralized training with hybrid execution in multi-agent reinforcement learning

Computational code for https://arxiv.org/abs/2210.06274.

Hybrid MARL is an extension over the EPyMARL library (https://github.com/uoe-agents/epymarl).

Contributors

  • Pedro P. Santos (@PPSantos)
  • Diogo Carvalho (@carvalhomm88)
  • Miguel Vasco (@miguelsvasco)

Installation instructions

To install the codebase, run (tested with python 3.8.10):

./install.sh

Running experiments

After installation, you can use the script run.sh to run experiments, where:

  1. ENV variable selects the environment to use:
    • SimpleSpreadXY-v0 - SpreadXY-2
    • SimpleSpreadXY4-v0 - SpreadXY-4;
    • SimpleSpreadBlind-v0 - SpreadBlindfold;
    • SimpleBlindDeaf-v0 - HearSee;
    • SimpleSpread-v0 - SimpleSpread;
    • SimpleSpeakerListener-v0 - SimpleSpeakerListener;
    • Foraging-2s-15x15-2p-2f-coop-v2 - LBF environment (modified);
  2. ALGO variable selects the RL algorithm to use (iql_ns, qmix_ns, ippo_ns, or mappo_ns for MPE environments; iql_ns_lbf, qmix_ns_lbf, ippo_ns_lbf, or mappo_ns_lbf for LBF environments).
  3. PERCEPTION variable selects the perceptual model to use:
    • obs - Obs.
    • joint_obs - Oracle;
    • joint_obs_drop_test - Masked joint obs.;
    • ablation_no_pred - MD baseline;
    • ablation_no_pred_masks - MD w/ masks baseline;
    • maro_no_training - MARO;
    • maro - MARO w/ dropout;
  4. TIME_LIMIT variable: 25 for MPE environments; 30 for LBF environments.

About

Centralized training with hybrid execution in multi-agent reinforcement learning

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published