Skip to content

LLNL/Abmarl

main
Switch branches/tags

Name already in use

A tag already exists with the provided branch name. Many Git commands accept both tag and branch names, so creating this branch may cause unexpected behavior. Are you sure you want to create this branch?
Code

Latest commit

 

Git stats

Files

Permalink
Failed to load latest commit information.
Type
Name
Latest commit message
Commit time
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Abmarl

Abmarl is a package for developing Agent-Based Simulations and training them with MultiAgent Reinforcement Learning (MARL). We provide an intuitive command line interface for engaging with the full workflow of MARL experimentation: training, visualizing, and analyzing agent behavior. We define an Agent-Based Simulation Interface and Simulation Manager, which control which agents interact with the simulation at each step. We support integration with popular reinforcement learning simulation interfaces, including gym.Env, MultiAgentEnv, and OpenSpiel. We define our own GridWorld Simulation Framework for creating custom grid-based Agent Based Simulations.

Abmarl leverages RLlib’s framework for reinforcement learning and extends it to more easily support custom simulations, algorithms, and policies. We enable researchers to rapidly prototype MARL experiments and simulation design and lower the barrier for pre-existing projects to prototype RL as a potential solution.

Build and Test Badge Sphinx docs Badge Lint Badge

Quickstart

To use Abmarl, install via pip: pip install abmarl

To develop Abmarl, clone the repository and install via pip's development mode. Note: Abmarl requires python3.7 or python3.8.

git clone git@github.com:LLNL/Abmarl.git
cd abmarl
pip install -r requirements.txt
pip install -e . --no-deps

Train agents in a multicorridor simulation:

abmarl train examples/multi_corridor_example.py

Visualize trained behavior:

abmarl visualize ~/abmarl_results/MultiCorridor-2020-08-25_09-30/ -n 5 --record

Note: If you install with conda, then you must also include ffmpeg in your virtual environment.

Documentation

You can find the latest Abmarl documentation on our ReadTheDocs page.

Documentation Status

Community

Citation

DOI

Abmarl has been published to the Journal of Open Source Software (JOSS). It can be cited using the following bibtex entry:

@article{Rusu2021,
  doi = {10.21105/joss.03424},
  url = {https://doi.org/10.21105/joss.03424},
  year = {2021},
  publisher = {The Open Journal},
  volume = {6},
  number = {64},
  pages = {3424},
  author = {Edward Rusu and Ruben Glatt},
  title = {Abmarl: Connecting Agent-Based Simulations with Multi-Agent Reinforcement Learning},
  journal = {Journal of Open Source Software}
}

Reporting Issues

Please use our issue tracker to report any bugs or submit feature requests. Great bug reports tend to have:

  • A quick summary and/or background
  • Steps to reproduce, sample code is best.
  • What you expected would happen
  • What actually happens

Contributing

Please submit contributions via pull requests from a forked repository. Find out more about this process here. All contributions are under the BSD 3 License that covers the project.

Release

LLNL-CODE-815883

About

Agent Based Modeling and Reinforcement Learning

Topics

Resources

License

Stars

Watchers

Forks

Packages

No packages published