Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[docs] Add POCA to major features #5122

Merged
merged 5 commits into from Mar 16, 2021
Merged

Conversation

ervteng
Copy link
Contributor

@ervteng ervteng commented Mar 15, 2021

Proposed change(s)

Added POCA to major features on front readme page.

Types of change(s)

  • Bug fix
  • New feature
  • Code refactor
  • Breaking change
  • Documentation update
  • Other (please describe)

Checklist

  • Added tests that prove my fix is effective or that my feature works
  • Updated the changelog (if applicable)
  • Updated the documentation (if applicable)
  • Updated the migration guide (if applicable)

Other comments

@ervteng ervteng requested a review from mmattar March 15, 2021 23:51
Copy link

@mmattar mmattar left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I'm cool w/ this change. However, taking a step back, I wonder if having 4 bullets around different trainers is just too much detail. One idea (don't love it tbh) would be to have one bullet that says something like:

  • Multiplate advanced training algorithms to support DRL and IL for single-agent and cooperative or competitive multi-agent environments.

This would replace these 4:

  • Training using two deep reinforcement learning algorithms, Proximal Policy
    Optimization (PPO) and Soft Actor-Critic (SAC)
  • Multi-agent cooperative training using a novel deep reinforcement learning algorithm,
    Multi-Agent POsthumous Credit Assignment (MA-POCA)
  • Built-in support for Imitation Learning through Behavioral Cloning (BC) or
    Generative Adversarial Imitation Learning (GAIL)
  • Self-play mechanism for training agents in adversarial scenarios

cc: @surfnerd for thoughts.

@ervteng
Copy link
Contributor Author

ervteng commented Mar 16, 2021

@mmattar what if we split it into two lines, one for RL and one for IL? e.g.

  • Support for training single-agent, multi-agent cooperative, and multi-agent competitive scenarios via several deep reinforcement learning algorithms (PPO, SAC, MA-POCA).
  • Support for learning from demonstrations through two imitation learning algorithms (BC and GAIL).

I think the callout of the algorithms is more for researchers TBH, they would want to know which algorithms the toolkit support at a glance. We might not need to spell out the entire acronym for this audience (though POCA isn't a known/Googleable algorithm).

@mmattar
Copy link

mmattar commented Mar 16, 2021

@ervteng - I dig it!

@ervteng ervteng merged commit 3f02528 into release_15_branch Mar 16, 2021
@delete-merged-branch delete-merged-branch bot deleted the release_15_add_to_readme branch March 16, 2021 21:20
@github-actions github-actions bot locked as resolved and limited conversation to collaborators Mar 17, 2022
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

3 participants