Skip to content
This repository has been archived by the owner on Jan 27, 2023. It is now read-only.

Revisit Agent classes #38

Closed
kngwyu opened this issue Nov 6, 2019 · 1 comment
Closed

Revisit Agent classes #38

kngwyu opened this issue Nov 6, 2019 · 1 comment
Labels
enhancement New feature or request

Comments

@kngwyu
Copy link
Owner

kngwyu commented Nov 6, 2019

Now we have two types of agents

  • OneStepAgent
    • for DQN-like algorithms
    • execute 1-step + stores transition to replay buffer + train agent by sampled transitions
  • NStepParallelAgent
    • for A2C-like algorithms
    • execute N-step in parallel environments + train the policy in an online manner
      These 2 divisions are practical but lack flexibility.
      E.g., we cannot extend OneStep algorithms to batched-parallel style without rewriting the whole process.

So we should re-define agent hierarchies using some important properties, like

  • Online/Offline(or use replay buffer or not)
  • MultiStep/OneStep
  • Not Parallel/Batch Parallel/ Async Parallel
@kngwyu kngwyu added the enhancement New feature or request label Nov 6, 2019
@kngwyu
Copy link
Owner Author

kngwyu commented Jan 1, 2020

Closed via #54

@kngwyu kngwyu closed this as completed Jan 1, 2020
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

1 participant