-
Notifications
You must be signed in to change notification settings - Fork 632
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Running exps with Dreamer-V3 #147
Comments
Wow! That's pretty cool, although return/reward is really only a tertiary measure of how well the agent does. Also, I haven't been running the Baseline version for quite a while. Would you be able to give the Pufferlib version a whirl so we can compare results? Specifically, we are interested in how far it can get through the game. This is visualized nicely on a weird coldmap (wandb.ai/jsuarez) or heatmap (wandb.ai/xinpw8). Check our current runs' Overviews for the run parameters. Clone https://github.com/PufferAI/Pufferlib (current branch is 0.5) and https://github.com/PufferAI/pokegym (current branch is main). Or grab the Dockerized version, Puffertank. |
You'd have to let it run for 11M-20M before you can really tell the status, see experiments here: |
Hi guys, first of all what an awesome video you've done on YT!
I'm one of the maintainers of sheeprl and I'm here just to tell you that we're running experiments with Dreamer-V3 on the standard env.
Right now I have modified your env code inside sheeprl and in the future we want to try out also the v2.
This is what I'm getting right now in terms of rewards:
This is the configuration I'm using:
I don't know if those are good results, but I wanted to share them.
If you wanna try out something with SheepRL let us now 🐑.
Thank you again!
The text was updated successfully, but these errors were encountered: