-
Notifications
You must be signed in to change notification settings - Fork 148
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Regarding inferencing the learnt policy #69
Comments
It is doable. However, MARLlib decides not to incorporate the loading and rendering functions as we find it hard to unify all ten environments to render in a similar pattern. We would like to provide you with instructions on how to implement this. There is a thorough solution provided by Sven: multiagent-load-only-one-policy-from-checkpoint. Any further question is welcome. We are happy to help you out. |
Hi Thanks for the previous answer. But we have been unable to use the learned policy to compute actions for our agents. Our objective is to compute agent's actions based on the learnt policy. But when we use the function agent.compute_single_action(obs), where obs = env.reset() It would be really helpful if you could provide some insights for the same. Also kindly let us know if we should be using some function other than agent.compute_single_action for the same purpose. |
Hi SiddSS, Sorry for the late reply. Check out mamujoco example and mpe example for loading the checkpoint and rendering the environment in MARLlib. Siyi |
New APIs are here for guiding how to render the pretrained model. |
Hi,
We have created our custom environment for and wrapped it in a gym class. After training using MAPPO, we got the .pkl files. Can you elaborate upon how to inference the learned policy ?
We already have a visualization of the env using pygame and just want to load the learned policies and see them play.
Thanks in advance.
The text was updated successfully, but these errors were encountered: