You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
However, according to Fig 2 in the original paper (https://arxiv.org/pdf/2006.03647.pdf), the performance of HalfCheetah should be around 6000, which is quite different from the evaluation results.
I wonder the parameter setting specified in the command above is the same as the setting of experiments in this paper? If not, could you let me know which hyper-parameter should be modified in order to reproduce the results reported in the paper? Or maybe there are some other reasons for this performance gap?
Thanks a lot!
The text was updated successfully, but these errors were encountered:
cheetah_run uses a different reward function (based on DM Control) from HalfCheetah (based on Open AI Gym MuJoCo). If you reproduce the main results in the paper, you should use HalfCheetah environments (also cheetah_run result is described in Appendix).
Hi,
I tried to run deployment-efficient experiments to reproduce the results reported in the paper with the following command:
After the training is finished, I observed the following evaluation results:
However, according to Fig 2 in the original paper (https://arxiv.org/pdf/2006.03647.pdf), the performance of HalfCheetah should be around 6000, which is quite different from the evaluation results.
I wonder the parameter setting specified in the command above is the same as the setting of experiments in this paper? If not, could you let me know which hyper-parameter should be modified in order to reproduce the results reported in the paper? Or maybe there are some other reasons for this performance gap?
Thanks a lot!
The text was updated successfully, but these errors were encountered: