Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Randomness of Evaluation #8

Open
coderlemon17 opened this issue Mar 4, 2024 · 3 comments
Open

Randomness of Evaluation #8

coderlemon17 opened this issue Mar 4, 2024 · 3 comments

Comments

@coderlemon17
Copy link

Hi, thanks for providing the code. However, when I evaluate the agent's performance on the town05-long-benchmark, there is some randomness in the evaluation results, even with a fixed seed.

After checking the visualization results, I believe some of the randomness comes from the different behaviors of the NPC vehicles, but I'm not sure how this will happen with a fixed random seed. Am I doing something wrong or it's just normal? Any help will be appreciated!

@jiaxiaosong1002
Copy link
Collaborator

@coderlemon17 Yes. Exacty, this has been discussed widely in the community about the randomness. It can not be controlled.

@coderlemon17
Copy link
Author

@jiaxiaosong1002 Thanks for your reply. If this randomness does exist, how many evaluations will you conduct to measure the model's performance under one seed? And will this randomness affect the model's final performance a lot?

@jiaxiaosong1002
Copy link
Collaborator

@coderlemon17 Hi, 3 runs are generally used. Yes, it could be and thus we need results on multiple benchmarks.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants