-
Notifications
You must be signed in to change notification settings - Fork 282
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
disable env.reset() after every episode #66
Comments
Commenting out line 147 would prevent the environment resetting during training. Commenting out line 130 would prevent the environment resetting during collection of data for validating Q-values. It seems that you might need to write a different environment and use a different set of functions if you want more control. |
In general validating Q-values and training should be consistent right (if one is reset another should also) |
Yes it makes sense to keep them consistent. |
when I commented the reset at line 130 as well, it gave me this error: when it ran line 132: Does this line just let agent make a random movement and what could be the possible reasons for the error? |
That line uses random actions to collect data for validating Q-values. I'm not sure why your edit is causing the error, so you will need to try and debug it yourself. |
Hi,
May I check if I would like to keep the environment as it is after each training episode, should I just comment line line 147 in main.py or should I also comment line 130? Besides what am I supposed to do if I just want to reset the agent's position but keep the environment as it is after each training episode?
Thank you.
The text was updated successfully, but these errors were encountered: