Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[RL-baseline] Model v5, experiment #3 #46

Open
wants to merge 3 commits into
base: RL-baseline-v5
Choose a base branch
from

Conversation

ziritrion
Copy link
Collaborator

Action set #2 was chosen for this experiment:
[0.0, 0.0, 0.0], # no action
[0.0, 0.8, 0.0], # throttle
[0.0, 0.0, 0.6], # break
[-0.9, 0.0, 0.0], # left
[0.9, 0.0, 0.0], # right

The Running Reward oscillated between 200 up until the 20k episode mark, where it suddenly dropped. Since the other experiments had experienced long intervals of low reward during training, I decided to train up to 30k episodes to see if the RR could improve, but the result was disappointing.

Final Running Reward was 72, with an achieved max of 379 around the 4k episode mark.

Tensorboard screenshots below:
TensorBoard
TensorBoard

Sample video below:
https://user-images.githubusercontent.com/1465235/113551628-44479480-95f5-11eb-9a6f-dc3a6e61ced5.mp4

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

1 participant