You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Running the Double-Dueling-DQN code results in a network, that stops learning after about 2000 episodes., i.e. the game results do not get better. Run the GridWorld example now four different times - and tried to adapt parameters: all result in mostly the same picture. The network has some 'good' learning curve at the beginning and then stops learning.
For some results see:
Running the Double-Dueling-DQN code results in a network, that stops learning after about 2000 episodes., i.e. the game results do not get better. Run the GridWorld example now four different times - and tried to adapt parameters: all result in mostly the same picture. The network has some 'good' learning curve at the beginning and then stops learning.
For some results see:
I also started using Breakout-v0 - with mostly the same result.
Does anybody have an idea? Which parameters can be adapted to get better results?
The text was updated successfully, but these errors were encountered: