DQN implemented in keras. Learning to play Cartpole using Deep Q Learning.
- Replay Buffer to train using de-correlated batches
- Separate Target Network to predict estimated q-values
- Huber Loss for stability
- Dueling Networks
- Prioritized Experience Replay