Explain the difference between Q-Learning and Deep Q-Learning
Q-Learning and Deep Q-Learning are both model-free reinforcement learning algorithms that take in a state, action, and reward to calculate a Q value which is used to determine the next action that should be taken to maximize the total return of an action-selection policy. They are model-free because neither algorithm contains a model that predicts future states given a Markov decision process.

Q-Learning uses a Q-Table to store the Q values for any given State-Action pair, which becomes costly in memory when the state or action space become large because it must store all possible combinations of both and update them as well at the end of any training episode.

Deep Q-Learning attempts to solve this memory limitation by using Deep Neural Networks to learn the Q values instead. This provides the benefits of being able to learn effectively larger Q tables, generalizing to states that were never seen during training, and being able to use continuous state spaces.