Three PyTorch experiments showing the Lottery Ticket Phenomena in three different contexts:
- Supervised Learning with the MNIST dataset and a FC network
- Reinforcement Learning with the Cartpole environment and a FC network
- Unsupervised Learning on genomics data with a Variational Autoencoder
Here is a plot of the average of 10 experiments in the context of reinforcement learning. Left are the pruned weights reinitialized with the same values than the unpruned weights before training. Right are the pruned weights randomly reinitialized. We clearly see that pruning and keeping the initialization values makes the agent closer to the solution while random reinitialization prevents the agent from converging to the solution.
Here is another example in context of unsupervised learning (code not included in this repo, see my deepsequence repo), where I prune the parameters of a variational autoencoder trained on genomics data and compute the Spearman coefficient between the prediction and the experimental measurement.