This repository has been archived by the owner on May 1, 2023. It is now read-only.
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Add finer control over the pruning logic, to accomodate more pruning
use-cases.
The full description of the new logic will be available as documentation.
In this commit:
Added a new callback to the CompressionScheduler:
compression_scheduler.before_parameter_optimization
which is invokedafter the gradients are are computed, but before the weights are updated
by the optimizer.
We skip the first mini-batch of the first epoch (global_mini_batch_id == 0)
because of PyTorch's SGD implementation (details later).
We register to the parameter backward hook in order to mask the gradients.
This gives us finer control over the parameter updates.
Add several DropFilter schedules
DropFilter is a method to regularize networks, and it can also be
used to prepare a network for permanent filter pruning.
Add documentation of pruning fine-control