-
Notifications
You must be signed in to change notification settings - Fork 331
[Collaborative optimization] Pruning-Clustering-preserving Quantization Aware Training #699
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks, a minor request below.
tensorflow_model_optimization/python/core/clustering/keras/cluster_wrapper.py
Outdated
Show resolved
Hide resolved
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thank you.
Hi @daverim, recently we have removed the dependency of this PR, so it is ready for review anytime. Could you please take a look when you have time? |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Some small linting issues.
...ion/keras/collaborative_optimizations/cluster_preserve/cluster_preserve_quantize_registry.py
Outdated
Show resolved
Hide resolved
...ion/keras/collaborative_optimizations/cluster_preserve/cluster_preserve_quantize_registry.py
Show resolved
Hide resolved
...ion/keras/collaborative_optimizations/cluster_preserve/cluster_preserve_quantize_registry.py
Outdated
Show resolved
Hide resolved
...ion/keras/collaborative_optimizations/cluster_preserve/cluster_preserve_quantize_registry.py
Outdated
Show resolved
Hide resolved
...ollaborative_optimizations/cluster_preserve/default_8bit_cluster_preserve_quantize_scheme.py
Outdated
Show resolved
Hide resolved
…on Aware Training (PCQAT)
Hi, Just curious why Pruned_Clustered Model - Mobilenet_v1 (ImageNet) - INT8 .tflite gzip compression (bytes) is so small? PCQAT model is larger than PC model? or missed a digit for this case? |
Hi @Xhark Yes, this is the mistake, last digit is missing. The compression ratio was around 2.3 in our experiments. |
Thanks, @Xhark, the number is now updated. |
Merging was blocked by build file strict dependencies -- resubmitting with
it fixed myself now, should be merged today.
…On Fri, Jun 11, 2021 at 5:39 AM Ruomei Yan ***@***.***> wrote:
Hi, @daverim <https://github.com/daverim> and @Xhark
<https://github.com/Xhark>, could you please also let us know if there is
anything we can do to help with the failed internal checks shown in this PR?
@akarmi <https://github.com/akarmi> @wwwind <https://github.com/wwwind>
for visibility
Thanks all!
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
<#699 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AASV4JJOCS5IS2OIGPIZV53TSEPIDANCNFSM433XELBA>
.
|
Brill, thanks a lot, David. |
This PR adds support for Pruning-Clustering-preserving Quantization Aware Training (PCQAT). Aiming at preserving the sparsity and unique weights in the output optimized model, fixed pruning masks and the stochastic updates of clustering training variables are enabled during quantization-aware training.
User API:
Main changes:
Initial results (pruning sparsity: 50%, number of clusters: 8 (DS-CNN-L), 16 (Mobilenet_v1)):