You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Prior to filing: check that this should be a bug instead of a feature request. Everything supported, including the compatible versions of TensorFlow, is listed in the overview page of each technique. For example, the overview page of quantization-aware training is here. An issue for anything not supported should be a feature request.
Describe the bug
Setting layer.trainable=False for a QuantizeWrapperV2 wrapped Dense layer doesn't convert all trainable weights to non-trainable.
System information
TensorFlow version (installed from source or binary):
2.12.0
TensorFlow Model Optimization version (installed from source or binary):
0.7.4
Python version:
3.10.11
Describe the expected behavior
Like the normal model, by setting layer.trainable=False for the quantized layer, the layer's weights should all become non-trainable.
Describe the current behavior
After setting layer.trainable=False for a quantized Dense layer, the Dense layer still contains trainable weights.
Screenshots
If applicable, add screenshots to help explain your problem.
Additional context
The output from the colab is below. Noted that by setting layer.trainable=False, the original model's Dense layer's weights become non-trainable. However, the dense_11/kernel:0 from the Dense layer in the quantized model is still trainable after setting layer.trainable=False.
trainable variables in model: 2
trainable variables in model after setting trainable to false: 0
trainable variables in quantized model: 1
trainable variables in quantized model after setting trainable to false: 1
dense_11/kernel:0
The text was updated successfully, but these errors were encountered:
Prior to filing: check that this should be a bug instead of a feature request. Everything supported, including the compatible versions of TensorFlow, is listed in the overview page of each technique. For example, the overview page of quantization-aware training is here. An issue for anything not supported should be a feature request.
Describe the bug
Setting
layer.trainable=False
for aQuantizeWrapperV2
wrappedDense
layer doesn't convert all trainable weights to non-trainable.System information
TensorFlow version (installed from source or binary):
2.12.0
TensorFlow Model Optimization version (installed from source or binary):
0.7.4
Python version:
3.10.11
Describe the expected behavior
Like the normal model, by setting
layer.trainable=False
for the quantized layer, the layer's weights should all become non-trainable.Describe the current behavior
After setting
layer.trainable=False
for a quantizedDense
layer, theDense
layer still contains trainable weights.Code to reproduce the issue
The colab contains the code to reproduce this bug.
https://colab.research.google.com/drive/1KYnZkBI_g3Pu9Vqz4UneXCtNOs3_kB39?usp=sharing
Screenshots
If applicable, add screenshots to help explain your problem.
Additional context
The output from the colab is below. Noted that by setting
layer.trainable=False
, the original model'sDense
layer's weights become non-trainable. However, thedense_11/kernel:0
from theDense
layer in the quantized model is still trainable after settinglayer.trainable=False
.The text was updated successfully, but these errors were encountered: