-
Notifications
You must be signed in to change notification settings - Fork 4.3k
Closed
Labels
arch-optimizationquantization, sparsity, nsquantization, sparsity, nsquantizationIssues relating to quantization tutorialsIssues relating to quantization tutorials
Description
In 3. Post Training Static Quantization
" This method converts both the weights and the activations to 8-bit integers beforehand so there won’t be on-the-fly conversion on the activations during the inference, as the dynamic quantization does, hence improving the performance significantly."
but I think the performance of Post Training Static Quantization is weaker than Post Training dynamic quantization since dynamic quantization can calibrate for each input.
Metadata
Metadata
Assignees
Labels
arch-optimizationquantization, sparsity, nsquantization, sparsity, nsquantizationIssues relating to quantization tutorialsIssues relating to quantization tutorials