Skip to content

Commit 7d0df27

Browse files
author
jmarin
committed
Improve paragraph following suggestions
1 parent 0ce9191 commit 7d0df27

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

recipes_source/quantization.rst

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -81,7 +81,7 @@ The full documentation of the `quantize_dynamic` API call is `here <https://pyto
8181
3. Post Training Static Quantization
8282
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
8383

84-
This method converts both the weights and the activations to 8-bit integers beforehand so there won’t be on-the-fly conversion on the activations during the inference, as the dynamic quantization does. While post-training static quantization can significantly enhance inference speed and reduce model size, this method may degrade the original model's accuracy. Converting weights and activation functions to 8-bit integers can slightly alter the network's behavior and activation responses, leading to a reduction in the model's original effectiveness.
84+
This method converts both the weights and the activations to 8-bit integers beforehand so there won’t be on-the-fly conversion on the activations during the inference, as the dynamic quantization does. While post-training static quantization can significantly enhance inference speed and reduce model size, this method may degrade the original model's accuracy more compared to post training dynamic quantization.
8585

8686
To apply static quantization on a model, run the following code:
8787

0 commit comments

Comments
 (0)