Skip to content

Commit 8c48ada

Browse files
jmarinturjmarin
and
jmarin
authored
Clarify post training quantization behaviour (#2778)
* Clarify dynamic quantization behaviour Co-authored-by: jmarin <javier.marin@satellogic.com>
1 parent 630c2e2 commit 8c48ada

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

recipes_source/quantization.rst

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -81,7 +81,7 @@ The full documentation of the `quantize_dynamic` API call is `here <https://pyto
8181
3. Post Training Static Quantization
8282
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
8383

84-
This method converts both the weights and the activations to 8-bit integers beforehand so there won't be on-the-fly conversion on the activations during the inference, as the dynamic quantization does, hence improving the performance significantly.
84+
This method converts both the weights and the activations to 8-bit integers beforehand so there wont be on-the-fly conversion on the activations during the inference, as the dynamic quantization does. While post-training static quantization can significantly enhance inference speed and reduce model size, this method may degrade the original model's accuracy more compared to post training dynamic quantization.
8585

8686
To apply static quantization on a model, run the following code:
8787

0 commit comments

Comments
 (0)