Skip to content

confusion in Tutorials > Quantization Recipe #2725

Closed
@XA23i

Description

@XA23i

In 3. Post Training Static Quantization
" This method converts both the weights and the activations to 8-bit integers beforehand so there won’t be on-the-fly conversion on the activations during the inference, as the dynamic quantization does, hence improving the performance significantly."

but I think the performance of Post Training Static Quantization is weaker than Post Training dynamic quantization since dynamic quantization can calibrate for each input.

cc @jerryzh168 @z-a-f @vkuzo

Metadata

Metadata

Assignees

No one assigned

    Labels

    arch-optimizationquantization, sparsity, nsquantizationIssues relating to quantization tutorials

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions