Skip to content

Commit e7e2525

Browse files
highlight the usage same as regular BF16 Autocast
1 parent c7398d2 commit e7e2525

File tree

1 file changed

+2
-0
lines changed

1 file changed

+2
-0
lines changed

prototype_source/pt2e_quant_ptq_x86_inductor.rst

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -175,6 +175,8 @@ In a more advanced scenario, int8-mixed-bf16 quantization comes into play. In th
175175
a Convolution or GEMM operator produces BFloat16 output data type instead of Float32 in the absence
176176
of a subsequent quantization node. Subsequently, the BFloat16 tensor seamlessly propagates through
177177
subsequent pointwise operators, effectively minimizing memory usage and potentially enhancing performance.
178+
The utilization of this feature mirrors that of regular BFloat16 Autocast, as simple as wrapping the
179+
script within the BFloat16 Autocast context.
178180

179181
::
180182

0 commit comments

Comments
 (0)