Skip to content

Commit 203724e

Browse files
authored
[Docs] add note on fp16 in fast diffusion (#6380)
add note on fp16
1 parent e7044a4 commit 203724e

File tree

1 file changed

+2
-0
lines changed

1 file changed

+2
-0
lines changed

docs/source/en/tutorials/fast_diffusion.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -96,6 +96,8 @@ bfloat16 reduces the latency from 7.36 seconds to 4.63 seconds:
9696

9797
</div>
9898

99+
_(We later ran the experiments in float16 and found out that the recent versions of torchao do not incur numerical problems from float16.)_
100+
99101
**Why bfloat16?**
100102

101103
* Using a reduced numerical precision (such as float16, bfloat16) to run inference doesn’t affect the generation quality but significantly improves latency.

0 commit comments

Comments
 (0)