From da7088df573766e510249b0c6c1007fc452c0d6f Mon Sep 17 00:00:00 2001 From: Jing Xu Date: Wed, 15 Nov 2023 20:13:32 +0900 Subject: [PATCH] add a note to fix https://github.com/pytorch/pytorch/issues/113522 --- intermediate_source/scaled_dot_product_attention_tutorial.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/intermediate_source/scaled_dot_product_attention_tutorial.py b/intermediate_source/scaled_dot_product_attention_tutorial.py index 2bfeb46b56c..083eb1a6656 100644 --- a/intermediate_source/scaled_dot_product_attention_tutorial.py +++ b/intermediate_source/scaled_dot_product_attention_tutorial.py @@ -119,7 +119,7 @@ def benchmark_torch_function_in_microseconds(f, *args, **kwargs): # # Depending on what machine you ran the above cell on and what hardware is # available, your results might be different. -# - If you don’t have a GPU and are running on CPU then the context manager +# - If you don’t have a GPU and are running on CPU then with FP32 the context manager # will have no effect and all three runs should return similar timings. # - Depending on what compute capability your graphics card supports # flash attention or memory efficient might have failed.