@@ -176,8 +176,7 @@ def evaluate(mod, inp):
176
176
177
177
######################################################################
178
178
# And indeed, we can see that running our model with ``torch.compile``
179
- # results in a significant speedup. On an NVIDIA A100 GPU, we observe a
180
- # ~1.5x speedup. Speedup mainly comes from reducing Python overhead and
179
+ # results in a significant speedup. Speedup mainly comes from reducing Python overhead and
181
180
# GPU read/writes, and so the observed speedup may vary on factors such as model
182
181
# architecture and batch size. For example, if a model's architecture is simple
183
182
# and the amount of data is large, then the bottleneck would be
@@ -234,8 +233,7 @@ def train(mod, data):
234
233
######################################################################
235
234
# Again, we can see that ``torch.compile`` takes longer in the first
236
235
# iteration, as it must compile the model, but in subsequent iterations, we see
237
- # significant speedups compared to eager. On an NVIDIA A100 GPU, we
238
- # observe a ~1.8x speedup.
236
+ # significant speedups compared to eager.
239
237
240
238
######################################################################
241
239
# Comparison to TorchScript and FX Tracing
0 commit comments