Skip to content

Commit 43dbe1e

Browse files
committed
Fixed typos
1 parent 3c4e84d commit 43dbe1e

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

recipes_source/recipes/tuning_guide.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -48,7 +48,7 @@
4848
# Disable bias for convolutions directly followed by a batch norm
4949
# ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
5050
# `torch.nn.Conv2d() <https://pytorch.org/docs/stable/generated/torch.nn.Conv2d.html#torch.nn.Conv2d>`_
51-
# has ``bias`` parameter which defaults to ``True`` (same is true for
51+
# has ``bias`` parameter which defaults to ``True`` (the same is true for
5252
# `Conv1d <https://pytorch.org/docs/stable/generated/torch.nn.Conv1d.html#torch.nn.Conv1d>`_
5353
# and
5454
# `Conv3d <https://pytorch.org/docs/stable/generated/torch.nn.Conv3d.html#torch.nn.Conv3d>`_
@@ -84,7 +84,7 @@
8484
###############################################################################
8585
# The second code snippet does not zero the memory of each individual parameter,
8686
# also the subsequent backward pass uses assignment instead of addition to store
87-
# gradients, this reduces number of memory operations.
87+
# gradients, this reduces the number of memory operations.
8888
#
8989
# Setting gradient to ``None`` has a slightly different numerical behavior than
9090
# setting it to zero, for more details refer to the
@@ -324,7 +324,7 @@ def fused_gelu(x):
324324
# ``DistributedDataParallel`` provides
325325
# `no_sync() <https://pytorch.org/docs/stable/generated/torch.nn.parallel.DistributedDataParallel.html#torch.nn.parallel.DistributedDataParallel.no_sync>`_
326326
# context manager which disables gradient all-reduce for particular iteration.
327-
# ``no_sync()`` should applied to first ``N-1`` iterations of gradient
327+
# ``no_sync()`` should be applied to first ``N-1`` iterations of gradient
328328
# accumulation, the last iteration should follow the default execution and
329329
# perform the required gradient all-reduce.
330330

0 commit comments

Comments
 (0)