diff --git a/recipes_source/recipes/tuning_guide.py b/recipes_source/recipes/tuning_guide.py index 43f67deffa0..fcc07595506 100644 --- a/recipes_source/recipes/tuning_guide.py +++ b/recipes_source/recipes/tuning_guide.py @@ -464,7 +464,7 @@ def fused_gelu(x): # perform the required gradient all-reduce. ############################################################################### -# Match the order of layers in constructors and during the execution if using ``DistributedDataParallel``(find_unused_parameters=True) +# Match the order of layers in constructors and during the execution if using ``DistributedDataParallel(find_unused_parameters=True)`` # ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ # `torch.nn.parallel.DistributedDataParallel `_ # with ``find_unused_parameters=True`` uses the order of layers and parameters