diff --git a/beginner_source/dist_overview.rst b/beginner_source/dist_overview.rst index 502961e20c3..584a5aa273a 100644 --- a/beginner_source/dist_overview.rst +++ b/beginner_source/dist_overview.rst @@ -70,7 +70,7 @@ When deciding what parallelism techniques to choose for your model, use these co #. Use `DistributedDataParallel (DDP) `__, if your model fits in a single GPU but you want to easily scale up training using multiple GPUs. - * Use `torchrun `__, to launch multiple pytorch processes if you are you using more than one node. + * Use `torchrun `__, to launch multiple pytorch processes if you are using more than one node. * See also: `Getting Started with Distributed Data Parallel <../intermediate/ddp_tutorial.html>`__