@@ -288,10 +288,10 @@ def catch_error(fn):
288
288
# Autograd support
289
289
# ----------------
290
290
#
291
- # Autograd currently supports named tensors in a limited manner: autograd
292
- # ignores names on all tensors. Gradient computation is still correct but we
293
- # lose the safety that names give us. It is on the roadmap to introduce
294
- # handling of names to autograd.
291
+ # Autograd currently ignores names on all tensors and just treats them like
292
+ # regular tensors. Gradient computation is correct but we lose the safety that
293
+ # names give us. It is on the roadmap to introduce handling of names to
294
+ # autograd.
295
295
296
296
x = torch .randn (3 , names = ('D' ,))
297
297
weight = torch .randn (3 , names = ('D' ,), requires_grad = True )
@@ -316,8 +316,8 @@ def catch_error(fn):
316
316
# Other supported (and unsupported) features
317
317
# ------------------------------------------
318
318
#
319
- # See here (link to be included) for a detailed breakdown of what is
320
- # supported with the 1.3 release.
319
+ # ` See here <https://pytorch.org/docs/stable/named_tensor.html>`_ for a
320
+ # detailed breakdown of what is supported with the 1.3 release.
321
321
#
322
322
# In particular, we want to call out three important features that are not
323
323
# currently supported:
@@ -346,7 +346,7 @@ def fn(x):
346
346
# Now we'll go through a complete example of implementing a common
347
347
# PyTorch ``nn.Module``: multi-head attention. We assume the reader is already
348
348
# familiar with multi-head attention; for a refresher, check out
349
- # `this explanation <https://nlp.seas.harvard.edu/2018/04/03/attention.html>` _
349
+ # `this explanation <https://nlp.seas.harvard.edu/2018/04/03/attention.html>`_
350
350
# or
351
351
# `this explanation <http://jalammar.github.io/illustrated-transformer/>`_.
352
352
#
0 commit comments