We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
instead of # self-attention layers in nn.TransformerEncoder are only allowed to attend,
nn.TransformerEncoder
nn.TransformerDecoder
Decoder rather than Encoder
cc @svekars @carljparker