We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
2 parents 5427598 + 9dbc8fc commit a912b81Copy full SHA for a912b81
beginner_source/examples_nn/two_layer_net_optim.py
@@ -33,7 +33,7 @@
33
34
# Use the optim package to define an Optimizer that will update the weights of
35
# the model for us. Here we will use Adam; the optim package contains many other
36
-# optimization algoriths. The first argument to the Adam constructor tells the
+# optimization algorithms. The first argument to the Adam constructor tells the
37
# optimizer which Tensors it should update.
38
learning_rate = 1e-4
39
optimizer = torch.optim.Adam(model.parameters(), lr=learning_rate)
0 commit comments