Skip to content

Commit 1bc2674

Browse files
authored
Merge pull request pytorch#545 from shiftlab/aws-keynote-updates
Update AWS getting started content
2 parents 536743a + 5d24230 commit 1bc2674

File tree

1 file changed

+9
-3
lines changed
  • _get_started/installation

1 file changed

+9
-3
lines changed

_get_started/installation/aws.md

Lines changed: 9 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -70,15 +70,21 @@ torch.cuda.is_available()
7070
## AWS Primer
7171
{: #aws-primer}
7272

73+
Generally, you will be using Amazon Elastic Compute Cloud (or [EC2](https://aws.amazon.com/ec2/?ec2-whats-new.sort-by=item.additionalFields.postDateTime&ec2-whats-new.sort-order=desc){:target="_blank"}) to spin up your instances. Amazon has various [instance types](https://aws.amazon.com/ec2/instance-types/){:target="_blank"}, each of which are configured for specific use cases. For PyTorch, it is highly recommended that you use the accelerated computing instances that feature GPUs or custom AI/ML accelerators as they are tailored for the high compute needs of machine learning.
74+
7375
In order to use AWS, you need to set up an [AWS account](https://aws.amazon.com/getting-started/){:target="_blank"}, if you do not have one already. You will create a username (your email address), password and an AWS account name (since you can create multiple AWS accounts for different purposes). You will also provide contact and billing information. The billing information is important because while AWS does provide what they call “free-tier” instances, to use PyTorch you will want more powerful, paid instances.
7476

75-
Once you are logged in, you will be brought to your [AWS console](https://aws.amazon.com/console/){:target="_blank"}. You can even learn more about AWS through a set of [simple tutorials](https://aws.amazon.com/getting-started/tutorials/){:target="_blank"}.
77+
Once you are logged in, you will be brought to your [AWS console](https://aws.amazon.com/console/){:target="_blank"}. You can even learn more about AWS through a set of [simple tutorials](https://aws.amazon.com/getting-started/tutorials/){:target="_blank"}.
78+
79+
### AWS Inferentia-based instances
80+
81+
[AWS Inferentia](https://aws.amazon.com/machine-learning/inferentia/){:target="_blank"} is a chip custom built by AWS to provide higher performance and low cost machine learning inference in the cloud. [Amazon EC2 Inf1 instances](https://aws.amazon.com/ec2/instance-types/inf1/){:target="_blank"} feature up to 16 AWS Inferentia chips, the latest second generation Intel Xeon Scalable processors, and up to 100 Gbps networking to enable high throughput and lowest cost inference in the cloud. You can use Inf1 instances with Amazon SageMaker for a fully managed workflow, or use the [AWS Neuron SDK](https://awsdocs-neuron.readthedocs-hosted.com/en/latest/){:target="_blank"} directly which is integrated with PyTorch.
7682

7783
### GPU-based instances
7884

79-
Generally, you will be using Amazon Elastic Compute Cloud (or [EC2](https://aws.amazon.com/ec2/){:target="_blank"}) to spin up your instances. Amazon has various [instance types](https://aws.amazon.com/ec2/instance-types/){:target="_blank"}, each of which are configured for specific use cases. For PyTorch, it is highly recommended that you use the accelerated computing, or [p3](https://aws.amazon.com/ec2/instance-types/p3/){:target="_blank"}, instances. They are tailored for the high compute needs of machine learning.
85+
[Amazon EC2 P4d instances](https://aws.amazon.com/ec2/instance-types/p4/) deliver the highest performance for machine learning training on AWS. They are powered by the latest NVIDIA A100 Tensor Core GPUs and feature first in the cloud 400 Gbps instance networking. P4d instances are deployed in hyperscale clusters called EC2 UltraClusters that are comprised of more than 4,000 NVIDIA A100 GPUs, Petabit-scale non-blocking networking, and scalable low latency storage with FSx for Lustre. Each EC2 UltraCluster provides supercomputer-class performance to enable you to solve the most complex multi-node ML training tasks.
8086

81-
The expense of your instance is directly correlated to the number of GPUs that it contains. The p3.2xlarge instance is the smallest, cheapest instance and can actually be suitable for many use cases.
87+
For ML inference, AWS Inferentia-based [Inf1](https://aws.amazon.com/ec2/instance-types/inf1/) instances provide the lowest cost inference in the cloud. Additionally, [Amazon EC2 G4dn instances](https://aws.amazon.com/ec2/instance-types/g4/) featuring NVIDIA T4 GPUs are optimized for GPU-based machine learning inference and small scale training that leverage NVIDIA libraries.
8288

8389
### Creating and Launching an Instance
8490

0 commit comments

Comments
 (0)