AWS EC2 Trn3 InstancesAmazon
|
Amazon Elastic InferenceAmazon
|
|||||
Related Products
|
||||||
About
Amazon EC2 Trn3 UltraServers are AWS’s newest accelerated computing instances, powered by the in-house Trainium3 AI chips and engineered specifically for high-performance deep-learning training and inference workloads. These UltraServers are offered in two configurations, a “Gen1” with 64 Trainium3 chips and a “Gen2” with up to 144 Trainium3 chips per UltraServer. The Gen2 configuration delivers up to 362 petaFLOPS of dense MXFP8 compute, 20 TB of HBM memory, and a staggering 706 TB/s of aggregate memory bandwidth, making it one of the highest-throughput AI compute platforms available. Interconnects between chips are handled by a new “NeuronSwitch-v1” fabric to support all-to-all communication patterns, which are especially important for large models, mixture-of-experts architectures, or large-scale distributed training.
|
About
Amazon Elastic Inference allows you to attach low-cost GPU-powered acceleration to Amazon EC2 and Sagemaker instances or Amazon ECS tasks, to reduce the cost of running deep learning inference by up to 75%. Amazon Elastic Inference supports TensorFlow, Apache MXNet, PyTorch and ONNX models. Inference is the process of making predictions using a trained model. In deep learning applications, inference accounts for up to 90% of total operational costs for two reasons. Firstly, standalone GPU instances are typically designed for model training - not for inference. While training jobs batch process hundreds of data samples in parallel, inference jobs usually process a single input in real time, and thus consume a small amount of GPU compute. This makes standalone GPU inference cost-inefficient. On the other hand, standalone CPU instances are not specialized for matrix operations, and thus are often too slow for deep learning inference.
|
|||||
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
|||||
Audience
AI researchers, data scientists, and enterprises looking for a solution for training and deploying large language models, generative-AI systems, and other deep-learning workloads
|
Audience
IT teams that need an advanced Infrastructure as a Service solution
|
|||||
Support
Phone Support
24/7 Live Support
Online
|
Support
Phone Support
24/7 Live Support
Online
|
|||||
API
Offers API
|
API
Offers API
|
|||||
Screenshots and Videos |
Screenshots and Videos |
|||||
Pricing
No information available.
Free Version
Free Trial
|
Pricing
No information available.
Free Version
Free Trial
|
|||||
Reviews/
|
Reviews/
|
|||||
Training
Documentation
Webinars
Live Online
In Person
|
Training
Documentation
Webinars
Live Online
In Person
|
|||||
Company InformationAmazon
Founded: 1994
United States
aws.amazon.com/ec2/instance-types/trn3/
|
Company InformationAmazon
Founded: 2006
United States
aws.amazon.com/machine-learning/elastic-inference/
|
|||||
Alternatives |
Alternatives |
|||||
|
|
|
|||||
|
|
||||||
|
|
|
|||||
|
|
||||||
Categories |
Categories |
|||||
Integrations
Amazon Web Services (AWS)
PyTorch
AWS Batch
AWS Inferentia
AWS ParallelCluster
AWS Trainium
Amazon EC2
Amazon EC2 G4 Instances
Amazon EKS
Amazon Elastic Container Service (Amazon ECS)
|
Integrations
Amazon Web Services (AWS)
PyTorch
AWS Batch
AWS Inferentia
AWS ParallelCluster
AWS Trainium
Amazon EC2
Amazon EC2 G4 Instances
Amazon EKS
Amazon Elastic Container Service (Amazon ECS)
|
|||||
|
|
|