AWS InferentiaAmazon
|
||||||
Related Products
|
||||||
About
AWS Inferentia accelerators are designed by AWS to deliver high performance at the lowest cost for your deep learning (DL) inference applications. The first-generation AWS Inferentia accelerator powers Amazon Elastic Compute Cloud (Amazon EC2) Inf1 instances, which deliver up to 2.3x higher throughput and up to 70% lower cost per inference than comparable GPU-based Amazon EC2 instances. Many customers, including Airbnb, Snap, Sprinklr, Money Forward, and Amazon Alexa, have adopted Inf1 instances and realized its performance and cost benefits. The first-generation Inferentia has 8 GB of DDR4 memory per accelerator and also features a large amount of on-chip memory. Inferentia2 offers 32 GB of HBM2e per accelerator, increasing the total memory by 4x and memory bandwidth by 10x over Inferentia.
|
About
TensorWave is an AI and high-performance computing (HPC) cloud platform purpose-built for performance, powered exclusively by AMD Instinct Series GPUs. It delivers high-bandwidth, memory-optimized infrastructure that scales with your most demanding models, training, or inference. TensorWave offers access to AMD’s top-tier GPUs within seconds, including the MI300X and MI325X accelerators, which feature industry-leading memory capacity and bandwidth, with up to 256GB of HBM3E supporting 6.0TB/s. TensorWave's architecture includes UEC-ready capabilities that optimize the next generation of Ethernet for AI and HPC networking, and direct liquid cooling that delivers exceptional total cost of ownership with up to 51% data center energy cost savings. TensorWave provides high-speed network storage, ensuring game-changing performance, security, and scalability for AI pipelines. It offers plug-and-play compatibility with a wide range of tools and platforms, supporting models, libraries, etc.
|
|||||
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
|||||
Audience
Companies searching for an advanced Deep Learning solution
|
Audience
AI infrastructure architects in need of a solution to support demanding AI and machine learning workloads
|
|||||
Support
Phone Support
24/7 Live Support
Online
|
Support
Phone Support
24/7 Live Support
Online
|
|||||
API
Offers API
|
API
Offers API
|
|||||
Screenshots and Videos |
Screenshots and Videos |
|||||
Pricing
No information available.
Free Version
Free Trial
|
Pricing
No information available.
Free Version
Free Trial
|
|||||
Reviews/
|
Reviews/
|
|||||
Training
Documentation
Webinars
Live Online
In Person
|
Training
Documentation
Webinars
Live Online
In Person
|
|||||
Company InformationAmazon
Founded: 2006
United States
aws.amazon.com/machine-learning/inferentia/
|
Company InformationTensorWave
United States
tensorwave.com
|
|||||
Alternatives |
Alternatives |
|||||
|
|
||||||
|
|
||||||
|
|
|
|||||
|
|
|
|||||
Categories |
Categories |
|||||
Integrations
AMD Radeon ProRender
AWS EC2 Trn3 Instances
AWS Parallel Computing Service
Amazon EC2 Inf1 Instances
Amazon EC2 Trn1 Instances
Anyscale
Axolotl
Hugging Face
LLaMA-Factory
Meta AI
|
Integrations
AMD Radeon ProRender
AWS EC2 Trn3 Instances
AWS Parallel Computing Service
Amazon EC2 Inf1 Instances
Amazon EC2 Trn1 Instances
Anyscale
Axolotl
Hugging Face
LLaMA-Factory
Meta AI
|
|||||
|
|
|