AWS InferentiaAmazon
|
||||||
Related Products
|
||||||
About
AWS Inferentia accelerators are designed by AWS to deliver high performance at the lowest cost for your deep learning (DL) inference applications. The first-generation AWS Inferentia accelerator powers Amazon Elastic Compute Cloud (Amazon EC2) Inf1 instances, which deliver up to 2.3x higher throughput and up to 70% lower cost per inference than comparable GPU-based Amazon EC2 instances. Many customers, including Airbnb, Snap, Sprinklr, Money Forward, and Amazon Alexa, have adopted Inf1 instances and realized its performance and cost benefits. The first-generation Inferentia has 8 GB of DDR4 memory per accelerator and also features a large amount of on-chip memory. Inferentia2 offers 32 GB of HBM2e per accelerator, increasing the total memory by 4x and memory bandwidth by 10x over Inferentia.
|
About
Deep learning frameworks such as TensorFlow, PyTorch, Caffe, Torch, Theano, and MXNet have contributed to the popularity of deep learning by reducing the effort and skills needed to design, train, and use deep learning models. Fabric for Deep Learning (FfDL, pronounced “fiddle”) provides a consistent way to run these deep-learning frameworks as a service on Kubernetes. The FfDL platform uses a microservices architecture to reduce coupling between components, keep each component simple and as stateless as possible, isolate component failures, and allow each component to be developed, tested, deployed, scaled, and upgraded independently. Leveraging the power of Kubernetes, FfDL provides a scalable, resilient, and fault-tolerant deep-learning framework. The platform uses a distribution and orchestration layer that facilitates learning from a large amount of data in a reasonable amount of time across compute nodes.
|
|||||
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
|||||
Audience
Companies searching for an advanced Deep Learning solution
|
Audience
Neural networks experts looking for a solution to run deep-learning frameworks as a service on Kubernetes
|
|||||
Support
Phone Support
24/7 Live Support
Online
|
Support
Phone Support
24/7 Live Support
Online
|
|||||
API
Offers API
|
API
Offers API
|
|||||
Screenshots and Videos |
Screenshots and Videos |
|||||
Pricing
No information available.
Free Version
Free Trial
|
Pricing
No information available.
Free Version
Free Trial
|
|||||
Reviews/
|
Reviews/
|
|||||
Training
Documentation
Webinars
Live Online
In Person
|
Training
Documentation
Webinars
Live Online
In Person
|
|||||
Company InformationAmazon
Founded: 2006
United States
aws.amazon.com/machine-learning/inferentia/
|
Company InformationIBM
Founded: 1911
United States
developer.ibm.com/open/projects/fabric-for-deep-learning-ffdl/
|
|||||
Alternatives |
Alternatives |
|||||
|
|
|
|||||
|
|
||||||
|
|
||||||
|
|
||||||
Categories |
Categories |
|||||
Integrations
AWS EC2 Trn3 Instances
AWS Parallel Computing Service
Amazon EC2 Inf1 Instances
Amazon EC2 Trn1 Instances
Anyscale
Caffe
Kubernetes
PyTorch
TensorFlow
Torch
|
Integrations
AWS EC2 Trn3 Instances
AWS Parallel Computing Service
Amazon EC2 Inf1 Instances
Amazon EC2 Trn1 Instances
Anyscale
Caffe
Kubernetes
PyTorch
TensorFlow
Torch
|
|||||
|
|
|