AWS Inferentia Description

AWS Inferentia Accelerators are designed by AWS for high performance and low cost for deep learning (DL), inference applications. The first-generation AWS Inferentia accelerator powers Amazon Elastic Compute Cloud, Amazon EC2 Inf1 instances. These instances deliver up to 2.3x more throughput and up 70% lower cost per input than comparable GPU-based Amazon EC2 instances. Inf1 instances have been adopted by many customers including Snap, Sprinklr and Money Forward. They have seen the performance and cost savings. The first-generation Inferentia features 8 GB of DDR4 memory per accelerator, as well as a large amount on-chip memory. Inferentia2 has 32 GB of HBM2e, which increases the total memory by 4x and memory bandwidth 10x more than Inferentia.

Integrations

No Integrations at this time

Reviews

Total
ease
features
design
support

No User Reviews. Be the first to provide a review:

Write a Review

Company Details

Company:
Amazon
Year Founded:
2006
Headquarters:
United States
Website:
aws.amazon.com/machine-learning/inferentia/

Media

AWS Inferentia Screenshot 1
Recommended Products
Top Rated Business VoIP Provider for 2024 for as low as $20/mo*! Icon
Top Rated Business VoIP Provider for 2024 for as low as $20/mo*!

Message, video, and phone on any device. Trusted by over 400,000 businesses.

- Includes 100+ Premium Features
- Unlimited Calling, Faxing, SMS, Conferencing.

Product Details

Platforms
SaaS
Type of Training
Documentation
Customer Support
Online

AWS Inferentia Features and Options

Deep Learning Software

Convolutional Neural Networks
Document Classification
Image Segmentation
ML Algorithm Library
Model Training
Neural Network Modeling
Self-Learning
Visualization

Infrastructure-as-a-Service (IaaS) Provider

Analytics / Reporting
Configuration Management
Data Migration
Data Security
Load Balancing
Log Access
Network Monitoring
Performance Monitoring
SLA Monitoring