Amazon Elastic Inference Description

Amazon Elastic Inference allows for low-cost GPU-powered acceleration to Amazon EC2 instances and Sagemaker instances, or Amazon ECS tasks. This can reduce the cost of deep learning inference by up 75%. Amazon Elastic Inference supports TensorFlow and Apache MXNet models. Inference is the process by which a trained model makes predictions. Inference can account for as much as 90% of total operational expenses in deep learning applications for two reasons. First, standalone GPU instances are usually used for model training and not inference. Inference jobs typically process one input at a time and use a smaller amount of GPU compute. Training jobs can process hundreds of data samples simultaneously, but inference jobs only process one input in real-time. This makes standalone GPU-based inference expensive. However, standalone CPU instances aren't specialized for matrix operations and are therefore often too slow to perform deep learning inference.

Integrations

Reviews

Total
ease
features
design
support

No User Reviews. Be the first to provide a review:

Write a Review

Company Details

Company:
Amazon
Year Founded:
2006
Headquarters:
United States
Website:
aws.amazon.com/machine-learning/elastic-inference/

Media

Recommended Products
Secure your business by securing your people. Icon
Secure your business by securing your people.

Over 100,000 businesses trust 1Password

Take the guesswork out of password management, shadow IT, infrastructure, and secret sharing so you can keep your people safe and your business moving.

Product Details

Platforms
SaaS
Type of Training
Documentation
Customer Support
Online

Amazon Elastic Inference Features and Options

Infrastructure-as-a-Service (IaaS) Provider

Analytics / Reporting
Configuration Management
Data Migration
Data Security
Load Balancing
Log Access
Network Monitoring
Performance Monitoring
SLA Monitoring