Q: What is Amazon Elastic Inference?
A: Amazon Elastic Inference (Amazon EI) is an accelerated compute service that allows you to attach just the right amount of GPU-powered inference acceleration to any Amazon EC2 or Amazon SageMaker instance type. This means you can now choose the instance type that is best suited to the overall compute, memory, and storage needs of your application, and then separately configure the amount of inference acceleration that you need.
Q: What are Amazon Elastic inference accelerators?
A: Amazon Elastic inference accelerators are GPU-powered hardware devices that are designed to work with any EC2 instance type to accelerate deep learning inference workloads at a low cost. When you launch an instance with Amazon Elastic Inference, an accelerator is provisioned and attached to the instance over the network. Deep learning tools and frameworks like TensorFlow Serving and Apache MXNet, that are enabled for Amazon Elastic Inference, can automatically detect and offload model computation to the attached accelerator.
Q: How do I provision Amazon Elastic Inference accelerators?
A: You can configure Amazon SageMaker endpoints or Amazon EC2 instances with Amazon Elastic Inference accelerators using the AWS management console, AWS command line interface (CLI), or the AWS SDK. There are two requirements for launching EC2 instances with accelerators. First, you will need to provision an AWS PrivateLink VPC Endpoint for the subnets where you plan to launch accelerators. Second, as you launch an instance, you need to provide an instance role with a policy that allows users accessing the instance to connect to accelerators. When you configure an instance to launch with Amazon EI, an accelerator is provisioned in the same Availability Zone behind the VPC endpoint.
Q: What model formats does Amazon Elastic Inference support?
A: Amazon Elastic Inference supports models trained using TensorFlow, Apache MXNet, and ONNX models.
Q: Can I deploy models on Amazon Elastic Inference using TensorFlow or Apache MXNet frameworks?
A: Yes, you can use AWS-enhanced TensorFlow Serving and Apache MXNet libraries to deploy models and make inference calls.
Q: How do I get access to AWS optimized frameworks?
A: The AWS Deep Learning AMIs include the latest releases of TensorFlow Serving and Apache MXNet that are optimized for use with Amazon Elastic Inference accelerators. You can also obtain the libraries via Amazon S3 to build your own AMIs or container images. Please see our documentation for more information.
Q: Can I use CUDA with Amazon Elastic Inference accelerators?
A: No. You can only use either the AWS-enhanced TensorFlow Serving or Apache MXNet libraries as an interface to Amazon Elastic Inference accelerators.
Pricing and billing
Q: How am I charged for Amazon Elastic Inference?
A: You pay only for the Amazon Elastic Inference accelerator hours you use. For more details, see the pricing page.
Q: Will I incur charges for AWS PrivateLink VPC Endpoints for the Amazon Elastic Inference service?
No. You will not incur charges for VPC Endpoints to the Amazon Elastic Inference service, as long as you have at least one instance configured with an accelerator, running in an Availability Zone where a VPC endpoint is provisioned.
See the Amazon Elastic Inference pricing page for detailed pricing information.
Instantly get access to the AWS Free Tier.
Get started with Amazon Elastic Inference on Amazon SageMaker or Amazon EC2.