DeepSparse Inference Runtime
Product Overview
Neural Magic's DeepSparse AMI is an inference runtime that allows you to create an EC2 instance capable of running state-of-the-art machine learning models with GPU-class performance on x86 instance types. This allows you to run your machine learning workloads without concern for specific hardware accelerators. Simply select from a broad range of instance types based on the performance and cost requirements of your use case and deploy.
The deployed DeepSparse instance also comes with built-in benchmarking capabilities to help you assess the performance and cost benefits of your deployed model in a variety of scenarios.
This AMI contains the DeepSparse Enterprise Cloud Distribution which allows for commercial deployments.
Version
By
Neural MagicCategories
Operating System
Linux/Unix, Amazon Linux 2
Delivery Methods