Overview
Deepgram powers end to end voice solutions on AWS, from real time transcription to lifelike speech synthesis and interruptible, human like voice agents. Deploy our STT/TTS and agent runtime where you need them: In SageMaker, in a Deepgram managed Dedicated environment or self hosted inside your AWS VPC for maximum control and compliance, with native touchpoints to Amazon Bedrock and Amazon Connect to compose complete voice workflows.
Procure through AWS Marketplace to accelerate onboarding with usage based pricing and consolidated billing on your AWS invoice, ideal for trials, POCs, and scaling to production while aligning to AWS commitments.
Deepgrams AWS alignment includes the AWS Generative AI Competency and a multi year strategic collaboration, giving teams confidence that integrations, cosell, and global scale on AWS are first class.
Use cases include: real time contact center transcription and automation with Amazon Connect + Lex, Bedrock powered voice agents with Deepgram STT/TTS, and streaming/batch analytics via S3, API Gateway, Lambda, and EKS/EC2, all built on the AWS patterns your teams already trust.
Highlights
- Real time STT for human like conversations: Sub 300 ms streaming latency with industry leading accuracy (Nova 3: 6.84% median WER streaming; 5.26% batch) to keep pace with fast, noisy speech.
- Natural, low latency TTS: Sub 250 ms responses with lifelike speech and streaming delivery for natural turn taking in real time.
- Production ready voice agents on AWS: Combine Deepgram STT/TTS with Amazon Bedrock for reasoning and Amazon Connect and Lex for contact center workflows supporting interruptible, human like dialogs at scale alongside real time STT voice agents and real time STT for transcription.
Details
Unlock automation with AI agent solutions

Features and programs
Financing for AWS Marketplace purchases
Pricing
Free trial
Dimension | Description | Cost/host/hour |
|---|---|---|
ml.m5.xlarge Inference (Batch) Recommended | Model inference on the ml.m5.xlarge instance type, batch mode | $3.90 |
ml.g6.2xlarge Inference (Real-Time) Recommended | Model inference on the ml.g6.2xlarge instance type, real-time mode | $16.90 |
ml.g6e.2xlarge Inference (Real-Time) | Model inference on the ml.g6e.2xlarge instance type, real-time mode | $32.50 |
ml.g5.2xlarge Inference (Real-Time) | Model inference on the ml.g5.2xlarge instance type, real-time mode | $14.95 |
ml.g4dn.2xlarge Inference (Real-Time) | Model inference on the ml.g4dn.2xlarge instance type, real-time mode | $3.90 |
Vendor refund policy
No refunds.
How can we make this page better?
Legal
Vendor terms and conditions
Content disclaimer
Delivery details
Amazon SageMaker model
An Amazon SageMaker model package is a pre-trained machine learning model ready to use without additional training. Use the model package to create a model on Amazon SageMaker for real-time inference or batch processing. Amazon SageMaker is a fully managed platform for building, training, and deploying machine learning models at scale.
Version release notes
V1 Release
Additional details
Inputs
- Summary
Binary audio payload for transcription and link to docs. https://developers.deepgram.com/reference/speech-to-text/listen-streamingÂ
- Input MIME type
- audio/*, *
Support
Vendor support
Basic support is provided through email (aws@deepgram.com ). Premium and VIP support packages are also available for enterprise clients.
AWS infrastructure support
AWS Support is a one-on-one, fast-response support channel that is staffed 24x7x365 with experienced and technical support engineers. The service helps customers of all sizes and technical abilities to successfully utilize the products and features provided by Amazon Web Services.
Similar products
