Overview
Most AI tools and applications only support OpenAI or Anthropic APIs and cannot connect to AWS Bedrock or AWS AI services like Transcribe, Polly, and Translate. stdapi.ai solves this with a production-ready OpenAI and Anthropic compatible AI gateway that makes AWS Bedrock and AI services instantly accessible to any existing tool - Open WebUI, n8n, Continue.dev, Cursor, LangChain, and 1000+ applications. No custom integration required.
WHY CHOOSE STDAPI.AI
Complete OpenAI & Anthropic API Support - Chat, embeddings, images, audio (speech/transcription/translation). Works with your existing applications out of the box.
80+ AI Models - Access Claude, Kimi K2, MiniMax M2.5, Qwen3, Nova 2, Llama 4, and more. Switch models instantly with no vendor lock-in.
Enterprise Compliance - All data stays in your AWS account. Configure allowed regions for HIPAA, GDPR, and FedRAMP compliance through AWS Bedrock certifications.
Purpose-Built for AWS Bedrock - Not a generic proxy. Deep integration with reasoning modes, prompt caching, guardrails, service tiers, inference profiles, and prompt routers. Built to leverage every Bedrock feature through standard API parameters.
MODELS AND CAPABILITIES
AI Models - Anthropic Claude 4.6+ (reasoning), Moonshot Kimi K2 (reasoning), MiniMax M2.5, Amazon Nova 2 (cost-effective), Meta Llama 4, Alibaba Qwen3, Zhipu GLM 5, Stability AI (images), Mistral, Cohere, Nvidia, AI21 Labs.
AWS Services - Polly (text-to-speech), Transcribe (speech-to-text with diarization), Translate (multi-language).
Advanced Features - Reasoning modes, prompt caching, guardrails, inference profiles, cross-region routing, prompt routers.
POPULAR USE CASES
Private ChatGPT Alternative - Deploy Open WebUI or LibreChat with enterprise security. Multi-modal chat, RAG document analysis, web search. Full data control.
Workflow Automation - Connect AI to 400+ services via n8n, Make, or Zapier. Automate support, content creation, and data processing without technical setup.
AI Coding Assistants - Use Continue.dev, Cline, Cursor, Claude Code, Windsurf in VS Code or JetBrains IDEs. Access Claude, Kimi K2, Qwen Coder via OpenAI or Anthropic-compatible interface.
Knowledge Management - Add AI to Obsidian, Notion, Logseq. Semantic search, auto-summarization, smart organization.
Team Bots - Deploy to Slack, Discord, Teams, Telegram. Q&A bots, documentation search, task automation.
Autonomous Agents - Build with OpenClaw, AutoGPT, LangGraph, CrewAI. Personal AI assistants, research agents, multi-agent systems, self-improving workflows.
KEY BENEFITS
Multi-Region Quota Multiplier - Each AWS region has independent quota. 3 regions means 3x tokens per minute and 3x daily limits. One endpoint with automatic failover and routing. Configure EU-only, US-only, or global for compliance.
Production Infrastructure - Terraform/OpenTofu deployment. ECS Fargate auto-scaling, ALB with HTTPS, WAF, VPC, KMS encryption. AWS Well-Architected Framework.
Built-in Security - CloudWatch monitoring, OpenTelemetry tracing, API key auth via Systems Manager, CORS, SSRF protection.
Cost Efficiency - Pay-per-use for models via Bedrock. No monthly minimums. Prompt caching and Fargate Spot reduce costs further.
No Vendor Lock-in - Standard OpenAI & Anthropic APIs. Switch providers anytime by changing one config value.
Automatic Model Fallback - When AWS retires a model, requests are transparently redirected to its replacement. Your applications survive model deprecations without code changes.
GET STARTED
Start with a 14-day free trial on AWS Marketplace. Deploy in minutes with Terraform templates. Comprehensive docs and step-by-step guides for Open WebUI, n8n, Continue.dev, and LangChain at https://stdapi.ai . Production-ready from day one.
Highlights
- Works instantly with 1000+ OpenAI and Anthropic compatible applications - Open WebUI, n8n, Continue.dev, Cursor, LangChain, Obsidian, Slack bots, and autonomous agent frameworks like OpenClaw, LangGraph, and CrewAI. Full API coverage: chat, embeddings, images, speech, and transcription. Switch between 80+ models including Claude, Kimi K2, MiniMax, Nova, Llama, and Qwen3 instantly, with no changes to your existing setup.
- All data stays in your AWS account and never reaches model providers. Configure allowed regions for GDPR, HIPAA, and FedRAMP compliance. Production-ready infrastructure via Terraform: ECS Fargate with auto-scaling, ALB with HTTPS, WAF, VPC, KMS encryption, CloudWatch monitoring, and OpenTelemetry tracing. Follows the AWS Well-Architected Framework.
- Pay only for what you use through AWS Marketplace - no monthly minimums, no subscriptions, and no markup on AWS Bedrock model pricing. Reduce costs further with prompt caching and Fargate Spot. Standard OpenAI and Anthropic APIs mean zero vendor lock-in: switch AI providers anytime with no impact on your applications.
Details
Introducing multi-product solutions
You can now purchase comprehensive solutions tailored to use cases and industries.
Features and programs
Financing for AWS Marketplace purchases
Pricing
Free trial
Dimension | Description | Cost/unit/hour |
|---|---|---|
Hours | Container Hours | $0.10 |
Vendor refund policy
stdapi.ai offers refunds on a case-by-case basis. We encourage you to try our free tier first to evaluate the product before purchasing.
To request a refund, contact support@stdapi.ai with your AWS Marketplace order ID and reason. Our team will review your request promptly.
For more information, visit https://stdapi.ai/operations_getting_started/
How can we make this page better?
Legal
Vendor terms and conditions
Content disclaimer
Delivery details
Amazon ECS Deployment with Terraform/OpenTofu
- Amazon ECS
- Amazon EKS
Container image
Containers are lightweight, portable execution environments that wrap server application software in a filesystem that includes everything it needs to run. Container applications run on supported container runtimes and orchestration services, such as Amazon Elastic Container Service (Amazon ECS) or Amazon Elastic Kubernetes Service (Amazon EKS). Both eliminate the need for you to install and operate your own container orchestration software by managing and scheduling containers on a scalable cluster of virtual machines.
Version release notes
See https://stdapi.ai/roadmap/ for more information
Additional details
Usage instructions
OVERVIEW
Hardened container via AWS Marketplace. Deploy on ECS or EKS for managed orchestration.
OPTION 1 - AUTOMATED (RECOMMENDED)
Quick Start:
- Subscribe on AWS Marketplace
- Install Terraform/OpenTofu
- Choose example: https://stdapi.ai/operations_getting_started/
- Configure settings
- Deploy: terraform init && terraform apply
- Access via ALB endpoint
Module Provisions: ECS Fargate (auto-scaling), ALB (HTTPS), VPC, S3, CloudWatch, IAM, KMS, WAF (optional), Route53 (optional)
Customize Via Variables: Domain/HTTPS, auto-scaling, allowed regions, API auth, WAF/monitoring, existing VPC integration
OPTION 2 - MANUAL
Requirements: Marketplace subscription, ECS/EKS cluster, IAM roles, networking for Bedrock/S3, S3 bucket, environment variables
Container: Access from Marketplace Registry. Regular security updates.
PRODUCTION BEST PRACTICES
Use ALB with HTTPS, enable auto-scaling, CloudWatch alarms, VPC endpoints, WAF, API keys, multi-region Bedrock.
EXAMPLES
https://stdapi.ai/operations_getting_started/
- Production: Single region, full security
- GDPR: Multi-region EU
- US: Multi-region US
- Open WebUI: ChatGPT alternative
- Cost-Optimized: Dev/test
MONITORING
CloudWatch logs/metrics, ECS health checks, OpenTelemetry/X-Ray optional
SUPPORT
Docs: https://stdapi.ai Email: support@stdapi.ai GitHub: bug reports/features
SECURITY
Vulnerability scans, prompt patches, minimal surface, non-root. Region-specific deployment, HIPAA/GDPR/FedRAMP compatible, CloudWatch audit logs.
Resources
Vendor resources
Support
Vendor support
Email support available at support@stdapi.ai for deployment assistance, configuration questions, and troubleshooting. Response within 1-2 business days.
Community support via GitHub Issues at https://github.com/stdapi-ai/stdapi.ai/issues for bug reports, feature requests, and discussions.
Comprehensive documentation at https://stdapi.ai includes Getting Started guides, API reference, integration examples, and troubleshooting guides.
AWS infrastructure support
AWS Support is a one-on-one, fast-response support channel that is staffed 24x7x365 with experienced and technical support engineers. The service helps customers of all sizes and technical abilities to successfully utilize the products and features provided by Amazon Web Services.