IBM & Red Hat on AWS
AWS and Red Hat at Red Hat Summit 2026: Accelerating AI, Innovation, and Open-Source Infrastructure
From agentic AI-powered development environments to open-source Kubernetes innovation, here’s what Red Hat and AWS and Red Hat brought to Atlanta.
Red Hat Summit 2026 marks a defining moment in the nearly 20-year AWS and Red Hat strategic collaboration. This year, our two companies will showcase the breadth and depth of a partnership that spans agentic AI development, AI inference optimization, open-source infrastructure innovation, and expanded enterprise software availability on AWS Marketplace. For the engineers, architects, and technology leaders who have been building on Red Hat platforms and AWS infrastructure, this Summit brings announcements that will materially change how they develop, deploy, and operate workloads — and how much it costs to do so.
AWS was proud to serve as a Platinum sponsor at Red Hat Summit 2026. Their presence reflects not just the depth of the commercial relationship, but a genuine technical partnership: co-engineering new capabilities, contributing to open-source projects, and building integrations that give customers choices about how they want to run AI and cloud-native workloads. Whether you are a developer looking to accelerate your daily workflow, a platform engineer responsible for managing Kubernetes infrastructure at scale, or a technology executive evaluating AI platform strategies, there is something meaningful in these announcements for you.
Here is a closer look at what we announced at Red Hat Summit 2026:
Kiro and Red Hat OpenShift Dev Spaces: Bringing Agentic AI to Cloud-Based Development
The modern software development environment is a study in tradeoffs. On one side, developers want access to the most powerful AI coding tools available — tools that can reason about a codebase, generate tests, and autonomously complete features from a high-level specification. On the other side, enterprise organizations need development environments that are consistent, secure, reproducible, and governed — with no configuration drift between team members, no sensitive credentials stored on individual laptops, and no debugging sessions lost to “works on my machine” failures. For years, teams have had to choose between these two priorities. That choice is no longer necessary.
Red Hat and AWS have integrated Kiro, the agentic AI-powered integrated development environment, with Red Hat OpenShift Dev Spaces. OpenShift Dev Spaces provides developers with consistent, secure, containerized workspaces that run directly on an OpenShift cluster.
Into this enterprise-grade environment, Kiro brings a new generation of AI-powered development capability. Unlike traditional code completion tools, Kiro uses a spec-driven development methodology: it converts high-level requirements into structured specifications before generating code, with autonomous AI agents that break features into sequenced tasks complete with unit tests, integration checks, and documentation. The integration uses a local-to-remote SSH workflow — developers run Kiro on their local desktop while connecting over SSH to the containerized workspace running in the OpenShift cluster. Kiro delivers AI-powered coding assistance inside a workspace already tailored to the project, leveraging the cluster’s compute and storage resources rather than the developer’s local machine.
For development teams, the practical impact is significant. New developers can onboard in minutes rather than days — navigating to Dev Spaces, selecting a repository, launching a workspace, and writing code with full AI assistance in an environment identical to every other team member’s. For platform engineering and security teams, centralized control over development environments is preserved: security policies, compliance requirements, resource limits, and network access remain under IT governance even as developers gain access to cutting-edge AI tooling. For organizations as a whole, the integration accelerates time-to-market by removing friction from the development process and redirecting developer attention from environment management to feature development.
OpenShift Dev Spaces 3.25 is available now, and the Kiro integration is available for teams running this release. Read more at: https://aws.amazon.com/blogs/ibm-redhat/cloud-development-meets-agentic-ai-kiro-and-red-hat-openshift-dev-spaces/
AI Platform and Inference
Moving AI from proof-of-concept to production is one of the defining challenges of this era. The collaboration between AWS and Red Hat extends well beyond developer tools into the core AI platform and inference stack — reducing procurement friction, optimizing inference costs, and enabling enterprises to run AI workloads more efficiently on purpose-built AWS silicon. The announcements in this theme reflect a comprehensive approach to the full AI deployment lifecycle.
AI Operator for AWS Inferentia: Enterprise AI Inference on Purpose-Built Silicon
Running AI inference workloads at production scale demands specialized hardware. General-purpose compute can serve many workloads, but inference tasks — particularly for large language models and deep learning applications — benefit substantially from purpose-built accelerators designed to optimize the specific mathematical operations these models perform. AWS Inferentia chips are engineered precisely for this: delivering high throughput, low latency inference at a cost point that makes production-scale AI economically viable.
The challenge, historically, has been integration. Enterprises running Red Hat OpenShift as their Kubernetes platform have invested significantly in building operational practices around OpenShift — monitoring, policy enforcement, RBAC, lifecycle management, and governance workflows. Incorporating specialized AI accelerators into that environment has traditionally required custom engineering: driver management, device plugin configuration, node labeling, and workload placement logic that sits outside standard OpenShift tooling.
The AI Operator for AWS Inferentia resolves this integration complexity. Red Hat and AWS are collaborating to bring Inferentia support directly into the OpenShift operator model — the same pattern OpenShift uses to manage databases, messaging systems, and other infrastructure components. The operator abstracts the configuration complexity of deploying and managing AI models on Inferentia instances, exposing Inferentia’s acceleration capabilities through the standard OpenShift interfaces that platform teams already manage.
For customers running production AI workloads, this means access to Inferentia’s purpose-built inference performance — delivering higher throughput and lower cost per inference compared to general-purpose GPU instances — without departing from the unified operational model they have established on OpenShift. AI workloads run alongside traditional containerized applications, managed through the same tools, policies, and workflows. For organizations at the stage of scaling AI from pilot to production, this unification of the operational model is as valuable as the hardware performance improvement itself.
Read more: https://aws.amazon.com/blogs/ibm-redhat/running-red-hat-ai-on-openshift-with-aws-neuron/
Red Hat AI Inference Server and AWS Trainium: Optimizing Inference at Scale
As organizations move deeper into production AI deployment, inference cost becomes the dominant variable in total cost of ownership. Training large models is expensive, but it happens once. Inference — serving those models to users, applications, and automated pipelines — happens millions or billions of times, and the cost accumulates accordingly. For enterprises running generative AI applications, recommendation systems, real-time content analysis, or autonomous agent workflows, inference infrastructure optimization is not an abstract concern. It is a line item that grows with adoption.
AWS Trainium chips were designed with this reality in mind. Built on custom silicon optimized for deep learning workloads, Trainium delivers compelling price-performance for inference at scale — particularly for transformer-based models that characterize the current generation of large language models and embedding systems. Red Hat AI Inference Server, in turn, provides the enterprise-grade model serving layer that organizations need: model lifecycle management, horizontal scaling, observability, and the operational consistency that enterprise support requires.
Red Hat and AWS are working together to bring Red Hat AI Inference Server support for AWS Trainium chips, creating a combination that addresses the inference cost problem with an enterprise operational model. When complete, customers will be able to deploy production AI inference workloads on Trainium’s purpose-built architecture while maintaining the Red Hat support, compliance posture, and operational consistency they rely on across their platform. For organizations running large language models, real-time AI applications, or high-volume inference pipelines, this collaboration represents a path to meaningfully reducing inference costs while preserving the enterprise operational framework they have built.
Red Hat AI Enterprise on AWS Marketplace: Simplifying Enterprise AI Procurement
For enterprises evaluating enterprise AI platforms, procurement complexity is a surprisingly significant barrier to adoption. Organizations that have invested in building a Red Hat OpenShift environment — and in establishing the operational practices, security controls, and compliance workflows that go with it — want an AI platform that integrates naturally into that environment. Organizations that have standardized on Red Hat Enterprise Linux for their most critical workloads know that RHEL’s security hardening, FIPS certification, Common Criteria evaluation, and long-term support lifecycle are not simply checkboxes — they are requirements that enable deployment in regulated industries and highly controlled environments.
What they do not want is a separate procurement process, a new vendor onboarding cycle, or an AI licensing commitment that sits outside their existing cloud financial management framework.
Red Hat AI Enterprise is now available on AWS Marketplace, addressing this friction directly. AWS customers can use their existing cloud spend — including committed spend agreements established through the AWS Enterprise Discount Program, private pricing agreements, and AWS Marketplace Private Offers — toward licensing Red Hat AI Enterprise.
For technology leaders, this availability changes the evaluation calculus. Organizations that have consolidated their spending commitments with AWS can now deploy Red Hat AI Enterprise without requiring a separate budget cycle or a standalone vendor negotiation. For procurement and finance teams, it simplifies reporting and chargeback. For platform engineers, it means Red Hat AI Enterprise can be provisioned and managed alongside existing AWS infrastructure with the familiar tools, dashboards, and billing workflows already in place. The path from AI experimentation to production deployment gets meaningfully shorter.
Cut costs, scale smarter with Red Hat OpenShift Service on AWS: Karpenter automates compute provisioning
Red Hat is improving compute capacity and node life cycle management for Red Hat OpenShift Service on AWS (ROSA) through Red Hat build of Karpenter, which is a fully-managed autoscaler based on the upstream Project Karpenter. This delivers significant infrastructure cost savings through right-sizing, spot optimization, and continuous consolidation, eliminating the operational burden of manual node scaling. Red Hat build of Karpenter will be generally available in the next minor release of Red Hat OpenShift in all AWS Regions where ROSA is generally available.
To learn more, read this AWS blog.
Looking Ahead
These announcements reflect the accelerating momentum of the AWS and Red Hat strategic collaboration. — and they represent a coherent arc, not a collection of independent projects. Taken together, these innovations address the full lifecycle of enterprise AI deployment: from the developer writing the first line of code to the platform engineer managing the infrastructure that serves it to the technology leader responsible for the cost and compliance posture of the entire stack. That comprehensive view reflects how AWS and Red Hat think about partnership — not as a collection of integrations, but as a shared commitment to making enterprise technology simpler, more capable, and more economically efficient.
Join us on May 28 for a hands-on experience for you to work directly with our Red Hat and Amazon Web Service (AWS) experts to try Red Hat® OpenShift® Service on AWS: https://www.redhat.com/en/events/webinar/redhat-openshift-service-on-aws-rosa?sc_cid=RHCTN1260000477281
To learn more about the AWS and Red Hat collaboration, visit www.redhat.com/aws.