Amazon EC2 P4 Instances

Highest performance for ML training and HPC applications in the cloud

Amazon EC2 P4d instances deliver the highest performance for machine learning (ML) training and high performance computing (HPC) applications in the cloud. P4d instances are powered by the latest NVIDIA A100 Tensor Core GPUs and deliver industry-leading high throughput and low latency networking. These instances are the first in the cloud to support 400 Gbps instance networking. P4d instances provide up to 60% lower cost to train ML models, including an average of 2.5x better performance for deep learning models compared to previous generation P3 and P3dn instances.

Amazon EC2 P4d instances are deployed in hyperscale clusters called EC2 UltraClusters that are comprised of the highest performance compute, networking, and storage in the cloud. Each EC2 UltraCluster is one of the most powerful supercomputers in the world, enabling customers to run their most complex multi-node ML training and distributed HPC workloads. Customers can easily scale from a few to thousands of NVIDIA A100 GPUs in the EC2 UltraClusters based on their ML or HPC project needs.

Researchers, data scientists, and developers can leverage P4d instances to train ML models for use cases such as natural language processing, object detection and classification, and recommendation engines, as well as run HPC applications such as pharmaceutical discovery, seismic analysis, and financial modeling. Unlike on-premises systems, customers can access virtually unlimited compute and storage capacity, scale their infrastructure based on business needs, and spin up a multi-node ML training job or a tightly coupled distributed HPC application in minutes, without any setup or maintenance costs.

Announcing the New Amazon EC2 P4d Instances (2:00)

Got questions or need help with EC2 UltraClusters?

Request help »

High Scale ML Training and HPC with EC2 P4d UltraClusters

EC2 UltraClusters of P4d instances combine high performance compute, networking, and storage into one of the most powerful supercomputers in the world. Each EC2 UltraCluster of P4d instances comprises more than 4,000 of the latest NVIDIA A100 GPUs, Petabit-scale non-blocking networking infrastructure, and high throughput low latency storage with FSx for Lustre. Any ML developer, researcher, or data scientist can spin up P4d instances in EC2 UltraClusters to get access to supercomputer-class performance with pay-as-you-go usage model to run their most complex multi-node ML training and HPC workloads.

For questions or assistance with EC2 UltraClusters, request help.

Available in some regions only. For information on which regions are supported and any additional information on EC2 UltraClusters contact us.


Reduce ML training time from days to minutes

With the latest generation NVIDIA A100 Tensor Core GPUs, each Amazon EC2 P4d instance delivers on average 2.5x better deep learning performance compared to previous generation P3 instances. EC2 UltraClusters of P4d instances enable everyday developers, data scientists, and researchers to run their most complex ML and HPC workloads by giving access to supercomputing-class performance without any upfront costs or long-term commitments. The reduced training time with P4d instances boosts productivity, enabling developers to focus on their core mission of building ML intelligence into business applications.

Run the most complex multi-node ML training with high efficiency

Developers can seamlessly scale to up to thousands of GPUs with EC2 UltraClusters of P4d instances. High throughput, low latency networking with support for 400 Gbps instance networking, Elastic Fabric Adapter (EFA), and GPUDirect RDMA technology, help rapidly train ML models using scale-out/distributed techniques. Elastic Fabric Adapter (EFA) uses the NVIDIA Collective Communications Library (NCCL) to scale to thousands of GPUs, and GPUDirect RDMA technology enables low latency GPU to GPU communication between P4d instances.

Lower the infrastructure costs for ML training and HPC

Amazon EC2 P4d instances deliver up to 60% lower cost to train ML models compared to P3 instances. Additionally, P4d instances are available for purchase as Spot Instances. Spot Instances take advantage of unused EC2 instance capacity and can lower your Amazon EC2 costs significantly with up to a 90% discount from On-Demand prices. With the lower cost of ML training with P4d instances, budgets can be reallocated to build more ML intelligence into business applications.

Easily get started and scale with AWS services

Deep learning AMIs and deep learning containers make it easy to deploy P4d deep learning environments in minutes as they contain the required deep learning framework libraries and tools. You can also easily add your own libraries and tools to these images. P4d instances support popular ML frameworks such as TensorFlow, PyTorch, and MXNet. Additionally, Amazon EC2 P4d instances are supported by major AWS services for ML, management, and orchestration such as Amazon SageMaker, Amazon Elastic Kubernetes Service (EKS), Amazon Elastic Container Service (ECS), AWS Batch, and AWS ParallelCluster.


Powered by NVIDIA A100 Tensor Core GPUs

NVIDIA A100 Tensor Core GPUs deliver unprecedented acceleration at scale for ML and high performance computing (HPC). NVIDIA A100’s third generation Tensor Cores accelerate every precision workload, speeding time to insight and time to market. Each A100 GPU offers over 2.5x the compute performance compared to the previous generation V100 GPU and comes with 40 GB HBM2 (in P4d instances) or 80 GB HBM2e (in P4de instances) of high-performance GPU memory. Higher GPU memory particularly benefits those workloads training on large datasets of high-resolution data. NVIDIA A100 GPUs leverage NVSwitch GPU interconnect throughput so each GPU can communicate with every other GPU in the same instance at the same 600GB/s bidirectional throughput and with single-hop latency.

High performance networking

P4d instances provide 400 Gbps networking to help customers better scale out their distributed workloads such as multi-node training more efficiently with high throughput networking between P4d instances as well as between a P4d instance and storage services such as Amazon S3 and FSx for Lustre. Elastic Fabric Adapter (EFA) is a custom network interface designed by AWS to help scale ML and HPC applications to thousands of GPUs. To further reduce latency, EFA is coupled with NVIDIA GPUDirect RDMA to enable low latency GPU to GPU communication between servers with OS bypass.

High throughput, low latency storage

Customers can access PetaByte-scale high throughput, low latency storage with FSx for Lustre or virtually unlimited cost-effective storage with Amazon S3 at 400 Gbps speeds. For workloads that need fast access to large datasets, each P4d instance also includes 8TB NVMe-based SSD storage with 16 GigaBytes/sec read throughput.

Built on AWS Nitro System

The P4d instances are built on the AWS Nitro System, which is a rich collection of building blocks that offloads many of the traditional virtualization functions to dedicated hardware and software to deliver high performance, high availability, and high security while also reducing virtualization overhead.

Customer stories

Toyota Research Institute (TRI), founded in 2015, is working to develop automated driving, robotics, and other human amplification technology for Toyota.

“At TRI, we're working to build a future where everyone has the freedom to move,” said Mike Garrison, Technical Lead, Infrastructure Engineering at TRI. "The previous generation P3 instances helped us reduce our time to train machine learning models from days to hours and we are looking forward to utilizing P4d instances, as the additional GPU memory and more efficient float formats will allow our machine learning team to train with more complex models at an even faster speed."

"At TRI-AD, we're working to build a future where everyone has the freedom to move and explore with a focus on reducing vehicle injuries and fatalities using adaptive driving and smart city. Through the use of Amazon EC2 P4d instances, we were able to reduce our training time for object recognition by 40% compared to previous generation GPU instances without any modification to existing codes." said Junya Inada, Director of Automated Driving (Recognition) at TRI-AD.

Jack Yan, Senior Director of Infrastructure Engineering at TRI-AD said, "Through the use of Amazon EC2 P4d instances, we were able to instantly reduce our cost to train compared to previous generation GPU instances enabling us to increase the number of teams working on model training. The networking improvements in P4d allowed us to efficiently scale to dozens of instances which gave us significant agility to rapidly optimize, retrain, and deploy models in test cars or simulation environments for further testing."

GE Healthcare is a leading global medical technology and digital solutions innovator. GE Healthcare enables clinicians to make faster, more informed decisions through intelligent devices, data analytics, applications and services, supported by its Edison intelligence platform.

“At GE Healthcare, we provide clinicians with tools that help them aggregate data, apply AI and analytics to that data and uncover insights that improve patient outcomes, drive efficiency and eliminate errors,” said Karley Yoder, VP & GM, Artificial Intelligence. “Our medical imaging devices generate massive amounts of data that need to be processed by our data scientists. With previous GPU clusters, it would take days to train complex AI models, such as Progressive GANs, for simulations and view the results. Using the new P4d instances reduced processing time from days to hours. We saw two- to three-times greater speed on training models with various image sizes, while achieving better performance with increased batch size and higher productivity with a faster model development cycle.”

HEAVY.AI is a pioneer in accelerated analytics. The HEAVY.AI platform is used in business and government to find insights in data beyond the limits of mainstream analytics tools.

“At HEAVY.AI, we’re working to build a future where Data Science and Analytics converge to break down and fuse data silos. Customers are leveraging their massive amounts of data that may include location and time to build a full picture of not only what is happening, but when and where through granular visualization of spatial temporal data. Our technology enables seeing both the forest and the trees.” said (Ray Falcione), VP of US Public Sector at HEAVY.AI. “Through the use of Amazon EC2 P4d instances, we were able to reduce the cost to deploy our platform significantly compared to previous generation GPU instances thus enabling us to cost-effectively scale massive data sets. The networking improvements on A100 has increased our efficiencies in how we scale to billions of rows of data and enabled our customers to glean insights even faster.”

Zenotech Ltd is redefining engineering online through the use of HPC Clouds delivering on demand licensing models together with extreme performance benefits by leveraging GPUs.

“At Zenotech we are developing the tools to enable designers to create more efficient and environmentally friendly products. We work across industries and our tools provide greater product performance insight through the use of large scale simulation.” said Jamil Appa, Director of Zenotech. “The use of AWS P4d instances, enables us to run our simulations 3.5 times faster compared to the previous generation of GPUs. This speed up cuts our time to solve significantly allowing our customers to get designs to market faster or to do higher fidelity simulations than were previously possible.”

Aon is a leading global professional services firm providing a broad range of risk, retirement and health solutions. Aon PathWise is a GPU-based and scalable HPC risk management solution that insurers and re-insurers, banks, and pension funds can use to address today’s key challenges such as hedge strategy testing, regulatory and economic forecasting, and budgeting. 

“At PathWise Solutions Group LLC, our product allows insurance companies, reinsurers and pension funds to access next generation technology to rapidly solve today’s key insurance challenges such as machine learning, hedge strategy testing, regulatory and financial reporting, business planning and economic forecasting, and new product development and pricing” said Peter Phillips, President and CEO, PathWise Solutions Group. "Through the use of Amazon EC2 P4d instances we are able to deliver amazing improvements in speed for single and double precision calculations over previous generation GPU instances for the most demanding calculations, allowing new range of calculations and forecasting to be done by clients for the very first time. Speed matters,” says Phillips, “and we continue to deliver meaningful value and the latest technology to our customers thanks to the new instances from AWS.”

Comprised of radiology and AI experts, Rad AI builds products that maximize radiologist productivity, ultimately making healthcare more widely accessible and improving patient outcomes.

“At Rad AI, our mission is to increase access to and quality of healthcare, for everyone. With a focus on medical imaging workflow, Rad AI saves radiologists time, reduces burnout, and enhances accuracy,” said Doktor Gurson, Co-founder of Rad AI. “We use AI to automate radiology workflows and help streamline radiology reporting. With the new EC2 P4d instances, we’ve seen faster inference and the ability to train models 2.4x faster, with higher accuracy than on previous generation P3 instances. This allows faster, more accurate diagnosis, and greater access to high quality radiology services provided by our customers across the US.“

Read case study to learn more »

Product Details

Instance Size vCPUs Instance Memory (GiB) GPU – A100 GPU memory Network Bandwidth (Gbps) GPUDirect RDMA GPU Peer to Peer Instance Storage (GB) EBS Bandwidth (Gbps) On-demand Price/hr 1-yr Reserved Instance Effective Hourly * 3-yr Reserved Instance Effective Hourly *
p4d.24xlarge 96 1152 8 320 GB
400 ENA and EFA Yes 600 GB/s NVSwitch 8 x 1000 NVMe SSD 19 $32.77 $19.22 $11.57
p4de.24xlarge (preview) 96 1152 8 640 GB
400 ENA and EFA Yes 600 GB/s NVSwitch 8 x 1000 NVMe SSD 19 $40.96 $24.01 $14.46
* - Prices shown are for Linux/Unix in the US East (Northern Virginia) AWS Region and rounded to the nearest cent. For full pricing details, see the Amazon EC2 pricing page.

Amazon EC2 P4d instances are available in the US East (N. Virginia and Ohio), US West (Oregon), Europe (Ireland and Frankfurt) and Asia Pacific (Tokyo and Seoul) regions and Amazon EC2 P4de instances are available in the AWS US East (N. Virginia) and US West (Oregon) regions.

Customers can purchase P4d and P4de instances as On-Demand Instances, Reserved Instances, Spot Instances, Dedicated Hosts or as part of Savings Plan.

Getting Started with Amazon EC2 P4d instances for machine learning

Using Amazon SageMaker

Amazon SageMaker is a fully-managed service for building, training, and deploying ML models. When used together with Amazon EC2 P4d instances, customers can easily scale to tens, hundreds, or thousands of GPUs to train a model quickly at any scale without worrying about setting up clusters and data pipelines.

Using AWS Deep Learning AMI or Deep Learning Containers

The AWS Deep Learning AMIs (DLAMI) provide ML practitioners and researchers with the infrastructure and tools to accelerate deep learning in the cloud, at any scale. AWS Deep Learning Containers are Docker images pre-installed with deep learning frameworks to make it easy to deploy custom ML environments quickly by letting you skip the complicated process of building and optimizing your environments from scratch.

Using Amazon Elastic Kubernetes Service (EKS) or Elastic Container Service (ECS)

Customers who prefer to manage their own containerized workloads via container orchestration services can deploy Amazon EC2 P4d instances with Amazon EKS or ECS.

Getting started with Amazon EC2 P4d instances for high performance computing

Amazon EC2 P4d instances are an ideal platform to run engineering simulations, computational finance, seismic analysis, molecular modeling, genomics, rendering, and other GPU-based high performance computing (HPC) workloads. HPC applications often require high network performance, fast storage, large amounts of memory, high compute capabilities, or all of the above. P4d instances support Elastic Fabric Adapter (EFA) that enables HPC applications using the Message Passing Interface (MPI) to scale to thousands of GPUs. AWS Batch and AWS ParallelCluster enable HPC developers to quickly build and scale distributed HPC applications.

Learn more »

Blog posts & articles

Additional resources

EC2 UltraClusters of P4d instances (1:59)
Amazon EC2 P4d vs P3: Natural Language Processing (0:34)
Amazon EC2 P4d vs P3: Speech to text (0:33)
Amazon EC2 P4d vs P3: Image Classification (0:37)

Get started with AWS

Sign up for an AWS account

Sign up for an AWS account

Instantly get access to the AWS Free Tier.

Learn with simple tutorials

Learn with 10-minute tutorials

Explore and learn with simple tutorials.

Start building with EC2 in the console

Start building in the console

Begin building with step-by-step guides to help you launch your AWS project.