In this post we’ll show you application performance and scaling results from Hpc7g, a new instance powered by AWS Graviton3E across a wide range of HPC workloads and disciplines.
The transition to a sustainable economy is a major goal of many organizations today. HPC plays a pivotal role and in this post we’ll explore how HPC on AWS is enabling the shift towards a sustainable future.
In this post we explore SeatGeek’s load testing system that simulates 50k simultaneous users. Originally built to prep SeatGeek for large-event traffic spikes, it now runs weekly to help them harden their code.
With AWS ParallelCluster 3.6, you can directly specify Slurm settings in the cluster config file – improving reproducibility and another step towards self-documentation for your HPC infrastructure.
In this post, we discuss how Novo Nordisk approached the deployment of a scale-out HPC platform for running AlphaFold, while meeting their enterprise IT requirements and keeping the user experience simple.
In this post, we describe Agent-Based Computational Economics (ACE), and how extreme scale computing makes it beneficial for policy design.
AWS ParallelCluster 3.6.0 can now detect GPU failures in HPC and AI/ML tasks. Health checks run at the start of Slurm jobs and if they fail, the job is requeued on another instance. This can increase reliability and prevent wasted spend.
AWS is developing new tools to enable easier and faster deployment of level 3 digital twin virtual sensors. in this post we show why L3 digital twins are needed for virtual sensors and how to elastically deploy one the cloud at scale.
In this blog post, we catch a glimpse into drug discovery to see how Evolvere Biosciences has deployed a customized architecture w/ AWS Batch and Nextflow to quickly and easily run its macromolecule design pipeline.
In this post, we discuss the “tipping point” problem, using HPC at a large scale to simulate the impact of deforestation to the risk of accelerating damage to the Amazon rainforest.