AWS HPC Blog
Category: Thought Leadership
How agent-based models powered by HPC are enabling large scale economic simulations
See how agent-based models, driven to scale by HPC in the cloud, are shedding new light on macroprudential policies with this post from Oxford’s Institute for New Economic Thinking.
How Amazon’s Search M5 team optimizes compute resources and cost with fair-share scheduling on AWS Batch
In this post, we share how Amazon Search optimizes their use of accelerated compute resources using AWS Batch fair-share scheduling to schedule distributed deep learning workloads.
Massively-scaling quantum chemistry to support a circular economy
As a part of AWS’s “Digital Technologies for a Circular Economy” initiative, we joined forces with Accenture, Intel and Good Chemistry to massively scale quantum chemistry simulations. This is the first and most complex step to discovering new pathways for PFAS destruction for a cleaner world.
Helping bioinformaticians transition to running workloads on AWS
Calling budding bioinformaticians! If you learn through hands-on practicals and walkthroughs, AWS and GIS have developed training and resources to help you increase the scale and productivity of your research using the AWS cloud.
Bridging research and HPC to tackle grand challenges
Today we announced the AWS Impact Computing Project at the Harvard Data Science Initiative (HDSI) to identify potential solutions that can improve the lives of humans, other species, and natural ecosystems. Deb Goldfarb describes its goals and our joint vision.
How AWS Batch developed support for Amazon Elastic Kubernetes Service
Today, we discuss AWS batch on Amazon EKS, and the initial motivation and design choices the team made when we developed the service, and some of the challenges to overcome.
Rearchitecting AWS Batch managed services to leverage AWS Fargate
AWS service teams continuously improve the underlying infrastructure and operations of managed services, and AWS Batch is no exception. The AWS Batch team recently moved most of their job scheduler fleet to a serverless infrastructure model leveraging AWS Fargate. I had a chance to sit with Devendra Chavan, Senior Software Development Engineer on the AWS Batch team, to discuss the move to AWS Fargate and its impact on the Batch managed scheduler service component.
Accelerating Genomics Pipelines Using Intel’s Open Omics Acceleration Framework on AWS
In this blog, we showcase the first version of Open Omics and benchmark three applications that are used in processing NGS data – sequence alignment tools BWA-MEM, minimap2, and single cell ATAC-Seq on Xeon-based Amazon Elastic Compute Cloud (Amazon EC2) Instances.
Choosing between AWS Batch or AWS ParallelCluster for your HPC Workloads
It’s an understatement that AWS has a lot of services (more than 200 at the time of this post!). We’re usually the first to point out that there’s more than one way to solve a problem. HPC is no different in this regard, because we offer a choice: customers can run their HPC workloads using AWS […]
The Convergent Evolution of Grid Computing in Financial Services
The Financial Services industry makes significant use of high performance computing (HPC) but it tends to be in the form of loosely coupled, embarrassingly parallel workloads to support risk modelling. The infrastructure tends to scale out to meet ever increasing demand as the analyses look at more and finer grained data. At AWS we’ve helped many customers tackle scaling challenges are noticing some common themes. In this post we describe how HPC teams are thinking about how they deliver compute capacity today, and highlight how we see the solutions converging for the future.