AWS HPC Blog
Category: Artificial Intelligence
Meet the Advanced Computing team of AWS at SC25 in St. Louis
We want to empower every scientist and engineer to solve hard problems by giving them access to the compute and analytical tools they need, when they need them. Cloud HPC can be a real human progress catalyst. If you run large scale simulations, tune complex models, or support researchers who consistently need more compute, the […]
AWS re:Invent 2025: Your Complete Guide to High Performance Computing Sessions
AWS re:Invent 2025 returns to Las Vegas, Nevada on December 1, uniting AWS builders, customers, partners, and IT professionals from across the globe. This year’s event offers you exclusive access to compelling customer stories and insights from AWS leadership as they tackle today’s most critical challenges in high-performance computing, from accelerating scientific discovery to optimizing […]
Leveraging LLMs as an Augmentation to Traditional Hyperparameter Tuning
When seeking to improve machine learning model performance, hyperparameter tuning is often the go-to recommendation. However, this approach faces significant limitations, particularly for complex models requiring extensive training times. In this post, we’ll explore a novel approach that combines gradient norm analysis with Large Language Model (LLM) guidance to intelligently redesign neural network architectures. This […]
Enhanced Performance for Whisper Audio Transcription on AWS Batch and AWS Inferentia
In this post, we’ll review the key optimizations and performance gains for our Whisper audio transcription solution powered by AWS Batch and AWS Inferentia.
Engineering at the speed of thought: Accelerating complex processes with multi-agent AI and Synera
In this post, we’ll examine how this multi-agent approach works, the architecture behind it, and the efficiency improvements it enables. While the focus is on an engineering use case, the principles apply broadly to any organization facing the challenge of coordinating specialized expertise to deliver faster, more consistent results.
Scale Reinforcement Learning with AWS Batch Multi-Node Parallel Jobs
Autonomous robots are increasingly used across industries, from warehouses to space exploration. While developing these robots requires complex simulation and reinforcement learning (RL), setting up training environments can be challenging and time-consuming. AWS Batch multi-node parallel (MNP) infrastructure, combined with NVIDIA Isaac Lab, offers a solution by providing scalable, cost-effective robot training capabilities for sophisticated behaviors and complex tasks.
Enhancing Equity Strategy Backtesting with Synthetic Data: An Agent-Based Model Approach
Developing robust investment strategies requires thorough testing, but relying solely on historical data can introduce biases and limit your insights. Learn how synthetic data from agent-based models can provide an unbiased testbed to systematically evaluate your strategies and prepare for future market scenarios. Part 1 of 2 covers the theoretical foundations of the approach.
Scaling your LLM inference workloads: multi-node deployment with TensorRT-LLM and Triton on Amazon EKS
LLMs are scaling exponentially. Learn how advanced technologies like Triton, TRT-LLM and EKS enable seamless deployment of models like the 405B parameter Llama 3.1. Let’s go large.
Deploying Generative AI Applications with NVIDIA NIM Microservices on Amazon Elastic Kubernetes Service (Amazon EKS) – Part 2
Learn how to deploy AI models at scale with @AWS using NVIDIA’s NIM and Amazon EKS! This step-by-step guide shows you how to create a GPU cluster for inference in this second post of a two-part series!
Whisper audio transcription powered by AWS Batch and AWS Inferentia
Transcribe audio files at scale for really low cost using Whisper and AWS Batch with Inferentia. Check out this post to deploy a cost-effective solution in minutes!








