Amazon Web Services

In this informative session, Raghu Ramesha, a senior machine learning architect at Amazon, delves into the powerful capabilities of Amazon SageMaker for high-performance, cost-effective machine learning inference. He explains why customers choose SageMaker over DIY solutions, highlighting its managed infrastructure, automatic scaling, and deployment strategies. Raghu explores various deployment options including real-time, serverless, asynchronous, and batch inference, each tailored for different use cases. The presentation also covers cost optimization techniques such as multi-model endpoints, instance right-sizing, and SageMaker Savings Plans. Viewers will gain insights into best practices for maximizing performance while minimizing costs in machine learning deployments using Amazon SageMaker.

product-information
skills-and-how-to
cost-optimization
ai-ml
cost-mgmt
Show 4 more

Up Next

VideoThumbnail
1:01:07

Accelerate ML Model Delivery: Implementing End-to-End MLOps Solutions with Amazon SageMaker

Nov 22, 2024
VideoThumbnail
9:30

Deploying ASP.NET Core 6 Applications on AWS Elastic Beanstalk Linux: A Step-by-Step Guide for .NET Developers

Nov 22, 2024
VideoThumbnail
47:39

Simplifying Application Authorization: Amazon Verified Permissions at AWS re:Invent 2023

Nov 22, 2024
VideoThumbnail
15:58

Revolutionizing Business Intelligence: Generative AI Features in Amazon QuickSight

Nov 22, 2024
VideoThumbnail
2:53:33

Streamlining Patch Management: AWS Systems Manager's Comprehensive Solution for Multi-Account and Multi-Region Patching Operations

Nov 22, 2024