Amazon Web Services

Amazon SageMaker provides fully managed deployment features for optimal machine learning inference performance and cost at scale. This workshop explores how to use SageMaker inference capabilities to quickly deploy ML models in production for various use cases, including hyper-personalization, Generative AI, and Large Language Models (LLMs). Learn about different SageMaker inference endpoint options and how to deploy LLMs for inference efficiently.

product-information
skills-and-how-to
cost-optimization
ai-ml
serverless
Show 4 more

Up Next

VideoThumbnail
1:01:07

Accelerate ML Model Delivery: Implementing End-to-End MLOps Solutions with Amazon SageMaker

Nov 22, 2024
VideoThumbnail
9:30

Deploying ASP.NET Core 6 Applications on AWS Elastic Beanstalk Linux: A Step-by-Step Guide for .NET Developers

Nov 22, 2024
VideoThumbnail
47:39

Simplifying Application Authorization: Amazon Verified Permissions at AWS re:Invent 2023

Nov 22, 2024
VideoThumbnail
15:58

Revolutionizing Business Intelligence: Generative AI Features in Amazon QuickSight

Nov 22, 2024
VideoThumbnail
39:31

AWS re:Invent 2023: What's New in AWS Amplify for Full-Stack Web and Mobile App Development

Nov 22, 2024