Amazon Web Services

Amazon SageMaker provides fully managed deployment features for optimal machine learning inference performance and cost at scale. This workshop explores how to use SageMaker inference capabilities to quickly deploy ML models in production for various use cases, including hyper-personalization, Generative AI, and Large Language Models (LLMs). Learn about different SageMaker inference endpoint options and how to deploy LLMs for inference efficiently.

product-information
skills-and-how-to
cost-optimization
ai-ml
serverless
Show 4 more

Up Next

VideoThumbnail
5:35

AWS WAF - Web Application Firewall protect your web applications from common web exploits

Jun 26, 2025
VideoThumbnail
16:03

Tọa đàm với anh Hiếu Trần - Co-founder của NAB Studio

Jun 26, 2025
VideoThumbnail
18:40

Thiết kế hạ tầng mạng chung trong môi trường sử dụng nhiều AWS account (Level 200)

Jun 26, 2025
VideoThumbnail
7:59

Triển khai và vận hành ứng dụng container trên môi trường nhiều AWS account (Level 300)

Jun 26, 2025
VideoThumbnail
7:06

Sử dụng Amazon S3 như thế nào? (Level 100)

Jun 26, 2025