Amazon Web Services
Amazon SageMaker provides fully managed deployment features for optimal machine learning inference performance and cost at scale. This workshop explores how to use SageMaker inference capabilities to quickly deploy ML models in production for various use cases, including hyper-personalization, Generative AI, and Large Language Models (LLMs). Learn about different SageMaker inference endpoint options and how to deploy LLMs for inference efficiently.