AWS Open Source Blog
Deploy Large Language Models Easily with the New ezsmdeploy Python SDK
AWS created the ezsmdeploy open source Python package to help easily deploy machine learning models and provide a variety of options such as passing one or more model files, automatic selection of instances, and autoscaling.
This blog post will cover the new features of the ezsmdeploy 2.0 SDK. We’ll also provide some code examples to demonstrate how to launch and interact with popular foundation models (FM). The new ezsmdeploy Python SDK from AWS makes it much simpler to deploy large language models (LLMs) from Hugging Face Hub and SageMaker Jumpstart as production-ready APIs on Amazon SageMaker.
Version 2.0 of the SDK has several new capabilities to further simplify deploying foundation models. With just a few lines of code, users can now deploy popular large language models such as Llama 2, Falcon, and Stable Diffusion. The SDK automatically selects instance types, configures autoscaling, and handles other details required to launch production-ready deployments.
Overall, the updates in ezsmdeploy 2.0 make it easier for developers to take large language models from Hugging Face Hub and SageMaker Jumpstart and launch them as production services. By handling instance selection, autoscaling, and other deployment details automatically, ezsmdeploy reduces the code needed to put state-of-the-art models into production from hundreds of lines to just a few.
New features of ezsmdeploy 2.0
Ezsmdeploy is the Python SDK that lets you go from local model files or S3 tarballs to fully-managed, autoscaling SageMaker endpoints. It handles everything from Docker builds to load testing so you can focus on your machine learning. You can test out multi-model endpoints by simply passing in a list of models, or low-cost serverless inference by passing in serverless=True
. You can even capture model data and monitor drift automatically with a single parameter, all without becoming a DevOps expert or changing how you do model prediction.
Version 2.0 of Ezsmdeploy adds new features related to support for SageMaker Jumpstart foundational models. SageMaker Jumpstart provides pre-built foundational models that can be used to bootstrap machine learning projects. This release allows users to deploy those pre-built models with one line of code.
The project has also added support for Hugging Face Hub models. Hugging Face Hub is a repository of machine learning models. Users can now utilize models easily from the Hugging Face Hub for SageMaker.
Additionally, OpenChatKit support has been included for conversational AI models. OpenChatKit is a toolkit for building chat models. Users can now leverage compatible OpenChatKit models for conversational applications. Testing has been performed on the variety of models and configurations.
Code examples
The following code can be executed in Amazon SageMaker notebooks. For more information on the machine learning development environments that SageMaker offers, please visit here.
To install the ezsmdeploy package on a SageMaker notebook instance, or any notebook environment with permissions to access AWS resources, run the following commands:
You can now deploy state-of-the-art models like GPT-3, Falcon, and Bloom directly from Hugging Face or Jumpstart to SageMaker, without having to build custom containers or write complex deployment code.
For example, to deploy the 40B parameter Falcon instruct model from Hugging Face, here is the Python code:
That’s it! ezsmdeploy will use an appropriate service provided Docker container (or build one), deploy the model on Amazon SageMaker, and provide an endpoint to query the model. This is considerably simpler than the standard way of deploying the same model (in this case, the Falcon 40B Instruct).
Here are some more examples that demonstrate deploying large models using ezsmdeploy.
1. Jumpstart LLM deployment
After importing ezsmdeploy, we can now call the Deploy
method passing in the model name, foundation model flag, and instance type. This will deploy the model to Amazon SageMaker’s hosted endpoints.
To test the deployed model, a payload is passed to the model with hyperparameters. Here, the query is about “steps to make a pizza.”
These are the steps to make a pizza
Make dough and let rise
Preheat oven to 500°F
Roll out dough thin
Top with sauce, cheese and toppings of choice
Bake on a hot stone or baking sheet for 10-15 minutes until crust is crisp and cheese is melted and brown
Allow to cool slightly
Slice into wedges
Serve warm and enjoy!
It is considered a best practice to not leave resources running on AWS. To delete the endpoint, we just call the delete_endpoint
API.
2. Hugging Face model deployment from hub
In the following tasks, we show how to deploy the state-of-the-art pre-trained Falcon 7B models from Hugging Face for Text2Text Generation. You can directly use a Falcon 7B model for many Natural Language Processing (NLP) tasks, without fine-tuning the model. To deploy the Falcon 7B model from Hugging Face, first import ezsmdeploy and then call the .Deploy
method with the model name, huggingface_model flag, foundation_model flag and instance type as follows:
import ezsmdeploy
The input is sent to the model using the predictor
method:
In this example, the input is “Paris is the capital of” and the Falcon model completes the sentence and responses the full text as “Paris is the capital of France.”
3. Huggingface model deployment on serverless
In this example, we show how to deploy a FM on a serverless instance easily by just enabling the serverless flag.
4. Chat interface with LLM chat models
Chat interfaces powered by large language models like RedPajama are becoming increasingly sophisticated. With the eszmdeploy library, developers can easily integrate LLM chat models into their applications, creating engaging and interactive experiences for users. These models can be deployed easily with just one line of code as shown in the code sample here:
5. Image generation with Stable Diffusion 2-1
Stable Diffusion is a text-to-image model that enables you to create photorealistic images from just a text prompt. A diffusion model trains by learning to remove noise that was added to a real image. This de-noising process generates a realistic image. These models can also generate images from text alone by conditioning the generation process on the text. For instance, Stable Diffusion is a latent diffusion where the model learns to recognize shapes in a pure noise image and gradually brings these shapes into focus if the shapes match the words in the input text.
In this example, you will learn how to use ezsmdeploy to deploy a Stable Diffusion model (“model-txt2img-stabilityai-stable-diffusion-v2-1-base
”) from JumpStart. Then you will use the predictor to generate artistic images and add prompts.
Now that the model is deployed, the prompt and image weight and height are determined to send as a payload to the model.
We can parse the image and prompt from the response and display the generated image.
Clean Up
Once we are done experimenting, we should delete the endpoint. In the last example, we simply call the delete_endpoint() method.
End-to-end examples
You can find more end-to-end examples in the GitHub repo for JumpStart, HuggingFace, PyTorch, and so on.
Conclusion
The new ezsmdeploy Python SDK introduced in this post enables deploying large machine learning models with a single API call. It significantly reduces the complexity of deploying large models in production. By handling deployment intricacies behind the scenes, data scientists can go from model to production API in minutes. The SDK drastically lowers deployment barriers so you can focus on your model while the SDK handles the infrastructure.