AWS Open Source Blog

Category: PyTorch on AWS

Deploy PyTorch model in TorchServe and host in Amazon SageMaker inference endpoint

Over the past few years, has become one of the most cutting-edge, open source, deep learning frameworks and the go-to choice for many machine learning use cases based on PyTorch. It has not only democratized deep learning and made it approachable to general audiences, but has also become a role model on how […]

Read More
How TalkingData leverage DJL with PyTorch for Large-Scale Offline Inference

How TalkingData uses AWS open source Deep Java Library with Apache Spark for machine learning inference at scale

This post is contributed by Xiaoyan Zhang, a Data Scientist from TalkingData. TalkingData is a data intelligence service provider that offers data products and services to provide businesses insights on consumer behavior, preferences, and trends. One of TalkingData’s core services is leveraging machine learning and deep learning models to predict consumer behaviors (e.g., likelihood of […]

Read More
workflow: how to deploy TorchServe on an Amazon EKS cluster for inference, which will allow you to quickly deploy a pre-trained machine learning model as a scalable, fault-tolerant web-service for low latency inference

Running TorchServe on Amazon Elastic Kubernetes Service

This article was contributed by Josiah Davis, Charles Frenzel, and Chen Wu. TorchServe is a model serving library that makes it easy to deploy and manage PyTorch models at scale in production environments. TorchServe removes the heavy lifting of deploying and serving PyTorch models with Kubernetes. TorchServe is built and maintained by AWS in collaboration […]

Read More