We use essential cookies and similar tools that are necessary to provide our site and services. We use performance cookies to collect anonymous statistics, so we can understand how customers use our site and make improvements. Essential cookies cannot be deactivated, but you can choose “Customize” or “Decline” to decline performance cookies.
If you agree, AWS and approved third parties will also use cookies to provide useful site features, remember your preferences, and display relevant content, including relevant advertising. To accept or decline all non-essential cookies, choose “Accept” or “Decline.” To make more detailed choices, choose “Customize.”
Customize cookie preferences
We use cookies and similar tools (collectively, "cookies") for the following purposes.
Essential
Essential cookies are necessary to provide our site and services and cannot be deactivated. They are usually set in response to your actions on the site, such as setting your privacy preferences, signing in, or filling in forms.
Performance
Performance cookies provide anonymous statistics about how customers navigate our site so we can improve site experience and performance. Approved third parties may perform analytics on our behalf, but they cannot use the data for their own purposes.
Allowed
Functional
Functional cookies help us provide useful site features, remember your preferences, and display relevant content. Approved third parties may set these cookies to provide certain site features. If you do not allow these cookies, then some or all of these services may not function properly.
Allowed
Advertising
Advertising cookies may be set through our site by us or our advertising partners and help us deliver relevant marketing content. If you do not allow these cookies, you will experience less relevant advertising.
Allowed
Blocking some types of cookies may impact your experience of our sites. You may review and change your choices at any time by selecting Cookie preferences in the footer of this site. We and selected third-parties use cookies or similar technologies as specified in the AWS Cookie Notice.
Your privacy choices
We display ads relevant to your interests on AWS sites and on other properties, including cross-context behavioral advertising. Cross-context behavioral advertising uses data from one site or app to advertise to you on a different company’s site or app.
To not allow AWS cross-context behavioral advertising based on cookies or similar technologies, select “Don't allow” and “Save privacy choices” below, or visit an AWS site with a legally-recognized decline signal enabled, such as the Global Privacy Control. If you delete your cookies or visit this site from a different browser or device, you will need to make your selection again. For more information about cookies and how we use them, please read our AWS Cookie Notice.
Dieser Inhalt steht in der ausgewählten Sprache nicht zur Verfügung. Wir arbeiten beständig daran, unsere Inhalte auch in der ausgewählten Sprache zur Verfügung zu stellen. Vielen Dank für Ihre Geduld.
PyTorch is an open-source deep learning framework that makes it easier to develop machine learning (ML) models and deploy them to production. You can use PyTorch on AWS to build, train, and deploy state-of-the-art deep learning models. PyTorch on AWS offers high-performance compute, storage, and networking services; open-source contributions to PyTorch, such as TorchElastic and TorchServe; and optimizations such as the Amazon S3 plugin for PyTorch. You can get started using AWS Deep Learning AMIs (DLAMIs), AWS Deep Learning Containers for containerized applications, or Amazon SageMaker for fully managed infrastructure, tools, and workflows.
Benefits
TorchServe
TorchServe is an open-source tool that makes it easier to deploy trained PyTorch models performantly at scale. TorchServe delivers lightweight serving with low latency, so you can deploy your models for high-performance inference. TorchServe also provides default handlers, such as object detection and text classification, for the most common applications, so you don’t have to write custom code to deploy your models. With powerful TorchServe features such as multimodal serving, model versioning for A/B testing, metrics for monitoring, and RESTful endpoints for application integration, you can quickly take your models from research to production. TorchServe supports any ML environment, including Amazon SageMaker, Kubernetes, Amazon Elastic Kubernetes Service (EKS), and Amazon Elastic Compute Cloud (EC2). To get started with TorchServe, see the documentation and our blog post.
TorchElastic Controller for Kubernetes
TorchElastic is a library for training large-scale deep learning models where it is critical to dynamically scale compute resources based on availability. Elastic and fault-tolerant training with TorchElastic can help you take ML models to production more quickly and adopt state-of-the-art approaches to model exploration as architectures continue to increase in size and complexity.
The TorchElastic Controller for Kubernetes is a native Kubernetes implementation for TorchElastic that automatically manages the lifecycle of the pods and services required for TorchElastic training. It allows you to start training jobs with a portion of the requested compute resources and dynamically scale as more resources become available, without having to stop and restart the jobs. In addition, jobs can recover from nodes that are replaced because of node failures or reclamation.
Amazon S3 plugin
Amazon S3 plugin for PyTorch is an open-source library intended for use with the PyTorch deep learning framework for streaming data from Amazon Simple Storage Service (S3). With this feature available in PyTorch Deep Learning Containers, you can use data from S3 buckets directly with PyTorch APIs without first needing to download data to local storage.
PyTorch support in the AWS Neuron SDK
The AWS Neuron SDK is integrated with PyTorch, providing developers with a familiar environment to run their machine learning inference on AWS Inferentia based Amazon EC2 Inf1 instances. The AWS Neuron SDK allows PyTorch models to execute on EC2 Inf1 instances and implements data parallelism on the models, which allows dynamic batching and parallelized inference for faster performance.