[SEO Subhead]
This Guidance illustrates a comprehensive approach for ingesting, processing, and deriving insights from satellite imagery data. AWS Ground Station is used to ingest the satellite data, and Amazon SageMaker is used to label the image data, train machine learning (ML) models, and deploy the trained models. From ingestion to ML analysis, the trained models can be integrated into your applications or dashboards for analysis and visualization of the satellite data. This holistic approach streamlines the entire lifecycle of satellite data processing and analytics so you can quickly develop and deploy intelligent applications powered by your own satellite data.
Note: [Disclaimer]
Architecture Diagram
[Architecture diagram description]
Step 1
Satellite sends data and imagery to the AWS Ground Station antenna.
Step 2
AWS Ground Station delivers baseband or digitized radio frequency (RF)-over-IP data to an Amazon Elastic Cloud Compute (Amazon EC2) instance.
Step 3
The Amazon EC2 instance receives and processes the data and then stores the data in an Amazon Simple Storage Service (Amazon S3) bucket.
Step 4
A data preparation notebook ingests data from the Amazon S3 bucket to prepare the data for model training.
Step 5
Amazon SageMaker Ground Truth labels the images.
Step 6
The labeled images are stored in the Amazon S3 bucket.
Step 7
The notebook hosts the training algorithm and code.
Step 8
Amazon SageMaker runs the training algorithm on the data and trains the ML model.
Step 9
SageMaker deploys the ML models to an endpoint.
Step 10
The SageMaker ML model processes the image data and stores the generated inferences and metadata in the Amazon DynamoDB database.
Step 11
Image data received into Amazon S3 automatically triggers an AWS Lambda function to run ML services on the image data.
Step 12
Applications interact with AWS Amplify to access the ML algorithm and database.
Get Started
Well-Architected Pillars
The AWS Well-Architected Framework helps you understand the pros and cons of the decisions you make when building systems in the cloud. The six pillars of the Framework allow you to learn architectural best practices for designing and operating reliable, secure, efficient, cost-effective, and sustainable systems. Using the AWS Well-Architected Tool, available at no charge in the AWS Management Console, you can review your workloads against these best practices by answering a set of questions for each pillar.
The architecture diagram above is an example of a Solution created with Well-Architected best practices in mind. To be fully Well-Architected, you should follow as many Well-Architected best practices as possible.
-
Operational Excellence
This Guidance uses an event-driven application pipeline where processing is automated based on Amazon S3 events. This approach relies on Lambda functions, which are triggered whenever new data becomes available in Amazon S3, eliminating the need for scheduled batch processes. The inference results and associated metadata, including quality and accuracy metrics for the ML model, persist in DynamoDB. By using this event-driven architecture, you can automate your data processing pipelines, reducing manual intervention and enabling real-time processing as new data becomes available.
-
Security
AWS Ground Station uses secure communication channels to transmit data between the ground stations and AWS. The data is encrypted in transit using industry-standard encryption protocols, such as TLS. Store your data in Amazon S3 and secure it from unauthorized access with encryption features and access management tools.
-
Reliability
Amazon S3 provides data durability and high availability, and Lambda is both a highly scalable and highly available compute resource that automatically scales up or down based on your workload demands. Each of these services can enhance reliability without additional overhead or management required from you or the user.
-
Performance Efficiency
SageMaker automatically captures and monitors various metrics during model training. These metrics can be key indicators to assist in rightsizing your infrastructure for future training jobs. In addition, Lambda automatically scales compute resources in response to incoming workloads for optimal utilization, eliminating the need to overprovision resources. Together, these services provide a managed experience and reduce the heavy lifting associated with orchestrating ML or compute tasks.
-
Cost Optimization
The Amplify service provides a simplified approach to building and scaling web and mobile applications while also helping to control costs. Amplify offers a pay-as-you-go pricing model, allowing the hosting of multiple sites and providing free SSL certificates, so you only pay for the resources you actually consume. Furthermore, the AWS Ground Station service allows you to pay solely for the actual antenna time used and utilize the global footprint of ground stations to download data when and where needed, without long-term contracts or hidden fees. You can access any antenna in the global AWS Ground Station network at a single price.
-
Sustainability
The use of Lambda removes the requirement for provisioning and maintaining physical servers, thereby reducing the energy consumption associated with traditional compute infrastructure. Additionally, the lifecycle policies of Amazon S3 allow for the automatic transition of data to colder, more cost-effective storage tiers, optimizing storage costs and resource utilization. Furthermore, the AWS Ground Station service eliminates the need for you to build and maintain your own ground station infrastructure, which can be resource-intensive and inefficient.
Related Content
[Title]
Disclaimer
The sample code; software libraries; command line tools; proofs of concept; templates; or other related technology (including any of the foregoing that are provided by our personnel) is provided to you as AWS Content under the AWS Customer Agreement, or the relevant written agreement between you and AWS (whichever applies). You should not use this AWS Content in your production accounts, or on production or other critical data. You are responsible for testing, securing, and optimizing the AWS Content, such as sample code, as appropriate for production grade use based on your specific quality control practices and standards. Deploying AWS Content may incur AWS charges for creating or using AWS chargeable resources, such as running Amazon EC2 instances or using Amazon S3 storage.
References to third-party services or organizations in this Guidance do not imply an endorsement, sponsorship, or affiliation between Amazon or AWS and the third party. Guidance from AWS is a technical starting point, and you can customize your integration with third-party services when you deploy the architecture.