[SEO Subhead]
The Guidance demonstrates how to create an enterprise-governed model in HighByte, ingest real-time, historical, and transactional data at scale from edge and cloud data sources into Snowflake, and interface with applications using REST APIs. It addresses challenges faced by operations leaders at manufacturing and industrial companies with disconnected, siloed data sources by enabling a scalable, unified, and integrated mechanism to harness data as an asset. By providing economical, secure, and easy access to high-quality datasets, the Guidance helps you build the foundation for digital industrial transformation and optimize operations across quality, maintenance, materials management, and process optimization.
Note: [Disclaimer]
Architecture Diagram
[Architecture diagram description]
Step 1
Define enterprise-standardized and contextualized data models using Central HighByte Intelligence Hub, and push the models to on-premises remote hubs.
Step 2
Ingest operational technology (OT) data using HighByte from industrial data sources including: programmable logic controller (PLC), supervisory control and data acquisition (SCADA), and message queuing telemetry transport (MQTT).
This may also include other sources, such as manufacturing execution system (MES), enterprise asset management (EAM), quality management system (QMS), historians, databases, files, and more.
Step 3
A direct, low-latency connection from HighByte to Snowpipe Streaming then connects to Snowflake tables to efficiently insert data.
Step 4
Publish to Snowpipe Streaming through Amazon Managed Streaming for Apache Kafka (Amazon MSK). Host a Kafka connector for Snowpipe Streaming on Amazon MSK Connect.
Step 5
Query data from Snowflake using direct table access with HighByte Intelligence Hub and the Snowflake SQL connector.
Step 6
Connect to Amazon Simple Storage Service (Amazon S3) using a HighByte S3 connector. Move large or historical datasets into S3 buckets for Snowflake Snowpipe services to automatically retrieve.
Step 7
Transform data using Snowpark to leverage the power of Java or Python. Data can then be aggregated and prepared using SQL for analytics.
Step 8
Create real-time dashboards in Amazon Managed Grafana using a native connector for Snowflake.
Step 9
Create Streamlit apps running on Amazon Elastic Compute Cloud (Amazon EC2) to display real-time dashboards from the raw data zone and historical analysis from the analytics zone in Snowflake using the Python or Snowpark connector.
Step 10
Analyze and visualize data using Amazon QuickSight with a native connector to Snowflake.
Get Started
Well-Architected Pillars
The AWS Well-Architected Framework helps you understand the pros and cons of the decisions you make when building systems in the cloud. The six pillars of the Framework allow you to learn architectural best practices for designing and operating reliable, secure, efficient, cost-effective, and sustainable systems. Using the AWS Well-Architected Tool, available at no charge in the AWS Management Console, you can review your workloads against these best practices by answering a set of questions for each pillar.
The architecture diagram above is an example of a Solution created with Well-Architected best practices in mind. To be fully Well-Architected, you should follow as many Well-Architected best practices as possible.
-
Operational Excellence
Serverless offerings such as Amazon S3, Amazon MSK, and Amazon Managed Grafana offload the burden of managing underlying infrastructure. This allows the focus to remain on core functionality and continuous improvement of your workloads. These AWS services provide scalability, elasticity, high availability, and durability, automatically scaling up or down based on demand.
-
Security
Certificate-based encryption enhances the security of this Guidance by authenticating communicating parties and helping to ensure confidentiality, integrity, and non-repudiation of data. This automated approach eliminates manual key exchange, establishing a trusted, encrypted channel for data transfer between systems like HighByte Intelligence Hub and Snowflake, safeguarding sensitive information.
-
Reliability
Services like Amazon S3, Amazon MSK, and Amazon Managed Grafana are fully managed by AWS, handling infrastructure, software updates, patches, and ongoing maintenance. These scalable services automatically adjust resources to meet fluctuating demand, helping to ensure consistent performance and availability during high traffic periods. Additionally, built-in disaster recovery and backup capabilities provide data durability and recovery from failures or outages, further enhancing reliability.
-
Performance Efficiency
The Guidance leverages purpose-built storage services and features like Amazon S3 cross-region replication (CRR) to reduce latency, increase throughput, and support scalability for data-driven workloads. By using geographically distributed storage across AWS Regions, the Guidance can provide lower-latency data access and match specific access patterns. These scalable services allow seamless capacity adjustments to handle fluctuating traffic and data demands for consistent performance.
-
Cost Optimization
This Guidance uses purpose-built storage services such as Amazon S3 and Amazon MSK to reduce latency and increase throughput while optimizing costs. By using managed services, it offloads infrastructure provisioning, configuration, and maintenance burdens, allowing focus on core application functionality. Services such as Amazon S3 and Amazon MSK are optimized for specific low-cost durable storage and high-throughput streaming, respectively. Additionally, managed services reduce operational costs through features like automatic scaling and software updates, minimizing operational overhead.
-
Sustainability
Scalable cloud services like Amazon S3 align with sustainability goals through on-demand usage and efficiency, paying only for consumed resources. Cloud providers' economies of scale and optimized infrastructure often result in lower carbon emissions per unit of computing power. Cloud infrastructure reduces hardware waste by eliminating constant provisioning and replacement of physical hardware.
Related Content
[Title]
Disclaimer
The sample code; software libraries; command line tools; proofs of concept; templates; or other related technology (including any of the foregoing that are provided by our personnel) is provided to you as AWS Content under the AWS Customer Agreement, or the relevant written agreement between you and AWS (whichever applies). You should not use this AWS Content in your production accounts, or on production or other critical data. You are responsible for testing, securing, and optimizing the AWS Content, such as sample code, as appropriate for production grade use based on your specific quality control practices and standards. Deploying AWS Content may incur AWS charges for creating or using AWS chargeable resources, such as running Amazon EC2 instances or using Amazon S3 storage.
References to third-party services or organizations in this Guidance do not imply an endorsement, sponsorship, or affiliation between Amazon or AWS and the third party. Guidance from AWS is a technical starting point, and you can customize your integration with third-party services when you deploy the architecture.