This Guidance showcases the versatility of the TetraScience Tetra Data Platform (TDP) and its seamless integration capabilities with other AWS services. TDP is a cloud-native solution that manages scientific data from various sources, such as instruments, contract research organizations, manufacturing facilities, and software systems. TDP centralizes this data into a scientific data lake hosted on AWS. The various ways you can integrate TDP with the rest of your AWS environment include high-performance computing (HPC), data analytics, data lakes, machine learning (ML), and AWS Partner Solutions. Using the multiple interfaces shown throughout this solution, TDP can accelerate the integration of a laboratory data mesh on AWS.

Please note: [Disclaimer]

Architecture Diagram

[Architecture diagram description]

Download the architecture diagram PDF 

Well-Architected Pillars

The AWS Well-Architected Framework helps you understand the pros and cons of the decisions you make when building systems in the cloud. The six pillars of the Framework allow you to learn architectural best practices for designing and operating reliable, secure, efficient, cost-effective, and sustainable systems. Using the AWS Well-Architected Tool, available at no charge in the AWS Management Console, you can review your workloads against these best practices by answering a set of questions for each pillar.

The architecture diagram above is an example of a Solution created with Well-Architected best practices in mind. To be fully Well-Architected, you should follow as many Well-Architected best practices as possible.

  • This Guidance uses Amazon CloudWatch, AWS CloudFormation, and Amazon Simple Notification Service (Amazon SNS) to enhance operational excellence. CloudFormation stacks contain custom alarms configured to automatically invoke actions based on system behavior, while CloudWatch dashboards visualize historical performance data. CloudWatch alarms can invoke notifications through Amazon SNS when thresholds are crossed. Together, these services enable real-time monitoring, alerting, and incident response to ensure the Tetra Data Platform runs smoothly, helping your team provide timely and effective support if unexpected issues arise.

    Read the Operational Excellence whitepaper 
  • When configuring this Guidance, we recommend using IAM, AWS KMS, and AWS Secrets Manager to augment your security posture. IAM enforces strict access controls over data and resources through policies and roles, following the principle of least privilege. AWS KMS enables centralized key management and 256-bit encryption using Advanced Encryption Standard (AES) 256 to protect sensitive data stored in Amazon S3. Secrets Manager centrally manages access credentials for external APIs and Amazon RDS databases. Together, these services allow you to build security best practices like encryption, access controls, and secret management directly into the Tetra Data Platform architecture.

    Read the Security whitepaper 
  • The fully-managed AWS services used in this architecture scale automatically as data volumes grow, maintaining 99.99% uptime. Specifically, Amazon S3 provides 99.999999999% (11 nines) durability and 99.99% availability through redundant storage. Amazon RDS deploys in multiple Availability Zones (AZs) with synchronous replication for high availability. OpenSearch Service auto-scales across AZs to sustain zone failure. By building your data lake architecture on the intrinsically reliable infrastructure of AWS, you offload responsibility for availability, backups, scaling, and disaster recovery to AWS.

    Read the Reliability whitepaper 
  • Athena and OpenSearch Service help make your workloads more efficient. Athena completes queries in parallel, so results return in seconds, while Amazon OpenSearch Serverless, an on-demand auto-scaling configuration for OpenSearch Service, auto-scales resources to maintain fast ingestion and query speeds as data volumes grow. Athena and OpenSearch Service relieve the burden of fine-tuning data pipelines and indexes, ensuring scientists get rapid responses to queries across exponentially growing datasets.

    Read the Performance Efficiency whitepaper 
  • This Guidance uses Amazon S3 and Athena to optimize costs, with Amazon S3 offering inexpensive, scalable object storage while Athena charges only for queries run. Together, these serverless services scale on demand, so you pay only for what you use. Furthermore, automatically tiering data across Amazon S3 storage classes optimizes price performance as access patterns change. Athena also allows tuning query patterns to minimize scanned data and costs. Building on this variable-spend infrastructure means you don't pay for unused capacity.

    Read the Cost Optimization whitepaper 
  • The centralized data lake of Amazon S3 eliminates redundant copies, while AWS Glue catalogs this data for analysis. Together, these on-demand services scale dynamically to workloads, maximizing resource utilization and minimizing energy demands. Avoiding overprovisioning with serverless architectures optimizes energy consumption to only what's needed for current workloads. Building on this variable-spend infrastructure means you don't leave unused capacity waiting idly. This saves energy for collecting insights rather than maintaining unused servers. AWS empowers innovating sustainably by matching compute costs to each workload's real-time interface patterns.

    Read the Sustainability whitepaper 

Implementation Resources

A detailed guide is provided to experiment and use within your AWS account. Each stage of building the Guidance, including deployment, usage, and cleanup, is examined to prepare it for deployment.

The sample code is a starting point. It is industry validated, prescriptive but not definitive, and a peek under the hood to help you begin.

[Content Type]


This [blog post/e-book/Guidance/sample code] demonstrates how [insert short description].


The sample code; software libraries; command line tools; proofs of concept; templates; or other related technology (including any of the foregoing that are provided by our personnel) is provided to you as AWS Content under the AWS Customer Agreement, or the relevant written agreement between you and AWS (whichever applies). You should not use this AWS Content in your production accounts, or on production or other critical data. You are responsible for testing, securing, and optimizing the AWS Content, such as sample code, as appropriate for production grade use based on your specific quality control practices and standards. Deploying AWS Content may incur AWS charges for creating or using AWS chargeable resources, such as running Amazon EC2 instances or using Amazon S3 storage.

References to third-party services or organizations in this Guidance do not imply an endorsement, sponsorship, or affiliation between Amazon or AWS and the third party. Guidance from AWS is a technical starting point, and you can customize your integration with third-party services when you deploy the architecture.

Was this page helpful?