AWS Database Blog

Category: Advanced (300)

Build secure multi-party computation (MPC) wallets using AWS Nitro Enclaves

Different types of blockchain applications and users demand varying types of private key management solutions, referred to as wallets. Custodial wallets are managed by third-party vendors such as a centralized crypto exchange, whereas non-custodial wallets give you full control and ownership over your private keys and funds. In this post, we focus on multi-party computation (MPC) wallets. We introduce the core concepts of MPC wallets, including the security features they offer that make MPC wallets uniquely well suited for institutional customers. We also detail the most critical aspects of implementing a distributed, highly secure MPC wallet on AWS, including a design for a single MPC cosigner that uses AWS Nitro Enclaves to protect the most sensitive information: the key shards.

Migrate from SAP ASE to SAP ASE on Amazon EC2 using AWS DMS and SAP ASE native methods

In this post, we provide different options for data migration from an SAP ASE on-premises database to SAP ASE on Amazon Elastic Compute Cloud (Amazon EC2) based on the size of data, application downtime, and data compliance. The migration methods include using AWS Database Migration Service (AWS DMS) and SAP ASE native features.

How to clone master user permissions in Amazon RDS for SQL Server

When you create a new Amazon Relational Database Service (Amazon RDS) for SQL Server instance, the master user is granted certain privileges for that database instance. We strongly recommend that you do not use the master user directly in your applications. Instead, adhere to the principles of least privilege and best practices by using a database user created with the minimal permissions required for your application. In this post, we discuss how to clone the master user to a new login and review it for the minimal permissions required.

Understanding Amazon Aurora MySQL storage space utilization

Storage in Amazon Aurora MySQL is managed differently from traditional MySQL databases. In this post, we explore the different types of storage available in Amazon Aurora MySQL, how Aurora uses those storage types, and how to monitor storage consumption. We also explore some of the database queries and Amazon CloudWatch metrics for Aurora that you can use to estimate Aurora storage billing.

How London Stock Exchange Group optimised blue/green deployments for Amazon Aurora PostgreSQL Global Database

In this post we share how the London Stock Exchange Group (LSEG) Capital Markets Business unit improved their Blue/Green software deployment methodology, by using continuous logical database replication. We show you the process of implementing a Blue/green deployment architecture using Aurora PostgreSQL Global Database. Specifically, we explore best practices and considerations when configuring the architecture. Blue/green deployment serves as a robust and efficient approach to make sure applications stay resilient and synchronized throughout the process.

AWS DMS homogenous migration from PostgreSQL to Amazon Aurora PostgreSQL

With AWS DMS homogenous migration, you can migrate data from your source database to an equivalent engine on AWS using native database tools. In this post, we show you an example of a complete homogeneous migration process and provide troubleshooting steps for migrating from PostgreSQL to Amazon Aurora PostgreSQL and Amazon RDS for PostgreSQL.

Use Spring Cloud to capture Amazon DynamoDB changes through Amazon Kinesis Data Streams

In this post, we demonstrate how you can use Spring Cloud to interact with Amazon DynamoDB and capture table-level changes using Kinesis Data Streams through familiar Spring constructs. We run you through a basic implementation and configuration that will help you get started.

Application Continuity for Oracle workloads with Amazon RDS Custom for Oracle

In this post, we show you how to implement Application Continuity in an RDS Custom for Oracle environment using a sample application. We also show you how to test the implementation to see that, when an outage occurs at the database tier, the application recovers and resumes without any data loss—automatically and transparently—along with the database failover. Finally, we show you how to verify the results before cleaning up the environment.