AWS Database Blog

Mask PII data using AWS DMS and Amazon Macie during migration

In this post, we present a solution to identify PII data using Amazon Macie, mask it using AWS Database Migration Service (AWS DMS), and migrate it from an Amazon Relational Database Service (Amazon RDS) for Oracle production source database to an RDS for Oracle development target database, before releasing the environment to users. This way, you can save time and make sure that sensitive data is protected.

Amazon Timestream for Amazon Connect real-time monitoring

Amazon Connect is an easy-to-use cloud contact center solution that helps companies of any size deliver superior customer service at a lower cost. Connect has many real-time monitoring capabilities. For requirements that go beyond those supported out of the box, Amazon Connect also provides you with data and APIs you can use to implement your […]

Powering Amazon RDS with AWS Graviton3: Benchmarks

In April 2023, AWS announced the ability to power Amazon Relational Database Service (Amazon RDS) instances with the AWS Graviton3 processor. AWS Graviton3 was announced in May 2022 and is built using a custom-designed ARM architecture using ARM Neoverse cores, optimized for delivering high performance and energy efficiency. AWS Graviton3 provides up to 25% better […]

Use Amazon DynamoDB incremental export to update Apache Iceberg tables

Amazon DynamoDB is a fully managed, serverless, key-value NoSQL database designed to run high-performance applications at any scale. DynamoDB recently launched a new feature: Incremental export to Amazon Simple Storage Service (Amazon S3). You can use incremental exports to update your downstream systems regularly using only the changed data. You no longer need to do a […]

Techniques to improve the state-of-the-art in Cloud FinOps using Amazon Neptune

Cloud computing has changed almost every business and industry by changing the delivery and consumption model. With the Cloud, businesses no longer need to plan for and procure servers and other IT infrastructure weeks or months in advance. This allows more flexibility and reliability, increased performance and efficiency, and helps to lower IT costs. The […]

How power utilities analyze and detect harmonics issues using power quality and customer usage data with Amazon Timestream: Part 2

In the first post of the series, we demonstrated how to use an Amazon Timestream database and its built-in time series functionalities to interpolate data and calculate the correlation between customer energy usage and power quality issues. In this post, we show you how to build a power quality analysis Proof of Concept (PoC) using […]

Impactful features in PostgreSQL 15

PostgreSQL is one of the most popular open-source relational database systems. The product of more than 30 years of development work, PostgreSQL has proven to be a highly reliable and robust database that can handle a large number of complicated data workloads. AWS offers services that make PostgreSQL database deployments straightforward to set up, manage, […]

Detect and fix low cardinality indexes in Amazon DocumentDB

Amazon DocumentDB (with MongoDB compatibility) is a fully managed native JSON document database that makes it easy and cost effective to operate critical document workloads at virtually any scale without managing infrastructure. It’s a best practice to create indexes to improve query performance especially when database size is very large. Without indexes, queries have to […]

Prewarm an Amazon RDS for Oracle database to reduce the impact of lazy loading

Amazon Relational Database Service (Amazon RDS) uses Amazon Elastic Block Store (Amazon EBS) as an underlying storage. RDS snapshots (automated and manual) are saved in Amazon Simple Storage Service (Amazon S3). For RDS instances that are restored from snapshots, the instances are made available as soon as the needed infrastructure is provisioned. However, there is […]

Implement Time to Live in Amazon Neptune, Part 1: Property Graph

Time to Live (TTL) is a mechanism that helps determine the longevity or lifespan of data, files, infrastructure, or even entire environments. When working with data, it could represent the amount of time a leaderboard expires in memory before being reloaded from storage, or how long a file must be kept for regulatory or compliance […]