AWS Database Blog
Category: Learning Levels
SQL to NoSQL: Modernizing data access layer with Amazon DynamoDB
The transition from SQL-based access patterns to a DynamoDB API-driven approach presents opportunities to optimize how your application interacts with its data layer. This final part of our series focuses on implementing an effective abstraction layer and handling various data access patterns in DynamoDB.
SQL to NoSQL: Modeling data in Amazon DynamoDB
In this post, we explore strategies for designing DynamoDB data models, including entity identification, table design decisions, and relationship modeling approaches. We examine practical scenarios comparing different modeling strategies, helping you make informed decisions for your specific use case.
SQL to NoSQL: Planning your application migration to Amazon DynamoDB
This is the first part of a series exploring how to effectively migrate from SQL to DynamoDB. We will examine how to analyze existing database structures and access patterns to prepare for migration, focusing on schema analysis, query patterns, and usage metrics that inform DynamoDB data model design.
AWS DMS validation: A custom serverless architecture
AWS DMS customers might choose not to use the data validation feature provided by the AWS DMS service due to the time it takes to complete validation after a load, a large dataset transfer or a data reload, where business requires rapid availability of data in the target environment. As a result, you might opt to perform validation manually or use a single pass full load only validation, which requires additional effort and time. In this post, we walk you through how to build a custom AWS DMS data validation solution with AWS serverless services.
Fluent Commerce’s approach to near-zero downtime Amazon Aurora PostgreSQL upgrade at 32 TB scale using snapshots and AWS DMS ongoing replication
Fluent Commerce, an omnichannel commerce platform, offers order management solutions that enable businesses to deliver seamless shopping experiences across various channels. Fluent uses Amazon Aurora PostgreSQL-Compatible Edition as its high-performance OLTP database engine to process their customers’ intricate search queries efficiently. Fluent Commerce strategically combined AWS-based upgrade approaches—including snapshot restores and AWS DMS ongoing replication—to seamlessly upgrade their 32 TB Aurora PostgreSQL databases with minimal downtime. In this post, we explore a pragmatic and cost-effective approach to achieve near-zero downtime during database upgrades. We explore the method of using the snapshot and restore method followed by continuous replication using AWS DMS.
Accelerate SQL Server to Amazon Aurora migrations with a customizable solution
Migrating from SQL Server to Amazon Aurora can significantly reduce database licensing costs and modernize your data infrastructure. To accelerate your migration journey, we have developed a migration solution that offers ease and flexibility. You can use this migration accelerator to achieve fast data migration and minimum downtime while customizing it to meet your specific business requirements. In this post, we showcase the core features of the migration accelerator, demonstrated through a complex use case of consolidating 32 SQL Server databases into a single Amazon Aurora instance with near-zero downtime, while addressing technical debt through refactoring.
Restore an Amazon RDS Custom for SQL Server instance using a backup from AWS Backup
AWS Backup supports the creation of on-demand backups of RDS Custom for SQL Server instances. However, the restoration of RDS Custom for SQL Server instances through AWS Backup is not natively supported at the time of writing this post. Nonetheless, this post presents a workaround solution that enables the successful restoration of RDS Custom for SQL Server instances using AWS Backup-created backups.
Better together: Amazon RDS for SQL Server and Amazon SageMaker Lakehouse, a generative AI data integration use case
Generative AI solutions are transforming how businesses operate worldwide. It has now become paramount for businesses to integrate generative AI capabilities into their customer-facing services and applications. The challenge they often face is the need to use massive amounts of relational data hosted on SQL Server databases to contextualize these new generative AI solutions. In this post, we demonstrate how you can address this challenge by combining Amazon RDS for SQL Server and Amazon SageMaker Lakehouse.
Announcing Valkey GLIDE 2.0 with support for Go, OpenTelemetry, and batching
AWS recently announced, in partnership with Google Cloud and the Valkey community, the general availability of Valkey General Language Independent Driver for the Enterprise (GLIDE) 2.0, the latest release. Valkey GLIDE is multi-language client library designed for reliability and performance. In this post, we discuss what Valkey GLIDE is and its key benefits, and then dive into its new enhancements.
Supercharging AWS database development with AWS MCP servers
Amazon Aurora, Amazon DynamoDB, and Amazon ElastiCache are popular choices for developers powering critical workloads, including global commerce platforms, financial systems, and real-time analytics applications. To enhance productivity, developers are supplementing everyday tasks with AI-assisted tools that understand context, suggest improvements, and help reason through system configurations. Model Context Protocol (MCP) is at the helm of this revolution, rapidly transforming how developers integrate AI assistants into their development pipelines. In this post, we explore the core concepts behind MCP and demonstrate how new AWS MCP servers can accelerate your database development through natural language prompts.