AWS Database Blog
Category: Intermediate (200)
How Iterate.ai uses Amazon MemoryDB to accelerate and cost-optimize their workforce management conversational AI agent
Iterate.ai is an enterprise AI platform company delivering innovative AI solutions to industries such as retail, finance, healthcare, and quick-service restaurants. Among its standout offerings is Frontline, a workforce management platform powered by AI, designed to support and empower Frontline workers. Available on both the Apple App Store and Google Play, Frontline uses advanced AI tools to streamline operational efficiency and enhance communication among dispersed workforces. In this post, we give an overview of durable semantic caching in Amazon MemoryDB, and share how Iterate used this functionality to accelerate and cost-optimize Frontline.
Diving deep into the new Amazon Aurora Global Database writer endpoint
On October 22, 2024, we announced the availability of the Aurora Global Database writer endpoint, a highly available and fully managed endpoint for your global database that Aurora automatically updates to point to the current writer instance in your global cluster after a cross-Region switchover or failover, alleviating the need for application changes and simplifying routing requests to the writer instance. In this post, we dive deep into the new Global Database writer endpoint, covering its benefits and key considerations for using it with your applications.
Use Amazon Neptune Analytics to analyze relationships in your data faster, Part 1: Introducing Parquet and CSV import and export
In this two-part series, we show how you can import and export using Parquet and CSV to quickly gather insights from your existing graph data. Part 1 introduces the import and export functionalities, and walks you through how to quickly get started with them. In Part 2, we show how you can use the new data mobility improvements in Neptune Analytics to enhance fraud detection.
How Skello uses AWS DMS to synchronize data from a monolithic application to microservices
Skello is a human resources (HR) software-as-a-service (SaaS) platform that focuses on employee scheduling and workforce management. It caters to various sectors, including hospitality, retail, healthcare, construction, and industry. In this post, we show how Skello uses AWS Database Migration Service (AWS DMS) to synchronize data from an monolithic architecture to microservices and perform data ingestion from the monolithic architecture and microservices to our data lake.
How Orca Security optimized their Amazon Neptune database performance
Orca Security, an AWS Partner, is an independent cybersecurity software provider whose patented agentless-first cloud security platform is trusted by hundreds of enterprises globally. At Orca Security, we use a variety of metrics to assess the significance of security alerts on cloud assets. Our Amazon Neptune database plays a critical role in calculating the exposure of individual assets within a customer’s cloud environment. By building a graph that maps assets and their connectivity between one another and to the broader internet, the Orca Cloud Security Platform can evaluate both how an asset is exposed as well as how an attacker could potentially move laterally within an account. In this post, we explore some of the key strategies we’ve adopted to maximize the performance of our Amazon Neptune database.
Vacasa’s migration to Amazon Aurora for a more efficient Property Management System
Vacasa is North America’s leading vacation rental management platform, revolutionizing the rental experience with advanced technology and expert teams. In the competitive short-term vacation property management industry, efficient systems are critical. To maintain its edge and continue providing top-notch service, Vacasa needed to modernize its primary transactional database to improve performance, provide high availability, and reduce costs. In this post, we share Vacasa’s journey from Amazon Relational Database Service (Amazon RDS) for MariaDB to Amazon RDS for MySQL, and finally to Amazon Aurora, highlighting the technical steps taken and the outcomes achieved.
Enhance the reliability of airlines’ mission-critical baggage handling using Amazon DynamoDB
In the world of air travel, baggage handling isn’t just about keeping track of baggage, but a seamless orchestration of different processes to improve the passenger baggage experience. A key component to make this happen is a strong database management strategy. In this post, we discuss how AWS Partner IBM Consulting developed an initiative to modernize a traditional baggage database architecture using Amazon DynamoDB and other Amazon Web Services (AWS) managed services, addressing the evolving needs of the airline industry.
Transition from AWS DMS to zero-ETL to simplify real-time data integration with Amazon Redshift
The zero-ETL integrations for Amazon Redshift are designed to automate data movement into Amazon Redshift, eliminating the need for traditional ETL pipelines. With zero-ETL integrations, you can reduce operational overhead, lower costs, and accelerate your data-driven initiatives. This enables organizations to focus more on deriving actionable insights and less on managing the complexities of data integration. In this post, we discuss the best practices for migrating your ETL pipeline from AWS DMS to zero-ETL integrations for Amazon Redshift.
How Monzo Bank reduced cost of TTL from time series index tables in Amazon Keyspaces
At Monzo, we use Amazon Keyspaces (for Apache Cassandra) as our main operational database. Today, we store over 350 TB of data across more than 2,000 tables in Amazon Keyspaces, handling over 2,000,000 reads and 100,000 writes per second at peak. In this post, we share how we used a different mechanism for row expiry than the Time to Live setting in Amazon Keyspaces to reduce our operating costs for an index while preserving its semantics.
Reduce latency and cost in read-heavy applications using Amazon DynamoDB Accelerator
Amazon DynamoDB Accelerator (DAX) is a fully managed, in-memory cache for DynamoDB. By using DAX with DynamoDB, you can improve the latency for read requests in your application. In this post, we discuss how to improve latency and reduce cost when using DynamoDB for your read-heavy applications.