Sprinklr Reduces Machine Learning Inference Costs on AWS Inferentia
Overview
Sprinklr provides a unified customer experience management (Unified-CXM) platform that combines different applications for marketing, advertising, research, customer care, sales, and social media engagement. As a cloud-first company, Sprinklr is always looking to improve efficiency and optimize its cost savings. To help it achieve its goals, the company uses Amazon Web Services (AWS)—specifically Amazon Elastic Compute Cloud (Amazon EC2), which provides secure, resizable compute capacity in the cloud.
In 2021, Sprinklr had the opportunity to try Amazon EC2 Inf1 Instances, which are powered by AWS Inferentia, a high-performance machine learning (ML) inference chip built from the ground up optimized for ML inference applications. By migrating its real-time workloads on its Unified-CXM platform from GPU-based Amazon EC2 instances onto AWS Inferentia, Sprinklr has realized a significant cost savings and has seen latency reduce by more than 30 percent on those workloads. Moreover, by reducing latency, the company has also improved the performance of its products and services for its customers.

About Sprinklr
With advanced artificial intelligence, Sprinklr’s Unified-CXM platform helps companies deliver human experiences to every customer, every time, across any modern channel. Headquartered in New York City, Sprinklr works with over 1,000 global enterprises and over 50 percent of the Fortune 100.

The goal is always to have lower latency, which means a better customer experience. Using Amazon EC2 Inf1 Instances, we are able to achieve that.
Jamal Mazhar
Vice President of Infrastructure and DevOps, SprinklrAWS Services Used
Get Started
Did you find what you were looking for today?
Let us know so we can improve the quality of the content on our pages