Accelerating Innovation for AI/ML Using AWS Services with Arcee AI
Learn how software startup Arcee AI accelerated innovation using Amazon EC2 Capacity Blocks for ML, AWS Inferentia2, and AWS Trainium.
Benefits
97.94%
reduction in training cost using AWS TrainiumReduced from 17 hours
to 1.6-hour training using AWS TrainiumOverview
Artificial intelligence (AI) is rapidly advancing, compelling many startups to develop sophisticated models to tackle a variety of use cases. The larger and more complex the model, the more the GPUs that are required to train it. Arcee AI develops domain-adapted small language models (SLMs) to help enterprises perform specialized tasks, such as analyzing legal documents. It needs to have reliable access to GPUs to manage its compute-intensive workloads.
When Arcee AI went to market, it needed a large amount of accelerated compute resources to train the many SLMs that it is continually developing. But because of ongoing GPU shortages, the startup lacked the compute that is required to effectively train its models and bring them into production. To navigate this obstacle, Arcee AI implemented a system for flexible compute reservation on Amazon Web Services (AWS) and adopted AI chips from AWS to optimize price performance.

About Arcee AI
Arcee AI develops domain-adapted small language models to help enterprises build, deploy, and manage generative artificial intelligence models. These specialized models operate entirely within a client’s virtual private cloud.

If you want to quickly develop an application that needs to be capable of different things in the future, AWS is a great place to build it.
Jacob Solawetz
Cofounder and Chief Technology Officer, Arcee AIAWS Services Used
Get Started
Did you find what you were looking for today?
Let us know so we can improve the quality of the content on our pages