
Overview
Databricks at AWS re:Invent 2024
Databricks at AWS re:Invent 2024

Product video
Get started today with up to $400 in usage credits during your 14-day free trial. Trial ends the earlier of when credits are consumed or the 14-day period expires. After your trial ends, you will be automatically enrolled into a Databricks pay-as-you-go plan using the payment method associated with your AWS Marketplace account, paying only for what you use and you can cancel anytime. You can view the full per-product rates for Databricks Units (DBUs) at https://www.databricks.com/product/pricing
The Databricks Data Intelligence Platform allows your entire organization to use data and AI. Its built on a lakehouse to provide an open, unified foundation for all your data and governance. And its powered by a Data Intelligence Engine that speaks the language of your organization so anyone can access the data and insights they need.
The Data Intelligence Platform simplifies your modern data stack by eliminating the data silos that traditionally separate and complicate data engineering, analytics, BI, data science and machine learning. Databricks is built on open source and open standards to maximize flexibility. And the platforms common approach to data management, security and governance helps you operate more efficiently and innovate faster across all analytics use cases.
Reach out to sales@databricks.com to get specialized configurations and pricing for Databricks on AWS Marketplace on a contract basis.
** Technical Support: For help setting up your account, connecting to data, or exploring the platform please reach out to awsmp-onboarding-help@databricks.com **
Highlights
- Simple: Databricks provides a simplified data architecture by unifying data, analytics and AI workloads on one common platform running on Amazon S3.
- Open: Built on top of the world's most successful open source data projects, the Lakehouse Platform unifies your data ecosystem with open standards and formats.
- Collaborative: With native collaboration capabilities, the Databricks Lakehouse Platform unifies data teams to collaborate across the entire data and AI workflow.
Details
Introducing multi-product solutions
You can now purchase comprehensive solutions tailored to use cases and industries.
Features and programs
Buyer guide

Financing for AWS Marketplace purchases
Pricing
Free trial
Dimension | Cost/unit |
|---|---|
Databricks Consumption Units | $1.00 |
Vendor refund policy
No refunds
Custom pricing options
How can we make this page better?
Legal
Vendor terms and conditions
Content disclaimer
Delivery details
Software as a Service (SaaS)
SaaS delivers cloud-based software applications directly to customers over the internet. You can access these applications through a subscription model. You will pay recurring monthly usage fees through your AWS bill, while AWS handles deployment and infrastructure management, ensuring scalability, reliability, and seamless integration with other AWS services.
Resources
Support
Vendor support
Please reach out to sales@databricks.com with any questions or for options on contract or pricing terms.
Technical Support: For help setting up your account, connecting to data, or exploring the platform please reach out to awsmp-onboarding-help@databricks.com
For additional training:
AWS infrastructure support
AWS Support is a one-on-one, fast-response support channel that is staffed 24x7x365 with experienced and technical support engineers. The service helps customers of all sizes and technical abilities to successfully utilize the products and features provided by Amazon Web Services.


Standard contract
Customer reviews
Databricks Simplifies Big Data Processing and Team Collaboration
Powerful Lakehouse Platform for Scalable Pipelines and Collaboration
Perfect for Cross-team Collaboration and Intensive Data Applications
The integrations are also excellent. It works smoothly with the broader cloud ecosystem and connects well with data sources, orchestration tools, model serving infrastructure, and external systems. That interoperability makes it much easier to move from prototype to deployed pipeline without constantly rebuilding connectors or managing glue code.
Performance has been consistently strong, especially when working with distributed workloads and large-scale feature engineering. Spark optimization, cluster management, and managed infrastructure significantly reduce operational overhead, which lets me focus more on model development and analysis rather than environment tuning. For iterative experimentation, spin-up times and overall responsiveness are noticeably better than many alternative managed platforms.
The AI integration is another area that still feels somewhat uneven. While there’s a clear push toward positioning the platform as an end-to-end AI/ML environment, some of the newer AI-focused features feel more like ecosystem additions than deeply integrated workflow improvements. In practice, there are still cases where custom tooling or external frameworks provide more flexibility and transparency, particularly for specialized model development, experimentation, and real-time inference use cases.
There can also be some complexity around tuning clusters and managing costs efficiently at scale. While the abstractions are helpful, getting the best performance-to-cost ratio sometimes requires deeper platform knowledge than the “fully managed” positioning might imply.
Overall, the platform is very strong technically, but pricing for always-on data-intensive workloads and the maturity of some AI-native capabilities are the two biggest areas where I’d like to see improvement.
For my work, the biggest benefit is real-time collaboration. It allows multiple people to work against the same datasets, notebooks, and pipelines without the usual friction of fragmented tooling or environment inconsistencies. That significantly speeds up experimentation, iteration, and knowledge sharing across projects, especially when moving quickly on model development or analyzing fast-changing data.
It also solves the challenge of scalable data access and processing. Working with high-volume time-series and transactional datasets requires infrastructure that can process large amounts of data efficiently without constant operational overhead. Databricks abstracts much of that complexity, making it possible to focus on analysis, feature engineering, and model development rather than spending time managing infrastructure.
The practical benefit is faster iteration cycles. I can move from raw data exploration to model experimentation and deployment much more quickly, which is especially valuable when working on real-time analytics, forecasting pipelines, and production-facing ML systems where speed of iteration directly impacts outcomes.
Overall, it reduces engineering friction and makes large-scale collaborative data work significantly more efficient, which translates into faster development, better experimentation, and more reliable deployment of data products.