
Overview
Databricks at AWS re:Invent 2024
Databricks at AWS re:Invent 2024

Product video
Get started today with up to $400 in usage credits during your 14-day free trial. Trial ends the earlier of when credits are consumed or the 14-day period expires. After your trial ends, you will be automatically enrolled into a Databricks pay-as-you-go plan using the payment method associated with your AWS Marketplace account, paying only for what you use and you can cancel anytime. You can view the full per-product rates for Databricks Units (DBUs) at https://www.databricks.com/product/pricing
The Databricks Data Intelligence Platform allows your entire organization to use data and AI. Its built on a lakehouse to provide an open, unified foundation for all your data and governance. And its powered by a Data Intelligence Engine that speaks the language of your organization so anyone can access the data and insights they need.
The Data Intelligence Platform simplifies your modern data stack by eliminating the data silos that traditionally separate and complicate data engineering, analytics, BI, data science and machine learning. Databricks is built on open source and open standards to maximize flexibility. And the platforms common approach to data management, security and governance helps you operate more efficiently and innovate faster across all analytics use cases.
Reach out to sales@databricks.com to get specialized configurations and pricing for Databricks on AWS Marketplace on a contract basis.
** Technical Support: For help setting up your account, connecting to data, or exploring the platform please reach out to awsmp-onboarding-help@databricks.com **
Highlights
- Simple: Databricks provides a simplified data architecture by unifying data, analytics and AI workloads on one common platform running on Amazon S3.
- Open: Built on top of the world's most successful open source data projects, the Lakehouse Platform unifies your data ecosystem with open standards and formats.
- Collaborative: With native collaboration capabilities, the Databricks Lakehouse Platform unifies data teams to collaborate across the entire data and AI workflow.
Details
Introducing multi-product solutions
You can now purchase comprehensive solutions tailored to use cases and industries.
Features and programs
Buyer guide

Financing for AWS Marketplace purchases
Pricing
Free trial
Dimension | Cost/unit |
|---|---|
Databricks Consumption Units | $1.00 |
Vendor refund policy
No refunds
Custom pricing options
How can we make this page better?
Legal
Vendor terms and conditions
Content disclaimer
Delivery details
Software as a Service (SaaS)
SaaS delivers cloud-based software applications directly to customers over the internet. You can access these applications through a subscription model. You will pay recurring monthly usage fees through your AWS bill, while AWS handles deployment and infrastructure management, ensuring scalability, reliability, and seamless integration with other AWS services.
Resources
Support
Vendor support
Please reach out to sales@databricks.com with any questions or for options on contract or pricing terms.
Technical Support: For help setting up your account, connecting to data, or exploring the platform please reach out to awsmp-onboarding-help@databricks.com
For additional training:
AWS infrastructure support
AWS Support is a one-on-one, fast-response support channel that is staffed 24x7x365 with experienced and technical support engineers. The service helps customers of all sizes and technical abilities to successfully utilize the products and features provided by Amazon Web Services.


Standard contract
Customer reviews
End-to-End Data Management with Databricks
All-in-One Powerhouse with Room for Pricing Clarity
Unified Data Engineering, Science, and Analytics in One Collaborative Platform
Another highlight is the integration with popular tools and cloud services that are widely used in the market today, which makes it easier to move data between them. The performance monitoring and job scheduling features help maintain visibility over pipelines, and the Delta Lake support for reliable data management has also been very useful.
Additionally, some of the more advanced features—such as fine-grained access controls and more complex job orchestration—can feel less intuitive. The documentation is extensive, but it occasionally leaves gaps that end up requiring trial and error.
Another challenge is ensuring data consistency and reliability across pipelines. With Delta Lake, Databricks provides ACID-compliant storage, versioned tables, and schema enforcement, which reduces data errors and simplifies data governance. This is especially beneficial when multiple teams are working on different stages of data pipelines at the same time.
Databricks also helps solve the problem of fragmented workflows for data scientists and engineers. Its unified environment supports multiple languages (Python, SQL, R, Scala) and includes integrated machine learning with MLFlow, making it easier to collaborate and move from data preparation to analytics and ML in one place.
Scalable Power with Manageable Trade-offs
It beats emailing notebooks back and forth or wrestling with merge conflicts; it feels like pair programming, but for data pipelines. And when you pair that with Delta Lake’s reliability for keeping my ETL jobs rock-solid on intermodal lane data, it ends up being a huge workflow saver.
Top notebook perks for me are the real-time editing and sharing that keeps everyone aligned during debugging, the built-in version history that lets me roll back mistakes quickly, and the seamless Spark integration so I’m not constantly context-switching when doing big data transforms.
Debugging intricate Spark job failures in notebooks often involves sifting through extensive log output, which extends resolution time considerably. Additionally, the UI experiences occasional performance delays under high workloads, impacting efficiency when responsiveness is essential.
This benefits me by streamlining pipelines that feed BI tools, reducing processing times from days to hours and minimizing manual infrastructure oversight. Collaborative notebooks further enhance team productivity through real-time editing, eliminating version control issues and accelerating development cycles.
Databricks Unifies Data and AI for Effortless ML at Scale
It makes building and scaling ML models feel much more straightforward, especially with built-in experiment tracking.
The integration with Apache Spark helps handle large datasets without extra setup.
Overall, it just reduces the friction between raw data and actually getting useful AI outcomes.
There’s also a bit of a learning curve, especially when working across notebooks, jobs, and cluster configs.
And for simpler use cases, it can feel like overkill compared to lighter-weight solutions.
It eliminates the need to move data across multiple systems, which reduces latency and pipeline complexity.
For me, that means faster experimentation and smoother deployment of AI models without worrying about infrastructure.
Overall, it helps focus more on solving business problems rather than managing tools.