Listing Thumbnail

    Databricks Data Intelligence Platform

     Info
    Deployed on AWS
    Free Trial
    The Databricks Data Intelligence Platform unlocks the power of data and AI for your entire organization. Enjoy up to $400 in usage credits during your 14-day free trial. Cancel anytime. After your trial ends, you will automatically be enrolled into a Databricks pay-as-you-go plan.
    4.6

    Overview

    Play video

    Get started today with up to $400 in usage credits during your 14-day free trial. Trial ends the earlier of when credits are consumed or the 14-day period expires. After your trial ends, you will be automatically enrolled into a Databricks pay-as-you-go plan using the payment method associated with your AWS Marketplace account, paying only for what you use and you can cancel anytime. You can view the full per-product rates for Databricks Units (DBUs) at https://www.databricks.com/product/pricing 

    The Databricks Data Intelligence Platform allows your entire organization to use data and AI. Its built on a lakehouse to provide an open, unified foundation for all your data and governance. And its powered by a Data Intelligence Engine that speaks the language of your organization so anyone can access the data and insights they need.

    The Data Intelligence Platform simplifies your modern data stack by eliminating the data silos that traditionally separate and complicate data engineering, analytics, BI, data science and machine learning. Databricks is built on open source and open standards to maximize flexibility. And the platforms common approach to data management, security and governance helps you operate more efficiently and innovate faster across all analytics use cases.

    Reach out to sales@databricks.com  to get specialized configurations and pricing for Databricks on AWS Marketplace on a contract basis.

    ** Technical Support: For help setting up your account, connecting to data, or exploring the platform please reach out to awsmp-onboarding-help@databricks.com **

    Highlights

    • Simple: Databricks provides a simplified data architecture by unifying data, analytics and AI workloads on one common platform running on Amazon S3.
    • Open: Built on top of the world's most successful open source data projects, the Lakehouse Platform unifies your data ecosystem with open standards and formats.
    • Collaborative: With native collaboration capabilities, the Databricks Lakehouse Platform unifies data teams to collaborate across the entire data and AI workflow.

    Details

    Delivery method

    Deployed on AWS
    New

    Introducing multi-product solutions

    You can now purchase comprehensive solutions tailored to use cases and industries.

    Multi-product solutions

    Features and programs

    Buyer guide

    Gain valuable insights from real users who purchased this product, powered by PeerSpot.
    Buyer guide

    Financing for AWS Marketplace purchases

    AWS Marketplace now accepts line of credit payments through the PNC Vendor Finance program. This program is available to select AWS customers in the US, excluding NV, NC, ND, TN, & VT.
    Financing for AWS Marketplace purchases

    Pricing

    Free trial

    Try this product free according to the free trial terms set by the vendor.

    Databricks Data Intelligence Platform

     Info
    Pricing is based on actual usage, with charges varying according to how much you consume. Subscriptions have no end date and may be canceled any time.
    Additional AWS infrastructure costs may apply. Use the AWS Pricing Calculator  to estimate your infrastructure costs.

    Usage costs (1)

     Info
    Dimension
    Cost/unit
    Databricks Consumption Units
    $1.00

    Vendor refund policy

    No refunds

    Custom pricing options

    Request a private offer to receive a custom quote.

    How can we make this page better?

    We'd like to hear your feedback and ideas on how to improve this page.
    We'd like to hear your feedback and ideas on how to improve this page.

    Legal

    Vendor terms and conditions

    Upon subscribing to this product, you must acknowledge and agree to the terms and conditions outlined in the vendor's End User License Agreement (EULA) .

    Content disclaimer

    Vendors are responsible for their product descriptions and other product content. AWS does not warrant that vendors' product descriptions or other product content are accurate, complete, reliable, current, or error-free.

    Usage information

     Info

    Delivery details

    Software as a Service (SaaS)

    SaaS delivers cloud-based software applications directly to customers over the internet. You can access these applications through a subscription model. You will pay recurring monthly usage fees through your AWS bill, while AWS handles deployment and infrastructure management, ensuring scalability, reliability, and seamless integration with other AWS services.

    Support

    Vendor support

    Please reach out to sales@databricks.com  with any questions or for options on contract or pricing terms.

    Technical Support: For help setting up your account, connecting to data, or exploring the platform please reach out to awsmp-onboarding-help@databricks.com 

    For additional training:

    AWS infrastructure support

    AWS Support is a one-on-one, fast-response support channel that is staffed 24x7x365 with experienced and technical support engineers. The service helps customers of all sizes and technical abilities to successfully utilize the products and features provided by Amazon Web Services.

    Product comparison

     Info
    Updated weekly

    Accolades

     Info
    Top
    10
    In Databases & Analytics Platforms, ML Solutions, Data Analytics
    Top
    10
    In ML Solutions
    Top
    10
    In Data Analysis

    Customer reviews

     Info
    Sentiment is AI generated from actual customer reviews on AWS and G2
    Reviews
    Functionality
    Ease of use
    Customer service
    Cost effectiveness
    Positive reviews
    Mixed reviews
    Negative reviews

    Overview

     Info
    AI generated from product descriptions
    Lakehouse Architecture
    Built on a lakehouse foundation providing unified data storage and governance across data engineering, analytics, BI, data science, and machine learning workloads
    Open Source Integration
    Constructed on open source data projects and open standards to maximize flexibility and interoperability across the data ecosystem
    Data Intelligence Engine
    Powered by a Data Intelligence Engine that enables organizational access to data and insights across diverse user roles and technical skill levels
    Unified Data Platform
    Consolidates data, analytics, and AI workloads on a single common platform running on Amazon S3, eliminating traditional data silos
    Collaborative Capabilities
    Provides native collaboration features enabling data teams to work together across the entire data and AI workflow
    AWS Service Integration
    Secure connectivity to Amazon S3, Amazon Redshift, and Amazon RDS with push-down computation capabilities
    Elastic Compute Scaling
    Distributed processing powered by Amazon EKS supporting Python, R, Spark, and other frameworks for data and ML workloads
    Pre-built AI Workflows
    Integration with AWS AI services including Amazon SageMaker and Amazon Comprehend for accelerated AI development
    Large Language Model Integration
    LLM Mesh connectivity to Amazon Bedrock enabling Chat, RAG, and Agentic workflow capabilities
    Visual Development Interface
    Low-code visual platform for data preparation, pipeline creation, and machine learning model development accessible to both technical and non-technical users
    Workload Auto-scaling
    Intelligently autoscales workloads up and down across hybrid and public cloud environments for optimized cloud infrastructure utilization.
    Multi-function Analytics Platform
    Provides integrated data warehouse, machine learning, and custom analytics capabilities with unified analytic functions to eliminate data silos.
    Shared Data Experience (SDX)
    Implements security and governance policies that are set once and applied consistently across all data and workloads, with portability across supported infrastructures.
    Data Lifecycle Management
    Manages complete data lifecycle functions including ingestion, transformation, querying, optimization, and predictive analytics across multiple cloud environments.
    Unified Security and Governance
    Ensures all workloads share common security, governance, and metadata with capabilities for data discovery, curation, and self-service access controls.

    Contract

     Info
    Standard contract
    No
    No
    No

    Customer reviews

    Ratings and reviews

     Info
    4.6
    702 ratings
    5 star
    4 star
    3 star
    2 star
    1 star
    77%
    21%
    1%
    0%
    0%
    10 AWS reviews
    |
    692 external reviews
    External reviews are from G2  and PeerSpot .
    Deeraj R.

    Databricks’ Unified Platform: Fast SQL, Streamlined Pipelines, and Context-Aware AI

    Reviewed on Mar 27, 2026
    Review provided by G2
    What do you like best about the product?
    The unified platform experience is what keeps me on Databricks. Having notebooks, pipelines, SQL warehouses, ML, and governance all in one place under Unity Catalog means I’m not constantly stitching together five different tools just to get work done.

    Lakeflow Pipelines (formerly DLT) makes it straightforward to build medallion-architecture pipelines, and the Photon engine delivers real performance gains on SQL workloads without requiring any code changes. Recent additions like Genie Code and background agents also show they’re serious about agentic AI—it doesn’t feel like a bolt-on copilot, because it can actually understand your data context through Unity Catalog. Serverless compute has been another big quality-of-life improvement as well, since I no longer have to wait for cluster spin-up when I just want to run quick, ad hoc queries.
    What do you dislike about the product?
    Cost management can be tricky—DBUs add up quickly if you’re not careful with cluster sizing and warehouse auto-scaling. The pricing model also isn’t always transparent, especially when you’re mixing serverless and classic compute.

    Unity Catalog is powerful, but the initial setup and the migration from legacy HMS can be painful, particularly for large orgs with years of existing Hive metastore objects. The documentation is generally good, yet it sometimes lags behind new feature releases. On top of that, the workspace UI can feel sluggish at times, especially when you’re working with a large number of assets.
    What problems is the product solving and how is that benefiting you?
    Before Databricks, our data stack was fragmented — separate tools for ETL, analytics, ML, and governance. That meant constant context-switching, duplicated data, and governance gaps. Databricks consolidates all of that into one lakehouse platform. Delta Lake gives us reliable ACID transactions on the data lake, Unity Catalog handles lineage and access control across the board, and SQL warehouses let our analysts self-serve without needing a separate data warehouse product. It's cut our pipeline development time significantly and made data governance something we can actually enforce consistently instead of hoping for the best.
    Naveena P.

    Databricks Unifies Data Engineering, Science, and Analytics Exceptionally Well

    Reviewed on Mar 27, 2026
    Review provided by G2
    What do you like best about the product?
    The ability to converge data engineering, data science, and analytics on a single platform without compromising on governance, performance, or flexibility is still rare in the industry. Databricks executes this exceptionally well.
    What do you dislike about the product?
    Reducing the spinning time of all purpose clusters and job clusters. It would be more usefula nd helpful if it starts as quick as serverless
    What problems is the product solving and how is that benefiting you?
    In enterprise banking, where regulatory compliance, data accuracy, and operational resilience are non-negotiable, Databricks is solving some of our most critical challenges. As a Lead Data Engineer managing end-to-end ETL pipelines, dashboard delivery, monitoring alerts, and data governance for a major banking client, the platform has become the backbone of our modern data architecture. Databricks unifies our fragmented data landscape through Delta Lake and Unity Catalog, giving us ACID-compliant transactions for reliable ETL, automated lineage for audit-ready governance, and fine-grained access controls to protect sensitive PII and financial data—all while enabling seamless schema evolution to handle the constant changes in source systems. This directly translates to faster, more trustworthy reporting: our dashboards in Power BI and Tableau now pull from a single source of truth, eliminating metric disputes between Risk, Finance, and Compliance teams. On the operational side, native alerting integrated with Slack and PagerDuty, combined with Databricks System Tables for observability, lets us proactively catch data quality issues or SLA breaches before they impact business decisions—reducing incident resolution time by over 60%. Ultimately, Databricks isn't just improving our engineering efficiency; it's enabling us to innovate responsibly in a highly regulated environment, delivering trusted insights at scale while keeping auditors confident and stakeholders aligned.
    Syed F.

    Unified Data Engineering, Analytics, and ML on a Scalable Databricks Platform

    Reviewed on Mar 27, 2026
    Review provided by G2
    What do you like best about the product?
    What I like most about Databricks is how it brings data engineering, analytics, and machine learning together in one platform. It streamlines the entire data pipeline—from ingestion and transformation through to serving—so I don’t have to rely on multiple separate tools to get end-to-end workflows done.

    Its integration with Spark and Delta Lake is another big plus, making it both scalable and dependable when working with large datasets.
    What do you dislike about the product?
    One challenge with Databricks is cost management and visibility. Since compute is abstracted through clusters and jobs, it can sometimes be difficult to track and optimize costs without additional monitoring or governance in place.
    What problems is the product solving and how is that benefiting you?
    Solves the problem of fragmented data ecosystems, where data engineering, analytics, and machine learning are handled in separate tools.
    Janakiraman K.

    Databricks Brings Spark, Delta, and ML Together with Effortless Auto-Scaling

    Reviewed on Mar 27, 2026
    Review provided by G2
    What do you like best about the product?
    Databricks is hands down my favorite platform for data engineering because it brings everything together in one place Spark processing, Delta Lake, and ML tools all play nice without the usual headaches. The auto-scaling clusters save tons of time on big ETL jobs, like the SAP integrations I've done, letting me focus on logic instead of babysitting resources. Unity Catalog has been a game changer for governance in our lakehouse setups too.
    What do you dislike about the product?
    Costs can sneak up fast if you're not watching usage closely, especially with premium features on large pipelines. The notebooks are great for prototyping but get messy in production without strict discipline. Setup for advanced stuff like custom Unity Catalog policies sometimes feels overly complex for what it delivers.
    What problems is the product solving and how is that benefiting you?
    Databricks tackles key data engineering headaches like scaling massive Spark jobs, data quality issues, and siloed teams by providing a unified lakehouse platform with Delta Lake for ACID transactions and reliable pipelines. When I have a large number of files or tables to process like in supply chain ETL from SAP systems it shines with optimized Delta processing, serverless compute, and Photon engine, slashing run times from days to hours while cutting costs through auto-scaling. This benefits me directly by speeding up project delivery, reducing debugging time on failures, and enabling seamless collaboration with analysts on notebooks without tool switches.
    Shyam s.

    Genie Code and Inline Assistant Dramatically Boosted My Debugging Productivity

    Reviewed on Mar 27, 2026
    Review provided by G2
    What do you like best about the product?
    Genie code and the inline Assistant were the most helpful tools for me on my project. They helped me debug a 2k-line codebase and clearly explained why I wasn’t getting accurate data. It also provided a query to run in my source system (SQLMI). By running the discrepancy script in parallel on the source and target, I was able to debug the entire code much faster and improve my productivity. Overall, it cut my work time from about 8 hours down to around 1 hour.
    What do you dislike about the product?
    In Delta Sharing, there isn’t a catalog-level SELECT permission, and I sometimes think having that would be helpful. Also, when I use the Genie code inside a VM, it can make the website unresponsive at times. These are areas that could be improved.
    What problems is the product solving and how is that benefiting you?
    In one of our claims-processing migration projects, the client needed near real-time data availability for downstream applications. Previously, the architecture used Amazon Redshift as the data warehouse, with Jasper and Sisense consuming the data for reporting and analytics. However, that setup didn’t support real-time or near real-time streaming efficiently, which led to delays in data availability for downstream systems.

    After migrating the platform to Databricks, we were able to substantially improve the data pipeline architecture. We implemented streaming along with optimized ETL pipelines, reducing the data refresh cycle to about 30 minutes. We also created a dedicated view that retains data from the previous run, so downstream systems always have a consistent dataset available while the next pipeline execution is still in progress.

    Before, we struggled with delayed refresh cycles and a limited ability to meet near real-time data needs in our Redshift-based architecture. After moving to Databricks, we enabled faster ETL processing and improved near real-time data availability.

    As a result, we reduced ETL refresh time to roughly 30 minutes and enabled near real-time access for downstream tools like Jasper and Sisense. Reliability also improved because the stable view continues to serve the previous run’s data during pipeline updates. Finally, the overall architecture became simpler by consolidating processing and analytics capabilities within Databricks.

    Overall, Databricks helped us build a more scalable and efficient near real-time data processing platform, significantly improving the timeliness and reliability of analytics for the claims-processing workflow.
    View all reviews