Sign in Agent Mode
Categories
Become a Channel Partner Sell in AWS Marketplace Amazon Web Services Home Help

dbt Platform

dbt Labs

Reviews from AWS customer

6 AWS reviews

External reviews

202 reviews
from and

External reviews are not included in the AWS star rating for the product.


    nayan S.

Best-in-Class SQL Transformations in dbt

  • April 08, 2026
  • Review provided by G2

What do you like best about the product?
SQL Transformation is the best thing which dbt have.
What do you dislike about the product?
Nothing as of now. but pricing seems high for model run
What problems is the product solving and how is that benefiting you?
ETL and data engineering portion. We develop metrics and fact tables in dbt transformation


    Harshwardhan Gullapalli

Data pipelines have improved financial accuracy and now build transparent audit-ready reports

  • April 07, 2026
  • Review provided by PeerSpot

What is our primary use case?

My main use case for dbt was transforming messy, extracted financial data into clean, production-ready datasets. Let me give you a concrete example. We would extract trial balance data from PDFs using OCR and then feed it through our GenFin mapping workflow. Before that data reached our GPT-4 model for classification, we used dbt to normalize account codes, handle currency conversions, work with UAE and multi-currency scenarios for UAE compliance, and create aggregations by cost center and account type. So dbt's job was essentially to take a raw extraction output, validate it against our schema, handle any missing values or duplicate entries, and then materialize clean tables that our LLMs could reliably work with.

There was one more important place beyond just transformation. dbt gave us visibility into data quality throughout the pipeline. We built tests into our models, including checking for null values in critical fields such as account numbers, ensuring amounts were numeric and within expected range, and validating that our transformed data matched the source record counts. This was critical because if there was a gap between extraction and transformation, we would catch it before it hit the LLM. The dependency graph in dbt was invaluable when we had issues downstream in our mapping or disclosure notes workflows, as we could trace back through the DAG.

The most concrete outcome was a significant reduction in data errors reaching our downstream AI models. Before dbt, we were catching bad extractions only after the LLM had already processed them, which meant manual rework. After implementing dbt's testing layer, we caught roughly 70% of those issues at the transformation stage itself, before they ever touched the model. Processing speed also improved because dbt's incremental models only processed the delta into new records. Our nightly reconciliation runs for UAE corporate tax workflows dropped from around 40 minutes to under 10 minutes. The less obvious win was team confidence. Our chartered accountant clients started trusting the outputs more because we could show them the full transformation chain.

The shift in client trust was really tangible. Before dbt, when we delivered a mapped disclosure notes document, clients would ask us to walk them through how we arrived at the specific numbers. We would have to manually trace back through extraction logs, which was time-consuming and sometimes incomplete. After dbt, we could literally show them the DAG, the dependency graph, and say, here is your source PDF, here is where we extracted the account code, here is the normalization rule we applied, and here is the aggregation, and here is your final mapped output. It is all documented and testable. That transparency eliminated a huge amount of back-and-forth. One specific example involved a client who was questioning why a particular cost center total did not match their internal records. With dbt's lineage, we traced it in minutes, found a rounding rule that needed adjustment in one model, fixed it, and reran the pipeline. The client saw the complete audit trail. That kind of visibility is what builds trust in a financial team.

One thing I would add positively was how well dbt's incremental models handled our late-arriving data scenarios. In financial processing, you often get corrections or amendments to documents days later, and incremental models let us efficiently merge those without full refreshes. That was really valuable for our UAE compliance workflows where reconciliation amendments came in batches.

What is most valuable?

dbt's incremental model let us handle late-arriving data efficiently without reprocessing everything. That clean output then fed directly into our GPT-4 structured output parser for financial mapping. The workflow fit really well because dbt gave us clear data lineage. We could trace any number of our final mapping output back to the source extract. That is critical in financial-based reporting, where auditors need to see the full transformation chain. For us, the standout features were the testing framework and the dependency graph. The tests, especially custom tests for financial data like validating that debits equal credits, caught a lot of our data quality issues early.

dbt's incremental model was key here. Rather than representing the entire dataset every run, dbt only touched the new or changed records, which kept our pipeline efficient as document volume grows. That said, dbt's scalability is fundamentally tied to your underlying database. In our setup, we managed performance by being deliberate about materialization strategies, using views for lightweight transformations and tables for incremental models for heavy aggregations.

What needs improvement?

As for something I wish we had, dbt's native support for Python transformations came later, and we did some complex financial classification calculations that felt clunky in pure SQL. We ended up writing Python in our n8n workflows and then fed the results back into dbt, which created a bit of a split-brain situation. If we would have had dbt Python models earlier, we could have kept that logic unified.

Managing multiple reporting standards was our biggest operational pain point with dbt. We were running UAE corporate tax compliance and IFRS disclosure workflows simultaneously for different clients, and dbt does not have a native concept of multi-tenant or multi-standard project organization. Everything lives in one flat structure, so we had to build more conventions: separate schema folders for IFRS models versus UACT models, custom macros to tag models by compliance regime, and environment variables to control which set of transformations run for which client.

For how long have I used the solution?

I have been working with dbt for approximately one and a half years.

What do I think about the stability of the solution?

dbt has been very stable.

What do I think about the scalability of the solution?

For our use case, dbt scaled well. We were processing large volumes of financial documents, hundreds of trial balances, balance sheets, and invoice sets, and dbt handled the transformation layer without issues.

How are customer service and support?

dbt's customer support varies depending on which tier you are using. We ran dbt Core, which is open-source, so there is no direct vendor support. The community support is actually quite good, and the dbt Slack community is also very supportive. We found answers to most of our questions through the docs and community. If we would have been on dbt Cloud, there would be dedicated support, but the pricing did not work for our team size. For a bootstrapped FinTech team in India working on UAE compliance workflows, the cost-benefit was not there.

Which solution did I use previously and why did I switch?

Before dbt, we were handling data transformation in a more ad-hoc way. We had Python scripts and custom SQL queries scattered across our n8n workflows, our automation platform. When we extracted financial data from PDFs using Tesseract OCR and LLM extraction, we would clean and transform it using Python code nodes within n8n. There was no clear lineage. If the number was wrong downstream, tracing back through n8n workflow logic was painful.

Which other solutions did I evaluate?

We did evaluate alternatives before committing to dbt. We looked at a few options. Apache Airflow was one. It is powerful for orchestration but felt like overkill for our use case and would have required significant DevOps overhead to manage. We also considered Talend and Informatica, but those were enterprise tools with enterprise pricing, and we were a lean team at Radiant Services.

What other advice do I have?

The most concrete outcome was a significant reduction in data errors reaching our downstream AI models. Before dbt, we were catching bad extractions only after the LLM had already processed them, which meant manual rework. After implementing dbt's testing layer, we caught roughly 70% of those issues at the transformation stage itself, before they ever touched the model. Processing speed also improved because dbt's incremental models only processed the delta into new records. Our nightly reconciliation runs for UAE corporate tax workflows dropped from around 40 minutes to under 10 minutes. I would rate this product an 8 out of 10.


    JohamAlvarez-Montoya

Centralized data transformations have improved workflows but integrations still need expansion

  • April 06, 2026
  • Review from a verified AWS customer

What is our primary use case?

I am currently working with dbt and use dbt's modular SQL models.

What is most valuable?

It is very convenient because at the end, I have the opportunity to orchestrate all my transformations in just one single place, rather than having them spread out. I can use SQL, which is very convenient for the sizes of data that I usually use in the day-to-day. Of course, some other deployments might require Spark, and in this case, it is not a good idea to use SQL plus dbt, but for most cases, it is very convenient. Because it is easy to set up, and also due to the cost, I have all the transformations in one single place with a very convenient tool.

It is very convenient; I can set up the expectations really easily, all integrated in the tool that I usually use for my data transformation, so it is very convenient.

What needs improvement?

With AI, everything is advancing so fast, so I would say that the most important thing is to try to integrate with more platforms. As of now, dbt has a strong integration with AWS and with Snowflake, but I have not seen other integrations. Having more partners and having more visibility on the things that can be done is important, because I see that competitors are doing great in that aspect. For example, Databricks and Snowflake itself are doing that, so more visibility, more partnerships, and more integrations would be helpful.

For how long have I used the solution?

I have been using dbt for five years as of now.

How are customer service and support?

I may not have enough information to respond about the technical support of dbt because I have not reached out to dbt support at any time, probably because I have never needed it. In the three or four organizations that I have been with using dbt, I have not contacted the support service directly, so it may be a good sign.

How was the initial setup?

I will say the deployment for dbt is very straightforward; once you have the experience, it is quite straightforward. For example, you can set up a Docker or Docker Compose and run it, or you can use directly the on-cloud version.

I would say that the deployment for dbt requires about a day; a day could work, and with a day, you can set up and deploy dbt.

What about the implementation team?

I find that with one person, it is enough to complete the deployment; of course, it can vary depending on the complexity of the project, but I would say that one person working one day is enough.

What's my experience with pricing, setup cost, and licensing?

I mentioned the cost as one of the advantages, specifically the license cost.

Which other solutions did I evaluate?

I think the pricing is very convenient; one of the barriers is that for example, some of the companies that I have been with, dbt is a normal solution or neutral in terms of cost. It is not cheap or more expensive, but the problem is that companies are really locked in with existing vendors, and those vendors offer alternatives that might be less expensive given that they have other products. dbt only offers data transformation services, making it hard to compete with vendors that have all the packages included, such as cloud and processing services. If you compare the cost of those packages with dbt alone, it is more expensive to use dbt alone.

What other advice do I have?

I am using a private cloud. I have used both on-prem and cloud versions of the product, but mainly the managed version, the on-cloud version. That is very convenient; of course with AI, that is being commoditized a little bit. But I like it; I used it more before. Now with AI, it is even easier to do documentation, but before AI, it was really convenient to generate documentation with that tool. My overall review rating for dbt is 7 out of 10.


    Information Technology and Services

Versatile Data Transformation Tool

  • March 27, 2026
  • Review provided by G2

What do you like best about the product?
I primarily value how dbt shifts data transformation into a software engineering workflow. By materializing code into tables and views automatically, it lets our team focus on the transformation logic rather than DDL boilerplate. The model selection syntax is incredibly efficient for running specific segments of our DAG without wasting warehouse resources.

Macros and Jinja integration have also been game-changers for modularizing our logic and reducing code repetition. I find the YAML-based unit testing to be a robust way to ensure data integrity before it reaches our BI layer. Between the two, I prefer dbt Cloud over Core because the IDE provides immediate visibility into query results and schema changes, which speeds up our development cycle.

I use it every workday. Customer support was quick and responsive when I ran into issues during the initial setup of dbt with Snowflake authentication. Implementation was also straightforward when connecting it to Snowflake, and once the connection was established, I didn’t have any ongoing issues aside from needing to reauthenticate every day.
What do you dislike about the product?
The most significant friction point is the authentication lifecycle with Snowflake. The session tokens expire frequently (often every few hours), forcing a manual re-authentication process that disrupts the flow of development.

Additionally, there is a noticeable feature gap between the versions. dbt Core lacks the native, instant result-set previewing that makes dbt Cloud so productive. Bringing a similar "live preview" or integrated results pane to the Core CLI experience would make it a much more viable option for local development.
What problems is the product solving and how is that benefiting you?
We use dbt to manage the entire T (Transform) layer of our ELT process within Snowflake. Before dbt, our transformations were often opaque and difficult to test. Now, we have:

Dryer Code: Macros help us maintain a "Don't Repeat Yourself" philosophy across hundreds of models.

Improved Data Quality: By implementing automated tests on primary keys and relationship constraints, we catch upstream data issues before they impact stakeholders.

Faster Onboarding: The combination of dbt Cloud’s UI and the structured project documentation makes it much easier for new analysts to understand our data lineage.


    Ahmed Shaaban

Data teams have streamlined code-driven pipelines and now collaborate faster on shared models

  • March 02, 2026
  • Review provided by PeerSpot

What is our primary use case?

I am working with one of our enterprise customers, managing their newly established cloud warehouse. They are using Snowflake and we are using dbt to manage all the transformation and views and tables in Snowflake. I am not currently working with Cribl, but I used to work with it for almost three years. Currently, I am working with dbt and Snowflake stack.

What is most valuable?

dbt is a tool that is basically SQL and a little bit of Python, which is somewhat low entry-level, so many of the engineers can use it as well as the analysts. Multiple teams from the business side can use it as well if we allow them. Performance-wise, it mainly depends on the platform that hosts it, whether it is Snowflake or Databricks or BigQuery. There is not much complication. Of course, there are the benefits of having code, so you have a software development lifecycle; you can use version control, testing, and documentation.

I would say the best feature or the most desirable feature for dbt is the ability to write everything in code. It is treating data the same way that Ansible did or Terraform did for infrastructure as code. Now you can code the pipeline instead of using SSIS and Apache NiFi and even Informatica PowerCenter. All of these tools are GUI-based tools. They have a low entry barrier, but you cannot really integrate them in a CI/CD pipeline, for example. For dbt, we can create those. More recently with the advances in AI, LLMs, and code assistant agents, we can hugely leverage those in dbt because I can simply ask the agent to write the code or write the model. However, you cannot really ask them to draw any SSIS package or an Apache NiFi flow, for example.

I think that dbt helps us quite a bit because it exposes a little bit of the functionality of Snowflake directly to us. We can use it with ease because we have some experience with Snowflake and we know what controls to adjust. Because we are a team of multiple individuals, we need to collaborate. Without version control, you have to manage the whole codebase one feature at a time, but what we do is we can use branching and different feature branches. Each one of us is working on their own feature branch. We collaborate, we merge our changes, and we can roll back in case we introduced some bugs. I would say the version control feature is a huge bonus or a huge plus.

What needs improvement?

We are still experimenting with testing, but not that much. We are not using some features yet. We are trying to introduce them because we are coming from a background of SSIS. The team used to work with SSIS, Microsoft SQL Server Integration Services. We are still adapting one feature at a time. Currently, we are working with the SQL modules and with the Jinja templating. We are experimenting with testing, but I would say towards the end of this year, we are planning to explore more of the documentation and the data lineage options as well.

I would say the benefits are coming from GUI-based tools like SSIS. We have more control on the codebase. We can create something of a system where we can use macros and templating, speeding up the development cycle. We are now trying to introduce a little testing, and also we are using some sort of a CI/CD cycle, so continuous integration and continuous deployment. I do not believe that these kinds of features are that common as a package as a whole package. dbt excels in that area.

I used to have a couple of notes about the performance, but lately I have discovered something called dbt Fusion, which, according to dbt Labs, they proclaim is much faster during the parsing of dbt models. However, I would love to see even more of an out-of-the-box solution regarding the testing. They are treating the testing in a good way so far, but I would love to see even more improvement because the whole data testing field is not very mature. It is not the same as software testing; for example, you have test suites, test tools, and profilers, but for data testing, it is not yet that advanced. I would love for dbt to take the lead on that.

For how long have I used the solution?

I have been using dbt since September 2024, so almost a year and a half.

What do I think about the stability of the solution?

I think that one of the issues with dbt is upgrading to later versions because we have some functionalities that we have designed that overcome default behaviors for dbt. Every upgrade is a little bit of a risk for us because we do not know if the workarounds that we developed will be available for the next version. However, in terms of stability, we have had no issue.

What do I think about the scalability of the solution?

I would say we have not experienced scalability issues so far. I am not aware of the scalability, but we are managing it on a very large scale. The bottlenecks that we have are not coming from dbt; they are coming from Snowflake. Once we scale up Snowflake, dbt has no issue whatsoever.

How are customer service and support?

Besides the issue with the upgrades and the default behaviors for the macros that we overwrote, we did not really need to reach out to dbt support.

Which solution did I use previously and why did I switch?

The team used to work with SSIS before I joined. They used to work with it, I believe, two or three years ago, but since I joined a year and a half ago, they switched to dbt. They switched to dbt just before I joined. I have also worked for some time with Apache NiFi as well.

How was the initial setup?

I am not aware of the initial setup because they set up everything before I joined, but they are using dbt Cloud. I do not think there are many difficulties or any hurdles to overcome during setup. You simply link your dbt Cloud account with the Snowflake account and that is it.

What other advice do I have?

In terms of metrics, I do not have exact metrics, but I get a sense of the speed of opening and closing data requests. I am not that familiar with the Scrum Master of our squad, but I believe our burn chart or something like that, which is an agile metric that measures the finished user stories, is the only sense or only kind of metrics that we have at the moment. However, you do get a sense of accomplishment and the speed of delivering value.

I would say just the testing is something to focus on. dbt Fusion is something I am not completely aware of, but I need to try it because I think it is a great feature, especially because we are dealing with multiple models. For our use case, we are dealing with 50 plus, almost 100 models. Many models are running at the same time. If you add up all the compile time and parsing time, it can add up to quite a bit. dbt Fusion promises that the parsing is much faster in one-tenth of the time, I believe.

I would say you really need to take care of your model and your data model because dbt gives you some freedom. If you do not really know what you are going to do, you can really mess things up. So you need to take care of the model, design your layers, define the responsibilities of each layer, define the criteria of each data layer, define the tests, and that is it.

I would rate this product an eight out of ten.


    reviewer2803965

Building medallion architecture has improved collaboration and streamlined data transformation

  • February 20, 2026
  • Review provided by PeerSpot

What is our primary use case?

I have primarily used dbt for data transformation. My team and I work with source systems from various clients, some using Snowflake, others using SQL Server, and some using their own legacy source systems. We extract and load all the data into our data lake on Azure. From there, we use dbt to transform that data. In technical terms, we have extracted and loaded the data into our data lake, and from there we are doing the transformation with dbt. We are creating different layers such as silver, bronze, and gold to build a medallion architecture with dbt.

What is most valuable?

From a developer point of view, I find the ease of development and the code to be the most useful capabilities of dbt. I use VS Code to run the dbt models, and since the end user is only concerned about their output and reports, the ease of development and the fact that it is free are significant advantages.

I assess the impact of dbt's version control system on team collaboration as great. I have used it extensively, especially when we had situations where the code broke, as we were able to roll back to earlier versions thanks to version control.

I find dbt's documentation site generator to be quite crisp and straightforward. It helps with project transparency and onboarding new team members because the documentation is excellent for addressing issues we face. I learned dbt concepts primarily using their website and their tutorials, which helped me significantly compared to other platforms such as YouTube and Udemy. The course content that dbt provides is free and excellent for anyone starting out.

What needs improvement?

dbt seems quite adequate currently, but if I needed to name a few areas for improvement, I would mention the migration of code to Git and GitHub, which sometimes fails and can be confusing for developers during handover. There are some glitches in the connection, but I am unsure if that is an issue from the dbt side or something else, so I cannot comment definitively.

For how long have I used the solution?

I have been working with dbt for approximately seven to eight months.

What do I think about the stability of the solution?

Regarding stability and reliability, I see the tool as quite good. In terms of use case, market presence, demand, learning, and performance, I believe dbt will continue to be in the market.

I would rate dbt's stability and reliability at a minimum of eight out of ten based on the limited experience I have. Comparing it to tools I have seen in the past, such as Informatica and Alteryx, dbt can easily match up to that rating, specifically for stability.

What do I think about the scalability of the solution?

I am not very certain how scalable dbt is from my experience, as I have had limited scope to work with it. I have not analyzed it deeply. I started as a developer and began with their free plan before moving to a paid plan, which was quite affordable at around one hundred or one hundred fifty dollars per month. We are currently focusing on report development and multi-tenant deployment, so we might consider scaling in the future.

How are customer service and support?

Earlier, we used technical support for dbt, but that was only valid for a month or fifteen days. We later moved to the paid version because I was working on the proof of concept of Qlik Sense and other tools, and we finalized dbt as well. Initially, I explored dbt for free for about ten days without trialing any further support.

So far, we have not interacted much with technical support because we usually get help from the community on their website. If you type your question, you will likely find that someone has already asked it, so we do not need to contact their support directly.

Which solution did I use previously and why did I switch?

I have not yet utilized dbt's testing framework.

How was the initial setup?

My experience with the initial setup and deployment of dbt involved using VS Code. I am not very confident here because I received some help from another data engineer to set it up on my machine. However, I have used VS Code in the past, and with some libraries, it was successfully done, but I am not entirely certain about every detail.

What about the implementation team?

I am both a customer and consultant for dbt because my company has bought the license, and as an experienced person, I work on a product for my company.

What's my experience with pricing, setup cost, and licensing?

The course content that dbt provides is free and excellent for anyone starting out.

Which other solutions did I evaluate?

I have not practically used a different solution for the same use cases, but I have been part of teams that used tools such as Alteryx, Informatica, and Talend, even though I did not work with them hands-on.

What other advice do I have?

From a developer point of view, I find the ease of development and the code to be the most useful capabilities of dbt. I use VS Code to run the dbt models, and since the end user is only concerned about their output and reports, the ease of development and the fact that it is free are significant advantages.

I assess the impact of dbt's version control system on team collaboration as great. I have used it extensively, especially when we had situations where the code broke, as we were able to roll back to earlier versions thanks to version control.

I find dbt's documentation site generator to be quite crisp and straightforward. It helps with project transparency and onboarding new team members because the documentation is excellent for addressing issues we face. I learned dbt concepts primarily using their website and their tutorials, which helped me significantly compared to other platforms such as YouTube and Udemy. The course content that dbt provides is free and excellent for anyone starting out.

dbt seems quite adequate currently, but if I needed to name a few areas for improvement, I would mention the migration of code to Git and GitHub, which sometimes fails and can be confusing for developers during handover. There are some glitches in the connection, but I am unsure if that is an issue from the dbt side or something else, so I cannot comment definitively.

I would rate my overall experience with dbt at eight out of ten.


    Syed A.

A Developer Friendly Transformation Tool

  • February 03, 2026
  • Review provided by G2

What do you like best about the product?
I like best about dbt is how it brings a clean, developer‑friendly structure to analytics work. It makes modeling and transforming data feel organized and predictable, thanks to its simple SQL‑first approach and clear project layout. I also really appreciate how dbt encourages good engineering practices such as version control, testing, documentation. So the entire workflow becomes more reliable and collaborative.
What do you dislike about the product?
I dislike about dbt is that some parts of the workflow can feel a bit inflexible, especially when you're trying to customize how tests or models behave in more complex projects. It also relies heavily on command‑line and configuration files, which can become demanding as the project grows. On top of that, dbt doesn’t handle ingestion or real‑time needs, so user often need additional tools to complete the pipeline, which makes the setup feel less seamless.
What problems is the product solving and how is that benefiting you?
dbt solves the problem of scattered, inconsistent transformation logic by giving user a clean, structured way to manage SQL models, tests, and documentation in one place. I no longer needs to deal with random queries or unclear business rules, everything becomes version‑controlled and easy to trace. Which helps me in my workflows to become productive.


    Scott J.

Reliable transformation practices at scale

  • January 27, 2026
  • Review provided by G2

What do you like best about the product?
One thing that I find impressive about dbt is that it promotes discipline in writing of transformations. It transformed my approach towards the way I deal with my work, as I now think twice before imposing changes. I use it on a regular basis, and it has enhanced teamwork since logic has less difficulty in reviewing and discussion. This has saved time on quick fixes and has assisted us in developing more confidence on outputs that may be shared.
What do you dislike about the product?
What I do not like about dbt is that there is a huge effect of little errors in the models. Some of them may break down under the pressure of having a few downstream pieces broken when there is a slight change. It is time consuming and can even bring several individuals into the same problem when it comes to debugging those chains. In my case, this retards progress and results in context switching which can be annoying when time lines are near.
What problems is the product solving and how is that benefiting you?
Dbt eliminates the issue of vague ownership and reasoning. It provides organization where responsibilities are clearly seen which enhances cooperation. In my case, it implies a reduced number of handoff problems and a streamlined collaboration. Co-workers become bolder in changes and tasks are less responsive on a daily basis. It has simplified our working process and made it more predictable in general.


    Shubham-Agarwal

Incremental data models have cut full refresh time and support trusted executive reporting

  • January 22, 2026
  • Review from a verified AWS customer

What is our primary use case?

I am currently working with dbt and Snowflake together. We use dbt for data transformation purposes. We obtain the data and store the raw data directly into Snowflake, then perform all transformations using dbt to prepare the data for reporting purposes.

We use dbt's modular SQL models. In dbt, we do not use full refresh or full data refresh. We have incremental strategies in place that only compute or transform incremental data, which operates in a CDC architecture. This approach is very fast, and we use it on a daily basis. We have scheduled all our dbt models using Airflow to run according to the scheduled time.

We use dbt's testing framework and the inbuilt functionality of dbt testing. For example, we use dbt's built-in tests to identify not null values and determine how many not null columns and values exist in each column. Beyond the built-in functionality, we have written custom SQL scripts to create external test cases on our models.

We ensure that incorrect or incomplete data does not go into the reporting layer because it can impact the business. We always perform dbt tests on our landing or raw data to ensure the correctness and completeness of the data before loading it into the final reporting layer. These reports are used by higher management, so we ensure that incorrect data is not published into the reporting layer for the Power BI reports.

We use dbt's documentation site generator. In dbt, we have YML file functionality, which can be used for creating documentation for each model. Whenever we make modifications to a model, we always update the YML file so we can track the history of how frequently we change the model. When new team members join, they can refer to this documentation to understand the data lineage and the data transformation strategy of the project.

What is most valuable?

dbt is very fast compared to the traditional tools. Previously, I worked on SSIS, which is provided by Microsoft, and data transformation took a considerable amount of time when dealing with large amounts of data. Since dbt works on the ELT architecture rather than the ETL architecture, it is much faster than traditional data transformation tools.

Previously, we were using SSIS packages, which were very slow. Recently we migrated all our SSIS packages to dbt models. After the migration, we moved the data from SQL Server to Snowflake. Previously, our data pipeline took around two days to load complete data when performing a full refresh. Since we migrated from SSIS to dbt model architecture, it takes around four hours only to complete a full refresh. The client is now happy because our downtime was drastically reduced when we perform a complete refresh of the data.

What needs improvement?

I am not very familiar with dbt's version control system.

I cannot identify any improvements in dbt because I am still exploring more functionality. I have been working with dbt for only three years, so I am exploring more functionalities and cannot see any limitations or improvement areas at this time.

In the past, I used the seed functionality, which is used to load raw files, individual files, or static files into the database. Going forward, if dbt can improve or implement more features on the seed side, that would be beneficial, especially when we have large files available that take time to load the data into Snowflake database.

For how long have I used the solution?

I have been working with dbt for the last three years.

What do I think about the stability of the solution?

I have not experienced any crashes, performance issues, or anything regarding stability and reliability.

What do I think about the scalability of the solution?

I find dbt very scalable.

How are customer service and support?

The dbt support team is very responsive. Whenever we have any issues on the dbt side, we always reach out to them. We did not face any challenges in the initial setup. I would rate the technical support a nine out of ten.

How would you rate customer service and support?

Positive

Which solution did I use previously and why did I switch?

Previously, we were using SSIS packages, which were very slow. Recently we migrated all our SSIS packages to dbt models. After the migration, we moved the data from SQL Server to Snowflake. Previously, our data pipeline took around two days to load complete data when performing a full refresh. Since we migrated from SSIS to dbt model architecture, it takes around four hours only to complete a full refresh. The client is now happy because our downtime was drastically reduced when we perform a complete refresh of the data.

The main reason we decided to switch to dbt is performance. As mentioned earlier, every quarter we perform a full refresh, and that refresh took considerable time on SQL Server. Since we had to migrate because our data is very large and growing daily, we adopted dbt because Snowflake is very fast. In Snowflake, the storage layer and the computation layer are separate, which is not present in the SQL Server traditional database. That is why we moved from SQL Server to Snowflake and from SSIS to dbt.

How was the initial setup?

We evaluated Databricks as well, but ultimately the client wanted to adopt Snowflake and dbt technologies only.

What about the implementation team?

We took help from Snowflake directly, the Snowflake company, for the Snowflake side. The dbt side is maintained or set up by our infrastructure team.

What was our ROI?

Since we migrated from SSIS to dbt model architecture, it takes around four hours only to complete a full refresh. The client is now happy because our downtime was drastically reduced when we perform a complete refresh of the data.

What's my experience with pricing, setup cost, and licensing?

The pricing, setup cost, and licensing cost are managed by our infrastructure teams. As data engineers, we are not familiar with these details.

I need to check with my infrastructure team on whether we purchased dbt through the AWS Marketplace or directly from the local vendor.

Which other solutions did I evaluate?

Since dbt has a license cost, if a company is small and does not have much budget, they can explore other tools because there are other tools that provide the same functionality at a lower cost. If an organization is small, they can explore other products as well.

What other advice do I have?

I am currently working with Power BI, Tableau, Python, Databricks, Snowflake, and PySpark in the current project. I would rate my overall experience with dbt a nine out of ten.

Which deployment model are you using for this solution?

Public Cloud

If public cloud, private cloud, or hybrid cloud, which cloud provider do you use?


    Joseph S.

Speedy but however it is quite pricey and resource hungry

  • January 20, 2026
  • Review provided by G2

What do you like best about the product?
The way it handles large amounts of data, as well as how it integrates into AWS (S3/Glue) is great. This allows me to avoid building custom pipelines which would have been very time consuming and caused additional headaches and due to its columnar database design, all of my complex query requests are processed in a timely manner which means I do not fall asleep while waiting for results.
What do you dislike about the product?
Vacuuming Tables… seriously, I have to manually vacuum and analyze tables to keep this thing running smoothly? It looks like 2005. Managing the clusters and nodes is also a pain – it’s not true serverless. If you’re not paying close attention to the costs, they will jump up way too high.
What problems is the product solving and how is that benefiting you?
This has allowed us to move away from a pandas-based reporting solution, which crashed consistently. We now can process billions of records from our retail business and have a working dashboard. Its architecture provides separate storage and compute, allowing us to scale our compute resources as much as needed based on the demand for reports by management. Most importantly, it has reduced the amount of yelling from our data team regarding slow query performance.