Apache Airflow vs. Databricks Data Intelligence Platform

Overview
ProductRatingMost Used ByProduct SummaryStarting Price
Apache Airflow
Score 8.7 out of 10
N/A
Apache Airflow is an open source tool that can be used to programmatically author, schedule and monitor data pipelines using Python and SQL.N/A
Databricks Data Intelligence Platform
Score 8.8 out of 10
N/A
Databricks offers the Databricks Lakehouse Platform (formerly the Unified Analytics Platform), a data science platform and Apache Spark cluster manager. The Databricks Unified Data Service provides a platform for data pipelines, data lakes, and data platforms.
$0.07
Per DBU
Pricing
Apache AirflowDatabricks Data Intelligence Platform
Editions & Modules
No answers on this topic
Standard
$0.07
Per DBU
Premium
$0.10
Per DBU
Enterprise
$0.13
Per DBU
Offerings
Pricing Offerings
Apache AirflowDatabricks Data Intelligence Platform
Free Trial
NoNo
Free/Freemium Version
YesNo
Premium Consulting/Integration Services
NoNo
Entry-level Setup FeeNo setup feeNo setup fee
Additional Details
More Pricing Information
Community Pulse
Apache AirflowDatabricks Data Intelligence Platform
Considered Both Products
Apache Airflow
Chose Apache Airflow
Apache Airflow is suited for a much wider set of use cases compared to Databricks. You can run it anywhere, and there is also no vendor lock-in. With Airflow, we can utilize almost any compute engine. Same thing we want to do with Databricks. There might be some level of …
Databricks Data Intelligence Platform

No answer on this topic

Features
Apache AirflowDatabricks Data Intelligence Platform
Workload Automation
Comparison of Workload Automation features of Product A and Product B
Apache Airflow
8.7
12 Ratings
5% above category average
Databricks Data Intelligence Platform
-
Ratings
Multi-platform scheduling9.312 Ratings00 Ratings
Central monitoring8.912 Ratings00 Ratings
Logging8.612 Ratings00 Ratings
Alerts and notifications9.312 Ratings00 Ratings
Analysis and visualization6.812 Ratings00 Ratings
Application integration9.412 Ratings00 Ratings
Best Alternatives
Apache AirflowDatabricks Data Intelligence Platform
Small Businesses

No answers on this topic

No answers on this topic

Medium-sized Companies
ActiveBatch Workload Automation
ActiveBatch Workload Automation
Score 7.5 out of 10
Snowflake
Snowflake
Score 8.7 out of 10
Enterprises
Control-M
Control-M
Score 9.3 out of 10
Snowflake
Snowflake
Score 8.7 out of 10
All AlternativesView all alternativesView all alternatives
User Ratings
Apache AirflowDatabricks Data Intelligence Platform
Likelihood to Recommend
8.8
(12 ratings)
10.0
(18 ratings)
Usability
8.2
(3 ratings)
10.0
(4 ratings)
Support Rating
-
(0 ratings)
8.7
(2 ratings)
Contract Terms and Pricing Model
-
(0 ratings)
8.0
(1 ratings)
Professional Services
-
(0 ratings)
10.0
(1 ratings)
User Testimonials
Apache AirflowDatabricks Data Intelligence Platform
Likelihood to Recommend
Apache
Airflow is well-suited for data engineering pipelines, creating scheduled workflows, and working with various data sources. You can implement almost any kind of DAG for any use case using the different operators or enforce your operator using the Python operator with ease. The MLOps feature of Airflow can be enhanced to match MLFlow-like features, making Airflow the go-to solution for all workloads, from data science to data engineering.
Read full review
Databricks
Medium to Large data throughput shops will benefit the most from Databricks Spark processing. Smaller use cases may find the barrier to entry a bit too high for casual use cases. Some of the overhead to kicking off a Spark compute job can actually lead to your workloads taking longer, but past a certain point the performance returns cannot be beat.
Read full review
Pros
Apache
  • Apache Airflow is one of the best Orchestration platforms and a go-to scheduler for teams building a data platform or pipelines.
  • Apache Airflow supports multiple operators, such as the Databricks, Spark, and Python operators. All of these provide us with functionality to implement any business logic.
  • Apache Airflow is highly scalable, and we can run a large number of DAGs with ease. It provided HA and replication for workers. Maintaining airflow deployments is very easy, even for smaller teams, and we also get lots of metrics for observability.
Read full review
Databricks
  • Process raw data in One Lake (S3) env to relational tables and views
  • Share notebooks with our business analysts so that they can use the queries and generate value out of the data
  • Try out PySpark and Spark SQL queries on raw data before using them in our Spark jobs
  • Modern day ETL operations made easy using Databricks. Provide access mechanism for different set of customers
Read full review
Cons
Apache
  • UI/Dashboard can be updated to be customisable, and jobs summary in groups of errors/failures/success, instead of each job, so that a summary of errors can be used as a starting point for reviewing them.
  • Navigation - It's a bit dated. Could do with more modern web navigation UX. i.e. sidebars navigation instead of browser back/forward.
  • Again core functional reorg in terms of UX. Navigation can be improved for core functions as well, instead of discovery.
Read full review
Databricks
  • Sometimes, when multiple jobs depend on each other in different environments, it is not always easy to see the full workflow in one place.
  • It is sometimes difficult to determine which job or cluster contributes more to the overall cost.
  • For beginners, cluster configuration may be a little difficult. So more recommendation in the platform can help.
Read full review
Usability
Apache
For its capability to connect with multicloud environments. Access Control management is something that we don't get in all the schedulers and orchestrators. But although it provides so many flexibility and options to due to python , some level of knowledge of python is needed to be able to build workflows.
Read full review
Databricks
Because it is an amazing platform for designing experiments and delivering a deep dive analysis that requires execution of highly complex queries, as well as it allows to share the information and insights across the company with their shared workspaces, while keeping it secured.

in terms of graph generation and interaction it could improve their UI and UX
Read full review
Support Rating
Apache
No answers on this topic
Databricks
One of the best customer and technology support that I have ever experienced in my career. You pay for what you get and you get the Rolls Royce. It reminds me of the customer support of SAS in the 2000s when the tools were reaching some limits and their engineer wanted to know more about what we were doing, long before "data science" was even a name. Databricks truly embraces the partnership with their customer and help them on any given challenge.
Read full review
Alternatives Considered
Apache
Multiple DAGs can be orchestrated simultaneously at varying times, and runs can be reproduced or replicated with relative ease. Overall, utilizing Apache Airflow is easier to use than other solutions now on the market. It is simple to integrate in Apache Airflow, and the workflow can be monitored and scheduling can be done quickly using Apache Airflow. We advocate using this tool for automating the data pipeline or process.
Read full review
Databricks
The most important differentiating factor for Databricks Lakehouse Platform from these other platforms is support for ACID transactions and the time travel feature. Also, native integration with managed MLflow is a plus. EMR, Cloudera, and Hortonworks are not as optimized when it comes to Spark Job Execution. Other platforms need to be self-managed, which is another huge hassle.
Read full review
Return on Investment
Apache
  • Impact Depends on number of workflows. If there are lot of workflows then it has a better usecase as the implementation is justified as it needs resources , dedicated VMs, Database that has a cost
  • Donot use it if you have very less usecases
Read full review
Databricks
  • The ability to spin up a BIG Data platform with little infrastructure overhead allows us to focus on business value not admin
  • DB has the ability to terminate/time out instances which helps manage cost.
  • The ability to quickly access typical hard to build data scenarios easily is a strength.
Read full review
ScreenShots