Apache Airflow vs. Databricks Data Intelligence Platform

Overview
ProductRatingMost Used ByProduct SummaryStarting Price
Apache Airflow
Score 8.7 out of 10
N/A
Apache Airflow is an open source tool that can be used to programmatically author, schedule and monitor data pipelines using Python and SQL. Created at Airbnb as an open-source project in 2014, Airflow was brought into the Apache Software Foundation’s Incubator Program 2016 and announced as Top-Level Apache Project in 2019. It is used as a data orchestration solution, with over 140 integrations and community support.N/A
Databricks Data Intelligence Platform
Score 8.7 out of 10
N/A
Databricks in San Francisco offers the Databricks Lakehouse Platform (formerly the Unified Analytics Platform), a data science platform and Apache Spark cluster manager. The Databricks Unified Data Service aims to provide a reliable and scalable platform for data pipelines, data lakes, and data platforms. Users can manage full data journey, to ingest, process, store, and expose data throughout an organization. Its Data Science Workspace is a collaborative environment for practitioners to run…
$0.07
Per DBU
Pricing
Apache AirflowDatabricks Data Intelligence Platform
Editions & Modules
No answers on this topic
Standard
$0.07
Per DBU
Premium
$0.10
Per DBU
Enterprise
$0.13
Per DBU
Offerings
Pricing Offerings
Apache AirflowDatabricks Data Intelligence Platform
Free Trial
NoNo
Free/Freemium Version
YesNo
Premium Consulting/Integration Services
NoNo
Entry-level Setup FeeNo setup feeNo setup fee
Additional Details
More Pricing Information
Community Pulse
Apache AirflowDatabricks Data Intelligence Platform
Top Pros
Top Cons
Features
Apache AirflowDatabricks Data Intelligence Platform
Workload Automation
Comparison of Workload Automation features of Product A and Product B
Apache Airflow
9.8
10 Ratings
17% above category average
Databricks Data Intelligence Platform
-
Ratings
Multi-platform scheduling9.910 Ratings00 Ratings
Central monitoring9.910 Ratings00 Ratings
Logging9.910 Ratings00 Ratings
Alerts and notifications9.910 Ratings00 Ratings
Analysis and visualization9.910 Ratings00 Ratings
Application integration9.010 Ratings00 Ratings
Best Alternatives
Apache AirflowDatabricks Data Intelligence Platform
Small Businesses

No answers on this topic

No answers on this topic

Medium-sized Companies
ActiveBatch Workload Automation
ActiveBatch Workload Automation
Score 8.0 out of 10
Amazon Athena
Amazon Athena
Score 9.0 out of 10
Enterprises
Redwood RunMyJobs
Redwood RunMyJobs
Score 9.3 out of 10
Amazon Athena
Amazon Athena
Score 9.0 out of 10
All AlternativesView all alternativesView all alternatives
User Ratings
Apache AirflowDatabricks Data Intelligence Platform
Likelihood to Recommend
9.0
(10 ratings)
10.0
(18 ratings)
Usability
10.0
(1 ratings)
10.0
(4 ratings)
Support Rating
-
(0 ratings)
8.7
(2 ratings)
Contract Terms and Pricing Model
-
(0 ratings)
8.0
(1 ratings)
Professional Services
-
(0 ratings)
10.0
(1 ratings)
User Testimonials
Apache AirflowDatabricks Data Intelligence Platform
Likelihood to Recommend
Apache
For a quick job scanning of status and deep-diving into job issues, details, and flows, AirFlow does a good job. No fuss, no muss. The low learning curve as the UI is very straightforward, and navigating it will be familiar after spending some time using it. Our requirements are pretty simple. Job scheduler, workflows, and monitoring. The jobs we run are >100, but still is a lot to review and troubleshoot when jobs don't run. So when managing large jobs, AirFlow dated UI can be a bit of a drawback.
Read full review
Databricks
Medium to Large data throughput shops will benefit the most from Databricks Spark processing. Smaller use cases may find the barrier to entry a bit too high for casual use cases. Some of the overhead to kicking off a Spark compute job can actually lead to your workloads taking longer, but past a certain point the performance returns cannot be beat.
Read full review
Pros
Apache
  • In charge of the ETL processes.
  • As there is no incoming or outgoing data, we may handle the scheduling of tasks as code and avoid the requirement for monitoring.
Read full review
Databricks
  • Process raw data in One Lake (S3) env to relational tables and views
  • Share notebooks with our business analysts so that they can use the queries and generate value out of the data
  • Try out PySpark and Spark SQL queries on raw data before using them in our Spark jobs
  • Modern day ETL operations made easy using Databricks. Provide access mechanism for different set of customers
Read full review
Cons
Apache
  • they should bring in some time based scheduling too not only event based
  • they do not store the metadata due to which we are not able to analyze the workflows
  • they only support python as of now for scripted pipeline writing
Read full review
Databricks
  • Connect my local code in Visual code to my Databricks Lakehouse Platform cluster so I can run the code on the cluster. The old databricks-connect approach has many bugs and is hard to set up. The new Databricks Lakehouse Platform extension on Visual Code, doesn't allow the developers to debug their code line by line (only we can run the code).
  • Maybe have a specific Databricks Lakehouse Platform IDE that can be used by Databricks Lakehouse Platform users to develop locally.
  • Visualization in MLFLOW experiment can be enhanced
Read full review
Usability
Apache
Easy to learn Easy to use Robust workflow orchestration framework Good in dependent job management
Read full review
Databricks
Because it is an amazing platform for designing experiments and delivering a deep dive analysis that requires execution of highly complex queries, as well as it allows to share the information and insights across the company with their shared workspaces, while keeping it secured.

in terms of graph generation and interaction it could improve their UI and UX
Read full review
Support Rating
Apache
No answers on this topic
Databricks
One of the best customer and technology support that I have ever experienced in my career. You pay for what you get and you get the Rolls Royce. It reminds me of the customer support of SAS in the 2000s when the tools were reaching some limits and their engineer wanted to know more about what we were doing, long before "data science" was even a name. Databricks truly embraces the partnership with their customer and help them on any given challenge.
Read full review
Alternatives Considered
Apache
There are a number of reasons to choose Apache Airflow over other similar platforms- Integrations—ready-to-use operators allow you to integrate Airflow with cloud platforms (Google, AWS, Azure, etc) Apache Airflow helps with backups and other DevOps tasks, such as submitting a Spark job and storing the resulting data on a Hadoop cluster It has machine learning model training, such as triggering a Sage maker job.
Read full review
Databricks
The most important differentiating factor for Databricks Lakehouse Platform from these other platforms is support for ACID transactions and the time travel feature. Also, native integration with managed MLflow is a plus. EMR, Cloudera, and Hortonworks are not as optimized when it comes to Spark Job Execution. Other platforms need to be self-managed, which is another huge hassle.
Read full review
Return on Investment
Apache
  • A lot of helpful features out-of-the-box, such as the DAG visualizations and task trees
  • Allowed us to implement complex data pipelines easily and at a relatively low cost
Read full review
Databricks
  • The ability to spin up a BIG Data platform with little infrastructure overhead allows us to focus on business value not admin
  • DB has the ability to terminate/time out instances which helps manage cost.
  • The ability to quickly access typical hard to build data scenarios easily is a strength.
Read full review
ScreenShots