Databricks offers the Databricks Lakehouse Platform (formerly the Unified Analytics Platform), a data science platform and Apache Spark cluster manager. The Databricks Unified Data Service provides a platform for data pipelines, data lakes, and data platforms.
$0.07
Per DBU
H2O.ai
Score 6.6 out of 10
N/A
An open-source end-to-end GenAI platform for air-gapped, on-premises or cloud VPC deployments. Users can Query and summarize documents or just chat with local private GPT LLMs using h2oGPT, an Apache V2 open-source project. And the commercially available Enterprise h2oGPTe provides information retrieval on internal data, privately hosts LLMs, and secures data.
N/A
TensorFlow
Score 7.7 out of 10
N/A
TensorFlow is an open-source machine learning software library for numerical computation using data flow graphs. It was originally developed by Google.
N/A
Pricing
Databricks Data Intelligence Platform
H2O.ai
TensorFlow
Editions & Modules
Standard
$0.07
Per DBU
Premium
$0.10
Per DBU
Enterprise
$0.13
Per DBU
No answers on this topic
No answers on this topic
Offerings
Pricing Offerings
Databricks Data Intelligence Platform
H2O.ai
TensorFlow
Free Trial
No
No
No
Free/Freemium Version
No
Yes
No
Premium Consulting/Integration Services
No
No
No
Entry-level Setup Fee
No setup fee
No setup fee
No setup fee
Additional Details
—
—
—
More Pricing Information
Community Pulse
Databricks Data Intelligence Platform
H2O.ai
TensorFlow
Considered Multiple Products
Databricks Data Intelligence Platform
Verified User
Team Lead
Chose Databricks Data Intelligence Platform
Databricks was picked among other competitors. Closest competition in our organization was H2O.ai and Databricks came out to be more useful for ROI and time to market in our internal research. We could have used AWS products, however Databricks notebooks and ability to launch …
Both are open source (though H2O only up to some level). Both comprise of deep learning, but H2O is not focused directly on deep learning, while Tensor Flow has a "laser" focus on deep learning. H2O is also more focused on scalability. H2O should be looked at not as a …
Medium to Large data throughput shops will benefit the most from Databricks Spark processing. Smaller use cases may find the barrier to entry a bit too high for casual use cases. Some of the overhead to kicking off a Spark compute job can actually lead to your workloads taking longer, but past a certain point the performance returns cannot be beat.
Most suited if in little time you wanted to build and train a model. Then, H2O makes life very simple. It has support with R, Python and Java, so no programming dependency is required to use it. It's very simple to use. If you want to modify or tweak your ML algorithm then H2O is not suitable. You can't develop a model from scratch.
TensorFlow is great for most deep learning purposes. This is especially true in two domains: 1. Computer vision: image classification, object detection and image generation via generative adversarial networks 2. Natural language processing: text classification and generation. The good community support often means that a lot of off-the-shelf models can be used to prove a concept or test an idea quickly. That, and Google's promotion of Colab means that ideas can be shared quite freely. Training, visualizing and debugging models is very easy in TensorFlow, compared to other platforms (especially the good old Caffe days). In terms of productionizing, it's a bit of a mixed bag. In our case, most of our feature building is performed via Apache Spark. This means having to convert Parquet (columnar optimized) files to a TensorFlow friendly format i.e., protobufs. The lack of good JVM bindings mean that our projects end up being a mix of Python and Scala. This makes it hard to reuse some of the tooling and support we wrote in Scala. This is where MXNet shines better (though its Scala API could do with more work).
Theano is perhaps a bit faster and eats up less memory than TensorFlow on a given GPU, perhaps due to element-wise ops. Tensorflow wins for multi-GPU and “compilation” time.
Because it is an amazing platform for designing experiments and delivering a deep dive analysis that requires execution of highly complex queries, as well as it allows to share the information and insights across the company with their shared workspaces, while keeping it secured.
in terms of graph generation and interaction it could improve their UI and UX
One of the best customer and technology support that I have ever experienced in my career. You pay for what you get and you get the Rolls Royce. It reminds me of the customer support of SAS in the 2000s when the tools were reaching some limits and their engineer wanted to know more about what we were doing, long before "data science" was even a name. Databricks truly embraces the partnership with their customer and help them on any given challenge.
Community support for TensorFlow is great. There's a huge community that truly loves the platform and there are many examples of development in TensorFlow. Often, when a new good technique is published, there will be a TensorFlow implementation not long after. This makes it quick to ally the latest techniques from academia straight to production-grade systems. Tooling around TensorFlow is also good. TensorBoard has been such a useful tool, I can't imagine how hard it would be to debug a deep neural network gone wrong without TensorBoard.
The most important differentiating factor for Databricks Lakehouse Platform from these other platforms is support for ACID transactions and the time travel feature. Also, native integration with managed MLflow is a plus. EMR, Cloudera, and Hortonworks are not as optimized when it comes to Spark Job Execution. Other platforms need to be self-managed, which is another huge hassle.
Both are open source (though H2O only up to some level). Both comprise of deep learning, but H2O is not focused directly on deep learning, while Tensor Flow has a "laser" focus on deep learning. H2O is also more focused on scalability. H2O should be looked at not as a competitor but rather a complementary tool. The use case is usually not only about the algorithms, but also about the data model and data logistics and accessibility. H2O is more accessible due to its UI. Also, both can be accessed from Python. The community around TensorFlow seems larger than that of H2O.
Keras is built on top of TensorFlow, but it is much simpler to use and more Python style friendly, so if you don't want to focus on too many details or control and not focus on some advanced features, Keras is one of the best options, but as far as if you want to dig into more, for sure TensorFlow is the right choice
Positive impact: saving in infrastructure expenses - compared to other bulky tools this costs a fraction
Positive impact: ability to get quick fixes from H2O when problems arise - compared to waiting for several months/years for new releases from other vendors
Positive impact: Access to H2O core team and able to get features that are needed for our business quickly added to the core H2O product