Likelihood to Recommend Hbase is well suited for large organizations with millions of operations performing on tables, real-time lookup of records in a table, range queries, random reads and writes and online analytics operations. Hbase cannot be replaced for traditional databases as it cannot support all the features, CPU and memory intensive. Observed increased latency when using with MapReduce job joins.
Read full review The most important thing when using ClickHouse is to be clear that the scenarios in which you want to use it really are the right ones. Many users think that when a database is very fast for a specific use case, it can be extrapolated to other contexts (most of the time different) in which a previous analysis has not been carried out.
ClickHouse is an analytical database, as such, it should be used for such purposes, where the information is stored correctly, the data volumes are really large and the queries to be performed are not the typical traditional queries on several columns with multiple aggregations. ClickHouse is not the solution for this.
On the other hand, if your case is not one of the above, it is quite possible that ClickHouse can help you. Where ClickHouse shines is when you are looking for aggregation over a particular column in large volumes of data.
Read full review Pros Scalability. HBase can scale to trillions of records. Fast. HBase is extremely fast to scan values or retrieve individual records by key. HBase can be accessed by standard SQL via Apache Phoenix. Integrated. I can easily store and retrieve data from HBase using Apache Spark. It is easy to set up DR and backups. Ingest. It is easy to ingest data into HBase via shell, Java, Apache NiFi, Storm, Spark, Flink, Python and other means. Read full review Their MergeTree table engine provide impressive performance for data insert in bulk Not only data insert but also the way MergeTree engine uses Primary Keys to sort the data and perform data skipping based on the granules its also their secret for ridiculous fast queries Data compression its also great They provide especial table engines that allow you to read data directly from other sources like S3 Since its written with C++ you have very granular data types and especial ones like enum, LowCardinality and etc, they save you a lot of storage since are stored as integer values ClickHouse functions besides the ones that respect ANSI Standards are also awesome and useful Read full review Cons There are very few commands in HBase. Stored procedures functionality is not available so it should be implemented. HBase is CPU and Memory intensive with large sequential input or output access while as Map Reduce jobs are primarily input or output bound with fixed memory. HBase integrated with Map-reduce jobs will result in random latencies. Read full review Avro data manipulation Kafka consistency DDL operations errors (by replica configuration) Read full review Likelihood to Renew There's really not anything else out there that I've seen comparable for my use cases. HBase has never proven me wrong. Some companies align their whole business on HBase and are moving all of their infrastructure from other database engines to HBase. It's also open source and has a very collaborative community.
Read full review Alternatives Considered Cassandra os great for writes. But with large datasets, depending, not as great as HBASE.
Cassandra does support parquet now. HBase still performance issues.
Cassandra has use cases of being used as time series. HBase, it fails miserably. GeoSpatial data, Hbase does work to an extent. HA between the two are almost the same.
Read full review ClickHouse outperforms, especially in costs, since its compression/indexing engines are so smart, and even with very low computing power, you can already perform huge analyses of the data.
Read full review Return on Investment As Hbase is a noSql database, here we don't have transaction support and we cannot do many operations on the data. Not having the feature of primary or a composite primary key is an issue as the architecture to be defined cannot be the same legacy type. Also the transaction concept is not applicable here. The way data is printed on console is not so user-friendly. So we had to use some abstraction over HBase (eg apache phoenix) which means there is one new component to handle. Read full review Queries that used to take more than 2 minutes now take less than 1 second Possibility to analyze use cases in real time (before was impossible) The applications are more complete and the users decisions are better Read full review ScreenShots