Apache Kafka is an open-source stream processing platform developed by the Apache Software Foundation written in Scala and Java. The Kafka event streaming platform is used by thousands of companies for high-performance data pipelines, streaming analytics, data integration, and mission-critical applications.
N/A
PostgreSQL
Score 8.7 out of 10
N/A
PostgreSQL (alternately Postgres) is a free and open source object-relational database system boasting over 30 years of active development, reliability, feature robustness, and performance. It supports SQL and is designed to support various workloads flexibly.
N/A
SAP HANA Cloud
Score 8.9 out of 10
N/A
SAP HANA is an application that uses in-memory database technology to process very large amounts of real-time data from relational databases, both SAP and non-SAP, in a very short time. The in-memory computing engine allows HANA to process data stored in RAM as opposed to reading it from a disk which means that the data can be accessed in real time by the applications using HANA. The product is sold both as an appliance and as a cloud-based software solution.
We were using PostgreSQL prior to SAP HANA. The biggest difference that is noticed from an end-user standpoint is the speed with which database transactions take place. Because of the growing scale of our application, we really needed something faster. PostgreSQL just wasn’t …
As discussed earlier, SAP HANA is one of its kind. With SAP HANA, we are much better equipped to handle and go even beyond the big data trend. Its machine learning and advanced analytic capabilities allow us to integrate with many external and internal resources. [Another] …
We compared Microsoft BI with SAP HANA. The reasons to go with SAP HANA were - 1. ability to ingest data into HANA from a non SAP database 2. in-memory database resulting in faster real time analytics 3. ability to scale up 4. ability to replicate data real time 5. very solid …
Apache Kafka is well-suited for most data-streaming use cases. Amazon Kinesis and Azure EventHubs, unless you have a specific use case where using those cloud PaAS for your data lakes, once set up well, Apache Kafka will take care of everything else in the background. Azure EventHubs, is good for cross-cloud use cases, and Amazon Kinesis - I have no real-world experience. But I believe it is the same.
PostgreSQL is best used for structured data, and best when following relational database design principles. I would not use PostgreSQL for large unstructured data such as video, images, sound files, xml documents, web-pages, especially if these files have their own highly variable, internal structure.
I think if you have a large organization, it's probably the product and the marketplace to go to. We're a large management consulting firm operating in four to seven countries. And generally speaking, I think that's the size and the scope where it scales best. I can't speak to smaller companies, but I can't see smaller companies leveraging the benefits as much as a larger organization can.
Really easy to configure. I've used other message brokers such as RabbitMQ and compared to them, Kafka's configurations are very easy to understand and tweak.
Very scalable: easily configured to run on multiple nodes allowing for ease of parallelism (assuming your queues/topics don't have to be consumed in the exact same order the messages were delivered)
Not exactly a feature, but I trust Kafka will be around for at least another decade because active development has continued to be strong and there's a lot of financial backing from Confluent and LinkedIn, and probably many other companies who are using it (which, anecdotally, is many).
Real-time reporting and analytics on data: because of its in-memory architecture, it is perfect for businesses that need to make quick decisions based on current information.
Managing workload with complex data: it can handle a vast range of data types, including relational, documental, geospatial, graph, vector, and time series data.
Developing and deploying intelligent data applications: it provides various tools for such applications and can be used for machine learning and artificial intelligence to automate tasks, gain insights from data, and make predictions.
Sometimes it becomes difficult to monitor our Kafka deployments. We've been able to overcome it largely using AWS MSK, a managed service for Apache Kafka, but a separate monitoring dashboard would have been great.
Simplify the process for local deployment of Kafka and provide a user interface to get visibility into the different topics and the messages being processed.
Learning curve around creation of broker and topics could be simplified
Requires higher processing power, otherwise it won't fly. How ever computing costs are lower. Incase you are migrating to cloud please do not select the highest config available in that series . Upgrading it later against a reserved instance can cost you dearly with a series change
Lack of clarity on licensing is one major challenge
Unless S/4 with additional features are enabled mere migration HANA DB is not a rewarding journey. Power is in S/4
We would rate our likelihood of renewing at 9/10. SAP HANA Cloud has proven to be a highly reliable and scalable data platform that consistently delivers strong performance. Its seamless integration with our overall SAP landscape, combined with improved analytics and real-time data capabilities, makes it a core part of our long-term technology strategy.
Apache Kafka is highly recommended to develop loosely coupled, real-time processing applications. Also, Apache Kafka provides property based configuration. Producer, Consumer and broker contain their own separate property file
Postgresql is the best tool out there for relational data so I have to give it a high rating when it comes to analytics, data availability and consistency, so on and so forth. SQL is also a relatively consistent language so when it comes to building new tables and loading data in from the OLTP database, there are enough tools where we can perform ETL on a scalable basis.
It is very useful solution which provides you speedier data processing, real-time analytics. It helps you manage diverse data types. It also offers you excellent disaster management. It has user friendly interface which helps you navigate system and transactions easily and perform task smoothly.
The data queries are relatively quick for a small to medium sized table. With complex joins, and a wide and deep table however, the performance of the query has room for improvement.
Support for Apache Kafka (if willing to pay) is available from Confluent that includes the same time that created Kafka at Linkedin so they know this software in and out. Moreover, Apache Kafka is well known and best practices documents and deployment scenarios are easily available for download. For example, from eBay, Linkedin, Uber, and NYTimes.
There are several companies that you can contract for technical support, like EnterpriseDB or Percona, both first level in expertise and commitment to the software.
But we do not have contracts with them, we have done all the way from googling to forums, and never have a problem that we cannot resolve or pass around. And for dozens of projects and more than 15 years now.
However, I am not the right person to answer this as we have another department to handle support and contact the service provider for any support required. Although i will say that they are the quick respondent and knows how to handle querry of the customers and provide quick and better support.
The online training is request based. Had there been recorded videos available online for potential users to benefit from, I could have rated it higher. The online documentation however is very helpful. The online documentation PDF is downloadable and allows users to pace their own learning. With examples and code snippets, the documentation is great starting point.
Professional GIS people are some of the most risk-averse there are, and it's difficult to get them to move to HANA in one step. Start with small projects building to 80% use of HANA spatial over time.
I used other messaging/queue solutions that are a lot more basic than Confluent Kafka, as well as another solution that is no longer in the market called Xively, which was bought and "buried" by Google. In comparison, these solutions offer way fewer functionalities and respond to other needs.
Although the competition between the different databases is increasingly aggressive in the sense that they provide many improvements, new functionalities, compatibility with complementary components or environments, in some cases it requires that it be followed within the same family of applications that performs the company that develops it and that is not all bad, but being able to adapt or configure different programs, applications or other environments developed by third parties apart is what gives PostgreSQL a certain advantage and this diversification in the components that can be joined with it, is the reason why it is a great option to choose.
I have deep knowledge of other disk based DBMSs. They are venerable technology, but the attempts to extend them to current architectures belie the fact they are built on 40 year old technology. There are some good columnar in-memory databases but they lack the completeness of capability present in the HANA platform.
Positive: Get a quick and reliable pub/sub model implemented - data across components flows easily.
Positive: it's scalable so we can develop small and scale for real-world scenarios
Negative: it's easy to get into a confusing situation if you are not experienced yet or something strange has happened (rare, but it does). Troubleshooting such situations can take time and effort.
Easy to administer so our DevOps team has only ever used minimal time to setup, tune, and maintain.
Easy to interface with so our Engineering team has only ever used minimal time to query or modify the database. Getting the data is straightforward, what we do with it is the bigger concern.