Apache Kafka is an open-source stream processing platform developed by the Apache Software Foundation written in Scala and Java. The Kafka event streaming platform is used by thousands of companies for high-performance data pipelines, streaming analytics, data integration, and mission-critical applications.
N/A
Db2
Score 8.7 out of 10
N/A
DB2 is a family of relational database software solutions offered by IBM. It includes standard Db2 and Db2 Warehouse editions, either deployable on-cloud, or on-premise.
$0
SAP HANA Cloud
Score 8.9 out of 10
N/A
SAP HANA is an application that uses in-memory database technology to process very large amounts of real-time data from relational databases, both SAP and non-SAP, in a very short time. The in-memory computing engine allows HANA to process data stored in RAM as opposed to reading it from a disk which means that the data can be accessed in real time by the applications using HANA. The product is sold both as an appliance and as a cloud-based software solution.
Implementation and administration complexity, user learning curves, cost considerations, migration difficulties, and possible support and documentation issues are some of the drawbacks of SAP HANA Cloud. With IBM Db2 it is also incredibly safe, effective, and user-friendly. …
DB2 and Oracle are more mature products, however, HANA stacks very well against it in terms of reliability and management. Cost is a huge factor in HANA's favor as well, especially given Oracle's excessive costs.
IBM has been a credible name for us as we have implemented
some of the IBM tools and are going great but when it comes to IBM DB2 it was
our not-so-good experience. We planned to save our time and cost with IBM DB2
Developer friendly environment and real time data access and processing
Verified User
Vice-President
Chose SAP HANA Cloud
Interestingly Workday financials is getting paired with Workday HCM.. Do not find it a comforting approach if one has to have tight integration with logistics operations
As discussed earlier, SAP HANA is one of its kind. With SAP HANA, we are much better equipped to handle and go even beyond the big data trend. Its machine learning and advanced analytic capabilities allow us to integrate with many external and internal resources. [Another] …
As users are comfortable using SAP HANA and now all solutions available with SAP HANA add-on modules the integration becomes much easier and cost effective else you need to have persons of different skill sets to maintain and operate the systems.
The choice of the SAP HANA solution was mainly determined by the choice of the new company ERP, which having been SAP, naturally led to the choice of its DB solution.
Similar to other big DBMS, but better or equal at stability and technical maintenance. Better or equal at documentation. There is room for improvement at SQL path analyzing.
We compared Microsoft BI with SAP HANA. The reasons to go with SAP HANA were - 1. ability to ingest data into HANA from a non SAP database 2. in-memory database resulting in faster real time analytics 3. ability to scale up 4. ability to replicate data real time 5. very solid …
Apache Kafka is well-suited for most data-streaming use cases. Amazon Kinesis and Azure EventHubs, unless you have a specific use case where using those cloud PaAS for your data lakes, once set up well, Apache Kafka will take care of everything else in the background. Azure EventHubs, is good for cross-cloud use cases, and Amazon Kinesis - I have no real-world experience. But I believe it is the same.
I have primarily used it as the basis for a SIS - but I have migrated more than a few systems from there database systems to DB2 (Filemaker, MySQL, etc.). DB2 does have a better structural approach, as opposed to Filemaker, which allows for more data consistency, but this can also lead to an inflexibility that can sometimes be counterintuitive when attempting to compensate for the flexibility of the work environment as Schools tend to have an all in one approach.
I think if you have a large organization, it's probably the product and the marketplace to go to. We're a large management consulting firm operating in four to seven countries. And generally speaking, I think that's the size and the scope where it scales best. I can't speak to smaller companies, but I can't see smaller companies leveraging the benefits as much as a larger organization can.
Really easy to configure. I've used other message brokers such as RabbitMQ and compared to them, Kafka's configurations are very easy to understand and tweak.
Very scalable: easily configured to run on multiple nodes allowing for ease of parallelism (assuming your queues/topics don't have to be consumed in the exact same order the messages were delivered)
Not exactly a feature, but I trust Kafka will be around for at least another decade because active development has continued to be strong and there's a lot of financial backing from Confluent and LinkedIn, and probably many other companies who are using it (which, anecdotally, is many).
Real-time reporting and analytics on data: because of its in-memory architecture, it is perfect for businesses that need to make quick decisions based on current information.
Managing workload with complex data: it can handle a vast range of data types, including relational, documental, geospatial, graph, vector, and time series data.
Developing and deploying intelligent data applications: it provides various tools for such applications and can be used for machine learning and artificial intelligence to automate tasks, gain insights from data, and make predictions.
Sometimes it becomes difficult to monitor our Kafka deployments. We've been able to overcome it largely using AWS MSK, a managed service for Apache Kafka, but a separate monitoring dashboard would have been great.
Simplify the process for local deployment of Kafka and provide a user interface to get visibility into the different topics and the messages being processed.
Learning curve around creation of broker and topics could be simplified
Requires higher processing power, otherwise it won't fly. How ever computing costs are lower. Incase you are migrating to cloud please do not select the highest config available in that series . Upgrading it later against a reserved instance can cost you dearly with a series change
Lack of clarity on licensing is one major challenge
Unless S/4 with additional features are enabled mere migration HANA DB is not a rewarding journey. Power is in S/4
The DB2 database is a solid option for our school. We have been on this journey now for 3-4 years so we are still adapting to what it can do. We will renew our use of DB2 because we don’t see. Major need to change. Also, changing a main database in a school environment is a major project, so we’ll avoid that if possible.
We would rate our likelihood of renewing at 9/10. SAP HANA Cloud has proven to be a highly reliable and scalable data platform that consistently delivers strong performance. Its seamless integration with our overall SAP landscape, combined with improved analytics and real-time data capabilities, makes it a core part of our long-term technology strategy.
Apache Kafka is highly recommended to develop loosely coupled, real-time processing applications. Also, Apache Kafka provides property based configuration. Producer, Consumer and broker contain their own separate property file
You have to be well versed in using the technology, not only from a GUI interface but from a command line interface to successfully use this software to its fullest.
It is very useful solution which provides you speedier data processing, real-time analytics. It helps you manage diverse data types. It also offers you excellent disaster management. It has user friendly interface which helps you navigate system and transactions easily and perform task smoothly.
I have never had DB2 go down unexpectedly. It just works solidly every day. When I look at the logs, sometimes DB2 has figured out there was a need to build an index. Instead of waiting for me to do it, the database automatically created the index for me. At my current company, we have had zero issues for the past 8 years. We have upgrade the server 3 times and upgraded the OS each time and the only thing we saw was that DB2 got better and faster. It is simply amazing.
The performances are exceptional if you take care to maintain the database. It is a very powerful tool and at the same time very easy to use. In our installation, we expect a DB machine on the mainframe with access to the database through ODBC connectors directly from branch servers, with fabulous end users experience.
Support for Apache Kafka (if willing to pay) is available from Confluent that includes the same time that created Kafka at Linkedin so they know this software in and out. Moreover, Apache Kafka is well known and best practices documents and deployment scenarios are easily available for download. For example, from eBay, Linkedin, Uber, and NYTimes.
Easily the best product support team. :) Whenever we have questions, they have answered those in a timely manner and we like how they go above and beyond to help.
However, I am not the right person to answer this as we have another department to handle support and contact the service provider for any support required. Although i will say that they are the quick respondent and knows how to handle querry of the customers and provide quick and better support.
Professional GIS people are some of the most risk-averse there are, and it's difficult to get them to move to HANA in one step. Start with small projects building to 80% use of HANA spatial over time.
I used other messaging/queue solutions that are a lot more basic than Confluent Kafka, as well as another solution that is no longer in the market called Xively, which was bought and "buried" by Google. In comparison, these solutions offer way fewer functionalities and respond to other needs.
DB2 was more scalable and easily configurable than other products we evaluated and short listed in terms of functionality and pricing. IBM also had a good demo on premise and provided us a sandbox experience to test out and play with the product and DB2 at that time came out better than other similar products.
I have deep knowledge of other disk based DBMSs. They are venerable technology, but the attempts to extend them to current architectures belie the fact they are built on 40 year old technology. There are some good columnar in-memory databases but they lack the completeness of capability present in the HANA platform.
By using DB2 only to support my IzPCA activities, my knowledge here is somewhat limited.
Anyway, from what I was able to understand, DB2 is extremely scallable.
Maybe the information below could serve as an example of scalability.
Customer have an huge mainframe environment, 13x z15 CECs, around 80 LPARs, and maybe more than 50 Sysplexes (I am not totally sure about this last figure...)
Today we have 7 IzPCA databases, each one in a distinct Syplex.
Plans are underway to have, at the end, an small LPAR, with only one DB2 sub-system, and with only one database, then transmit the data from a lot of other LPARs, and then process all the data in this only one database.
The IzPCA collect process (read the data received, manipulate it, and insert rows in the tables) today is a huge process, demanding many elapsed hours, and lots of CPU.
Almost 100% of the tables are PBR type, insert jobs run in parallel, but in 4 of the 7 database, it is a really a huge and long process.
Combining the INSERTs loads from the 7 databases in only one will be impossible.......,,,,
But, IzPCA recently introduced a new feature, called "Continuous Collector".
By using that feature, small amounts of data will be transmited to the central LPAR at every 5 minutes (or even less), processed immediately,in a short period of time, and withsmall use of CPU, instead of one or two transmissions by day, of very large amounts of data and the corresponding collect jobs occurring only once or twice a day, with long elapsed times, and huge comsumption of CPU
I suspect the total CPU seconds consumed will be more or less the same in both cases, but in the new method it will occur insmall bursts many times a day!!
Positive: Get a quick and reliable pub/sub model implemented - data across components flows easily.
Positive: it's scalable so we can develop small and scale for real-world scenarios
Negative: it's easy to get into a confusing situation if you are not experienced yet or something strange has happened (rare, but it does). Troubleshooting such situations can take time and effort.