Confluent Cloud is a cloud-native service for Apache Kafka used to connect and process data in real time with a fully managed data streaming platform. Confluent Platform is the self-managed version.
$385
per month
Databricks Data Intelligence Platform
Score 8.8 out of 10
N/A
Databricks offers the Databricks Lakehouse Platform (formerly the Unified Analytics Platform), a data science platform and Apache Spark cluster manager. The Databricks Unified Data Service provides a platform for data pipelines, data lakes, and data platforms.
$0.07
Per DBU
IBM Cognos Analytics
Score 7.4 out of 10
N/A
IBM Cognos is a full-featured business intelligence suite by IBM, designed for larger deployments. It comprises Query Studio, Reporting Studio, Analysis Studio and Event Studio, and Cognos Administration along with tools for Microsoft Office integration, full-text search, and dashboards.
$10
per month per user
Pricing
Confluent
Databricks Data Intelligence Platform
IBM Cognos Analytics
Editions & Modules
Basic
$0
Standard
Starting at ~$385
per month
Enterprise
Starting at ~$1,150
per month
Standard
$0.07
Per DBU
Premium
$0.10
Per DBU
Enterprise
$0.13
Per DBU
On Demand - Standard
USD 10.00
per month per user
On Demand - Premium
USD 42.40
per month per user
On Demand - Standard
USD 10.60
per month per user
Offerings
Pricing Offerings
Confluent
Databricks Data Intelligence Platform
IBM Cognos Analytics
Free Trial
No
No
Yes
Free/Freemium Version
Yes
No
No
Premium Consulting/Integration Services
No
No
Yes
Entry-level Setup Fee
No setup fee
No setup fee
Optional
Additional Details
Confluent monthly bills are based upon resource consumption, i.e., you are only charged for the resources you use when you actually use them:
Stream: Kafka clusters are billed for eCKUs/CKUs ($/hour), networking ($/GB), and storage ($/GB-hour).
Connect: Use of connectors is billed based on throughput ($/GB) and a task base price ($/task/hour).
Process: Use of stream processing with Confluent Cloud for Apache Flink is calculated based on CFUs ($/minute).
Govern: Use of Stream Governance is billed based on environment ($/hour).
Confluent storage and throughput is calculated in binary gigabytes (GB), where 1 GB is 2^30 bytes. This unit of measurement is also known as a gibibyte (GiB). Please also note that all prices are stated in United States Dollars unless specifically stated otherwise.
All billing computations are conducted in Coordinated Universal Time (UTC).
For our use case it was very important that the technology we were working with fit into our Azure architecture, and met our data processing size requirements to stream data within certain SLAs. Confluent more than met our performance requirements and compared to the others …
If you have a need to stream data, real time or segmented structured data then Confluent is a great platform to do so with. You won't run into packet transfer size limitations that other platforms have. Flexibility in on-prem, cloud, and managed cloud offerings makes it very flexible no matter how you choose to implement.
Medium to Large data throughput shops will benefit the most from Databricks Spark processing. Smaller use cases may find the barrier to entry a bit too high for casual use cases. Some of the overhead to kicking off a Spark compute job can actually lead to your workloads taking longer, but past a certain point the performance returns cannot be beat.
Well suited: Financial reporting - It can handle complex, pixel perfect, muti-page reports with scheduled delivery to stakeholders (like sales report by region on quarterly periodicity) Operational dashboard across departments - It can combine multiple data sources (ERP, CRM, excels etc) with filters, and embedded AI insights Less appropriate: Live dashboards - As stated earlier as well, IBM Cognos Analytics doesn't suit well for live dashboards or event driven data. For ex: live web traffic data or IOT device data, etc Data science - Although IBM Cognos Analytics is great tool for data exploration but it should not be used as a substitute for Python or R, which has edge over advanced modelling and stats based workflows like predictive modelling or clustering
IBM Cognos Analytics enables customer data segmentation, which is essential for marketing, improving and streamlining purchasing behavior and preferences. This helps companies create more targeted and effective marketing campaigns.
Our clients Through data analysis, we can identify and observe trends in the behavior of other clients, allowing us to anticipate needs and adjust strategies to avoid consequences.
For an existing solution, renewing licenses does provide a good return on investment. Additionally, while rolling out scorecards and dashboards with little adhoc capabilities, to end users, cognos is very easily scalable. It also allows to create a solution that has a mix of OLAP and relational data-sources, which is a limitation with other tools. Synchronizing with existing security setup is easy too.
Because it is an amazing platform for designing experiments and delivering a deep dive analysis that requires execution of highly complex queries, as well as it allows to share the information and insights across the company with their shared workspaces, while keeping it secured.
in terms of graph generation and interaction it could improve their UI and UX
We have a strong user base (3500 users) that are highly utilizing this tool. Basic users are able to consume content within the applied security model. We have a set of advanced users that really push the limits of Cognos with Report and Query Studio. These users have created a lot of personal content and stored it in 'My Reports'. Users enjoy this flexibility.
Reports can typically be viewed through any browser that can access the server, so the availability is ultimately up to what the company utilizing it is comfortable with allowing, though report development tends to be more picky about browsers and settings as mentioned above. It also has an optional iPad app and general mobile browsing support, but dashboards lack the mobile compatibility. What keeps it from getting a higher score is the desktop tools that are vital to the development process. The compatibility with only Windows when the server has a wide range of compatibility can be a real sore point for a company that outfits its employees exclusively with Mac or Linux machines. Of course, if they are planning on outsourcing the development anyways, it's a rather moot point
Overall no major complaints but it doesn't handle DMR (Dimensionally Modeled for Relational) very well. DMR modelling is a capability that IBM Cognos Framework Manager provides allowing you to specify dimensional information for relational metadata and allows for OLAP-style queries. However, the capability is not very efficient and, for example, if I'm using only 2 columns on a 20-column model, the software is not smart enough to exclude 18 columns and the query side gets progressively larger and larger until it's effectively unusable.
The support from the Confluent platform is great and satisfying. We have been working with Confluent for more than a year now. They sent out resident architects to help us set up Confluent cluster on our cloud and help us troubleshoot problems we have encountered. Overall, it has been a great experience working with the Confluent Platform.
One of the best customer and technology support that I have ever experienced in my career. You pay for what you get and you get the Rolls Royce. It reminds me of the customer support of SAS in the 2000s when the tools were reaching some limits and their engineer wanted to know more about what we were doing, long before "data science" was even a name. Databricks truly embraces the partnership with their customer and help them on any given challenge.
Why is their web application not working as fast as you think it should? They never know, and it is always a a bunch of shots in the dark to find out. Trying to download software from them is like trying to find a book at the library before computers were invented.
Onsite training provided by IBM Cognos was effective and as expected. They did not perform training with our data which was a bit difficult for our end-users.
The online courses they offer are thorough and presented in such a way that someone who isn't already familiar with the general design methodologies used in this field will be capable of making a good design. The training environments are provided as a fully self contained virtual machine with everything needed already to create the environments. We've had some persisting issues with the environments becoming unavailable, but support has been responsive when these issues arise and straightening them out for us
Make sure that any custom tables that you have, are built into your metadata packages. You can still access them via SQL queries in Cognos, but it is much easier to have them as a part of the available metadata packages.
For our use case it was very important that the technology we were working with fit into our Azure architecture, and met our data processing size requirements to stream data within certain SLAs. Confluent more than met our performance requirements and compared to the others scale options and cost to run it was more than financially viable as a platform solution to our global operations.
The most important differentiating factor for Databricks Lakehouse Platform from these other platforms is support for ACID transactions and the time travel feature. Also, native integration with managed MLflow is a plus. EMR, Cloudera, and Hortonworks are not as optimized when it comes to Spark Job Execution. Other platforms need to be self-managed, which is another huge hassle.
Power BI is stronger for quick ad-hoc analysis and dashboards, but IBM Cognos Analytics is better when consistency, precision, and mass distribution matter. Tableau is best for interactive analysis, while IBM Cognos Analytics is better for standardized, repeatable enterprise reporting. Sigma shines for customizable dashboards and drill-down analysis while IBM Cognos Analytics holds an edge in data discovery and visualization.
The Cognos architecture is well suited for scalability. However, the architecture must be designed with scalability in mind from day one of the implementation. We recently upgraded from 10.1 to 10.2.1 and took the opportunity to revamp our architecture. It is now poised for future growth and scalability.