IBM Cloud Databases are open source data stores for enterprise application development. Built on a Kubernetes foundation, they offer a database platform for serverless applications. They are designed to scale storage and compute resources seamlessly without being constrained by the limits of a single server. Natively integrated and available in the IBM Cloud console, these databases are now available through a consistent consumption, pricing, and interaction model. They aim to provide a cohesive…
N/A
SAP HANA Cloud
Score 8.9 out of 10
N/A
SAP HANA is an application that uses in-memory database technology to process very large amounts of real-time data from relational databases, both SAP and non-SAP, in a very short time. The in-memory computing engine allows HANA to process data stored in RAM as opposed to reading it from a disk which means that the data can be accessed in real time by the applications using HANA. The product is sold both as an appliance and as a cloud-based software solution.
$0.95
per month Capacity Units
SingleStore
Score 7.8 out of 10
N/A
SingleStore aims to enable organizations to scale from one to one million customers, handling SQL, JSON, full text and vector workloads in one unified platform.
We decided to go with SAP HANA Cloud because it provides more features and better pricing than IBM cloud database. We are more familiar with SAP HANA Cloud than IBM Cloud Databases.
We adopt SAP HANA Cloud because it excels in in-memory processing which provides faster access to data. It offers real-time performance, deep integration with SAP software, and advanced analytics & visualization capabilities.
Less Appropriate Scenario: 1) Small Scale or Low Budget Projects 2) Organizations with limited expertise in cloud technologies may find the learning curve steep, especially if they are not familiar with the IBM Cloud platform 3) If database requirements are highly dynamic and change frequently, the comprehensive features and management provided by IBM Cloud Databases might be overkill. A more flexible, self-managed solution could be preferable for adapting to rapid changes.
I think if you have a large organization, it's probably the product and the marketplace to go to. We're a large management consulting firm operating in four to seven countries. And generally speaking, I think that's the size and the scope where it scales best. I can't speak to smaller companies, but I can't see smaller companies leveraging the benefits as much as a larger organization can.
Good for Applications needing instant insights on large, streaming datasets. Applications processing continuous data streams with low latency. When a multi-cloud, high-availability database is required When NOT to Use Small-scale applications with limited budgets Projects that do not require real-time analytics or distributed scaling Teams without experience in distributed databases and HTAP architectures.
The ease of setup was effortless. For anyone with development experience, a few simple questions such as name and login data will get you set up.
The web application to manage cluster settings, billing settings and even introspect the data was simple and most importantly worked all the time. This can not always be said for web interfaces of other products.
Real-time reporting and analytics on data: because of its in-memory architecture, it is perfect for businesses that need to make quick decisions based on current information.
Managing workload with complex data: it can handle a vast range of data types, including relational, documental, geospatial, graph, vector, and time series data.
Developing and deploying intelligent data applications: it provides various tools for such applications and can be used for machine learning and artificial intelligence to automate tasks, gain insights from data, and make predictions.
Better cost reports, before just increasing to another tier, thus increasing the price. This is critical for early stage startups, where budget is tight.
Add more data center options. As a comparison, a similar service, Aiven.io has dozen more options than Compose (basically all big cloud providers). We moved from AWS to Digital Ocean, which made us stop using Compose, since Compose forces us to be either on IBM or AWS.
Requires higher processing power, otherwise it won't fly. How ever computing costs are lower. Incase you are migrating to cloud please do not select the highest config available in that series . Upgrading it later against a reserved instance can cost you dearly with a series change
Lack of clarity on licensing is one major challenge
Unless S/4 with additional features are enabled mere migration HANA DB is not a rewarding journey. Power is in S/4
It does not release a patch to have back porting; it just releases a new version and stops support; it's difficult to keep up to that pace.
Support engineers lack expertise, but they seem to be improving organically.
Lacks enterprise CDC capability: Change data capture (CDC) is a process that tracks and records changes made to data in a database and then delivers those changes to other systems in real time.
For enterprise-level backup & restore capability, we had to implement our model via Velero snapshot backup.
IBM is our trusted partner which never failed to meet our expectations. Stability, efficiency, usability and security is a must have for our business which is fully provided by IBM Cloud Databases
We would rate our likelihood of renewing at 9/10. SAP HANA Cloud has proven to be a highly reliable and scalable data platform that consistently delivers strong performance. Its seamless integration with our overall SAP landscape, combined with improved analytics and real-time data capabilities, makes it a core part of our long-term technology strategy.
IBM Cloud Databases' pricing structure is easy to understand, and if you choose the right product, you can operate your system at minimal cost. Although there is ample documentation available, there doesn't seem to be a user community running on it, so specific usage know-how and troubleshooting can sometimes take longer than expected.
It is very useful solution which provides you speedier data processing, real-time analytics. It helps you manage diverse data types. It also offers you excellent disaster management. It has user friendly interface which helps you navigate system and transactions easily and perform task smoothly.
[Until it is] supported on AWS ECS containers, I will reserve a higher rating for SingleStore. Right now it works well on EC2 and serves our current purpose, [but] would look forward to seeing SingleStore respond to our urge of feature in a shorter time period with high quality and security.
Solutions are based around a business needs and even when implementing such solution, real time insights are also followed through showing the updates the business are implementing while informing the end users as what is new with technology.
SingleStore excels in real-time analytics and low-latency transactions, making it ideal for operational analytics and mixed workloads. Snowflake shines in batch analytics and data warehousing with strong scalability for large datasets. SingleStore offers faster data ingestion and query execution for real-time use cases, while Snowflake is better for complex analytical queries on historical data.
Support is helpful enough, but we haven't always had questions answered in a satisfactory manner. At one time we realized that Compose had stopped taking database snapshots on its two-per-day schedule, and had in fact not taken one for many days. Support recognized the problem and it was fixed, but the lack of proactive checks and the inability to share exactly what happened has caused us to look elsewhere for production work loads
However, I am not the right person to answer this as we have another department to handle support and contact the service provider for any support required. Although i will say that they are the quick respondent and knows how to handle querry of the customers and provide quick and better support.
The support deep dives into our most complexed queries and bizarre issues that sometimes only we get comparing to other clients. Our special workload (thousands of Kafka pipelines + high concurrency of queries). The response match to the priority of the request, P1 gets immediate return call. Missing features are treated, they become a client request and being added to the roadmap after internal consideration on all client needs and priority. Bugs are patched quite fast, depends on the impact and feasible temporary workarounds. There is no issue that we haven't got a proper answer, resolution or reasoning
Professional GIS people are some of the most risk-averse there are, and it's difficult to get them to move to HANA in one step. Start with small projects building to 80% use of HANA spatial over time.
We allowed 2-3 months for a thorough evaluation. We saw pretty quickly that we were likely to pick SingleStore, so we ported some of our stored procedures to SingleStore in order to take a deeper look. Two SingleStore people worked closely with us to ensure that we did not have any blocking problems. It all went remarkably smoothly.
The reason why I choose IBM Cloud Databases is that the IBM cloud toolset is already being used in other functions of the company and by using IBM Cloud Databases, the other cloud tools are better embedded and integrated. If the company is set to use amazon tools, I would go for rds.
I have deep knowledge of other disk based DBMSs. They are venerable technology, but the attempts to extend them to current architectures belie the fact they are built on 40 year old technology. There are some good columnar in-memory databases but they lack the completeness of capability present in the HANA platform.
Greenplum is good in handling very large amount of data. Concurrency in Greenplum was a major problem. Features available in SingleStore like Pipelines and in memory features are not available in Greenplum. Gemfire was not scaling well like SingleStore. Support of both Greenplum and Gemfire was not good. Product team did not help us much like the ones in SingleStore who helped us getting started on our first cluster very fast.
As the overall performance and functionality were expanded, we are able to deliver our data much faster than before, which increases the demand for data.
Metadata is available in the platform by default, like metadata on the pipelines. Also, the information schema has lots of metadata, making it easy to load our assets to the data catalog.