SAP Datasphere, the next generation of SAP Data Warehouse Cloud, is a comprehensive data service that enables data professionals to deliver seamless and scalable access to mission-critical business data. It provides a unified experience for data integration, data cataloging, semantic modeling, data warehousing, data federation, and data virtualization. SAP Datasphere enables users to distribute mission-critical business data — with business context and logic preserved — across the data…
N/A
SingleStore
Score 8.3 out of 10
N/A
SingleStore aims to enable organizations to scale from one to one million customers, handling SQL, JSON, full text and vector workloads in one unified platform.
$0.69
per hour
Pricing
SAP Datasphere
SingleStore
Editions & Modules
No answers on this topic
OnDemand
$0.69
per hour
Offerings
Pricing Offerings
SAP Datasphere
SingleStore
Free Trial
Yes
Yes
Free/Freemium Version
No
Yes
Premium Consulting/Integration Services
No
Yes
Entry-level Setup Fee
No setup fee
Optional
Additional Details
SAP Datasphere is available as a subscription or consumption-based model. The SAP Datasphere capacity unit (CU) offers an adaptable approach to pricing that enables any workload on any hyperscaler. The number of CUs required is determined by the unique workload, with the ability to tailor the combination of required services within SAP Datasphere utilizing a flexible tenant configuration. The services that contribute to CU consumption are the core application (compute and storage), data lake, BW bridge, data integration, and data catalog (crawling and storage).
SAP Datasphere is well suited for scalable cloud based data integration scenarios which also opens up the doors for AI driven insights which are much harder to achieve with on-prem data warehouses. Considering the licensing model of SAP Datasphere being based on consumption driven capacity units cost can be a big consideration for organizations with large volumes of data that can be a pre-requisite for data mining and AI use cases. So this can be a bottleneck or not so well adopted scenario for SAP Datasphere.
Good for Applications needing instant insights on large, streaming datasets. Applications processing continuous data streams with low latency. When a multi-cloud, high-availability database is required When NOT to Use Small-scale applications with limited budgets Projects that do not require real-time analytics or distributed scaling Teams without experience in distributed databases and HTAP architectures.
SAP Data Warehouse Cloud offers free trial for 90 days with free 128 GB of storage and 64 GB memory.
Availability of self-service data modeling and analytics on SAP Data Warehouse Cloud enables users to access and analyze data without getting support from the IT team.
Without zero coding while collecting, connecting, analyzing and modeling data, it saves us time and operational costs of partnering with external IT support experts.
It does not release a patch to have back porting; it just releases a new version and stops support; it's difficult to keep up to that pace.
Support engineers lack expertise, but they seem to be improving organically.
Lacks enterprise CDC capability: Change data capture (CDC) is a process that tracks and records changes made to data in a database and then delivers those changes to other systems in real time.
For enterprise-level backup & restore capability, we had to implement our model via Velero snapshot backup.
We are moving into using SAP datasphere heavily and replacing all of the SAP HANA native calc view logic to the sap datasphere graphical view which will reduce the legacy SAP BW data warehouse. Also need some more features such as debugging, sql preview and prompts enhancements so that we can generate the reports.
It is one of the best tools and a boon to Logistics teams across the globe. One tends to actually process warehousing data so smoothly and the way demonstration is made while in programs it makes it user friendly. The Inventory touch points that one identify is simply awesome and is best part.
[Until it is] supported on AWS ECS containers, I will reserve a higher rating for SingleStore. Right now it works well on EC2 and serves our current purpose, [but] would look forward to seeing SingleStore respond to our urge of feature in a shorter time period with high quality and security.
SingleStore excels in real-time analytics and low-latency transactions, making it ideal for operational analytics and mixed workloads. Snowflake shines in batch analytics and data warehousing with strong scalability for large datasets. SingleStore offers faster data ingestion and query execution for real-time use cases, while Snowflake is better for complex analytical queries on historical data.
I would greatly acknowledge the services of Sap Data [warehouse Cloud] because we were struggling before its arrival where we used to get manual data connections and this used to consume a lot of time but after its use, we now are able to connect data easily saving a lot of time and finances.
The support deep dives into our most complexed queries and bizarre issues that sometimes only we get comparing to other clients. Our special workload (thousands of Kafka pipelines + high concurrency of queries). The response match to the priority of the request, P1 gets immediate return call. Missing features are treated, they become a client request and being added to the roadmap after internal consideration on all client needs and priority. Bugs are patched quite fast, depends on the impact and feasible temporary workarounds. There is no issue that we haven't got a proper answer, resolution or reasoning
We allowed 2-3 months for a thorough evaluation. We saw pretty quickly that we were likely to pick SingleStore, so we ported some of our stored procedures to SingleStore in order to take a deeper look. Two SingleStore people worked closely with us to ensure that we did not have any blocking problems. It all went remarkably smoothly.
Each of these listed software has its own unique strength and capacity that scales well. SAP Datasphere on its end up against them with more suitability for large establishments with complex data ecosystems with scalability support. Also, it avails a pay-as-you-go pricing for users, and it is widely up for data quality, data governance, and data discovery.
Greenplum is good in handling very large amount of data. Concurrency in Greenplum was a major problem. Features available in SingleStore like Pipelines and in memory features are not available in Greenplum. Gemfire was not scaling well like SingleStore. Support of both Greenplum and Gemfire was not good. Product team did not help us much like the ones in SingleStore who helped us getting started on our first cluster very fast.
Despite the pricing model being expensive for small businesses, it provides decent features and capabilities for organizations of different sizes and it's an appropriate investment in today's business environment where there is constant pressure to build a scalable and flexible analytics service
Ever since we implemented SAP Data Warehouse Cloud, we have been able to reduce the additional costs of hiring third-party service providers by incorporating professional services offered by the vendor.
Preserving data quality has enhanced governance on data by having a single source that is accessible to every business user via self-service capabilities.
Operational cost is lowered by connecting data in one integrated solution hence making it easy to access information without having to keeping logging to other applications. Additionally, no external IT support is needed since SAP Data Warehouse Cloud has no-coding modeling tools.
SAP Data Warehouse Cloud has enabled every business user to understand different data by transforming data to real insights.
As the overall performance and functionality were expanded, we are able to deliver our data much faster than before, which increases the demand for data.
Metadata is available in the platform by default, like metadata on the pipelines. Also, the information schema has lots of metadata, making it easy to load our assets to the data catalog.