SingleStore aims to enable organizations to scale from one to one million customers, handling SQL, JSON, full text and vector workloads in one unified platform.
$0.69
per hour
Teradata Vantage
Score 8.4 out of 10
N/A
Teradata Vantage is presented as a modern analytics cloud platform that unifies everything—data lakes, data warehouses, analytics, and new data sources and types. Supports hybrid multi-cloud environments and priced for flexibility, Vantage delivers unlimited intelligence to build the future of business.
Users can deploy Vantage on public clouds (such as AWS, Azure, and GCP), hybrid multi-cloud environments, on-premises with Teradata IntelliFlex, or on commodity hardware with VMware.
Good for Applications needing instant insights on large, streaming datasets. Applications processing continuous data streams with low latency. When a multi-cloud, high-availability database is required When NOT to Use Small-scale applications with limited budgets Projects that do not require real-time analytics or distributed scaling Teams without experience in distributed databases and HTAP architectures.
Teradata Vantage is well suited for large scale ETL pipelines like the ones we developed for anti money laundering risk matrices. It handles heavy joins, aggregations, and transformations on transactional data efficiently. We generate alert variables, adjust for inflation, and monitor establishments monthly with it, all integrated with Python and Control-M for a centralised automation across the company. For less appropriate, I would say that heavy resource demands might slow down experimentation for iterative work.
It does not release a patch to have back porting; it just releases a new version and stops support; it's difficult to keep up to that pace.
Support engineers lack expertise, but they seem to be improving organically.
Lacks enterprise CDC capability: Change data capture (CDC) is a process that tracks and records changes made to data in a database and then delivers those changes to other systems in real time.
For enterprise-level backup & restore capability, we had to implement our model via Velero snapshot backup.
Teradata is an excellent option but only for a massive amount of data warehousing or analysis. If your data is not that big then it could be a misfit for your company and cost you a lot. The cost associated is quite extensive as compared to some other alternative RDBMS systems available in the market.
Migration of data from Teradata to some other RDBMS systems is quite painful as the transition is not that smooth and you need to follow many steps and even if one of them fails. You need to start from the beginning almost.
Last but not least the UI is pretty outdated and needs a revamp. Though it is simple, it needs to be presented in a much better way and more advanced options need to bee presented on the front page itself.
Teradata is a mature RDBMS system that expands its functionality towards the current cloud capabilities like object storage and flexible compute scale.
[Until it is] supported on AWS ECS containers, I will reserve a higher rating for SingleStore. Right now it works well on EC2 and serves our current purpose, [but] would look forward to seeing SingleStore respond to our urge of feature in a shorter time period with high quality and security.
Teradata Vantage allows us to create a scalable infrastructure to support our strategic initiatives. The dedicated compute power ensures reliable performance with isolated workloads and dedicated resources, optimizing workflows for faster, more efficient data transfers. The compute clusters support ETL processes and OSF’s developers and data science team with the flexibility to create self-service analytics, to spin up/down at any time, driving better performance and minimizing costs.
SingleStore excels in real-time analytics and low-latency transactions, making it ideal for operational analytics and mixed workloads. Snowflake shines in batch analytics and data warehousing with strong scalability for large datasets. SingleStore offers faster data ingestion and query execution for real-time use cases, while Snowflake is better for complex analytical queries on historical data.
The support deep dives into our most complexed queries and bizarre issues that sometimes only we get comparing to other clients. Our special workload (thousands of Kafka pipelines + high concurrency of queries). The response match to the priority of the request, P1 gets immediate return call. Missing features are treated, they become a client request and being added to the roadmap after internal consideration on all client needs and priority. Bugs are patched quite fast, depends on the impact and feasible temporary workarounds. There is no issue that we haven't got a proper answer, resolution or reasoning
We have meetings at the beginning with the technical team to explain our requirements to them and they were really putting in a lot of effort to come up with a solution which will address all our needs. They implemented the software and also trained a few of our resources on the same too. We can get in touch with them now as well whenever we run into a roadblock but it's very less now.
We allowed 2-3 months for a thorough evaluation. We saw pretty quickly that we were likely to pick SingleStore, so we ported some of our stored procedures to SingleStore in order to take a deeper look. Two SingleStore people worked closely with us to ensure that we did not have any blocking problems. It all went remarkably smoothly.
Greenplum is good in handling very large amount of data. Concurrency in Greenplum was a major problem. Features available in SingleStore like Pipelines and in memory features are not available in Greenplum. Gemfire was not scaling well like SingleStore. Support of both Greenplum and Gemfire was not good. Product team did not help us much like the ones in SingleStore who helped us getting started on our first cluster very fast.
Teradata is way ahead of its competitor because of its unique features of ensuring data privacy and data never gets corrupted even in worst case scenario. In most cases, the data corruption is a major issue if left unused and it leads to important data being wiped off which in ideal case should be stored for 3 years
As the overall performance and functionality were expanded, we are able to deliver our data much faster than before, which increases the demand for data.
Metadata is available in the platform by default, like metadata on the pipelines. Also, the information schema has lots of metadata, making it easy to load our assets to the data catalog.
Moving to Teradata in the Cloud-enabled a level of agility that previously didn't exist in the organization. It also enabled a level of analytic competency that was not achievable using other options on the aggressive timeline that was required. We didn't want to settle for reinventing a wheel when we had a super tuned performance capable beast readily available in Teradata. Teradata lets us focus on our business rather than spending money and effort trying to design software or database foundations features on an open source or lower performance platform.