PostgreSQL (alternately Postgres) is a free and open source object-relational database system boasting over 30 years of active development, reliability, feature robustness, and performance. It supports SQL and is designed to support various workloads flexibly.
N/A
Presto
Score 10.0 out of 10
N/A
Presto is an open source SQL query engine designed to run queries on data stored in Hadoop or in traditional databases.
Teradata supported development of Presto followed the acquisition of Hadapt and Revelytix.
N/A
SAP HANA Cloud
Score 8.9 out of 10
N/A
SAP HANA is an application that uses in-memory database technology to process very large amounts of real-time data from relational databases, both SAP and non-SAP, in a very short time. The in-memory computing engine allows HANA to process data stored in RAM as opposed to reading it from a disk which means that the data can be accessed in real time by the applications using HANA. The product is sold both as an appliance and as a cloud-based software solution.
We were using PostgreSQL prior to SAP HANA. The biggest difference that is noticed from an end-user standpoint is the speed with which database transactions take place. Because of the growing scale of our application, we really needed something faster. PostgreSQL just wasn’t …
PostgreSQL is best used for structured data, and best when following relational database design principles. I would not use PostgreSQL for large unstructured data such as video, images, sound files, xml documents, web-pages, especially if these files have their own highly variable, internal structure.
Presto is for interactive simple queries, where Hive is for reliable processing. If you have a fact-dim join, presto is great..however for fact-fact joins presto is not the solution.. Presto is a great replacement for proprietary technology like Vertica
I think if you have a large organization, it's probably the product and the marketplace to go to. We're a large management consulting firm operating in four to seven countries. And generally speaking, I think that's the size and the scope where it scales best. I can't speak to smaller companies, but I can't see smaller companies leveraging the benefits as much as a larger organization can.
Linking, embedding links and adding images is easy enough.
Once you have become familiar with the interface, Presto becomes very quick & easy to use (but, you have to practice & repeat to know what you are doing - it is not as intuitive as one would hope).
Organizing & design is fairly simple with click & drag parameters.
Real-time reporting and analytics on data: because of its in-memory architecture, it is perfect for businesses that need to make quick decisions based on current information.
Managing workload with complex data: it can handle a vast range of data types, including relational, documental, geospatial, graph, vector, and time series data.
Developing and deploying intelligent data applications: it provides various tools for such applications and can be used for machine learning and artificial intelligence to automate tasks, gain insights from data, and make predictions.
Presto was not designed for large fact fact joins. This is by design as presto does not leverage disk and used memory for processing which in turn makes it fast.. However, this is a tradeoff..in an ideal world, people would like to use one system for all their use cases, and presto should get exhaustive by solving this problem.
Resource allocation is not similar to YARN and presto has a priority queue based query resource allocation..so a query that takes long takes longer...this might be alleviated by giving some more control back to the user to define priority/override.
UDF Support is not available in presto. You will have to write your own functions..while this is good for performance, it comes at a huge overhead of building exclusively for presto and not being interoperable with other systems like Hive, SparkSQL etc.
Requires higher processing power, otherwise it won't fly. How ever computing costs are lower. Incase you are migrating to cloud please do not select the highest config available in that series . Upgrading it later against a reserved instance can cost you dearly with a series change
Lack of clarity on licensing is one major challenge
Unless S/4 with additional features are enabled mere migration HANA DB is not a rewarding journey. Power is in S/4
We would rate our likelihood of renewing at 9/10. SAP HANA Cloud has proven to be a highly reliable and scalable data platform that consistently delivers strong performance. Its seamless integration with our overall SAP landscape, combined with improved analytics and real-time data capabilities, makes it a core part of our long-term technology strategy.
Postgresql is the best tool out there for relational data so I have to give it a high rating when it comes to analytics, data availability and consistency, so on and so forth. SQL is also a relatively consistent language so when it comes to building new tables and loading data in from the OLTP database, there are enough tools where we can perform ETL on a scalable basis.
It is very useful solution which provides you speedier data processing, real-time analytics. It helps you manage diverse data types. It also offers you excellent disaster management. It has user friendly interface which helps you navigate system and transactions easily and perform task smoothly.
The data queries are relatively quick for a small to medium sized table. With complex joins, and a wide and deep table however, the performance of the query has room for improvement.
There are several companies that you can contract for technical support, like EnterpriseDB or Percona, both first level in expertise and commitment to the software.
But we do not have contracts with them, we have done all the way from googling to forums, and never have a problem that we cannot resolve or pass around. And for dozens of projects and more than 15 years now.
However, I am not the right person to answer this as we have another department to handle support and contact the service provider for any support required. Although i will say that they are the quick respondent and knows how to handle querry of the customers and provide quick and better support.
The online training is request based. Had there been recorded videos available online for potential users to benefit from, I could have rated it higher. The online documentation however is very helpful. The online documentation PDF is downloadable and allows users to pace their own learning. With examples and code snippets, the documentation is great starting point.
Professional GIS people are some of the most risk-averse there are, and it's difficult to get them to move to HANA in one step. Start with small projects building to 80% use of HANA spatial over time.
Although the competition between the different databases is increasingly aggressive in the sense that they provide many improvements, new functionalities, compatibility with complementary components or environments, in some cases it requires that it be followed within the same family of applications that performs the company that develops it and that is not all bad, but being able to adapt or configure different programs, applications or other environments developed by third parties apart is what gives PostgreSQL a certain advantage and this diversification in the components that can be joined with it, is the reason why it is a great option to choose.
Presto is good for a templated design appeal. You cannot be too creative via this interface - but, the layout and options make the finalized visual product appealing to customers. The other design products I use are for different purposes and not really comparable to Presto.
I have deep knowledge of other disk based DBMSs. They are venerable technology, but the attempts to extend them to current architectures belie the fact they are built on 40 year old technology. There are some good columnar in-memory databases but they lack the completeness of capability present in the HANA platform.
Easy to administer so our DevOps team has only ever used minimal time to setup, tune, and maintain.
Easy to interface with so our Engineering team has only ever used minimal time to query or modify the database. Getting the data is straightforward, what we do with it is the bigger concern.