Hadoop is an open source software from Apache, supporting distributed processing and data storage. Hadoop is popular for its scalability, reliability, and functionality available across commoditized hardware.
N/A
PostgreSQL
Score 8.7 out of 10
N/A
PostgreSQL (alternately Postgres) is a free and open source object-relational database system boasting over 30 years of active development, reliability, feature robustness, and performance. It supports SQL and is designed to support various workloads flexibly.
N/A
SAP HANA Cloud
Score 8.9 out of 10
N/A
SAP HANA is an application that uses in-memory database technology to process very large amounts of real-time data from relational databases, both SAP and non-SAP, in a very short time. The in-memory computing engine allows HANA to process data stored in RAM as opposed to reading it from a disk which means that the data can be accessed in real time by the applications using HANA. The product is sold both as an appliance and as a cloud-based software solution.
I haven't worked with other Big Data aggregation services like Hadoop. As far as I know, Hadoop is the leading choice in this field with good cause. There is a lot of community support, custom modules, paid consultants, free and paid training. All this makes it an ideal choice …
In my experience using all of these products over many years, PostgreSQL is better than any of them in reliability, performance, productivity, cost, scalability and interoperability across operating systems.
We were using PostgreSQL prior to SAP HANA. The biggest difference that is noticed from an end-user standpoint is the speed with which database transactions take place. Because of the growing scale of our application, we really needed something faster. PostgreSQL just wasn’t …
We compared Microsoft BI with SAP HANA. The reasons to go with SAP HANA were - 1. ability to ingest data into HANA from a non SAP database 2. in-memory database resulting in faster real time analytics 3. ability to scale up 4. ability to replicate data real time 5. very solid …
Altogether, I want to say that Apache Hadoop is well-suited to a larger and unstructured data flow like an aggregation of web traffic or even advertising. I think Apache Hadoop is great when you literally have petabytes of data that need to be stored and processed on an ongoing basis. Also, I would recommend that the software should be supplemented with a faster and interactive database for a better querying service. Lastly, it's very cost-effective so it is good to give it a shot before coming to any conclusion.
PostgreSQL is best used for structured data, and best when following relational database design principles. I would not use PostgreSQL for large unstructured data such as video, images, sound files, xml documents, web-pages, especially if these files have their own highly variable, internal structure.
I think if you have a large organization, it's probably the product and the marketplace to go to. We're a large management consulting firm operating in four to seven countries. And generally speaking, I think that's the size and the scope where it scales best. I can't speak to smaller companies, but I can't see smaller companies leveraging the benefits as much as a larger organization can.
Real-time reporting and analytics on data: because of its in-memory architecture, it is perfect for businesses that need to make quick decisions based on current information.
Managing workload with complex data: it can handle a vast range of data types, including relational, documental, geospatial, graph, vector, and time series data.
Developing and deploying intelligent data applications: it provides various tools for such applications and can be used for machine learning and artificial intelligence to automate tasks, gain insights from data, and make predictions.
Requires higher processing power, otherwise it won't fly. How ever computing costs are lower. Incase you are migrating to cloud please do not select the highest config available in that series . Upgrading it later against a reserved instance can cost you dearly with a series change
Lack of clarity on licensing is one major challenge
Unless S/4 with additional features are enabled mere migration HANA DB is not a rewarding journey. Power is in S/4
Hadoop is organization-independent and can be used for various purposes ranging from archiving to reporting and can make use of economic, commodity hardware. There is also a lot of saving in terms of licensing costs - since most of the Hadoop ecosystem is available as open-source and is free
We would rate our likelihood of renewing at 9/10. SAP HANA Cloud has proven to be a highly reliable and scalable data platform that consistently delivers strong performance. Its seamless integration with our overall SAP landscape, combined with improved analytics and real-time data capabilities, makes it a core part of our long-term technology strategy.
As Hadoop enterprise licensed version is quite fine tuned and easy to use makes it good choice for Hadoop administrators. It’s scalability and integration with Kerberos is good option for authentication and authorisation. installation can be improved. logging can be improved so that it become easier for debugging purposes. parallel processing of data is achieved easily.
Postgresql is the best tool out there for relational data so I have to give it a high rating when it comes to analytics, data availability and consistency, so on and so forth. SQL is also a relatively consistent language so when it comes to building new tables and loading data in from the OLTP database, there are enough tools where we can perform ETL on a scalable basis.
It is very useful solution which provides you speedier data processing, real-time analytics. It helps you manage diverse data types. It also offers you excellent disaster management. It has user friendly interface which helps you navigate system and transactions easily and perform task smoothly.
The data queries are relatively quick for a small to medium sized table. With complex joins, and a wide and deep table however, the performance of the query has room for improvement.
It's a great value for what you pay, and most Data Base Administrators (DBAs) can walk in and use it without substantial training. I tend to dabble on the analyst side, so querying the data I need feels like it can take forever, especially on higher traffic days like Monday.
There are several companies that you can contract for technical support, like EnterpriseDB or Percona, both first level in expertise and commitment to the software.
But we do not have contracts with them, we have done all the way from googling to forums, and never have a problem that we cannot resolve or pass around. And for dozens of projects and more than 15 years now.
However, I am not the right person to answer this as we have another department to handle support and contact the service provider for any support required. Although i will say that they are the quick respondent and knows how to handle querry of the customers and provide quick and better support.
The online training is request based. Had there been recorded videos available online for potential users to benefit from, I could have rated it higher. The online documentation however is very helpful. The online documentation PDF is downloadable and allows users to pace their own learning. With examples and code snippets, the documentation is great starting point.
Professional GIS people are some of the most risk-averse there are, and it's difficult to get them to move to HANA in one step. Start with small projects building to 80% use of HANA spatial over time.
Not used any other product than Hadoop and I don't think our company will switch to any other product, as Hadoop is providing excellent results. Our company is growing rapidly, Hadoop helps to keep up our performance and meet customer expectations. We also use HDFS which provides very high bandwidth to support MapReduce workloads.
Although the competition between the different databases is increasingly aggressive in the sense that they provide many improvements, new functionalities, compatibility with complementary components or environments, in some cases it requires that it be followed within the same family of applications that performs the company that develops it and that is not all bad, but being able to adapt or configure different programs, applications or other environments developed by third parties apart is what gives PostgreSQL a certain advantage and this diversification in the components that can be joined with it, is the reason why it is a great option to choose.
I have deep knowledge of other disk based DBMSs. They are venerable technology, but the attempts to extend them to current architectures belie the fact they are built on 40 year old technology. There are some good columnar in-memory databases but they lack the completeness of capability present in the HANA platform.
There are many advantages of Hadoop as first it has made the management and processing of extremely colossal data very easy and has simplified the lives of so many people including me.
Hadoop is quite interesting due to its new and improved features plus innovative functions.
Easy to administer so our DevOps team has only ever used minimal time to setup, tune, and maintain.
Easy to interface with so our Engineering team has only ever used minimal time to query or modify the database. Getting the data is straightforward, what we do with it is the bigger concern.