Apache Spark is a multi-language engine for executing data engineering, data science, and machine learning on single-node machines or clusters.
N/A
Presto
Score 10.0 out of 10
N/A
Presto is an open source SQL query engine designed to run queries on data stored in Hadoop or in traditional databases.
Teradata supported development of Presto followed the acquisition of Hadapt and Revelytix.
N/A
Talend Open Studio (discontinued)
Score 9.0 out of 10
N/A
Talend Open Studio was an open source integration software, used to build basic data pipelines or execute simple ETL and data integration tasks. Qlik and Talend discontinued the service in early 2024, and it is no longer available.
All the above systems work quite well on big data transformations whereas Spark really shines with its bigger API support and its ability to read from and write to multiple data sources. Using Spark one can easily switch between declarative versus imperative versus functional …
I think Presto is one of the best solutions out there today at the cutting edge for interactive query analysis. One of the challenges is presto is a niche tool for the interactive query use case and doesn't have the knobs and whistles as much as Spark. In the foreseeable future …
Well suited: To most of the local run of datasets and non-prod systems - scalability is not a problem at all. Including data from multiple types of data sources is an added advantage. MLlib is a decently nice built-in library that can be used for most of the ML tasks. Less appropriate: We had to work on a RecSys where the music dataset that we used was around 300+Gb in size. We faced memory-based issues. Few times we also got memory errors. Also the MLlib library does not have support for advanced analytics and deep-learning frameworks support. Understanding the internals of the working of Apache Spark for beginners is highly not possible.
Presto is for interactive simple queries, where Hive is for reliable processing. If you have a fact-dim join, presto is great..however for fact-fact joins presto is not the solution.. Presto is a great replacement for proprietary technology like Vertica
For quick daily integrations Talend is a very good tool and it makes development time so short and easy. Citizen developers who are not great programmers can pick up and start using Talend Open Studio within weeks. It's well suited for all kinds of data migration between various systems. It is less appropriate for smaller synchronous services where you need to trace the complete transaction and how data moved between them. It's also less appropriate for small data movements where other tools can be easier to use and manage.
Linking, embedding links and adding images is easy enough.
Once you have become familiar with the interface, Presto becomes very quick & easy to use (but, you have to practice & repeat to know what you are doing - it is not as intuitive as one would hope).
Organizing & design is fairly simple with click & drag parameters.
Presto was not designed for large fact fact joins. This is by design as presto does not leverage disk and used memory for processing which in turn makes it fast.. However, this is a tradeoff..in an ideal world, people would like to use one system for all their use cases, and presto should get exhaustive by solving this problem.
Resource allocation is not similar to YARN and presto has a priority queue based query resource allocation..so a query that takes long takes longer...this might be alleviated by giving some more control back to the user to define priority/override.
UDF Support is not available in presto. You will have to write your own functions..while this is good for performance, it comes at a huge overhead of building exclusively for presto and not being interoperable with other systems like Hive, SparkSQL etc.
The community is not that up to date and forum is not that great in response. Probably we should make people aware of the tool more on how to use and its implementations.
Talend crashes when transforming a lot of data (millions of rows).
Proper training documentation is a must for talend which is currently lagging. This will help users to learn more about Talend and use it effectively.
There is no licence requirement for Talend Open Studio. So, this is not relevant question. However, if you are asking whether we will use Talend in future. Yes. We will continue to use it. It's very powerful free tool which caters to all our extra, transform, load capabilities. We just love Talend for it's great functionality and ease of use.
If the team looking to use Apache Spark is not used to debug and tweak settings for jobs to ensure maximum optimizations, it can be frustrating. However, the documentation and the support of the community on the internet can help resolve most issues. Moreover, it is highly configurable and it integrates with different tools (eg: it can be used by dbt core), which increase the scenarios where it can be used
Talend Open Studio is based on Eclipse and is full of redundant procedures to do one thing, like when installing libraries. Sometimes I cannot manually download the libraries that it can't find.
Many times, Talend freezes. When you give a cancel command, it takes several minutes to stop. It also takes a great toll on our PC with 16 GB of ram and I7 CPU, even in idle status. If you are downloading Maven Jar/Libraries, you cannot do anything and have to wait until the task is finished.
1. It integrates very well with scala or python. 2. It's very easy to understand SQL interoperability. 3. Apache is way faster than the other competitive technologies. 4. The support from the Apache community is very huge for Spark. 5. Execution times are faster as compared to others. 6. There are a large number of forums available for Apache Spark. 7. The code availability for Apache Spark is simpler and easy to gain access to. 8. Many organizations use Apache Spark, so many solutions are available for existing applications.
Talend Open Studio is free and we are not using the enterprise version which comes with licence and support. So, mostly depend on the open source community for any issues that we face. The document is good and we didn't have to use any support so far. We did evaluate the enterprise version and so far sticking to the free version.
Spark in comparison to similar technologies ends up being a one stop shop. You can achieve so much with this one framework instead of having to stitch and weave multiple technologies from the Hadoop stack, all while getting incredibility performance, minimal boilerplate, and getting the ability to write your application in the language of your choosing.
Presto is good for a templated design appeal. You cannot be too creative via this interface - but, the layout and options make the finalized visual product appealing to customers. The other design products I use are for different purposes and not really comparable to Presto.
Informatica has a limited number of components that you can use. This places a heavy limitation on the capabilities of Informatica. On the other hand, Talend allows you to create your own custom components using Java. For businesses that need to perform a wide variety of data operations, it can be quite useful to have the option of creating your own custom components to satisfy business needs.
I delivered projects the client did not believe were possible, and I provided intermediate value by providing visibility to hidden data problems in their systems they could not detect before.
I was able to work 3 projects at a time, pausing gracefully in one while switching to the other, with minimal effort.