Apache Spark is a multi-language engine for executing data engineering, data science, and machine learning on single-node machines or clusters.
N/A
IBM DataStage
Score 7.7 out of 10
N/A
IBM® DataStage® is a data integration tool that helps users to design, develop and run jobs that move and transform data. At its core, the DataStage tool supports extract, transform and load (ETL) and extract, load and transform (ELT) patterns. A basic version of the software is available for on-premises deployment, and the cloud-based DataStage for IBM Cloud Pak® for Data offers automated integration capabilities in a hybrid or multicloud environment.
Well suited: To most of the local run of datasets and non-prod systems - scalability is not a problem at all. Including data from multiple types of data sources is an added advantage. MLlib is a decently nice built-in library that can be used for most of the ML tasks. Less appropriate: We had to work on a RecSys where the music dataset that we used was around 300+Gb in size. We faced memory-based issues. Few times we also got memory errors. Also the MLlib library does not have support for advanced analytics and deep-learning frameworks support. Understanding the internals of the working of Apache Spark for beginners is highly not possible.
DataStage is somewhat outdated for an ETL. I guess that's what makes it a bit lagged behind its competitors. It can be used for data processing, sure, but its performance seems to be lagging behind or quite slow given the server it is running from. I won’t depend on this application if it's handling a lot of mission-critical banking and business data.
Technical support is a key area IBM should improve for this product. Sometimes our case is assigned to a support engineer and he has no idea of the product or services.
Provide custom reports for datastage jobs and performance such as job history reports, warning messages or error messages.
Make it fully compatible with Oracle and users can direct use of Oracle ODBC drivers instead of Data Direct driver. Same for SQL server.
If the team looking to use Apache Spark is not used to debug and tweak settings for jobs to ensure maximum optimizations, it can be frustrating. However, the documentation and the support of the community on the internet can help resolve most issues. Moreover, it is highly configurable and it integrates with different tools (eg: it can be used by dbt core), which increase the scenarios where it can be used
Because it is robust, and it is being continuously improved. DS is one of the most used and recognized tools in the market. Large companies have implemented it in the first instance to develop their DW, but finding the advantages it has, they could use it for other types of projects such as migrations, application feeding, etc.
It could load thousands of records in seconds. But in the Parallel version, you need to understand how to particionate the data. If you use the algorithms erroneously, or the functionalities that it gives for the parsing of data, the performance can fall drastically, even with few records. It is necessary to have people with experience to be able to determine which algorithm to use and understand why.
1. It integrates very well with scala or python. 2. It's very easy to understand SQL interoperability. 3. Apache is way faster than the other competitive technologies. 4. The support from the Apache community is very huge for Spark. 5. Execution times are faster as compared to others. 6. There are a large number of forums available for Apache Spark. 7. The code availability for Apache Spark is simpler and easy to gain access to. 8. Many organizations use Apache Spark, so many solutions are available for existing applications.
IBM offers different levels of support but in my experience being and IBM shop helps to get direct support from more knowledgeable technicians from IBM. Not sure on the cost of having this kind of support, but I know there's also general support and community blogs and websites on the Internet make it easy to troubleshoot issues whenever there's need for that.
Spark in comparison to similar technologies ends up being a one stop shop. You can achieve so much with this one framework instead of having to stitch and weave multiple technologies from the Hadoop stack, all while getting incredibility performance, minimal boilerplate, and getting the ability to write your application in the language of your choosing.
With effective capabilities and easy to manipulate the features and easy to produce accurate data analytics and the Cloud services Automation, this IBM platform is more reliable and easy to document management. The features on this platform are equipped with excellent big data management and easy to provide accurate data analytics.
It’s hard to say at this point, it delivers, but not quite as I expected. It takes a lot of resources to manage and sort this out (manpower, financial).
Definitely, I don’t have the exact numbers, but given the data it processes, it is A LOT. So props to the developer of this application.
Again, based on my experience, I’d choose other ETL apps if there is one that's more user-friendly.