Apache Airflow is an open source tool that can be used to programmatically author, schedule and monitor data pipelines using Python and SQL.
N/A
SnapLogic
Score 8.6 out of 10
N/A
SnapLogic is a cloud integration platform with a self-service capacity supported by over 450 prebuilt modifiable connectors. SnapLogic also offers real-time and batch integration processes for interfacing with external data sources, a drag-and-drop interface, and use of the vendors’ Iris AI.
Other providers found it difficult to allow us to use their services on our cloud premises (exclusively in our AWS accounts) which we need for compliance. SnapLogic was able to give us this guarantee.
Airflow is well-suited for data engineering pipelines, creating scheduled workflows, and working with various data sources. You can implement almost any kind of DAG for any use case using the different operators or enforce your operator using the Python operator with ease. The MLOps feature of Airflow can be enhanced to match MLFlow-like features, making Airflow the go-to solution for all workloads, from data science to data engineering.
Snaplogic is unique from other IPASS tools if you're very sensitive about data security as they have an on-premise option where your data never needs to leave your data center. And data pipelines can be quickly created if Snaplogic has the requisite connector to your data sources. On the downside, if you're transforming a large amount of data for example in training machine learning models, a tool with elastic compute capability is more appropriate.
Apache Airflow is one of the best Orchestration platforms and a go-to scheduler for teams building a data platform or pipelines.
Apache Airflow supports multiple operators, such as the Databricks, Spark, and Python operators. All of these provide us with functionality to implement any business logic.
Apache Airflow is highly scalable, and we can run a large number of DAGs with ease. It provided HA and replication for workers. Maintaining airflow deployments is very easy, even for smaller teams, and we also get lots of metrics for observability.
UI/Dashboard can be updated to be customisable, and jobs summary in groups of errors/failures/success, instead of each job, so that a summary of errors can be used as a starting point for reviewing them.
Navigation - It's a bit dated. Could do with more modern web navigation UX. i.e. sidebars navigation instead of browser back/forward.
Again core functional reorg in terms of UX. Navigation can be improved for core functions as well, instead of discovery.
This has been hands down the BEST software company I have ever used and dealt with. I am a 25 year IT veteran at this college. They go above and beyond in soliciting our feedback/input and proactively follow up about bugs, issues, etc. I have given multiple potential clients my thoughts and after seeing the SL demo they all sign up. I appreciate their support model, it's REFRESHING!
For its capability to connect with multicloud environments. Access Control management is something that we don't get in all the schedulers and orchestrators. But although it provides so many flexibility and options to due to python , some level of knowledge of python is needed to be able to build workflows.
They can be prompt but they have not been as useful as I've wanted. We had a bug that affected many of our customers through an API connection between SnapLogic and our platform. Eventually they were able to figure it out, but it took a long time of negotiating between our engineering team and theirs. Additionally, we installed the SnapLogic groundplex for our customers and we've run into a bunch of problems of connectivity. If SnapLogic offered to be on those calls with our clients to troubleshoot how to fix these problems, I would give them a better grade here.
Multiple DAGs can be orchestrated simultaneously at varying times, and runs can be reproduced or replicated with relative ease. Overall, utilizing Apache Airflow is easier to use than other solutions now on the market. It is simple to integrate in Apache Airflow, and the workflow can be monitored and scheduling can be done quickly using Apache Airflow. We advocate using this tool for automating the data pipeline or process.
We opted for SnapLogic due its ease of use and the flexibility it offers, it was the platform that was strongest in both application integration and data integration and both were use cases we wanted to be able to cover.
Impact Depends on number of workflows. If there are lot of workflows then it has a better usecase as the implementation is justified as it needs resources , dedicated VMs, Database that has a cost