Apache Airflow is an open source tool that can be used to programmatically author, schedule and monitor data pipelines using Python and SQL.
N/A
AWS Batch
Score 7.8 out of 10
N/A
With AWS Batch, users package the code for batch jobs, specify dependencies, and submit batch jobs using the AWS Management Console, CLIs, or SDKs. AWS Batch allows users to specify execution parameters and job dependencies, and facilitates integration with a broad range of popular batch computing workflow engines and languages (e.g., Pegasus WMS, Luigi, Nextflow, Metaflow, Apache Airflow, and AWS Step Functions).
N/A
JAMS
Score 8.1 out of 10
N/A
JAMS is a centralized workload automation and job scheduling solution that runs, monitors, and manages jobs and workflows. Reliably orchestrate the critical IT processes that run your business from a single pane of glass.
$9,996
per year
Pricing
Apache Airflow
AWS Batch
JAMS
Editions & Modules
No answers on this topic
No answers on this topic
Core
9,996.00
per year
Advanced
Customized Pricing
per year
Offerings
Pricing Offerings
Apache Airflow
AWS Batch
JAMS
Free Trial
No
No
Yes
Free/Freemium Version
Yes
No
No
Premium Consulting/Integration Services
No
No
Yes
Entry-level Setup Fee
No setup fee
No setup fee
No setup fee
Additional Details
—
—
- Core: For small teams getting started with automation. Core Integrations: PowerShell, SQL, Azure Data Factory, Python, 20+ others.
- Advanced: Comprehensive solution for large-scale operations. Core Integrations: SAP, JDE, Ellucian Banner, Informatica, Mainframe and Power Systems.
Airflow is well-suited for data engineering pipelines, creating scheduled workflows, and working with various data sources. You can implement almost any kind of DAG for any use case using the different operators or enforce your operator using the Python operator with ease. The MLOps feature of Airflow can be enhanced to match MLFlow-like features, making Airflow the go-to solution for all workloads, from data science to data engineering.
More appropriate if you have a tech group that can use more of the AWS Batch rather than one or 2 things. It works great for me, but there was a huge learning curve the first week of using it. Now, I love it - and I hope to dig deep into other parts not just S3.
There's probably better schedulers out there. JAMS is good for an on-premises/classic IT implementation. JAMS is well-suited for use cases such as collecting a file from a shared location and uploading it to an API, running scripts on servers, and handling middleware tasks that process files or data between handoffs across different locations. JAMS is best suited to environments with IT staff who can develop, test, implement, maintain, and troubleshoot scripts; it does not use natural language processing (so it requires in-depth knowledge of the scripting language in use), and it does not appear to offer native dashboarding or reporting that is easily accessible to all users.
Apache Airflow is one of the best Orchestration platforms and a go-to scheduler for teams building a data platform or pipelines.
Apache Airflow supports multiple operators, such as the Databricks, Spark, and Python operators. All of these provide us with functionality to implement any business logic.
Apache Airflow is highly scalable, and we can run a large number of DAGs with ease. It provided HA and replication for workers. Maintaining airflow deployments is very easy, even for smaller teams, and we also get lots of metrics for observability.
UI/Dashboard can be updated to be customisable, and jobs summary in groups of errors/failures/success, instead of each job, so that a summary of errors can be used as a starting point for reviewing them.
Navigation - It's a bit dated. Could do with more modern web navigation UX. i.e. sidebars navigation instead of browser back/forward.
Again core functional reorg in terms of UX. Navigation can be improved for core functions as well, instead of discovery.
JAMS is a critical resource free up people to do other things and ensuring that processes and tasks are run consistently. We are also confident that procedures are run consistently and on time or as soon as the necessary data is available. With automated job failure notification, we are not required to check that jobs are running correctly.
For its capability to connect with multicloud environments. Access Control management is something that we don't get in all the schedulers and orchestrators. But although it provides so many flexibility and options to due to python , some level of knowledge of python is needed to be able to build workflows.
Key advantages include cost-effectiveness through dynamic resource provisioning and the use of spot instances. It auto-scales to meet workload demands, allowing easy job submission via the AWS Management Console or SDKs. It integrates seamlessly with other services like S3 and CloudWatch. It features automatic retries for failed jobs. It allows for a custom computing environment tailored to specific needs
JAMS is very user friendly; you hardly need to do coding. The only thing that I would say a challenge is setting it up, but that's because you barely know the product yet and, in every processing, setting up is the difficult part. But once you've set it up and you are going to use it, you will really feel that it is worth to invest in this kind of software solution, it really does it job very well.
We didnt really encounter any downtime, no issues encountered during 2 years of use of JAMs also our client barely raise an issue with JAMS, mostly the issues is on the batch jobs that jams executes. So I would gave it a perfect 10, very reliable hardly encounters any error and bug
JAMS performance is very great, there are no issues raised with the performance, it just like nothing happens on the job after integration it gives you this monitoring capability, no reports and bugs raised on the performance, we didnt do integration with other software only database and with use of JAMS agent to different servers
I've never had to wait more than a day for a response to any email queries submitted. We had a very positive experience using support hours during out migration process from v6 to v7. We've also recently had a weeklong group training course where all attendees were positive about the learning outcomes, a shoutout to Jose who did both the migration and the weeklong course!
People that were involved in the POC found the training a lot easier to follow. I think most people would have preferred to just get the training material and run through themselves.
I Was not part of the original Implementation, and the persons did that are no longer with the Organization. But I was part of the recent Upgrade process a year ago and I am the JAMS admin and was very pleased
Multiple DAGs can be orchestrated simultaneously at varying times, and runs can be reproduced or replicated with relative ease. Overall, utilizing Apache Airflow is easier to use than other solutions now on the market. It is simple to integrate in Apache Airflow, and the workflow can be monitored and scheduling can be done quickly using Apache Airflow. We advocate using this tool for automating the data pipeline or process.
JAMS is WAY more advanced, it isn't a fair comparison. The history is easy to get through. It is easy to get alerts of complete to failed and with a log. Adding jobs is extremely easy that even my teammates who do not manange the software are able to set them up. With the new web component we are very excited for the future of JAMS advancements.
The product is quite flexible. There are a number of features and functions that we use on a daily basis, and there are many features that are available that we have not yet needed or explored (like setting up jobs with the ability to do FTP or Sftp file transfers).
Impact Depends on number of workflows. If there are lot of workflows then it has a better usecase as the implementation is justified as it needs resources , dedicated VMs, Database that has a cost
Our company depends on the JAMS Scheduler in executing a large number of SAP Jobs. In never having had a true Enterprise Scheduler such as JAMS before now, we are very happy with the results.
There a number of features in JAMS for setting up schedules and dependencies on other jobs. This helps our company achieve the necessary workflows for Job execution. This optimization saves on system resources and keeps Jobs flowing smoothly.
We are very happy that JAMS is a robust solution with High Availability. This is necessary for Enterprise products, to reduce downtime, which we have not had as a result. This feature definitely saves our company money by reducing or eliminating unexpected downtime.